Archives For

shepherd-joannaTruth on the Market is delighted to welcome our newest blogger, Joanna Shepherd. Joanna is a Professor of Law at Emory School of Law and holds an adjunct position in the Emory Economics Department (where she also earned her PhD). At the law school she teaches Torts, Law and Economics, Analytical Methods for Lawyers, and Legal and Economic Issues in Health Policy. She also frequently teaches economics courses to law professors and federal and state judges.

Joanna is also a senior scholar at the International Center for Law and Economics.

Joanna’s research focuses on various law & econ topics. Her recent research has examined issues related to the healthcare and pharmaceutical industries, tort reform, litigation practice, and judicial behavior. Her works has appeared in the Michigan Law Review, Vanderbilt Law Review, Southern California Law Review, New York University Law ReviewDuke Law JournalUCLA Law Review, The Journal of Legal Studies, The Journal of Law & Economics, The American Law & Economics Review, Health Matrix, and The American Journal of Law & Medicine, among others. Joanna is also an author of the textbook, Economic Analysis for Lawyers, with Henry Butler and Christopher Drahozal. Her research has been discussed in numerous newspapers, including the Wall Street Journal and the New York Times, and has been cited by several courts including the Supreme Court.

You can find links to Joanna’s scholarship on her SSRN page.

Welcome Joanna!

Last week, FCC General Counsel Jonathan Sallet pulled back the curtain on the FCC staff’s analysis behind its decision to block Comcast’s acquisition of Time Warner Cable. As the FCC staff sets out on its reported Rainbow Tour to reassure regulated companies that it’s not “hostile to the industries it regulates,” Sallet’s remarks suggest it will have an uphill climb. Unfortunately, the staff’s analysis appears to have been unduly speculative, disconnected from critical market realities, and decidedly biased — not characteristics in a regulator that tend to offer much reassurance.

Merger analysis is inherently speculative, but, as courts have repeatedly had occasion to find, the FCC has a penchant for stretching speculation beyond the breaking point, adopting theories of harm that are vaguely possible, even if unlikely and inconsistent with past practice, and poorly supported by empirical evidence. The FCC’s approach here seems to fit this description.

The FCC’s fundamental theory of anticompetitive harm

To begin with, as he must, Sallet acknowledged that there was no direct competitive overlap in the areas served by Comcast and Time Warner Cable, and no consumer would have seen the number of providers available to her changed by the deal.

But the FCC staff viewed this critical fact as “not outcome determinative.” Instead, Sallet explained that the staff’s opposition was based primarily on a concern that the deal might enable Comcast to harm “nascent” OVD competitors in order to protect its video (MVPD) business:

Simply put, the core concern came down to whether the merged firm would have an increased incentive and ability to safeguard its integrated Pay TV business model and video revenues by limiting the ability of OVDs to compete effectively, especially through the use of new business models.

The justification for the concern boiled down to an assumption that the addition of TWC’s subscriber base would be sufficient to render an otherwise too-costly anticompetitive campaign against OVDs worthwhile:

Without the merger, a company taking action against OVDs for the benefit of the Pay TV system as a whole would incur costs but gain additional sales – or protect existing sales — only within its footprint. But the combined entity, having a larger footprint, would internalize more of the external “benefits” provided to other industry members.

The FCC theorized that, by acquiring a larger footprint, Comcast would gain enough bargaining power and leverage, as well as the means to profit from an exclusionary strategy, leading it to employ a range of harmful tactics — such as impairing the quality/speed of OVD streams, imposing data caps, limiting OVD access to TV-connected devices, imposing higher interconnection fees, and saddling OVDs with higher programming costs. It’s difficult to see how such conduct would be permitted under the FCC’s Open Internet Order/Title II regime, but, nevertheless, the staff apparently believed that Comcast would possess a powerful “toolkit” with which to harm OVDs post-transaction.

Comcast’s share of the MVPD market wouldn’t have changed enough to justify the FCC’s purported fears

First, the analysis turned on what Comcast could and would do if it were larger. But Comcast was already the largest ISP and MVPD (now second largest MVPD, post AT&T/DIRECTV) in the nation, and presumably it has approximately the same incentives and ability to disadvantage OVDs today.

In fact, there’s no reason to believe that the growth of Comcast’s MVPD business would cause any material change in its incentives with respect to OVDs. Whatever nefarious incentives the merger allegedly would have created by increasing Comcast’s share of the MVPD market (which is where the purported benefits in the FCC staff’s anticompetitive story would be realized), those incentives would be proportional to the size of increase in Comcast’s national MVPD market share — which, here, would be about eight percentage points: from 22% to under 30% of the national market.

It’s difficult to believe that Comcast would gain the wherewithal to engage in this costly strategy by adding such a relatively small fraction of the MVPD market (which would still leave other MVPDs serving fully 70% of the market to reap the purported benefits instead of Comcast), but wouldn’t have it at its current size – and there’s no evidence that it has ever employed such strategies with its current market share.

It bears highlighting that the D.C. Circuit has already twice rejected FCC efforts to impose a 30% market cap on MVPDs, based on the Commission’s inability to demonstrate that a greater-than-30% share would create competitive problems, especially given the highly dynamic nature of the MVPD market. In vacating the FCC’s most recent effort to do so in 2009, the D.C. Circuit was resolute in its condemnation of the agency, noting:

In sum, the Commission has failed to demonstrate that allowing a cable operator to serve more than 30% of all [MVPD] subscribers would threaten to reduce either competition or diversity in programming.

The extent of competition and the amount of available programming (including original programming distributed by OVDs themselves) has increased substantially since 2009; this makes the FCC’s competitive claims even less sustainable today.

It’s damning enough to the FCC’s case that there is no marketplace evidence of such conduct or its anticompetitive effects in today’s market. But it’s truly impossible to square the FCC’s assertions about Comcast’s anticompetitive incentives with the fact that, over the past decade, Comcast has made massive investments in broadband, steadily increased broadband speeds, and freely licensed its programming, among other things that have served to enhance OVDs’ long-term viability and growth. Chalk it up to the threat of regulatory intervention or corporate incompetence if you can’t believe that competition alone could be responsible for this largesse, but, whatever the reason, the FCC staff’s fears appear completely unfounded in a marketplace not significantly different than the landscape that would have existed post-merger.

OVDs aren’t vulnerable, and don’t need the FCC’s “help”

After describing the “new entrants” in the market — such unfamiliar and powerless players as Dish, Sony, HBO, and CBS — Sallet claimed that the staff was principally animated by the understanding that

Entrants are particularly vulnerable when competition is nascent. Thus, staff was particularly concerned that this transaction could damage competition in the video distribution industry.

Sallet’s description of OVDs makes them sound like struggling entrepreneurs working in garages. But, in fact, OVDs have radically reshaped the media business and wield enormous clout in the marketplace.

Netflix, for example, describes itself as “the world’s leading Internet television network with over 65 million members in over 50 countries.” New services like Sony Vue and Sling TV are affiliated with giant, well-established media conglomerates. And whatever new offerings emerge from the FCC-approved AT&T/DIRECTV merger will be as well-positioned as any in the market.

In fact, we already know that the concerns of the FCC are off-base because they are of a piece with the misguided assumptions that underlie the Chairman’s recent NPRM to rewrite the MVPD rules to “protect” just these sorts of companies. But the OVDs themselves — the ones with real money and their competitive futures on the line — don’t see the world the way the FCC does, and they’ve resolutely rejected the Chairman’s proposal. Notably, the proposed rules would “protect” these services from exactly the sort of conduct that Sallet claims would have been a consequence of the Comcast-TWC merger.

If they don’t want or need broad protection from such “harms” in the form of revised industry-wide rules, there is surely no justification for the FCC to throttle a merger based on speculation that the same conduct could conceivably arise in the future.

The realities of the broadband market post-merger wouldn’t have supported the FCC’s argument, either

While a larger Comcast might be in a position to realize more of the benefits from the exclusionary strategy Sallet described, it would also incur more of the costs — likely in direct proportion to the increased size of its subscriber base.

Think of it this way: To the extent that an MVPD can possibly constrain an OVD’s scope of distribution for programming, doing so also necessarily makes the MVPD’s own broadband offering less attractive, forcing it to incur a cost that would increase in proportion to the size of the distributor’s broadband market. In this case, as noted, Comcast would have gained MVPD subscribers — but it would have also gained broadband subscribers. In a world where cable is consistently losing video subscribers (as Sallet acknowledged), and where broadband offers higher margins and faster growth, it makes no economic sense that Comcast would have valued the trade-off the way the FCC claims it would have.

Moreover, in light of the existing conditions imposed on Comcast under the Comcast/NBCU merger order from 2011 (which last for a few more years) and the restrictions adopted in the Open Internet Order, Comcast’s ability to engage in the sort of exclusionary conduct described by Sallet would be severely limited, if not non-existent. Nor, of course, is there any guarantee that former or would-be OVD subscribers would choose to subscribe to, or pay more for, any MVPD in lieu of OVDs. Meanwhile, many of the relevant substitutes in the MVPD market (like AT&T and Verizon FiOS) also offer broadband services – thereby increasing the costs that would be incurred in the broadband market even more, as many subscribers would shift not only their MVPD, but also their broadband service, in response to Comcast degrading OVDs.

And speaking of the Open Internet Order — wasn’t that supposed to prevent ISPs like Comcast from acting on their alleged incentives to impede the quality of, or access to, edge providers like OVDs? Why is merger enforcement necessary to accomplish the same thing once Title II and the rest of the Open Internet Order are in place? And if the argument is that the Open Internet Order might be defeated, aside from the completely speculative nature of such a claim, why wouldn’t a merger condition that imposed the same constraints on Comcast – as was done in the Comcast/NBCU merger order by imposing the former net neutrality rules on Comcast – be perfectly sufficient?

While the FCC staff analysis accepted as true (again, contrary to current marketplace evidence) that a bigger Comcast would have more incentive to harm OVDs post-merger, it rejected arguments that there could be countervailing benefits to OVDs and others from this same increase in scale. Thus, things like incremental broadband investments and speed increases, a larger Wi-Fi network, and greater business services market competition – things that Comcast is already doing and would have done on a greater and more-accelerated scale in the acquired territories post-transaction – were deemed insufficient to outweigh the expected costs of the staff’s entirely speculative anticompetitive theory.

In reality, however, not only OVDs, but consumers – and especially TWC subscribers – would have benefitted from the merger by access to Comcast’s faster broadband speeds, its new investments, and its superior video offerings on the X1 platform, among other things. Many low-income families would have benefitted from expansion of Comcast’s Internet Essentials program, and many businesses would have benefited from the addition of a more effective competitor to the incumbent providers that currently dominate the business services market. Yet these and other verifiable benefits were given short shrift in the agency’s analysis because they “were viewed by staff as incapable of outweighing the potential harms.”

The assumptions underlying the FCC staff’s analysis of the broadband market are arbitrary and unsupportable

Sallet’s claim that the combined firm would have 60% of all high-speed broadband subscribers in the U.S. necessarily assumes a national broadband market measured at 25 Mbps or higher, which is a red herring.

The FCC has not explained why 25 Mbps is a meaningful benchmark for antitrust analysis. The FCC itself endorsed a 10 Mbps baseline for its Connect America fund last December, noting that over 70% of current broadband users subscribe to speeds less than 25 Mbps, even in areas where faster speeds are available. And streaming online video, the most oft-cited reason for needing high bandwidth, doesn’t require 25 Mbps: Netflix says that 5 Mbps is the most that’s required for an HD stream, and the same goes for Amazon (3.5 Mbps) and Hulu (1.5 Mbps).

What’s more, by choosing an arbitrary, faster speed to define the scope of the broadband market (in an effort to assert the non-competitiveness of the market, and thereby justify its broadband regulations), the agency has – without proper analysis or grounding, in my view – unjustifiably shrunk the size of the relevant market. But, as it happens, doing so also shrinks the size of the increase in “national market share” that the merger would have brought about.

Recall that the staff’s theory was premised on the idea that the merger would give Comcast control over enough of the broadband market that it could unilaterally impose costs on OVDs sufficient to impair their ability to reach or sustain minimum viable scale. But Comcast would have added only one percent of this invented “market” as a result of the merger. It strains credulity to assert that there could be any transaction-specific harm from an increase in market share equivalent to a rounding error.

In any case, basing its rejection of the merger on a manufactured 25 Mbps relevant market creates perverse incentives and will likely do far more to harm OVDs than realization of even the staff’s worst fears about the merger ever could have.

The FCC says it wants higher speeds, and it wants firms to invest in faster broadband. But here Comcast did just that, and then was punished for it. Rather than acknowledging Comcast’s ongoing broadband investments as strong indication that the FCC staff’s analysis might be on the wrong track, the FCC leadership simply sidestepped that inconvenient truth by redefining the market.

The lesson is that if you make your product too good, you’ll end up with an impermissibly high share of the market you create and be punished for it. This can’t possibly promote the public interest.

Furthermore, the staff’s analysis of competitive effects even in this ersatz market aren’t likely supportable. As noted, most subscribers access OVDs on connections that deliver content at speeds well below the invented 25 Mbps benchmark, and they pay the same prices for OVD subscriptions as subscribers who receive their content at 25 Mbps. Confronted with the choice to consume content at 25 Mbps or 10 Mbps (or less), the majority of consumers voluntarily opt for slower speeds — and they purchase service from Netflix and other OVDs in droves, nonetheless.

The upshot? Contrary to the implications on which the staff’s analysis rests, if Comcast were to somehow “degrade” OVD content on the 25 Mbps networks so that it was delivered with characteristics of video content delivered over a 10-Mbps network, real-world, observed consumer preferences suggest it wouldn’t harm OVDs’ access to consumers at all. This is especially true given that OVDs often have a global focus and reach (again, Netflix has 65 million subscribers in over 50 countries), making any claims that Comcast could successfully foreclose them from the relevant market even more suspect.

At the same time, while the staff apparently viewed the broadband alternatives as “limited,” the reality is that Comcast, as well as other broadband providers, are surrounded by capable competitors, including, among others, AT&T, Verizon, CenturyLink, Google Fiber, many advanced VDSL and fiber-based Internet service providers, and high-speed mobile wireless providers. The FCC understated the complex impact of this robust, dynamic, and ever-increasing competition, and its analysis entirely ignored rapidly growing mobile wireless broadband competition.

Finally, as noted, Sallet claimed that the staff determined that merger conditions would be insufficient to remedy its concerns, without any further explanation. Yet the Commission identified similar concerns about OVDs in both the Comcast/NBCUniversal and AT&T/DIRECTV transactions, and adopted remedies to address those concerns. We know the agency is capable of drafting behavioral conditions, and we know they have teeth, as demonstrated by prior FCC enforcement actions. It’s hard to understand why similar, adequate conditions could not have been fashioned for this transaction.

In the end, while I appreciate Sallet’s attempt to explain the FCC’s decision to reject the Comcast/TWC merger, based on the foregoing I’m not sure that Comcast could have made any argument or showing that would have dissuaded the FCC from challenging the merger. Comcast presented a strong economic analysis answering the staff’s concerns discussed above, all to no avail. It’s difficult to escape the conclusion that this was a politically-driven result, and not one rigorously based on the facts or marketplace reality.

As the organizer of this retrospective on Josh Wright’s tenure as FTC Commissioner, I have the (self-conferred) honor of closing out the symposium.

When Josh was confirmed I wrote that:

The FTC will benefit enormously from Josh’s expertise and his error cost approach to antitrust and consumer protection law will be a tremendous asset to the Commission — particularly as it delves further into the regulation of data and privacy. His work is rigorous, empirically grounded, and ever-mindful of the complexities of both business and regulation…. The Commissioners and staff at the FTC will surely… profit from his time there.

Whether others at the Commission have really learned from Josh is an open question, but there’s no doubt that Josh offered an enormous amount from which they could learn. As Tim Muris said, Josh “did not disappoint, having one of the most important and memorable tenures of any non-Chair” at the agency.

Within a month of his arrival at the Commission, in fact, Josh “laid down the cost-benefit-analysis gauntlet” in a little-noticed concurring statement regarding a proposed amendment to the Hart-Scott-Rodino Rules. The technical details of the proposed rule don’t matter for these purposes, but, as Josh noted in his statement, the situation intended to be avoided by the rule had never arisen:

The proposed rulemaking appears to be a solution in search of a problem. The Federal Register notice states that the proposed rules are necessary to prevent the FTC and DOJ from “expend[ing] scarce resources on hypothetical transactions.” Yet, I have not to date been presented with evidence that any of the over 68,000 transactions notified under the HSR rules have required Commission resources to be allocated to a truly hypothetical transaction.

What Josh asked for in his statement was not that the rule be scrapped, but simply that, before adopting the rule, the FTC weigh its costs and benefits.

As I noted at the time:

[I]t is the Commission’s responsibility to ensure that the rules it enacts will actually be beneficial (it is a consumer protection agency, after all). The staff, presumably, did a perfectly fine job writing the rule they were asked to write. Josh’s point is simply that it isn’t clear the rule should be adopted because it isn’t clear that the benefits of doing so would outweigh the costs.

As essentially everyone who has contributed to this symposium has noted, Josh was singularly focused on the rigorous application of the deceptively simple concept that the FTC should ensure that the benefits of any rule or enforcement action it adopts outweigh the costs. The rest, as they say, is commentary.

For Josh, this basic principle should permeate every aspect of the agency, and permeate the way it thinks about everything it does. Only an entirely new mindset can ensure that outcomes, from the most significant enforcement actions to the most trivial rule amendments, actually serve consumers.

While the FTC has a strong tradition of incorporating economic analysis in its antitrust decision-making, its record in using economics in other areas is decidedly mixed, as Berin points out. But even in competition policy, the Commission frequently uses economics — but it’s not clear it entirely understands economics. The approach that others have lauded Josh for is powerful, but it’s also subtle.

Inherent limitations on anyone’s knowledge about the future of technology, business and social norms caution skepticism, as regulators attempt to predict whether any given business conduct will, on net, improve or harm consumer welfare. In fact, a host of factors suggests that even the best-intentioned regulators tend toward overconfidence and the erroneous condemnation of novel conduct that benefits consumers in ways that are difficult for regulators to understand. Coase’s famous admonition in a 1972 paper has been quoted here before (frequently), but bears quoting again:

If an economist finds something – a business practice of one sort or another – that he does not understand, he looks for a monopoly explanation. And as in this field we are very ignorant, the number of ununderstandable practices tends to be very large, and the reliance on a monopoly explanation, frequent.

Simply “knowing” economics, and knowing that it is important to antitrust enforcement, aren’t enough. Reliance on economic formulae and theoretical models alone — to say nothing of “evidence-based” analysis that doesn’t or can’t differentiate between probative and prejudicial facts — doesn’t resolve the key limitations on regulatory decisionmaking that threaten consumer welfare, particularly when it comes to the modern, innovative economy.

As Josh and I have written:

[O]ur theoretical knowledge cannot yet confidently predict the direction of the impact of additional product market competition on innovation, much less the magnitude. Additionally, the multi-dimensional nature of competition implies that the magnitude of these impacts will be important as innovation and other forms of competition will frequently be inversely correlated as they relate to consumer welfare. Thus, weighing the magnitudes of opposing effects will be essential to most policy decisions relating to innovation. Again, at this stage, economic theory does not provide a reliable basis for predicting the conditions under which welfare gains associated with greater product market competition resulting from some regulatory intervention will outweigh losses associated with reduced innovation.

* * *

In sum, the theoretical and empirical literature reveals an undeniably complex interaction between product market competition, patent rules, innovation, and consumer welfare. While these complexities are well understood, in our view, their implications for the debate about the appropriate scale and form of regulation of innovation are not.

Along the most important dimensions, while our knowledge has expanded since 1972, the problem has not disappeared — and it may only have magnified. As Tim Muris noted in 2005,

[A] visitor from Mars who reads only the mathematical IO literature could mistakenly conclude that the U.S. economy is rife with monopoly power…. [Meanwhile, Section 2’s] history has mostly been one of mistaken enforcement.

It may not sound like much, but what is needed, what Josh brought to the agency, and what turns out to be absolutely essential to getting it right, is unflagging awareness of and attention to the institutional, political and microeconomic relationships that shape regulatory institutions and regulatory outcomes.

Regulators must do their best to constantly grapple with uncertainty, problems of operationalizing useful theory, and, perhaps most important, the social losses associated with error costs. It is not (just) technicians that the FTC needs; it’s regulators imbued with the “Economic Way of Thinking.” In short, what is needed, and what Josh brought to the Commission, is humility — the belief that, as Coase also wrote, sometimes the best answer is to “do nothing at all.”

The technocratic model of regulation is inconsistent with the regulatory humility required in the face of fast-changing, unexpected — and immeasurably valuable — technological advance. As Virginia Postrel warns in The Future and Its Enemies:

Technocrats are “for the future,” but only if someone is in charge of making it turn out according to plan. They greet every new idea with a “yes, but,” followed by legislation, regulation, and litigation…. By design, technocrats pick winners, establish standards, and impose a single set of values on the future.

For Josh, the first JD/Econ PhD appointed to the FTC,

economics provides a framework to organize the way I think about issues beyond analyzing the competitive effects in a particular case, including, for example, rulemaking, the various policy issues facing the Commission, and how I weigh evidence relative to the burdens of proof and production. Almost all the decisions I make as a Commissioner are made through the lens of economics and marginal analysis because that is the way I have been taught to think.

A representative example will serve to illuminate the distinction between merely using economics and evidence and understanding them — and their limitations.

In his Nielson/Arbitron dissent Josh wrote:

The Commission thus challenges the proposed transaction based upon what must be acknowledged as a novel theory—that is, that the merger will substantially lessen competition in a market that does not today exist.

[W]e… do not know how the market will evolve, what other potential competitors might exist, and whether and to what extent these competitors might impose competitive constraints upon the parties.

Josh’s straightforward statement of the basis for restraint stands in marked contrast to the majority’s decision to impose antitrust-based limits on economic activity that hasn’t even yet been contemplated. Such conduct is directly at odds with a sensible, evidence-based approach to enforcement, and the economic problems with it are considerable, as Josh also notes:

[I]t is an exceedingly difficult task to predict the competitive effects of a transaction where there is insufficient evidence to reliably answer the[] basic questions upon which proper merger analysis is based.

When the Commission’s antitrust analysis comes unmoored from such fact-based inquiry, tethered tightly to robust economic theory, there is a more significant risk that non-economic considerations, intuition, and policy preferences influence the outcome of cases.

Compare in this regard Josh’s words about Nielsen with Deborah Feinstein’s defense of the majority from such charges:

The Commission based its decision not on crystal-ball gazing about what might happen, but on evidence from the merging firms about what they were doing and from customers about their expectations of those development plans. From this fact-based analysis, the Commission concluded that each company could be considered a likely future entrant, and that the elimination of the future offering of one would likely result in a lessening of competition.

Instead of requiring rigorous economic analysis of the facts, couched in an acute awareness of our necessary ignorance about the future, for Feinstein the FTC fulfilled its obligation in Nielsen by considering the “facts” alone (not economic evidence, mind you, but customer statements and expressions of intent by the parties) and then, at best, casually applying to them the simplistic, outdated structural presumption – the conclusion that increased concentration would lead inexorably to anticompetitive harm. Her implicit claim is that all the Commission needed to know about the future was what the parties thought about what they were doing and what (hardy disinterested) customers thought they were doing. This shouldn’t be nearly enough.

Worst of all, Nielsen was “decided” with a consent order. As Josh wrote, strongly reflecting the essential awareness of the broader institutional environment that he brought to the Commission:

[w]here the Commission has endorsed by way of consent a willingness to challenge transactions where it might not be able to meet its burden of proving harm to competition, and which therefore at best are competitively innocuous, the Commission’s actions may alter private parties’ behavior in a manner that does not enhance consumer welfare.

Obviously in this regard his successful effort to get the Commission to adopt a UMC enforcement policy statement is a most welcome development.

In short, Josh is to be applauded not because he brought economics to the Commission, but because he brought the economic way of thinking. Such a thing is entirely too rare in the modern administrative state. Josh’s tenure at the FTC was relatively short, but he used every moment of it to assiduously advance his singular, and essential, mission. And, to paraphrase the last line of the movie The Right Stuff (it helps to have the rousing film score playing in the background as you read this): “for a brief moment, [Josh Wright] became the greatest [regulator] anyone had ever seen.”

I would like to extend my thanks to everyone who participated in this symposium. The contributions here will stand as a fitting and lasting tribute to Josh and his legacy at the Commission. And, of course, I’d also like to thank Josh for a tenure at the FTC very much worth honoring.

Tomorrow (August 24, 2015) marks once and future TOTM’er Josh Wright’s last day as an FTC Commissioner. Starting tomorrow and continuing throughout the week, Truth on the Market will be hosting a symposium —  a collection of commentaries and contributions — honoring Josh’s tenure at the FTC. We’ve invited contributions from a range of luminaries, including academics, practitioners, former FTC officials, and the like. Watch this space for the contributions, and feel free to add your own thoughts in the comments to the posts. Links to the posts will be collected here.

Monday’s posts will commence with contributions from

  • Richard Epstein,
  • Jon Jacobson,
  • Tom Hazlett, and
  • Keith Hylton

— with many more to come!


Today, for the first time in its 100-year history, the FTC issued enforcement guidelines for cases brought by the agency under the Unfair Methods of Competition (“UMC”) provisions of Section 5 of the FTC Act.

The Statement of Enforcement Principles represents a significant victory for Commissioner Joshua Wright, who has been a tireless advocate for defining and limiting the scope of the Commission’s UMC authority since before his appointment to the FTC in 2013.

As we’ve noted many times before here at TOTM (including in our UMC Guidelines Blog Symposium), FTC enforcement principles for UMC actions have been in desperate need of clarification. Without any UMC standards, the FTC has been free to leverage its costly adjudication process into settlements (or short-term victories) and businesses have been left in the dark as to what what sorts of conduct might trigger enforcement. Through a series of unadjudicated settlements, UMC unfairness doctrine (such as it is) has remained largely within the province of FTC discretion and without judicial oversight. As a result, and either by design or by accident, UMC never developed a body of law encompassing well-defined goals or principles like antitrust’s consumer welfare standard.

Commissioner Wright has long been at the forefront of the battle to rein in the FTC’s discretion in this area and to promote the rule of law. Soon after joining the Commission, he called for Section 5 guidelines that would constrain UMC enforcement to further consumer welfare, tied to the economically informed analysis of competitive effects developed in antitrust law.

Today’s UMC Statement embodies the essential elements of Commissioner Wright’s proposal. Under the new guidelines:

  1. The Commission will make UMC enforcement decisions based on traditional antitrust principles, including the consumer welfare standard;
  2. Only conduct that would violate the antitrust rule of reason will give rise to enforcement, and the Commission will not bring UMC cases without evidence demonstrating that harm to competition outweighs any efficiency or business justifications for the conduct at issue; and
  3. The Commission commits to the principle that it is more appropriate to bring cases under the antitrust laws than under Section 5 when the conduct at issue could give rise to a cause of action under the antitrust laws. Notably, this doesn’t mean that the agency gets to use UMC when it thinks it might lose under the Sherman or Clayton Acts; rather, it means UMC is meant only to be a gap-filler, to be used when the antitrust statutes don’t apply at all.

Yes, the Statement is a compromise. For instance, there is no safe harbor from UMC enforcement if any cognizable efficiencies are demonstrated, as Commissioner Wright initially proposed.

But by enshrining antitrust law’s consumer welfare standard in future UMC caselaw, by obligating the Commission to assess conduct within the framework of the well-established antitrust rule of reason, and by prioritizing antitrust over UMC when both might apply, the Statement brings UMC law into the world of modern antitrust analysis. This is a huge achievement.

It’s also a huge achievement that a Statement like this one would be introduced by Chairwoman Ramirez. As recently as last year, Ramirez had resisted efforts to impose constraints on the FTC’s UMC enforcement discretion. In a 2014 speech Ramirez said:

I have expressed concern about recent proposals to formulate guidance to try to codify our unfair methods principles for the first time in the Commission’s 100 year history. While I don’t object to guidance in theory, I am less interested in prescribing our future enforcement actions than in describing our broad enforcement principles revealed in our recent precedent.

The “recent precedent” that Ramirez referred to is precisely the set of cases applying UMC to reach antitrust-relevant conduct that led to Commissioner Wright’s efforts. The common law of consent decrees that make up the precedent Ramirez refers to, of course, are not legally binding and provide little more than regurgitated causes of action.

But today, under Congressional pressure and pressure from within the agency led by Commissioner Wright, Chairwoman Ramirez and the other two Democratic commissioners voted for the Statement.

Competitive Effects Analysis Under the Statement

As Commissioner Ohlhausen argues in her dissenting statement, the UMC Statement doesn’t remove all enforcement discretion from the Commission — after all, enforcement principles, like standards in law generally, have fuzzy boundaries.

But what Commissioner Ohlhausen seems to miss is that, by invoking antitrust principles, the rule of reason and competitive effects analysis, the Statement incorporates by reference 125 years of antitrust law and economics. The Statement itself need not go into excessive detail when, with only a few words, it brings modern antitrust jurisprudence embodied in cases like Trinko, Leegin, and Brooke Group into UMC law.

Under the new rule of reason approach for UMC, the FTC will condemn conduct only when it causes or is likely to cause “harm to competition or the competitive process, taking into account any associated cognizable efficiencies and business justifications.” In other words, the evidence must demonstrate net harm to consumers before the FTC can take action. That’s a significant constraint.

As noted above, Commissioner Wright originally proposed a safe harbor from FTC UMC enforcement whenever cognizable efficiencies are present. The Statement’s balancing test is thus a compromise. But it’s not really a big move from Commissioner Wright’s initial position.

Commissioner Wright’s original proposal tied the safe harbor to “cognizable” efficiencies, which is an exacting standard. As Commissioner Wright noted in his Blog Symposium post on the subject:

[T]he efficiencies screen I offer intentionally leverages the Commission’s considerable expertise in identifying the presence of cognizable efficiencies in the merger context and explicitly ties the analysis to the well-developed framework offered in the Horizontal Merger Guidelines. As any antitrust practitioner can attest, the Commission does not credit “cognizable efficiencies” lightly and requires a rigorous showing that the claimed efficiencies are merger-specific, verifiable, and not derived from an anticompetitive reduction in output or service. Fears that the efficiencies screen in the Section 5 context would immunize patently anticompetitive conduct because a firm nakedly asserts cost savings arising from the conduct without evidence supporting its claim are unwarranted. Under this strict standard, the FTC would almost certainly have no trouble demonstrating no cognizable efficiencies exist in Dan’s “blowing up of the competitor’s factory” example because the very act of sabotage amounts to an anticompetitive reduction in output.

The difference between the safe harbor approach and the balancing approach embodied in the Statement is largely a function of administrative economy. Before, the proposal would have caused the FTC to err on the side of false negatives, possibly forbearing from bringing some number of welfare-enhancing cases in exchange for a more certain reduction in false positives. Now, there is greater chance of false positives.

But the real effect is that more cases will be litigated because, in the end, both versions would require some degree of antitrust-like competitive effects analysis. Under the Statement, if procompetitive efficiencies outweigh anticompetitive harms, the defendant still wins (and the FTC is to avoid enforcement). Under the original proposal fewer actions might be brought, but those that are brought would surely settle. So one likely outcome of choosing a balancing test over the safe harbor is that more close cases will go to court to be sorted out. Whether this is a net improvement over the safe harbor depends on whether the social costs of increased litigation and error are offset by a reduction in false negatives — as well as the more robust development of the public good of legal case law.  

Reduced FTC Discretion Under the Statement

The other important benefit of the Statement is that it commits the FTC to a regime that reduces its discretion.

Chairwoman Ramirez and former Chairman Leibowitz — among others — have embraced a broader role for Section 5, particularly in order to avoid the judicial limits on antitrust actions arising out of recent Supreme Court cases like Trinko, Leegin, Brooke Group, Linkline, Weyerhaeuser and Credit Suisse.

For instance, as former Chairman Leibowitz said in 2008:

[T]he Commission should not be tied to the more technical definitions of consumer harm that limit applications of the Sherman Act when we are looking at pure Section 5 violations.

And this was no idle threat. Recent FTC cases, including Intel, N-Data, Google (Motorola), and Bosch, could all have been brought under the Sherman Act, but were brought — and settled — as Section 5 cases instead. Under the new Statement, all four would likely be Sherman Act cases.

There’s little doubt that, left unfettered, Section 5 UMC actions would only have grown in scope. Former Chairman Leibowitz, in his concurring opinion in Rambus, described UMC as

a flexible and powerful Congressional mandate to protect competition from unreasonable restraints, whether long-since recognized or newly discovered, that violate the antitrust laws, constitute incipient violations of those laws, or contravene those laws’ fundamental policies.

Both Leibowitz and former Commissioner Tom Rosch (again, among others) often repeated their views that Section 5 permitted much the same actions as were available under Section 2 — but without the annoyance of those pesky, economically sensible, judicial limitations. (Although, in fairness, Leibowitz also once commented that it would not “be wise to use the broader [Section 5] authority whenever we think we can’t win an antitrust case, as a sort of ‘fallback.’”)

In fact, there is a long and unfortunate trend of FTC commissioners and other officials asserting some sort of “public enforcement exception” to the judicial limits on Sherman Act cases. As then Deputy Director for Antitrust in the Bureau of Economics, Howard Shelanski, told Congress in 2010:

The Commission believes that its authority to prevent “unfair methods of competition” through Section 5 of the Federal Trade Commission Act enables the agency to pursue conduct that it cannot reach under the Sherman Act, and thus avoid the potential strictures of Trinko.

In this instance, and from the context (followed as it is by a request for Congress to actually exempt the agency from Trinko and Credit Suisse!), it seems that “reach” means “win.”

Still others have gone even further. Tom Rosch, for example, has suggested that the FTC should challenge Patent Assertion Entities under Section 5 merely because “we have a gut feeling” that the conduct violates the Act and it may not be actionable under Section 2.

Even more egregious, Steve Salop and Jon Baker advocate using Section 5 to implement their preferred social policies — in this case to reduce income inequality. Such expansionist views, as Joe Sims recently reminded TOTM readers, hearken back to the troubled FTC of the 1970s:  

Remember [former FTC Chairman] Mike Pertschuck saying that Section 5 could possibly be used to enforce compliance with desirable energy policies or environmental requirements, or to attack actions that, in the opinion of the FTC majority, impeded desirable employment programs or were inconsistent with the nation’s “democratic, political and social ideals.” The two speeches he delivered on this subject in 1977 were the beginning of the end for increased Section 5 enforcement in that era, since virtually everyone who heard or read them said:  “Whoa! Is this really what we want the FTC to be doing?”

Apparently, for some, it is — even today. But don’t forget: This was the era in which Congress actually briefly shuttered the FTC for refusing to recognize limits on its discretion, as Howard Beales reminds us:

The breadth, overreaching, and lack of focus in the FTC’s ambitious rulemaking agenda outraged many in business, Congress, and the media. Even the Washington Post editorialized that the FTC had become the “National Nanny.” Most significantly, these concerns reverberated in Congress. At one point, Congress refused to provide the necessary funding, and simply shut down the FTC for several days…. So great were the concerns that Congress did not reauthorize the FTC for fourteen years. Thus chastened, the Commission abandoned most of its rulemaking initiatives, and began to re-examine unfairness to develop a focused, injury-based test to evaluate practices that were allegedly unfair.

A truly significant effect of the Policy Statement will be to neutralize the effort to use UMC to make an end-run around antitrust jurisprudence in order to pursue non-economic goals. It will now be a necessary condition of a UMC enforcement action to prove a contravention of fundamental antitrust policies (i.e., consumer welfare), rather than whatever three commissioners happen to agree is a desirable goal. And the Statement puts the brakes on efforts to pursue antitrust cases under Section 5 by expressing a clear policy preference at the FTC to bring such cases under the antitrust laws.

Commissioner Ohlhausen’s objects that

the fact that this policy statement requires some harm to competition does little to constrain the Commission, as every Section 5 theory pursued in the last 45 years, no matter how controversial or convoluted, can be and has been couched in terms of protecting competition and/or consumers.

That may be true, but the same could be said of every Section 2 case, as well. Commissioner Ohlhausen seems to be dismissing the fact that the Statement effectively incorporates by reference the last 45 years of antitrust law, too. Nothing will incentivize enforcement targets to challenge the FTC in court — or incentivize the FTC itself to forbear from enforcement — like the ability to argue Trinko, Leegin and their ilk. Antitrust law isn’t perfect, of course, but making UMC law coextensive with modern antitrust law is about as much as we could ever reasonably hope for. And the Statement basically just gave UMC defendants blanket license to add a string of “See Areeda & Hovenkamp” cites to every case the FTC brings. We should count that as a huge win.

Commissioner Ohlhausen also laments the brevity and purported vagueness of the Statement, claiming that

No interpretation of the policy statement by a single Commissioner, no matter how thoughtful, will bind this or any future Commission to greater limits on Section 5 UMC enforcement than what is in this exceedingly brief, highly general statement.

But, in the end, it isn’t necessarily the Commissioners’ self-restraint upon which the Statement relies; it’s the courts’ (and defendants’) ability to take the obvious implications of the Statement seriously and read current antitrust precedent into future UMC cases. If every future UMC case is adjudicated like a Sherman or Clayton Act case, the Statement will have been a resounding success.

Arguably no FTC commissioner has been as successful in influencing FTC policy as a minority commissioner — over sustained opposition, and in a way that constrains the agency so significantly — as has Commissioner Wright today.

Yesterday, the International Center for Law & Economics, together with Professor Gus Hurwitz, Nebraska College of Law, and nine other scholars of law and economics, filed an amicus brief in the DC Circuit explaining why the court should vacate the FCC’s 2015 Open Internet Order.

A few key points from ICLE’s brief follow, but you can read a longer summary of the brief here.

If the 2010 Order was a limited incursion into neighboring territory, the 2015 Order represents the outright colonization of a foreign land, extending FCC control over the Internet far beyond what the Telecommunications Act authorizes.

The Commission asserts vast powers — powers that Congress never gave it — not just over broadband but also over the very ‘edge’ providers it claims to be protecting. The court should be very skeptical of the FCC’s claims to pervasive powers over the Internet.

In the 2015 Order, the FCC Invoked Title II, admitted that it was unworkable for the Internet, and then tried to ‘tailor’ the statute to avoid its worst excesses.

That the FCC felt the need for such sweeping forbearance should have indicated to it that it had ‘taken an interpretive wrong turn’ in understanding the statute Congress gave it. Last year, the Supreme Court blocked a similar attempt by the EPA to ‘modernize’ old legislation in a way that gave it expansive new powers. In its landmark UARG decision, the Court made clear that it won’t allow regulatory agencies to rewrite legislation in an effort to retrofit their statutes to their preferred regulatory regimes.

Internet regulation is a question of ‘vast economic and political significance,’ yet the FCC  didn’t even bother to weigh the costs and benefits of its rule. 

FCC Chairman Tom Wheeler never misses an opportunity to talk about the the Internet as ‘the most important network known to Man.’ So why did he and the previous FCC Chairman ignore requests from other commissioners for serious, independent economic analysis of the supposed problem and the best way to address it? Why did the FCC rush to adopt a plan that had the effect of blocking the Federal Trade Commission from applying its consumer protection laws to the Internet? For all the FCC’s talk about protecting consumers, it appears that its real agenda may be simply expanding its own power.

Joining ICLE on the brief are:

  • Richard Epstein (NYU Law)
  • James Huffman (Lewis & Clark Law)
  • Gus Hurwitz (Nebraska Law)
  • Thom Lambert (Missouri Law)
  • Daniel Lyons (Boston College Law)
  • Geoffrey Manne (ICLE)
  • Randy May (Free State Foundation)
  • Jeremy Rabkin (GMU Law)
  • Ronald Rotunda (Chapman Law)
  • Ilya Somin (GMU Law)

Read the brief here, and the summary here.

Read more of ICLE’s work on net neutrality and Title II, including:

  • Highlights from policy and legal comments filed by ICLE and TechFreedom on net neutrality
  • “Regulating the Most Powerful Network Ever,” a scholarly essay by Gus Hurwitz for the Free State Foundation
  • “How to Break the Internet,” an essay by Geoffrey Manne and Ben Sperry, in Reason Magazine
  • “The FCC’s Net Neutrality Victory is Anything But,” an op-ed by Geoffrey Manne, in Wired
  • “The Feds Lost on Net Neutrality, But Won Control of the Internet,” an op-ed by Geoffrey Manne and Berin Szoka in Wired
  • “Net Neutrality’s Hollow Promise to Startups,” an op-ed by Geoffrey Manne and Berin Szoka in Computerworld
  • Letter signed by 32 scholars urging the FTC to caution the FCC against adopting per se net neutrality rules by reclassifying ISPs under Title II
  • The FCC’s Open Internet Roundtables, Policy Approaches, Panel 3, Enhancing Transparency, with Geoffrey Manne​

On Thursday I will be participating in an ABA panel discussion on the Apple e-books case, along with Mark Ryan (former DOJ attorney) and Fiona Scott-Morton (former DOJ economist), both of whom were key members of the DOJ team that brought the case. Details are below. Judging from the prep call, it should be a spirited discussion!

Readers looking for background on the case (as well as my own views — decidedly in opposition to those of the DOJ) can find my previous commentary on the case and some of the issues involved here:

Other TOTM authors have also weighed in. See, e.g.:


ABA Section of Antitrust Law

Federal Civil abaantitrustEnforcement Committee, Joint Conduct, Unilateral Conduct, and Media & Tech Committees Present:

“The 2d Cir.’s Apple E-Books decision: Debating the merits and the meaning”

July 16, 2015
12:00 noon to 1:30 pm Eastern / 9:00 am to 10:30 am Pacific

On June 30, the Second Circuit affirmed DOJ’s trial victory over Apple in the Ebooks Case. The three-judge panel fractured in an interesting way: two judges affirmed the finding that Apple’s role in a “hub and spokes” conspiracy was unlawful per se; one judge also would have found a rule-of-reason violation; and the dissent — stating Apple had a “vertical” position and was challenging the leading seller’s “monopoly” — would have found no liability at all. What is the reasoning and precedent of the decision? Is “marketplace vigilantism” (the concurring judge’s phrase) ever justified? Our panel — which includes the former DOJ head of litigation involved in the case — will debate the issues.


  • Ken Ewing, Steptoe & Johnson LLP


  • Geoff Manne, International Center for Law & Economics
  • Fiona Scott Morton, Yale School of Management
  • Mark Ryan, Mayer Brown LLP

Register HERE

The FTC recently required divestitures in two merger investigations (here and here), based largely on the majority’s conclusion that

[when] a proposed merger significantly increases concentration in an already highly concentrated market, a presumption of competitive harm is justified under both the Guidelines and well-established case law.” (Emphasis added).

Commissioner Wright dissented in both matters (here and here), contending that

[the majority’s] reliance upon such shorthand structural presumptions untethered from empirical evidence subsidize a shift away from the more rigorous and reliable economic tools embraced by the Merger Guidelines in favor of convenient but obsolete and less reliable economic analysis.

Josh has the better argument, of course. In both cases the majority relied upon its structural presumption rather than actual economic evidence to make out its case. But as Josh notes in his dissent in In the Matter of ZF Friedrichshafen and TRW Automotive (quoting his 2013 dissent in In the Matter of Fidelity National Financial, Inc. and Lender Processing Services):

there is no basis in modern economics to conclude with any modicum of reliability that increased concentration—without more—will increase post-merger incentives to coordinate. Thus, the Merger Guidelines require the federal antitrust agencies to develop additional evidence that supports the theory of coordination and, in particular, an inference that the merger increases incentives to coordinate.

Or as he points out in his dissent in In the Matter of Holcim Ltd. and Lafarge S.A.

The unifying theme of the unilateral effects analysis contemplated by the Merger Guidelines is that a particularized showing that post-merger competitive constraints are weakened or eliminated by the merger is superior to relying solely upon inferences of competitive effects drawn from changes in market structure.

It is unobjectionable (and uninteresting) that increased concentration may, all else equal, make coordination easier, or enhance unilateral effects in the case of merger to monopoly. There are even cases (as in generic pharmaceutical markets) where rigorous, targeted research exists, sufficient to support a presumption that a reduction in the number of firms would likely lessen competition. But generally (as in these cases), absent actual evidence, market shares might be helpful as an initial screen (and may suggest greater need for a thorough investigation), but they are not analytically probative in themselves. As Josh notes in his TRW dissent:

The relevant question is not whether the number of firms matters but how much it matters.

The majority in these cases asserts that it did find evidence sufficient to support its conclusions, but — and this is where the rubber meets the road — the question remains whether its limited evidentiary claims are sufficient, particularly given analyses that repeatedly come back to the structural presumption. As Josh says in his Holcim dissent:

it is my view that the investigation failed to adduce particularized evidence to elevate the anticipated likelihood of competitive effects from “possible” to “likely” under any of these theories. Without this necessary evidence, the only remaining factual basis upon which the Commission rests its decision is the fact that the merger will reduce the number of competitors from four to three or three to two. This is simply not enough evidence to support a reason to believe the proposed transaction will violate the Clayton Act in these Relevant Markets.

Looking at the majority’s statements, I see a few references to the kinds of market characteristics that could indicate competitive concerns — but very little actual analysis of whether these characteristics are sufficient to meet the Clayton Act standard in these particular markets. The question is — how much analysis is enough? I agree with Josh that the answer must be “more than is offered here,” but it’s an important question to explore more deeply.

Presumably that’s exactly what the ABA’s upcoming program will do, and I highly recommend interested readers attend or listen in. The program details are below.

The Use of Structural Presumptions in Merger Analysis

June 26, 2015, 12:00 PM – 1:15 PM ET


  • Brendan Coffman, Wilson Sonsini Goodrich & Rosati LLP


  • Angela Diveley, Office of Commissioner Joshua D. Wright, Federal Trade Commission
  • Abbott (Tad) Lipsky, Latham & Watkins LLP
  • Janusz Ordover, Compass Lexecon
  • Henry Su, Office of Chairwoman Edith Ramirez, Federal Trade Commission

In-person location:

Latham & Watkins
555 11th Street,NW
Ste 1000
Washington, DC 20004

Register here.

Profile-Pic-3-professional-200x300Truth On the Market is pleased to announce that Kristian Stout of the International Center for Law and Economics (“ICLE”) has joined our team of writers. Kristian was recently hired by ICLE as Associate Director for Innovation Policy, bringing with him over ten years of experience as a technology professional and entrepreneur. In his role at ICLE, Kristian’s work is focused on the areas of Innovation, Data, Privacy, Telecom, and Intellectual Property.

Kristian has previously been a lecturer in the computer science department of Rutgers University,  is frequently invited to speak on law and technology topics, and has been published in law journals and legal treatises on intellectual property and innovation policy. Kristian is an attorney licensed to practice law in New Jersey and Pennsylvania, is a partner at A&S Technologies, a software services firm, and sits on the board of CodedByKids, a nonprofit organization that provides STEM education to underprivileged children.

Kristian graduated magna cum laude from the Rutgers University School of law, served on the editorial board of the Rutgers Journal of Law and Public Policy, and was awarded a Governor’s Executive Fellowship from the Eagleton Institute of Politics.

He is excited to join the TOTM team, bringing with him a fusion of technological-optimism and a belief in the power of free markets to enhance the welfare of all humanity.

Recently, Commissioner Pai praised the introduction of bipartisan legislation to protect joint sales agreements (“JSAs”) between local television stations. He explained that

JSAs are contractual agreements that allow broadcasters to cut down on costs by using the same advertising sales force. The efficiencies created by JSAs have helped broadcasters to offer services that benefit consumers, especially in smaller markets…. JSAs have served communities well and have promoted localism and diversity in broadcasting. Unfortunately, the FCC’s new restrictions on JSAs have already caused some stations to go off the air and other stations to carry less local news.

fccThe “new restrictions” to which Commissioner Pai refers were recently challenged in court by the National Association of Broadcasters (NAB), et. al., and on April 20, the International Center for Law & Economics and a group of law and economics scholars filed an amicus brief with the D.C. Circuit Court of Appeals in support of the petition, asking the court to review the FCC’s local media ownership duopoly rule restricting JSAs.

Much as it did with with net neutrality, the FCC is looking to extend another set of rules with no basis in sound economic theory or established facts.

At issue is the FCC’s decision both to retain the duopoly rule and to extend that rule to certain JSAs, all without completing a legally mandated review of the local media ownership rules, due since 2010 (but last completed in 2007).

The duopoly rule is at odds with sound competition policy because it fails to account for drastic changes in the media market that necessitate redefinition of the market for television advertising. Moreover, its extension will bring a halt to JSAs currently operating (and operating well) in nearly 100 markets.  As the evidence on the FCC rulemaking record shows, many of these JSAs offer public interest benefits and actually foster, rather than stifle, competition in broadcast television markets.

In the world of media mergers generally, competition law hasn’t yet caught up to the obvious truth that new media is competing with old media for eyeballs and advertising dollars in basically every marketplace.

For instance, the FTC has relied on very narrow market definitions to challenge newspaper mergers without recognizing competition from television and the Internet. Similarly, the generally accepted market in which Google’s search conduct has been investigated is something like “online search advertising” — a market definition that excludes traditional marketing channels, despite the fact that advertisers shift their spending between these channels on a regular basis.

But the FCC fares even worse here. The FCC’s duopoly rule is premised on an “eight voices” test for local broadcast stations regardless of the market shares of the merging stations. In other words, one entity cannot own FCC licenses to two or more TV stations in the same local market unless there are at least eight independently owned stations in that market, even if their combined share of the audience or of advertising are below the level that could conceivably give rise to any inference of market power.

Such a rule is completely unjustifiable under any sensible understanding of competition law.

Can you even imagine the FTC or DOJ bringing an 8 to 7 merger challenge in any marketplace? The rule is also inconsistent with the contemporary economic learning incorporated into the 2010 Merger Guidelines, which looks at competitive effects rather than just counting competitors.

Not only did the FCC fail to analyze the marketplace to understand how much competition there is between local broadcasters, cable, and online video, but, on top of that, the FCC applied this outdated duopoly rule to JSAs without considering their benefits.

The Commission offers no explanation as to why it now believes that extending the duopoly rule to JSAs, many of which it had previously approved, is suddenly necessary to protect competition or otherwise serve the public interest. Nor does the FCC cite any evidence to support its position. In fact, the record evidence actually points overwhelmingly in the opposite direction.

As a matter of sound regulatory practice, this is bad enough. But Congress directed the FCC in Section 202(h) of the Telecommunications Act of 1996 to review all of its local ownership rules every four years to determine whether they were still “necessary in the public interest as the result of competition,” and to repeal or modify those that weren’t. During this review, the FCC must examine the relevant data and articulate a satisfactory explanation for its decision.

So what did the Commission do? It announced that, instead of completing its statutorily mandated 2010 quadrennial review of its local ownership rules, it would roll that review into a new 2014 quadrennial review (which it has yet to perform). Meanwhile, the Commission decided to retain its duopoly rule pending completion of that review because it had “tentatively” concluded that it was still necessary.

In other words, the FCC hasn’t conducted its mandatory quadrennial review in more than seven years, and won’t, under the new rules, conduct one for another year and a half (at least). Oh, and, as if nothing of relevance has changed in the market since then, it “tentatively” maintains its already suspect duopoly rule in the meantime.

In short, because the FCC didn’t conduct the review mandated by statute, there is no factual support for the 2014 Order. By relying on the outdated findings from its earlier review, the 2014 Order fails to examine the significant changes both in competition policy and in the market for video programming that have occurred since the current form of the rule was first adopted, rendering the rulemaking arbitrary and capricious under well-established case law.

Had the FCC examined the record of the current rulemaking, it would have found substantial evidence that undermines, rather than supports, the FCC’s rule.

Economic studies have shown that JSAs can help small broadcasters compete more effectively with cable and online video in a world where their advertising revenues are drying up and where temporary economies of scale (through limited contractual arrangements like JSAs) can help smaller, local advertising outlets better implement giant, national advertising campaigns. A ban on JSAs will actually make it less likely that competition among local broadcasters can survive, not more.

OfficialPaiCommissioner Pai, in his dissenting statement to the 2014 Order, offered a number of examples of the benefits of JSAs (all of them studiously ignored by the Commission in its Order). In one of these, a JSA enabled two stations in Joplin, Missouri to use their $3.5 million of cost savings from a JSA to upgrade their Doppler radar system, which helped save lives when a devastating tornado hit the town in 2011. But such benefits figure nowhere in the FCC’s “analysis.”

Several econometric studies also provide empirical support for the (also neglected) contention that duopolies and JSAs enable stations to improve the quality and prices of their programming.

One study, by Jeff Eisenach and Kevin Caves, shows that stations operating under these agreements are likely to carry significantly more news, public affairs, and current affairs programming than other stations in their markets. The same study found an 11 percent increase in audience shares for stations acquired through a duopoly. Meanwhile, a study by Hal Singer and Kevin Caves shows that markets with JSAs have advertising prices that are, on average, roughly 16 percent lower than in non-duopoly markets — not higher, as would be expected if JSAs harmed competition.

And again, Commissioner Pai provides several examples of these benefits in his dissenting statement. In one of these, a JSA in Wichita, Kansas enabled one of the two stations to provide Spanish-language HD programming, including news, weather, emergency and community information, in a market where that Spanish-language programming had not previously been available. Again — benefit ignored.

Moreover, in retaining its duopoly rule on the basis of woefully outdated evidence, the FCC completely ignores the continuing evolution in the market for video programming.

In reality, competition from non-broadcast sources of programming has increased dramatically since 1999. Among other things:

  • VideoScreensToday, over 85 percent of American households watch TV over cable or satellite. Most households now have access to nearly 200 cable channels that compete with broadcast TV for programming content and viewers.
  • In 2014, these cable channels attracted twice as many viewers as broadcast channels.
  • Online video services such as Netflix, Amazon Prime, and Hulu have begun to emerge as major new competitors for video programming, leading 179,000 households to “cut the cord” and cancel their cable subscriptions in the third quarter of 2014 alone.
  • Today, 40 percent of U.S. households subscribe to an online streaming service; as a result, cable ratings among adults fell by nine percent in 2014.
  • At the end of 2007, when the FCC completed its last quadrennial review, the iPhone had just been introduced, and the launch of the iPad was still more than two years away. Today, two-thirds of Americans have a smartphone or tablet over which they can receive video content, using technology that didn’t even exist when the FCC last amended its duopoly rule.

In the face of this evidence, and without any contrary evidence of its own, the Commission’s action in reversing 25 years of agency practice and extending its duopoly rule to most JSAs is arbitrary and capricious.

The law is pretty clear that the extent of support adduced by the FCC in its 2014 Rule is insufficient. Among other relevant precedent (and there is a lot of it):

The Supreme Court has held that an agency

must examine the relevant data and articulate a satisfactory explanation for its action, including a rational connection between the facts found and the choice made.

In the DC Circuit:

the agency must explain why it decided to act as it did. The agency’s statement must be one of ‘reasoning’; it must not be just a ‘conclusion’; it must ‘articulate a satisfactory explanation’ for its action.


[A]n agency acts arbitrarily and capriciously when it abruptly departs from a position it previously held without satisfactorily explaining its reason for doing so.


The FCC ‘cannot silently depart from previous policies or ignore precedent’ . . . .”

And most recently in Judge Silberman’s concurrence/dissent in the 2010 Verizon v. FCC Open Internet Order case:

factual determinations that underly [sic] regulations must still be premised on demonstrated — and reasonable — evidential support

None of these standards is met in this case.

It will be noteworthy to see what the DC Circuit does with these arguments given the pending Petitions for Review of the latest Open Internet Order. There, too, the FCC acted without sufficient evidentiary support for its actions. The NAB/Stirk Holdings case may well turn out to be a bellwether for how the court views the FCC’s evidentiary failings in that case, as well.

The scholars joining ICLE on the brief are:

  • Babette E. Boliek, Associate Professor of Law, Pepperdine School of Law
  • Henry N. Butler, George Mason University Foundation Professor of Law and Executive Director of the Law & Economics Center, George Mason University School of Law (and newly appointed dean).
  • Richard Epstein, Laurence A. Tisch Professor of Law, Classical Liberal Institute, New York University School of Law
  • Stan Liebowitz, Ashbel Smith Professor of Economics, University of Texas at Dallas
  • Fred McChesney, de la Cruz-Mentschikoff Endowed Chair in Law and Economics, University of Miami School of Law
  • Paul H. Rubin, Samuel Candler Dobbs Professor of Economics, Emory University
  • Michael E. Sykuta, Associate Professor in the Division of Applied Social Sciences and Director of the Contracting and Organizations Research Institute, University of Missouri

The full amicus brief is available here.