Cite as: 592 U. S. ____ (2020) 1
Statement of THOMAS, J.
SUPREME COURT OF THE UNITED STATES
MALWAREBYTES, INC. v. ENIGMA SOFTWARE
GROUP USA, LLC
ON PETITION FOR WRIT OF CERTIORARI TO THE UNITED
STATES COURT OF APPEALS FOR THE NINTH CIRCUIT
No. 19–1284. Decided October 13, 2020
The petition for a writ of certiorari is denied.
Statement of JUSTICE THOMAS respecting the denial of
certiorari.
This petition asks us to interpret a provision commonly
called §230, a federal law enacted in 1996 that gives Inter-
net platforms immunity from some civil and criminal
claims. 47 U. S. C. §230. When Congress enacted the stat-
ute, most of today’s major Internet platforms did not exist.
And in the 24 years since, we have never interpreted this
provision. But many courts have construed the law broadly
to confer sweeping immunity on some of the largest compa-
nies in the world.
This case involves Enigma Software Group USA and Mal-
warebytes, two competitors that provide software to enable
individuals to filter unwanted content, such as content pos-
ing security risks. Enigma sued Malwarebytes, alleging
that Malwarebytes engaged in anticompetitive conduct by
reconfiguring its products to make it difficult for consumers
to download and use Enigma products. In its defense, Mal-
warebytes invoked a provision of §230 that states that a
computer service provider cannot be held liable for provid-
ing tools “to restrict access to material” that it “considers to
be obscene, lewd, lascivious, filthy, excessively violent, har-
assing, or otherwise objectionable.” §230(c)(2). The Ninth
Circuit relied heavily on the “policy” and “purpose” of §230
to conclude that immunity is unavailable when a plaintiff
alleges anticompetitive conduct.
2 MALWAREBYTES, INC. v. ENIGMA SOFTWARE
GROUP USA, LLC
Statement of THOMAS, J.
The decision is one of the few where courts have relied on
purpose and policy to deny immunity under §230. But the
court’s decision to stress purpose and policy is familiar.
Courts have long emphasized nontextual arguments when
interpreting §230, leaving questionable precedent in their
wake.
I agree with the Court’s decision not to take up this case.
I write to explain why, in an appropriate case, we should
consider whether the text of this increasingly important
statute aligns with the current state of immunity enjoyed
by Internet platforms.
I
Enacted at the dawn of the dot-com era, §230 contains
two subsections that protect computer service providers
from some civil and criminal claims. The first is defini-
tional. It states, “No provider or user of an interactive com-
puter service shall be treated as the publisher or speaker of
any information provided by another information content
provider.” §230(c)(1). This provision ensures that a com-
pany (like an e-mail provider) can host and transmit third-
party content without subjecting itself to the liability that
sometimes attaches to the publisher or speaker of unlawful
content. The second subsection provides direct immunity
from some civil liability. It states that no computer service
provider “shall be held liable” for (A) good-faith acts to re-
strict access to, or remove, certain types of objectionable
content; or (B) giving consumers tools to filter the same
types of content. §230(c)(2). This limited protection enables
companies to create community guidelines and remove
harmful content without worrying about legal reprisal.
Congress enacted this statute against specific back-
ground legal principles. See Stewart v. Dutra Constr. Co.,
543 U. S. 481, 487 (2005) (interpreting a law by looking to
the “backdrop against which Congress” acted). Tradition-
ally, laws governing illegal content distinguished between
Cite as: 592 U. S. ____ (2020) 3
Statement of THOMAS, J.
publishers or speakers (like newspapers) and distributors
(like newsstands and libraries). Publishers or speakers
were subjected to a higher standard because they exercised
editorial control. They could be strictly liable for transmit-
ting illegal content. But distributors were different. They
acted as a mere conduit without exercising editorial control,
and they often transmitted far more content than they
could be expected to review. Distributors were thus liable
only when they knew (or constructively knew) that content
was illegal. See, e.g., Stratton Oakmont, Inc. v. Prodigy Ser-
vices Co., 1995 WL 323710, *3 (Sup. Ct. NY, May 24, 1995);
Restatement (Second) of Torts §581 (1976); cf. Smith v. Cal-
ifornia, 361 U. S. 147, 153 (1959) (applying a similar prin-
ciple outside the defamation context).
The year before Congress enacted §230, one court blurred
this distinction. An early Internet company was sued for
failing to take down defamatory content posted by an uni-
dentified commenter on a message board. The company
contended that it merely distributed the defamatory state-
ment. But the company had also held itself out as a family-
friendly service provider that moderated and took down of-
fensive content. The court determined that the company’s
decision to exercise editorial control over some content “ren-
der[ed] it a publisher” even for content it merely distri-
buted. Stratton Oakmont, 1995 WL 323710, *3–*4.
Taken at face value, §230(c) alters the Stratton Oakmont
rule in two respects. First, §230(c)(1) indicates that an In-
ternet provider does not become the publisher of a piece of
third-party content—and thus subjected to strict liability—
simply by hosting or distributing that content. Second,
§230(c)(2)(A) provides an additional degree of immunity
when companies take down or restrict access to objectiona-
ble content, so long as the company acts in good faith. In
short, the statute suggests that if a company unknowingly
leaves up illegal third-party content, it is protected from
publisher liability by §230(c)(1); and if it takes down certain
4 MALWAREBYTES, INC. v. ENIGMA SOFTWARE
GROUP USA, LLC
Statement of THOMAS, J.
third-party content in good faith, it is protected by
§230(c)(2)(A).
This modest understanding is a far cry from what has
prevailed in court. Adopting the too-common practice of
reading extra immunity into statutes where it does not be-
long, see Baxter v. Bracey, 590 U. S. —— (2020) (THOMAS,
J., dissenting from denial of certiorari), courts have relied
on policy and purpose arguments to grant sweeping protec-
tion to Internet platforms. E.g., 1 R. Smolla, Law of Defa-
mation §4:86, p. 4–380 (2d ed. 2019) (“[C]ourts have ex-
tended the immunity in §230 far beyond anything that
plausibly could have been intended by Congress); accord,
Rustad & Koenig, Rebooting Cybertort Law, 80 Wash. L.
Rev. 335, 342–343 (2005) (similar). I address several areas
of concern.
A
Courts have discarded the longstanding distinction be-
tween “publisher” liability and “distributor” liability. Al-
though the text of §230(c)(1) grants immunity only from
“publisher” or “speaker” liability, the first appellate court to
consider the statute held that it eliminates distributor lia-
bility too—that is, §230 confers immunity even when a com-
pany distributes content that it knows is illegal. Zeran v.
America Online, Inc., 129 F. 3d 327, 331–334 (CA4 1997).
In reaching this conclusion, the court stressed that permit-
ting distributor liability “would defeat the two primary pur-
poses of the statute,” namely, “immuniz[ing] service provid-
ers” and encouraging “selfregulation.” Id., at 331, 334. And
subsequent decisions, citing Zeran, have adopted this hold-
ing as a categorical rule across all contexts. See, e.g., Uni-
versal Communication Systems, Inc. v. Lycos, Inc., 478 F.
3d 413, 420 (CA1 2007); Shiamili v. Real Estate Group of
NY, Inc., 17 N. Y. 3d 281, 288–289, 952 N. E. 2d 1011, 1017
(2011); Doe v. Bates, 2006 WL 3813758, *18 (ED Tex.,
Dec. 27, 2006).
Cite as: 592 U. S. ____ (2020) 5
Statement of THOMAS, J.
To be sure, recognizing some overlap between publishers
and distributors is not unheard of. Sources sometimes use
language that arguably blurs the distinction between pub-
lishers and distributors. One source respectively refers to
them as “primary publishers” and “secondary publishers
or disseminators,” explaining that distributors can be
“charged with publication.” W. Keeton, D. Dobbs, R.
Keeton, & D. Owen, Prosser and Keeton on Law of Torts
799, 803 (5th ed. 1984).
Yet there are good reasons to question this interpretation.
First, Congress expressly imposed distributor liability in
the very same Act that included §230. Section 502 of the
Communications Decency Act makes it a crime to “know-
ingly . . . display” obscene material to children, even if a
third party created that content. 110 Stat. 133–134 (codi-
fied at 47 U. S. C. §223(d)). This section is enforceable by
civil remedy. 47 U. S. C. §207. It is odd to hold, as courts
have, that Congress implicitly eliminated distributor liabil-
ity in the very Act in which Congress explicitly imposed it.
Second, Congress enacted §230 just one year after Strat-
ton Oakmont used the terms “publisher” and “distributor,”
instead of “primary publisher” and “secondary publisher.”
If, as courts suggest, Stratton Oakmont was the legal back-
drop on which Congress legislated, e.g., FTC v. Accusearch
Inc., 570 F. 3d 1187, 1195 (CA10 2009), one might expect
Congress to use the same terms Stratton Oakmont used.
Third, had Congress wanted to eliminate both publisher
and distributor liability, it could have simply created a cat-
egorical immunity in §230(c)(1): No provider “shall be held
liable” for information provided by a third party. After all,
it used that exact categorical language in the very next sub-
section, which governs removal of content. §230(c)(2).
Where Congress uses a particular phrase in one subsection
and a different phrase in another, we ordinarily presume
that the difference is meaningful. Russello v. United States,
464 U. S. 16, 23 (1983); cf. Doe v. America Online, Inc., 783
6 MALWAREBYTES, INC. v. ENIGMA SOFTWARE
GROUP USA, LLC
Statement of THOMAS, J.
So. 2d 1010, 1025 (Fla. 2001) (Lewis, J., dissenting) (relying
on this rule to reject the interpretation that §230 eliminated
distributor liability).
B
Courts have also departed from the most natural reading
of the text by giving Internet companies immunity for their
own content. Section 230(c)(1) protects a company from
publisher liability only when content is “provided by an-
other information content provider.” (Emphasis added.)
Nowhere does this provision protect a company that is itself
the information content provider. See Fair Housing Coun-
cil of San Fernando Valley v. Roommates.Com, LLC, 521 F.
3d 1157, 1165 (CA9 2008). And an information content pro-
vider is not just the primary author or creator; it is anyone
“responsible, in whole or in part, for the creation or devel-
opment” of the content. §230(f )(3) (emphasis added).
But from the beginning, courts have held that §230(c)(1)
protects the “exercise of a publisher’s traditional editorial
functions—such as deciding whether to publish, withdraw,
postpone or alter content.” E.g., Zeran, 129 F. 3d, at 330
(emphasis added); cf. id., at 332 (stating also that §230(c)(1)
protects the decision to “edit”). Only later did courts wres-
tle with the language in §230(f )(3) suggesting providers are
liable for content they help develop “in part.” To harmonize
that text with the interpretation that §230(c)(1) protects
“traditional editorial functions,” courts relied on policy ar-
guments to narrowly construe §230(f )(3) to cover only sub-
stantial or material edits and additions. E.g., Batzel v.
Smith, 333 F. 3d 1018, 1031, and n. 18 (CA9 2003) (“[A] cen-
tral purpose of the Act was to protect from liability service
providers and users who take some affirmative steps to edit
the material posted”).
Under this interpretation, a company can solicit thou-
sands of potentially defamatory statements, “selec[t] and
edi[t] . . . for publication” several of those statements, add
Cite as: 592 U. S. ____ (2020) 7
Statement of THOMAS, J.
commentary, and then feature the final product promi-
nently over other submissions—all while enjoying immun-
ity. Jones v. Dirty World Entertainment Recordings LLC,
755 F. 3d 398, 403, 410, 416 (CA6 2014) (interpreting “de-
velopment” narrowly to “preserv[e] the broad immunity
th[at §230] provides for website operators’ exercise of tradi-
tional publisher functions”). To say that editing a state-
ment and adding commentary in this context does not
“creat[e] or develo[p]” the final product, even in part, is
dubious.
C
The decisions that broadly interpret §230(c)(1) to protect
traditional publisher functions also eviscerated the nar-
rower liability shield Congress included in the statute. Sec-
tion 230(c)(2)(A) encourages companies to create content
guidelines and protects those companies that “in good faith
. . . restrict access to or availability of material that the pro-
vider or user considers to be obscene, lewd, lascivious,
filthy, excessively violent, harassing, or otherwise objec-
tionable.” Taken together, both provisions in §230(c) most
naturally read to protect companies when they unknow-
ingly decline to exercise editorial functions to edit or remove
third-party content, §230(c)(1), and when they decide to ex-
ercise those editorial functions in good faith, §230(c)(2)(A).
But by construing §230(c)(1) to protect any decision to
edit or remove content, Barnes v. Yahoo!, Inc., 570 F. 3d
1096, 1105 (CA9 2009), courts have curtailed the limits
Congress placed on decisions to remove content, see e-ven-
tures Worldwide, LLC v. Google, Inc., 2017 WL 2210029, *3
(MD Fla., Feb. 8, 2017) (rejecting the interpretation that
§230(c)(1) protects removal decisions because it would
“swallo[w] the more specific immunity in (c)(2)”). With no
limits on an Internet company’s discretion to take down ma-
terial, §230 now apparently protects companies who ra-
cially discriminate in removing content. Sikhs for Justice,
8 MALWAREBYTES, INC. v. ENIGMA SOFTWARE
GROUP USA, LLC
Statement of THOMAS, J.
Inc. v. Facebook, Inc., 697 Fed. Appx. 526 (CA9 2017), aff ’g
144 F. Supp. 3d 1088, 1094 (ND Cal. 2015) (concluding that
“ ‘ any activity that can be boiled down to deciding whether
to exclude material that third parties seek to post online is
perforce immune’ ” under §230(c)(1)).
D
Courts also have extended §230 to protect companies
from a broad array of traditional product-defect claims. In
one case, for example, several victims of human trafficking
alleged that an Internet company that allowed users to post
classified ads for “Escorts” deliberately structured its web-
site to facilitate illegal human trafficking. Among other
things, the company “tailored its posting requirements to
make sex trafficking easier,” accepted anonymous pay-
ments, failed to verify e-mails, and stripped metadata from
photographs to make crimes harder to track. Jane Doe No.
1 v. Backpage.com, LLC, 817 F. 3d 12, 16–21 (CA1 2016).
Bound by precedent creating a “capacious conception of
what it means to treat a website operator as the publisher
or speaker,” the court held that §230 protected these web-
site design decisions and thus barred these claims. Id., at
19; see also M. A. v. Village Voice Media Holdings, LLC, 809
F. Supp. 2d 1041, 1048 (ED Mo. 2011).
Consider also a recent decision granting full immunity to
a company for recommending content by terrorists. Force
v. Facebook, Inc., 934 F. 3d 53, 65 (CA2 2019), cert. denied,
590 U. S. —— (2020). The court first pressed the policy ar-
gument that, to pursue “Congress’s objectives, . . . the text
of Section 230(c)(1) should be construed broadly in favor of
immunity.” 934 F. 3d, at 64. It then granted immunity,
reasoning that recommending content “is an essential re-
sult of publishing.” Id., at 66. Unconvinced, the dissent
noted that, even if all publisher conduct is protected by
§230(c)(1), it “strains the English language to say that in
Cite as: 592 U. S. ____ (2020) 9
Statement of THOMAS, J.
targeting and recommending these writings to users . . . Fa-
cebook is acting as ‘the publisher of . . . information pro-
vided by another information content provider.’ ” Id., at 76–
77 (Katzmann, C. J., concurring in part and dissenting in
part) (quoting §230(c)(1)).
Other examples abound. One court granted immunity on
a design-defect claim concerning a dating application that
allegedly lacked basic safety features to prevent harass-
ment and impersonation. Herrick v. Grindr LLC, 765 Fed.
Appx. 586, 591 (CA2 2019), cert. denied, 589 U. S. ——
(2019). Another granted immunity on a claim that a social
media company defectively designed its product by creating
a feature that encouraged reckless driving. Lemmon v.
Snap, Inc., 440 F. Supp. 3d 1103, 1107, 1113 (CD Cal. 2020).
A common thread through all these cases is that the
plaintiffs were not necessarily trying to hold the defendants
liable “as the publisher or speaker” of third-party content.
§230(c)(1). Nor did their claims seek to hold defendants li-
able for removing content in good faith. §230(c)(2). Their
claims rested instead on alleged product design flaws—that
is, the defendant’s own misconduct. Cf. Accusearch, 570
F. 3d, at 1204 (Tymkovich, J., concurring) (stating that
§230 should not apply when the plaintiff sues over a defend-
ant’s “conduct rather than for the content of the infor-
mation”). Yet courts, filtering their decisions through the
policy argument that “Section 230(c)(1) should be construed
broadly,” Force, 934 F. 3d, at 64, give defendants immunity.
II
Paring back the sweeping immunity courts have read into
§230 would not necessarily render defendants liable for
online misconduct. It simply would give plaintiffs a chance
to raise their claims in the first place. Plaintiffs still must
prove the merits of their cases, and some claims will un-
doubtedly fail. Moreover, States and the Federal Govern-
ment are free to update their liability laws to make them
10 MALWAREBYTES, INC. v. ENIGMA SOFTWARE
GROUP USA, LLC
Statement of THOMAS, J.
more appropriate for an Internet-driven society.
Extending §230 immunity beyond the natural reading of
the text can have serious consequences. Before giving com-
panies immunity from civil claims for “knowingly host[ing]
illegal child pornography,” Bates, 2006 WL 3813758, *3, or
for race discrimination, Sikhs for Justice, 697 Fed. Appx.,
at 526, we should be certain that is what the law demands.
Without the benefit of briefing on the merits, we need not
decide today the correct interpretation of §230. But in an
appropriate case, it behooves us to do so.