Section 230 protected Twitter from liability for deleting Senate candidate’s accounts

Plaintiff (an Arizona senate candidate) sued Twitter after it suspended four of plaintiff’s accounts. He brought claims for (1) violation of the First Amendment; (2) violation of federal election law; (3) breach of contract; (4) conversion, (5) antitrust; (6) negligent infliction of emotional distress; (7) tortious interference; and (8) promissory estoppel.

Twitter moved to dismiss on multiple grounds, including that Section 230(c)(1) of the Communications Decency Act (“CDA”), 47 U.S.C. § 230, rendered it immune from liability for each of plaintiff’s claims that sought to treat it as a publisher of third-party content.

The CDA protects from liability (1) any provider of an interactive computer service (2) whom a plaintiff seeks to treat as a publisher or speaker (3) of information provided by another information content provider.

The court granted the motion to dismiss, on Section 230 grounds, all of the claims except the antitrust claim (which it dismissed for other reasons). It held that Twitter is a provider of an interactive computer service. And plaintiff sought to treat Twitter as a publisher or speaker by trying to pin liability on it for deleting accounts, which is a quintessential activity of a publisher. The deleted accounts were comprised of information provided by another information content provider (i.e., not Twitter, but plaintiff himself).

Brittain v. Twitter, 2019 WL 2423375 (N.D. Cal. June 10, 2019)

Section 230 protected Google in lawsuit over blog post

Defendant used Google’s Blogger service to write a post – about plaintiffs’ business practices – that plaintiffs found objectionable. So plaintiffs sued Google in federal court for defamation, tortious interference with a business relationship, and intentional infliction of emotional distress. The lower court dismissed the case on grounds that the Communications Decency Act (at 47 U.S.C. §230) immunized Google from liability for the publication of third party content.

Plaintiffs sought review with the U.S. Court of Appeals for the District of Columbia. On appeal, the court affirmed the dismissal. Applying a three part test the court developed in Klayman v. Zuckerberg, 753 F.3d 1354 (D.C. Cir. 2014) (which in turn applied analysis from the leading case of Zeran v. America Online, Inc., 129 F.3d 327 (4th Cir. 1997)), the court held that Section 230 entitled Google to immunity because: (1) Google was a “provider or user of an interactive computer service,” (2) the relevant blog post contained “information provided by another information content provider,” and (3) the complaint sought to hold Google liable as “the publisher or speaker” of the blog post.

The court rejected defendant’s argument that in establishing and enforcing its Blogger Content Policy, Google influenced and thereby created the content it published. It held that Google’s role was strictly one of “output control” – because Google’s choice was limited to a “yes” or a “no” decision whether to remove the post, its action constituted “the very essence of publishing.” Since Section 230 immunizes online defendants against complaints seeking to hold them as the publisher of content, the lower court properly dismissed the action.

Bennett v. Google, LLC, 882 F.3d 1163 (D.C. Cir., February 23, 2018)

About the Author: Evan Brown is a Chicago technology and intellectual property attorney. Call Evan at (630) 362-7237, send email to ebrown [at] internetcases.com, or follow him on Twitter @internetcases. Read Evan’s other blog, UDRP Tracker, for information about domain name disputes.

Ninth Circuit upholds decision in favor of Twitter in terrorism case

Tamara Fields and Heather Creach, representing the estates of their late husbands and joined by Creach’s two minor children, sued Twitter, Inc. Plaintiffs alleged that the platform knowingly provided material support to ISIS, enabling the terrorist organization to carry out the 2015 attack in Jordan that killed their loved ones. The lawsuit sought damages under the Anti-Terrorism Act (ATA), which allows U.S. nationals injured by terrorism to seek compensation.

Plaintiffs alleged that defendant knowingly and recklessly provided ISIS with access to its platform, including tools such as direct messaging. Plaintiffs argued that these services allowed ISIS to spread propaganda, recruit followers, raise funds, and coordinate operations, ultimately contributing to the attack. Defendant moved to dismiss the case, arguing that plaintiffs failed to show a direct connection between its actions and the attack. Defendant also invoked Section 230 of the Communications Decency Act, which shields platforms from liability for content created by users.

The district court agreed with defendant and dismissed the case, finding that plaintiffs had not established proximate causation under the ATA. Plaintiffs appealed, but the Ninth Circuit upheld the dismissal. The appellate court ruled that plaintiffs failed to demonstrate a direct link between defendant’s alleged support and the attack. While plaintiffs showed that ISIS used defendant’s platform for various purposes, the court found no evidence connecting those activities to the specific attack in Jordan. The court emphasized that the ATA requires a clear, direct relationship between defendant’s conduct and the harm suffered.

The court did not address defendant’s arguments under Section 230, as the lack of proximate causation was sufficient to resolve the case. Accordingly, this decision helped clarify the legal limits of liability for platforms under the ATA and highlighted the challenges of holding technology companies accountable for how their services are used by third parties.

Three Reasons Why This Case Matters:

  • Sets the Bar for Proximate Cause: The ruling established that a direct causal link is essential for liability under the Anti-Terrorism Act.
  • Limits Platform Liability: The decision underscores the difficulty of holding online platforms accountable for misuse of their services by bad actors.
  • Reinforces Section 230’s Role: Although not directly addressed, the case highlights the protections Section 230 offers to tech companies.

Fields v. Twitter, Inc., 881 F.3d 739 (9th Cir. 2018)

Anti-malware provider immune under CDA for calling competitor’s product a security threat

kills_bugs

Plaintiff anti-malware software provider sued defendant – who also provides software that protects internet users from malware, adware etc. – bringing claims for false advertising under the Section 43(a) of Lanham Act, as well as other business torts. Plaintiff claimed that defendant wrongfully revised its software’s criteria to identify plaintiff’s software as a security threat when, according to plaintiff, its software is “legitimate” and posed no threat to users’ computers.

Defendant moved to dismiss the complaint for failure to state a claim upon which relief may be granted. It argued that the provisions of the Communications Decency Act at Section 230(c)(2) immunized it from plaintiff’s claims.

Section 230(c)(2) reads as follows:

No provider or user of an interactive computer service shall be held liable on account of—

(A) any action voluntarily taken in good faith to restrict access to or availability of material that the provider or user considers to be obscene, lewd, lascivious, filthy, excessively violent, harassing, or otherwise objectionable, whether or not such material is constitutionally protected; or

(B) any action taken to enable or make available to information content providers or others the technical means to restrict access to material described in [paragraph (A)].

Specifically, defendant argued that the provision of its software using the criteria it selected was an action taken to make available to others the technical means to restrict access to malware, which is objectionable material.

The court agreed with defendant’s argument that the facts of this case were “indistinguishable” from the Ninth Circuit’s opinion in in Zango, Inc. v. Kaspersky, 568 F.3d 1169 (9th Cir. 2009), in which the court found that Section 230 immunity applied in the anti-malware context.

Here, plaintiff had argued that immunity should not apply because malware is not within the scope of “objectionable” material that it is okay to seek to filter in accordance with 230(c)(2)(B). Under plaintiff’s theory, malware is “not remotely related to the content categories enumerated” in Section 230(c)(2)(A), which (B) refers to. In other words, the objectionableness of malware is of a different nature than the objectionableness of material that is obscene, lewd, lascivious, filthy, excessively violent, harassing. The court rejected this argument on the basis that the determination of whether something is objectionable is up to the provider’s discretion. Since defendant found plaintiff’s software “objectionable” in accordance with its own judgment, the software qualifies as “objectionable” under the statute.

Plaintiff also argued that immunity should not apply because defendant’s actions taken to warn of plaintiff’s software were not taken in good faith. But the court applied the plain meaning of the statute to reject this argument – the good faith requirement only applies to conduct under Section 230(c)(2)(A), not (c)(2)(B).

Finally, plaintiff had argued that immunity should not apply with respect to its Lanham Act claim because of Section 230(e)(2), which provides that “nothing in [Section 230] shall be construed to limit or expand any law pertaining to intellectual property.” The court rejected this argument because although the claim was brought under the Lanham Act, which includes provisions concerning trademark infringement (which clearly relates to intellectual property), the nature of the Lanham Act claim here was for unfair competition, which is not considered to be an intellectual property claim.

Enigma Software Group v. Malwarebytes Inc., 2017 WL 5153698 (N.D. Cal., November 7, 2017)

About the Author: Evan Brown is a Chicago technology and intellectual property attorney. Call Evan at (630) 362-7237, send email to ebrown [at] internetcases.com, or follow him on Twitter @internetcases. Read Evan’s other blog, UDRP Tracker, for information about domain name disputes.

Google can, at least for now, disregard Canadian court order requiring deindexing worldwide

U.S. federal court issues preliminary injunction, holding that enforcement of Canadian order requiring Google to remove search results would run afoul of the Communications Decency Act (at 47 U.S.C. 230)

canada-us
Canadian company Equustek prevailed in litigation in Canada against rival Datalink on claims relating to trade secret misappropriation and unfair competition. After the litigation, Equustek asked Google to remove Datalink search results worldwide. Google initially refused altogether, but after a Canadian court entered an injunction against Datalink, Google removed Datalink results from google.ca. Then a Canadian court ordered Google to delist worldwide, and Google complied. Google objected to the order requiring worldwide delisting, and took the case all the way up to the Canadian Supreme Court, which affirmed the lower courts’ orders requiring worldwide delisting.

So Google filed suit in federal court in the United States, seeking a declaratory judgment that being required to abide by the Canadian order would, among other things, be contrary to the protections afforded to interactive computer service providers under the Communications Decency Act, at 47 U.S.C. 230.

The court entered the preliminary injunction (i.e., it found in favor of Google pending a final trial on the merits), holding that (1) Google would likely succeed on its claim under the Communications Decency Act, (2) it would suffer irreparable harm in the absence of preliminary relief, (3) the balance of equities weighed in its favor, and (4) an injunction was in the public interest.

Section 230 of the Communications Decency Act immunizes providers of interactive computer services against liability arising from content created by third parties. It states that “[n]o provider or user of an interactive computer service shall be treated as the publisher or speaker of any information provided by another information content provider.” [More info about Section 230]

The court found that there was no question Google is a “provider” of an “interactive computer service.” Also, it found that Datalink—not Google—“provided” the information at issue. And finally, it found that the Canadian order would hold Google liable as the “publisher or speaker” of the information on Datalink’s websites. So the Canadian order treated Google as a publisher, and would impose liability for failing to remove third-party content from its search results. For these reasons, Section 230 applied.

Summarizing the holding, the court observed that:

The Canadian order would eliminate Section 230 immunity for service providers that link to third-party websites. By forcing intermediaries to remove links to third-party material, the Canadian order undermines the policy goals of Section 230 and threatens free speech on the global internet.

The case provides key insight into the evolving legal issues around global enforcement and governance.

Google, Inc. v. Equustek Solutions, Inc., 2017 WL 5000834 (N.D. Cal. November 2, 2017)

YouTube not liable for aiding ISIS in Paris attack

The Communications Decency Act provided immunity to Google in a suit brought against it by the family of an American college student killed in the November 2015 attack.

Plaintiffs filed suit against Google (as operator of YouTube) alleging violation of federal laws that prohibit providing material support to terrorists, arising from the November 2015 Paris attack that ISIS carried out. Plaintiffs argued that the YouTube platform, among other things, aided in recruitment and provided ISIS the means to distribute messages about its activities.

Google moved to dismiss the lawsuit, arguing that Section 230 of the Communications Decency Act (47 U.S.C. 230) provided immunity from suit. The court granted the motion to dismiss.

Section 230 Generally

Section 230(c) provides that “[n]o provider or user of an interactive computer service shall be treated as the publisher or speaker of any information provided by another information content provider.” Accordingly, Section 230 precludes liability that treats a website as the publisher or speaker of information users provide on the website, protecting websites from liability for material posted on the website by someone else.

JASTA Did Not Repeal Section 230 Immunity

In response to Google’s arguments in favor of Section 230 immunity, plaintiffs first argued that a recent federal statute – the Justice Against Sponsors of Terrorism Act, or “JASTA” – effectively repealed the immunity conferred to interactive computer services by Section 230. Plaintiffs focused on language in the statute that stated that its purpose “is to provide civil litigants with the broadest possible basis, consistent with the Constitution of the United States, to seek relief” against terrorists and those who assist them.

The court rejected plaintiffs’ arguments that JASTA repealed Section 230 immunity. Significantly, the statute did not expressly repeal Section 230’s protections, nor did it do so implicitly by evincing any “clear and manifest” congressional intent to repeal any part of the Communications Decency Act.

Section 230 Need Not Be Applied Outside the United States

Plaintiffs also argued that Section 230 immunity did not arise because the Communications Decency Act should not apply outside the territorial jurisdiction of the United States. According to plaintiffs, Google provided support and resources to ISIS outside the United States (in Europe and the Middle East), ISIS’s use of Google’s resources was outside the United States, and the Paris attacks and plaintiffs’ relative’s death took place outside the United States.

The court rejected this argument, holding that Section 230’s focus is on limiting liability. The application of the statute to achieve that objective must occur where the immunity is needed, namely, at the place of litigation. Since the potential for liability, and the application of immunity was occurring in the United States, there was no need to apply Section 230 “extraterritorially”.

Immunity Protected Google

Google argued that plaintiffs’ claims sought to treat it as the publisher or speaker of the offending ISIS content, thus satisfying one of the requirements for Section 230 immunity. Plaintiffs countered that their lawsuit did not depend on the characterization of Google as the publisher or speaker of ISIS’s content, because their claims focused on Google’s violations of the federal criminal statutes that bar the provision of material support to terrorists.

But the court found that the conduct Google was accused of — among other things, failing to ensure that ISIS members who had been kicked off could not re-establish accounts — fit within the traditional editorial functions of a website operator. Accordingly, despite the plaintiffs’ characterization of its claims, the court found such claims to be an attempt to treat Google as the publisher or speaker of the ISIS videos.

The court similarly rejected plaintiffs’ arguments that Section 230 immunity should not apply because, by appending advertisements to some of the ISIS videos, Google became an “information content provider” itself, and thus responsible for the videos. This argument failed primarily because the content of the advertisements (which themselves were provided by third parties) did not contribute to the unlawfulness of the content of the videos.

Gonzalez v. Google, Inc., — F.Supp.3d —, 2017 WL 4773366 (N.D. Cal., October 23, 2017)

Evan_BrownAbout the Author: Evan Brown is a Chicago technology and intellectual property attorney. Call Evan at (630) 362-7237, send email to ebrown [at] internetcases.com, or follow him on Twitter @internetcases. Read Evan’s other blog, UDRP Tracker, for information about domain name disputes.

Google and YouTube protected by Section 230

The case of Weerahandi v. Shelesh is a classic example of how Section 230 (a provision of the Communications Decency Act (CDA), found at 47 USC 230) shielded online intermediaries from alleged tort liability occasioned by their users.

Background Facts

Plaintiff was a YouTuber and filed a pro se lawsuit for, among other things, defamation, against a number of other YouTubers as well as Google and YouTube. The allegations arose from a situation back in 2013 in which one of the individual defendants sent what plaintiff believed to be a “false and malicious” DMCA takedown notice to YouTube. One of the defendants later took the contact information plaintiff had to provide in the counter-notification and allegedly disseminated that information to others who were alleged to have published additional defamatory YouTube videos.

Google and YouTube also got named as defendants for “failure to remove the videos” and for not taking “corrective action”. These parties moved to dismiss the complaint, claiming immunity under Section 230. The court granted the motion to dismiss.

Section 230’s Protections

Section 230 provides, in pertinent part that “[n]o provider or user of an interactive computer service shall be treated as the publisher or speaker of any information provided by another information content provider.” 47 U.S.C. § 230(c)(1). Section 230 also provides that “[n]o cause of action may be brought and no liability may be imposed under any State or local law that is inconsistent with this section.” 47 U.S.C. § 230(e)(3).

The CDA also “proscribes liability in situations where an interactive service provider makes decisions ‘relating to the monitoring, screening, and deletion of content from its network.’ ” Obado v. Magedson, 612 Fed.Appx. 90, 94–95 (3d Cir. 2015). Courts have recognized Congress conferred broad immunity upon internet companies by enacting the CDA, because the breadth of the internet precludes such companies from policing content as traditional media have. See Jones v. Dirty World Entm’t Recordings LLC, 755 F.3d 398, 407 (6th Cir. 2014); Batzel v Smith, 333 F.3d 1018, 1026 (9th Cir. 2003); Zeran v. Am. Online, Inc., 129 F.3d 327, 330 (4th 1997); DiMeo v. Max, 433 F. Supp. 2d 523, 528 (E.D. Pa. 2006).

How Section 230 Applied Here

In this case, the court found that the CDA barred plaintiff’s claims against Google and YouTube. Both Google and YouTube were considered “interactive computer service[s].” Parker v. Google, Inc., 422 F. Supp. 2d 492, 551 (E.D. Pa. 2006). Plaintiff did not allege that Google or YouTube played any role in producing the allegedly defamatory content. Instead, Plaintiff alleged both websites failed to remove the defamatory content, despite his repeated requests.

Plaintiff did not cite any authority in his opposition to Google and YouTube’s motion, and instead argued that the CDA did not bar claims for the “failure to remove the videos” or to “take corrective action.” The court held that to the contrary, the CDA expressly protected internet companies from such liability. Under the CDA, plaintiff could not assert a claim against Google or YouTube for decisions “relating to the monitoring, screening, and deletion of content from its network. ” Obado, 612 Fed.Appx. at 94–95 (3d Cir. 2015); 47 U.S.C. § 230(c)(1) (“No provider or user of an interactive computer service shall be treated as the publisher or speaker of any information provided by another information content provider.”). For these reasons, the court found the CDA barred plaintiff’s claims against Google and YouTube.

Weerahandi v. Shelesh, 2017 WL 4330365 (D.N.J. September 29, 2017)

Evan_BrownAbout the Author: Evan Brown is a Chicago technology and intellectual property attorney. Call Evan at (630) 362-7237, send email to ebrown [at] internetcases.com, or follow him on Twitter @internetcases. Read Evan’s other blog, UDRP Tracker, for information about domain name disputes.

Quora gets Section 230 victory in the Tenth Circuit

Pro se plaintiff Silver filed suit in federal court in New Mexico against the online question-and-answer website Quora, alleging that statements made by two different individuals concerning his professional services were defamatory. Quora moved to dismiss, arguing that the immunity provisions of the Communications Decency Act, at 47 U.S.C. 230 shielded it from liability arising from content posted by its users. The district court granted the motion to dismiss. Plaintiff sought review with the Tenth Circuit Court of Appeals. On review, the court affirmed the lower court’s dismissal of the case.

Citing to its previous Section 230 precedent, Ben Ezra, Weinstein, & Co. v. Am. Online Inc., 206 F.3d 980 (10th Cir. 2000), the court held that Quora was a provider of “an interactive computer service,” that its actions forming the basis of alleged liability, namely, in hosting the content, were that of a “publisher or speaker,” and that the content giving rise to the alleged liability was from “another information content provider,” i.e., the users who posted the content.

Silver v. Quora, Inc., 2016 WL 6892146 (10th Circuit, November 23, 2016)

Evan_BrownAbout the Author: Evan Brown is a Chicago technology and intellectual property attorney. Call Evan at (630) 362-7237, send email to ebrown [at] internetcases.com, or follow him on Twitter @internetcases. Read Evan’s other blog, UDRP Tracker, for information about domain name disputes.

Yelp not liable for allegedly defamatory customer reviews

In a recent case having an outcome that should surprise no one, the United States Court of Appeals for the Ninth Circuit has affirmed a lower court’s decision that held Yelp immune from liability under the Communications Decency Act (47 U.S.C. 230 – the “CDA”) over customer reviews that were allegedly defamatory.

Plaintiff sued Yelp for violations under RICO and the Washington Consumer Protection Act, as well as libel under Washington law. Yelp moved to dismiss for failure to state to claim upon which relief may be granted. The lower court found that plaintiff had failed to allege any facts that plausibly suggested Yelp was responsible for the content, and therefore dismissed the case. Plaintiffs sought review with the Ninth Circuit. On appeal, the court affirmed.

The appellate court observed that plaintiff’s complaint, which he filed pro se, “pushed the envelope” of creative pleading. The court observed that plaintiff cryptically – “to the point of opacity” – alleged that Yelp was the one that created and developed the offending content. The court declined to open the door to such “artful skirting” of the Communications Decency Act’s safe harbor provision.

The key question before the court was whether the alleged defamatory reviews were provided by Yelp or by another information content provider. CDA immunity does not extend to situations where the web site itself is responsible for the creation or development of the offending content. The immunity protects providers or users of interactive computer services when the claims being made against them seek to treat them as a publisher or speaker of the information provided by another information content provider.

In this case, the court found that a careful reading of plaintiff’s complaint revealed that he never specifically alleged that Yelp created the content of the allegedly defamatory posts. Rather, plaintiff pled that Yelp adopted them from another website and transformed them into its own stylized promotions. The court found that these “threadbare” allegations of Yelp’s fabrication of allegedly defamatory statements were implausible on their face and were insufficient to avoid immunity under the Communications Decency Act. The court was careful to note that CDA immunity does not extend to content created or developed by an interactive computer service. “But the immunity in the CDA is broad enough to require plaintiffs alleging such a theory to state the facts plausibly suggesting the defendant fabricated content under a third party’s identity.”

The plaintiff had alleged in part that Yelp’s rating system and its use by the author of the allegedly defamatory content resulted in the creation or development of information by Yelp. The court rejected this argument, finding that the rating system did “absolutely nothing to enhance the defamatory sting of the message beyond the words offered by the user.” The court further observed that the star rating system was best characterized as a neutral tool operating on voluntary inputs that did not amount to content development or creation.

Finally, the court addressed plaintiff’s cryptic allegations that Yelp should be held liable for republishing the alleged defamatory content as advertisements or promotions on Google. A footnote in the opinion states that plaintiff was not clear whether the alleged republication was anything more than the passive indexing of Yelp reviews by the Google crawler. The decision’s final outcome, however, does not appear to depend on whether Google indexed that content as Yelp passively stood by or whether Yelp affirmatively directed the content to Google. “Nothing in the text of the CDA indicates that immunity turns on how many times an interactive computer service publishes information provided by another information content provider.” In the same way that Yelp would not be liable for posting user generated content on its web site, it would not be liable for disseminating the same content in essentially the same format to a search engine. “Simply put, proliferation and dissemination of content does not equal creation or development of content.”

Kimzey v. Yelp! Inc., — F.3d —, 2016 WL 4729492 (9th Cir. September 12, 2016)

Evan_BrownAbout the Author: Evan Brown is a Chicago technology and intellectual property attorney. Call Evan at (630) 362-7237, send email to ebrown [at] internetcases.com, or follow him on Twitter @internetcases. Read Evan’s other blog, UDRP Tracker, for information about domain name disputes.

Twitter avoids liability in terrorism lawsuit

Update 1/31/2018: The Ninth Circuit upheld the court’s decision discussed below.

The families of two U.S. contractors killed in Jordan sued Twitter, accusing the platform of providing material support to the terrorist organization ISIS. Plaintiffs alleged that by allowing ISIS to create and maintain Twitter accounts, the company violated the Anti-Terrorism Act (ATA). Plaintiffs further claimed this support enabled ISIS to recruit, fundraise, and promote extremist propaganda, ultimately leading to the deaths of the contractors. The lawsuit aimed to hold Twitter responsible for the actions of ISIS and to penalize it for facilitating the organization’s digital presence.

Twitter moved to dismiss, arguing that the claims were barred under the Communications Decency Act (CDA) at 47 U.S.C. §230. Section 230 provides immunity to internet platforms from being treated as the publisher or speaker of content posted by third parties. The court had to decide whether Twitter’s role in allowing ISIS to use its platform made it liable for the consequences of ISIS’s acts.

The court dismissed the case, finding that Section 230 shielded Twitter from liability. The court ruled that plaintiffs’ claims attempted to treat Twitter as the publisher of content created by ISIS, which is precisely the type of liability Section 230 was designed to prevent. The court also concluded that plaintiffs failed to establish a plausible connection, or proximate causation, between Twitter’s actions and the deaths. Importantly, in the court’s view, plaintiffs could not demonstrate that ISIS’s use of Twitter directly caused the attack in Jordan or that the shooter had interacted with ISIS content on the platform.

The court further addressed plaintiffs’ argument regarding private messages sent through Twitter’s direct messaging feature. It ruled that these private communications were also protected under Section 230, as the law applies to all publishing activities, whether public or private.

Three reasons why this case matters:

  • Expanding the scope of Section 230: The case reinforced the broad immunity provided to tech companies under Section 230, including their handling of controversial or harmful content.
  • Clarifying proximate causation in ATA claims: The ruling highlighted the challenges of proving a direct causal link between a platform’s operations and acts of terrorism.
  • Balancing tech innovation and accountability: The decision underscored the ongoing debate about how to balance the benefits of open platforms with the need for accountability in preventing misuse.

Fields v. Twitter, Inc., 200 F. Supp. 3d 964 (N.D. Cal., August 10, 2016).

Scroll to top