Fixing Big Tech’s Complicity in Misinformation and Crime

By Chris Ricchetti | 13 February 2021

The urgent crisis of misinformation demands a comprehensive, all-of-society response. The ideas examined here could lead to significant improvements. But many more perspectives and resources, along with as much national soul-searching and determination as we can muster are needed now. Our deep divisions are being hard wired into our brains. And we know that the road we are on now ends at “checkmate humanity.”

Overwhelmed by Technology
Defining the Problem with Big Tech
Addressing the Three Complaints
Political Discrimination
Criminality
Misinformation
Multilateral Working Groups
Fact Checking
More Accountability
Don’t Wreck the Internet
Give Humans a Fighting Chance
Toward an Algorithmic Fairness Doctrine
Carpe Diem
End Notes

➢ To monitor proposed and pending legislation related to Section 230, check the Disruptive Competition Project.

➢ For research into the sources and tactics of organized misinformation campaigns worldwide, check the Project on Computational Propaganda at Oxford.

➢ For the latest from the front lines of technology activism, follow the Center for Humane Technology.

This discussion builds upon a previous article that examined the enactment, consequences since enactment, and pending proposals for amendment or repeal of Section 230 of Title 47 of the United States Code, a 1996 statute that provides internet companies with broad immunity from civil liability arising from content that users post to their platforms.

Overwhelmed by Technology

We live in a world where sophisticated algorithms process gargantuan amounts of data to drive an unending stream of custom-curated content into the mind of each individual technology user. This hyper-automated bias confirmation complex plays upon the neural plasticity of the human brain, continuously reinforcing our tribal beliefs. The colossal rift between the “Two Americas” is being hardwired into our brains at warp speed.

As Tristan Harris, co-founder of the Center for Humane Technology, has observed, futurists have long anticipated with trepidation some distant point in time when technology would become so advanced that it would overwhelm human strength (i.e., perform all tasks—from labor to critical thinking—better than humans), effectively rendering human beings economically “obsolete.” What no one saw coming was the earlier point—according to Harris, we are already there—when technology would have the capacity to overwhelm human weakness.

User “engagement” drives revenue for social media platforms. The more time users spend scrolling their feeds and the more interactions (e.g., likes, shares, comments, etc.) they transact, the more data platforms can collect and the more paid advertising they can display. That is how the business model works. Consequently, content delivery on social media is optimized for user engagement—for “hooking” users and holding their attention, without regard for any of the detrimental effects on them or on the societies in which they live. Playing upon our fear, outrage, and other primordial instincts powerfully and reliably sustains user engagement (that and cat videos).

Evan Williams, a co-founder of both Twitter and Blogger and currently the CEO of Medium, explains that content-curation algorithms generally do not distinguish between looking at content and preferring it. He observes that “everyone” feels compelled to look at car crashes as they pass. He goes on to explain, “The internet interprets behavior like this to mean everyone is asking for car crashes, so it tries to supply them.” This leads to an all-car-crashes-all-the-time environment and incentivizes many users and content creators to post content that is ever more extreme. Tristan Harris calls this phenomenon the “race to the bottom of the brain stem.”

Through perpetual arousal of our most destructive instincts (i.e., fear and anger) the algorithmic technology deployed on social media platforms is overwhelming human weakness on a massive scale. Our country is being ripped apart by the polarizing effects of tech-enabled echo chambers. This is an imminent and ongoing threat to national security, indeed to our continued existence as a democratic republic. The United States of America as we know it cannot and will not survive if we allow this to continue.

What actions should Americans and their government take to save ourselves from this dire situation?

Defining the Problem with Big Tech

Not surprisingly, there is no consensus around what specific problem(s) need solving. The current broad-based outcry against “Big Tech” is a confusing cacophony of at least three categories of frustrations: (1) complaints about lies, misinformation, and conspiracy theories, (2) complaints about incitement to political violence and other illegal activity, (3) complaints about political discrimination, censorship, and the Freedom of Speech.

Conservatives complain that Big Tech censors their voices (i.e., over-moderates conservative content). To punish this perceived injustice, they say they want social media companies to be legally responsible[1] for anything and everything that billions of users post on their platforms—which, obviously, would incentivize a much greater degree of content moderation (censorship), not less.

One must assume that conservative lawmakers well know that removing Section 230 immunity from social media platforms would not fix their censorship complaint and would, in fact, make the problem worse. They muddy the waters further by ranting about “violations” of their First Amendment “rights”—an argument that rests upon a (presumably intentional) distortion of the right to Freedom of Speech. They know that, to the extent the First Amendment can properly be applied to content moderation at all, it protects platforms, not users. They know that the whole of Corporate America relies on immunity from liability for third-party content.

Confronting the very real existential crisis that we face is not made easier when politicians and pundits knowingly cast their complaints within an improper legal framework and pretend to advocate for solutions that are both legally and politically untenable.

Liberals, for their part, are concerned mostly with misinformation, violence, and other criminal activity. They see in the proliferation of untruth a dangerous, cultural revolt against science, expertise, and empirical facts, and they argue that without a shared set of basic facts, there can be no functioning democracy. While there has always been a certain amount of misinformation in circulation, technology has fundamentally changed the media landscape. Propaganda may be as old as time, but the reach and repetitiveness of today’s tech-enabled propaganda machines are unprecedented in human history.

The use of technology to bring communities of terrorists together, to link predators with potential victims, and to facilitate every kind of crime and corruption is also novel. Public officials across the political spectrum accuse social media and other internet services of willful blindness and negligence in under-moderating unconscionable activity on their platforms.

One must assume that left-leaning lawmakers also realize that the full repeal of Section 230 would not be a sufficient remedy for their complaints and would precipitate a multitude of unintended consequences.

Neither party has yet proposed a targeted, workable strategy to correct Big Tech’s complicity in the misinformation crisis. Both parties seem to agree that social media platforms should not function as arbiters of truth. Platforms have never wanted this responsibility anyway, though some have recently indicated their willingness to take it on, provided that the Congress provides them with “cover” in the form of clear and detailed guidance.

Addressing the Three Complaints

Let us address the three categories of complaints against Big Tech (see above) in reverse order.

Political Discrimination

With respect to Category 3—political discrimination, censorship, and the Freedom of Speech—this set of complaints is the source of much of confusion. Under current law, social media platforms are private spaces, and there is no “right” to the Freedom of Speech in any private space. A compelling case could be made that the largest social media platforms have become de facto public spaces where discrimination should be prohibited by law, as it is in other public accommodations. However, not all forms of discrimination are illegal, even in public spaces. Fixing the perceived problem of political discrimination on the part of social media platforms would also require that political views be added to the list of Federally protected traits.[2] The types of discrimination that are currently illegal under Federal law are limited to discrimination on the basis of race, color, religion, sex, or national origin.[3]

Adding political views to the list of Federally protected traits and expanding the statutory definition of public accommodations to include the largest social media platforms would be constructive improvements from a fairness perspective and would focus moderation efforts on content that is demonstrably false or misleading and content that facilitates (potential or actual) criminal activity—Categories 1 and 2. ✔ Done.

Criminality

With respect to Category 2—incitement to political violence and other illegal activity—this set of complaints could be addressed, in part, by enacting exceptions to Section 230 immunity for any of the following: (i) willful blindness to criminality, (ii) negligence in moderating criminality, and (iii) intentional criminality.[4] Of course, existing criminal statues and case law already prohibit criminal activity online. However, overly broad interpretations of Section 230 immunity frequently impede the robust prosecution of criminal offenses.[5] This clear injustice could be fixed simply by adding clarifying language to Section 230(c)(1) to the effect that the broad immunity granted thereunder shall not impede enforcement of any criminal law or prohibit Federal and state governments or injured parties from bringing civil actions in connection with underlying criminal conduct.

Past efforts to correct the Category 2 problem have been piecemeal, focusing on specific types of heinous criminality, such as sex trafficking or the sexual exploitation of children. Making it clear that Section 230 provides no protection whatsoever for any criminal conduct should eliminate the need to specify every conceivable type of criminality in the amended language of the statue.

In addition to clarifying Section 230 as it relates to criminal conduct (including negligence in moderating and willful blindness to criminal conduct), new legislation is needed to establish mandatory minimum standards and practices for moderating criminality, reporting possible criminal activity, cooperating with law enforcement, responding to court orders in a timely manner, etc. Content distributors that demonstrate faithful compliance should be entitled to “safe harbor” protection from civil liability in connection with what should then be low incidences of crime on their platforms, except when mens rea can be established. ✔ Done.

Misinformation

Category 1—lies, misinformation, and conspiracy theories—is the most urgent of the three, as it constitutes a clear and present existential threat to democracy. Addressing the problem of misinformation will be extraordinarily difficult, given that the Two Americas live in vastly different and largely incompatible “realities.” But if we are to believe Tristan Harris’ prophetic warning that technology has already pushed us dangerously close to the point he calls “checkmate humanity,” then we have no choice but to develop systems to curtail the spread of misinformation that will be both constitutional and widely accepted across the Two Americas—even at the cost of some of our liberty,[6] and even at the cost of an internet that may become somewhat less “fun” and less convenient.

Multilateral Working Groups

As a starting point, multiple interdisciplinary working groups should be convened to study the crisis of misinformation[7] and to develop systems, procedures, technologies, educational programs, legislation, and other solutions to combat it. It is essential that the solutions ultimately implemented should not be perceived as emanating from government. However, a presidential and/or congressional commission could lead by establishing a timeline for brainstorming, experimentation, and evaluation of potential solutions, as well as target dates for introducing legislation based on the most promising evidence-based ideas.[8]

Working groups could be convened by Federal and state governments, nonprofit organizations, science and academia, Big Tech, smaller tech, grassroots activists, and other stakeholders. Presumably, many of the solutions arising from this process would not require legislative action and could be implemented by the private and nonprofit sectors as they are developed.[9]

Fact Checking

Although “truth arbitration” by itself will not be sufficient to bridge the canyon separating the Two Americas, any honest examination of the misinformation crisis should lead to the conclusion that some degree of empirical fact checking is urgently needed.

In our pluralistic, democratic society, the range of content subjected to mandatory fact checking would have to be narrow and limited to that which can be transparently and empirically verified, then confirmed from a multitude of diverse perspectives. As philosophers have observed for thousands of years, objectivity is elusive. On the other hand, facts are not relative.

As a society, we must reach a consensus about demarcation of the line that separates alternative points of view from “alternative facts.” And we must agree that empirical verification of basic facts is essential for democracy, even when, inevitably, not everyone may agree with every determination made by fact checking institutions.[10]

More Accountability

As discussed at length in a previous article, it is fashionable now for lawmakers to threaten Big Tech with the full repeal of Section 230 (or major modifications to it), as if using this blunt instrument would somehow solve all three categories of complaints against social media platforms once and for all. For all the reasons considered there, this is a terrible idea that would have far reaching, undesirable consequences.

Don’t Wreck the Internet

General protection from civil liability arising from third-party content is a fundamental necessity for doing business on the internet, not a special privilege. Section 230 immunity must be the default for internet platforms that host user-generated content[11]—unless they knowingly facilitate or are negligent in moderating clear criminality. Changes to the law and new regulations should target specific outcomes and be tied to a comprehensive strategy for dealing with the misinformation crisis and the Two-Americas divide.

Give Humans a Fighting Chance

Foremost in every effort to fight the misinformation crisis should be the consideration of what to do about Big Tech’s reckless use of algorithms to exploit fear and outrage, for the ultimate purpose of monetizing users. This is an urgent public health[12] and national security threat. The use of algorithmic technology to amplify or to limit the reach of user-generated and third-party content should be categorically prohibited as a strict-liability tort and should disqualify offenders from Section 230 immunity across all social media platforms they own or operate. Severe mandatory civil penalties and regulatory sanctions should also be imposed. Clearly, this would be a new constraint on First Amendment rights on which the Supreme Court would ultimately have to rule. In the face of imploding democracy, this constraint on liberty is not only justified but critical.

The use of algorithms for most purposes should be acceptable. Deploying user data and algorithms for the purpose of serving micro-targeted advertising content is a defensible practice (that should be monitored for potential regulation as the technology continues to evolve).[13] The utilization of algorithms as moderation tools (e.g., auto-removal of content containing the “N” word) is another constructive use. And there are, of course, many others.

As noted above, the First Amendment does not confer upon users any “right” to the Freedom of Speech on private social media platforms. Nonetheless, the freedom to expose oneself to a broad range of ideas is one of the key benefits—for individuals and for democratic society—of our First Amendment rights. (Of course, with or without technology, democracy depends on citizens having the will and making the effort to exercise this freedom.) The abuse of content-curation algorithms is antithetical to this core American value in that the technology drives a customized set of preferred ideas into each user’s brain without relent, sabotaging critical thought and making us all less free. All individual rights are subject to limits that serve the public good. The greater good that will result from prohibiting destructive uses of algorithms far outweighs the cost to tech companies of limiting their right to deploy them.

Toward an Algorithmic “Fairness Doctrine”

Instead of using algorithms to turn up the volume in the echo chamber, another partial solution to the misinformation crisis might be to require platforms to deploy algorithms that would present users with content that represents a different point of view on the content they are consuming.

For example, immediately following the 2020 presidential election, several analyses[14] that purported to demonstrate the “statistical impossibility” of a Biden win were widely shared and swallowed whole by millions. Rebuttal analyses quickly arose that were circulated in different echo chambers. It would have been good for the citizens of both Americas to have been presented with convenient links to opposing viewpoints as they were consuming content that reflected their default perspectives.

Such “alternative viewpoint algorithms” might also have helped to lessen the intensity of the debate over mask wearing. There is a legitimate, ongoing inquiry in academic medicine into the efficacy of surgical masks in the operating theatre. The science is not as conclusive as their nearly universal deployment in surgical facilities might suggest. Nonetheless, most guardians of public health, in full awareness of the lingering uncertainty about mask efficacy in the surgical context, have rendered their professional judgment in favor of public mask wearing as a means of reducing the rate of transmission of the virus that causes COVID-19. And research during the pandemic seems to indicate that masks do provide imperfect but substantial protection. Given that most of us are not scientists, our consumption of scientific information calls for a lot more humility, on all sides.

It is dead wrong (pun intended) to assert that masks “don’t work.”[15] It is also not correct, though far less dangerous, to assert that the scientific understanding of mask efficacy is complete and conclusive. Both Americas would have benefited from exposure to a fuller picture. This might have helped some accept the judgment of experts and ratify their participation in the social contract of mask wearing, despite their personal reservations.

Carpe Diem

The urgent crisis of misinformation demands a comprehensive, all-of-society response. The ideas examined here could lead to significant improvements. But many more perspectives and resources, along with as much national soul-searching and determination as we can muster are needed now. Our deep divisions are being hard wired into our brains. And we know that the road we are on now ends at “checkmate humanity.”

End Notes

  1. i.e., by removing the immunity from civil liability provided under Section 230.

  2. Some countries already prohibit discrimination on the basis of politics.

  3. Legislation proposed by former Attorney General William Barr in September 2020 attempts to “fix” the perceived problem of political discrimination by establishing good faith standards applicable to content moderation. Barr would require platforms to disclose their content moderation policies in detail, apply them consistently, and avoid any appearance of political discrimination. Otherwise, they could be sued by users, whose claims of harm would be legally tenuous, likely rendering Barr’s good faith standards an exercise in futility.

  4. Several bills introduced in the 116th and 117th Congresses, as well as Barr’s proposed legislation (see footnote 3), would eliminate Section 230 immunity for platforms that knowingly facilitate criminal content or activity.

  5. While Section 230(e)(1) clearly states that the statute was intended to have no effect on criminal law, some courts misapply Section 230(c)(1) to shield guilty parties from criminal prosecution. More frequently, Section 230(c)(1) bars law enforcement and injured parties from civil actions in connection with criminal activity.

  6. Most Americans were willing to accept significant limits on their civil liberties following the tragic events of September 11, 2001, ostensibly in service of the public good.

  7. A great deal of new research into the psychology and social psychology of misinformation has been conducted since 2016, augmenting the already considerable body of knowledge. The sources and tactics of organized misinformation campaigns is another active area of research.

  8. The constitutionality of regulating platforms’ moderation practices is unclear—and effectively unknowable until such time as relevant litigation provides the Supreme Court with an opportunity to decide this. One should not assume that the publisher-distributor distinction is a strict dualism. It is conceivable the Supreme Court would decide that, for purposes of content moderation authorized under Section 230(c)(2), platforms are publishers (of a special kind, given their immunity) and, therefore, beyond the reach of government regulation; but that, for all other purposes, platforms are mere distributors, pursuant to Section 230(c)(1). Or the Court could decide that platforms are distributors for all purposes and, therefore, regulating their content moderation practices would not be a violation of their First Amendment rights. Instead, by establishing content moderation standards to be operationalized by platforms, the government would likely be infringing the First Amendment rights of platform users (i.e., the actual publishers of user generated content).

  9. An independent oversight board established by Facebook in May 2020 is already offering to advise other social media platforms regarding their content moderation practices.

  10. The Federal government might consider establishing the framework for a new type of nonprofit organization—call them “independent fact checking institutions”—with a governance structure that includes robust checks and balances designed to minimize conflicts of interest and the risk of corruption. Such entities could be required by law to be fully transparent, with every financial and operational detail made public in real time. The Federal government could incentivize the nonprofit sector to create these organizations and require social media platforms to “subscribe” to one or more of them. The fact checking entities would have the power to moderate all user-generated and third-party content published on a subscribing platform. Perhaps fact checking bodies could be crowd funded, similar to public radio and television, except with contribution limits.

  11. Section 230 protects commerce and innovation across the entire internet, not just Big Tech. According to the Internet Association, the parties that have relied on Section 230 as a legal defense include “internet service providers and website hosts, newspapers, universities, libraries, search engines, employers, bloggers, website moderators and listserv owners, marketplaces, app stores, spam protection and anti-fraud tools, domain name registrars, and social media companies.”

  12. The misuse of algorithms to maximize user “engagement” (i.e., addiction) is not unlike Big Tobacco’s use of additives to enhance addiction and mask the negative effects of cigarettes.

  13. Tristan Harris makes a compelling case that the use of personal data and algorithms to serve advertising content is similarly abusive in its power to undermine independent thought. He argues that the power dynamic between humans and algorithms is so far out-of-balance that the interaction cannot be fair. Many others agree.

  14. Examples include an analysis by Charles J. Cicchetti, Ph.D. comparing Clinton’s (2016) and Biden’s (2020) results in key states, a claim that the election results violate Benford’s Law, an argument based on Biden’s loss of all but one of the traditional “bellwether counties,” etc.

  15. Anti-mask arguments based on the extremely small size of the virus relative to the distance between mask fibers are incomplete and inadequate. And there are straightforward explanations as to why the guidance from public health officials regarding mask wearing has changed over time.

It’s the Algorithms, Stupid! – Understanding the Section 230 Debate

By Chris Ricchetti | 27 January 2021

Our country is being torn apart by the polarizing effects of tech-enabled echo chambers… Strategic curation of the content presented to users is the very definition of what it means to be a speaker or a publisher… Both amplifying and constraining the reach of any user-generated content algorithmically should be illegal.

PREFACE Up-Front Disclosure
PART I Background
What is Section 230
Common Misconceptions
Section 230 In Context
Provisions of Section 230
Modifications Since 1996
PART II The Current Debate
Political “Neutrality”
Truth and Lies
Complete Repeal
Mixed Metaphors
Barr’s Two Cents
Getting Warmer
Right Tool, Right Job
It’s the Algorithms, Stupid!
AFTERTHOUGHTS Lingering Legal Ambiguities
END NOTES

Updated 13 February 2021: To monitor the latest proposed and pending legislation related to Section 230, check the Disruptive Competition Project.

Center for Humane Technology Logo

To learn more about the effects of technology on human well-being, visit the Center for Humane Technology.

PREFACE – Up-Front Disclosure

This article provides a straightforward and relatively “objective” review of Section 230, the circumstances that gave rise to it, the intent behind it, the intended and unintended consequences of its enactment, judicial interpretations, and the robust public debate about potential changes to the law—none of which address the most serious and urgent tech-enabled threat to democracy: the reckless misuse of algorithmic technology.

Also included—at no charge—are the author’s point of view and analysis, which can be summarized as follows:

  1. There are three broad categories of problematic, user-generated content: (1) content that is clearly criminal, (2) content that is “objectionable” but not clearly criminal, and (3) content that is false, distorted, misleading, etc., but not clearly criminal. One (1) is the only category that government is constitutionally empowered to restrict. This does not make the other two any less destructive or absolve Americans and our government from seeking ways to lessen their detrimental effects.
  2. General protection from civil liability for third-party content must be the default position for internet platforms that host user-generated content—unless they knowingly facilitate or are negligent in moderating clear criminality. Section 230 protects commerce and innovation across the entire internet. It does not only benefit “Big Tech.” [1]
  3. However, the existing protections conferred under Subsection (c)(1) are too broad. Section 230 should not, as courts have frequently held that it does, protect platforms that knowingly facilitate clearly criminal content or activity, that are willfully blind, or that are negligent in moderating clear criminality. Such platforms should not be immune from liability and should be shut down.
  4. Lawmakers should define in detail the minimum required standards and practices for moderating clear criminality, including penalties for non-compliance (i.e., something other than the loss of Section 230 protections across the offender’s entire business).
  5. To the extent that overly broad interpretations of Section 230(c)(1) may continue to impede the government and injured parties from bringing civil actions in connection with criminal conduct, clarifying language should be added to Subsection (c)(1), without disturbing its vitally important broad civil liability protections. Obviously, the law should never be applied as a shield from culpability for any underlying criminal conduct.
  6. Lawmakers should not attempt to regulate the moderation of content that is “objectionable” but not clearly criminal (category “2” above). The largest platforms do this reasonably well, despite innumerable (and inevitable) errors, omissions, and inconsistencies.
  7. The tech-enabled amplification of lies, distortions, and misinformation (category “3” above) on a global scale is—by far—the most serious and destructive consequence of Big Tech’s ubiquity in American life.
  8. The use of algorithms to auto-curate user-generated content (including shared content not authored by the sharing user) should be categorically prohibited as a strict-liability tort. Offenders should be excluded from all liability protections under Section 230(c)(1) and be deemed publishers of all third-party content posted to their platforms.
  9. Withdrawing the liability protections granted under Section 230 is not the appropriate remedy for correcting perceived injustices related to discrimination on the basis of politics or the Freedom of Speech. Civil rights law is the proper forum in which to fight those battles, for those who are inclined to fight them.
  10. Withdrawing the liability protections granted under Section 230 is not the appropriate remedy for correcting anticompetitive markets. Antitrust law is the proper forum in which to fight those battles, for those who are inclined to fight them.
  11. Keeping children physically, emotionally, and morally safe and secure should always be of paramount importance in policymaking. There is no higher purpose under the social contract on which democracy is based. However, whenever social conservatives purport to do anything in the name of “protecting children,” there is cause for skepticism and scrutiny. Evangelical and Catholic leaders have a decades-long track record of appealing to the universally shared value of protecting children to justify legislation and law enforcement actions that were, in fact, intended to chill or to criminalize constitutionally protected adult speech. In retrospect, most of these efforts have failed to withstand judicial scrutiny and have been partially or fully struck down by the courts—including, in more than one instance, by the Supreme Court of the United States. Consequently, the producers of some of the most extreme and harmful content have either been punished very lightly or not punished at all. Society would be better served by the enactment of more narrowly tailored statutes that provide muscular protections for kids without advancing a broader, moralistic “hidden agenda.”

PART I – Background

The current debate about Section 230, part of the Communications Decency Act of 1996, is taking place at a time when Americans are deeply polarized. Technology has brought to light and greatly amplified our differences to such an extent that there are now, quite literally, two “Americas”—each with its own values, priorities, “facts,” and perceptions of reality. That Big Tech companies have become “too big” and “too powerful” is among the rarest of declarative statements that both Americas deem to be “true,” albeit for different reasons.

What is Section 230?

Title V, Section 509 of the Communications Decency Act of 1996,[2] entitled “Online Family Empowerment,” amends the Communications Act of 1934 by adding Section 230 at the end of Title II.

The heading of Section 230 reads, “Protection for Private Blocking and Screening of Offensive Material.”

In short, Section 230 encourages tech companies to censor content that is “obscene, lewd, lascivious, filthy, excessively violent, harassing, or otherwise objectionable” and shields them from civil liability when they do. In other words, content creators—including users who post comments—may not sue tech companies for labeling or removing their content in “good faith.” (Establishing “bad faith” is neither inexpensive, nor easy.)

Most of the 1996 Act was struck down by the Supreme Court[3] subsequent to its enactment, but the part of the Act commonly referred to as “Section 230” was deemed constitutional and remains in effect.

Common Misconceptions

Before examining the current debate about Section 230, let us look at some of the common legal misconceptions—among both left and right leaning Americans—that are clouding the debate.

The First Amendment to the Constitution of the United States prohibits the government from infringing upon Americans’ Right to Freedom of Speech, except when doing so serves a clear and compelling public interest. The First Amendment does not restrict citizens from taking action to limit, suppress, distort, or otherwise interfere with each other’s speech. There are many laws that do protect Americans from each other, but the First Amendment is not one of them. Social media platforms are private spaces, and no one has a constitutional “right” to freedom of speech on social media. In fact, the First Amendment protects the rights of social media platforms to moderate their private spaces without government interference. Their moderation decisions will ultimately be judged in the marketplace.

There are other important features of American civic life that many Americans have either forgotten or may never have learned. Many Millennial and Gen-Z Americans, by no fault of their own, were raised with a set of values from which they could easily and understandably infer certain inaccuracies about American law, particularly as it relates to the freedom of speech and discrimination. Consider the following examples:

Speech that is commonly referred to as “hate speech”—as immoral, destructive, and despicable as it may be—is not categorically illegal. More precisely, it is illegal only when it is direct, personal, and can be shown to be “truly threatening” or reasonably likely to incite others to immediate, illegal action, especially violence. As in all areas of the law, there are many nuances and an extensive body of case law in which the contours of these general principles have been worked out and applied to a wide variety of facts and circumstances. Consequently, the latitude to spew hatred is not as wide as it may seem. But there is a great deal of highly offensive speech, including some that many would consider to be “triggering,” that is nonetheless legal and protected under the First Amendment. While the subset of all hate speech that is prohibited by law is narrower than many Americans realize, Section 230 does enable tech companies, if they choose, to censor the complementary subset of all hate speech that is hateful but not illegal.

Discrimination—as immoral, destructive, and despicable as it may be—is also not categorically illegal. More precisely, it is illegal only in certain contexts and only when it affects a legally protected class of persons. The laws restricting discrimination are somewhat different in the areas of employment, housing, education, and public accommodations. The types of discrimination that are always illegal include discrimination on the basis of “race, color, religion, sex, or national origin.” In some parts of the country, private clubs and religious organizations are free to discriminate—even on the basis of these otherwise unlawful forms of discrimination. However, if a business owner has a personal distaste for rich people, or “latte liberals,” or people who eat meat, or doctors who perform abortions, or dog owners who don’t clean up after their dogs, or alumni of Michigan State, it is altogether legal for her to deny service for any of these reasons—provided that the reason given for the denial of service is not a “cover” for some other form of discrimination that is actually illegal. In June 2018, former White House Press Secretary, Sarah Huckabee Sanders, was ejected from the Red Hen restaurant in Lexington, Virginia because of her association with President Donald Trump. This discrimination was legal, and Ms. Sanders had no legal recourse. Likewise, social media platforms and other internet spaces are fully within their rights to deny service or to censor anyone on the basis of politics. And they are under no obligation to exercise discriminatory bias on a consistent basis.[4]

Those who have suggested (perhaps not without merit) that social media platforms of a certain size are de facto public spaces could pursue expanding the statutory meaning of “public accommodations” to include large internet platforms, and they could pursue adding “political views” to the list of protected traits[5] under the Civil Rights Act of 1964. Some countries already prohibit discrimination on the basis of politics. But Section 230 was not intended to deal with issues of discrimination and is not the proper legal tool to correct any perceived injustice related to discrimination.

Section 230 is also not about monopoly power. To the extent that the enormity, economies of scale, reach, and other market advantages of Big Tech companies may constitute nearly insurmountable barriers to entry and stifle competition, antitrust litigation is the proper legal tool to correct these anticompetitive market dynamics. Threats to withdraw Section 230 liability protections as means of “punishing” Big Tech for their success are misguided and will not change their oligarchical market positions or lessen their market dominance.

Lastly, it should be self-evident that Big Tech is wholly uninterested in functioning as arbiters of truth. We know this because they resisted public pressure to involve themselves in “truth moderation” and politics for a long time and, until recently, had made few attempts to censor even the most egregious lies and oppressive political content from despicable actors (e.g., foreign dictators). Worse yet, some smaller players have done exactly nothing upon learning that their platforms were facilitating horrific crimes. Big Tech is interested in one thing only: maximizing user engagement. Big Tech simply wants users on their social media platforms consuming content—pretty much any content—for as long as possible, every day, because that is what drives their revenue. From their perspective, robust content moderation is an enormously expensive and risky distraction from making money.

Big Tech is not in collusion with the “deep state,” or “antifa,” or George Soros, or the “squad,” or the “Illuminati,” or “a cabal of cannibalistic democrat celebrity pedophiles” to thwart democracy and establish a one-world communist government in fulfillment of Biblical prophecy. To the extent that “shadowbanning” was ever a “thing,” the practice was an early effort by Big Tech to quietly extract themselves from the irreconcilable demands of the Two Americas. Their moral failing has been their intentional amplification of polarizing falsehoods—not their reluctant efforts to censor lies, distortions, and actual criminal content in response to a sustained public outcry and threats from lawmakers of both parties to cut them off at the knees. As with Corporate America generally, Big Tech’s obsession with maximizing shareholder value without regard for any other moral precepts defines its depravity, not “censorship.”

Section 230 In Context

When Section 230 was enacted in 1996, public access to and commercial use of the internet were new, as was the graphical user interface. Section 230 was intended to address two specific—and closely related—issues of concern at the time.

First, before Section 230 was enacted, some courts held[6] that if internet companies restricted access to any third-party content, they would then become liable for all third-party content accessed via their services—to the same extent as if they had been the actual creators of the content. Few internet companies would have survived if they could have been sued for every instance of racist, sexist, discriminatory, libelous, obscene, or otherwise offensive content that users might encounter. The clear legal incentive before Section 230 was for internet companies to do no moderation at all. Many Americans wanted internet companies to take a more active role in “policing” and restricting access to offensive content. Section 230 was intended to correct the perceived problem of under-moderation.

Second, in the early years of the public internet, many Americans were concerned about the proliferation of pornography online (perhaps they were on to something) and they wanted to ensure that children would not have access to or inadvertently land on adult content. As the title of the legislation (Communications Decency Act of 1996) suggests, preventing minors from encountering adult content was the overarching and primary purpose of Section 230 and of the Act as a whole. In fact, the part of the Act that establishes Section 230 is entitled “Online Family Empowerment.” Specifically, the legislation was intended to enable internet access software companies—such as CompuServe and America Online—to deploy technologies that would empower parents to block adult websites and other content not appropriate for children, without being deemed publishers of and without assuming potential liability for all the content on all the websites across the entire internet that users accessed via their services.

The technology environment in 1996 was vastly different from what it is today. Users accessed the internet via the existing land-based telephone lines. There were no internet service providers (ISPs) apart from the land-based phone carriers.[7] There were no social media platforms. Online communities—called “bulletin boards”—were much smaller and tended to be populated by a small subculture of “brainiacs” and “geeks.” Most Americans were not even aware that they existed. There was almost no (non-adult) entertainment online. There was no streaming video—neither prerecorded nor live (content had to be downloaded, then viewed or played locally). There were no lifestyle or affinity websites, limited news, little commerce, and few platforms (other than the aforementioned “bulletin boards”) where users could post comments or engage in open dialogue. There were no mobile devices capable of accessing the internet or any of its content. All of this developed rapidly in the years following 1996, but none of it was known or contemplated at the time Section 230 was enacted.

Provisions of Section 230

Section 230 is both clear and concise.[8]

Subsections (a) and (b) of Section 230 state that the internet is super cool and that it is the policy of the United States to promote its ongoing development.

The substance of the legislation is found at Subsection (c), which reads as follows:

(c) Protection for “Good Samaritan” blocking and screening of offensive material

(1) Treatment of publisher or speaker

No provider or user of an interactive computer service shall be treated as the publisher or speaker of any information provided by another information content provider.

(2) Civil liability

No provider or user of an interactive computer service shall be held liable on account of—

(A) any action voluntarily taken in good faith to restrict access to or availability of material that the provider or user considers to be obscene, lewd, lascivious, filthy, excessively violent, harassing, or otherwise objectionable, whether or not such material is constitutionally protected; or

(B) any action taken to enable or make available to information content providers or others the technical means to restrict access to material described in paragraph (1).[9]

Subsection (d) simply states that internet companies should notify users of any parental control features available on their platforms, and Subsection (e) confirms that Section 230 is not intended to affect the interpretation or enforcement of certain existing laws.

In short, Section 230 resolved the “Moderator’s Dilemma” (see footnote 6). Subsection (c)(1) says, in effect, “There is no need to censor anything and everything that could possibly be offensive to somebody somewhere, because no one can sue you anyway (i.e., don’t over-moderate).” The intent here was to keep the internet a place for vibrant and open discussion. Subsection (c)(2) encourages moderation where it truly matters by saying, in effect, “On the other hand, we don’t want you to completely ignore the third-party content on your platform either (i.e., don’t under-moderate). These are the types of content that we’d like you to make a good-faith effort to moderate…” The intent here was to make the internet safe for kids. The language of Section 230 enables, indeed, encourages, platforms to restrict user access to objectionable content without the risk of being deemed the creators of any content created by a third party.

Modifications Since 1996

In response to the Supreme Court’s decision to strike down all but Section 230 of the Communications Decency Act of 1996, Congress passed, and President Clinton signed into law, a more narrowly tailored effort to modify the liability protections of Section 230 called the Child Online Protection Act (1998). This legislation was also overturned by the High Court, on the grounds that it was still too overly broad, vague, and burdensome and was likely to chill legal speech between adults.

The Digital Millennium Copyright Act (1998), without directly modifying the language of Section 230, created new requirements for internet companies to respond to infringement complaints related to copyrighted content posted on their platforms, and opened them to potential liability for third-party content if they fail to comply with the “safe-harbor” procedures for processing infringement complaints set forth in the Act.

While Section 230 explicitly states that the legislation was not intended to affect criminal law,[10] it has nonetheless been cited by courts as the basis for shielding internet companies and their responsible agents from civil actions in cases where they had been knowing enablers of, or the principal actors in, criminal activity. In response, Congress passed, and President Trump signed into law, the Allow States and Victims to Fight Online Sex Trafficking Act of 2017 (FOSTA-SESTA). The Act clarifies that Section 230 does not prohibit the enforcement of Federal or state criminal or civil law when internet companies “knowingly assist, facilitate or support sex trafficking,” thus creating the first explicit statutory exception to the liability protections afforded by Section 230.

It appears to some that forcing sex trafficking offline may have had the unintended effect of making it more difficult for law enforcement to prosecute sex trafficking crimes and may also be making sex workers less safe.[11] Others have claimed that the Act stifles innovation in that it creates an oversized risk for tech entrepreneurs who could inadvertently run afoul of the law with devastating consequences. Add these to the laundry list of concerns that should be considered when evaluating proposals for further legislative action.

In the early days of the public internet, before anyone could conceive of the advances in technology that would enable its vast commercial potential or the development of a “digital divide,” broadband (i.e., internet-access infrastructure) was treated as an “information service” for regulatory purposes. This designation limited the powers of the Federal Communications Commission (FCC) to regulate broadband.

In February 2015, in response to growing public demands to preserve the neutrality and openness that made the internet great, the Obama FCC issued the Open Internet Order, that reclassified broadband as a “telecommunications” service under Title II of the Communications Act of 1934, subjecting it to the more stringent regulatory regime applicable to “common carriers” (e.g., phone companies). Among other consequences, the FCC was then able to ban internet service providers (ISPs) from “picking internet winners and losers” via practices such as blocking, throttling, and paid prioritization. Additionally, ISPs would now have a duty to work toward universal access to more uniform internet access services at more uniform prices. Ironically, ISPs would not now need the protections of Section 230 because their status as common carriers came with its own protections from civil and criminal liability. Internet actors higher on the “stack” than ISPs, including websites, and computer and mobile applications, remained very much reliant upon Section 230 protections.

In 2018, the Trump FCC issued the Restoring Internet Freedom Order (you can’t make this stuff up) restoring ISPs to their original, lightly regulated status as “information services,” thus defeating the Obama Administration’s noble efforts to maintain net neutrality.

In May 2020, former President Trump signed an executive order purported to strip Twitter, Facebook, and Google of Section 230 liability protections. Like most of Trump’s reactive, petty, and entirely uninformed initiatives, it went nowhere. It did, however, dovetail with a more substantive examination by the Barr Justice Department of all things related to Section 230, already under way, that culminated in proposed legislation being sent to Congress by Attorney General Barr (see below).

PART II – The Current Debate

With all this in mind, let us now turn our attention to the current debate about the content moderation practices of Big Tech and the arguments being made for amending or repealing Section 230.

Political “Neutrality”

We’ll start with an easy one—the Ending Support for Internet Censorship Act. In June 2020, insurrectionist and Dunning-Kruger poster boy, Senator Josh Hawley (R-MO) proposed requiring tech companies to “prove” their “political neutrality” (i.e., that their moderation practices reflect no “discernable viewpoint”) to the Federal Trade Commission every two years as a condition for Section 230 immunity. Really, Josh? Government control of private speech? Flag on the play. First Amendment violation. 6-year penalty. Next!

Senator Hawley’s proposed legislation and his messaging around it are cynical, misleading and in bad faith. As a Yale Law graduate, he well understands that his proposed “solution” to the “problem” of Big Tech “censorship” is unconstitutional on its face. Senator Hawley simply happens to be, for the moment, one of the more visible performers carrying out the standard Republican modus operandi: flagrantly lying to their own constituents, for whom they have neither respect, nor concern. Instead of leading by telling the truth about issues and proposing workable solutions, they opportunistically seize upon every grievance du jour, and milk it dry to augment their own power. None but a small handful of Republicans stand for any discernible consistent ideology beyond their own self-interest. Hawley has no regard for the freedom of speech as a shared value, and no concern for the insidious injustices of discrimination, except in this one instance where he perceives it to be affecting him. He is plenty well prepared by education and experience to lead his constituents back to reality and to come up with meaningful proposals to address their real concerns. And he knows that Section 230 is not the right tool for the job.

As previously stated, it may well be that large social media platforms should be regarded as public accommodations[12] under the law, and it may well be that discrimination on the basis of politics should be legally prohibited. However, each would represent a fundamental change to American law that would have implications far beyond the scope of Section 230. If these more foundational changes were to be enacted, cleaning up the language of Section 230 to bring it into conformity would then become little more than a ministerial function.

Truth and Lies

Concerns about “fact checking” and “truth arbitration” (i.e., moderating intentionally misleading or factually incorrect content) should be kept separate and distinct from concerns about discrimination on the basis of politics and the freedom to express dissenting opinions (see previous section). They are two different things altogether.

In the current deeply polarized environment, solving the problem of pervasive misinformation is surely the more challenging of the two. No one trusts Big Tech to do it, and everyone seems to believe that the slightest intrusion by government would take us down a steep slippery slope to authoritarianism—even though most Americans were willing to accept significant limits on their civil rights following the tragic events of September 11, 2001, ostensibly in service of the public good.

Beyond the excruciating question as to who should be charged with truth arbitration, lies the equally daunting challenging of defining the line between alternative points of view and “alternative facts.

The difficulty of these challenges should not preclude trying to solve them—at least to some extent that would make a meaningful difference. In a pluralistic, democratic society, the scope of publicly accepted truth arbitration would have to be narrow and limited to that which can be transparently and empirically verified, then confirmed from a multitude of diverse perspectives. As philosophers have observed for thousands of years, objectivity is elusive. But facts are not relative. Our deeply divided house will not stand unless we can achieve some degree of reconciliation between the binary “realities” of the Two Americas.

Many participants on all sides of the current debate do not seem to realize that all of this is beyond the scope of anything that Section 230 was ever intended to address. A Section 230 makeover will not deliver us from the existential crisis that we face.

Complete Repeal

Both President Biden and former President Trump have advocated for the complete repeal of Section 230. Speaker Pelosi has called it a “gift” that “could be removed.” This is a terrible, “surgery-with-a-chainsaw” remedy that would have numerous far-reaching consequences, both good and bad.

While Section 230 has unquestioningly been misused to shield some very “bad hombres” from accountability for serious and harmful crimes (this can be fixed), its effect on the growth and development of the internet has been largely positive. Section 230 has even been credited with singularly shaping the development of the internet as we know it.[13]

Section 230 generally protects tech companies—as well as startups and hobbyist bloggers who host comments—from liability for user-generated content. This essential protection is a fundamental necessity for doing business on the internet, not a special privilege. We all benefit from innovation and competition among the providers of tech-enabled services—from home delivery to entertainment to local transportation and travel to the availability of vast information and educational resources to matchmaking. Removing this liability protection would, in the long run, harm the “little guy” more than it would harm Big Tech.

As understandably frustrated as so many are with Big Tech and its enormous influence on American life, the total repeal of Section 230 is not the proper remedy. To maintain an online environment that is vibrant, open, dynamic, and convenient, there must be reasonable limits on the financial risks associated with developing new applications and launching new ventures. Hobbyists publishing recipes and cooking videos on a private URL should not be held liable for offensive content that passersby may decide to post on a whim. Startups that are fortunate enough to experience sudden geometric growth should not be forced to prioritize and allocate disproportionate resources to human moderation of user-generated content during a make-it-or-break-it growth phase. Imposing this burden on “small tech” by law would create yet another barrier to entry and further shield Big Tech from meaningful competition.

Mixed Metaphors

A bill introduced in the Senate during the second session of the 116th Congress by Senators Graham (R-SC) and Blumenthal (D-CT) called the Eliminating Abusive and Rampant Neglect of Interactive Technologies Act of 2020 (EARN IT) would establish a National Commission on Online Child Sexual Exploitation Prevention, under the control of the Attorney General, to recommend best practices that platforms may (or may not) choose to implement to prevent and respond to instances of child sexual abuse on their platforms. The original bill would also have carved out a second statutory exception to Section 230 liability protections for tech companies that fail to comply with the Commission’s recommended best practices or demonstrate to the Commission that they have in place their own reasonable practices for preventing child sexual exploitation on their platforms. That provision was not included in the substitute version of the bill reported to the Judiciary Committee in July 2020.

The original version of the bill came under fire for an entirely different reason. The original bill would also have guaranteed “legal access” by law enforcement to any digital message, a provision that critics claimed would effectively ban the use of encryption technology in the United States—yet another debate for which Section 230 is not the proper forum. Not only was the new requirement regarding encryption removed from the substitute version of the bill, the substitute bill also expressly protects platforms from both civil and criminal liability for child sexual exploitation based solely on their use of encryption technology—even when it impedes law enforcement from gathering evidence.

Barr’s Two Cents

In June 2020, Attorney General William Barr sent proposed legislation to Congress that would condition immunity from liability for moderation decisions upon compliance with the moderation parameters set forth in his amended Subsection (c)(2). Specifically, moderation decisions would be shielded from liability only when platforms have an “objectively reasonable belief” that the moderated content is “obscene, lewd, lascivious, filthy, excessively violent, promoting terrorism or violent extremism, harassing, promoting self-harm, or unlawful, whether or not such material is constitutionally protected,” and only when they moderate in “good faith,” a requirement that Barr’s proposed legislation defines in detail.[14] Any other moderation decisions would not be shielded from liability under Section 230, although the general exemption from civil liability for unmoderated non-criminal content would remain in effect. The proposal is clearly intended to discourage “censorship” of political content. It might discourage some platforms from moderating most or all “category 2” (objectionable but not illegal) content.

Second, the Barr proposal would remove liability protections in circumstances where platforms purposefully “promote, solicit, or facilitate” third-party content or activity in violation of any Federal criminal law (called the “Bad Samaritan Carve-Out”).[15] Third, just as FOSTA-SESTA created an exception for sex trafficking content (see above), the Barr proposal would create three additional, specific exceptions—for child abuse and exploitation, terrorism, and cyberstalking content. Fourth, the Barr proposal confirms that the Federal and state governments could bring civil actions against platforms in connection with any failure to remove criminally unlawful content under one of the statutory exceptions to Section 230.

Barr’s proposed legislation is not unreasonable. Platforms of all sizes would continue to enjoy broad protections under Section 230, except when they cannot provide a “reasonable factual basis” for their moderation decisions. The addition of a “Bad Samaritan Carve-Out” would certainly be positive to the extent that it reinforced the ability of law enforcement to prosecute actual criminality that causes serious harm.

The “good faith” standards set forth in the Barr proposal are quite superior to those included in the Limiting Section 230 Immunity to Good Samaritans Act introduced at about the same time by Senators Josh Hawley (R-MO), Marco Rubio (R-FL), Mike Braun (R-IN), Tom Cotton (R-AR) and Kelly Loeffler (R-GA).[16] The Senators introduced the bill in deference to former President Trump’s Executive Order aimed at exacting revenge on Twitter, Facebook, and Google (see above). Undoubtedly, the Senators’ principled effort was further inspired by a series of lawsuits brought against social media platforms by aggrieved “influencers,” such as far-right sensation Laura Loomer, among many others, that had failed to yield any relief for the plaintiffs from the pervasive tyranny of anti-conservative bias.

As previously mentioned, under current law, social media platforms are private spaces, and discrimination on the basis of politics is not prohibited. Moreover, efforts by the government to regulate the moderation of legal content are likely unconstitutional under the First Amendment. It may well be that exceptionally large social media platforms have become de facto public spaces and should be treated as such for public accommodations and civil rights purposes. But Section 230 is not the proper forum for resolving such foundational and far-reaching issues.

Getting Warmer

The Platform Accountability and Consumer Transparency Act (PACT Act), introduced in June 2020 by Senators Brian Schatz (D-HI) and John Thune (R-SD), appears to be a sincere effort at meaningful reform that is uncomplicated by hidden motives. First, the Act would establish new transparency and reporting requirements that most major platforms already practice. Platforms would have to publish detailed acceptable use policies, detailed moderation standards, and quarterly statistics tracking their moderation decisions. All good. Second, platforms would have to provide users with mechanisms for reporting acceptable use violations and potentially illegal content, and for appealing moderation decisions.[17] Also good. Third, the Act would establish standards for processing, acting upon and responding to content-related complaints. Under certain circumstances, if platforms failed to comply with their own policies or applied them inconsistently, they could be subject to Federal Trade Commission enforcement actions for “unfair or deceptive acts and practices.” Points for creativity! Fourth, Section 230 protections would not apply in the case of certain violations, such as failing to remove illegal content within 24 hours of being notified by courts or users.

The proposed legislation would seem to incentivize platforms to adopt lenient acceptable use policies and to over-moderate, making it easier for malicious users to submit frivolous complaints to silence other users. Also, the Act would provide exceptions for smaller internet businesses—generally a fantastic idea! However, specifically with respect to criminality that causes serious harm, the smaller players probably commit some of the most egregious offenses.

Overall, the PACT Act would be a step in the right direction—a scalpel, rather than a chain saw. The Act deals primarily with the moderation of content that is actually illegal, as well as transparency in moderation practices generally, and it does not seek to misapply Section 230 to solve unrelated popular frustrations.

Right Tool, Right Job

The Digital Millennium Copyright Act (1998) and the Stop Internet Sexual Exploitation Act—a bill intended to rein in platforms that facilitate distribution of non-consensual porn (AKA “revenge porn”) introduced in December 2020 by Senators Jeff Merkley (D-OR) and Ben Sasse (R-NE)[18]—are two examples of targeted legislation that creates appropriate and effective regulation to solve specific problems related to third-party content while leaving the broad general liability protections of Section 230 untouched.

It’s the Algorithms, Stupid!

If any of the advancements in computer technology since 1996 rise to the level of a paradigm shift, it is the capacity to compile gargantuan amounts of data about individual users, married with sophisticated algorithms that drive an unending stream of custom-curated content into the mind of each individual technology user. This hyper-automated bias confirmation complex plays upon the neural plasticity of the human brain, continuously reinforcing of our tribal beliefs. The colossal rift between the Two Americas is being hardwired into our brains at warp speed. Our country is being torn apart by the polarizing effects of tech-enabled echo chambers.

It is not their good-faith efforts to fact-check and remove egregious lies—and certainly not their efforts to prevent clear criminality—that renders the extension of Section 230 immunity to large social media platforms indefensible. Their destructive use of algorithms is what harms the public good! Big Tech’s reckless use of machine learning to maximize user engagement time by exploiting fear and outrage, for the ultimate purpose of monetizing users, in willful blindness to the devastating consequences, should be categorically prohibited as a strict-liability tort.[19]

Strategic curation of the content presented to users is the very definition of what it means to be a speaker or a publisher.

The use of algorithms for some purposes should be acceptable. Deploying user data and algorithms for the purpose of serving micro-targeted advertising content is a defensible practice. Advertising content is reviewed and approved by platforms prior to publication and, therefore, is not third-party content and, therefore, is not shielded from liability under Section 230 anyway. The utilization of algorithms as moderation tools (e.g., auto-removal of content containing the “N” word) is another constructive use.

But when user data and algorithms are deployed for the purpose of custom curating third-party content—to any extent—the offending platform should lose all liability protections under Section 230 and be liable for all third-party content.

Tech companies should exercise no influence whatsoever on the reach of any non-commercial, third-party content on their platforms. Two factors only should affect the third-party content to which users are exposed: (1) their own independent efforts to seek and to select the content they prefer to consume and, (2) the sharing of content by other users with whom they have consented to connect on a platform—displayed in a non-curated format (e.g., reverse chronological order). Both amplifying and constraining the reach of any user-generated content algorithmically should be illegal.

This proposed change would only lessen, not eliminate, the problem of echo chambers.[20] Since it is not a legitimate interest of government to constrain the rights of users to consume any legal content of their choice, the proposed change strictly targets platforms, not users. People are people. But the tech-driven bias confirmation machine is overwhelming our human frailty and adding fuel to the fire of our innate tribalism. This is not making us freer and will not end well, if allowed to continue. Algorithm abuse is the ultimate product liability atrocity!

The Congress should find that the misuse of algorithmic curation technology causes severe harm to the public and should legislate against any use of such technology—either to amplify or to limit the reach of—third-party content, under penalty of the immediate loss of all Section 230 protections across all social media platforms owned or operated by the offender. Severe mandatory civil penalties and regulatory sanctions should also be imposed.

AFTERTHOUGHTS – Lingering Legal Ambiguities

(1) While it is certainly true that Section 230 has shielded internet companies and their responsible agents from civil actions in cases where they had been knowing enablers of, or the principal actors in, criminal activity, this commentator has been unable to verify that the law has ever been misapplied as a shield against culpability for the underlying criminal conduct. Any such misuse would obviously constitute a serious injustice.

(2) It is by no means clear that all content moderation practices are of the same kind. For example, when Twitter moderators attach a warning label to a questionable Tweet, is Twitter engaging in constitutionally protected counter-speech (i.e., beyond the reach of government to regulate or sanction) or is it revising the speech of the original speaker (i.e., as co-speaker), despite being treated as merely a distributor with moderation privileges under Section 230?

(3) Given the technically legal rights of platforms to discriminate on the basis of politics and to engage in arbitrary and inconsistent moderation, what, exactly, under the existing language of Section 230(c)(2) would constitute moderation in “bad faith” (assuming that the moderated content is merely objectionable to someone, and not illegal)?

END NOTES

  1. According to the Internet Association, a not-disinterested lobbying group representing “medium-big-to-big tech,” the parties that have “relied on” Section 230 as a legal defense include “internet service providers and website hosts, newspapers, universities, libraries, search engines, employers, bloggers, website moderators and listserv owners, marketplaces, app stores, spam protection and anti-fraud tools, domain name registrars, and social media companies.” The percentage of said parties who have successfully relied on Section 230 immunity was not reported. Regardless, the law does protect any website, computer application, or mobile application—from bloggers to news and sports apps, to fin-tech, to e-commerce, to content aggregators, to affinity websites, to social media platforms, etc.

  2. The Communications Decency Act of 1996 refers to Title V of the Telecommunications Act of 1996.

  3. See Reno v. American Civil Liberties Union (1997)

  4. It is more than trivially ironic that free-market conservatives who have long asserted—with quasi-religious fervor—that markets are “self-correcting” and that they lead naturally to the optimal allocation of resources are now incensed that Big Tech’s market-driven moderation practices are impacting them in ways they did not expect. When entire classes of workers are displaced by tectonic shifts in the structure of the economy, those externalities are dismissed as “natural workings of the market.” But any constraints on their “right” to post content to social media platforms is tyranny, “Yes, tyranny, I tell you!” SMDH

  5. In the case of Boy Scouts of America v. Dale (2000), the Supreme Court of the United States, relying on an earlier precedent, held that a private organization’s First Amendment Right of Free Association (citing its “implicit” right of non-association) trumps the State of New Jersey’s public accommodations law prohibiting discrimination on the basis of sexual orientation, which, like political beliefs, is not a Federally protected trait. Also note that, like social media platforms, private organizations such as the Boy Scouts are not public accommodations, under Federal law.

  6. Two early cases illustrate the “Moderator’s Dilemma” that Section 230 was intended to resolve. In Cubby, Inc. v. CompuServe Inc. (SDNY 1991) the Court held that CompuServe was not liable for allegedly defamatory content posted by a third party because it had made no effort to moderate the large volume of third-party content on its platforms and, therefore, was presumed unaware of the defamatory nature of the content and, therefore, was deemed to be a mere distributor of it. However, in Stratton Oakmont, Inc. v. Prodigy Services Co. (SCONY 1995) the Court held that Prodigy’s established moderation practices were a form of editorial control and that Prodigy, therefore, was liable as the publisher of allegedly defamatory content posted by a third party. The takeaway from both cases was that liability risk is minimized when platforms do not moderate third-party content at all. But refraining from content moderation altogether is problematic because, without it, platforms would eventually be overrun by offensive and criminal content that users would find objectionable. Section 230 was enacted to release platforms from this no-win situation for the public good.

  7. Because of their classification as “common carriers” under Title II of the Communications Act of 1934, the telephone companies that provided the land-wire infrastructure over which most internet traffic was transmitted (in 1996) were already shielded from content-related liability and exempted from any potential duty to moderate content.

  8. The outdated language of the statute has subsequently been clarified by the courts. “Interactive computer service” has been interpreted as including essentially any service or platform through which content is accessed. “Information” means content, broadly defined. “Information content provider” refers to the creators and publishers of any type of content—text, audio, images, video, etc. “Restrict access to” includes labeling, curating, removing, blocking, deleting, and, of course, restricting access to content.

  9. Here, “(1)” must be an overlooked typo that made it into the final legislation because the reference is nonsensical. The intended reference must surely have been: “…material described in paragraph (A).”

  10. Section 230, Subsection (e)(1)

  11. In December 2019, Senator Elizabeth Warren (D-MA) and Representative Ro Khanna (D-CA) introduced bills that would have commissioned the Department of Health and Human Services to study the unintended harmful effects of FOSTA-SESTA on sex workers. The House version of the bill was called the SAFE SEX Workers Act.

  12. Perhaps the most accurate way to characterize social media platforms is to say that they are advertising and personal data collection platforms. If so, it is not clear how this may affect the “public accommodations” debate.

  13. Referring to Section 230, Subsection (c)(1), Jeff Kosseff, attorney, and Assistant Professor of Cybersecurity Law at the United States Naval Academy, calls Section 230 “The Twenty-Six Words that Created the Internet,” which is also the title of his 2019 book recounting the history of the legislation’s impact since enactment.

  14. Platforms would be acting in “good faith” if they established and disclosed detailed moderation practices that complied with the Barr proposal’s narrow definition of permissible moderation, implemented those practices consistently and without pretext, and provided affected users with explanations for their moderation decisions and a mechanism for appealing those decisions.

  15. The “Bad Samaritan Carve-Out” is a necessary and important proposal. There are many loathsome actors who operate websites and applications for the express purpose of promoting, soliciting, and facilitating criminal content and activity that causes serious harm.

  16. Certain provisions of the Senate bill were not included in the Barr proposal, such as establishing statutory damages of $5,000 per user and limiting the applicability of the good-faith requirement to platforms with at least 30 million US users and $1.5 billion in global revenue.

  17. Platforms would be required to provide affected users with a detailed explanation for each decision to label, to remove, to demonetize, or to limit algorithmically the reach of any content within 14 days of making such decisions.

  18. The bill has been criticized for requiring independent adult content creators to disclose their personal information to platforms who may be run by despicable actors who should not be trusted with the information.

  19. Intending no hyperbole, former Google design ethicist, Tristan Harris, has characterized algorithmic technology’s capacity to overwhelm human weakness as “Checkmate Humanity”—an urgent and profound existential crisis that ultimately will lead to our demise.

  20. For clarity, bona fide content creators (i.e., publishers, for liability purposes), such as Breitbart “News,” would still be allowed to pound falsehoods into the brains of users on their own platforms, but Facebook and Twitter (mere distributors of falsehood) would be prohibited from amplifying or limiting the reach of Breitbart content on their platforms.