The song that was inspired by this article is available here.
My reading this article as an audio podcast is available here.
This week saw a dramatic turn in our nation’s desperate efforts to clean up the increasingly poisonous online sewers that we call social media.
First, the backstory:
If you publish a newspaper or newsletter (like this one) and you publish “illegal” content — encouragement of crimes or homicide, offer to sell drugs, display child porn, advocate overthrowing the government by violence — you can go to jail. If you publish things that defame or lie about persons or corporations, you can be sued into bankruptcy.
If you own a bookstore or newsstand and distribute books, magazines, and newspapers and offer for sale illegal content — child or snuff porn, stolen copyrighted material, instructions for making illegal drugs or weapons — you can also go to jail. And if you sell materials that openly defame individuals or corporations, you can also be sued into bankruptcy.
In the first category, you’d be a publisher. In the second, you’d be a distributor.
But what is social media? Particularly those types of social media that use an algorithm to push user-produced content out to people who haven’t explicitly asked for it?
Twenty-eight years ago, social media sites like CompuServe and AOL were regulated as if they were publishers, with the occasional/secondary oversight as if they were distributors. They had an obligation to make sure that illegal or defamatory content wasn’t published on their sites, or, if it was, to remove it within a reasonable time period.
The internet back then was a relatively safe and peaceful place. I know, as I ran a large part of one of the largest social media sites that existed at the time.
But then things got weird.
Back in 1996, some geniuses in Congress thought, “Hey, let’s do away with the entire concept of the publisher or distributor having responsibility for what happens in their place if that publisher or distributor happens to be operating online as what is called social media.”
Seriously. Selling drugs, trading in guns and ammunition, human trafficking, planning terrorist attacks, overthrowing governments, sparking genocides, promoting open lies and naked defamation. All good. No problem.
No matter what happens on a social media site, Congress said, its owners and managers bear no responsibility whatsoever for what’s published or distributed on and from the site. None. They’re untouchable. They take the profits but never have to worry about being answerable for the criminality and damage their site is doing.
Sounds crazy, right?
But that’s almost exactly what Congress did with Section 230 of the Communications Decency Act of 1996 in what they thought, at the time, was a good-faith effort to help the brand-new internet grow in a way they hoped would eventually become an important and useful social good.
It sure hasn’t worked out that way. And, as noted, it wasn’t at all this way in the years before 1996.
Back when the internet started, but before hypertext markup language (HTML) was invented, there were really only two big “houses” on the internet: CompuServe and AOL. My old friend and business partner Nigel Peacock and I ran “forums” on CompuServe starting around 1979, right up until 1996.
We ran the IBM PC Forum, the Macintosh Forum, and about two dozen other “social media forums” where people interested in ADHD, UFOs, the Kennedy assassination, international trade, spirituality, and a bunch of other topics were discussed.
CompuServe paid us well, because we had to make sure nothing criminal happened in any of the forums we ran. We even had to carry our own liability insurance. And we split that revenue with the 20 or so people who worked with us.
We kept those places open and safe, as did hundreds of other “Sysops” or Systems Operators who ran other forums on CompuServe and AOL. After all, it was CompuServe’s and AOL’s “publications” or “bookstores,” and they were paying us to make sure nothing illegal happened inside them.
Until 1996, that is. That year, after Section 230 became law, CompuServe decided they no longer needed to pay Sysop moderators to keep their forums clean and crime-free, so they quit paying us. Most of us left.
The result of Section 230 of the Communications Decency Act of 1996 is obvious today. The attack on our Capitol was largely planned on social media, where you can also buy ghost guns, other people’s credit card numbers, drugs, and illegal porn.
Social media sites now run algorithms that choose specific content from users to push out to other users to keep them “engaged” so they’ll have maximum exposure to advertisements, which have made the owners of the sites into billionaires.
In 1997, in the case Zeran v. America Online, the Fourth Circuit Court of Appeals ruled that Section 230 is written so tightly that even when an online service knowingly allows lawbreaking, they can’t be held accountable.
More recently, last month in Moody v NetChoice, the Supreme Court held that social media companies are also protected, like newspaper publishers are, by the First Amendment. They essentially affirmed Zeran, writing:
“[A} platform’s algorithm that reflects ‘editorial judgments’ about ‘compiling the third-party speech it wants in the way it wants’ is the platform’s own ‘expressive product’ and is therefore protected by the First Amendment.”
Mark Zuckerberg, who owns one of those “publications,” has become one of the richest men on the planet because he doesn’t have to pay for content moderation. Twitter made a few billionaires, too, before Elon Musk turned it into a rightwing disinformation machine.
Nonetheless, Section 230 lives on. I wrote a book that covers it, The Hidden History of Big Brother: How the Death of Privacy and the Rise of Surveillance Threaten Us and Our Democracy.
So did Josh Hawley, the Republican Senator from Missouri who hopes to be the next Trumpy president, and his book’s take is pretty much the same as mine: section 230 is extremely problematic, at the very least.
Which brings us to this week’s big news. For the first time, a federal appeals court (the Third Circuit) has ruled that because Section 230 largely deals with social media sites as “publishers,” it does not protect them as “distributors” (like bookstores).
In this case, a 10-year-old girl received a TikTok video — pushed out to her by that company’s algorithm — for a thing called the “blackout challenge” where people see how long they can cut off their own breathing or blood supply before blacking out. Little Nylah Anderson tried the challenge and accidentally asphyxiated herself to death.
Her mother sued in Pennsylvania for negligence and wrongful death, using state product liability laws as the basis for her suit. From there it went to federal court, where Anderson v ByteDance ended up before the Third Circuit.
Two Trump-appointed and one Obama-appointed judges ruled unanimously that ByteDance, which owns TikTok, isn’t “publishing” a social media site but — because their algorithm “curates” content and sends its choices, unsolicited, to users on the site — is actually “distributing” content.
In other words, social media sites are bookstores, not newspapers. And online “bookstores,” they ruled, are not protected by Section 230.
The case is far from settled; from here it’ll go to the Supreme Court, and its fate there is hard to predict given the Court’s embrace of the First Amendment argument in previous cases.
And the social media companies, raking in billions in profits, have so far stymied all congressional efforts to do away with Section 230 or regulate their behavior by taking advantage of the Supreme Court’s legalization of political bribery: Silicon Valley is one of the larger players in the lobbying and campaign contributions game.
Nonetheless, there’s now a very real possibility, even absent of congressional action, that social media companies may end up back where AOL and CompuServe were before 1996, having to hire thousands of content moderators to protect themselves from both criminal and civil liability. They’d also, no doubt, use AI to do much of the job.
Europe is moving in this direction too, with the arrest in France last week of the Russian founder of Telegram, a social media channel where human trafficking and other criminal activity were both common and known to the Systems Operators. He’s being charged with complicity in that activity on Telegram, even though he didn’t directly participate.
If Zuckerberg and his peers have to start hiring people to do what Nigel and I did for CompuServe years ago, it may reduce their income from tens of billions to mere billions. It certainly won’t make a difference in their lifestyles.
And society will be the better for it, our political landscape will stabilize, and fewer children will die.
That’s a tradeoff that’s well worth making. But why wait for the Supreme Court (which may not agree with the Third Circuit)? Congress can also repeal or substantially rewrite Section 230. And there’s a growing bipartisan consensus around that effort.
If you agree, call your members of Congress at 202-224-3121 and let them know you believe it’s time to consign Section 230 to the trash heap of history and begin to hold accountable and regulate these toxic behemoths to the same simple standards to which magazines, newspapers, newsletters like this, and bookstores must adhere.
We’re talking about our children and our democracy, after all. Section 230 has put both at risk, and its only “upside” has been to create social media billionaires. It’s time to return to rational regulation of social media and that begins with gutting or ending altogether Section 230.