Online Speech Is Now An Existential Question For Tech
Share Button

Online Speech Is Now An Existential Question For Tech

Content moderation rules used to be a question of taste. Now, they can determine a service’s prospects for survival.

By Christopher Mims
Wed, Feb 24, 2021 3:15amGrey Clock 6 min

Every public communication platform you can name—from Facebook, Twitter and YouTube to Parler, Pinterest and Discord—is wrestling with the same two questions:

How do we make sure we’re not facilitating misinformation, violence, fraud or hate speech?

At the same time, how do we ensure we’re not censoring users?

The more they moderate content, the more criticism they experience from those who think they’re over-moderating. At the same time, any statement on a fresh round of moderation provokes some to point out objectionable content that remains. Like any question of editorial or legal judgment, the results are guaranteed to displease someone, somewhere—including Congress, which this week called the chief executives of Facebook, Google and Twitter to a hearing on March 25 to discuss misinformation on their platforms.

For many services, this has gone beyond a matter of user experience, or growth rates, or even ad revenue. It’s become an existential crisis. While dialling up moderation won’t solve all of a platform’s problems, a look at the current winners and losers suggests that not moderating enough is a recipe for extinction.

Facebook is currently wrestling with whether it will continue its ban of former president Donald Trump. Pew Research says 78% of Republicans opposed the ban, which has contributed to the view of many in Congress that Facebook’s censorship of conservative speech justifies breaking up the company—something a decade of privacy scandals couldn’t do.

Parler, a haven for right-wing users who feel alienated by mainstream social media, was taken down by its cloud service provider, Amazon Web Services, after some of its users live-streamed the riot at the U.S. Capitol on Jan. 6. Amazon cited Parler’s apparent inability to police content that incites violence. While Parler is back online with a new service provider, it’s unclear if it has the infrastructure to serve a large audience.

During the weeks Parler was offline, the company implemented algorithmic filtering for a few content types, including threats and incitement, says a company spokesman. The company also has an automatic filter for “trolling” that detects such content, but it’s up to users whether to turn it on or not. In addition, those who choose to troll on Parler are not penalized in Parler’s algorithms for doing so, “in the spirit of First Amendment,” says the company’s guidelines for enforcement of its content moderation policies. Parler recently fired its CEO, who said he experienced resistance to his vision for the service, including how it should be moderated.

Now, just about every site that hosts user-generated content is carefully weighing the costs and benefits of updating their content moderation systems, using a mix of human professionals, algorithms and users. Some are even building rules into their services to pre-empt the need for increasingly costly moderation.

The saga of gaming-focused messaging app Discord is instructive: In 2018, the service, which is aimed at children and young adults, was one of those used to plan the Charlottesville riots. A year later, the site was still taking what appeared to be a deliberately laissez-faire approach to content moderation.

By this January, however, spurred by reports of hate speech and lurking child predators, Discord had done a complete 180. It now has a team of machine-learning engineers building systems to scan the service for unacceptable uses, and has assigned 15% of its overall staff to trust and safety issues.

This newfound attention to content moderation helped keep Discord away from the controversy surrounding the Capitol riot, and caused it to briefly ban a chat group associated with WallStreetBets during the GameStop stock runup. Discord’s valuation doubled to $7 billion over roughly the same period, a validation that investors have confidence in its moderation strategy.

The prevalence problem

The challenge successful platforms face is moderating content “at scale,” across millions or billions of pieces of shared content.

Before any action can be taken, services must decide what should be taken down, an often slow and deliberative process.

Imagine, for example, that a grass-roots movement gains momentum in a country, and begins espousing extreme and potentially dangerous ideas on social media. While some language might be caught by algorithms immediately, a decision about whether discussion of a particular movement, like QAnon, should be banned completely, could take months on a service such as YouTube, says a Google spokesman.

One reason it can take so long is the global nature of these platforms. Google’s policy team might consult with experts in order to consider regional sensitivities before making a decision. After a policy decision is made, the platform has to train AI and write rules for human moderators to enforce it—then make sure both are carrying out the policies as intended, he adds.

While AI systems can be trained to catch individual pieces of problematic content, they’re often blind to the broader meaning of a body of posts, says Tracy Chou, founder of content-moderation startup Block Party and former tech lead at Pinterest.

Take the case of the “Stop the Steal” protest, which led to the deadly attack on the U.S. Capitol. Individual messages used to plan the attack, like “Let’s meet at location X,” would probably look innocent to a machine-learning system, says Ms Chou, but “the context is what’s key.” Facebook banned all content mentioning “Stop the Steal” after the riot.

Even after Facebook has identified a particular type of content as harmful, why does it seem constitutionally unable to keep it off its platform?

It’s the “prevalence problem.” On a truly gigantic service, even if only a tiny fraction of content is problematic, it can still reach millions of people. Facebook has started publishing a quarterly report on its community standards enforcement. During the last quarter of 2020, Facebook says users saw seven or eight pieces of hate speech out of every 10,000 views of content. That’s down from 10 or 11 pieces the previous quarter. The company said it will begin allowing third-party audits of these claims this year.

While Facebook has been leaning heavily on AI to moderate content, especially during the pandemic, it currently has about 15,000 human moderators. And since every new moderator comes with a fixed additional cost, the company has been seeking more efficient ways for its AI and existing humans to work together.

In the past, human moderators reviewed content flagged by machine learning algorithms in more or less chronological order. Content is now sorted by a number of factors, including how quickly it’s spreading on the site, says a Facebook spokesman. If the goal is to reduce the number of times people see harmful content, the most viral stuff should be top priority.

A content moderator in every pot

Companies that aren’t Facebook or Google often lack the resources to field their own teams of moderators and machine-learning engineers. They have to consider what’s within their budget, which includes outsourcing the technical parts of content moderation to companies such as San Francisco-based startup Spectrum Labs.

Through its cloud-based service, Spectrum Labs shares insights it gathers from any one of its clients with all of them—which include Pinterest and Riot Games, maker of League of Legends—in order to filter everything from bad words and human trafficking to hate speech and harassment, says CEO Justin Davis.

Mr Davis says Spectrum Labs doesn’t say what clients should and shouldn’t ban. Beyond illegal content, every company decides for itself what it deems acceptable, he adds.

Pinterest, for example, has a mission rooted in “inspiration,” and this helps it take a clear stance in prohibiting harmful or objectionable content that violates its policies and doesn’t fit its mission, says a company spokeswoman.

Services are also attempting to reduce the content-moderation load by reducing the incentives or opportunity for bad behaviour. Pinterest, for example, has from its earliest days minimized the size and significance of comments, says Ms Chou, the former Pinterest engineer, in part by putting them in a smaller typeface and making them harder to find. This made comments less appealing to trolls and spammers, she adds.

The dating app Bumble only allows women to reach out to men. Flipping the script of a typical dating app has arguably made Bumble more welcoming for women, says Mr Davis, of Spectrum Labs. Bumble has other features designed to pre-emptively reduce or eliminate harassment, says Chief Product Officer Miles Norris, including a “super block” feature that builds a comprehensive digital dossier on banned users. This means that if, for example, banned users attempt to create a new account with a fresh email address, they can be detected and blocked based on other identifying features.

The ‘supreme court of content’

Facebook CEO Mark Zuckerberg recently described Facebook as something between a newspaper and a telecommunications company. For it to continue being a global town square, it doesn’t have the luxury of narrowly defining the kinds of content and interactions it will allow. For its toughest content moderation decisions, it has created a higher power—a financially independent “oversight board” that includes a retired U.S. federal judge, a former prime minister of Denmark and a Nobel Peace Prize laureate.

In its first decision, the board overturned four of the five bans Facebook brought before it.

Facebook has said that it intends the decisions made by its “supreme court of content” to become part of how it makes everyday decisions about what to allow on the site. That is, even though the board will make only a handful of decisions a year, these rulings will also apply when the same content is shared in a similar way. Even with that mechanism in place, it’s hard to imagine the board can get to more than a tiny fraction of the types of situations content moderators and their AI assistants must decide every day.

But the oversight board might accomplish the goal of shifting the blame for Facebook’s most momentous moderation decisions. For example, if the board rules to reinstate the account of former President Trump, Facebook could deflect criticism of the decision by noting it was made independent of its own company politics.

Meanwhile, Parler is back up, but it’s still banned from the Apple and Google app stores. Without those essential routes to users—and without web services as reliable as its former provider, Amazon—it seems unlikely that Parler can grow anywhere close to the rate it otherwise might have. It’s not clear yet whether Parler’s new content filtering algorithms will satisfy Google and Apple. How the company balances its enhanced moderation with its stated mission of being a “viewpoint neutral” service will determine whether it grows to be a viable alternative to Twitter and Facebook or remains a shadow of what it could be with such moderation.



MOST POPULAR

Rugged coastal drives and fireside drams define a slow, indulgent journey through Scotland’s far north.

A haven for hedge-fund titans and Hollywood grandees, Greenwich is one of the world’s most expensive residential enclaves, where eye-watering prices meet unapologetic grandeur.

Related Stories
Lifestyle
The Workers Opting to Retire Instead of Taking on AI
By Lauren Weber & Ray A. Smith 07/04/2026
Lifestyle
ROLLS-ROYCE UNVEILS YACHT-INSPIRED CULLINAN SERIES FOR BESPOKE CLIENTS
By Staff Writer 30/03/2026
Lifestyle
BMW’s Electric i3 and iX3 Raise the EV Standard With a 400-Plus-Mile Range
By Jim Motavalli 26/03/2026

Their careers spanned the personal computing, internet and smartphone waves. But some older workers see AI’s arrival as the cue to exit. 

By Lauren Weber & Ray A. Smith
Tue, Apr 7, 2026 4 min

Luke Michel has already lived through two technology overhauls in his career, first desktop publishing in the 1980s and online publishing later on. But AI? He’s had enough. 

So when his employer, the Dana-Farber Cancer Institute, made an early-retirement offer to some staff last year, the 68-year-old content strategist decided to speed up his exit. Before, he had expected to work a couple more years. 

“The time and energy you have to devote to learning a whole new vocabulary and a whole new skill set, it wasn’t worth it,” he said. 

It isn’t that he’s shunning artificial intelligence—he is learning Spanish with the help of Anthropic’s Claude. But, at this point, he’s less than eager to endure all the ways the technology promises to upend work. 

“I just want to use it for my own purposes and not someone else’s,” he said. 

After rising for decades and then hovering around 40% in the 2010s, the share of Americans over 55 years old in the workforce has slipped to 37.2%, the lowest level in more than 20 years.  

The financial cushion of rising home equity and stock-market returns is driving some of the decline, economists and retirement advisers say. 

But for some older professionals, money is only part of the equation.  

They say they don’t want to spend the last years of their career going through the tumult of AI adoption, which has brought new tools, new expectations and a lot of uncertainty.  

Many people retire when key elements of their work lives are disrupted at once, said Robert Laura , co-founder of the Retirement Coaches Association and an expert on the psychology of retirement. 

“Maybe their autonomy is being challenged or changed, their friends are leaving the workplace, or they disagree with the company’s direction,” he said.  

“When two or three of these things show up, that’s when people start to opt out.”  

“AI is a big one,” he adds. “It disrupts their autonomy, their professionalism.” 

Michel, whose work required overseeing and strategizing on website content, has been here before.  

When desktop publishing arrived in the 1980s, he was a graphic designer using triangles and rubber cement.  

The internet’s arrival changed everything again. Both developments required new skills, and he was energized by the challenge of learning alongside colleagues and peers. 

It felt different this time around. “Your battery doesn’t hold a charge as long as it used to,” he said. 

He would rather spend his energy volunteering, making art, going to operas and chairing the Council on Aging in North Andover, Mass., where he lives. 

In an AARP survey last summer of 5,000 people 50 and over, 25% of those who planned to retire sooner than expected counted work stress and burnout as factors.  

About half of those retired said they had left work at least partly because they had the financial security to do so. 

In general, older Americans are less likely than younger counterparts to use AI, research shows.  

About 30% of people from ages 30 to 49 said they used ChatGPT on the job, nearly double the share of those 50 and older, according to a 2025 Pew Research Center survey of more than 5,000 adults. 

Baby boomers and members of Generation X also experienced the sharpest declines in confidence using AI technology, according to a ManpowerGroup survey of more than 13,900 workers in 19 countries. 

“We as employers aren’t doing a good enough job saying (to older workers), we value the skills that you already have, so much so that we want to invest in you to help you do your job better,” says Becky Frankiewicz , ManpowerGroup’s chief strategy officer. 

Jennifer Kerns’s misgivings about AI contributed to her departure last month from GitHub, where the 60-year-old worked as a program manager.  

Coming from a family of artists, she said, it offends her that AI models train on the creative work of people who aren’t compensated for their intellectual property. And she worries about AI’s effect on people’s critical-thinking skills. 

So she was dismayed when GitHub, a Microsoft-owned hosting service for software projects, began investing heavily in AI products and expecting employees to incorporate AI into much of their work. In employee-engagement surveys, the company had begun asking them to rate their AI usage on a scale of 1 to 5. 

When it came time to write reports and reviews, colleagues would suggest that she use ChatGPT.  

“I’d be like, ‘I have no idea how to use that and I have no interest in using AI to write anything for me,’” she said. 

It would have been more prudent to work until she was closer to Medicare eligibility, she said. But by waiting until her children were out of college and some of her stock grants had vested, the math worked. 

Her first act as a nonworking person: a solo trip to Scotland, where she took a darning workshop and learned how to repair sweaters.  

“The opposite of AI,” she said. 

Employers already under pressure to cut workers—such as in the tech industry—may welcome some of these retirements, said Gad Levanon , chief economist at Burning Glass Institute, which studies labor-market data. 

“The more people retire, the fewer they have to let go,” he said. 

Some of the savviest tech users are also balking at sticking around for the AI upheaval. Terry Grimm, who worked in IT for 40 years, retired from his senior software consultant role at 65 last May.  

His firm had just been acquired by a bigger firm, which meant learning and integrating the parent company’s AI and other tech tools into his work.   

Until then, Grimm expected he might work a couple more years, though he felt that he probably had enough saved to retire. 

“I just got to the point where I was spending 40 hours at work and then 20 hours training and studying,” said Grimm, who has since moved with his wife from the Dallas area to a housing development on a golf course in El Dorado, Ark.  

“I’m like, ‘I’ll let the younger guys do this.’”