Skip to main content

The context you need, when you need it

When news breaks, you need to understand what actually matters — and what to do about it. At Vox, our mission to help you make sense of the world has never been more vital. But we can’t do it on our own.

We rely on readers like you to fund our journalism. Will you support our work and become a Vox Member today?

Join now

Misinformation is winning the war on misinformation 

Online falsehoods are as bad as they‘ve ever been. Does anyone care?

US-PANDEMIC-VACCINE-MISINFORMATION-HEALTH
US-PANDEMIC-VACCINE-MISINFORMATION-HEALTH
Online misinformation about everything from elections to vaccines still finds audiences online.
Chris Delmas/AFP via Getty Images
A.W. Ohlheiser
A.W. Ohlheiser is a senior technology reporter at Vox, writing about the impact of technology on humans and society. They have also covered online culture and misinformation at the Washington Post, Slate, and the Columbia Journalism Review, among other places. They have an MA in religious studies and journalism from NYU.

Misinformation on the internet has never been worse. Or at least that’s my analysis of it, based on vibes.

People on TikTok are eating up videos saying a bunch of inaccurate things about the dangers of sunscreen, while the platform’s on-app Shop propels obscure books containing bogus cures for cancer onto the Amazon bestseller list. Meanwhile, the presumed Republican nominee for president is fresh off what appears to be a successful push to neuter efforts to address disinformation campaigns about the election. Also, Google’s AI Overview search results told people to put glue on pizza.

But this is all anecdotal. Can I prove my hunch with data? Sadly, no. The data I — or more accurately, researchers with actual expertise on this — would need to do that is locked behind the opaque doors of the companies that run the platforms and services on which the internet’s worst nonsense is hosted. Evaluating the reach of misinformation, in the present day, is a grueling and indirect process with imperfect results.

For my final newsletter contribution, I wanted to find a way to assess the state of misinformation online. As I’ve been covering this topic over and over for the past while, there’s one question that keeps popping into my head: Do companies like Google, Meta, and TikTok even care about meaningfully tackling this problem?

The answer to this question, too, is imperfect. But there are some things that might lead to an educated guess.

Ways to measure misinformation are disappearing

One of the most important things a journalist can do while writing about the spread of bad information online is to find a way to measure its reach. There’s a huge difference between a YouTube video with 1,000 views, and one with 16 million, for instance. But lately, some of the key metrics used to put supposedly “viral” misinformation into context have been disappearing from public view.

TikTok disabled view counts for popular hashtags earlier this year, shifting instead to simply showing the number of posts made on TikTok using the hashtag. Meta is shutting down CrowdTangle, a once-great tool for researchers and journalists looking to closely examine how information spreads across social media platforms, in August. That’s just a couple months before the 2024 election. And Elon Musk decided to make “likes” private on the platform, a decision that, to be fair, is bad for accountability but could have some benefits for normal users of X.

Between all this and declining access to platform APIs, researchers are limited in how much they can really track or speak to what’s going on.

“How do we track things over time? Apart from relying on the platform’s word,” said Ananya Sen, an assistant professor of information technology and management at Carnegie Mellon University, whose recent research looks at how companies inadvertently fund misinformation-laden sites when they use large ad tech platforms.

Disappearing metrics is basically the opposite of what a lot of experts on manipulated information recommend. Transparency and disclosure are “key” components of reform efforts like the Digital Services Act in the EU, said Yacine Jernite, machine learning and society lead for Hugging Face, an open-source data science and machine learning platform.

“We’ve seen that people who use [generative AI] services for information about elections may get misleading outputs,” Jernite added, “so it’s particularly important to accurately represent and avoid over-hyping the reliability of those services.”

It’s generally better for an information ecosystem when people know more about what they’re using and how it works. And while some aspects of this fall under media literacy and information hygiene efforts, a portion of this has to come from the platforms and their boosters. Hyping up an AI chatbot as a next-generation search tool sets expectations that aren’t fulfilled by the service itself.

Platforms don’t have much incentive to care

Platforms aren’t just amplifying bad information, they’re making money off it. From TikTok Shop purchases to ad sales, if these companies take meaningful, systemic steps to change how disinformation circulates on their platforms, they might work against their business interests.

Social media platforms are designed to show you things you want to engage with and share. AI chatbots are designed to give the illusion of knowledge and research. But neither of these models are great for evaluating veracity, and doing so often requires limiting the scope of a platform working as intended. Slowing or narrowing how a platform like this works means less engagement, which means no growth, which means less money.

“I personally can’t imagine that they would ever be as aggressively interested in addressing this as the rest of us are,” said Evan Thornburg, a bioethicist who posts on TikTok as @gaygtownbae. “The thing that they’re able to monetize is our attention, our interest, and our buying power. And why would they whittle that down to a narrow scope?“

Many platforms begrudgingly began efforts to take on misinformation after the 2016 US elections, and again at the beginning of the Covid pandemic. But since then, there’s been kind of a pullback. Meta laid off employees from teams involved with content moderation in 2023, and rolled back its Covid-era rules. Maybe they’re sick of being held responsible for this stuff at this point. Or, as technology changes, they see an opportunity to move on from it.

So do they care?

Again, it’s hard to quantify the efforts by major platforms to curb misinformation, which leaves me leaning once again on informed vibes. For me, it feels like major platforms are backing away from prioritizing the fight against misinformation and disinformation, and that there’s a general kind of fatigue out there on the topic more broadly. That doesn’t mean that nobody is doing anything.

Prebunking, which involves preemptively fact-checking rumors and lies before they gain traction, is super promising, especially when applied to election misinformation. Crowdsourced fact-checking is also an interesting approach. And to the credit of platforms themselves, they do continue to update their rules as new problems emerge.

There’s a way in which I have some sympathy for the platforms here. This is an exhausting topic, and it’s tough to be told, over and over, that you’re not doing enough. But pulling back and moving on doesn’t stop bad information from finding audiences over and over. While these companies assess how much they care about moderating and addressing their platform’s capacity to spread lies, the people targeted by those lies are getting hurt.

More in Technology

Technology
The case for AI realismThe case for AI realism
Technology

AI isn’t going to be the end of the world — no matter what this documentary sometimes argues.

By Shayna Korol
Politics
OpenAI’s oddly socialist, wildly hypocritical new economic agendaOpenAI’s oddly socialist, wildly hypocritical new economic agenda
Politics

The AI company released a set of highly progressive policy ideas. There’s just one small problem.

By Eric Levitz
Future Perfect
Human bodies aren’t ready to travel to Mars. Space medicine can help.Human bodies aren’t ready to travel to Mars. Space medicine can help.
Future Perfect

Protecting astronauts in space — and maybe even Mars — will help transform health on Earth.

By Shayna Korol
Podcasts
The importance of space toilets, explainedThe importance of space toilets, explained
Podcast
Podcasts

Houston, we have a plumbing problem.

By Peter Balonon-Rosen and Sean Rameswaram
Technology
What happened when they installed ChatGPT on a nuclear supercomputerWhat happened when they installed ChatGPT on a nuclear supercomputer
Technology

How they’re using AI at the lab that created the atom bomb.

By Joshua Keating
Future Perfect
Humanity’s return to the moon is a deeply religious missionHumanity’s return to the moon is a deeply religious mission
Future Perfect

Space barons like Jeff Bezos and Elon Musk don’t seem religious. But their quest to colonize outer space is.

By Sigal Samuel