Community Notes Is A Useful Tool For Some Things… But Not As A Full Replacement For Trust & Safety

from the it's-a-tool,-not-the-tool dept

When Twitter first launched what it called “Birdwatch,” I was hopeful that it would turn into a useful alternative approach to helping with trust & safety/content moderation questions, but I noted that there were many open questions, in particular with how it would deal with malicious actors seeking to game the system. When Elon took over Twitter, he really seemed to embrace Birdwatch, though he changed the name to the pointlessly boring “Community Notes.”

I still think the concept is a good one, and think it’s one of Elon’s few good moves. I think other social media sites should experiment with some similar ideas as well.

The problem, though, is that Elon seems to think that Community Notes is an effective replacement for a comprehensive trust & safety program. At the heart of so many of Elon’s decisions in firing the vast majority of the company’s trust & safety staff was that “Community Notes can handle it.”

As we’re in the midst of a series of major crises around the globe, where the flow of information has proven incredibly important, one thing we’re clearly learning is that Community Notes is not up to the task. Just to drive this point home, over the weekend Elon himself posted some fucking nonsense (as he’s prone to do) and many hours later Community Notes pointed out it was hogwash. Elon, as he’s done in the past when he’s been “Noted,” claimed he was happy it happened to himself… before claiming that his post was “obviously a joke meme” and that “there is more than a grain of truth to it.”

Image

So, first of all, there isn’t “more than a grain of truth to it.” The whole thing is simply false. But, more importantly, if you look at the top replies to his “obviously a joke meme,” suggests that Elon’s biggest fans did not, even remotely, think that this was “obviously a joke meme,” but rather took it entirely seriously, cheering him on for “telling the truth.” Here’s just one of the top replies to his original tweet:

Image

Also, it took quite some time for the note to appear on Elon’s account. And, look, content moderation at scale is impossible to do well and all that, but Community Notes seems like the exact wrong approach in situations like this one. Especially at a time when the accounts pushing out the most viewed news these days seem to be made up by a combination of grifters and idiots:

Online we have seen many users of X describe their experience of this crisis as different. Some of that may result from the more ambiguous nature of the larger conflict, especially as the news cycle moves from the unambiguous horror of the initial attack to concerns about Israel’s response. However, our investigation here suggests an additional factor: in Musk’s short tenure as owner of the platform, a new set of news elites has emerged. These elites post frequently, many sharing unvetted content and emotionally charged media. While sharing no single political ideology, many embrace a similar culture of rapid production of unlinked or ambiguously sourced content, embracing a “firehose of media” ethos that places the onus of verification on the end-user. This occurs in an environment that has been shorn of many of the “credibility signals” that served to ground users in the past — checkmarks that indicated notability, fact-checks distributed through Twitter Trends, and Twitter/X-based labeling of deceptive content. Even fundamental affordances of the web — such as simple sourcing through links — have been devalued by the platform, and, perhaps as a result, by the new elites that now direct its users’ attention.

Leaving aside the significant concern of taking away professional, trained trust & safety employees, and replacing them with random (often hand-picked) untrained volunteers, there are serious concerns coming to light about how Community Notes actually works in practice.

Multiple reports have come out lately highlighting the limitations of Community Notes on important breaking news in the midst of various conflicts around the world, where you have malicious actors seeking to deliberately spread misinformation. A report at Wired found that Community Notes is actually making some of the problems worse, rather than better.

On Saturday, the company wrote on its own platform that “notes across the platform are now being seen tens of millions of times per day, generating north of 85 million impressions in the last week.” It added that thousands of new contributors had been enrolled in the system. However, a WIRED investigation found that Community Notes appears to be not functioning as designed, may be vulnerable to coordinated manipulation by outside groups, and lacks transparency about how notes are approved. Sources also claim that it is filled with in-fighting and disinformation, and there appears to be no real oversight from the company itself.

“I understand why they do it, but it doesn’t do anything like what they say it does,” one Community Notes contributor tells WIRED. “It is prone to manipulation, and it is far too slow and cumbersome. It serves no purpose as far as I can see. I think it’s probably making the disinformation worse, to be honest.”

The report isn’t just based on random Community Notes users, but looking more closely at how the program works, and the ability for it to be gamed. Wired found that it wasn’t difficult to set up multiple accounts controlled by one person which all had access to Community Notes, meaning that you could manipulate support for a position with just a small group of users controlling multiple accounts.

It also points to earlier (pre-Elon) research that showed that (then) Birdwatch wasn’t used nearly as much for standard fact checking, but was used in political debates by users who disagreed politically with someone who had tweeted.

Back during the summer, the Poynter Institute had a good analysis of the limitations of Community Notes for dealing with real-time misinformation campaigns during crises. Specifically, the design of the current Community Notes has some, well, questionable assumptions built in. Apparently, it looks over your tweeting history and assigns you to a camp as being either “left” or “right” and then only allows a Community Note to go public if enough of the “left” people and the “right” people agree on a note.

“It has to have ideological consensus,” he said. “That means people on the left and people on the right have to agree that that note must be appended to that tweet.”

Essentially, it requires a “cross-ideological agreement on truth,” and in an increasingly partisan environment, achieving that consensus is almost impossible, he said. 

Another complicating factor is the fact that a Twitter algorithm is looking at a user’s past behavior to determine their political leanings, Mahadevan said. Twitter waits until a similar number of people on the political right and left have agreed to attach a public Community Note to a tweet. 

While that may work on issues where there isn’t any kind of culture war, it’s completely useless for culture war issues, where plenty of disinformation flows. Indeed, the Poynter report notes that a huge percentage of the highest rated Community Notes inside the Community Notes system are never seen by the public because they don’t have “cross-ideological agreement.”

The problem is that regular Twitter users might never see that note. Sixty percent of the most-rated notes are not public, meaning the Community Notes on “the tweets that most need a Community Note” aren’t public, Mahadevan said. 

The setup with “cross-ideological” consensus basically seems almost perfectly designed to make sure that the absolute worst nonsense will never have Community Notes shown publicly.

Meanwhile, a report from NBC News also highlights how even when Community Notes is able to help debunk false information, it often comes way too late.

NBC News focused on two prominent pieces of Israel-Hamas misinformation that have already been debunked: a fake White House news release that was posted to X claiming the Biden administration had granted Israel $8 billion in emergency aid and false reports that St. Porphyrius Orthodox Church in Gaza was destroyed.

Only 8% of 120 posts related to those stories had published community notes, while 26% had unpublished notes from volunteers that had yet to be approved. About two-thirds of the top posts NBC News reviewed had no proposed or published Community Notes on them.

The findings echo what a Community Notes volunteer said was X’s lack of response to efforts to debunk misleading posts.

“All weekend we were furiously vetting, writing, and approving Community Notes on hundreds of posts which were demonstrably fake news,” Kim Picazio, a Community Notes volunteer, wrote on Instagram’s Threads. “It took 2+ days for the backroom to press whatever button to finally make all our warnings publicly viewable. By that time… You know the rest of that sentence.”

And when the Community Notes don’t show up until much later, a ton of nonsense can spread:

A post about the debunked White House news release published by a verified account had nearly 500,000 views and no proposed or appended note Tuesday afternoon.The Community Notes system also showed that a user tried to submit a fact-check Sunday on another post including the same known misinformation but that it had yet to be approved, saying, “Needs more ratings.” The post had accrued 80,000 views since Sunday.

In a search for St. Porphyrius Orthodox Church in Gaza, only five Community Notes had been applied to the top 42 posts echoing the debunked misinformation. Several posts from verified users with no notes repeated the claim and got over 100,000 views, while 13 Community Notes had been proposed on posts of the debunked claims but had not yet been approved for publishing.

Another deep dive look at how Community Notes handled the first few days of the Israel/Palestine mess showed just how ineffective it was:

During the first 5 days of the conflict, just 438 Community Notes (attached to 309 posts from 223 unique accounts) earned a “HELPFUL” rating and ended up being displayed publicly to users. Although it’s impossible to know what percentage of content about the war this represents, the fact that trending topics related to the conflict have routinely involved hundreds of thousands or even millions of posts suggests that a few hundred posts is just a drop in the bucket. The visible notes were generally attached to popular posts — the 309 posts in question earned a combined total of 2147081 likes, an average of 6948 likes per post. The majority of the posts that earned Community Notes (222 of 309 posts, 71.8%) came from paid X Premium/Twitter Blue subscribers, and the majority of the accounts posting them (147 of 223, 65.9%) are X Premium subscribers, who are potentially earning a share of X’s ad revenue based on the number of times their posts are seen and who therefore have a financial motive to never delete misleading content. (Overall, roughly 7% of posts that received Community Notes were deleted during the period studied, but there’s no reliable way of knowing how many of these posts were related to the Israel/Hamas war.)

Again, I really like the concept of Community Notes. I think it’s a very useful tool — and one example (of many) of trust & safety tools beyond simply “taking down” content. But it needs to be part of a wider strategy, not the only strategy. And, the program can’t be setup with such a huge blindspot for culture war issues.

But, that’s exactly how things currently work, and it’s a shame, in part because I fear it’s going to discourage others from creating their own versions of Community Notes.

Filed Under: , , , , , ,
Companies: twitter, x

Rate this comment as insightful
Rate this comment as funny
You have rated this comment as insightful
You have rated this comment as funny
Flag this comment as abusive/trolling/spam
You have flagged this comment
The first word has already been claimed
The last word has already been claimed
Insightful Lightbulb icon Funny Laughing icon Abusive/trolling/spam Flag icon Insightful badge Lightbulb icon Funny badge Laughing icon Comments icon

Comments on “Community Notes Is A Useful Tool For Some Things… But Not As A Full Replacement For Trust & Safety”

Subscribe: RSS Leave a comment
24 Comments
This comment has been deemed insightful by the community.
Somewhat Less Anonymous Coward (profile) says:

Somehow i get the feeling that the fundamental shortcoming of the community notes – that it’s a voting system that cares little (or possibly not at all) about actual authority on any subject – is exactly why it’s pushed by Musk as an end all be all solution. The chaos that he and those he allies with are creating is essentially a mob rule, one they hope to be leading. This system is exactly that – a large enough crowd can drown anything, regardless of how truthful is is.

This comment has been flagged by the community. Click here to show it.

This comment has been flagged by the community. Click here to show it.

Matthew Bennett says:

Re: Re:

Ah, my autistic Bar Weiss.

They weren’t censors,

Yes, they were.

and the issue is with dealing with objectively wrong information.

1) Not “objective” anything since people obviously diagreed.
2) they shadow-banned real doctors saying things that turned out to be objectively true
3) This supposed “misinformation” (much of it was not that) they dealt with via censorship, you idiot.

And that dystopian censorship regime, using the excuse of supposed “misinformation” and “hate speech”, regardless of whether anyone else thinks those accusations are accurate, is exactly what Masnick wants back.

No. Fuck you.

Strawb (profile) says:

Re: Re: Re:

Yes, they were.

No, they weren’t.
Now you.

1) Not “objective” anything since people obviously diagreed.

People disagree about the shape of the planet we live on, too. That doesn’t suddenly make it subjective.

2) they shadow-banned real doctors saying things that turned out to be objectively true

Example?

3) This supposed “misinformation” (much of it was not that) they dealt with via censorship, you idiot.

How did they stop people from saying the same stuff that got them suspended/banned on other platforms? Because that’s what censorship is.

And that dystopian censorship regime, using the excuse of supposed “misinformation” and “hate speech”, regardless of whether anyone else thinks those accusations are accurate, is exactly what Masnick wants back.

No, that’s just your MDS talking again.

bhull242 (profile) says:

Re: Re: Re:

1) Not “objective” anything since people obviously diagreed.

People can be objectively wrong. People are also capable of lying. Your argument is invalid.

2) they shadow-banned real doctors saying things that turned out to be objectively true

That’s not the moderation I’m talking about, but at any rate, that happens with any kind of moderation done at scale. It is fundamentally impossible to avoid either false positives or false negatives.

Now, what I was referring to was things like people making up fake news websites that were made to look like they came from, say, CBS News even though they weren’t, and stuff like whether it was Israel, Hamas, or someone else whose rocket ended up in the parking lot of a Palestinian hospital. Obviously, that wasn’t the only moderation going on, but that’s the sort of thing you’d need people to moderate.

3) This supposed “misinformation” (much of it was not that)

Again, I’m only focusing on the stuff that’s more clear-cut. Obviously, it was far from perfect; I’m not interested in discussing that. The fact is that there was objective disinformation, and that needs to be handled somehow.

they dealt with via censorship, you idiot.

I really don’t care to play semantic games right now. It really doesn’t matter what you call it.

And that dystopian censorship regime, using the excuse of supposed “misinformation” and “hate speech”, regardless of whether anyone else thinks those accusations are accurate, is exactly what Masnick wants back.

Here’s the thing; Mike is saying that that is better for Twitter’s business and a number of its users than what Elon is doing. He has said multiple times that Dorsey’s Twitter was far from perfect, so he’s not saying that that’s exactly how he wants it to be. He’s just comparing two different ways of handling it and expressing which is better (in his opinion or in terms of making money).

Also, there’s nothing “dystopian” about it. Facebook was worse than Twitter using your measures as Twitter was far more leery about bans or removing posts, but that’s not a dystopia. You’re exaggerating what actually happened.

As I see it, the difference between you and Mike on this is how much you each weigh false positives and false negatives in moderation by social media companies (whether or not you want to call it “censorship”, we both appear to agree that it was a form of moderation). Apparently, you think that any false positives are unacceptable, and reducing false positives far outweighs reducing false negatives. Mike (along with others, myself included) clearly disagrees with that. He’s not on the opposite side of the spectrum; he just prefers more of a balance between the two.

The sooner you can acknowledge that much, the sooner we can get anywhere with this.

This comment has been flagged by the community. Click here to show it.

Anonymous Coward says:

I really like the concept of Community Notes

Until the “community” is… the Twitter community. It’s not the first time (of countless times) that misinformation has plagued Twitter, and now virtually everybody can have the audience he deserves he paid for, community notes seem more like cool aid on black plague, putting a tweet in context, but drown in dozen of unrelated tweets, links, images, videos, scam, ads…
It’ll be like Reddit, where a few motivated admins would just give up because the service is asking too much from them, and in only one direction.

Somewhat Less Anonymous Coward (profile) says:

Re:

The note from the screenshot did not make through voting. However, Imgur post does contain a link to a Xeet which originated the pic, and this is the only reply to it with likes:

Luke loves Nazis and snuggles Commies

Of note, the reply is by someone with the blue tick. At a glance, this appears to be the average quality of the replies to the man’s Xeets in which he criticizes Elmo.
“Tell me who your friends are, and I will tell you who you are” seems to be working in Musk’s case, his only defenders being toxic assholes or worse.

This comment has been flagged by the community. Click here to show it.

Matthew Bennett says:

It's not "Trust and Saftey", it's censorship.

Also very explictly stifling of dissent. Giving it an Orwellian name doesn’t make it any better yopu gaslighting tool.

Community Notes is 100x better than the dystopian “let your betters decide what is true” bull crap that you want.

Musk said a stupid thing and got told. That’s great! It took about half a day, also great! It’s also an objectively fair and nonpartisan system uninfluenced by government, and the former system was none of those things.

It’s meme, grow up, people get to say things you think are wrong. It’s not like an emergency that it took a few hours to get tagged.

It also points to earlier (pre-Elon) research that showed that (then) Birdwatch

Yeah, I get it, you want to compare Community Notes that is working great to a similar system that worked less great so that you can establish some guilt by association. But the fact is CN is working great.

But it needs to be part of a wider strategy, not the only strategy.

No, thank you, we tried that and it sucked. It led to ideological filtering, silencing of dissent, and censorship at gov request. And you continue to lie about all of that but we’re not going back. Even FB has largely ditched its “fact-checking” if you didn’t notice, cuz it didn’t work, like at all.

You’re not getting the Gadde and Roth Ministry of Truth back, go pound sand.

This comment has been flagged by the community. Click here to show it.

This comment has been flagged by the community. Click here to show it.

Benjamin Jay Barber says:

Mike Masnick Malding Again

Everyone who is up to current events, and knows how to read a map, knows that the Taliban control Afghanistan. The average person knows that the US has been intervening in the middle east, and invaded both Iraq, Afghanistan and Syria, and gave weapons to Saddam to attack Iran with.

The fact that some people are idiots, or that you think the meme is hyperbolic, is only because you don’t know how to meme. The same can have been said since the beginning of political cartooning.

Add Your Comment

Your email address will not be published. Required fields are marked *

Have a Techdirt Account? Sign in now. Want one? Register here

Comment Options:

Make this the or (get credits or sign in to see balance) what's this?

What's this?

Techdirt community members with Techdirt Credits can spotlight a comment as either the "First Word" or "Last Word" on a particular comment thread. Credits can be purchased at the Techdirt Insider Shop »

Follow Techdirt

Techdirt Daily Newsletter

Ctrl-Alt-Speech

A weekly news podcast from
Mike Masnick & Ben Whitelaw

Subscribe now to Ctrl-Alt-Speech »
Techdirt Deals
Techdirt Insider Discord
The latest chatter on the Techdirt Insider Discord channel...
Loading...