Two GCHQ Employees Suggest The Solution To CSAM Distribution Is… More Client-Side Scanning

from the offloading-law-enforcement-work-to-the-private-sector dept

The font of not-great ideas continues to overflow at Lawfare. To be fair, this often-overflowing font is due to its contributors, which are current and former members of spy agencies that have violated rights, broken laws, and otherwise done what they can to make internet communications less secure.

We’ve heard from these contributors before. Ian Levy and Crispin Robinson are both GCHQ employees. A few years ago, as companies like Facebook started tossing around the idea of end-to-end encryption, Levy and Robinson suggested a workaround that would have done the same amount of damage as mandated backdoors, even if the pitch was slightly different than the suggestions offered by consecutive FBI directors.

What was suggested then was some sort of parallel communication network that would allow spies and law enforcement to eavesdrop on communications. The communications would still be encrypted. It’s just that the “good guys” would have their own encrypted channel to listen in on these communications. Theoretically, communications would still be secure, unable to be accessed by criminals. But opening a side door is not a whole lot different than opening a back door. A blind CC may be a bit more secure than undermining encryption entirely, but it’s still opens up another communication channel — one that might be left open and unguarded by the interceptors who would likely feel whatever bad things might result from that is acceptable because (lol) spy agencies only target dangerous enemies of the state.

The pair are back at it. In this post for Lawfare, Levy and Robinson suggest a “solution” that has already been proposed (and discarded) by the company that attempted it first: Apple. The “solution” is apparently trivially easy to exploit and prone to false positives/negatives, but that isn’t stopping these GCHQ reps from suggesting it be given another spin.

According to the paper [PDF] published by these two GCHQ employees, the key to fighting CSAM (Child Sexual Assault Material) in the era of end-to-end encryption is… more client-side scanning of content. And it goes beyond matching local images to known hashes stored by agencies that combat the sexual exploitation of children.

For example, one of the approaches we propose is to have language models running entirely locally on the client to detect language associated with grooming. If the model suggests that a conversation is heading toward a risky outcome, the potential victim is warned and nudged to report the conversation for human moderation. Since the models can be tested and the user is involved in the provider’s access to content, we do not believe this sort of approach attracts the same vulnerabilities as others.

Well, no vulnerabilities except for the provider’s access to what are supposed to be end-to-end encrypted communications. If this is the solution, the provider may as well not offer encryption at all, since it apparently won’t actually be encrypted at both ends. The provider will have access to the client side in some form, which opens a security hole that would not be present otherwise. The only mitigating factor is that the provider will not have its own copy of the communications. And if it doesn’t have that, of what use is it to law enforcement?

The proposal (which the authors note is not to be viewed as representative of GCHQ or the UK government) operates largely on faith.

[W]e believe that a robust evidence-based approach to this problem can lead to balanced solutions that ensure privacy and safety for all. We also believe that a framework for evaluating the benefits and disbenefits is needed.

“Disbenefits” is a cool word. If one were more intellectually honest, they might use a word like “drawback” or “flaw” or “negative side effects.” But that’s the Newspeak offered by the two GCHQ employees.

The following sentence makes it clear the authors don’t know whether any of their proposals will work, nor how many [cough] disbenefits they will cause. Just spitballing, I guess, but with the innate appeal to authority that comes from their positions and a tastefully-formatted PDF.

We don’t provide one in this paper but note that the U.K.’s national Research Centre on Privacy, Harm Reduction and Adversarial Influence Online (REPHRAIN) is doing so as part of the U.K. government’s Safety Tech Challenge Fund, although this will require interpretation in the context of national data protection laws and, in the U.K., guidance from the Information Commissioner’s Office.

[crickets.wav]

The authors do admit client-side scanning (whether of communications or content) is far from flawless. False negatives and false positives will be an ongoing problem. The system can easily be duped into ok’ing CSAM. That’s why they want to add client-side scanning of written communications to the mix, apparently in hopes that a combination of the two will reduce the “disbenefits.”

Supposedly this can be accomplished with tech magic crafted by people nerding harder and a system of checks and balances that will likely always remain theoretical, even if it’s hard-coded into moderation guidelines and law enforcement policies.

For example, offenders often send existing sexually explicit images of children to potential victims to try to engender trust (hoping that victims reciprocate by sending explicit images of themselves). In this case, there is no benefit whatsoever in an offender creating an image that is classified as child abuse material (but is not), since they are trying to affect the victim, not the system. This weakness could also be exploited by sending false-positive images to a target, hoping they are somehow investigated or tracked. This is mitigated by the reality of how the moderation and reporting process works, with multiple independent checks before any referral to law enforcement.

This simply assumes such “multiple independent checks” exist or will exist. They may not. It may be policy for tech companies to simply forward everything questionable to law enforcement and allow the “pros” to sort it out. That “solution” is easiest for tech companies and since they’ll be operating in good faith, legal culpability for adverse law enforcement reactions will be minimal.

That assumptive shrug that robust policies exist, will exist, or will be followed thousands of times a day leads directly into another incorrect assumption: that harm to innocent people will be mitigated because of largely theoretical checks and balances on both ends of the equation.

The second issue is that there is no way of proving which images a client-side scanning algorithm is seeking to detect, leaving the possibility of “mission creep” where other types of images (those not related to child sexual abuse) are also detected. We believe this is relatively simple to fix through a small change to how the global child protection non-governmental organizations operate. We would have a consistent list of known bad images, with cryptographic assurances that the databases contain only child sexual abuse images that can be attested to publicly and audited privately. We believe these legitimate privacy concerns can be mitigated technically and the legal and policy challenges are likely harder, but we believe they are soluble.

The thing is, we already have a “consistent list of known bad images.” If we’re not already doing the other things in that sentence (a verifiable database that can be “attested to publicly and audited privately”), then the only thing more client-side content scanning can do is produce more false positives and negatives. Again, the authors assume these things are already in place. And they use these assumptions to buttress their claims that the “disbenefits” will be limited by what they assume will happen (“multiple independent checks”) or assume has already happened (an independently verifiable database of known CSAM images).

That’s a big ask. The other big ask is the paper proposes the private sector do all of the work. Companies will be expected to design and implement client-side scanning. They will be expected to hire enough people to provide human backstops for AI-guided flagging of content. They will need to have specialized personnel in place to act as law enforcement liaisons. And they will need to have solid legal teams in place to deal with the blowback (I’m sorry, “disbenefits”) of false positives and negatives.

If all of this is in place, and law enforcement doesn’t engage in mission creep, it will work the way the authors suggest it will work: a non-encryption-breaking solution to the distribution of CSAM via end-to-end encrypted communications platforms. That’s not to say the paper does not admit all the pieces need to come together to make this work. But this proposal raises far more questions than it answers. And yet the authors seem to believe it will work because it’s merely possible.

Through our research, we’ve found no reason as to why client-side scanning techniques cannot be implemented safely in many of the situations society will encounter. That is not to say that more work is not needed, but there are clear paths to implementation that would seem to have the requisite effectiveness, privacy, and security properties.

It’s still a long way from probable, though. And that’s not even in the same neighborhood as theoretically possible if everything else goes right.

Filed Under: , , , ,

Rate this comment as insightful
Rate this comment as funny
You have rated this comment as insightful
You have rated this comment as funny
Flag this comment as abusive/trolling/spam
You have flagged this comment
The first word has already been claimed
The last word has already been claimed
Insightful Lightbulb icon Funny Laughing icon Abusive/trolling/spam Flag icon Insightful badge Lightbulb icon Funny badge Laughing icon Comments icon

Comments on “Two GCHQ Employees Suggest The Solution To CSAM Distribution Is… More Client-Side Scanning”

Subscribe: RSS Leave a comment
18 Comments
Rico R. (profile) says:

Artificial Intelligence is as reliable as they are sentient...

YouTube can’t even get its AI-flagging bots to behave. At one point, they took Thomas the Tank Engine parodies combining George Carlin’s Thomas narration with his adult comedy bits, videos that YouTube themselves age-restricted in the past, and marked them as “Made for Kids”. Now, the UK wants to expand similar technology to texting and chatting apps for “client-side scanning” of encrypted communications? To protect children and stop the spread of CSAM?

Even Apple’s proposal (which is very flawed) doesn’t go as far as this proposal! “But we won’t abuse or allow the abuse of this technology to target innocent people, we pinky-promise!” Give me a break! At least here in the US, where we have a SCOTUS determined to take away rights and states willing to criminalize the former rights, we need more protection of our digital civil liberties, not less. Besides, AI is far from perfect. It will likely NEVER be 100% perfect, so don’t pretend like it is, or that it can be if we “nerd harder”.

Naughty Autie says:

Re:

YouTube can’t even get its AI-flagging bots to behave. At one point, they took Thomas the Tank Engine parodies combining George Carlin’s Thomas narration with his adult comedy bits, videos that YouTube themselves age-restricted in the past, and marked them as “Made for Kids”.

For which any parent whose kids saw those vids can sue Alphabet. By having its bots make decisions as to the suitability of content for a specified age group, YouTube is acting as a publisher and thus can’t rely on Section 230 protections in regards to that content.

That One Guy (profile) says:

False positives carry no personal risk. False negatives though...

Other than the gross violation of privacy probably the biggest flaw I see in the ‘there will be checks in place to ensure innocent people aren’t caught up and run through the mean-grinder’ is that no-one will want to be the one who gave a pass to potential CSAM lest it come back positive and they’re left explaining their in-action, so the idea that everyone involved will carefully make sure that particular content/speech is super-duper illegal before flagging it as such and passing it up the chain is laughable.

That Anonymous Coward (profile) says:

Re:

False positives have huge risks to people.
Some gunho cops gonna knock on the door expecting a scumbag and well we know how that plays out.

We have evidence that people merely accused of being pedophiles by questionable people, still results in people assuming it must be true because no would say that otherwise. (Kash Hills story about that horrible canadian woman)

I mean its not like cops have shot people for just opening the front door, once they were sure something bad was happening despite any evidence either way.

Naughty Autie says:

For example, one of the approaches we propose is to have language models running entirely locally on the client to detect language associated with grooming. If the model suggests that a conversation is heading toward a risky outcome, the potential victim is warned and nudged to report the conversation for human moderation. Since the models can be tested and the user is involved in the provider’s access to content, we do not believe this sort of approach attracts the same vulnerabilities as others.

Translation: “We want everyone to live in glass houses so no one can throw stones. Don’t worry, we’ll build the houses out of toughened glass just in case a stone is thrown accidentally.”

That One Guy (profile) says:

Re:

Minor correction: “We want everyone but us to live in glass houses so no one can throw stones.”

You can be very sure that the likes of GCHQ, politicians and others that love to argue against privacy would have a huge problem were anyone to suggest that they and/or their employees have text scanning tech on their electronic devices because clearly they would never do anything that might be questionable.

Anonymous Coward says:

Re:

Also, it implies that we have control over all user devices so that our software is always run. If users can control their own devices, bad people will replace the scanner with one that does nothing. Further allowing scanning for CSAM is allowing the camels nose into the tent, and it will be followed by other camels looking for drug dealers, political protesters, etc.

Naughty Autie says:

Re:

Maybe there should be a law passed in every country that whatever security protocol with backdoors, sidedoors, whatever, the people are expected to use, the spy organisations such as GCHQ, NSA, etc. also have to use it, and if they won’t because it’s too insecure, then the people don’t have to use it either.

Anonymous Coward says:

Anyone remember Operation Avalance/Operation Ore?

In which the FBI/UK police joint shut down a child pornography website back at the turn of the millenium?

Well, a lot of the people who were accused of accessing the website, were in fact people who had their credit card information stolen.

Now, I know people crap all over the U.S. police a lot, but at least they took the time to find out if people had their credit cards stolen.

The UK? Well, they couldn’t even be bothered to do that, and just accussed actual innocent people of being pedophiles and child molesters, a good majority had their lives ruined and some actually commited suicide as a result.

I honestly can’t say I have any expections not to fuck this up either.

There’s also lots of news reports recently of police officers in the UK caught having sex with minors.

The whole thing stinks.

Add Your Comment

Your email address will not be published. Required fields are marked *

Have a Techdirt Account? Sign in now. Want one? Register here

Comment Options:

Make this the or (get credits or sign in to see balance) what's this?

What's this?

Techdirt community members with Techdirt Credits can spotlight a comment as either the "First Word" or "Last Word" on a particular comment thread. Credits can be purchased at the Techdirt Insider Shop »

Follow Techdirt

Techdirt Daily Newsletter

Techdirt Deals
Techdirt Insider Discord
The latest chatter on the Techdirt Insider Discord channel...
Loading...