Yes, Section 230 Should Protect ChatGPT And Other Generative AI Tools

from the no-this-wasn't-written-by-chatgpt dept

Question Presented: Does Section 230 Protect Generative AI Products Like ChatGPT?

As the buzz around Section 230 and its application to algorithms intensifies in anticipation of the Supreme Court’s response, ‘generative AI’ has soared in popularity among users and developers, begging the question: does Section 230 protect generative AI products like ChatGPT? Matt Perault, a prominent technology policy scholar and expert, thinks not, as he discussed in his recently published Lawfare article: Section 230 Won’t Protect ChatGPT.

Perault’s main argument follows as such: because of the nature of generative AI, ChatGPT operates as a co-creator (or material contributor) of its outputs and therefore could be considered the ‘information content provider’ of problematic results, ineligible for Section 230 protection. The co-authors of Section 230, former Representative Chris Cox and Sen. Ron Wyden, have also suggested that their law doesn’t grant immunity to generative AI. 

I respectfully disagree with both the co-authors of Section 230 and Perault, and offer the counter argument: Section 230 does (and should) protect products like ChatGPT.

It is my opinion that generative AI does not demand exceptional treatment. Especially since, as it currently stands, generative AI is not exceptional technology; an understandably provocative take to which we’ll soon return. 

But first, a refresher on Section 230.

Section 230 Protects Algorithmic Curation and Augmentation of Third-Party Content 

Recall that Section 230 says websites and users are not liable for the content they did not create, in whole or in part. To evaluate whether the immunity applies, the Barnes v. Yahoo! Court provided a widely accepted three-part test:

  1. The defendant is an interactive computer service; 
  2. The plaintiff’s claim treats the defendant as a publisher or speaker; and
  3. The plaintiff’s claim derives from content the defendant did not create. 

The first prong is not typically contested. Indeed, the latter prongs are usually the flashpoint(s) of most Section 230 cases. And in the case of ChatGPT, the third prong seems especially controversial. 

Section 230’s statutory language states that a website becomes an information content provider when it is “responsible, in whole or in part, for the creation or development” of the content at issue. In their recent Supreme Court case challenging Section 230’s boundaries, the Gonzalez Petitioners assert that the use of algorithms to manipulate and display third-party content precludes Section 230 protection because the algorithms, as developed by the defendant website, convert the defendant into an information content provider. But existing precedent suggests otherwise.

For example, the Court in Fair Housing Council of San Fernando Valley v. Roommate.com (aka ‘the Roommates case’)—a case often invoked to evade Section 230—held that it is not enough for a website to merely augment the content at issue to be considered a co-creator or developer. Rather, the website must have materially contributed to the content’s alleged unlawfulness.  Or, as the majority put it, “[i]f you don’t encourage illegal content, or design your website to require users to input illegal content, you will be immune.” 

The majority also expressly distinguished Roomates.com from “ordinary search engines,” noting that unlike Roommates.com, search engines like Google do not use unlawful criteria to limit the scope of searches conducted (or results delivered), nor are they designed to achieve illegal ends. In other words, the majority suggests that websites retain immunity when they provide neutral tools to facilitate user expression. 

While “neutrality” brings about its own slew of legal ambiguities, the Roommates Court offers some clarity suggesting that websites with a more hands-off approach to content facilitation are safer than websites that guide, encourage, coerce, or demand users produce unlawful content. 

For example, while the Court rejected Roommate’s Section 230 defense for its allegedly discriminatory drop-down options, the Court simultaneously upheld Section 230’s application to the “additional comments” option offered to Roommates.com users. The “additional comments” were separately protected because Roommates did not solicit, encourage, or demand their users provide unlawful content via the web form. In other words, a blank web form that simply asks for user input is a neutral tool, eligible for Section 230 protection, regardless of how the user actually uses the tool. 

The Barnes Court would later reiterate the neutral tools argument, noting that the provision of neutral tools to carry out what may be unlawful or illicit content does not amount to ‘development’ for the purposes of Section 230. Hence, while the ‘material contribution’ test is rather nebulous (especially for emerging technologies), it is relatively clear that a website must do something more than just augmenting, curating, and displaying content (algorithmically or otherwise) to transform into the creator or developer of third-party content.

The Court in Kimzey v. Yelp offers further clarification: 

“the material contribution test makes a “‘crucial distinction between, on the one hand, taking actions (traditional to publishers) that are necessary to the display of unwelcome and actionable content and, on the other hand, responsibility for what makes the displayed content illegal or actionable.’”).”

So, what does this mean for ChatGPT?

The Case For Extending Section 230 Protection to ChatGPT

In his line of questioning during the Gonzalez oral arguments, Justice Gorsuch called into question Section 230’s application to generative AI technologies. But before we can even address the question, we need to spend some time understanding the technology. 

Products like ChatGPT use large language models (LLMs) to produce a reasonable continuation of human-sounding responses. In other words, as discussed here by Stephen Wolfram, renown computer scientist, mathematician, and creator of WolframAlpha, ChatGPT’s core function is to “continue text in a reasonable way, based on what it’s seen from the training it’s had (which consists in looking at billions of pages of text from the web, etc).” 

While ChatGPT is impressive, the science behind it is not necessarily remarkable. Computing technology reduces complex mathematical computations into step-by-step functions that the computer can then solve at tremendous speeds. As humans, we do this all the time, just much slower than a computer. For example, when we’re asked to do non-trivial calculations in our heads, we start by breaking up the computation into smaller functions on which mental math is easily performed until we arrive at the answer.

Tasks that we assume are fundamentally impossible for computers to solve are said to involve ‘irreducible computations’ (i.e. computations that cannot be simply broken up into smaller mathematical functions, unaided by human input). Artificial intelligence relies on neural networks to learn and then ‘solve’ said computations. ChatGPT approaches human queries the same way. Except, as  Wolfram notes, it turns out that said queries are not as sophisticated to compute as we may have thought: 

“In the past there were plenty of tasks—including writing essays—that we’ve assumed were somehow “fundamentally too hard” for computers. And now that we see them done by the likes of ChatGPT we tend to suddenly think that computers must have become vastly more powerful—in particular surpassing things they were already basically able to do (like progressively computing the behavior of computational systems like cellular automata).

But this isn’t the right conclusion to draw. Computationally irreducible processes are still computationally irreducible, and are still fundamentally hard for computers—even if computers can readily compute their individual steps. And instead what we should conclude is that tasks—like writing essays—that we humans could do, but we didn’t think computers could do, are actually in some sense computationally easier than we thought.

In other words, the reason a neural net can be successful in writing an essay is because writing an essay turns out to be a “computationally shallower” problem than we thought. And in a sense this takes us closer to “having a theory” of how we humans manage to do things like writing essays, or in general deal with language.”

In fact, ChatGPT is even less sophisticated when it comes to its training. As Wolfram asserts:

“ChatGPT as it currently is, the situation is actually much more extreme, because the neural net used to generate each token of output is a pure “feed-forward” network, without loops, and therefore has no ability to do any kind of computation with nontrivial “control Flow.””

Put simply, ChatGPT uses predictive algorithms and an array of data made up entirely of publicly available information online to respond to user-created inputs. The technology is not sophisticated enough to operate outside of human-aided guidance and control. Which means that ChatGPT (and similarly situated generative AI products) are functionally akin to “ordinary search engines” and predictive technology like autocomplete. 

Now we apply Section 230. 

For the most part, the courts have consistently applied Section 230 to algorithmically generated outputs. For example, the Sixth Circuit in O’Kroley v. Fastcase Inc. upheld Section 230 for Google’s automatically generated snippets that summarize and accompany each Google result. The Court notes that even though Google’s snippets could be considered a separate creation of content, the snippets derive entirely from third-party information found at each result. Indeed, the Court concludes that contextualization of third-party content is in fact a function of an ordinary search engine. 

Similarly, in Obado v. Magedson, Section 230 applies to search result snippets. The Court says: 

Plaintiff also argues that Defendants displayed through search results certain “defamatory search terms” like “Dennis Obado and criminal” or posted allegedly defamatory images with Plaintiff’s name. As Plaintiff himself has alleged, these images at issue originate from third-party websites on the Internet which are captured by an algorithm used by the search engine, which uses neutral and objective criteria. Significantly, this means that the images and links displayed in the search results simply point to content generated by third parties. Thus, Plaintiff’s allegations that certain search terms or images appear in response to a user-generated search for “Dennis Obado” into a search engine fails to establish any sort of liability for Defendants. These results are simply derived from third-party websites, based on information provided by an “information content provider.” The linking, displaying, or posting of this material by Defendants falls within CDA immunity.

The Court also nods to Roommates

“None of the relevant Defendants used any sort of unlawful criteria to limit the scope of searches conducted on them; “[t]herefore, such search engines play no part in the ‘development’ of the unlawful searches” and are acting purely as an interactive computer service…

The Court goes further, extending Section 230 to autocomplete (i.e. when the service at issue uses predictive algorithms to suggest and preempt a user’s query): 

“suggested search terms auto-generated by a search engine do not remove that search engine from the CDA’s broad protection because such auto-generated terms “indicates only that other websites and users have connected plaintiff’s name” with certain terms.”

Like Google Search, ChatGPT is entirely driven by third-party input. In other words, ChatGPT does not invent, create, or develop outputs absent any prompting from an information content provider  (i.e. a user). Further, nothing on the service expressly or impliedly encourages users to submit unlawful queries. In fact, OpenAI continues to implement guardrails that force ChatGPT to ignore requests that would demand problematic and / or unlawful responses. Compare this to Google Search which may actually still provide a problematic or even unlawful result. Perhaps ChatGPT actually improves the baseline for ordinary search functionality. 

Indeed, ChatGPT essentially functions like the “additional comments” web form in Roommates. And while ChatGPT may “transform” user input into a result that responds to the user-driven query, that output is entirely composed of third-party information scraped from the web. Without more, this transformation is simply an algorithmic augmentation of third-party content (much like Google’s snippets). And as discussed, algorithmic compilations or augmentations of third-party content are not enough to transform the service into an information content provider (e.g. Roommates; Batzel v. Smith; Dyroff v. The Ultimate Software Group, Inc.; Force v. Facebook). 

The Limit Does Exist

Of course, Section 230’s coverage is not without its limits. There’s no doubt that future generative AI defendants, like OpenAI, will face an uphill battle in persuading a court. Not only do defendants have the daunting challenge of explaining generative AI technologies for less technologically savvy judges, the current judicial swirl around Section 230 and algorithms does defendants no favors. 

For example, the Supreme Court could very well hand-down a convoluted opinion in Gonzalez that introduces ambiguity as to when Section 230 applies to algorithmic curation / augmentation. Such an opinion would only serve to undermine the precedence discussed above. Indeed, future defendants may find themselves embroiled in convoluted debate about AI’s capacity for neutrality. In fact, it would be intellectually dishonest to ignore emerging common law developments that preclude Section 230 from claims alleging dangerous / defective product designs (e.g. Lemmon v. Snap, A.M. v. Omegle, Oberdorf v. Amazon). 

Further, the Fourth Circuit’s recent decision in Henderson v. Public Data could also prove to be problematic for future AI defendants as it imposes contributive liability for publisher activities that go beyond those of “traditional editorial functions” (which could include any and all publisher functions done via algorithms). 

Lastly, as we saw in the Meta / DOJ settlement regarding Meta’s discriminatory practices involving algorithmic targeting of housing advertisements, AI companies cannot easily avoid liability when they materially contribute to the unlawfulness of the result. If OpenAI were to hard-code ChatGPT with unlawful responses, Section 230 will likely be unavailable. However, as you might imagine, this is a non-trivial distinction. 

Public Policy Demands Section 230 Protections for Generative AI Technologies

Section 230 was initially established with the recognition that the online world would undergo frequent advancements, and that the law must accommodate these changes to promote a thriving digital ecosystem. 

Generative AI is the latest iteration of web technology that has enormous potential to bring about substantial benefits for society and transform the way we use the Internet. And it’s already doing good. Generative AI is currently used in the healthcare industry, for instance, to improve medical imaging and to speed up drug discovery and development. 

As discussed, courts have developed precedence in favor of Section 230 immunity for online services that solicit or encourage users to create and provide content. Courts have also extended the immunity to online services that facilitate the submission of user-created content. From a legal standpoint, generative AI tools are not unique from any other online service that encourages user interaction and contextualizes third-party results. 

From a public policy perspective, it is crucial that courts uphold Section 230 immunity for generative AI products. Otherwise, we risk foreclosing on the technology’s true potential. Today, there are tons of variations of ChatGPT-like products offered by independent developers and computer scientists who are likely unequipped to deal with an inundation of litigation that Section 230 typically preempts. 

In fact, generative AI products are arguably more vulnerable to frivolous lawsuits because they depend entirely upon whatever query or instructions its users may provide, malicious or otherwise. Without Section 230, developers of generative AI services must anticipate and guard against every type of query that could cause harm. 

Indeed, thanks to Section 230, companies like OpenAI are doing just that by providing guardrails that limit ChatGPT’s responses to malicious queries. But those guardrails are neither comprehensive nor perfect. And like with all other efforts to moderate awful online content, the elimination of Section 230 could discourage generative AI companies from implementing said guardrails in the first place; a countermove that would enable users to prompt LLMs with malicious queries to bait out unlawful responses subject to litigation. In other words, plaintiffs could transform ChatGPT into their very own personal perpetual litigation machine. 

And as Perault rightfully warns: 

“If a company that deploys an LLM can be dragged into lengthy, costly litigation any time a user prompts the tool to generate text that creates legal risk, companies will narrow the scope and scale of deployment dramatically. Without Section 230 protection, the risk is vast: Platforms using LLMs would be subject to a wide array of suits under federal and state law. Section 230 was designed to allow internet companies to offer uniform products throughout the country, rather than needing to offer a different search engine in Texas and New York or a different social media app in California and Florida. In the absence of liability protections, platforms seeking to deploy LLMs would face a compliance minefield, potentially requiring them to alter their products on a state-by-state basis or even pull them out of certain states entirely…

…The result would be to limit expression—platforms seeking to limit legal risk will inevitably censor legitimate speech as well. Historically, limits on expression have frustrated both liberals and conservatives, with those on the left concerned that censorship disproportionately harms marginalized communities, and those on the right concerned that censorship disproportionately restricts conservative viewpoints.

The risk of liability could also impact competition in the LLM market. Because smaller companies lack the resources to bear legal costs like Google and Microsoft may, it is reasonable to assume that this risk would reduce startup activity.”

Hence, regardless of how we feel about Section 230’s applicability to AI, we will be forced to reckon with the latest iteration of Masnick’s Impossibility Theorem: there is no content moderation system that can meet the needs of all users. The lack of limitations on human awfulness mirrors the constant challenge that social media companies encounter with content moderation. The question is whether LLMs can improve what social media cannot.

Filed Under: , , , , , ,
Companies: openai

Rate this comment as insightful
Rate this comment as funny
You have rated this comment as insightful
You have rated this comment as funny
Flag this comment as abusive/trolling/spam
You have flagged this comment
The first word has already been claimed
The last word has already been claimed
Insightful Lightbulb icon Funny Laughing icon Abusive/trolling/spam Flag icon Insightful badge Lightbulb icon Funny badge Laughing icon Comments icon

Comments on “Yes, Section 230 Should Protect ChatGPT And Other Generative AI Tools”

Subscribe: RSS Leave a comment
47 Comments
Anonymous Coward says:

If you cannot copyright “AI” outputs, then the hosting sites or code developers shouldn’t be responsible for natural outputs either.

If they are intentionally or inadvertently providing algorithms, other code, or constraints that queer the output, they might be “responsible”, but very likely not in any legal sense, as they seem to strive to avoid not only illegal, but also otherwise socially negative output. (Unless, you know, future bad rulings and laws.)

Hopefully devs will strive to avoid WEIRD, and related, centrisms, but there shouldn’t be anything illegal in failing to do so.

This comment has been flagged by the community. Click here to show it.

Anonymous Coward says:

So put lies about yourself on the internet, let ChatGPT pick them up, then people who read it will repeat it, without section 230 protection, and get sued.

Pretty easy money for an unethical lawyer against defenseless individuals whose only mistake was believing what they read on the internet to the point of repeating it.

This comment has been flagged by the community. Click here to show it.

This comment has been flagged by the community. Click here to show it.

P. Orin Zack (user link) says:

Remixed Internet Content

If I understand the implied basis for protection, it seems to rest on the fact that the LLMs base their output on material from the Internet which was produced ultimately by some other person and then used to train the LLM.

What happens when a LLM has the ability to apply a pattern observed in one context to content drawn from a different context? This would make it possible for the AI to use parallel construction to present a conclusion that may not have already been present on the Internet. What it produced in this instance would not be a remix of existing content. Would that make the AI the creator of original content?

jmiers230 (profile) says:

Re:

yeah — it’s a good Q. The remixing of content argument isn’t the strongest IMO, especially as courts are increasingly reticent to apply 230 to other similar examples.

The strongest argument in favor of 230 protection, even as applied to your example, is that the content creation rests with the user that provides the query to begin with (i.e. the original information content provider).

Now…once the AI starts acting entirely on its own, absent any human prompting…who knows.

P. Orin Zack (user link) says:

Re: Re: Seeding Content

Hmmm. If content creation rests with the user who provided the prompt, that raises other possibilities. It would in essence be the same as seeding a random number generator. Who is the creator if the AI’s output becomes the next prompt? (Ask the AI to pose a recursive question about X.) This issue gets very twisty.

Anonymous Coward says:

Why was Jess Miers’ affiliation with the tech lobbying group Chamber Of Progress not disclosed, nor her former employment at Google and Twitter? It would be a big help knowing that this comes from an organization that lobbies on the side of huge corporations a ton of the time, and the author’s prior employment at two huge tech firms.

In regards to the article: I don’t think that AI should be protected by 230. Generative AI is moving way too fast, and harms are already manifesting. Deepfake AI revenge porn is a thing, as the article “QTCinderella Finds She Cannot Sue The Creator Of Deepfake Site” on TheGamer discusses. Steph Sterling’s YouTube video, “How AI Takes The Art Out Of The Artist”, also discusses how generative AI technology can be used, and has been used, against sex workers.

The “The Limit Does Exist” segment of the article also comes off as a list of situations where Miers laments the idea that Section 230 actually has limits, with one part at the end where legal limits feel like they’re begrudgingly acknowledged. It feels icky.

Also in that segment: Talking about Henderson v. Public Data and linking to an Eric Goldman article about it, while Miers does not also disclose that she is a teaching assistant for Goldman, as pointed out on Ctrl Alt Dissent, is something that is pretty glaring.

Goldman’s analysis of why Henderson v. Public Data is a bad decision is flawed. Right here on Techdirt, the initial bunch of comments to the article “Fourth Circuit Goes On The Attack Against Section 230 In A Lawsuit Over Publication Of Third Party Data” point out the issue: Public Data is definitely the publisher here, and they reformatted and edited the content published. There was no harm to 230 there because quite frankly shouldn’t apply.

We need to maintain some semblance of shared reality if we are to function as a society. Regulating AI tech like OpenAI, ChatGPT, and the tools that people are using to create the adult materials mentioned above, as well as enforcing liability on said tech and tools and those that use them for harm, would be nice.

Anonymous Coward says:

Re:

We need to maintain some semblance of shared reality if we are to function as a society. Regulating AI tech like OpenAI, ChatGPT, and the tools that people are using to create the adult materials mentioned above, as well as enforcing liability on said tech and tools and those that use them for harm, would be nice.

It’a like the Criminal Codes of every single fucking country in the entire fucking world does not exist for the very sole purpose of limiting the use of tools like ChatGPT for HARASSMENT and whatnot!

Rocky says:

Re: Re: Re:

Because it’s two entirely different things. Scraping the internet for training material might fall under copyright law whereas section 230 is about how websites deal liability for content and how content from an AI might not qualify as speech/expression within that context – regardless on what material it was trained on. We can frame the above as two questions: Should we treat learning/training as different things legally depending on if the neural network used is artificial (AI) or natural (human)? and Should we treat speech/expressions from an AI the same way as for a human?

The last question is actually framing it a bit dishonestly for the simple reason that anything published by an AI is at the behest of a human or a human controlled entity in some manner and the question then becomes “Is this specific output from an AI affected by Section 230 or not?” – because every time liability for speech comes up we must look at the context and the specific “offending” speech, just like how earlier cases have looked at liability for content that was automatically produced and drawn from multiple sources.

Anonymous Coward says:

Re: Re: Re:2

We can frame the above as two questions: Should we treat learning/training as different things legally depending on if the neural network used is artificial (AI) or natural (human)? and Should we treat speech/expressions from an AI the same way as for a human?

I agree. And I think that if the answer to one of those questions is yes, then logically the answer to the other must be no.

Anonymous Coward says:

Re:

If I understand the situation, it’s not the program itself that can be sued, but its owners.

The GRIFTERS who created ChatGPT would be ones subject to lawsuits regarding the training dataset.

Also, ChatGPT can’t do shit without human input. Assuming the grifters who own ChatGPT want to play by the rules (and they should if they want to have a shred of respect from the rest of us), for S230 purposes, if shown a warrant, they should comply at the bare minimum.

Anonymous Coward says:

Re: Re:

I mean, if we take OpenAI at their word, they should all be in jail for slavery by now. “It’s basically human” and “Yes, we own it” are tricky positions to maintain at the same time.

and they should if they want to have a shred of respect from the rest of us

That’s a loadbearing ‘if’

Anonymous Coward says:

Re: Re: Re:

I don’t have a lot of respect for the former NFT grifters and assorted scam artists running the hot new AI companies.

Granted, the current wave of AIs might make for useful tools and could drastically shorten development times for software, apps and games, they ARE being marketed as “replacements” for creative labor…

…which, sadly, isn’t just happening anytime soon, it’s also being pushed by people who know how to grift the grant and investment system.

And they also happen to be the same folk who WERE pushing cryptocurrency and NFTs. At least, in my opinion.

Suffice to say, I am assuming they do not have even that bare minimum of humanity.

Anonymous Coward says:

I’m not convinced. Yes, it only provides information based on user prompts. So what? The commenters here also only comment based on the prompt of an article; does that mean the people who write the articles are responsible for what we write here? Obviously not.

Over at the Volokh Conspiracy they gave an example of ChatGPT, in response to a user asking about what a certain person was accused of, also telling the user that the person had pled guilty. When pressed, it presented quotes allegedly from The Guardian, Reuters, and The Washington Post. It also claimed to be referencing a Department of Justice press release, although it backed off on that claim when pressed further. But the majority of what it was saying was false! While the guy had been accused by some people, he hadn’t even been prosecuted, let alone pled guilty, and the quotes simply didn’t exist.

In this case, it’s making stuff up. I don’t think you can call this something “entirely composed of third-party information scraped from the web.” Perhaps the individual words were scraped from the Web, but it created something entirely new from them – something libelous. And I don’t think you can say the person asking the question was the one who “provided” this, either. He didn’t provide anything about prosecution or a false guilty plea.

This AI is categorically different from a Google search. Google might show you a snippet, but it won’t make one up to justify why it showed you a page!

Rocky says:

Re:

I’m not convinced. Yes, it only provides information based on user prompts. So what? The commenters here also only comment based on the prompt of an article; does that mean the people who write the articles are responsible for what we write here? Obviously not.

That’s a very bad analogy. Techdirt doesn’t go up to people and prompt them with a headline to get comments, but that’s how language model AI’s are used – someone prompts them for a response from an input.

Over at the Volokh Conspiracy they gave an example of ChatGPT, in response to a user asking about what a certain person was accused of, also telling the user that the person had pled guilty. When pressed, it presented quotes allegedly from The Guardian, Reuters, and The Washington Post. It also claimed to be referencing a Department of Justice press release, although it backed off on that claim when pressed further. But the majority of what it was saying was false! While the guy had been accused by some people, he hadn’t even been prosecuted, let alone pled guilty, and the quotes simply didn’t exist.

They where asking something that doesn’t have a shred of intuition or intelligence to produce a text on a given topic, and then they pressed it for more?! What did they expect to happen? AI isn’t some magical tech, it doesn’t really have any kind of concept of what is true of false.

I don’t think you can call this something “entirely composed of third-party information scraped from the web.”

But it is, but it’s all jumbled up in the training set and if the material it was trained on contained the text from several cases they wont be spat out verbatim – it’ll come out in a mish-mash that seems to make sense but it’s actually made up of information from multiple sources because the texts become correlated within the context of crime for example.

Perhaps the individual words were scraped from the Web, but it created something entirely new from them – something libelous.

There has to be intent to lie and knowledge of the real truth for something to be considered libelous.

And I don’t think you can say the person asking the question was the one who “provided” this, either. He didn’t provide anything about prosecution or a false guilty plea.

Taking the output from an “AI” as the gospel truth is just plain stupid. Now, if someone takes that unvetted output and knowingly publishes it without a disclaimer they should be entirely responsible for that publication.

This AI is categorically different from a Google search. Google might show you a snippet, but it won’t make one up to justify why it showed you a page!

Yes, because an AI isn’t a search engine and it doesn’t actually give you links to factual information – it makes shit up that may be factual. In a sense it makes everything up and some of it is true, some of it isn’t. It’s like asking someone to tell a story, it may be based on real events or it may be entirely fictional with real elements to make it more believable.

This is the kind of debate I see every time some new tech becomes available, the general public assumes a lot of shit based on articles from main stream media that were so dumbed down it bears no resemblance to what the tech actually does or how it functions.

Anonymous Coward says:

Re: Re:

it makes shit up

Indeed. It makes it up. It is therefore the creator of the content it made up, I would argue, and therefore doesn’t get 230 protection on that content.

There has to be intent to lie and knowledge of the real truth for something to be considered libelous.

You can libel someone recklessly. The standard for “actual malice” is “with knowledge that it was false or with reckless disregard of whether it was false or not.” And that standard is only applied if it’s a public figure; otherwise mere negligence could suffice. How good are the disclaimers of the people who made this AI? Are they making sure everyone knows its limitations, so a reasonable person would know not to trust it? Or are they touting things like “It passed the bar!” and making a reasonable person think it’s accurate?

Rocky says:

Re: Re: Re:

Indeed. It makes it up. It is therefore the creator of the content it made up, I would argue, and therefore doesn’t get 230 protection on that content.

And if it doesn’t get protections and someone want to sue it for libel, how do you propose that would work?! You can’t sue software.

You can libel someone recklessly. The standard for “actual malice” is “with knowledge that it was false or with reckless disregard of whether it was false or not.” And that standard is only applied if it’s a public figure; otherwise mere negligence could suffice.

An AI doesn’t know what malice or with reckless disregard of whether it was false or not, or negligence for that matter.

How good are the disclaimers of the people who made this AI? Are they making sure everyone knows its limitations, so a reasonable person would know not to trust it? Or are they touting things like “It passed the bar!” and making a reasonable person think it’s accurate?

OpenAI has this disclaimer:

3. Content

. . .
*(d) Accuracy. Artificial intelligence and machine learning are rapidly evolving fields of study. We are constantly working to improve our Services to make them more accurate, reliable, safe and beneficial. Given the probabilistic nature of machine learning, use of our Services may in some situations result in incorrect Output that does not accurately reflect real people, places, or facts. You should evaluate the accuracy of any Output as appropriate for your use case, including by using human review of the Output. *

Ie, if someone takes the output and just post it somewhere and it turns out to be false and libelous – guess who’s on the hook for that?

nasch (profile) says:

Re:

I think you’re right on. The AI is only “remixing” in the sense that everything it says is English words that have all been seen before. Its output, as a whole, is new and unique. If a human did what ChatGPT does it could not credibly be accused of plagiarism.

However, the second question is, should generative AI be covered by section 230? That’s trickier IMO because if it isn’t, nobody will offer such services in the US. Is that a good outcome or bad? Probably some of each.

Rocky says:

Re: Re: Re:2

No more than you can sue any other computer program. But you can perhaps sue the company offering its services.

And how is that different from the current situation?

You always have a “publisher”, or as 230 specifies it: information content provider. That person or entity (as in company, corporation or other associations of people) is always responsible for the content they produce and disseminate, and that means that 230 will cover AI output because otherwise a site could suddenly be responsible for content generated by an AI that a 3rd party posted.

Anonymous Coward says:

Re: Re: Re:3

The buck has to stop somewhere. The companies and orgs making these generative AIs should be liable for their output when it’s used to cause harm. Perpetually asking victims to seek out anonymous users who may be halfway across the world and untouchable by the average person seeking legal recourse, having that be the status-quo, is fucked.

Rocky says:

Re: Re: Re:4

The buck has to stop somewhere. The companies and orgs making these generative AIs should be liable for their output when it’s used to cause harm.

Tell me, how do you tell if some particularly libelous piece of text was written by a person or an AI? Some anonymous asshole can just post something they made up and say it was made by an AI from company X and that company must then deal with a libel-suit. If you think that’s the way to go, prepare to be bled dry in court.

Perpetually asking victims to seek out anonymous users who may be halfway across the world and untouchable by the average person seeking legal recourse, having that be the status-quo, is fucked.

So the solution is to seek out a third party and pretend it’s a money piñata while ignoring the real perpetrator?! I’m sure that’ll make the real perpetrators change their behavior and stop posting shit on the internet.

Anonymous Coward says:

Re: Re: Re:5

Scenario: Someone uses AI generation to create a fake audio clip of a real person in the U.S. saying some racist shit. It gets posted online and the victim is hounded relentlessly thanks to it. Their life gets wrecked for a while until they can prove to everyone that it was fake. The victim then tries to find the person who originally uploaded it to pin them for defamation. Oh, it looks like they used a throwaway email and a VPN and can’t be found. Or, they’re in a country that’s not gonna help out the victim at all. But the victim manages to find the website of the folks that made the AI model that let the uploader make that clip. They’re in the U.S., and there’s enough info where the victim can file suit.

Do you think the victim should be allowed to sue the creators of the AI model for letting it be used for defamation?

Another Scenario: Someone uploads an AI-generated photo that makes it look like someone is wearing blackface. The victim has to deal with the fallout as people think the photo is real. The same deal: The original poster can’t be located, or they’re somewhere the victim can’t reach. But the company that made the model used to make the blackface deepfake, they can be found.

Should the victim be allowed to sue the company that made the model and allowed it to be used that way?

Victims of deepfakes and ensuing harassment borne of those deepfakes, they deserve some manner of legal recourse or recompense for being abused by fly-by-night anonymous actors using those tools for nasty shit.

Anonymous Coward says:

Re: Re: Re:6

No, they shouldn’t be able to sue the AI (or rather, the company behind it) in those cases any more than they should be able to sue the person who laid the Internet cable the harasser used. If someones tells an AI “Change this photo so the person is wearing blackface” then it’s not the AI’s fault. That’s not the same as the AI falsely claiming that someone had pled guilty to a crime.

It’s like the difference between the maker of a car being sued because the accelerator pedal stuck and the car hit someone versus the maker of a car being sued because someone intentionally floored the accelerator and the car hit someone. If the product was being used in a reasonable manner and it does something unreasonable, that’s the product’s fault. If the product was intentionally misused, that’s the user’s fault.

Zach says:

This is a terrible take. The black-box nature of AI makes regulation more important, not less. AI authoritatively making libelous claims should open up its creators to liability.

The the content moderation standard isn’t germane. It doesn’t have to meet the needs of all users, it just has to cross the low bar of non-libelous. If it isn’t up to that challenge it shouldn’t be used without actual moderation.

Rocky says:

Re:

This is a terrible take. The black-box nature of AI makes regulation more important, not less. AI authoritatively making libelous claims should open up its creators to liability.

No, only those who post the output online without vetting it.

If it isn’t up to that challenge it shouldn’t be used without actual moderation.

BINGO!

Jack the Nonabrasive (profile) says:

It’s the tokens that matter

In using an analogy to Google snippets, the author inadvertently undermines his own argument.

The snippets are packages of meaning, created by a 3rd party, clearly covered by the statute.

The tokens used by large text models (here, I’m using computational linguist David Schlangen’s term for them, recognizing that language is more than just text) are much smaller units of meaning than the excerpts used in snippets.

The work done by the website in labeling (that’s original content), creating training datasets (admittedly, editorial content), providing human feedback (if using reinforcement learning with human feedback for training), and finally generating the content token-by-token using copyright-protected models they’ve generated through that act of creation seems to demonstrate that it’s their own content, not curated 3rd party content.

It’s the size of the units of meaning that are mixed and matched using my expertise that matters.

To use an analogy: if i make you a martini, and it’s horrible, you’ll blame the rotgut gin I used.

I can take your water, his yeast, her malt and ferment a tasty beverage or an undrinkable mess, but it will be my beer you either enjoy or not, not yours, his, and hers.

Size matters, especially in units of meaning with respect to this statute.

Leave a Reply to P. Orin Zack Cancel reply

Your email address will not be published. Required fields are marked *

Have a Techdirt Account? Sign in now. Want one? Register here

Comment Options:

Make this the or (get credits or sign in to see balance) what's this?

What's this?

Techdirt community members with Techdirt Credits can spotlight a comment as either the "First Word" or "Last Word" on a particular comment thread. Credits can be purchased at the Techdirt Insider Shop »

Follow Techdirt

Techdirt Daily Newsletter

Ctrl-Alt-Speech

A weekly news podcast from
Mike Masnick & Ben Whitelaw

Subscribe now to Ctrl-Alt-Speech »
Techdirt Deals
Techdirt Insider Discord
The latest chatter on the Techdirt Insider Discord channel...
Loading...