When Words Aren’t Speech
AI can generate perfect sentences. The First Amendment is about something else.
Why trying to answer this question helped me feel (slightly) less existentially terrified by machines that sound like humans
The first time I used ChatGPT, I didn’t sleep the entire night. It was so much better, so much more realistic, and so much more uncannily human than I ever thought I’d see in my lifetime that I was convinced humanity was over (in the long run), that I was out of a job as a data scientist (in the medium run), and that we were all about to be overrun by AI-generated nonsense (in the short-to-right-now run).
Since then, I’ve grudgingly come to accept that large language models, or LLMs, which are the backbone of generative AI, are here to stay (if nothing else, I refuse to become a Millennial version of a Boomer asking what “a Twitter” is). I’ve come to appreciate that they make me at least feel like I’m coding more quickly. And as an insecure person with zero self-esteem, I don’t even mind their sycophancy, even if it’s probably bad for me — and science — in the long run.
But in addition to having concerns about the impact of LLMs on the environment, copyright, critical thinking, and more, I still found myself most deeply unsettled by the fact that AI-generated speech sounded a lot like, well, speech. So when The Preamble took on the topic of the First Amendment, I figured, why not see what it has to say about the creepiest speech of all: robots’?
In doing so, I came to appreciate a deeper nuance to the First Amendment, human speech, and what makes us different from the machines, even if they sound more like us every day.
Speech is about much more than words
The first shock to me as someone who is in no way a legal scholar but fancies herself informed about politics and governance is that the First Amendment begins with a focus on religion rather than speech. Speech is included alongside various other ways we might express our beliefs, such as using the press, assembling, and petitioning the government.
Out of the gates, this made me rethink what we’re doing when we are speaking. We’re not just saying random words, or even words in some grammatically correct order. We’re expressing a belief, an idea, a preference, an agenda — something. By declaring that Congress cannot in any way compel or forbid us to say or listen to whatever we want (with some exceptions around violence, defamation, breaking the law, and obscenities), the First Amendment protects our right to have, share, and learn about points of view. It’s not about words nearly as much as it’s about what’s behind those words. Thus, if speech is really about ideas and expressing them, then words that sound human are not speech any more than a parrot or a creepy doll with a string is.
This also raises the question of whether the incoherent babbling of a baby or the rambling of someone with dementia is speech. Even if we cannot understand what they are saying, if the baby or the ill person is trying to communicate something, then perhaps their utterances are closer to “speech” than the crystal-clear words of a machine.
This points to a few different kinds of speech:
speech that is coherent and expresses a belief (most human-generated content, even if that “belief” is something mundane, like the view that pretzel salad is a salad);
speech that is incoherent but does attempt to express a belief (the mutterings of babies and the infirm);
speech that is coherent but does not express a belief (LLMs, parrots, creepy dolls);
and speech that neither is coherent nor attempts to express a belief (most of the papers I wrote in college).
Intentions matter
That brings us to a second helpful distinction. Much of the law, whether about speech or otherwise, hinges on “intent.” If a person kills someone, it matters a lot in the eyes of the law (and the associated true crime podcast) whether they did it on purpose (murder) or not (manslaughter). If I say a bunch of false but inflammatory things on the internet, it matters whether I thought they were true when I said them (innocent mistake) or I knew they were false but said them anyway (Alex Jones).
Thus, as two legal scholars have recently written, AI cannot itself have intentions in the way that humans do. Yes, AI models like LLMs may have “goals” like finding out how likely two words are to appear together or minimizing the gap between predictions and real data, but there is no intent behind any of its output. LLMs are not producing speech because they wish to convey something, any more than a Magic 8 Ball wishes to say something when it tells you that “sources say yes” or instructs you to “ask again later” — even though the (unintended) consequence of the latter is absolute fury. They are producing sequences of words that look like speech because they have been programmed to do so. From this perspective, LLM output (even when surprisingly hard to predict) cannot be protected or unprotected speech because it’s not said with any intent.
Okay, but is it someone’s speech?
If LLMs are not expressing beliefs and they can’t have intent, then are they just really sophisticated tools? If I (hypothetically) write and circulate a manifesto declaring that all cheese should be replaced by vegan cheese, it shouldn’t matter whether I wrote it using an LLM, a typewriter, or a quill and ink. These are my beliefs, expressed with intent, and thus are my own protected speech — regardless of how unpopular my idea might be, or what tools I used to communicate it.
But suppose I use an LLM for said manifesto and somewhere along the way it produces language that incites pro-vegan-cheese lawlessness. If I share that language without reading it first, is it still my speech? Some legal scholars would say yes, because an LLM can be thought of as an assistant or “agent” acting on behalf of the user. Thus, just as a hospital could conceivably (though not necessarily) be held responsible for the behavior of one of its nurses, I might be responsible for language I circulate even if an LLM created it for me.
Things get more complicated, however, if we consider that an LLM might produce speech that harms me, the user. For example, LLMs have been documented to “hallucinate” (or generate and confidently present false facts). If I ask an LLM whether a poisonous berry is safe, it incorrectly says yes, and then I eat the berry and get sick, whose speech was it that led to my behavior? It starts to be rather brain-breaking to think of this as my own speech that led me to eat the berry.
This tension is at the heart of some disturbing cases about LLMs that are in courts now. If, when I interact with LLMs, I run the risk of incurring harm, then shouldn’t we consider that the responsibility for the harm lies at least in part with the creator of the LLM? Perhaps the engineers who made it, or the CEO who oversaw the process and product deployment, should have built in safeguards to prevent certain kinds of dangerous output. In fact, it’s increasingly possible for regular people to build their own LLMs. In that case, is it their speech?
Indeed, some experts argue that we should consider LLM output to be a form of expression of those who built the LLM. If you think of a very simple LLM, perhaps a bot that answers yes/no questions on behalf of someone who built it, then it isn’t so hard to see the output of more complex LLMs in the same way, and to consider it the protected speech of LLM creators.
What if it’s… everyone’s speech?
There are yet two more lenses to look through as we try to wrap our legal minds around what on earth (and in space) LLMs are producing.
One is to consider the inputs to LLMs, or the data that they are “trained” on. LLMs look for patterns in massive amounts of text from the internet. Much of this text is not copyrighted, such as posts on Reddit or Wikipedia, but meaningful quantities are subject to copyright law. This material includes the work of journalists, authors, comedians, musicians, screenwriters, and more. A legal debate is underway as to whether the use of all of this data by the companies that built the LLMs can be considered “fair use.” But the way LLMs are trained also raises the question of whether their output is the complicated, combined speech of all of us. If so, then once again we might see LLM output as protected speech.
Our final angle is related to the idea that the First Amendment is not just about our right to express beliefs. It’s also about our right to hear them. Regardless of who is producing the LLM content, or whose speech we decide it is, we can imagine a world where any attempt to curb or regulate its output needs to take into account our First Amendment right to be exposed to any ideas we want (within the existing rules around obscenities and lawlessness).
Now what?
Over the course of working on this article I’ve become slightly more persuaded by arguments that LLM output is not speech, primarily on the rationale that LLMs don’t intend to produce meaning. But I’ve also come to appreciate more deeply that there is more to the First Amendment than I originally realized, including its emphasis on freedom to have, express, and hear lots of different beliefs. And if an LLM is capable of producing an idea or solution that could really improve our lives, I’m certainly sympathetic to arguments that we ought to be careful before we start regulating it.
It could be that the legal world makes decisions about LLMs that resemble its treatment of social media platforms, wherein the platforms are not considered responsible for what users post and are also free to moderate content as they see fit.
But — unlike social media or even search engines, which can be thought of as editors or curators of information — the output from LLMs is not user-generated. Instead, it’s the result of the algorithm itself, which brings us back to the tricky question of whether the output of a machine is the speech of its creator. (“Actually, it’s Frankenstein’s monster’s speech.”)
For now, one of the contributions of the debate about LLMs and the First Amendment, regardless of where it ultimately lands, is that it has forced a lot of us to think carefully about what we’re even doing when we’re speaking, why we’re doing it, and exactly what about it feels human (or at least non-machine, as I am convinced my dogs speak with absolutely clear beliefs and intentions, especially at four in the morning).
For now, because I am a human with free will, I will leave you with a haiku that summarizes this article, and not just because that seems exactly like something I’d use an LLM to write (I didn’t!).
Is it human speech?
Without intent and belief
Perhaps just slick words.








I was supposed to publish an essay this morning but I think it needs another week of work, and after reading this I’m reconsidering some of what I was writing based on your insights. So thank you, Andrea! In all my research on where our founding documents have holes that need filling, I hadn’t considered how slippery LLMs are within existing speech paradigms. But I did appreciate your category of speech that neither is coherent nor attempts to express a belief, because I had considered my own college writing with the same disdain and horror that you seem to have for yours. Kids these days won’t know what it’s like to submit an essay that was three pages long only because the assignment said it needed to be three pages long, but having less than a page of ideas in your brain. Or I guess they will, but maybe their non-ideas will read like a masterpiece based on the world’s written thoughts? 😭
My other takeaway is how much the First Amendment needs more clarity for today’s world. It’s the reason unlimited money gets dumped into our elections through Citizens United, where the Supreme Court decided that spending money is a form of protected speech. It’s why a billionaire effectively gets a billion times more political influence than someone who can’t maintain a savings account because of the cost of living. It’s even being used to block data privacy laws, with companies arguing that collecting and selling your personal information is their protected “speech,” as if data about you is somehow their expression.
Here’s one that blew my mind: I was watching Hulu and saw Serena Williams, one of the greatest athletes in human history, telling me she needed GLP-1 medication to get her body weight under control. The person I was watching with, like me recovering from holiday overeating, half-joked about maybe needing to get on the drug too. And that’s the point, right? If a world-class athlete needs pharmaceuticals to lose a few pounds, what hope do the rest of us have? To be clear: Serena’s truth and circumstances are none of my business, usually. I’m just talking about the messages we receive from powerful companies. Then at a party I mentioned this to a friend from New Zealand, and she told me the US and NZ are the only two developed countries that allow pharmaceutical companies to advertise directly to consumers. Everyone else figured out that letting drug companies manufacture demand is insane. But here, it’s protected speech, so we get manipulated into asking our doctors for medications they wouldn’t otherwise prescribe, which drives up healthcare costs for everyone, which makes insurance more expensive for all of us. Overmedicated and under-cared for. The First Amendment is doing a lot of counterproductive work that the founders definitely did not intend.
Section 230 is another place where First Amendment principles have been stretched into something the founders couldn’t have imagined. It was written in the 90s when we were talking about not holding AOL responsible for the misguided words we’d send each other as instant messages or post on discussion boards. That made sense. But now the same law applies when Facebook finds some random person you’ve never heard of, who you’re not friends with or following, and shoves their ragebait political post above the photos of your friend’s dogs that you actually signed up for. The argument that platforms are just neutral conduits falls apart when you realize their engagement-driven systems are making editorial choices millions of times per second, but those choices benefit from legal protections that no newspaper or TV station would ever receive. And when lawmakers try to require algorithmic transparency or accountability, platforms are increasingly arguing that their recommendation algorithms are protected editorial speech, meaning we might not even be allowed to know why certain content gets pushed to millions of people while other content gets buried. Meanwhile, actual people harmed by coordinated harassment campaigns or defamatory content or deepfakes have almost no recourse.
And speaking of defamation, it’s wild that our libel laws make it nearly impossible for public figures to win suits against actual lies, but wealthy men can weaponize those same courts to silence accusers and journalists. For instance the president suing WSJ for publishing the existence of a book in which he alluded to collaborating on “secrets” with Jeffrey Epstein, which absolutely happened, but is part of a consistent pattern of legal abuse against a free press. The legal costs alone can destroy someone or a newspaper financially, even if the case eventually gets dismissed. The First Amendment is supposed to protect the powerless speaking truth to power, but too often it’s been interpreted to protect the powerful from accountability.
Sorry for the tangents… maybe this is why I take weeks to write an essay. All of this is to say: your piece made me think that “what is speech” might be a red herring for what I’m working on. Maybe the better question is “who benefits from calling something speech?” Because right now, that answer is increasingly corporations and billionaires rather than regular people trying to participate in democracy. The founders wrote the First Amendment to protect a pamphleteer criticizing the king. I don’t think they were trying to protect Novo Nordisk’s right to convince skinny people they need Ozempic, or TikTok’s right to sell your location data to scammers, or Elon Musk’s right to radicalize teenagers without anyone being allowed to ask how Grok works. But here we are. We need several new amendments to clarify what the old one actually meant.
Let me make this a lot simpler.
The First Amendment is all about freedom of conscience. Period. No other factors need be introduced in order to completely undestand the purpose of the Amendment.
Conscience very often involves religious belief. But certainly there is nothing about religious belief that is special among all the different ways human beings concieve of ethical or moral behavior.
Because the First Amendment is about freedome of conscience, it can only apply to entities which HAVE a conscience; which is human beings, so far as we know. It CANNOT appy to a machine. Or to a corporation.
Declaring that a machine or a corporation has "rights" under the First Amendment is a perversion of that part of the Constitution. It is a manipulation, designed to further select economic interests, and has nothing whatsoever to do with freedom or liberty. Quite the opposite, actually.