(This post’s featured image is not a photo of the idyllic California vineyard my wife and I visited in 2015, but a similar looking AI-generated fiction)
The excitement around large language models (LLMs) continues. Often just called “AI,” this new technology takes instructions in plain English, and generates new text or images so good you’d think a human made them. There are some serious concerns about the ethics of how this is done (see the Dangers of Stochastic Parrots), and many articles warning people that LLMs aren’t as human as they seem. Still, these LLMs are clearly doing something smart, and it’s weirdly compelling. What’s going on here? What can LLMs teach us about the human mind, our strengths and weaknesses, and why we’re so easily fooled and mesmerized by this tech?
There are lots of articles about how LLMs work and their limitations (I even wrote one a while back), so I won’t go into much detail here. What matters for now is that LLMs are prediction engines. Given some text, they try to guess how a human would respond, based on a statistical analysis of all the content on the internet. They do this extremely well, but not perfectly. LLMs don’t think, perceive, or interact with the world in a human-like way, so sometimes they make weird mistakes a human never would.
In some ways, the human brain works a lot like an LLM. From day one, the brain is looking for language. It automatically builds up vast networks of words, concepts, and their relationships. When I hear someone talking, my mind is suddenly filled with the speaker’s ideas and their associations, building up an image in my mind’s eye. This is basically what an LLM does.
However, I also have many other intelligent faculties that join in. I don’t just know that “dog” is a word related to “cat.” I have memories of specific dogs, what their fur felt like, and the experiences we shared. I have common sense, logic, and theory of mind to decide whether something I hear is truth, fantasy, error, or deception. I monitor my own thoughts to gauge my level of confidence and correct my mistakes. I anticipate the future, make plans, and use language to achieve my goals. LLMs don’t do any of that.
Of course, there are AI researchers working to approximate these other kinds of human intelligence (though progress here is limited compared with LLMs themselves). Rebooting AI is a great book exploring that work, which argues its just a matter of time before AI can do everything a human can. Personally, I’m skeptical that we’ll ever reverse engineer all the subtlety of human thought, but I think it’s safe to say that AIs will become more powerful and well-rounded in the future. Perhaps the more important question is whether building increasingly realistic human simulations is a good idea at all.
For now, LLMs are a bit of a one-trick pony. What’s scary is that one trick is often good enough to fool humans and do useful work. In particular, even though LLMs were designed to be text prediction engines, they are surprisingly good at general problem solving. They can paint pictures, do math, solve brain teasers, and even write and simulate computer programs. Maybe those “extra” faculties of the human mind aren’t so important after all?
LLMs “think” in terms of words and relationships and patterns they’ve seen before. In human terms, that means stereotypes, cliches, generalizing from past “experience,” and repeating what they are told. We like to think that human thought is more sophisticated than that, but it often isn’t. We sometimes don’t see people as individuals, but in terms of the role they play in society (ie, “barista” or “mom”). We make decisions based on rules of thumb or gut feeling, without the need for logic and reasoning. We talk about things we don’t fully comprehend. We repeat talking points in order to fit in with our tribe. We confidently make up nonsense just to satisfy each other and move on. It’s surprising how LLM-like humans can be sometimes.
And that’s not meant to be derogatory! Those ways of thinking can be very effective. A lot of language isn’t about complex ideas, comprehension, and reasoning, but just putting one word after another to evoke an image in someone else’s mind. Past experience often is a highly effective and low-effort way to predict the future. Lying is anti-social, but “fake it ‘til you make it” works. One of the fastest ways to pick up a new skill is to boldly make mistakes, get feedback, and learn from that experience. The main difference is that today’s LLMs don’t learn from their mistakes, they never doubt their “intuition,” and they have no alternative ways of “thinking” when these techniques fall short.
So, yeah, LLMs only do part of what humans do, but it’s a big and important part. Occasionally we do need facts, critical thinking, self-doubt, and all the rest to do the right thing, but they don’t come up as often as we like to think. The real danger of LLMs, then, is that 80% of the time they might be good enough, but 20% of the time we need fancy human judgment to notice they screwed up and decide what to do. This is a serious problem. Humans are bad at vigilance, and we have a strong instinct to trust language, which in this case is exactly the wrong response.
Language is a defining feature of our species. We aren’t just capable of language, it’s a human universal. Every culture has language. Babies attend to speech from the moment they’re born, and start to babble in a few months. When there isn’t a common language spoken around them, children raised together will spontaneously invent one. Language is a biological imperative for us. It’s in our DNA.
When we perceive language, our minds automatically assume that it’s communication. We imagine another mind behind the words, usually with good intent and a desire to cooperate. Up until recently, this was a pretty safe assumption, so it was totally reasonable for the brain to immediately and automatically translate language into meaning. But now this instinct is backfiring. LLMs create realistic text and imagery without any intentional meaning. They don’t produce “answers,” “opinions,” or “art,” just random content that looks like those things. It’s both very difficult and important to remember that.
We’re still working to understand how these LLMs work, what their limitations are, and what they’re good for. As we do that, I hope we’ll come to understand ourselves better, too. What do you think? Have LLMs made you think about minds any differently? Have you seen any interesting examples of AIs acting strange or foolish? What about people acting like LLMs? Any thoughts or fears about computers gradually inching toward human-like abilities? I’d love to hear from you in the comments.