(The image for this post is a photo of a jacquard loom I took in the workshop of Luigi Bevilacqua in Venice. It’s a heavy wooden frame covered in pulleys and string, in the midst of weaving fine velvet in gold and red. The loom is automated using punch cards, and served as a model for early computers. It’s not something we’d usually think of as “AI,” but it is a symbolic mechanism that automates a complex human behavior.)
The term “Artificial Intelligence” has been around since the 1950s, and it’s always been ambiguous. Generally, AI is about reproducing the intelligence of living things using math and machines, and there have been many different approaches to that problem. However, in the past decade or so, the word AI has become synonymous with a class of algorithms known as Deep Learning (DL). This technology has produced stunning results and has rapidly been integrated into software of all kinds. This has led to mixed reactions, including ethical concerns and active debates about what “human-level general intelligence” means and whether or not AI has it. But what does DL actually do? How is it like and unlike human intelligence? Let’s at least scratch the surface of this important question.
Personally, I’m frustrated by the AI field’s obsession with DL. We act like brains are the secret to intelligence, DL is just a “brain in a computer,” and anything else is of marginal interest. But, in truth, the brain is just one small part of a vast and diverse intelligent system. Consider:
- Evolution “designed” sophisticated solutions to real-life challenges without the use of top-down engineering or even conscious thought.
- Organisms of all kinds perceive, analyze, decide, and react to the world in real time, with or without a brain.
- Brains come in all shapes and sizes, from simple to complex, with many species-specific architectures and special-purpose modules for things like sensory perception, emotions, memory, and motion planning.
- Mammals have an additional brain structure called the neocortex (birds have an analogous structure called the dorsal ventricular ridge) which provides a layer of abstract cognition on top of the older, more specialized parts.
- Individuals compete and collaborate to form ecosystems, colonies, and societies that are intelligent in their own right.
So the brain itself is only a small part of a vast network of intelligence. But also, DL is only sorta like one part of the brain. Neural networks as they exist in DL are very loosely inspired by the fine structure of the neocortex. I like to think of that as the “cognitive fabric” from which the neocortex is built. Evolution has shaped that fabric into special-purpose brain regions, each tuned to solve different problems. These regions are networked together into a particular architecture, providing multiple layers of analysis, careful mixing of perceptions and cognitive faculties, and multiple kinds of self-monitoring. All of that is orchestrated by the lower-level brain structures, which still define the basic emotions, modes of thinking, flow of thought, and the relationship between abstract mental activity and the concrete needs of the body. Generally speaking, DL ignores all of that structure.
If DL captures just one facet of our mind’s intelligence, then what does that part do? It observes data, finds patterns, and learns stereotypes about what it sees. It can apply those stereotypes to extrapolate rich and coherent scenes from noisy fragments of data, filling in gaps with reasonable guesses. The brain uses this tool everywhere, and it’s a crucial ingredient for how humans perceive and think about the world. DL makes that tool available to software developers.
What makes our minds “human like” is the evolved structure that applies this pattern matching / stereotyping faculty in particular ways that generate our perceptions, intuition, biases, self-awareness, train of thought, attention, and dreams. When people make algorithms using DL, we provide the structure that determines how the AI uses that faculty, and thus how it behaves. They resemble the human mind only as much as we try to reproduce the human thought process into our code. We typically don’t, and that’s probably a good thing. Attempting to bring something human-like to life in a computer sounds even more ethically problematic than cloning. Recent experiments into “chain of thought” for language models are a step in this direction, though they aren’t really trying to make the model “think like a person” so much as to get high scores on tests for “reasoning skills,” which is not the same thing.
This raises some interesting questions. Can DL algorithms really understand the world? Sorta. Large language models like GPT provide an interesting example. By consuming vast quantities of text, these algorithms can master nuanced patterns in words that reflect human ideas and the the physical world. They understand these ideas well enough to use them, interpreting and generating both text and images. Yet, they only experience the physical world indirectly, so in some sense they don’t fully understand, and they get many details wrong. It’s an open philosophical question just how different human and machine understanding really are.
Do DL algorithms think or have desires? Generally, no. Most often DL is used to implement a function, in the mathematical sense. They take some input (i.e., an image) and produce some output (i.e., a label for that image). That is the entire scope of their existence. They don’t reflect, compare alternatives, or make decisions. They have no needs to fulfill, and no way to perceive themselves or their environment. More complex DL architectures start to blur the line, though. We give them analogs of memory and attention. In reinforcement learning, we even use DL to make agents that inhabit virtual worlds, have a sense of self, and make their own choices. Perhaps these algorithms could be said to “think,” but their “minds” are alien, adapted to a world of experiences totally unlike our own.
Do we need to worry about AI taking over the world? No, but also yes. The Terminator scenario seems unlikely. Those evil robots are human-like, in ways current AI cannot even begin to approach. In particular, they want to destroy humanity, and take the initiative to act on that. Today’s AI has no desires, and does nothing until prompted. However, there are other, more realistic concerns. Today, we mostly use ML for two purposes: to help computers understand human expression, and to automate human behaviors. Both of these can be problematic, especially if we (incorrectly) assume these algorithms think like people do.
The real danger is trusting these algorithms too much. DL is incredibly good at one thing: stereotyping. It does not have any notion of cause and effect, common sense, morality, or logic. Stereotypes can be effective shortcuts to solving hard problems, but they can cause real harm. Think of Microsoft’s racist chatbot, Google’s smart camera that can’t see Black people, or the tyranny of “the algorithm” in social media. When we allow DL algorithms to understand data for us, or make decisions that influence our lives, we’re trusting a system that has no judgment, sense of consequences, or accountability. That’s taking a big risk, and usually it’s just a few people at a tech company making the decision for millions of others around the world.
I have mixed feelings about DL. It’s an incredible tool, it does some really cool stuff, and it has already created tremendous value for society. It has also done a lot of damage, especially to minority communities. I’m concerned about all the hype, and how rapidly we’ve integrated DL into every facet of life. We don’t understand this technology well enough to know what the consequences will be. I also hate that our focus on DL has blinded the field of AI to other opportunities. Life is full of brilliant designs! By exploring more broadly, we might find other useful tools, but also come to understand ourselves better and how we fit into the bigger picture of living intelligence. Isn’t that more important?