(This post’s featured image is Slicing the Connectome by Alfred Anwander, a visualization of the deep structure of the human brain. It shows flowing bundles of thin fibers, color coded by their direction of travel and which parts of the brain they integrate. Used under the Creative Commons ShareAlike 4.0 International license)
Artificial Neural Networks (ANNs) are a revolutionary technology. They’re the foundation of Deep Learning (DL), the technique that’s brought us AI systems that understand images, text, and speech, can beat us playing chess, go, and StarCraft, or even drive a car. These models are supposedly “brain-like.” We call them “neural” after all, and they can do many things people do, but how much are they like us, really? Some argue that in a few years we’ll scale this technology up and find that AI is smarter than we are! But if we take a closer look at just how “brain-like” these systems are, there’s good reason to doubt that. Is “bigger” all we need, or is a human mind something fundamentally different?
It’s worth starting with how brains and ANNs really are similar to each other. ANNs were inspired by observations of actual brains, both studying their anatomy under a microscope, and watching patterns of electrical activity in living subjects. Our understanding of how brains work is still very crude, but there are some general principles we can glean that seem very important. The brain is a vast network of small, relatively simple modules (neurons). Each one is connected to a few or even hundreds of other neurons in a vast tangled network. These neurons sometimes “spike” with electrical activity, and when one does it often nudges its neighbors to spike, too. Information can flow, transform, and cascade as it propagates through the network. Often many neurons fire together, and it seems that the meaning they represent is actually held in the collective firing patterns across the whole network. That suggests what matters in terms of scale isn’t how many neurons there are, but how many different combinations of neurons, which is vastly more. Learning seems to involve the neurons “re-wiring” with each other, adding or removing connections (synapses) to other neurons, or changing their strength.
Computer scientists were inspired by this picture of the brain, seeing it as a sort of programmable circuit. Each neuron does just a small, simple operation, but by wiring them together into a large network you can make more complex functions. Importantly, such a network could represent nearly any function, depending on the particular way the neurons get wired up together. For this reason, ANNs are often known as “universal function approximators.” Another common way of understanding ANNs is through the math used to implement them. Each neuron does a simple operation: it takes in signals from all the synapses that feed into it, multiplies those signals by “weights” that determine how much this neuron listens to each of its neighbors, adds them up, then performs an “activation function” on the result. Often, that’s as simple as returning the final sum, or zero if that sum is negative. ANNs use large arrays of identical neurons, and the final output is basically just a math expression that does all those additions and multiplications. The synapse weights are variables (or “parameters”) in this expression, and “training” a neural network is simply finding the right values for all the variables so that the math expression resolves to the answer you want, rather than something else.
But there are an infinite number of possible functions. If an ANN can represent any of them, how do you get it to be the function you want and not just some random garbage? To start with, you need to have a whole bunch of data. Either you have some task you want the neural network to do and correct answers you want it to learn (say, the name of what appears in an image, or a translation from one language to another), or you have many examples of something you want the network to generate more of (like text or images). Then, you can use an external process that evaluates the network, compares the results to the target, tweaks the connections between neurons to make the result a little better, and then does that over and over again. Researchers have found clever ways to make this training process as efficient and reliable as possible, but it can still be painfully slow and expensive, using vast amounts of data and computing power. Generally you never get a perfect result, so you carry on training until the outputs are “good enough” to satisfy the programmer. Then training stops. All the trainable parameters in the ANN are locked down. What you have left doesn’t learn at all, it’s just a math function that does… whatever it was configured to do.
This already suggests a massive difference with the human brain. For a human being, there is no “right answer” to most challenges we face. There may be serious consequences for getting the answer wrong, but this may not be obvious in the moment. This makes learning how to respond vastly harder, because we have to set our own goals, and intuit for ourselves if we got it right, now and across time. There’s no “external” training process to rewire the neurons, either. Each neuron manages its own connections, somehow knowing which ones to strengthen or weaken without any top-down picture of the network they are making or the problem they are trying to solve. There’s no “training phase” where the brain gets configured once and for all. No, we learn continually as we act, which means we adapt continuously, changing our understanding and our strategy in context (unlike ANNs). It’s even more complicated than that. Animals with brains seem to have an “offline” learning process, too; dreams play an important role in memory consolidation and mastering new skills. It’s not clear how that relates to other kinds of learning and it seems very different from the training process we use for ANNs.
Another major difference is how ANNs are organized. DL’s breakout success came from computer vision models inspired by the visual cortex. Brain researchers showed that animal vision depends on layers of neurons, arranged in a hierarchy. The low-level ones detect primitive patterns like edges of a particular orientation. Then higher layers use those edge detectors to find higher-level patterns, like shapes and objects. For a long time, computer vision systems worked this way, too (though, “vision transformers” are starting to change that). Much effort was put into making systems with more layers in order to learn more complex functions, which is where the “deep” in “Deep Learning” comes from. The hope was that if we made them much larger, they would perform much better, and so far that’s mostly been true.
But are deep, hierarchical ANNs really brain-like? They seem to be vastly smaller and simpler, actually. For one thing, it’s estimated that the human brain has 86 billion neurons with 150 trillion connections between them. By contrast, GPT5 is estimated to have 2-5 trillion parameters, but this is likely not a fair comparison. Each living neuron is vastly more complicated than the simple “weighted sum” of a neuron in an ANN. We don’t fully understand what neurons do (not to mention all the other kinds of cells in the brain, which may also play critical roles in thinking and learning), but they are living cells that autonomously manage their activations and their relationships with other cells. Unlike ANNs, where each neuron produces a single value and all neurons activate at the same time, biological neurons communicate in complex ways across multiple channels with patterns that change over time, at multiple frequencies. That means each biological neuron may be doing much larger and more complex tasks than their artificial counterparts. But aside from being a few orders of magnitude bigger than an ANN, the brain also has a much more complex structure.
The outermost part of the human brain is called the neocortex, and it’s where most of our “higher thought” happens. It’s broken down into regions which each serve a specialized function, like vision, language, logic, or social interaction. But these regions don’t look very different from each other, anatomically speaking. We mostly discovered this structure by measuring electrical activity while a subject sits in a scanner thinking about something. The neocortex is more or less uniform in its structure, but different parts are assigned different roles, and the deep structure of the brain connects them up in a very particular way. Human brains have a highly stereotyped layout, with roughly the same set of faculties interacting with each other in roughly the same way in every person. Except, of course, when something goes wrong. Brains are also very plastic, and can reorganize around damage, and adapt to patterns of use. For instance, people born blind often show activation in the visual cortex when they read braille. I like to think of the neocortex as a sort of general purpose “cognitive fabric.” Evolution, development, and learning subdivide it into modules, then compose those modules into a particular cognitive architecture that suits a species’ lifestyle. Brain cells are like an evolved platform for building brains and minds of all sorts.
If the “universal function approximator” metaphor applies to brains, then brains have vast numbers of them, arranged into a complex network, where information flows through paths that diverge and converge and loop back on each other in very particular ways. It would be like a whole computer, specialized for the purpose of generating the human mind, built from these programmable circuits. If so, then what makes our minds “human-like” is not the fine structure of the brain that inspired ANNs, but the coarse structure that determines the shape and flow of our thoughts. In nature, that design is evolved, but flexible, and it continuously adapts to meet demands. By comparison, a programmer working with ANNs chooses an architecture by intuition, it’s usually very simple, and then it never changes. Researchers are continually exploring different kinds of neural modules and ways of composing them together, but they’re all much simpler than actual brain networks, which really aren’t the inspiration any more. AI researchers have found ways to do useful work with ANNs, and don’t care whether they resemble brains or not.
But we shouldn’t commit to the “function approximator” metaphor too quickly. Some would argue the brain is nothing at all like a function! An ANN is a function. It is a math expression that describes how to turn inputs to outputs. When a programmer evaluates that expression, they get an answer, but otherwise the ANN does nothing. It’s just a pile of math. By contrast, the mind is always active, whether it’s observing the senses, controlling the body, thinking hard, or just daydreaming. In some sense, the brain has a narrow interface with the body and the rest of the world through the brain stem. If the brain implements a function, this could be where the inputs and outputs pass… except, it’s not clear where the mind ends and the body begins! The nervous system extends throughout the body, sensing, controlling, and interacting with the various organ systems, often without checking in with the brain at all. The human gut (and the microbes that live there) play a role in cognition and can shape our mood and behavior. It’s probably not the only organ system that does that, either. The mind is deeply integrated with the body, driven by its needs, cravings, and instincts. The mind is also creative, generating its own inputs, setting goals, identifying problems to solve and strategies to solve them. This doesn’t seem very function-like at all!
By using the same low-level principle as the brain, ANNs let us automatically design functions to solve a wide range of difficult real-world problems. Often, we do this by observing humans solve some problem, and copying what they do. In short, we choose what human skills to automate, and we make functions that simulate those skills. The AI doesn’t “figure out” the answers; we have to already know the answers, and then bake them into the AI. Using many processor cores running in parallel, we spend the equivalent of hundreds of years of compute time to train our models to approximate what a human can do effortlessly by intuition. Once we pay that up front cost, then that knowledge is baked into the network, unchanging until we decide to rebuild it. In short, all the things that make the human mind open-ended, creative, and continuously adaptive are missing from ANNs. They don’t have the high-level architecture that makes human minds so robust, capable, and well-rounded, and we don’t know how to make anything like that. We have no idea what “thought” even is, let alone how to create it.
There’s no reason to think AI couldn’t become more brain-like and mind-like over time. There’s nothing magical about our biology. But it seems absurd to think that scaling up our current designs will get us there. Modern AIs are very good at mimicking human data. They appear smart, because they reflect and reshape the intelligence we feed into them. But they simply do not do all the extraordinary things our brains do. Most importantly, they do not continuously set their own goals, improve their design, or adapt to changing needs through evolution, development, and learning. No, every innovation in AI is an act of human cleverness imposed onto the machine. This is why I think we have a long way to go. What we’ve built is a toy compared to an actual human brain. We’ve successfully borrowed a few of nature’s tricks, but there’s vastly more there, and we hardly even know what we’re missing. We will need many more clever insights, and those do not come often or easily.













