is there a limit to how much you can learn? Is there a point where a human neural brain won’t be able to store any more new information?

2.68K views

Obviously currently this doesn’t happen, but assuming there would be some techniques or devices for learning more – would there be a limit to how much our neural networks in the brain can store?

In: 802

92 Answers

Anonymous 0 Comments

While the exact neural network topology used in mammal brains for something like memory storage isn’t really fully known, you can take an area of the brain like the hippocampus in isolation and see that its behavior should follow at least some models of artificial neural nets (ANNs). (ANNs can be a good model for many systems within brains that are well-studied, and this is knowable since now there are connectomes — complete neuron wiring maps, and thus neuron-by-neuron simulations — available for some animals.) ([There are several kinds of memories stored in several areas of the brain](https://qbi.uq.edu.au/brain-basics/memory/where-are-memories-stored), so one must always be cautious about over-generalizing anything.)

So let’s look at an ANN designed explicitly around memory storage — the [modern Hopfield net](https://en.wikipedia.org/wiki/Hopfield_network#Dense_associative_memory_or_modern_Hopfield_network), which is a generalized version of the classic model of computational neuroscience. If you push it to the limits of storage capacity, with or without error tolerance (discussed in ref #10), the number of memories you can store scales exponentionally with the number of (feature, or the input-output) nodes/neurons (*N*) as

[*Max number of memories*] =~ 2^(*N*/2)

Of course, there are trade-offs when you optimize a network for storage like this — of particular concern is the reliability and error-tolerance in recall, which drops significantly in the modern Hopfield net as you push for more capacity. (However, it doesn’t become zero, which is the point.) This is in extreme contrast to what is taught (or what I was taught) for the “typical” neural net (multilayer perceptron), which is to expect a capacity somewhere from Log *N* to *N*^2 (iirc, and I’m skipping details), for which architecture the number of neurons in the hippocampus may not be able to explain the number of memories we humans seem to accurately consistently recall.

(For a super-silly estimate, [the human hippocampus contains about 40 million neurons](https://pubmed.ncbi.nlm.nih.gov/2358525/) — though only parts are used in memory storage, it’s still gonna be millions, so let’s say *N* = 1 million. If the network architecture for storage is some sort of perceptron (the typical ANN I mentioned) then that would maybe allow 1 trillion memories. If it’s a modern Hopfield net, however, then the 2^(500k) memory capacity overflows a basic big-number calculator — it is effectively infinite.)

One issue with applying optimized models like these to biological systems is that a neural network in the brain is limited to the functionality that the cells and glia come with. Changing the update function in something like a modern Hopfield net is a single line of code (and no retraining/relearning), while an analogous change for neurons in the brain would probably require creating a new kind of neuron altogether for that purpose. (Keep in mind that there are many different types of neurons in the brain and periphery, and many more types of glial cells, all with different response “functions” to incoming signals).

You are viewing 1 out of 92 answers, click here to view all answers.