Progress in understanding of how the brain works
Yehouda Harpaz
yh@maldoo.com
last modified 10 Jul 2025
related texts 

Progress in understanding of how the brain works

Most of the discussion in this site of current research are highly critical (Errors, Myths). This page is intended to highlight what looks to me like progress.

[ 10 Jul 2025 ]: By now I am not sure if any the cases I wtite about here is actual progress. Each case gets over some of the logical and methodological errors, but then get stuck on other errors.

1. The brain is always active

[6 Aug 2007]

1.1) In my model I wrote as one of the major hypotheses that the System (i.e. the brain) is always active (Major Hypothseis 6, 4.5.3). I also wrote that it looks to me obvious (6.3.2.6).

1.2) At least 99% of published articles in the area are effectively based on the assumption that the activity of the brain when it does not explicitly do something is of no interest. However, this seems to change. For example, in this article (Spontaneous Activity Associated with Primary Visual Cortex: A Resting-State fMRI Study, Wang et al, Cerebral Cortex Advance Access published online on June 29, 2007 ; Full text here) they look at activity of the resting brain, and say that "This confirmation supports the perspective that brain is a system intrinsically operating on its own, and sensory information interacts with rather than determines the operation of the system."

1.3) A much stronger support to their conclusion is the observation that the activity of the "resting brain" is far larger than the changes that are normally reported, which they discuss in the paragraph following the abstract. As the references that they mention (one from 1955) show, that is an old fact. What is progress is the fact that they are actually looking at a resting brain, and interpret it as ".. operating on its own..".

1.4) They (and apparently the papers by Raichle that they quote) are still worried of "going too far". They say "Therefore, as suggested by Raichle and colleagues, in terms of overall brain functions, the ongoing intrinsic activity within various brain systems may be at least as important as the activity evoked by external stimuli (Raichle and Gusnard 2005; Raichle and Mintun 2006)." Thus the intrinsic activity is only ".. at least as important..", rather than the obvious "much more important", so we still have some distance to go. But there is a progress in the right direction.

2. Do neurons "represent" anything?

[6 Oct 2007]

In this article (Churchland MM, Shenoy KV (2007) Temporal complexity and heterogeneity of single-neuron activity in premotor and motor cortex. Journal of Neurophysiology. 97:4235-4257doi ), they find that the neurons in the premotor and motor cortex show complex and heterogenous activity. From this, they suggest the possibility that these neurons don't represent anything.

As discussed here, neurons don't represent anything, but many cognitive scientists and neurosceintists seem to be unable to comprehend this possibility. The authors of the article above are clrearly capable of comprehending it. They present alternative view, but seem to regard their observations as showing that the neurons that they look at do not represent anything. They present the same view in other articles (list of publications, for example the one about "Reference frames for reach planning in macaque dorsal premotor cortex").

The fact that they positively think that the neurons do not represent is progress. The progress is limited, however, because it is still based on the assumption that showing correlation between neural activity and something else (behaviour, stimulus) shows representation. Therefore "representationalists" can still believe that these neurons represent something else.

3. looking at ensembles of neurons.

[2 Dec 2007]

In this article (Natural stimuli evoke dynamic sequences of states in sensory cortical ensembles, Jones et al, PNAS | November 20, 2007 | vol. 104 | no. 47 | 18772-18777 (open access article)), they analyze ensembles of neurons, and show that the ensemble response correlates much better with the input than single-neuron analysis. They stress that the single-neuron analysis that is normally used loses information.

Their data is about taste in rats, and it is not obvious how it is going to generalize to other senses. The number of neurons in each "ensemble" is also pretty small (10). But it is encouraging to see researchers that look at ensembles, and explicitly state that single-neuron analysis loses information. They explicitly state that the coherent state sequences that they see do not represent sensory codes, which is also progress.

4. "The Brain Activity Map Project and the Challenge of Functional Connectomics"

[29 Aug 2012]

That is the title of this article (doi) (A. Paul Alivisatos, Miyoung Chun, George M. Church, Ralph J. Greenspan, Michael L. Roukes, Rafael Yuste; Neuron - 21 June 2012 (Vol. 74, Issue 6, pp. 970-974)). The main point about this is that they recognize that you need to look at the network rather than individual neurons to understand what it does. For example, they start the summmary by saying: "The function of neural circuits is an emergent property that arises from the coordinated activity of large numbers of neurons."

They criticize current studies, and for example say (end of first paragraph of "Emergent Properties of Brain Circuits"):

However, neural circuits can involve millions of neurons, so it is probable that neuronal ensembles operate at a multineuronal level of organization, one that will be invisible from single neuron recordings, just as it would be pointless to view an HDTV program by looking just at one or a few pixels on a screen.
Thus they suggest quite strongly that single neuron studies are useless. That is definitely progress.

As to their suggestion, I think they are over-optimistic about our ability to measure activity in living networks. I suspect that recognizing this limit is the reason that they and other reasearchers did not advocate network research like that until now, and perfomed and supported single-neuron research even when it is clear that it is useless (in complex animals, that means almost always), and that they would not feel able to say it without the optimistic predictions.

Because of the technical issues, it is not obvious that their suggestion is the best approach. But thinking about the actual real networks rather than about single neurons or artifical networks is a large step forward.

4. "The importance of mixed selectivity in complex cognitive tasks"

[1 Jun 2013]

See here. Looks like real progress.

5. "Small sample sizes reduce the replicability of task-based fMRI studies"

[29 Dec 2018]

See here. Looks like real progress.

6. "In vitro neurons learn and exhibit sentience when embodied in a simulated game-world"

[23 Oct 2022]

Here they play around with millions of neurons on a system with many electrodes for stimualtion and some reading channels. They then teach this system to play Pong. They show that the teaching improved the performance of it in playing Pong. It should be noted that this research is done mainly by a commercial entity (Cortical Labs).

"Playing Pong" here is a very simple repsonse. It means just increasing/descreasing the activity in some areas of the system, in response to the input about the state of the game. The teaching is done by giving the system random feedback when it fails, and fixed (i.e. the same each time) input when it succeed. They call the fixed input "predictable", but since their system clearly doesn't predict anything, the predictability is irrelevant.

The improvement in performance is quite unimpresive (see their figure 5), but it does look significant. It is diffcult to say how real it is.

The main progress in this article is probably the demonstartion of the technical fit of working with large number of neurons over a long period of time.

On the theoretical side, it is a large progress that they are trying to do "computation" with neurons themselves, rather than using models of neurons. As a result, they cannot use any unrealistic feature, which is on its own a significant progress.

The feedback thay they give (random for failure vs. fixed for success) can be compared to the "Cognitive Positive Outcome" which I hypothesized as the driver for learning (See Major Hypothesis 12 in my model). The negative outcome in my model is speeding up the random activation of micornodes, while in this paper it is random input. That is essentailly the same.

The positive output on my model is simply slowing down random activity, while in this paper they give some fixed feedback. That is not obviously the same, but if we assume that the fixed feedback cause their system to to tend to stay in a specifc place in the "activity space" all the time (not only when it is on), the fixed feedback will tend to not cause their system large disturbance as the random input, so in this sense it is similar to my idea. They didn't actually check if the feedback has a large or small effect on the activity in the sytsem, which would have given us a hint if this assumption is correct.

[26 Oct 2022] I sent an email to the first author asking whether they measured the effect of the freedback on the activity, or maybe can compute it. He replied that they didn't compute it, and it would be difficult to do it now, but it is something they may do in the future. Apparently they haven't thought about it.

That is on its own a quite interesting observation, because the feedback affects the behaviour of the system by the way it changes the activity of it. So it looks obvious that it is something that it would be useful to measure, but apparently it is not obvious, and these people haven't considered it.

In their theoretical discussion they try to wrap the results in theoretical ideas of entropy and surprise, which are just rubbish (second and third paragraphs of the Discussion). But they seem to realize that they don't actually have a proper model.

It is not obvious how much impact this kind of research will have, which will depend on how attractive researchers will find it. On one hand, it shows that you can try to look at computations in actual neurons, shich is attractive. But the other hand, you can do much less this way than you can do with computer models (at least in the near future), which makes it less attractive.

7. "Beyond binding: from modular to natural vision"

[6 Jul 2025]

In this "opinion" they argue against modularity in vision, mainly based on studies of brain damage (leison studies). The idea of modularity in the human cerebral cortex (where visual input is processed) is obviously non-sense, as I wrote long time ago here: The modularity assumption, but it is holding (wikipedia entry), so actually publishing refutation of it is significant progress.

It is not obvious if the authors of this article realize the dumbness of the idea of midularity or not. They give the impression that they believe it was sensible to believe it until lately, when (in their words):

However, as our experimental and analytical tools have become more sophisticated, evidence has emerged that challenges this framework’s fundamental assumptions.
It is possible they write it like that because if they actually point the dumbness of "modularity" their "opinion" will be rejected.

Exchanges with the CoPilot of Bing suggests that the obsfucation of the discussion by using a quasy-redefinition of "modularity" a to mean "biased response", which I wrote about in The modularity assumption, is by now more common. Not obvious whether this is progress or not.

The "modules" in vision modularity are supposed to be associated with specific input attributes. The advance of machine learning system based on activations networks (for example, AlphaFold, LLM and its various derivatives) should make it much easier to see that complex processing does not require such modules. That may make it easier to realize how stupid is the assumption of vision modularity.

8. "On the origin of memory neurons in the human hippocampus"

[10 Jul 2025]

The description of the appearance of memory of an episode (i.e. the ability to recall it) in this article is not that far from my ideas in the description of the ERS (Major Hypothesis 8). That is at least some progress.

The main difference is that I write in terms of neurons in a network only, while they also used various junk concepts, like "representation", "engram", "allocation (of neurons)", "neurons compete". They have a problem with labeling neurons (good), but still use labels anyway(bad). For example, in the "Concluding remarks and future perspectives" (last part) they predict (third paragraph):

"the transformation from index neurons to concept neurons should be accompanied by specific changes, such as shifts in firing patterns or connectivity, "
So they could not bring themselves to the idea that that the labels are useless and do not contribute anything. Of the changes they mention, connectivity is clearly attribute of the network, not a single neuron, and the firing patterns will change as result of changes in the connectivity (assuming "changes in the connectivity" here include changes in synapse strengths too).

If we translate their prediction to rpoper terms, i.e. neurons in the hippocampus will change to be permannetly linked to some episodic memory, I think it is wrong. The permanent memory is going to be in the cortex, we know that from people that lost their hippocampus and lost alomost all of their ability to form new memories, but retailed old ones.