Our brain interprets visual information by combining what we see with what we already know. A study published in the journal Neuron by researchers at the Champalimaud Foundation, and supported by the CaixaResearch Health Call of the la Caixa Foundation, reveals a mechanism for learning and storing this existing knowledge about the world. They found that neurons are wired to connect seemingly unrelated concepts. This wiring may be crucial for enhancing the brain’s ability to predict what we see based on past experiences, and bring us a step closer to understanding how this process goes awry in mental health disorders.
Credit: David Marr (Book: Vision by David Marr, MIT Press)
Our brain interprets visual information by combining what we see with what we already know. A study published in the journal Neuron by researchers at the Champalimaud Foundation, and supported by the CaixaResearch Health Call of the la Caixa Foundation, reveals a mechanism for learning and storing this existing knowledge about the world. They found that neurons are wired to connect seemingly unrelated concepts. This wiring may be crucial for enhancing the brain’s ability to predict what we see based on past experiences, and bring us a step closer to understanding how this process goes awry in mental health disorders.
How do we learn to make sense of our environment? Over time, our brain builds a hierarchy of knowledge, with higher-order concepts linked to the lower-order features that comprise them. For instance, we learn that cabinets contain drawers and that Dalmatian dogs have black-and-white patches, and not vice versa. This interconnected framework shapes our expectations and perception of the world, allowing us to identify what we see based on context and experience.
“Take an elephant”, says Leopoldo Petreanu, senior author of the la Caixa-funded study. “Elephants are associated with lower-order attributes such as colour, size, and weight, as well as higher-order contexts like jungles or safaris. Connecting concepts helps us understand the world and interpret ambiguous stimuli. If you’re on a safari, you may be more likely to spot an elephant behind the bushes than you would otherwise. Similarly, knowing it’s an elephant makes you more likely to perceive it as grey even in the dim light of dusk. But where in the fabric of the brain is this prior knowledge stored, and how is it learned?”.
The brain’s visual system consists of a network of areas that work together, with lower areas handling simple details (e.g. small regions of space, colours, edges) and higher areas representing more complex concepts (e.g. larger regions of space, animals, faces). Cells in higher areas send “feedback” connections to lower areas, putting them in a position to learn and embed real-world relationships shaped by experience. For instance, cells encoding an “elephant” might send feedback to cells processing features like “grey”, “big” and “heavy”. The researchers therefore set about investigating how visual experience influences the organisation of these feedback projections, whose functional role remains largely unknown.
“We wanted to understand how these feedback projections store information about the world”, says Rodrigo Dias, one of the study’s first authors. “To do this, we examined the effects of visual experience on feedback projections to a lower visual area called V1 in mice. We raised two groups of mice differently: one in a normal environment with regular light exposure, and the other in darkness. We then observed how the feedback connections, and cells they target in V1, responded to different regions of the visual field”.
In mice raised in darkness, the feedback connections and V1 cells directly below them both represented the same areas of visual space. First author Radhika Rajan picks up the story, “It was amazing to see how well the spatial representations of higher and lower areas matched up in the dark-reared mice. This suggests that the brain has an inherent, genetic blueprint for organising these spatially aligned connections, independent of visual input”. However, in normally-reared mice, these connections were less precisely matched, and more feedback inputs conveyed information from surrounding areas of the visual field.
Rajan continues, “We found that with visual experience, feedback provides more contextual and novel information, enhancing the ability of V1 cells to sample information from a broader area of the visual scene”. This effect depended on the origin within the higher visual area: feedback projections from deeper layers were more likely to convey surround information compared to those from superficial layers.
Moreover, the team discovered that in normally-reared mice, deep-layer feedback inputs to V1 become organised according to the patterns they “prefer” to see, such as vertical or horizontal lines. “For instance”, Dias says, “inputs that prefer vertical lines avoid sending surround information to areas located along the vertical direction. In contrast, we found no such bias in connectivity in dark-reared mice”.
“This suggests that visual experience plays a crucial role in fine-tuning feedback connections and shaping the spatial information transmitted from higher to lower visual areas”, notes Petreanu. “We developed a computational model that shows how experience leads to a selection process, reducing connections between feedback and V1 cells whose representations overlap too much. This minimises redundancy, allowing V1 cells to integrate a more diverse range of feedback”.
Perhaps counterintuitively, the brain might encode learned knowledge by connecting cells that represent unrelated concepts, and that are less likely to be activated together based on real-world patterns. This could be an energy-efficient way to store information, so that when encountering a novel stimulus, like a pink elephant, the brain’s preconfigured wiring maximises activation, enhancing detection and updating predictions about the world.
Identifying this brain interface where prior knowledge combines with new sensory information could be valuable for developing interventions in cases where this integration process malfunctions. As Petreanu concludes, “Such imbalances are thought to occur in conditions like autism and schizophrenia. In autism, individuals may perceive everything as novel because prior information isn’t strong enough to influence perception. Conversely, in schizophrenia, prior information could be overly dominant, leading to perceptions that are internally generated rather than based on actual sensory input. Understanding how sensory information and prior knowledge are integrated can help address these imbalances”.
Journal
Neuron
DOI
10.1016/j.neuron.2024.07.009
Method of Research
Experimental study
Subject of Research
Animals
Article Title
Visual experience reduces the spatial redundancy between cortical feedback inputs and primary visual cortex neurons
Article Publication Date
12-Aug-2024