November 22, 2024

Decoding Visual Deceptions: How the Brain Predicts and Rewrites Reality

Decoding Visual Deceptions: How The Brain Predicts And Rewrites RealityNeuroscience Visual Perception Art - Decoding Visual Deceptions: How The Brain Predicts And Rewrites Reality

A groundbreaking study reveals how the brain’s complex visual system not only perceives but also predicts and alters our view of the world. This in-depth look shows that our brain integrates past experiences with present visual information to construct reality, which can lead to deceptions or illusions. Credit: SciTechDaily.com

Experience shapes neural connectivity: Our brain interprets visual information by combining what we see with what we already know.

Researchers at the Champalimaud Foundation have uncovered how our brains blend visual stimuli with prior knowledge to enhance perception, a key process that could have implications for treating mental health disorders like autism and schizophrenia. Their study, supported by the la Caixa Foundation, reveals that neurons connect disparate concepts, refining our ability to predict and interpret visual information based on past experiences.

Building a Hierarchy of Knowledge

How do we learn to make sense of our environment? Over time, our brain builds a hierarchy of knowledge, with higher-order concepts linked to the lower-order features that comprise them. For instance, we learn that cabinets contain drawers and that Dalmatian dogs have black-and-white patches, and not vice versa. This interconnected framework shapes our expectations and perception of the world, allowing us to identify what we see based on context and experience.

“Take an elephant,” says Leopoldo Petreanu, senior author of the la Caixa-funded study. “Elephants are associated with lower-order attributes such as color, size, and weight, as well as higher-order contexts like jungles or safaris. Connecting concepts helps us understand the world and interpret ambiguous stimuli. If you’re on a safari, you may be more likely to spot an elephant behind the bushes than you would otherwise. Similarly, knowing it’s an elephant makes you more likely to perceive it as grey even in the dim light of dusk. But where in the fabric of the brain is this prior knowledge stored, and how is it learned?”

Meaningless Black and White Patches Dalmatian Dog - Decoding Visual Deceptions: How The Brain Predicts And Rewrites RealityMeaningless Black and White Patches Dalmatian Dog - Decoding Visual Deceptions: How The Brain Predicts And Rewrites Reality

Learned concepts and prior knowledge of what a Dalmatian dog looks like helps you to interpret this seemingly meaningless pattern of black and white patches. Credit: David Marr (Book: Vision by David Marr, MIT Press)

The Role of Feedback in Visual Processing

The brain’s visual system consists of a network of areas that work together, with lower areas handling simple details (e.g. small regions of space, colors, edges) and higher areas representing more complex concepts (e.g. larger regions of space, animals, faces). Cells in higher areas send “feedback” connections to lower areas, putting them in a position to learn and embed real-world relationships shaped by experience. For instance, cells encoding an “elephant” might send feedback to cells processing features like “grey,” “big,” and “heavy.” The researchers therefore set about investigating how visual experience influences the organisation of these feedback projections, whose functional role remains largely unknown.

Visual Experience and Feedback Connections

“We wanted to understand how these feedback projections store information about the world,” says Rodrigo Dias, one of the study’s first authors. “To do this, we examined the effects of visual experience on feedback projections to a lower visual area called V1 in mice. We raised two groups of mice differently: one in a normal environment with regular light exposure, and the other in darkness. We then observed how the feedback connections, and cells they target in V1, responded to different regions of the visual field.”

The Influence of Visual Experience on Brain Connectivity

In mice raised in darkness, the feedback connections and V1 cells directly below them both represented the same areas of visual space. First author Radhika Rajan picks up the story, “It was amazing to see how well the spatial representations of higher and lower areas matched up in the dark-reared mice. This suggests that the brain has an inherent, genetic blueprint for organizing these spatially aligned connections, independent of visual input.” However, in normally-reared mice, these connections were less precisely matched, and more feedback inputs conveyed information from surrounding areas of the visual field.

Rajan continues, “We found that with visual experience, feedback provides more contextual and novel information, enhancing the ability of V1 cells to sample information from a broader area of the visual scene.” This effect depended on the origin within the higher visual area: feedback projections from deeper layers were more likely to convey surround information compared to those from superficial layers.

Experience-Driven Adaptation in Visual Feedback

Moreover, the team discovered that in normally-reared mice, deep-layer feedback inputs to V1 become organized according to the patterns they “prefer” to see, such as vertical or horizontal lines. “For instance,” Dias says, “inputs that prefer vertical lines avoid sending surround information to areas located along the vertical direction. In contrast, we found no such bias in connectivity in dark-reared mice.”

“This suggests that visual experience plays a crucial role in fine-tuning feedback connections and shaping the spatial information transmitted from higher to lower visual areas,” notes Petreanu. “We developed a computational model that shows how experience leads to a selection process, reducing connections between feedback and V1 cells whose representations overlap too much. This minimizes redundancy, allowing V1 cells to integrate a more diverse range of feedback.”

Implications for Mental Health Disorders

Perhaps counterintuitively, the brain might encode learned knowledge by connecting cells that represent unrelated concepts, and that are less likely to be activated together based on real-world patterns. This could be an energy-efficient way to store information, so that when encountering a novel stimulus, like a pink elephant, the brain’s preconfigured wiring maximizes activation, enhancing detection and updating predictions about the world.

Identifying this brain interface where prior knowledge combines with new sensory information could be valuable for developing interventions in cases where this integration process malfunctions. As Petreanu concludes, “Such imbalances are thought to occur in conditions like autism and schizophrenia. In autism, individuals may perceive everything as novel because prior information isn’t strong enough to influence perception. Conversely, in schizophrenia, prior information could be overly dominant, leading to perceptions that are internally generated rather than based on actual sensory input. Understanding how sensory information and prior knowledge are integrated can help address these imbalances.”

Reference: “Visual experience reduces the spatial redundancy between cortical feedback inputs and primary visual cortex neurons” 12 August 2024, Neuron.
DOI: 10.1016/j.neuron.2024.07.009