November 2, 2024

Revolutionizing Neuroscience: Stanford AI Mirrors Brain Organization

Revolutionizing Neuroscience: Stanford AI Mirrors Brain OrganizationHuman Brain Artificial Intelligence Drawing - Revolutionizing Neuroscience: Stanford AI Mirrors Brain Organization

Stanford’s Wu Tsai Neurosciences Institute has developed an AI model called a topographic deep artificial neural network (TDANN) that mimics the brain’s organization of visual information. This model, which uses naturalistic inputs and spatial constraints, has successfully replicated the brain’s functional maps and could significantly impact both neuroscience research and artificial intelligence. The findings, published after seven years of research, highlight the potential for more energy-efficient AI and enhanced virtual neuroscience experiments that could revolutionize medical treatments and AI’s visual processing capabilities.

Stanford researchers have developed an AI that replicates brain-like responses to visual stimuli, potentially transforming neuroscience and AI development with implications for energy efficiency and medical advancements.

A team at Stanford’s Wu Tsai Neurosciences Institute has achieved a significant breakthrough in using AI to mimic the way the brain processes sensory information to understand the world, paving the way for advancements in virtual neuroscience.

Watch the seconds tick by on a clock and, in visual regions of your brain, neighboring groups of angle-selective neurons will fire in sequence as the second hand sweeps around the clock face. These cells form beautiful “pinwheel” maps, with each segment representing a visual perception of a different angle. Other visual areas of the brain contain maps of more complex and abstract visual features, such as the distinction between images of familiar faces vs. places, which activate distinct neural “neighborhoods.”

Such functional maps can be found across the brain, both delighting and confounding neuroscientists, who have long wondered why the brain should have evolved a map-like layout that only modern science can observe.

To address this question, the Stanford team developed a new kind of AI algorithm — a topographic deep artificial neural network (TDANN) — that uses just two rules: naturalistic sensory inputs and spatial constraints on connections; and found that it successfully predicts both the sensory responses and spatial organization of multiple parts of the human brain’s visual system.

Seven Years of Research Culminate in Publication

After seven years of extensive research, the findings were published in a new paper — “A unifying framework for functional organization in the early and higher ventral visual cortex” — in the journal Neuron.

The research team was led by Wu Tsai Neurosciences Institute Faculty Scholar Dan Yamins, an assistant professor of psychology and computer science; and

Institute affiliate Kalanit Grill-Spector, a professor in psychology.

Unlike conventional neural networks, the TDANN incorporates spatial constraints, arranging its virtual neurons on a two-dimensional “cortical sheet” and requiring nearby neurons to share similar responses to sensory input. As the model learned to process images, this topographical structure caused it to form spatial maps, replicating how neurons in the brain organize themselves in response to visual stimuli. Specifically, the model replicated complex patterns such as the pinwheel structures in the primary visual cortex (V1) and the clusters of neurons in the higher ventral temporal cortex (VTC) that respond to categories like faces or places.

Eshed Margalit, the study’s lead author, who completed his PhD working with Yamins and Grill-Spector, said the team used self-supervised learning approaches to help the accuracy of training models that simulate the brain.

“It’s probably more like how babies are learning about the visual world,” Margalit said. “I don’t think we initially expected it to have such a big impact on the accuracy of the trained models, but you really need to get the training task of the network right for it to be a good model of the brain.”

Implications for Neuroscience and AI

The fully trainable model will help neuroscientists better understand the rules of how the brain organizes itself, whether for vision, like in this study, or other sensory systems such as hearing.

“When the brain is trying to learn something about the world — like seeing two snapshots of a person — it puts neurons that respond similarly in proximity in the brain and maps form,” said Grill-Spector, who is the Susan S. and William H. Hindle Professor in the School of Humanities and Sciences. “We believe that principle should be translatable to other systems, as well.”

This innovative approach has significant implications for both neuroscience and artificial intelligence. For neuroscientists, the TDANN provides a new lens to study how the visual cortex develops and operates, potentially transforming treatments for neurological disorders. For AI, insights derived from the brain’s organization can lead to more sophisticated visual processing systems, akin to teaching computers to ‘see’ as humans do.

The findings could also help explain how the human brain operates with such stellar energy efficiency. For example, the human brain can compute a billion-billion math operations with only 20 watts of power, compared with a supercomputer that requires a million times more energy to do the same math. The new findings emphasize that neuronal maps — and the spatial or topographic constraints that drive them — likely serve to keep the wiring connecting the brain’s 100 billion neurons as simple as possible. These insights could be key to designing more efficient artificial systems inspired by the elegance of the brain.

“AI is constrained by power,” Yamins said. “In the long run, if people knew how to run artificial systems at a much lower power consumption, that could fuel AI’s development.”

More energy-efficient AI could help grow virtual neuroscience, where experiments could be done more quickly and at a larger scale. In their study, the researchers demonstrated as a proof of principle that their topographical deep artificial neural network reproduced brain-like responses to a wide range of naturalistic visual stimuli, suggesting that such systems could, in the future, be used as fast, inexpensive playgrounds for prototyping neuroscience experiments and rapidly identifying hypotheses for future testing.

Virtual neuroscience experiments could also advance human medical care. For example, better training an artificial visual system in the same way a baby visually learns about the world might help an AI see the world like a human, where the center of gaze is sharper than the rest of a field of view. Another application could help develop prosthetics for vision or simulate exactly how diseases and injuries affect parts of the brain.

“If you can do things like make predictions that are going to help develop prosthetic devices for people that have lost vision, I think that’s really going to be an amazing thing,” Grill-Spector said.

Reference: “A unifying framework for functional organization in early and higher ventral visual cortex” by Eshed Margalit, Hyodong Lee, Dawn Finzi, James J. DiCarlo, Kalanit Grill-Spector and Daniel L.K. Yamins, 10 May 2024, Neuron.
DOI: 10.1016/j.neuron.2024.04.018