New theory about how the brain decodes sensory information

brain

New research out of the University of Pennsylvania is filling in gaps between two prevailing theories about how the brain generates our perception of the world.

One, called the Bayesian decoding theory, says that to best recognize what's in front of us, the brain combines the it receives, a scene we're looking at, for example, with our preconceived notions. Experience says that a school bus is typically yellow, so logically the one in front of us is more likely yellow than blue.

The second theory, called efficient coding, explains that, for sensory input resources like neurons in the brain and retinas in the eyes to efficiently do their jobs, they process and store more precise information about the most frequently viewed objects. If you see a yellow bus every day, your brain encodes an accurate picture of the vehicle so on next viewing you'll know its color. Conversely, the brain won't expend many resources to remember a blue bus.

Alan Stocker, an assistant professor in the psychology and the electrical and systems engineering departments, and Xue-Xin Wei, a psychology graduate student, combined these two concepts to create a new theory about how we perceive our world: How often we observe an object or scene shapes both what we expect of something similar in the future and how accurately we'll see it.

"There are two forces that determine what we perceive," Stocker said.

According to his research, those sometimes work against each other; we end up observing the opposite of what our experience tells us should happen. Keeping with the bus example, a bus that is actually blue would look even bluer than it is.

Stocker and Wei also discovered that the accuracy of the object or scene itself changes the viewer's perception. "If [the stimulus] is cloudy, for example, that would have a different effect than if the 'noise' was at the neural level," Stocker said. "If I show you something briefly, the neurons in the brain only have time to respond with a few spikes. That gives a low accuracy representation in the ."

Given the prevalence of the Bayesian model, Stocker said there's sometimes hesitance to discuss the notion of repulsive bias, something he said he hopes their research makes more acceptable.

"It's a big step forward bringing efficient coding into the picture," said Stocker. "It opens up the door to a whole new class of data considered 'anti-Bayesian' before."

More information: "A Bayesian observer model constrained by efficient coding can explain 'anti-Bayesian' percepts." Nature Neuroscience 18, 1509–1517 (2015) DOI: 10.1038/nn.4105

Journal information: Nature Neuroscience
Citation: New theory about how the brain decodes sensory information (2015, October 7) retrieved 29 March 2024 from https://medicalxpress.com/news/2015-10-theory-brain-decodes-sensory.html
This document is subject to copyright. Apart from any fair dealing for the purpose of private study or research, no part may be reproduced without the written permission. The content is provided for information purposes only.

Explore further

Shapes, lines and movements are in the eye of the beholder

173 shares

Feedback to editors