A password will be e-mailed to you.

Mapping Auditory Brain Cells for Maximum Hearing Precision

Summary: Researchers have identified the specific synaptic and post-synaptic characteristics that allow auditory neurons to compute with temporal precision.

Source: Lehigh University.

Team identifies optimal synaptic and post-synaptic characteristics that allow cells of the cochlear nucleus to compute with temporal precision; Method reveals features of input-output optimization that apply to all neurons that process time varying input.

When it comes to hearing, precision is important. Because vertebrates, such as birds and humans, have two ears–and sounds from either side travel different distances to arrive at each one – localizing sound involves discerning subtle differences in when sounds arrive. The brain has to keep time better than a Swiss watch in order to locate where sound is coming from.

In fact, the quality of this sound processing precision is a limiting factor in how well one detects the location of sound and perceives speech.

A team of researchers led by R. Michael Burger, neuroscientist and associate professor in Lehigh University’s Department of Biological Sciences, have identified the specific synaptic and post-synaptic characteristics that allow auditory neurons to compute with temporal precision–ultimately revealing the optimal arrangement of both input and electrical properties needed for neurons to process their “preferred” frequency with maximum precision.

In order for birds and mammals to hear, hair cells in the cochlea–the auditory portion of the inner ear–vibrate in response to sounds and thereby convert sound into electrical activity. Each hair cell is tuned to a unique frequency tone, which humans ultimately experience as pitch.

Every hair cell in the cochlea is partnered with several neurons that convey information from the ear to the brain in an orderly way. The tone responses in the cochlea are, essentially, “remapped” to the cochlear nucleus, the first brain center to process sounds.

This unique spatial arrangement of how sounds of different frequencies are processed in the brain is called tonotopy. It can be visualized as a kind of sound map: tones that are close to each other in terms of frequency are represented by neighboring neurons of the cochlear nucleus.

Timing precision is important to cochlear nucleus neurons because their firing pattern is specific for each sound frequency. That is, their output pattern is akin to a digital code that is unique for each tone.

“In the absence of sound, neurons fire randomly and at a high rate,” says Burger. “In the presence of sound, neurons fire in a highly stereotyped manner known as phase-locking–which is the tendency for a neuron to fire at a particular phase of a periodic stimulus or sound wave.”

Previous research by Burger and Stefan Oline–a former Ph.D. candidate at Lehigh, now a postdoctoral fellow at New York University Medical School–demonstrated for the first time that synaptic inputs–the messages being sent between cells–are distinct across frequencies and that these different impulse patterns are “mapped” onto the cells of the cochlear nucleus. They further established the computational processes by which neurons “tuned” to process low frequency sound actually improve the phase-locking precision of the impulses they receive. However, the mechanisms that allow neurons to respond properly to these frequency-specific incoming messages remained poorly understood.

In new research, published in an article in The Journal of Neuroscience, Burger and Oline–along with Dr. Go Ashida of the University of Oldenburg in Germany–have investigated auditory brain cell membrane selectivity and observed that the neurons “tuned” to receive high-frequency sound preferentially select faster input than their low-frequency-processing counterparts–and that this preference is tolerant of changes to the inputs being received.

“A low frequency cell will tolerate a slow input and still be able to fire–but a high frequency cell requires a very rapid input and rejects slow input,” says Burger. “The neurons essentially demand that the high-frequency input be more precise.”

“What I find really striking is that the tuning of these neurons helps them uniquely deal with the constraints of the ear,” says Oline. “Neurons responding to low frequency input can average their inputs from hair cells to improve their resolution. But hair cells aren’t very good at responding to high frequency tones as they introduce a lot of timing errors. Because of this, and because they occur at such a high rate, averaging these inputs is impossible and would smear information across multiple sound waves. So, instead, the high-frequency-processing cells use an entirely different strategy: they are as picky as possible to avoid averaging at all costs.”

Burger and his colleagues built a computer simulation of low frequency and high frequency neurons, based on observations of physiological activity. They then used these computational models to test which combinations of properties are crucial to phase-locking. The model predicted that the optimal arrangement of synaptic and cell membrane properties for phase-locking is specific to stimulus frequency. These computational predictions were then tested physiologically in the neurons.

Diagram of the auditory system.

The differential processing between high and low characteristic frequency neurons suggests that synaptic integration may differ along the tonotopy. Neurosciencenews image is credited to R. Michael Burger and Stefan Oline.

The team’s model is not only useful for determining how the brain responds to sounds, but also reveals general features of input-output optimization that apply to any brain cell that processes time varying input.

Paving the way to more precise hearing

Understanding the mechanisms that allow cells of the cochlear nucleus to compute with temporal precision has implications for understanding the evolution of the auditory system.

“It’s really the high frequency-processing cells that have uniquely evolved in mammals,” explains Burger.

Understanding these processes may also be important for advancing the technology used to make cochlear implants. A cochlear implant is an electronic medical device that helps provide a sense of sound to someone who is deaf or has severe hearing loss. It replaces the function of the damaged inner ear by sending electrical impulses directly to the auditory nerve. These impulses, in turn, are interpreted by the brain as sound.

Though an established and effective treatment for many, cochlear implants cannot currently simulate the precision of sound experienced by those with a naturally-developed auditory system. The sound processing lacks the clarity of natural hearing, especially across frequencies.

“Ideally, what you want–whether in your natural auditory system or through a cochlear implant–is the most precise representation in the brain of the various frequencies,” says Burger.

Burger and his colleagues have assembled what is known about the optimal electrical properties and synaptic inputs into a single cohesive model, laying the groundwork needed to investigate some of the big questions in the field of auditory neuroscience. Resolving these questions may someday lead scientists and medical professionals to a better understanding of how to preserve the natural organization of the auditory structures in the brain for those who are born with profound hearing loss.

About this neuroscience research article

Funding: NIH//National Institute on Deafness and Other Communication Disorders funded study.

Source: Lori Friedman – Lehigh University
Image Source: This NeuroscienceNews.com image is credited to R. Michael Burger and Stefan Oline.
Original Research: Abstract for “Tonotopic Optimization for Temporal Processing in the Cochlear Nucleus” by Stefan N. Oline, Go Ashida, and R. Michael Burger in Journal of Neuroscience. Published online August 10 2016 doi:10.1523/JNEUROSCI.4449-15.2016

Cite This NeuroscienceNews.com Article
Lehigh University. “Mapping Auditory Brain Cells for Maximum Hearing Precision.” NeuroscienceNews. NeuroscienceNews, 12 September 2016.
<http://neurosciencenews.com/auditory-neurons-hearing-mapping-5016/>.
Lehigh University. (2016, September 12). Mapping Auditory Brain Cells for Maximum Hearing Precision. NeuroscienceNews. Retrieved September 12, 2016 from http://neurosciencenews.com/auditory-neurons-hearing-mapping-5016/
Lehigh University. “Mapping Auditory Brain Cells for Maximum Hearing Precision.” http://neurosciencenews.com/auditory-neurons-hearing-mapping-5016/ (accessed September 12, 2016).

Abstract

Tonotopic Optimization for Temporal Processing in the Cochlear Nucleus

In the auditory system, sounds are processed in parallel frequency-tuned circuits, beginning in the cochlea. Auditory nerve fibers reflect this tonotopy and encode temporal properties of acoustic stimuli by “locking” discharges to a particular stimulus phase. However, physiological constraints on phase-locking depend on stimulus frequency. Interestingly, low characteristic frequency (LCF) neurons in the cochlear nucleus improve phase-locking precision relative to their auditory nerve inputs. This is proposed to arise through synaptic integration, but the postsynaptic membrane’s selectivity for varying levels of synaptic convergence is poorly understood. The chick cochlear nucleus, nucleus magnocellularis (NM), exhibits tonotopic distribution of both input and membrane properties. LCF neurons receive many small inputs and have low input thresholds, whereas high characteristic frequency (HCF) neurons receive few, large synapses and require larger currents to spike. NM therefore presents an opportunity to study how small membrane variations interact with a systematic topographic gradient of synaptic inputs. We investigated membrane input selectivity and observed that HCF neurons preferentially select faster input than their LCF counterparts, and that this preference is tolerant of changes to membrane voltage. We then used computational models to probe which properties are crucial to phase-locking. The model predicted that the optimal arrangement of synaptic and membrane properties for phase-locking is specific to stimulus frequency and that the tonotopic distribution of input number and membrane excitability in NM closely tracks a stimulus-defined optimum. These findings were then confirmed physiologically with dynamic-clamp simulations of inputs to NM neurons.

SIGNIFICANCE STATEMENT One way that neurons represent temporal information is by phase-locking, which is discharging in response to a particular phase of the stimulus waveform. In the auditory system, central neurons are optimized to retain or improve phase-locking precision compared with input from the auditory nerve. However, the difficulty of this computation varies systematically with stimulus frequency. We examined properties that contribute to temporal processing both physiologically and in a computational model. Neurons processing low-frequency input benefit from integration of many weak inputs, whereas those processing higher frequencies progressively lose precision by integration of multiple inputs. Here, we reveal general features of input-output optimization that apply to all neurons that process time varying input.

“Tonotopic Optimization for Temporal Processing in the Cochlear Nucleus” by Stefan N. Oline, Go Ashida, and R. Michael Burger in Journal of Neuroscience. Published online August 10 2016 doi:10.1523/JNEUROSCI.4449-15.2016

Feel free to share this Neuroscience News.
Join our Newsletter
Sign up to receive the latest neuroscience headlines and summaries sent to your email daily from NeuroscienceNews.com
We hate spam. Your email address will not be sold or shared with anyone else.
No more articles

Pin It on Pinterest

Share This

Share This

Share this neuroscience news with your friends!