Featured Research

from universities, journals, and other organizations

Computation Provides A Virtual Recording Of Auditory Signaling

Date:
February 3, 2005
Source:
Public Library Of Science
Summary:
Tim Gollisch and Andreas Herz turned to computational methods and showed that it's possible to reveal the individual steps of complex signal processing by analyzing the output activity alone. Using grasshopper auditory receptors as models, the authors identified the individual signal-processing steps from eardrum vibrations to electrical potential within a sub-millisecond time frame and propose a model for auditory signaling.

A schematic representation of auditory signaling. (Graphic courtesy of Public Library Of Science)

A small rodent rustles through a field in the still night, making just enough noise to betray its location to a circling barn owl. A female frog sits on the bank of a pond amid a cacophony of courting bullfrogs, immune to the mating calls of all but her own species. Thanks to a sophisticated sensory processing system, animals can cut through a vast array of ambient auditory stimuli to extract meaningful information that allows them to tell where a sound came from, for example, or whether they should respond to a particular mating call.

An acoustic stimulus arrives at the ear as sound energy in the form of air pressure fluctuations. The sound signal triggers oscillations in mechanical resonators such as the eardrum and hair sensilla. These oscillations convert sound energy into mechanical energy, opening ion channels in auditory receptor cells and producing electrical currents that change the neuron's membrane potential. This, in turn, produces the action potential that carries the sound signal to the brain. This multistep signal transduction process takes less than a millisecond, but exactly how it occurs at this time scale remains obscure. Direct measurements of the individual steps can't be made without destroying the mechanical structure; consequently, most measurements are taken downstream of the mechanical oscillations at locations like the auditory nerve. Likewise, the temporal resolution of most stimulus–response trials is far too imprecise to analyze processing at the sub-millisecond level.

Given these experimental limitations, Tim Gollisch and Andreas Herz turned to computational methods and showed that it's possible to reveal the individual steps of complex signal processing by analyzing the output activity alone. Using grasshopper auditory receptors as models, the authors identified the individual signal-processing steps from eardrum vibrations to electrical potential within a sub-millisecond time frame and propose a model for auditory signaling.

The crucial step in their study is the search for those sets of inputs (stimuli) that would yield a given fixed output (response). To get the parameters to describe the final output, the authors generated a sound stimulus (two short clicks) and recorded axon responses of receptor neurons in a grasshopper auditory nerve. From these recordings, they defined the fixed output as the probability of a receptor neuron firing a single action potential. They then asked how the various parameters, which were associated with different time scales, could produce the same predefined firing probability.

By varying the stimulus parameters and comparing the obtained values within their mathematical framework—and making certain assumptions, for example, that the steps signal through a “feedforward” process—they could then tease out the individual processing steps that contribute to the desired output within the required time frame. With this approach, Gollisch and Herz disentangled individual steps of two consecutive integration processes—which they conclude are the mechanical resonance of the eardrum and the electrical integration of the receptor neuron—down to the microsecond level. Surprisingly, this fine temporal resolution is achieved even though the neuron's action potentials jitter by about one millisecond.

Thus, using just the final output, this approach can extract the temporal details of the individual processes that contribute to the chain of auditory transduction events. While this method is best-suited for deconstructing unidirectional pathways, the authors suggest it could also help separate “feedforward” from feedback signaling components, especially when feedback is triggered by the final steps. But since many sensory systems share the same basic signal-processing steps, this method is likely applicable to a broad range of problems.


Story Source:

The above story is based on materials provided by Public Library Of Science. Note: Materials may be edited for content and length.


Cite This Page:

Public Library Of Science. "Computation Provides A Virtual Recording Of Auditory Signaling." ScienceDaily. ScienceDaily, 3 February 2005. <www.sciencedaily.com/releases/2005/02/050201185558.htm>.
Public Library Of Science. (2005, February 3). Computation Provides A Virtual Recording Of Auditory Signaling. ScienceDaily. Retrieved August 20, 2014 from www.sciencedaily.com/releases/2005/02/050201185558.htm
Public Library Of Science. "Computation Provides A Virtual Recording Of Auditory Signaling." ScienceDaily. www.sciencedaily.com/releases/2005/02/050201185558.htm (accessed August 20, 2014).

Share This




More Mind & Brain News

Wednesday, August 20, 2014

Featured Research

from universities, journals, and other organizations


Featured Videos

from AP, Reuters, AFP, and other news services

Charter Schools Alter Post-Katrina Landscape

Charter Schools Alter Post-Katrina Landscape

AP (Aug. 20, 2014) Nine years after Hurricane Katrina, charter schools are the new reality of public education in New Orleans. The state of Louisiana took over most of the city's public schools after the killer storm in 2005. (Aug. 20) Video provided by AP
Powered by NewsLook.com
Researcher Testing on-Field Concussion Scanners

Researcher Testing on-Field Concussion Scanners

AP (Aug. 19, 2014) Four Texas high school football programs are trying out an experimental system designed to diagnose concussions on the field. The technology is in response to growing concern over head trauma in America's most watched sport. (Aug. 19) Video provided by AP
Powered by NewsLook.com
Kids' Drawings At Age 4 Linked To Intelligence At Age 14

Kids' Drawings At Age 4 Linked To Intelligence At Age 14

Newsy (Aug. 19, 2014) A study by King's College London says there's a link between how well kids draw at age 4 and how intelligent they are later in life. Video provided by Newsy
Powered by NewsLook.com
Mental, Neurological Disabilities Up 21% Among Kids

Mental, Neurological Disabilities Up 21% Among Kids

Newsy (Aug. 18, 2014) New numbers show a decade's worth of changes in the number of kids with disabilities. They suggest mental disabilities are up; physical ones are down. Video provided by Newsy
Powered by NewsLook.com

Search ScienceDaily

Number of stories in archives: 140,361

Find with keyword(s):
Enter a keyword or phrase to search ScienceDaily for related topics and research stories.

Save/Print:
Share:

Breaking News:
from the past week

In Other News

... from NewsDaily.com

Science News

Health News

Environment News

Technology News



Save/Print:
Share:

Free Subscriptions


Get the latest science news with ScienceDaily's free email newsletters, updated daily and weekly. Or view hourly updated newsfeeds in your RSS reader:

Get Social & Mobile


Keep up to date with the latest news from ScienceDaily via social networks and mobile apps:

Have Feedback?


Tell us what you think of ScienceDaily -- we welcome both positive and negative comments. Have any problems using the site? Questions?
Mobile: iPhone Android Web
Follow: Facebook Twitter Google+
Subscribe: RSS Feeds Email Newsletters
Latest Headlines Health & Medicine Mind & Brain Space & Time Matter & Energy Computers & Math Plants & Animals Earth & Climate Fossils & Ruins