More Than Meets The Ear In Successful Cocktail Party Conversations
- Date:
- March 8, 2008
- Source:
- Public Library of Science
- Summary:
- Just picture the scene: you're at a cocktail party, talking to someone you would like to get to know better but the background noise is making it hard to concentrate. Luckily, humans are very gifted at listening to someone speaking while many other people are talking loudly at the same time. This so-called cocktail-party-phenomenon is based on the ability of the human auditory system to decompose the acoustic world into discrete objects of perception. Now researchers have a better understanding of how this occurs. Different speakers have different temporal fine structure in their voiced speech and such signals are represented in different areas of the auditory cortex according to this different time structure. This provides a new solution for the cocktail party problem whereby people are able to hear and understand their conversation partner at a party in spite of considerable background noise.
- Share:
Just picture the scene: you're at a cocktail party, talking to someone you would like to get to know better but the background noise is making it hard to concentrate. Luckily, humans are very gifted at listening to someone speaking while many other people are talking loudly at the same time. This so-called cocktail-party-phenomenon is based on the ability of the human auditory system to decompose the acoustic world into discrete objects of perception.
It was originally believed that the major acoustic cue used by the auditory system to solve this task was directional information of the sound source, but even though localisation of different sound sources with two ears improves the performance, it can be achieved monaurally, for example in telephone conversations, where no directional information is available.
Scientists led by Holger Schulze at the Leibniz-Institute for Neurobiology in Magdeburg, and the Universities of Ulm, Newcastle and Erlangen have now found a neuronal mechanism in the auditory system that is able to solve the task based on the analysis of the temporal fine structure of the acoustic scene. Different speakers have different temporal fine structure in their voiced speech and that such signals are represented in different areas of the auditory cortex according to this different time structure.
By means of a so-called winner-take-all algorithm, one of these representations gains control over all other representations. This means that only the voice of the speaker to whom you wish to listen is still represented in the auditory cortex and can thus be followed over time. This predominance of the representation of one speaker's voice over the representations of all other speakers is achieved by long-range inhibitory interactions that are first described by Schulze and colleagues using functional neurophysiological, pharmacological and anatomical methods.
The findings provide a deeper understanding of how the parcellation of sensory input into perceptually distinct objects is realised in the brain, and may help to improve the auditory experience of hearing aid wearers at cocktail parties.
Journal reference: Kurt S, Deutscher A, Crook JM, Ohl FW, Budinger E, et al (2008) Auditory Cortical Contrast Enhancing by Global Winner-Take-All Inhibitory Interactions. PLoS One 3(3): e1735. doi:10.1371/journal.pone.0001735 http://www.plosone.org/doi/pone.0001735
Story Source:
Materials provided by Public Library of Science. Note: Content may be edited for style and length.
Cite This Page: