New model for speech and sound recognition

People are adept at recognizing sensations such as sounds or smells, even when many stimuli appear simultaneously. But how the association works between the current event and memory is still poorly understood. Scientists at the Bernstein Center and the Ludwig-Maximilians Universität (LMU) München have developed a mathematical model that accurately mimics this process with little computational effort and may explain experimental findings that have so far remained unclear. (PLoS ONE, September 14, 2011)

The so-called ‘cocktail party-problem’ has already kept scientists busy for decades. How is it possible for the brain to filter familiar voices out of background noise? It is a long-standing hypothesis that we create a kind of sound library in the auditory cortex of the brain during the course of our lives. Professor Christian Leibold and Dr. Gonzalo Otazu at LMU Munich who are also members of the Bernstein Center Munich now show in a new model how the brain can compare stored and perceived sounds in a particularly efficient manner. Figuratively speaking, current models operate on the following principle: An archivist (possibly the brain region thalamus) compares the incoming sound with the individual entries in the library, and receives the degree of matching for each entry. Usually, however, several entries fit similarly well, so the archivist does not know which result is actually the right one.

The new model is different: as previously the archivist compares the sound with the library entries, this time getting back only a few really relevant records and information about how much the archived and heard elements differ. Therefore, only in the case of unknown or little matching inputs are large amounts of data sent back.

The researchers now want to incorporate their findings into other models that are more biologically detail-oriented, and finally test it in psychoacoustic experiments.