Could not connect to DB: 1040: Too many connectionsCould not execute 'UPDATE pressrelease SET r_hits = r_hits+ 1, r_total_hits = r_total_hits+ 1, r_pub_hits = r_pub_hits+ 1, r_total_pub_hits = r_total_pub_hits+ 1 WHERE r_id = 188219' on database eurekalert:
2002: Can't connect to local MySQL server through socket '/tmp/mysql.sock' (2) [ Back to EurekAlert! ] Public release date: 22-Jun-2011
[ | E-mail Share Share ]

Contact: Karen Mallet
km463@georgetown.edu
Georgetown University Medical Center

Researchers identify components of speech recognition pathway in humans

Finding suggests speech perception evolved from animals

Washington, D.C. Neuroscientists at Georgetown University Medical Center (GUMC) have defined, for the first time, three different processing stages that a human brain needs to identify sounds such as speech and discovered that they are the same as ones identified in non-human primates.

In the June 22 issue of the Journal of Neuroscience, the researchers say their discovery made possible with the help of 13 human volunteers who spent time in a functional MRI machine could potentially offer important insights into what can go wrong when someone has difficulty speaking, which involves hearing voice-generated sounds, or understanding the speech of others.

But more than that, the findings help shed light on the complex, and extraordinarily elegant, workings of the "auditory" human brain, says Josef Rauschecker, PhD, a professor in the departments of physiology/ biophysics and neuroscience and a member of the Georgetown Institute for Cognitive and Computational Sciences at GUMC.

"This is the first time we have been able to identify three discrete brain areas that help people recognize and understand the sounds they are hearing," says Rauschecker. "These sounds, such as speech, are vitally important to humans, and it is critical that we understand how they are processed in the human brain."

Rauschecker and his colleagues at Georgetown have been instrumental in building a unified theory about how the human brain processes speech and language. They have shown that both human and non-human primates process speech along two parallel pathways, each of which run from lower to higher functioning neural regions.

These pathways are dubbed the "what" and "where" streams and are roughly analogous to how the brain processes sight, but in different regions. The "where" stream localizes sound and the "what" pathway identifies the sound.

Both pathways begin with the processing of signals in the auditory cortex, located inside a deep fissure on the side of the brain underneath the temples - the so-called "temporal lobe." Information processed by the "what" pathway then flows forward along the outside of the temporal lobe, and the job of that pathway is to recognize complex auditory signals, which include communication sounds and their meaning (semantics). The "where" pathway is mostly in the parietal lobe, above the temporal lobe, and it processes spatial aspects of a sound its location and its motion in space but is also involved in providing feedback during the act of speaking.

Auditory perception - the processing and interpretation of sound information - is tied to anatomical structures; signals move from lower to higher brain regions, Rauschecker says. "Sound as a whole enters the ear canal and is first broken down into single tone frequencies, then higher-up neurons respond only to more complex sounds, including those used in the recognition of speech, as the neural representation of the sound moves through the various brain regions," he says.

In this study, Rauschecker and his colleagues computational neuroscientist Maximilian Riesenhuber, Ph.D., and Mark Chevillet, a student in the Interdisciplinary Program in Neuroscience identified the three distinct areas in the "what" pathway in humans that had been seen in non-human primates. Only two had been recognized before in previous human studies.

The first, and most primary, is the "core" which analyzes tones at the basic level of simple frequencies. The second area, the "belt", wraps around the core, and integrates several tones, "like buzz sounds," that lie close to each other, Rauschecker says. The third area, the "parabelt," responds to speech sounds such as vowels, which are essentially complex bursts of multiple frequencies.

Rauschecker is fascinated by the fact that although speech and language are considered to be uniquely human abilities, the emerging picture of brain processing of language suggests "in evolution, language must have emerged from neural mechanisms at least partially available in animals," he says. "There appears to be a conservation of certain processing pathways through evolution in humans and nonhuman primates."

###

The study was funded by a National Science Foundation grant awarded to Rauschecker and Riesenhuber.

About Georgetown University Medical Center

Georgetown University Medical Center is an internationally recognized academic medical center with a three-part mission of research, teaching and patient care (through MedStar Health). GUMC's mission is carried out with a strong emphasis on public service and a dedication to the Catholic, Jesuit principle of cura personalis -- or "care of the whole person." The Medical Center includes the School of Medicine and the School of Nursing & Health Studies, both nationally ranked; Georgetown Lombardi Comprehensive Cancer Center, designated as a comprehensive cancer center by the National Cancer Institute; and the Biomedical Graduate Research Organization (BGRO), which accounts for the majority of externally funded research at GUMC including a Clinical Translation and Science Award from the National Institutes of Health. In fiscal year 2009-2010, GUMC accounted for nearly 80 percent of Georgetown University's extramural research funding.



[ Back to EurekAlert! ] [ | E-mail Share Share ]

 


AAAS and EurekAlert! are not responsible for the accuracy of news releases posted to EurekAlert! by contributing institutions or for the use of any information through the EurekAlert! system.