Science News
from research organizations

How do we combine faces and voices?

Date:
March 10, 2011
Source:
Elsevier
Summary:
Faces and voices are known to be some of the key features that enable us to identify individual people, and they are rich in information such as gender, age, and body size, that lead to a unique identity for a person. A large body of neuropsychological and neuroimaging research has already determined the various brain regions responsible for face recognition and voice recognition separately, but exactly how our brain goes about combining the two different types of information (visual and auditory) is still unknown. Now a new study has revealed the brain networks involved in this "cross-modal" person recognition.
Share:
       
FULL STORY

Human social interactions are shaped by our ability to recognise people. Faces and voices are known to be some of the key features that enable us to identify individual people, and they are rich in information such as gender, age, and body size, that lead to a unique identity for a person. A large body of neuropsychological and neuroimaging research has already determined the various brain regions responsible for face recognition and voice recognition separately, but exactly how our brain goes about combining the two different types of information (visual and auditory) is still unknown.

Now a new study, published in the March 2011 issue of Elsevier's Cortex, has revealed the brain networks involved in this "cross-modal" person recognition.

A team of researchers in Belgium used functional magnetic resonance imaging (fMRI) to measure brain activity in 14 participants while they performed a task in which they recognised previously learned faces, voices, and voice-face associations. Dr Frédéric Joassin, Dr Salvatore Campanella, and colleagues compared the brain areas activated when recognising people using information from only their faces (visual areas), or only their voices (auditory areas), to those activated when using the combined information. They found that voice-face recognition activated specific "cross-modal" regions of the brain, located in areas known as the left angular gyrus and the right hippocampus. Further analysis also confirmed that the right hippocampus was connected to the separate visual and auditory areas of the brain.

Recognising a person from the combined information of their face and voice therefore relies not only on the same brain networks involved in using only visual or only auditory information, but also on brain regions associated with attention (left angular gyrus) and memory (hippocampus). According to the authors, the findings support a dynamic vision of cross-modal interactions in which the areas involved in processing both face and voice information are not simply the final stage of a hierarchical model, but rather, they may work in parallel and influence each other.


Story Source:

The above post is reprinted from materials provided by Elsevier. Note: Materials may be edited for content and length.


Journal Reference:

  1. Frédéric Joassin, Mauro Pesenti, Pierre Maurage, Emilie Verreckt, Raymond Bruyer, Salvatore Campanella. Cross-modal interactions between human faces and voices involved in person recognition. Cortex, 2011; 47 (3): 367 DOI: 10.1016/j.cortex.2010.03.003

Cite This Page:

Elsevier. "How do we combine faces and voices?." ScienceDaily. ScienceDaily, 10 March 2011. <www.sciencedaily.com/releases/2011/03/110309101809.htm>.
Elsevier. (2011, March 10). How do we combine faces and voices?. ScienceDaily. Retrieved September 5, 2015 from www.sciencedaily.com/releases/2011/03/110309101809.htm
Elsevier. "How do we combine faces and voices?." ScienceDaily. www.sciencedaily.com/releases/2011/03/110309101809.htm (accessed September 5, 2015).

Share This Page: