Science News
from research organizations

Musical brain-reading sheds light on neural processing of music

Date:
December 17, 2013
Source:
Suomen Akatemia (Academy of Finland)
Summary:
Finnish and Danish researchers have developed a new method that performs decoding, or brain-reading, during continuous listening to real music. Based on recorded brain responses, the method predicts how certain features related to tone color and rhythm of the music change over time, and recognizes which piece of music is being listened to. The method also allows pinpointing the areas in the brain that are most crucial for the processing of music.
Share:
       
Total shares:  
FULL STORY

Finnish and Danish researchers have developed a new method that performs decoding, or brain-reading, during continuous listening to real music. Based on recorded brain responses, the method predicts how certain features related to tone color and rhythm of the music change over time, and recognizes which piece of music is being listened to. The method also allows pinpointing the areas in the brain that are most crucial for the processing of music. The study was published in the journal NeuroImage.

Using functional magnetic resonance imaging (fMRI), the research team at the Finnish Centre of Excellence in Interdisciplinary Music Research in the Universities of Jyväskylä and Helsinki, and the Center for Functionally Integrative Neuroscience in Aarhus University, Denmark, recorded the brain responses of participants while they were listening to a 16-minute excerpt of the album Abbey Road by the Beatles. Following this, they used computational algorithms to extract a collection of musical features from the musical recording. Subsequently, they employed a collection of machine-learning methods to train a computer model that predicts how the features of the music change over time. Finally, they develop a classifier that predicts which part of the music the participant was listening to at each time.

The researchers found that most of the musical features included in the study could be reliably predicted from the brain data. They also found that the piece being listened to could be predicted significantly better than chance. Fairly large differences were however found between participants in terms of the prediction accuracy. An interesting finding was that areas outside of the auditory cortex, including motor, limbic, and frontal areas, had to be included in the models to obtain reliable predictions, providing thus evidence for the important role of these areas in the processing of musical features.

"We believe that decoding provides a method that complements other existing methods to obtain more reliable information about the complex processing of music in the brain," says Prof. Petri Toiviainen from the University of Jyväskylä. "Our results provide additional evidence for the important involvement of emotional and motor areas in music processing."


Story Source:

The above story is based on materials provided by Suomen Akatemia (Academy of Finland). Note: Materials may be edited for content and length.


Journal Reference:

  1. Petri Toiviainen, Vinoo Alluri, Elvira Brattico, Mikkel Wallentin, Peter Vuust. Capturing the musical brain with Lasso: Dynamic decoding of musical features from fMRI data. NeuroImage, 2013; DOI: 10.1016/j.neuroimage.2013.11.017

Cite This Page:

Suomen Akatemia (Academy of Finland). "Musical brain-reading sheds light on neural processing of music." ScienceDaily. ScienceDaily, 17 December 2013. <www.sciencedaily.com/releases/2013/12/131217085058.htm>.
Suomen Akatemia (Academy of Finland). (2013, December 17). Musical brain-reading sheds light on neural processing of music. ScienceDaily. Retrieved May 22, 2015 from www.sciencedaily.com/releases/2013/12/131217085058.htm
Suomen Akatemia (Academy of Finland). "Musical brain-reading sheds light on neural processing of music." ScienceDaily. www.sciencedaily.com/releases/2013/12/131217085058.htm (accessed May 22, 2015).

Share This Page:


Mind & Brain News
May 22, 2015

Latest Headlines
updated 12:56 pm ET