A new study uses a cappella to show how the brain decodes music and speech in different hemispheres.
Researchers created 100 a cappella recordings, each of a soprano singing a sentence. They then distorted the recordings along two fundamental auditory dimensions: spectral and temporal dynamics, and had 49 participants distinguish the words or the melodies of each song.
They conducted the experiment with two groups of English and French speakers to enhance reproducibility and generalizability. They demonstrate the experiment here.
The researchers found that for both languages, when the temporal information was distorted, participants had trouble distinguishing the speech content, but not the melody. Conversely, when spectral information was distorted, they had trouble distinguishing the melody, but not the speech. This shows that speech and melody depend on different acoustical features.
Left and right
To test how the brain responds to these different sound features, the researchers then scanned participants with functional magnetic resonance imaging (fMRI) while they distinguished the sounds. The researchers found that speech processing occurred in the left auditory cortex, while melodic processing occurred in the right auditory cortex.
Next, they set out to test how degradation in each acoustic dimension would affect brain activity. They found that degradation of the spectral dimension only affected activity in the right auditory cortex, and only during melody perception, while degradation of the temporal dimension affected only the left auditory cortex, and only during speech perception. This shows that the differential response in each hemisphere depends on the type of acoustical information in the stimulus.
Previous studies in animals have found that neurons in the auditory cortex respond to particular combinations of spectral and temporal energy, and are highly tuned to sounds that are relevant to the animal in its natural environment, such as communication sounds.
Music and speech
For humans, both speech and music are important means of communication. This study shows that music and speech exploit different ends of the spectro-temporal continuum, and that hemispheric specialization may be the nervous system’s way of optimizing the processing of these two communication methods.
“It has been known for decades that the two hemispheres respond to speech and music differently, but the physiological basis for this difference remained a mystery,” says first author Philippe Albouy, a postodoctoral fellow at the Montreal Neurological Institute of McGill University.
“Here we show that this hemispheric specialization is linked to basic acoustical features that are relevant for speech and music, thus tying the finding to basic knowledge of neural organization.”
The results appear in Science.
A Banting fellowship to Albouy and the Canadian Institutes for Health Research and from the Canadian Institute for Advanced Research supported the work.
Source: McGill University