Psychoacoustic cues to emotion in speech prosody and music

Cogn Emot. 2013;27(4):658-84. doi: 10.1080/02699931.2012.732559. Epub 2012 Oct 12.

Abstract

There is strong evidence of shared acoustic profiles common to the expression of emotions in music and speech, yet relatively limited understanding of the specific psychoacoustic features involved. This study combined a controlled experiment and computational modelling to investigate the perceptual codes associated with the expression of emotion in the acoustic domain. The empirical stage of the study provided continuous human ratings of emotions perceived in excerpts of film music and natural speech samples. The computational stage created a computer model that retrieves the relevant information from the acoustic stimuli and makes predictions about the emotional expressiveness of speech and music close to the responses of human subjects. We show that a significant part of the listeners' second-by-second reported emotions to music and speech prosody can be predicted from a set of seven psychoacoustic features: loudness, tempo/speech rate, melody/prosody contour, spectral centroid, spectral flux, sharpness, and roughness. The implications of these results are discussed in the context of cross-modal similarities in the communication of emotion in the acoustic domain.

Publication types

  • Research Support, Non-U.S. Gov't

MeSH terms

  • Acoustic Stimulation
  • Adolescent
  • Adult
  • Auditory Perception
  • Cues*
  • Emotions*
  • Female
  • Humans
  • Male
  • Middle Aged
  • Models, Psychological
  • Music / psychology*
  • Neural Networks, Computer
  • Psychoacoustics*
  • Speech*