Cross-modal speech perception in adults and infants using nonspeech auditory stimuli

J Exp Psychol Hum Percept Perform. 1991 Aug;17(3):829-40. doi: 10.1037//0096-1523.17.3.829.

Abstract

Adults and infants were tested for the capacity to detect correspondences between nonspeech sounds and real vowels. The /i/ and /a/ vowels were presented in 3 different ways: auditory speech, silent visual faces articulating the vowels, or mentally imagined vowels. The nonspeech sounds were either pure tones or 3-tone complexes that isolated a single feature of the vowel without allowing the vowel to be identified. Adults perceived an orderly relation between the nonspeech sounds and vowels. They matched high-pitched nonspeech sounds to /i/ vowels and low-pitched nonspeech sounds to /a/ vowels. In contrast, infants could not match nonspeech sounds to the visually presented vowels. Infants' detection of correspondence between auditory and visual speech appears to require the whole speech signal; with development, an isolated feature of the vowel is sufficient for detection of the cross-modal correspondence.

Publication types

  • Research Support, U.S. Gov't, P.H.S.

MeSH terms

  • Adult
  • Attention
  • Female
  • Humans
  • Language Development*
  • Lipreading*
  • Male
  • Middle Aged
  • Phonetics*
  • Pitch Discrimination
  • Speech Acoustics
  • Speech Perception*