The perception of consonants has been investigated in various studies and shown to critically depend on fine details in the stimuli. The present study proposes a microscopic speech perception model that combines an auditory processing front end with a correlation-based template-matching back end to predict consonant recognition and confusions. The model represents an extension of the auditory signal processing model by Dau et al. [(1997). J. Acoust. Soc. Am. 102, 2892–2905] towards predicting microscopic speech perception data. Model predictions were computed for the extensive consonant perception data set provided by Zaar and Dau [(2015). J. Acoust. Soc. Am. 138, 1253–1267], obtained with consonant-vowels (CVs) in white noise. The predictions were in good agreement with the perceptual data both in terms of consonant recognition and confusions. The model was further evaluated with respect to perceptual artifacts induced by (i) different hearing-aid signal processing strategies and (ii) simulated cochlear-implant processing, based on data from DiNino et al. [(2016). J. Acoust. Soc. Am., 140, 4404–4418]. The model successfully predicted the strong consonant confusions measured in these conditions. Overall, the results suggest that the proposed model may provide a valuable framework for assessing acoustic transmission channels and hearing-instrument signal processing.
|Journal||Journal of the Acoustical Society of America|
|Publication status||Published - 2018|
|Event||The 3rd Joint Meeting of the Acoustical Society of America and the European Acoustics Association - John B. Hynes Veterans Memorial Convention Center, Boston, United States|
Duration: 25 Jun 2017 → 29 Jun 2017
|Conference||The 3rd Joint Meeting of the Acoustical Society of America and the European Acoustics Association|
|Location||John B. Hynes Veterans Memorial Convention Center|
|Period||25/06/2017 → 29/06/2017|