Noise-robust cortical tracking of attended speech in real-world acoustic scenes

Søren Fuglsang, Torsten Dau, Jens Hjortkjær

Research output: Contribution to journalJournal articleResearchpeer-review

1 Downloads (Pure)


Selectively attending to one speaker in a multi-speaker scenario is thought to synchronize low-frequency cortical activity to the attended speech signal. In recent studies, reconstruction of speech from single-trial electroencephalogram (EEG) data has been used to decode which talker a listener is attending to in a two-talker situation. It is currently unclear how this generalizes to more complex sound environments. Behaviorally, speech perception is robust to the acoustic distortions that listeners typically encounter in everyday life, but it is unknown whether this is mirrored by a noise-robust neural tracking of attended speech. Here we used advanced acoustic simulations to recreate real-world acoustic scenes in the laboratory. In virtual acoustic realities with varying amounts of reverberation and number of interfering talkers, listeners selectively attended to the speech stream of a particular talker. Across the different listening environments, we found that the attended talker could be accurately decoded from single-trial EEG data irrespective of the different distortions in the acoustic input. For highly reverberant environments, speech envelopes reconstructed from neural responses to the distorted stimuli resembled the original clean signal more than the distorted input. With reverberant speech, we observed a late cortical response to the attended speech stream that encoded temporal modulations in the speech signal without its reverberant distortion. Single-trial attention decoding accuracies based on 40-50s long blocks of data from 64 scalp electrodes were equally high (80-90% correct) in all considered listening environments and remained statistically significant using down to 10 scalp electrodes and short (
Original languageEnglish
Pages (from-to)435-444
Number of pages10
Publication statusPublished - 2017


  • Journal Article
  • Acoustic simulations
  • Auditory attention
  • Cortical entrainment
  • Decoding
  • Delta rhythms
  • EEG
  • Speech
  • Theta rhythms


Dive into the research topics of 'Noise-robust cortical tracking of attended speech in real-world acoustic scenes'. Together they form a unique fingerprint.

Cite this