File Download
  Links for fulltext
     (May Require Subscription)
Supplementary

Article: Robust EEG-based decoding of auditory attention with high-RMS-level speech segments in noisy conditions

TitleRobust EEG-based decoding of auditory attention with high-RMS-level speech segments in noisy conditions
Authors
KeywordsEEG
Temporal response function (TRF)
Auditory attention decoding
Speech RMS-level segments
Signal-to-noise ratio
Issue Date2020
PublisherFrontiers Research Foundation. The Journal's web site is located at http://www.frontiersin.org/humanneuroscience/
Citation
Frontiers in Human Neuroscience, 2020, v. 14, article no. 557534 How to Cite?
AbstractThe attended speech stream can be detected robustly, even in adverse auditory scenarios with auditory attentional modulation, and can be decoded using electroencephalographic (EEG) data. Speech segmentation based on the relative root-mean-square (RMS) intensity can be used to estimate segmental contributions to perception in noisy conditions. High-RMS-level segments contain crucial information for speech perception. Hence, this study aimed to investigate the effect of high-RMS-level speech segments on auditory attention decoding performance under various signal-to-noise ratio (SNR) conditions. Scalp EEG signals were recorded when subjects listened to the attended speech stream in the mixed speech narrated concurrently by two Mandarin speakers. The temporal response function was used to identify the attended speech from EEG responses of tracking to the temporal envelopes of intact speech and high-RMS-level speech segments alone, respectively. Auditory decoding performance was then analyzed under various SNR conditions by comparing EEG correlations to the attended and ignored speech streams. The accuracy of auditory attention decoding based on the temporal envelope with high-RMS-level speech segments was not inferior to that based on the temporal envelope of intact speech. Cortical activity correlated more strongly with attended than with ignored speech under different SNR conditions. These results suggest that EEG recordings corresponding to high-RMS-level speech segments carry crucial information for the identification and tracking of attended speech in the presence of background noise. This study also showed that with the modulation of auditory attention, attended speech can be decoded more robustly from neural activity than from behavioral measures under a wide range of SNR.
Persistent Identifierhttp://hdl.handle.net/10722/304225
ISSN
2023 Impact Factor: 2.4
2023 SCImago Journal Rankings: 0.787
PubMed Central ID
ISI Accession Number ID

 

DC FieldValueLanguage
dc.contributor.authorWang, L-
dc.contributor.authorWu, EX-
dc.contributor.authorChen, F-
dc.date.accessioned2021-09-23T08:57:01Z-
dc.date.available2021-09-23T08:57:01Z-
dc.date.issued2020-
dc.identifier.citationFrontiers in Human Neuroscience, 2020, v. 14, article no. 557534-
dc.identifier.issn1662-5161-
dc.identifier.urihttp://hdl.handle.net/10722/304225-
dc.description.abstractThe attended speech stream can be detected robustly, even in adverse auditory scenarios with auditory attentional modulation, and can be decoded using electroencephalographic (EEG) data. Speech segmentation based on the relative root-mean-square (RMS) intensity can be used to estimate segmental contributions to perception in noisy conditions. High-RMS-level segments contain crucial information for speech perception. Hence, this study aimed to investigate the effect of high-RMS-level speech segments on auditory attention decoding performance under various signal-to-noise ratio (SNR) conditions. Scalp EEG signals were recorded when subjects listened to the attended speech stream in the mixed speech narrated concurrently by two Mandarin speakers. The temporal response function was used to identify the attended speech from EEG responses of tracking to the temporal envelopes of intact speech and high-RMS-level speech segments alone, respectively. Auditory decoding performance was then analyzed under various SNR conditions by comparing EEG correlations to the attended and ignored speech streams. The accuracy of auditory attention decoding based on the temporal envelope with high-RMS-level speech segments was not inferior to that based on the temporal envelope of intact speech. Cortical activity correlated more strongly with attended than with ignored speech under different SNR conditions. These results suggest that EEG recordings corresponding to high-RMS-level speech segments carry crucial information for the identification and tracking of attended speech in the presence of background noise. This study also showed that with the modulation of auditory attention, attended speech can be decoded more robustly from neural activity than from behavioral measures under a wide range of SNR.-
dc.languageeng-
dc.publisherFrontiers Research Foundation. The Journal's web site is located at http://www.frontiersin.org/humanneuroscience/-
dc.relation.ispartofFrontiers in Human Neuroscience-
dc.rightsThis work is licensed under a Creative Commons Attribution-NonCommercial-NoDerivatives 4.0 International License.-
dc.subjectEEG-
dc.subjectTemporal response function (TRF)-
dc.subjectAuditory attention decoding-
dc.subjectSpeech RMS-level segments-
dc.subjectSignal-to-noise ratio-
dc.titleRobust EEG-based decoding of auditory attention with high-RMS-level speech segments in noisy conditions-
dc.typeArticle-
dc.identifier.emailWu, EX: ewu@eee.hku.hk-
dc.identifier.authorityWu, EX=rp00193-
dc.description.naturepublished_or_final_version-
dc.identifier.doi10.3389/fnhum.2020.557534-
dc.identifier.pmid33132874-
dc.identifier.pmcidPMC7576187-
dc.identifier.scopuseid_2-s2.0-85093854220-
dc.identifier.hkuros325428-
dc.identifier.volume14-
dc.identifier.spagearticle no. 557534-
dc.identifier.epagearticle no. 557534-
dc.identifier.isiWOS:000579833200001-
dc.publisher.placeSwitzerland-

Export via OAI-PMH Interface in XML Formats


OR


Export to Other Non-XML Formats