File Download
There are no files associated with this item.
Links for fulltext
(May Require Subscription)
- Publisher Website: 10.1109/ISCSLP.2018.8706690
- Scopus: eid_2-s2.0-85065880011
- WOS: WOS:000469313700014
Supplementary
- Citations:
- Appears in Collections:
Conference Paper: An end-to-end approach to automatic speech assessment for people with aphasia
Title | An end-to-end approach to automatic speech assessment for people with aphasia |
---|---|
Authors | |
Keywords | Cantonese Pathological speech assessment End-to-end |
Issue Date | 2018 |
Citation | 2018 11th International Symposium on Chinese Spoken Language Processing, ISCSLP 2018 - Proceedings, 2018, p. 66-70 How to Cite? |
Abstract | Conventionally, automatic assessment of pathological speech involves two main steps: (1) extraction of pathology-specific features; (2) classification or regression of extracted features. Given the great variety of speech and language disorders, feature design is never a straightforward task, and yet it is most critical to the performance of assessment. This paper presents an end-to-end approach to automatic speech assessment for Cantonese-speaking people with aphasia (PWA). The assessment is formulated as a binary classification problem to differentiate PWA with high scores of subjective assessment from those with low scores. The sequence-to-one GRU-RNN and CNN models are applied to realize the end-to-end mapping from speech signals to the classification result. The speech features used for assessment are learned implicitly by the neural network model. Preliminary experimental results show that the end-to-end approach could reach a performance level comparable to conventional two-step approach. The experimental results also suggest that CNN performs better than sequence-to-one GRU-RNN in this specific task. |
Persistent Identifier | http://hdl.handle.net/10722/307430 |
ISI Accession Number ID |
DC Field | Value | Language |
---|---|---|
dc.contributor.author | Qin, Ying | - |
dc.contributor.author | Lee, Tan | - |
dc.contributor.author | Wu, Yuzhong | - |
dc.contributor.author | Kong, Anthony Pak Hin | - |
dc.date.accessioned | 2021-11-03T06:22:35Z | - |
dc.date.available | 2021-11-03T06:22:35Z | - |
dc.date.issued | 2018 | - |
dc.identifier.citation | 2018 11th International Symposium on Chinese Spoken Language Processing, ISCSLP 2018 - Proceedings, 2018, p. 66-70 | - |
dc.identifier.uri | http://hdl.handle.net/10722/307430 | - |
dc.description.abstract | Conventionally, automatic assessment of pathological speech involves two main steps: (1) extraction of pathology-specific features; (2) classification or regression of extracted features. Given the great variety of speech and language disorders, feature design is never a straightforward task, and yet it is most critical to the performance of assessment. This paper presents an end-to-end approach to automatic speech assessment for Cantonese-speaking people with aphasia (PWA). The assessment is formulated as a binary classification problem to differentiate PWA with high scores of subjective assessment from those with low scores. The sequence-to-one GRU-RNN and CNN models are applied to realize the end-to-end mapping from speech signals to the classification result. The speech features used for assessment are learned implicitly by the neural network model. Preliminary experimental results show that the end-to-end approach could reach a performance level comparable to conventional two-step approach. The experimental results also suggest that CNN performs better than sequence-to-one GRU-RNN in this specific task. | - |
dc.language | eng | - |
dc.relation.ispartof | 2018 11th International Symposium on Chinese Spoken Language Processing, ISCSLP 2018 - Proceedings | - |
dc.subject | Cantonese | - |
dc.subject | Pathological speech assessment | - |
dc.subject | End-to-end | - |
dc.title | An end-to-end approach to automatic speech assessment for people with aphasia | - |
dc.type | Conference_Paper | - |
dc.description.nature | link_to_subscribed_fulltext | - |
dc.identifier.doi | 10.1109/ISCSLP.2018.8706690 | - |
dc.identifier.scopus | eid_2-s2.0-85065880011 | - |
dc.identifier.spage | 66 | - |
dc.identifier.epage | 70 | - |
dc.identifier.isi | WOS:000469313700014 | - |