File Download
  Links for fulltext
     (May Require Subscription)
Supplementary

Conference Paper: Multi-task self-supervised learning for disfluency detection

TitleMulti-task self-supervised learning for disfluency detection
Authors
Issue Date2020
PublisherAAAI press
Citation
34th AAAI Conference on Artificial Intelligence (AAAI 2020), New York, 7-12 February 2020. In AAAI 2020 - 34th AAAI Conference on Artificial Intelligence, 2020, p. 9193-9200 How to Cite?
AbstractMost existing approaches to disfluency detection heavily rely on human-annotated data, which is expensive to obtain in practice. To tackle the training data bottleneck, we investigate methods for combining multiple self-supervised tasksi. e., supervised tasks where data can be collected without manual labeling. First, we construct large-scale pseudo training data by randomly adding or deleting words from unlabeled news data, and propose two self-supervised pre-training tasks: (i) tagging task to detect the added noisy words. (ii) sentence classification to distinguish original sentences from grammatically-incorrect sentences. We then combine these two tasks to jointly train a network. The pre-trained network is then fine-tuned using human-annotated disfluency detection training data. Experimental results on the commonly used English Switchboard test set show that our approach can achieve competitive performance compared to the previous systems (trained using the full dataset) by using less than 1% (1000 sentences) of the training data. Our method trained on the full dataset significantly outperforms previous methods, reducing the error by 21% on English Switchboard.
Persistent Identifierhttp://hdl.handle.net/10722/322791
ISBN
ISI Accession Number ID

 

DC FieldValueLanguage
dc.contributor.authorWang, S-
dc.contributor.authorChe, W-
dc.contributor.authorLiu, Q-
dc.contributor.authorQin, P-
dc.contributor.authorLiu, T-
dc.contributor.authorWang, WY-
dc.date.accessioned2022-11-16T06:31:36Z-
dc.date.available2022-11-16T06:31:36Z-
dc.date.issued2020-
dc.identifier.citation34th AAAI Conference on Artificial Intelligence (AAAI 2020), New York, 7-12 February 2020. In AAAI 2020 - 34th AAAI Conference on Artificial Intelligence, 2020, p. 9193-9200-
dc.identifier.isbn9781577358350-
dc.identifier.urihttp://hdl.handle.net/10722/322791-
dc.description.abstractMost existing approaches to disfluency detection heavily rely on human-annotated data, which is expensive to obtain in practice. To tackle the training data bottleneck, we investigate methods for combining multiple self-supervised tasksi. e., supervised tasks where data can be collected without manual labeling. First, we construct large-scale pseudo training data by randomly adding or deleting words from unlabeled news data, and propose two self-supervised pre-training tasks: (i) tagging task to detect the added noisy words. (ii) sentence classification to distinguish original sentences from grammatically-incorrect sentences. We then combine these two tasks to jointly train a network. The pre-trained network is then fine-tuned using human-annotated disfluency detection training data. Experimental results on the commonly used English Switchboard test set show that our approach can achieve competitive performance compared to the previous systems (trained using the full dataset) by using less than 1% (1000 sentences) of the training data. Our method trained on the full dataset significantly outperforms previous methods, reducing the error by 21% on English Switchboard.-
dc.languageeng-
dc.publisherAAAI press-
dc.relation.ispartofAAAI 2020 - 34th AAAI Conference on Artificial Intelligence-
dc.titleMulti-task self-supervised learning for disfluency detection-
dc.typeConference_Paper-
dc.description.naturelink_to_OA_fulltext-
dc.identifier.doi10.1609/aaai.v34i05.6456-
dc.identifier.scopuseid_2-s2.0-85106599788-
dc.identifier.hkuros700004137-
dc.identifier.spage9193-
dc.identifier.epage9200-
dc.identifier.isiWOS:000668126801078-
dc.publisher.placeWashington, D.C.-

Export via OAI-PMH Interface in XML Formats


OR


Export to Other Non-XML Formats