File Download

There are no files associated with this item.

Supplementary

Conference Paper: CoNT: Contrastive Neural Text Generation

TitleCoNT: Contrastive Neural Text Generation
Authors
Issue Date2022
PublisherCurran Associates, Inc..
Citation
Thirty-Sixth Conference on Neural Information Processing Systems (NeurIPS) (Hybrid), New Orleans, Louisiana, United States of America, November 28-December 9, 2022. In Advances in Neural Information Processing Systems 35 (NeurIPS 2022) How to Cite?
AbstractRecently, contrastive learning attracts increasing interests in neural text generation as a new solution to alleviate the exposure bias problem. It introduces a sequence-level training signal which is crucial to generation tasks that always rely on auto-regressive decoding. However, previous methods using contrastive learning in neural text generation usually lead to inferior performance. In this paper, we analyse the underlying reasons and propose a new Contrastive Neural Text generation framework, CoNT. CoNT addresses bottlenecks that prevent contrastive learning from being widely adopted in generation tasks from three aspects -- the construction of contrastive examples, the choice of the contrastive loss, and the strategy in decoding. We validate CoNT on five generation tasks with ten benchmarks, including machine translation, summarization, code comment generation, data-to-text generation and commonsense generation. Experimental results show that CoNT clearly outperforms its baseline on all the ten benchmarks with a convincing margin. Especially, CoNT surpasses previous the most competitive contrastive learning method for text generation, by 1.50 BLEU on machine translation and 1.77 ROUGE-1 on summarization, respectively. It achieves new state-of-the-art on summarization, code comment generation (without external data) and data-to-text generation.
Persistent Identifierhttp://hdl.handle.net/10722/317986

 

DC FieldValueLanguage
dc.contributor.authorAn, C-
dc.contributor.authorFeng, J-
dc.contributor.authorLv, K-
dc.contributor.authorKong, L-
dc.contributor.authorQiu, X-
dc.contributor.authorHuang, X-
dc.date.accessioned2022-10-07T10:30:35Z-
dc.date.available2022-10-07T10:30:35Z-
dc.date.issued2022-
dc.identifier.citationThirty-Sixth Conference on Neural Information Processing Systems (NeurIPS) (Hybrid), New Orleans, Louisiana, United States of America, November 28-December 9, 2022. In Advances in Neural Information Processing Systems 35 (NeurIPS 2022)-
dc.identifier.urihttp://hdl.handle.net/10722/317986-
dc.description.abstractRecently, contrastive learning attracts increasing interests in neural text generation as a new solution to alleviate the exposure bias problem. It introduces a sequence-level training signal which is crucial to generation tasks that always rely on auto-regressive decoding. However, previous methods using contrastive learning in neural text generation usually lead to inferior performance. In this paper, we analyse the underlying reasons and propose a new Contrastive Neural Text generation framework, CoNT. CoNT addresses bottlenecks that prevent contrastive learning from being widely adopted in generation tasks from three aspects -- the construction of contrastive examples, the choice of the contrastive loss, and the strategy in decoding. We validate CoNT on five generation tasks with ten benchmarks, including machine translation, summarization, code comment generation, data-to-text generation and commonsense generation. Experimental results show that CoNT clearly outperforms its baseline on all the ten benchmarks with a convincing margin. Especially, CoNT surpasses previous the most competitive contrastive learning method for text generation, by 1.50 BLEU on machine translation and 1.77 ROUGE-1 on summarization, respectively. It achieves new state-of-the-art on summarization, code comment generation (without external data) and data-to-text generation.-
dc.languageeng-
dc.publisherCurran Associates, Inc..-
dc.relation.ispartofAdvances in Neural Information Processing Systems 35 (NeurIPS 2022)-
dc.titleCoNT: Contrastive Neural Text Generation-
dc.typeConference_Paper-
dc.identifier.emailKong, L: lpk@cs.hku.hk-
dc.identifier.authorityKong, L=rp02775-
dc.identifier.hkuros337877-
dc.publisher.placeUnited States-

Export via OAI-PMH Interface in XML Formats


OR


Export to Other Non-XML Formats