File Download

There are no files associated with this item.

Supplementary

Conference Paper: Stable and Effective Trainable Greedy Decoding for Sequence to Sequence Learning

TitleStable and Effective Trainable Greedy Decoding for Sequence to Sequence Learning
Authors
KeywordsNLP
NMT
Seq2Seq
Beam search
Issue Date2018
Citation
Sixth International Conference on Learning Representations (ICLR) Workshop, Vancouver, Canada, 30 April - 3 May 2018 How to Cite?
AbstractWe introduce a fast, general method to manipulate the behavior of the decoder in a sequence to sequence neural network model. We propose a small neural network actor that observes and manipulates the hidden state of a previously-trained decoder. We evaluate our model on the task of neural machine translation. In this task, we use beam search to decode sentences from the plain decoder for each training set input, rank them by BLEU score, and train the actor to encourage the decoder to generate the highest-BLEU output in a single greedy decoding operation without beam search. Experiments on several datasets and models show that our method yields substantial improvements in both translation quality and translation speed over its base system, with no additional data.
Persistent Identifierhttp://hdl.handle.net/10722/261954

 

DC FieldValueLanguage
dc.contributor.authorChen, Y-
dc.contributor.authorCho, K-
dc.contributor.authorBowman, SR-
dc.contributor.authorLi, VOK-
dc.date.accessioned2018-09-28T04:50:54Z-
dc.date.available2018-09-28T04:50:54Z-
dc.date.issued2018-
dc.identifier.citationSixth International Conference on Learning Representations (ICLR) Workshop, Vancouver, Canada, 30 April - 3 May 2018-
dc.identifier.urihttp://hdl.handle.net/10722/261954-
dc.description.abstractWe introduce a fast, general method to manipulate the behavior of the decoder in a sequence to sequence neural network model. We propose a small neural network actor that observes and manipulates the hidden state of a previously-trained decoder. We evaluate our model on the task of neural machine translation. In this task, we use beam search to decode sentences from the plain decoder for each training set input, rank them by BLEU score, and train the actor to encourage the decoder to generate the highest-BLEU output in a single greedy decoding operation without beam search. Experiments on several datasets and models show that our method yields substantial improvements in both translation quality and translation speed over its base system, with no additional data.-
dc.languageeng-
dc.relation.ispartofInternational Conference on Learning Representations (ICLR) Workshop-
dc.subjectNLP-
dc.subjectNMT-
dc.subjectSeq2Seq-
dc.subjectBeam search-
dc.titleStable and Effective Trainable Greedy Decoding for Sequence to Sequence Learning-
dc.typeConference_Paper-
dc.identifier.emailLi, VOK: vli@eee.hku.hk-
dc.identifier.authorityLi, VOK=rp00150-
dc.identifier.hkuros292172-
dc.publisher.placeVancouver, Canada-

Export via OAI-PMH Interface in XML Formats


OR


Export to Other Non-XML Formats