File Download

There are no files associated with this item.

  Links for fulltext
     (May Require Subscription)
Supplementary

Conference Paper: INK: Injecting kNN Knowledge in Nearest Neighbor Machine Translation

TitleINK: Injecting kNN Knowledge in Nearest Neighbor Machine Translation
Authors
Issue Date11-Jul-2023
Abstract

Neural machine translation has achieved promising results on many translation tasks. However, previous studies have shown that neural models induce a non-smooth representation space, which harms its generalization results. Recently, kNN-MT has provided an effective paradigm to smooth the prediction based on neighbor representations during inference. Despite promising results, kNN-MT usually requires large inference overhead. We propose an effective training framework INK to directly smooth the representation space via adjusting representations of kNN neighbors with a small number of new parameters. The new parameters are then used to refresh the whole representation datastore to get new kNN knowledge asynchronously. This loop keeps running until convergence. Experiments on four benchmark datasets show that INK achieves average gains of 1.99 COMET and 1.0 BLEU, outperforming the state-of-the-art kNN-MT system with 0.02x memory space and 1.9x inference speedup.


Persistent Identifierhttp://hdl.handle.net/10722/333813

 

DC FieldValueLanguage
dc.contributor.authorZhu, Wenhao-
dc.contributor.authorXu, Jingjing-
dc.contributor.authorHuang, Shujian-
dc.contributor.authorKong, Lingpeng-
dc.contributor.authorChen, Jiajun-
dc.date.accessioned2023-10-06T08:39:17Z-
dc.date.available2023-10-06T08:39:17Z-
dc.date.issued2023-07-11-
dc.identifier.urihttp://hdl.handle.net/10722/333813-
dc.description.abstract<p>Neural machine translation has achieved promising results on many translation tasks. However, previous studies have shown that neural models induce a non-smooth representation space, which harms its generalization results. Recently, kNN-MT has provided an effective paradigm to smooth the prediction based on neighbor representations during inference. Despite promising results, kNN-MT usually requires large inference overhead. We propose an effective training framework INK to directly smooth the representation space via adjusting representations of kNN neighbors with a small number of new parameters. The new parameters are then used to refresh the whole representation datastore to get new kNN knowledge asynchronously. This loop keeps running until convergence. Experiments on four benchmark datasets show that INK achieves average gains of 1.99 COMET and 1.0 BLEU, outperforming the state-of-the-art kNN-MT system with 0.02x memory space and 1.9x inference speedup.<br></p>-
dc.languageeng-
dc.relation.ispartofAnnual Meeting of the Association for Computational Linguistics (ACL 2023) (11/07/2023-18/07/2023)-
dc.titleINK: Injecting kNN Knowledge in Nearest Neighbor Machine Translation-
dc.typeConference_Paper-
dc.identifier.doi10.18653/v1/2023.acl-long.888-

Export via OAI-PMH Interface in XML Formats


OR


Export to Other Non-XML Formats