File Download
There are no files associated with this item.
Links for fulltext
(May Require Subscription)
- Publisher Website: 10.1007/978-3-030-32236-6_23
- Scopus: eid_2-s2.0-85075829968
- Find via
Supplementary
-
Citations:
- Scopus: 0
- Appears in Collections:
Conference Paper: Neural Machine Translation with Bilingual History Involved Attention
Title | Neural Machine Translation with Bilingual History Involved Attention |
---|---|
Authors | |
Keywords | Attention mechanism Bilingual history information Neural machine translation |
Issue Date | 2019 |
Publisher | Springer |
Citation | 8th CCF International Conference, NLPCC 2019, Dunhuang, China, October 9–14, 2019. In Tang, J, Kan, MY, Zhao, D, et al. (Eds), Natural Language Processing and Chinese Computing: 8th CCF International Conference, NLPCC 2019, Dunhuang, China, October 9–14, 2019, Proceedings, Part II, p. 265-275. Cham, Switzerland: Springer, 2019 How to Cite? |
Abstract | The using of attention in neural machine translation (NMT) has greatly improved translation performance, but NMT models usually calculate attention vectors independently at different time steps and consequently suffer from over-translation and under-translation. To mitigate the problem, in this paper we propose a method to consider the translated source and target information up to now related to each source word when calculating attentions. The main idea is to keep track of the translated source and target information assigned to each source word at each time step and then accumulate these information to get the completion degree for each source word. In this way, in the later calculation of the attention, the model can adjust the attention weights to give a reasonable final completion degree for each source word. Experimental results show that our method can outperform the strong baseline systems significantly both on the Chinese-English and English-German translation tasks and produce better alignment on the human aligned data set. |
Persistent Identifier | http://hdl.handle.net/10722/312057 |
ISBN | |
ISSN | 2023 SCImago Journal Rankings: 0.606 |
Series/Report no. | Lecture Notes in Computer Science ; 11839 Lecture Notes in Computer Science. Lecture Notes in Artificial Intelligence LNCS sublibrary. SL 7, Artificial Intelligence |
DC Field | Value | Language |
---|---|---|
dc.contributor.author | Xue, Haiyang | - |
dc.contributor.author | Feng, Yang | - |
dc.contributor.author | You, Di | - |
dc.contributor.author | Zhang, Wen | - |
dc.contributor.author | Li, Jingyu | - |
dc.date.accessioned | 2022-04-06T04:32:05Z | - |
dc.date.available | 2022-04-06T04:32:05Z | - |
dc.date.issued | 2019 | - |
dc.identifier.citation | 8th CCF International Conference, NLPCC 2019, Dunhuang, China, October 9–14, 2019. In Tang, J, Kan, MY, Zhao, D, et al. (Eds), Natural Language Processing and Chinese Computing: 8th CCF International Conference, NLPCC 2019, Dunhuang, China, October 9–14, 2019, Proceedings, Part II, p. 265-275. Cham, Switzerland: Springer, 2019 | - |
dc.identifier.isbn | 9783030322359 | - |
dc.identifier.issn | 0302-9743 | - |
dc.identifier.uri | http://hdl.handle.net/10722/312057 | - |
dc.description.abstract | The using of attention in neural machine translation (NMT) has greatly improved translation performance, but NMT models usually calculate attention vectors independently at different time steps and consequently suffer from over-translation and under-translation. To mitigate the problem, in this paper we propose a method to consider the translated source and target information up to now related to each source word when calculating attentions. The main idea is to keep track of the translated source and target information assigned to each source word at each time step and then accumulate these information to get the completion degree for each source word. In this way, in the later calculation of the attention, the model can adjust the attention weights to give a reasonable final completion degree for each source word. Experimental results show that our method can outperform the strong baseline systems significantly both on the Chinese-English and English-German translation tasks and produce better alignment on the human aligned data set. | - |
dc.language | eng | - |
dc.publisher | Springer | - |
dc.relation.ispartof | Natural Language Processing and Chinese Computing: 8th CCF International Conference, NLPCC 2019, Dunhuang, China, October 9–14, 2019, Proceedings, Part II | - |
dc.relation.ispartofseries | Lecture Notes in Computer Science ; 11839 | - |
dc.relation.ispartofseries | Lecture Notes in Computer Science. Lecture Notes in Artificial Intelligence | - |
dc.relation.ispartofseries | LNCS sublibrary. SL 7, Artificial Intelligence | - |
dc.subject | Attention mechanism | - |
dc.subject | Bilingual history information | - |
dc.subject | Neural machine translation | - |
dc.title | Neural Machine Translation with Bilingual History Involved Attention | - |
dc.type | Conference_Paper | - |
dc.description.nature | link_to_subscribed_fulltext | - |
dc.identifier.doi | 10.1007/978-3-030-32236-6_23 | - |
dc.identifier.scopus | eid_2-s2.0-85075829968 | - |
dc.identifier.spage | 265 | - |
dc.identifier.epage | 275 | - |
dc.identifier.eissn | 1611-3349 | - |
dc.publisher.place | Cham, Switzerland | - |