Coverage-based Neural Machine TranslationDownload PDF

20 Apr 2024 (modified: 15 Feb 2016)ICLR 2016 workshop submissionReaders: Everyone
Abstract: Attention mechanism advanced state-of-the-art neural machine translation (NMT) by jointly learning to align and translate. However, attentional NMT ignores past alignment information, which leads to over-translation and under-translation problems. In response to this problem, we maintain a coverage vector to keep track of the attention history. The coverage vector is fed to the attention model to help adjust the future attention, which guides NMT to pay more attention to the untranslated source words. Experiments show that coverage-based NMT significantly improves both translation and alignment qualities over NMT without coverage.
Conflicts: huawei.com, ict.ac.cn
2 Replies

Loading