• No results found

A Character level Decoder without Explicit Segmentation for Neural Machine Translation

N/A
N/A
Protected

Academic year: 2020

Share "A Character level Decoder without Explicit Segmentation for Neural Machine Translation"

Copied!
11
0
0

Loading.... (view fulltext now)

Full text

Loading

Figure

Figure 1: Bi-scale recurrent neural network
Figure 2:(left) The BLEU scores on En-Csw.r.t. the length of source sentences. (right) Thedifference of word negative log-probabilities be-tween the subword-level decoder and either of thecharacter-level base or bi-scale decoder.
Table 1: BLEU scores of the subword-level, character-level base and character-level bi-scale decodersfor both single models and ensembles
Figure 3: Alignment matrix of a test example from En-De using the BPE→Char (bi-scale) model.

References

Related documents

framework for machine translation ( Kalchbrenner and Blunsom , 2013 ; Cho et al. , 2014 ), which employs a recurrent neural net- work (RNN) encoder to model the source

A Character Aware Encoder for Neural Machine Translation Proceedings of COLING 2016, the 26th International Conference on Computational Linguistics Technical Papers, pages 3063?3070,

Nonparametric Word Segmentation for Machine Translation Proceedings of the 23rd International Conference on Computational Linguistics (Coling 2010), pages 815?823, Beijing, August

6.1 Translation Task: Large Track NIST We first report the experiments using our mono- lingual unigram Dirichlet Process model for word segmentation on the NIST machine translation

End-to-end training makes the neural ma- chine translation (NMT) architecture sim- pler, yet elegant compared to traditional statistical machine translation (SMT). However, little

We did not consider phrase-based statistical machine translation (PBSMT) and hierarchical phrase-based statistical machine translation (HPBSMT), because the OSM approach achieved

We experiment in two different scenarios: 1) a bilin- gual setting where we train a model on data from a single language pair; and 2) a multilingual setting where the task

Tibetan texts are all word segmentation pre-processed in traditional Tibetan machine translations (Guan, 2015). In this article, the traditional method of Tibetan