• No results found

Neural Machine Translation between Myanmar (Burmese) and Rakhine (Arakanese)

N/A
N/A
Protected

Academic year: 2020

Share "Neural Machine Translation between Myanmar (Burmese) and Rakhine (Arakanese)"

Copied!
9
0
0

Loading.... (view fulltext now)

Full text

(1)

Proceedings of VarDial, pages 80–88 80

Neural Machine Translation

Between Myanmar (Burmese) and Rakhine (Arakanese)

Thazin Myint Oo , Ye Kyaw Thu and Khin Mar Soe

Natural Language Processing Lab., University of Computer Studies Yangon, Myanmar Language and Semantic Technology Research Team (LST), NECTEC, Thailand

Language and Speech Science Research Lab., Waseda University, Japan {thazinmyintoo, khinmarsoe}@ucsy.edu.mm, ka2pluskha2@gmail.com

Abstract

This work explores neural machine trans-lation between Myanmar (Burmese) and Rakhine (Arakanese). Rakhine is a lan-guage closely related to Myanmar, often considered a dialect. We implemented three prominent neural machine transla-tion (NMT) systems: recurrent neural networks (RNN), transformer, and con-volutional neural networks (CNN). The systems were evaluated on a Myanmar-Rakhine parallel text corpus developed by us. In addition, two types of word seg-mentation schemes for word embeddings were studied: Word-BPE and Syllable-BPE segmentation. Our experimental re-sults clearly show that the highest quality NMT and statistical machine translation (SMT) performances are obtained with Syllable-BPE segmentation for both types of translations. If we focus on NMT, we find that the transformer with Word-BPE segmentation outperforms CNN and RNN for both Myanmar-Rakhine and Rakhine-Myanmar translation. However, CNN with Syllable-BPE segmentation obtains a higher score than the RNN and trans-former.

1 Introduction

The Myanmar language includes a number of mutually intelligible Myanmar dialects, with a largely uniform standard dialect used by most Myanmar standard speakers. Speakers of the standard Myanmar may find the dialects hard to follow. The alternative phonology, mor-phology, and regional vocabulary cause some problems in communication. Machine trans-lation (MT) has so far neglected the impor-tance of properly handling the spelling, lexi-cal, and grammar divergences among language varieties. In the Republic of the Union of Myanmar, there are many ethnical groups, and dialectal varieties exist within the stan-dard Myanmar language.

To address this problem, we are developing a Myanmar and Rakhine dialectal corpus with monolingual and parallel text. We conducted

statistical machine translation (SMT) experi-ments and obtained results similar to previous research (Oo et al.,2018).

Deep learning revolution brings rapid and dramatic change to the field of machine trans-lation. The main reason for moving from SMT to neural machine translation (NMT) is that it achieved the fluency of translation that was a huge step forward compared with the previ-ous models. In a trend that carries over from SMT, the strongest NMT systems benefit from subtle architecture modifications and hyperpa-rameter tuning.

NMT models have advanced the state of the art by building a single neural net-work that can learn representations better (Sutskever et al.,2014a). Other authors (Rik-ters et al., 2018) conducted experiments with different NMTs for less-resourced and mor-phologically rich languages, such as Estonian and Russian. They compared the multi-way model performance to one-way model perfor-mance, by using different NMT architectures that allow achieving state-of-the-art transla-tion. For the multiway model trained using the transformer network architecture, the re-ported improvement over the baseline meth-ods was +3.27 bilingual evaluation understudy (BLEU) points.

(Honnet et al.,2017) proposed solutions for the machine translation of a family of dialects, Swiss German, for which parallel corpora are scarcee. The authors presented three strate-gies for normalizing Swiss German input to address the regional and spelling diversity. The results show that character-based neural machine translation was the most promising strategy for text normalization and that in combination with phrase-based statistical ma-chine translation it achieved 36% BLEU score. In their study, NMT outperformed SMT.

(2)

neural network (CNN), and transformer. We investigated the translation quality of the corresponding hyper-parameters (batch size, learning rate, cell type, and activation func-tion) in machine translation between the stan-dard Myanmar and national varieties of the same group of languages. In addition, we used two types of segmentation schemes: word byte pair encoding (Word-BPE) segmenta-tion and syllable byte pair encoding (Syllable-BPE) segmentation. We compared the per-formance of this method to SMT and NMT experiments with the RNN, transformer, and CNN. We found that the transformer with Word-BPE segmentation outperformed both CNN and RNN for both Myanmar-Rakhine and Myanmar-Rakhine translations. We also found that CNN with Syllable-BPE segmen-tation obtained a higher score compared with RNN and the transformer.

2 Rakhine Language

Rakhine (Arakanese) is one of the eight national ethnic groups in the Republic of the Union of Myanmar. The Arakan was officially altered to “Rakhine” in 1989 and is located on a narrow coastal strip on the west of Myanmar, 300 miles long and 50 to 20 miles wide. The total population in all countries is nearly 3 million. The Rakhine language has been studied by researchers. L.F-Taylor’s “The Dialects of Burmese” described compar-ative pronunciation, sentence construction, and grammar usage in Rakhine, Dawei, In-tha, Taung-yoe, Danu, and Yae. Professor Denise Bernot, in “The vowel system of Arakanese and Tavoyan,” mainly emphasized the vowels of standard Myanmar and Tavoyan (Dawei) in 1965. In “Three Burmese Dialects” (1969), the linguist John Okell studied the spoken language of Myanmar, Dawei, and In-tha: specifically, usage of grammar and vowel differences (OKELL, 1995). Although the Rakhine language used the script as Arakanese or Rakkhawanna Akkhara before at least the 8th century A.D., the current Rakhine script is nearly the same as the Myanmar script. Generally, the Arakanese language is mutually intelligible with the Myanmar language and has the same word order (namely, subject-object-verb (SOV)). Examples of parallel sentences in Myanmar (my) and Rakhine (rk) are given as follows.

rk: ဒေယာ တစ် ထည် ဇာေလာက်ေလး ။ my: လုံချည် တစ် ထည် ဘယ်ေလာက်လဲ ။

(“How much for a longyi?” in English)

rk: ကေလေချ တိ ေဘာလုံး ေကျာက် နီေရ ။ my: ေကာင်ေလး ေတွ ေဘာလုံး ကန် ေနတယ် ။ (“Boys are playing football” in English)

rk: ဇာ ေြပာ နီချင့် ယင်းသူရို ့ ။ my: သူတို့ ဘာ ေြပာ ေနတာလဲ ။

(“What are they talking about” in English)

rk: အေဘာင်သျှင် စျီး က သပုံ ဝယ် လာတယ် ။ my: အဘွား ေစျး က ဆပ်ြပာ ဝယ် လာတယ် ။ (“The grandmother buys soap from the market” in English)

3 Difference between the Rakhine and standard Myanmar language

(3)

4 Segmentation

4.1 Word Segmentation

In both Myanmar and Rakhine texts, spaces are used to separate the phrases for easier reading. The spaces are not strictly necessary and are rarely used in short sentences. There are no clear rules for using spaces. Thus, spaces may (or may not) be inserted between words, phrases, and even between root words and their affixes. Although Myanmar sen-tences of ASEAN-MT corpus (Boonkwan and Supnithi, 2013) are already segmented, we have to consider some rules for manual word segmentation of Rakhine sentences. We de-fined Rakhine “word” to be a meaningful unit. Affix, root word, and suffix (s) are separated such as "စား ဗျာယ်", "စား ပီးဗျာယ်", "စား ဖို့ဗျာယ်". Here, "စား" (“eat” in English) is a root word and the others are suffixes for past and future tenses. As Myanmar language, Rakhine plural nouns are identified by the following particle. We added a space between the noun and the following particle: for example a Rakhine word "ကလိန့်ေမေချ တိ" (ladies) is segmented as two words "ကလိန့်ေမေချ" and the particle "တိ". In Rakhine grammar, particles describe the type of noun and are used after a number or text number. For example, a Rakhine word "အ‌ြွကီေစ့နှစ်ခတ်"(“two coins” in English) is segmented as "အ‌ြွကီေစ့ နှစ် ခတ်". In our manual word segmentation rules, compound nouns are considered as one word. Thus, a Rakhine compound word "ေဖ့သာ" + "အိတ်"

(“money” + “bag” in English) is written as one word "ေဖ့သာအိတ်" (“wallet” in English). Rakhine adverb words such as "အဂေယာင့်" (“really” in English), "အြမန်" (“quickly” in English) are also considered as one word. The following is an example of word segmentation for a Rakhine sentence in our corpus, and the meaning is “Among the four air conditioners in our room, two are out of order.”

Unsegmented sentence:

အကျွန်ရို ့အခန်းထဲမှာဟိေရလီအီးစက်ေလးလုံးမှာနှစ်လုံးပျက် နီေရ ။

Segmented sentence:

အကျွန်ရို ့ အခန်း ထဲမှာ ဟိ ေရ လီအီးစက် ေလး လုံး မှာ နှစ် လုံး ပျက် နီေရ ။

4.2 Syllable Segmentation

Generally, Myanmar words are composed of multiple syllables, and most of the syllables are composed of more than one character. Syl-lables are also the basic units for pronuncia-tion of Myanmar words. If we only focus on consonant-based syllables, the structure of the syllable can be described with Backus normal form (BNF) as follows:

Syllable:=CM W[CK][D]

Here, C stands for consonants, M for me-dials, V for vowels, K for vowel killer char-acter, and D for diacritic characters. Myan-mar syllable segmentation can be done with a rule-based approach (Maung and Makami, 2008;Thu et al.,2013), finite state automaton (FSA) (Hlaing, 2012), or regular expressions (RE) (https://github.com/ye-kyawthu/sylbreak). In our experiments, we used RE-based Myan-mar syllable segmentation tool named “syl-break.”

4.3 Byte-Pair-Encoding

(Sennrich et al., 2016) proposed a method to enable open-vocabulary translation of rare and unknown words as a sequence of sub-word units representing BPE algorithm (Gage, 1994). The input is a monolingual corpus for a language (one side of the parallel training data, in our case) and starts with an initial vocabulary, the characters in the text corpus. The vocabulary is updated using an iterative greedy algorithm. In every iteration, the most frequent bigram (based on the current vocabu-lary) in the corpus is added to the vocabulary (the merge operation). The corpus is again en-coded using the updated vocabulary, and this process is repeated for a predetermined num-ber of merge operations. The numnum-ber of merge operations is the only hyperparameter of the system that needs to be tuned. A new word can be segmented by looking up the learnt vo-cabulary. For instance, a new word “rocket,” ဒုံးပျံmay be segmented asဒ@@◌ုံး ပျံafter look-ing up the learnt vocabulary, assumlook-ing ဒand ◌ုံး ပျံ as BPE units learnt during the training. 5 Encoder-Decoder Models for

NMT

The core idea is to encode a variable-length input sequence of tokens into a sequence of vector representations, and then decode these representations into a sequence of output to-kens. Formally, with a given sentence X =

(4)

NMT system models p(Y|X) as a target lan-guage sequence model, conditioning the prob-ability of target word yt on target history

Y1:t−1 and source sentence X. Both xi and

yt are integer IDs given by the source and

target vocabulary mapping, Vsrc and Vtrg,

built from the training data tokens and rep-resented as one-hot vectors xi ∈ {0,1}|Vsrc|

and yt ∈ {0,1}|Vtrg|. These are

embed-ded into e-dimensional vector representations (Vaswani et al., 2017) ESxi and ETyt,

us-ing embeddus-ing matrices Re×|Vsrc| and ET

Re×|Vtrg|. The target sequence is factorized as p(Y|X;θ) = ∏mt=1p(yt|Y1:t−1, X;θ). The

model, parameterized by θ, consists of an en-coder and deen-coder part, which vary depending on the model architecture. p(yt|Y1:t−1, X;θ)is

parameterized via a softmax output layer over some decoder representations S¯t:

p(yt|Y1:t−1, X;θ) =sof tmax(Wo¯st+bo), (1)

whereWoscales to the dimension of the target

vocabulary Vtrg.

5.1 Stacked RNN with attention

The encoder consists of a bidirectional RNN followed by a stack of unidirectional RNNs. An RNN at layerlproduces a sequence of hid-den stateshl

1. . .hln:

hl

i=fenc(hli−1,h l

i−1), (2)

wherefrnnis some non-linear function, such

as a gated recurrent unit (GRU) or long short-term memory (LSTM) cell, and hl−1

i is the

output from the lower layer at step i. The bidirectional RNN on the lowest layer uses em-beddings ESxi as input and concatenates the

hidden states of a forward and a reverse RNN:

h0 i = [

−→

h0 i;

←−

h0

i]. With deeper networks,

learn-ing becomes increaslearn-ingly difficult (Hochreiter et al.,2001;Pascanu et al.,2012), and residual connections of the formhl

i=hli−1+fenc(h l−1 i , hl

i−1) become essential (He et al.,2016).

The decoder consists of an RNN to predict one target word at a time through a state vec-tors:

st=fdec([ETyt−1; ¯st−1],st−1), (3)

where fdec is a multilayer RNN, st−1 the

previous state vector, and ¯st−1 the

source-dependent attentional vector. Providing the attentional vector as an input to the first de-coder layer is also called input feeding (Lu-ong et al., 2015). The initial decoder hidden state is a non-linear transformation of the last

encoder hidden state: s0 = tanh(Winithn +

binit). The attentional vector¯st combines the

decoder state with acontext vector ct:

¯st=tanh(W¯s[st;ct]), (4)

where ct is a weighted sum of encoder

hid-den states: ct = ∑n

i=1αtihi. The attention

vectorαtis computed by an attention network

(Bahdanau et al.,2014;Luong et al.,2015):

αti=softmax(score(st,hi))

score(s,h) =va tanh(Wus+Wvh).

(5)

5.2 Self-Attentional Transformer

The transformer model (Vaswani et al.,2017) uses attention to replace recurrent dependen-cies, making the representation at time step i independent from the other time steps. This requires the explicit encoding of positional in-formation in the sequence by adding fixed or learned positional embeddings to the embed-ding vectors.

The encoder uses several identical blocks consisting of two core sublayers: self-attention and a feedforward network. The self-attention mechanism is a variation of the dot-product attention (Luong et al., 2015) generalized to three inputs: query matrixQRn×d, key

ma-trix KRn×d, and value VRn×d, where d

denotes the number of hidden units. (Vaswani et al., 2017) further extend attention to mul-tiple heads, allowing for focusing on different parts of the input. A singlehead uproduces a context matrix

Cu =softmax (

QWQu

(

KWK

u

)T

du

)

VWV

u,

(6) where matrices WQu, WKu, and WVu are in

Rd×du. The final context matrix is given by concatenating the heads, followed by a linear transformation: C = [C1;. . .;Ch]WO. The

form in Equation6suggests parallel computa-tion across all time steps in a single large ma-trix multiplication. Given a sequence of hid-den stateshi (or input embeddings),

concate-nated toHRn×d, the encoder computes

self-attention usingQ=K=V=H. The second subnetwork of an encoder block is a feedfor-ward network with ReLU activation defined as

(5)

which is also easily parallelizable across time steps. Each sublayer, self-attention and feed-forward network, is followed by a postprocess-ing stack of dropout, layer normalization, and residual connection.

The decoder uses the same self-attention and feedforward networks subnetworks. To maintain auto-regressiveness of the model, self-attention is masked out on future time steps according to (Vaswani et al., 2017). In addition to self-attention, a source attention layer, which uses the encoder hidden states as key and value inputs, is added. Given decoder hidden states SRm×s and the encoder

hid-den states of the final encoder layerHl, source

attention is computed as in Equation 5 with

Q = S,K = Hl,V = Hl. As in the

en-coder, each sublayer is followed by a postpro-cessing stack of dropout, layer normalization (Ba et al.,2016), and residual connection.

5.3 Fully Convolutional Models

The convolutional model (Gehring et al.,2017) uses convolutional operations and also dis-penses with recurrence. Hence, input embed-dings are again augmented with explicit posi-tional encodings.

The convolutional encoder applies a set of (stacked) convolutions that are defined as

hl

i =v(Wl[hli−−⌊1k/2;. . .;h l−1

i+⌊k/2] +b

l) +hl−1 i ,

(8) wherevis a non-linearity such as a gated lin-ear unit (Gehring et al.,2017;Dauphin et al., 2016), and Wl Rdcnn×kd are the

convolu-tional filters. To increase the context window captured by the encoder architecture, multiple layers of convolutions are stacked. To main-tain sequence length across multiple stacked convolutions, inputs are padded with zero vec-tors.

The decoder is similar to the encoder but adds an attention mechanism to every layer. The output of the target side convolution

sl∗

t =v(Wlstl−−1k+1;. . .; ¯s l−1

t ] +bl) (9)

is combined to form S and then fed as an input to the attention mechanism of Equa-tion 6 with a single attention head and Q =

S,K = Hl,V = Hl, resulting in a set of

context vectors ct. The full decoder hidden

state is a residual combination with the con-text such that

¯

sl

t=slt∗+ct+ ¯slt−1 (10)

To avoid convolving over future time steps at timet, the input is padded to the left.

6 Experiments

6.1 Corpus Preparation and Statistics

We used 18,373 Myanmar sentences (with no name entity tags) of the ASEAN-MT Paral-lel Corpus (Boonkwan and Supnithi, 2013), which is a parallel corpus in the travel do-main. It contains six main categories: peo-ple (greeting, introduction, and communica-tion), survival (transportation, accommoda-tion, and finance), food (food, beverages, and restaurants), fun (recreation, traveling, shop-ping, and nightlife), resource (number, time, and accuracy), special needs (emergency and health). Manual translation into the Rakhine language was done by native Rakhine students from two Myanmar universities, and the trans-lated corpus was checked by the editor of a Rakhine newspaper. Word segmentation for Rakhine was done manually, and there are ex-actly 123,018 words in total. We used 14,076 sentences for training, 2,485 sentences for de-velopment, and 1,812 sentences for evaluation.

6.2 Moses SMT system

(6)

Batch Size RNN Transformer CNN

my-rk rk-my my-rk rk-my my-rk rk-my

128 79.86 81.44 79.64 82.01 80.82 83.59

256 80.76 82.94 79.47 81.37 80.33 83.54

[image:6.595.143.457.61.125.2]

512 80.00 82.26 79.47 80.79 79.86 81.38

Table 1: BLEU scores of Syllable-BPE segmentation with different batch sizes for three NMT models

Batch Size RNN Transformer CNN

my-rk rk-my my-rk rk-my my-rk rk-my

128 60.02 44.44 72.70 72.82 69.03 72.24

256 60.31 46.47 73.39 72.45 65.61 68.26

[image:6.595.93.506.406.510.2]

512 42.76 34.93 73.30 72.95 67.89 71.68

Table 2: BLEU scores of Word-BPE segmentation with different batch sizes for three NMT models

Learning rate

RNN Transformer

GRU LSTM GRU LSTM

my-rk rk-my my-rk rk-my my-rk rk-my my-rk rk-my 0.0001 79.47 81.37 79.48 80.88 80.76 82.94 80.26 83.02 0.0002 79.82 81.65 82.85 82.07 80.88 81.54 80.90 82.99 0.0003 80.22 82.23 80.24 82.13 80.92 82.63 81.78 83.30 0.0004 80.65 82.66 80.85 82.33 81.25 82.54 81.92 84.06

0.0005 80.41 81.46 81.98 83.86 80.57 82.30 80.65 82.51

Table 3: BLEU scores for batch size 256 of Syllable-BPE segmentation with different learning rates and two memory cell types on RNN and the transformer

Learning rate

Batch Size 128 Batch Size 256

ReLu Soft-ReLu ReLu Soft-ReLu

my-rk rk-my my-rk rk-my my-rk rk-my my-rk rk-my 0.0001 81.37 83.29 80.00 81.97 80.26 81.97 80.03 81.08 0.0002 81.01 82.24 79.89 82.50 80.07 82.29 80.01 81.51 0.0003 80.99 81.59 80.11 83.34 81.16 81.69 82.14 84.08

0.0004 N/A N/A N/A N/A 79.74 80.87 83.75 83.06

0.0005 N/A N/A N/A N/A 79.05 82.43 81.44 83.31

Table 4: BLEU scores for batch sizes 128 and 256 of Syllable-BPE segmentation with different learning rates and two activation functions on CNN

Segmentation OSM RNN Transformer CNN

my-rk rk-my my-rk rk-my my-rk rk-my my-rk rk-my Syllable-BPE 82.71 84.36 82.03 83.98 82.85 82.65 83.75 84.08 Word-BPE 77.12 75.27 60.31 46.47 73.39 72.95 69.03 72.24

Table 5: Comparison of SMT and NMT (top BLEU scores) on two segmentation schemes

6.3 Framework for NMT

An open-source sequence-to-sequence toolkit for NMT written in Python (Hieber et al., 2017) and built on Apache MXNET (Chen et al., 2015), the toolkit offers scalable train-ing and inference for the three most prominent encoder-decoder architectures: attentional recurrent neural network (Schwenk, 2012;

Kalchbrenner and Blunsom, 2013; Sutskever et al., 2014b), self-attentional transformers (Vaswani et al.,2017), and fully convolutional networks (Gehring et al.,2017).

6.4 Training Details

[image:6.595.93.506.555.608.2]
(7)

per-formance on the validation set has improved for 8 checkpoints, the learning rate is multi-plied by 32 checkpoints. All the neural net-works have eight layers. For RNN Seq2Seq, the encoder has one bi-directional LSTM and six stacked unidirectional LSTMs, and the en-coder is a stack of eight unidirectional LSTMs. The size of hidden states is 512. We apply layer-normalization and label smoothing (0.1) in all models. We tie the source and target em-beddings. The dropout rate of the embeddings and transformer blocks is set to (0.1). The dropout rate of RNNs is (0.2). The attention mechanism in the transformer has eight heads. We applied three different batch sizes (128, 256, and 512) for RNN, Transformer, and CNN network architectures. The learning rate varies from 0.0001 to 0.0005. Two memory cell types GRU and LSTM were used for the RNN and transformer. Moreover, two activa-tion funcactiva-tions were applied to the CNN ar-chitecture. The comparison between Syllable-BPE and Word-Syllable-BPE segmentation schemes was done for both SMT (i.e., OSM) and NMT (RNN, Transformer, and CNN) techniques. All experiments are run on NVIDIA Tesla K80 24GB GDDR5. We trained all mod-els for the maximum number of epochs us-ing the AdaGrad and adaptive moment esti-mation (Adam) optimizer. The BPE segmen-tation models were trained with a vocabulary of 8,000.

6.5 Evaluation

We used automatic criteria to evaluate the ma-chine translation output. The metric BLEU (Papineni et al.,2002) measures the adequacy of the translation between language pairs, such as Myanmar and English. The Higher BLEU scores are better. Before computing BLEU, the translations were decomposed into their constituent syllables to ensure that the results are cross-comparable.

7 Results and Discussion

The BLEU score results for three NMT approaches (RNN, Transformer, and CNN) with three batch sizes (128, 256, and 512) for Syllable-BPE segmentation scheme are shown in Table 1. Bold numbers indi-cate the highest BLEU score among differ-ent batch sizes. CNN achieved the high-est BLEU scores for both Myanmar-Rakhine and Rakhine-Myanmar translations. How-ever, the transformer architecture achieved the top BLEU scores for Word-BPE segmen-tation schemes for both Myanmar-Rakhine

and Rakhine-Myanmar neural machine trans-lations (see Table 2).

From the experimental results of Table 1 and Table2, we noticed that RNN and Trans-former NMT with Syllable-BPE have a de-creased translation performance for batch size 512. Thus, we used batch size 256 for fur-ther experiments with the RNN and trans-former architectures. The NMT performance of the RNN and transformer with Syllable-BPE segmentation schemes together with dif-ferent learning rates (from 0.0001 to 0.0005) and two different memory cell types (GRU and LSTM) can be seen in Table 3. From these BLEU scores of the RNN and transformer ap-proaches, LSTM gave the highest NMT per-formance for both Myanmar-Rakhine dialect translation and vice versa.

To observe the maximum translation per-formance of CNN architecture, we extended experiments by using two activation functions (ReLu and Soft-ReLu), two batch sizes (128 and 256), and five learning rates (from 0.0001 to 0.0005) (see Table 4). Here, bold num-bers indicate the highest BLEU scores of each batch size. From these results, we can clearly see that Soft-ReLu achieved the highest BLEU scores for both Myanmar to Rakhine and Rakhine to Myanmar translations. We found that the training processes with learning rate 0.0004 and 0.0005 were stopped for the batch size 128 for both ReLu and Soft-ReLu activa-tion funcactiva-tions.

We also made a comparison between SMT and NMT, and the results can be seen in Ta-ble 5. In this study, we run only OSM ap-proach for the SMT experiments based on the our previous SMT work (Oo et al.,2018). The Table5presents that although CNN achieved the top BLEU score (83.75) for Myanmar to Rakhine translation, OSM gave the best BLEU (84.36) score for Rakhine to Myanmar translation. Furthermore, we also found that Syllable-BPE segmentation scheme is working well for both SMT and NMT for Myanmar-Rakhine dialect language pair.

(8)

same word order of SOV and also share a lot of vocabulary. For these reasons, we assume that both SMT and NMT systems reach a very high machine translation performance.

8 Conclusion

This paper presents the first study of the neural machine translation between Stan-dard Myanmar and Rakhine (a spoken Myan-mar dialect). We implemented three promi-nent NMT systems: RNN, transformer, and CNN. The systems were evaluated on a Myanmar-Rakhine parallel text corpus that we are developing. We also investigated two types of segmentation schemes (Word-BPE segmentation and Syllable-(Word-BPE seg-mentation). Our results clearly show that the highest performance of SMT and NMT was obtained with Syllable-BPE segmenta-tion for both Myanmar-Rakhine and Rakhine-Myanmar translation. If we only focus on NMT, we find that the transformer with Word-BPE segmentation outperforms CNN and RNN for both Myanmar-Rakhine and Rakhine-Myanmar. We also find that CNN with syllable-BPE segmentation obtains a higher BLEU score compared with the RNN and transformer. In the near future, we plan to conduct a further study with a focus on NMT models with one more subword segmen-tation scheme SentencePiece for Myanmar-Rakhine NMT. Moreover, we intend to inves-tigate SMT and NMT approaches for other Myanmar dialect languages, such as Myeik and Dawei.

References

Lei Jimmy Ba, Ryan Kiros, and Geoffrey E. Hinton. 2016. Layer normalization. CoRR, abs/1607.06450.

Dzmitry Bahdanau, Kyunghyun Cho, and Yoshua Bengio. 2014. Neural machine translation by jointly learning to align and translate. CoRR, abs/1409.0473.

Prachya Boonkwan and Thepchai Supnithi. 2013. Technical report for the network-based asean language translation public service project. On-line Materials of Network-based ASEAN Lan-guages Translation Public Service for Members,

NECTEC.

Tianqi Chen, Mu Li, Yutian Li, Min Lin, Naiyan Wang, Minjie Wang, Tianjun Xiao, Bing Xu, Chiyuan Zhang, and Zheng Zhang. 2015. Mxnet: A flexible and efficient machine learning library for heterogeneous distributed systems. CoRR, abs/1512.01274.

Yann N. Dauphin, Angela Fan, Michael Auli, and David Grangier. 2016. Language model-ing with gated convolutional networks. CoRR, abs/1612.08083.

Philip Gage. 1994. A new algorithm for data com-pression. C Users J., 12(2):23–38.

Jonas Gehring, Michael Auli, David Grangier, De-nis Yarats, and Yann N. Dauphin. 2017. Convo-lutional sequence to sequence learning. CoRR, abs/1705.03122.

Kaiming He, Xiangyu Zhang, Shaoqing Ren, and Jian Sun. 2016. Deep residual learning for image recognition. In2016 IEEE Conference on Com-puter Vision and Pattern Recognition, CVPR 2016, Las Vegas, NV, USA, June 27-30, 2016, pages 770–778.

Kenneth Heafield. 2011. Kenlm: Faster and smaller language model queries. In Proceedings of the Sixth Workshop on Statistical Machine Translation, WMT ’11, pages 187–197, Strouds-burg, PA, USA. Association for Computational Linguistics.

Felix Hieber, Tobias Domhan, Michael Denkowski, David Vilar, Artem Sokolov, Ann Clifton, and Matt Post. 2017. Sockeye: A toolkit for neural machine translation. CoRR, abs/1712.05690.

Tin Htay Hlaing. 2012. Manually constructed context-free grammar for myanmar syllable structure. InIn Proceedings of the Student Re-search Workshop at the 13 th Conference of the European Chapter of the Association for Com-putational Linguistics (EACL’12), pages 32–37.

Sepp Hochreiter, Yoshua Bengio, and Paolo Fras-coni. 2001. Gradient flow in recurrent nets: the difficulty of learning long-term dependencies. In J. Kolen and S. Kremer, editors,Field Guide to Dynamical Recurrent Networks. IEEE Press.

Pierre-Edouard Honnet, Andrei Popescu-Belis, Claudiu Musat, and Michael Baeriswyl. 2017. Machine translation of low-resource spoken di-alects: Strategies for normalizing swiss german. CoRR, abs/1710.11035.

Hideki Isozaki, Tsutomu Hirao, Kevin Duh, Kat-suhito Sudoh, and Hajime Tsukada. 2010. Au-tomatic evaluation of translation quality for dis-tant language pairs. In Proceedings of the 2010 Conference on Empirical Methods in Nat-ural Language Processing, pages 944–952, Cam-bridge, MA. Association for Computational Lin-guistics.

(9)

Philipp Koehn, Hieu Hoang, Alexandra Birch, Chris Callison-Burch, Marcello Federico, Nicola Bertoldi, Brooke Cowan, Wade Shen, Christine Moran, Richard Zens, Chris Dyer, Ondřej Bo-jar, Alexandra Constantin, and Evan Herbst. 2007. Moses: Open source toolkit for statisti-cal machine translation. In Proceedings of the 45th Annual Meeting of the ACL on Interactive Poster and Demonstration Sessions, ACL ’07, pages 177–180, Stroudsburg, PA, USA. Associ-ation for ComputAssoci-ational Linguistics.

Minh-Thang Luong, Hieu Pham, and Christo-pher D. Manning. 2015. Effective approaches to attention-based neural machine translation. CoRR, abs/1508.04025.

Zin Maung Maung and Yoshiki Makami. 2008. A rule-based syllable segmentation of myanmar text. InProceedings of IJCNLP-08 work-shop of NLP for Less Privileged Language, pages 51–58.

Franz Josef Och. 2003. Minimum error rate train-ing in statistical machine translation. In Pro-ceedings of the 41st Annual Meeting on Associa-tion for ComputaAssocia-tional Linguistics - Volume 1, ACL ’03, pages 160–167, Stroudsburg, PA, USA. Association for Computational Linguistics.

John OKELL. 1995. Three burmese dialects. Pa-pers in Southeast Asian Linguistics No.13, Stud-ies in Burmese Languages, 13:1–138.

Thazin Myint Oo, Ye Kyaw Thu, and Khin Mar Soe. 2018. Statistical machine translation between myanmar (burmese) and rakhine (arakanese). InProceedings of ICCA2018, pages 304–311.

Kishore Papineni, Salim Roukos, Todd Ward, and Wei-Jing Zhu. 2002. Bleu: A method for au-tomatic evaluation of machine translation. In Proceedings of the 40th Annual Meeting on Asso-ciation for Computational Linguistics, ACL ’02, pages 311–318, Stroudsburg, PA, USA. Associ-ation for ComputAssoci-ational Linguistics.

Razvan Pascanu, Tomas Mikolov, and Yoshua Ben-gio. 2012. Understanding the exploding gradient problem. CoRR, abs/1211.5063.

Matīss Rikters, Mārcis Pinnis, and Rihards Krišlauks. 2018. Training and Adapting Mul-tilingual NMT for Less-resourced and Morpho-logically Rich Languages. In Proceedings of the Eleventh International Conference on Lan-guage Resources and Evaluation (LREC 2018), Miyazaki, Japan. European Language Resources Association (ELRA).

Holger Schwenk. 2012. Continuous space transla-tion models for phrase-based statistical machine translation. In Proceedings of COLING 2012: Posters, pages 1071–1080. The COLING 2012 Organizing Committee.

Rico Sennrich, Barry Haddow, and Alexandra Birch. 2016. Neural machine translation of rare words with subword units. In Proceedings of the 54th Annual Meeting of the Association for Computational Linguistics (Volume 1: Long Pa-pers), pages 1715–1725. Association for Compu-tational Linguistics.

Yi Mon Shwe Sin and Khin Mar Soe. 2018. Syllable-based myanmar-english neural machine translation. InProceedings of ICCA2018, pages 228–233.

Ilya Sutskever, Oriol Vinyals, and Quoc V. Le. 2014a. Sequence to sequence learning with neu-ral networks. In Proceedings of the 27th In-ternational Conference on Neural Information Processing Systems - Volume 2, NIPS’14, pages 3104–3112, Cambridge, MA, USA. MIT Press.

Ilya Sutskever, Oriol Vinyals, and Quoc V. Le. 2014b. Sequence to sequence learning with neu-ral networks. CoRR, abs/1409.3215.

Ye Kyaw Thu, Finch Andrew, Win Pa Pa, and Sumita Eiichiro. 2016. A large-scale study of statistical machine translation meth-ods for myanmar language. In Proceedings of SNLP2016.

Ye Kyaw Thu, Finch Andrew, Sagisaka Yoshinori, and Sumita Eiichiro. 2013. A study of myanmar word segmentation schemes for statistical ma-chine translation. InProceedings of ICCA2013, pages 167–179.

Figure

Table 1: BLEU scores of Syllable-BPE segmentation with different batch sizes for three NMT models

References

Related documents

A Monthly Double-Blind Peer Reviewed Refereed Open Access International e-Journal - Included in the International Serial Directories. GE- International Journal of Engineering

In a previous study (Fuchs et al., 2011), we showed that in the absence of sensory feedback from legs, the deafferented motor centres can produce fictive rhythmic activity, which

Since we find that prolonged starvation suppresses cryophilic movement, we can make specific predictions regarding the movement of worms on spatial thermal gradients after growing

Purpose: To evaluate diagnostic reliability and to establish optimal scanning techniques of a recently developed Fast Spin-echo MR pulse sequence that allows rapid

A comparison of enzyme activities of the glycolytic pathway, the citric acid cycle and the pentose phosphate shunt measured in homogenates of swimbladder epithelium, rete

Lack of steady-state running has forced the authors of the invertebrate experiments to calculate a value called net cost of transport to compare with data on the minimum cost

For example, if there was an excitatory input arriving from the contralateral antenna in the unit included in the (IL ( + ) , —) group, the unit showed a larger magnitude of

Response* recorded from the right pedal nerve in the whole-animal preparation when the light stimuli were applied to each photoreceptor separately, (o) Dorsal eye illumi- nation;