Low-Resource Machine Translation for Low-Resource Languages: Leveraging Comparable Data, Code-Switching and Compute Resources
←
→
Page content transcription
If your browser does not render page correctly, please read the page content below
Low-Resource Machine Translation for Low-Resource Languages: Leveraging Comparable Data, Code-Switching and Compute Resources Garry Kuwanto∗ Afra Feyza Akyürek∗ Institut Teknologi Bandung Boston University gkuwanto@students.itb.ac.id akyurek@bu.edu Isidora Chara Tourni∗ Siyang Li∗ Derry Wijaya Boston University {isidora, siyangli, wijaya}@bu.edu Abstract With large language models such as BERT and GPT-3 prevalent in NLP, one can achieve good We conduct an empirical study of unsuper- low resource NMT performance by pretraining vised neural machine translation (NMT) for these on a large amount of data: hundreds of mil- arXiv:2103.13272v1 [cs.CL] 24 Mar 2021 truly low resource languages, exploring the case when both parallel training data and lions of sentences including monolingual and paral- compute resource are lacking, reflecting the lel data from related higher resource languages, reality of most of the world’s languages and on large compute resources: hundreds of large the researchers working on these languages. memory GPUs, for a long period of time: weeks We propose a simple and scalable method to (Liu et al., 2020). However, the assumption of improve unsupervised NMT, showing how abundant monolingual data, available parallel data, adding comparable data mined using a bilin- and compute resources, remains unsubstantiated gual dictionary along with modest additional compute resource to train the model can sig- when it comes to NMT for truly low-resource lan- nificantly improve its performance. We also guages. The assumption that parallel data from re- demonstrate how the use of the dictionary to lated higher resource languages is available (Zoph code-switch monolingual data to create more et al., 2016; Nguyen and Chiang, 2017; Dabre et al., comparable data can further improve perfor- 2017; Kocmi and Bojar, 2018), such as Russian for mance. With this weak supervision, our best Kazakh, only applies to some languages, making method achieves BLEU scores that improve the provided approaches inapplicable to hundreds over supervised results for English→Gujarati of left-behind languages (Joshi et al., 2020). (+18.88), English→Kazakh (+5.84), and English→Somali (+1.16), showing the The lack of parallel data motivates unsupervised promise of weakly-supervised NMT for many NMT. However, its best performance for low re- low resource languages with modest compute source languages that are distant from English is resource in the world. To the best of our spectacularly low (BLEU scores of less than 1), knowledge, our work is the first to quantita- giving an impression of its uselessness (Kim et al., tively showcase the impact of different modest 2020; Liu et al., 2020). We show there are compa- compute resource in low resource NMT. rable data that can be easily mined to provide weak 1 Introduction supervision to NMT, and that unsupervised NMT, combined with these comparable data is not useless Approaches in Unsupervised NMT research have for truly low resource languages. Together, they seen significant progress recently, facilitating trans- outperform their separate scores as well as super- lation with no parallel training data, using tech- vised NMT performance from English to these lan- niques such as back-translation and auto-encoding guages, which is promising given the overwhelm- (Lample et al., 2018; Artetxe et al., 2017b; Sen ing need for automatic translation from English to et al., 2019; Artetxe et al., 2019; Sun et al., 2019). other languages, as information in the former is However, most works in this area have focused on ample compared to the latter (Nekoto et al., 2020). simulating low-resource scenarios either for high- In this work, among the low resource languages resource languages (e.g., German), for which un- in MT for which there are test corpora, we focus our supervised methods are redundant, or for medium- attention to three: Gujarati (gu) and Somali (so), resource and similar to English ones (e.g., Roma- the scraping-by languages for which some monolin- nian) rather than truly low-resource languages. gual data are available, and Kazakh (kk) for which ∗ Contributed equally more monolingual data are available (Joshi et al.,
2020). These are distant languages that are diverse for low-resource languages with no dependency on in terms of script, morphological complexity and available parallel data from related languages, and word order from English (en) and for which unsu- quantify the effects of available computational re- pervised NMT has either never been explored (for sources. Finally, we achieve the state-of-the-art re- so) or has been shown to perform poorly i.e., 0.6 sults without parallel data from English to these lan- and 0.8 BLEU for en→gu and en→kk. guages, with improvements up to 21 BLEU points We mine comparable sentences from Gujarati, from previous unsupervised results and up to 18 Somali, and Kazakh Wikipedia, that are linked to BLEU points from supervised results (Kim et al., their corresponding English articles. Wikipedia 2020). We believe that the improvements presented has pages in over 300 languages, both high and here are remarkable, especially considering the fact low resource, allowing our approaches presented that these languages have been treated as “hope- here to scale to other languages. We empirically less cases” of unsupervised NMT in recent works. evaluate various lightweight methods to mine com- While making no language-specific assumptions, parable sentences from these linked pages. We find such as availability of a related high-resource lan- that a simple method of using a bilingual lexicon guage, or parallel data we provide state-of-the-art dictionaries to mine sentences with lexical overlap results, hoping our insights will motivate similar results in comparable data that achieves the best work in other languages. NMT performance in all three languages. We also use the dictionary to code-switch monolingual data 2 Related Work harnessing more comparable data. Despite the rapid progress in language technolo- Unless otherwise indicated, we assume limited gies, research efforts in NLP have only incorpo- access to compute resource and restrict our com- rated about 6% of all 7000 world languages (Joshi pute expenses by the average monthly income for et al., 2020). In their study, the authors develop the language speaking region. Our assumption of a taxonomy of six categories depending on the limited access to compute resource is realistic and data available in each language (labeled and unla- driven by personal experiences and previous works beled), and study extensively their resource dispar- that have observed how indeterminate access to a ities, and their representation in NLP conferences. large number of GPUs is unrealistic beyond large Their analysis highlights that NLP methods and industrial/academic environments (Ahmed and Wa- venues need to further focus on under-explored hed, 2020). We observe that without this constraint, and typologically diverse languages. This moti- adding even a small compute power from 1 GPU vates our choice of languages in this paper since (16GB) to 4 GPUs (32GB) and training for a longer Somali, Gujarati, and Kazakh are low-resource and period of time can already improve NMT perfor- typologically diverse languages that are also under- mance. We note how pivotal, even misguiding explored in unsupervised NMT. when not accounted for, compute resources can It is important to note that aside from the grow- be in achieving compelling results. Furthermore, ing gap in available data, a growing phenomenon in previous works have shown that increasing com- the AI community is the so-called “compute divide” pute resources carries with it a high environmental between large firms, elite and non-elite universities, cost (Strubell et al., 2019; Bender et al., 2021). To caused by the large computational requirements in this end, we echo their suggestions and urge the GPU usage and the researchers’ unequal access to NLP community to factor in the compute resources computing resources (Ahmed and Wahed, 2020; when reporting results. In addition, by using lim- Strubell et al., 2019). We believe that a discussion ited compute resources, we hope to shed light on of compute requirement should always be factored the true state of low-resource NMT when both data in, especially in the study of low resource NMT; and access to compute resource are lacking. since for many of these languages the lack of com- The contributions of our paper are manifold. We pute resource, infrastructure, and time constraints empirically evaluate lightweight methods for min- can hinder communities in low resourced societies ing comparable sentences for training an NMT from working and publishing on these languages; model. We explore different ways of leveraging and can render the use of techniques developed in monolingual data for NMT, using code switch- high resourced societies inapplicable in this low ing. We introduce guidelines for training NMT resource setting (Joshi et al., 2020; Nekoto et al.,
Wikipedia WMT1 (2018/2019) Leipzig Corpora2 (2016) soWaC163 (2016) Total gu 243K 531K 600K - 1.36M kk 1M 7.5M 1M - 9.51M so 32K 123K - 1.83M 1.97M 843K(gu) 517K(gu) 1.36M(gu) en 4.44M(kk) 5.07M(kk) - - 9.51M(kk) 654K(so) 1.32M(so) 1.97M(so) Table 1: Monolingual data sources (with data size in number of sentences) 2020). et al. (2019); Pavlick et al. (2014)) or can be in- For low resource languages where parallel train- duced from sentence alignment of comparable cor- ing data is little to none, unsupervised NMT can pora such as Wikipedia titles (Ramesh and Sankara- play a crucial role. However, previous works have narayanan, 2018; Pourdamghani et al., 2018; Wi- only focused on high-resource and/or similar-to- jaya et al., 2017). English languages. Most recently, several works have questioned the universal usefulness of unsu- 3 Model pervised NMT and showed its poor results for low- Since transformer-based architectures have proven resource languages (Kim et al., 2020; Marchisio successful in MT numerous times in the past (Bar- et al., 2020). They reason that this is because fac- rault et al., 2019), for all of our experiments we use tors that are important for good unsupervised NMT XLM (Conneau and Lample, 2019). We first pre- such as linguistic similarity between source and train a bilingual Language Model (LM) using the target, and domain proximity along with size and Masked Language Model (MLM) objective (De- quality of the monolingual corpora are hard to sat- vlin et al., 2019) on the monolingual corpora of isfy in the case of low resource languages. In this two languages (e.g. Somali and English for en- work we show that we can improve unsupervised so). For both the LM pretraining and NMT model performance by leveraging comparable data. fine-tuning, unless otherwise noted, we follow the There is a large body of work in mining these hyper-parameter settings suggested in the XLM pseudo parallel sentences (Munteanu et al., 2004; repository4 . For every language pair we extract a Munteanu and Marcu, 2006; Zweigenbaum et al., shared 60,000 subword vocabulary using Byte-Pair 2017; Guo et al., 2018; Grover and Mitra, 2017; Encoding (BPE) (Sennrich et al., 2015b). Apart Schwenk, 2018; Hangya et al., 2018; Hangya and from MLM, Conneau and Lample (2019) intro- Fraser, 2019; Wu et al., 2019b); yet most ap- duced the Translation Language Model (TLM) for proaches have not been widely applied in the low- which we use utilize mined comparable data intro- resource scenarios we aim to examine. More- duced in Section 4.2. over, some of the recent works rely on supervised After pretraining the LM, we train a NMT systems trained on parallel data(Schwenk et al., model in an unsupervised or weakly-supervised 2019a,b; Pourdamghani et al., 2019) and/or require (using comparable data) manner. We again follow extensive compute resource for iterative mining of the setup recommended in Conneau and Lample comparable sentences (Tran et al., 2020), hence (2019), where both encoder and decoder are initial- may not be applicable in low resource setting. ized using the same pretrained encoder block. For The most recent unsupervised bitext extraction unsupervised NMT, we use back-translation (BT) method utilizes multilingual contextual embed- and denoising auto-encoding (AE) losses (Lample dings retrieved from mBERT Keung et al. (2020); et al., 2018), and the same monolingual data as in Devlin et al. (2019). Unlike these approaches that LM pretraining, elucidated in Section 4.1. Lastly, require expensive computation of contextual em- for our experiments that leverage mined compara- bedding similarity between sentences, we employ ble data, we follow BT+AE with BT+MT, where a simpler method of using bilingual dictionaries to MT stands for supervised machine translation ob- mine sentences with lexical overlap. Our method jective for which we use the mined data. also does not require a multilingual model con- 1 taining the language but simply an existing dic- 2 http://data.statmt.org/news-crawl/ https://wortschatz.uni-leipzig.de/en/download/ tionary which is available from various sources 3 http://habit-project.eu/wiki/SetOfEthiopianWebCorpora 4 (Wiktionary, Kamholz et al. (2014); Thompson http://github.com/facebookresearch/XLM
DATE FIRST gu (Kawardha is an important town in the Indian state of Chhattisgarh.) en Kawardha is a city and a municipality in Kabirdham district in the Indian state of Chhattisgarh. gu (These matches were played from Monday 5th March to Friday 9th March.) en The matches were played from Monday 5 March until Friday 9 March. (PROJECTED ) DICT gu (Former President of Gujarati Sahitya Parishad, Dr.) en He also served as the president of Gujarati Sahitya Parishad. LENGTH kk (Otrau is a municipality in Hesse, Germany.) en Ottrau is a community in the Schwalm-Eder-Kreis in Hesse, Germany. LINK kk (The largest islands: Saaremaa, Hiiumaa, Muhu, Vorsmi, etc.) en Two of them are large enough to constitute separate counties: Saaremaa and Hiiumaa. ( PROJECTED ) DICT kk (Abdur-Rahman ibn Auf is another name for ’Abd al-Rahman ibn Ghauf.) en His name has also been transliterated as Abdel Rahman Ibn Auf. Table 2: Sample comparable sentence pairs in Gujarati and Kazakh. Translations in parentheses are obtained using Google Translate which may be imperfect. 4 Dataset source to English. Word overlap is defined as the ratio of matching words divided by the 4.1 Monolingual Data average number of words in the sentences. The languages we study are Gujarati, Kazakh and • MERGED(X): This category includes all of Somali. They are spoken by 55M, 22M and 16M the above with the corresponding DICT(X). speakers worldwide, respectively, and are distant In order to curate dict(X), we first sample 10k from English, in terms of writing scripts and al- most frequent words from Wikipedia dumps of gu, phabets. Additionally, these languages have few kk and so and obtain translations of these words parallel but some comparable and/or monolingual from Amazon MTurk dictionary (Pavlick et al., data available, which makes them ideal candidates 2014), dict(MTurk), and using Google Translate for our low-resource unsupervised NMT research. API, dict(Google). For Kazakh, because MTurk Our monolingual data (see Table 1) are carefully provides very few translations, we add entries chosen from the same domain of news data and from NorthEuraLex bilingual lexicon (Dellert et al., from similar time periods (late 2010s) to mitigate 2020). We also create a higher coverage dictionary domain discrepancy between source and target lan- based on dict(Google) by first training monolin- guages as per previous research (Kim et al., 2020). gual word embeddings for each language using fast- For English data, we use Wikipedia pages linked Text’s skipgram model (Bojanowski et al., 2017) to gu, kk and so pages, respectively, combined with on the monolingual data of the language (Table the randomly downsampled WMT NewsCrawl cor- 1); then learn a linear mapping between the source pus so that target and source data are equal in size. and target word embeddings, using dict(Google) as seed translations with MUSE (Conneau et al., 4.2 Comparable Data Mining 2017). Based on this mapping, we find translations Our comparable data which is delineated in Table 3 of up to 200k most frequent words from each lan- comes from the linked Wikipedia pages in differ- guage monolingual data by projecting the source ent languages obtained using the langlinks from embedding (gu, kk, and so) to the target language Wikimedia dumps and six lightweight methods to embedding (en) and taking as translation the target extract comparable sentence pairs: word that has the highest cosine similarity based • DATE: Pairs with at least one identical date Cross-Domain Similarity Local Scaling (CSLS) item (e.g. December). metric (Conneau et al., 2017), which adjusts cosine • FIRST: The first sentence of both articles. similarity values of a word based on the density • LENGTH: Pairs in the first paragraph whose of the area where its embedding lies. We call this lengths are within ±1 of each other. higher coverage dictionary dict(Projected). • LINK: Pairs with at least two identical URLs. 4.3 Data Augmentation Using • NUMBER: Pairs with at least two identical Code-Switching numbers. • DICT(X): Pairs having at least 20% word To create more comparable data for training, simi- overlap after word-by-word translation from lar to the idea of back-translation to automatically
Method # sentences BLEU gu kk so en-gu gu-en en-kk kk-en en-so so-en DATE 3758 3079 618 0.11 1.47 0.09 0.00 0.00 0.00 DICT ( MTURK ) 44492 5832 19754 6.17 1.82 0.17 0.30 0.38 0.00 DICT ( GOOGLE ) 65974 53211 9927 6.39 1.98 0.55 1.34 0.73 0.95 DICT ( PROJECTED ) 71424 58961 9880 5.86 1.79 0.66 1.30 1.17 1.13 FIRST 8671 103002 4118 0.00 0.00 0.00 0.23 0.52 1.03 LENGTH 9487 52553 1870 1.11 0.21 0.00 0.17 0.25 0.00 LINK 3795 18294 683 1.27 0.00 0.21 0.31 0.00 0.00 NUMBER 926 2604 254 0.85 0.00 0.00 0.00 0.13 0.17 MERGED ( MTURK ) 64904 156871 25101 6.08 1.90 0.32 0.48 0.00 0.53 MERGED ( GOOGLE ) 85032 196382 15361 6.73 2.14 0.55 1.28 1.05 0.76 MERGED ( PROJECTED ) 90229 202147 15377 7.15 2.56 0.69 1.19 0.52 1.03 Table 3: Different methods of mining comparable sentences used, number of comparable sentence pairs extracted with each technique and BLEU scores of models trained using only MT objective with these pairs. There is no LM pretraining. Best results are highlighted. translate the monolingual text to and from the tar- for which we use the mined data as before. get language (Sennrich et al., 2015a), we use our dict(Projected) dictionary to automatically trans- 5 Experiments and Results late monolingual Wikipedia sentences to and from 5.1 Experimental Setup the target language, creating real (i.e., monolin- gual) source to pseudo (i.e., dictionary-translated) We use WMT 2019 news dataset for evaluation of target sentence pairs and pseudo source to real tar- Gujarati ↔ English and Kazakh ↔ English. We get sentence pairs, respectively. Since the coverage use DARPA’s LORELEI (Tracey et al., 2019) vali- of our dictionary is limited, there are words in the dation and test data sets for Somali. monolingual sentence that it cannot translate. In Because we are interested in simulating a lim- those cases, we leave the words as they are, result- ited compute resource setting, instead of training ing in an imperfect, code-switched translation of our language models for prolonged times, we take the monolingual sentence. Previous works on text into account the average monthly income1 of each classification has shown that fine-tuning multilin- region the language is primarily spoken in (Gujarat gual models such as multilingual BERT on code- for Gujarati, Kazakhstan for Kazakh and Somalia switched data can improve performance on few- for Somali) and use Amazon AWS EC2 rate2 as an shot and zero-shot classification tasks (Akyürek estimate on how long we should train. Our calcula- et al., 2020; Qin et al., 2020) ranging from frame tion yields 40, 60 and 72 hours of training time on classification (Liu et al., 2019a) to natural language 1 GPU (16GB) for Gujarati, Somali and Kazakh, inference (Conneau et al., 2018), sentiment clas- respectively. While we train our language models sification (Barnes et al., 2018), document classi- only up to this calculated time, we do not specifi- fication (Schwenk and Li, 2018), dialogue state cally enforce training time limits for NMT training, tracking (Mrkšić et al., 2017), and spoken language which stops only if BLEU score on validation does understanding (Schuster et al., 2019). We include not improve after 10 epochs. We observe that NMT code-switched sentences that have at least 20% of training lasts less than 24 hours in average. words translated from their original sentences to augment our training. Despite its imperfect nature, 5.2 Which type of comparable data is the we believe that such (code-switched, real) sentence best? pairs can provide some weak supervision to the In order to empirically evaluate which method pro- unsupervised NMT model to learn to translate. In vides the best quality comparable data, we train our experiments that leverage code-switched sen- NMTs from a randomly initialized LM using only tences, we follow the unsupervised NMT training the MT objective, hence no expensive pre-training step i.e., BT+AE with BT+MTcs and then BT+MT, on monolingual data. This is an efficient way to where MTcs stands for supervised MT objective, see how the comparable sentences fare when used for which we use the (code-switched, real) sentence 1 https://www.numbeo.com/cost-of-living/ pairs, and MT stands for supervised MT objective 2 https://calculator.aws/#/
in other training scenarios. In Table 3 we see the 5.4 Accounting for Compute Resources numbers of sentence pairs mined using a particular Previous research demonstrates that LM pretrain- method, and the BLEU scores obtained using only ing significantly improves performance in many these extracted corpora in NMT. We observe that downstream natural language processing tasks, in- for Kazakh-English and Somali-English, sentence cluding NMT (Lample et al., 2018). Moreover, pairs from the dict(Projected) yield the best BLEU LMs benefit from large batch sizes during pretrain- score when both directions i.e., X-en and en-X are ing (Liu et al., 2019b). In this section, we attempt considered together. Moreover, because all types of to quantify how increased availability of compute sentence pairs in English-Gujarati, except (perhaps resources (practically enabling large batch sizes) af- surprisingly) FIRST are generally better than other fect the language model perplexity as well as trans- language pairs, merging all of the mined corpora lation performance, even when data is scarce. Note provides a better result. that using gradient accumulation one can mimic larger number of GPUs. However, in this case, 5.3 Leveraging Comparable Data to Boost replicating a setup of four GPUs with only one can Unsupervised MT quickly become prohibitive time-wise, especially considering the already prolonged times required to Kim et al. (2020) pinpoint the impracticality of train transformer-based LMs (Devlin et al., 2019). unsupervised NMT due to the large monolingual In Table 5 we provide LM perplexities for differ- corpora required and the difficulty of satisfying ent numbers of GPUs for both Kazakh and Somali, other, domain and linguistic similarity constraints and BLEU scores in Table 6. We use NVIDIA for low-resource languages such as Kazakh and V100 32GB GPUs, setting the batch size to 64 Gujarati. As seen in Table Table 4, they report re- for LM pretraining, and the tokens per batch to 3k markably low-scores for these languages in both for MT fine-tuning, per GPU. We keep all other directions, and conclude that only(!) 50k paired parameters the same for simplicity. The results sup- sequences will suffice to surpass the performance port the hypothesis that enhanced compute resource of unsupervised NMT. We find the call to create bears significant potential to boost LM quality. We 50k parallel sentences for languages which do not observe that the perplexity consistently drops as even have sizable annotated data for testing to be the number of GPUs increases. In our experiments beyond reach in the near future. Therefore, in this we found a strong correlation between translation work, we show that by leveraging inexpensive re- performance and a better LM pretraining. sources such as lexical translations, it is possible to Notably, without any hyper-parameter tuning achieve satisfactory results which are on par, if not other than increasing the number of GPUs, the un- better, than supervised NMT. supervised translation scores improve consistently In Table 4, we provide the BLEU scores using in general, if not dramatically (Table 6). For So- the best comparable data (from Table 3) to train mali, BLEU scores almost double in both en-so and weakly-supervised NMT models. We observe sig- so-en directions, rising from 8.35 to 14.76 for en-so nificant improvements in all three languages and and from 8.02 to 14.83 for so-en. Kazakh proved to both directions. In addition, in Table 4 we also pro- be a more challenging case for low-resource NMT, vide the BLEU scores for the TLM model, which nonetheless, simply utilizing more resources even makes use of comparable data from the beginning. for Kazakh results in consistent improvements in While the TLM model is still much better than us- Table 6 in any translation direction. ing no mined sequences, it is not as good as adding For weakly-supervised NMT, MT objective uses comparable data after training an unsupervised comparable data mined with dict(Projected). Ex- NMT, especially for Kazakh and Somali models. cept for one case, we observe the same pattern even This signals that adding comparable data, which after the MT step in Table 6 where the scores are are not of great quality, too early into training may already significantly higher compared to the unsu- not be the optimal practice. Letting the models first pervised case. We think the drop in en-kk direction learn from monolingual data for simpler language in the 4 GPUs case could be remedied with adjust- features and possibly grammar, and then adding ing the learning rate (Puri et al., 2018), which we comparable data later on is potentially a more effi- avoided for easy interpretation of the results. In cient way to utilize mined data. these experiments, we simply highlight how a priv-
Method en-gu gu-en en-kk kk-en en-so so-en Other methods Multilingual, Supervised, Google Translate 31.4 26.2 23.1 28.9 22.7 27.7 Multilingual, Supervised, mBART252 0.1 0.3 2.5 7.4 - - Bilingual, Unsupervised1 0.6∗ 0.6∗ 0.8∗ 2.0∗ - - Bilingual, Supervised 3.5∗1 9.9∗1 2.4∗1 10.3∗1 - 25.363 Bilingual, Semi-supervised1 4.0∗ 14.2∗ 3.1∗ 12.5∗ - - Ours Bilingual, Unsupervised (MLM + BT + AE) 0.44 0.31 1.11 1.60 8.53 7.99 Bilingual, Weakly-supervised (MLM + TLM) + (BT + MT) 18.05 14.51 4.06 5.65 9.06 10.33 Bilingual, Weakly-supervised MLM + (BT + AE) 0.44 0.31 1.11 1.6 8.53 7.99 (BT + MT) 18.13 14.36 5.86 8.02 14.92 14.31 4 GPUs 22.38 21.04 7.26 10.63 16.96 15.83 Table 4: BLEU scores for previous supervised and unsupervised results from 1 Kim et al. (2020) (∗ report Sacre- BLEU), 2 Liu et al. (2020), and 3 Liu and Kirchhoff (2018) and our weakly-supervised models which leverage comparable data mined using lexical translations. Test and validation sets are from WMT19 for Gujarati and Kazakh and from Tracey et al. (2019) for Somali. MLM, AE, BT and MT stand for Masked Language Model, Auto-Encoding loss, Back Translation loss and Machine Translation loss, respectively. Bolded means best score. MT and TLM objectives use the best comparable data as paired data according to Table 3. Our models, except for the last row, use a single 32GB GPU. Unlike previous results, we use the standard XLM parameters and do not conduct extensive hyperparameter tuning, which can further improve our performance. For completeness sake, Lakew et al. (2020) reports supervised results for Somali on a pooled test set from sources that doesn’t include the test corpus studied here. They achieve 9.16 and 13.38 for en-so and so-en, respectively. GPUs en-kk en-so 2017; Wu et al., 2019a; Song et al., 2019; Pour- en kk en so damghani et al., 2019), very few works studied the 1 15.17 15.37 16.65 10.41 low-resource languages we examine in this paper. 2 14.28 14.46 14.62 9.07 In Table 4, we compare the best results we achieved 4 12.70 12.16 12.44 7.78 in this paper to the previous works (Liu and Kirch- Table 5: Test set perplexities of the bilingual LMs hoff, 2018; Kim et al., 2020), breaking down the trained using different number of GPUs while every- specific components that contributed to the results, thing else is fixed. Best results are highlighted. including the compute power. Except for two cases, our comparable-data-powered model remarkably GPUs en-kk kk-en en-so so-en Unsupervised NMT: MLM + (BT+AE) outperforms previous results, including supervised 1 0.58 1.32 8.35 8.02 ones which utilize tens of thousands of parallel se- 2 1.02 1.73 12.05 11.35 quences, if not more. Improvements gained in this 4 2.91 3.86 14.76 14.83 Weakly-supervised NMT: MLM + (BT+AE) + (BT+MT) paper over the latest reported unsupervised method 1 7.30 10.23 14.79 14.93 (Kim et al., 2020) range from around at least 6.5 2 8.24 10.60 16.11 15.72 BLEU in en-kk to 21.8 in en-gu! Moreover, it is 4 7.26 10.63 16.96 15.83 critical to note that we provide the first unsuper- Table 6: BLEU scores using different number of 32GB vised translation benchmark for English translation GPUs. MT objective uses the best comparable data ac- to and from Somali, as, to our knowledge, all re- cording to Table 3. For both languages best compara- sults reported in the literature so far conduct MT in ble data was yielded using dict(Projected). Tokens per a supervised way. batch set at 3k which is different than our basic 1 GPU (16GB) setup. Best results are highlighted. 5.6 When life hits you hard, hit back harder! ileged access to compute resources can drastically Among the three languages, Kazakh suffers the affect translation performance without resorting most from relatively poor crosslinguality, which is to much engineering. Likewise, as LM pretrain- reflected in Table 4. We believe that this may be ing has become a common practice for many lan- due to its agglutinative nature, where each of its guage tasks (Yang et al., 2019; Bao et al., 2019), we root words can produce hundreds or thousands of strongly believe that accounting for the compute variant word forms (Tolegen et al., 2020). Hence, resources is essential when creating benchmarks. Kazakh may require many more training exam- ples to generalize compared to other, less morpho- 5.5 Benchmark Results logically complex, languages. We observe that Surprisingly, regardless of unsupervised MT being adding more comparable data in the form of code- in the spotlight for a few years now (Conneau et al., switched sentences as discussed in Section 4.3
Somali Sentence Warbaahinta Ruushka ayaa sheegtay in diyaarad ay leedahay shirkadda diyaaradaha Malaysia oo ka duushay Amsterdam kuna socotay Kuala Lumpur ay ku burburtay bariga Ukraine. Ours MLM + (BT + AE) The Russian government said that a company from Malaysia was on board at Amsterdam kuna dhacday Kuala Lumpur in a burburtay in bariga Ukraine. (BT + MT) Russian media said a Malaysia Airlines jet that landed in Amsterdam at Kuala Lumpur crashed in eastern Ukraine. 4 GPUs Russian media said a Malaysian airline flight from Amsterdam to Kuala Lumpur crashed in eastern Ukraine. Reference Translation Russian media reported that a Malaysian Airlines plane flying from Amsterdam to Kuala Lumpur has crashed in Eastern Ukraine. Table 7: Sample translations from Somali → English by different models. en-kk kk-en even without additional compute power. Further, Unsupervised 1.11 1.6 in Table 8 we introduce code-switched training Weakly-supervised 5.86 8.02 even before comparable data is used. In addition Code-switched 7.26 9.64 to back-translation, the code-switched words act Table 8: BLEU scores for en-kk and kk-en translation as anchors that help align the embedding space from different NMT models, using one 16GB GPU. of the two languages, yielding a state-of-the-art can help, adding a further 1.4 and 1.62 BLEU performance for Kazakh without any parallel data points for translation to and from Kazakh under and with mediocre compute resources. Nonethe- the constrained compute resource setting (Table less, since there is still room for improvement, we 8). In the future, we believe that for morpho- deem the need to create more parallel resources for logically complex languages such as Kazakh, we languages such as Kazakh (Joshi et al., 2020) as need to conduct morphological segmentation pre- well as other low-resource and/or highly complex processing (Sánchez-Cartagena et al., 2019) to seg- languages with millions of speakers crucial. ment words into subword units, which has been Unlike the canonical languages studied in unsu- shown to outperform BPE for highly inflected lan- pervised NMT i.e. German, French, and Romanian, guages (Sánchez-Cartagena and Toral, 2016; Huck the core importance of our work lies in that the lan- et al., 2017; Sánchez-Cartagena et al., 2018). guages we examine are truly low-resource as well as linguistically, geographically and morpholog- 6 Discussion and Future Work ically diverse from English. Given that in none of our experiments do we assume a related high- The different scenarios studied have various effects resource language to aid in translation, or use any when it comes to the quality of the output trans- parallel sequences, we set the ground for similar lations. Unlike Somali, Gujarati does not benefit analyses and extension of our approaches to other much from unsupervised training, potentially due low-resource languages. We also show that trans- to small monolingual data along with a different lation scores can almost be doubled by enhancing script. Nevertheless, mined comparable data for the compute power used in the experiments. Hence, Gujarati is of high quality (see Table 3). Hence, we urge the community to factor in this significant adding these to BT+AE model results in BLEU parameter when reporting results. scores of 18.13 and 14.36, which is further boosted with more computational power. Moreover, our So- It is also worth noting that, for some languages, mali model using mined comparable data surpasses even lexical translations may be difficult to ob- the supervised score in en-so direction. Table 7 tain. Therefore, future work can explore using lists sample translations for Somali using different unsupervised methods to obtain lexical translations models, showing how translation gets increasingly in order to yield comparable sequences for MT. better with addition of mined data and compute re- Several previous works study harnessing word vec- source. In general, we observe that Kazakh BLEU tors trained on monolingual corpora in unsuper- percentages are lower compared to those of Somali vised manner (Conneau et al., 2017; Irvine and and Gujarati. We believe this is due to the mor- Callison-Burch, 2017), with small bilingual seeds phologically more complex structure of Kazakh (Artetxe et al., 2017a), or with other modalities (Briakou and Carpuat, 2019). Yet, the introduc- or knowledge resources (Hewitt et al., 2018; Wi- tion of comparable data substantially increases the jaya et al., 2017). Others have explored extracting BLEU scores for Kazakh, by 4.7 and 6.4 points, lexical translations from word alignment of compa-
rable corpora such as Bible, Universal Declaration References of Human Rights, or Wikipedia Titles (Ramesh Nur Ahmed and Muntasir Wahed. 2020. The de- and Sankaranarayanan, 2018; Pourdamghani et al., democratization of ai: Deep learning and the com- 2018) as Wikipedia is a rich lexical semantic re- pute divide in artificial intelligence research. arXiv source (Zesch et al., 2007; Wijaya et al., 2015). We preprint arXiv:2010.15581. are also interested in exploring other sources for Afra Feyza Akyürek, Lei Guo, Randa Elanwar, Prakash mining comparable sentences outside of Wikipedia Ishwar, Margrit Betke, and Derry Tanti Wijaya. such as international news sites (Voice of America, 2020. Multi-label and multilingual news framing etc.) analysis. In Proceedings of the 58th Annual Meet- ing of the Association for Computational Linguistics, pages 8614–8624. 7 Conclusion Mikel Artetxe, Gorka Labaka, and Eneko Agirre. In this work, we explore ways to mine, create, and 2017a. Learning bilingual word embeddings with utilize comparable data to improve performance (almost) no bilingual data. In Proceedings of the of unsupervised NMT, which can scale to many 55th Annual Meeting of the Association for Compu- low resource languages; while shedding light at tational Linguistics (Volume 1: Long Papers), pages 451–462, Vancouver, Canada. Association for Com- another relevant but often overlooked factor that putational Linguistics. is compute resource. Without using any parallel data aside from bilingual lexicons and without as- Mikel Artetxe, Gorka Labaka, and Eneko Agirre. 2019. An effective approach to unsupervised machine suming any availability of similar, higher resource translation. arXiv preprint arXiv:1902.01313. languages, we lift up NMT results for truly low- resource languages from disconcerting scores (i.e. Mikel Artetxe, Gorka Labaka, Eneko Agirre, and fractional BLEU points) to two-digit BLEU scores. Kyunghyun Cho. 2017b. Unsupervised neural ma- chine translation. arXiv preprint arXiv:1710.11041. We hope this work will set the ground for applica- tion of our methods to other languages, as well as Siqi Bao, Huang He, Fan Wang, Hua Wu, and Haifeng further exploration of similar approaches. Wang. 2019. Plato: Pre-trained dialogue generation model with discrete latent variable. arXiv preprint arXiv:1910.07931. Jeremy Barnes, Roman Klinger, and Sabine Schulte im Walde. 2018. Bilingual sentiment embeddings: Joint projection of sentiment across languages. arXiv preprint arXiv:1805.09016. Loïc Barrault, Ondřej Bojar, Marta R Costa-Jussà, Christian Federmann, Mark Fishel, Yvette Gra- ham, Barry Haddow, Matthias Huck, Philipp Koehn, Shervin Malmasi, et al. 2019. Findings of the 2019 conference on machine translation (wmt19). In Proceedings of the Fourth Conference on Machine Translation (Volume 2: Shared Task Papers, Day 1), pages 1–61. Emily M. Bender, Timnit Gebru, Angelina McMillan- Major, and Shmargaret Shmitchell. 2021. On the dangers of stochastic parrots: Can language models be too big? In Proceedings of the 2021 ACM Confer- ence on Fairness, Accountability, and Transparency, FAccT ’21, page 610–623, New York, NY, USA. As- sociation for Computing Machinery. Piotr Bojanowski, Edouard Grave, Armand Joulin, and Tomas Mikolov. 2017. Enriching word vectors with subword information. Transactions of the Associa- tion for Computational Linguistics, 5:135–146. Eleftheria Briakou and Marine Carpuat. 2019. The university of maryland’s kazakh-english neural ma- chine translation system at wmt19. In Proceedings of the Fourth Conference on Machine Translation
(Volume 2: Shared Task Papers, Day 1), pages 134– John Hewitt, Daphne Ippolito, Brendan Callahan, Reno 140. Kriz, Derry Tanti Wijaya, and Chris Callison-Burch. 2018. Learning translations via images with a mas- Alexis Conneau and Guillaume Lample. 2019. Cross- sively multilingual image dataset. In Proceedings lingual language model pretraining. In Advances of the 56th Annual Meeting of the Association for in Neural Information Processing Systems, pages Computational Linguistics (Volume 1: Long Papers), 7059–7069. pages 2566–2576. Alexis Conneau, Guillaume Lample, Marc’Aurelio Matthias Huck, Simon Riess, and Alexander Fraser. Ranzato, Ludovic Denoyer, and Hervé Jégou. 2017. 2017. Target-side word segmentation strategies for Word translation without parallel data. arXiv neural machine translation. In Proceedings of the preprint arXiv:1710.04087. Second Conference on Machine Translation, pages 56–67. Alexis Conneau, Guillaume Lample, Ruty Rinott, Ad- ina Williams, Samuel R Bowman, Holger Schwenk, Ann Irvine and Chris Callison-Burch. 2017. A com- and Veselin Stoyanov. 2018. Xnli: Evaluating cross- prehensive analysis of bilingual lexicon induction. lingual sentence representations. arXiv preprint Computational Linguistics, 43(2):273–310. arXiv:1809.05053. Pratik Joshi, Sebastin Santy, Amar Budhiraja, Kalika Raj Dabre, Tetsuji Nakagawa, and Hideto Kazawa. Bali, and Monojit Choudhury. 2020. The state and 2017. An empirical study of language relatedness fate of linguistic diversity and inclusion in the NLP for transfer learning in neural machine translation. world. In Proceedings of the 58th Annual Meet- In Proceedings of the 31st Pacific Asia Conference ing of the Association for Computational Linguistics, on Language, Information and Computation, pages pages 6282–6293, Online. Association for Computa- 282–286. tional Linguistics. Johannes Dellert, Thora Daneyko, Alla Münch, Alina Ladygina, Armin Buch, Natalie Clarius, Ilja Grigor- David Kamholz, Jonathan Pool, and Susan M Colow- jew, Mohamed Balabel, Hizniye Isabella Boga, Za- ick. 2014. Panlex: Building a resource for panlin- lina Baysarova, et al. 2020. Northeuralex: A wide- gual lexical translation. In LREC, pages 3145–3150. coverage lexical database of northern eurasia. Lan- guage resources and evaluation, 54(1):273–301. Phillip Keung, Julian Salazar, Yichao Lu, and Noah A Smith. 2020. Unsupervised bitext mining and trans- Jacob Devlin, Ming-Wei Chang, Kenton Lee, and lation via self-trained contextual embeddings. arXiv Kristina Toutanova. 2019. BERT: Pre-training of preprint arXiv:2010.07761. deep bidirectional transformers for language under- standing. In Proceedings of the 2019 Conference Yunsu Kim, Miguel Graça, and Hermann Ney. 2020. of the North American Chapter of the Association When and why is unsupervised neural machine trans- for Computational Linguistics: Human Language lation useless? arXiv preprint arXiv:2004.10581. Technologies, Volume 1 (Long and Short Papers), pages 4171–4186, Minneapolis, Minnesota. Associ- Tom Kocmi and Ondřej Bojar. 2018. Trivial transfer ation for Computational Linguistics. learning for low-resource neural machine translation. arXiv preprint arXiv:1809.00357. Jeenu Grover and Pabitra Mitra. 2017. Bilingual word embeddings with bucketed cnn for parallel sentence Surafel M Lakew, Matteo Negri, and Marco Turchi. extraction. In Proceedings of ACL 2017, Student Re- 2020. Low resource neural machine translation: search Workshop, pages 11–16. A benchmark for five african languages. arXiv preprint arXiv:2003.14402. Mandy Guo, Qinlan Shen, Yinfei Yang, Heming Ge, Daniel Cer, Gustavo Hernandez Abrego, Keith Guillaume Lample, Alexis Conneau, Ludovic Denoyer, Stevens, Noah Constant, Yun-Hsuan Sung, Brian and Marc’Aurelio Ranzato. 2018. Unsupervised ma- Strope, et al. 2018. Effective parallel corpus min- chine translation using monolingual corpora only. ing using bilingual sentence embeddings. arXiv In International Conference on Learning Represen- preprint arXiv:1807.11906. tations. Viktor Hangya, Fabienne Braune, Yuliya Kala- Angli Liu and Katrin Kirchhoff. 2018. Context models souskaya, and Alexander Fraser. 2018. Unsuper- for oov word translation in low-resource languages. vised parallel sentence extraction from comparable arXiv preprint arXiv:1801.08660. corpora. In Proc. IWSLT. Siyi Liu, Lei Guo, Kate Mays, Margrit Betke, and Viktor Hangya and Alexander Fraser. 2019. Unsuper- Derry Tanti Wijaya. 2019a. Detecting frames in vised parallel sentence extraction with parallel seg- news headlines and its application to analyzing news ment detection helps machine translation. In Pro- framing trends surrounding us gun violence. In Pro- ceedings of the 57th Annual Meeting of the Asso- ceedings of the 23rd Conference on Computational ciation for Computational Linguistics, pages 1224– Natural Language Learning (CoNLL), pages 504– 1234. 514.
Yinhan Liu, Jiatao Gu, Naman Goyal, Xian Li, Sergey Nima Pourdamghani, Marjan Ghazvininejad, and Edunov, Marjan Ghazvininejad, Mike Lewis, and Kevin Knight. 2018. Using word vectors to improve Luke Zettlemoyer. 2020. Multilingual denoising word alignments for low resource machine transla- pre-training for neural machine translation. Transac- tion. In Proceedings of the 2018 Conference of the tions of the Association for Computational Linguis- North American Chapter of the Association for Com- tics, 8:726–742. putational Linguistics: Human Language Technolo- gies, Volume 2 (Short Papers), pages 524–528. Yinhan Liu, Myle Ott, Naman Goyal, Jingfei Du, Man- dar Joshi, Danqi Chen, Omer Levy, Mike Lewis, Raul Puri, Robert Kirby, Nikolai Yakovenko, and Luke Zettlemoyer, and Veselin Stoyanov. 2019b. Bryan Catanzaro. 2018. Large scale language mod- Roberta: A robustly optimized bert pretraining ap- eling: Converging on 40gb of text in four hours. proach. arXiv preprint arXiv:1907.11692. In 2018 30th International Symposium on Com- puter Architecture and High Performance Comput- Kelly Marchisio, Kevin Duh, and Philipp Koehn. 2020. ing (SBAC-PAD), pages 290–297. IEEE. When does unsupervised machine translation work? arXiv preprint arXiv:2004.05516. Libo Qin, Minheng Ni, Yue Zhang, and Wanxiang Che. 2020. Cosda-ml: Multi-lingual code-switching data Nikola Mrkšić, Ivan Vulić, Diarmuid Ó Séaghdha, Ira augmentation for zero-shot cross-lingual nlp. arXiv Leviant, Roi Reichart, Milica Gašić, Anna Korho- preprint arXiv:2006.06402. nen, and Steve Young. 2017. Semantic special- Sree Harsha Ramesh and Krishna Prasad Sankara- ization of distributional word vector spaces using narayanan. 2018. Neural machine translation for monolingual and cross-lingual constraints. Transac- low resource languages using bilingual lexicon in- tions of the association for Computational Linguis- duced from comparable corpora. arXiv preprint tics, 5:309–324. arXiv:1806.09652. Dragos Stefan Munteanu, Alexander Fraser, and Daniel Víctor M Sánchez-Cartagena, Marta Bañón, Ser- Marcu. 2004. Improved machine translation perfor- gio Ortiz Rojas, and Gema Ramírez-Sánchez. 2018. mance via parallel sentence extraction from compa- Prompsit’s submission to wmt 2018 parallel corpus rable corpora. In Proceedings of the Human Lan- filtering shared task. In Proceedings of the Third guage Technology Conference of the North Ameri- Conference on Machine Translation: Shared Task can Chapter of the Association for Computational Papers, pages 955–962. Linguistics: HLT-NAACL 2004, pages 265–272. Víctor M Sánchez-Cartagena, Juan Antonio Pérez- Dragos Stefan Munteanu and Daniel Marcu. 2006. Ex- Ortiz, and Felipe Sánchez-Martínez. 2019. The tracting parallel sub-sentential fragments from non- universitat d’alacant submissions to the english-to- parallel corpora. In Proceedings of the 21st Interna- kazakh news translation task at wmt 2019. In tional Conference on Computational Linguistics and Proceedings of the Fourth Conference on Machine 44th Annual Meeting of the Association for Compu- Translation (Volume 2: Shared Task Papers, Day 1), tational Linguistics, pages 81–88. pages 356–363. Wilhelmina Nekoto, Vukosi Marivate, Tshinondiwa Víctor M Sánchez-Cartagena and Antonio Toral. 2016. Matsila, Timi Fasubaa, Tajudeen Kolawole, Taiwo Abu-matran at wmt 2016 translation task: Deep Fagbohungbe, Solomon Oluwole Akinola, Sham- learning, morphological segmentation and tuning suddee Hassan Muhammad, Salomon Kabongo, on character sequences. In Proceedings of the Salomey Osei, et al. 2020. Participatory re- First Conference on Machine Translation: Volume search for low-resourced machine translation: A 2, Shared Task Papers, pages 362–370. case study in african languages. arXiv preprint Tal Schuster, Ori Ram, Regina Barzilay, and Amir arXiv:2010.02353. Globerson. 2019. Cross-lingual alignment of con- Toan Q Nguyen and David Chiang. 2017. Trans- textual word embeddings, with applications to fer learning across low-resource, related languages zero-shot dependency parsing. arXiv preprint for neural machine translation. arXiv preprint arXiv:1902.09492. arXiv:1708.09803. Holger Schwenk. 2018. Filtering and mining parallel data in a joint multilingual space. arXiv preprint Ellie Pavlick, Matt Post, Ann Irvine, Dmitry Kachaev, arXiv:1805.09822. and Chris Callison-Burch. 2014. The language de- mographics of amazon mechanical turk. Transac- Holger Schwenk, Vishrav Chaudhary, Shuo Sun, tions of the Association for Computational Linguis- Hongyu Gong, and Francisco Guzmán. 2019a. tics, 2:79–92. Wikimatrix: Mining 135m parallel sentences in 1620 language pairs from wikipedia. arXiv preprint Nima Pourdamghani, Nada Aldarrab, Marjan arXiv:1907.05791. Ghazvininejad, Kevin Knight, and Jonathan May. 2019. Translating translationese: A two-step Holger Schwenk and Xian Li. 2018. A corpus for mul- approach to unsupervised machine translation. tilingual document classification in eight languages. arXiv preprint arXiv:1906.05683. arXiv preprint arXiv:1805.09821.
Holger Schwenk, Guillaume Wenzek, Sergey Edunov, Chau Tran, Yuqing Tang, Xian Li, and Jiatao Gu. 2020. Edouard Grave, and Armand Joulin. 2019b. Cc- Cross-lingual retrieval for iterative self-supervised matrix: Mining billions of high-quality paral- training. Advances in Neural Information Process- lel sentences on the web. arXiv preprint ing Systems, 33. arXiv:1911.04944. Derry Tanti Wijaya, Brendan Callahan, John Hewitt, Sukanta Sen, Kamal Kumar Gupta, Asif Ekbal, and Jie Gao, Xiao Ling, Marianna Apidianaki, and Chris Pushpak Bhattacharyya. 2019. Multilingual unsu- Callison-Burch. 2017. Learning translations via ma- pervised nmt using shared encoder and language- trix completion. In Proceedings of the 2017 Con- specific decoders. In Proceedings of the 57th An- ference on Empirical Methods in Natural Language nual Meeting of the Association for Computational Processing, pages 1452–1463. Linguistics, pages 3083–3089. Derry Tanti Wijaya, Ndapandula Nakashole, and Tom Mitchell. 2015. “a spousal relation begins with Rico Sennrich, Barry Haddow, and Alexandra Birch. a deletion of engage and ends with an addition 2015a. Improving neural machine translation of divorce”: Learning state changing verbs from models with monolingual data. arXiv preprint wikipedia revision history. In Proceedings of the arXiv:1511.06709. 2015 conference on empirical methods in natural language processing, pages 518–523. Rico Sennrich, Barry Haddow, and Alexandra Birch. 2015b. Neural machine translation of rare Jiawei Wu, Xin Wang, and William Yang Wang. 2019a. words with subword units. arXiv preprint Extract and edit: An alternative to back-translation arXiv:1508.07909. for unsupervised neural machine translation. arXiv preprint arXiv:1904.02331. Kaitao Song, Xu Tan, Tao Qin, Jianfeng Lu, and Tie- Yan Liu. 2019. Mass: Masked sequence to sequence Lijun Wu, Jinhua Zhu, Di He, Fei Gao, QIN Tao, Jian- pre-training for language generation. arXiv preprint huang Lai, and Tie-Yan Liu. 2019b. Machine trans- arXiv:1905.02450. lation with weakly paired documents. In Proceed- ings of the 2019 Conference on Empirical Methods Emma Strubell, Ananya Ganesh, and Andrew McCal- in Natural Language Processing and the 9th Inter- lum. 2019. Energy and policy considerations for national Joint Conference on Natural Language Pro- deep learning in NLP. In Proceedings of the 57th cessing (EMNLP-IJCNLP), pages 4366–4375. Annual Meeting of the Association for Computa- tional Linguistics, pages 3645–3650, Florence, Italy. Wei Yang, Yuqing Xie, Aileen Lin, Xingyu Li, Luchen Association for Computational Linguistics. Tan, Kun Xiong, Ming Li, and Jimmy Lin. 2019. End-to-end open-domain question answering with bertserini. arXiv preprint arXiv:1902.01718. Haipeng Sun, Rui Wang, Kehai Chen, Masao Utiyama, Eiichiro Sumita, and Tiejun Zhao. 2019. Unsuper- Torsten Zesch, Iryna Gurevych, and Max Mühlhäuser. vised bilingual word embedding agreement for un- 2007. Analyzing and accessing wikipedia as a lexi- supervised neural machine translation. In Proceed- cal semantic resource. Data Structures for Linguis- ings of the 57th Annual Meeting of the Association tic Resources and Applications, 197205. for Computational Linguistics, pages 1235–1245. Barret Zoph, Deniz Yuret, Jonathan May, and Brian Thompson, Rebecca Knowles, Xuan Zhang, Kevin Knight. 2016. Transfer learning for low- Huda Khayrallah, Kevin Duh, and Philipp Koehn. resource neural machine translation. arXiv preprint 2019. Hablex: Human annotated bilingual lexicons arXiv:1604.02201. for experiments in machine translation. In Proceed- ings of the 2019 Conference on Empirical Methods Pierre Zweigenbaum, Serge Sharoff, and Reinhard in Natural Language Processing and the 9th Inter- Rapp. 2017. Overview of the second bucc shared national Joint Conference on Natural Language Pro- task: Spotting parallel sentences in comparable cor- cessing (EMNLP-IJCNLP), pages 1382–1387. pora. In Proceedings of the 10th Workshop on Build- ing and Using Comparable Corpora, pages 60–67. Gulmira Tolegen, Alymzhan Toleu, Orken Mamyr- bayev, and Rustam Mussabayev. 2020. Neural named entity recognition for kazakh. arXiv preprint arXiv:2007.13626. Jennifer Tracey, Stephanie Strassel, Ann Bies, Zhiyi Song, Michael Arrigo, Kira Griffitt, Dana Delgado, Dave Graff, Seth Kulick, Justin Mott, et al. 2019. Corpus building for low resource languages in the darpa lorelei program. In Proceedings of the 2nd Workshop on Technologies for MT of Low Resource Languages, pages 48–55.
You can also read