Gezmu, A. M., Seyoum, B. E., Gasser, M., and Nürnberger,
A. (2018). Contemporary Amharic corpus:
Automatically morpho-syntactically tagged Amharic
corpus. In Proceedings of the First Workshop on
Linguistic Resources for Natural Language Processing,
pages 65–70, Santa Fe, New Mexico, USA. Association
for Computational Linguistics.
Heafield, K. (2011). KenLM: Faster and smaller language
model queries. In Proceedings of the Sixth Workshop
on Statistical Machine Translation, pages 187–197,
Edinburgh, Scotland. Association for Computational
Linguistics.
Huang, L. and Chiang, D. (2007). Forest rescoring: Faster
decoding with integrated language models. In
Proceedings of the 45th Annual Meeting of the
Association of Computational Linguistics, pages 144–
151, Prague, Czech Republic. Association for
Computational Linguistics.
Kingma, D. P. and Ba, J. (2014). Adam: A method for
stochastic optimization. Computing Research
Repository, arXiv:1412.6980. version 9.
Kneser, R., and Ney, H. (1995). Improved backing-off for
m-gram language modeling. In 1995 International
Conference on Acoustics, Speech, and Signal
Processing, Volume 1, pp. 181–184). IEEE.
Koehn, P. and Haddow, B. (2009). Edinburgh’s submission
to all tracks of the WMT 2009 shared task with
reordering and speed improvements to Moses. In
Proceedings of the Fourth Workshop on Statistical
Machine Translation, pages 160–164, Athens, Greece.
Association for Computational Linguistics.
Koehn, P., Hoang, H., Birch, A., Callison-Burch, C.,
Federico, M., Bertoldi, N., Cowan, B., Shen, W.,
Moran, C., Zens, R., Dyer, C., Bojar, O., Constantin,
A., and Herbst, E. (2007). Moses: Open source toolkit
for statistical machine translation. In Proceedings of the
45th Annual Meeting of the Association for
Computational Linguistics, Companion Volume
Proceedings of the Demo and Poster Sessions, pages
177–180, Prague, Czech Republic. Association for
Computational Linguistics.
Koehn, P. and Knowles, R. (2017). Six challenges for
neural machine translation. In Proceedings of the First
Workshop on Neural Machine Translation, pages 28–
39, Vancouver. Association for Computational
Linguistics.
Koehn, P., Och, F. J., and Marcu, D. (2003). Statistical
phrase-based translation. In Proceedings of the 2003
Human Language Technology Conference of the North
American Chapter of the Association for
Computational Linguistics, pages 127–133.
Kudo, T. (2018). Subword regularization: Improving neural
network translation models with multiple subword
candidates. In Proceedings of the 56th Annual Meeting
of the Association for Computational Linguistics
(Volume 1: Long Papers), pages 66–75, Melbourne,
Australia. Association for Computational Linguistics.
Kudo, T. and Richardson, J. (2018). SentencePiece: A
simple and language independent subword tokenizer
and detokenizer for neural text processing. In
Proceedings of the 2018 Conference on Empirical
Methods in Natural Language Processing: System
Demonstrations, pages 66–71, Brussels, Belgium.
Association for Computational Linguistics.
Kumar, S. and Byrne, W. (2004). Minimum Bayes-risk
decoding for statistical machine translation. In
Proceedings of the Human Language Technology
Conference of the North American Chapter of the
Association for Computational Linguistics: HLT-
NAACL 2004, pages 169–176, Boston, Massachusetts,
USA. Association for Computational Linguistics.
Lample, G., Ott, M., Conneau, A., Denoyer, L., and
Ranzato, M. (2018). Phrase-based & neural
unsupervised machine translation. In Proceedings of
the 2018 Conference on Empirical Methods in Natural
Language Processing, pages 5039–5049, Brussels,
Belgium. Association for Computational Linguistics.
Morishita, M., Oda, Y., Neubig, G., Yoshino, K., Sudoh,
K., and Nakamura, S. (2017). An empirical study of
minibatch creation strategies for neural machine
translation. In Proceedings of the First Workshop on
Neural Machine Translation, pages 61–68, Vancouver.
Association for Computational Linguistics.
Neishi, M., Sakuma, J., Tohda, S., Ishiwatari, S.,
Yoshinaga, N., and Toyoda, M. (2017). A bag of useful
tricks for practical neural machine translation:
Embedding layer initialization and large batch size. In
Proceedings of the 4th Workshop on Asian Translation
(WAT2017), pages 99–109, Taipei, Taiwan. Asian
Federation of Natural Language Processing.
Nguyen, T. and Chiang, D. (2018). Improving lexical
choice in neural machine translation. In Proceedings of
the 2018 Conference of the North American Chapter of
the Association for Computational Linguistics: Human
Language Technologies, Volume 1 (Long Papers),
pages 334–343, New Orleans, Louisiana. Association
for Computational Linguistics.
Och, F. J. (2003). Minimum error rate training in statistical
machine translation. In Proceedings of the 41st Annual
Meeting of the Association for Computational
Linguistics, pages 160–167, Sapporo, Japan.
Association for Computational Linguistics.
Och, F. J. and Ney, H. (2003). A systematic comparison of
various statistical alignment models. Computational
Linguistics, 29(1):19–51.
Papineni, K., Roukos, S., Ward, T., and Zhu, W.-J. (2002).
BLEU: a method for automatic evaluation of machine
translation. In Proceedings of the 40th Annual Meeting
of the Association for Computational Linguistics, pages
311–318, Philadelphia, Pennsylvania, USA.
Association for Computational Linguistics.
Post, M. (2018). A call for clarity in reporting BLEU scores.
In Proceedings of the Third Conference on Machine
Translation: Research Papers, pages 186–191,
Brussels, Belgium. Association for Computational
Linguistics.
Reiter, E. (2018). A structured review of the validity of
BLEU. Computational Linguistics, 44(3):393–401.
Sennrich, R., Haddow, B., and Birch, A. (2016). Neural
machine translation of rare words with subword units.