Sequence Transduction: Generalization and Challenges

by Thamme Gowda in Note tags: NMT
Sequence to sequence transduction is a general problem, for which many other problems are special cases. I also highlight some challenges of this general problem. more…

Many-to-English Machine Translation Tools, Data, and Pretrained Models

by Thamme Gowda in Paper tags: NMT
We present useful tools for machine translation research: MTData, NLCodec, and RTG. We demonstrate their usefulness by creating a multilingual neural machine translation model capable of translating from 500 source languages to English. We make this multilingual model readily downloadable and usable as a service, or as a parent model for transfer-learning to even lower-resource languages. more…

Macro-Average: Rare Types Are Important Too

by Thamme Gowda in Paper tags: NMT
We explore the simple type-based classifier metric, \maf1, and study its applicability to MT evaluation. We find that MacroF1 is competitive on direct assessment, and outperforms others in indicating downstream cross-lingual information retrieval task performance. Further, we show that MacroF1 can be used to effectively compare supervised and unsupervised neural machine translation, and reveal significant qualitative differences in the methods' outputs. more…

Finding the Optimal Vocabulary for Neural Machine Translation

by Thamme Gowda in Paper tags: NMT
We cast neural machine translation (NMT) as a classification task in an autoregressive setting and analyze the limitations of both classification and autoregression components. Classifiers are known to perform better with balanced class distributions during training. Since the Zipfian nature of languages causes imbalanced classes, we explore its effect on NMT. We analyze the effect of various vocabulary sizes on NMT performance on multiple languages with many data sizes, and reveal an explanation for _why_ certain vocabulary sizes are better than others. more…