On Long-Tailed Phenomena in Neural Machine Translation

Vikas Raunak, Siddharth Dalmia, Vivek Gupta, Florian Metze

4th Workshop on Structured Prediction for NLP Workshop Paper

You can open the pre-recorded video in a separate window.

Abstract: State-of-the-art Neural Machine Translation (NMT) models struggle with generating low-frequency tokens, tackling which remains a major challenge. The analysis of long-tailed phenomena in the context of structured prediction tasks is further hindered by the added complexities of search during inference. In this work, we quantitatively characterize such long-tailed phenomena at two levels of abstraction, namely, token classification and sequence generation. We propose a new loss function, the Anti-Focal loss, to better adapt model training to the structural dependencies of conditional text generation by incorporating the inductive biases of beam search in the training process. We show the efficacy of the proposed technique on a number of Machine Translation (MT) datasets, demonstrating that it leads to significant gains over cross-entropy across different language pairs, especially on the generation of low-frequency words. We have released the code to reproduce our results.
NOTE: Video may display a random order of authors. Correct author list is at the top of this page.