Comparing Unsupervised Word Translation Methods Step by Step

Research output: Chapter in Book/Report/Conference proceedingArticle in proceedingsResearchpeer-review

Cross-lingual word vector space alignment is the task of mapping the vocabularies of two languages into a shared semantic space, which can be used for dictionary induction, unsupervised machine translation, and transfer learning. In the unsupervised regime, an initial seed dictionary is learned in the absence of any known correspondences between words, through {\bf distribution matching}, and the seed dictionary is then used to supervise the induction of the final alignment in what is typically referred to as a (possibly iterative) {\bf refinement} step. We focus on the first step and compare distribution matching techniques in the context of language pairs for which mixed training stability and evaluation scores have been reported. We show that, surprisingly, when looking at this initial step in isolation, vanilla GANs are superior to more recent methods, both in terms of precision and robustness. The improvements reported by more recent methods thus stem from the refinement techniques, and we show that we can obtain state-of-the-art performance combining vanilla GANs with such refinement techniques.
Original languageEnglish
Title of host publicationAdvances in Neural Information Processing Systems 32 (NIPS 2019)
Number of pages11
Publication date2019
Publication statusPublished - 2019
Event33rd Conference on Neural Information Processing Systems (NeurIPS 2019) - Vancouver, Canada
Duration: 8 Dec 201914 Dec 2019


Conference33rd Conference on Neural Information Processing Systems (NeurIPS 2019)

ID: 240315759