1
views
0
recommends
+1 Recommend
0 collections
    0
    shares
      • Record: found
      • Abstract: found
      • Article: found
      Is Open Access

      Multilingual Neural Machine Translation with Knowledge Distillation

      Preprint
      , , , , ,

      Read this article at

      Bookmark
          There is no author summary for this article yet. Authors can add summaries to their articles on ScienceOpen to make them more accessible to a non-specialist audience.

          Abstract

          Multilingual machine translation, which translates multiple languages with a single model, has attracted much attention due to its efficiency of offline training and online serving. However, traditional multilingual translation usually yields inferior accuracy compared with the counterpart using individual models for each language pair, due to language diversity and model capacity limitations. In this paper, we propose a distillation-based approach to boost the accuracy of multilingual machine translation. Specifically, individual models are first trained and regarded as teachers, and then the multilingual model is trained to fit the training data and match the outputs of individual models simultaneously through knowledge distillation. Experiments on IWSLT, WMT and Ted talk translation datasets demonstrate the effectiveness of our method. Particularly, we show that one model is enough to handle multiple languages (up to 44 languages in our experiment), with comparable or even better accuracy than individual models.

          Related collections

          Most cited references3

          • Record: found
          • Abstract: not found
          • Conference Proceedings: not found

          A Gift from Knowledge Distillation: Fast Optimization, Network Minimization and Transfer Learning

            Bookmark
            • Record: found
            • Abstract: not found
            • Conference Proceedings: not found

            Multi-Task Learning for Multiple Language Translation

              Bookmark
              • Record: found
              • Abstract: not found
              • Conference Proceedings: not found

              Sequence-Level Knowledge Distillation

                Bookmark

                Author and article information

                Journal
                27 February 2019
                Article
                1902.10461
                3a912f08-8447-4853-a8b9-6a28c959a271

                http://arxiv.org/licenses/nonexclusive-distrib/1.0/

                History
                Custom metadata
                Accepted to ICLR 2019
                cs.CL

                Theoretical computer science
                Theoretical computer science

                Comments

                Comment on this article