10
views
0
recommends
+1 Recommend
0 collections
    0
    shares
      • Record: found
      • Abstract: found
      • Article: found
      Is Open Access

      Regularizing Recurrent Neural Networks via Sequence Mixup

      Preprint
      , ,

      Read this article at

      Bookmark
          There is no author summary for this article yet. Authors can add summaries to their articles on ScienceOpen to make them more accessible to a non-specialist audience.

          Abstract

          In this paper, we extend a class of celebrated regularization techniques originally proposed for feed-forward neural networks, namely Input Mixup (Zhang et al., 2017) and Manifold Mixup (Verma et al., 2018), to the realm of Recurrent Neural Networks (RNN). Our proposed methods are easy to implement and have a low computational complexity, while leverage the performance of simple neural architectures in a variety of tasks. We have validated our claims through several experiments on real-world datasets, and also provide an asymptotic theoretical analysis to further investigate the properties and potential impacts of our proposed techniques. Applying sequence mixup to BiLSTM-CRF model (Huang et al., 2015) to Named Entity Recognition task on CoNLL-2003 data (Sang and De Meulder, 2003) has improved the F-1 score on the test stage and reduced the loss, considerably.

          Related collections

          Author and article information

          Journal
          27 November 2020
          Article
          2012.07527
          aeca22dc-56db-439a-8167-ab1e940bfeb9

          http://creativecommons.org/licenses/by/4.0/

          History
          Custom metadata
          17 pages
          cs.CL cs.LG stat.ML

          Theoretical computer science,Machine learning,Artificial intelligence
          Theoretical computer science, Machine learning, Artificial intelligence

          Comments

          Comment on this article