0
views
0
recommends
+1 Recommend
0 collections
    0
    shares
      • Record: found
      • Abstract: found
      • Article: found
      Is Open Access

      Teaching Syntax by Adversarial Distraction

      Preprint
      , ,

      Read this article at

      Bookmark
          There is no author summary for this article yet. Authors can add summaries to their articles on ScienceOpen to make them more accessible to a non-specialist audience.

          Abstract

          Existing entailment datasets mainly pose problems which can be answered without attention to grammar or word order. Learning syntax requires comparing examples where different grammar and word order change the desired classification. We introduce several datasets based on synthetic transformations of natural entailment examples in SNLI or FEVER, to teach aspects of grammar and word order. We show that without retraining, popular entailment models are unaware that these syntactic differences change meaning. With retraining, some but not all popular entailment models can learn to compare the syntax properly.

          Related collections

          Most cited references11

          • Record: found
          • Abstract: not found
          • Conference Proceedings: not found

          Neural Machine Translation of Rare Words with Subword Units

            Bookmark
            • Record: found
            • Abstract: not found
            • Conference Proceedings: not found

            Supervised Learning of Universal Sentence Representations from Natural Language Inference Data

              Bookmark
              • Record: found
              • Abstract: found
              • Article: found
              Is Open Access

              A large annotated corpus for learning natural language inference

              Understanding entailment and contradiction is fundamental to understanding natural language, and inference about entailment and contradiction is a valuable testing ground for the development of semantic representations. However, machine learning research in this area has been dramatically limited by the lack of large-scale resources. To address this, we introduce the Stanford Natural Language Inference corpus, a new, freely available collection of labeled sentence pairs, written by humans doing a novel grounded task based on image captioning. At 570K pairs, it is two orders of magnitude larger than all other resources of its type. This increase in scale allows lexicalized classifiers to outperform some sophisticated existing entailment models, and it allows a neural network-based model to perform competitively on natural language inference benchmarks for the first time.
                Bookmark

                Author and article information

                Journal
                25 October 2018
                Article
                1810.11067
                eb760df1-f07a-4c31-bf38-44875cb09c73

                http://arxiv.org/licenses/nonexclusive-distrib/1.0/

                History
                Custom metadata
                Juho Kim, Christopher Malon, and Asim Kadav. 2018. "Teaching Syntax by Adversarial Distraction." Proceedings of the EMNLP First Workshop on Fact Extraction and Verification
                To appear at the EMNLP 2018 First Workshop on Fact Extraction and Verification (FEVER)
                cs.CL

                Theoretical computer science
                Theoretical computer science

                Comments

                Comment on this article