3
views
0
recommends
+1 Recommend
0 collections
    0
    shares
      • Record: found
      • Abstract: found
      • Article: found
      Is Open Access

      Beyond the Benchmarks: Toward Human-Like Lexical Representations

      review-article

      Read this article at

      ScienceOpenPublisherPMC
          There is no author summary for this article yet. Authors can add summaries to their articles on ScienceOpen to make them more accessible to a non-specialist audience.

          Abstract

          To process language in a way that is compatible with human expectations in a communicative interaction, we need computational representations of lexical properties that form the basis of human knowledge of words. In this article, we concentrate on word-level semantics. We discuss key concepts and issues that underlie the scientific understanding of the human lexicon: its richly structured semantic representations, their ready and continual adaptability, and their grounding in crosslinguistically valid conceptualization. We assess the state of the art in natural language processing (NLP) in achieving these identified properties, and suggest ways in which the language sciences can inspire new approaches to their computational instantiation.

          Related collections

          Most cited references208

          • Record: found
          • Abstract: found
          • Article: not found

          Long Short-Term Memory

          Learning to store information over extended time intervals by recurrent backpropagation takes a very long time, mostly because of insufficient, decaying error backflow. We briefly review Hochreiter's (1991) analysis of this problem, then address it by introducing a novel, efficient, gradient-based method called long short-term memory (LSTM). Truncating the gradient where this does not do harm, LSTM can learn to bridge minimal time lags in excess of 1000 discrete-time steps by enforcing constant error flow through constant error carousels within special units. Multiplicative gate units learn to open and close access to the constant error flow. LSTM is local in space and time; its computational complexity per time step and weight is O(1). Our experiments with artificial data involve local, distributed, real-valued, and noisy pattern representations. In comparisons with real-time recurrent learning, back propagation through time, recurrent cascade correlation, Elman nets, and neural sequence chunking, LSTM leads to many more successful runs, and learns much faster. LSTM also solves complex, artificial long-time-lag tasks that have never been solved by previous recurrent network algorithms.
            • Record: found
            • Abstract: not found
            • Conference Proceedings: not found

            Glove: Global Vectors for Word Representation

              • Record: found
              • Abstract: found
              • Article: not found

              BERT: Pre-training of Deep Bidirectional Transformers for Language Understanding

              We introduce a new language representation model called BERT, which stands for Bidirectional Encoder Representations from Transformers. Unlike recent language representation models, BERT is designed to pre-train deep bidirectional representations from unlabeled text by jointly conditioning on both left and right context in all layers. As a result, the pre-trained BERT model can be fine-tuned with just one additional output layer to create state-of-the-art models for a wide range of tasks, such as question answering and language inference, without substantial task-specific architecture modifications. BERT is conceptually simple and empirically powerful. It obtains new state-of-the-art results on eleven natural language processing tasks, including pushing the GLUE score to 80.5% (7.7% point absolute improvement), MultiNLI accuracy to 86.7% (4.6% absolute improvement), SQuAD v1.1 question answering Test F1 to 93.2 (1.5 point absolute improvement) and SQuAD v2.0 Test F1 to 83.1 (5.1 point absolute improvement).

                Author and article information

                Contributors
                Journal
                Front Artif Intell
                Front Artif Intell
                Front. Artif. Intell.
                Frontiers in Artificial Intelligence
                Frontiers Media S.A.
                2624-8212
                24 May 2022
                2022
                : 5
                : 796741
                Affiliations
                [1] 1Department of Computer Science, University of Toronto , Toronto, ON, Canada
                [2] 2Linguistics Department, University of Geneva , Geneva, Switzerland
                Author notes

                Edited by: Alessandro Lenci, University of Pisa, Italy

                Reviewed by: Nasredine Semmar, CEA Saclay, France; James Pustejovsky, Brandeis University, United States

                *Correspondence: Suzanne Stevenson suzanne@ 123456cs.toronto.edu

                This article was submitted to Frontiers in Artificial Intelligence, a section of the journal Frontiers in Artificial Intelligence

                Article
                10.3389/frai.2022.796741
                9170951
                35685444
                c60841bd-221b-4942-a8c6-d63c181fc405
                Copyright © 2022 Stevenson and Merlo.

                This is an open-access article distributed under the terms of the Creative Commons Attribution License (CC BY). The use, distribution or reproduction in other forums is permitted, provided the original author(s) and the copyright owner(s) are credited and that the original publication in this journal is cited, in accordance with accepted academic practice. No use, distribution or reproduction is permitted which does not comply with these terms.

                History
                : 17 October 2021
                : 19 April 2022
                Page count
                Figures: 0, Tables: 0, Equations: 0, References: 208, Pages: 14, Words: 13160
                Categories
                Artificial Intelligence
                Conceptual Analysis

                computational linguistics,natural language processing,lexical semantics,lexicon structure,human lexical representations,cross-linguistic generalization

                Comments

                Comment on this article

                Related Documents Log