6
views
0
recommends
+1 Recommend
0 collections
    0
    shares
      • Record: found
      • Abstract: found
      • Article: found
      Is Open Access

      Unsupervised Learning of Syntactic Structure with Invertible Neural Projections

      Preprint
      , ,

      Read this article at

      Bookmark
          There is no author summary for this article yet. Authors can add summaries to their articles on ScienceOpen to make them more accessible to a non-specialist audience.

          Abstract

          Unsupervised learning of syntactic structure is typically performed using generative models with discrete latent variables and multinomial parameters. In most cases, these models have not leveraged continuous word representations. In this work, we propose a novel generative model that jointly learns discrete syntactic structure and continuous word representations in an unsupervised fashion by cascading an invertible neural network with a structured generative prior. We show that the invertibility condition allows for efficient exact inference and marginal likelihood computation in our model so long as the prior is well-behaved. In experiments we instantiate our approach with both Markov and tree-structured priors, evaluating on two tasks: part-of-speech (POS) induction, and unsupervised dependency parsing without gold POS annotation. On the Penn Treebank, our Markov-structured model surpasses state-of-the-art results on POS induction. Similarly, we find that our tree-structured model achieves state-of-the-art performance on unsupervised dependency parsing for the difficult training condition where neither gold POS annotation nor punctuation-based constraints are available.

          Related collections

          Most cited references2

          • Record: found
          • Abstract: not found
          • Conference Proceedings: not found

          Language as a Latent Variable: Discrete Generative Models for Sentence Compression

            Bookmark
            • Record: found
            • Abstract: found
            • Article: found
            Is Open Access

            Unsupervised Neural Hidden Markov Models

            In this work, we present the first results for neuralizing an Unsupervised Hidden Markov Model. We evaluate our approach on tag in- duction. Our approach outperforms existing generative models and is competitive with the state-of-the-art though with a simpler model easily extended to include additional context.
              Bookmark

              Author and article information

              Journal
              28 August 2018
              Article
              1808.09111
              a9902739-63c2-4ba9-bb08-2ba24f1aa8c2

              http://arxiv.org/licenses/nonexclusive-distrib/1.0/

              History
              Custom metadata
              EMNLP 2018
              cs.CL cs.LG

              Theoretical computer science,Artificial intelligence
              Theoretical computer science, Artificial intelligence

              Comments

              Comment on this article