1
views
0
recommends
+1 Recommend
0 collections
    0
    shares
      • Record: found
      • Abstract: found
      • Article: found
      Is Open Access

      PanGu-Coder: Program Synthesis with Function-Level Language Modeling

      Preprint

      Read this article at

      Bookmark
          There is no author summary for this article yet. Authors can add summaries to their articles on ScienceOpen to make them more accessible to a non-specialist audience.

          Abstract

          We present PanGu-Coder, a pretrained decoder-only language model adopting the PanGu-Alpha architecture for text-to-code generation, i.e. the synthesis of programming language solutions given a natural language problem description. We train PanGu-Coder using a two-stage strategy: the first stage employs Causal Language Modelling (CLM) to pre-train on raw programming language data, while the second stage uses a combination of Causal Language Modelling and Masked Language Modelling (MLM) training objectives that focus on the downstream task of text-to-code generation and train on loosely curated pairs of natural language program definitions and code functions. Finally, we discuss PanGu-Coder-FT, which is fine-tuned on a combination of competitive programming problems and code with continuous integration tests. We evaluate PanGu-Coder with a focus on whether it generates functionally correct programs and demonstrate that it achieves equivalent or better performance than similarly sized models, such as CodeX, while attending a smaller context window and training on less data.

          Related collections

          Author and article information

          Journal
          22 July 2022
          Article
          2207.11280
          7084be75-d33c-44fd-8853-13418040f0cb

          http://arxiv.org/licenses/nonexclusive-distrib/1.0/

          History
          Custom metadata
          27 pages
          cs.LG cs.AI cs.CL cs.PL cs.SE

          Software engineering,Theoretical computer science,Programming languages,Artificial intelligence

          Comments

          Comment on this article