0
views
0
recommends
+1 Recommend
0 collections
    0
    shares
      • Record: found
      • Abstract: found
      • Article: found
      Is Open Access

      Pretraining Strategies for Structure Agnostic Material Property Prediction

      research-article

      Read this article at

      Bookmark
          There is no author summary for this article yet. Authors can add summaries to their articles on ScienceOpen to make them more accessible to a non-specialist audience.

          Abstract

          In recent years, machine learning (ML), especially graph neural network (GNN) models, has been successfully used for fast and accurate prediction of material properties. However, most ML models rely on relaxed crystal structures to develop descriptors for accurate predictions. Generating these relaxed crystal structures can be expensive and time-consuming, thus requiring an additional processing step for models that rely on them. To address this challenge, structure-agnostic methods have been developed, which use fixed-length descriptors engineered based on human knowledge about the material. However, the fixed-length descriptors are often hand-engineered and require extensive domain knowledge and generally are not used in the context of learnable models which are known to have a superior performance. Recent advancements have proposed learnable frameworks that can construct representations based on stoichiometry alone, allowing the flexibility of using deep learning frameworks as well as leveraging structure-agnostic learning. In this work, we propose three different pretraining strategies that can be used to pretrain these structure-agnostic, learnable frameworks to further improve the downstream material property prediction performance. We incorporate strategies such as self-supervised learning (SSL), fingerprint learning (FL), and multimodal learning (ML) and demonstrate their efficacy on downstream tasks for the Roost architecture, a popular structure-agnostic framework. Our results show significant improvement in small data sets and data efficiency in the larger data sets, underscoring the potential of our pretrain strategies that effectively leverage unlabeled data for accurate material property prediction.

          Related collections

          Most cited references55

          • Record: found
          • Abstract: not found
          • Article: not found

          Commentary: The Materials Project: A materials genome approach to accelerating materials innovation

            Bookmark
            • Record: found
            • Abstract: found
            • Article: not found

            BERT: Pre-training of Deep Bidirectional Transformers for Language Understanding

            We introduce a new language representation model called BERT, which stands for Bidirectional Encoder Representations from Transformers. Unlike recent language representation models, BERT is designed to pre-train deep bidirectional representations from unlabeled text by jointly conditioning on both left and right context in all layers. As a result, the pre-trained BERT model can be fine-tuned with just one additional output layer to create state-of-the-art models for a wide range of tasks, such as question answering and language inference, without substantial task-specific architecture modifications. BERT is conceptually simple and empirically powerful. It obtains new state-of-the-art results on eleven natural language processing tasks, including pushing the GLUE score to 80.5% (7.7% point absolute improvement), MultiNLI accuracy to 86.7% (4.6% absolute improvement), SQuAD v1.1 question answering Test F1 to 93.2 (1.5 point absolute improvement) and SQuAD v2.0 Test F1 to 83.1 (5.1 point absolute improvement).
              Bookmark
              • Record: found
              • Abstract: not found
              • Article: not found

              SchNet – A deep learning architecture for molecules and materials

                Bookmark

                Author and article information

                Journal
                J Chem Inf Model
                J Chem Inf Model
                ci
                jcisd8
                Journal of Chemical Information and Modeling
                American Chemical Society
                1549-9596
                1549-960X
                01 February 2024
                12 February 2024
                : 64
                : 3
                : 627-637
                Affiliations
                []Department of Material Science and Engineering, Carnegie Mellon University , Pittsburgh, Pennsylvania 15213, United States
                []Department of Mechanical Engineering, Carnegie Mellon University , Pittsburgh, Pennsylvania 15213, United States
                Author notes
                Author information
                https://orcid.org/0000-0002-7405-7689
                https://orcid.org/0000-0001-6216-0518
                https://orcid.org/0000-0002-2952-8576
                Article
                10.1021/acs.jcim.3c00919
                10865364
                38301621
                3dfdf5c4-e5f9-4141-b35c-e84f17d4d951
                © 2024 The Authors. Published by American Chemical Society

                Permits the broadest form of re-use including for commercial purposes, provided that author attribution and integrity are maintained ( https://creativecommons.org/licenses/by/4.0/).

                History
                : 20 June 2023
                : 12 January 2024
                : 11 January 2024
                Categories
                Article
                Custom metadata
                ci3c00919
                ci3c00919

                Computational chemistry & Modeling
                Computational chemistry & Modeling

                Comments

                Comment on this article