9
views
0
recommends
+1 Recommend
0 collections
    0
    shares
      • Record: found
      • Abstract: found
      • Article: found
      Is Open Access

      RPEMHC: improved prediction of MHC–peptide binding affinity by a deep learning approach based on residue–residue pair encoding

      research-article

      Read this article at

      Bookmark
          There is no author summary for this article yet. Authors can add summaries to their articles on ScienceOpen to make them more accessible to a non-specialist audience.

          Abstract

          Motivation

          Binding of peptides to major histocompatibility complex (MHC) molecules plays a crucial role in triggering T cell recognition mechanisms essential for immune response. Accurate prediction of MHC peptide binding is vital for the development of cancer therapeutic vaccines. While recent deep learning-based methods have achieved significant performance in predicting MHC peptide binding affinity, most of them separately encode MHC molecules and peptides as inputs, potentially overlooking critical interaction information between the two.

          Results

          In this work, we propose RPEMHC, a new deep learning approach based on residue–residue pair encoding to predict the binding affinity between peptides and MHC, which encode an MHC molecule and a peptide as a residue–residue pair map. We evaluate the performance of RPEMHC on various MHC-II-related datasets for MHC peptide binding prediction, demonstrating that RPEMHC achieves better or comparable performance against other state-of-the-art baselines. Moreover, we further construct experiments on MHC-I-related datasets, and experimental results demonstrate that our method can work on both two MHC classes. These extensive validations have manifested that RPEMHC is an effective tool for studying MHC–peptide interactions and can potentially facilitate the vaccine development.

          Availability

          The source code of the method along with trained models is freely available at https://github.com/lennylv/RPEMHC.

          Related collections

          Most cited references32

          • Record: found
          • Abstract: found
          • Article: not found

          Long Short-Term Memory

          Learning to store information over extended time intervals by recurrent backpropagation takes a very long time, mostly because of insufficient, decaying error backflow. We briefly review Hochreiter's (1991) analysis of this problem, then address it by introducing a novel, efficient, gradient-based method called long short-term memory (LSTM). Truncating the gradient where this does not do harm, LSTM can learn to bridge minimal time lags in excess of 1000 discrete-time steps by enforcing constant error flow through constant error carousels within special units. Multiplicative gate units learn to open and close access to the constant error flow. LSTM is local in space and time; its computational complexity per time step and weight is O(1). Our experiments with artificial data involve local, distributed, real-valued, and noisy pattern representations. In comparisons with real-time recurrent learning, back propagation through time, recurrent cascade correlation, Elman nets, and neural sequence chunking, LSTM leads to many more successful runs, and learns much faster. LSTM also solves complex, artificial long-time-lag tasks that have never been solved by previous recurrent network algorithms.
            Bookmark
            • Record: found
            • Abstract: not found
            • Article: not found

            ImageNet classification with deep convolutional neural networks

              Bookmark
              • Record: found
              • Abstract: found
              • Article: not found

              Attention Is All You Need

              The dominant sequence transduction models are based on complex recurrent or convolutional neural networks in an encoder-decoder configuration. The best performing models also connect the encoder and decoder through an attention mechanism. We propose a new simple network architecture, the Transformer, based solely on attention mechanisms, dispensing with recurrence and convolutions entirely. Experiments on two machine translation tasks show these models to be superior in quality while being more parallelizable and requiring significantly less time to train. Our model achieves 28.4 BLEU on the WMT 2014 English-to-German translation task, improving over the existing best results, including ensembles by over 2 BLEU. On the WMT 2014 English-to-French translation task, our model establishes a new single-model state-of-the-art BLEU score of 41.8 after training for 3.5 days on eight GPUs, a small fraction of the training costs of the best models from the literature. We show that the Transformer generalizes well to other tasks by applying it successfully to English constituency parsing both with large and limited training data. 15 pages, 5 figures
                Bookmark

                Author and article information

                Contributors
                Role: Associate Editor
                Journal
                Bioinformatics
                Bioinformatics
                bioinformatics
                Bioinformatics
                Oxford University Press
                1367-4803
                1367-4811
                January 2024
                04 January 2024
                04 January 2024
                : 40
                : 1
                : btad785
                Affiliations
                School of Computer Science and Technology, Soochow University, Suzhou, Jiangsu 215006, China
                School of Computer Science and Technology, Soochow University, Suzhou, Jiangsu 215006, China
                Province Key Lab for Information Processing Technologies, Soochow University, Suzhou, Jiangsu 215006, China
                Collaborative Innovation Center of Novel Software Technology and Industrialization , Nanjing, Jiangsu 210000, China
                School of Computer Science and Technology, Soochow University, Suzhou, Jiangsu 215006, China
                School of Computer Science and Technology, Soochow University, Suzhou, Jiangsu 215006, China
                School of Computer Science and Technology, Soochow University, Suzhou, Jiangsu 215006, China
                School of Computer Science and Technology, Soochow University, Suzhou, Jiangsu 215006, China
                School of Computer Science and Technology, Soochow University, Suzhou, Jiangsu 215006, China
                School of Computer Science and Technology, Soochow University, Suzhou, Jiangsu 215006, China
                Province Key Lab for Information Processing Technologies, Soochow University, Suzhou, Jiangsu 215006, China
                Collaborative Innovation Center of Novel Software Technology and Industrialization , Nanjing, Jiangsu 210000, China
                School of Computer Science and Technology, Soochow University, Suzhou, Jiangsu 215006, China
                Province Key Lab for Information Processing Technologies, Soochow University, Suzhou, Jiangsu 215006, China
                Collaborative Innovation Center of Novel Software Technology and Industrialization , Nanjing, Jiangsu 210000, China
                Author notes
                Corresponding authors. School of Computer Science and Technology, Soochow University, 333 Ganjiangdong Road, Suzhou 215006, China. E-mail: tfwu@ 123456suda.edu.cn (T.W.); E-mail: qiang@ 123456suda.edu.cn (Q.L.)

                Equal contribution from Xuejiao Wang and Tingfang Wu.

                Author information
                https://orcid.org/0000-0001-8137-2436
                https://orcid.org/0000-0002-8103-555X
                Article
                btad785
                10.1093/bioinformatics/btad785
                10796178
                38175759
                64c04e61-066d-4c1c-bccf-a5178431bbeb
                © The Author(s) 2024. Published by Oxford University Press.

                This is an Open Access article distributed under the terms of the Creative Commons Attribution License ( https://creativecommons.org/licenses/by/4.0/), which permits unrestricted reuse, distribution, and reproduction in any medium, provided the original work is properly cited.

                History
                : 04 May 2023
                : 20 December 2023
                : 26 December 2023
                : 28 December 2023
                : 17 January 2024
                Page count
                Pages: 13
                Funding
                Funded by: National Natural Science Foundation of China, DOI 10.13039/501100001809;
                Award ID: 62002251
                Award ID: 62272335
                Funded by: Natural Science Foundation of Jiangsu Province Youth Fund;
                Award ID: BK20200856
                Funded by: Priority Academic Program Development of Jiangsu Higher Education Institutions, DOI 10.13039/501100012246;
                Funded by: Collaborative Innovation Center of Novel Software Technology and Industrialization;
                Categories
                Original Paper
                Sequence Analysis
                AcademicSubjects/SCI01060

                Bioinformatics & Computational biology
                Bioinformatics & Computational biology

                Comments

                Comment on this article