4
views
0
recommends
+1 Recommend
0 collections
    0
    shares
      • Record: found
      • Abstract: not found
      • Article: not found

      Improving fine-tuning of self-supervised models with Contrastive Initialization

      , , , , ,
      Neural Networks
      Elsevier BV

      Read this article at

      ScienceOpenPublisher
      Bookmark
          There is no author summary for this article yet. Authors can add summaries to their articles on ScienceOpen to make them more accessible to a non-specialist audience.

          Related collections

          Most cited references52

          • Record: found
          • Abstract: not found
          • Conference Proceedings: not found

          Momentum Contrast for Unsupervised Visual Representation Learning

            Bookmark
            • Record: found
            • Abstract: found
            • Article: not found

            Training products of experts by minimizing contrastive divergence.

            It is possible to combine multiple latent-variable models of the same data by multiplying their probability distributions together and then renormalizing. This way of combining individual "expert" models makes it hard to generate samples from the combined model but easy to infer the values of the latent variables of each expert, because the combination rule ensures that the latent variables of different experts are conditionally independent when given the data. A product of experts (PoE) is therefore an interesting candidate for a perceptual system in which rapid inference is vital and generation is unnecessary. Training a PoE by maximizing the likelihood of the data is difficult because it is hard even to approximate the derivatives of the renormalization term in the combination rule. Fortunately, a PoE can be trained using a different objective function called "contrastive divergence" whose derivatives with regard to the parameters can be approximated accurately and efficiently. Examples are presented of contrastive divergence learning using several types of expert on several types of data.
              Bookmark
              • Record: found
              • Abstract: not found
              • Article: not found

              Shortcut learning in deep neural networks

                Bookmark

                Author and article information

                Journal
                Neural Networks
                Neural Networks
                Elsevier BV
                08936080
                February 2023
                February 2023
                : 159
                : 198-207
                Article
                10.1016/j.neunet.2022.12.012
                1f7ce635-ef51-420c-83b1-062a73654900
                © 2023

                https://www.elsevier.com/tdm/userlicense/1.0/

                https://doi.org/10.15223/policy-017

                https://doi.org/10.15223/policy-037

                https://doi.org/10.15223/policy-012

                https://doi.org/10.15223/policy-029

                https://doi.org/10.15223/policy-004

                History

                Comments

                Comment on this article