12
views
0
recommends
+1 Recommend
0 collections
    0
    shares
      • Record: found
      • Abstract: found
      • Article: found
      Is Open Access

      Multiple imputation for continuous variables using a Bayesian principal component analysis

      Preprint
      , ,

      Read this article at

      Bookmark
          There is no author summary for this article yet. Authors can add summaries to their articles on ScienceOpen to make them more accessible to a non-specialist audience.

          Abstract

          We propose a multiple imputation method based on principal component analysis (PCA) to deal with incomplete continuous data. To reflect the uncertainty of the parameters from one imputation to the next, we use a Bayesian treatment of the PCA model. Using a simulation study and real data sets, the method is compared to two classical approaches: multiple imputation based on joint modelling and on fully conditional modelling. Contrary to the others, the proposed method can be easily used on data sets where the number of individuals is less than the number of variables and when the variables are highly correlated. In addition, it provides unbiased point estimates of quantities of interest, such as an expectation, a regression coefficient or a correlation coefficient, with a smaller mean squared error. Furthermore, the widths of the confidence intervals built for the quantities of interest are often smaller whilst ensuring a valid coverage.

          Related collections

          Most cited references14

          • Record: found
          • Abstract: not found
          • Article: not found

          Fully conditional specification in multivariate imputation

            Bookmark
            • Record: found
            • Abstract: not found
            • Article: not found

            What to Do about Missing Values in Time-Series Cross-Section Data

              Bookmark
              • Record: found
              • Abstract: found
              • Article: not found

              Cross-validation of component models: a critical look at current methods.

              In regression, cross-validation is an effective and popular approach that is used to decide, for example, the number of underlying features, and to estimate the average prediction error. The basic principle of cross-validation is to leave out part of the data, build a model, and then predict the left-out samples. While such an approach can also be envisioned for component models such as principal component analysis (PCA), most current implementations do not comply with the essential requirement that the predictions should be independent of the entity being predicted. Further, these methods have not been properly reviewed in the literature. In this paper, we review the most commonly used generic PCA cross-validation schemes and assess how well they work in various scenarios.
                Bookmark

                Author and article information

                Journal
                2014-01-22
                2015-08-19
                Article
                1401.5747
                591841af-366e-443c-b654-599cee0796e4

                http://arxiv.org/licenses/nonexclusive-distrib/1.0/

                History
                Custom metadata
                62H25, 62F10, 62F40, 62F15
                16 pages
                stat.ME

                Methodology
                Methodology

                Comments

                Comment on this article