28
views
0
recommends
+1 Recommend
0 collections
    0
    shares
      • Record: found
      • Abstract: not found
      • Article: not found

      Perspective: Materials informatics and big data: Realization of the “fourth paradigm” of science in materials science

      ,
      APL Materials
      AIP Publishing

      Read this article at

      ScienceOpenPublisher
      Bookmark
          There is no author summary for this article yet. Authors can add summaries to their articles on ScienceOpen to make them more accessible to a non-specialist audience.

          Related collections

          Most cited references40

          • Record: found
          • Abstract: not found
          • Article: not found

          The random subspace method for constructing decision forests

          Tin Ho (1998)
            Bookmark
            • Record: found
            • Abstract: not found
            • Article: not found

            Instance-based learning algorithms

              Bookmark
              • Record: found
              • Abstract: found
              • Article: not found

              Rotation forest: A new classifier ensemble method.

              We propose a method for generating classifier ensembles based on feature extraction. To create the training data for a base classifier, the feature set is randomly split into K subsets (K is a parameter of the algorithm) and Principal Component Analysis (PCA) is applied to each subset. All principal components are retained in order to preserve the variability information in the data. Thus, K axis rotations take place to form the new features for a base classifier. The idea of the rotation approach is to encourage simultaneously individual accuracy and diversity within the ensemble. Diversity is promoted through the feature extraction for each base classifier. Decision trees were chosen here because they are sensitive to rotation of the feature axes, hence the name "forest." Accuracy is sought by keeping all principal components and also using the whole data set to train each base classifier. Using WEKA, we examined the Rotation Forest ensemble on a random selection of 33 benchmark data sets from the UCI repository and compared it with Bagging, AdaBoost, and Random Forest. The results were favorable to Rotation Forest and prompted an investigation into diversity-accuracy landscape of the ensemble models. Diversity-error diagrams revealed that Rotation Forest ensembles construct individual classifiers which are more accurate than these in AdaBoost and Random Forest, and more diverse than these in Bagging, sometimes more accurate as well.
                Bookmark

                Author and article information

                Journal
                AMPADS
                APL Materials
                APL Mater.
                AIP Publishing
                2166-532X
                May 01 2016
                May 01 2016
                : 4
                : 5
                : 053208
                Article
                10.1063/1.4946894
                35586923-f421-41fd-ac07-4d8a182296fb
                © 2016
                History

                Comments

                Comment on this article