11
views
0
recommends
+1 Recommend
0 collections
    0
    shares
      • Record: found
      • Abstract: found
      • Article: found
      Is Open Access

      Random Feature Stein Discrepancies

      Preprint
      ,

      Read this article at

      Bookmark
          There is no author summary for this article yet. Authors can add summaries to their articles on ScienceOpen to make them more accessible to a non-specialist audience.

          Abstract

          Computable Stein discrepancies have been deployed for a variety of applications, including sampler selection in posterior inference, approximate Bayesian inference, and goodness-of-fit testing. Existing convergence-determining Stein discrepancies admit strong theoretical guarantees but suffer from a computational cost that grows quadratically in the sample size. While linear-time Stein discrepancies have been proposed for goodness-of-fit testing, they exhibit avoidable degradations in testing power---even when power is explicitly optimized. To address these shortcomings, we introduce feature Stein discrepancies (\(\Phi\)SDs), a new family of quality measures that can be cheaply approximated using importance sampling. We show how to construct \(\Phi\)SDs that provably determine the convergence of a sample to its target and develop high-accuracy approximations---random \(\Phi\)SDs (R\(\Phi\)SDs)---which are computable in near-linear time. In our experiments with sampler selection for approximate posterior inference and goodness-of-fit testing, R\(\Phi\)SDs typically perform as well or better than quadratic-time KSDs while being orders of magnitude faster to compute.

          Related collections

          Most cited references5

          • Record: found
          • Abstract: found
          • Article: found
          Is Open Access

          Equivalence of distance-based and RKHS-based statistics in hypothesis testing

          We provide a unifying framework linking two classes of statistics used in two-sample and independence testing: on the one hand, the energy distances and distance covariances from the statistics literature; on the other, maximum mean discrepancies (MMD), that is, distances between embeddings of distributions to reproducing kernel Hilbert spaces (RKHS), as established in machine learning. In the case where the energy distance is computed with a semimetric of negative type, a positive definite kernel, termed distance kernel, may be defined such that the MMD corresponds exactly to the energy distance. Conversely, for any positive definite kernel, we can interpret the MMD as energy distance with respect to some negative-type semimetric. This equivalence readily extends to distance covariance using kernels on the product space. We determine the class of probability distributions for which the test statistics are consistent against all alternatives. Finally, we investigate the performance of the family of distance kernels in two-sample and independence tests: we show in particular that the energy distance most commonly employed in statistics is just one member of a parametric family of kernels, and that other choices from this family can yield more powerful tests.
            Bookmark
            • Record: found
            • Abstract: not found
            • Article: not found

            Control functionals for Monte Carlo integration

              Bookmark
              • Record: found
              • Abstract: not found
              • Article: not found

              VECTOR VALUED REPRODUCING KERNEL HILBERT SPACES AND UNIVERSALITY

                Bookmark

                Author and article information

                Journal
                20 June 2018
                Article
                1806.07788
                95b389ff-c2f1-4f2c-8668-aab7d03de294

                http://arxiv.org/licenses/nonexclusive-distrib/1.0/

                History
                Custom metadata
                20 pages, 4 figures
                stat.ML cs.LG

                Machine learning,Artificial intelligence
                Machine learning, Artificial intelligence

                Comments

                Comment on this article