Journal
JOURNAL OF MACHINE LEARNING RESEARCH
Volume 22, Issue -, Pages -Publisher
MICROTOME PUBL
Keywords
Adaptation; goodness of fit; maximum mean discrepancy; optimal rates of Convergence; reproducing kernel Hilbert space
Funding
- National Science Foundation (NSF) [DMS-1803450, DMS-2015285]
Ask authors/readers for more resources
The study examines the statistical performance of reproducing kernel Hilbert space (RKHS) embedding in testing problems, showing that a basic version of kernel embedding test may not be optimal, especially when chi(2) distance is used as the separation metric. The authors propose a simple modification to address this issue, demonstrating that the moderated approach offers optimal tests for various deviations from null hypotheses and can adapt over multiple interpolation spaces. Numerical experiments support the effectiveness of the approach.
The reproducing kernel Hilbert space (RKHS) embedding of distributions offers a general and flexible framework for testing problems in arbitrary domains and has attracted considerable amount of attention in recent years. To gain insights into their operating characteristics, we study here the statistical performance of such approaches within a minimax framework. Focusing on the case of goodness-of-fit tests, our analyses show that a vanilla version of the kernel embedding based test could be minimax suboptimal, when considering chi(2) distance as the separation metric. Hence we suggest a simple remedy by moderating the embedding. We prove that the moderated approach provides optimal tests for a wide range of deviations from the null and can also be made adaptive over a large collection of interpolation spaces. Numerical experiments are presented to further demonstrate the merits of our approach.
Authors
I am an author on this paper
Click your name to claim this paper and add it to your profile.
Reviews
Recommended
No Data Available