Abstract
We investigate the nonparametric, composite hypothesis testing problem for arbitrary unknown distributions in the asymptotic regime where both the sample size and the number of hypothesis grow exponentially large. Such asymptotic analysis is important in many practical problems, where the number of variations that can exist within a family of distributions can be countably infinite. We introduce the notion of discrimination capacity, which captures the largest exponential growth rate of the number of hypothesis relative to the sample size so that there exists a test with asymptotically vanishing probability of error. Our approach is based on various distributional distance metrics in order to incorporate the generative model of the data. We provide analyses of the error exponent using the maximum mean discrepancy and Kolmogorov-Smirnov distance and characterize the corresponding discrimination rates, i.e., lower bounds on the discrimination capacity, for these tests. Finally, an upper bound on the discrimination capacity based on Fano's inequality is developed. Numerical results are presented to validate the theoretical results.
Original language | English (US) |
---|---|
Article number | 8438996 |
Pages (from-to) | 1005-1014 |
Number of pages | 10 |
Journal | IEEE Journal on Selected Topics in Signal Processing |
Volume | 12 |
Issue number | 5 |
DOIs | |
State | Published - Oct 2018 |
Keywords
- Channel coding
- Kolmogorov-Smirnov distance
- discrimination rate
- error exponent
- maximum mean discrepancy
ASJC Scopus subject areas
- Signal Processing
- Electrical and Electronic Engineering