Estimation of Rényi Entropy and Mutual Information Based on Generalized Nearest-Neighbor Graphs
NIPS(2010)
摘要
We present simple and computationally efficient nonparametric estimators of
Rényi entropy and mutual information based on an i.i.d. sample drawn from an
unknown, absolutely continuous distribution over ^d. The estimators are
calculated as the sum of p-th powers of the Euclidean lengths of the edges of
the `generalized nearest-neighbor' graph of the sample and the empirical copula
of the sample respectively. For the first time, we prove the almost sure
consistency of these estimators and upper bounds on their rates of convergence,
the latter of which under the assumption that the density underlying the sample
is Lipschitz continuous. Experiments demonstrate their usefulness in
independent subspace analysis.
更多查看译文
关键词
rate of convergence,nearest neighbor graph,upper bound,lipschitz continuity,absolute continuity,mutual information
AI 理解论文
溯源树
样例
生成溯源树,研究论文发展脉络