DocumentCode :
2026072
Title :
Divergence estimation for machine learning and signal processing
Author :
Sugiyama, Masakazu
Author_Institution :
Dept. of Comput. Sci., Tokyo Inst. of Technol. Tokyo, Tokyo, Japan
fYear :
2013
fDate :
18-20 Feb. 2013
Firstpage :
12
Lastpage :
13
Abstract :
Approximating a divergence between two probability distributions from their samples is a fundamental challenge in the statistics, information theory, and machine learning communities, because a divergence estimator can be used for various purposes such as two-sample homogeneity testing, change-point detection, and class-balance estimation. Furthermore, an approximator of a divergence between the joint distribution and the product of marginals can be used for independence testing, which has a wide range of applications including feature selection and extraction, clustering, object matching, independent component analysis, and causality learning. In this talk, we review recent advances in direct divergence approximation that follow the general inference principle advocated by Vladimir Vapnik-one should not solve a more general problem as an intermediate step. More specifically, direct divergence approximation avoids separately estimating two probability distributions when approximating a divergence. We cover direct approximators of the Kullback-Leibler (KL) divergence, the Pearson (PE) divergence, the relative PE (rPE) divergence, and the L2-distance. Despite the overwhelming popularity of the KL divergence, we argue that the latter approximators are more useful in practice due to their computational efficiency, high numerical stability, and superior robustness against outliers.
Keywords :
approximation theory; independent component analysis; learning (artificial intelligence); signal processing; statistical distributions; statistical testing; KL divergence; Kullback-Leibler divergence; L2-distance; Pearson divergence; causality learning; change-point detection; class-balance estimation; clustering application; divergence approximation; divergence estimation; feature extraction; feature selection; independence testing; independent component analysis; inference principle; information theory; machine learning; object matching; probability distribution; relative PE divergence; signal processing; statistics; two-sample homogeneity testing; Approximation methods; Estimation; Feature extraction; Mutual information; Neural networks; Probability distribution; Testing; Kullback-Leibler divergence; L 2-distance; Pearson divergence; relative Pearson divergence;
fLanguage :
English
Publisher :
ieee
Conference_Titel :
Brain-Computer Interface (BCI), 2013 International Winter Workshop on
Conference_Location :
Gangwo
Print_ISBN :
978-1-4673-5973-3
Type :
conf
DOI :
10.1109/IWW-BCI.2013.6506611
Filename :
6506611
Link To Document :
بازگشت