DocumentCode :
2386573
Title :
On learning mixtures of heavy-tailed distributions
Author :
Dasgupta, Anirban ; Hopcroft, John ; Kleinberg, Jon ; Sandler, Mark
fYear :
2005
fDate :
23-25 Oct. 2005
Firstpage :
491
Lastpage :
500
Abstract :
We consider the problem of learning mixtures of arbitrary symmetric distributions. We formulate sufficient separation conditions and present a learning algorithm with provable guarantees for mixtures of distributions that satisfy these separation conditions. Our bounds are independent of the variances of the distributions; to the best of our knowledge, there were no previous algorithms known with provable learning guarantees for distributions having infinite variance and/or expectation. For Gaussians and log-concave distributions, our results match the best known sufficient separation conditions by D. Achlioptas and F. McSherry (2005) and S. Vempala and G. Wang (2004). Our algorithm requires a sample of size O˜(dk), where d is the number of dimensions and k is the number of distributions in the mixture. We also show that for isotropic power-laws, exponential, and Gaussian distributions, our separation condition is optimal up to a constant factor.
Keywords :
Gaussian distribution; exponential distribution; learning (artificial intelligence); log normal distribution; normal distribution; Gaussian distribution; Gaussians distributions; arbitrary symmetric distributions; exponential distribution; heavy-tailed distributions; isotropic power-laws; learning mixture; log-concave distributions; Computer science; Data mining; Gaussian distribution; Iterative methods; Machine learning; Parameter estimation; Polynomials; Probability distribution; Statistical analysis; Statistical distributions;
fLanguage :
English
Publisher :
ieee
Conference_Titel :
Foundations of Computer Science, 2005. FOCS 2005. 46th Annual IEEE Symposium on
Print_ISBN :
0-7695-2468-0
Type :
conf
DOI :
10.1109/SFCS.2005.56
Filename :
1530741
Link To Document :
بازگشت