DocumentCode :
3688598
Title :
A rate-distortion framework for supervised learning
Author :
Matthew Nokleby;Ahmad Beirami;Robert Calderbank
Author_Institution :
Department of Electrical and Computer Engineering, Duke University, Durham, NC 27708, USA
fYear :
2015
Firstpage :
1
Lastpage :
6
Abstract :
An information-theoretic framework is presented for bounding the number of samples needed for supervised learning in a parametric Bayesian setting. This framework is inspired by an analogy with rate-distortion theory, which characterizes tradeoffs in the lossy compression of random sources. In a parametric Bayesian environment, the maximum a posteriori classifier can be viewed as a random function of the model parameters. Labeled training data can be viewed as a finite-rate encoding of that source, and the excess loss due to using the learned classifier instead of the MAP classifier can be viewed as distortion. A strict bound on the loss-measured in terms of the expected total variation-is derived, providing a minimum number of training samples needed to drive the expected total variation to within a specified tolerance. The tightness of this bound is demonstrated on the classification of Gaus-sians, for which one can derive closed-form expressions for the bound.
Keywords :
"Interpolation","Training","Entropy","Yttrium","Zinc","Rate-distortion","Mutual information"
Publisher :
ieee
Conference_Titel :
Machine Learning for Signal Processing (MLSP), 2015 IEEE 25th International Workshop on
Type :
conf
DOI :
10.1109/MLSP.2015.7324319
Filename :
7324319
Link To Document :
بازگشت