DocumentCode :
3216210
Title :
PAC=PAExact and other equivalent models in learning
Author :
Bshouty, Nader H. ; Gavinsky, Dmitry
Author_Institution :
Dept. of Comput. Sci., Technion-Israel Inst. of Technol., Haifa, Israel
fYear :
2002
fDate :
2002
Firstpage :
167
Lastpage :
176
Abstract :
The probably almost exact model (PAExact) can be viewed as the exact model relaxed so that: 1. The counterexamples to equivalence queries are distributionally drawn rather than adversarially chosen. 2. The output hypothesis is equal to the target with negligible error (1/ω(poly) for any poly). This model allows studying (almost) exact learnability of infinite classes and is in some sense analogous to the Exact-learning model for finite classes. It is known that PAExact-learnable⇒PAC-learnable [BJT02]. In this paper we show that if a class is PAC-learnable (in polynomial time) then it is PAExact-learnable (in polynomial time). Therefore, PAExact-learnable=PAC-learnable. It follows from this result that if a class is PAC-learnable then it is learnable in the probabilistic prediction model from examples with an algorithm that runs in polynomial time for each prediction (polynomial in log(the number of trials)) and that after polynomial number of mistakes achieves a hypothesis that predicts the target with probability 1-1/2poly. We also show that if a class is PAC-learnable in parallel then it is PAExact-learnable in parallel.
Keywords :
learning (artificial intelligence); probability; PAExact-learnable; equivalence queries; equivalent models; learning; probabilistic prediction model; probably almost exact model; Boosting; Computational modeling; Computer errors; Computer science; Computer simulation; Distributed computing; Error correction; Polynomials; Predictive models;
fLanguage :
English
Publisher :
ieee
Conference_Titel :
Foundations of Computer Science, 2002. Proceedings. The 43rd Annual IEEE Symposium on
ISSN :
0272-5428
Print_ISBN :
0-7695-1822-2
Type :
conf
DOI :
10.1109/SFCS.2002.1181893
Filename :
1181893
Link To Document :
بازگشت