DocumentCode :
3031661
Title :
Bayesian state combining for context models
Author :
Bunton, Suzanne
Author_Institution :
Dept. of Comput. Sci. & Eng., Washington Univ., Seattle, WA, USA
fYear :
1998
fDate :
30 Mar-1 Apr 1998
Firstpage :
329
Lastpage :
338
Abstract :
The best-performing on-line methods for estimating probabilities of symbols in a sequence (required for computing minimal codes) use context trees with either information-theoretic state selection or context-tree weighting. This paper derives de novo from Bayes´ theorem, a novel technique for modeling sequences on-line with context trees, which we call “Bayesian state combining” or BSC. BSC is comparable in function to both information-theoretic state selection and context-tree weighting. However, it is a truly distinct alternative to either of these techniques, which like BSC, can be viewed as “dispatchers” of probability estimates from the set of competing, memoryless models represented by the context tree. The resulting technique handles sequences over m-ary input alphabets for arbitrary m and may employ any probability estimator applicable to context models (e.g., Laplace, Krichevsky-Trofimov, blending, and more generally, mixtures). In experiments that control other (256-ary) context-tree model features such as Markov order and probability estimators, we compare the performance of BSC and information-theoretic state selection. The background notation and concepts are reviewed, as required to understand the modeling problem and application of our result. The leading notion of the paper is derived, which dynamically maps certain states in context-models to a set of mutually exclusive hypotheses and their prior and posterior probabilities. The efficient sequential computation of the posterior probabilities of the hypotheses, which was made possible via a non-obvious application of the percolating description-length update mechanism introduced by Bunton (see Proceedings Data Compression Conference, IEEE Computer Society Press, 1997) is described. The preliminary empirical performance of the technique on the Calgary Corpus is presented, the relationship of BSC to information-theoretic state selection and context-tree weighting is discussed
Keywords :
Bayes methods; Markov processes; information theory; probability; state estimation; trees (mathematics); Bayes theorem; Bayesian state combining; Calgary Corpus; Markov order; context models; context tree; context trees; context-tree weighting; experiments; information-theoretic state selection; m-ary input alphabets; memoryless models; minimal codes; on-line methods; percolating description-length update; posterior probability; prior probability; probability estimates; probability estimator; sequences; Automatic control; Bayesian methods; Computer science; Context modeling; Data compression; Frequency; Probability; State estimation; Stochastic processes; Writing;
fLanguage :
English
Publisher :
ieee
Conference_Titel :
Data Compression Conference, 1998. DCC '98. Proceedings
Conference_Location :
Snowbird, UT
ISSN :
1068-0314
Print_ISBN :
0-8186-8406-2
Type :
conf
DOI :
10.1109/DCC.1998.672161
Filename :
672161
Link To Document :
بازگشت