DocumentCode
659130
Title
Information-theoretic limits on the classification of Gaussian mixtures: Classification on the Grassmann manifold
Author
Nokleby, Matthew ; Calderbank, R. ; Rodrigues, Miguel R. D.
Author_Institution
Duke Univ., Durham, NC, USA
fYear
2013
fDate
9-13 Sept. 2013
Firstpage
1
Lastpage
5
Abstract
Motivated by applications in high-dimensional signal processing, we derive fundamental limits on the performance of compressive linear classifiers. By analogy with Shannon theory, we define the classification capacity, which quantifies the maximum number of classes that can be discriminated with low probability of error, and the diversity-discrimination tradeoff, which quantifies the tradeoff between the number of classes and the probability of classification error. For classification of Gaussian mixture models, we identify a duality between classification and communications over non-coherent multiple-antenna channels. This duality allows us to characterize the classification capacity and diversity-discrimination tradeoff using existing results from multiple-antenna communication. We also identify the easiest possible classification problems, which correspond to low-dimensional subspaces drawn from an appropriate Grassmann manifold.
Keywords
Gaussian processes; error statistics; mixture models; signal classification; Gaussian mixture models; Grassmann manifold; Shannon theory; classification capacity; compressive linear classifiers; diversity-discrimination tradeoff; high-dimensional signal processing; information-theoretic limits; low-dimensional subspaces; multiple-antenna communication; noncoherent multiple-antenna channels; probability of classification error; Capacity planning; Coherence; Indexes; MIMO; Manifolds; Receiving antennas; Signal to noise ratio;
fLanguage
English
Publisher
ieee
Conference_Titel
Information Theory Workshop (ITW), 2013 IEEE
Conference_Location
Sevilla
Print_ISBN
978-1-4799-1321-3
Type
conf
DOI
10.1109/ITW.2013.6691253
Filename
6691253
Link To Document