• DocumentCode
    659167
  • Title

    Learning joint quantizers for reconstruction and prediction

  • Author

    Raginsky, Maxim

  • Author_Institution
    Dept. of Electr. & Comput. Eng., Univ. of Illinois, Urbana, IL, USA
  • fYear
    2013
  • fDate
    9-13 Sept. 2013
  • Firstpage
    1
  • Lastpage
    5
  • Abstract
    We consider the problem of empirical design of variable-rate quantizers for reconstruction and prediction. When a discriminative model (conditional distribution of the unobserved output given the observed input) is known or can be accurately estimated from a separate training set, we show that this problem reduces to designing a certain type of a generalized quantizer by means of empirical risk minimization on unlabeled input samples only. We derive a high-probability upper bound on the resulting expected performance of such a quantizer in terms of the training sample size and the complexity parameters of the reconstruction and the prediction problems. We also discuss two illustrative examples: binary classification with absolute loss and the information bottleneck.
  • Keywords
    probability; quantisation (signal); signal reconstruction; binary classification; complexity parameter; generalized quantizer; high-probability; joint quantizers; prediction problems; reconstruction problem; risk minimization; variable-rate quantizer; Decoding; Distortion measurement; Image reconstruction; Joints; Loss measurement; Quantization (signal); Rate distortion theory;
  • fLanguage
    English
  • Publisher
    ieee
  • Conference_Titel
    Information Theory Workshop (ITW), 2013 IEEE
  • Conference_Location
    Sevilla
  • Print_ISBN
    978-1-4799-1321-3
  • Type

    conf

  • DOI
    10.1109/ITW.2013.6691290
  • Filename
    6691290