• DocumentCode
    814512
  • Title

    Any reasonable cost function can be used for a posteriori probability approximation

  • Author

    Saerens, Marco ; Latinne, Patrice ; Decaestecker, Christine

  • Author_Institution
    IRIDIA Lab., Univ. Libre de Bruxelles, Brussels, Belgium
  • Volume
    13
  • Issue
    5
  • fYear
    2002
  • fDate
    9/1/2002 12:00:00 AM
  • Firstpage
    1204
  • Lastpage
    1210
  • Abstract
    In this paper, we provide a straightforward proof of an important, but nevertheless little known, result obtained by Lindley in the framework of subjective probability theory. This result, once interpreted in the machine learning/pattern recognition context, puts new light on the probabilistic interpretation of the output of a trained classifier. A learning machine, or more generally a model, is usually trained by minimizing a criterion-the expectation of the cost function-measuring the discrepancy between the model output and the desired output. In this letter, we first show that, for the binary classification case, training the model with any "reasonable cost function" can lead to Bayesian a posteriori probability estimation. Indeed, after having trained the model by minimizing the criterion, there always exists a computable transformation that maps the output of the model to the Bayesian a posteriori probability of the class membership given the input. Then, necessary conditions allowing the computation of the transformation mapping the outputs of the model to the a posteriori probabilities are derived for the multioutput case. Finally, these theoretical results are illustrated through some simulation examples involving various cost functions.
  • Keywords
    Bayes methods; decision theory; learning (artificial intelligence); Bayes decision making; a posteriori probability; class membership; cost function; learning machine; loss function; subjective probability theory; Artificial intelligence; Artificial neural networks; Bayesian methods; Computational modeling; Cost function; Decision making; Input variables; Laboratories; Machine learning; Mean square error methods;
  • fLanguage
    English
  • Journal_Title
    Neural Networks, IEEE Transactions on
  • Publisher
    ieee
  • ISSN
    1045-9227
  • Type

    jour

  • DOI
    10.1109/TNN.2002.1031952
  • Filename
    1031952