Any reasonable cost function can be used for a posteriori probability approximation

IEEE Transactions on Neural Networks - Tập 13 Số 5 - Trang 1204-1210 - 2002
M. Saerens1,2, P. Latinne1, C. Decaestecker3
1IRIDIA Laboratory (Artificial Intelligence Laboratory, Université Libre de Bruxelles, Brussels, Belgium
2SmalS-MvM, Brussels, Belgium
3Belgian Research Founds (FNRS), Université Libre de Bruxelles, Brussels, Belgium

Tóm tắt

In this paper, we provide a straightforward proof of an important, but nevertheless little known, result obtained by Lindley in the framework of subjective probability theory. This result, once interpreted in the machine learning/pattern recognition context, puts new light on the probabilistic interpretation of the output of a trained classifier. A learning machine, or more generally a model, is usually trained by minimizing a criterion-the expectation of the cost function-measuring the discrepancy between the model output and the desired output. In this letter, we first show that, for the binary classification case, training the model with any "reasonable cost function" can lead to Bayesian a posteriori probability estimation. Indeed, after having trained the model by minimizing the criterion, there always exists a computable transformation that maps the output of the model to the Bayesian a posteriori probability of the class membership given the input. Then, necessary conditions allowing the computation of the transformation mapping the outputs of the model to the a posteriori probabilities are derived for the multioutput case. Finally, these theoretical results are illustrated through some simulation examples involving various cost functions.

Từ khóa

#Cost function #Laboratories #Machine learning #Bayesian methods #Mean square error methods #Artificial intelligence #Computational modeling #Decision making #Artificial neural networks #Input variables

Tài liệu tham khảo

10.2307/1402456 hampshire, 1990, equivalence proofs for multi-layer perceptron classifiers and the bayesian discriminant function, Proc 1990 Connectionist Models Summer School, 159 10.1109/72.883416 mccullagh, 1990, Generalized Linear Models 10.1162/neco.1991.3.4.461 wolfram, 1999, The Mathematica Book 10.1109/18.243457 10.1109/IJCNN.1991.155295 10.1007/978-1-4419-8746-4 saerens, 1996, non mean square error criteria for the training of learning machines, Proc 13th Int Conf Machine Learning (ICML), 427 bishop, 1995, Neural Networks for Pattern Recognition