Minimax strategies for training classifiers under unknown priors
Tóm tắt
Most supervised learning algorithms are based on the assumption that the training data set reflects the underlying statistical model of the real data. However, this stationarity assumption is not always satisfied in practice: quite frequently, class prior probabilities are not in accordance with the class proportions in the training data set. The minimax approach is based on selecting the classifier that minimize the error probability under the worst case conditions. We propose a two-step learning algorithm to train a neural network in order to estimate the minimax classifier that is robust to changes in the class priors. During the first step, posterior probabilities based on training data priors are estimated. During the second step, class priors are modified in order to minimize a cost function that is asymptotically equivalent to the worst-case error rate. This procedure is illustrated on a softmax-based neural network. Several experimental results show the advantages of the proposed method with respect to other approaches.
Từ khóa
#Minimax techniques #Neural networks #Training data #Error analysis #Robustness #Supervised learning #Error probability #Cost function #ProposalsTài liệu tham khảo
10.1145/312129.312285
guerrero-curieses, 2000, A Natural Approach to Sample Selection in Binary Classification, Proc of the Learning'2000 International Conference
weiss, 2001, The effect of class distribution on classifier learning, Tech Report ML-TR 43
lawrence, 1998, Neural Networks Classification and Prior Class Probabilities, Tricks of the Trade Lecture Notes in Computer Science State-of-the-Art Surveys, 299, 10.1007/3-540-49430-8_15
kubat, 1997, Addressing the curse of imbalanced training sets: One-sided selection, Proceedings of the 14th International Conference on Machine Learning, 179
10.1162/089976602753284446
10.1109/18.243457
10.1109/72.925549
van trees, 1968, Detection Estimation and Modulation Theory
10.1109/NNSP.1999.788145
