A novel Bagged Naïve Bayes-Decision Tree approach for multi-class classification problems

Journal of Intelligent & Fuzzy Systems - Tập 36 Số 3 - Trang 2261-2271 - 2019
Namrata Singh1, Pradeep Singh1
1Department of Computer Science and Engineering, National Institute of Technology Raipur, Chhattisgarh, India

Tóm tắt

Từ khóa


Tài liệu tham khảo

Kavakiotis, 2017, Machine learning and data mining methods in diabetes research, Comput Struct Biotechnol J, 15, 104, 10.1016/j.csbj.2016.12.005

Liao, 2012, Data mining techniques and applications - A decade review from 2000 to 2011, Expert Syst Appl, 39, 11303, 10.1016/j.eswa.2012.02.063

Langley, 1994, Induction of Selective Bayesian Classifiers, In: Uncertainty Proceedings, 1994, 399

Zhou, 2009, In: Encyclopedia of Biometrics, 270, 10.1007/978-0-387-73003-5_293

Rocha, 2014, Multiclass from binary: Expanding One-versus-all, one-versus-one and ECOC-based approaches, IEEE Trans Neural Networks Learn Syst, 25, 289, 10.1109/TNNLS.2013.2274735

Chaudhary, 2016, A hybrid ensemble for classification in multiclass datasets: An application to oilseed disease dataset, Comput Electron Agric, 124, 65, 10.1016/j.compag.2016.03.026

Aly, 2005, Survey on multiclass classification methods, Neural Netw, 19, 1

Silva-Palacios, 2018, Probabilistic class hierarchies for multiclass classification, J Comput Sci, 26, 254, 10.1016/j.jocs.2018.01.006

Chaudhary, 2016, An improved random forest classifier for multi-class classification, Inf Process Agric, 3, 215

García-Pedrajas, 2011, An empirical study of binary classifier fusion methods for multiclass classification, Inf Fusion, 12, 111, 10.1016/j.inffus.2010.06.010

Mousavi, 2015, A new ensemble learning methodology based on hybridization of classifier ensemble selection approaches, Appl Soft Comput J, 37, 652, 10.1016/j.asoc.2015.09.009

Agarwal, 2018, Improving multiclass classification by deep networks using DAGSVM and Triplet Loss, Pattern Recognit Lett, 112, 184, 10.1016/j.patrec.2018.06.034

Sarker, 2018, In: Australasian Conference on Data Mining, 72

Ren, 2014, Incremental naïve bayesian learning algorithm based on classification contribution degree, J Comput, 9, 1967, 10.4304/jcp.9.8.1967-1974

A. Frank and A. Asuncion , UCI Machine Learning Repository. University of California, School of Information and Computer Science, http://archive.ics.uci.edu/ml.

De Caigny, 2018, A new hybrid classification algorithm for customer churn prediction based on logistic regression and decision trees, Eur J Oper Res, 269, 760, 10.1016/j.ejor.2018.02.009

Kotsiantis, 2014, A hybrid decision tree classifier, J Intell Fuzzy Syst, 26, 327, 10.3233/IFS-120741

Carvalho, 2004, A hybrid decision tree/genetic algorithm method for data mining, Inf Sci (Ny), 163, 13, 10.1016/j.ins.2003.03.013

Panhalkar, 2017, In: Proceedings of the International Conference on Data Engineering and Communication Technology, 85, 10.1007/978-981-10-1678-3_8

Wang, 2006, Combining decision tree and Naive Bayes for classification, Knowledge-Based Syst, 19, 511, 10.1016/j.knosys.2005.10.013

Polat, 2009, A novel hybrid intelligent method based on C4. 5 decision tree classifier and one-against-all approach for multi-class classification problems, Expert Syst Appl, 36, 1587, 10.1016/j.eswa.2007.11.051

Lee, 2017, A novel bagging C4. 5 algorithm based on wrapper feature selection for supporting wise clinical decision making, J Biomed Inform, 78, 144, 10.1016/j.jbi.2017.11.005

Singh, 2018, Multi-classifier model for software fault prediction, Int Arab J Inf Technol, 15, 912

Ala’raj, 2016, Classifiers consensus system approach for credit scoring, Knowledge-Based Syst, 104, 89, 10.1016/j.knosys.2016.04.013

Sun, 2018, Imbalanced enterprise credit evaluation with DTE-SBD: Decision tree ensemble based on SMOTE and bagging with differentiated sampling rates, Inf Sci (Ny), 425, 76, 10.1016/j.ins.2017.10.017

Wu, 2015, Self-adaptive attribute weighting for Naive Bayes classification, Expert Syst Appl, 42, 1487, 10.1016/j.eswa.2014.09.019

Chandra, 2011, Robust approach for estimating probabilities in Naïve–Bayes Classifier for gene expression data, Expert Syst Appl, 38, 1293, 10.1016/j.eswa.2010.06.076

Wei, 2015, A cascade random forests algorithm for predicting protein-protein interaction sites, IEEE Trans Nanobioscience, 14, 746, 10.1109/TNB.2015.2475359

Lin, 2017, Random forests-based extreme learning machine ensemble for multi-regime time series prediction, Expert Syst Appl, 83, 164, 10.1016/j.eswa.2017.04.013

Silva-Palacios, 2017, Improving performance of multiclass classification by inducing class hierarchies, Procedia Comput Sci, 108, 1692, 10.1016/j.procs.2017.05.218

Friedman, 1997, Bayesian network classifiers, Mach Learn, 29, 131, 10.1023/A:1007465528199

L. Breiman , Random Forests, 2001, pp. 1–33.

Opitz, 1999, Popular ensemble methods: An empirical study, J Artif Intell Res, 11, 169, 10.1613/jair.614

C.L. Blake and C.J. Merz , UCI repository of machine learning databases, http://archive.ics.uci.edu/ml/index.php.

C.A. Kulikowski and S.M. Weiss , Computer systems that learn: classification and prediction methods from statistics neural nets, machine learning, and expert systems, ,Morgan Kaufmann Publishers, San Francisco, 1991.

C.E. Rasmussen , R.M. Neal , G. Hinton , D. van Camp , M. Revow , Z. Ghahramani , R. Kustra and R. Tibshirani , Delve Datasets, http://www.cs.toronto.edu/~delve/data/datasets.html.

J.H. Friedman , R. Tibshirani and T. Hastie , Datasets for “The Elements of Statistical Learning,” https://statweb.stanford.edu/~tibs/ElemStatLearn/data.html.

Breiman, 1996, Bagging predictors, Mach Learn, 24, 123, 10.1007/BF00058655

L. Breiman , Friedman , H. Jerome , R. Olshen and C.J. Stone , Classification and Regression Trees, Chapman and Hall, New York, 1984.

Breiman, 2001, Random forests, Mach Learn, 45, 5, 10.1023/A:1010933404324