ABSTRACT
Adaptive boosting (AdaBoost) is a method for building classification ensemble, which combines multiple classifiers built in an iterative process of reweighting instances. This method proves to be a very effective classification method, therefore it was the major part of our evolutionary inspired classification algorithm.
In this paper, we introduce the Genetic Programming with AdaBoost (GPAB) which combines the induction of classification trees with genetic programming (GP) and AdaBoost for multiple class problems. Our method GPAB builds the ensemble of classification trees and uses AdaBoost through the evolution to weight instances and individual trees.
To evaluate the potential of the proposed evolutionary method, we made an experiment where we compared the GPAB with Random Forest and AdaBoost on several standard UCI classification benchmarks. The results show that GPAB improves classification accuracy in comparison to other two classifiers.
- Rodrigo Coelho Barros, Marcio Porto Basgalupp, ACPLF De Carvalho, and Alex Alves Freitas. 2012. A survey of evolutionary algorithms for decision-tree induction. Systems, Man, and Cybernetics, Part C: Applications and Reviews, IEEE Transactions on 42, 3 (2012), 291--312. Google ScholarDigital Library
- Leo Breiman. 1996. Bias, Variance, and Arcing Classifiers. Technical Report. Statistics Department, University of California at Berkeley.Google Scholar
- Tianqi Chen and Carlos Guestrin. 2016. XGBoost: A Scalable Tree Boosting System. In Proceedings of the 22nd ACM SIGKDD International Conference on Knowledge Discovery and Data Mining. ACM, New York, NY, USA, 785--794. Google ScholarDigital Library
- P.G. Espejo, S. Ventura, and F. Herrera. 2010. A Survey on the Application of Genetic Programming to Classification. IEEE Transactions on Systems, Man, and Cybernetics, Part C (Applications and Reviews) 40, 2 (March 2010), 121--144. Google ScholarDigital Library
- Yoav Freund and Robert E. Schapire. 1996. Experiments with a New Boosting Algorithm. In Proceedings of the Thirteenth International Conference on International Conference on Machine Learning (ICML'96). Morgan Kaufmann Publishers Inc., San Francisco, CA, USA, 148--156. Google ScholarDigital Library
- Jerome Friedman, Trevor Hastie, Robert Tibshirani, et al. 2000. Additive logistic regression: a statistical view of boosting (with discussion and a rejoinder by the authors). The annals of statistics 28, 2 (2000), 337--407.Google Scholar
- Muhammad Iqbal, Bing Xue, Harith Al-Sahaf, and Mengjie Zhang. 2017. Cross-domain reuse of extracted knowledge in genetic programming for image classification. IEEE Transactions on Evolutionary Computation 21, 4 (2017), 569--587.Google ScholarDigital Library
- Andy Liaw and Matthew Wiener. 2002. Classification and regression by random-Forest. R news 2, 3 (2002), 18--22.Google Scholar
- V. Podgorelec, S. Karakatič, R. C. Barros, and M. P. Basgalupp. 2015. Evolving balanced decision trees with a multi-population genetic algorithm. In 2015 IEEE Congress on Evolutionary Computation (CEC). IEEE, Piscataway, NJ, 54--61.Google Scholar
- Vili Podgorelec, Matej Šprogar, and Sandi Pohorec. 2013. Evolutionary design of decision trees. Wiley Interdisciplinary Reviews: Data Mining and Knowledge Discovery 3, 2 (2013), 63--82.Google ScholarCross Ref
Index Terms
- Building boosted classification tree ensemble with genetic programming
Recommendations
An automated ensemble learning framework using genetic programming for image classification
GECCO '19: Proceedings of the Genetic and Evolutionary Computation ConferenceAn ensemble consists of multiple learners and can achieve a better generalisation performance than a single learner. Genetic programming (GP) has been applied to construct ensembles using different strategies such as bagging and boosting. However, no GP-...
Ensemble Genetic Programming
Genetic ProgrammingAbstractEnsemble learning is a powerful paradigm that has been used in the top state-of-the-art machine learning methods like Random Forests and XGBoost. Inspired by the success of such methods, we have developed a new Genetic Programming method called ...
Comparing ensemble learning approaches in genetic programming for classification with unbalanced data
GECCO '13 Companion: Proceedings of the 15th annual conference companion on Genetic and evolutionary computationThis paper compares three approaches to evolving ensembles in Genetic Programming (GP) for binary classification with unbalanced data. The first uses bagging with sampling, while the other two use Pareto-based multi-objective GP (MOGP) for the trade-off ...
Comments