skip to main content
10.1145/1389095.1389229acmconferencesArticle/Chapter ViewAbstractPublication PagesgeccoConference Proceedingsconference-collections
research-article

Pareto analysis for the selection of classifier ensembles

Published: 12 July 2008 Publication History

Abstract

The overproduce-and-choose strategy involves the generation of an initial large pool of candidate classifiers and it is intended to test different candidate ensembles in order to select the best performing solution. The ensemble's error rate, ensemble size and diversity measures are the most frequent search criteria employed to guide this selection. By applying the error rate, we may accomplish the main objective in Pattern Recognition and Machine Learning, which is to find high-performance predictors. In terms of ensemble size, the hope is to increase the recognition rate while minimizing the number of classifiers in order to meet both the performance and low ensemble size requirements. Finally, ensembles can be more accurate than individual classifiers only when classifier members present diversity among themselves. In this paper we apply two Pareto front spread quality measures to analyze the relationship between the three main search criteria used in the overproduce-and-choose strategy. Experimental results conducted demonstrate that the combination of ensemble size and diversity does not produce conflicting multi-objective optimization problems. Moreover, we cannot decrease the generalization error rate by combining this pair of search criteria. However, when the error rate is combined with diversity or the ensemble size, we found that these measures are conflicting objective functions and that the performances of the solutions are much higher.

References

[1]
M. Aksela and J. Laaksonen. Using diversity of errors for selecting members of a committee classifier. Pattern Recognition, 39(4):608--623, 2006.]]
[2]
S. Ando and E. Suzuki. Distributed multi-objective ga for generating comprehensive pareto front in deceptive optimization problems. In Proceedings of the IEEE CEC, pages 1569--1576, 2006.]]
[3]
G. Brown, J. Wyatt, R. Harris, and X. Yao. Diversity creation methods: a survey and categorisation. Information Fusion, 6(1):5--20, 2005.]]
[4]
K. Deb. Multi-Objective Optimization using Evolutionary Algorithms. John Wiley & Sons, LTD, 2001.]]
[5]
K. Deb. Unveiling innovative design principles by means of multiple conflicting objectives. Engineering Optimization, 35(5):445--470, 2003.]]
[6]
B. Gabrys and D. Ruta. Genetic algorithms in classifier fusion. Applied Soft Computing, 6(4):337--347, 2006.]]
[7]
L. Kuncheva and C. Whitaker. Measures of diversity in classifier ensembles an their relationship with the ensemble accuracy. Machine Learning, 51(2):181--207, 2003.]]
[8]
L. Oliveira, R. Sabourin, F. Bortolozzi, and C. Suen. Automatic recognition of handwritten numerical strings: A recognition and verification strategy. IEEE TPAMI, 24(11):1438--1454, 2002.]]
[9]
F. Roli, G. Giacinto, and G. Vernazza. Methods for designing multiple classifier systems. In Proceedings of MCS, pages 78--87, 2001.]]
[10]
D. Ruta and B. Gabrys. Classifier selection for majority voting. Information Fusion, 6(1):163--168, 2005.]]
[11]
G. Tremblay, R. Sabourin, and P. Maupin. Optimizing nearest neighbour in random subspaces using a multi-objective genetic algorithm. In Proceedings of ICPR, pages 208--211, Cambridge, UK, 2004.]]
[12]
C. Whitaker and L. Kuncheva. Examining the relationship between majority vote accuracy and diversity in bagging and boosting. Technical report, School of Informatics, Unversity of Wales, 2003.]]
[13]
J. Wu and S. Azarm. Metrics for quality assessment of a multiobjective design optimization solution set. Transactions ASME, Journal of Mechanical Design, 123:18--25, 2001.]]
[14]
G. Zenobi and P. Cunningham. Using diversity in preparing ensembles of classifiers based on differente feature subsets to minimize generalization error. In Proceedings of XII European Conference on Machine Learning, pages 576--587, Freiburg, Germany, 2001.]]

Cited By

View all
  • (2023)A survey of evolutionary algorithms for supervised ensemble learningThe Knowledge Engineering Review10.1017/S026988892300002438Online publication date: 1-Mar-2023
  • (2019)A Multi-objective Meta-Analytic Method for Customer Churn PredictionBusiness and Consumer Analytics: New Ideas10.1007/978-3-030-06222-4_20(781-813)Online publication date: 31-May-2019
  • (2018)Variable Hidden Neuron Ensemble for Mass Classification in Digital Mammograms [Application Notes]IEEE Computational Intelligence Magazine10.1109/MCI.2012.22285988:1(68-76)Online publication date: 17-Dec-2018
  • Show More Cited By

Index Terms

  1. Pareto analysis for the selection of classifier ensembles

    Recommendations

    Comments

    Information & Contributors

    Information

    Published In

    cover image ACM Conferences
    GECCO '08: Proceedings of the 10th annual conference on Genetic and evolutionary computation
    July 2008
    1814 pages
    ISBN:9781605581309
    DOI:10.1145/1389095
    • Conference Chair:
    • Conor Ryan,
    • Editor:
    • Maarten Keijzer
    Permission to make digital or hard copies of all or part of this work for personal or classroom use is granted without fee provided that copies are not made or distributed for profit or commercial advantage and that copies bear this notice and the full citation on the first page. Copyrights for components of this work owned by others than ACM must be honored. Abstracting with credit is permitted. To copy otherwise, or republish, to post on servers or to redistribute to lists, requires prior specific permission and/or a fee. Request permissions from [email protected]

    Sponsors

    Publisher

    Association for Computing Machinery

    New York, NY, United States

    Publication History

    Published: 12 July 2008

    Permissions

    Request permissions for this article.

    Check for updates

    Author Tags

    1. Pareto analysis
    2. classifier ensembles
    3. diversity measures
    4. ensemble selection

    Qualifiers

    • Research-article

    Conference

    GECCO08
    Sponsor:

    Acceptance Rates

    Overall Acceptance Rate 1,669 of 4,410 submissions, 38%

    Contributors

    Other Metrics

    Bibliometrics & Citations

    Bibliometrics

    Article Metrics

    • Downloads (Last 12 months)5
    • Downloads (Last 6 weeks)0
    Reflects downloads up to 08 Mar 2025

    Other Metrics

    Citations

    Cited By

    View all
    • (2023)A survey of evolutionary algorithms for supervised ensemble learningThe Knowledge Engineering Review10.1017/S026988892300002438Online publication date: 1-Mar-2023
    • (2019)A Multi-objective Meta-Analytic Method for Customer Churn PredictionBusiness and Consumer Analytics: New Ideas10.1007/978-3-030-06222-4_20(781-813)Online publication date: 31-May-2019
    • (2018)Variable Hidden Neuron Ensemble for Mass Classification in Digital Mammograms [Application Notes]IEEE Computational Intelligence Magazine10.1109/MCI.2012.22285988:1(68-76)Online publication date: 17-Dec-2018
    • (2018)Stability investigation of multi-objective heuristic ensemble classifiersInternational Journal of Machine Learning and Cybernetics10.1007/s13042-018-0789-6Online publication date: 25-Jan-2018
    • (2016)Ensemble classifiers with improved overfitting2016 1st Conference on Swarm Intelligence and Evolutionary Computation (CSIEC)10.1109/CSIEC.2016.7482130(93-97)Online publication date: Mar-2016
    • (2013)Overproduce-and-select: The grim reality2013 IEEE Symposium on Computational Intelligence and Ensemble Learning (CIEL)10.1109/CIEL.2013.6613140(52-59)Online publication date: Apr-2013
    • (2012)A multilayered ensemble architecture for the classification of masses in digital mammogramsProceedings of the 25th Australasian joint conference on Advances in Artificial Intelligence10.1007/978-3-642-35101-3_8(85-94)Online publication date: 4-Dec-2012
    • (2010)Iterative Boolean combination of classifiers in the ROC spacePattern Recognition10.1016/j.patcog.2010.03.00643:8(2732-2752)Online publication date: 1-Aug-2010
    • (2009)Multi-objective evolution of the Pareto optimal set of neural network classifier ensembles2009 International Conference on Machine Learning and Cybernetics10.1109/ICMLC.2009.5212485(74-79)Online publication date: Jul-2009

    View Options

    Login options

    View options

    PDF

    View or Download as a PDF file.

    PDF

    eReader

    View online with eReader.

    eReader

    Figures

    Tables

    Media

    Share

    Share

    Share this Publication link

    Share on social media