skip to main content
10.1145/775047.775093acmconferencesArticle/Chapter ViewAbstractPublication PageskddConference Proceedingsconference-collections
Article

Efficient handling of high-dimensional feature spaces by randomized classifier ensembles

Authors Info & Claims
Published:23 July 2002Publication History

ABSTRACT

Handling massive datasets is a difficult problem not only due to prohibitively large numbers of entries but in some cases also due to the very high dimensionality of the data. Often, severe feature selection is performed to limit the number of attributes to a manageable size, which unfortunately can lead to a loss of useful information. Feature space reduction may well be necessary for many stand-alone classifiers, but recent advances in the area of ensemble classifier techniques indicate that overall accurate classifier aggregates can be learned even if each individual classifier operates on incomplete "feature view" training data, i.e., such where certain input attributes are excluded. In fact, by using only small random subsets of features to build individual component classifiers, surprisingly accurate and robust models can be created. In this work we demonstrate how these types of architectures effectively reduce the feature space for submodels and groups of sub-models, which lends itself to efficient sequential and/or parallel implementations. Experiments with a randomized version of Adaboost are used to support our arguments, using the text classification task as an example.

References

  1. I. Aleksander and T. J. Stonham. A guide to pattern recognition using random-access memories, IEE Proceedings-E Computers and Digital Techniques, 2(1):29--40, 1979.]]Google ScholarGoogle ScholarCross RefCross Ref
  2. I. Aleksander, W. Thomas, and P. Bowden. WISARD, a radical new step forward in image recognition. Sensor Rev., 4(3):120--124, 1984.]]Google ScholarGoogle Scholar
  3. Y. Amit, G. Blanchard, and K. Wilder. Multiple randomized classifiers: MRCL. Technical Report 446, Depertment of Statistics, University of Chicago, 2000.]]Google ScholarGoogle Scholar
  4. C. Apté, F. Damerau, and S. M. Weiss. Automated learning of decision rules for text categorization. ACM Transactions on Information Systems, 12(3):233--251, 1994.]] Google ScholarGoogle ScholarDigital LibraryDigital Library
  5. W. Bledsoe and I. Browning. Pattern recognition and reading by machine. In IRE Joint Computer Conference, pages 225--232, 1959.]]Google ScholarGoogle ScholarDigital LibraryDigital Library
  6. L. Breiman. Bagging predictors. Machine Learning, 24(2):123--140, 1996.]] Google ScholarGoogle ScholarCross RefCross Ref
  7. L. Breiman. Arcing classifiers. The Annals of Statistics, 26(3):801--849, 1998.]]Google ScholarGoogle ScholarCross RefCross Ref
  8. L. Breiman. Random forests. Machine Learning, 24(2):5--32, 2001.]] Google ScholarGoogle ScholarDigital LibraryDigital Library
  9. T. G. Dietterich. An experimental comparison of three methods of constructing ensembles of decision trees: Bagging, boosting, and randomization. Machine Learning, 40(2):139--157, 2000.]] Google ScholarGoogle ScholarDigital LibraryDigital Library
  10. C. Domingo and O. Watanabe. Scaling up a boosting-based learner via adaptive sampling. In Proceedings of the 2000 Pacific-Asia Conference on Knowledge Discovery and Data Mining(PAKDD-2000), pages 317--328, 2000.]] Google ScholarGoogle ScholarDigital LibraryDigital Library
  11. Y. Freund and R. E. Schapire. Experiments with a new boosting algorithm. In Proceedings of the Thirteenth International Machine Learning Conference, pages 148--156, 1996.]]Google ScholarGoogle ScholarDigital LibraryDigital Library
  12. J. Friedman, T. Hastie, and R. Tibshirani. Additive logistic regression: a statistical view of boosting. The Annals of Statistics, 38(2):337--374, 2000.]]Google ScholarGoogle ScholarCross RefCross Ref
  13. D. Pavlov, J. Mao, and B. Dom. Scaling-up support vector machines using boosting algorithm. In Proceedings of the 2000 International Conference on Pattern Recognition, 2000.]]Google ScholarGoogle ScholarCross RefCross Ref
  14. J. A. Reichler and H. D. Harris. Parallel online continuous arcing and a new framework for wrapping parallel ensembles. In Proceedings of IJCAI 2001: International Joint Conference on Artificial Intelligence, Workshop on Wrappers for Performance Enhancement in Knowledge Discovery in Databases, pages 148--156, 2001.]]Google ScholarGoogle Scholar
  15. R. Rohwer and M. Morciniec. The theoretical and experimental status of the N-tuple classifier. Neural Networkas, 11(1):1--14, 1998.]] Google ScholarGoogle ScholarDigital LibraryDigital Library
  16. R. E. Schapire and Y. Singer. BoosTexter: A boosting-based system for text categorization. Machine Learning, 39(2--3):135--168, 2000.]] Google ScholarGoogle ScholarDigital LibraryDigital Library
  17. V. N. Vapnik. Statistical Learning Theory. John Wiley, New York, 1998.]]Google ScholarGoogle ScholarDigital LibraryDigital Library
  18. Y. Yang and X. Liu. A re-examination of text categorization methods. In Proceedings of the 22nd International ACM SIGIR Conference on Research and Development in Information Retrieval, pages 42--49, 1999.]] Google ScholarGoogle ScholarDigital LibraryDigital Library
  19. Y. Yang and J. P. Pedersen. A comparative study on feature selection in text categorization. In Proceedings of the Fourteenth International Conference on Machine Learning (ICML'97), pages 412--420, 1997.]] Google ScholarGoogle ScholarDigital LibraryDigital Library
  20. C. Yu and D. B. Skillicorn. Parallelizing boosting and bagging. Technical Report 2001-442, Depertment of Computing and Information Science, Queen's University, Kingston, Canada, 2001.]]Google ScholarGoogle Scholar

Index Terms

  1. Efficient handling of high-dimensional feature spaces by randomized classifier ensembles

              Recommendations

              Comments

              Login options

              Check if you have access through your login credentials or your institution to get full access on this article.

              Sign in
              • Published in

                cover image ACM Conferences
                KDD '02: Proceedings of the eighth ACM SIGKDD international conference on Knowledge discovery and data mining
                July 2002
                719 pages
                ISBN:158113567X
                DOI:10.1145/775047

                Copyright © 2002 ACM

                Permission to make digital or hard copies of all or part of this work for personal or classroom use is granted without fee provided that copies are not made or distributed for profit or commercial advantage and that copies bear this notice and the full citation on the first page. Copyrights for components of this work owned by others than ACM must be honored. Abstracting with credit is permitted. To copy otherwise, or republish, to post on servers or to redistribute to lists, requires prior specific permission and/or a fee. Request permissions from [email protected]

                Publisher

                Association for Computing Machinery

                New York, NY, United States

                Publication History

                • Published: 23 July 2002

                Permissions

                Request permissions about this article.

                Request Permissions

                Check for updates

                Qualifiers

                • Article

                Acceptance Rates

                KDD '02 Paper Acceptance Rate44of307submissions,14%Overall Acceptance Rate1,133of8,635submissions,13%

                Upcoming Conference

                KDD '24

              PDF Format

              View or Download as a PDF file.

              PDF

              eReader

              View online with eReader.

              eReader