skip to main content
10.1145/1277741.1277811acmconferencesArticle/Chapter ViewAbstractPublication PagesirConference Proceedingsconference-collections
Article

Feature selection for ranking

Authors Info & Claims
Published:23 July 2007Publication History

ABSTRACT

Ranking is a very important topic in information retrieval. While algorithms for learning ranking models have been intensively studied, this is not the case for feature selection, despite of its importance. The reality is that many feature selection methods used in classification are directly applied to ranking. We argue that because of the striking differences between ranking and classification, it is better to develop different feature selection methods for ranking. To this end, we propose a new feature selection method in this paper. Specifically, for each feature we use its value to rank the training instances, and define the ranking accuracy in terms of a performance measure or a loss function as the importance of the feature. We also define the correlation between the ranking results of two features as the similarity between them. Based on the definitions, we formulate the feature selection issue as an optimization problem, for which it is to find the features with maximum total importance scores and minimum total similarity scores. We also demonstrate how to solve the optimization problem in an efficient way. We have tested the effectiveness of our feature selection method on two information retrieval datasets and with two ranking models. Experimental results show that our method can outperform traditional feature selection methods for the ranking task.

References

  1. R. Battiti. Using mutual information for selecting features in supervised neural net learning. IEEE Transactions on Neural Networks. vol. 5, NO.4, July 1994.Google ScholarGoogle ScholarDigital LibraryDigital Library
  2. P. Borlund. The concept of relevance in IR. Journal of the American Society for Information Science and Technology 54(10): 913--925, 2003. Google ScholarGoogle ScholarDigital LibraryDigital Library
  3. L. Breiman, J. H. Friedman, R. A. Olshen, and C.J.Stone. Classification and regression trees. Wadsworth and Brooks, 1984.Google ScholarGoogle Scholar
  4. C. Burges, T. Shaked, E. Renshaw, A .Lazier, M. Deeds, N. Hamilton, G. Hullender. Learning to rank using gradient descent. ICML 2005. Google ScholarGoogle ScholarDigital LibraryDigital Library
  5. A. Blum and P. Langley. Selection of relevant features and examples in machine learning. AI, 97(1-2), 1997. Google ScholarGoogle ScholarDigital LibraryDigital Library
  6. Y. Cao, J. Xu, T. Y. Liu, H. Li, Y. Huang, H. W. Hon. Adapting ranking SVM to document retrieval. SIGIR 2006. Google ScholarGoogle ScholarDigital LibraryDigital Library
  7. G. Forman. An extensive empirical study of feature selection metrics for text classification. Journal of Machine Learning Research, 2003. Google ScholarGoogle ScholarDigital LibraryDigital Library
  8. I. Guyon, A. Elisseeff. An introduction to variable and feature selection. Journal of Machine Learning Research, 2003. Google ScholarGoogle ScholarDigital LibraryDigital Library
  9. W. Hersh, C. Buckley, T. J. Leone, and D. Hick-man. OHSUMED: an interactive retrieval evaluation and new large text collection for research. SIGIR 1994. Google ScholarGoogle ScholarDigital LibraryDigital Library
  10. R. Herbrich, T. Graepel, and K. Obermayer. Large margin rank boundaries for ordinal regression. Advances in Large Margin Classifiers, MIT Press, Pages: 115--132, 2000.Google ScholarGoogle Scholar
  11. K. Jarvelin and J. Kekalainen. Cumulated grain-based evaluation of IR techniques, ACM Transactions on Information Systems, 2002. Google ScholarGoogle ScholarDigital LibraryDigital Library
  12. T. Joachims. Making large-scale SVM learning practical. Advances in Kernel Methods - Support Vector Learning, B. Schölkopf and C. Burges and A. Smola (ed.), MIT-Press, 1999. Google ScholarGoogle ScholarDigital LibraryDigital Library
  13. T. Joachims. Optimizing search engines using clickthrough data. KDD 2002. Google ScholarGoogle ScholarDigital LibraryDigital Library
  14. N. Kwak, C. H. Choi. Input feature selection for classification problems. Neural Networks, IEEE Transactions on Neural Networks, vol.13, No.1, January 2002. Google ScholarGoogle ScholarDigital LibraryDigital Library
  15. R. Kohavi, G. H. John. Wrappers for feature selection. Artificial Intelligence, 1997. Google ScholarGoogle ScholarDigital LibraryDigital Library
  16. M. Kendall. Rank correlation methods. Oxford University Press, 1990.Google ScholarGoogle Scholar
  17. J. Lafferty and C. Zhai. Document language models, query models, and risk minimization for Information Retrieval. SIGIR 2001. Google ScholarGoogle ScholarDigital LibraryDigital Library
  18. A. M. Liebetrau. Measures of association, volume 32 of Quantitative Applications in the Social Sciences. Sage Publications, Inc., 1983.Google ScholarGoogle Scholar
  19. W. Lior, S. Bileschi. Combining variable selection with dimensionality reduction. CVPR 2005.Google ScholarGoogle Scholar
  20. D. Mladenic and M. Grobelnik. Feature selection for unbalanced class distribution and Naíve Bayes. ICML 1999. Google ScholarGoogle ScholarDigital LibraryDigital Library
  21. R. Nallapati. Discriminative models for information retrieval. SIGIR 2004. Google ScholarGoogle ScholarDigital LibraryDigital Library
  22. A. Y. Ng. Feature selection, L1 vs. L2 regularization, and rotational invariance. ICML 2004. Google ScholarGoogle ScholarDigital LibraryDigital Library
  23. J. Ponte and W. B. Croft. A language model approach to information retrieval. SIGIR 1998. Google ScholarGoogle ScholarDigital LibraryDigital Library
  24. T. Qin, T. Y. Liu, X. D. Zhang, Z. Chen, and W. Y. Ma. A study of relevance propagation for web search. SIGIR 2005. Google ScholarGoogle ScholarDigital LibraryDigital Library
  25. S. Robertson. Overview of the okapi projects, Journal of Documentation, Vol. 53, No. 1, pp. 3--7, 1997.Google ScholarGoogle ScholarCross RefCross Ref
  26. S. Robertson and D.A. Hull. The TREC-9 Filtering Track Final Report. Proceeding of the 9th Text Retrieval Conference, pages 25--40, 2000.Google ScholarGoogle Scholar
  27. S. Theodoridis, K. Koutroumbas. Pattern recognition. Academic Press, New York, 1999. Google ScholarGoogle ScholarDigital LibraryDigital Library
  28. E.M. Voorhees and D.K. Harman. TREC: experiment and evaluation in Information Retrieval. MIT Press, 2005. Google ScholarGoogle ScholarDigital LibraryDigital Library
  29. J. Weston, S. Mukherjee, O. Chapelle, M. Pontil, T. Poggio and V. Vapnik. Feature selection for SVMs. NIPS 2001.Google ScholarGoogle Scholar
  30. G. R. Xue, Q. Yang, H. J. Zeng, Y. Yu, and Z. Chen. Exploiting the hierarchical structure for link analysis. SIGIR 2005. Google ScholarGoogle ScholarDigital LibraryDigital Library
  31. Y. Yang and Jan O. Pedersen. A comparative study on feature selection in text categorization. ICML 1997. Google ScholarGoogle ScholarDigital LibraryDigital Library
  32. R.B. Yates, B.R. Neto. Modern information retrieval, Addison Wesley, 1999. Google ScholarGoogle ScholarDigital LibraryDigital Library
  33. MSN, http://www.msn.com.Google ScholarGoogle Scholar

Index Terms

  1. Feature selection for ranking

      Recommendations

      Comments

      Login options

      Check if you have access through your login credentials or your institution to get full access on this article.

      Sign in
      • Published in

        cover image ACM Conferences
        SIGIR '07: Proceedings of the 30th annual international ACM SIGIR conference on Research and development in information retrieval
        July 2007
        946 pages
        ISBN:9781595935977
        DOI:10.1145/1277741

        Copyright © 2007 ACM

        Permission to make digital or hard copies of all or part of this work for personal or classroom use is granted without fee provided that copies are not made or distributed for profit or commercial advantage and that copies bear this notice and the full citation on the first page. Copyrights for components of this work owned by others than ACM must be honored. Abstracting with credit is permitted. To copy otherwise, or republish, to post on servers or to redistribute to lists, requires prior specific permission and/or a fee. Request permissions from [email protected]

        Publisher

        Association for Computing Machinery

        New York, NY, United States

        Publication History

        • Published: 23 July 2007

        Permissions

        Request permissions about this article.

        Request Permissions

        Check for updates

        Qualifiers

        • Article

        Acceptance Rates

        Overall Acceptance Rate792of3,983submissions,20%

      PDF Format

      View or Download as a PDF file.

      PDF

      eReader

      View online with eReader.

      eReader