skip to main content
10.1145/2766462.2767787acmconferencesArticle/Chapter ViewAbstractPublication PagesirConference Proceedingsconference-collections
short-paper

Challenges of Mathematical Information Retrievalin the NTCIR-11 Math Wikipedia Task

Authors Info & Claims
Published:09 August 2015Publication History

ABSTRACT

Mathematical Information Retrieval concerns retrieving information related to a particular mathematical concept. The NTCIR-11 Math Task develops an evaluation test collection for document sections retrieval of scientific articles based on human generated topics. Those topics involve a combination of formula patterns and keywords. In addition, the optional Wikipedia Task provides a test collection for retrieval of individual mathematical formula from Wikipedia based on search topics that contain exactly one formula pattern. We developed a framework for automatic query generation and immediate evaluation. This paper discusses our dataset preparation, topic generation and evaluation methods, and summarizes the results of the participants, with a special focus on the Wikipedia Task.

References

  1. Formats for topics and submissions for the math2 task at ntcir-11. Technical report, NTCIR, 2014.Google ScholarGoogle Scholar
  2. Akiko Aizawa, Michael Kohlhase, and Iadh Ounis. NTCIR-10 Math Pilot Task Overview. In Proceedings of the 10th NTCIR Conference on Evaluation of Information Access Technologies, pages 654--661, Tokyo, Japan, 2013.Google ScholarGoogle Scholar
  3. Akiko Aizawa, Michael Kohlhase, Iadh Ounis, and Moritz Schubotz. NTCIR-11 Math-2 Task Overview. In Proceedings of the 11th NTCIR Conference on Evaluation of Information Access Technologies, pages 88--98, 2014.Google ScholarGoogle Scholar
  4. Michael Kohlhase, Helena Mihaljevic-Brandt, Wolfram Sperber, and Olaf Teschke. Mathematical Formula Search. pages 56--57, September 2013.Google ScholarGoogle Scholar
  5. Michael Kohlhase, Corneliu Prodescu, and Christian Liguda. Xlsearch: A search engine for spreadsheets. In Simon Thorne et. al, editor, Proceedings of the EuSpRIG 2013 Conference "Spreadsheet Risk Management". July 4--5, London, United Kingdom, pages 47--58. Five Star Printing Ldt, Claydon, 2013.Google ScholarGoogle Scholar
  6. Matthias S. Reichenbach, Anurag Agarwal, and Richard Zanibbi. Rendering expressions to improve accuracy of relevance assessment for math search. Proceedings of the 37th international ACM SIGIR conference on Research & development in information retrieval - SIGIR '14, pages 851--854, 2014. Google ScholarGoogle ScholarDigital LibraryDigital Library
  7. Moritz Schubotz and Gabriel Wicke. Mathoid: Robust, scalable, fast and accessible math rendering for wikipedia. In Stephen Watt et al., editor, Intelligent Computer Mathematics, volume 8543 of Lecture Notes in Computer Science, pages 224--235. Springer International Publishing, 2014.Google ScholarGoogle ScholarCross RefCross Ref
  8. Heinrich Stamerjohanns, Michael Kohlhase, Deyan Ginev, Catalin David, and Bruce Miller. Transforming large collections of scientific publications to xml. Mathematics in Computer Science, 3(3):299--307, 2010.Google ScholarGoogle ScholarCross RefCross Ref
  9. Ellen M. Voorhees. The TREC-8 Question Answering Track Report. TREC, 1999.Google ScholarGoogle Scholar
  10. Keita Del Valle Wangari, Richard Zanibbi, and Anurag Agarwal. Discovering real-world use cases for a multimodal math search interface. Proceedings of the 37th international ACM SIGIR conference on Research & development in information retrieval - SIGIR '14, pages 947--950, 2014. Google ScholarGoogle ScholarDigital LibraryDigital Library

Recommendations

Comments

Login options

Check if you have access through your login credentials or your institution to get full access on this article.

Sign in
  • Published in

    cover image ACM Conferences
    SIGIR '15: Proceedings of the 38th International ACM SIGIR Conference on Research and Development in Information Retrieval
    August 2015
    1198 pages
    ISBN:9781450336215
    DOI:10.1145/2766462

    Copyright © 2015 ACM

    Permission to make digital or hard copies of all or part of this work for personal or classroom use is granted without fee provided that copies are not made or distributed for profit or commercial advantage and that copies bear this notice and the full citation on the first page. Copyrights for components of this work owned by others than the author(s) must be honored. Abstracting with credit is permitted. To copy otherwise, or republish, to post on servers or to redistribute to lists, requires prior specific permission and/or a fee. Request permissions from [email protected].

    Publisher

    Association for Computing Machinery

    New York, NY, United States

    Publication History

    • Published: 9 August 2015

    Permissions

    Request permissions about this article.

    Request Permissions

    Check for updates

    Qualifiers

    • short-paper

    Acceptance Rates

    SIGIR '15 Paper Acceptance Rate70of351submissions,20%Overall Acceptance Rate792of3,983submissions,20%

PDF Format

View or Download as a PDF file.

PDF

eReader

View online with eReader.

eReader