skip to main content
10.1145/2909132.2909252acmconferencesArticle/Chapter ViewAbstractPublication PagesaviConference Proceedingsconference-collections
research-article
Public Access

Assessing Topic Representations for Gist-Forming

Authors Info & Claims
Published:07 June 2016Publication History

ABSTRACT

As topic modeling has grown in popularity, tools for visualizing the process have become increasingly common. Though these tools support a variety of different tasks, they generally have a view or module that conveys the contents of an individual topic. These views support the important task of gist-forming: helping the user build a cohesive overall sense of the topic's semantic content that can be generalized outside the specific subset of words that are shown. There are a number of factors that affect these views, including the visual encoding used, the number of topic words included, and the quality of the topics themselves. To our knowledge, there has been no formal evaluation comparing the ways in which these factors might change users' interpretations. In a series of crowdsourced experiments, we sought to compare features of visual topic representations in their suitability for gist-forming. We found that gist-forming ability is remarkably resistant to changes in visual representation, though it deteriorates with topics of lower quality.

References

  1. E. Alexander, J. Kohlmann, R. Valenza, M. Witmore, and M. Gleicher. Serendip: Topic model-driven visual exploration of text corpora. In Visual Analytics Science and Technology (VAST), 2014 IEEE Conference on, pages 173--182. IEEE, 2014.Google ScholarGoogle ScholarCross RefCross Ref
  2. L. AlSumait, D. Barbará, J. Gentle, and C. Domeniconi. Topic significance ranking of lda generative models. In Machine Learning and Knowledge Discovery in Databases, pages 67--82. Springer, 2009.Google ScholarGoogle ScholarCross RefCross Ref
  3. D. Blei, A. Ng, and M. Jordan. Latent dirichlet allocation. J. Machine Learning Research, 3:993--1022, 2003. Google ScholarGoogle ScholarDigital LibraryDigital Library
  4. M. Bostock, V. Ogievetsky, and J. Heer. D3: Data-driven documents. IEEE TVCG, 2011. Google ScholarGoogle ScholarDigital LibraryDigital Library
  5. A. Chaney and D. Blei. Visualizing topic models. In Proc. AAAI on Weblogs and Social Media, 2012.Google ScholarGoogle Scholar
  6. J. Chang, S. Gerrish, C. Wang, J. L. Boyd-graber, and D. M. Blei. Reading tea leaves: How humans interpret topic models. In Advances in neural information processing systems, pages 288--296, 2009.Google ScholarGoogle ScholarDigital LibraryDigital Library
  7. J. Chuang, C. Manning, and J. Heer. Termite: visualization techniques for assessing textual topic models. In Proc. Advanced Visual Interfaces, pages 74--77. ACM, 2012. Google ScholarGoogle ScholarDigital LibraryDigital Library
  8. C. Collins, F. B. Viégas, and M. Wattenberg. Parallel tag clouds to explore and analyze facted text corpora. In Proc. of the IEEE Symp. on Visual Analytics Science and Technology (VAST), 2009.Google ScholarGoogle Scholar
  9. M. Correll and M. Gleicher. Error bars considered harmful: Exploring alternate encodings for mean and error. IEEE Transactions on Visualization and Computer Graphics, 20(12):2142--2151, dec 2014. IEEE Vis Conference, InfoVis track, to appear.Google ScholarGoogle ScholarCross RefCross Ref
  10. W. Cui, S. Liu, L. Tan, C. Shi, Y. Song, Z. Gao, H. Qu, and X. Tong. Textow: Towards better understanding of evolving topics in text. IEEE TVCG, 17(12):2412--2421, 2011. Google ScholarGoogle ScholarDigital LibraryDigital Library
  11. J. Davies. d3-cloud. https://github.com/jasondavies/d3-cloud, 2015.Google ScholarGoogle Scholar
  12. M. J. Halvey and M. T. Keane. An assessment of tag presentation techniques. In Proceedings of the 16th international conference on World Wide Web, pages 1313--1314. ACM, 2007. Google ScholarGoogle ScholarDigital LibraryDigital Library
  13. J. Harris. Word clouds considered harmful, blog, http://www.niemanlab.org/2011/10/word-clouds-considered-harmful/, 2011.Google ScholarGoogle Scholar
  14. M. A. Hearst and D. Rosner. Tag clouds: Data analysis tool or social signaller? In Hawaii International Conference on System Sciences, Proceedings of the 41st Annual, pages 160--160. IEEE, 2008. Google ScholarGoogle ScholarDigital LibraryDigital Library
  15. S. Lohmann, J. Ziegler, and L. Tetzlaff. Comparison of tag cloud layouts: Task-related performance and visual exploration. In Human-Computer Interaction--INTERACT 2009, pages 392--404. Springer, 2009. Google ScholarGoogle ScholarDigital LibraryDigital Library
  16. E. Meeks. Using word clouds for topic modeling results, blog, https://dhs.stanford.edu/algorithmic-literacy/using-word-clouds-for-topic-modeling-results/, 2012.Google ScholarGoogle Scholar
  17. R. Řehůřek and P. Sojka. Software Framework for Topic Modelling with Large Corpora. In Proceedings of the LREC 2010 Workshop on New Challenges for NLP Frameworks, pages 45--50, Valletta, Malta, May 2010. ELRA. http://is.muni.cz/publication/884893/en.Google ScholarGoogle Scholar
  18. A. W. Rivadeneira, D. M. Gruen, M. J. Muller, and D. R. Millen. Getting our head in the clouds: toward evaluation studies of tagclouds. In Proceedings of the SIGCHI conference on Human factors in computing systems, pages 995--998. ACM, 2007. Google ScholarGoogle ScholarDigital LibraryDigital Library
  19. E. Sandhaus. The New York Times Annotated Corpus LDC2008T19. DVD. Philadelphia: Linguistic Data Consortium, 2008.Google ScholarGoogle Scholar
  20. A. J. Torget, R. Mihalcea, J. Christensen, and G. McGhee. Mapping texts: Combining text-mining and geo-visualization to unlock the research potential of historical newspapers. 2011.Google ScholarGoogle Scholar
  21. T. van der Geest and R. van Dongelen. What is beautiful is useful-visual appeal and expected information quality. In Professional Communication Conference, 2009. IPCC 2009. IEEE International, pages 1--5. IEEE, 2009.Google ScholarGoogle ScholarCross RefCross Ref
  22. F. B. Viégas and M. Wattenberg. Timelines tag clouds and the case for vernacular visualization. interactions, 15(4):49--52, 2008. Google ScholarGoogle ScholarDigital LibraryDigital Library
  23. F. Wei, S. Liu, Y. Song, S. Pan, M. X. Zhou, W. Qian, L. Shi, L. Tan, and Q. Zhang. Tiara: a visual exploratory text analytic system. In Proc. ACM Knowledge discovery and data mining, pages 153--162. ACM, 2010. Google ScholarGoogle ScholarDigital LibraryDigital Library

Recommendations

Comments

Login options

Check if you have access through your login credentials or your institution to get full access on this article.

Sign in
  • Published in

    cover image ACM Conferences
    AVI '16: Proceedings of the International Working Conference on Advanced Visual Interfaces
    June 2016
    400 pages
    ISBN:9781450341318
    DOI:10.1145/2909132

    Copyright © 2016 ACM

    Permission to make digital or hard copies of all or part of this work for personal or classroom use is granted without fee provided that copies are not made or distributed for profit or commercial advantage and that copies bear this notice and the full citation on the first page. Copyrights for components of this work owned by others than the author(s) must be honored. Abstracting with credit is permitted. To copy otherwise, or republish, to post on servers or to redistribute to lists, requires prior specific permission and/or a fee. Request permissions from [email protected].

    Publisher

    Association for Computing Machinery

    New York, NY, United States

    Publication History

    • Published: 7 June 2016

    Permissions

    Request permissions about this article.

    Request Permissions

    Check for updates

    Qualifiers

    • research-article
    • Research
    • Refereed limited

    Acceptance Rates

    AVI '16 Paper Acceptance Rate20of96submissions,21%Overall Acceptance Rate128of490submissions,26%

PDF Format

View or Download as a PDF file.

PDF

eReader

View online with eReader.

eReader