skip to main content
10.1145/3170427.3188597acmconferencesArticle/Chapter ViewAbstractPublication PageschiConference Proceedingsconference-collections
abstract

Towards Fixation Extraction in Corneal Imaging Based Eye Tracking Data

Published:20 April 2018Publication History

ABSTRACT

Humans sense most of the environment through their eyes. Hence, gaze is a powerful way to estimate visual attention. Head-mounted or mobile eye tracking is an established tool to analyze the visual behavior of people. Since these systems usually require some kind of calibration prior to usage, a new generation of mobile eye tracking devices based on corneal imaging has been investigated. However, little attention has been given on how to analyze corneal imaging specific eye tracking data. A classic approach in state-of-the-art systems is to extract different eye movements (e.g., fixations, saccades and pursuits movements). So far, there is no approach for applying these methods to corneal imaging data. We present a proof-of-concept method for fixation extraction and clustering of corneal imaging data. With this method we can compress the eye tracking data and make it ready for further analysis (e.g., attention measurement and object detection).

References

  1. Jason S. Babcock and Jeff B. Pelz. 2004. Building a Lightweight Eyetracking Headgear. In Proceedings of the 2004 Symposium on Eye Tracking Research & Applications (ETRA '04). ACM, New York, NY, USA, 109--114. Google ScholarGoogle ScholarDigital LibraryDigital Library
  2. Andreas Bulling and Hans Gellersen. 2010. Toward Mobile Eye-Based Human-Computer Interaction. IEEE Pervasive Computing 9, 4 (Oct. 2010), 8--12. Google ScholarGoogle ScholarDigital LibraryDigital Library
  3. Edwin S. Dalmaijer, Sebastiaan Mathôt, and Stefan Van der Stigchel. 2014. PyGaze: An Open-Source, Cross-platform Toolbox for Minimal-effort Programming of Eyetracking Experiments. Behavior Research Methods 46, 4 (01 Dec 2014), 913--921.Google ScholarGoogle Scholar
  4. Andrew T. Duchowski. 2007. Eye Tracking Methodology: Theory and Practice. Springer-Verlag New York, Inc., Secaucus, NJ, USA. Google ScholarGoogle ScholarDigital LibraryDigital Library
  5. Dan Witzner Hansen and Qiang Ji. 2010. In the Eye of the Beholder: A Survey of Models for Eyes and Gaze. IEEE Trans. Pattern Anal. Mach. Intell. 32, 3 (March 2010), 478--500. Google ScholarGoogle ScholarDigital LibraryDigital Library
  6. Christian Lander, Felix Kosmalla, Frederik Wiehr, and Sven Gehring. 2017. Using Corneal Imaging for Measuring a Human's Visual Attention. In Proceedings of the 2017 ACM International Joint Conference on Pervasive and Ubiquitous Computing and Proceedings of the 2017 ACM International Symposium on Wearable Computers (UbiComp '17). ACM, New York, NY, USA, 947--952. Google ScholarGoogle ScholarDigital LibraryDigital Library
  7. Christian Lander, Antonio Krüger, and Markus Löchtefeld. 2016. "The Story of Life is Quicker Than the Blink of an Eye": Using Corneal Imaging for Life Logging. In Proc. of the 2016 ACM International Joint Conference on Pervasive and Ubiquitous Computing: Adjunct (UbiComp '16). ACM, New York, NY, USA, 1686--1695. Google ScholarGoogle ScholarDigital LibraryDigital Library
  8. Christian Lander, Markus löchtefeld, and Antonio Krüger. 2018. hEYEbrid: A Hybrid Approach for Mobile Calibration-Free Gaze Estimation. Proc. ACM Interact. Mob. Wearable Ubiquitous Technol. 1, 4, Article 149 (Jan. 2018), 29 pages. Google ScholarGoogle ScholarDigital LibraryDigital Library
  9. Dongheng Li, Jason Babcock, and Derrick J. Parkhurst. 2006. openEyes: A Low-Cost Head-Mounted Eye-Tracking Solution. In Proceedings of the 2006 Symposium on Eye Tracking Research & Applications (ETRA '06). ACM, New York, NY, USA, 95--100. Google ScholarGoogle ScholarDigital LibraryDigital Library
  10. Atsushi Nakazawa, Christian Nitschke, and Toyoaki Nishida. 2015. Non-Calibrated and Real-Time Human View Estimation Using a Mobile Corneal Imaging Camera. In Multimedia & Expo Workshops (ICMEW), 2015 IEEE International Conference on. IEEE, 1--6.Google ScholarGoogle ScholarCross RefCross Ref
  11. Ko Nishino, Peter N. Belhumeur, and Shree K. Nayar. 2005. Using Eye Reflections for Face Recognition Under Varying Illumination. In Proceedings of the Tenth IEEE International Conference on Computer Vision Volume 01 (ICCV '05). IEEE Computer Society, Washington, DC, USA, 519--526. Google ScholarGoogle ScholarDigital LibraryDigital Library
  12. Ko Nishino and Shree K. Nayar. 2004a. Eyes for Relighting. In ACM SIGGRAPH 2004 Papers (SIGGRAPH '04). ACM, New York, NY, USA, 704--711. Google ScholarGoogle ScholarDigital LibraryDigital Library
  13. Ko Nishino and Shree K. Nayar. 2004b. The World in an Eye. In Proceedings of the 2004 IEEE Computer Society Conference on Computer Vision and Pattern Recognition, Vol. 1. IEEE, I--444.Google ScholarGoogle Scholar
  14. Ko Nishino and Shree K. Nayar. 2006. Corneal Imaging System: Environment from Eyes. Int. J. Comput. Vision 70, 1 (Oct. 2006), 23--40. Google ScholarGoogle ScholarDigital LibraryDigital Library
  15. Christian Nitschke, Atsushi Nakazawa, and Toyoaki Nishida. 2013. I See What You See: Point of Gaze Estimation from Corneal Images. In Proceedings of the 2013 IAPR Asian Conference on Pattern Recognition (ACPR '13). IEEE Computer Society, Washington, DC, USA, 298--304. Google ScholarGoogle ScholarDigital LibraryDigital Library
  16. Dario D. Salvucci and Joseph H. Goldberg. 2000. Identifying Fixations and Saccades in Eye-tracking Protocols. In Proceedings of the 2000 Symposium on Eye Tracking Research & Applications (ETRA '00). ACM, New York, NY, USA, 71--78. Google ScholarGoogle ScholarDigital LibraryDigital Library
  17. Javier San Agustin, Henrik Skovsgaard, Emilie Mollenbach, Maria Barret, Martin Tall, Dan Witzner Hansen, and John Paulin Hansen. 2010. Evaluation of a Low-cost Open-source Gaze Tracker. In Proceedings of the 2010 Symposium on Eye-Tracking Research & Applications (ETRA '10). ACM, New York, NY, USA, 77--80. Google ScholarGoogle ScholarDigital LibraryDigital Library
  18. Roel Vertegaal. 2003. Attentive User Interfaces. Commun. ACM 46, 3 (2003), 30--33.Google ScholarGoogle ScholarDigital LibraryDigital Library

Index Terms

  1. Towards Fixation Extraction in Corneal Imaging Based Eye Tracking Data

    Recommendations

    Comments

    Login options

    Check if you have access through your login credentials or your institution to get full access on this article.

    Sign in
    • Published in

      cover image ACM Conferences
      CHI EA '18: Extended Abstracts of the 2018 CHI Conference on Human Factors in Computing Systems
      April 2018
      3155 pages
      ISBN:9781450356213
      DOI:10.1145/3170427

      Copyright © 2018 Owner/Author

      Permission to make digital or hard copies of part or all of this work for personal or classroom use is granted without fee provided that copies are not made or distributed for profit or commercial advantage and that copies bear this notice and the full citation on the first page. Copyrights for third-party components of this work must be honored. For all other uses, contact the Owner/Author.

      Publisher

      Association for Computing Machinery

      New York, NY, United States

      Publication History

      • Published: 20 April 2018

      Check for updates

      Qualifiers

      • abstract

      Acceptance Rates

      CHI EA '18 Paper Acceptance Rate1,208of3,955submissions,31%Overall Acceptance Rate6,164of23,696submissions,26%

      Upcoming Conference

      CHI '24
      CHI Conference on Human Factors in Computing Systems
      May 11 - 16, 2024
      Honolulu , HI , USA

    PDF Format

    View or Download as a PDF file.

    PDF

    eReader

    View online with eReader.

    eReader