skip to main content
10.1145/3119881.3119890acmconferencesArticle/Chapter ViewAbstractPublication PagessapConference Proceedingsconference-collections
research-article

Subtle gaze guidance for immersive environments

Published:16 September 2017Publication History

ABSTRACT

Immersive displays allow presentation of rich video content over a wide field of view. We present a method to boost visual importance for a selected - possibly invisible - scene part in a cluttered virtual environment. This desirable feature enables to unobtrusively guide the gaze direction of a user to any location within the immersive 360° surrounding. Our method is based on subtle gaze direction which did not include head rotations in previous work. For covering the full 360° environment and wide field of view, we contribute an approach for dynamic stimulus positioning and shape variation based on eccentricity to compensate for visibility differences across the visual field. Our approach is calibrated in a perceptual study for a head-mounted display with binocular eye tracking. An additional study validates the method within an immersive visual search task.

References

  1. Sean Andrist, Michael Gleicher, and Bilge Mutlu. 2017. Looking Coordinated: Bidirectional Gaze Mechanisms for Collaborative Interaction with Virtual Characters. Computer Human Interaction (CHI) (2017). Google ScholarGoogle ScholarDigital LibraryDigital Library
  2. Reynold Bailey, Ann McNamara, Nisha Sudarsanam, and Cindy Grimm. 2009. Subtle gaze direction. ACM Transactions on Graphics (TOG) 28, 4 (2009), 100. Google ScholarGoogle ScholarDigital LibraryDigital Library
  3. Erhardt Barth, Michael Dorr, Martin Böhme, Karl Gegenfurtner, and Thomas Martinetz. 2006. Guiding the mind's eye: improving communication and vision by external control of the scanpath. In Electronic Imaging 2006. International Society for Optics and Photonics, 60570D--60570D.Google ScholarGoogle Scholar
  4. Frank Biocca, Charles Owen, Arthur Tang, and Corey Bohil. 2007. Attention issues in spatial information systems: Directing mobile users' visual attention using augmented reality. Journal of Management Information Systems 23, 4 (2007), 163--184. Google ScholarGoogle ScholarDigital LibraryDigital Library
  5. Thomas Booth, Srinivas Sridharan, Ann McNamara, Cindy Grimm, and Reynold Bailey. 2013. Guiding attention in controlled real-world environments. In Proceedings of the ACM Symposium on Applied Perception. ACM, 75--82. Google ScholarGoogle ScholarDigital LibraryDigital Library
  6. Forrester Cole, Douglas DeCarlo, Adam Finkelstein, Kenrick Kin, R Keith Morley, and Anthony Santella. 2006. Directing Gaze in 3D Models with Stylized Focus. Rendering Techniques 2006 (2006), 17th. Google ScholarGoogle ScholarDigital LibraryDigital Library
  7. Björn B de Koning and Halszka Jarodzka. 2017. Attention guidance strategies for supporting learning from dynamic visualizations. In Learning from Dynamic Visualization. Springer, 255--278.Google ScholarGoogle Scholar
  8. Michael Dorr, Eleonora Vig, Karl R Gegenfurtner, Thomas Martinetz, and Erhardt Barth. 2008. Eye movement modelling and gaze guidance. In Fourth International Workshop on Human-Computer Conversation.Google ScholarGoogle Scholar
  9. Carl Gutwin and Chris Fedak. 2004. A comparison of fisheye lenses for interactive layout tasks. In Proceedings of Graphics interface 2004. Canadian Human-Computer Communications Society, 213--220. Google ScholarGoogle ScholarDigital LibraryDigital Library
  10. Hajime Hata, Hideki Koike, and Yoichi Sato. 2016. Visual Guidance with Unnoticed Blur Effect. In Proceedings of the International Working Conference on Advanced Visual Interfaces. ACM, 28--35. Google ScholarGoogle ScholarDigital LibraryDigital Library
  11. Raphael Hoffmann, Patrick Baudisch, and Daniel S Weld. 2008. Evaluating visual cues for window switching on large screens. In Proceedings of the SIGCHI Conference on Human Factors in Computing Systems. ACM, 929--938. Google ScholarGoogle ScholarDigital LibraryDigital Library
  12. Halszka Jarodzka, Tamara van Gog, Michael Dorr, Katharina Scheiter, and Peter Gerjets. 2013. Learning to see: Guiding students' attention via a model's eye movements fosters learning. Learning and Instruction 25 (2013), 62--70.Google ScholarGoogle ScholarCross RefCross Ref
  13. Azam Khan, Justin Matejka, George Fitzmaurice, and Gordon Kurtenbach. 2005. Spotlight: directing users' attention on large displays. In Proceedings of the SIGCHI conference on Human factors in computing systems. ACM, 791--798. Google ScholarGoogle ScholarDigital LibraryDigital Library
  14. Robert Kosara, Silvia Miksch, and Helwig Hauser. 2002. Focus + context taken literally. IEEE Computer Graphics and Applications 22, 1 (2002), 22--29. Google ScholarGoogle ScholarDigital LibraryDigital Library
  15. Yen-Chen Lin, Yung-Ju Chang, Hou-Ning Hu, Hsien-Tzu Cheng, Chi-Wen Huang, and Min Sun. 2017. Tell Me Where to Look: Investigating Ways for Assisting Focus in 360° Video. In Proceedings of the 2017 CHI Conference on Human Factors in Computing Systems. ACM, 2535--2545. Google ScholarGoogle ScholarDigital LibraryDigital Library
  16. Andrei Lintu and Noelle Carbonell. 2009. Gaze Guidance through Peripheral Stimuli. (2009).Google ScholarGoogle Scholar
  17. Weiquan Lu, Henry Been-Lirn Duh, Steven Feiner, and Qi Zhao. 2014. Attributes of subtle cues for facilitating visual search in augmented reality. IEEE transactions on visualization and computer graphics 20, 3 (2014), 404--412. Google ScholarGoogle ScholarDigital LibraryDigital Library
  18. Ann McNamara, Reynold Bailey, and Cindy Grimm. 2009. Search task performance using subtle gaze direction with the presence of distractions. ACM Transactions on Applied Perception (TAP) 6, 3 (2009), 17. Google ScholarGoogle ScholarDigital LibraryDigital Library
  19. Ann McNamara, Thomas Booth, Srinivas Sridharan, Stephen Caffey, Cindy Grimm, and Reynold Bailey. 2012. Directing gaze in narrative art. In Proceedings of the ACM Symposium on Applied Perception. ACM, 63--70. Google ScholarGoogle ScholarDigital LibraryDigital Library
  20. Laura Pomarjanschi, Michael Dorr, and Erhardt Barth. 2012. Gaze guidance reduces the number of collisions with pedestrians in a driving simulator. ACM Transactions on Interactive Intelligent Systems (TiiS) 1, 2 (2012), 8. Google ScholarGoogle ScholarDigital LibraryDigital Library
  21. Yoichi Sato, Yusuke Sugano, Akihiro Sugimoto, Yoshinori Kuno, and Hideki Koike. 2016. Sensing and Controlling Human Gaze in Daily Living Space for Human-Harmonized Information Environments. In Human-Harmonized Information Technology, Volume 1. Springer, 199--237.Google ScholarGoogle Scholar
  22. Alia Sheikh, Andy Brown, Zillah Watson, and Michael Evans. 2016. Directing attention in 360-degree video. (2016).Google ScholarGoogle Scholar
  23. Michael Stengel, Steve Grogorick, Martin Eisemann, Elmar Eisemann, and Marcus A Magnor. 2015. An affordable solution for binocular eye tracking and calibration in head-mounted displays. In Proceedings of the 23rd ACM international conference on Multimedia. ACM, 15--24. Google ScholarGoogle ScholarDigital LibraryDigital Library
  24. Hans Strasburger, Ingo Rentschler, and Martin Jüttner. 2011. Peripheral vision and pattern recognition: A review. 11, 5 (2011), 13--13.Google ScholarGoogle Scholar
  25. Eduardo E Veas, Erick Mendez, Steven K Feiner, and Dieter Schmalstieg. 2011. Directing attention and influencing memory with visual saliency modulation. In Proceedings of the SIGCHI Conference on Human Factors in Computing Systems. ACM, 1471--1480. Google ScholarGoogle ScholarDigital LibraryDigital Library
  26. Eleonora Vig, Michael Dorr, and Erhardt Barth. 2011. Learned saliency transformations for gaze guidance. In IS&T/SPIE Electronic Imaging. International Society for Optics and Photonics, 78650W--78650W.Google ScholarGoogle Scholar
  27. N. Waldin, M. Waldner, and I. Viola. 2017. Flicker Observer Effect: Guiding Attention Through High Frequency Flicker in Images. Computer Graphics Forum 36, 2 (2017), 467--476. Google ScholarGoogle ScholarDigital LibraryDigital Library
  28. Martin Weier, Michael Stengel, Thorsten Roth, Piotr Didyk, Elmar Eisemann, Martin Eisemann, Steve Grogorick, André Hinkenjann, E Kruijff, M Magnor, et al. 2017. Perception-driven Accelerated Rendering. In Computer Graphics Forum, Vol. 36. Wiley Online Library, 611--643. Google ScholarGoogle ScholarDigital LibraryDigital Library

Index Terms

  1. Subtle gaze guidance for immersive environments

        Recommendations

        Comments

        Login options

        Check if you have access through your login credentials or your institution to get full access on this article.

        Sign in
        • Published in

          cover image ACM Conferences
          SAP '17: Proceedings of the ACM Symposium on Applied Perception
          September 2017
          101 pages
          ISBN:9781450351485
          DOI:10.1145/3119881

          Copyright © 2017 ACM

          Permission to make digital or hard copies of all or part of this work for personal or classroom use is granted without fee provided that copies are not made or distributed for profit or commercial advantage and that copies bear this notice and the full citation on the first page. Copyrights for components of this work owned by others than the author(s) must be honored. Abstracting with credit is permitted. To copy otherwise, or republish, to post on servers or to redistribute to lists, requires prior specific permission and/or a fee. Request permissions from [email protected].

          Publisher

          Association for Computing Machinery

          New York, NY, United States

          Publication History

          • Published: 16 September 2017

          Permissions

          Request permissions about this article.

          Request Permissions

          Check for updates

          Qualifiers

          • research-article

          Acceptance Rates

          Overall Acceptance Rate43of94submissions,46%

        PDF Format

        View or Download as a PDF file.

        PDF

        eReader

        View online with eReader.

        eReader