skip to main content
10.1145/1180995.1181051acmconferencesArticle/Chapter ViewAbstractPublication Pagesicmi-mlmiConference Proceedingsconference-collections
Article

Recognizing gaze aversion gestures in embodied conversational discourse

Published:02 November 2006Publication History

ABSTRACT

Eye gaze offers several key cues regarding conversational discourse during face-to-face interaction between people. While a large body of research results exist to document the use of gaze in human-to-human interaction, and in animating realistic embodied avatars, recognition of conversational eye gestures - distinct eye movement patterns relevant to discourse - has received less attention. We analyze eye gestures during interaction with an animated embodied agent and propose a non-intrusive vision-based approach to estimate eye gaze and recognize eye gestures. In our user study, human participants avert their gaze (i.e. with "look-away" or "thinking" gestures) during periods of cognitive load. Using our approach, an agent can visually differentiate whether a user is thinking about a response or is waiting for the agent or robot to take its turn.

References

  1. M. Argyle and M. Cook. Gaze and mutual gaze. Cambridge University Press, 1976.]]Google ScholarGoogle Scholar
  2. AT&T. Natural Voices. http://www.naturalvoices.att.com.]]Google ScholarGoogle Scholar
  3. S. Basu, I. Essa, and A. Pentland. Motion regularization for model-based head tracking. In In Intl. Conf. on Pattern Recognition (ICPR '96), 1996.]] Google ScholarGoogle ScholarDigital LibraryDigital Library
  4. Tim Bickmore and Justine Cassell. J. van Kuppevelt, L. Dybkjaer, and N. Bernsen (eds.), Natural, Intelligent and Effective Interaction with Multimodal Dialogue Systems, chapter Social Dialogue with Embodied Conversational Agents. Kluwer Academic, 2004.]]Google ScholarGoogle Scholar
  5. Breazeal, Hoffman, and A. Lockerd. Teaching and working with robots as a collaboration. In The Third International Conference on Autonomous Agents and Multi-Agent Systems AAMAS 2004, pages 1028--1035. ACM Press, July 2004.]] Google ScholarGoogle ScholarDigital LibraryDigital Library
  6. De Carolis, Pelachaud, Poggi, and F. de Rosis. Behavior planning for a reflexive agent. In Proceedings of IJCAI, Seattle, September 2001.]]Google ScholarGoogle ScholarDigital LibraryDigital Library
  7. M. R. M. Mimica C. H. Morimoto. Eye gaze tracking techniques for interactive applications. Computer Vision and Image Understanding, 98:52--82, 2005.]]Google ScholarGoogle ScholarCross RefCross Ref
  8. R. Alex Colburn, Michael F. Cohen, and Steven M. Ducker. The role or eye gaze in avatar mediated conversational interfaces. Technical Report MSR-TR-2000-81, Microsoft Research, July 2000.]]Google ScholarGoogle Scholar
  9. Dillman, Becher, and P. Steinhaus. ARMAR II - a learning and cooperative multimodal humanoid robot system. International Journal of Humanoid Robotics, 1(1):143--155, 2004.]]Google ScholarGoogle ScholarCross RefCross Ref
  10. Dillman, Ehrenmann, Steinhaus, Rogalla, and R. Zoellner. Human friendly programming of humanoid robots - the German Collaborative Research Center. In The Third IARP Intenational Workshop on Humanoid and Human-Friendly Robotics, Tsukuba Research Centre, Japan, December 2002.]]Google ScholarGoogle Scholar
  11. Atsushi Fukayama, Takehiko Ohno, Naoki Mukawa, Minako Sawaki, and Norihiro Hagita. Messages embedded in gaze of interface agents - impression management with agent's gaze. In CHI '02: Proceedings of the SIGCHI conference on Human factors in computing systems, pages 41--48, 2002.]] Google ScholarGoogle ScholarDigital LibraryDigital Library
  12. A. M. Glenberg, J. L. Schroeder, and D.A. Robertson. Averting the gaze disengages the environment and facilitates remembering. Memory and cognition, 26(4):651--658, July 1998.]]Google ScholarGoogle ScholarCross RefCross Ref
  13. Z. M. Griffin and K. Bock. What the eyes say about speaking. Psychological Science, 11(4):274--279, 2000.]]Google ScholarGoogle ScholarCross RefCross Ref
  14. Haptek. Haptek Player. http://www.haptek.com.]]Google ScholarGoogle Scholar
  15. Craig Hennessey, Borna Noureddin, and Peter Lawrence. A single camera eye-gaze tracking system with free head motion. In ETRA '06: Proceedings of the 2006 symposium on Eye tracking research & applications, pages 87--94, 2006.]] Google ScholarGoogle ScholarDigital LibraryDigital Library
  16. A. Kendon. Some functions of gaze direction in social interaction. Acta Psyghologica, 26:22--63, 1967.]]Google ScholarGoogle ScholarCross RefCross Ref
  17. Sooha Park Lee, Jeremy B. Badler, and Norman I. Badler. Eyes alive. In SIGGRAPH '02: Proceedings of the 29th annual conference on Computer graphics and interactive techniques, pages 637--644, 2002.]] Google ScholarGoogle ScholarDigital LibraryDigital Library
  18. Michael Li and Ted Selker. Eye pattern analysis in intelligent virtual agents. In Conference on Intelligent Virutal Agents(IVA02), pages 23--35, 2001.]] Google ScholarGoogle ScholarDigital LibraryDigital Library
  19. P. Majaranta and K. J. Raiha. Twenty years of eye typing: systems and design issues. In ETRA '02: Proceedings of the 2002 symposium on Eye tracking research & applications, pages 15--22, 2002.]] Google ScholarGoogle ScholarDigital LibraryDigital Library
  20. Louis-Philippe Morency, Ali Rahimi, and Trevor Darrell. Adaptive view-based appearance model. In Proceedings IEEE Conf. on Computer Vision and Pattern Recognition, volume 1, pages 803--810, 2003.]]Google ScholarGoogle ScholarCross RefCross Ref
  21. Louis-Philippe Morency, Candace Sidner, Christopher Lee, and Trevor Darrell. Contextual recognition of head gestures. In Proceedings of the International Conference on Multi-modal Interfaces, October 2005.]] Google ScholarGoogle ScholarDigital LibraryDigital Library
  22. Louis-Philippe Morency, Patrik Sundberg, and Trevor Darrell. Pose estimation using 3d view-based eigenspaces. In ICCV Workshop on Analysis and Modeling of Faces and Gestures, pages 45--52, Nice, France, 2003.]] Google ScholarGoogle ScholarDigital LibraryDigital Library
  23. Nakano, Reinstein, Stocky, and Justine Cassell. Towards a model of face-to-face grounding. In Proceedings of the Annual Meeting of the Association for Computational Linguistics, Sapporo, Japan, July 2003.]] Google ScholarGoogle ScholarDigital LibraryDigital Library
  24. D. G. Novick, B. Hansen, and K. Ward. Coordinating turn-taking with gaze. In Proceedings of the Fourth International Conference on Spoken Language, volume 3, pages 1888--1891, 1996.]]Google ScholarGoogle ScholarCross RefCross Ref
  25. Takehiko Ohno and Naoki Mukawa. A free-head, simple calibration, gaze tracking system that enables gaze-based interaction. In ETRA '04: Proceedings of the 2004 symposium on Eye tracking research & applications, pages 115--122, 2004.]] Google ScholarGoogle ScholarDigital LibraryDigital Library
  26. C. Pelachaud and M. Bilvi. Modelling gaze behavior for conversational agents. In International Working Conference on Intelligent Virtual Agents, pages 15--17, September 2003.]]Google ScholarGoogle ScholarCross RefCross Ref
  27. A. Pentland, B. Moghaddam, and T. Starner. View-based and modular eigenspaces for face recognition. In Proc. of IEEE Conf. on Computer Vision and Pattern Recognition, Seattle, WA, June 1994.]]Google ScholarGoogle ScholarCross RefCross Ref
  28. Pernilla Qvarfordt and Shumin Zhai. Conversing with the user based on eye-gaze patterns. In CHI '05: Proceedings of the SIGCHI conference on Human factors in computing systems, pages 221--230, 2005.]] Google ScholarGoogle ScholarDigital LibraryDigital Library
  29. Daniel C. Richardson and Michael J. Spivey. Representation, space and hollywood squares: looking at things that aren't there anymore. Cognition, 76:269--295, 2000.]]Google ScholarGoogle ScholarCross RefCross Ref
  30. D. C. Richardson and R. Dale. Looking to understand: The coupling between speakers' and listeners' eye movements and its relationship to discourse comprehension. In Proceedings of the 26th Annual Meeting of the Cognitive Science Society, pages 1143--1148, 2004.]]Google ScholarGoogle Scholar
  31. C. Sidner, C. Lee, C. D. Kidd, N. Lesh, and C. Rich. Explorations in engagement for humans and robots. Artificial Intelligence, 166(1--2):140--164, August 2005.]]Google ScholarGoogle Scholar
  32. M. J. Spivey and J. J. Geng. Oculomotor mechanisms activated by imagery and memory: Eye movements to absent objects. Psychological Research/Psychologische Forschung, 65(4):235--241, 2001.]]Google ScholarGoogle ScholarCross RefCross Ref
  33. G. Stein and A. Shashua. Direct estimation of motion and extended scene structure from moving stereo rig. In Proc. of CVPR, June 1998.]] Google ScholarGoogle ScholarDigital LibraryDigital Library
  34. R. Stiefelhagen, J. Yang, and A. Waibel. Tracking eyes and monitoring eye gaze. In Proceedings of the Workshop on Perceptual User Interfaces (PUI'97), pages 98--100, Alberta, Canada, 1997.]]Google ScholarGoogle Scholar
  35. D. Traum and J. Rickel. Embodied agents for multi-party dialogue in immersive virtual world. In Proceedings of the International Joint Conference on Autonomous Agents and Multi-agent Systems (AAMAS 2002), pages 766--773, July 2002.]] Google ScholarGoogle ScholarDigital LibraryDigital Library
  36. David Traum and Jeff Rickel. Embodied agents for multi-party dialogue in immersive virtual worlds. In AAMAS '02: Proceedings of the first international joint conference on Autonomous agents and multiagent systems, pages 766--773, 2002.]] Google ScholarGoogle ScholarDigital LibraryDigital Library
  37. B. M. Velichkovsky and J. P. Hansen. New technological windows in mind: There is more in eyes and brains for human-computer interaction. In CHI '96: Proceedings of the SIGCHI conference on Human factors in computing systems, 1996.]] Google ScholarGoogle ScholarDigital LibraryDigital Library
  38. Roel Vertegaal, Robert Slagter, Gerrit van der Veer, and Anton Nijholt. Eye gaze patterns in conversations: there is more to conversational agents than meets the eyes. In CHI '01: Proceedings of the SIGCHI conference on Human factors in computing systems, pages 301--308, 2001.]] Google ScholarGoogle ScholarDigital LibraryDigital Library
  39. P. Viola and M. Jones. Robust real-time face detection. In ICCV, page II: 747, 2001.]] Google ScholarGoogle ScholarDigital LibraryDigital Library
  40. X. Xie, R. Sudhakar, and H. Zhuang. A cascaded scheme for eye tracking and head movement compensation. IEEE Transactions on Systems, Man and Cybernetics, 28(4):487--490, July 1998.]]Google ScholarGoogle ScholarDigital LibraryDigital Library
  41. L. Young and D. Sheena. Survey of eye movement recording methods. Behavior Research Methods and Instrumentation, 7:397--429, 1975.]]Google ScholarGoogle ScholarCross RefCross Ref
  42. S. Zhai, C. Morimoto, and S. Ihde. Manual and gaze input cascaded (magic) pointing. In CHI99, pages 246--253, 1999.]] Google ScholarGoogle ScholarDigital LibraryDigital Library

Index Terms

  1. Recognizing gaze aversion gestures in embodied conversational discourse

          Recommendations

          Comments

          Login options

          Check if you have access through your login credentials or your institution to get full access on this article.

          Sign in
          • Published in

            cover image ACM Conferences
            ICMI '06: Proceedings of the 8th international conference on Multimodal interfaces
            November 2006
            404 pages
            ISBN:159593541X
            DOI:10.1145/1180995

            Copyright © 2006 ACM

            Permission to make digital or hard copies of all or part of this work for personal or classroom use is granted without fee provided that copies are not made or distributed for profit or commercial advantage and that copies bear this notice and the full citation on the first page. Copyrights for components of this work owned by others than ACM must be honored. Abstracting with credit is permitted. To copy otherwise, or republish, to post on servers or to redistribute to lists, requires prior specific permission and/or a fee. Request permissions from [email protected]

            Publisher

            Association for Computing Machinery

            New York, NY, United States

            Publication History

            • Published: 2 November 2006

            Permissions

            Request permissions about this article.

            Request Permissions

            Check for updates

            Qualifiers

            • Article

            Acceptance Rates

            Overall Acceptance Rate453of1,080submissions,42%

          PDF Format

          View or Download as a PDF file.

          PDF

          eReader

          View online with eReader.

          eReader