skip to main content
10.1145/2522848.2532202acmconferencesArticle/Chapter ViewAbstractPublication Pagesicmi-mlmiConference Proceedingsconference-collections
keynote

Hands and speech in space: multimodal interaction with augmented reality interfaces

Published:09 December 2013Publication History

ABSTRACT

Augmented Reality (AR) is technology that allows virtual imagery to be seamlessly integrated into the real world. Although first developed in the 1960's it has only been recently that AR has become widely available, through platforms such as the web and mobile phones. However most AR interfaces have very simple interaction, such as using touch on phone screens or camera tracking from real images. New depth sensing and gesture tracking technologies such as Microsoft Kinect or Leap Motion have made is easier than ever before to track hands in space. Combined with speech recognition and AR tracking and viewing software it is possible to create interfaces that allow users to manipulate 3D graphics in space through a natural combination of speech and gesture. In this paper I will review previous research in multimodal AR interfaces and give an overview of the significant research questions that need to be addressed before speech and gesture interaction can become commonplace.

References

  1. Azuma, R. 1997. A Survey of Augmented Reality. Presence, 6 (4), 355--385.Google ScholarGoogle ScholarDigital LibraryDigital Library
  2. Bai, H., Lee, G. A., and Billinghurst, M. 2012. Freeze view touch and finger gesture based interaction methods for handheld augmented reality interfaces. In Proceedings of the 27th Conference on Image and Vision Computing New Zealand (pp. 126--131). ACM. Google ScholarGoogle ScholarDigital LibraryDigital Library
  3. Billinghurst, M. and Lee, M (2012). Multimodal Interfaces for Augmented Reality. In Dill, John, Earnshaw, Rae, Kasik, David et al (editors), Expanding the Frontiers of Visual Analytics and Visualization. London : Springer.Google ScholarGoogle Scholar
  4. Chu CP, Dani TH, and Gadh R. 1997. Multimodal Interface for a virtual reality based computer aided design system. Proceedings of 1997 IEEE International Conference on Robotics and Automation 2: 1329--1334Google ScholarGoogle Scholar
  5. Cohen, P. R., Dalrymple, M., Moran, D. B., Pereira, F. C., and Sullivan, J. W. 1989. Synergistic use of direct manipulation and natural language. In ACM SIGCHI Bulletin (Vol. 20, No. SI, pp. 227--233). ACM. Google ScholarGoogle ScholarDigital LibraryDigital Library
  6. Irawati, S., Green, S., Billinghurst, M., Duenser, A., and Ko, H. 2006. An evaluation of an augmented reality multimodal interface using speech and paddle gestures. In Advances in Artificial Reality and Tele-Existence (pp. 272--283). Springer Berlin Heidelberg. Google ScholarGoogle ScholarDigital LibraryDigital Library
  7. Microsoft Kinect Website: http://www.xbox.com/KINECTGoogle ScholarGoogle Scholar
  8. Krum, D. M., Omoteso, O., Ribarsky, W., Starner, T., and Hodges, L. F. 2002. Speech and gesture multimodal control of a whole Earth 3D visualization environment. In Proceedings of the symposium on Data Visualisation 2002 (pp. 195--200). Eurographics Association. Google ScholarGoogle ScholarDigital LibraryDigital Library
  9. Leap Motion Website: https://www.leapmotion.com/Google ScholarGoogle Scholar
  10. Olwal, A., Benko, H., and Feiner, S. 2003. SenseShapes: Using statistical geometry for object selection in a multimodal augmented reality system. In Proceedings of the 2nd IEEE/ACM International Symposium on Mixed and Augmented Reality (p. 300). IEEE Computer Society. Google ScholarGoogle ScholarDigital LibraryDigital Library

Index Terms

  1. Hands and speech in space: multimodal interaction with augmented reality interfaces

        Recommendations

        Comments

        Login options

        Check if you have access through your login credentials or your institution to get full access on this article.

        Sign in
        • Published in

          cover image ACM Conferences
          ICMI '13: Proceedings of the 15th ACM on International conference on multimodal interaction
          December 2013
          630 pages
          ISBN:9781450321297
          DOI:10.1145/2522848

          Copyright © 2013 Owner/Author

          Permission to make digital or hard copies of part or all of this work for personal or classroom use is granted without fee provided that copies are not made or distributed for profit or commercial advantage and that copies bear this notice and the full citation on the first page. Copyrights for third-party components of this work must be honored. For all other uses, contact the Owner/Author.

          Publisher

          Association for Computing Machinery

          New York, NY, United States

          Publication History

          • Published: 9 December 2013

          Check for updates

          Qualifiers

          • keynote

          Acceptance Rates

          ICMI '13 Paper Acceptance Rate49of133submissions,37%Overall Acceptance Rate453of1,080submissions,42%

        PDF Format

        View or Download as a PDF file.

        PDF

        eReader

        View online with eReader.

        eReader