skip to main content
10.1145/1124772.1124951acmconferencesArticle/Chapter ViewAbstractPublication PageschiConference Proceedingsconference-collections
Article

Comparing remote gesture technologies for supporting collaborative physical tasks

Published: 22 April 2006 Publication History

Abstract

The design of remote gesturing technologies is an area of growing interest. Current technologies have taken differing approaches to the representation of remote gesture. It is not clear which approach has the most benefit to task performance. This study therefore compared performance in a collaborative physical (assembly) task using remote gesture systems constructed with combinations of three different gesture formats (unmediated hands only, hands and sketch and digital sketch only) and two different gesture output locations (direct projection into a worker's task space or on an external monitor). Results indicated that gesturing with an unmediated representation of the hands leads to faster performance with no loss of accuracy. Comparison of gesture output locations did not find a significant difference between projecting gestures and presenting them on external monitors. These results are discussed in relation to theories of conversational grounding and the design of technologies from a 'mixed ecologies' perspective.

References

[1]
Argyle, M. Bodily Communication. (1988). Routledge.
[2]
Bekker, M. M., Olson, J. S., & Olson, G. M. Analysis of gestures in face-to-face design teams provides guidance for how to use groupware in design. In Proceedings of DIS 95. ACM Press. (1995). 157--166.
[3]
Chapanis, A., Ochsman, R. B., Parrish, R. N., &Weeks, G. D. Studies in interactive communication: I. The effects of four communication modes on the behavior of teams during cooperative problem-solving. Human Factors, 14(6), (1972). 487--509.
[4]
Clark, H. H. Using language. (1996). Cambridge: Cambridge University Press.
[5]
Clark, H. H., & Krych, M. A. Speaking while monitoring addressees for understanding. Journal of Memory & Language, 50(1), (2004). 62--81.
[6]
Clark, H. H., & Wilkes-Gibbs, D. Referring as a collaborative process. Cognition, 22(1), (1986). 1--39.
[7]
Clark, H. H., & Brennan, S.E. Grounding in Communication. In L.B. Resnick, R.M. Levine, & S.D. Teasley (Eds.). Perspectives on socially shared cognition, (1991). 127--149. Washington, DC: APA.
[8]
Fraser, M. Working with Objects in Collaborative Environments. (2001) Unpublished PhD thesis, University of Nottingham.
[9]
Fussell, S. R., Setlock, L. D., Yang, J., Ou, J., Mauer, E. M., & Kramer, A. Gestures over video streams to support remote collaboration on physical tasks. Human-Computer Interaction, 19, (2004). 273--309.
[10]
Fussell, S.R., Kraut, R.E., & Siegel, J. Coordination of communication: Effects of shared visual context on collaborative work. Proceedings of CSCW 2000, ACM Press (2000). 21--30.
[11]
Fussell, S.R., Setlock, L.D., & Parker, E.M. Where do helpers look? Gaze targets during collaborative physical tasks. In Proceedings of CHI 2003 (Extended Abstracts), ACM Press (2003).768--769.
[12]
Gergle, D., Kraut, R.E., & Fussell, S.R. Action as language in a shared visual space. Proceedings of CSCW 2004. ACM Press (2004). 487--496.
[13]
Heath, C. C. and Luff P. K. Convergent activities: collaborative work and multimedia technology in London Underground Line Control Rooms, in D. Middleton and Y. Engestrom (eds.), Cognition and Communication at Work: Distributed Cognition in the Workplace. Cambridge University Press (1996). 96--130.
[14]
Karsenty, L. Cooperative work and shared visual context: An empirical study of comprehension problems and in side-by-side and remote help dialogues. Human-Computer Interaction, 14(3), (1999). 283--315.
[15]
Kirk, D. S. & Stanton Fraser, D. The Effects of Remote Gesturing on Distance Instruction. In Proceedings of CSCL 2005. LEA (2005).
[16]
Kirk, D., Crabtree, A. & Rodden, T. Ways of the Hands. In Proceedings of ECSCW 2005. Kluwer (2005).
[17]
Kraut, R. E., Fussell, S. R., & Siegel, J. Visual information as a conversational resource in collaborative physical tasks. Human Computer Interaction, 18(1), (2003). 13-49.
[18]
Kraut, R. E., Gergle, D., & Fussell, S. R. The use of visual information in shared visual spaces: Informing the development of virtual co-presence. In Proceedings of CSCW 2002, ACM Press. (2002). 31--40.
[19]
Kraut, R. E., Miller, M. D., & Siegel, J. Collaboration in performance of physical tasks: Effects on outcomes and communication. Proceedings of CSCW 1996 ACM Press (1996). 57--66.
[20]
Kuzuoka, H. Spatial workspace collaboration: A Sharedview video support system for remote collaboration capability. Proceedings of CHI'92 ACM Press (1992). 533--540.
[21]
Kuzuoka, H., Kosaka, J., Yamazaki, K., Suga, Y., Yamazaki, A., Luff, P., & Heath, C. Mediating Dual Ecologies. Proceedings of CSCW 2004 ACM Press (2004). 477--486).
[22]
Kuzuoka, H., Kosuge, T., & Tanaka, K. GestureCam: A video communication system for sympathetic remote collaboration. Proceedings of CSCW 1994 ACM Press (1994) 35--43.
[23]
Kuzuoka, H., Oyama, S., Yamazaki, K., Suzuki, K., & Mitsuishi, M. GestureMan: A mobile robot that embodies a remote instructor's actions. Proceedings of CSCW 2000 ACM Press (2000). 155--162.
[24]
Luff, P., Heath, C., Kuzuoka, H., Hindmarsh, J., Yamazaki, K., and Oyama, S., Fractured ecologies: creating environments for collaboration, Special Issue of the HCI Journal: 'Talking About Things: Mediated Conversation about Objects', 18, 1 & 2, (2003). 51--84.
[25]
McNeill, D. Hand and Mind. What gestures reveal about thought. (1996) Chicago: University of Chicago Press.
[26]
Ou, J., Fussell, S. R., Chen, X., Setlock, L. D., & Yang, J. Gestural communication over video stream: Supporting multimodal interaction for remote collaborative physical tasks. In Proceedings of ICMI 2003, (2003)
[27]
Ou, J., Min, L., Yang, J. and Fussell, S. R. Effects of Task Properties, Partner Actions and Message Content on Eye Gaze Patterns in a Collaborative Task. In Proceedings of CHI 2005. ACM Press (2005). 231 -- 240.
[28]
Rauscher, F. H., Krauss, R. M. & Chen, Y. (1996) Gesture, Speech and Lexical Access: The Role of Lexical Movements in Speech Production. Psychological Science. 7 4 226--231.
[29]
Sakata, N., Kurata, T., Kato, T., Kourogi, M. and Kuzuoka, H. WACL: Supporting Telecommunications Using Wearable Active Camera with Laser Pointer. Proceedings of Int. Symp. on Wearable Computers 2003. IEEE Press (2003).
[30]
Tang, A., Boyle, M. and Greenberg, S. Display and Presence Disparity in Mixed Presence Groupware. Proceedings of Australasian user Interface. ACM Press (2004). 73--82.
[31]
Tang, J. C. Findings from observational studies of collaborative work. International Journal of Man-Machine Studies, 34, (1991). 143--160.

Cited By

View all
  • (2024)Field Trial of a Tablet-based AR System for Intergenerational Connections through Remote ReadingProceedings of the ACM on Human-Computer Interaction10.1145/36536968:CSCW1(1-28)Online publication date: 26-Apr-2024
  • (2024)SharedNeRF: Leveraging Photorealistic and View-dependent Rendering for Real-time and Remote CollaborationProceedings of the 2024 CHI Conference on Human Factors in Computing Systems10.1145/3613904.3642945(1-14)Online publication date: 11-May-2024
  • (2024)Effect of the Hand and Gaze Pointers in Remote CollaborationIEEE Access10.1109/ACCESS.2024.350006512(172774-172784)Online publication date: 2024
  • Show More Cited By

Index Terms

  1. Comparing remote gesture technologies for supporting collaborative physical tasks

      Recommendations

      Comments

      Information & Contributors

      Information

      Published In

      cover image ACM Conferences
      CHI '06: Proceedings of the SIGCHI Conference on Human Factors in Computing Systems
      April 2006
      1353 pages
      ISBN:1595933727
      DOI:10.1145/1124772
      Permission to make digital or hard copies of all or part of this work for personal or classroom use is granted without fee provided that copies are not made or distributed for profit or commercial advantage and that copies bear this notice and the full citation on the first page. Copyrights for components of this work owned by others than ACM must be honored. Abstracting with credit is permitted. To copy otherwise, or republish, to post on servers or to redistribute to lists, requires prior specific permission and/or a fee. Request permissions from [email protected]

      Sponsors

      Publisher

      Association for Computing Machinery

      New York, NY, United States

      Publication History

      Published: 22 April 2006

      Permissions

      Request permissions for this article.

      Check for updates

      Author Tags

      1. CSCW
      2. hands
      3. mixed ecologies
      4. remote gesturing

      Qualifiers

      • Article

      Conference

      CHI06
      Sponsor:
      CHI06: CHI 2006 Conference on Human Factors in Computing Systems
      April 22 - 27, 2006
      Québec, Montréal, Canada

      Acceptance Rates

      Overall Acceptance Rate 6,199 of 26,314 submissions, 24%

      Upcoming Conference

      CHI 2025
      ACM CHI Conference on Human Factors in Computing Systems
      April 26 - May 1, 2025
      Yokohama , Japan

      Contributors

      Other Metrics

      Bibliometrics & Citations

      Bibliometrics

      Article Metrics

      • Downloads (Last 12 months)49
      • Downloads (Last 6 weeks)3
      Reflects downloads up to 09 Feb 2025

      Other Metrics

      Citations

      Cited By

      View all
      • (2024)Field Trial of a Tablet-based AR System for Intergenerational Connections through Remote ReadingProceedings of the ACM on Human-Computer Interaction10.1145/36536968:CSCW1(1-28)Online publication date: 26-Apr-2024
      • (2024)SharedNeRF: Leveraging Photorealistic and View-dependent Rendering for Real-time and Remote CollaborationProceedings of the 2024 CHI Conference on Human Factors in Computing Systems10.1145/3613904.3642945(1-14)Online publication date: 11-May-2024
      • (2024)Effect of the Hand and Gaze Pointers in Remote CollaborationIEEE Access10.1109/ACCESS.2024.350006512(172774-172784)Online publication date: 2024
      • (2024)Gesturing in the Air in Supporting Full MobilityComputer‐Supported Collaboration10.1002/9781119719830.ch8(195-213)Online publication date: 24-May-2024
      • (2024)Supporting Remote Hand Gestures over the Workspace VideoComputer‐Supported Collaboration10.1002/9781119719830.ch7(173-193)Online publication date: 24-May-2024
      • (2024)Communication Cues for Remote GuidanceComputer‐Supported Collaboration10.1002/9781119719830.ch4(81-114)Online publication date: 24-May-2024
      • (2024)Supporting Tailorability to Meet Individual Task NeedsComputer‐Supported Collaboration10.1002/9781119719830.ch11(269-298)Online publication date: 24-May-2024
      • (2024) Augmenting Hand Gestures in 3D Mixed Reality Computer‐Supported Collaboration10.1002/9781119719830.ch10(243-268)Online publication date: 24-May-2024
      • (2023)VRoxy: Wide-Area Collaboration From an Office Using a VR-Driven Robotic ProxyProceedings of the 36th Annual ACM Symposium on User Interface Software and Technology10.1145/3586183.3606743(1-13)Online publication date: 29-Oct-2023
      • (2023)Using Virtual Replicas to Improve Mixed Reality Remote CollaborationIEEE Transactions on Visualization and Computer Graphics10.1109/TVCG.2023.324711329:5(2785-2795)Online publication date: 22-Feb-2023
      • Show More Cited By

      View Options

      Login options

      View options

      PDF

      View or Download as a PDF file.

      PDF

      eReader

      View online with eReader.

      eReader

      Figures

      Tables

      Media

      Share

      Share

      Share this Publication link

      Share on social media