ABSTRACT
Three-dimensional printed models have the potential to serve as powerful accessibility tools for blind people. Recently, researchers have developed methods to further enhance 3D prints by making them interactive: when a user touches a certain area in the model, the model speaks a description of the area. However, these interactive models were limited in terms of their functionalities and interaction techniques. We conducted a two-section study with 12 legally blind participants to fill in the gap between existing interactive model technologies and end users' needs, and explore design opportunities. In the first section of the study, we observed participants' behavior as they explored and identified models and their components. In the second section, we elicited user-defined input techniques that would trigger various functions from an interactive model. We identified five exploration activities (e.g., comparing tactile elements), four hand postures (e.g., using one hand to hold a model in the air), and eight gestures (e.g., using index finger to strike on a model) from the participants' exploration processes and aggregate their elicited input techniques. We derived key insights from our findings including: (1) design implications for I3M technologies, and (2) specific designs for interactions and functionalities for I3Ms.
- Animal Cell - Thingiverse: http://www.thingiverse.com/thing:689381. Accessed: 2017-04-04.Google Scholar
- Autodesk Meshmixer: http://www.meshmixer.com/. Accessed: 2017-05-04.Google Scholar
- Brown, C. and Hurst, A. 2012. VizTouch: automatically generated tactile visualizations of coordinate spaces. TEI '12, 131-138. Google ScholarDigital Library
- Brule, E., Bailly, G., Brock, A., Valentin, F., Denis, G. and Jouffrais, C. 2016. MapSense: Multi-Sensory Interactive Maps for Children Living with Visual Impairments. CHI '16, 445-457. Google ScholarDigital Library
- Buehler, E., Hurst, A. and Hofmann, M. 2014. Coming to grips: 3D printing for accessibility. ASSETS '14, 291-292. Google ScholarDigital Library
- Buehler, E., Kane, S.K. and Hurst, A. 2014. ABC and 3D: opportunities and obstacles to 3D printing in special education environments. ASSETS '14, 107-114. Google ScholarDigital Library
- Davidson, P.W. 1972. Haptic judgments of curvature by blind and sighted humans. Journal of Experimental Psychology. 93, 1 (1972), 43-55.Google ScholarCross Ref
- Good, M.D., Whiteside, J.A., Wixon, D.R. and Jones, S.J. 1984. Building a user-derived interface. Communications of the ACM. 27, 10 (Oct. 1984), 1032-1043. Google ScholarDigital Library
- Götzelmann, T. 2016. LucentMaps: 3D Printed Audiovisual Tactile Maps for Blind and Visually Impaired People. ASSETS '16, 81-90. Google ScholarDigital Library
- Hatch, M. 2013. The Maker Movement Manifesto: Rules for Innovation in the New World of Crafters, Hackers, and Tinkerers. McGraw Hill Professional.Google Scholar
- Hatwell, Y. 1987. Motor and cognitive functions of the hand in infancy and childhood. International Journal of Behavioral Development. 10, 4 (Dec. 1987), 509-526.Google ScholarCross Ref
- Hatwell, Y., Arlette, S. and Edouard, G. 2003. Touching for knowing: cognitive psychology of haptic manual perception. John Benjamins Publishing.Google Scholar
- Hofmann, M., Harris, J., Hudson, S.E. and Mankoff, J. 2016. Helping Hands: Requirements for a Prototyping Methodology for Upper-limb Prosthetics Users. CHI '16, 1769-1780. Google ScholarDigital Library
- Kane, S.K. and Bigham, J.P. 2014. Tracking@ stemxcomet: teaching programming to blind students via 3D printing, crisis management, and twitter. SIGCSE '14, 247-252. Google ScholarDigital Library
- Kim, J. and Yeh, T. 2015. Toward 3D-Printed Movable Tactile Pictures for Children al Impairments with Visual Impairments. CHI '15, 2815-2824. Google ScholarDigital Library
- Klatzky, R.L., Loomis, J.M. and Lederman, S.J. 1993. Haptic identification of objects and their depictions. Perception & Psychophysics. 54, 2 (1993), 170-178.Google ScholarCross Ref
- Kolitsky, M.A. 2016. 3D printing makes virtual world more real for blind learners. e-mentor. 1 (63) (2016), 65-70.Google Scholar
- Lederman, S.J. and Klatzky, R.L. 1993. Extracting object properties through haptic exploration. Acta Psychologica. 84, 1 (1993), 29-40.Google ScholarCross Ref
- Lederman, S.J. and Klatzky, R.L. 1987. Hand movements: A window into haptic object recognition. Cognitive Psychology. 19, 3 (1987), 342-368.Google ScholarCross Ref
- Lipson, H. and Kurman, M. 2013. Fabricated: The new world of 3D printing. John Wiley & Sons. Google ScholarDigital Library
- McDonald, S., Dutterer, J., Abdolrahmani, A., Kane, S.K. and Hurst, A. 2014. Tactile aids for visually impaired graphical design education. ASSETS '14, 275-276. Google ScholarDigital Library
- Morash, V.S., Pensky, A.E.C., Tseng, S.T.W. and Miele, J.A. 2014. Effects of using multiple hands and fingers on haptic performance in individuals who are blind. Perception. 43, 6 (2014), 569-588.Google ScholarCross Ref
- Morris, M.R. 2012. Web on the Wall: Insights from a Multimodal Interaction Elicitation Study. ITS '12, 95-104. Google ScholarDigital Library
- OpenStreetMap Buildings: https://osmbuildings.org/. Accessed: 2017-04-04.Google Scholar
- Reichinger, A., Fuhrmann, A., Maierhofer, S. and Purgathofer, W. 2016. Gesture-Based Interactive Audio Guide on Tactile Reliefs. ASSETS '16, 91-100. Google ScholarDigital Library
- Rovira, K., Deschamps, L. and Baena-Gomez, D. 2011. Mental rotation in blind and sighted adolescents: The effects of haptic strategies. Revue Europeene de Psychologie Appliquee. 61, 3 (2011), 153-160.Google ScholarCross Ref
- Saldana, J. 2015. The Coding Manual for Qualitative Researchers. Sage.Google Scholar
- Scouten, E.L. 1967. The Rochester method, an oral multisensory approach for instructing prelingual deaf children. American annals of the deaf. 112, 2 (Mar. 1967), 50-5.Google Scholar
- Shen, H., Edwards, O., Miele, J. and Coughlan, J.M. 2013. CamIO: A 3D computer vision system enabling audio/haptic interaction with physical objects by blind users. ASSETS '13, 1-2. Google ScholarDigital Library
- Shi, L. 2015. Talkabel: A Labeling Method for 3D Printed Models. ASSETS '15, 361-362. Google ScholarDigital Library
- Shi, L., McLachlan, R., Zhao, Y. and Azenkot, S. 2016. Magic Touch: Interacting with 3D Printed Graphics. ASSETS '16, 329-330. Google ScholarDigital Library
- Shi, L., Zelzer, I., Feng, C. and Azenkot, S. 2016. Tickers and Talker: An Accessible Labeling Toolkit for 3D Printed Models. CHI '16, 4896-4907. Google ScholarDigital Library
- Shi, L., Zhao, Y. and Azenkot, S. 2017. Markit and Talkit: A Low-Barrier Toolkit to Augment 3D Printed Models with Audio Annotations. UIST '17. Google ScholarDigital Library
- Stangl, A., Kim, J. and Yeh, T. 2014. 3D printed tactile picture books for children with visual impairments. IDC '14, 321-324. Google ScholarDigital Library
- Symmons, M. and Richardson, B. 2000. Raised Line Drawings are Spontaneously Explored with a Single Finger. Perception. 29, 5 (May 2000), 621-626.Google ScholarCross Ref
- Szpiro, S., Zhao, Y. and Azenkot, S. 2016. Finding a store, searching for a product: a study of daily challenges of low vision people. UbiComp '16, 61-72. Google ScholarDigital Library
- Szpiro, S.F.A., Hashash, S., Zhao, Y. and Azenkot, S. 2016. How People with Low Vision Access Computing Devices. ASSETS '16, 171-180. Google ScholarDigital Library
- Taylor, B., Dey, A., Siewiorek, D. and Smailagic, A. 2016. Customizable 3D Printed Tactile Maps as Interactive Overlays. ASSETS '16, 71-79. Google ScholarDigital Library
- Teibrich, A., Mueller, S., Guimbretière, F., Kovacs, R., Neubert, S. and Baudisch, P. 2015. Patching Physical Objects. UIST '15, 83-91. Google ScholarDigital Library
- Textured Earth - Thingiverse: http://www.thingiverse.com/thing:17336. Accessed: 2017-04-04.Google Scholar
- The Peachy Printer - The First $100 3D Printer: https://www.kickstarter.com/projects/117421627/the-peachy-printer-the-first-100-3d-printer-and-sc. Accessed: 2017-05-05.Google Scholar
- Wobbrock, J.O., Morris, M.R. and Wilson, A.D. 2009. User-defined gestures for surface computing. CHI '09, 1083. Google ScholarDigital Library
Index Terms
- Designing Interactions for 3D Printed Models with Blind People
Recommendations
Designing Interactive 3D Printed Models with Teachers of the Visually Impaired
CHI '19: Proceedings of the 2019 CHI Conference on Human Factors in Computing SystemsStudents with visual impairments struggle to learn various concepts in the academic curriculum because diagrams, images, and other visual are not accessible to them. To address this, researchers have design interactive 3D printed models (I3Ms) that ...
Designing Conversational Multimodal 3D Printed Models with People who are Blind
DIS '23: Proceedings of the 2023 ACM Designing Interactive Systems Conference3D printed models have been used to improve access to graphical information by people who are blind, offering benefits over conventional accessible graphics. Here we investigate an interactive 3D printed model (I3M) that combines a conversational ...
Comments