Abstract
Eyes-free operation of mobile devices is critical in situations where the visual channel is either unavailable or attention is needed elsewhere. In such situations, vibrotactile tracing along paths or lines can help users to navigate and identify symbols and shapes without visual information. In this paper, we investigated the applicability of different metrics that can measure the effectiveness of vibrotactile line tracing methods on touch screens. In two user studies, we compare trace Length Error, Area Error, and Fréchet Distance as alternatives to commonly used trace Time. Our results show that a lower Fréchet distance is correlated better with the comprehension of a line trace. Furthermore, we show that distinct feedback methods perform differently with varying geometric features in lines and propose a segmented line design for tactile line tracing studies. We believe the results will inform future designs of eyes-free operation techniques and studies.
Supplemental Material
- Paul Bach-y Rita and Stephen W. Kercel. 2003. Sensory substitution and the human-machine interface. Trends in Cognitive Sciences , Vol. 7, 12 (Dec. 2003), 541--546. https://doi.org/10.1016/j.tics.2003.10.013Google Scholar
- Andrea Bianchi, Ian Oakley, Jong Keun Lee, Dong Soo Kwon, and Vassilis Kostakos. 2011. Haptics for tangible interaction: A vibro-tactile prototype. In Proceedings of the fifth international conference on Tangible, embedded, and embodied interaction - TEI '11. ACM Press, New York, New York, USA, 283. https://doi.org/10.1145/1935701.1935764Google Scholar
Digital Library
- Stephen Brewster and Lorna M. Brown. 2004. Tactons: Structured Tactile Messages for Non-Visual Information Display. In Proceedings of the Fifth Conference on Australasian User Interface - Volume 28 (Dunedin, New Zealand) (AUIC '04). Australian Computer Society, Inc., AUS, 15--23. https://dl.acm.org/doi/10.5555/976310.976313Google Scholar
Digital Library
- Stephen A. Brewster, Peter C. Wright, and Alistair D. N. Edwards. 1993. An Evaluation of Earcons for Use in Auditory Human-Computer Interfaces. In Proceedings of the INTERACT '93 and CHI '93 Conference on Human Factors in Computing Systems (Amsterdam, The Netherlands) (CHI '93). Association for Computing Machinery, New York, NY, USA, 222--227. https://doi.org/10.1145/169059.169179Google Scholar
Digital Library
- S. Choi and K. J. Kuchenbecker. 2013. Vibrotactile Display: Perception, Technology, and Applications. Proc. IEEE , Vol. 101, 9 (2013), 2093--2104. https://doi.org/10.1109/JPROC.2012.2221071Google Scholar
Cross Ref
- Ashley Colley, Lasse Virtanen, Timo Ojala, and Jonna H"a kkil"a. 2016. Guided touch screen: enhanced eyes-free interaction. In Proceedings of the 5th ACM International Symposium on Pervasive Displays - PerDis '16 . ACM Press, New York, New York, USA, 80--86. https://doi.org/10.1145/2914920.2915008Google Scholar
Digital Library
- Heather Culbertson, Julie M. Walker, Michael Raitor, and Allison M. Okamura. 2017. WAVES: A Wearable Asymmetric Vibration Excitation System for Presenting Three-Dimensional Translation and Rotation Cues. In Proceedings of the 2017 CHI Conference on Human Factors in Computing Systems , Vol. 2017-May. ACM, New York, NY, USA, 4972--4982. https://doi.org/10.1145/3025453.3025741Google Scholar
Digital Library
- Kiran Dandekar, Balasundar I. Raju, and Mandayam A. Srinivasan. 2003. 3-D Finite-Element Models of Human and Monkey Fingertips to Investigate the Mechanics of Tactile Sense . Journal of Biomechanical Engineering , Vol. 125, 5 (oct 2003), 682--691. https://doi.org/10.1115/1.1613673Google Scholar
Cross Ref
- Ronald Ecker, Verena Broy, Andreas Butz, and Alexander De Luca. 2009. pieTouch: a direct touch gesture interface for interacting with in-vehicle information systems. In Proceedings of the 11th International Conference on Human-Computer Interaction with Mobile Devices and Services - MobileHCI '09. ACM Press, New York, New York, USA, 1. https://doi.org/10.1145/1613858.1613887Google Scholar
Digital Library
- Nicholas A Giudice, Hari Prasath Palani, Eric Brenner, and Kevin M Kramer. 2012. Learning non-visual graphical information using a touch-based vibro-audio interface. In Proceedings of the 14th international ACM SIGACCESS conference on Computers and accessibility (Boulder, Colorado, USA) (ASSETS '12). Association for Computing Machinery, New York, NY, USA, 103--110. https://doi.org/10.1145/2384916.2384935Google Scholar
Digital Library
- L. H. Goldish and H. E. Taylor. 1974. The Optacon: a valuable device for blind persons. New Outlook for the Blind , Vol. 68, 2 (1974), 49--56. https://doi.org/10.1177/0145482X7406800201Google Scholar
- Daniel Goldreich and Ingrid M. Kanics. 2006. Performance of blind and sighted humans on a tactile grating detection task. Perception & Psychophysics , Vol. 68, 8 (Nov. 2006), 1363--1371. https://doi.org/10.3758/BF03193735Google Scholar
Cross Ref
- Cagatay Goncu and Kim Marriott. 2011. GraVVITAS: Generic Multi-touch Presentation of Accessible Graphics. In Human-Computer Interaction -- INTERACT 2011, Pedro Campos, Nicholas Graham, Joaquim Jorge, Nuno Nunes, Philippe Palanque, and Marco Winckler (Eds.). Springer Berlin Heidelberg, Berlin, Heidelberg, 30--48. https://doi.org/10.1007/978--3--642--23774--4_5Google Scholar
Cross Ref
- Jenna L Gorlewicz, Jennifer L Tennison, P Merlin Uesbeck, Margaret E Richard, Hari P Palani, Andreas Stefik, Derrick W Smith, and Nicholas A Giudice. 2020. Design Guidelines and Recommendations for Multimodal, Touchscreen-based Graphics. ACM Trans. Access. Comput. , Vol. 13, 3 (Aug. 2020), 1--30. https://doi.org/10.1145/3403933Google Scholar
Digital Library
- Jay Henderson, Jeff Avery, Laurent Grisoni, and Edward Lank. 2019. Leveraging Distal Vibrotactile Feedback for Target Acquisition. In Proceedings of the 2019 CHI Conference on Human Factors in Computing Systems (Glasgow, Scotland Uk) (CHI '19). Association for Computing Machinery, New York, NY, USA, 1--11. https://doi.org/10.1145/3290605.3300715Google Scholar
Digital Library
- Jonggi Hong, Alisha Pradhan, Jon E Froehlich, and Leah Findlater. 2017. Evaluating Wrist-Based Haptic Feedback for Non-Visual Target Finding and Path Tracing on a 2D Surface. In Proceedings of the 19th International ACM SIGACCESS Conference on Computers and Accessibility (Baltimore, Maryland, USA) (ASSETS '17). Association for Computing Machinery, New York, NY, USA, 210--219. https://doi.org/10.1145/3132525.3132538Google Scholar
Digital Library
- R D Jacobson. 1998. Navigating maps with little or no sight: An audio-tactile approach. Proceedings of the Workshop on Content Visualization and Intermedia Representations (CVIR) (1998), 95--102. https://aclanthology.org/W98-0214Google Scholar
- Charles F. Jekel, Gerhard Venter, Martin P. Venter, Nielen Stander, and Raphael T. Haftka. 2019. Similarity measures for identifying material parameters from hysteresis loops using inverse analysis. International Journal of Material Forming , Vol. 12, 3 (May 2019), 355--378. https://doi.org/10.1007/s12289-018--1421--8Google Scholar
Cross Ref
- Hwan Kim, HyeonBeom Yi, Hyein Lee, and Woohun Lee. 2018. HapCube: A Wearable Tactile Device to Provide Tangential and Normal Pseudo-Force Feedback on a Fingertip .Association for Computing Machinery, New York, NY, USA, 1--13. https://doi.org/10.1145/3173574.3174075Google Scholar
Digital Library
- Roberta L. Klatzky, Nicholas A. Giudice, Christopher R. Bennett, and Jack M. Loomis. 2014. Touch-Screen Technology for the Dynamic Display of 2D Spatial Information Without Vision: Promise and Progress. Multisensory Research , Vol. 27, 5--6 (2014), 359 -- 378. https://doi.org/10.1163/22134808-00002447Google Scholar
Cross Ref
- Pedro Lopes, Douaa Yüksel, Franccois Guimbretière, and Patrick Baudisch. 2016. Muscle-Plotter: An Interactive System Based on Electrical Muscle Stimulation That Produces Spatial Output. In Proceedings of the 29th Annual Symposium on User Interface Software and Technology (Tokyo, Japan) (UIST '16). Association for Computing Machinery, New York, NY, USA, 207--217. https://doi.org/10.1145/2984511.2984530Google Scholar
Digital Library
- Christopher McAdam and Stephen Brewster. 2009. Distal Tactile Feedback for Text Entry on Tabletop Computers. In Proceedings of the 23rd British HCI Group Annual Conference on People and Computers: Celebrating People and Technology (Cambridge, United Kingdom) (BCS-HCI '09). BCS Learning & Development Ltd., Swindon, GBR, 504--511. https://doi.org/10.14236/ewic/HCI2009.63Google Scholar
Cross Ref
- Hari P. Palani, Paul D. S. Fink, and Nicholas A. Giudice. 2020. Design Guidelines for Schematizing and Rendering Haptically Perceivable Graphical Elements on Touchscreen Devices . International Journal of Human--Computer Interaction , Vol. 36, 15 (Sept. 2020), 1393--1414. https://doi.org/10.1080/10447318.2020.1752464 Publisher: Taylor & Francis.Google Scholar
Cross Ref
- Sylvain Pauchet, Catherine Letondal, Jean-Luc Vinot, Mickaë l Causse, Mathieu Cousy, Valentin Becquet, and Guillaume Crouzet. 2018. GazeForm: Dynamic gaze-adaptive touch surface for eyes-free interaction in airliner cockpits. In Proceedings of the 2018 on Designing Interactive Systems Conference 2018 - DIS '18. ACM Press, New York, New York, USA, 1193--1205. https://doi.org/10.1145/3196709.3196712Google Scholar
Digital Library
- Martin Pielot, Benjamin Poppinga, Wilko Heuten, and Susanne Boll. 2011. A Tactile Compass for Eyes-Free Pedestrian Navigation. In Human-Computer Interaction -- INTERACT 2011, Pedro Campos, Nicholas Graham, Joaquim Jorge, Nuno Nunes, Philippe Palanque, and Marco Winckler (Eds.). Springer Berlin Heidelberg, Berlin, Heidelberg, 640--656. https://doi.org/10.1007/978--3--642--23771--3_47Google Scholar
Digital Library
- Jun Rekimoto. 2013. Traxion: A Tactile Interaction Device with Virtual Force Sensation. In Proceedings of the 26th Annual ACM Symposium on User Interface Software and Technology (St. Andrews, Scotland, United Kingdom) (UIST '13). Association for Computing Machinery, New York, NY, USA, 427--432. https://doi.org/10.1145/2501988.2502044Google Scholar
Digital Library
- Marzia Taibbi, Cristian Bernareggi, Andrea Gerino, Dragan Ahmetovic, and Sergio Mascetti. 2014. AudioFunctions: Eyes-Free Exploration of Mathematical Functions on Tablets. In Computers Helping People with Special Needs, Klaus Miesenberger, Deborah Fels, Dominique Archambault, Petr Pevn áz, and Wolfgang Zagler (Eds.). Springer International Publishing, Cham, 537--544. https://doi.org/10.1007/978--3--319-08596--8_84Google Scholar
- Jennifer L. Tennison and Jenna L. Gorlewicz. 2016. Toward Non-visual Graphics Representations on Vibratory Touchscreens: Shape Exploration and Identification. In Haptics: Perception, Devices, Control, and Applications, Fernando Bello, Hiroyuki Kajimoto, and Yon Visell (Eds.). Springer International Publishing, Cham, 384--395. https://doi.org/10.1007/978--3--319--42324--1_38Google Scholar
- Jennifer L Tennison and Jenna L Gorlewicz. 2019. Non-visual Perception of Lines on a Multimodal Touchscreen Tablet. ACM Trans. Appl. Percept. , Vol. 16, 1 (Feb. 2019), 1--19. https://doi.org/10.1145/3301415Google Scholar
Digital Library
- E Yoo, E Park, and B Chung. 2001. Mental practice effect on line-tracing accuracy in persons with hemiparetic stroke: a preliminary study. Arch. Phys. Med. Rehabil. , Vol. 82, 9 (Sept. 2001), 1213--1218. https://doi.org/10.1053/apmr.2001.25095Google Scholar
Cross Ref
- Stanislaw Zabramski. 2011. Careless Touch: A Comparative Evaluation of Mouse, Pen, and Touch Input in Shape Tracing Task. In Proceedings of the 23rd Australian Computer-Human Interaction Conference (Canberra, Australia) (OzCHI '11). Association for Computing Machinery, New York, NY, USA, 329--332. https://doi.org/10.1145/2071536.2071588Google Scholar
Digital Library
- Stanislaw Zabramski, Suman Shrestha, and Wolfgang Stuerzlinger. 2013. Easy vs. Tricky: The Shape Effect in Tracing, Selecting, and Steering With Mouse, Stylus, and Touch. In Proceedings of International Conference on Making Sense of Converging Media (Tampere, Finland) (AcademicMindTrek '13). Association for Computing Machinery, New York, NY, USA, 99--103. https://doi.org/10.1145/2523429.2523444Google Scholar
Digital Library
- Kaixing Zhao, Marcos Serrano, Bernard Oriola, and Christophe Jouffrais. 2020. VibHand: On-Hand Vibrotactile Interface Enhancing Non-Visual Exploration of Digital Graphics. Proc. ACM Hum.-Comput. Interact. , Vol. 4, ISS, Article 207 (Nov. 2020), bibinfonumpages19 pages. https://doi.org/10.1145/3427335Google Scholar
Digital Library
- Shengdong Zhao, Pierre Dragicevic, Mark Chignell, Ravin Balakrishnan, and Patrick Baudisch. 2007. Earpod: Eyes-free menu selection using touch input and reactive audio feedback. In Proceedings of the SIGCHI Conference on Human Factors in Computing Systems - CHI '07 . ACM Press, New York, New York, USA, 1395--1404. https://doi.org/10.1145/1240624.1240836Google Scholar
Digital Library
Index Terms
Composite Line Designs and Accuracy Measurements for Tactile Line Tracing on Touch Surfaces
Recommendations
Guided touch screen: enhanced eyes-free interaction
PerDis '16: Proceedings of the 5th ACM International Symposium on Pervasive DisplaysTouch screen based interfaces are pervading into areas such as car and industrial machinery control systems, where more tangible, physical interfaces have historically been used. This often results in a reduction in the possibility to operate the ...
Smart Touch: Improving Touch Accuracy for People with Motor Impairments with Template Matching
CHI '16: Proceedings of the 2016 CHI Conference on Human Factors in Computing SystemsWe present two contributions toward improving the accessibility of touch screens for people with motor impairments. First, we provide an exploration of the touch behaviors of 10 people with motor impairments, e.g., we describe how touching with the back ...
Direct manipulation video navigation on touch screens
MobileHCI '14: Proceedings of the 16th international conference on Human-computer interaction with mobile devices & servicesDirect Manipulation Video Navigation (DMVN) systems allow a user to directly drag an object of interest along its motion trajectory and have been shown effective for space-centric video browsing tasks. This paper designs touch-based interface techniques ...






Comments