Abstract
One common task when controlling smart displays is the manipulation of menu items. Given current examples of smart displays that support distant bare hand control, in this paper we explore menu item selection tasks with three different mappings of barehand movement to target selection. Through a series of experiments, we demonstrate that Positional mapping is faster than other mappings when the target is visible but requires many clutches in large targeting spaces. Rate-based mapping is, in contrast, preferred by participants due to its perceived lower effort, despite being slightly harder to learn initially. Tradeoffs in the design of target selection in smart tv displays are discussed.
- Anand Agarawala and Ravin Balakrishnan. 2006. Keepin' It Real: Pushing the Desktop Metaphor with Physics, Piles and the Pen .Association for Computing Machinery, New York, NY, USA, 1283--1292. https://doi.org/10.1145/1124772.1124965Google Scholar
Digital Library
- Takeshi Asano, Ehud Sharlin, Yoshifumi Kitamura, Kazuki Takashima, and Fumio Kishino. 2005. Predictive Interaction Using the Delphian Desktop. In Proceedings of the 18th Annual ACM Symposium on User Interface Software and Technology (Seattle, WA, USA) (UIST '05). Association for Computing Machinery, New York, NY, USA, 133--141. https://doi.org/10.1145/1095034.1095058Google Scholar
Digital Library
- Ravin Balakrishnan. 2004. "Beating" Fitts' law: virtual enhancements for pointing facilitation. International Journal of Human-Computer Studies , Vol. 61, 6 (2004), 857--874. https://doi.org/10.1016/j.ijhcs.2004.09.002 Fitts' law 50 years later: applications and contributions from human-computer interaction.Google Scholar
Digital Library
- Olivier Bau and Wendy E. Mackay. 2008. OctoPocus: A Dynamic Guide for Learning Gesture-Based Command Sets. In Proceedings of the 21st Annual ACM Symposium on User Interface Software and Technology (Monterey, CA, USA) (UIST '08). Association for Computing Machinery, New York, NY, USA, 37--46. https://doi.org/10.1145/1449715.1449724Google Scholar
Digital Library
- R. Bolt. 1980. "Put-that-there": Voice and gesture at the graphics interface. In SIGGRAPH '80 .Google Scholar
Digital Library
- Andy Cockburn and Joshua Savage. 2004. Comparing Speed-dependent Automatic Zooming with Traditional Scroll, Pan and Zoom Methods. In People and Computers XVII -- Designing for Society, Eamonn O'Neill, Philippe Palanque, and Peter Johnson (Eds.). Springer London, London, 87--102.Google Scholar
- Clifton Forlines, Daniel Vogel, and Ravin Balakrishnan. 2006. HybridPointing: Fluid Switching between Absolute and Relative Pointing with a Direct Input Device. In Proceedings of the 19th Annual ACM Symposium on User Interface Software and Technology (Montreux, Switzerland) (UIST '06). Association for Computing Machinery, New York, NY, USA, 211--220. https://doi.org/10.1145/1166253.1166286Google Scholar
Digital Library
- Tovi Grossman and Ravin Balakrishnan. 2005. The Bubble Cursor: Enhancing Target Acquisition by Dynamic Resizing of the Cursor's Activation Area. In Proceedings of the SIGCHI Conference on Human Factors in Computing Systems (Portland, Oregon, USA) (CHI '05). Association for Computing Machinery, New York, NY, USA, 281--290. https://doi.org/10.1145/1054972.1055012Google Scholar
Digital Library
- Yves Guiard, Renaud Blanch, and Michel Beaudouin-Lafon. 2004. Object pointing: a complement to bitmap pointing in GUIs. In Proceedings of Graphics Interface 2004 . Citeseer, 9--16.Google Scholar
- Sandra G Hart and Lowell E Staveland. 1988. Development of NASA-TLX (Task Load Index): Results of empirical and theoretical research. In Advances in psychology . Vol. 52. Elsevier, 139--183.Google Scholar
- Juan David Hincapié-Ramos, Xiang Guo, Paymahn Moghadasian, and Pourang Irani. 2014. Consumed endurance: a metric to quantify arm fatigue of mid-air interactions. Proceedings of the SIGCHI Conference on Human Factors in Computing Systems (2014).Google Scholar
Digital Library
- Takeo Igarashi and Ken Hinckley. 2000. Speed-Dependent Automatic Zooming for Browsing Large Documents. In Proceedings of the 13th Annual ACM Symposium on User Interface Software and Technology (San Diego, California, USA) (UIST '00). Association for Computing Machinery, New York, NY, USA, 139--148. https://doi.org/10.1145/354401.354435Google Scholar
Digital Library
- Keiko Katsuragawa, Krzysztof Pietroszek, James R Wallace, and Edward Lank. 2016. Watchpoint: Freehand pointing with a smartwatch in a ubiquitous display environment. In Proceedings of the International Working Conference on Advanced Visual Interfaces . 128--135.Google Scholar
Digital Library
- Panayiotis Koutsabasis and Panagiotis Vogiatzidakis. 2019. Empirical research in mid-air interaction: A systematic review. International Journal of Human-Computer Interaction , Vol. 35, 18 (2019), 1747--1768.Google Scholar
Cross Ref
- I. MacKenzie. 1992. Fitts' law as a performance model in human-computer interaction.Google Scholar
- Sven Mayer, Valentin Schwind, Robin Schweigert, and Niels Henze. 2018. The effect of offset correction and cursor on mid-air pointing in real and virtual environments. In Proceedings of the 2018 CHI Conference on Human Factors in Computing Systems. 1--13.Google Scholar
Digital Library
- Michael McGuffin and Ravin Balakrishnan. 2002. Acquisition of Expanding Targets. In Proceedings of the SIGCHI Conference on Human Factors in Computing Systems (Minneapolis, Minnesota, USA) (CHI '02). Association for Computing Machinery, New York, NY, USA, 57--64. https://doi.org/10.1145/503376.503388Google Scholar
Digital Library
- Mathieu Nancel, Emmanuel Pietriga, Olivier Chapuis, and Michel Beaudouin-Lafon. 2015. Mid-air pointing on ultra-walls. ACM Transactions on Computer-Human Interaction (TOCHI) , Vol. 22, 5 (2015), 1--62.Google Scholar
Digital Library
- Krzysztof Pietroszek, Liudmila Tahai, James R Wallace, and Edward Lank. 2017. Watchcasting: Freehand 3D interaction with off-the-shelf smartwatch. In 2017 IEEE Symposium on 3D User Interfaces (3DUI). IEEE, 172--175.Google Scholar
Cross Ref
- Shaishav Siddhpuria, Sylvain Malacria, Mathieu Nancel, and Edward Lank. 2018. Pointing at a distance with everyday smart devices. In Proceedings of the 2018 CHI Conference on Human Factors in Computing Systems. 1--11.Google Scholar
Digital Library
- Rafael Veras, Gaganpreet Singh, Farzin Farhadi-Niaki, Ritesh Udhani, Parth Pradeep Patekar, Wei Zhou, Pourang Irani, and Wei Li. 2021. Elbow-Anchored Interaction: Designing Restful Mid-Air Input .Association for Computing Machinery, New York, NY, USA. https://doi.org/10.1145/3411764.3445546Google Scholar
Digital Library
- Daniel Vogel and Ravin Balakrishnan. 2005. Distant Freehand Pointing and Clicking on Very Large, High Resolution Displays. In Proceedings of the 18th Annual ACM Symposium on User Interface Software and Technology (Seattle, WA, USA) (UIST '05). Association for Computing Machinery, New York, NY, USA, 33--42. https://doi.org/10.1145/1095034.1095041Google Scholar
Digital Library
- Fan Zhang, Valentin Bazarevsky, Andrey Vakunov, Andrei Tkachenka, George Sung, Chuo-Ling Chang, and Matthias Grundmann. 2020. MediaPipe Hands: On-device Real-time Hand Tracking. arxiv: 2006.10214 [cs.CV]Google Scholar
- Futian Zhang, Sachi Mizobuchi, Wei Zhou, Taslim Arefin Khan, Wei Li, and Edward Lank. 2021. Leveraging CD Gain for Precise Barehand Video Timeline Browsing on Smart Displays. In Human-Computer Interaction -- INTERACT 2021 . Springer International Publishing, Cham, 72--91.Google Scholar
Index Terms
Analyzing Midair Object Pointing Mappings for Smart Display Input
Recommendations
Crossing-based selection with direct touch input
CHI '14: Proceedings of the SIGCHI Conference on Human Factors in Computing SystemsFundamental performance results for crossing-based selec-tion tasks with direct touch input are presented. A close adaptation of Accot and Zhai's indirect stylus crossing ex-periment reveals similar trends for direct touch input: touch crossing task ...
Midair touch display
SIGGRAPH '15: ACM SIGGRAPH 2015 Emerging TechnologiesWe demonstrate a system which enables users with bare hands and naked eyes to tactually interact with midair three-dimensional objects. The displayed midair images are three-dimensional and can be seen and touched from multiple angles. By creating ...
Midair displays: exploring the concept of free-floating public displays
CHI EA '14: CHI '14 Extended Abstracts on Human Factors in Computing SystemsDue to advances in technology, displays could replace literally any surface in the future, including walls, windows, and ceilings. At the same time, midair remains a relatively unexplored domain for the use of displays as of today, particularly in ...






Comments