Abstract

The human hand dexterity provides a natural and rich interface for computer interaction. Technologies for tracking hand and finger motion are now accessible and can be used with a minimum computation power for the manipulation of virtual puppets in real-time. However creating an intuitive hand-based motion con- trol interface presents problems such as, mapping the hand to an object that demands more degrees of freedom or to assign gestures that are difficult to memorize or to execute. We propose an ergonomic hand- mapping model for digital puppetry, based on the human hand anatomy and biomechanics, adapting tradi- tional puppetry methods. A cinematic virtual puppetry application was developed supporting distinct in- teraction styles based on the hand dexterity skills. An experiment using the Leap Motion controller was conducted to evaluate the hand mapping feasibility. The participants considered that the proposed interface provides a good level of directness. Furthermore, a custom interface for smartphones was combined with the application to extend the puppet manipulation and show control operations.
- Avola, D., Cinque, L., Levialdi, S., Petracca, A., Placidi, G. and Spezialetti, M. 2014. Markerless Hand Gesture Interface Based on LEAP Motion Controller. In Proceedings of the International Conference on Distributed Multimedia Systems, 260--266.Google Scholar
- Ben, S. 1982. The future of interactive systems and the emergence of direct manipulation. Behaviour & Information Technology 1, 3: 237--256.Google Scholar
- Brand, P.W. and Hollister, A. 1993. Clinical mechanics of the hand. Mosby Year Book, Maryland Heights.Google Scholar
- Bullock, I.M., Ma, R.R. and Dollar, A.M. 2013. A Hand-Centric Classification of Human and Robot Dexterous Manipulation. IEEE Transactions on Haptics 6, 2: 129--144. Google Scholar
Digital Library
- Currell, D. 2004. Making and Manipulating Marionettes. The Crowood Press, Wiltshire.Google Scholar
- Dipietro, L., Sabatini, A.M. and Dario, P. 2008. A Survey of Glove-Based Systems and Their Applications. Trans. Sys. Man Cyber Part C 38, 4: 461--482. Google Scholar
Digital Library
- Guna, J., Jakus, G., Pogačnik, M., Tomazič, S. and Sodnik, J. 2014. An Analysis of the Precision and Reliability of the Leap Motion Sensor and Its Suitability for Static and Dynamic Tracking. Sensors 14, 2: 3702--3720.Google Scholar
Cross Ref
- Hamill, J. and Knutzen, K.M. 2003. Biomechanical Basis of Human Movement. Lippincott Williams & Wilkins, Philadelphia.Google Scholar
- Hutchins, E.L., Hollan, J.D. and Norman, D.A. 1985. Direct Manipulation Interfaces. Human-Computer Interaction 1, 4: 311--338. Google Scholar
Digital Library
- Igarashi, T., Moscovich, T. and Hughes, J.F.J. 2005. As-Rigid-As-Possible Shape Manipulation. Transactions on Computer Graphics 24, 3. Google Scholar
Digital Library
- Jones, L.A. and Lederman, S.J. 2006. Human Hand Function. Oxford University Press, New York City.Google Scholar
- Kazi, R.H., Chua, K.C., Zhao, S., Davis, R. and Low, K.-L. 2011. SandCanvas: A Multi-touch Art Medium Inspired by Sand Animation. In Proceedings of the Proceedings of the SIGCHI Conference on Human Factors in Computing Systems, New York, NY, USA, 1283--1292. Google Scholar
Digital Library
- LaViola, J.J., Jr. 1999. A Survey of Hand Posture and Gesture Recognition Techniques and Technology. Brown University, Providence, RI, USA.Google Scholar
- Leite, L. and Orvalho, V. 2012. Shape Your Body: Control a Virtual Silhouette Using Body Motion. In Proceedings of the CHI '12 Extended Abstracts on Human Factors in Computing Systems, New York, NY, USA, 1913--1918. Google Scholar
Digital Library
- Marionette Zoo. 2014. Retrieved April 21, 2016 from http://mndl.hu/2014-01--26-marionette-zoo.htmlGoogle Scholar
- Momeni, A. and Rispoli, Z. 2015. Dranimate: Rapid Real-time Gestural Rigging and Control of Animation. In Proceedings of the Adjunct Proceedings of the 28th Annual ACM Symposium on User Interface Software Technology, New York, NY, USA, 61--62. Google Scholar
Digital Library
- Nitsche, M. and Nayak, S. 2012. Cell Phone Puppets: Turning Mobile Phones into Performing Objects. In Proceedings of the International Conference on Entertainment Computing, Berlin, Heidelberg, 363--372. Google Scholar
Digital Library
- Norkin, C.C. and White, D.J. 2009. Measurement of Joint Motion: A Guide to Goniometry. F.A. Davis, Philadelphia.Google Scholar
- Okada, Y., Etou, H. and Niijima, K. 2004. Intuitive Interfaces for Motion Generation and Search. In Proceedings of the Intuitive Human Interfaces for Organizing and Accessing Intellectual Assets, 49--67. Google Scholar
Digital Library
- Oshita, M. 2012. Multi-Touch Interface for Character Motion Control Using Example-Based Posture Synthesis. In Proceedings of the International Conference on Computer Graphics, Visualization and Computer Vision, 1--9.Google Scholar
- Palastanga, N., Field, D. and Soames, R.W. 2013. Anatomy and Human Movement: Structure and Function. Elsevier Science.Google Scholar
- Premaratne, P. and Premaratne, P. 2014. Cognitive Science and Technology. Springer Singapore, Singapore.Google Scholar
- Puppet in Motion. 2013. Sixense. Retrieved May 19, 2015 from http://sixense.com/puppetinmotionGoogle Scholar
- Ryu, J., Cooney, W.P., III, Askew, L.J., An, K.-N. and Chao, E.Y.S. 2015. Functional ranges of motion of the wrist joint. Journal of Hand Surgery 16, 3: 409--419.Google Scholar
Cross Ref
- Sridhar, S., Feit, A.M., Theobalt, C. and Oulasvirta, A. 2015. Investigating the Dexterity of Multi-Finger Input for Mid-Air Text Entry. In Proceedings of the ACM Conference on Human Factors in Computing Systems, New York, NY, USA, 3643--3652. Google Scholar
Digital Library
- Stringless - Remote Control for Unity3D:. 2016. Stringless - Remote Control for Unity3D. github.com. Retrieved April 2016 from https://github.com/grifu/StringlessUnityGoogle Scholar
- Sturman, D. 1992. Whole-hand Input. Massachuster Institute of Technology.Google Scholar
- Sturman, D. and Zeltzer, D. 1994. A Survey of Glove-based lnput. IEEE Computer Graphics and Applications: 1--10. Google Scholar
Digital Library
- Vardy, A. 1998. Articulated Human Hand Model with Inter-Joint Dependency Constraints. Computer Science 6752.Google Scholar
Index Terms
Mani-Pull-Action: Hand-based Digital Puppetry
Recommendations
Shape your body: control a virtual silhouette using body motion
CHI EA '12: CHI '12 Extended Abstracts on Human Factors in Computing SystemsIn this paper we propose to use our body as a puppetry controller, giving life to a virtual silhouette through acting. A framework was deployed based on Microsoft Kinect using OpenNI and Unity to animate in real-time a silhouette. This was used to ...
Virtual marionette
IUI '12: Proceedings of the 2012 ACM international conference on Intelligent User InterfacesVirtual Marionette is a research on digital puppetry, an interdisciplinary approach that brings the art of puppetry into the world of digital animation. Inspired in the traditional marionette technology our intention is to study novel interfaces as an ...
Anim-actor: understanding interaction with digital puppetry using low-cost motion capture
ACE '11: Proceedings of the 8th International Conference on Advances in Computer Entertainment TechnologyCharacter animation has traditionally relied on complex and time-consuming key frame animation or expensive motion capture techniques. These methods are out of reach for low budget animation productions. We present a low-cost performance-driven ...






Comments