Abstract

The rapid proliferation of Massive Open Online Courses (MOOC) has resulted in many-fold increase in sharing the global classrooms through customized online platforms, where a student can participate in the classes through her personal devices, such as personal computers, smartphones, tablets, etc. However, in the absence of direct interactions with the students during the delivery of the lectures, it becomes difficult to judge their involvements in the classroom. In academics, the degree of student's attention can indicate whether a course is efficacious in terms of clarity and information. An automated feedback can hence be generated to enhance the utility of the course. The precision of discernment in the context of human attention is a subject of surveillance. However, visual patterns indicating the magnitude of concentration can be deciphered by analyzing the visual emphasis and the way an individual visually gesticulates, while contemplating the object of interest. In this paper, we develop a methodology called Gestsatten which captures the learner's attentiveness from his visual gesture patterns. In this approach, the learner's visual gestures are tracked along with the region of focus. We consider two aspects in this approach -- first, we do not transfer learner's video outside her device, so we apply in-device computing to protect her privacy; second, considering the fact that a majority of the learners use handheld devices like smartphones to observe the MOOC videos, we develop a lightweight approach for in-device computation. A three level estimation of learner's attention is performed based on these information. We have implemented and tested Gestatten over 48 participants from different age groups, and we observe that the proposed technique can capture the attention level of a learner with high accuracy (average absolute error rate is 8.68%), which meets her ability to learn a topic as measured through a set of cognitive tests.
- Henny Admoni and Brian Scassellati. 2017. Social eye gaze in human-robot interaction: A review. Journal of Human-Robot Interaction, Vol. 6, 1 (2017), 25--63.Google Scholar
Digital Library
- Jung-Jung Chang, Wen-Shen Lin, and Hong-Ren Chen. 2018. How attention level and cognitive style affect learning in a MOOC environment? Based on the perspective of brainwave analysis. Computers in Human Behavior (2018).Google Scholar
- Dan Chen, Xusheng Tang, Zongying Ou, and Ning Xi. 2006. A Hierarchical Floatboost and MLP Classifier for Mobile Phone Embedded Eye Location System. In Proceedings of the Third International Conference on Advnaces in Neural Networks - Volume Part II. 20--25.Google Scholar
Digital Library
- Yimin Chen, Tao Li, Rui Zhang, Yanchao Zhang, and Terri Hedgpeth. 2018. EyeTell: Video-Assisted Touchscreen Keystroke Inference from Eye Movements. In Proceedings of the IEEE Symposium on Security and Privacy. 144--160. https://doi.org/10.1109/SP.2018.00010Google Scholar
Cross Ref
- Derrick Coetzee, Armando Fox, Marti A Hearst, and Björn Hartmann. 2014. Should your MOOC forum use a reputation system?. In Proceedings of the 17th ACM conference on Computer supported cooperative work & social computing. ACM, 1176--1187.Google Scholar
Digital Library
- A. Dasgupta, A. George, S. L. Happy, and A. Routray. 2013. A Vision-Based System for Monitoring the Loss of Attention in Automotive Drivers. IEEE Transactions on Intelligent Transportation Systems, Vol. 14, 4 (Dec 2013), 1825--1838. https://doi.org/10.1109/TITS.2013.2271052Google Scholar
Digital Library
- Alexander De Luca, Martin Denzel, and Heinrich Hussmann. 2009. Look into My Eyes!: Can You Guess My Password?. In Proceedings of the 5th Symposium on Usable Privacy and Security. Article 7, 12 pages.Google Scholar
Digital Library
- Heiko Drewes, Alexander De Luca, and Albrecht Schmidt. 2007. Eye-gaze Interaction for Mobile Phones. In Proceedings of the 4th International Conference on Mobile Technology, Applications, and Systems and the 1st International Symposium on Computer Human Interaction in Mobile Technology. 364--371.Google Scholar
Digital Library
- Morten Lund Dybdal, Javier San Agustin, and John Paulin Hansen. 2012. Gaze Input for Mobile Devices by Dwell and Gestures. In Proceedings of the Symposium on Eye Tracking Research and Applications. 225--228.Google Scholar
Digital Library
- Inger Ekman, Laura Ermi, Jussi Lahti, Jani Nummela, Petri Lankoski, and Frans Mayra. 2005. Designing sound for a pervasive mobile game. In Proceedings of the 2005 ACM SIGCHI International Conference on Advances in computer entertainment technology. ACM, 110--116.Google Scholar
Digital Library
- Karen Emmorey, Robin Thompson, and Rachael Colvin. 2008. Eye gaze during comprehension of American Sign Language by native and beginning signers. Journal of Deaf Studies and Deaf Education, Vol. 14, 2 (2008), 237--243.Google Scholar
Cross Ref
- Carrie Demmans Epp, Cosmin Munteanu, Benett Axtell, Keerthika Ravinthiran, Yomna Aly, and Elman Mansimov. 2017. Finger tracking: facilitating non-commercial content production for mobile e-reading applications. In Proceedings of the 19th International Conference on Human-Computer Interaction with Mobile Devices and Services. ACM, 34.Google Scholar
Digital Library
- Augusto Esteves, Eduardo Velloso, Andreas Bulling, and Hans Gellersen. 2015. Orbits: Gaze interaction for smart watches using smooth pursuit eye movements. In Proceedings of the 28th Annual ACM Symposium on User Interface Software & Technology. ACM, 457--466.Google Scholar
Digital Library
- Seongwon Han, Sungwon Yang, Jihyoung Kim, and Mario Gerla. 2012. EyeGuardian: a framework of eye tracking and blink detection for mobile device users. In Proceedings of the Twelfth Workshop on Mobile Computing Systems & Applications. ACM, 6.Google Scholar
Digital Library
- Andrew G. Howard, Menglong Zhu, Bo Chen, Dmitry Kalenichenko, Weijun Wang, Tobias Weyand, Marco Andreetto, and Hartwig Adam. 2017. MobileNets: Efficient Convolutional Neural Networks for Mobile Vision Applications. CoRR, Vol. abs/1704.04861 (2017). arxiv: 1704.04861 http://arxiv.org/abs/1704.04861Google Scholar
- Jonathan Huang, Vivek Rathod, Chen Sun, Menglong Zhu, Anoop Korattikara, Alireza Fathi, Ian Fischer, Zbigniew Wojna, Yang Song, Sergio Guadarrama, and Kevin Murphy. 2016. Speed/accuracy trade-offs for modern convolutional object detectors. CoRR, Vol. abs/1611.10012 (2016). arxiv: 1611.10012 http://arxiv.org/abs/1611.10012Google Scholar
- Han Jiang, Karmen Dykstra, and Jacob Whitehill. 2018. Predicting When Teachers Look at Their Students in 1-on-1 Tutoring Sessions. In Proceedings of the 2018 13th IEEE International Conference on Automatic Face & Gesture Recognition. IEEE, 593--598.Google Scholar
Cross Ref
- Despina Kamilali and Chryssa Sofianopoulou. 2015. Microlearning as Innovative Pedagogy for Mobile Learning in MOOCs. International Association for Development of the Information Society (2015).Google Scholar
- Jari Kangas, Deepak Akkil, Jussi Rantala, Poika Isokoski, P"aivi Majaranta, and Roope Raisamo. 2014. Gaze gestures and haptic feedback in mobile devices. In Proceedings of the SIGCHI Conference on Human Factors in Computing Systems. ACM, 435--438.Google Scholar
Digital Library
- Moritz Kassner, William Patera, and Andreas Bulling. 2014. Pupil: an open source platform for pervasive eye tracking and mobile gaze-based interaction. In Proceedings of the 2014 ACM international joint conference on pervasive and ubiquitous computing: Adjunct publication. ACM, 1151--1160.Google Scholar
Digital Library
- Gregor Kennedy, Carleton Coffrin, Paula De Barba, and Linda Corrin. 2015. Predicting success: how learners' prior knowledge, skills and activities predict MOOC performance. In Proceedings of the fifth international conference on learning analytics and knowledge. ACM, 136--140.Google Scholar
Digital Library
- Kyle Krafka, Aditya Khosla, Petr Kellnhofer, Harini Kannan, Suchendra Bhandarkar, Wojciech Matusik, and Antonio Torralba. 2016. Eye Tracking for Everyone. IEEE Conference on Computer Vision and Pattern Recognition (CVPR) (2016).Google Scholar
- Abhishek Kumar, Kushal Srivastava, Kuldeep Yadav, and Om Deshmukh. 2017. Multi-faceted Index driven Navigation for Educational Videos in Mobile Phones. In Proceedings of the 22nd International Conference on Intelligent User Interfaces. ACM, 357--361.Google Scholar
Digital Library
- Manu Kumar, Jeff Klingner, Rohan Puranik, Terry Winograd, and Andreas Paepcke. [n. d.]. Improving the accuracy of gaze input for interaction. In Proceedings of the 2008 symposium on Eye tracking research & applications. 65--68.Google Scholar
- Tsung-Yi Lin, Michael Maire, Serge Belongie, James Hays, Pietro Perona, Deva Ramanan, Piotr Dollár, and C Lawrence Zitnick. 2014. Microsoft coco: Common objects in context. In Proceedings of the European Conference on Computer Vision. Springer, 740--755.Google Scholar
Cross Ref
- Dachuan Liu, Bo Dong, Xing Gao, and Haining Wang. 2015b. Exploiting Eye Tracking for Smartphone Authentication. Applied Cryptography and Network Security (2015), 457--477.Google Scholar
- Wei Liu, Dragomir Anguelov, Dumitru Erhan, Christian Szegedy, Scott E. Reed, Cheng-Yang Fu, and Alexander C. Berg. 2015a. SSD: Single Shot MultiBox Detector. CoRR, Vol. abs/1512.02325 (2015). arxiv: 1512.02325 http://arxiv.org/abs/1512.02325Google Scholar
- Alexander Mariakakis, Mayank Goel, Md Tanvir Islam Aumi, Shwetak N. Patel, and Jacob O. Wobbrock. 2015. SwitchBack: Using Focus and Saccade Tracking to Guide Users' Attention for Mobile Task Resumption. In Proceedings of the 33rd Annual ACM Conference on Human Factors in Computing Systems. 2953--2962.Google Scholar
- Ann McNamara, Katerina Mania, George Koulieris, and Laurent Itti. 2014. Attention-aware rendering, mobile graphics and games. In ACM SIGGRAPH 2014 Courses. ACM, 6.Google Scholar
Digital Library
- Emiliano Miluzzo, Tianyu Wang, and Andrew T. Campbell. 2010. EyePhone: Activating Mobile Phones with Your Eyes. In Proceedings of the Second ACM SIGCOMM Workshop on Networking, Systems, and Applications on Mobile Handhelds. 15--20.Google Scholar
- Takashi Nagamatsu, Michiya Yamamoto, and Hiroshi Sato. 2010. MobiGaze: Development of a Gaze Interface for Handheld Mobile Devices. In CHI '10 Extended Abstracts on Human Factors in Computing Systems. 3349--3354.Google Scholar
Digital Library
- Mohammad Othman, Telmo Amaral, Róisin McNaney, Jan D Smeddinck, John Vines, and Patrick Olivier. 2017. CrowdEyes: crowdsourcing for robust real-world mobile eye tracking. In Proceedings of the 19th International Conference on Human-Computer Interaction with Mobile Devices and Services. ACM, 18.Google Scholar
Digital Library
- Alexandra Papoutsaki, Patsorn Sangkloy, James Laskey, Nediyana Daskalova, Jeff Huang, and James Hays. 2016. Webgazer: Scalable webcam eye tracking using user interactions. In Proceedings of the Twenty-Fifth International Joint Conference on Artificial Intelligence-IJCAI 2016 .Google Scholar
- Tomislav Pejsa, Sean Andrist, Michael Gleicher, and Bilge Mutlu. 2015. Gaze and attention management for embodied conversational agents. ACM Transactions on Interactive Intelligent Systems, Vol. 5, 1 (2015), 3.Google Scholar
Digital Library
- Ken Pfeuffer, Melodie Vidal, Jayson Turner, Andreas Bulling, and Hans Gellersen. 2013. Pursuit calibration: Making gaze calibration less tedious and more flexible. In Proceedings of the 26th annual ACM symposium on User interface software and technology. ACM, 261--270.Google Scholar
Digital Library
- Phuong Pham and Jingtao Wang. 2016. Adaptive review for mobile MOOC learning via implicit physiological signal sensing. In Proceedings of the 18th ACM International Conference on Multimodal Interaction. ACM, 37--44.Google Scholar
Digital Library
- Yasuhito Sawahata, Rajiv Khosla, Kazuteru Komine, Nobuyuki Hiruma, Takayuki Itou, Seiji Watanabe, Yuji Suzuki, Yumiko Hara, and Nobuo Issiki. 2008. Determining comprehension and quality of TV programs using eye-gaze tracking. Pattern Recognition, Vol. 41, 5 (2008), 1610--1626.Google Scholar
Digital Library
- Kshitij Sharma. 2015. Gaze analysis methods for learning analytics. Technical Report. EPFL.Google Scholar
- Kshitij Sharma, Hamed S. Alavi, Patrick Jermann, and Pierre Dillenbourg. 2016. A Gaze-based Learning Analytics Model: In-video Visual Feedback to Improve Learner's Attention in MOOCs. In Proceedings of the Sixth International Conference on Learning Analytics & Knowledge (LAK '16). ACM, New York, NY, USA, 417--421. https://doi.org/10.1145/2883851.2883902Google Scholar
Digital Library
- Manbir Sodhi, Bryan Reimer, and Ignacio Llamazares. 2002. Glance analysis of driver eye movements to evaluate distraction. Behavior Research Methods, Instruments, & Computers 34.4 (2002), 529--538.Google Scholar
Cross Ref
- Ayse Saliha Sunar, Su White, Nor Aniza Abdullah, and Hugh C Davis. 2017. How learners' interactions sustain engagement: a MOOC case study. IEEE Transactions on Learning Technologies, Vol. 10, 4 (2017), 475--487.Google Scholar
Digital Library
- Fabian Timm and Erhardt Barth. 2011. Accurate Eye Centre Localisation by Means of Gradients. Visapp 11 (2011), 125--130.Google Scholar
- Marc Tonsen, Julian Steil, Yusuke Sugano, and Andreas Bulling. 2017. Invisibleeye: Mobile eye tracking using multiple low-resolution cameras and learning-based gaze estimation. Proceedings of the ACM on Interactive, Mobile, Wearable and Ubiquitous Technologies, Vol. 1, 3 (2017), 106.Google Scholar
Digital Library
- P Viola and M Jones. 2001. Rapid object detection using a boosted cascade of simple features. In Proceedings of the IEEE Computer Society Conference on Computer Vision and Pattern Recognition, Vol. 1. IEEE Comput. Soc, I--511.Google Scholar
Cross Ref
- Dirk Walther, Ueli Rutishauser, Christof Koch, and Pietro Perona. 2005. Selective visual attention enables learning and recognition of multiple objects in cluttered scenes. Computer Vision and Image Understanding, Vol. 100, 1--2 (2005), 41--63.Google Scholar
Cross Ref
- Jian-Gang Wang, Eric Sung, and Ronda Venkateswarlu. 2003. Eye Gaze Estimation from a Single Image of One Eye. In Proceedings of the IEEE International Conference on Computer Vision .Google Scholar
Cross Ref
- Jacob Whitehill, Kiran Mohan, Daniel Seaton, Yigal Rosen, and Dustin Tingley. 2017. MOOC dropout prediction: How to measure accuracy?. In Proceedings of the Fourth (2017) ACM Conference on Learning@ Scale. ACM, 161--164.Google Scholar
Digital Library
- Phillip Ian Wilson and John Fernandez. 2006. Facial feature detection using Haar classifiers. Journal of Computing Sciences in Colleges, Vol. 21, 4 (2006), 127--133.Google Scholar
Digital Library
- Dan Witzner Hansen and Qiang Ji. 2010. In the Eye of the Beholder- A Survey of Models for Eyes and Gaze. IEEE Transactions on Pattern Analysis and Machine Intelligence, Vol. 32, 3 (2010), 478 -- 500. https://doi.org/10.1109/TPAMI.2009.30Google Scholar
Digital Library
- Erroll Wood and Andreas Bulling. 2014. EyeTab: Model-based Gaze Estimation on Unmodified Tablet Computers. In Proceedings of the Symposium on Eye Tracking Research and Applications. 207--210.Google Scholar
Digital Library
- Xiang Xiao and Jingtao Wang. 2015. Towards attentive, bi-directional MOOC learning on mobile devices. In Proceedings of the 2015 ACM on International Conference on Multimodal Interaction. ACM, 163--170.Google Scholar
Digital Library
- Xiang Xiao and Jingtao Wang. 2016. Context and cognitive state triggered interventions for mobile MOOC learning. In Proceedings of the 18th ACM International Conference on Multimodal Interaction. ACM, 378--385.Google Scholar
Digital Library
- Xiang Xiao and Jingtao Wang. 2017. Understanding and detecting divided attention in mobile MOOC learning. In Proceedings of the ACM Conference on Human Factors in Computing Systems. 2411--2415.Google Scholar
- Wei Zhang, Xujun Huang, Shengming Wang, Jiangbo Shu, Hai Liu, and Hao Chen. 2017a. Student performance prediction via online learning behavior analytics. In 2017 International Symposium on Educational Technology (ISET). IEEE, 153--157.Google Scholar
Cross Ref
- Xucong Zhang, Yusuke Sugano, and Andreas Bulling. 2017b. Everyday Eye Contact Detection Using Unsupervised Gaze Target Discovery. Proceedings of the 30th Annual ACM Symposium on User Interface Software and Technology (2017), 193--203. https://doi.org/10.1145/3126594.3126614Google Scholar
Digital Library
- Yue Zhao, Tarmo Robal, Christoph Lofi, and Claudia Hauff. 2018. Stationary vs. Non-stationary Mobile Learning in MOOCs. In Proceedings of the Adjunct Publication of the 26th Conference on User Modeling, Adaptation and Personalization. ACM, 299--303.Google Scholar
Digital Library
- Zhiwei Zhu and Qiang Ji. 2004. Eye and gaze tracking for interactive graphic display. Machine Vision and Applications, Vol. 15, 3 (01 Jul 2004), 139--148. https://doi.org/10.1007/s00138-004-0139--4Google Scholar
Digital Library
Index Terms
Gestatten: Estimation of User's Attention in Mobile MOOCs From Eye Gaze and Gaze Gesture Tracking
Recommendations
Eye Gaze Tracking under Natural Head Movements
CVPR '05: Proceedings of the 2005 IEEE Computer Society Conference on Computer Vision and Pattern Recognition (CVPR'05) - Volume 1 - Volume 01Most available remote eye gaze trackers based on Pupil Center Corneal Reflection (PCCR) technique have two characteristics that prevent them from being widely used as an important computer input device for human computer interaction. First, they must ...
Gaze Speedup: Eye Gaze Assisted Gesture Typing in Virtual Reality
IUI '23: Proceedings of the 28th International Conference on Intelligent User InterfacesMid-air text input in augmented or virtual reality (AR/VR) is an open problem. One proposed solution is gesture typing where the user performs a gesture trace over the keyboard. However, this requires the user to move their hands precisely and ...
Supportiveness of language MOOCs for self-regulated learning: a review of commercial language MOOCs on the market
MOOCs have been frequently applied as an effective approach to language education, especially when they can support self-regulated learning. However, few studies have discussed the supportiveness of using MOOCs for language education (i.e. language MOOCs) ...






Comments