Location via proxy:   [ UP ]  
[Report a bug]   [Manage cookies]                
skip to main content
research-article

Gliding and saccadic gaze gesture recognition in real time

Published: 13 January 2012 Publication History

Abstract

Eye movements can be consciously controlled by humans to the extent of performing sequences of predefined movement patterns, or 'gaze gestures'. Gaze gestures can be tracked noninvasively employing a video-based eye tracking system. Gaze gestures hold the potential to become an emerging input paradigm in the context of human-computer interaction (HCI) as low-cost eye trackers become more ubiquitous. The viability of gaze gestures as an innovative way to control a computer rests on how easily they can be assimilated by potential users and also on the ability of machine learning algorithms to discriminate in real time intentional gaze gestures from typical gaze activity performed during standard interaction with electronic devices. In this work, through a set of experiments and user studies, we evaluate the performance of two different gaze gestures modalities, gliding gaze gestures and saccadic gaze gestures, and their corresponding real-time recognition algorithms, Hierarchical Temporal Memory networks and the Needleman-Wunsch algorithm for sequence alignment. Our results show that a specific combination of gaze gesture modality, namely saccadic gaze gestures, and recognition algorithm, Needleman-Wunsch, allows for reliable usage of intentional gaze gestures to interact with a computer with accuracy rates higher than 95% and completion speeds of around 1.5 to 2.5 seconds per gesture. The optimal gaze gesture modality and recognition algorithm do not interfere with otherwise standard human-computer gaze interaction, generating very few false positives during real time recognition and positive feedback from the users. These encouraging results and the low cost eye tracking equipment used, open up a new HCI paradigm for the fields of accessibility and interaction with smartphones, tablets, projected displays and traditional desktop computers.

References

[1]
Bee, N. and André, E. 2008. Writing with your eye: A dwell time free writing system adapted to the nature of human eye gaze. In Proceedings of the 4th IEEE Tutorial and Research Workshop on Perception and Interactive Technologies for Speech-Based Systems. Springer-Verlag, 111--122.
[2]
Bleiweiss, A., Eshar, D., Kutliroff, G., Lerner, A., Oshrat, Y., and Yanai, Y. 2010. Enhanced interactive gaming by blending full-body tracking and gesture animation. In Proceedings of ACM SIGGRAPH ASIA Sketches (SA'10). ACM, New York, NY, 34:1--34:2.
[3]
Bulling, A., Roggen, D., and Troster, G. 2008. It's in your eyes: towards context-awareness and mobile hci using wearable eog goggles. In Proceedings of the 10th International Conference on Ubiquitous Computing (UbiComp'08). ACM, New York, NY, 84--93.
[4]
Chau, M. and Betke, M. 2005. Real time eye tracking and blink detection with usb cameras. Boston U. Compu. Sci. 12, 1--10.
[5]
Drewes, H. 2010. Eye gaze tracking for human computer interaction. Ph.D. thesis, Ludwig-Maximilians-Universität München.
[6]
Drewes, H. and Schmidt, A. 2007. Interacting with the computer using gaze gestures. In Proceedings of the 11th IFIP TC 13 International Conference on Human-Computer Interaction (INTERACT'07). Springer-Verlag, Berlin, 475--488.
[7]
Duchowski, A. 2007. Eye Tracking Methodology: Theory and Practice. Springer-Verlag, Berlin, USA.
[8]
George, D. and Hawkins, J. 2005. A hierarchical bayesian model of invariant pattern recognition in the visual cortex. In Proceedings of the IEEE International Joint Conference on Neural Networks (IJCNN'05). Vol. 3. 1812--1817.
[9]
George, D. and Hawkins, J. 2009. Towards a mathematical theory of cortical micro-circuits. PLoS Comput. Biol 5, 10, e1000532.
[10]
George, D. and Jarosy, B. 2007. The HTM learning algorithms. Tech. rep., Numenta.
[11]
Gupta, L. and Suwei, M. 2001. Gesture-based interaction and communication: automated classification of hand gesture contours. IEEE Trans. Syst. Man, Cyber. Part C: Appl. Rev. 31, 1, 114--120.
[12]
Hennessey, C., Noureddin, B., and Lawrence, P. 2008. Fixation precision in high-speed noncontact eye-gaze tracking. IEEE Trans. Syst. Man, Cyber. Part B: Cybern. 38, 2, 289--298.
[13]
Holden, E., Owens, R., and Geoffrey, G. 1999. Adaptive fuzzy expert system for sign recognition. In Proceedings of the International Conference on Signal and Image Processing (SIP'00).
[14]
Isokoski, P. 2000. Text input methods for eye trackers using off-screen targets. In Proceedings of the Symposium on Eye Tracking Research and Applications (ETRA'00). ACM, New York, NY, 15--21.
[15]
Istance, H., Hyrskykari, A., Immonen, L., Mansikkamaa, S., and Vickers, S. 2010. Designing gaze gestures for gaming: an investigation of performance. In Proceedings of the Symposium on Eye-Tracking Research and Applications (ETRA'10). ACM, New York, NY, 323--330.
[16]
Jacob, R. 1991. The use of eye movements in human-computer interaction techniques: what you look at is what you get. ACM Trans. Inf. Syst 9, 2, 152--169.
[17]
Li, D., Babcock, J., and Parkhurst, D. 2006. openEyes: a low-cost head-mounted eye-tracking solution. In Proceedings of the Symposium on Eye Tracking Research and Applications. ACM, 95--100.
[18]
Mackay, D. 2003. Dasher - an efficient keyboard alternative. Adv. Clin. Newr. Rehab. 3, 2, 24.
[19]
Mollenbach, E. 2010. Selection strategies in gaze interaction. Ph.D. thesis, Loughborough University.
[20]
Mollenbach, E., Hansen, J., and Lillholm, M.G., A. G. 2009. Single stroke gaze gestures. In Proceedings of the 27th International Conference Extended Abstracts on Human Factors in Computing Systems (CHI'09). ACM, New York, NY, 4555--4560.
[21]
Mollenbach, E., Lillholm, M., Gail, A., and Hansen, J. 2010. Single gaze gestures. In Proceedings of the Symposium on Eye-Tracking Research and Applications (ETRA'10). 177--180.
[22]
Morency, L.-P., Quattoni, A., and Darrell, T. 2007. Latent-dynamic discriminative models for continuous gesture recognition. Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition. 1--8.
[23]
Needleman, S. and Wunsch, C. 1970. A general method applicable to the search for similarities in the amino acid sequence of two proteins. J. Molec. Biol. 48, 3, 443--453.
[24]
Näätänen, R. and Koskinen, P. 1975. Simple reaction time with very small imperative- stimulus probabilities. Acta Psychologica 39, 1, 43--50.
[25]
Qvarfordt, P. and Zhai, S. 2005. Conversing with the user based on eye-gaze patterns. In Proceedings of the SIGCHI Conference on Human Factors in Computing Systems. ACM, New York, NY, 221--230.
[26]
Robinson, D. 1965. The mechanics of human smooth pursuit eye movement. J. Physiol. 180, 3, 569--591.
[27]
Rozado, D., Rodriguez, F., and Varona, P. 2010. Optimizing hierarchical temporal memory for multivariable time series. In Proceedings of the International Conference on Artificial Neural Networks (ICANN'10). K. Diamantaras, W. Duch, and L. Iliadis, Eds., Lecture Notes in Computer Science, vol. 6353. Springer Berlin, 506--518.
[28]
Rozado, D., Rodriguez, F., and Varona, P. 2011. Gaze gesture recognition with hierarchical temporal memory networks. In Advances in Computational Intelligence, J. Cabestany, I. Rojas, and G. Joya, Eds. Lecture Notes in Computer Science, vol. 6691. Springer Berlin, 1--8.
[29]
San Agustin, J. 2010. Off-the-shelf gaze interaction. Ph.D. thesis, IT University of Copenhagen.
[30]
San Agustin, J., Skovsgaard, H., Mollenbach, E., Barret, M., Tall, M., Hansen, D., and Hansen, J. 2010. Evaluation of a low-cost open-source gaze tracker. In Proceedings of the Symposium on Eye-Tracking Research and Applications (ETRA'10). ACM, New York, NY, 77--80.
[31]
Velichkovsky, B. and Hansen, J. P. 1996. New technological windows into mind: There is more in eyes and brains for human-computer interaction. In Proceedings of the International Conference on Computer-Human Interaction (CHI'96). 496--503.
[32]
Ward, D. J. and Mackay, D. J. C. 2002. Fast hands-free writing by gaze direction. Nature 418, 6900.
[33]
Wobbrock, J., Rubinstein, J., Sawyer, M., and Duchowski, A. 2007. Gaze-based creativity not typing but writing: Eye-based text entry using letter-like gestures. In Proceedings of the 3rd Conference on Communication by Gaze Interaction (COGAIN'07).
[34]
Wobbrock, J., Rubinstein, J., Sawyer, M., and Duchowski, A. 2008. Longitudinal evaluation of discrete consecutive gaze gestures for text entry. In Proceedings of the Symposium on Eye Tracking Research and Applications (ETRA'08). ACM, New York, NY, 11--18.

Cited By

View all
  • (2024)Evaluating the performance of gaze interaction for map target selectionCartography and Geographic Information Science10.1080/15230406.2024.233533152:1(82-102)Online publication date: 9-Apr-2024
  • (2024)Guiding gaze gestures on smartwatchesInternational Journal of Human-Computer Studies10.1016/j.ijhcs.2023.103196183:COnline publication date: 14-Mar-2024
  • (2023)Digital Alternative Communication for Individuals with Amyotrophic Lateral Sclerosis: What We HaveJournal of Clinical Medicine10.3390/jcm1216523512:16(5235)Online publication date: 11-Aug-2023
  • Show More Cited By

Recommendations

Comments

Information & Contributors

Information

Published In

cover image ACM Transactions on Interactive Intelligent Systems
ACM Transactions on Interactive Intelligent Systems  Volume 1, Issue 2
January 2012
157 pages
ISSN:2160-6455
EISSN:2160-6463
DOI:10.1145/2070719
Issue’s Table of Contents
Permission to make digital or hard copies of all or part of this work for personal or classroom use is granted without fee provided that copies are not made or distributed for profit or commercial advantage and that copies bear this notice and the full citation on the first page. Copyrights for components of this work owned by others than ACM must be honored. Abstracting with credit is permitted. To copy otherwise, or republish, to post on servers or to redistribute to lists, requires prior specific permission and/or a fee. Request permissions from [email protected]

Publisher

Association for Computing Machinery

New York, NY, United States

Publication History

Published: 13 January 2012
Accepted: 01 October 2011
Revised: 01 August 2011
Received: 01 December 2010
Published in TIIS Volume 1, Issue 2

Permissions

Request permissions for this article.

Check for updates

Author Tags

  1. Gaze gestures
  2. dynamic programming
  3. gaze tracking
  4. hierarchical temporal memory

Qualifiers

  • Research-article
  • Research
  • Refereed

Funding Sources

Contributors

Other Metrics

Bibliometrics & Citations

Bibliometrics

Article Metrics

  • Downloads (Last 12 months)15
  • Downloads (Last 6 weeks)1
Reflects downloads up to 08 Feb 2025

Other Metrics

Citations

Cited By

View all
  • (2024)Evaluating the performance of gaze interaction for map target selectionCartography and Geographic Information Science10.1080/15230406.2024.233533152:1(82-102)Online publication date: 9-Apr-2024
  • (2024)Guiding gaze gestures on smartwatchesInternational Journal of Human-Computer Studies10.1016/j.ijhcs.2023.103196183:COnline publication date: 14-Mar-2024
  • (2023)Digital Alternative Communication for Individuals with Amyotrophic Lateral Sclerosis: What We HaveJournal of Clinical Medicine10.3390/jcm1216523512:16(5235)Online publication date: 11-Aug-2023
  • (2023)An End-to-End Review of Gaze Estimation and its Interactive Applications on Handheld Mobile DevicesACM Computing Surveys10.1145/3606947Online publication date: 30-Jun-2023
  • (2023)Study on the brightness and graphical display object directions of the Single-Gaze-Gesture user interfaceDisplays10.1016/j.displa.2023.10253780(102537)Online publication date: Dec-2023
  • (2022)Toward Gaze-Based Map Interactions: Determining the Dwell Time and Buffer Size for the Gaze-Based Selection of Map FeaturesISPRS International Journal of Geo-Information10.3390/ijgi1102012711:2(127)Online publication date: 10-Feb-2022
  • (2021)Exploring Social Acceptability and Users’ Preferences of Head- and Eye-Based Interaction with Mobile DevicesProceedings of the 20th International Conference on Mobile and Ubiquitous Multimedia10.1145/3490632.3490636(12-23)Online publication date: 5-Dec-2021
  • (2019)Research on Multi-modal Interactive Control for Quadrotor UAV2019 IEEE 16th International Conference on Networking, Sensing and Control (ICNSC)10.1109/ICNSC.2019.8743337(329-334)Online publication date: May-2019
  • (2019)Design and Realization of Eye Control System for Small Ground Unmanned Platform2019 Chinese Automation Congress (CAC)10.1109/CAC48633.2019.8996238(4303-4308)Online publication date: Nov-2019
  • (2019)Priority Order of Single Gaze Gestures in Eye Control SystemHuman Systems Engineering and Design II10.1007/978-3-030-27928-8_45(299-305)Online publication date: 14-Aug-2019
  • Show More Cited By

View Options

Login options

Full Access

View options

PDF

View or Download as a PDF file.

PDF

eReader

View online with eReader.

eReader

Figures

Tables

Media

Share

Share

Share this Publication link

Share on social media