Location via proxy:   [ UP ]  
[Report a bug]   [Manage cookies]                
skip to main content
10.1145/3588015.3588407acmconferencesArticle/Chapter ViewAbstractPublication PagesetraConference Proceedingsconference-collections
research-article

Multi-Rate Sensor Fusion for Unconstrained Near-Eye Gaze Estimation

Published: 30 May 2023 Publication History

Abstract

The power requirements of video-oculography systems can be prohibitive for high-speed operation on portable devices. Recently, low-power alternatives such as photosensors have been evaluated, providing gaze estimates at high frequency with a trade-off in accuracy and robustness. Potentially, an approach combining slow/high-fidelity and fast/low-fidelity sensors should be able to exploit their complementarity to track fast eye motion accurately and robustly. To foster research on this topic, we introduce OpenSFEDS, a near-eye gaze estimation dataset containing approximately 2M synthetic camera-photosensor image pairs sampled at 500 Hz under varied appearance and camera position. We also formulate the task of sensor fusion for gaze estimation, proposing a deep learning framework consisting in appearance-based encoding and temporal eye-state dynamics. We evaluate several single- and multi-rate fusion baselines on OpenSFEDS, achieving 8.7% error decrease when tracking fast eye movements with a multi-rate approach vs. a gaze forecasting approach operating with a low-speed sensor alone.

Supplemental Material

MP4 File
Presentation video - short version
ZIP File
Supplementary report and video.

References

[1]
Pankaj K Agarwal, Jiří Matoušek, and Subhash Suri. 1992. Farthest neighbors, maximum spanning trees and related problems in higher dimensions. Computational Geometry 1, 4 (1992), 189–201.
[2]
AN Angelopoulos, JNP Martel, AP Kohli, J Conradt, and G Wetzstein. 2021. Event-Based Near-Eye Gaze Tracking Beyond 10,000 Hz.IEEE Transactions on Visualization and Computer Graphics (2021).
[3]
Pradeep K Atrey, M Anwar Hossain, Abdulmotaleb El Saddik, and Mohan S Kankanhalli. 2010. Multimodal fusion for multimedia analysis: a survey. Multimedia systems 16, 6 (2010), 345–379.
[4]
Tadas Baltrušaitis, Chaitanya Ahuja, and Louis-Philippe Morency. 2018. Multimodal machine learning: A survey and taxonomy. IEEE transactions on pattern analysis and machine intelligence 41, 2 (2018), 423–443.
[5]
Aayush K Chaudhary, Nitinraj Nair, Reynold J Bailey, Jeff B Pelz, Sachin S Talathi, and Gabriel J Diaz. 2022. Temporal RIT-Eyes: From real infrared eye-images to synthetic sequences of gaze behavior. IEEE Transactions on Visualization and Computer Graphics 28, 11 (2022), 3948–3958.
[6]
Yihua Cheng, Haofei Wang, Yiwei Bao, and Feng Lu. 2021. Appearance-based gaze estimation with deep learning: A review and benchmark. arXiv preprint arXiv:2104.12668 (2021).
[7]
Jamil Fayyad, Mohammad A Jaradat, Dominique Gruyer, and Homayoun Najjaran. 2020. Deep learning sensor fusion for autonomous vehicle perception and localization: A review. Sensors 20, 15 (2020), 4220.
[8]
Wolfgang Fuhl and Enkelejda Kasneci. 2021. A multimodal eye movement dataset and a multimodal eye movement segmentation analysis. In ACM Symposium on Eye Tracking Research and Applications. 1–7.
[9]
Chao Gou, Yue Wu, Kang Wang, Kunfeng Wang, Fei-Yue Wang, and Qiang Ji. 2017. A joint cascaded framework for simultaneous eye detection and eye state estimation. Pattern Recognition 67 (2017), 23–31.
[10]
Henry Griffith, Dmytro Katrychuk, and Oleg Komogortsev. 2019. Assessment of Shift-Invariant CNN Gaze Mappings for PS-OG Eye Movement Sensors. In Proc. IEEE/CVF International Conference on Computer Vision Workshops. 0–0.
[11]
Henry Griffith, Dillon Lohr, Evgeny Abdulin, and Oleg Komogortsev. 2021. GazeBase, a large-scale, multi-stimulus, longitudinal eye movement dataset. Scientific Data 8, 1 (2021), 1–9.
[12]
Dan Witzner Hansen and Qiang Ji. 2009. In the eye of the beholder: A survey of models for eyes and gaze. IEEE transactions on pattern analysis and machine intelligence 32, 3 (2009), 478–500.
[13]
Kenneth Holmqvist, Marcus Nyström, Richard Andersson, Richard Dewhurst, Halszka Jarodzka, and Joost Van de Weijer. 2011. Eye tracking: A comprehensive guide to methods and measures. OUP Oxford.
[14]
Tapas Kanungo, David M Mount, Nathan S Netanyahu, Christine D Piatko, Ruth Silverman, and Angela Y Wu. 2002. An efficient k-means clustering algorithm: Analysis and implementation. IEEE transactions on pattern analysis and machine intelligence 24, 7 (2002), 881–892.
[15]
Joohwan Kim, Michael Stengel, Alexander Majercik, Shalini De Mello, David Dunn, Samuli Laine, Morgan McGuire, and David Luebke. 2019. Nvgaze: An anatomically-informed dataset for low-latency, near-eye gaze estimation. In Proc. 2019 CHI Conference on Human Factors in Computing Systems. 1–12.
[16]
Oleg V Komogortsev, Denise V Gobert, Sampath Jayarathna, Sandeep M Gowda, 2010. Standardization of automated analyses of oculomotor fixation and saccadic behaviors. IEEE Transactions on biomedical engineering 57, 11 (2010), 2635–2645.
[17]
Oleg V Komogortsev, Young Sam Ryu, and Do Hyong Koh. 2012. Fast Target Selection via Saccade-driven Methods. (2012).
[18]
Rakshit Kothari, Zhizhuo Yang, Christopher Kanan, Reynold Bailey, Jeff B Pelz, and Gabriel J Diaz. 2020. Gaze-in-wild: A dataset for studying eye and head coordination in everyday activities. Scientific reports 10, 1 (2020), 1–18.
[19]
Richard Li, Eric Whitmire, Michael Stengel, Ben Boudaoud, Jan Kautz, David Luebke, Shwetak Patel, and Kaan Akşit. 2020. Optical Gaze Tracking with Spatially-Sparse Single-Pixel Detectors. In 2020 IEEE International Symposium on Mixed and Augmented Reality. IEEE, 117–126.
[20]
Tianxing Li, Qiang Liu, and Xia Zhou. 2017. Ultra-low power gaze tracking for virtual reality. In Proc. 15th ACM Conference on Embedded Network Sensor Systems. 1–14.
[21]
Conny Lu, Praneeth Chakravarthula, Yujie Tao, Steven Chen, and Henry Fuchs. 2020. Improved vergence and accommodation via purkinje image tracking with multiple cameras for ar glasses. In 2020 IEEE International Symposium on Mixed and Augmented Reality. IEEE, 320–331.
[22]
Diederick C Niehorster, Thiago Santini, Roy S Hessels, Ignace TC Hooge, Enkelejda Kasneci, and Marcus Nyström. 2020. The impact of slippage on the data quality of head-worn eye trackers. Behavior Research Methods 52, 3 (2020), 1140–1160.
[23]
Cristina Palmero, Oleg V Komogortsev, and Sachin S Talathi. 2020. Benefits of temporal information for appearance-based gaze estimation. In ACM Symposium on Eye Tracking Research and Applications. 1–5.
[24]
Cristina Palmero, Abhishek Sharma, Karsten Behrendt, Kapil Krishnakumar, Oleg V Komogortsev, and Sachin S Talathi. 2021. OpenEDS2020 Challenge on Gaze Tracking for VR: Dataset and Results. Sensors 21, 14 (2021), 4769.
[25]
Seonwook Park, Xucong Zhang, Andreas Bulling, and Otmar Hilliges. 2018. Learning to find eye region landmarks for remote gaze estimation in unconstrained settings. In ACM Symposium on Eye Tracking Research and Applications. 1–10.
[26]
Anjul Patney, Marco Salvi, Joohwan Kim, Anton Kaplanyan, Chris Wyman, Nir Benty, David Luebke, and Aaron Lefohn. 2016. Towards foveated rendering for gaze-tracked virtual reality. ACM Transactions on Graphics 35, 6 (2016), 1–12.
[27]
Sonia Porta, Benoit Bossavit, Rafael Cabeza, Andoni Larumbe-Bergera, Gonzalo Garde, and Arantxa Villanueva. 2019. U2Eyes: a binocular dataset for eye tracking and gaze estimation. In Proc. IEEE/CVF International Conference on Computer Vision Workshops.
[28]
Wasifur Rahman, Md Kamrul Hasan, Sangwu Lee, Amir Zadeh, Chengfeng Mao, Louis-Philippe Morency, and Ehsan Hoque. 2020. Integrating multimodal information in large pretrained transformers. In Proc. Association for Computational Linguistics. Meeting, Vol. 2020. NIH Public Access, 2359.
[29]
Ioannis Rigas, Hayes Raffle, and Oleg V. Komogortsev. 2017. Hybrid ps-v technique: A novel sensor fusion approach for fast mobile eye-tracking with sensor-shift aware correction. IEEE Sensors Journal 17, 24 (2017), 8356–8366.
[30]
Ioannis Rigas, Hayes Raffle, and Oleg V Komogortsev. 2018. Photosensor oculography: survey and parametric analysis of designs using model-based simulation. IEEE Transactions on Human-Machine Systems 48, 6 (2018), 670–681.
[31]
Marc Tonsen, Julian Steil, Yusuke Sugano, and Andreas Bulling. 2017. Invisibleeye: Mobile eye tracking using multiple low-resolution cameras and learning-based gaze estimation. Proc. ACM on Interactive, Mobile, Wearable and Ubiquitous Technologies 1, 3 (2017), 1–21.
[32]
Erroll Wood, Tadas Baltrušaitis, Louis-Philippe Morency, Peter Robinson, and Andreas Bulling. 2016a. A 3D Morphable Eye Region Model for Gaze Estimation. In Proc. European Conference on Computer Vision. 297–313.
[33]
Erroll Wood, Tadas Baltrušaitis, Louis-Philippe Morency, Peter Robinson, and Andreas Bulling. 2016b. Learning an appearance-based gaze estimator from one million synthesised images. In ACM Symp. Eye Tracking Research and Applications. 131–138.
[34]
Erroll Wood, Tadas Baltrušaitis, Xucong Zhang, Yusuke Sugano, Peter Robinson, and Andreas Bulling. 2015. Rendering of Eyes for Eye-Shape Registration and Gaze Estimation. In Proc. IEEE International Conference on Computer Vision. 3756–3764.
[35]
Yuk-Hoi Yiu, Moustafa Aboulatta, Theresa Raiser, Leoni Ophey, Virginia L Flanagin, Peter Zu Eulenburg, and Seyed-Ahmad Ahmadi. 2019. DeepVOG: Open-source pupil segmentation and gaze estimation in neuroscience using deep learning. Journal of neuroscience methods 324 (2019), 108307.
[36]
Raimondas Zemblys and Oleg Komogortsev. 2018. Making stand-alone PS-OG technology tolerant to the equipment shifts. In Proc. 7th Workshop on Pervasive Eye Tracking and Mobile Eye-Based Interaction. 1–9.

Cited By

View all
  • (2024)Towards Invisible Eye Tracking with Lens-Coupled Lateral PhotodetectorsProceedings of the 2024 Symposium on Eye Tracking Research and Applications10.1145/3649902.3656354(1-7)Online publication date: 4-Jun-2024

Recommendations

Comments

Information & Contributors

Information

Published In

cover image ACM Conferences
ETRA '23: Proceedings of the 2023 Symposium on Eye Tracking Research and Applications
May 2023
441 pages
ISBN:9798400701504
DOI:10.1145/3588015
Permission to make digital or hard copies of all or part of this work for personal or classroom use is granted without fee provided that copies are not made or distributed for profit or commercial advantage and that copies bear this notice and the full citation on the first page. Copyrights for components of this work owned by others than the author(s) must be honored. Abstracting with credit is permitted. To copy otherwise, or republish, to post on servers or to redistribute to lists, requires prior specific permission and/or a fee. Request permissions from [email protected].

Sponsors

Publisher

Association for Computing Machinery

New York, NY, United States

Publication History

Published: 30 May 2023

Permissions

Request permissions for this article.

Check for updates

Author Tags

  1. Deep learning
  2. Eye tracking
  3. Gaze estimation
  4. Photosensors.
  5. Sensor fusion
  6. Virtual reality

Qualifiers

  • Research-article
  • Research
  • Refereed limited

Data Availability

Funding Sources

  • MINECO/FEDER, UE
  • ICREA Academia Program

Conference

ETRA '23

Acceptance Rates

Overall Acceptance Rate 69 of 137 submissions, 50%

Upcoming Conference

Contributors

Other Metrics

Bibliometrics & Citations

Bibliometrics

Article Metrics

  • Downloads (Last 12 months)125
  • Downloads (Last 6 weeks)8
Reflects downloads up to 11 Jan 2025

Other Metrics

Citations

Cited By

View all
  • (2024)Towards Invisible Eye Tracking with Lens-Coupled Lateral PhotodetectorsProceedings of the 2024 Symposium on Eye Tracking Research and Applications10.1145/3649902.3656354(1-7)Online publication date: 4-Jun-2024

View Options

Login options

View options

PDF

View or Download as a PDF file.

PDF

eReader

View online with eReader.

eReader

HTML Format

View this article in HTML Format.

HTML Format

Media

Figures

Other

Tables

Share

Share

Share this Publication link

Share on social media