Papers
Topics
Authors
Recent
Gemini 2.5 Flash
Gemini 2.5 Flash
158 tokens/sec
GPT-4o
7 tokens/sec
Gemini 2.5 Pro Pro
45 tokens/sec
o3 Pro
4 tokens/sec
GPT-4.1 Pro
38 tokens/sec
DeepSeek R1 via Azure Pro
28 tokens/sec
2000 character limit reached

GazeTrak: Exploring Acoustic-based Eye Tracking on a Glass Frame (2402.14634v2)

Published 22 Feb 2024 in cs.HC

Abstract: In this paper, we present GazeTrak, the first acoustic-based eye tracking system on glasses. Our system only needs one speaker and four microphones attached to each side of the glasses. These acoustic sensors capture the formations of the eyeballs and the surrounding areas by emitting encoded inaudible sound towards eyeballs and receiving the reflected signals. These reflected signals are further processed to calculate the echo profiles, which are fed to a customized deep learning pipeline to continuously infer the gaze position. In a user study with 20 participants, GazeTrak achieves an accuracy of 3.6{\deg} within the same remounting session and 4.9{\deg} across different sessions with a refreshing rate of 83.3 Hz and a power signature of 287.9 mW. Furthermore, we report the performance of our gaze tracking system fully implemented on an MCU with a low-power CNN accelerator (MAX78002). In this configuration, the system runs at up to 83.3 Hz and has a total power signature of 95.4 mW with a 30 Hz FPS.

Definition Search Book Streamline Icon: https://streamlinehq.com
References (67)
  1. Tobii AB. 2022a. Tobii Pro Fusion. Retrieved Sept. 13, 2022 from https://www.tobiipro.com/product-listing/fusion/
  2. Tobii AB. 2022b. Tobii Pro Glasses 3. Retrieved Sept. 13, 2022 from https://www.tobiipro.com/product-listing/tobii-pro-glasses-3/
  3. Real-time Pupil Tracking from Monocular Video for Digital Puppetry. CoRR abs/2006.11341 (2020). arXiv:2006.11341 https://arxiv.org/abs/2006.11341
  4. Christer Ahlstrom and Tania Dukic. 2010. Comparison of Eye Tracking Systems with One and Three Cameras. In Proceedings of the International Conference on Methods and Techniques in Behavioral Research (MB). Article 3, 4 pages. https://doi.org/10.1145/1931344.1931347
  5. Analog Devices AI. 2022. MaximIntegratedAI. Retrieved Aug 23, 2023 from https://github.com/MaximIntegratedAI
  6. Analog Devices AI. 2023. MAX7800x Power Monitor and Energy Benchmarking Guide. Retrieved Aug 23, 2023 from https://github.com/MaximIntegratedAI/MaximAI_Documentation/blob/master/Guides/MAX7800x%20Power%20Monitor%20and%20Energy%20Benchmarking%20Guide.md
  7. EyeTell: Tablet-based Calibration-free Eye-typing using Smooth-pursuit movements. In ACM Symposium on Eye Tracking Research and Applications. 1–6.
  8. Frank H Borsato and Carlos H Morimoto. 2016. Episcleral surface tracking: challenges and possibilities for using mice sensors for wearable eye tracking. In Proceedings of the Ninth Biennial ACM Symposium on Eye Tracking Research & Applications. 39–46.
  9. It’s in your eyes: Towards context-awareness and mobile HCI using wearable EOG goggles. In Proceedings of the 10th international conference on Ubiquitous computing. 84–93.
  10. Wearable EOG Goggles: Eye-Based Interaction in Everyday Environments. In CHI Extended Abstracts on Human Factors in Computing Systems. 3259–3264. https://doi.org/10.1145/1520340.1520468
  11. Communication by Gaze Interaction Association. 2010. Woking copy of definitions and terminology for Eye Tracker accuracy and precision. Retrieved Sept. 13, 2022 from http://old.cogain.org/forums/eye-tracker-accuracy-and-precision-general-discussion/eye-tracker-accuracy-terms-and-definiti.html
  12. Gaze Detection by Wearable Eye-Tracking and NIR LED-Based Head-Tracking Device Based on SVR. Etri Journal 34, 4 (2012), 542–552.
  13. Analog Devices. 2022. MAX78002 Evaluation Kit. Retrieved Aug 23, 2023 from https://www.analog.com/media/en/technical-documentation/data-sheets/MAX78002EVKIT.pdf
  14. DigiKey. 2023. OWR-05049T-38D. Retrieved Mar 13, 2023 from https://www.digikey.com/en/products/detail/ole-wolff-electronics-inc/OWR-05049T-38D/13683703
  15. GazBy: Gaze-Based BERT Model to Incorporate Human Attention in Neural Information Retrieval. 182–192. https://doi.org/10.1145/3539813.3545129
  16. Ergoneers. 2022. Dikablis Glasses 3. Retrieved Sept. 13, 2022 from https://www.ergoneers.com/en/mobile-eye-tracker-dikablis-glasses-3/?gclid=CjwKCAiA9tyQBhAIEiwA6tdCrFd7F7xBwNa4XVP09wRHlBATh_jafRuH5ErUVdhJt5WVLK2_FdVs7RoCpK4QAvD_BwE
  17. Centers for Disease Control and Prevention (CDC). 2023. NIOSH Sound Level Meter App. Retrieved Mar 13, 2023 from https://www.cdc.gov/niosh/topics/noise/app.html
  18. SonicFace: Tracking Facial Expressions Using a Commodity Microphone Array. Proc. ACM Interact. Mob. Wearable Ubiquitous Technol. 5, 4, Article 156 (dec 2022), 33 pages. https://doi.org/10.1145/3494988
  19. GazeRecorder. 2022. GazeRecorder Webcam Eye Tracking. Retrieved Sept. 13, 2022 from https://gazerecorder.com/webcam-eye-tracking-accuracy/
  20. Andre Golard and Sachin S Talathi. 2021. Ultrasound for Gaze Estimation—A Modeling and Empirical Study. Sensors 21, 13 (2021), 4502.
  21. Brown HCI Group. 2021. WebGazer.js: Democratizing Webcam Eye Tracking on the Browser. Retrieved Sept. 13, 2022 from https://webgazer.cs.brown.edu/#publication
  22. Craig Hennessey and Jacob Fiset. 2012. Long range eye tracking: bringing eye tracking into the living room. In Proceedings of the Symposium on Eye Tracking Research and Applications. 249–252.
  23. iMotions. 2022. SMI Eye Tracking Glasses. Retrieved Sept. 13, 2022 from https://imotions.com/hardware/smi-eye-tracking-glasses/
  24. Pupil: an open source platform for pervasive eye tracking and mobile gaze-based interaction. In Proceedings of the 2014 ACM international joint conference on pervasive and ubiquitous computing: Adjunct publication. 1151–1160.
  25. Eye Tracking for Everyone. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition (CVPR).
  26. Pinpointing: Precise Head- and Eye-Based Target Selection for Augmented Reality. In Proceedings of the CHI Conference on Human Factors in Computing Systems. 1–14. https://doi.org/10.1145/3173574.3173655
  27. Low Power Lab. 2018. Introduction to Current Ranger. Retrieved Mar 13, 2023 from https://lowpowerlab.com/guide/currentranger/
  28. Pupil Labs. 2022. Pupil Invisible. Retrieved Sept. 13, 2022 from https://pupil-labs.com/products/
  29. Robust head mounted wearable eye tracking system for dynamical calibration. Journal of Eye Movement Research 8, 5 (2015).
  30. Yaxiong Lei. 2021. Eye Tracking Calibration on Mobile Devices. In ACM Symposium on Eye Tracking Research and Applications (ETRA Adjunct). Article 4, 4 pages. https://doi.org/10.1145/3450341.3457989
  31. Identification and classification of construction equipment operators’ mental fatigue using wearable eye-tracking technology. Automation in Construction 109 (2020), 103000.
  32. EyeEcho: Continuous and Low-power Facial Expression Tracking on Glasses. arXiv:2402.12388 [cs.HC]
  33. EarIO: A Low-power Acoustic Sensing Earable for Continuously Tracking Detailed Facial Movements. Proceedings of the ACM on Interactive, Mobile, Wearable and Ubiquitous Technologies 6, 2 (2022), 1–24.
  34. Ultra-Low Power Gaze Tracking for Virtual Reality. In Proceedings of the ACM Conference on Embedded Network Sensor Systems (SenSys). Article 25, 14 pages. https://doi.org/10.1145/3131672.3131682
  35. Tianxing Li and Xia Zhou. 2018. Battery-Free Eye Tracker on Glasses. In Proceedings of the Annual International Conference on Mobile Computing and Networking (MobiCom). 67–82. https://doi.org/10.1145/3241539.3241578
  36. Bhanuka Mahanama. 2022. Multi-User Eye-Tracking. In Symposium on Eye Tracking Research and Applications (ETRA). Article 36, 3 pages. https://doi.org/10.1145/3517031.3532197
  37. PoseSonic: 3D Upper Body Pose Estimation Through Egocentric Acoustic Sensing on Smartglasses. Proc. ACM Interact. Mob. Wearable Ubiquitous Technol. 7, 3, Article 111 (sep 2023), 28 pages. https://doi.org/10.1145/3610895
  38. Extraction of read text using a wearable eye tracker for automatic video annotation. In Adjunct Proceedings of the 2015 ACM International Joint Conference on Pervasive and Ubiquitous Computing and Proceedings of the 2015 ACM International Symposium on Wearable Computers. 849–854.
  39. IShadow: Design of a Wearable, Real-Time Mobile Gaze Tracker. In Proceedings of the 12th Annual International Conference on Mobile Systems, Applications, and Services (MobiSys ’14). 82–94. https://doi.org/10.1145/2594368.2594388
  40. CIDER: Enabling Robustness-Power Tradeoffs on a Computational Eyeglass. In Proceedings of the 21st Annual International Conference on Mobile Computing and Networking (MobiCom ’15). 400–412. https://doi.org/10.1145/2789168.2790096
  41. Contactless sleep apnea detection on smartphones. In Proceedings of the 13th annual international conference on mobile systems, applications, and services. 45–57.
  42. A wearable gaze tracking system for children in unconstrained environments. Computer Vision and Image Understanding 115, 4 (2011), 476–486.
  43. U.S. Department of Health and Human Services. 1998. Criteria for a recommended standard: occupational noise exposure. DHHS (NIOSH) Publication No. 98–126 (1998). https://www.cdc.gov/niosh/docs/98-126/
  44. Just blink your eyes: A head-free gaze tracking system. In CHI’03 extended abstracts on Human factors in computing systems. 950–957.
  45. Smartphone Eye Tracking Toolbox: Accurate Gaze Recovery on Mobile Displays. In Proceedings of the Symposium on Eye Tracking Research and Applications (ETRA). 367–68. https://doi.org/10.1145/2578153.2628813
  46. Alexandra Papoutsaki. 2015. Scalable Webcam Eye Tracking by Learning from User Interactions. In Proceedings of the Annual ACM Conference Extended Abstracts on Human Factors in Computing Systems (CHI EA). 219–222. https://doi.org/10.1145/2702613.2702627
  47. Searchgazer: Webcam eye tracking for remote studies of web search. In Proceedings of the 2017 conference on conference human information interaction and retrieval. 17–26.
  48. WearCam: A head mounted wireless camera for monitoring gaze attention and for the diagnosis of developmental disorders in young children. In IEEE International Symposium on Robot and Human Interactive Communication (RO-MAN). 594–598. https://doi.org/10.1109/ROMAN.2007.4415154
  49. PJRC. 2023a. Audio Adaptor Boards for Teensy 3.x and Teensy 4.x. Retrieved Mar 13, 2023 from https://www.pjrc.com/store/teensy3_audio.html
  50. PJRC. 2023b. Teensy 4.1 Development Board. Retrieved Mar 13, 2023 from https://www.pjrc.com/store/teensy41.html
  51. Limbus/pupil switching for wearable eye tracking under variable lighting conditions. In Proceedings of the 2008 symposium on Eye tracking research & applications. 61–64.
  52. Towards Efficient Calibration for Webcam Eye-Tracking in Online Experiments. In Symposium on Eye Tracking Research and Applications (ETRA). Article 27, 7 pages. https://doi.org/10.1145/3517031.3529645
  53. RealEye sp. z o.o. 2022. RealEye Webcam Eye-Tracking. Retrieved Sept. 13, 2022 from https://www.realeye.io/
  54. EchoNose: Sensing Mouth, Breathing and Tongue Gestures inside Oral Cavity using a Non-contact Nose Interface. In Proceedings of the 2023 ACM International Symposium on Wearable Computers (Cancun, Quintana Roo, Mexico) (ISWC ’23). Association for Computing Machinery, New York, NY, USA, 22–26. https://doi.org/10.1145/3594738.3611358
  55. TDK. 2023. ICS-43434. Retrieved Mar 13, 2023 from https://invensense.tdk.com/products/ics-43434/
  56. A wearable head-mounted sensor-based apparatus for eye tracking applications. In 2008 IEEE Conference on Virtual Environments, Human-Computer Interfaces and Measurement Systems. IEEE, 136–139.
  57. A head-mounted sensor-based eye tracking device: eye touch system. In Proceedings of the 2008 symposium on Eye tracking research & applications. 87–90.
  58. Wearable Eye Tracking for Mental Health Monitoring. Computer Communications 35, 11 (jun 2012), 1306–1311. https://doi.org/10.1016/j.comcom.2011.11.002
  59. Thermographic Eye Tracking. In Proceedings of the Biennial ACM Symposium on Eye Tracking Research & Applications. 307–310. https://doi.org/10.1145/2857491.2857543
  60. C-FMCW based contactless respiration detection using acoustic signal. Proceedings of the ACM on Interactive, Mobile, Wearable and Ubiquitous Technologies 1, 4 (2018), 1–20.
  61. EyeContact: scleral coil eye tracking for virtual reality. In Proceedings of the 2016 ACM International Symposium on Wearable Computers. 184–191.
  62. Comparison of Webcam and Remote Eye Tracking. In Symposium on Eye Tracking Research and Applications (ETRA). Article 32, 7 pages. https://doi.org/10.1145/3517031.3529615
  63. Detecting eye contact using wearable eye-tracking glasses. In Proceedings of the 2012 ACM conference on ubiquitous computing. 699–704.
  64. HPSpeech: Silent Speech Interface for Commodity Headphones. In Proceedings of the 2023 ACM International Symposium on Wearable Computers (Cancun, Quintana Roo, Mexico) (ISWC ’23). Association for Computing Machinery, New York, NY, USA, 60–65. https://doi.org/10.1145/3594738.3611365
  65. EchoSpeech: Continuous Silent Speech Recognition on Minimally-obtrusive Eyewear Powered by Acoustic Sensing. In Proceedings of the 2023 CHI Conference on Human Factors in Computing Systems (Hamburg, Germany) (CHI ’23). Association for Computing Machinery, New York, NY, USA, Article 852, 18 pages. https://doi.org/10.1145/3544548.3580801
  66. Discrimination of gaze directions using low-level eye image features. In Proceedings of the 1st international workshop on pervasive eye tracking & mobile eye-based interaction. 9–14.
  67. MobiET: A New Approach to Eye Tracking for Mobile Device. In Proceedings of the ACM International Joint Conference and International Symposium on Pervasive and Ubiquitous Computing and Wearable Computers (UbiComp). 862–869. https://doi.org/10.1145/3267305.3274174
Citations (8)

Summary

  • The paper introduces an innovative acoustic sensor approach that uses FMCW signals and deep learning to estimate gaze direction.
  • It empirically evaluated the system with 20 participants, achieving 4.9° cross-session and 3.6° in-session tracking accuracy at 83.3 Hz.
  • The design’s low power consumption, reducing usage to as little as 95.4 mW, highlights its potential for wearable real-time applications.

An Analysis of GazeTrak: Acoustic-based Eye Tracking for Smart Glasses

The paper "GazeTrak: Exploring Acoustic-based Eye Tracking on a Glass Frame" introduces a pioneering system for eye tracking using acoustic methods in a glasses form factor. Utilizing a custom setup of speakers and microphones, GazeTrak employs frequency-modulated continuous-wave (FMCW) acoustic signals to track eye movements. This setup stands out due to its low power consumption compared to traditional camera-based systems, making it particularly suitable for integration with devices like smart glasses, which often have limited battery capacity.

Key Approaches and Findings

  1. Innovative Use of Acoustic Sensors: GazeTrak leverages the distortion of acoustic signals as they interact with the eyes and surrounding features to infer gaze direction. The system's hardware components include one speaker and four microphones per glass arm, utilizing encoded sound waves to generate an "echo profile." This profile is processed with a deep learning model to interpret gaze positions.
  2. Deep Learning Integration: A specialized deep learning architecture based on the ResNet-18 framework was employed to handle the complex task of translating echo profiles into gaze estimations. The integration of machine learning allows for adapting to individual user characteristics and improving predictive accuracy over time.
  3. Empirical Evaluation: The empirical section involved a paper with 20 participants, demonstrating that GazeTrak can achieve a cross-session tracking accuracy of 4.9 degrees and an even higher accuracy of 3.6 degrees within the same session. Such precision is achieved with a refresh rate of 83.3 Hz and modest power consumption, affirming the feasibility of real-time eye tracking on glasses.
  4. Low Power Consumption: A significant achievement of GazeTrak is its low power signature — 287.9 mW on a Teensy 4.1 setup and down to 95.4 mW when fully implemented on a MAX78002 microcontroller with a custom low-power CNN accelerator. This represents a power reduction of over 95% compared to some existing camera-based solutions.
  5. Robustness to Noise: The paper also confirmed GazeTrak's resilience under various environmental noise conditions, demonstrating its potential flexibility for diverse real-world applications without significant degradation of performance.

Practical Implications

The development of GazeTrak can be considered a substantial advancement in the domain of wearable eye-tracking technology. Its ability to operate effectively at low power levels increases the potential for prolonged active use without frequent recharging, something researchers and developers have long pursued in wearable devices. This makes it particularly suited to applications in mobile health, augmented reality, and assistive technologies.

Theoretical Implications and Future Work

Theoretically, GazeTrak encourages a shift in the paradigm of eye-tracking methodologies from camera-centric systems to acoustics-based ones. This transition leverages the innate benefits of acoustic methods, like low power requirements and unobtrusive setup, while presenting challenges such as maintaining precision across users and dynamic environmental conditions. Future research should focus on refining the model's accuracy further and reducing the amount of calibration data needed. Additionally, expanding the dataset for training the model and exploring diverse feature extraction methods could bolster generalizability to broader demographics without exhaustive personalization.

Conclusion

"GazeTrak: Exploring Acoustic-based Eye Tracking on a Glass Frame" demonstrates a promising alternative to traditional eye-tracking systems, pivoting on acoustic sensing's inherent advantages. Its utility in continuous, low-power, real-time tracking opens up various applications and sets a new benchmark for wearable technologies. As further innovations are made, GazeTrak and similar approaches could reshape the landscape of how gaze tracking is integrated into everyday technology.

Youtube Logo Streamline Icon: https://streamlinehq.com