Papers
Topics
Authors
Recent
Gemini 2.5 Flash
Gemini 2.5 Flash
162 tokens/sec
GPT-4o
7 tokens/sec
Gemini 2.5 Pro Pro
45 tokens/sec
o3 Pro
4 tokens/sec
GPT-4.1 Pro
38 tokens/sec
DeepSeek R1 via Azure Pro
28 tokens/sec
2000 character limit reached

UVAGaze: Unsupervised 1-to-2 Views Adaptation for Gaze Estimation (2312.15644v1)

Published 25 Dec 2023 in cs.CV

Abstract: Gaze estimation has become a subject of growing interest in recent research. Most of the current methods rely on single-view facial images as input. Yet, it is hard for these approaches to handle large head angles, leading to potential inaccuracies in the estimation. To address this issue, adding a second-view camera can help better capture eye appearance. However, existing multi-view methods have two limitations. 1) They require multi-view annotations for training, which are expensive. 2) More importantly, during testing, the exact positions of the multiple cameras must be known and match those used in training, which limits the application scenario. To address these challenges, we propose a novel 1-view-to-2-views (1-to-2 views) adaptation solution in this paper, the Unsupervised 1-to-2 Views Adaptation framework for Gaze estimation (UVAGaze). Our method adapts a traditional single-view gaze estimator for flexibly placed dual cameras. Here, the "flexibly" means we place the dual cameras in arbitrary places regardless of the training data, without knowing their extrinsic parameters. Specifically, the UVAGaze builds a dual-view mutual supervision adaptation strategy, which takes advantage of the intrinsic consistency of gaze directions between both views. In this way, our method can not only benefit from common single-view pre-training, but also achieve more advanced dual-view gaze estimation. The experimental results show that a single-view estimator, when adapted for dual views, can achieve much higher accuracy, especially in cross-dataset settings, with a substantial improvement of 47.0%. Project page: https://github.com/MickeyLLG/UVAGaze.

Definition Search Book Streamline Icon: https://streamlinehq.com
References (33)
  1. Social eye gaze in human-robot interaction: a review. Journal of Human-Robot Interaction, 6(1): 25–63.
  2. Robust real-time multi-view eye tracking. arXiv preprint arXiv:1711.05444.
  3. Generalizing gaze estimation with rotation consistency. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, 4207–4216.
  4. Utilizing VR and gaze tracking to develop AR solutions for industrial maintenance. In CHI.
  5. Deep semantic gaze embedding and scanpath comparison for expertise classification during OPT viewing. In ACM symposium on eye tracking research and applications, 1–10.
  6. Puregaze: Purifying gaze feature for generalizable gaze estimation. In Proceedings of the AAAI Conference on Artificial Intelligence, volume 36, 436–443.
  7. A coarse-to-fine adaptive network for appearance-based gaze estimation. In Proceedings of the AAAI Conference on Artificial Intelligence, volume 34, 10623–10630.
  8. DVGaze: Dual-View Gaze Estimation. In Proceedings of the IEEE/CVF International Conference on Computer Vision, 20632–20641.
  9. Appearance-based gaze estimation via evaluation-guided asymmetric regression. In Proceedings of the European Conference on Computer Vision (ECCV), 100–115.
  10. Demiris, Y. 2007. Prediction of intent in robotics and multi-agent systems. Cognitive processing, 8(3): 151–158.
  11. Unsupervised multi-view gaze representation learning. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, 5001–5009.
  12. Domain adaptation gaze estimation by embedding with prediction consistency. In Proceedings of the Asian Conference on Computer Vision.
  13. Deep residual learning for image recognition. In Proceedings of the IEEE conference on computer vision and pattern recognition, 770–778.
  14. Gaze360: Physically unconstrained gaze estimation in the wild. In Proceedings of the IEEE/CVF International Conference on Computer Vision, 6912–6921.
  15. A Preliminary Study on Performance Evaluation of Multi-View Multi-Modal Gaze Estimation under Challenging Conditions. In Extended Abstracts of the 2020 CHI Conference on Human Factors in Computing Systems, 1–7.
  16. Gaze-contingent ocular parallax rendering for virtual reality. ACM Transactions on Graphics (TOG).
  17. Eye tracking for everyone. In Proceedings of the IEEE conference on computer vision and pattern recognition, 2176–2184.
  18. Multiview multitask gaze estimation with deep convolutional neural networks. IEEE transactions on neural networks and learning systems, 30(10): 3010–3023.
  19. Generalizing Gaze Estimation with Outlier-guided Collaborative Adaptation. In Proceedings of the IEEE/CVF International Conference on Computer Vision, 3835–3844.
  20. Eye tracking and eye-based human–computer interaction. In Advances in physiological computing, 39–65. Springer.
  21. Predicting primary gaze behavior using social saliency fields. In Proceedings of the IEEE International Conference on Computer Vision, 3503–3510.
  22. Few-shot adaptive gaze estimation. In Proceedings of the IEEE/CVF International Conference on Computer Vision, 9368–9377.
  23. Deep pictorial gaze estimation. In Proceedings of the European Conference on Computer Vision (ECCV), 721–738.
  24. Designing social cues for collaborative robots: the role of gaze and breathing in human-robot collaboration. In Proceedings of the 2020 ACM/IEEE International Conference on Human-Robot Interaction, 343–357.
  25. Hybrid gaze/EEG brain computer interface for robot arm control on a pick and place task. In 2015 37th Annual International Conference of the IEEE Engineering in Medicine and Biology Society (EMBC), 1476–1479. IEEE.
  26. Generalizing eye tracking with bayesian adversarial learning. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, 11907–11916.
  27. Wheelchair-Centered Omnidirectional Gaze-Point Estimation in the Wild. IEEE Transactions on Human-Machine Systems.
  28. Comparing Single-modal and Multimodal Interaction in an Augmented Reality System. In ISMAR. IEEE.
  29. Improving few-shot user-specific gaze adaptation via gaze redirection synthesis. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, 11937–11946.
  30. Gazeonce: Real-time multi-person gaze estimation. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, 4197–4206.
  31. ETH-XGaze: A large scale dataset for gaze estimation under extreme head pose and gaze variation. In European Conference on Computer Vision, 365–381. Springer.
  32. It’s written all over your face: Full-face appearance-based gaze estimation. In Computer Vision and Pattern Recognition Workshops (CVPRW), 2017 IEEE Conference on, 2299–2308. IEEE.
  33. Mpiigaze: Real-world dataset and deep appearance-based gaze estimation. IEEE transactions on pattern analysis and machine intelligence, 41(1): 162–175.
Citations (3)

Summary

We haven't generated a summary for this paper yet.

Github Logo Streamline Icon: https://streamlinehq.com

GitHub