Papers
Topics
Authors
Recent
Gemini 2.5 Flash
Gemini 2.5 Flash 78 tok/s
Gemini 2.5 Pro 43 tok/s Pro
GPT-5 Medium 23 tok/s
GPT-5 High 29 tok/s Pro
GPT-4o 93 tok/s
GPT OSS 120B 470 tok/s Pro
Kimi K2 183 tok/s Pro
2000 character limit reached

Normalized Validity Scores for DNNs in Regression based Eye Feature Extraction (2403.11665v1)

Published 18 Mar 2024 in cs.CV

Abstract: We propose an improvement to the landmark validity loss. Landmark detection is widely used in head pose estimation, eyelid shape extraction, as well as pupil and iris segmentation. There are numerous additional applications where landmark detection is used to estimate the shape of complex objects. One part of this process is the accurate and fine-grained detection of the shape. The other part is the validity or inaccuracy per landmark, which can be used to detect unreliable areas, where the shape possibly does not fit, and to improve the accuracy of the entire shape extraction by excluding inaccurate landmarks. We propose a normalization in the loss formulation, which improves the accuracy of the entire approach due to the numerical balance of the normalized inaccuracy. In addition, we propose a margin for the inaccuracy to reduce the impact of gradients, which are produced by negligible errors close to the ground truth.

Definition Search Book Streamline Icon: https://streamlinehq.com
References (47)
  1. Eye-tracking technology in medical education: A systematic review. Medical teacher, 40(1):62–69, 2018.
  2. A. Bulat and G. Tzimiropoulos. How far are we from solving the 2d & 3d face alignment problem and a dataset of 230,000 3d facial landmarks. in ICCV, 1:4, 2017.
  3. Face alignment by explicit shape regression. Journal of Computer Vision, 2014.
  4. Eye tracking in virtual reality. Journal of eye movement research, 12(1), 2019.
  5. Active appearance models. PAMI, 6, 2001.
  6. Active shape models-their training and application. Computer Vision and Image Understanding, 61(1):38–59, 1995.
  7. Style aggregated network for facial landmark detection. in CVPR, 2:6, 2018.
  8. An image is worth 16x16 words: Transformers for image recognition at scale. arXiv preprint arXiv:2010.11929, 2020.
  9. Andrew T Duchowski. A breadth-first survey of eye-tracking applications. Behavior Research Methods, Instruments, & Computers, 34(4):455–470, 2002.
  10. Wing loss for robust facial landmark localisation with convolutional neural networks. in CVPR, 1:3, 2018.
  11. Random cascaded-regression copse for robust facial landmark detection. Signal Processing Letters, 22(1):76–80, 2015.
  12. Learning to validate the quality of detected landmarks. In Twelfth International Conference on Machine Vision (ICMV 2019), volume 11433, pages 97–104. SPIE, 2020.
  13. Teyed: Over 20 million real-world eye images with pupil, eyelid, and iris 2d and 3d segmentations, 2d and 3d landmarks, 3d eyeball, gaze vector, and eye movement types. In 2021 IEEE International Symposium on Mixed and Augmented Reality (ISMAR), pages 367–375. IEEE, 2021.
  14. Excuse: Robust pupil detection in real-world scenarios. In Computer Analysis of Images and Patterns: 16th International Conference, CAIP 2015, Valletta, Malta, September 2-4, 2015 Proceedings, Part I 16, pages 39–51. Springer, 2015.
  15. Pupilnet: Convolutional neural networks for robust pupil detection. arXiv preprint arXiv:1601.04902, 2016.
  16. Else: Ellipse selection for robust pupil detection in real-world environments. In Proceedings of the ninth biennial ACM symposium on eye tracking research & applications, pages 123–130, 2016.
  17. Pistol: Pupil invisible supportive tool to extract pupil, iris, eye opening, eye movements, pupil and iris gaze vector, and 2d as well as 3d gaze. In Proceedings of the 18th International Joint Conference on Computer Vision, Imaging and Computer Graphics Theory and Applications (VISIGRAPP 2023) - HUCAPP, pages 27–38. INSTICC, SciTePress, 2023.
  18. Using eye tracking to trace a cognitive process: Gaze behaviour during decision making in a natural environment. Journal of eye movement research, 6(1), 2013.
  19. Deep pyramidal residual networks. In Proceedings of the IEEE conference on computer vision and pattern recognition, pages 5927–5935, 2017.
  20. Application of eye tracking in medicine: A survey, research issues and challenges. Computerized Medical Imaging and Graphics, 65:176–190, 2018.
  21. Eye tracking: empirical foundations for a minimal reporting guideline. Behavior research methods, 55(1):364–416, 2023.
  22. Densely connected convolutional networks. in CVPR, 1:3, 2017.
  23. Adam: A method for stochastic optimization. arXiv preprint arXiv:1412.6980, 2014.
  24. Gaussian mixture 3d morphable face model. Pattern Recognition, 74:617–628, 2018.
  25. Ellseg: An ellipse segmentation framework for robust gaze tracking. IEEE Transactions on Visualization and Computer Graphics, 27(5):2757–2767, 2021.
  26. Reliable crowdsourcing and deep locality-preserving learning for expression recognition in the wild. in CVPR, pages 2584–2593, 2017.
  27. Sphereface: Deep hypersphere embedding for face recognition. in CVPR, 1:1, 2017.
  28. Mining design heuristics for additive manufacturing via eye-tracking methods and hidden markov modeling. Journal of Mechanical Design, 142(12):124502, 2020.
  29. Robust facial landmark detection via a fully-convolutional local-global context network. in CVPR, pages 781–790, 2018.
  30. Pupil detection and tracking using multiple light sources. Image and vision computing, 18(4):331–335, 2000.
  31. Supporting exploration of eye tracking data: Identifying changing behaviour over long durations. In Proceedings of the Sixth Workshop on Beyond Time and Errors on Novel Evaluation Methods for Visualization, pages 70–77, 2016.
  32. The use of eye tracking technology to explore learning and performance within virtual reality and mixed reality settings: a scoping review. Interactive Learning Environments, 30(7):1338–1350, 2022.
  33. U-net: Convolutional networks for biomedical image segmentation. In Medical Image Computing and Computer-Assisted Intervention–MICCAI 2015: 18th International Conference, Munich, Germany, October 5-9, 2015, Proceedings, Part III 18, pages 234–241. Springer, 2015.
  34. Pure: Robust pupil detection for real-time pervasive eye tracking. Computer Vision and Image Understanding, 170:40–50, 2018.
  35. Deep convolutional network cascade for facial point detection. in CVPR, pages 3476–3483, 2013.
  36. Inception-v4, inception-resnet and the impact of residual connections on learning. In Proceedings of the AAAI conference on artificial intelligence, volume 31, 2017.
  37. Efficientnet: Rethinking model scaling for convolutional neural networks. In International conference on machine learning, pages 6105–6114. PMLR, 2019.
  38. Mnemonic descent method: A recurrent process applied for end-to-end face alignment. in CVPR, pages 4177–4187, 2016.
  39. Deepeye: Deep convolutional network for pupil detection in real environments. Integrated Computer-Aided Engineering, 26(1):85–95, 2019.
  40. Copula ordinal regression for joint estimation of facial action unit intensity. in CVPR, pages 4902–4910, 2016.
  41. Aggregated residual transformations for deep neural networks. In Proceedings of the IEEE conference on computer vision and pattern recognition, pages 1492–1500, 2017.
  42. Stacked hourglass network for robust facial landmark localisation. in CVPR Workshops, pages 2025–2033, 2017.
  43. Neural aggregation network for video face recognition. in CVPR, 4:7, 2017.
  44. Deepvog: Open-source pupil segmentation and gaze estimation in neuroscience using deep learning. Journal of neuroscience methods, 324:108307, 2019.
  45. Coarse-to-fine auto-encoder networks (cfan) for real-time face alignment. in ECCV, 1–16, Springer, 2014.
  46. Opportunities for using eye tracking technology in manufacturing and logistics: Systematic literature review and research agenda. Computers & Industrial Engineering, page 108444, 2022.
  47. Unpaired image-to-image translation using cycle-consistent adversarial networks. in ICCV, pages 2223–2232, 2017.
List To Do Tasks Checklist Streamline Icon: https://streamlinehq.com

Collections

Sign up for free to add this paper to one or more collections.

Summary

We haven't generated a summary for this paper yet.

Ai Generate Text Spark Streamline Icon: https://streamlinehq.com

Paper Prompts

Sign up for free to create and run prompts on this paper using GPT-5.

Dice Question Streamline Icon: https://streamlinehq.com

Follow-up Questions

We haven't generated follow-up questions for this paper yet.

Authors (1)

Don't miss out on important new AI/ML research

See which papers are being discussed right now on X, Reddit, and more:

“Emergent Mind helps me see which AI papers have caught fire online.”

Philip

Philip

Creator, AI Explained on YouTube