Papers
Topics
Authors
Recent
Gemini 2.5 Flash
Gemini 2.5 Flash
194 tokens/sec
GPT-4o
7 tokens/sec
Gemini 2.5 Pro Pro
45 tokens/sec
o3 Pro
4 tokens/sec
GPT-4.1 Pro
38 tokens/sec
DeepSeek R1 via Azure Pro
28 tokens/sec
2000 character limit reached

Boosting Facial Action Unit Detection Through Jointly Learning Facial Landmark Detection and Domain Separation and Reconstruction (2310.05207v5)

Published 8 Oct 2023 in cs.CV, cs.AI, cs.LG, and cs.MM

Abstract: Recently how to introduce large amounts of unlabeled facial images in the wild into supervised Facial Action Unit (AU) detection frameworks has become a challenging problem. In this paper, we propose a new AU detection framework where multi-task learning is introduced to jointly learn AU domain separation and reconstruction and facial landmark detection by sharing the parameters of homostructural facial extraction modules. In addition, we propose a new feature alignment scheme based on contrastive learning by simple projectors and an improved contrastive loss, which adds four additional intermediate supervisors to promote the feature reconstruction process. Experimental results on two benchmarks demonstrate our superiority against the state-of-the-art methods for AU detection in the wild.

Definition Search Book Streamline Icon: https://streamlinehq.com
References (20)
  1. What the face reveals: Basic and applied studies of spontaneous expression using the Facial Action Coding System (FACS), Oxford University Press, 2020.
  2. “Self-supervised learning of a facial attribute embedding from video,” arXiv preprint arXiv:1808.06882, 2018.
  3. “Learning representations for facial actions from unlabeled videos,” IEEE Transactions on Pattern Analysis and Machine Intelligence, vol. 44, no. 1, pp. 302–317, 2022.
  4. “Contrastive learning of person-independent representations for facial action unit detection,” IEEE Transactions on Image Processing, 2023.
  5. “Domain-adversarial training of neural networks,” The journal of machine learning research, vol. 17, no. 1, pp. 2096–2030, 2016.
  6. “Adversarial discriminative domain adaptation,” in Proc. CVPR, 2017, pp. 7167–7176.
  7. “Diverse image-to-image translation via disentangled representations,” in Proc. ECCV, 2018, pp. 35–51.
  8. “T2net: Synthetic-to-realistic translation for solving single-image depth estimation tasks,” in Proc. ECCV, 2018, pp. 767–783.
  9. “Unconstrained facial action unit detection via latent feature domain,” IEEE Transactions on Affective Computing, vol. 13, no. 2, pp. 1111–1126, 2022.
  10. “Jaa-net: joint facial action unit detection and face alignment via adaptive attention,” International Journal of Computer Vision, vol. 129, pp. 321–340, 2021.
  11. “Dual learning for joint facial landmark detection and action unit recognition,” IEEE Transactions on Affective Computing, 2021.
  12. “Multi-order networks for action unit detection,” IEEE Transactions on Affective Computing, 2022.
  13. “Cbam: Convolutional block attention module,” in Proc. ECCV, 2018, pp. 3–19.
  14. “Knowledge distillation with the reused teacher classifier,” in Proc. CVPR, 2022, pp. 11933–11942.
  15. “Fitnets: Hints for thin deep nets,” Proc. ICLR, vol. 2, pp. 3, 2015.
  16. “Knowledge distillation via softmax regression representation learning,” in Proc. ICLR, 2020.
  17. “Mma-net: Multi-view mixed attention mechanism for facial action unit detection,” Pattern Recognition Letters, 2023.
  18. “Multi-level adaptive region of interest and graph learning for facial action unit recognition,” in Proc. ICASSP. IEEE, 2021, pp. 2005–2009.
  19. “Bp4d-spontaneous: a high-resolution spontaneous 3d dynamic facial expression database,” Image and Vision Computing, vol. 32, no. 10, pp. 692–706, 2014.
  20. “Emotionet: An accurate, real-time algorithm for the automatic annotation of a million facial expressions in the wild,” in Proc. CVPR, 2016, pp. 5562–5570.

Summary

We haven't generated a summary for this paper yet.