Papers
Topics
Authors
Recent
Gemini 2.5 Flash
Gemini 2.5 Flash
97 tokens/sec
GPT-4o
53 tokens/sec
Gemini 2.5 Pro Pro
44 tokens/sec
o3 Pro
5 tokens/sec
GPT-4.1 Pro
47 tokens/sec
DeepSeek R1 via Azure Pro
28 tokens/sec
2000 character limit reached

Learning Gaze-aware Compositional GAN (2405.20643v1)

Published 31 May 2024 in cs.CV and cs.AI

Abstract: Gaze-annotated facial data is crucial for training deep neural networks (DNNs) for gaze estimation. However, obtaining these data is labor-intensive and requires specialized equipment due to the challenge of accurately annotating the gaze direction of a subject. In this work, we present a generative framework to create annotated gaze data by leveraging the benefits of labeled and unlabeled data sources. We propose a Gaze-aware Compositional GAN that learns to generate annotated facial images from a limited labeled dataset. Then we transfer this model to an unlabeled data domain to take advantage of the diversity it provides. Experiments demonstrate our approach's effectiveness in generating within-domain image augmentations in the ETH-XGaze dataset and cross-domain augmentations in the CelebAMask-HQ dataset domain for gaze estimation DNN training. We also show additional applications of our work, which include facial image editing and gaze redirection.

User Edit Pencil Streamline Icon: https://streamlinehq.com
Authors (7)
  1. Nerea Aranjuelo (6 papers)
  2. Siyu Huang (50 papers)
  3. Ignacio Arganda-Carreras (10 papers)
  4. Luis Unzueta (1 paper)
  5. Oihana Otaegui (9 papers)
  6. Hanspeter Pfister (131 papers)
  7. Donglai Wei (46 papers)

Summary

We haven't generated a summary for this paper yet.