Papers
Topics
Authors
Recent
Gemini 2.5 Flash
Gemini 2.5 Flash
102 tokens/sec
GPT-4o
59 tokens/sec
Gemini 2.5 Pro Pro
43 tokens/sec
o3 Pro
6 tokens/sec
GPT-4.1 Pro
50 tokens/sec
DeepSeek R1 via Azure Pro
28 tokens/sec
2000 character limit reached

Establishing an Evaluation Metric to Quantify Climate Change Image Realism (1910.10143v1)

Published 22 Oct 2019 in cs.LG, cs.CV, and stat.ML

Abstract: With success on controlled tasks, generative models are being increasingly applied to humanitarian applications [1,2]. In this paper, we focus on the evaluation of a conditional generative model that illustrates the consequences of climate change-induced flooding to encourage public interest and awareness on the issue. Because metrics for comparing the realism of different modes in a conditional generative model do not exist, we propose several automated and human-based methods for evaluation. To do this, we adapt several existing metrics, and assess the automated metrics against gold standard human evaluation. We find that using Fr\'echet Inception Distance (FID) with embeddings from an intermediary Inception-V3 layer that precedes the auxiliary classifier produces results most correlated with human realism. While insufficient alone to establish a human-correlated automatic evaluation metric, we believe this work begins to bridge the gap between human and automated generative evaluation procedures.

User Edit Pencil Streamline Icon: https://streamlinehq.com
Authors (5)
  1. Sharon Zhou (15 papers)
  2. Alexandra Luccioni (11 papers)
  3. Gautier Cosne (4 papers)
  4. Michael S. Bernstein (47 papers)
  5. Yoshua Bengio (601 papers)
Citations (6)