Papers
Topics
Authors
Recent
Gemini 2.5 Flash
Gemini 2.5 Flash
97 tokens/sec
GPT-4o
53 tokens/sec
Gemini 2.5 Pro Pro
43 tokens/sec
o3 Pro
4 tokens/sec
GPT-4.1 Pro
47 tokens/sec
DeepSeek R1 via Azure Pro
28 tokens/sec
2000 character limit reached

LatentExplainer: Explaining Latent Representations in Deep Generative Models with Multi-modal Foundation Models (2406.14862v4)

Published 21 Jun 2024 in cs.LG, cs.CL, and cs.CV

Abstract: Deep generative models like VAEs and diffusion models have advanced various generation tasks by leveraging latent variables to learn data distributions and generate high-quality samples. Despite the field of explainable AI making strides in interpreting machine learning models, understanding latent variables in generative models remains challenging. This paper introduces \textit{LatentExplainer}, a framework for automatically generating semantically meaningful explanations of latent variables in deep generative models. \textit{LatentExplainer} tackles three main challenges: inferring the meaning of latent variables, aligning explanations with inductive biases, and handling varying degrees of explainability. Our approach perturbs latent variables, interpreting changes in generated data, and uses multi-modal LLMs (MLLMs) to produce human-understandable explanations. We evaluate our proposed method on several real-world and synthetic datasets, and the results demonstrate superior performance in generating high-quality explanations for latent variables. The results highlight the effectiveness of incorporating inductive biases and uncertainty quantification, significantly enhancing model interpretability.

User Edit Pencil Streamline Icon: https://streamlinehq.com
Authors (6)
  1. Mengdan Zhu (6 papers)
  2. Raasikh Kanjiani (2 papers)
  3. Jiahui Lu (3 papers)
  4. Andrew Choi (9 papers)
  5. Qirui Ye (1 paper)
  6. Liang Zhao (353 papers)

Summary

We haven't generated a summary for this paper yet.