Papers
Topics
Authors
Recent
Gemini 2.5 Flash
Gemini 2.5 Flash
97 tokens/sec
GPT-4o
53 tokens/sec
Gemini 2.5 Pro Pro
44 tokens/sec
o3 Pro
5 tokens/sec
GPT-4.1 Pro
47 tokens/sec
DeepSeek R1 via Azure Pro
28 tokens/sec
2000 character limit reached

Conditional Generative Modeling via Learning the Latent Space (2010.03132v2)

Published 7 Oct 2020 in cs.LG and cs.CV

Abstract: Although deep learning has achieved appealing results on several machine learning tasks, most of the models are deterministic at inference, limiting their application to single-modal settings. We propose a novel general-purpose framework for conditional generation in multimodal spaces, that uses latent variables to model generalizable learning patterns while minimizing a family of regression cost functions. At inference, the latent variables are optimized to find optimal solutions corresponding to multiple output modes. Compared to existing generative solutions, in multimodal spaces, our approach demonstrates faster and stable convergence, and can learn better representations for downstream tasks. Importantly, it provides a simple generic model that can beat highly engineered pipelines tailored using domain expertise on a variety of tasks, while generating diverse outputs. Our codes will be released.

User Edit Pencil Streamline Icon: https://streamlinehq.com
Authors (5)
  1. Sameera Ramasinghe (36 papers)
  2. Kanchana Ranasinghe (21 papers)
  3. Salman Khan (244 papers)
  4. Nick Barnes (81 papers)
  5. Stephen Gould (104 papers)
Citations (9)

Summary

We haven't generated a summary for this paper yet.