Papers
Topics
Authors
Recent
Gemini 2.5 Flash
Gemini 2.5 Flash
97 tokens/sec
GPT-4o
53 tokens/sec
Gemini 2.5 Pro Pro
44 tokens/sec
o3 Pro
5 tokens/sec
GPT-4.1 Pro
47 tokens/sec
DeepSeek R1 via Azure Pro
28 tokens/sec
2000 character limit reached

GLSR-VAE: Geodesic Latent Space Regularization for Variational AutoEncoder Architectures (1707.04588v1)

Published 14 Jul 2017 in cs.LG, cs.AI, and stat.ML

Abstract: VAEs (Variational AutoEncoders) have proved to be powerful in the context of density modeling and have been used in a variety of contexts for creative purposes. In many settings, the data we model possesses continuous attributes that we would like to take into account at generation time. We propose in this paper GLSR-VAE, a Geodesic Latent Space Regularization for the Variational AutoEncoder architecture and its generalizations which allows a fine control on the embedding of the data into the latent space. When augmenting the VAE loss with this regularization, changes in the learned latent space reflects changes of the attributes of the data. This deeper understanding of the VAE latent space structure offers the possibility to modulate the attributes of the generated data in a continuous way. We demonstrate its efficiency on a monophonic music generation task where we manage to generate variations of discrete sequences in an intended and playful way.

User Edit Pencil Streamline Icon: https://streamlinehq.com
Authors (3)
  1. Gaëtan Hadjeres (24 papers)
  2. Frank Nielsen (125 papers)
  3. François Pachet (14 papers)
Citations (64)

Summary

We haven't generated a summary for this paper yet.