Papers
Topics
Authors
Recent
Gemini 2.5 Flash
Gemini 2.5 Flash
110 tokens/sec
GPT-4o
56 tokens/sec
Gemini 2.5 Pro Pro
44 tokens/sec
o3 Pro
6 tokens/sec
GPT-4.1 Pro
47 tokens/sec
DeepSeek R1 via Azure Pro
28 tokens/sec
2000 character limit reached

VAEM: a Deep Generative Model for Heterogeneous Mixed Type Data (2006.11941v1)

Published 21 Jun 2020 in cs.LG and stat.ML

Abstract: Deep generative models often perform poorly in real-world applications due to the heterogeneity of natural data sets. Heterogeneity arises from data containing different types of features (categorical, ordinal, continuous, etc.) and features of the same type having different marginal distributions. We propose an extension of variational autoencoders (VAEs) called VAEM to handle such heterogeneous data. VAEM is a deep generative model that is trained in a two stage manner such that the first stage provides a more uniform representation of the data to the second stage, thereby sidestepping the problems caused by heterogeneous data. We provide extensions of VAEM to handle partially observed data, and demonstrate its performance in data generation, missing data prediction and sequential feature selection tasks. Our results show that VAEM broadens the range of real-world applications where deep generative models can be successfully deployed.

User Edit Pencil Streamline Icon: https://streamlinehq.com
Authors (5)
  1. Chao Ma (187 papers)
  2. Sebastian Tschiatschek (43 papers)
  3. José Miguel Hernández-Lobato (151 papers)
  4. Richard Turner (11 papers)
  5. Cheng Zhang (388 papers)
Citations (59)

Summary

We haven't generated a summary for this paper yet.