Papers
Topics
Authors
Recent
Gemini 2.5 Flash
Gemini 2.5 Flash
41 tokens/sec
GPT-4o
59 tokens/sec
Gemini 2.5 Pro Pro
41 tokens/sec
o3 Pro
7 tokens/sec
GPT-4.1 Pro
50 tokens/sec
DeepSeek R1 via Azure Pro
28 tokens/sec
2000 character limit reached

Task-customized Masked AutoEncoder via Mixture of Cluster-conditional Experts (2402.05382v1)

Published 8 Feb 2024 in cs.CV and cs.LG

Abstract: Masked Autoencoder~(MAE) is a prevailing self-supervised learning method that achieves promising results in model pre-training. However, when the various downstream tasks have data distributions different from the pre-training data, the semantically irrelevant pre-training information might result in negative transfer, impeding MAE's scalability. To address this issue, we propose a novel MAE-based pre-training paradigm, Mixture of Cluster-conditional Experts (MoCE), which can be trained once but provides customized pre-training models for diverse downstream tasks. Different from the mixture of experts (MoE), our MoCE trains each expert only with semantically relevant images by using cluster-conditional gates. Thus, each downstream task can be allocated to its customized model pre-trained with data most similar to the downstream data. Experiments on a collection of 11 downstream tasks show that MoCE outperforms the vanilla MAE by 2.45\% on average. It also obtains new state-of-the-art self-supervised learning results on detection and segmentation.

User Edit Pencil Streamline Icon: https://streamlinehq.com
Authors (7)
  1. Zhili Liu (20 papers)
  2. Kai Chen (512 papers)
  3. Jianhua Han (49 papers)
  4. Lanqing Hong (72 papers)
  5. Hang Xu (204 papers)
  6. Zhenguo Li (195 papers)
  7. James T. Kwok (65 papers)
Citations (23)