Papers
Topics
Authors
Recent
Gemini 2.5 Flash
Gemini 2.5 Flash
119 tokens/sec
GPT-4o
56 tokens/sec
Gemini 2.5 Pro Pro
43 tokens/sec
o3 Pro
6 tokens/sec
GPT-4.1 Pro
47 tokens/sec
DeepSeek R1 via Azure Pro
28 tokens/sec
2000 character limit reached

Hiera: A Hierarchical Vision Transformer without the Bells-and-Whistles (2306.00989v1)

Published 1 Jun 2023 in cs.CV and cs.LG

Abstract: Modern hierarchical vision transformers have added several vision-specific components in the pursuit of supervised classification performance. While these components lead to effective accuracies and attractive FLOP counts, the added complexity actually makes these transformers slower than their vanilla ViT counterparts. In this paper, we argue that this additional bulk is unnecessary. By pretraining with a strong visual pretext task (MAE), we can strip out all the bells-and-whistles from a state-of-the-art multi-stage vision transformer without losing accuracy. In the process, we create Hiera, an extremely simple hierarchical vision transformer that is more accurate than previous models while being significantly faster both at inference and during training. We evaluate Hiera on a variety of tasks for image and video recognition. Our code and models are available at https://github.com/facebookresearch/hiera.

User Edit Pencil Streamline Icon: https://streamlinehq.com
Authors (13)
  1. Chaitanya Ryali (4 papers)
  2. Yuan-Ting Hu (12 papers)
  3. Daniel Bolya (14 papers)
  4. Chen Wei (72 papers)
  5. Haoqi Fan (33 papers)
  6. Po-Yao Huang (31 papers)
  7. Vaibhav Aggarwal (8 papers)
  8. Arkabandhu Chowdhury (8 papers)
  9. Omid Poursaeed (19 papers)
  10. Judy Hoffman (75 papers)
  11. Jitendra Malik (211 papers)
  12. Yanghao Li (43 papers)
  13. Christoph Feichtenhofer (52 papers)
Citations (99)

Summary

Overview of "Hiera: A Hierarchical Vision Transformer without the Bells-and-Whistles"

Introduction

The paper presents Hiera, a simplified hierarchical vision transformer model that cuts down on the complexity prevalent in many state-of-the-art models by employing Masked Autoencoder (MAE) pretraining. This strategy enables the model to learn spatial biases inherently, eliminating the need for specialized architectural components that often slow down these models.

Key Contributions

  • Simplicity and Efficiency: Hiera is designed to eschew the cumbersome modules found in models like MViT and Swin. By leveraging MAE pretraining, Hiera maintains competitive accuracy while being significantly faster during both inference and training.
  • Sparse MAE Pretraining: Extensive experiments demonstrate the efficacy of sparse MAE pretraining. The authors cleverly utilize separate mask units for aligning MAE with the hierarchical transformers, addressing the challenges of rigid 2D grids.
  • Model Simplification: The paper systematically removes architectural complexities such as relative position embeddings, attention residuals, and convolutional layers, illustrating that these are unnecessary when spatial biases are learned through pretraining.

Empirical Results

  • Strong Numerical Performance: Hiera surpasses the state-of-the-art on various image recognition tasks and performs exceptionally well on video recognition tasks with substantial improvements over previous works on datasets like Kinetics-400 and Something-Something-v2.
  • Speed Improvements: The model achieves up to 2.4x faster performance compared to MViTv2 on images, and significantly reduces inference times on video tasks, making it highly efficient for practical deployment.

Implications and Future Directions

The results suggest that the removal of complex components from vision transformers is feasible, provided that robust pretraining techniques like MAE are employed. This raises questions about the necessity of certain architectural features traditionally deemed essential in transformer-based models.

Future developments could focus on further integrating Hiera-style architectures with additional pretraining paradigms, exploring the generalizability of this simplification strategy across other domains in AI.

Conclusion

"Hiera: A Hierarchical Vision Transformer without the Bells-and-Whistles" takes a bold step towards simplifying vision transformers without compromising performance. By demonstrating that learned biases can effectively replace manually designed modules, this research opens avenues for more efficient model designs in computer vision and beyond.

Github Logo Streamline Icon: https://streamlinehq.com
X Twitter Logo Streamline Icon: https://streamlinehq.com

Tweets