Papers
Topics
Authors
Recent
Gemini 2.5 Flash
Gemini 2.5 Flash
41 tokens/sec
GPT-4o
59 tokens/sec
Gemini 2.5 Pro Pro
41 tokens/sec
o3 Pro
7 tokens/sec
GPT-4.1 Pro
50 tokens/sec
DeepSeek R1 via Azure Pro
28 tokens/sec
2000 character limit reached

LieRE: Generalizing Rotary Position Encodings (2406.10322v2)

Published 14 Jun 2024 in cs.CV and cs.LG

Abstract: While Rotary Position Embeddings (RoPE) for LLMs have become widely adopted, their application for other modalities has been slower. Here, we introduce Lie group Relative position Encodings (LieRE) that goes beyond RoPE in supporting n-dimensional inputs. We evaluate the performance of LieRE on 2D and 3D image classification tasks and observe that LieRE leads to marked relative improvements in performance (up to 9.7% for 2D and up to 25.5% for 3D), training efficiency (3.5x reduction), data efficiency (30%) compared to the baselines of DeiT III, RoPE-Mixed and Vision-Llama. https://github.com/Stanford-AIMI/LieRE

User Edit Pencil Streamline Icon: https://streamlinehq.com
Authors (5)
  1. Sophie Ostmeier (11 papers)
  2. Brian Axelrod (11 papers)
  3. Michael E. Moseley (1 paper)
  4. Akshay Chaudhari (34 papers)
  5. Curtis Langlotz (24 papers)
X Twitter Logo Streamline Icon: https://streamlinehq.com

Tweets

Youtube Logo Streamline Icon: https://streamlinehq.com