Papers
Topics
Authors
Recent
Assistant
AI Research Assistant
Well-researched responses based on relevant abstracts and paper content.
Custom Instructions Pro
Preferences or requirements that you'd like Emergent Mind to consider when generating responses.
Gemini 2.5 Flash
Gemini 2.5 Flash 134 tok/s
Gemini 2.5 Pro 41 tok/s Pro
GPT-5 Medium 25 tok/s Pro
GPT-5 High 28 tok/s Pro
GPT-4o 86 tok/s Pro
Kimi K2 203 tok/s Pro
GPT OSS 120B 445 tok/s Pro
Claude Sonnet 4.5 37 tok/s Pro
2000 character limit reached

Line of Sight: On Linear Representations in VLLMs (2506.04706v1)

Published 5 Jun 2025 in cs.CV and cs.AI

Abstract: LLMs can be equipped with multimodal capabilities by fine-tuning on embeddings of visual inputs. But how do such multimodal models represent images in their hidden activations? We explore representations of image concepts within LlaVA-Next, a popular open-source VLLM. We find a diverse set of ImageNet classes represented via linearly decodable features in the residual stream. We show that the features are causal by performing targeted edits on the model output. In order to increase the diversity of the studied linear features, we train multimodal Sparse Autoencoders (SAEs), creating a highly interpretable dictionary of text and image features. We find that although model representations across modalities are quite disjoint, they become increasingly shared in deeper layers.

Summary

  • The paper demonstrates that VLLMs encode semantic features linearly, as evidenced by competitive ImageNet classifications using linear probes.
  • It employs steering vectors with Contrastive Activation Addition to causally manipulate model outputs at early transformer layers.
  • Sparse autoencoders reveal a hierarchical structure in multimodal features, highlighting both interpretability and limitations for adversarial use.

Line of Sight: On Linear Representations in VLLMs

Introduction

The paper "Line of Sight: On Linear Representations in VLLMs" explores how vision-LLMs (VLLMs), specifically LlaVA-Next, represent images in their hidden activations. The paper focuses on the linear decodability of image concepts, specifically ImageNet classes, from the residual stream of transformer models. The paper utilizes various techniques, including linear probes, steering vectors, and Sparse Autoencoders (SAEs), to decrypt and manipulate these hidden representations.

Linear Probes and Their Efficacy

Linear probe techniques are a primary method explored to assess the presence of decodable features in the residual streams of VLLMs. To determine if the ImageNet class information is linearly encodable, linear classifiers trained on the mean embeddings of image tokens at different layers are applied. Figure 1

Figure 1: ImageNet Accuracy. We benchmark linear probes trained on mean embeddings across image tokens on ImageNet Classification.

The performance of these probes is found to be competitive with state-of-the-art self-supervised learning (SSL) techniques, suggesting that a significant portion of semantic information is indeed encoded linearly.

Steering Vectors: Causal Manipulation of Model Representations

Steering vectors emerge as a pivotal method for causally testing the influence of linearly encodable features within model representations. By using Contrastive Activation Addition (CAA), the paper investigates model behavior manipulation. Steering vectors are employed to test the causal effect of such linear decodings on model outputs. Figure 2

Figure 2: Steering Vector Interventions. Overview and evaluation of steering vector effects.

The paper reinforces that steering at earlier layers is most fruitful, indicating an "information transfer" point where image semantics significantly influence text tokens. This highlights the potential for targeted interventions to dynamically control model outputs.

Sparse Autoencoders: Expanding Multimodal Feature Interpretability

The paper also explores the use of Sparse Autoencoders (SAEs) to extend the exploration of linear feature spaces beyond simple probing. SAEs provide a scalable method for discovering interpretable features within a model's residual streams. Figure 3

Figure 3: Sparse Autoencoder Training and Evaluation. Training overview and loss recovery measures.

The deployment of SAEs in VLLMs unveils a hierarchy of information across different layers, confirming that while individual neurons are informative, they are less structured than SAE features. This approach underscores SAE's effectiveness in decoding complex feature sets like those in ImageNet.

Adversarial Implications and Limitations

Although the linear representations offer significant insights and manipulative control, the paper acknowledges the inherent limitations in their deployment for adversarial attacks. The linear features provide a comparatively brittle basis for such applications, indicating areas for further exploration and refinement. Figure 4

Figure 4: Analysis of noise robustness in adversarial settings.

Conclusion

The research conducted in "Line of Sight: On Linear Representations in VLLMs" contributes valuable insights into the role and potential of linear feature representations in vision-LLMs. Through robust methodologies like linear probes, steering vectors, and sparse autoencoders, the paper navigates the hidden intricacies of multimodal models, revealing both the promise and limitations of current frameworks. Future work should aim to deepen the understanding of information conversion from image to text in VLLMs, potentially revolutionizing how such models are trained and deployed in real-world scenarios.

Dice Question Streamline Icon: https://streamlinehq.com

Open Problems

We haven't generated a list of open problems mentioned in this paper yet.

List To Do Tasks Checklist Streamline Icon: https://streamlinehq.com

Collections

Sign up for free to add this paper to one or more collections.

X Twitter Logo Streamline Icon: https://streamlinehq.com

Tweets

This paper has been mentioned in 1 tweet and received 11 likes.

Upgrade to Pro to view all of the tweets about this paper: