Papers
Topics
Authors
Recent
Gemini 2.5 Flash
Gemini 2.5 Flash
119 tokens/sec
GPT-4o
56 tokens/sec
Gemini 2.5 Pro Pro
43 tokens/sec
o3 Pro
6 tokens/sec
GPT-4.1 Pro
47 tokens/sec
DeepSeek R1 via Azure Pro
28 tokens/sec
2000 character limit reached

Your representations are in the network: composable and parallel adaptation for large scale models (2303.04105v2)

Published 7 Mar 2023 in cs.LG and cs.CV

Abstract: We propose InCA, a lightweight method for transfer learning that cross-attends to any activation layer of a pre-trained model. During training, InCA uses a single forward pass to extract multiple activations, which are passed to external cross-attention adapters, trained anew and combined or selected for downstream tasks. We show that, even when selecting a single top-scoring adapter, InCA achieves performance comparable to full fine-tuning, at a cost comparable to fine-tuning just the last layer. For example, with a cross-attention probe 1.3% the size of a pre-trained ViT-L/16 model, we achieve performance within 0.2% of the full fine-tuning paragon at a computational training cost of 51% of the baseline, on average across 11 downstream classification. Unlike other forms of efficient adaptation, InCA does not require backpropagating through the pre-trained model, thus leaving its execution unaltered at both training and inference. The versatility of InCA is best illustrated in fine-grained tasks, which may require accessing information absent in the last layer but accessible in intermediate layer activations. Since the backbone is fixed, InCA allows parallel ensembling as well as parallel execution of multiple tasks. InCA achieves state-of-the-art performance in the ImageNet-to-Sketch multi-task benchmark.

User Edit Pencil Streamline Icon: https://streamlinehq.com
Authors (10)
  1. Yonatan Dukler (10 papers)
  2. Alessandro Achille (60 papers)
  3. Hao Yang (328 papers)
  4. Varsha Vivek (3 papers)
  5. Luca Zancato (21 papers)
  6. Benjamin Bowman (10 papers)
  7. Avinash Ravichandran (35 papers)
  8. Charless Fowlkes (35 papers)
  9. Ashwin Swaminathan (18 papers)
  10. Stefano Soatto (179 papers)
Citations (2)

Summary

We haven't generated a summary for this paper yet.