Papers
Topics
Authors
Recent
Gemini 2.5 Flash
Gemini 2.5 Flash
119 tokens/sec
GPT-4o
56 tokens/sec
Gemini 2.5 Pro Pro
43 tokens/sec
o3 Pro
6 tokens/sec
GPT-4.1 Pro
47 tokens/sec
DeepSeek R1 via Azure Pro
28 tokens/sec
2000 character limit reached

Universal Representation Learning from Multiple Domains for Few-shot Classification (2103.13841v1)

Published 25 Mar 2021 in cs.CV

Abstract: In this paper, we look at the problem of few-shot classification that aims to learn a classifier for previously unseen classes and domains from few labeled samples. Recent methods use adaptation networks for aligning their features to new domains or select the relevant features from multiple domain-specific feature extractors. In this work, we propose to learn a single set of universal deep representations by distilling knowledge of multiple separately trained networks after co-aligning their features with the help of adapters and centered kernel alignment. We show that the universal representations can be further refined for previously unseen domains by an efficient adaptation step in a similar spirit to distance learning methods. We rigorously evaluate our model in the recent Meta-Dataset benchmark and demonstrate that it significantly outperforms the previous methods while being more efficient. Our code will be available at https://github.com/VICO-UoE/URL.

User Edit Pencil Streamline Icon: https://streamlinehq.com
Authors (3)
  1. Wei-Hong Li (14 papers)
  2. Xialei Liu (35 papers)
  3. Hakan Bilen (62 papers)
Citations (76)

Summary

Universal Representation Learning from Multiple Domains for Few-Shot Classification

The paper addresses the challenge of few-shot classification, which involves learning classifiers for novel classes from minimal labeled samples, particularly in scenarios where these classes belong to entirely new domains. While traditional machine learning models typically rely on abundant labeled data, the goal of few-shot classification is to enhance data efficiency to achieve human-level adaptability, whereby models can generalize from limited information. Existing methods typically employ adaptation mechanisms to align feature representations to novel domains or utilize a plethora of domain-specific feature extractors for feature selection.

The authors propose an innovative method for learning universal deep representations by leveraging knowledge distillation from multiple individually trained models across diverse domains. This process involves co-aligning domain-specific features using small task-specific adapters and Centered Kernel Alignment (CKA), ultimately yielding a singular "universal" set of deep representations which can be further refined for previously unseen domains using techniques analogous to distance learning methods.

The paper's experimental evaluations are notable, conducted rigorously on the Meta-Dataset benchmark, a challenging environment comprising ten datasets encompassing distinct domains for meta-training and meta-testing. Remarkably, the proposed method surmounts prior state-of-the-art techniques, not only excelling in accuracy but also demonstrating superior computational efficiency. This efficiency arises due to the reliance on a single universal model compared to prior methods requiring multiple domain-specific models.

Key technical insights include the use of task-specific adapters to diminish negative transfer during feature alignment. Centered Kernel Alignment (CKA) is utilized to measure non-linear similarities between intermediate representations, which is crucial given the varied nature of the domains involved. These techniques enhance the robustness of the universal feature set, making it better suited to handle the inherent domain shift that occurs during few-shot learning.

The implications of this research are profound. Practically, it promises advancements in few-shot learning applications across fields requiring rapid adaptation to new data with minimal supervision, such as natural language processing and computer vision tasks, including those pertinent to autonomous systems and healthcare diagnostics. Theoretically, it contributes to the ongoing discourse on universal representation learning, presenting a scalable solution that challenges the necessity of domain-specific networks in favor of a singular universal network.

Future research could further explore non-linear transformation methods for feature alignment beyond linear adapters and delve into hybrid models that combine the benefits of both universal and domain-specific representations. Additionally, exploring adaptive scheduling of the distillation process could enhance the model’s capacity to balance generality and specificity across domains.

In summary, this paper significantly progresses the domain of few-shot classification by demonstrating the efficacy of universal representation learning in a multi-domain setting, opening pathways for more adaptable, efficient AI systems.