Coarse-To-Fine Few-Shot Class-Incremental Learning
- The paper introduces a novel hyperbolic embedding framework that aligns with hierarchical data to improve fine-grained recognition in few-shot class-incremental learning.
- It leverages contrastive optimization, weight normalization, and freezing in the Poincaré ball to preserve coarse-class knowledge while integrating new fine classes.
- Empirical benchmarks demonstrate significant improvements in classification accuracy and scalability, underscoring the method's practical impact on hierarchical and incremental learning tasks.
Coarse-To-Fine Few-Shot Class-Incremental Learning (C2FSCIL) addresses the challenge of incrementally learning a sequence of novel, fine-grained classes with only a few annotated examples per new class, while preserving recognition ability for previously learned, coarser-grained (e.g., superclass) labels. Standard approaches to few-shot class-incremental learning are limited when underlying data and label structures are hierarchical. Recent methods leverage hyperbolic embeddings to align model geometry with hierarchical semantics and introduce novel optimization and augmentation techniques adapted to the hyperbolic context, achieving improved scalability and performance under few-shot and incremental constraints.
1. Hyperbolic Space Representation for Hierarchical Data
Hierarchical structures such as coarse-to-fine (superclass/subclass) taxonomies naturally require an embedding space that can represent exponentially expanding relationships efficiently. Hyperbolic space, characteristically of negative curvature, provides a geometric growth rate well-matched to such semantic hierarchies. The Poincaré ball model is commonly used, where each point lies in an -dimensional ball of radius and the Riemannian metric is
where is the curvature. The hyperbolic distance between vectors and is given by
which reflects the exponential expansion away from the ball center. This property yields compact and discriminative representations for both coarse (central, generic) and fine (peripheral, specific) class semantics (Dai et al., 23 Sep 2025).
2. The Knowe Approach: Learning, Normalizing, and Freezing Weights
The Knowe algorithm forms the foundation of C2FSCIL (Xiang et al., 2021). It begins by learning a general-purpose embedding via supervised and contrastive learning on coarse labels. The base session constructs an embedding space wherein samples from the same superclass are mapped nearby, with inter-class repulsion. After this contrastive pre-training, incrementally learned fine classes are added via classifier weights (prototypes), which are -normalized and then frozen—ensuring stability (retention of old knowledge) and plasticity (adaptation to new classes). In the hyperbolic variant, these steps are performed in the Poincaré ball, and the classifier weights are normalized and frozen with respect to the geometry of the space. This strict freezing avoids catastrophic forgetting and prevents bias in the evolving classifier as new fine-grained categories are encountered in later sessions (Dai et al., 23 Sep 2025).
3. Hyperbolic Embedding and Model Architecture
Rather than extracting features in a Euclidean space, the model incorporates a hyperbolic mapping layer, typically using the exponential map to endow deep feature vectors with hyperbolic structure. The mapping transforms a tangent space vector (often from a learned base point ) into the Poincaré ball as
yielding vectors confined to . Downstream, classification is handled by hyperbolic fully connected (HypFC) layers. These layers employ Möbius addition and multiplication to retain hyperbolic compatibility: with and denoting Möbius matrix multiplication and addition. All operations, including the comparison of features and prototypes, use hyperbolic distance rather than Euclidean or dot-product similarity, maintaining geometry-consistency throughout the network (Dai et al., 23 Sep 2025).
4. Hyperbolic Contrastive Optimization and Few-Shot Augmentation
To exploit the advantages of hyperbolic geometry during learning, the standard contrastive loss is reformulated for hyperbolic space: where is a query and is its positive sample, with a temperature parameter. Classification and optimization are thus fully aligned with the underlying negative curvature. In few-shot settings, the method estimates the fine-class feature distribution in hyperbolic space using the maximum entropy principle. The wrapped normal distribution in the Poincaré ball is used for this purpose: Augmented features are sampled from this distribution, supplementing the meager set of real examples for each novel fine class, thus alleviating overfitting and improving classifier generalization under few-shot constraints (Dai et al., 23 Sep 2025).
5. Empirical Evaluation and Impact on C2FSCIL Benchmarks
Experimental results on benchmarks including CIFAR-100, tieredImageNet, and OpenEarthSensing demonstrate that hyperbolic embedding and contrastive loss yield marked improvements in both coarse-class and fine-class accuracy (Dai et al., 23 Sep 2025). Specifically, embeddings in hyperbolic space lead to superior hierarchical separation, which manifests as gains under incremental conditions (e.g., higher overall and fine-grained accuracy after multiple incremental sessions and a stable forgetting rate). Augmentation via maximum entropy sampling is especially effective for mitigating overfitting in the regime of very few labeled samples per new class. This suggests that the hyperbolic geometry is well suited to the hierarchical, coarse-to-fine progression inherent in C2FSCIL: it allows both stable partitioning for base (coarse) classes and flexible allocation for novel (fine) subcategories.
6. Discussion and Future Directions
The integration of hyperbolic geometry into C2FSCIL offers several advantages: it aligns model representation with the underlying data hierarchy, naturally supports scalable partitioning for increasingly fine labels, reduces catastrophic forgetting through weight normalization and freezing, and leverages geometric data augmentation to support generalization from scant supervision. Future directions include extending these techniques to more complex, multi-modal hierarchical tasks, devising adaptive curvature adjustment schemes, and exploring variants of hyperbolic metric augmentation tailored to specific hierarchical domains. Applications are anticipated in domains where data is both incremental and hierarchical, such as remote sensing, taxonomic biology, and real-time object recognition for robotics or autonomous vehicles (Dai et al., 23 Sep 2025).