Dualistic Meta-Learning for Open-Set DG (MEDIC)
- The paper introduces MEDIC, a framework that leverages dualistic meta-learning to jointly optimize inter-domain and inter-class splits for robust decision boundaries.
- It employs episodic training with domain-wise and class-wise splits to mitigate class imbalance and enhance cross-domain transferability.
- Experimental evaluations on PACS and Digits-DG datasets show that MEDIC outperforms ERM-bcls and MVDG-bcls in both OSCR and H-score metrics.
Dualistic MEta-learning with joint DomaIn-Class matching (MEDIC) is a meta-learning framework designed for the Open-Set @@@@1@@@@ (OSDG) problem. OSDG extends standard Domain Generalization (DG) to the more realistic regime where both domain distribution shift and previously unseen class categories may appear at test time. MEDIC establishes generalizable decision boundaries through dualistic meta-learning that jointly leverages inter-domain and inter-class data splits, optimized to handle the inherent class imbalance of open-set recognition and the challenges of cross-domain transferability (Wang et al., 2023).
1. Problem Formulation: Open-Set Domain Generalization
In OSDG, the training data comprises source domains , each drawn from a distinct distribution over input-label pairs , with the set of known classes. During evaluation, a model faces unseen target domains defined by a distribution over with labels coming from , where is a set of unknown classes disjoint from . Crucially, samples with are never observed during training.
The objective is to train a model to (i) correctly classify target-domain data from and (ii) reliably reject data from as unknown, despite never seeing during training. This scenario encompasses both distribution (domain) shift and class novelty, exceeding the complexity of standard DG or closed-set OOD detection.
2. One-vs-All Boundaries and Class Imbalance Challenge
The open-set nature of OSDG motivates the use of one-vs-all classifiers. For each , a binary classifier outputs ; a test example is declared "unknown" if for some detection threshold . This approach directly partitions the feature space, allowing explicit rejection of outliers.
However, each one-vs-all binary classifier faces extreme class imbalance: positives are rare (samples from class ), while negatives—conceptually the union of all other classes—are much more numerous and diverse. This imbalance tends to bias decision boundaries—either crowding them against the positive region (yielding over-confident, under-generalized boundaries) or expanding them too far (reducing detection sensitivity), which degrades both known-class accuracy and unknown rejection. These limitations are accentuated under domain shift, further complicating OSDG.
3. Dualistic Meta-Learning and Joint Domain-Class Matching
MEDIC addresses OSDG via a dualistic, episodic meta-learning protocol combining inter-domain and inter-class splits:
- Domain-wise split: Randomly divide source domains into two disjoint sets, and .
- Class-wise split: Partition known classes into two subsets, and .
This induces four training groups (episodes):
| Episode | Domains | Classes |
|---|---|---|
Meta-training uses and (differing in both domain and class), while meta-testing uses the crossed pair and .
At each iteration:
- Mini-batches , , , are sampled from the four splits.
- The meta-train loss is , with defined below.
- Take a virtual gradient step: .
- Compute the meta-test loss at : .
- The final update is .
A first-order Taylor expansion reveals an alignment term in the resulting objective, which encourages the model to discover parameter updates that generalize across both domains and classes. MEDIC's regularizer combines four inner products, each linking episodes that are either same-domain/different-class or same-class/different-domain, which explicitly enforces domain and class generalization and prevents decision boundary skew from class imbalance.
4. Objective Functions, Losses, and Decision Strategy
The core losses are:
- Closed-set cross-entropy: (softmax classifier).
- One-vs-all binary loss (Ovanet style):
This encourages the correct class output to be high while pushing down the strongest negative class.
- Combined open-set loss: .
The per-iteration meta-objective is:
- ,
- ,
- ,
- Update toward minimizing .
Rejection at test time: For an input , compute class confidence by either or where . Classify as "known" if , "unknown" otherwise.
5. Addressing Class Imbalance and Generalization
MEDIC's distinct class-wise and domain-wise splits during episodic meta-learning enforce decision boundaries that are robust to the relative proportion of positives and negatives—mitigating the dominant-negative effect and ensuring that boundaries are not distorted due to imbalance. Simultaneous domain-wise splitting fosters invariance to domain shift, preventing collapse of known classes when faced with novel target domains.
In practice, this yields emphatic improvements on open-set metrics such as OSCR (Open Set Classification Rate) and H-score, while maintaining or very marginally improving closed-set recognition compared to standard DG algorithms. A plausible implication is that MEDIC's principled matching approach is particularly suited for applications where previously unseen categories are expected under domain shift (Wang et al., 2023).
6. Experimental Evaluation
Performance metrics are reported primarily on PACS (6 known, 1 unknown class splits; Deep ResNet-18/50) and Digits-DG (6 known, 4 unknown classes; ConvNet). Key results are summarized as follows:
| Method | Dataset | OSCR | H-score |
|---|---|---|---|
| ERM-bcls | PACS-R18 | 70.3% | 70.6% |
| MVDG-bcls | PACS-R18 | 75.82% | 70.9% |
| MEDIC-bcls | PACS-R18 | 77.89% | 75.98% |
| MVDG-bcls | PACS-R50 | 83.21% | — |
| MEDIC-bcls | PACS-R50 | 84.85% | — |
| ERM-bcls | Digits-DG | 66.03% | — |
| MEDIC-bcls | Digits-DG | 71.15% | — |
MEDIC consistently outperforms ERM-bcls and MVDG-bcls on both OSCR and H-score. In all considered regimes, it also matches or slightly outperforms strong closed-domain generalization (SWAD) in closed-set recognition (Wang et al., 2023).
7. Algorithm Workflow Summary
The per-iteration workflow is as follows:
- Randomly split source domains into and classes into .
- Sample mini-batches: from , from , from , from .
- Compute meta-train loss .
- Take a virtual update: .
- Compute meta-test loss .
- Final update: .
- Repeat until convergence.
This episodic, dual-split procedure underpins MEDIC's generalizability to novel domains and unknown classes (Wang et al., 2023).