Papers
Topics
Authors
Recent
Assistant
AI Research Assistant
Well-researched responses based on relevant abstracts and paper content.
Custom Instructions Pro
Preferences or requirements that you'd like Emergent Mind to consider when generating responses.
Gemini 2.5 Flash
Gemini 2.5 Flash 134 tok/s
Gemini 2.5 Pro 41 tok/s Pro
GPT-5 Medium 26 tok/s Pro
GPT-5 High 27 tok/s Pro
GPT-4o 100 tok/s Pro
Kimi K2 204 tok/s Pro
GPT OSS 120B 433 tok/s Pro
Claude Sonnet 4.5 37 tok/s Pro
2000 character limit reached

Zero/One-Shot DDA Techniques

Updated 26 October 2025
  • Zero/One-Shot DDA is a set of techniques that enable model adaptation using minimal to no target examples through generative models, metric learning, and semantic representation.
  • It employs methods like GAN-based augmentation, meta-learning, and parameter regression to synthesize data and facilitate robust performance across unseen tasks.
  • These approaches are applied in fields such as speech recognition, robotic grasping, and NAS, demonstrating significant improvements in accuracy and efficiency with extremely limited training data.

Zero/One-Shot DDA refers to methodologies for data-driven augmentation, adaptation, or learning in scenarios where only zero (no target examples) or one (a single target example) from a new distribution, task, or class is available. These paradigms underpin advanced forms of zero-shot and one-shot learning and are critical in domains ranging from speech recognition for rare clinical populations to generative modeling, neural architecture search, and interactive perception systems. The unifying thread is the ability to robustly transfer, synthesize, or adapt models and data for new classes or tasks with extreme data scarcity, often via explicit domain adaptation, generative models, semantic representation learning, or specialized augmentation strategies.

1. Foundations and Scope

Zero/One-Shot DDA encompasses several mechanisms:

  • Zero-shot learning: Generalization to new classes, domains, or tasks without any explicit examples from the target domain; knowledge is transferred from auxiliary semantic sources or via global property sharing.
  • One-shot learning: Efficient adaptation or augmentation for new entities or tasks using only a single reference; methods typically exploit exemplar information via metric learning, generative modeling, or parameter regression.
  • Data-driven augmentation (DDA): Techniques for expanding limited data scenarios by synthesizing or generating pseudo-samples, simulating new domains, or leveraging compositional semantic diversity.

Such strategies are essential in high-impact applications where collecting large datasets is infeasible or the potential data diversity is not reflected in available resources.

2. Methodological Approaches

Several technical frameworks implementing zero/one-shot DDA have been developed:

Generative Modeling and Augmentation

  • One-shot domain adaptation with generative adversarial networks, e.g., StyleGAN-based approaches shift a pre-trained generator's output distribution towards a single target example using iterative optimization and style-mixing, enabling unlimited synthetic sample generation reflecting both generic and target-specific statistics (Yang et al., 2020). Manifold-shifting and style-layer transfer ensure both global structure and low-level details are preserved, critical for downstream data augmentation (e.g., face manipulation detection).
  • Diverse and faithful one-shot adaptation leverages global CLIP embedding difference (domain gap) for coarse adaptation and attentive style loss for token-wise local alignment. Selective cross-domain consistency in the GAN latent space retains diversity (Zhang et al., 2022). The framework can be extended to zero-shot adaptation using CLIP text encodings.

Semantic Representation and Consolidation

  • Attribute-driven subpopulation modeling replaces single class prototype vectors with sets of attribute-inferred vectors via LLM inference. Classification is performed by consolidating predictions nonlinearly (top-k selection/averaging) over these subpopulations, enhancing performance on atypical and diverse instances without retraining (Moayeri et al., 25 Apr 2024). The approach also enables transparent, interpretable prediction and debugging.

Task-Oriented Parameter Regression

  • Zero-shot task transfer with meta-learning regresses model parameters for unseen tasks from the weight-space of known tasks, guided by explicit task correlation matrices derived from crowdsourcing or other sources (Pal et al., 2019). This approach facilitates the synthesis of effective models for novel target tasks in the absence of ground truth.

Modular, Multi-Stage Recognition

  • Task-oriented grasping (TOG) with hybrid zero-/one-shot pipelines: Modular systems segment and recognize objects using zero-shot vision-LLMs (e.g., CLIP, SAM), select affordance regions with one-shot reference matching (AffCorrs), and find grasps within the affordance mask via geometric filtering (Holomjova et al., 5 Jun 2025). This design achieves 68.9% grasp accuracy in multi-object scenes and illustrates the trade-off between annotation burden and subcategory identification accuracy.

3. Benchmark Datasets and Evaluation Protocols

The development of zero/one-shot DDA algorithms has spurred creation of highly annotated benchmarks:

  • TD-TOG dataset includes 1,449 real-world annotated RGB-D scenes spanning 30 object categories, 120 subcategories, with masks, grasps, and affordances (Holomjova et al., 5 Jun 2025). It allows rigorous comparison of zero-shot and one-shot pipelines for TOG and highlights the demand for datasets that support unseen subcategory generalization.
  • Dysarthric speech recognition datasets (TORGO, CDSD): These are used in evaluating zero/one-shot data augmentation strategies for sentence-level ASR in clinical communication settings (Wang et al., 19 Oct 2025).

Empirical metrics include task-oriented grasp accuracy, F₁ scores for object segmentation/recognition, intersection-over-union for affording region detection, CIDEr and RefPAC for captioning, and word/character error rate for speech recognition.

4. Specialized Strategies for Extreme Data Scarcity

Text-Coverage Data Augmentation for Speech

  • Zero-shot DDA for dysarthric speech leverages a pre-defined set of representative texts, using generative models (VITS for zero-shot; Fish-Speech for one-shot) to synthesize speech samples aligned to the actual communication or rehabilitation domain (Wang et al., 19 Oct 2025). Fine-tuning the recognizer on these samples yields substantial error rate reduction for unseen speakers, clearly outperforming shallow fusion with text-only domain adaptation.

Neural Architecture Search via Ranking Distillation

  • RD-NAS framework combines zero-cost proxies (which rapidly score architectures with no training) as "teacher" with one-shot NAS supernet as "student", using margin ranking loss and group-aware subnet sampling to distill reliable architectural ranking signals (Dong et al., 2023). This yields up to a 10.7% improvement in ranking consistency, providing highly efficient search even under data constraints.

Patch-Centric Unified Captioning

  • Patch-centric zero-shot captioning frameworks aggregate dense, semantically-rich patch embeddings (from DINO/Talk2DINO backbones) for arbitrary region captioning. Captioners trained on text datasets only can generate descriptions for any region, including noncontiguous mouse traces, without region-level supervision, accomplishing state-of-the-art zero-shot dense and region-set captioning performance (Bianchi et al., 3 Oct 2025).

Zero-Shot Audio Effect Modeling

  • Zero-shot amplifier modeling via contrastive tone embeddings enables conditioning a generator for unseen amplifier tones at inference by computing reference-based embeddings (direct, nearest-retrieval, or mean) and modulating generator activations via FiLM layers (Chen et al., 15 Jul 2024). This one-to-many strategy, validated by STFT loss and t-SNE/qualitative spectrogram analyses, allows dynamic emulation for arbitrary analog devices.

5. Performance, Trade-offs, and Interpretability

Zero/one-shot DDA methods demonstrate notable performance improvements:

  • Substantial accuracy gains on conventional, generalized, and few-shot benchmarks by modular designs, automatic selection/fusion mechanisms, and balance between seen/unseen class diversity (Rahman et al., 2017).
  • Interpretable consolidation in subpopulation-driven classification delivers transparent explanations, insightfully revealing failure modes and enabling robust fairness/accuracy trade-offs (Moayeri et al., 25 Apr 2024).
  • Cross-task adaptation and ranking via meta-learning and distillation frameworks equip decision-making systems to extrapolate model parameters, search architectures, or adapt effect models without retraining or data.
  • Metrics such as harmonic mean, F₁ score, CIDEr, WER, and STFT loss are widely used for quantifying performance in benchmarking settings and highlight the comparative effectiveness of zero/one-shot DDA over manual or heuristics-based methods (Or et al., 2021).

A plausible implication is that adopting attribute-rich, patch-centric, or generative augmentative strategies will become standard practice in domains requiring robust generalization under data scarcity and diversity.

6. Practical Impact and Application Areas

Zero/One-Shot DDA supports broad high-value applications:

  • Clinical assistive technologies for dysarthric speech recognition and rehabilitation (Wang et al., 19 Oct 2025).
  • Automated dense image/video captioning, region tracing, and indexing for retrieval and assistive technology (Bianchi et al., 3 Oct 2025).
  • Digital emulation and creative sound design in audio, supporting zero-shot effect modeling and amplifier recreation (Chen et al., 15 Jul 2024).
  • Robotics and automated manipulation via task-oriented grasping with flexible object/affordance recognition pipelines (Holomjova et al., 5 Jun 2025).
  • Game design with deep-learned dynamic difficulty adjustment systems that optimally balance user experience and completion constraints in a one-shot scenario (Or et al., 2021).
  • Efficient NAS in resource-constrained environments, marrying fast zero-cost ranking with one-shot supernet training (Dong et al., 2023).

This suggests growing cross-domain applicability, especially as semantic knowledge bases, generative models, and vision-language representation frameworks mature.

7. Implications for Future Research

Current findings:

  • Scalable and interpretable augmentation is achievable via automated attribute inference, semantic consolidation, and patch-based representations without needing retraining, supporting fair and transparent AI system design.
  • Hybrid modular approaches (e.g., combining zero-shot VLM segmentation, one-shot visual reference affordance detection, and geometric filtering for grasps) enable robust adaptation across unseen object subcategories.
  • Meta-learning with parameter regression, coupled with explicit task-correlation quantification, enable structured prediction for entirely new tasks.
  • Contrastive and content-based condition embeddings in audio and vision provide generalizable, flexible conditioning for novel domains.

A plausible implication is that future systems will increasingly integrate LLM-driven semantic diversity, patch-centric representation, generative augmentation, and meta-adaptive pipelines to manage real-world category, style, and domain shifts under severe data constraints, especially as applications demand higher transparency, fairness, and robust generalization.

Forward Email Streamline Icon: https://streamlinehq.com

Follow Topic

Get notified by email when new papers are published related to Zero/One-Shot DDA.