Papers
Topics
Authors
Recent
2000 character limit reached

Two-Stage Pretraining Process

Updated 28 November 2025
  • Two-stage pretraining is a sequential approach that first leverages broad, weakly-labeled data with self-supervised objectives before transitioning to domain-specific fine-tuning.
  • It improves adaptation and efficiency by separating universal feature extraction from task-specific specialization, reducing overfitting and handling data scarcity.
  • Empirical results show gains such as up to 10–20% accuracy improvements in vision-language tasks and significant reductions in WER for ASR through strategic parameter tuning.

A two-stage pretraining process is a sequential approach for model initialization, representation learning, or adaptation, wherein a model is first optimized on one objective or data regime (“stage one”), then further refined using a second objective, modality, or data source (“stage two”). This strategy is prevalent across vision, language, multi-modal, and scientific domains, addressing challenges such as data scarcity, distribution mismatch, compute efficiency, compositional generalization, and parameter-efficient transfer. The two-stage methodology exploits the complementary benefits of distinct learning phases—for example, universal feature abstraction followed by domain/task-specific specialization, or unsupervised pretraining followed by supervised adaptation on auxiliary or target labels.

1. Motivations and Core Paradigms

Two-stage pretraining methods capitalize on (i) the availability of broad, often weakly-labeled or unlabeled corpora, and (ii) the necessity of efficient or robust adaptation to specialized tasks, domains, or distributions. Key paradigms include:

  • Distribution Bridging: Aligning a model to reduce feature or covariate shift before fine-grained adaptation (e.g., LayerNorm tuning prior to task tuning) (Zhao et al., 2023).
  • Task/Modality Curriculum: Solving easier or generic objectives (e.g., contrastive alignment, masking, denoising) prior to more specialized or difficult ones (Jamal et al., 5 Aug 2024, Wijaya et al., 5 Nov 2024).
  • Auxiliary Label Bootstrapping: Self-supervised or cheap-label pretraining before adaptation with few real or expensive labels (Wijaya et al., 5 Nov 2024).
  • Parameter-Efficient Adaptation: Structuring pretraining such that only a small set of parameters are tuned in the second stage, improving adaptation speed and generalization (Zhao et al., 2023).
  • Compute and Data Efficiency: Training universal representations with either dense, multi-source data or a single encoder, then optimizing a lightweight fusion or domain-specific module in the second phase (Li et al., 2019).

The structure and rationale of these approaches are tailored to the statistical and computational properties of the application, yielding significant improvements over monolithic or single-stage pretraining.

2. Representative Architectures and Methodological Variants

A diverse array of two-stage pretraining implementations has been proposed, each architecturally aligned to their application:

  • Vision-LLMs: "Chinese CLIP" (Yang et al., 2022) implements stage 1 as locked-image tuning (LiT: frozen vision encoder, text encoder adaptation) and stage 2 as full contrastive tuning (entire model joint optimization).
  • Video-to-Text: VideoOFA (Chen et al., 2023) applies initial large-scale image-text pretraining, followed by video-text-specific adaptation with temporal embedding injection and stage-dependent architectural options (full or per-frame encoding).
  • Multimodal/Multistream Speech: In end-to-end ASR, a universal feature extractor (UFE) is trained on all data in stage 1, then a stream-fusion attention network (e.g., HAN) on frozen UFE outputs in stage 2 (Li et al., 2019).
  • Molecular Property Prediction: In MoleVers, self-supervised masked atom prediction plus dynamic denoising (with a branching encoder) is followed by supervised adaptation to auxiliary quantum-calculated properties (Wijaya et al., 5 Nov 2024).
  • Parameter-Efficient Transfer: TTC-Tuning (Zhao et al., 2023) first tunes only LayerNorm parameters for distribution adaptation, then applies Taylor-based channel selection and trains only thin adapters on task-specific data.

These designs frequently involve:

  • Frozen and trainable module scheduling,
  • Curriculum over objectives or data,
  • Layer-wise or head-specific adaptation,
  • Use of auxiliary tasks or synthetic data, and
  • Constraint of per-stage compute or parameter count.

3. Objectives, Loss Functions, and Optimization Criteria

The two stages normally employ different, but complementary, optimization objectives:

Mathematically, the process is expressed as sequential minimization of stage-wise losses, with possible freezing and re-initialization of parameter subsets. For example, in parameter-efficient transfer: Stage 1:minγ,β(x,y)LCE(hbackbone(x;γ,β),y)\text{Stage 1:} \quad \min_{\gamma,\beta} \sum_{(x,y)} \mathcal{L}_{\text{CE}}(h_{\text{backbone}}(x;\gamma,\beta), y)

Stage 2:minΘTTC(x,y)LCE(TTC(hbackbone(x;γ,β)),y)\text{Stage 2:} \quad \min_{\Theta_{\text{TTC}}} \sum_{(x,y)} \mathcal{L}_{\text{CE}}(\text{TTC}(h_{\text{backbone}}(x;\gamma^*,\beta^*)), y)

where only γ,β\gamma,\beta (LayerNorm params) are updated in stage 1, and only the TTC module in stage 2.

Two-stage pretraining has shown robust gains across diverse settings, with empirical improvements documented in literature:

Application Area Stage 1 Stage 2 Empirical Gain Example Reference
Vision-Language CLIP LiT (frozen vision) Joint contrastive fine-tune +5–10 MR (retrieval), +10–20% cls accuracy (Yang et al., 2022)
Video Captioning Image–text pretrain Video–text adaptation +9.7 CIDEr vs. prior SOTAs (Chen et al., 2023)
ASR (multistream) UFE single-stream HAN stream attention fusion 8.2–32.4% WER reduction, ~50% fewer params (Li et al., 2019)
Molecular Prop. Pred. MAP+denoising (self-sup) Auxiliary DFT labels 20/22 SOTA in low-data real-world assays (Wijaya et al., 5 Nov 2024)
PETL for Vision (ViT) LayerNorm tuning Taylor-selected adapters +1.7 pp over Adapter, SOTA with <0.2M params (Zhao et al., 2023)

These gains are frequently attributed to improved sample complexity, accelerated convergence, increased robustness to distribution/domain shift, and reduced overfitting in resource-limited or highly specialized downstream regimes.

5. Limitations, Negative Transfer, and Design Caveats

Despite widespread success, two-stage pretraining is subject to several limitations and pitfalls:

  • Objective Inconsistency: When stage 1 and 2 losses operate in different domains (e.g., velocity L² fit vs. seismic waveform data fit in FWI), parameter updates exhibit negative transfer and loss of plasticity, causing the model to stagnate in local optima (Chen et al., 5 Jun 2025).
  • Overfitting to Stage 1 Biases: If the foundational model “locks in” suboptimal representations, stage 2 adaptation may be ineffective. This issue is pronounced in settings with a small number of fine-tuning samples or large objective dissimilarity.
  • Phase Scheduling: Over-extending the fine-tuning or phase 2 duration beyond 40–50% of training can degrade performance due to over-specialization and catastrophic forgetting (Feng et al., 18 Dec 2024).
  • Parameter Initialization and Freezing: Careful parameter freezing and initialization (block-wise head copying, partial reuse) is critical to avoid catastrophic drift, especially in transfer and causal learning contexts (Zhou et al., 15 Jan 2025).

Designing compatible objectives, regularization regimes, and staged optimization schedules are essential to realize the intended advantages of two-stage curricula.

6. Design Patterns, Best Practices, and Theoretical Insights

Established best practices and actionable recipes have emerged:

  • Curriculum over Data or Modality: Start with broad/diverse or easy objectives, transitioning to more specific, high-quality, or harder tasks as second-stage fine-tuning.
  • Quality-weighted Data Blends: In LLM pretraining, “two-phase” mixtures leverage quality scores to upsample high-Q sources in later phases, with empirical scaling validated up to 15T tokens and 25B parameters (Feng et al., 18 Dec 2024).
  • Partial Parameter Reuse: For CATE or parameter-efficient transfer, initialize fine-tuning heads by block-copying pretrained weights and incrementally expand only the necessary subnetworks (Zhou et al., 15 Jan 2025, Zhao et al., 2023).
  • Ablation and Sensitivity Analysis: Use ablations to determine where to insert adapters, which channels to fine-tune, optimal masking/noise rates, and phase ratios (typically 60:40 for phase1:phase2) (Zhao et al., 2023, Feng et al., 18 Dec 2024).
  • Robustness to Data Regime: Two-stage pretraining especially excels for regimes with scarce or expensive labeled data, as demonstrated by state-of-the-art downstream performance on “in the wild” molecular property assays with ≤50 samples (Wijaya et al., 5 Nov 2024).

Theoretical insights include empirical confirmation that feature-level distribution alignment (via normalization parameter tuning) substantially reduces Jensen–Shannon divergence between pretraining and downstream distributions (Zhao et al., 2023), and that parameter updates between mismatched objectives can exhibit cosine similarity near zero, flagging loss of plasticity and potential for negative transfer (Chen et al., 5 Jun 2025).

7. Extensions, Open Problems, and Future Directions

Current work raises several open research avenues:

  • Multi-Phase and Curriculum Learning: Extending beyond two phases, integrating adaptive or curriculum-based schedules, and managing phase transitions dynamically (Feng et al., 18 Dec 2024).
  • Interaction with Data Selection and Augmentation: Jointly optimizing two-stage pretraining with data selection methods (e.g., DoReMi, curriculum clustering) and augmentation strategies remains an open area.
  • Objective Harmonization: Designing more consistent or harmonizable objective pairs to avoid negative transfer, especially when moving between self-supervised and physically-constrained or task-specific losses (Chen et al., 5 Jun 2025).
  • Interpretability and Analytical Diagnostics: Probing the emergence and transfer of linguistic, structural, or semantic information across stages, leveraging ablations and case-studies in attention patterns (Kuo et al., 2023).
  • Scaling and Cross-Task Generalization: Empirically, two-phase data blending and representation learning consistently scale to larger model sizes, but systematic guidance for blend design and phase length in new domains remains incomplete (Feng et al., 18 Dec 2024).

Future research is likely to focus on formalization of multi-phase interaction, diagnostic tools for detecting and mitigating negative transfer, and development of general recipes for phase design and parameter initialization across domains.

Whiteboard

Follow Topic

Get notified by email when new papers are published related to Two-Stage Pretraining Process.