Papers
Topics
Authors
Recent
Search
2000 character limit reached

Private Mask Pre-Training (PMP)

Updated 7 February 2026
  • PMP is a framework that embeds a hidden sparse binary mask during pre-training to restrict unauthorized fine-tuning in foundation models.
  • It uses an early-bird lottery ticket algorithm to identify and stabilize the mask, ensuring only the optimized subnetwork is updated in subsequent training.
  • Empirical findings show that PMP maintains base performance while reducing adaptation gains from unauthorized fine-tuning by up to 20 points.

Private Mask Pre-Training (PMP) is a pre-training framework developed to embed intrinsic barriers against unauthorized downstream fine-tuning in open-sourced foundation models. PMP achieves this by identifying and privatizing a sparse subnetwork during pre-training via a binary mask, which is hidden prior to model release. As a result, unauthorized fine-tuning—absent knowledge of the mask—provably yields limited gains and is destabilized by a geometrical mismatch between the pre-training and adaptation subspaces. The PMP framework is architecture-agnostic and preserves base model usability, while granting model owners the ability to retain adaptation control without requiring architectural or policy obfuscation (Wang et al., 31 Jan 2026).

1. Formalization and Core Principles

Let W∈RdW\in\mathbb{R}^d denote the dense parameter vector of a foundation model. During PMP, a sparse binary mask m∈{0,1}dm\in\{0,1\}^d is selected, with sparsity ratio pp (∥m∥0=p d\|m\|_0 = p\,d), partitioning WW into Wm=W⊙mW_m=W\odot m (the active "ticket") and Wmˉ=W⊙(1−m)W_{\bar m}=W\odot(1-m) (the inactive complement). Pre-training proceeds such that, after a brief warm-up, all learning updates are confined to WmW_m, with WmˉW_{\bar m} frozen at initial values. The final released model is W∗=Wm∗+Wmˉ0W^* = W_m^* + W_{\bar m}^0, and only the dense weights are published—mm remains secret.

The central PMP principle is that unauthorized fine-tuning without access to mm must indiscriminately update both WmW_m and WmˉW_{\bar m}. Since only WmW_m was optimized during pre-training, downstream SGD in the orthogonal frozen subspace encounters a loss surface with high curvature and misaligned gradients, inducing instability and bounding any net fine-tuning gain.

2. Mask Identification Procedure

PMP adopts an early-bird lottery ticket-based algorithm to derive the private mask mm. The procedure involves a warm-up phase of TEBT_{\rm EB} steps (typically 500) on a pre-training subset Dpre\mathcal{D}_{\rm pre}. At each step, the absolute parameter gradients g(t)g^{(t)} are computed. The top-kk (k=⌊p d⌋k = \lfloor p\,d\rfloor) entries by magnitude define candidate support m(t)m^{(t)}. Mask stability is assessed using intersection-over-union (IoU) thresholds and a hit counter. Once the candidate stabilizes for Δ\Delta consecutive steps (e.g., τ=0.99\tau=0.99, Δ=5\Delta=5), mm is fixed. The main pre-training then proceeds, updating only WmW_m.

The process is summarized in the following table:

Step Operation Hyperparameters
Gradient evaluation Compute g(t)=∣∇Wℓ(f(W(t));B(t))∣g^{(t)}=|\nabla_W \ell(f(W^{(t)});B^{(t)})|
Top-K mask assignment mi(t)=1m^{(t)}_i = 1 if gi(t)≥γg^{(t)}_i \geq \gamma else $0$ k=⌊p d⌋k=\lfloor p\,d\rfloor
Mask stabilization IoU(m(t),m(t−1))≥τ(m^{(t)}, m^{(t-1)}) \geq \tau ⇒\Rightarrow increase hits τ=0.99\tau=0.99, Δ=5\Delta=5

After mask selection, all subsequent training is strictly in the masked subspace.

3. Theoretical Analysis of Fine-Tuning Instability

PMP provides formal guarantees that unauthorized fine-tuning is globally destabilized by the mask secrecy. The analysis rests on a local geometrical assumption: in the vicinity of the trained parameters, the loss landscape along WmW_m is flat (null Hessian), while along WmˉW_{\bar m}, the Hessian is strictly positive definite—representing steep directions untouched during pre-training.

Let the downstream objective be Lft(Wm,Wmˉ)=Lpre(Wm,Wmˉ)+Δ(Wm,Wmˉ)L_{\rm ft}(W_m, W_{\bar m}) = L_{\rm pre}(W_m, W_{\bar m}) + \Delta(W_m, W_{\bar m}), where Δ\Delta encodes task shift. For any standard SGD step with step-size η\eta not restricted to WmW_m: $\E[L_{\rm pre}(W_m', W_{\bar m}')] \geq L_{\rm pre}(W_m^*, W_{\bar m}^0) + c\eta^2,$ for some c>0c>0, so long as the gradient in the frozen subspace is non-zero with positive probability. Thus, arbitrary adaptation not guided by mm systematically increases the pre-training loss, bounding downstream gains over many steps. The proof adapts second-order Taylor expansion to show that the quadratic curvature along WmˉW_{\bar m} dominates.

4. Implementation Aspects

The PMP pipeline comprises three principal stages:

  1. Mask Discovery: Early-bird mask selection (500 steps).
  2. Sparse-Subspace Pre-Training: Training proceeds with gradients confined to WmW_m using standard optimizers (AdamW, cosine lr decay, batch size 4×84\times8, gradient norm clipping).
  3. Release: Only W∗W^* is released, with mm withheld.

Experiments adopt TinyLlama-1.1B (22 layers, hidden size 2048, 32 heads) and GPT-2 architectures, pre-trained on SlimPajama-6B (tokenized to 256 tokens, causal LM loss). Authorized fine-tuning can be enabled for select users by providing mm such that only WmW_m updates. Storage and release do not expose mm. Empirically, adversaries are unable to reconstruct mm from observed gradients or outputs, as gradient-magnitude distributions between masked and unmasked weights overlap.

5. Empirical Findings

PMP is evaluated on base (zero-shot, pre-finetuning) capabilities and unauthorized fine-tuning across GLUE tasks (CoLA, SST-2, MRPC, QQP, STS-B, MNLI, QNLI, RTE). The experimental protocol holds all training and adaptation hyperparameters constant across non-PMP and PMP settings.

Key outcomes:

  • Base performance is unchanged with PMP (e.g., TinyLlama: ∼50%\sim50\% w/o PMP vs. ∼51%\sim51\% w/ PMP).
  • Unauthorized fine-tuning accuracy is greatly reduced (TinyLlama: 76%76\% w/o PMP vs. 46%46\% w/ PMP).
  • Varying the mask ratio pp modulates the barrier: base accuracy is stable, but unauthorized fine-tuning drops from ∼87%\sim87\% at p=1p=1 to ∼60%\sim60\% at p=0.7p=0.7.
  • Across a grid of learning rates and epochs, PMP consistently suppresses adaptation gains by ∼20\sim20 points, while non-PMP models are robust.
  • Authorized fine-tuning (with mm) recovers high post-adaptation accuracy (∼80%\sim80\% on GLUE).

6. Discussion and Open Problems

PMP establishes non-fine-tunability by imposing a pre-training-level geometry mismatch, graphically supported by loss-landscape sweeps indicating flat valleys along WmW_m and steep walls in orthogonal directions. The mask ratio pp operates as a tunable control on the model's resilience to adaptation, trading off adaptation difficulty for training speed.

Empirical findings suggest that the mask cannot be reliably inferred from black-box queries, due to the statistical overlap of gradient magnitudes between masked and unmasked weights.

Several limitations and open directions are recognized:

  • Theoretical guarantees focus on single-step adaptation; multi-step SGD dynamics, especially for sophisticated adversaries, warrant deeper investigation.
  • The secrecy of mm is central; partial leakage or white-box access could present vulnerabilities, necessitating future robustness analyses.
  • Current experiments focus on GLUE; effectiveness on instruction-tuning, multimodal, or adversarial tasks remains to be evaluated.
  • Interactions with other release strategies, such as quantization or differential privacy, are unexplored.

PMP provides an architecture-agnostic, low-overhead method to regulate foundation model adaptation post-release, preserving base utility and authorized fine-tuning while bounding unauthorized reuse through theoretically and empirically supported mechanisms (Wang et al., 31 Jan 2026).

Definition Search Book Streamline Icon: https://streamlinehq.com
References (1)

Topic to Video (Beta)

No one has generated a video about this topic yet.

Whiteboard

No one has generated a whiteboard explanation for this topic yet.

Follow Topic

Get notified by email when new papers are published related to Private Mask Pre-Training (PMP).