Papers
Topics
Authors
Recent
Search
2000 character limit reached

Masked Next-Token Prediction (MNTP)

Updated 19 November 2025
  • Masked next-token prediction is defined by fitting a sigmoidal curve to model surprisal curves, determining the 50% acquisition threshold as a key metric.
  • The methodology tracks word predictions across training steps, allowing robust comparisons between different LM architectures using metrics such as perplexity and KL divergence.
  • Empirical results reveal that frequency and local distributional statistics drive acquisition in LMs, contrasting with human language learning which relies more on conceptual grounding.

Neural Age of Acquisition (nAoA) is a quantitative metric designed to characterize when neural LMs acquire the ability to predict individual words during training. Introduced by Chang & Bergen (2021), nAoA operationalizes the step in training, on a logarithmic scale, at which a model's uncertainty about a word decreases to a specific threshold—analogue to developmental age-of-acquisition metrics in child language research. By tracking the evolution of model predictions across hundreds of words during training, nAoA enables rigorous comparative analysis between different LM architectures and human word learning dynamics, providing diagnostic insight into the mechanisms of lexical acquisition in artificial systems (Chang et al., 2021).

1. Formal Definition of Neural Age of Acquisition

nAoA is defined through the learning curve of a LLM’s surprisal for a given word ww at training step ss:

  • Surprisal is computed as S(w,s)=log2Ps(w)S(w,s) = -\log_2 P_s(w), where Ps(w)P_s(w) is the model’s average predicted probability of ww, evaluated over up to 512 held-out contexts.
  • A sigmoidal curve, S^(w,s)=L+UL1+exp[k(ss0)]\hat{S}(w,s) = L + \frac{U - L}{1 + \exp[-k(s - s_0)]}, is fit to the sequence of surprisals across training:
    • UU: Upper asymptote (surprisal under a uniform-chance prediction),
    • LL: Lower asymptote (minimal surprisal ever observed for ww),
    • kk: Slope parameter,
    • s0s_0: Midpoint of the sigmoid in training steps.
  • The acquisition cutoff surprisal is S=(U+L)/2S^* = (U + L) / 2, set analogously to the “50% acquisition” threshold in child studies.
  • Neural Age of Acquisition for word ww is nAoA(w)=log10snAoA(w) = \log_{10} s^*, with ss^* defined such that S^(w,s)=S\hat{S}(w, s^*) = S^*.

This provides a standardized, model- and context-agnostic scale for comparing acquisition timing of lexical items.

2. Computation and Modeling Workflow

To compute nAoA, Chang & Bergen (2021) utilized multiple LM architectures and a large, systematically chosen set of target words:

  • Models: Unidirectional LSTM (37M parameters), Bidirectional LSTM (51M), GPT-2-style Transformer (108M), and BERT base (109M), evaluated at perplexities 54.8, 9.0, 30.2, and 7.2, respectively.
  • Training Data: 25.6M sentence pairs (BookCorpus + WikiText-103), tokenized using a unigram-trained SentencePiece model (vocabulary ≈32,000 subword units, lowercased inputs).
  • Evaluation Regime: At ~200 checkpoints, more densely sampled earlier in training (100–1,000,000 steps), each target word ww was masked in held-out contexts and average surprisal S(w,s)S(w,s) was recorded.
  • Parameter Recovery: Sigmoid parameters {UU, LL, kk, s0s_0} were fit via least squares to each word’s learning curve.
  • Vocabulary Coverage: Targeted 651 English items from the MacArthur–Bates Communicative Development Inventory (CDI), with 611 mapped to single SentencePiece tokens.

This protocol allows direct mapping from raw training dynamics to interpretable acquisition profiles for substantial subsets of the lexicon.

3. Predictors of nAoA in Neural LLMs

Linear regressions were applied to predict nAoA for each word, using the following predictors: log-frequency, mean length of utterance (MLU), word length (number of characters), concreteness, and lexical class. Key findings include:

Predictor Effect in LMs Effect in Children Statistical Detail
Log-frequency Strong negative (βfreq<0\beta_{freq}<0, R2=0.910.94R^2=0.91-0.94) Weak (R2=0.01R^2=0.01) More frequent words learned earlier
MLU (context length) Positive (βMLU>0\beta_{MLU}>0; GPT-2, BiLSTM, BERT) Positive Words in longer utterances learned later (p<.001p<.001), unidirectional LSTM: n.s.
Word length Negative (βchars<0\beta_{chars}<0, p<.001p<.001) Positive (βchars>0\beta_{chars}>0, p<.001p<.001) LMs: Longer words learned earlier; Children: shorter words earlier
Concreteness No significant effect Negative (βconc<0\beta_{conc}<0, p<.001p<.001) Children acquire more concrete words earlier
Lexical class Nouns and function words later (uniLSTM, GPT-2), no effect (biLMs) Nouns < verbs < function words p<.01; ordering reversed in children

This analysis reveals pronounced reliance of LMs on frequency and local distributional statistics, diverging from human reliance on semantic or conceptual content.

4. Characteristic Training Dynamics

Detailed tracking of masked-token distributions via average KL divergence with respect to various reference distributions at each checkpoint illuminates the acquisition process:

  1. Early Training: The model’s output distribution for masked tokens collapses towards the corpus unigram distribution; the average KL divergence to the unigram baseline diminishes.
  2. Intermediate Training: Model predictions transition to approximating bigram statistics; KL divergence to bigram statistics drops after the unigram phase.
  3. Late Training: Divergence from unigram and bigram statistics increases, with model predictions converging on the full contextual, highly specific empirical distributions; KL divergence with respect to the one-hot “true” distribution decreases.

This sequential transition holds across all tested architectures, including both unidirectional and bidirectional models (Chang et al., 2021).

5. Interpretive Context: LLMs Versus Children

Comparisons with developmental child data reveal key contrasts:

  • LMs rely far more on frequency as a predictor of word acquisition (substantially higher R2R^2) than children.
  • The effect of mean utterance length is robust in both systems, with words in longer utterances being acquired later.
  • LMs show a reversed effect for word length and null effect for concreteness, contrary to child data—children acquire shorter and more concrete words earlier.
  • Lexical class effects in LMs diverge by architecture (strong noun delay in uniLSTM/GPT-2; null in biLMs), with opposite ordering to that observed in children.

These results underscore the importance of sensorimotor and social grounding in children, which is not captured by distributional learning from text alone.

6. Implications and Research Trajectories

The findings on nAoA carry broader implications for both the study and development of neural LLMs:

  • Distributional Mechanisms: LMs manifest a strong two-stage acquisition regime shaped by frequency and n-gram statistics, in contrast to the multifaceted grounding of child language learning.
  • Model Evaluation: nAoA profiles constitute a fine-grained diagnostic for pretraining regimes and architecture selection, supplementing aggregate metrics such as perplexity.
  • Toward Human-like LM Acquisition: Future research may explore integration of multimodal grounding (e.g., vision, action), interactive or curriculum-based learning, and alternative nAoA definitions (e.g., based on other surprisal quantiles or non-sigmoid fits).
  • Limitations: Current nAoA methodology is restricted to text-only training and does not inherently account for the conceptual or perceptual dimensions of word meaning.

This suggests that achieving human-like vocabulary acquisition in LMs may require explicit inclusion of sensorimotor grounding and functionally relevant learning objectives beyond next-token prediction (Chang et al., 2021).

Definition Search Book Streamline Icon: https://streamlinehq.com
References (1)

Topic to Video (Beta)

No one has generated a video about this topic yet.

Whiteboard

No one has generated a whiteboard explanation for this topic yet.

Follow Topic

Get notified by email when new papers are published related to Masked Next-Token Prediction (MNTP).