Papers
Topics
Authors
Recent
2000 character limit reached

SmolLM Family: Efficient Transformer Models

Updated 28 November 2025
  • SmolLM family is a collection of small- and mid-sized decoder-only transformer models designed for high efficiency in computationally constrained environments.
  • Its progressive low-rank decomposition method enables significant parameter compression with minimal retraining while preserving performance.
  • Data-centric pretraining and targeted post-training alignment empower these models to rival larger LMs in tasks, semantic novelty, and instruction-following.

The SmolLM family encompasses a spectrum of small- and mid-sized decoder-only transformer LLMs and their variants, developed for high efficiency and broad applicability in computationally constrained environments. SmolLM and its successors are distinguished by systematic model scaling (via progressive compression or data-centric pretraining), explicit paper of optimization dynamics, and empirical demonstrations that small models can approach or match the performance of larger LLMs when equipped with principled data curation and post-training alignment protocols. The family additionally serves as a reference suite in semantic novelty research.

1. Model Families, Architectures, and Parameterization

The canonical SmolLM family includes sub-2B parameter decoder-only (GPT-style) transformers. Three main sizes are frequently cited: SmolLM-135M (≈135M params), SmolLM-360M (≈360M), and SmolLM-1.7B (≈1.7B) (Vij et al., 4 Feb 2025). Each model instantiates a standard transformer LM stack: multiple decoder layers, each with multi-head self-attention and MLP subblocks, with embedding dimension dmodeld_\mathrm{model} and feedforward inner dimension 4dmodel4d_\mathrm{model}, commonly using approximate parameter allocations:

Variant #Layers Hidden Dim (dmodeld_\mathrm{model}) #Params
SmolLM-135M ≈12 ≈768 135M
SmolLM-360M ≈24 ≈1024 360M
SmolLM-1.7B ≈48 ≈2048 1.7B
SmolLM2-1.7B 24 2048 1.7B

SmolLM2-1.7B, a significant descendant, is architecturally aligned with Llama 2-1.7B: 24 layers, 2048 hidden size, 32 attention heads, rotary positional encoding (RoPE), SwiGLU activations, context length up to 2048 or (with extension) 8192 tokens (Allal et al., 4 Feb 2025). The models are pretrained with a causal language modeling objective: LLM=t=1Tlogpθ(wtw<t)\mathcal{L}_{LM} = -\sum_{t=1}^T \log p_\theta(w_t \mid w_{<t})

2. Model Construction and Compression via Progressive Low-Rank Decomposition

A defining characteristic of the SmolLM family is construction not only from scratch but via Progressive Low-Rank Decomposition (PLRD) (Hajimolahoseini et al., 28 Jun 2024). PLRD generates a fine-grained "spectrum" of model sizes from a single pretrained foundation model (e.g., Mistral-7B, Llama-2-7B), addressing deployment under diverse compute or memory budgets.

PLRD applies iterated truncated SVD to each dense matrix WRdin×doutW \in \mathbb{R}^{d_\mathrm{in} \times d_\mathrm{out}}:

  • Initialization: SVD factorizations yield W0=UΣ1/2W_0 = U\Sigma^{1/2} and W1=Σ1/2VW_1 = \Sigma^{1/2} V^\top, introducing a controllable inner rank Rmin(din,dout)R \ll \min(d_\mathrm{in}, d_\mathrm{out}).
  • Progression: The rank is reduced by a schedule Ri+1=αRiR_{i+1} = \lfloor \alpha R_i \rfloor, with α0.75\alpha\approx 0.75; factorization/fine-tuning steps repeat until target size is reached.
  • Minimal retraining: Each step is followed by limited fine-tuning (≈250M tokens/step, 1B tokens total), preserving accuracy with two orders of magnitude lower compute (≈0.1% of pretraining tokens).

Table of representative PLRD-generated family members:

Variant Backbone Final Rattn_{attn},Rmlp_{mlp} Params Train Tokens Zero-shot avg.
PLRD-Mistral-3.1B Mistral-7B (64,256) 3.1B 1B 45.6%
PLRD-LLaMA2-3.3B LLaMA-2-7B (64,256) 3.3B 1B 45.4%
Open-LLaMA-3B-v2 3.0B 1T 45.5%

PLRD yields performance on par with from-scratch training but using ≈1% or less of energy/resources. This enables deployment of near-arbitrary size variants, subject to the chosen rank schedule and available hardware constraints (Hajimolahoseini et al., 28 Jun 2024).

3. Training Regimes: Data-Centric Strategies and Stagewise Curation

SmolLM2 exemplifies a rigorous, data-centric pretraining protocol over a multi-stage process totaling ~11T tokens (Allal et al., 4 Feb 2025):

  • Stage 1 (0–6T): High-quality web (FineWeb-Edu 54%, DCLM 36%, StarCoderData 10%).
  • Stage 2 (6–8T): Increased code (20%) and nascent math augmentation.
  • Stage 3 (8–10T): Co-injection of Stack-Edu code and InfiMM-WebMath, rebalanced web mix.
  • Stage 4 (10–11T): "Upsampling" of curated math/code (FineMath4+, Stack-Edu), textbooks, and synthetic data.

Each stage includes manual data weight refinement based on interim benchmark feedback (e.g., upweighting FineMath4+ when GSM8K scores improved). Custom datasets such as FineMath (math reasoning, classified via high-threshold Llama scoring) and Stack-Edu (classifier-filtered code, 125B tokens across 15 languages) were introduced to overcome deficiencies in public corpora.

SmolTalk, the instruction-following corpus, aggregates 1.1M pairs with focus on constraint adherence, math, code, and multi-turn conversation. Ablations indicate that targeted upsampling and data curation critically improve downstream math and code benchmarks without sacrificing general language modeling performance (Allal et al., 4 Feb 2025).

4. Post-Training and Alignment: Optimization Dynamics in the Sub-2B Regime

SmolLM post-training and instruction tuning advances are embodied in SmolTulu, which adapts AllenAI TüLU 3’s multi-stage post-training to the SmolLM2-1.7B scale using three major alignment steps (Alrashed, 11 Dec 2024):

  1. Supervised Finetuning (SFT): On mixed instruction datasets (covering ARC, BBH, GSM8K, HellaSwag, IFEval, MMLU-Pro, PIQA).
  2. Direct Preference Optimization (DPO): On pairwise preference datasets, applying a KL-penalized, length-normalized DPO loss:

LDPO=E(x,yc,yr)D[logσ(βlogπθ(ycx)πref(ycx)βlogπθ(yrx)πref(yrx))]\mathcal{L}_{\text{DPO}} = \mathbb{E}_{(x,y_c,y_r)\sim\mathcal{D}} \left[\log \sigma\left(\beta \log \frac{\pi_\theta(y_c|x)}{\pi_\text{ref}(y_c|x)} - \beta \log \frac{\pi_\theta(y_r|x)}{\pi_\text{ref}(y_r|x)} \right)\right]

(β = 5, policy πref\pi_\text{ref} is the frozen SFT posteriors).

  1. Reward Modeling / RLVR (optional): Further refinement on DPO pairs and synthetic reward signals.

Crucially, SmolTulu introduces empirical guidance for the learning-rate-to-batch-size ratio R=η/BR = \eta / B. High RR (e.g., R0.5×107R \gtrsim 0.5 \times 10^{-7} for DPO) is found to be optimal for reasoning tasks at sub-2B scale. Comparative ablations demonstrate substantial gains: for SmolTulu DPO-1130 (high RR), IFEval = 67.7%, GSM8K = 51.6%, ARC = 51.5%; moderate-RR variants can further boost ARC and PIQA (Alrashed, 11 Dec 2024). These results challenge the "one-size-fits-all" scaling laws and highlight the need for task-dependent optimization ratio tuning in small LMs.

5. Empirical Characterization and Domain-Specific Evaluation

A variety of domain benchmarks and measurement protocols have been applied to SmolLM variants (Vij et al., 4 Feb 2025, Alrashed, 11 Dec 2024):

  • Standard NLG metrics: ROUGE, BLEU, PPL, and zero-shot/few-shot test suites (ARC, GSM8K, HellaSwag, MMLU, HumanEval, etc.).
  • Domain-specific metrics: In recipe generation, customized measures include ingredient coverage, step complexity, coherence, and LLM-as-judge rubrics targeting allergen safety and task relevance (Vij et al., 4 Feb 2025).
  • Empirical findings: SmolLM-360M and SmolLM-1.7B achieve highly similar downstream quality even with a ×5 parameter disparity; for structured NLG tasks, architectural efficiency and fine-tuning dominate scale.
Model ROUGE-1 BLEU-1 Ing. Cov. Step Comp. Allergen Safety
SmolLM-360M ft 0.11 0.07 0.16 0.98 2.57
SmolLM-1.7B ft 0.11 0.07 0.27 0.97 2.54
Phi-2 ft 0.17 0.11 0.30 0.99 2.44

Instruction-tuned variants demonstrate higher originality—reinforced by external evaluation via LLM-as-judge (Vij et al., 4 Feb 2025).

6. Semantic Novelty and Memorization Analysis

The notion of "semantic novelty," or un-attributability, has been rigorously quantified for the SmolLM family (Davydov et al., 31 Oct 2025). Novelty is measured as the absence of any semantically similar context in the pretraining corpus, operationalized via a two-stage retrieval pipeline:

  1. Retrieval: Pretraining corpus is chunked and indexed using GIST embeddings (L2-normalized, 512-token granularity). Fast retrieval is performed via FAISS cosine similarity.
  2. Reranking: ColBERTv2 computes contextual similarity between model outputs (split at varied chunk sizes) and the retrieved candidates. Per-chunk similarity is normalized against human references

N(k)=medianchunks(s~(q(k),Cq(k))μB)N^{(k)} = \text{median}_{\text{chunks}} \left( \frac{\widetilde{s}(q^{(k)},C_q^{(k)})}{\mu_B} \right )

N(k)<1N^{(k)}<1 implies model outputs are more novel than human-written text.

Results demonstrate that SmolLM2-Instruct models generate more novel output than their base counterparts (≈10–20pp across chunk sizes). 360M models are more novel than 1.7B, and instruction tuning increases semantic divergence in both open-domain and task-specific settings (GSM8K, TruthfulQA, OpenRewrite) (Davydov et al., 31 Oct 2025).

7. Limitations, Trade-offs, and Best Practices

The SmolLM family provides flexible solutions across compute regimes but involves several practical and theoretical caveats (Hajimolahoseini et al., 28 Jun 2024, Allal et al., 4 Feb 2025, Alrashed, 11 Dec 2024):

  • PLRD dependency: Requires a public, sufficiently large parent. Aggressive rank compression (R<32R<32) can result in unrecoverable loss.
  • Data curation overhead: Manual refinement for data-centric pipelines remains labor-intensive and costly (SmolLM2 pretraining >>\$250k).
  • Task adaptation: While small models approach larger ones on many tasks, outlier domains and extreme compression may still demand bespoke architectures or more extensive domain alignment.
  • Optimization regime: Task-dependent R=η/BR = \eta/B selection is critical at the sub-2B scale; benchmarks must include both reasoning and pattern recognition to avoid overfitting to a single regime.
  • Evaluation granularity: Both generic and domain-specific metrics are required; LLM-as-judge and semantic novelty scoring provide complementary assurances.

PLRD models can be further compressed (quantization/pruning), supporting deployment on a diverse hardware landscape. Reports systematically recommend explicit, task-verified tuning of optimization hyperparameters, and always benchmarking both memorization and generative novelty. Usage in NLG (recipe generation) and benchmarking suites demonstrates the robustness and competitive performance of SmolLM-family models at low compute and energy budgets, even under highly specialized constraints (Vij et al., 4 Feb 2025, Davydov et al., 31 Oct 2025).


In summary, the SmolLM family and its variants (notably SmolLM2 and SmolTulu) demonstrate that principled model compression, data-centric pretraining, and task-specific post-training alignment enable small LLMs to achieve competitive performance on diverse tasks, semantic novelty, and instruction-following, at a fraction of the computational and environmental cost typical of larger models (Hajimolahoseini et al., 28 Jun 2024, Alrashed, 11 Dec 2024, Allal et al., 4 Feb 2025, Davydov et al., 31 Oct 2025, Vij et al., 4 Feb 2025).

Slide Deck Streamline Icon: https://streamlinehq.com

Whiteboard

Forward Email Streamline Icon: https://streamlinehq.com

Follow Topic

Get notified by email when new papers are published related to SmolLM Family.