Papers
Topics
Authors
Recent
2000 character limit reached

Generative Pre-Training

Updated 4 December 2025
  • Generative pre-training is a self-supervised learning paradigm that models massive unlabeled data to learn hierarchical representations for various downstream tasks.
  • It employs architectures such as autoregressive Transformers, denoising autoencoders, and flow-based models to capture joint, marginal, or conditional data distributions.
  • Its adaptability across modalities—spanning language, vision, speech, graphs, and molecules—enables robust performance in tasks like classification, reasoning, and generation.

Generative pre-training is a class of self-supervised representation learning techniques in which a model—typically based on an autoregressive Transformer, denoising autoencoder, or flow-based architecture—is first trained to model the generative structure of massive unlabeled data via explicit or implicit likelihood maximization. Once pre-training is complete, the parameters can be adapted, directly or via light fine-tuning, to specialized downstream tasks (classification, prediction, reasoning, editing, retrieval, and more). This paradigm, initially dominant in language modeling, now permeates vision, speech, graphs, code, molecules, document analysis, and multimodal domains. The following sections outline foundational principles, model architectures, diverse domain applications, empirical advances, and current limitations.

1. Fundamental Principles and Objectives

The defining feature of generative pre-training is large-scale self-supervised learning by fitting the joint, marginal, or conditional distribution of input data. The most prevalent objectives are:

  • Autoregressive Language Modeling: The model learns P(x1,...,xT)=t=1TP(xtx<t)P(x_1, ..., x_T) = \prod_{t=1}^T P(x_t | x_{<t}) by minimizing the negative log-likelihood or cross-entropy loss over tokenized data. This setup underpins GPT, PhoGPT, and many molecular and vision-text models (Nguyen et al., 2023, Liu et al., 2023, Zhu et al., 2023).
  • Masked Modeling: Bidirectional, denoising autoencoder objectives, where inputs are randomly masked and the model predicts masked tokens conditioned on the visible context (as in GPTFace's MILM and span-masked code objectives) (Li et al., 21 Oct 2025).
  • Explicit Generation of Structured Data: Pre-training on the generation of graphs, 3D renders, scene layouts, or molecular sequences establishes priors over spatial or topological structures (Cui et al., 2023, Wang et al., 2023, Xie et al., 24 Apr 2024, Xie et al., 2023).
  • Flow-based Matching and Diffusion: Certain domains (e.g., speech, molecular graphs) apply flow-matching objectives, training continuous-time invertible mappings between noise and data (Liu et al., 2023, Pan et al., 2023).

The overarching aim is to obtain parameterizations capable of capturing hierarchical, compositional, and cross-modal dependencies—thereby encoding rich representations transferable to a wide spectrum of downstream tasks.

2. Model Architectures and Data Modalities

While the foundational architecture is a multi-layer Transformer, the specifics are extended to match data domain intricacies:

  • Unimodal LLMs: Standard decoder-only (GPT, PhoGPT) or encoder-decoder (BART, T5) Transformers with large token and context sizes model natural language (Nguyen et al., 2023).
  • Multimodal Extensions: VL-GPT and ERNIE-ViLG handle both images and text by hybridizing continuous (visual embedding) and discrete token streams, employing image tokenizers (ViT, VQ-VAE), detokenizers (diffusion decoders), and specialized cross-modal attention blocks (Zhu et al., 2023, Zhang et al., 2021).
  • Graph-based Models: Generative GNNs (GPT-GNN) factorize graph generation into node attribute and edge structure prediction, often using autoregressive or SE(3)-equivariant models (Hu et al., 2020).
  • Code and Structured Data Models: Seq2seq Transformers are adapted via tailored input-output corruption (e.g., NatGen's "naturalizing" by applying semantic-preserving code transformations and training the model to reconstruct idiomatic original forms) (Chakraborty et al., 2022).
  • Tokenization and Discretization Strategies: Many domains implement explicit tokenization. Visual data (VisorGPT, Take-A-Photo, Learning Long-form Video Prior) discretize bounding box/pose/mask coordinates; molecular models tokenize SMILES with regex-based or subword strategies; document intelligence models interleave text and quantized spatial location tokens (Xie et al., 2023, Wang et al., 2023, Xie et al., 24 Apr 2024, Mao et al., 25 Mar 2024, Liu et al., 2023).

Architectural variants often introduce hierarchical decoders, cross-modal fusion layers, span-masking modules, or hybrid pipelines (autoencoder+autoregressive).

3. Domain-Specific Implementations

Natural Language and Multilinguality

Monolingual models such as PhoGPT (Nguyen et al., 2023) conduct generative pre-training over massive tokenized corpora (102B tokens), using scalable architectures (32 decoder blocks, 3.7B parameters, 8192 context). Pre-training is strictly autoregressive and adapted to linguistic specifics via custom BPE tokenization.

Molecules and Chemistry

MolXPT (Liu et al., 2023) entwines scientific text and molecular SMILES in a unified decoder-only Transformer, pre-training on pure text (30M PubMed abstracts), pure SMILES (30M), and "wrapped" mixed sequences (8M), with molecule names replaced by their SMILES. The absence of explicit modality-type embeddings and cross-modal losses enables seamless bi-directional information flow.

Vision, Layout, and 3D/Video

  • VisorGPT (Xie et al., 2023) models visual prior (object/pose/layout distributions) by tokenizing spatial coordinates, employing prompt engineering to control generative outputs in image synthesis and scene layout tasks.
  • VL-GPT (Zhu et al., 2023) and ERNIE-ViLG (Zhang et al., 2021) enable seamless joint modeling and conditional generation across image and language via unified token sequences and joint objectives.
  • Video and 3D Models: Take-A-Photo (Wang et al., 2023) and Learning Long-form Video Prior (Xie et al., 24 Apr 2024) tokenize rendered 3D or video content (bounding boxes, keypoints) for autoregressive Transformer modeling, incorporating special position embeddings, and leveraging datasets with dense spatio-temporal annotation.

Graphs and Scientific Data

  • GPT-GNN (Hu et al., 2020) pre-trains GNN encoders for attributed graphs using autoregressive generation of node attributes and masked-edges, enhancing transfer learning for node classification and link prediction.
  • EG³P (Cui et al., 2023) bridges text and explanation-graph generation for reasoning over synthetic knowledge graphs, employing large-scale synthetic corpora and maximum-likelihood training.

Document Intelligence and OCR

ViTLP (Mao et al., 25 Mar 2024) designs a generative pre-training regime that interleaves language tokens with explicit spatial location markers, employing a hierarchical decode (global→local: [LOC] tokens and their bounding-box coordinates) and multi-segment strategies for arbitrarily long documents.

Music and Speech

  • Jukebox-powered Melody Transcription: Features learned via generative pre-training (hierarchical VQ-VAE + Transformer) encode musical structure; these features drive downstream melody transcription surpassing classic EM/Turing-test-based baselines (Donahue et al., 2022).
  • SpeechFlow (Liu et al., 2023) pre-trains a flow-based model with masked-audio conditions on large raw speech corpora, enabling versatile adaptation to enhancement, separation, and synthesis.

4. Empirical Performance and Evaluation

Empirical validation consistently demonstrates significant transfer and sample efficiency advantages:

Domain/Model Downstream Tasks Notable Results
PhoGPT (Nguyen et al., 2023) ViTruthfulQA, instruction Outperforms GPT-3.5-turbo on Vietnamese
MolXPT (Liu et al., 2023) MoleculeNet, CheBI-20 ROC-AUC 81.9%, best Text2Mol (0.578 vs 0.554)
GPT-GNN (Hu et al., 2020) Node/link, multi-domain +9.1% vs. state-of-art on OAG; robust transfer
ViTLP (Mao et al., 25 Mar 2024) OCR, DocVQA, classification >95% recog. F1; matches discriminative baselines
VisorGPT (Xie et al., 2023) Conditional image synthesis High layout prior fidelity (KL divergence)
Jukebox (Donahue et al., 2022) Melody transcription +20% over spectrogram features, 0.744 F1
SpeechFlow (Liu et al., 2023) Enhancement, separation, TTS Surpasses expert-specific models
BootRet (Tang et al., 16 Jul 2024) Generative retrieval MRR@20=42.79 (MS MARCO), better than NOVO

The gains derive from three attributes: (i) structured pre-training objectives that align with downstream inference modes, (ii) exploitation of extremely large unlabeled datasets (often billions of tokens), and (iii) architectural modifications for domain specifics (e.g., masking, span-based heads, scene-level masks).

5. Methodological Innovations and Design Choices

  • Cross-Modal and Multimodal Fusion: Wrapping or interleaving tokens (MolXPT, VL-GPT) achieves implicit alignment without explicit auxiliary objectives; others (GPTFace) use explicit image-text matching (ITM) losses (Li et al., 21 Oct 2025, Zhu et al., 2023).
  • Tokenization Granularity and Discretization: Granular BPE for language, regex tokenization for SMILES, m=512 bins for 2D/3D coordinates, PQ codes for document identifiers (BootRet) (Liu et al., 2023, Xie et al., 2023, Tang et al., 16 Jul 2024).
  • Prompt Engineering and Conditional Generation: Unified prompt templates for controlling output structure (VisorGPT), in-context learning (VL-GPT), and instruction tuning extend adaptability to custom tasks.
  • Synthetic Data and Corpus Augmentation: EG³P and BootRet construct synthetic graphs, queries, and document variants via LLMs or controlled generation, building large-scale pseudo-labeled pre-training corpora (Cui et al., 2023, Tang et al., 16 Jul 2024).
  • Hierarchical and Multi-Segment Decoding: Many models (ViTLP, GPD-1) decouple global and local decoding or segment long sequences to process arbitrarily sized contexts (Mao et al., 25 Mar 2024, Xie et al., 11 Dec 2024).

6. Limitations, Open Challenges, and Prospects

Limitations and areas for future research include:

  • Scalability and Model Size: Many published models remain in the 100M–1B parameter regime; scaling to multi-billion parameters (language, multimodal, graph) is ongoing and may alter in-context learning and cross-modal alignment (Nguyen et al., 2023).
  • Alignment Across Modalities: Some models (MolXPT) do not employ explicit cross-modal alignment objectives; future work may incorporate contrastive or co-attention modules (Liu et al., 2023).
  • Synthetic Data Bias: Synthetic corpus construction may introduce distributional biases, potentially limiting real-world generalization (EG³P, BootRet) (Cui et al., 2023, Tang et al., 16 Jul 2024).
  • Editability and Controllability: Controllable generation and adaptive sampling (e.g., with ITM gradient guidance or prompt engineering) are effective but may require high compute or specialized sampling strategies (Li et al., 21 Oct 2025).
  • Domain Adaptation and Diversity: Many generative pre-training models remain confined to single domains or languages; broad generalization to new domains, scripts, or modalities requires further research (Liu et al., 2023, Chakraborty et al., 2022).
  • Evaluation Metrics and Application Breadth: Standard metrics may hide nuances in structure, creativity, and semantic fidelity, particularly for structured outputs (graphs, code, long-form video), necessitating richer evaluation (Cui et al., 2023, Donahue et al., 2022).

7. Significance and Impact

Generative pre-training serves as the cornerstone for current and emerging foundation models across modalities. By shifting the learning paradigm from task-labeled datasets to scalable, generative modeling of unannotated data, it has enabled the rise of broadly capable, transferable, and adaptable architectures. Empirical results show that such models can match or surpass purpose-built task-specific systems, especially where downstream data is limited or highly variable. The generality of the generative objectives—when paired with innovations in data wrangling, architecture, and training—promises continued advances in performance, generalization, and task breadth in machine learning, AI, and scientific discovery.

Slide Deck Streamline Icon: https://streamlinehq.com

Whiteboard

Forward Email Streamline Icon: https://streamlinehq.com

Follow Topic

Get notified by email when new papers are published related to Generative Pre-Training.