Papers
Topics
Authors
Recent
2000 character limit reached

Generative AI: Foundations & Impact

Updated 8 December 2025
  • Generative AI is a collection of machine learning techniques that create novel artifacts such as text, images, and audio by learning data distributions.
  • It employs architectures like transformers, VAEs, GANs, and diffusion models to model and sample high-dimensional data for various applications.
  • Its applications range from creative industries and autonomous systems to education, while addressing challenges in ethics, bias, and sustainability.

Generative artificial intelligence (GenAI) encompasses machine learning techniques capable of synthesizing new, meaningful artifacts—such as text, images, audio, code, or 3D objects—that are statistically similar to data observed during training but novel in their composition. GenAI systems are typically instantiated as deep neural networks that learn and model the joint or marginal distribution of high-dimensional data, allowing them to sample and generate new content rather than merely classify, regress, or retrieve. Foundation models—large-scale, pre-trained architectures like transformers—underpin much of contemporary GenAI, supporting a wide array of modalities and downstream applications across research, industry, education, and creative sectors (Feuerriegel et al., 2023, Jauhiainen et al., 22 Aug 2025, Storey et al., 25 Feb 2025).

1. Technical Foundations: Model Families and Mathematical Objectives

GenAI models are unified by the principle of learning a data-generating process, operationalized via several deep generative model families:

  • Autoregressive Models: Factorize the joint probability of a sequence into conditionals, enabling recursive sampling:

p(x1:T)=t=1Tp(xtx1:t1)p(x_{1:T}) = \prod_{t=1}^T p(x_t \mid x_{1:t-1})

The dominant architecture is the transformer, where next-token prediction is trained by maximum-likelihood estimation (MLE) (Tewari, 7 Sep 2025, Feuerriegel et al., 2023).

LVAE(θ,ϕ)=Eqϕ(zx)[logpθ(xz)]DKL(qϕ(zx)p(z))\mathcal{L}_{\text{VAE}}(\theta, \phi) = \mathbb{E}_{q_\phi(z\mid x)}[\log p_\theta(x\mid z)] - D_{\mathrm{KL}}(q_\phi(z\mid x) \| p(z))

(Feuerriegel et al., 2023, Tewari, 7 Sep 2025).

  • Normalizing Flows: Learn invertible mappings x=fθ(z)x = f_\theta(z) such that the density of xx can be exactly computed via the change-of-variables formula (Tewari, 7 Sep 2025).
  • Generative Adversarial Networks (GANs): Formulate a minimax game between a generator GG and a discriminator DD:

minGmaxDExpdata[logD(x)]+Ezpz[log(1D(G(z)))]\min_G \max_D\, \mathbb{E}_{x \sim p_{\rm data}}[\log D(x)] + \mathbb{E}_{z \sim p_z}[\log(1 - D(G(z)))]

(Feuerriegel et al., 2023, Ning et al., 5 Nov 2025).

  • Diffusion Models: Model a forward noising process and a learned reverse denoising process:

q(xtxt1)=N(xt;1βtxt1,βtI)q(x_t \mid x_{t-1}) = \mathcal{N}(x_t; \sqrt{1-\beta_t}x_{t-1}, \beta_t I)

pθ(xt1xt)=N(xt1;μθ(xt,t),Σθ(xt,t))p_\theta(x_{t-1}\mid x_t) = \mathcal{N}(x_{t-1}; \mu_\theta(x_t,t), \Sigma_\theta(x_t,t))

Denoising is trained via a weighted MSE loss (Ning et al., 5 Nov 2025, Feuerriegel et al., 2023, Tewari, 7 Sep 2025).

2. Architectures and Emergent Capabilities

The transformer architecture, introduced by Vaswani et al., is foundational to modern GenAI due to its scalability, parallelizable self-attention, and capacity to model long-range dependencies. Multi-head self-attention computes:

Attention(Q,K,V)=softmax(QKdk)V\mathrm{Attention}(Q, K, V) = \mathrm{softmax}\left(\frac{QK^\top}{\sqrt{d_k}}\right)V

Embedding layers, residual connections, and layer normalization allow depth and expressivity. Pretrained "foundation models" such as GPT-4/5, BERT, DALL·E-2, and diffusion backbones (e.g., Stable Diffusion) encode broad structural priors, later adapted or prompted for specific generative tasks through supervised fine-tuning, instruction tuning, and RLHF (Jauhiainen et al., 22 Aug 2025, Storey et al., 25 Feb 2025, Ning et al., 5 Nov 2025).

GenAI systems demonstrate emergent behaviors, including:

  • Few-shot and zero-shot generalization
  • Autoregressive coherence over large context windows
  • Multimodal composition (text, image, audio, code, 3D mesh)
  • Agentic workflow orchestration (planning, tool-use, and reasoning) (Acharya, 12 Oct 2025, Jauhiainen et al., 22 Aug 2025)

3. Domains and Application Patterns

GenAI applications span a broad array of technical and domain settings. Representative categories include:

4. Performance, Evaluation, and System Integration

Model assessment uses both objective and subjective measures:

Model optimization and deployment emphasize model compression, prompt tuning, retrieval-augmented generation (RAG), and hardware-aware search for efficient inference in edge and real-time scenarios (Ning et al., 5 Nov 2025, Stappen et al., 2023).

5. Systemic, Ethical, and Socio-Technical Considerations

As GenAI systems permeate real-world pipelines, several critical issues emerge:

  • Hallucinations & Reliability: Outputs can be plausible but factually inaccurate, raising misinformation risks (Feuerriegel et al., 2023).
  • Bias & Fairness: Training data often reflect and amplify societal and demographic biases; evaluation employs demographic parity, equal opportunity, and false-positive rate difference metrics. Real-world harms can include under-representation and inequitable access (Healy, 2023, Ning et al., 5 Nov 2025).
  • Intellectual Property and Attribution: Models may reproduce or remix protected data; watermarks and provenance mechanisms are deployed, but legal regimes remain unsettled (Feuerriegel et al., 2023, Tewari, 7 Sep 2025).
  • Environmental and Economic Impact: Training large models consumes significant energy (e.g., GPT-3 emits ~552 t CO2_2 for pre-training); cost and accessibility are major determinants of adoption (Feuerriegel et al., 2023, Acharya, 12 Oct 2025).
  • Human-AI Agency: Blurred boundaries between human learners/workers and AI systems, especially in education, raise questions about authenticity and responsibility (Yan et al., 2023, Healy, 2023).
  • Governance and Accountability: Calls for explainability, audit trail frameworks, regulatory approaches tailored to domain risks (e.g., XR-specific legal paradigms and compliance) (Ning et al., 5 Nov 2025).

6. Future Directions and Open Research Problems

Current research agendas and open issues delineate prospective advances:

7. Integrated Overview and Socio-Technical Impact

GenAI represents a confluence of advances in connectionist neural architectures, large-scale training, and algorithmic innovation in density modeling, adversarial learning, and multimodal fusion. It has transformed the information ecosystem, enabling scalable co-creation—humans and AI iteratively produce, refine, and curate artifacts in a feedback loop. Organizational structures, educational workflows, and creative industries are fundamentally reshaped as GenAI tools become integrated across research, commerce, design, and daily life.

The socio-technical ecosystem is characterized by its layered configuration: core model development, infrastructural deployment, human–machine interface, application embedding, and regulatory oversight. Emergent properties—such as trust, agency allocation, transparency, and unexpected behaviors—arise only in the context of these full systems (Storey et al., 25 Feb 2025).

Achieving responsible, inclusive, and sustainable deployment of generative artificial intelligence demands advances not only in technical sophistication, but also in theory-driven modeling of human–AI interaction, rigorous evaluation under uncertainty, equitable access, and robust governance. Contemporary research frames GenAI not as a mere algorithmic substrate, but as a central agent in co-creative, adaptive, and ethically contested socio-technical networks (Jauhiainen et al., 22 Aug 2025, Feuerriegel et al., 2023, Ning et al., 5 Nov 2025).

Slide Deck Streamline Icon: https://streamlinehq.com

Whiteboard

Forward Email Streamline Icon: https://streamlinehq.com

Follow Topic

Get notified by email when new papers are published related to Generative Artificial Intelligence.