Papers
Topics
Authors
Recent
Search
2000 character limit reached

Dynamic Generative Memory (DGM)

Updated 7 April 2026
  • Dynamic Generative Memory is a framework that combines generative models with dynamic memory operations to support continual learning and adaptive task performance.
  • It employs mechanisms such as Bayesian updates, attractor dynamics, and replay-driven rehearsal to enhance robust recall and prevent catastrophic forgetting.
  • Implementations like DM-GAN, MemGen, and diffusion-based techniques validate DGM's efficacy in diverse applications, from text-to-image synthesis to dynamic video modeling.

Dynamic Generative Memory (DGM) encompasses a family of frameworks and architectures that integrate generative modeling and dynamic memory operations to enable learning, retrieval, and creative synthesis across continual and dynamic environments. DGM is instantiated in various forms across deep generative modeling, continual learning, diffusion models, memory-augmented neural architectures, and LLM-based agents—unifying mechanisms from distributed associative memory with end-to-end differentiable learning. DGM enables models to store, retrieve, and dynamically update representations or patterns, supporting robust prediction, generative replay, and task-adaptive expansion beyond traditional static memory or feed-forward paradigms.

1. Core Principles and Definitions

DGM systems fuse the mechanisms of generative models (e.g., GANs, VAEs, diffusion, autoregressive LMs) with dynamic, addressable memory structures that may be explicit (matrices, buffers, memory tokens) or implicit (distributed attractors in parameter space). In all cases, memory is not static: it is actively written to, reorganized, or refreshed as new data, tasks, or reasoning steps unfold. This dynamic behavior underpins:

Dynamicity is realized via mechanisms such as Bayesian online updates, replay-driven rehearsal, synaptic-plasticity masking, episodic clustering, reward-gated memory triggering, attention over latent tokens, and attractor dynamics.

2. Canonical DGM Architectures and Mechanisms

Representative DGM instantiations differ in their memory organization, generative process, and memory update rules. The following table summarizes key forms:

DGM Variant Memory Representation Generative Process
Conditional GAN + Replay External generator + buffer Sample pseudo-inputs; mix replay & real for training
Variational Kanerva Machine Matrix-normal global memory Bayesian update, distributed addressing, VAE-style
Attractor-based (DKM) Latent matrix + dynamic addresses Iterative coordinate-descent; Lyapunov energy
Diffusion Score Network Implicit attractors in weights Denoising SDE/ODE, unified retrieval/generation
Tokenized Video/Latents Spatiotemporal tokens; attention Attentional read/write, relevance-driven selection
LLM Latent Memory (MemGen) LoRA adapters, latent input tokens Triggered active weaving of token memories

In continual learning, DGM is often implemented as a deep generator (GAN or VAE) that, after encountering each new task, generates pseudo-samples corresponding to prior tasks for rehearsal. Memory expansion is driven by synaptic masks that count parameter saturation, instantiating plasticity-inspired dynamic growth (Ostapenko et al., 2019, Wu et al., 2018). In distributed memory (Kanerva Machine, DKM), memory is a latent matrix updated analytically via Bayes rule, and retrieval proceeds by coordinate descent on an energy/Lyapunov function (Wu et al., 2018, Wu et al., 2018). In modern agentic/LLM settings, memory is either tokenized and inserted as latent context via specialized modules or deployed conditioned on metacognitive trigger signals (Zhang et al., 29 Sep 2025).

3. Continual Learning: Dynamic Generative Replay

Catastrophic forgetting in continual learning is addressed by DGM through generative replay and memory-augmented solvers (Shin et al., 2017, Ostapenko et al., 2019, Rios et al., 2018). In this paradigm:

  • A task solver network is coupled with a deep generative model (GAN, AC-GAN, VAE) constituting the memory module.
  • At each new task, pseudo-examples are generated conditioned on all previous tasks' classes (via old generator), and replayed together with current real data to update both generator and solver.
  • Memory units (episodic buffers or latent cluster selectors) store a minimal set of real patterns for stabilizing the generator and anchor diversity.
  • Dynamic network expansion, governed by binary plasticity masks, allows the architecture to grow when capacity thresholds are surpassed (Ostapenko et al., 2019).
  • Losses incorporate adversarial, task, and mask sparsity components; empirical results consistently show superior retention and lower forgetting than regularization-based approaches (EWC, SI, HAT), even with fixed-size memory budgets (Shin et al., 2017, Rios et al., 2018).

4. Distributed and Attractor-Based DGM

The Kanerva Machine (Wu et al., 2018) and the Dynamic Kanerva Machine (Wu et al., 2018) realize DGM as analytically tractable, distributed memories:

  • Memory is a global matrix MM with a matrix-normal prior. Observations are read and written via soft addressing vectors produced from latent keys.
  • Writing corresponds to Bayesian updating of MM given observation codes; reading is distributed (softmax-based), enabling smooth interpolation and high capacity.
  • DKM introduces attractor dynamics at retrieval: coordinate-descent alternates between recomputing optimal address weights and reconstructing the observation, leading to convergence at stored patterns.
  • A Lyapunov function derived from the ELBO ensures stability of attractor dynamics; at training time, no backpropagation is performed through retrieval steps, preventing vanishing gradients.
  • These models demonstrate enhanced capacity and generalization compared to slot-based (DNC) or sequential replay-based architectures, and fast online adaptation via analytic updates (Wu et al., 2018, Wu et al., 2018).

5. Diffusion Models and Associative DGM

Score-based diffusion models constitute a modern DGM instance with attractor interpretation (Ambrogioni, 2023):

  • The learned time-indexed score network s(x,t;W)s(x, t; W) models xlogpt(x)\nabla_x \log p_t(x).
  • The SDE dx=σ(t)2xlogpt(x)dt+σ(t)dWtdx = \sigma(t)^2 \nabla_x \log p_t(x) dt + \sigma(t) dW_t drives both recall (denoising corrupted patterns) and generation (sampling from noise to data manifold).
  • The corresponding energy function EDM(x,t)=σ(t)2logpt(x)+CE_{DM}(x, t) = -\sigma(t)^2 \log p_t(x) + C exhibits minima at stored data patterns as tTt \to T, homologous to modern Hopfield networks.
  • Training by denoising score matching instantiates a Hebbian-like synaptic update; thus, memory is distributed in network weights, with attractor basins facilitating associative retrieval and creative synthesis in a unified process.
  • Empirical and theoretical results show that DGM so instantiated inherits the exponential capacity of modern Hopfield nets while supporting continual, end-to-end differentiable learning (Ambrogioni, 2023).

6. Tokenized and Agent-Centric Dynamic Generative Memory

Recent agentic paradigms introduce DGM as generative, on-demand latent memory interwoven into reasoning loops (Zhang et al., 29 Sep 2025, Chen et al., 26 Mar 2026):

  • In MemGen (Zhang et al., 29 Sep 2025), memory modules consist of (i) a trigger network that detects when additional memory should be injected into the LLM context, and (ii) a memory weaver that generates latent token sequences as machine-native memory, prepended to the LLM’s internal state.
  • Memory accrual and invocation are learned via RL or supervised signals, and post-hoc analysis shows that different clusters of latent memory spontaneously specialize into planning, procedural, or working memory faculties.
  • In video world modeling (HyDRA) (Chen et al., 26 Mar 2026), memory sequences are tokenized using 3D convolutions over latent spaces and are attended to via spatiotemporal relevance-driven mechanisms. This hybrid approach attends to both static background and dynamic subject features, enabling robust subject tracking and continuity even with occlusions.
  • These DGM variants demonstrate superior performance across domains such as embodied action, cross-domain reasoning, code generation, and long-term temporal prediction, with significant gains over retrieval- and parametric-memory baselines.

7. Text-to-Image Synthesis and Dynamic Memory GANs

The Dynamic Memory Generative Adversarial Network (DM-GAN) (Zhu et al., 2019) operationalizes DGM principles in conditional generation:

  • DM-GAN interposes a dynamic memory module at each refinement stage in a cascaded generator for text-to-image synthesis.
  • Memory slots are dynamically computed via a memory writing gate that weights each word feature by its relevance to the image's current state, and a response gate adaptively fuses memory readouts with image features at every spatial location.
  • The memory is not persistent; all slots are recomputed at each refinement stage based on the current image and textual context.
  • This dynamic adaptation yields significant gains in Inception Score, FID, and text-image alignment over prior models, especially when initial conditionals are poor; performance improvement is robust to ablations removing memory, writing, or response gates (Zhu et al., 2019).

References

Topic to Video (Beta)

No one has generated a video about this topic yet.

Whiteboard

No one has generated a whiteboard explanation for this topic yet.

Follow Topic

Get notified by email when new papers are published related to Dynamic Generative Memory (DGM).