Regenerative Logic-Core Protocol
- RLCP is a dual-stream framework that decouples LLMs' internal logic from stored factual associations to achieve a low-entropy reasoning core.
- The protocol uses adversarial unlearning with gradient reversal layers and KL-divergence anchoring to purge factual data while preserving computational logic.
- Empirical results reveal a sharp phase transition in factual retention and emergent chain-of-thought scaffolding, advancing modular neural architectures.
The Regenerative Logic-Core Protocol (RLCP) is a dual-stream adversarial unlearning framework for LLMs designed to decouple internal representations of general reasoning (logic) from those encoding specific factual knowledge (facts). RLCP implements a hypothesis termed "digital metabolism," positing that active forgetting of stored facts is necessary for distilling a low-entropy core for abstract logic, yielding models that retain algorithmic reasoning but discard entangled factual associations. Empirically, RLCP exhibits a sharp phase transition in factual retention and prompts emergent structural and behavioral phenomena, including crystallization of latent manifolds and spontaneous adoption of chain-of-thought reasoning. RLCP provides a dynamic, weight-level alternative to modular architectures that physically separate memory and computation, such as DeepSeek Engram, and constitutes a concrete step towards a "Neural CPU + Symbolic RAM" paradigm (Peng et al., 15 Jan 2026).
1. Digital Metabolism: Theoretical Motivation and Formulation
LLMs typically encode both reasoning operations and factual associations within shared weights, leading to "parameter entanglement." This superposition results in the "memory wall," where additional parameters serve mainly to store low-entropy facts rather than enhance computational capacity. Forgetting a memorized fact at inference causes the model to hallucinate, as it cannot explicitly signal retrieval failure. RLCP introduces a thermodynamic analogy—digital metabolism—treating factual associations as high-energy states to be actively purged, leaving a core of abstract logical operators.
The underlying information-theoretic objective refines the standard bottleneck formulation:
where denotes latent representations, the factual component, and the logical component. This objective penalizes factual retention in while explicitly preserving or boosting logical information.
A key empirical observation is that the gradients of factual-recall loss () are nearly orthogonal to those of logical-reasoning loss ():
Given this, purging facts via gradient steps has at most effect on logical capabilities.
2. RLCP Framework: Architecture and Dual-Stream Training
RLCP employs a dual-stream training architecture comprising three interlocked objectives:
- Metabolic Stream: Adversarial unlearning of facts via a Gradient Reversal Layer (GRL) attached at a designated deep layer (, e.g., layer 20), with a linear probe trained to recover entity identity from hidden states. The GRL inverts gradients for the probe’s loss so that the backbone model is optimized to destroy linearly decodable factual signals.
- Survival Stream: A retrieval-augmented generation (RAG)-style objective ensures the model can answer factual questions when relevant context is presented, preserving "contextual reasoning" capacity.
- Homeostatic Repair: A KL-divergence anchoring term stabilizes the unlearning process by maintaining the overall output distribution close to a reference, preventing catastrophic collapse.
The composite batch loss is:
where is the probe cross-entropy (maximized with respect to probe, minimized with respect to model), addresses context-based QA, is an unlikelihood loss penalizing correct fact recall in absence of context, and anchors the model to the reference.
RLCP Algorithm (Key Steps)
| Step | Operation | Details |
|---|---|---|
| 1 | Probe | GRL inverts probe gradient at |
| 2 | Losses | Composite of RAG, probe, unlikelihood, KL |
| 3 | Update | via |
The adversarial weight is set dynamically as a function of training progress. On each batch, the probe is updated to maximize factual decoding, while the backbone is updated to minimize it and satisfy the other objectives.
3. Mathematical Properties and Metrics
The gradient reversal layer operates as follows: during backpropagation, the hidden state is passed forward unchanged, but probe gradients are multiplied by during backward propagation.
Parameter updates decompose:
The impact on logic loss is bounded (Corollary 2.4):
Empirical metrics defined for assessment include zero-shot recall (fraction of facts correctly answered without context), probe accuracy (fraction of correct entity classifications by the linear probe), and RAG accuracy (QA with context).
4. Empirical Phenomena: Phase Transition, Structural Crystallization, and Attention Dynamics
Upon applying RLCP to Qwen2.5-0.5B with 15 city–country facts, a sharp recall phase transition is observed:
| Model | Zero-Shot Recall | Probe Accuracy (Layer 20) | RAG Performance |
|---|---|---|---|
| Original | 100% | 93.3% | 100% |
| Just-RAG Baseline | 95% | 88.5% | 100% |
| RLCP | 0% | 6.7% (chance) | 100% |
The reduction in probe accuracy to chance level (1/15 ≈ 6.7%) marks a transition to a factual "tabula rasa" state, with factual associations rendered linearly undecodable. t-SNE analysis reveals "structural crystallization," where semantic subspaces (e.g., cities, fruits) collapse into type centroids, erasing linear separability of individual facts but preserving categorical manifolds.
Attention entropy per head, defined as (with the attention weight), cools from (diffuse) in the baseline to (sharp, focused) with RLCP. Attention allocation shifts: RLCP models assign ≈70% of attention to the supporting context token (cf. <10% in baseline), indicating a reallocation of attention capacity from internal memory to external context.
5. Emergent Chain-of-Thought Scaffolding
On GSM8K mathematical reasoning tasks, RLCP models spontaneously adopt explicit chain-of-thought (CoT) scaffolding—stepwise problem-solving—in contrast to the shortcutting or direct associative recall () observed in standard models. The "metabolized" RLCP model executes explicit algorithmic decompositions, such as:
- Step 1: divide,
- Step 2: add,
- Step 3: sum...
This suggests that RLCP, by erasing direct factual associations, compels the model to reconstruct answers algorithmically. The behavioral change is interpreted as the emergence of a pure logic core free from fact-memorization, but the causal mechanism (e.g., whether due purely to resource reallocation or regularization effects) warrants further investigation.
6. Modular Architectures and Open Research Questions
RLCP achieves soft decoupling of memory and computation by dynamically unlearning facts without architectural modification of the Transformer. This complements hard decoupling approaches, such as DeepSeek Engram, which instantiate distinct modules for symbolic RAM and neural compute. RLCP findings suggest that logical computation and factual storage can be separated at the level of weights through targeted adversarial training.
Open questions include:
- Scalability: efficacy of RLCP for unlearning thousands of facts in large models (≥10B parameters) without compromising reasoning.
- Task generalization: effects of in-domain unlearning (e.g., removing arithmetic facts) on task-specific performance.
- Probing: potential for nonlinear or deeper analyses to confirm total factual erasure.
- Mechanism: disambiguating the cause of emergent chain-of-thought.
- Objective design: efficiency of alternative objectives for approximate layerwise metabolism loss.
Overall, RLCP demonstrates that targeted adversarial unlearning can carve out a neural logic core, evidencing a phase transition in factual retention, structural crystallization, focused attention, and the emergence of stepwise reasoning, advancing the modular separation of memory and computation in deep learning systems (Peng et al., 15 Jan 2026).