Papers
Topics
Authors
Recent
2000 character limit reached

Cooperative Training with Slow Thinking Solvers

Updated 5 January 2026
  • The paper presents a cooperative training paradigm where a fast solver quickly proposes solutions while a slow solver refines them through iterative deliberation.
  • It details methodologies like chain-of-thought reasoning, MCMC-based energy minimization, and meta-controller gating to boost sample efficiency and robustness.
  • Empirical findings show significant gains in navigation, vision segmentation, and language reasoning, achieving improved accuracy and reduced computational steps.

Cooperative training with slow thinking solvers refers to machine learning frameworks in which a “fast” component proposes candidate solutions rapidly, while a “slow” component iteratively refines, evaluates, or trains the fast component using more deliberative processes. This paradigm, inspired by human dual-process cognitive theories (System 1: fast, intuitive; System 2: slow, analytic), has been realized in diverse settings, including vision, language, navigation, generative modeling, and multi-modal problems. Crucially, the interaction is not just at deployment: cooperative training mechanisms allow the fast solver to learn from, and be shaped by, the slow solver’s outputs—resulting in models that combine rapid inference with improved robustness, sample efficiency, or reasoning ability.

1. Dual-Process Framework: Fast and Slow Solvers

The dual-process framework formalizes reasoning as the interplay between a fast-thinking module (“System I”) and a slow-thinking module (“System II”) (Chung et al., 27 May 2025, Su et al., 2024, Saeed et al., 27 Jun 2025, Ganapini et al., 2022, Xie et al., 2019). The fast module is typically a feed-forward (or direct-mapping) neural network that generates candidate solutions with minimal computation. The slow module operates by iterative search, optimization, or deliberation, often guided by an explicit or implicit objective and having access to mechanisms (such as energy functions, chain-of-thought reasoning, or self-play) that allow deeper exploration or verification.

Generic structure:

  • Fast solver: Provides immediate proposals or predictions, e.g., selecting actions in grid navigation, generating initial image segmentations, proposing answers in QA.
  • Slow solver: Runs a more computationally intensive process, such as search (A*, self-play RL, MCMC, or explicit chain-of-thought), to iteratively refine or verify outputs.
  • Cooperation: The slow solver either serves as a source of high-quality supervision or refines the fast solver’s outputs; in certain settings, the fast solver is explicitly trained to mimic or distill the slow solver’s behavior.

2. Cooperative Training Algorithms

Cooperative training schemes differ in implementation but share fundamental elements:

  • Bootstrapping: The fast solver’s outputs serve as initializations for the slow solver, which then yields improved outputs via iterative, slow reasoning.
  • Feedback: Gradients or differences from the slow solver's refinements are used to update the fast solver’s parameters. This encourages the fast solver to internalize patterns and regularities that the slow solver uncovers.
  • Teacher-student and self-distillation: The slow solver acts as a teacher, generating trajectories, plans, or refined answers that augment the fast solver's training set.

Representative Approaches and Technical Realizations

Framework Fast Solver Slow Solver Cooperative Mechanism
Thinker Direct answer in <1000 tokens 4-stage RL w/ deliberation Stage-wise PPO, refinement and summarization (Chung et al., 27 May 2025)
Dualformer Transformer, plan-only Transformer, full trace Randomized trace dropping, joint cross-entropy (Su et al., 2024)
CoopNets Conditional generator MCMC under EBM Joint updates via mapping/objective shift (Xie et al., 2019)
SOFAI Lookup table policy MDFT deliberative RL Meta-controller orchestrates, S2 data updates S1 (Ganapini et al., 2022)
SOPHIA On-policy visual encoder Off-policy LLM reasoning Semi-off-policy RL, reward propagation (Shen et al., 22 Jul 2025)
Vision Dual-Process Feed-forward prediction Self-play RL refinement Iterative self-improvement, pseudo-labeling (Saeed et al., 27 Jun 2025)

The cooperative interaction often proceeds via alternating or joint updates—e.g., the fast initializer is trained to map inputs to the slow solver’s outputs, while the solver is refined via contrastive divergence between data and fast-initialized samples (Xie et al., 2019).

3. Architectures and Mathematical Formalisms

Formalism varies by domain, but key mathematical features include:

  • Energy-Based Modeling with Cooperative Updates: The slow solver defines an energy or value function, and refines fast-initialized predictions via MCMC or Langevin dynamics; the fast initializer is trained to minimize the distance to the slow solver’s refinements (Xie et al., 2019).
  • Chain-of-Thought and Randomized Trace-Dropping: Transformers trained on randomized reasoning traces learn both slow (full trace) and fast (plan-only) modes. Training samples are corrupted at various granularities to force the model to interpolate between deliberation and intuition (Su et al., 2024).
  • Multi-Stage RL (Thinker Task): RL is cast as a multi-stage MDP with distinct policies and rewards for fast answering, verification, slow refinement, and summarization, with intra- and inter-stage discounting to enforce proper credit assignment (Chung et al., 27 May 2025).
  • Meta-Controller Gating: Hand-designed or learned meta-controllers allocate compute between fast and slow solvers, based on experience, state visitation, confidence, and expected reward (Ganapini et al., 2022).
  • Self-Play and Adversarial Discrimination: Iterated self-play among slow solver refiners with rewards assigned via a discriminative model enables continuous improvement with additional “thinking time,” especially valuable under limited labeled data (Saeed et al., 27 Jun 2025).

4. Empirical Findings and Efficiency Gains

Cooperative frameworks demonstrate significant improvements in key metrics:

  • Efficiency–Quality Trade-offs: Dualformer solves 30×30 maze navigation tasks with an optimal rate of 97.6% in slow mode (surpassing full-trace baselines) and 80% in fast mode (compared to 30% for plan-only training), while reducing reasoning step count by up to 59.9% (Su et al., 2024).
  • Sample Efficiency and Robustness: Cooperative training in segmentation (FTN + STN) yields 15% relative gains in Dice over standard baselines with only 10 training subjects, and consistently better performance under distribution shift and artifacts (Chen et al., 2021).
  • Iterative Improvement: In machine vision tasks, performance increases monotonically with additional self-play inference steps, surpassing both foundation models and large supervised nets, with gains >3–21 percentage points in cancer localization, and robustness with as few as 4–16 labeled cases per new domain (Saeed et al., 27 Jun 2025).
  • Language Reasoning Accuracy: The Thinker 4-stage task yields accuracy improvements from 45.9% to 51% on DeepSeek-R1-Qwen-1.5B, with “Thinker-Fast” alone using 1k tokens approaching full-length QA accuracy at much lower cost (Chung et al., 27 May 2025).
  • Transfer and Distillation: Cooperative frameworks facilitate generalization across reasoning, planning, and non-verbal tasks, with slow solver refinements enabling fast solvers to internalize shortcuts while maintaining high-fidelity outputs (Su et al., 2024).

5. Domain-Specific Variants and Extensions

Language and Multimodal Reasoning

Frameworks such as SOPHIA integrate on-policy LVLMs (for vision-language alignment) with off-policy, slow-thinking LLMs to propagate refined reasoning traces, crucial for challenging tasks like MathVision and OlympiadBench, where SOPHIA closes the performance gap to closed-source GPT-4.1 (Shen et al., 22 Jul 2025).

Vision (Segmentation, Diagnosis, Navigation)

Cooperative dual-process designs are applied to high-dimensional solution spaces: image segmentation (decoupled encoder–decoder paired with denoising autoencoder for refinement), medical image analysis (adversarially trained predictors with discriminative self-play for iterative improvement), and constrained navigation (combining lookup-based policies and deliberative RL solvers) (Chen et al., 2021, Saeed et al., 27 Jun 2025, Ganapini et al., 2022).

Generative Modeling

Energy-based cooperative training allows stable, diverse generation across conditional tasks (e.g., sketch-to-photo, edge-to-image), with fast generators amortizing MCMC refinements and energy solvers guaranteeing sample diversity (Xie et al., 2019).

6. Comparative Analysis and Theoretical Insights

Relative to cascaded or meta-controller approaches (where fast and slow solvers are separate models with explicit gating logic), single-backbone cooperative methods (as in Dualformer) avoid the complexity of multi-stage pipelines and controller tuning, yet still allow controllable trade-offs between speed and depth of reasoning (Su et al., 2024).

Theoretically, “objective shift” (contrastive divergence on slow-refined fast-generated samples) and “mapping shift” (distilling slow solver corrections into the initializer) underpin convergence and mutual regularization (Xie et al., 2019). Self-play with discriminator-based rewards replaces explicit loss shaping with adversarial refinement, enabling applicability to non-verbal tasks without rule-based reward decomposition (Saeed et al., 27 Jun 2025).

Cooperative feedback loops also provide resilience to overfitting, improve generalization under distribution shifts, and generate robust representations, particularly when paired with latent-space augmentations or stochastic corruption during training (Chen et al., 2021, Su et al., 2024).

7. Limitations and Future Directions

Current cooperative training frameworks face computational challenges due to iterative steps or the requirement for high-quality slow solver outputs. MCMC-based refinement can be slow for high-dimensional targets, and slow-solver reliance can bottleneck online adaptation in dynamic environments (Xie et al., 2019, Su et al., 2024). A plausible implication is that further advances in short-run MCMC, efficient chain-of-thought distillation, or hybrid policy optimization may unlock even broader applicability.

Potential extensions include generalization to unpaired data, symbolic reasoning tasks, multi-agent negotiation, and domains lacking explicit reward signals but equipped with learnable discriminators. The template of learning a verifier and coupling it to iterative self-play refinement (vision), or randomized trace-based distillation (language), points toward increasingly general, adaptive reasoning agents that can trade compute for rigor and extract maximal information from limited supervision (Saeed et al., 27 Jun 2025, Su et al., 2024).


In summary, cooperative training with slow thinking solvers is an empirically validated and theoretically grounded paradigm that supports fast, robust, and high-quality prediction in domains ranging from navigation and vision to language and multimodal reasoning. By explicitly modeling and leveraging the complementary strengths of fast and slow processes, it enables the construction of systems capable of balancing efficiency, accuracy, and adaptive reasoning across diverse tasks.

Whiteboard

Topic to Video (Beta)

Follow Topic

Get notified by email when new papers are published related to Cooperative Training with Slow Thinking Solvers.