OpenGame: Open Agentic Coding for Games
Abstract: Game development sits at the intersection of creative design and intricate software engineering, demanding the joint orchestration of game engines, real-time loops, and tightly coupled state across many files. While LLMs and code agents now solve isolated programming tasks with ease, they consistently stumble when asked to produce a fully playable game from a high-level design, collapsing under cross-file inconsistencies, broken scene wiring, and logical incoherence. We bridge this gap with OpenGame, the first open-source agentic framework explicitly designed for end-to-end web game creation. At its core lies Game Skill, a reusable, evolving capability composed of a Template Skill that grows a library of project skeletons from experience and a Debug Skill that maintains a living protocol of verified fixes - together enabling the agent to scaffold stable architectures and systematically repair integration errors rather than patch isolated syntax bugs. Powering this framework is GameCoder-27B, a code LLM specialized for game engine mastery through a three-stage pipeline of continual pre-training, supervised fine-tuning, and execution-grounded reinforcement learning. Since verifying interactive playability is fundamentally harder than checking static code, we further introduce OpenGame-Bench, an evaluation pipeline that scores agentic game generation along Build Health, Visual Usability, and Intent Alignment via headless browser execution and VLM judging. Across 150 diverse game prompts, OpenGame establishes a new state-of-the-art. We hope OpenGame pushes code agents beyond discrete software engineering problems and toward building complex, interactive real-world applications. Our framework will be fully open-sourced.
Paper Prompts
Sign up for free to create and run prompts on this paper using GPT-5.
Top Community Prompts
Explain it Like I'm 14
OpenGame: Turning Ideas Into Playable Web Games with AI
1) Big picture: What is this paper about?
This paper introduces OpenGame, an AI system that can turn a plain‑English idea (like “make a simple space shooter where I dodge asteroids”) into a working 2D web game. It doesn’t just write bits of code—it plans the project, builds it, tests it, fixes mistakes, and adds art and sounds so the result is actually playable in a browser.
2) What questions are the researchers trying to answer?
The authors focus on three simple questions:
- Can an AI reliably make a full game from a written description, not just small snippets of code?
- What kind of “skills” help an AI keep large projects organized and fix common mistakes on its own?
- How do we fairly test whether a generated game really works and matches the user’s idea?
3) How does OpenGame work? (Explained with everyday analogies)
Making a full game is like building a tiny theme park: lots of moving parts must work together (graphics, physics, input, scenes, assets). OpenGame tackles this with three main pieces:
- A specialized coding model (GameCoder‑27B)
- Think of this like a chef trained specifically in game recipes, not just general cooking. The model learns:
- Continual Pre‑Training (CPT): It “reads” lots of open game code (especially Phaser 3, a popular web game engine) to learn patterns like game loops and physics.
- Supervised Fine‑Tuning (SFT): It practices with teacher‑provided examples and step‑by‑step solutions, so it can follow instructions better.
- Reinforcement Learning (RL): It does trial‑and‑error on smaller game tasks (like collision or state machines), gets a score when the code runs correctly, and learns from that feedback—like a coach giving it pointers after scrimmages.
- An “agentic” coding process (the AI acts step by step)
- “Agentic” here means the AI plans, writes, runs, checks, and fixes its own code in a loop. The process has six phases:
- 1) Understand your request and classify the game by physics (e.g., “gravity platformer” vs “top‑down movement”), which helps choose the right plan.
- 2) Scaffolding: Create a clean project skeleton—the basic folders, files, and structure—before adding details (like building the stage before installing rides).
- 3) Design a GDD (Game Design Document): A simple blueprint listing rules, characters, goals, and assets.
- 4) Generate assets: Use AI to create art, animations, and sounds that match the plan.
- 5) Code implementation: Fill in specific “hooks” in template files rather than writing everything from scratch, so the structure stays stable.
- 6) Verify and fix: Build and run the game in a test browser and automatically repair errors until it plays.
- “Game Skills” the agent learns and reuses
- Template Skill: A growing library of reliable starting blueprints (templates) for common game types—like “side‑view with gravity,” “top‑down movement,” “grid‑based logic,” “path/wave enemies,” and “UI‑driven” games. Picking the right blueprint early keeps the whole project consistent.
- Debug Skill: A living repair guide the AI updates over time. It records common problems (e.g., wrong asset names, missing configs) and their proven fixes—like a mechanic’s notebook—so it can solve issues faster in the future.
Important tools and terms explained simply:
- Phaser 3: A JavaScript toolbox for making web games. It’s popular and friendly to code.
- Headless browser: A “robot” web browser without a visible screen, used to automatically run and test games.
- Vision‑LLM (VLM) judge: An AI that “looks” at screenshots and “reads” instructions to check if the game’s visuals and behaviors match the request.
4) What did they find, and why does it matter?
The team built a new test called OpenGame‑Bench that scores generated games on:
- Build Health: Does it compile and run without crashing?
- Visual Usability: Does it display clear, animated, interactable scenes (not just a blank screen)?
- Intent Alignment: Does the game actually do what the user asked?
On 150 different game prompts, OpenGame beat strong baselines. In one top setup, it achieved about:
- Build Health ≈ 72/100
- Visual Usability ≈ 67/100
- Intent Alignment ≈ 65/100
Why this is important:
- It shows an AI can create full, interactive games, not just code snippets.
- The biggest gains were in Intent Alignment, meaning the AI stuck closer to the user’s idea. That’s crucial for creativity and usefulness.
- The specialized “Template Skill” and “Debug Skill,” plus step‑by‑step verification, made a real difference—especially compared to general‑purpose code models.
Extra insights:
- Iterative debugging (fixing in a few rounds) greatly improves results compared to one‑shot generation.
- Physics‑heavy genres (like platformers) worked best; more abstract logic games (like some strategy or puzzle types) were harder because mistakes are less visible and harder for the AI to detect.
5) What could this change in the future?
- Lower barrier to creating games: Teachers, students, and hobbyists could turn ideas into playable games quickly, even without expert coding skills.
- Better AI software builders: The techniques—reusable templates, living debugging guides, and real play‑testing—can help AI build other complex interactive apps, not just games.
- Fairer evaluations: OpenGame‑Bench focuses on whether things actually work when you play them, not just whether code “looks right,” which is a step toward more realistic testing of AI‑written software.
In short, OpenGame is a big step toward AI that can handle creative, complex projects end‑to‑end. It blends solid “blueprints,” a growing “repair handbook,” a game‑smart code model, and real testing—so more people can bring their game ideas to life.
Knowledge Gaps
Knowledge gaps, limitations, and open questions
Below is a single, concrete list of what remains missing, uncertain, or unexplored in the paper, framed to guide future research:
- Generalization beyond Phaser 3 and 2D web: How well do Template/Debug Skills and GameCoder-27B transfer to other engines (Godot, Unity, Unreal), to 3D games, or to mobile deployment targets?
- Multi-file integration RL: The RL stage trains on single-file modules via unit tests; how to design execution-grounded RL environments that reward multi-file coordination, scene wiring, and temporal gameplay correctness?
- Evaluation validity and robustness: The reliance on VLM judges for Visual Usability and Intent Alignment lacks calibration, inter-judge agreement, and robustness analysis under adversarial or ambiguous prompts.
- Ground-truth requirement extraction: The automatic conversion of free-form prompts into structured requirement specs (for IA scoring) is not validated; how accurate and consistent is this step?
- Human-centered evaluation: No human playtesting or UX studies on fun, engagement, clarity of controls, or perceived quality; how do automated scores correlate with human judgments?
- Weakness on logic-heavy genres: Strategy and puzzle/UI titles show markedly lower IA; which diagnostic tools (e.g., property-based tests, invariant checks, state instrumentation) best surface silent logic desynchronization?
- Debug protocol governance: How are conflicting rules resolved, how is rule quality assessed over time, and how is P versioned/rolled back to avoid brittle heuristic accumulation?
- Template library evolution risks: What safeguards prevent overfitting templates to benchmark distributions, and how are fragment extraction criteria validated for safety and reusability?
- Dataset transparency and leakage: Precise composition, size, and licenses of CPT/SFT corpora are not detailed; given prompts sourced from public repos, what measures ensure train–eval contamination is avoided?
- Synthetic supervision provenance: SFT relies on proprietary models (e.g., MiniMax 2.5) to generate “ground truth”; how reliable are these targets, and will the synthetic dataset be released for reproducibility?
- Proprietary backend dependence: Best results require Claude Sonnet 4.6; what performance can be guaranteed in a fully open-source stack, and how to reduce dependence on closed models?
- Cost and scalability: Wall-clock time, compute cost, and memory footprint per game (especially across T debugging iterations) are not reported; how do costs scale with game scope and asset complexity?
- Asset pipeline limitations: Audio generation/quality, consistency across asset styles, and content moderation/IP compliance are unaddressed; how to ensure legally safe, coherent, and performant assets?
- Asset licensing and IP: Examples reference copyrighted franchises (e.g., Marvel); how does the system prevent infringing content or trademarked assets during generation?
- Non-visual mechanics evaluation: VU favors visible motion/entropy; how to detect correctness of non-visual mechanics (e.g., inventory logic, hidden timers) that may not be captured by screenshots?
- Performance/runtime metrics: No reporting on FPS, input latency, memory usage, or mobile/browser variability; what runtime constraints emerge for larger scenes and animation-heavy games?
- Multiplatform packaging: Lack of experiments on packaging, deployment, and compatibility across browsers and devices; can the pipeline produce PWA/mobile-friendly builds?
- Interactive clarification loop: The agent never queries the user when prompts are ambiguous; would active clarification improve IA and reduce misclassification by the Physics-First classifier?
- Classifier/tool reliability: No accuracy/error analysis for classify-game-type and other tools; how often do misclassifications route tasks to suboptimal template families?
- Security and sandboxing: Executing generated code/assets in headless browsers raises security concerns; what isolation/permission models are necessary for safe, at-scale evaluation?
- Long-context handling: The Three-Layer Reading Strategy is beneficial, but scalability with very large projects and long GDDs is unclear; how do different context management schemes affect outcomes?
- Debug iteration policy: The impact of iteration budgets (T) beyond small values and adaptive stopping criteria is underexplored; could learned stopping or confidence estimates reduce unnecessary cycles?
- Maintainability and code quality: Metrics for readability, modularity, test coverage, and code smells are absent; how does template-driven code fare under static analysis and long-term maintenance?
- Collaboration workflows: No support for multi-agent or mixed-initiative workflows with artists/designers; how to integrate human feedback loops, version control, and asset revisions mid-generation?
- Multiplayer and persistence: The framework targets single-player, single-session games; how to extend to networked mechanics, save/load systems, and persistent state across sessions?
- Physics and AI opponents: Evaluations do not cover advanced physics (ragdolls, joints), pathfinding, or NPC AI behavior correctness; how to benchmark such capabilities reliably?
- Benchmark fairness and scope: OpenGame-Bench enforces Phaser usage; does this bias results toward Phaser-centric templates and hinder general agent comparisons?
- Failure taxonomy and reporting: Pipeline errors are reported separately but not categorized; which failure classes dominate (build vs runtime vs interaction), and what targeted mitigations are most effective?
- Continual learning stability: As £ and P evolve, how to prevent catastrophic forgetting, rule drift, or degradation on earlier tasks; can formal tests guard against regression?
- Creativity/originality: Template reuse may reduce diversity; how to measure and foster novelty while maintaining stability and correctness?
- Accessibility and localization: No consideration of color-blind modes, screen-reader cues, or multi-language UI; how can accessibility be specified and validated in the GDD and evaluation?
- Reproducibility details: Hyperparameters, training schedules, and exact data splits for CPT/SFT/RL are not fully specified; what artifacts and seeds will be released to ensure reproducibility?
- Generalization under domain shift: Performance under out-of-distribution prompts (novel mechanics, hybrid genres, unusual control schemes) remains untested; what stress tests best probe robustness?
Practical Applications
Overview
Based on the paper’s contributions—OpenGame (agentic framework for end-to-end web game creation), Game Skill (Template Skill + Debug Skill), GameCoder-27B (domain-specialized code LLM), and OpenGame-Bench (dynamic evaluation)—the following applications map these findings to concrete, real-world use cases across industry, academia, policy, and daily life.
Immediate Applications
These are deployable now with the current capabilities (web-based 2D games using Phaser; automated scaffolding, asset synthesis, and iterative debugging; headless-browser/VLM evaluation).
- Classroom “lesson-to-game” authoring for teachers (education)
- What: Turn lesson plans or quizzes into playable browser games (e.g., buzz-in quiz games, vocabulary puzzles, physics challenges) using Template Skill to scaffold genre-appropriate structures and asset synthesis for visuals/audio.
- How: Workflow—paste lesson text → generate GDD → asset synthesis → build → OpenGame-Bench sanity check → classroom deployment via a simple web host or LMS embed.
- Tools/Products: “EduGame Builder” plugin for LMSs; Google Classroom/Canvas integration; template packs (quiz, flashcards, top-down lab).
- Assumptions/Dependencies: School IT must allow hosting/embedding web games; moderation for age-appropriate content; accessibility (WCAG) not guaranteed out-of-the-box; internet connectivity for asset generation.
- Rapid prototyping for indie studios and game jams (software/gaming)
- What: Create vertical slices and prototypes from natural-language briefs; reuse template families (platformer, top-down, tower-defense/path-and-wave) to accelerate iteration.
- How: Brief → classify-game-type → scaffold → implement with hook-driven methods → iterate using Debug Skill; validate with OpenGame-Bench.
- Tools/Products: “JamKit” starter with prebuilt archetype libraries and a one-click deploy-to-Itch workflow.
- Assumptions/Dependencies: Best suited to 2D web games; advanced art/audio quality and polish still need human refinement; IP clearance for any third-party assets/themes.
- Advergame and campaign microsites for marketers (media/advertising)
- What: Generate seasonal or branded mini-games (e.g., product launch, event engagement) quickly for web embeds.
- How: Brand brief → tailored GDD → style-constrained asset generation → build → embed script for CMS sites.
- Tools/Products: “AdverGame-as-a-Service” with template/preset brand palettes; analytics hooks for engagement funnels.
- Assumptions/Dependencies: Brand approvals for generated assets; legal review for likeness/IP; basic analytics SDK integration required.
- Creator/streamer engagement games (content platforms)
- What: Lightweight meme or audience-participation games for livestreams or social posts.
- How: Incorporate chat triggers/leaderboards in a Phaser project; auto-generate art/audio consistent with theme.
- Tools/Products: OBS/Streamlabs overlay-ready builds; simple “chat commands to events” middleware.
- Assumptions/Dependencies: Platform APIs for chat/overlays; moderation for user-generated content.
- Serious mini-games for public outreach (policy/public sector)
- What: Browser-based interactive explainers (public health behaviors, recycling, safety drills) delivered as simple games.
- How: Agency brief → template selection (UI-driven or puzzle) → asset generation with agency branding → deploy to public site.
- Tools/Products: “CivicGame Kit” with accessibility-first UI templates; multilingual asset generation.
- Assumptions/Dependencies: Compliance with branding and accessibility standards; legal review for messaging accuracy.
- Gamified patient education and engagement (healthcare)
- What: Educational games explaining procedures, medication adherence, or rehab routines.
- How: Clinician-provided scripts → GDD → gentle mechanics (UI-driven puzzle/quiz) → deploy to patient portals or kiosks.
- Tools/Products: “HealthEdu Game Pack” with HIPAA-friendly hosting patterns (static content; no PHI).
- Assumptions/Dependencies: No storage of PHI in game code/assets; hospital IT approvals; clinical content validation.
- Automated UI/playability smoke testing for web games/apps (software QA)
- What: Use OpenGame-Bench’s headless execution + VLM judging for visual usability and basic intent checks in CI.
- How: Run builds in CI → capture frames → compute entropy/motion → VLM assertions (e.g., “Start button visible”).
- Tools/Products: GitHub Actions/CircleCI or self-hosted runners with OpenGame-Bench adapters for non-game SPAs.
- Assumptions/Dependencies: Stable headless browser environment; well-specified visual/intent assertions; VLM inference costs.
- Coding education and bootcamps (education)
- What: Teach game loops, state management, and debugging using generated projects; students modify hook methods.
- How: Instructor presets archetype → students extend scripts in designated extension points → automatic evaluation with OpenGame-Bench.
- Tools/Products: Classroom-ready template packs; graded rubrics tied to Intent Alignment metrics.
- Assumptions/Dependencies: Developer environment setup; alignment of curriculum with Phaser/TypeScript.
- Reusable debugging knowledge in CI pipelines (software tooling)
- What: Incorporate the “living debug protocol” to auto-detect and fix recurring issues (e.g., asset key mismatches, invalid scene transitions) before human review.
- How: Pre-execution validators + common fix recipes run as pre-commit hooks or CI steps.
- Tools/Products: “Debug Protocol Runner” CLI; VS Code extension surfacing suggested fixes.
- Assumptions/Dependencies: Error signatures must be mappable to rules; best results within Phaser/JS ecosystems.
- Accessibility and localization starter games (education/NGOs)
- What: Generate localized UI-driven games with text/audio variants for outreach programs.
- How: Asset synthesis conditioned on locale; auto-populate transcripts/captions; simple language-switch UI.
- Tools/Products: “L10n Game Starter” with multi-language JSON and alt-text scaffolds.
- Assumptions/Dependencies: Machine translation quality varies; manual accessibility testing still required.
Long-Term Applications
These require further research, scaling, or productization beyond current web-based 2D scope or need broader ecosystem integration.
- Extension to professional engines (Unity/Unreal/Godot) (gaming/software)
- What: Apply Game Skill to engines using proprietary GUIs and binary assets to produce richer 2D/3D games.
- How: Combine code generation with GUI automation and asset pipeline handling; learn engine-specific serialization.
- Dependencies: Robust computer-use agents for editor workflows; engine API coverage; licensing constraints; compute for builds.
- Production-grade end-to-end game generation (gaming)
- What: From prompt to store-ready titles (mobile/web), including performance tuning, analytics, and monetization.
- How: Integrate ad/SDKs, payments, crash reporting, telemetry-driven balancing; human-in-the-loop art and narrative passes.
- Dependencies: SDK integrations, privacy and platform policy compliance, scalable asset pipelines, QA/regression testing.
- General-purpose interactive-app generation (beyond games) (software/enterprise)
- What: Use Template Skill + Debug Skill to scaffold complex, interactive web applications (dashboards, training simulators) with dynamic evaluation similar to OpenGame-Bench.
- How: New template families for forms/flows, stateful widgets, data-binding; VLM-based intent checks for UI/UX criteria.
- Dependencies: Domain-specific component libraries, test oracles for business logic, data governance/security reviews.
- Automated level design, balancing, and A/B optimization loops (gaming/analytics)
- What: Integrate telemetry to auto-tune difficulty, pacing, and rewards through RL or Bayesian optimization.
- How: Continuous deploy → collect play data → propose code/parameter changes → evaluate via OpenGame-Bench + live metrics.
- Dependencies: Data pipelines and privacy compliance; experiment frameworks; guardrails to prevent negative UX.
- Serious game platforms for workforce training (healthcare, energy, public safety, finance)
- What: Generate training scenarios (e.g., triage, grid incidents, phishing response) as interactive simulations.
- How: Domain ontologies feed GDD; scenario scripting; assessment hooks for competency tracking.
- Dependencies: SME-validated content; regulatory alignment (e.g., OSHA, FINRA); secure hosting.
- No-/low-code “prompt-to-deploy” platforms (SaaS)
- What: A hosted service where users describe a game/app and get a live URL, with built-in templates, asset libraries, and evaluation gates.
- How: Managed OpenGame backend with one-click publish, versioning, and team collaboration features.
- Dependencies: Multi-tenant security, compute cost management, content moderation, uptime SLAs.
- Multimodal co-creation pipelines with DCC tools (creative software)
- What: Seamless round-trips with Figma, Aseprite, Spine, or Blender: GDD → code → asset edits → re-integration.
- How: Import/export bridges and schema-aware adapters; preserve animation rigs and atlases; map design tokens to code.
- Dependencies: Stable file format support; API access; version control/merge strategies for assets.
- Advanced QA for interactive systems with richer judges (software QA/HCI)
- What: Extend OpenGame-Bench to evaluate UX flows, accessibility, and non-functional requirements using specialized VLMs and agents.
- How: Scripted interaction sequences; heuristic + learned metrics (latency, responsiveness); accessibility audits.
- Dependencies: High-fidelity UI understanding; cost-effective multimodal inference; robust test data generation.
- Safety/ethics and IP-aware content generation (policy/compliance)
- What: Guardrails for copyrighted characters, violent themes for minors, or disallowed content; automatic brand/IP checks.
- How: Content filters and licensing checkers integrated into asset synthesis and GDD; policy-aware prompts.
- Dependencies: Reliable IP detection databases; false-positive/negative management; jurisdiction-specific policies.
- Lightweight simulators for robotics/HRI or science education (robotics/education)
- What: Generate 2D physics or logic simulators for algorithm prototyping or instructional labs (e.g., kinematics, planning).
- How: Physics-first templates extended with sensor models and scripted tasks; automated scoring within OpenGame-Bench.
- Dependencies: Adequate fidelity for target domain; bridging to real-world data/hardware when needed.
- Community-driven template and debug-protocol ecosystems (open source)
- What: Marketplace/repository of vetted template families (genres/domains) and shared debugging rules across frameworks.
- How: Contribution guidelines; provenance and quality badges; automatic regression checks using OpenGame-Bench.
- Dependencies: Governance/maintainers; compatibility matrices; long-term sustainability.
Cross-cutting Assumptions and Dependencies
- Engine scope: Current strengths are in 2D web games using Phaser; 3D or proprietary engines require significant extension.
- Asset generation: Quality, licensing, and brand compliance of generated images/audio vary; professional pipelines may be needed.
- Evaluation: OpenGame-Bench relies on headless browsers and VLMs; compute cost and determinism need management.
- Security and privacy: Generated code must be sandboxed; avoid embedding secrets/PII; adhere to school/enterprise IT policies.
- Accessibility and localization: Baseline support is limited; human QA remains critical for compliance and quality.
- Human-in-the-loop: Creative direction, narrative, and fine art typically still require designer oversight for production releases.
Glossary
- Agentic framework: A system that coordinates autonomous tools and reasoning steps to accomplish complex tasks end-to-end. "the first open-source agentic framework explicitly designed for end-to-end web game creation."
- Archetype: A canonical gameplay/physics pattern used to classify tasks and select suitable templates (e.g., platformer, grid-based). "archetype-specific API constraints"
- Asset pipeline: The processes and tooling that manage creation, packaging, and loading of art and audio resources in a game. "update loops, physics, event handling, asset pipelines, and tightly coupled state"
- Asset registry: A structured list of required assets and their keys used to coordinate code with generated resources. "from the GDD's asset registry."
- Build Health (BH): An evaluation metric that measures whether a project compiles, loads, and runs without critical errors. "Build Health (BH) measures whether the project compiles, loads, and renders without critical errors."
- Continual Pre-Training (CPT): Further pretraining of a model on domain-specific corpora to instill specialized knowledge. "three-stage training pipeline: Continual Pre-Training (CPT), Supervised Fine-Tuning (SFT), and Reinforcement Learning (RL)."
- Debug Skill: An agent capability that accumulates and applies verified fixes via a shared debugging protocol to repair projects. "Debug Skill maintains a living debugging protocol (P)"
- Deterministic game engines: Engines that produce predictable, repeatable outcomes for the same inputs, aiding testing and reproducibility. "which requires deterministic game engines."
- Discrete grid logic: A gameplay regime where movement and interactions occur on a discrete grid rather than continuous space. "discrete grid logic, path-and-wave dynamics, and UI-driven gameplay."
- Execution-grounded reinforcement learning: RL that uses actual code execution results (e.g., tests) to provide feedback and rewards. "execution-grounded reinforcement learning."
- Game Design Document (GDD): A technical specification of mechanics, assets, and systems that guides implementation. "produce a technical Game Design Document (GDD)."
- Game loop: The continuous cycle that updates game state and rendering in real time. "loses track of global state across the game loop"
- Headless browser: A browser environment without a graphical UI used for automated execution and testing. "via headless browser execution and VLM judging."
- Hook methods: Predefined extension points in a base class meant to be overridden to inject custom behavior. "overrides designated hook methods (e.g., setupCustomCollisions)"
- Intent Alignment (IA): An evaluation metric that gauges how well the generated game satisfies the natural-language requirements. "Intent Alignment (IA) derives a weighted pass rate from per-requirement verdicts"
- Living debugging protocol: A continually updated repository of error signatures, root causes, and verified fixes used during repair. "a living debugging protocol (P)"
- Meta template (M0): A minimal, game-agnostic project skeleton that defines universal structure for a playable game. "starting from a single game-agnostic meta template (M0)"
- Path-and-wave dynamics: A template family for games (e.g., tower defense) where entities follow paths and spawn in waves. "path-and-wave dynamics"
- Phaser 3: A popular web-based 2D game framework with a programmatic API surface suited to LLMs. "use the Phaser 3 framework."
- Physics-First Classification: A routing strategy that categorizes tasks by physical and spatial mechanics to select an appropriate archetype. "Physics-First Classification rule"
- Project scaffolding: Automatically creating a stable initial project structure and boilerplate before adding game-specific logic. "stabilizes project scaffolding and resolves recurrent cross-file failures."
- Scene wiring: The configuration and connections among scenes, assets, and initialization that ensure correct runtime flow. "broken scene wiring"
- State-machine transitions: Changes between well-defined states (e.g., idle→run→jump) governed by a state machine. "state-machine transitions"
- Template Method Pattern: An OOP design pattern where a base algorithm defines steps and subclasses override hooks to customize parts. "Template Method Pattern: rather than writing the project from scratch, the agent copies template files and overrides designated hook methods"
- Template Skill: The capability that curates and applies reusable project skeletons to stabilize structure and reduce search space. "Template Skill grows an evolving library of project skeletons"
- Template family: A group of specialized templates capturing recurring physics/interaction regimes for reuse. "template families such as gravity-based side view and top-down continuous motion."
- Three-Layer Reading Strategy: A staged context-loading approach that prioritizes API summary, target source, and implementation guide to reduce context drift. "we introduce a Three-Layer Reading Strategy."
- Tilemap: A grid-based map representation where tiles are described in data (often JSON) for rendering and collision. "generate-tilemap converts ASCII layouts into structured JSON tilemaps."
- Top-down continuous motion: A gameplay regime where entities move continuously in a top-down view rather than grid-stepped. "top-down continuous motion"
- Unit tests: Automated tests that validate small components or modules against specified behaviors. "evaluated against predefined unit tests"
- Vision-LLM (VLM): A model that jointly processes images and text, used here to judge visual quality and requirement satisfaction. "Vision-LLM (VLM) judge score"
Collections
Sign up for free to add this paper to one or more collections.