The WalkXR Simulation System is an automated method for rapidly testing how different people might emotionally and cognitively respond to any part of a WalkXR experience. It simulates interactions using structured inputs and custom persona profiles, then sends them through a dedicated large language model (LLM) deployed via OpenRouter. This LLM has been parameterized for our simulations, allowing it to generate nuanced, emotionally attuned simulation output at scale.
Each simulation returns structured insight: how people might feel across a moment or entire journey, what resonates or causes friction, where curiosity or fatigue emerges, and how AI might support that experience. In addition to emotional arcs, the simulations provide feedback on prompt design, rituals, tone, inclusivity, storytelling effectiveness, and specific AI feature opportunities. They also generate potential training data for future LLM fine-tuning or RAG integration. The goal is to enable thoughtful iteration, adaptive design, and emotionally intelligent system development across both general WalkXR content and AI-based companion features.
Each walk is modular, broken into individual moments called “modules, " including prompts, rituals, reflections, or interactive storytelling. We simulate them using four modes, ranging from focused (one persona, one module) to comprehensive (all personas, full walk). With automation now fully integrated, the simulation system can rapidly ingest parameters, generate results, and organize outputs into an analysis-ready format inside a central spreadsheet.
All data structures and insight types are informed by design decisions made by Ben and Roman to support scalable, real-time WalkXR iteration and AI prototyping. Each simulation mode is tuned to extract specific feedback relevant to its scope. If you'd like new insight types or dimensions to be captured, contact Roman Di Domizio via Discord. All simulation components, including prompt templates, Apps Script logic, and output schema, must remain in sync to ensure the system works as intended.
| Walk | Modules | Personas |
|---|---|---|
| Small Moments | 8 modules based on the 5/14/25 Miro demo | 10 diverse personas with detailed backgrounds generated by ChatGPT. |
| WalkXR Public Service Experience (PSE) for Financial Scams | 10 Modules | 10 diverse personas with detailed backgrounds generated by ChatGPT. |
| WalkXR Diabetes | 8 Modules | 10 diverse personas with detailed backgrounds generated by ChatGPT. |
Simulation Modes
The WalkXR Simulation System supports four simulation modes, each designed to simulate different combinations of personas and walk/module scope. While the structure of the output remains consistent, each mode serves a distinct purpose—surfacing emotional dynamics, design friction, and opportunities for AI integration across different contexts.
Modes 2 and 3 are currently the most effective for generating high-quality, structured training data. They allow us to scale testing across multiple modules and personas while maintaining clean, analyzable results for tagging, tuning, and agent development.
All modes now use the same 25-field JSON output structure, where each object represents one persona engaging with one specific module. This enables precise mapping across emotional responses, cognitive load, design clarity, and AI readiness. The output is optimized for both human review and machine learning workflows.
Simulates how one persona responds to a specific module (e.g., prompt, ritual, reflection).
Best for testing micro-level tone, prompt clarity, or emotional activation. Useful for quick iteration or isolated module tuning.
Simulates one persona across every module in a full walk.
Best for surfacing pacing patterns, emotional flow, and cumulative experience insights. Highly recommended for generating full-run training data for agent behavior across a complete experience.
Simulates all personas responding to the same single module.
Best for comparing how diverse backgrounds interpret and react to the same prompt. Ideal for inclusivity testing, identifying resonance/failure patterns, and tuning agent adaptability.
Simulates the full journey for all personas across every module.
Best for comprehensive pattern detection, agent personalization training, and roadmap-level experience analysis. Use with caution due to the large output size and generation cost.
Each mode returns an array of structured JSON objects, one object per persona per module, using the following 25-field schema:
This schema is shared across all modes to ensure data consistency and allow multi-modal comparison. It is optimized for both agent design and emotional pattern analysis.
All simulations are powered by a custom-configured LLM hosted via OpenRouter (deepseek/deepseek-r1-0528-qwen3-8b:free). It handles longer context, performs better with structured JSON, and is more consistent for simulation chaining and reasoning. That helps a lot when running long walks or full persona batches. Guided by carefully authored prompt templates and a structured Apps Script backend. Output is automatically parsed, validated, and injected into the All Output sheet, ready for analysis.
IMPORTANT: Do not change any field names, prompt schemas, or column headers without explicit coordination. The entire system, prompt templates, App Script logic, output schema, and UI must remain synchronized.
⭐You can add new walks, modules, and personas freely, just ensure they follow the formatting of existing entries.
To simulate a new walk not yet inside the Simulation System Sheets file, follow these steps:
NOTE: The more detailed and specific the modules and personas are, the more accurate and meaningful the simulation output will be.
To run a simulation:
There’s no need to copy/paste prompts or outputs anymore, everything happens in one click.
When you click “Run Simulation” in the User Interface sheet, everything runs automatically:
deepseek/deepseek-r1-0528-qwen3-8b:freeIn real time:
💡 While the raw response is helpful, it’s best to explore results in the All Output sheet.
There, you can:
If you see “❌ JSON parsing or validation failed. See logs.”, it means the LLM response didn’t match the required format (usually missing fields). This happens occasionally; just click Run Simulation again.
Do not edit column headers or structure in All Output. The simulation system depends on this schema to work correctly. For custom formats, contact Roman.