Overview
A simulation orchestrates multiple conversations between your agent and AI-generated personas based on a specific scenario. Each simulation has clear goals, constraints, and success criteria that guide the evaluation process.How Simulations Work
Simulations bring together an agent, a scenario, and evaluation objectives to create comprehensive testing environments. You define:- The scenario context that guides all conversations
- Target numbers for personas and conversations you want to generate
- Turn limits to prevent conversations from running too long
- Approval settings to control persona quality
- Objectives that define what success looks like
Simulation Lifecycle
1
Configuration
Set up simulation parameters, define scenario, and select objectives for evaluation
2
Persona Generation
Create AI-generated personas with diverse profiles and characteristics
3
Approval Process
Review and approve generated personas (if auto_approve is false)
4
Conversation Assignment
Assign conversations between approved personas and the agent
5
Conversation Execution
Run conversations with turn limits and scenario context
6
Evaluation
Score completed conversations against defined objectives
7
Analysis
Review results and performance metrics
Status Tracking
Simulations progress through several states as they execute:- pending: Simulation created but not yet started
- queued: Simulation is queued for execution
- in_progress: Personas are being generated or conversations are running
- completed: All target conversations have been finished and evaluated
- failed: Simulation encountered errors and stopped
- canceled: Simulation was manually canceled
- canceling: Simulation is in the process of being canceled
- expired: Simulation exceeded its time limit
Example Simulation
Scenario Design
Effective scenarios provide clear context for conversations:Support Scenarios
“Customer experiencing login issues after recent password reset”
Sales Scenarios
“Prospective customer interested in enterprise pricing for team of 50”
Technical Scenarios
“Developer struggling with API integration and receiving timeout errors”
Onboarding Scenarios
“New user setting up their first project and configuring team permissions”
Planning Your Simulation
Determine Conversation Volume
- Quick Test: 10-20 conversations for basic functionality validation
- Standard Evaluation: 50-100 conversations for reliable metrics
- Comprehensive Assessment: 200+ conversations for statistical significance
Set Realistic Targets
- Consider your agent’s response time when setting conversation targets
- Factor in evaluation time if using manual scoring
- Plan for potential failures or retries
- Ensure persona count supports your target conversation volume
Choose Turn Limits
- Short Interactions: 3-5 turns for quick queries
- Standard Support: 10-15 turns for typical problem resolution
- Complex Issues: 20+ turns for detailed troubleshooting
Best Practices
Use descriptive scenario text that provides clear context for both persona generation and conversation flow.