Overview
Cekura is a simulation and production monitoring platform for voice AI agents. The broad idea is to use the same evaluation system across the full lifecycle of a Pipecat agent: simulate conversations before deployment, validate infrastructure continuously in CI, monitor real calls after deployment with metrics, dashboards, and alerts, and turn production issues back into new simulations so you can fix regressions and verify them before shipping again.- Treat simulations and monitoring as one continuous QA workflow
- Connect to Pipecat through telephony, SIP, or native Pipecat WebRTC flows
- Run scenario-based tests from the dashboard or API with automated or manual session control
- Track quality with predefined metrics and custom metrics tailored to your agent’s workflow
Simulations
Testing in Cekura is built around scenarios, metrics, and repeatable runs. The main design choice is how you want Cekura to connect to your Pipecat agent:- Use telephony or SIP when you want to validate the full voice and network path
- Use native Pipecat connections when you want faster, more direct WebRTC-based testing
Automated Pipecat connection
For Pipecat Cloud agents, configure the following fields in Cekura:- Provider: Select
Pipecatas the voice integration provider - Assistant ID: Optional Pipecat assistant identifier
- Pipecat Cloud API Key: Used by Cekura to create and manage sessions
- Pipecat Agent Name: Identifier used for your agent in Pipecat
- Agent Configuration (JSON): Optional request body sent to Pipecat’s start endpoint
- Room Properties (JSON): Optional Daily room configuration for the WebRTC session
Automated Pipecat Testing
Configure Pipecat Cloud credentials once and let Cekura create and manage
sessions for each test run.
Manual Pipecat Testing
Provide your own room URL and token when you want explicit control over
Pipecat session creation and lifecycle.
Manual Pipecat connection
If you want to control the session yourself, Cekura also supports a manual Pipecat connection flow. In this mode, you provide the Pipecat room URL and token for each run, and Cekura joins the session you created instead of provisioning one automatically. This mode is useful when you need custom session orchestration, want to test nonstandard environments, or already manage Pipecat rooms in your own infrastructure.The automated flow is the best default for Pipecat Cloud. Use the manual flow
when you need explicit control over room creation and access tokens.
Infrastructure Suite
For infrastructure-focused validation, Cekura provides an Infrastructure Suite with 18+ pre-built scenarios for latency, audio quality, interruption handling, language support, and failure cases such as packet loss and rapid-fire speech. Instead of building these tests from scratch, you can add the suite to your project from the Cekura dashboard and run the scenarios as a group. This makes it a good fit for validating the parts of your stack around the LLM itself, such as transport behavior, turn-taking, silence handling, and voice pipeline stability.Pre-Built Infra Tests
Add Cekura’s predefined infrastructure scenarios to your project with the
appropriate metric mappings already set up.
GitHub Actions CI
Run tagged infrastructure scenarios automatically on pull requests, pushes,
or scheduled workflows.
infrastructure-suite and execute them on every pull request:
Production observability
Cekura can also monitor production calls after they complete. For Pipecat, the core pattern is simple: send completed calls to Cekura, use metrics and dashboards to track quality, infrastructure health, and workflow performance over time, then turn failed or interesting real calls into new simulations so the same issues are covered in your future test suite.Send calls to Cekura
If you want observability for Pipecat traffic, send recordings and transcripts to Cekura’s observability API:transcript_type: "pipecat" along with either an agent ID or assistant_id, plus a voice_recording or voice_recording_url.
Example: send a Pipecat production call
Metrics and performance insights
Cekura’s metrics system can be used in both simulations and production monitoring. For reference, see the metrics overview, predefined metrics, and custom metrics. Relevant predefined metrics for production monitoring include:- Latency with percentile statistics for response-time monitoring
- Infrastructure Issues to detect failures to respond within a configured timeout
- Transcription Accuracy for STT quality checks
- AI Interrupting User and Stop Time After User Interruption for turn-taking quality
- Gibberish Detection, Average Pitch, and other speech-quality metrics for audio analysis
- CSAT, Sentiment, and Dropoff Node for agent-performance and user-experience insights
Next steps
Pipecat Integration Guide
Step-by-step setup for running automated Pipecat tests in Cekura.
Testing Overview
Start with Cekura’s broader guide to scenario creation, execution, and testing
workflows.
Observability Overview
Learn how Cekura structures production monitoring, observability, and call
review workflows.
Metrics Overview
Review predefined metrics and custom metrics for workflow, logic, speech,
and infrastructure evaluation.