Skip to main content

The Simulation & Evaluation Engine for Voice AI

Roark is the evaluation layer for voice AI agents. Monitor every production call in real-time, run metrics powered by Roark Prism — our purpose-built evaluation model — and stress-test agents with simulations before they reach customers.

How It Works

1

Get calls into Roark

Connect a voice platform (Vapi, Retell, ElevenLabs, LiveKit, others) and calls sync automatically. Or upload recordings directly via the SDK.
2

Calls are analyzed automatically

Every call is transcribed, then run through Roark’s voice analysis models — detecting sentiment, 64+ emotions, interruptions, speech pauses, and vocal cues. Active metric policies collect any additional metrics you’ve configured.
3

Evaluate with metrics and thresholds

Define what good looks like. Use built-in metrics like response time or create custom evaluations with Roark Prism (e.g., “Did the agent verify the caller’s identity?”). Add thresholds to turn values into pass/fail outcomes.
4

Test before you ship

Run simulations against your agent with synthetic callers. Attach metrics with thresholds to your run plan — if the agent doesn’t meet your bar, you know before customers do.

Quick Start

Send a call recording and let Roark handle the rest:
import Roark from '@roarkanalytics/sdk'

const client = new Roark({
  bearerToken: process.env.ROARK_API_BEARER_TOKEN,
})

const call = await client.call.create({
  recordingUrl: 'https://example.com/recording.wav',
  startedAt: '2025-01-15T10:00:00Z',
  interfaceType: 'PHONE',
  callDirection: 'INBOUND',
  agent: { name: 'Support Agent' },
  customer: { phoneNumberE164: '+15551234567' },
})
The call appears in Call History with full transcription, sentiment, and emotion analysis. Any active metric policies run automatically.

Explore the Docs