A simulation is a simulated conversation between our agent and your voice or chat agent. You can define the environment on how to test your agent within test sets and Templates. Metrics define the success or failure criteria for your tests.

Once you’ve run evaluations through simulating conversations, you will see them in:

Runs

A Run is an evaluation. A Run can consist of multiple conversations (e.g., if the test set consists of multiple scenarios/transcripts).

You can also schedule runs, which will then be shown under “Scheduled”

Overview

The Overview tab consists of all individual conversations. It helps you get an overview of your agent’s performance by creating your own summary graphs and see aggregated performance over time.

Review

Use Coval’s Human-in-the-loop review capabilities to label runs for review.