# Evaluators, Guardrails & Conversation Logs

## Evaluators & Guardrails

### 📌 Overview

**Evaluators** (or evals) are automated tools in indigo.ai that **analyze conversations**.\
They help measure key aspects such as response relevance, user sentiment, tone, and topics, providing objective, timely, and scalable evaluations without relying on manual reviews.

Evaluators work together with **Guardrails**, which act as **preventive checks** during live conversations.&#x20;

Together, they provide a complete framework for monitoring, improving, and governing the virtual assistant's quality.

{% hint style="info" %}
From a technical perspective, evals run [LLM models](https://guide.indigo.ai/getting-started/ai-knowledge-hub/large-language-models-llms-available-on-our-platform) *after* a response to check its quality and correctness, while guardrails run *before and during* generation to constrain what the virtual assistant can or cannot say.

In simple terms:

* **Evals = post-checks** ✅ (evaluate answers after they’re produced).
* **Guardrails = pre-checks + live constraints**🚦 (control behavior before and during the reply).
  {% endhint %}

<figure><img src="https://4264750556-files.gitbook.io/~/files/v0/b/gitbook-x-prod.appspot.com/o/spaces%2F4dIFGOeVbHW1PRwnG0Z8%2Fuploads%2FMRHnpCOfw5kLWdusESvs%2Fadd%20evaluator%20step%201.png?alt=media&#x26;token=9507da6f-98a6-45ae-bcdf-19943e63ad4d" alt=""><figcaption></figcaption></figure>

### ✅ Benefits

* **Objective and fast evaluations** → continuous quality monitoring, independent from human judgment.
* **Operational efficiency** → reduce manual effort, freeing resources for higher-value tasks.
* **Trend and pattern detection** → uncover recurring issues, user sentiment trends, or escalation needs.
* **Improved perceived quality** → proactive monitoring strengthens brand reputation and service trustworthiness.

### Built-in vs Custom

indigo.ai provides both built-in evaluators and guardrails (ready-to-use “black boxes”), and the **ability to design custom ones for specific use cases**.

{% hint style="info" %}
For more details read the full guide: [evaluators-and-guardrails](https://guide.indigo.ai/getting-started/workspace/utilities/evaluators-and-guardrails "mention")
{% endhint %}

## Conversation Logs

### 📌 Overview

Conversation Logs provide a centralized place in the indigo.ai platform where you can view the **full list of conversations and their corresponding Evaluator or Guardrail results**.

This section allows teams to analyze every single conversation, understand how evaluators and guardrails performed, and start any necessary **debugging** flows.

<figure><img src="https://4264750556-files.gitbook.io/~/files/v0/b/gitbook-x-prod.appspot.com/o/spaces%2F4dIFGOeVbHW1PRwnG0Z8%2Fuploads%2FSyY0i1FJVUSWfrXEVWtD%2Fconversation%20log%203.png?alt=media&#x26;token=9a062053-f15f-42f6-8455-71d025e7d54d" alt=""><figcaption></figcaption></figure>

### ✅ Benefits

* **Detailed visibility** → review all conversations in one place, with full evaluator and guardrail outcomes.
* **Operational efficiency** → quickly identify issues, user feedback, or API errors without switching tools.
* **Debug-ready** → inspect conversations in detail, including errors and triggers, to improve the chatbot’s performance.

### 💬 What is a Conversation Log?

A Conversation Log represents one interaction between an AI Agent and an end-user.

* A conversation begins when the user starts the chat and ends when they click Close Chat.
* Multiple conversations together form a chat.

In the Conversation Logs view, you can browse the list of conversations and check how evaluators and guardrails performed for each one.

{% hint style="info" %}
For more details read the full guide: [conversation-logs](https://guide.indigo.ai/getting-started/workspace/utilities/conversation-logs "mention")
{% endhint %}
