By Question Evaluation Metrics¶
The Evaluation Metrics feature is a core component of the Quality AI module, allowing supervisors to define, tailor, and monitor performance indicators that assess the quality of interactions between agents and customers. This process is driven entirely by individual questions. Users can create and customize evaluation criteria using various measurement types, which are managed within the Evaluation Forms section.
This By Question configuration evaluates adherence to specific questions asked or answered during customer-agent interactions.
What It Offers For Supervisors¶
-
Standardized quality assessment framework.
-
Customizable evaluation criteria based on business needs.
-
Multi-language support for global operations.
-
Automated evaluation suggestions through AI.
When to Use This Metric¶
Primary Use Cases:
-
Quality Assurance: When supervisors need to systematically evaluate agent adherence to specific protocols, scripts, or expected responses.
-
Training Assessment: To measure how well agents follow prescribed interaction patterns during customer conversations.
-
Compliance Monitoring: For ensuring agents deliver critical information like disclaimers, privacy policies, or regulatory statements (privacy policies, disclaimers).
-
Performance Standardization: When you need consistent evaluation criteria across different agents and interactions.
How It Works¶
The metric operates through a question-driven evaluation process with two main approaches:
Static Adherence¶
-
Evaluates agent responses across all conversations without conditional requirements.
-
Best for universal standards like greetings or standard procedures.
-
No trigger conditions needed.
Dynamic Adherence¶
-
Evaluates agent responses only when specific triggers occur.
-
Trigger-based detection using customer or agent utterances.
-
Ideal for conditional scenarios where specific responses are required only in certain contexts.
Configure By Question Metrics¶
-
Navigate to Contact Center AI > Quality AI > Configure > Evaluation Forms > Evaluation Metrics.
-
Click + New Evaluation Metric.
-
From the Evaluation Metrics Measurement Type dropdown, select By Question.
-
Enter a descriptive Name for future reference of the metrics. (for example, agent’s warm greeting).
-
Select a preferred language from the dropdown to evaluate the form and configure the required metric.
-
Enter an evaluation Question to help supervisors assess agent adherence as a reference for the audits and interaction reviews.
-
Select the agent Adherence Type (Static or Dynamic).
Note
-
For Static, you must configure at least one agent answer utterance for the Adherence Type.
-
For Dynamic, you must configure at least one trigger and one agent answer utterance for the adherence type.
Adherence Type Configuration¶
The following are the key trigger components that you can configure:
-
Customer Utterance: Evaluation initiated by customer statements.
-
Agent Utterance: Evaluation triggered by agent responses.
-
Multiple Triggers: Support for complex conditional scenarios.
-
GenAI Suggestions: Automated variation generation.
Static Adherence¶
Static Adherence measures whether agents say the required phrase, regardless of what triggers the conversation. It applies to all calls, without any condition or contextual trigger.
-
Use this adherence when you need a consistent, universal check (for example, mandatory greeting scripts, regulatory disclaimers).
-
You should define acceptable utterances for a queue.
-
You set a similarity threshold to evaluate whether the agent’s actual response matches the pre‑defined acceptable utterance.
-
No triggers or contextual conditions needed; it is ideal for fixed, non‑situational compliance items.
-
You must configure at least one agent utterance template for this adherence type.
Dynamic Adherence¶
Dynamic Adherence is a context-sensitive method to evaluate agent performance based on specific triggers rather than monitoring every interaction. It is ideal for scenarios where agent behavior should be assessed only when certain conditions are met.
-
Adherence is checked only when a configured trigger (agent or customer utterance) is detected.
-
You must define at least one trigger (either customer or agent utterances) and one acceptable agent response that activate the adherence check.
-
Evaluates agent behavior only in the context of the detected trigger.
-
Determines how closely an agent's response must match a predefined response.
-
Adjustable by criticality of the adherence similarity use case:
-
Lower Threshold: Should close to 60% (Yellow) for casual interactions, greetings.
-
Higher Threshold: Must close to 100% (Green) for critical topics, such as legal disclaimers or privacy policies.
-
Once a trigger is detected, the system evaluates whether the agent responded with one of the pre-defined acceptable utterances.
Note
For Dynamic, configure at least one Trigger and one agent Answer utterance for the adherence type.
-
-
-
Choose utterance source for Trigger (who initiates the trigger).
Trigger Configuration¶
Provides the following two options to select based on the trigger created by Agent Utterance or Customer Utterance for evaluation.
-
Trigger: Select and evaluate responses based on triggers created from agent or customer utterances.
-
You can add multiple trigger utterances for conditional checks (for example, AND/OR conditions) to serve as the trigger condition or refine when the trigger is activated.
-
You can define a trigger for either a customer or agent utterance and configure the corresponding responses for that scenario.
-
Customer Utterance: Configure the Customer Utterance that triggers the adherence check. You can enter more than one utterance using Generative AI Assistants that are similar to utterances with the same meaning. For example, customer says “I need a refund”.
-
Agent Utterance: Configure the Agent Utterance for triggers initiated by the agent. Enter the utterances using generative AI Assistants’ suggestions that have similar utterances with the same meaning. You can add or delete multiple utterances for the Customer and the Agent. For example, agent says “let me transfer you to support”.
-
-
-
Choose a Trigger Detection Method.
-
GenAI-Based Adherence: Enter a descriptive prompt to define intent. This allows contextual detection of whether the agent’s response aligns with the intended goal, without relying on predefined samples. This uses LLMs to understand context and intent and do not require sample utterances or thresholds.
-
Deterministic Adherence: This relies on predefined sample utterances and detects adherence based on semantic similarity.
-
Utterance: Enter sample training utterances to detect adherence and define similarity thresholds accordingly.
Enablement of GenAI-Based Features (Pre-requisite)¶
The GenAI-based features are activated only when the following conditions are enabled:
a. Navigate to Manage> Generative AI> GenAI Features.
-
Choose an Agent Answer adherence type.
Agent Answer¶
-
GenAI-Based Adherence: Use AI and natural language understanding (LLM) to detect meaning, context, and intent. This evaluates whether the agent's answer fulfills the intents, even if phrased differently.
-
Deterministic Adherence: This relies on an ML-based method using the semantic similarity.
-
Description: Enter a prompt Description to check the agent's intent accuracy, phrasing response, and adherence.
GenAI-Based Adherence¶
Uses Generative AI to automatically evaluate agent responses by understanding natural language, including intent and context, even when phrased differently.
Agent Answer
a. Select an Answer Detection Method to evaluate whether agents respond according to the prompt’s intent. This uses a probabilistic LLM-based method and requires no model training.
b. Enable the GenAI-Based Adherence. Use a Large Language Model (LLM) to detect trigger phrases and evaluate adherence using contextual understanding.
c. Enter a prompt Description explaining the metric’s intent or details behind the adherence metric. This applies to all selected languages.
Note
-
Before assessing GenAI-based adherence responses, ensure that the supported model and features, including GenAI-based agent answer adherence and customer trigger detection, are enabled for the respective products in the GenAI features section. Learn more.
-
No example utterances or similarity thresholds are required; LLMs evaluate adherence contextually using zero-shot prompts.
-
For effective prompts and LLM-based adherence detection, Learn more.
Deterministic Adherence¶
Evaluates agent responses based on semantic similarity to predefined sample utterances or answers.
Agent Answer Configuration¶
a. Select Deterministic Adherence to assess responses based on similarity to sample answers. Encoder-based mode lets you define expected replies.
b. Define an Answer as a set of acceptable utterances for each queue, using Generative AI to generate the following automated response variations:
-
Set a similarity threshold to determine how closely user input must match expected utterances to get agent answers.
-
Add language-specific, prompt-based evaluation parameters.
-
Helps you find relevant answers by suggesting different ways to ask the same question.
-
Provides expected answers that match the meaning of your question, even if it is asked in different ways.
-
Delete AI-suggested answers that are not required.
- Set the Similarity percentage for the metric based on the defined use case and attribute.
Similarity Thresholds¶
Evaluates agent responses based on semantic similarity to predefined sample utterances or answers.
-
You can set the Similarity percentage for the desired evaluation metrics. Whether it is Static or Dynamic, you can configure the expected Similarity threshold.
-
You should set a lower adherence similarity threshold (for example, 60%) for soft skills like greetings and etiquette, and a higher adherence similarity threshold (for example, 100%) for compliance-critical (Policy Privacy or Disclaimer) statements, because it is critical for the user to follow the adherence depending on the use cases.
Note
The Similarity threshold option is available only when GenAI-Based Adherence is enabled. You can configure thresholds for both Static and Dynamic evaluations.
-
-
Choose a Count Type based on the selected Adherence Type.
Count Type Configuration¶
The following are the key count types that you can configure:
- Entire Conversation: Full interaction evaluation
- Time Bound: Specific time ranges or message counts
Entire Conversation¶
Evaluates adherence throughout the complete interaction. This allows you to check for adherence at different points of conversation. It does not matter where the agent wants to check adherence throughout the conversation.
Time Bound¶
Focuses on specific timeframes (first or last X seconds or messages). This evaluates adherence within a specific time range or number of messages in the interaction. It can occur at the start or end of the conversation, either for a defined number of seconds or a set number of chat messages.
-
Parameter: Choose between the First Part of Conversation or the Last Part of Conversation, and configure the relevant criteria to evaluate the metric.
-
Voice: Enter the number of seconds from the start or end of the interaction to evaluate this metric.
-
Chat: Enter the number of messages from the start or end of the interaction to evaluate this metric.
-
Click Create to save and activate the By Question configured adherence metric.
Managing Evaluation Metrics¶
Edit or Delete Evaluation Metrics¶
Steps to edit or delete existing Evaluation Metrics:
-
Right-click to select any of the existing Evaluation Metrics.
-
Click Edit to update the required Edit By Question Metrics dialog box fields.
-
Click Delete to remove the selected By Question evaluation metric.
-
Click Update to save the changes.
Language Dependency Warnings¶
This section outlines the limitations and dependencies associated with modifying language settings in evaluation metrics.
Modification Warnings¶
-
You cannot remove a language if any evaluation form or attribute currently uses it.
-
Remove the language from all associated evaluation forms and attributes before modifying their language settings.
-
You can safely remove languages that are not linked to any forms or metrics.
Delete Warnings¶
This section describes the warnings and prerequisites you must address before deleting a metric.
-
If the metric is used in any evaluation form, the system displays a warning message.
-
Remove the metric from all associated evaluation forms before you delete it.
-
If any attributes are linked to the metric, assign a different metric to those attributes before proceeding with deletion.
-
The system allows you to delete the metric only after resolving all dependencies.