Agent Node: Frequently Asked Questions¶
Language Models¶
What types of language models are available on the XO Platform?¶
The XO Platform offers three types of language models:
- System or Out-of-Box Integrations: Includes models from providers like OpenAI, Azure OpenAI, and Anthropic.
- Custom models (Bring Your Own Models): Allows users to bring their own models to the platform.
- Fine-tuned XO GPT models: Used for specific features like DialogGPT, Answer Generation, Query Rephrase, Rephrase Dialog Response, and Vector Generation.
How do I configure out-of-box (OOB) integrations?¶
To configure OOB integrations:
- Use your API keys for models such as OpenAI, Azure OpenAI, or Anthropic.
- The platform offers various models, but you can also connect to new ones by entering the correct model names and API key.
- The platform verifies these by calling the provider's endpoints.
- OOB system models come with default prompts, while dynamically added models require custom prompts.
How do I set up a custom model (Bring Your Own Model)?¶
To set up a custom model, follow these steps:
- Enter the integration name and model name (customizable to your preferences).
- Provide the endpoint where your models are hosted and the required authentication and headers. Create a test prompt to verify the credentials.
- The platform will validate the credentials and configure the model.
- Once configured, the custom model can be utilized with tailored prompts for various GenAI and LLM features.
Are XO GPT models compatible with Agent Nodes?¶
No. XO GPT models are not compatible with Agent Nodes. You'll need to choose from other available system or custom models.
Tools¶
What are the tools in the Agent Node?¶
Tools are special capabilities you can give to language models that allow them to interact with external systems. They enable the LLM to reach beyond its own knowledge and interact with the outside world during a conversation.
Which LLM providers support tool calling?¶
All major LLM providers support some version of tool calling, including:
- Anthropic
- Cohere
- Mistral
- OpenAI
- Azure OpenAI
What are the key components of a tool?¶
There are five key components to creating and using tools:
- Name: A clear and descriptive name helps the LLM understand the tool's purpose.
- Description: The most crucial part; it explains what the tool can do, when to use it, and any limitations.
- Parameters: The specific pieces of information necessary for your tool to function.
- Action Nodes: The components that execute when the tool is called, including script nodes, service nodes, or search AI nodes.
- Transitions: Defines what happens after the tool executes, such as Default, Exit Out, or Jump to a Node.
What transition options are available after a tool executes?¶
The XO Platform provides three options:
- Default transition: Sends the output back to the language model for incorporation into the conversation.
- Exit Out: The agent node completes, and the flow advances to the next node in your dialog.
- Jump to a Node: Redirects the conversation flow to a specific node within your dialog.
Prompts¶
What types of prompts does the XO Platform offer?¶
The XO Platform offers two main categories of prompts:
- System or Default Prompts: Ready-made prompts for each feature come with the platform.
- Custom Prompts: These prompts you create yourself, tailored to your specific needs.
What versions of custom prompts are available?¶
For the agent node specifically, there are two different versions of custom prompts:
- Version 1: Enables multi-turn conversations with or without tools (using a more traditional prompt structure).
- Version 2: A newer, more advanced prompt framework that supports tools (enhances accuracy and simplifies agent development).
Can I use tool calling with system prompts?¶
No, all system prompts for the agent node are categorized under Version 1 and do not permit tool-calling capabilities. If you require those advanced features, you must create a custom prompt.
How do I create a custom prompt?¶
To create a custom prompt:
- Go to the Prompts Library.
- Choose to create a New Prompt.
- Decide whether to start fresh or use a template.
- Select the appropriate version based on your needs.
- Use the editor to craft your instructions.
- Define the necessary output keys and test your prompt.
Agent Node Components¶
What are the main components of an Agent Node?¶
The main components of an Agent Node include:
- Language Model Selection: Select the model that will power your agent node.
- Prompts: Guidelines that direct your language model's behavior and responses. System Context: Defines your agent's role, communication style, and operational parameters.
- Entities: Information that your agent needs to gather from users.
- Business Rules: Guidelines that your agent must follow when interacting with users.
- Tools: Allow your LLM to engage with external resources.
- Exit Scenarios: Specify when to alter direction in the conversation.
- Pre/Post-Processor Scripts offer greater control over your agent's behavior.
What is system context?¶
System context is your agent's comprehensive framework that defines:
- Identity and purpose
- Behavioral guidelines
- Processing parameters
- Role and communication style
- Knowledge domains
- Response patterns and decision-making
- Operational workflows
- Context evolution throughout conversations
What are business rules?¶
Business rules are guidelines your agent follows when interacting with users. They define:
- The scope of what your agent can and should do.
- Validation logic and process requirements.
- Data quality standards.
- Compliance with organizational policies.
What are exit scenarios?¶
Exit scenarios refer to conditions that enable you to determine when the conversation should take a different path, such as:
- When a user expresses frustration.
- When you've gathered enough entities to proceed.
- When a user explicitly requests to end the conversation.
- When a user provides invalid information multiple times.
- When a user asks to speak with a human agent.
Version Differences¶
What's the difference between Version 1 and Version 2 prompts?¶
- Version 1: The legacy prompt system functions effectively for straightforward tasks but has certain limitations. It supports both tool calling and text generation.
- Version 2: The newer system is designed around tool calling functionality. It integrates better with other systems and manages more complex workflows. However, it only works with JavaScript prompts, not JSON mode.
How does entity collection differ between Version 1 and Version 2?¶
- Version 1: Requires separate definitions of entity collection logic and manages the complexity of mapping entities to business processes.
- Version 2: Integrates entity collection directly into the tool-calling framework. Instead of juggling separate systems, you define custom tools with parameters that naturally capture the entities needed.
What types of tools does Version 2 include in its prompts?¶
Version 2 provides two types of tools:
- System tool: Included by default. Includes "End_Orchestration" to help the platform understand when to end the conversation with the language model.
- Custom tools: Entirely defined by you based on your specific use cases.
Implementation¶
How does the execution flow work with pre and post-processor scripts?¶
When configuring pre- and post-processor scripts at both the node and prompt levels, the execution order is: Node Pre-processor → Prompt Pre-processor → Prompt Execution → Prompt Post-processor → Node Post-processor.
Can I import templates for custom prompts?¶
Yes, the XO Platform provides templates that you can use as a starting point. Version 2 includes a template specifically designed for GPT-4o, which comes pre-configured with a post-processor script and all the necessary output keys.
How does the conversation history function with Agent Nodes?¶
You can configure the duration of the conversation history, which determines how many of the most recent messages exchanged between the user and the bot are sent to the language model. This allows the model to maintain context throughout the interaction.
How can I test my custom prompt?¶
After setting up your prompt and entering test values for any variables, click "Test" to verify if the structure aligns with expectations. This will display the response from the language model and assist you in debugging any issues.