Creating an AI Agent

Prev Next

Overview

Feature Details
Introduced in Version9.20
Modified in Version9.23
LocationManage < Integrations < AI < Agents

This is a step-by-step guide to creating an AI Agent inside the Decisions Platform.

What is an AI Agent?

An AI Agent is a specific configuration of an AI interaction that involves tool usage. Tool usage allows AI to perform code execution on the application that called the LLM. 

For example, an AI cannot send an email. With the use of AI tools in conjunction with AI itself, you can enable the LLM to send an email. It can get much more complex than that. By taking the complex cognitive thinking capabilities of AI and a workflow application, AI can be empowered to execute workflows inside the Decisions application to be connected to data that is outside of its reach and access to actionable capabilities to empower an LLM to be functional, not just informational.


Agent-to-Agent Communication

Decisions supports defining relationships between Agents by allowing a primary Agent to be associated with one or more sub-agents. This enables coordinated task execution across multiple Agents.

When configured, the primary Agent evaluates incoming requests and determines whether a task should be handled directly or delegated to a sub-agent. This behavior is influenced by:

  • The system prompt defined for the primary Agent
  • The system prompts defined for each sub-agent
  • The input prompt provided to the primary Agent

Sub-agents execute their assigned responsibilities and return results to the primary Agent, which then compiles and returns the final response.

This approach allows complex workflows to be broken into smaller, specialized tasks. For example, a primary Agent responsible for sending emails may delegate user retrieval to a sub-agent that manages user data.

Note
Sub-agents cannot operate independently and must be associated with a primary Agent. Deleting a sub-agent that is actively linked within a hierarchy may be restricted or require reassignment.


Use Case

Example Use CasesDescription
Sales EnablementBy adding tools to provide data to an LLM, AI can be used to interpret and answer questions regarding data that is important to the Sales team. Does sales management need a way to quickly get answers around upcoming planned deals? Connect an LLM (using an Agent) to your sales pipeline data, and now questions get answered quickly.
Informational ChatbotsUsing tool flows to provide chatbots with more information. This will provide the responses from LLMs to be more consistent and correct. 

Example

This section demonstrates how to create and configure an AI Agent.

  1. Create a System Prompt

    Navigate to the System Prompt folder (Manage > Integrations > AI > System Prompts) to create a system prompt. This is used to create context for the AI interaction. As a simple example, a system prompt may define behavior such as assisting with sending emails or controlling tone.

    System prompts can influence both the functional behavior of the Agent and the end-user experience. Prompt groups can also be used to organize prompts, including a Default Prompt Group. 

    Adding Dependencies to System/User Prompts
    In v9.23+, Dependencies have been added to System/User Prompts. This will prevent System prompts, chat prompts, and agents that are used by the chat completion step from being deleted.

    System Prompting 101
    A good AI Agent implementation depends on system prompts and information access. Prompts can include rules and directives to control behavior and improve consistency.

  2. Create Tool Flows for the Agent

    An AI Tool flow is an executable flow used by the Agent to complete tasks. These flows include descriptions that are sent to the LLM to help it understand how to use the tool.

    Tool flows require a response message, which can return data or confirm that an action has been completed.

    Function Calling = Decisions Flow Execution
    AI Agents rely on function calling to execute workflows. The process follows: Decisions -> LLM -> Workflow Execution -> LLM -> Decisions.

    The following example tool flows are used:

    • team-email-fetch: Retrieves a distribution email based on department name
    • send-email: Sends an email using address, subject, and body
  3. Configure the Agent

    Navigate to the Agents folder (Manage > Integrations > AI > Agents) to create and configure an AI Agent. Configure the Agent to use the appropriate system prompt and tool flows.

  4. Execute the Agent

    Agents can be executed using the Chat Completion step in two ways:

    • Configure the Agent directly within the step (system prompt, user prompt, tools, provider, model). This method would allow for complete flexibility in the AI Agent.
    • Use the Use Agent option to select a pre-configured AI Agent.
    • Example request: “Please send an email to the IT team telling them that their outfits look good today.”
      The Agent will:
    • Call the team-email-fetch tool to retrieve the email address
    • Call the send-email tool to send the message

Feature Changes

DescriptionVersionRelease DateDeveloper Task
Dependencies have been added to System/User Prompts. This will prevent System prompts, chat prompts, and agents that are used by the chat completion step from being deleted. 
9.23April 2026[DT-047210]
Model Restrictions
With this technology moving as fast as it is, you might experience some limitations/issues with using certain models. Currently, our AI Agents do not work for OpenAI GPT-5. This is the only limitation we are aware of at the moment.