Core platform components
Tars operates through six main components that work together to deliver intelligent conversational experiences:AI Agents
- Intelligent Conversation Handlers. Human-like conversation understanding and context awareness.Automatic tool selection and workflow execution
Knowledge Bases
- Intelligent Information Retrieval
- Vector-based semantic search using Weaviate and Qdrant
- Multiple content source support with automatic processing
Tools Integration
- External System Connectivity. 600+ pre-built integrations for business platforms
Data Collector
- User Query Extraction. Extracts relevant information from user queries
Categorizer
- Query Categorization. Categorizes user queries into different branches.Facilitates targeted response generation.Improves conversation flow management
User-input
- Flexible Input Collection. Supports diverse input types including text, star ratings, buttons, interactive cards, auto-suggestions, calendar/time selection, file uploads, geolocation, language selection, and auto page redirection.
How does an AI Agent work in Tars?
AI Agents in Tars use advanced language models (LLMs) to intelligently manage conversations. The LLM dynamically decides whether to retrieve knowledge, execute tools, or generate direct responses based on user input:How does a Question Answering Agent work in Tars?
A Question Answering Agent is a specialized agent optimized for information retrieval. It uses RAG (Retrieval-Augmented Generation) technology to search your knowledge base and provide accurate, context-aware answers:1
User input
User asks a question or requests information through any interface (web, WhatsApp, direct link).
2
Retrieve
Performs vector database search using semantic similarity to find relevant
knowledge from the Knowledge Base.
3
Augment
Combines retrieved knowledge with conversation context and user history for
comprehensive understanding.
4
Generate
Creates accurate, knowledge-based responses with source attribution and contextual relevance.
How does the Knowledge Base work in Tars?
The Knowledge Base System transforms unstructured content from various sources into intelligent, searchable knowledge that your AI Agent can access instantly. The system operates in three phases:Load
-> Index
-> Retrieve
.
Load Phase
Data Ingestion Connect to 12+ data sources including websites, PDFs, Slack, Notion, Confluence, and more. Content is extracted, cleaned, and stored for processing.
Index Phase
Embedding Creation Documents are split into optimal chunks and converted into vector embeddings using OpenAI’s models, then stored in dual vector databases (Qdrant & Weaviate).
Retrieve Phase
Semantic Search User queries are converted to vectors and matched against stored embeddings. AI-powered reranking returns the top 4 most relevant results in under 500ms.
Knowledge processing workflow
1
Content ingestion
Your data sources (websites, PDFs, Slack channels, Notion pages, etc.) are connected and content is extracted using specialized loaders. Raw content is cleaned and prepared for processing.
2
Text chunking & embedding
Documents are split into 1,000-character chunks with overlap to maintain context. Each chunk is converted into a 1,536-dimensional vector using OpenAI’s embedding model, capturing semantic meaning.
3
Vector storage
Generated embeddings are stored in both Qdrant and Weaviate vector databases with metadata (source URL, document ID). Multi-tenant isolation ensures your data remains private and secure.
4
Query processing
When users ask questions, queries are converted into the same vector format and compared against millions of stored vectors using semantic similarity search.
5
AI reranking & response
Retrieved chunks are reranked using Cohere’s AI model to evaluate relevance. The top 4 results are returned with source attribution, enabling your Agent to provide accurate, contextual answers.
Fast & scalable: The system processes 500-1,000 chunks per minute during indexing and delivers search results in under 500ms, supporting millions of vectors across multiple knowledge bases.
How tool execution is done in Tars?
Tool execution at Tars is managed through a configured AI Agent, which interacts with connected external systems securely and intelligently. The architecture ensures that all actions, authentications, and data exchanges happen entirely within backend systems, keeping credentials and sensitive information fully protected.Tool execution architecture
Backend-Only Execution
Secure Processing All tool interactions and processing occur exclusively within backend systems. Tool credentials remain encrypted in backend databases and are never exposed to frontend clients.
AI-Driven Intelligence
Smart Tool Selection The AI Agent automatically determines which tools to use based on user intent and connected tool capabilities, executing actions within the LangGraph workflow.
Encrypted Communication
End-to-End Security Every payload transmitted between services is encrypted, and the AI Agent can only execute tools that are explicitly connected and authorized.
Tool execution workflow
1
User interaction
The user interacts with the CX Agent interface, submitting queries or requests that may require tool execution.
2
Backend parsing & payload preparation
The Chatbot Backend receives the user input, parses it, and prepares a structured payload. All sensitive data within the payload is securely encrypted before being sent to the AI Service Backend.
3
AI Agent execution
Inside the AI Service Backend, the configured AI Agent processes the request. If the agent has connected tools, it executes the appropriate tool action within the LangGraph workflow. If no tools are connected, the agent continues reasoning and generates a response without performing any tool action.
4
Response handling
The AI Service Backend processes the tool output (if any) and formulates the final agent response. This response is securely returned to the Chatbot Backend, which parses and extracts only the necessary information for display.
5
User display
The processed response is sent to the CX Agent interface, where it’s shown to the user. Throughout the flow, raw credentials or sensitive data are never exposed.
Security first: Tool execution at Tars is secure, backend-driven, and fully agent-controlled. Each execution is logged for transparency, monitoring, and governance, ensuring that sensitive data never leaves trusted backend environments.
How does WhatsApp integration work in Tars?
TARS enables businesses to deploy AI Agents on WhatsApp through a seamless integration architecture. The system abstracts provider differences, normalizes message formats, and ensures reliable, sequential message delivery while maintaining conversation context and state.WhatsApp integration architecture
The integration is built on three key abstraction layers that work together to deliver AI-powered conversations on WhatsApp:360Dialog Provider
WhatsApp API Integration Receives webhooks from WhatsApp Business API and delivers messages back to users.
Tars Chatbot Backend
Message Orchestration Handles API routing, validation, credit checking, message transformation (inbound/outbound), conversation engine processing, and response delivery.
AI Service
AI Agent Processing Handles AI Agent dependent requests from the Chatbot Backend, processes with AI logic, and returns intelligent responses.
WhatsApp integration workflow
1
Webhook reception
A user sends a message via WhatsApp. The message is received by 360Dialog and forwarded to the Tars Chatbot Backend via webhook.
2
Validation & transformation
The Tars Chatbot Backend validates credentials, checks credits, and transforms the WhatsApp message into Tars internal format.
3
Conversation processing
The Conversation Engine (within Chatbot backend) processes the message, executes action functions, and determines the next gambit. If AI Agent processing is needed, the request is sent to the AI Service.
4
AI processing (if needed)
The AI Service handles AI Agent dependent requests, processes with AI logic, and returns the intelligent response back to the Tars Chatbot Backend.
5
Response transformation & delivery
The Tars Chatbot Backend transforms the response into WhatsApp-compatible format (including buttons, lists, media), queues it in Redis for sequential delivery, updates conversation state in MongoDB, and sends it to the user via 360Dialog and WhatsApp.
Seamless integration: The system supports multi-turn conversations, media handling, interactive elements, live chat handover, and automatic error handling with retry logic.
How does Live Chat work in Tars?
Tars provides a standalone Live Chat feature that enables direct conversations between users and human agents. The platform includes a native Live Chat dashboard where human agents can manage and respond to conversations in real-time.Live Chat components
AI Agent Interface
Users interact through the AI Agent interface. Messages flow to the Chatbot backend for processing.
Tars Chatbot Backend
When a Live Chat Gambit is triggered, the backend coordinates the handoff to the Live Chat Server and handles session management.
Live Chat Server
Creates Live Chat sessions, notifies agents via the Admin Dashboard, and routes messages between users and human agents.
Admin Dashboard
Human agents receive notifications, view conversations, and respond to users in real-time through the native dashboard.
Live Chat workflow
1
User message
A user sends a message through the AI Agent Interface, which forwards it to the Tars Chatbot Backend for processing.
2
Live Chat Gambit triggered
When the conversation reaches a Live Chat Gambit, the Tars Chatbot Backend triggers the handoff process and sends the conversation to the Live Chat Server.
3
Session creation & agent notification
The Live Chat Server creates a new Live Chat session and notifies human agents through the Admin Dashboard.
4
Human agent response
Human agents view the conversation in the Admin Dashboard and respond. The response flows back through the Live Chat Server to the AI Agent Interface.
5
Response delivery
The AI Agent Interface displays the human agent’s response to the user. The conversation continues with real-time message exchange between user and agent.
6
Session end
When the conversation is resolved, the human agent can end the session from the Admin Dashboard, returning control to the automated flow.
Standalone capability: Live Chat operates independently from AI Agent flows. You can build conversation paths that lead directly to human agents using the Live Chat Gambit, or keep it completely separate from automated workflows.