Chat Interface
The Chat Interface is how you interact with Support Bot’s AI Copilot. It provides a natural, conversational way to search incidents, analyze patterns, and get resolution recommendations with streaming responses and context awareness.Key Features
Streaming Responses
See answers as they’re generated in real-time using Server-Sent Events (SSE)
Conversation Memory
The AI remembers your entire conversation using PostgreSQL checkpointing
Smart Caching
Frequently asked questions are cached for instant responses
Markdown Support
Rich formatting for code blocks, tables, lists, and more
How It Works
Streaming Architecture
The chat interface uses Server-Sent Events for real-time streaming:Event Types
status - Processing Updates
status - Processing Updates
Shows what the AI is currently doing:
final_answer - Response Chunks
final_answer - Response Chunks
The actual AI response, streamed token by token:
title - Conversation Title
title - Conversation Title
Auto-generated title for the conversation:
complete - Final Response
complete - Final Response
Signals the end of streaming:
Smart Caching
Frequently asked questions are cached for instant responses:Cache Benefits
Instant Responses
Cached queries return in milliseconds instead of seconds
Reduced LLM Costs
Skip LLM API calls for common questions
Consistent Answers
Same question always gets the same verified answer
Lower Latency
No vector search or LLM inference needed
The cache only works for self-contained queries without conversation context. Follow-up questions always go through the full pipeline.
Conversation Memory
The chat interface maintains context using LangGraph’s PostgreSQL checkpointing:Clarification System
The chat interface detects ambiguous queries and asks for clarification:Clarification Triggers
The system asks for clarification when:- You use pronouns without context (“tell me about it”)
- You reference “this” or “that” without prior context
- You ask follow-up questions in a new chat
Title Generation
Conversations are automatically titled for easy reference:- “Show me payment failures” → “Payment Failures”
- “Database timeout in loan processing” → “Loan Processing Timeout”
- “Swift transfer issues EU region” → “EU Swift Transfers”
Managing Chats
Listing Chats
Get all your conversations:Viewing Messages
Retrieve a full conversation:Renaming Chats
Change the conversation title:Archiving Chats
Remove chats from your active list:Non-Streaming Mode
For simple integrations, use the non-streaming endpoint:LLM Provider Override
Override the default LLM provider per message:Prompt Guardrails
The chat interface validates messages before processing:- Injection attacks
- Inappropriate content
- Sensitive data leakage
Best Practices
Writing Good Queries
Be Specific
Be Specific
Provide clear context in your questions:✅ “Show me HTTP 500 errors in the PaymentAPI from last week”❌ “Show me errors”
Use Incident IDs
Use Incident IDs
Reference specific incidents when available:✅ “What was the mitigation for INC-2025-001?”❌ “What did we do for that payment issue?”
Build on Context
Build on Context
Take advantage of conversation memory:
- “Find database timeout incidents”
- “Which application was affected most?”
- “Show me the mitigation steps for the LoanAPI ones”
Start Fresh When Needed
Start Fresh When Needed
Create a new chat for unrelated topics:Chat 1: Payment gateway issues
Chat 2: Database performance problems
Chat 3: Deployment failures
Performance Tips
Use Streaming
Streaming provides faster perceived performance as users see responses immediately
Keep Chats Focused
Shorter conversations with fewer messages perform better
Archive Old Chats
Regularly archive completed investigations to keep your list clean
Cache Common Questions
Frequently asked questions are automatically cached for instant responses
Troubleshooting
Streaming Connection Drops
If the SSE connection is interrupted:- Check timeout settings: Ensure your reverse proxy doesn’t timeout SSE connections
- Verify network: Test with a stable connection
- Use reconnection logic: Implement automatic reconnection in your client
Context Not Maintained
If follow-up questions don’t work:- Verify chat_id: Ensure you’re passing the same chat_id for follow-ups
- Check PostgreSQL: The checkpointer requires a working database
- Review logs: Look for “thread_id” in the backend logs
Slow Response Times
If responses are slow:- Check cache hit rate: Common questions should be cached
- Monitor LLM latency: Some providers are faster than others
- Review incident count: Large knowledge bases may need optimization
Next Steps
AI Copilot
Learn how the underlying AI agent works
API Reference
Complete API documentation for chat endpoints
LLM Providers
Configure different AI models for your chats
Guardrails
Set up prompt validation and safety controls