Deploy & Monitor
Production Operations & Feedback Loops
Why Monitoring Matters
Deploying an AI support agent is not the finish line — it's the starting line. The critical question isn't "does it work?" but "how well does it work, and is it getting better?"
The key metric is auto-resolution rate — what percentage of tickets does the AI resolve without human intervention? Industry average is 20-30%. A well-tuned system with good KB coverage hits 60-80%.
Key Concepts
Multi-Channel Deployment
A support agent that only works in a web UI misses most interactions. Production deployment means:
| Channel | Integration | Key Consideration |
|---|---|---|
| Web chat | Next.js app (Module 5) | SSE streaming, rich formatting |
| Slack | @slack/bolt | Ephemeral messages for sensitive data |
| Inbound parsing | Auto-respond for high confidence only | |
| API | MCP server | Headless access for other tools |
The same pipeline (classify → retrieve → respond → escalate) serves all channels. Only the delivery format changes.
Slack Integration
The Slack bot has three modes:
/support "I can't log in" — quick queriesCritical: Use ephemeral messages for sensitive topics. A billing inquiry response with account details should only be visible to the requester, not the entire channel.
Monitoring Dashboard
Four categories of metrics:
Resolution Metrics:
Quality Metrics:
Operational Metrics:
Trend Metrics:
The Feedback Loop
The most important operational pattern:
Deploy → Monitor → Identify Gaps → Fix → Redeploy
↑ │
└──────────────────────────────────────────┘Weekly cycle:
Each iteration makes the system better. After 4-8 weeks, auto-resolution rate typically jumps from 40% to 65%+.
KB Coverage Gaps
The highest-leverage improvement: adding missing KB articles. When the retrieval system returns no good results (top score < 0.3), log the query. After a week, cluster these "no answer" queries by topic. The top clusters are your missing KB articles.
Example: if 50 queries per week ask about "API pagination" but there's no KB article for it, writing one article resolves all 50 going forward.
MCP Server
The Model Context Protocol exposes the support agent to other AI tools:
| MCP Tool | Purpose |
|---|---|
| `search_kb` | Search knowledge base from any MCP client |
| `classify_ticket` | Run classification from external workflows |
| `check_escalation` | Evaluate escalation rules programmatically |
| `draft_response` | Generate response drafts from other tools |
This is the "headless support agent" pattern — the intelligence is decoupled from any specific UI.
Operations Runbook
Essential documentation for production:
What You'll Build
Glossary
| Term | Meaning |
|---|---|
| Auto-resolution rate | % of tickets resolved by AI without human help |
| Escalation rate | % of tickets handed to humans |
| CSAT | Customer Satisfaction score (1-5) |
| MCP | Model Context Protocol — universal tool interface |
| Feedback loop | Continuous cycle of monitoring, diagnosing, and improving |
| KB coverage gap | Customer questions with no matching KB article |
| Ephemeral message | Slack message only visible to the requester |
This is chapter 6 of AI Customer Support Agent.
Get the full hands-on course — free during early access. Build the complete system. Your projects become your portfolio.
View course details