Skip to main content
The Ticketing view on your metrics dashboard covers ticketing channels-Zendesk and Salesforce. Unlike chat, ticketing workflows expect human-AI collaboration. Use the channel filter to compare Zendesk and Salesforce performance, and the label filter to segment by customer tier or product area.

Why Involvement Rate, Not Resolution Rate

In chat, the AI either resolves a conversation or it doesn’t-Resolution Rate captures this cleanly. In ticketing, human finishing is a normal, valuable outcome. A ticket where the AI drafts a response and a human sends it still saved significant agent time, but Resolution Rate counts this as a failure of autonomy. Every ticket falls into one of two buckets: involved (the AI participated) or not involved (the AI didn’t participate). Within involved tickets, three levels describe how the AI participated:
LevelWhat happensCustomer sees AI?
AutonomousThe AI handled the entire ticket without any human interventionYes
PublicThe AI generated customer-visible responses, then a human also participatedYes
PrivateThe AI suggested responses internally, but a human sent all customer-facing messagesNo
Involvement Rate counts all three levels together-it tells you what share of tickets the AI touches at all. Relative Autonomous Rate narrows the lens to just involved tickets and measures how many the AI handled fully on its own. Together they answer: “Does the AI participate?” and “When it does, how often does it resolve tickets without help?”

Single metrics

CardDescriptionInterpretation
Involvement RatePercentage of tickets where the AI participated (autonomous, public, or private)80%+ means the AI assists with most tickets
Involved TicketsAbsolute count of tickets with AI participation, with trend comparisonMultiply by average handling time to estimate agent hours saved
Relative Autonomous RatePercentage of AI-involved tickets handled fully autonomously (excludes human-only tickets)60%+ indicates strong autonomous performance among involved tickets
Better Monday ScorePercentage of weekend tickets where the AI provided at least one customer-visible response70%+ means strong weekend coverage, reducing Monday morning backlogs
High involvement + low autonomy means the AI engages frequently but needs human finishing-focus on knowledge gaps. Low involvement + high autonomy means the AI is effective but underutilized-expand coverage to more ticket types.

Charts and Use Cases

ChartUse case
Involvement Flow (Sankey)Identify optimization paths. A wide autonomous → escalated flow reveals knowledge gaps to fix. Maximize the autonomous → resolved flow. A large not involved column means tickets the AI could handle.
AI Involvement vs. SuccessCompare resolution rates across involvement types. If autonomous and public resolution rates are similar, humans aren’t adding much-expand autonomous handling. High escalation in autonomous tickets means the AI correctly recognizes its limits but needs better knowledge.
Involvement Rate Over TimeTrack adoption. Growing autonomous (green) and shrinking not-involved (gray) sections indicate improving coverage and knowledge.
Involvement Rate EvolutionSpot inflection points. Autonomous rising while public falls means the AI takes over tickets that previously needed human finishing. A plateau in the autonomous line signals you’ve hit current knowledge limits.

Identifying Issues

Convert public involvement to autonomous. Filter the Sankey diagram to public → resolved. Review what information humans added that the AI lacked. Extract those patterns into your knowledge base. Monitor whether those ticket types shift to autonomous → resolved. Reduce Monday backlog. Check the Better Monday Score. If it’s below 50%, filter to unanswered weekend tickets and review them by topic. Add knowledge for the most common weekend inquiry types and consider less aggressive escalation rules outside business hours. Find topics the AI can’t handle alone. Filter the Sankey to autonomous → escalated and group by topic on the topics dashboard. Decide for each topic: add knowledge (if information gap), create specific escalation rules (if legitimately complex), or improve guidance (if judgment issue). Measure ROI. Multiply the Involved Tickets count by your average ticket handling time to estimate agent hours saved. For a more precise calculation, use the Relative Autonomous Rate to isolate tickets that needed zero human time.
Don’t optimize Involvement Rate or Better Monday Score at the expense of answer quality. A high score with poor responses frustrates customers. Monitor CSAT alongside automation metrics.

Next Steps

  • Metrics - Return to the dashboard overview and cross-channel comparison
  • Conversations - Drill into individual tickets to understand metric patterns
  • Topics - Segment ticket performance by topic
  • Improve Answers - Use ticket insights to refine knowledge
  • Data Providers - Add knowledge to increase autonomous resolution