Login Get started

The Support Leader's Playbook for Scaling Without Burning Out Your Team

Damien Mulhall
Damien Mulhall
Strategic Project Manager & Operations Lead
13 min read
Materials
The Support Leader's Playbook for Scaling Without Burning Out Your Team

Every support leader I talk to says some version of the same thing:

"I need to do more with less. But I refuse to burn out my team doing it."

According to HubSpot's 2024 State of Customer Service report, 75% of support reps said 2024 had the highest ticket volume they'd ever seen. Budgets aren't growing to match. Hiring freezes keep extending. And somewhere in the middle, your team absorbs the difference.

The old playbook of 'hire more people or push the existing team harder' doesn't work anymore. There's no budget for the first option and no capacity for the second.

But there's a question most leaders skip: what if some of the tickets in your queue don't need your team at all?

The Question Worth Asking #

Most leaders frame capacity as binary: either hire more people (expensive) or push the current team harder (unsustainable).

The third option gets overlooked because it requires admitting something uncomfortable: a significant portion of your queue is work that doesn't require human judgment. Not deflected with "sorry, I can't help." Actually resolved the customer's problem solved without a human ever touching it.

This isn't about chatbots that send help articles and hope customers go away. It's about AI that does the work: processes the refund, tracks the order, updates the account, answers the policy question accurately.

The distinction matters. Deflection frustrates customers and creates repeat contacts. Resolution solves the problem.

The 5 Ticket Types That Don't Need Your Team #

After analysing ticket data across dozens of support operations, patterns emerge. These five categories consistently don't require human judgment but they do require different AI capabilities.

1. WISMO (Where Is My Order) #

E-commerce data from Malomo's 2024 Shipping Experience Report suggests 30-50% of support volume is customers asking where their package is. That range is wide because it depends on your shipping reliability, delivery windows, and proactive communication. Companies with poor tracking visibility sit at the high end; those with automated shipping updates sit lower.

These customers aren't angry. They're not confused. They just want tracking information your systems already have.

Why it doesn't need humans: The answer exists in your order management system. The work is retrieval, not judgment.

What AI needs to handle it: Read access to your OMS and carrier tracking APIs. This isn't a chatbot answering questions it's an integration that pulls live data.

💡 WISMO is the canary in your support coal mine. If it's above 40% of volume, your proactive communication is failing. Before automating the response, ask why customers are asking in the first place. Often the fix is sending better shipping notifications, not answering more WISMO tickets.

2. Policy Lookups #

"What's your return policy?" "Do you ship to Portugal?" "How long is the warranty?"

These questions have documented answers. Your agents are copying from a knowledge base and pasting into a chat window. No interpretation required.

Why it doesn't need humans: The answer is fixed. It doesn't depend on the customer's specific situation.

What AI needs to handle it: Access to your knowledge base and the ability to match questions to the right policy. The quality depends entirely on your documentation—garbage in, garbage out.

💡 Policy lookups expose documentation debt. If AI can't answer a policy question accurately, it's usually because the policy isn't documented clearly—or at all. Use AI failures as a diagnostic: every question it can't answer reveals a gap in your knowledge base.

3. Account Changes #

Password resets. Address updates. Subscription pauses. Payment method changes.

These are transactional actions with clear rules. If the customer is authenticated and the request is valid, it gets processed.

Why it doesn't need humans: The logic is deterministic. Authenticate → validate → execute. No judgment calls.

What AI needs to handle it: This is where most AI deployments fall short. Answering "how do I change my address" is easy. Actually changing the address requires write access to your systems CRM, billing platform, subscription management. The AI needs to be more than a chatbot; it needs to be an agent with permissions to act.

💡 Account changes are the test case for "action-taking" versus "answer-suggesting" AI. If your AI can explain how to update a payment method but can't actually update it, you haven't automated the ticket, you've just made the customer do the work themselves.

4. Status Checks #

"Did you receive my return?" "Has my refund been processed?" "When will my replacement ship?"

Information retrieval. The answers exist in your systems. Someone just needs to look them up.

Why it doesn't need humans: Same as WISMO, it's data lookup, not problem-solving.

What AI needs to handle it: Read access to returns management, refund processing, and fulfilment systems. The integration complexity depends on how fragmented your backend is.

5. Tier-1 Troubleshooting #

"The app won't load." "My discount code isn't working." "I can't log in."

Basic diagnostic trees with predictable paths. If X, try Y. Still not working? Try Z.

Why it doesn't need humans: The troubleshooting follows a script. There's a finite decision tree with known solutions.

What makes a path "predictable": The problem has fewer than 5 common causes, each with a documented fix. You can draw the decision tree on a whiteboard. Your agents handle it the same way every time. If any of these aren't true, it's not Tier-1, it's a judgment call.

What AI needs to handle it: Documented troubleshooting flows and, ideally, the ability to check system status (is there an outage? is the discount code actually valid?). Without system access, AI can only walk through generic steps.

💡 The trap with Tier-1 troubleshooting: it often morphs mid-conversation. Customer starts with 'app won't load' (Tier-1) but it becomes 'I've tried everything and I'm furious' (human required). AI needs to detect this shift and escalate gracefully, not keep pushing diagnostic steps at a frustrated customer.

What Should Stay Human #

Automation isn't the goal. Resolution is. Some conversations genuinely require your team.

Angry customers who need to feel heard #

When someone is frustrated and needs to vent, a human being who listens and validates their experience is irreplaceable. AI can detect frustration (sentiment analysis is reasonably mature), but it can't provide genuine empathy.

The nuance: AI should be detecting frustration and escalating, not attempting to resolve. The failure mode is AI that keeps offering solutions to someone who's past wanting solutions, they want acknowledgment.

Complex multi-issue tickets #

When a customer has three interconnected problems, they need someone who can hold context across all of them and find a solution that addresses the whole situation. "My order was wrong, I returned it, the refund went to the wrong card, and now I need to reorder but the item is out of stock." That's four issues with dependencies. AI can handle each individually; humans see the whole picture.

High-value accounts #

High-value customers deserve relationship management, not transactional efficiency. The definition of "high-value" depends on your business: enterprise contracts in B2B, top-percentile LTV in consumer, strategic accounts in any context. Your best agents should be spending time here.

💡 Define "high-value" explicitly before deploying AI. Is it accounts over $10K annual contract value? Top 5% by lifetime spend? Customers flagged by sales as strategic? If you can't define it, AI can't route around it.

Edge cases and exceptions #

Anything that doesn't fit neatly into your rules. The weird situations, the exceptions, the "I've never seen this before" tickets. These require human judgment and creativity.

The test: if you couldn't write a rule to handle it, AI can't follow a rule to handle it.

Customers who prefer humans #

Some customers want to talk to a person even for simple queries. This is a legitimate preference, not a failure to educate them about self-service. Offer the option to reach a human without friction. Forcing AI interaction on customers who don't want it creates resentment that outweighs any efficiency gain.

The Hybrid Reality #

Clean categories are useful for planning. Reality is messier.

Tickets morph. A WISMO query becomes a complaint when the customer learns their package is delayed. A simple troubleshooting ticket becomes complex when the first three solutions don't work and the customer is now frustrated. A policy question becomes an edge case when the customer's situation doesn't quite fit the documented policy.

AI needs to handle these transitions gracefully:

  • Detect when a conversation has shifted from a category it can handle to one it can't

  • Escalate with context so the human agent doesn't start from zero

  • Know its limits and admit them rather than confidently providing wrong answers

The worst AI deployments are the ones that try to resolve everything and escalate nothing. They frustrate customers, create repeat contacts, and make your metrics look better than your actual service quality.

When AI Gets It Wrong #

It will. Plan for it.

AI will occasionally provide wrong information, execute an action incorrectly, or fail to escalate when it should. The question isn't whether this happens, it's how you detect and recover.

Monitor resolution quality, not just resolution rate. A 70% resolution rate means nothing if 20% of those 'resolutions' are wrong answers the customer accepted because they didn't know better.

Sample AI-resolved conversations regularly. Have humans review a random subset weekly. Look for wrong answers, missed escalations, and customer frustration that wasn't detected.

Make escalation easy. Customers should be able to reach a human at any point without friction. "I want to talk to a person" should never be met with more AI.

Have a recovery playbook. When AI gives wrong information that creates a problem, how do you make it right? The answer can't be "apologise and hope they don't churn."

The Transparency Problem #

Here's where most AI deployments stall: the black box problem.

Support leaders want to automate. But when they bring it to Legal or Security, the questions start:

  • "How does it decide what to say?"

  • "Can we audit the conversations?"

  • "What if it says something wrong?"

  • "How do we explain this to regulators?"

Most AI tools can't answer these questions well. The models are opaque. The decision-making is hidden.

Open-source architecture changes this in specific ways:

  • Audit trails: Every decision the AI makes can be logged with the reasoning chain that led to it. When Legal asks "why did it say that?", you can show them.

  • Customisation without waiting: If the AI is handling a ticket type incorrectly, you can modify the logic yourself rather than submitting a feature request and waiting six months.

  • Security review: Your security team can inspect the code. They can verify data handling, identify risks, and approve deployment based on actual evidence rather than vendor promises.

  • Regulatory explanation: When regulators ask how automated decisions are made, you can show them the decision logic rather than shrugging and pointing at a vendor's black box.

Open-source doesn't automatically mean compliant or auditable. It means you have the access to make it compliant and auditable. The work is still yours to do.

Getting Started: A Realistic Timeline #

If you're ready to explore this, here's what the process actually looks like.

Week 1-2: Audit #

Pull your last 500 tickets. Tag each one: could this have been resolved without human judgment? Be honest, the temptation is to over-estimate what AI can handle.

Categorise by the five types above. Calculate the percentage in each bucket. Most teams find 40-60% fall into automatable categories, but your number might be higher or lower depending on your product complexity and customer base.

Week 3-4: Pick One #

Identify your single highest-volume automatable ticket type. Usually it's WISMO or a policy question.

Document the rules exhaustively. How would you train a new agent to handle this ticket? Write it down. Include the edge cases, the exceptions, the "if this then that" logic.

"If you can write the rules" is doing a lot of work in that sentence. Most teams discover their rules aren't as documented as they thought. The gaps become visible when you try to write them down comprehensively.

Week 5-8: Implement and Test #

Deploy AI for that one ticket type. Run it in parallel with human handling initially AI resolves, but a human reviews before the response goes out.

Measure accuracy, not just resolution rate. Did the AI get it right? Did customers escalate afterward? Did resolution quality match or exceed human handling?

Week 9+: Expand Based on Evidence #

Once you've proven it works for one ticket type and only then add another. Build confidence incrementally rather than deploying broadly and hoping.

Timeline reality check: most teams take 2-3 months to get one ticket type automated well. Anyone promising results in two weeks is overselling.

The Goal #

Your agents should spend their time on conversations that genuinely need them: the angry customer who needs empathy, the complex problem that needs creativity, the VIP account that needs relationship management.

"Where's my order?" doesn't need your team. Neither does "what's your return policy?" Neither does "can you update my address?"

These tickets consume agent time without using agent skills. They create volume without creating value. And they're the tickets that burn teams out not because they're hard, but because they're endless.

The tools to handle this exist. The question is whether you'll deploy them thoughtfully starting small, measuring honestly, expanding based on evidence or whether you'll either ignore them entirely or deploy them recklessly and create new problems.

There's a middle path. It takes longer than the vendors promise but delivers more than the skeptics expect.

Hay is running a pilot program through Q1 2026 for support teams handling 2,000+ tickets monthly. We'll audit your ticket mix, identify automation candidates, and deploy action-taking AI for your highest-volume ticket type. Limited to 10 teams.

Apply at hay.chat/pilot.

About the Author

Damien Mulhall

Damien Mulhall

Strategic Project Manager & Operations Lead

Damien spent 10+ years managing support operations and project delivery for global brands including Dell, Microsoft, Intel, and Google. He's PMP-certified and brings structure, process, and operational clarity to everything Hay builds.