Skip to content
All blueprints
Blueprint

Channel-Specific Shrinkage and Concurrency Blueprint

A step-by-step method for replacing one-size-fits-all staffing assumptions with channel-specific shrinkage, concurrency, and workload math.

Audience

WFM analysts, planning managers, and support leaders running blended voice, chat, and async teams

Time

60 to 90 minutes for the first model refresh

Before you start

Use this blueprint when

  • Your staffing model looks safe but voice still misses service level during peaks
  • You use the same shrinkage figure for voice, chat, and email
  • Chat utilization looks strong in theory but AHT grows after blending
  • Async teams look overstaffed on paper but backlog still behaves unpredictably
  • You need a practical math correction before changing schedules or headcount

Prerequisites

  • Two weeks of interval-level operational data
  • Scheduled time and actual available time by channel
  • Chat concurrency observations by interval or time band
  • A current staffing spreadsheet or WFM model you can edit
  • Agreement on what counts as shrinkage versus productive wait time

Inputs needed

  • Voice arrivals and AHT by 15-minute interval
  • Chat arrivals, AHT, and concurrency by interval
  • Email or ticket arrivals, backlog, and average work time
  • Scheduled hours versus actual available hours by assignment type
  • Same-day re-contact rate by channel if available
  • Peak-period occupancy and queue response targets

Steps

1

Separate the assumptions before you separate the channels

Identify which parts of your current staffing model are actually shared assumptions, then split them by channel.

Most omnichannel staffing models are not truly channel-specific. They usually keep one shared shrinkage rate, one shared occupancy comfort zone, and one concurrency assumption that gets applied too broadly.

Before changing the staffing math, list the assumptions that are currently universal in your model. In most teams, those are shrinkage, AHT treatment, occupancy, concurrency, and schedule flexibility.

  • voice assumptions
  • chat assumptions
  • async assumptions

This creates the foundation for a model that reflects real operating behavior instead of carrying voice defaults into every other queue.

2

Measure shrinkage from actual behavior, not benchmark habit

Calculate shrinkage separately for each channel or assignment type using your own scheduled versus available time.

Start with a two-week audit. For each assignment type, compare scheduled time with actual available time after planned and unplanned exceptions.

  1. Pull scheduled hours by agent or team
  2. Pull actual available hours during the same period
  3. Segment the result by voice, chat, async, or blended assignment
  4. Replace your universal shrinkage value with those observed rates

This is where many teams discover that chat and async queues are not actually carrying the same shrinkage burden as voice. The correction improves staffing accuracy immediately, even before you change forecasting or routing. If your team still needs a shared baseline, connect the exercise to the live contact center shrinkage tool and the shrinkage glossary entry so the planning language stays consistent.

3

Treat productive wait time differently in chat and async work

Do not classify all non-talking or non-typing time as lost capacity.

Voice is highly sensitive to real-time absence. Chat is not identical because some waiting time inside a session can be reused productively across other chats. Async work is different again because the queue tolerates more pacing flexibility.

Instead of forcing every idle-looking minute into shrinkage, agree on what counts as productive wait time in each channel. That decision has a direct effect on staffing inputs and often explains why chat appears chronically overstaffed in spreadsheet models.

4

Build a concurrency curve, not a concurrency slogan

Replace one fixed concurrency ratio with interval-based ranges that reflect peak and shoulder behavior.

A concurrency assumption like three chats per agent is usually a business-case shortcut, not an operating truth. At peak, that number often collapses because customers reply faster, queues are hotter, and agents have less recovery time between interactions.

  • peak: lower concurrency
  • shoulder: higher concurrency
  • trough: stable or experimentally increased concurrency

If you can only make one improvement, make it this one. A time-banded concurrency curve usually outperforms a flat ratio because it reflects where the day actually breaks.

5

Adjust AHT by assignment mix, not just by channel

Measure how handle time changes when agents are dedicated versus blended.

Chat AHT for dedicated chat agents is not the same number you should use for agents who are also taking voice work. The same pattern can appear in email when agents are repeatedly pulled into real-time queues.

Split AHT into at least two views: dedicated assignment and blended assignment. That gives you a correction factor you can apply before staffing math turns a false efficiency assumption into a service-level miss.

6

Route each channel into the right staffing method

Use the right planning logic for each type of work instead of trying to force one formula across everything.

Once your assumptions are cleaned up, route each channel into the planning method that actually fits. Voice belongs in real-time queue staffing. Chat belongs in adjusted real-time staffing with concurrency and AHT corrections. Async work belongs in backlog and drain-rate planning.

If you want these assumptions reflected in an actual planning workflow, the strongest stable internal destinations today are forecasting for interval demand modeling and intraday management for day-of-operation corrections.

7

Review the math monthly, not once a year

Refresh the assumptions that drift fastest so the model stays useful.

Concurrency, blended AHT, and assignment-specific shrinkage are living inputs. They move when staffing models change, when routing changes, and when the team takes on new work.

Review them on a fixed cadence, usually monthly or quarterly depending on how volatile your environment is. A simple refresh loop prevents old assumptions from becoming invisible sources of systematic error.

Implementation checklist

0/7

This blueprint is most useful when treated as a modeling reset, not a theoretical argument. If you are building a broader planning workflow, keep it connected to the rest of your resources and link it with adjacent blueprints that cover peak scheduling, overtime control, and intraday response. The strongest hub for that cluster is the support-team WFM blueprint.

The goal is simple: make the staffing math reflect how the operation actually behaves. Once that happens, schedule conversations become more honest and performance problems become easier to diagnose.

Related resource

Forecasting and Staffing Planning

Open resource

Your next schedule could take 2 minutes.

Import your team, set your rules, hit auto-fill. Most teams are live the same day.

Try Soon free

30 days free · No credit card required

Already have an account? Sign in