Models

What Is Workforce Capacity Planning?

The process of determining how many people, agents, or AI fleet replicas are needed to meet forecast demand at a target service level.

What Is Workforce Capacity Planning?

Workforce capacity planning is the process of determining how many people, contact-center agents, or AI fleet replicas are needed to meet forecast demand at a target service level. For human teams it relies on Erlang C and Erlang A formulas, historical volume curves, shrinkage factors (breaks, training, attrition), and target metrics like average speed of answer. For 2026 contact centers it now includes AI-fleet capacity — voice agents, chat agents, workflow-AI agents — sized via session-concurrency autoscaling and cost-per-handled-contact targets. FutureAGI does not run capacity planning itself; it provides the AI-fleet quality metrics those plans depend on.

Why It Matters in Production LLM and Agent Systems

Capacity planning is unforgiving. Plan low and customers wait, abandon, or escalate, and CSAT drops. Plan high and labor cost balloons or cloud-inference cost balloons. The math is well-developed for human queues — Erlang gives you a confident answer if your forecast is good — but adding AI fleets changes the equation in ways the standard math doesn’t yet handle.

The complications are real. AI fleet capacity is elastic in seconds (autoscaling) but constrained by inference-engine cold start, model availability, and rate limits at upstream providers. AI fleet quality varies with load — under contention, time-to-first-audio rises and ASR accuracy can drop. AI fleets also handle the easier portion of the demand mix; what flows through to human queues is harder, longer, and needs different staffing. A 1:1 substitution of AI for humans does not work; the demand mix shifts.

The pain shows up unevenly. A capacity planner over-staffs the human queue because they did not account for AI deflection lift. A finance lead sees a 40% rise in inference cost during a launch event because the AI fleet did not scale predictably. A product owner sees AI quality degrade during peak hours and traces it to fleet contention. By 2026, capacity planning is a hybrid discipline: WFM for humans, autoscaling for AI fleets, and FutureAGI evaluations to verify that adding capacity actually preserves quality.

How FutureAGI Handles Workforce Capacity Planning

FutureAGI does not produce capacity plans, but it produces the inputs every capacity decision needs. AI-fleet quality under load: when autoscaling adds replicas, FutureAGI’s ConversationResolution, ASRAccuracy, and AudioQualityEvaluator continue scoring; if quality drops as concurrency rises, the autoscaler target is wrong. Demand-mix verification: FutureAGI’s per-route eval-fail-rate-by-cohort tells the planner whether the AI fleet is actually deflecting the routine cases or pushing borderline cases to humans incorrectly. Pre-launch capacity tests: the simulate SDK’s Scenario and LiveKitEngine run synthetic load against the AI stack, capturing quality at target concurrency before production.

A concrete example: a retailer plans for Black Friday: forecast 8× normal volume, AI deflection target 45%, human staffing scaled to 1.6×. The team uses LiveKitEngine to replay 4× normal volume through the voice-agent fleet and watches WER and resolution scores; at 4× concurrency, ASRAccuracy drops 6 points and ConversationResolution drops 9 points. The autoscaler is reconfigured to ramp earlier, and the simulation is re-run until quality holds. The capacity plan then incorporates not just “we can handle the load” but “we can handle it at quality.” The WFM tool plans the human side; FutureAGI verifies the AI side.

For ongoing capacity, FutureAGI feeds quality metrics into the planner’s spreadsheet via API; the planner reconciles WFM occupancy, AI-fleet concurrency, cost-per-contact, and quality scores into one operating budget.

How to Measure or Detect It

Capacity planning depends on quality-validated demand mix:

  • Forecast accuracy (planner metric): historical forecast vs actual; the input to every plan.
  • AI deflection rate — share of demand handled by AI; verified via ConversationResolution.
  • Concurrent-sessions per replica (autoscaling target) — the AI side of occupancy.
  • ConversationResolution under load — does AI quality hold as concurrency rises?
  • ASRAccuracy by cohort — voice-fleet quality per language, accent, audio path.
  • Cost-per-handled-contact — AI side via token cost, human side via payroll; reconcile to total operating cost.
  • Human-queue follow-through quality — when AI escalates, does the human queue still hit SLA?
from fi.evals import ConversationResolution, ASRAccuracy

# Run continuously to surface quality-vs-capacity tradeoffs.
resolution = ConversationResolution()
asr = ASRAccuracy()
result = resolution.evaluate(transcript=session, user_goal=goal)
print(result.score)

Common Mistakes

  • 1:1 AI-for-human substitution math. AI handles the easier mix; what flows to humans is harder and needs more handle time.
  • No quality check on AI capacity. Adding replicas can preserve throughput while degrading quality; verify with FutureAGI evals.
  • Skipping pre-launch simulation. Production traffic is the wrong place to discover the AI fleet does not scale.
  • One forecast for both surfaces. Demand mix shifts as AI deflection improves; refresh forecasts quarterly.
  • Static cost-per-contact assumptions. Inference-cost-per-token changes; rebuild the cost model when models change.

Frequently Asked Questions

What is workforce capacity planning?

Workforce capacity planning determines how many people, agents, or AI fleet replicas are needed to meet forecast contact volume at a target service level — using Erlang formulas, historical volume, and shrinkage.

How does AI change capacity planning?

AI shifts a portion of demand from human queues to AI agent fleets. The plan now needs two capacity tracks: human staffing (sized via WFM) and AI fleet replicas (sized via autoscaling on session-concurrency targets).

Does FutureAGI do capacity planning?

FutureAGI does not run capacity planning — that lives in WFM tools and autoscaling configs. It provides the quality metrics that capacity decisions depend on: ConversationResolution, ASRAccuracy, and per-cohort eval-fail-rate that flag when adding capacity won't fix the problem.