Runrate Framework
AI Workforce P&L
Treat AI agents like employees: cost structure, productivity target, and retirement trigger per agent.
Read the full framework →Every human on your team has a lifecycle. Hire, onboard, perform, develop, retire. When a team member announces they're leaving, you don't scramble—you have a succession plan. You know how to backfill their work. You understand the replacement cost.
AI agents have lifecycles too. But most organizations don't treat them that way. They deploy an agent, it runs until it breaks, and then someone notices performance degrading or cost exploding six months later. By then, the damage is done. The better approach is to plan for the agent's lifecycle from day one—and that includes planning for its retirement.
The five stages of an AI agent lifecycle
An agent's lifecycle parallels a human's, with some important differences.
Stage 1: Hiring (Procurement). You identify a need (high-volume, repetitive work) and decide to fulfill it with an agent. Evaluation phase: does this use case justify agent deployment? Is the work volume high enough? Is the manual cost baseline clear? Decision: which model? Claude Sonnet, GPT-4, open-source LLaMA? This phase costs $10K–$30K in evaluation and spike work.
Stage 2: Onboarding (Deployment). You integrate the model with your systems, build the prompt, set up logging and observability, establish quality metrics. The agent goes live. Ramp time: 2–4 weeks from "approved" to "in production." Deployment cost: $20K–$50K depending on complexity (integration depth, knowledge base size, third-party API calls).
Stage 3: Performance (Active Operation). The agent runs at steady-state, handling work items at a target cost-per-outcome. Monthly tracking: Is the agent at or above the cost-per-outcome target? Is accuracy stable? Is human review rate in bounds? Typical duration: 6–18 months depending on the model lifecycle and business stability.
Stage 4: Optimization (Performance Plan). After the agent has been in production for 4–6 months, you start to see improvement opportunities. Better prompts, better knowledge base, better tool integration could cut cost by 15–25% while holding accuracy constant. This is your development phase—you invest engineering time to improve the agent. Cost: $5K–$15K. If successful, you extend the agent's productive life by 6–12 months. If unsuccessful, you move to retirement.
Stage 5: Retirement (Offboarding or Migration). The agent is sunset for one of several reasons: (a) a new model version is available that's cheaper or more capable, (b) a newer agent version has been developed that supersedes the old one, (c) the work process it handles is being automated away or shifted to a different system, or (d) accuracy/cost metrics have degraded beyond recovery. You don't just turn it off—you have a sunset plan. Parallel run the new agent for a month to build confidence, migrate traffic gradually, shut down the old agent once stability is confirmed. Cost: $5K–$15K, downtime risk: medium-to-high (need careful orchestration).
Real examples of agent retirement
Claude 2 to Claude 3 Opus transition. Teams that deployed Claude 2-based agents in late 2023 faced a choice in March 2024 when Claude 3 (Opus, Sonnet, Haiku) shipped. Claude 2 was still available and working fine, but Claude 3 Opus was demonstrably more capable at the same price. What's the business case for migrating?
If your agent had been running for 8 months at $1.50 per ticket (cost per outcome), and Claude 3 Opus could hit $1.35 per ticket (you estimate this from benchmark testing), the migration pays for itself in 2–3 months ($15K migration cost ÷ $0.15 per-ticket savings × 50K tickets/month = 2 months). You migrate.
But if your agent had been heavily optimized for Claude 2 and Claude 3 Opus's cost savings are only $0.05 per ticket (5% improvement), the payback is 30 months. You don't migrate. You wait for the next opportunity.
GPT-4 to GPT-4 Turbo to GPT-4o migration. OpenAI's customers experienced three model transitions in 2024. Teams running GPT-4 (Nov 2023) saw GPT-4 Turbo ship with 3× context window and 30% cost reduction. Payback on migration: 1–2 months. Most teams migrated. Then GPT-4o shipped with another 50% cost reduction and better speed. Payback on migrating from Turbo to 4o: 3–4 weeks. Most teams migrated again.
Those are clean cases. Cost improvement drives the decision. Payback is short. The business case is obvious.
Sunsetting GPT-3.5-based agents. OpenAI stopped maintaining GPT-3.5 in mid-2024 after the API and token limits degraded. Teams running production agents on GPT-3.5 couldn't ignore it. They had to migrate to GPT-4 or GPT-4o. Cost impact: GPT-4 was 10× more expensive than GPT-3.5 per call. For a team running 100K tickets/month on GPT-3.5 at $0.15 per ticket, that's $1,500/month in API cost. Migrating to GPT-4o would have cost $5,000/month. The cost increase was catastrophic, so most teams either (a) deprecated the agent entirely, (b) restricted its use to high-value customers, or (c) invested heavily in optimization to cut cost by 70% through better prompting.
The decision framework: when to migrate, optimize, or retire
When you're six months into an agent's lifecycle and you're considering an upgrade or replacement, use this framework.
Step 1: Calculate the payback period for migration.
Migration cost (engineering + testing + downtime risk): $10K–$20K. Expected cost per outcome improvement: (old cost − new cost). Monthly work volume: 50,000 items.
Payback period = migration cost ÷ (cost savings per item × monthly volume).
If payback is less than 3 months, migrate. If payback is 3–6 months, migrate unless you have other priorities. If payback is 6+ months, don't migrate unless forced (e.g., the old model is being sunsetted).
Step 2: Assess accuracy and stability.
Is the agent's accuracy declining? Has it degraded more than 2–3 points year-over-year? If yes, you have a performance problem. You need to either optimize (invest in better prompts, better tooling) or retire the agent.
Are there new models available that are demonstrably more accurate on your specific use case? If yes and payback is reasonable, migrate.
Step 3: Consider external factors.
Is the old model being sunsetted by the vendor? (Claude 2 was deprecated. GPT-3.5 API degraded. This happens.) If yes, you have a forced migration timeline. Plan 4–6 weeks before sunset.
Are there compliance or security concerns with the old model? If your regulated industry requires using the latest model version, that's a compliance-driven retirement, not optional.
Is there competitive pressure to upgrade? If a competitor's agent is running on GPT-4o and yours is on GPT-3.5, you're at a cost and quality disadvantage. That's business pressure.
Step 4: Make the decision.
Retire (migrate to a new agent): if payback period is < 3 months OR the old model is being sunsetted. Optimize (invest in prompt improvements): if payback period is 3–6 months AND there's a clear optimization path (identified through spike work). Hold (do nothing): if payback period is > 6 months AND the agent is performing well AND there's no external pressure.
The mechanics of retirement: the parallel-run playbook
When you decide to retire an agent, don't flip the switch. Run both agents in parallel.
Week 1-2: Parallel run with 10% traffic to new agent, 90% to old agent. Route 10% of live work items to the new agent. Monitor: Does it hit the cost-per-outcome target? Does accuracy match or exceed the old agent? Is there latency degradation? Are there integration issues?
If pass rate is 95%+ on all metrics, proceed to week 3.
Week 3-4: Parallel run with 50% traffic to new agent, 50% to old agent. If you hit issues at 50%, they'll surface. If you don't, traffic can go to 50% with lower risk than the 10% run.
Week 5: Move to 90% new agent, 10% old agent (reverse of the original). You're now fully confident that the new agent works. The 10% running on the old agent is a safety valve.
Week 6: Shut down the old agent. Delete the old infrastructure, retire the model subscription if you're per-model billed, document the deprecation.
This playbook takes 6 weeks and costs $5K–$10K (engineering and testing time). That seems like overhead, but it's cheap insurance against deploying a broken agent to production and losing a month of productivity.
Planning for retirement from day one
The best organizations build retirement planning into day one of agent deployment.
In the initial RFP or vendor evaluation, ask: What's your model sunset timeline? Claude? 18–24 months before a version is deprecated. OpenAI? Similar. Open-source models? Depends on the foundation, but usually you own the lifecycle.
In the deployment phase, document: Which model version is this agent on? What's the next-gen model in the vendor's roadmap? What are the estimated cost and accuracy improvements? Build a simple migration roadmap: "If Claude 4 ships in 2025, we'll evaluate migration in Q1 2025. If payback is < 3 months, we'll execute in Q2."
In the operational phase, track: Model release dates, performance benchmarks of new models on similar work, cost trajectory for models you're using. Every quarter, ask: Is there a newer model that we should evaluate? Has the payback period changed?
The P&L implication
Planned retirement is cheaper than unplanned deprecation.
If you run an agent into the ground (accuracy degrades to 60%, cost creeps to $3.50 per ticket, users complain), you're forced into a panic migration at high cost and high risk. That's a fire drill.
If you plan for retirement, you migrate on your schedule, at lower cost, with time to test. You get 6 months of operational benefit from the optimization and upgrade cycle before retiring the agent.
The CFO difference: Planned retirement shows up as a predictable operational cost ($5K–$15K per agent per year for planned upgrades). Unplanned retirement shows up as a crisis and a write-off.
What to do next
Audit your current agent fleet. For each agent in production, document: (1) what model version is it running, (2) when was that model released, (3) what's the roadmap for the next model version, (4) if we need to migrate, what's the payback period? Once you have that spreadsheet, you've planned the next 12 months of agent lifecycle management.
If you're building the CFO's case for AI cost attribution, the 40-page CFO Field Guide to AI Costs walks through the line-item model and the board-deck talking points.
Go deeper with the field guide.
A step-by-step PDF for implementing AI cost attribution.
Was this article helpful?