Runrate Framework
The AI Cost Iceberg
Visible API spend (10%) vs hidden inference, storage, observability, retries, human review (90%).
Read the full framework →Runrate Framework
AI Workforce P&L
Treat AI agents like employees: cost structure, productivity target, and retirement trigger per agent.
Read the full framework →According to MIT NANDA's "GenAI Divide" study, 95% of AI pilots fail to deliver measurable P&L impact. Yet the 5% that succeed follow a repeatable pattern: they don't measure tokens or API costs—they measure outcomes. They treat AI agents the way a CFO treats headcount. And crucially, they capture both the visible cost (the API bill) and the hidden cost (inference overhead, retries, human review, integration debt) in a single model. This pillar walks through the exact methodology the 5% use to build defensible AI ROI cases.
The MIT Finding: Why 95% of Pilots Fail
The MIT NANDA study interviewed 1,420 executives at organizations running AI pilots. The finding was stark: 95% of pilots generated no measurable profit-and-loss impact. Most organizations could not even say whether a pilot succeeded or failed. No baseline. No attribution. No clear path from "we deployed an AI agent" to "here's what we saved."
This isn't a technology problem. It's a measurement problem.
McKinsey's State of AI 2025 found that while 88% of large organizations now use AI in at least one function, only 39% see meaningful EBIT impact, and just 5.5% qualify as "AI high performers" with repeatable AI value creation. The gap isn't in the AI models themselves—it's in the financial infrastructure to measure them.
Most CFOs inherited this blind spot. They have sophisticated machinery to track headcount ROI: payroll systems, time tracking, productivity dashboards, cost-to-hire models, attrition analytics. But when an AI agent does work, that same financial machinery goes silent. The agent's cost is buried in the AI Cost Iceberg—visible API spend over hidden inference overhead, retries, observability, and human-in-the-loop review time. And the agent's outcome (a resolved ticket, an adjudicated claim) is often unattributed to any revenue or margin line.
The 5% who succeed build an AI Workforce P&L. They stop asking "what did our API cost?" and start asking "what did this agent cost to deliver an outcome?"
The AI Cost Iceberg: Separating Visible from Hidden Cost
Most organizations budget for the 10% of AI cost they can see: the monthly API bill from OpenAI, Anthropic, or Google. A CFO looks at the invoice and sees "$4,200 this month for Claude API calls."
The other 90% of cost is invisible—until it blows up the budget.
When you run an AI agent at scale, you pay for far more than the inference itself. You pay for:
- Inference at scale: The per-token cost compounds as volume grows. A $0.003 cost per token at 100,000 daily requests becomes $300,000/month at scale.
- Vector database and embedding storage: Semantic search and retrieval-augmented generation require persistent vector stores. These are not free.
- Retries on failure: When an agent hallucmates or errors, the system retries. Each retry is a new API call.
- Tool calls to third-party services: An AI agent solving a customer support ticket might call Stripe (to look up a payment), Twilio (to send an SMS), Salesforce (to create a case record). Each tool call costs time and money.
- Human-in-the-loop review: Most production agents require humans to review, approve, or correct AI decisions. That review time is often not counted as an AI cost, but it absolutely is—and it often dwarfs the API cost.
- Observability and monitoring: Logging, tracing, and monitoring AI pipelines at scale requires infrastructure. Many teams reinvent this internally and hide the cost.
- Prompt engineering and evaluation: Building good prompts takes human labor. Evaluating agent quality takes human labor and often requires calling the model thousands of times.
- Caching infrastructure: Prompt caching (Anthropic's or OpenAI's) reduces API costs but requires infrastructure investment and complexity.
- Security and compliance overhead: Tokenization, encryption at rest and in transit, audit logging—these layers add cost and latency.
The shorthand: the AI Cost Iceberg—visible API spend over hidden inference, retries, observability, and human review.
Most CFOs are looking at the tip and budgeting against the iceberg.
Moving from Token Cost to Cost Per Outcome
The pivot that separates the 5% from the 95% is this: stop measuring token cost. Measure cost per outcome.
An outcome is the discrete work item your agent completed: a customer support ticket resolved, an insurance claim adjudicated, a loan application processed, a legal contract reviewed. These outcomes map cleanly to your P&L because they map to revenue or cost avoidance.
The formula is simple:
Cost Per Outcome = (Total AI spend + Human review spend + Infrastructure spend) / Number of outcomes
Here's a worked example. Suppose you deploy an AI customer support agent:
- Monthly total AI spend (visible + hidden): $18,000 (includes API, inference, retries, vector database, observability, prompt caching)
- Monthly human review cost: $12,000 (two FTEs at $36k each spend 25% of their time reviewing AI decisions)
- Monthly infrastructure and observability: $4,000
- Total monthly cost: $34,000
- Tickets resolved per month: 2,000
- Cost per resolved ticket: $34,000 / 2,000 = $17 per ticket
Now you can compare this to the baseline. Before AI, how much did it cost to resolve a ticket? If your fully-loaded cost per CSR was $50k/year ($4,167/month) and each CSR handled 200 tickets per month, your baseline cost-per-ticket was $20.84.
AI brought it to $17. You've improved the unit economics by 18%. That's a credible, board-grade ROI case.
The AI Workforce P&L Framework
The methodological leap is to treat AI agents like you treat employees. Build a payroll-equivalent infrastructure for AI.
In your payroll system, every employee has:
- A timecard (how much time they spent on which work)
- An attribution path (their salary and benefits charged to which cost center)
- A classification (W-2 employee, contractor, temp)
- A retirement trigger (when to offboard them)
In your AI Workforce P&L, every agent needs the same:
- A timecard: A detailed log of which outcomes it delivered (which tickets it resolved, which claims it adjudicated). This is event-level granularity—not aggregated monthly but work-by-work.
- An attribution path: The agent's cost is charged to the same P&L line it serves (customer support cost, claims processing cost, loan origination cost), not buried in "AI infrastructure."
- A classification: Is the agent a third-party API (like Claude API—a contractor model with per-token billing) or self-hosted (like an LLM on your own GPU—an employee model with fixed cost)? This changes the ROI profile.
- A retirement trigger: When does the agent stop making economic sense? If a human CSR can now resolve tickets faster than the AI, it's time to redeploy the agent.
This isn't metaphorical. The teams seeing repeatable 5% success are literally building payroll-equivalent ledgers for AI. They have a "payroll number" (total AI agent cost), a "headcount equivalent" (cost per agent divided by annual cost-per-FTE), and a "margin contribution" (revenue generated or cost saved per agent).
Worked Example: Customer Support AI ROI
Let's walk through a complete customer support ROI model to show how the 5% do it.
Current state (baseline):
- Customer support team: 12 FTEs
- Average compensation (salary + benefits + overhead): $52,000/year = $4,333/month per person
- Total monthly cost: 12 × $4,333 = $52,000
- Average tickets handled per person per month: 200
- Total capacity: 2,400 tickets/month
- Cost per resolved ticket: $21.67
Proposed state (with AI agent):
- AI agent deployment: One customer support AI agent that handles first-contact resolution on 70% of tickets. Remaining 30% escalate to human.
- Monthly AI agent cost (visible + hidden): $6,000
- Monthly human review cost: $8,000 (one FTE at 50% utilization reviewing escalations and edge cases)
- Monthly infrastructure: $1,000
- Total monthly AI cost: $15,000
- Tickets handled per month: 2,400 (same total, but AI handles 1,680 and human handles 720)
- Cost for AI-handled tickets: ($6,000 + $8,000 + $1,000) × (1,680 / 2,400) = $10,500
- Cost for human-handled tickets: $52,000 × (720 / 2,400) = $15,600
- Total monthly cost: $26,100
- New cost per ticket: $10.88
- Cost reduction: 50%
- Annual savings: ($52,000 - $26,100) × 12 = $311,000
That's a credible ROI model you can walk into a board meeting with. It's built on transparent assumptions (escalation rates, human review overhead, AI cost), and it maps directly to the P&L (you spend less to resolve the same tickets).
McKinsey's High-Performer Profile
McKinsey's analysis of 5% "AI high performers" identified a repeatable pattern in how they build ROI. These teams share four characteristics:
-
A clear outcome metric before deployment. They define the P&L line they're targeting (customer support cost, claims processing cost), and they measure baseline cost-per-outcome before deploying the agent. No baseline, no ROI.
-
Integrated infrastructure for measuring cost and outcome together. They don't separate "what did we spend on AI?" from "what did we accomplish with it?" The infrastructure (data pipeline, dashboard, reporting) ties cost to outcome at the work-item level.
-
Attribution clarity. Every dollar of AI spend is attributed to a business outcome and a P&L line. There's no "miscellaneous AI infrastructure" bucket. If the agent serves customer support, its cost is in customer support P&L.
-
Feedback loops. They measure continuously and adjust. If the cost per outcome drifts, they know within a week. If an agent's accuracy drops, they intervene. This is the same discipline a CFO applies to headcount productivity metrics.
The Three Numbers Every CFO Needs
To build an AI ROI case that survives board scrutiny, you need exactly three numbers:
-
Baseline cost per outcome. How much does it cost today (with humans) to do this work? This is your control case. If you can't answer this, stop—you can't measure ROI.
-
Total cost per outcome with AI. This includes the visible API cost, the hidden infrastructure cost, and the human review cost, divided by the number of outcomes. This is the new numerator in your ROI formula.
-
Confidence interval on #2. Your CFO asks: "How sure are you about that number?" You should have a sensitivity analysis showing what happens if AI accuracy is 10% lower, if human review time is 2x higher, if API costs spike 25%. The 5% don't pretend their forecast is perfect; they show the range.
With these three numbers, you can calculate:
- Year-1 ROI = (Baseline cost per outcome - AI cost per outcome) × Annual volume / AI infrastructure investment
- Payback period = AI infrastructure investment / ((Baseline cost per outcome - AI cost per outcome) × Monthly volume)
- Headcount equivalent = Total monthly AI cost / Average monthly cost per FTE
Why Most Pilots Fail: Five Root Causes
The 95% fail for predictable reasons. If you avoid these five traps, you're already in the 5%.
1. No baseline. The team deploys an AI agent and then asks "did it work?" with no measurement of how much the old process cost. Without a baseline, you can't calculate ROI. The first step is always: measure cost-per-outcome for the current manual process.
2. Attribution gap. The AI agent cost sits in "AI infrastructure" on one P&L line, while the human cost reduction sits in "customer support" on another line. Finance can't connect them. They ask: "Did we actually save money or did we just add cost?"
3. Hidden cost blindness. The team celebrates a $0.03/token API cost and forgets about the vector database ($2,000/month), the retries (25% of calls fail and retry), the human review (an FTE worth of time), and the observability stack. When the true cost emerges, the ROI evaporates.
4. Sandbox-to-production gap. A pilot in a low-volume environment works great: 98% accuracy, $0.50 per outcome, no human review needed. Then you go to production with 10x volume. Accuracy drops to 88%, human review overhead explodes, and the cost per outcome jumps to $8. The lesson: test at scale before you declare ROI.
5. No feedback loop. The pilot launches, the team moves on, and cost-per-outcome is never measured again. Six months later, the model has drifted, the infrastructure cost is higher, and no one knows whether the agent is still creating value. Success requires continuous measurement.
From Iceberg to Outcome: The Attribution Layer
The infrastructure that enables the 5% is an attribution layer—a system that connects the AI agent's cost to its outcome. This is what Runrate builds: work-item-level cost attribution for AI agents.
Without it, you're stuck. Your invoice from OpenAI shows $12,000/month in API costs, but you have no idea which business outcome that paid for. Was it customer support? Claims processing? Loan origination? The CFO can't allocate it to P&L.
With it, you can answer: "Our customer support AI agent cost $3.50 per resolved ticket last month—specifically, on 4,200 resolutions. Our claims agent cost $18 per adjudicated claim, on 1,800 claims. Our loan agent cost $120 per processed application, on 340 applications." Now you can make real optimization decisions. You can retire the underperforming agent. You can scale the profitable one.
This is the move from token cost to cost per outcome. And it's non-negotiable for the 5%.
What to Do Next
Build a baseline cost-per-outcome for the process you're targeting before you deploy any AI. Interview the operations leader. Ask: "How much does it cost today to resolve a ticket / adjudicate a claim / process an application?" Get the answer in dollars. Then, when you deploy the agent, measure cost-per-outcome again—including all AI costs, human review, and infrastructure. If the new number is lower, you have an ROI. If it's higher, you learn why. Run the numbers yourself with the AI ROI Calculator, which walks through this model step-by-step. Or use this framework as the spine of a board conversation with your finance team—the math is simple enough that a CFO can follow it in a meeting.
Calculate your AI ROI.
See what your agents actually cost — and what they're returning.