Runrate Framework
The AI Cost Iceberg
Visible API spend (10%) vs hidden inference, storage, observability, retries, human review (90%).
Read the full framework →Building a credible AI ROI case shouldn't require a spreadsheet PhD. The Runrate AI ROI Calculator takes the guesswork out of AI economics by walking you through a transparent, CFO-grade methodology. You input your current cost structure, AI agent assumptions, and volume forecasts—then the calculator handles the hard math: baseline vs. AI cost per outcome, year-1 ROI, payback period, and sensitivity scenarios. This article explains what the calculator does, what inputs it needs, and why finance leaders trust the methodology.
What the Calculator Does
The AI ROI Calculator is built around a single principle: cost per outcome, not token cost.
Most AI tools measure success in tokens, latency, or accuracy. Those are engineering metrics. The calculator asks the question a CFO asks: "What does this agent cost to deliver an outcome, and is that cheaper than doing it manually?"
You start by defining your use case. Are you building a customer support agent? A claims processor? A loan originator? Each has different baseline costs and different outcome metrics.
Then you input five critical numbers:
-
Current cost per work item (baseline). How much does it cost today, with humans, to resolve one support ticket or adjudicate one claim? This is your control case. If you can't answer this, stop—you can't measure ROI.
-
Monthly volume. How many work items do you process per month? This drives the scale of savings.
-
Total AI agent cost (all-in). Not just the API bill. Include inference overhead, retries, vector database, human review, observability, everything. The calculator provides a breakdown so you don't forget hidden costs.
-
Escalation rate and human review percentage. What percent of AI decisions require human review or escalation to a human? Most agents don't reach 100% accuracy. A realistic forecast (60-85% first-touch accuracy) is better than an optimistic one (95%) that doesn't survive contact with production.
-
AI infrastructure and training investment. What's the one-time cost to build, test, and deploy the agent? This determines payback period.
The calculator outputs:
- Cost per outcome: AI vs. baseline. Your new cost per resolved ticket with AI, compared to your current manual cost. The gap is where value lives.
- Year-1 savings. Total dollar savings in year 1 if you deploy the agent at your forecast volume.
- Payback period. How many months until the investment is recovered.
- Year-1, 2, and 3 ROI. Return on investment by year, accounting for scaling volume.
- Sensitivity analysis. What happens if your forecast is off? If accuracy is 10% lower, if escalation is 2x higher, if API costs spike 25%—the calculator shows the range of outcomes.
Why the Methodology Works
The calculator's strength is transparency. It doesn't hide assumptions. Every input is visible. Every output is auditable. A finance partner or board member can follow the math.
Most AI ROI models fail because they either:
- Forget hidden costs. They use only the OpenAI API invoice and ignore retries, infrastructure, human review, and observability—which together often exceed the API cost.
- Assume perfect accuracy. They model 95% first-touch resolution when reality delivers 70%. When the real numbers emerge, the ROI collapses.
- Use engineering metrics as proxies for business value. They celebrate "25% latency improvement" and forget to convert it to "cost per outcome."
- Hide assumptions. The model is a black box, and the CFO has no way to challenge it. When ROI doesn't materialize, there's nowhere to look.
The Runrate calculator avoids all four traps.
On hidden costs: The calculator explicitly asks you to model retries, escalation, human review, infrastructure, observability, and vector database cost. If you don't include them, the forecast is incomplete.
On accuracy: The calculator doesn't assume 95% accuracy. It asks you: "What's your realistic first-touch resolution rate?" and models human review and escalation as explicit cost lines. If your agent is 80% accurate and requires 20% human review, that review cost is in the model.
On business value: The calculator translates everything to cost per outcome, which is the business metric. It doesn't care about tokens or latency. It cares about: "What did we pay to deliver this result?"
On transparency: Every input drives the output. The CFO can see: "Our baseline is $22 per ticket. With AI at 80% FCR and 20% escalation, we're at $8 per ticket. That's a 64% reduction. At 2,000 tickets per month, that's $28,000/month saved. Our agent infrastructure cost is $12,000/month, for a net savings of $16,000/month."
Methodology: The Cost Model
Here's the exact math the calculator runs.
Step 1: Calculate baseline cost per outcome.
Baseline = (Total annual cost of current process) / (Annual volume of outcomes)
Example: 12 CSRs at $50k salary + 30% benefits = $65k fully loaded. 12 × $65k = $780k annually. 2,400 tickets per month = 28,800 per year. Baseline = $780k / 28,800 = $27.08 per ticket.
Step 2: Model total AI cost per outcome.
This is the tricky part because AI costs are multi-layered.
Total AI cost per outcome = (API inference cost + retry overhead + human review cost + infrastructure cost + observability cost + vector database cost) / (number of outcomes)
Let's say you deploy an agent for customer support:
- API inference cost per call: You model the average number of tokens per call and the cost per token. If your agent averages 800 input tokens + 200 output tokens, and the rate is $0.001 per 1K tokens, cost per call = (800 + 200) × 0.001 / 1,000 = $0.001 per call. But you also need the model's cost per token—Claude 3.5 Sonnet, for instance, costs $0.003 per input token and $0.015 per output token, for a typical 2,000-token conversation = about $0.015 per call.
- Retry overhead: If 25% of calls fail and retry, multiply API cost by 1.25. If you use prompt caching and get a 40% cache hit rate, multiply by 0.8. Net: $0.015 × 1.25 × 0.8 = $0.015 per call (in this case they wash).
- Human review cost: If 20% of decisions require human review at an average review time of 2 minutes, that's 2 minutes × monthly volume × 20% / 60 = human hours. At $75/hour loaded, that's $75 / 60 × 2 × volume × 0.2 = $0.50 per reviewed outcome.
- Infrastructure cost: $3,000/month for gateway, rate limiting, caching. At 2,000 outcomes/month, that's $1.50 per outcome.
- Observability cost: $1,200/month. At 2,000 outcomes/month, that's $0.60 per outcome.
- Vector database cost: $800/month. At 2,000 outcomes/month, that's $0.40 per outcome.
Total AI cost per outcome = $0.015 + $0.50 + $1.50 + $0.60 + $0.40 = $3.015 per outcome.
Step 3: Calculate cost reduction and annual savings.
Cost reduction = (Baseline - AI cost) / Baseline = ($27.08 - $3.02) / $27.08 = 89%.
Annual savings (gross) = (Baseline - AI cost) × annual volume = $24.06 × 28,800 = $692,928.
Step 4: Calculate payback and ROI.
Infrastructure investment (one-time): $50,000 (for building, testing, deploying the agent).
Payback period = Infrastructure investment / (monthly savings) = $50,000 / ($692,928 / 12) = $50,000 / $57,744 = 0.87 months (about 3 weeks).
Year-1 ROI = (Annual savings - annual recurring cost) / Infrastructure investment = ($692,928 - $144,000) / $50,000 = 1,098% ROI (or 11x return).
Step 5: Sensitivity analysis.
The calculator re-runs the math for:
- Baseline +/- 10%
- AI accuracy -5%, -10%, -15% (driving higher escalation)
- API costs +15%, +25%, +50%
- Volume +/- 20%
This shows the CFO the range of plausible outcomes, not a false point estimate.
What Inputs You Need to Gather
To use the calculator, you'll need to work with your operations team to gather:
- Current headcount and cost for the function. Total annual payroll + benefits + overhead for the team doing this work.
- Annual and monthly volume of outcomes. How many tickets, claims, or applications per month?
- Current cost per outcome. Total cost ÷ volume.
- Your target use case. What process are you automating?
- AI accuracy estimate from testing. What's your realistic first-touch resolution rate?
- Escalation and human review percentage. What percent of cases need human intervention?
- Expected AI infrastructure cost. API spend + retries + observability + vector database. The calculator provides templates.
If you don't have some of these (especially current cost per outcome), the calculator will guide you through the math to derive it.
Common Scenarios
The calculator comes pre-loaded with templates for four common use cases:
- Customer Support. First-touch resolution agent. Baseline: $22/ticket. Realistic AI cost: $4-8/ticket (depending on accuracy and review overhead).
- Claims Processing. Triage and initial adjudication agent. Baseline: $35/claim. Realistic AI cost: $10-18/claim.
- Loan Origination. Initial screening and document processing. Baseline: $150/application. Realistic AI cost: $25-50/application.
- Back-office Processing. Invoice processing, reconciliation, etc. Baseline: $12/invoice. Realistic AI cost: $2-5/invoice.
Each template shows typical accuracy rates, escalation percentages, and infrastructure costs so you have realistic starting points.
Running the Calculator
The calculator lives at /tools/ai-roi-calculator. It's interactive—you enter inputs in a form, and it updates outputs in real time. You can download the results as a PDF (great for board decks) or as an Excel file if you want to extend the model.
The entire calculation is transparent. If your CFO wants to know "how did you get to $4.2 per outcome," you can trace every line of the formula. That transparency is where trust comes from.
Most finance leaders run the calculator 2-3 times with different scenarios (conservative, realistic, optimistic) and use the realistic case as the basis for a board recommendation. The conservative case becomes the risk disclosure.
Start with a use case where you have clear cost data today. The easier the baseline is to defend, the more credible the AI case will be.
Calculate your AI ROI.
See what your agents actually cost — and what they're returning.
Was this article helpful?