Runrate Framework
5-Stage AI Cost Maturity Curve
From Invisible → Tracked → Allocated → Optimized → Governed — where does your org sit?
Read the full framework →FinOps Foundation built a discipline that works brilliantly for cloud infrastructure cost: you can drill from total spend down to a specific compute instance and see its cost line-by-line. But AI agents break that model. A single resolved support ticket touches OpenAI, a vector database, observability infrastructure, and a human reviewer. Traditional FinOps tools see these as disconnected cost centers. They don't see the ticket. This gap is why 73% of companies deploying AI agents still can't answer "what did this work item cost?" and why CFOs end up watching the iceberg instead of the tip.
The problem with applying traditional FinOps to AI
Traditional FinOps was built for cloud infrastructure. The hierarchy is clean: Organization > Business Unit > Project > Environment > Service > Instance. Each level has a cost tag. You can ask "how much did production EU cost in March?" and the tool gives you the answer. The cost is proportional to consumption: more compute hours = more cost. More storage gigabytes = more cost. You can forecast by historical burn rate and apply discount factors for commitments.
AI agents invert this model. A single agent work item (one support ticket, one claims decision) is the unit of business value. But that work item may trigger costs across multiple vendors and cost centers simultaneously:
- The API call to OpenAI (visible cost, usually 5-15% of total)
- Retrieval-augmented generation (RAG) vector database queries (hidden cost)
- Observability platform logging (hidden cost)
- If the agent fails and retries, a retry-storm cost spike (hidden, bursty)
- Tool calls to third-party APIs—Stripe, Twilio, Salesforce (hidden, variable)
- Human-in-the-loop review for ambiguous cases (hidden, overlapped with salary)
A traditional FinOps platform sees five disconnected expenses. It doesn't see one work item that generated all five. It can't answer "cost per resolved ticket" because it has no concept of "resolved ticket." It knows API cost, storage cost, observability cost. It doesn't know if the agent actually worked.
Why the cost hierarchy breaks
Traditional FinOps uses an organizational cost hierarchy: cost flows down from the org to business units to teams to projects. Your CFO sets a "cloud budget of $2M/month," allocates $400k to sales, $800k to product, $800k to ops. Each team knows their budget. Each team can drill down to their instances and optimize.
This works for cloud because cost is proportional to resource consumption and resource consumption is relatively stable. A sales team's Salesforce infrastructure (servers, storage, networking) has predictable cost. An ops team's data warehouse has measurable cost per GB.
AI agents need a work-item cost hierarchy: cost flows backward from the outcome. Your agent completes 50,000 support tickets per month. Those 50,000 tickets collectively cost $250,000. Cost per ticket: $5. That's your unit economics. If you run 100,000 tickets next month, cost may be $450,000 (not $500k) because you hit better cache hit rates and reuse prompts. Cost per ticket dropped to $4.50.
The traditional CFO doesn't think this way about cloud. They think "we allocated $X to cloud, and consumption is up or down." But this is exactly how they think about headcount. "We budgeted 50 FTEs in support. They cost us $4M. We process 200k tickets per year. Cost per ticket is $20 in salary." AI agents are the new payroll. They need payroll-grade cost tracking, not cloud-grade.
Three ways traditional FinOps fails at agent cost visibility
1. It can't attribute cost to an outcome. Your FinOps tool logs: "OpenAI API call used $0.08 in tokens on timestamp 2026-05-02T14:32:15Z, user_id=agent-support-v2." That's all the tool knows. It doesn't know whether the call resulted in a ticket being resolved, a customer being satisfied, or a retry loop that wasted $0.40. Traditional FinOps has no concept of "outcome." It only has "resource consumed."
2. It can't cross-vendor cost attribution. Your AI agent architecture looks like this: OpenAI (inference) → Pinecone (vector DB) → Datadog (observability) → Stripe (toll transaction) → manual review (human cost, embedded in salary). Each vendor has a different cost model and billing schedule. OpenAI bills by token count. Pinecone bills by stored vectors. Datadog bills by ingested logs. Your traditional FinOps platform probably talks to 1-2 of these vendors, not all 5. The tool can't draw a line connecting all five costs to one work item. So the CFO sees five separate line items on five different vendor invoices and has no way to know if the agent was expensive or cheap.
3. It can't handle human-in-the-loop cost. If your claims agent processes 100 claims per day and hands 30 of them back to human reviewers (because they were ambiguous or high-risk), those 30 have a labor cost embedded in FTE cost, salary, benefits. The agent didn't process those claims end-to-end. But traditional FinOps has no way to track "this claim cost $8 in API inference, $0.50 in vector DB, and $12 in human review time." It sees the API cost and the salary line. It has no bridge.
The maturity curve problem: You can't move past stage 2 without agent-level thinking
The 5-Stage AI Cost Maturity Curve shows this gap:
Stage 1 (Invisible): AI spend is buried in shadow charges and contractor invoices. Traditional FinOps can't help you here; you have no data.
Stage 2 (Tracked): AI spend is on your tech bill under "API subscriptions" or "cloud platforms." Traditional FinOps can get you here. You have a line item. You can see it trending. CloudZero is comfortable here.
Stage 3 (Allocated): AI spend is split by team. "Sales spent $40k on AI, product spent $120k." Traditional FinOps can approximate this if you tag instances or add chargeback logic. But the cost is still not tied to outcomes, so the allocation feels arbitrary.
Stage 4 (Optimized): AI spend is tied to a specific work item with a cost-per-outcome KPI. "Support runs at $0.42 per ticket, claims runs at $7.20 per claim." This is where traditional FinOps breaks entirely. There is no framework for "what did this ticket cost?" in traditional FinOps because traditional FinOps was never built for outcomes. It was built for resources. Runrate exists because this stage requires agent-native thinking.
Stage 5 (Governed): Cost SLOs, automated anomaly detection, board reporting, policy enforcement. You can't get here without stage 4.
Most CFOs are trying to jump from stage 2 to stage 4 using traditional FinOps tools. They're using a cloud-infrastructure framework to answer an agent-economics question. It doesn't work.
Why CloudZero and Vantage can't fill the gap alone
CloudZero is the engineer-beloved FinOps platform. It's excellent at cloud cost visibility: "show me EC2 cost by region by environment by team." But their strength is also their constraint. They were built for engineers asking "how do I optimize compute?" They're starting to add AI cost visibility, but they still frame the problem as "tokens per dollar" and "model efficiency"—engineering questions, not business questions.
Vantage has the best UX in the cloud cost space. But like CloudZero, they were built for the cloud-infrastructure hierarchy. Their reporting is built around "which service cost the most?" not "which work item?" They don't have a concept of a claims processor that costs $6 per claim.
Neither platform has the agent-attribution DNA. They're like trying to use a cloud load balancer to route support tickets—you're asking the wrong tool to solve the wrong problem. CloudZero and Vantage will tell you what you spent on inference. They won't tell you what you got for that spend.
What agent-native cost attribution requires
To answer "what did this work item cost?" you need:
-
An outcome definition. What is a "resolved ticket," a "processed claim," a "qualified lead"? How do you count it? Does a ticket that gets escalated to manual review count as "resolved"?
-
Cross-vendor cost tracking. Pull cost data from OpenAI, Anthropic, Pinecone, Weaviate, Datadog, and your internal service mesh. Correlate them by timestamp and session ID.
-
Attribution logic. Decide how to distribute cost when multiple work items use the same vector database record, or the same cached prompt, or the same observability infrastructure. This is non-trivial.
-
Human cost modeling. If a human reviews 20% of the agent's output, what portion of the human's cost gets allocated to the agent?
-
Anomaly detection tied to outcomes. Flag when a work item costs 3x the historical median, not just when a single cost center spikes.
This is agent-native. Traditional FinOps can't do it because it has no concept of agents or outcomes. This is why Runrate exists: to be the operational layer that connects FinOps Foundation principles to the agentic enterprise.
Explore the full FinOps for AI framework in the pillar article.
Go deeper with the field guide.
A step-by-step PDF for implementing AI cost attribution.
Was this article helpful?