Runrate Framework
AI Workforce P&L
Treat AI agents like employees: cost structure, productivity target, and retirement trigger per agent.
Read the full framework →Runrate Framework
The AI Cost Iceberg
Visible API spend (10%) vs hidden inference, storage, observability, retries, human review (90%).
Read the full framework →Runrate Framework
5-Stage AI Cost Maturity Curve
From Invisible → Tracked → Allocated → Optimized → Governed — where does your org sit?
Read the full framework →Today's CFOs treat AI the way accountants treated electricity in 1910: as an opaque infrastructure cost buried in overhead. But AI agents don't behave like infrastructure—they behave like labor. The thesis is simple: if you're deploying AI agents to do work, you need payroll-equivalent systems to track, price, and govern that work.
The current frame is broken
For the past two years, CFOs have tried to fit AI into existing mental models. The software model made sense at first: you buy a model license (or API credits), you integrate it, you maintain it. CapEx + operational expense. Predictable, contained, auditable.
But that frame collapses when you actually deploy agents at scale.
A single Anthropic Claude Sonnet API call costs $0.003. Runrate's customers see individual agent deployments that run 15,000 API calls per day. At that volume, you're paying $45/day, or $1,350/month, for one agent. Add human review overhead (a person reviewing 10% of the agent's work at $35/hour), retries on failure, vector database storage, observability tooling, and integration tax (Stripe calls, Twilio lookups, Salesforce API cost), and that same agent costs your company $8,500/month to operate.
The software model would have you budget this as "AI API spend" and move on. But if your customer service team spans 8 agents, you've just added $68,000/month to your OpEx without a clear attribution line. The FinOps team can tell you that the API calls went up. The CFO can't tell you which customer segment is actually being served.
This is the problem the workforce thesis solves.
The new frame: AI as labor, not infrastructure
Here's the reframe. An AI agent is not a software license. It's a worker.
That worker has a compensation cost (the API bill + the inference overhead), a productivity level (resolutions per day, accuracy rate), a skill set (can it handle edge cases? does it need human review?), and a useful lifespan (when does the newer model version make this agent obsolete?).
Treat it exactly as you'd treat a human CSR on your team.
A human CSR costs your company roughly $52,000/year in salary, plus $20,000 in benefits, plus $15,000 in overhead (office space, software licenses, training), plus another $8,000 in hiring and turnover friction. That's $95,000 loaded cost per FTE.
An AI agent—Claude-powered, running 50,000 resolutions per year at $1.50 per resolution (the true cost, not just the token cost), plus $30,000 in operational overhead, plus $20,000 in integration and training costs amortized—comes in around $125,000 per year fully loaded.
They're in the same cost ballpark. But they operate under completely different unit economics. The human scales linearly with customer growth (one CSR per 50 tickets). The AI agent scales nonlinearly (one agent can handle 1,000 tickets after you invest in prompt engineering). The human costs the same whether she's busy or idle. The AI agent scales with actual work performed.
This is why the software frame fails. You can't build financial governance around something that looks like infrastructure but has the economics of labor.
What payroll-equivalent infrastructure looks like
Every payroll system has four core components. AI labor needs the same.
Timecards — In payroll, a timecard tracks which project a person spent her 40 hours on. In AI, you need to track which work item (a support ticket, an insurance claim, a loan application) each agent interaction belongs to. This is work-item-level cost attribution. Without it, you have a cost but no P&L owner.
Payroll system — In payroll, the payroll system (ADP, Workday) is the single source of truth for compensation liability, tax withholding, and period-end reporting. In AI, you need an equivalent: a cost attribution system that ingests API logs, matches them to work items, adds human review and infrastructure overhead, and produces P&L reports by business unit, customer, or campaign. This is what Runrate does.
1099 vs. W-2 classification — In payroll, a 1099 contractor costs less than a W-2 employee because you don't manage benefits or compliance. In AI, the analog is: third-party API (like OpenAI's API, which you pay per call) vs. self-hosted (like running an open-source LLaMA model on your own GPU cluster, where you pay for infrastructure upfront and inference is "free" incrementally). The financial model is completely different. You need to know which agents are third-party and which are self-hosted, and you need to cost them differently.
Performance reviews — In payroll, you review whether a person is earning their salary through productivity, quality, and skill growth. In AI, you review whether an agent is earning its cost-per-outcome target. If accuracy slips from 94% to 87%, or if the cost per resolution creeps from $1.40 to $2.10, those are performance signals that the agent needs retraining, retirement, or replacement.
Add a fifth: retirement triggers. When GPT-4 shipped, teams running GPT-3.5-based agents had to decide: do we keep paying $0.005 per token for the old model, or do we migrate to $0.015 per token for the new one? Migrating costs engineering effort and testing; not migrating means you're running outdated models. Those decisions need to be made at the CFO layer, not the engineering layer, and they require clear cost-per-outcome baselines to justify.
Why this isn't a metaphor—the financial implications are real
Some executives hear "AI is the new payroll" and think it's a cute analogy. It's not.
The reason payroll systems exist is because labor cost is the largest line item on most operating P&Ls (often 50-70% of revenue in service businesses), it's highly visible to auditors, and the tax and compliance risk is enormous. You don't manage payroll by gut feel because the stakes are too high.
AI spend will hit the same magnitude. CloudZero's 2025 survey found that AI spend per company rose 36% year-over-year, from $62,964/month to $85,521/month. For a $100M revenue company running 50+ agents in production, AI labor cost could easily exceed $2M annually.
At that scale, the governance requirements become non-negotiable. Your board will ask:
- How much did it cost to serve this customer with AI agents?
- What's our cost per resolved work item by agent, by skill level, by business unit?
- Which agents are above or below their cost-per-outcome target?
- When does the newer model version justify migrating old agents?
Without a payroll-equivalent system, you can't answer any of those questions. You're left with "our AI API spend is $85,521/month" and no idea whether it's profitable or not.
What changes for the CFO function
If you accept the thesis—that AI agents are labor, not infrastructure—then the CFO function changes in four ways.
First: you hire a new operating model. Your CFO needs an "AI Payroll Manager" equivalent: a finance leader who understands agent deployment, can read API logs, and can translate engineering metrics (accuracy, latency, tokens per call) into business metrics (cost per outcome, ROIC per agent, ROI per campaign).
Second: you build a new line on the P&L. Instead of burying AI cost in "software and cloud," you create a specific "AI labor" or "agent workforce" line, broken down by customer segment, product line, or business unit.
Third: you establish a cost-per-outcome KPI for every agent in production. Runrate's average customer baseline is $1.80 per resolution for a claims adjudication agent, $0.92 per response for a customer service agent. You need to benchmark against these and set improvement targets.
Fourth: you create a monthly board report that answers: how much did AI labor cost this month, by business unit? What's the trend? Are agents improving or deteriorating on cost per outcome? Which agents are candidates for retirement or retraining?
The transition starts with visibility
This doesn't happen overnight. Most teams are still at stage 1 or 2 of the AI Cost Maturity Curve—AI spend is either invisible or tracked at the account level only, with no work-item attribution.
The first step is the same as any payroll implementation: visibility. You need to see, in one dashboard, what every AI agent is costing you on a per-work-item basis. Until you have that, all other governance is guessing.
Once you have visibility, the rest follows: allocation to P&L owners, optimization toward cost-per-outcome targets, and finally, board-grade governance with SLOs and automated anomaly detection.
The AI Workforce P&L framework—detailed in our financial model guide—walks through the line items and the operational model that makes this work.
What to do next
Start with a single high-volume agent: your customer service chatbot, your claims adjudication AI, your loan origination assistant. Map every call that agent makes to the work item it serves, add up the full cost (API + human review + overhead), and calculate the cost per outcome. That one number—$1.42 per claim, $0.87 per ticket—becomes your baseline. Everything else flows from that.
Curious where your team sits on the maturity curve? Take the 15-question self-assessment and get a personalized report.
Go deeper with the field guide.
A step-by-step PDF for implementing AI cost attribution.
Was this article helpful?