Cost per Action vs Cost per Token: Which LLM Pricing Model Saves You Money?

Cost per Action vs Cost per Token: Which LLM Pricing Model Saves You Money?

When you use an LLM like GPT-4 or Claude, you’re not paying for a conversation. You’re paying for tokens-tiny chunks of text the model processes. But what if you could just pay for the result instead? Like paying $5 to get a contract reviewed, not $0.03 for every 1,000 words you type in? That’s the difference between cost per token and cost per action. One is technical. The other is business-friendly. And right now, most companies are stuck with the first one-whether they like it or not.

How Cost per Token Actually Works

Cost per token is the default pricing model for every major LLM API: OpenAI, Anthropic, Google, and others. You’re charged for every token your prompt uses and every token the model generates. A token isn’t a word. It’s often a piece of a word. In English, 1,000 tokens roughly equals 750 words. But it’s not consistent. "Don’t" is one token. "Unbelievable" might be three. Numbers and punctuation count too.

Here’s the catch: input and output tokens cost different amounts. Anthropic’s Claude Sonnet 4.5 charges $3 per million input tokens but $15 per million output tokens. That’s a five-times difference. Why? Because generating text requires the model to predict each token one after another-slow, compute-heavy work. Reading your prompt? That’s faster, parallelizable. So the math isn’t fair, and it’s not intuitive.

And then there are "reasoning tokens." These are internal steps the model takes to think through a problem-like working out a legal argument or solving a math equation. Anthropic and others charge extra for these, often at higher rates than output tokens. A user on Reddit reported their bill jumped 5x after processing legal documents because the model was using dozens of reasoning tokens per response. They didn’t even know these existed until their invoice arrived.

For a legal tech startup processing 1,000 contracts a month, each averaging 10,000 input tokens and 500 output tokens, the monthly cost is around $37.50. Sounds low? Now imagine 10,000 contracts. Or 100,000. Or 1 million. Costs scale fast. And if your output suddenly goes from 500 tokens to 5,000 because the model over-explains? Your bill triples overnight.

Why Cost per Token Feels Like a Trap

The biggest complaint from business teams? Unpredictability. You can’t budget for it. You don’t know how long a summary will be. You don’t know if the model will get stuck in a loop. You don’t know if it’ll add 200 extra tokens of fluff just to sound thorough.

Traceloop surveyed 247 enterprise AI teams in April 2025. 68% said their biggest cost challenge was tracking token usage across features, users, and departments. They needed to know: "Is this $1,200 bill coming from customer support chats? Or from internal reports?" But token data doesn’t map to business outcomes. It maps to code.

Developers can optimize. They can rewrite prompts to be shorter. They can use caching so repeat questions don’t cost anything. One SaaS company cut their monthly LLM bill from $1,200 to $380 just by trimming prompts and reusing responses. But that’s not something a marketing manager or paralegal can do. They just want a contract summary. They shouldn’t need a computer science degree to get one.

And the documentation? It’s messy. OpenAI’s cost guide rates 4.2/5 on Capterra. Anthropic’s? 3.8. That gap isn’t about quality-it’s about clarity. One explains how tokens work. The other doesn’t tell you why your bill spiked.

A single contract under a lamp with a  price tag floating above, tokens fading into smoke, metalpoint illustration.

What Cost per Action Could Look Like

Cost per action flips the script. Instead of paying for tokens, you pay for a task. Contract review: $5. Summarize a 50-page report: $3. Extract all names and dates from an email: $1. No counting. No surprises.

This isn’t science fiction. Jasper.ai rolled out "Content Generation Packs" in April 2025: $20 for 10 blog posts. Harvey AI launched "Legal Task Units" in May: $12 per contract reviewed. These aren’t experiments-they’re products. And they’re selling.

The upside? Budgeting becomes simple. A legal team can say, "We allocated $5,000 for contract reviews this quarter," and know exactly how many they can process. No one needs to monitor API logs. No one needs to learn what a "token" is. The finance team doesn’t need to justify a $1,200 spike-they just see $12 per task.

But there’s a flip side. If your contract is short and simple, you still pay $12. If the model gets it right in 200 tokens, you still pay $12. If it takes 2,000 tokens to get it right? Still $12. For the provider, that’s risky. For the customer, it’s a win-unless your task is complex and the model fails. Then you pay again.

That’s why per-action pricing only works for well-defined tasks. You can’t say "write me a novel" and expect a fixed price. But "extract all dates from this invoice"? That’s perfect. "Classify this support ticket as billing, technical, or account"? Perfect.

Which Model Fits Your Use Case?

Here’s how to pick:

  • Use cost per token if: You’re building something exploratory. You need flexible outputs. You’re testing ideas. Your prompts vary wildly. You have engineers who can optimize prompts and monitor usage. Startups and developers often fall here.
  • Use cost per action if: You’re automating repetitive, well-defined tasks. You have non-technical teams using the system. You need predictable monthly costs. You’re in legal, HR, healthcare, or compliance. Enterprises with high-volume, low-variability workflows prefer this.

Google’s Vertex AI Model Optimizer is trying to bridge the gap. You don’t pick a model. You pick an outcome: "Low cost," "High accuracy," or "Balanced." The system picks the best model and pricing behind the scenes. It’s not per-action yet-but it’s moving that way.

A tree with two trunks: one tangled with tokens, the other bearing dollar sign fruit, drawn in metalpoint.

The Future: Hybrid Models Are Coming

Per-token pricing isn’t going away. Not yet. 92% of LLM providers still use it as their main model, according to Grumatic’s March 2025 data. But the tide is turning.

Why? Because costs are dropping. Andreessen Horowitz found per-million-token prices fell 68% from late 2023 to early 2025. That means providers can afford to experiment. They can start bundling. They can offer flat-rate tiers for common tasks.

Look at what’s happening: Anthropic now lets you set "effort controls"-limit how many reasoning tokens the model uses. That’s not per-action, but it’s a step toward outcome-based pricing. Providers are learning that customers don’t care about tokens. They care about results.

By 2028, Binadox predicts 35% of enterprise LLM usage will run on per-action or hybrid pricing. Legal, healthcare, and finance will lead the shift. Why? Because those industries need audit trails, fixed budgets, and compliance. Tokens don’t give you that. Task pricing does.

What You Should Do Right Now

If you’re using LLMs today:

  1. Track your token usage. Use tools like Traceloop or Langfuse. Know your cost per task, even if you’re on token pricing.
  2. Look for providers offering fixed-price task bundles. Jasper and Harvey aren’t the only ones-check what Anthropic and Google are testing.
  3. Ask your vendor: "Can I pay per contract review, not per token?" If they say no, ask when they’ll offer it.
  4. For simple, repeatable tasks, start building your own "action" layer. Write a script that says: "If input is this, output is that, cost is $X." It’s not perfect, but it’s better than guessing.

The goal isn’t to avoid tokens. It’s to stop letting them control your budget. The best LLM strategy isn’t about finding the cheapest model. It’s about aligning your pricing model with your business goals. Tokens are the engine. Outcomes are the destination. Stop paying for the engine. Start paying for the ride.

Comments

  • Ben De Keersmaecker
    Ben De Keersmaecker
    December 22, 2025 AT 21:39

    I’ve been tracking token usage for our legal docs and honestly? The surprise spikes are insane. One day we did 200 contract reviews, next day it jumped to 1,200 because someone pasted a 50-page PDF and the model started summarizing every footnote. No one told us it’d do that. Now we’re using Langfuse to cap output tokens-saved us 40% last month.

  • Aaron Elliott
    Aaron Elliott
    December 23, 2025 AT 01:09

    One must ask: is the commodification of linguistic output not a profound epistemological regression? We have reduced the sublime act of reasoning to a commodity ledger, where the soul of language is measured in subword units. The token is the new dollar bill-arbitrary, opaque, and violently divorced from meaning. The per-action model is not progress-it is capitulation to the aesthetics of efficiency.

  • Chris Heffron
    Chris Heffron
    December 23, 2025 AT 04:28

    lol at the 5x cost difference between input and output tokens 😅
    Also, "reasoning tokens"?? Who came up with that? Sounds like a marketing buzzword to justify higher bills. I just want my contract summary without a PhD in AI accounting.

  • Adrienne Temple
    Adrienne Temple
    December 23, 2025 AT 05:37

    My team just switched to Harvey AI’s contract review packs-$12 per doc, no questions asked. We used to have engineers spending hours optimizing prompts. Now our paralegals just upload and go. No more "why is our bill $1,800 this month?!" emails. I wish all tech tools were this simple. 😊

  • Sandy Dog
    Sandy Dog
    December 23, 2025 AT 07:08

    Okay but imagine this: you pay $5 for a contract review… and the AI gives you a 3000-word essay on medieval contract law instead of the summary you asked for. You’re STILL charged $5. No refund. No apology. Just a robotic voice saying "task completed." Meanwhile, your boss is yelling because the client got a 10-page philosophy paper instead of a clause list. This isn’t pricing-it’s a trap wrapped in a shiny UI. I’m not even mad… I’m just disappointed. 💔

Write a comment

By using this form you agree with the storage and handling of your data by this website.