Optimize guide10 minUpdated 2026-05-06

What OpenClaw actually costs — and how to cut the bill.

OpenClaw itself is free. The LLM API calls it makes are not. A casual personal-assistant setup runs $10–30/mo. A 24/7 multi-agent setup with browser automation can hit $300–600. Here's what determines which end of that range you land on, and the seven settings that move the needle most.

Quick answers

  • How much does OpenClaw cost per month?

    $10–$30 for personal use on Sonnet with reasonable settings. $30–$80 for an active team agent. $300–$600 for an unoptimized 24/7 multi-agent setup. The runtime itself is free; the bills are LLM API calls.
  • How can I make OpenClaw cheaper?

    Seven levers: route heartbeats to Haiku or Ollama, route by task complexity, enable Anthropic prompt caching, trim context, use the Batch API for non-urgent jobs, set hard quotas, run local models for 60–80% of traffic. Real bills typically drop 60–90%.
  • What costs the most in OpenClaw?

    Heartbeats — the periodic check-ins the gateway sends. They run whether you're using the agent or not, and on a default setup are 60–80% of total token volume. Routing them to Haiku or a local model is the single biggest win.
  • Does OpenClaw have a free tier?

    The runtime is free. For free LLM inference, use Ollama with a local model — no API costs at all. For free hosting, Oracle Cloud Free Tier (2 ARM vCPU, 12 GB RAM) genuinely works. Combined: $0 setup.
  • Which model is cheapest for OpenClaw?

    Free: any Ollama model (zero per-token cost). Cheapest cloud: Claude Haiku 4.5 ($1/$5 per M input/output) or GPT-4o-mini ($0.15/$0.60). Use these as defaults; escalate to Sonnet only when reasoning quality matters.

Anatomy

Where the cost goes

The bill is almost entirely LLM API tokens. The runtime, the gateway, the storage — all free. So the question is just: which API calls are you making, with which model, how often, and how much context per call?

60–80%

Heartbeats

10–20%

User chat

10–25%

Tool calls + browser

<5%

Memory / scheduled

Most people assume their bill is driven by chat volume. It almost never is. The single biggest line item on a default setup is heartbeats — periodic check-ins the gateway sends so the agent can react to scheduled events. They run whether you're using the agent or not.

Your numbers

Cost calculator

Drag the sliders to match your situation. The model dropdown uses published 2026 list prices.

OpenClaw cost calculator

50/day
2,500 tokens
400 tokens
50%

Estimated monthly cost

$41.11

  • Input tokens$6.19
  • Output tokens$9.00
  • Heartbeats$25.92

Estimates based on published 2026 list prices for each model. Real bills land within ±25% in our experience.

Reality check

Real bills land within ±25% of the calculator's estimate in our experience. The biggest miss factor is browser snapshots — they're token-heavy and hard to predict.

Lever 1

Heartbeats: the silent killer

The default heartbeat interval is 3 minutes. That's 480 calls per day, ~14,400 per month. Each one consumes 1–2k input tokens (the agent's running context). On Sonnet 4.6, that's ~$15/mo on heartbeats alone — before any actual work.

The fix is two-part:

  • Route heartbeats to a cheap model. Haiku, GPT-4o-mini, or a local Ollama model. Quality doesn't matter; this is just a "is anything pending?" check.
  • Increase the interval. If your agent doesn't have scheduled jobs, 15 minutes is fine. If it does, match the heartbeat to the cron resolution.
~/.openclaw/openclaw.jsonjson
{
  "heartbeat": {
    "intervalMinutes": 15,
    "model": "claude-haiku-4-5"
  }
}

Lever 2

Model routing

The single most impactful change after fixing heartbeats: stop using the same flagship model for everything. 70–80% of an agent's calls don't need flagship-quality reasoning.

TaskDefault modelOptimized
HeartbeatsSonnetHaiku or local
Inbox triageSonnetHaiku
Daily digestSonnetHaiku
Web search summarySonnetHaiku
Browser tasksSonnetSonnet (keep)
Long-form draftingSonnetSonnet (keep)
Code generationSonnetSonnet (keep)

Full routing setup including local models in the Ollama guide.

Lever 3

Prompt caching

Anthropic's prompt cache gives you 0.1x input pricing on repeated content. OpenClaw loads the same MEMORY.md and system prompt every turn — those are perfect cache candidates.

~/.openclaw/openclaw.jsonjson
{
  "anthropic": {
    "promptCaching": true,
    "cacheBreakpoints": ["system", "memory"]
  }
}

In benchmarks, this drops monthly bills 30–50% by itself for memory-heavy agents. The cache TTL is 5 minutes, so it works best when the agent is actively used (multiple turns close together) rather than once-an-hour pings.

Lever 4

Context discipline

Every file you load into the agent's context costs money. Most people over-load context because it's easy and the cost isn't visible per-call.

  • Don't dump everything into MEMORY.md. Use memory_search for the things the agent needs occasionally; reserve MEMORY.md for facts loaded every session.
  • Trim system prompts. If your prompt is 4k tokens, half of that is probably redundant.
  • Compact aggressively. The default compaction threshold is generous; bring it down to 50% for active sessions to keep average context smaller.
  • Use vector recall over file dumps. Loading a 50k-token doc into context vs searching it for the 500 relevant tokens is a 100x cost difference.

Lever 5

Batch API for non-urgent jobs

Anthropic's Batch API gives you 50% off both input and output if you can wait up to 24 hours for the result. OpenClaw cron jobs that don't need immediate output should use it.

WorkflowLatency tolerable?Batch fit?
Daily 09:00 digest24h is fineYes — submit at midnight
Weekly research compilationDays fineYes
Inbox triage every hour<5minNo
User chatReal-timeNo
Archive analysis24h fineYes
cron jobjson
{
  "schedule": "0 0 * * *",
  "task": "Compile daily digest",
  "options": { "batch": true, "deliverBy": "09:00" }
}

Lever 6 + 7

Set hard quotas

Two quotas, two layers. The first is OpenClaw's soft cap — it warns and pauses when hit, but a misbehaving agent can route around it. The second is the provider's hard cap on the API key itself, which stops calls dead.

OpenClaw soft capjson
{
  "quotas": {
    "maxMonthlySpend": 50,
    "warnAt": 0.8,
    "pauseAt": 1.0
  }
}

Then in the Anthropic console: Settings → API keys → spending limit. Set it 50% above your soft cap as a safety net.

Set both

The soft cap saves you from accidents. The hard cap saves you from disasters. We've seen runaway loops burn $200 in 20 minutes — set the hard cap.

FAQ

How much does OpenClaw really cost per month?
$10–$30 for personal use on Sonnet with reasonable settings. $30–$80 for an active team agent. $300–$600 for an unoptimized 24/7 multi-agent setup. The calculator above gives you a real estimate from your usage.
What's the single biggest cost lever?
Heartbeat traffic. They run every few minutes whether you're chatting or not, and over a month they're often 60–80% of total tokens. Routing them to Haiku or a local model drops bills 40–70% on its own.
Does prompt caching actually work?
Yes — Anthropic charges 0.1x base input rate for cached tokens. Cache hits become huge for agents that load the same MEMORY.md and system prompt every turn. Real-world cache hit rates of 50–80% are normal once enabled.
Should I use the Batch API for OpenClaw?
For non-urgent work yes. Anthropic offers 50% off both input and output for batched requests with up to 24h latency. Daily digests, weekly research compilations, archive analysis — all great fits.
Are subscription tokens (Claude Pro/Max) usable?
Briefly they were. As of April 2026, Anthropic excludes third-party tools from subscription quotas. OpenAI never allowed it. Plan on real API billing.
How do I set a hard cap that actually stops spending?
Set the spending limit on the API key itself in your provider's console — that's the brake the provider enforces. OpenClaw's internal `quotas.maxMonthlySpend` is a soft warning, not a hard stop.

Want OpenClaw without the ops?

Provision is the managed OpenClaw cloud — agents, channels, browser, and skills, all running. $99/mo. 48-hour free trial.