- what OpenClaw is,
- why model costs can grow quickly,
- why SaladCloud is a practical way to host models for OpenClaw.
What OpenClaw is
OpenClaw is the orchestration layer around your model, tools, memory, and channels. The model is the “brain,” while OpenClaw handles:- conversation and session state,
- tool execution and workflow logic,
- channel integrations,
- scheduling and background activity.
Why costs increase with API-hosted models
OpenClaw itself is free, but model inference is not. With traditional API providers, billing is typically token-based. Costs tend to grow because of:- accumulated conversation and tool context,
- repeated background/heartbeat interactions,
- sub-agent and multi-step task execution,
- always-on usage patterns.
Why host OpenClaw models on SaladCloud
SaladCloud changes the cost model from token billing to compute-time billing for your deployment. For many OpenClaw workloads, this has three main advantages:- Predictability: pay for runtime, not each generated token.
- Control: scale replicas up/down, including scheduled scale-to-zero windows.
- Privacy posture: keep more prompt/context flow within infrastructure you control.
Inference options on SaladCloud
You can run OpenClaw-compatible model endpoints with:- Ollama: easiest path for most users.
- vLLM: higher throughput and advanced serving optimizations.
- TGI: strong text-generation serving option for many Hugging Face models.
Recommended operating pattern
A common production pattern is:- run OpenClaw locally (or in your controlled environment),
- host one or more model endpoints on SaladCloud,
- set model fallbacks in OpenClaw,
- apply autoscaling/scheduling to align runtime with active hours.