Alongside its security hardening, OpenClaw v2026.4.15-beta.1 delivers four meaningful capability additions. Cloud-backed memory indexes, GitHub Copilot as an embedding provider, an OAuth health card in Control UI, and a new lean mode for local model deployments — here is what each one does.
Control UI: OAuth Token Health at a Glance
PR #66211
A new Model Auth status card in the Control UI Overview shows OAuth token health and provider rate-limit pressure at a glance. It surfaces attention callouts when OAuth tokens are expiring or have already expired. The underlying models.authStatus gateway method strips credentials before sending and caches results for 60 seconds to avoid hammering providers on every refresh. Thanks to @omarshahine.
This is a practical quality-of-life improvement for anyone running multiple model providers. Silent token expiry has historically been a silent failure mode — you'd only discover the problem when a request failed, not proactively.
Memory/LanceDB: Cloud Storage Support
PR #63502
The memory-lancedb plugin can now store durable memory indexes on remote object storage instead of local disk only. This unlocks persistent, portable memory for cloud-hosted OpenClaw deployments where writing to local disk is impractical — think serverless environments, multi-instance setups, or stateless containers.
The implementation keeps the local-disk path intact as the default. Remote storage is opt-in via configuration. Thanks to @rugvedS07.
GitHub Copilot Embedding Provider
PR #61718
OpenClaw memory search can now use GitHub Copilot as an embedding provider. A dedicated Copilot embedding host helper is exposed for plugins to reuse the transport while honoring remote overrides, token refresh, and safer payload validation.
For developers already authenticated with GitHub Copilot through their IDE, this means memory-search embeddings can run through that same credential without managing a separate OpenAI or other embedding API key. Thanks to @feiskyer and @vincentkoc.
Experimental: localModelLean Mode
PR #66495
A new experimental flag, agents.defaults.experimental.localModelLean: true, drops heavyweight default tools — including browser, cron, and message — from the agent's tool set when running on local models. This reduces prompt size significantly for setups where a weaker local model is being used and the full tool surface is unnecessary overhead.
The flag leaves the normal path completely unchanged. It is opt-in and documented as experimental. Thanks to @ImLukeF.
This is a thoughtful addition. Local models often run on constrained hardware and have smaller context windows. Trimming 10–15 tool definitions from the system prompt can meaningfully improve response quality on models like Ollama's smaller variants.
Packaging: Leaner Builds
PR #67099
Plugin runtime dependencies are now localized to their owning extensions, and the published docs payload has been trimmed. Install and package-manager guardrails are tighter, so published builds stay leaner and the core package stops carrying extension-owned runtime baggage. Thanks to @vincentkoc.
Status
All of these changes are in v2026.4.15-beta.1 — a pre-release. The stable v2026.4.15 has not yet been tagged. See the full pre-release changelog for the complete list of fixes also included in this build.