OpenClaw v2026.4.2, released April 2, 2026, included one feature that stands out from the usual infrastructure and plugin work: native Google Assistant integration on Android. Users can now invoke their OpenClaw agent directly from the Android assistant trigger and hand prompts straight into the chat composer — without unlocking their phone and opening an app.
What Changed in PR #59596
The work, contributed by @obviyus, added two distinct pieces to the Android companion app:
- Assistant-role entrypoints — These register OpenClaw as an assistant-compatible app at the Android OS level, so the system knows it can receive voice or text prompts from the assistant surface.
- Google Assistant App Actions metadata — This is the declarative layer that tells Google Assistant what OpenClaw can do, what intents it handles, and how to route prompts into the app's chat composer.
Together, these changes mean Android users can say something like "Hey Google, ask OpenClaw to check my emails" or invoke it from a Pixel shortcut, and the prompt lands directly in the OpenClaw chat session.
Why This Is a Meaningful Shift
Until now, OpenClaw on mobile was primarily a messaging-app experience. You'd open Telegram, WhatsApp, or Discord, and talk to your agent there. The agent was capable, but the interaction required deliberate context-switching — you had to go to a specific app.
Google Assistant integration changes that interaction model. OpenClaw becomes part of the ambient computing layer on Android. It can be invoked from the lock screen, from a driving mode shortcut, or from a custom routine. For users who rely on OpenClaw for calendar management, reminders, or quick web lookups, this is a qualitative improvement in how quickly they can reach their agent.
It also positions OpenClaw alongside established smart assistants rather than as a separate niche tool.
Fits a Broader Pattern
This isn't happening in isolation. The macOS Voice Wake feature (PR #58490, also in v2026.4.2) brought a similar ambient-invocation pattern to Apple platforms — users can now trigger Talk Mode on macOS without clicking anything. The Android and macOS changes are clearly part of a coordinated push to make OpenClaw feel native to each platform's input model.
On Android specifically, deeper integration is possible. With App Actions metadata in place, future contributors could add deeper Shortcuts integration, Bixby support on Samsung devices, or voice-specific system prompts that adapt when the agent is invoked hands-free.
How to Enable It
The feature is part of the Android companion app update that ships alongside v2026.4.2. Once you've updated the companion app, the Google Assistant App Actions metadata registers automatically. No manual configuration is required beyond having a connected gateway.
If your gateway is v2026.4.2 or later and your Android app is updated, you should be able to invoke OpenClaw through the Google Assistant trigger immediately.
Full release notes are available on GitHub.