Turn your Meta Ray-Ban smart glasses into an AI-powered assistant. Uses Gemini for real-time voice + vision — ask what you're looking at, add items to lists, search the web, control smart home devices. Streams ~1fps from glasses camera with bidirectional audio. Based on the viral VisionClaw project.
Features
Bidirectional audio with Gemini Live API. Camera streams ~1fps from your glasses — ask what you're looking at and get instant spoken answers.
Route voice commands to WhatsApp, Telegram, iMessage, shopping lists, smart home, web search, reminders, and more through OpenClaw.
Primary mode uses Meta Ray-Ban smart glasses. Testing mode falls back to your iPhone camera — no special hardware required to try it.
Say 'add milk to my shopping list' or 'text Sarah I'm on my way' — the agent handles the rest through connected integrations.
How it works
Camera captures frames from your Meta Ray-Bans (or iPhone) and compresses to JPEG at ~1fps
Audio streams bidirectionally — your voice goes to Gemini, responses play back through your glasses speaker
Gemini Live API processes both video and audio in real-time via WebSocket
Tool calls route through OpenClaw gateway for task execution (messaging, search, smart home, etc.)
Use cases
Integrations
Configure your agent, connect your integrations, and deploy in under 60 seconds. No infrastructure to manage.
Plans start at $49/mo · Cancel anytime