Parlor
Full voice + vision AI running locally on your Mac — no cloud needed
Parlor is an on-device real-time multimodal AI application that runs an end-to-end audio+video understanding and voice response loop entirely on local hardware — no API keys, no servers, no data leaving the machine. The creator built it to power a free English-learning platform without incurring ongoing server costs. It captures microphone and camera input, sends them through Gemma 4 E2B via LiteRT-LM on the GPU for comprehension, and returns synthesized speech via Kokoro TTS — all with an end-to-end latency of 2.5 to 3 seconds on an Apple M3 Pro. The stack is deliberately lean: browser-based voice activity detection (VAD), streaming audio output to minimize perceived latency, mid-response interruption support, and a total model download of roughly 2.6 GB. It's written in Python and requires no special setup beyond downloading the models. Apache 2.0 licensed. Parlor surfaced on Hacker News with over 280 points — an unusually strong signal for a one-developer demo project. The reaction reflects a broader shift: multimodal voice AI that required server-grade hardware six months ago now runs on consumer MacBooks, and open-source developers are starting to ship production-ready applications built entirely on that foundation.
Panel Reviews
The Builder
Developer Perspective
“2.5–3 second end-to-end latency for full voice + vision on a MacBook is genuinely remarkable. The architecture is clean — VAD in the browser, LiteRT-LM on GPU for the heavy lifting, Kokoro for TTS. This is a solid foundation for building privacy-first voice assistants, tutors, or accessibility tools without any ongoing API costs.”
The Skeptic
Reality Check
“Three-second latency is still noticeably clunky for natural conversation — OpenAI and Google's voice APIs run in under a second. On older Macs or non-Apple hardware the latency will be worse. It's a proof of concept, not a daily driver, and the model quality gap between Gemma 4 E2B and GPT-4o voice is real.”
The Futurist
Big Picture
“The trajectory here is the story. If M3 Pro hits 3 seconds today, M5 will hit under 1 second in 18 months. Every capability improvement in edge chips directly translates to closed-loop multimodal AI as a baseline feature of devices. Parlor is one of the first working demos of where all consumer devices are headed.”
The Creator
Content & Design
“For language tutoring, creative storytelling tools, or interactive audio-visual demos, having no cloud dependency means total privacy for learners and zero recurring costs for creators. The English-learning use case the creator shipped it for is exactly the kind of high-impact low-resource application this technology should be enabling.”
Community Sentiment
“Latency vs cloud voice AI — where on-device stands today”
“Apple Silicon as the practical platform for local multimodal”
“No API costs for education apps”