Bonsai-8B
1-bit quantized 8B LLM — 1.15GB, runs on-device at 368 tok/s
Bonsai-8B is a 1-bit quantized language model from Prism ML, based on Qwen3-8B, that compresses a full 8B parameter model down to just 1.15 gigabytes. Running at 368 tokens per second on an RTX 4090, it achieves a 6.2x throughput speedup over FP16 equivalents while scoring 70.5 average across standard benchmarks — maintaining competitive quality despite the extreme compression. The model uses end-to-end 1-bit quantization rather than post-training quantization applied to a pretrained FP16 model. This means all weights are trained natively as ternary values {-1, 0, +1}, enabling the 14x size reduction versus FP16 without the quality cliff typical of aggressive post-training quants. Bonsai-8B targets the edge and on-device inference market: robotics, mobile apps, offline-capable applications, and scenarios where privacy and latency requirements make cloud inference impractical. The 1.15GB size fits in phone RAM and runs on consumer CPUs. Apache 2.0 license means it's deployable anywhere.
Panel Reviews
The Builder
Developer Perspective
“1.15GB for an 8B model that runs at 368 tok/s is genuinely remarkable. Fitting LLM intelligence into a package that runs on a phone CPU opens use cases that were completely impractical months ago. For offline apps, robotics, or privacy-sensitive deployments, this changes the calculus entirely.”
The Skeptic
Reality Check
“70.5 average benchmark score sounds reasonable until you remember that 1-bit quantization makes the model brittle on tasks requiring numerical precision, long-context reasoning, and nuanced instruction following. The gap between 'competitive on benchmarks' and 'usable for complex tasks' is still significant for ultra-compressed models.”
The Futurist
Big Picture
“1-bit LLMs running on-device are the foundation for truly private, always-available AI. When an 8B model fits in 1GB and runs on a phone, every app becomes AI-capable without cloud dependencies. Bonsai-8B is a milestone in the long march toward AI that runs everywhere.”
The Creator
Content & Design
“For most creative workflows, you need quality over tiny model size — image-gen and writing assistance benefits from more capable models. Bonsai-8B is impressive engineering, but for production creative tools the quality trade-off of aggressive quantization is still real. Great for quick drafts, not polished work.”
Community Sentiment
“1.15GB footprint and on-device deployment”
“368 tok/s throughput on RTX 4090”
“End-to-end 1-bit training vs post-training quant”