
Building tiny frontier AI models that run on edge devices
Report from 24 days ago
Stellon Labs is a two-person AI research group building tiny, open-source neural models that run on CPUs on phones, Raspberry Pis, and in browsers. Today they ship KittenTTS, a ~15M‑parameter text‑to‑speech model with an ONNX weight file under 25 MB, published as code, weights, and a pip/ONNX package designed to run without a GPU GitHub Hugging Face.
Developers are already running KittenTTS in browser demos and small self‑hosted servers on low‑power devices, and the launch attracted community traction on Hacker News and Hugging Face web demo server example HN thread HF space.
Top-down context:
Analysts estimate the global TTS market at roughly USD 3.5–4.0B with ~12–14% CAGR, while on‑device/edge AI markets are in the multi‑tens of billions and growing quickly MarketsandMarkets Mordor Intelligence Grand View Research — On‑Device AI Grand View Research — Edge AI. Stellon targets the on‑device TTS slice for apps/devices where cloud is impractical.
Bottom-up calculation:
Using a USD ~3.5–4.0B TTS base, if 5–10% of usage is on‑device near‑term, TAM ≈ USD 175–400M; if 15–25% shifts on‑device in 2–4 years, TAM ≈ USD 525M–1.0B MarketsandMarkets. Expansion into adjacent on‑device speech/multimodal tooling taps parts of the broader edge AI market (tens of billions) Grand View Research.
Assumptions: