Yooz Labs
Sovereign Intelligence. Built for the skeptical.
Privacy-first AI that runs entirely on your devices. No cloud, no tracking, no compromises.
What we're building
We're the privacy infrastructure for the AI decade. Every Yooz product is designed for the 70% of consumers who don't trust cloud AI but lack consumer-grade alternatives.
- Yooz Engine β unified local AI service for macOS (STT, LLM, grammar, VAD, TTS).
- Yooz Whisper β voice keyboard for macOS.
- Yooz Notes β note-taking with private AI memory.
- Remi β Claude Code's distant friend. Secure peer-to-peer remote sessions for Claude Code (and soon Codex), with an iPhone app and local auto-approve.
- Yooz Vault β privacy hardware (home server).
- Universal AI Platform Layer β one API across Apple Core ML, Android ML Kit, Windows DirectML.
What lives on this Hugging Face org
The model weights β Apache 2.0, fully open source. The Yooz product code is source-available on GitHub under PolyForm Shield, but the weights stay open so the research community can build on them, audit them, and remix them.
Model categories
| Category | What it is |
|---|---|
| ASR | Speech-to-text checkpoints (Qwen3-ASR Swift port, Parakeet derivatives) |
| LLMs (Touchup) | Fine-tuned small LLMs that fix/clean speech-to-text transcripts |
| Distillations | Small students distilled from larger teachers for on-device inference |
| Adapters | LoRA / DoRA adapters published alongside their fused checkpoints |
All checkpoints document their lineage (base model + Hugging Face link), eval numbers (real benchmarks, not vibes), and Swift / Python usage snippets.
Why open weights?
The competitive moat in privacy-first AI lives in the product, not the weights:
- Multi-device orchestration (phone β PC β Vault) over WireGuard mesh.
- Universal platform abstraction across Apple, Android, Windows AI APIs.
- Private AI memory: encrypted, local, with permissioned cross-app context.
- Beautiful, consumer-grade UX.
The weights themselves should be open so the research community can audit privacy claims, reproduce evals, and build on top. Releases follow the standard "ship the artifact + the recipe to reproduce it."
Provenance
We never train on user data without opt-in. All training corpora and synthetic data sources are documented in the model card for each checkpoint. Where we fine-tune from a base model (Qwen, Gemma, etc.), the lineage is preserved and the upstream license is respected.
Source-available, not closed
The product code lives at github.com/yooz-labs under PolyForm Shield. You can read it, fork it, and build on it for non-competing use cases. We chose this path because we want to stay community-aligned without enabling AWS-style "managed Yooz" competing services.
Get in touch
- Engineering & research: dev@yooz.info
- Bugs and feature requests: file on the relevant GitHub repo under yooz-labs
- Mailing list / news: coming soon
We're building the privacy infrastructure for the AI decade. Every decision prioritizes user sovereignty, data privacy, and beautiful simplicity.