Run LLMs on-device with Onde Inference, with first-class support for Apple silicon.
Swift SDK · Kotlin Multiplatform SDK · Flutter SDK · React Native SDK · Website
Onde is already shipping in real App Store apps. Chat runs fully on-device, so there is no server round trip and no user data leaving the device. For SDK docs, platform notes, and setup details, see https://ondeinference.com/sdk. If you want to test downloads, model selection, or GGUF export before wiring the engine into app code, use Onde CLI.
Onde is dual-licensed under MIT and Apache 2.0. You may use it under either license at your option.
| Dependency | License | Author |
|---|---|---|
| mistral.rs | MIT | Eric Buehler |
| UniFFI | MPL-2.0 | Mozilla |
| tokio | MIT | Tokio contributors |
Models downloaded by Onde have their own licenses independent of this crate. By using Onde, you are also subject to the license of the model you load:
| Model | Size | License | Commercial use |
|---|---|---|---|
| Qwen 2.5 1.5B Instruct (GGUF Q4_K_M) | ~941 MB | Qwen Community License | ✅ with conditions¹ |
| Qwen 2.5 3B Instruct (GGUF Q4_K_M) | ~1.93 GB | Qwen Community License | ✅ with conditions¹ |
| Qwen 2.5 Coder 7B Instruct (GGUF Q4_K_M) | ~4.4 GB | Qwen Community License | ✅ with conditions¹ |
| Qwen 3 1.7B (GGUF Q4_K_M) | ~1.3 GB | Apache 2.0 | ✅ |
| Qwen 3 4B (GGUF Q4_K_M) | ~2.7 GB | Apache 2.0 | ✅ |
| Qwen 3 8B (GGUF Q4_K_M) | ~5 GB | Apache 2.0 | ✅ |
| Qwen 3 14B (GGUF Q4_K_M) | ~8.4 GB | Apache 2.0 | ✅ |
| DeepSeek Coder 6.7B Instruct (GGUF Q4_K_M) | ~3.8 GB | DeepSeek License v1.0 | ✅ with conditions² |
¹ Qwen Community License conditions: no training of competing models, attribution required, no misrepresentation of origin. Organisations with more than 100 million monthly active users must obtain a separate commercial licence from Alibaba Cloud.
² DeepSeek License v1.0 conditions: use-based restrictions apply (see Attachment A of the license). Prohibits military use, generation of disinformation, and certain other uses. Governing law is PRC law.
Onde's own license (MIT OR Apache-2.0) is independent of these model licenses. If you build an application on top of Onde, you are responsible for complying with the license of whichever model your users load.
© 2026 Onde Inference (Splitfire AB).