-
Updated
May 16, 2026 - TypeScript
on-device-inference
Here are 39 public repositories matching this topic...
TinyML & Edge AI: On-device inference, model quantization, embedded ML, ultra-low-power AI for microcontrollers and IoT devices.
-
Updated
Nov 10, 2025 - Python
AI for Apple silicon devices.
-
Updated
May 15, 2026 - Rust
Auditable offline edge intelligence for low-cost edge devices, with benchmark evidence and public board proof on ESP32-C3.
-
Updated
Mar 23, 2026 - Python
Flutter starter example app to get started with NobodyWho, a library designed to run LLMs locally and efficiently on any device.
-
Updated
May 12, 2026 - Dart
The Private Agent OS — search files, run AI agents, connect to 10,000+ tools via the complete protocol stack (MCP, AG-UI, A2UI, A2A). Zero cloud. Zero telemetry. On-device inference.
-
Updated
May 14, 2026 - Rust
React Native starter example app to get started with NobodyWho, a library designed to run LLMs locally and efficiently on any device.
-
Updated
May 12, 2026 - TypeScript
Custom llama.cpp fork with character intelligence engine: control vectors, attention bias, head rescaling, attention temperature, fast weight memory
-
Updated
May 16, 2026 - C++
iOS + Android app that runs local LLMs on-device + routstr cloud LLMs for anonymous inference
-
Updated
Sep 18, 2025 - TypeScript
Mobile AI: iOS CoreML, Android TFLite, on-device inference, ONNX, TensorRT, and ML deployment for smartphones.
-
Updated
Nov 10, 2025 - Python
High-performance Android SDK for on-device LLM inference (GGUF). Privacy-focused, offline-first, and powered by llama.cpp with a clean Kotlin Coroutines API.
-
Updated
Mar 27, 2026 - Kotlin
Production Android AI with ExecuTorch 1.0 - Deploy PyTorch models to mobile with NPU acceleration and 50KB footprint
-
Updated
Nov 14, 2025 - Python
Unofficial Swift SDK for Google's LiteRT-LM — run Gemma 4 on-device with text, vision, audio, and tool calling. CPU + GPU (Metal). iOS 17+ / macOS 14+.
-
Updated
May 2, 2026 - Swift
-
Updated
May 14, 2026 - JavaScript
On-device AI inference for Swift apps. Run LLMs locally on iOS, macOS, visionOS and watchOS.
-
Updated
May 15, 2026 - Swift
Neural acoustic echo cancellation for Apple platforms using CoreML — Swift package with 128/256/512-unit DTLN-aec models
-
Updated
May 1, 2026 - Swift
iOS & macOS starter example apps to get started with NobodyWho, a library designed to run LLMs locally and efficiently on any device.
-
Updated
May 13, 2026 - Swift
Swift wrapper for Apple's BNNS graph API — run compiled CoreML models (.mlmodelc) on CPU with zero-copy buffer management
-
Updated
Mar 9, 2026 - Swift
React Native SDK for local LLM inference and on-device AI on iOS and Android.
-
Updated
Mar 14, 2026 - TypeScript
Ad generation via offline LLMs with on-device inference, optionally managed by a self-hosted CMS.
-
Updated
May 13, 2026
Improve this page
Add a description, image, and links to the on-device-inference topic page so that developers can more easily learn about it.
Add this topic to your repo
To associate your repository with the on-device-inference topic, visit your repo's landing page and select "manage topics."