TinyML & Edge AI: On-device inference, model quantization, embedded ML, ultra-low-power AI for microcontrollers and IoT devices.
-
Updated
Nov 10, 2025 - Python
TinyML & Edge AI: On-device inference, model quantization, embedded ML, ultra-low-power AI for microcontrollers and IoT devices.
iOS + Android app that runs local LLMs on-device + routstr cloud LLMs for anonymous inference
Mobile AI: iOS CoreML, Android TFLite, on-device inference, ONNX, TensorRT, and ML deployment for smartphones.
Production Android AI with ExecuTorch 1.0 - Deploy PyTorch models to mobile with NPU acceleration and 50KB footprint
Real-time SAM2 segmentation on edge devices - 40x faster C++ inference with ONNX Runtime for iOS/Android deployment
Run small LLMs directly on your device, no cloud computing needed.
A free, 100% on-device ChatGPT Application. Your data never leaves your hardware. Private. Local. Limitless.
Add a description, image, and links to the on-device-inference topic page so that developers can more easily learn about it.
To associate your repository with the on-device-inference topic, visit your repo's landing page and select "manage topics."