Machine Learning Systems
-
Updated
Apr 20, 2026 - JavaScript
Machine Learning Systems
Production Android AI with ExecuTorch 1.0 - Deploy PyTorch models to mobile with NPU acceleration and 50KB footprint
LLM inference on mobile via Capacitor — run quantized GGUF models on-device
Qualcomm® AI Hub Models is our collection of state-of-the-art machine learning models optimized for performance (latency, memory etc.) and ready to deploy on Qualcomm® devices.
📱 Optimized ML for edge devices. Showcasing efficient model deployment, GPU-CPU memory transfer optimization, and real-world edge AI applications. 🤖
Standalone ONNX runtime session management and preprocessing for Dust — iOS/macOS
INT8 quantization of MobileNetV2 for learning and production-oriented iOS mobile inference.
Model download and serving orchestration for Dust — Capacitor bridge
On-device text embedding generation for iOS and Android via Capacitor
Android ONNX runtime session management and preprocessing for Dust
Claude Code skill for Google LiteRT - on-device AI/ML deployment framework
Android ML model server — download management, session caching, accelerator probing
Standalone tokenizers and embedding runtime primitives for on-device text embeddings
magnitude-based pruning of MobileNetV2 for learning and production-oriented iOS mobile inference.
Standalone model server business logic for iOS — download, caching, accelerator probing
Add a description, image, and links to the mobile-ml topic page so that developers can more easily learn about it.
To associate your repository with the mobile-ml topic, visit your repo's landing page and select "manage topics."