feat(local-apps): Add Nexa SDK integration
Introduction
Hello! Friends from Huggingface π
Thank you for maintaining this amazing library and ecosystem.
Following a recent discussion with Hugging Face (see here), your CEO expressed interest in exploring collaboration to bring on-device model inference to the community. This PR proposes adding Nexa SDK to the local apps section as a new runtime integration option.
About Nexa SDK
Nexa SDK is an on-device inference framework that runs any model on any device, across any backend. It runs on CPUs, GPUs, NPUs with backend support for CUDA, Metal, Vulkan, and Qualcomm / Intel / AMD NPU. It handles multiple input modalities including text π, image πΌοΈ, and audio π§. The SDK includes an OpenAI-compatible API server with support for JSON schema-based function calling and streaming. It supports model formats such as GGUF, MLX, Nexa AI's own .nexa format, enabling efficient quantized inference across diverse platforms.
Example Demos
πΌοΈ Multi-Image Reasoning |
π€ Image + Text β Function Call |
πΆ Multi-Audio Comparison |
|
Related Links
- π€ Nexa AI on Hugging Face
- π» Nexa SDK GitHub Repository
- π Official Documentation
- π§© Latest Releases
Logo File
Thank you for your time reviewing this PR!
Weβre excited to explore how Nexa SDK can extend Hugging Face models to mobile and edge platforms. π
@SBrandeis @Wauplin @julien-c @pcuenca look forward to your review feedbacks!