This is an automated archive made by the Lemmit Bot.
The original was posted on /r/surface by /u/Different-Effect-724 on 2025-11-03 19:43:14+00:00.
Surface X Elite users — your NPU can now run real local AI models with NexaSDK
Surface users with the new Snapdragon X Elite chip — you’ve actually got a full AI engine sitting idle in there.
NexaSDK just made it usable. It’s a local inference SDK that runs models like:
- Qwen3 (LLM)
- Qwen3-VL (vision-language)
- Parakeet (speech)
- OmniNeural (multimodal agent)
All on the Qualcomm NPU — 9× more energy efficient and 2× faster than CPU/GPU.
Install is dead simple:
nexa config set license '<access_token>'
- Run the model:
nexa infer NexaAI/OmniNeural-4B
Video demo shows the model reading a poster and scheduling an event locally — no cloud involved.
https://reddit.com/link/1onlrw5/video/6p290iu3i3zf1/player
Repo: github.com/NexaAI/nexa-sdk
Would love to see what Surface owners build with this — anyone else trying to tap the NPU for local AI?