Qwen3.5 and gpt-oss are good for local inference with tool use capabilities.
this post was submitted on 21 Feb 2026
3 points (100.0% liked)
Ollama - Local LLMs for everyone!
340 readers
1 users here now
A place to discuss Ollama, from basic use, extensions and addons, integrations, and using it in custom code to create agents.
founded 10 months ago
MODERATORS
I tried Qwen3.5 9b & 4b versions for the last couple of days and my experience has been really bad. Qwen3.5 regardless the size are slow and seems like overthinking too much. I'm getting better results with ministral-3.
I’ve had decent results with Qwen3.5-122B-A10B-AWQ-4bit. But I’m planning to try the latest unsloth version to see how it performs. There are times when I think I’m talking to a child especially when I come off of using Opus.