Framework Choice Critical: llama.cpp and vLLM Outperform Ollama for Qwen 3.5 Testing
1 min readA community member highlighted critical differences in how various local inference frameworks handle Qwen 3.5 models, particularly regarding chain-of-thought reasoning, tool calling, and response quality. Testing revealed that llama.cpp, transformers, vLLM, and SGLang reliably support Qwen 3.5's advanced features, while Ollama and similar wrapper frameworks showed degraded performance with broken tool calls and incomplete reasoning outputs.
This finding is crucial for practitioners because it explains discrepancies in reported results across the community and emphasizes the importance of framework selection for model evaluation. Ollama's abstraction layer, while convenient for casual users, may mask or introduce issues when running cutting-edge models with complex reasoning capabilities. Practitioners planning to deploy Qwen 3.5 for production use should prioritize lower-level frameworks that provide direct control over inference parameters and avoid potentially problematic abstraction layers.
The recommendation to use battle-tested frameworks like llama.cpp and vLLM becomes especially important when evaluating models for specific tasks like agentic RAG pipelines or tool-using applications.
Source: r/LocalLLaMA · Relevance: 9/10