Qwen 3.5-35B Unsloth Dynamic GGUFs Achieve SOTA Quantisation Benchmarks
1 min readUnsloth has released optimised quantisations for Qwen 3.5-35B that represent significant progress in making larger models practical for local deployment. The Unsloth Dynamic quants achieve state-of-the-art performance metrics across nearly all bit depths, with comprehensive research backing the release through over 150 KL Divergence benchmarks and a total of 9TB of GGUF variants.
What makes this release particularly valuable for practitioners is the empirical rigour—the team benchmarked extensively to demonstrate quality preservation across different quantisation levels, enabling users to make informed trade-offs between model quality and inference speed. Additionally, the release fixes a critical tool calling chat template bug that was affecting all quantisation uploaders, improving compatibility across the ecosystem.
For local LLM practitioners, this represents a solid foundation for deploying Qwen 3.5-35B at scale on consumer hardware, with the benchmark data providing transparency into quality/performance trade-offs.
Source: r/LocalLLaMA · Relevance: 9/10