Qwen3.5 Small: 0.8B, 2B, 4B, 9B Released

(huggingface.co)

6 points | by syx 5 hours ago ago

3 comments

  • throwaway2027 4 hours ago ago

    So 27B at Q3 or 9B at Q8?

  • powera 5 hours ago ago

    This looks like somebody re-releasing QWEN models to promote their own company. https://news.ycombinator.com/item?id=47217305 is the link to QWEN's repo.

    • cpburns2009 4 hours ago ago

      If you want to have a chance at running a large model, it needs to be quantized. The unsloth user on Huggingface manages popular quantizations for many models, Qwen included, and I think he developed dynamic GGUF quantization.

      Take Qwen/Qwen3.5-35B-A3B for example. It's 72 GB. While unsloth/Qwen3.5-35B-A3B-GGUF has quantizations from 9-38 GB.