Product Thumbnail

QwQ-32B

Matching R1 reasoning yet 20x smaller

Open Source
Artificial Intelligence

QwQ-32B, from Alibaba Qwen team, is a new open-source 32B LLM achieving DeepSeek-R1 level reasoning via scaled Reinforcement Learning. Features a "thinking mode" for complex tasks.

Top comment

Hi everyone!

Check out QwQ-32B, a new open-source language model from the Qwen team. It's achieving something remarkable: reasoning performance comparable to DeepSeek-R1, but with a model that's 20 times smaller (32B parameters vs. 671B)!

This is a big deal because:

🤯 Size/Performance Ratio: It punches way above its weight class in reasoning, math, and coding tasks.
🧠 Scaled Reinforcement Learning: They achieved this by scaling up Reinforcement Learning (RL) on a strong foundation model (Qwen2.5-32B).
🤔 "Thinking Mode": Like some other recent models, it has a special "thinking mode" (activated with tags) that allows for longer chains of thought.
✅ Open Source: Model weights are available under Apache 2.0.

Available now in Qwen Chat and HF Spaces.

The implications of getting this level of reasoning performance from a 32B model are huge. It opens up possibilities for deploying powerful AI on less powerful hardware, reducing costs, and making advanced reasoning capabilities more accessible.

Comment highlights

This score and size are amazing, I have a few questions

  1. What is the difference between it and DeepSeek R1?

  2. Do you have plans to make a larger size reasoning model?

Thank you for your contribution to the open source community!

I use hugging face every day for my job. I didn't know that models there can be launched on PH. Very cool!