Qwen Tongyi Deep Research
a 32B model that beats other SOTA deep research on many benchmarks
tongyi-agent.github.io/blog/introdu...
Qwen Tongyi Deep Research
View original thread
25
3
- fully synthetic
- continued pre-training (mid) on agent traces
i’ve been saying this! data! data! data!
- continued pre-training (mid) on agent traces
i’ve been saying this! data! data! data!
3
this is for dataset construction 🤯
1
from github: they describe how they synthesize successful single step traces, and then combine them together into successful multi-step traces
🤯
generally, long traces = higher quality
they can sythesize their way into the highest quality
github.com/Alibaba-NLP/...
🤯
generally, long traces = higher quality
they can sythesize their way into the highest quality
github.com/Alibaba-NLP/...
3
btw i fucked up the labeling earlier. this is a 30B-A3B, so yeah, extremely compact and cheap
3