1. 📘 Topic and Domain: The paper presents Tongyi DeepResearch, an open-source large language model designed specifically for autonomous deep information-seeking research tasks.
2. 💡 Previous Research and New Ideas: Based on previous work in LLMs and agent systems, it introduces a novel end-to-end agentic training framework combining mid-training and post-training phases, along with automated data synthesis and customized environments.
3. ❓ Problem: The paper aims to develop an efficient, open-source AI research agent capable of conducting complex, multi-step reasoning and information seeking tasks that would typically take humans several hours.
4. 🛠️ Methods: The authors used a combination of agentic mid-training, post-training, automated data synthesis pipeline, and stage-specific environments, built on a 30.5B parameter model that activates only 3.3B parameters per token.
5. 📊 Results and Evaluation: The model achieved state-of-the-art performance across multiple benchmarks, including 32.9 on Humanity's Last Exam, 43.4 on BrowseComp, 72.2 on WebWalkerQA, and others, outperforming both open-source and proprietary systems.