Alibaba-NLP / DeepResearch
- четверг, 18 сентября 2025 г. в 00:00:04
Tongyi DeepResearch, the Leading Open-source DeepResearch Agent
We present Tongyi DeepResearch, an agentic large language model featuring 30.5 billion total parameters, with only 3.3 billion activated per token. Developed by Tongyi Lab, the model is specifically designed for long-horizon, deep information-seeking tasks. Tongyi DeepResearch demonstrates state-of-the-art performance across a range of agentic search benchmarks, including Humanity's Last Exam, BrowserComp, BrowserComp-ZH, WebWalkerQA,xbench-DeepSearch, FRAMES and SimpleQA.
Tongyi DeepResearch builds upon our previous work on the
WebAgent project.
More details can be found in our 📰 Tech Blog.
You can directly download the model by following the links below.
Model | Download Links | Model Size | Context Length |
---|---|---|---|
Tongyi-DeepResearch-30B-A3B | 🤗 HuggingFace 🤖 ModelScope |
30B-A3B | 128K |
[2025/09/17]🔥 We have released Tongyi-DeepResearch-30B-A3B.
This guide provides instructions for setting up the environment and running inference scripts located in the inference folder.
conda
or virtualenv
.# Example with Conda
conda create -n react_infer_env python=3.10.0
conda activate react_infer_env
Install the required dependencies:
pip install -r requirements.txt
eval_data/
in the project root.eval_data/example.jsonl
.{"question": "...","answer": "..."}
eval_data
folder for reference.question
field and place the referenced file inside the eval_data/file_corpus/
directory.run_react_infer.sh
and modify the following variables as instructed in the comments:
MODEL_PATH
- path to the local or remote model weights.DATASET
- path to the evaluation set, e.g. example
.OUTPUT_PATH
- path for saving the prediction results, e.g. ./outputs
.API_KEY
, BASE_URL
, or other credentials. Each key is explained inline in the bash script.bash run_react_infer.sh
With these steps, you can fully prepare the environment, configure the dataset, and run the model. For more details, consult the inline comments in each script or open an issue.
We provide benchmark evaluation scripts for various datasets. Please refer to the evaluation scripts directory for more details.
Tongyi DeepResearch also has an extensive deep research agent family. You can find more information in the following paper:
[1] WebWalker: Benchmarking LLMs in Web Traversal
[2] WebDancer: Towards Autonomous Information Seeking Agency
[3] WebSailor: Navigating Super-human Reasoning for Web Agent
[4] WebShaper: Agentically Data Synthesizing via Information-Seeking Formalization
[5] WebWatcher: Breaking New Frontier of Vision-Language Deep Research Agent
[6] WebResearcher: Unleashing unbounded reasoning capability in Long-Horizon Agents
[7] ReSum: Unlocking Long-Horizon Search Intelligence via Context Summarization
[8] WebWeaver: Structuring Web-Scale Evidence with Dynamic Outlines for Open-Ended Deep Research
[9] WebSailor-V2: Bridging the Chasm to Proprietary Agents via Synthetic Data and Scalable Reinforcement Learning
[10] Scaling Agents via Continual Pre-training
[11] Towards General Agentic Intelligence via Environment Scaling
🔥🔥🔥 We are hiring! Research intern positions are open (based in Hangzhou、Beijing、Shanghai)
📚 Research Area:Web Agent, Search Agent, Agent RL, MultiAgent RL, Agentic RAG
☎️ Contact:yongjiang.jy@alibaba-inc.com
For communications, please contact Yong Jiang (yongjiang.jy@alibaba-inc.com).
@misc{tongyidr,
author={Tongyi DeepResearch Team},
title={Tongyi-DeepResearch},
year={2025},
howpublished={\url{https://github.com/Alibaba-NLP/DeepResearch}}
}