π€ WebShaperQA ο½
WebShaperQA ο½
π€ WebSailor-3B ο½
ModelScope WebSailor-3B |
π€ WebDancer-QwQ-32B |
ModelScope WebDancer-QwQ-32B |
π€ WebWalkerQA
You can check the paper of WebDancer and WebWalker and WebSailor and WebShaper.
π₯ π₯ π₯ Stay tuned for more updates! We are working on building native agentic model based on the Browser and more open-domain environments!
- WebShaper (Preprint 2025) - WebShaper: Agentically Data Synthesizing via Information-Seeking Formalization
- WebSailor (Preprint 2025) - WebSailor: Navigating Super-human Reasoning for Web Agent
- WebDancer (Preprint 2025) - WebDancer: Towards Autonomous Information Seeking Agency
- WebWalker (ACL 2025) - WebWalker: Benchmarking LLMs in Web Traversal
2025.07.22
π₯π₯π₯We release WebShaper: Agentically Data Synthesizing via Information-Seeking Formalization.2025.07.11
π₯π₯π₯WebSailor-3B is released. You can deploy it with one click usingAlibaba Cloud's FunctionAI in ten minutes!
2025.07.03
π₯π₯π₯We release WebSailor, an agentic search model specialized in performing extremely complex information seeking tasks, achieving open-source SOTA on some of the most difficult browsing benchmarks. WebSailor topped the HuggingFace daily papers.2025.06.23
π₯π₯π₯The model, interactive demo, and some of the data of WebDancer have been open-sourced. You're welcome to try them out!2025.05.29
π₯π₯π₯We release WebDancer, a native agentic search model towards autonomous information seeking agency and Deep Research-like model.2025.05.15
WebWalker is accepted by ACL 2025 main conference.2025.01.14
We release WebWalker, a benchmark for LLMs in web traversal and a multi-agent framework for information seeking.
- A
formalization-driven
data synthesis method for information-seeking agents, grounded in our proposed task formalization. Leveraging this method, we construct the WebShaper dataset, which enables systematic generation of IS instances. - We propose an agentic Expander that iteratively generates and validates questions in alignment with the formalization.
- We conduct extensive experiments across multiple benchmarks to evaluate the effectiveness of WebShaper. We achieve new state-of-the-art results on GAIA (60.19) and WebWalkerQA (52.50) benchmarks.
- A complete post-training methodology enabling models to engage in extended thinking and information seeking, ultimately allowing them to successfully complete extremely complex tasks previously considered unsolvable.
- Introduces SailorFog-QA, a scalable QA benchmark with high uncertainty and difficulty, curated with a novel data synthesis method through graph sampling and information obfuscation. Example SailorFog-QA data samples can be found at:
WebSailor/dataset/sailorfog-QA.jsonl
- Effective post-training pipeline consisting of (1) high-quality reconstruction of concise reasoning from expert trajectories for clean supervision, (2) a two-stage training process involving an RFT cold start stage, followed by Duplicating Sampling Policy Optimization (DUPO), an efficient agentic RL algorithm excelling in effectiveness and efficiency.
- WebSailor-72B significantly outperforms all open-source agents and frameworks while closing the performance gap with leading proprietary systems, achieving a score of 12.0% on BrowseComp-en, 30.1% on BrowseComp-zh, and 55.4% on GAIA.
- The checkpoint is coming soon.
- Native agentic search reasoning model using ReAct framework towards autonomous information seeking agency and Deep Research-like model.
- We introduce a four-stage training paradigm comprising browsing data construction, trajectory sampling, supervised fine-tuning for effective cold start, and reinforcement learning for improved generalization, enabling the agent to autonomously acquire autonomous search and reasoning skills.
- Our data-centric approach integrates trajectory-level supervision fine-tuning and reinforcement learning (DAPO) to develop a scalable pipeline for training agentic systems via SFT or RL.
- WebDancer achieves a Pass@3 score of 64.1% on GAIA and 62.0% on WebWalkerQA.
You need to enter the WebDancer
folder for the following commands.
conda create -n webdancer python=3.12
pip install -r requirements.txt
Download the WebDancer model from π€ HuggingFace and deploy it using the provided scripts with sglang.
cd scripts
bash deploy_model.sh WebDancer_PATH
Note: Replace
WebDancer_PATH
with the actual path to the downloaded model.
Edit the following keys in WebDancer/scripts/run_demo.sh
:
GOOGLE_SEARCH_KEY
, you can get it from serper.JINA_API_KEY
, you can get it from jina.DASHSCOPE_API_KEY
, you can get it from dashscope.
Then, launch the demo with Gradio to interact with the WebDancer model:
cd scripts
bash run_demo.sh
We provide demos for BrowseComp-en, BrowseComp-zh and Daily Use. Our model can complete highly difficult and uncertain tasks requiring massive information acquisition and complex reasoning.
We provide demos for WebWalkerQA, GAIA and Daily Use. Our model can execute the long-horizon tasks with multiple steps and complex reasoning, such as web traversal, information seeking and question answering.
The content of this project itself is licensed under LICENSE.
If this work is helpful, please kindly cite as:
@misc{tao2025webshaper,
title={WebShaper: Agentically Data Synthesizing via Information-Seeking Formalization},
author={Zhengwei Tao and Jialong Wu and Wenbiao Yin and Junkai Zhang and Baixuan Li and Haiyang Shen and Kuan Li and Liwen Zhang and Xinyu Wang and Yong Jiang and Pengjun Xie and Fei Huang and Jingren Zhou},
year={2025},
eprint={2507.15061},
archivePrefix={arXiv},
primaryClass={cs.CL},
url={https://arxiv.org/abs/2507.15061},
}
@misc{li2025websailor,
title={WebSailor: Navigating Super-human Reasoning for Web Agent},
author={Kuan Li and Zhongwang Zhang and Huifeng Yin and Liwen Zhang and Litu Ou and Jialong Wu and Wenbiao Yin and Baixuan Li and Zhengwei Tao and Xinyu Wang and Weizhou Shen and Junkai Zhang and Dingchu Zhang and Xixi Wu and Yong Jiang and Ming Yan and Pengjun Xie and Fei Huang and Jingren Zhou},
year={2025},
eprint={2507.02592},
archivePrefix={arXiv},
primaryClass={cs.CL},
url={https://arxiv.org/abs/2507.02592},
}
@misc{wu2025webdancer,
title={WebDancer: Towards Autonomous Information Seeking Agency},
author={Jialong Wu and Baixuan Li and Runnan Fang and Wenbiao Yin and Liwen Zhang and Zhengwei Tao and Dingchu Zhang and Zekun Xi and Yong Jiang and Pengjun Xie and Fei Huang and Jingren Zhou},
year={2025},
eprint={2505.22648},
archivePrefix={arXiv},
primaryClass={cs.CL},
url={https://arxiv.org/abs/2505.22648},
}
@misc{wu2025webwalker,
title={WebWalker: Benchmarking LLMs in Web Traversal},
author={Jialong Wu and Wenbiao Yin and Yong Jiang and Zhenglin Wang and Zekun Xi and Runnan Fang and Deyu Zhou and Pengjun Xie and Fei Huang},
year={2025},
eprint={2501.07572},
archivePrefix={arXiv},
primaryClass={cs.CL},
url={https://arxiv.org/abs/2501.07572},
}
π₯π₯π₯ We are hiring! Research intern positions are open (based in HangzhouγBeijingγShanghai)
π Research AreaοΌWeb Agent, Search Agent, Agent RL, MultiAgent RL, Agentic RAG
βοΈ ContactοΌ[email protected]
For communications, please contact Yong Jiang ([email protected]).