Deep Research (DR) agents have rapidly gained popularity in both research and industry, thanks to recent progress in LLMs. However, most popular public DR agents are not designed with human thinking and writing processes in mind. They often lack structured steps that support human researchers, such as drafting, searching, and using feedback. Current DR agents compile test-time algorithms and various tools without cohesive frameworks, highlighting the critical need for purpose-built frameworks that can match or excel human research capabilities. The absence of human-inspired cognitive processes in current methods creates a gap between how humans do research and how AI agents handle complex research tasks.
Existing works, such as test-time scaling, utilize iterative refinement algorithms, debate mechanisms, tournaments for hypothesis ranking, and self-critique systems to generate research proposals. Multi-agent systems utilize planners, coordinators, researchers, and reporters to produce detailed responses, while some frameworks enable human co-pilot modes for feedback integration. Agent tuning approaches focus on training through multitask learning objectives, component-wise supervised fine-tuning, and reinforcement learning to improve search and browsing capabilities. LLM diffusion models attempt to break autoregressive sampling assumptions by generating complete noisy drafts and iteratively denoising tokens for high-quality outputs.
Researchers at Google introduced Test-Time Diffusion Deep Researcher (TTD-DR), inspired by the iterative nature of human research through repeated cycles of searching, thinking, and refining. It conceptualizes research report generation as a diffusion process, starting with a draft that serves as an updated outline and evolving foundation to guide research direction. The draft undergoes iterative refinement through a “denoising” process, dynamically informed by a retrieval mechanism that incorporates external information at each step. This draft-centric design makes report writing more timely and coherent while reducing information loss during iterative search processes. TTD-DR achieves state-of-the-art results on benchmarks that require intensive search and multi-hop reasoning.
The TTD-DR framework addresses limitations of existing DR agents that employ linear or parallelized processes. The proposed backbone DR agent contains three major stages: Research Plan Generation, Iterative Search and Synthesis, and Final Report Generation, each containing unit LLM agents, workflows, and agent states. The agent utilizes self-evolving algorithms to enhance the performance of each stage, helping it to find and preserve high-quality context. The proposed algorithm, inspired by recent self-evolution work, is implemented in a parallel workflow along with sequential and loop workflows. This algorithm can be applied to all three stages of agents to improve overall output quality.
In side-by-side comparisons with OpenAI Deep Research, TTD-DR achieves 69.1% and 74.5% win rates for long-form research report generation tasks, while outperforming by 4.8%, 7.7%, and 1.7% on three research datasets with short-form ground-truth answers. It shows strong performance in Helpfulness and Comprehensiveness auto-rater scores, especially on LongForm Research datasets. Moreover, the self-evolution algorithm achieves 60.9% and 59.8% win rates against OpenAI Deep Research on LongForm Research and DeepConsult. The correctness score shows an enhancement of 1.5% and 2.8% on HLE datasets, though the performance on GAIA remains 4.4% below OpenAI DR. The incorporation of Diffusion with Retrieval leads to substantial gains over OpenAI Deep Research across all benchmarks.
In conclusion, Google presents TTD-DR, a method that addresses fundamental limitations through human-inspired cognitive design. The framework’s approach conceptualizes research report generation as a diffusion process, utilizing an updatable draft skeleton that guides research direction. TTD-DR, enhanced by self-evolutionary algorithms applied to each workflow component, ensures high-quality context generation throughout the research process. Moreover, evaluations demonstrate that TTD-DR’s state-of-the-art performance across various benchmarks that require intensive search and multi-hop reasoning, with superior results in both comprehensive long-form research reports and concise multi-hop reasoning tasks.
Check out the Paper here. Feel free to check our Tutorials page on AI Agent and Agentic AI for various applications. Also, feel free to follow us on Twitter and don’t forget to join our 100k+ ML SubReddit and Subscribe to our Newsletter.
Sajjad Ansari is a final year undergraduate from IIT Kharagpur. As a Tech enthusiast, he delves into the practical applications of AI with a focus on understanding the impact of AI technologies and their real-world implications. He aims to articulate complex AI concepts in a clear and accessible manner.
Leave a comment