DREAM: Deep Research Evaluation with Agentic Metrics
Key Summary
- ā¢Deep research agents write long reports, but old tests often judge only how smooth they sound and whether they add links, not whether the facts are true today or the logic really holds.
- ā¢The paper spots the Mirage of Synthesis: reports can look great on the surface while hiding factual mistakes, outdated info, or shaky reasoning.
- ā¢It introduces a four-part map of quality: Presentation Quality, Task Compliance, Analytical Depth, and Source Quality.
- ā¢The big gap is capability mismatch: static judges canāt browse, check dates, or verify facts like research agents can.
- ā¢DREAM fixes this by making the judge an agent too, using tools to search, verify, and think step by step (capability parity).
- ā¢DREAM builds a custom checklist for each question (adaptive metrics) and also uses universal checks (static metrics).
- ā¢Key-Information Coverage (KIC) and Reasoning Quality (RQ) are agent-built tests that catch missing updates and flawed logic.
- ā¢DREAMās factuality test checks truth against the live web, not just against the reportās citations.
- ā¢Experiments show DREAM is much more sensitive to time decay and hidden reasoning or factual errors than popular benchmarks.
- ā¢This gives a scalable, reference-free way to judge deep research that better matches real-world needs.
Why This Research Matters
Real decisionsāabout health, finance, policy, and educationādepend on research that is current, correct, and logically sound. If evaluations reward only smooth writing and matching citations, people can be misled by confident but outdated or false reports. DREAMās agentic judging brings web tools and verification into the evaluation loop, catching time-sensitive changes and subtle reasoning errors. This reduces the risk of acting on misinformation, especially in fast-moving domains like law, markets, and technology. It also gives builders a clearer signal for improving their agents, accelerating practical progress. In short, better judging leads to safer, smarter uses of AI in daily life.
Detailed Explanation
Tap terms for definitions01Background & Problem Definition
š Hook: Imagine your class writes science reports. Two students might write very different reports about the same topic, and both could still be great. So, how do you grade fairly when there isnāt just one correct answer?
š„¬ The Concept (AI Agents): An AI agent is a computer helper that reads, searches, and writes to complete tasks on its own. How it works:
- It receives a goal (like āresearch TikTokās legal status in the U.S.ā).
- It plans steps: search, read, compare, summarize.
- It uses tools (like web search) and then produces a report. Why it matters: If we donāt evaluate well, we might trust reports that look good but arenāt actually correct. š Anchor: Like a student doing a library project: the agent looks up books (web pages), takes notes, and writes a paper.
š Hook: You know how grades need clear rules? In research, we also need clear ways to judge quality.
š„¬ The Concept (Research Evaluation): Research evaluation is judging how good a report is across several dimensions, not just if a single answer is right. How it works:
- Decide what āgoodā means (clarity, completeness, logic, truth, sources).
- Measure each part with specific checks.
- Combine the checks into a final score. Why it matters: Without the right checks, we reward style over substance. š Anchor: Teachers use rubrics (clarity, evidence, organization) to grade essays; research needs rubrics too.
š Hook: Picture a shiny poster that looks amazing but has the wrong facts. It fools you at first glance.
š„¬ The Concept (Mirage of Synthesis): The Mirage of Synthesis is when a report seems excellentāsmooth writing and matching citationsābut hides outdated info, wrong facts, or shaky logic. How it works:
- The report sounds fluent and authoritative.
- It includes citations that match the text.
- But the facts may be old, misinterpreted, or logically weak. Why it matters: We can be tricked into trusting something that isnāt true or current. š Anchor: A history project with perfect design but wrong dates still deserves a low score.
š Hook: Think of four pillars holding up a building. If even one cracks, the building is unsafe.
š„¬ The Concept (Four-Vertical Taxonomy): A taxonomy is a tidy way to group the pieces of report quality into four verticals. How it works:
- Presentation Quality: Is the writing clear and well-organized?
- Task Compliance: Did it follow the instructions and cover the needed topics?
- Analytical Depth: Is the reasoning strong and insightful?
- Source Quality: Are sources used properly, faithfully, and crediblyāand are claims actually true? Why it matters: Checking all four catches surface polish and deeper truth. š Anchor: Like grading an essay for writing, following the prompt, strong arguments, and trustworthy references.
š Hook: Imagine a ref who canāt see in the dark trying to judge a night game.
š„¬ The Concept (Capability Mismatch): Capability mismatch is when the evaluator canāt do what the research agent can doālike browse, update facts, and verify claims. How it works:
- Static judges only read the report.
- They donāt search the web or check dates.
- They miss time-sensitive mistakes and ungrounded logic. Why it matters: Reports that āsound rightā slip by even when theyāre wrong. š Anchor: A spelling bee judge who canāt hear letters clearly will score unfairly.
š Hook: When both teams use the same rules and gear, the game is fair.
š„¬ The Concept (Capability Parity): Capability parity means the evaluator should have similar abilities as the agent it judges, including tool use and reasoning. How it works:
- Equip the evaluator with web search and other tools.
- Let it check facts and timelines directly.
- Have it probe the logic, not just the surface. Why it matters: Fair judging needs equal power to verify. š Anchor: A math contest grader who can also use a calculator checks answers more accurately.
Before this paper, most evaluation was static: LLM judges or citation alignment workflows. These caught fluency and whether a citation text matched the claim. But they missed two big issues: time decay (old facts) and external truth (a claim can match its citation but still be false today). That creates the Mirage of Synthesis. The missing piece was an evaluator that can actively research like the agent.
This paper fills that gap with DREAM, which makes the evaluation itself agentic. It uses universal checks for writing and source behavior and builds a custom, up-to-date checklist and reasoning probes for each query. The stakes are real: students, journalists, doctors, and business analysts rely on reports that must be not just pretty, but correct now and logically sound. If evaluation canāt tell the difference, people make decisions on shaky ground.
02Core Idea
š Hook: Imagine grading a current events report without the internetāhow would you know if yesterdayās facts changed today?
š„¬ The Concept (DREAM): DREAM turns the evaluator into an active research agent so it can verify facts, dates, and reasoningājust like the report-writing agent. How it works:
- Build a custom evaluation protocol for the question: some checks are universal (static), others are freshly built for this query (adaptive).
- Use the right ājudgeā for each check: a plain LLM when reading is enough, an agent with tools when you must verify, and a workflow for citations and domain credibility.
- Score writing, factuality, coverage of key info, reasoning quality, citation integrity, and source authority. Why it matters: Without an agentic evaluator, we keep rewarding well-written but wrong or outdated reports. š Anchor: Like a science fair judge who brings a thermometer and scale to re-check claims instead of only reading the poster.
Three analogies for the same idea:
- Detective vs. Desk Clerk: A desk clerk reads and stamps forms; a detective leaves the desk, gathers evidence, and cross-checks alibis. DREAM is the detective.
- Fresh Groceries: Old rubrics are like a shopping list from last month. DREAM re-checks whatās fresh today before deciding if your meal plan (report) is good.
- Ref with Replay: A ref using instant replay (web tools) sees what really happened, not just what it looked like at full speed. DREAM uses replay for facts and logic.
Before vs. After:
- Before: Judges rewarded fluency and citation text match. Time decay and extrinsic truth slipped by.
- After: Judges test time-sensitive facts (KIC), probe reasoning steps (RQ), and check truth beyond the given citations (Factuality), while still grading writing and citation behavior.
Why it works (intuition):
- If the task itself needs web search, then the judge must also search. Thatās capability parity.
- Turning ācoverageā into yes/no questions from up-to-date sources (KIC) forces reports to mention what matters now.
- Turning āreasoningā into step-checked plans (RQ) exposes circular logic and missing support.
- Checking truth against the live web (Factuality) defeats the trap where a wrong claim cites a matching but outdated page.
š Hook: You know how a good game plan breaks a big goal into clear plays?
š„¬ The Concept (Building Blocks): DREAM is built from smaller pieces that each guard a failure point. How it works:
- Static Metrics: Writing Quality (readability/organization), Factuality (truth vs. the world), Citation Integrity (attribution + faithfulness), Domain Authoritativeness (source reputation).
- Adaptive Metrics: Key-Information Coverage (KIC) builds a fresh checklist of must-mention facts; Reasoning Quality (RQ) builds challenge questions plus a verification plan.
- Evaluators: LLM Evaluator (no tools) for reading judgments; Agent Evaluator (with tools) for RQ; Workflow Evaluator for citation checks and domain credibility. Why it matters: Each block fixes a blind spotātogether they see what old benchmarks missed. š Anchor: Like checking a bike: pump the tires (writing), test the brakes (reasoning), tighten the chain (citations), and confirm the helmet is certified (domain authority).
03Methodology
At a high level: Research Query ā Phase 1: Protocol Creation (Static + Adaptive metrics) ā Phase 2: Protocol Execution (route each metric to the right evaluator) ā Final Scores.
Prerequisite concepts with sandwich explanations:
š Hook: Think of a standard set of classroom rules used for every assignment. š„¬ The Concept (Static Metrics): Static metrics are universal checks that apply to any report. How it works: They include Writing Quality, Factuality (truth vs. world), Citation Integrity (attribution + faithfulness), and Domain Authoritativeness (source reputation). Why it matters: Without them, weād miss basic presentation and source hygiene. š Anchor: Every essay needs clear writing and reliable referencesāno matter the topic.
š Hook: For a field trip, you bring a custom checklist (raincoat if it might rain) instead of the same list every time. š„¬ The Concept (Adaptive Metrics): Adaptive metrics are built fresh for the exact question, capturing time-sensitive and topic-specific expectations. How it works: The agent researches the query and converts key facts into yes/no checks (KIC), and it designs hard reasoning probes plus a verification plan (RQ). Why it matters: Without adaptation, rubrics stay blind to what changed this week. š Anchor: Packing sunscreen for a sunny trip and umbrella for a rainy oneācustomized prep wins.
Phase 1: Protocol Creation (how the recipe starts)
- Static Metrics are set up:
- Writing Quality (WQ): A fixed rubric scores clarity of ideas and content, organization, and sentence fluency.
- Factuality: Extract top claims, create neutralized web queries (to avoid confirmation bias), collect supporting and opposing evidence, then judge each claim as supported, partial, contradicted, or unverifiable.
- Citation Integrity (CI): ⢠Claim Attribution (CA): What fraction of verifiable claims have explicit citations? ⢠Citation Faithfulness (CF): For cited claims, does the source text really support the claim? ⢠Combine CA and CF so you must both cite and cite correctly.
- Domain Authoritativeness (DA): Check if cited domains are reputable (e.g., government, academic, established news) and average their credibility.
What breaks without each step:
- Without WQ: Good content gets lost in messy writing.
- Without Factuality: Plausible but false claims sneak through.
- Without CI: Either no citations (opaque) or lots of wrong citations (misleading) go unpunished.
- Without DA: Shaky blogs count the same as gold-standard institutions.
- Adaptive Metrics are created by a tool-using agent:
- Key-Information Coverage (KIC): ⢠The agent searches up-to-date sources. ⢠It identifies essential, time-sensitive facts and turns each into a yes/no checklist item. ⢠This converts ādid you cover what matters now?ā into a grounded test.
- Reasoning Quality (RQ): ⢠The agent writes challenging, query-specific questions. ⢠It prepares a validation plan: extract the reportās reasoning chain, check external sources, and compare carefully. ⢠The plan ensures we evaluate the logic itself, not just its surface.
Example with real data (TikTok legal status): KIC asks, āDoes the report mention the current divestiture deadline (Jan 23, 2026)?ā If the report uses old dates, it fails that checklist item.
Phase 2: Protocol Execution (running the tests)
š Hook: Different tools for different jobsāuse a ruler to measure, a scale to weigh. š„¬ The Concept (Evaluator Routing): DREAM sends each metric to the simplest evaluator that has the needed capability. How it works:
- LLM Evaluator (no tools): Grades Writing Quality and checks KIC presence in the text.
- Agent Evaluator (with tools): Executes RQ validation plans, retrieving fresh evidence to test logic.
- Workflow Evaluator: Runs pipelines for Factuality, Citation Integrity, and Domain Authoritativeness. Why it matters: Using the right tool keeps judging accurate and efficient. š Anchor: You donāt use scissors to hammer a nail; you pick the proper tool.
Detailed, step-by-step ārecipesā
- Writing Quality: Read the report; score ideas/content, organization, sentence fluency using a fixed rubric; average the parts.
- Factuality:
- Extract 30 salient claims.
- Write neutral search queries (e.g., ācurrent divestiture deadline for TikTokā instead of copying the claim).
- Retrieve multiple sources; pull both supporting and opposing snippets.
- Judge each claim (supported/partial/contradicted/unverifiable) and average across claims.
- Citation Integrity:
- Extract verifiable claims and their citations.
- Compute Claim Attribution: fraction with citations.
- For cited ones, test source text vs. claim for faithfulness.
- Combine attribution and faithfulness so agents must both cite and cite correctly.
- Domain Authoritativeness:
- Collect unique domains from citations.
- Judge each domainās reputation with a rubric (e.g., academic/government high; personal blogs low).
- Average to get a task score.
- KIC (Adaptive):
- Agent searches the live web.
- Extracts essential current facts.
- Writes yes/no checklist items grounded in those facts.
- LLM checks if the report covers each item.
- RQ (Adaptive):
- Agent drafts challenging questions.
- Builds a validation plan: extract reasoning chain, gather external evidence, and compare.
- Executes the plan with tools and deducts points for logical gaps.
The secret sauce
- Convert vague goals (ābe comprehensiveā) into concrete, verifiable items (KIC checklist) built from todayās facts.
- Evaluate logic by plan-and-verify steps (RQ), not just by how convincing the prose sounds.
- Separate ādoes the source text match?ā from āis the claim actually true in the world?ā so faithful-but-false claims are caught.
Concrete anchors
- If a report on TikTok omits āJan 23, 2026 deadline,ā KIC flags it.
- If a report argues āA causes Bā but sources donāt support the link, RQ downgrades it.
- If a report cites a neat-looking blog for a medical claim, DA lowers the score even if the wording matches.
04Experiments & Results
š Hook: If a smoke alarm canāt detect smoke, itās not much of an alarm. We tested whether DREAM actually āsmellsā the hidden problems old tests miss.
- The Test: What and why we measured
- We checked if the agent-built protocols (KIC and RQ) are clear, relevant, and verifiable to humans.
- We tested time awareness: does KIC penalize outdated reports?
- We tested reasoning detection: does RQ catch subtle logic errors hidden in fluent writing?
- We tested truth beyond citations: can Factuality catch well-cited but false claims?
- We checked whether writing can be graded reference-free yet align with human preference.
-
The Competition: We compared DREAM to popular benchmarks like DeepResearch Bench (RACE) and citation alignment (FACT).
-
The Scoreboard (with context)
- Human study on protocols: Annotators gave the full agent-with-retrieval version top scores (about 0.92ā0.93 on a 0ā1 scale) for relevance, clarity, verifiability, and validation quality. Thatās like getting an A when the simpler LLM baseline gets a CāB range.
- Temporal awareness: On 20 time-volatile queries, DREAMāKIC dropped strongly as info got older (e.g., from about 79 for current to ~22 for Jan 2024), clearly signaling staleness. DRBāRACE barely movedālike a thermometer stuck near room temperature.
- Reasoning flaws: We created pairs of reports where one had injected logical errors but kept great writing. RACEās scores barely changed (~9% drop), sometimes even praising the flawed one. DREAMāRQ cut scores by ~40% on averageālike differentiating an A essay from a D when reasoning breaks.
- Factuality vs. citation: We fed pairs with correct claims and plausible-but-false claims that still matched their citations. The citation-alignment metric stayed high (fooled by matching text). DREAMāFactuality fell steadily as we added more false claims, tracking the true error rate. Thatās exactly what a real fact-checker should do.
- Writing Quality alignment: DREAMās reference-free writing score correlated well (Kendallās Ļ ā 0.6) with DRBās human-validated rankingsāsolidly within typical human agreement.
- Surprising findings
- Fluent, well-structured reports can trick static judges into high scores even when important facts are missing or old.
- Citation behavior showed two opposite failure modes in open-source agents: (a) cite often but unfaithfully, or (b) cite rarely, leaving claims unattributed. Both are risky.
- Benchmarking three open agents (LangChain Open DR, Smolagents Open DR, Tongyi Deep Research)
- All struggled with Citation Integrity for different reasons (low attribution or low faithfulness).
- Smolagents often led on Writing Quality, Factuality, KIC, and RQ, yet still had near-zero citation discipline on some datasets.
- Relative rankings stayed stable across different judge backbones (Claude, DeepSeek, Kimi), showing robustness.
Bottom line: DREAM is much more sensitive to what truly mattersātime, truth, and reasoningāwhile staying practical and scalable without needing a gold āanswerā report.
05Discussion & Limitations
š Hook: Even the best toolbox has limitsāyou still need power and materials to build the house.
- Limitations (be specific)
- Tool dependency: If search engines or APIs are slow, down, or biased, agentic checks can suffer.
- Cost and latency: Multi-step verification takes more compute and time than a single static judge call.
- Scope: DREAM grades final outputs, not the agentās research journey (like search efficiency or source discovery skill).
- Required resources
- Reliable web search and content retrieval (news, papers, docs).
- An LLM backbone for judging and an agent framework for tool use.
- Budget and time for multi-step evaluations, plus caching to save repeated checks.
- When NOT to use DREAM
- Fully closed-book tasks where truth is defined by a fixed, static dataset (no live updates needed).
- Extremely time- or cost-constrained settings where only a quick, shallow signal is acceptable.
- Highly subjective evaluations (e.g., pure creative style) where external truth checking is irrelevant.
- Open questions
- Process evaluation: How to score the research path (good source discovery, efficient browsing) in addition to the final product?
- Optimization: Can we smartly skip or cache steps to cut cost while keeping sensitivity high?
- Adversarial robustness: How to resist coordinated misinformation or SEO spam targeting evaluators?
- Domain adaptation: How to tune DA (domain credibility) across specialized fields without penalizing niche but authoritative sources?
š Anchor: Think of DREAM as a high-precision lab testāmore accurate than a quick strip test, but it needs equipment, time, and careful handling.
06Conclusion & Future Work
-
Three-sentence summary: Deep research reports can look excellent while hiding outdated facts or weak logicāthe Mirage of Synthesis. DREAM fixes this by giving the evaluator agent-like powers (capability parity) to search, verify, and probe reasoning using both static and adaptive metrics. Experiments show DREAM is far more sensitive to time decay, factual errors, and reasoning flaws than popular static benchmarks.
-
Main achievement: Turning evaluation itself into an agentic, tool-using process that builds query-specific checks (KIC, RQ) and verifies truth beyond citation alignment.
-
Future directions: Add process-level scoring for research trajectories, reduce cost with smart caching and selective checks, and harden against adversarial misinformation. Also refine domain authority scoring for specialized niches.
-
Why remember this: As AI agents research on the live web, judging them with frozen rubrics is like grading current events with last yearās newspaperāDREAM updates the judge so it stays fair, factual, and future-ready.
Practical Applications
- ā¢Newsrooms: Vet AI-generated briefings against the live web to avoid publishing outdated or false claims.
- ā¢Healthcare content: Check medical claims beyond citations to ensure they align with current guidelines.
- ā¢Corporate research: Validate market analyses for up-to-date figures and sound causal reasoning before decisions.
- ā¢Education: Grade student reports with adaptive checklists that reflect the latest developments.
- ā¢Policy analysis: Ensure legal timelines, rulings, and stakeholder positions are current and accurately represented.
- ā¢Product documentation: Confirm technical guides cite authoritative sources and present correct, current info.
- ā¢Search/RAG systems: Benchmark retrieval and synthesis quality with KIC and RQ to guide model improvements.
- ā¢Compliance teams: Monitor regulation changes with time-aware coverage checks to prevent risk from stale info.
- ā¢Agent development: Use DREAM scores to pinpoint weaknesses (e.g., citation faithfulness vs. attribution).
- ā¢Fact-checking platforms: Automate reference-free truth verification that catches well-cited falsehoods.