pith. machine review for the scientific record. sign in

arxiv: 2604.26180 · v1 · submitted 2026-04-28 · 💻 cs.DB · cs.AI· cs.CL

Recognition: unknown

Evergreen: Efficient Claim Verification for Semantic Aggregates

Alexander W. Lee, Anupam Datta, Benjamin Han, Sam Yeom, Shayak Sen, Ugur Cetintemel

Pith reviewed 2026-05-07 12:18 UTC · model grok-4.3

classification 💻 cs.DB cs.AIcs.CL
keywords semanticevergreencostlatencyverificationloweraggregateclaim
0
0 comments X

The pith

Evergreen verifies claims from semantic aggregates by compiling them into optimized semantic queries, achieving F1=1.0 at 3.2x lower cost and 4.0x lower latency than unoptimized verification on restaurant review benchmarks.

A machine-rendered reading of the paper's core claim, the machinery that carries it, and where it could break.

Semantic aggregation lets LLMs turn large tables of data, such as restaurant reviews, into short natural language statements that include counts, averages, or comparisons. These statements can include claims that do not match the actual data. Checking them directly with an LLM is expensive and often fails because the full table exceeds the model's memory limit. Evergreen turns each claim into a structured verification query that runs inside the same semantic engine. It adds shortcuts that stop early once enough evidence appears, sort data by relevance first, and estimate confidence without full scans. It also records exactly which data rows support the final yes or no answer using mathematical tracking of evidence. On real restaurant review collections, the system reached perfect accuracy scores while using far less computation than asking the LLM directly or using retrieval agents.

Core claim

On a benchmark of real-world restaurant review datasets reflecting production-inspired workloads, Evergreen achieves excellent verification quality (F1 = 1.00) with a strong LLM while reducing cost by 3.2x and latency by 4.0x compared to unoptimized verification.

Load-bearing premise

That the underlying semantic query engine can correctly compile and execute the verification queries for claims involving quantifiers, groupings, and comparisons without exceeding practical limits on LLM context or introducing errors in provenance tracking.

read the original abstract

With recent semantic query processing engines, semantic aggregation has become a primitive operator, enabling the reduction of a relation into a natural language aggregate using an LLM. However, the resulting semantic aggregate may contain claims that are not grounded in the underlying relation. Verifying such claims is challenging: they often involve quantifiers, groupings, and comparisons over relations that far exceed LLM context windows and require a costly combination of semantic and symbolic processing. We present Evergreen, a system that recasts claim verification as a semantic query processing task with tailored optimizations and provenance capture. Evergreen compiles each claim into a declarative semantic verification query and executes it on the same engine that produced the aggregate. To reduce cost and latency, Evergreen avoids unnecessary LLM calls through verification-aware optimizations (early stopping, relevance sorting, and estimation with confidence sequences) and general-purpose optimizations for semantic queries (operator fusion, similarity filtering, and prompt caching). Each verdict is accompanied by citations that identify a minimal set of tuples justifying the result, with semantics based on semiring provenance for first-order logic. On a benchmark of real-world restaurant review datasets reflecting production-inspired workloads, Evergreen achieves excellent verification quality (F1 = 1.00) with a strong LLM while reducing cost by 3.2x and latency by 4.0x compared to unoptimized verification. Even with a significantly weaker LLM, Evergreen outperforms a strong LLM-as-a-judge baseline in F1 at 48x lower cost and 2.3x lower latency. Relative to a retrieval-augmented agent, Evergreen compares favorably in F1 and latency with similar cost when both use a strong LLM; yet, with a much weaker LLM, it achieves the same F1 at 63x lower cost and 4.2x lower latency.

Editorial analysis

A structured set of objections, weighed in public.

Desk editor's note, referee report, simulated authors' rebuttal, and a circularity audit. Tearing a paper down is the easy half of reading it; the pith above is the substance, this is the friction.

Referee Report

3 major / 2 minor

Summary. The manuscript presents Evergreen, a system that recasts verification of claims in LLM-generated semantic aggregates as a declarative semantic query processing task. Claims are compiled into verification queries executed on the same engine, augmented with verification-aware optimizations (early stopping, relevance sorting, confidence-sequence estimation) and general semantic-query optimizations (operator fusion, similarity filtering, prompt caching), plus semiring-based provenance for minimal tuple citations. On a benchmark of real-world restaurant review datasets, it reports F1=1.00 with a strong LLM at 3.2x lower cost and 4.0x lower latency versus unoptimized verification, and favorable comparisons to LLM-as-a-judge and retrieval-augmented agent baselines even with weaker LLMs.

Significance. If the empirical claims hold under rigorous validation, the work would be significant for database systems research on semantic query processing: it demonstrates a practical path to reliable, low-cost verification of complex aggregates involving quantifiers and groupings, while preserving provenance. The ability to outperform strong baselines at much lower cost with weaker LLMs is a notable strength that could broaden adoption of LLM-based aggregates in production workloads.

major comments (3)
  1. [§5] §5 (Experiments) and benchmark description: the headline F1=1.00, 3.2x cost, and 4.0x latency results rest on a benchmark whose construction, claim generation process, ground-truth annotation, workload selection criteria, and controls against post-hoc tuning are not described in sufficient detail. Without these, it is impossible to assess whether the reported gains are robust or could be artifacts of benchmark choice.
  2. [§3] §3 (System Design) and §4 (Optimizations): the central assumption that claim-to-query compilation plus the listed optimizations (early stopping, operator fusion, provenance tracking) preserve exact semantics for quantifiers, groupings, and comparisons is not independently validated. No experiments or checks are reported for context-window overflow, dropped tuples in semiring provenance, or fidelity of the compiled verification queries against the original claims.
  3. [§5.2] Table 2 / §5.2 (Baseline comparisons): the exact prompts, model versions, and implementation details for the 'unoptimized verification', 'LLM-as-a-judge', and 'retrieval-augmented agent' baselines are not provided, nor are statistical significance tests or variance across runs for the F1, cost, and latency deltas. This makes the 48x/63x cost claims difficult to reproduce or generalize.
minor comments (2)
  1. [§3.1] Notation for semiring provenance and confidence sequences could be clarified with a small running example in §3.1 to aid readers unfamiliar with the formalism.
  2. [Figure 3] Figure 3 (latency/cost plots) would benefit from error bars or explicit mention of the number of runs averaged.

Simulated Author's Rebuttal

3 responses · 0 unresolved

We thank the referee for their constructive and detailed feedback on our manuscript. We address each of the major comments point by point below, indicating where we agree that additional material is warranted and outlining the corresponding revisions.

read point-by-point responses
  1. Referee: [§5] §5 (Experiments) and benchmark description: the headline F1=1.00, 3.2x cost, and 4.0x latency results rest on a benchmark whose construction, claim generation process, ground-truth annotation, workload selection criteria, and controls against post-hoc tuning are not described in sufficient detail. Without these, it is impossible to assess whether the reported gains are robust or could be artifacts of benchmark choice.

    Authors: We appreciate the referee highlighting the need for greater detail. While §5 characterizes the benchmark as real-world restaurant review datasets chosen to reflect production-inspired workloads, we agree that the claim generation process, ground-truth annotation protocol, workload selection criteria, and safeguards against post-hoc tuning require explicit description to support reproducibility and robustness claims. In the revised manuscript we will add a dedicated subsection to §5 that specifies (i) how claims involving quantifiers, groupings, and comparisons were synthesized, (ii) the annotation procedure and inter-annotator agreement for ground truth, (iii) the criteria used to select workloads independently of optimization development, and (iv) the controls employed to prevent tuning on the evaluation set. revision: yes

  2. Referee: [§3] §3 (System Design) and §4 (Optimizations): the central assumption that claim-to-query compilation plus the listed optimizations (early stopping, operator fusion, provenance tracking) preserve exact semantics for quantifiers, groupings, and comparisons is not independently validated. No experiments or checks are reported for context-window overflow, dropped tuples in semiring provenance, or fidelity of the compiled verification queries against the original claims.

    Authors: The compilation rules in §3 are defined to translate each claim element (quantifiers, groupings, comparisons) into the corresponding declarative operators of the underlying semantic query engine, thereby preserving semantics by construction. The semiring provenance formalism adopted in §4 is the standard provenance semantics for first-order logic and therefore tracks every derivation without dropping tuples. Nevertheless, we acknowledge that the submitted version does not contain separate validation experiments addressing context-window overflow, provenance completeness, or end-to-end fidelity between compiled queries and original claims. We will add a validation subsection (either in §4 or §5) that reports (a) fidelity comparisons on a held-out set of small claims where Evergreen outputs are cross-checked against direct LLM verification, and (b) provenance coverage statistics confirming that no relevant tuples are omitted. revision: partial

  3. Referee: [§5.2] Table 2 / §5.2 (Baseline comparisons): the exact prompts, model versions, and implementation details for the 'unoptimized verification', 'LLM-as-a-judge', and 'retrieval-augmented agent' baselines are not provided, nor are statistical significance tests or variance across runs for the F1, cost, and latency deltas. This makes the 48x/63x cost claims difficult to reproduce or generalize.

    Authors: We concur that reproducibility requires the exact prompts, model versions, and implementation details of all baselines. In the revised manuscript we will append a supplementary section containing (i) the verbatim prompts used for each baseline, (ii) the precise model identifiers and temperature settings, and (iii) pseudocode or repository references for the retrieval-augmented agent. In addition, we will augment Table 2 and the accompanying text with per-run variance (standard deviation across five independent executions) and the results of paired statistical significance tests (e.g., Wilcoxon signed-rank) for all reported F1, cost, and latency differences. revision: yes

Axiom & Free-Parameter Ledger

0 free parameters · 1 axioms · 0 invented entities

The central claim rests on the assumption that a semantic query engine exists and can be extended with the described optimizations, plus standard assumptions about LLM behavior on structured prompts.

axioms (1)
  • domain assumption The semantic query processing engine can correctly execute compiled verification queries involving quantifiers and groupings.
    Invoked when the paper states that claims are compiled and executed on the same engine.

pith-pipeline@v0.9.0 · 5640 in / 1243 out tokens · 81719 ms · 2026-05-07T12:18:28.876908+00:00 · methodology

discussion (0)

Sign in with ORCID, Apple, or X to comment. Anyone can read and Pith papers without signing in.