Recognition: 1 theorem link
· Lean TheoremFEVER: a large-scale dataset for Fact Extraction and VERification
Pith reviewed 2026-05-13 19:46 UTC · model grok-4.3
The pith
The FEVER dataset introduces 185,445 claims to benchmark fact verification against textual sources.
A machine-rendered reading of the paper's core claim, the machinery that carries it, and where it could break.
Core claim
FEVER consists of 185,445 claims generated by altering sentences extracted from Wikipedia and subsequently verified without knowledge of the sentence they were derived from. The claims are classified as Supported, Refuted or NotEnoughInfo by annotators achieving 0.6841 in Fleiss κ. For the first two classes, the annotators also recorded the sentence(s) forming the necessary evidence for their judgment. To characterize the challenge, a pipeline approach achieves 31.87% accuracy on labeling with correct evidence and 50.91% without.
What carries the argument
The FEVER dataset of Wikipedia-altered claims with blind verification labels and evidence annotations.
If this is right
- Automated claim verification systems must handle both evidence retrieval and classification to succeed on this benchmark.
- The gap between pipeline performance and potential oracles shows that current methods have substantial room for improvement.
- The dataset can drive research on verifying claims in large text corpora like Wikipedia.
- High annotator agreement establishes a reliable human baseline for the task.
Where Pith is reading between the lines
- End-to-end models might close the performance gap by learning retrieval and verification jointly.
- The approach could be extended to claims requiring multiple evidence sentences or cross-document verification.
- This benchmark may help evaluate systems for detecting misinformation in real-world textual sources.
Load-bearing premise
Annotators without knowledge of the original sentence can reliably determine the correct label and evidence for claims created by altering Wikipedia sentences.
What would settle it
If new annotators given the original sentences produce substantially different labels or evidence sets, the blind annotation process would not be a valid test of verification.
read the original abstract
In this paper we introduce a new publicly available dataset for verification against textual sources, FEVER: Fact Extraction and VERification. It consists of 185,445 claims generated by altering sentences extracted from Wikipedia and subsequently verified without knowledge of the sentence they were derived from. The claims are classified as Supported, Refuted or NotEnoughInfo by annotators achieving 0.6841 in Fleiss $\kappa$. For the first two classes, the annotators also recorded the sentence(s) forming the necessary evidence for their judgment. To characterize the challenge of the dataset presented, we develop a pipeline approach and compare it to suitably designed oracles. The best accuracy we achieve on labeling a claim accompanied by the correct evidence is 31.87%, while if we ignore the evidence we achieve 50.91%. Thus we believe that FEVER is a challenging testbed that will help stimulate progress on claim verification against textual sources.
Editorial analysis
A structured set of objections, weighed in public.
Referee Report
Summary. The manuscript introduces the FEVER dataset of 185,445 claims generated by altering Wikipedia sentences and verified by annotators without access to the source sentences. Claims are labeled Supported, Refuted, or NotEnoughInfo with an overall Fleiss' κ of 0.6841; evidence sentences are recorded for the first two classes. A pipeline baseline achieves 31.87% accuracy when supplied with the correct evidence and 50.91% when evidence is ignored, leading the authors to conclude that FEVER is a challenging testbed for claim verification against textual sources.
Significance. If the annotation quality is adequately demonstrated, the public release of this large-scale dataset with associated evidence would constitute a substantial contribution to NLP research on fact extraction and verification. The concrete reporting of inter-annotator agreement and two baseline accuracies provides a clear starting point for future work and could stimulate measurable progress on the task.
major comments (1)
- [Abstract] Abstract: The claim that FEVER forms a valid benchmark rests on the reliability of the three-way labels and evidence annotations. Only an aggregate Fleiss' κ of 0.6841 is reported; no class-wise agreement, no evidence-sentence selection agreement, and no information on the number of annotators per claim or disagreement resolution are provided. Because claims are artificially generated by sentence alteration, this moderate agreement level directly affects the interpretation of the 31.87% and 50.91% baseline figures and requires additional quantitative support.
Simulated Author's Rebuttal
We thank the referee for their constructive review and for acknowledging the potential significance of the FEVER dataset. We address the single major comment below and will make the requested revisions to strengthen the reporting of annotation quality.
read point-by-point responses
-
Referee: [Abstract] Abstract: The claim that FEVER forms a valid benchmark rests on the reliability of the three-way labels and evidence annotations. Only an aggregate Fleiss' κ of 0.6841 is reported; no class-wise agreement, no evidence-sentence selection agreement, and no information on the number of annotators per claim or disagreement resolution are provided. Because claims are artificially generated by sentence alteration, this moderate agreement level directly affects the interpretation of the 31.87% and 50.91% baseline figures and requires additional quantitative support.
Authors: We agree that the abstract would benefit from expanded reporting on annotation reliability. The full manuscript (Section 3) details that each claim was annotated by three crowd workers, with final labels determined by majority vote and evidence sentences selected by the same workers; disagreements were resolved via discussion among annotators. To directly address the concern, we will revise the abstract to report class-wise Fleiss' κ values, inter-annotator agreement on evidence sentence selection, and the per-claim annotator count. We will also add a brief clarification that the moderate aggregate κ is consistent with the inherent subjectivity of claim verification and does not invalidate the baselines: the 31.87% accuracy with gold evidence already demonstrates the difficulty of the verification step, while the 50.91% figure without evidence underscores the need for retrieval. revision: yes
Circularity Check
No circularity: empirical dataset construction with independent baselines
full rationale
The paper constructs a dataset by altering Wikipedia sentences and collecting human annotations (Fleiss κ=0.6841), then reports direct empirical baseline accuracies from a pipeline (31.87% with evidence, 50.91% without). No equations, derivations, fitted parameters renamed as predictions, or self-citations appear in the provided text. The central claim that FEVER is a challenging testbed follows from these reported numbers without any reduction by construction to the inputs.
Axiom & Free-Parameter Ledger
axioms (1)
- domain assumption Wikipedia sentences can be altered to create claims that annotators can classify as Supported, Refuted or NotEnoughInfo without seeing the original sentence
Lean theorems connected to this paper
-
IndisputableMonolith.Foundation.RealityFromDistinctionreality_from_one_distinction unclearIt consists of 185,445 claims generated by altering sentences extracted from Wikipedia and subsequently verified without knowledge of the sentence they were derived from. The claims are classified as Supported, Refuted or NotEnoughInfo by annotators achieving 0.6841 in Fleiss κ.
Forward citations
Cited by 22 Pith papers
-
Needle-in-RAG: Prompt-Conditioned Character-Level Traceback of Poisoned Spans in Retrieved Evidence
RAGCharacter localizes poisoned character spans in RAG evidence via prompt-conditioned counterfactual masking and achieves the best accuracy-over-attribution trade-off across tested attacks and models.
-
InvEvolve: Evolving White-Box Inventory Policies via Large Language Models with Performance Guarantees
InvEvolve evolves white-box inventory policies from LLMs with statistical safety guarantees and outperforms classical and deep learning methods on synthetic and real retail data.
-
HeadRank: Decoding-Free Passage Reranking via Preference-Aligned Attention Heads
HeadRank improves decoding-free passage reranking by preference-aligning attention heads to increase discriminability in middle-context documents, outperforming baselines on 14 benchmarks with only 211 training queries.
-
Retrieval-Augmented Generation for Knowledge-Intensive NLP Tasks
RAG models set new state-of-the-art results on open-domain QA by retrieving Wikipedia passages and conditioning a generative model on them, while also producing more factual text than parametric baselines.
-
From Articles to Premises: Building PrimeFacts, an Extraction Methodology and Resource for Fact-Checking Evidence
PrimeFacts extracts decontextualized premises from fact-check articles, raising evidence retrieval MRR by up to 30% and verdict prediction Macro-F1 by 10-20 points over baselines.
-
CAR: Query-Guided Confidence-Aware Reranking for Retrieval-Augmented Generation
CAR reranks documents in RAG by promoting those that increase generator confidence (via answer consistency sampling) and demoting those that decrease it, yielding NDCG@5 gains on BEIR datasets that correlate with F1 i...
-
When Embedding-Based Defenses Fail: Rethinking Safety in LLM-Based Multi-Agent Systems
Embedding-based defenses fail against attacks that align malicious message embeddings with benign ones in LLM multi-agent systems, but token-level confidence scores improve robustness by enabling better pruning of sus...
-
InvEvolve: Evolving White-Box Inventory Policies via Large Language Models with Performance Guarantees
InvEvolve uses LLMs and RL to generate certified inventory policies that outperform classical and deep learning methods on synthetic and real data while providing multi-period performance guarantees.
-
Learning to Route Queries to Heads for Attention-based Re-ranking with Large Language Models
RouteHead trains a lightweight router to dynamically select optimal LLM attention heads per query for improved attention-based document re-ranking.
-
Guaranteeing Knowledge Integration with Joint Decoding for Retrieval-Augmented Generation
GuarantRAG improves RAG accuracy up to 12.1% and cuts hallucinations 16.3% by decoupling parametric reasoning from evidence integration via contrastive DPO and joint decoding.
-
Data, Not Model: Explaining Bias toward LLM Texts in Neural Retrievers
Bias toward LLM texts in neural retrievers arises from artifact imbalances between positive and negative documents in training data that are absorbed during contrastive learning.
-
NV-Embed: Improved Techniques for Training LLMs as Generalist Embedding Models
NV-Embed achieves first place on the MTEB leaderboard across 56 tasks by combining a latent attention layer, causal-mask removal, two-stage contrastive training, and data curation for LLM-based embedding models.
-
Unsupervised Dense Information Retrieval with Contrastive Learning
Contrastive learning trains unsupervised dense retrievers that beat BM25 on most BEIR datasets and support cross-lingual retrieval across scripts.
-
An Information-Geometric Framework for Stability Analysis of Large Language Models under Entropic Stress
A thermodynamic-inspired information-geometric framework defines a composite LLM stability score that outperforms a utility-entropy baseline by 0.0299 on average across 80 observations, with gains increasing at higher...
-
Learning Uncertainty from Sequential Internal Dispersion in Large Language Models
SIVR detects LLM hallucinations by learning from token-wise and layer-wise variance patterns in internal hidden states, outperforming baselines with better generalization and less training data.
-
Verify Before You Commit: Towards Faithful Reasoning in LLM Agents via Self-Auditing
SAVeR adds self-auditing of internal beliefs in LLM agents via persona-based candidates and constraint-guided repairs, improving faithfulness on six benchmarks without hurting task performance.
-
Hallucination Basins: A Dynamic Framework for Understanding and Controlling LLM Hallucinations
LLM hallucinations arise from task-dependent basins in latent space, with separability varying by task and geometry-aware steering reducing their probability.
-
Multilingual E5 Text Embeddings: A Technical Report
Open-source multilingual E5 embedding models are trained via contrastive pre-training on 1 billion text pairs and fine-tuning, with an instruction-tuned model matching English SOTA performance.
-
Text Embeddings by Weakly-Supervised Contrastive Pre-training
E5 text embeddings trained with weakly-supervised contrastive pre-training on CCPairs outperform BM25 on BEIR zero-shot and achieve top results on MTEB, beating much larger models.
-
Trust but Verify: Introducing DAVinCI -- A Framework for Dual Attribution and Verification in Claim Inference for Language Models
DAVinCI combines claim attribution to model internals and external sources with entailment-based verification to improve LLM factual reliability by 5-20% on fact-checking datasets.
-
Understanding the planning of LLM agents: A survey
A survey that provides a taxonomy of methods for improving planning in LLM-based agents across task decomposition, plan selection, external modules, reflection, and memory.
-
Hypencoder Revisited: Reproducibility and Analysis of Non-Linear Scoring for First-Stage Retrieval
Reproducibility study confirms Hypencoder's non-linear query-specific scoring improves retrieval over bi-encoders on standard benchmarks but standard methods remain faster and hard-task results are mixed due to implem...
discussion (0)
Sign in with ORCID, Apple, or X to comment. Anyone can read and Pith papers without signing in.