Towards Effective Extraction and Evaluation of Factual Claims
ACL 2025 Main Conference |
A common strategy for fact-checking long-form content generated by Large Language Models (LLMs) is extracting simple claims that can be verified independently. Since inaccurate or incomplete claims compromise fact-checking results, ensuring claim quality is critical. However, the lack of a standardized evaluation framework impedes assessment and comparison of claim extraction methods. To address this gap, we propose a framework for evaluating claim extraction in the context of fact-checking along with automated, scalable, and replicable methods for applying this framework, including novel approaches for measuring coverage and decontextualization. We also introduce Claimify, an LLM-based claim extraction method, and demonstrate that it outperforms existing methods under our evaluation framework. A key feature of Claimify is its ability to handle ambiguity and extract claims only when there is high confidence in the correct interpretation of the source text.
Claimify: Extracting high-quality claims from language model outputs
Dasha Metropolitansky, Research Data Scientist, Microsoft Research Special Projects, introduces Claimify, a new method for extracting simple, verifiable claims from LLM outputs. Claim extraction is a key step in fact-checking LLM-generated content. Claimify outperforms prior techniques, ensuring that extracted claims are accurate, verifiable, and preserve critical context.
Paper: Towards Effective Extraction and Evaluation of Factual Claims
Blog post: Claimify: Extracting high-quality claims from language model outputs