Back to Main Conference 2026
LREC 2026main

SciClaimEval: Cross-modal Claim Verification in Scientific Papers

Proceedings of the Fifteenth Language Resources and Evaluation Conference (LREC 2026)

DOI:10.63317/4ap9rg2gnwmf

Abstract

We present SciClaimEval, a new scientific dataset for the claim verification task. Unlike existing resources, SciClaimEval features authentic claims, including refuted ones, directly extracted from published papers. To create refuted claims, we introduce a novel approach that modifies the supporting evidence (figures and tables), rather than altering the claims or relying on large language models (LLMs) to fabricate contradictions. The dataset provides cross-modal evidence with diverse representations: figures are available as images, while tables are provided in multiple formats, including images, LaTeX source, HTML, and JSON. SciClaimEval contains 1,664 annotated samples from 180 papers across three domains, machine learning, natural language processing, and medicine, validated through expert annotation. We benchmark 11 multimodal foundation models, both open-source and proprietary, across the dataset. Results show that figure-based verification remains particularly challenging for all models, as a substantial performance gap remains between the best system and human baseline.

Details

Paper ID
lrec2026-main-864
Pages
pp. 11060-11071
BibKey
ho-etal-2026-sciclaimeval
Editor
N/A
Publisher
European Language Resources Association (ELRA)
ISSN
2522-2686
ISBN
978-2-493814-49-4
Conference
The Fifteenth Language Resources and Evaluation Conference (LREC 2026)
Location
Palma, Mallorca, Spain
Date
11 May 2026 16 May 2026

Authors

  • XH

    Xanh Ho

  • YW

    Yun-Ang Wu

  • SK

    Sunisth Kumar

  • TX

    Tian Cheng Xia

  • FB

    Florian Boudin

  • AG

    Andre Greiner-Petter

  • AA

    Akiko Aizawa

Links