Back to Main Conference 2026
LREC 2026main

Format Matters: A Critical Evaluation of Output Formats for Prompting LLMs in SLU and NER

Proceedings of the Fifteenth Language Resources and Evaluation Conference (LREC 2026)

DOI:10.63317/3osjjdr778fh

Abstract

Output format is often an unreported factor in LLM evaluations for structured NLP tasks such as Slot Filling or Named Entity Recognition. This work proposes to explore the impact of the output structured format generated by LLMs. We show that measured performance and reliability depend on the requested format (JSON, XML or inline Key-Values). A study is performed across four SLU and three NER benchmarks and considering 13 instruction-tuned open-weight LLMs, using standardized and open-source prompts and parsers. This format-specific evaluation reveals statistically significant swings of 2-46 F1 points depending on model and dataset. Additionally, we propose a lightweight selection procedure to determine the best format per model-dataset combination using only a small development slice; thus reducing trial-and-error in practice.

Details

Paper ID
lrec2026-main-593
Pages
pp. 7485-7497
BibKey
lepagnol-etal-2026-format
Editor
N/A
Publisher
European Language Resources Association (ELRA)
ISSN
2522-2686
ISBN
978-2-493814-49-4
Conference
The Fifteenth Language Resources and Evaluation Conference (LREC 2026)
Location
Palma, Mallorca, Spain
Date
11 May 2026 16 May 2026

Authors

  • PL

    Pierre Lepagnol

  • SG

    Sahar Ghannay

  • TG

    Thomas Gerald

  • CS

    Christophe Servan

  • SR

    Sophie Rosset

Links