Back to Main Conference 2026
LREC 2026main

ObfusQAte: A Proposed Framework to Evaluate LLM Robustness on Obfuscated Factual Question Answering

Proceedings of the Fifteenth Language Resources and Evaluation Conference (LREC 2026)

DOI:10.63317/4bcqprdhjoxv

Abstract

The rapid proliferation of Large Language Models (LLMs) has significantly contributed to the development of equitable AI systems capable of factual question-answering (QA). However, no known study tests the LLMs’ robustness when presented with obfuscated versions of questions. To systematically evaluate these limitations, we propose a novel technique, ObfusQAte and leveraging the same, introduce ObfusQA, a comprehensive, first of its kind, framework, with multi-tiered obfuscation levels designed to examine LLM capabilities across three distinct dimensions: (i) Named-Entity Indirection, (ii) Distractor Indirection, and (iii) Contextual Overload. By capturing these fine-grained distinctions in language, ObfusQA provides a comprehensive benchmark for evaluating LLM robustness and adaptability. Our study observes that LLMs exhibit a tendency to fail or generate hallucinated responses, when confronted with these increasingly nuanced variations. To foster research in this direction, we make ObfusQAte publicly available.

Details

Paper ID
lrec2026-main-401
Pages
pp. 5129-5145
BibKey
ghosh-etal-2026-obfusqate
Editor
N/A
Publisher
European Language Resources Association (ELRA)
ISSN
2522-2686
ISBN
978-2-493814-49-4
Conference
The Fifteenth Language Resources and Evaluation Conference (LREC 2026)
Location
Palma, Mallorca, Spain
Date
11 May 2026 16 May 2026

Authors

  • SG

    Shubhra Ghosh

  • AB

    Abhilekh Borah

  • AG

    Aditya Kumar Guru

  • KG

    Kripabandhu Ghosh

Links