Back to Main Conference 2026
LREC 2026main

Bulgarian Massive Multitask Language Understanding Benchmark

Proceedings of the Fifteenth Language Resources and Evaluation Conference (LREC 2026)

DOI:10.63317/57jy4fetjmc4

Abstract

Assessing the broad general knowledge of Large Language Models (LLMs) across multiple domains in Bulgarian remains challenging due to the limited availability of Bulgarian evaluation benchmarks. To address this gap, we introduce the Bulgarian Massive Multitask Language Understanding benchmark (MMLU-BG), designed to evaluate whether LLMs possess generalised knowledge capabilities beyond simple text prediction in Bulgarian. This paper presents the structure, the development protocol, and the size of the MMLU-BG benchmark. It is tested in comparison with the original MMLU for English across seven LLMs selected according to specific criteria. The experiments demonstrate that the MMLU-BG benchmark assesses multi-domain versatility and highlights the models’ strengths and weaknesses across different subject areas.

Details

Paper ID
lrec2026-main-366
Pages
pp. 4658-4672
BibKey
koeva-etal-2026-bulgarian
Editor
N/A
Publisher
European Language Resources Association (ELRA)
ISSN
2522-2686
ISBN
978-2-493814-49-4
Conference
The Fifteenth Language Resources and Evaluation Conference (LREC 2026)
Location
Palma, Mallorca, Spain
Date
11 May 2026 16 May 2026

Authors

  • SK

    Svetla Peneva Koeva

  • IS

    Ivelina Stoyanova

  • DG

    Dimiter Georgiev

  • SL

    Svetlozara Leseva

  • VS

    Valentina Stefanova

  • MT

    Maria Todorova

  • TD

    Tsvetana Ivanova Dimitrova

  • HK

    Hristina Kukova

  • MM

    Mihaela Moskova

  • TT

    Tinko Tinchev

Links