Back to Main Conference 2022
LREC 2022main

LuxemBERT: Simple and Practical Data Augmentation in Language Model Pre-Training for Luxembourgish

Proceedings of the Thirteenth International Conference on Language Resources and Evaluation (LREC 2022)

DOI:10.63317/2tmtoonjyfz8

Abstract

Pre-trained Language Models such as BERT have become ubiquitous in NLP where they have achieved state-of-the-art performance in most NLP tasks. While these models are readily available for English and other widely spoken languages, they remain scarce for low-resource languages such as Luxembourgish. In this paper, we present LuxemBERT, a BERT model for the Luxembourgish language that we create using the following approach: we augment the pre-training dataset by considering text data from a closely related language that we partially translate using a simple and straightforward method. We are then able to produce the LuxemBERT model, which we show to be effective for various NLP tasks: it outperforms a simple baseline built with the available Luxembourgish text data as well the multilingual mBERT model, which is currently the only option for transformer-based language models in Luxembourgish. Furthermore, we present datasets for various downstream NLP tasks that we created for this study and will make available to researchers on request.

Details

Paper ID
lrec2022-main-543
Pages
pp. 5080-5089
BibKey
lothritz-etal-2022-luxembert
Editor
N/A
Publisher
European Language Resources Association (ELRA)
ISSN
2522-2686
ISBN
79-10-95546-38-2
Conference
Thirteenth Language Resources and Evaluation Conference
Location
Marseille, France
Date
20 June 2022 25 June 2022

Authors

  • CL

    Cedric Lothritz

  • BL

    Bertrand Lebichot

  • KA

    Kevin Allix

  • LV

    Lisa Veiber

  • TB

    Tegawende Bissyande

  • JK

    Jacques Klein

  • AB

    Andrey Boytsov

  • CL

    Clément Lefebvre

  • AG

    Anne Goujon

Links