Back to Main Conference 2022
LREC 2022main

Evaluation of Transfer Learning for Polish with a Text-to-Text Model

Proceedings of the Thirteenth International Conference on Language Resources and Evaluation (LREC 2022)

DOI:10.63317/3nw9bnnbis3i

Abstract

We introduce a new benchmark for assessing the quality of text-to-text models for Polish. The benchmark consists of diverse tasks and datasets: KLEJ benchmark adapted for text-to-text, en-pl translation, summarization, and question answering. In particular, since summarization and question answering lack benchmark datasets for the Polish language, we describe in detail their construction and make them publicly available. Additionally, we present plT5 - a general-purpose text-to-text model for Polish that can be fine-tuned on various Natural Language Processing (NLP) tasks with a single training objective. Unsupervised denoising pre-training is performed efficiently by initializing the model weights with a multi-lingual T5 (mT5) counterpart. We evaluate the performance of plT5, mT5, Polish BART (plBART), and Polish GPT-2 (papuGaPT2). The plT5 scores top on all of these tasks except summarization, where plBART is best. In general (except summarization), the larger the model, the better the results. The encoder-decoder architectures prove to be better than the decoder-only equivalent.

Details

Paper ID
lrec2022-main-466
Pages
pp. 4374-4394
BibKey
chrabrowa-etal-2022-evaluation
Editor
N/A
Publisher
European Language Resources Association (ELRA)
ISSN
2522-2686
ISBN
79-10-95546-38-2
Conference
Thirteenth Language Resources and Evaluation Conference
Location
Marseille, France
Date
20 June 2022 25 June 2022

Authors

  • AC

    Aleksandra Chrabrowa

  • ŁD

    Łukasz Dragan

  • KG

    Karol Grzegorczyk

  • DK

    Dariusz Kajtoch

  • MK

    Mikołaj Koszowski

  • RM

    Robert Mroczkowski

  • PR

    Piotr Rybak

Links