Back to Main Conference 2024
LREC-COLING 2024main

How to Encode Domain Information in Relation Classification

Proceedings of the 2024 Joint International Conference on Computational Linguistics, Language Resources and Evaluation (LREC-COLING 2024)

DOI:10.63317/2o54weekh99y

Abstract

Current language models require a lot of training data to obtain high performance. For Relation Classification (RC), many datasets are domain-specific, so combining datasets to obtain better performance is non-trivial. We explore a multi-domain training setup for RC, and attempt to improve performance by encoding domain information. Our proposed models improve > 2 Macro-F1 against the baseline setup, and our analysis reveals that not all the labels benefit the same: The classes which occupy a similar space across domains (i.e., their interpretation is close across them, for example “physical”) benefit the least, while domain-dependent relations (e.g., “part-of”) improve the most when encoding domain information.

Details

Paper ID
lrec2024-main-0728
Pages
pp. 8301-8306
BibKey
bassignana-etal-2024-encode
Editor
N/A
Publisher
European Language Resources Association (ELRA) and ICCL
ISSN
2522-2686
ISBN
979-10-95546-34-4
Conference
Joint International Conference on Computational Linguistics, Language Resources and Evaluation
Location
Turin, Italy
Date
20 May 2024 25 May 2024

Authors

  • EB

    Elisa Bassignana

  • VG

    Viggo Unmack Gascou

  • FL

    Frida Nøhr Laustsen

  • GK

    Gustav Kristensen

  • MP

    Marie Haahr Petersen

  • Rv

    Rob van der Goot

  • BP

    Barbara Plank

Links