Back to Main Conference 2024
LREC-COLING 2024main

Improving Cross-lingual Transfer with Contrastive Negative Learning and Self-training

Proceedings of the 2024 Joint International Conference on Computational Linguistics, Language Resources and Evaluation (LREC-COLING 2024)

DOI:10.63317/2jac3495z7j2

Abstract

Recent studies improve the cross-lingual transfer learning by better aligning the internal representations within the multilingual model or exploring the information of the target language using self-training. However, the alignment-based methods exhibit intrinsic limitations such as non-transferable linguistic elements, while most of the self-training based methods ignore the useful information hidden in the low-confidence samples. To address this issue, we propose CoNLST (Contrastive Negative Learning and Self-Training) to leverage the information of low-confidence samples. Specifically, we extend the negative learning to the metric space by selecting negative pairs based on the complementary labels and then employ self-training to iteratively train the model to converge on the obtained clean pseudo-labels. We evaluate our approach on the widely-adopted cross-lingual benchmark XNLI. The experiment results show that our method improves upon the baseline models and can serve as a beneficial complement to the alignment-based methods.

Details

Paper ID
lrec2024-main-0769
Pages
pp. 8781-8791
BibKey
li-etal-2024-improving-cross-lingual
Editor
N/A
Publisher
European Language Resources Association (ELRA) and ICCL
ISSN
2522-2686
ISBN
979-10-95546-34-4
Conference
Joint International Conference on Computational Linguistics, Language Resources and Evaluation
Location
Turin, Italy
Date
20 May 2024 25 May 2024

Authors

  • GL

    Guanlin Li

  • XZ

    Xuechen Zhao

  • AJ

    Amir Jafari

  • WS

    Wenhao Shao

  • RF

    Reza Farahbakhsh

  • NC

    Noel Crespi

Links