Back to WILDRE 2024
LREC-COLING 2024workshop

FZZG at WILDRE-7: Fine-tuning Pre-trained Models for Code-mixed, Less-resourced Sentiment Analysis

Proceedings of the 7th Workshop on Indian Language Data: Resources and Evaluation

DOI:10.63317/2tfssxkcostk

Abstract

This paper describes our system used for a shared task on code-mixed, less-resourced sentiment analysis for Indo-Aryan languages. We are using the large language models (LLMs) since they have demonstrated excellent performance on classification tasks. In our participation in all tracks, we use unsloth/mistral-7b-bnb-4bit LLM for the task of code-mixed sentiment analysis. For track 1, we used a simple fine-tuning strategy on PLMs by combining data from multiple phases. Our trained systems secured first place in four phases out of five. In addition, we present the results achieved using several PLMs for each language.

Details

Paper ID
lrec2024-ws-wildre-09
Pages
pp. 59-65
BibKey
thakkar-etal-2024-fzzg
Editor
N/A
Publisher
European Language Resources Association (ELRA) and ICCL
ISSN
N/A
ISBN
N/A
Workshop
Proceedings of the 7th Workshop on Indian Language Data: Resources and Evaluation
Location
undefined, undefined
Date
20 May 2024 25 May 2024

Authors

  • GT

    Gaurish Thakkar

  • MT

    Marko Tadić

  • NM

    Nives Mikelic Preradovic

Links