Tehničko veleučilište u Zagrebu · Zagreb

FZZG at WILDRE-7 : fine-tuning pre-trained models for code-mixed, less-resourced sentiment analysis

izvorni znanstveni rad

izvorni znanstveni rad

FZZG at WILDRE-7 : fine-tuning pre-trained models for code-mixed, less-resourced sentiment analysis

Vrsta prilog sa skupa (u zborniku)
Tip izvorni znanstveni rad
Godina 2024
Nadređena publikacija Proceedings of the 7th Workshop on Indian Language Data: Resource and Evaluation @LREC-COLING-2024 (WILDRE-7)
Stranice str. 59-65
ISSN 2951-2093
EISSN 2522-2686
Status objavljeno

Sažetak

This paper describes our system used for a shared task on code-mixed, less-resourced sentiment analysis for Indo-Aryan languages. We are using the large language models (LLMs) since they have demonstrated excellent performance on classification tasks. In our participation in all tracks, we use unsloth/mistral-7b-bnb-4bit LLM for the task of code-mixed sentiment analysis. For track 1, we used a simple fine-tuning strategy on PLMs by combining data from multiple phases. Our trained systems secured first place in four phases out of five. In addition, we present the results achieved using several PLMs for each language.

Ključne riječi

sentiment analysis; code-mixed; LLM; Indo-Aryan