Transferring sentiment cross-lingually within and across same-family languages

izvorni znanstveni rad

izvorni znanstveni rad

Transferring sentiment cross-lingually within and across same-family languages

Vrsta prilog u časopisu
Tip izvorni znanstveni rad
Godina 2024
Časopis Applied sciences (Basel)
Volumen 14
Svesčić 13, 5652
Stranice str. 1-21
DOI 10.3390/app14135652
EISSN 2076-3417
Status objavljeno

Sažetak

Natural language processing for languages with limited resources is hampered by a lack of data. Using English as a hub language for such languages, cross-lingual sentiment analysis has been developed. The sheer quantity of English language resources raises questions about its status as the primary resource. This research aims to examine the impact on sentiment analysis of adding data from same-family versus distant-family languages. We analyze the performance using low-resource and high-resource data from the same language family (Slavic), investigate the effect of using a distant-family language (English) and report the results for both settings. Quantitative experiments using multi-task learning demonstrate that adding a large quantity of data from related and distant-family languages is advantageous for cross-lingual sentiment transfer.

Ključne riječi

sentiment analysis; language models; transfer learning