Samuel Chen


2025

pdf bib
United We Fine-Tune: Structurally Complementary Datasets for Hope Speech Detection
Priya Dharshini Krishnaraj | Tulio Ferreira Leite da Silva | Gonzalo Freijedo Aduna | Samuel Chen | Farah Benamara | Alda Mari
Proceedings of the First Workshop on Comparative Performance Evaluation: From Rules to Language Models

We propose a fine-tuning strategy for English Multi-class Hope Speech Detection using Mistral, leveraging two complementary datasets: PolyHope and CDB, a new unified framework for hope speech detection. While the former provides nuanced hope-related categories such as GENERALIZED, REALISTIC, and UNREALISTIC HOPE, the later introduces linguistically grounded dimensions including COUNTERFACTUAL, DESIRE, and BELIEF. By fine-tuning Mistral on both datasets, we enable the model to capture deeper semantic representations of hope. In addition to fine-tuning, we developed advanced prompting strategies which provide interpretable, zero-shot alternatives and further inform annotation and classification designs. Our approach achieved third place in the multi-class (Macro F1=71.77) and sixth in the binary (Macro F1=85.35) settings.