Elena Filatova
2026
The Impact of Highlighting Subjective Language on Perceived News Trustworthiness
Mohammad Shokri | Vivek Sharma | Emily Klapper | Shweta Jain | Elena Filatova | Sarah Ita Levitan
The Proceedings for the 15th Workshop on Computational Approaches to Subjectivity, Sentiment Social Media Analysis (WASSA 2026)
Mohammad Shokri | Vivek Sharma | Emily Klapper | Shweta Jain | Elena Filatova | Sarah Ita Levitan
The Proceedings for the 15th Workshop on Computational Approaches to Subjectivity, Sentiment Social Media Analysis (WASSA 2026)
The rise of misinformation and opinionated articles has made understanding how misleading or biased content influences readers an increasingly important problem. While most prior work focuses on detecting misinformation or deceptive language in real time, far less attention has been paid to how such content is perceived by readers, which is an essential component of misinformation’s effectiveness. In this study, we examine whether highlighting subjective sentences in news articles affects perceived trustworthiness. Using a controlled user experiment and 1,334 article–reader evaluations, we find that highlighting subjective content produces a modest yet statistically significant decrease in trust, with substantial variation across articles and participants. To explain this variation, we model trust change after highlighting subjective language as a function of article-level linguistic features and reader-level attitudes. Our findings suggest that readers’ reactions to highlighted subjective language are driven primarily by characteristics of the text itself, and that highlighting subjective language offers benefits for may help readers better assess the reliability of potentially misleading news articles.
Council of LLMs: Evaluating Capability of Large Language Models to Annotate Propaganda
Vivek Sharma | Shweta Jain | Mohammad Shokri | Sarah Ita Levitan | Elena Filatova
The Proceedings for the 15th Workshop on Computational Approaches to Subjectivity, Sentiment Social Media Analysis (WASSA 2026)
Vivek Sharma | Shweta Jain | Mohammad Shokri | Sarah Ita Levitan | Elena Filatova
The Proceedings for the 15th Workshop on Computational Approaches to Subjectivity, Sentiment Social Media Analysis (WASSA 2026)
Data annotation is essential for supervised natural language processing tasks but remains labor-intensive and expensive. Large language models (LLMs) have emerged as promising alternatives, capable of generating high-quality annotations either autonomously or in collaboration with human annotators. However their use in autonomous annotations is often questioned for their ethical take on subjective matters. This study investigates the effectiveness of LLMs in a autonomous, and hybrid annotation setups in propaganda detection. We evaluate GPT and open-source models on two datasets from different domains, namely, Propaganda Techniques Corpus (PTC) for news articles and the Journalist Media Bias on X (JMBX) for social media. Our results show that LLMs, in general, exhibit high recall but lower precision in detecting propaganda, often over-predicting persuasive content. Multi-annotator setups did not outperform the best models in single-annotator setting although it helped reasoning models boost their performance. Hybrid annotation, combining LLMs and human input, achieved the highest overall accuracy than LLM-only settings. We further analyze misclassifications and found that LLM have higher sensitivity towards certain propaganda techniques like loaded language, name calling, and doubt. Finally, using error typology analysis, we explore the reasoning provided on misclassifications by the LLM. Our result shows that although some studies report LLM outperforming manual annotations and it could prove useful in hybrid annotation, its incorporation in the human annotation pipeline must be implemented with caution.
2025
Can LLMs Generate and Solve Linguistic Olympiad Puzzles?
Neh Majmudar | Elena Filatova
Proceedings of the 2025 Conference on Empirical Methods in Natural Language Processing
Neh Majmudar | Elena Filatova
Proceedings of the 2025 Conference on Empirical Methods in Natural Language Processing
In this paper, we introduce a combination of novel and exciting tasks: the solution and generation of linguistic puzzles. We focus on puzzles used in Linguistic Olympiads for high school students. We first extend the existing benchmark for the task of solving linguistic puzzles. We explore the use of Large Language Models (LLMs), including recent state-of-the-art models such as OpenAI’s o1, for solving linguistic puzzles, analyzing their performance across various linguistic topics. We demonstrate that LLMs outperform humans on most puzzles types, except for those centered on writing systems, and for the understudied languages. We use the insights from puzzle-solving experiments to direct the novel task of puzzle generation. We believe that automating puzzle generation, even for relatively simple puzzles, holds promise for expanding interest in linguistics and introducing the field to a broader audience. This finding highlights the importance of linguistic puzzle generation as a research task: such puzzles can not only promote linguistics but also support the dissemination of knowledge about rare and understudied languages.
2024
Subjectivity Detection in English News using Large Language Models
Mohammad Shokri | Vivek Sharma | Elena Filatova | Shweta Jain | Sarah Levitan
Proceedings of the 14th Workshop on Computational Approaches to Subjectivity, Sentiment, & Social Media Analysis
Mohammad Shokri | Vivek Sharma | Elena Filatova | Shweta Jain | Sarah Levitan
Proceedings of the 14th Workshop on Computational Approaches to Subjectivity, Sentiment, & Social Media Analysis
Trust in media has reached a historical low as consumers increasingly doubt the credibility of the news they encounter. This growing skepticism is exacerbated by the prevalence of opinion-driven articles, which can influence readers’ beliefs to align with the authors’ viewpoints. In response to this trend, this study examines the expression of opinions in news by detecting subjective and objective language. We conduct an analysis of the subjectivity present in various news datasets and evaluate how different language models detect subjectivity and generalize to out-of-distribution data. We also investigate the use of in-context learning (ICL) within large language models (LLMs) and propose a straightforward prompting method that outperforms standard ICL and chain-of-thought (CoT) prompts.
2012
Irony and Sarcasm: Corpus Generation and Analysis Using Crowdsourcing
Elena Filatova
Proceedings of the Eighth International Conference on Language Resources and Evaluation (LREC'12)
Elena Filatova
Proceedings of the Eighth International Conference on Language Resources and Evaluation (LREC'12)
The ability to reliably identify sarcasm and irony in text can improve the performance of many Natural Language Processing (NLP) systems including summarization, sentiment analysis, etc. The existing sarcasm detection systems have focused on identifying sarcasm on a sentence level or for a specific phrase. However, often it is impossible to identify a sentence containing sarcasm without knowing the context. In this paper we describe a corpus generation experiment where we collect regular and sarcastic Amazon product reviews. We perform qualitative and quantitative analysis of the corpus. The resulting corpus can be used for identifying sarcasm on two levels: a document and a text utterance (where a text utterance can be as short as a sentence and as long as a whole document).
2010
Rethinking Grammatical Error Annotation and Evaluation with the Amazon Mechanical Turk
Joel Tetreault | Elena Filatova | Martin Chodorow
Proceedings of the NAACL HLT 2010 Fifth Workshop on Innovative Use of NLP for Building Educational Applications
Joel Tetreault | Elena Filatova | Martin Chodorow
Proceedings of the NAACL HLT 2010 Fifth Workshop on Innovative Use of NLP for Building Educational Applications
2009
Directions for Exploiting Asymmetries in Multilingual Wikipedia
Elena Filatova
Proceedings of the Third International Workshop on Cross Lingual Information Access: Addressing the Information Need of Multilingual Societies (CLIAWS3)
Elena Filatova
Proceedings of the Third International Workshop on Cross Lingual Information Access: Addressing the Information Need of Multilingual Societies (CLIAWS3)
2008
An Unsupervised Approach to Biography Production Using Wikipedia
Fadi Biadsy | Julia Hirschberg | Elena Filatova
Proceedings of ACL-08: HLT
Fadi Biadsy | Julia Hirschberg | Elena Filatova
Proceedings of ACL-08: HLT
2006
Automatic Creation of Domain Templates
Elena Filatova | Vasileios Hatzivassiloglou | Kathleen McKeown
Proceedings of the COLING/ACL 2006 Main Conference Poster Sessions
Elena Filatova | Vasileios Hatzivassiloglou | Kathleen McKeown
Proceedings of the COLING/ACL 2006 Main Conference Poster Sessions
2005
Tell Me What You Do and I’ll Tell You What You Are: Learning Occupation-Related Activities for Biographies
Elena Filatova | John Prager
Proceedings of Human Language Technology Conference and Conference on Empirical Methods in Natural Language Processing
Elena Filatova | John Prager
Proceedings of Human Language Technology Conference and Conference on Empirical Methods in Natural Language Processing
2004
A Formal Model for Information Selection in Multi-Sentence Text Extraction
Elena Filatova | Vasileios Hatzivassiloglou
COLING 2004: Proceedings of the 20th International Conference on Computational Linguistics
Elena Filatova | Vasileios Hatzivassiloglou
COLING 2004: Proceedings of the 20th International Conference on Computational Linguistics
Event-Based Extractive Summarization
Elena Filatova | Vasileios Hatzivassiloglou
Text Summarization Branches Out
Elena Filatova | Vasileios Hatzivassiloglou
Text Summarization Branches Out