Elena Musi
2025
Toward Reasonable Parrots: Why Large Language Models Should Argue with Us by Design
Elena Musi | Nadin Kökciyan | Khalid Al Khatib | Davide Ceolin | Emmanuelle Dietz | Klara Maximiliane Gutekunst | Annette Hautli-Janisz | Cristián Santibáñez | Jodi Schneider | Jonas Scholz | Cor Steging | Jacky Visser | Henning Wachsmuth
Proceedings of the 12th Argument mining Workshop
Elena Musi | Nadin Kökciyan | Khalid Al Khatib | Davide Ceolin | Emmanuelle Dietz | Klara Maximiliane Gutekunst | Annette Hautli-Janisz | Cristián Santibáñez | Jodi Schneider | Jonas Scholz | Cor Steging | Jacky Visser | Henning Wachsmuth
Proceedings of the 12th Argument mining Workshop
In this position paper, we advocate for the development of conversational technology that is inherently designed to support and facilitate argumentative processes. We argue that, at present, large language models (LLMs) are inadequate for this purpose, and we propose an ideal technology design aimed at enhancing argumentative skills. This involves re-framing LLMs as tools to exercise our critical thinking skills rather than replacing them. We introduce the concept of reasonable parrots that embody the fundamental principles of relevance, responsibility, and freedom, and that interact through argumentative dialogical moves. These principles and moves arise out of millennia of work in argumentation theory and should serve as the starting point for LLM-based technology that incorporates basic principles of argumentation.
2022
Multitask Instruction-based Prompting for Fallacy Recognition
Tariq Alhindi | Tuhin Chakrabarty | Elena Musi | Smaranda Muresan
Proceedings of the 2022 Conference on Empirical Methods in Natural Language Processing
Tariq Alhindi | Tuhin Chakrabarty | Elena Musi | Smaranda Muresan
Proceedings of the 2022 Conference on Empirical Methods in Natural Language Processing
Fallacies are used as seemingly valid arguments to support a position and persuade the audience about its validity. Recognizing fallacies is an intrinsically difficult task both for humans and machines. Moreover, a big challenge for computational models lies in the fact that fallacies are formulated differently across the datasets with differences in the input format (e.g., question-answer pair, sentence with fallacy fragment), genre (e.g., social media, dialogue, news), as well as types and number of fallacies (from 5 to 18 types per dataset). To move towards solving the fallacy recognition task, we approach these differences across datasets as multiple tasks and show how instruction-based prompting in a multitask setup based on the T5 model improves the results against approaches built for a specific dataset such as T5, BERT or GPT-3. We show the ability of this multitask prompting approach to recognize 28 unique fallacies across domains and genres and study the effect of model size and prompt choice by analyzing the per-class (i.e., fallacy type) results. Finally, we analyze the effect of annotation quality on model performance, and the feasibility of complementing this approach with external knowledge.
2020
Interpreting Verbal Irony: Linguistic Strategies and the Connection to theType of Semantic Incongruity
Debanjan Ghosh | Elena Musi | Smaranda Muresan
Proceedings of the Society for Computation in Linguistics 2020
Debanjan Ghosh | Elena Musi | Smaranda Muresan
Proceedings of the Society for Computation in Linguistics 2020
2019
Rubric Reliability and Annotation of Content and Argument in Source-Based Argument Essays
Yanjun Gao | Alex Driban | Brennan Xavier McManus | Elena Musi | Patricia Davies | Smaranda Muresan | Rebecca J. Passonneau
Proceedings of the Fourteenth Workshop on Innovative Use of NLP for Building Educational Applications
Yanjun Gao | Alex Driban | Brennan Xavier McManus | Elena Musi | Patricia Davies | Smaranda Muresan | Rebecca J. Passonneau
Proceedings of the Fourteenth Workshop on Innovative Use of NLP for Building Educational Applications
We present a unique dataset of student source-based argument essays to facilitate research on the relations between content, argumentation skills, and assessment. Two classroom writing assignments were given to college students in a STEM major, accompanied by a carefully designed rubric. The paper presents a reliability study of the rubric, showing it to be highly reliable, and initial annotation on content and argumentation annotation of the essays.
2018
A Multi-layer Annotated Corpus of Argumentative Text: From Argument Schemes to Discourse Relations
Elena Musi | Tariq Alhindi | Manfred Stede | Leonard Kriese | Smaranda Muresan | Andrea Rocci
Proceedings of the Eleventh International Conference on Language Resources and Evaluation (LREC 2018)
Elena Musi | Tariq Alhindi | Manfred Stede | Leonard Kriese | Smaranda Muresan | Andrea Rocci
Proceedings of the Eleventh International Conference on Language Resources and Evaluation (LREC 2018)
ChangeMyView Through Concessions: Do Concessions Increase Persuasion?
Elena Musi | Debanjan Ghosh | Smaranda Muresan
Dialogue Discourse Volume 9
Elena Musi | Debanjan Ghosh | Smaranda Muresan
Dialogue Discourse Volume 9
In Discourse Studies concessions are considered among those argumentative strategies that increase persuasion. We aim to empirically test this hypothesis by calculating the distribution of argumentative concessions in persuasive vs. non-persuasive comments from the the ChangeMyView subreddit. This constitutes a challenging task since concessions do not always bear an argumentative role and are expressed through polysemous lexical markers. Drawing from a theoretically-informed typology of concessions, we first conduct a crowdsourcing task to label a set of polysemous lexical markers as introducing an argumentative concession relation or not. Second, we present a self-training method to automatically identify argumentative concessions using linguistically motivated features. While we achieve a moderate F1 of 57.4% via the self-training method, our subsequent error analysis highlights that the self training method is able to generalize and identify other types of concessions that are argumentative, but were not considered in the annotation guidelines. Our findings from the manual labeling and the classification experiments indicate that the type of argumentative concessions we investigated is almost equally likely to be used in winning and losing arguments. While this result seems to contradict theoretical assumptions, we provide some reasons related to the ChangeMyView subreddit.
2017
Analyzing the Semantic Types of Claims and Premises in an Online Persuasive Forum
Christopher Hidey | Elena Musi | Alyssa Hwang | Smaranda Muresan | Kathy McKeown
Proceedings of the 4th Workshop on Argument Mining
Christopher Hidey | Elena Musi | Alyssa Hwang | Smaranda Muresan | Kathy McKeown
Proceedings of the 4th Workshop on Argument Mining
Argumentative text has been analyzed both theoretically and computationally in terms of argumentative structure that consists of argument components (e.g., claims, premises) and their argumentative relations (e.g., support, attack). Less emphasis has been placed on analyzing the semantic types of argument components. We propose a two-tiered annotation scheme to label claims and premises and their semantic types in an online persuasive forum, Change My View, with the long-term goal of understanding what makes a message persuasive. Premises are annotated with the three types of persuasive modes: ethos, logos, pathos, while claims are labeled as interpretation, evaluation, agreement, or disagreement, the latter two designed to account for the dialogical nature of our corpus. We aim to answer three questions: 1) can humans reliably annotate the semantic types of argument components? 2) are types of premises/claims positioned in recurrent orders? and 3) are certain types of claims and/or premises more likely to appear in persuasive messages than in non-persuasive messages?
2016
Search
Fix author
Co-authors
- Smaranda Muresan 7
- Debanjan Ghosh 3
- Tariq Alhindi 2
- Davide Ceolin 1
- Tuhin Chakrabarty 1
- Patricia Davies 1
- Emmanuelle Dietz 1
- Alex Driban 1
- Yanjun Gao 1
- Klara Maximiliane Gutekunst 1
- Annette Hautli 1
- Christopher Hidey 1
- Alyssa Hwang 1
- Khalid Al Khatib 1
- Leonard Kriese 1
- Nadin Kökciyan 1
- Kathleen McKeown 1
- Rebecca J. Passonneau 1
- Andrea Rocci 1
- Cristián Santibáñez 1
- Jodi Schneider 1
- Jonas Scholz 1
- Manfred Stede 1
- Cor Steging 1
- Jacky Visser 1
- Henning Wachsmuth 1
- Brennan Xavier McManus 1