Juanwen Pan
2025
Enhancing LLM-Based Persuasion Simulations with Cultural and Speaker-Specific Information
Weicheng Ma
|
Hefan Zhang
|
Shiyu Ji
|
Farnoosh Hashemi
|
Qichao Wang
|
Ivory Yang
|
Joice Chen
|
Juanwen Pan
|
Michael Macy
|
Saeed Hassanpour
|
Soroush Vosoughi
Findings of the Association for Computational Linguistics: EMNLP 2025
Large language models (LLMs) have been used to synthesize persuasive dialogues for studying persuasive behavior. However, existing approaches often suffer from issues such as stance oscillation and low informativeness. To address these challenges, we propose reinforced instructional prompting, a method that ensures speaker characteristics consistently guide all stages of dialogue generation. We further introduce multilingual prompting, which aligns language use with speakers’ native languages to better capture cultural nuances. Our experiments involving speakers from eight countries show that continually reinforcing speaker profiles and cultural context improves argument diversity, enhances informativeness, and stabilizes speaker stances. Moreover, our analysis of inter-group versus intra-group persuasion reveals that speakers engaging within their own cultural groups employ more varied persuasive strategies than in cross-cultural interactions. These findings underscore the importance of speaker and cultural awareness in LLM-based persuasion modeling and suggest new directions for developing more personalized, ethically grounded, and culturally adaptive LLM-generated dialogues.
A Generalizable Rhetorical Strategy Annotation Model Using LLM-based Debate Simulation and Labelling
Shiyu Ji
|
Farnoosh Hashemi
|
Joice Chen
|
Juanwen Pan
|
Weicheng Ma
|
Hefan Zhang
|
Sophia Pan
|
Ming Cheng
|
Shubham Mohole
|
Saeed Hassanpour
|
Soroush Vosoughi
|
Michael Macy
Findings of the Association for Computational Linguistics: EMNLP 2025
Rhetorical strategies are central to persuasive communication, from political discourse and marketing to legal argumentation. However, analysis of rhetorical strategies has been limited by reliance on human annotation, which is costly, inconsistent, difficult to scale. Their associated datasets are often limited to specific topics and strategies, posing challenges for robust model development. We propose a novel framework that leverages large language models (LLMs) to automatically generate and label synthetic debate data based on a four-part rhetorical typology (causal, empirical, emotional, moral). We fine-tune transformer-based classifiers on this LLM-labeled dataset and validate its performance against human-labeled data on this dataset and on multiple external corpora. Our model achieves high performance and strong generalization across topical domains. We illustrate two applications with the fine-tuned model: (1) the improvement in persuasiveness prediction from incorporating rhetorical strategy labels, and (2) analyzing temporal and partisan shifts in rhetorical strategies in U.S. Presidential debates (1960–2020), revealing increased use of affective over cognitive argument in U.S. Presidential debates.
Search
Fix author
Co-authors
- Joice Chen 2
- Farnoosh Hashemi 2
- Saeed Hassanpour 2
- Shiyu Ji 2
- Weicheng Ma 2
- show all...