Fernando Gonzalez Adauto
2024
Do LLMs Think Fast and Slow? A Causal Study on Sentiment Analysis
Zhiheng Lyu
|
Zhijing Jin
|
Fernando Gonzalez Adauto
|
Rada Mihalcea
|
Bernhard Schölkopf
|
Mrinmaya Sachan
Findings of the Association for Computational Linguistics: EMNLP 2024
Sentiment analysis (SA) aims to identify the sentiment expressed in a piece of text, often in the form of a review. Assuming a review and the sentiment associated with it, in this paper we formulate SA as a combination of two tasks: (1) a causal discovery task that distinguishes whether a review “primes” the sentiment (Causal Hypothesis C1), or the sentiment “primes” the review (Causal Hypothesis C2); and (2) the traditional prediction task to model the sentiment using the review as input. Using the peak-end rule in psychology, we classify a sample as C1 if its overall sentiment score approximates an average of all the sentence-level sentiments in the review, and as C2 if the overall sentiment score approximates an average of the peak and end sentiments. For the prediction task, we use the discovered causal mechanisms behind the samples to improve the performance of LLMs by proposing causal prompts that give the models an inductive bias of the underlying causal graph, leading to substantial improvements by up to 32.13 F1 points on zero-shot five-class SA.
Analyzing the Role of Semantic Representations in the Era of Large Language Models
Zhijing Jin
|
Yuen Chen
|
Fernando Gonzalez Adauto
|
Jiarui Liu
|
Jiayi Zhang
|
Julian Michael
|
Bernhard Schölkopf
|
Mona Diab
Proceedings of the 2024 Conference of the North American Chapter of the Association for Computational Linguistics: Human Language Technologies (Volume 1: Long Papers)
Traditionally, natural language processing (NLP) models often use a rich set of features created by linguistic expertise, such as semantic representations. However, in the era of large language models (LLMs), more and more tasks are turned into generic, end-to-end sequence generation problems. In this paper, we investigate the question: what is the role of semantic representations in the era of LLMs? Specifically, we investigate the effect of Abstract Meaning Representation (AMR) across five diverse NLP tasks. We propose an AMR-driven chain-of-thought prompting method, which we call AMRCOT, and find that it generally hurts performance more than it helps. To investigate what AMR may have to offer on these tasks, we conduct a series of analysis experiments. We find that it is difficult to predict which input examples AMR may help or hurt on, but errors tend to arise with multi-word expressions, named entities, and in the final inference step where the LLM must connect its reasoning over the AMR to its prediction. We recommend focusing on these areas for future work in semantic representations for LLMs. Our code: https://github.com/causalNLP/amr_llm
Search
Co-authors
- Zhijing Jin 2
- Bernhard Schölkopf 2
- Zhiheng Lyu 1
- Rada Mihalcea 1
- Mrinmaya Sachan 1
- show all...