@inproceedings{cercas-curry-cercas-curry-2023-computer,
title = "Computer says {``}No{''}: The Case Against Empathetic Conversational {AI}",
author = "Cercas Curry, Alba and
Cercas Curry, Amanda",
editor = "Rogers, Anna and
Boyd-Graber, Jordan and
Okazaki, Naoaki",
booktitle = "Findings of the Association for Computational Linguistics: ACL 2023",
month = jul,
year = "2023",
address = "Toronto, Canada",
publisher = "Association for Computational Linguistics",
url = "https://aclanthology.org/2023.findings-acl.515",
doi = "10.18653/v1/2023.findings-acl.515",
pages = "8123--8130",
abstract = "Emotions are an integral part of human cognition and they guide not only our understanding of the world but also our actions within it. As such, whether we soothe or flame an emotion is not inconsequential. Recent work in conversational AI has focused on responding empathetically to users, validating and soothing their emotions without a real basis. This AI-aided emotional regulation can have negative consequences for users and society, tending towards a one-noted happiness defined as only the absence of {``}negative{''} emotions. We argue that we must carefully consider whether and how to respond to users{'} emotions.",
}
<?xml version="1.0" encoding="UTF-8"?>
<modsCollection xmlns="http://www.loc.gov/mods/v3">
<mods ID="cercas-curry-cercas-curry-2023-computer">
<titleInfo>
<title>Computer says “No”: The Case Against Empathetic Conversational AI</title>
</titleInfo>
<name type="personal">
<namePart type="given">Alba</namePart>
<namePart type="family">Cercas Curry</namePart>
<role>
<roleTerm authority="marcrelator" type="text">author</roleTerm>
</role>
</name>
<name type="personal">
<namePart type="given">Amanda</namePart>
<namePart type="family">Cercas Curry</namePart>
<role>
<roleTerm authority="marcrelator" type="text">author</roleTerm>
</role>
</name>
<originInfo>
<dateIssued>2023-07</dateIssued>
</originInfo>
<typeOfResource>text</typeOfResource>
<relatedItem type="host">
<titleInfo>
<title>Findings of the Association for Computational Linguistics: ACL 2023</title>
</titleInfo>
<name type="personal">
<namePart type="given">Anna</namePart>
<namePart type="family">Rogers</namePart>
<role>
<roleTerm authority="marcrelator" type="text">editor</roleTerm>
</role>
</name>
<name type="personal">
<namePart type="given">Jordan</namePart>
<namePart type="family">Boyd-Graber</namePart>
<role>
<roleTerm authority="marcrelator" type="text">editor</roleTerm>
</role>
</name>
<name type="personal">
<namePart type="given">Naoaki</namePart>
<namePart type="family">Okazaki</namePart>
<role>
<roleTerm authority="marcrelator" type="text">editor</roleTerm>
</role>
</name>
<originInfo>
<publisher>Association for Computational Linguistics</publisher>
<place>
<placeTerm type="text">Toronto, Canada</placeTerm>
</place>
</originInfo>
<genre authority="marcgt">conference publication</genre>
</relatedItem>
<abstract>Emotions are an integral part of human cognition and they guide not only our understanding of the world but also our actions within it. As such, whether we soothe or flame an emotion is not inconsequential. Recent work in conversational AI has focused on responding empathetically to users, validating and soothing their emotions without a real basis. This AI-aided emotional regulation can have negative consequences for users and society, tending towards a one-noted happiness defined as only the absence of “negative” emotions. We argue that we must carefully consider whether and how to respond to users’ emotions.</abstract>
<identifier type="citekey">cercas-curry-cercas-curry-2023-computer</identifier>
<identifier type="doi">10.18653/v1/2023.findings-acl.515</identifier>
<location>
<url>https://aclanthology.org/2023.findings-acl.515</url>
</location>
<part>
<date>2023-07</date>
<extent unit="page">
<start>8123</start>
<end>8130</end>
</extent>
</part>
</mods>
</modsCollection>
%0 Conference Proceedings
%T Computer says “No”: The Case Against Empathetic Conversational AI
%A Cercas Curry, Alba
%A Cercas Curry, Amanda
%Y Rogers, Anna
%Y Boyd-Graber, Jordan
%Y Okazaki, Naoaki
%S Findings of the Association for Computational Linguistics: ACL 2023
%D 2023
%8 July
%I Association for Computational Linguistics
%C Toronto, Canada
%F cercas-curry-cercas-curry-2023-computer
%X Emotions are an integral part of human cognition and they guide not only our understanding of the world but also our actions within it. As such, whether we soothe or flame an emotion is not inconsequential. Recent work in conversational AI has focused on responding empathetically to users, validating and soothing their emotions without a real basis. This AI-aided emotional regulation can have negative consequences for users and society, tending towards a one-noted happiness defined as only the absence of “negative” emotions. We argue that we must carefully consider whether and how to respond to users’ emotions.
%R 10.18653/v1/2023.findings-acl.515
%U https://aclanthology.org/2023.findings-acl.515
%U https://doi.org/10.18653/v1/2023.findings-acl.515
%P 8123-8130
Markdown (Informal)
[Computer says “No”: The Case Against Empathetic Conversational AI](https://aclanthology.org/2023.findings-acl.515) (Cercas Curry & Cercas Curry, Findings 2023)
ACL