2024
pdf
bib
abs
The Illusion of Competence: Evaluating the Effect of Explanations on Users’ Mental Models of Visual Question Answering Systems
Judith Sieker
|
Simeon Junker
|
Ronja Utescher
|
Nazia Attari
|
Heiko Wersing
|
Hendrik Buschmeier
|
Sina Zarrieß
Proceedings of the 2024 Conference on Empirical Methods in Natural Language Processing
We examine how users perceive the limitations of an AI system when it encounters a task that it cannot perform perfectly and whether providing explanations alongside its answers aids users in constructing an appropriate mental model of the system’s capabilities and limitations. We employ a visual question answer and explanation task where we control the AI system’s limitations by manipulating the visual inputs: during inference, the system either processes full-color or grayscale images. Our goal is to determine whether participants can perceive the limitations of the system. We hypothesize that explanations will make limited AI capabilities more transparent to users. However, our results show that explanations do not have this effect. Instead of allowing users to more accurately assess the limitations of the AI system, explanations generally increase users’ perceptions of the system’s competence – regardless of its actual performance.
pdf
bib
Revisiting the Phenomenon of Syntactic Complexity Convergence on German Dialogue Data
Yu Wang
|
Hendrik Buschmeier
Proceedings of the 20th Conference on Natural Language Processing (KONVENS 2024)
pdf
bib
abs
What Can I Do with this Data Point? Towards Modeling Legal and Ethical Aspects of Linguistic Data Collection and (Re-)use
Annett Jorschick
|
Paul T. Schrader
|
Hendrik Buschmeier
Proceedings of the Workshop on Legal and Ethical Issues in Human Language Technologies @ LREC-COLING 2024
Linguistic data often inherits characteristics that limit open science practices such as data publication, sharing, and reuse. Part of the problem is researchers’ uncertainty about the legal requirements, which need to be considered at the beginning of study planning, when consent forms for participants, ethics applications, and data management plans need to be written. This paper presents a newly funded project that will develop a research data management infrastructure that will provide automated support to researchers in the planning, collection, storage, use, reuse, and sharing of data, taking into account ethical and legal aspects to encourage open science practices.
pdf
bib
abs
How Much Does Nonverbal Communication Conform to Entropy Rate Constancy?: A Case Study on Listener Gaze in Interaction
Yu Wang
|
Yang Xu
|
Gabriel Skantze
|
Hendrik Buschmeier
Findings of the Association for Computational Linguistics: ACL 2024
According to the Entropy Rate Constancy (ERC) principle, the information density of a text is approximately constant over its length. Whether this principle also applies to nonverbal communication signals is still under investigation. We perform empirical analyses of video-recorded dialogue data and investigate whether listener gaze, as an important nonverbal communication signal, adheres to the ERC principle. Results show (1) that the ERC principle holds for listener gaze; and (2) that the two linguistic factors syntactic complexity and turn transition potential are weakly correlated with local entropy of listener gaze.
pdf
bib
abs
Conversational Feedback in Scripted versus Spontaneous Dialogues: A Comparative Analysis
Ildiko Pilan
|
Laurent Prévot
|
Hendrik Buschmeier
|
Pierre Lison
Proceedings of the 25th Annual Meeting of the Special Interest Group on Discourse and Dialogue
Scripted dialogues such as movie and TV subtitles constitute a widespread source of training data for conversational NLP models. However, there are notable linguistic differences between these dialogues and spontaneous interactions, especially regarding the occurrence of communicative feedback such as backchannels, acknowledgments, or clarification requests. This paper presents a quantitative analysis of such feedback phenomena in both subtitles and spontaneous conversations. Based on conversational data spanning eight languages and multiple genres, we extract lexical statistics, classifications from a dialogue act tagger, expert annotations and labels derived from a fine-tuned Large Language Model (LLM). Our main empirical findings are that (1) communicative feedback is markedly less frequent in subtitles than in spontaneous dialogues and (2) subtitles contain a higher proportion of negative feedback. We also show that dialogues generated by standard LLMs lie much closer to scripted dialogues than spontaneous interactions in terms of communicative feedback.
2023
pdf
bib
abs
Does Listener Gaze in Face-to-Face Interaction Follow the Entropy Rate Constancy Principle: An Empirical Study
Yu Wang
|
Hendrik Buschmeier
Findings of the Association for Computational Linguistics: EMNLP 2023
It is generally assumed that language (written and spoken) follows the entropy rate constancy (ERC) principle, which states that the information density of a text is constant over time. Recently, this has also been found for nonverbal gestures used in monologue, but it is still unclear whether the ERC principle also applies to listeners’ nonverbal signals. We focus on listeners’ gaze behaviour extracted from video-recorded conversations and trained a transformer-based neural sequence model to process the gaze data of the dialogues and compute its information density. We also compute the information density of the corresponding speech using a pre-trained language model. Our results show (1) that listeners’ gaze behaviour in dialogues roughly follows the ERC principle, as well as (2) a congruence between information density of speech and listeners’ gaze behaviour.
2021
pdf
bib
abs
Decoding, Fast and Slow: A Case Study on Balancing Trade-Offs in Incremental, Character-level Pragmatic Reasoning
Sina Zarrieß
|
Hendrik Buschmeier
|
Ting Han
|
Simeon Schüz
Proceedings of the 14th International Conference on Natural Language Generation
Recent work has adopted models of pragmatic reasoning for the generation of informative language in, e.g., image captioning. We propose a simple but highly effective relaxation of fully rational decoding, based on an existing incremental and character-level approach to pragmatically informative neural image captioning. We implement a mixed, ‘fast’ and ‘slow’, speaker that applies pragmatic reasoning occasionally (only word-initially), while unrolling the language model. In our evaluation, we find that increased informativeness through pragmatic decoding generally lowers quality and, somewhat counter-intuitively, increases repetitiveness in captions. Our mixed speaker, however, achieves a good balance between quality and informativeness.
2018
pdf
bib
Proceedings of the Workshop on NLG for Human–Robot Interaction
Mary Ellen Foster
|
Hendrik Buschmeier
|
Dimitra Gkatzia
Proceedings of the Workshop on NLG for Human–Robot Interaction
2014
pdf
bib
abs
ALICO: a multimodal corpus for the study of active listening
Hendrik Buschmeier
|
Zofia Malisz
|
Joanna Skubisz
|
Marcin Wlodarczak
|
Ipke Wachsmuth
|
Stefan Kopp
|
Petra Wagner
Proceedings of the Ninth International Conference on Language Resources and Evaluation (LREC'14)
The Active Listening Corpus (ALICO) is a multimodal database of spontaneous dyadic conversations with diverse speech and gestural annotations of both dialogue partners. The annotations consist of short feedback expression transcription with corresponding communicative function interpretation as well as segmentation of interpausal units, words, rhythmic prominence intervals and vowel-to-vowel intervals. Additionally, ALICO contains head gesture annotation of both interlocutors. The corpus contributes to research on spontaneous human–human interaction, on functional relations between modalities, and timing variability in dialogue. It also provides data that differentiates between distracted and attentive listeners. We describe the main characteristics of the corpus and present the most important results obtained from analyses in recent years.
pdf
bib
Situationally Aware In-Car Information Presentation Using Incremental Speech Generation: Safer, and More Effective
Spyros Kousidis
|
Casey Kennington
|
Timo Baumann
|
Hendrik Buschmeier
|
Stefan Kopp
|
David Schlangen
Proceedings of the EACL 2014 Workshop on Dialogue in Motion
2012
pdf
bib
Referring in Installments: A Corpus Study of Spoken Object References in an Interactive Virtual Environment
Kristina Striegnitz
|
Hendrik Buschmeier
|
Stefan Kopp
INLG 2012 Proceedings of the Seventh International Natural Language Generation Conference
pdf
bib
Combining Incremental Language Generation and Incremental Speech Synthesis for Adaptive Information Presentation
Hendrik Buschmeier
|
Timo Baumann
|
Benjamin Dosch
|
Stefan Kopp
|
David Schlangen
Proceedings of the 13th Annual Meeting of the Special Interest Group on Discourse and Dialogue
2010
pdf
bib
Middleware for Incremental Processing in Conversational Agents
David Schlangen
|
Timo Baumann
|
Hendrik Buschmeier
|
Okko Buß
|
Stefan Kopp
|
Gabriel Skantze
|
Ramin Yaghoubzadeh
Proceedings of the SIGDIAL 2010 Conference
2009
pdf
bib
An Alignment-Capable Microplanner for Natural Language Generation
Hendrik Buschmeier
|
Kirsten Bergmann
|
Stefan Kopp
Proceedings of the 12th European Workshop on Natural Language Generation (ENLG 2009)