Evaluating Human Perception and Bias in AI-Generated Humor

Narendra Nath Joshi


Abstract
This paper explores human perception of AI-generated humor, examining biases and the ability to distinguish between human and AI-created jokes. Through a between-subjects user study involving 174 participants, we tested hypotheses on quality perception, source identification, and demographic influences. Our findings reveal that AI-generated jokes are rated comparably to human-generated ones, with source blindness improving AI humor ratings. Participants struggled to identify AI-generated jokes accurately, and repeated exposure led to increased appreciation. Younger participants showed more favorable perceptions, while technical background had no significant impact. These results challenge preconceptions about AI’s humor capabilities and highlight the importance of addressing biases in AI content evaluation. We also suggest pathways for enhancing human-AI creative collaboration and underscore the need for transparency and ethical considerations in AI-generated content.
Anthology ID:
2025.chum-1.9
Volume:
Proceedings of the 1st Workshop on Computational Humor (CHum)
Month:
January
Year:
2025
Address:
Online
Editors:
Christian F. Hempelmann, Julia Rayz, Tiansi Dong, Tristan Miller
Venues:
chum | WS
SIG:
Publisher:
Association for Computational Linguistics
Note:
Pages:
79–87
Language:
URL:
https://aclanthology.org/2025.chum-1.9/
DOI:
Bibkey:
Cite (ACL):
Narendra Nath Joshi. 2025. Evaluating Human Perception and Bias in AI-Generated Humor. In Proceedings of the 1st Workshop on Computational Humor (CHum), pages 79–87, Online. Association for Computational Linguistics.
Cite (Informal):
Evaluating Human Perception and Bias in AI-Generated Humor (Joshi, chum 2025)
Copy Citation:
PDF:
https://aclanthology.org/2025.chum-1.9.pdf