Tobias Scheffer


2024

pdf bib
Fine-Tuning Pre-Trained Language Models with Gaze Supervision
Shuwen Deng | Paul Prasse | David Reich | Tobias Scheffer | Lena Jäger
Proceedings of the 62nd Annual Meeting of the Association for Computational Linguistics (Volume 2: Short Papers)

Human gaze data provide cognitive information that reflect human language comprehension and has been effectively integrated into a variety of natural language processing (NLP) tasks, demonstrating improved performance over corresponding plain text-based models. In this work, we propose to integrate a gaze module into pre-trained language models (LMs) at the fine-tuning stage to improve their capabilities to learn representations that are grounded in human language processing. This is done by extending the conventional purely text-based fine-tuning objective with an auxiliary loss to exploit cognitive signals. The gaze module is only included during training, retaining compatibility with existing pre-trained LM-based pipelines. We evaluate the proposed approach using two distinct pre-trained LMs on the GLUE benchmark and observe that the proposed model improves performance compared to both standard fine-tuning and traditional text augmentation baselines.

2023

pdf bib
Pre-Trained Language Models Augmented with Synthetic Scanpaths for Natural Language Understanding
Shuwen Deng | Paul Prasse | David Reich | Tobias Scheffer | Lena Jäger
Proceedings of the 2023 Conference on Empirical Methods in Natural Language Processing

Human gaze data offer cognitive information that reflects natural language comprehension. Indeed, augmenting language models with human scanpaths has proven beneficial for a range of NLP tasks, including language understanding. However, the applicability of this approach is hampered because the abundance of text corpora is contrasted by a scarcity of gaze data. Although models for the generation of human-like scanpaths during reading have been developed, the potential of synthetic gaze data across NLP tasks remains largely unexplored. We develop a model that integrates synthetic scanpath generation with a scanpath-augmented language model, eliminating the need for human gaze data. Since the model’s error gradient can be propagated throughout all parts of the model, the scanpath generator can be fine-tuned to downstream tasks. We find that the proposed model not only outperforms the underlying language model, but achieves a performance that is comparable to a language model augmented with real human gaze data. Our code is publicly available.

2014

pdf bib
A Model of Individual Differences in Gaze Control During Reading
Niels Landwehr | Sebastian Arzt | Tobias Scheffer | Reinhold Kliegl
Proceedings of the 2014 Conference on Empirical Methods in Natural Language Processing (EMNLP)

2005

pdf bib
Predicting Sentences using N-Gram Language Models
Steffen Bickel | Peter Haider | Tobias Scheffer
Proceedings of Human Language Technology Conference and Conference on Empirical Methods in Natural Language Processing