2024
pdf
bib
Proceedings of GermEval 2024 Shared Task on Statement Segmentation in German Easy Language (StaGE)
Thorben Schomacker
|
Miriam Anschütz
|
Regina Stodden
Proceedings of GermEval 2024 Shared Task on Statement Segmentation in German Easy Language (StaGE)
pdf
bib
Overview of the GermEval 2024 Shared Task on Statement Segmentation in German Easy Language (StaGE)
Thorben Schomacker
|
Miriam Anschütz
|
Regina Stodden
|
Georg Groh
|
Marina Tropmann-Frick
Proceedings of GermEval 2024 Shared Task on Statement Segmentation in German Easy Language (StaGE)
pdf
bib
abs
Images Speak Volumes: User-Centric Assessment of Image Generation for Accessible Communication
Miriam Anschütz
|
Tringa Sylaj
|
Georg Groh
Proceedings of the Third Workshop on Text Simplification, Accessibility and Readability (TSAR 2024)
Explanatory images play a pivotal role in accessible and easy-to-read (E2R) texts. However, the images available in online databases are not tailored toward the respective texts, and the creation of customized images is expensive. In this large-scale study, we investigated whether text-to-image generation models can close this gap by providing customizable images quickly and easily. We benchmarked seven, four open- and three closed-source, image generation models and provide an extensive evaluation of the resulting images. In addition, we performed a user study with people from the E2R target group to examine whether the images met their requirements. We find that some of the models show remarkable performance, but none of the models are ready to be used at a larger scale without human supervision. Our research is an important step toward facilitating the creation of accessible information for E2R creators and tailoring accessible images to the target group’s needs.
pdf
bib
abs
Simpler Becomes Harder: Do LLMs Exhibit a Coherent Behavior on Simplified Corpora?
Miriam Anschütz
|
Edoardo Mosca
|
Georg Groh
Proceedings of the Workshop on DeTermIt! Evaluating Text Difficulty in a Multilingual Context @ LREC-COLING 2024
Text simplification seeks to improve readability while retaining the original content and meaning. Our study investigates whether pre-trained classifiers also maintain such coherence by comparing their predictions on both original and simplified inputs. We conduct experiments using 11 pre-trained models, including BERT and OpenAI’s GPT 3.5, across six datasets spanning three languages. Additionally, we conduct a detailed analysis of the correlation between prediction change rates and simplification types/strengths. Our findings reveal alarming inconsistencies across all languages and models. If not promptly addressed, simplified inputs can be easily exploited to craft zero-iteration model-agnostic adversarial attacks with success rates of up to 50%.
2023
pdf
bib
abs
Language Models for German Text Simplification: Overcoming Parallel Data Scarcity through Style-specific Pre-training
Miriam Anschütz
|
Joshua Oehms
|
Thomas Wimmer
|
Bartłomiej Jezierski
|
Georg Groh
Findings of the Association for Computational Linguistics: ACL 2023
Automatic text simplification systems help to reduce textual information barriers on the internet. However, for languages other than English, only few parallel data to train these systems exists. We propose a two-step approach to overcome this data scarcity issue. First, we fine-tuned language models on a corpus of German Easy Language, a specific style of German. Then, we used these models as decoders in a sequence-to-sequence simplification task. We show that the language models adapt to the style characteristics of Easy Language and output more accessible texts. Moreover, with the style-specific pre-training, we reduced the number of trainable parameters in text simplification models. Hence, less parallel data is sufficient for training. Our results indicate that pre-training on unaligned data can reduce the required parallel data while improving the performance on downstream tasks.
pdf
bib
abs
This is not correct! Negation-aware Evaluation of Language Generation Systems
Miriam Anschütz
|
Diego Miguel Lozano
|
Georg Groh
Proceedings of the 16th International Natural Language Generation Conference
Large language models underestimate the impact of negations on how much they change the meaning of a sentence. Therefore, learned evaluation metrics based on these models are insensitive to negations. In this paper, we propose NegBLEURT, a negation-aware version of the BLEURT evaluation metric. For that, we designed a rule-based sentence negation tool and used it to create the CANNOT negation evaluation dataset. Based on this dataset, we fine-tuned a sentence transformer and an evaluation metric to improve their negation sensitivity. Evaluating these models on existing benchmarks shows that our fine-tuned models outperform existing metrics on the negated sentences by far while preserving their base models’ performances on other perturbations.
2022
pdf
bib
abs
TUM Social Computing at GermEval 2022: Towards the Significance of Text Statistics and Neural Embeddings in Text Complexity Prediction
Miriam Anschütz
|
Georg Groh
Proceedings of the GermEval 2022 Workshop on Text Complexity Assessment of German Text
In this paper, we describe our submission to the GermEval 2022 Shared Task on Text Complexity Assessment of German Text. It addresses the problem of predicting the complexity of German sentences on a continuous scale. While many related works still rely on handcrafted statistical features, neural networks have emerged as state-of-the-art in other natural language processing tasks. Therefore, we investigate how both can complement each other and which features are most relevant for text complexity prediction in German. We propose a fine-tuned German DistilBERT model enriched with statistical text features that achieved fourth place in the shared task with a RMSE of 0.481 on the competition’s test data.