2024
pdf
bib
abs
Sign Language Translation with Sentence Embedding Supervision
Yasser Hamidullah
|
Josef van Genabith
|
Cristina España-Bonet
Proceedings of the 62nd Annual Meeting of the Association for Computational Linguistics (Volume 2: Short Papers)
State-of-the-art sign language translation (SLT) systems facilitate the learning process through gloss annotations, either in an end2end manner or by involving an intermediate step. Unfortunately, gloss labelled sign language data is usually not available at scale and, when available, gloss annotations widely differ from dataset to dataset. We present a novel approach using sentence embeddings of the target sentences at training time that take the role of glosses. The new kind of supervision does not need any manual annotation but it is learned on raw textual data. As our approach easily facilitates multilinguality, we evaluate it on datasets covering German (PHOENIX-2014T) and American (How2Sign) sign languages and experiment with mono- and multilingual sentence embeddings and translation systems. Our approach significantly outperforms other gloss-free approaches, setting the new state-of-the-art for data sets where glosses are not available and when no additional SLT datasets are used for pretraining, diminishing the gap between gloss-free and gloss-dependent systems.
2022
pdf
bib
abs
Spatio-temporal Sign Language Representation and Translation
Yasser Hamidullah
|
Josef Van Genabith
|
Cristina España-bonet
Proceedings of the Seventh Conference on Machine Translation (WMT)
This paper describes the DFKI-MLT submission to the WMT-SLT 2022 sign language translation (SLT) task from Swiss German Sign Language (video) into German (text). State-of-the-art techniques for SLT use a generic seq2seq architecture with customized input embeddings. Instead of word embeddings as used in textual machine translation, SLT systems use features extracted from video frames. Standard approaches often do not benefit from temporal features. In our participation, we present a system that learns spatio-temporal feature representations and translation in a single model, resulting in a real end-to-end architecture expected to better generalize to new data sets. Our best system achieved 5 ± 1 BLEU points on the development set, but the performance on the test dropped to 0.11 ± 0.06 BLEU points.
2021
pdf
bib
abs
AVASAG: A German Sign Language Translation System for Public Services (short paper)
Fabrizio Nunnari
|
Judith Bauerdiek
|
Lucas Bernhard
|
Cristina España-Bonet
|
Corinna Jäger
|
Amelie Unger
|
Kristoffer Waldow
|
Sonja Wecker
|
Elisabeth André
|
Stephan Busemann
|
Christian Dold
|
Arnulph Fuhrmann
|
Patrick Gebhard
|
Yasser Hamidullah
|
Marcel Hauck
|
Yvonne Kossel
|
Martin Misiak
|
Dieter Wallach
|
Alexander Stricker
Proceedings of the 1st International Workshop on Automatic Translation for Signed and Spoken Languages (AT4SSL)
This paper presents an overview of AVASAG; an ongoing applied-research project developing a text-to-sign-language translation system for public services. We describe the scientific innovation points (geometry-based SL-description, 3D animation and video corpus, simplified annotation scheme, motion capture strategy) and the overall translation pipeline.