Logo image
Using Sign Language Production as Data Augmentation to enhance Sign Language Translation
Conference proceeding   Peer reviewed

Using Sign Language Production as Data Augmentation to enhance Sign Language Translation

IVA Adjunct '25: Adjunct Proceedings of the 25th ACM International Conference on Intelligent Virtual Agents, pp.1-10
25th ACM International Conference on Intelligent Virtual Agents (Berlin, Germany)
30/09/2025

Abstract

Computing methodologies Sign Language Translation Data Augmentation Sign Language Production Generative Models Computer Vision
Machine learning models fundamentally rely on large quantities of high-quality data. Collecting the necessary data for these models can be challenging due to cost, scarcity, and privacy restrictions. Signed languages are visual languages used by the deaf community and are considered low-resource languages. Sign language datasets are often orders of magnitude smaller than their spoken language counterparts. Sign Language Production (SLP) is the task of generating sign language videos from spoken language sentences, while Sign Language Translation (SLT) is the reverse translation task. Here, we propose leveraging recent advancements in SLP to augment existing sign language datasets and enhance the performance of SLT models. For this, we utilize three techniques: a skeleton-based approach to production, sign stitching, and two photo-realistic generative models, SignGAN and SignSplat. We evaluate the effectiveness of these techniques in enhancing the performance of SLT models by generating variation in the signer's appearance and the motion of the skeletal data. Our results demonstrate that the proposed methods can effectively augment existing datasets and enhance the performance of SLT models by up to 19%, paving the way for more robust and accurate SLT systems, even in resource-constrained environments.

Metrics

36 Record Views

Has preprint

Details

Logo image

Usage Policy