@inproceedings{745e98cd3b994044acd53d4a21259eae,
title = "Generating Synthetic Sign Language Datasets Using Conditional Generative Adversarial Networks",
abstract = "Approximately 70 million individuals globally utilize sign languages as a means of communication due to their hearing impairment. The study undertaken in sign languages is extensive and fruitful. However, there are over 300 sign languages worldwide, but most research focuses on a single language [1]. Creating AI models for sign language challenges sometimes requires large datasets, which can be difficult to produce. Various research has produced datasets for sign language using various methodologies; nevertheless, they often focus on specific sign languages. Developing hand skeleton templates for sign languages provides a more efficient method than creating numerous instances of distinct signs. By creating a basic framework or structure, it becomes much simpler to utilize generative models, like GANs[2], to generate a wide range of different versions of the framework. These generative models can effectively reproduce and adjust the fundamental structures into many sign language forms, capturing the diversity in hand shapes, orientations, and movements necessary for precise sign representation. The main objective of our research is to develop a conditional generative adversarial network (cGAN) model that can generate hand images based on hand skeletons; this approach not only improves the capacity to generate sign language data on a larger scale, but also guarantees uniformity across different versions of signs. This makes it easier to create sign language recognition systems that are more reliable and flexible. To train this model, we devised a web scraping technique that produced a significant collection of hand photos taken from TED lecture recordings, together with their corresponding skeletons. Our created cGAN-based model allows researchers to generate artificial hand images by employing target skeleton inputs. This enables the creation of extensive datasets for sign language. Our contribution is expected to streamline the exploration of additional sign languages that encounter challenges in collecting datasets.",
keywords = "cGAN, GAN, generative models, sign language, WGAN",
author = "Yehuda Yadid and Sarel Cohen and Raid Saabni",
note = "Publisher Copyright: {\textcopyright} 2025 SPIE.; 17th International Conference on Machine Vision, ICMV 2024 ; Conference date: 10-10-2024 Through 13-10-2024",
year = "2025",
month = jan,
day = "1",
doi = "10.1117/12.3055193",
language = "English",
series = "Proceedings of SPIE - The International Society for Optical Engineering",
publisher = "SPIE",
editor = "Wolfgang Osten",
booktitle = "Seventeenth International Conference on Machine Vision, ICMV 2024",
address = "United States",
}