Benchmarking Data Augmentation for Contrastive Learning in Static Sign Language Recognition

Research output: Contribution in Book/Catalog/Report/Conference proceedingConference contribution

Abstract

Sign language (SL) is a communication method used by deaf people. Static sign language recognition (SLR) is a challenging task aimed at identifying signs in images, for which acquisition of annotated data is time-consuming. To leverage unannotated data, practitioners have turned to unsupervised methods. Contrastive representation learning proved to be effective in capturing important features from unannotated data. It is known that the performance of the contrastive model depends on the data augmentation technique used during training. For various applications, a set of effective data augmentation has been identified, but it is not yet the case for SL. This paper identifies the most effective augmentation
for static SLR. The results show a difference in accuracy of up to 30% between appearance-based augmentations combined with translations and
augmentations based on rotations, erasing, or vertical flips.
Original languageFrench
Title of host publicationESANN 2024
Subtitle of host publication32nd European Symposium on Artificial Neural Networks, Computational Intelligence and Machine Learning
Publisheri6doc.com
Publication statusAccepted/In press - 15 Jan 2025

Cite this