Benchmarking Data Augmentation for Contrastive Learning in Static Sign Language Recognition

Résultats de recherche: Contribution dans un livre/un catalogue/un rapport/dans les actes d'une conférenceArticle dans les actes d'une conférence/un colloque

Résumé

Sign language (SL) is a communication method used by deaf people. Static sign language recognition (SLR) is a challenging task aimed at identifying signs in images, for which acquisition of annotated data is time-consuming. To leverage unannotated data, practitioners have turned to unsupervised methods. Contrastive representation learning proved to be effective in capturing important features from unannotated data. It is known that the performance of the contrastive model depends on the data augmentation technique used during training. For various applications, a set of effective data augmentation has been identified, but it is not yet the case for SL. This paper identifies the most effective augmentation
for static SLR. The results show a difference in accuracy of up to 30% between appearance-based augmentations combined with translations and
augmentations based on rotations, erasing, or vertical flips.
langue originaleFrançais
titreESANN 2024
Sous-titre32nd European Symposium on Artificial Neural Networks, Computational Intelligence and Machine Learning
Editeuri6doc.com
Etat de la publicationAccepté/sous presse - 15 janv. 2025

Contient cette citation