Carregant...
Fitxers
Tipus de document
ArticleVersió
Versió publicadaData de publicació
Llicència de publicació
Si us plau utilitzeu sempre aquest identificador per citar o enllaçar aquest document: https://hdl.handle.net/2445/157458
Multi-Modal Deep Hand Sign Language Recognition in Still Images Using Restricted Boltzmann Machine
Títol de la revista
Director/Tutor
ISSN de la revista
Títol del volum
Recurs relacionat
Resum
In this paper, a deep learning approach, Restricted Boltzmann Machine (RBM), is used to perform automatic hand sign language recognition from visual data. We evaluate how RBM, as a deep generative model, is capable of generating the distribution of the input data for an enhanced recognition of unseen data. Two modalities, RGB and Depth, are considered in the model input in three forms: original image, cropped image, and noisy cropped image. Five crops of the input image are used and the hand of these cropped images are detected using Convolutional Neural Network (CNN). After that, three types of the detected hand images are generated for each modality and input to RBMs. The outputs of the RBMs for two modalities are fused in another RBM in order to recognize the output sign label of the input image. The proposed multi-modal model is trained on all and part of the American alphabet and digits of four publicly available datasets. We also evaluate the robustness of the proposal against noise. Experimental results show that the proposed multi-modal model, using crops and the RBM fusing methodology, achieves state-of-the-art results on Massey University Gesture Dataset 2012, American Sign Language (ASL). and Fingerspelling Dataset from the University of Surrey's Center for Vision, Speech and Signal Processing, NYU, and ASL Fingerspelling A datasets.
Matèries
Matèries (anglès)
Citació
Citació
RASTGOO, Razieh, KIANI, Kourosh, ESCALERA GUERRERO, Sergio. Multi-Modal Deep Hand Sign Language Recognition in Still Images Using Restricted Boltzmann Machine. _Sensors_. 2018. Vol. 20, núm. 11, pàgs. 809. [consulta: 22 de gener de 2026]. ISSN: 1424-8220. [Disponible a: https://hdl.handle.net/2445/157458]