ChaLearn Looking at People: IsoGD and ConGD Large-Scale RGB-D Gesture Recognition
View/ Open
Access
info:eu-repo/semantics/openAccessDate
MAY 2022Author
Wan, JunLin, Chi
Wen, Longyin
Li, Yunan
Miao, Qiguang
Escalera, Sergio
Anbarjafari, Gholamreza
Guyon, Isabelle
Guo, Guodong
Li, Stan Z.
Metadata
Show full item recordCitation
Wan, J., Lin, C., Wen, L., Li, Y., Miao, Q., Escalera, S., Anbarjafari, G., Li, S. Z. (May 01, 2022). ChaLearn Looking at People: IsoGD and ConGD Large-Scale RGB-D Gesture Recognition. Ieee Transactions on Cybernetics, 52, 5, 3422-3433.Abstract
The ChaLearn large-scale gesture recognition challenge has run twice in two workshops in conjunction with the International Conference on Pattern Recognition (ICPR) 2016 and International Conference on Computer Vision (ICCV) 2017, attracting more than 200 teams around the world. This challenge has two tracks, focusing on isolated and continuous gesture recognition, respectively. It describes the creation of both benchmark datasets and analyzes the advances in large-scale gesture recognition based on these two datasets. In this article, we discuss the challenges of collecting large-scale ground-truth annotations of gesture recognition and provide a detailed analysis of the current methods for large-scale isolated and continuous gesture recognition. In addition to the recognition rate and mean Jaccard index (MJI) as evaluation metrics used in previous challenges, we introduce the corrected segmentation rate (CSR) metric to evaluate the performance of temporal segmentation for continuous gesture recognition. Furthermore, we propose a bidirectional long short-term memory (Bi-LSTM) method, determining video division points based on skeleton points. Experiments show that the proposed Bi-LSTM outperforms state-of-the-art methods with an absolute improvement of 8.1% (from 0.8917 to 0.9639) of CSR.