ABSTRACT
We present the LaRED, a Large RGB-D Extensible hand gesture Dataset, recorded with an Intel's newly-developed short range depth camera. This dataset is unique and differs from the existing ones in several aspects. Firstly, the large volume of data recorded: 243, 000 tuples where each tuple is composed of a color image, a depth image, and a mask of the hand region. Secondly, the number of different classes provided: a total of 81 classes (27 gestures in 3 different rotations). Thirdly, the extensibility of dataset: the software used to record and inspect the dataset is also available, giving the possibility for future users to increase the number of data as well as the number of gestures. Finally, in this paper, some experiments are presented to characterize the dataset and establish a baseline as the start point to develop more complex recognition algorithms. The LaRED dataset is publicly available at: http://mclab.citi.sinica.edu.tw/dataset/lared/lared.html.
- H. Birk, T. B. Moeslund, and C. B. Madsen. Real-time recognition of hand alphabet gestures using principal component analysis. In 10th Scandinavian Conference on Image Analysis, 1997.Google Scholar
- A. Kurakin, Z. Zhang, and Z. Liu. A real-time system for dynamic hand gesture recognition with a depth sensor. In European Signal Processing Conference, 2012.Google Scholar
- T. kyun Kim, S. fai Wong, and R. Cipolla. Tensor canonical correlation analysis for action classification. In IEEE Computer Vision and Pattern Recognition, 2007.Google ScholarCross Ref
- Y. LeCun, K. Kavukcuoglu, and C. Farabet. Convolutional networks and applications in vision. In IEEE International Symposium on Circuits and Systems, 2010.Google Scholar
- L. Liu and L. Shao. Learning discriminative representations from rgb-d video data. In International Joint Conferences on Artificial Intelligence, 2013. Google ScholarDigital Library
- S. Marcel and O. Bernier. Hand posture recognition in a body-face centered space. In Proceedings of the Conference on Human Factors in Computer Systems (CHI), 1999. Google ScholarDigital Library
- S. Marcel, O. Bernier, J.-E. Viallet, and D. Collobert. Hand gesture recognition using input-output hidden markov models. In Proceedings of the 4th International Conference on Automatic Face and Gesture Recognition, 2000. Google ScholarDigital Library
- E. Ohn-Bar and M. M. Trivedi. The power is in your hands: 3d analysis of hand gestures in naturalistic video. In IEEE Computer Vision and Pattern Recognition Workshop on Analysis and Modeling of Faces and Gestures, 2013. Google ScholarDigital Library
- R. B. Palm. Prediction as a candidate for learning deep hierarchical models of data. Master's thesis, 2012.Google Scholar
- Z. Ren, J. Yuan, and Z. Zhang. Robust hand gesture recognition based on finger-earth mover's distance with a commodity depth camera. In Proceedings of the 19th ACM international conference on Multimedia, 2011. Google ScholarDigital Library
- J. Triesch and C. von der Malsburg. Robust classification of hand postures against complex backgrounds. In Proceedings of the Second International Conference on Automatic Face and Gesture Recognition, 1996. Google ScholarDigital Library
- J. Triesch and C. von der Malsburg. A system for person-independent hand posture recognition against complex backgrounds. IEEE Transactions Pattern Analysis Machine Intelligence, 23(12):1449--1453, 2001. Google ScholarDigital Library
Index Terms
- LaRED: a large RGB-D extensible hand gesture dataset
Recommendations
VRGestures: Controller and Hand Gesture Datasets for Virtual Reality
Advances in Computer GraphicsAbstractGesture Recognition is attracting increasingly more attention over the years and has been adopted in main applications in the real world and the Virtual one. New generation Virtual Reality (VR) headsets like the Meta Quest 2 support hand tracking ...
Gesture Recognition and Multi-modal Fusion on a New Hand Gesture Dataset
Pattern Recognition Applications and MethodsAbstractWe present a baseline for gesture recognition using state-of-the-art sequence classifiers on a new freely available multi-modal dataset of free-hand gestures. The dataset consists of roughly 100,000 samples, grouped into six classes of typical and ...
An eye tracking dataset for point of gaze detection
ETRA '12: Proceedings of the Symposium on Eye Tracking Research and ApplicationsThis paper presents a new, publicly available eye tracking dataset, aimed to be used as a benchmark for Point of Gaze (PoG) detection algorithms. The dataset consists of a set of videos recording the eye motion of human test subjects as they were ...
Comments