Di Wu
Multi-modality gesture detection and recognition with un-supervision, randomization and discrimination
Wu, Di; Chen, Guang; Clarke, Daniel; Weikersdorfer, David; Giuliani, Manuel; Gaschler, Andre; Knoll, Alois
Authors
Guang Chen
Daniel Clarke
David Weikersdorfer
Manuel Giuliani Manuel.Giuliani@uwe.ac.uk
Co- Director Bristol Robotics Laboratory
Andre Gaschler
Alois Knoll
Contributors
Lourdes Agapito
Editor
Michael Bronstein
Editor
Carsten Rother
Editor
Abstract
© Springer International Publishing Switzerland 2015. We describe in this paper our gesture detection and recognition system for the 2014 ChaLearn Looking at People (Track 3: Gesture Recognition) organized by ChaLearn in conjunction with the ECCV 2014 conference. The competition’s task was to learn a vacabulary of 20 types of Italian gestures and detect them in sequences. Our system adopts a multi-modality approach for detecting as well as recognizing the gestures. The goal of our approach is to identify semantically meaningful contents from dense sampling spatio-temporal feature space for gesture recognition. To achieve this, we develop three concepts under the random forest framework: un-supervision; discrimination; and randomization. Un-supervision learns spatio-temporal features from two channels (grayscale and depth) of RGB-D video in an unsupervised way. Discrimination extracts the information in dense sampling spatio-temporal space effectively. Randomization explores the dense sampling spatio-temporal feature space efficiently. An evaluation of our approach shows that we achieve a mean Jaccard Index of 0.6489, and a mean average accuracy of 90.3% over the test dataset.
Citation
Wu, D., Chen, G., Clarke, D., Weikersdorfer, D., Giuliani, M., Gaschler, A., & Knoll, A. (2015). Multi-modality gesture detection and recognition with un-supervision, randomization and discrimination. Lecture Notes in Artificial Intelligence, 8925, 608-622. https://doi.org/10.1007/978-3-319-16178-5_43
Journal Article Type | Conference Paper |
---|---|
Conference Name | ChaLearn Looking at People Workshop, European Conference on Computer Vision (ECCV2014) |
Acceptance Date | Sep 6, 2014 |
Publication Date | Jan 1, 2015 |
Journal | Lecture Notes in Computer Science (including subseries Lecture Notes in Artificial Intelligence and Lecture Notes in Bioinformatics) |
Print ISSN | 0302-9743 |
Electronic ISSN | 1611-3349 |
Publisher | Springer Verlag |
Peer Reviewed | Peer Reviewed |
Volume | 8925 |
Pages | 608-622 |
DOI | https://doi.org/10.1007/978-3-319-16178-5_43 |
Keywords | multi-modality gesture, unsupervised learning, random forest, discriminative training |
Public URL | https://uwe-repository.worktribe.com/output/812125 |
Publisher URL | http://dx.doi.org/10.1007/978-3-319-16178-5_43 |
Additional Information | Title of Conference or Conference Proceedings : ChaLearn Looking at People Workshop, European Conference on Computer Vision (ECCV2014) |
You might also like
Usability study of a novel triple-arm mixed-reality robot teleoperation system
(2023)
Conference Proceeding
Measuring visual social engagement from proxemics and gaze
(2022)
Conference Proceeding
User requirements for a robot teleoperation system for general medical examination
(2022)
Conference Proceeding
Demo: Untethered haptic teleoperation for nuclear decommissioning using a low-power wireless control technology
(2022)
Conference Proceeding
Downloadable Citations
About UWE Bristol Research Repository
Administrator e-mail: repository@uwe.ac.uk
This application uses the following open-source libraries:
SheetJS Community Edition
Apache License Version 2.0 (http://www.apache.org/licenses/)
PDF.js
Apache License Version 2.0 (http://www.apache.org/licenses/)
Font Awesome
SIL OFL 1.1 (http://scripts.sil.org/OFL)
MIT License (http://opensource.org/licenses/mit-license.html)
CC BY 3.0 ( http://creativecommons.org/licenses/by/3.0/)
Powered by Worktribe © 2024
Advanced Search