|3/2013 - 12|
3D Hand Gesture Recognition using the Hough TransformOPRISESCU, S. , BARTH, E.
|Click to see author's profile in SCOPUS, IEEE Xplore, Web of Science|
|Download PDF (715 KB) | Citation | Downloads: 492 | Views: 2,625|
image processing, computer vision, gesture recognition, Kinect camera, Hough transform
Blue keywords are present in both the references section and the paper title.
About this article
Date of Publication: 2013-08-31
Volume 13, Issue 3, Year 2013, On page(s): 71 - 76
ISSN: 1582-7445, e-ISSN: 1844-7600
Digital Object Identifier: 10.4316/AECE.2013.03012
Web of Science Accession Number: 000326321600012
SCOPUS ID: 84884965434
This paper presents an automatic 3D dynamic hand gesture recognition algorithm relying on both intensity and depth information provided by a Kinect camera. Gesture classification consists of a decision tree constructed on six parameters delivered by the Hough transform of projected 3D points. The Hough transform is originally applied, for the first time, on the projected gesture trajectories to obtain a reliable decision. The experimental data obtained from 300 video sequences with different subjects validate the proposed recognition method.
|References|||||Cited By «-- Click to see who has cited this paper|
| A. Kolb, E. Barth, R. Koch, R. Larsen, "Time-of-Flight Cameras in Computer Graphics," In Computer Graphics Forum, 29(1), pp 141-159, 2010 |
[CrossRef] [Web of Science Times Cited 120]
 X. Liu and K. Fujimura, "Hand gesture recognition using depth data," Proc. of the 6th IEEE international conf. on Automatic face and gesture recognition (FGR' 04), Washington, DC, USA, 529-534, 2004.
 S. Mitra and T. Acharya, "Gesture Recognition: A Survey," IEEE Trans. on Syst., man, and cybernetics, Part C: Applications and Reviews, pp. 311-324, vol. 37, no. 3, may 2007
[CrossRef] [Web of Science Times Cited 706]
 M. B. Holte, T. B. Moeslund, and P. Fihl, "View-invariant gesture recognition using 3D optical flow and harmonic motion context," Comput. Vis. Image Underst. 114, 12, pp. 1353-1361, 2010.
[CrossRef] [Web of Science Times Cited 38]
 P. Doliotis, A. Stefan, C. McMurrough, D. Eckhard, and V. Athitsos, "Comparing gesture recognition accuracy using color and depth information," in Proceedings of PETRA, pp. 20:1-20:7, 2011.
 C. Keskin, A. T. Cemgil, and L. Akarun, "DTW Based Clustering to Improve Hand Gesture Recognition," in Proceedings of HBU'11, pp. 72-81, Amsterdam, 2011.
 L. Gallo, A.P. Placitelli, and M. Ciampi, "Controller-free exploration of medical image data: experiencing the Kinect," Proc. of. 24th IEEE CMBS'11, Piscataway, NJ, USA, 2011.
[CrossRef] [Web of Science Times Cited 71]
 S. Soutschek, J. Penne and J. Hornegger, 3D gesture-based scene navigation in medical imaging applications using time-of-flight cameras, IEEE Conf. on Computer Vision & Pattern Recogn., Workshop on ToF-Camera based Computer Vision (2008).
 P. Yanik et al., "Use of Kinect Depth Data and Growing Neural Gas for Gesture Based Robot Control," in Proc. of PervaSense, pp. 283-290, 2012.
 Prodan, R.-C., Pentiuc, S.-G., Vatavu, R.-D., "An Efficient Solution for Hand Gesture Recognition from Video Sequence," Advances in Electrical and Computer Engineering, vol. 12, no. 3, pp. 85-88, 2012,
[CrossRef] [Full Text] [Web of Science Times Cited 2]
 K. Lai, J. Konrad, and P. Ishwar, "A gesture-driven computer interface using Kinect camera," in Proc. Southwest Symposium on Image Analysis and Interpretation, Apr. 2012.
 Q. Munib, M. Habeeb, B. Takruri and H. A. Al-Malik, "American sign language (ASL) recognition based on Hough transform and neural networks," Expert Systems with Applications, vol. 32, 1, pp. 24-37, 2007.
[CrossRef] [Web of Science Times Cited 38]
 O. Altun, S. Albayrak, "Turkish fingerspelling recognition system using Generalized Hough Transform, interest regions, and local descriptors," Patt. Rec. Letters, vol. 32, 13, pp. 1626-1632, 2011.
[CrossRef] [Web of Science Times Cited 9]
 N.N. Bhat, "Real time robust hand gesture recognition and visual servoing," India Conference (INDICON), Annual IEEE, pp.1153-1157, 7-9 December 2012
Web of Science® Citations for all references: 984 TCR
SCOPUS® Citations for all references: 0
Web of Science® Average Citations per reference: 70 ACR
SCOPUS® Average Citations per reference: 0
TCR = Total Citations for References / ACR = Average Citations per Reference
We introduced in 2010 - for the first time in scientific publishing, the term "References Weight", as a quantitative indication of the quality ... Read more
Citations for references updated on 2019-03-19 05:23 in 81 seconds.
Note1: Web of Science® is a registered trademark of Clarivate Analytics.
Note2: SCOPUS® is a registered trademark of Elsevier B.V.
Disclaimer: All queries to the respective databases were made by using the DOI record of every reference (where available). Due to technical problems beyond our control, the information is not always accurate. Please use the CrossRef link to visit the respective publisher site.
Faculty of Electrical Engineering and Computer Science
Stefan cel Mare University of Suceava, Romania
All rights reserved: Advances in Electrical and Computer Engineering is a registered trademark of the Stefan cel Mare University of Suceava. No part of this publication may be reproduced, stored in a retrieval system, photocopied, recorded or archived, without the written permission from the Editor. When authors submit their papers for publication, they agree that the copyright for their article be transferred to the Faculty of Electrical Engineering and Computer Science, Stefan cel Mare University of Suceava, Romania, if and only if the articles are accepted for publication. The copyright covers the exclusive rights to reproduce and distribute the article, including reprints and translations.
Permission for other use: The copyright owner's consent does not extend to copying for general distribution, for promotion, for creating new works, or for resale. Specific written permission must be obtained from the Editor for such copying. Direct linking to files hosted on this website is strictly prohibited.
Disclaimer: Whilst every effort is made by the publishers and editorial board to see that no inaccurate or misleading data, opinions or statements appear in this journal, they wish to make it clear that all information and opinions formulated in the articles, as well as linguistic accuracy, are the sole responsibility of the author.