Click to open the HelpDesk interface
AECE - Front page banner

Menu:


FACTS & FIGURES

JCR Impact Factor: 0.595
JCR 5-Year IF: 0.661
Issues per year: 4
Current issue: Feb 2018
Next issue: May 2018
Avg review time: 105 days


PUBLISHER

Stefan cel Mare
University of Suceava
Faculty of Electrical Engineering and
Computer Science
13, Universitatii Street
Suceava - 720229
ROMANIA

Print ISSN: 1582-7445
Online ISSN: 1844-7600
WorldCat: 643243560
doi: 10.4316/AECE


TRAFFIC STATS

1,911,304 unique visits
529,094 downloads
Since November 1, 2009



No robots online now


SJR SCImago RANK

SCImago Journal & Country Rank


SEARCH ENGINES

aece.ro - Google Pagerank




TEXT LINKS

Anycast DNS Hosting
MOST RECENT ISSUES

 Volume 18 (2018)
 
     »   Issue 1 / 2018
 
 
 Volume 17 (2017)
 
     »   Issue 4 / 2017
 
     »   Issue 3 / 2017
 
     »   Issue 2 / 2017
 
     »   Issue 1 / 2017
 
 
 Volume 16 (2016)
 
     »   Issue 4 / 2016
 
     »   Issue 3 / 2016
 
     »   Issue 2 / 2016
 
     »   Issue 1 / 2016
 
 
 Volume 15 (2015)
 
     »   Issue 4 / 2015
 
     »   Issue 3 / 2015
 
     »   Issue 2 / 2015
 
     »   Issue 1 / 2015
 
 
  View all issues  


FEATURED ARTICLE

Wind Speed Prediction with Wavelet Time Series Based on Lorenz Disturbance, ZHANG, Y., WANG, P., CHENG, P., LEI, S.
Issue 3/2017

AbstractPlus






LATEST NEWS

2017-Jun-14
Thomson Reuters published the Journal Citations Report for 2016. The JCR Impact Factor of Advances in Electrical and Computer Engineering is 0.595, and the JCR 5-Year Impact Factor is 0.661.

2017-Apr-04
We have the confirmation Advances in Electrical and Computer Engineering will be included in the EBSCO database.

2017-Jan-30
We have the confirmation Advances in Electrical and Computer Engineering will be included in the Gale database.

Read More »


    
 

  1/2013 - 8

Novel Interface for Simulation of Assembly Operations in Virtual Environments

CRACIUN, E.-G. See more information about CRACIUN, E.-G. on SCOPUS See more information about CRACIUN, E.-G. on IEEExplore See more information about CRACIUN, E.-G. on Web of Science, GRISONI, L. See more information about  GRISONI, L. on SCOPUS See more information about  GRISONI, L. on SCOPUS See more information about GRISONI, L. on Web of Science, PENTIUC, S.-G. See more information about  PENTIUC, S.-G. on SCOPUS See more information about  PENTIUC, S.-G. on SCOPUS See more information about PENTIUC, S.-G. on Web of Science, RUSU, I. See more information about RUSU, I. on SCOPUS See more information about RUSU, I. on SCOPUS See more information about RUSU, I. on Web of Science
 
Click to see author's profile on See more information about the author on SCOPUS SCOPUS, See more information about the author on IEEE Xplore IEEE Xplore, See more information about the author on Web of Science Web of Science

Download PDF pdficon (812 KB) | Citation | Downloads: 371 | Views: 2,460

Author keywords
assembly operation, assembly simulation, gesture interface, interaction, virtual environment

References keywords
recognition(11), vision(8), user(8), virtual(6), tracking(6), human(6), technology(5), systems(5), pattern(5), interface(5)
Blue keywords are present in both the references section and the paper title.

About this article
Date of Publication: 2013-02-28
Volume 13, Issue 1, Year 2013, On page(s): 47 - 52
ISSN: 1582-7445, e-ISSN: 1844-7600
Digital Object Identifier: 10.4316/AECE.2013.01008
Web of Science Accession Number: 000315768300008
SCOPUS ID: 84875351874

Abstract
Quick view
Full text preview
The objective of this paper is to propose a new interface based on human gestures for simulation of assembly operation inside virtual environments. Using as a guide the current context for assembly simulation, we have analyzed the existing techniques used in developing gesture-based interface and extracted the constituent elements for an assembly gesture interface. The interface we propose offers support for assembly/disassembly operation, replace or delete components inside a virtual scene. To improve the current practice in virtual assembly simulation, our interface is an efficient method for assembly operations and a competitive approach for the current assembly simulation techniques.


References | Cited By  «-- Click to see who has cited this paper

[1] CATIA Handbook: Assembly Design, [Online] Available: Temporary on-line reference link removed - see the PDF document

[2] A. Kulik, "Building on realism and magic for designing 3D interaction techniques", IEEE Comput. Graph. Appl., vol. 29, pp. 22-33, 2009,
[CrossRef] [Web of Science Times Cited 14] [SCOPUS Times Cited 16]


[3] Z. Liu, J. Tan, "Constrained behavior manipulation for interactive assembly in a virtual environment", International Journal of Advanced Manufacturing Technology, vol. 32(7), pp. 797-810, 2007,
[CrossRef] [Web of Science Times Cited 23] [SCOPUS Times Cited 39]


[4] C. Chapman, "The future of user interfaces", 2010, [Online] Available: Temporary on-line reference link removed - see the PDF document

[5] J. Jain, A. Lund, D. Wixon, "The future of natural user interfaces", in Proceedings of the 2011 Annual conference on Human factors in computing systems, pp. 211-214, 2011,
[CrossRef] [SCOPUS Times Cited 22]


[6] H. Zhu, D. Wu, X. Fan, "Interactive assembly tool planning based on assembly semantics in virtual environment", International Journal of Advanced Manufacturing Technology, vol. 51, pp. 739-755, 2010,
[CrossRef] [Web of Science Times Cited 8] [SCOPUS Times Cited 17]


[7] R. D. Vatavu, S. G. Pentiuc, C. Chaillou, L. Grisoni, Samuel Degrande, "Visual Recognition of Hand Postures for Interacting with Virtual Environments", Advances in Electrical and Computer Engineering, vol. 6 (13), pp. 55-58, 2006.

[8] L. Struijk, N. S. Andreasen, "A tongue based control for disabled people", in Proceedings of the 10th international conference on Computers Helping People with Special Needs, pp. 913-918, 2006,
[CrossRef]


[9] A. Karpov, A. Ronzhin, I. Kipyatkova , "An assistive bi-modal user interface integrating multi-channel speech recognition and computer vision", Proceedings of the 14th international conference on Human-computer interaction: interaction techniques and environments, pp. 454-463, 2011.
[CrossRef] [SCOPUS Times Cited 17]


[10] L. Yu, P. Svetachov, P. Isenberg, E. Maarten, T. Isenberg, "FI3D: Direct-Touch Interaction for the Exploration of 3D Scientific Visualization Spaces", IEEE Transactions on Visualization and Computer Graphics, vol. 16(6), pp. 1613-1622, 2010,
[CrossRef] [Web of Science Times Cited 24] [SCOPUS Times Cited 42]


[11] C. Stern, H. Noser, J. Weissmann, P. Stucki, "Application scenarios for scientific visualization and virtual reality using a CAVE infrastructure", Proceedings of the workshop on Virtual environments, pp.319-320, 2003,
[CrossRef] [SCOPUS Times Cited 2]


[12] A. Sepehri, Y. Yacoob, L. S. Davis, "Parametric Hand Tracking for Recognition of Virtual Drawings", Proceedings of the Fourth IEEE International Conference on Computer Vision Systems, pp.6, 2006,
[CrossRef] [SCOPUS Times Cited 3]


[13] Q. Chen, N.D. Georganas, E.M. Petriu, "Hand Gesture Recognition Using Haar-Like Features and a Stochastic Context-Free Grammar", IEEE Transactions on Instrumentation and Measurement, vol. 57(8), pp. 1562-1571, 2008,
[CrossRef] [Web of Science Times Cited 61] [SCOPUS Times Cited 97]


[14] R. Wang, S. Paris, J. Popovici, "6D hands: markerless hand-tracking for computer aided design", Proceedings of the 24th annual ACM symposium on User interface software and technology, pp. 549-558, 2011,
[CrossRef] [SCOPUS Times Cited 113]


[15] L. Sun, U. Klank, M. Beetz, "Eyewatchme: 3D hand and object tracking for inside out activity analysis", First Workshop on Egocentric Vision, Computer Vision and Pattern Recognition, pp. 9-16, 2009,
[CrossRef]


[16] B. Stenger, A. Thayananthan, P. Torr, R. Cipolla, "Model-Based Hand Tracking Using a Hierarchical Bayesian Filter", IEEE Trans. Pattern Anal. Mach. Intell., vol. 28, no.9, pp. 1372-1384, 2006,
[CrossRef] [Web of Science Times Cited 173] [SCOPUS Times Cited 256]


[17] M. Raptis, D. Kirovski, H. Hoppe, "Real-time classification of dance gestures from skeleton animation", Proceedings of the 2011 ACM SIGGRAPH/Eurographics Symposium on Computer Animation, pp. 147-156, 2011,
[CrossRef] [SCOPUS Times Cited 188]


[18] L. R. Rabiner, "A Tutorial on Hidden Markov Models and Selected Applications in Speech Recognition", Proceedings of the IEEE, vol. 77(2), pp.257-286, 1989,
[CrossRef] [Web of Science Times Cited 9211] [SCOPUS Times Cited 12485]


[19] S. Zhong, J. Ghosh, "A New Formulation of Coupled Hidden Markov Models", Dept. of Electrical and Computer Eng., Univ. of Texas at Austin, 2001.

[20] A. D. Wilson, A. F. Bobick, "Recognition and Interpretation of Parametric Gesture", Proceedings of the Sixth International Conference on Computer Vision, pp. 329-336, 1998,
[CrossRef] [Web of Science Times Cited 3]


[21] J. O. Wobbrock, A. D. Wilson, Y. Li, "Gestures without libraries, toolkits or training: a $1 recognizer for user interface prototypes", Proceedings of the 20th annual ACM symposium on User interface software and technology, pp. 159-168, 2007 [invalid ISBN in registered DOI - 10.1145/1294211.1294238]

[22] Y. Li, "Protractor: a fast and accurate gesture recognizer", Proceedings of the 28th international conference on Human factors in computing systems, Atlanta, Georgia, USA, pp. 2169-2172, 2010,
[CrossRef] [SCOPUS Times Cited 84]


[23] S. Malacria, E. Lecolinet, Y. Guiard, "Clutch-free panning and integrated pan-zoom control on touch-sensitive surfaces: the cyclostar approach", Proceedings of the 28th international conference on Human factors in computing systems, pp. 2615-2624, 2010,
[CrossRef] [SCOPUS Times Cited 47]


[24] M. Baglioni, S. Malacria, E. Lecolinet, Y. Guiard ,"Flick-and-brake: finger control over inertial/sustained scroll motion", Proceedings of the 2011 annual conference extended abstracts on Human factors in computing systems, pp. 2281-2286, 2011,
[CrossRef] [SCOPUS Times Cited 12]


[25] S. Goldin-Meadow, "Talking and Thinking With Our Hands", Current Directions in Psychological Science, vol. 15, no. 1, pp. 34-39, 2006,
[CrossRef] [Web of Science Times Cited 36] [SCOPUS Times Cited 40]


[26] S. C. W. Ong, S. Ranganath, Y. V. Venkatesh, "Understanding gestures with systematic variations in movement dynamics", Pattern Recogn., vol. 39, no. 9, pp. 1633-1648, 2006,
[CrossRef] [Web of Science Times Cited 9] [SCOPUS Times Cited 12]


[27] P. Viola, M. Jones, "Rapid Object Detection using a Boosted Cascade of Simple Features", Computer Vision and Pattern Recognition, IEEE Computer Society Conference, vol. 1, pp. 511-518, 2001,
[CrossRef]


[28] R. E. Schapire, "A brief introduction to boosting", in Proceedings of the 16th international joint conference on Artificial intelligence, pp. 1401-1406, 1999.

[29] G. R. Bradski, "Computer vision face tracking for use in a perceptual user interface", Intel Technology Journal, vol. 2, no. 2, pp. 12-21, 1998,
[CrossRef]


[30] S. Xinmeng, "Image Analysis by Moments", Ph.D. Thesis. Department of Electrical and Computer Engineering, Univ. of Manitoba, 1993.

[31] J. Hershberger, J. Snoeyink, "Speeding Up the Douglas-Peucker Line-Simplification Algorithm", Department of Computer Science, Univ. of British Columbia, 1992.

[32] M. Wright, C. Lin, E. O'Neill, D. Cosker, P. Johnson,"3D gesture recognition: an evaluation of user and system performance", Proceedings of the 9th International Conference on Pervasive Computing, San Francisco, USA, vol. 6696 LNCS. Heidelberg: Springer-Verlag, pp. 294-313, 2011,
[CrossRef] [SCOPUS Times Cited 9]


[33] M. Bourne, "Radius of Curvature", [Online] Available: Temporary on-line reference link removed - see the PDF document

[34] M. Kolsch, M. Turk, "Flocks of Features for Tracking Articulated Objects", Real-Time Vision for Human-Computer Interaction, pp. 67-83, 2005,
[CrossRef] [SCOPUS Times Cited 6]


[35] M. Elmezain, A. Al-Hamadi, J. Appenrodt, B. Michaelis, "A Hidden Markov Model-based continuous gesture recognition system for hand motion trajectory", International Conference on Pattern Recognition, pp.519-522, 2008,
[CrossRef]




References Weight

Web of Science® Citations for all references: 9,562 TCR
SCOPUS® Citations for all references: 13,507 TCR

Web of Science® Average Citations per reference: 266 ACR
SCOPUS® Average Citations per reference: 375 ACR

TCR = Total Citations for References / ACR = Average Citations per Reference

We introduced in 2010 - for the first time in scientific publishing, the term "References Weight", as a quantitative indication of the quality ... Read more

Citations for references updated on 2018-04-19 14:46 in 195 seconds.




Note1: Web of Science® is a registered trademark of Clarivate Analytics.
Note2: SCOPUS® is a registered trademark of Elsevier B.V.
Disclaimer: All queries to the respective databases were made by using the DOI record of every reference (where available). Due to technical problems beyond our control, the information is not always accurate. Please use the CrossRef link to visit the respective publisher site.

Copyright ©2001-2018
Faculty of Electrical Engineering and Computer Science
Stefan cel Mare University of Suceava, Romania


All rights reserved: Advances in Electrical and Computer Engineering is a registered trademark of the Stefan cel Mare University of Suceava. No part of this publication may be reproduced, stored in a retrieval system, photocopied, recorded or archived, without the written permission from the Editor. When authors submit their papers for publication, they agree that the copyright for their article be transferred to the Faculty of Electrical Engineering and Computer Science, Stefan cel Mare University of Suceava, Romania, if and only if the articles are accepted for publication. The copyright covers the exclusive rights to reproduce and distribute the article, including reprints and translations.

Permission for other use: The copyright owner's consent does not extend to copying for general distribution, for promotion, for creating new works, or for resale. Specific written permission must be obtained from the Editor for such copying. Direct linking to files hosted on this website is strictly prohibited.

Disclaimer: Whilst every effort is made by the publishers and editorial board to see that no inaccurate or misleading data, opinions or statements appear in this journal, they wish to make it clear that all information and opinions formulated in the articles, as well as linguistic accuracy, are the sole responsibility of the author.




Website loading speed and performance optimization powered by: