Click to open the HelpDesk interface
AECE - Front page banner

Menu:


FACTS & FIGURES

JCR Impact Factor: 0.699
JCR 5-Year IF: 0.674
Issues per year: 4
Current issue: May 2019
Next issue: Aug 2019
Avg review time: 82 days


PUBLISHER

Stefan cel Mare
University of Suceava
Faculty of Electrical Engineering and
Computer Science
13, Universitatii Street
Suceava - 720229
ROMANIA

Print ISSN: 1582-7445
Online ISSN: 1844-7600
WorldCat: 643243560
doi: 10.4316/AECE


TRAFFIC STATS

2,256,816 unique visits
592,532 downloads
Since November 1, 2009



No robots online now


SJR SCImago RANK

SCImago Journal & Country Rank


SEARCH ENGINES

aece.ro - Google Pagerank




TEXT LINKS

Anycast DNS Hosting
MOST RECENT ISSUES

 Volume 19 (2019)
 
     »   Issue 2 / 2019
 
     »   Issue 1 / 2019
 
 
 Volume 18 (2018)
 
     »   Issue 4 / 2018
 
     »   Issue 3 / 2018
 
     »   Issue 2 / 2018
 
     »   Issue 1 / 2018
 
 
 Volume 17 (2017)
 
     »   Issue 4 / 2017
 
     »   Issue 3 / 2017
 
     »   Issue 2 / 2017
 
     »   Issue 1 / 2017
 
 
 Volume 16 (2016)
 
     »   Issue 4 / 2016
 
     »   Issue 3 / 2016
 
     »   Issue 2 / 2016
 
     »   Issue 1 / 2016
 
 
  View all issues  








LATEST NEWS

2018-May-31
Starting today, the minimum number a pages for a paper is 8, so all submitted papers should have 8, 10 or 12 pages. No exceptions will be accepted.

2018-Jun-27
Clarivate Analytics published the InCites Journal Citations Report for 2017. The JCR Impact Factor of Advances in Electrical and Computer Engineering is 0.699, and the JCR 5-Year Impact Factor is 0.674.

2017-Apr-04
We have the confirmation Advances in Electrical and Computer Engineering will be included in the EBSCO database.

2017-Feb-16
With new technologies, such as mobile communications, internet of things, and wide applications of social media, organizations generate a huge volume of data, much faster than several years ago. Big data, characterized by high volume, diversity and velocity, increasingly drives decision making and is changing the landscape of business intelligence, from governments to private organizations, from communities to individuals. Big data analytics that discover insights from evidences has a high demand for computing efficiency, knowledge discovery, problem solving, and event prediction. We dedicate a special section of Issue 4/2017 to Big Data. Prospective authors are asked to make the submissions for this section no later than the 31st of May 2017, placing "BigData - " before the paper title in OpenConf.

2017-Jan-30
We have the confirmation Advances in Electrical and Computer Engineering will be included in the Gale database.

Read More »


    
 

  1/2018 - 18
View TOC | « Previous Article | Next Article »

The Passive Operating Mode of the Linear Optical Gesture Sensor

CZUSZYNSKI, K. See more information about CZUSZYNSKI, K. on SCOPUS See more information about CZUSZYNSKI, K. on IEEExplore See more information about CZUSZYNSKI, K. on Web of Science, RUMINSKI, J. See more information about  RUMINSKI, J. on SCOPUS See more information about  RUMINSKI, J. on SCOPUS See more information about RUMINSKI, J. on Web of Science, WTOREK, J. See more information about WTOREK, J. on SCOPUS See more information about WTOREK, J. on SCOPUS See more information about WTOREK, J. on Web of Science
 
Click to see author's profile in See more information about the author on SCOPUS SCOPUS, See more information about the author on IEEE Xplore IEEE Xplore, See more information about the author on Web of Science Web of Science

Download PDF pdficon (1,858 KB) | Citation | Downloads: 215 | Views: 656

Author keywords
gesture recognition, human computer interaction, photodiodes, interactive system, wearable sensors

References keywords
gesture(26), sensor(15), recognition(12), sensors(6), ruminski(6), interaction(6), devices(6), czuszynski(6), comput(6), wtorek(5)
Blue keywords are present in both the references section and the paper title.

About this article
Date of Publication: 2018-02-28
Volume 18, Issue 1, Year 2018, On page(s): 145 - 156
ISSN: 1582-7445, e-ISSN: 1844-7600
Digital Object Identifier: 10.4316/AECE.2018.01018
Web of Science Accession Number: 000426449500018
SCOPUS ID: 85043282873

Abstract
Quick view
Full text preview
The study evaluates the influence of natural light conditions on the effectiveness of the linear optical gesture sensor, working in the presence of ambient light only (passive mode). The orientations of the device in reference to the light source were modified in order to verify the sensitivity of the sensor. A criterion for the differentiation between two states - possible gesture and no gesture - was proposed. Additionally, different light conditions and possible features were investigated, relevant for the decision of switching between the passive and active modes of the device. The criterion was evaluated based on the specificity and sensitivity analysis of the binary ambient light condition classifier. The elaborated classifier predicts ambient light conditions with the accuracy of 85.15%. Understanding the light conditions, the hand pose can be detected. The achieved accuracy of the hand poses classifier trained on the data obtained in the passive mode in favorable light conditions was 98.76%. It was also shown that the passive operating mode of the linear gesture sensor reduces the total energy consumption by 93.34%, resulting in 0.132mA. It was concluded that optical linear sensor could be efficiently used in various lighting conditions.


References | Cited By  «-- Click to see who has cited this paper

[1] J. Ioannidis, G. Q. M. Jr, I. Ben-shaul, M. Levedopoulos, M. Liu, "Porting AIX onto the Student Electronic Notebook," Newsl. ACM SIGSMALL/PC Notes. vol. 17, pp. 5-11, 1991.
[CrossRef]


[2] T. Starner, S. Mann, B. Rhodes, J. Healey, K.B. Russell, J. Levine, A. Pentland, "Wearable Computing and Augmented Reality," MIT Media Lab Vision and Modeling Group Technical Report, 1995, vol. 355. [Online] Available: Temporary on-line reference link removed - see the PDF document

[3] K. Hinckley, J. Pierce, M. Sinclair, E. Horvitz, "Sensing Techniques for Mobile Interaction," in UIST ’00 Proc. 13th Annu. ACM Symp. User Interface Softw. Technol., San Diego, 2000, pp. 91-100.
[CrossRef]


[4] C. Metzger, M. Anderson, T. Starner, "FreeDigiter?: A Contact - free Device for Gesture Control," in ISWC ’04 Proc. Eighth Int. Symp. Wearable Comput., Washington, 2004, pp. 18-21.
[CrossRef] [Web of Science Times Cited 14]


[5] A. Pasarica, C. Miron, D. Arotari?ei, Gladiola Andruseac, H. Costin, C. Rotariu, "Remote Control of a Robotic Platform Based on Hand Gesture Recognition", Proc. of the IEEE 6th Int. Conference on E-Health and Bioengineering – "EHB 2017", Sinaia, Romania, 22-24 June 2017, pp. 643-646.
[CrossRef]


[6] T. Kopinski, U. Handmann, "Touchless Interaction for Future Mobile Applications," in Int. Conf. Comput. Netw. Commun. (ICNC), Work. Comput. Netw. Commun. (CNC), Kauai, 2016.
[CrossRef]


[7] H. Lahiani, M. Elleuch, "Real Time Hand Gesture Recognition System for Android Devices," in 15th Int. Conf. Intell. Syst. Des. Appl., 2015, pp. 591-596.
[CrossRef]


[8] K. Oh, S.-H. Hwang, S. You, J. Cho, M. Jeon, M.-K. Kim, "Gesture Sensor for Mobile Devices," System LSI Business, 2013, pp. 1-12. [Online] Available: Temporary on-line reference link removed - see the PDF document

[9] C. Zhang, J. Tabor, J. Zhang, X. Zhang, "Extending Mobile Interaction Through Near-Field Visible Light Sensing," in Proc. 21st Annu. Int. Conf. Mob. Comput. Netw. - MobiCom ’15, Paris, 2015, pp. 345-357.
[CrossRef]


[10] Y. Kim, K. Baek, "A motion gesture sensor using photodiodes with limited field-of-view", in Opt. Express. vol. 21, no. 8, pp. 555-560, 2013.
[CrossRef] [Web of Science Times Cited 9]


[11] J.S. Kim, S.J. Yun, Y.S. Kim, "Low-power motion gesture sensor with a partially open cavity package," Opt. Express. vol. 24, no. 10, pp. 10537-10546, 2016.
[CrossRef] [Web of Science Times Cited 3]


[12] G. Batchuluun, B. Odgerel, C.H. Lee, "Hand Gesture Recognition Using an Infrared Proximity Sensor Array," Int. J. Fuzzy Log. Intell. Syst. 15, 2015, pp. 186-191.
[CrossRef]


[13] A. Withana, R. Peiris, N. Samarasekara, S. Nanayakkara, "zSense?: Enabling Shallow Depth Gesture Recognition for Greater Input Expressivity on Smart Wearables," in CHI ’15 Proc. 33rd Annu. ACM Conf. Hum. Factors Comput. Syst., 2015, pp. 3661-3670.
[CrossRef] [Web of Science Times Cited 17]


[14] Z. Zivkovic, "Air Gesture Control Using 5-Pixel Light Sensor", in Consumer Electronics (ICCE), 2014 IEEE Int. Conf. on, Las Vegas, 2014.
[CrossRef]


[15] J. Kim, J. He, K. Lyons, T. Starner, "The Gesture Watch: A wireless contact-free Gesture based wrist interface", in Proc. of the Int. Symp. on Wearable Computers, ISWC, Boston, 2007, pp. 15-22.
[CrossRef]


[16] C. Chuang, T. Chang, P. Jau, F. Chang, "Touchless Positioning System Using LED Sensors", in IEEE Int. Conf. on System Science and Engineering (ICSSE), 2014, Shanghai, pp. 1-6.
[CrossRef]


[17] S. Tang, W. Tseng, W. Luo, K. Chiu, S. Lin, Y. Liu, "Virtual Mouse : A Low Cost Proximity-Based Gestural Pointing Device", in Human-Computer Interaction. Interaction Techniques and Environments. HCI 2011. Lecture Notes in Computer Science, vol. 6762. Springer, 2011, pp. 491-499.
[CrossRef]


[18] A. Butler, S. Izadi, S. Hodges, "SideSight: multi-"touch" interaction around small devices", in Proc. of the 21st annual ACM Symp. on User interface software and technology (UIST), 2008, Monterey, pp. 201-204.
[CrossRef]


[19] T. Fan, C. Ma, Z. Gu, Q. Lv, J. Chen, D. Ye, et al., "Wireless Hand Gesture Recognition Based on Continuous-Wave Doppler Radar Sensors," IEEE Trans. Microw. Theory Tech. vol. 64, no. 11, pp. 4012-4020, 2016.
[CrossRef] [Web of Science Times Cited 20]


[20] T. Fan, D. Ye, J. Hangfu, Y. Sun, C. Li, L. Ran, "Hand Gesture Recognition Based on Wi-Fi Chipsets," in Radio Wirel. Symp. (RWS), 2017 IEEE, Phoenix, 2017, pp. 98-100.
[CrossRef]


[21] S. Kim, H. Han, J. Kim, S. Lee, T. Kim, "A Hand Gesture Recognition Sensor Using Reflected Impulses", in IEEE Sensors Journal, 2017, vol. 17, issue 10, pp. 2975-2976.
[CrossRef] [Web of Science Times Cited 9]


[22] M. Tang, F. Wang, T. Horng, "Human Gesture Sensor Using Ambient Wireless Signals Based on Passive Radar Technology," in Microw. Symp. (IMS), 2015 IEEE MTT-S Int., Phoenix, 2015, pp. 2-5.
[CrossRef]


[23] F. Wang, M. Tang, Y. Chiu, T. Horng, "Gesture Sensing Using Retransmitted Wireless Communication Signals Based on Doppler Radar Technology," IEEE Trans. Microw. Theory Tech. vol. 63, no. 12, pp. 4592-4602, 2015.
[CrossRef] [Web of Science Times Cited 13]


[24] R. Bainbridge, J.A. Paradiso, "Wireless Hand Gesture Capture Through Wearable Passive Tag Sensing," in Body Sens. Networks (BSN), 2011 Int. Conf., Dallas, 2011, pp. 200-204.
[CrossRef]


[25] R. Krigslund, P. Popovski, G.F. Pedersen, "3D Gesture Recognition Using Passive RFID Tags," in Antennas Propag. Soc. Int. Symp. (APSURSI), 2013 IEEE, Orlando, 2013, pp. 2307-2308.
[CrossRef]


[26] B. Kellogg, V. Talla, S. Gollakota, "Bringing Gesture Recognition To All Devices," in Proceeding NSDI’14 Proc. 11th USENIX Conf. Networked Syst. Des. Implement., Seattle, 2014, pp. 303-316. [Online] Available: Temporary on-line reference link removed - see the PDF document

[27] F. Aezinia, Y. Wang, B. Bahreyni, "Touchless Capacitive Sensor for Hand Gesture Detection," in SENSORS, 2011 IEEE, Limerick, 2011, pp. 3-6.
[CrossRef]


[28] M. Pulkkinen, J. Salomaa, M. M. Moayer, T. Haapala, K. Halonen, "462-nW 2-Axis Gesture Sensor Interface Based On Capacitively Controlled Ring Oscillators", in Circuits and Systems (ISCAS), IEEE Int. Symp. on, Baltimore, 2017.
[CrossRef]


[29] P. Wojtczuk, A. Armitage, T.D. Binnie, T. Chamberlain, "PIR Sensor Array for Hand Motion Recognition," in SENSOR DEVICES 2011 Second Int. Conf. Sens. Device Technol. Appl., 2011, pp. 99-102.

[30] P. Wojtczuk, D. Binnie, A. Armitage, T. Chamberlain, C. Giebeler, "A Touchless Passive Infrared Gesture Sensor," in UIST ’13 Adjun. Proc. 26th Annu. ACM Symp. User Interface Softw. Technol., St. Andrews, 2013, pp. 67-68.
[CrossRef]


[31] A. Holmes, S. Desai, A. Nahapetian, "LuxLeak?: Capturing Computing Activity Using Smart Device Ambient Light Sensors," in Proceeding SmartObjects ’16 Proc. 2nd Work. Exp. Des. Implement. Smart Objects, New York, 2016, pp. 47-52.
[CrossRef]


[32] M. Kaholokula, "Reusing Ambient Light to Recognize Hand Gestures," Dartmouth College, Tech. Rep., 2016. [Online] Available: Temporary on-line reference link removed - see the PDF document

[33] K. Czuszynski, J. Ruminski, J. Wtorek, A. Vogl, M. Haller, "Interactions using passive optical proximity detector," in Hum. Syst. Interact. (HSI), 2015 8th Int. Conf., IEEE Xplore, Warsaw, 2015, pp. 180-186.
[CrossRef]


[34] A. Bujnowski, K. Czuszynski, J. Ruminski, J. Wtorek, R. McCall, A. Popleteev, N. Louveton, T. Engel, "Comparison of active proximity radars for the wearable devices," in Hum. Syst. Interact. (HSI), 2015 8th Int. Conf., IEEE Xplore, 2015, pp. 158-165.
[CrossRef]


[35] K. Czuszynski, J. Ruminski, A. Bujnowski, J. Wtorek, "Semi complex navigation with an active optical gesture sensor," in UbiComp ’16 Proc. 2016 ACM Int. Jt. Conf. Pervasive Ubiquitous Comput. Adjun., Heidelberg, 2016, pp. 269-272.
[CrossRef] [Web of Science Times Cited 4]


[36] K. Czuszynski, J. Ruminski, J. Wtorek, "Analysis of the properties of the active linear gesture sensor," Metrol. Meas. Syst. vol. 24. no. 4, 2017, pp. 617-630.
[CrossRef] [Web of Science Times Cited 2]


[37] K. Czuszynski, J. Ruminski, J. Wtorek, "Pose classification in the gesture recognition using the linear optical sensor," in Hum. Syst. Interact. (HSI), 2017 10th Int. Conf., IEEE Xplore, Ulsan, 2017, pp. 18-24.
[CrossRef]


[38] S. Berman, H. Stern, "Sensors for gesture recognition systems," IEEE Trans. Syst. Man Cybern. Part C Appl. and Rev., vol. 42, no. 3, pp. 277-290, 2012.
[CrossRef] [Web of Science Times Cited 33]


[39] K. Czuszynski, J. Ruminski, T. Kocejko, "Septic safe interactions with smart glasses in health care", in 37th Annual Int. Conf. of the IEEE Engineering in Medicine and Biology Society (EMBC), 2015, Milano, pp. 1604-1607.
[CrossRef]




References Weight

Web of Science® Citations for all references: 124 TCR
SCOPUS® Citations for all references: 0

Web of Science® Average Citations per reference: 3 ACR
SCOPUS® Average Citations per reference: 0

TCR = Total Citations for References / ACR = Average Citations per Reference

We introduced in 2010 - for the first time in scientific publishing, the term "References Weight", as a quantitative indication of the quality ... Read more

Citations for references updated on 2019-06-17 08:12 in 231 seconds.




Note1: Web of Science® is a registered trademark of Clarivate Analytics.
Note2: SCOPUS® is a registered trademark of Elsevier B.V.
Disclaimer: All queries to the respective databases were made by using the DOI record of every reference (where available). Due to technical problems beyond our control, the information is not always accurate. Please use the CrossRef link to visit the respective publisher site.

Copyright ©2001-2019
Faculty of Electrical Engineering and Computer Science
Stefan cel Mare University of Suceava, Romania


All rights reserved: Advances in Electrical and Computer Engineering is a registered trademark of the Stefan cel Mare University of Suceava. No part of this publication may be reproduced, stored in a retrieval system, photocopied, recorded or archived, without the written permission from the Editor. When authors submit their papers for publication, they agree that the copyright for their article be transferred to the Faculty of Electrical Engineering and Computer Science, Stefan cel Mare University of Suceava, Romania, if and only if the articles are accepted for publication. The copyright covers the exclusive rights to reproduce and distribute the article, including reprints and translations.

Permission for other use: The copyright owner's consent does not extend to copying for general distribution, for promotion, for creating new works, or for resale. Specific written permission must be obtained from the Editor for such copying. Direct linking to files hosted on this website is strictly prohibited.

Disclaimer: Whilst every effort is made by the publishers and editorial board to see that no inaccurate or misleading data, opinions or statements appear in this journal, they wish to make it clear that all information and opinions formulated in the articles, as well as linguistic accuracy, are the sole responsibility of the author.




Website loading speed and performance optimization powered by: