Click to open the HelpDesk interface
AECE - Front page banner

Menu:


FACTS & FIGURES

JCR Impact Factor: 0.699
JCR 5-Year IF: 0.674
Issues per year: 4
Current issue: May 2018
Next issue: Aug 2018
Avg review time: 108 days


PUBLISHER

Stefan cel Mare
University of Suceava
Faculty of Electrical Engineering and
Computer Science
13, Universitatii Street
Suceava - 720229
ROMANIA

Print ISSN: 1582-7445
Online ISSN: 1844-7600
WorldCat: 643243560
doi: 10.4316/AECE


TRAFFIC STATS

1,993,772 unique visits
539,162 downloads
Since November 1, 2009



Robots online now
SemrushBot


SJR SCImago RANK

SCImago Journal & Country Rank


SEARCH ENGINES

aece.ro - Google Pagerank




TEXT LINKS

Anycast DNS Hosting
MOST RECENT ISSUES

 Volume 18 (2018)
 
     »   Issue 2 / 2018
 
     »   Issue 1 / 2018
 
 
 Volume 17 (2017)
 
     »   Issue 4 / 2017
 
     »   Issue 3 / 2017
 
     »   Issue 2 / 2017
 
     »   Issue 1 / 2017
 
 
 Volume 16 (2016)
 
     »   Issue 4 / 2016
 
     »   Issue 3 / 2016
 
     »   Issue 2 / 2016
 
     »   Issue 1 / 2016
 
 
 Volume 15 (2015)
 
     »   Issue 4 / 2015
 
     »   Issue 3 / 2015
 
     »   Issue 2 / 2015
 
     »   Issue 1 / 2015
 
 
  View all issues  








LATEST NEWS

2018-Jun-27
Clarivate Analytics published the InCites Journal Citations Report for 2017. The JCR Impact Factor of Advances in Electrical and Computer Engineering is 0.699, and the JCR 5-Year Impact Factor is 0.674.

2017-Jun-14
Thomson Reuters published the Journal Citations Report for 2016. The JCR Impact Factor of Advances in Electrical and Computer Engineering is 0.595, and the JCR 5-Year Impact Factor is 0.661.

2017-Feb-16
With new technologies, such as mobile communications, internet of things, and wide applications of social media, organizations generate a huge volume of data, much faster than several years ago. Big data, characterized by high volume, diversity and velocity, increasingly drives decision making and is changing the landscape of business intelligence, from governments to private organizations, from communities to individuals. Big data analytics that discover insights from evidences has a high demand for computing efficiency, knowledge discovery, problem solving, and event prediction. We dedicate a special section of Issue 4/2017 to Big Data. Prospective authors are asked to make the submissions for this section no later than the 31st of May 2017, placing "BigData - " before the paper title in OpenConf.

Read More »


    
 

  2/2012 - 8

Toward Automatic Recognition of Children's Affective State Using Physiological Parameters and Fuzzy Model of Emotions

SCHIPOR, O.-A. See more information about SCHIPOR, O.-A. on SCOPUS See more information about SCHIPOR, O.-A. on IEEExplore See more information about SCHIPOR, O.-A. on Web of Science, PENTIUC, S.-G. See more information about  PENTIUC, S.-G. on SCOPUS See more information about  PENTIUC, S.-G. on SCOPUS See more information about PENTIUC, S.-G. on Web of Science, SCHIPOR, M.-D. See more information about SCHIPOR, M.-D. on SCOPUS See more information about SCHIPOR, M.-D. on SCOPUS See more information about SCHIPOR, M.-D. on Web of Science
 
Click to see author's profile in See more information about the author on SCOPUS SCOPUS, See more information about the author on IEEE Xplore IEEE Xplore, See more information about the author on Web of Science Web of Science

Download PDF pdficon (722 KB) | Citation | Downloads: 560 | Views: 2,528

Author keywords
assisted speech therapy, emotion recognition, fuzzy model, physiological parameters

References keywords
emotion(11), speech(8), schipor(8), recognition(8), therapy(6), user(5), system(5), pentiuc(5), physiological(4), affect(4)
Blue keywords are present in both the references section and the paper title.

About this article
Date of Publication: 2012-05-30
Volume 12, Issue 2, Year 2012, On page(s): 47 - 50
ISSN: 1582-7445, e-ISSN: 1844-7600
Digital Object Identifier: 10.4316/AECE.2012.02008
Web of Science Accession Number: 000305608000008
SCOPUS ID: 84865304575

Abstract
Quick view
Full text preview
Affective computing - the ability of a system to recognize, understand and simulate human emotional intelligence - is one of the most dynamic fields of HCI - Human Computer Interaction. These characteristics find their applicability in those areas where it is necessary to extend traditional cognitive communication with emotional features. That is why, Computer Based Speech Therapy Systems (CBST), and especially those involving children with speech disorders, require this qualitative shift. So in this paper we propose an original emotional framework recognition as an extension for our previous developed system - Logomon. A fuzzy model is used in order to interpret the values of specific physiological parameters and to obtain the emotional state of the subject. Moreover, an experiment that indicates the emotion pattern (average fuzzy sets) for each therapeutic sequence is also presented. The obtained results encourage us to continue working on automatic emotion recognition and provide important clues regarding the future development of our CBST.


References | Cited By  «-- Click to see who has cited this paper

[1] S. G. Pentiuc, I. Tobolcea, O. A. Schipor, M. Danubianu, M. D. Schipor, "Translation of the Speech Therapy Programs in the Logomon Assisted Therapy System," Advances in Electrical and Computer Engineering, vol. 10, no. 2, pp. 48-52, 2010
[CrossRef] [Full Text] [Web of Science Times Cited 8] [SCOPUS Times Cited 12]


[2] O. A. Schipor, S. G. Pentiuc, M. D. Schipor, "Using a Fuzzy Emotion Model in Computer Assisted Speech Therapy," in Proc. 3rd International Conference on Software, Services and Semantic Technologies, Bourgas, Bulgaria, 2011, pp. 189-194
[CrossRef]


[3] M. Cerlinca, A. Graur, S. G. Pentiuc, T. I. Cerlinca, "Developing a Logopaedic Mobile Device Using a FPGA," in Proc. of SACI '07, Romania, pp. 89-92, 2007.

[4] O. A. Schipor, S. G. Pentiuc, M. D. Schipor, "Improving Computer Based Speech Therapy Using a Fuzzy Expert System," Computing and Informatics, vol. 29, no. 2, pp. 303-318, 2010.

[5] M. H. Zaharia, F. Leon, "Speech Therapy Based on Expert System," Advances in Electrical and Computer Engineering, vol. 9, no. 1, pp. 74-77, 2009
[CrossRef] [Full Text] [Web of Science Times Cited 5] [SCOPUS Times Cited 6]


[6] M. D. Schipor, S. G. Pentiuc, O. A. Schipor, "End-User Recommendations on LOGOMON - a Computer Based Speech Therapy System for Romanian Language," Advances in Electrical and Computer Engineering, vol. 10, no. 4, pp. 57-60, 2010
[CrossRef] [Full Text] [Web of Science Times Cited 5] [SCOPUS Times Cited 7]


[7] A. Mehrabian. Nonverbal Communication. Aldine-Atherton, 1972.

[8] S.M. Lajevardi, Z. M. Hussain, "Feature Extraction for Facial Expression Recognition based on Hybrid Face Regions," Advances in Electrical and Computer Engineering, vol. 9, no. 3, pp. 63-67, 2009
[CrossRef] [Full Text] [Web of Science Times Cited 9] [SCOPUS Times Cited 17]


[9] M. Paleari, B. Huet, R. Chellali, "Towards multimodal emotion recognition : A new approach," in Proc. of ACM International Conference on Image and Video Retrieval on Medical & Biological Engineering & Computing, Xi'an, China, 2010, pp. 174-181.
[CrossRef] [SCOPUS Times Cited 14]


[10] E. Mower, A. Metallinou, C. C. Lee, A. Kazemzadeh, C. Busso, S. Lee, S. Narayanan, "Interpreting Ambiguous Emotional Expressions," in Proc. of ACII, Amsterdam, 2009.

[11] G. Castellano, L. Kessous, G. Caridakis, "Emotion Recognition through Multiple Modalities: Face, Body Gesture, Speech," Affect and Emotion in Human-Computer, vol. 4868, pp. 92-103, 2008
[CrossRef] [SCOPUS Times Cited 107]


[12] M. S. Hussain, R. A. Calvo, "A Framework for Multimodal Affect Recognition," in Proc. of HCSNet Perception and Action Workshop, Brisbane, Australia, 2009.

[13] J. Kim, E. André, "Emotion Recognition Based on Physiological Changes in Listening Music," IEEE Trans. on Pattern Analysis and Machine Intelligence, vol. 30, pp. 2067-2083, 2008
[CrossRef] [Web of Science Times Cited 311] [SCOPUS Times Cited 417]


[14] C. L. Lisetti, F. Nasoz, "MAUI: A Multimodal Affective User Interface," in Proc. of the ACM Multimedia International Conference, Juan les Pins, France, 2002.

[15] G. N. Yannakakis, H. P. Martinez, A. Jhala, "Towards affective camera control in games," User Modeling and User-Adapted Interaction, vol. 20, pp. 313-340, 2010
[CrossRef] [Web of Science Times Cited 41] [SCOPUS Times Cited 56]


[16] K. H. Kim, S. W. Bang, S. R. Kim, "Emotion Recognition System Using Short-Term Monitoring of Physiological Signals," in Proc. on Medical & Biological Engineering & Computing, 2004, pp. 419-427.

[17] O. Grigore, V. Velican, "Self-Organizing Maps For Identifying Impaired Speech," Advances in Electrical and Computer Engineering, vol. 11, no. 3, pp. 41-48, 2011
[CrossRef] [Full Text] [Web of Science Times Cited 3] [SCOPUS Times Cited 3]


[18] T. Polzehl, S. Sundaram, H. Ketabdar, M. Wagner, F. Metze, "Emotion Classification in Children's Speech Using Fusion of Acoustic and Linguistic Features," in Proc. of INTERSPEECH-2009, Brighton, UK, pp. 340-343, 2009.

[19] A. Ortony, T. J. Turner, "What's basic about basic emotions?" Psychological Review, vol. 97, pp. 315-331, 1990
[CrossRef] [Web of Science Times Cited 692] [SCOPUS Times Cited 825]


[20] W. Parrot. Emotions in Social Psychology. Philadelphia: Psychology, 2001.

[21] J. A. Russell, "A circumplex model of affect," vol. 39, pp. 1161-1178, 1980.

[22] H. Schlossberg, "Three dimensions of emotion," Psychological review, vol. 61, pp. 81-88, 1954
[CrossRef] [SCOPUS Times Cited 458]


[23] K. De Verena, L. C. Cooper. Managing emotions in mergers and acquisition. Edward Elgar, 2011.

[24] P. Petta, C. Pelachaud, R. Cowie. Emotion-Oriented Systems: The HUMAINE Handbook. Springer, 2011.

[25] H. Leng, Y. Lin, L. A. Zanzi, "An experimental study on physiological parameters toward driver emotion recognition," in Proc. of EHAWC, Beijing, pp. 237-246, 2007.
[CrossRef]


[26] F. Nasoz, C. L. Lissetti, K. Alvarez, N. Finkelstein, "Emotion Recognition from Physiological Signals for User Modeling of Affect," in Proc. of M'2003, Johnstown, 2003.



References Weight

Web of Science® Citations for all references: 1,074 TCR
SCOPUS® Citations for all references: 1,922 TCR

Web of Science® Average Citations per reference: 40 ACR
SCOPUS® Average Citations per reference: 71 ACR

TCR = Total Citations for References / ACR = Average Citations per Reference

We introduced in 2010 - for the first time in scientific publishing, the term "References Weight", as a quantitative indication of the quality ... Read more

Citations for references updated on 2018-07-20 03:52 in 97 seconds.




Note1: Web of Science® is a registered trademark of Clarivate Analytics.
Note2: SCOPUS® is a registered trademark of Elsevier B.V.
Disclaimer: All queries to the respective databases were made by using the DOI record of every reference (where available). Due to technical problems beyond our control, the information is not always accurate. Please use the CrossRef link to visit the respective publisher site.

Copyright ©2001-2018
Faculty of Electrical Engineering and Computer Science
Stefan cel Mare University of Suceava, Romania


All rights reserved: Advances in Electrical and Computer Engineering is a registered trademark of the Stefan cel Mare University of Suceava. No part of this publication may be reproduced, stored in a retrieval system, photocopied, recorded or archived, without the written permission from the Editor. When authors submit their papers for publication, they agree that the copyright for their article be transferred to the Faculty of Electrical Engineering and Computer Science, Stefan cel Mare University of Suceava, Romania, if and only if the articles are accepted for publication. The copyright covers the exclusive rights to reproduce and distribute the article, including reprints and translations.

Permission for other use: The copyright owner's consent does not extend to copying for general distribution, for promotion, for creating new works, or for resale. Specific written permission must be obtained from the Editor for such copying. Direct linking to files hosted on this website is strictly prohibited.

Disclaimer: Whilst every effort is made by the publishers and editorial board to see that no inaccurate or misleading data, opinions or statements appear in this journal, they wish to make it clear that all information and opinions formulated in the articles, as well as linguistic accuracy, are the sole responsibility of the author.




Website loading speed and performance optimization powered by: