Deep Learning based Automatic Facial Emotion Recognition
Keywords:
Face emotion recognition, feature extraction, Deep neural network, Convolutional Neural Network (CNN), Machine learning method.Abstract
Automatic feeling acknowledgment reliant upon look is a captivating assessment field, which has presented and applied in a couple of districts like security, prosperity and in human machine interfaces. The task of video synopsis has been playing a significant role in domain of video surveillance based systems. Various automation-based system relies on the identification of human emotions exhibited in running video frames. In this work we have investigated the methods to enhance the performance of deep learning model for the task of facial emotion recognition. Models are trained on Facial Emotion Recognition (FER) database with a keen focus to identify the optimal learning parameters using fit one cycle policy. Transfer learning is applied on popular models viz VGG and Resnet to identify the delineating decision boundary for human emotions. The model has successfully achieved a permissible classification rate of 70.2 % on FER dataset respectively
Downloads
References
Chu, H.C.; Tsai, W.W.; Liao, M.J.; Chen, Y.M. Facial feeling acknowledgment with change identification for understudies with advanced chemical imbalance in versatile e-learning. Delicate Comput. 2017, 22, 2973-2999.
Chloé, C.; Vasilescu, I.; Devillers, L.; Richard, G.; Ehrette, T. Fear-type emotion recognition for future audio-based surveillance systems. Speech Commun. 2008, 50, 487–503.
Saste, T.S.; Jagdale, S.M. Emotion recognition from speech using MFCC and DWT for security system. In Proceedings of the IEEE 2017 International Conference of Electronics, Communication and Aerospace Technology (ICECA), Coimbatore, India, 20–22 April 2017; pp. 701–704.
Marco, L.; Carcagnì, P.; Distante, C.; Spagnolo, P.; Mazzeo, P.L.; Rosato, A.C.; Petrocchi, S. Computational assessment of facial expression production in ASD children. Sensors 2018, 18, 3993. [CrossRef]
Meng, Q.; Hu, X.; Kang, J.; Wu, Y. On the effectiveness of facial expression recognition for evaluation of urban sound perception. Sci. Total Environ. 2020, 710, 135484.
Ali, M.; Chan, D.; Mahoor, M.H. Going deeper in facial expression recognition using deep neural networks. In Proceedings of the IEEE 2016 IEEE Winter Conference on Applications of Computer Vision (WACV), Lake Placid, NY, USA, 7–10 March 2016.
Liu, P.; Han, S.; Meng, Z.; Tong, Y. Facial expression recognition via a boosted deep belief network. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, Columbus, OH, USA, 23–28 June 2014; pp. 1805–1812.
E. Sariyanidi, H. Gunes, et A. Cavallaro, « Automatic Analysis of Facial Affect: A Survey of Registration, Representation, and Recognition », IEEE Trans. Pattern Anal. Mach. Intell., oct. 2014.
C.-N. Anagnostopoulos, T. Iliou, I. Giannoukos Features and classifiers for emotion recognition from speech: a survey from 2000 to 2011 Artif. Intell. Rev., 43 (2) (2015), pp. 155-177 févr.
L. Shu A Review of Emotion Recognition Using Physiological Signals Sensors, 18 (7) (2018), p. 2074 juill., doi: 10.3390/s18072074.
J. Kołodziej, H. González-Vélez (Eds.), High-Performance Modelling and Simulation for Big Data Applications: Selected Results of the COST Action IC1406 cHiPSet, Springer International Publishing, Cham (2019), pp. 307H. Alkawaz, D. Mohamad, A.H. Basori, T. Saba Blend Shape Interpolation and FACS for Realistic Avatar 3D Res., 6 (1) (2015), p. 6.
P. V. Rouast, M. Adam, et R. Chiong, « Deep Learning for Human Affect Recognition: Insights and New Developments », IEEE Trans. Affect. Comput., p. 1-1, 2018.
C. Shan, S. Gong, P.W. McOwan Facial expression recognition based on Local Binary Patterns: A comprehensive study Image Vis. Comput., 27 (6) (2009), pp. 803-816.
T. Jabid, M.H. Kabir, O. ChaeRobust Facial Expression Recognition Based on Local Directional Pattern ETRI J., 32 (5) (2010), pp. 784-794.
S. Zhang, L. Li, et Z. Zhao, « Facial expression recognition based on Gabor wavelets and sparse representation », in 2012 IEEE 11th International Conference on Signal Processing, oct. 2012, vol. 2, p. 816-819.
R. Gross, I. Matthews, J. Cohn, T. Kanade, S. Baker Multi-PIE Proc. Int. Conf. Autom. Face Gesture Recognit. Int. Conf. Autom. Face Gesture Recognit., 28 (5) (2010), pp. 807-813.
Downloads
Published
How to Cite
Issue
Section
License
This work is licensed under a Creative Commons Attribution-ShareAlike 4.0 International License.
All papers should be submitted electronically. All submitted manuscripts must be original work that is not under submission at another journal or under consideration for publication in another form, such as a monograph or chapter of a book. Authors of submitted papers are obligated not to submit their paper for publication elsewhere until an editorial decision is rendered on their submission. Further, authors of accepted papers are prohibited from publishing the results in other publications that appear before the paper is published in the Journal unless they receive approval for doing so from the Editor-In-Chief.
IJISAE open access articles are licensed under a Creative Commons Attribution-ShareAlike 4.0 International License. This license lets the audience to give appropriate credit, provide a link to the license, and indicate if changes were made and if they remix, transform, or build upon the material, they must distribute contributions under the same license as the original.