Optimal Deep Convolutional Neural Network Based Face Detection and Emotion Recognition Model
Keywords:
Computer vision, Deep learning, Face detection, Facial emotion recognition, RMSProp optimizerAbstract
Face detection and emotion recognition are two closely connected tasks in computer vision that include analysing facial images to identify faces and detect the emotions expressed by the individual. Face detection is the way of localizing and locating faces within image or video frames. The objective is to detect the presence and position of faces, by drawing bounding boxes around them. Facial emotion recognition (FER) aims to detect and classify the emotions expressed by individuals based on facial expressions. Typically, this task can be done after face detection, where the faces detected are analysed further for emotional cues. Emotion recognition can be advanced by means of classical deep learning (DL) or machine learning (ML) techniques. Contemporary research on emotion classification has accomplished grand performance over DL based approaches. This article introduces an Optimal Deep Convolutional Neural Network based Face Detection and Emotion Recognition model (ODCNN-FDER) technique. The aim of the ODCNN-FDER technique is to detect faces and identify the existence of different emotions in them. To achieve this, the ODCNN-FDER technique initially employs Multi-Task Cascaded Convolutional Neural Network (MCCNN) model. Next, the fusion based feature extraction process is involved using two DL models namely EfficientNetB3 and InceptionResNetV2. For emotion recognition, Convolutional Attention Gated Recurrent Neural Network (CAGRNN) model is used. Lastly, root mean square propagation (RMSProp) optimizer was exploited for the optimal hyperparameter tuning of the CAGRNN approach. The performance validation of the ODCNN-FDER methodology was tested on the FER-2013 database. The experimental values highlighted the improved face detection and FER results of the ODCNN-FDER technique over other models.
Downloads
References
S. Gupta, P. Kumar, and R. K. Tekchandani, “Facial emotion recognition based real-time learner engagement detection system in online learning context using deep learning models,”. Multimedia Tools and Applications, 82(8), pp.11365-11394, 2023.
M. A. H. Akhand, S. Roy, N. Siddique, M. A. S. Kamal, and T. Shimamura, “Facial emotion recognition using transfer learning in the deep CNN,” Electronics, 10(9), p.1036, 2021.
S. Zhang, X. Pan, Y. Cui, X. Zhao, and L. Liu, L. “Learning affective video features for facial expression recognition via hybrid deep learning,” IEEE Access, 7, pp.32297-32304, 2019.
D. Wu, X. Han, Z. Yang, and R. Wang, “Exploiting transfer learning for emotion recognition under cloud-edge-client collaborations,” IEEE Journal on Selected Areas in Communications, 39(2), pp.479-490, 2020.
A. Jaiswal, A. K. Raju, and S. Deb, “Facial emotion detection using deep learning,” In 2020 international conference for emerging technology (INCET) (pp. 1-5). IEEE, 2020.
L. Schoneveld, A. Othmani, and H. Abdelkawy, “Leveraging recent advances in deep learning for audio-visual emotion recognition,” Pattern Recognition Letters, 146, pp.1-7, 2021.
D. Canedo, and A. J. Neves, “Facial expression recognition using computer vision: A systematic review,” Applied Sciences, 9(21), p.4678, 2019.
D. Ruan, Y. Yan, S. Lai, Z. Chai, C. Shen, and H. Wang, “Feature decomposition and reconstruction learning for effective facial expression recognition,” In Proceedings of the IEEE/CVF conference on computer vision and pattern recognition pp. 7660-7669, 2021.
S. Minaee, M. Minaei, and A. Abdolrashidi, A, “Deep-emotion: Facial expression recognition using attentional convolutional network,” Sensors, 21(9), p.3046, 2021.
T. S. Gunawan, A. Ashraf, B. S. Riza, E. V. Haryanto, R. Rosnelly, M. Kartiwi, and Z. Janin, “Development of video-based emotion recognition using deep learning with Google Colab,” TELKOMNIKA (Telecommunication Computing Electronics and Control), 18(5), pp.2463-2471, 2020.
D. K. Jain, A. K. Dutta, E. Verdú, S. Alsubai, and A. R. W. Sait, “An automated hyperparameter tuned deep learning model enabled facial emotion recognition for autonomous vehicle drivers,” Image and Vision Computing, 133, p.104659, 2023.
M. M. T. Zadeh, M. Imani, and B. Majidi, “Fast facial emotion recognition using convolutional neural networks and Gabor filters,” In 2019 5th Conference on Knowledge Based Engineering and Innovation (KBEI) (pp. 577-581). IEEE, 2019.
M. A. Nasri, M. A. Hmani, A. Mtibaa, D. Petrovska-Delacretaz, M. B. Slima, and A. B. Hamida, “Face emotion recognition from static image based on convolution neural networks,” In 2020 5th International Conference on Advanced Technologies for Signal and Image Processing (ATSIP) (pp. 1-6). IEEE, 2020.
Z. Gao, Y. Li, Y. Yang, X. Wang, N. Dong, and H. D. Chiang, “A GPSO-optimized convolutional neural networks for EEG-based emotion recognition,” Neurocomputing, 380, pp.225-235, 2020.
J. Shao, and Y. Qian, “Three convolutional neural network models for facial expression recognition in the wild,” Neurocomputing, 355, pp.82-92, 2019.
K. Mohan, A. Seal, O. Krejcar, and A. Yazidi, “Facial expression recognition using local gravitational force descriptor-based deep convolution neural networks,” IEEE Transactions on Instrumentation and Measurement, 70, pp.1-12, 2020.
H. Cui, A. Liu, X. Zhang, X. Chen, K. Wang, and X. Chen, “EEG-based emotion recognition using an end-to-end regional-asymmetric convolutional neural network,” Knowledge-Based Systems, 205, p.106243, 2020.
K. Zhang, Z. Zhang, Z. Li, and Y. Qiao, “Joint face detection and alignment using multitask cascaded convolutional networks,” IEEE signal processing letters, 23(10), pp.1499-1503, 2016.
S. A. Shah, G. M. Lakho, H. A. Keerio, M. N. Sattar, G. Hussain, M. Mehdi, R. B. Vistro, E. A. Mahmoud, and H. O. Elansary, “Application of Drone Surveillance for Advance Agriculture Monitoring by Android Application Using Convolution Neural Network,” Agronomy, 13(7), p.1764, 2023.
R. Florez, F. Palomino-Quispe, R. J. Coaquira-Castillo, J. C. Herrera-Levano, T. Paixão, and A. B. Alvarez, “A CNN-Based Approach for Driver Drowsiness Detection by Real-Time Eye State Identification,” Applied Sciences, 13(13), p.7849, 2023.
N. Parveen, P. Chakrabarti, B. T. Hung, and A. Shaik, “Twitter sentiment analysis using hybrid gated attention recurrent network,” Journal of Big Data, 10(1), pp.1-29, 2023.
D. V. Babu, C. Karthikeyan, and A. Kumar, “Performance analysis of cost and accuracy for whale swarm and RMSprop optimizer,” In IOP Conference Series: Materials Science and Engineering, Vol. 993, No. 1, p. 012080, 2020. IOP Publishing.
https://www.kaggle.com/datasets/deadskull7/fer2013
M. F. Alsharekh, “Facial Emotion Recognition in Verbal Communication Based on Deep Learning,” Sensors, 22(16), p.6105, 2022.
Ms. Mohini Dadhe, Ms. Sneha Miskin. (2015). Optimized Wireless Stethoscope Using Butterworth Filter. International Journal of New Practices in Management and Engineering, 4(03), 01 - 05. Retrieved from http://ijnpme.org/index.php/IJNPME/article/view/37
Banerjee, S. ., Chakraborty, S. ., & Mondal, A. C. . (2023). Machine Learning Based Crop Prediction on Region Wise Weather Data. International Journal on Recent and Innovation Trends in Computing and Communication, 11(1), 145–153. https://doi.org/10.17762/ijritcc.v11i1.6084
Joshi, K., Kumar, V., Sundaresan, V., Ashish Kumar Karanam, S., Dhabliya, D., Daniel Shadrach, F., & Ramachandra, A. C. (2022). Intelligent fusion approach for MRI and CT imaging using CNN with wavelet transform approach. Paper presented at the IEEE International Conference on Knowledge Engineering and Communication Systems, ICKES 2022, doi:10.1109/ICKECS56523.2022.10060322 Retrieved from www.scopus.com
Downloads
Published
How to Cite
Issue
Section
License
This work is licensed under a Creative Commons Attribution-ShareAlike 4.0 International License.
All papers should be submitted electronically. All submitted manuscripts must be original work that is not under submission at another journal or under consideration for publication in another form, such as a monograph or chapter of a book. Authors of submitted papers are obligated not to submit their paper for publication elsewhere until an editorial decision is rendered on their submission. Further, authors of accepted papers are prohibited from publishing the results in other publications that appear before the paper is published in the Journal unless they receive approval for doing so from the Editor-In-Chief.
IJISAE open access articles are licensed under a Creative Commons Attribution-ShareAlike 4.0 International License. This license lets the audience to give appropriate credit, provide a link to the license, and indicate if changes were made and if they remix, transform, or build upon the material, they must distribute contributions under the same license as the original.