Abstract

Emotion recognition is defined as identifying human emotion and is directly related to different fields such as human–computer interfaces, human emotional processing, irrational analysis, medical diagnostics, data-driven animation, human–robot communication, and many more. This paper proposes a new facial emotional recognition model using a convolutional neural network. Our proposed model, “ConvNet”, detects seven specific emotions from image data including anger, disgust, fear, happiness, neutrality, sadness, and surprise. The features extracted by the Local Binary Pattern (LBP), region based Oriented FAST and rotated BRIEF (ORB) and Convolutional Neural network (CNN) from facial expressions images were fused to develop the classification model through training by our proposed CNN model (ConvNet). Our method can converge quickly and achieves good performance which the authors can develop a real-time schema that can easily fit the model and sense emotions. Furthermore, this study focuses on the mental or emotional stuff of a man or woman using the behavioral aspects. To complete the training of the CNN network model, we use the FER2013 databases at first, and then apply the generalization techniques to the JAFFE and CK+ datasets respectively in the testing stage to evaluate the performance of the model. In the generalization approach on the JAFFE dataset, we get a 92.05% accuracy, while on the CK+ dataset, we acquire a 98.13% accuracy which achieve the best performance among existing methods. We also test the system’s success by identifying facial expressions in real-time. ConvNet consists of four layers of convolution together with two fully connected layers. The experimental results show that the ConvNet is able to achieve 96% training accuracy which is much better than current existing models. However, when compared to other validation methods, the suggested technique was more accurate. ConvNet also achieved validation accuracy of 91.01% for the FER2013 dataset. We also made all the materials publicly accessible for the research community at: https://github.com/Tanoy004/Emotion-recognition-through-CNN.

Details

Title
Four-layer ConvNet to facial emotion recognition with minimal epochs and the significance of data diversity
Author
Debnath, Tanoy 1 ; Reza, Md. Mahfuz 1 ; Rahman, Anichur 2 ; Beheshti, Amin 3 ; Band, Shahab S. 4 ; Alinejad-Rokny, Hamid 5 

 Mawlana Bhashani Science and Technology University, Department of Computer Science and Engineering, Tangail, Bangladesh (GRID:grid.443019.b) (ISNI:0000 0004 0479 1356) 
 Mawlana Bhashani Science and Technology University, Department of Computer Science and Engineering, Tangail, Bangladesh (GRID:grid.443019.b) (ISNI:0000 0004 0479 1356); National Institute of Textile Engineering and Research (NITER), Constituent Institute of Dhaka University, Department of Computer Science and Engineering, Savar, Bangladesh (GRID:grid.8198.8) (ISNI:0000 0001 1498 6059) 
 Macquarie University, Department of Computing, Sydney, Australia (GRID:grid.1004.5) (ISNI:0000 0001 2158 5405) 
 National Yunlin University of Science and Technology, Future Technology Research Center, CollegeofFuture, Douliou, Taiwan (GRID:grid.412127.3) (ISNI:0000 0004 0532 0820) 
 The Graduate School of Biomedical Engineering, UNSW Sydney, BioMedical Machine Learning Lab (BML), Sydney, Australia (GRID:grid.1005.4) (ISNI:0000 0004 4902 0432); UNSW Data Science Hub, The University of New South Wales (UNSW Sydney), Sydney, Australia (GRID:grid.1005.4) (ISNI:0000 0004 4902 0432); AI-Enabled Processes (AIP) Research Centre, Macquarie University, Health Data Analytics Program, Department of Computing, Sydney, Australia (GRID:grid.1004.5) (ISNI:0000 0001 2158 5405) 
Publication year
2022
Publication date
2022
Publisher
Nature Publishing Group
e-ISSN
20452322
Source type
Scholarly Journal
Language of publication
English
ProQuest document ID
2656454676
Copyright
© Crown 2022. This work is published under http://creativecommons.org/licenses/by/4.0/ (the “License”). Notwithstanding the ProQuest Terms and Conditions, you may use this content in accordance with the terms of the License.