Facial Emotion Detection Using Deep Learning
Facial Emotion Detection Using Deep Learning
net/publication/343414057
CITATIONS READS
190 3,443
3 authors, including:
Akriti Jaiswal
Indian Institute of Technology Guwahati
5 PUBLICATIONS 195 CITATIONS
SEE PROFILE
All content following this page was uploaded by Akriti Jaiswal on 24 August 2023.
2
Authorized licensed use limited to: INDIAN INSTITUTE OF TECHNOLOGY GUWAHATI. Downloaded on August 24,2023 at 06:47:25 UTC from IEEE Xplore. Restrictions apply.
Fig. 1: Network Architecture
The datasets primarily vary in the amount, consistency, C. Results using Proposed Model
and cleanness of the images. For example, the FERC-2013
collection has about 32,000 low-resolution images. It can also In emotion detection we are using three steps, i.e., face
be noted that the facial expressions in the JAFFE (i.e. further detection, features extraction and emotion classification using
extended as CK+ ) are posed (i.e. clean), while the FERC-2013 deep learning with our proposed model which gives better
set displays “in the wild” emotions. This makes it harder to result than previous model. In the proposed method, computa-
interpret the images from the FERC 2013 set, but given the tion time reduces, validation accuracy increases and loss also
large size of the dataset, a model’s robustness can be beneficial decreases, and further performance evaluation achieved which
for the diversity. compares our model with previous existing model. We tested
our neural network architectures on FERC-2013 and JAFFE
database which contains seven primary emotions like sad, fear,
B. Training Details happiness, angry, neutral, surprised, disgust.
We train the network using GPU for 100 epochs to ensure Fig.2 shows the proportions of detected emotions in a single
that the precision converges to the optimum. The network will image of FER dataset. Fig.2(a) shows the image, whereas
be trained on a larger set than the one previously described the detected emotion proportions are shown in Fig.2(b). It
in an attempt to improve the model even more. Training will is clearly observable that neutral has higher proportion than
take place with 20,000 pictures from the FERC-2013 dataset other emotions. That means, the emotion detected for this
instead of 9,000 pictures. The FERC-2013 database also uses image (in Fig.2(a)) is neutral. Similarly, Fig.3 show another
newly designed verification (2000 images) and sample sets image and corresponding emotion proportions. From Fig.3(b),
(1000 images). It shows number of emotions in the final testing it is observable that happy emotion has higher proportion than
and validation set after training and testing our model. The others. That suggests that image of Fig.3(a) detects happy
accuracy will be higher on all validation and test sets than in emotions.
previous runs, emphasizing that emotion detection using deep
convolutional neural networks can improve the performance Similarly, performance is evaluated for all the test images
of a network with more information. of the dataset. We have achieved 95 percentage for happy, 75
3
Authorized licensed use limited to: INDIAN INSTITUTE OF TECHNOLOGY GUWAHATI. Downloaded on August 24,2023 at 06:47:25 UTC from IEEE Xplore. Restrictions apply.
Fig. 2: (a) Image, (b) Proportion of emotions.
4
Authorized licensed use limited to: INDIAN INSTITUTE OF TECHNOLOGY GUWAHATI. Downloaded on August 24,2023 at 06:47:25 UTC from IEEE Xplore. Restrictions apply.
TABLE I: Confusion Matrix (%) for emotion detection using proposed model
Emotions Angry Sad Happy Disgust Fear Neutral Surprise
Angry 56 12 3 9 8 11 1
Sad 10 69 2 6 9 2 2
Happy 0 0 95 0 0 3 2
Disgust 7 13 0 63 8 5 4
Fear 9 8 3 2 65 10 3
Neutral 2 1 8 1 7 75 6
Surprise 7 3 11 0 3 8 68
TABLE II: Qualitative assessment of our proposed model for emotion detection
Model (Dataset) Validation accuracy (%) Validation loss Computation time per step (msec.)
100 and 50 epochs respectively shown in TABLE II. The aim [2] E. Correa, A. Jonker,M.Ozo, andR.Stolk, “Emotion recognition using
of the training step is to determine the correct configuration deep convolutional neural networks,” Tech. Report IN4015, 2016.
[3] Y. I. Tian, T. Kanade, and J.F.Cohn, “Recognizing action units for facial
parameters for the neural network which are: number of nodes expression analysis,” IEEE Transactions on pattern analysis and machine
in the hidden layer (HL), rate of learning (LR), momentum intelligence, vol. 23, no. 2, pp. 97–115, 2001.
(Mom), and epoch (Ep). Different combinations of these [4] C. R. Darwin. The expression of the emotions in man and animals. John
Murray, London, 1872.
parameters have been tested to find out how to achieve the [5] P. Ekman and W. V. Friesen. Constants across cultures in the face and
better recognition rate. emotion. Journal of personality and social psychology, 17(2):124, 1971.
From Table II, it is observed that our proposed model shows [6] J. Nicholson, K. Takahashi, and R. Nakatsu. Emotion recognition in
speech using neural networks. Neural computing applications, 9(4):
70.14% average accuracy compared to the 67.02% average 290–296, 2000.
accuracy reported in model B FOR FER dataset. In this case of [7] B. Fasel and J. Luettin. Automatic facial expression analysis: a survey.
JAFFE database, we achieved average accuracy 98.65% which Pattern recognition, 36(1):259–275, 2003.
[8] A. Krizhevsky and G. Hinton. Learning multiple layers of features from
is also higher than model B. tiny images, 2009.
[9] J. Deng, W. Dong, R. Socher, L.-J. Li, K. Li, and L. Fei-Fei. Imagenet: A
VI. C ONCLUSION large-scale hierarchical image database. In Computer Vision and Pattern
Recognition, 2009. CVPR 2009. IEEE Conference on, pages 248–255.
In this paper, we have proposed a deep learning based IEEE, 2009.
facial emotion detection method from image. We discuss [10] A. Krizhevsky, I. Sutskever, and G. E. Hinton. Imagenet classification
with deep convolutional neural networks. In Advances in neural infor-
our proposed model using two different datasets, JAFFE and mation processing systems, pages 1097–1105, 2012.
FERC-2013. The performance evaluation of the proposed [11] Y. Lv, Z. Feng, and C. Xu. Facial expression recognition via deep
facial emotion detection model is carried out in terms of learning. In Smart Computing (SMARTCOMP), 2014 International
Conference on, pages 303–308. IEEE, 2014.
validation accuracy, computational complexity, detection rate, [12] TFlearn. Tflearn: Deep learning library featuring a higher-level api for
learning rate, validation loss, computational time per step. We tensorflow. URL http://tflearn.org/.
analyzed our proposed model using trained and test sample [13] Open Source Computer Vision Face detection using haar cascades. URL
http://docs.opencv.org/master/d7/d8b/tutorialpyfacedetection.html.
images, and evaluate their performance compare to previous [14] P. J. Werbos et al., “Backpropagation through time: what it does and
existing model. Results of the experiment show that the model how to do it,” Proceedings of the IEEE, vol. 78, no. 10, pp. 1550–1560,
proposed is better in terms of the results of emotion detection 1990.
[15] P. Lucey, J. F. Cohn, T. Kanade, J. Saragih, Z. Ambadar, and I.
to previous models reported in the literature. The experiments Matthews. The extended cohn-kanade dataset (ck+): A complete dataset
show that the proposed model is producing state-of-the-art for action unit and emotion-specified expression. In Computer Vision
effects on both two datasets. and Pattern Recognition Workshops (CVPRW), 2010 IEEE Computer
Society Conference on, pages 94–101. IEEE, 2010.
[16] Kaggle. Challenges in representation learning: Facial expression recog-
R EFERENCES nition challenge, 2013.
5
Authorized licensed use limited to: INDIAN INSTITUTE OF TECHNOLOGY GUWAHATI. Downloaded on August 24,2023 at 06:47:25 UTC from IEEE Xplore. Restrictions apply.