Professional Documents
Culture Documents
Presented By:
Sudesha Basu Majumder Parichay Singh Sreyan Ghosh
19BEE0228 19BEE0229 19BEE0232
1. Problem Statement
Even though humans have had an easy time detecting emotions from
facial expressions, performing the same feat with a computer program
is rather difficult. It is now possible to discern emotions from photos
because of recent advances in computer vision and machine learning.
In this paper, we propose a unique facial emotion identification
methodology based on convolutional neural networks.
2. Introduction
Other uses of Affectiva include helping children with autism, helping blind
people read facial expressions, helping the robot interact more with people
intelligently, and monitor attention signs while driving to improve driver
safety.
3. Literature Survey
For still images, the probability of predicted expressions can be given. This
feature has not been included in our project. Our goal was to predict the
expression of a face in the image as accurately as possible. The higher the
test accuracy, the better our model will perform in the real world.
Data
The data consists of 48x48 pixel grayscale images of faces. The faces have
been automatically registered so that the face is more or less centered and
occupies about the same amount of space in each image. The task is to
categorize each face based on the emotion shown in the facial expression into
one of seven categories (0=Angry, 1=Disgust, 2=Fear, 3=Happy, 4=Sad,
5=Surprise, 6=Neutral).
Fig 1.
Model
The model being used is going to be created using Tensorflow 2.0 and Keras.
It is a convolutional neural network (CNN) consisting of 4 Conv2D layers
followed by MaxPooling2D layers and 2 Dense layers at the end to output the
probability distribution. The activation function in the Conv2D layers is ReLU
and in the last dense layer is softmax. We will take the padding of images into
consideration as a hyperparameter and see if the model performs better with
or without it.
Fig 2.
Algorithm
Workflow
5. Future Work
Moving forward on the groundwork laid down by our project, we can, in the
future, aim towards:
● Making the prediction model more robust and have lower latency.
● Increasing the data fed into the model and incorporating deeper neural
networks to make more accurate predictions.
● Deploying the model to edge devices or on the web to be of use easily.
Fig 3.
6. References
● Pourmirzaei, M., Esmaili, F. and Montazer, G.A., 2021. Using Self-Supervised Co-Training to
Improve Facial Representation. arXiv preprint arXiv:2105.06421.
● Goodfellow, I.J., Erhan, D., Carrier, P.L., Courville, A., Mirza, M., Hamner, B., Cukierski, W.,
Tang, Y., Thaler, D., Lee, D.H. and Zhou, Y., 2013, November. Challenges in representation
learning: A report on three machine learning contests. In International conference on neural
information processing (pp. 117-124). Springer, Berlin, Heidelberg.
● Zhou, H., Meng, D., Zhang, Y., Peng, X., Du, J., Wang, K. and Qiao, Y., 2019, October.
Exploring emotion features and fusion strategies for audio-video emotion recognition. In 2019
International Conference on Multimodal Interaction (pp. 562-566).
● Meng, D., Peng, X., Wang, K. and Qiao, Y., 2019, September. Frame attention networks for
facial expression recognition in videos. In 2019 IEEE International Conference on Image
Processing (ICIP) (pp. 3866-3870). IEEE.
● Shi, J. and Zhu, S., 2021. Learning to Amend Facial Expression Representation via De-albino
and Affinity. arXiv preprint arXiv:2103.10189.
● T. Vo, G. Lee, H. Yang and S. Kim, "Pyramid With Super Resolution for In-the-Wild Facial
Expression Recognition," in IEEE Access, vol. 8, pp. 131988-132001, 2020, doi:
10.1109/ACCESS.2020.3010018.
● Acharya, D., Huang, Z., Pani Paudel, D. and Van Gool, L., 2018. Covariance pooling for facial
expression recognition. In Proceedings of the IEEE Conference on Computer Vision and
Pattern Recognition Workshops (pp. 367-374).
● Burkert, P., Trier, F., Afzal, M.Z., Dengel, A. and Liwicki, M., 2015. Dexpression: Deep
convolutional neural network for expression recognition. arXiv preprint arXiv:1509.05371.
● Ming, Z., Xia, J., Luqman, M.M., Burie, J.C. and Zhao, K., 2019. Dynamic multi-task learning
for face recognition with facial expression. arXiv preprint arXiv:1911.03281.
● Wang, K., Peng, X., Yang, J., Meng, D. and Qiao, Y., 2020. Region attention networks for
pose and occlusion robust facial expression recognition. IEEE Transactions on Image
Processing, 29, pp.4057-4069.
● Minaee, S., Minaei, M. and Abdolrashidi, A., 2021. Deep-emotion: Facial expression
recognition using attentional convolutional network. Sensors, 21(9), p.3046.
● Gacav, C., Benligiray, B. and Topal, C., 2017, March. Greedy search for descriptive spatial
face features. In 2017 IEEE International Conference on Acoustics, Speech and Signal
Processing (ICASSP) (pp. 1497-1501). IEEE.
● Bulat, A., Cheng, S., Yang, J., Garbett, A., Sanchez, E. and Tzimiropoulos, G., 2021.
Pre-training strategies and datasets for facial representation learning. arXiv preprint
arXiv:2103.16554.
● ICEMI, Yangzhou, 2017 Human face detection algorithm via Haar cascade classifier
combined with three additional classifiers: L. Cuimei, Q. Zhiliang, J. Nan and W. Jianhua