Professional Documents
Culture Documents
Technical Seminar
Presentation
on
INTRODUCTION.
PROBLEM DEFINITION.
LITERATURE SURVEY.
BLOCK DIAGRAM.
METHODOLOGY.
RESULT.
CONCLUSION.
REFERENCES.
2
INTRODUCTION
• Humanoid robots are expected to be socially assistive robots.
• Face recognition is an important matter in human–machine interaction.
• Feature extraction is crucial in determining the success of recognition.
• Emotion recognition can be performed based on speech, text, and facial expressions.
• Face recognition and emotion recognition should be implemented as a unit to improve the capacity for human–
robot interaction.
• Deep learning (DL)makes use of artificial neural networks which implements the technical thoughts of the
human brain into a machine.
• CNN algorithm is used to analyze data and oversees feature extraction and classification of images.
• The distance between the object’s face and the robot determines where the human is.
PROBLEM DEFINITION
Implementation of facial recognition on real-time humanoid robot systems is an important aspect of human-machine
interaction. Face and emotion recognition have been treated as independent challenges which hampers a humanoid
robot’s capability to interact with humans. The combination of facial and emotion. Recognition as a single unit in
real-time humanoid robots using convolutional neural network architectures gives better functionality at various
tasks.
BLOCK DIAGRAM
METHODOLOGY
1. Hardware design
Webcam, JX Servo, 60KG, Arduino, Raspberry Pi, and Dot matrix are
used for implementation.
Face images are captured via webcam embedded in the robot’s eyes.
A dot matrix is used to present characters such as lines and circles.
JX Servo functions as the neck of the robot to facilitate motion for
continuous detection.
2. FACE AND EMOTION RECOGNITION ALGORITHM
SYSTEM DESIGN
8
RESULT
the VGG16 architecture has a lower training
loss than AlexNet.
The training losses of the proposed
architecture and AlexNet are 0.011, 0.151 and
0.052, respectively.
the accuracy shows the VGG16 architecture
performed much better than AlexNet.
accuracy of 95% and 86% respectivly.
The cause of the error is the similarity of the
facial features.
9
CONCLUSION
the VGG16model (model B) is superior to other models.
the success rates in detecting and recognizing faces and emotions of 100% and 73%, respectively.
However, the training process is much longer than AlexNet.
The number of layers used is much larger in VGG16.
Both models provide similar results in realtime humanoid robot implementation.
The distance and illumination are also important factorsin recognition.
Necessary to upgrade the size of the training dataset while considering illumination in future
10
REFERENCES
1. D. S. Trigueros, L. Meng, and M. Hartnett, ‘‘Face recognition: From traditional to deep learning methods,’’ 2018,
arXiv:1811.00116.
2. A. E. Omer and A. Khurran, ‘‘Facial recognition using principal component analysis based dimensionality reduction,’’
in Proc. Int. Conf. Comput., Control, Netw., Electron. Embedded Syst. Eng. (ICCNEEE), Sep. 2015, pp. 434–439.
3. H. M. Ebeid, ‘‘Using MLP and RBF neural networks for face recognition: An insightful comparative case study,’’ in
Proc. Int. Conf. Comput. Eng. Syst., Nov. 2011, pp. 123–128.
4. W. S. M. Sanjaya, D. Anggraeni, K. Zakaria, A. Juwardi, and M. Munawwaroh, ‘‘The design of face recognition and
tracking for humanrobot interaction,’’ in Proc. 2nd Int. Conf. Inf. Technol., Inf. Syst. Electr. Eng. (ICITISEE), Nov.
2017, pp. 315–320.
5. Y. Gizlenmistir, ‘‘Design and implementation of real time face tracking humanoid robot,’’ in Proc. 26th Signal Process.
Commun. Appl. Conf. (SIU), May 2018, pp. 1–6.
6. X. Zhao and C. Wei, ‘‘A real-time face recognition system based on the improved LBPH algorithm,’’ in Proc. IEEE 2nd
Int. Conf. Signal Image Process. (ICSIP), Aug. 2017, pp. 72–76. 11
Thanks!
Any questions?
12