Professional Documents
Culture Documents
Mohanty 2019
Mohanty 2019
using Dlib
1
Shruti Mohanty, 1Shruti V Hegde, 1Supriya Prasad, 2J. Manikandan
1
Dept of ECE, PES University, 100-Feet Ring Road, BSK Stage III, Bengaluru - 85, Karnataka, India
2
Dept of ECE, Crucible of Research and Innovation (CORI)
PES University, 100-Feet Ring Road, BSK Stage III, Bengaluru - 85, Karnataka, India
{email : shrutimohanty998@gmail.com, shrutihegde98@gmail.com, supriya.sprasad@gmail.com,
manikandanj@pes.edu}
Alert if drowsiness
Dlib facial Landmarks Facial Signs of detected
detector Frame pre- landmarks
predictor drowsiness
initialized processing mapped
created monitored
cases
(inset text: aspect ratios of both eyes and mouth and drowsiness count)
III. EXPERIMENTAL RESULTS
The results of real-time detection are lower as the model
A. Dataset Description currently works exceedingly well under good to perfect light
conditions like those found in the dataset videos, whereas
Table I provides a description of the 2 datasets used for the real-time testing was performed under a variety of
testing purposes. lighting conditions.
TABLE I: DATASET DESCRIPTION
Future work will focus on enhancing the model to work
Feature YawDD Dataset[10] MRL Eye Dataset[11] under poor to mediocre lighting conditions, and including
Number of 29 84,898 more drowsiness signs such as head nodding for the
videos/images Videos Images
drowsiness detection model.
Number of males 16 33
Number of females 13 4
Actions performed Without talking or Closed or open eyes REFERENCES
talking/singing or
yawning [1] M. Matsugu., K. Mori., Y. Mitari, and Y. Kaneda, “Subject
Resolution 640x480 RGB (24-bit 640x480, 1280x1024, independent facial expression recognition with robust face detection
true colour) 752x480 using a convolutional neural network,” Neural networks: the official
journal of the Intern. Neural Network Society, vol. 16, no. 5-6, pp.
555-559, June 2003.
B. Results Obtained from Comparison [2] P. Lucey, J. Cohn, S. Lucey, I. Matthews, S. Sridharan and K. M.
Prkachin, "Automatically detecting pain using facial actions," in 3rd
Table II below describes the recognition accuracy obtained Intern. Conf. on Affective Computing and Intelligent Interaction and
from two approaches (eye closure and yawn detection), on Workshops, Amsterdam, 2009, pp. 1-8.
using standard datasets and real-time. [3] L. Liu, D. Preotiuc-Pietro., Z.R. Saman, M.E. Moghaddam, and L.H.
Ungar, “Analyzing Personality through Social Media Profile Picture
Real-time computational results were calculated by taking Choice,” Intern. AAAI Conf. on Web and Social Media, Cologne,
the average of 5 trials each of 12 subjects (including 5 males Germany, May 2016, pp. 214.
and 7 females) recorded at different locations. Average [4] P. K. Stanley, T. Jaya Prakash, S. Sibin Lal and P. V. Daniel,
result included cases with and without glasses. Video frames "Embedded based drowsiness detection using EEG signals," IEEE
with instances of 2 states (sleepy and non-sleepy) for every Intern. Conf. on Power, Control, Signals and Instrumentation
Engineering, Chennai, India, Sept. 2017, pp. 2596-2600.
trial. Highest percentage accuracy obtained is 96.71 % for
[5] T. Danisman, I. M. Bilasco, C. Djeraba and N. Ihaddadene, "Drowsy
yawn detection followed by 93.25% for drowsy blink driver detection system using eye blink patterns," Intern. Conf. on
detection. Machine and Web Intelligence, Algiers, Algeria, Oct. 2010, pp. 230-
TABLE II: RESULTS OBTAINED
233.
Recognition Accuracy [6] L. Li, Y. Chen and Z. Li, "Yawning detection for monitoring driver
Features fatigue based on two cameras," 12th Intern. IEEE Conf. on Intelligent
Real Time Dataset Transportation Systems, St. Louis, USA, Oct. 2009, pp. 1-6.
Eye 82.02 % 93.25 % [7] Davis E. King, Dlib-models [Online].
Available: https://github.com/davisking/dlib-models
Yawn 85.44 % 96.71 %
[8] L. Anzalone, Training Alternative Dlib Shape Predictor models using
Python, Oct. 2018. Accessed on July 16, 2019. [Online]. Available:
https://medium.com/datadriveninvestor/training-alternative-dlib-
shape-predictor-models-using-python-d1d8f8bd9f5c
IV. CONCLUSIONS
[9] N. Dalal, B. Triggs “Histogram of oriented gradients for human
In the Dlib approach, the library’s pre-trained 68 facial detection”, IEEE Computer Society Conf. on Computer Vision and
landmark detector is used. The face detector which is based Pattern Recognition, San Diego, USA, June 2005, pp. 63-69.
on Histogram of Oriented Gradients (HOG) was [10] S. Abtahi, M. Omidyeganeh, S. Shirmohammadi, and B. Hariri. 2014.
implemented. The quantitative metric used in the proposed YawDD: a yawning detection dataset, 5th ACM Multimedia Systems
algorithm was the Eye Aspect Ratio (EAR) to monitor the Conf., New York, USA, Mar. 2014, pp. 24-28.
[11] R. Fusek, MRL Eye Dataset, Accessed on May 28th, 2019. [Online].
driver’s blinking pattern and Mouth Aspect Ratio (MAR) to
Available: http://mrl.cs.vsb.cz/eyedataset
determine if the driver yawned in the frames of the
continuous video stream. The average real-time test
accuracies obtained using Dlib for eyes and yawn were
82.02% and 85.44% respectively and 93.25% and 96.71%
respectively for pre-recorded videos.