Professional Documents
Culture Documents
BACHELOR OF ENGINEERING
IN
Computer Science
Chandigarh University
February - June 2023
Table of Contents
CHAPTER 3. DESIGN FLOW/PROCESS ...............................................................21
Real-time detection of suspicious activity in video surveillance systems requires quick response times
to detect and alert security personnel about potential threats. CNN models can be computationally
expensive and require a large number of computations to process each input frame, which can be a
significant constraint in resource-constrained environments. The computational cost of a CNN model is
determined by its architecture, size, and complexity. Larger models with more layers and parameters
require more computations, and hence, more computational resources. In resource-constrained
environments, such as embedded systems, IoT devices, or edge computing platforms, the limited
processing power and memory can make it challenging to deploy complex CNN models for real-time
detection.
The model architecture needs to be designed next. In the proposed model, we will use multiple
models and combine their output for the desired result alongside keeping nodes of various
parameters such as motion pattern, object detection, facial recognition, time of the day etc. Some
of the models that will be used are CNN(convolutional neural network), MobileNet model,
ResNet (Residual Network),a deep ensemble CNN architecture. This architecture consists of
three independent CNNs, each with a different architecture and number of layers. The outputs of
these three CNNs are then combined to form the final output of the model. The choice of CNN
architecture depends on the type of data being used and the complexity of the problem being
solved.
Once the model architecture is designed, the model needs to be trained. The training process
involves feeding the training data into the model and adjusting the model's parameters to
minimize the error between the predicted output and the actual output. The training process can
take a long time, depending on the complexity of the model and the size of the data.
After the model is trained, it needs to be evaluated using the testing data. The model's
performance is measured using various metrics, including accuracy, sensitivity, specificity, and
precision. In the proposed model, an accuracy of 66.6%, with an loss of 82% , this can be
improved by increasing the dataset and epochs size. The model will have an flow of program as
given below in the figure:
Finally, the model needs to be deployed. The model can be deployed in various ways, including
as a web application or a mobile application. The deployment process involves integrating the
model into the desired platform and ensuring that it runs efficiently and accurately.
In conclusion, implementing a deep ensemble CNN model for suspicious detection requires
careful planning and execution. It involves collecting and preprocessing the data, designing the
model architecture, training and evaluating the model, and finally deploying the model. The
proposed model achieved high accuracy, sensitivity, specificity, and precision and has the
potential to improve outcomes and early detection of crime.
3.5.1 Block diagram:
TensorFlow is an open-source platform for building and training machine learning models. It
was developed by Google and is used widely in the industry for tasks such as image and speech
recognition, natural language processing, and predictive analytics.
4. Scikit-learn:
Scikit-learn is a free and open-source machine learning library for the Python programming
language. It provides tools for classification, regression, clustering, and dimensionality reduction,
among other tasks.
5. Google Colaboratory:
Google Colaboratory is a free cloud-based platform that provides access to Python programming
environments, including Jupyter Notebooks and TensorFlow. It is widely used by data scientists
and machine learning practitioners to experiment with code and algorithms without the need for
a local machine.
6. DICOM file format:
DICOM (Digital Imaging and Communications in Medicine) is a standard format for medical
images such as X-rays, CT scans, and MRIs. It includes metadata such as patient information,
imaging parameters, and pixel data.
7. LIDC-IDRI dataset:
The Lung Image Database Consortium and Image Database Resource Initiative (LIDC-IDRI) is a
publicly available dataset of CT scans for lung cancer diagnosis. It contains over 1000 CT scans
with annotations for lung nodules by multiple radiologists, making it a valuable resource for
developing and testing machine learning algorithms for lung cancer detection.
8. FER(2013) dataset:
The FER2013 dataset is a widely-used facial expression recognition dataset that consists of
35,887 grayscale images in 48x48-pixel resolution. The images are categorized into seven
emotions: anger, disgust, fear, happiness, sadness, surprise, and neutral. The dataset was created
by Pierre-Luc Carrier and Aaron Courville and made available to the public in 2013. The images
in the dataset were collected from a range of sources, including the internet and manually-labeled
facial expression datasets. FER2013 has become a popular benchmark dataset for evaluating
facial expression recognition models and has been used in numerous studies in the field.
9. OpenCV:
OpenCV is the huge open-source library for computer vision, machine learning, and image
processing and now it plays a major role in real-time operation which is very important in
today’s systems. By using it, one can process images and videos to identify objects, faces, or
even handwriting of a human. When integrated with various libraries, such as NumPy, python is
capable of processing the OpenCV array structure for analysis. To Identify image pattern and its
various features we use vector space and perform mathematical operations on these features.
10. OS module:
The functions OS module provides allows us to operate on underlying Operating System tasks,
irrespective of it being a Windows Platform, Macintosh or Linux. In this lesson, we will review
these functions and what we can do with them. Let us start our journey with these functions and
what information they offer.
11. Kaggle:
Kaggle is a popular platform for data scientists, machine learning engineers, and enthusiasts to
compete in data science challenges, collaborate on projects, and share knowledge. Kaggle
provides a variety of resources and tools for data analysis, such as datasets, Jupyter notebooks,
and competitions, to enable users to explore and analyze data, build models, and share their
results with the community.
12. Github:
GitHub is a web-based platform for version control and collaboration that allows developers to
store and manage their code, as well as track changes and collaborate with other team members.
It provides a variety of tools for developers, including the ability to host and review code,
manage projects and software releases, and collaborate on open-source software development.
With GitHub, developers can work on code together, share feedback, and make changes to code
in a controlled manner.