Professional Documents
Culture Documents
Abstract—In the proposed work, initially, the YOLO algorithm this algorithm. Vehicles are assumed to have similar fixed
is used to extract and classify objects in a frame. In the sequence sizes and the ground is assumed to be flat [7], which is
of frames, due to various reasons the confidence measure sud- practically not possible since the area of the vehicle depends
denly drops. This changes the class of an object in consecutive
frames which affects the object tracking and counting process on the position of the camera [8]. In the literature, many
severely. To overcome this limitation of the YOLO algorithm, it techniques such as the nearest neighbour, multiple hypothesis
is modified to enable and track the same object efficiently in the tracking, joint probabilistic method are available, where data
sequence of frames. This will in turn increase object tracking is available for different applications, but the drawbacks of the
and counting accuracy. In the proposed work drastic change in above-mentioned algorithms remain as it is. These drawbacks
confidence scores and class change of an object in consecutive
frames are identified by tracking the confidence of a particular include, lack of robustness for the nearest neighbour, the
object in the sequence of frames. These outliers are detected requirement of prior data increases complications for data
and removed using the RANSAC algorithm. After the removal association in joint probabilistic algorithm and multiple hy-
of the outliers, interpolation is applied to get the new confidence pothesis implementation require large computations [9]. As
score at that point. By applying the proposed method a smooth per the author, in traditional methods like template matching
confidence measure variation is obtained across the frames. Using
this, average counting accuracy has been increased from 66% to or kernel-based methods, object detection gives the initial
87% and overall average object classification accuracy is in the position of the object and tracking can be considered a process
range of 94 - 96% for various standard datasets. of detecting a target based on initial information. In the long
Index Terms—Object detection, YOLO, Motion Tracking, range application, it increases computation and thus degrades
RANSAC. the output due to changes caused by shadows and illumination
effects [10]. Careful and smooth use of big data-set, as well
as the powerful Graphics Processing Unit with quality results
I. I NTRODUCTION increases the importance of deep learning in the field of
he object tracking system is an interesting field for computer vision applications like object detection and classifi-
T researchers in different applications like robot navigation,
unmanned aerial vehicle [1], driver assistance with vehicle
cation. The backbone of deep learning is a convolutional neural
network that has an automatic feature extraction which gives
tracking module [2], hand gesture recognition [3] in human- quality results [11]. So from the deep learning colony YOLO
computer interaction and music transcription using finger module is preferred for object detection and classification.
tracking [4]. Recent studies reflect many complex problems ef- The N-YOLO module divides the image instead of resizing
fectively handled by deep learning techniques such as organoid it and the correlation-based tracking algorithm is used after
in vitro model detection and tracking for the reduction of labor merging. The computation time for detection and tracking can
cost and precise surveillance [5]. The advancement of deep be decreased here. But the limitation of the above model is that
learning brings exclusive improvement in object detection and when the tracker gets the object back, new ID gets assigned
improves tracking approach, which includes object extraction and the algorithm is not able to handle the re-entering of
and tracking of the extracted target. Correct object detection the object [12]. YOLOv3 is a modified version of the YOLO
is at high peaks in many applications. Change detection of a algorithm which can detect and classify multiple objects with
particular target through video frames using deep learning de- a single inference and hence the computation time required
velopments has improved a lot in video detection technology. for the same is less and also the accuracy of the algorithm is
To link the same object throughout the frames, a data associ- increases as compared to YOLO [13].
ation algorithm based on spatial information or appearance is Specifically in Vehicle tracking for vehicle counting ap-
enforced here. Output from object detection is used as input plication, YOLO detects target accurately, however it is a
credentials for the tracker [6]. A part-based particle filter is challenging task since 1. The area of the vehicles varies with
implemented using the hidden state of center of the vehicle for the position of camera. 2. Sometimes, the color of the vehicle
vehicle tracking. Here a pre-trained geometric model with rich and color of the road matches and fewer features are available
information from invalid parts makes precise predictions. From due to gray level variation. 3. While entering or leaving the
motion patterns, accuracy of tracking vehicle is improved for video, a little part of the vehicle appears in the video which
can affect the result of classification. 4. If a heavy vehicle
Varsha Kshirsagar Deshpande, Ph.D Scholar,Raghavendra Bhalerao, As- covers a small vehicle, the class of the vehicle changes.
sistant Professor, Manish Chaturvedi, Assistant Professor, Institute of
Infrastructure Technology Research and Management Ahmedabad e- 5. Occlusion and surface variation can affect the output in
mail:varsha.kshirsagar.19pe@iitram.ac.in terms of false-positive value 6. Morphological variation of the
Authorized licensed use limited to: SRM Institute of Science and Technology. Downloaded on January 08,2024 at 04:35:00 UTC from IEEE Xplore. Restrictions apply.
390 IEEE LATIN AMERICA TRANSACTIONS, VOL. 21, NO. 3, MARCH 2023
vehicles between two frames is significant. Due to all these by replacing last layer of alexnet with regression layer for
reasons the confidence score of the detected object reduces the object detection as well localization. With deep multiBox
below the threshold and the class of the object changes for method multiple objects are detected with its localization [27].
such frames which can affect accuracy of the object counting. In the next version sliding window approach with multi-scaling
The vehicles at junctions can take different directions and approach is used for detection, classification and localization
need to be tracked along the entire roundabout to find the [28]. In the upgraded version the image is divided into small
entry and exit. Generally an object starts tracking at entry regions and the application of deep CNN gives feature vector
point and is tracked until exit point and once it crosses a where support vector machine is used for classification [29].
specific entry and exit it is counted in that direction. In such "You Only Look Once (YOLO) is popular and now frequently
case if the confidence score is not identical at entry and exit used module for object detection due to its features [13].
point it will affect the counting accuracy. Based on this, the YOLOv3 gives a trade-off between accuracy and speed [16].
primary research objective of the proposed work is to design an In the field of object detection still results of small object
efficient classification tracking module for the extracted target detection are not satisfactory by learning shallow features at
with an acceptable confidence score in the video streams. the shallow level and deep features at the deep level, the
The initial stage of the work focused on object recognition proposed Multi-Scale feature YOLO learning (MSFYOLO)
and classification, which are taken care by YOLO module. tool is used for better results [30]. YOLOv4 is modified
Influence of variation in confidence score, area, velocity due with denseNet framework which is used to study echocar-
to occlusion and shadow is pulled off by rejecting outliers diographic images for diagnosing congenital heart diseases
using the RANSAC algorithm and these rejected points are (CHDs) [19]. Tracking is to establish association between the
updated by curve fitting using linear interpolation. same target through the video by crucial parameters. Initially
Our main contributions in the proposed work are; 1. Tracing an association between same target is maintained by a point
target object path by importing features such as confidence correspondence and the method is comfortable for the small
score. 2. Incorporate tracing by improving the low confidence objects, where points need to be tracked in every frame. The
score of the extracted classified target with smooth curve fitting deterministic and probabilistic are considered as two types of
using the RANSAC algorithm and linear interpolation. 3. To point tracking. As per literature here, calculation of the missing
improve counting score by rectifying misclassified vehicles. In point is handled by hypothetical point [31]. In another method,
the proposed work section 1 and 2 summarises the introduction the background subtraction is used for target extraction, then
and related work of the proposed module. Section 3 contains centroid trajectory is utilized for establishment of relation
the working flow of the algorithm under the heading of between target for tracking [32]. The Kernel Tracking is the
methodology. Section 4 consists of experiment and result another way of data association which is based on the template
discussion. Section 5 comprises Conclusion. matching or appearance of the target and can be used for
single tracking or the multi-tracking. The kernel correlation
filter utilizes number of training samples and improve quality
II. L ITERATURE S URVEY
of tracker. This method is classified into parametric and
Literature survey reviews development of object detection non parametric. In the parametric distribution at each frame,
and tracking algorithm in last decades. A systematic survey of the target location is upgraded through statistical approach
counting objects from digital images is discussed by dividing such as a mixture of gaussians or its upgraded version with
available techniques and main tools into eight sections [22]. variable parameters and spatial mixture method. The color
Object detection techniques can be broadly divided into tradi- identification in the Hue Saturation Value (HSV) color space
tional approach and deep learning approach. In traditional ap- and active contour models with open source software is used
proach, appearance based method detects object based on color for association and recognition of object in spite of variation
or depth. Whereas in motion based approach, background in size and shape [33]. Expectation maximization is another
subtraction, spatio temporal filtering and optical flow are the mathematical tool used in the literature. In the extended ver-
strategies used to extract object from image. For extraction and sion, methods like poisson distribution, dirichlet distribution
automatic classification of scene images, modified binary local and the regression models are also considered as statistical
and global descriptor is used for effective results [23]. Segmen- model for capturing motion parameters. But to reduce the high
tation of similar regions can be utilized for object extraction computation cost and complexity between the object module
with mean shift clustering and camshift algorithm for object with hypothesized position for a non parametric model, the
tracking which is recognized for good processing efficiency mean shift algorithm is used and the module is upgraded with
and its simplicity [24]. Highlighted background subtraction a weighted histogram. In the modified module a spatial- color
methods from the literature are, frame difference, region based, histogram is used, but here it is required that some part of the
texture based background frame, gaussian mixture based and object should be inside the selected shape whose location is
markov model. These can be upgraded for adaptive nature to defined by previous position of the object. To wipe out such
update background frame after certain interval. In optical flow requirement, the kalman filter or the particle filter can be used
technique image optical flow field is calculated and clustering to predict the location of the target in next frame [34].
is performed as per optical flow distribution of the image [25]. The silhouette tracking is observed when the target is avail-
Introduction of deep learning in the era of the object detection able in the form of complex shape. Here the data association
is as a regression module [26], afterwords it is modified, is obtained through a model, developed using previous frame
Authorized licensed use limited to: SRM Institute of Science and Technology. Downloaded on January 08,2024 at 04:35:00 UTC from IEEE Xplore. Restrictions apply.
DESHPANDE et al.: MODIFIED YOLO MODULE FOR EFFICIENT OBJECT TRACKING 391
TABLE I
S UMMERY OF VARIOUS METHODS
Sr No. YOLO-version Application Remark
A Modified YOLO Model for On-Road Vehicle Detection The ML-based classifier is used to classify Vehicles and 16 convolutional layers of
1 YOLO
in Varying Weather Conditions YOLO has been used along with two fully connected layers at the end [14].
Here clustering of the bounding box of the vehicle logo database, reconstructing
2 YOLOv2 Vehicle Logo Detection Based on Modified YOLOv2 network pre-training and multi-scale detection training are used in a modified
version [15].
An Accurate and Fast Object Detector Using Localization using Gaussian Improvement in accuracy is obtained through Gaussian modeling, loss function
3 YOLOv3
Yolov3 reconstruction and the utilization of localization in YOLOv3network [16].
Here YOLOv3 provides an advanced solution in face recognition for thermal and visible
4 YOLOv3 Robust Thermal-Visible Heterogeneous for face recognition imagery for security purposes also modified CycleGYAN module is used to translate
LWIR images to visible images with good robustness and efficiency [17].
YOLOv4 module along with the image super-resolution module (ISR) and
5 YOLOv4 Face Recognition Approach Based on a Cycle
AI method is used to detect the wearing of a helmet [18].
Here YOLOv4 is used for object detection on street trees where the coverage ratio
Using YOLOv4 version accuracy using Coverage Ratio of Street Trees
7 YOLOv4 can be estimated using parameters. Integration of remote sensing images is utilized
improved.
to improve the accuracy of the coverage ratio [20].
Application of an Improved YOLOv5 Algorithm in Real-Time Detection The small object detection problem of YOLOv5 is improved by the modified
8 YOLOv5
of Foreign Objects by Ground Penetrating Radar. network structure [21]
Authorized licensed use limited to: SRM Institute of Science and Technology. Downloaded on January 08,2024 at 04:35:00 UTC from IEEE Xplore. Restrictions apply.
392 IEEE LATIN AMERICA TRANSACTIONS, VOL. 21, NO. 3, MARCH 2023
Authorized licensed use limited to: SRM Institute of Science and Technology. Downloaded on January 08,2024 at 04:35:00 UTC from IEEE Xplore. Restrictions apply.
DESHPANDE et al.: MODIFIED YOLO MODULE FOR EFFICIENT OBJECT TRACKING 393
START
Read Video
Frames of Video 1:frameno(n)
qk, pk are objects and Ak(n-1) , Akn area of object
in a frame(n) &(n-1)
Is
pk<qk
pk++ qk++
no
Is
classname [pk]=classname[qk]
yes
dist=dist,between object[pk]
and object [qk]
can have more than one class, among which class having
highest confidence score is considered. This information of previous object = pk
tracked as qk
YOLO object detection algorithm drops the confidence score Classification Inaccuracy>0.5
below threshold for that particular frame which is normally Save Class id
Authorized licensed use limited to: SRM Institute of Science and Technology. Downloaded on January 08,2024 at 04:35:00 UTC from IEEE Xplore. Restrictions apply.
394 IEEE LATIN AMERICA TRANSACTIONS, VOL. 21, NO. 3, MARCH 2023
Confidence Score
shadow, illumination .etc
if classification Inaccuracy > 0.5 then 0.6
Class id(n) = class id(n-1) 0.4
Append Confidence Score for tracking curve
0.2 Inlier data
in RANSAC algorithm as a inlier point Outlier data
Else Robust line model by RANSAC
0.0
Outlier point 0 20 40 60 80 100 120 140 160 180
Frame no.
Replace outlier points with 0
Linear interpolation on output Fig. 7. Outlier rejection by RANSAC.
Plot
Smooth Curve with improved confidence score 1.2
end if
1.0
As explained in the YOLO module, confidence score is the
0.8
Confidence Score
multiplication of probability of object and intersection over
union.
0.6
The value of both the factors lies between 0 and 1. Hence the
product of both will give a value ranging between 0 and 1. So 0.4
the maximum value of Confidence score is 1. Threshold value
can be varied to set objectness score, which can be defined as 0.2
probability of presence or absence of an object in bounding 0.0
box. In the given work if its value is greater than 0.5 then it 0 20 40 60 80 100 120 140 160 180
Frame no.
is considered as inlier point otherwise it is considered as an
outlier point as shown in Figure 7. For every frame of video in Fig. 8. Track approximation by RANSAC and Interpolation.
which extracted object travels, confidence score is appended
in vector with the flow of video as in equation (4),
best appropriate curve by rejecting outlier points obtained by
confidence score. From the graph, modified values can be
D. Track Approximation Through RANSAC Algorithm updated for missing and mis-classified frame, Hence tracking
RANSAC algorithm gives a solution to minimize effect of curve can be improved with modified values.
outlier points and gives the best fitted model with inlier points. (
The idea of the algorithm is to propose a model from the ex- > 0.5 Outlier point
ϵ (4)
tracted data with variation in confidence score due to shadow, < 0.5 Inlier Point
occlusion and many more parameters. RANSAC assists to
search a perfect curve from extracted data of confidence score E. Linear Interpolation
by YOLO. We can write, Γt ⊆ Rn where Γt denotes tracking
approximations with inlier points. Measurement Space is given Linear interpolation can identify the nearest point from
by ϕ ⊆ Rn and confidence score countable data with respect the collected information of outlier and inlier points from
to time is considered as Cs ⊆ ϕ for tracking approximations RANSAC approximation. Initially outlier points are replaced
Cs with parameter vector. Measurement Space is given by by zero. Afterwards, linear interpolation is used to fill these
ϕ ⊆ Rn and confidence score countable data with respect to zeroes by appropriate values. By using these values a model
time is considered as Cs ⊆ ϕ. Tracking approximations are is built as shown in Figure 9 and 10. For all inlier points
given by Cs with parameter vector. An outlier point is consid- model is tested. The number of iterations are finalized when
ered as datum that is distant from considered model affected the algorithm gets sufficient number of inlier points to support
by variation in parameters. Following equation distinguishes the model.
inliers and outliers.
q
dθ t(Cs ) = (Cs − Cθ t)T .(Cs − Cθ t) (2) F. Object Tracking/Counting
Where whereCθ t is the orthogonal projection of Cs onto the As mentioned in section II, YOLO provides excellent results
surface defined by model θt . for object detection but in case of classified counting, the
T output gets affected due to sudden change in class which is
θt = [Cs 1, Cs 2, · · Cs n] ⊆ Rn (3)
an impact of decrease in confidence score. After the vehicle is
Data preprocessing produces a data set as shown in equation detected by YOLO, marker lines are inserted at entry or exit
(4), As shown in Figure 8, RANSAC algorithm will give of a given video as shown in Figure 12 for the counting of
Authorized licensed use limited to: SRM Institute of Science and Technology. Downloaded on January 08,2024 at 04:35:00 UTC from IEEE Xplore. Restrictions apply.
DESHPANDE et al.: MODIFIED YOLO MODULE FOR EFFICIENT OBJECT TRACKING 395
0.8
Confidence score
0.6
0.4
0.2
Y-coordinate
200
00 50 100 150 200 250 300 350 400 450 500 550
X-coordinate
1
Fig. 12. Frame with marker lines.
Confidence Score
0.8
0.6
0.4
Y-coordinate
200
00 50 100 150 200 250 300 350 400 450 500 550
X-coordinate
Number of frames
0 200 400 600 800 1000 1200 1400
Intersection of vehicle
1
and marker line
Authorized licensed use limited to: SRM Institute of Science and Technology. Downloaded on January 08,2024 at 04:35:00 UTC from IEEE Xplore. Restrictions apply.
396 IEEE LATIN AMERICA TRANSACTIONS, VOL. 21, NO. 3, MARCH 2023
14
3 0 0 0
Bus
12
10
car Two Wheeler Others
0 12 14 0
Predicted_class
8
6
0 0 14 0 Fig. 18. Confusion matrix by YOLO and Proposed algorithm for
4 OUR Dataset.
0 0 2 6 2
0
TABLE III
Bus Others Two Wheeler car C OMPARISON FOR VARIOUS DATASET
Actual_class
Dataset YOLOv3 Modified YOLO
Fig. 15. Confusion Matrix by YOLO. Class
Bus
Accuracy
1
Precision
1
Recall
1
F1 Score
1
Accuracy Precision
1 1
Recall
1
F1 Score
1
car1 1 1 1 1 1 1 1 1
car2 1 1 1 1 1 1 1 1
car3 1 1 1 1 1 1 1 1
car4 1 1 1 1 1 1 1 1
car5 1 1 1 1 1 1 1 1
car6 1 1 1 1 1 1 1 1
3 0 0 0
Bus
person1 1 1 1 1 1 1 1 1
person2 1 1 1 1 1 1 1 1
person3 1 1 1 1 1 1 1
Two Wheeler
0 0 0 6 apple 1 1 1 1 1 1 1 1
car
TABLE IV
Table III shows experimental analysis of various datasets
C OMPARISON C HART
along with our dataset. Precision, Recall, F-measure and Ac-
curacy these are the four metrics used for analysis of proposed OVERALL ACCURACY
algorithm. Dataset RCNN YOLOv3 YOLOv5 YOLOv6 Modified YOLO
Implementation of the proposed algorithm reflects incre- HUMAN 0.7712 0,7767 0.7982 0.8230 0.9823
SUBWAY 0.9541 0.9513 0.9481 0.9564 0.9904
ments in above mentioned parameters for the extracted ob- KITESURF 0.7984 0.8168 0.8064 0.8423 0.9306
ject throughout the video and increases object counting and OWN 0.8435 0.8864 0.8657 0.8989 0.9743
classification accuracy from average 80% to 96.94%. From
Authorized licensed use limited to: SRM Institute of Science and Technology. Downloaded on January 08,2024 at 04:35:00 UTC from IEEE Xplore. Restrictions apply.
DESHPANDE et al.: MODIFIED YOLO MODULE FOR EFFICIENT OBJECT TRACKING 397
Authorized licensed use limited to: SRM Institute of Science and Technology. Downloaded on January 08,2024 at 04:35:00 UTC from IEEE Xplore. Restrictions apply.
398 IEEE LATIN AMERICA TRANSACTIONS, VOL. 21, NO. 3, MARCH 2023
[20] W. Han, L. Cao, and S. Xu, “A method of the coverage ratio of Varsha Kshirsagar Deshpande persuing Ph.D from
street trees based on deep learning.,” International Journal of Interactive IITRAM Ahmedabad, Working as a Assistant Pro-
Multimedia & Artificial Intelligence, vol. 7, no. 5, 2022. fessor and Head in RMD Sinhgad School of En-
[21] Z. Qiu, Z. Zhao, S. Chen, J. Zeng, Y. Huang, and B. Xiang, “Application gineering, Pune. She has completed Master of En-
of an improved yolov5 algorithm in real-time detection of foreign objects gineering from VESIT Chembur, Mumbai in 2010.
by ground penetrating radar,” Remote Sensing, vol. 14, no. 8, p. 1895, Her research areas are Image Processing, Wireless
2022. Sensor Network, Robotics and its application.
[22] J. G. A. Barbedo, “A review on methods for automatic counting of
objects in digital images,” IEEE Latin America Transactions, vol. 10,
no. 5, pp. 2112–2124, 2012.
[23] S. Cervantes, A. Mexicano, J.-A. Cervantes, R. Rodríguez, and
J. Fuentes-Pacheco, “Binary pattern descriptors for scene classification,”
IEEE Latin America Transactions, vol. 18, no. 01, pp. 83–91, 2020. Raghavendra Bhalerao holds the position of As-
[24] X. Chen, H. Wu, X. Li, X. Luo, and T. Qiu, “Real-time visual object sistant Professor, Electrical and Computer Science
tracking via camshift-based robust framework.,” International Journal Engineering department in Institute of Infrastructure
of Fuzzy Systems, vol. 14, no. 2, 2012. Technology Research and Management (IITRAM).
[25] S. Kumar and J. S. Yadav, “Video object extraction and its tracking He completed M.Tech in Spatial Information Tech-
using background subtraction in complex environments,” Perspectives nology from School of Electronics, Devi Ahliya
in Science, vol. 8, pp. 317–322, 2016. Vishwavidyalaya Indore, in 2010 ( Gold Medallist
[26] C. Szegedy, A. Toshev, and D. Erhan, “Deep neural networks for object ). He received Ph.D.from the Center of Studies
detection,” 2013. in Resources Engineering (CSRE,IIT Bombay) in
[27] D. Erhan, C. Szegedy, A. Toshev, and D. Anguelov, “Scalable object 2016. His area of research are Tri-Stereo Image
detection using deep neural networks,” in Proceedings of the IEEE Analysis, Digital Image processing, Applications of
conference on computer vision and pattern recognition, pp. 2147–2154, IP to Remote Sensing, Medical Image analysis.
2014.
[28] P. Sermanet, D. Eigen, X. Zhang, M. Mathieu, R. Fergus, and Y. Le-
Cun, “Overfeat: Integrated recognition, localization and detection using Manish Chaturvedi is working as a Assistant Pro-
convolutional networks,” arXiv preprint arXiv:1312.6229, 2013. fessor, Electrical and Computer Science Engineering
[29] R. Girshick, J. Donahue, T. Darrell, and J. Malik, “Rich feature department in IITRAM. He completed his MTech
hierarchies for accurate object detection and semantic segmentation,” and Ph.D. from Dhirubhai Ambani Institute of In-
in Proceedings of the IEEE conference on computer vision and pattern formation and Communication Technology (DA-
recognition, pp. 580–587, 2014. IICT). His research interests include the design of
[30] Z. Song, Y. Zhang, Y. Liu, K. Yang, and M. Sun, “Msfyolo: Feature Intelligent Transportation Systems, Embedded Sys-
fusion-based detection for small objects,” IEEE Latin America Transac- tems and IoT, Scalable protocol design for large
tions, vol. 20, no. 5, pp. 823–830, 2022. distributed systems, and the application of ICT for
[31] C. J. Veenman, M. J. Reinders, and E. Backer, “Resolving motion solving problems of societal importance. Recently,
correspondence for densely moving points,” IEEE Transactions on he developed interest in distributed data structures,
Pattern Analysis and Machine Intelligence, vol. 23, no. 1, pp. 54–72, peer to peer content sharing, and Blockchain framework.
2001.
[32] K. Shafique and M. Shah, “A noniterative greedy algorithm for multi-
frame point correspondence,” IEEE transactions on pattern analysis and
machine intelligence, vol. 27, no. 1, pp. 51–65, 2005.
[33] A. S. Silva, F. M. Q. Severgnini, M. L. Oliveira, V. M. S. Mendes,
and Z. A. Peixoto, “Object tracking by color and active contour
models segmentation,” IEEE Latin America Transactions, vol. 14, no. 3,
pp. 1488–1493, 2016.
[34] R. Xia, Y. Chen, and B. Ren, “Improved anti-occlusion object track-
ing algorithm using unscented rauch-tung-striebel smoother and kernel
correlation filter,” Journal of King Saud University-Computer and Infor-
mation Sciences, 2022.
[35] A. Yilmaz, O. Javed, and M. Shah, “Object tracking: A survey,” Acm
computing surveys (CSUR), vol. 38, no. 4, pp. 13–es, 2006.
[36] M. Mandal and S. K. Vipparthi, “An empirical review of deep learning
frameworks for change detection: Model design, experimental frame-
works, challenges and research needs,” IEEE Transactions on Intelligent
Transportation Systems, 2021.
[37] M. Adimoolam, S. Mohan, G. Srivastava, et al., “A novel technique to
detect and track multiple objects in dynamic video surveillance systems,”
2022.
[38] P. Jiang, D. Ergu, F. Liu, Y. Cai, and B. Ma, “A review of yolo algorithm
developments,” Procedia Computer Science, vol. 199, pp. 1066–1073,
2022.
[39] W. Ben, “Real-time photographing and filming detection on mobile
devices,” B.S. thesis, University of Twente, 2022.
Authorized licensed use limited to: SRM Institute of Science and Technology. Downloaded on January 08,2024 at 04:35:00 UTC from IEEE Xplore. Restrictions apply.