Advance Driver Warning and Alert System Using R-CNN
Rekha S. Kotwal1 Mukesh Narwade 2
JSPMs’ Bhivarabai Sawant Institute of JSPMs’ Bhivarabai Sawant Institute of Technology and Research Technology and Research rekhaaher21@gmail.com EMAIL
Priyank Nikalje 3 Shaikh Sahil 4
JSPMs’ Bhivarabai Sawant Institute of JSPMs’ Bhivarabai Sawant Institute of Technology and Research Technology and Research EMAIL EMAIL
Shaikh Zaheer 5 JSPMs’ Bhivarabai Sawant Institute of Technology and Research EMAIL
and alert systems have become integral Detector), ADAS(Advanced Driver components of modern vehicles, Assistance Systems) enhancing road safety and driver assistance. This review paper provides an in-depth analysis of the latest advancements in driving warning and alert systems. It covers various technologies, their effectiveness, and their impact on road safety. Additionally, it explores the challenges and future directions in this field. Our review encompasses an analysis of the datasets commonly used for evaluation, model performance metrics, and benchmark results achieved by these models. Furthermore, we discuss the real- world implications of these detection systems, including their potential in enhancing road safety, autonomous navigation, and traffic management.
Keywords: YOLOv5 (You Only Look
I. Introduction
This review paper embarks on an
exploration of three leading object detection models— YOLOv5 (You Only Look Once version 5), R-CNN (Region- based Convolutional Neural Network), and Single Shot MultiBox Detector (SSD)—as they pertain to the tasks of lane, speed bump, and pothole detection. The significance of these tasks cannot be understated; they serve as the foundation upon which autonomous vehicles and advanced driver-assistancesystems (ADAS) rely.
As we delve into the heart of this
examination, we shall unearth the intricacies of each detection model, illuminating their unique architectural paradigms, and highlighting their respective strengths and weaknesses. Through a meticulous survey of the existing literature, we offer a comprehensive overview of how these models have been harnessed for the detection of lanes, speed bumps, and potholes, drawing from both academic research and real-world algorithms are frequently used, lane detection applications. is a crucial problem that needs to be addressed. The main lane feature used by This review paper is more than a mere conventional algorithms is gradient technical exposition; it is a roadmap to information, but they perform poorly when understanding the potential and limitations there are different lighting conditions of these object detection models in shaping present.The research offers a unique the future of transportation. We dissect the technique for illumination-robust lane datasets used for evaluation, scrutinize recognition called the "gradient-enhancing performance metrics, and present conversion method" to address this drawback. benchmark results, providing a holistic view This technique uses linear discriminant of their effectiveness. analysis to convert RGB color images into Furthermore, we pivot our gaze towards the fresh gray-level images. Large gradients at practical implications of these detection lane boundaries are highlighted in the systems. Their deployment has the potential converted photos that are the outcome. The gray-level conversion vector is dynamically to revolutionize road safety, making updated, increasing the system's robustness as autonomous navigation safer and more it adjusts to shifting lighting circumstances reliable while aiding traffic management in [2]. In an Internet of Things-based framework urban environments. However, the road to (IBF), this study presents a unique lane seamless integration is not without its identification and tracking strategy for obstacles, which we shall explore in detail. autonomous vehicles. The Vehicle Board (VB), Cloud Module (CM), and Vehicle II. Literature Survey Remote Controller are the three main modules that make up the IBF. LD and This paper focuses on a crucial feature of tracking are initially handled by the VB, automated and self-driven transportation, lane however in tough circumstances or during and speed breaker identification, which system outages, data is forwarded to the CM strongly depends on the effectiveness of for processing. The results are subsequently detection algorithms in the setting of vehicle- forwarded to the VB so that it may take the mounted cameras. In order to make necessary action, like operating a remote autonomous driving possible, these vehicle in the event of a lane departure.An algorithms' performance is crucial. The study illumination invariance technique for makes use of computer vision and machine knowing lane markings in a range of lighting learning approaches to improve detection situations is also included in the research. effectiveness.The researchers use OpenCV Real-world data-based simulation findings for lane detection and combine a Fully show strong lane-keeping rates of 95.3% in Convolutional Network (FCN) with gradient tunnels and 95.2% on roads [3]. This paper and color thresholds. In spite of changing offers a novel method for improving the environmental conditions, the FCN is skilled precision and stability of trajectory tracking in locating the area of interest, the lane. This control for autonomous cars by combining method guarantees reliable and robust lane lane identification and Model Predictive detection [1]. In advanced driving assistance Control (MPC). The research utilizes image systems (ADAS), where vision-based recognition algorithms, such as dynamic Region of Interest (ROI) extraction, edge algorithms have had trouble in difficult real- detection, and Hough straight line life circumstances due to things like weak identification, to extract lane lines from road markings, obstructions, and guardrails. photographs taken by autonomous vehicles This study introduces a strong multi-lane equipped with cameras.The work uses MPC, detection process as a solution to these which is based on a vehicle dynamics model concerns.To deal with these problems, the with three degrees of freedom, to enable suggested method integrates three crucial accurate tracking of the extracted lane lines. technologies. In order to extract strong lane Additionally, a fuzzy controller that takes into features from photos, even when there are account both yaw angle and yaw rate is used barriers and barely discernible lanes present, to adjust front wheel steering angle. The goal it first uses an adaptive threshold. In order to of this lane identification and control avoid false lane recognition brought on by algorithm integration is to make the vehicle incorrect lane characteristics, an improved more adept at adhering to the specified Random Sample Consensus (RANSAC) trajectory [4].One of the major research technique is introduced. This approach makes difficulties brought on by the quick use of feedback from past lane curvature data development of autonomous vehicles and as well as information from lane edge angles. smart transportation is lane detection. The system also employs a lane classification Although Convolutional Neural Networks method that only chooses the validated lanes (CNNs) have become a top method for this for enhanced performance [6].Advanced use, research is still being done to improve driver assistance systems are essential to their performance. In order to simplify CNNs assuring driving safety and require an for lane detection, the Dynamic Mode efficient lane-detection algorithm. However, Decomposition (DMD) approach is there are several difficulties that lane introduced in this study.The research uses a detection algorithms must overcome, such as CNN model to conduct lane detection and imprecise lane markers, poor vision caused focuses on minimizing the number of layers by bad weather, problems with illumination while preserving model accuracy, which is and light reflection, the existence of shadows, made possible by including DMD. The and complicated road-based instructions. This optimized model exhibits accuracy values study presents a robust and real-time vision- that are comparable to those of the baseline based lane detection system with aneffective model and is implemented using the region of interest (ROI) that handles high TensorFlow library. The suggested approach noise levels and processing time in response outperforms the baseline model in terms of to these problems.To increase accuracy, the speed since layer reduction results in a suggested method integrates gradient and reduction in the amount of computations color cues. In order to validate the properties needed. Both models are thoroughly tested of the lane markings, it combines a line [5].The vital task of lane detection is crucial clustering procedure with scan-line tests, in the global efforts made by the automotive efficiently eliminating fake lane markings. industry to prevent traffic accidents and Additionally, the program tracks actual lane create autonomous vehicles and advanced markers using gathered statistical data. The driver assistance systems (ADAS). For algorithm's correctness and real-time vehicles to stay in their allotted lanes, functionality are demonstrated by accurate lane detection is crucial, but current experimental findings, which make it ideal for embedded devices with constrained IV. Proposed System computational resources [8]. Detecting potholes in road infrastructure is III. Existing System a critical task, and a systematic approach is essential for accuracy and reliability. The The autonomous vehicle system relies on process begins with dataset annotation, real-time video data captured by sensors and where the location and boundaries of cameras mounted on the vehicle. It employs potholes are marked within images or video two specialized deep learning models for frames. This annotation can be done critical tasks. First, for lane detection, a manually or with the assistance of semi- Fully Convolutional Network (FCN) is automatic tools. Data preprocessing utilized, trained on diverse data sources such follows, ensuring uniformity by resizing as Udacity. During the training phase, the images to a consistent resolution, FCN learns to segment and identify lane normalizing pixel values, and verifying boundaries from input images. In testing, annotation correctness.For training, an the trained FCN model processes test appropriate deep learning model is selected, images to predict the number of lanes with options such as Faster R-CNN, YOLO, present, with a clever mechanism to SSD, or custom architectures designed for calculate the mean average of multiple specific needs. Hyperparameter tuning, detected lanes to avoid confusion during including adjustments to learning rates, autonomous driving. The FCN model batch sizes, and data augmentation architecture leverages semantic techniques, is crucial for optimizing model segmentation through convolution, up- performance. sampling via deconvolution, and information fusion from shallower layers. Validation and testing phases are critical, with the model's performance assessed on For speed breaker detection, the system separate validation and test datasets to employs YOLOv5, a cutting-edge object ensure it generalizes well and performs detection algorithm. A specialized dataset effectively in real-world scenarios. During for speed breaker detection is meticulously testing, real- world data, often captured by crafted using Roboflow, including crucial vehicle- mounted cameras or other sources, preprocessing steps like data augmentation is input into the model for pothole and annotation. YOLOv5l, a specific detection. YOLOv5 model variant, is then trained on this preprocessed dataset to effectively In the inference stage, the trained model detect speed breakers. During testing, the analyzes each frame to identify the model leverages learned weights to predict presence and location of potholes. the presence of speed breakers in the test Depending on the model architecture, the data. output may include bounding boxes indicating pothole position and size or Both the lane detection and speed breaker pixel-wise masks segmenting potholes detection models undergo rigorous from the background. performance evaluation. The successful implementation of this methodology for pothole detection plays a crucial role in addressing road hazards as lane- keeping, adaptive cruise control, promptly, enhancing road safety. Specific emergency braking, and more. The focus is tools, models, and strategies can be adapted on collision prevention, reducing driver based on project requirements and workload, and improving overall driving available resources. Ongoing monitoring experience. and model maintenance are essential to sustain detection accuracy over time and Complexity and Integration: Lane Pothole ensure the continued safety of road and Speed Breaker Detection: While this is infrastructure. a valuable safety feature, it is relatively simpler in terms of sensor requirements Detecting lane potholes, speed breakers, or and computational complexity compared to other road hazards using image detection is a full-fledged ADAS. It can often be a specific application within the broader implemented using a single camera sensor field of advanced driver-assistance systems and specialized image processing (ADAS). Here's how it differs from a algorithms. typical ADAS: In summary, lane pothole and speed Specific Focus:Lane Pothole and Speed breaker detection using image processing is Breaker Detection: This application a specialized application within the realm focuses on identifying and warning the of ADAS. While it serves a specific driver or the vehicle's control system about purpose of hazard avoidance and driver specific road hazards like potholes, speed comfort, ADAS encompasses a broader breakers, or debris within the lane. The range of features aimed at improving primary goal is to ensure driver safety and overall road safety and driving experience. comfort by avoiding these obstacles.
Detection Method: Lane Pothole and Speed
Breaker Detection: This application often relies on image processing and computer vision techniques to identify specific visual cues associated with potholes and speed breakers. It may involve detecting changes in road surface texture, depth discontinuities, or other visual features that indicate a hazard.
ADAS: ADAS incorporates a variety of
sensors, including cameras, radar, lidar, ultrasonic sensors, and more. These sensors work together to provide a comprehensive view of the vehicle's surroundings and detect a wide range of objects, including other vehicles, pedestrians, cyclists, and road markings.ADAS aims to enhance overall road safety and driving comfort by assisting the driver with various tasks, such V.Conclusion smart city infrastructure,’’ Complex Intel. In conclusion, the integration of speed Syst. May 2021 breaker, pothole, and lane detection systems [4] Hu, J., Xiong, S., Zha, J. et al. Lane in cars using machine learning has ushered Detection and Trajectory Tracking Control of in remarkable advancements in road safety, Autonomous Vehicle Based on Model driver comfort, and overall vehicle Predictive Control. Int.J Automot. Technol. performance. These systems play a pivotal (2020).IEEE role in enhancing safety by providing timely [5] Aravind, H., Sivraj, P. and alerts to drivers about potential hazards like Ramachandran, K.I., 2020, July. Design and speed breakers, potholes, and lane Optimization of CNN for Lane Detection. deviations, consequently reducing accidents [6] Zou, Q., Jiang, H., Dai, Q., Yue, Y., and vehicle damage. Moreover, they elevate Chen,L. and Wang, Q., 2019. Robust lane passenger comfort by facilitating smoother detection from continuous driving scenes rides over road irregularities. Beyond safety using deep neural networks. IEEE and comfort, these systems contribute to transactions on vehicular technology, vehicle health, improve fuel efficiency, and [7] C. Lee and J. H. Moon, ‘‘Robust Lane enhance traffic flow, ultimately creating a detection and tracking for real time more convenient and environmentally applications,’’ IEEE Trans. Intel. Transp. friendly driving experience. Syst 2018. [8] Bruls, T., Maddern, W., Morye, A.A. References and Newman, P., 2018, May. Mark yourself: Road marking segmentation via weakly- [1] H. G. C, R. P and R. S, "Detection of supervised annotations from multimodal Lane and Speed Breaker Warning System for data. In 2018 IEEE . Autonomous Vehicles using Machine [9] Komori, H. and Onoguchi, K., 2018, Learning Algorithm," 2022 IEEE. December. Driving Lane Detection Based on [2] F. Munir, S. Azam, M. Jeon, B. -G. Lee Recognition of Road Boundary Situation. In and W. Pedrycz, "LDNet: End-to-End Lane 2018 . Marking Detection Approach Using a [10] H. Zhu, K.-V. Yuen, L. Mikhaylova, Dynamic Vision Sensor," in IEEE and H. Leung, “Overview of Environment Transactions on Intelligent Transportation Perception for Intelligent Vehicles,” IEEE Systems, July 2022 Transactions on Intelligent Transportation [3] S. Ghanem, P. Kanungo, G. Panda, S. C. Systems. Satapathy, and R. Sharma, ‘‘Lane detection [11] J. Guan, F. An, X. Zhang, L. Chen, H. under artificial colored light in tunnels and Mattausch, Real-time straight-line detection on highways: An IoT-based framework for for XGAsize videos by Hough Transform with parallelized voting procedures, Sensors.