You are on page 1of 24

Application of Image Classification in Road Infrastructure

Maintenance: A Literature Review

I. INTRODUCTION Additionally, the review examines practical


In today's rapidly evolving world, applications of image classification in
maintaining road infrastructure is crucial for real-world scenarios, such as detecting
facilitating smooth transportation networks potholes, cracks, surface defects, and
and fostering economic development. As monitoring the condition of road signs and
technology advances, image classification markings. By automating inspection
techniques have become valuable tools for processes, image classification saves time
evaluating and managing road and resources and allows for timely
infrastructure. This literature review maintenance interventions, ultimately
explores the use of image classification in improving road safety and operational
road infrastructure maintenance, examining efficiency.
methodologies, challenges, and
advancements in this field. However, despite its potential benefits, the
adoption of image classification in road
Road infrastructure includes various infrastructure maintenance faces
elements, such as pavement conditions, challenges. Concerns about data quality,
signage, markings, and roadside assets. scalability, and algorithm robustness need
Traditional inspection methods often involve to be addressed to ensure reliable
manual surveys or expensive equipment, performance across different environmental
which are labor-intensive and conditions and geographic locations.
time-consuming. Image classification offers
a compelling alternative by employing By carefully examining existing literature,
computer vision algorithms to automatically this review aims to identify gaps in current
analyze images captured by research and propose directions for future
vehicle-mounted or drone-mounted investigations. By enhancing our
cameras. understanding of the application of image
classification in road infrastructure
The review aims to provide a maintenance, we can contribute to the
comprehensive overview of research efforts development of more efficient and
in this area, highlighting different sustainable transportation systems.
approaches used for image classification in
road infrastructure maintenance. II. LITERATURE SURVEY
Researchers have explored various
methods, from basic pixel-based A. Winter Road Condition Recognition
classification to advanced deep learning Using Video Image Classification
models, to accurately identify and assess
This study focuses on classifying winter
different road features.
road conditions based on video images to
enhance maintenance decisions. The accuracy of 84.25% achieved by a method
dataset used consists of 69 color images utilizing engineered visual features. Both of
captured during winter and spring filming these methods significantly outperformed a
sessions with a handheld Hi-8 camcorder, simple baseline classifier, which achieved
depicting various road states such as dry, an accuracy of 74.23%. The unsupervised
wet, snowy, icy, and snowy with tracks. The learning approach involved the extraction of
performance of the model developed in the random patches from training images,
study shows promising results, with typical followed by the application of k-means
rates of correct classification ranging from clustering to learn a dictionary of visual
40% to 50%. The neural network approach features. Subsequently, features were
demonstrates the potential to distinguish extracted from novel images by encoding
between different road states using features patches based on the learned dictionary
such as mean and standard deviation of and pooling encoded vectors over a grid to
image colors, response to filters, and ratios form final image representations. Finally, a
of image properties. The study showcases linear SVM classifier was trained on the
the feasibility of using video images and extracted features and labels to effectively
neural networks for winter road classify road images in the test set into the
maintenance decision-making. Further two distinct classes.
enhancements such as refining feature
extraction, collecting more data, and
exploring advanced network architectures
are suggested to improve the model's C. Classification of Paved and Unpaved
accuracy and utility in real-world Road Image Using Convolutional Neural
applications. Network for Road Condition Inspection
System

The research focuses on developing a road


B. Image-Based Road Type Classification condition inspection system using
Convolutional Neural Network (CNN) for
The study utilized a comprehensive road classifying paved and unpaved road
image dataset comprising 20,000 real-world images. The dataset comprises 16,372 road
samples collected from Google Street View. images collected from urban and rural roads
These images were extracted from GPS in Timor Leste, categorized into paved and
trajectories and geo-located, resulting in a unpaved classes. These images depict
dataset with two distinct classes: paved various environmental conditions (wet,
roads and unpaved roads, each containing muddy, dry, dusty, shady) and road surfaces
10,000 images. The images were manually (ground, rocks, sands, asphalt). The dataset
extracted from road surfaces and also includes diverse features commonly
standardized to 32x32 pixel sub-images. In found on roads, such as buildings, trees,
terms of performance, the evaluation of the vehicles, and human elements. Images are
classification methods revealed that the sized 1920 x 1080 pixels with three color
proposed unsupervised feature learning channels (red, green, blue), simulating
approach achieved a notable test accuracy real-world road conditions for effective CNN
of 85.3% using 5-fold cross-validation. This training. The proposed CNN model
performance was comparable to the achieved high performance in classifying
road images, with precision, recall, and scenes from inner city to rural areas, with
F1-score values of 98.0%, 98.4%, and sequences ranging from 200-400m in length
98.2%, respectively. This indicates its ability and recorded at speeds between 20-50
to accurately differentiate between paved km/h. The results demonstrate that the road
and unpaved roads with confidence. The curve can be accurately extracted without
CNN architecture includes a feature user input in under a minute for sequences
extractor part with convolutional and pooling up to a minute long. Quantitative evaluation
layers, followed by fully connected layers against ground truth revealed that 50% of
acting as a classifier. Leveraging deep the extracted road contour was within 3.5cm
learning techniques, the model of ground truth, and 88% was within 9cm.
outperformed conventional machine The road labeling achieved a precision of
learning algorithms like Random Forest and 99% and a recall of 96%, while the curb
Support Vector Machine (SVM). Specifically, labeling achieved a precision of 45% and a
the technique involves utilizing deep recall of 79%. The key technique involves
learning algorithms to automatically extract using geometric cues from LiDAR data
features from road images, enabling exclusively to semantically label the 3D
effective classification between paved and scene, thereby avoiding reliance on image
unpaved roads. This approach capitalizes features. This ensures that the generated
on the CNN's capability to learn labeled data is suitable for training image
discriminative features directly from the classifiers without biasing them to the
images, eliminating the need for generation method. Overall, the proposed
handcrafted feature engineering and technique offers a promising solution for
enhancing classification accuracy. automating the generation of labeled
training data for road scene image
classification, with robust performance
across diverse real-world scenarios.
D. Automatic Generation of Training
Data for Image Classification of Road
Scenes
E. Classification of Road Surfaces Based
This study introduces a novel technique for on CNN Architecture and Tire Acoustical
automating the generation of labeled Signals
training data for image classification of road
scenes. This approach leverages LiDAR The study utilizes a convolutional neural
(light detection and ranging) data to network (CNN) architecture to classify road
semantically label a 3D point cloud scene, surfaces based on tire acoustical signals.
detect curb locations, and extract road The dataset consists of images derived from
boundary shapes using an active contour these signals collected on various road
(snake) algorithm. The labeled 3D point surfaces (asphalt and snowy) using different
cloud is then projected into camera images, types of tires (tire-A and tire-B) on both front
and morphological operations are applied to and rear wheels. Data preparation involves
achieve dense labeling. The technique was creating balanced sets for training,
tested on 16 datasets recorded with a validation, and testing. In terms of
Velodyne HDL-64e LiDAR and performance, the CNN-based model
high-resolution camera, covering diverse demonstrates promising results in
accurately categorizing road surfaces, the automated method, a line extraction
showcasing the efficacy of the approach. module automatically detected linear
Specifically, the classification accuracy of features. In the semi-automated method,
road surfaces serves as the metric for preprocessing filters such as Prewitt and
evaluating the model's performance. To Sobel high pass filters were applied to
process the data, the study employs enhance roads before extraction, followed
techniques such as transforming 1D noise by maximum likelihood classification for
signals into 2D time-frequency image analysis.
representations using Continuous Wavelet
Transform (CWT) to generate input images
for the CNN. This high-resolution CWT
facilitates capturing intricate signal G. Road signs classification by ANN for
characteristics, enhancing the classification real-time implementation
performance of the CNN.
This study focuses on traffic sign
classification and leverages a dataset
comprising 177 images captured in urban
F. Road Detection from High and Low and countryside areas in Tunisia. These
Resolution Satellite Images images encompass various lighting and
weather conditions, including sunny, cloudy,
The study focused on road extraction from rainy, and nighttime scenarios. The dataset
satellite images in the western area of contains traffic signs representing four
Ankara, Turkey. Five satellite images with shape classes: triangular, inverted
varying resolutions were used: QUICKBIRD triangular, circular, and octagonal. Over 100
(2.4m), IKONOS (4m), SPOT-PAN (10m), images were employed for training the
ASTER (15m), and LANDSAT-ETM (30m). neural networks. Regarding model
Automated and semi-automated methods performance, the Multi-Layer Perceptron
were employed to detect roads from these (MLP) neural network achieved an overall
images. The automated extraction was shape classification rate of 94.5%. However,
performed using the line module of PCI the classification rate for inverted triangles
Geomatica software. The study aimed to was slightly lower at 90%. For content
assess the impact of resolution on road classification of circular and triangular signs
extraction accuracy. The automated method using the Single Layer Perceptron (SLP)
was most effective for extracting primary neural network, an overall classification rate
roads in bare rural areas from of 89% was attained, with triangular signs
high-resolution images. In contrast, the exhibiting a slightly higher rate (90%)
semi-automated method utilized compared to circular signs (88%).
preprocessing filters and image Confusion primarily occurred between
classification techniques. Performance and inverted and non-inverted triangles, as well
accuracy were assessed by comparing as between circular and octagonal shapes.
extracted roads with ground truth GIS road Notably, confusion in circular sign content
data. Higher resolution images, such as classification was observed between speed
Quickbird and Ikonos, demonstrated limit signs of 50km/h and 30km/h. The study
superior road network extraction results employed two neural networks: an MLP for
compared to lower resolution images. For shape classification and an SLP for content
classification. Features extracted from each softmax activation in the output layer for
traffic sign image included the number of converting logits to normalized probabilities
black pixels, distance transforms, and sizes for different classes. Additionally, a
of lines and shapes. The MLP network cross-entropy loss function was used to
consisted of 6 input nodes, 20 hidden measure divergence between predicted and
nodes, and 4 output nodes for shape true distributions during training.
classification, while the SLP network utilized
pixel values of extracted sign contents as I. The Separation of the Unpaved Roads
input and featured 5 output nodes to classify and Prioritization of Paving These Roads
into 32 sign classes. Both networks were Using UAV Images
trained using the backpropagation algorithm
and Widrow-Hoff learning rule, respectively. The study utilized high-resolution UAV
images covering a 332-hectare area in Yazd
city, Iran. These images consisted of three
bands (Red, Green, Blue) and were
H. Road Damage Detection using CNN in employed to classify four land cover
Machine Learning classes: built-up land, paved roads, bare
land, and vegetation cover. Each land cover
The researchers trained their Convolutional class had over 500 training data points
Neural Network (CNN) model on a publicly collected. In terms of model performance,
available dataset of road images containing five object-based classification techniques
various types of damage. The dataset were evaluated, with the Support Vector
contains 6,000 images, with 4,500 images Machine (SVM) achieving the highest
used for training and 1,500 images used for overall accuracy of 89% and a kappa
testing. The images show cracks, potholes, coefficient of 80% for classifying the land
and patches on road surfaces. Their covers. SVM outperformed other
proposed CNN model for road damage techniques, demonstrating its suitability for
detection consists of six convolutional layers this specific dataset and analysis. The SVM
followed by three fully connected layers. technique was then employed to classify the
Rectified Linear Unit (ReLU) activation land covers from the UAV images,
function is used in all layers except the effectively distinguishing between paved
output layer, which uses softmax. and unpaved roads. Unpaved roads were
Evaluation of the CNN model on the test prioritized for paving based on the ratio of
dataset yielded an accuracy of 94% in built-up land to bare land in each road,
detecting and classifying different types of resulting in the assignment of five priority
road damage from images. The precision levels, with 13% of roads identified as
and recall of the model were 92% and 96% needing medium-high priority upgrades.
respectively. The core techniques employed
include convolutional layers for feature
extraction from road images, ReLU
activation for introducing non-linearity, max
pooling for dimensionality reduction while
retaining important information, fully J. Road condition assessment by OBIA
connected layers for integrating spatial and feature selection techniques using
features and performing classification, and
very high-resolution WorldView-2 dataset (AISD) and the ISPRS 2D Semantic
imagery Labeling Contest dataset. The AISD, known
as Potsdam-A, consists of 24 original
This study utilized Very High Spatial images, later cropped into 512x512 patches
Resolution (VHR) WorldView-2 (WV-2) and augmented to a total of 4307 images.
satellite imagery, with a 0.5m pan The ISPRS dataset, Potsdam-B, comprises
sharpened multispectral resolution, acquired 38 accurately labeled images,
over diverse locations including Universiti encompassing five foreground classes:
Putra Malaysia (UPM), Kuala Lumpur (KL), impervious surfaces, buildings, low
and Bandar Baru Bangi, Malaysia. The vegetation, trees, and cars. The study
objective was to extract and classify road introduces the Hybrid Multi-resolution and
conditions (good vs. poor) from the imagery, Transformer semantic extraction Network
employing Object-Based Image Analysis (HMRT), which outperforms existing models
(OBIA) and feature selection techniques. like FCN-8S, U-Net, PSPNet, and
Several classification methods were DeeplabV3+. On the Potsdam-A dataset,
evaluated, including Support Vector HMRT achieves an 85.32% recall, 84.88%
Machine (SVM), Random Forest (RF), and F1 score, 85.99% overall accuracy, and a
Chi-square. Chi-square emerged as the mean intersection-over-union (IoU) of
optimal performer, identifying 19 pertinent 74.19%. Moreover, HMRT demonstrates its
attributes spanning spatial, spectral, texture, effectiveness in generalization testing on
and color features. Classification using the Potsdam-B dataset, achieving a recall of
Chi-square achieved an overall accuracy of 91.29%, an F1 score of 90.41%, an overall
83.19% at the UPM training site and accuracy of 91.32%, and a mean IoU of
maintained accuracy at new sites (83.44% 84%. To enhance semantic segmentation,
for KL and 87.80% for Bandar Baru Bangi) the study employs several innovative
without additional training data. OBIA techniques. It introduces a multi-resolution
involved image segmentation, attribute semantic extraction branch to maintain
computation, and object-based resolution during downsampling, preserving
classification, surpassing pixel-level crucial feature information. Additionally, a
analysis. Feature selection techniques Transformer sequence feature extraction
played a crucial role in identifying attributes network is integrated to obtain a global
critical for classification accuracy. The receptive field for each pixel, enhancing
Chi-square technique particularly stood out contextual understanding. Transformer
for selecting attributes essential for the encoding and decoding are applied post
OBIA classification model, highlighting its convolutional feature extraction to facilitate
pivotal role in the process. global context understanding. Furthermore,
the fusion of feature maps from different
K. Multi-Resolution Transformer Network resolutions and network branches enriches
for Building and Road Segmentation of semantic information, contributing to the
Remote Sensing Image superior performance of the HMRT model.
This study focuses on semantic L. Near Real-time Map Building with
segmentation of remote sensing images in Multi-class Image Set Labeling and
the Potsdam region. It utilizes two main
datasets: the aerial image segmentation
Classification of Road Conditions Using Networks (CNNs) and transfer learning
Convolutional Neural Networks techniques. The dataset utilized in the study
consists of images depicting various traffic
This study focuses on the classification of signs collected from diverse locations,
road conditions using Convolutional Neural ensuring the model's robustness by
Networks (CNNs) in near real-time. The exposing it to signs in different
dataset used consisted of 47,000 images environments and conditions. These images
captured by street and highway cameras are carefully preprocessed and divided into
across North America. These images were training, validation, and test sets, with
manually labeled with five classes: dry, wet, respective proportions of 60%, 20%, and
snow/ice, poor, and offline road conditions. 20%. Transfer learning is employed to adapt
The dataset was meticulously curated to pre-trained CNN models, specifically
include a variety of road conditions, VGG-16 and ResNet-34 architectures, for
ensuring the effectiveness of the the task of classifying traffic sign images.
classification models. Various configurations Among these architectures, ResNet-34
of CNNs were evaluated in the study, with emerges as the top performer, achieving an
the EfficientNet-B4 framework achieving the impressive accuracy of 99.74% on the test
highest validation accuracy of 90.6%. set, indicating strong generalization
Despite its slightly lower accuracy, capabilities. Additionally, the Confusion
EfficientNet-B0 performed well with an Matrix and Kappa coefficient demonstrate
accuracy of 90.3% and faster execution the model's ability to accurately identify
time. The results demonstrated the images with excellent agreement between
effectiveness of CNNs in classifying road predicted and actual classes. The study
conditions when trained on a diverse and further examines the model's performance
well-labeled dataset. Convolutional Neural on new images, revealing high accuracy
Networks (CNNs) served as the primary rates for predicting various types of traffic
technique for classifying road conditions. signs. It suggests potential improvements
Different CNN architectures, including by exploring deeper models such as
VGG-16, ResNet50, Xception, ResNet-101/150. Transfer learning serves
InceptionResNetV2, and EfficientNet-B0, as the primary technique for leveraging
were employed and trained with specific pre-trained CNNs, fine-tuning the top layers
hyperparameters and optimization on the target dataset. To optimize model
techniques. Additionally, data augmentation training, various optimizers including SGD,
and regularization methods were utilized to Adam, and RMSProp are explored, with
enhance model performance and SGD with Momentum identified as the most
generalization. effective for ResNet-34, achieving
convergence in 50 epochs with 97%
accuracy. The study employs
comprehensive methodologies including
M. Improving the transfer learning
data preprocessing, model architecture
performances in the classification of the
selection, hyperparameter tuning, and
automotive traffic roads signs
evaluation metrics to develop and analyze
This study focuses on the classification of the best-performing model for traffic sign
traffic road signs using Convolutional Neural classification.
O. Classification of Road Damage from
Digital Image Using Backpropagation
Neural Network

The research aims to reduce traffic


accidents caused by road damage through
N. Pothole Classification Model Using early detection. By utilizing image
Edge Detection in Road Image processing techniques and
Backpropagation Neural Network, the study
This study investigates the detection and
explores the classification of road conditions
classification of road damages, specifically
into good road and damaged road
potholes, leveraging data from the Global
categories. The dataset comprises images
Road Damage Detection Challenge 2020.
representing good road conditions, including
The dataset encompasses 13,376
unmarked good road and good road
pre-processed road damage images
markings, as well as damaged road
collected from Japan, India, and the Czech
conditions. A total of 30 data points were
Republic. To enhance model accuracy,
used for testing the model's performance.
various preprocessing techniques are
employed, including grayscale conversion The developed model achieved an accuracy
and YOLO object detection to remove rate of 83%, surpassing the results obtained
non-pothole objects. The dataset is divided using other algorithms such as region
into training (70%), validation (10%), and split-merger and fractal dimension. The
test (20%) sets for evaluation. Performance Backpropagation Neural Network algorithm
evaluation of the model involves metrics effectively recognized unmarked good road
such as Mean Squared Error (MSE), Peak conditions but showed partial success in
Signal-to-Noise Ratio (PSNR), and distinguishing between good road markings
Structural Similarity Index (SSIM) to and damaged road. The research employed
measure image distortion and restoration a systematic process divided into training
rates. Pothole classification accuracy on the and testing subprocesses. The training
test data ranges from 60.5% to 96.2%, with process involved resizing, grayscaling, edge
an Area Under the Curve (AUC) of 0.9 for detection, thresholding, and training with
single potholes and 0.733 for multiple Backpropagation Neural Network. Similarly,
potholes. The proposed model exhibits the testing process included preprocessing
promising results, achieving a precision of steps followed by testing with the neural
83.45% and an accuracy of 77.86%, network.
surpassing existing methods in pothole
detection. Techniques such as
pre-processing, object detection, and edge
detection contribute to the model's P. An integrated machine learning model
effectiveness in detecting and classifying for automatic road crack detection and
road damages. classification in urban areas

This study presents an integrated machine


learning model designed for the automatic
detection and classification of road cracks
from image data. The model aims to provide underrepresented markings, negative
a comprehensive and accurate system for images, and a "NULL" class. The algorithm
timely pavement inspection, ultimately demonstrated remarkable performance,
reducing maintenance costs compared to achieving classifier accuracies ranging from
manual methods. The dataset used for 94.7% to 99.1% and detection/recognition
testing the proposed model consists of 400 accuracies between 90.8% and 93.1%,
images of asphalt pavement collected from surpassing existing baseline methods. To
urban areas in Tehran, Iran. These images achieve these results, the algorithm
were captured using smartphone cameras employed a blend of computer vision and
and depict various types of cracks, including machine learning techniques. These
transverse, longitudinal, diagonal, and block techniques included inverse perspective
cracks. The proposed model involves seven mapping for generating bird's-eye views,
stages, including image acquisition, contrast enhancement, and texture removal
segmentation, denoising, enhancement, line using bilateral filtering for illumination
detection using Hough transform, feature robustness. Additionally, the algorithm
extraction, and classification. Heuristic utilized MSER for detecting regions of
algorithms are developed for the interest, density-based clustering for
segmentation and denoising stages to merging closely grouped regions, and deep
improve noise removal. The Hough learning models such as PCANet and
transform and equations are utilized to shallow CNNs for classification. The
extract features representing crack algorithm also incorporated iterative
structural details. The model's performance refinement of regions and parallel
was evaluated using six classification processing of image pairs to enhance
models: neural network, SVM, decision tree, robustness under varying lighting
KNN, Bagged Trees, and a proposed hybrid conditions. Overall, this study underscores
model combining Bagged Trees and SVM. the effectiveness of machine learning
The hybrid model achieved the best approaches in achieving accurate and
performance, with an overall accuracy of reliable road marking detection and
93.86% for classifying cracks into the four classification.
types.

R. Capsule-Based Networks for Road


Q. Robust Road Marking Detection and Marking Extraction and Classification
Recognition Using Density-Based From Mobile LiDAR Point Clouds
Grouping and Machine Learning
Techniques This study investigates road marking
extraction and classification using mobile
This study evaluates an algorithm for road LiDAR point cloud data collected from urban
marking detection and classification, roads, highways, and underground garages.
utilizing a publicly available dataset The dataset, comprising over 7,000
comprising 1443 road images. These samples annotated at the pixel level on
images were manually labeled and intensity images, covers seven road
categorized into 11 classes, encompassing marking categories. The data was
various road marking types, partitioned into training, validation, and test
sets. For road marking extraction, the study ground truths. Convolutional Neural
employs a U-shaped Networks (CNNs) served as the primary
convolution-deconvolution capsule network technique, utilizing pretrained models like
guided by Intersection over Union (IoU) Resnet-18, VGG-16, AlexNet, MobileNet,
loss, achieving an average precision of and ShuffleNet as backbones.
94.11%, recall of 90.52%, and F1-score of
92.43%. For classification, a hybrid capsule A classification and regression CNN model
network is utilized, incorporating a revised detected longitudinal road markings, while a
dynamic routing algorithm and large-margin U-Net architecture-based segmentation
Softmax loss, achieving an average model accurately delineated regions if
misclassification rate of 3.42%. Notably, the present. Optimization was achieved using
models demonstrate robust performance the Adaptive Moment Estimation (Adam)
even on low-quality data from underground technique, with mean squared error
garages, addressing challenges such as low employed as the loss function.
point density, poor illumination, and low
intensity contrast. Leveraging capsule T. Road Vehicle Classification using
networks, the models effectively capture Support Vector Machines
pose and spatial information, providing a
This study presents an approach for vehicle
reliable solution for road marking extraction
classification using support vector machines
and classification from Mobile LiDAR
(SVMs) and simple features. The study
Scanning (MLS) point clouds.
utilized vehicle images from two distinct
S. Classification and Segmentation of sources: the COIL database, providing
Longitudinal Road Marking using controlled test data with color images of
Convolutional Neural Networks for various vehicle types and colors, and real
Dynamic Retroreflection Estimation video data captured from road-side
cameras, offering more challenging
The study utilized a dataset comprising real-world test data. The techniques
29,635 luminance images captured by a employed included foreground
vehicle-installed camera, with 4958 positive segmentation using a Gaussian mixture
images containing road markings and model (GMM) based background
reference plates, and 24,677 negative subtraction method, feature extraction
images. The dataset was split into training involving shape and color features, and
(24,270 images) and evaluation (5365 classification using SVMs with various
images) sets. The Resnet-18 model kernels. The performance evaluation
achieved the highest accuracy of 96.96% revealed that the approach achieved 100%
and the lowest Mean Absolute Error (MAE) recognition for common vehicle colors on
of 0.0230 for regression in the classification the COIL database and an overall accuracy
and regression model. In the segmentation exceeding 80% for classifying vehicles into
model, Intersection over Union (IoU) scores different types and colors on real video
for background, reference plate, and data, despite some confusion between
longitudinal road marking were 0.9860, specific vehicle types. This pragmatic
0.9349, and 0.9495, respectively, with an approach demonstrated promising
average of 0.9568, indicating high performance on both controlled and
performance and close alignment with real-world data, showcasing the potential of
SVMs and simple features for vehicle
classification.

Author/s
Title Dataset used Techniques used Performance of the
Model

Kuehnle, Winter Road It comprises 69 color Feature Extraction The developed neural
A., & Condition images showing network model shows
Burghout Recognition various road Neural Network promising results,
, W. Using Video conditions like dry, Architecture through achieving correct
Image wet, snowy, icy, and Feature Combination classification rates
Classification snowy with tracks, between 40% to 50%.
providing diverse It can distinguish
training and testing between different road
samples. states, except for
certain combinations
like ice/wet and
ice/tracks.

Slavkovi Image-Based Utilized a Manual Extraction This performance was


kj, V., Road Type comprehensive comparable to the
Verstockt Classification dataset of 20,000 Standardization to accuracy of 84.25%
, S., De real-world road 32x32 pixels achieved by a method
Neve, images sourced from utilizing engineered
W., Van Google Street View, Comparison: visual features. Both of
Hoecke, containing two Unsupervised vs. these methods
S., & classes: paved Engineered Features significantly
Van De roads and unpaved outperformed a simple
Walle, R. roads, each Unsupervised baseline classifier,
comprising 10,000 Learning: Patch which achieved an
images. Extraction, K-means accuracy of 74.23%.
Clustering

Feature Extraction
from Novel Images

Classifier Training
(Linear SVM)
V.
Pereira, Classification The dataset includes CNN Architecture The proposed CNN
S. of Paved and 16,372 road images model achieved high
Tamura, Unpaved from urban and rural Feature Extractor performance in
S. Road Image areas in Timor Leste, classifying road
Hayamiz Using categorized into Support Vector images, with precision,
u, & H. Convolutiona paved and unpaved Machine recall, and F1-score
Fukai. l Neural classes. It values of 98.0%,
Network for showcases various Pooling 98.4%, and 98.2%,
Road environmental respectively.
Condition conditions (wet,
Inspection muddy, dry, dusty,
System shady) and road
surfaces (ground,
rocks, sands,
asphalt), along with
common road
features like
buildings, trees,
vehicles, and
humans. Images are
1920x1080 pixels
with RGB channels.
Kühner,
T., Automatic The technique was LiDAR Semantic The road labeling
Wirges, Generation of tested on 16 Labeling achieved a precision of
S., & Training Data datasets recorded 99% and a recall of
Lauer, for Image with a Velodyne Curb Detection & 96%, while the curb
M. Classification HDL-64 LiDAR and Road Boundary labeling achieved a
of Road high-resolution Extraction precision of 45% and a
Scenes camera, covering recall of 79%.
diverse scenes from Point Cloud
inner city to rural Projection for Dense
areas, with Labeling
sequences ranging
from 200-400m in Geometric Cue
length and recorded Utilization
at speeds between
20-50 km/h.
Yoo, J.,
Lee, Classification The dataset consists CNN for Road The CNN-based model
C.-H., of Road of images derived Surface demonstrates
Jea, Surfaces from these signals Classification promising results in
H.-M., Based on collected on various accurately categorizing
Lee, CNN road surfaces Balanced Data Sets road surfaces,
S.-K., Architecture (asphalt and snowy) Preparation showcasing the
Yoon, Y., and Tire using different types efficacy of the
Lee, J., Acoustical of tires (tire-A and Promising Results in approach. Specifically,
Yum, K., Signals tire-B) on both front Categorization the classification
& and rear wheels. accuracy of road
Hwang, Continuous Wavelet surfaces serves as the
S.-U. Transform (CWT) for metric for evaluating
Signal the model's
Transformation performance.

High-Resolution
CWT for Detailed
Signal Analysis
Gecen,
R., & ROAD The study focused Automated & Performance and
Sarp, G DETECTION on road extraction Semi-Automated accuracy were
FROM HIGH from satellite images Road Extraction assessed by
AND LOW in the western area comparing extracted
RESOLUTIO of Ankara, Turkey. PCI Geomatica's roads with ground truth
N SATELLITE Five satellite images Line Module GIS road data. Higher
IMAGES with varying resolution images,
resolutions were Preprocessing such as Quickbird and
used: QUICKBIRD Filters: Prewitt, Sobel Ikonos, demonstrated
(2.4m), IKONOS superior road network
(4m), SPOT-PAN Accuracy Evaluation extraction results
(10m), ASTER with Ground Truth compared to lower
(15m), and GIS Data resolution images.
LANDSAT-ETM
(30m).
Hamdi,
S., Road signs Traffic sign CNN architecture for The MLP neural
Faiedh, classification classification using a road surface network achieved
H., by ANN for dataset of 177 classification 94.5% shape
Souani, real-time images from urban classification accuracy,
C., & implementati and rural areas in Balanced datasets with slightly lower
Besbes, on Tunisia. Images for training, accuracy (90%) for
K. cover diverse validation, testing inverted triangles. The
lighting and weather SLP neural network
conditions: sunny, Accurate attained an 89%
cloudy, rainy, and categorization of overall classification
nighttime. Dataset road surfaces rate for content
includes traffic signs classification, with
of four shape Utilization of triangular signs slightly
classes: triangular, Continuous Wavelet outperforming circular
inverted triangular, Transform (CWT) signs (90% vs. 88%).
circular, and
octagonal. Enhanced CNN
classification with
high-resolution CWT
Shinde,
A., Road Publicly available CNN model with: Evaluation of the CNN
Borgave, Damage dataset of road model on the test
H., Detection damage images, Architecture: Six dataset yielded an
Anand, using CNN in comprising 6,000 convolutional layers, accuracy of 94% in
A., & Machine images. 4,500 three fully connected detecting and
Thigale, Learning images for training layers classifying different
S. S. and 1,500 for types of road damage
testing. Images Activation: ReLU in from images. The
depict cracks, all layers, softmax in precision and recall of
potholes, and output the model were 92%
patches on road and 96% respectively.
surfaces. Loss: Cross-entropy
for training
The
Mansour Separation of These images High-resolution UAV In terms of model
moghad the Unpaved consisted of three image usage performance, five
dam, M., Roads and bands (Red, Green, object-based
Ghafaria Prioritization Blue) and were RGB band extraction classification
n of Paving employed to classify techniques were
Malamiri, These Roads four land cover Four land cover class evaluated, with the
H. R., Using UAV classes: built-up classification Support Vector
Arabi Images land, paved roads, Machine (SVM)
Aliabad, bare land, and Evaluation of five achieving the highest
F., Fallah vegetation cover. classification overall accuracy of
Tafti, M., Each land cover techniques 89% and a kappa
Haghani, class had over 500 coefficient of 80% for
M., & training data points Support Vector classifying the land
Shojaei, collected Machine (SVM) for covers.
S. road classification

Unpaved road
prioritization based
on land cover ratios

Identification of roads
needing upgrades

Shahi,
K., Road The dataset used in Object-Based Image Classification using
Shafri, condition this study included Analysis (OBIA) Chi-square achieved
H. Z. M., assessment Very High Spatial an overall accuracy of
& by OBIA and Resolution (VHR) Feature Selection 83.19% at the UPM
Hamedia feature WorldView-2 (WV-2) Techniques training site and
nfar, A. selection satellite imagery with maintained accuracy at
techniques 0.5m pansharpened Classification new sites (83.44% for
using very multispectral Methods: Support KL and 87.80% for
high-resoluti resolution. It covered Vector Machine Bandar Baru Bangi)
on diverse locations (SVM), Random without additional
WorldView-2 such as Universiti Forest (RF), training data. OBIA
imagery Putra Malaysia Chi-square involved image
(UPM), Kuala segmentation, attribute
Lumpur (KL), and computation, and
Bandar Baru Bangi, object-based
Malaysia. classification,
surpassing pixel-level
analysis.
Sun, Z.,
Wangpin Multi-Resolut It uses two datasets: Hybrid The study introduces
g, Z., ion aerial image Multi-resolution the Hybrid
Chen, Transformer segmentation Transformer Network Multi-resolution and
D., & Network for dataset (AISD) and (HMRT) Transformer semantic
Min, X. Building and ISPRS 2D Semantic extraction Network
Road Labeling Contest Aerial Image (HMRT), which
Segmentatio dataset. AISD Segmentation outperforms existing
n of Remote (Potsdam-A) has 24 Dataset (AISD) models like FCN-8S,
Sensing original images U-Net, PSPNet, and
Image cropped into ISPRS 2D Semantic DeeplabV3+. On the
512x512 patches, Labeling Contest Potsdam-A dataset,
augmented to 4307 Dataset HMRT achieves an
images. ISPRS 85.32% recall, 84.88%
dataset (Potsdam-B) Outperforms existing F1 score, 85.99%
contains 38 labeled models like FCN-8S, overall accuracy, and a
images with five U-Net, PSPNet, and mean
classes: impervious DeeplabV3+ intersection-over-union
surfaces, buildings, (IoU) of 74.19%.
low vegetation, Multi-resolution Moreover, HMRT
trees, and cars. The semantic extraction demonstrates its
study introduces branch effectiveness in
Hybrid generalization testing
Multi-resolution and Transformer on the Potsdam-B
Transformer sequence feature dataset, achieving a
semantic extraction extraction network recall of 91.29%, an F1
Network (HMRT), score of 90.41%, an
surpassing FCN-8S Transformer overall accuracy of
U-Net, PSPNet, and encoding and 91.32%, and a mean
DeeplabV3+. decoding IoU of 84%.

Fusion of feature
maps

Semantic
segmentation of
remote sensing
images

Ramann Near The dataset used CNNs for road EfficientNet-B4


a, S., Real-time consisted of 47,000 condition achieved the highest
Sengoz, Map Building images captured by classification validation accuracy of
C., with street and highway 90.6%
Kehler, Multi-class cameras across Evaluation: VGG-16,
S., & Image Set North America. ResNet50, Xception, EfficientNet-B0
Pham, Labeling and These images were InceptionResNetV2, exhibited an accuracy
D. Classification manually labeled EfficientNet-B0 of 90.3% with faster
of Road with five classes: dry, execution time
Conditions wet, snow/ice, poor, Data augmentation,
Using and offline road regularization for
Convolutiona conditions. performance
l Neural
Network
Barodi,
A., Bajit, Improving Images depicting Convolutional Neural Dataset preprocessed,
A., the transfer various traffic signs Networks (CNNs) split into
Benbrahi learning collected from used for train/validation/test
m, M., & performance diverse locations classification sets (60%, 20%, 20%).
Tamtaoui s in the Transfer learning on
, A. classification Carefully Transfer learning VGG-16 and
of the preprocessed and employed to adapt ResNet-34 for traffic
automotive divided into training, pre-trained CNN sign classification.
traffic roads validation, and test models (VGG-16 and ResNet-34 achieves
signs sets ResNet-34 99.74% accuracy on
architectures) test set, demonstrating
strong generalization.
Optimizers explored Confusion Matrix and
include SGD, Adam, Kappa coefficient
and RMSProp confirm accuracy and
agreement between
Model architecture predicted and actual
selection, classes.
hyperparameter
tuning, and
evaluation metrics
employed for model
development and
analysis
Baek,
J.-W., & Pothole The dataset Leveraged The proposed model
Chung, Classification encompasses pre-processing, exhibits promising
K. Model Using 13,376 object detection, and results, achieving a
Edge pre-processed road edge detection precision of 83.45%
Detection in damage images techniques. and an accuracy of
Road Image collected from 77.86%, surpassing
Japan, India, and the existing methods in
Czech Republic. pothole detection.
Sutikno,
H. A. W., Classification The dataset Backpropagation The developed model
& of Road comprises images Neural Network for achieved an accuracy
Budiarto, Damage from representing good road condition rate of 83%,
P. Y. Digital Image road conditions, classification. surpassing the results
Using including unmarked obtained using other
Backpropaga good road and good Systematic training algorithms such as
tion Neural road markings, as and testing region split-merger and
Network well as damaged subprocesses. fractal dimension.
road conditions. A
total of 30 data Training: Resize,
points were used for grayscale, edge
testing the model's detection,
performance. thresholding,
Backpropagation
Neural Network.

Testing:
Preprocessing,
neural network
testing.
Ahmadi,
A., An integrated The dataset used for Integrated machine The proposed model
Khalesi, machine testing the proposed learning model involves seven stages,
S., & learning model consists of including image
Golroo, model for 400 images of Seven-stage acquisition,
A. automatic asphalt pavement process: image segmentation,
road crack collected from urban acquisition, denoising,
detection and areas in Tehran, segmentation, enhancement, line
classification Iran. These images denoising, detection using Hough
in urban were captured using enhancement, line transform, feature
areas smartphone detection (Hough extraction, and
cameras and depict transform), feature classification. Heuristic
various types of extraction, algorithms are
cracks, including classification developed for the
transverse, segmentation and
longitudinal, Heuristic algorithms denoising stages to
diagonal, and block for segmentation and improve noise removal.
cracks. denoising

Hough transform for


feature extraction

Evaluation of six
classification models
Bailo, O.,
Lee, S., Robust Road Publicly available Road marking The algorithm
Rameau, Marking dataset comprising detection & demonstrated
F., Yoon, Detection 1443 road images. classification remarkable
J. S., & and These images were algorithm performance, achieving
Kweon, Recognition manually labeled classifier accuracies
I. Using and categorized into Computer vision & ranging from 94.7% to
Density-Base 11 classes, machine learning 99.1% and
d Grouping encompassing detection/recognition
and Machine various road marking Inverse perspective accuracies between
Learning types, mapping, contrast 90.8% and 93.1%,
Techniques underrepresented enhancement, surpassing existing
markings, negative texture removal baseline methods. To
images, and a (bilateral filtering) achieve these results,
"NULL" class. the algorithm employed
MSER region a blend of computer
detection, vision and machine
density-based learning techniques.
clustering, deep
learning classification

Iterative refinement,
parallel processing
Ma, L., Capsule-Bas
Li, Y., Li, ed Networks The dataset, U-shaped capsule For road marking
J., Yu, for Road comprising over network: Road extraction, the study
Y., Marking 7,000 samples marking extraction employs a U-shaped
Junior, J. Extraction annotated at the convolution-deconvolut
M., and pixel level on Hybrid capsule ion capsule network
Gonçalv Classification intensity images, network: Road guided by Intersection
es, W. From Mobile covers seven road marking classification over Union (IoU) loss,
N., & LiDAR Point marking categories. achieving an average
Chapma Clouds The IoU loss: Extraction precision of 94.11%,
n, M. A. model recall of 90.52%, and
F1-score of 92.43%.
Revised routing,
Softmax loss:
Classification model

Robust on diverse
datasets: Urban
roads, highways

Effective pose
capture: Capsule
networks
Chun,
C., Lee, Classification The dataset was Density-based In the segmentation
T., Kwon, and split into training separation model, Intersection
S., & Segmentatio (24,270 images) and techniques Chemical over Union (IoU)
Ryu, S. n of evaluation (5365 analysis Microscopy scores for background,
Longitudinal images) sets. The Pyrolysis-GC reference plate, and
Road techniques longitudinal road
Marking marking were 0.9860,
using 0.9349, and 0.9495,
Convolutiona respectively, with an
l Neural average of 0.9568,
Networks for indicating high
Dynamic performance and close
Retroreflectio alignment with ground
n Estimation truths.

Chen, Z.,
Pears, Road Vehicle The dataset merges Foreground
Classification Achieved 100%
N., vehicle images from Segmentation: recognition on the
Freeman using two sources: the GMM-based vehicle
Support COIL database, even
, M., & COIL (Columbia segmentation. with added noise.
Austin, J. Vector Object Image
Machines Library) database, Feature Extraction:
comprising color Attained an overall
Shape (size, aspect
images of 100 accuracy of over 80%
ratio, width, solidity)
objects, including on real video for
and color (8-bin 3D
various vehicle types classifying vehicles into
color histogram in
and colors (source different types and
YCbCr space).
1), and real video colors, with some
footage from confusion between
Classification: SVMs
roadside cameras types like van and
with linear, RBF,
(source 2). This HGV.
polynomial kernels.
amalgamation Explored OVO, OVA
provides diverse test strategies.
data for vehicle Parameters
classification optimized via grid
algorithms, search.
encompassing both
controlled
environments and
real-world scenarios,
thus facilitating
comprehensive
testing and
evaluation.
CONCLUSION REFERENCES

This study provides a comprehensive A. Kuehnle, A., & Burghout, W. (1998).


overview of the application of image Winter Road Condition Recognition Using
classification techniques in road Video Image Classification. *Transportation
infrastructure maintenance. It highlights the Research Record*, *1627*(1), 29-33.
efficacy of Convolutional Neural Networks doi.org/10.3141/1627-05
(CNNs) in accurately identifying and https://journals.sagepub.com/doi/abs/10.314
categorizing various elements crucial for 1/1627-05
road management, including road surfaces,
markings, damage, and traffic signs. B. Slavkovikj, V., Verstockt, S., De Neve,
Through a systematic review of relevant W., Van Hoecke, S., & Van De Walle, R.
literature, the study elucidates the diverse (2014). Image-Based Road Type
applications of image classification in tasks Classification. In *2014 22nd International
such as road surface inspection, pothole Conference on Pattern Recognition* (pp.
detection, and traffic sign recognition. 2359-2364). Stockholm, Sweden.
https://ieeexplore.ieee.org/abstract/docume
The findings underscore the consistent high nt/6977121
performance of CNNs across different
datasets and environmental conditions, C. V. Pereira, S. Tamura, S. Hayamizu, & H.
indicating their potential to significantly Fukai. (2018). Classification of Paved and
enhance road safety and infrastructure Unpaved Road Image Using Convolutional
management practices. Additionally, the Neural Network for Road Condition
study emphasizes the importance of Inspection System. In *2018 5th
preprocessing methods, advanced model International Conference on Advanced
architectures, and optimization techniques Informatics: Concept Theory and
in improving the accuracy and efficiency of Applications (ICAICTA)* (pp. 165-169).
image classification models for road Krabi, Thailand. doi:
maintenance tasks. [10.1109/ICAICTA.2018.8541284]
(https://ieeexplore.ieee.org/abstract/docu
Despite the promising results showcased in ment/8541284)
the reviewed literature, the study identifies
several challenges that warrant further D. Kühner, T., Wirges, S., & Lauer, M.
attention. These include the need for more (2019). Automatic Generation of Training
diverse datasets, improved model Data for Image Classification of Road
interpretability, and the development of Scenes. In *2019 IEEE Intelligent
real-time implementation strategies. Transportation Systems Conference (ITSC)*
Addressing these challenges will be crucial (pp. 1097-1103). Auckland, New Zealand.
for advancing the practical application of doi:10.1109/ITSC.2019.8917089
image classification in road infrastructure https://ieeexplore.ieee.org/abstract/docume
maintenance and realizing its transformative nt/8917089
potential in reshaping transportation
systems for increased safety and efficiency. E. Yoo, J., Lee, C.-H., Jea, H.-M., Lee,
S.-K., Yoon, Y., Lee, J., Yum, K., & Hwang,
S.-U. (2022). Classification of Road
Surfaces Based on CNN Architecture and H. Shinde, A., Borgave, H., Anand, A., &
Tire Acoustical Signals. *Applied Sciences*, Thigale, S. S. (2023). Road Damage
12(19), 9521. Detection using CNN in Machine Learning.
https://doi.org/10.3390/app12199521 *International Journal of Engineering
https://www.mdpi.com/2076-3417/12/19/952 Research & Technology (IJERT)*, 12(04),.
1?fbclid=IwAR3dq8-DI_IMRftI-U64r16cEWj https://www.ijert.org/research/road-damage-
oF6rrOYMRJVxzkE7_4tScfRbWtvSrRXk detection-using-cnn-in-machine-learning-IJE
RTV12IS040232.pdf
F. Gecen, R., & Sarp, G. (Commission
PS-13, WgS IV/3). Road detection from high I. Mansourmoghaddam, M., Ghafarian
and low resolution satellite images. The Malamiri, H. R., Arabi Aliabad, F., Fallah
International Archives of the Tafti, M., Haghani, M., & Shojaei, S. (2022).
Photogrammetry, Remote Sensing and The Separation of the Unpaved Roads and
Spatial Information Sciences, 37(B4), Prioritization of Paving These Roads Using
355–357. UAV Images. *Air, Soil and Water Research,
https://d1wqtxts1xzle7.cloudfront.net/55840 15*. doi:10.1177/11786221221086285
664/ROAD_DETECTION_FROM_HIGH_A
ND_LOW_RESOLUTION_SATELLITE_IMA https://journals.sagepub.com/doi/full/10.117
GES-libre.pdf?1519046417=&response-con 7/11786221221086285
tent-disposition=inline%3B+filename%3DR
OAD_DETECTION_FROM_HIGH_AND_LO J. Shahi, K., Shafri, H. Z. M., &
W_RESOLUT.pdf&Expires=1708758903&Si Hamedianfar, A. (2017). Road condition
gnature=PcDEvb3eBSxkievSgEE3Tujp1gn assessment by OBIA and feature selection
mBoAAoxpLpRjSaFA~F0-1cX37YDk8ixy12I techniques using very high-resolution
thkh4ne~KZqADanf-VM2wLv-uLGIEIOgr1e WorldView-2 imagery. *Geocarto
34YX1P8~IEr6B2lqkPdw2DTj-uyS05wOMG International, 32*(12), 1389-1406.
~eKH7josuhoTRgQqvMANfc0lTiJ11ciSRON https://doi.org/10.1080/10106049.2016.121
KOeJyVKswTHEsqFY4Mg~tYAF-9MiXqv5l5 3888
WL1BR5I3vv9QmcysduAHDRg48vDvMEkis https://www.tandfonline.com/doi/abs/10.108
SsEtrgY21qS8rytfvgTVw0JZTG13oFGmrEb 0/10106049.2016.1213888
4oC3yn4vltUWMYLutgU3TSjW7TpomHEM
UKWgheyPkViYN4JGk29WbSFMI7nRDmq K. Sun, Z., Wangping, Z., Chen, D., & Min,
N0MhF7kvyZg__&Key-Pair-Id=APKAJLOH X. (2022). Multi-Resolution Transformer
F5GGSLRBV4ZA Network for Building and Road
Segmentation of Remote Sensing Image.
G. Hamdi, S., Faiedh, H., Souani, C., & *ISPRS International Journal of
Besbes, K. (2017). Road signs classification Geo-Information, 11*, 165.
by ANN for real-time implementation. In https://doi.org/10.3390/ijgi11030165
*2017 International Conference on Control, https://www.researchgate.net/publication/35
Automation and Diagnosis (ICCAD)* (pp. 8865989_Multi-Resolution_Transformer_Ne
328-332). Hammamet, Tunisia. twork_for_Building_and_Road_Segmentatio
doi:10.1109/CADIAG.2017.8075679 n_of_Remote_Sensing_Image/link/624afd0
https://ieeexplore.ieee.org/abstract/docume c7931cc7ccf1583b4/download?_tp=eyJjb25
nt/8075679 0ZXh0Ijp7ImZpcnN0UGFnZSI6InB1YmxpY
2F0aW9uIiwicGFnZSI6InB1YmxpY2F0aW9 159-165.
uIn19 https://doi.org/10.11591/ijai.v6.i4.pp159-165
https://ijai.iaescore.com/index.php/IJAI/articl
L. Ramanna, S., Sengoz, C., Kehler, S., & e/view/8176?fbclid=IwAR0wWdyCtmpQVh1
Pham, D. (2021). Near Real-time Map 4-aNkNLVEI94Z1qnLFC1xfPccWkd9OcE7t
Building with Multi-class Image Set Labeling utAXHM2qSA
and Classification of Road Conditions Using
Convolutional Neural Networks. *Applied P. Ahmadi, A., Khalesi, S., & Golroo, A.
Artificial Intelligence, 35*(11), 803-833. (2021). An integrated machine learning
https://doi.org/10.1080/08839514.2021.193 model for automatic road crack detection
5590 and classification in urban areas.
https://www.tandfonline.com/doi/full/10.1080 *International Journal of Pavement
/08839514.2021.1935590?fbclid=IwAR2xHn Engineering, 23*(1), 1-17.
fDzxEctWHxAwNAj3ip8IEqAsSEcAwIaBRcs https://doi.org/10.1080/10298436.2021.190
Sf6Q0abumt3lgE4qLo 5808
https://www.researchgate.net/publication/35
M. Barodi, A., Bajit, A., Benbrahim, M., & 2412295_An_integrated_machine_learning
Tamtaoui, A. (2021). Improving the transfer _model_for_automatic_road_crack_detectio
learning performances in the classification n_and_classification_in_urban_areas
of the automotive traffic roads signs. *E3S
Web Conf., 234*, 00064. Q. Bailo, O., Lee, S., Rameau, F., Yoon, J.
doi:10.1051/e3sconf/202123400064 S., & Kweon, I. (2017). *Robust Road
https://www.e3s-conferences.org/articles/e3 Marking Detection and Recognition Using
sconf/abs/2021/10/e3sconf_icies2020_0006 Density-Based Grouping and Machine
4/e3sconf_icies2020_00064.html Learning Techniques*.
doi:10.1109/WACV.2017.90
N. Baek, J.-W., & Chung, K. (2020, https://www.researchgate.net/publication/31
September 23). Pothole Classification 2286941_Robust_Road_Marking_Detection
Model Using Edge Detection in Road _and_Recognition_Using_Density-Based_G
Image. *Applied Sciences, 10*(19), 6662. rouping_and_Machine_Learning_Technique
https://doi.org/10.3390/app10196662 s
https://www.researchgate.net/publication/34
5399617_Pothole_Classification_Model_Usi R. Ma, L., et al. (2021). Capsule-Based
ng_Edge_Detection_in_Road_Image/link/60 Networks for Road Marking Extraction and
9d750a458515c2658ca34b/download?_tp= Classification From Mobile LiDAR Point
eyJjb250ZXh0Ijp7ImZpcnN0UGFnZSI6InB1 Clouds. *IEEE Transactions on Intelligent
YmxpY2F0aW9uIiwicGFnZSI6InB1YmxpY2 Transportation Systems, 22*(4), 1981-1995.
F0aW9uIn19 doi:10.1109/TITS.2020.2990120
https://ieeexplore.ieee.org/document/90878
O. Sutikno, H. A. W., & Budiarto, P. Y. 59?fbclid=IwAR07t8UIuydU4WdBnLnVhMwj
(2017). Classification of Road Damage from 9EfKD5YUhHn1VjoPT8_ylA__w7sxEGkzw
Digital Image Using Backpropagation R8
Neural Network. *IAES International Journal
of Artificial Intelligence (IJ-AI), 6*(4),
S. Chun, C., Lee, T., Kwon, S., & Ryu, S.
(2020). Classification and Segmentation of
Longitudinal Road Marking Using
Convolutional Neural Networks for Dynamic
Retroreflection Estimation. *Sensors,
20*(19), 5560. doi:10.3390/s20195560
https://www.researchgate.net/publication/34
5869913_Classification_and_Segmentation
_of_Longitudinal_Road_Marking_Using_Co
nvolutional_Neural_Networks_for_Dynamic
_Retroreflection_Estimation/link/60a02b18a
6fdcccacb59ba16/download?_tp=eyJjb250Z
Xh0Ijp7ImZpcnN0UGFnZSI6InB1YmxpY2F
0aW9uIiwicGFnZSI6InB1YmxpY2F0aW9uIn
19

T. Chen, Z., Pears, N., Freeman, M., &


Austin, J. (2009). Road Vehicle
Classification using Support Vector
Machines. In Proceedings - 2009 IEEE
International Conference on Intelligent
Computing and Intelligent Systems, ICIS
2009 (Vol. 4, pp. 214-218).
doi:10.1109/ICICISYS.2009.5357707
https://www.researchgate.net/publication/22
4092629_Road_Vehicle_Classification_usin
g_Support_Vector_Machines

You might also like