Algorithms 17 00103 v2
Algorithms 17 00103 v2
Review
Object Detection in Autonomous Vehicles under Adverse
Weather: A Review of Traditional and Deep
Learning Approaches
Noor Ul Ain Tahir 1,† , Zuping Zhang 1, *,† , Muhammad Asim 2,3, *,† , Junhong Chen 3,4,†
and Mohammed ELAffendi 2
1 School of Computer Science and Engineering, Central South University, Changsha 410083, China;
214718021@csu.edu.cn
2 EIAS Data Science and Blockchain Laboratory, College of Computer and Information Sciences, Prince Sultan
University, Riyadh 11586, Saudi Arabia; affendi@psu.edu.sa
3 School of Computer Science and Technology, Guangdong University of Technology,
Guangzhou 510006, China; junhong.chen@uhasselt.be
4 Expertise Centre for Digital Media, Flanders Make, Hasselt University, 3500 Hasselt, Belgium
* Correspondence: zpzhang@csu.edu.cn (Z.Z.); asimpk@gdut.edu.cn or masim@psu.edu.sa (M.A.)
† These authors contributed equally to this work.
1. Introduction
Copyright: © 2024 by the authors. The World Health Organization (WHO) releases statistics each year on the number
Licensee MDPI, Basel, Switzerland. of people who have been hurt in traffic accidents. Around 1.3 million people worldwide
This article is an open access article die each year, and 20 to 50 million suffer serious injuries [1], with young males under
distributed under the terms and
25 constituting the majority, 73%, of fatal traffic accidents. Extensive research on the wor-
conditions of the Creative Commons
rying prevalence of traffic fatalities has led to the development of upgraded intelligent
Attribution (CC BY) license (https://
transportation systems (ITS) with expanded capabilities. The cutting-edge of contem-
creativecommons.org/licenses/by/
porary automotive technology is represented by automated driving systems (ADS) and
4.0/).
autonomous vehicles (AVs). AVs, especially self-driving cars, have generated a great deal
of interest in the fields of intelligent transportation, computer vision, and artificial intel-
ligence [2]. No other invention since the creation of vehicles has had such a dramatic
impact on the automotive industry as self-driving vehicles. As per the forecasts presented
in Figure 1 [3], the market for self-driving cars is expected to grow from 23.80 million
units in 2022 to 80.43 million units by 2032. These vehicles primarily rely on cameras
and sensors, and examples include LiDAR (light detection and ranging), radar, ultrasonic,
etc., which have a number of autonomous features, including reduced use of fuel, cutting
carbon dioxide emissions by up to 10%, decreased traffic congestion [4], decreased road
accidents, and utilizing roads more efficiently to optimize the transportation system. AVs
can assist in managing and studying traffic by gathering effective and valuable traffic
data. Along with other road-related capabilities, they are capable of recognizing static
and non-static objects such as vehicles, pedestrians, and road lanes. Object detection (OD)
is critical for AVs as it ensures safety by identifying and avoiding moving obstacles in
their path. Numerous research works have investigated methods for identifying objects
in AVs, broadly divided into manual, semi-automated, and fully automated systems [5],
as shown in Figure 2. Roadside artifacts are visually assessed by human inspectors using
manual procedures; nonetheless, subjectivity and labor intensity are drawbacks. While
semi-automated methods gather data from moving cars and process them afterward, they
are still labor-intensive [6]. High-resolution cameras and sensors on cars record pictures
and videos in completely autonomous systems. Software-based models that have been
trained to detect objects evaluate the data either after they have been collected or in real
time. These techniques seek to solve safety problems while improving the precision and
effectiveness of item identification on roadways and streets. Vehicle-based surveillance is a
widely used technique since it is a quick and efficient way to examine objects. However,
it has always been difficult for autonomous programs to identify objects under adverse
weather conditions such as rain, fog, snow, haze, storms, and low lighting effects.
81.00 $80.43
72.90 $70.93
64.80 $62.6
56.70 $55.3
$48.9
48.60
$43.27
40.50 $38.33
$33.98
32.40 $30.15
$26.78
24.30 $23.8
16.20
8.10
0.00
2022 2023 2024 2025 2026 2027 2028 2029 2030 2031 2032
Figure 1. The global self-driving cars market size 2022–2032 in USD Million [3].
The weather has a number of negative consequences for mobility and traffic. Pre-
cipitation is recorded roughly 11.0% of the time worldwide on average [7]. Studies have
unequivocally shown that, in comparison to typical weather, rainfall can cause a 70%
increase in the likelihood of accidents [8]. Moreover, snowfall occurs in 77% of the countries
worldwide. For example, according to US national data, annually, 24% of vehicle accidents
related to weather conditions take place on slippery, snowy, slushy, or icy roads, while
15% occur when snow is actively falling or mixed with sleet [9], highlighting the real risks
associated with snowy weather. Visibility is greatly reduced by environmental variables
such as fog, haze, sandstorms, and strong sunlight, which presents real difficulties for
drivers [10].
Algorithms 2024, 17, 103 3 of 36
Semi Fully
Manual
Automated Automated
Traditional and deep learning (DL) algorithms are crucial for enhancing OD in au-
tonomous vehicles, especially in adverse weather. While traditional AI, based on statistical
learning, has been foundational, it faces limitations due to manual feature engineering
and limited adaptability to dynamic environments, requiring extensive retraining and
hindering its effectiveness in real-time applications.
DL offers a powerful alternative to traditional machine learning methods for OD
in AVs, addressing the limitations of manual feature engineering and adaptability in
dynamic environments. DL’s multi-layer neural networks automatically extract features
and learn from data, providing a more flexible and adaptable solution that enhances the
performance of surveillance systems, self-driving vehicles, and smart city applications [11].
DL’s neural network-based approach is particularly adept at handling complex models that
traditional techniques cannot [12]. In AV object detection, hybrid models combining one-
stage and two-stage detectors are effective, with two-stage models focusing on accuracy
and one-stage models on speed, although the reasons for the latter’s lower accuracy are
not fully understood.
The research significance of this paper lies in its comprehensive analysis of the chal-
lenges and advancements in OD for AVs in adverse weather conditions. It provides a
critical review of both traditional and DL approaches, offering insights into the limitations
and potential improvements of the current detection algorithms. The paper contributes to
the broader field of intelligent transportation systems by emphasizing the need for robust
and reliable detection systems that can operate effectively in a variety of weather scenarios,
which is crucial for the safe deployment of AVs in real-world conditions. Building on this
significance, this paper also contributes to the discussion regarding AV architecture and
challenges in adverse weather, and reviews the literature on detecting pedestrians, vehicles,
and road lanes using traditional and DL methods. It also summarizes common evaluation
metrics for OD. In this paper, we contribute to the field by examining the fundamental
architecture of AVs and the specific challenges they face in adverse weather conditions. We
have compiled comprehensive datasets, leveraging real-world statistics from LiDAR and
camera sensors, to provide a robust foundation for our analysis. We detail the core structure
of OD systems and elucidate both traditional and DL methodologies for AVs. Building
upon these approaches, we provide a critical review of the existing literature, focusing
on the detection of three primary objects—pedestrians, vehicles, and road lanes—under
challenging weather conditions. The structure of this paper is shown in Figure 3.
Algorithms 2024, 17, 103 4 of 36
Section 1
Introduction Sensor
Technologies
Section 2
RelatedSurvey Actuators
O Section 3
Overview of AVs Perception
R
G Section 4 Planning and
A Overview of OD Control
N Section 5 Challenges for
I Overviewof AVs under
Traditional / Adverse
Z DLApproaches weather
A
T Section 6
Previous work in
I OD
O Vehicle Pedestrian Road Lane
Section 7
N Discussion /
Detection Detection Detection
Limitation/ Future
Work
Section 8
Conclusion
knowledges that detecting smaller, obscured, or partially visible objects remains a challenge,
indicating a need for further research in this area. The study in [18] serves as a tutorial
on advanced techniques addressing the impact of rainy conditions on AV object detection.
The study evaluates the performance of detection methods under clear and rainy scenarios,
highlighting the efficacy of deraining methods, DL-based domain adaptation, and image
translation frameworks. However, it does not comprehensively address the limitations of
these methods in other adverse weather conditions. The study in [19] examines vehicle
detection approaches in video-based traffic surveillance, highlighting the challenges posed
by dim illumination, variable weather, occlusion, and shadows. The study’s exploration
of traditional approaches is valuable, but it does not fully integrate the advancements
in DL methods for improved detection. Other [20] research focuses on DL methods for
detecting on-road vehicles, providing a valuable list of works. However, the study’s limited
scope does not encompass the full range of challenges faced by AVs in adverse weather
conditions. Another paper [21] exclusively examines 3D methods for vehicle detection,
offering insights into a specific detection approach. Yet, the study’s narrow focus does
not address the broader context of AVs and their performance under various weather
conditions.
In summary, while the surveyed literature provides a valuable foundation for un-
derstanding AV sensor technologies and detection techniques, there is a critical gap in
the holistic examination of both traditional and DL approaches under adverse weather
conditions. This survey paper focuses on traditional approaches, deep learning approaches,
and their applications in object detection in AVs under adverse weather conditions. Fur-
thermore, we have detailed our literature review methodology, which involves a systematic
search across reputable digital repositories such as Springer Link, Elsevier, IEEE, MDPI, and
Wiley Online Library. We utilized targeted keywords to identify relevant research articles
and focused on studies published in peer-reviewed journals and international conferences
from 2001 to 2023. This approach ensured a comprehensive and high-quality selection of
literature for our review.
3. An overview of AVs
The Society of Automotive Engineers (SAE) has established a classification system
that divides Autonomous Driving Systems (ADS) into six levels of automation, ranging
from Level 0 (No Automation) to Level 5 (Full Automation), as illustrated in Figure 4 [22].
This system is pivotal in understanding the evolution of autonomous vehicles (AVs), which
are equipped with advanced sensors and software to navigate independently, thereby en-
hancing safety technologies and autonomy. Such a progression necessitates a collaborative
effort among scientists and engineers across various disciplines to address the complex
challenges associated with AV development [23]. The SAE automation levels provide a
standardized framework that is critical for evaluating the capabilities and limitations of
AVs at different stages of their development. By categorizing AVs into these distinct levels,
it becomes possible to systematically assess the technological advancements and hurdles
encountered at each level. This structured approach is instrumental in setting realistic
expectations about AV performance, shaping regulatory frameworks, and educating the
public on the operational capabilities of AVs under various scenarios. Thus, the classifi-
cation of AVs according to SAE levels is essential for advancing the field of autonomous
driving, guiding its regulatory landscape, and informing societal understanding of these
emerging technologies.
Algorithms 2024, 17, 103 6 of 36
Initially, in the 1920s, AVs were referred to as “phantom autos” due to their lack of a
human driver, relying on remote control [24]. Significant advancements in AV technology
emerged in the 1980s, notably with Pomerleau’s “Autonomous Land Vehicle in a Neural
Network” project [25]. This project demonstrated the potential of neural networks to signif-
icantly enhance autonomous navigation systems. The Defense Advanced Research Projects
Agency (DARPA) initiated the inaugural Grand Challenge in 2004 with the primary aim
of stimulating research and the advancement of AVs. Subsequent to the 2004 challenge,
DARPA hosted additional grand challenge events in 2005 and 2007, famously recognized
as the Urban Challenge [25]. Developments in recent times have resulted in the global
implementation of AVs. Some manufacturers have achieved Level 4 SAE standard [26]
autonomy through the use of LiDAR technology. Google started secretly building its first
AV in 2009, and, on May 1, 2012, in Las Vegas, they successfully completed their first
autonomous driving test [27]. The UK government launched a competition in 2014 with
the goal of encouraging and developing AVs [28]. The Mcity driverless shuttle project was
started recently by the University of Michigan. It is the first Level 4 automated shuttle
project in the US [29]. However, due to safety considerations, AVs cannot operate in severe
rain or snow. Although testing and research have been carried out in inclement weather,
problems such as the persistent usage of windshield wipers may result in an operational
stoppage. Low temperatures impacted charging during the night and limited daily operat-
ing hours for the Sohjoa Baltic Project [30]. Tesla’s autopilot works well in mild rain and
snow and has obvious lane markers, but it struggles in inclement weather [31]. Regulation
(EU) 2019/2144 was introduced by the European Parliament and Council in 2019. This was
the first time that guidelines pertaining to automated and fully autonomous vehicles had
been established [32]. When Waymo (formerly Google’s self-driving division, rebranded
as Waymo in 2016) announced that their “Waymo Driver” had successfully completed
20 million self-driven miles and carried out simulations that equated to 15 billion miles, it
marked an important turning point in the growth of AVs [33]. While researchers are striving
to develop sensors for different weather scenarios, thorough studies that address every
facet are still absent. The AVs use environment-sensing skills to function with the least
amount of human interaction [34]. Using an autonomous driving framework, the weather
has a significant impact on the data obtained from car sensors and the general state of the
surroundings. Adverse weather conditions create obstacles to object identification, moni-
toring, and localization, necessitating modifications to control and planning techniques.
Furthermore, the efficiency of the vehicle can be directly impacted by meteorological factors
like wind, general weather, and road conditions.
This interaction results in an ongoing cycle whereby changes in the environment and
vehicle states impact one another, highlighting the weather’s crucial role in autonomous
driving. Figure 5 shows the architecture of AVs.
Algorithms 2024, 17, 103 7 of 36
External Factors
Perception
Sensors Planning
Autonomous Car
Static & non static
objects Actuators Control
Weather External
forecasts V2V, V2I, VLC, V2X
3.1.1. LiDAR
LiDAR is considered the primary perceptive sensor in the self-driving car industry.
Even though 3D-LiDAR has only been used in automobiles for slightly over ten years, it has
already been shown to be crucial to the development of AVs and ADS [36]. Its exceptional
ability to measure things precisely and its resilience to changing lighting conditions high-
light how vital it is. Critical features of 3D scanning using lasers include its form aspect,
affordability, tolerance to external variables, measuring range, precision, pinpoint density,
scanning acceleration, configurational versatility, and spectral qualities [37]. These sensors
work well in a range of weather conditions by using laser beams to measure distances and
produce accurate 3D point clouds. However, there could be problems when there is a great
deal of rain, snow, or fog since laser beams can disperse and potentially skew the data.
Older types, such as the Velodyne HDL64-S2 [38], continue to function better in dense fog,
while the performance of more modern laser scanners decreases. While making manual
modifications can increase performance, it is still difficult to perceive reliably in deep fog.
For this reason, alternative technologies, including gated imaging, should be considered
for better performance in adverse conditions [39].
Algorithms 2024, 17, 103 8 of 36
3.1.2. Radar
The fundamental components of ADS and self-driving automobiles are automotive
radars and a variety of additional sensors. The successful use of these technical advances
depends on complex systems with long signal-processing chains that link the radars and
sensors to the controller. The vital tasks of identifying objects and obstructions, pinpointing
their exact location, and evaluating their relative speed in relation to the vehicle fall to
automotive radar equipment. Advances in millimeter-wave (mm-wave) semiconductor
technology, along with the development of signal processing techniques, form the basis
for the progress of car radar systems. A variety of signal processing methods have been
developed to improve resolution and reliability in a range of measuring parameters, includ-
ing target velocities in the vicinity, azimuth elevation perspectives, and the radius of the
vehicle [40]. Frequency Modulated Continuous Wave radar, or FMCW radar, modifies the
broadcast signal’s frequency continually at a predetermined pace. By measuring the gap
between the sent and reflecting signals, which is inversely related to the amount of time
required for the electrical signal to make its way to and from the object being targeted, we
can determine the range due to this modulation of frequency. FMCW radar has excellent
accuracy and range resolution, as well as speed measurement features, which make it
especially useful for autonomous driving applications [41].
3.1.3. Ultrasonic
One typical vehicle sensor that is frequently left out of discussions regarding ADS
is the ultrasonic sensor. Despite this error, it is positioned strategically on the vehicle’s
bumpers and throughout its body, making it indispensable for parking assistance and
blind spot surveillance. Surprisingly, ultrasonic sensors have proven to be dependable and
reasonably priced over a long period of time, which makes them a useful sensor choice [42].
The frequency range of ultrasound sounds, which are audible only to humans, is normally
between 30 and 480 kHz. In the field of ultrasonic sensing, the frequency range that is most
frequently utilized is between 40 and 70 kHz. The resolution and sensor range are highly
dependent on the selected frequency. Longer sensing ranges are correlated with the lower
ones. For example, the measuring range is up to 11 m, and the accuracy is one centimeter
(cm) at the commonly used frequency of 58 kHz. Conversely, higher frequencies such as
300 kHz provide amazing resolution, maybe down to one millimeter, but at the penalty of
a shorter range, capped at about 30 cm [43]. Ultrasonic sensors are useful for close-quarters
applications like parking because their normal operating range is 11 m [44]. However, they
can be used in autonomous driving; for example, Tesla’s “summon” capability can be used
to navigate garage doors and parking lots [45].
3.1.4. Camera
As the eyes and storytellers of automobiles, cameras are essential components of ADS.
They are a crucial component of ADS, capturing the dynamic story of the surroundings,
even though they are technologically more advanced than LiDAR. Installed on Windows,
dashcams record continuously and have made a substantial contribution to the initial ADS
datasets. Fisheye-lens professional camera setups nowadays increase the volume of data
that may be collected. On the other hand, adverse weather can cause visual problems for
cameras due to rain, snow, and fog. In low light, specialized cameras improve visibility,
such as night vision versions. The potential to overcome weather-related constraints in
sensor fusion methods through the integration of these cameras is encouraging. The
function of cameras will change further as autonomous driving technology develops,
influencing sensor technologies and how they interact with the outside world. In summary,
cameras are the storytellers of autonomous cars; they are sensitive to bad weather yet
versatile enough to be essential to safer autonomous systems.
Algorithms 2024, 17, 103 9 of 36
3.2. Actuators
AVs use data from the planning stage to carry out the vehicle’s motions in the control
stage. This entails giving actuators instructions for steering, braking, accelerating, and
signaling. A drive-by-wire method is frequently used to transmit commands effectively.
The control system is in charge of creating and monitoring trajectories, making sure that
intended routes are taken. Methods for generating trajectory include sensor-based, which
is appropriate for robotics, and dynamics-based, which is more appropriate for vehicles.
The majority of the time, trajectory tracking uses geometrical or model-based techniques.
Although they have limits and only react to errors as they happen [46], feedback con-
trollers like the proportional-integral-derivative (PID) are employed to prevent deviations
from intended trajectories. Two-degree independent controllers combine feedforward and
feedback controllers to circumvent these restrictions.
3.3. Perception
To highlight the importance of perception for AVs, it is important to emphasize how
it helps the AV determine its location and its surroundings. In order to extract road fea-
tures, detect objects, anticipate their behavior, and carry out Simultaneous Localization and
Mapping (SLAM) activities, the perception module uses data from sensors and commu-
nication equipment [47]. For detecting the surroundings, a variety of active and passive
sensors, such as cameras, LIDAR, and RADAR, can be employed; each has advantages
and disadvantages of its own, as shown in Table 1. Emphasized are the benefits and
drawbacks of various sensors, illustrating the compromises made in terms of resolution,
weather sensitivity, and accuracy. One way to overcome the limits of individual sensors
is to introduce sensor fusion, a technique that combines data from several sensors. There
are two main types of AV systems discussed: one is sensor fusion, which uses various
sensors such as LIDAR, RADAR, and cameras, and the other is pure vision, which uses
cameras and computer vision algorithms. It is mentioned that a pure vision-based system
has advantages like affordability and simplicity of implementation. In the parts that follow,
this research concentrates on the benefits of a pure vision-based method [48].
Technical parameters
General Weather Road Weather
Slow down when Extended
it's foggy outside. headways are
necessary in
foggy scenarios.
Sensor performance
Driver behavior General Weather Road Weather
to problems like solar flares and thunderstorms. Communication interruptions can result
from bad weather such as fog, heavy rain, snow, and hail obstructing the visible light
communication (VLC) line of sight.
The process of OD in AVs unfolds in several key steps, beginning with the collection
of image data by the sensor system. Following this, a Region of Interest (ROI) or regional
proposal strategy is employed, marking a critical phase in object tracking. This approach is
commonly utilized in devices such as cameras and stereoscopic systems, where ROIs serve
as the foundational elements for tracking objects. The subsequent step involves the use of
classifiers to meticulously analyze and refine shapes, edges, and lines. This analysis aids in
the classification of targets and their identification based on human categories, embedding
these ROIs within the image for object recognition.
Various techniques, including sliding windows, Locally De-correlated Channel Fea-
tures (LDCF), and constrained searches, are applied to locate ROIs efficiently. The extraction
of features from these ROIs represents the third stage of the process. At this juncture, ei-
ther DL-based methods or traditional manual OD algorithms can be employed for the
extraction and classification of features. While manual feature extraction techniques can
be advantageous for models that rely on lower-level features by facilitating manual ROI
suggestions [52], their efficacy diminishes when dealing with complex features, rendering
them less reliable. However, to derive useful and relevant data from video or photographs
using a computing device, a great deal of computation has to be completed [53]. In contrast,
DL methodologies enable systems to autonomously learn feature representations, thereby
enhancing the accuracy and adaptability of feature extraction in dynamic and complex
driving environments.
The evolution of research on OD in bad weather depends heavily on datasets. They
act as the building blocks for creating, refining, and testing algorithms and models. In this
part, we are going to discuss a few of the datasets that are frequently utilized in this field
of study. One of the most often used benchmarks for autonomous driving research is the
KITTI [54] dataset. It features a range of various metropolitan settings. The collection offers
LiDAR data, excellent-quality pictures, and annotations for cars. A2D2 [55] is a sizable
dataset and delivers extensive annotations for several items, including vehicles, together
with high-resolution camera and LiDAR data. Waymo’s dataset [56], which is renowned for
its comprehensive and varied gathering of data from AVs and includes information from
a range of weather and lighting circumstances, is a useful tool for research on inclement
weather. Among the noteworthy datasets, NuScenes [57], BDD100K [58], ApolloScape [59],
and SYNTHIA [60] datasets offer diverse and extensive resources for exploring vehicle
detection in challenging conditions. A quick overview of the weather conditions consid-
ered and the particular sensors utilized for gathering data for each dataset can be found in
Table 2. The University of Michigan used a Segway robot to collect four seasons of LIDAR
data for the NCLT [61] on campus. The first AV dataset, called the Canadian Adverse
Driving Conditions (CADC) [62] dataset, focuses on snow conditions. The Challenging
Conditions Dataset with Correspondences (ACDC) [54], created for training and testing lex-
ical techniques for segmentation in difficult visual conditions, was later introduced by the
same team. ACDC includes a high-quality pixel-level resolution collection of well-labeled
photos encompassing various categories of mist, at night, rain, and snowfall. The public’s
Digital Video Collection now includes the IUPUI [63] Driving Video/Image Benchmark.
Example views of varied lighting and roadway conditions captured by in-car webcams
are included in this baseline. In these movies, patrolling and disaster response situations
involving safety automobiles are shown. Together, these datasets provide priceless re-
sources for developing the discipline and promoting creativity in the creation of reliable
and weather-adaptive vehicle detection systems.
Algorithms 2024, 17, 103 12 of 36
3. Road Lane Detection: Road lane detection is the ability of AVs to identify and under-
stand the position and orientation of road lanes. This information is essential for the
vehicle to navigate within its designated lane, follow traffic rules, and make correct
turns, ensuring a smooth and safe driving experience.
1 2 3
a) Sensor
(a) (b)
b) ROI
(c)
c) Feature Extraction
(d)
d) Classification
Deep learning
performance
Traditional
approaches
Data
The Viola–Jones approach (Viola and Jones, [84] was developed for human face identi-
fication using Haar-like feature extractors. It consists of four primary parts; the primary
image algorithm for effective image participation is AdaBoost. Haar-like feature extraction
is used to uncover biased features. Cascade classifier implementation is used to filter out
background data and concentrate on regions that are more likely to contain the object
of interest. The algorithm performs remarkably well, taking only 0.067 s to process a
384 × 288 pixel image. Its detection rate is 91.4%, and it produces 50 false positives. Never-
theless, it is not well adapted for the generic recognition of objects and has slow periods
for training. The HOG method was created for identifying humans in digital photos and
was first presented in [85]. In order to determine the gradient’s length and direction, this
method examines input images employing gradient filters, both horizontal and vertical.
The filtered images are divided into 8 × 8 pixel cells, which are then further divided into
2 × 2 cell blocks that have 50% overlap. For every cell, orientation histograms are produced,
quantizing gradient directions into nine bins ranging from 0◦ to 180◦ . The magnitudes of
these bins are used as votes. A vector representing the HOG is produced by concatenating
these histograms for every cell in a block. To take into consideration changes in illumination
and contrast, these vectors are standardized. The final HOG descriptor, which consists of
all the normalized blocks, is fed into a Support Vector Machine (SVM) to determine if an
image is human or not. Although the HOG algorithm’s dense grid method results in a
larger processing burden and less efficient performance in cases involving obscured objects,
it is effective in decreasing false positives when compared with Haar wavelets.
These AI techniques, which have their roots in statistical principles, have traditionally
formed the basis of machine learning models. These techniques enable computers to
recognize trends and forecast outcomes based on past data. However, even with their
historical importance, traditional methods have clear drawbacks. Reliance on manual
feature engineering, as shown in Figure 9, in which experts carefully craft features for
the model necessitates domain knowledge and frequently fails to capture the nuances of
intricate datasets. Furthermore, these models’ ability to adjust to unfavorable weather
or dynamic surroundings is hindered by their inability to quickly incorporate new data
without requiring significant retraining, which limits their usefulness in situations that
change quickly. The development of more sophisticated methods, or DL approaches, to
overcome these obstacles and improve the power of AI systems has been made possible by
this realization of their limitations.
Car/not
Features
a car
Car/not
a car
5.2. DL Approaches
DL emerged as a subfield of machine learning and artificial intelligence in the 1990s [86].
Unlike traditional methods, DL offers distinct advantages, including the ability to achieve
higher levels of abstraction, improved precision, and faster execution. These advantages
make DL a valuable choice for OD. The OD algorithm employing DL typically consists of
Algorithms 2024, 17, 103 15 of 36
three core components: a recurrent neural network (RNN), a depth belief network (DBN),
and a convolutional neural network (CNN). Overall, object detectors based on CNN [87]
can be categorized into two primary types.
1. Single or one-stage detectors are known as the “non-regional proposal method” and
“dense prediction”.
2. A two-stage detector is known as the “regional proposal method” and “sparse
prediction,” as shown in Figure 10. In a single-stage detector, all the tasks are integrated
into a unified system structure. Conversely, a two-stage detector separates the system
into distinct stages for region selection, classification, and localization. Some regional
proposal techniques consist of region-based CNN (R-CNN) [88], Fast R-CNN [89], Faster
R-CNN [90], Spatial Pyramid Pooling Networks (SPPNet) [91], and Feature Pyramid Net-
work (FPN) [92]. On the other side, the non-regional category includes Single Shot Multi-
box Detector (SSD) [93], You Only Look Once (YOLO 1-8) [94], EfficientDet [95], DEtection
TRansformer (DETR) [96], and Fully Convolutional One-Stage (FCOS) [97]. Since these
methods constitute the foundation of CNN, they have emerged as the standard for OD. The
amalgamation of one-stage and two-stage OD techniques has gained prominence in the
field of OD within AVs. These hybrid approaches have demonstrated effectiveness across
diverse scenarios, achieving promising results in precise OD and localization. Two-stage
algorithms tend to deliver superior accuracy, while one-stage algorithms offer faster pro-
cessing speeds. Notably, the reasons behind the lower accuracy of one-stage algorithms
remain unclear. A study examining the drawbacks of one-stage algorithms, especially those
with dense sampling, was conducted in Ref. [98]. The study found that there was a major
problem with the unbalanced background samples (negative instances) and foreground
values (positive examples).
Object Detection
To solve this problem, the authors created Focal Loss, which alters the traditional
cross-entropy loss function to give problematic cases more weight and thereby puts the
emphasis on positive samples. After this adjustment, the RetinaNet network was created,
showing better accuracy and quicker detection times than one- and two-stage methods.
It should be mentioned too that the approach necessitates adjusting the focusing value,
an extra hyperparameter [99]. However, the two primary state-of-the-art methods for
DL-based detection these days are the YOLOv8 from the YOLO series (YOLOv1–YOLOv8)
and the Faster R-CNN from the R-CNN family.
Algorithms 2024, 17, 103 16 of 36
In order to speed up the region proposal step and allow the system to acquire knowl-
edge of optimal region proposals, Ren et al. [90] proposed Faster R-CNN as an improve-
ment to Fast R-CNN. The architecture of Faster R-CNN is described in Figure 11. This
was achieved mainly by integrating the region proposal network (RPN). RPN’s main ob-
jective is to produce proposals with various ratios of aspects and sizes. Similar to Fast
R-CNN, it employs high-resolution feature maps as input for RPN to identify the regions
of interest in an image rather than depending on selective search for region recommenda-
tions. While Faster R-CNN improves the precision of detection and reaches an evaluation
time of 0.2 s (five frames per second), it still requires a large amount of processing power
and cannot fully meet the real-time system demands. It is also more complex because
it requires training two interconnected neural networks, which makes system optimiza-
tion difficult [100]. While Faster R-CNN algorithms demonstrate satisfactory accuracy on
COCO and ImageNet datasets, they encounter difficulties when it comes to detecting small
objects and objects at varying scales. This approach, while being more memory-efficient
and faster, faces challenges in achieving consistent performance during both training and
test-time inferences.
Feature maps ROI Pooling
Classifier
Proposals
Jocher [100] implemented YOLOv8; however, that work has not yet been published in
a journal. The structure of YOLOv8 is shown in Figure 12. Different versions are available
with different scales of YOLO (N, S, M, L, and X) according to the scaling coefficient. The
two main differences are the variety of kernels at different locations in the network and
the number of feature extraction modules. These modifications are intended to meet the
unique requirements of various situations. The backbone, neck, and head are the three
primary modules that make up the YOLOv8 networks. First, the backbone module extracts
features from the input photos. The neck component then performs feature fusion to
create features in three different sizes: large, medium, and small. The final output is the
detection results. Eventually, these characteristics of various sizes are sent to the head
section for identification of the object. To improve deep image extraction features, the
YOLOv8 system uses a layered C2f technique inside the backbone section. The main
function of the SPPF layer, which is the tenth layer in the backbone module, is to transform
input feature maps—regardless of their original size—into a fixed-size feature vector. The
goal of this change is to increase the responsiveness of the network. YOLOv8 uses the
traditional top-down as well as bottom-up feature fusion technique referred to as PAN
(Path Aggregation Network) in the neck component. This tactic successfully combines
elements from several scales. YOLOv8 adopts the widely used decoupled head architecture
in the head section, which divides the categorization head from the identification head.
Furthermore, there is a noticeable change in the network’s OD technique: the anchor-free
strategy replaces the anchor-based method.
Algorithms 2024, 17, 103 17 of 36
Conv
C2f
C2f Detect
Conv
Conv
Concat
C2f
upsample
Conv Concat
C2f
C2f Concat
C2f Detect
Conv
upsample Conv
C2f
SPPF Concat
C2f Detect
Backbone
Figure 12. Structure of YOLOv8.
to exterior objects, adverse weather, and lighting variations. Even though they help to
identify objects, biased classifiers like SVM and ANN might perform poorly in scenarios
involving complicated traffic, which can result in false positives and make them susceptible
to changes in the background, obstacles, lighting, and dimensions.
One study [101] investigated issues with physical detection techniques and equipment
maintenance by examining the application of SVM and Pyramid Histogram of Words
(PHOW) for traffic and road condition detection. The experimental findings show an
accuracy of more than 82.5% on average under different weather circumstances. The
suggested technique efficiently divides photos into four categories based on the type of
traffic and road conditions. However, the paper does not detail the dataset’s diversity, which
could impact the model’s robustness and generalizability. The research in [102] tackled the
requirement for a reliable traffic surveillance system that takes into consideration variables
like headlights, reflections, and shadows and operates well in a variety of weather and
lighting circumstances. The suggested method is based on pattern classification and uses a
distinctive image’s histogram as a distinguishing characteristic for classification. Because
of its excellent generalization skills, SVM is used as a classifier. In tests, the system can
discriminate between headlights and shadows with good accuracy and recognize moving
cars in a variety of weather and lighting scenarios. However, this paper does not address
the system’s performance under high-density traffic or its computational efficiency, which
could be critical limitations for real-time applications. Using webcams as weather sensors,
researchers [85] came up with a novel technique that allowed them to identify over 180,000
annotated photographs depending on indoor or outdoor circumstances with a 98% accuracy
rate using SVM. The results of the research [103] indicate that DPM effectively addresses
occlusion difficulties, even though it might not be required for ordinary detection settings.
Accuracy is improved by adding image context information, and system speed is increased
by adding other features. On the other hand, the multi-scale system only slightly improves
accuracy, and the classifier has little effect on detector quality [21,104,105].
The vehicle detection method in [106], which utilizes pseudo-visual pursuit and HOG-
LBP feature fusion, showcases the superiority of traditional machine learning techniques in
achieving high detection accuracy and speed in severe weather conditions. This approach,
with its reliance on radar information, demonstrates a clear advantage over methods that
do not incorporate such advanced sensor data, particularly in scenarios where visibility
is compromised. In contrast, the work by Abdullah et al. [107], while also effective, high-
lights the limitations of traditional methods when solely relying on MOG 2 background
subtraction and histogram equalization. Their system’s high effectiveness rate of 96.4%
across various weather conditions suggests that traditional techniques can still perform
well. The vehicle detection system by Wu and Juang [108], which uses histogram extension
and the Gray-level Differential Value Method (GDVM), shows that traditional methods
can be effective in complex traffic conditions. However, the system’s reduced accuracy in
rainy and dark conditions suggests a limitation of traditional approaches when dealing
with significant environmental variations. The authors in [109] developed a system for con-
current vehicle and weather detection using optical flow and color histogram approaches.
While this system demonstrates the utility of traditional methods in distinguishing between
raindrops and normal conditions, it also indicates that these methods may struggle with
the nuanced recognition tasks required in adverse weather.
Tian et al. [110] proposed a method for vehicle detection and tracking at night using
HOG features and SVM for object recognition, effectively leveraging traditional machine
learning techniques to enhance video surveillance in low-light conditions. Kuang et al. [111]
achieved a high detection rate of 95.82% with minimal false positives in their multiclass
system for nighttime conditions, employing tensor decomposition and object proposal
methods. Furthermore, the research in [112] presents a system that integrates rule-based
algorithms with a shallow neural network for nighttime vehicle detection, demonstrat-
ing significant improvements over traditional methods in terms of detection speed and
proactive capabilities. These studies highlight the strengths of both traditional and hybrid
Algorithms 2024, 17, 103 19 of 36
approaches in the realm of nighttime vehicle detection, providing a foundation for future
advancements in this field.
A vehicle tracking and counting system with camera shaking was presented in [113]
for difficult sandy weather conditions. The method used particle filter tracking for vehicle
detection, integrated headlight analysis, and background subtraction. While the method
worked well in difficult situations, such as evenings and sandy weather, with a vibrating
camera, it had trouble identifying white automobiles in snowy settings. Similarly, the
background subtraction method encountered difficulties when there was movement in
the backdrop or vibrations from the camera. A vehicle detection and tracking system for
complex traffic was described in [114]. It used a background reduction technique with
“low rank + sparse” disassembly and performed well in both qualitative and quantitative
evaluations. Table 3 tabulates these works.
various batch sizes (128, 256, 512, and 1024) is recommended to enhance accuracy and
generalization. Testing on larger datasets enhances classifier accuracy and widens learning
scope. In the realm of two-stage detectors, Ref. [131] introduces a practical approach for
vehicle detection in challenging conditions like fog, low light, and sun glare, leveraging
three trained CNNs (ResNet, GoogleNet, and SqueezeNet) with transfer learning. No-
tably, ResNet achieves a validation accuracy of 100%, while GoogleNet and SqueezeNet
achieve 65.50% and 90.0%, respectively. Additionally, Ref. [132] employs Fast R-CNN
for day and night vehicle detection, yielding impressive results under adverse conditions,
validated by high recall (98.44%), accuracy (94.20%), and precision (90%). Furthermore,
the on-road vehicle detection method proposed in [133] utilizes multiple region proposal
networks (RPNs) in Faster R-CNN to identify vehicles across diverse sizes and weather
scenarios, outperforming the existing techniques with high average precision rates on
various datasets, including DAWN (89.48%), CDNet 2014 (91.20%), and LISA (95.16%). On
the other hand, regarding one-stage detectors, another study [134] introduces an improved
SSD-based front vehicle recognition algorithm for smart automobiles. Tested on the KITTI
dataset, it achieves a high mAP of 92.18% and processes frames quickly at 15 ms per
frame. The system enhances smart car safety in congested traffic and adverse weather
conditions, prioritizing both accuracy and real-time performance. Another study [135]
enhances a YOLO-based algorithm for precise vehicle detection in foggy weather by inte-
grating a dehazing module with multi-scale retinex. This enhanced model, trained with
augmented data, surpasses traditional YOLO in foggy conditions. Additionally, Ref. [136]
proposes a modified YOLO model for on-road vehicle detection and tracking across various
weather conditions. Utilizing a single CNN, it exhibits robustness and outperforms the
existing techniques in intelligent transportation systems. Furthermore, Miao et al. [137]
developed a nighttime vehicle detection method using fine-tuned YOLOv3, outperforming
Faster R-CNN and SSD in terms of accuracy and efficiency. They achieved an average
precision of 93.66%. Another study [138] utilized YOLOv4 with SPP-NET layers for ve-
hicle detection, achieving an 81% mAP. In contrast, the study in [139] focused on harsh
weather conditions, introducing YOLOv4 with an anchor-free and decoupled head, albeit
achieving a 60.3% mAP and focusing exclusively on a single class. Moreover, the goal
of [140] was to enhance self-driving vehicle detection in adverse weather using YOLOv5
with Transformer and CBAM modules, achieving an impressive mAP of 94.7% and FPS of
199.86. The DL approach proposed in [141] for nighttime vehicle detection in autonomous
cars, combining a Generative Adversarial Network for image translation and YOLOv5
for detection, achieved a high accuracy of 96.75%, significantly enhancing the reliability
of AV recognition models for night conditions. This study in [12] presents a DL-based
intelligent AV weather-detecting system. Using a combined dataset from MCWDS2018 and
DAWN2020, the performance of three deep convolutional neural networks was evaluated
by categorizing six weather categories: overcast, rainy, snowy, sandy, sunny, and sunrise.
The CNNs are SqueezeNet, ResNet-50, and EfficientNet-b0. The ResNet-50, EfficientNet-
b0, and SqueezeNet models achieved 98.48%, 97.78%, and 96.05% accuracy, respectively,
in the experiments, demonstrating remarkable accuracy rates while preserving quick in-
ference times using GPU acceleration. By combining previously disparate datasets, the
study’s novel methodology makes a significant addition to DL applications for autonomous
systems’ weather detection.
In simple traffic situations, the aforementioned extraction and classification algorithms
have shown themselves to be successful at detecting vehicles. However, issues including
scale sensitivity, occlusion, and a high number of false positive results limit their efficacy
in more complex traffic conditions [142]. These algorithms perform well in simple traffic
scenarios, but they are limited in complex traffic conditions due to a significant number of
false positive detections. As a result, an explanation connected with vehicle detection is
intended for this part, and the difficulties associated are also highlighted.
Algorithms 2024, 17, 103 22 of 36
stereo-vision technology with two cameras. The integration of data from various sensors,
such as radar or LIDAR, is proposed to enhance autonomous driving performance. The
research also advocates for the application of DL methods to refine feature extraction and
categorization, achieving an AP of 89.85% and a mAP of 90% on the CVC-09 and LSIFIR
databases. In the realm of nighttime pedestrian recognition, Ref. [150] introduces an en-
hanced Faster R-CNN architecture, which delivers significant improvements, particularly
for distant pedestrians. The study in [151] tackles the issue of optical camera-based pedes-
trian recognition under adverse weather conditions by employing a cascaded classification
technique that incorporates both local and global features, resulting in increased detection
rates. The effectiveness of this strategy is demonstrated in challenging weather conditions
such as snow or rain. The use of YOLO-v5 in [152] introduces a novel multispectral method
for pedestrian detection, combining independent subnetworks for RGB and IR images
to create an effective detector. The method’s performance on open databases, with high
mAP scores and IoU of 0.5 across various datasets, is encouraging. and another study [153]
evaluates YOLOv7, Faster R-CNN, SSD, and HoG for both vehicle and pedestrian detection
in different weather conditions. YOLOv7’s superior performance in terms of accuracy,
precision, recall, and mAP underscores the critical role of weather conditions in selecting
appropriate DL methods for autonomous driving.
Using already installed roadside webcams, real-time road weather and road state
detection algorithms were developed in [163]. Transfer learning was utilized to train
detection models utilizing three previously learned CNN architectures. When it came
to accuracy, ResNet18 outperformed the other models, scoring about 97% for weather
and 99% for surface conditions. These models can be useful in improving road safety
since they can automatically recognize and relate real-time circumstances to road networks
without the need for human intervention. They can be incorporated into advanced traveler
information systems (ATIS) for drivers and utilized to optimize winter maintenance routes.
Additionally, the models could replace manual reporting in snowplows, increasing driver
safety and accuracy. Research [164] presents a novel approach that combines computer
vision and DL to extract meteorological data from street-level photos without any image
restrictions. With recognition accuracy ranging from 91% to 95.6% across several categories,
the study presents four deep CNN models designed to identify a variety of visibility
circumstances, including dawn and dusk, daytime, evening, glare, and weather elements
like rain and snow. A study [165] presents a DL-based method for lane marker detection
in order to address the difficulties associated with classifying road markers during rainy
weather. Specifically designed for bad weather, the approach gives priority to the best
feature selection from video frames in order to counteract rain-induced blurriness, leading
to adequate classification precision even in difficult weather circumstances. In other
research [166], two different models for AV lane prediction in highway traffic situations are
presented. The study’s main focus is on the use of AI for lane prediction using a sizable
dataset from NGSIM. To showcase the effectiveness of these models, two distinct subsets
with 5000 car samples each were used. The strategy employed a range of classifiers in
the Identification Learner and different methods in Neural Net Fitting, emphasizing the
methodology’s importance in accomplishing good lane prediction without diving into
particular accuracy measures.
Due to different illumination circumstances, global thresholds for lane marker iden-
tification frequently produce disappointing outcomes. An efficient lane departure (LD)
technique is presented, enabling lane marker identification that may be used in daylight
highway scenarios as well as tunnel situations with artificial lighting. In order to obtain
lane features, the usual LD technique begins by performing pre-processing to eradicate
perspective-related distortion and isolate an ROI. In order to identify suitable lane markers
using color, shape, alignment, or geometrical measurements using the road scenario data,
two categorization methods, model-based and feature-based, are used [17]. To minimize
false positives, prospective lane markers are further refined and validated using curved
or linear lane fitting. This helps with features like adaptive speed control, automatic lane
focusing, and lane departure alert. LD devices may not be as successful due to a number of
issues, such as road obstructions, ambient lighting sources, and noisy weather. To address
these issues, a shadow-resistant methodology that makes use of Fourier transformation
and the maximally stable extreme region (MSER) approach in the blue color stream is used
as a reference [167]. The research findings [168] suggest that the minimal safe separation
between the self-absorbed car and the vehicle should be used to estimate the extraction
of ROI. According to the study, a car traveling at 110 km/h may cover a distance of 35 m
with a height of just 150 pixels. Usually, a predetermined model or set of features is used to
construct lane markers using the ROI.
In [169], the authors modified YOLOv3 and presented a brand-new BGRU-Lane
(BGRU-L) model; the method integrates spatial distribution with visual data. High accuracy
(90.28 mAP) and real-time detection speed (40.20 fps) are achieved in difficult settings
through integration utilizing the Dempster–Shafer algorithm, as demonstrated by datasets
from KIT, Toyota Technological Institute, and Euro Truck Simulator 2.
The dynamic environment of research and innovation in the field of road lane detection
for autonomous cars is shown by our thorough analysis of the literature. The abundance of
information and developing techniques indicates the ongoing commitment to increasing
the potential of self-governing systems.
Algorithms 2024, 17, 103 25 of 36
32% 40%
28%
25
20
Number of Papers
15
10
0
Traditional Approaches DL Approaches
Vehicles Pedestrians Road Lanes
14 2
12
Number of Papers
4
10
2
6
4
9 3
2
1
3
2
0
In order to improve OD, Faster R-CNN was integrated into the RPN. This resulted
in shorter evaluation times, but it also required a significant amount of processing power
and had difficulties maintaining consistent performance for small and variously scaled
objects.YOLOv8 is designed to cater to specific requirements. The YOLOv8 structure
comprises backbone, neck, and head modules, enhancing feature extraction with C2f and
employing the PAN feature fusion technique. It also adopts a decoupled head architecture
and introduces an anchor-free strategy for OD. Furthermore, we analyzed high-quality
datasets related to AVs, and some standard evaluation metrics were also reviewed and
summarized. Our discussion underscores the advancements in OD for autonomous vehicles
under adverse weather, paving the way for enhanced safety and reliability in ADs.
7.2. Limitations
1. Our literature review primarily focuses on the detection of pedestrians, vehicles, and
road lanes, which may not encompass all possible objects and scenarios relevant to
AVs in adverse weather conditions. There may be other critical elements, such as
traffic signals or animals, that warrant further investigation.
2. The detection algorithms discussed in our review may have inherent limitations, such
as difficulties in detecting small or occluded objects, which could impact the accuracy
and reliability of AVs in certain situations.
3. Our study primarily considers a range of adverse weather conditions, but there may
be other environmental factors, such as dust storms or heavy fog, that were not
extensively covered in the reviewed literature.
4. The field of AV technology is rapidly evolving, and new detection algorithms and
sensor technologies may emerge that could significantly impact the findings of our
review. Our study may not capture the most recent advancements in this area.
5. While our study includes datasets that simulate adverse weather conditions, the
simulation environments may not perfectly replicate real-world driving scenarios.
The complexity of real-world conditions, including unpredictable human behavior
and dynamic traffic patterns, can introduce additional challenges not fully captured
in simulated datasets.
6. The ethical considerations and societal acceptance of AVs, especially in challenging
conditions, are not addressed in our study. Public trust and the ethical use of AV
technology are essential factors for their successful integration into smart cities.
Algorithms 2024, 17, 103 29 of 36
8. Conclusions
This paper reviewed the traditional and DL approaches for vehicles, pedestrians,
and road lane detection in AVs under adverse weather conditions. We first studied the
architecture of AVs with sensor technologies and other components and also discussed
the challenges for AVs in adverse weather. After an overview of almost all the datasets
related to AVs covering different weather conditions, we explained the basic structure of
OD and the evaluation matrices used for it. Then, we explained the traditional approaches
and DL approaches and discussed the traditional feature extraction methods that were
prevalent for detection and classification but had limitations in adverse conditions. Manual
feature extraction made them less suitable for complex applications. DL has become
much more popular and prominent than regular algorithms in more recent times. DL
approaches explain the structure of YOLOV8 (one-stage detectors) and Faster R-CNN (two-
stage detectors). Two-stage algorithms tend to deliver superior accuracy, while one-stage
algorithms offer faster processing speeds. Notably, the reasons behind the lower accuracy
of one-stage algorithms remain unclear. In addition, the statistics about the status quo of
traditional and DL approaches for OD in AVs were provided based on the works collected
in this survey paper. We found that DL was intensively used for vehicles, pedestrians,
and road lane detection in AVs compared with the traditional approaches. Specifically,
one-stage detectors were frequently used for vehicle detection compared with two-stage
Algorithms 2024, 17, 103 30 of 36
detectors. In addition, vehicle detection was frequently studied using both traditional and
DL approaches followed by road lane and pedestrian detection. Finally, we presented a
useful discussion along with some future research directions.
Author Contributions: All authors have equally contributed. All authors have read and agreed to
the published version of the manuscript.
Funding: This work was supported by EIAS Data Science Lab, CCIS, Prince Sultan University and
also by the Special Research Fund (BOF) of Hasselt University (No. BOF23DOCBL11), and Chen
Junhong was sponsored by the China Scholarship Council (No. 202208440309).
Acknowledgments: The authors would like to thank Prince Sultan University for their support.
Conflicts of Interest: The authors declare no conflicts of interest.
References
1. World Health Organization. WHO: Global Status Report on Road Safety: Summary; Technical Report. Available online:
https://www.who.int/health-topics/road-safety#tab=tab_1 (accessed on 18 February 2024).
2. Iftikhar, S.; Asim, M.; Zhang, Z.; El-Latif, A.A.A. Advance generalization technique through 3D CNN to overcome the false
positives pedestrian in autonomous vehicles. Telecommun. Syst. 2022, 80, 545–557. [CrossRef]
3. Self-Driving Cars Global Market Size. Available online: https://precedenceresearch.com/self-driving-cars-market (accessed on
18 December 2023).
4. Badue, C.; Guidolini, R.; Carneiro, R.V.; Azevedo, P.; Cardoso, V.B.; Forechi, A.; Jesus, L.; Berriel, R.; Paixao, T.M.; Mutz, F.; et al.
Self-driving cars: A survey. Expert Syst. Appl. 2021, 165, 113816. [CrossRef]
5. Bansal, P.; Kockelman, K.M. Are we ready to embrace connected and self-driving vehicles? A case study of Texans. Transportation
2018, 45, 641–675. [CrossRef]
6. Chehri, A.; Sharma, T.; Debaque, B.; Duclos, N.; Fortier, P. Transport systems for smarter cities, a practical case applied to traffic
management in the city of Montreal. In Sustainability in Energy and Buildings 2021; Springer: Singapore, 2022; pp. 255–266.
7. Trenberth, K.E.; Zhang, Y. How often does it really rain? Bull. Am. Meteorol. Soc. 2018, 99, 289–298. [CrossRef]
8. Andrey, J.; Yagar, S. A temporal analysis of rain-related crash risk. Accid. Anal. Prev. 1993, 25, 465–472. [CrossRef]
9. National Oceanic and Atmospheric Administration, Getting Traction: Tips for Traveling in Winter Weather. Available online:
https://www.weather.gov/wrn/getting_traction (accessed on 24 October 2023).
10. Mehra, A.; Mandal, M.; Narang, P.; Chamola, V. ReViewNet: A fast and resource optimized network for enabling safe autonomous
driving in hazy weather conditions. IEEE Trans. Intell. Transp. Syst. 2020, 22, 4256–4266. [CrossRef]
11. Abu Al-Haija, Q.; Krichen, M.; Abu Elhaija, W. Machine-learning-based darknet traffic detection system for IoT applications.
Electronics 2022, 11, 556. [CrossRef]
12. Al-Haija, Q.A.; Gharaibeh, M.; Odeh, A. Detection in adverse weather conditions for autonomous vehicles via deep learning. AI
2022, 3, 303–317. [CrossRef]
13. Zang, S.; Ding, M.; Smith, D.; Tyler, P.; Rakotoarivelo, T.; Kaafar, M.A. The impact of adverse weather conditions on autonomous
vehicles: How rain, snow, fog, and hail affect the performance of a self-driving car. IEEE Veh. Technol. Mag. 2019, 14, 103–111.
[CrossRef]
14. Mohammed, A.S.; Amamou, A.; Ayevide, F.K.; Kelouwani, S.; Agbossou, K.; Zioui, N. The perception system of intelligent
ground vehicles in all weather conditions: A systematic literature review. Sensors 2020, 20, 6532. [CrossRef]
15. Yoneda, K.; Suganuma, N.; Yanase, R.; Aldibaja, M. Automated driving recognition technologies for adverse weather conditions.
IATSS Res. 2019, 43, 253–262. [CrossRef]
16. Zhang, Y.; Carballo, A.; Yang, H.; Takeda, K. Perception and sensing for autonomous vehicles under adverse weather conditions:
A survey. ISPRS J. Photogramm. Remote Sens. 2023, 196, 146–177. [CrossRef]
17. Arthi, V.; Murugeswari, R.; Nagaraj, P. Object Detection of Autonomous Vehicles under Adverse Weather Conditions. In
Proceedings of the 2022 International Conference on Data Science, Agents & Artificial Intelligence (ICDSAAI), Chennai, India,
8–10 December 2022; Volume 1, pp. 1–8.
18. Hnewa, M.; Radha, H. Object detection under rainy conditions for autonomous vehicles: A review of state-of-the-art and
emerging techniques. IEEE Signal Process. Mag. 2020, 38, 53–67. [CrossRef]
19. Abbas, A.F.; Sheikh, U.U.; Al-Dhief, F.T.; Mohd, M.N.H. A comprehensive review of vehicle detection using computer vision.
TELKOMNIKA Telecommun. Comput. Electron. Control 2021, 19, 838–850. [CrossRef]
20. Yang, Z.; Pun-Cheng, L.S. Vehicle detection in intelligent transportation systems and its applications under varying environments:
A review. Image Vis. Comput. 2018, 69, 143–154. [CrossRef]
21. Muneeswaran, V.; Nagaraj, P.; Rajasekaran, M.P.; Reddy, S.U.; Chaithanya, N.S.; Babajan, S. IoT based Multiple Vital Health
Parameter Detection and Analyzer System. In Proceedings of the 2022 7th International Conference on Communication and
Electronics Systems (ICCES), Coimbatore, India, 22–24 June 2022; pp. 473–478. [CrossRef]
22. SAE International. Available online: https://www.sae.org/standards/content/j3016_202104/ (accessed on 11 November 2023).
Algorithms 2024, 17, 103 31 of 36
23. Bimbraw, K. Autonomous cars: Past, present and future a review of the developments in the last century, the present scenario and
the expected future of autonomous vehicle technology. In Proceedings of the 2015 12th International Conference on Informatics
in Control, Automation and Robotics (ICINCO), Colmar, France, 21–23 July 2015; Volume 1, pp. 191–198.
24. Lafrance, A. Your Grandmother’s Driverless Car. The Atlantic, 29 June 2016. Available online: https://www.theatlantic.com/
technology/archive/2016/06/beep-beep/489029/ (accessed on 27 October 2023).
25. Pomerleau, D.A. Alvinn: An autonomous land vehicle in a neural network. In NIPS’88: Proceedings of the 1st International
Conference on Neural Information Processing Systems, 1 January 1988; MIT Press: Cambridge, MA, USA; pp. 305–313.
26. SAE International. Surface Vehicle Recommended Practice. Joint SAE/RCCC Fuel Consumption Test Procedure (Short Term
in-Service Vehicle) Type I. 1986. Available online: https://ca-times.brightspotcdn.com/54/02/2d5919914cfe9549e79721b12e66/j3
016-202104.pdf (accessed on 1 May 2021)
27. How Google’s Autonomous Car Passed the First US State Self-Driving Test. 2014. Available online: https://spectrum.ieee.org/
how-googles-autonomous-car-passed-the-first-us-state-selfdriving-test (accessed on 31 December 2023).
28. Claire, P. The Pathway to Driverless Cars: Summary Report and Action Plan; OCLC: London, UK, 2015.
29. Briefs, U. Mcity Grand Opening. Res. Rev. 2015, 46, 1–3.
30. Bellone, M.; Ismailogullari, A.; Müür, J.; Nissin, O.; Sell, R.; Soe, R.M. Autonomous driving in the real-world: The weather
challenge in the Sohjoa Baltic project. In Towards Connected and Autonomous Vehicle Highways: Technical, Security and Social
Challenges; Springer International Publishing: Cham, Switzerland, 2021; pp. 229–255.
31. Lambert, F. Watch Tesla Autopilot Go through a Snowstorm. Available online: https://electrek.co/2019/01/28/tesla-autopilot-
snow-storm/ (accessed on 2 January 2023).
32. EUR-Lex. E.L. EUR-Lex-32019R2144-EN-EUR-Lex. 2014. Available online: https://eur-lex.europa.eu/legal-content/en/ALL/
?uri=CELEX%3A32019R2144 (accessed on 13 October 2023).
33. Waymo, W. Waymo Safety Report. 2020. Available online: https://waymo.com/safety/ (accessed on 18 August 2023).
34. Gehrig, S.K.; Stein, F.J. Dead reckoning and cartography using stereo vision for an autonomous car. In Proceedings of the 1999
IEEE/RSJ International Conference on Intelligent Robots and Systems. Human and Environment Friendly Robots with High
Intelligence and Emotional Quotients (Cat. No.99CH36289), Kyongju, Republic of Korea, 17–21 October 1999; Volume 3, pp.
1507–1512. [CrossRef]
35. Rawashdeh, N.A.; Bos, J.P.; Abu-Alrub, N.J. Drivable path detection using CNN sensor fusion for autonomous driving in the
snow. In Proceedings of the SPIE 11748, Autonomous Systems: Sensors, Processing, and Security for Vehicles and Infrastructure
2021, Online, FL, USA, 12–17 April 2021; Volume 1174806. [CrossRef]
36. Thrun, S.; Montemerlo, M.; Dahlkamp, H.; Stavens, D.; Aron, A.; Diebel, J.; Fong, P.; Gale, J.; Halpenny, M.; Hoffmann, G.; et al.
Stanley: The robot that won the DARPA Grand Challenge. J. Field Robot. 2006, 23, 661–692. [CrossRef]
37. Carballo, A.; Lambert, J.; Monrroy, A.; Wong, D.; Narksri, P.; Kitsukawa, Y.; Takeuchi, E.; Kato, S.; Takeda, K. LIBRE: The multiple
3D LiDAR dataset. In Intelligent Vehicles Symposium (IV), Las Vegas, NV, USA, 19 October–13 November 2020; pp. 1094–1101.
[CrossRef]
38. Velodyne, HDL-64E Spec Sheet. Available online: https://velodynesupport.zendesk.com/hc/en-us/articles/115003632634
-HDL-64E-Spec-Sheet (accessed on 13 October 2023).
39. Bijelic, M.; Gruber, T.; Ritter, W. A benchmark for lidar sensors in fog: Is detection breaking down? In Proceedings of the
Intelligent Vehicles Symposium (IV), Changshu, China, 26–30 June 2018; pp. 760–767. [CrossRef]
40. Patole, S.M.; Torlak, M.; Wang, D.; Ali, M. Automotive radars: A review of signal processing techniques. IEEE Signal Process.
Mag. 2017, 34, 22–35. [CrossRef]
41. Navtech Radar, FMCW Radar. Available online: https://navtechradar.com/explore/fmcw-radar/ (accessed on 22 November 2021).
42. Carullo, A.; Parvis, M. An ultrasonic sensor for distance measurement in automotive applications. IEEE Sens. J. 2001, 1, 143.
[CrossRef]
43. Frenzel, L. Ultrasonic Sensors: A Smart Choice for Shorter-Range Applications. 2018. Available online: https:
//www.electronicdesign.com/industrialautomation/article/21806202/ultrasonic-sensors-a-smart-choicefor-shorterrange-
applications (accessed on 14 August 2023).
44. Kamemura, T.; Takagi, H.; Pal, C.; Ohsumi, A. Development of a long-range ultrasonic sensor for automotive application. SAE
Int. J. Passeng. Cars-Electron. Electr. Syst. 2008, 1, 301–306. [CrossRef]
45. Tesla, Summon Your Tesla from Your Phone. Available online: https://www.tesla.com/blog/summon-your-tesla-your-phone
(accessed on 2 November 2023).
46. Pendleton, S.D.; Andersen, H.; Du, X.; Shen, X.; Meghjani, M.; Eng, Y.H.; Rus, D.; Ang, M.H., Jr. Perception, planning, control,
and coordination for autonomous vehicles. Machines 2017, 5, 6. [CrossRef]
47. Nasseri, A.; Shlomit, H. Autonomous Vehicle Technology Report. 2020. Available online: https://www.wevolver.com/article/
(accessed on 28 October 2023).
48. Rosique, F.; Navarro, P.J.; Fernández, C.; Padilla, A. A systematic review of perception system and simulators for autonomous
vehicles research. Sensors 2019, 19, 648. [CrossRef] [PubMed]
49. Paden, B.; Čáp, M.; Yong, S.Z.; Yershov, D.; Frazzoli, E. A survey of motion planning and control techniques for self-driving
urban vehicles. IEEE Trans. Intell. Veh. 2016, 1, 33–55. [CrossRef]
Algorithms 2024, 17, 103 32 of 36
50. Shladover, S.E.; Desoer, C.A.; Hedrick, J.K.; Tomizuka, M.; Walrand, J.; Zhang, W.B.; McMahon, D.H.; Peng, H.; Sheikholeslam,
S.; McKeown, N. Automated vehicle control developments in the PATH program. IEEE Trans. Veh. Technol. 1991, 40, 114–130.
[CrossRef]
51. Geiger, A.; Lenz, P.; Stiller, C.; Urtasun, R. Vision meets robotics: The kitti dataset. Int. J. Robot. Res. 2013, 32, 1231–1237.
[CrossRef]
52. Sun, H.; Zhang, W.; Yu, R.; Zhang, Y. Motion planning for mobile robots—Focusing on deep reinforcement learning: A systematic
review. IEEE Access 2021, 9, 69061–69081. [CrossRef]
53. Rai, M.; Khosla, B.; Dhawan, Y.; Kharotia, H.; Kumar, N.; Bandi, A. CLEAR: An Efficient Traffic Sign Recognition Technique for
Cyber-Physical Transportation Systems. In Proceedings of the 2022 IEEE/ACM 15th International Conference on Utility and
Cloud Computing (UCC), Vancouver, WA, USA, 6 December 2022; pp. 418–423.
54. Sakaridis, C.; Dai, D.; Van Gool, L. ACDC: The adverse conditions dataset with correspondences for semantic driving scene
understanding. In Proceedings of the IEEE/CVF International Conference on Computer Vision, Montreal, QC, Canada, 10–17
October 2021; pp. 10765–10775.
55. Geyer, J.; Kassahun, Y.; Mahmudi, M.; Ricou, X.; Durgesh, R.; Chung, A.S.; Hauswald, L.; Pham, V.H.; Mühlegg, M.; Dorn, S.; et al.
A2d2: Audi autonomous driving dataset. arXiv 2020, arXiv:2004.06320.
56. Sun, P.; Kretzschmar, H.; Dotiwalla, X.; Chouard, A.; Patnaik, V.; Tsui, P.; Guo, J.; Zhou, Y.; Chai, Y.; Caine, B.; et al. Scalability in
perception for autonomous driving: Waymo open dataset. In Proceedings of the 2020 IEEE/CVF Conference on Computer Vision
and Pattern Recognition (CVPR), Seattle, WA, USA, 13–19 June 2020; pp. 2443–2451. [CrossRef]
57. Caesar, H.; Bankiti, V.; Lang, A.H.; Vora, S.; Liong, V.E.; Xu, Q.; Krishnan, A.; Pan, Y.; Baldan, G.; Beijbom, O. nuscenes: A
multimodal dataset for autonomous driving. In Proceedings of the 2020 IEEE/CVF Conference on Computer Vision and Pattern
Recognition (CVPR), Seattle, WA, USA, 13–19 June 2020; pp. 11621–11631. [CrossRef]
58. Yu, F.; Chen, H.; Wang, X.; Xian, W.; Chen, Y.; Liu, F.; Madhavan, V.; Darrell, T. Bdd100k: A diverse driving dataset for
heterogeneous multitask learning. In Proceedings of the 2020 IEEE/CVF Conference on Computer Vision and Pattern Recognition
(CVPR), Seattle, WA, USA, 13–19 June 2020; pp. 2636–2645. [CrossRef]
59. Huang, X.; Wang, P.; Cheng, X.; Zhou, D.; Geng, Q.; Yang, R. The apolloscape open dataset for autonomous driving and its
application. IEEE Trans. Pattern Anal. Mach. Intell. 2019, 42, 2702–2719. [CrossRef] [PubMed]
60. Ros, G.; Sellart, L.; Materzynska, J.; Vazquez, D.; Lopez, A.M. The synthia dataset: A large collection of synthetic images for
semantic segmentation of urban scenes. In Proceedings of the 2016 IEEE Conference on Computer Vision and Pattern Recognition
(CVPR), Las Vegas, NV, USA, 26 June–1 July 2016; pp. 3234–3243. [CrossRef]
61. Carlevaris-Bianco, N.; Ushani, A.K.; Eustice, R.M. University of Michigan North Campus long-term vision and lidar dataset. Int.
J. Robot. Res. 2016, 35, 1023–1035. [CrossRef]
62. Pitropov, M.; Garcia, D.E.; Rebello, J.; Smart, M.; Wang, C.; Czarnecki, K.; Waslander, S. Canadian adverse driving conditions
dataset. Int. J. Robot. Res. 2021, 40, 681–690. [CrossRef]
63. Zheng, J.Y. IUPUI driving videos and images in all weather and illumination conditions. arXiv 2021, arXiv:2104.08657.
64. Bos, J.P.; Chopp, D.; Kurup, A.; Spike, N. Autonomy at the end of the earth: An inclement weather autonomous driving data set.
In Proceedings of the Autonomous Systems: Sensors, Processing, and Security for Vehicles and Infrastructure 2020, Online, CA,
USA, 27 April–9 May 2020; Volume 11415, pp. 36–48. [CrossRef]
65. Vora, S.; Lang, A.H.; Helou, B.; Beijbom, O. Pointpainting: Sequential fusion for 3d object detection. In Proceedings of the 2020
IEEE/CVF Conference on Computer Vision and Pattern Recognition (CVPR), Seattle, WA, USA, 13–19 June 2020; pp. 4603–4611.
[CrossRef]
66. Chang, M.F.; Lambert, J.; Sangkloy, P.; Singh, J.; Bak, S.; Hartnett, A.; Wang, D.; Carr, P.; Lucey, S.; Ramanan, D.; et al. Argoverse:
3d tracking and forecasting with rich maps. In Proceedings of the 2019 IEEE/CVF Conference on Computer Vision and Pattern
Recognition (CVPR), Long Beach, CA, USA, 15–20 June 2019; pp. 8748–8757. [CrossRef]
67. Meyer, M.; Kuschk, G. Automotive radar dataset for deep learning based 3d object detection. In Proceedings of the 2019 16th
European Radar Conference (EuRAD), Paris, France, 2–4 October 2019; pp. 129–132.
68. Bijelic, M.; Gruber, T.; Mannan, F.; Kraus, F.; Ritter, W.; Dietmayer, K.; Heide, F. Seeing through fog without seeing fog: Deep
multimodal sensor fusion in unseen adverse weather. In Proceedings of the 2020 IEEE/CVF Conference on Computer Vision and
Pattern Recognition (CVPR), Seattle, WA, USA, 15–20 June 2020; pp. 11682–11692.
69. Sakaridis, C.; Dai, D.; Van Gool, L. Semantic foggy scene understanding with synthetic data. Int. J. Comput. Vis. 2018, 126, 973–992.
[CrossRef]
70. Kim, J.; Rohrbach, A.; Darrell, T.; Canny, J.; Akata, Z. Textual Explanations for Self-Driving Vehicles. In Computer Vision–ECCV
2018; Ferrari, V., Hebert, M., Sminchisescu, C., Weiss, Y., Eds.; Lecture Notes in Computer Science; Springer: Cham, Switzerland,
2018; Volume 11206.
71. Neuhold, G.; Ollmann, T.; Rota Bulo, S.; Kontschieder, P. The mapillary vistas dataset for semantic understanding of street
scenes. In Proceedings of the 2017 IEEE International Conference on Computer Vision (ICCV), Venice, Italy, 22–29 October 2017;
pp. 4990–4999.
72. Braun, M.; Krebs, S.; Flohr, F.; Gavrila, D.M. Eurocity persons: A novel benchmark for person detection in traffic scenes. IEEE
Trans. Pattern Anal. Mach. Intell. 2019, 41, 1844–1861. [CrossRef]
Algorithms 2024, 17, 103 33 of 36
73. Maddern, W.; Pascoe, G.; Linegar, C.; Newman, P. 1 year, 1000 km: The oxford robotcar dataset. Int. J. Robot. Res. 2017, 36, 3–15.
[CrossRef]
74. Pham, Q.H.; Sevestre, P.; Pahwa, R.S.; Zhan, H.; Pang, C.H.; Chen, Y.; Mustafa, A.; Chandrasekhar, V.; Lin, J. A 3D dataset:
Towards autonomous driving in challenging environments. In Proceedings of the 2020 IEEE International Conference on Robotics
and Automation (ICRA), Paris, France, 31 May–31 August 2020; pp. 2267–2273.
75. Wenzel, P.; Wang, R.; Yang, N.; Cheng, Q.; Khan, Q.; von Stumberg, L.; Zeller, N.; Cremers, D. 4Seasons: A cross-season dataset
for multi-weather SLAM in autonomous driving. In Pattern Recognition; Akata, Z., Geiger, A., Sattler, T., Eds.; DAGM GCPR 2020.
Lecture Notes in Computer Science; Springer: Cham, Switzerland, 2021; Volume 12544, pp. 404–417.
76. Zendel, O.; Honauer, K.; Murschitz, M.; Steininger, D.; Dominguez, G.F. Wilddash-creating hazard-aware benchmarks. In Pro-
ceedings of the Computer Vision–ECCV 2018: 15th European Conference, Munich, Germany, 8–14 September 2018; Proceedings,
Part VI, pp. 402–416.
77. Choi, Y.; Kim, N.; Hwang, S.; Park, K.; Yoon, J.S.; An, K.; Kweon, I.S. KAIST multi-spectral day/night data set for autonomous
and assisted driving. IEEE Trans. Intell. Transp. Syst. 2018, 19, 934–948. [CrossRef]
78. Sheeny, M.; De Pellegrin, E.; Mukherjee, S.; Ahrabian, A.; Wang, S.; Wallace, A. RADIATE: A radar dataset for automotive
perception in bad weather. In Proceedings of the 2021 IEEE International Conference on Robotics and Automation (ICRA), Xi’an,
China, 30 May–5 June 2021; pp. 1–7.
79. Yan, Z.; Sun, L.; Krajník, T.; Ruichek, Y. EU long-term dataset with multiple sensors for autonomous driving. In Proceedings of
the 2020 IEEE/RSJ International Conference on Intelligent Robots and Systems (IROS), Las Vegas, NV, USA, 24 October 2020–24
January 2021; pp. 10697–10704.
80. Burnett, K.; Yoon, D.J.; Wu, Y.; Li, A.Z.; Zhang, H.; Lu, S.; Qian, J.; Tseng, W.K.; Lambert, A.; Leung, K.Y.; et al. Boreas: A
multi-season autonomous driving dataset. Int. J. Robot. Res. 2023, 42, 33–42. [CrossRef]
81. Kenk, M.A.; Hassaballah, M. DAWN: Vehicle detection in adverse weather nature dataset. arXiv 2020, arXiv:2008.05402.
82. Saralajew, S.; Ohnemus, L.; Ewecker, L.; Asan, E.; Isele, S.; Roos, S. A dataset for provident vehicle detection at night. In
Proceedings of the 2021 IEEE/RSJ International Conference on Intelligent Robots and Systems (IROS), Prague, Czech Republic,
27 September–1 October 2021; pp. 9750–9757.
83. Lowe, D.G. Object recognition from local scale-invariant features. In Proceedings of the Seventh IEEE International Conference
on Computer Vision, Kerkyra, Greece, 20–25 September 1999; Volume 2, pp. 1150–1157.
84. Viola, P.; Jones, M. Rapid object detection using a boosted cascade of simple features. In Proceedings of the 2001 IEEE Computer
Society Conference on Computer Vision and Pattern Recognition. CVPR 2001, Kauai, HI, USA, 8–14 December 2001. [CrossRef]
85. Dalal, N.; Triggs, B. Histograms of oriented gradients for human detection. In Proceedings of the 2005 IEEE Computer Society
Conference on Computer Vision and Pattern Recognition (CVPR’05), San Diego, CA, USA, 20–26 June 2005; Volume 1, pp. 886–893.
86. Asim, M.; Wang, Y.; Wang, K.; Huang, P.Q. A review on computational intelligence techniques in cloud and edge computing.
IEEE Trans. Emerg. Top. Comput. Intell. 2020, 4, 742–763. [CrossRef]
87. LeCun, Y.; Bottou, L.; Bengio, Y.; Haffner, P. Gradient-based learning applied to document recognition. Proc. IEEE 1998,
86, 2278–2324. [CrossRef]
88. Girshick, R.; Donahue, J.; Darrell, T.; Malik, J. Rich feature hierarchies for accurate object detection and semantic segmentation.
In Proceedings of the 2014 IEEE Conference on Computer Vision and Pattern Recognition, Columbus, OH, USA, 23–28 June 2014;
pp. 580–587.
89. Girshick, R. Fast r-cnn. In Proceedings of the 2015 IEEE International Conference on Computer Vision (ICCV), Santiago, Chile,
7–13 December 2015; pp. 1440–1448. [CrossRef]
90. Ren, S.; He, K.; Girshick, R.; Sun, J. Faster R-CNN: Towards real-time object detection with region proposal networks. IEEE Trans.
Pattern Anal. Mach. Intell. 2016, 39, 1137–1149. [CrossRef]
91. He, K.; Zhang, X.; Ren, S.; Sun, J. Spatial pyramid pooling in deep convolutional networks for visual recognition. IEEE Trans.
Pattern Anal. Mach. Intell. 2015, 37, 1904–1916. [CrossRef]
92. Lin, T.Y.; Dollár, P.; Girshick, R.; He, K.; Hariharan, B.; Belongie, S. Feature pyramid networks for object detection. In
Proceedings of the 2017 IEEE Conference on Computer Vision and Pattern Recognition (CVPR), Honolulu, HI, USA, 21–26 July
2017; pp. 936–944. [CrossRef]
93. Liu, W.; Anguelov, D.; Erhan, D.; Szegedy, C.; Reed, S.; Fu, C.Y.; Berg, A.C. Ssd: Single shot multibox detector. In Computer Vision–
ECCV 2016; Leibe, B., Matas, J., Sebe, N., Welling, M., Eds.; Lecture Notes in Computer Science; Springer: Cham, Switzerland,
2016; Volume 9905, pp. 21–37.
94. Terven, J.; Cordova-Esparza, D. A comprehensive review of YOLO: From YOLOv1 to YOLOv8 and beyond. arXiv 2023,
arXiv:2304.00501.
95. Tan, M.; Pang, R.; Le, Q.V. EfficientDet: Scalable and efficient object detection. In Proceedings of the 2020 IEEE/CVF Conference
on Computer Vision and Pattern Recognition (CVPR), Seattle, WA, USA, 13–19 June 2020; pp. 10778–10787. [CrossRef]
96. Carion, N.; Massa, F.; Synnaeve, G.; Usunier, N.; Kirillov, A.; Zagoruyko, S. End-to-end object detection with transformers. In
Computer Vision–ECCV 2020; Vedaldi, A., Bischof, H., Brox, T., Frahm, J.M., Eds.; Lecture Notes in Computer Science; Springer:
Cham, Switzerland, 2020; Volume 12346, pp. 213–229.
Algorithms 2024, 17, 103 34 of 36
97. Tian, Z.; Shen, C.; Chen, H.; He, T. Fcos: Fully convolutional one-stage object detection. In Proceedings of the 2019 IEEE/CVF
International Conference on Computer Vision (ICCV), Seoul, Republic of Korea, 27 October–2 November 2019; pp. 9626–9635.
[CrossRef]
98. Lin, T.Y.; Goyal, P.; Girshick, R.; He, K.; Dollár, P. Focal loss for dense object detection. In Proceedings of the 2017 IEEE
International Conference on Computer Vision (ICCV), Venice, Italy, 22–29 October 2017; pp. 2999–3007. [CrossRef]
99. Galvao, L.G.; Abbod, M.; Kalganova, T.; Palade, V.; Huda, M.N. Pedestrian and vehicle detection in autonomous vehicle
perception systems—A review. Sensors 2021, 21, 7267. [CrossRef]
100. Jocher, G. Ultralytics/Yolov8, GitHub, GitHub Repository. 2023. Available online: https://github.com/ultralytics/ultralytics
(accessed on 15 December 2023).
101. Jia, S.J.; Zhai, Y.T.; Jia, X.W. Detection of Traffic and Road Condition Based on SVM and PHOW. Appl. Mech. Mater. 2014,
513, 3651–3654. [CrossRef]
102. Gao, D.; Zhou, J.; Xin, L. SVM-based detection of moving vehicles for automatic traffic monitoring. In Proceedings of the ITSC
2001. 2001 IEEE Intelligent Transportation Systems. Proceedings (Cat. No.01TH8585), Oakland, CA, USA, 25–29 August 2001;
pp. 745–749. [CrossRef]
103. Zheng, F.; Luo, S.; Song, K.; Yan, C.W.; Wang, M.C. Improved lane line detection algorithm based on Hough transform. Pattern
Recognit. Image Anal. 2018, 28, 254–260. [CrossRef]
104. Vamsi, A.M.; Deepalakshmi, P.; Nagaraj, P.; Awasthi, A.; Raj, A. IOT based autonomous inventory management for warehouses.
In EAI International Conference on Big Data Innovation for Sustainable Cognitive Computing; Haldorai, A., Ramu, A., Mohanram, S.,
Onn, C., Eds.; EAI/Springer Innovations in Communication and Computing; Springer: Cham, Switzerland, 2020; pp. 371–376.
105. Nagaraj, P.; Muneeswaran, V.; Sudar, K.M.; Ali, R.S.; Someshwara, A.; Kumar, T.S. Internet of Things based smart hospital saline
monitoring system. In Proceedings of the 2021 5th International Conference on Computer, Communication and Signal Processing
(ICCCSP), Chennai, India, 24–25 May 2021; pp. 53–58. [CrossRef]
106. Wang, Z.; Zhan, J.; Duan, C.; Guan, X.; Yang, K. Vehicle detection in severe weather based on pseudo-visual search and HOG–LBP
feature fusion. Proc. Inst. Mech. Eng. Part D J. Automob. Eng. 2022, 236, 1607–1618. [CrossRef]
107. Abdullah, M.N.; Ali, Y.H. Vehicles detection system at different weather conditions. Iraqi J. Sci. 2021, 62, 2040–2052. [CrossRef]
108. Wu, B.F.; Juang, J.H. Adaptive vehicle detector approach for complex environments. IEEE Trans. Intell. Transp. Syst. 2012,
13, 817–827. [CrossRef]
109. Padilla, D.A.; Villaverde, J.F.; Magdaraog, J.J.T.; Oconer, A.J.L.; Ranjo, J.P. Vehicle and Weather Detection Using Real Time Image
Processing Using Optical Flow and Color Histogram. In Proceedings of the 2019 5th International Conference on Control,
Automation and Robotics (ICCAR), Beijing, China, 19–22 April 2019; pp. 880–883. [CrossRef]
110. Tian, Q.; Zhang, L.; Wei, Y.; Zhao, W.; Fei, W. Vehicle Detection and Tracking at Night in Video Surveillance. Int. J. Online Eng.
2013, 9, 60–64. [CrossRef]
111. Kuang, H.; Chen, L.; Chan, L.L.H.; Cheung, R.C.; Yan, H. Feature selection based on tensor decomposition and object proposal
for night-time multiclass vehicle detection. IEEE Trans. Syst. Man Cybern. Syst. 2018, 49, 71–80. [CrossRef]
112. Ewecker, L.; Asan, E.; Ohnemus, L.; Saralajew, S. Provident vehicle detection at night for advanced driver assistance systems.
Auton. Robot. 2023, 47, 313–335. [CrossRef]
113. Yaghoobi Ershadi, N. Improving vehicle tracking rate and speed estimation in dusty and snowy weather conditions with a
vibrating camera. PLoS ONE 2017, 12, e0189145. [CrossRef]
114. Yang, H.; Qu, S. Real-time vehicle detection and counting in complex traffic scenes using background subtraction model with
low-rank decomposition. IET Intell. Transp. Syst. 2018, 12, 75–85. [CrossRef]
115. Govardhan, P. Night Time Pedestrian Detection for Advanced Driving Assistance Systems (ADAS) Using Near Infrared Images.
Ph.D. Thesis, National Institute of Technology, Rourkela, India, 2014.
116. Sasaki, Y.; Emaru, T.; Ravankar, A.A. SVM based pedestrian detection system for sidewalk snow removing machines. In
Proceedings of the 2021 IEEE/SICE International Symposium on System Integration (SII), Iwaki, Japan, 11–14 January 2021;
pp. 700–701.
117. Kang, S.; Byun, H.; Lee, S.W. Real-time pedestrian detection using support vector machines. In Pattern Recognition with Support
Vector Machines. SVM 2002; Lee, S.W., Verri, A., Eds.; Lecture Notes in Computer Science; Springer: Berlin/Heidelberg, Germany,
2002; Volume 2388, pp. 268–277.
118. Jegham, I.; Khalifa, A.B. Pedestrian detection in poor weather conditions using moving camera. In Proceedings of the 2017
IEEE/ACS 14th International Conference on Computer Systems and Applications (AICCSA), Hammamet, Tunisia, 30 October–3
November 2017; pp. 358–362.
119. Fan, Y. Research of Pedestrian Tracking Based on HOG Feature and Haar Feature. Comput. Sci. 2013, 40, 199–203.
120. Ding, B.; Liu, Z.; Sun, Y. Pedestrian detection in haze environments using dark channel prior and histogram of oriented gradient.
In Proceedings of the 2018 Eighth International Conference on Instrumentation & Measurement, Computer, Communication and
Control (IMCCC), Harbin, China, 19–21 July 2018; pp. 1003–1008.
121. Sotelo, M.; Parra, I.; Fernandez, D.; Naranjo, E. Pedestrian detection using SVM and multi-feature combination. In Proceedings
of the 2006 IEEE Intelligent Transportation Systems Conference, Toronto, ON, Canada, 17–20 September 2006; pp. 103–108.
[CrossRef]
Algorithms 2024, 17, 103 35 of 36
122. Satyawan, A.S.; Fuady, S.; Mitayani, A.; Sari, Y.W. HOG Based Pedestrian Detection System for Autonomous Vehicle Operated
in Limited Area. In Proceedings of the 2021 International Conference on Radar, Antenna, Microwave, Electronics, and
Telecommunications (ICRAMET), Bandung, Indonesia, 23–24 November 2021; pp. 147–152.
123. Jang, G.; Park, J.; Kim, M. Cascade-Adaboost for Pedestrian Detection Using HOG and Combined Features. In Advances in
Computer Science and Ubiquitous Computing; Springer: Singapore, 2016; pp. 430–435.
124. Zhang, X.; Zhu, X. Autonomous path tracking control of intelligent electric vehicles based on lane detection and optimal preview
method. Expert Syst. Appl. 2019, 121, 38–48. [CrossRef]
125. Gern, A.; Moebus, R.; Franke, U. Vision-based lane recognition under adverse weather conditions using optical flow. In
Proceedings of the Intelligent Vehicle Symposium 2002, Versailles, France, 17–21 June 2002; Volume 2, pp. 652–657.
126. Tran, T.T.; Son, J.H.; Uk, B.J.; Lee, J.H.; Cho, H.M. An adaptive method for detecting lane boundary in night scene. In Advanced
Intelligent Computing Theories and Applications. With Aspects of Artificial Intelligence; Huang, D.S., Zhang, X., Reyes García, C.A.,
Zhang, L., Eds.; ICIC 2010. Lecture Notes in Computer Science; Springer: Berlin/Heidelberg, Germany, 2010; Volume 6216,
pp. 301–308.
127. Chen, T.Y.; Chen, C.H.; Luo, G.M.; Hu, W.C.; Chern, J.C. Vehicle detection in nighttime environment by locating road lane
and taillights. In Proceedings of the 2015 International Conference on Intelligent Information Hiding and Multimedia Signal
Processing (IIH-MSP), Adelaide, SA, Australia, 23–25 September 2015; pp. 60–63.
128. Guo, J.; Wei, Z.; Miao, D. Lane detection method based on improved RANSAC algorithm. In Proceedings of the 2015 IEEE
Twelfth International Symposium on Autonomous Decentralized Systems, Taichung, Taiwan, 25–27 March 2015; pp. 285–288.
129. Antonio, J.A.; Romero, M. Pedestrians’ Detection Methods in Video Images: A Literature Review. In Proceedings of the 2018
International Conference on Computational Science and Computational Intelligence (CSCI), Las Vegas, NV, USA, 12–14 December
2018; pp. 354–360.
130. Papadimitriou, O.; Kanavos, A.; Mylonas, P.; Maragoudakis, M. Advancing Weather Image Classification using Deep Convolu-
tional Neural Networks. In Proceedings of the 2023 18th International Workshop on Semantic and Social Media Adaptation &
Personalization (SMAP) 18th International Workshop on Semantic and Social Media Adaptation & Personalization (SMAP 2023),
Limassol, Cyprus, 25–26 September 2023; pp. 1–6.
131. Alhammadi, S.A.; Alhameli, S.A.; Almaazmi, F.A.; Almazrouei, B.H.; Almessabi, H.A.; Abu-Kheil, Y. Thermal-Based Vehicle
Detection System using Deep Transfer Learning under Extreme Weather Conditions. In Proceedings of the 2022 8th International
Conference on Information Technology Trends (ITT), Dubai, United Arab Emirates, 25–26 May 2022; pp. 119–123.
132. Arora, N.; Kumar, Y.; Karkra, R.; Kumar, M. Automatic vehicle detection system in different environment conditions using fast
R-CNN. Multimed. Tools Appl. 2022, 81, 18715–18735. [CrossRef]
133. Ghosh, R. On-road vehicle detection in varying weather conditions using faster R-CNN with several region proposal networks.
Multimed. Tools Appl. 2021, 80, 25985–25999. [CrossRef]
134. Cao, J.; Song, C.; Song, S.; Peng, S.; Wang, D.; Shao, Y.; Xiao, F. Front vehicle detection algorithm for smart car based on improved
SSD model. Sensors 2020, 20, 4646. [CrossRef] [PubMed]
135. Li, W. Vehicle detection in foggy weather based on an enhanced YOLO method. In Journal of Physics: Conference Series; IOP
Publishing: Bristol, UK, 2022; Volume 2284, p. 012015.
136. Ghosh, R. A modified yolo model for on-road vehicle detection in varying weather conditions. In Intelligent Computing and
Communication Systems; Springer: Singapore, 2021; pp. 45–54.
137. Miao, Y.; Liu, F.; Hou, T.; Liu, L.; Liu, Y. A nighttime vehicle detection method based on YOLO v3. In Proceedings of the 2020
Chinese Automation Congress (CAC), Shanghai, China, 6–8 November 2020; pp. 6617–6621.
138. Humayun, M.; Ashfaq, F.; Jhanjhi, N.Z.; Alsadun, M.K. Traffic management: Multi-scale vehicle detection in varying weather
conditions using yolov4 and spatial pyramid pooling network. Electronics 2022, 11, 2748. [CrossRef]
139. Wang, R.; Zhao, H.; Xu, Z.; Ding, Y.; Li, G.; Zhang, Y.; Li, H. Real-time vehicle target detection in inclement weather conditions
based on YOLOv4. Front. Neurorobot. 2023, 17, 1058723. [CrossRef]
140. Yao, J.; Fan, X.; Li, B.; Qin, W. Adverse Weather Target Detection Algorithm Based on Adaptive Color Levels and Improved
YOLOv5. Sensors 2022, 22, 8577. [CrossRef] [PubMed]
141. Razzok, M.; Badri, A.; Mourabit, I.E.; Ruichek, Y.; Sahel, A. Pedestrian detection under weather conditions using conditional
generative adversarial network. Int. J. Artif. Intell. 2023, 12, 1557–1568. [CrossRef]
142. Pawełczyk, M.; Wojtyra, M. Real world object detection dataset for quadcopter unmanned aerial vehicle detection. IEEE Access
2020, 8, 174394–174409. [CrossRef]
143. Yahiaoui, M.; Rashed, H.; Mariotti, L.; Sistu, G.; Clancy, I.; Yahiaoui, L.; Kumar, V.R.; Yogamani, S. Fisheyemodnet: Moving object
detection on surround-view cameras for autonomous driving. arXiv 2019, arXiv:1908.11789.
144. Ragesh, N.; Rajesh, R. Pedestrian detection in automotive safety: Understanding state-of-the-art. IEEE Access 2019, 7, 47864–47890.
[CrossRef]
145. Kim, J.H.; Hong, H.G.; Park, K.R. Convolutional neural network-based human detection in nighttime images using visible light
camera sensors. Sensors 2017, 17, 1065. [CrossRef] [PubMed]
146. Hsia, C.H.; Peng, H.C.; Chan, H.T. All-Weather Pedestrian Detection Based on Double-Stream Multispectral Network. Electronics
2023, 12, 2312. [CrossRef]
Algorithms 2024, 17, 103 36 of 36
147. Lai, K.; Zhao, J.; Liu, D.; Huang, X.; Wang, L. Research on pedestrian detection using optimized mask R-CNN algorithm in
low-light road environment. In Journal of Physics: Conference Series; IOP Publishing: Bristol, UK, 2021; Volume 1777, p. 012057.
148. Tumas, P.; Nowosielski, A.; Serackis, A. Pedestrian detection in severe weather conditions. IEEE Access 2020, 8, 62775–62784.
[CrossRef]
149. Shakeri, A.; Moshiri, B.; Garakani, H.G. Pedestrian detection using image fusion and stereo vision in autonomous vehicles. In
Proceedings of the 2018 9th International Symposium on Telecommunications (IST), Tehran, Iran, 17–19 December 2018; pp.
592–596. [CrossRef]
150. Galarza-Bravo, M.A.; Flores-Calero, M.J. Pedestrian detection at night based on faster R-CNN and far infrared images. In
Intelligent Robotics and Applications; Chen, Z., Mendes, A., Yan, Y., Chen, S., Eds.; ICIRA 2018. Lecture Notes in Computer Science;
Springer: Cham, Switzerland, 2018; Volume 10985, pp. 335–345.
151. Xu, Y.; Cao, X.; Qiao, H. Optical camera based pedestrian detection in rainy or snowy weather. In Fuzzy Systems and
Knowledge Discovery; Wang, L., Jiao, L., Shi, G., Li, X., Liu, J., Eds.; FSKD 2006. Lecture Notes in Computer Science; Springer:
Berlin/Heidelberg, Germany, 2006; Volume 4223, pp. 1182–1191.
152. Montenegro, B.; Flores-Calero, M. Pedestrian detection at daytime and nighttime conditions based on YOLO-v5. Ingenius. Rev.
Cienc. Tecnol. 2022, 85–95. [CrossRef]
153. Zaman, M.; Saha, S.; Zohrabi, N.; Abdelwahed, S. Deep Learning Approaches for Vehicle and Pedestrian Detection in Adverse
Weather. In Proceedings of the 2023 IEEE Transportation Electrification Conference & Expo (ITEC), Detroit, MI, USA, 21–23 June
2023; pp. 1–6.
154. Son, J.; Yoo, H.; Kim, S.; Sohn, K. Real-time illumination invariant lane detection for lane departure warning system. Expert Syst.
Appl. 2015, 42, 1816–1824. [CrossRef]
155. Kortli, Y.; Marzougui, M.; Atri, M. Efficient implementation of a real-time lane departure warning system. In Proceedings of the
2016 International Image Processing, Applications and Systems (IPAS), Hammamet, Tunisia, 5–7 November 2016; pp. 1–6.
156. Sun, T.Y.; Tsai, S.J.; Chan, V. HSI color model based lane-marking detection. In Proceedings of the 2006 IEEE Intelligent
Transportation Systems Conference, Toronto, ON, Canada, 7–12 September 2006; pp. 1168–1172.
157. Chiu, K.Y.; Lin, S.F. Lane detection using color-based segmentation. In Proceedings of the IEEE Proceedings. Intelligent Vehicles
Symposium, Las Vegas, NV, USA, 6–8 June 2005; pp. 706–711.
158. Shin, S.; Shim, I.; Kweon, I.S. Combinatorial approach for lane detection using image and LIDAR reflectance. In Proceedings of
the 2015 12th International Conference on Ubiquitous Robots and Ambient Intelligence (URAI), Goyangi, Republic of Korea,
28–30 October 2015; pp. 485–487.
159. Rose, C.; Britt, J.; Allen, J.; Bevly, D. An integrated vehicle navigation system utilizing lane-detection and lateral position
estimation systems in difficult environments for GPS. IEEE Trans. Intell. Transp. Syst. 2014, 15, 2615–2629. [CrossRef]
160. Hsieh, J.W.; Chen, L.C.; Chen, D.Y. Symmetrical SURF and its applications to vehicle detection and vehicle make and model
recognition. IEEE Trans. Intell. Transp. Syst. 2014, 15, 6–20. [CrossRef]
161. Yusuf, M.M.; Karim, T.; Saif, A.S. A robust method for lane detection under adverse weather and illumination conditions using
convolutional neural network. In Proceedings of the ICCA 2020: International Conference on Computing Advancements, Dhaka,
Bangladesh, 10–12 January 2020; pp. 1–8.
162. Nguyen, V.; Kim, H.; Jun, S.; Boo, K. A study on real-time detection method of lane and vehicle for lane change assistant system
using vision system on highway. Eng. Sci. Technol. Int. J. 2018, 21, 822–833. [CrossRef]
163. Khan, M.N.; Ahmed, M.M. Weather and surface condition detection based on road-side webcams: Application of pre-trained
convolutional neural network. Int. J. Transp. Sci. Technol. 2022, 11, 468–483. [CrossRef]
164. Ding, Y.; Xu, Z.; Zhang, Y.; Sun, K. Fast lane detection based on bird’s eye view and improved random sample consensus
algorithm. Multimed. Tools Appl. 2017, 76, 22979–22998. [CrossRef]
165. Ab Ghani, H.; Daud, A.M.; Besar, R.; Sani, Z.M.; Kamaruddin, M.N.; Syahali, S. Lane Detection Using Deep Learning for Rainy
Conditions. In Proceedings of the 2023 9th International Conference on Computer and Communication Engineering (ICCCE),
Kuala Lumpur, Malaysia, 15–16 August 2023; pp. 373–376.
166. Raj, N.; Dutta, K.K.; Kamath, A. Lane Prediction by Autonomous Vehicle in Highway Traffic using Artificial Neural Networks. In
Proceedings of the 2021 Fourth International Conference on Electrical, Computer and Communication Technologies (ICECCT),
Erode, India, 15–17 September 2021; pp. 1–6.
167. Krishnaveni, P.; Sutha, J. Novel deep learning framework for broadcasting abnormal events obtained from surveillance
applications. J. Ambient. Intell. Humaniz. Comput. 2020, 13, 4123. [CrossRef]
168. Ouyang, W.; Wang, X. A discriminative deep model for pedestrian detection with occlusion handling. In Proceedings of the 2012
IEEE Conference on Computer Vision and Pattern Recognition, Providence, RI, USA, 16–21 June 2012; pp. 3258–3265.
169. Haris, M.; Hou, J.; Wang, X. Lane lines detection under complex environment by fusion of detection and prediction models.
Transp. Res. Rec. 2022, 2676, 342–359. [CrossRef]
Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual
author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to
people or property resulting from any ideas, methods, instructions or products referred to in the content.