You are on page 1of 16

Renewable and Sustainable Energy Reviews 188 (2023) 113873

Contents lists available at ScienceDirect

Renewable and Sustainable Energy Reviews


journal homepage: www.elsevier.com/locate/rser

Review article

Electric vehicle charging service operations: A review of machine learning


applications for infrastructure planning, control, pricing and routing
Nilgun Fescioglu-Unver ∗, Melike Yıldız Aktaş
Department of Industrial Engineering, TOBB University of Economics and Technology, Ankara, Turkey

ARTICLE INFO ABSTRACT

Keywords: The majority of global road transportation emissions come from passenger and freight vehicles. Electric vehicles
Electric vehicle (EV) provide a sustainable transportation way, but customers’ charging service related concerns affect the
Charging service operations EV adoption rate. Effective infrastructure planning, charge scheduling, charge pricing, and electric vehicle
Machine learning
routing strategies can help relieve customer perceived risks. The number of studies using machine learning
Infrastructure planning
algorithms to solve these problems is increasing daily. Forecasting, clustering, and reinforcement based models
Charge scheduling
Pricing
are frequently the main solution methods or provide valuable inputs to other solution procedures. This study
Routing reviews the studies that apply machine learning models to improve EV charging service operations and provides
future research directions.

1. Introduction fast charging stations, while users are accepting to make detours, they
strongly reject waiting times at the station [11].
Road transportation carbon emissions accounted for 16% of global When charging is completed at home or the office, the most critical
emissions in 2022 [1]. Passenger and freight vehicles are responsible concerns are the charging price related to the time of charge and
for the majority of this emission. Electric vehicles (EV) provide a getting the required state of charge (SoC) before the departure time.
sustainable transportation way, and ensuring a widespread deployment In this case, charge control and scheduling models aid EV users charge
requires overcoming the adoption barriers. their vehicles at the right time and price. When consumers use public
Various factors affect EV adoption, including economic factors such charging infrastructure, the most critical concerns become locating a
as EV purchase/operation/maintenance and charging cost, charging nearby available station, time spent while driving to and waiting at
infrastructure readiness (i.e., station availability, range anxiety), con- the station, and the charging price. Adequate infrastructure planning,
sumer psychological aspects, government policies, vehicle performance, dynamic pricing, and routing models improve public charging service
and environmental perspectives [2]. Especially charging related con- operations.
cerns are among the highest perceived risks of EV adoption [3]. Charg- There are several review studies on electric vehicle charging stations
ing takes place at homes, offices, publicly accessible places like malls,
and charging services. Review topics include charging infrastructure
and charging stations on roads. Surveys show that home charging and
control and impact on grid [12,13], smart charging [14], station lo-
office charging are the most used places for charging, and around a
cation and sizing optimization [15,16], battery swapping stations [17],
total of 95% of charging events occur at these locations. Consumers are
pricing [18], electric vehicle routing [19] and the optimization studies
mostly concerned about charging location and charging costs. Public
on electric vehicle service operations [20].
charging is the least used infrastructure type; however, it is essential
Machine learning models are frequently applied to electric vehicle
for long journeys and is seen as a safety net. [4].
charging related problems. For the smart grid, machine learning helps
Charging duration and charging station density influence
grid power quality and condition monitoring, load forecasting and bal-
consumers’ public charging infrastructure evaluation significantly [5].
The time spent searching for a public charging station and waiting ancing [21,22], renewable energy, energy storage system integration,
at the station is one of the highest perceived risks of electric vehicle and security [23]. Hybrid electric vehicle (HEV) energy management
adoption [3]. Fast charging can effectively reduce range anxiety [6]. strategies also use machine learning to control HEV’s different en-
Fast charging infrastructure expansion is required for the widespread ergy sources [24]. Data repositories provide machine learning models
adoption of electric vehicles [7–9] and is a more cost efficient solution with real-world data about EV market growth, infrastructure, user
when compared with increasing battery capacities [10]. In the case of behavior [25], and electric vehicle load [26].

∗ Corresponding author.
E-mail addresses: nfunver@etu.edu.tr (N. Fescioglu-Unver), melikeyildiz95@gmail.com (M. Yıldız Aktaş).

https://doi.org/10.1016/j.rser.2023.113873
Received 10 March 2023; Received in revised form 2 October 2023; Accepted 9 October 2023
Available online 17 October 2023
1364-0321/© 2023 Elsevier Ltd. All rights reserved.
N. Fescioglu-Unver and M. Yıldız Aktaş Renewable and Sustainable Energy Reviews 188 (2023) 113873

optimization problems. Several review studies investigate the related


Abbreviations studies from different perspectives. Al-Ogaili et al. [27] review the
AC Alternate Current EV charge control strategies and focus on the studies using schedul-
ANN Artificial Neural Networks ing, clustering, and forecasting methods. They cover mathematical
programming and metaheuristic charge scheduling approaches, clus-
BEV Battery Electric Vehicles
tering methods for load profiling, and forecast studies for short/long
BPNN Back Propagation Neural Network
term load forecasting, published before 2020. Abdullah et al.’s review
CART Classification and Regression Trees study [28] concentrates on reinforcement learning (RL) based EV charg-
CFSFDP Clustering by Fast Search-and-Find of Den- ing coordination literature. They list the reinforcement learning specific
sity Peaks parameters such as the state, reward (objective), and constraints, and
CHAID Chi-square Automatic Interaction Detector the specific RL algorithm type each study uses. Deb [29] review the
CS Charging Station studies that use machine learning approaches for charging demand
DBSCAN Density-Based Spatial Clustering of Appli- prediction, scheduling, charging station placement, and charger uti-
cations with Noise lization prediction. In another review, Shahriar et al. [30] focus on
DC Direct Current predicting charging behavior such as EV arrival/departure times, charg-
DDPG Deep Deterministic Policy Gradient ing duration, and energy consumption and review the studies using
DDQN Double Deep Q-Network supervised/unsupervised machine learning methods. To the best of
our knowledge, there are no review studies investigating the machine
DEVRP Dynamic Electric Vehicle Routing Problem
learning applications in dynamic pricing and EV routing.
DNN Deep Neural Network
EV charging service related topics are interrelated with each other.
DQL Deep Q-Learning The infrastructure plan provides inputs to charge control/scheduling
DQN Deep Q-Network models, charge coordination requires dynamic pricing information and
DRL Deep Reinforcement Learning the outputs of these layers are inputs to the routing algorithms. Gov-
ER-EV Extended Range EVs ernment policies on electric vehicle support and renewable energy
EV Electric Vehicle generation affect the energy demand and supply. Policymakers and
EVCS Electric Vehicle Charging Station Selection engineers can use machine learning algorithms to obtain valuable in-
Problem formation in all EV charging service operations. This study differs from
EVEC Electric Vehicle Energy Consumption Prob- the existing ones with its focus and coverage. The key contributions of
lem this study are:
EVRP Electric Vehicle Routing Problem • Focusing on charging service related topics that mostly influence
EVSC EV Smart Charging the EV user’s EV adoption decisions.
FCEV Fuel Cell Electric Vehicles • Covering a broad range of EV charging service related areas,
FCLM Flow Capturing Location Model including; Infrastructure planning, charge control and scheduling,
GB Gradient Boosting pricing, and routing.
GRU Gated Recurrent Unit • Covering studies that use machine learning algorithms both as
HEV Hybrid Electric Vehicle the primary solution method and as an aid providing input to the
main method.
kNN k-Nearest Neighbor
• Including the most recent literature in its coverage area.
LSTM Long Short Term Memory
• Highlighting the findings and future research directions for each
MCLM Maximal Covering Location Model
charging service topic.
NN Neural Networks
PCA Principal Component Analysis This paper is organized as follows. Section 2 introduces the electric
PHEV Plug-in Hybrid Electric Vehicles vehicle types and charging modes. Section 3 starts with infrastructure
planning and covers the studies that use machine learning for station
PPO Proximal Policy Optimization
placement and demand/charging station (CS) utilization forecasting.
QoS Quality of Service
Next, Section 4 reviews the charge control and schedule literature from
RF Random Forest
the aggregator and user perspectives. Sections 5 and 6 review the ma-
RL Reinforcement Learning chine learning based dynamic pricing and routing studies, respectively.
SAC Soft Actor Critic Finally Section 7 summarizes the findings and Section 8 concludes the
SARSA State Action Reward State Action study.
SoC State of Charge
SVR Support Vector Regressions 2. Electric vehicle charging
TOU Time of Use
V2B Vehicle to Building Electric vehicles can be classified as battery electric vehicles (BEV),
plug-in hybrid electric vehicles (PHEV), hybrid electric vehicles (HEV),
V2G Vehicle to Grid
fuel cell electric vehicles (FCEVs), and extended range EVs (ER-EV),
V2H Vehicle to Home
which are essentially BEVs equipped with a combustion engine to
V2V Vehicle to Vehicle charge the battery when needed. PHEVs and HEVs include both an
V2X Vehicle to Everything electric engine and an internal combustion engine. PHEVs’ electric
VRP Vehicle Routing Problem engine can be charged by an external electric source (by plugging in the
main grid) whereas the HEVs are charged by the internal combustion
engine’s energy [31,32].
Electric vehicle batteries can be charged by conductive and wireless
Looking from the EV users’ perspective, charging service location, charging methods. Conductive charging is the most widely used method
duration, and cost are important factors in EV adoption decisions. Ma- and is classified into three types according to the power transfer levels
chine learning algorithms also aid in service operations planning and — Level 1, Level 2, and Level 3. Level 1 charging is often called slow

2
N. Fescioglu-Unver and M. Yıldız Aktaş Renewable and Sustainable Energy Reviews 188 (2023) 113873

charging and is primarily used at homes, where the vehicles connect sizing of charging stations. Deb et al. [40] classify the charging sta-
directly to the electrical outlet. Since the power transfer rate is low tion placement studies into three groups according to their modeling
(AC single-phase 3.3 kW), the battery charging time required for the considerations: Models considering the distribution network, trans-
vehicle’s 100 km journey reaches an average of 6-8 h. Level 2 charging portation network, or both. In the distribution network approach,
enables faster charging through a charging unit. Level-2 charging is the main concern is the power system’s benefits and costs. In the
primarily available at public long term parking locations like shopping transportation network approach, the main concern is optimal CS dis-
centers, hotels, airports, and workplace parking lots, and EV owners tribution in the road network. The transportation network approach
can also install these chargers in their homes. Depending on the power is divided into node-based, tour-based, and path-based approaches.
level (AC single-phase 7.4 kW, AC three-phase 22 kW), the charging Node-based approaches model the CS placement problem as a facility
time for 100 km varies from 4 h to 1 h [33]. location problem where the facilities are located at nodes to satisfy
Level 3 charging (also called fast or DC charging) decreases the the nodes’ demand. Node-based solution methods include P-median
charging time for a 100 km road to 20–30 min with a 50 kW power based approaches and set covering models. Path-based approaches
DC charging unit. With DC Rapid (150 kW) and DC Ultra-Fast (350 kW) concentrate on the vehicle flow along the paths and require traffic
charging modes, the charging time will decrease to 10 min at dedicated flow data. Tour-based approaches include the round-trip possibilities
charging stations. Level 3 fast charging stations generally meet the to the problem. Path-based solutions frequently use flow capturing
charge needs on long highway corridors. Fast charging is the model location model (FCLM) and maximal covering location model (MCLM)
that most closely resembles the refueling process and suits the current variations. Metais et al. [41] further classify the infrastructure planning
behavioral pattern of vehicle owners. Therefore, locating fast charging studies with respect to their coverage as location, sizing, or both. Sizing
stations in urban areas to meet the daily charging needs of EV users studies can optimize the number of vehicles that can be accommodated
can improve the EV transition rate [33]. or the station’s power level.
Wireless charging allows the vehicle to be charged without a direct Accurate charging demand and station availability/utilization fore-
connection between the vehicle and the grid. During wireless charging, casting improve the infrastructure planning process outcomes. Machine
electricity transfers from the transmitter pads placed on the ground to learning models help the infrastructure planning problem by improving
the receiver pads placed under the vehicle. Wireless charging technolo- the charging station placement and sizing heuristics and the prediction
gies can be classified according to electricity transfer methods or the of long/short term charging demand and charger utilization. Fig. 2
transmitted distance (near field–far field). Near-field charging methods summarizes the infrastructure planning sub-problems machine learning
include inductive, magnetic resonance, and capacitative charging. Far algorithms are frequently applied.
field methods include microwave, radio wave, and laser technologies, This section investigates the infrastructure related studies in two
and are still in the prototype stage [34]. In addition, wireless charging sub-sections. Section 3.1 includes studies using machine learning as
can be static or dynamic. In the static case, the vehicle is stationary, part of the station placement/sizing problem solution. Section 3.2
whereas in the dynamic case, the vehicle is charged en-route while investigates long/short term charging demand and station utilization
in motion [35]. Vehicle to grid (V2G) technology is also applicable to forecasting studies.
wireless charging [36].
Other electric vehicle charging related technologies include vehicle 3.1. Charging station placement and sizing optimization
to everything (V2X), battery swapping, and mobile charging. Vehicle
to everything (V2X) concept, which connects vehicles to the grid, to Charging station placement and sizing studies determine optimal
homes, and other vehicles, enables using the energy in EV batteries locations and capacities for charging stations. Several studies use clus-
outside the vehicles [37,38]. The control of charging can be coordi- tering algorithms to determine the candidate locations and select the
nated by unidirectional charging, such as smart charging, including station locations/sizes according to an EV user specific, or CS/grid op-
V1G, or bidirectional power flows like vehicle-to-grid (V2G), vehicle- erator specific objective, or a combination. Examples of EV user specific
to-home (V2H), vehicle-to-building (V2B) or vehicle-to-vehicle (V2V) objectives include minimizing the distance to the station, minimizing
as in Fig. 1. In a V2G setting, the grid operators benefit from frequency queuing time, or maximizing the demand coverage. CS or grid operator
regulation, voltage support, and spinning reserves in case of unplanned specific objectives include maximizing profit, minimizing power loss,
events. The EV owner benefits from the energy arbitrage by buying voltage fluctuation, load fluctuation, and minimizing cost. Cost items
energy at low priced times and selling it back to the grid at high priced include operation, construction, and maintenance costs. Some studies
peak usage times. In a V2H setting, the EV acts as a backup power combine EV user and CS operator benefits in a single objective or use
storage during blackouts and aids household load management during multiple objectives. Infrastructure planning studies frequently use K-
day time. V2B is a setting similar to V2H where the vehicle acts as a means and hierarchical clustering algorithms. K-means is a partition
storage, but in this case, the vehicle owners can receive a payment from based clustering algorithm that iteratively update the cluster centers
the building operators. In a V2V setting, the vehicle acts as a mobile until convergence. It has high computing efficiency but is sensitive
charging unit, mainly to rescue vehicles on highways [38]. to the preset number of clusters and outliers. Hierarchical clustering
Battery swapping enables drivers to replace the discharged batteries is based on forming a hierarchical relationship between datasets. It
with a full battery within a few minutes in the battery swap stations has higher time complexity but is suitable for data sets with arbitrary
and eliminates the charging wait time [17]. Mobile charging enables shapes [42].
EV users to charge their vehicles at their preferred time and location. K-means algorithm and its variations are frequently used as a clus-
A mobile charging station can be a truck that carries a charging unit to tering model for location determination. Andrenacci et al. [43] develop
connect the EVs to the main grid, can have a separate battery to charge a fuzzy K-means clustering approach to determine the candidate loca-
the EVs, or can act as a mobile battery swap station. Vehicle to vehicle tions and estimate the energy demand of each cluster through vehicle
(V2V) power transfer is another type of mobile charging [39]. trajectory analysis. Karezmi et al. [44] integrate K-means clustering
with Genetic Algorithm to determine the zone borders while maximiz-
3. Infrastructure planning ing the distributor’s profit and considering EV users’ welfare. Wang
et al. [45] plan charging stations and scattered piles for electric vehicle
The electric vehicle adoption rate growth highly depends on the sharing systems. They use K-means clustering to determine candidate
electric vehicle charging infrastructure expansion. Building an ade- station locations and mean shift clustering algorithm for planning scat-
quate infrastructure is an expensive process with several constraints. tered pile locations. Their objective is to minimize the total cost, which
Infrastructure planning problem includes selecting the location and includes the costs related to queueing time, operation, construction,

3
N. Fescioglu-Unver and M. Yıldız Aktaş Renewable and Sustainable Energy Reviews 188 (2023) 113873

Fig. 1. Vehicle to everything.

Fig. 2. Infrastructure planning.

and maintenance. Marino and Marufuzzaman [46] determine potential population density. Momtazpour et al. [53] propose a coordinated
CS locations by principal component analysis (PCA) guided K-means clustering approach to cluster locations with low electricity load and
clustering. Bae et al. [47] propose the K-means based clustering al- high-income people. After determining the candidate locations, they
gorithm k-GRAPE (K-means group agent partitioning and placement assign EVs to stations to determine the demand at each location and
event) and maximize the EV users’ individual utilities (including dis- compute the required storage capacity. Wang et al. [54] plan the
tance and brand preferences) with a game theoretical approach. Liu charging facilities of electric buses and use AP clustering to plan the
et al. [48] consider renewable energy usage and use the K-means midway supply stations. Clio and Babacan [55] use Density-based
method to improve user satisfaction in clusters. Sánchez et al. [49] Spatial Clustering of Applications with Noise (DBSCAN) to identify the
use the K-means algorithm to solve the combined charging station smallest number of CS locations sufficient for a large EV taxi fleet,
placement and electric vehicle routing problem. considering the distances from the historic breakpoints of the drivers.
Hierarchical clustering algorithm has the benefit of working without Other machine learning algorithms such as reinforcement learning are
a pre-defined cluster number. Zeng et al. [50] determine the battery rarely used for infrastructure planning. As an example Ko [56] combine
swap station locations using K-means and hierarchical clustering meth- genetic algorithm and reinforcement learning to optimize electric tram
ods. They map GPS data to the road network by hidden Markov models, wireless charging infrastructure.
estimate the electricity consumption and the location of swap demand, In some studies, machine learning algorithms provide demand
characterize the spatial demand distribution by clustering methods, and prediction information to the main algorithm that solves the capac-
select the centroid of the clusters as the station locations. Bitencourt ity/sizing problem. Hu et al. [57] propose a multiple periods location
et al. [51] employ hierarchical clustering to determine semi-fast charg- and capacity expansion model. They employ a bayesian regularization
ing stations’ zones and determine each zone’s CS location/capacity back propagation neural network (Bayesian BPNN) model for electric
using an evolutionary search algorithm. vehicle quantity prediction and solve the model with greedy algorithm.
There are several other clustering algorithms used for solving the Jiang et al. [58] evaluate charging station location options for electric
infrastructure planning problem. Yao and Luo [52] use clustering by taxis. They use a multi agent simulation framework, employ Q-Learning
fast search-and-find of density peaks (CFSFDP) algorithm and consider to train the agents, and evaluate different charging station location

4
N. Fescioglu-Unver and M. Yıldız Aktaş Renewable and Sustainable Energy Reviews 188 (2023) 113873

Table 1
Infrastructure planning.
Ref Year Journal Type Objective ML method
[50] 2019 IEEE Access Location Min. distance K-means , Hierarchical clust.
[43] 2016 Applied Energy Location Min. distance Fuzzy K-means
[44] 2016 Applied energy Loc. & Cap. Max. profit K-means
[45] 2020 IEEE Transactions on Industry Applications Loc. & Cap. Min. cost K-means
[46] 2020 Journal of Cleaner Production Loc. & Cap. Min. cost PCA K-means
[48] 2018 Energy Loc. & Cap. Min voltage/load fluctuation, cost K-means
[47] 2020 IEEE Transactions on Intelligent Transportation Systems Location Max. EV utility k-GRAPE
[49] 2022 Energies Location K-means
[54] 2021 IEEE Transactions on Intelligent Transportation Systems Loc. & Cap. Min. cost AP clustering
[51] 2021 Electric Power Systems Research Loc. & Cap. Min. electrical losses, Min. zone center deviation Hierarchical clust.
[53] 2014 ACM Transactions on Intelligent Systems and Technology (TIST Loc. & Cap. Min. cost, distance Coordinated clust.
[52] 2018 Kybernetes Location Max. demand coverage CFSFDP clust.
[55] 2021 Applied Energy Location Min. number of CS locations DBSCAN
[57] 2020 Neurocomputing Loc. & Cap. Max. demand coverage Bayesian BPNN
[58] 2018 IET Generation, Transmission & Distribution Location Min. cost, Min. power loss/voltage deviation Multi-step Q-Learning

scenarios through evidential reasoning to optimize the benefits of algorithm (RF), k-nearest neighbor (kNN), and Chi-square automatic in-
both the taxi owners and the power grid. Table 1 summarizes the teraction detector (CHAID), and predict EV charging within households
studies which determine the infrastructure location (type:Location), through energy usage data. Kim and Kim [65] compare LSTM and ANN
and both location and capacity (type: Loc. & Cap.). Table 1 shows with time-series models in power consumption forecasts for short and
that studies have various objectives. Most frequent objectives are long term EV power consumption (features: Geographical data, special
minimizing distance [43,43,50], minimizing cost [45,45,46,48,54,58], day indicators, temperature). Table 2 lists the forecasting studies with
minimizing cost and distance together [53] and maximizing demand the machine learning algorithms they use.
coverage [52,57].
4. Charge control and coordination
3.2. Charging demand and charging station utilization forecasting
Uncoordinated EV charging can increase the EV owners’ electricity
Forecasting studies in this area can be grouped as long term purchase price and bring significant loads to the power grid. EV Smart
(monthly, yearly) and short term (daily, weekly). Short term fore- Charging (EVSC) uses information and communication technologies to
casting helps EV users and grip operators’ daily planning process. coordinate the EV charging system. The stakeholders in an EV charging
Charging station utilization prediction is generally handled as a short service system are the power grid operator, the aggregator, and the EV
term forecasting problem. Month, day, time interval, temperature, and owners.
precipitation are some of the features that machine learning algorithms The grid operator collects energy from power plants and renew-
use. Studies frequently use LSTM, random forest and kNN algorithms able energy sources. The aggregators are responsible for managing
for classification and/or regression. LSTM algorithm is based on re- the interaction between a group of EV owners and the grid operator.
current neural networks and frequently used in time-series prediction. Aggregators collect the charging demand and connection times from
Random forest (RF) combines multiple decision trees and it is easier to the EVs via smart meters and make energy buy/sell decisions and
evaluate the importance of different variables in prediction. K-nearest negotiations with the power grid [13]. An aggregator can manage the
neighbor (kNN) algorithm is a non parametric method which uses charge/discharge schedule of connected EVs such that the power grid,
a selected distance measure to determine the nearest neighbors. The EV owners, and the aggregator benefit from this operation. The grid
algorithm performances vary by the problem type. Most studies select operator can benefit from services such as peak load reduction/shifting,
several forecasting algorithms and compare their performances on the frequency and voltage regulation [14,66]. Meanwhile, EV owners ben-
same data set. efit from reduced charging costs. Fig. 3 shows the interactions between
Hecht et al. [59] use gradient boosting (GB) and random forest the grid operator, aggregator, and EVs.
(RF) classification algorithms to predict the next day CS occupation in Charge scheduling studies frequently use linear, non linear, mixed
five classes from very low to very high (features: Temperature, pre- integer, dynamic programming, game theory and metaheuristics ap-
cipitation, weekday, holiday, traffic). Ostermann et al. [60] compared proaches [67]. Nimalsiri et al. [68] analyze the charge control architec-
different forecasting algorithms to predict the short term occupancy tures in three categories: Centralized, decentralized, and hierarchical.
of charging points (features: Month, day, weekday, interval, holiday, In centralized architectures, a central decision maker controls all EV
bringing day, time). Friese et al. [61] use agglomerative clustering to charging schedules. Decentralized architectures include cases where the
predict charging point occupancy and random forest classification to EVs communicate with each other or use an aggregator as a communi-
analyze CS usage patterns (features: CP occupancy — weekday and cation bridge while determining their charging schedules. Hierarchical
time, CS pattern — the number of detached or row houses, commercial architectures use multiple aggregators, each coordinating a group of
buildings, households with a monthly net income of 900 EUR, and the EVs. In this case, the aggregator may act as a decision maker for the
buying power per inhabitant). Aduama et al. [62] use LSTM to fore- connected EVs, communicate with a central aggregator and control the
cast charging station energy demand (features: Temperature, humidity, charging schedule of EVs directly or indirectly (e.g., through electricity
wind speed). price).
Charging demand long term forecasting helps the infrastructure Fig. 4 depicts an architecture with different aggregator
and expansion planning. Yi et al. [63] forecast monthly commercial types. Aggregator-1 acts as a main aggregator for the sub aggregators
EV charging demand on a state/city scale with sequence to sequence (Aggregators 5 & 6), which are the charging stations. Aggregator-2 is a
(Seq2Seq) deep learning and long short term memory (LSTM) models utility company sending/collecting information to/from smart chargers
(features: Number of monthly check-ins, charging site rating, number at homes. Aggregator-3 is a parking lot manager. Aggregator-4 is a
and type of plugs, opening hours, parking spots, neighbor point of charging station directly connected to the grid operator. In addition,
interest). Verma et al. [64] compare several classification algorithms, aggregators may have their own renewable energy sources and storage
including classification and regression trees (CART), Random forest units.

5
N. Fescioglu-Unver and M. Yıldız Aktaş Renewable and Sustainable Energy Reviews 188 (2023) 113873

Table 2
Charging demand and CS utilization forecasting.
Ref Year Journal Type Term ML method
[59] 2021 Energies CS utilization Short term GB, RF
[60] 2022 Energies CS utilization Short term Linear and logistic regression, kNN, XGBoost, RF
[61] 2021 Sustainability CS utilization, CS usage patterns Short term Agglomerative Clust., RF
[62] 2023 Energies CS energy demand Short term LSTM
[63] 2022 Journal of Intelligent Transportation Systems Demand Long term Seq2Seq, LSTM, XGBoost, ARIMA, Prophet
[64] 2019 Computers & Industrial Engineering Demand Long term CART, RF, kNN, CHAID
[65] 2021 Energies Demand Long/short term ARIMA, ANN, LSTM

Fig. 3. Interactions between EV, Aggregator, and Grid Operator.

Fig. 4. EV, Aggregator, Grid Network Architecture.

6
N. Fescioglu-Unver and M. Yıldız Aktaş Renewable and Sustainable Energy Reviews 188 (2023) 113873

Wang et al. [66] classify the smart charging algorithms from the problem of a shared electric vehicle fleet using reinforcement learning.
perspectives of the grid operator, aggregator, and user (customer). In this case, the model’s user is the fleet operator, and the objective is
The grid operator objectives include grid stability and power quality. to maximize the profit by maximizing the fleet revenue and minimizing
Comprehensive reviews [69,70] investigating electric vehicle charging the charging costs. Vandael et al. [83] plan day ahead charge schedule
station effects on the grid exist in the literature. This review study of an EV fleet using a fitted Q-iteration based reinforcement learning.
mainly concentrates on the end user’s charging experience and handle Looking from the aggregators’ perspective, Arif et al. [84] solve
the user and aggregator perspectives. For user type scheduling algo- the scheduling problem of a parking lot aggregator with interruptable
rithms, the objective is minimizing the total cost while satisfying the charging through Q-Learning. Cao et al. [85] develop an actor-critic
charging demand of the vehicles on time — before the vehicle departs. learning based scheduling model with continuous charging amount
The disadvantage of user type algorithms is that when the number of action space for charge scheduling at a CS. In another study, they work
vehicles using the same scheduling algorithm increase, it may create with shared autonomous EVs, which can more easily be scheduled to
new peak points in low price regions. The users can be fleet owners act- participate in V2G services [86].
ing as aggregators who schedule their vehicles’ charge times with cost Lee et al. [87] handle a single public charger, model its usage
minimization objectives. Electric vehicle fleet charging problems show pattern through kernel density estimation, and use this pattern for
different energy consumption requirement characteristics depending on training the DQN (deep Q-network) for charge/discharge scheduling
the type (ultra-light, light, heavy), number of vehicles, and drivers [71]. with cost minimization objective in a dynamic pricing environment.
Aggregators can directly control charging schedules or indirectly Wan et al. [88] predict energy prices using neural networks and min-
affect the schedules through dynamic pricing (pricing models inves- imize the total cost of a group of EVs using multiagent DRL. Ye
tigated in Section 5). Aggregator type scheduling algorithms aim to et al. [89] propose a deep Q-Learning algorithm for charge/discharge
minimize the operating cost or maximize the profit of the aggregator. scheduling and centrally allocate chargers to vehicles. In this study,
In both the user and aggregator perspectives, the primary cost item is the chargers make their charge/discharge decisions individually to
the electricity purchasing cost, while battery degradation cost is also maximize the charging station’s profit. Aljafari et al. [90] propose
included in several studies. Meeting the user demand on time can be a a charge/discharge scheduling model using multi agent deep neural
constraint or can be part of the cost with an associated penalty. networks. Wang et al. [91] propose a reinforcement learning based
Uncertainties in the charge control problem include EV arrival/ charge scheduling model for a photovoltaic and energy storage inte-
departure times, charging time and quantity preferences, and elec- grated charging station. Shin et al. [92] develop a multiagent deep
tricity prices. The energy quantity available is also uncertain if the reinforcement model for distributed charging stations with photovoltaic
energy source is partly or fully photovoltaic or wind. Machine learning energy and storage systems. Jin and Xu [93] introduce a nodal multi
algorithms provide effective online solutions under these uncertain- target policy for scheduling a large scale EV fleet under renewable en-
ties. Charge/discharge scheduling models frequently use reinforcement ergy and price uncertainties, and determine the target levels and apply
learning (RL) based algorithms. Studies employ different RL tech- the policy through a soft actor critic RL model (SAC). Liu et al. [94]
niques, such as Q-Learning, deep Q-networks, and deep deterministic use neural networks for demand forecasting and proposes a mixed-
policy gradient methods, or propose new RL based scheduling algo- integer nonlinear programming model and a heuristic model for charge
rithms and test their effectiveness. Q-Learning is a model free, value scheduling. Yan et al. [95] propose a multi agent DRL based soft actor
based reinforcement learning (RL) model that uses the Q-table to track critic (SAC) model to minimize the charging cost and range anxiety
the value of taking an action in a given state. Q-Learning requires while avoiding the transformer overload in a residential area. Ming
discrete state variables and discrete actions. When the state variables et al. [96] propose a two level coordinated charge/discharge scheduling
take continuous values, a Q-table is not sufficient. Deep Q-Learning model that decides the total charge/discharge energy to be distributed
algorithm uses the deep Q-network (DQN) instead of the Q-table and by a constrained DRL (constrained SAC and PPO) and distributed the
can handle continuous state space. Deep deterministic policy gradient energy with a priority rule based on the energy demand amount.
(DDPG) is a model free actor-critic model and enables continuous state Aggregators selling electricity to the local market should determine
and action space. Both DQN and DDPG models belong to the Deep re- their bidding volume and price in addition to the charging schedule.
inforcement learning (DRL) algorithms class. LSTM method frequently Tao et al. [97] develop a deep reinforcement based bidding strategy
accompanies the scheduling models by electricity price forecasting. for EV aggregators and determine the quantity and price to purchase
Looking from the user’s perspective, Li et al. [72] propose a con- and sell electricity at each time slot in the local electricity market under
strained deep learning algorithm to guarantee a fully charged battery information asymmetry. Lin et al. [98] use an LSTM model to predict
before departure. Yan et al. [73] consider drivers’ range anxiety and the EV departure/arrival times, which provides the EV load in each
combines supervised learning with soft actor critic reinforcement learn- period to the mixed integer linear programming model that determines
ing for charge scheduling. Several studies employ LSTM to predict the optimal bidding volume at the electricity market.
future electricity prices and make EV owners’ charge/discharge sched- Battery swap stations also act as aggregators in charge control
ules with deep reinforcement learning. The main objective of these systems. Asadi and Pinkley [99] handle a battery swap station where
studies is cost minimization while avoiding on trip battery depletion. battery replacement cost comes in addition to the electricity purchase
Some studies use cost minimization as the single performance met- and battery degradation costs. They propose a reinforcement learning
ric [74–76], while others evaluate user satisfaction [77] and consider based algorithm to generate a charge, discharge, and battery replace-
range anxiety factor [78]. Zhang et al. [79] combine the user and grid ment schedule. Gao et al. [100] propose a deep deterministic policy
operator perspectives and propose a multi agent deep reinforcement gradient (DDPG) algorithm to serve batteries to electric buses while
model and LSTM model to minimize the cost for users and avoid minimizing the operating cost in a charge/discharge environment.
grid transformer overload. Chiş et al. [80] predict electricity prices Wang et al. [101] use battery swap stations as fast frequency regu-
through a bayesian neural network and uses fitted Q-iteration model lation services and use deep Q-Network for charge scheduling. Liang
for charge scheduling. While reinforcement learning is frequently used et al. [102] handle a system where a central battery changing station
for scheduling, Lopez et al. [81] propose different machine learn- charges the batteries and send them to swap stations via trucks. They
ing methods (kNN, shallow neural network-SNN and a deep neural combine multi agent deep reinforcement learning (DRL) and binary
network-DNN) to determine the charge schedule of a connected EV integer programming for joint scheduling of charging, swapping, and
with a cost minimization goal. During training the DNN, they use op- truck routing operations. Tan et al. [103] combine LSTM, DRL, and
timal solutions obtained with dynamic programming. Liang et al. [82] mixed integer linear programming model to schedule swap operations,
solve the joint charge scheduling, order dispatching, and rebalancing where LSTM predicts the electricity prices and demand, DRL allocates

7
N. Fescioglu-Unver and M. Yıldız Aktaş Renewable and Sustainable Energy Reviews 188 (2023) 113873

Table 3
Charge control and scheduling.
Ref Year Journal Type Objective ML method
[72] 2019 IEEE Transactions on Smart Grid User Min. cost Safe DRL
[73] 2021 IEEE Transactions on Smart Grid User Min. cost and anxiety SAC
[77] 2020 IEEE Internet of Things Journal User Min. cost DDPG, LSTM
[74] 2021 Journal of Modern Power Systems User Min. cost DDPG, LSTM
and Clean Energy
[78] 2018 IEEE Transactions on Smart Grid User Min. cost DQN, LSTM
[76] 2020 IEEE Transactions on Vehicular User Min. cost Dyna-Q RL, LSTM
Technology
[75] 2020 Computers and Electrical Engineering User Min. cost Q-Learning, LSTM
[79] 2022 IEEE Transactions on Intelligent User and grid Min. cost and power load Multiagent DRL, LSTM
Transportation Systems
[80] 2016 IEEE Transactions on Vehicular User Min. cost Fitted Q-iteration, Bayesian NN
Technology
[81] 2018 IEEE Transactions on Smart Grid User Min. cost KNN, SNN, DNN
[82] 2020 IEEE Transactions on Smart Grid User Max. profit DRL
[83] 2015 IEEE Transactions on Smart Grid User Min. cost Fitted Q-iteration
[84] 2016 Sustainable Energy, Grids and Aggregator Min. cost Q-learning
Networks
[86] 2020 IEEE Internet of Things Journal Aggregator Min. cost Q-Learning
[85] 2021 IEEE Internet of Things Journal Aggregator Min. cost Actor-critic learning
[87] 2020 Energies Aggregator Min cost DQN, Kernel density estimation
[88] 2022 Journal of the Franklin Institute Aggregator Min. cost Multiagent DRL, NN
[90] 2023 Computers and Electrical Engineering Aggregator Max. profit Multiagent DNN
[89] 2022 IEEE Transactions on Smart Grid Aggregator Max. Profit DQN
[91] 2022 Sustainability Aggregator Min. cost and improve clean Double DQN, Dueling DQN
energy consumption
[92] 2020 IEEE Transactions of Industrial Aggregator Min. cost DQN
Informatics
[93] 2020 IEEE Transactions on Smart Grid Aggregator Min. cost SAC
[94] 2022 International Journal of Electrical Aggregator Min. cost NN
Power and Energy Systems
[95] 2022 IEEE Transactions on Industrial Aggregator Min. cost and anxiety Multi agent DRL, SAC
Informatics
[96] 2023 Applied Energy Aggregator Min. cost Const. SAC, PPO
[97] 2021 IEEE Transactions on Industrial Aggregator Max. profit Async. DRL, DDPG
Informatics
[98] 2020 IEEE Access Aggregator Max profit LSTM
[99] 2021 Transportation Research Part E: Aggregator Max. profit RL
Logistics and Transportation Review
[100] 2020 IEEE Transactions on Industry Aggregator Min. cost DDPG
Applications
[101] 2020 IEEE Transactions on Industrial Aggregator Max. profit DQN
Informatics
[102] 2022 IEEE Transactions on Smart Grid Aggregator Max. profit Multi-agent DRL
[103] 2023 Engineering Applications of Artificial Aggregator Min. cost LSTM, DRL
Intelligence
[104] 2023 IEEE Transactions on Smart Grid Aggregator Min. cost PPO

power to chargers, and MILP model allocates charger power to batter- waiting times or decreased charging costs. Time varying charging pric-
ies. Jin et al. [104] determine the total charge amount bu proximal ing models include time of use (TOU) pricing and dynamic pricing. TOU
policy optimization (PPO) and allocate this amount among waiting bat- pricing depends on time and does not change dynamically according
teries using the least demand first priority rule, under renewable energy to the operating conditions. Dynamic pricing allows service operators
and electricity price uncertainties. Table 3 summarizes the machine dynamically vary the price EV users pay for charging [105]. Charging
learning methods and objectives of the charge control and scheduling service dynamic pricing techniques frequently used in the literature
studies. The objectives are frequently minimizing cost [72–79,81,83, include evolutionary algorithms, mixed integer linear/nonlinear pro-
85–88,91–94,100,103] or maximizing profit [82,84,89,90,97–99,101, gramming, control theory, and dynamic programming [18]. Machine
102], sometimes with additional objectives such as improving clean learning, especially reinforcement learning use, is increasing in this
energy consumption [91] or minimizing user anxiety [73,95]. area.
The dynamic pricing problem requires decision making in a stochas-
5. Dynamic pricing tic and continuously changing environment with uncertain EV ar-
rival/departure times, charge demand, and utility company electric
An efficient time varying pricing strategy creates benefits for service sales prices changing in real time. The charging location can be a
providers and customers. Service providers’ benefits include grid load parking lot where the vehicles are parked for a long duration or a fast
balancing and increased profits. Customers benefit from decreased charging station where the owners wait alongside their vehicles. The

8
N. Fescioglu-Unver and M. Yıldız Aktaş Renewable and Sustainable Energy Reviews 188 (2023) 113873

same price can be valid for all vehicles, or stations can charge different consumption. A realistic energy consumption computation requires
prices with a differentiated service. taking the environmental conditions, road conditions, driver behavior
Charging parked electric vehicles requires joint determination of and vehicle type/load into account [117,118].
charging price and charge schedule. Parked EVs’ demand should be The total travel time of a vehicle includes the driving time, waiting
satisfied before they leave, and EVs will pay from the price rate valid at time at the station, and charging time (if charging is part of the plan).
the time they enter the parking area. Wang et al. [106] and Paraskevas Uncertainties include traffic and road conditions, energy consumption,
et al. [107] determine the charging price and the total charging rate at changing driver behavior, and waiting times at the stations. Fig. 5
each slot through state–action-reward-state–action (SARSA) and deep summarizes an electric vehicle navigation problem where vehicles
Q-Learning models respectively. The total charging rate is later dis- use different routes and charging stations depending on the traffic
tributed among EVs using the least laxity first algorithm. Moghaddam condition, charging price, and congestion at the stations.
et al. [108] introduce an EV load-shifting mechanism with fast adaptive Studies that use machine learning as a part of the solution approach
heuristic critic (Fast-AHC) based reinforcement learning and aim to employ it either for the primary path planning tasks or as an aid for pro-
increase the station’s profit while maintaining the utility’s stability. viding input forecasts to the path planning models. This section divides
In the fast charging stations case, EV owners have a limited time, the EV routing/navigation studies into three subsections. Sections 6.1
and they can quit waiting for charging after a while in the queue. and 6.2 investigate the EVRP and EV navigation studies that use ML
Fang et al. [109] propose a Q-Learning model for dynamic pricing as the main path planning algorithm, respectively. Section 6.3 lists the
at a fast charging station and show that as the backlog rate (ratio of studies that use ML to provide an input forecast to the path planning
people accepting to wait) increases, the revenue benefit of dynamic model.
pricing wrt. stationary pricing also increases. Quian et al. [110] handle
multiple charging stations in a competitive environment and propose 6.1. Electric vehicle routing problem and variations
a multi agent deep reinforcement learning model for determining the
optimal pricing strategy for each charging station. Electric vehicle routing problem (EVRP) is finding a set of vehicle
In the differentiated service case, Quality of Service (QoS) level is routes that service a set of customer nodes and depart/end at a given
part of the objective or a constraint. QoS metric may include items depot node. EVRP is a generalization of the VRP (Vehicle routing
like EV waiting time, service rejection rate, and EV owner’s satisfaction problem) and is considered NP-hard in the strong sense [117]. Solution
with charging price. Abdalrahman and Zhuang [111] investigate a approaches for EVRP include exact algorithms like branch-price-and-
charging service infrastructure with multiple stations with multiple cut and heuristic algorithms like genetic algorithm, tabu search, and
service types (AC charging, fast DC charging, wireless charging) and iterated local search. Solving the EVRP problem in real time brings
develop a twin delayed DDPG based differentiated pricing model that additional uncertainties, such as dynamic customer requests. Machine
maximizes station utilization while maintaining the target QoS level. learning solutions can anticipate and prepare for future demand and
Zhao and Lee [112] consider a charging station with multiple charging changing conditions and accelerate solving large scale problems.
service levels that share the same waiting area and propose a deep rein- There are several variants of the EVRP. Machine learning algorithms
forcement based dynamic pricing framework considering differentiated are especially used in online routing problems with dynamic customer
service requirement levels and maximizing the cumulative QoS. request arrivals (DEVRP — Dynamic Electric Vehicle Routing Problem).
Looking from the EV owners’ perspective, Zhang et al. [113] pro- Depending on the problem type, the models determine when to relocate
pose a cooperative multi agent-DQN based bidding strategy for an EV the EVs, when to assign them to which customers, and when to send
fleet. The EV users determine their charging quantity and unit price them to charging. The charging location can be the depot or a nearby
offer through this strategy to maximize their utilities, and charging charging station, and the charging level can be full or partial. James
stations determine the winning bids with an auction mechanism. Kim et al. [119] propose a computationally fast vehicle routing model for
et al. [114] propose two Q-Learning algorithms. The first is a dynamic online transportation in large networks where requests arrive/get can-
pricing model for the service provider, and the second is a charge celed dynamically. They use neural combinatorial optimization for tour
scheduling model for the EV owners. Simultaneous use of these models planning and deep reinforcement learning for determining the neural
decreases costs for both parties. network model parameters. Basso et al. [120] handle dynamic customer
Using the V2G technology, EV aggregators and service providers can requests and use Q-Learning for route planning with minimum energy
act as prosumers and both buy and sell electricity at the local energy consumption and battery depletion risk. Shi et al. [121] solves a com-
market. Qui et al. [115] combine the prioritized experience replay munity operated fleet’s ride hailing problem using deep reinforcement
method with deep deterministic policy gradient (DDPG) to determine learning. Kullman et al. [122] use a deep Q-Learning variant - dueling
the prices the aggregator offers to a population of EVs with and without double DQN (D3QN) with attention mechanism for the ride hailing
V2G flexibility. Liu et al. [116] increase the price change frequency and problem. Lin et al. [123] use RL model with an attention mechanism
establish a quarter-hourly pricing strategy using a DDPG based rein- to solve the EVRP with time windows. For this static problem type, RL
forcement model. Table 4 summarizes these studies. The main objective helps solve very large problem instances.
of most studies is maximizing profit or minimizing cost [106–110,114–
116], while maximizing stability [108], quality of service [111,112], 6.2. Electric vehicle navigation with charging station selection
and EV user utility [113] are additional objectives in some studies.
Electric vehicle navigation problems route electric vehicles from an
6. Electric vehicle routing, navigation and charging station selec- origin to a destination point. Unlike the EVRP type problems, there
tion are no customer nodes to stop and serve at. Navigating with charging
station selection (EVCS) studies, route the vehicles so that each ve-
Electric vehicle routing studies can be divided into two main classes: hicle can charge at a charging station. The charging station selection
Electric vehicle routing problems (EVRP) and electric vehicle naviga- decision may increase the total distance traveled depending on the
tion problems. EVRP can be described as the set of problems that route location of the CS or may increase the total travel time depending
one or more electric vehicles to serve one or more customer nodes. on the queuing time at the selected station. Charging station selec-
Electric vehicle navigation problems do not include customer stops tion brings additional uncertainties to the routing problem: charging
but they route the electric vehicles from an origin to a destination prices and random arrivals/waiting times at CSs. Studies in this area
point. Some of these studies include charging stops to the plan, while frequently use reinforcement learning based algorithms to overcome
others focus on arriving at the destination point with minimum energy these uncertainties. Zhang et al. [124] use the Q-Learning algorithm to

9
N. Fescioglu-Unver and M. Yıldız Aktaş Renewable and Sustainable Energy Reviews 188 (2023) 113873

Table 4
Dynamic pricing.
Ref Year Journal Type Objective ML method
[106] 2019 IEEE Transactions on Industrial Informatics Price and schedule Max. profit SARSA
[107] 2022 Energies Price and schedule Max. profit DNN, DQN
[108] 2022 IEEE Access Price Max. profit and stability Actor-critic, Fast-AHC
[109] 2020 IEEE Transactions on Intelligent Transportation Systems Price Max. profit Q-Learning
[110] 2021 IEEE Transactions on Power Systems Price Max. profit Multi agent DRL
[111] 2020 IEEE Transactions on Intelligent Transportation Systems Differentiated service Max. demand w. QoS DRL
[112] 2021 IEEE Transactions on Transportation Electrification Differentiated service Max. QoS DNN, DRL
[114] 2015 IEEE Transactions on smart grid Price and schedule Min. cost Q-Learning, PDS learning
[113] 2020 Neurocomputing Price and schedule Max. EV util. Multi agent DQN
[115] 2020 IEEE Transactions on Industry Applications Price and schedule Max. profit DDPG
[116] 2021 IEEE Access Price Max. profit DDPG

Fig. 5. Electric vehicle navigation and charging station selection.

allocate EVs to charging stations while minimizing the station queueing planning process by energy consumption, charging price, speed, and
times and the travel time, including road congestion. Qian et al. [125] travel time forecasts.
included the varying charging costs at CSs and proposed a deep Q- Thibault et al. [131] plans the EV path with energy consumption
Learning based model to minimize the total travel time (road and and travel time minimization objective and uses ANN to approximate
station waiting times) and charging costs. Lee et al. [126] introduce the optimal vehicle speed for real time online usage. Shukla et al. [132]
a central management system that takes charging requests from EVs, use DNN for traffic volume and station wait time prediction and allo-
current state information (i.e., charging/waiting EVs) from CSs, and cate CSs to EVs to minimize travel time. Huang et al. [133,134] propose
recommends/reserves CSs to minimize the total travel time of multiple a central route and charge planner for EVs to prevent traffic congestion.
EVs, using deep Q-network (DQN) based model. Xu et al. [127] propose The first study [133] estimates the travel time by Support Vector
a graph reinforcement learning based fast charging station recommen- Regressions, using weather, day of the week, traffic density, light wait
dation system that minimizes the total travel and station waiting times time, and passenger pick up/drop off time information. The second
while considering the stability of nodal voltages. Tao et al. [128] study [134] uses support vector regressions (SVR) to compute the EV
charging times and travel times and propose algorithms to avoid rush
handles a multi-type charging service recommendation system with
hour congestion. Basso et al. [135] use linear programming for EVRP
fast charging, slow charging, and battery swap options and uses deep
path planning while forecasting energy usage with Bayesian Regression
Q-Learning with mixed integer linear programming to dispatch the
techniques. Lu et al. [136] use A-star algorithm for path planning with
charging/swapping orders. Suanpang et al. [129] propose a multi agent
charging station selection of EV users with different travel objectives
deep reinforcement learning model for charging station selection, with
(travel distance, travel time, wait time, charge cost) and integrates
waiting time, charging price, and charging failure (incomplete charge)
gated recurrent unit neural network (GRU NN) for charging price
rate lowering objective. Jin and Xu [130] develop an advantage actor-
forecasts.
critic (A2C) algorithm for CS selection and path planning. They use the Another group of studies navigate EVs while minimizing their en-
shortest path algorithm to determine the route to the CSs and select a ergy consumption (EVEC). In these studies, vehicle charging either does
CS through the actor-critic model. not exist or is done using on-route charging stations with no waiting
time. They consider energy consumption factors such as acceleration,
6.3. Forecasting for path planning deceleration, hill climbing, rolling resistance, and aerodynamic drag
force. Although these studies do not explicitly handle the charging
In several studies, path planning is completed by a linear program- service, they provide valuable insight into how machine learning meth-
ming or heuristic approach. Machine learning algorithms help the path ods aid in minimizing energy consumption by road planning or by

10
N. Fescioglu-Unver and M. Yıldız Aktaş Renewable and Sustainable Energy Reviews 188 (2023) 113873

Table 5
Electric vehicle routing, navigation, and charging station selection.
Ref Year Journal Type Objective ML method
[119] 2019 IEEE Transactions on Intelligent Transportation Systems EVRP Min. distance, Max. delivered requests DRL, DNN
[120] 2022 Transportation Research Part E EVRP Min. energy, Min. battery depletion risk Q-Learning
[121] 2019 IEEE Transactions on Intelligent Transportation Systems EVRP Min. distance, waiting time DRL
[122] 2022 Transportation Science EVRP Max. profit D3QN
[123] 2021 IEEE Transactions on Intelligent Transportation Systems EVRP Min. distance DRL
[124] 2019 Sustainability EVCS Min. total time Q-Learning
[125] 2019 IEEE Transactions on Smart Grid EVCS Min. total time, cost DQN
[126] 2020 Energies EVCS Min. total time DQN
[127] 2022 International Journal of Electrical Power and Energy Systems EVCS Min. total time, Balance power system Graph RL
[128] 2022 Applied Energy EVCS Min. total time, Max. revenue DQN
[129] 2023 Sustainability EVCS Min wait time, price, failure Multi agent DRL
[130] 2022 IEEE Internet of Things Journal EVCS Min total time and cost A2C
[131] 2018 IEEE Transactions on Intelligent Vehicles Forecast Min. energy ANN
[133] 2021 Information Technology and Control Forecast Min. travel time SVR
[134] 2021 International Journal of Computational Intelligence Systems Forecast Min. distance, total time, cost SVR
[132] 2020 Internet of Things Forecast Min. total time DNN
[135] 2021 Transportation Research Part B: Methodological Forecast Min. energy Bayesian Regression
[136] 2019 IEEE Access Forecast Min. total time, cost GRU NN
[137] 2020 IEEE Transactions on Vehicular Technology EVEC Min. energy Q-Learning
[138] 2021 Electric Power Systems Research EVEC Min. energy DDQN
[140] 2021 Frontiers in Big Data EVEC Min. energy Q-Learning
[141] 2021 Applied Energy EVEC Min. total time, energy RNN
[139] 2022 Vehicles EVEC Min. energy SARSA
[142] 2017 IEEE Transactions on Intelligent Vehicles Forecast Min. total time, energy K-means, Naive Bayes
[143] 2021 ACM Transactions on Internet Technology Forecast Min. energy DNN, Linear regression, Decision tree

energy consumption forecasting. Zhang et al. [137] use the Q-Learning


method to minimize the total energy (fuel + electricity) spent by a
PHEV considering the distance traveled, road slope, and vehicle speed.
Aljohani et al. utilize double deep Q-Learning (DDQN) [138] and
SARSA [139] to develop routing models that minimize EV’s energy
utilization. Dorokhova et al. [140] use Q-Learning approach to generate
energy feasible paths while including energy recuperation of EVs and
allowing charging on the route (station availability-wait times not
considered). Bi et al. [141] propose a central controller model which
uses a bilevel optimization algorithm to minimize travel time and
energy consumption of fuel and electric vehicles on the roads. They
use a random neural network (RNN) model for route planning and a
nonlinear optimization model for generating incentives that motivate
road users. Bozorgi et al. [142] employ data mining techniques to
Fig. 6. Frequently used machine learning techniques.
extract the desired speed profiles that minimize travel time or energy
consumption and use this information to select a route among the
candidate shortest time routes. Ayman et al. [143] employs DNN,
linear regression, and decision tree for energy consumption forecast conference proceedings are not included in the search. Lastly, studies
and assigns mixed fleets to fixed routes by Genetic algorithm. Detailed in languages other than English are not covered in this review.
reviews exist on energy consumption [144] and range prediction [145] Most of the studies covered (90%) belong to the years 2018–2022.
studies. Fig. 6 provides a visual insight into frequently used machine learning
Table 5 summarizes the studies with their problem types, methods, techniques for charging service area.
and objectives. The main objectives can be categorized as minimizing Machine learning algorithms aid different planning/service opera-
distance [119,121,123,134], minimizing time (travel, waiting or total tions in multiple ways. The first step in charging service planning is
time) [124–130,132–134,136,141,142], minimizing energy [120,131, building a cost-efficient infrastructure with sufficient area coverage.
135,137–143] and minimizing cost/maximizing profit [122,125,128, Forecasting models provide valuable input to infrastructure planning
130,134,136]. algorithms through long term and short term demand prediction. Clus-
tering algorithms aid in determining the location and capacity of the
7. Findings and research directions charging stations.
Once the infrastructure is planned, the problem is how to control
Electric vehicle charging service optimization research is growing and coordinate the dynamic charging requests of electric vehicles.
continuously. Machine learning methods have an increasing share in There are several uncertainties in the charge control problem, including
the literature. This study reviews the machine learning applications the electric vehicle arrival/departure times, vehicles’ state of charge
related to charging service operations. The study covers journal pub- and charge amounts requested, and the grid electricity prices. In addi-
lications indexed in Web of Science database, published between 2014 tion, when the primary energy source is solar or wind, the available
and 2023. The search was conducted separately for each section subject energy is also uncertain. Forecasting algorithms provide vehicle de-
- Infrastructure planning, charge control/coordination/scheduling, dy- parture/arrival time, charge demand, electricity price, and renewable
namic pricing, and routing, using different keyword combinations. This energy amount predictions. For charge control and coordination, rein-
study has several limitations. First of all, although the search covered a forcement learning based models provide effective scheduling solutions
broad range, there may still be relevant articles uncovered. In addition, in real time under uncertainty.

11
N. Fescioglu-Unver and M. Yıldız Aktaş Renewable and Sustainable Energy Reviews 188 (2023) 113873

Dynamic pricing models determine the aggregators’ electricity sell- • Long term demand forecasting studies provide essential
ing price in real time under the same uncertainties. Reinforcement inputs to infrastructure planning problems. Including new
learning based multi agent and single agent models determine the features, such as government policy predictions, may im-
aggregators’ selling price in real time while minimizing the cost, max- prove these forecasts’ accuracy.
imizing the profit, or maximizing the quality of service. Finally, nav- • Infrastructure resilience during natural disasters such as
igating an electric vehicle requires energy consumption, travel time, earthquakes or floods is another important research area.
and speed forecasts. Forecasting algorithms provide these inputs to The infrastructure requirements differ according to the
the navigation models. Navigation plans may include stops at charging disaster type [148]. EVs both act as a resilience resource
stations or direct the vehicle from departure to destination points with for infrastructures, and require additional energy storage
the available charge. Reinforcement learning based models navigate the systems/renewable energy for usage during power out-
vehicles successfully under several constraints. ages [149]. Along with mixed integer [150] and multi
Charging service optimization studies covered in this study in- objective programming [151] based infrastructure opti-
clude stand alone machine learning applications and machine learn- mization studies, capacity simulation [152], EV prioritiza-
ing/mathematical optimization hybrid models. Machine learning mod- tion [153] and evacuation routing [154,155] studies exist
els provide fast but not necessarily optimal responses in real time under in the literature. Scheduling and routing the EVs and
uncertainty. On the other hand, mathematical optimization models mobile charging units are some of the problems that ma-
(like mixed integer or dynamic programming) can provide optimal chine learning algorithms can help with under this highly
solutions to many problem types. However, depending on the type uncertain environment that requires a very fast response in
and scale of the problem, their response time may not be fast enough real time.
to use in real time decision making. Combining machine learning 2. Charge control and coordination
algorithms with mathematical optimization models and creating hybrid
models enables capturing the advantages of both worlds. Hybridiza- • Scheduling and control algorithms generally aim to max-
tion methods include; using machine learning models to provide in- imize the profit or minimize the costs of the aggrega-
puts/constraints to mathematical models, separating the main problem tors/users. As the number of renewable energy facilities
into sub-problems and solving each problem with the most appropriate and individual homes integrating solar energy grows, the
model type, and training machine learning models with mathematical studies focusing on minimizing the carbon footprint of EVs
model solutions. Charging service problems like scheduling and rout- through renewable energy usage should also increase.
ing are combinatorial optimization problems where machine learning • When considering the V2G setting, another uncertainty
can improve the already available optimization algorithms in several that arises is the willingness of EVs to participate in the
ways [146]. Therefore, increasing hybrid studies can speed up EV V2G system [14].
charging service planning research advancements. • Different business models like police and military have
different requirements and uncertainties than regular EV
Although they look separated, all charging service related research
fleets. Machine learning models can aid in this area.
areas are connected with each other. Current studies take each prob-
lem as a separate optimization problem to be solved with several 3. Dynamic pricing
assumptions related to the other parts of the bigger picture. Better
solutions are possible with a holistic approach that interconnects in- • Dynamic pricing models can promote renewable energy
frastructure planning with coordination, pricing, and navigation. A top usage and maximize renewable energy utilization.
down or bottom up iterative design framework can reduce the number • Integrating quality of service based dynamic pricing to fast
of assumptions required for each problem type and improve the real charging stations is another important research area with
world resemblance of the studies and the reliability of the solutions. additional uncertainties.
Simulation is a useful tool for determining model outcomes under • Priority service at fast charging stations requires different
resource management models [156,157]. Dynamic pric-
different conditions. Each step’s output/emergent requirements can be
ing models can improve their performance by balancing
feedback to the previous solution step. Fig. 7 proposes a framework for
different service type requests.
this type of interconnected study.
New technologies and business models come with their own un- 4. Routing - Navigation
certainties and response time requirements. Some research areas are
in their development phase, and the number of studies in these areas • Fast charging stations are mostly preferred for the shorter
needs to be expanded. Some of the areas that can benefit from machine charge time they offer. However, when congestion oc-
learning applications are as follows: curs, an accurate waiting time prediction is important to
determine the expected total time at each station.
1. Infrastructure planning • Wireless charging infrastructure brings additional uncer-
tainties to the routing problem.
• Wireless charging is considered an emergent technology
that will accelerate EV adoption [147]. Especially dynamic 8. Conclusion
charging, charging when the vehicle is en route, is a game
changing technology that will relieve the users’ range anx- Consumer perceived risks regarding the availability and quality of
iety. Dynamic charging technology brings its own uncer- electric vehicle charging services are important barriers in front of EV
tainties and requires additional infrastructure plans and adoption. Machine learning algorithms have a continuously growing
dynamic routing algorithms. However, studies in this area application area, including planning and service operations for electric
are still limited. vehicle charging.
• The number and capacity of renewable energy facilities are This study reviews electric vehicle charging infrastructure plan-
increasing continuously. Including predicted regional re- ning, charge control and coordination, dynamic pricing, and naviga-
newable resource availability and charging stations’ ability tion/routing studies that use machine learning algorithms as a part
to deploy and use their renewable resources could improve of the solution procedure. The study has several limitations. The
the infrastructure plans. search was limited to the journal publications listed in Web of science

12
N. Fescioglu-Unver and M. Yıldız Aktaş Renewable and Sustainable Energy Reviews 188 (2023) 113873

Fig. 7. Iterative design framework connecting the research areas.

database, and it includes only the publications in English language. This The future research areas that can benefit from machine learning
study improves the current review studies literature in several ways. applications can be summarized as follows:
The primary focus is on charging service operations that effect the
• Infrastructure planning: Renewable energy usage comes with its
users’ EV adoption decisions. Therefore this study has a broad coverage
own uncertainties. Using governments’ renewable energy policies
area that includes several topics interrelated with each other. It covers
to forecast charging stations’ renewable energy storage/usage
the studies that use machine learning as the main problem solving
potential can effect infrastructure plans. Government policies on
method or as an aid to the main method. This study also proposes an
promoting EV deployment is an important factor in predicting
iterative design framework that connects these research topics.
long term demand. Infrastructure planning for resilience is an-
This review summarizes how machine learning aids problem solving
other topic that has limited studies but very important for pro-
in the current literature and provides future research directions. The
viding mobility during times that require immediate response.
key findings of this study shows that depending on the problem type,
• Charge control and coordination: Along with the rise in renewable
machine learning algorithms serve as the main solution method or
energy deployment, studies focusing on EV charge control for car-
provide inputs/constraints to the main method. When the problem
bon footprint reduction should also increase. In the V2G setting,
to be solved requires fast decision making in an environment with
forecasting the EV’s that will accept participation will improve
uncertainty, reinforcement learning class models provide successful
solution accuracy.
solutions. When an input to the main method has uncertainty and
• Dynamic pricing: Studies promoting renewable energy usage
depends on several other features, forecasting algorithms aid in pre-
through pricing, considering quality of service and different ser-
dicting it. When the problem scale gets bigger and it is not possible
vice level requirements for fast charging stations are important.
to solve with mathematical models within a reasonable time limit,
• Routing/Navigation: Fast charging station waiting time predic-
machine learning algorithms provide satisfactory solutions. Creating
tion is an important but frequently ignored topic. Wireless charg-
hybrid models with machine learning and mathematical models cap-
ing is another emergent research area with its additional uncer-
tures the advantages of both methodologies. Although infrastructure
tainties.
planning, charge control, pricing, and routing studies are interrelated,
they are treated separately with additional assumptions regarding other This study shows that machine learning is a powerful tool that
layers. This study proposes an iterative design framework that con- aids in all EV charging service operation levels. Government policies
nects these areas through simulation. Simulation is a valuable tool on EV support, renewable energy promotions, electricity pricing reg-
to determine the new requests and requirements that will emerge ulations are important factors that affect the demand and resource
under different conditions. When the EV charging service operations’s inputs/constraints of all stages. A holistic approach is required to
interrelated research layers give each other feedback, it will be possible enable EV widespread adoption. Policymakers should use an iterative
to create a holistic design with less assumptions and more resemblance design framework and follow a top down or bottom up approach
to the real-life. to evaluate the effects of their decisions before implementing them.

13
N. Fescioglu-Unver and M. Yıldız Aktaş Renewable and Sustainable Energy Reviews 188 (2023) 113873

Researchers can include policy forecasts to obtain more realistic results. [22] Kumar NM, Chand AA, Malvoni M, Prasad KA, Mamun KA, Islam F, Chopra SS.
Engineers can follow the advances in machine learning algorithms and Distributed energy resources and the application of AI, IoT, and blockchain in
smart grids. Energies 2020;13(21):5739.
use hybrid models for successful designs.
[23] Ali SS, Choi BJ. State-of-the-art artificial intelligence techniques for distributed
smart grids: A review. Electronics 2020;9(6):1030.
Declaration of competing interest [24] Song C, Kim K, Sung D, Kim K, Yang H, Lee H, Cho GY, Cha SW. A review of
optimal energy management strategies using machine learning techniques for
The authors declare that they have no known competing finan- hybrid electric vehicles. Int J Automot Technol 2021;22:1437–52.
[25] Pevec D, Babic J, Podobnik V. Electric vehicles: A data science perspective
cial interests or personal relationships that could have appeared to
review. Electronics 2019;8(10):1190.
influence the work reported in this paper. [26] Amara-Ouali Y, Goude Y, Massart P, Poggi J-M, Yan H. A review of electric
vehicle load open data and models. Energies 2021;14(8):2233.
Data availability [27] Al-Ogaili AS, Hashim TJT, Rahmat NA, Ramasamy AK, Marsadek MB, Faisal M,
Hannan MA. Review on scheduling, clustering, and forecasting strategies for
controlling electric vehicle charging: Challenges and recommendations. IEEE
No data was used for the research described in the article.
Access 2019;7:128353–71.
[28] Abdullah HM, Gastli A, Ben-Brahim L. Reinforcement learning based EV
References charging management systems–a review. IEEE Access 2021;9:41506–31.
[29] Deb S. Machine learning for solving charging infrastructure planning problems:
[1] Electric vehicles report. Paris: International Energy Agency IEA; 2022, https: A comprehensive review. Energies 2021;14(23):7833.
//www.iea.org/reports/electric-vehicles (accessed Feb, 2023). [30] Shahriar S, Al-Ali A-R, Osman AH, Dhou S, Nijim M. Machine learning
[2] Kumar RR, Alok K. Adoption of electric vehicle: A literature review and approaches for EV charging behavior: A review. IEEE Access 2020;8:168980–93.
prospects for sustainability. J Clean Prod 2020;253:119911. [31] Sanguesa JA, Torres-Sanz V, Garrido P, Martinez FJ, Marquez-Barja JM.
[3] Featherman M, Jia SJ, Califf CB, Hajli N. The impact of new technologies on A review on electric vehicles: Technologies and challenges. Smart Cities
consumers beliefs: Reducing the perceived risks of electric vehicle adoption. 2021;4(1):372–404.
Technol Forecast Soc Change 2021;169:120847. [32] Leijon J, Boström C. Charging electric vehicles today and in the future. World
[4] Hardman S, Jenn A, Tal G, Axsen J, Beard G, Daina N, Figenbaum E, Electr Veh J 2022;13(8):139.
Jakobsson N, Jochem P, Kinnear N, et al. A review of consumer preferences [33] LaMonaca S, Ryan L. The state of play in electric vehicle charging services–A
of and interactions with electric vehicle charging infrastructure. Transp Res D review of infrastructure provision, players, and policies. Renew Sustain Energy
2018;62:508–23. Rev 2022;154:111733.
[5] Globisch J, Plötz P, Dütschke E, Wietschel M. Consumer preferences for public [34] Mohammed SAQ, Jung J-W. A comprehensive state-of-the-art review of
charging infrastructure for electric vehicles. Transp Policy 2019;81:54–63. wired/wireless charging technologies for battery electric vehicles: Classifica-
[6] Zhang B, Niu N, Li H, Wang Z, He W. Could fast battery charging effectively tion/common topologies/future research issues. IEEE Access 2021;9:19572–85.
mitigate range anxiety in electric vehicle usage? Evidence from large-scale data [35] Ahmad A, Alam MS, Chabaan R. A comprehensive review of wireless charging
on travel and charging in Beijing. Transp Res D 2021;95:102840. technologies for electric vehicles. IEEE Trans Transp Electrif 2017;4(1):38–63.
[7] Baumgarte F, Kaiser M, Keller R. Policy support measures for widespread [36] Triviño A, Gonzalez-Gonzalez JM, Castilla M. Review on control techniques for
expansion of fast charging infrastructure for electric vehicles. Energy Policy EV bidirectional wireless chargers. Electronics 2021;10(16):1905.
2021;156:112372. [37] Wang J, Shao Y, Ge Y, Yu R. A survey of vehicle to everything (V2X) testing.
[8] Neaimeh M, Salisbury SD, Hill GA, Blythe PT, Scoffield DR, Francfort JE. Sensors 2019;19(2):334.
Analysing the usage and evidencing the importance of fast chargers for the
[38] Pearre NS, Ribberink H. Review of research on V2X technologies, strategies,
adoption of battery electric vehicles. Energy Policy 2017;108:474–86.
and operations. Renew Sustain Energy Rev 2019;105:61–70.
[9] Burnham A, Dufek EJ, Stephens T, Francfort J, Michelbacher C, Carl-
[39] Afshar S, Macedo P, Mohamed F, Disfani V. Mobile charging stations for electric
son RB, Zhang J, Vijayagopal R, Dias F, Mohanpurkar M, et al. Enabling
vehicles—A review. Renew Sustain Energy Rev 2021;152:111654.
fast charging–Infrastructure and economic considerations. J Power Sources
[40] Deb S, Tammi K, Kalita K, Mahanta P. Review of recent trends in charging
2017;367:237–49.
infrastructure planning for electric vehicles. Wiley Interdiscip Rev: Energy
[10] Funke SÁ, Plötz P, Wietschel M. Invest in fast-charging infrastructure or in
Environ 2018;7(6):e306.
longer battery ranges? A cost-efficiency comparison for Germany. Appl Energy
[41] Metais M-O, Jouini O, Perez Y, Berrada J, Suomalainen E. Too much or not
2019;235:888–99.
enough? Planning electric vehicle charging infrastructure: A review of modeling
[11] Philipsen R, Schmidt T, Van Heek J, Ziefle M. Fast-charging station here,
options. Renew Sustain Energy Rev 2022;153:111719.
please! user criteria for electric vehicle fast-charging locations. Transp Res F
[42] Xu D, Tian Y. A comprehensive survey of clustering algorithms. Ann Data Sci
2016;40:119–29.
2015;2:165–93.
[12] Solanke TU, Khatua PK, Ramachandaramurthy VK, Yong JY, Tan KM. Control
[43] Andrenacci N, Ragona R, Valenti G. A demand-side approach to the optimal
and management of a multilevel electric vehicles infrastructure integrated with
deployment of electric vehicle charging stations in metropolitan areas. Appl
distributed resources: A comprehensive review. Renew Sustain Energy Rev
Energy 2016;182:39–46.
2021;144:111020.
[13] Das HS, Rahman MM, Li S, Tan C. Electric vehicles standards, charging [44] Kazemi MA, Sedighizadeh M, Mirzaei MJ, Homaee O. Optimal siting and
infrastructure, and impact on grid integration: A technological review. Renew sizing of distribution system operator owned EV parking lots. Appl Energy
Sustain Energy Rev 2020;120:109618. 2016;179:1176–84.
[14] Sadeghian O, Oshnoei A, Mohammadi-Ivatloo B, Vahidinasab V, Anvari- [45] Wang N, Wang C, Niu Y, Yang M, Yu Y. A two-stage charging facilities
Moghaddam A. A comprehensive review on electric vehicles smart charg- planning method for electric vehicle sharing systems. IEEE Trans Ind Appl
ing: Solutions, strategies, technologies, and challenges. J Energy Storage 2020;57(1):149–57.
2022;54:105241. [46] Marino CA, Marufuzzaman M. Unsupervised learning for deploying smart
[15] Rahman I, Vasant PM, Singh BSM, Abdullah-Al-Wadud M, Adnan N. Review of charging public infrastructure for electric vehicles in sprawling cities. J Clean
recent trends in optimization techniques for plug-in hybrid, and electric vehicle Prod 2020;266:121926.
charging infrastructures. Renew Sustain Energy Rev 2016;58:1039–47. [47] Bae S, Jang I, Gros S, Kulcsár B, Hellgren J. A game approach for charging
[16] Bilal M, Rizwan M. Electric vehicles in a smart grid: a comprehensive survey station placement based on user preferences and crowdedness. IEEE Trans Intell
on optimal location of charging station. IET Smart Grid 2020;3(3):267–79. Transp Syst 2020;23(4):3654–69.
[17] Ahmad F, Saad Alam M, Saad Alsaidan I, Shariff SM. Battery swapping [48] Liu J-p, Zhang T-x, Zhu J, Ma T-n. Allocation optimization of electric vehicle
station for electric vehicles: opportunities and challenges. IET Smart Grid charging station (EVCS) considering with charging satisfaction and distributed
2020;3(3):280–6. renewables integration. Energy 2018;164:560–74.
[18] Saharan S, Bawa S, Kumar N. Dynamic pricing techniques for Intelligent [49] Sánchez DG, Tabares A, Faria LT, Rivera JC, Franco JF. A clustering approach
Transportation System in smart cities: A systematic review. Comput Commun for the optimal siting of recharging stations in the electric vehicle routing
2020;150:603–25. problem with time windows. Energies 2022;15(7):2372.
[19] Dündar H, Ömürgönülşen M, Soysal M. A review on sustainable urban vehicle [50] Zeng M, Pan Y, Zhang D, Lu Z, Li Y. Data-driven location selection for battery
routing. J Clean Prod 2021;285:125444. swapping stations. IEEE Access 2019;7:133760–71.
[20] Shen Z-JM, Feng B, Mao C, Ran L. Optimization models for electric vehicle [51] Bitencourt L, Abud TP, Dias BH, Borba BS, Maciel RS, Quirós-Tortós J. Optimal
service operations: A literature review. Transp Res B 2019;128:462–77. location of EV charging stations in a neighborhood considering a multi-objective
[21] Elbouchikhi E, Zia MF, Benbouzid M, El Hani S. Overview of signal pro- approach. Electr Power Syst Res 2021;199:107391.
cessing and machine learning for smart grid condition monitoring. Electronics [52] Yao CZ, Luo J. Deployment of a charging network based on a cost-benefit model
2021;10(21):2725. and the CFSFDP algorithm. Kybernetes 2018.

14
N. Fescioglu-Unver and M. Yıldız Aktaş Renewable and Sustainable Energy Reviews 188 (2023) 113873

[53] Momtazpour M, Butler P, Ramakrishnan N, Hossain MS, Bozchalui MC, [82] Liang Y, Ding Z, Ding T, Lee W-J. Mobility-aware charging scheduling for shared
Sharma R. Charging and storage infrastructure design for electric vehicles. ACM on-demand electric vehicle fleet using deep reinforcement learning. IEEE Trans
Trans Intell Syst Technol 2014;5(3):1–27. Smart Grid 2020;12(2):1380–93.
[54] Wang C, Xie D, Gu C, Fan Y. Planning of regional urban bus charging [83] Vandael S, Claessens B, Ernst D, Holvoet T, Deconinck G. Reinforcement
facility: A case study of Fengxian, Shanghai. IEEE Trans Intell Transp Syst learning of heuristic EV fleet charging in a day-ahead electricity market. IEEE
2021;23(8):13592–603. Trans Smart Grid 2015;6(4):1795–805.
[55] Cilio L, Babacan O. Allocation optimisation of rapid charging stations in large [84] Arif A, Babar M, Ahamed TI, Al-Ammar E, Nguyen P, Kamphuis IR, Malik N.
urban areas to support fully electric taxi fleets. Appl Energy 2021;295:117072. Online scheduling of plug-in vehicles in dynamic pricing schemes. Sustain
[56] Ko YD. An efficient integration of the genetic algorithm and the reinforcement Energy Grids Netw 2016;7:25–36.
learning for optimal deployment of the wireless charging electric tram system. [85] Cao Y, Wang H, Li D, Zhang G. Smart online charging algorithm for elec-
Comput Ind Eng 2019;128:851–60. tric vehicles via customized actor–critic learning. IEEE Internet Things J
[57] Hu D, Zhang J, Liu Z-W. Charging stations expansion planning under govern- 2021;9(1):684–94.
ment policy driven based on Bayesian regularization backpropagation learning. [86] Cao Y, Li D, Zhang Y, Chen X. Joint optimization of delay-tolerant autonomous
Neurocomputing 2020;416:47–58. electric vehicles charge scheduling and station battery degradation. IEEE
[58] Jiang C, Jing Z, Ji T, Wu Q. Optimal location of PEVCSs using MAS and ER Internet Things J 2020;7(9):8590–9.
approach. IET Gener Transm Distrib 2018;12(20):4377–87. [87] Lee J, Lee E, Kim J. Electric vehicle charging and discharging algorithm
[59] Hecht C, Figgener J, Sauer DU. Predicting electric vehicle charging station based on reinforcement learning with data-driven approach in dynamic pricing
availability using ensemble machine learning. Energies 2021;14(23):7834. scheme. Energies 2020;13(8):1950.
[60] Ostermann A, Fabel Y, Ouan K, Koo H. Forecasting charging point occupancy [88] Wan Y, Qin J, Ma Q, Fu W, Wang S. Multi-agent DRL-based data-driven
using supervised learning algorithms. Energies 2022;15(9):3409. approach for PEVs charging/discharging scheduling in smart grid. J Franklin
[61] Friese PA, Michalk W, Fischer M, Hardt C, Bogenberger K. Charging point Inst B 2022;359(4):1747–67.
usage in germany—automated retrieval, analysis, and usage types explained. [89] Ye Z, Gao Y, Yu N. Learning to operate an electric vehicle charg-
Sustainability 2021;13(23):13046. ing station considering vehicle-grid integration. IEEE Trans Smart Grid
[62] Aduama P, Zhang Z, Al-Sumaiti AS. Multi-feature data fusion-based load 2022;13(4):3038–48.
forecasting of electric vehicle charging stations using a deep learning model. [90] Aljafari B, Jeyaraj PR, Kathiresan AC, Thanikanti SB. Electric vehicle optimum
Energies 2023;16(3):1309. charging-discharging scheduling with dynamic pricing employing multi agent
[63] Yi Z, Liu XC, Wei R, Chen X, Dai J. Electric vehicle charging demand forecasting deep neural network. Comput Electr Eng 2023;105:108555.
using deep learning model. J Intell Transp Syst 2022;26(6):690–703. [91] Wang R, Chen Z, Xing Q, Zhang Z, Zhang T. A modified rainbow-based deep
[64] Verma A, Asadi A, Yang K, Maitra A, Asgeirsson H. Analyzing household reinforcement learning method for optimal scheduling of charging station.
charging patterns of Plug-in electric vehicles (PEVs): A data mining approach. Sustainability 2022;14(3):1884.
Comput Ind Eng 2019;128:964–73. [92] Shin M, Choi D-H, Kim J. Cooperative management for PV/ESS-enabled electric
[65] Kim Y, Kim S. Forecasting charging demand of electric vehicles using time-series vehicle charging stations: A multiagent deep reinforcement learning approach.
models. Energies 2021;14(5):1487. IEEE Trans Ind Inf 2019;16(5):3493–503.
[66] Wang Q, Liu X, Du J, Kong F. Smart charging for electric vehicles: [93] Jin J, Xu Y. Optimal policy characterization enhanced actor-critic approach
A survey from the algorithmic perspective. IEEE Commun Surv Tutor for electric vehicle charging scheduling in a power distribution network. IEEE
2016;18(2):1500–17. Trans Smart Grid 2020;12(2):1416–28.
[67] Yang Z, Li K, Foley A. Computational scheduling methods for integrating plug- [94] Liu J, Lin G, Rehtanz C, Huang S, Zhou Y, Li Y. Data-driven intelligent EV
in electric vehicles with power systems: A review. Renew Sustain Energy Rev charging operating with limited chargers considering the charging demand
2015;51:396–416. forecasting. Int J Electr Power Energy Syst 2022;141:108218.
[68] Nimalsiri NI, Mediwaththe CP, Ratnam EL, Shaw M, Smith DB, Halgamuge SK. [95] Yan L, Chen X, Chen Y, Wen J. A cooperative charging control strategy for
A survey of algorithms for distributed charging control of electric vehicles in electric vehicles based on multiagent deep reinforcement learning. IEEE Trans
smart grid. IEEE Trans Intell Transp Syst 2019;21(11):4497–515. Ind Inf 2022;18(12):8765–75.
[69] Khalid MR, Khan IA, Hameed S, Asghar MSJ, Ro J-S. A comprehensive review [96] Ming F, Gao F, Liu K, Li X. A constrained DRL-based bi-level coordinated
on structural topologies, power levels, energy storage systems, and standards method for large-scale EVs charging. Appl Energy 2023;331:120381.
for electric vehicle charging stations and their impacts on grid. IEEE Access [97] Tao Y, Qiu J, Lai S. Deep reinforcement learning based bidding strategy for
2021;9:128069–94. EVAs in local energy market considering information asymmetry. IEEE Trans
[70] Shaukat N, Khan B, Ali S, Mehmood C, Khan J, Farid U, Majid M, Anwar S, Ind Inf 2021;18(6):3831–42.
Jawad M, Ullah Z. A survey on electric vehicle transportation within smart grid [98] Lin J, Dong P, Liu M, Huang X, Deng W. Research on demand response of
system. Renew Sustain Energy Rev 2018;81:1329–49. electric vehicle agents based on multi-layer machine learning algorithm. IEEE
[71] Gonçalves F, de Abreu Borges L, Batista R. Electric vehicle charging data Access 2020;8:224224–34.
analytics of corporate fleets. World Electr Veh J 2022;13(12):237. [99] Asadi A, Pinkley SN. A stochastic scheduling, allocation, and inven-
[72] Li H, Wan Z, He H. Constrained EV charging scheduling based on safe deep tory replenishment problem for battery swap stations. Transp Res E
reinforcement learning. IEEE Trans Smart Grid 2019;11(3):2427–39. 2021;146:102212.
[73] Yan L, Chen X, Zhou J, Chen Y, Wen J. Deep reinforcement learning for [100] Gao Y, Yang J, Yang M, Li Z. Deep reinforcement learning based optimal
continuous electric vehicles charging control with dynamic user behaviors. IEEE schedule for a battery swapping station considering uncertainties. IEEE Trans
Trans Smart Grid 2021;12(6):5124–34. Ind Appl 2020;56(5):5775–84.
[74] Li S, Hu W, Cao D, Dragičević T, Huang Q, Chen Z, Blaabjerg F. Electric vehicle [101] Wang X, Wang J, Liu J. Vehicle to grid frequency regulation capacity optimal
charging management based on deep reinforcement learning. J Mod Power Syst scheduling for battery swapping station using deep Q-network. IEEE Trans Ind
Clean Energy 2021;10(3):719–30. Inf 2020;17(2):1342–51.
[75] Chang F, Chen T, Su W, Alsafasfeh Q. Control of battery charging based on [102] Liang Y, Ding Z, Zhao T, Lee W-J. Real-time operation management for battery
reinforcement learning and long short-term memory networks. Comput Electr swapping-charging system via multi-agent deep reinforcement learning. IEEE
Eng 2020;85:106670. Trans Smart Grid 2022;14(1):559–71.
[76] Wang F, Gao J, Li M, Zhao L. Autonomous PEV charging scheduling using [103] Tan M, Dai Z, Su Y, Chen C, Wang L, Chen J. Bi-level optimization of charging
Dyna-Q reinforcement learning. IEEE Trans Veh Technol 2020;69(11):12609– scheduling of a battery swap station based on deep reinforcement learning. Eng
20. Appl Artif Intell 2023;118:105557.
[77] Zhang F, Yang Q, An D. CDDPG: A deep-reinforcement-learning-based approach [104] Jin J, Mao S, Xu Y. Optimal priority rule enhanced deep reinforcement learning
for electric vehicle charging control. IEEE Internet Things J 2020;8(5):3075–87. for charging scheduling in an electric vehicle battery swapping station. IEEE
[78] Wan Z, Li H, He H, Prokhorov D. Model-free real-time EV charging Trans Smart Grid 2023.
scheduling based on deep reinforcement learning. IEEE Trans Smart Grid [105] Limmer S. Dynamic pricing for electric vehicle charging—A literature review.
2018;10(5):5246–57. Energies 2019;12(18):3574.
[79] Zhang Z, Wan Y, Qin J, Fu W, Kang Y. A deep rl-based algorithm for [106] Wang S, Bi S, Zhang YA. Reinforcement learning for real-time pric-
coordinated charging of electric vehicles. IEEE Trans Intell Transp Syst ing and scheduling control in EV charging stations. IEEE Trans Ind Inf
2022;23(10):18774–84. 2021;17(2):849–59.
[80] Chiş A, Lundén J, Koivunen V. Reinforcement learning-based plug-in elec- [107] Paraskevas A, Aletras D, Chrysopoulos A, Marinopoulos A, Doukas DI. Opti-
tric vehicle charging with forecasted price. IEEE Trans Veh Technol mal management for EV charging stations: A win–win strategy for different
2016;66(5):3674–84. stakeholders using constrained deep Q-learning. Energies 2022;15(7):2323.
[81] López KL, Gagné C, Gardner M-A. Demand-side management using deep [108] Moghaddam V, Yazdani A, Wang H, Parlevliet D, Shahnia F. An online rein-
learning for smart charging of electric vehicles. IEEE Trans Smart Grid forcement learning approach for dynamic pricing of electric vehicle charging
2018;10(3):2683–91. stations. IEEE Access 2020;8:130305–13.

15
N. Fescioglu-Unver and M. Yıldız Aktaş Renewable and Sustainable Energy Reviews 188 (2023) 113873

[109] Fang C, Lu H, Hong Y, Liu S, Chang J. Dynamic pricing for electric vehicle [133] Huang C-J, Hu K-W, Ho H-Y, Chuang H-W. Congestion-preventing routing and
extreme fast charging. IEEE Trans Intell Transp Syst 2020;22(1):531–41. charging scheduling mechanism for electric vehicles in dense urban areas. Inf
[110] Qian T, Shao C, Li X, Wang X, Chen Z, Shahidehpour M. Multi-agent deep Technol Control 2021;50(2):284–307.
reinforcement learning method for EV charging station game. IEEE Trans Power [134] Huang C-J, Hu K-W, Ho HY, Xie BZ, Feng C-C, Chuang H-W. A distributed urban
Syst 2021;37(3):1682–94. traffic congestion prevention mechanism for mixed flow of human-driven and
[111] Abdalrahman A, Zhuang W. Dynamic pricing for differentiated PEV charging autonomous electric vehicles. Int J Comput Intell Syst 2021;14(1):1714–27.
services using deep reinforcement learning. IEEE Trans Intell Transp Syst [135] Basso R, Kulcsár B, Sanchez-Diaz I. Electric vehicle routing problem with
2020;23(2):1415–27. machine learning for energy prediction. Transp Res B 2021;145:24–55.
[112] Zhao Z, Lee CK. Dynamic pricing for EV charging stations: A deep reinforcement [136] Lu Y, Gu J, Xie D, Li Y. Integrated route planning algorithm based on spot price
learning approach. IEEE Trans Transp Electrif 2021;8(2):2456–68. and classified travel objectives for EV users. IEEE Access 2019;7:122238–50.
[113] Zhang Y, Zhang Z, Yang Q, An D, Li D, Li C. EV charging bidding by multi- [137] Zhang Q, Wu K, Shi Y. Route planning and power management for PHEVs with
DQN reinforcement learning in electricity auction market. Neurocomputing reinforcement learning. IEEE Trans Veh Technol 2020;69(5):4751–62.
2020;397:404–14. [138] Aljohani TM, Ebrahim A, Mohammed O. Real-time metadata-driven routing
[114] Kim B-G, Zhang Y, Van Der Schaar M, Lee J-W. Dynamic pricing and energy optimization for electric vehicle energy consumption minimization using deep
consumption scheduling with reinforcement learning. IEEE Trans Smart Grid reinforcement learning and Markov chain model. Electr Power Syst Res
2015;7(5):2187–98. 2021;192:106962.
[115] Qiu D, Ye Y, Papadaskalopoulos D, Strbac G. A deep reinforcement learning [139] Aljohani TM, Mohammed O. A real-time energy consumption minimiza-
method for pricing electric vehicles with discrete charging levels. IEEE Trans tion framework for electric vehicles routing optimization based on SARSA
Ind Appl 2020;56(5):5901–12. reinforcement learning. Vehicles 2022;4(4):1176–94.
[116] Liu D, Wang W, Wang L, Jia H, Shi M. Dynamic pricing strategy of electric [140] Dorokhova M, Ballif C, Wyrsch N. Routing of electric vehicles with interme-
vehicle aggregators based on DDPG reinforcement learning algorithm. IEEE diary charging stations: A reinforcement learning approach. Front Big Data
Access 2021;9:21556–66. 2021;4:586481.
[117] Kucukoglu I, Dewil R, Cattrysse D. The electric vehicle routing problem and its [141] Bi H, Shang W-L, Chen Y, Wang K, Yu Q, Sui Y. GIS aided sustainable urban
variations: A literature review. Comput Ind Eng 2021;161:107650. road management with a unifying queueing and neural network model. Appl
[118] Xiao Y, Zhang Y, Kaku I, Kang R, Pan X. Electric vehicle routing problem: Energy 2021;291:116818.
A systematic review and a new comprehensive model with nonlinear energy [142] Bozorgi AM, Farasat M, Mahmoud A. A time and energy efficient routing
recharging and consumption. Renew Sustain Energy Rev 2021;151:111567. algorithm for electric vehicles based on historical driving data. IEEE Trans Intell
[119] James J, Yu W, Gu J. Online vehicle routing with neural combinatorial Veh 2017;2(4):308–20.
optimization and deep reinforcement learning. IEEE Trans Intell Transp Syst [143] Ayman A, Sivagnanam A, Wilbur M, Pugliese P, Dubey A, Laszka A. Data-driven
2019;20(10):3806–17. prediction and optimization of energy use for transit fleets of electric and ICE
[120] Basso R, Kulcsár B, Sanchez-Diaz I, Qu X. Dynamic stochastic electric vehicle vehicles. ACM Trans Internet Technol (TOIT) 2021;22(1):1–29.
routing with safe reinforcement learning. Transp Res E 2022;157:102496. [144] Di Martino A, Miraftabzadeh SM, Longo M. Strategies for the modelisation of
[121] Shi J, Gao Y, Wang W, Yu N, Ioannou PA. Operating electric vehicle fleet for electric vehicle energy consumption: A review. Energies 2022;15(21):8115.
ride-hailing services with reinforcement learning. IEEE Trans Intell Transp Syst [145] Varga BO, Sagoian A, Mariasiu F. Prediction of electric vehicle range:
2019;21(11):4822–34. A comprehensive review of current issues and challenges. Energies
[122] Kullman ND, Cousineau M, Goodson JC, Mendoza JE. Dynamic ride-hailing with 2019;12(5):946.
electric vehicles. Transp Sci 2022;56(3):775–94. [146] Bengio Y, Lodi A, Prouvost A. Machine learning for combinatorial optimization:
[123] Lin B, Ghaddar B, Nathwani J. Deep reinforcement learning for the electric a methodological tour d’horizon. European J Oper Res 2021;290(2):405–21.
vehicle routing problem with time windows. IEEE Trans Intell Transp Syst [147] Taiebat M, Xu M. Synergies of four emerging technologies for accelerated
2021;23(8):11528–38. adoption of electric vehicles: Shared mobility, wireless charging, vehicle-to-grid,
[124] Zhang L, Gong K, Xu M. Congestion control in charging stations allocation with and vehicle automation. J Clean Prod 2019;230:794–7.
Q-learning. Sustainability 2019;11(14):3900. [148] Adderly SA, Manukian D, Sullivan TD, Son M. Electric vehicles and natural
[125] Qian T, Shao C, Wang X, Shahidehpour M. Deep reinforcement learning for EV disaster policy implications. Energy Policy 2018;112:437–48.
charging navigation by coordinating smart grid and intelligent transportation [149] Hussain A, Musilek P. Resilience enhancement strategies for and through
system. IEEE Trans Smart Grid 2019;11(2):1714–23. electric vehicles. Sustainable Cities Soc 2022;80:103788.
[126] Lee K-B, A. Ahmed M, Kang D-K, Kim Y-C. Deep reinforcement learning based [150] Wang H, Fang Y-P, Zio E. Resilience-oriented optimal post-disruption
optimal route and charging station selection. Energies 2020;13(23):6255. reconfiguration for coupled traffic-power systems. Reliab Eng Syst Saf
[127] Xu P, Zhang J, Gao T, Chen S, Wang X, Jiang H, Gao W. Real-time fast charging 2022;222:108408.
station recommendation for electric vehicles in coupled power-transportation [151] Ebadat-Parast M, Nazari MH, Hosseinian SH. Distribution system resilience
networks: A graph reinforcement learning method. Int J Electr Power Energy enhancement through resilience-oriented optimal scheduling of multi-microgrids
Syst 2022;141:108030. considering normal and emergency conditions interlink utilizing multi-objective
[128] Tao Y, Qiu J, Lai S, Sun X, Zhao J, Zhou B, Cheng L. Data-driven programming. Sustainable Cities Soc 2022;76:103467.
on-demand energy supplement planning for electric vehicles considering [152] MacDonald CD, Kattan L, Layzell D. Modelling electric vehicle charging network
multi-charging/swapping services. Appl Energy 2022;311:118632. capacity and performance during short-notice evacuations. Int J Disaster Risk
[129] Suanpang P, Jamjuntr P, Kaewyong P, Niamsorn C, Jermsittiparsert K. An intel- Reduct 2021;56:102093.
ligent recommendation for intelligently accessible charging stations: Electronic [153] Hussain A, Kim H-M. EV prioritization and power allocation during outages: A
vehicle charging to support a sustainable smart tourism city. Sustainability lexicographic method-based multiobjective optimization approach. IEEE Trans
2023;15(1):455. Transp Electrif 2021;7(4):2474–87.
[130] Jin J, Xu Y. Shortest-path-based deep reinforcement learning for EV charging [154] Li Q, Soleimaniamiri S, Li X. Optimal mass evacuation planning for electric
routing under stochastic traffic condition and electricity prices. IEEE Internet vehicles before natural disasters. Transp Res D 2022;107:103292.
Things J 2022;9(22):22571–81. [155] Zhang J, Zhang X. A multi-trip electric bus routing model considering equity
[131] Thibault L, De Nunzio G, Sciarretta A. A unified approach for electric vehicles during short-notice evacuations. Transp Res D 2022;110:103397.
range maximization via eco-routing, eco-driving, and energy consumption [156] Kakillioglu EA, Aktaş MY, Fescioglu-Unver N. Self-controlling resource man-
prediction. IEEE Trans Intell Veh 2018;3(4):463–75. agement model for electric vehicle fast charging stations with priority service.
[132] Shukla RM, Sengupta S. Cop: An integrated communication, optimization, and Energy 2022;239:122276.
prediction unit for smart plug-in electric vehicle charging. Internet Things [157] Fescioglu-Unver N, Aktaş MY, Kasnakoğlu C. Feedback controlled resource
2020;9:100148. management model for express service in electric vehicle charging stations.
J Clean Prod 2021;311:127629.

16

You might also like