ming zhu , xiao-yang liu , and xiaodong wang

15
IEEE TRANSACTIONS ON INTELLIGENT TRANSPORTATION SYSTEMS, VOL. XX, NO. XX, XXX 20XX 1 Deep Reinforcement Learning for Unmanned Aerial Vehicle-Assisted Vehicular Networks Ming Zhu * , Xiao-Yang Liu * , and Xiaodong Wang Abstract—Unmanned aerial vehicles (UAVs) are envisioned to complement the 5G communication infrastructure in future smart cities. Hot spots easily appear in road intersections, where effective communication among vehicles is challenging. UAVs may serve as relays with the advantages of low price, easy deployment, line-of-sight links, and flexible mobility. In this paper, we study a UAV-assisted vehicular network where the UAV jointly adjusts its transmission control (power and channel) and 3D flight to maximize the total throughput. First, we formulate a Markov decision process (MDP) problem by modeling the mobility of the UAV/vehicles and the state transitions. Secondly, we solve the tar- get problem using a deep reinforcement learning method under unknown or unmeasurable environment variables especially in 5G, namely, the deep deterministic policy gradient (DDPG), and propose three solutions with different control objectives. Envi- ronment variables are unknown and unmeasurable, therefore, we use a deep reinforcement learning method. Moreover, considering the energy consumption of 3D flight, we extend the proposed solutions to maximize the total throughput per energy unit by encouraging or discouraging the UAV’s mobility. To achieve this goal, the DDPG framework is modified. Thirdly, in a simplified model with small state space and action space, we verify the optimality of proposed algorithms. Comparing with two baseline schemes, we demonstrate the effectiveness of proposed algorithms in a realistic model. Index Terms—Unmanned aerial vehicle, vehicular networks, smart cities, Markov decision process, deep reinforcement learn- ing, power control, channel control. I. I NTRODUCTION Intelligent transportation system [1] [2] [3] [4] is a key component of smart cities, which employs real-time data communication for traffic monitoring, path planning, entertain- ment and advertisement [5]. High speed vehicular networks [6] emerge as a key component of intelligent transportation systems that provide cooperative communications to improve data transmission performance. With the increasing number of vehicles, the current com- munication infrastructure may not satisfy data transmission re- quirements, especially when hot spots (e.g., road intersections) appear during rush hours. Unmanned aerial vehicles (UAVs) or drones [7] can complement the 4G/5G communication infrastructure, including vehicle-to-vehicle (V2V) communi- cations, and vehicle-to-infrastructure (V2I) communications. * Equal contribution. M. Zhu is with the Shenzhen Institutes of Advanced Technol- ogy, Chinese Academy of Sciences, Shenzhen 518055, China. E-mail: [email protected]. X.-Y. Liu and X. Wang are with the Department of Electrical Engineer- ing, Columbia University, New York, NY 10027, USA E-mail: {xiaoyang, wangx}@ee.columbia.edu. Downlink Base station Fig. 1. The scenario of a UAV-assisted vehicular network. Qualcomm has received a certification of authorization allow- ing for UAV testing below 400 feet [8]; Huawei will cooperate with China Mobile to build the first cellular test network for regional logistics UAVs [9]. A UAV-assisted vehicular network in Fig. 1 has several advantages. First, the path loss will be much lower since the UAV can move nearer to vehicles compared with stationary base stations. Secondly, the UAV is flexible in adjusting the transmission control [10] based on the mobility of vehicles. Thirdly, the quality of UAV-to-vehicle links is generally better than that of terrestrial links [11], since they are mostly line- of-sight (LoS). Maximizing the total throughput of UAV-to-vehicle links has several challenges. First, the communication channels vary with the UAV’s three-dimensional (3D) positions. Secondly, the joint adjustment of the UAV’s 3D flight and transmission control (e.g., power control) cannot be solved directly using conventional optimization methods, especially when the envi- ronment variables are unknown and unmeasurable. Thirdly, the channel conditions are hard to acquire, e.g., the path loss from the UAV to vehicles is closely related to the height/density of buildings and street width. In this paper, we propose deep reinforcement learning [12] based algorithms to maximize the total throughput of UAV-to- vehicle communications, which jointly adjusts the UAV’s 3D flight and transmission control by learning through interacting with the environment. The main contributions of this paper can be summarized as follows: 1) We formulate the problem as a Markov decision process (MDP) problem to maximize the to- tal throughput with the constraints of total transmission power and total channel; 2) We apply a deep reinforcement learning method, the deep deterministic policy gradient (DDPG), to arXiv:1906.05015v10 [cs.LG] 4 Mar 2020

Upload: others

Post on 14-May-2022

14 views

Category:

Documents


0 download

TRANSCRIPT

Page 1: Ming Zhu , Xiao-Yang Liu , and Xiaodong Wang

IEEE TRANSACTIONS ON INTELLIGENT TRANSPORTATION SYSTEMS, VOL. XX, NO. XX, XXX 20XX 1

Deep Reinforcement Learning for Unmanned AerialVehicle-Assisted Vehicular Networks

Ming Zhu∗, Xiao-Yang Liu∗, and Xiaodong Wang

Abstract—Unmanned aerial vehicles (UAVs) are envisionedto complement the 5G communication infrastructure in futuresmart cities. Hot spots easily appear in road intersections, whereeffective communication among vehicles is challenging. UAVs mayserve as relays with the advantages of low price, easy deployment,line-of-sight links, and flexible mobility. In this paper, we studya UAV-assisted vehicular network where the UAV jointly adjustsits transmission control (power and channel) and 3D flight tomaximize the total throughput. First, we formulate a Markovdecision process (MDP) problem by modeling the mobility of theUAV/vehicles and the state transitions. Secondly, we solve the tar-get problem using a deep reinforcement learning method underunknown or unmeasurable environment variables especially in5G, namely, the deep deterministic policy gradient (DDPG), andpropose three solutions with different control objectives. Envi-ronment variables are unknown and unmeasurable, therefore, weuse a deep reinforcement learning method. Moreover, consideringthe energy consumption of 3D flight, we extend the proposedsolutions to maximize the total throughput per energy unit byencouraging or discouraging the UAV’s mobility. To achieve thisgoal, the DDPG framework is modified. Thirdly, in a simplifiedmodel with small state space and action space, we verify theoptimality of proposed algorithms. Comparing with two baselineschemes, we demonstrate the effectiveness of proposed algorithmsin a realistic model.

Index Terms—Unmanned aerial vehicle, vehicular networks,smart cities, Markov decision process, deep reinforcement learn-ing, power control, channel control.

I. INTRODUCTION

Intelligent transportation system [1] [2] [3] [4] is a keycomponent of smart cities, which employs real-time datacommunication for traffic monitoring, path planning, entertain-ment and advertisement [5]. High speed vehicular networks[6] emerge as a key component of intelligent transportationsystems that provide cooperative communications to improvedata transmission performance.

With the increasing number of vehicles, the current com-munication infrastructure may not satisfy data transmission re-quirements, especially when hot spots (e.g., road intersections)appear during rush hours. Unmanned aerial vehicles (UAVs)or drones [7] can complement the 4G/5G communicationinfrastructure, including vehicle-to-vehicle (V2V) communi-cations, and vehicle-to-infrastructure (V2I) communications.

∗Equal contribution.M. Zhu is with the Shenzhen Institutes of Advanced Technol-

ogy, Chinese Academy of Sciences, Shenzhen 518055, China. E-mail:[email protected].

X.-Y. Liu and X. Wang are with the Department of Electrical Engineer-ing, Columbia University, New York, NY 10027, USA E-mail: xiaoyang,[email protected].

Downlink

Base station

Fig. 1. The scenario of a UAV-assisted vehicular network.

Qualcomm has received a certification of authorization allow-ing for UAV testing below 400 feet [8]; Huawei will cooperatewith China Mobile to build the first cellular test network forregional logistics UAVs [9].

A UAV-assisted vehicular network in Fig. 1 has severaladvantages. First, the path loss will be much lower since theUAV can move nearer to vehicles compared with stationarybase stations. Secondly, the UAV is flexible in adjusting thetransmission control [10] based on the mobility of vehicles.Thirdly, the quality of UAV-to-vehicle links is generally betterthan that of terrestrial links [11], since they are mostly line-of-sight (LoS).

Maximizing the total throughput of UAV-to-vehicle linkshas several challenges. First, the communication channels varywith the UAV’s three-dimensional (3D) positions. Secondly,the joint adjustment of the UAV’s 3D flight and transmissioncontrol (e.g., power control) cannot be solved directly usingconventional optimization methods, especially when the envi-ronment variables are unknown and unmeasurable. Thirdly, thechannel conditions are hard to acquire, e.g., the path loss fromthe UAV to vehicles is closely related to the height/density ofbuildings and street width.

In this paper, we propose deep reinforcement learning [12]based algorithms to maximize the total throughput of UAV-to-vehicle communications, which jointly adjusts the UAV’s 3Dflight and transmission control by learning through interactingwith the environment. The main contributions of this paper canbe summarized as follows: 1) We formulate the problem as aMarkov decision process (MDP) problem to maximize the to-tal throughput with the constraints of total transmission powerand total channel; 2) We apply a deep reinforcement learningmethod, the deep deterministic policy gradient (DDPG), to

arX

iv:1

906.

0501

5v10

[cs

.LG

] 4

Mar

202

0

Page 2: Ming Zhu , Xiao-Yang Liu , and Xiaodong Wang

IEEE TRANSACTIONS ON INTELLIGENT TRANSPORTATION SYSTEMS, VOL. XX, NO. XX, XXX 20XX 2

solve the problem. DDPG is suitable to solve MDP problemswith continuous states and actions. We propose three solutionswith different control objectives to jointly adjust the UAV’s 3Dflight and transmission control. Then we extend the proposedsolutions to maximize the total throughput per energy unit.To encourage or discourage the UAV’s mobility, we modifythe reward function and the DDPG framework; 3) We verifythe optimality of proposed solutions using a simplified modelwith small state space and action space. Finally, we provideextensive simulation results to demonstrate the effectiveness ofthe proposed solutions compared with two baseline schemes.

The remainder of the paper is organized as follows. SectionII discusses related works. Section III presents system modelsand problem formulation. Solutions are proposed in SectionIV. Section V presents the performance evaluation. SectionVI concludes this paper.

II. RELATED WORKS

The dynamic control for the UAV-assisted vehicular net-works includes flight control and transmission control. Flightcontrol mainly includes the planning of flight path, time, anddirection. Yang et al. [13] proposed a joint genetic algorithmand ant colony optimization method to obtain the best UAVflight paths to collect sensory data in wireless sensor networks.To further minimize the UAVs’ travel duration under certainconstraints (e.g., energy limitations, fairness, and collision),Garraffa et al. [14] proposed a two-dimensional (2D) pathplanning method based on a column generation approach. Liuet al. [15] proposed a deep reinforcement learning approachto control a group of UAVs by optimizing the flying directionsand distances to achieve the best communication coverage inthe long run with limited energy consumption.

The transmission control of UAVs mainly concerns resourceallocations, e.g., access selection, transmission power andbandwidth/channel allocation. Wang et al. [16] presented apower allocation strategy for UAVs considering communica-tions, caching, and energy transfer. In a UAV-assisted commu-nication network, Yan et al. [17] studied a UAV access selec-tion and base station bandwidth allocation problem, where theinteraction among UAVs and base stations was modeled as aStackelberg game, and the uniqueness of a Nash equilibriumwas obtained.

Joint control of both UAVs’ flight and transmission hasalso be considered. Wu et al. [18] considered maximizing theminimum achievable rates from a UAV to ground users byjointly optimizing the UAV’s 2D trajectory and power alloca-tion. Zeng et al. [19] proposed a convex optimization methodto optimize the UAV’s 2D trajectory to minimize its missioncompletion time while ensuring each ground terminal recoversthe file with high probability when the UAV disseminates acommon file to them. Zhang et al. [20] considered the UAVmission completion time minimization by optimizing its 2Dtrajectory with a constraint on the connectivity quality frombase stations to the UAV. However, most existing researchworks neglected adjusting UAVs’ height to obtain better qual-ity of links by avoiding various obstructions or non-line-of-sight (NLoS) links.

4

1

2

3

0Light 2

Light 1

Flow 1

Flow 2 Flow 2

Flow 1

Fig. 2. A one-way-two-flow road intersection.

Fan et al. [21] optimized the UAV’s 3D flight and transmis-sion control together; however, the 3D position optimizationwas converted to a 2D position optimization by the LoS linkrequirement. The existing deep reinforcement learning basedmethodd only handle UAVs’ 2D flight and simple transmissioncontrol decisions. For example, Challita et al. [22] proposed adeep reinforcement learning based method for a cellular UAVnetwork by optimizing the 2D path and cell association toachieve a tradeoff between maximizing energy efficiency andminimizing both wireless latency and the interference on thepath. A similar scheme is applied to provide intelligent trafficlight control in [23].

In addition, most existing works assumed that the groundterminals are stationary; whereas in reality, some groundterminals move with certain patterns, e.g., vehicles move underthe control of traffic lights. This work studies a UAV-assistedvehicular network where the UAV’s 3D flight and transmissioncontrol can be jointly adjusted, considering the mobility ofvehicles in a road intersection.

III. SYSTEM MODELS AND PROBLEM FORMULATION

In this section, we first describe the traffic model and com-munication model, and then formulate the target problem as aMarkov decision process. The variables in the communicationmodel are listed in Table I for easy reference.

A. Traffic Model

We start with a one-way-two-flow road intersection, asshown in Fig. 2, while a much more complicated scenarioin Fig. 7 will be described in Section V-B. Five blocks arenumbered as 0, 1, 2, 3, and 4, where block 0 is the intersection.We assume that each block contains at most one vehicle,indicated by binary variables n = (n0, ..., n4) ∈ 0, 1. Thereare two traffic flows in Fig. 2,• “Flow 1”: 1→ 0→ 3;• “Flow 2”: 2→ 0→ 4.The traffic light L has four configurations:• L=0: red light for flow 1 and green light for flow 2;• L=1: red light for flow 1 and yellow light for flow 2;• L=2: green light for flow 1 and red light for flow 2;

Page 3: Ming Zhu , Xiao-Yang Liu , and Xiaodong Wang

IEEE TRANSACTIONS ON INTELLIGENT TRANSPORTATION SYSTEMS, VOL. XX, NO. XX, XXX 20XX 3

TABLE IVARIABLES IN COMMUNICATION MODEL

hit, Hit channel power gain and channel state from the UAV

to a vehicle in block i in time slot t.ψit SINR from the UAV to a vehicle in block i in time slot t.dit, D

it horizontal distance and Euclidean distance between

the UAV and a vehicle in block i.P,C, b total transmission power, total number of channels, and band-

width of each channel.ρit, c

it transmission power and number of channels allocated for the

vehicle in block i in time slot t.

L=0 L=0

Time slots:

L=0 L=2 L=2 L=2

...

...

Traffic light state transitions

0 1 N-1 N+1 N+2 2N... ...

... ...

N

L=1 L=3

2N+1Flow 2:

Flow 1: 0 1 N-1 N N+1 N+2 2N 2N+1

Fig. 3. Traffic light states along time.

• L=3: yellow light for flow 1 and red light for flow 2.Time is partitioned into slots with equal duration. The durationof a green or red light occupies N time slots, and the durationof a yellow light occupies a time slot, which are shown inFig. 3. We assume that each vehicle moves one block in atime slot if the traffic light is green.

B. Communication Model

We focus on the downlink communications (UAV-to-vehicle), since they are directly controlled by the UAV. Thereare two channel states of each UAV-to-vehicle link, line-of-sight (LoS) and non-line-of-sight (NLoS). Let x and zdenote the block (horizontal position) and height of the UAVrespectively, where x ∈ 0, 1, 2, 3, 4 corresponds to these fiveblocks in Fig. 2, and z is discretized to multiple values. Weassume that the UAV stays above the five blocks since theUAV trends to get nearer to vehicles. Next, we describe thecommunication model, including the channel power gain, thesignal to interference and noise ratio (SINR), and the totalthroughput.

First, the channel power gain between the UAV and avehicle in block i in time slot t is hit with a channel stateHit ∈ NLoS,LoS. hi is formulated as [10] [24]

hit =

(Di

t)−β1 , if Hi

t = LoS,β2(Di

t)−β1 , if Hi

t = NLoS,(1)

where Dit is the Euclidean distance between the UAV and the

vehicle in block i in time slot t, β1 is the path loss exponent,and β2 is an additional attenuation factor caused by NLoSconnections.

The probabilities of LoS and NLoS links between the UAVand a vehicle in block i in time slot t are [25]

p(Hit =LoS)=

1

1 + α1 exp(−α2( 180π arctan z

dit− α1))

, (2)

p(Hit =NLoS)=1− p(Hi

t =LoS), i ∈ 0, 1, 2, 3, 4, (3)

where α1 and α2 are system parameters depending on theenvironment (height/density of buildings, and street width,etc.). We assume that α1, α2, β1, and β2 have fixed valuesamong all blocks in an intersection. dit is the horizontaldistance in time slot t. The angle 180

π arctan zdit

is measuredin “degrees” with the range 0 ∼ 90. Both dit and zt arediscrete variables, therefore, Di

t =√

(dit)2 + z2

t is also adiscrete variable.

Secondly, the SINR ψit in time slot t from the UAV to avehicle in block i is characterized as [26]

ψit =ρith

it

bcitσ2, i ∈ 0, 1, 2, 3, 4, (4)

where b is the equal bandwidth of each channel, ρit and cit arethe allocated transmission power and number of channels forthe vehicle in block i in time slot t, respectively, and σ2 isthe additive white Gaussian noise (AWGN) power spectrumdensity, and hi is formulated by (1). We assume that theUAV employs orthogonal frequency division multiple access(OFDMA) [27]; therefore, there is no interference among thesechannels.

Thirdly, the total throughput (reward) of UAV-to-vehiclelinks is formulated as [28]∑i∈0,1,2,3,4

bcit log(1+ψit)=∑

i∈0,1,2,3,4

bcit log(1+ρith

it

bcitσ2

). (5)

C. MDP Formulation

The UAV aims to maximize the total throughput with theconstraints of total transmission power and total channels:∑

i∈0,1,2,3,4

ρit ≤ P,∑

i∈0,1,2,3,4

cit ≤ C,

0 ≤ ρit ≤ ρmax, 0 ≤ cit ≤ cmax, i ∈ 0, 1, 2, 3, 4,

where P is the total transmission power, C is the total numberof channels, ρmax is the maximum power allocated to a vehicle,cmax is the maximum number of channels allocated to avehicle, ρit is a discrete variable, and cit is a nonnegative integervariable.

The UAV-assisted communication is modeled as a Markovdecision process (MDP). On one hand, from (2) and (3), weknow that the channel state of UAV-to-vehicle links follows astochastic process. On the other hand, the arrival of vehiclesfollows a stochastic process under the control of the trafficlight, e.g., (12) and (13).

Under the MDP framework, the state space S, actionspace A, reward r, policy π, and state transition probabilityp(st+1|st, at) of our problem are defined as follows.• State S = (L, x, z,n,H), where L is the traffic

light state, (x, z) is the UAV’s 3D position with x ∈0, 1, 2, 3, 4 being the block and z being the height, andH = (H0, ...,H4) is the channel state from the UAV toeach block i ∈ 0, 1, 2, 3, 4 with Hi ∈ NLoS,LoS.Let z ∈ [zmin, zmax], where zmin and zmax are the UAV’sminimum and maximum height, respectively. The blockx is the location projected from UAV’s 3D position tothe road.

Page 4: Ming Zhu , Xiao-Yang Liu , and Xiaodong Wang

IEEE TRANSACTIONS ON INTELLIGENT TRANSPORTATION SYSTEMS, VOL. XX, NO. XX, XXX 20XX 4

S3

S2 S4

S1

0

0

0

0

S0

6

6

1

3

Fig. 4. The position state transition diagram when the UAV’s height is fixed.

• Action A = (f ,ρ, c) denotes the action set. fx de-notes the horizontal flight, and fz denotes the verticalflight, both of which constitute the UAV’s 3D flightf = (fx, fz). With respect to horizontal flight, weassume that the UAV can hover or flight to its adjacentblock in a time slot, thus fx ∈ 0, 1, ..., 7 in Fig. 4.With respect to vertical flight, we assume

fz ∈ −5, 0, 5, (6)

which means that the UAV can flight downward 5 meters,horizontally, and up 5 meters in a time slot. The UAV’sheight changes as

zt+1 = fzt + zt. (7)

ρ = (ρ0t , ..., ρ

4t ) and c = (c0t , ..., c

4t ) are the transmis-

sion power and channel allocation actions for those fiveblocks, respectively. At the end of time slot t, the UAVmoves to a new 3D position according to action f , andover time slot t, the transmission power and number ofchannels are ρ and c, respectively.

• Reward r(st, at) =∑i∈0,1,2,3,4 bn

itcit log(1 +

ρithit

bcitσ2 )

is the total throughput after a transition from state stto st+1 taking action at. Note that the total throughputover the t-th time slot is measured at the state st =(Lt, xt, zt,nt,Ht).

• Policy π is the strategy for the UAV, which maps states toa probability distribution over the actions π : S → P(A),where P(·) denotes probability distribution. In time slot t,the UAV’s state is st = (Lt, xt, zt,nt,Ht), and its policyπt outputs the probability distribution over the action at.We see that the policy indicates the action preference ofthe UAV.

• State transition probability p(st+1|st, at) formulated in(8) is the probability of the UAV entering the new statest+1, after taking the action at at the current state st.At the current state st = (Lt, xt, zt,nt,Ht), after takingthe 3D flight and transmission control at = (f ,ρ, c), theUAV moves to the new 3D position (xt+1, zt+1), andthe channel state changes to Ht+1, with the traffic light

changes to Lt+1 and the number of vehicles in each blockchanges to nt+1.

The state transitions of the traffic light along time are shownin Fig. 3. The transition of the channel state for UAV-to-vehiclelinks is a stochastic process, which is reflected by (2) and (3).

Next, we discuss the MDP in three aspects: the statetransition probability, the state transitions of the number ofvehicles in each block, and the UAV’s 3D position. Notethat the transmission power control and channel control donot affect the traffic light, the channel state, the number ofvehicles, and the UAV’s 3D position.

First, we discuss the state transition probabilityp(st+1|st, at) = p((Lt+1, xt+1, zt+1,nt+1,Ht+1)|(Lt, xt, zt,nt,Ht), (ft,ρt, ct)). The UAV’s 3D fightonly affects the UAV’s 3D position state and the channelstate, the traffic light state of the next time slot relies onthe current traffic light state, and the number of vehicles ineach block of the next time slot relies on the current numberof vehicles and the traffic light state. Therefore, the statetransition probability is

p(st+1|st, at) = p(xt+1, zt+1|xt, zt,ft)×p(Ht+1|xt, zt,ft)× p(Lt+1|Lt)×p(nt+1|Lt,nt), (8)

where p(xt+1, zt+1|xt, zt,ft) is easily obtained by the 3Dposition state transition based on the UAV’s flight actions inFig. 4, p(Ht+1|xt, zt,ft) is easily obtained by (2) and (3),p(Lt+1|Lt) is obtained by the traffic light state transition inFig. 3, and p(nt+1|Lt,nt) is easily obtained by (9) ∼ (13).

Secondly, we discuss the state transitions of the number ofvehicles in each block. It is a stochastic process. The UAV’sstates and actions do not affect the number of vehicles of allblocks. Let λ1 and λ2 be the probabilities of the arrivals ofnew vehicles in flow 1 and 2, respectively.

The state transitions for the number of vehicles in block 0,3, and 4 are

n0t+1 =

n2t , if Lt = 0,

n1t , if Lt = 2,

0, otherwise,(9)

n3t+1 =

n0t , if Lt = 2, 3,

0, otherwise,(10)

n4t+1 =

n0t , if Lt = 0, 1,

0, otherwise.(11)

The transition probability is 1 in (9), (10) and (11) sincethe transitions are deterministic in block 0, 3, and 4. Whilethe state transition probabilities for the number of vehicles inblock 1 and 2 are nondeterministic, moreover, both of themare affected by their current number of vehicles and the trafficlight. Taking block 1 when the traffic light state Lt = 2 as anexample, the probability for the number of vehicles is

p(n1t+1 = 1|Lt = 2) = λ1, (12)

p(n1t+1 = 0|Lt = 2) = 1− λ1. (13)

Page 5: Ming Zhu , Xiao-Yang Liu , and Xiaodong Wang

IEEE TRANSACTIONS ON INTELLIGENT TRANSPORTATION SYSTEMS, VOL. XX, NO. XX, XXX 20XX 5

When (n1t = 0, Lt 6= 2) and (n1

t = 1, Lt 6= 2), the probabilityfor the number of vehicles will be obtained in a similar way.

Thirdly, we discuss the state transition of the UAV’s 3Dposition. It includes block transitions and height transitions.The UAV’s height transition is formulated in (7). If the UAV’sheight is fixed, the corresponding position state transitiondiagram is shown in Fig. 4, where Sii∈0,1,2,3,4 denotesthe block of the UAV: 0 denotes staying in the current block;1, 2, 3, 4 denotes a flight from block 0 to the other blocks(1, 2, 3, and 4); 5 denotes an anticlockwise flight; 6 denotes aflight from block 1, 2, 3, or 4 to block 0; 7 denotes a clockwiseflight.

IV. PROPOSED SOLUTIONS

In this section, we first describe the motivation, and thenpresent an overview of Q-learning and the deep deterministicpolicy gradient algorithm, and then propose solutions withdifferent control objectives, and finally present an extensionof solutions that takes into account the energy consumption of3D flight.

A. Motivation

Deep reinforcement learning methods are suitable for thetarget problem since environment variables are unknown andunmeasurable. For example, α1 and α2 are affected by theheight/density of buildings, and the height and size of vehicles,etc. β1 and β2 are time dependent and are affected by thecurrent environment such as weather [29]. Although UAVscan detect the LoS/NloS links using equipped cameras, itis very challenging to detect them accurately for severalreasons. First, the locations of recievers on vehicles should belabeled for detection. Secondly, it is hard to detect receiversaccurately using computer vision technology since receiversare much small than vehicles. Thirdly, it requires automobilemanufacturers to label the loactions of receivers, which maynot be satisfied in several years. Therefore, it requires a largeamount of labor to test these environment variables accurately.

It is hard to obtain the optimal strategies even all environ-ment variables are known. Existing works [30] [31] obtain thenear-optimal strategies in the 2D flight scenario when users arestationary, however, they are not capable of solving our targetproblem since the UAV adjusts its 3D position and vehiclesmove with their patterns under the control of traffic lights.

B. Q-learning

The state transition probabilities of MDP are unknown inour problem, since some variables are unknown, e.g., α1,α2, λ1, and λ2. Our problem cannot be solved directly us-ing conventional MDP solutions, e.g., dynamic programmingalgorithms, policy iteration and value iteration algorithms.Therefore, we apply the reinforcement learning (RL) ap-proach. The return from a state is defined as the sum ofdiscounted future reward

∑Ti=t γ

i−tr(si, ai), where T is thetotal number of time slots, and γ ∈ (0, 1) is a discountfactor that diminishes the future reward and ensures thatthe sum of an infinite number of rewards is still finite. Let

Qπ(st, at) = Eai∼π[∑Ti=t γ

i−tr(si, ai)|st, at] represents theexpected return after taking action at in state st under policyπ. The Bellman equation gives the optimality condition inconventional MDP solutions [32]:

Qπ(st, at)=∑st+1,rt

p(st+1, rt|st, at)[rt+γmax

at+1

Qπ(st+1, at+1)

].

Q-learning [33] is a classical model-free RL algorithm [34].Q-learning with the essence of exploration and exploitationaims to maximize the expected return by interacting with theenvironment. The update of Q(st, at) is

Q(st, at)← Q(st, at) + α[rt + γmaxat+1

Q(st+1, at+1)

−Q(st, at)], (14)

where α is a learning rate.Q-learning uses the ε-greedy strategy [35] to select an

action, so that the agent behaves greedily most of the time, butselects randomly among all the actions with a small probabilityε. The ε-greedy strategy is defined as follows

at =

arg maxaQ(st, a), with probability 1− ε,a random action, with probability ε.

(15)

The Q-learning algorithm [32] is shown in Alg. 1. Line 1is initialization. In each episode, the inner loop is executed inlines 4 ∼ 7. Line 5 selects an action using (15), and then theaction is executed in line 6. Line 7 updates the Q-value.

Q-learning cannot solve our problem because of severallimitations. 1) Q-learning can only solve MDP problems withsmall state space and action space. However, the state spaceand action space of our problem are very large. 2) Q-learningcannot handle continuous state or action space. The UAV’stransmission power allocation actions are continuous. Thetransmission power control is a continuous action in reality. Ifwe discretize the transmission power allocation actions, anduse Q-learning to solve it, the result may be far from theoptimum. 3) Q-learning will converge slowly using too manycomputational resources [32], and this is not practical in ourproblem. Therefore, we adopt the deep deterministic policygradient algorithm to solve our problem.

C. Deep Deterministic Policy Gradient

The deep deterministic policy gradient (DDPG) method[36] uses deep neural networks to approximate both actionpolicy π and value function Q(s, a). This method has twoadvantages: 1) it uses neural networks as approximators, essen-tially compressing the state and action space to much smallerlatent parameter space, and 2) the gradient descent methodcan be used to update the network weights, which greatlyspeeds up the convergence and reduces the computationaltime. Therefore, the memory and computational resources arelargely saved. In real systems, DDPG exploits the powerfulskills introduced in AlphaGo zero [37] and Atari game playing[38], including experience replay buffer, actor-critic approach,soft update, and exploration noise.

1) Experience replay buffer Rb stores transitions that willbe used to update network parameters. At each time slot t,

Page 6: Ming Zhu , Xiao-Yang Liu , and Xiaodong Wang

IEEE TRANSACTIONS ON INTELLIGENT TRANSPORTATION SYSTEMS, VOL. XX, NO. XX, XXX 20XX 6

Algorithm 1: Q-learning-based algorithmInput: the number of episodes K, the learning rate α, parameter ε.

1: Initialize all states. Initialize Q(s, a) for all state-action pairs randomly.2: for episode k = 1 to K3: Observe the initial state s1.4: for each slot t = 1 to T5: Select the UAV’s action at from state st using (15).6: Execute the UAV’s action at, receive reward rt, and observe a new state st+1 from the environment.7: Update Q-value function: Q(st, at)← Q(st, at) + α

[rt + γmaxat+1

Q(st+1, at+1)−Q(st, at)].

a transition (st, at, rt, st+1) is stored in Rb. After a certainnumber of time slots, each iteration samples a mini-batchof M = |Ω| transitions (sj , aj , rj , sj)j∈Ω to train neuralnetworks, where Ω is a set of indices of sampled transitionsfrom Rb. “Experience replay buffer” has two advantages: 1)enabling the stochastic gradient decent method [39]; and 2)removing the correlations between consecutive transitions.

2) Actor-critic approach: the critic approximates the Q-value, and the actor approximates the action policy. Thecritic has two neural networks: the online Q-network Q withparameter θQ and the target Q-network Q′ with parameter θQ

′.

The actor has two neural networks: the online policy networkµ with parameter θµ and the target policy network µ′ withparameter θµ

′. The training of these four neural networks are

discussed in the next subsection.3) Soft update with a low learning rate τ 1 is introduced

to improve the stability of learning. The soft updates of thetarget Q-network Q′ and the target policy network µ′ are asfollows

θQ′← τθQ + (1− τ)θQ

′= θQ

′+ τ(θQ − θQ

′), (16)

θµ′← τθµ + (1− τ)θµ

′= θµ

′+ τ(θµ − θµ

′). (17)

4) Exploration noise is added to the actor’s target policyto output a new action

at = µ′(st|θµ′) +Nt. (18)

There is a tradeoff between exploration and exploitation, andthe exploration is independent from the learning process.Adding exploration noise in (18) ensures that the UAV hasa certain probability of exploring new actions besides the onepredicted by the current policy µ′(st|θµ

′), and avoids that the

UAV is trapped in a local optimum.

D. Deep Reinforcement Learning-based Solutions

The UAV has two transmission controls, power and channel.We use the power allocation as the main control objectivefor two reasons. 1) Once the power allocation is determined,the channel allocation will be easily obtained in OFDMA.According to Theorem 4 of [40], in OFDMA, if all linkshave the equal weights just as our reward function (5), thetransmitter should send messages to the receiver with thestrongest channel in each time slot. In our problem, thestrongest channel is not determined since the channel state(LoS or NLoS) is a random process. DDPG trends to allocatemore power to the strongest channels with large probabilities,

therefore, channel allocation will be easily obtained based onpower allocation actions. 2) Power allocation is continuous,and DDPG is suitable to handle these actions. However, if weuse DDPG for the channel allocation, the number of actionvariables will be very large and the convergence will be veryslow, since the channel allocation is discrete and the number ofchannels is generally large (e.g., 200) especially in rush hours.We choose power control or flight as control objectives sincecontrolling power and flight is more efficient than controllingchannel. Moreover, the best channel allocation strategy canbe obtained indirectly if the power is allocated in OFDMA.Based on the above analysis, we propose three algorithms:• PowerControl: the UAV adjusts the transmission power

allocation using the actor network at a fixed 3D position,and the channels are allocated to vehicles by Alg.2 ineach time slot.

• FlightControl: the UAV adjusts its 3D flight using theactor network, and the transmission power and channelallocation are equally allocated to each vehicle in eachtime slot.

• JointControl: the UAV adjusts its 3D flight and thetransmission power allocation using the actor network,and the channels are allocated to vehicles by Alg.2 ineach time slot.

To allocate channels among blocks, we introduce a variabledenoting the average allocated power of a vehicle in block i:

ρit =

ρitnit, if nit 6= 0,

0, otherwise.(19)

The channel allocation algorithm is shown in Alg. 2, whichis executed after obtaining the power allocation actions. Asthe above description, it achieves the best channel allocationin OFDMA if the power allocation is known [40]. Line 1is the initialization. Lines 2 ∼ 3 calculate and sort ρt =ρiti∈0,1,2,3,4. Line 5 assigns the maximum number ofchannels to the current possibly strongest channel, and line6 updates the remaining total number of channels.

The DDPG-based algorithms are given in Alg. 3. Thealgorithm has two parts: initializations, and the main process.First, we describe the initializations in lines 1 ∼ 3. In line 1, allstates are initialized: the traffic light L is initialized as 0, thenumber of vehicles n in all blocks is 0, the UAV’s block andheight are randomized, and the channel state Hi for each blocki is set as LoS or NLoS with the same probability. Note that theaction space DDPG controls in PowerControl, FlightControl,and JointControl is different. Line 2 initializes the parameters

Page 7: Ming Zhu , Xiao-Yang Liu , and Xiaodong Wang

IEEE TRANSACTIONS ON INTELLIGENT TRANSPORTATION SYSTEMS, VOL. XX, NO. XX, XXX 20XX 7

: actor s online policy network

Experience

replay buffer

Agent (UAV)

Input

layer

Fully-

connected

layer

Fully-

connected

layer

Output

layerInput

layer

Output

layer

Fully-

connected

layer

Fully-

connected

layer

Input

layerFully-

connected

layer

Fully-

connected

layer

Output

layer

: actor s target policy network

: critic s online Q-network

Input

layer

Output

layer

Fully-

connected

layer

Fully-

connected

layer

: critic s target Q-network

Actor Critic

Soft update Soft update

Environment

Communication links

Fig. 5. Framework of the DDPG algorithm.

Algorithm 2: Channel allocation in time slot tInput: the power allocation ρ, the number of vehicles in all blocks n, the maximum number of channels allocated to

a vehicle cmax, the total number of channels C.Output: the channel allocation ct for all blocks.

1: Initialize the remaining total number of channels Cr ← C.2: Calculate the average allocated power for each vehicle in all blocks ρt by (19).3: Sort ρt by the descending order, and obtain a sequence of block indices J .4: for block j ∈ J5: cjt ← min(Cr, n

jtcmax).

7: Cr ← Cr − cjt .8: Return ct.

Page 8: Ming Zhu , Xiao-Yang Liu , and Xiaodong Wang

IEEE TRANSACTIONS ON INTELLIGENT TRANSPORTATION SYSTEMS, VOL. XX, NO. XX, XXX 20XX 8

Algorithm 3: DDPG-based algorithms: PowerControl, FlightControl, and JointControlInput: the number of episodes K, the number of time slots T in an episode, the mini-batch size M , the learning rate τ .

1: Initialize all states, including the traffic light state L, the UAV’s 3D position (x, z), the number of vehicles n and thechannel state H in all blocks.

2: Randomly initialize critic’s online Q-network parameters θQ and actor’s online policy network parameters θµ, andinitialize the critic’s target Q-network parameters θQ

′ ← θQ and actor’s target policy network parameters θµ′ ← θµ.

3: Allocate an experience replay buffer Rb.4: for episode k = 1 to K5: Initialize a random process (a standard normal distribution) N for the UAV’s action exploration.6: Observe the initial state s1.7: for t = 1 to T8: Select the UAV’s action at = µ′(st|θµ

′) +Nt according to the policy of µ′ and the exploration noise Nt.

9: if PowerControl10: Combine the channel allocation in Alg. 2 and at as the UAV’s action at at a fixed 3D position.11: if FlightControl12: Combine the equal transmission power, equal channel allocation and at (3D flight) as the UAV’s action at.13: if JointControl14: Combine the 3D flight action, the channel allocation in Alg. 2 and at as the UAV’s action at.15: Execute the UAV’s action at, and receive reward rt, and observe new state st+1 from the environment.16: Store transition (st, at, rt, st+1) in the UAV’s experience replay buffer Rb.17: Sample Rb to obtain a random mini-batch of M transitions (sjt , a

jt , r

jt , s

jt+1)j∈Ω ⊆ Rb, where Ω is a set of

indices of sampled transitions with |Ω| = M .18: The critic’s target Q-network Q′ calculates and outputs yjt = rjt + γQ′(sjt+1, µ

′(sjt+1|θµ′)|θQ′

) to the critic’sonline Q-network Q.

19: Update the critic’s online Q-network Q to make its Q-value fit yjt by minimizing the loss function:∇θQLosst(θQ) = ∇θQ [ 1

M

∑Mj=1(yjt −Q(sjt , a

jt |θQ))2].

20: Update the actor’s online policy network µ based on the input ∇aQ(s, a|θQ)|s=sjt ,a=µ(sjt)j∈Ω from Q using the

policy gradient by the chain rule:1M

∑j∈Ω Est [∇aQ(s, a|θQ)|s=st,a=µ(st)∇θµµ(s|θµ)|s=st ].

21: Soft update the critic’s target Q-network Q′ and actor’s target policy network µ′ to make the evaluation of theUAV’s actions and the UAV’s policy more stable: θQ

′ ← τθQ + (1− τ)θQ′, θµ

′ ← τθµ + (1− τ)θµ′.

of the critic and actor. Line 3 allocates an experience replaybuffer Rb.

Secondly, we present the main process. Line 5 initializesa random process for action exploration. Line 6 receives aninitial state s1. Let at be the action DDPG controls, and atbe the UAV’s all action. Line 8 selects an action according toat and an exploration noise Nt. Lines 9 ∼ 10 combine thechannel allocation actions in Alg. 2 and at as at at a fixed 3Dposition in PowerControl. Lines 11 ∼ 12 combine the equaltransmission power, equal channel allocation actions and at(3D flight) as at in FlightControl. Lines 13 ∼ 14 combinethe 3D flight action, the channel allocation actions in Alg.2 and at as at in JointControl. Line 15 executes the UAV’saction at, and then the UAV receives a reward and all statesare updated. Line 16 stores a transition into Rb. In line 17, arandom mini-batch of transitions are sampled from Rb. Line18 sets the value of yj for the critic’s online Q-network. Lines19 ∼ 21 update all network parameters.

The DDPG-based algorithms in Alg. 3 in essence arethe approximated Q-learning method in Alg. 1. The ex-ploration noise in line 8 approximates the second caseof (15) in Q-learning. Lines 18 ∼ 19 in Alg. 3 make[rt + γmaxat+1

Q(st+1, at+1)−Q(st, at)]

in line 7 of Alg.1 converge. Line 20 of Alg. 3 approximates the first case of

(15) in Q-learning, since both of them aim to obtain the policyof the maximum Q-value. The soft update of Q′ in line 21 ofAlg. 3 is exactly (14) in Q-learning, where τ and α are learningrates. Next, we discuss the training and test stages of proposedsolutions.

1) In the training stage, we train the actor and the critic, andstore the parameters of their neural networks. Fig. 5 illustratesthe data flow and parameter update process. The training stagehas two parts. First, Q and µ are trained through a randommini-batch of transitions sampled from the experience replaybuffer Rb. Secondly, Q′ and µ′ are trained through soft update.

The training process is as follows. A mini-batch of Mtransitions (sjt , a

jt , r

jt , s

jt+1)j∈Ω are sampled from Rb, where

Ω is a set of indices of sampled transitions from Rb with|Ω| = M . Then two data flows are outputted from Rb:rjt , s

jt+1j∈Ω → µ′, and sjt , a

jtj∈Ω → Q. µ′ outputs

rjt , sjt+1, µ

′(sjt+1|θµ′)j∈Ω to Q′ to calculate yjt j∈Ω. Then

Q calculates and outputs ∇aQ(s, a|θQ)|s=sjt ,a=µ(sjt)j∈Ω to

µ. µ updates its parameters by (22). Then two soft updates areexecuted for Q′ and µ′ in (16) and (17), respectively.

The data flow of the critic’s target Q-network Q′

and online Q-network Q are as follows. Q′ takes(rjt , s

jt+1, µ

′(sjt+1|θµ′))j∈Ω as the input and outputs

Page 9: Ming Zhu , Xiao-Yang Liu , and Xiaodong Wang

IEEE TRANSACTIONS ON INTELLIGENT TRANSPORTATION SYSTEMS, VOL. XX, NO. XX, XXX 20XX 9

yjt j∈Ω to Q. yjt is calculated by

yjt = rjt + γQ′(sjt+1, µ′(sjt+1|θµ

′)|θQ

′). (20)

Q takessjt , a

jtj∈Ω,

as the input and outputs

∇aQ(s, a|θQ)|s=sjt ,a=µ(sjt)j∈Ω to µ for updating

parameters in (22), where sjtj∈Ω are sampled fromRb, and µ(sjt ) = arg maxaQ(sjt , a).

The data flows of the actor’s online policy network µ andtarget policy network µ′ are as follows. After Q outputs∇aQ(s, a|θQ)|s=sjt ,a=µ(sjt)

j∈Ω to µ, µ updates its param-eters by (22). µ′ takes rjt , s

jt+1j∈Ω as the input and outputs

rjt , sjt+1, µ

′(sjt+1|θµ′)j∈Ω to Q′ for calculating yjt j∈Ω in

(20), where rjt , sjt+1j∈Ω are sampled from Rb.

The updates of parameters of four neural networks (Q,Q′, µ, and µ′) are as follows. The online Q-network Q up-dates its parameters by minimizing the L2-norm loss functionLosst(θQ) to make its Q-value fit yjt :

∇θQLosst(θQ) = ∇θQ [1

M

M∑j=1

(yjt −Q(sjt , ajt |θQ))2]. (21)

The target Q-network Q′ updates its parameters θQ′

by (16).The online policy network µ updates its parameters followingthe chain rule with respect to θµ:

Est [∇θµQ(s, a|θQ)|s=st,a=µ(st|θµ)]

= Est [∇aQ(s, a|θQ)|s=st,a=µ(st)∇θµµ(s|θµ)|s=st ]. (22)

The target policy network µ′ updates its parameters θµ′

by(17).

In each time slot t, the current state st from the environ-ment is delivered to µ′, and µ′ calculates the UAV’s targetpolicy µ′(st)|θµ

′. Finally, an exploration noise N is added to

µ′(st|θµ′) to get the UAV’s action in (18).

2) In the test stage, we restore the neural network ofthe actor’s target policy network µ′ based on the storedparameters. This way, there is no need to store transitions tothe experience replay buffer Rb. Given the current state st, weuse µ′ to obtain the UAV’s optimal action µ′(st|θµ

′). Note that

there is no noise added to µ′(st|θµ′), since all neural networks

have been trained and the UAV has got the optimal actionthrough µ′. Finally, the UAV executes the action µ′(st|θµ

′).

E. Extension on Energy Consumption of 3D Flight

The UAV’s energy is used in two parts, communication and3D flight. The above proposed solutions in Alg. 3 do not con-sider the energy consumption of 3D flight. In this subsection,we discuss how to incorporate the energy consumption of 3Dflight into Alg. 3. To encourage or discourage the UAV’s 3Dflight actions in different directions with different amount ofenergy consumption, we modify the reward function and theDDPG framework.

The UAV aims to maximize the total throughput per energyunit since the UAV’s battery has limited capacity. For example,the UAV DJI Mavic Air [41] with full energy can only fly21 minutes. Given that the UAV’s energy consumption of 3Dflight is much larger than that of communication, we only use

the former part as the total energy consumption. Thus, thereward function (5) is modified as follows

r(st, at) =1

e(at)

∑i∈0,1,2,3,4

bnitcit log(1 +

ρithit

bcitσ2

), (23)

where e(at) is the energy consumption of taking action atin time slot t. Our energy consumption setups follow theUAV DJI Mavic Air [41]. The UAV has three vertical flightactions per time slot just as in (6). If the UAV keeps movingdownward, horizontally, or upward until the energy for 3Dflight is used up, the flight time is assumed to be 27, 21, and17 minutes, respectively. If the duration of a time slot is set to6 seconds, so the UAV can fly 270, 210, and 170 time slots,respectively. Therefore, the formulation of e(at) is given by

e(at) =

1

270Efull, if moving downward 5 meters,1

210Efull, if moving horizontally,1

170Efull, if moving upward 5 meters,(24)

where Efull is the total energy if the UAV’s battery is full.Let δ(t) be a prediction error as follows

δ(t) = r(st, at)−Q(st, at), (25)

where δ(t) evaluates the difference between the actual rewardr(st, at) and the expected return Q(st, at). To make the UAVlearn from the prediction error δ(t), not the difference betweenthe new Q-value and old Q-value in (14), the Q-value isupdated by the following rule

Q(st, at)← Q(st, at) + αδ(t)⇔Q(st, at)← Q(st, at) + α(r(st, at)−Q(st, at)), (26)

where α is a learning rate similar to (14).We introduce α+ and α− to represent the learning rate when

δ(t) ≥ 0 and δ(t) < 0, respectively. Therefore, the UAV canchoose to be active or inactive by properly setting the values ofα+ and α−. The update of Q-value in Q-learning is modifiedas follows, inspired by [42]

Q(st, at)← Q(st, at) +

α+δ(t), if δ(t) ≥ 0,

α−δ(t), if δ(t) < 0.(27)

We define the prediction error δ(t) as the difference betweenthe actual reward and the output of the critic’s online Q-network Q:

δ(t) = r(st, at)−Q(st, at|θQ). (28)

We use τ+ and τ− to denote the weights when δ(t) ≥ 0and δ(t) < 0, respectively. The update of the critic’s targetQ-network Q′ is

θQ′←

τ+θQ + (1− τ+)θQ

′, if δ(t) ≥ 0,

τ−θQ + (1− τ−)θQ′, if δ(t) < 0.

(29)

The update of the actor’s target policy network µ′ is

θµ′←

τ+θµ + (1− τ+)θµ

′, if δ(t) ≥ 0,

τ−θµ + (1− τ−)θµ′, if δ(t) < 0.

(30)

If τ+ > τ−, the UAV is active and prefers to move. Ifτ+ < τ−, the UAV is inactive and prefers to stay. If τ+ = τ−,

Page 10: Ming Zhu , Xiao-Yang Liu , and Xiaodong Wang

IEEE TRANSACTIONS ON INTELLIGENT TRANSPORTATION SYSTEMS, VOL. XX, NO. XX, XXX 20XX 10

the UAV is neither active nor inactive. To approximate theQ-value, we introduce yjt similar to (20) and then make thecritic’s online Q-network Q to fit it. We optimize the lossfunction

∇θQLosst(θQ) = ∇θQ [1

M

M∑j=1

(yjt −Q(sjt , ajt |θQ))2], (31)

where yjt = rjt .We modify the MDP, DDPG framework, and DDPG-based

algorithms by considering the energy consumption of 3Dflight:• The MDP is modified as follows. The state space S =

(L, x, z, n,H,E), where E is the energy in the UAV’sbattery. The energy changes as follows

Et+1 = maxEt − e(at), 0. (32)

The other parts of MDP formulation and state transitionsare the same as in Section III-C.

• There are three modifications in the DDPG framework:a) The critic’s target Q-network Q′ feeds yj = rj to thecritic’s online Q-network Q instead of yj in (20). b) Theupdate of the critic’s target Q-network Q′ is (29) insteadof (16). c) The update of the actor’s target policy networkµ′ is (30) instead of (17).

• The DDPG-based algorithms are modified from Alg. 3.Initialize the energy state of the UAV as full in the start ofeach episode. In each time step of an episode, the energystate is updated by (32), and this episode terminates ifthe energy state Et ≤ 0. The reward function is replacedby (23).

V. PERFORMANCE EVALUATION

For a one-way-two-flow road intersection in Fig. 2, wepresent the optimality verification of deep reinforcement learn-ing algorithms. Then, we study a more realistic road intersec-tion as shown in Fig. 7, and present our simulation results.

Our simulations are executed on a server with LinuxOS, 200 GB memory, two Intel(R) Xeon(R) Gold [email protected] GHz, a Tesla V100-PCIE GPU.

The implementation of Alg. 3 includes two parts: buildingthe environment (including traffic and communication models)for our scenarios, and using the DDPG algorithm in Tensor-Flow [43].

A. Optimality Verification of Deep Reinforcement Learning

The parameter settings are summarized in Table II.In the simulations, there are three types of parameters:DDPG algorithm parameters, communication parameters, andUAV/vehicle parameters.

First, we describe the DDPG algorithm parameters. Thenumber of episodes is 256, and the number of time slots in anepisode is 256, so the number of total time slots is 65,536. Theexperience replay buffer capacity is 10,000, and the learningrate of target networks τ is 0.001. The mini-batch size M is512. The training data set is full in the 10, 000th time slot,and is updated in each of the following 256 × 256 - 10,000

TABLE IIVALUES OF PARAMETERS IN SIMULATION SETTINGS

α1 α2 β1 β2 σ2 d9.6 0.28 3 0.01 -130 dBm/Hz 3P C N γ zmin zmax

1 ∼ 6 10 10 0.9 10 200M λ gsi gli gri b512 0.1 ∼ 0.7 0.4 0.3 0.3 100 KHzτ ρmax cmax

0.001 3 W 5

= 55,536 time slots. The test data set is real-time among allthe 256 × 256 = 65,536 time slots.

Secondly, we describe communication parameters. α1 andα2 are set to 9.6 and 0.28, which are common values in urbanareas [44]. β1 is 3, and β2 is 0.01, which are widely usedin path loss modeling. The duration of a time slot is set to 6seconds, and the number of occupied red or green traffic lightN is 10, i.e., 60 seconds constitute a red/green duration, whichis commonly seen in cities and can ensure that the vehicles inblocks can get the next block in a time slot. The white powerspectral density σ2 is set to -130 dBm/Hz. The total UAVtransmission power P is set to 6 W in consideration of thelimited communication ability. The total number of channels Cis 10. The bandwidth of each channel b is 100 KHz. Therefore,the total bandwidth of all channels is 1 MHz. The maximumpower allocated to a vehicle ρmax is 3 W, and the maximumnumber of channels allocated to a vehicle cmax is 5. We assumethat the power control for each vehicle has 4 discrete values(0, 1, 2, 3).

Thirdly, we describe UAV/vehicle parameters. λ is set to 0.1∼ 0.7. The length of a road block d is set to 3 meters. Theblocks’ distance is easily calculated as follows: D(1, 0) = d,and D(1, 3) = 2d, where D(i, j) is the Euclidean distancefrom block i to block j. We assume the arrival of vehicles inblock 1 and 2 follows a binomial distribution with the sameparameter λ in the range 0.1 ∼ 0.7. The discount factor γ is0.9.

The assumptions of the simplified scenario in Fig. 2 are asfollows. To keep the state space small for verification purpose,we assume the channel states of all communication links areLoS, and the UAV’s height is fixed as 150 meters, so thatthe UAV can only adjusts its horizontal flight control andtransmission control. The traffic light state is assumed to havetwo values (red or green).

The configure of neural networks in proposed solutions isbased on the configure of the DDPG action space. A neuralnetwork consists of an input layer, fully-connected layers, andan output layer. The number of fully-connected layers in actoris set to 4.

Theoretically, it is well-known that deep reinforcementlearning algorithms (including DDPG algorithms) solve MDPproblems and achieve the optimal results with much lessmemory and computational resources. We provide the opti-mality verification of DDPG-based algorithms in Alg. 3 ina one-way-two-flow road intersection in Fig. 2. The reasonsare as follows: (i) the MDP problem in such a simplifiedscenario is explicitly defined and the theoretically optimal

Page 11: Ming Zhu , Xiao-Yang Liu , and Xiaodong Wang

IEEE TRANSACTIONS ON INTELLIGENT TRANSPORTATION SYSTEMS, VOL. XX, NO. XX, XXX 20XX 11

0.1 0.2 0.3 0.4 0.5 0.6 0.7

Vehicle arrival probability λ

8

10

12

14

16

Tota

l th

roughput

(Mbps)

PowerControl

FlightControl

JointControl

Opt PowerControl

Opt FlightControl

Opt JointControl

Fig. 6. Total throughput vs. vehicle arrival probability λ in optimalityverification.

policy can be obtained using the Python MDP Toolbox [45];and (ii) this optimality verification process also serves a goodcode debugging process before we apply the DDPG algorithmin TensorFlow [43] to the more realistic road intersectionscenario in Fig. 7.

The result of DDPG-based algorithms matches that of thepolicy iteration algorithm using Python MDP Toolbox [45](serving as the optimal policy). The total throughput obtainedby the policy iteration algorithm and DDPG-based algorithmsare shown as dashed lines and solid lines in Fig. 6. Therefore,DDPG-based algorithms achieve near optimal policies. We seethat, the total throughput in JointControl is the largest, which ismuch higher than PowerControl and FlightControl. This is inconsistent with our believes that the JointControl of power andflight allocation will be better than the control of either of both.The performance of PowerControl is better than FlightControl.The throughput increases with the increasing of vehicle arrivalprobability λ in all algorithms, and it saturates when λ ≥ 0.6due to traffic congestion.

B. More Realistic Traffic Model

We consider a more realistic road intersection model inFig. 7. There are totally 33 blocks with four entrances (block26, 28, 30, and 32), and four exits (block 25, 27, 29, and31). Vehicles in block i ∈ 2, 4, 6, 8 go straight, turn left,turn right with the probabilities gsi , gli, and gri , such thatgsi + gli + gri = 1. We assume vehicles can turn right when thetraffic light is green.

Now, we describe the settings different from the last sub-section. The discount factor γ is 0.4 ∼ 0.9. The total UAVtransmission power P is set to 1 ∼ 6 W. The total number ofchannels C is 100 ∼ 200, which is much larger than that insubsection V-A since there are more vehicles in the realisticmodel. The bandwidth of each channel b is 5 KHz, therefore,the total bandwidth of all channels is 0.5 ∼ 1 MHz. Themaximum power allocated to a vehicle ρmax is 0.9 W, and

6 5

1 2

4

3

7

8

9 10

12

11

14 13

15

160

17 18

20

19

22 21

23

24

25

28

30

31

26

29

32 27

Fig. 7. Realistic road intersection model.

the maximum number of channels allocated to a vehicle cmaxis 50. The minimum and maximum height of the UAV is 10meters and 200 meters. The probability of a vehicle goingstraight, turning left, and turning right (gsi , gli, and gri ) is setto 0.4, 0.3, and 0.3, respectively, and each of them is assumedto be the same in block 2, 4, 6, and 8. We assume the arrivalof vehicles in block 26, 28, 30, and 32 follows a binomialdistribution with the same parameter λ in the range 0.1 ∼ 0.7.

The UAV’s horizontal and vertical flight actions are asfollows. We assume that the UAV’s block is 0 ∼ 8 since thenumber of vehicles in the intersection block 0 is generallythe largest and the UAV will not move to the block far fromthe intersection block. Moreover, within a time slot we assumethat the UAV can stay or only move to its adjacent blocks. TheUAV’s vertical flight action is set by (6). In PowerControl, theUAV stays at block 0 with the height of 150 meters.

C. Baseline Schemes

We compare with two baseline schemes. Generally, theequal transmission power and channels allocation is commonin communication systems for fairness. Therefore, they areused in baseline schemes.

The first baseline scheme is Cycle, i.e., the UAV cyclesanticlockwise at a fixed height (e.g., 150 meters), and the UAVallocates the transmission power and channels equally to eachvehicle in each time slot. The UAV moves along the fixedtrajectory periodically, without considering the vehicle flows.

The second baseline scheme is Greedy, i.e., at a fixed height(e.g., 150 meters), the UAV greedily moves to the block withthe largest number of vehicles. If a nonadjacent block has thelargest number of vehicles, the UAV has to move to block0 and then move to that block. The UAV also allocates thetransmission power and the channels equally to each vehicle

Page 12: Ming Zhu , Xiao-Yang Liu , and Xiaodong Wang

IEEE TRANSACTIONS ON INTELLIGENT TRANSPORTATION SYSTEMS, VOL. XX, NO. XX, XXX 20XX 12

10000 10200 10400 10600 10800 1100076543210

-Loss

-Loss of PowerControl

10000 10200 10400 10600 10800 1100076543210

-Loss

-Loss of FlightControl

10000 10200 10400 10600 10800 11000Time slot

76543210

-Loss

-Loss of JointControl

Fig. 8. Convergence of loss functions in training stage.

in each time slot. The UAV tries to serve the block with thelargest number of vehicles by moving nearer to them.

D. Simulation Results

The training time is about 4 hours, and the test time isalmost real-time, since it only uses the well trained targetpolicy network. Next, we first show the convergence of lossfunctions, and then show total throughput vs. discount factor,total transmission power, total number of channels and vehiclearrival probability, and finally present the total throughput andthe UAV’s flight time vs. energy percent for 3D flight.

The convergence of loss functions in training stage forPowerControl, FlightControl, and JointControl indicates thatthe neural network is well-trained. It is shown in Fig. 8 whenP = 6, C = 200, λ = 0.5 and γ = 0.9 during time slots10,000 ∼ 11,000. The first 10,000 time slots are not shownsince during the 0 ∼ 10,000, the experience replay buffer hasnot achieved its capacity. We see that, the loss functions inthree algorithms converge after time slot 11,000. The othermetrics in the paper are measured in test stage by default.

Total throughput vs. discount factor γ is drawn in Fig. 9when P = 6, C = 200, and λ = 0.5. We can see that, whenγ changes, the throughput of three algorithms is steady; andJointControl achieves higher total throughput, comparing withPowerControl and FlightControl, respectively. PowerControlachieves higher throughput than FlightControl since Power-Control allocates power and channel to strongest channelswhile FlightControl only adjusts the UAV’s 3D position toenhance the strongest channel and the equal power and channelallocation is far from the best strategy in OFDMA.

Total throughput vs. total transmission power (P = 1 ∼ 6)and total number of channels (C = 100 ∼ 200) are shownby Fig. 10 and Fig. 11, where we set λ = 0.5 and γ = 0.9.We see that JointControl achieves the best performance fordifferent transmission power and channel budgets, respectively.Moreover, the total throughput of all algorithms increaseswhen the total transmission power or total number of channelsincreases. PowerControl and FlightControl only adjust thetransmission power or 3D flight, while JointControl jointly

0.4 0.5 0.6 0.7 0.8 0.9

Discount factor γ

10

12

14

16

18

20

Tota

l th

roughput

(Mbps)

PowerControl

FlightControl

JointControl

Fig. 9. Throughput vs. discount factor γ.

1 2 3 4 5 6

Total transmission power (W)

11

12

13

14

15

16

17

18

19

Tota

l th

roughput

(Mbps)

PowerControl

FlightControl

JointControl

Cycle

Greedy

Fig. 10. Total throughput vs. total transmission power (C = 200).

adjusts both of them, so its performance is the best. The totalthroughput of DDPG-based algorithms is improved greatlythan that of Cycle and Greedy. The performance of Greedyis a little better than Cycle, since Greedy tries to get nearer tothe block with the largest number of vehicles.

Total throughput vs. vehicle arrival probability λ is shownin Fig. 12. Note that the road intersection has a capacity of2 units, i.e., it can serve at most two traffic flows at thesame time, therefore, it cannot serve traffic flows where λis very high, e.g., λ = 0.8 and λ = 0.9. We see that,when λ increases, i.e., more vehicles arrive at the intersection,the total throughput increases. However, when λ gets higher,e.g., λ = 0.6, the total throughput saturates due to trafficcongestion.

Next, we test the metrics considering of the energy con-sumption of 3D flight. The total throughput vs. energy percent

Page 13: Ming Zhu , Xiao-Yang Liu , and Xiaodong Wang

IEEE TRANSACTIONS ON INTELLIGENT TRANSPORTATION SYSTEMS, VOL. XX, NO. XX, XXX 20XX 13

100 120 140 160 180 200

Total number of channels

6

8

10

12

14

16

18

20Tota

l th

roughput

(Mbps)

PowerControl

FlightControl

JointControl

Cycle

Greedy

Fig. 11. Total throughput vs. total number of channels (P = 6).

0.1 0.2 0.3 0.4 0.5 0.6 0.7

Vehicle arrival probability λ

6

8

10

12

14

16

18

20

Tota

l th

roughput

(Mbps)

PowerControl

FlightControl

JointControl

Cycle

Greedy

Fig. 12. Total throughput vs. vehicle arrival probability λ.

for 3D flight in JointControl is shown in Fig. 13. When τ+

increases, the total throughput almost increases. We get thatif the UAV is more active in 3D flight, it will help to improvethe total throughput. However, the improvement of the totalthroughput is not very clear since the UAV has to considerthe energy consumption in the new reward function (23). Inaddition, when τ+ is higher, the total throughput has morevariance since the UAV prefers to get higher reward throughmore ventures.

The UAV’s flight time vs. energy percent for 3D flight inJointControl is shown in Fig. 14. When τ− = 0.001 and τ+ =0.0008, the UAV’s flight time is the longest since the UAV isinactive. When τ− = 0.001 and τ+ = 0.0012, the UAV’sflight time is the shortest, since the UAV is active and prefersto flight. When τ− = τ+ = 0.001, the UAV’s flight timeis between the other two cases. If the energy percent for 3D

40 50 60 70 80 90

Energy percent for 3D flight (%)

14.0

14.5

15.0

15.5

16.0

16.5

17.0

17.5

Tota

l th

roughput

(Mbps)

τ − = 0. 001, τ + = 0. 0008

τ − = 0. 001, τ + = 0. 001

τ − = 0. 001, τ + = 0. 0012

Fig. 13. Total throughput vs. energy percent for 3D flight in JointControl (P= 6, C = 200).

40 50 60 70 80 90

Energy percent for 3D flight (%)

6

8

10

12

14

16

18

20

22

24U

AV

's f

light

tim

e (

min

ute

s)

τ − = 0. 001, τ + = 0. 0008

τ − = 0. 001, τ + = 0. 001

τ − = 0. 001, τ + = 0. 0012

Fig. 14. UAV’s flight time vs. energy percent for 3D flight in JointControl(P = 6, C = 200).

flight increases, the UAV’s flight time increases linearly in thethree cases.

VI. CONCLUSIONS

We studied a UAV-assisted vehicular network where theUAV acted as a relay to maximize the total throughputbetween the UAV and vehicles. We focused on the downlinkcommunication where the UAV could adjust its transmissioncontrol (power and channel) under 3D flight. We formulatedour problem as a MDP problem, explored the state transitionsof UAV and vehicles under different actions, and then pro-posed three deep reinforcement learning schemes based on theDDPG algorithms, and finally extended them to account forthe energy consumption of the UAV’s 3D flight by modifying

Page 14: Ming Zhu , Xiao-Yang Liu , and Xiaodong Wang

IEEE TRANSACTIONS ON INTELLIGENT TRANSPORTATION SYSTEMS, VOL. XX, NO. XX, XXX 20XX 14

the reward function and the DDPG framework. In a simplifiedscenario with small state space and action space, we verifiedthe optimality of DDPG-based algorithms. Through simulationresults, we demonstrated the superior performance of thealgorithms under a more realistic traffic scenario comparedwith two baseline schemes.

In the future, we will consider the scenario where multipleUAVs constitute a relay network to assist vehicular networksand study the coverage overlap/probability, relay selection, en-ergy harvesting communications, and UAV cooperative com-munication protocols. We pre-trained the proposed solutionsusing servers, and we hope the UAV trains the neural netwroksin the future if light and low energy consumption GPUs areapplied at the edge.

REFERENCES

[1] M. Chaqfeh, H. El-Sayed, and A. Lakas, “Efficient data disseminationfor urban vehicular environments,” IEEE Transactions on IntelligentTransportation Systems (TITS), no. 99, pp. 1–11, 2018.

[2] M. Zhu, X.-Y. Liu, F. Tang, M. Qiu, R. Shen, W. Shu, and M.-Y. Wu,“Public vehicles for future urban transportation,” IEEE Transactions onIntelligent Transportation Systems (TITS), vol. 17, no. 12, pp. 3344–3353, 2016.

[3] M. Zhu, X.-Y. Liu, and X. Wang, “Joint transportation and chargingscheduling in public vehicle systems - a game theoretic approach,” IEEETransactions on Intelligent Transportation Systems (TITS), vol. 19, no. 8,pp. 2407–2419, 2018.

[4] M. Zhu, X.-Y. Liu, and X. Wang, “An online ride-sharing path-planningstrategy for public vehicle systems,” IEEE Transactions on IntelligentTransportation Systems (TITS), vol. 20, no. 2, pp. 616–627, 2019.

[5] K. Li, C. Yuen, S. S. Kanhere, K. Hu, W. Zhang, F. Jiang, andX. Liu, “An experimental study for tracking crowd in smart cities,”IEEE Systems Journal, 2018.

[6] F. Cunha, L. Villas, A. Boukerche, G. Maia, A. Viana, R. A. Mini,and A. A. Loureiro, “Data communication in VANETs: protocols,applications and challenges,” Elsevier Ad Hoc Networks, vol. 44, pp. 90–103, 2016.

[7] H. Sedjelmaci, S. M. Senouci, and N. Ansari, “Intrusion detection andejection framework against lethal attacks in UAV-aided networks: abayesian game-theoretic methodology,” IEEE Transactions on IntelligentTransportation Systems (TITS), vol. 18, no. 5, pp. 1143–1153, 2017.

[8] “Paving the path to 5G: optimizing commercial LTE networks fordrone communication (2018).”https://www.qualcomm.cn/videos/paving-path-5g-optimizing-commercial-lte-networks-drone-communication.

[9] “Huawei signs MoU with China Mobile Sichuan and Fonair aviation tobuild cellular test networks for logistics drones (2018).”https://www.huawei.com/en/press-events/news/2018/3/MoU-ChinaMobile-FonairAviation-Logistics.

[10] M. Alzenad, A. El-Keyi, F. Lagum, and H. Yanikomeroglu, “3-Dplacement of an unmanned aerial vehicle base station (UAV-BS) forenergy-efficient maximal coverage,” IEEE Wireless CommunicationsLetters (WCL), vol. 6, no. 4, pp. 434–437, 2017.

[11] M. Giordani, M. Mezzavilla, S. Rangan, and M. Zorzi, “An efficientuplink multi-connectivity scheme for 5G mmWave control plane appli-cations,” IEEE Transactions on Wireless Communications (TWC), 2018.

[12] S. Wang, H. Liu, P. H. Gomes, and B. Krishnamachari, “Deep reinforce-ment learning for dynamic multichannel access in wireless networks,”IEEE Transactions on Cognitive Communications and Networking(TCCN), vol. 4, no. 2, pp. 257–265, 2018.

[13] Q. Yang and S.-J. Yoo, “Optimal UAV path planning: sensing dataacquisition over IoT sensor networks using multi-objective bio-inspiredalgorithms,” IEEE Access, vol. 6, pp. 13671–13684, 2018.

[14] M. Garraffa, M. Bekhti, L. Letocart, N. Achir, and K. Boussetta,“Drones path planning for WSN data gathering: a column generationheuristic approach,” in IEEE Wireless Communications and NetworkingConference (WCNC), pp. 1–6, 2018.

[15] C. H. Liu, Z. Chen, J. Tang, J. Xu, and C. Piao, “Energy-efficientUAV control for effective and fair communication coverage: A deepreinforcement learning approach,” IEEE Journal on Selected Areas inCommunications (JSAC), vol. 36, no. 9, pp. 2059–2070, 2018.

[16] H. Wang, G. Ding, F. Gao, J. Chen, J. Wang, and L. Wang, “Power con-trol in UAV-supported ultra dense networks: communications, caching,and energy transfer,” IEEE Communications Magazine, vol. 56, no. 6,pp. 28–34, 2018.

[17] S. Yan, M. Peng, and X. Cao, “A game theory approach for joint accessselection and resource allocation in UAV assisted IoT communicationnetworks,” IEEE Internet of Things Journal (IOTJ), 2018.

[18] Y. Wu, J. Xu, L. Qiu, and R. Zhang, “Capacity of UAV-enabledmulticast channel: joint trajectory design and power allocation,” in IEEEInternational Conference on Communications (ICC), pp. 1–7, 2018.

[19] Y. Zeng, X. Xu, and R. Zhang, “Trajectory design for completiontime minimization in UAV-enabled multicasting,” IEEE Transactions onWireless Communications (TWC), vol. 17, no. 4, pp. 2233–2246, 2018.

[20] S. Zhang, Y. Zeng, and R. Zhang, “Cellular-enabled UAV communi-cation: trajectory optimization under connectivity constraint,” in IEEEInternational Conference on Communications (ICC), pp. 1–6, 2018.

[21] R. Fan, J. Cui, S. Jin, K. Yang, and J. An, “Optimal node placement andresource allocation for UAV relaying network,” IEEE CommunicationsLetters, vol. 22, no. 4, pp. 808–811, 2018.

[22] U. Challita, W. Saad, and C. Bettstetter, “Deep reinforcement learningfor interference-aware path planning of cellular-connected UAVs,” inIEEE International Conference on Communications (ICC), 2018.

[23] X.-Y. Liu, Z. Ding, S. Borst, and A. Walid, “Deep reinforcement learningfor intelligent transportation systems,” in NeurIPS Workshop on MachineLearning for Intelligent Transportation Systems, 2018.

[24] A. Al-Hourani, S. Kandeepan, and S. Lardner, “Optimal LAP altitudefor maximum coverage,” IEEE Wireless Communications Letters (WCL),vol. 3, no. 6, pp. 569–572, 2014.

[25] M. Mozaffari, W. Saad, M. Bennis, and M. Debbah, “Unmanned aerialvehicle with underlaid device-to-device communications: Performanceand tradeoffs,” IEEE Transactions on Wireless Communications (TWC),vol. 15, no. 6, pp. 3949–3963, 2016.

[26] D. Oehmann, A. Awada, I. Viering, M. Simsek, and G. P. Fettweis,“SINR model with best server association for high availability studiesof wireless networks,” IEEE Wireless Communications Letters (WCL),vol. 5, no. 1, pp. 60–63, 2015.

[27] N. Gupta and V. A. Bohara, “An adaptive subcarrier sharing schemefor ofdm-based cooperative cognitive radios,” IEEE Transactions onCognitive Communications and Networking (TCCN), vol. 2, no. 4,pp. 370–380, 2016.

[28] P. Ramezani and A. Jamalipour, “Throughput maximization in dual-hop wireless powered communication networks,” IEEE Transactions onVehicular Technology (TVT), vol. 66, no. 10, pp. 9304–9312, 2017.

[29] M. Agiwal, A. Roy, and N. Saxena, “Next generation 5G wirelessnetworks: A comprehensive survey,” IEEE Communications Surveys &Tutorials, vol. 18, no. 3, pp. 1617–1655, 2016.

[30] Q. Wu and R. Zhang, “Common throughput maximization in uav-enabled ofdma systems with delay consideration,” IEEE Transactionson Communications (TOC), vol. 66, no. 12, pp. 6614–6627, 2018.

[31] S. Zhang, Y. Zeng, and R. Zhang, “Cellular-enabled uav communication:Trajectory optimization under connectivity constraint,” in IEEE Interna-tional Conference on Communications (ICC), pp. 1–6, 2018.

[32] R. S. Sutton and A. G. Barto, Reinforcement learning: An introduction.MIT press, 2018.

[33] C. J. Watkins and P. Dayan, “Q-learning,” Springer Machine learning,vol. 8, no. 3-4, pp. 279–292, 1992.

[34] C. Wirth and G. Neumann, “Model-free preference-based reinforcementlearning,” in AAAI Conference on Artificial Intelligence, 2016.

[35] H. Van Hasselt, A. Guez, and D. Silver, “Deep reinforcement learningwith double q-learning,” in AAAI Conference on Artificial Intelligence,2016.

[36] T. P. Lillicrap, J. J. Hunt, A. Pritzel, N. Heess, T. Erez, Y. Tassa,D. Silver, and D. Wierstra, “Continuous control with deep reinforcementlearning,” in International Conference on Learning Representations(ICLR), 2016.

[37] D. Silver, J. Schrittwieser, K. Simonyan, I. Antonoglou, A. Huang,A. Guez, T. Hubert, L. Baker, M. Lai, A. Bolton, et al., “Masteringthe game of go without human knowledge,” Nature, vol. 550, no. 7676,p. 354, 2017.

[38] V. Mnih, K. Kavukcuoglu, D. Silver, A. Graves, I. Antonoglou, D. Wier-stra, and M. Riedmiller, “Playing atari with deep reinforcement learn-ing,” https://arxiv.org/pdf/1312.5602, 2013.

[39] A. Daniely, “SGD learns the conjugate kernel class of the network,” inAdvances in Neural Information Processing Systems (NIPS), pp. 2422–2430, 2017.

Page 15: Ming Zhu , Xiao-Yang Liu , and Xiaodong Wang

IEEE TRANSACTIONS ON INTELLIGENT TRANSPORTATION SYSTEMS, VOL. XX, NO. XX, XXX 20XX 15

[40] Z. Wang, V. Aggarwal, and X. Wang, “Joint energy-bandwidth allocationin multiple broadcast channels with energy harvesting,” IEEE Transac-tions on Communications (TOC), vol. 63, no. 10, pp. 3842–3855, 2015.

[41] “Homepage of DJI Mavic Air (2019).”https://www.dji.com/cn/mavic-air?site=brandsite&from=nav.

[42] G. Lefebvre, M. Lebreton, F. Meyniel, S. Bourgeois-Gironde, andS. Palminteri, “Behavioural and neural characterization of optimisticreinforcement learning,” Nature Human Behaviour, vol. 1, no. 4, p. 0067,2017.

[43] M. Abadi, P. Barham, J. Chen, Z. Chen, A. Davis, J. Dean, M. Devin,S. Ghemawat, G. Irving, M. Isard, et al., “Tensorflow: a system for large-scale machine learning,” in USENIX Symposium on Operating SystemsDesign and Implementation (OSDI), pp. 265–283, 2016.

[44] M. Mozaffari, W. Saad, M. Bennis, and M. Debbah, “Drone small cellsin the clouds: design, deployment and performance analysis,” in IEEEGlobal Communications Conference (GLOBECOM), pp. 1–6, 2015.

[45] “Python Markov decision process (MDP) Toolbox (2019).”https://pymdptoolbox.readthedocs.io/en/latest/api/mdptoolbox.html.

Ming Zhu received the Ph.D. degree in ComputerScience and Engineering in Shanghai Jiao TongUniversity, Shanghai, China. He is now a Post-Doctoral Researcher and an Assistant Professor inShenzhen Institutes of Advanced Technology, Chi-nese Academy of Sciences, Shenzhen, China.

His research interests are in the area of big data,intelligent transportation systems, smart cities, andartificial intelligence.

Xiao-Yang Liu received the B.Eng. degree in Com-puter Science from Huazhong University of Scienceand Technology, and the PhD degree in the Depart-ment of Computer Science and Engineer, ShanghaiJiao Tong University, China. He is currently a PhD inthe Department of Electrical Engineering, ColumbiaUniversity.

His research interests include tensor theory, deeplearning, non-convex optimization, big data analysisand IoT applications.

Xiaodong Wang (S’98-M’98-SM’04-F’08) receivedthe Ph.D. degree in electrical engineering fromPrinceton University. He is currently a Professorof electrical engineering with Columbia University,New York NY, USA. His research interests fall inthe general areas of computing, signal processing,and communications. He has authored extensivelyin these areas. He has authored the book entitledWireless Communication Systems: Advanced Tech-niques for Signal Reception, (Prentice Hall, 2003).His current research interests include wireless com-

munications, statistical signal processing, and genomic signal processing.He has served as an Associate Editor of the IEEE TRANSACTIONS ONCOMMUNICATIONS, IEEE TRANSACTIONS ON WIRELESS COMMU-NICATIONS, IEEE TRANSACTIONS ON SIGNAL PROCESSING, andIEEE TRANSACTIONS ON INFORMATION THEORY. He is an ISI HighlyCited Author. He received the 1999 NSF CAREER Award, the 2001 IEEECommunications Society and Information Theory Society Joint Paper Award,and the 2011 IEEE Communication Society Award for Outstanding Paper onNew Communication Topics.