This paper was converted on www.awesomepapers.org from LaTeX by an anonymous user.
Want to know more? Visit the Converter page.

UAV-enabled Collaborative Beamforming via Multi-Agent Deep Reinforcement Learning

Saichao Liu, Geng Sun*,  Jiahui Li*,  Shuang Liang,
Qingqing Wu,  Pengfei Wang,  and Dusit Niyato, 
This study is supported in part by the National Natural Science Foundation of China (62272194, 62172186), and in part by the Science and Technology Development Plan Project of Jilin Province (20230201087GX). (Corresponding authors: Geng Sun and Jiahui Li.) Saichao Liu is with the College of Computer Science and Technology, Jilin University, Changchun 130012, China (e-mail: [email protected]). Geng Sun is with the College of Computer Science and Technology, Jilin University, Changchun 130012, China, and with Key Laboratory of Symbolic Computation and Knowledge Engineering of Ministry of Education, Jilin University, Changchun 130012, China; he is also affiliated with the College of Computing and Data Science, Nanyang Technological University, Singapore 639798 (e-mail: [email protected]). Jiahui Li is with the College of Computer Science and Technology, Jilin University, Changchun 130012, China (e-mail: [email protected]). Shuang Liang is with the School of Information Science and Technology, Northeast Normal University, Changchun 130117, China (e-mail: [email protected]). Qingqing Wu is with the Department of Electronic Engineering, Shanghai Jiao Tong University, Shanghai 200240, China (e-mail: [email protected]). Pengfei Wang is with the School of Computer Science and Technology, Dalian University of Technology, Dalian 116024, China (e-mail: [email protected]). Dusit Niyato is with the College of Computing and Data Science, Nanyang Technological University, Singapore 639798 (email: [email protected]).
Abstract

In this paper, we investigate an unmanned aerial vehicle (UAV)-assistant air-to-ground communication system, where multiple UAVs form a UAV-enabled virtual antenna array (UVAA) to communicate with remote base stations by utilizing collaborative beamforming. To improve the work efficiency of the UVAA, we formulate a UAV-enabled collaborative beamforming multi-objective optimization problem (UCBMOP) to simultaneously maximize the transmission rate of the UVAA and minimize the energy consumption of all UAVs by optimizing the positions and excitation current weights of all UAVs. This problem is challenging because these two optimization objectives conflict with each other, and they are non-concave to the optimization variables. Moreover, the system is dynamic, and the cooperation among UAVs is complex, making traditional methods take much time to compute the optimization solution for a single task. In addition, as the task changes, the previously obtained solution will become obsolete and invalid. To handle these issues, we leverage the multi-agent deep reinforcement learning (MADRL) to address the UCBMOP. Specifically, we use the heterogeneous-agent trust region policy optimization (HATRPO) as the basic framework, and then propose an improved HATRPO algorithm, namely HATRPO-UCB, where three techniques are introduced to enhance the performance. Simulation results demonstrate that the proposed algorithm can learn a better strategy compared with other methods. Moreover, extensive experiments also demonstrate the effectiveness of the proposed techniques.

Index Terms:
Unmanned aerial vehicle, collaborative beamforming, energy efficiency, trust region learning, multi-agent deep reinforcement learning.

1 Introduction

Unmanned aerial vehicles (UAVs) are aircraft without pilots, which are flying by remote or autonomous control [1]. With the advantages of high mobility, low cost and line-of-sight (LoS) communications, UAVs can be applied in many wireless communication scenarios. For instance, a UAV can act as an aerial base station (BS) to provide emergency communication service for ground users when the ground BSs are unavailable[2, 3]. Moreover, a UAV can serve as a relay to improve the coverage and connectivity of wireless networks [4]. In addition, the UAVs can be regarded as the new aerial users which is able to deliver packages [5].

Despite the application prospect of UAVs being promising, many challenges still exist. Specifically, the limited battery capacity will affect the work efficiency of a UAV [6] [7], thus some relevant optimizations (e.g., flight trajectory [8] and mission execution time [9]) are necessary to prolong the working time of UAVs. Besides, a UAV is difficult to communicate with remote BSs due to its limited transmission power [10], and it cannot fly close to the BSs because of the finite energy capacity. Therefore, it is significant to improve the transmission ability of a UAV.

Collaborative beamforming (CB) has been justified to efficiently improve the transmission performance of distributed systems with limited resources [11]. Specifically, the UAVs can form a virtual antenna array (VAA), where each UAV is regarded as a single antenna element, then the UAV-assisted VAA (UVAA) can produce a high-gain mainlobe toward the remote BS by utilizing CB, such that achieving the long distance air-to-ground (A2G) communication between the UAVs and remote BS. To obtain an optimal beam pattern so that achieving a higher receiving rate of the BS, all UAVs need to fly to more appropriate positions and adjust their excitation current weights [12]. However, for communicating with different BSs, all UAVs need to adjust their positions constantly, which will consume extra energy. Hence, in addition to improving the transmission rate, the optimization to the motion energy consumption of UVAA is also required so as to prolong the lifespan of UVAA.

Previous studies usually optimize the communication performance or energy efficiency separately by using the traditional methods (e.g., convex optimization, or evolutionary computation). However, it is difficult to achieve optimal performance of these two optimization objectives simultaneously because they conflict with each other. Furthermore, since the cooperation among UAVs is very complex and the number of UAVs is usually large in the UVAA system, traditional methods will take much time to compute the optimization results. Besides, the system is dynamic, causing the currently obtained optimization results will become invalid when the system status changes. Therefore, it is inappropriate to apply traditional methods to simultaneously optimize the two objectives of the UVAA system.

Deep reinforcement learning (DRL) is an effective approach to solve complex and dynamic optimization problems without prior knowledge, and it has been proven to be an effective method to deal with complex optimization problems with high-dimensional continuous spaces [13]. Several studies have applied DRL to solve single-UAV communication optimization problems [14] [15]. However, DRL methods usually consider only single-agent learning frameworks, which are inappropriate for dealing with the UAV-enabled CB optimization problem. Thus, multi-agent deep reinforcement learning (MADRL) can be regarded as a potential approach to obtain the optimal policies since the learned policy of each agent involves the cooperation with other agents in MADRL. Specifically, all agents use the deep neural network (DNN) for policy learning. Benefiting from the powerful learning ability of DNN, all agents can learn excellent strategies. Meanwhile, compared with traditional methods, all agents can make corresponding decision directly without consuming much computing resource when the environment changes.

In this work, we utilize MADRL to deal with the UAV-enabled CB optimization problem. Consider that the cooperation is very complex among UAVs, we use the heterogeneous-agent trust region policy optimization (HATRPO) [16] as the main MADRL framework. Besides, the number of UAVs is usually large in the UVAA, which may cause the curse of dimensionality for MADRL algorithm. Furthermore, the policy gradient estimation problem may exists. Therefore, we propose an improved algorithm for solving our constructed problem well. The main contributions are summarized as follows:

  • UAV-enabled CB Multi-objective Optimization Formulation: We study a UAV-assistant A2G communication system, where a swarm of UAVs form a UVAA to perform CB for communicating with remote BSs. Then, we formulate a UAV-enabled CB multi-objective optimization problem (UCBMOP), aiming at simultaneously maximizing the transmission rate of UVAA and minimizing the energy consumption of the UAV elements by optimizing the positions and the excitation current weights of all UAVs. The problem is difficult to address because the two optimization objectives conflict with each other, and they are non-concave with respect to the optimization variables, such as the coordinates and the excitation current weights of UAVs.

  • New MADRL Approach: We model a multi-agent Markov decision process, called Markov game, for characterizing the formulated UCBMOP. Then, we propose an improved HATRPO algorithm, namely HATRPO-UCB, to solve the UCBMOP. Specifically, three techniques are proposed to enhance the performance of conventional HATRPO, which are observation enhancement, agent-specific global state and Beta distribution for policy, respectively. The first technique is to combine the algorithm with the problem well, so as to learn the better strategy for UCBMOP. The other two techniques are used to enhance the learning performance of critic and actor network, respectively.

  • Simulation and Performance Analysis: Simulation results show that the proposed HATRPO-UCB learns the best strategy compared with two classic antenna array solutions, three baseline MADRL algorithms and the conventional HATRPO. Besides, ablation experiments also demonstrate the effectiveness of our proposed techniques.

The rest of the paper is organized as follows. Section 2 discusses the related works. The system model and problem formulation are presented in Section 3. Section 4 proposes HATRPO-UCB. Simulation results and analysis are provided in Section 5. Finally, Section 7 concludes the paper.

2 Related Work

Several previous works have studied the energy optimization in UAV-based communication scenarios. For example, Zeng et al. [8] considered a scenario where a UAV needs to communicate with multiple ground nodes, then proposed an efficient algorithm to achieve the energy minimization by jointly optimizing the hovering locations and durations, as well as the flying trajectory connecting these hovering locations. Cai et al. [17] studied the energy-efficient secure downlink UAV communication problem, with the purpose of maximizing the system energy efficiency. The authors divided the optimization problem into two subproblems, and proposed an alternating optimization-based algorithm to jointly optimize the resource allocation strategy, the trajectory of information UAV, and the jamming policy of jammer UAV. Wang et al. [18] investigated a multiple UAVs-assisted mobile edge computing (MEC) system, where a natural disaster has damaged the edge server. The authors proposed two game theory-based algorithms, where the first algorithm is used to minimize the total energy consumption of multiple UAVs, and the second one is designed to achieve the utility trade-off between the UAV-MEC server and mobile users.

Except for adopting traditional methods, some works also use DRL to optimize the energy consumption of UAV. For optimizing the energy consumption of cellular-connected UAV, Zhan et al. [9] developed a DRL algorithm based on dueling deep Q network (DQN) with multi-step double Q-learning to jointly optimize the mission completion time, trajectory of UAV and associations to BSs. Abedin et al. [19] propose an agile deep reinforcement learning algorithm to achieve an energy-efficient trajectory optimization for multiple UAVs so as to improve the data freshness and connectivity to the Internet of Things (IoT) devices. Liu et al. [20] proposed a decentralized DRL-based framework to control multiple UAVs for achieving the long-term communication coverage in an area. The goal is to simultaneously maximize the temporal average coverage score achieved by all UAVs in a task, maximize the geographical fairness of all considered point-of-interests (PoIs), and minimize the total energy consumption.

UAV communications enabled by CB are also investigated in previous works. For example, Garza et al. [12] proposed the differential evolution method for a UAV-based 3D antenna array, aiming at achieving a maximum performance with respect to directivity and sidelobe level (SLL). Mozaffari et al. [21] considered a UAV-based linear antenna array (LAA) communication system and proposed two optimization algorithms, where the first algorithm is used to minimize the transmission time for ground users, and the second one is designed to optimize the control time of UAV-based LAA. Sun et al. [22] developed a multi-objective optimization algorithm to simultaneously minimize the total transmission time, total performing time of UVAA, and total motion and hovering energy consumption of UAVs. Moreover, the authors in [23] studied a secure and energy-efficient relay communication problem, where the UVAA serves as an aerial relay for providing communications to the blocked or low-quality terrestrial networks. The authors designed an improved evolutionary computation method to jointly circumvent the effects of the known and unknown eavesdroppers and minimize the propulsion energy consumption of UAVs.

Likewise, Sun et al[24] considered a UAV-enabled secure communication scenario, where multiple UAVs form a UVAA to perform CB for communicating with the remote BSs, and multiple known and unknown eavesdroppers exist to wiretap the information. Li et al[25] studied the UAV-assisted IoT and introduced CB into IoTs and UAVs simultaneously to achieve energy and time-efficient data harvesting and dissemination from multiple IoT clusters to remote BSs. Huang et al[26] studied a dual-UAVs jamming-aided system to implement physical layer encryption in maritime wireless communication. Li et al[27] considered a UAV-enabled data collection scenario, where CB is adopted to mitigate the interference of the LoS channel of UAV to the terrestrial network devices. Moorthy et al[28] aimed at achieving the high data rate of CB of UAV swarm, and then designed a distributed solution algorithm based on a combination of echo state network learning and online reinforcement learning to maximize the throughput of UAV swarm. However, all of these abovementioned methods will take much time to compute the solution for a task. More seriously, when the task changes, the previous optimization solution is invalid, causing a longer latency for recalculation.

DRL has been an effective way to solve the online computation problems about UAV communications. For example, Wang et al. [14] investigated a UAV-assisted IoT system, and they proposed a twin-delayed deep deterministic policy gradient (TD3)-based method that optimizes the trajectory of UAV to efficiently collect data from IoT ground nodes. Xie et al. [15] proposed a multi-step dueling double DQN (multi-step D3QN)-based method to optimize the trajectory of cellular-connected UAV while guaranteeing the stable connectivity of UAV to the cellular network during its flight. Besides, MADRL can be also applied to solve the communication problems involving multiple UAVs. For instance, Zhang et al. [29] proposed a continuous action attention multi-agent deep deterministic policy gradient (CAA-MADDPG) algorithm to operate a UAV transmitter to communicate with ground users and multiple UAV jammers. Dai et al. [30] studied the joint decoupled uplink-downlink association and trajectory design problem for full-duplex UAV network, with the purpose of maximizing the sum-rate of user equipments in both uplink and downlink. The authors proposed a MADRL-based approach and developed an improved clip and count-based proximal policy optimization (PPO) algorithm. However, in the above works, the cooperation pattern among UAVs is only the coordination, which is less complex than CB, and the completion of CB involves more UAVs. Hence, these abovementioned MADRL algorithms are not suitable for solving our formulated optimization problem.

In summary, different from the above works, we investigate a joint problem about communication performance and energy consumption under a scenario where multiple UAVs perform CB. Then, the MADRL is adopted to efficiently solve our formulated optimization problem.

Refer to caption
Figure 1: Sketch map of the considered UVAA communication system.

3 System Model and Problem Formulation

As shown in Fig. 1, a UAV-assisted A2G wireless communication system is considered. Specifically, there are NN UAVs flying in a square area with length LL, where the set of UAVs is denoted as 𝒩{i=1,2,,N}\mathcal{N}\triangleq\{i=1,2,\dots,N\}. Some BSs are distributed in the far-field region, and their positions are assumed to be known by all UAVs. Each UAV is equipped with a single omni-directional antenna, and they will perform the communication tasks with these BSs. For example, their stored data or some generated emergency data needs to be uploaded to BSs. However, each UAV is not able to support such the long-distance communication due to the limited transmission ability, and they also can not fly to these BSs because of their own limited battery energy capacity. Note that for all UAVs, the distances among them are much shorter than those to BSs, which means that they are close to each other. Thus, all UAVs can fly closer to form an UVAA for performing CB to complete communication tasks. Moreover, the UVAA only communicates with one BS at each time as performing CB.

In the considered system, the 3D Cartesian coordinate system is utilized. Specifically, the positions of the ii-th UAV and the target BS are denoted as (xi,yi,zi)(x_{i},y_{i},z_{i}) and (xBS,yBS,0)(x_{\mathrm{BS}},y_{\mathrm{BS}},0), respectively. Furthermore, the elevation and azimuth angles from each UAV to the target BS can be computed by θi=cos1(0zi(xBSxi)2+(yBSyi)2+(0zi)2)\theta_{i}=\cos^{-1}(\frac{{0-z_{i}}}{{\sqrt{(x_{\mathrm{BS}}-x_{i})^{2}+(y_{\mathrm{BS}}-y_{i})^{2}+(0-z_{i})^{2}}}}) and ϕi=sin1(yBSyi(xBSxi)2+(yBSyi)2)\phi_{i}=\sin^{-1}(\frac{y_{\mathrm{BS}}-y_{i}}{\sqrt{(x_{\mathrm{BS}}-x_{i})^{2}+(y_{\mathrm{BS}}-y_{i})^{2}}}), respectively.

TABLE I: Summary of main notations
Notation Definition
Notation used in system model
AF(θ,ϕ)AF(\theta,\phi) Array factor
BB Transmission bandwidth
dBSd_{\mathrm{BS}} Distance from origin of UVAA to target BS
dmind_{min} Minimal distance between two UAVs
EE Total energy consumption of UAV
EGE_{G} Energy consumed by gravity during UAV flight
EKE_{K} Energy consumption from kinetic energy during UAV flight
fcf_{c} Carrier frequency
G()G(\cdot) Array gain of UVAA
gcg_{c} Channel power gain
HminH_{min} Minimum flight altitude of UAV
HmaxH_{max} Maximum flight altitude of UAV
IiI_{i} Excitation current weight of the ii-th UAV
LL Length of monitor area
mUAVm_{\mathrm{UAV}} Mass of UAV
NN Number of UAVs
PLoSP_{\mathrm{LoS}} LoS probability
PNLoSP_{\mathrm{NLoS}} NLoS probability
PtP_{t} Total transmit power of UVAA
R()R(\cdot) Transmission rate
w(θ,ϕ)w(\theta,\phi) Magnitude of far-field beam pattern of UVA element
α\alpha Path loss exponent
η\eta Antenna array efficiency
θBS\theta_{\mathrm{BS}} Elevation angle between UVAA and target BS
λ\lambda Wavelength
μLoS\mu_{\mathrm{LoS}} Attenuation factor for LoS link
μNLoS\mu_{\mathrm{NLoS}} Attenuation factor for NLoS link
σ2\sigma^{2} Noise power
ϕBS\phi_{\mathrm{BS}} Azimuth angle between UVAA and target BS
Ψi\Psi_{i} Initial phase of the ii-th UAV
Notation used in reinforcement learning
aia_{i} Action of the ii-th agent
𝒂\bm{a} Joint action of all agents
𝒜i\mathcal{A}_{i} action space of the ii-th agent
A𝝅(st,𝒂t)A^{\bm{\pi}}(s^{t},\bm{a}^{t}) Advantage function for joint policy 𝝅\bm{\pi}
DKLmax(,)\mathrm{D^{max}_{KL}}(\cdot,\cdot) Maximal KL-divergence between two policies
𝒈ik\bm{g}^{k}_{i} Gradient of maximization objective of the ii-th agent
𝑯ik\bm{H}^{k}_{i} Hessian of expected KL-divergence of the ii-th agent
J(𝝅)J(\bm{\pi}) Expected total reward for all agents
L1:i𝝅(,)L^{\bm{\pi}}_{1:i}(\cdot,\cdot) Surrogate equation of the ii-th agent
oio_{i} Local observation of the ii-th agent
𝒪i\mathcal{O}_{i} Observation space of the ii-th agent
Q𝝅(st,𝒂t)Q^{\bm{\pi}}(s^{t},\bm{a}^{t}) State-action value function for joint policy 𝝅\bm{\pi}
rir_{i} Reward of the ii-th agent
ss Global state
𝒮\mathcal{S} Global state space
V𝝅(st)V^{\bm{\pi}}(s^{t}) State value function for joint policy 𝝅\bm{\pi}
αj\alpha^{j} A coefficient is found by backtracking line search
γ\gamma Discount factor
δ\delta KL constraint threshold
ρ𝝅\rho_{\bm{\pi}} Marginal state distribution for joint policy 𝝅\bm{\pi}

3.1 UVAA Communication Model

The goal of UVAA is to enable multiple UAVs to simulate the classical beamforming for achieving the far-field communication with remote BSs. The gain towards the target BS will be enhanced by the superposition of electromagnetic waves emitted from all UAVs through CB. Before the CB-enabled communication, all UAVs can be synchronized in terms of the carrier frequency, time and initial phase [31], and then send the same data to the target BS.

Benefiting from the high altitude of UAV, a higher LoS probability between the UVAA and BS can be achieved. Additionally, the LoS probability depends also on the propagation environment and the statistic modeling of the building density. Hence, the LoS probability can be modeled as [32]

PLoS=11+Cexp(D[θrecC]),P_{\mathrm{LoS}}=\frac{1}{1+C\operatorname{exp}(-D[\theta_{\mathrm{rec}}-C])}, (1)

where CC and DD are the parameters that depend on the propagation environment. θrec\theta_{\mathrm{rec}} is the elevation angle from the UVAA to the target BS. Then, the non-LoS (NLoS) probability is given by PNLoS=1PLoSP_{\mathrm{NLoS}}=1-P_{\mathrm{LoS}}.

According to the probability of LoS and NLoS, the channel power gain can be expressed as

gc=Ko1dBSα[PLoSμLoS+PNLoSμNLoS]1,g_{c}=K^{-1}_{o}d^{-\alpha}_{\mathrm{BS}}[P_{\mathrm{LoS}}\mu_{\mathrm{LoS}}+P_{\mathrm{NLoS}}\mu_{\mathrm{NLoS}}]^{-1}, (2)

where Ko=(4πfcc)2K_{o}=(\frac{4\pi f_{c}}{c})^{2}, dBSd_{\mathrm{BS}} is the distance from the origin of UVAA to the target BS, μLoS\mu_{\mathrm{LoS}} and μNLoS\mu_{\mathrm{NLoS}} are different attenuation factors for the LoS and NLoS links, respectively, α\alpha is a constant representing the path loss exponent, fcf_{c} is the carrier frequency, and cc is the speed of light.

Combining the above channel model, the UVAA will perform CB to generate a beam pattern with a sharp mainlobe for the target BS. Then, the transmission rate can be expressed as [21]

R(𝐗,𝐘,𝐙,𝐈)=Blog2(1+gcPtG(𝐗,𝐘,𝐙,𝐈)σ2),R(\bm{\mathrm{X}},\bm{\mathrm{Y}},\bm{\mathrm{Z}},\bm{\mathrm{I}})=B\log_{2}\left(1+\frac{g_{c}P_{t}G(\bm{\mathrm{X}},\bm{\mathrm{Y}},\bm{\mathrm{Z}},\bm{\mathrm{I}})}{\sigma^{2}}\right), (3)

where 𝐗={xi}i=1N,𝐘={yi}i=1N,𝐙={zi}i=1N,𝐈={Ii}i=1N\bm{\mathrm{X}}=\{x_{i}\}^{N}_{i=1},\bm{\mathrm{Y}}=\{y_{i}\}^{N}_{i=1},\bm{\mathrm{Z}}=\{z_{i}\}^{N}_{i=1},\bm{\mathrm{I}}=\{I_{i}\}^{N}_{i=1} represent the 3D coordinates and excitation current weights of all UAVs while communicating with the BS, BB is the transmission bandwidth, PtP_{t} is the total transmit power of UVAA, and σ2\sigma^{2} is the noise power. In addition, G(𝐗,𝐘,𝐙,𝐈)G(\bm{\mathrm{X}},\bm{\mathrm{Y}},\bm{\mathrm{Z}},\bm{\mathrm{I}}) is the array gain of UVAA toward the location of BS, which is given by [21]

G(𝐗,𝐘,𝐙,𝐈)=4π|AF(θBS,ϕBS)|2w(θBS,ϕBS)202π0π|AF(θ,ϕ)|2w(θ,ϕ)2sinθdθdϕη,\displaystyle G(\bm{\mathrm{X}},\bm{\mathrm{Y}},\bm{\mathrm{Z}},\bm{\mathrm{I}})=\frac{4\pi|AF(\theta_{\mathrm{BS}},\phi_{\mathrm{BS}})|^{2}w(\theta_{\mathrm{BS}},\phi_{\mathrm{BS}})^{2}}{\displaystyle\int_{0}^{2\pi}\int_{0}^{\pi}|AF(\theta,\phi)|^{2}w(\theta,\phi)^{2}\sin{\theta}d\theta d\phi}\eta, (4)

where η[0,1]\eta\in[0,1] is the antenna array efficiency, (θBS,ϕBS)(\theta_{\mathrm{BS}},\phi_{\mathrm{BS}}) represents the direction toward the target BS, and w(θ,ϕ)w(\theta,\phi) is the magnitude of the far-field beam pattern of each UAV element. Note that the array gain is equal to the array directivity multiplied by η\eta. The array directivity D=U(θBS,ϕBS)/U0D=U(\theta_{\mathrm{BS}},\phi_{\mathrm{BS}})/U_{0}, where U(θBS,ϕBS)=|AF(θBS,ϕBS)|2w(θBS,ϕBS)2U(\theta_{\mathrm{BS}},\phi_{\mathrm{BS}})=|AF(\theta_{\mathrm{BS}},\phi_{\mathrm{BS}})|^{2}w(\theta_{\mathrm{BS}},\phi_{\mathrm{BS}})^{2} is the maximum radiation intensity produced by the array factor (AF) to the BS, and U0=1/4π02π0π|AF(θ,ϕ)|2w(θ,ϕ)2sinθdθdϕU_{0}=1/4\pi\int_{0}^{2\pi}\int_{0}^{\pi}|AF(\theta,\phi)|^{2}w(\theta,\phi)^{2}\sin{\theta}d\theta d\phi is the average radiation intensity. Besides, w(θ,ϕ)w(\theta,\phi) is always equal to 0 dB in this work because the omni-directional antenna equipped on the UAV has the identical power constraints.

As for the AF, it is an important index to describe the status of beam pattern, which is written as [12] [33]

AF(θ,ϕ)=i=1NIiejΨiej[2πλ(xisinθcosϕ+yisinθsinϕ+zicosθ)],\displaystyle AF(\theta,\phi)=\sum_{i=1}^{N}I_{i}e^{j\Psi_{i}}e^{j[\frac{2\pi}{\lambda}(x_{i}\sin{\theta}\cos{\phi}+y_{i}\sin{\theta}\sin{\phi}+z_{i}\cos{\theta})]}, (5)

where θ[0,π]\theta\in[0,\pi] and ϕ[π,π]\phi\in[-\pi,\pi] are the elevation and azimuth angles, respectively. λ\lambda is the wavelength and 2πλ\frac{2\pi}{\lambda} denotes the wave number. Additionally, IiI_{i} and Ψi\Psi_{i} are the excitation current weight and initial phase of the ii-th UAV, where the phase synchronization is completed by compensating the distance between an UAV and the origin of UVAA [34]. Hence, Ψi\Psi_{i} is expressed as

Ψi=\displaystyle\Psi_{i}= (6)
2πλ(xisinθBScosϕBS+yisinθBSsinϕBS+zicosθBS).\displaystyle-\frac{2\pi}{\lambda}(x_{i}\sin{\theta_{\mathrm{BS}}}\cos{\phi_{\mathrm{BS}}}+y_{i}\sin{\theta_{\mathrm{BS}}}\sin{\phi_{\mathrm{BS}}}+z_{i}\cos{\theta_{\mathrm{BS}}}).

As can be seen, the locations and excitation current weights of UAVs are directly related to the AF.

3.2 UAV Energy Consumption Model

In this work, we adopt a typical rotary-wing UAV. This is due to the fact that rotary-wing UAVs have the inherent benefit of hovering capabilities, which simplifies the difficulty of beam alignment and handling Doppler effects during transmission. In this work, the considered communication system is required to provide a stable and high-rate communication service for the far-field BSs. In this case, rotary-wing UAVs are more suitable for the considered scenario since they can hover for collecting and forwarding data. In general, the total energy consumption of a UAV includes two main components that are the propulsion and communication [8]. Specifically, the propulsion energy consumption is related to the hovering and movement of the UAVs, and communication-related energy consumption is used for signal processing, circuits, transmitting and receiving, respectively. Typically, the propulsion energy consumption is much more than communication-related energy consumption (e.g., hundreds of watts (w) versus a few w) [8] [35]. Thus, the communication-related energy consumption is neglected in this work. Then, the propulsion power consumption of UAV with speed vv during the straight and horizontal flight is calculated as [8]

P(v)=\displaystyle P(v)= PB(1+3v2vtip2)+PI(1+v44v04v22v02)1/2\displaystyle P_{B}\left(1+\frac{3v^{2}}{v^{2}_{tip}}\right)+P_{I}\left(\sqrt{1+\frac{v^{4}}{4v^{4}_{0}}}-\frac{v^{2}}{2v^{2}_{0}}\right)^{1/2} (7)
+12d0ρsAv3,\displaystyle+\frac{1}{2}d_{0}\rho sAv^{3},

where PBP_{B} and PIP_{I} are blade profile power and induced power under the hovering condition, respectively. vtipv_{tip} is the tip speed of rotor blade, and v0v_{0} is mean rotor induced velocity in hovering. d0d_{0}, ρ\rho, ss, and AA denote the fuselage drag ratio, air density, rotor solidity and rotor disc area, respectively. Besides, the additional/less energy consumption caused by the acceleration/deceleration of horizontal flight is ignored because it only takes a small proportion of the total operation time of UAV manoeuvring duration [8].

Refer to caption
Figure 2: Energy consumption of UAV in the horizontal and vertical flights under different speeds.

For arbitrary 3D UAV trajectory with UAV climbing and descending over time, the heuristic closed-form approximation of energy consumption is derived as [1]

E(T)0TP(v(t))𝑑t+12\displaystyle E(T)\approx\int_{0}^{T}P(v(t))dt+\frac{1}{2} mUAV(v(T)2v(0)2)\displaystyle m_{\mathrm{UAV}}(v(T)^{2}-v(0)^{2}) (8)
+mUAVg(h(T)h(0)),\displaystyle+m_{\mathrm{UAV}}g(h(T)-h(0)),

where TT is the total flight time. v(t)v(t) is the instantaneous UAV speed at time tt. The second and third parts represent the kinetic energy consumption and potential energy consumption, respectively. mUAVm_{\mathrm{UAV}} is the mass of UAV, and gg is the gravitational acceleration. As can be seen, Eq. (8) has more terms than Eq. (7), meaning that more energy will be consumed in the vertical direction. Moreover, Fig. 2 shows that the motion energy consumption of UAV per second in the vertical directioon is more than that in the horizontal direction under different speeds.

In this work, the UAV follows the flight rule that it first flies along the horizontal direction, then in the vertical direction, which is a common adopted flight strategy [36][37][38]. Therefore, when a UAV reaches the target position, the corresponding energy consumption can be expressed as

E=PH(vH)tH+PC(vC)tC+PD(vD)tD,\displaystyle E=P_{H}(v_{H})t_{H}+P_{C}(v_{C})t_{C}+P_{D}(v_{D})t_{D}, (9)

where tH=Δd/vHt_{H}=\Delta d/v_{H}, tC=|ΔhC|/vCt_{C}=|\Delta h_{C}|/v_{C} and tD=|ΔhD|/vDt_{D}=|\Delta h_{D}|/v_{D}. {PH,PC,PD}\{P_{H},P_{C},P_{D}\} are the horizontal, climbing and descending flight powers, respectively, and {Δd,ΔhC,ΔhD}\{\Delta d,\Delta h_{C},\Delta h_{D}\} are the corresponding flying distance during the horizontal, climbing and descending flights. Furthermore, {vH,vC,vD}\{v_{H},v_{C},v_{D}\} represent the maximum-endurance (ME) speed of the horizontal, climbing and descending flights, respectively, which are the optimal UAV speed that maximizes the UAV endurance with any given onboard energy [1].

Remark 1.

The challenge of adopting this type of rotary-wing UAVs lies in their weak transmission abilities and limited service time. To solve this issue, we correspondingly introduce the CB and give the problem formulation as follows.

3.3 Problem Formulation

In the system, the UVAA is required to communicate with more than one BS before the battery runs out. To achieve the maximum transmission performance, all UAVs need to fly to better locations and adjust their excitation current weights. However, the motion energy consumption of UAVs will be increased during their movements, which will reduce the lifespan of UVAA. Therefore, the optimization problems about transmission rate and energy consumption should be considered simultaneously.

In this work, we aim to maximize the transmission rate of UVAA for communicating with the BS and minimize the motion energy consumption of all UAVs simultaneously, through optimizing the hovering positions and excitation current weights of all UAVs. Then, the UCBMOP can be formulated as

max𝐗,𝐘,𝐙,𝐈\displaystyle\mathop{\operatorname{max}}\limits_{\bm{\mathrm{X}},\bm{\mathrm{Y}},\bm{\mathrm{Z}},\bm{\mathrm{I}}} f={RT,Etotal}\displaystyle f=\{R_{T},-E_{total}\} (10a)
s.t. C1:L/2xiL/2,i𝒩,\displaystyle C1:-L/2\leq x_{i}\leq L/2,\forall i\in\mathcal{N}, (10b)
C2:L/2yiL/2,i𝒩,\displaystyle C2:-L/2\leq y_{i}\leq L/2,\forall i\in\mathcal{N}, (10c)
C3:HminziHmax,i𝒩,\displaystyle C3:H_{min}\leq z_{i}\leq H_{max},\forall i\in\mathcal{N}, (10d)
C4:0Ii1,i𝒩,\displaystyle C4:0\leq I_{i}\leq 1,\forall i\in\mathcal{N}, (10e)
C5:di,jdmin,i,j𝒩,\displaystyle C5:d_{i,j}\geq d_{min},\forall i,j\in\mathcal{N}, (10f)

where RTR_{T} is the transmission rate of UVAA to a specific BS defined by Eq. (3). The motion energy consumption of all UAVs is Etotal=i=1NEiE_{total}=\sum_{i=1}^{N}E_{i}, where EiE_{i} is the motion energy consumption of UAV ii that can be obtained according to the Eq. (9). Constraints C1, C2 and C3 together specify the flight area of UAV, where LL is the length of monitor area, and HminH_{min} and HmaxH_{max} are altitude constraints of the UAVs. Constraint C4 specifies the adjustment range of excitation current weight. Moreover, Constraint C5 requires any two adjacent UAVs to keep a certain distance to avoid the collision, where dmind_{min} is set as the minimum distance.

4 MADRL-Based UAV-enabled Collaborative Beamforming

The formulated UCBMOP is challenging and cannot be solved by traditional optimization methods, and the reasons are as follows:

  • UCBMOP is a non-convex and long-term optimization problem. Specifically, due to the complex non-linear relationship between UAV coordinates and excitation current weights with antenna gain, the optimization of the virtual antenna array transmission performance is a non-convex optimization problem. Meanwhile, UCBMOP involves long-term optimization objectives and needs an effective algorithm that can achieve long-term optimal solutions. As such, UCBMOP is challenging and cannot be solved by traditional convex optimization.

  • UCBMOP is an NP-hard problem. Specifically, UCBMOP can be simplified as a typical nonlinear multidimensional 0-1 knapsack problem that has been demonstrated to be NP-hard [39]. Therefore, the proposed UCBMOP is also NP-hard. Such NP-hard problems are challenging for traditional optimization methods to solve in polynomial time.

  • UCBMOP is a complex large-scale optimization problem. Specifically, the considered scenarios often involve a large number of UAVs, while each UAV also has multiple variables to be optimized (i.e., 3D coordinates and excitation current weights). In addition, the completion of CB also needs the complex cooperation of all UAVs. Hence, the complexity and massive dimension of decision variables make the traditional optimization method unable to solve it efficiently.

  • UCBMOP operates within a dynamic environment. In particular, UAVs in these scenarios must make decisions to adapt to highly dynamic channel conditions and the unpredictability of natural surroundings. Additionally, real-time response is crucial for UAVs, as prolonged hovering in the air results in significant energy consumption. Consequently, conventional offline algorithms or approaches reliant on temporary calculations cannot solve this problem.

Different from the traditional optimization methods, deep reinforcement learning can have real-time responses and handle the complexity and large-scale decision variables of UCBMOP. However, in the considered system, UAVs need to collaborate to achieve transmission gain while simultaneously competing to minimize their individual energy consumption. Such tasks are distributed and require multiple UAVs to work together to achieve common objectives. Standard reinforcement learning assumes individual agents to be isolated, making it less suitable for the scenario involving multiple decision-makers.

Remark 2.

In this case, MADRL has the capability to decentralize decision-making and adaptability for the changing environments, and enabling effective responses to complex interactions among multiple agents [40]. Moreover, MADRL allows agents to access the global state of the UAV collaborative beamforming system, providing them with information about the entire system and reducing non-stationarity. In addition, the scalability and robustness of MADRL to partial observability also enhance its applicability in the considered dynamic and distributed scenarios. Thus, we propose an MADRL-based method that is able to solve the formulated UCBMOP by learning the strategy in the considered dynamic scenario.

4.1 Markov Game for UCBMOP

In this section, the formulated UCBMOP is transformed into a Markov game [41] that is defined as 𝒩,𝒮,{𝒪i}i𝒩,{𝒜i}i𝒩,P,{Ri}i𝒩,γ\langle\mathcal{N},\mathcal{S},\{\mathcal{O}_{i}\}_{i\in\mathcal{N}},\{\mathcal{A}_{i}\}_{i\in\mathcal{N}},P,\{R_{i}\}_{i\in\mathcal{N}},\gamma\rangle. At time slot tt, all agents are with state st𝒮s^{t}\in\mathcal{S}, each agent uses its policy πθi\pi_{\theta_{i}} to take an action ait𝒜ia^{t}_{i}\in\mathcal{A}_{i} according to its local observation oit𝒪io^{t}_{i}\in\mathcal{O}_{i}, and it will receive a reward rit=Ri(st,𝒂t)r^{t}_{i}=R_{i}(s^{t},\bm{a}^{t}) from the environment, where 𝒂t={a1t,,aNt}\bm{a}^{t}=\{a^{t}_{1},\dots,a^{t}_{N}\} denotes the joint action. Then, all agents move to the next state st+1s^{t+1} with probability P(st+1|st,𝒂t)P(s^{t+1}|s^{t},\bm{a}^{t}) and each of them receives its local observation oit+1o^{t+1}_{i} correlated with the state. All agents aim to maximize their own expected total reward J(πθi)=𝔼s0,𝒂0,[t=0γtrit]J(\pi_{\theta_{i}})=\mathbb{E}_{s^{0},\bm{a}^{0},\dots}[\sum^{\infty}_{t=0}\gamma^{t}r^{t}_{i}]. The details about the state and observation space, the action space and the reward function are introduced as follows.

4.1.1 State and Observation Space

At each time slot, every agent observes its own local state information to make the decision. In the system, the local observation of any agents (i.e., UAVs) includes three parts that are its own related information, the relevant information of other UAVs and the related information of BS. Specifically, the local observation of the ii-th agent can be expressed as

oi\displaystyle o_{i} ={θi,ϕi,di,di,o,Ii,{dj,i,Ij}j𝒩,ji,x,y,z},\displaystyle=\{\theta_{i},\phi_{i},d_{i},d_{i,o},I_{i},\{d_{j,i},I_{j}\}_{j\in\mathcal{N},j\neq i},x^{\prime},y^{\prime},z^{\prime}\}, (11)

where (θi,ϕi,di)(\theta_{i},\phi_{i},d_{i}) is the spherical coordinate of the target BS relative to the ii-th UAV. However, when the UVAA will communicate with more than one BS, then the gap among different BSs regarding did_{i} is very large. Furthermore, all UAVs only fly in the fixed area. To eliminate the gap and let the UAV make decisions more efficiently, we compute a reference point in the monitor area that is closest to the target BS, where the coordinate of reference point is denoted as (x,y,z)(x^{\prime},y^{\prime},z^{\prime}). Then, the distance between the ii-th UAV and the reference point is regarded as did_{i}. Moreover, di,od_{i,o} is the distance between the ii-th UAV and the origin of UVAA, and IiI_{i} is the excitation current weight of the ii-th UAV. The information of other UAVs contains their own excitation current weights and the distances to the ii-th UAV. As for the information about the target BS, its coordinate is replaced by the reference point, which will be beneficial for the agent to learn the better strategy.

As for the global state of the ii-th agent sis_{i}, instead of using the concatenation of local observations or the environment-provided global state, we propose an agent-specific global state that combines some agent-related features (e.g., (θi,ϕi,di)(\theta_{i},\phi_{i},d_{i}), di,od_{i,o}, dj,id_{j,i}) and the global information provided from the environment which includes the coordinates and excitation current weights of all UAVs, as well as the coordinate of reference point, which is inspired from [42].

Refer to caption
Figure 3: The transmission rate of UVAA to BSs at different distances when performing CB communication. The parameters about LoS probability CC and DD are set as 10 and 0.6, respectively.

4.1.2 Action Space

According to the local observation of UAVs, each UAV will take an action to fly to a certain location with the optimized excitation current weight for performing CB. Then, the action of each agent can be defined by

ai={xi,yi,zi,Ii},\displaystyle a_{i}=\{x_{i},y_{i},z_{i},I_{i}\}, (12)

where (xi,yi,zi)(x_{i},y_{i},z_{i}) is the location to which the ii-th UAV is going to fly, and IiI_{i} is the excitation current weight of the ii-th UAV in the next time slot.

4.1.3 Reward Function

The reward function is used to evaluate how well an agent executes the actions. Specifically, a suitable reward function can make the agent learn the best strategy. UCBMOP involves two optimization objectives, which are maximizing the transmission rate of UVAA for communicating with the BS, i.e., RTR_{T}, and minimizing the motion energy consumption of all UAVs, i.e., EtotalE_{total}. To this end, we design a reward function by considering the following aspects:

  1. 1.

    Reward Function for Optimizing RTR_{T}. We find that there are three key components that affect the transmission rate RTR_{T}, which are the transmit power and gain of UVAA, transmission distance, and A2G communication angle. Specifically, the transmit power and gain of UVAA determine the transmission performance of UVAA, the transmission distance indicates the signal fading degree, and the air-to-ground communication angle affects the A2G channel condition. Note that we do not directly set the transmission rate as a reward since it changes significantly with the transmission distance, making the DRL method unstable and decreasing the applicability of the model.

    According to the analysis above, first, we regard the multiplication of the array gain and the total transmit powers as a reward, i.e., rTR=GPtr^{\mathrm{TR}}=GP_{t}. Note that this reward is shared by all agents. Then, we set the distance between the UAV and BS as a penalty for each agent. However, the UAV only flies within a fixed area, and then the aforementioned distance is replaced by the one between the UAV and the reference point of target BS, which is denoted as riU2B=di,rr^{\mathrm{U2B}}_{i}=-d_{i,r}. Finally, as shown in Eq. (1), the LoS probability is determined mainly by the height of the UAVs. Therefore, we set a reward to optimize the flight altitude of each UAV, denoted as riH=zir_{i}^{\mathrm{H}}=z_{i}. As such, the designed reward function considers all controllable components of the transmission rate, thereby achieving the optimization of the first objective.

  2. 2.

    Reward Function for Optimizing EtotalE_{total}. To achieve this goal, we regard the energy consumption of each UAV as a penalty, which is denoted as riEC=Eir_{i}^{\mathrm{EC}}=-E_{i}. As such, the optimization of the second objective can be achieved by the reward function.

  3. 3.

    Reward Function for Coordinating RTR_{T} and EtotalE_{total}. To this end, we aim to enhance the connectivity among UAVs. Compared with RTR_{T}, the optimization of EtotalE_{total} is more directly tied to the action space (i.e., reducing the movement of the UAVs), which may lead the algorithm to prioritize the optimization of EtotalE_{total} and under-optimize RTR_{T}. To overcome this issue, we try to concentrate the array elements (i.e., the UAVs in UVAA) on an applicable distance to achieve a higher CB performance [43] [23], thereby facilitating the optimization of RTR_{T}. Specifically, we design a reward to coordinate the UAVs by minimizing the cumulative distances between the current UAV and other UAVs, as well as between the current UAV and the origin of the UVAA, i.e.,

    riU2U=1κ(j𝒩,jidj,i+di,o),\displaystyle r^{\mathrm{U2U}}_{i}=-\frac{1}{\kappa}(\sum_{j\in\mathcal{N},j\neq i}d_{j,i}+d_{i,o}), (13)

    where κ\kappa is a large constant for scaling. Since this reward decreases the difficulty of optimizing RTR_{T}, the designed reward function can well coordinate the optimization processes of the two optimization objectives.

In summary, the whole reward function of each agent is

ri=ω1rTR+ω2riH+ω3riEC+ω4riU2B+ω5riU2U,\displaystyle r_{i}=\omega_{1}r^{\mathrm{TR}}+\omega_{2}r^{\mathrm{H}}_{i}+\omega_{3}r^{\mathrm{EC}}_{i}+\omega_{4}r^{\mathrm{U2B}}_{i}+\omega_{5}r^{\mathrm{U2U}}_{i}, (14)

where ω1\omega_{1}, ω2\omega_{2}, ω3\omega_{3}, ω4\omega_{4}, and ω5\omega_{5} represent weight coefficients of different parts. Moreover, it is also worth noting that when any two UAVs violate the minimum separation constraint, they will receive a negative reward. These weighting coefficient values can be determined by considering their value ranges, importance, and characteristics. Specifically, larger weights are assigned to rewards with wider ranges to appropriately reflect their significance in optimization. Moreover, higher weights are allocated to rewards that play a more critical role in achieving the optimization objectives. In addition, the weights can also be adjusted based on the unique characteristics of each reward, ensuring that the optimization process addresses their specific complexities.

Refer to caption
Figure 4: Flowchart of HATRPO-UCB algorithm for UAV-enabled CB.

4.2 HATRPO-UCB Algorithm

In this section, we propose a HATRPO-UCB, which extends from conventional HATRPO to find the solution strategy for UCBMOP. Specifically, each UAV acts as an agent that has an actor network and a critic network consisting of DNNs. The algorithm uses the sequential policy update scheme to train all agents. The scheme can make the update of the current agent includes the updates of previous agents, and thus our algorithm can make the actor networks of all agents learn the joint policy. Moreover, each agent has not only the individual rewards (i.e., riHr_{i}^{\mathrm{H}}, riECr_{i}^{\mathrm{EC}}, etc.) but also the shared rewards (i.e., rTRr^{\mathrm{TR}}). Meanwhile, we design an agent-specific global state that contains global information provided by the environment and some features from the local observation. As such, each UAV is guided by the shared rewards, and considers the global information and local observation, thereby obtaining a policy associated to the global state. We begin with introducing the conventional HATRPO.

4.2.1 Conventional HATRPO

HATRPO was proposed by Kuba et al. [16], which has the better performance than other baseline MADRL algorithms in many multi-agent tasks (e.g., Multi-Agent MuJoCo [44], StarCraftII Multi-Agent Challenge (SMAC) [45]). HATRPO applies the trust region learning [46] to MADRL successfully, achieving the monotonic improvement guarantee for joint policy of multiple agents such as the trust region policy optimization (TRPO) algorithm [47]. The detailed theory of HATRPO is introduced in the following.

In the beginning, a cooperative Markov game with NN agents exists. A joint policy 𝝅=(π1,,πN)\bm{\pi}=(\pi_{1},\dots,\pi_{N}) to all agents exists. At time slot tt, the agents are at state sts^{t}. Each agent utilizes its own policy πi\pi_{i} to take an action aita^{t}_{i}, and then the actions of all agents form the joint action 𝒂t=(a1t,,aNt)\bm{a}^{t}=(a^{t}_{1},\dots,a^{t}_{N}). Afterwards, all agents receive a joint reward rtr^{t}, and move to a new state st+1s^{t+1} with probability P(st+1|st,𝒂t)P(s^{t+1}|s^{t},\bm{a}^{t}). The goal of all agents is to maximize the expected total reward:

J(𝝅)=𝔼s0:ρ𝝅0:,𝒂0:𝝅[t=0γtrt],\displaystyle J(\bm{\pi})=\mathbb{E}_{s^{0:\infty}\sim\rho^{0:\infty}_{\bm{\pi}},\bm{a}^{0:\infty}\sim\bm{\pi}}\left[\sum^{\infty}_{t=0}\gamma^{t}r^{t}\right], (15)

where γ[0,1)\gamma\in[0,1) is the discount factor and ρ𝝅\rho_{\bm{\pi}} is the marginal state distribution [16]. Besides, the state value function and the state-action value function are defined as V𝝅(st)=𝔼𝒂t:𝝅,st+1:P[l=0γlrt+l]V^{\bm{\pi}}(s^{t})=\mathbb{E}_{\bm{a}^{t:\infty}\sim\bm{\pi},s^{t+1:\infty}\sim P}\left[\sum_{l=0}^{\infty}\gamma^{l}r^{t+l}\right] and Q𝝅(st,𝒂t)=𝔼st+1:P,𝒂t+1:𝝅[l=0γlrt+l]Q^{\bm{\pi}}(s^{t},\bm{a}^{t})=\mathbb{E}_{s^{t+1:\infty}\sim P,\bm{a}^{t+1:\infty}\sim\bm{\pi}}\left[\sum_{l=0}^{\infty}\gamma^{l}r^{t+l}\right]. The advantage function is written as A𝝅(st,𝒂t)=Q𝝅(st,𝒂t)V𝝅(st)A^{\bm{\pi}}(s^{t},\bm{a}^{t})=Q^{\bm{\pi}}(s^{t},\bm{a}^{t})-V^{\bm{\pi}}(s^{t}).

To extend the key idea of TRPO to MADRL, a sequential policy update scheme [16] is introduced. Then, the following inequality can be derived to achieve the monotonic improvement guarantee for joint policy.

J(𝝅¯)J(𝝅)+i=1N[L1:i𝝅(𝝅¯1:i1,π¯i)CDKLmax(πi,π¯i)].\displaystyle J(\bm{\bar{\pi}})\geq J(\bm{\pi})+\sum^{N}_{i=1}[L^{\bm{\pi}}_{1:i}(\bm{\bar{\pi}}_{1:i-1},\bar{\pi}_{i})-C\mathrm{D^{max}_{KL}}(\pi_{i},\bar{\pi}_{i})]. (16)

where 𝝅¯\bm{\bar{\pi}} is the next candidate joint policy. DKLmax(πi,π¯i)=maxsDKL(πi(|s),π¯(|s))\mathrm{D^{max}_{KL}}(\pi_{i},\bar{\pi}_{i})=\mathrm{max}_{s}\mathrm{D_{KL}}(\pi_{i}(\cdot|s),\bar{\pi}(\cdot|s)) is the maximal KL-divergence to measure the gap between two policies. C=4γmaxs,𝒂|A𝝅(s,𝒂)|(1γ)2C=\frac{4\gamma\mathrm{max}_{s,\bm{a}}|A^{\bm{\pi}}(s,\bm{a})|}{(1-\gamma)^{2}} is the penalty coefficient. Moreover, L1:i𝝅(𝝅¯1:i1,π¯i)L^{\bm{\pi}}_{1:i}(\bm{\bar{\pi}}_{1:i-1},\bar{\pi}_{i}) is a surrogate equation of the ii-th agent, which is expressed as

L1:i𝝅(\displaystyle L^{\bm{\pi}}_{1:i}( 𝝅¯1:i1,π¯i)\displaystyle\bm{\bar{\pi}}_{1:i-1},\bar{\pi}_{i}) (17)
=𝔼sρ𝝅,𝒂1:i1𝝅¯1:i1,aiπ¯i[Ai𝝅(s,𝒂1:i1,ai)],\displaystyle=\mathbb{E}_{s\sim\rho_{\bm{\pi}},\bm{a}_{1:i-1}\sim\bm{\bar{\pi}}_{1:i-1},a_{i}\sim\bar{\pi}_{i}}[A^{\bm{\pi}}_{i}(s,\bm{a}_{1:i-1},a_{i})],

where Ai𝝅(s,𝒂1:i1,ai)A^{\bm{\pi}}_{i}(s,\bm{a}_{1:i-1},a_{i}) is the local advantage function of each agent. Based on the above theory, all agents can update their policies sequentially, and the updating procedure is given as

πik+1=argmaxπi(L1:i𝝅k(𝝅1:i1k+1,πi)CDKLmax(πik,πi)).\displaystyle\pi^{k+1}_{i}=\underset{\pi_{i}}{\operatorname{arg\,max}}(L^{\bm{\pi}^{k}}_{1:i}(\bm{\pi}^{k+1}_{1:i-1},\pi_{i})-C\mathrm{D^{max}_{KL}}(\pi^{k}_{i},\pi_{i})). (18)

Furthermore, the sequential updating scheme does not require that the updating order of all agents is fixed, meaning that the updating order can be flexibly adjusted at each iteration. As such, the algorithm can adapt to dynamic changes in the environment and progressively assimilate new information in non-static conditions.

To implement the above procedure for parameterized joint policy 𝜽=(θ1,,θN)\bm{\theta}=(\theta_{1},\dots,\theta_{N}) in practice, in HATRPO, the amendment similar to TRPO is performed. As the maximal KL-divergence penalty DKLmax(πθik,πθi)\mathrm{D^{max}_{KL}}(\pi_{\theta^{k}_{i}},\pi_{\theta_{i}}) is hard to compute, it is replaced by the expected KL-divergence constraint 𝔼sρ𝝅𝜽k[DKL(πθik(|s),πθi(|s))]δ\mathbb{E}_{s\sim\rho_{\bm{\pi}_{\bm{\theta}^{k}}}}[\mathrm{D_{KL}}(\pi_{\theta^{k}_{i}}(\cdot|s),\pi_{\theta_{i}}(\cdot|s))]\leq\delta, where δ\delta is a threshold hyperparameter. Then, at the kk+11-th iteration, given a permutation, all agents can optimize their policy parameters sequentially according to the method as follows:

θik+1=\displaystyle\theta^{k+1}_{i}= (19)
argmaxθi𝔼sρ𝝅𝜽k,𝒂1:i1𝝅𝜽1:i1k+1,aiπθi[Ai𝝅𝜽k(s,𝒂1:i1,ai)],\displaystyle\underset{\theta_{i}}{\operatorname{arg\,max}}\mathbb{E}_{s\sim\rho_{\bm{\pi}_{\bm{\theta}^{k}}},\bm{a}_{1:i-1}\sim\bm{\pi}_{\bm{\theta}^{k+1}_{1:i-1}},a_{i}\sim\pi_{\theta_{i}}}[A^{\bm{\pi}_{\bm{\theta}^{k}}}_{i}(s,\bm{a}_{1:i-1},a_{i})],
s.t.𝔼sρ𝝅𝜽k[DKL(πθik(|s),πθi(|s))]δ.\displaystyle\operatorname{s.t.}\mathbb{E}_{s\sim\rho_{\bm{\pi}_{\bm{\theta}^{k}}}}[\mathrm{D_{KL}}(\pi_{\theta^{k}_{i}}(\cdot|s),\pi_{\theta_{i}}(\cdot|s))]\leq\delta.

For the computation of the above equation, similar to TRPO, the linear approximation to the objective function and the quadratic approximation to KL constraint are applied, derivating a closed-form updating scheme that is shown as

θik+1=θik+αj2δ𝒈ik(𝑯ik)1𝒈ik(𝑯ik)1𝒈ik.\displaystyle\theta^{k+1}_{i}=\theta^{k}_{i}+\alpha^{j}\sqrt{\frac{2\delta}{\bm{g}^{k}_{i}(\bm{H}^{k}_{i})^{-1}\bm{g}^{k}_{i}}}(\bm{H}^{k}_{i})^{-1}\bm{g}^{k}_{i}. (20)

αj(0,1)\alpha^{j}\in(0,1) is a coefficient that is found via backtracking line search. 𝑯ik=θi2𝔼sρ𝝅𝜽k[DKL(πθik(|s),πθi(|s))]|θi=θik\bm{H}^{k}_{i}=\nabla^{2}_{\theta_{i}}\mathbb{E}_{s\sim\rho_{\bm{\pi}_{\bm{\theta}^{k}}}}[\mathrm{D_{KL}}(\pi_{\theta^{k}_{i}}(\cdot|s),\pi_{\theta_{i}}(\cdot|s))]|_{\theta_{i}=\theta^{k}_{i}} is the Hessian of the expected KL-divergence. 𝒈ik\bm{g}^{k}_{i} is the gradient of the objective in Eq. (19), whose computation requires the estimation about 𝔼𝒂1:i1𝝅𝜽1:i1k+1,aiπθi[Ai𝝅𝜽k(s,𝒂1:i1,ai)]\mathbb{E}_{\bm{a}_{1:i-1}\sim\bm{\pi}_{\bm{\theta}^{k+1}_{1:i-1}},a_{i}\sim\pi_{\theta_{i}}}[A^{\bm{\pi}_{\bm{\theta}^{k}}}_{i}(s,\bm{a}_{1:i-1},a_{i})]. According to the derivation in [16], given a batch BB of trajectories with length TT, 𝒈ik\bm{g}^{k}_{i} can be computed as

𝒈ik=1Bb=1Bt=0TM1:i(st,𝒂t)θilogπθi(ait|st)|θi=θik,\displaystyle\bm{g}^{k}_{i}=\frac{1}{B}\sum\limits^{B}_{b=1}\sum\limits^{T}_{t=0}M_{1:i}(s^{t},\bm{a}^{t})\nabla_{\theta_{i}}\operatorname{log}\pi_{\theta_{i}}(a^{t}_{i}|s^{t})|_{\theta_{i}=\theta^{k}_{i}}, (21)

where

M1:i(st,𝒂t)=𝝅𝜽1:i1k+1(𝒂1:i1t|st)𝝅𝜽1:i1k(𝒂1:i1t|st)A𝝅𝜽k(st,𝒂t).\displaystyle M_{1:i}(s^{t},\bm{a}^{t})=\frac{\bm{\pi}_{\bm{\theta}^{k+1}_{1:i-1}}(\bm{a}^{t}_{1:i-1}|s^{t})}{\bm{\pi}_{\bm{\theta}^{k}_{1:i-1}}(\bm{a}^{t}_{1:i-1}|s^{t})}A^{\bm{\pi}_{\bm{\theta}^{k}}}(s^{t},\bm{a}^{t}). (22)

𝝅𝜽1:i1k+1(𝒂1:i1t|st)𝝅𝜽1:i1k(𝒂1:i1t|st)\frac{\bm{\pi}_{\bm{\theta}^{k+1}_{1:i-1}}(\bm{a}^{t}_{1:i-1}|s^{t})}{\bm{\pi}_{\bm{\theta}^{k}_{1:i-1}}(\bm{a}^{t}_{1:i-1}|s^{t})} is the compound policy ratio about the previous agents 1:i11:i-1, which can be computed easily after these agents complete their updates.

Although HATRPO has the excellent performance, it still faces many challenges in solving the UCBMOP. First, the formulated problem involves the complex cooperation among UAVs. Furthermore, only one time slot exists for each episode in the problem, which is different from other traditional MADRL tasks. Thus, it is necessary to take measures to connect the algorithm with the problem closely. Second, there are many UAVs in the problem, which may increase the input dimension of critic network substantially, influencing the critic learning. Finally, the conventional HATRPO uses the Gaussian distribution to sample actions. However, all actions have the finite range in the problem. Thus, using the Gaussian distribution may induce the bias to the policy gradient estimation. Accordingly, these reasons motivate us to propose the HATRPO-UCB, and the details are described in the following sections.

4.2.2 Algorithm Design of HATRPO-UCB

In the proposed HATRPO-UCB, we take conventional HATRPO as the foundation algorithm framework, and propose three techniques that are observation enhancement, agent-specific global state and Beta distribution for policy, to enhance the performance of the algorithm.

Observation enhancement: In MADRL, the environmental information is usually the state that guides the agents to take action. However, in UCBMOP, each episode only has one time slot, and the UVAA may communicate with different remote BSs at various episodes. In this case, the state consisting of the simple Cartesian coordinates of the UAVs and BSs may not precisely characterize their positional relationship. This is because the distance between the UVAA and BSs is significantly larger than the relative distance between the UAVs, and even the same set of UAV coordinates may represent different environmental states when serving various BSs (e.g., transmission angle and distance). To overcome this issue, we propose two observation enhancement strategies as follows. First, we design a spherical coordinate enhancement strategy with dynamically changing origins to represent the UAV positions. Instead of the Cartesian coordinate of each UAV, we set the state to the spherical coordinates with the BS in the current episode as the origin. In this way, the state can adequately express the information about the orientation and distance between the BS and the UVAA. Second, we propose a position representation based on reference points to characterize the BS locations. Specifically, we regard the closest point to the target BS in the monitor area as the reference point and then adopt this reference point to replace the target BS as the state. This adjustment retains the BS information but amplifies the position differences between the UAVs, which better expresses the relative positions of the UAVs. The enhanced state and observation contain representative and adaptive information about the environment, which ensures the proposed method remains effective when the location of the served BS at different episodes changes significantly.

Agent-specific global state: In general, the global state employed by MADRL algorithms has two forms, which are the concatenation of all local observations and environment-provided global state. However, in the considered scenarios, the number of agents (i.e., UAVs) is often large, which means that the input dimension of the critic network grows a lot when using the concatenation of all local observations. This condition will increase the learning burden, thereby making the performance of critic learning degrade. Additionally, the environment-provided global state is also not an appropriate choice because its contained information is not sufficient, which only contains the coordinates and excitation current weights of all UAVs, as well as the coordinates of the reference point. Inspired by [42], we design an agent-specific global state that combines the global information provided by the environment and some features from local observation, such as the spherical coordinate of the target BS that is relative to the UAV (θi,ϕi,di)(\theta_{i},\phi_{i},d_{i}) and the distance between two UAVs dj,id_{j,i}, as the input of the critic network. This agent-specific global state can improve the fitting performance of critic learning and thus achieve better MADRL performance.

Beta distribution for policy: In the conventional HATRPO, the actor uses the Gaussian distribution to sample actions. However, the Gaussian distribution has been demonstrated to have the negative impact on reinforcement learning [48]. Specifically, for an action with the finite interval, the Gaussian distribution with infinite support will define the action range out of its boundary. Furthermore, the probability density of all actions is greater than 0, but the probability of actions beyond boundary should equal to 0. Hence, in our scenario where all actions of UAV have the finite interval, it is inevitable that some actions sampled from the Gaussian distribution will be truncated, leading to the boundary effect. Then, the boundary effect may cause bias about policy gradient estimation.

To handle the above issue, the authors of [48] conducted the research on Beta distribution for reinforcement learning. The study indicates that the Beta distribution does not have a bias because the Beta distribution has a finite range (i.e., [0, 1]), and thus the probability density of actions beyond the boundary is guaranteed to be 0. Furthermore, experimental results show that the Beta distribution can make reinforcement learning algorithms converge faster and obtain higher rewards. Thus, in this paper, we choose the Beta distribution instead of the original Gaussian distribution, and then the Beta distribution is defined as

f(x;α,β)=Γ(α+β)Γ(α)Γ(β)xα1(1x)β1,\displaystyle f(x;\alpha,\beta)=\frac{\Gamma(\alpha+\beta)}{\Gamma(\alpha)\Gamma(\beta)}x^{\alpha-1}(1-x)^{\beta-1}, (23)

where α\alpha and β\beta are the shape parameters, and Γ()\Gamma(\cdot) is the Gamma function that extends factorial to real numbers. In addition, α\alpha and β\beta are set to be not less than 1 (i.e., α,β1\alpha,\beta\geq 1) in the work, which are modeled by softplus and then add a constant 1 [48].

4.2.3 Algorithm Workflow

Fig. 4 displays the framework of the proposed HATRPO-UCB algorithm, which has two phases that are the training phase and the implementation phase. Each UAV as an agent has a replay buffer and two networks (i.e., the actor network and critic network). The replay buffer is used to collect the data about the interaction between the UAV and the environment. Then, a central server is deployed with the replay buffers and networks of all agents for training.

Algorithm 1 Heterogeneous-Agent Trust Region Policy Optimization for UCBMOP (HATRPO-UCB)

Input: Number of episodes NEpiN^{\mathrm{Epi}}, number of agents NN,
  batch size BB, stepsize α\alpha, possible steps in line search LL, line
  search acceptance threshold κ\kappa.
Initialize: Actor networks {θi0,i𝒩}\{\theta^{0}_{i},\forall i\in\mathcal{N}\}, Critic networks
{ϕi0,i𝒩}\{\phi^{0}_{i},\forall i\in\mathcal{N}\}, Replay buffers {i,i𝒩}\{\mathcal{B}_{i},\forall i\in\mathcal{N}\}.

1:for episode = 1,,NEpi1,\dots,N^{\mathrm{Epi}} do
2:     Initialize the environment, acquire local observations
3:     and global states of all agents {(oi,si),i𝒩}\{(o_{i},s_{i}),\forall i\in\mathcal{N}\}.
4:     Each UAV selects its own action ai=πθik(oi)a_{i}=\pi_{\theta^{k}_{i}}(o_{i}) from
5:     the Beta distribution.
6:     All UAVs execute actions to complete CB and then
7:     receive their rewards {ri,i𝒩}\{r_{i},\forall i\in\mathcal{N}\}.
8:     Push transition of each agent oi,si,ai,ri\langle o_{i},s_{i},a_{i},r_{i}\rangle into
9:     its own replay buffer i\mathcal{B}_{i}.
10:     if the training condition is met then
11:         Draw a random permutation of agents i1:Ni_{1:N}.
12:         for agent i=1,,Ni=1,\dots,N do
13:              Sample a random mini-batch of BB transitions
14:              from i\mathcal{B}_{i}.
15:              Compute advantage function Ai(si,𝒂)A_{i}(s_{i},\bm{a}) based
16:              on the critic network.
17:              Obtain the compound policy ratio of the
18:              previous agents l=1i1πθlk+1(al|ol)πθlk(al|ol)\prod^{i-1}_{l=1}\frac{\pi_{\theta^{k+1}_{l}}(a_{l}|o_{l})}{\pi_{\theta^{k}_{l}}(a_{l}|o_{l})}, set
19:              Mi(si,𝒂)=Ai(si,𝒂)l=1i1πθlk+1(al|ol)πθlk(al|ol)M_{i}(s_{i},\bm{a})=A_{i}(s_{i},\bm{a})\prod^{i-1}_{l=1}\frac{\pi_{\theta^{k+1}_{l}}(a_{l}|o_{l})}{\pi_{\theta^{k}_{l}}(a_{l}|o_{l})}.
20:              Use Mi(si,𝒂)M_{i}(s_{i},\bm{a}) to estimate the gradient of
21:              maximization objective of agent 𝒈ik\bm{g}^{k}_{i} according
22:              to Eq. (21).
23:              Compute the Hessian of the expected
24:              KL-divergence 𝑯ik\bm{H}^{k}_{i} to approximate the KL
25:              constraint.
26:              𝒈ik\bm{g}^{k}_{i} and 𝑯ik\bm{H}^{k}_{i} are introduced to update the
27:              parameter of actor network θik+1\theta^{k+1}_{i} according to
28:              Eq. (20).
29:              Update critic network by the following
30:              equation:
31:              ϕik+1=argminϕi1Bb=1B(Vϕi(sib)Rib)2\phi^{k+1}_{i}=\operatorname{argmin}_{\phi_{i}}\frac{1}{B}\sum\limits^{B}_{b=1}(V_{\phi_{i}}(s^{b}_{i})-R^{b}_{i})^{2},
32:              where RibR^{b}_{i} is the discounted return at state sibs^{b}_{i}.
33:         end for
34:     end if
35:end for

Output: Trained models of all agents.

In the beginning, the central server will initialize the parameters of networks. Then, at each time slot, the central server will exchange the information with UAVs and then control them, where the amount of the information is very small because it only includes the observation information of UAVs and actions required to be done by UAVs, and the communication distance is very short compared with that between the UAVs and the BS. Hence, for performing the communication tasks, the above communication overhead is very small and acceptable. Then, for illustrating the specific process at each time slot, taking the ii-th agent as an example, after the central server receives the information, the actor network πθi\pi_{\theta_{i}} utilizes the observation oio_{i} to generate a Beta distribution about action, and then an action aia_{i} is sampled from the Beta distribution and sent to the UAV. The critic network is used to predict the Q-value of current global state sis_{i}. After executing the sampled action, the ii-th agent receives a reward rir_{i} from the environment and observes the next local state oio^{\prime}_{i} and global state sis^{\prime}_{i}. After that, the above information tuple oi,si,ai,ri\langle o_{i},s_{i},a_{i},r_{i}\rangle will be stored in the replay buffer i\mathcal{B}_{i}.

All agents will execute more time slots until their replay buffers accumulate enough data, and then the replay buffers will randomly select a mini-batch of tuples for training networks. The sequential updating scheme is adopted to update policies of all agents. In each training, all agents are updated sequentially in a randomly generated order. For the ii-th agent, the first step is to estimate the gradient 𝒈ik\bm{g}^{k}_{i}, before that, the advantage function of ii-th agent Ai(si,𝒂)A_{i}(s_{i},\bm{a}) and the compound policy ratio of the previous agents l=1i1πθlk+1(al|ol)πθlk(al|ol)\prod^{i-1}_{l=1}\frac{\pi_{\theta^{k+1}_{l}}(a_{l}|o_{l})}{\pi_{\theta^{k}_{l}}(a_{l}|o_{l})} are required to compute. Then, the Hessian of the expected KL-divergence 𝑯ik\bm{H}^{k}_{i} is computed to approximate the KL constraint. After completing the computation of 𝒈ik\bm{g}^{k}_{i} and 𝑯ik\bm{H}^{k}_{i}, the parameter of actor network θi\theta_{i} can be updated via backtracking line search. As for the critic network VϕiV_{\phi_{i}}, the update of network parameter can be achieved by optimizing the following loss function: 1Bb=1B(Vϕi(sib)Rib)2\frac{1}{B}\sum^{B}_{b=1}(V_{\phi_{i}}(s^{b}_{i})-R^{b}_{i})^{2}. Therefore, the above updating procedure will be repeated until HATRPO-UCB converges. The whole training process can be seen in Algorithm 1.

During the implementation stage, only the trained actor network is deployed to the UAV. In each communication mission, all UAVs receive the local observations from the environment, and then select actions through the actor networks. Afterwards, according to the selected actions, all UAVs move to the target locations and adjust the excitation current weights for performing CB. Compared with the training stage, the implementation stage is completed online and does not need much computational resource. In contrast, the training stage costs much computational resource, which is conducted offline in a central server.

4.2.4 Analysis of HATRPO-UCB

In this section, the computational complexity of the training and implementation stage of the proposed HATRPO-UCB is discussed.

Complexity of training. For nuavn_{\mathrm{uav}} agents, each agent is equipped with an actor network and a critic network that are formed by DNNs. A DNN consists of an input layer, LL fully connected layers and an output layer, where z0z_{0}, ziz_{i} and zL+1z_{L+1} denote the number of neurons of input layer, ii-th fully connected layer and output layer, respectively. Then, the computational complexity of DNN at each time slot is O(i=1L+1zi1zi)O(\sum^{L+1}_{i=1}z_{i-1}\cdot z_{i}).

At each training process, a mini-batch of data with nepin_{\mathrm{epi}} episodes is sampled from the replay buffer, but each episode only contains one time slot. Additionally, the critic network is updated by the stochastic gradient methods, thus the total computational complexity of critic network is O(nepi(i=1L+1zi1zi))O(n_{\mathrm{epi}}\cdot(\sum^{L+1}_{i=1}z_{i-1}\cdot z_{i})). However, the actor network is updated using the backtracking line search. In Eq. (20), the gradient 𝒈ik\bm{g}^{k}_{i} is first computed, whose computational complexity is O(nepi(i=1L+1zi1zi))O(n_{\mathrm{epi}}\cdot(\sum^{L+1}_{i=1}z_{i-1}\cdot z_{i})). Then the Hessian of expected KL-divergence 𝑯ik\bm{H}^{k}_{i} is computed to approximate the KL constraint. The computational complexity of expected KL-divergence is O(nepi(i=1L+1zi1zi))O(n_{\mathrm{epi}}\cdot(\sum^{L+1}_{i=1}z_{i-1}\cdot z_{i})), and the computational one of Hessian matrix is O((i=1L+1zi1zi)2)O((\sum^{L+1}_{i=1}z_{i-1}\cdot z_{i})^{2}). Hence, the computational complexity of 𝑯ik\bm{H}^{k}_{i} is O(nepi(i=1L+1zi1zi)+(i=1L+1zi1zi)2)O(n_{\mathrm{epi}}\cdot(\sum^{L+1}_{i=1}z_{i-1}\cdot z_{i})+(\sum^{L+1}_{i=1}z_{i-1}\cdot z_{i})^{2}). For computing (𝑯ik)1𝒈ik(\bm{H}^{k}_{i})^{-1}\bm{g}^{k}_{i}, the conjugate gradient algorithm is adopted, which obtains result by searching iteratively until convergence. The computational complexity of each iteration is O((i=1L+1zi1zi)2)O((\sum^{L+1}_{i=1}z_{i-1}\cdot z_{i})^{2}). Then the total computational complexity of (𝑯ik)1𝒈ik(\bm{H}^{k}_{i})^{-1}\bm{g}^{k}_{i} is O(nepi(i=1L+1zi1zi)+lconj(i=1L+1zi1zi)2)O(n_{\mathrm{epi}}\cdot(\sum^{L+1}_{i=1}z_{i-1}\cdot z_{i})+l_{\mathrm{conj}}\cdot(\sum^{L+1}_{i=1}z_{i-1}\cdot z_{i})^{2}), where lconjl_{\mathrm{conj}} is the number of iterations. During the backtracking line search, the computational complexity is O(lbtnepi(i=1L+1zi1zi))O(l_{\mathrm{bt}}\cdot n_{\mathrm{epi}}\cdot(\sum^{L+1}_{i=1}z_{i-1}\cdot z_{i})), where lbtl_{\mathrm{bt}} is the search number and O(nepi(i=1L+1zi1zi))O(n_{\mathrm{epi}}\cdot(\sum^{L+1}_{i=1}z_{i-1}\cdot z_{i})) is the computational complexity in each search. Hence, the total computational complexity of actor network is O(lbtnepi(i=1L+1zi1zi)+lconj(i=1L+1zi1zi)2)O(l_{\mathrm{bt}}\cdot n_{\mathrm{epi}}\cdot(\sum^{L+1}_{i=1}z_{i-1}\cdot z_{i})+l_{\mathrm{conj}}\cdot(\sum^{L+1}_{i=1}z_{i-1}\cdot z_{i})^{2}). Assume that HATRPO-UCB needs to be trained ltl_{\mathrm{t}} times for convergence, the overall computational complexity in the training phase is O(ltnuav(lbtnepi(i=1L+1zi1zi)+lconj(i=1L+1zi1zi)2))O(l_{\mathrm{t}}\cdot n_{\mathrm{uav}}\cdot(l_{\mathrm{bt}}\cdot n_{\mathrm{epi}}\cdot(\sum^{L+1}_{i=1}z_{i-1}\cdot z_{i})+l_{\mathrm{conj}}\cdot(\sum^{L+1}_{i=1}z_{i-1}\cdot z_{i})^{2})).

Complexity of inference. In the implementation phase, all UAVs only use the trained actor networks to make decisions. Therefore, the computational complexity of HATRPO-UCB in the implementation stage is O(nuav(i=1L+1zi1zi))O(n_{\mathrm{uav}}\cdot(\sum^{L+1}_{i=1}z_{i-1}\cdot z_{i})).

Remark 3.

Note that determining the theoretical bounds and convergence of the proposed MADRL-based algorithm confronts significant challenges [49]. Specifically, MADRL involves tuning numerous hyperparameters, such as learning rates, network architectures, and exploration strategies, resulting in an explosion of potential combinations. Thus, analyzing theoretical bounds and convergence becomes nearly infeasible given the vast parameter space. Moreover, DNNs introduce approximation errors when modeling complex functions, particularly in high-dimensional state spaces of the formulated optimization problem. These errors can lead to suboptimal performance in certain states, complicating the analysis of theoretical bounds and convergence. In addition, MADRL models interact with environments that introduce uncertainty factors, including noise and randomness. These uncertainties can yield different outcomes in various interaction trajectories, further complicating the analysis of the theoretical bounds and convergence [50]. Thus, we evaluate the performance of the proposed HATRPO-UCB by conducting extensive simulations like the works in [51, 40] in the following.

5 Simulation Results

In this section, we evaluate the performance of the proposed HATRPO-UCB algorithm for UCBMOP.

5.1 Simulation Configuration

We implement simulations in an environment with Python 3.8 and Pytorch 1.10, and perform all experiments on a server with AMD EPYC 7642 48-Core CPU, NVIDIA GeForce RTX 3090 GPU and 128GB RAM.

In the simulation, a 100 m ×\times 100 m square area is considered, where 16 UAVs are flying in the area. The mass (mUAVm_{UAV}), the minimum and maximum flying heights (HminH_{min} and HmaxH_{max}) of each UAV are set as 2 kg, 100 m and 120 m [22], respectively. The minimum distance between any UAVs (dmind_{min}) is 0.5 m. Moreover, the transmit power of each UAV, the carrier frequency (fcf_{c}) and the total noisy power spectral density are 0.1 W, 2.4 GHz and -157 dBm/Hz, respectively [21] [52]. The path loss exponent (α\alpha), as well as the attenuation factors of LoS and NLoS links (μLoS\mu_{LoS}, μNLoS\mu_{NLoS}) are 2, 3 dB and 23 dB, respectively [53]. Other parameter configurations about system model are summarized in Table II. Based on these settings and the mathematical models shown in Section 3, we build a simulation environment for the DRL model to interact and collect data.

In the MADRL stage, an environment is built according to the UCBMOP. The MADRL algorithm learns by interacting with the environment. During the training stage, the environment will run 4×1054\times 10^{5} episodes, where each episode only has one time slot. For the settings of reward weight coefficients, we first set the weight coefficients by normalizing each reward to make the rewards on the same order of magnitude. Then, we fine-tune the reward weights by considering their importance and characteristics (the tuning process and results are shown in Appendix A of the supplemental material). As such, the reward weight coefficients ω1\omega_{1}, ω2\omega_{2}, ω3\omega_{3}, ω4\omega_{4}, and ω5\omega_{5} are set to 100, 4, 30, 12 and 5, respectively.

In HATRPO-UCB, all actor and critic networks are three-layer fully connected neural networks (i.e., two hidden layers and one output layer). Each hidden layer contains 64 neurons, which is initialized by orthogonal initialization, and it is equipped with ReLU activation function. The actor network is updated by the backtracking line search, meanwhile, the KL-divergence between the new actor network and the old one must meet the KL-threshold which is set as 0.001. The Adam optimizer [54] is employed to update the critic network, where the learning rate is 0.005. The value setting of other related parameters is also presented in Table II.

TABLE II: Parameter settings
Parameter Value
Fuselage drag ratio (d0d_{0}) 0.6
Air density (ρ\rho) 1.225 km/m3\mathrm{km/m^{3}}
Rotor solidity (ss) 0.05
Rotor disc area (AA) 0.503 m2\mathrm{m^{2}}
Tip speed of the rotor blade (vtipv_{tip}) 120 m/s\mathrm{m/s}
Mean rotor induced velocity in hover (v0v_{0}) 4.03
Blade profile power (PBP_{B}) 79.76
Induced power (PIP_{I}) 88.66
Discount factor (γ\gamma) 0.99
Number of line searches 10
Clip parameter for loss value 0.2
Max norm of gradients 10
Accept ratio of loss improve 0.5
Weight initialization gain for actor network 0.01
Weight initialization for neural network Orthogonal

In addition, to verify the effectiveness and performance of HATRPO-UCB, except for conventional HATRPO, the following baseline methods are also introduced for comparison.

  • LAA: All UAVs form a linear antenna array (LAA), and they are symmetrically excited and located about the origin of the array [21].

  • RAA: A rectangular antenna array (RAA) is formed by all UAVs, where all UAVs are also symmetrically excited and located about the origin of the array.

  • MADDPG: Multi-Agent Deep Deterministic Policy Gradient (MADDPG) is based on the centralized training with decentralized execution (CTDE) paradigm, extending Deep Deterministic Policy Gradient (DDPG) to MADRL [55].

  • IPPO: Independent Proximal Policy Optimization (IPPO) is that the single-agent PPO algorithm is adopted directly to solve multi-agent tasks [56]. The IPPO has been justified that it can achieve excellent performance on SMAC.

  • MAPPO: Multi-Agent Proximal Policy Optimization (MAPPO) is an extension of PPO [42]. MAPPO utilizes parameter-sharing trick, then all agents jointly use a policy network and a value network. The global state is the input of the value network.

Note that the proposed observation enhancement technique are added to these abovementioned MADRL algorithms, so that making sure that they can solve the UCBMOP properly. Furthermore, they are also tuned continuously and achieve the convergence.

5.2 Convergence Analysis

In practice, DRL-based models are usually deployed only after achieving convergence through training. Even if the practical environment changes dynamic, we also need to re-train and fine-tune the model before deploying it. Thus, the convergence performance of HATRPO-UCB is vital. As such, in the section, we compare the convergence performance of HATRPO-UCB with other algorithms. Fig. 5 shows the convergence performance of all algorithms. It can be observed that all algorithms converge successfully. HATRPO-UCB achieves the fastest convergence at approximately 750 epoch, then the second algorithm is HATRPO. MAPPO, IPPO and MADDPG are slowest, which all begin to converge at about 1400 epoch. However, in the performance, the reward obtained by HATRPO is higher than that of the HATRPO-UCB, reaching about 85. IPPO and MAPPO are both worse than HATRPO-UCB. MADDPG has the poorest performance, which only converges to around 80. The reasons are that the introduced observation enhancement provides representative and adaptive information for the agents to take action, the agent-specific global state is able to improve the fitting performance of critic learning, and the Beta distribution overcomes the boundary effect issue. Such improvements could effectively address the major issues in the training process, thus facilitating the algorithm in achieving reliable and high-performance outcomes and converging rapidly.

Refer to caption
Figure 5: Convergence performance of different methods.
Refer to caption
Figure 6: Optimization results of the UAV energy consumption obtained by different methods.
TABLE III: Numerical optimization results of different methods
First BS Second BS
Method Transmission rate (bps) Energy consumption (J) Transmission rate (bps) Energy consumption (J)
LAA 1.020×1061.020\times 10^{6} 15208 1.793×1071.793\times 10^{7} /
RAA 0.998×1060.998\times 10^{6} 14217 1.788×1071.788\times 10^{7} /
MADDPG 1.027×1061.027\times 10^{6} 22319 1.904×𝟏𝟎𝟕\bm{1.904\times 10^{7}} 17913
IPPO 1.022×1061.022\times 10^{6} 15790 1.867×1071.867\times 10^{7} 21110
MAPPO 1.035×𝟏𝟎𝟔\bm{1.035\times 10^{6}} 15025 1.896×1071.896\times 10^{7} 23252
HATRPO 1.032×1061.032\times 10^{6} 13765 1.838×1071.838\times 10^{7} 11779
HATRPO-UCB 1.029×1061.029\times 10^{6} 𝟏𝟑𝟒𝟎𝟏\bm{13401} 1.825×1071.825\times 10^{7} 𝟏𝟎𝟐𝟔𝟏\bm{10261}
Refer to caption
(a)
Refer to caption
(b)
Refer to caption
(c)
Figure 7: Effectiveness of different techniques. (Observation enhancement, agent-specific global state and Beta distribution for policy).

To further analyze these algorithms, we also depict the optimization process of all alogrithms to the UAV energy consumption in Fig. 6. It can be observed that HATRPO-UCB obtains the best optimization performance, making the energy consumption of UAV reduce to about 700 J and achieving the convergence at approximately 1000 epochs. Then, HATRPO has the certain optimization performance, but it does not have the obvious optimization effect like HATRPO-UCB and the ultimate optimization result is not better than that of HATRPO-UCB. In addition, MAPPO, IPPO and MADDPG have no optimization effect, whose optimization performance becomes worse continously with the training process. The more details about performance of all algorithms will be introduced in the following sections.

5.3 Performance Comparison

The practical performance of all approaches is compared in the section. In our system, all UAVs have other kinds of tasks after completing CB, but they may continue to perform CB for communicating with another BS. Thus, we will analyze the actual performance of UVAA communicating continuously with two BSs. Table III shows the numerical optimization results of various methods. It can be observed that HATRPO-UCB achieves the best performance on the optimization of energy consumption of UVAA, and obtains the outstanding results on the transmission rate optimization. The highest transmission rates to the two BSs are achieved by MAPPO and MADDPG, respectively, but they consume more energy. Furthermore, remaining MADRL approaches also outperform HATRPO-UCB in terms of the optimization of transmission rate. The reason may be that all UAVs optimized by these algorithms fly higher and closer to BS, thus the UVAA can get the lower path loss and shorter distance with BS. However, the more energy consumption will be induced.

We also show the flight paths of UAVs in the Appendix B of supplemental material. Benefiting from the DNN, the UVAA optimized by MADRL methods can directly perform CB for the second BS without consuming the much computation resource. Besides, we can also observe that all UAVs optimized by other MADRL approaches can achieve the higher altitude and the shorter distance with BS compared with HATRPO-UCB, and thus, they can obtain the higher transmission rate but will consume more energy as shown in Table III. In summary, HATRPO-UCB learns the best strategy, making UVAA obtain the great transmission rate and save the energy at the same time.

5.4 Ablation Analysis

For verifying the effectiveness of proposed techniques, we conduct some ablation experiments. Fig. 7a displays some convergence performance as HATRPO-UCB is not implemented with some techniques. When HATRPO-UCB is not equipped with observation enhancement, we can observe that the performance is decreased. For further analysis, we present the practical performance of UVAA whether adopting observation enhancement or not in Fig. 8. As can be seen, when using observation enhancement, all UAVs can cooperate better, and they can also fly toward the BS to shorten the communication distance. On the contrary, without observation enhancement, all UAVs are scattered and have no clear flight targets. Furthermore, the CB performance is also influenced. The UVAA gets 24.85 reward about the transmission rate that is smaller than 25.05 obtained by the algorithm with observation enhancement. Therefore, the observation enhancement technique can significantly enable UVAA to learn the better strategy.

For testing the effectiveness of agent-specific global state, we replace the agent-specific global state, then use the environment-provided global state to evaluate the algorithm performance. As shown in Fig. 7a, HATRPO-UCB suffers from the performance degradation, which is because the information contained in the environment-provided global state is insufficient. Likewise, the critic learning is also influenced. Fig. 7b shows the learning curves of the critic network based on different global states, where the concatenation of all local observations is also added for comparison. It can be observed that the loss based on the agent-specific global state is lower than those under the other two global states. The reason is that the concatenation of local observations contains much redundant information, increasing the training burden, thus the learning process becomes very unstable and performance degrades. Then, the environment-provided global state also influences the training efficiency due to the insufficient information. However, according to the two figures, the influence produced by the two global states looks small. The reason could be that each episode only has one time slot in the UCBMOP, which distinctly differs from the regular MADRL tasks that require sequential decision-making. Hence, the role of critic network is not as significant as that in the regular tasks.

Refer to caption
(a)
Refer to caption
(b)
Figure 8: The impact of the first technique to UVAA when communicating with a BS. (a) Without observation enhancement; (b) With observation enhancement.

As for Beta distribution for policy, an action sampled from the Beta distribution is not beyond its setting range, which is beneficial for the policy gradient estimation so that no bias occurs. We show the impact of Beta distribution for policy to the optimization about the energy consumption of UAV in Fig. 7c. It can be seen that Beta distribution for policy can make HATRPO-UCB optimize the UAV energy consumption well, reducing the energy consumption to around 700 J. However, without Beta distribution for policy, the optimization effect is substantially weakened, and the ultimate performance is not better than the algorithm with Beta distribution for policy. Therefore, according to [48], adopting the Beta distribution can help the agent learn the better strategy compared with the Gaussian distribution.

Refer to caption
(a)
Refer to caption
(b)
Refer to caption
(c)
Refer to caption
(d)
Figure 9: Impact of different hyperparameter settings.

5.5 Impact of Hyperparameter Setting

The appropriate hyperparameter setting is critical to the algorithm. In HATRPO-UCB, the following hyperparameters are closely related to the algorithm performance that are the Optimizer, Learning rate, KL-threshold and Number of neurons. Hence, we analyze the impact of different values of each hyperparameter to HATRPO-UCB and simultaneously verify the reasonableness of our setting.

Algorithm performance under different optimizers. We first analyze the influence of various optimizers. Four optimizers are adopted to optimize the critic network, which are SGD[57], AdaGrad[58], RMSProp[59] and Adam[54], respectively. Their optimization results about the critic loss are shown in Fig. 9a. We can observe that, following the above order of optimizers, the optimized result becomes better. The performance of SGD is the worst, and the performance degradation appears in the following training. In the other optimizers, the training process also becomes unstable, resulting in an oscillating training curve. Then, when using Adam, the optimization performance is best and the training curve is smoothest. Therefore, Adam is choosen as the optimizer of HATRPO-UCB.

Algorithm performance under different learning rates. Except for determining the optimizer, the critic network is also optimized based on a certain learning rate. Hence, Fig. 9b shows the learning process of critic network based on Adam under four learning rates. It can be observed that the optimization performance is best and the critic loss curve is smoothest as the learning rate is 0.005. When equaling to 0.05 or 0.0005, the curve oscillates slightly, which demonstrates that they are not appropriate under the current hyperparameter configuration. However, as the learning rate is 0.5, the training process gets very unstable and the performance degrades seriously at the end of training. The probable reason is that the learning rate is too large to produce the fluctuation.

Algorithm performance under different KL-thresholds. As for the KL-threshold, it is related to the update of actor network, but the KL-threshold also influences the learning speed of the critic network. Fig. 9c describes the critic loss curve under various KL-thresholds. We can observe that critic network begins to converge at approximately 1000 epoch as the KL-threshold is 0.001. Then, when the KL-threshold equals 0.1 and 0.01, the convergence speed of critic network becomes faster. However, these large KL-thresholds also cause the instability to the critic learning in the second half of the whole process, and then also make the actor learning unstable. At last, 0.0001 KL-threshold makes the critic network learn so slowly that it converges at about 2000 epoch. Therefore, 0.001 KL-threshold is the best choice, which can make the critic network converge properly.

Algorithm performance under different neuron numbers. The number of neurons of hidden layer has the important effect to the DNN. Hence, Fig. 9d displays the performance of HATRPO-UCB under various numbers of neurons. We can observe that the convergence of critic network becomes slower as the number of neurons increases. The reason is that the large number of neurons makes the DNN big and complex, increasing the training burden. Hence, the number of neurons should be set to 64. In conclusion, according to all the above analysis, the hyperparameter setting with Adam, 0.005 learning rate, 0.001 KL-threshold and 64 neurons can make HATRPO-UCB achieve the best performance.

Refer to caption
Figure 10: Transmission rate of UVAA under different phase errors.

5.6 Impact of Imperfect Synchronization

The imperfect synchronization exists in CB, where the phase errors may be generated to influence the CB performance. Thus, we evaluate the impact of phase errors to the transmission performance of UVAA in the section.

The phase error at ii-th UAV antenna is denoted as ϵi\epsilon_{i}, and then the AF of UVAA can be rewritten as [60]

AF(θ,ϕ)\displaystyle AF(\theta,\phi) =\displaystyle= (24)
i=1NIiejΨiej[2πλ(xisinθcosϕ+yisinθsinϕ+zicosθ)+ϵi],\displaystyle\sum_{i=1}^{N}I_{i}e^{j\Psi_{i}}e^{j[\frac{2\pi}{\lambda}(x_{i}\sin{\theta}\cos{\phi}+y_{i}\sin{\theta}\sin{\phi}+z_{i}\cos{\theta})+\epsilon_{i}]},

which ϵi\epsilon_{i} is assumed to follow a Tikhonov (or von Mises) distribution [60] [61]. The Tikhonov distribution is described as

fϵ(ϵ)=12πI0(γ)eγcos(ϵ),\displaystyle f_{\epsilon}(\epsilon)=\frac{1}{2\pi I_{0}(\gamma)}e^{\gamma cos(\epsilon)}, (25)

where |ϵ|<π|\epsilon|<\pi, I0(x)I_{0}(x) is the zero-th order modified Bessel function of the first kind, and γ\gamma is the inverse of the variance of the phase error. Then, the impact of phase errors to the transmission rate of UVAA under different values of γ\gamma is shown in Fig. 10. It can be seen that the phase errors make the UVAA transmission performance decline. However, the generated phase errors become smaller as γ\gamma gets large, weakening the influence. For tackling the imperfect synchronization, there has been many works proposing various closed-loop or open-loop methods [31]. Furthermore, as the better synchronization algorithms are being proposed continuously, the impact of the imperfect synchronization will become less and less.

6 Discussion

In this section, we discuss possible issues and solutions when deploying the proposed method in actual scenarios.

6.1 The Impact of UAV Collision

In this part, we discuss the impact of UAV collisions. As shown in Eq. (10f), we set the minimum separation between two UAVs in UCBMOP. Following this, we set a penalty in the reward. Specifically, when any two UAVs violate the minimum separation constraint, they will get a tiny negative reward. As such, the proposed algorithm is able to make UAVs keep the minimum separation.

Moreover, many mature methods of UAV collision avoidance can also help the UAVs avoid violating the minimum separation in reality. For example, cameras, infrared, radar, LiDAR, sonar, etc, can be utilized by UAVs for obstacle detection. Based on this, the existing methods such as sense & avoid [62] can be appropriately embedded in the proposed optimization framework. This type of method focuses on reducing the computational cost with a short response time, deviating the UAVs from their original paths when needed, and then turning the UAVs to the previous flight state quickly. Since this method only consumes very little computational resources, energy, and time, it does not have a significant effect on the results of our work. As such, UAV collisions do not affect the effectiveness of the proposed method.

6.2 Additional Energy Consumption in Synchronization

In the considered system, there is additional energy consumption in the synchronization process for beacon signal reception (for closed loop) and location estimation (for open loop). However, this additional energy consumption is minor and can be ignored, and the reasons are as follows.

First, the closed-loop synchronization methods are based on feedback, whose energy consumption is quite small compared to the transmission or motion energy consumption of the UAVs. Specifically, the closed-loop synchronization approaches mainly include two types that are the iterative bit feedback and rich feedback methods. In the considered UAV-enabled CB scenarios, both of these closed-loop methods result in negligible energy consumption. This is because the recently proposed methods, such as iterative bit feedback approaches and rich feedback methods, can achieve rapid convergence with minimal energy consumption (less than 1 Joule in most cases). As such, this additional energy consumption is much smaller than the motion energy of UAVs (often several hundred Joules each second).

To demonstrate, we select a representative method, i.e., D1BF [63], to implement and evaluate. Following this, we introduce the energy consumption of point-to-point sending and broadcast receiving models from [64]. As shown in Fig. 11(a), the energy consumption of the closed-loop synchronization is not larger than even 1 Joule. Thus, we can demonstrate that the additional energy consumption of closed-loop synchronization in the considered system is quite small and can be ignored.

Second, compared to the closed-loop synchronization methods, the open-loop synchronization methods are not based on iterative feedback and have less energy consumption. Specifically, the open-loop synchronization approaches can be divided into two categories that are the intra-node communication and blind method. On the one hand, the intra-node methods are mostly based on the master-slave architecture [65]. We also use the energy model above and assume that the master node sends a reference beacon to all slave nodes and then the slaves synchronize their own phases. In this case, the energy consumption of synchronization of the UAVs can be shown in Fig. 11(b). As can be seen, it is also not larger than 1 Joule and can be omitted. On the other hand, blind methods allow the nodes to synchronize among themselves and do not require feedback from the receiver or the reference nodes. Hence, the overhead of the blind method can be smaller. Thus, the additional energy consumption of open-loop synchronization in the considered system is also small and can be ignored.

Overall, the additional energy consumption for synchronization does not have an evident impact on the availability of the proposed method.

Refer to caption
(a)
Refer to caption
(b)
Figure 11: Energy consumption of the UAVs for the synchronization process. (a) Closed loop. (b) Open loop. These synchronization methods need no more than one hundred times generation to complete synchronization tasks. As can be seen, the energy consumption of each generation is very small and can be ignored.

7 Conclusion

In this paper, a UAV-assistant A2G communication system is investigated, where multiple UAVs form a UVAA to perform CB for communicating with remote BSs. Then, we formulate a UCBMOP, aiming at simultaneously maximizing the transmission rate of UVAA and minimizing the energy consumption of all UAVs. Consider that the system is dynamic and the cooperation among UAVs is complex, we propose the HATRPO-UCB, which is an MADRL algorithm to address the problem. Except for combining conventional HATRPO, three techniques are proposed to enhance the performance of the proposed algorithm. Simulation results demonstrate that the proposed HATRPO-UCB learns the better strategy than other baseline methods including LAA, RAA, MADDPG, IPPO, MAPPO and conventional HATRPO, making UVAA achieve the best transmission performance and save the energy consumption simultaneously. Furthermore, the effectiveness of three techniques is also verified by the ablation experiments.

References

  • [1] Y. Zeng, Q. Wu, and R. Zhang, “Accessing from the sky: A tutorial on UAV communications for 5G and beyond,” Proc. IEEE, vol. 107, no. 12, pp. 2327–2375, 2019.
  • [2] N. Zhao, W. Lu, M. Sheng, Y. Chen, J. Tang, F. R. Yu, and K.-K. Wong, “UAV-assisted emergency networks in disasters,” IEEE Wirel. Commun., vol. 26, no. 1, pp. 45–51, 2019.
  • [3] S. Chen, J. Zhang, E. Bjornson, J. Zhang, and B. Ai, “Structured massive access for scalable cell-free massive MIMO systems,” IEEE J. Sel. Areas Commun., vol. 39, pp. 1086–1100, Apr. 2021.
  • [4] S. Ahmed, M. Z. Chowdhury, and Y. M. Jang, “Energy-efficient uav relaying communications to serve ground nodes,” IEEE Commun. Lett., vol. 24, no. 4, pp. 849–852, 2020.
  • [5] M. Khosravi and H. Pishro-Nik, “Unmanned aerial vehicles for package delivery and network coverage,” in Proc. VTC2020-Spring, pp. 1–5, IEEE, 2020.
  • [6] Y. Zeng, R. Zhang, and T. J. Lim, “Wireless communications with unmanned aerial vehicles: Opportunities and challenges,” IEEE Commun. Mag., vol. 54, no. 5, pp. 36–42, 2016.
  • [7] M. Li, L. Liu, Y. Gu, Y. Ding, and L. Wang, “Minimizing energy consumption in wireless rechargeable UAV networks,” IEEE Internet Things J., vol. 9, no. 5, pp. 3522–3532, 2021.
  • [8] Y. Zeng, J. Xu, and R. Zhang, “Energy minimization for wireless communication with rotary-wing UAV,” IEEE Trans. Wireless Commun., vol. 18, no. 4, pp. 2329–2345, 2019.
  • [9] C. Zhan and Y. Zeng, “Energy minimization for cellular-connected uav: From optimization to deep reinforcement learning,” IEEE Trans. Wireless Commun., 2022.
  • [10] M. Mozaffari, W. Saad, M. Bennis, Y.-H. Nam, and M. Debbah, “A tutorial on UAVs for wireless networks: Applications, challenges, and open problems,” IEEE Commun. Surv. Tutorials, vol. 21, no. 3, pp. 2334–2360, 2019.
  • [11] S. Liang, Z. Fang, G. Sun, Y. Liu, G. Qu, S. Jayaprakasam, and Y. Zhang, “A joint optimization approach for distributed collaborative beamforming in mobile wireless sensor networks,” Ad Hoc Netw., vol. 106, p. 102216, 2020.
  • [12] J. Garza, M. A. Panduro, A. Reyna, G. Romero, and C. d. Rio, “Design of UAVs-based 3D antenna arrays for a maximum performance in terms of directivity and SLL,” Int. J. Antenn. Propag., vol. 2016, 2016.
  • [13] S. Fujimoto, H. Hoof, and D. Meger, “Addressing function approximation error in actor-critic methods,” in Proc. ICML, pp. 1587–1596, PMLR, 2018.
  • [14] Y. Wang, Z. Gao, J. Zhang, X. Cao, D. Zheng, Y. Gao, D. W. K. Ng, and M. Di Renzo, “Trajectory design for UAV-based internet of things data collection: A deep reinforcement learning approach,” IEEE Internet Things J., vol. 9, no. 5, pp. 3899–3912, 2021.
  • [15] H. Xie, D. Yang, L. Xiao, and J. Lyu, “Connectivity-aware 3D UAV path design with deep reinforcement learning,” IEEE Trans. Veh. Technol., vol. 70, no. 12, pp. 13022–13034, 2021.
  • [16] J. G. Kuba, R. Chen, M. Wen, Y. Wen, F. Sun, J. Wang, and Y. Yang, “Trust region policy optimisation in multi-agent reinforcement learning,” arXiv preprint arXiv:2109.11251, 2021.
  • [17] Y. Cai, Z. Wei, R. Li, D. W. K. Ng, and J. Yuan, “Joint trajectory and resource allocation design for energy-efficient secure uav communication systems,” IEEE Trans. Commun., vol. 68, no. 7, pp. 4536–4553, 2020.
  • [18] M. Wang, L. Zhang, P. Gao, X. Yang, K. Wang, and K. Yang, “Stackelberg game-based intelligent offloading incentive mechanism for a multi-UAV-assisted mobile edge computing system,” IEEE Internet Things J., 2023.
  • [19] S. F. Abedin, M. S. Munir, N. H. Tran, Z. Han, and C. S. Hong, “Data freshness and energy-efficient UAV navigation optimization: A deep reinforcement learning approach,” IEEE Trans. Intell. Transp. Syst., vol. 22, no. 9, pp. 5994–6006, 2020.
  • [20] C. H. Liu, X. Ma, X. Gao, and J. Tang, “Distributed energy-efficient multi-uav navigation for long-term communication coverage by deep reinforcement learning,” IEEE Trans. Mob. Comput., vol. 19, no. 6, pp. 1274–1285, 2019.
  • [21] M. Mozaffari, W. Saad, M. Bennis, and M. Debbah, “Communications and control for wireless drone-based antenna array,” IEEE Trans. Commun., vol. 67, no. 1, pp. 820–834, 2018.
  • [22] G. Sun, J. Li, Y. Liu, S. Liang, and H. Kang, “Time and energy minimization communications based on collaborative beamforming for UAV networks: A multi-objective optimization method,” IEEE J. Sel. Areas Commun., vol. 39, no. 11, pp. 3555–3572, 2021.
  • [23] G. Sun, J. Li, A. Wang, Q. Wu, Z. Sun, and Y. Liu, “Secure and energy-efficient UAV relay communications exploiting collaborative beamforming,” IEEE Trans. Commun., vol. 70, no. 8, pp. 5401–5416, 2022.
  • [24] G. Sun, X. Zheng, Z. Sun, Q. Wu, J. Li, Y. Liu, and V. C. Leung, “UAV-enabled secure communications via collaborative beamforming with imperfect eavesdropper information,” IEEE Trans. Mob. Comput., vol. 23, no. 4, pp. 3291–3308, 2024.
  • [25] J. Li, G. Sun, L. Duan, and Q. Wu, “Multi-objective optimization for UAV swarm-assisted iot with virtual antenna arrays,” IEEE Trans. Mob. Comput., pp. 1–18, 2024.
  • [26] J. Huang, A. Wang, G. Sun, and J. Li, “Jamming-aided maritime physical layer encrypted dual-UAVs communications exploiting collaborative beamforming,” in Proc. CSCWD, IEEE, 2023.
  • [27] H. Li, D. Wei, G. Sun, J. Wang, J. Li, and H. Kang, “Interference mitigation via collaborative beamforming in UAV-enabled data collections: A multi-objective optimization method,” 2022.
  • [28] S. Krishna Moorthy, N. Mastronarde, S. Pudlewski, E. S. Bentley, and Z. Guan, “Swarm UAV networking with collaborative beamforming and automated ESN learning in the presence of unknown blockages,” Comput. Netw., vol. 231, p. 109804, 2023.
  • [29] Y. Zhang, Z. Mou, F. Gao, J. Jiang, R. Ding, and Z. Han, “UAV-enabled secure communications by multi-agent deep reinforcement learning,” IEEE Trans. Veh. Technol., vol. 69, no. 10, pp. 11599–11611, 2020.
  • [30] C. Dai, K. Zhu, and E. Hossain, “Multi-agent deep reinforcement learning for joint decoupled user association and trajectory design in full-duplex multi-UAV networks,” IEEE Trans. Mob. Comput., 2022.
  • [31] S. Jayaprakasam, S. K. A. Rahim, and C. Y. Leow, “Distributed and collaborative beamforming in wireless sensor networks: Classifications, trends, and research directions,” IEEE Commun. Surv. Tutorials, vol. 19, no. 4, pp. 2092–2116, 2017.
  • [32] A. Al-Hourani, S. Kandeepan, and S. Lardner, “Optimal lap altitude for maximum coverage,” IEEE Wireless Commun. Lett., vol. 3, no. 6, pp. 569–572, 2014.
  • [33] G. Sun, Y. Liu, Z. Chen, A. Wang, Y. Zhang, D. Tian, and V. C. M. Leung, “Energy efficient collaborative beamforming for reducing sidelobe in wireless sensor networks,” IEEE Trans. Mob. Comput., vol. 20, no. 3, pp. 965–982, 2021.
  • [34] S. Jayaprakasam, S. K. Abdul Rahim, C. Y. Leow, and T. O. Ting, “Sidelobe reduction and capacity improvement of open-loop collaborative beamforming in wireless sensor networks,” PloS one, vol. 12, no. 5, p. e0175510, 2017.
  • [35] Y. Zeng, X. Xu, and R. Zhang, “Trajectory design for completion time minimization in UAV-enabled multicasting,” IEEE Trans. Wireless Commun., vol. 17, no. 4, pp. 2233–2246, 2018.
  • [36] S.-F. Chou, A.-C. Pang, and Y.-J. Yu, “Energy-aware 3D unmanned aerial vehicle deployment for network throughput optimization,” IEEE Trans. Wireless Commun., vol. 19, no. 1, pp. 563–578, 2019.
  • [37] Z. Yang, W. Xu, and M. Shikh-Bahaei, “Energy efficient UAV communication with energy harvesting,” IEEE Trans. Veh. Technol., vol. 69, no. 2, pp. 1913–1927, 2019.
  • [38] C. You and R. Zhang, “Hybrid offline-online design for UAV-enabled data harvesting in probabilistic LoS channels,” IEEE Trans. Wireless Commun., vol. 19, no. 6, pp. 3753–3768, 2020.
  • [39] P. Goos, U. Syafitri, B. Sartono, and A. Vazquez, “A nonlinear multidimensional knapsack problem in the optimal design of mixture experiments,” Eur. J. Oper. Res., vol. 281, no. 1, pp. 201–221, 2020.
  • [40] A. Feriani and E. Hossain, “Single and multi-agent deep reinforcement learning for ai-enabled wireless networks: A tutorial,” IEEE Commun. Surv. Tutorials, vol. 23, no. 2, pp. 1226–1252, 2021.
  • [41] M. L. Littman, “Markov games as a framework for multi-agent reinforcement learning,” in Machine learning proceedings 1994, pp. 157–163, Elsevier, 1994.
  • [42] C. Yu, A. Velu, E. Vinitsky, Y. Wang, A. Bayen, and Y. Wu, “The surprising effectiveness of ppo in cooperative, multi-agent games,” arXiv preprint arXiv:2103.01955, 2021.
  • [43] D. Salama, T. K. Sarkar, M. N. Abdallah, X. Yang, and M. Salazar-Palma, “Adaptive processing at multiple frequencies using the same antenna array consisting of dissimilar nonuniformly spaced elements over an imperfectly conducting ground,” IEEE Trans. Antennas Propag., vol. 67, pp. 622–625, Jan. 2019.
  • [44] C. S. de Witt, B. Peng, P.-A. Kamienny, P. Torr, W. Böhmer, and S. Whiteson, “Deep multi-agent reinforcement learning for decentralized continuous cooperative control,” arXiv preprint arXiv:2003.06709, 2020.
  • [45] M. Samvelyan, T. Rashid, C. S. De Witt, G. Farquhar, N. Nardelli, T. G. Rudner, C.-M. Hung, P. H. Torr, J. Foerster, and S. Whiteson, “The starcraft multi-agent challenge,” arXiv preprint arXiv:1902.04043, 2019.
  • [46] S. Kakade and J. Langford, “Approximately optimal approximate reinforcement learning,” in Proc. of ICML, p. 267–274, 2002.
  • [47] J. Schulman, S. Levine, P. Abbeel, M. Jordan, and P. Moritz, “Trust region policy optimization,” in Proc. ICML, pp. 1889–1897, 2015.
  • [48] P.-W. Chou, D. Maturana, and S. Scherer, “Improving stochastic policy gradients in continuous control with deep reinforcement learning using the beta distribution,” in Proc. ICML, pp. 834–843, 2017.
  • [49] W. Zhou, Z. Cao, N. Deng, K. Jiang, and D. Yang, “Identify, estimate and bound the uncertainty of reinforcement learning for autonomous driving,” IEEE Trans. Intell. Transp. Syst., vol. 24, no. 8, pp. 7932–7942, 2023.
  • [50] K. Zhang, Z. Yang, and T. Başar, Multi-Agent Reinforcement Learning: A Selective Overview of Theories and Algorithms, pp. 321–384. Springer International Publishing, 2021.
  • [51] T. Li, K. Zhu, N. C. Luong, D. Niyato, Q. Wu, Y. Zhang, and B. Chen, “Applications of multi-agent reinforcement learning in future internet: A comprehensive survey,” IEEE Commun. Surv. Tutorials, vol. 24, no. 2, pp. 1240–1279, 2022.
  • [52] J. Li, H. Kang, G. Sun, S. Liang, Y. Liu, and Y. Zhang, “Physical layer secure communications based on collaborative beamforming for UAV networks: A multi-objective optimization approach,” in Proc. IEEE INFOCOM 2021, pp. 1–10, IEEE, 2021.
  • [53] M. Mozaffari, W. Saad, M. Bennis, and M. Debbah, “Wireless communication using unmanned aerial vehicles (UAVs): Optimal transport theory for hover time optimization,” IEEE Trans. Wireless Commun., vol. 16, no. 12, pp. 8052–8066, 2017.
  • [54] D. P. Kingma and J. Ba, “Adam: A method for stochastic optimization,” arXiv preprint arXiv:1412.6980, 2014.
  • [55] R. Lowe, Y. I. Wu, A. Tamar, J. Harb, O. Pieter Abbeel, and I. Mordatch, “Multi-agent actor-critic for mixed cooperative-competitive environments,” Advances in neural information processing systems, vol. 30, 2017.
  • [56] C. S. de Witt, T. Gupta, D. Makoviichuk, V. Makoviychuk, P. H. Torr, M. Sun, and S. Whiteson, “Is independent learning all you need in the starcraft multi-agent challenge?,” arXiv preprint arXiv:2011.09533, 2020.
  • [57] L. Bottou et al., “Stochastic gradient learning in neural networks,” Proceedings of Neuro-Nımes, vol. 91, no. 8, p. 12, 1991.
  • [58] J. Duchi, E. Hazan, and Y. Singer, “Adaptive subgradient methods for online learning and stochastic optimization.,” Journal of machine learning research, vol. 12, no. 7, 2011.
  • [59] T. Tieleman, G. Hinton, et al., “Lecture 6.5-rmsprop: Divide the gradient by a running average of its recent magnitude,” COURSERA: Neural networks for machine learning, vol. 4, no. 2, pp. 26–31, 2012.
  • [60] A. Minturn, D. Vernekar, Y. L. Yang, and H. Sharif, “Distributed beamforming with imperfect phase synchronization for cognitive radio networks,” in Proc. ICC, pp. 4936–4940, IEEE, 2013.
  • [61] Y. S. Shmaliy, “Von mises/tikhonov-based distributions for systems with differential phase measurement,” Signal Process., vol. 85, no. 4, pp. 693–703, 2005.
  • [62] Y. Zeng, Y. Hu, S. Liu, J. Ye, Y. Han, X. Li, and N. Sun, “Rt3d: Real-time 3-d vehicle detection in lidar point cloud for autonomous driving,” IEEE Robot. Autom. Lett., vol. 3, no. 4, pp. 3434–3440, 2018.
  • [63] I. Thibault, G. E. Corazza, and L. Deambrogio, “Random, deterministic, and hybrid algorithms for distributed beamforming,” in 2010 5th Advanced Satellite Multimedia Systems Conference and the 11th Signal Processing for Space Communications Workshop, IEEE, 2010.
  • [64] J. Feng, Y.-H. Lu, B. Jung, and D. Peroulis, “Energy efficient collaborative beamforming in wireless sensor networks,” in 2009 IEEE International Symposium on Circuits and Systems, IEEE, May 2009.
  • [65] F. Quitin, M. M. Ur Rahman, R. Mudumbai, and U. Madhow, “Distributed beamforming with software-defined radios: Frequency synchronization and digital feedback,” in Proc. IEEE GLOBECOM, 2012.
[Uncaptioned image] Saichao Liu received a BS degree and an MS degree in Software Engineering from Henan University, China, in 2019 and 2022, respectively. He is currently pursuing the Ph.D. degree with the College of Computer Science and Technology, Jilin University, Changchun, China. His research interests include wireless communications, UAV networks, antenna arrays and reinforcement learning.
[Uncaptioned image] Geng Sun (S’17-M’19) received the B.S. degree in communication engineering from Dalian Polytechnic University, and the Ph.D. degree in computer science and technology from Jilin University, in 2011 and 2018, respectively. He was a Visiting Researcher with the School of Electrical and Computer Engineering, Georgia Institute of Technology, USA. He is an Associate Professor in College of Computer Science and Technology at Jilin University, and His research interests include wireless networks, UAV communications, collaborative beamforming and optimizations.
[Uncaptioned image] Jiahui Li (S’21) received a BS degree in Software Engineering, and an MS degree in Computer Science and Technology from Jilin University, Changchun, China, in 2018 and 2021, respectively. He is currently studying Computer Science at Jilin University to get a Ph.D. degree, and also a visiting Ph. D. at Singapore University of Technology and Design (SUTD), Singapore. His current research focuses on UAV networks, antenna arrays, and optimization.
[Uncaptioned image] Shuang Liang received the B.S. degree in Communication Engineering from Dalian Polytechnic University, China in 2011, the M.S. degree in Software Engineering from Jilin University, China in 2017, and the Ph.D. degree in Computer Science from Jilin University, China in 2022. She is a post-doctoral in the School of Information Science and Technology, Northeast Normal University, and her research interests focus on wireless communication and UAV networks.
[Uncaptioned image] Qingqing Wu (S’13-M’16-SM’21) received the B.Eng. and the Ph.D. degrees in Electronic Engineering from South China University of Technology and Shanghai Jiao Tong University (SJTU) in 2012 and 2016, respectively. From 2016 to 2020, he was a Research Fellow in the Department of Electrical and Computer Engineering at National University of Singapore. He is currently an Associate Professor with Shanghai Jiao Tong University. His current research interest includes intelligent reflecting surface (IRS), unmanned aerial vehicle (UAV) communications, and MIMO transceiver design. He has coauthored more than 100 IEEE journal papers with 26 ESI highly cited papers and 8 ESI hot papers, which have received more than 18,000 Google citations. He was listed as the Clarivate ESI Highly Cited Researcher in 2022 and 2021, the Most Influential Scholar Award in AI-2000 by Aminer in 2021 and World’s Top 2% Scientist by Stanford University in 2020 and 2021.
[Uncaptioned image] Pengfei Wang (Member, IEEE) received the B.S., M.S., and Ph.D. degrees in software engineering from Northeastern University (NEU), China, in 2013, 2015, and 2020, respectively. From 2016 to 2018, he was a Visiting Ph.D. Student with the Department of Electrical Engineering and Computer Science, Northwestern University, IL, USA. He is currently an Associate Professor with the School of Computer Science and Technology, Dalian University of Technology (DUT), China. He has authored more than 30 papers on high-quality journals and conferences, such as IEEE/ACM TRANSACTIONS ON NETWORKING, IEEE INFOCOM, IEEE TRANSACTIONS ON INTELLIGENT TRANSPORTATION SYSTEMS, DAC, IEEE ICNP, IEEE ICDCS, IEEE INTERNET OF THINGS JOURNAL, and JSA. He also holds a series of patents in U.S. and China. His research interests are ubiquitous computing, big data, and AIoT.
[Uncaptioned image] Dusit Niyato (Fellow, IEEE) received the B.Eng. degree from the King Mongkuts Institute of Technology Ladkrabang (KMITL), Thailand, in 1999, and the Ph.D. degree in electrical and computer engineering from the University of Manitoba, Canada, in 2008. He is currently a Professor with the School of Computer Science and Engineering, Nanyang Technological University, Singapore. His research interests include the Internet of Things (IoT), machine learning, and incentive mechanism design.