This paper was converted on www.awesomepapers.org from LaTeX by an anonymous user.
Want to know more? Visit the Converter page.

\old@ps@headings

Beam Selection for Energy-Efficient mmWave Network Using Advantage Actor Critic Learning

Ycaro Dantas1, Pedro Enrique Iturria-Rivera1, Hao Zhou1, Majid Bavand2, Medhat Elsayed2,
Raimundas Gaigalas2, and Melike Erol-Kantarci1, 2Ericsson Inc., Ottawa, Canada
Emails:{ydant103, pitur008, hzhou098, melike.erolkantarci}@uottawa.ca,
{majid.bavand, medhat.elsayed, raimundas.gaigalas}@ericsson.com
1School of Electrical Engineering and Computer Science, University of Ottawa, Ottawa, Canada
Abstract

The growing adoption of mmWave frequency bands to realize the full potential of 5G, turns beamforming into a key enabler for current and next-generation wireless technologies. Many mmWave networks rely on beam selection with Grid-of-Beams (GoB) approach to handle user-beam association. In beam selection with GoB, users select the appropriate beam from a set of pre-defined beams and the overhead during the beam selection process is a common challenge in this area. In this paper, we propose an Advantage Actor Critic (A2C) learning-based framework to improve the GoB and the beam selection process, as well as optimize transmission power in a mmWave network. The proposed beam selection technique allows performance improvement while considering transmission power improves Energy Efficiency (EE) and ensures the coverage is maintained in the network. We further investigate how the proposed algorithm can be deployed in a Service Management and Orchestration (SMO) platform. Our simulations show that A2C-based joint optimization of beam selection and transmission power is more effective than using Equally Spaced Beams (ESB) and fixed power strategy, or optimization of beam selection and transmission power disjointly. Compared to the ESB and fixed transmission power strategy, the proposed approach achieves more than twice the average EE in the scenarios under test and is closer to the maximum theoretical EE.

Index Terms:
beamforming, beam selection, energy efficiency, reinforcement learning, advantage actor critic.

I Introduction

mmWave technology significantly empowered the fifth-generation wireless networks (5G) due to the availability of ultra-wide frequency bands in the spectrum bands above 6 GHz. mmWave is also expected to take an important role in Beyond 5G (B5G) technologies, and even the range of THz communications is considered as a potential enabler in the next generations [1]. The gains of using mmWave can only be unlocked with the use of large antenna arrays that generate ultra-narrow beams and overcome propagation losses. This gives beam management a central role in modern wireless systems.

Beam selection is the process to establish the best beam pair for the transmitter and the receiver among the available beams on both sides. The exhaustive search over all possible beam pairs could not be efficient in some cases, as in high mobility scenarios. The overhead in this process could consume a significant fraction of the channel coherence interval and leave little time for utilization [2]. Different solutions were applied to address this problem, such as the use of out-of-band information [2], hierarchical beamforming [3], and data-driven methods based on contextual information [4].

While efficient beam management procedures are envisioned, global goals for wireless network systems, such as Energy Efficiency (EE), remain on the horizon. EE was pointed out by Next Generation Mobile Networks as a key factor in minimizing the Total Cost of Ownership (TCO) and the environmental footprint of the network infrastructure [5]. The commitment to EE improvement goals tends to stay on focus for B5G. In addition, Open Radio Access Network (O-RAN) signatory operators have recently released their technical priorities about EE [6]. Not only efficient hardware and software designs are targeted, but there is a focus on EE performance indicators to be reported at different system levels, and on taking advantage of Machine Learning (ML) algorithms to automate its improvement.

Intelligent resource management through ML algorithms is seen as a way to achieve the needs of future wireless networks [7]. Automated network optimization gained priority in recent virtualized and disaggregated RAN architectures, such as Cloud RAN and Open RAN. The abstraction layers designed for network control, such as the Service Management and Orchestration (SMO) platform, are responsible for that. SMO allows the use of RAN control applications (rApps) and could be implemented on top of Open RAN, Cloud RAN or purpose-built RAN environments, where it could assume Self-Organizing Network (SON) functions[8]. Furthermore, O-RAN Alliance sees the Grid-of-Beams (GoB) optimization through ML models as one of the main use cases for MIMO and beamforming-related applications to be deployed on the SMO [9]. It could be used to implement intelligent control over beamforming functions, allowing a flexible and dynamic optimization of those parameters to address the beam selection, for instance.

This paper builds upon the idea of local feature optimization to contribute to a global goal. Based on the GoB optimization use case, a Reinforcement Learning (RL) solution is proposed in a way to select the best beam subsets and reduce the overhead of the beam selection process. The Advantage Actor Critic (A2C) framework is applied to perform the joint optimization of beam selection and gNB transmission power, targeting to improve EE without producing coverage roles. In particular, intrinsic information from the network (UE SINR levels) is leveraged to feed the RL algorithm. The algorithm will act over the GoB and power configurations according to an objective function designed to reward EE and penalize coverage holes. In this scenario, the use of an ML model can provide flexible design, improved performance, and feasible implementation in the SMO.

The main contribution of this work is the proposed coverage-aware A2C-based beam selection method for energy-efficient mmWave networks. In addition, we investigate how the proposed algorithm can be deployed in the SMO to interact with recent RAN architectures.

The remaining sections are organized as follows. In Section II, the related work in the area is presented. Section III details the system model. Section IV describes the proposed EE and coverage-aware beam selection algorithm. The details about the simulations and results are shown in Section V. Finally, Section VI concludes this work.

II Related Work

ML models have been applied to solve a wide range of wireless network problems including beam management procedures [4], [7], [10] - [14]. In order to leverage the gains of data-driven methods, many solutions proposed the use of external information sources to feed ML algorithms and take better beam management decisions. In [4], the authors demonstrated the advantages of using GPS information to support the beam selection process in a vehicle-to-infrastructure (V2I) scenario.

The inherent error associated with geolocation data was targeted in [10], which explored the Uncertainty K-Means (UK-means) algorithm to improve the accuracy in user grouping during the beam association. In [11], the joint sensing and communication paradigm was extended into a vision-aided approach. Satellite images, beyond location information, were used to support beam management decisions. Moreover, a federated learning model was proposed to address the beam selection problem in [12]. The authors have showed that the collaborative learning process improved performance on throughput and best beam pairs detection compared to their baselines.

Beamforming codebook design using clustering algorithms and RL methods was the focus of [13]. The efficient codebook design was proved to provide good performance with a reduced number of beams, thus addressing the overhead in the beam selection problem. In [14], Deep Q-Networks (DQN) were applied to jointly solve beamforming, power control, and interference coordination problem, but no EE analysis was conducted.

Different than existing works, we propose a coverage-aware intelligent beam selection method to improve EE in mmWave networks. The proposed method investigates the trade-off between EE metrics and mmWave coverage. Moreover, this work discusses how the proposed intelligent beam selection method can be implemented in an SMO platform.

III System Model

III-A Network Model

The network configuration used in this work consists of a 3D beamforming capable gNB and NUE{N_{UE}} stationary User Equipment (UE). We consider analog beamforming and assume a UE could experience line of sight (LOS) or non-line of sight (NLOS) propagation conditions. Finally, the presence of UDP traffic in the downlink direction is considered.

III-B Grid-of-Beams Model

The 3D beamforming capability is represented by a regular GoB, that can be described through the set of its elevation angles Θ={θ1,θ2,,θNΘ}\Theta=\{\theta_{1},\theta_{2},…,\theta_{N_{\Theta}}\} and azimuth angles Φ={ϕ1,ϕ2,,ϕNΦ}\Phi=\{\phi_{1},\phi_{2},…,\phi_{N_{\Phi}}\}. θ\theta and ϕ\phi are, respectively, the elevation and azimuth angles of the maximum gain in a beam. Each beam is then described by the pair (θi,ϕj)(\theta_{i},\phi_{j}), where i[1,NΘ]i\in[1,N_{\Theta}] and j[1,NΦ]j\in[1,N_{\Phi}]. NΘN_{\Theta} and NΦN_{\Phi} are the number of values that θ\theta and ϕ\phi could assume. The maximum number of beams available to the system, here defined as NBN_{B}, is NB=NΘNΦN_{B}=N_{\Theta}\cdot N_{\Phi}. The full GoB, named Υ\Upsilon, is obtained in terms of Θ\Theta and Φ\Phi: if Υ=Θ×Φ\Upsilon=\Theta\times\Phi, then:

Υ=[(θ1,ϕ1)(θ1,ϕ2)(θ1,ϕNΦ)(θ2,ϕ1)(θ2,ϕ2)(θ2,ϕNΦ)(θNΘ,ϕ1)(θNΘ,ϕ2)(θNΘ,ϕNΦ)]\Upsilon=\begin{bmatrix}(\theta_{1},\phi_{1})&(\theta_{1},\phi_{2})&\dots&(\theta_{1},\phi_{N_{\Phi}})\\ (\theta_{2},\phi_{1})&(\theta_{2},\phi_{2})&\dots&(\theta_{2},\phi_{N_{\Phi}})\\ \vdots&\vdots&\ddots&\vdots\\ (\theta_{N_{\Theta}},\phi_{1})&(\theta_{N_{\Theta}},\phi_{2})&\dots&(\theta_{N_{\Theta}},\phi_{N_{\Phi}})\end{bmatrix} (1)

The 3D GoB beamforming model is depicted in Fig. 1. The overhead in the beam selection process derives from the requirement to exchange pilot signals and evaluate the quality of each beam pair. In highly changing scenarios, it could impact the initial access and mobility procedures and consume channel utilization time. Due to this overhead, constraints on the use of NBN_{B} beams in the beam selection could be imposed. A subset of the GoB composed of KBK_{B} beams, where KB<NBK_{B}<N_{B}, must be required in some cases. This constraint can be attended if we limit the number of elevation and azimuth angle options to a subset of Θ\Theta and Φ\Phi: Θ\Theta^{\prime} and Φ\Phi^{\prime}, where ΘΘ\Theta^{\prime}\subset\Theta and ΦΦ\Phi^{\prime}\subset\Phi. Consider that KB=KΘKΦK_{B}=K_{\Theta}\cdot K_{\Phi}, where KΘK_{\Theta} and KΦK_{\Phi} are, respectively, the number of options that each beam elevation angle θ\theta and azimuth angle ϕ\phi could assume after the constraint, so as KΘ=|Θ|K_{\Theta}=|\Theta^{\prime}| and KΦ=|Φ|K_{\Phi}=|\Phi^{\prime}|.

Refer to caption

Figure 1: Conceptual design of the 3D GoB beamforming model.

IV Proposed Approach

This work proposes joint optimization of beam selection and gNB transmission power, driven by EE and coverage awareness. The aim is to use GoB considering overhead constraints in the beam selection process in mmWave networks.

For the case of KΘ<NΘK_{\Theta}<N_{\Theta} and KΦ<NΦK_{\Phi}<N_{\Phi}, Θ\Theta^{\prime} and Φ\Phi^{\prime} should be determined. One solution is to sample Θ\Theta and Φ\Phi according to KΘK_{\Theta} and KΦK_{\Phi}, and consider equally spaced values for θi\theta_{i} and ϕj\phi_{j} within its ranges. This approach, referred to as Equally Spaced Beams (ESB), chooses Θ\Theta^{\prime} and Φ\Phi^{\prime} to provide fair spacial coverage for the network area.

Another option is to determine Θ\Theta^{\prime} and Φ\Phi^{\prime} according to the characteristics of the scenario. Moreover, a dynamic choice over the subset of Θ\Theta and Φ\Phi could prove to be more efficient. Fig. 2 illustrates different ways to determine GoB subsets in the case of overhead constraints.

Refer to caption

Figure 2: Conceptual azimuth view of GoB subsets. a) KΦ=NΦK_{\Phi}=N_{\Phi}, then Φ=Φ\Phi^{\prime}=\Phi: the full GoB is in use; b) KΦ<NΦK_{\Phi}<N_{\Phi}: Φ\Phi^{\prime} is chosen from Equally Spaced Beams (ESB) to provide fair coverage; c) KΦ<NΦK_{\Phi}<N_{\Phi}: Φ\Phi^{\prime} is chosen from the beams that could provide the best service for the UE in this scenario.

IV-A The Advantage Actor Critic (A2C) Framework

The optimization of beam selection and transmission power is done using the RL framework A2C. The foundational goal in RL is to learn a policy based on the sequential actions and environment states that maximize a cumulative future reward. If an action has a high expected reward, it should be given a higher probability of selection for an observed state. A2C differs from other Deep Reinforcement Learning (DRL) frameworks because it proposes a clear separation between two tasks performed by the agent. Each block within the agent (actor and critic) has its dedicated neural network engine to perform the corresponding task.The critic’s task is to judge the actor’s decisions based on the current state of the environment and the outcomes provided as a reward. The actor’s task is to define the best action to be performed on the environment based on the current state and the policy values sent by the critic [17].

IV-B Markov Decision Process (MDP) Formulation

IV-B1 State Space

The proposed environment is the network model described in Section III and presented to the algorithm through the state space. sts_{t} consists of the SINR values Γ\Gamma observed for each UE, quantized and truncated over the range [0,10][0,10], here represented as Γ^\hat{\Gamma}.

For each UE,

Γ^i={0, for Γi<=0Γi, for 0<Γi<1010, for Γi>=10\hat{\Gamma}_{i}=\begin{cases}0,$ for $\Gamma_{i}<=0\\ \lfloor\Gamma_{i}\rceil,$ for $0<\Gamma_{i}<10\\ 10,$ for $\Gamma_{i}>=10\end{cases} (2)

Thus, the state space is formulated as:

st=[Γ^1,Γ^2,,Γ^NUE].s_{t}=[\hat{\Gamma}_{1},\hat{\Gamma}_{2},...,\hat{\Gamma}_{N_{UE}}]. (3)

IV-B2 Action Space

Following the GoB model described in Section III, the action space is here defined as:

at=[Θl,Φm,δn],a_{t}=[\Theta_{l}^{\prime},\Phi_{m}^{\prime},\delta_{n}], (4)

where ΘlΘ\Theta_{l}^{\prime}\subset\Theta and ΦmΦ\Phi_{m}^{\prime}\subset\Phi. ll and mm are subset indexes so as l[1,L]l\in[1,L] and m[1,M]m\in[1,M], where LL and MM represent the number of different subset options from Θ\Theta and Φ\Phi that account for KΘK_{\Theta} and KΦK_{\Phi} constraints. Here, LL and MM are defined as: L=(NΘKΘ)L=\binom{N_{\Theta}}{K_{\Theta}} and M=(NΦKΦ)M=\binom{N_{\Phi}}{K_{\Phi}}. Consider that ΘL=(ΘKΘ)\Theta^{L}=\binom{\Theta}{K_{\Theta}} and ΦM=(ΦKΦ)\Phi^{M}=\binom{\Phi}{K_{\Phi}}, so ΘL\Theta^{L} and ΦM\Phi^{M} could represent the sets of subsets from Θ\Theta and Φ\Phi that account for KΘK_{\Theta} and KΦK_{\Phi} constraints. Hence, ΘlΘL\Theta_{l}^{\prime}\in\Theta^{L} and ΦmΦM\Phi_{m}^{\prime}\in\Phi^{M}.

δn\delta_{n} accounts for a power level change over the baseline power level PBP_{B}, chosen from the set of power level change options Δ\Delta. nn is a change index, so as n[1,NP]n\in[1,N_{P}], and NP=|Δ|N_{P}=|\Delta|.

IV-B3 Reward Function

The reward function is proposed in a way to reinforce learning towards the best achievable EE and to avoid the presence of coverage holes in the scenario. It is defined as follows:

rt=εεMAXωCρ,r_{t}={\frac{\varepsilon}{\varepsilon_{MAX}}}-\omega_{C}\cdot\rho, (5)

where ε\varepsilon represents the EE in terms of gNB throughput (τ\tau) and transmission power (PP), so as ε=τP\varepsilon=\frac{\tau}{P}. Here, εMAX\varepsilon_{MAX} is the maximum theoretical EE, derived from the data rate provided by the server application in the system (τMAX\tau_{MAX}) and the minimum amount of power available to be selected (PMINP_{MIN}), i.e. εMAX=τMAXPMIN\varepsilon_{MAX}=\frac{\tau_{MAX}}{P_{MIN}}. Furthermore, ρ\rho determines the UE outage ratio, based on an SINR threshold Γ^T\hat{\Gamma}^{T} and ρ=nUENUE\rho=\frac{n_{UE}}{N_{UE}}, where nUEn_{UE} is the number of UE served by the gNB with Γ^<=Γ^T\hat{\Gamma}<=\hat{\Gamma}^{T}. Since the second term in the reward function determines the penalties according to the UE outage ratio, ωC\omega_{C} is used as an out-of-coverage density factor weight, customizable to tune the severity of the penalty and balance the two goals within the reward function: EE and coverage awareness. In summary, rtr_{t} is proportional to EE, and it seeks to maintain a greater number of UEs above the SINR threshold.

IV-C Baseline Algorithms

Two other versions of the MDP described above are evaluated as baselines. The first one, named A2C-P, intends to cover the case where only power optimization is conducted by the algorithm. In this case, ESB is used to define the constrained GoB. It uses the same state space and reward function presented above, however, the action space is defined as:

atP=[δn].a^{P}_{t}=[\delta_{n}]. (6)

A second variant named A2C-B is proposed with the intent to cover the case where only beam selection optimization is conducted. In this case, the baseline power level is used. As A2C-P, it uses the same state space and reward function as A2C, but the action space is defined as:

atB=[Θl,Φm].a^{B}_{t}=[\Theta_{l}^{\prime},\Phi_{m}^{\prime}]. (7)

Finally, we consider the use of ESB and fixed transmission power as a third baseline, covering the case of no dynamic optimization.

IV-D Algorithm Deployment

rApps are applications designed to automate and control RAN functions and run on top of the SMO platform. rApps are enabled by the Non-Real Time RAN Intelligent Controller (Non-RT RIC), an SMO entity, and operate in control loops of 1 second or more [15]. In accordance with the GoB optimization use case presented in [9], the approach proposed in this work may be deployed as an rApp in the SMO. Note that, the implementation as an rApp is feasible because the global GoB and power optimization proposed here do not require near real-time feedback and interventions.

Fig. 3-a) shows a potential deployment where the proposed solution is used as an rApp in the SMO for Open RAN architecture. In O-RAN, network data gathered by the Open Decentralized Unit (O-DU) is sent to the SMO and Non-RT RIC Framework through O1 interface. Non-RT RIC Framework uses R1 communication to transfer the data to the rApp.

In our case, the A2C-based GoB and power configurations (actions) can be applied to the Open Radio Unit (O-RU) through Open FH M-Plane interface, and to O-DU through O1 interface. Note that, Non-RT RIC rApps are not limited to the Open RAN environment. With the use of proper interfaces, similar deployment could be done on top of Cloud RAN or purpose-built network architectures.

Fig. 3-b) details the functioning of the proposed scheme when it is developed as an rApp. The algorithm configuration is done based on Θ\Theta, Φ\Phi, ωC\omega_{C}, KΘK_{\Theta} and KΦK_{\Phi}. In each loop, data collected from O-DU will enable the reward function computation (feeds the critic) and the state report (feeds the actor and the critic). The action selection will be done by the actor and then sent as an rApp policy to be applied to O-RU and O-DU.

Refer to caption

Figure 3: The proposed solution deployed as an rApp in the SMO. a) Data flow through the rApp and Open RAN entities (SMO: Service Management and Orchestration; RIC: RAN Intelligent Controller; O-RU: Open Radio Unit; O-DU: Open Distributed Unit; O-CU: Open Centralized Unit; R1, O1, A1, E2, Open FH M-Plane, and Open FH CUS-Plane are open interfaces; F1-c and F1-u are 3GPP-defined interfaces); b) A2C based rApp internal functioning (ata_{t}: action taken at time tt; sts_{t}: state reported at time tt; rtr_{t}: reward value at time tt).

V Simulations and Results

V-A Simulation Environment and Settings

We implement our proposed scheme using the 5G LENA module for ns-3 [16]. The main system parameters for the simulation environment are summarized in Table I.

Each simulation is set to last 150 ms, where only 100 ms are used for data traffic purposes. Different scenarios were set according to the intended number of UE, since NUE={5,10,15,20}N_{UE}=\{5,10,15,20\}. The results shown in Section V refer to the average value and confidence interval of 40 different seed runs for each scenario configuration. In each run, new channel parameters and UE distribution are set.

A regular GoB where NΘ=3N_{\Theta}=3 and NΦ=5N_{\Phi}=5, so as NB=15N_{B}=15, is considered here. The set of available angles for each axis are the following: Θ={15,45,60}\Theta=\{-15^{\circ},-45^{\circ},-60^{\circ}\} and Φ={0,45,90,135,180}\Phi=\{0^{\circ},45^{\circ},90^{\circ},135^{\circ},180^{\circ}\}. KΘ=1K_{\Theta}=1 and KΦ=3K_{\Phi}=3 are defined considering a constraint of KB=3K_{B}=3. For the A2C based approach, consider L=3L=3 and M=10M=10. For the ESB strategy with KB=3K_{B}=3, Θ={45}\Theta^{\prime}=\{-45^{\circ}\} and Φ={0,90,180}\Phi^{\prime}=\{0^{\circ},90^{\circ},180^{\circ}\} are considered. Both ESB strategies use the baseline power level (PBP_{B}), so as A2C-B. PBP_{B} is set as 35 dBm. Five power level change options δ\delta are defined, so as NP=5N_{P}=5. The set of power level change options is the following: Δ={3,1.5,0,+1.5,+3}\Delta=\{-3,-1.5,0,+1.5,+3\} dB.

A2C implementation was done using the python package Stable Baselines 3 [19]. Each A2C episode corresponds to a new simulation round, fed with the action set by A2C, and that returns the states and reward values needed to feed the RL algorithm’s next iteration. The convergence curve for the reward values obtained by A2C in the case where NUE=5N_{UE}=5 is depicted in Fig. 4.

TABLE I: System Parameters
Parameter Value
Simulation Area 50m x 50m
gNB Antenna Array 4x32
Number of gNB 1
Number of UE {5, 10, 15, 20}
Carrier Frequency 28GHz
Bandwidth 100MHz
Numerology 2
Component Carriers 1
Propagation Model 3GPP TR 38.900 Urban Macro
gNB TX Power {32, 33.5, 35, 36.5, 38}dBm
UE TX Power 10dBm
Traffic Type DL UDP
Packet Size 400KB
Data Rate (User Application) 1Mbps
Refer to caption
Figure 4: A2C convergence curve for the scenario with ωC=1\omega_{C}=1 and NUE=5N_{UE}=5. The shaded area represents the range of the confidence interval out of 40 simulation runs.

V-B Results

Fig. 5 shows the EE values obtained for the proposed approach (A2C) when compared to the ESB for KB=3K_{B}=3 and KB=15K_{B}=15. The cases with only the power or the beam subset optimization (A2C-P and A2C-B) are also shown. KB=15K_{B}=15 represents the case where Υ=Υ\Upsilon^{\prime}=\Upsilon, and the beam search over the full GoB is performed.

A2C outperforms ESB for the same number of beams. More importantly, it can improve EE even when compared to the case of greater KBK_{B} with no power optimization. The ratio of εMAX\varepsilon_{MAX} achieved by each strategy is summarized in Table II.

Refer to caption

Figure 5: gNB Energy Efficiency. KBK_{B} is the number of beams in the grid. δ\delta is the power level change from the baseline power level.
TABLE II: Max Energy Efficiency Ratio
ε/εMAX\varepsilon/\varepsilon_{MAX}
NUEN_{UE}
Algorithm 55 1010 1515 2020 Avg.
A2C (KB=3,δ=δnK_{B}=3,\delta=\delta_{n}) 0.74 0.55 0.50 0.43 0.56
A2C-P (KB=3,δ=δnK_{B}=3,\delta=\delta_{n}) 0.51 0.36 0.32 0.32 0.38
A2C-B (KB=3,δ=0K_{B}=3,\delta=0) 0.42 0.34 0.31 0.27 0.33
ESB (KB=3,δ=0K_{B}=3,\delta=0) 0.33 0.28 0.24 0.23 0.27
ESB (KB=15,δ=0K_{B}=15,\delta=0) 0.44 0.42 0.39 0.35 0.40

For the same number of beams (KB=3K_{B}=3), the use of A2C does not cause a significant impact on throughput, as shown in Fig. 6. The results for the complete original GoB (KB=15K_{B}=15) are shown to demonstrate the achievable performance when no constraints to NBN_{B} are applied. In the case of throughput, KB=15K_{B}=15 achieves higher values due to the gains in SINR derived from the use of more beams. However, it is not compliant with the constraints defined in our problem (KB=3K_{B}=3). KB=15K_{B}=15 with fixed power is also less energy-efficient than the proposed solution.

Refer to caption

Figure 6: gNB Throughput. KBK_{B} is the number of beams in the grid. δ\delta is the power level change from the baseline power level.

Fig. 7 shows that A2C could converge to minimize transmission power at the gNB. If only power optimization is performed (A2C-P), throughput is not improved, and the power levels in use are diminished but are higher than for the global A2C case (Fig. 7). Thus, A2C-P presents poor EE improvement. When only the beam subset optimization is performed (A2C-B), throughput is improved, but since the power is fixed and not optimized, EE gains are discrete. Moreover, A2C-P and A2C-B have similar performances on EE. It is the joint optimization of beam selection and transmission power that produces the best EE results. The proposed approach leveraged the gains from the beam subset optimization with throughput to minimize transmission power and improve EE at the gNB.

Refer to caption

Figure 7: gNB Power Usage. KBK_{B} is the number of beams in the grid. δ\delta is the power level change from the baseline power level.

As shown in Section IV, the reward function of the proposed approach is designed to improve EE and avoid the presence of coverage holes. Fig. 8 shows how the proposed approach reduces the number of UE with Γ^<0\hat{\Gamma}<0 in the scenario when compared to the ESB case with the same number of beams. It happened even though the use of A2C translated into lower levels of power applied in the system.

Refer to caption

Figure 8: Percentage of UE with SINR Above 0 dB. KBK_{B} is the number of beams in the grid. δ\delta is the power level change from the baseline power level.

VI Conclusions

This paper showed the design and evaluation of an A2C-based strategy for GoB optimization. The A2C framework was applied to promote EE improvement and coverage awareness, deciding jointly on the beam selection and transmission power at the gNB. A deployment option on an SMO platform was also shown. The proposed approach is able to optimize GoB beamforming and introduce significant gains in EE. The algorithm could place good decisions in terms of beam subset choice and transmission power with the use of intrinsic information only (UE SINR levels). The gains from the beam selection were leveraged to minimize the transmission power in the system. No throughput or coverage losses were required to improve EE, and the balance between EE and coverage was explored within the reward function, which accounts for the flexibility of this proposal. Future directions for this work include the use of the proposed approach on multi-gNB and mobility scenarios.

Acknowledgment

This work is partly supported by MITACS and NSERC programs. The work of Ycaro Dantas is partly supported by the Vector Scholarship in Artificial Intelligence, provided through the Vector Institute.

References

  • [1] C. Chaccour, M. N. Soorki, W. Saad, M. Bennis, P. Popovski and M. Debbah, “Seven defining features of terahertz (THz) wireless systems: a fellowship of communication and sensing,” IEEE Communications Surveys and Tutorials, vol. 24, no. 2, pp. 967-993, second quarter 2022.
  • [2] A. Ali, N. González-Prelcic and R. W. Heath, “Millimeter Wave Beam-Selection Using Out-of-Band Spatial Information,” IEEE Transactions on Wireless Communications, vol. 17, no. 2, pp. 1038-1052, Feb. 2018.
  • [3] J. Wang, “Beam codebook based beamforming protocol for multi-Gbps millimeter-wave WPAN systems”, IEEE J. Sel. Areas Commun., vol. 27, no. 8, pp. 1390-1399, Oct. 2009.
  • [4] A. Klautau, P. Batista, N. González-Prelcic, Y. Wang and R. W. Heath, “5G MIMO Data for Machine Learning: Application to Beam-Selection Using Deep Learning,” in 2018 Information Theory and Applications Workshop (ITA), pp. 1-9, Feb. 2018.
  • [5] NGMN Alliance, “NGMN 5G white paper,” March 2015.
  • [6] Deutsche Telekom, Orange, Telefonica, TIM, and Vodafone “Open RAN Technical Priorities, Focus on Energy Efficiency,” June 2021.
  • [7] M. Elsayed and M. Erol-Kantarci, ”AI-Enabled Future Wireless Networks: Challenges, Opportunities, and Open Issues,” in IEEE Vehicular Technology Magazine, vol. 14, no. 3, pp. 70-77, Sept. 2019.
  • [8] Ericsson, “Intelligent Optimization Guide,” June 2022. Available at https://www.ericsson.com/en/ran/intelligent-ran-automation.
  • [9] O-RAN Alliance Working Group 1, “O-RAN Massive MIMO Use Cases Technical Report 1.0,” July 2022. Available at https://www.o-ran.org/.
  • [10] Y. Yao, H. Zhou and M. Erol-Kantarci, “Deep reinforcement learning-based radio resource allocation and beam management under location uncertainty in 5G mmWave networks,” arXiv:2204.10984, Apr. 2022.
  • [11] Y. Yao, H. Zhou and M. Erol-Kantarci, “Joint Sensing and Communications for Deep Reinforcement Learning-based Beam Management in 6G,” arXiv:2208.01880, Aug. 2022.
  • [12] M. B. Mashhadi, M. Jankowski, T. -Y. Tung, S. Kobus and D. Gündüz, ”Federated mmWave Beam Selection Utilizing LIDAR Data,” in IEEE Wireless Communications Letters, vol. 10, no. 10, pp. 2269-2273, Oct. 2021.
  • [13] Y. Zhang, M. Alrabeiah and A. Alkhateeb, “Reinforcement learning of beam codebooks in millimeter wave and terahertz MIMO systems,” IEEE Transactions on Communications, vol. 70, no. 2, pp. 904-919, Feb. 2022.
  • [14] F. B. Mismar, B. L. Evans and A. Alkhateeb, “Deep reinforcement learning for 5G networks: joint beamforming, power control, and interference coordination,” IEEE Transactions on Communications, vol. 68, no. 3, pp. 1581-1592, March 2020.
  • [15] M. Polese, L. Bonati, S. D’Oro, S. Basagni, and T. Melodia, “Understanding O-RAN: architecture, interfaces, algorithms, security, and research challenges,” arXiv:2202.01032, Feb. 2022.
  • [16] Natale Patriciello, Sandra Lagen, Biljana Bojovic, Lorenza Giupponi, “An E2E simulator for 5G NR networks”, Simulation Modelling Practice and Theory (SIMPAT), Vol. 96, 101933, Nov. 2019.
  • [17] M. Sewak, “Actor-critic models and the A3C,” Deep Reinforcement Learning, pp. 141–152, Springer, 2019.
  • [18] 3rd Generation Partnership Project (3GPP), “Study on channel model for frequency spectrum above 6 GHz,” 3GPP TR 38.900 version 14.2.0 Release 14, 2017.
  • [19] A. Raffin, A. Hill, A. Gleave, A. Kanervisto, M. Ernestus, and N. Dormann, “Stable-Baselines3: reliable reinforcement learning implementations”, Journal of Machine Learning Research, 2021.