34
Learning to Optimize in Smart Radio Environments A. Zappone LANEAS, CentraleSupelec, Paris, France

Learning to Optimize in Smart Radio Environmentsyuenchau/6GWorkshop... · Learning to Optimize 80 AI provides the tools to address C.2 and C.3. Complexity crunch in communication

  • Upload
    others

  • View
    0

  • Download
    0

Embed Size (px)

Citation preview

Page 1: Learning to Optimize in Smart Radio Environmentsyuenchau/6GWorkshop... · Learning to Optimize 80 AI provides the tools to address C.2 and C.3. Complexity crunch in communication

Learning to Optimize in Smart Radio Environments

A. Zappone

LANEAS, CentraleSupelec, Paris, France

Page 2: Learning to Optimize in Smart Radio Environmentsyuenchau/6GWorkshop... · Learning to Optimize 80 AI provides the tools to address C.2 and C.3. Complexity crunch in communication

For more info

A. Zappone, M. Di Renzo, and M. Debbah, ”Wireless NetworksDesign in the Era of Deep Learning: Model-Based, AI-Based, orBoth?”, IEEE Transactions on Communincations, submitted, 2019(invited paper), arXiv preprint arXiv:1902.02647

M. Di Renzo, M. Debbah, D.-T. Phan-Huy, A. Zappone, M.-S.Alouini, C. Yuen, V. Sciancalepore, G. C. Alexandropoulos, J. Hoydis,H. Gacanin, J. de Rosny, A. Bounceu, G. Lerosey, M. Fink. ”SmartRadio Environments Empowered by AI Reconfigurable Meta-Surfaces:An Idea Whose Time Has Come”, Eurasip Journal on WirelessCommunications and Networking, submitted 2019, arXiv preprintarXiv:1903.08925

Page 3: Learning to Optimize in Smart Radio Environmentsyuenchau/6GWorkshop... · Learning to Optimize 80 AI provides the tools to address C.2 and C.3. Complexity crunch in communication

Why Deep Learning in Wireless Communications?

Deep Learning in Wireless Communications – Why ?

56

HISILICON SEMICONDUCTORHUAWEI TECHNOLOGIES CO., LTD.

6GVision 510~20Gbps

: :10Mbps/m2

-500km/h

1Millions/km2+

6G

Tbps

1/m2 10nsD10cm +40dB

100x

1ms

:3~5 X

+0.1~1Gbps

10000x

q Increasing network complexity! Typically we have Cases 2 or 3.q Rapidly reaching the scenario where traditional models do notprovide the accuracy required / are too complex to optimize.

Faster transmission technologies are not enough. Wireless networks architectures need their own (artificial) brain

Because of complexity and modeling issues!

Page 4: Learning to Optimize in Smart Radio Environmentsyuenchau/6GWorkshop... · Learning to Optimize 80 AI provides the tools to address C.2 and C.3. Complexity crunch in communication

Why Deep Learning in Wireless Communications?

Deep Learning in Wireless Communications – Why ?

56

HISILICON SEMICONDUCTORHUAWEI TECHNOLOGIES CO., LTD.

6GVision 510~20Gbps

: :10Mbps/m2

-500km/h

1Millions/km2+

6G

Tbps

1/m2 10nsD10cm +40dB

100x

1ms

:3~5 X

+0.1~1Gbps

10000x

q Increasing network complexity! Typically we have Cases 2 or 3.q Rapidly reaching the scenario where traditional models do notprovide the accuracy required / are too complex to optimize.

Faster transmission technologies are not enough. Wireless networks architectures need their own (artificial) brain

Because of complexity and modeling issues!

Page 5: Learning to Optimize in Smart Radio Environmentsyuenchau/6GWorkshop... · Learning to Optimize 80 AI provides the tools to address C.2 and C.3. Complexity crunch in communication

Deep learning in wireless

Enablers:

Deep learning requires a lot of data to process

A lot of data is available over the air (Big Data Era)

Improved computing abilities (GPUs)

Challenges:

How to acquire, store, and process this much data?

Artificial neural networks in the cloud or in each device?

How to integrate artificial neural networks into wireless networks?

Page 6: Learning to Optimize in Smart Radio Environmentsyuenchau/6GWorkshop... · Learning to Optimize 80 AI provides the tools to address C.2 and C.3. Complexity crunch in communication

Deep learning in wireless

Enablers:

Deep learning requires a lot of data to process

A lot of data is available over the air (Big Data Era)

Improved computing abilities (GPUs)

Challenges:

How to acquire, store, and process this much data?

Artificial neural networks in the cloud or in each device?

How to integrate artificial neural networks into wireless networks?

Page 7: Learning to Optimize in Smart Radio Environmentsyuenchau/6GWorkshop... · Learning to Optimize 80 AI provides the tools to address C.2 and C.3. Complexity crunch in communication

Smart radio environments and AI

Smart Radio Environments ⇒ AI

Meta-surfaces provide data storage and processing abilities

Smart radio environments allow handling big data

Smart Radio Environments ⇐ AI

A lot of degrees of freedom to optimize

AI provides a framework for low-complexity system design

Page 8: Learning to Optimize in Smart Radio Environmentsyuenchau/6GWorkshop... · Learning to Optimize 80 AI provides the tools to address C.2 and C.3. Complexity crunch in communication

Smart radio environments and AI

Smart Radio Environments ⇒ AI

Meta-surfaces provide data storage and processing abilities

Smart radio environments allow handling big data

Smart Radio Environments ⇐ AI

A lot of degrees of freedom to optimize

AI provides a framework for low-complexity system design

Page 9: Learning to Optimize in Smart Radio Environmentsyuenchau/6GWorkshop... · Learning to Optimize 80 AI provides the tools to address C.2 and C.3. Complexity crunch in communication

MISO downlink with intelligent surfaces

C. Huang, A. Zappone, G. Alexandropoulos, M. Debbah, C. Yuen, ”LargeIntelligent Surface for Energy Efficiency in Wireless Communication”,

IEEE Transactions on Wireless Communications, submitted (minorrevision), 2019, https://arxiv.org/abs/1810.06934

Page 10: Learning to Optimize in Smart Radio Environmentsyuenchau/6GWorkshop... · Learning to Optimize 80 AI provides the tools to address C.2 and C.3. Complexity crunch in communication

Problem statement: EE Maximization

Large Intelligent Surface (LIS)

Base station

User

User

User

Reflecting Element

1

𝑘

𝐾

Upon transmit zero-forcing

maxΦ,P

∑Kk=1 log2

(1 + pkσ

−2)

ξ∑K

k=1 pk + PBS + KPUE + NPn(b)

s.t. log2

(1 + pkσ

−2)≥ Rmin,k ∀k = 1, 2, . . . ,K ,

tr((H2ΦH1)+P(H2ΦH1)+H) ≤ Pmax,

|φn| = 1 ∀n = 1, 2, . . . ,N,

Page 11: Learning to Optimize in Smart Radio Environmentsyuenchau/6GWorkshop... · Learning to Optimize 80 AI provides the tools to address C.2 and C.3. Complexity crunch in communication

Problem statement: EE Maximization

Large Intelligent Surface (LIS)

Base station

User

User

User

Reflecting Element

1

𝑘

𝐾

Upon transmit zero-forcing

maxΦ,P

∑Kk=1 log2

(1 + pkσ

−2)

ξ∑K

k=1 pk + PBS + KPUE + NPn(b)

s.t. log2

(1 + pkσ

−2)≥ Rmin,k ∀k = 1, 2, . . . ,K ,

tr((H2ΦH1)+P(H2ΦH1)+H) ≤ Pmax,

|φn| = 1 ∀n = 1, 2, . . . ,N,

Page 12: Learning to Optimize in Smart Radio Environmentsyuenchau/6GWorkshop... · Learning to Optimize 80 AI provides the tools to address C.2 and C.3. Complexity crunch in communication

Performance analysis

Alternating maximization of P and Φ

Optimization of P performed in closed-form

Two iterative methods to optimize Φ

(0,0)

Base Station

LIS

100m

100m

200m

ObstaclesUsers

19

Table I

SIMULATION AND ALGORITHMIC PARAMETERS

Parameters Values

RIS central element placement: (100m, 100m)

BS central element placement: (0, 0)

Small scale fading model 8i, k, j: [H1]ij , [h2,k]i ⇠ CN (0, 1)

Large scale fading model at distance d: 10�3.53

d3.76

Transmission bandwidth BW: 180kHz

Circuit dissipated power at BS PBS: 9dBW

Circuit dissipated power coefficients at BS ⇠ and AF relay ⇠AF: 1.2

Maximum transmit power at BS and AF relay Pmax=PR,max: 20dBW

Dissipated power at each user PUE: 10dBm

Dissipated power at the n-th RIS element Pn(b): 10dBm

Dissipated power at each AF relay transmit-receive antenna PR: 10dBm

Algorithmic convergence parameter: ✏ = 10�3

positions and channel realizations, generated according to the 3GPP propagation environment

described in [62], whose parameters are summarized in Table I. Therein, [H1]ij and [h2,k]i with

i = 1, 2, . . . , N , k = 1, 2, . . . , K, and j = 1, 2, . . . , M denote the (i, j)-th and i-th elements of

the respective matrices. In the table above, we also include the hardware dissipation parameters

of [10], [13] for BS, RIS, and the mobile users, as well as for the AF relay that will be used

for performance comparisons purposes. The relay is assumed to transmit with maximum power

PR,max, which is considered in all performance results equal to Pmax.

Without loss of generality, in the figures that follow we assume equal individual rate constraints

for all K users, i.e., Rmin,k = Rmin 8k. In addition, we set Rmin to a fraction of the rate that

each user would have in the genie case of mutually orthogonal channels and uniform power

allocation. In particular, this genie rate for each k-th mobile user is given by

R = log2

✓1 +

Pmax

K�2

◆. (36)

Thus, the QoS constraints depend on Pmax, which ensures that the minimum rate is commensurate

to the maximum power that can be used, in turn leading to a feasibility rate of Problem (8) that

is approximately constant with Pmax. Table II below shows the feasibility rate obtained for

Pmax = 20 dBW and different fractions R.

Page 13: Learning to Optimize in Smart Radio Environmentsyuenchau/6GWorkshop... · Learning to Optimize 80 AI provides the tools to address C.2 and C.3. Complexity crunch in communication

Comparison with AF relaying: Spectral Efficiency

M BS antennas, K mobile users, N reflecting elements

Page 14: Learning to Optimize in Smart Radio Environmentsyuenchau/6GWorkshop... · Learning to Optimize 80 AI provides the tools to address C.2 and C.3. Complexity crunch in communication

Comparison with AF relaying: Energy Efficiency

M BS antennas, K mobile users, N reflecting elements

Page 15: Learning to Optimize in Smart Radio Environmentsyuenchau/6GWorkshop... · Learning to Optimize 80 AI provides the tools to address C.2 and C.3. Complexity crunch in communication

Complexity crunch in communication networks

Relatively small system with ZF, but iterative algorithms needed

If multi-user interference is considered, problem is much harder

Optimal online resource allocation not feasible for larger systems

Learning to Optimize

80AI provides the tools to address C.2 and C.3

Page 16: Learning to Optimize in Smart Radio Environmentsyuenchau/6GWorkshop... · Learning to Optimize 80 AI provides the tools to address C.2 and C.3. Complexity crunch in communication

Complexity crunch in communication networks

Relatively small system with ZF, but iterative algorithms needed

If multi-user interference is considered, problem is much harder

Optimal online resource allocation not feasible for larger systems

Learning to Optimize

80AI provides the tools to address C.2 and C.3

Page 17: Learning to Optimize in Smart Radio Environmentsyuenchau/6GWorkshop... · Learning to Optimize 80 AI provides the tools to address C.2 and C.3. Complexity crunch in communication

Deep Learning

Learning by ANN

The distinctive trait of deep learning is to implement the learning process byartificial neural networks (ANN).

ANNs are universal function approximators (under very mild assumptions.)

ANNs exploit large datasets better than other machine learning techniques.

Page 18: Learning to Optimize in Smart Radio Environmentsyuenchau/6GWorkshop... · Learning to Optimize 80 AI provides the tools to address C.2 and C.3. Complexity crunch in communication

Artificial Neural Networks

ANN model

ANNs are organized hierarchically in layers of elementary processing units,called neurons.

An input layer forwards the input data to the rest of the network.

One or more hidden layers process the input data.

An output layer applies a final processing to the data before outputting it.

Weights and biases model the strength of the connections among neurons.

Hidden layerInput layer Output layer

Page 19: Learning to Optimize in Smart Radio Environmentsyuenchau/6GWorkshop... · Learning to Optimize 80 AI provides the tools to address C.2 and C.3. Complexity crunch in communication

Fully-connected networks

Hidden layerInput layer Output layer

The input of Layer ` is a vector x`−1. The output x`(n) of neuron i is:

x`(i) = fn,`(zi,`)

zi,` = wTi,`x`−1 + bi,` .

w i,` is a vector weighting the inputs from the previous layer.

bi,` is a bias term.

fi,`(zi,`) is called activation function of the neuron (elementary function).

Each neuron simply takes an affine combination of the inputs, computesthe value of the activation function, and propagates the result.

Page 20: Learning to Optimize in Smart Radio Environmentsyuenchau/6GWorkshop... · Learning to Optimize 80 AI provides the tools to address C.2 and C.3. Complexity crunch in communication

Training a neural network

ANNs are trained in a supervised fashion.

Training problem

Given the training set STR = {(xn, y∗n )}NTRn=1 , optimize all weights and bias of the

ANN, i.e. θ = {W tot , btot}, in order to minimize the training error.

minθ

NTR∑n=1

L(y n(θ), y∗n)

What about hyperparameters? (e.g. number of layers, neurons, etc.)Trial and error procedure (with some insight)

Page 21: Learning to Optimize in Smart Radio Environmentsyuenchau/6GWorkshop... · Learning to Optimize 80 AI provides the tools to address C.2 and C.3. Complexity crunch in communication

Training a neural network

ANNs are trained in a supervised fashion.

Training problem

Given the training set STR = {(xn, y∗n )}NTRn=1 , optimize all weights and bias of the

ANN, i.e. θ = {W tot , btot}, in order to minimize the training error.

minθ

NTR∑n=1

L(y n(θ), y∗n)

What about hyperparameters? (e.g. number of layers, neurons, etc.)Trial and error procedure (with some insight)

Page 22: Learning to Optimize in Smart Radio Environmentsyuenchau/6GWorkshop... · Learning to Optimize 80 AI provides the tools to address C.2 and C.3. Complexity crunch in communication

EE Maximization in Interference Networks

B. Matthiesen, A. Zappone, E. A. Jorswieck, M. Debbah, ”Deep Learningfor Optimal Energy-Efficient Power Control in Wireless Interference

Networks”, submitted, 2018, https://arxiv.org/abs/1812.06920

Page 23: Learning to Optimize in Smart Radio Environmentsyuenchau/6GWorkshop... · Learning to Optimize 80 AI provides the tools to address C.2 and C.3. Complexity crunch in communication

General Interference-limited network model

Interference network with K transmitters and M receivers.

Rx1

RxMpK

AF

h1

hK

g1

gM

p1 Rx1

RxMpK

p1

AF Rx1

pK

p1

The receive filter used by receiver m to decode the data from user k iscm,k .

All receivers have N receive antennas and thermal noise power σ2.

Each transmitter k has a single antenna and transmits with power pk .

The channel between transmitter k and receiver m is hk,m.

The SINR for transmitter k at receiver m is:

γk,m =|cH

m,khk,m|2pkσ2 +

∑j 6=k pj |cH

m,khj,m|2=

akpkσ2 +

∑j 6=k pjbk,j

Page 24: Learning to Optimize in Smart Radio Environmentsyuenchau/6GWorkshop... · Learning to Optimize 80 AI provides the tools to address C.2 and C.3. Complexity crunch in communication

Sum-EE maximization

Sum-EE maximization is the toughest EE maximization problem:

Sum-EE(p) =K∑

k=1

log2(1 + γk)

Pc + pk

pk ∈ [0,Pmax,k ] for all k.

B is the transmission bandwidth.

Pc is the total hardware power dissipated in all network nodes.

Alternative problem formulation

We can write the problem as the computation of the function:

F (a, b,Pmax) = arg maxp∈S Sum-EE(p, a, b,Pmax)

with a = {ak}k , b = {bj,k}j,k , Pmax = {Pmax,k}k .

Feedforward neural networks are universal function approximators.An FFN with input (a, b,Pmax) and output p can learn F.

Page 25: Learning to Optimize in Smart Radio Environmentsyuenchau/6GWorkshop... · Learning to Optimize 80 AI provides the tools to address C.2 and C.3. Complexity crunch in communication

Sum-EE maximization

Sum-EE maximization is the toughest EE maximization problem:

Sum-EE(p) =K∑

k=1

log2(1 + γk)

Pc + pk

pk ∈ [0,Pmax,k ] for all k.

B is the transmission bandwidth.

Pc is the total hardware power dissipated in all network nodes.

Alternative problem formulation

We can write the problem as the computation of the function:

F (a, b,Pmax) = arg maxp∈S Sum-EE(p, a, b,Pmax)

with a = {ak}k , b = {bj,k}j,k , Pmax = {Pmax,k}k .

Feedforward neural networks are universal function approximators.An FFN with input (a, b,Pmax) and output p can learn F.

Page 26: Learning to Optimize in Smart Radio Environmentsyuenchau/6GWorkshop... · Learning to Optimize 80 AI provides the tools to address C.2 and C.3. Complexity crunch in communication

Sum-EE maximization

Sum-EE maximization is the toughest EE maximization problem:

Sum-EE(p) =K∑

k=1

log2(1 + γk)

Pc + pk

pk ∈ [0,Pmax,k ] for all k.

B is the transmission bandwidth.

Pc is the total hardware power dissipated in all network nodes.

Alternative problem formulation

We can write the problem as the computation of the function:

F (a, b,Pmax) = arg maxp∈S Sum-EE(p, a, b,Pmax)

with a = {ak}k , b = {bj,k}j,k , Pmax = {Pmax,k}k .

Feedforward neural networks are universal function approximators.An FFN with input (a, b,Pmax) and output p can learn F.

Page 27: Learning to Optimize in Smart Radio Environmentsyuenchau/6GWorkshop... · Learning to Optimize 80 AI provides the tools to address C.2 and C.3. Complexity crunch in communication

Implementation and complexity

Algorithm

Generate a training set {(ai , bi ,Pmax), pi}i by maximizing the Sum-EE.

Train the FNN to adjust parameters and hyperparameters.

Use the trained network to obtain the optimal power allocation for new(not in the training set) channels.

Complexity

The trained FNN provides a formula for the optimal power allocation.

Training the FNN can be done offline and sporadically.

Complexity of training reduced by the optimization approach in [1].

References[1] B. Matthiesen, A. Zappone, E. A. Jorswieck, and M. Debbah, “Deep learning for optimal energy-efficient

power control in wireless interference networks,” http://export.arxiv.org/pdf/1812.06920, 2018

Page 28: Learning to Optimize in Smart Radio Environmentsyuenchau/6GWorkshop... · Learning to Optimize 80 AI provides the tools to address C.2 and C.3. Complexity crunch in communication

Implementation and complexity

Algorithm

Generate a training set {(ai , bi ,Pmax), pi}i by maximizing the Sum-EE.

Train the FNN to adjust parameters and hyperparameters.

Use the trained network to obtain the optimal power allocation for new(not in the training set) channels.

Complexity

The trained FNN provides a formula for the optimal power allocation.

Training the FNN can be done offline and sporadically.

Complexity of training reduced by the optimization approach in [1].

References[1] B. Matthiesen, A. Zappone, E. A. Jorswieck, and M. Debbah, “Deep learning for optimal energy-efficient

power control in wireless interference networks,” http://export.arxiv.org/pdf/1812.06920, 2018

Page 29: Learning to Optimize in Smart Radio Environmentsyuenchau/6GWorkshop... · Learning to Optimize 80 AI provides the tools to address C.2 and C.3. Complexity crunch in communication

Neural network architecture

Layer Type Size Activation functionInput 20 –

Layer 1 (fully-connected) 128 eluLayer 2 (fully-connected) 64 reluLayer 3 (fully-connected) 32 eluLayer 4 (fully-connected) 16 reluLayer 5 (fully-connected) 8 eluLayer 6 (fully-connected) 4 linear

Supervised learning (Keras)

ADAM with MSE

Training set: NTR = 102, 000. Validation set of NV = 10, 200.

Testing set: NTest = 10, 000 for each considered value ofPmax = −30, . . . , 20 dB, with 1 dB step.

The total number of generated data samples is 622,000.

Interference channel with K = 4 links in a square area with edge of 2km.

Single-antenna transmitters; receivers with N = 2 antennas each.

Rayleigh fast-fading, path-loss, (and shadowing).

Page 30: Learning to Optimize in Smart Radio Environmentsyuenchau/6GWorkshop... · Learning to Optimize 80 AI provides the tools to address C.2 and C.3. Complexity crunch in communication

Numerical analysis: Training performance

The 622,000 data samples were generated in 8.4 CPU hourson Intel Haswell nodes with Xeon E5-2680 v3 CPUs running at 2.50 GHz.

0 20 40 60 80 1000

0.2

0.4

0.6

0.8

1

Epoch

Mea

nSq

uare

dE

rror

Training LossValidation Loss

Both training and validation error monotonically decrease.Neither underfitting nor overfitting occurs.

Page 31: Learning to Optimize in Smart Radio Environmentsyuenchau/6GWorkshop... · Learning to Optimize 80 AI provides the tools to address C.2 and C.3. Complexity crunch in communication

Numerical analysis: Testing performance

−30 −20 −10 0 10 200

0.5

1

1.5

2

2.5

Pmax [dBW]

WSE

E[M

bit/J

oule

]

Optimal ANN SCASCAos Max. Power Best only

The FNN achieves optimal performance.

SCA performs close to the FNN only with a sophisticated initialization rule.

Page 32: Learning to Optimize in Smart Radio Environmentsyuenchau/6GWorkshop... · Learning to Optimize 80 AI provides the tools to address C.2 and C.3. Complexity crunch in communication

For more applications

A. Zappone, M. Di Renzo, M. Debbah, T. T. Lam, X. Qian,”Model-Aided Wireless Artificial Intelligence: Embedding ExpertKnowledge in Deep Neural Networks Towards Wireless SystemsOptimization”, submitted, 2018, https://arxiv.org/abs/1808.01672

A. Zappone, L. Sanguinetti, M. Debbah, ”User Association (and LoadBalancing) for Massive MIMO through Deep Learning”, IEEEAsilomar 2018

L. Sanguinetti, A. Zappone, M. Debbah, ”On-line power allocation inMassive MIMO”, IEEE Asilomar 2018

Page 33: Learning to Optimize in Smart Radio Environmentsyuenchau/6GWorkshop... · Learning to Optimize 80 AI provides the tools to address C.2 and C.3. Complexity crunch in communication

Thank you for listening

Alessio Zappone

LANEAS, CentraleSupelec, Paris, France

[email protected]

Page 34: Learning to Optimize in Smart Radio Environmentsyuenchau/6GWorkshop... · Learning to Optimize 80 AI provides the tools to address C.2 and C.3. Complexity crunch in communication

B. Matthiesen, A. Zappone, E. A. Jorswieck, and M. Debbah, “Deeplearning for optimal energy-efficient power control in wireless interferencenetworks,” http://export.arxiv.org/pdf/1812.06920, 2018.