a class of separable convex optimization problems in

22
A class of separable convex optimization problems in communication Arun Padakandla joint work with Dr. Rajesh Sundaresan Dept of Electrical Communication Engineering, Indian Institute of Science, Bangalore.

Upload: others

Post on 14-Nov-2021

4 views

Category:

Documents


0 download

TRANSCRIPT

Page 1: A class of separable convex optimization problems in

A class of separable convex optimization

problems in communication

Arun Padakandlajoint work with Dr. Rajesh Sundaresan

Dept of Electrical Communication Engineering,

Indian Institute of Science, Bangalore.

Page 2: A class of separable convex optimization problems in

A simple separable convex optimization problem in

geometry

Let L be a natural number.

min∑L

l=1 x2l

subject to 0 ≤ xl ≤ βl∑

L

l=1 xl = K

For illustration consider L = 2,Applications in MVU estimator of parameters in wireless sensor networks(Zacharias and Sundaresan [2007]).

Page 3: A class of separable convex optimization problems in

Single user parallel Gaussian channel

L independent channels with noise variancesσ2

1 ≤ · · · ≤ σ2L

Y = X + N

Sum power constraint P Joules per channel use.

Communicating independently on the L channels is optimal

Goal : Identify powers P1,P2, · · · ,PL satisfying∑

L

l=1 Pl = P toachieve maximum rate?

Allocating power Pl on channel with noise variance σ2l

yields a rate12

log(

1 + Pl

σ2l

)

Page 4: A class of separable convex optimization problems in

Single user parallel Gaussian channel

max∑

L

l=112log

(

1 + Pl

σ2l

)

subject to Pl ≥ 0 for l = 1, 2, · · · ,L

∑L

l=1 Pl ≤ P

The solution is a technique popularly called water-filling.

The problem involves

◮ maximizing a separable concave function

◮ subject to positivity and a linear constraint

Page 5: A class of separable convex optimization problems in

Multiple users parallel Gaussian channel

K user multiple access channel.

Bandwidth constraint restricts signals to a vectorspace of L dimensions, K > L.

Users assigned vectors (directions in RL) that

are modulated.

Consider an allocation of vectors and let(µ1, · · · , µL) =eig( signal + interference matrix).

Each user is assigned a single vector and hencewill operate along one direction in contrast tothe single user parallel Gaussian case.

Page 6: A class of separable convex optimization problems in

Multiple users parallel Gaussian channelµ must satisfy the followingpowers are allocated in every direction i.e.,

µl − σ2l≥ 0 for l = 1, 2, · · · , L

the first user (largest power constraint) can beaccomodated in the least noisy dimension i.e.

µ1 − σ21 ≥ P1,

the first n users (large power constraint) can beaccomodated in the least n noisy dimension i.e.

n∑

l=1

µl − σ2l ≥

n∑

l=1

Pl , for n = 1, 2, · · · , L− 1

the sum power constraint is obeyed

L∑

l=1

µl − σ2l

=

K∑

k=1

Pk .

Page 7: A class of separable convex optimization problems in

Multiple users parallel Gaussian channel

µ satisfies above constraints ⇒ an allocation of directions to theusers.For simplicity let xl

def= µl − σ2

l

L∑

l=1

1

2log (µl)

︸ ︷︷ ︸

log det (S + I)

L∑

l=1

1

2log

(σ2

l

)

︸ ︷︷ ︸

log det (noise)

=∑

L

l=112

log(

1 + xl

σ2l

)

This is the sum rate of the K users.

Page 8: A class of separable convex optimization problems in

Multiple users parallel Gaussian channel

Sum rate maximization can be stated as

maximize∑

L

l=1 log(

1 + xl

σ2l

)

subject to xl ≥ 0 for l = 1, 2, · · · ,L

∑n

l=1 xl ≥∑

n

k=1 Pk , for n = 1, 2, · · · ,L− 1

∑L

l=1 xl =∑

K

k=1 Pk .

Single user parallel Gaussian channel power allocation problem hadonly the first positivity and the last equality constraint.

Page 9: A class of separable convex optimization problems in

Abstraction of the problem

Suppose gl , l = 1, · · · , L be functions that satisfy the following:

◮ gl : (al , bl)→ R where 0 ∈ (al , bl)

◮ gl strictly convex, continuously differentiable in (al , bl);

◮ g ′

l(0) are in increasing order with respect to the index l i.e.:

g ′

1(0) ≤ g ′

2(0) ≤ · · · ≤ g ′

L(0); (1)

min∑L

l=1 gl(xl )

subject to xl ≥ 0 for l = 1, · · · , L

∑n

l=1 xl ≥∑

n

k=1 αk , for n = 1, · · · , L− 1

∑L

l=1 xl =∑K

k=1 αk .

Page 10: A class of separable convex optimization problems in

An explanation of the finite step algorithm

min∑

L

l=1 gl(xl ) gl strictly convex, continuously differentiable

subject to∑

L

l=1 xl =∑

K

k=1 αk .

Page 11: A class of separable convex optimization problems in

An explanation of the finite step algorithm

min∑

L

l=1 gl(xl ) gl strictly convex, continuously differentiable

subject to∑

L

l=1 xl =∑

K

k=1 αk .

Suppose you find (x1, x2, · · · , xL) ∈ RL+ such that

x1 + x2 + · · ·+ xL = α1 + α2 + · · ·+ αK

and

h1(x1) = h2(x2) = · · · = hL(xL)def= ΘL

1

Page 12: A class of separable convex optimization problems in

An explanation of the finite step algorithm

min∑

L

l=1 gl(xl ) gl strictly convex, continuously differentiable

subject to∑

L

l=1 xl =∑

K

k=1 αk .

Suppose you find (x1, x2, · · · , xL) ∈ RL+ such that

x1 + x2 + · · ·+ xL = α1 + α2 + · · ·+ αK

and

h1(x1) = h2(x2) = · · · = hL(xL)def= ΘL

1

Tempted to declare (x1, x2, · · · , xL) =(h−1

1 (ΘL1), h

−12 (ΘL

2), · · · , h−1L

(ΘL1)

)

as the solution

Page 13: A class of separable convex optimization problems in

An explanation of the finite step algorithm

Wait!!Suppose there exists (x ′

1, x′

2, · · · , x′

n) ∈ Rn+(n < L) such that

x ′

1 + x ′

2 + · · ·+ x ′

n = α1 + α2 + · · · + αn

and

h1(x′

1) = h2(x′

2) = · · · = hn(x′

n)def= θn

1 < ΘL1

Page 14: A class of separable convex optimization problems in

An explanation of the finite step algorithm

Wait!!Suppose there exists (x ′

1, x′

2, · · · , x′

n) ∈ Rn+(n < L) such that

x ′

1 + x ′

2 + · · ·+ x ′

n = α1 + α2 + · · · + αn

and

h1(x′

1) = h2(x′

2) = · · · = hn(x′

n)def= θn

1 < ΘL1

You are in trouble. Note that

x1 + x2 + · · ·+ xn =∑

n

l=1 h−1m (ΘL

1)

>∑n

l=1 h−1m

(θn1) h−1

mis strictly increasing

= x ′

1 + x ′

2 + · · ·+ x ′

n

= α1 + α2 + · · ·+ αn

(2)

(x1, · · · , xL) =(h−1

1 (ΘL1), h

−12 (ΘL

2), · · · , h−1L

(ΘL1)

)is not feasible!!!

Page 15: A class of separable convex optimization problems in

An explanation of the finite step algorithm

Let us look at

ξ1 = max{ΘL

1 , hL(0), θn

1 ; n = 1, 2, · · · , L− 1}

where θn

1 defines a (x ′

1, x′

2, · · · , x′

n)

θn

1

def= h1(x

1) = h2(x′

2) = · · · = hn(x′

n)

and

x ′

1 + x ′

2 + · · ·+ x ′

n= α1 + α2 + · · ·+ αn.

Page 16: A class of separable convex optimization problems in

An explanation of the finite step algorithm

Let us look at

ξ1 = max{ΘL

1 , hL(0), θn

1 ; n = 1, 2, · · · , L− 1}

where θn

1 defines a (x ′

1, x′

2, · · · , x′

n)

θn

1

def= h1(x

1) = h2(x′

2) = · · · = hn(x′

n)

and

x ′

1 + x ′

2 + · · ·+ x ′

n= α1 + α2 + · · ·+ αn.

ξ1 is picked max to ensure feasibility.

Page 17: A class of separable convex optimization problems in

An explanation of the finite step algorithm

Let us look at

ξ1 = max{ΘL

1 , hL(0), θn

1 ; n = 1, 2, · · · , L− 1}

where θn

1 defines a (x ′

1, x′

2, · · · , x′

n)

θn

1

def= h1(x

1) = h2(x′

2) = · · · = hn(x′

n)

and

x ′

1 + x ′

2 + · · ·+ x ′

n= α1 + α2 + · · ·+ αn.

ξ1 is picked max to ensure feasibility.

If hL(0) is the largest candidate ⇒ marginal cost too large.

Set xL = 0 and focus on reduced problem with L now equal to L− 1

Page 18: A class of separable convex optimization problems in

An explanation of the finite step algorithm

Suppose ξ1 = ΘL

1 . The intermediate ladder constraints did not matter.

Page 19: A class of separable convex optimization problems in

An explanation of the finite step algorithm

Suppose ξ1 = ΘL

1 . The intermediate ladder constraints did not matter.

Suppose ξ1 = θn

1 for some n < L. Set xl ← x ′

lfor l = 1, 2, · · · , n

Page 20: A class of separable convex optimization problems in

An explanation of the finite step algorithm

Suppose ξ1 = ΘL

1 . The intermediate ladder constraints did not matter.

Suppose ξ1 = θn

1 for some n < L. Set xl ← x ′

lfor l = 1, 2, · · · , n

By definition of θn

1 we have met

n∑

l=1

xl ≥

n∑

k=1

αk ,

constraint with equality

Page 21: A class of separable convex optimization problems in

An explanation of the finite step algorithm

Suppose ξ1 = ΘL

1 . The intermediate ladder constraints did not matter.

Suppose ξ1 = θn

1 for some n < L. Set xl ← x ′

lfor l = 1, 2, · · · , n

By definition of θn

1 we have met

n∑

l=1

xl ≥

n∑

k=1

αk ,

constraint with equality and furthermore θn1 being maximum of the

candidates the assignment respects

l∑

l=1

xl ≥

l∑

k=1

αk , for l ≥ 1, · · · , n − 1.

Thus focus on the setting the variables xl+1, · · · , xL.

Page 22: A class of separable convex optimization problems in

Concluding remarks

We focused on a separable convex optimization problem with linearascending constraints.

Applications in optimal sequence design for CDMA.

We provided a finite step algorithm to identify the optimal solution.