• 沒有找到結果。

Convex optimization problems

N/A
N/A
Protected

Academic year: 2022

Share "Convex optimization problems"

Copied!
24
0
0

加載中.... (立即查看全文)

全文

(1)

Convex optimization problems

! optimization problem in standard form

! convex optimization problems

! linear optimization

! quadratic optimization

! geometric programming

! quasiconvex optimization

! generalized inequality constraints

! semidefinite programming

! vector optimization

IOE 611: Nonlinear Programming, Winter 2006 4. Convex optimization problems Page 4–1

Optimization problem in standard form

minimize f0(x)

subject to fi(x)≤ 0, i = 1, . . . , m hi(x) = 0, i = 1, . . . , p

! x ∈ Rn is the optimization variable

! f0 : Rn → R is the objective or cost function

! fi : Rn → R, i = 1, . . . , m, are the inequality constraint functions

! hi : Rn → R are the equality constraint functions optimal value:

p! = inf{f0(x)| fi(x) ≤ 0, i = 1, . . . , m, hi(x) = 0, i = 1, . . . , p}

! p! = ∞ if problem is infeasible (no x satisfies the constraints)

! p! = −∞ if problem is unbounded below

IOE 611: Nonlinear Programming, Winter 2006 4. Convex optimization problems Page 4–2

(2)

Optimal and locally optimal points

! x is feasible if x ∈ dom f0 and it satisfies the constraints

! a feasible x is optimal if f0(x) = p!; Xopt is the set of optimal points

! x is locally optimal if there is an R > 0 such that x is optimal for

minimize (over z) f0(z)

subject to fi(z) ≤ 0, i = 1, . . . , m, hi(z) = 0, i = 1, . . . , p

&z − x&2 ≤ R examples (with n = 1, m = p = 0)

! f0(x) = 1/x, dom f0 = R++: p! = 0, no optimal point

! f0(x) = − log x, dom f0 = R++: p! = −∞

! f0(x) = x log x, dom f0 = R++: p! = −1/e, x = 1/e is optimal

! f0(x) = x3 − 3x, p! = −∞, local optimum at x = 1

IOE 611: Nonlinear Programming, Winter 2006 4. Convex optimization problems Page 4–3

Implicit constraints

The standard form optimization problem has an implicit constraint

x ∈ D =

!m i=0

dom fi

!p i=1

dom hi,

! we call D the domain of the problem

! the constraints fi(x)≤ 0, hi(x) = 0 are the explicit constraints

! a problem is unconstrained if it has no explicit constraints (m = p = 0)

example:

minimize f0(x) =−"k

i=1log(bi − aTi x) is an unconstrained problem with implicit constraints aTi x < bi, i = 1, . . . , k

IOE 611: Nonlinear Programming, Winter 2006 4. Convex optimization problems Page 4–4

(3)

Feasibility problem

find x

subject to fi(x)≤ 0, i = 1, . . . , m hi(x) = 0, i = 1, . . . , p

can be considered a special case of the general problem with f0(x) = 0:

minimize 0

subject to fi(x)≤ 0, i = 1, . . . , m hi(x) = 0, i = 1, . . . , p

! p! = 0 if constraints are feasible; any feasible x is optimal

! p! = ∞ if constraints are infeasible

IOE 611: Nonlinear Programming, Winter 2006 4. Convex optimization problems Page 4–5

Convex optimization problem

standard form convex optimization problem minimize f0(x)

subject to fi(x)≤ 0, i = 1, . . . , m aTi x = bi, i = 1, . . . , p

! f0, f1, . . . , fm are convex; equality constraints are affine

! problem is quasiconvex if f0 is quasiconvex (and f1, . . . , fm

convex) often written as

minimize f0(x)

subject to fi(x)≤ 0, i = 1, . . . , m Ax = b

important property: feasible set of a convex optimization problem is convex

IOE 611: Nonlinear Programming, Winter 2006 4. Convex optimization problems Page 4–6

(4)

example

minimize f0(x) = x12 + x22

subject to f1(x) = x1/(1 + x22) ≤ 0 h1(x) = (x1 + x2)2 = 0

! f0 is convex; feasible set {(x1, x2)| x1 = −x2 ≤ 0} is convex

! not a convex problem (according to our definition): f1 is not convex, h1 is not affine

! equivalent (but not identical) to the convex problem minimize x12 + x22

subject to x1 ≤ 0 x1 + x2 = 0

IOE 611: Nonlinear Programming, Winter 2006 4. Convex optimization problems Page 4–7

Local and global optima

Any locally optimal point of a convex problem is (globally) optimal proof: suppose x is locally optimal and y is optimal with

f0(y ) < f0(x).

“x locally optimal” means there is an R > 0 such that z feasible, &z − x&2 ≤ R = f0(z) ≥ f0(x).

Consider z = θy + (1− θ)x with θ = R/(2&y − x&2)

! &y − x&2 > R, so 0 < θ < 1/2

! z is a convex combination of two feasible points, hence also feasible

! &z − x&2 = R/2 and

f0(z)≤ θf0(x) + (1− θ)f0(y ) < f0(x),

which contradicts our assumption that x is locally optimal

IOE 611: Nonlinear Programming, Winter 2006 4. Convex optimization problems Page 4–8

(5)

Optimality criterion for differentiable f

0

x is optimal if and only if it is feasible and

∇f0(x)T(y − x) ≥ 0 for all feasible y

Optimality criterion for differentiable f

0

x is optimal if and only if it is feasible and

∇f0(x)T(y− x) ≥ 0 for all feasible y

PSfrag replacements

−∇f0(x)

X x

if nonzero, ∇f0(x) defines a supporting hyperplane to feasible set X at x

Convex optimization problems 4–9

if nonzero, ∇f0(x) defines a supporting hyperplane to feasible set X at x

IOE 611: Nonlinear Programming, Winter 2006 4. Convex optimization problems Page 4–9

Optimality criterion: special cases

! unconstrained problem: x is optimal if and only if x ∈ dom f0, ∇f0(x) = 0

! equality constrained problem

minimize f0(x) subject to Ax = b x is optimal if and only if there exists a ν such that

x ∈ dom f0, Ax = b, ∇f0(x) + ATν = 0

! minimization over nonnegative orthant

minimize f0(x) subject to x + 0 x is optimal if and only if

x ∈ dom f0, x + 0,

# ∇f0(x)i ≥ 0 xi = 0

∇f0(x)i = 0 xi > 0

IOE 611: Nonlinear Programming, Winter 2006 4. Convex optimization problems Page 4–10

(6)

Equivalent convex problems

two problems are (informally) equivalent if the solution of one is readily obtained from the solution of the other, and vice-versa.

Some common transformations that preserve convexity:

! eliminating equality constraints minimize f0(x)

subject to fi(x)≤ 0, i = 1, . . . , m Ax = b

is equivalent to

minimize (over z) f0(Fz + x0)

subject to fi(Fz + x0) ≤ 0, i = 1, . . . , m where F and x0 are such that

Ax = b ⇐⇒ x = Fz + x0 for some z

IOE 611: Nonlinear Programming, Winter 2006 4. Convex optimization problems Page 4–11

! introducing equality constraints minimize f0(A0x + b0)

subject to fi(Aix + bi) ≤ 0, i = 1, . . . , m is equivalent to

minimize (over x, yi) f0(y0)

subject to fi(yi) ≤ 0, i = 1, . . . , m

yi = Aix + bi, i = 0, 1, . . . , m

! introducing slack variables for linear inequalities minimize f0(x)

subject to aTi x ≤ bi, i = 1, . . . , m is equivalent to

minimize (over x, s) f0(x)

subject to aTi x + si = bi, i = 1, . . . , m si ≥ 0, i = 1, . . . m

IOE 611: Nonlinear Programming, Winter 2006 4. Convex optimization problems Page 4–12

(7)

! epigraph form: standard form convex problem is equivalent to

minimize (over x, t) t

subject to f0(x)− t ≤ 0

fi(x)≤ 0, i = 1, . . . , m Ax = b

! minimizing over some variables minimize f0(x1, x2)

subject to fi(x1) ≤ 0, i = 1, . . . , m is equivalent to

minimize f˜0(x1)

subject to fi(x1) ≤ 0, i = 1, . . . , m where ˜f0(x1) = infx2 f0(x1, x2)

IOE 611: Nonlinear Programming, Winter 2006 4. Convex optimization problems Page 4–13

Linear program (LP)

minimize cTx + d subject to Gx - h

Ax = b

! convex problem with affine objective and constraint functions

! feasible set is a polyhedron

Linear program (LP)

minimize cTx + d subject to Gx ! h Ax = b

• convex problem with affine objective and constraint functions

• feasible set is a polyhedron

PSfrag replacements P x!

−c

Convex optimization problemsIOE 611: Nonlinear Programming, Winter 2006 4. Convex optimization problems Page 4–14 4–17

(8)

Examples

diet problem: choose quantities x1, . . . , xn of n foods

! one unit of food j costs cj, contains amount aij of nutrient i

! healthy diet requires nutrient i in quantity at least bi to find cheapest healthy diet,

minimize cTx

subject to Ax + b, x + 0

piecewise-linear minimization

minimize maxi=1,...,m(aTi x + bi) equivalent to an LP

minimize t

subject to aTi x + bi ≤ t, i = 1, . . . , m

IOE 611: Nonlinear Programming, Winter 2006 4. Convex optimization problems Page 4–15

Chebyshev center of a polyhedron

Chebyshev center of

P = {x | aTi x ≤ bi, i = 1, . . . , m} is center of largest inscribed ball

B = {xc + u | &u&2 ≤ r}

Chebyshev center of a polyhedron Chebyshev center of

P = {x | aTi x≤ bi, i = 1, . . . , m} is center of largest inscribed ball

B = {xc+ u| "u"2 ≤ r}PSfrag replacements

xcheb

xcheb

• aTi x≤ bi for all x∈ B if and only if

sup{aTi (xc+ u)| "u"2 ≤ r} = aTi xc+ r"ai"2 ≤ bi

• hence, xc, r can be determined by solving the LP maximize r

subject to aTi xc+ r"ai"2 ≤ bi, i = 1, . . . , m

Convex optimization problems 4–19

! aTi x ≤ bi for all x ∈ B if and only if

sup{aTi (xc + u)| &u&2 ≤ r} = aTi xc + r&ai&2 ≤ bi

! hence, xc, r can be determined by solving the LP maximizexc,r r

subject to aTi xc + r&ai&2 ≤ bi, i = 1, . . . , m

IOE 611: Nonlinear Programming, Winter 2006 4. Convex optimization problems Page 4–16

(9)

Quadratic program (QP)

minimize (1/2)xTPx + qTx + r subject to Gx - h

Ax = b

! P ∈ Sn+, so objective is convex quadratic

! minimize a convex quadratic function over a polyhedron Quadratic program (QP)

minimize (1/2)xTP x + qTx + r subject to Gx! h

Ax = b

• P ∈ Sn+, so objective is convex quadratic

• minimize a convex quadratic function over a polyhedron

PSfrag replacements

P x!

−∇f0(x!)

Convex optimization problems 4–22

IOE 611: Nonlinear Programming, Winter 2006 4. Convex optimization problems Page 4–17

Examples

least-squares

minimize &Ax − b&22

! analytical solution x! = Ab (A is pseudo-inverse)

! can add linear constraints, e.g., l - x - u linear program with random cost

minimize ¯cTx + γxTΣx = E cTx + γ var(cTx) subject to Gx - h, Ax = b

! c is random vector with mean ¯c and covariance Σ

! hence, cTx is random variable with mean ¯cTx and variance xTΣx

! γ > 0 is risk aversion parameter; controls the trade-off between expected cost and variance (risk)

IOE 611: Nonlinear Programming, Winter 2006 4. Convex optimization problems Page 4–18

(10)

Quadratically constrained quadratic program (QCQP)

minimize (1/2)xTP0x + q0Tx + r0

subject to (1/2)xTPix + qiTx + ri ≤ 0, i = 1, . . . , m Ax = b

! Pi ∈ Sn+; objective and constraints are convex quadratic

! if P1, . . . , Pm ∈ Sn++, feasible region is intersection of m ellipsoids and an affine set

IOE 611: Nonlinear Programming, Winter 2006 4. Convex optimization problems Page 4–19

Second-order cone programming

minimize fTx

subject to &Aix + bi&2 ≤ ciTx + di, i = 1, . . . , m Fx = g

(Ai ∈ Rni×n, F ∈ Rp×n)

! inequalities are called second-order cone (SOC) constraints:

(Aix + bi, ciTx + di) ∈ second-order cone in Rni+1

! for ni = 0, reduces to an LP; if ci = 0, reduces to a QCQP

! more general than QCQP and LP

IOE 611: Nonlinear Programming, Winter 2006 4. Convex optimization problems Page 4–20

(11)

Robust linear programming

the parameters in optimization problems are often uncertain, e.g., in an LP

minimize cTx

subject to aTi x ≤ bi, i = 1, . . . , m, there can be uncertainty in c, ai, bi

two common approaches to handling uncertainty (in ai, for simplicity)

! deterministic model: constraints must hold for all ai ∈ Ei

minimize cTx

subject to aTi x ≤ bi for all ai ∈ Ei, i = 1, . . . , m,

! stochastic model: ai is random variable; constraints must hold with probability η

minimize cTx

subject to prob(aiTx ≤ bi) ≥ η, i = 1, . . . , m

IOE 611: Nonlinear Programming, Winter 2006 4. Convex optimization problems Page 4–21

Deterministic approach via SOCP

! choose an ellipsoid as Ei:

Ei = {¯ai + Piu | &u&2 ≤ 1} (¯ai ∈ Rn, Pi ∈ Rn×n) center is ¯ai, semi-axes determined by singular values/vectors of Pi

! robust LP

minimize cTx

subject to aTi x ≤ bi ∀ai ∈ Ei, i = 1, . . . , m is equivalent to the SOCP

minimize cTx

subject to ¯aiTx +&PiTx&2 ≤ bi, i = 1, . . . , m (follows from sup"u"2≤1(¯ai + Piu)Tx = ¯aTi x +&PiTx&2)

IOE 611: Nonlinear Programming, Winter 2006 4. Convex optimization problems Page 4–22

(12)

Stochastic approach via SOCP

! assume ai is Gaussian with mean ¯ai, covariance Σi (ai ∼ N (¯ai, Σi))

! aTi x is Gaussian r.v. with mean ¯aiTx, variance xTΣix; hence

prob(aTi x ≤ bi) = Φ

%bi − ¯aTi x

1/2i x&2

&

where Φ(x) = (1/√

2π)'x

−∞e−t2/2dt is CDF of N (0, 1)

! robust LP

minimize cTx

subject to prob(aiTx ≤ bi) ≥ η, i = 1, . . . , m, with η ≥ 1/2, is equivalent to the SOCP

minimize cTx

subject to ¯aTi x + Φ−1(η)&Σ1/2i x&2 ≤ bi, i = 1, . . . , m

IOE 611: Nonlinear Programming, Winter 2006 4. Convex optimization problems Page 4–23

Geometric programming

monomial function

f (x) = cx1a1x2a2 · · · xnan, dom f = Rn++

with c > 0; exponent αi can be any real number posynomial function: sum of monomials

f (x) =

$K k=1

ckx1a1kx2a2k · · · xnank, dom f = Rn++

geometric program (GP) minimize f0(x)

subject to fi(x)≤ 1, i = 1, . . . , m hi(x) = 1, i = 1, . . . , p with fi posynomial, hi monomial

IOE 611: Nonlinear Programming, Winter 2006 4. Convex optimization problems Page 4–24

(13)

Geometric program in convex form

change variables to yi = log xi, and take logarithm of cost, constraints

! monomial f (x) = cx1a1· · · xnan transforms to

log f (ey1, . . . , eyn) = aTy + b (b = log c)

! posynomial f (x) = "K

k=1ckx1a1kx2a2k · · · xnank transforms to log f (ey1, . . . , eyn) = log

% K

$

k=1

eaTky +bk

&

(bk = log ck)

! geometric program transforms to convex problem minimize log("K

k=1exp(aT0ky + b0k)) subject to log("K

k=1exp(aTiky + bik))

≤ 0, i = 1, . . . , m Gy + d = 0

IOE 611: Nonlinear Programming, Winter 2006 4. Convex optimization problems Page 4–25

Design of cantilever beam Design of cantilever beam

PSfrag replacements

F segment 4 segment 3 segment 2 segment 1

• N segments with unit lengths, rectangular cross-sections of size w

i

× h

i

• given vertical force F applied at the right end design problem

minimize total weight

subject to upper & lower bounds on w

i

, h

i

upper bound & lower bounds on aspect ratios h

i

/w

i

upper bound on stress in each segment

upper bound on vertical deflection at the end of the beam variables: w

i

, h

i

for i = 1, . . . , N

Convex optimization problems 4–31

! N segments with unit lengths, rectangular cross-sections of size wi × hi

! given vertical force F applied at the right end design problem

minimize total weight

subject to upper & lower bounds on wi, hi

upper bound & lower bounds on aspect ratios hi/wi upper bound on stress in each segment

upper bound on vertical deflection at the end of the beam variables: wi, hi for i = 1, . . . , N

IOE 611: Nonlinear Programming, Winter 2006 4. Convex optimization problems Page 4–26

(14)

Objective and constraint functions

! total weight w1h1 +· · · + wNhN is posynomial

! aspect ratio hi/wi and inverse aspect ratio wi/hi are monomials

! maximum stress in segment i is given by 6iF /(wih2i), a monomial

! the vertical deflection yi and slope vi of central axis at the right end of segment i are defined recursively as

vi = 12(i − 1/2) F

Ewih3i + vi+1

yi = 6(i − 1/3) F

Ewihi3 + vi+1+ yi+1

for i = N, N − 1, . . . , 1, with vN+1 = yN+1 = 0 (E is Young’s modulus)

vi and yi are posynomial functions of w , h

IOE 611: Nonlinear Programming, Winter 2006 4. Convex optimization problems Page 4–27

Formulation as a GP

minimize w1h1 +· · · + wNhN

subject to wmax−1 wi ≤ 1, wminwi−1 ≤ 1, i = 1, . . . , N h−1maxhi ≤ 1, hminh−1i ≤ 1, i = 1, . . . , N

Smax−1 wi−1hi ≤ 1, Sminwih−1i ≤ 1, i = 1, . . . , N 6iF σmax−1 wi−1h−2i ≤ 1, i = 1, . . . , N

ymax−1 y1 ≤ 1 note

! we write wmin ≤ wi ≤ wmax and hmin ≤ hi ≤ hmax

wmin/wi ≤ 1, wi/wmax ≤ 1, hmin/hi ≤ 1, hi/hmax ≤ 1

! we write Smin ≤ hi/wi ≤ Smax as

Sminwi/hi ≤ 1, hi/(wiSmax) ≤ 1

! The number of monomials appearing in y1 grows approximately as N2.

IOE 611: Nonlinear Programming, Winter 2006 4. Convex optimization problems Page 4–28

(15)

Minimizing spectral radius of nonnegative matrix

Perron-Frobenius eigenvalue λpf(A)

! Consider (elementwise) positive A∈ Rn×n that is irreducible:

(I + A)n−1 > 0.

! P-F Theorem: there is a real, positive eigenvalue of A, λpf, equal to spectral radius maxi i(A)|

! determines asymptotic growth (decay) rate of Ak: Ak ∼ λkpf as k → ∞

! alternative characterization:

λpf(A) = inf{λ | Av - λv for some v 2 0}

minimizing spectral radius of matrix of posynomials

! minimize λpf(A(x)), where the elements A(x)ij are posynomials of x

! equivalent geometric program:

minimize λ subject to "n

j=1A(x)ijvj/(λvi)≤ 1, i = 1, . . . , n variables λ, v , x

IOE 611: Nonlinear Programming, Winter 2006 4. Convex optimization problems Page 4–29

Quasiconvex optimization

minimize f0(x)

subject to fi(x)≤ 0, i = 1, . . . , m Ax = b

with f0 : Rn → R quasiconvex, f1, . . . , fm convex

can have locally optimal points that are not (globally) optimal Quasiconvex optimization

minimize f0(x)

subject to fi(x) ≤ 0, i = 1, . . . , m Ax = b

with f0 : Rn → R quasiconvex, f1, . . . , fm convex

can have locally optimal points that are not (globally) optimal

PSfrag replacements

(x, f0(x))

Convex optimization problems 4–14

IOE 611: Nonlinear Programming, Winter 2006 4. Convex optimization problems Page 4–30

(16)

convex representation of sublevel sets of f

0

if f0 is quasiconvex, there exists a family of functions φt such that:

! φt(x) is convex in x for fixed t

! t-sublevel set of f0 is 0-sublevel set of φt, i.e., f0(x)≤ t ⇐⇒ φt(x)≤ 0 example

f0(x) = p(x) q(x)

with p convex, q concave, and p(x) ≥ 0, q(x) > 0 on dom f0

can take φt(x) = p(x)− tq(x):

! for t ≥ 0, φt convex in x

! p(x)/q(x)≤ t if and only if φt(x) ≤ 0

IOE 611: Nonlinear Programming, Winter 2006 4. Convex optimization problems Page 4–31

quasiconvex optimization via convex feasibility problems

φt(x)≤ 0, fi(x) ≤ 0, i = 1, . . . , m, Ax = b (1)

! for fixed t, a convex feasibility problem in x

! if feasible, we can conclude that t ≥ p!; if infeasible, t ≤ p!

Bisection method for quasiconvex optimization given l ≤ p!, u ≥ p!, tolerance * > 0.

repeat

1. t := (l + u)/2.

2. Solve the convex feasibility problem (1).

3. if (1) is feasible, u := t; else l := t.

until u − l ≤ *.

requires exactly 3log2((u− l)/*)4 iterations (where u, l are initial values)

IOE 611: Nonlinear Programming, Winter 2006 4. Convex optimization problems Page 4–32

(17)

(Generalized) linear-fractional program

minimize f0(x) subject to Gx - h

Ax = b linear-fractional program

f0(x) = cTx + d

eTx + f , dom f0(x) ={x | eTx + f > 0}

! a quasiconvex optimization problem; can be solved by bisection

! also equivalent to the LP (variables y , z) minimize cTy + dz subject to Gy - hz

Ay = bz eTy + fz = 1 z ≥ 0

IOE 611: Nonlinear Programming, Winter 2006 4. Convex optimization problems Page 4–33

Generalized linear-fractional program

f0(x) = max

i=1,...,r

ciTx + di

eiTx + fi , dom f0(x) ={x | eiTx+fi > 0, i = 1, . . . , r} a quasiconvex optimization problem; can be solved by bisection

example: Von Neumann model of a growing economy maximize (over x, x+) mini=1,...,nxi+/xi

subject to x+ + 0, Bx+ - Ax

! x, x+ ∈ Rn: activity levels of n sectors, in current and next period

! (Ax)i, (Bx+)i: produced, resp. consumed, amounts of good i

! xi+/xi: growth rate of sector i

allocate activity to maximize growth rate of slowest growing sector

IOE 611: Nonlinear Programming, Winter 2006 4. Convex optimization problems Page 4–34

(18)

Generalized inequality constraints

convex problem with generalized inequality constraints minimize f0(x)

subject to fi(x) -Ki 0, i = 1, . . . , m Ax = b

! f0 : Rn → R convex; fi : Rn → Rki Ki-convex w.r.t. proper cone Ki

! same properties as standard convex problem (convex feasible set, local optimum is global, etc.)

conic form problem: special case with affine objective and constraints

minimize cTx

subject to Fx + g -K 0 Ax = b

extends linear programming (K = Rm+) to nonpolyhedral cones

IOE 611: Nonlinear Programming, Winter 2006 4. Convex optimization problems Page 4–35

Semidefinite program (SDP)

minimize cTx

subject to x1F1 + x2F2+· · · + xnFn + G - 0 Ax = b

with Fi, G ∈ Sk

! inequality constraint is called linear matrix inequality (LMI)

! includes problems with multiple LMI constraints: for example, x1Fˆ1+· · · + xnFˆn + ˆG - 0, x1F˜1+· · · + xnF˜n + ˜G - 0 is equivalent to single LMI

x1

* ˆF1 0 0 F˜1

+ +x2

* ˆF2 0 0 F˜2

+

+· · ·+xn* ˆFn 0 0 F˜n

+

+* ˆG 0 0 G˜

+ - 0

IOE 611: Nonlinear Programming, Winter 2006 4. Convex optimization problems Page 4–36

(19)

LP and SOCP as SDP

LP and equivalent SDP LP: minimize cTx

subject to Ax - b

SDP: minimize cTx

subject to diag(Ax − b) - 0 (note different interpretation of generalized inequality -)

SOCP and equivalent SDP SOCP: minimize fTx

subject to &Aix + bi&2 ≤ ciTx + di, i = 1, . . . , m SDP: minimize fTx

subject to

* (ciTx + di)I Aix + bi

(Aix + bi)T ciTx + di

+

+ 0, i = 1, . . . , m

IOE 611: Nonlinear Programming, Winter 2006 4. Convex optimization problems Page 4–37

Eigenvalue minimization

minimize λmax(A(x))

where A(x) = A0+ x1A1+· · · + xnAn (with given Ai ∈ Sk)

equivalent SDP

minimize t

subject to A(x) - tI

! variables x ∈ Rn, t ∈ R

! follows from

λmax(A) ≤ t ⇐⇒ A- tI

IOE 611: Nonlinear Programming, Winter 2006 4. Convex optimization problems Page 4–38

(20)

Matrix norm minimization

minimize &A(x)&2 = ,

λmax(A(x)TA(x))-1/2

where A(x) = A0+ x1A1+· · · + xnAn (with given Ai ∈ Sp×q) equivalent SDP

minimize t subject to

* tI A(x)

A(x)T tI +

+ 0

! variables x ∈ Rn, t ∈ R

! constraint follows from

&A&2 ≤ t ⇐⇒ ATA - t2I , t ≥ 0

⇐⇒

* tI A AT tI

+ + 0

IOE 611: Nonlinear Programming, Winter 2006 4. Convex optimization problems Page 4–39

Convexity of vector-valued functions

f : Rn → Rm is K -convex if dom f is convex and f (θx + (1− θ)y) -K θf (x) + (1− θ)f (y) for x, y ∈ dom f , 0 ≤ θ ≤ 1

example f : Sm → Sm, f (X ) = X2 is Sm+-convex

proof: for fixed z ∈ Rm, zTX2z = &Xz&22 is convex in X , i.e., zT(θX + (1− θ)Y )2z ≤ θzTX2z + (1− θ)zTY2z for X , Y ∈ Sm, 0 ≤ θ ≤ 1

therefore (θX + (1− θ)Y )2 - θX2+ (1− θ)Y2

IOE 611: Nonlinear Programming, Winter 2006 4. Convex optimization problems Page 4–40

(21)

Vector optimization

general vector optimization problem minimize (w.r.t. K ) f0(x)

subject to fi(x)≤ 0, i = 1, . . . , m hi(x) ≤ 0, i = 1, . . . , p vector objective f0 : Rn → Rq, minimized w.r.t. proper cone K ∈ Rq

convex vector optimization problem minimize (w.r.t. K ) f0(x)

subject to fi(x)≤ 0, i = 1, . . . , m Ax = b

with f0 K -convex, f1, . . . , fm convex

IOE 611: Nonlinear Programming, Winter 2006 4. Convex optimization problems Page 4–41

Optimal and Pareto optimal points

set of achievable objective values

O = {f0(x)| x feasible}

! feasible x is optimal if f0(x) is a minimum value of O

! feasible x is Pareto optimal if f0(x) is a minimal value of O

Optimal and Pareto optimal points

set of achievable objective values

O = {f0(x) | x feasible}

• feasible x is optimal if f0(x) is a minimum value of O

• feasible x is Pareto optimal if f0(x) is a minimal value of O

PSfrag replacements

O

f0(x!)

x! is optimal

PSfrag replacements

O f0(xpo)

xpo is Pareto optimal

Convex optimization problems 4–41

Optimal and Pareto optimal points

set of achievable objective values

O = {f0(x)| x feasible}

• feasible x is optimal if f0(x) is a minimum value of O

• feasible x is Pareto optimal if f0(x) is a minimal value of O

PSfrag replacements

O

f0(x!)

x! is optimal

PSfrag replacements

O f0(xpo)

xpo is Pareto optimal

Convex optimization problems 4–41

IOE 611: Nonlinear Programming, Winter 2006 4. Convex optimization problems Page 4–42

(22)

Multicriterion optimization

vector optimization problem with K = Rq+ f0(x) = (F1(x), . . . , Fq(x))

! q different objectives Fi; roughly speaking we want all Fi’s to be small

! feasible x! is optimal if

y feasible = f0(x!) - f0(y ) if there exists an optimal point, the objectives are noncompeting

! feasible xpo is Pareto optimal if

y feasible, f0(y ) - f0(xpo) = f0(xpo) = f0(y ) if there are multiple Pareto optimal values, there is a trade-off between the objectives

IOE 611: Nonlinear Programming, Winter 2006 4. Convex optimization problems Page 4–43

Regularized least-squares

multicriterion problem with two objectives

F1(x) = &Ax − b&22, F2(x) =&x&22

! example with A∈ R100×10

! shaded region is O

! heavy line is formed by Pareto optimal points

Regularized least-squares

multicriterion problem with two objectives

F1(x) =!Ax − b!22, F2(x) =!x!22

• example with A ∈ R100×10

• shaded region is O

• heavy line is formed by Pareto optimal points

PSfrag replacements

F1(x) = !Ax − b!22 F2(x)=!x!2 2

0 5 10 15

0 5 10 15

Convex optimization problems 4–43

IOE 611: Nonlinear Programming, Winter 2006 4. Convex optimization problems Page 4–44

(23)

Risk return trade-off in portfolio optimization

minimize (w.r.t. R2+) (−¯pTx, xTΣx) subject to 1Tx = 1, x + 0

! x ∈ Rn is investment portfolio; xi is fraction invested in asset i

! p ∈ Rn is vector of relative asset price changes; modeled as a random variable with mean ¯p, covariance Σ

! ¯pTx = E r is expected return; xTΣx = var r is return variance example

Risk return trade-off in portfolio optimization

minimize (w.r.t. R2+) (−¯pTx, xTΣx) subject to 1Tx = 1, x" 0

• x ∈ Rn is investment portfolio; xiis fraction invested in asset i

• p ∈ Rn is vector of relative asset price changes; modeled as a random variable with mean ¯p, covariance Σ

• ¯pTx = E r is expected return; xTΣx = var r is return variance example

PSfrag replacements

meanreturn

standard deviation of return

0% 10% 20%

0%

5%

10%

15%

PSfrag replacements

standard deviation of return

allocationx

x(1) x(2) x(3)

x(4)

0% 10% 20%

0 0.5 1

Convex optimization problems 4–44

IOE 611: Nonlinear Programming, Winter 2006 4. Convex optimization problems Page 4–45

Scalarization

to find Pareto optimal points: choose λ 2K 0 and solve scalar problem

minimize λTf0(x)

subject to fi(x)≤ 0, i = 1, . . . , m hi(x) = 0, i = 1, . . . , p

if x is optimal for scalar problem, then it is Pareto-optimal for vector optimization problem

Scalarization

to find Pareto optimal points: choose λ!K0 and solve scalar problem minimize λTf0(x)

subject to fi(x)≤ 0, i = 1, . . . , m hi(x) = 0, i = 1, . . . , p

if x is optimal for scalar problem, then it is Pareto-optimal for vector optimization problem

PSfrag replacements O

f0(x1)

λ1

f0(x2) λ2 f0(x3)

for convex vector optimization problems, can find (almost) all Pareto optimal points by varying λ!K0

Convex optimization problems 4–45

for convex vector optimization problems, can find (almost) all Pareto optimal points by varying λ 2K 0

IOE 611: Nonlinear Programming, Winter 2006 4. Convex optimization problems Page 4–46

(24)

examples

! for multicriterion problem, find Pareto optimal points by minimizing positive weighted sum

λTf0(x) = λ1F1(x) +· · · + λqFq(x)

! regularized least-squares of page 111 (with λ = (1, γ)) minimize &Ax − b&22 + γ&x&22

for fixed γ > 0, a least-squares problem

! risk-return trade-off of page 112 (with λ = (1, γ)) minimize −¯pTx + γxTΣx subject to 1Tx = 1, x + 0 for fixed γ > 0, a QP

IOE 611: Nonlinear Programming, Winter 2006 4. Convex optimization problems Page 4–47

參考文獻

相關文件

Wang, Solving pseudomonotone variational inequalities and pseudo- convex optimization problems using the projection neural network, IEEE Transactions on Neural Network,

Numerical experiments are done for a class of quasi-convex optimization problems where the function f (x) is a composition of a quadratic convex function from IR n to IR and

For different types of optimization problems, there arise various complementarity problems, for example, linear complemen- tarity problem, nonlinear complementarity problem

For different types of optimization problems, there arise various complementarity problems, for example, linear complementarity problem, nonlinear complementarity problem,

In section 4, based on the cases of circular cone eigenvalue optimization problems, we study the corresponding properties of the solutions for p-order cone eigenvalue

It is well-known that, to deal with symmetric cone optimization problems, such as second-order cone optimization problems and positive semi-definite optimization prob- lems, this

This kind of algorithm has also been a powerful tool for solving many other optimization problems, including symmetric cone complementarity problems [15, 16, 20–22], symmetric

In this paper, motivated by Chares’s thesis (Cones and interior-point algorithms for structured convex optimization involving powers and exponentials, 2009), we consider