• 沒有找到結果。

半正定規劃

N/A
N/A
Protected

Academic year: 2021

Share "半正定規劃"

Copied!
30
0
0

加載中.... (立即查看全文)

全文

(1)

國立交通大學

應用數學系

碩士論文

半正定規劃

Semidefinite Programming Problems

研究生:葉彬

指導教授:翁志文 教授

(2)

半正定規劃

Semidefinite Programming Problems

研究生:葉彬

指導教授:翁志文 教授

Student: Bin Yeh

Advisor: Chih-Wen Weng

國立交通大學 應用數學系

碩士論文

A Thesis

Submitted to Department of Applied Mathematics College of Science

National Chiao Tung University in Partial Fulfillment of the Requirements

for the Degree of Master

in

Applied Mathematics

June 2010

Hsinchu, Taiwan, Republic of China

(3)

半正定規劃

研究生:葉彬

指導教授:翁志文

國立交通大學應用數學系碩士班

摘要

在半正定規劃的問題中,我們要求一些對稱矩陣的彷射組合必須是半正 定,在這樣的限制下試圖將目標線性函數最小化。這些限制未必是線性, 但它們具有中凸的性質故半正定規劃是一種中凸規劃。在這篇論文中我們 探討了一些半正定規劃的基本性質與基礎理論並給出證明。

(4)

Semidefinite Programming Problems

Student: Bin Yeh

Advisor: Chih-Wen Weng

Department of Applied Mathematics

National Chiao Tung University

Abstract

In semidefinite programming problems one minimizes a linear func-tion subject to some constraints which requires an affine combinafunc-tion of symmetric matrices to be positive semidefinite. The constraints may not be linear but it is convex so semidefinite programming prob-lems are convex optimization probprob-lems. In this paper we give some basic properties and fundamental theorems with their proofs regrading semidefinite programming problems.

(5)

誌謝

這一路上最感謝我的指導教授翁志文老師,因為我的不成熟而給他添了 很多麻煩,實在是非常的對不起老師,在這邊要跟他說一聲對不起與謝謝 您。 兩年多來還有其他很多很照顧我的教授們,包括傅恆霖老師、白啟光老 師、黃大原老師、陳秋媛老師、王夏聲老師,真的非常感謝他們的指導與 照顧。 我還要感謝我同師門的學長鎬文和喻培,當然還有其他給我許多幫助的學 長姐,像是芳婷學姐、敏筠學姐、惠蘭學姐、國元學長、智懷學長,他們 給了我很多的幫助。 也非常謝謝眾多好同學們,永潔、光祥、文昱、易萱、宜君、舜婷、玉雯、 慧棻,以及研究室的夥伴們:育生、建文、思綸、健峰、思賢,還有我在 求學過程中認識的許多學弟妹:侖欣、彥伶、權益、湧昇、昇華、育慈、 定國、冠緯、閔豪、瑩晏、瑜堯、紹鵬、昌翰,謝謝你們陪我渡過快樂的 時光。 最後我要感謝我的家人,爸爸,媽媽與弟弟。無論是什麼情況,他們總是 支持著我、鼓勵著我,萬分的感謝你們。 要感謝的人很多,很幸運我能得到這麼多人的幫助。感恩之心難以言喻, 就僅在此聊表感激之情,沒有各位就沒有今日的我。

(6)

Contents

1 Introduction 1 2 Semidefinite programing problems 2 3 Dual problem 5 4 A few lemmas 8 5 Property of Semidefinite programming Problems 21 6 References 24

(7)

1 Introduction

This paper is based on the survey paper by Vandenberghe and Boyd [1] and some details, mostly proofs, which are omitted in that survey. In this paper we focus on fundamental theory about semidefinite programming problem, rather than its application.

Semidefinite programming (SDP) is an optimization problem concerned with the optimization of a linear objective function over the intersection of the cone of positive semidefinite matrices. Linear programming, a well-used mathematical model which is used to obtain the best outcome with some certain restriction in the form of linear equations, is actually a special case of semidefinite programming problem. There are many different types of algorithms to solve semidefinite programming problem and these algorithms are capable of getting the result in polynomial time.

There are a lot of applications of semidefinite programming. In opera-tions research and combinatorial optimization many problems are modeled as semidefinite programming problem so they can be well approximated. For example, Goemans and Williamson found an algorithm to obtain maximum cut using semidefinite programming [3].

In section 2 we give some basic definition about semidefinite programming problem and an example for better understanding. In section 3 the maybe most important property of semidefinite programming problem, duality, is discussed. In the further sections some more lemmas and propositions are

(8)

given and eventually in section 5 a critical theorem is proved.

Some proofs in this paper are based on the lecture note written by L´aszl´o Lov´asz [2] and the book by Abraham Berman and Naomi Shaked-Monderer [4] and the note by Konstantin Aslanidi [5]. The main idea of proof of lemma 4.13 is provided by Renato Paes Leme.

2

Semidefinite programing problems

Throughout, we useRn to denote the set of column vectors of size n and

Rn×nto denote the set of n×n symmetric matrices over R. For a real matrix

M , we use M ≥ 0 when M has nonnegative entries, and use M ≥p 0 when

M is symmetric and positive semidefinite, i.e. zTM z ≥ 0 for all z ∈ Rn.

Similarly, we will denote positive definite by M >p 0. The following is the

general form of a semidefinite programming problem.

Problem 2.1. (SDPn(c, F0, F1, . . . , Fm))

Given a column vector c ∈ Rm and m + 1 n × n symmetric matrices

F0, F1, . . . , Fm ∈ Rn×n. Find min x c Tx for x = [x1, x2, . . . , xm]T ∈ Rm subject to F0 + mi=1 xiFi ≥p 0.

Give a SDPn(c, F0, F1, . . . , Fm) and let F (x) = F0+

m

i=1xiFi. A vector

x∈ Rmis said to be feasible if F (x)

(9)

called the feasible region of the problem. A vector xoptin the feasible region

is called an optimal point if cTx

opt reaches the minimum of cTx among all

feasible points x. In this case the value cTx

opt is called the minimum of the

problem. The number

inf

x c T

x∈ R ∪ {−∞},

where the infimum is taking for feasible x, is called the infimum of the prob-lem. If the infimum is not equal to the minimum, then SDPn(c, F0, F1, . . . , Fm)

is said to be infeasible. The following is a simple example of semidefinite programming problem. Example 2.2. Consider SDP2(   0 1   ,   0 1 1 0   ,   1 0 0 0   ,   0 0 0 1  ). The feasible region is

     x1 x2   |   0 1 1 0   + x1   1 0 0 0   + x2   0 0 0 1   ≥p 0    =      x1 x2   |   x1 1 1 x2   ≥p 0   . Note that   x1 1 1 x2 

 is semidefinite if and only if both x1 and x1x2− 1

are nonnegative, i.e. x1 ≥ 0 and x1x2− 1 ≥ 0. Thus the feasible region is

     x1 x2   |x1 ≥ 0, x1x2 ≥ 1   . (1)

(10)

The minimum of (0, 1)(x1, x2)T = x2 does not exist in the feasible region since

x2 tends to zero as x1 tends to infinity. However the infimum of this problem,

obtained by above discussion, is zero. Thus, this example is infeasible. The following is the general form of a linear programming problem, which we will show to be a special case of semidefinite programming problem.

Problem 2.3. (LPPm(c, b, A))

Given b, c∈ Rm and a symmetric matrix A∈ Rm×m. Find

min

x c T

x

subject to all x ∈ Rm satisfying

Ax + b≥ 0.

Note that a linear programming problem is a special case of semidefinite programming problem. In fact by setting F0 = diag(b), Fi = diag(Ai) for

i = 1, . . . , m, where Ai is the ith column of A, the LPPm(c, b, A) becomes

the SDPn(c, F0, F1, . . . , Fm).

Similar to the semidefinite programming problem, in LPPm(c, b, A), a

vector x ∈ Rm is said to be feasible if Ax + b ≥ 0 and the set {x ∈

Rm | Ax + b ≥ 0} is called the feasible region. A vector x

opt in the feasible

region is called an optimal point if cTx

opt reaches the minimum of cTx for

all feasible points x.

We now show Example 2.2 is no way to be interpreted as a linear program-ming problem. To the contrary assume Example 2.2 is also a LPm(c, b, A).

(11)

Then from the definition m = 2 and A is a 2 × 2 matrix. Let V denote the nullspace of A. Then referring to (1) as the feasible region Ω := {x ∈ R2 | Ax + b ≥ 0} also as a feasible region of LP

m(c, b, A) we must have

Ω + V ⊆ Ω. This only happens when V = 0, i.e. A is invertible. But then the feasible region of LPm(c, b, A) is {x ∈ R2 | x ≥ −A−1b}, which is clearly

not to be Ω, a contradiction.

3

Dual problem

To solve a classical optimization problem, the problem and its dual

prob-lem play an important role. The following probprob-lem is the dual of SDPn(c, F0, F1, . . . , Fm).

Problem 3.1. (SDPn(c, F0, F1, . . . , Fm))

Given c∈ Rmand m+1 symmetric matrices F

0, F1, . . . , Fm ∈ Rn×n. Find

max

Z −tr(F0Z)

subject to all symmetric n× n matrices Z with

Z ≥p 0,

tr(FiZ) = ci for 1≤ i ≤ m, (2)

where tr(M ) is the trace of M .

Give a SDPn(c, F0, F1, . . . , Fm), a symmetric matrix Z is said to be

fea-sible if Z ≥p 0 and tr(FiZ) = ci for 1 ≤ i ≤ m. The set {Z ∈ Rn×n|Z ≥p

(12)

problem. A symmetric matrix Zopt in the feasible region is called an

opti-mal point if−tr(F0Z) reaches the maximum of−tr(F0Z) among all feasible

points Z. In this case the value cTx

opt is called the maximum of the

prob-lem. The number

sup

Z −tr(F

0Z)∈ R ∪ {∞},

where the supremum is taking for feasible Z, is called the supremum of the

problem. If the supremum is not equal to the maximum, then SDPn(c, F0, F1, . . . , Fm)

is said to be infeasible. The original semidefinite programming problem will be referred as the primal problem.

In Section 5, we will show that the supremum of SDPn(c, F0, F1, . . . , Fm)

is no larger than the infimum of SDPn(c, F0, F1, . . . , Fm). This explains their

dual relation. Here we give an example of dual semidefinite programming problem.

Example 3.2. The dual problem of

SDP2(   0 1   ,   0 1 1 0   ,   1 0 0 0   ,   0 0 0 1  ) in Example 2.2 is SDP2(   0 1   ,   0 1 1 0   ,   1 0 0 0   ,   0 0 0 1  ). That is to maximize −tr(   0 1 1 0   Z)

(13)

for all 2× 2 symmetric matrices Z subject to tr(   1 0 0 0   Z) = 0, tr(   0 0 0 1   Z) = 1, and Z ≥p 0. Let Z =   Z11 Z12 Z21 Z22 

. The first condition says

0 = tr(   1 0 0 0     Z11 Z12 Z21 Z22  ) = tr(   Z11 Z12 0 0  ) = Z11,

and the second condition implies

1 = tr(   0 0 0 1     Z11 Z12 Z21 Z22  ) = T r(   0 0 Z21 Z22  ) = Z22.

Thus we only need to consider matrices of the form 

 0 Z12

Z21 1

 . The third condition Z p 0 is equivalent to

( x1 x2 ) Z   x1 x2   ≥ 0 for all x1, x2 ∈ R. As 0 ( x1 x2 ) Z   x1 x2   =( x2Z21 x1Z12+ x2 )  x1 x2   = x1x2(Z21+ Z12) + x22 = 2x1x2Z12+ x22,

Z ≥p 0 is equivalent to Z12= 0. Then our feasible region is

     0 Z12 Z21 1   |Z21 = Z12 = 0   .

(14)

The goal is to maximize −tr(   0 1 1 0     Z11 Z12 Z21 Z22  ) = −tr(   Z21 Z22 Z11 Z12  ) = −(Z21+ Z12),

which is always zero in the feasible region. Then the maximum of

SDP2(   0 1   ,   0 1 1 0   ,   1 0 0 0   ,   0 0 0 1  )

is 0 which, as we showed before, is also the infimum but not the minimum of

SDP2(   0 1   ,   0 1 1 0   ,   1 0 0 0   ,   0 0 0 1  ).

The general theory of relation between a primal problem and its dual problem will be given the section 5.

4

A few lemmas

Before proceeding, we introduce some basic properties of positive semidef-inite matrices. Positive semidefinite matrices have some great properties in the way of geometry space. To see this we need some definition first. A

convex cone C is a set of vectors in a vector space such that (i) for any

vector v in C, rv is also in C for any r > 0, and (ii) for any two vectors v,

v′ ∈ C, v + v′ ∈ C.

For any convex cone C ⊆ Rn×n, the polar cone C is defined by

(15)

where A· B is the inner product for matrices, defined as

A· B = tr(AtB) =

1≤i,j≤n

Ai,jBi,j.

Note that a polar cone is also a convex cone, which is trivial to prove by applying the definition of convex cone.

Lemma 4.1. A positive multiple of a positive semidefinite symmetric matrix

is still positive semidefinite, and sum of two positive semidefinite symmetric matrices is still positive semidefinite.

Proof. Let symmetric matrices A, B ∈ Rn×n be positive semidefinite and

r > 0. Then for any vector x∈ Rn, we have xT(rA)x = r(xTAx)≥ 0

and

xT(A + B)x = (xTAx) + (xTBx)≥ 0,

which finished the proof.

The above lemma shows that the set of all symmetric positive semidefinite matrices Pn is actually a convex cone inRn×n. For next lemma, we will use

Hadamard product of matrices; that is, for two matrices A, B ∈ Rn×n, the

Hadamard product A◦ B is defined as

(A◦ B)ij = AijBij

for all 1≤ i, j ≤ n.

A property of Hadamard product will be given later. To prove the prop-erty we first show the following proposition.

(16)

Proposition 4.2. Let A be a symmetric n×n real matrix. Then A is positive

semidefinite if and only if there exists an n×n matrix B such that A = BBT.

Proof. (⇐)

For all x∈ Rn,

xTAx = xT(BBT)x = (BTx)T(BTx) = (BTx)· (BTx)≥ 0.

(⇒)

Since A is symmetric, it follows that A is normal and therefore A is unitary similar to a diagonal matrix, i.e. P APT = D for some n× n matrix P with

PTP = I and some diagonal matrix D. Then A = PTDP . Note since A

and D are similar, they have same eigenvalues λ1, λ2, . . . , λn and they are

all nonnegative because A is positive semidefinite. Then we may assume

D = diag(λ1, λ2, . . . , λn). Now let

D′ =         λ1 0 λ2 . .. 0 √λn         , thus A = PTDP = PTD′D′P = (PTD′)(PTD′)T.

Corollary 4.3. Let A be symmetric n× n real matrix. Then A is positive

semidefinite if and only if there exist vectors v1, v2, ..., vn ∈ Rn such that

(17)

Proof. (⇒)

By last lemma A = BBT for some n× n matrix B. Let v

i be the ith column

vector of B for 1 ≤ i ≤ n. Then (A)ij = (BBT)ij = nh=1 (vh)i(vh)j = nh=1 (vhvhT)ij = ( nh=1 vhvhT)ij. (⇐)

Let A =ni=1vivTi . Let B =

( v1 v2 . . . vn ) . Then (A)ij = ( nh=1 vhvhT)ij = nh=1 (vhvhT)ij = nh=1 (vh)i(vh)j = (BBT)ij.

Then by last proposition, A is positive semidefinite.

Corollary 4.4. The Hadamard product of two positive semidefinite matrices

is still positive semidefinite.

Proof. Let A =ki=1vivTi and B =

l j=1wjw T j, then A◦ B = k,li=1,j=1 (vi◦ wj)(vi◦ wj)T.

Therefore by last corollary A◦ B is positive semidefinite. Now the lemma mentioned earlier can be proved.

Lemma 4.5. The polar cone of Pn is itself, i.e. Pn∗ = Pn.

Proof. Let symmetric matrices A, B ∈ Rn×n be positive semidefinite. Note

tr(AB) = ni=1 (AB)ii= ni=1 ∑ 1≤j≤n AijBji = ni=1 ∑ 1≤j≤n AjiBji = eT(A◦ B)e

(18)

where e ∈ Rn is the all-1 vector. Then by corollary 4.4, A◦ B is still positive

semidefinite and then eT(A◦ B)e ≥ 0. Therefore, P n ⊇ Pn.

On the other hand, suppose A∈ Pn∗. For any column vector x∈ Rn, the

n× n symmetric matrix xxT is positive semidefinite since

yT(xxT)y = (xTy)T(xTy) = ( ni=1 xiyi)( ni=1 xiyi) = ( ni=1 xiyi)2 ≥ 0

for any y∈ Rn. Thus

xAxT = tr(x(AxT)) = tr(AxTx)≥ 0

for any x∈ Rn. Then A∈ P n.

So the polar cone of Pn is itself, which leads to the next proposition.

Proposition 4.6. A symmetric matrix A is positive semidefinite if and only

if tr(AB)≥ 0 for all symmetric positive semidefinite matrix B.

We immediately have the following corollary.

Corollary 4.7. Let A,B ∈ Rn×n be symmetric matrices. If A,B

p 0, then

tr(AB)≥ 0.

Proposition 4.8. Let A ∈ Pn and B ∈ Rn×n be a positive definite matrix.

Then tr(AB) = 0 if and only if A = 0.

Proof. It’s clearly tr(0B) = 0. On the other hand, since A ≥p 0 by taking

ei, i.e. the vector with 1 at the ith entry and zero for the rest, we have

(19)

Furthermore there exists a n× n matrix P such that PtP = P Pt = I and

A′ = PtAP = diag(λ1, ..., λn) is diagonal. Note A′ is still positive

semidefi-nite. Then

0 = tr(AB) = tr(PtP AB) = tr(PtABP ) = tr(PtAP PtBP )

= tr(A′PtBP ) = ∑ 1≤i≤n λi(PtBP )ii = ∑ 1≤i≤n λi(etiP tBP e i) = ∑ 1≤i≤n λi((P ei)tB(P ei))≥ 0

Since (P ei)tB(P ei) > 0 for all 1 ≤ i ≤ n, it follows that λi = 0 for all

1≤ i ≤ n, A′ = 0. Therefore A = 0.

Now we need something else to prove next lemma. A set C ⊆ Rn is

convex if

(1− t)x + ty ∈ C for all x, y ∈ C, t ∈ [0, 1]. Let |x0| =

x· x for all x ∈ Rn, where x · x is the inner product of x and

x. For any C ⊆ Rn, let cl(C) ={x|x ∈ Rn and for any r > 0, there is a y

C such that |x − y| < r}. A subset M of Rn is called an affine set if M =

{x|Bx = b} for some B ̸= 0, B ∈ Rn×n and b∈ Rn

A hyperplane inRn is a subset which can be written as {x|x ∈ Rn, x· b =

β} for some β ∈ R and b ̸= 0, b ∈ Rn. A hyperplane separates Rn into two

parts {x|x ∈ Rn, x· b ≥ β} and {x|x ∈ Rn, x· b < β}. The sets C

1 and C2 are

separated if there exist a hyperplane H such that C1 and C2 are in different

(20)

Proposition 4.9. Let C ⊂ Rn be a convex set and x

0 ∈ Rn with x0 ∈ cl(C)./

Then there is a hyperplane H separates C and x0.

Proof. Since Rn is compact, we have

|y∗− x

0| = inf

y∈C|y − x0|

for some y∗ ∈ cl(C). Note |y∗− x0| ̸= 0 since x0 ∈ cl(C). Let/

h(x) = 1 |y∗− x0|2(2(x· (y − x 0)) +|x0|2− |y∗|2) Then h(x0) = 1 |y∗ − x0|2(2(x0· y )− 2(x 0· x0) +|x0|2− |y∗|2) = 1 |y∗− x0|2(2(x0· y )− (x 0· x0)− |y∗|2) = 1 |y∗− x0|2(2(x0· y )− (x 0· x0)− (y∗· y∗)) = −1 |y∗− x0|2(x0− y )· (x 0− y∗) = −1 And h(y∗) = 1 |y∗− x0|2(2(y · (y− x 0)) +|x0|2− |y∗|2) = 1 |y∗− x0|2(2(y · y)− 2(y· x 0) +|x0|2 − |y∗|2) = 1 |y∗− x0|2((y · y)− 2(y· x 0) + (x0· x0)) = 1 |y∗− x0|2(y − x 0)· (y∗− x0) = 1

Thus the hyperplane {x|h(x) = 0} separates C and x0 since C is convex.

Proposition 4.10. Let C ⊆ Rnbe a convex set. Let M ⊆ Rn be a nonempty

affine set with C ∩ M = ∅. Then there is an hyperplane H containing M such that C is contained in one of the two parts that H separates Rn into.

(21)

Proof. By induction on dim(M ). First let S be a subspace with M = S + a

for some a ∈ Rn. If dim(S) = n− 1 then M itself is such a hyperplane, so

we are done.

Suppose it holds for all S with dimension larger than k < n− 1 for some

k. When dim(S) = k− 1, we have dim(S⊥) ≥ 2 and therefore contains a subspace T of dimension 2. The set C− M = {x − y|x ∈ C, y ∈ M} does not contain 0, so we can find a subset L⊂ T of dimension 1 such that 0 ∈ L and

L∩(C −M) = ∅. We now can add the basis of L into the basis of S obtaining

a new subspace S′. Then by induction hypothesis, there is a hyperplane H containing S′ and C is contained in one of two parts that H separates Rn

into. Thus we are done.

Proposition 4.11. Let C1, C2 ⊆ Rnbe nonempty convex sets with C1∩C2 =

∅. Then they are separated.

Proof. Taking C = C1− C2 = {x − y|x ∈ C1, y ∈ C2} and M = {0} in last

proposition. Thus there is a hyperplane H contains 0 and H ∩ C = ∅. Let

H = {x|x ∈ Rn, x· b = 0} for some b ∈ Rn. By using −b to replace b if necessary, we have inf x∈Cx· b ≥ 0, sup x∈C x· b > 0. Then 0≤ inf x∈Cx· b = infx1∈C1 x1· b − sup x2∈C1 x2· b, that is inf x1∈C1 x1· b ≥ sup x2∈C1 x2· b.

(22)

Then C1 and C2 are separated by hyperplane

H′ ={x|x · b = inf

x1∈C1

x1· b}.

The next lemma is well-known as the semidefinite version of Farkas’ Lemma, which is a similar theorem regarding linear programming.

Lemma 4.12 (Homogenous Version). Let A1, A2, . . ., Am be symmetric

matrices in Rn×n. Then the system

x1A1+ . . . + xmAm >p 0

has no solution in x1, x2, . . ., xm if and only if there exists a symmetric

matrix Y ̸= 0 such that Ai· Y = tr(AiY ) = 0 for all 1≤ i ≤ m and Y ≥p 0.

Proof. (⇒)

Since the system has no solution, we have that

{

i

xiAi|xi ∈ R} ∩ int(Pn) =

where int(Pn) is the interior of Pn. Moreover, by directly checking the

def-inition, {ixiAi|xi ∈ R} is a convex cone and therefore a convex set. We

can consider Rn×n asRn2

, thus by last lemma, there is a hyperplane

H ={x|x ∈ Rn×n, x· Y = β}

separates {ixiAi|xi ∈ R} and Pnfor some Y ∈ Rn×n and β∈ R, which we

may assume

β = inf

P∈Pn

(23)

by the definition of hyperplane. Furthermore, we may assume inf

P∈Pn

P · Y ≥ 0

by replacing Y with −Y if necessary. But since Pn is a convex cone, for any

x∈ Pn, rx∈ Pn for all r > 0. By taking r small enough we have

inf

P∈Pn

P · Y = 0

Then our hyperplane

H ={x|x · Y = 0}

implies that (∑ixiAi)· Y ≤ 0 and P · Y > 0 for all P ∈ int(Pn). Now for a

fixed 1≤ i ≤ m, let xi = 1 and xj = 0 for all i̸= j. Then

0≥ (

i

xiAi)· Y = Ai· Y.

We can also let xi =−1 and xj = 0 for all i ̸= j. Then

0≥ (

i

xiAi)· Y = −Ai· Y.

These two inequalities show that Ai· Y = 0 for all 1 ≤ i ≤ m. On the other

hand, P · Y ≥ 0 for all P ∈ Pn. Then by proposition 4.6, Y ≥p 0.

(⇐)

Suppose there is a solution of x1A1+ . . . + xmAm >p 0. Then

(∑ i xiAi)· Y =i xi(Ai· Y ) = 0 By proposition 4.8, Y = 0, a contradiction.

(24)

Lemma 4.13 (Nonhomogeneous Version). Let A1, A2, . . ., Am, B be

sym-metric matrices in Rn×n. Then the system

x1A1+ . . . + xmAm− B >p 0

has no solution in x1, x2, . . ., xm if and only if there exists a symmetric

matrix Y ̸= 0 such that tr(AiY ) = 0 for all 1 ≤ i ≤ m, tr(BY ) ≥ 0 and

Y ≥p 0.

Proof. This is done by applying last lemma to following matrices

  A1 0 0 0   ,   A2 0 0 0   , ...,   Ak 0 0 0   , −B 0 0 1   . We claim that if the system

x1   A1 0 0 0  +x2   A2 0 0 0  +...+xk   Ak 0 0 0  +xk+1 −B 0 0 1   >p 0

has a solution, then so does the system

x1A1 + . . . + xmAm− B >p 0.

To prove this claim, suppose v1, ..., vk, vk+1 is a solution of

x1   A1 0 0 0  +x2   A2 0 0 0  +...+xk   Ak 0 0 0  +xk+1 −B 0 0 1   >p 0. Let M = ( ∑ 1≤i≤k vi   Ai 0 0 0  ) + vk+1 −B 0 0 1   . Then M >p 0. Thus 0 < etk+1M ek+1 = vk+1.

(25)

Let ui = vvi

k+1 for all 1≤ i ≤ k + 1, and

M′ = ( ∑ 1≤i≤k ui   Ai 0 0 0  ) +   −B 0 0 1   = 1 vk+1 M

which is still positive definite. Note that all principal submatrices of a positive definite matrix is again positive definite. Therefore the following principal submatrix of M′ (∑ 1≤i≤k uiAi)− B >p 0. Thus u1, u2, ..., uk is a solution of x1A1 + . . . + xmAm− B >p 0. (⇒)

The claim we proved is equivalent to that if

x1A1+ . . . + xmAm− B >p 0

has no solution, then either the system

x1   A1 0 0 0  +x2   A2 0 0 0  +...+xk   Ak 0 0 0  +xk+1 −B 0 0 1   >p 0.

So by previous lemma, there exists Y ∈ R(n+1)×(n+1) such that Y

p 0 and

  Ai 0

0 0 

 · Y = 0 for all 1 ≤ i ≤ k, and −B 0 0 1   · Y = 0. Let Y =   Y′ y yT y 0  

(26)

for some y ∈ Rn, y

0 ∈ R, then

Ai· Y′ = 0 for all 1≤ i ≤ k.

Since Y ≥p 0, we take en+1 and we have

0≤ eTn+1Y en+1 = y0. Together with 0 =   −B 0 0 1   · Y = −BY′+ y 0, we conclude that BY′ = y0 ≥ 0.

Note Y′ ≥p 0 since Y′ is a principal submatrix of Y ≥p 0.

(⇐)

By assumption there exists Y ̸= 0 such that Ai·Y = tr(AiY ) = 0 for all 1

i≤ m, Y ≥p 0 and tr(BY ) ≥ 0. Let

Z =   Y 0 0 B· Y .

Suppose there is a solution of x1A1 + . . . + xmAm− B >p 0. Then so does

the system x1   A1 0 0 0  +x2   A2 0 0 0  +...+xk   Ak 0 0 0  +xk+1 −B 0 0 1   >p 0. But Z ·   A1 0 0 0   =   Y 0 0 B· Y ·   A1 0 0 0   = 0 for all 1 ≤ i ≤ m,

(27)

and  −B 0 0 1   =   Y 0 0 B· Y · −B 0 0 1   = −B · Y + B · Y = 0. Note Z p 0. However, since the system

x1   A1 0 0 0  +x2   A2 0 0 0  +...+xk   Ak 0 0 0  +xk+1 −B 0 0 1   >p 0

has solution, such Z should not exist by last lemma, a contradiction.

5 Property of Semidefinite programming

Prob-lems

A SDPn(c, F0, F1, . . . , Fm) is said to have feasible solution if there exists

x with F (x) ≥p 0, and is said to be strictly feasible if F (x) >p 0 for some

x in the feasible region, where as before F (x) = F0+

m

i=1xiFi. In this case

x is called a strictly feasible solution. Similarly SDP∗n(c, F0, F1, . . . , Fm)

is said to have feasible solution if there is some Z in the feasible region, and Z is said to be strictly feasible if Z >p 0. In this case Z is called a

strictly feasible solution.

Theorem 5.1. Let p and d∗ be the infimum of SDPn(c, F0, F1, . . . , Fm) and

supremum of SDP∗n(c, F0, F1, . . . , Fm) respectively, and assume p∗, d∗ <∞.

Then p∗ ≥ d∗. Moreover, suppose either of the following conditions (i)-(ii) holds.

(28)

(i) The primal problem is strictly feasible.

(ii) The dual problem is strictly feasible.

Then p∗ = d∗.

Proof. Let x be a vector in the feasible region of primal problem and Z be a

symmetric matrix in the feasible region of its dual problem. Then referring to (2) and by corollary 4.7, cTx + tr(ZF0) = mi=1 tr(ZFixi) + tr(ZF0) = tr(ZF (x))≥ 0. Thus cTx≥ −tr(ZF 0) which shows p∗ ≥ d∗.

Now the system

cTx < p∗ F (x)≥p 0

has no solution x ∈ Rm by the definition of p. Therefore if we define the

matrices F0 =   p∗ 0 0 F0   and F′ i =   −ci 0 0 Fi for 1 ≤ i ≤ m,

F0 + x1F1 + . . . + xmFm′ >p 0 has no solution in Rn. Thus by lemma 4.13

there is a positive semidefinite matrix Y ̸= 0 such that

tr(F0′Y )≥ 0 and tr(Fi′Y ) = 0 for 1≤ i ≤ m. By letting Y =   y00 y yT Z

(29)

we obtain

tr(F0Z)≥ y00p∗ and tr(FiZ) = y00ci for 1≤ i ≤ m.

We claim that y00̸= 0. Suppose not. Then

tr(F0Z)≥ 0 and tr(FiZ) = 0 for 1≤ i ≤ m.

Therefore by lemma 4.13, the existence of Z implies F0 + x1F1 + . . . +

xmFm >p 0 has no solution inRn, contradicts to the hypothesis that primal

problem is strictly feasible. Thus, y00 ̸= 0.

Now y00̸= 0 and since Y is positive semidefinite, y00> 0. By scaling we may

assume y00= 1. But then Z satisfies tr(ZF0)≥ y00p∗ = p∗ and d∗ ≥ tr(ZF0)

by definite of d∗, thus d∗ ≥ p. Together with first part of the proof we have

p∗ = d∗.

The case that condition (ii) holds is similar to prove. Therefore the proof is complete.

(30)

6 References

[1] Lieven Vandenberghe and Stephen Boyd, Semidefinite Programming, SIAM Review, Vol. 38, No. 1. (Mar., 1996), pp. 49-95.

[2] L´aszl´o Lov´asz, Semidefinite programs and combinatorial optimization, lecture notes, 1995-2001.

[3] Michel X. Goemans and David P. Williamson, Improved approximation

algorithms for maximum cut and satisfiability problems using semidefi-nite programming, J. ACM, Vol. 42, pp. 1115-1145, 1995.

[4] Abraham Berman and Naomi Shaked-Monderer, Completely positive

matrices, World Scientific Publishing Company, 2003.

[5] Konstantin Aslanidi, Notes on Quantitative Analysis in Finance, Avail-able: http://www.opentradingsystem.com, 2007.

參考文獻

相關文件

Numerical results are reported for some convex second-order cone programs (SOCPs) by solving the unconstrained minimization reformulation of the KKT optimality conditions,

Chen, The semismooth-related properties of a merit function and a descent method for the nonlinear complementarity problem, Journal of Global Optimization, vol.. Soares, A new

Particularly, combining the numerical results of the two papers, we may obtain such a conclusion that the merit function method based on ϕ p has a better a global convergence and

Then, it is easy to see that there are 9 problems for which the iterative numbers of the algorithm using ψ α,θ,p in the case of θ = 1 and p = 3 are less than the one of the

By exploiting the Cartesian P -properties for a nonlinear transformation, we show that the class of regularized merit functions provides a global error bound for the solution of

It is well known that second-order cone programming can be regarded as a special case of positive semidefinite programming by using the arrow matrix.. This paper further studies

In section 4, based on the cases of circular cone eigenvalue optimization problems, we study the corresponding properties of the solutions for p-order cone eigenvalue

volume suppressed mass: (TeV) 2 /M P ∼ 10 −4 eV → mm range can be experimentally tested for any number of extra dimensions - Light U(1) gauge bosons: no derivative couplings. =&gt;