• 沒有找到結果。

4 Continuation approach for the CMFWP

N/A
N/A
Protected

Academic year: 2022

Share "4 Continuation approach for the CMFWP"

Copied!
21
0
0

加載中.... (立即查看全文)

全文

(1)

to appear in Journal of Global Optimization, 2011

A continuation approach for the capacitated multi-facility Weber problem based on nonlinear SOCP reformulation

Jein-Shan Chen 1 Department of Mathematics National Taiwan Normal University

Taipei 11677, Taiwan

Shaohua Pan2

School of Mathematical Sciences South China University of Technology

Guangzhou 510640, China

Chun-Hsu Ko 3

Department of Electrical Engineering I-Shou University

Kaohsiung 840, Taiwan

September 4, 2009

(revised on July 21, 2010, December 2, 2010)

Abstract. We propose a primal-dual continuation approach for the capacitated multi- facility Weber problem (CMFWP) based on its nonlinear second-order cone program (SOCP) reformulation. The main idea of the approach is to reformulate the CMFWP as a nonlinear SOCP with a nonconvex objective function, and then introduce a logarithmic barrier term and a quadratic proximal term into the objective to construct a sequence of convexified subproblems. By this, this class of nondifferentiable and nonconvex optimization problems is converted into the solution of a sequence of nonlinear convex SOCPs. In this paper, we employ the semismooth Newton method proposed in [17] to solve the KKT system of the resulting convex SOCPs. Preliminary numerical results are reported for eighteen test

1Member of Mathematics Division, National Center for Theoretical Sciences, Taipei Office. The author’s work is partially supported by National Science Council of Taiwan. E-mail: jschen@math.ntnu.edu.tw, FAX: 886-2-29332342.

2The author’s work is supported by the Fundamental Research Funds for the Central Universities (SCUT), Guangdong Natural Science Foundation (No. 9251802902000001) and National Young Natural Science Foundation (No. 10901058). E-mail: shhpan@scut.edu.cn

3E-mail: chko@isu.edu.tw

(2)

instances, which indicate that the continuation approach is promising to find a satisfying suboptimal solution, even a global optimal solution for some test problems.

Key words: Capacitated multi-facility Weber problem, nondifferentiable, nonconvex, second-order cone program, semismooth Newton method.

1 Introduction

There are many different kinds of facility location problems for which various methods have been proposed; see [10, 15, 21, 20, 22, 23, 26, 32, 35, 36] and references therein. The web-site maintained by EWGLA (Euro Working Group on Location Analysis) also serves as a good resource to look for related literature including survey papers, books, and journals. In this paper, we consider the newer but more difficult capacitated multi-facility Weber problem (CMFWP) which plays an important role in the operation and management science.

The CMFWP, also called the capacitated Euclidean distance location-allocation prob- lem in other contexts, is concerned with locating a set of facilities and allocating their capacity to satisfy the demand of a set of customers with known locations so that the total transportation cost is minimized. Supply centers such as plants and warehouses may constitute the facilities, while retailers and dealers may be considered as customers. The mathematical model of CMFWP can be stated as follows:

min

m i=1

n j=1

cijwij∥xi− aj

s.t.

n j=1

wij = si, i = 1, 2, . . . , m (1)

m i=1

wij = dj, j = 1, 2, . . . , n

wij ≥ 0, i = 1, 2, . . . , m; j = 1, 2, . . . , n xi ∈ IR2, i = 1, 2, . . . , m.

In the formulation of CMFWP, m is the number of facilities to be located, n is the number of customers, si is the capacity of facility i, and djis the demand of customer j. Throughout this paper, we without loss of generality assume that the total capacity of all facilities equals the total demand of all customers, i.e.,

m i=1

si =

n j=1

dj. (2)

In addition, the allocations wij are unknown variables denoting the amount to be shipped from facility i to customer j with the unit shipment cost per unit distance being cij. If all

(3)

wij are known, the CMFWP reduces to the traditional convex multi-facility location prob- lem for which many efficient algorithms (see [10, 13, 22, 33, 34, 26]) have been proposed, whereas if all xi are fixed, it reduces to the ordinary transportation problem. For the sake of notation, in the sequel, we denote by xi = (xi1, xi2) the unknown coordinates of facility i, and by aj = (aj1, aj2) the given coordinates of customer j.

We see that the objective function of (1) is nondifferentiable at any points where xi = aj for some i ∈ {1, 2, . . . , m} and j ∈ {1, 2, . . . , n}, which precludes a direct application of effective gradient-based methods for finding its solution. Also, it is nonconvex since wij are unknown decision variables. Therefore, the CMFWP belongs to a class of nondifferentiable nonconvex optimization problems subject to (m + n) linear constraints and mn nonnega- tivity constraints. In fact, Sherali and Nordai [30] have shown that this class of problems is NP-hard, even if all demand points aj are located on a straight line.

For this class of problems, Cooper in his seminal work [7] first proposed an exact solu- tion method by explicit enumeration of all extreme points of the transportation polytope, defined by the first three groups of constraints of (1). Selim [29] later presented a biconvex cutting plane procedure in his unpublished dissertation. The exact solution method, simi- lar to Cooper’s complete enumeration, can effectively deal with very small instances only.

Recently, Sherali and Nordai [28] developed a branch-and-bound algorithm which is based on a partitioning of all the allocation space and finitely converges to a global optimum within a specified percentage tolerance. Apart from these exact methods, some heuristic methods have also been proposed by the reformulation linearization technique [27] or an approximating mixed integer linear programming formulation [1]. We observe that most of these methods are combinatorial primal ones which are designed by exploiting the structure of the problem itself and do not provide any information about the dual solution.

In this paper, we propose a continuous primal-dual approach by converting the CMFWP into the solution of a sequence of nonlinear convex second-order cone programs (SOCPs).

Specifically, we first reformulate the CMFWP as a nonlinear SOCP with a nonconvex cost function, and then introduce a logarithmic barrier term and a quadratic proximal term into the objective to circumvent the nonconvex difficulty. Among others, the strict convexity of logarithmic function and the strong convexity of quadratic proximal term are fully used to convexify the objective function of the resulting SOCP. Such a technique is not new which is often used in the literature; see [4] for example. The SOCP reformulation has recently attracted much attention for engineering and operations research problems. However, to our best of knowledge, they are all formed into linear SOCPs for which some softwares using interior point methods [18, 31] can be applied. In contrast, the nonlinear SOCP reformula- tion has little been used since all the aforementioned softwares are only able to solve linear SOCPs. This paper is concerned with the application of the nonlinear SOCP reformulation in the CMFWP, and its main purpose is to propose an alternative continuous approximate

(4)

method to handle the class of difficult problems, instead of introducing a highly specialized method in a competition for the best solution and the fasted computation time.

This paper is organized as follows. In Section 2, we review the general convex SOCP and the semismooth Newton method [17] for solving it. Section 3 presents a detailed process of reformulating (1) as a nonlinear SOCP. Section 4 proposes a primal-dual continuation approach for the CMFWP by solving approximately a sequence of convexified SOCPs. In Section 5, we report the preliminary numerical results for some test problems from [3, 28]

with the continuation approach, and compare the final objective values with those yielded by the global optimization methods [28]. Finally, we conclude this paper.

Throughout this paper, ∥ · ∥ denotes the Euclidean norm, IRn denotes the space of n- dimensional real column vectors, and IRn1× · · · × IRnm is identified with IRn1+···+nm. Thus, (x1, . . . , xm)∈ IRn1×· · ·×IRnm is viewed as a column vector in IRn1+···+nm. The notations I and 0 denote an identity matrix and a zero matrix of suitable dimension, respectively, and diag(x1, . . . , xn) means a diagonal matrix with x1, . . . , xn as the diagonal elements. Given a finite number of square matrices Q1, . . . , Qn, we denote the block diagonal matrix with these matrices as block diagonal by diag(Q1, . . . , Qn). For a differentiable function f , we denote by∇f(x) and ∇2xxf (x) the gradient and the Hessian matrix of f at x, respectively.

For a differentiable mapping G : IRn → IRm, we denote by G(x)∈ IRm×n the Jacobian of G at x. LetO be an open set in IRn. If G :O → IRnis a locally Lipschitz continuous, then

BG(x) :={H ∈ IRm×n| ∃{xk} ⊆ DG : xk → x, G(xk)→ H}

is nonempty and called the B-subdifferential of G at x ∈ O, where DG denotes the set of points at which G is differentiable. We assume that the reader is familiar with the concepts of (strongly) semismooth functions, and refer to [24, 25] for details.

2 Preliminaries

The convex SOCP is to minimize a convex function over the intersection of an affine linear manifold with the Cartesian product of second-order cones, which can be described as

minimize g(x)

subject to Ax = b, x∈ K, (3)

where g : IRn → IR is a twice continuously differentiable convex function, A is an m × n matrix with full row rank, b ∈ IRm and K is the Cartesian product of second-order cones (SOCs), also called Lorentz cones. In other words,

K = Kn1 × Kn2 × · · · × Knq, (4)

(5)

where q, n1, . . . , nq ≥ 1, n1+· · · + nq = n, and Kni denotes the SOC in IRni defined by Kni :=

{

xi = (xi1, xi2, . . . , xini)∈ IRni | x2i2+ . . . + x2in

i ≤ xi1

}

(5) with K1 denoting the nonnegative real number set IR+. A special case of (4) corresponds to the nonnegative orthant cone IRn+, i.e., q = n and n1 =· · · = nq = 1. When g is linear, clearly, (3) becomes a linear SOCP which has been investigated in many previous works and the interested reader is referred to the survey papers [2, 19] and the books [5, 6] for many important applications and theoretical properties.

The treatment of nonlinear convex SOCPs is much more recent and mainly focuses on the research of effective solution methods. Notice that K is a closed convex cone which is self-dual in the sense that K equals its dual cone K := {y ∈ IRn| ⟨y, x⟩ ≥ 0 ∀x ∈ K}.

Thus, it is easy to write the Karush-Kuhn-Tucker (KKT) optimality conditions of (3) as

∇g(x) − ATy− λ = 0 Ax− b = 0

⟨x, λ⟩ = 0, x ∈ K, λ ∈ K.

(6)

These conditions are also sufficient for optimality since g is convex. Based on the KKT system (6), there have been several methods proposed for solving (3), which include the smoothing Newton methods [8, 12, 16], the merit function method [9], and the semismooth Newton method [17]. As mentioned in the introduction, this paper is concerned with the application of the nonlinear SOCP in the CMFWP.

Since the resulting convex SOCPs in Sec. 4 will be solved with the semismooth Newton method in [17], we next review it. Let PK : IRn → IRn denote the Euclidean projection operator onto the cone K, i.e., PK(z) := argminy∈K{∥z − y∥} for any z ∈ IRn. Then, from [12, Prop. 4.1], we have that

x− PK(x− λ) = 0 ⇐⇒ x ∈ K, λ ∈ K, ⟨x, λ⟩ = 0. (7) Consequently, the solution of the convex SOCP (3) is equivalent to finding the zeros of

Φ(ω) := Φ(x, y, λ) :=

∇g(x) − ATy− λ Ax− b x− PK(x− λ)

. (8)

Since PK is strongly semismooth by [16, Prop. 4.5], using [11, Theorem 19] then yields that the operator Φ is at least semismooth, and furthermore, it is strongly semismooth if

2xxg(x) is locally Lipschitz continuous at any x∈ IRn. The semismooth Newton method in [17] finds a zero of Φ by applying the nonsmooth Newton method [24, 25] to the semismooth system Φ(ω) = 0. In other words, it generates the iterate sequence k = (xk, yk, λk)} by

ωk+1 := ωk− Wk−1Φ(ωk), (9)

(6)

where Wk is an arbitrary element from the B-subdifferential ∂BΦ(ωk) and has the form of

Wk=

2xxg(xk) −AT −I

A 0 0

I− Vk 0 Vk

for a suitable block diagonal matrix Vk= diag(V1k, . . . , Vqk) with Vik ∈ ∂BPKni(xki − λki).

The following two technical lemmas respectively provide the formula to compute the value of PK at any point and the representation of each element in V ∈ ∂BPKn(z).

Lemma 2.1 [17, Lemma 2.2] For any given z = (z1, z2)∈ IR × IRn−1, it holds that PKn(z) = max{0, µ1(z)}u(1)z + max{0, µ2(z)}u(2)z ,

where µ1(z), µ2(z) and u(1)z , u(2)z are the spectral values and the spectral vectors of z, respec- tively, given by

µ1(z) = z1− ∥z2∥, µ2(z) = z1+∥z2∥;

u(1)z = 1 2

(

1,−¯z2

)

, u(2)z = 1 2

(

1, ¯z2) with ¯z2 = ∥zz2

2 if z2 ̸= 0 and otherwise ¯z2 being any vector in IRn−1 satisfying ∥¯z2∥ = 1.

Lemma 2.2 [17, Lemma 2.6] Given a general point z = (z1, z2)∈ IR×IRn−1, each element V ∈ ∂BPKn(z) has the following representation:

(a) If z1 ̸= ±∥z2∥, then PKn(z) is continuously differentiable with

V = PKn(z) =

0 if z1 <−∥z2 I if z1 >∥z2 1

2

( 1 z¯2T

¯ z2 H

)

if − ∥z2∥ < z1 <∥z2∥, where

¯

z2 := z2

∥z2∥, H :=

(

1 + z1

∥z2

)

I− z1

∥z2∥z¯2z¯2T. (b) If z2 ̸= 0 and z1 =∥z2∥, then

V

{

I, 1 2

( 1 z¯2T

¯ z2 H

)}

, where ¯z2 := z2

∥z2 and H := 2I− ¯z2z¯2T.

(7)

(c) If z2 ̸= 0 and z1 =−∥z2∥, then V

{

0, 1 2

( 1 z¯T2

¯ z2 H

)}

, where ¯z2 := z2

∥z2 and H := ¯z2z¯2T. (d) If z = 0, then either V = 0 or V = I or V belongs to the set

{1 2

( 1 z¯2T

¯ z2 H

) H = (w0+ 1)I− w0z¯2z¯2T for some |w0| ≤ 1 and ∥¯z2∥ = 1

}

.

3 Nonlinear SOCP reformulation

In this section, we will present the detailed process of reformulating the capacitated multi- facility Weber problem (1) as a nonlinear SOCP with the form of (3). First, by introducing mn new variables tij for i = 1, 2, . . . , m, j = 1, 2, . . . , n, we can transform (1) into

min

m i=1

n j=1

cijwijtij

s.t. ∥xi− aj∥ ≤ tij, i = 1, 2, . . . , m; j = 1, 2, . . . , n

n j=1

wij = si, i = 1, 2, . . . , m (10)

m i=1

wij = dj, j = 1, 2, . . . , n

wij ≥ 0, i = 1, 2, . . . , m; j = 1, 2, . . . , n xi ∈ IR2, i = 1, 2, . . . , m.

We write out all the constraints of ∥xi− aj∥ ≤ tij, i = 1, 2, . . . , m; j = 1, 2, . . . , n as below:

(xi1− a11)2+ (xi2− a12)2 ≤ ti1, i = 1, 2, . . . , m,

(xi1− a21)2+ (xi2− a22)2 ≤ ti2, i = 1, 2, . . . , m, (11) ... ... ... ...

(xi1− an1)2+ (xi2− an2)2 ≤ tin, i = 1, 2, . . . , m.

Let {

uij := xi1− aj1, i = 1, 2, . . . , m, j = 1, 2, . . . , n;

vij := xi2− aj2, i = 1, 2, . . . , m, j = 1, 2, . . . , n. (12) Then, the constraints in (11) turn into

(ui1)2+ (vi1)2 ≤ (ti1)2, ti1≥ 0, i = 1, 2, . . . , m,

(ui2)2+ (vi2)2 ≤ (ti2)2, ti2≥ 0, i = 1, 2, . . . , m, (13)

... ... ... ... ...

(uin)2+ (vin)2 ≤ (tin)2, tin ≥ 0, i = 1, 2, . . . , m.

(8)

Let

ˆ xij :=

tij

uij vij

∈ IR3, i = 1, 2, . . . , m, j = 1, 2, . . . , n. (14)

From (13) and the definition ofK3, we have ˆxij ∈ K3 for i = 1, 2, . . . , m and j = 1, 2, . . . , n.

It should be pointed out that we have created additional constraints through the above reformulation procedure. In particular, from (12), it follows that

u11− u1k = ak1− a11, k = 2, 3, . . . , n, ... ...

um1− umk = ak1− a11, k = 2, 3, . . . , n (15) and

v11− v1k = ak2− a12, k = 2, 3, . . . , n, ... ...

vm1− vmk = ak2− a12, k = 2, 3, . . . , n. (16) We will see that (15)–(16) can be recast as a linear system. Note that (15) and (16) are equivalent to

[0 1 0]

t11 u11 v11

+ [0 −1 0]

t1k u1k v1k

= ak1− a11, k = 2, 3, . . . , n,

... ... ...

[0 1 0]

tm1 um1 vm1

+ [0 −1 0]

tmk umk vmk

= ak1− a11, k = 2, 3, . . . , n

and

[0 0 1]

t11 u11

v11

+ [0 0 −1]

t1k u1k

v1k

= ak2− a12, k = 2, 3, . . . , n

... ... ...

[0 0 1]

tm1 um1 vm1

+ [0 0 −1]

tmk umk vmk

= ak2− a12, k = 2, 3, . . . , n.

We can simplify (15)–(16) by introducing the following notations. More specifically, let

Au :=

0 1 0 0 −1 0

0 1 0 0 −1 0

... . ..

0 1 0 0 −1 0

∈ IR(n−1)×3n, (17)

(9)

Al :=

0 0 1 0 0 −1

0 0 1 0 0 −1

... . ..

0 0 1 0 0 −1

∈ IR(n−1)×3n, (18)

and denote

bu :=

a21− a11

a31− a11

... an1− a11

∈ IRn−1, bl :=

a22− a12

a32− a12

... an2− a12

∈ IRn−1.

Then, equations (15)–(16) can be recast as the following system of linear constraints

Au

Au

. ..

Au

−− − − − − − − − − − Al

Al

. ..

Al

x11] ... [ˆx1n]

−−

x21] ... [ˆx2n]

−−...

−−

xm1] ... [ˆxmn]

=

bu

bu ... bu

−−

bl bl ... bl

, (19)

where the dimensions in the linear system are 2m(n− 1) × 3mn, 3mn × 1, 2m(n − 1) × 1 for the matrix, column of variables, and column of constants, respectively.

We next look into the constraints on demand and capacity, namely the two groups of

(10)

equality constraints in (10). We notice that they can be recast as a linear system as below:

[1 1 1 · · · 1]

[1 1 1 · · · 1]

. ..

[1 1 1 · · · 1]

− − − − − − − − − − − − − − − − − − − − [1 0 0· · · 0] [1 0 0 · · · 0] · · · [1 0 0· · · 0]

[0 1 0· · · 0] [0 1 0 · · · 0] · · · [0 1 0· · · 0]

... ... ... ...

[0 0 0· · · 1] [0 0 0 · · · 1] · · · [0 0 0· · · 1]

w11 ... w1n

−−

w21 ... w2n

−−...

−−

wm1 ... wmn

=

s1 s2 ... sm

−−

d1 d2 ... dn

. (20)

Again, we point it out that the dimensions of the system (20) are (m + n)×(mn), (mn)×1, (m + n)× 1 for the matrix, column of variables and column of constants, respectively.

Moreover, the coefficient matrix has not full row rank due to the assumption (2), but its any m + n− 1 rows are all linear independent. For convenience, in the rest of this paper, Aw denotes the matrix composed of the first m + n−1 rows in the coefficient matrix of (20).

In summary, we reformulate the CMFWP (1) as the following nonlinear SOCP:

minimize

m i=1

n j=1

cijwijtij subject to Ax = b

x∈ (K3)mn× (K1)mn

(21)

where (K3)mn× (K1)mn denotes the Cartesian product of mn K3 and mnK1, and

x := (ˆx11, . . . , ˆx1n, . . . , ˆxm1, . . . ˆxmn, w11, . . . , w1n, . . . , wm1, . . . , wmn) , (22)

A :=

Au |

Au |

. .. |

Au |

− − − − − − − − − − − − | 0

Al |

Al |

. .. |

Al |

− − − − − − − − − − − − | − − −

0 | Aw

, (23)

(11)

b := (bu, bu, . . . , bu, bl, bl, . . . , bl, s1, . . . , sm, d1, . . . , dn−1) . (24) Notice that, by the expression (22) of x, the objective function of (21) can be rewritten as a quadratic function xTQx, where Q = [Qkl]4mn×4mn is a nonsymmetric matrix with

Qkl=

{ cij if k = 3(i− 1)n + 3(j − 1) + 1, l = 3mn + (i − 1)n + j;

0 otherwise (25)

for k, l = 1, 2, . . . , 4mn and i = 1, 2, . . . , m; j = 1, 2, . . . , n. Hence, (21) is equivalent to minimize xTQx

subject to Ax = b

x∈ (K3)mn× (K1)mn

(26)

where Q is a 4mn× 4mn matrix given by (25), A is a (2mn − m + n − 1) × 4mn matrix with full row rank (2mn− m + n − 1), and the dimension of b is (2mn − m + n − 1) × 1.

Since xTQx = xTQTx for any x∈ IR4mn, the SOCP (21) is further equivalent to minimize f (x) := 1

2xT(Q + QT)x subject to Ax = b

x∈ (K3)mn× (K1)mn

(27)

The SOCP (27) has an advantage over (26) that its Hessian matrix is symmetric, although its objective function is still nonconvex. In the next section, we develop a primal-dual algorithm for solving the CMFWP based on the nonlinear SOCP reformulation (27).

We should point out that the Hessian matrix of the function f , i.e., ¯Q = (Q + QT) and the constraint coefficient matrix A of (27) are both sparse. For example, when m = n = 2, Q = [ ¯¯ Qkl]16×16 has only eight nonzero entries: Q¯1,13 = ¯Q13,1 = c11, ¯Q4,14 = ¯Q14,4 = c12, ¯Q7,15 = ¯Q15,7 = c21, ¯Q10,16 = ¯Q16,10 = c22, whereas A = [Aij]7×16 has sixteen nonzero entries: A12 = 1, A15 =−1, A28 = 1, A2,11 =−1, A33 = 1, A36 =−1, A49 = 1, A4,12 =

−1, A5,13 = A5,14 = 1, A5,15 = A5,16 = 1, A6,13 = A6,15 = 1, A7,14 = A7,16 = 1. In addition, we observe that each row of ¯Q has at most a nonzero entry ci1,j1 with ci1,j1 {c11, . . . , c1n, . . . , cm1, . . . , cmn}, and all diagonal entries are zero.

4 Continuation approach for the CMFWP

This section designs a primal-dual approximate algorithm for the nonconvex SOCP (27).

The main idea is transforming (27) into the solution of a sequence of convexified SOCP subproblems.

(12)

First, note that the variables wij are nonnegative and restricted by the linear constraints

n

j=1wij = si and mi=1wij = dj, and hence the nonconvex SOCP (27) is equivalent to min f (x)

s.t. Ax = b

0≤ wij ≤ min{si, dj}, i = 1, . . . , m; j = 1, . . . , n x∈ (K3)mn× (K1)mn.

(28)

Since the logarithmic barrier function mi=1nj=1[ln(wij) + ln(min{si, dj} − wij)] is well defined when 0 < wij < min{si, dj} for all i = 1, 2, . . . , m and j = 1, 2, . . . , n, and moreover, its limit is +∞ if some wij tends to 0 or min{si, dj}, we can dispense with the bound constraints on wij in (28) and obtain the following transformed problem:

min f (x)− τm

i=1

n j=1

[

ln(wij) + ln(min{si, dj} − wij)] s.t. Ax = b

x∈ (K3)mn× (K1)mn

(29)

where τ > 0 is a barrier parameter that finally tends to 0. The above operation seems to purposely make the original problem (27) more complex. However, we will see that the introduction of the logarithmic barrier term gives a help in locating the feasible solution of (27), as well as plays a certain role in convexifying f (x).

Now, we introduce a quadratic proximal term 1

2∥x − z∥2 with z ∈ IR4mn being a given vector into the objective of (29) to further convexify f (x). Define

g(x, z, τ, ε) := f (x)− τm

i=1

n j=1

[

ln(wij) + ln(min{si, dj} − wij)]+ε

2∥x − z∥2 (30) where ε > 0 is a proximal parameter that will become larger until over some threshold.

Proposition 4.1 Given a vector z ∈ IR4mn, let g be defined as in (30) for any τ, ε > 0.

Then, there exists ε0 > 0 such that g is strictly convex for any ε > ε0 and τ > 0.

Proof. For given z∈ IR4mn and any τ, ε > 0, we compute the Hessian matrix of g as

2xxg(x, z, τ, ε) = Q + QT + diag(ε, . . . , ε

| {z }

3mn

, P11+ε, . . . , P1n+ε, . . . , Pm1+ε, . . . , Pmn+ε)

with Pij for i = 1, 2, . . . , m and j = 1, 2, . . . , n given by Pij = τ

(wij)2 + τ

(min{si, dj} − wij)2 > 0.

(13)

Since all diagonal entries of Q + QT are zero and each row of Q + QT has at most a nonzero entry ci1,j1 with ci1,j1 ∈ {c11, . . . , c1n, c21, . . . , c2n, . . . , cm1, . . . , cmn}, we have that

[2xxg(x, z, τ, ε)]

kk >

4mn

l=1,l̸=k

[2xxg(x, z, τ, ε)]

kl

for all k = 1, 2, . . . , 4mn whenever ε > ε0 with ε0 := max

1≤i≤m,1≤j≤n{cij}. (31)

In other words, the matrix 2xxg(x, z, τ, ε) is strictly diagonally dominant for any ε > ε0 and τ > 0. From Corollary 7.2.3 of [14], it then follows that 2xxg(x, z, τ, ε) is positive definite, and consequently g(x, z, τ, ε) is strictly convex, for any ε > ε0 and τ > 0. 2

Proposition 4.1 states that for a given z ∈ IR4mn, the function g(x, z, τ, ε) is strictly convex for any ε > ε0 and τ > 0. In fact, it is also strongly convex when ε > ε0. In view of this, our continuation approach seeks for an approximate optimal solution of the CMFWP by solving a sequence of the following subproblems

min g(x, ˆxk, τk, εk) s.t. Ax = b

x∈ (K3)mn× (K1)mn

(32)

with a decreasing sequencek} and a increasing sequence {εk}, where ˆxk is a vector given by the solution of the last subproblem. For a fixed k, since the subproblem (32) is a convex SOCP whenever εk > ε0, its solution is easy, which from Section 2 is equivalent to solving

xg(x, ˆxk, τk, εk)− ATy− λ = 0 Ax− b = 0

x− PK(x− λ) = 0

(33)

with K = (K3)mn× (K1)mn. Define the mapping Φk : IR10mn−m+n−1→ IR10mn−m+n−1 by

Φk(ω) = Φk(x, y, λ) :=

xg(x, ˆxk, τk, εk)− ATy− λ Ax− b

x− PK(x− λ)

. (34)

Then, solving the nonsmooth system (33) is equivalent to finding the zero of the operator Φk(ω). We will attain this goal by using the semismooth Newton method in [17].

Next we describe the iteration steps of the continuation approach in which two fixed constants c1 ∈ (0, 1) and c2 > 1 are used to reduce and increase the dynamic parameters τ and ε. Let Ψk(w) := 1

2∥Φk(ω)∥2 denote the natural merit function of system Φk(ω) = 0.

Algorithm 4.1 (Continuation Approach )

(14)

(S.0) Given the constants c1 ∈ (0, 1), c2 > 1 and ˆτ , ˆε > 0. Select τ0, ε0 > 0 and a starting point ¯ω0 = (¯x0, ¯y0, ¯λ0) with the last mn elements ¯w0ij of ¯x0 satisfying 0 < ¯w0ij <

min{si, dj}. Let ˆx0 := ¯x0, and set k := 0.

(S.1) Compute (by Algorithm 4.2 below) an approximate optimal solution ωk = (xk, yk, λk) of (33) with the starting point ¯ωk = (¯xk, ¯yk, ¯λk).

(S.2) If τk < ˆτ and εk > ˆε, then stop, and otherwise go to (S.3).

(S.3) Modify the parameters τk and εk by τk+1 := c1τk and εk+1 := c2εk, and let

¯

xk+1:= xk, y¯k+1 := yk, λ¯k+1 := zk, xˆk+1:= xk. (S.4) Set k := k + 1, and go to (S.1).

Note that the continuation approach is a primal-dual one and the last mn components of the final iterate yk provide an approximate dual solution of the CMFWP, which usually has a certain economic meaning associated with this class of transportation problems. In addition, the main computation work of Algorithm 4.1 is to seek an approximate solution of (33). Such a solution can be easily obtained when εk is large enough since, on the one hand, the mappingxg(x, ˆxk, τk, εk) is strongly monotone in this case, which together with Lemma 3.1 of [16] implies that the function Ψk(ω) has bounded level sets; and on the other hand, from Section 2, the operator Φk is at least semismooth which guarantees in theory a fast algorithm with a superlinear (or quadratic) convergence rate, to seek the solution of (33). Of course, we should emphasize that, when computing the approximate solution of (33), we must restrict the maximum steplength of the iterates (see Algorithm 4.2 below).

We next describe the specific iteration steps of the semismooth Newton method [17]

when applying it for solving the semismooth system (33). For a given k, from Section 2 it follows that the main iteration step of the method is as follows:

ωl+1 := ωl− Wl−1Φkl), Wl ∈ ∂BΦkl) for l = 0, 1, 2, . . . ,

where ∂BΦkl) denotes the B-subdifferential of the semismooth mapping Φk at the point ωl and any element Wl of ∂BΦkl) has the expression

Wl :=

2xxg(xl, ˆxk, εk, τk) −AT I

A 0 0

I− Vl 0 Vl

(35)

for a suitable block diagonal matrix Vl = diag(V1l, . . . , V4mnl ) with Vil ∈ ∂BPK3(xl− λl) for i = 1, 2, . . . , 3mn and Vil ∈ ∂BPK1(xl− λl) for i = 3mn + 1, . . . , 4mn.

Algorithm 4.2 (Solving the subproblem (33) for Step (S.1) of Algorithm 4.1)

參考文獻

相關文件

Section 3 is devoted to developing proximal point method to solve the monotone second-order cone complementarity problem with a practical approximation criterion based on a new

Abstract Based on a class of smoothing approximations to projection function onto second-order cone, an approximate lower order penalty approach for solving second-order cone

Based on a class of smoothing approximations to projection function onto second-order cone, an approximate lower order penalty approach for solving second-order cone

Although we have obtained the global and superlinear convergence properties of Algorithm 3.1 under mild conditions, this does not mean that Algorithm 3.1 is practi- cally efficient,

In section 4, based on the cases of circular cone eigenvalue optimization problems, we study the corresponding properties of the solutions for p-order cone eigenvalue

Since the generalized Fischer-Burmeister function ψ p is quasi-linear, the quadratic penalty for equilibrium constraints will make the convexity of the global smoothing function

Abstract We investigate some properties related to the generalized Newton method for the Fischer-Burmeister (FB) function over second-order cones, which allows us to reformulate

Fukushima, On the local convergence of semismooth Newton methods for linear and nonlinear second-order cone programs without strict complementarity, SIAM Journal on Optimization,