• 沒有找到結果。

String-averaging expectation-maximization for maximum likelihood estimation in emission tomography

N/A
N/A
Protected

Academic year: 2021

Share "String-averaging expectation-maximization for maximum likelihood estimation in emission tomography"

Copied!
21
0
0

加載中.... (立即查看全文)

全文

(1)

This content has been downloaded from IOPscience. Please scroll down to see the full text.

Download details:

IP Address: 140.113.38.11

This content was downloaded on 25/12/2014 at 02:49

Please note that terms and conditions apply.

String-averaging expectation-maximization for maximum likelihood estimation in emission

tomography

View the table of contents for this issue, or go to the journal homepage for more 2014 Inverse Problems 30 055003

(2)

Inverse Problems 30 (2014) 055003 (20pp) doi:10.1088/0266-5611/30/5/055003

String-averaging expectation-maximization

for maximum likelihood estimation in

emission tomography

Elias Salom ˜ao Helou

1

, Yair Censor

2

, Tai-Been Chen

3

,

I-Liang Chern

4,5

, ´

Alvaro Rodolfo De Pierro

1

, Ming Jiang

6

and Henry Horng-Shing Lu

7

1Department of Applied Mathematics and Statistics, State University of S˜ao Paulo,

Postal Box 668, S˜ao Carlos, SP, Brazil

2Department of Mathematics, University of Haifa, Mt. Carmel, Haifa 3190501, Israel 3Department of Medical Imaging and Radiological Sciences, I-Shou University,

Kaohsiung City, Taiwan 82445, ROC

4Department of Applied Mathematics, Center of Mathematical Modeling and

Scientific Computing, National Chiao Tung University, Hsin Chu, Taiwan 30010, ROC

5Department of Mathematics, National Taiwan University, Taipei, Taiwan 10617,

ROC

6LMAM, School of Mathematical Sciences, Beijing International Center for

Mathematical Research, Peking University, Beijing 100871, People’s Republic of China

7Institute of Statistics, National Chiao Tung University, 1001 University Road,

Hsinchu, Taiwan 30010, ROC

E-mail:elias@icmc.usp.br,yair@math.haifa.ac.il,ctb@isu.edu.tw,

chern@math.ntu.edu.tw,alvaro@ime.unicamp.br,ming-jiang@ieee.organd

hslu@stat.nctu.edu.tw

Received 8 May 2013, revised 31 January 2014 Accepted for publication 5 February 2014 Published 17 April 2014

Abstract

We study the maximum likelihood model in emission tomography and propose a new family of algorithms for its solution, called string-averaging expectation-maximization (SAEM). In the string-averaging algorithmic regime, the index set of all underlying equations is split into subsets, called ‘strings’, and the algorithm separately proceeds along each string, possibly in parallel. Then, the end-points of all strings are averaged to form the next iterate. SAEM algorithms with several strings present better practical merits than the classical row-action maximum-likelihood algorithm. We present numerical experiments showing the effectiveness of the algorithmic scheme, using data of image reconstruction problems. Performance is evaluated from the computational

(3)

cost and reconstruction quality viewpoints. A complete convergence theory is also provided.

Keywords: positron emission tomography (PET), string-averaging, block-iterative, expectation-maximization (EM) algorithm, ordered subsets expectation maximization (OSEM) algorithm, relaxed EM, string-averaging EM algorithm

(Some figures may appear in colour only in the online journal) 1. Introduction

The expectation-maximization (EM) algorithm (see, e.g., [1, chapter 7] or [2]) has become a household tool for maximum likelihood estimation in emission tomography (ET). The original EM algorithm is simultaneous since when passing from a current iterate xkto the next one xk+1, all equations ai, x = b

i, i = 1, 2, . . . , m, in the linear system Ax = b of the underlying

problem are used. In contrast to this, there is the row-action maximum likelihood algorithm (RAMLA) (of Browne and De Pierro [3]) which is structurally sequential since each iteration

k involves only one equationai(k), x = bi(k), where i(k) is one of the indices {1, 2, . . . , m}

chosen at the kth iteration, from the linear system of the underlying problem. In-between these two structural extremes we have also a block-RAMLA structure [3, page 689] and the ordered subsets EM (OSEM) algorithm (of Hudson and Larkin [4]), both allowing to process in each iteration a ‘block’ (i.e., a subset) of the m underlying equations, by applying to the equations of the block the simultaneous original EM iterative step. Section3brings a deeper discussion on alternative methods and related works.

In this article we propose and experiment with a new variant of the EM algorithm which uses string-averaging (SA), thus we call the resulting algorithm the SAEM algorithm. In the SA algorithmic regime the index set of the m underlying equations is again split into subsets, now called ‘strings’, and from a current iterate xkthe algorithm first proceeds sequentially along

the indices of each string separately (which can be done in parallel) and then the endpoints of all strings are averaged to form the next iterate xk+1. This is in stark contrast with how

the above mentioned OSEM and block-RAMLA treat the equations of a block. Full details regarding the algorithm will be given in section4.

We advocate SAEM as a theoretically sound way to provide better algorithmic behaviour in ET image reconstruction. In order to support our claims, simulation studies performed with the intention to evaluate image quality and computational effort are reported in section6and a theoretical study of the convergence properties of the method can be found in section 5. Next section introduces the fundamentals of the problem and section7brings our concluding remarks.

2. ET image reconstruction: the problem, approaches and related works

Positron Emission Tomography (PET) and Single Photon Emission Tomography (SPECT) are

ideal modalities for in vivo functional imaging. Thus, ET can be used to monitor the effects of therapy inside the living body or in oncological studies. The image reconstruction problem in ET can be fully-discretized and modeled by a system of linear equations [10]:

Ax= b (1)

where the measured data is b= (bi)mi=1 ∈ R m

, the image vector is x= (xj)nj=1 ∈ R n

, and A is an m× n real matrix, and the problem is to estimate the image vector x from the measured

(4)

data b. Solution of this system by classical direct elimination methods will likely be impractical because of the huge data and image dimensions, ill-posedness of A coupled with noisy and/or incomplete data b, unstructured imaging matrix A, among several other well-known reasons [10,11].

Instead, iterative methods, especially row-action methods, such as the algebraic reconstruction technique (ART) [17,18], are usually applied for computationally efficient high-quality reconstruction [16]. Iterative algorithms have been important in this field because of their superior reconstructions over analytical methods in many instances, and their ability to handle the very large and sparse data associated with fully-discretized image reconstruction, see, e.g., [5,11–13]. For reviews see, e.g., [14] and [15].

Another key advantage of iterative methods is their flexibility. For example, least-squares and I-divergence optimization approaches are commonly used for image reconstruction. In these approaches, the true image is estimated by the minimizer of an objective function, i.e., as a solution of a convex optimization problem. The least-squares functional is given by

LLS(x) := 12b − Ax2, ∀x ∈ Rn. (2)

ART is a prominent algorithm in this approach, which can be turned to be efficient for ET [16] even though it was originally proposed for x-ray tomography [17,18]. The I-divergence (also called the Kullback–Leibler distance) functional is given by

LKL(x) := I(b, Ax) = m  i=1 bilog bi ai, x+ a i, x − b i, ∀x ∈ Rn+, (3)

where the vector ai=ai j

n

j=1is the transpose of the ith row of A. From a statistical perspective,

the least-squares approach is equivalent to finding an image as the maximum likelihood estimate from data with Gaussian noise, while the I-divergence approach does the same with Poissonian noise.

For the image reconstruction problem in ET, the I-divergence approach is more appropriate because Poissonian noise is dominant. The following iterative maximum likelihood expectation maximization recursion, termed the MLEM (or the classical EM) algorithm, is popular since the 1980s [9] for the minimization of (3):

xkj+1:= x k j 1 m i=1aij m  i=1 aij bi ai, xk, for j = 1, 2, . . . , n. (4)

It was first proposed independently by Richardson [19] and Lucy [20] and for that reason is still known as the Richardson–Lucy (RL) algorithm in applications to astronomy and microscopy. It was later rediscovered in [7] and in [8] by applying the general EM framework from statistics [21,22] to the ET reconstruction problem.

In addition to the above two approaches, there are those which utilize other Bregman divergences (distances) and the Bayesian framework for image reconstruction [10,11,31], of which the above least-squares and I-divergence approaches are special cases. In fact, SAEM can be readily generalized to cover these cases, because our analysis of convergence applies to the minimization of a general convex function over the nonnegative orthant.

From yet another perspective, the image reconstruction task can also be viewed as a

convex feasibility problem (CFP) see [14]. Consequently the method based on projection onto convex sets (POCS) provides significant inspiration when developing iterative reconstruction algorithms, see, e.g., [11, chapter 5] or [23]. There have been various iterative projection algorithms based on different kinds of projections such as the Euclidean, oblique and Bregman

(5)

projections. Among them, the string-averaging projections (SAP) algorithmic scheme has attracted attention recently since its presentation in [24]. Therefore, the present paper brings an algorithmic bridge for this SA scheme, going from the set theoretic to the maximum likelihood framework.

3. EM-type algorithms

3.1. The classical EM algorithm

For A, ai, b, and bi, as defined above for i∈ I := {1, 2, . . . , m}, we define RiEM: Rn → Rnas

the ith row operator for the EM algorithm,

RiEM(x) := bi

ai, xx. (5)

For any index subset B ⊆ I, let |B| denote the number of elements in B and let RB be the

averaging operator RB: R|B|×n→ Rndefined componentwise as (RB({yi}i∈B))j:= 1 m i=1a i j  i∈B aijyij, for j = 1, 2, . . . , n. (6) The classical EM algorithm and many of its variants for image reconstruction are simultaneous algorithms, see, e.g., [11, section 1.3] on the classification of iterative algorithms from the perspective of the algorithmic parallelization strategy. In pseudo-code, the classical EM algorithm, that uses (4), is as follows:

Algorithm 3.1. The classical EM algorithm

f or k= 0, 1, . . . f or i∈ I xk+1,i:= RiEM(x k), (7) end xk+1 := RI({xk+1,i}i∈I). (8) end

It first executes simultaneously (in parallel) row-action operations on all rows from the same current iterate xk. Then the intermediate iteration vectors{xk+1,i}mi=1are combined by the

averaging operator RIto obtain the next iterate.

3.2. The block iterative EM algorithm

The classical EM algorithm can be accelerated by its block-iterative version [4, 25, 26]. A block-iterative version of a simultaneous algorithm, see [11, section 1.3], employs the algorithmic operators Ri

EMand RBas the simultaneous algorithm does, but it first breaks up

the index set I = {1, 2, . . . , m} into ‘blocks’ of indices so that for t = 1, 2, . . . , T, the block

Btis a subset of I of the form Bt :=



it1, it2, . . . , itm(t)



. (9)

with m(t) denoting the number of elements in Bt. The block-iterative version of the classical

(6)

Algorithm 3.2. The block-iterative EM algorithm f or k= 0, 1, . . . t= t(k) ∈ {1, 2, . . . , T} (10) f or i∈ Bt xk+1,i:= RiEM(x k), (11) end xk+1 := RBt({x k+1,i} i∈Bt). (12) end

At the kth iteration, the active block Bt is determined by the control sequence t = t(k).

Then, for each block the algorithm performs a simultaneous step as if the classical EM method was applied to this block alone and the iteration is updated. In the literature of maximum likelihood reconstruction for ET, block-iterative EM algorithms are called OSEM algorithms as they were named in the work of Hudson and Larkin [4] see also [30,33]. For block-iterative projection methods see, e.g., [27–29].

3.3. EM-type algorithms with relaxation

The OSEM algorithm has an improved experimental convergence rate, but may not converge when the system (1) is inconsistent [3]. This can be resolved by introducing relaxation as in the least-squares approach [35]. The first block-iterative EM algorithm with relaxation is the ‘row-action maximum likelihood algorithm’ (RAMLA) in [3]. See [34] for row-action methods in general. After statistical study of the noise propagation from the projection data to the reconstructed image, a modified version of the RAMLA, called ‘dynamic RAMLA’ (DRAMA), using variable relaxation parameters with blocks, was proposed in [36]. A formula extending RAMLA and DRAMA was proposed in [37] as follows,

xk+1= xk− λk,tD(xk)∇LBKLt (xk), (13)

whereλk,tare stepsize parameters,∇LBKLt is the gradient of the partial negative log-likelihood LBt KL(x) :=  i∈Bt bilog bi ai, x+ a i, x − b i, (14)

and the pre-conditioner D is given by

D(x) := diag  xj pj 1  j  n , (15)

with positive constants pj> 0 . Possible choices of pjare pj= max 1tT  i∈Bt aij  , (16) or pj= m i=1a i j T . (17)

(7)

Convergence results established in [37] require that the following conditions are met. Condition A: x0∈ Rn

+, b∈ Rm+, A∈ Rm+×n,Tt=1L Bt

KL(x) = LKL(x).

Condition B: rank(W (x)12A) = n, where W (x) 1

2 is the component-wise square root of the

diagonal matrix W(x) = diag  bi  ai, x2  1  j  n  . (18)

This is equivalent to the strict convexity of the function LKL(x).

Condition C: 0< λk,t λ, where λ is chosen such that λ < min pj  i∈Bta i j 1  j  n,1  t  T  , (19)

and, denotingλk= λk,1, such that ∞  k=0 λk= ∞; ∞  k=0 λ2 k< ∞; ∞  k=0 λk− λk,t < ∞; (20) λk,t λk → 1; and λk λk+1 < constant. (21) The above assumptions on the relaxation parameters are very general and cover several situations of interest [32,36,40,41]. If each block size is one, i.e., each block corresponds to one equation, then by using (13) and (15), we get the relaxed row-action iteration of RAMLA:

xkj+1= xkj+ λk,i ai j pj bi ai, xk − 1 xkj, for j = 1, 2, . . . , n. (22)

4. The string-averaging scheme

4.1. The string-averaging prototypical scheme

The SA algorithmic regime was originally formulated in [24] in general terms and applied there for solving the CFP with iterative projection algorithms, see, e.g., [24,42,43]. In the SA paradigm, the index set I = {1, 2, . . . , m} is split into ‘strings’ of indices. From the current iterate xk, certain algorithmic operators (we shall call them step operators in the following) are

applied sequentially along the indices of each string and the end-points of all strings are then combined by an additional algorithmic operator (which we name the combination operator from now on) to yield the next iteration vector.

To define SA algorithms precisely, the same decomposition as in (9) for the index set I is utilized. The index set

St:=  it1, it2, . . . , itm(t)  =its| 1  s  m(t)  (23) is now serving as the string of indices in the current context, for t= 1, 2, . . . , T. Viewed like that, strings and blocks are just names for index subsets Bt ⊆ I or St ⊆ I for t = 1, 2, . . . , T.

Interleaving of strings and blocks is possible, leading to algorithms with a tree-like parallelism structure, whose convergence properties can be almost directly devised from our analysis below.

Let us consider a set Q⊂ Rnand family of operators{Ri}m

i=1mapping Q into itself, and

an additional operator R which maps QT (i.e., the product of T copies of Q) into Q. In the

SA paradigm these operators{Ri}m

i=1 are the step operators and the operator R serves as the

(8)

Algorithm 4.1. The string-averaging prototypical scheme [24] Initialization x0∈ Q is an arbitrary starting point. Iterative Step Given the current iterate xk,

(i) for all t = 1, 2, . . . , T, compute in parallel as follows: apply successively the step

operator along the string St,

xk+1,t := Ritm(t)◦ . . . ◦ Rit2◦ Rit1(xk), (24)

(ii) apply the combination operator

xk+1:= R{xk+1,i}Tt=1



. (25)

For every t = 1, 2, . . . , T, this algorithmic scheme first applies to xk successively the step

operators Ri whose indices i belong to the tth string S

t. This can be done in parallel for all

strings because the jobs of going along each string from (one and the same current iterate) xk

to the end-point xk+1,t are independent. Then the combination operator R maps all end-points

onto the next iterate xk+1. The iteration from xkto xk+1is called one cycle of iteration, whereas

the iteration from xk to xk+1,t is called the tth sub-iteration in the kth cycle. Notice that we

can always obtain from this framework a fully-sequential algorithm by the choice T = 1 and

S1 = I or a fully-simultaneous algorithm by the choice T = m and St = {t}, t = 1, 2, . . . , T.

4.2. The string-averaging EM algorithm

Here we merge the SA algorithmic structure described above with the maximum likelihood estimator in order to create the new string-averaging EM (SAEM) algorithm. To this end we adopt the row-action operation of RAMLA as step operators (22), namely:

 Riλk,i(xk)j:= xkj+ λk,i ai j pj bi ai, xk − 1 xkj, for j = 1, 2, . . . , n. (26)

To combine end-points of strings we use convex combinations, thus, our algorithm takes the following form.

Algorithm 4.2. The string-averaging-EM (SAEM)

Initialization Choose parameters pj, for j = 1, 2, . . . , n, and parameters λk,i. Choose x0 > 0 as an arbitrary initial vector, construct a family of strings {S

t}tT=1, choose a weight system{wt}tT=1such thatwt > 0 for all t = 1, 2, . . . , T, and

T

t=1wt = 1.

Iterative Step Given the current iterate xk,

(i) for all t= 1, 2, . . . , T, compute (possibly in parallel) as follows: apply successively

the RAMLA row-action iterative step given by (26) along the string St,

xk+1,t := RSt(x) := Ritm(t)◦ . . . ◦ Rit2◦ Rit1(xk), (27) where the dependence of each step operator in (27) on the relaxation parameterλk,it

s,

s= 1, 2, . . . , m(t) has been left out for clarity.

(ii) then combine the end-points by

xk+1= T



t=1

wtxk+1,t (28)

using the weights system{wt}Tt=1.

5. Theoretical justification

In this section we provide a general justification for why SAEM algorithms present a convergent behavior. We base our proofs on the fact that the convex combination operator, used in the

(9)

averaging step, preserves certain asymptotic characteristics of the step operator and, therefore, we can expect convergence whenever the step operator comes from a convergent algorithm as is the case here.

Incremental algorithms [38], such as those used for the stringing by the step operator of the SAEM algorithm, satisfy an approximation given by

RSt(x, λ) = x − λD(x)∇LSt(x) + O(λ2), (29)

according to proposition5.3below. In the current theory, the general form (29) allows us to prove convergence under suitable hypotheses for non-averaged algorithms, mainly because

LSt = L and, therefore, for such iterations we have:

xk+1= xk− λkD(xk)∇L(xk) + O  λ2 k  . (30)

But any conclusion drawn from equation (30) should hold as well for a properly averaged algorithm, by replacing L by the following ˜L:

˜L(x) :=T

t=1

wtLSt(x), (31)

wherewt> 0 are the weights. For the SAEM algorithm4.2with

 wi= 1, we have xk+1= xk− λkD(xk)∇ ˜L(xk) + O  λ2 k  . (32)

If the strings Stare pairwise disjoint and such that

T t=1St= {1, 2, . . . , m} and wt = 1/T, we have xk+1= xkλk T D(x k)∇L(xk) + Oλ2 k  (33) for the averaged iteration, and convergence will be toward the optimizer of ˜L= L. In order to

make the discussion more precise, from now on we consider the following general form of a string-averaging algorithm:

Algorithm 5.1. General string-averaging algorithm

f or k= 0, 1, . . . f or t= 1, 2, . . . , T xk+1,t := RSt(xk, λ k), (34) end xk+1 := T  t=1 ωtxk+1,t (35) end

By considering this algorithm we handle step operators RSt of a rather general form, but stick to the concrete realization of the combination operator. For this kind of iteration we have the following result, wherein is a real nonnegative function of λ.

Proposition 5.2. Suppose thatωt > 0 and

T

t=1ωt = 1 in algorithm5.1, and that for every t∈ {1, 2, . . . , T} the following equality holds for all k  0,

RSt(xk, λ

k) = xk− λkD(xk)∇LSt(xk) + O((λk)). (36) Then we have

xk+1= xk− λkD(xk)∇ ˜L(xk) + O((λk)), (37)

(10)

Proof. The result is verified by observing that xk+1= T  t=1 ωtRSt(xk, λk) = T  t=1 ωtxk− λkD(xk)∇ T  t=1 ωtLSt(xk) + T  t=1 ωtO((λk)) = xk− λ kD(xk)∇ ˜L(xk) + O((λk)), (38)

where we have used the definition of the algorithm for the first equality, hypothesis (36) for the second, and then appliedTt=1ωt = 1, the definition of ˜L, and a trivial property of the O

notation to obtain the third equation. 

We now prove the claim that our stringing operators RSt do satisfy an equation such as (36) with(λk) = λ2k, as long as Lipschitz continuity holds for each parcel of D(x

k)LSt(xk) used during each stringing operation, according to the next statement.

Proposition 5.3. Let LSt = i∈StL iand RSt(x, λ) := Ri t m(t) λ ◦ . . . ◦ R it 2 λ ◦ R it 1 λ(x), where each Riλ for i∈ {1, 2, . . . , m} is given by Riλ(x) := x − λD(x)∇Li(x) (39) For k= 1, 2, . . ., denote yk,0:= xk, (40) yk,s:= Rits λk(y k,s−1), s = 1, 2, . . . , m(t), (41) xk+1,t := yk,m(t) (42) xk+1:= T  t=1 ωtxk+1,t. (43)

Assume that each D(·)∇Lit

j(·) is Lipschitz continuous with a Lipschitz constant M and bounded

by an upper bound N on the set{xk, yk,s}

k∈N,s∈I, then the operator RSt satisfies RSt(xk, λ k) = xk− λkD(xk)∇LSt(xk) + O  λ2 k  . (44)

Proof. Unfolding of the definition of the operator leads to

RSt(xk, λ k) = xk− λk m(t)  s=1 D(yk,s−1)∇Lits(yk,s−1). (45)

Consider the magnitude of the following difference,

k:= D(xk)∇LSt(xk) − λ k m(t)  s=1 D(yk,s−1)∇Lit s(yk,s−1), (46) which we can estimate as follows,

k =  m(t)  s=1 (D(xk)∇Lit s(xk) − D(yk, j−1)∇Lits(yk, j−1))    (47)  m(t)  s=1 D(xk)∇Lit s(xk) − D(yk, j−1)∇Lits(yk, j−1) (48)  m(t)  s=1 Mxk− yk,s−1, (49)

(11)

where the first inequality follows from the triangular inequality, and the latter from Lipschitz continuity. Now we notice that if s> 1, then

xk− yk,s = λk s  l=1 D(yk,l−1)∇Litj(yk,l−1)  λkmN (50)

because of the boundedness hypothesis. Putting (49) and (50) together leads tok  λ kK

for some large enough constant K, i.e.,k= O(λ

k). On the other hand, from (45) and (46) RSt(xk, λ

k) = xk− λkD(xk)∇LSt(xk) + λkk, (51)

which implies the assertion, sinceλO(λ) = O(λ2). 

We conclude by making the following connection.

Corollary 5.4. Under the assumptions of propositions5.3and5.2, algorithm5.1satisfies, for all k 0, xk+1= xk− λkD(xk)∇ ˜L(xk) + O  λ2 k  . (52)

Proof. Use proposition5.3followed by proposition5.2.  So far we have shown how the string-averaging iteration approximates a scaled gradient iteration. From now on we study how to obtain convergence from these approximate steps, which requires conditions on the sequence of parameters{λk} ⊂ R+. We will need



k=1λk = ∞ because we will usually assume some sort of boundedness on D(xk)∇L(xk)

and, consequently, we would never reach optimal points if they happen to be too far away from the initial guess. On the other hand, if we do not imposeλk → 0, the error O(λ2k) in

the approximation does not necessarily vanish and will eventually dominate the computations whenD(xk)∇L(xk) becomes small. This must happen if we wish to converge to an optimal point since a necessary condition on a solution x∗for the non-negatively constrained problem is D(x)∇L(x) = 0. We now show that, with no further hypotheses, at least one subsequence generated by the algorithm is of interest.

We will need the following lemma to prove the next proposition.

Lemma 5.5 (lemma 2, [44]). Assume that{ek}, {νk} and {dk} are sequences of nonnegative numbers satisfying for all k 0,

ek+1 ek− νkdk (53)

and that∞k=0νk= +∞. Then 0 is a cluster point of {dk}.

Proposition 5.6. Suppose that the algorithm generating{xk}k∈Nsatisfies (52),



k∈Nλk= ∞, λk→ 0+, ˜L is smoothly differentiable and ˜L(xk) is well-defined. Assume also that {xk} ⊂ Rn+ is bounded. Then either lim ˜L(xk) = −∞ (i.e., the sequence of objective values is unbounded from below) or there is a subsequence lksuch that

lim

k→∞D(x

lk)∇ ˜L(xlk) = 0. (54)

Proof. We first estimate how much does the objective function improve at each iteration. By using (52) followed by a first-order Taylor expansion we get:

˜L(xk+1) = ˜L xk− λkD(xk)∇ ˜L(xk) + O  λ2 k  = ˜L(xk) − λ k∇ ˜L(xk)TD(xk)∇ ˜L(xk) + O  λ2 k  . (55)

(12)

Then we rewrite the above equality as the following inequality, holding for some M> 0, by definition of O(λ2):

˜L(xk+1)  ˜L(xk) − λ

k(∇ ˜L(xk)TD(xk)∇ ˜L(xk) − Mλk). (56)

Now suppose that∇ ˜L(xk)TD(xk)∇ ˜L(xk) is bounded from below by a positive β. Then, for k

large enough, the factor between parentheses in the second term on the right-hand side of (56) will be nonnegative, satisfying the nonnegativity hypotheses of lemma5.5. Therefore, if the sequence of objective values is bounded from below, it follows from lemma5.5that there is a subsequence lksuch that

lim k→∞{∇ ˜L(x lk)TD(xlk)∇ ˜L(xlk) + O(λ lk)} = 0. (57) Becauseλk→ 0+, we have lim k→∞∇ ˜L(x lk)TD(xlk)∇ ˜L(xlk) = 0. (58)

Consequently, for each j= 1, 2, . . . , n, lim

k→∞Dj(x lk)(∂

j˜L(xlk))2 = 0. (59)

By the boundeness assumption, we can assume that, with the same subsequence{lk},

lim

k→∞D(x

lk) = d, (60)

for some d∈ Rn×n. Hence, we obtain, for each j= 1, 2, . . . , n, lim

k→∞(Dj(x lk))2(∂

j˜L(xlk))2= dj j· 0 = 0. (61)

Therefore, for each j= 1, 2, . . . , n, lim

k→∞Dj(x lk)∂

j˜L(xlk) = 0, (62)

from which the conclusion follows. 

In particular, if it is known beforehand that ˜L is bounded from below, then we have shown

that the algorithm does provide, at least, a useful subsequence. This is likely to be the case, since in real optimization problems (like in tomographic reconstruction), one will not be able to obtain an unboundedly good solution. Using the above result with some further assumptions (which include strict convexity) we obtain the following global convergence result.

Corollary 5.7. Suppose that the assumptions of propositions5.3and5.2hold, assume also that ˜L is smoothly differentiable and strictly convex and that ˜L(xk) is well-defined and that

{xk} ⊂ Rn

+ is bounded. If ˜L(xk) converges, then any sequence generated by algorithm 5.1 converges to the solution of

min

x∈Rn

+ ˜L(x).

(63) Proof. If{xk} is bounded then we may assume, without loss of generality, that the subsequence

which we have proven to exist in proposition5.6is convergent, say xlk → x. If x∈ Rn

+and ˜L

is convex, then D(x) ˜L(x) = 0 is a necessary and sufficient condition for x∗to be an optimal solution of problem (63).

By continuity, because ˜L(xlk) converges, it must converge to ˜L(x). Therefore, the objective value at every accumulation point of{xk} equals ˜L(x). However, since we assumed that ˜L is strictly convex, this reasoning implies that all accumulation points of the bounded sequence {xk} are the same, namely x. Thus, we have proven that xk→ x. 

We now remove some of the hypotheses used in the last corollary by referring to works available in the literature. This leads to our main result, stated below.

(13)

Theorem 5.8. Assume that all components of x0 ∈ Rn

+ are positive, 0 < λk  λ for some suitably smallλ > 0,∞k=0λk= ∞ and



k=0λ2k< ∞. Then, if L is the negative Poisson log-likelihood and if ˜L is strictly convex, then any sequence generated by algorithm5.1converges to the solution of (63).

Proof. Boundedness of{xk}, for small enough λk, is a consequence of [3, proposition 1].

Convergence of ˜L(xk) may then be obtained by assuming that D(x)∇ ˜L(x) is Lipschitz continuous on the closure of the convex hull of the iterates and that∞k=0λ2k < ∞, as in

[32, lemma 3]. Furthermore, we see that, if λk  λ for some suitably small λ > 0, it is

possible to adapt the ideas leading to [37, corollary 1] to our case, yielding the conclusion that D(x) ˜L(x) is Lipschitz continuous within the closure of the convex hull of the iterates (and subiterates) of algorithm 5.1. Therefore, since ˜L is naturally bounded from below, we can apply

corollary5.7. 

6. Experimental work

This section is devoted to the experimental setup we have used in order to test the practical performance of SAEM algorithms and the results obtained. We will base our conclusions on two different figures of merit, the first is the squared error from the ground truth image, and the second being a total-variation based analysis, which puts on firmer grounds our claims about image quality. In what follows we provide a detailed description of the experimental setup. Subsection6.1will report the experimental results.

There are certain sources of pseudo-randomness in the numerical experiments we have performed, such as the choice of strings and the noise simulations, but the presented results are representative of the typical situation, as we have noticed little deviation from this behavior from one run to another of the experiments. Moreover, to enable any interested reader to reproduce our research, the full source code of the numerical algorithms used in this paper is available upon request.

(a) Data generation

In our simulated studies we used the modified Shepp–Logan head phantom [7] in order to compare the quality of images reconstructed by RAMLA and SAEM. Let us define the so-called Radon transformR[ f ] of a function f : R2→ R:

R[ f ](θ, t) :=

 R

f(t(cos θ, sin θ )T+ s(− sin θ, cos θ )T) ds. (64) Figure1 shows a schematic representation of the basic geometry of the Radon transform. The Shepp–Logan phantom is shown, together with its sinogram, i.e., its Radon transform presented as an image in theθ × t coordinate system.

We have used a parallel beam geometry where the samples were measured at the pairs(θ, t) in the set1, θ2, . . . θv}×{t1, t2, . . . , tr} where θi= π(i−1)/v and tj= −1+2( j−1)/(r−1)

for 1  i  v and 1  j  r. That is, we uniformly sample the Radon transform on the box [0, π ) × [−1, 1]. Data was generated as samples of a Poisson random variable having as parameter the exact Radon transform of the scaled phantom:

bi∼ Poisson(κR[ f ](θvi, tri)), 1  i  vr, (65) whereviis one plus the largest integer smaller than(i − 1)/r, ri= i − r(vi− 1), and where

(14)

t t R[f ](θ, t) θ −1 0 1 t 0 π2 π θ

Figure 1.Left: schematic representation of the Radon transform. In the definition,θ is

the angle between the normal to the integration path and the vertical axis, while t is the displacement from origin of the line of integration. Right: image of the Radon transform of the image shown on the left in theθ × t coordinate system.

will also denote the vector bas the one with ideal data: bi = κR[ f ](θvi, tri). In these studies, we have usedv = 288 and r = 256 in order to reconstruct images consisting of 256 × 256 pixels.

We also experiment using data from a human cardiac SPECT study. In this set of experiments, the data set hadv = 60 and r = 64 and we reconstructed images with 64 × 64 pixels. No attenuation correction has been applied to the data.

(b) String selection and weights

We denote by SAEM-T an SAEM algorithm with T strings, and the strings were built by randomly shuffling the data and then partitioning it in T contiguous chunks of uniform size, each of which would become a string. It was done like that for two reasons. First, it is often recognized that a random order performs as well, if not better, than most others of practical viability in sequential algorithms. Second, this avoids unfair comparisons that could arise from specific choices of subsets, which could possibly be more suitable to one EM variant than to others.

We wanted to keep the classical maximum likelihood model, so we used non-weighted averaging, that isωi = 1/T. More sophisticated weight selection schemes are conceivable,

but this subject is beyond the scope of the present paper.

(c) Scaling matrix

In the algorithms that we use, the scaling matrix D(x) is a diagonal matrix with nonzero elements given by xj/pjfor a fixed, but unspecified, set of weights pj > 0. We have chosen pj=

n

i=1ai j. Several other forms where possible, including some dependent on the particular

method being used, but we chose to keep the experiments understandable and to have a uniform treatment among the algorithms, rather than to complicate the results introducing another varying parameter.

(15)

RAMLA SAEM-4 SAEM-2 SAEM-5 SAEM-3 SAEM-6 0.00 3.57 7.15 10.72 14.29 5.55 6.27 6.99 7.72 8.44 LKL (x k)· 10 3 Time (seconds) RAMLA SAEM-4 SAEM-2 SAEM-5 SAEM-3 SAEM-6 0.00 3.58 7.17 10.75 14.34 2.79 3.10 3.41 3.72 4.03 LKL (x k)· 10 4 Time (seconds) RAMLA SAEM-4 SAEM-2 SAEM-5 SAEM-3 SAEM-6 0.00 3.59 7.19 10.78 14.38 2.09 2.27 2.45 2.63 2.82 LKL (x k)· 10 4 Time (seconds) RAMLA SAEM-4 SAEM-2 SAEM-5 SAEM-3 SAEM-6 0.00 3.59 7.19 10.78 14.37 2.23 2.32 2.42 2.51 2.61 LKL (x k)· 10 4 Time (seconds)

Figure 2. Convergence speed of several SAEM variations under different noise

condition. From top left in clockwise direction: 0.00%, 3.96%, 7.94% and 25.03% of relative noiseb− b/b. Notice that if up to T

maxstrings can be run in parallel, then

at any given time point, SAEM-T has similar log-likelihood values for T= 1, . . . , Tmax.

Running time was estimated by accumulating high-definition clock information of iteration computation, ignoring input/output time.

(d) Stepsize

We are required to determine the nonsummable stepsize sequence{λk} ⊂ R+. We use the rule,

for SAEM-T : λk= λT 0 k0.51/T + 1, (66) where λT

0 was selected as the largest value for which SAEM-T would not lead to negative

values in the first iteration. This technique for the starting parameter leads to the fastest possible convergence in terms of objective function decrease, and the scaling by the inverse of the number of strings accounts for the 1/T factor, caused by the averaging, in (33). Figure2

shows that this is indeed a well balanced parameter rule, because it leads all SAEM variations to have similar log-likelihood values at any time point, as long as all the strings can be run simultaneously in hardware. It can also be seen that this behavior remains consistent under several noise regimes. This same kind of plot is shown in the left-hand side of figure6for the, smaller, real data example, where the comparison is no longer as balanced, because parallelism overhead does not pay off with small images, but in this case computation time is negligible anyway.

(16)

RAMLA SAEM-4 SAEM-2 SAEM-5 SAEM-3 SAEM-6 2.79 3.10 3.41 3.72 4.03 0.08 0.10 0.12 0.14 0.16 LKL(xk) · 10−4 MS E (x k) RAMLA SAEM-4 SAEM-2 SAEM-5 SAEM-3 SAEM-6 2.79 3.10 3.41 3.72 4.03 0.43 0.68 0.92 1.17 1.42 LKL(xk) · 10−4 TV (x k)· 10 7

Figure 3.Relative squared error and total variation as functions of the log-likelihood for

iterations of SAEM variants. Notice that both the relative error and the total variation values, for a fixed likelihood level, are decreasing functions of the number of strings. The solid vertical lines show the log-likelihood level of the images of figure4.

(e) Starting image

Initial guess for the algorithms was a uniform image with an expected photon count equal to the data. That is, x0

j = α where α is such that m  i=0 (Ax0) i= m  i=0 bi. (67)

The value ofα can be easily obtained as α =mi=0bi/

m

i=0(A1)i, where 1 is the vector whose

components are all equal to 1.

6.1. Image reconstruction analysis

Figure 2 shows that all SAEM-T algorithms present similar convergence speeds when measured as the ratio of objective function decrease and computation time. Notice that at iteration k, SAEM-T has a smaller log-likelihood value than SAEM-(T + 1). Hence, more strings means slower algorithms iteration-wise because there is less incrementalism present. It is widely recognized, however, that log-likelihood cannot be used as an image quality measure because over-fitting the image to the data will amplify high frequency components of the noise. This leads to the need for other ways of assessing reconstruction quality.

In our simulated studies, we make use of the mean squared error, which requires full knowledge about the ideal sought-after image:

MSE(x) :=x − x2 2 x2 2 , (68)

where xis the noise-free discretization of the Shepp–Logan head phantom, properly scaled

byκ. Another functional we have used to measure image quality was the total variation [39]: TV(x) := r1  i=1 r2  j=1  (xi, j− xi, j−1)2+ (xi, j− xi−1, j)2, (69)

where r1r2 = n and we use the convention xi, j = xi(r2−1)+ j for i ∈ {1, 2, . . . , r1} and j ∈ {1, 2, . . . , r2} with the boundary condition x0,k = xk,0 = 0. Unlike the MSE, the total

(17)

RAMLA SAEM-6

Figure 4.Details of reconstruction obtained by RAMLA (left) and SAEM-6 (right).

Both images where these details taken from have, up to linear interpolation errors, the log-likelihood level indicated by the solid vertical lines in figure3. It is noticeable that the smaller total variance is reflected in both smoother large areas and better resolved sharp transitions. The horizontal solid colored lines show where the profiles of figure5

where taken from.

RAMLA SAEM-6 TRUE IMAGE 0 0.25 0.5 0.75 1 −1 −0.5 0 0.5 1

Figure 5.Profile lines from images in figure4. Notice how SAEM-6 reconstruction

presents less overshoot and more smoothness than RAMLA reconstruction.

variation (TV) alone cannot be used as a measure of image quality. However, given two reconstructed images of piecewise constant emission rates with the same Poisson likelihood, the one with the smaller total variation is more likely to have better image quality. Since

(18)

RAMLA SAEM-4 SAEM-2 SAEM-5 SAEM-3 SAEM-6 0.00 0.11 0.23 0.34 0.45 2.14 2.32 2.50 2.69 2.87 LKL (x k)· 10 3 Time (seconds) RAMLA SAEM-4 SAEM-2 SAEM-5 SAEM-3 SAEM-6 2.14 2.32 2.50 2.69 2.87 2.56 3.92 5.28 6.65 8.01 LKL(xk) · 10−3 TV (x k)· 10 4

Figure 6.Speed convergence and total variation results for the human SPECT study.

The solid line in the graphic at the right indicates the log-likelihood level of the images in figure7.

RAMLA SAEM-6

Figure 7.Male human cardiac SPECT study. The log-likelihood level of the images is

indicated by the solid lines in the graphic at the right in figure6. Data kindly provided by Roberto Isoardi—Fundaci´on Escuela de Medicina Nuclear (FUESMEN)—Mendoza, Argentina.

(19)

neither the log-likelihood nor the total variation require knowledge of the ideal image, the combination of both can be used also in the evaluation of the SPECT reconstruction.

Figure3shows plots of both figures of merit as functions of the log-likelihood for one of the simulated experiments. On the left we have MSE and on the right TV. The remarkable feature of these plots is the fact that the quality of the image obtained by SAEM-T , at any given fixed likelihood level is an increasing function of T , the number of strings. This behavior is unaltered as we vary the noise level. Notice the same effect in the TV versus LKLcurve shown

in the graphic at the right of figure6, and that the difference in the total variation translates into significant visual improvement in the real data study (figure7), even more than it does in the simulated case (figures4and5).

7. Conclusions

We have presented the new string-averaging expectation-maximization family of algorithms. The theoretical convergence properties of the method were studied and preliminary experimental evidence was provided of the technique’s suitability for good quality reconstruction.

The theory we have developed here can handle more general regularized objective functions, as they can achieve a better balance between smoothness and adherence to the data than traditional techniques currently used in tomographic scanners.

Future work could focus on comparative evaluation of SAEM against other paradigms, such as OS-EM or block-RAMLA, in light of the TV versus LKLplots, which we have shown

to be useful to assess image quality. It is potentially worth investigating a posteriori stopping criteria based on such plots, in the spirit of the L-curve [6].

Additional medical task oriented, and statistically based, experimental work is called for to asses the SAEM method’s potential in realistic situations.

Acknowledgments

We are sincerely grateful to the anonymous reviewers whose insightful comments encouraged us to extend some of the experimental work in the paper. This resulted in discovering further advantages of the proposed SAEM algorithm in the final version. Work by EH was partially supported by FAPESP grant 2013/16508-3, Brazil. YC was partially supported by the United States-Israel Binational Science Foundation (BSF) grant number 200912 and by US Department of Army award number W81XWH-10-1-0170. TB was partially supported by the National Science Council of the Republic of China, Taiwan (NSC 97-2118-M-214-001-MY2). IC was partially supported by the National Center for Theoretical Sciences (Taipei Office) and the National Science Council of the Republic of China (NSC 99-2115-M-002-003-MY3). AD was partially supported by CNPq grant no 301064/2009-1, Brazil. MJ was partially supported by the National Basic Research and Development Program of China (973 Program) (2011CB809105), National Science Foundation of China (61121002, 10990013, 60325101). HL was supported by grants from National Science Council, National Center for Theoretical Sciences, Center of Mathematical Modeling and Scientific Computing at National Chiao Tung University in Taiwan.

References

[1] Lange K 2004 Optimization (New York: Springer)

[2] McLachlan G J and Krishnan T 2008 The EM Algorithm and Extensions 2nd edn (Hoboken, NJ: Wiley-Interscience)

(20)

[3] Browne J and Pierro A R D 1996 A row-action alternative to the EM algorithm for maximizing likelihood in emission tomography IEEE Trans. Med. Imaging15 687–99

[4] Hudson H M and Larkin R S 1994 Accelerated image reconstruction using ordered subsets of projection data IEEE Trans. Med. Imaging13 601–9

[5] Leahy R and Byrne C 2000 Recent developments in iterative image reconstruction for PET and SPECT IEEE Trans. Med. Imaging19 257–60

[6] Hansen P C 1992 Analysis of discrete ill-posed problems by means of the L-curve SIAM

Rev.34 561–80

[7] Shepp L A and Vardi Y 1982 Maximum likelihood restoration for emission tomography IEEE

Trans. Med. Imaging1 113–22

[8] Lange K and Carson R 1984 EM reconstruction algorithms for emission and transmission tomography J. Comput. Assist. Tomogr. 8 306–16

[9] Vardi Y, Shepp L A and Kaufman L 1985 A statistical model for positron emission tomography

J. Am. Stat. Assoc.80 8–20(with discussion)

[10] Herman G T 2009 Fundamentals of Computerized Tomography: Image Reconstruction from

Projections 2nd edn (New York: Springer)

[11] Censor Y and Zenios S A 1997 Parallel Optimization: Theory, Algorithms and Applications (New York: Oxford University Press)

[12] Fessler J A 2000 Statistical image reconstruction methods for transmission tomography Handbook

of Medical Imaging ed M Sonka and J M Fitzpatrick (Bellingham: SPIE) chapter 1 pp 1–70

[13] Jiang M and Wang G 2002 Development of iterative algorithms for image reconstruction J. X-Ray

Sci. Technol. 10 77–86

[14] Bauschke H H and Borwein J M 1996 On projection algorithms for solving convex feasibility problems SIAM Rev.38 367–426

[15] Censor Y, Chen W, Combettes P L, Davidi R and Herman G T 2012 On the effectiveness of projection methods for convex feasibility problems with linear inequality constraints Comput.

Optim. Appl.51 1065–88

[16] Herman G T and Meyer L B 1993 Algebraic reconstruction techniques can be made computationally efficient IEEE Trans. Med. Imaging12 600–9

[17] Kaczmarz S 1937 Angen¨aherte Aufl¨osung von Systemen linearer Gleichungen Bull. Int. Acad. Pol.

Sci. Lett. A35 355–7

[18] Gordon R, Bender R and Herman G T 1970 Algebraic reconstruction techniques (ART) for three-dimensional electron microscopy and x-ray photography J. Theor. Biol.29 471–82

[19] Richardson W H 1972 Bayesin-based iterative method of image restoration J. Opt. Soc. Am. A62 55–59

[20] Lucy L B 1974 An iterative technique for the rectification of observed distribution Astron.

J.79 745–54

[21] Dempster A P, Laird N M and Rubin D B 1977 Maximum likelihood from incomplete data via the EM algorithm J. R. Stat. Soc. Ser. B 39 1–38

[22] Bertero M, Boccacci P, Desidera G and Vicidomini G 2009 Image deblurring with poisson data: from cells to galaxies Inverse Problems25 123006

[23] Stark H and Yang Y 1998 Vector Space Projections: A Numerical Approach to Signal and Image

Processing, Neural Nets, and Optics (New York: Wiley)

[24] Censor Y, Elfving T and Herman G T 2001 Averaging strings of sequential iterations for convex feasibility problems Inherently Parallel Algorithms in Feasibility and Optimization

and their Applications ed D Butnariu, Y Censor and S Reich (Amsterdam: North-Holland)

pp 101–13

[25] Zhu H, Zhou J, Shu H, Li S and Luo L 2004 Improved SAGE algorithm for PET image reconstruction using rescaled block-iterative method IEMBS’04: 26th Annu. Int. Conf. of the IEEE Engineering

in Medicine and Biology Society vol 1 pp 1353–6

[26] Byrne C 2005 Choosing parameters in block-iterative or ordered subset reconstruction algorithms

IEEE Trans. Image Process.14 321–7

[27] Censor Y and Elfving T 2002 Block-iterative algorithms with diagonally scaled oblique projections for the linear feasibility problem SIAM J. Matrix Anal. Appl.24 40–58

[28] Censor Y and Herman G T 2002 Block-iterative algorithms with underrelaxed Bregman projections

SIAM J. Optim.13 283–97

[29] Davidi R, Herman G T and Censor Y 2009 Perturbation-resilient block-iterative projection methods with application to image reconstruction from projections Int. Trans. Oper. Res.16 505–24

(21)

[30] Byrne C L 1998 Accelerating the EMML algorithm and related iterative algorithms by rescaled block-iterative methods IEEE Trans. Image Process.7 100–9

[31] Pierro A R De and Yamagishi M E B 2001 Fast EM-like methods for maximum ‘a posteriori’ estimates in emission tomography IEEE Trans. Med. Imaging20 280–8

[32] Ahn S and Fessler J A 2003 Globally convergent image reconstruction for emission tomography using relaxed ordered subset s algorithms IEEE Trans. Med. Imaging22 613–26

[33] Hsiao I T, Rangarajan A, Khurd P and Gindi G 2004 An accelerated convergent ordered subsets algorithm for emission tomography Phys. Med. Biol.49 2145–56

[34] Censor Y 1981 Row-action methods for huge and sparse systems and their applications SIAM

Rev.23 444–66

[35] Jiang M and Wang G 2003 Convergence studies on iterative algorithms for image reconstruction

IEEE Trans. Med. Imaging22 569–79

[36] Tanaka E and Kudo H 2003 Subset-dependent relaxation in block-iterative algorithms for image reconstruction in emission tomography Phys. Med. Biol.48 1405–22

[37] Helou E S and Pierro A R De 2005 Convergence results for scaled gradient algorithms in positron emission tomography Inverse Problems21 1905–14

[38] Helou E S and Pierro A R De 2010 Incremental subgradients for constrained convex optimization: a unified framework and new methods SIAM J. Optim.20 1547–72

[39] Rudin L I, Osher S and Fatemi E 1992 Nonlinear total variation based noise removal algorithms

Physica D60 259–68

[40] Wang G and Jiang M 2004 Ordered-subset simultaneous algebraic reconstruction techniques (OS-SART) J. X-Ray Sci. Technol. 12 169–77

[41] Tanaka E and Kudo H 2010 Optimal relaxation parameters of DRAMA (dynamic RAMLA) aiming at one-pass image reconstruction for 3D-PET Phys. Med. Biol.55 2917–39

[42] Censor Y and Tom E 2003 Convergence of string-averaging projection schemes for inconsistent convex feasibility problems Optim. Methods Softw.18 543–54

[43] Penfold S N, Schulte R W, Censor Y, Bashkirov V, McAllister S, Schubert K E and Rosenfeld A 2010 Block-iterative and string-averaging projection algorithms in proton computed tomography image reconstruction Biomedical Mathematics: Promising Directions in Imaging, Therapy

Planning and Inverse Problems ed Y Censor, M Jiang and G Wang (Madison, WI: Medical

Physics Publishing) pp 347–67

[44] Trummer M R 1981 Reconstructing pictures from projections: on the convergence of the ART algorithm with relaxation Computing26 189–95

數據

Figure 1. Left: schematic representation of the Radon transform. In the definition, θ is
Figure 2. Convergence speed of several SAEM variations under different noise
Figure 3. Relative squared error and total variation as functions of the log-likelihood for
Figure 4. Details of reconstruction obtained by RAMLA (left) and SAEM-6 (right).
+2

參考文獻

相關文件

Keywords: accuracy measure; bootstrap; case-control; cross-validation; missing data; M -phase; pseudo least squares; pseudo maximum likelihood estimator; receiver

We are not aware of any existing methods for identifying constant parameters or covariates in the parametric component of a semiparametric model, although there exists an

In particular, we present a linear-time algorithm for the k-tuple total domination problem for graphs in which each block is a clique, a cycle or a complete bipartite graph,

Then, it is easy to see that there are 9 problems for which the iterative numbers of the algorithm using ψ α,θ,p in the case of θ = 1 and p = 3 are less than the one of the

We further want to be able to embed our KK GUTs in string theory, as higher dimensional gauge theories are highly non-renormalisable.. This works beautifully in the heterotic

It should be stressed that the four eigenvalues obtained here do not change even if we include other field outside KBc subalgebra or outside the dressed B 0 gauge, since such fields

„ „ The The extended nature extended nature of string theory introduces of string theory introduces additional degrees of freedom?. additional degrees of freedom localized

S15 Expectation value of the total spin-squared operator h ˆ S 2 i for the ground state of cationic n-PP as a function of the chain length, calculated using KS-DFT with various