• 沒有找到結果。

Analysis of Switching Dynamics with Competing Support Vector Machines

N/A
N/A
Protected

Academic year: 2021

Share "Analysis of Switching Dynamics with Competing Support Vector Machines"

Copied!
16
0
0

加載中.... (立即查看全文)

全文

(1)

Analysis of Switching Dynamics with

Competing Support Vector Machines

Ming-Wei Chang and Chih-Jen Lin

Department of Computer Science and

Information Engineering

National Taiwan University

Taipei 106, Taiwan (

cjlin@csie.ntu.edu.tw

)

Ruby C. Weng

Department of Statistics

National Chenechi University

Taipei 116, Taiwan

Abstract

We present a framework for the unsupervised segmentation of switching dynamics using support vector machines. Following the architecture by Pawelzik et al. [21] where annealed competing neural networks were used to segment a non-stationary time series, in this article we exploit the use of support vector machines, a well-known learning technique. First, a new formulation of support vector regression is proposed. Second, an expectation-maximization (EM) step is suggested to adaptively adjust the annealing parameter. Results indicate that the proposed approach is promising.

Index Terms

Annealing, competing experts, expectation maximization, support vector machines, unsupervised time series segmentation.

I. INTRODUCTION

Recently support vector machines (SVMs) [26] have been a promising method for data classification and regression. For an account of various applications of the method, see [15], [16], and references therein. However, its application to unsupervised learning problems has not been exploited much. In this paper we aim to apply it to unsupervised segmentation of time series. Practical applications of unsupervised segmentation of time series include, for example, speech recognition [23], signal classification [6], and brain data [20], [9].

(2)

The topic of unsupervised segmentation has been investigated by researchers in various fields. See for example, an early survey on various approaches in [18], the fuzzy c-regression models (FCRM) [5] with applications to models with known forms and known distribution of the noise term, a combination of supervised and unsupervised learning using hidden Markov models based on neural networks [4], and other competing neural networks approaches in [14], [21], [8], [11], [9].

In [21], [14], annealed competing neural networks were used to segment a non-stationary time series, where non-stationarities are caused by switching dynamics. This method is called “Annealed Competition of Experts” (ACE). Unlike the mixtures of experts architecture [7] which used an input-dependent gating-network, the ACE method drives the competition of experts by an evaluation of prediction performance so that the underlying dynamics can have overlapping input domains. Two main features of this approach are memory derived from a slow switching rate and deterministically annealed competition of the experts during training. The assumption of slow switching rate is imposed to resolve problems caused by overlapping input-output relations. The idea of annealing is to avoid getting stuck in local minima and resolve the underlying dynamics in a hierarchical manner. (The deterministic annealing method was described in the context of clustering [24].) The neural network used is a Radial Basis Function (RBF) network of the Moody-Darken type [13].

The present paper aims to solve the same unsupervised segmentation problem as in [21], [14]. We propose a framework using competing support vector machines (SVMs). The standard SVMs assume equal weights on all error terms, which mean that each data point is equally important. However, due to the switching nature of the problem, the data points actually come from different sources and therefore, the contribution of each data point to each predictor would not be the same. In order to solve this problem, we propose a modified formulation of SVMs which allows different weights on the error terms. The dual problem and the implementation for this formulation are also new. Here the weights are adjusted by relative prediction performance, as in [21], [14]. In addition to the new formulation, this paper is novel in presenting an adaptive annealing method. A key observation is that the annealing parameter characterizes some statistical property of the error terms. Therefore, at each iteration during training, we treat the annealing parameter as an unknown parameter and estimate it based on the current weighting coefficients and error terms. This estimate is essentially the maximum likelihood estimate, which can be obtained by an expectation-maximization (EM) step. We shall call this annealing method adaptive deterministic annealing.

The paper is organized as follows. In Section II, we briefly describe the framework of [21], [14] and present a modified SVM formulation. In Section III, we give motivation behind the proposed annealing method and derive an estimate of the annealing parameter. The implementation of the modified SVM formulation is in Section IV. Section V demonstrates experimental results on some data sets. Then we give discussion in Section VI.

(3)

II. A MODIFIEDSVM FORMULATION

First, we describe the segmentation problem using input-output pairs and outline the approach of [21]. Then we review the standard SVMs and introduce a modified formulation.

Let (xt, yt) = (xt, frt(xt)) be generated by m unknown functions, where t = 1, . . . , l, rt ∈ {1, . . . , m}, and rt = j if the t-th data point is generated by fj. If we set xt+1= frt(xt), we get a time series {xt}. The task is to determine rt, t = 1, . . . , l, and functions fj, j = 1, . . . , m. For

notational simplicity, we take yt= frt(xt) in this section. The readers should notice that the method presented below does not restrict to time series with embedding order 1. The extension to higher embedding order can be done by simply replacing the scalar xt by vectors from the method of the

time delay embedding of the time series [10], [21].

Define pti = 1 if rt = i and 0 otherwise. Then pti, fi, i = 1, . . . , m, t = 1, . . . , l is an optimal

solution of the following non-convex optimization problem:

min p,f l X t=1 m X i=1 pti(yt− fi(xt))2 subject to m X i=1 pti = 1, p t i ≥ 0, t = 1, . . . , l. (1)

The main computations for solving (1) can be divided into two steps: (i) minimize the objective function in (1) over fi with pti fixed; (ii) adjust pti given current fi. In [21], the authors use RBF

networks for step(i); i.e. for i = 1, . . . , m, they solve

min ˆ fi l X t=1 pti(yt− ˆfi(xt))2, (2)

where ˆfiare RBF networks. For step(ii), they assume that ˆfi(xt) are distributed according to Gaussian

and then by Bayes rules

pti= exp (−β(eti)2) Pm j=1exp (−β(e t j)2) , (3) where eti= yt− ˆfi(xt) (4)

and β is the annealing parameter which controls the degree of competition among predictors. To incorporate the property of low switching rates, the sequence ((xt−∆, yt−∆), . . . , (xt+∆, yt+∆)) are

assumed to come from the same source. Using Bayes rules again, the weighting coefficients pti can be updated as pti = exp (−β P∆ δ=−∆(e t−δ i ) 2) Pm j=1exp (−β P∆ δ=−∆(e t−δ j )2) . (5)

We will further discuss β in Section III and the choice of ∆ in Section V. We summarize the algorithm in [21] as follows:

(4)

Algorithm II.1

1) Given ∆ and parameters of RBF networks. Randomly pick initial pt

iwhich satisfies m

X

i=1

pti = 1; pti≥ 0, t = 1, . . . , l. 2) Solve (2) using the RBF networks and obtain ˆfi.

3) Increase β slightly and update pt iby(5).

4) If certain stopping criteria are met, terminate the algorithm. Otherwise, go to step 2.

Next we propose to modify support vector regression (SVR) for solving (2). The idea of SVR is to map input data into a higher dimensional space and find a function which approximates the hidden relationships of the given data. The standard formulation is as follows.

min w,b,ξ,ξ∗ 1 2w Tw + C l X t=1 (ξt+ ξt,∗) subject to −  − ξt,∗≤ yt− (wTφ(xt) + b) ≤  + ξt, (6) ξt≥ 0, ξt,∗≥ 0, t = 1, . . . , l,

where w and b are unknown parameters to be estimated, ξ and ξ∗indicate errors,  is the tolerance, and φ is the function that maps data xtto a higher dimensional space. SVR uses the so called “regularization

term”12wTw in (6) to avoid overfitting. Without this term, for certain mapping functions φ, the solution

of (6) always satisfies − ≤ yt− (wTφ(xt) + b) ≤ , see, for example, [2, Corollary1]. Thus it overfits

the data. If we take

f (x) = wTφ(x) + b, then (6) can be rewritten as

min w,b 1 2w Tw + C l X t=1 |yt− f (xt)| (7)

where | · | is the -insensitive loss function.

The standard SVR considers a uniform penalty parameter C for all data. By comparing (2) and (7), we propose to use different weights Cpti in each error term of the objective function. Thus for each i = 1, . . . , m, we have the following modification.

min wi,bi,ξ,ξ∗ 1 2w T iwi+ C l X t=1 pti(ξti+ ξ t,∗ i ) subject to −  − ξit,∗≤ yt− (wiTφ(xt) + bi) ≤  + ξti, (8) ξit≥ 0, ξt,∗i ≥ 0, t = 1, . . . , l. Note that if we remove the term 12wT

i wi and set  = 0, then (8) is equivalent to (2) with `2-norm

(5)

SVR. Without it, we will overfit the data because in this case − ≤ yt− (wiTφ(xt) + bi) ≤  for all

t with pti> 0.

The stopping criterion of our algorithm is as follows | ˆobj − obj|

|obj| ≤ 0.05, (9)

where ˆobj and obj are the objective values of two consecutive iterations. Here the objective function is defined as l X t=1 m X i=1 pti|yt− fi(xt)|.

A summary of our changes from Algorithm II.1 is in the following:

Algorithm II.2

1’ Given ∆, SVR parameters C and , and SVR mapping function φ, solve (8) and obtain ˆfi.

3’ Update β by the method described in Section III. If (9) is satisfied, stop the algorithm.

For an account on the relation between RBF networks and SVRs with RBF kernels, see [3], [25] and references therein.

III. ADJUSTMENT OFβ USINGMAXIMUMLIKELIHOODESTIMATION

Recall that the annealing parameter β in (3) and (5) controls the degree of competition. If β is large, pt

i≈ 1 for i = argminj

P∆

δ=−∆(e t−δ

j )2 and 0 otherwise. This is the so-called hard competition

(winner-takes-all). If one uses hard competition right from the beginning, it is very likely to get stuck in an undesired local minimum. In [21], this problem was resolved by using deterministic annealing.

In this section we describe an adaptive method for adjusting β. Let et

i, defined in (4), denote

the difference between the observed value and the estimated value. From (3), β can be viewed as a parameter that characterizes certain statistical property about et

i. Therefore, at each iteration we suggest

to update β by maximizing the conditional likelihood function given current ˆfi. To do so, we assume

that, given current ˆfi, yt follows a mixture of Gaussian distributions pi with mean ˆfi(xt), a common

unknown variance τ , and mixing coefficients ai. Then, the density of yt given ˆfi is

p(yt| ˆfi, τ ) = m X i=1 aipi(yt| ˆfi, τ ) = m X i=1 ai √ 2πτ exp{−(e t i) 2/(2τ )}. (10)

Let ˆτ be an estimate of τ . Then using (10) we can estimate pt

i by Bayes rules ˆ pti ≡ p(rt= i| ˆf , yt, ˆτ ) = p(yt, rt= i| ˆf , ˆτ ) Pm k=1p(yt, rt= k| ˆf , ˆτ ) = aiexp{−(e t i)2/(2ˆτ )} Pm k=1akexp{−(etk)2/(2ˆτ )} . (11)

(6)

By comparing (5) and (11), we suggest to choose 1/(2ˆτ ) as our next β. Since ˆτ is the measure of the variation of eti, it is intuitively clear that the next ˆτ will decrease if ˆfi in the next iteration can better

fit the data. So the new β is likely to increase, which corresponds to the annealing property used in [24], [21].

We propose to estimate τ in each iteration by the maximum likelihood principle. This can be obtained by Expectation Maximization method. To see how, let rtbe as in Section II, which is a latent

variable (unobservable). We augment yt with rt and call {(yt, rt) : t = 1, . . . , l} the complete data.

Then the complete-data log-likelihood function of τ is

L(τ ) = l X t=1 logp(yt, rt| ˆfi, τ ) = l X t=1 logartprt(yt| ˆfi, τ ). (12) Let τ(g) and pt(g)

i ≡ p(rt = i| ˆf , yt, τ(g)) be current estimates. Let Y ≡ (y1, . . . , yl) and R ≡

(r1, . . . , rl). The E-step calculates the expectation of the conditional log-likelihood function of (Y, R)

given Y and τ(g): Q(τ, τ(g)) (13) ≡ E[logp(Y, R| ˆf , τ )| ˆf , Y, τ(g)] = l X t=1 E[log p(yt, rt| ˆf , τ )| ˆf , Y, τ(g)] = l X t=1 m X i=1 log(aipi(yt| ˆfi, τ ))p t(g) i = l X t=1 m X i=1 {(log ai)p t(g) i } + l X t=1 m X i=1 {[log pi(yt| ˆf , τ )]p t(g) i }, (14)

where the second equality follows from the independence of each observation and the third equality follows from (12) and the definition of pt(g)i .

The M-step finds the maximizer of Q:

τ(g+1)= arg maxτQ(τ, τ g

). (15)

By (14) and Gaussian assumption, we find that the maximum of (14) occurs at

τ(g+1) = Pl t=1 Pm i=1{(e t i)2p t(g) i } Pl t=1 Pm i=1p t(g) i = Pl t=1 Pm i=1{(e t i)2p t(g) i } l , (16)

where pt(g)i is obtained by replacing ˆτ in (11) by τ(g).

To apply the above method, the Gaussian assumption in (10) is not necessary. For example, if we replace the mixture of Gaussians by a mixture of Laplace (double exponential) distributions with a

(7)

common scale parameter τ , i.e. pi(yt| ˆfi, τ ) ∝ τ−1exp(−|yt− ˆfi|/τ ), the expectation and maximization

steps can be carried out similarly. The resulting estimate of τ is similar to (16), but with (eti)2replaced by |eti|. Because here a linear loss function is used for SVR, we suggest to use |eti| instead of (eti)2

when updating pti. Therefore, at the (g + 1)st iteration of our implementation, we update pti by (5)

with β replaced by 1/τ(g+1) and (et

i)2 replaced by |eti|. Note that we use (5) because the parameter

∆ incorporates the slowly switching property [21] while ai are less important in the implementation.

IV. SOLVING THEMODIFIEDSUPPORTVECTORREGRESSION

In this section we discuss how to solve the new SVR formulation (8). As SVRs map data into higher dimensional spaces, w is a long vector variable. Thus, usually a dual problem is easier to be solved. The Lagrangian dual of (8) can be derived by a similar way as that of standard SVR so here we give only the resulting formulation:

min α,α∗ 1 2(α − α ∗)TK(α − α) +  l X t=1 (αt+ α∗t) + l X t=1 yt(αt− α∗t) subject to l X t=1 (αt− α∗t) = 0, (17) 0 ≤ αt, α∗t ≤ Cp t i, t = 1, . . . , l,

where K is a square matrix with Kt,o = K(xt, xo) ≡ φ(xt)Tφ(x0). Here we mainly consider the

RBF kernel K(xi, xj) = e−γ||xi−xj||

2

. The main difference is that each dual variable has its own upper bound Cpt

i. At the optimal solution wi=Plt=1(αt− α∗t)φ(xt). So

ˆ fi(x) = l X t=1 (αt− α∗t)K(xt, x).

If αt− α∗t 6= 0, then xtis a support vector of the modified SVR (17). The main difficulty on solving

(17) is that K is a large dense matrix. This issue has occurred for the case of classification and some methods such as the decomposition method (e.g. [19]) have been proposed. The decomposition method avoids the memory problem by iteratively working on few variables. The extreme is the Sequential Minimal Optimization (SMO) [22] where in each iteration only two variables are updated. Here we consider the SMO-type implementation inLIBSVM[1]. Originally inLIBSVMthe two variables are selected by the following rule: The first element is the αt (or αt∗) which has the smallest value of the

following numbers:

−(K(α − α∗))t−  − yt, if αt> 0 and

−(K(α − α∗))t+  − yt, if α∗t < C.

The second element is the αt(or αt∗) which has the largest value of

−(K(α − α∗))t−  − yt, if αt< C and

(8)

They are derived from the optimality condition of the dual of (7). We change C in the above formula to Cpti while the convergence of the algorithm still holds. The modified code ofLIBSVMis available upon request.

V. EXPERIMENTS

A. Four Chaotic Time Series

We test the case of completely overlapping input manifold used in [21]. For all (xt, yt), yt= frt(xt). They consider x ∈ [0, 1] and four different functions: f1(x) = 4x(1 − x), f2(x) = 2x if x ∈ [0, 0.5)

and 2(1 − x) if x ∈ [0.5, 1], f3(x) = f1(f1(x)), and f4 = f2(f2(x)). It is easily seen that all the fi

map x from [0, 1] to [0, 1]. They set xt+1= yt= frt(xt) and activate the four functions consecutively, each for 100 time steps, giving an overall 400 time steps. They then repeat the procedure three times and obtain a time series of 1,200 points. An illustration of these functions is in Figure 1. As the number of functions is considered unknown, we start from six competing SVMs. In the end some SVMs represent the same function and the six SVMs represent four different functions. If we use only four competing SVMs in the beginning, we find it is easier to fall into local minima than using six SVRs.

For SVR parameters, we simply set C = 1 and the width of the insensitive tube  to be 0.03 (actually, we may set  to be any nonnegative small number, say 0 to 0.05). We set the parameter γ of the RBF kernel to be 50. Since a smaller γ means greater smoothness of predictors, in order to avoid getting trapped in a local minimum, we may start with a smaller γ and gradually increase it during training. Therefore, we also try to adaptively adjust γ during training by setting γ(g)= 1/(2τ(g)), where τ(g)

is as in (16). It works well too. The choice of ∆ is the same as that in [8]: as the small switching rate 1/100 guarantees a large probability for short sequences of length 7 to contain no switching event, a choice of ∆ = 3 would work. We also find that this parameter is not crucial during training. For this case the algorithm stops in about six iterations and the data points are well separated. See Figure 1 for the first four iterations. Here the initial pt

i are chosen to be around 1/6. So in the beginning,

the six predictors are about the same. As the iteration goes, the underlying structure resolves in a hierarchical manner. We then add noise 0.1N (0, 0.5) to these four functions. The algorithm stops in seven iterations. In Figure 2 we present the first six iterations. For this case we randomly assign the initial pti to be 0 or 1 under the condition

P6

i=1p t

i = 1, t = 1, . . . , 1200. By comparing the first

graphs in Figures 1 and 2, we find that if the initial pt

i are chosen to be around 1/6, the six predictors

(9)

Fig. 1. First four iterations (data without noise)

B. Mackey-Glass chaotic system

Next we consider a high-dimensional chaotic time series obtained from the Mackey-Glass delay-differential equation [12]: dx(t) dt = −0.1x(t) + 0.2x(t − td) 1 + x(t − td)10 .

Following earlier experiments, points are selected every six time steps. We generate 300 points with delay parameter td = 23. Then we switch td be to 17, 23, and 30, and generate also 300 points

for each of the td. Totally there are 1,200 points from three underlying dynamics. The embedding

dimension is d = 6. That is, ytis the one-step ahead value of xt= (yt−1, . . . , yt−6). For this problem

we set the γ = 1 and leave the other parameters the same as in the previous example. As can be seen in Figure 3, the underlying dynamics are well segmented.

C. Santa Fe Time Series: Data Set D

We also consider the benchmark Data Set D from the Santa Fe Time Series Prediction Competition [27]. This is an artificial data generated from a nine-dimensional periodically driven system with an asymmetrical four-well potential and a drift on the parameters. The system operates in one well for

(10)

Fig. 2. First six iterations (data with noise)

some time and then switch to another well with a different dynamical behavior. Each participant is asked to predict the first 25 steps and the evaluation is based on the root mean squared error (RMSE) of the prediction. The RMSE obtained by the winner of the Santa Fe Competition, Zhang and Hutchinson [27, pp219-241], is 0.0665.

One may train a model on the whole dataset and then conduct predictions. But if data are actually from different sources, it would be better to do segmentation before prediction. [21] proposes to use the “Annealed Competition of Experts” (ACE) method to segment the time series into regimes of

(11)

Fig. 3. pt

i, i = 1, . . . , 3 at each time point t (x-axis: time; y-axis: pti)

approximately stationary dynamics by using six RBF networks. Then the prediction is simply done by iterating the particular predictor that is responsible for the generation of the latest training data. Similar to [21], in [17] they uses ACE with six RBF networks to segment the time series. Instead of RBF networks, they then use SVR to train the particular class of data that includes the data points at the end of the full training set. They also use SVR to train the full dataset and compare its predicting performance with that obtained by using segmentation.

Here we start with the standard SVR (6) and train a model on the whole dataset. We set the embedding dimension as 20 and the kernel parameter γ = 1.5 as in [17]. The SVR parameter  is set to be 0 and C is determined by five fold cross validation on {2−4, 2−3, . . . , 25}. By five fold cross validation we mean that after data are transformed into (xt, yt), they are randomly separated to five

groups. Sequentially one group is validated by training the rest. We then use this model to predict the first 25 steps. Unfortunately, the C value with the smallest validation error does not always produce the best prediction. For example, a typical run shows that: the smallest validation error 0.0351 occurs at C = 23 and the prediction error is 0.0748; the smallest prediction error 0.0521 is attained at C = 24,

but it has a larger validation error 0.0357. As the result of a single run is not always representative, we decide to investigate the long-run performance of this method by repeating the same procedure 30 times with different cross validations chosen randomly. The distribution of the selected C values is shown in Figure 4 (the light bars). In average, the RMSE of the 25-step prediction is 0.0871.

Next we use Algorithm II.1 with six competing SVRs to segment the time series. The parameters in the modified SVR (8) are set as γ = 0.05 (as in [17]),  = 0, and C = 0.01. Then we evaluate the pti values for each predictor at the last 25 points of the training data. The predictor with the

(12)

Fig. 4. The Distribution of the Selected C Values

largest pti for the majority of the 25 points is chosen. After doing so, we use the standard SVR (6)

to train this particular class of data. Here we set  = 0 and select C by five fold cross validation on {2−4, 2−3, . . . , 25}. The resulting model is used to predict the first 25 steps of the testing data.

To study the long-run performance of the method we repeat this procedure 30 times. The average of the RMSEs of the prediction errors is 0.0656. The distribution of the best C selected is given in Figure 4 (the dark bars). Of the 30 experiments, 15 of them outperform the result obtained by the winner. Due to different initial pti, the number of data points chosen for after the segmentation varies. For our 30 runs, this number ranges from 400 to 1,100. The best prediction performances (RMSE = 0.038 to 0.042) occur at numbers around 800 to 1,000. Our finding is that the prediction result using segmentation is better than that using the whole dataset directly.

As mentioned in [17], determining the SVR parameters (C, ) is computationally intensive. They suggest to determine them at the minimum of the one step prediction error measured on a randomly chosen validation set. To have a more stable parameter selection, here we use five fold cross validation to select C. Moreover, instead of setting  = 0, we also tried positive , but the results are not better. It is desirable to compare our parameter settings with those in [17]. However, they did not indicate their final choice of the parameters so we cannot make the comparison.

VI. DISCUSSION

In this paper we present a framework for the unsupervised segmentation of non-stationary time series using SVMs. The problems we consider here are the same as those in [21], [14]. The method used in this paper is novel in two aspects. First, a new formulation of SVM and its implementation are proposed for the switching problems. Second, we use statistical reasoning to adjust the annealing parameter. The annealing property of the proposed method is given intuitively in Section II. From the

(13)

figures we obtained in Section V, our results are comparable to those in [21]. Moreover, as we adjust β adaptively, our algorithm requires much fewer iterations than theirs.

Several parameters in our algorithm must be chosen in the beginning. For the mapping function φ, we use the RBF kernel K(xi, xj) = e−γ||xi−xj||

2

, which is the most popular choice. The selection of C, the width of the insensitive tube  in (8), and the kernel parameter γ is usually by try and error. For the two chaotic time series examples in Section V, we can easily select suitable C, , and γ, see discussions in that section. However, for the prediction of Data Set D from the Santa Fe Time Series Competition, it takes quite a long time to tune the parameter.

For a slowly changing dynamics, [21] is the first paper that proposes to use the parameter ∆ for updating pti. Intuitively, this method would work better for smaller switching rates. The switching rates of the first two examples in Section V are set to be 1/100 and 1/300 respectively, the same as those in [21]. For the chaotic time series example in Section V-A, an unreported experiment shows that our algorithm also works very well for a switching rate as high as 1/5.

As to the computational time, the major cost in each iteration of the proposed approach is on solving m different SVM problems. It is known that for most SVM implementations fewer support vectors (i.e. αt− α∗t 6= 0, where (αt, α∗t) is optimal for (8)) leads to less training time. We also consider

free support vectors (i.e. −Cpt

i < αt− α∗t < 0 or 0 < αt− αt∗ < Cpti) because for some SVM

implementations they are more related to the training time than the number of support vectors. The reason is that these implementations are iterative procedures and in final iterations mainly those free support vectors are still considered. More discussions can be found in [1]. Figure 5 presents the number of support vectors and the number of free support vectors at each iteration for the chaotic time series example in Section V-A. Here the data is generated without adding the noise. To interpret, in the first iteration the average number of support vectors of the six SVMs is 133, the average number of free support vectors is 11; in the second iteration the average number of support vectors is 701, that of free support vectors is 26, and so on.

Remember that there are 1,200 training data points and six competing SVRs. In this section, we randomly assign the initial pt

i to be 0 or 1 under the condition

P6

i=1p t

i = 1, t = 1, . . . , 1200.

Therefore, at the first iteration, approximately 200 data points are assigned to each predictor and the average number of support vectors of the six SVRs are less than 200. After updating pt

i, for all t with

pti > 0 the corresponding data points are assigned to the i-th SVR. So the number of support vectors substantially increased at the second iteration. Then as the underlying structure gradually resolved, the number of support vectors reduces again. An important trick we used here is to set all pti ≤ 0.01 to be zero. This effectively decreases the number of support vectors.

As for the numbers of free support vectors, they are all small compared to support vectors. At an optimal solution of an SVM, some bounded variables have small values (i.e. small corresponding pt

(14)

Fig. 5. Number of support vectors (i.e. αt− α∗t 6= 0) and number of free support vectors (i.e. −Cpti< αt− α ∗ t < 0 or

0 < αt− α∗t < Cpti)

but some are large (i.e. corresponding pti close to 1). For variables with small bounds, our current implementation,LIBSVM, can quickly identify them, so the training time of the first iteration is similar to that of the second iteration (less than two seconds on a Pentium III-500 machine). In other words, though the number of support vectors is dramatically increased, at the second iteration, many pt

i are

small soLIBSVMquickly identifies bounded variables. Then the decomposition method mainly works on free variables, so the difference between the two iterations is not that much.

ACKNOWLEDGEMENTS

This work was supported in part by the National Science Council of Taiwan via the grants NSC 90-2213-E-002-111 and NSC 91-2118-M-004-003. The second author thanks Danil Prokhorov and other organizers of IJCNN Challenge 2001 for bringing him to this subject. The authors also thank Jens Kohlmorgen for some helpful comments.

REFERENCES

[1] C.-C. Chang and C.-J. Lin. LIBSVM: a library for support vector machines, 2001. Software available at

http://www.csie.ntu.edu.tw/˜cjlin/libsvm.

[2] C.-C. Chang and C.-J. Lin. Training ν-support vector classifiers: Theory and algorithms. Neural Computation, 13(9):2119– 2147, 2001.

[3] T. Evgeniou, M. Pontil, and T. Poggio. Regularization networks and support vector machines. Advances in Computational Mathematics, 13:1–50, 2000.

[4] L. A. Feldkamp, T. M. Feldkamp, and D. V. Prokhorov. An approach to adaptive classification. In S. Haykin and B. Kosko, editors, Intelligent signal processing. IEEE Press, 2001.

[5] R. J. Hathaway and J. C. Bezdek. Switching regression models and fuzzy clustering. IEEE Transactions on Fuzzy Systems, 1(3):195–204, 1993.

(15)

[6] S. Haykin and D. Cong. Classification of radar clutter using neural networks. IEEE Transactions on Neural Networks, 2:589–600, 1991.

[7] R. A. Jacobs, M. I. Jordan, S. J. Nowlan, and G. E. Hinton. Adaptive mixtures of local experts. Neural Computation, 3(1):79–87, 1991.

[8] A. Kehagias and V. Petridis. Time-series segmentation using predictive modular neural networks. Neural Computation, 9:1691–1709, 1997.

[9] J. Kohlmorgen, K.-R. M¨uller, J. Rittweger, and K. Pawelzik. Identification of nonstationary dynamics in physiological recordings. Biological Cybernetics, 83:73–84, 2000.

[10] W. Liebert, K. Pawelzik, and H. G. Schuster. Optimal embeddings of chaotic attractors from topological considerations. Europhys. Lett., 14:521, 1991.

[11] S. Liehr, K. Pawelzik, J. Kohlmorgen, and K. R. M¨uller. Hidden markov mixtures of experts with an application to EEG recordings from sleep. Theory in Biosci., 118(3-4):246–260, 1999.

[12] M. C. Mackey and L. Glass. Oscillation and chaos in physiological control systems. Science, 197:287–289, 1977. [13] J. Moody and C. J. Darken. Fast learning in networks of locally–tuned processing units. Neural Computation, 1:281–293,

1989.

[14] K.-R. M¨uller, J. Kohlmorgen, and K. Pawelzik. Analysis of switching dynamics with competing neural networks. IEICE Transactions on Fundamentals of Electronics, Communications and Computer Sciences, E78–A(10):1306–1315, 1995. [15] K.-R. M¨uller, S. Mika, G. R¨atsch, K. Tsuda, and B. Sch¨olkopf. An introduction to kernel-based learning algorithms. IEEE

Transactions on Neural Networks, 12(2):181–201, 2001.

[16] K.-R. M¨uller, A. Smola, G. R¨atsch, B. Sch¨olkopf, J. Kohlmorgen, and V. Vapnik. Predicting time series with support vector machines. In W. Gerstner, A. Germond, M. Hasler, and J.-D. Nicoud, editors, Proceedings of ICANN ´97: : Proc. of the Int. Conf. on Artificial Neural Networks, pages 999–1004, Berlin, 1997. Springer.

[17] K.-R. M¨uller, A. Smola, G. R¨atsch, B. Sch¨olkopf, J. Kohlmorgen, and V. Vapnik. Predicting time series with support vector machines. In B. Sch¨olkopf, C. J. C. Burges, and A. J. Smola, editors, Advances in Kernel Methods - Support Vector Learning, pages 243–254, Cambridge, MA, 1999. MIT Press.

[18] R. Murray-Smith and T. A. J. (Eds.). Multiple Model Approaches to Modelling and Control. Taylor and Francis, London, 1997.

[19] E. Osuna, R. Freund, and F. Girosi. Training support vector machines: An application to face detection. In Proceedings of CVPR’97, pages 130–136, New York, NY, 1997. IEEE.

[20] K. Pawelzik. Detecting coherence in neuronal data. In L. Van Hemmen and K. Schulten, editors, Physics of neural networks. Springer, 1994.

[21] K. Pawelzik, J. Kohlmorgen, and K.-R. M¨uller. Annealed competition of experts for a segmentation and classification of switching dynamics. Neural Computation, 8(2):340–356, 1996.

[22] J. C. Platt. Fast training of support vector machines using sequential minimal optimization. In B. Sch¨olkopf, C. J. C. Burges, and A. J. Smola, editors, Advances in Kernel Methods - Support Vector Learning, Cambridge, MA, 1998. MIT Press.

[23] L. Rabiner. A tutorial on hidden markov models and selected applications in speech recognition. Proceedings of the IEEE, 77(2):257–285, 1989.

[24] K. Rose, F. Gurewitz, and G. Fox. Statistical mechanics and phase transitions in clustering. Physical Rev. Letters, 65(8):945–948, 1990.

[25] B. Sch¨olkopf, K.-K. Sung, C. J. C. Burges, F. Girosi, P. Niyogi, T. Poggio, and V. Vapnik. Comparing support vector machines with gaussian kernels to radial basis function classiers. IEEE Transactions on Signal Processing, 45(11):2758– 2765, 1997.

[26] V. Vapnik. Statistical Learning Theory. Wiley, New York, NY, 1998.

[27] A. S. Weigend and N. A. Gershenfeld, editors. Time series prediction: forcasting the future and understanding the past. Addison-Wesley, 1994.

(16)

PLACE PHOTO HERE

Mathematics in 2001 from National Taiwan University, where in 2003 he received his M.S. degree in Computer Science & Information Engineering. His research interests include machine learning, numerical optimization, and algorithm analysis.

PLACE PHOTO HERE

Chih-Jen Lin (S’91-M’98) received his B.S. degree in Mathematics from National Taiwan University

in 1993. From 1995 to 1998 he studied at the Department of Industrial and Operations Engineering at the University of Michigan and received his M.S. and Ph.D. degrees. He is currently associate professor in the Department of Computer Science and Information Engineering at National Taiwan University. His research interests include machine learning, numerical optimization, and applications of Operations Research.

PLACE PHOTO HERE

Ruby Chiu-Hsing Weng received her B.S. and M.S. degrees in Mathematics from National Taiwan

University in 1993 and 1995, and Ph.D. degree in Statistics from the University of Michigan in 1999. She is Assistant Professor, Department of Statistics, National Chengchi University, Taipei. Her research interests include sequential analysis and time series analysis.

數據

Fig. 1. First four iterations (data without noise)
Fig. 2. First six iterations (data with noise)
Fig. 4. The Distribution of the Selected C Values
Fig. 5. Number of support vectors (i.e. α t − α ∗ t 6= 0) and number of free support vectors (i.e

參考文獻

相關文件

– stump kernel: succeeded in specific applications infinite ensemble learning could be better – existing AdaBoost-Stump applications may switch. not the

2 Distributed classification algorithms Kernel support vector machines Linear support vector machines Parallel tree learning.. 3 Distributed clustering

2 Distributed classification algorithms Kernel support vector machines Linear support vector machines Parallel tree learning?. 3 Distributed clustering

•providing different modes of support in learning tasks (e.g.. You are a summer intern in the publicity team of Go Green, a non- governmental organisation committed to

support vector machine, ε-insensitive loss function, ε-smooth support vector regression, smoothing Newton algorithm..

Abstract In this paper, we study the parabolic second-order directional derivative in the Hadamard sense of a vector-valued function associated with circular cone.. The

“Transductive Inference for Text Classification Using Support Vector Machines”, Proceedings of ICML-99, 16 th International Conference on Machine Learning, pp.200-209. Coppin

Solving SVM Quadratic Programming Problem Training large-scale data..