• 沒有找到結果。

Optimizing 0/1 Loss for Perceptrons by Random Coordinate Descent

N/A
N/A
Protected

Academic year: 2022

Share "Optimizing 0/1 Loss for Perceptrons by Random Coordinate Descent"

Copied!
6
0
0

加載中.... (立即查看全文)

全文

(1)

Proceedings of International Joint Conference on Neural Networks, Orlando, Florida, USA, August 12–17, 2007

Optimizing 0/1 Loss for Perceptrons by Random Coordinate Descent

Ling Li and Hsuan-Tien Lin

Abstract—The 0/1 loss is an important cost function for perceptrons. Nevertheless it cannot be easily minimized by most existing perceptron learning algorithms. In this paper, we propose a family of random coordinate descent algorithms to directly minimize the 0/1 loss for perceptrons, and prove their convergence. Our algorithms are computationally efficient, and usually achieve the lowest 0/1 loss compared with other algo- rithms. Such advantages make them favorable for nonseparable real-world problems. Experiments show that our algorithms are especially useful for ensemble learning, and could achieve the lowest test error for many complex data sets when coupled with AdaBoost.

I. INTRODUCTION

The perceptron was first introduced by Rosenblatt [1] as a probabilistic model for information processing in the brain.

It is simply a linear threshold classifier, which can be thought as a hyperplane in the input space.

Given a set of examples with binary labels, an important task of perceptron learning is to find a hyperplane that clas- sifies the examples with the smallest number of mislabeling errors. A data set that can be classified by some perceptron without any mislabeling errors is called linearly separable, or simply separable. For a separable set, the task of learning is relatively easy and can be carried out by many existing algorithms. For example, the perceptron learning rule [2]

is guaranteed to converge to a separating hyperplane in a finite number of iterations. The hard-margin support vector machine (SVM) can even find the separating hyperplane that maximizes the minimal example margin [3].

However, these algorithms behave poorly when the data set is nonseparable, which is a more common situation in real- world problems. In such a situation, the perceptron learning rule will not converge, and is very unstable in the sense that the learned hyperplane might change from an optimal one to a worst-possible one within a single trial [4]. The optimization problem of the hard-margin SVM becomes infeasible, and hence cannot be solved without modifications.

To tackle the nonseparable cases, many different algorithms have been proposed (see Section II). Although those algo- rithms appear quite different, they usually try to minimize some cost functions of example margins. Note that the number of mislabeling errors is proportional to a specific cost function, the 0/1 loss.

The 0/1 loss is an important criterion for perceptron learning. It captures the discrete nature of the binary clas- sification (correct or incorrect), and partially indicates the

The authors are with the Learning Systems Group, Computer Science Option, California Institute of Technology, Pasadena, CA 91125, USA (email: {ling, htlin}@caltech.edu).

prediction power. A good 0/1 loss minimization algorithm can be used either to obtain standalone perceptrons, or to build more complex classifiers with many perceptrons.

However, the 0/1 loss cannot be directly minimized by most existing perceptron algorithms, which is both because the minimization problem is NP-complete [5], and because the loss is neither convex nor smooth.

In this paper, we propose a family of new perceptron algorithms to directly minimize the 0/1 loss. The central idea is random coordinate descent, i.e., iteratively search- ing along randomly chosen directions. An efficient update procedure is used to exactly minimize the 0/1 loss along the chosen direction. Both the randomness and the exact minimization procedure help escape from local minima.

Theoretical analyses indicate that our algorithms globally minimize the 0/1 loss with arbitrarily high probability under simple settings, and perform random search towards an optimal hyperplane efficiently. Experimental results further demonstrate that our algorithms achieve the best 0/1 loss most of the time when compared with other perceptron algorithms, and are thus favorable base learners for ensemble learning methods such as AdaBoost [6].

The paper is organized as follows. We discuss some of the existing algorithms in Section II. We then introduce our random coordinate descent algorithms, as well as their convergence analyses, in Section III. Our algorithms are compared with existing ones, both as standalone learners and as base learners of AdaBoost, in Section IV. Finally, we conclude in Section V.

II. RELATEDWORK

We assume that the input space is a subset of Rm. A perceptron consists of a weight vector (w1, . . . , wm) and a bias term b (i.e., the negative threshold). To avoid treating them separately, we use the notation w = (w0, w1, . . . , wm) with w0 = b, and accordingly expand the original input vector to x ∈ Rm+1 with x0 = 1. Then, the perceptron performs classification by the sign of the inner product between w and x, i.e., sign (hw, xi).

For a given training set {(xi, yi)}Ni=1 with yi ∈ {−1, 1}

being the labels, let ρi = yihw, xii be the unnormalized margin of the i-th example. Most perceptron algorithms try to minimize some cost function based on ρi:

C(w) =

N

X

i=1

ϕic (ρi) , (1) where ϕi is the sample weight, and c (ρ) is a margin cost function. Some common margin cost functions are listed

(2)

TABLE I

SOME COMMON MARGIN COST FUNCTIONS

cost function c(ρ)

perceptron criterion max {0, −ρ}

SVM hinge loss max {0, 1 − ρ}

modified least-squares (max {0, 1 − ρ})2

0/1 loss [ρ ≤ 0]

in Table I. Note that many margin cost functions, including the first three in Table I, can be viewed as monotonic and continuous approximations of the 0/1 loss [7], which is a direct measure of the classification performance.

The well-known perceptron learning rule (PLR) [2] per- forms gradient descent on the perceptron criterion associated with some individual (xi, yi). It updates the weight vector w when it predicts wrongly on xi, i.e., when yihw, xii ≤ 0,

w⇐ w + yixi. (2)

This update rule is applied repeatedly to every example in the training set. If the training set is separable, the perceptron convergence theorem [2] guarantees that a separating hyper- plane can be found in finite time. However, if the training set is nonseparable, the algorithm will never converge and there is no guarantee for obtaining good perceptrons in terms of the 0/1 loss.

The pocket algorithm [4] solves the convergence problem of PLR at the price of much more computation. It runs PLR while keeping “in its pocket” the best-till-now weight vector in terms of the 0/1 loss. Although it can find an optimal weight vector that minimizes the 0/1 loss with arbitrarily high probability, the number of epochs required is prohibitively large in practice [4]. This is partly because the algorithm aims at minimizing the 0/1 loss, but only adopts (2) from PLR to update the weight vector. Since PLR is very unstable when the training set is nonseparable, much computation is wasted on bad weight vectors.

In contrast to the pocket algorithm, which uses only the best weight vector, Freund and Schapire [8] suggested com- bining all the weight vectors that occurred in the trials of PLR by a majority vote. One variant, which uses averaging instead of voting, can produce a single perceptron. Since it was shown that the voted- and the averaged-perceptron models perform similarly in practice [8], we will only consider the averaged-perceptron algorithm in this paper. The averaged- perceptron algorithm operates with the perceptron criterion, but does not explicitly minimize any cost functions. Thus, the obtained perceptron is usually not the best minimizer of either the perceptron criterion or the 0/1 loss.

When the margin cost function c(ρ) satisfies certain smoothness assumptions, which is the case of all cost func- tions in Table I except the 0/1 loss, the stochastic gradient descent algorithm (SGD) [9] can be used to minimize the associated C(w). For example, PLR is just a special case of SGD with the perceptron criterion. However, because the 0/1 loss has sharp transitions at ρ = 0, and zero gradients elsewhere, it cannot be directly minimized by SGD.

Minimizing the 0/1 loss for perceptrons is a challenging

task. The problem is NP-complete [5], and hence determinis- tic optimization is likely to be inefficient. In addition, general numerical optimization cannot work because of the zero gradients, non-convexity, and non-smoothness. Practically, smooth and preferably convex approximations of the 0/1 loss are usually considered instead.

Nevertheless, the 0/1 loss is important because it captures the discrete nature of the binary classification (correct or incorrect). It is thus interesting to see whether a decent minimizer of the 0/1 loss could outperform minimizers of other cost functions. In addition, a good 0/1 loss minimizer can be useful in some practical cases. For example, adaptive boosting (AdaBoost), one of the most popular ensemble learning algorithms, expects an efficient and decent 0/1 loss minimizer as the base learner [6]. Such a good base learner could help AdaBoost in training speed and algorithmic convergence. However, existing perceptron algorithms men- tioned above usually cannot be good base learners (see Section IV), because of slowness (e.g., pocket) and/or cost function mismatch (e.g., averaged-perceptron).

III. RANDOMCOORDINATEDESCENT

From now on, we will focus on the 0/1 loss for perceptrons and propose a family of algorithms to directly minimize it. The notation E(w), or the word error, will be used to specifically represent C(w) with the 0/1 loss. Similar to (2), our algorithms updates the perceptron weight vector iteratively by choosing an update direction d and a proper descent step α,

w⇐ w + αd. (3)

We first show how to determine α for a given direction d, and then discuss how we could choose d.

A. Finding Optimal Descent Step

Assume that an update direction d has been chosen. To have the most decrease of E(w) along the direction d, we determine the best step size α by solving

minα∈RE(w + αd) =

N

X

i=1

ϕi[yihw + αd, xii ≤ 0] . (4) We can take a closer look at the error that (w + αd) makes on an individual example (xi, yi). Let δi= hd, xii.

When δi 6= 0, hw + αd, xii = δi δ−1i hw, xii + α. The error of (w + αd) on (xi, yi) is the same as the error of a 1-D decision stump [10] with bias α on the example δ−1i hw, xii , yisign (δi).

When δi = 0, hw + αd, xii = hw, xii. Thus, the error does not change with α.

There exists a deterministic and efficient algorithm for mini- mizing the training error for decision stumps [10]. Hence, we can transform all training examples with δi 6= 0 us- ing (xi, yi) 7→ δ−1i hw, xii , yisign (δi), and then apply the decision stump learning algorithm on the transformed training set to decide the optimal descent step α. Such an update procedure is illustrated in Fig. 1. Note that α is not restricted to be positive, and hence the direction d does not need to be strictly descent.

(3)

Input: A training set {(xi, yi)}Ni=1, its sample weight {ϕi}Ni=1, the current w, and an update direction d

1: fori= 1to N do {generate the 1-D training set}

2: δi⇐ hd, xii 3: ifδi6= 0 then

4: x0i⇐ δ−1i hw, xii, yi0⇐ yisign (δi) 5: end if

6: end for

7: Find the optimal decision stump only for those {(x0i, yi0)}

and {ϕi}with δi6= 0: α⇐ argmin

α∈R

X

i: δi6=0

ϕiˆy0i· sign`x0i+ α´ ≤ 0˜

8: w⇐ w + αd

Fig. 1. The update procedure that solves (4)

Input: A training set, sample weight, and the number of epochs T 1: Initialize w(1)

2: fort= 1to T do

3: Generate a random update direction d(t)∈ Rm+1 4: Update w(t)to w(t+1) with the algorithm in Fig. 1 5: end for

6: return w(T +1) as the perceptron weight vector Fig. 2. Random coordinate descent (RCD) for perceptrons

B. Choosing Update Directions

There are many ways to choose an update direction. The simplest way might be the cyclic coordinate descent (CCD).

It focuses on one basis vector in each epoch, and rotates through the (m + 1) basis vectors repeatedly. Such a tech- nique is commonly used when the cost function is not differentiable.

To avoid getting stuck in some local minima potentially caused by using fixed directions in a fixed order, we may choose the update directions randomly. This method, together with the update procedure, is called random coordinate descent (RCD) and is depicted in Fig. 2. Note that CCD can be thought as a degenerate variant of RCD. Below we will discuss some other representative variants of RCD algorithms.1

The simplest variant is called RCD-plain, in which the update directions are independent and identically-distributed random vectors. We have used two common distributions for generating the random vectors. The first one, called the uniform random vectors, picks each component of the vector from a uniform distribution spanned over the corresponding feature range. The other one uses Gaussian distribution estimated from the corresponding feature statistics, and is called the Gaussian random vectors.

Note that the final value of the bias, w0, can be in quite different ranges from the other components of w, due to the setting x0 = 1. Thus, it might be helpful to have an update direction devoted to adjusting w0 only. If the zeroth basis vector is additionally adopted as the update direction every (m + 1) epochs,RCD-plainbecomesRCD-bias.

Inspired by what PLR does, we may also use the gradient of the perceptron criterion associated with a randomly picked

1More variants can be found in our earlier technical report [11].

example (xi, yi) as the update direction. We call this vari- antRCD-grad. The difference between PLR andRCD-grad is that PLR does not pursue the optimal descent step.

C. Convergence of RCD

Next we analyze the convergence properties of RCD algorithms, which follows from random search techniques in optimization [12]. Note that both E(w) and RCD algorithms are invariant to the magnitude of vectors w or d. For easier analysis, we assume that every nonzero w(t) or d(t) is normalized to unit length.

First, we show a general result of global convergence that applies to RCD-plain and RCD-bias, with either uniform or Gaussian random vectors.

Definition 1: We call d(t)

t=1 sufficiently random if Qt=1 1 − Prd(t)∈ A = 0for all A ⊆ S ∪ {0} with nonzero measure, where S = {w: kwk = 1}.

Theorem 1: For an RCD algorithm with sufficiently ran- dom directions d(t)

t=1,

t→∞lim Prh

w(t) minimizes E(w)i

= 1. (5)

Proof: We extend from the global search convergence theorem [12, page 20], which, together with the property that E(w) is lower bounded and evaluates to discrete values, implies the correctness of (5) under the following conditions:

1) S ∪ {0} is a measurable set; E(w) is measurable.

2) The directions d(t)

t=1 are sufficiently random.

3) The sequence E(w(t))

t=1 is non-increasing, and E(w(t+1)) ≤ E(d(t)).

The proof relies on the third condition, which holds be- cause α is the minimizer of E(w(t)+ αd(t))in Fig. 1.

We can easily verify that the assumption of Theorem 1 is satisfied byRCD-plainandRCD-bias, with either the uni- form or the Gaussian random vectors. However, Theorem 1 does not fully reveal their efficiency. Next, we take RCD- plainwith the Gaussian random vectors d(t)∼ N (0, 1)as an example, and show how it performs efficient random search.

Define B(w, R) as a hypersphere in Rm+1 with cen- ter w and radius R. Let

A(w, R) = {w/ kwk : w ∈ B(w, R) \ {0}} ⊂ S.

An integration on S shows that when kwk = 1, the probability of the direction d(t)∈ A(w, R)is

m · vm Z R

0

rm−2dp

1 − r2≥ vmRm, (6) where vm is some constant [13]. Note that in the space of weight w, the N hyperplanes hxi, wi = 0 cut the surface S into different regions, each of which represents a specific prediction pattern on the training set. Following the analysis of Dunagan and Vempala [14], each region can be approximately modeled by some A(w, R)with kwk = 1.2 Assume that w is from a region where E(w) < E(w(t)).

2Within the region, we choose a wsuch that R is maximized.

(4)

Equation (6) gives a lower bound for locating the region in a naive random search. However,RCD-plaincan perform the search more efficiently from the current w(t):

Theorem 2: For the piece A(w, R)defined above, if the angle between w and w(t) is φ ∈ [0,π2], then RCD-plain with the standard Gaussian random vectors satisfies

Prh

E(w(t+1)) ≤ E(w)i

≥ vmRm sinmφ.

Proof: The scale invariance of E(w) implies that E(w) = E(w)for any k > 0 and w ∈ B(kw, kR).

Consider d(t) such that the line w(t) + αd(t) intersects B(kw, kR), which is equivalent to d(t) ∈ Ak

with Ak = A(kw− w(t), kR). Thus, for any k > 0, Prh

E(w(t+1)) ≤ E(w)i

≥ Prh

d(t)∈ Aki . The theorem is proved by setting k = 1/ cos φ, which results in a piece Ak with the maximum measure.

Similar analysis can be carried out for the case of the uniform or other Gaussian random vectors. When w(t) and w are close, the angle φ between them is small, and RCD-plain greatly improves the lower bound of the probability for decreasing E(w). In other words,RCD-plainperforms fast local search for a better w nearby. Note that because of the structure of the w-space, a region with global minima is close to regions with small training errors. With the implicit use of the structure, and the ability to escape from local minima, RCD-plain (and similarly RCD-bias) can hence perform global minimization efficiently.

IV. EXPERIMENTS

We first compare the performance of different RCD vari- ants in minimizing the training error. Some better variants are further compared with the existing perceptron algorithms.

We use nine real-world data sets3 from the UCI machine learning repository [15], with 80% of randomly chosen examples for training and the rest for testing. Three artificial data sets4 are also randomly generated, with 600 examples for training and 4400 for testing. The 0/1 loss (error) is measured over 500 runs on the training and testing sets, and is presented using the mean and the standard error.

a) Data Preprocessing.: The features in the training set are first linearly normalized to [−1, 1] solely based on the training examples. Then the test examples are normalized accordingly.

b) Initial Seeding.: We initialize w(1) with either the zero vector or the Fisher’s linear discriminant (FLD, see for example [17]). For the latter case, any singular estimate of the within-class covariance matrix is regularized with an eigenvalue shrinkage parameter of 10−10 [18].

A. Comparisons within RCD Variants

In Fig. 3, we compare different RCD variants using their training errors on the pima data set. The results shown are

3They areaustralian,breast,cleveland,german,heart,ionosphere, pima,sonar, andvotes84. See [11] for details.

4They areringnorm,threenorm[16], andyinyang. See [11] for details.

100 101 102 103

20 22 24 26 28 30 32 34 36

Number of epochs

Training error (%)

CCD RCD−grad RCD−plain FLD, CCD FLD, RCD−grad FLD, RCD−plain

100 101 102 103

19.5 20 20.5 21 21.5 22

Number of epochs

Training error (%)

FLD, CCD FLD, RCD−grad FLD, RCD−plain FLD, RCD−bias

Fig. 3. Training errors of RCD algorithms onpima

100 101 102 103

19.5 20 20.5 21 21.5 22 22.5 23 23.5 24

Number of epochs

Training error (%)

FLD, RCD−bias FLD, pocket FLD, ave−perc FLD, SGD−mls

100 101 102 103

19.5 20 20.5 21 21.5 22 22.5 23 23.5 24

Number of epochs

Test error (%)

FLD, RCD−bias FLD, pocket FLD, ave−perc FLD, SGD−mls

Fig. 4. Training and test errors of stochastic algorithms onpima

based on uniform random vectors, while similar results have been observed for Gaussian random vectors. Some important observations are:

With the same number of epochs, training errors of RCD algorithms using the FLD seeding are significantly lower than those using the zero seeding.

BothCCDandRCD-graddo not work as well asRCD- plain and RCD-bias. This result confirms that suffi- ciently random directions are important for RCD.

RCD-biasis better than RCD-plain, especially at the beginning. However, the edge gets smaller with more training epochs.

We obtain similar observations in some other data sets.

Thus,RCD-plainandRCD-bias, with the FLD seeding, are more promising. Next, we compare them with some existing perceptron algorithms. For a fair comparison, we also equip the other algorithms with the FLD seeding.

B. Comparisons as Standalone Learners

We consider the pocket algorithm with ratchet (denoted as pocket) [4], an improved variant of the averaged- perceptron algorithm, in which examples are presented com- pletely randomly during training (ave-perc) [8, 11], SGD with a learning rate 0.002 on the modified least- squares (SGD-mls) [9], and the linear soft-margin SVM with parameter selection (soft-SVM) [19, 20]. The first three stochastic algorithms can be directly compared to RCD algorithms by allowing the same T = 2000 epochs, and the last one is included for reference.

Fig. 4 presents the performance of the stochastic algo- rithms on the pima data set.5 In the competition for low training errors, RCD-bias is clearly the best, and pocket

5The curves ofRCD-plainare very close to those ofRCD-bias,and are thus not shown.

(5)

TABLE II

TRAINING ERRORS(%)OF PERCEPTRON ALGORITHMS USING THEFLDSEEDING

data set RCD-plain RCD-bias pocket ave-perc SGD-mls soft-SVM

australian 10.12± 0.03 9.98± 0.03 10.81± 0.03 12.19± 0.03 12.70± 0.04 14.33± 0.03 breast 1.68± 0.01 1.68± 0.01 1.86± 0.01 2.87± 0.02 2.77± 0.02 2.70± 0.02 cleveland 10.57± 0.05 10.62 ± 0.05 12.07± 0.05 14.40± 0.06 14.48± 0.06 14.74± 0.05 german 19.16± 0.04 18.80± 0.03 21.10± 0.03 21.31± 0.04 22.18± 0.05 21.48± 0.04 heart 9.48± 0.05 9.49± 0.05 11.22± 0.05 13.64± 0.06 13.82± 0.06 14.20± 0.06 ionosphere 3.88± 0.04 3.97± 0.04 3.41± 0.05 4.92± 0.06 5.14± 0.05 6.95± 0.10 pima 19.60± 0.04 19.60 ± 0.03 20.34± 0.03 21.99± 0.04 22.25± 0.04 22.09± 0.04 ringnorm 27.61± 0.07 27.36± 0.08 30.46± 0.07 35.49± 0.11 34.52± 0.13 31.82± 0.09 sonar 2.56± 0.04 2.62± 0.04 0.00± 0.00 0.37± 0.02 1.42± 0.06 11.58± 0.20 threenorm 11.41± 0.06 11.39 ± 0.06 13.53± 0.06 14.43± 0.06 14.51± 0.06 14.47± 0.06 votes84 1.32± 0.02 1.31± 0.02 1.46± 0.02 2.42± 0.03 2.48± 0.03 3.02± 0.04 yinyang 15.33± 0.05 15.36 ± 0.05 15.61± 0.05 19.10± 0.07 19.03± 0.07 18.89± 0.08

(results within one standard error of the best are marked in bold) TABLE III

TEST ERRORS(%)OF PERCEPTRON ALGORITHMS USING THEFLDSEEDING

data set RCD-plain RCD-bias pocket ave-perc SGD-mls soft-SVM

australian 14.24± 0.12 13.92± 0.12 14.31± 0.12 13.64± 0.12 13.87± 0.12 14.78± 0.12 breast 3.65± 0.07 3.61± 0.07 3.43± 0.06 3.36± 0.06 3.28± 0.06 3.22± 0.06 cleveland 18.68± 0.22 18.57± 0.21 18.49± 0.21 16.74± 0.20 16.76 ± 0.20 16.72 ± 0.20 german 24.45± 0.12 23.70± 0.13 25.24± 0.13 23.24± 0.12 24.05± 0.13 23.64± 0.12 heart 18.13± 0.21 18.20± 0.22 17.63± 0.20 16.51± 0.20 16.49 ± 0.20 16.45 ± 0.20 ionosphere 13.91± 0.17 14.72± 0.18 12.87± 0.18 12.76± 0.18 12.63± 0.18 12.57 ± 0.17 pima 23.79± 0.14 23.50± 0.14 23.50± 0.14 22.79± 0.14 23.07± 0.14 23.19± 0.14 ringnorm 35.83± 0.04 35.65± 0.04 36.59± 0.04 39.27± 0.08 38.38± 0.10 35.70± 0.05 sonar 25.98± 0.29 26.20± 0.29 25.20± 0.25 25.09± 0.26 24.90± 0.28 23.89± 0.27 threenorm 16.82± 0.03 16.86± 0.03 17.65± 0.04 16.14± 0.02 16.18± 0.02 16.08± 0.02 votes84 5.21± 0.09 5.00± 0.10 5.24± 0.10 4.52± 0.10 4.70± 0.11 4.39± 0.09 yinyang 17.71± 0.02 17.75± 0.02 17.74± 0.02 19.25± 0.02 19.21± 0.02 19.21± 0.02

(results within one standard error of the best are marked in bold)

follows. However, when the test error is concerned, the other three methods, especially ave-perc, are the winners. The final performance of all perceptron algorithms are shown in Tables II and III. Similarly, RCD-plain and RCD-bias achieve the lowest training errors for most of the data sets.

On the other hand, soft-SVM and ave-perc, which are known to be regularized, could usually obtain better test errors. Thus, the 0/1 loss itself may not be the best cost function, and overfitting without regularization shall explain the inferior test performance of RCD algorithms.

We also observe that pocket is much slower thanRCD- bias, because with nonseparable data sets, checks on whether a new weight vector should replace the in-pocket one may happen very often. In addition, pocket usually also needs more epochs in order to achieve the same level of training error asRCD-bias.

C. Comparisons as AdaBoost Base Learners

AdaBoost expects its base learners to efficiently find a hypothesis with low weighted training error. The details for plugging the sample weights ϕi into existing perceptron algorithms are listed in our technical report [11]. We use the reweighting technique for pocket and ave-perc. Then, we set T = 200 for all perceptron algorithms, which seems to be sufficient for all the data sets, and apply them as AdaBoost base learners. We run AdaBoost for up to 200 iterations, and report the results with the zero seeding in Table IV, while similar results have been obtained with the FLD seeding.

We observe that ave-perc, SGD-mls, and soft-SVM, which are not designed for the weighted training error, usually fail to return a decent perceptron, and cause Ada- Boost to stop at some early iteration [11]. That is, their performance is mainly associated with a few regularized perceptrons rather than with AdaBoost. On the other hand, AdaBoost with RCD-plain, RCD-bias, or pocket always runs through all 200 iterations, and hence their performance is fully connected to AdaBoost.

To further compare perceptron algorithms as a base learn- ers to AdaBoost, we mark the entries with in Table IV to denote a significant improvement from the best entry of Table III. There are some data sets without, which means that they are simple enough, and can be modeled sufficiently well with one single perceptron (e.g.,australian). For those data sets, ave-perc, SGD-mls, and soft-SVM have good performance both as standalone learners and with AdaBoost.

However, since they already perform well as standalone learners, there is no need in binding them with AdaBoost on those data sets.

On the other hand, for complex data sets that can be better modeled by AdaBoost with perceptrons (e.g.,yinyang), base learners that aim at minimizing the 0/1 loss (RCD-plain, RCD-bias, or pocket) perform significantly better than others. In addition,RCD-plain and RCD-bias are usually better thanpocketin terms of both the training speed and the performance. Thus,RCD-plainandRCD-biasfit the needs

(6)

TABLE IV

TEST ERRORS(%)OFADABOOST WITH PERCEPTRON ALGORITHMS USING THE ZERO SEEDING

data set RCD-plain RCD-bias pocket ave-perc SGD-mls soft-SVM

australian 15.45± 0.12 15.49± 0.12 15.75± 0.12 13.61± 0.12 14.00 ± 0.12 15.65± 0.13 breast 3.21± 0.06 3.34± 0.06 3.41± 0.07 3.35± 0.06 3.24± 0.06 3.20± 0.06 cleveland 18.00± 0.21 18.22± 0.21 18.95± 0.20 16.81± 0.20 16.74 ± 0.20 16.69 ± 0.21 german 25.17± 0.13 25.37± 0.12 25.57± 0.13 23.25± 0.12 23.96 ± 0.13 23.48± 0.12 heart 17.60± 0.21 17.58± 0.22 18.94± 0.21 16.55± 0.20 16.54 ± 0.20 16.57 ± 0.21 ionosphere10.36± 0.1610.30± 0.1611.65± 0.17 13.21± 0.17 12.67± 0.17 10.83± 0.16 pima 24.87± 0.14 24.79± 0.14 25.15± 0.14 22.77± 0.14 23.01 ± 0.14 23.06± 0.13 ringnorm 8.60± 0.05 12.22± 0.07 7.12± 0.04 39.29± 0.08 38.32± 0.09 15.49± 0.17 sonar 16.44± 0.25 16.06± 0.2525.02± 0.27 25.77± 0.27 25.37± 0.27 21.52± 0.25 threenorm 14.51± 0.02 15.34± 0.03 14.95± 0.02 16.14± 0.02 16.17± 0.02 15.97± 0.02 votes84 4.25± 0.09 4.24± 0.09 4.54± 0.10 4.74± 0.10 4.68± 0.10 4.82± 0.10 yinyang 3.95± 0.03 3.98± 0.03 4.87± 0.02 19.25± 0.02 19.23± 0.02 19.14± 0.02

(results within one standard error of the best entry in the row are marked in bold) (results better than the best of Table III by one standard error are marked with)

of AdaBoost, and can almost always achieve the best test error in this situation. The results demonstrate the usefulness of RCD algorithms as base learners of AdaBoost.

V. CONCLUSION

We have proposed a family of new learning algorithms to directly optimize the 0/1 loss for perceptrons. The main ingredients are random coordinate descent (RCD) and an update procedure to exactly minimize the 0/1 loss along the update direction. We have proved the convergence of our RCD algorithms, and have also analyzed their speedup heuristic. Experimental results have confirmed that RCD algorithms are efficient, and usually achieve the lowest in- sample 0/1 loss compared with several existing perceptron learning algorithms.

RCD algorithms can be used not only as standalone learners, but also as favorable base learners for ensemble learning. In terms of the improvement for the out-of-sample performance, our experimental results have demonstrated that AdaBoost works better with RCD algorithms than with other existing perceptron algorithms.

On some data sets on which a single perceptron works well, our results indicate that RCD-based 0/1 loss minimizer may not be the best choice. On the other hand, regularized algorithms such as the averaged-perceptron and the soft- margin SVM could achieve better out-of-sample perfor- mance. Future work will be focused on RCD algorithms for some regularized 0/1 loss.

ACKNOWLEDGMENTS

We wish to thank Yaser Abu-Mostafa and Amrit Pratap for many valuable discussions. This work was mainly supported by the Caltech SISL Graduate Fellowship.

REFERENCES

[1] F. Rosenblatt, “The perceptron: A probabilistic model for information storage and organization in the brain,” Psychological Review, vol. 65, pp. 386–408, 1958.

[2] ——, Principles of Neurodynamics: Perceptrons and the Theory of Brain Mechanisms. Spartan, 1962.

[3] V. N. Vapnik, The Nature of Statistical Learning Theory. Springer- Verlag, 1995.

[4] S. I. Gallant, “Perceptron-based learning algorithms,” IEEE Transac- tions on Neural Networks, vol. 1, pp. 179–191, 1990.

[5] P. Marcotte and G. Savard, “Novel approaches to the discrimination problem,” Mathematical Methods of Operations Research, vol. 36, pp.

517–545, 1992.

[6] Y. Freund and R. E. Schapire, “Experiments with a new boosting algorithm,” in Machine Learning: Proceedings of the Thirteenth In- ternational Conference (ICML ’96), L. Saitta, Ed., 1996, pp. 148–156.

[7] T. Hastie, R. Tibshirani, and J. Friedman, The Elements of Statistical Learning: Data Mining, Inference, and Prediction, ser. Springer Series in Statistics. Springer-Verlag, 2001.

[8] Y. Freund and R. E. Schapire, “Large margin classification using the perceptron algorithm,” Machine Learning, vol. 37, pp. 277–296, 1999.

[9] T. Zhang, “Solving large scale linear prediction problems using stochastic gradient descent algorithms,” in ICML 2004: Proceedings of the Twenty-First International Conference on Machine Learning, R. Greiner and D. Schuurmans, Eds., 2004.

[10] R. C. Holte, “Very simple classification rules perform well on most commonly used datasets,” Machine Learning, vol. 11, pp. 63–91, 1993.

[11] L. Li, “Perceptron learning with random coordinate descent,” Cal- ifornia Institute of Technology, Pasadena, CA, Computer Science Technical Report CaltechCSTR:2005.006, 2005.

[12] F. J. Solis and R. J.-B. Wets, “Minimization by random search techniques,” Mathematics of Operations Research, vol. 6, pp. 19–30, 1981.

[13] S. Axler, P. Bourdon, and W. Ramey, Harmonic Function Theory, 2nd ed., ser. Graduate Texts in Mathematics. Springer-Verlag, 2001.

[14] J. Dunagan and S. Vempala, “A simple polynomial-time rescaling algorithm for solving linear programs,” in Proceedings of the 36th Annual ACM Symposium on Theory of Computing, L. Babai, Ed., 2004, pp. 315–320.

[15] S. Hettich, C. L. Blake, and C. J. Merz, “UCI repository of ma- chine learning databases,” 1998, available at http://www.ics.uci.edu/

mlearn/MLRepository.html.

[16] L. Breiman, “Arcing classifiers,” The Annals of Statistics, vol. 26, pp.

801–824, 1998.

[17] C. M. Bishop, Neural Networks for Pattern Recognition. Oxford University Press, 1995.

[18] J. H. Friedman, “Regularized discriminant analysis,” Journal of the American Statistical Association, vol. 84, pp. 165–175, 1999.

[19] C.-C. Chang and C.-J. Lin, LIBSVM: A library for support vector ma- chines, 2001, software available at http://www.csie.ntu.edu.tw/cjlin/

libsvm.

[20] C.-W. Hsu, C.-C. Chang, and C.-J. Lin, “A practical guide to support vector classification,” National Taiwan University, Tech. Rep., 2003.

參考文獻

相關文件

On the other hand, by (22), the triangle inequality, and the definition of the operator norm... On the other hand, by (22), the triangle inequality, and the definition of the

HAMR =⇒ lower 0/1 loss, similar Hamming loss BCH =⇒ even lower 0/1 loss, but higher Hamming loss to improve Binary Relevance, use. HAMR or BCH =⇒ lower 0/1 loss, lower

• Color the nodes representing literals by their truth values (color 0 for false and color 1 for true).. – We are dealing with the a-triangles here, not the

Most existing machine learning algorithms are designed by assuming that data can be easily accessed.. Therefore, the same data may be accessed

a single instruction.. Thus, the operand can be modified before it can be modified before it is used. Useful for fast multipliation and dealing p g with lists, table and other

Now given the volume fraction for the interface cell C i , we seek a reconstruction that mimics the sub-grid structure of the jump between 0 and 1 in the volume fraction

Bootstrapping is a general approach to statistical in- ference based on building a sampling distribution for a statistic by resampling from the data at hand.. • The

The hashCode method for a given class can be used to test for object equality and object inequality for that class. The hashCode method is used by the java.util.SortedSet