• 沒有找到結果。

Support Vector Machines and Kernel Methods

N/A
N/A
Protected

Academic year: 2022

Share "Support Vector Machines and Kernel Methods"

Copied!
79
0
0

加載中.... (立即查看全文)

全文

(1)

Support Vector Machines and Kernel Methods

Chih-Jen Lin

Department of Computer Science National Taiwan University

Talk at International Workshop on Recent Trends in Learning,

(2)

Outline

Basic concepts: SVM and kernels Training SVM

Practical use of SVM Multi-class classification

Research directions: large-scale training Research directions: linear SVM

Research directions: others Conclusions

(3)

Outline

Basic concepts: SVM and kernels Training SVM

Practical use of SVM Multi-class classification

Research directions: large-scale training Research directions: linear SVM

Research directions: others Conclusions

(4)

Support Vector Classification

Training vectors : xi, i = 1, . . . , l Feature vectors. For example, A patient = [height, weight, . . .]T

Consider a simple case with two classes:

Define an indicator vector y yi =

 1 if xi in class 1

−1 if xi in class 2 A hyperplane which separates all data

(5)

wTx + b = h+1

−10

i

A separating hyperplane: wTx + b = 0 (wTxi) + b ≥ 1 if yi = 1 (wTxi) + b ≤ −1 if yi = −1

Decision function f (x) = sgn(wTx + b), x: test data Many possible choices of w and b

(6)

Maximal Margin

Distance between wTx + b = 1 and −1:

2/kwk = 2/

√ wTw

A quadratic programming problem (Boser et al., 1992)

minw,b

1 2wTw

subject to yi(wTxi + b) ≥ 1, i = 1, . . . , l .

(7)

Data May Not Be Linearly Separable

An example:

Allow training errors

Higher dimensional ( maybe infinite ) feature space φ(x) = [φ1(x), φ2(x), . . .]T.

(8)

Standard SVM (Boser et al., 1992; Cortes and Vapnik, 1995)

min

w,b,ξ

1

2wTw +C

l

X

i =1

ξi

subject to yi(wTφ(xi)+ b) ≥ 1 −ξi, ξi ≥ 0, i = 1, . . . , l .

Example: x ∈ R3, φ(x) ∈ R10 φ(x) = [1,√

2x1,√

2x2,√

2x3, x12, x22, x32,√

2x1x2,√

2x1x3,√

2x2x3]T

(9)

Finding the Decision Function

w: maybe infinite variables

The dual problem: finite number of variables minα

1

TQα − eTα

subject to 0 ≤ αi ≤ C , i = 1, . . . , l yTα = 0,

where Qij = yiyjφ(xi)Tφ(xj) and e = [1, . . . , 1]T At optimum

w =Pl

i =1αiyiφ(xi)

(10)

Kernel Tricks

Qij = yiyjφ(xi)Tφ(xj) needs a closed form Example: xi ∈ R3, φ(xi) ∈ R10

φ(xi) = [1,√

2(xi)1,√

2(xi)2,√

2(xi)3, (xi)21, (xi)22, (xi)23,√

2(xi)1(xi)2,√

2(xi)1(xi)3,√

2(xi)2(xi)3]T Then φ(xi)Tφ(xj) = (1 + xTi xj)2.

Kernel: K (x, y) = φ(x)Tφ(y); common kernels:

e−γkxi−xjk2, (Radial Basis Function) (xTi xj/a + b)d (Polynomial kernel)

(11)

Can be inner product in infinite dimensional space Assume x ∈ R1 and γ > 0.

e−γkxi−xjk2 = e−γ(xi−xj)2 = e−γxi2+2γxixj−γxj2

=e−γxi2−γxj2 1 + 2γxixj

1! + (2γxixj)2

2! + (2γxixj)3

3! + · · ·

=e−γxi2−γxj2 1 · 1+

r2γ 1!xi ·

r2γ 1!xj+

r(2γ)2 2! xi2 ·

r(2γ)2 2! xj2 +

r(2γ)3 3! xi3 ·

r(2γ)3

3! xj3 + · · · = φ(xi)Tφ(xj), where

φ(x ) = e−γx2

 1,

r2γ 1!x ,

r(2γ)2 2! x2,

r(2γ)3

3! x3, · · ·

T

.

(12)

Issues

So what kind of kernel should I use?

What kind of functions are valid kernels?

How to decide kernel parameters?

Some of these issues will be discussed later

(13)

Decision function

At optimum

w =Pl

i =1αiyiφ(xi) Decision function

wTφ(x) + b

=

l

X

i =1

αiyiφ(xi)Tφ(x) + b

=

l

X

i =1

αiyiK (xi, x) + b

(14)

Support Vectors: More Important Data

Only φ(xi) of αi > 0 used ⇒ support vectors

-0.2 0 0.2 0.4 0.6 0.8 1 1.2

-1.5 -1 -0.5 0 0.5 1

(15)

We have roughly shown basic ideas of SVM A 3-D demonstration

http://www.csie.ntu.edu.tw/~cjlin/

libsvmtools/svmtoy3d More about dual problems

We omit detailed derivations here

Quite a few people think that for any optimization problem

⇒ Lagrangian dual exists Wrong! We usually need

Convex programming; Constraint qualification We have them for SVM

(16)

Outline

Basic concepts: SVM and kernels Training SVM

Practical use of SVM Multi-class classification

Research directions: large-scale training Research directions: linear SVM

Research directions: others Conclusions

(17)

Large Dense Quadratic Programming

minα

1

TQα − eTα

subject to 0 ≤ αi ≤ C , i = 1, . . . , l yTα = 0

Qij 6= 0, Q : an l by l fully dense matrix 50,000 training points: 50,000 variables:

(50, 0002 × 8/2) bytes = 10GB RAM to store Q Traditional optimization methods:

(18)

Decomposition Methods

Working on some variables each time (e.g., Osuna et al., 1997; Joachims, 1998; Platt, 1998)

Similar to coordinate-wise minimization Working set B , N = {1, . . . , l }\B fixed Sub-problem at the kth iteration:

minαB

1

2αTBkN)TQBB QBN

QNB QNN

 αB αkN



eTB (ekN)TαB αkN



subject to 0 ≤ αt ≤ C , t ∈ B, yBTαB = −yTNαkN

(19)

Avoid Memory Problems

The new objective function 1

TBQBBαB + (−eB + QBNαkN)TαB + constant Only B columns of Q needed (|B| ≥ 2)

Calculated when used Trade time for space

(20)

How Decomposition Methods Perform?

Convergence not very fast

But, no need to have very accurate α Prediction not affected much

In some situations, # support vectors  # training points

Initial α1 = 0, some instances never used

(21)

An example of training 50,000 instances using LIBSVM

$svm-train -c 16 -g 4 -m 400 22features Total nSV = 3370

Time 79.524s

On a Xeon 2.0G machine

Calculating the whole Q takes more time

#SVs = 3,370  50,000

A good case where some remain at zero all the time

(22)

Issues of Decomposition Methods

Techniques for faster decomposition methods store recently used kernel elements working set size/selection

theoretical issues: convergence

and others (details not discussed here) Major software:

LIBSVM

http://www.csie.ntu.edu.tw/~cjlin/libsvm SVMlight

http://svmlight.joachims.org

(23)

Caching and Shrinking

Speed up decomposition methods Caching (Joachims, 1998)

Store recently used kernel columns in computer memory

100K Cache

$ time ./svm-train -m 0.01 -g 0.01 a6a 13.62s

40M Cache

$ time ./svm-train -m 40 -g 0.01 a6a 11.40s

(24)

Shrinking (Joachims, 1998)

Some elements remain bounded until the end Heuristically resized to a smaller problem See -h 1 option in LIBSVM software

After certain iterations, most bounded elements identified and not changed (Lin, 2002)

So caching and shrinking are useful

(25)

Outline

Basic concepts: SVM and kernels Training SVM

Practical use of SVM Multi-class classification

Research directions: large-scale training Research directions: linear SVM

Research directions: others Conclusions

(26)

Let’s Try a Practical Example

A problem from astroparticle physics

1 1:2.61e+01 2:5.88e+01 3:-1.89e-01 4:1.25e+02 1 1:5.70e+01 2:2.21e+02 3:8.60e-02 4:1.22e+02 1 1:1.72e+01 2:1.73e+02 3:-1.29e-01 4:1.25e+02 ...

0 1:2.39e+01 2:3.89e+01 3:4.70e-01 4:1.25e+02 0 1:2.23e+01 2:2.26e+01 3:2.11e-01 4:1.01e+02 0 1:1.64e+01 2:3.92e+01 3:-9.91e-02 4:3.24e+01 Training and testing sets available: 3,089 and 4,000

Sparse format: zero values not stored

(27)

The Story Behind this Data Set

User:

I am using libsvm in a astroparticle physics application .. First, let me congratulate you to a really easy to use and nice package. Unfortunately, it gives me astonishingly bad results...

OK. Please send us your data

I am able to get 97% test accuracy. Is that good enough for you ?

User:

(28)

Training and Testing

Training

$./svm-train train.1

optimization finished, #iter = 6131 nSV = 3053, nBSV = 724

Total nSV = 3053 Testing

$./svm-predict test.1 train.1.model test.1.out Accuracy = 66.925% (2677/4000)

nSV and nBSV: number of SVs and bounded SVs (αi = C ).

(29)

Why this Fails

After training, nearly 100% support vectors Training and testing accuracy different

$./svm-predict train.1 train.1.model o Accuracy = 99.7734% (3082/3089)

Most kernel elements:

Kij = e−kxi−xjk2/4 (

= 1 if i = j ,

→ 0 if i 6= j . Some features in rather large ranges

(30)

Data Scaling

Without scaling

Attributes in greater numeric ranges may dominate Linearly scale each feature to [0, 1] by:

feature value − min max − min , There are other ways

Scaling generally helps, but not always

(31)

Data Scaling: Same Factors

A common mistake

$./svm-scale -l -1 -u 1 train.1 > train.1.scale

$./svm-scale -l -1 -u 1 test.1 > test.1.scale Same factor on training and testing

$./svm-scale -s range1 train.1 > train.1.scale

$./svm-scale -r range1 test.1 > test.1.scale

(32)

After Data Scaling

Train scaled data and then predict

$./svm-train train.1.scale

$./svm-predict test.1.scale train.1.scale.model test.1.predict

Accuracy = 96.15%

Training accuracy now is

$./svm-predict train.1.scale train.1.scale.model o Accuracy = 96.439%

Default parameter: C = 1, γ = 0.25

(33)

Different Parameters

If we use C = 20, γ = 400

$./svm-train -c 20 -g 400 train.1.scale

$./svm-predict train.1.scale train.1.scale.model o Accuracy = 100% (3089/3089)

100% training accuracy but

$./svm-predict test.1.scale train.1.scale.model o Accuracy = 82.7% (3308/4000)

Very bad test accuracy Overfitting happens

(34)

Overfitting

In theory

You can easily achieve 100% training accuracy This is useless

When training and predicting a data, we should Avoid underfitting: small training error

Avoid overfitting: small testing error

(35)

l and s: training; and 4: testing

(36)

Parameter Selection

Need to select suitable parameters C and kernel parameters

Example:

γ of e−γkxi−xjk2 a, b, d of (xTi xj/a + b)d How to select them?

So performance better?

(37)

Performance Evaluation

Available data ⇒ training and validation Train the training; test the validation k-fold cross validation (CV):

Data randomly separated to k groups

Each time k − 1 as training and one as testing Select parameters/kernels with best CV result

(38)

Selecting Kernels

RBF, polynomial, or others?

For beginners, use RBF first

Linear kernel: special case of RBF

Performance of linear the same as RBF under certain parameters (Keerthi and Lin, 2003) Polynomial: numerical difficulties

(< 1)d → 0, (> 1)d → ∞ More parameters than RBF

(39)

A Simple Procedure

1. Conduct simple scaling on the data 2. Consider RBF kernel K (x, y) = e−γkx−yk2

3. Use cross-validation to find the best parameter C and γ

4. Use the best C and γ to train the whole training set 5. Test

For beginners only, you can do a lot more

(40)

Contour of Parameter Selection

(41)

The good region of parameters is quite large SVM is sensitive to parameters, but not that sensitive

Sometimes default parameters work

but it’s good to select them if time is allowed

(42)

Outline

Basic concepts: SVM and kernels Training SVM

Practical use of SVM Multi-class classification

Research directions: large-scale training Research directions: linear SVM

Research directions: others Conclusions

(43)

Multi-class Classification

k classes

One-against-the rest: Train k binary SVMs:

1st class vs. (2, · · · , k)th class 2nd class vs. (1, 3, . . . , k)th class

...

k decision functions

(w1)Tφ(x) + b1 ...

(wk)Tφ(x) + b

(44)

Prediction:

arg max

j (wj)Tφ(x) + bj

Reason: If x ∈ 1st class, then we should have (w1)Tφ(x) + b1 ≥ +1

(w2)Tφ(x) + b2 ≤ −1 ...

(wk)Tφ(x) + bk ≤ −1

(45)

Multi-class Classification (Cont’d)

One-against-one: train k(k − 1)/2 binary SVMs (1, 2), (1, 3), . . . , (1, k), (2, 3), (2, 4), . . . , (k − 1, k) If 4 classes ⇒ 6 binary SVMs

yi = 1 yi = −1 Decision functions class 1 class 2 f12(x) = (w12)Tx + b12 class 1 class 3 f13(x) = (w13)Tx + b13 class 1 class 4 f14(x) = (w14)Tx + b14 class 2 class 3 f23(x) = (w23)Tx + b23 class 2 class 4 f24(x) = (w24)Tx + b24 class 3 class 4 f34(x) = (w34)Tx + b34

(46)

For a testing data, predicting all binary SVMs Classes winner

1 2 1

1 3 1

1 4 1

2 3 2

2 4 4

3 4 3

Select the one with the largest vote class 1 2 3 4

# votes 3 1 1 1 May use decision values as well

(47)

More Complicated Forms

Solving a single optimization problem (Weston and Watkins, 1999; Crammer and Singer, 2002; Lee et al., 2004)

There are many other methods A comparison in Hsu and Lin (2002)

RBF kernel: accuracy similar for different methods But 1-against-1 fastest for training

(48)

Outline

Basic concepts: SVM and kernels Training SVM

Practical use of SVM Multi-class classification

Research directions: large-scale training Research directions: linear SVM

Research directions: others Conclusions

(49)

SVM doesn’t Scale Up

Yes, if using kernels

Training millions of data is time consuming

Cases with many support vectors: quadratic time bottleneck on

QSV, SV

For noisy data: # SVs increases linearly in data size (Steinwart, 2003)

Some solutions Parallelization Approximation

(50)

Parallelization

Multi-core/Shared Memory/GPU

• One line change of LIBSVM

Multicore Shared-memory

1 80 1 100

2 48 2 57

4 32 4 36

8 27 8 28

50,000 data (kernel evaluations: 80% time)

• GPU (Catanzaro et al., 2008) Distributed Environments

• Chang et al. (2007); Zanni et al. (2006); Zhu et al.

(2009). All use MPI; reasonably good speed-up

(51)

Approximately Training SVM

Can be done in many aspects Data level: sub-sampling Optimization level:

Approximately solve the quadratic program Other non-intuitive but effective ways I will show one today

Many papers have addressed this issue

(52)

Approximately Training SVM (Cont’d)

Subsampling

Simple and often effective More advanced techniques

Incremental training: (e.g., Syed et al., 1999)) Data ⇒ 10 parts

train 1st part ⇒ SVs, train SVs + 2nd part, . . . Select and train good points: KNN or heuristics For example, Bakır et al. (2005)

(53)

Approximately Training SVM (Cont’d)

Approximate the kernel; e.g., Fine and Scheinberg (2001); Williams and Seeger (2001)

Use part of the kernel; e.g., Lee and Mangasarian (2001); Keerthi et al. (2006)

Early stopping of optimization algorithms Tsang et al. (2005) and others

And many more

Some simple but some sophisticated

(54)

Approximately Training SVM (Cont’d)

Sophisticated techniques may not be always useful Sometimes slower than sub-sampling

covtype: 500k training and 80k testing rcv1: 550k training and 14k testing

covtype rcv1

Training size Accuracy Training size Accuracy

50k 92.5% 50k 97.2%

100k 95.3% 100k 97.4%

500k 98.2% 550k 97.8%

(55)

Approximately Training SVM (Cont’d)

Sophisticated techniques may not be always useful Sometimes slower than sub-sampling

covtype: 500k training and 80k testing rcv1: 550k training and 14k testing

covtype rcv1

Training size Accuracy Training size Accuracy

50k 92.5% 50k 97.2%

100k 95.3% 100k 97.4%

500k 98.2% 550k 97.8%

(56)

Discussion: Large-scale Training

We don’t have many large and well labeled sets Expensive to obtain true labels

Specific properties of data should be considered We will illustrate this point using linear SVM The design of software for very large data sets should be application different

(57)

Outline

Basic concepts: SVM and kernels Training SVM

Practical use of SVM Multi-class classification

Research directions: large-scale training Research directions: linear SVM

Research directions: others Conclusions

(58)

Linear SVM

Data not mapped to another space

w,b,ξmin 1

2wTw + C

l

X

i =1

ξi

subject to yi(wTxi + b) ≥ 1 −ξi, ξi ≥ 0, i = 1, . . . , l .

In theory, RBF kernel with certain parameters ⇒ as good as linear (Keerthi and Lin, 2003):

Test accuracy of linear ≤ Test accuracy of RBF But can be an approximation to nonlinear

Recently linear SVM an important research topic

(59)

Linear SVM for Large Document Sets

Bag of words model (TF-IDF or others) A large # of features

Accuracy similar with/without mapping vectors What if training is much faster?

A very effective approximation to nonlinear SVM

(60)

A Comparison: LIBSVM and LIBLINEAR

rcv1: # data: > 600k, # features: > 40k Using LIBSVM (linear kernel): > 10 hours Using LIBLINEAR (same stopping condition) Computation: < 5 seconds; I/O: 60 seconds Accuracy similar to nonlinear; more than 100x speedup

Training millions of data in a few seconds

See some results in Hsieh et al. (2008) by running LIBLINEAR

http:

//www.csie.ntu.edu.tw/~cjlin/liblinear

(61)

Testing Accuracy versus Training Time

news20 yahoo-japan

(62)

Why Training Linear SVM Is Faster?

In optimization, each iteration often needs

if (α) = (Qα)i − 1 Nonlinear SVM

if (α) = Xl

j =1yiyjK (xi, xjj − 1 cost: O(nl ); n: # features, l : # data

Linear: use w ≡ Xl

j =1yjαjxj and ∇if (α) = yiwTxi − 1 Only O(n) cost if w is maintained

(63)

Extension: Training Explicit Form of Nonlinear Mappings

Linear-SVM method to train φ(x1), . . . , φ(xl) Kernel not used

Applicable only if dimension of φ(x) not too large Low-degree Polynomial Mappings

K (xi, xj) = (xTi xj + 1)2 = φ(xi)Tφ(xj) φ(x) = [1,√

2x1, . . . ,√

2xn, x12, . . . , xn2,

2x1x2, . . . ,

2xn−1xn]T

(64)

Testing Accuracy and Training Time

Data set

Degree-2 Polynomial Accuracy diff.

Training time (s)

Accuracy Linear RBF LIBLINEAR LIBSVM

a9a 1.6 89.8 85.06 0.07 0.02

real-sim 59.8 1,220.5 98.00 0.49 0.10

ijcnn1 10.7 64.2 97.84 5.63 −0.85

MNIST38 8.6 18.4 99.29 2.47 −0.40

covtype 5,211.9 NA 80.09 3.74 −15.98

webspam 3,228.1 NA 98.44 5.29 −0.76

Training φ(xi) by linear: faster than kernel, but sometimes competitive accuracy

(65)

Discussion: Directly Train φ(x

i

), ∀i

See details in our work (Chang et al., 2010) A related development: Sonnenburg and Franc (2010)

Useful for certain applications

(66)

Linear Classification: Data Larger than Memory

Existing methods cannot easily handle this situation See our recent KDD work (Yu et al., 2010)

KDD 2010 best paper award

Training several million data (or more) on your laptop

(67)

Linear Classification: Online Learning

For extremely large data, cannot keep all data

After using new data to update the model; may not need them any more

Online learning instead of offline learning

Training often by stochastic gradient descent methods

They use only a subset of data at each step Now an important research topic (e.g.,

Shalev-Shwartz et al., 2007; Langford et al., 2009;

Bordes et al., 2009)

(68)

Linear Classification: L1 Regularization

1-norm versus 2-norm

kwk1 = |w1| + · · · + |wn|, kwk22 = w12 + · · · + wn2

w

|w |

w w2

2-norm: all wi are non-zeros; 1-norm: some wi may be zeros; useful for feature selection

Recently a hot topic; see our survey (Yuan et al., 2010)

(69)

Outline

Basic concepts: SVM and kernels Training SVM

Practical use of SVM Multi-class classification

Research directions: large-scale training Research directions: linear SVM

Research directions: others Conclusions

(70)

Multiple Kernel Learning (MKL)

How about using

t1K1 + t2K2 + · · · + trKr, where t1 + · · · + tr = 1 as the kernel

Related to parameter/kernel selection

If K1 better ⇒ t1 close to 1, others close to 0 Earlier development (Lanckriet et al., 2004): high computational cost

Many subsequent works (e.g., Rakotomamonjy et al., 2008).

Still ongoing; so far MKL has not been a practical tool yet

(71)

Ranking

Labels become ranking information e.g., x1 ranks higher than x2

RankSVM (Joachims, 2002): add constraint wTxi ≥ wTxj + ξij if xi ranks better than xj Many subsequent works

However, whether SVM is the most suitable method for ranking is an issue

(72)

Other Directions

Semi-supervised learning

Use information from unlabeled data Active learning

Needs cost to obtain labels of data Cost sensitive learning

For unbalanced data Structured Learning

Data instance not an Euclidean vector Maybe a parse tree of a sentence Feature selection

(73)

Outline

Basic concepts: SVM and kernels Training SVM

Practical use of SVM Multi-class classification

Research directions: large-scale training Research directions: linear SVM

Research directions: others Conclusions

(74)

Discussion and Conclusions

SVM and kernel methods are rather mature areas But still quite a few interesting research issues Many are extensions of standard classification (e.g., semi-supervised learning)

It is possible to identify more extensions through real applications

(75)

References I

G. H. Bakır, L. Bottou, and J. Weston. Breaking svm complexity with cross-training. In L. K.

Saul, Y. Weiss, and L. Bottou, editors, Advances in Neural Information Processing Systems 17, pages 81–88. MIT Press, Cambridge, MA, 2005.

A. Bordes, L. Bottou, and P. Gallinari. SGD-QN: Careful quasi-Newton stochastic gradient descent. Journal of Machine Learning Research, 10:1737–1754, 2009.

B. E. Boser, I. Guyon, and V. Vapnik. A training algorithm for optimal margin classifiers. In Proceedings of the Fifth Annual Workshop on Computational Learning Theory, pages 144–152. ACM Press, 1992.

B. Catanzaro, N. Sundaram, and K. Keutzer. Fast support vector machine training and classification on graphics processors. In Proceedings of the Twenty Fifth International Conference on Machine Learning (ICML), 2008.

E. Chang, K. Zhu, H. Wang, H. Bai, J. Li, Z. Qiu, and H. Cui. Parallelizing support vector machines on distributed computers. In NIPS 21, 2007.

Y.-W. Chang, C.-J. Hsieh, K.-W. Chang, M. Ringgaard, and C.-J. Lin. Training and testing low-degree polynomial data mappings via linear SVM. Journal of Machine Learning Research, 11:1471–1490, 2010. URL

http://www.csie.ntu.edu.tw/~cjlin/papers/lowpoly_journal.pdf.

(76)

References II

K. Crammer and Y. Singer. On the learnability and design of output codes for multiclass problems. Machine Learning, (2–3):201–233, 2002.

S. Fine and K. Scheinberg. Efficient svm training using low-rank kernel representations.

Journal of Machine Learning Research, 2:243–264, 2001.

C.-J. Hsieh, K.-W. Chang, C.-J. Lin, S. S. Keerthi, and S. Sundararajan. A dual coordinate descent method for large-scale linear SVM. In Proceedings of the Twenty Fifth International Conference on Machine Learning (ICML), 2008. URL

http://www.csie.ntu.edu.tw/~cjlin/papers/cddual.pdf.

C.-W. Hsu and C.-J. Lin. A comparison of methods for multi-class support vector machines.

IEEE Transactions on Neural Networks, 13(2):415–425, 2002.

T. Joachims. Making large-scale SVM learning practical. In B. Sch¨olkopf, C. J. C. Burges, and A. J. Smola, editors, Advances in Kernel Methods - Support Vector Learning, Cambridge, MA, 1998. MIT Press.

T. Joachims. Optimizing search engines using clickthrough data. In Proceedings of the eighth ACM SIGKDD international conference on Knowledge discovery and data mining, 2002.

S. S. Keerthi and C.-J. Lin. Asymptotic behaviors of support vector machines with Gaussian kernel. Neural Computation, 15(7):1667–1689, 2003.

S. S. Keerthi, O. Chapelle, and D. DeCoste. Building support vector machines with reduced classifier complexity. Journal of Machine Learning Research, 7:1493–1515, 2006.

(77)

References III

G. Lanckriet, N. Cristianini, P. Bartlett, L. El Ghaoui, and M. Jordan. Learning the Kernel Matrix with Semidefinite Programming. Journal of Machine Learning Research, 5:27–72, 2004.

J. Langford, L. Li, and T. Zhang. Sparse online learning via truncated gradient. Journal of Machine Learning Research, 10:771–801, 2009.

Y. Lee, Y. Lin, and G. Wahba. Multicategory support vector machines. Journal of the American Statistical Association, 99(465):67–81, 2004.

Y.-J. Lee and O. L. Mangasarian. RSVM: Reduced support vector machines. In Proceedings of the First SIAM International Conference on Data Mining, 2001.

C.-J. Lin. A formal analysis of stopping criteria of decomposition methods for support vector machines. IEEE Transactions on Neural Networks, 13(5):1045–1052, 2002. URL http://www.csie.ntu.edu.tw/~cjlin/papers/stop.ps.gz.

E. Osuna, R. Freund, and F. Girosi. Training support vector machines: An application to face detection. In Proceedings of CVPR’97, pages 130–136, New York, NY, 1997. IEEE.

J. C. Platt. Fast training of support vector machines using sequential minimal optimization. In B. Sch¨olkopf, C. J. C. Burges, and A. J. Smola, editors, Advances in Kernel Methods - Support Vector Learning, Cambridge, MA, 1998. MIT Press.

A. Rakotomamonjy, F. Bach, S. Canu, and Y. Grandvalet. SimpleMKL. Journal of Machine

(78)

References IV

S. Shalev-Shwartz, Y. Singer, and N. Srebro. Pegasos: primal estimated sub-gradient solver for SVM. In Proceedings of the Twenty Fourth International Conference on Machine Learning (ICML), 2007.

S. Sonnenburg and V. Franc. COFFIN : A computational framework for linear SVMs. In Proceedings of the Twenty Seventh International Conference on Machine Learning (ICML), 2010.

I. Steinwart. Sparseness of support vector machines. Journal of Machine Learning Research, 4:

1071–1105, 2003.

N. A. Syed, H. Liu, and K. K. Sung. Incremental learning with support vector machines. In Workshop on Support Vector Machines, IJCAI99, 1999.

I. Tsang, J. Kwok, and P. Cheung. Core vector machines: Fast SVM training on very large data sets. Journal of Machine Learning Research, 6:363–392, 2005.

J. Weston and C. Watkins. Multi-class support vector machines. In M. Verleysen, editor, Proceedings of ESANN99, pages 219–224, Brussels, 1999. D. Facto Press.

C. K. I. Williams and M. Seeger. Using the Nystr¨om method to speed up kernel machines. In T. Leen, T. Dietterich, and V. Tresp, editors, Neural Information Processing Systems 13, pages 682–688. MIT Press, 2001.

(79)

References V

H.-F. Yu, C.-J. Hsieh, K.-W. Chang, and C.-J. Lin. Large linear classification when data cannot fit in memory. In Proceedings of the 16th ACM SIGKDD International Conference on Knowledge Discovery and Data Mining, 2010. URL

http://www.csie.ntu.edu.tw/~cjlin/papers/kdd_disk_decomposition.pdf.

G.-X. Yuan, K.-W. Chang, C.-J. Hsieh, and C.-J. Lin. A comparison of optimization methods and software for large-scale l1-regularized linear classification. Journal of Machine Learning Research, 2010. URL http://www.csie.ntu.edu.tw/~cjlin/papers/l1.pdf. To appear.

L. Zanni, T. Serafini, and G. Zanghirati. Parallel software for training large scale support vector machines on multiprocessor systems. Journal of Machine Learning Research, 7:

1467–1492, 2006.

Z. A. Zhu, W. Chen, G. Wang, C. Zhu, and Z. Chen. P-packSVM: Parallel primal gradient descent kernel SVM. In Proceedings of the 2009 edition of the IEEE International Conference on Data Mining, 2009.

參考文獻

相關文件

1 Embedding Numerous Features: Kernel Models Lecture 1: Linear Support Vector Machine.

2 Distributed classification algorithms Kernel support vector machines Linear support vector machines Parallel tree learning.. 3 Distributed clustering

2 Distributed classification algorithms Kernel support vector machines Linear support vector machines Parallel tree learning?. 3 Distributed clustering

Keywords Support vector machine · ε-insensitive loss function · ε-smooth support vector regression · Smoothing Newton algorithm..

support vector machine, ε-insensitive loss function, ε-smooth support vector regression, smoothing Newton algorithm..

英國人,自我教育的學者。獨 立發現四元數可簡化為空間向 量,而仍然滿足電磁學的需求. Heaviside

“Transductive Inference for Text Classification Using Support Vector Machines”, Proceedings of ICML-99, 16 th International Conference on Machine Learning, pp.200-209. Coppin

Solving SVM Quadratic Programming Problem Training large-scale data..