• 沒有找到結果。

Machine Learning Techniques (ᘤᢈ)

N/A
N/A
Protected

Academic year: 2022

Share "Machine Learning Techniques (ᘤᢈ)"

Copied!
147
0
0

加載中.... (立即查看全文)

全文

(1)

Machine Learning Techniques

( 機器學習技法)

Lecture 7: Blending and Bagging

Hsuan-Tien Lin (林軒田) htlin@csie.ntu.edu.tw

Department of Computer Science

& Information Engineering

National Taiwan University

( 國立台灣大學資訊工程系)

Hsuan-Tien Lin (NTU CSIE) Machine Learning Techniques 0/23

(2)

Blending and Bagging

Roadmap

1 Embedding Numerous Features: Kernel Models

Lecture 6: Support Vector Regression kernel ridge regression

(dense) via ridge regression +

representer theorem;

support vector regression

(sparse) via regularized

tube

error +

Lagrange dual

2

Combining Predictive Features: Aggregation Models

Lecture 7: Blending and Bagging

Motivation of Aggregation Uniform Blending

Linear and Any Blending

Bagging (Bootstrap Aggregation)

3 Distilling Implicit Features: Extraction Models

(3)

Blending and Bagging Motivation of Aggregation

An Aggregation Story

Your T friends g

1

, · · · ,g

T

predicts whether stock will go up as g

t

(x).

You can . . .

select

the most trust-worthy friend from their

usual performance

—validation!

mix

the predictions from all your friends

uniformly

—let them

vote!

mix

the predictions from all your friends

non-uniformly

—let them vote, but

give some more ballots

combine

the predictions

conditionally

—if

[t satisfies some condition]

give some ballots to friend t

...

aggregation

models:

mix

or

combine

hypotheses (for better performance)

Hsuan-Tien Lin (NTU CSIE) Machine Learning Techniques 2/23

(4)

Blending and Bagging Motivation of Aggregation

An Aggregation Story

Your T friends g

1

, · · · ,g

T

predicts whether stock will go up as g

t

(x).

You can . . .

select

the most trust-worthy friend from their

usual performance

—validation!

mix

the predictions from all your friends

uniformly

—let them

vote!

mix

the predictions from all your friends

non-uniformly

—let them vote, but

give some more ballots

combine

the predictions

conditionally

—if

[t satisfies some condition]

give some ballots to friend t

...

aggregation

models:

mix

or

combine

hypotheses (for better performance)

(5)

Blending and Bagging Motivation of Aggregation

An Aggregation Story

Your T friends g

1

, · · · ,g

T

predicts whether stock will go up as g

t

(x).

You can . . .

select

the most trust-worthy friend from their

usual performance

—validation!

mix

the predictions from all your friends

uniformly

—let them

vote!

mix

the predictions from all your friends

non-uniformly

—let them vote, but

give some more ballots

combine

the predictions

conditionally

—if

[t satisfies some condition]

give some ballots to friend t

...

aggregation

models:

mix

or

combine

hypotheses (for better performance)

Hsuan-Tien Lin (NTU CSIE) Machine Learning Techniques 2/23

(6)

Blending and Bagging Motivation of Aggregation

An Aggregation Story

Your T friends g

1

, · · · ,g

T

predicts whether stock will go up as g

t

(x).

You can . . .

select

the most trust-worthy friend from their

usual performance

—validation!

mix

the predictions from all your friends

uniformly

—let them

vote!

mix

the predictions from all your friends

non-uniformly

—let them vote, but

give some more ballots

combine

the predictions

conditionally

—if

[t satisfies some condition]

give some ballots to friend t

...

aggregation

models:

mix

or

combine

hypotheses (for better performance)

(7)

Blending and Bagging Motivation of Aggregation

An Aggregation Story

Your T friends g

1

, · · · ,g

T

predicts whether stock will go up as g

t

(x).

You can . . .

select

the most trust-worthy friend from their

usual performance

—validation!

mix

the predictions from all your friends

uniformly

—let them

vote!

mix

the predictions from all your friends

non-uniformly

—let them vote, but

give some more ballots

combine

the predictions

conditionally

—if

[t satisfies some condition]

give some ballots to friend t

...

aggregation

models:

mix

or

combine

hypotheses (for better performance)

Hsuan-Tien Lin (NTU CSIE) Machine Learning Techniques 2/23

(8)

Blending and Bagging Motivation of Aggregation

An Aggregation Story

Your T friends g

1

, · · · ,g

T

predicts whether stock will go up as g

t

(x).

You can . . .

select

the most trust-worthy friend from their

usual performance

—validation!

mix

the predictions from all your friends

uniformly

—let them

vote!

mix

the predictions from all your friends

non-uniformly

—let them vote, but

give some more ballots

combine

the predictions

conditionally

—if

[t satisfies some condition]

give some ballots to friend t

...

aggregation

models:

mix

or

combine

hypotheses (for better performance)

(9)

Blending and Bagging Motivation of Aggregation

An Aggregation Story

Your T friends g

1

, · · · ,g

T

predicts whether stock will go up as g

t

(x).

You can . . .

select

the most trust-worthy friend from their

usual performance

—validation!

mix

the predictions from all your friends

uniformly

—let them

vote!

mix

the predictions from all your friends

non-uniformly

—let them vote, but

give some more ballots

combine

the predictions

conditionally

—if

[t satisfies some condition]

give some ballots to friend t

...

aggregation

models:

mix

or

combine

hypotheses (for better performance)

Hsuan-Tien Lin (NTU CSIE) Machine Learning Techniques 2/23

(10)

Blending and Bagging Motivation of Aggregation

An Aggregation Story

Your T friends g

1

, · · · ,g

T

predicts whether stock will go up as g

t

(x).

You can . . .

select

the most trust-worthy friend from their

usual performance

—validation!

mix

the predictions from all your friends

uniformly

—let them

vote!

mix

the predictions from all your friends

non-uniformly

—let them vote, but

give some more ballots

combine

the predictions

conditionally

—if

[t satisfies some condition]

give some ballots to friend t

...

aggregation

models:

mix

or

combine

hypotheses (for better performance)

(11)

Blending and Bagging Motivation of Aggregation

Aggregation with Math Notations

Your T friends g

1

, · · · ,g

T

predicts whether stock will go up as g

t

(x).

select

the most trust-worthy friend from their

usual performance

G(x) = g

t

(x) with t

=argmin

t∈{1,2,··· ,T } E val (g t )

mix

the predictions from all your friends

uniformly

G(x) = sign P

T

t=1 1

· g

t

(x)

mix

the predictions from all your friends

non-uniformly

G(x) = sign P

T

t=1 α t

· g

t

(x)

with

α t ≥ 0

• include select: α

t

= q

E

val

(g

t

) smallest y

• include uniformly: α

t

=

1

combine

the predictions

conditionally

G(x) = sign P

T

t=1 q t (x)

· g

t

(x)

with

q t (x) ≥ 0

• include non-uniformly: q

t

(x) =

α

t

aggregation models: a

rich family

Hsuan-Tien Lin (NTU CSIE) Machine Learning Techniques 3/23

(12)

Blending and Bagging Motivation of Aggregation

Aggregation with Math Notations

Your T friends g

1

, · · · ,g

T

predicts whether stock will go up as g

t

(x).

select

the most trust-worthy friend from their

usual performance

G(x) = g

t

(x) with t

=argmin

t∈{1,2,··· ,T } E val (g t )

mix

the predictions from all your friends

uniformly

G(x) = sign P

T

t=1 1

· g

t

(x)

mix

the predictions from all your friends

non-uniformly

G(x) = sign P

T

t=1 α t

· g

t

(x)

with

α t ≥ 0

• include select: α

t

= q

E

val

(g

t

) smallest y

• include uniformly: α

t

=

1

combine

the predictions

conditionally

G(x) = sign P

T

t=1 q t (x)

· g

t

(x)

with

q t (x) ≥ 0

• include non-uniformly: q

t

(x) =

α

t

aggregation models: a

rich family

(13)

Blending and Bagging Motivation of Aggregation

Aggregation with Math Notations

Your T friends g

1

, · · · ,g

T

predicts whether stock will go up as g

t

(x).

select

the most trust-worthy friend from their

usual performance

G(x) = g

t

(x) with t

=argmin

t∈{1,2,··· ,T } E val (g t )

mix

the predictions from all your friends

uniformly

G(x) = sign

P

T

t=1 1

· g

t

(x)



mix

the predictions from all your friends

non-uniformly

G(x) = sign P

T

t=1 α t

· g

t

(x)

with

α t ≥ 0

• include select: α

t

= q

E

val

(g

t

) smallest y

• include uniformly: α

t

=

1

combine

the predictions

conditionally

G(x) = sign P

T

t=1 q t (x)

· g

t

(x)

with

q t (x) ≥ 0

• include non-uniformly: q

t

(x) =

α

t

aggregation models: a

rich family

Hsuan-Tien Lin (NTU CSIE) Machine Learning Techniques 3/23

(14)

Blending and Bagging Motivation of Aggregation

Aggregation with Math Notations

Your T friends g

1

, · · · ,g

T

predicts whether stock will go up as g

t

(x).

select

the most trust-worthy friend from their

usual performance

G(x) = g

t

(x) with t

=argmin

t∈{1,2,··· ,T } E val (g t )

mix

the predictions from all your friends

uniformly

G(x) = sign

P

T

t=1 1

· g

t

(x)



mix

the predictions from all your friends

non-uniformly

G(x) = sign

P

T

t=1 α t

· g

t

(x)

with

α t ≥ 0

• include select: α

t

= q

E

val

(g

t

) smallest y

• include uniformly: α

t

=

1

combine

the predictions

conditionally

G(x) = sign P

T

t=1 q t (x)

· g

t

(x)

with

q t (x) ≥ 0

• include non-uniformly: q

t

(x) =

α

t

aggregation models: a

rich family

(15)

Blending and Bagging Motivation of Aggregation

Aggregation with Math Notations

Your T friends g

1

, · · · ,g

T

predicts whether stock will go up as g

t

(x).

select

the most trust-worthy friend from their

usual performance

G(x) = g

t

(x) with t

=argmin

t∈{1,2,··· ,T } E val (g t )

mix

the predictions from all your friends

uniformly

G(x) = sign

P

T

t=1 1

· g

t

(x)



mix

the predictions from all your friends

non-uniformly

G(x) = sign

P

T

t=1 α t

· g

t

(x)

with

α t ≥ 0

• include select: α

t

= q

E

val

(g

t

) smallest y

• include uniformly: α

t

=

1

combine

the predictions

conditionally

G(x) = sign P

T

t=1 q t (x)

· g

t

(x)

with

q t (x) ≥ 0

• include non-uniformly: q

t

(x) =

α

t

aggregation models: a

rich family

Hsuan-Tien Lin (NTU CSIE) Machine Learning Techniques 3/23

(16)

Blending and Bagging Motivation of Aggregation

Aggregation with Math Notations

Your T friends g

1

, · · · ,g

T

predicts whether stock will go up as g

t

(x).

select

the most trust-worthy friend from their

usual performance

G(x) = g

t

(x) with t

=argmin

t∈{1,2,··· ,T } E val (g t )

mix

the predictions from all your friends

uniformly

G(x) = sign

P

T

t=1 1

· g

t

(x)



mix

the predictions from all your friends

non-uniformly

G(x) = sign

P

T

t=1 α t

· g

t

(x)

with

α t ≥ 0

• include select: α

t

= q

E

val

(g

t

) smallest y

• include uniformly: α

t

=

1

combine

the predictions

conditionally

G(x) = sign P

T

t=1 q t (x)

· g

t

(x)

with

q t (x) ≥ 0

• include non-uniformly: q

t

(x) =

α

t

aggregation models: a

rich family

(17)

Blending and Bagging Motivation of Aggregation

Aggregation with Math Notations

Your T friends g

1

, · · · ,g

T

predicts whether stock will go up as g

t

(x).

select

the most trust-worthy friend from their

usual performance

G(x) = g

t

(x) with t

=argmin

t∈{1,2,··· ,T } E val (g t )

mix

the predictions from all your friends

uniformly

G(x) = sign

P

T

t=1 1

· g

t

(x)



mix

the predictions from all your friends

non-uniformly

G(x) = sign

P

T

t=1 α t

· g

t

(x)

with

α t ≥ 0

• include select: α

t

= q

E

val

(g

t

) smallest y

• include uniformly: α

t

= 1

combine

the predictions

conditionally

G(x) = sign P

T

t=1 q t (x)

· g

t

(x)

with

q t (x) ≥ 0

• include non-uniformly: q

t

(x) =

α

t

aggregation models: a

rich family

Hsuan-Tien Lin (NTU CSIE) Machine Learning Techniques 3/23

(18)

Blending and Bagging Motivation of Aggregation

Aggregation with Math Notations

Your T friends g

1

, · · · ,g

T

predicts whether stock will go up as g

t

(x).

select

the most trust-worthy friend from their

usual performance

G(x) = g

t

(x) with t

=argmin

t∈{1,2,··· ,T } E val (g t )

mix

the predictions from all your friends

uniformly

G(x) = sign

P

T

t=1 1

· g

t

(x)



mix

the predictions from all your friends

non-uniformly

G(x) = sign

P

T

t=1 α t

· g

t

(x)

with

α t ≥ 0

• include select: α

t

= q

E

val

(g

t

) smallest y

• include uniformly: α

t

= 1

combine

the predictions

conditionally

G(x) = sign

P

T

t=1 q t (x)

· g

t

(x)

with

q t (x) ≥ 0

• include non-uniformly: q

t

(x) =

α

t

aggregation models: a

rich family

(19)

Blending and Bagging Motivation of Aggregation

Aggregation with Math Notations

Your T friends g

1

, · · · ,g

T

predicts whether stock will go up as g

t

(x).

select

the most trust-worthy friend from their

usual performance

G(x) = g

t

(x) with t

=argmin

t∈{1,2,··· ,T } E val (g t )

mix

the predictions from all your friends

uniformly

G(x) = sign

P

T

t=1 1

· g

t

(x)



mix

the predictions from all your friends

non-uniformly

G(x) = sign

P

T

t=1 α t

· g

t

(x)

with

α t ≥ 0

• include select: α

t

= q

E

val

(g

t

) smallest y

• include uniformly: α

t

= 1

combine

the predictions

conditionally

G(x) = sign

P

T

t=1 q t (x)

· g

t

(x)

with

q t (x) ≥ 0

• include non-uniformly: q

t

(x) =

α

t

aggregation models: a

rich family

Hsuan-Tien Lin (NTU CSIE) Machine Learning Techniques 3/23

(20)

Blending and Bagging Motivation of Aggregation

Aggregation with Math Notations

Your T friends g

1

, · · · ,g

T

predicts whether stock will go up as g

t

(x).

select

the most trust-worthy friend from their

usual performance

G(x) = g

t

(x) with t

=argmin

t∈{1,2,··· ,T } E val (g t )

mix

the predictions from all your friends

uniformly

G(x) = sign

P

T

t=1 1

· g

t

(x)



mix

the predictions from all your friends

non-uniformly

G(x) = sign

P

T

t=1 α t

· g

t

(x)

with

α t ≥ 0

• include select: α

t

= q

E

val

(g

t

) smallest y

• include uniformly: α

t

= 1

combine

the predictions

conditionally

G(x) = sign

P

T

t=1 q t (x)

· g

t

(x)

with

q t (x) ≥ 0

• include non-uniformly: q

t

(x) = α

t

aggregation models: a

rich family

(21)

Blending and Bagging Motivation of Aggregation

Aggregation with Math Notations

Your T friends g

1

, · · · ,g

T

predicts whether stock will go up as g

t

(x).

select

the most trust-worthy friend from their

usual performance

G(x) = g

t

(x) with t

=argmin

t∈{1,2,··· ,T } E val (g t )

mix

the predictions from all your friends

uniformly

G(x) = sign

P

T

t=1 1

· g

t

(x)



mix

the predictions from all your friends

non-uniformly

G(x) = sign

P

T

t=1 α t

· g

t

(x)

with

α t ≥ 0

• include select: α

t

= q

E

val

(g

t

) smallest y

• include uniformly: α

t

= 1

combine

the predictions

conditionally

G(x) = sign

P

T

t=1 q t (x)

· g

t

(x)

with

q t (x) ≥ 0

• include non-uniformly: q

t

(x) = α

t

aggregation models: a

rich family

Hsuan-Tien Lin (NTU CSIE) Machine Learning Techniques 3/23

(22)

Blending and Bagging Motivation of Aggregation

Recall: Selection by Validation

G(x) = g

t

(x) with t

= argmin

t∈{1,2,··· ,T }

E val (g t )

simple

and popular

what if use E

in

(g

t

)instead of

E val (g t )?

complexity price on d

VC

, remember? :-)

need

one strong

g

t

to guarantee small

E val

(and small E

out

)

selection:

rely on one strong hypothesis

aggregation:

can we do better with many (possibly weaker) hypotheses?

(23)

Blending and Bagging Motivation of Aggregation

Recall: Selection by Validation

G(x) = g

t

(x) with t

= argmin

t∈{1,2,··· ,T }

E val (g t )

simple

and popular

what if use E

in

(g

t

)instead of

E val (g t )?

complexity price on d

VC

, remember? :-)

need

one strong

g

t

to guarantee small

E val

(and small E

out

)

selection:

rely on one strong hypothesis

aggregation:

can we do better with many (possibly weaker) hypotheses?

Hsuan-Tien Lin (NTU CSIE) Machine Learning Techniques 4/23

(24)

Blending and Bagging Motivation of Aggregation

Recall: Selection by Validation

G(x) = g

t

(x) with t

= argmin

t∈{1,2,··· ,T }

E val (g t )

simple

and popular

what if use E

in

(g

t

)instead of

E val (g t )?

complexity price on d

VC

, remember? :-)

need

one strong

g

t

to guarantee small

E val

(and small E

out

)

selection:

rely on one strong hypothesis

aggregation:

can we do better with many (possibly weaker) hypotheses?

(25)

Blending and Bagging Motivation of Aggregation

Recall: Selection by Validation

G(x) = g

t

(x) with t

= argmin

t∈{1,2,··· ,T }

E val (g t )

simple

and popular

what if use E

in

(g

t

)instead of

E val (g t )?

complexity price on d

VC

, remember? :-)

need

one strong

g

t

to guarantee small

E val

(and small E

out

)

selection:

rely on one strong hypothesis

aggregation:

can we do better with many (possibly weaker) hypotheses?

Hsuan-Tien Lin (NTU CSIE) Machine Learning Techniques 4/23

(26)

Blending and Bagging Motivation of Aggregation

Recall: Selection by Validation

G(x) = g

t

(x) with t

= argmin

t∈{1,2,··· ,T }

E val (g t )

simple

and popular

what if use E

in

(g

t

)instead of

E val (g t )?

complexity price on d

VC

, remember? :-)

need

one strong

g

t

to guarantee small

E val

(and small E

out

)

selection:

rely on one strong hypothesis

aggregation:

can we do better with many (possibly weaker) hypotheses?

(27)

Blending and Bagging Motivation of Aggregation

Recall: Selection by Validation

G(x) = g

t

(x) with t

= argmin

t∈{1,2,··· ,T }

E val (g t )

simple

and popular

what if use E

in

(g

t

)instead of

E val (g t )?

complexity price on d

VC

, remember? :-)

need

one strong

g

t

to guarantee small

E val

(and small E

out

)

selection:

rely on one strong hypothesis

aggregation:

can we do better with many (possibly weaker) hypotheses?

Hsuan-Tien Lin (NTU CSIE) Machine Learning Techniques 4/23

(28)

Blending and Bagging Motivation of Aggregation

Why Might Aggregation Work?

mix

different weak hypotheses

uniformly

—G(x) ‘strong’

aggregation

=⇒

feature transform (?)

mix

different random-PLA hypotheses

uniformly

—G(x) ‘moderate’

aggregation

=⇒

regularization (?)

proper aggregation =⇒

better performance

(29)

Blending and Bagging Motivation of Aggregation

Why Might Aggregation Work?

mix

different weak hypotheses

uniformly

—G(x) ‘strong’

aggregation

=⇒

feature transform (?)

mix

different random-PLA hypotheses

uniformly

—G(x) ‘moderate’

aggregation

=⇒

regularization (?)

proper aggregation =⇒

better performance

Hsuan-Tien Lin (NTU CSIE) Machine Learning Techniques 5/23

(30)

Blending and Bagging Motivation of Aggregation

Why Might Aggregation Work?

mix

different weak hypotheses

uniformly

—G(x) ‘strong’

aggregation

=⇒

feature transform (?)

mix

different random-PLA hypotheses

uniformly

—G(x) ‘moderate’

aggregation

=⇒

regularization (?)

proper aggregation =⇒

better performance

(31)

Blending and Bagging Motivation of Aggregation

Why Might Aggregation Work?

mix

different weak hypotheses

uniformly

—G(x) ‘strong’

aggregation

=⇒

feature transform (?)

mix

different random-PLA hypotheses

uniformly

—G(x) ‘moderate’

aggregation

=⇒

regularization (?)

proper aggregation =⇒

better performance

Hsuan-Tien Lin (NTU CSIE) Machine Learning Techniques 5/23

(32)

Blending and Bagging Motivation of Aggregation

Why Might Aggregation Work?

mix

different weak hypotheses

uniformly

—G(x) ‘strong’

aggregation

=⇒

feature transform (?)

mix

different random-PLA hypotheses

uniformly

—G(x) ‘moderate’

aggregation

=⇒

regularization (?)

proper aggregation =⇒

better performance

(33)

Blending and Bagging Motivation of Aggregation

Why Might Aggregation Work?

mix

different weak hypotheses

uniformly

—G(x) ‘strong’

aggregation

=⇒

feature transform (?)

mix

different random-PLA hypotheses

uniformly

—G(x) ‘moderate’

aggregation

=⇒

regularization (?)

proper aggregation =⇒

better performance

Hsuan-Tien Lin (NTU CSIE) Machine Learning Techniques 5/23

(34)

Blending and Bagging Motivation of Aggregation

Why Might Aggregation Work?

mix

different weak hypotheses

uniformly

—G(x) ‘strong’

aggregation

=⇒

feature transform (?)

mix

different random-PLA hypotheses

uniformly

—G(x) ‘moderate’

aggregation

=⇒

regularization (?)

proper aggregation =⇒

better performance

(35)

Blending and Bagging Motivation of Aggregation

Fun Time

Consider three decision stump hypotheses from R to {−1, +1}:

g

1

(x ) = sign(1 − x ), g

2

(x ) = sign(1 + x ), g

3

(x ) = −1. When mixing the three hypotheses uniformly, what is the resulting G(x )?

1

2J|x | ≤ 1K − 1

2

2J|x | ≥ 1K − 1

3

2Jx ≤ −1K − 1

4

2Jx ≥ +1K − 1

Reference Answer: 1

The ‘region’ that gets two positive votes from g

1

and g

2

is |x | ≤ 1, and thus G(x ) is positive within the region only. We see that the three decision stumps g

t

can be aggregated to form a more sophisticated hypothesis G.

Hsuan-Tien Lin (NTU CSIE) Machine Learning Techniques 6/23

(36)

Blending and Bagging Motivation of Aggregation

Fun Time

Consider three decision stump hypotheses from R to {−1, +1}:

g

1

(x ) = sign(1 − x ), g

2

(x ) = sign(1 + x ), g

3

(x ) = −1. When mixing the three hypotheses uniformly, what is the resulting G(x )?

1

2J|x | ≤ 1K − 1

2

2J|x | ≥ 1K − 1

3

2Jx ≤ −1K − 1

4

2Jx ≥ +1K − 1

Reference Answer: 1

The ‘region’ that gets two positive votes from g

1

and g

2

is |x | ≤ 1, and thus G(x ) is positive within the region only. We see that the three decision stumps g

t

can be aggregated to form a more sophisticated hypothesis G.

(37)

Blending and Bagging Uniform Blending

Uniform Blending (Voting) for Classification

uniform

blending: known g t

, each with

1

ballot

G(x) = sign

T

X

t=1

1 · g

t

(x)

!

same

g t

(autocracy): as good as one single

g t

very different

g t

(diversity+

democracy):

majority can

correct

minority

similar results with uniform voting for multiclass

G(x) = argmax

1≤k ≤K T

X

t=1

Jg

t

(x) = kK

how about

regression?

Hsuan-Tien Lin (NTU CSIE) Machine Learning Techniques 7/23

(38)

Blending and Bagging Uniform Blending

Uniform Blending (Voting) for Classification

uniform blending: known g t

, each with

1

ballot

G(x) = sign

T

X

t=1

1 · g

t

(x)

!

same

g t

(autocracy): as good as one single

g t

very different

g t

(diversity+

democracy):

majority can

correct

minority

similar results with uniform voting for multiclass

G(x) = argmax

1≤k ≤K T

X

t=1

Jg

t

(x) = kK

how about

regression?

(39)

Blending and Bagging Uniform Blending

Uniform Blending (Voting) for Classification

uniform blending: known g t

, each with

1

ballot

G(x) = sign

T

X

t=1

1 · g

t

(x)

!

same

g t

(autocracy): as good as one single

g t

very different

g t

(diversity+

democracy):

majority can

correct

minority

similar results with uniform voting for multiclass

G(x) = argmax

1≤k ≤K T

X

t=1

Jg

t

(x) = kK

how about

regression?

Hsuan-Tien Lin (NTU CSIE) Machine Learning Techniques 7/23

(40)

Blending and Bagging Uniform Blending

Uniform Blending (Voting) for Classification

uniform blending: known g t

, each with

1

ballot

G(x) = sign

T

X

t=1

1 · g

t

(x)

!

same

g t

(autocracy):

as good as one single

g t

very different

g t

(diversity+

democracy):

majority can

correct

minority

similar results with uniform voting for multiclass

G(x) = argmax

1≤k ≤K T

X

t=1

Jg

t

(x) = kK

how about

regression?

(41)

Blending and Bagging Uniform Blending

Uniform Blending (Voting) for Classification

uniform blending: known g t

, each with

1

ballot

G(x) = sign

T

X

t=1

1 · g

t

(x)

!

same

g t

(autocracy):

as good as one single

g t

very different

g t

(diversity+

democracy):

majority can

correct

minority

similar results with uniform voting for multiclass

G(x) = argmax

1≤k ≤K T

X

t=1

Jg

t

(x) = kK

how about

regression?

Hsuan-Tien Lin (NTU CSIE) Machine Learning Techniques 7/23

(42)

Blending and Bagging Uniform Blending

Uniform Blending (Voting) for Classification

uniform blending: known g t

, each with

1

ballot

G(x) = sign

T

X

t=1

1 · g

t

(x)

!

same

g t

(autocracy):

as good as one single

g t

very different

g t

(diversity+

democracy):

majority can

correct

minority

similar results with uniform voting for multiclass

G(x) = argmax

1≤k ≤K T

X

t=1

Jg

t

(x) = kK

how about

regression?

(43)

Blending and Bagging Uniform Blending

Uniform Blending (Voting) for Classification

uniform blending: known g t

, each with

1

ballot

G(x) = sign

T

X

t=1

1 · g

t

(x)

!

same

g t

(autocracy):

as good as one single

g t

very different

g t

(diversity+

democracy):

majority can

correct

minority

similar results with uniform voting for multiclass

G(x) = argmax

1≤k ≤K T

X

t=1

Jg

t

(x) = kK

how about

regression?

Hsuan-Tien Lin (NTU CSIE) Machine Learning Techniques 7/23

(44)

Blending and Bagging Uniform Blending

Uniform Blending for Regression

G(x) =

1 T

T

X

t=1

g t

(x)

same

g t

(autocracy): as good as one single

g t

very different

g t

(diversity+

democracy):

=⇒

some

g t

(x) > f (x), some

g t

(x) < f (x)

=⇒average

could be

more accurate than individual

diverse hypotheses:

even simple

uniform blending

can be better than any

single hypothesis

(45)

Blending and Bagging Uniform Blending

Uniform Blending for Regression

G(x) = 1 T

T

X

t=1

g t

(x)

same

g t

(autocracy): as good as one single

g t

very different

g t

(diversity+

democracy):

=⇒

some

g t

(x) > f (x), some

g t

(x) < f (x)

=⇒average

could be

more accurate than individual

diverse hypotheses:

even simple

uniform blending

can be better than any

single hypothesis

Hsuan-Tien Lin (NTU CSIE) Machine Learning Techniques 8/23

(46)

Blending and Bagging Uniform Blending

Uniform Blending for Regression

G(x) = 1 T

T

X

t=1

g t

(x)

same

g t

(autocracy):

as good as one single

g t

very different

g t

(diversity+

democracy):

=⇒

some

g t

(x) > f (x), some

g t

(x) < f (x)

=⇒average

could be

more accurate than individual

diverse hypotheses:

even simple

uniform blending

can be better than any

single hypothesis

(47)

Blending and Bagging Uniform Blending

Uniform Blending for Regression

G(x) = 1 T

T

X

t=1

g t

(x)

same

g t

(autocracy):

as good as one single

g t

very different

g t

(diversity+

democracy):

=⇒

some

g t

(x) > f (x), some

g t

(x) < f (x)

=⇒average

could be

more accurate than individual

diverse hypotheses:

even simple

uniform blending

can be better than any

single hypothesis

Hsuan-Tien Lin (NTU CSIE) Machine Learning Techniques 8/23

(48)

Blending and Bagging Uniform Blending

Uniform Blending for Regression

G(x) = 1 T

T

X

t=1

g t

(x)

same

g t

(autocracy):

as good as one single

g t

very different

g t

(diversity+

democracy):

=⇒

some

g t

(x) > f (x), some

g t

(x) < f (x)

=⇒average

could be

more accurate than individual

diverse hypotheses:

even simple

uniform blending

can be better than any

single hypothesis

(49)

Blending and Bagging Uniform Blending

Uniform Blending for Regression

G(x) = 1 T

T

X

t=1

g t

(x)

same

g t

(autocracy):

as good as one single

g t

very different

g t

(diversity+

democracy):

=⇒

some

g t

(x) > f (x), some

g t

(x) < f (x)

=⇒average

could be

more accurate than individual

diverse hypotheses:

even simple

uniform blending

can be better than any

single hypothesis

Hsuan-Tien Lin (NTU CSIE) Machine Learning Techniques 8/23

(50)

Blending and Bagging Uniform Blending

Theoretical Analysis of Uniform Blending

G(x)

=

1 T

T

X

t=1

g t (x)

avg (g

t

(x) − f (x))

2



= avg

g

2t

− 2g

t

f + f

2



= avg g

2t



− 2Gf + f

2

= avg g

2t



− G

2

+ (

G − f

)

2

= avg g

2t



− 2G

2

+ G

2

+ (G − f )

2

= avg g

2t

− 2g

t

G

+ G

2

 + (G − f )

2

=

avg

(g

t

− G)

2



+ (G − f )

2

avg

(E

out

(g

t

)) =

avg



E(g

t

G) 2



+E

out

(G)

avg

E(g

t

G) 2



+E

out

(G)

(51)

Blending and Bagging Uniform Blending

Theoretical Analysis of Uniform Blending

G(x)

=

1 T

T

X

t=1

g t (x)

avg (g

t

(x) − f (x))

2



= avg

g

2t

− 2g

t

f + f

2



= avg g

2t



− 2Gf + f

2

= avg g

2t



− G

2

+ (

G − f

)

2

= avg g

2t



− 2G

2

+ G

2

+ (G − f )

2

= avg g

2t

− 2g

t

G

+ G

2

 + (G − f )

2

=

avg

(g

t

− G)

2



+ (G − f )

2

avg

(E

out

(g

t

)) =

avg



E(g

t

G) 2



+E

out

(G)

avg

E(g

t

G) 2



+E

out

(G)

Hsuan-Tien Lin (NTU CSIE) Machine Learning Techniques 9/23

(52)

Blending and Bagging Uniform Blending

Theoretical Analysis of Uniform Blending

G(x)

=

1 T

T

X

t=1

g t (x)

avg (g

t

(x) − f (x))

2



= avg g

2t

− 2g

t

f + f

2



= avg g

2t



− 2Gf + f

2

= avg g

2t



− G

2

+ (

G − f

)

2

= avg g

2t



− 2G

2

+ G

2

+ (G − f )

2

= avg g

2t

− 2g

t

G

+ G

2

 + (G − f )

2

=

avg

(g

t

− G)

2



+ (G − f )

2

avg

(E

out

(g

t

)) =

avg



E(g

t

G) 2



+E

out

(G)

avg

E(g

t

G) 2



+E

out

(G)

(53)

Blending and Bagging Uniform Blending

Theoretical Analysis of Uniform Blending

G(x)

=

1 T

T

X

t=1

g t (x)

avg (g

t

(x) − f (x))

2



= avg g

2t

− 2g

t

f + f

2



= avg g

2t

 − 2Gf + f

2

= avg g

2t



− G

2

+ (

G − f

)

2

= avg g

2t



− 2G

2

+ G

2

+ (G − f )

2

= avg g

2t

− 2g

t

G

+ G

2

 + (G − f )

2

=

avg

(g

t

− G)

2



+ (G − f )

2

avg

(E

out

(g

t

)) =

avg



E(g

t

G) 2



+E

out

(G)

avg

E(g

t

G) 2



+E

out

(G)

Hsuan-Tien Lin (NTU CSIE) Machine Learning Techniques 9/23

(54)

Blending and Bagging Uniform Blending

Theoretical Analysis of Uniform Blending

G(x)

=

1 T

T

X

t=1

g t (x)

avg (g

t

(x) − f (x))

2



= avg g

2t

− 2g

t

f + f

2



= avg g

2t

 − 2Gf + f

2

= avg g

2t

 − G

2

+ (G − f )

2

= avg g

2t



− 2G

2

+ G

2

+ (G − f )

2

= avg g

2t

− 2g

t

G

+ G

2

 + (G − f )

2

=

avg

(g

t

− G)

2



+ (G − f )

2

avg

(E

out

(g

t

)) =

avg



E(g

t

G) 2



+E

out

(G)

avg

E(g

t

G) 2



+E

out

(G)

(55)

Blending and Bagging Uniform Blending

Theoretical Analysis of Uniform Blending

G(x)

=

1 T

T

X

t=1

g t (x)

avg (g

t

(x) − f (x))

2



= avg g

2t

− 2g

t

f + f

2



= avg g

2t

 − 2Gf + f

2

= avg g

2t

 − G

2

+ (G − f )

2

= avg g

2t

 − 2G

2

+ G

2

+ (G − f )

2

= avg g

2t

− 2g

t

G

+ G

2

 + (G − f )

2

=

avg

(g

t

− G)

2



+ (G − f )

2

avg

(E

out

(g

t

)) =

avg



E(g

t

G) 2



+E

out

(G)

avg

E(g

t

G) 2



+E

out

(G)

Hsuan-Tien Lin (NTU CSIE) Machine Learning Techniques 9/23

(56)

Blending and Bagging Uniform Blending

Theoretical Analysis of Uniform Blending

G(x)

=

1 T

T

X

t=1

g t (x)

avg (g

t

(x) − f (x))

2



= avg g

2t

− 2g

t

f + f

2



= avg g

2t

 − 2Gf + f

2

= avg g

2t

 − G

2

+ (G − f )

2

= avg g

2t

 − 2G

2

+ G

2

+ (G − f )

2

= avg g

2t

− 2g

t

G + G

2

 + (G − f )

2

= avg

(g

t

− G)

2

 + (G − f )

2

avg

(E

out

(g

t

)) =

avg



E(g

t

G) 2



+E

out

(G)

avg

E(g

t

G) 2



+E

out

(G)

(57)

Blending and Bagging Uniform Blending

Theoretical Analysis of Uniform Blending

G(x)

=

1 T

T

X

t=1

g t (x)

avg (g

t

(x) − f (x))

2



= avg g

2t

− 2g

t

f + f

2



= avg g

2t

 − 2Gf + f

2

= avg g

2t

 − G

2

+ (G − f )

2

= avg g

2t

 − 2G

2

+ G

2

+ (G − f )

2

= avg g

2t

− 2g

t

G + G

2

 + (G − f )

2

= avg (g

t

− G)

2

 + (G − f )

2

avg

(E

out

(g

t

)) =

avg



E(g

t

G) 2



+E

out

(G)

avg

E(g

t

G) 2



+E

out

(G)

Hsuan-Tien Lin (NTU CSIE) Machine Learning Techniques 9/23

(58)

Blending and Bagging Uniform Blending

Theoretical Analysis of Uniform Blending

G(x)

=

1 T

T

X

t=1

g t (x)

avg (g

t

(x) − f (x))

2



= avg g

2t

− 2g

t

f + f

2



= avg g

2t

 − 2Gf + f

2

= avg g

2t

 − G

2

+ (G − f )

2

= avg g

2t

 − 2G

2

+ G

2

+ (G − f )

2

= avg g

2t

− 2g

t

G + G

2

 + (G − f )

2

= avg (g

t

− G)

2

 + (G − f )

2

avg

(E

out

(g

t

)) =

avg



E(g

t

G) 2



+E

out

(G)

avg

E(g

t

G) 2



+E

out

(G)

(59)

Blending and Bagging Uniform Blending

Theoretical Analysis of Uniform Blending

G(x)

=

1 T

T

X

t=1

g t (x)

avg (g

t

(x) − f (x))

2



= avg g

2t

− 2g

t

f + f

2



= avg g

2t

 − 2Gf + f

2

= avg g

2t

 − G

2

+ (G − f )

2

= avg g

2t

 − 2G

2

+ G

2

+ (G − f )

2

= avg g

2t

− 2g

t

G + G

2

 + (G − f )

2

= avg (g

t

− G)

2

 + (G − f )

2

avg

(E

out

(g

t

)) =

avg



E(g

t

G) 2



+E

out

(G)

avg

E(g

t

G) 2



+E

out

(G)

Hsuan-Tien Lin (NTU CSIE) Machine Learning Techniques 9/23

(60)

Blending and Bagging Uniform Blending

Some Special g t

consider a

virtual

iterative process that for t = 1, 2, . . . , T

1

request size-N data D

t

from P

N

(i.i.d.)

2

obtain

g t

by A(D

t

)

g ¯

= lim

T →∞

G

=

T →∞

lim

1 T

T

X

t=1

g t

=

E

D A(D)

avg

(E

out

(g

t

)) =

avg



E(g

t

¯ g) 2



+E

out

(

g) ¯

expected

performance of A

=

expected deviation

to

consensus

+

performance of

consensus

performance of

consensus: called bias

• expected deviation

to

consensus: called variance

uniform blending:

reduces

variance

for more stable performance

(61)

Blending and Bagging Uniform Blending

Some Special g t

consider a

virtual

iterative process that for t = 1, 2, . . . , T

1

request size-N data D

t

from P

N

(i.i.d.)

2

obtain

g t

by A(D

t

)

g ¯

= lim

T →∞

G

=

T →∞

lim

1 T

T

X

t=1

g t

=

E

D A(D)

avg

(E

out

(g

t

)) =

avg



E(g

t

¯ g) 2



+E

out

(

g) ¯

expected

performance of A

=

expected deviation

to

consensus

+

performance of

consensus

performance of

consensus: called bias

• expected deviation

to

consensus: called variance

uniform blending:

reduces

variance

for more stable performance

Hsuan-Tien Lin (NTU CSIE) Machine Learning Techniques 10/23

(62)

Blending and Bagging Uniform Blending

Some Special g t

consider a

virtual

iterative process that for t = 1, 2, . . . , T

1

request size-N data D

t

from P

N

(i.i.d.)

2

obtain

g t

by A(D

t

)

g ¯

= lim

T →∞

G

=

T →∞

lim

1 T

T

X

t=1

g t

=

E

D A(D)

avg

(E

out

(g

t

)) =

avg



E(g

t

¯ g) 2



+E

out

(

g) ¯

expected

performance of A

=

expected deviation

to

consensus

+

performance of

consensus

performance of

consensus: called bias

• expected deviation

to

consensus: called variance

uniform blending:

reduces

variance

for more stable performance

(63)

Blending and Bagging Uniform Blending

Some Special g t

consider a

virtual

iterative process that for t = 1, 2, . . . , T

1

request size-N data D

t

from P

N

(i.i.d.)

2

obtain

g t

by A(D

t

)

g ¯

= lim

T →∞

G

=

T →∞

lim

1 T

T

X

t=1

g t

=

E

D A(D)

avg

(E

out

(g

t

)) =

avg



E(g

t

¯ g) 2



+E

out

(

g) ¯

expected

performance of A

=

expected deviation

to

consensus

+

performance of

consensus

performance of

consensus: called bias

• expected deviation

to

consensus: called variance

uniform blending:

reduces

variance

for more stable performance

Hsuan-Tien Lin (NTU CSIE) Machine Learning Techniques 10/23

(64)

Blending and Bagging Uniform Blending

Some Special g t

consider a

virtual

iterative process that for t = 1, 2, . . . , T

1

request size-N data D

t

from P

N

(i.i.d.)

2

obtain

g t

by A(D

t

)

g ¯

=

T →∞

lim

G

= lim

T →∞

1 T

T

X

t=1

g t

=

E

D A(D)

avg

(E

out

(g

t

)) =

avg



E(g

t

¯ g) 2



+E

out

(

g) ¯

expected

performance of A

=

expected deviation

to

consensus

+

performance of

consensus

performance of

consensus: called bias

• expected deviation

to

consensus: called variance

uniform blending:

reduces

variance

for more stable performance

(65)

Blending and Bagging Uniform Blending

Some Special g t

consider a

virtual

iterative process that for t = 1, 2, . . . , T

1

request size-N data D

t

from P

N

(i.i.d.)

2

obtain

g t

by A(D

t

)

g ¯

=

T →∞

lim

G

= lim

T →∞

1 T

T

X

t=1

g t

=

E

D A(D)

avg

(E

out

(g

t

)) =

avg



E(g

t

¯ g) 2



+E

out

(

g) ¯

expected

performance of A

=

expected deviation

to

consensus

+

performance of

consensus

performance of

consensus: called bias

• expected deviation

to

consensus: called variance

uniform blending:

reduces

variance

for more stable performance

Hsuan-Tien Lin (NTU CSIE) Machine Learning Techniques 10/23

(66)

Blending and Bagging Uniform Blending

Some Special g t

consider a

virtual

iterative process that for t = 1, 2, . . . , T

1

request size-N data D

t

from P

N

(i.i.d.)

2

obtain

g t

by A(D

t

)

g ¯

= lim

T →∞ G

= lim

T →∞

1 T

T

X

t=1

g t

=

E

D A(D)

avg

(E

out

(g

t

)) =

avg



E(g

t

¯ g) 2



+E

out

(

g) ¯

expected

performance of A

=

expected deviation

to

consensus

+

performance of

consensus

performance of

consensus: called bias

• expected deviation

to

consensus: called variance

uniform blending:

reduces

variance

for more stable performance

(67)

Blending and Bagging Uniform Blending

Some Special g t

consider a

virtual

iterative process that for t = 1, 2, . . . , T

1

request size-N data D

t

from P

N

(i.i.d.)

2

obtain

g t

by A(D

t

)

g ¯

= lim

T →∞ G

= lim

T →∞

1 T

T

X

t=1

g t

=

E

D A(D)

avg

(E

out

(g

t

)) =

avg



E(g

t

¯ g) 2



+E

out

(

g) ¯

expected

performance of A

=

expected deviation

to

consensus

+

performance of

consensus

performance of

consensus: called bias

• expected deviation

to

consensus: called variance

uniform blending:

reduces

variance

for more stable performance

Hsuan-Tien Lin (NTU CSIE) Machine Learning Techniques 10/23

(68)

Blending and Bagging Uniform Blending

Some Special g t

consider a

virtual

iterative process that for t = 1, 2, . . . , T

1

request size-N data D

t

from P

N

(i.i.d.)

2

obtain

g t

by A(D

t

)

g ¯

= lim

T →∞ G

= lim

T →∞

1 T

T

X

t=1

g t

=

E

D A(D)

avg

(E

out

(g

t

)) =

avg



E(g

t

¯ g) 2



+E

out

(

g) ¯ expected

performance of A =

expected deviation

to

consensus

+

performance of

consensus

performance of

consensus: called bias

• expected deviation

to

consensus: called variance

uniform blending:

reduces

variance

for more stable performance

參考文獻

相關文件

3 Distilling Implicit Features: Extraction Models Lecture 14: Radial Basis Function Network. RBF

Lecture 4: Soft-Margin Support Vector Machine allow some margin violations ξ n while penalizing them by C; equivalent to upper-bounding α n by C Lecture 5: Kernel Logistic

Hsuan-Tien Lin (NTU CSIE) Machine Learning Techniques 5/22.. Decision Tree Decision Tree Hypothesis. Disclaimers about

1 Embedding Numerous Features: Kernel Models Lecture 1: Linear Support Vector Machine.. linear SVM: more robust and solvable with quadratic programming Lecture 2: Dual Support

1 Embedding Numerous Features: Kernel Models Lecture 1: Linear Support Vector Machine.

Principle Component Analysis Denoising Auto Encoder Deep Neural Network... Deep Learning Optimization

For a deep NNet for written character recognition from raw pixels, which type of features are more likely extracted after the first hidden layer.

Hsuan-Tien Lin (NTU CSIE) Machine Learning Techniques 3/24.:. Deep Learning Deep