• 沒有找到結果。

Spoken Lecture Summarization by Random Walk

N/A
N/A
Protected

Academic year: 2022

Share "Spoken Lecture Summarization by Random Walk"

Copied!
45
0
0

加載中.... (立即查看全文)

全文

(1)

Author: Yun-Nung Chen, Yu Huang , Ching-Feng Yeh, and Lin-Shan Lee Speaker: Hung-Yi Lee

Spoken Lecture Summarization by Random Walk over a Graph Constructed with Automatically

Extracted Key Terms

National Taiwan University

(2)

Outline

O

Introduction

O

Graph-based summarization approach

 A spoken document is transformed into a graph structure

Nodes: sentences in a spoken document

Edge weight: topical similarities of sentences

 Random walk is used to select indicative sentences

 all sentences in a document can be jointly considered

O

Experiments

O

Conclusion

(3)

Introduction –

Extractive Summarization (1/2)

O

Extractive speech summarization

O

Select the indicative sentences in a spoken document

O

Cascade the sentences to form a summary

O

The number of sentences selected as summary is

decided by a predefined ratio

(4)

Introduction –

Extractive Summarization (2/2)

O

Each sentence S in a spoken document d is given an importance score I(S,d)

O

Select the indicative sentences based on I(S,d)

n

i

t

t t

t

S

1 2

   

    

n

i

i

d

t s d

1

] ,

[ ,

S

I    

sentence term

term statistical measure

Importance score

(5)

Introduction – PLSA

……

……

……

……

……

……

……

……

……

……

……

……

T1

t

1

t

2

t

3

t

i

t

M

S1

S2

Sj

SJ

Sentences Latent

Topics

Terms

T2

Tk

TK

P(Tk|Sj)

P(Tk|Sj): weight of latent topic Tk for sentence Sj

(6)

Proposed Approach (1/2)

• Basic idea

Not only the sentences with high importance score

based on statistical measure should be considered

as indicative sentence

(7)

Proposed Approach (1/2)

• Basic idea

Not only the sentences with high importance score based on statistical measure should be considered as indicative sentence

But the sentences topically similar to the indicative

sentences should also be considered as indicative

(8)

Proposed Approach (2/2)

• Graph-based approach

▫ Sentences in a spoken document are nodes on a graph, and topical similarities of sentences are weights of

edges.

▫ Use random walk to obtain new scores for summary selection

▫ → all sentences in the document can be jointly considered rather than individually.

(9)

Spoken Document d

S1

S1 S2 S3 S4 S5 S6

Graph Construction (1/2)

S2

S3 S4

S5

S6

Each sentence Si in the spoken document d is a node on the graph.

(10)

W( i , j ) (Si

Sj):

Topical similarity from sentence Si to Sj

(based on PLSA latent topics of sentences)

Graph Construction (1/2)

Spoken Document d

S1

S1 S2 S3 S4 S5

S6 S2

S3 S4

S5

S6 W( 3, 4): topical

similarities from sentence S3 to S4

(11)

O

Topical Similarity from sentences S

i

to S

j

O Edge weight W(i , j) (sentence Si → sentence Sj)

Graph Construction (2/2) - Topical Similarities

tj t1

tm

Si

T1 T2

Tk TK

Sj P(Tk|Si) P(Tk|Sj)

t2

ti t1

tn t2

 W( i , j ): evaluated by the latent topic similarities of sentences Si to Sj based on PLSA model

(12)

Find a set of new scores based on graph structure

{G(i) for each sentence Si in document d} which satisfies

         

 

i in S i

j

i j j

d

i 1 I S , G W ˆ ,

G  

Mathematical Formulation

 G(i) for sentence Si would be a new importance score for summary selection

(13)

Find a set of new scores based on graph structure

{G(i) for each sentence Si in document d} which satisfies

         

 

i in S i

j

i j j

d

i 1 I S , G W ˆ ,

G  

Mathematical Formulation

The original importance score of node Si

Scores propagate from other nodes to node Si ( weighted by 1-α ) ( weighted by α )

(14)

         

 

i in S i

j

i j j

d

i 1 I S , G W ˆ ,

G  

Mathematical Formulation

    

n

i

i

d

t s d

1

] ,

[ ,

S

I    

term statistical measure

Importance score

Find a set of new scores based on graph structure

{G(i) for each sentence Si in document d} which satisfies

(15)

         

 

i in S i

j

i j j

d

i 1 I S , G W ˆ ,

G  

Si

Mathematical Formulation

Sj

Find a set of new scores based on graph structure

{G(i) for each sentence Si in document d} which satisfies

(16)

         

 

i in S i

j

i j j

d

i 1 I S , G W ˆ ,

G  

Si

Mathematical Formulation

Sj

Find a set of new scores based on graph structure

{G(i) for each sentence Si in document d} which satisfies

(17)

         

 

i in S i

j

i j j

d

i 1 I S , G W ˆ ,

G  

   

 

 

j out

Sk j k

i i j

j

W ,

, , W

W ˆ

Mathematical Formulation

Si Sj

Find a set of new scores based on graph structure

{G(i) for each sentence Si in document d} which satisfies

(18)

         

 

i in S i

j

i j j

d

i 1 I S , G W ˆ ,

G  

Mathematical Formulation

Sj Si

Find a set of new scores based on graph structure

{G(i) for each sentence Si in document d} which satisfies

   

 

 

j out

Sk j k

i i j

j

W ,

, , W

W ˆ

The scores propagate from a node to all other

nodes sums to unity.

Sa

(19)

         

 

i in S i

j

i j j

d

i 1 I S , G Wˆ ,

G  

Mathematical Formulation

 G(i) can obtain higher score when 1) I(Si,d) is high.

2) More sentences topically similar to Si

(20)

         

 

i in S i

j

i j j

d

i 1 I S , G Wˆ ,

G  

Mathematical Formulation

 G(i) can obtain higher score when 1) I(Si,d) is high.

2) More sentences topically similar to Si

(21)

         

 

i in S i

j

i j j

d

i 1 I S , G Wˆ ,

G  

Mathematical Formulation

 All sentences in the documents are considered jointly

 Rather than individually

(22)

    

4 1- I S ,

    

6 ˆ 6,4

   

3 ˆ 3,4

G   4 d G W G W

Mathematical Formulation – an Example

S1

S2

S3 S4

S5

S6

Find G(1), G(2), G(3), G(4), G(5), G(6) such that

G(4) G(3)

G(6)

(23)

    

3 1- I S ,

    

1 ˆ 1,3

   

4 ˆ 4,3

G   3 d G W G W

    

4 1- I S ,

    

6 ˆ 6,4

   

3 ˆ 3,4

G   4 d G W G W

    

1 1- I S ,

    

2 ˆ 2,1

G   1 d G W

    

2 1- I S ,

    

3 ˆ 3,2

G   2 d G W

    

5 1- I S ,

    

4 ˆ 4,5

G   5 d G W

    

6 1- I S ,

    

4 ˆ 4,6

   

5 ˆ 5,6

G   6 d G W G W

Mathematical Formulation – Equations to be solved

S1

S2

S3 S4

S5

S6

Find G(1), G(2), G(3), G(4), G(5), G(6) such that

 How to solve these equations to obtain G(1), G(2), …… G(6)?

G(4) G(3)

G(6)

 solve the problem iteratively (random walk)

(24)

Random Walk Solution

S1

S2

S3 S4

S5

S6

G0(4) G0(3)

G0(1)

G0(2)

G0(5)

G0(6)

 Each sentence is assigned an initial value G0(i)

G0(i) = I(Si,d)

(25)

Random Walk Solution

S1

S2

S3 S4

S5

S6

G0(4) G0(3)

G0(1)

G0(2)

G0(5)

G0(6)

 Update the score for each sentence ……

    

4 1- I S ,

    

6 ˆ 6,4

   

3 ˆ 3,4

G1   4 d G0 W G0 W

(26)

Random Walk Solution

S1

S2

S3 S4

S5

S6

G1(4) G1(3)

G1(1)

G1(2)

G1(5)

G1(6)

 Update the score for each sentence ……

    

4 1- I S ,

    

6 ˆ 6,4

   

3 ˆ 3,4

G2   4 d G1 W G1 W

(27)

Random Walk Solution

S1

S2

S3 S4

S5

S6

G2(4) G2(3)

G2(1)

G2(2)

G2(5)

G2(6)

 The process is repeated ……

    

4 1- I S ,

    

6 ˆ 6,4

   

3 ˆ 3,4

G3   4 d G2 W G2 W

(28)

Random Walk Solution

S1

S2

S3 S4

S5

S6

G(4) G(3)

G(1)

G(2)

G(5)

G(6)

 The process is repeated ……

The score of each node would finally converge.

 According to the theory of random walk:

 The converged score G(i) is actually G(i) satisfying

         

 

i in S i

j

i j j

d

i 1 I S , G W ˆ ,

G  

(29)

I’(S

i

,d) = I(S

i

,d)

1- δ

G(i)

δ

New scores: Consider graph structure Original importance

score based on terms in the sentences For summary

selection

Graph-based Summarization Approach

         

 

i in S i

j

i j j

d

i 1 I S , G W ˆ ,

G  

Find a set of new scores based on graph structure

{G(i) for each sentence Si in document d} which satisfies

    

n

i

i

d t s d

1

] ,

[ ,

S

I    

term statistical measure

Importance score

(30)

Experimental Setup (1/2)

O Corpus: course offered in National Taiwan University

O Mandarin Chinese embedded by English words

O Single speaker

O 45.2 hours

O ASR System

O Bilingual AM with model adaptation [1]

O LM with adaptation using random forests [2]

Language Mandarin English Overall

Acc (%) 78.15 53.44 76.26

[1] Ching-Feng Yeh, et al., “Bilingual Acoustic Model Adaptation by Unit Merging on Different Levels and Cross-level Integration, ” Interspeech, 2011.

[2] Ching-Feng Yeh, et al. , “An Integrated Framework for Transcribing Mandarin-English Code-mixed Lectures with Improved Acoustic and Language Modeling,” ISCSLP, 2010.

(31)

Experimental Setup (2/2)

O

Spoken Documents

▫ We segmented the whole lecture into 155 documents by topic segmentation

▫ 34 documents out of the 155 were tested.

▫ The average length of each document was about 17.5 minutes

▫ Human produced reference summaries for each document

O

Evaluation

O ROUGE-1, ROUGE-2, ROUGE-3

O ROUGE-L: Longest Common Subsequence (LCS)

(32)

Experimental Results

41 46 51 56

10% 20% 30%

18 23 28

10% 20% 30%

9 14 19

10% 20% 30%

40 45 50

10% 20% 30%

ROUGE-1 ROUGE-2 ROUGE-3 ROUGE-L

Summarization ratio Summarization ratio Summarization ratio Summarization ratio

Baseline1 Baseline1 + Proposed (Graph)

Baseline1: I(Si,d) – importance score using latent topic entropy term statistical measure

Baseline1+Proposed: I(Si,d)G(i)

(33)

Experimental Results

41 46 51 56

10% 20% 30%

18 23 28

10% 20% 30%

9 14 19

10% 20% 30%

40 45 50

10% 20% 30%

ROUGE-1 ROUGE-2 ROUGE-3 ROUGE-L

Summarization ratio Summarization ratio Summarization ratio Summarization ratio

Baseline1 Baseline1 + Proposed (Graph)

 The proposed approach outperformed the first baseline in most cases.

(Compare blue and red bars)

(34)

Experimental Results

41 46 51 56

10% 20% 30%

18 23 28

10% 20% 30%

9 14 19

10% 20% 30%

40 45 50

10% 20% 30%

ROUGE-1 ROUGE-2 ROUGE-3 ROUGE-L

Baseline2 Baseline2 + Proposed (Graph)

Summarization ratio Summarization ratio Summarization ratio Summarization ratio 41

46 51 56

10% 20% 30%

18 23 28

10% 20% 30%

9 14 19

10% 20% 30%

40 45 50

10% 20% 30%

ROUGE-1 ROUGE-2 ROUGE-3 ROUGE-L

Summarization ratio Summarization ratio Summarization ratio Summarization ratio

Baseline1 Baseline1 + Proposed (Graph)

Baseline2: I(Si,d) – importance score using key-term based statistical measure

Baseline2+Proposed: I(Si,d)G(i)

(35)

Experimental Results

41 46 51 56

10% 20% 30%

18 23 28

10% 20% 30%

9 14 19

10% 20% 30%

40 45 50

10% 20% 30%

ROUGE-1 ROUGE-2 ROUGE-3 ROUGE-L

Baseline2 Baseline2 + Proposed (Graph)

Summarization ratio Summarization ratio Summarization ratio Summarization ratio 41

46 51 56

10% 20% 30%

18 23 28

10% 20% 30%

9 14 19

10% 20% 30%

40 45 50

10% 20% 30%

ROUGE-1 ROUGE-2 ROUGE-3 ROUGE-L

Summarization ratio Summarization ratio Summarization ratio Summarization ratio

Baseline1 Baseline1 + Proposed (Graph)

 The proposed approach always outperformed the second baseline.

(Compare green and orange bars)

(36)

Conclusions

• The performance of summarization can be improved by

▫ Graph-based approach considering topical similarity

 This offers a way to globally consider all sentences in a document for

summarization rather than considers each

sentence individually

(37)

37

Master Defense, National Taiwan University

(38)

Random Walk Solution

S1

S2

S3 S4

S5

S6

αG0(4) αG0(3)

αG0(1)

αG0(2)

αG0(5)

αG0(6)

(39)

Random Walk Solution

S1

S2

S3 S4

S5

S6

αG0(4) αG0(3)

αG0(1)

αG0(2)

αG0(5)

αG0(6)

   

 

 

j out

Sk j k

i i j

j

W ,

, , W

W ˆ

   

4 Wˆ 4,3

G0

   

4 Wˆ 4,6

G0

   

4 Wˆ 4,5

G0

(40)

Random Walk Solution

S1

S2

S3 S4

S5

S6

αG0(4) αG0(3)

αG0(1)

αG0(2)

αG0(5)

αG0(6)

   

6 Wˆ 6,4

G0

   

3 Wˆ 3,4

G0

(41)

Random Walk Solution

S1

S2

S3 S4

S5

S6

   

6 Wˆ 6,4

G0

   

3 Wˆ 3,4

G0

     

4 1- 4 G

   

3 Wˆ 3,4 G

   

6 Wˆ 6,4

G1    0  0

1-

  

4

(42)

         

 

i in S i

j

i j j

d

i 1 I S , G W ˆ ,

G  

Mathematical Formulation

Sj Si

Find a set of new scores based on graph structure

{G(i) for each sentence Si in document d} which satisfies

 

W ˆ   ,

1

Skout j j k

   

 

 

j out

Sk j k

i i j

j

W ,

, , W

W ˆ

(43)

         

 

i in S i

j

i j j

d

i 1 I S , G W ˆ ,

G  

Mathematical Formulation

Find a set of new scores based on graph structure

{G(i) for each sentence Si in document d} which satisfies

Sj Si

Sa

   

 

j i W

 

j a

W

i j j W

, ,

G ,

The amount of score Sj propagate to Si is

(44)

         

 

i in S i

j

i j j

d

i 1 I S , G W ˆ ,

G  

Mathematical Formulation

Find a set of new scores based on graph structure

{G(i) for each sentence Si in document d} which satisfies

Sj Si

Sa

   

 

j i W

 

j a

W

i j j W

, ,

G ,

The amount of score Sj propagate to Si is

   

 

j i W

 

j a

W

a j j W

, ,

G ,

The amount of score Sj propagate to Sa is

(45)

    

4 1- I S ,

    

6 ˆ 6,4

   

3 ˆ 3,4

G   4 d G W G W

   

 

3,2 3,4

 

3,4

4 , ˆ 3

W W

W W

 

6,4

 

6,4

ˆ W

W

Mathematical Formulation – an Example

S1

S2

S3 S4

S5

S6

Find G(1), G(2), G(3), G(4), G(5), G(6) such that

G(4) G(3)

G(6)

depends on S4 itself

Depends on topically similar sentences (S3 and S6)

參考文獻

相關文件

In this paper, we have studied a neural network approach for solving general nonlinear convex programs with second-order cone constraints.. The proposed neural network is based on

Regarding the importance of these aspects, employers generally perceived all aspects were of above average importance, with Management Skills receiving the lowest score of 3.75 and

The above information is for discussion and reference only and should not be treated as investment

As n increases, not only does the fixed locality bound of five become increasingly negligible relative to the size of the search space, but the probability that a random

Fayun’s annotation is according to Kumarajiva’s original translation, not only sentences by the strict branch demonstrates it to the Lotus Sutra, also aware of

As a result banks should be so proactive as if they are doing the marketing job to make their employees feel the importance of internal marketing, who can only in

Furthermore, based on the temperature calculation in the proposed 3D block-level thermal model and the final region, an iterative approach is proposed to reduce

By seminars and case participation of domestic and foreign experts and scholars, optimized development and content of elderly leisure activities will be proposed as not