• 沒有找到結果。

Spoken Lecture Summarization by Random Walk over a Graph Constructed with Automatically Extracted Key Terms

N/A
N/A
Protected

Academic year: 2022

Share "Spoken Lecture Summarization by Random Walk over a Graph Constructed with Automatically Extracted Key Terms"

Copied!
4
0
0

加載中.... (立即查看全文)

全文

(1)

Spoken Lecture Summarization by Random Walk over a Graph Constructed with Automatically Extracted Key Terms

Yun-Nung Chen

, Yu Huang

, Ching-Feng Yeh

#

, and Lin-Shan Lee

†#

Graduate Institute of Computer Science and Information Engineering

#

Graduate Institute of Communication Engineering

†#

National Taiwan University, Taiwan

vivian.ynchen@gmail.com, lslee@gate.sinica.edu.tw

Abstract

This paper proposes an improved approach for spoken lecture summarization, in which random walk is performed on a graph constructed with automatically extracted key terms and proba- bilistic latent semantic analysis (PLSA). Each sentence of the document is represented as a node of the graph and the edge be- tween two nodes is weighted by the topical similarity between the two sentences. The basic idea is that sentences topically similar to more important sentences should be more important.

In this way all sentences in the document can be jointly consid- ered more globally rather than individually. Experimental re- sults showed significant improvement in terms of ROUGE eval- uation.

Index Terms: summarization, course lecture, probabilistic la- tent semantic analysis (PLSA), random walk, key term

1. Introduction

In the Internet era, digital content over the network includes all the information and activities of human life. The most attractive form of network content is multimedia that may include speech.

Such speech information usually tells the subjects, topics, and core concepts of the content. However, multimedia/spoken doc- uments are just video/audio signals, usually much more difficult to retrieve and browse, because they cannot be easily displayed on the screen, and the user cannot simply ”skim through” each of them from the beginning to the end. Hence, spoken docu- ment summarization becomes very important [1].

Automatic summarization of spoken documents have been actively investigated. While most work was focused primarily on news content, recent effort has been increasingly directed to new domains such as lectures and meeting recordings [2, 3].

This paper takes course lecture as an example in the experi- ments. Many approaches selected a number of indicative sen- tences or passages from the original spoken documents accord- ing to a target summarization ratio, and sequenced them to form a summary. Some approaches tried to identify sentences carry- ing concepts closer to the complete documents [4]. The spo- ken document summarization actually carry intrinsic difficul- ties such as the recognition errors, problems with spontaneous speech, and lack of correct sentence or paragraph boundaries.

A general approach has been found very successful [5, 6], in which each sentence in the document d, S = t1t2...ti...tn, rep-

resented as a sequence of terms ti, is given an importance score:

I(S, d) = 1 n

n

X

i=1

1s(ti, d) + λ2l(ti)

3c(ti) + λ4g(ti)] + λ5b(S), (1) where s(ti, d), l(ti), c(ti), g(ti) are respectively some statisti- cal measure (such as TF-IDF), linguistic measure (e.g., differ- ent parts-of-speech (PoSs) are given different weights), confi- dence score and N-gram score for the term ti, and b(S) is cal- culated from the grammatical structure of the sentence S, and λ1, λ2, λ3, λ4and λ5are weighting parameters. For each doc- ument d, sentences used in the summary is then selected based on this importance score I(S, d).

Prior work showed that the topical information obtained by probabilistic latent semantic analysis (PLSA) was very useful in estimating the statistical measure s(ti, d) in (1) above to iden- tify the important sentences [7, 8]. Such topical information from PLSA is also used in this paper. We further rescore I(S, d) above in (1) by random walk over a graph to consider not only the importance of terms but the similarity between sentences in the whole document, so that sentences topically similar to more important sentences are given higher scores.

2. Proposed Approach

2.1. Probabilistic Latent Semantic Analysis (PLSA) PLSA [10] has been widely used to analyze the semantics of documents based on a set of latent topics. Given a set of documents {dj, j = 1, 2, ..., J } and all terms {ti, i = 1, 2, ..., M } they include, PLSA uses a set of latent topic variables, {Tk, k = 1, 2, ..., K}, to characterize the ”term- document” co-occurrence relationships. The probability of ob- serving a term tigiven a document djcan be parameterized by

P (ti| dj) =

K

X

k=1

P (ti| Tk)P (Tk| dj). (2)

The PLSA model can be optimized with EM algorithm by max- imizing a likelihood function [10].

2.1.1. Latent Topic Significance (LTS)

Latent Topic Significance (LTS) for a given term tiwith respect to a topic Tkcan be defined [7, 8] as

LTSti(Tk) = P

dj∈Dn(ti, dj)P (Tk| dj) P

dj∈Dn(ti, dj)[1 − P (Tk| dj)], (3)

(2)

where n(ti, dj) is the occurrence count of term tiin a document dj. In the numerator of (3), the count of the given term tiin each document dj, n(ti, dj), is weighted by the likelihood that the given topic Tkis addressed by the document dj, P (Tk| dj), and then summed over all documents djin the training corpus D. Therefore the numerator is the total count of the given term ti used for the given topic Tkover the whole training corpus, as estimated by PLSA model. The denominator is very similar except for latent topics other than Tk, so P (Tk| dj) is replaced by [1 − P (Tk | dj)]. Thus, a higher LTSti(Tk) indicates the term tiis more significant for the latent topic Tk.

2.1.2. Latent Topic Entropy (LTE)

Latent Topic Entropy (LTE), LTE(ti), for a given term tican be calculated as (4) from the topic distribution P (Tk | ti) for each term ti,

LTE(ti) = −

K

X

k=1

P (Tk| ti) log P (Tk| ti), (4)

where the topic distribution P (Tk | ti) can be estimated as follows [7, 8],

P (Tk| ti) = P (ti| Tk) · P (Tk)

P (ti) , (5)

where the probability P (Tk) is left out because a good approach to estimate it is not yet available, while P (ti) can be obtained from a large corpus. LTE(ti) is a measure of how the term tiis focused on a few topics, so a lower latent topic entropy implies the term carries more topical information.

2.2. Automatic Key Term Extraction

Key terms are the terms used in the documents carrying core concepts of the content. They are useful for indexing, retrieving, and browsing. There are in general two types of key terms: key- words (single words) and key phrases (such as ”hidden Markov model”). Automatically extracting key terms from spoken con- tent is still a difficult problem, but some initial approaches have been shown to be successful in recent experiments [9]. Such approaches include use of right/left branching entropy derived from PAT-Trees to extract frequently occurring patterns includ- ing two or more words, and identifying or verifying key terms (including key phrases) by prosodic (pitch, duration, energy), lexical, and semantic (from PLSA) features with unsupervised techniques or supervised training [9]. Here well show such au- tomatically extracted key terms are very helpful in summariza- tion.

2.3. Statistical Measures of a Term

Here in this work, the statistical measure of a term ti, s(ti, d) in (1) can be defined in two different ways below.

2.3.1. LTE-Based Statistical Measure

The statistical measure s(ti, d) in (1) can be defined based on LTE(ti) in (4) as

sLTE(ti, d) = βn(ti, d)

LTE(ti), (6)

where β is a scaling factor, so the score sLTE(ti, d) is inversely proportion to the latent topic entropy LTE(ti). Some previ- ous works [7, 8] showed that this measure outperformed the

very successful ”significance score” in speech summarization [6], and here we use sLTE(ti, d) as the baseline.

2.3.2. Key-Term-Based Statistical Measure

With automatically extracted key terms (with some errors) as mentioned above, we can estimate a new latent topic probability PKEY(Tk | d) hopefully better than P (Tk | d) directly from PLSA,

PKEY(Tk| d) = P

ti∈keyn(ti, d)P (Tk| ti) PK

k=1

P

ti∈keyn(ti, d)P (Tk| ti), (7) where key is the set of automatically extracted key terms, and P (Tk| ti) is in (5). Therefore only the automatically extracted key terms ti in d are considered, avoiding the influence from other insignificant terms. We can then define the statistical mea- sure s(ti, d) as

sKEY(ti, d) =

K

X

k=1

LTSti(Tk)PKEY(Tk| d), (8)

where the information of LTSti(Tk) in (3) is used here.

2.4. Random Walk on a Graph

We formulate the sentence selection problem as random walk on a directed graph, in which each sentence is a node and the edges between them are weighted by topical similarity. The basic idea is that a sentence similar to more important sentences should be more important [12, 13]. In this way all sentences in the document can be jointly considered more globally rather than individually. We define two directed edges between each pair of nodes with two directions, weighted by an asymmetric topical similarity between them, We then keep only the top N outgoing edges with the highest weights from each node, while consider incoming edges to each node for importance propagation in the graph. A simplified example for such a graph is in Figure 1, in which Ai and Biare the sets of neighbors of the node Si

connected respectively by outgoing and incoming edges.

2.4.1. Topical Similarity between Sentences

Within a document d, we can first compute the probability that the topic Tkis addressed by a sentence Si,

P (Tk| Si) = P

t∈Sin(t, Si)P (Tk| t) P

t∈Sin(t, Si) . (9) Then an asymmetric topical similarity sim(Si, Sj) for sen- tences Si to Sj (with direction Si → Sj) can be defined by accumulating LTSt(Tk) in (3) weighted by P (Tk| Si) for all terms t in Sjover all latent topics,

sim(Si, Sj) = X

t∈Sj K

X

k=1

LTSt(Tk)P (Tk| Si), (10)

We normalize this similarity by the total similarity summed over the top N sentences Skwith edges outgoing from Si, or the set Ai, to produce the weight p(i, j) for the edge from Sito Sjon the graph,

p(i, j) = sim(Si, Sj) P

Sk∈Aisim(Si, Sk). (11)

(3)

p(4, 3) S

1

S

3

S

4

S

5

S

6

S

2

p(3, 4)

A

4

= {S

3

, S

5

, S

6

} B

4

= {S

3

, S

6

}

Figure 1: A simplified example of the graph considered. Each sentence is represented as a node on the graph.AiandBiare the neighbors of the nodeSiconnected respectively by outgoing and incoming edges.

2.4.2. Random Walk

Random walk [12, 13] is now performed over the graph ob- tained above. v(i) is the new score for node Si, which is the interpolation of two scores, the normalized initial importance r(i) for node Siand the score contributed by all neighboring nodes Sjof node Siweighted by p(j, i),

v(i) = (1 − α)r(i) + α X

Sj∈Bi

p(j, i)v(j), (12)

where α is the interpolation weight, Biis the set of neighbors connected to node Sivia incoming edges, and

r(i) = I(Si, d) P

SjI(Sj, d) (13)

is the importance score of sentence Si, I(Si, d), defined in (1), but normalized by the sum of such scores for all sen- tences Sjin the document d. (12) can be solved with the ap- proach very similar to that for the PageRank problem [11]. Let v = [v(i), i = 1, 2, ..., L]Tand r = [r(i), i = 1, 2, ..., L]T be the column vectors for v(i) and r(i) for all sentences in the document, where L is the total number of sentences in the docu- ment d and T represents transpose. (12) then has a vector from below,

v = (1 − α)r + αPv

= ((1 − α)reT+ αP)v = P0v, (14) where P is an L × L matrix of p(j, i), and e = [1, 1, ..., 1]Tis an L-dimensional vector with all components being 1. Because P

iv(i) = 1 from (12) and (13), eTv = 1.

It has been shown that the solution v of (14) is the dominant eigenvector of P0[14], or the eigenvector corresponding to the largest absolute eigenvalue (which is 1) of P0. The solution v(i) can then be integrated with the original importance score I(Si, d) using LTE-based statistical measure (6) or key-term- based statistical measure (8),

I(Sˆ i, d) = I(Si, d)(v(i))δ, (15) where δ is a weighting parameter.

3. Experimental Setup

3.1. Corpus

The corpus used in this research is the lectures for a course of- fered in National Taiwan University. The lectures were given

in the host language of Mandarin Chinese but with almost all terminologies produced in the guest language of English and embedded in the Mandarin utterances. There is a total of 17 chapters, while the lecture is 45.2 hours long. We segmented the whole lecture into 155 documents by topic segmentation [15], and extracted the summary for each document.

34 documents out of the 155 were tested. The average length of each document was about 17.5 minutes, and manual transcriptions without word errors and ASR results were both used. For ASR, the acoustic models were trained with the AST- MIC corpus for Mandarin and the Sinica Taiwan English cor- pus for English respectively, and then adapted by 25.2 minutes bilingual corpus from the target speaker (the course instructor) [16]. The language model was trained by two other courses of- fered by the same instructor and adapted by the course slides.

The accuracies for the ASR transcriptions are 78.15% for Man- darin characters, 53.44% for English words, and 76.26% for overall.

The automatically key term extraction approach mentioned in Section 2.2 was used and both keywords and key phrases were extracted [9]. The F-measures for key terms were 62.70%

and 67.31% for ASR and manual transcriptions respectively.

We set the number of topics for PLSA as 32, considering the course of 17 chapters. The value of α is set to 0.9, which is em- pirically better choice [12, 13]. Two human subjects (students at National Taiwan University) were requested to produce two ref- erence summaries for each document by ranking the importance of the sentences in each document from ”the most important” to

”of average importance.”

3.2. Evaluation Metrics

The well-known evaluation package called ROUGE [17] was used in this research, including ROUGE-N (N = 1, 2, 3) and ROUGE-L scores. ROUGEr-N is an N-gram recall between the automatically generated summary and a set of manually generated reference summaries calculated as

ROUGEr-N = P

S∈S

P

gramN ∈SCountmatch(gramN) P

S∈S

P

gramN ∈SCount(gramN) , (16) where N stands for the length of N-gram considered, gramN; S is an individual reference summary, and S is a set of reference summaries. Countmatch(gramN) is the maximum number of N-grams co-occurring in the automatically generated summary and the reference summary. Count(gramN) is then the to- tal number of N-grams in the reference summary. ROUGE-L is similarly obtained but counting the ”longest common sub- sequence” (LCS) between the automatically generated sum- mary and the reference summary. F-measures for ROUGE-N (N = 1, 2, 3) and ROUGE-L can be evaluated in exactly the same way, which are used in the following results.

4. Results and Discussion

In the experiments to be presented below, the summarization ratio was set to be 10%, 20%, and 30% respectively. The key phrases (with more than one word) automatically extracted were taken as individual terms in PLSA modeling and all following processes.

Figure 2 shows the results for ROUGE-N and ROUGE-L for ASR (Figure 2 (a)-(d)) and manual transcriptions (Figure 2 (e)-(h)). In each case the three groups of bars are for 10%, 20%, and 30% summarization ratios, and in each group the four bars

(4)

10% 20% 30%

45 50 55

(a) ASR:ROUGE−1

F−measure

10% 20% 30%

20 25 30

(b) ASR:ROUGE−2

F−measure

10% 20% 30%

10 15 20 25

(c) ASR:ROUGE−3

F−measure

10% 20% 30%

40 45 50 55

(d) ASR:ROUGE−L

F−measure

10% 20% 30%

45 50 55

F−measure

(e) Manual:ROUGE−1 1:LTE 2:LTE+RW 3:Key 4:Key+RW

10% 20% 30%

20 25 30

(f) Manual:ROUGE−2

F−measure

10% 20% 30%

10 15 20 25

(g) Manual:ROUGE−3

F−measure

10% 20% 30%

40 45 50 55

(h) Manual:ROUGE−L

F−measure

Figure 2: The results of different choices of parameters: LTE- based (1, 2) or key-term-based (3, 4), without (1, 3) or with (2, 4) random walk, for ASR ((a)-(d)) or manual ((e)-(h)) transcrip- tions at summarization ratios of10%, 20%, and 30%.

are respectively for LTE-based statistical measure sLTE(ti, d) in (6) (LTE, the baseline), that followed by random walk (LTE + RW), key-term-based statistical measure in (8) (Key) and that followed by random walk (Key + RW). In all cases, the key- term-based statistical measure (bar 3) always outperformed the baseline (bar 1), or the LTE-based statistical measure. Clearly the key term knowledge is very helpful, especially for manual transcriptions. This is probably because in manual transcrip- tions all key terms are correctly transcribed (although may be incorrectly extracted) so that the key-term-based statistical mea- sures were much more accurately estimated. Similar improve- ments can also be observed for ASR transcriptions, but slightly less significant.

In all cases, random walk based on topical similarity al- ways helped the LTE-based statistical measure (bar 2 vs bar 1) For ASR transcriptions (Figure 2 (a)-(d)), random walk also improved the performance of key-term-based statistical mea- sure (bar 4 vs bar 3). However, with 10% summarization ra- tio for manual transcriptions (Figure 2 (e)-(h)), random walk wasn’t able to similarly help the key-term-based statistical mea- sure (bar 4 vs bar 3). The reason is probably that for manual transcriptions the important sentences within 10% were already very well selected by the key-term-based statistical measure, so adding extra topical similarity among many sentences cannot further improve the performance. However, the important sen- tences ranked between top 10% to 30% are less clear and there- fore random walk can help the key term knowledge (bar 4 vs bar 3) for 20% and 30% of summarization ratio (Figure 2 (e)-(h)).

Table 1 lists the maximum relative improvement (RI) achievable with respect to the baseline in all cases discussed above. For ASR transcriptions, the RI with different summa- rization ratios are similar, and the results are all from key-term- based statistical measure with random walk, probably because the topical similarity can compensate recognition errors and help include important sentences. For manual transcriptions, on the other hand, the results are from LTE-based statistical mea- sure with or without random walk, probably because without recognition errors the key terms can accurately include impor-

Table 1: Maximum relative improvement (RI) with respect to the baseline achievable in all cases (%).

MAX RI ASR Transcriptions Manual Transcriptions

10% 20% 30% 10% 20% 30%

ROUGE-1 5.30 7.19 7.13 14.37 8.27 6.74

ROUGE-2 13.94 15.48 17.45 33.31 18.33 15.33 ROUGE-3 24.98 22.23 21.33 50.98 28.21 22.25

ROUGE-L 6.23 7.51 7.92 14.52 9.12 7.49

tant sentences so that topical similarity cannot improve the per- formance in all cases; also, the RI with 10% summarization ratio is highest, probably because key terms are very helpful to identify top 10% important sentences; but with roughly 67%

of F-measure of key terms, the top 20% or 30% important sen- tences selected did include some noisy sentences.

5. Conclusions

Extensive experiments and evaluation with ROUGE metrics showed key-term-based statistical measure is good for speech summarization, and random walk also improves the perfor- mance. The random walk approach helps give higher scores to sentences topically similar to more important sentences, and thus consider all sentences in the document more globally.

6. References

[1] L. Lee, B. Chen, ”Spoken document understanding and organiza- tion, in Special Section, IEEE Signal Processing Megazine, 2005.

[2] J. Glass and et al., ”Recent progress in the MIT spoken lecture processing project,” in InterSpeech, 2007.

[3] S. Banerjee, A. Rudnicky, ”An extractive-summarizaion baseline for the automatic detection of noteworthy utterances in multi- party human-human dialog,” in SLT, 2008.

[4] Y. Gong, X. Liu, ”Generic text summarization using relevance measure and latent semantic analysis,” in ACM SIGIR Conference on R&D in Information Retrieval, 2001.

[5] J. Goldstein, M. Kantrowitz, and J. Carbonell, ”Summarizing text documents: Sentence selection and evaluation metrics,” in ACM SIGIR Conference on R&D in Information Retrieval, 1999.

[6] S. Furui and et al., ”Speech-to-text and speech-to-speech summa- rization of spontaneous speech,” in IEEE Trans. on Speech and Audio Processing, 2004.

[7] S. Kong, L. Lee, ”Improved spoken document summarization us- ing probabilistic latent semantic analysis (PLSA),” in ICASSP, 2006.

[8] S. Kong, L. Lee, ”Improved summarization of chinese spoken documents by probabilistic latent semantic analysis (PLSA) with further analysis and integrated scoring,” in SLT, 2006.

[9] Y. Chen and et al., ”Automatic key term extraction from spoken course lectures using branching entropy and prosodic/semantic features,” in SLT, 2010.

[10] T. Hofmann, ”Probabilistic latent semantic analysis,” in Univer- sity in AI, 1999.

[11] L. Page and et al, ”The pagerank citation ranking: bringing order to the web,” in Technical Report, Stanford Digital Library Tech- nologies Project, 1998.

[12] Y. Chen and et al, ”Improved spoken term detection with graph- based re-ranking in feature space,” in ICASSP, 2011.

[13] W. Hsu, L. Kennedy, ”Video search reranking through random walk over document-level context graph,” in MM, 2007.

[14] A. Langville, C. Meyer, ”A survey of eigenvector methods for web information retrieval,” in SIAM Review, 2005.

[15] S. Hsu, ”Topic segmentation on lecture corpus and its applica- tion,” in Master’s thesis of NTU, 2008.

[16] C. Yeh and et al, ”Bilingual acoustic modeling with state map- ping and three-stage adaptation for transcribing unbalanced code- mixed lectures,” in ICASSP, 2011.

[17] C. Lin, ”Rouge: A package for automatic evaluation of sum- maries,” in Workshop on Text Summarization Branches Out, 2004.

參考文獻

相關文件

o Construct a two-layer graph to represent 1) the utterance nodes in utterance-layer. 2) the speaker nodes

• More important utterances have higher weights corresponding to the prosodic feature P.  Unsupervised learn important

My paper on key term extraction was accepted for the Workshop on Spoken Language Technologies (SLT 2010, Student Best Paper Award), and as an undergraduate I

This paper presents a graph-based model that integrates prosodic features into an unsupervised speech summarization framework without any lexical informa- tion.. In particular it

One is the posterior prob- ability of query Q in the x i lattice (the first term on the right side of (19)), and the other is the relevance scores of the similar utterances (the

Experimental results show that this new approach offers significantly better performance than the previously proposed pseudo-relevance feedback approach, which considers primarily

Between-Layer Relation via Hidden Parameters Given a set of utterances {U i , i = 1, 2, ..., |U |} and a set of speakers {S j , j = 1, 2, |S|}, where a speaker node in the graph

For manual transcripts, row (e) cannot perform better by combing two types of similarity, which means topical similarity can dominate lexical similarity, since without