• 沒有找到結果。

Mining Web informative structures and contents based on entropy analysis

N/A
N/A
Protected

Academic year: 2021

Share "Mining Web informative structures and contents based on entropy analysis"

Copied!
15
0
0

加載中.... (立即查看全文)

全文

(1)

Hung-Yu Kao, Shian-Hua Lin, Member, IEEE Computer Society, Jan-Ming Ho, Member, IEEE, and

Ming-Syan Chen, Senior Member, IEEE

Abstract—In this paper, we study the problem of mining the informative structure of a news Web site that consists of thousands of hyperlinked documents. We define the informative structure of a news Web site as a set of index pages (or referred to as TOC, i.e., table of contents, pages) and a set of article pages linked by these TOC pages. Based on the Hyperlink Induced Topics Search (HITS) algorithm, we propose an entropy-based analysis (LAMIS) mechanism for analyzing the entropy of anchor texts and links to eliminate the redundancy of the hyperlinked structure so that the complex structure of a Web site can be distilled. However, to increase the value and the accessibility of pages, most of the content sites tend to publish their pages with intrasite redundant information, such as navigation panels, advertisements, copy announcements, etc. To further eliminate such redundancy, we propose another mechanism, called InfoDiscoverer, which applies the distilled structure to identify sets of article pages. InfoDiscoverer also employs the entropy information to analyze the information measures of article sets and to extract informative content blocks from these sets. Our result is useful for search engines, information agents, and crawlers to index, extract, and navigate significant information from a Web site. Experiments on several real news Web sites show that the precision and the recall of our approaches are much superior to those obtained by conventional methods in mining the informative structures of news Web sites. On the average, the augmented LAMIS leads to prominent performance improvement and increases the precision by a factor ranging from 122 to 257 percent when the desired recall falls between 0.5 and 1. In comparison with manual heuristics, the precision and the recall of InfoDiscoverer are greater than 0.956.

Index Terms—Informative structure, link analysis, hubs and authorities, anchor text, entropy, information extraction.

æ

1

I

NTRODUCTION

R

ECENTLY, there has been explosive progress in the

development of the World Wide Web. This progress creates numerous and various information contents pub-lished as HTML pages on the Internet. Furthermore, for the purpose of maintenance, flexibility, and scalability of Web sites, Web publishing techniques are migrating from writing static pages to deploying dynamic application programs, which generate contents on requests based on predefined templates and contents stored in back-end databases. Most commercial Web sites, such as portal sites, search engines, e-commerce stores, news, etc., apply the dynamic technique to adapt diverse requests from numer-ous Web users. Such Web sites are called systematic Web sites in this paper. A news Web site that generates pages with daily hot news and archives historic news is a typical example of the systematic Web site.

Due to the evolution of automatic generation of Web pages, the number of Web pages grows explosively [9], [17]. However, there is a lot of redundant and irrelevant information on the Internet [8], such as contents of mirror

sites or identical pages with different URLs [4], [5]. We call this kind of redundancy intersite redundancy. Also, a lot of redundant information exists within a Web site, especially in pages automatically generated by systematic Web sites. Such redundancy is referred to as intrasite redundancy. Examples of intrasite redundancy include company logos and texts, navigation panels, advertisements, catalogs of services, and announcements of copyright and privacy policies. These contents are frequently texts or hyperlinks irrelevant to the meaning of the page, while said hyperlinks are used for easy access to other pages that are semantically irrelevant to the original page. In a systematic news Web site, adding such irrelevant links makes it convenient for users to browse other news articles with fewer clicks by following shortcuts in the page. However, these irrelevant links increase the difficulty for Web site analyzers, search engines, and Web miners to perform their tasks. Those systems try to analyze, index, and mine information from the whole site, including redundant and irrelevant informa-tion. However, the performance of these systems is unavoidably degraded by the redundant and irrelevant information.

Consider the example in Fig. 1. We divide the root page of Washington Post1into several parts with different styles and contents, i.e.,

1. a banner with links “news,” “OnPolitics,” “Enter-tainment,“ “Live Online,” etc., at the top,

2. a menu with 22 links of news categories on the left, 3. a banner with advertisement links,

. H.-Y. Kao and M.-S. Chen are with the Electrical Engineering Department, National Taiwan University, Taipei, Taiwan, ROC. E-mail: bobby@arbor.ee.ntu.edu.tw and mschen@cc.ee.ntu.edu.tw. . S.-H. Lin is with the Computer Science and Information Science

Department, National Chi Nan University, Nantou Hsien, Taiwan, ROC. E-mail: shlin@csie.ncnu.edu.tw.

. J.-M. Ho is with the Institute of Information Science, Academia Sinica, Taipei, Taiwan, ROC. E-mail: hoho@iis.sinica.edu.tw.

Manuscript received 1 Sept. 2002; revised 1 Apr. 2003; accepted 10 Apr. 2003. For information on obtaining reprints of this article, please send e-mail to:

(2)

4. general announcements about Washington Post, 5. a block with promoted hot news and

advertise-ments,

6. a TOC block, and

7. a list with headline news.

In this case, items 1) and 2) are distributed among most pages in the site and are therefore redundant for users. We call these identical blocks redundant content blocks. However, they are still indexed by search engines. Such indexing induces an increase of the index size and is useless for users and harmful for the quality of search results. Items 3), 4) and 5) are irrelevant to the context of the page and are called irrelevant content blocks. These parts will make the topic of the page drift when terms in these parts are indexed. The last two items, 6) and 7), draw more attention from users and are called informative content blocks, in which users are able to read news articles via one click from anchors. For a user who is visiting to read news, items except for 6) and 7) are insignificant since they are used for visual and traversal purposes. The following examples describe their impacts with more detail:

Example 1. After searching “game hardware tech jobs” in Google (http://www.google.com)2, one of the most popular search engines, we found 20 pages of CNET (http://www.cnet.com, a news and product review Web site for computing and technology) in the top 100 results. However, none of these pages came from the Web pages categorized in CNET Job Seeker,3 which contains the desired information. There are matched query terms in redundant parts of pages among all these 20 pages, however, three of these pages do not contain any matched terms in the informative parts of pages. The matched terms in redundant parts of a page will increase the rank of that page, even though they are usually ignored by users. Note that six of the 20 pages are ranked as top 16 and the page with the highest ranking does not

contain any desirable information in its informative parts.

Example 2. We estimate the redundant rate of news Web sites from our news search engine (NSE) that collects articles pages from several news Web sites in Taiwan.4 According to the hand-coded rules, NSE merely crawls informative structures (TOC and article pages) and indexes the daily updated informative contents (article pages) from news Web sites. In a certain news Web site, one month of online news articles are reachable through their Web links. Based on these rules, it suffices to index only 5 percent of their pages and the index size of their informative parts is about 12.14 percent of the original page size. This implies that a typical search engine always indexes too much redundant information. Also, Web miners spend more effort in analyzing the whole site rather than focusing on the informative parts. In a systematic Web site, we define a set of TOC (Table of Content) pages and a set of article pages linked by these TOC pages as the informative structure of a Web site. Furthermore, both kinds of pages are analyzed to extract informative content blocks. The informative structure of a Web site is therefore represented as a set of TOC blocks pointing to a set of article blocks. An example of the informative structure is illustrated in Fig. 2. In the page TOC1, the content block enclosed by red line is the root of the structure and points to three article pages each contain-ing a news article enclosed in a so called informative block. News agents or other applications may then access all the news article pages by traversing the informative structure. In this paper, we define the problem of mining informative structures from Web sites and present our solutions. We comment that the informative structure is useful in the following applications:

Fig. 1. A sample page of news Web sites.

2. The result was queried from www.google.com on Febraury 5, 2002. 3. CNET Job Seeker: http://dice.cnet.com/seeker.epl?rel_code=1&op=1.

4. http://nse.yam.com/. The news search engine collects news page of 15 news Web sites in Taiwan.

(3)

. Crawlers and Information Agents can focus on the informative structure to precisely and efficiently extract information for further analysis.

. Search Engines may index only the informative parts of an article page rather than indexing every page in the whole Web site. As a consequence, we are able to reduce the index size and increase the retrieval precision when the query terms match terms in nonindexed redundant blocks.

. Previous researches on Web miners (information extraction systems, e.g., WIEN [26], Stalker [31], IEPAD [16], and [35] expect the input Web pages to possess a high degree of regularity so that structured information, e.g., metadata, encoded in these pages can be retrieved. Informative structure is a key to automatically locate target informative content blocks containing useful repeated patterns in the whole site.

To illustrate the difficulty of the problem, we consider a real Web site in Fig. 3, which shows a small graph of a Web site that contains only 270 pages and 1,515 links. The graph looks like a messy “Web.” The graph shown in Fig. 4 is a subgraph (informative structure) extracted from the origi-nal graph in Fig. 3 based on manually labeling TOC and article pages of the Web site. That is the answer set of the informative structure. This concise graph in Fig. 4 only contains TOC pages and the corresponding article pages, which excludes 76.8 percent of the links and 6.7 percent of the pages from the original graph. Fig. 4 consists of several

trees whose root nodes are TOC pages. In this paper, we propose methods to reduce complicated Web site structures such as the one in Fig. 3 into concise informative structures as in Fig. 4.

Explicitly, we propose in the paper mechanisms to automatically discover informative structure of a Web site and contents of pages to reduce intrasite redundancy. We also present an entropy-based analysis to estimate the information quality of links and content blocks. This new entropy measure is used in the approaches for discovering informative structure. In the rest of the paper, we first present a literature survey on related studies of the paper in Section 2. In Section 3, we present a formal model on the problem and develop our entropy-based analysis. Follow-ing the system design and implementation in Section 4, we perform several experiments on real data sets to evaluate our methods in Section 5. Finally, we conclude the paper and describe the direction of future research in Section 6.

2

R

ELATED

W

ORK

The Hyperlink Induced Topics Search (HITS) algorithm [25] and Google’s PageRank [7] are widely applied to analyze the structure of the Web. HITS estimates the authority and hub values of hyperlinked pages in the Web and Google merely ranks pages. Both methods are applied to ranking the search result. Based on a mutual reinforcement relation-ship, HITS provides an innovative methodology for Web searching and topics distillation. According to the definition in [25], a Web page is an authority on a topic if it provides

Fig. 3. The linked graph of CDN (www.cdn.com.tw). Page numbers N = 270 and total links L = 1,515.

(4)

good information and is a hub if it provides links to good authorities. HITS uses the mutual reinforcement operation to propagate hub and authority values to represent the linking characteristic. In recent research work on link analysis of hyperlinked documents, HITS is applied to the research area of topic distillation and several kinds of link weights are proposed to enhance the significance of links in hyperlinked documents. In the Clever system [14], weights tuned empirically are added to distinguish same-site links and others. In [3], the similarity between the document and the linked document is taken as the link weight for analysis. Another study that uses the similarity between the surrounding text of a link and the linked document to determine the link weight is conducted in [13]. Considering the distribution of terms in documents, Chakrabarti et al. [11] combines the TFIDF-weighted model and microhub to represent the significance of links in regions with informa-tion needed.

Intuitively, HITS and its related methods applied to the topic distillation are useful in the analysis of Web informative structures. However, the topic distillation is different from the informative structure mining in several aspects:

. The former distills hubs and authorities from a set of pages retrieved from search engines with a given query. These pages are not restricted to be published from the same Web site. However, the latter mines the informative structure from all pages of a given Web site.

. With different targets of data sets, studies of topic distillation usually omit intralinks and nepotistic links to perform the mutual reinforcement between sites. However, these links are important while the link analysis is focused on a Web site.

. Most adaptive topic distillation algorithms based on HITS take the relationship between queries and documents into consideration. However, these algo-rithms do not work well on mining informative structures because of the absent of a target query. Furthermore, as described in [12], [27], the link analysis algorithms, e.g., HITS, are vulnerable to the effect of nepotistic clique attack and Tightly-Knit Community (TKC). This is caused by the reason that the values of hub and authority of nodes will be self-prompted by the mutual reinforcement propaga-tions in a highly connecting community. The effects will be more significant for mining informative structures of Web sites since we observed that nepotistic links and cliques appear more frequently in a Web site [24].

Based on mining the informative structure of a Web site, the complex structure is reduced to a concise one. However, if we look into pages of the structure, many redundant content blocks are not meaningful for the pages content. In [20] and [23], studies provide learning mechanisms to recognize advertisements and redundant/irrelevant links of Web pages. However, these methods need to build the training data first and related domain knowledge must be included to extract features for generation of classification rules. Therefore, both methods are difficult to be applied to

automatically extract the informative structures of systema-tic Web sites.

Studies of Information Extraction (IE) [21], [22], [26], [35] aim to mine structure values (metadata) of pages from Web sites. Although being able to extract valuable metadata from pages, most of these IE systems need labor-intensive work. Cardie [10] defines five pipelined processes for an IE system: tokenization and tagging, sentence analysis, extraction, merging, and template generation. Machine learning is usually applied to learn, generalize, and generate rules in the last three processes based on manually generated domain-specific knowledge such as concept dictionaries and templates. Training instances applied to learning processes are also artificially selected and labeled. For example, in Wrapper induction [26], the author manually defines six wrapper classes, which consist of knowledge to extract data by recognizing delimiters to match one or more of the classes. The richer a wrapper class, the more likely it will work with any new site [15]. SoftMealy [22] provides a GUI that allows a user to open a Web site, define the attributes and label the tuples in the Web page. The common disadvantages of IE systems are the cost of templates, domain-dependent knowledge, or annotations of corpora generated by hand. This is the very reason that these systems are merely applied to specific Web applica-tions, which extract the structural information from pages of specific Web sites or pages generated by CGI. Conse-quently, the general IE systems are not scalable and, therefore, cannot be applied to resolve the problem of redundant content blocks in pages.

3

T

HE

E

NTROPY

-B

ASED

A

NALYSIS

In this section, we propose an entropy-based analysis to remedy the deficit of HITS-related algorithms. In light of this analysis, we devise two mechanisms to mine the informative structure of a Web site. We first develop a mechanism of analyzing the entropy of anchor texts and links, namely, Link Analysis of Mining Informative Struc-ture (LAMIS), to reduce a complex Web site to a distilled concise Web structure. Then, we devise another mechanism on analyzing the entropy of content blocks, namely, InfoDiscoverer, to identify informative (significant) content blocks from pages in the concise structure.

3.1 LAMIS—Analyzing Entropy of Anchor Texts and Links

Given an entrance of a Web site, HITS is applied to measure hub and authority values of all pages. Intuitively, a TOC page is expected to have a high hub value, and an article page is to have a high authority value. That is, HITS is designed to provide a reasonable solution to mining the informative structures of news Web sites. However, the existence of redundant and irrelevant links usually causes the phenomenon of topic drift in pages published in systematic Web sites. Therefore, the analysis of authority and hub cannot solely depend on the linked structures with the page granularity. We apply the entropy to measure the significance of anchor texts and page content while using HITS on the link analysis. For example, when the link entropy is applied to shrink the CDN graph shown in Fig. 3,

(5)

the thresholds of entropy values, 0.8 and 0.4, reduce that graph to those shown in Fig. 5 and Fig. 6, respectively. In comparison with the answer set shown in Fig. 4, Fig. 6 reduced by the threshold 0.4 indeed approaches to the optimal solution. Explicitly, the objective of our first mechanism, LAMIS, is to reduce a complex graph to an optimal subgraph that represents the informative TOC and article structure. The detail is described in the following sections.

3.2 InfoDiscoverer—Analyzing Entropy of Content Blocks

The entropy-based analysis can further be applied to analyze page content blocks and discover informative contents from article pages clustered by LAMIS. Our second mechanism, InfoDiscoverer, standing for “discover-ing informative content blocks,” is designed to analyze the information measure of content blocks (in a page set) and dynamically select the entropy-threshold to classify a page’s blocks into either informative or redundant. By partitioning a page into blocks based on HTML tags, InfoDiscoverer calculates entropy values of features (i.e., terms, keywords, or phrases) according to the probability distribution of features in the page set. Entropy values of content blocks are derived from their features. Based on the answer set

generated from 13 manually tagged news Web sites with a total of 26,518 Web pages, experiments show that both recall and precision rates are greater than 0.956.

Consequently, LAMIS applies the link entropy to dis-cover the informative structure and InfoDisdis-coverer employs the content entropy to determine the content property, informative or redundant. In the following, we describe the derivation of content and link entropy values. Then, we develop mechanisms to apply link entropy values to enhance the link analysis and mine the informative structure.

3.3 The Entropy of Content and Link and Enhanced Link Analysis

The text is an important clue for users to determine the meaning of a page. Therefore, we extract features (terms) from the page text to represent its corresponding signifi-cance. For the same reason, the anchor text is an important clue for users to click the link. Therefore, we also extract features (terms) from an anchor text to represent the significance of the link. In this paper, a term corresponds to a meaningful keyword. The idea is that features frequently appearing in most pages are redundant and carry less information to users. In contrast, features appear-ing in fewer pages are more informative. That is, we can

Fig. 5. The graph of CDN: links with entropy values smaller than 0:8½N ¼ 241; L ¼ 569.

(6)

apply the probability distribution of terms in pages to estimate the information measures of features. First, we calculate the feature entropy and deduce the entropy values of content blocks and links.

Given a set of HTML pages or a Web site, we can parse all pages into content blocks and links based on HTML tags. Also, texts appearing in contents and links can be parsed into a set of terms with corresponding blocks and links. In the page set, those features form a feature-document matrix with weights in matrix entries. The feature weight is calculated based on the calculation of feature frequency in the page set [29]. Then, Shannon’s information entropy [34] is applied to calculate the feature entropy based on the feature-document matrix. By definition, the entropy E can be expressed as Pn

i¼1pilog pi, where pi is the probability of eventi and

n is number of events. By normalizing the weight of a feature to be [0, 1], the entropy of feature (term) Ti is:

EðTiÞ ¼ 

Xn j¼1

wijlog2wij;

in which wijis the value of normalized feature frequency in

the page set. To normalize the entropy to the range [0, 1], the base of the logarithm is chosen to be the number of pages, and the previous equation becomes:

EðTiÞ ¼ 

Xn j¼1

wijlognwij;

where n¼ jDj; D is the set of pages:

Entropy values of content blocks and links are derived from the average of their features entropies. For the example of calculating the entropy of a content block, intuitively, feature entropies contribute to the semantic measure of a content block that owns these features, i.e., the entropy value of a content block is the summation of its features entropies:

HðCBiÞ ¼

Xk j¼1

EðTjÞ;

where Tjis a feature ofCBiwith k features:

Since content blocks contain different numbers of features, the equation is normalized as:

HðCBiÞ ¼

Pk j¼1EðTjÞ

k :

That is, the entropy of a content block, HðCBÞ, is the average of all feature entropies in the block. The link entropy is measured analogously.

In the link graph of a Web site G ¼ ðV ; EÞ, HITS algorithm computes two scores for each node v in V , i.e., the hub score HðvÞ and the authority score AðvÞ by the mutual reinforcement propagation between connecting nodes. In our approach, we incorporate entropy values of links as link weights to present the significance of links in a Web site structure. Therefore, the original HITS algorithm is modified as follows: AðvÞ ¼ X ðu;vÞ2E HðuÞ uvand HðvÞ ¼ X ðv;uÞ2E AðuÞ uv;

where uv is the weight of the link from u to v, denoted

ANuv. According to the definition of entropy of a link, uvis

defined as follows:

uv¼ 1  EðANuvÞ:

It can be seen that the more information a link carries, the larger the link weight is, i.e., the lower the link entropy is. The weight uv can also be expressed as the average of

the weights of features contained, i.e., uv¼

Pk

j¼1WðTjÞ

k :

According to the characteristic of the entropy, the weight of feature Tj, i.e., W ðTjÞ, is defined as WðTjÞ ¼ 1  EðTjÞ. The

weight of a feature is similar to its inverse document frequency, IDF, which is defined as logndfnj, where dfjis the

document frequency of Tj. IDF is usually applied to

represent the discriminability of a term in a set of documents. According to the definition, there are the following relationships between W ðTjÞ and IDFj: 1) if Tj

is uniformly distributed among some pages, W ðTjÞ ¼ IDFj,

and 2) if Tjis not uniformly distributed, then W ðTjÞ > IDFj

and the more skewed the distribution of Tj is, the larger

WðTjÞ is. The relationship 1) can be proven as follows:

Relationship 1. If Tj is uniformly distributed among some of

pages, W ðTjÞ ¼ IDFj.

Proof. Assume n ¼ jDj, tfj¼ the total frequency of Tj,

be-cause Tj is uniformly distributed, tfij, i.e., the term

frequency of Tjin pagei, is equal to tfj

dfj. According to the

definition of the entropy-based weighting,

(7)

WðTjÞ ¼ 1  EðTjÞ ¼ 1   X dfj tfij tfj logn tfij tfj 0 @ 1 A

and tfj¼ tfijdfj. We then get

WðTjÞ ¼ 1 þ logn 1 dfj ¼ logn n dfj ¼ IDFj:

When the distribution of Tjis more skewed, EðTjÞ will

decrease and W ðTjÞ will increase. The relationship 2) is

therefore conformed. tu Benefiting from these two relationships, the weight of a term attained from the entropy value is more representative for the importance of a term than from IDF. We will give the statistical result on the real Web site in Section 4.1 to show these relationships.

Moreover, the SALSA proposed in [27] is designed to resist effects of TKC and cliques and we also apply entropy weights on SALSA to remedy the effects similarly. The improvement will be empirically evaluated by our experi-mental studies later.

3.4 An Illustrated Example

Considering the example of a simple news Web site shown in Fig. 7, page P0is the homepage of the Web site. Page P1is

the TOC page with two anchors linking to news article pages, P3 and P4. Page P2is an advertisement page linked

by the other four pages. Most pages contain anchor texts, i.e., “home,” “hot news,” and “sales,” linking to P0, P1, and

P2, respectively. P3 and P4 have anchors linking to each

other. It is regarded as a cross-reference to present the related news. The Web site structure is also widely used in commercial Web sites. Based on the terms in each page, the feature entropy of the Web site is calculatedas below:

EðT0Þ ¼  X3 j¼1 1 3log5 1 3¼ 0:682; EðTiÞ ¼  X2 j¼1 2 5log5 2 5 1 5¼ 0:655; EðT2Þ ¼ EðT5Þ ¼  X4 j¼1 1 4log5 1 4¼ 0:861; and EðT3Þ ¼ EðT4Þ ¼  X2 j¼1 1 2log5 1 2¼ 0:430:

The entropy values of links derived from feature entropies are listed in Table 1. According to the definition of entropy, the most informative links are AN10and AN11,

which link to article pages P3and P4from TOC page P1.

Based on the link entropy, we use our entropy-based HITS algorithm to calculate values of hub and authority. In comparison with HITS shown in Table 2, hub and authority values are obtained after 10 iterations. P1 is ranked as the

top 1 hub page by the entropy-based HITS, and P3 and P4

are ranked with the highest authority. However, HITS ranks the advertisement page (P2) as the best authoritative page,

and news article pages (P3and P4) as good hub ones. It can

be seen that the link entropy is effective to enhance the link significance in link analysis algorithms.

4

T

HE

S

YSTEM

D

ESIGN

In this section, we will describe the design of LAMIS and InfoDiscoverer. The LAMIS system is designed to explore hyperlink information to extract and identify the hub (TOC) and authority (article) pages. InfoDiscoverer then measures the entropy values of content blocks among clustered pages (a set of article or TOC pages) and dynamically selects the entropy-threshold to extract informative content blocks. Given an entrance URL of a Web site, without manual inventions and prior knowledge about the Web site, LAMIS and InfoDiscoverer are capable of crawling all pages, analyzing the structure, and extracting the informative structures and content blocks of the site. In this section, we describe the system architecture and processes of LAMIS and InfoDiscoverer.

4.1 The System Architecture

Our Web mining system, shown in Fig. 8, consists of three parts:

1. Web Crawler which crawls pages, parses them into blocks, and builds the link graph of the Web site. 2. Feature Extractor which extracts features (terms),

in-degrees, out-in-degrees, text lengths, and links as metadata of pages. Feature Extractor also calculates entropy values of features to derive entropies of links and content blocks.

3. Informative structure mining module which is composed of LAMIS and InfoDiscoverer.

First, a starting URL of a site is given to Crawler. In our system, we can assign crawl depths to different paths

TABLE 2

(8)

(subtrees) of the site. Once a site has been crawled, a page is represented by some content blocks, and a site structure is built, and terms appearing in all pages are extracted. Terms appearing in links or content blocks are recorded for the subsequent estimation of feature entropy. As we know, extracting English terms is relatively simple. Applying stemming algorithms and removing stop words based on a stop-list, English keywords (terms) can be extracted [33]. Extracting terms used in oriental languages is, in general, more difficult because of the lack of separators in these languages. However, many studies have applied statistical approaches to extracting keywords of oriental languages [18]. In our system, we use an algorithm to extract keywords from Chinese sentences based on a Chinese term base generated via collecting hot queries, excluding stop words, from our search engine.5 After extracting features, the system maintains a feature-document matrix to represent the feature frequency corresponding to each document. According to the matrix, we can calculate the entropy values of features and derive links and content blocks entropies to be used as inputs to LAMIS and InfoDiscoverer.

The distribution of term weights assigned by entropy values and IDF values in Fig. 9 shows that their correlations conform to the two relationships described in Section 3.3. 4.2 LAMIS: Mining Informative Structures

To employ the link entropy in enhancing the link analysis algorithms, the most important issue is to prove the revised algorithm is convergent. In HITS, hub and authority will converge to the principal eigenvector of the link matrix [25]. Also, the weighted HITS algorithm is converged, if the multiplication of both weight matrices has no negative entry values [3]. In LAMIS, the weighting factor uv, is bounded in

[0,1], and we use the same weight matrix in hub and authority. Therefore, LAMIS will be converged after constant

iterations. In our experiments on various sites, 91.7 percent of the hub values will converge to zero after 10 iterations. Fig. 10 shows one of the experimental results in which hub values of all pages are sorted in descending order. It can be seen that they decrease sharply and most of them become zero. From our observation, TOC pages tend to hold high hub values and we use the top-N threshold to extract TOC pages from the ranked list of the whole page sets.

During the iteration of HITS, the converged curves are slightly undulated as shown in Fig. 11. This phenomenon is due to the effect of multiple propagation paths of mutual reinforcement relationships in HITS algorithm and is called nonuniqueness in [30]. Consider Fig. 12, for example. It is seen that two propagation paths are independent and the two converged authority and hub sets of one page, i.e., ðA2k; H2kÞ

and ðA2k¼1; H2kþ1Þ, will hence be generated. While the

authority and hub sets of all pages are considered, the two independent sets must be combined alternately. In Fig. 12, if these values converge at iteration 3, two authority and hub sets are

fðAa3; Ha3Þ; ðAb2; Hb2Þ; ðAc3; Hc3Þg

and

fðAa2; Ha2Þ; ðAb3; Hb3Þ; ðAc2; Hc2Þg

In general cases, these two independent sets will converge to the same one and we may select one of them to be used subsequently.

Fig. 8. The system architecture.

Fig. 9. 1-EN versus IDF for www.cnn.com.

Fig. 10. Distribution of hub values. 5. The searching service is a project sponsored by Yam, (http://

(9)

4.3 InfoDiscoverer: Mining Informative Contents After identifying TOC pages from a Web site, we can find article pages by following links in TOC pages. However, the redundant and irrelevant links in article pages are not easy to discovered. Therefore, we apply InfoDiscoverer to extract informative content blocks of the set of TOC pages. Article pages are defined as pages linked by anchors appearing in the informative blocks of a TOC page. Also, these article pages form a new data set from which InfoDiscoverer extracts informative blocks as the meaningful content article pages.

Given a set of TOC or article pages, InfoDiscoverer classifies content blocks into two categories, redundant and informative, based on the entropy values of content blocks as follows:

. If the entropy of a content block is higher than a defined threshold or close to 1, the block is absolutely redundant since most of the block’s features appear in every page.

. If the entropy of a content block is less than a defined threshold, the block is informative because features of the page are distinguishable from others, i.e., these features of the page seldom appear in other pages.

The threshold is not easy to determine since it would vary for different page sets. If the higher threshold is chosen, the higher recall rate is expected. However, the precision rate may become lower. To get a balanced recall-precision rate, we apply the greedy approach to dynami-cally determine the threshold for different page sets. If the

threshold is increased, more informative features (in informative content blocks) will also be included. The basic idea of the greedy approach is described as the following heuristic:

. Starting the entropy-threshold from 0 to 1.0 with an interval such as 0.1, increasing the threshold value will include more features since more content blocks are probably included. If the increase of the threshold never includes more features, the boundary between informative and redundant blocks is reached.

5

E

XPERIMENTS AND

E

VALUATION

In this section, we describe experiments on several news Web sites to evaluate the performance and improvement of our approaches. We first describe the data sets used in experiments. Then, we describe the evaluation of LAMIS and InfoDiscoverer and assess the performance of both methods.

5.1 The Data Sets

In our experiments, the data sets6 contain 14 Chinese and five English news Web sites as described in Table 3. All of these news sites provide real-time news and historical news browsing services including several domains: politics, finance, sports, life, international issues, entertainment, health, cultures, etc. In our experiments, the crawl depth is set to 3, and after pages have been crawled, the domain experts inspect the content of each page in Chinese news Web sites. They labeled these pages as TOC or article to build the answer set of data sets used in the following experiments. We found that the percentages of TOC pages vary among sites, i.e., different sites have different policies to organize their site structure. In fact, several sites have particular information structures. The diversity of informa-tion structures in data sets demonstrates the general applicability of LAMIS.

5.2 Evaluation on LAMIS

After extracting features from crawled pages, we compute the entropy values of content blocks and anchors. We found that entropy values of 71.6 percent of links are larger than 0.8, and they are probably redundant. As we expect, they appear in links or contents of navigation

Fig. 11. The convergence of authority and hub values.

Fig. 12. Two independent propagation paths of mutual reinforcement relationships in HITS.

6. Pages of Web sites in data sets are crawled at 2001/12/27 and 2002/4/11. The data sets can be retrieved in our research site http:// kp06.iis.sinica.edu.tw/isd/index.html.

(10)

panels, advertisements, and copyright announcements, etc. We first compare the performances of three link analysis algorithms: HITS, SALSA, and LAMIS. Basically, SALSA is the HITS algorithm with link normalization (LN) [6] and is therefore symbolized by HITS-LN. Similarly, the link normalization can be employed in LAMIS to become LAMIS-LN. We also compare these algorithms with a simple heuristic algorithm Out-link (OLOL), which ranks pages according to their counts of out-link rather than hub values. The idea of algorithm OL comes from that TOC pages have more links than article pages in general. Therefore, these methods are denoted by: OL, HITS, HITS-LN (SALSA), LAMIS, and LAMIS-HITS-LN. Based on the data sets (with answer sets) of Table 3, we conduct experiments to evaluate the performance of the previous five methods. By ranking pages according to their hub values or out-link counts, we examine the precision at 11 standard recall levels [2] for each site. Fig. 13 shows the precision

histograms of these methods based on the average precision rates of Web sites in data sets. We observe that LAMIS-LN emerges as the winner, followed by heuristic OL. HITS does not work well in these data sets.

Note, however, that these methods do not suffice to simultaneously render high recall and precision rates. In view of this, we shall devise some techniques to enhance LAMIS in the following.

5.2.1 Augmented Features of LAMIS

After investigating the raw pages in these data sets, we found that informative links are merely located in one or several informative content blocks. In previous experi-ments, we only consider hub values. The authority value is probably a compensation of the hub. Considering the anchor texts of links, redundant links are usually used for presentation purpose and their text lengths are thus usually very short. However, informative links tend to describe the title of the linked pages for the readable purpose. Therefore, we also consider the length of anchor text and, conse-quently, propose the following techniques to enhance our methods:

. Page mode (PA) versus Content block mode (CB), . Hybrid ranking (HR) of authority and hub, and . Anchor text length of links, which is linear to the

number (weight) of term counts (TW).

Page Mode versus Content Block Mode. In Fig. 14, we can see the difference of mutual reinforcement propagation between the page mode and the content block mode. In the page mode, authority of P2, i.e., Ap2, can affect the value Ap3

through hub of P1, Hp1. If P2 is authoritative, Ap3will also

be promoted, even though it is not an authoritative page. In

TABLE 3

Data Sets and Related Information

(11)

the content block mode, we divide P1 into two blocks, one contains a link to P2 and the other contains a link to P3. They are treated as separate nodes. Hence, the propagation of high authority of P2 will now be terminated at CB1 and P2 will not be incorrectly promoted. In our approach, blocks in the content block mode are extracted based on the <TABLE> HTML tag.

Hybrid Ranking (HR) of Authority and Hub. When the contexts of pages are complex, pages may contain more hybrid characteristics of hub and authority. To reduce the effect of hybridization of hubs and authorities in a page, we take into consideration the authority and use the hybrid ranking of hubs and authority. The idea is motivated by the observation that TOC pages hold not only the higher hub values, but also the lower authority values. To capture this notion, by assuming that the hub is inversely proportional to the authority, we employ the following equation in our experiments,

Rank¼ hub  nauthority;

where n is a Web site dependent coefficient with its value determined by

log2

the number of links in the W eb site 1; 000

 

:

The factor is motivated from the observation from experi-ments that, when the number of links in a Web site increases, the difference between the hub value and the authority value of a page also increases.

Applying Anchor Text Length to Reduce the Effect of Redundant Links. In our definition, redundant links are those links associated with navigation panels, advertise-ments, and others not relevant to the topic of the page. According to the information theory, terms in these links usually hold high entropy values so that these links are less weighted in link analyses. In TOC pages, informative links are usually described by a sentence to summarize the content of the linked page. For example, the anchor text of a TOC page’s link is usually the title of the linked article page. However, a readable sentence frequently consists of key-words and stop key-words, and the link entropy is therefore diluted. Therefore, the length of anchor text is taken into the consideration to enhance the performance of LAMIS. The anchor length is linear to the number of terms extracted from the anchor text; we define the term weight (TW) to evaluate the effect of the anchor text length:

uv¼ uvð1 þ log10ðterm countÞÞ:

This equation means that, if the number of terms in an anchor is 10, the weight of the link is doubled. The factor is motivated from the observation that, when the length of the anchortext is longer, the anchor contains richer information. 5.2.2 Performance Evaluation on LAMIS with

Augmented Features

Considering the augmented features mentioned previously, we have many optional combinations of methods and experiments. For a typical example, LAMIS-LN-CB-HR-TW means we integrate the “link normalization,” “hybrid ranking,” and “the term weight induced from the length of the anchor text” to LAMIS by analyzing the values of authority and hub in “the content block mode.” In Table 4, we show some more experiments for combinations of these

Fig. 14. Propagations of mutual reinforcement on different modes.

TABLE 4

R-Precision of All Experiments

Fig. 15. R-Precision improvement of augmented LAMIS.

(12)

methods based on the R-Precision, i.e., the equal point of recall and precision. The average R-Precision, 0.82, of LAMIS-LN-CB-HR-TW (the optimal LAMIS) is the best result. The optimal LAMIS is ranked first for R-Precision in six of eight Chinese data sets and ranked second in the other two. As compared to HITS, LAMIS-LN-CB-HR-TW improves the R-Precision by a factor of 2.03.

We also select the major methods and draw the graph from the experimental result. It can be seen that LAMIS-LN-CB-HR-TW outperforms others in all nine Chinese news Web sites in Fig. 15.

To evaluate our methods in English Web sites, we conducted several experiments on five English news Web sites and one nonnews Web site and compared the top 20 precision rates as shown in Fig. 16. We note that similar as in Chinese news Web sites, LAMIS-LN-CB-HR-TW still performs very well, indicating the robustness of this method. For comparison, we check the top 20 ranked pages and manually assign their properties, TOC or article. We list

the top 20 results of LAMIS-LN-CB-HR-TW and HITS for Washington Post in Table 5. Clearly, the augmented LAMIS ranks TOC pages with high hub values and HITS ranks 12 article pages with top 20 hub values. We also apply experiments on the nonnews Web site and the result is similar to those of news sites, indicating the applicability of our method to different Web sites.

5.3 Evaluation on InfoDiscoverer

We choose several Chinese news sites from the data sets. Since news articles of different categories may be published with different presentation styles, we choose one category from each site as shown in Table 6. That is, each site’s category indicates a page set used in InfoDiscoverer. For efficient execution, we do not run InfoDiscoverer for all instances in the page set since some sites may contain thousands of pages. Instead, 10 training pages are randomly selected from each cluster in the first experiment.

TABLE 5

Top 20 Results of LAMIS-LN-CB-HR-TW and HITS for www.washingtonpost.com

TABLE 6

(13)

To evaluate the performance of InfoDiscoverer, we estimate recall and precision rates of extracted features based on features in the answer set. Regarding features extracted from hand-coding informative content blocks as desired features (the answer set), measures of recall and precision are shown in Fig. 17. Clearly, the ideal is that both recall and precision rates equal ones.

Based on the greedy heuristic defined in Section 4, InfoDiscoverer dynamically finds the optimal threshold of each site, collects features of extracted informative content blocks, and calculates the corresponding recall and preci-sion rates based on the answer set. The result is shown in the last three columns of Table 6. The result shows that all sites, except for UDN, have very high recall rates (at least 0.956). These optimal thresholds of sites are distributed from 0.1 to 0.7. For example, CNet is converged at 0.5 with recall 0.956. That is, optimal thresholds vary among different sites. The recall of UDN (0.760) is not perfect. By tracing back to the training pages and the corresponding hand-coding data, we found that the hand-coding data of UDN is incorrectly classified because of the inclusion of the title information of news categories. The precision of TTIMES is 0.530 at the optimal threshold 0.7. We checked news articles of TTIMES and found that each page includes an extra content block consisting of “anchors of related news,” which are news pages related to the current article. Since the block consists of too many anchors, the text length of the block is even longer than that of the article in many pages. Therefore, these included noisy features affect the decision of the threshold. Consequently, InfoDiscoverer is able to dynamically find the optimal threshold of content block entropy for different page sets.

from five to 40 with interval 5. The result is shown in Fig. 18, in which the dotted line denotes the recall rate (R) and the solid line represents the precision (P). Most clusters have perfect recall and precision rates approaching to 1 (many R or P lines are overlapped at the highest value 1.0), but precision rates of few clusters (solid lines) are not when the number of randomly selected examples is increased. It is noted that the number of example has an influence on the precision rate since the precision rates of CTS, ET, and CTimes are degraded below 0.9 when the number is increased. In contrast, the random number has little effect on the recall rate since most dotted lines have recall rates larger than 0.956, except for CNet’s 0.942. Intuitively, if contents of a cluster are similar, the more examples involved, the higher entropy-threshold would be selected for filtering informative content blocks. Consequently, more training examples do not imply higher precision. However, the recall rate is not affected because a higher threshold means more features included.

6

C

ONCLUSIONS AND

F

UTURE

W

ORK

In the paper, we propose a system, composed of LAMIS and InfoDiscoverer, to mine informative structures and contents from Web sites. Given an entrance URL of a Web site, our system is able to crawl the site, parse its pages, analyze entropies of features, links and contents, and mine the informative structures and contents of the site. With a fully automatic flow, the system is useful to serve as a preprocessor of search engines and Web miners (information extraction systems). The system can also be applied to various Web applications with its capability of reducing a complex Web site structure to a concise one with informative contents.

During performing experiments of LAMIS, we found that the HITS-related algorithms are not good enough to be applied in mining the informative structures, even when the link entropy is considered. Therefore, we developed and investigated several techniques to enhance the mechanism proposed. We conducted several experiments and showed

Fig. 17. Recall and precision rates of content block evaluation.

(14)

that LAMIS-LN-CB-HR-TW was able to achieve the optimal solution in most cases. The R-Precision 0.82 indicates that the enhanced LAMIS performs very well in mining the informative structure of a Web site. The result of InfoDis-coverer also shows that both recall and precision rates are larger than 0.956, which is very close to the hand-coding result. In the future, we are interested in the further enhancement of our system. For example, the concept of generalization/specialization can be applied to find the optimal granularity of blocks to be utilized in LAMIS and InfoDiscoverer. Also, our proposed mechanisms are sig-nificant for and are worthy of further deployment in several Web domain-specific studies, including those for Web miners and intelligent agents. These are matters of future research.

R

EFERENCES

[1] B. Amento, L. Terveen, and W. Hill, “Does “Authority” Mean Quality? Predicting Expert Quality Ratings of Web Documents,” Proc. 23th ACM SIGIR, 2000.

[2] R. Baeza-Yates and B. Ribeiro-Neto, Modern Information Retrieval. Addision Wesley, 1999.

[3] K. Bharat and M.R. Henzinger, “Improved Algorithms for Topic Distillation in a Hyperlinked Environment,” Proc. 21st ACM SIGIR, 1998.

[4] K. Bharat and A. Broder, “Mirror and Mirror and on the Web: A Study of Host Pairs with Replicated Content,” Proc. Eighth Int’l World Wide Web Conf., May 1999.

[5] K. Bharat, A. Broder, J. Dean, and M.R. Henzinger, “A Comparison of Techniques to Find Mirrored Hosts on the WWW,” IEEE Data Eng. Bull., vol. 23, no. 4, pp. 21-26, 2000. [6] A. Borodin, G.O. Roberts, J.S. Rosenthal, and P. Tsaparas,

“Finding Authorities and Hubs from Link Structures on the World Wide Web,” Proc. 10th World Wide Web Conf., 2001. [7] S. Brin and L. Page, “The Anatomy of a Large-Scale Hypertextual

Web Search Engine,” Proc. Seventh World Wide Web Conf., 1998. [8] A. Broder, S. Glassman, M. Manasse, and G. Zweig, “Syntactic

Clustering of the Web,” Proc. Sixth World Wide Web Conf., 1997. [9] A. Broder, R. Kumar, F. Maghoul, P. Raghavan, S. Rajagopalan,

R. Stata, A. Tomkins, and J. Wiener, “Graph Structure in the Web,” Proc. Ninth World Wide Web Conf., 2000.

[10] C. Cardie, “Empirical Methods in Information Extraction,” AI Magazine, vol. 18, no. 4, pp. 5-79, 1997.

[11] S. Chakrabarti, M. Joshi, and V. Tawde, “Enhanced Topic Distillation Using Text, Markup Tags, and Hyperlinks,” Proc. 24th ACM SIGIR, 2001.

[12] S. Chakrabarti, “Integrating the Document Object Model with Hyperlinks for Enhanced Topic Distillation and Information Extraction,” Proc. 10th World Wide Web Conf., 2001.

[13] S. Chakrabarti, B. Dom, P. Raghavan, S. Rajagopalan, D. Gibson, and J.M. Kleinberg, “Automatic Resource Compilation by Analyzing Hyperlink Structure and Associated Text,” Proc. Seventh World Wide Web Conf., 1998.

[14] S. Chakrabarti, B. Dom, S. Kumar, P. Raghavan, S. Rajagopalan, A. Tomkins, D. Gibson, and J.M. Kleinberg, “Mining the Web’s Link Structure,” Computer, vol. 32, no. 8, pp. 60-67, Aug. 1999. [15] B. Chidlovskii, “Wrapper Generation by k-Reversible Grammar

Induction,” Proc. Workshop Machine Learning for Information Extraction, Aug. 2000.

[16] C.H. Chang and S.C. Lui, “IEPAD: Information Extraction Based on Pattern Discovery,” Proc. 10th World Wide Web Conf., 2001. [17] M.-S. Chen, J.-S. Park, and P.S. Yu, “Efficient Data Mining for Path

Traversal Patterns,” IEEE Trans. Knowledge and Data Eng., vol. 10, no. 2, pp. 209-221, Apr. 1998.

[18] L.F. Chien, “PAT-Tree-Based Keyword Extraction for Chinese Information Retrieval,” Proc. ACM SIGIR Conf. Research and Development in Information Retrieval, 1997.

[19] V. Crescenzi, G. Mecca, and P. Merialdo, “RoadRunner: Towards Automatic Data Extraction from Large Web Sites,” Proc. 27th Int’l Conf. Very Large Data Bases, 2001.

[20] B.D. Davison, “Recognizing Nepotistic Links on the Web,” Proc. Nat’l Conf. Artificial Intelligence (AAAI), 2000.

[21] D. Freitag, “Machine Learning for Information Extraction,” PhD Dissertation, Computer Science Dept., Carnegie Mellon Univ., Pittsburgh, PA, 1998.

[22] C.N. Hsu and M.T. Dung, “Generating Finite-State Transducers for Semi-Structured Data Extraction from the Web,” Information Systems, vol. 23, no. 8, pp. 521-538, 1998.

[23] N. Jushmerick, “Learning to Remove Internet Advertisements,” Proc. Third Int’l Conf. Autonomous Agents, 1999.

[24] H.Y. Kao, S.H. Lin, J.M. Ho, and M.S. Chen, “Entropy-Based Link Analysis for Mining Web Informative Structures,” Proc. The 11th ACM CIKM, 2002.

[25] J.M. Kleinberg, “Authoritative Sources in a Hyperlinked Environ-ment,” Proc. ACM-SIAM Symp. Discrete Algorithms, 1998. [26] N. Kushmerick, D. Weld, and R. Doorenbos, “Wrapper Induction

for Information Extraction,” Proc. 15th Int’l Joint Conf. Artificial Intelligence (IJCAI), 1997.

[27] R. Lempel and S. Moran, “The Stochastic Approach for Link-Structure Analysis (SALSA) and the TKC Effect,” Proc. Ninth Int’l World Wide Web Conf., 2000.

[28] W.S. Li, N.F. Ayan, O. Kolak, and Q. Vu, “Constructing Multi-Granular and Topic-Focused Web Site Maps,” Proc. 10th World Wide Web Conf., 2001.

[29] S.H. Lin and J.M. Ho, “Discovering Informative Content Blocks from Web Documents,” Proc. Eighth ACM SIGKDD, 2002. [30] J.C. Miller, G. Rae, and F. Schaefer, “Modifications of Kleinberg’s

HITS Algorithm Using Matrix Exponentiation and Web Log Records,” Proc. 24th ACM SIGIR Conf. Research and Development in Information Retrieval, 2001.

[31] I. Muslea, S. Minton, and C. Knoblock, “A Hierarchical Approach to Wrapper Induction,” Proc. Third Int’l Conf. Autonomous Agents (Agents ’99), 1999.

[32] P. Pirolli, J. Pitkow, and R. Rao, “Silk from a Sow’s Ear: Extracting Usable Structures from the Web,” Proc. ACM SIGCHI Conf. Human Factors in Computing, 1996.

[33] G. Salton, Automatic Text Processing: The Transformation, Analysis, and Retrieval of Information by Computer. Addison Wesley, 1989. [34] C.E. Shannon, “A Mathematical Theory of Communication,” Bell

System Technical J., vol. 27, pp. 398-403, 1948.

[35] K. Wang and H. Liu, “Discovering Structural Association of Semistructured Data,” IEEE Trans. Knowledge and Data Eng., vol. 12, no. 3, pp. 353-371, 2000.

[36] W3C DOM, Document Object Model (DOM), http:// www.w3.org/DOM/, 2003.

Hung-Yu Kao received the BS and MS degree from the Department of Computer Science at National Tsing Hua University (NTHU), Hsinchu, Taiwan, in 1994 and 1996, respectively. In July 2003, he received the PhD degree from the Electrical Engineering Department, National Tai-wan University, Taipei, TaiTai-wan. His research interests include information retrieval, knowledge management, data mining, and bioinformatics.

Shian-Hua Lin received the PhD degree in engineering science from National Cheng-Kung University, Taiwan, in 2000. He joined the Institute of Information Science, Academia Sinica, Taiwan, as a postdoctoral fellow. Since 2002, he has been with the Department of Computer Science and Information Engineering, National Chi-Nan University, Taiwan. Currently, he is an assistant professor. His research interests include Web document classification and retrieval, data mining, bioinformatics systems, and digital library. He is a member of IEEE Computer Society.

(15)

Sinica, Taiwan, R.O.C, as a associate research fellow in 1989 and was promoted to research fellow in 1994. He visited IBM T.J. Watson Research Center in the summers of 1987 and 1988, Leonardo Fibonacci Institute for the Foundations of Computer Science, Italy, in summer the of 1992, and Dagstuhl-Seminar on “Combinatorial Methods for Integrated Circuit Design,” IBFI-Geschaftsstelle, Schlo£ Dagstuhl, Fachbereich Informatik, Bau 36, Universitat des Saarlandes, Germany, in October 1993. He is a member of the IEEE and ACM. His research interests target at the integration of theoretical and application-oriented research, including mobile computing, environment for management and presentation of digital archive, management, retrieval, and classification of Web documents, continuous video streaming and distribution, video confer-encing, real-time operating systems with applications to continuous media systems, computational geometry, combinatorial optimization, VLSI design algorithms, and implementation and testing of VLSI algorithms on real designs. He is associate editor of IEEE Transactions on Multimedia. He was program chair of the Symposium on Real-Time Media Systems, Taipei, 1994-1998, general cochair of the International Symposium on Multi-Technology Information Processing, 1997, and general cochair of IEEE RTAS 2001. He was also a steering committee member of the VLSI Design/CAD Symposium, and program committee member of several previous conferences including ICDCS 1999, and IEEE Workshop on Dependable and Real-Time E-Commerce Systems (DARE’98), etc. In domestic activities, he is program chair of the Digital Archive Task Force Conference, the First Workshop on Digital Archive Technology, a steering committee member of the 14th VLSI Design/ CAD Symposium and the International Conference on Open Source 2002, and is also a program committee member of the 13th Workshop on Object-Oriented Technology and Applications, the Eighth Workshop on Mobile Computing, the 2001 Summer Institute on Bio-Informatics, Workshop on Information Society and Digital Divide, the 2002 International Conference on Digital Archive Technologies (ICDAT2002), the APEC Workshop on e-Learning and Digital Archives (APEC2002), and the 2003 Workshop on e-Commerce, e-Business, and e-Service (EEE’03).

nication Engineering and a professor in the Electrical Engineering Department, National Taiwan University, Taipei, Taiwan. He was a research staff member at IBM Thomas J. Watson Research Center, Yorktown Heights, New York, from 1988 to 1996. His research interests include database systems, data mining, mobile computing systems, and multimedia networking, and he has published more than 160 papers in his research areas. In addition to serving as a program committee member of many conferences, Dr. Chen served as an associate editor of the IEEE Transactions on Knowledge and Data Engineering on data mining and parallel database areas from 1997 to 2001, he is on the editorial board of the VLDB Journal, the Journal of Information Science and Engineering, and the Journal of the Chinese Institute of Electrical Engineering, was a distinguished visitor of the IEEE Computer Society for Asia-Pacific from 1998 to 2000, and program chair of PAKDD-02 (Pacific Area Knowledge Discovery and Data Mining), program vice-chair of VLDB-2002 (Very Large Data Bases) and ICPP 2003, general chair of Real-Time Multimedia System Workshop in 2001, program chair of IEEE ICDCS Workshop on Knowledge Discovery and Data Mining in the World Wide Web in 2000, and program cochair ofthe International Conference on Mobile Data Management (MDM) in 2003, International Computer Symposium (ICS) on Computer Networks, Internet and Multimedia in 1998 and 2000, and ICS on Databases and Software Engineering in 2002. He was a keynote speaker on Web data mining at the International Computer Congress in Hong Kong, 1999, a tutorial speaker on Web data mining at DASFAA-1999 and on parallel databases at the 11th IEEE International Conference on Data Engineering in 1995 and also a guest coeditor for the IEEE Transactions on Knowledge and Data Engineering on a special issue for data mining in December 1996. He holds, or has applied for, 18 US patents and seven ROC patents in the areas of data mining, Web applications, interactive video playout, video server design, and concurrency and coherency control protocols. He is a recipient of the NSC (National Science Council) Distinguished Research Award in Taiwan and the Outstanding Innovation Award from IBM Corporate for his contribution to a major database product, and also received numerous awards for his research, teaching, inventions, and patent applications. He was a coauthor with his students on a paper which received the ACM SIGMOD Research Student Award and Long-Term Thesis Award. Dr. Chen is a senior member of IEEE and a member of ACM.

.For more information on this or any computing topic, please visit our Digital Library at http://computer.org/publications/dlib.

數據

Fig. 1. A sample page of news Web sites.
Fig. 3. The linked graph of CDN (www.cdn.com.tw). Page numbers N = 270 and total links L = 1,515.
Fig. 6. The graph of CDN: links with entropy values smaller than 0:4½N ¼ 236; L ¼ 353.
Fig. 7. A simple Web site, jDj ¼ 5.
+6

參考文獻

相關文件

obtained by the Disk (Cylinder ) topology solutions. When there are blue and red S finite with same R, we choose the larger one. For large R, it obeys volume law which is same

Corpus-based information ― The grammar presentations are based on a careful analysis of the billion-word Cambridge English Corpus, so students and teachers can be

Based on historical documents and archeological evidence, this thesis provides an analysis of, raises some worth-noting questions on, the development of Western Qin Buddhism

CAST: Using neural networks to improve trading systems based on technical analysis by means of the RSI financial indicator. Performance of technical analysis in growth and small

CAST: Using neural networks to improve trading systems based on technical analysis by means of the RSI financial indicator. Performance of technical analysis in growth and small

We also propose a Unified Code Management Schemes to eliminate code blocking completely and the reassignment cost will be reduced as far as possible based on CIDP.. Our schemes

In this chapter, the results for each research question based on the data analysis were presented and discussed, including (a) the selection criteria on evaluating

Based on the tourism and recreational resources and lodging industry in Taiwan, this paper conducts the correlation analysis on spatial distribution of Taiwan