• 沒有找到結果。

Named Entity Recognition From Spoken Documents Using Global Evidences and External Knowledge Sources with Applications on Mandarin Chinese

N/A
N/A
Protected

Academic year: 2021

Share "Named Entity Recognition From Spoken Documents Using Global Evidences and External Knowledge Sources with Applications on Mandarin Chinese"

Copied!
6
0
0

加載中.... (立即查看全文)

全文

(1)

NAMED ENTITY RECOGNITION FROM SPOKEN DOCUMENTS USING GLOBAL

EVIDENCES AND EXTERNAL KNOWLEDGE SOURCES WITH APPLICATIONS ON

MANDARIN CHINESE

Yi-cheng Pan

, Yu-ying Liu

, and Lin-shan Lee

Graduate Institute of Computer Science and Information Engineering, National Taiwan University

Taipei, Taiwan, Republic of China

{thomas,lsl}@speech.ee.ntu.edu.tw,

tammy@realtek.com.tw

ABSTRACT

In this paper, we propose two efficient approaches for Named Entity recognition (NER) from spoken documents. The first approach used a very efficient data structure, the PAT trees, to extract global evidences from the whole spoken docu-ments, to be used with the well-known local (internal and external) evidences popularly used by conventional approaches. The basic idea is that a Named Entity (NE) may not be eas-ily recognized in certain contexts, but may become much more easily recognized when its repeated occurrences in all the different sentences in the same spoken document are considered jointly. This approach is equally useful for NER from text and spoken documents. The second approach is to try to recover some Named Entities (NEs) which are out-of-vocabulary (OOV) words and thus can’t be obtained in the transcriptions. The basic idea is to use reliable and important words in the transcription to construct queries to retrieve relevant text documents from external knowl-edge sources (such as Internet). Matching the NEs obtained from these retrieved relevant text documents with some se-lected sections of the phone lattice of the spoken document can recover some NEs which are OOV words. The exper-iments were performed on Mandarin Chinese by incorporat-ing these two approaches to a conventional hybrid statistic/rule-based NER system for Chinese language. Very significant performance improvements were obtained.

1. INTRODUCTION

Named Entities (NEs) such as person names, location names and organization names usually carry the core information of spoken documents, and are usually the key in understand-ing spoken documents. Therefore, Named Entity recog-nition (NER) has been the key technique in applications such as information retrieval, information extraction, ques-tion answering, and machine translaques-tion for spoken docu-ments. In the last decades, substantial efforts have been made and impressive achievements have been obtained in

the area of Named Entity recognition (NER) for text docu-ments. Most of these works, either rule-based or statistics-based , rely on two major categories of evidences to extract the Named Entities (NEs) [1]. The first is the internal evi-dences within the Named Entities (NEs) such as the Capi-talization of the characters, while the second is the external evidences such as the word co-occurrences observed from its context. However, these two categories of evidences are definitely not enough. For example, in the sentence “Even Language Technologies can’t handle such problem”, it is very difficult to recognize that “Language Technologies” in-stead of “Even Language Technologies” is an organization name with these two evidences alone. Considering the case of spoken documents, NER becomes more difficult since we can’t simply rely on text NER techniques to be applied on the transcriptions of the spoken documents. The automatic speech recognition (ASR) outputs always have many speech recognition errors, and many NEs themselves are out-of-vocabulary (OOV) words and therefore can’t be obtained in the transcriptions at the beginning.

In this paper, we propose two important new approaches to alleviate the above difficulties. The first is an efficient data structure, the PAT trees, to extract the global evidences from the whole spoken document to help the NER task. This approach is equally useful to text and spoken docu-ments. The second is to retrieve relevant text documents from some external knowledge sources (such as the Inter-net) using queries constructed by reliable and important words in the transcription, and recover some NEs which are OOV words by matching with the NE candidates obtained from these relevant text documents. These approaches have been applied on Chinese broadcast news, and the initial experi-ments indicated that these approaches are very useful.

Below, in Section 2 we will briefly present the approach of using PAT trees to extract global evidences for NER. This approach can be applied to text documents or ASR outputs of spoken documents. In Section 3, we describe the approach of recovering OOV words by retrieving rele-vant text documents from some external knowledge sources.

(2)

In Section 4, a baseline NER system for Chinese language is briefly summarized, on which the above two new ap-proaches were integrated and with which all the experiments were performed. The experiments with Chinese broadcast news and discussions are then given in Section 5, with con-cluding remarks finally given in Section 6.

2. EXTRACTION AND ORGANIZATION OF GLOBAL EVIDENCES WITH PAT TREES Many NEs may appear repeatedly many times in different sentences in the same documents. It may be difficult to rec-ognize it from each single sentence individually, but if we can record such repetitive occurrences along with their dif-ferent contexts in difdif-ferent sentences, we can consider the internal/external evidences for them in different sentences jointly. Such repetitive occurrences and the associated in-ternal/external evidences contexts in different sentences in the same document are referred to as global evidences in this paper. Almost all NER approaches can be decomposed into two stages, NE candidate generation and NE verifica-tion and classificaverifica-tion. The global evidences proposed here can be very helpful in NE candidate generation.

2.1. PAT tree

We need to efficiently organize and record the global evi-dences for a text document or the transcription of a spoken document. This is not trivial since there are a huge number of different segments of symbols (words or characters) in a document or a transcription. That’s why we use the PAT trees to do this job. A PAT tree is an efficient data structure that has been successfully used in the area of information retrieval [2]. It is well known that the binary search tree is a very simple data structure for data search, but it requires more space and time because all the branch nodes and leaf nodes need to be constructed, stored, and traversed during search. As in the example in Figure 1, a complete tree is needed even if only 6 data are recorded. To reduce the space and time requirements, a comparison bit (CB) is usually used to indicate the bit differentiating the data to construct a compressed binary search tree as shown in Figure 2 (CB = 2 indicating the second bit is the comparison bit and so on), in which only necessary branch/leaf nodes are kept. The PAT tree, developed based on the PATRICIA algorithm [3], is conceptually equivalent to a compressed binary search tree, but further reduced using augmented branch nodes which can also serve as leaf nodes. As shown in Figure 3, the leaf nodes are actually absorbed by the branch nodes and only 6 nodes are needed for the 6 data in Figures 1 and 2. When the PAT tree is used here for NER, a PAT tree can be constructed for each document, and the data recorded in the PAT tree are the symbol (e.g. character for Chinese or

Branch Nodes Leaf Nodes 0 0 0 1 A B C D E F 0 0 0 0 A:0010 B:0011 C:0110 D:1000 E:1011 F:1110 0

Fig. 1. An example of a complete binary search tree CB=1 CB=2 CB=2 CB=3 CB=4 A:0010 B:0011 C:0110 D:1000 E:1011 F:1110

Fig. 2. An example of a compressed binary search tree

word for English) strings in the document, probably divided by punctuation marks such as “.” and “,”, together with all possible suffix strings. Also, CB in Figure 3 now plays the role as checking symbols (character or word) rather than bits. For example, consider a Chinese sentence or character string “.αβγν ,ωε.”, where each Greek character represents a Chinese character. It is first divided into two character strings, “αβγν” and “ωε”, and then all possible different suffix strings, i.e., “αβγν”, “βγν”, “γν”, “ν”, “ωε”, “ε” are recorded in the PAT tree. During the construction of the PAT tree, each distinct suffix string is represented as a node in the PAT tree, but the repeated suffix strings are recorded on the same node with frequency counted. In this way, not only all suffix strings are recorded, but also the repeated suf-fix strings are counted in the PAT tree. At the same time the number of branches or child nodes split from a node indi-cates the variety of the nodes right context symbols. As in the example shown in the left half of Figure 4, where A, B, C are three distinct right context symbols for the symbol segment “αβγ” appearing in the document. So the PAT tree not only efficiently records all the symbol strings and possi-ble suffix strings with minimum storage and fast search, but offer a means to count all the right contexts for a specific symbol segment. Similarly we can construct an inverse PAT tree in a reversed order as the right half of Figure 4, i.e.,

(3)

CB=2 CB=1 CB=4 CB=2 CB=3 F:1110 C:0110 B:0011 A:0010 E:1011 D:1000 root

Fig. 3. An example of a PAT tree, recording 6 data:A, B, C, D, E, and F

including “νγβα”, “γβα”, “βα”, “α”, “ε”, “ω” as in the above example, so as to count the left contexts of a specific symbol segment, where D, E, F are the left context symbols for the symbol segment “αβγ” appearing in the document. 2.2. NE candidate generation using global evidences The process to generate NE candidates based on global ev-idences using Pat tree is easy. We first construct a PAT tree and an inverse PAT tree for the whole document. Each sym-bol segment s that has frequency count more than 2 will be first selected along with its number of child nodes for right/left context counts. But these selected symbol seg-mentss are just segments of symbols, not necessarily rep-resent complete entities. Therefore all these selected sym-bol segmentss should be tested by the criteria as given in inequalities (1) and (2) below to make sure they represent a complete entity with a clear concept and correct boundaries, |LC| > tc and |RC| > tc, (1) max l∈LC f(l, s) f(s) < tb and r∈RCmax f(s, r) f(s) < tb, (2) whereLC is the set of distinct left contexts of the symbol segments and RC the set of distinct right contexts, |·| is the number of elements in the set,f(s) is the frequency count of s, and f(l, s), f(s, r) are the frequency counts for the sym-bol segmentss preceded by the left context l or followed by the right contextr. The meaning of inequalities (1) and (2) is that in order fors to represent a complete entity with a clear concept and correct boundaries, it should have enough variety in its left and right contexts (inequality (1)), and its left and right contexts can’t be dominated by a certain specific context (inequality (2)), i.e.,maxl∈LC f (l,s)f (s) < tb means the counts for the most frequent left contextl for s can’t exceed a given ratio. The thresholdstcandtb can be empirically determined. Note that the tests in inequalities

(1) and (2) can be very efficiently performed with the PAT tree. Those symbol segments satisfying these criteria are then the NE candidates extracted. In this way, the global evidences for NEs can be efficiently extracted and recorded by PAT trees. Y1 X1 DE : 1 Y ... A DEJ DEJB... JE : 2 Y ... C DEJ DEJ : 1 X JED : 2 X Y2 X2 ... D

JED JEDE... JEDF...

The Pat tree The Inverse Pat tree

Fig. 4. An example of the PAT tree and the inverse PAT tree

2.3. Practical issues

Note that when a PAT tree is constructed for a document, all the n-gram statistics of the document are automatically in-cluded in the PAT tree. But the PAT tree also records much more context information about the document in addition to the n-gram statistics, in particular those required by inequal-ities (1) and (2). This is why PAT trees are proposed here as a special approach to NER. Also, PAT tree is specially ef-ficient in using very limited space to record huge quantities of context information. But the time complexity to search through all the nodes in a PAT tree grows exponentially with the length of the document. A simple approach to this prob-lem is to divide a long document into several parts, and con-struct sub-PAT trees for each part of it. In the experiments discussed below with broadcast news, all news stories are not long, and the search through the PAT trees can always be completed in real time. In fact, in real implementation it may not be necessary to check all the nodes in the PAT tree and the inverse PAT tree with inequalities (1) and (2). Some simple rules, such as a person name must include a surname or be followed by some cue words like “point out” or “say”, are very helpful in deleting the impossible nodes and making the search through PAT trees very efficient.

Also, those NEs appearing only once in the whole doc-ument can’t be extracted as NE candidates by the PAT tree. But the PAT tree doesn’t hurt anyway and the correct recog-nition of such NEs is determined by the stage of NE verifi-cation and classifiverifi-cation in any case. PAT tree helps as long as some NEs repeat themselves in a document with some difficult context as the example mentioned in Section 1.

(4)

3. CONFIDENCE MEASURES AND RECOVERY OF OOV WORDS

For spoken documents, the basic process of NER is to per-form speech recognition first and then perper-form NER on the transcription obtained using NER approaches for text docu-ments. But there are always recognition errors in the tran-scription. Also,many NEs themselves are out-of-vocabulary (OOV) words and thus can’t be obtained in the transcription. Confidence measures are helpful to both issues.

3.1. Use of confidences measure to identify reliable words in the transcription

We first carefully evaluated the confidence measureC(w) for each word w in the obtained word graph for the spo-ken document. Two levels of confidence measure are incor-porated here. The lower level is from the posterior word probability computed for each word hypothesisw from the word graph [4]. We then clustered together all word hy-pothesis on the word graph corresponding to the same word w with close enough begin-time and end-time to construct a sausage [5] In the sausage the posterior probabilities for all entries merged into a single entry were summed up. The posterior probabilities are regarded as the lower level con-fidence measure. The higher level concon-fidence measure, on the other hand, is based on the probabilistic latent seman-tic analysis (PLSA) using the knowledge about the words in the transcription [6] In this approach we evaluated the probability of observing each wordw in the transcription given the transcription of the whole spoken document D, p(w|D), and assign a score related to this probability. It was found that such high-level confidence measure really helped. We then linearly combined the two different levels of confidence measures into a single valueC(w). It was used to identify the reliable words in the transcription. 3.2. Extra NE candidates from external knowledge sources The basic idea here is that the OOV NEs can’t be obtained in the transcription, but may exist in some other relevant text documents, which can be retrieved from external knowledge sources (such as the INternet) using reliable and important words in the transcription of the spoken document. The NEs recognized from those text documents may then be good NE candidates for the spoken document. Therefore the first step is to construct good queries from the transcription of the spoken document to be used on some search engine, scuh as Google, to retrieve relevant text documents. There are several ways to construct such queries. A simple approach was to take every three adjacent words in the transcription of the spoken document and use them to construct disjunctive (Ored) queries. But queries obtained this way may bring many noisy text documents. A more efficient approach,

however, was to find more reliable and important words to construct the conjunctive (Anded) queries. In this ap-proach, we traversed the word graph and selected the word sequences within the graph with component words having relatively higher confidence measure and relatively higher term frequency (tf)· inverse document frequency (idf) scores, and use them to construct the queries [7]. As an example shown in Figure 5, the solid edges of the word sequence ABC and the word D are two selected queries. Less num-ber of more relevant text documents were retrieved with queries constructed in this way. NER processes including using global evidences with PAT trees were then performed on these documents to produce a set of NE candidates to be considered.

A

B

C

D

Fig. 5. Query construction based on the word graph by con-fidence measures and tf·idf scores

3.3. NE matching with phone lattices in the spoken doc-uments

With NE candidates obtained from the relevant text docu-ments as mentioned above, there are several ways to use them. The first is to add them to the lexicon, if they are not in the original lexicon, and then perform a second-phase transcription. Hopefully this time some NEs are not OOV any longer. A problem with this approach is these new NEs also need language model scores in the recognition process. Such language model scores can be estimated [8], but the results may not be very satisfactory. Here we used a differ-ent approach. The basic idea is that those word segmdiffer-ents in the transcription with relatively lower confidence mea-sure are likely to be recognition errors due to OOV NEs. So we can match the recognized phone lattices for these seg-ments with the NE candidates obtained from the relevant text documents. If the similarity is higher than a threshold, we then include the matched NE as a new NE candidate for the spoken document to go through the standard NE ver-ification/classification procedure. In order to perform the matching between two phone sequences, we define a phone similarity matrix, which is based on both the acoustic dis-tance (from the Mahalanobis Disdis-tance between the HMM models of the phones) and the pronunciation distance (from the probability that a phone is likely to be pronounced as another phone in the pronunciation model [9]). In this ma-trix, every pair of phones has a distance between 0 and 1

(5)

between them. The phone sequence matching is then based on the total distance normalized with the number of phones in the sequences.

4. THE BASELINE NER SYSTEM FOR CHINESE TEXT DOCUMENTS

The above two new approaches were incorporated with a baseline NER system for Chinese text documents briefly presented here. It is a hybrid Chinese system, primarily based on a previously proposed statistics-based framework [10], but integrated with many carefully developed rules. In this system, we not only recognize NEs as person names (PN), location names (LOC), and organization names (ORG), but simultaneously solve the problem of Chinese word seg-mentation (because there are no blanks serving as word bound-aries in Chinese texts).

Given a sequence os Chinese charactersS = s1, ..., sn, there exist many possible word sequencesW = w1, ..., wm (each word is composed of one to several characters) with corresponding class sequencesC = c1, ..., cm, (the classes here are PN, LOC, ORG and all other words in the vocabu-lary) and the purpose here is to find the best word sequence W∗ = w∗1, ..., wm∗ and its corresponding class sequence C∗= c∗1, ..., c∗mthat maximizeP (C, W ) as shown in equa-tion 3

(C∗, W) = arg

C,WP (C, W )

= argC,WP (W, C)P (C). (3) Here P(C) can be estimated by a class trigram language model according to equation (4),

P (C) ∼= P (c1) · P (c2|c1) m  i=3

P (ci|ci−2ci−1), (4) andP (W |C) can be approximated as given below,

P (W |C) = P (w1. . . wm|c1. . . cm) ∼= m  i=1

P (wi|ci). (5) The class trigramP (ci|ci−2ci−1) and the probabilities P (wi|ci) were trained using an NE labeled corpus.

In the hybrid approach here, we first generated all pos-sible NE candidate classes to construct a word lattice and evaluate the corresponding probabilitiesP (wi|ci). At this stage, quite many carefully developed rules were utilized. For example, for a Chinese person name candidate, we re-quired that it included a surname predefined in a surname list, or has some person name related keywords in its con-text. We also incorporated the global evidences obtained with the PAT tree. For all character segments that fulfill the inequalities (1) and (2) as defined in Section 2.3 but not

yet constructed as a NE candidate, we tag them as NE can-didates with all possible class types PN, LOC, ORG. We finally assembled all candidates into a word lattice and per-formed the Viterbi search based on equation (3). The Viterbi search was multi-layered, because a Chinese NE can be a concatenation of several NEs.

5. EXPERIMENT RESULTS

We performed two sets of experiments for NER from text and spoken documents respectively in the evaluation. For the text documents, the standard Chinese test set of MET (Multilingual Entity Task)-2 defined by MUC-7 (Seventh Message Understanding Conference) was used, which in-cludes 100 text documents. For the spoken documents, we used a total of 200 Chinese broadcast news recorded from radio stations at Taipei in 8 days in Sept. 2002 as the test corpus. NEs manually identified from these news stories, including 315 person names, 457 location names and 500 organization names, were taken as references. The extended knowledge sources to be retrieved for relevant documents are the Chinese text news available at “Yahoo! Kimo News Portal” [11] for the whole month of Sept. 2002, including about 38,800 text news stories. The evaluation metric we used was Precision and Recall rates plus F1 score, as de-fined by MUC-7. The results are listed in Table 1. For

Experiment

Cases NE Recall Precision F1 score Overall F1 PER 94 96 95.0 LOC 89 93 91.0 (A) ORG 87 96 91.3 89.5 PER 95 96 95.5 LOC 94 92 93.0 Text Docu-ments (B) ORG 89 95 91.9 91.1 PER 71 86 77.8 LOC 86 91 88.4 (C) ORG 64 95 76.5 77.6 PER 73 85 78.5 LOC 87 91 89.0 (D) ORG 67 95 78.6 80.0 PER 76 87 81.1 LOC 87 90 88.5 Spoken Docu-ments (E) ORG 68 95 79.3 80.9

Table 1. baseline in part (A), with global evidences by PAT trees in part (B) for text documents; and baseline in part (C), with global evidences in part (D) and OOV recovering in part (E) for spoken documents.

the text document, part (A) lists the results for the baseline Chinese NER system presented in Section 4, while part (B) are the results when the global evidences obtained in the PAT tree were used. Very significant improvements were

(6)

Character Accuracy for ASR Baseline 87.99% After NER approach 88.26% Table 2. Character Accuracy for ASR.

obtained (91.1 vs. 89.5) and the help from the global ev-idences was verified. Note that here the global evev-idences were used for NE candidate generation only, so very signif-icant improvements in recall rates were obtained, but at the price of slight degradation in precision rates. Such degrada-tion can actually be avoided if more deliberate NE verifica-tion techniques were applied, but left out here.

For the spoken documents, the results in part (C) are obtained with the baseline system presented in Section 4 di-rectly performed on the transcriptions of the 200 news sto-ries without global evidences considered, and the overall F1 score was 77.62. Part (D) are then the results after the global evidences recorded in the PAT tree were utilized. Very sim-ilar to the situation for the text documents, recall rates were significantly improved, resulting in a much higher overall F1 score of 80.01. The final results after filtering with con-fidence measures and the OOV recovery techniques as men-tioned in Section 3 are listed in part (E) after careful selec-tions of the three thresholds, including those for the con-fidence measure, for tf/idf score and for phone sequence matching. Comparing parts (D) and (E), both recall and pre-cision can be improved by the OOV recovery scheme, with the overall F1 score achieving 80.93. In fact, the character accuracy for the 200 news stories was also improved from 87.99% (baseline system for part (C)) to 88.26% (part(E)). This may look marginal, but the point here is that the key information in NEs was extracted much better.

6. CONCLUSION

In this paper, we propose two new approaches for NER for spoken documents. In the first approach, the PAT tree was used to extract and organize the global evidences. Such ev-idences exist everywhere and are independent of languages and domains. The second approach tried to recover the OOV NEs in the spoken documents using external knowl-edge sources by finding NE candidates in retrieved relevant text documents. Preliminary experiments on Chinese broad-cast news indicated that significant improvements were ob-tained.

7. REFERENCES

[1] D. McDonald, “Internal and external evidence in the identification and semantic categorization of proper names,” in Corpus Processing for Lexical Acquisition. 1996, MIT Press. Cambridge, MA.

[2] L.-F. Chien, “Pat-tree-based keyword extraction for chinese information retrieval,” in Proceedings of the 20th Annual International ACM SIGIR Conference on Research and Developement in Information Retrieval, 1997, pp. 50–58.

[3] G.H Gonnet, R.A. Baeza-Yates, and T. Snider, “New indices for text: Pat trees and pat arrays,” in Informa-tion Retrieval: Data Structure and Algorithms. 1992, Prentice Hall, Inc.

[4] F. Wessel, R. Schl¨uter, K. Macherey, and H. Ney, “Confidence measures for large vocabulary contin-uous speech recognition,” IEEE Transactions on Speech and Audio Processing, vol. 3, no. 9, pp. 288– 298, 2001.

[5] L. Mangu, E. Brill, and A. Stocke, “Finding consen-sus among words: Lattice-based word error minimiza-tion,” in Proc. Eurospeech99, 1999, pp. 495–498. [6] S. Cox and S. Dasmahhapatra, “High-level approaches

to confidence estimation in speech recognition,” IEEE Transactions on Speech and Audio Processing, vol. 10, no. 7, pp. 460–471, 2002.

[7] P.-C. Chang and L.-S. Lee, “Improved language model adaptation using existing and derived exter-nal resources,” in Proceedings of the IEEE Auto-matic Speech Recognition and Understanding Work-shop, 2003, pp. 531–536.

[8] D. M. Bikel, R. schwartz, and R. M. Weischedel, “An algorithm that learns what’s in a name,” Machine Learning (Special Issue on Natural Language Pro-cessing), vol. 34, no. 3, pp. 211–231, 1999.

[9] M.-Y Tsai and L.-S Lee, “Pronunciation variation analysis based on acoustic and phonemic distance measures with application examples on mandarin chi-nese,” in IEEE Wordkshop on Automatic Speech Recognition and Understanding, 2003, pp. 117–122. [10] J. Sun, M. Zhou, and J.-F. Guo, “A class-based

lan-guage model approach to chinese named entity identi-fication,” International Journal of Computaional lin-guistics and Chinese Language Processing, vol. 8, no. 2, pp. 1–28, 2003.

數據

Fig. 1. An example of a complete binary search tree
Fig. 4. An example of the PAT tree and the inverse PAT tree
Fig. 5. Query construction based on the word graph by con- con-fidence measures and tf ·idf scores
Table 1. baseline in part (A), with global evidences by PAT trees in part (B) for text documents; and baseline in part (C), with global evidences in part (D) and OOV recovering in part (E) for spoken documents.

參考文獻

相關文件

Faced with the external impact such as the US-China trade war and the Covid-19 pandemic, many companies in Yunlin County express that orders will not be affected

In this paper, we would like to characterize non-radiating volume and surface (faulting) sources for the elastic waves in anisotropic inhomogeneous media.. Each type of the source

Results for such increasing stability phenomena in the inverse source problems for the acoustic, electromagnetic, and elastic waves can be found in [ABF02, BLT10, BHKY18, BLZ20,

Teachers may consider the school’s aims and conditions or even the language environment to select the most appropriate approach according to students’ need and ability; or develop

Robinson Crusoe is an Englishman from the 1) t_______ of York in the seventeenth century, the youngest son of a merchant of German origin. This trip is financially successful,

In view of the large quantity of information that can be obtained on the Internet and from the social media, while teachers need to develop skills in selecting suitable

The entire moduli space M can exist in the perturbative regime and its dimension (∼ M 4 ) can be very large if the flavor number M is large, in contrast with the moduli space found

When using this image in external sources it can be cited as:Blausen.com staff. Wikiversity Journal