• 沒有找到結果。

An EM based multiple instance learning method for image classification

N/A
N/A
Protected

Academic year: 2021

Share "An EM based multiple instance learning method for image classification"

Copied!
5
0
0

加載中.... (立即查看全文)

全文

(1)

An EM based multiple instance learning method for image classification

H.T. Pao

a,*

, S.C. Chuang

b

, Y.Y. Xu

b

, Hsin-Chia Fu

b

aDepartment of Management Science, National Chiao Tung University, Hsinchu, Taiwan, ROC bDepartment of Computer Science, National Chiao Tung University, Hsinchu, Taiwan, ROC

Abstract

In this paper, we propose an EM based learning algorithm to provide a comprehensive procedure for maximizing the measurement of diverse density on given multiple Instances. Furthermore, the new EM based learning framework converts an MI problem into a single-instance treatment by using EM to maximize the single-instance responsibility for the corresponding label of each bag. To learn a desired image class, a user may select a set of exemplar images and label them to be conceptual related (positive) or conceptual unrelated (negative) images. A positive image consists of at least one object that the user may be interested, and a negative image should not contain any object that the user may be interested. By using the proposed EM based learning algorithm, an image retrieval prototype system is imple-mented. Experimental results show that for only a few times of relearning cycles, the prototype system can retrieve user’s favor images from WWW over Internet.

 2007 Published by Elsevier Ltd.

Keywords: Multiple-instance learning; Image retrieve; WWW; EM method

1. Introduction

Due to rapid advances in computer hardware and Inter-net capability, visual information, incorporating image storage and retrieval, has become a preferred source of information for many consumers. Information in a visual form differs from traditional database contents in many important ways. It requires more space for storage, is highly unstructured, and needs some kind of decoding pro-cess to determine its semantic content (Leung, Hibler, &

Mwara, 1992; Sethi, 1995). The limitation of the current

image analysis techniques necessitates that most image retrieval systems use some form of text description pro-vided by users as the basis to index and retrieve images. These techniques are rather primitive at present and they need further development and refinement (Hwarth &

Bux-ton, 2000; Lin, Wang, & Yang, 1999). Many different

methods and techniques have been recently proposed for

modeling and retrieving images (Chua, Pung, Lu, & Jong, 1994; Gamanya, Maeyer, & Dapper, 2007; Gudivada & Raghavan, 1997; Jorgensen, 1998; Shakir & Nagao, 1996;

Yang & Wu, 1997). They largely fall into two categories,

namely, content-based methods and concept-based meth-ods. Content-based image retrieval systems have been pro-posed to index and search by their low-level features (contents) such as color, texture, shape, spectral, etc. Photobook Pentland, Picard, and Sclaroff (1994), QBIC

Ashley et al. (1995), NETRA Ma and Manjunath (1997)

andTsai (2007)are such attempts to enable users to query

based on these features.Smith and Chang (1999)present a general framework for querying and retrieving images by spatial and feature attributes. The spatial and feature (SaFe) system integrates content-based techniques with spatial query methods in order to search for image by arrangement of regions. Usually, the contents of an image are very complicated, so an image can be seen as the com-bination of the small subimages, in which each subimage has its own content. For example, if an image contains an interested subimage such as ‘‘Waterfall’’ and a few

unin-0957-4174/$ - see front matter  2007 Published by Elsevier Ltd. doi:10.1016/j.eswa.2007.08.055

*

Corresponding author. Tel.: +886 35 731578; fax: +886 35 724176. E-mail address:htpao@cc.nctu.edu.tw(H.T. Pao).

www.elsevier.com/locate/eswa Expert Systems with Applications 35 (2008) 1468–1472

Expert Systems with Applications

(2)

terested subimages. One would like to identify this image as ‘‘The Image Containing a Waterfall’’. In traditional meth-ods, feature vectors are extracted from the whole image. It is very hard to extract suitable feature vectors from the whole image just to properly represent ‘‘The Image Con-taining a Waterfall’’. Thus, some methods proposed to first segment interested subimages from an image, and then extract feature vectors from the interested subimages. In fact, it is very difficult to segment the interested subimages precisely. Beside, the interested subimages in an image may be different for different users, as a consequence, different feature vectors may be extracted from different interested subimages when the same image are queried by different users. Thus, this approach complicates the system design and confuses users in selecting proper query subimages. In order to represent an image correctly, multiple-instances learning (MIL) is a way to model ambiguity in semi-super-vised learning setting, where each training example is a bag of instances and the labels are assigned on the bags instead of on the instances. After being introduced byDietterich,

Lathrop, and Lozano-Perez (1997), MIL has become an

active research area and a number of MIL algorithms have been proposed (Andrews, Tsochantaridis, & Hofmann,

2004, Wang & Zucker, 2000), etc. Maron and Ratan

(1998) proposed the Multiple-Instance learning method

to learn several instances with various diverse densities, and to maximize diverse density (DD) by Quasi-Newton method. In addition,Zhang and Goldman (2001)proposed an EM for multiple-instance learning by using Quasi-New-ton search to maximize DD. In this paper, we propose an EM based learning algorithm, which provides a compre-hensive procedure to maximizing the measurement of diverse density of the given multiple Instances. Further-more, the new EM based learning framework converts MI problem into a single-instance treatment by using EM to maximize the instance responsibility for the correspond-ing label of each bag of instances. As we can see, the pro-posed algorithm provides comprehensive procedures to maximizing the measurement of diverse density of the given multiple instances.

2. EM based multiple-instance learning algorithm

In the Multiple-Instance learning, conceptual related (positive) and conceptual unrelated (negative) images are used for reinforced and antireinforced learning of a user’s desired image class. Each positive training image contains at least one interested subimage related to the desired image class, and each negative training image should not contain any subimage related to the desired image class. The target of the Multiple-instance learning is to search the optimal point of the image class in the feature space, where the optimal point is close to the intersection of the feature vectors extracted from the subimages of the positive training images and is far from the union of the feature vec-tors extracted from the subimages of the negative training images.

For example, if one wants to train an image class t with Pt positive images and Ntnegative images. Each positive

image has Vþ

t subimages, and each negative image has

Vt subimages. We denote the kth feature vector extracted from the kth subimage of the ith positive image as Xþik, and the kth feature vector extracted from the kth subimage of the ith negative example as Xik. The probability that Xþik belongs to class t is Pðt j XþikÞ, and the probability that

Xik belongs to class t is Pðt j XikÞ. A measurement called

Diverse Density is used to evaluate that how many different positive images have feature vectors near a point t, and how far the negative feature vectors are from a point t. The Diverse Density for a class t is defined asPearl (1998)

DDt¼ YPt i¼1 ð1 Y Vþt k¼1 ð1  P ðt j XþikÞÞÞ YNt i¼1 ðY V t k¼1 ð1  P ðt j XikÞÞÞ: ð1Þ

The optimal point of the class t is appeared where the di-verse density is maximized. By taking the first partial deriv-atives of Eq. (1) with respect to parameters of the class t and setting the partial derivatives to zero, the optimal point of the class t can be obtained. Suppose the density function of the class t is a D-dimensional Gaussian mixture with uncorrelated features. The parameters are the mean ltcd,

the variance r2

tcd, and the cluster prior probability ptc of

each cluster in the class t. The estimating parameters of the class t can be derived by o

oltcdDDt¼ 0, o ortcdDDt¼ 0, and o optcDDt¼ 0. Thus ltcd¼ X Pt i¼1 Pti 1 Pti   XVþ t k¼1 QtcðXþikÞxþ ikd ! " X Nt i¼1 XVt k¼1 QtcðX ikÞx  ikd # XPt i¼1 Pti 1 Pti   Xt k¼1 QtcðXþikÞ ! " , X Nt i¼1 XVt k¼1 QtcðXikÞ # ; ð2Þ r2 tcd¼ XPt i¼1 Pti 1 Pti   Xt k¼1 QtcðXþikÞkx þ ikd ltcdk 2 ! " X Nt i¼1 XV t k¼1 QtcðXikÞkx ikd ltcdk 2 # XPt i¼1 Pti 1 Pti   Xt k¼1 QtcðXþ ikÞ ! " , X Nt i¼1 XV t k¼1 QtcðXikÞ # ; ð3Þ

(3)

ptc¼ X Pt i¼1 Pti 1 Pti   Xt k¼1 QtcðXþikÞ ! " X Nt i¼1 XVt k¼1 QtcðXikÞ # XPt i¼1 Pti 1 Pti   Xt k¼1 PðtjXþ ikÞ 1 P ðtjXþikÞ  ! " , X Nt i¼1 XV t k¼1 PðtjXikÞ 1 P ðtjXikÞ  # ; ð4Þ where PðcjXI ik; tÞ ¼ ptc P ðXI ikjc; tÞ PðtjXI ikÞ ; ð5Þ PðtjXI ikÞ ðlÞ¼PðX I ik ðlÞ jtÞPt PðXI ikÞ ; ð6Þ Pti¼ YPi k¼1 1 P tjXI ik     ; QtcðXI ikÞ ¼ PðtjXI ikÞP ðcjX I ik; tÞ 1 P ðtjXI ikÞ ; PðXI ikjc; tÞ ¼ 1 QD d¼1ð2pr2tcdÞ 1 2  exp 1 2 XD d¼1 xI ikd ltcd rtcd  2! ;

and the notation q represents + or.

According to Eqs. (2)–(4), we proposed an EM based Multiple-Instance learning algorithm to learn these param-eters. The EM based Multiple-Instance learning algorithm contains two steps: the expectation step (E-step) and the maximization step (M-step). The algorithm is described as follows.

1. Choose an initial point in the feature space, and let its parameters are lð0Þtcd, r2ð0Þtcd , and pð0Þtc .

2. E-Step: Using the calculated model parameters lðlÞtcd, r2ðlÞtcd, pðlÞtc, Eqs. (5) and (6), estimate PðlÞðcjX

I ik; tÞ and

PðlÞðtjXI

ikÞ.M-Step: Using the estimated PðlÞðcjX I ik; tÞ and

PðlÞðtjXI

ikÞ, compute the new model parameters l ðlþ1Þ tcd ,

r2ðlþ1Þtcd , and pðlþ1Þtc according to Eqs.(2)–(4).

3. Calculate the diverse density DD(l+1). If (DD(l+1) DDl) is smaller than a predefined threshold , then stop the process. Otherwise, loop step 2.

As we can see, the proposed algorithm provides compre-hensive procedures to maximizing the measurement of diverse density of the given multiple instances. Further-more, the new EM based learning framework converts MI problem into a single-instance treatment by using EM to estimate and to maximize the instance responsibility for the corresponding label of each bag of instances.

3. Image feature extraction and indexing

Before training the system for indexing images, multiple feature vectors are extracted from the multiple instances of several exemplar training images. Then, the system are trained according to the proposed EM based Multiple-Instance learning algorithm. Finally, the testing images are evaluated using Bayesian decision rule for indexing and classification.

3.1. Image feature extraction

The image features extraction we used are similar to the method proposed in (Pearl, 1998). First, a number of instances are randomly selected from an image. Then, the feature vectors are extracted from the instances as shown

inFig. 1. The feature vector in the position (i,j) is defined

as X = {x1,  ,x15}, where

• {x1, x2, x3} is the average YUV values of C.

• {x4, x5, x6} is the average YUV values of A minus

aver-age YUV values of C.

• {x7, x8, x9} is the average YUV values of B minus

aver-age YUV values of C.

• {x10, x11, x12} is the average YUV values of D minus

average YUV values of C.

• {x13, x14, x15} is the average YUV values of E minus

average YUV values of C.

It is clear to see that the proposed feature extraction provides not only the color information but also some of the spatial information.

(i,j ) A B C D E

Fig. 1. Feature vectors are extracted from ‘‘+’’ shaped subimage (instance). The instance consists of 5 subregions: A, B, C, D and E. Each subregion is composed of 2· 2 pixels. The feature vector, X = {x1, . . . , x15}, is the YUV value of C, and the difference values of C

(4)

3.2. Image indexing

After the feature vectors are extracted from the sub-images of the training sub-images, the system can be trained to perform indexing and classification for the images with respect to the class t. First, a user needs to select some related and unrelated images for a class as training images, then feature vectors of these images are extracted. Once the training feature vectors are ready, the EM based Multiple-Instance learning algorithm is used to compute the mean ltcd, the variance r2tcd, and the cluster prior probability

ptc. By using these parameters, the posterior probabilities

P(tjXi) of an unindex image can be computed for each

class. The unindex image is indexed to the class f if the P(fjXi) is the highest among all the P(tjXi).

4. Experimental results

We have built a prototype system to evaluate the pro-posed image classification and indexing method. This sys-tem is called the ‘‘Intelligent Multimedia Information Processing Systems’’ (IMIPS) (IMIPS, xxxx). This system has been used as a video search engine over the WWW. Once a new video file is found by a video spider in IMIPS,

the system will download the file and save several key frames in a database.

When a desired image class is to be trained, the positive and the negative exemplar images are selected from the stored key frames. Then, the system learns the desired image class using the proposed EM based Multiple-Instance learning algorithm. When the optimal model for desired class is trained, each key frame in the database is indexed by its posterior probability associated with the desired class.

A web-based user interface is depicted inFig. 2. Each of the title of the trained classes are display in a pulldown menu: ‘‘Select Query Type’’. When a user is trying to search an image of a certain class, one can use ‘‘Select Query Type’’ to select a class. Then, the associated images will be shown. Suppose, the class ‘‘Human’’ is selected, the system responds with all the images belonging to ‘‘Human’’ from the database in a descending order according to their computed value of posterior probability. As we can see, most of the shown images are human-related. If the user wants to create a new class, one can click the ‘‘Random’’ button, then a set of randomly selected images will be shown. One can select the ‘‘Positive’’ button to include con-ceptual related images, and select the ‘‘Negative’’ button to exclude conceptual unrelated images. When the ‘‘Refine’’

Fig. 2. When a user enters the IMIPS, one can select a class on interested of images, which are key frames of a video over the WWW. Then, the user can click on an image to view its associated video program.

(5)

button is pressed, the system will train a new class accord-ing to the selections.

In this prototype system, we have trained five classes of nature scenes: ‘‘Human’’, ‘‘Star’’, ‘‘Sky’’, ‘‘Animal’’, and ‘‘Fire’’. The experimental results are shown in Table 1. The number in the intersection of row ‘‘Animal’’ and col-umn ‘‘10’’ indicates that the correctly retrieved images is 10 out of the total 10 retrieved images. The correctness of retrieve is judged according to human perception.

5. Conclusion

In this paper, we propose an EM based Multiple-Instance learning algorithm and implement a user friendly video search engine (IMIPS) over WWW. The multiple-instances learning (MIL) is a way to model ambiguity in semi-supervised learning setting, where each training exam-ple is a bag of instances and the labels are assigned on the bags instead of on the instances. In this paper, we propose an EM based learning algorithm, which provides a compre-hensive procedure to maximizing the measurement of diverse density of the given multiple Instances. Further-more, the new EM based learning framework converts MI problem into a single-instance treatment by using EM to maximize the instance responsibility for the correspond-ing label of the bag. The experimental results show that the retrieved images is quite match with human perception. How to properly determine the correct number of the clus-ters in the mixture Gaussian model of each class is a prob-lem we want to solve in the future. In order to build a more powerful model, some features, such as shapes, tex-tures,locations and domain knowledge, etc., will also be included in the future systems.

References

Andrews, S., Tsochantaridis, I., & Hofmann, T. (2004). Support vector machines for multiple-instance learning. In NIPS.

Ashley, J., Flickner, M., Hafner, J. L., Lee, D., Niblack, W., & Petkovic, D. (1995). The query by image content (QBIC) system. In SIGMOD Conference, (p. 475).

Chua, T. S., Pung, H. K., Lu, G. J., & Jong, H. S. (1994). A concept-based image retrieval system. IEEE Computer, 590–598.

Dietterich, T. G., Lathrop, R. H., & Lozano-Perez, T. (1997). Solving the multiple-instance problem with axis-parallel rectangles. Artificial Intelligence Journal, 89.

Gamanya, Ruvimbo, Maeyer, Philippe De, & Dapper, Morgan De (2007). An automated satellite image classification design using object-oriented segmentation algorithms: A move towards standardization. Expert System with Applications, 32(2), 616–624.

Gudivada, Venkat N., & Raghavan, Vijay V. (1997). Modeling and retrieving images content system. Information Processing and Man-agement, 33(4), 427–452.

Hwarth, R. J., & Buxton, H. (2000). Conceptual-description from monitoring and watching image sequences. Image and Vision Comput-ing, 18, 105–135.

IMIPS: The intelligent multimedia information processing system,http:// imips.csie.nctu.edu.tw/imips/imips.html.

Jorgensen, Corinne (1998). Corinne jorgensen: Attributes of images in describing tasks. ACM Transactions on Information Systems, 34(2), 61–174.

Leung, C. H. C., Hibler, D., & Mwara, N. (1992). Image retrieval by content description. Journal of Information Science, 18, 111–119. Lin, H. C., Wang, L. L., & Yang, S. N. (1999). Regular-texture image

retrieval based on texture-primitive extraction. Image and Vision Computing, 17, 51–63.

Ma, W.Y., & Manjunath, B.S. (1997).NETRA: A toolbox for navigating

large image databases. In Proceedings of the IEEE International Conference Image Processing 1, (Vol. 1, pp. 568, V571).

Maron, O., & Lakshmi Ratan, A. (1998). Multiple-instance learning for natural scene classification. In Machine learning: Proceedings of the 15th international conference.

Pearl, Judea (1998). Probablistic reasoning in intelligent systems: Networks of plausible inference. Morgan Kaufmann.

Pentland, A., Picard, R. W., & Sclaroff, S. (1994). Photobook: Content-based manipulation of image databases. In Proceedings of the SPIE Storage Retrieval Image Video Databases II, (pp. 34–47).

Sethi, Ishwar K. (1995). Image computing for visual information systems. IEEE Computer, 06–10.

Shakir, Hussain Sabir, & Nagao, Makoto (1996). Context-sensitive processing of semantic queries in an image database system. Informa-tion Processing and Management, 32(5), 573–600.

Smith, J. R., & Chang, S. F. (1999). Integrated spatial and feature image query, http://disney.ctr.columbia.edu/safe/. Multimedia System, 7(2):129–140.

Tsai, Chih-Fong (2007). Image mining by spectral features: A case study of scenery image classification l. Expert system with applications, 32, 135–142.

Wang, J. & Zucker, J. D. (2000). Solving the multiple-instance problem: A lazy learning approach. In Proceedings of the 17th international conference on machine learning, (pp. 1119–1125).

Yang, Li, & Wu, Jiankang (1997). Towards a semantic image database system. Data and Knowledge Engineering, 22, 207–227.

Zhang, Q. & Goldman, S. A. (2001) Em-dd: An improved multiple-instance learning technique. In Neural Information Processing Systems. Table 1

Results of natural scenes classification Total of retrieve Templates 10 20 30 40 50 Animal 10 20 30 39 48 Human 10 15 18 22 23 Sky 7 13 19 26 29 Star 10 20 30 35 40

Fire 9 13 N/A N/A N/A

The number in each row indicates the correctly retrieved images of each classes with respect to total retrieved images.

數據

Fig. 1. Feature vectors are extracted from ‘‘+’’ shaped subimage (instance). The instance consists of 5 subregions: A, B, C, D and E
Fig. 2. When a user enters the IMIPS, one can select a class on interested of images, which are key frames of a video over the WWW

參考文獻

相關文件

volume suppressed mass: (TeV) 2 /M P ∼ 10 −4 eV → mm range can be experimentally tested for any number of extra dimensions - Light U(1) gauge bosons: no derivative couplings. =>

For pedagogical purposes, let us start consideration from a simple one-dimensional (1D) system, where electrons are confined to a chain parallel to the x axis. As it is well known

Courtesy: Ned Wright’s Cosmology Page Burles, Nolette & Turner, 1999?. Total Mass Density

The observed small neutrino masses strongly suggest the presence of super heavy Majorana neutrinos N. Out-of-thermal equilibrium processes may be easily realized around the

incapable to extract any quantities from QCD, nor to tackle the most interesting physics, namely, the spontaneously chiral symmetry breaking and the color confinement.. 

• Formation of massive primordial stars as origin of objects in the early universe. • Supernova explosions might be visible to the most

The difference resulted from the co- existence of two kinds of words in Buddhist scriptures a foreign words in which di- syllabic words are dominant, and most of them are the

• elearning pilot scheme (Four True Light Schools): WIFI construction, iPad procurement, elearning school visit and teacher training, English starts the elearning lesson.. 2012 •