• 沒有找到結果。

Global Coordination of Local Linear Models

N/A
N/A
Protected

Academic year: 2022

Share "Global Coordination of Local Linear Models"

Copied!
8
0
0

加載中.... (立即查看全文)

全文

(1)

Global Coordination of Local Linear Models

Sam Roweisy, Lawrence K. Saulyy, and Geoffrey E. Hintony

yDepartment of Computer Science, University of Toronto

yyDepartment of Computer and Information Science, University of Pennsylvania

Abstract

High dimensional data that lies on or near a low dimensional manifold can be de- scribed by a collection of local linear models. Such a description, however, does not provide a global parameterization of the manifold—arguably an important goal of unsupervised learning. In this paper, we show how to learn a collection of local linear models that solves this more difficult problem. Our local linear models are represented by a mixture of factor analyzers, and the “global coordi- nation” of these models is achieved by adding a regularizing term to the standard maximum likelihood objective function. The regularizer breaks a degeneracy in the mixture model’s parameter space, favoring models whose internal coor- dinate systems are aligned in a consistent way. As a result, the internal coor- dinates change smoothly and continuously as one traverses a connected path on the manifold—even when the path crosses the domains of many different local models. The regularizer takes the form of a Kullback-Leibler divergence and illustrates an unexpected application of variational methods: not to perform ap- proximate inference in intractable probabilistic models, but to learn more useful internal representations in tractable ones.

1 Manifold Learning

Consider an ensemble of images, each of which contains a face against a neutral back- ground. Each image can be represented by a point in the high dimensional vector space of pixel intensities. This representation, however, does not exploit the strong correlations between pixels of the same image, nor does it support many useful operations for reasoning about faces. If, for example, we select two images with faces in widely different locations and then average their pixel intensities, we do not obtain an image of a face at their average location. Images of faces lie on or near a low-dimensional, curved manifold, and we can represent them more usefully by the coordinates on this manifold than by pixel intensi- ties. Using these “intrinsic coordinates”, the average of two faces is another face with the average of their locations, poses and expressions.

To analyze and manipulate faces, it is helpful to imagine a “magic black box” with levers or dials corresponding to the intrinsic coordinates on this manifold. Given a setting of the levers and dials, the box generates an image of a face. Given an image of a face, the box deduces the appropriate setting of the levers and dials. In this paper, we describe a fairly general way to construct such a box automatically from an ensemble of high-dimensional vectors. We assume only that there exists an underlying manifold of low dimensionality and that the relationship between the raw data and the manifold coordinates is locally linear and smoothly varying. Thus our method applies not only to images of faces, but also to many other forms of highly distributed perceptual and scientific data (e.g., spectrograms of speech, robotic sensors, gene expression arrays, document collections).

(2)

2 Local Linear Models

The global structure of perceptual manifolds (such as images of faces) tends to be highly nonlinear. Fortunately, despite their complicated global structure, we can usually char- acterize these manifolds as locally linear. Thus, to a good approximation, they can be represented by collections of simpler models, each of which describes a locally linear neighborhood[3, 6, 8]. For unsupervised learning tasks, a probabilistic model that nicely captures this intuition is a mixture of factor analyzers (MFA)[5]. The model is used to describe high dimensional data that lies on or near a lower dimensional manifold. MFAs parameterize a joint distribution over observed and hidden variables:

P(x;s;z

s

)=P(xjs;z

s )P(z

s

js)P(s); (1)

where the observed variable,x2RD, represents the high dimensional data; the discrete hidden variables,s2f1;2;:::;Sg, indexes different neighborhoods on the manifold; and the continuous hidden variables, zs

2R

d, represent low dimensional local coordinates.

The model assumes that data is sampled from different neighborhoods on the manifold with prior probabilitiesP(s)=ps, and that within each neighborhood, the data’s local coordinates are normally distributed1as:

P(z

s

js)=(2) d=2

exp



1

2 z

s

>

z

s



: (2)

Finally, the model assumes that the data’s high and low dimensional coordinates are related by linear processes parameterized by centers

s

, loading matricessand noise levels s:

P(xjs;z

s

)=j2

s j

1=2

exp



1

2 [x 

s



s z

s

>

1

s

[x 

s



s z

s



: (3) The marginal data distribution,P(x), is obtained by summing/integrating out the model’s discrete and continuous latent variables. The result is a mixture of Gaussian distributions with parameterized covariance matrices of the form:

P(x)= X

s p

s j2(

s



>

s +

s )j

1=2

exp



1

2 [x 

s

>

(

s



>

s +

s )

1

[x 

s



:

(4) The learning problem for MFAs is to estimate the centers

s

, transformations s, and noise levels sof these linear processes, as well as the prior probabilitiespsof sampling data from different parts of the manifold. Parameter estimation in MFAs can be handled by an Expectation-Maximization (EM) algorithm[5] that attempts to maximize the log- probability,logP(x), averaged over training examples.

Note that the parameter space of this model exhibits an invariance: takings

!

s R

s, whereRsareddorthogonal matrices (Rs

R

>

s

=I), does not change the marginal dis- tribution,P(x). The transformationss

!

s R

s correspond to arbitrary rotations and reflections of the local coordinates in each linear model. The objective function for the EM algorithm is unchanged by these transformations. Thus, maximum likelihood estimation in MFAs does not favor any particular alignment; instead, it produces models whose inter- nal representations change unpredictably as one traverses connected paths on the manifold.

Can we encourage models whose local coordinate systems are aligned in a consistent way?

3 Global Coordination

Suppose the data lie near a smooth manifold with a locally flat (developable) structure.

Then there exist a single set of “global coordinates”g which parametrize the manifold

1Although in principle each neighborhood could have a different prior on its local coordinates, without loss of generality we have made the standard assumption thatP(zsjs)is the same for all settings ofsand absorbed the shape of each local Gaussian model into the matricess.

(3)

s,z

g x

global hidden variables

coordinates data

Figure 1: Graphical model for globally coordinated MFAs. Al- though global coordinatesgare unobserved, they affect the learn- ing through a regularization term. After learning, inferences about the global variables are made by computing posterior distributions,

P(gjx). Likewise, data can easily be generated by sampling from the conditional distribution,P(xjg). All these operations are partic- ularly tractable due to the conditional independencies of the model.

everywhere. Furthermore, to a good approximation, these global coordinates can be related to the local coordinates of different neighborhoods (in their region of validity) by linear2 transformations:

g(s;z

s ) = A

s z

s +

s

: (5)

What does it mean to say that the coordinatesg(s;zs

)provide a global parameterization of the manifold? Intuitively, if a data point belongs to overlapping neighborhoods, then the global coordinates computed from their local coordinate systems, given by eq. (5), should agree. We can formalize this “global coordination” of different local models by treating the coordinatesgas unobserved variables and incorporating them into the probabilistic model:

P(gjs;z

s

) =Æ(g A

s z

s



s

); (6)

(Here we posit a deterministic relationship between local and global coordinates, although it is possible to add noise to this mapping as well.) The globally coordinated MFA is rep- resented by the graphical model in Fig. 1. We can appeal to its conditional independencies to make other useful inferences. In particular:

P(gjx

n

;s) = Z

dz

s

P(gjs;z

s )P(z

s js;x

n

) (7)

P(gjx

n ) =

X

s P(sjx

n )P(gjx

n

;s): (8)

Now, if two or more mixture components—say,s1ands2—explain a data pointxnwith non-negligible probability, then the posterior distributions for the global coordinates of this data point, as induced by eq. (8), should be nearly identical: that is,P(gjxn

;s

1 ) 

P(gjx

n

;s

2

). To enforce this criterion of agreement, we need to penalize models whose posterior distributionsP(gjxn

)given by eq. (8) are multimodal, since multiple modes only arise when different mixture components give rise to inconsistent global coordinates. While directly penalizing multimodality ofP(gjxn

)is difficult, a penalty which encourages con- sistency can be easily incorporated into the learning algorithm. We introduce a family of unimodal distributions over bothgands, and encourage the true posteriors,P(g;sjxn

), to be close to some member,Q(g;sjxn

), of this family.

Developing this idea further, we introduce a new objective function for unsupervised learn- ing in MFAs. The new objective function incorporates a regularizer to encourage the global consistency of local models:

= X

n

logP(x

n ) 

X

ns Z

dgQ(g;sjx

n )log



Q(g;sjx

n )

P(g;sjx

n )



; (9)

The first term in this objective function computes the log-probability of the data. The second term computes a sum of Kullback-Leibler (KL) divergences; these are designed to

2Without loss of generality, the matricesAscan be taken to be symmetric and positive-definite, by exploiting the polar factorization and absorbing reflection and rotation into the local coordinate systems. (In practice, though, it may be easier to optimize the objective function without constraining the matrices to be of this form.) In the experiments reported below, we have further restricted them to be diagonal. Together, then, the coordination matricesAsand vectorssaccount for an axis-aligned scaling and uniform translation between the global and local coordinate systems.

(4)

penalize MFAs whose posterior distributions over global coordinates are not unimodal. The twin goals of density estimation and manifold learning in MFAs are pursued by attempting to balance these terms in the objective function. The factorcontrols the tradeoff between density modeling and global coordination: as!0only strict invariances (which do not affect likelihood) are exploited in order to achieve submodel agreement. In what follows we have set=1arbitrarily; further optimization is possible.

The most convenient way to parameterize the family of unimodal distributions is a factor- ized form involving a Gaussian density and a multinomial:

Q(g;sjx

n

)=Q(gjx

n )Q(sjx

n

); Q(gjx

n

)N(g

n

;

n

); Q(sjx

n )=q

ns (10) Note that the distributionQ(g;sjxn

)in eq. (10) factorizes oversandg, implying that—

according to this family of models—the global coordinategis independent of the mixture componentsgiven the data point xn. Also,Q(gjxn

)is Gaussian, and thus unimodal.

These are exactly the constraints we wish to impose on the posteriorP(g;sjxn

). At each iteration of learning, the meansg

n

, covariance matricesn, and mixture weightsqnsare determined separately for each data point,xn so as to maximize the objective function in eq. (9): this amounts to computing the unimodal distributions,Q(g;sjxn

), best matched to the true posterior distributions,P(g;sjxn

).

4 Learning Algorithm

Latent variable models are traditionally estimated by maximum likelihood or Bayesian methods whose objective functions do not reward the interpretability of their internal rep- resentations. Note how the goal of developing more useful internal representations has changed the learning problem in a fundamental way. Now we have additional “coordina- tion” parameters–the offsetssand weightsAs–that must also be learned from examples.

We also have auxiliary parameters for each data point—the meansg

n

, covariance matri- cesn, and mixture weightsqns—that determine the target distributions,Q(g;sjxn

). All these parameters, as well as the MFA model parametersfps

;

s

;

s

;

s

g, must be chosen to “stitch together” the local coordinates systems in a smooth way and to learn internal representations easily coordinated by the local-to-global mapping in eq. (6).

Optimization of the objective function in eq. (9) is reminiscent of so-called “variational”

methods for approximate learning[7]. In these methods, an approximation to an exact (but intractable) posterior distribution is fitted by minimizing a KL divergence between the two distributions. The auxiliary parameters of the approximating distribution are known as variational parameters. Our objective function illustrates an unexpected application of such variational methods: not to perform approximate inference in intractable probabilistic models, but to learn more useful internal representations in tractable ones. We introduce the unimodal and factorized distributionsQ(g;sjxn

)to regularize the multimodal distributions

P(g;sjx

n

). Penalizing the KL divergence between these distributions lifts a degeneracy in the model’s parameter space and favors local linear models that can be globally aligned.

4.1 Computing and optimizing the objective function

Evaluating the objective function in eq. (9) requires a sum and integral over the latent vari- ables of the model. These operations are simplified by rewriting the objective function as:

= X

ns Z

dgQ(g;sjx

n

)[ logQ(g;sjx

n

)+logP(x

n

;g;s)℄: (11) The factored form of the distributionsQ(g;sjxn

)makes it straightforward to perform the required sums and integrals. The final result is a simple form in terms of entropiesSnsand

(5)

energiesEnsassociated with thenth data point:

 =

X

ns q

ns (S

ns E

ns

); (12)

S

ns

= 1

2 logj

n

j logq

ns +

d

2

log(2); (13)

E

ns

= 1

2 g

>

ns V

s g

ns +

1

2 x

>

ns

1

s x

ns g

>

ns A

>

s



>

s

1

s x

ns +

1

2 Tr[

n V

s

+ 1

2 logj

s

j+logjA

s

j logp

s +

D+d

2

log(2); (14)

where we have introduced simplifying notation for the vector differencesxns

=(x

n



s )

andgns

=(g

n



s

)and the local precision matricesVs

=A 1

s

I+

>

s

1

s



s



A 1

s

. Iteratively maximizing the objective function by coordinate ascent now leads to a learning algorithm of the same general style as EM.

4.2 E-step

Maximizing the objective function, eq. (9), with respect to the regularizing parameters

fg

n

;

n

;q

ns

g(and subject to the constraint

P

s q

ns

=1) leads to the fixed point equations:



n

=

"

X

s q

ns V

s

#

1

g

n

=

n

"

X

s q

ns m

ns

#

q

ns

= e

Ens

P

s 0

e E

ns 0

: (15) wheremns

= V

s



s +A

>

s



>

s

1

s x

ns. These equations can be solved by iteration with initializationqns

=p

s. Notice thatVsandmnsonly need to be computed once before iter- ating the fixed point equations. The objective function is completely invariant to translation and rescaling ofg

n

ands(sinceAs,sandg

n

appear only in the form(g

n



s )A

>

s

). To remove this degeneracy, after solving the equations above we further constrain the global coordinates to have mean zero and unit variance in each direction. These constraints are enforced without changing the value of the objective function by simply translating the offsetssand rescaling the diagonal matricesAs.

4.3 M-step

The M-step consists of maximizing the objective function, eq. (9), with respect to the generative model parameters. Let us denote the updated parameter estimates by

f~p

s

;~

s

;~

s

;

~



s

;

~

s

;

~

A

s

g. Letting qs

= P

n q

ns, the M-step updates for the first three of these are:

~ p

s q

s Æ

X

s 0

q

s 0

~



s q

1

s X

n q

ns g

n

~



s q

1

s X

n q

ns x

n

: (16) The remaining updates, to be performed in the order shown, are given in terms of updated difference vectorsx~ns

=x

n

~



s

; ~g

ns

=g

n

~



s, the correlationsCs

= P

n q

ns

~ x

ns

~ g

>

ns

, and the variancesGs

= P

n q

ns [

n +g~

ns

~ g

>

ns

.

~



s

C

s G

1

s A

s

; (17)

h

~

s i

i q

1

s X

n q

ns



h

~ x

ns

~



s A

1

s

~ g

ns i

2

i +

h

~



s A

1

s



n A

>

s

~



>

s i

i



(18)

~

A 1

s

(I+

>

s

1

s



s )

1 n

A

>

s q

s +

>

s

~

1

s C

s o

G 1

s (19)

At the optimum, the coordination weightsAssatisfy an algebraic Riccati equation which can be solved by iterating the update shown above. (Such equations can also be solved by much more sophisticated methods well known in the engineering community. Most approaches involve inverting the previous value ofAs which may be expensive for full matrices but is fast in our diagonal implementation.)

(6)

Figure 2: Global coordination of local lin- ear models. (left) A model trained using maxi- mum likelihood, with the arrows indicating the direction of increase for each factor analyzer’s local coordinate system. (right) A coordinated model; arrows indicate the direction in the data space corresponding to increasing the global coordinategas inferred by the algorithm. The ellipses show the one standard deviation con- tour of the density of each analyzer.

5 Experiments

We have tested our model on simple synthetic manifolds whose structure is known as well as on collections of images of handwritten digits and faces. Figure 2 illustrates the basic concept of coordination, as achieved by our learning rule. In the coordinated model, the global coordinate always points in the same direction along the data manifold, as defined by the composition of the transformationssandAs. In the model trained with maxi- mum likelihood, the density is well captured but each local latent variable has a random orientation along the manifold.

We also applied the algorithm to collections of images of handwritten digits and of faces.

The representation ofxwas an unprocessed vector of raw 8-bit grayscale pixel intensities for each image (of dimensionality 256 for the1616digits and 560 for the2820faces.) The MFAs had 64 local models and the global coordinates were two dimensional. After training, the coordinated MFAs had learned a smooth, continuous mapping from the plane to images of digits or of faces. This allows us both to infer a two-dimensional location given any image by computingP(gjx)and to generate new images from any point in the plane by computingP(xjg). (Precisely what we wanted from the magic box.) In general, both of these conditional distributions have the form of a mixture of Gaussians. Figure 3 shows the inferred global coordinatesg

n

(i.e. the means of the unimodal distributionsQ(gjxn )) of the training points after the last iteration of training as well as examples of new images from the generative model, created by evaluating the mean ofP(xjg)along straight line paths in the global coordinate space. In the case of digits, it seems as though our models have captured tilt/shape and identity and represented them as the two axes of thegspace; in the case of the faces the axes seem to capture pose and expression. (For the faces, the final

gspace was rotated by hand to align interpretable directions with the coordinate axes.) As with all EM algorithms, the coordinated MFA learning procedure is susceptible to lo- cal optima. Crucial to the success of our experiments is a good initialization, which was provided by the Locally Linear Embedding algorithm[9]. We clampedg

n

equal to the em- bedding coordinate provided by LLE andnto a small value and trained until convergence (typically 30-100 iterations). Then we proceeded with training using the full EM equations to updateg

n

, again until convergence (usually 5-10 more iterations). Note, however, that LLE and other embedding algorithms such as Isomap[10] are themselves unsupervised, so the overall procedure, including this initial phase, is still unsupervised.

6 Discussion

Mixture models provide a simple way to approximate the density of high dimensional data that lies on or near a low dimensional manifold. However, their hidden representations do not make explicit the relationship between dissimilar data vectors. In this paper, we have shown how to learn global coordinates that can act as an encapsulating interface, so that other parts of a learning system do not need to interact with the individual compo- nents of a mixture. This should improve generalization as well as facilitate the propagation and exchange of information when these models are incorporated into a larger (perhaps

(7)

Figure 3: Automat- ically constructed two dimensional global param- eterizations of manifolds of digits and faces. Each plot shows the global coordinate space discov- ered by the unsupervised algorithm; points indicate the inferred means gn for each training item at the end of learning. The image stacks on the borders are not from the training set but are generated from the model itself and represent the mean of the predictive distribution P(xjg) at the corresponding open circles (sampled along the straight lines in the global space).

The models provide both a two degree-of-freedom gen- erator for complex images via P(xjg) as well as a pose/slant recognition sys- tem viaP(gjx).

For the handwritten digits, the training set consisted of 1100 examples of the digit “2” (shown as crosses above) mixed with 1100 ex- amples of “3”s (shown as triangles). The digits are from the NIST dataset, dig- itized at 16x16 pixels. For the faces, we used 2000 im- ages of a single person with various poses and expres- sions taken from consecu- tive frames of a video digi- tized at 20x20 pixels. Bren- dan Frey kindly provided the face data.

hierarchical) architecture for probabilistic reasoning.

Two variants of our purely unsupervised proposal are possible. The first is to use an em- bedding algorithm (such as LLE or Isomap) not only as an initialization step but to provide clamped values for the global coordinates. While this supervised approach may work in practice, unsupervised coordination makes clear the objective function that is being opti-

(8)

Figure 4: A situation in which an un-coordinated mix- ture model–trained to do density estimation–cannot be “post- coordinated”. Noise has caused one of the local density mod- els to orient orthogonal to the manifold. In globally coordi- nated learning, there is an additional pressure to align with neighbouring models which would force the local model to lie in the correct subspace.

mized, which unifies the goals of manifold learning and density estimation. Another variant is to train an unsupervised mixture model (such as a MFA) using a traditional maximum likelihood objective function and then to “post-coordinate” its parameters by applying local reflections/rotations and translations to create global coordinates. As illustrated in figure 4, however, this two-step procedure can go awry because of noise in the original training set.

When both density estimation and coordination are optimized simultaneously there is extra pressure for local experts to fit the global structure of the manifold.

Our work can be viewed as a synthesis of two long lines of research in unsupervised learning. In the first are efforts at learning the global structure of nonlinear manifolds [1, 4, 9, 10]; in the second are efforts at developing probabilistic graphical models for rea- soning under uncertainty[5, 6, 7]. Our work proposes to model the global coordinates on manifolds as latent variables, thus attempting to combine the representational advantages of both frameworks. It differs from embedding by providing a fully probabilistic model valid away from the training set, and from work in generative topographic mapping[2] by not requiring a uniform discretized gridding of the latent space. Moreover, by extending the usefulness of mixture models,it further develops an architecture that has already proved quite powerful and enormously popular in applications of statistical learning.

Acknowledgements

We thank Mike Revow for sharing his unpublished work (at the University of Toronto) on coordinat- ing mixtures, and Zoubin Ghahramani, Peter Dayan, Jakob Verbeek and two anonymous reviewers for helpful comments and corrections.

References

[1] D. Beymer & T. Poggio. Image representations for visual learning. pringerScience 272 (1996).

[2] C. Bishop, M. Svensen, and C. Williams. GTM: The generative topographic mapping.

Neural Computation 10 (1998).

[3] C. Bregler & S. Omohundro. Nonlinear image interpolation using manifold learning.

Advances in Neural Information Processing Systems 7 (1995).

[4] D. DeMers & G.W. Cottrell. Nonlinear dimensionality reduction.

Advances in Neural Information Processing Systems 5 (1993).

[5] Ghahramani, Z. and Hinton, G. The EM algorithm for mixtures of factor analyzers.

University of Toronto Technical Report CRG-TR-96-1 (1996).

[6] Hinton, G., Dayan, P., and Revow, M. Modeling the manifolds of images of handwritten digits.

IEEE Transactions on Neural Networks 8 (1997).

[7] M. Jordan, Z. Ghahramani, T. Jaakkola, and L. Saul. An introduction to variational methods for graphical models. Machine Learning 37(2) (1999).

[8] N. Kambhatla and T. K. Leen. Dimension reduction by local principal component analysis.

Neural Computation 9 (1997).

[9] S. T. Roweis & L. K. Saul. Nonlinear dimensionality reduction by locally linear embedding.

Science 290 (2000).

[10] J. B. Tenenbaum, V. de Silva, and J. C. Langford. A global geometric framework for nonlinear dimensionality reduction. Science 290 (2000).

參考文獻

相關文件

The integral of the mean squared error of an estimator of a re- gression function is used as a criterion for defining an optimal design measure in the context of local

Regard G as an additive group, then G is a vector space over finite field Z/2Z (§4.13) and an automorphism is just a nonsingular linear transformation... In this case, using

An additional senior teacher post, to be offset by a post in the rank of CM or Assistant Primary School Master/Mistress (APSM) as appropriate, is provided to each primary

An additional senior teacher post, to be offset by a post in the rank of Certificated Master/Mistress or Assistant Primary School Master/ Mistress as appropriate, is provided

An additional senior teacher post, to be offset by a post in the rank of CM or Assistant Primary School Master/Mistress (APSM) as appropriate, is provided to each primary

An additional senior teacher post, to be offset by a post in the rank of CM or Assistant Primary School Master/Mistress (APSM) as appropriate, is provided to each primary

An additional senior teacher post, to be offset by a post in the rank of Certificated Master/Mistress or Assistant Primary School Master/Mistress as appropriate, is provided to

An additional senior teacher post, to be offset by a post in the rank of CM or APSM as appropriate, is provided to each primary special school/special school with