• 沒有找到結果。

1.Introduction 2.PreviousWork Abstract ash,picard @media.mit.edu AshishKapoorandRosalindW.PicardMITMediaLaboratory20AmesStreet,Cambridge,MA02139 Real-Time,FullyAutomaticUpperFacialFeatureTracking

N/A
N/A
Protected

Academic year: 2022

Share "1.Introduction 2.PreviousWork Abstract ash,picard @media.mit.edu AshishKapoorandRosalindW.PicardMITMediaLaboratory20AmesStreet,Cambridge,MA02139 Real-Time,FullyAutomaticUpperFacialFeatureTracking"

Copied!
6
0
0

加載中.... (立即查看全文)

全文

(1)

MIT. Media Laboratory Affective Computing Technical Report #550

Appears in Proceedings from 5th International Conference on Automatic Face and Gesture Recognition, May 2002

Real-Time, Fully Automatic Upper Facial Feature Tracking

Ashish Kapoor and Rosalind W. Picard MIT Media Laboratory

20 Ames Street, Cambridge, MA 02139 ash, picard



@media.mit.edu

Abstract

Robust, real-time, fully automatic tracking of facial fea- tures is required for many computer vision and graphics ap- plications. In this paper, we describe a fully automatic sys- tem that tracks eyes and eyebrows in real time. The pupils are tracked using the red eye effect by an infrared sensitive camera equipped with infrared LEDs. Templates are used to parameterize the facial features. For each new frame, the pupil coordinates are used to extract cropped images of eyes and eyebrows. The template parameters are recovered by PCA analysis on these extracted images using a PCA basis, which was constructed during the training phase with some example images. The system runs at 30 fps and requires no manual initialization or calibration. The system is shown to work well on sequences with considerable head motions and occlusions.

1. Introduction

Facial feature tracking has attracted a lot of attention due to its applications in fields of computer vision and graphics and a lot of research has been done to capture the commu- nicative ability of the face. Applications like facial expres- sion analysis, animation and coding need to track the facial features robustly and efficiently. Also, rather then just track- ing the positions of facial features, we need to get the shape information as well. The variability in appearance of facial features changes due to pose, lighting, facial expressions etc making the task difficult and complex.

Even harder is the task of tracking the facial features ro- bustly in real time, without any manual alignment or cal- ibration. Many previous approaches have focused just on tracking the location of the facial features or require some manual initialization/intervention. In this paper, we de- scribe a fully automatic system that requires no manual in- tervention and robustly tracks upper facial features in detail

using templates in real time at 30 fps.

2. Previous Work

There is much prior work on detecting and tracking the facial features. Many feature extraction methods are based on deformable templates [10], which are difficult to use for real-time tracking and have to be initialized properly to achieve a good performance. Tian et al [8, 9] use multiple- state templates to track the facial features. Feature point tracking together with masked edge filtering is used to track the upper facial features. The system requires that tem- plates be manually initialized in the first frame of the se- quence, which prevents it from being automatic. Essa et al [4] analyze the facial expressions using optical flow in an estimation and control framework coupled with a physical model describing the skin and muscle structure of face. The shortcoming of their system is that it does not recover the detailed description of the facial features.

The tracking of facial features in detail requires recover- ing the parameters that drive the shape of the feature. A lot of research has been directed towards recovering shape us- ing image matching techniques. Jones and Poggio [6] used a stochastic gradient-descent based technique to recover the shape and the texture parameters. Cootes et al [1] use active appearance models, which are statistical appearance models generated by combining a model of shape variation with a model of texture variation, for the purpose of image match- ing. The iterative nature of both these approaches and per- formance dependence on the initialization makes it difficult to use them in real time. Covell et al [2, 3] exploit the cou- pling between the shape and the texture parameters using example images labeled with control points. The shape is recovered from the image appearance in an non-iterative way using eigen analysis. Given an image of the face, this approach first needs to estimate the location of features of interest, which is again very difficult to do robustly in real time.

1

(2)

Morimoto et al [7] have described a system to detect and track pupils using the red-eye effect. Haro et al [5] have extended this system to detect and track the pupils using a Kalman filter and probabilistic PCA. Our infrared camera equipped with infrared LEDs, which is used to highlight and track pupils, is an in-house built version of the IBM Blue Eyes camera (http://www.almaden.ibm.com/cs/blueeyes).

Our system aims to combine the best of these methods, with several new improvements. We start with a simpli- fied version of Covell’s approach of estimating both shape and texture [2, 3]. Her method has a problem with requir- ing a good first estimate of the feature points; we solve this problem by using an infrared camera equipped with infrared LEDs, an in-house built version of the IBM Blue Eyes camera (http://www.almaden.ibm.com/cs/blueeyes), together with a combined detection/tracking procedure we developed. Thus, the part of our approach that locates the pupils is perhaps most similar to that of Morimoto et al [7]

and Haro et al [5], while the feature tracking is most simi- lar to that of Covell. However, the combination is new, and involves some simplifications that allow it to work in real time without any initialization. The method presented here is the first we know of that combines the strengths of robust IR pupil detection with flexible learning of estimators for shape and texture for feature point tracking.

The next section describes the overall architecture of the system, including our approach for detection and tracking of facial features. Followed by that, we demonstrate track- ing results and discuss some experimental evaluations.

3. The Overall System

The overall system is shown in figure 1. An infrared sensitive camera synchronized with infrared LEDs is used as a sensor and produces an image with highlighted pupils. The image obtained from the sensor is used to extract the position of the pupils. The location of pupils are used to locate other facial features and recover their the shape information using a PCA based technique. The whole system is very efficient and runs in real time at 30 fps.

The pupil tracking system is shown in figure 2. The whole unit is placed under the monitor pointing towards the users face. The system has an infrared sensitive camera coupled with two concentric rings of infrared LEDs. One set of LEDs is on the optical axis and produces the red-eye effect. The other set of LEDs, which are off axis, keeps the scene at about the same illumination. The two sets of LEDs are synchronized with the camera and are switched

Tracked Pupils Tracked Templates Infrared camera

capture video

Pupil Tracking Facial Feature Tracking with IR LEDs to

 !"$#%'&)(*%,+.-/-103254768%9 !

;:3!< +.9=>?+@67A;6B,+%C5DFEE/-4HGEI-/+>C5JKLIJI%

6B#.M9FALIN68A%8!

on and off to generate two interlaced images for a single frame. The image where the on-axis LEDs are on has white pupils whereas the image where the off-axis LEDs are on has black pupils. These two images are subtracted to get a difference image, which is used to track the pupils. Figure 3 shows a sample image, the de-interlaced images and the difference image obtained using the system.

PORQ;S TVUWX)5Y1Z%[U\

The pupils are detected and tracked using the difference image, which is noisy due to the interlacing and motion ar- tifacts. Also, objects like glasses and earrings can show up as bright spots in the difference image due to their specu- larity. To remove this noise we first threshold the difference image using an adaptive thresholding algorithm [5]. First, the algorithm computes the histogram and then thresholds the image keeping only 0.1 % of the brightest pixels. All the non-zero pixels in the resulting image are set to 255 (max- val). The thresholded image is used to detect and to track the pupil. The pupil tracker is either in a detection mode or a tracking mode. Whenever there is information about the pupils in the previous frame the tracker is in tracking mode and whenever the previous frame has no information about the pupils the tracker switches to the detection mode. The pupil tracking algorithm is shown in figure 4.

During the tracking mode the tracker maintains a state

(3)

vector, comprised of the spatial information about the pupils. Specifically, the average distance between the pupils during the current tracking phase and their x, y coordi- nates in the previous frames is maintained. To obtain the new positions of pupils a search for the largest connected component is limited to a bounding box centered on previ- ous pupils. The new connected components are accepted as valid pupils when they satisfy a number of spatial con- straints. If the area is greater and the displacement of their centers from previous pupil position lies below a certain threshold, the connected components are considered valid.

Also if a connected component is found for both the eyes then the distance between these pupils is also compared with the average distance maintained in the state space to rule out false detections. Once the connected components are identified as valid pupil regions, the state vector is up- dated.

The tracker switches to the detection mode whenever there is no information about the pupils. In this mode the tracker simply selects the two largest connected com- ponents that have an area greater then a certain threshold.

Again, to validate the regions, we apply some spatial con- straints. This approach allows us to track the pupils effi- ciently. Head movements during head nods and head shakes do produce motion artifacts but due to the nature of our al- gorithm to spatially constrain the search space, it tracks the pupils well. In extreme cases when head movements are too fast, the pupils are lost as motion artifacts overpower the red-eye effect and the pupils are absent from the differ- ence image altogether. We found this tracking algorithm to be fairly reliable.

  Y1ZUYW  $>Y1Z%[U \

We use templates to represent the detailed shape infor- mation of the facial features. Eye and eyebrow templates, used in our system, are shown in figure 5. A set of 8 points placed on the eye contour describes the shape of an eye.

Two of these points correspond to the eye corners and the rest are equidistant on the contour. Similarly 3 points are used to describe the shape of an eyebrow. A total of 22 points (8 for each eye and 3 for each eyebrow) are used to describe the positions and shapes of upper facial features.

Tian et al [8, 9] use a template that consists of the iris loca- tion and two parameterized parabolas, which would fit the lower and the upper eyelid. They use the corners of eyes and center points on the eyelids to extract the template parame- ters. As our system tracks more points just than eye-corners and center points on the eyelid, it can represent more de- tailed and accurate information about the shape.

The pupil positions obtained are used to crop out regions of interest: two 140 x 80 pixel images of the eyes and two 170 x 80 pixel images of eyebrows. The template parame-

Image captured by the IR camera

The difference image

De−interlaced sampled image, when the on−axis LEDs are off when the on−axis LEDs are on

De−interlaced sampled image,

3!EI->67?+>C>D5L%F%4 L. 6B#. L ,,+55J@C+I9

>,+!

ters for the features are recovered by PCA analysis on these extracted images using a PCA basis, which was constructed during the training phase with some example images. We describe the method to recover these parameters below in detail.

3.3.1 Recovering Template Parameters

Our system exploits the fact that it can estimate the location of facial features very robustly in the image. Once the fa- cial features are located, regions of interest corresponding to eyes and eyebrows are cropped out and analyzed to re- cover the shape description. Given some example images, as training data, with hand marked control points that de- scribe the shapes, our approach tries to express the new fea- ture image as a linear combination of examples. We apply the same linear combination to the marked control points to recover the control points for the new image. Since the number of example images is large, we use principal com- ponent analysis (PCA) to recover the template parameters.

Given example images , let  ( ) be vectors corresponding to the marked control points on each image.

If is the mean image, then the covariance matrix of the training images can be expressed as:



! where"$# &%(' *)+ ,(' *)-). /' 10

The eigenvectors of can be computed by first computing the eigenvectors for( 2 . If344#5 % ).5 , )&-5 / 0 where56

(4)

NONE Initialize state vector to

Start in detection mode

Obtain new difference image

Tracker in tracking mode?

Search for the largest connected component around

All spatial constraints satisfied?

Set state vector to NONE Set mode to detect Set mode to tracking

Update state vector

in the whole image connected components

Search for two largest previous

pupil position only each

No

No Yes

Yes

NI !"$#. E/- 6B,+%C5D>L.$-A%?6 #9 !

represents the eigenvectors of!  , then the eigenvectors

  of



can be computed as:

  # % )  , )&-)

 / 0 = 3

As the eigenvectors are expressed as a linear combination of example images, we can place the control points on the eigen images using the same linear combination. Let be the mean of the vectors corresponding to the control points in example images and let    (   ) be the vector corresponding to the control points on an eigenvector, then:

#!2% )! , )-)!/ 0 #2%!'  )-). / '  03

To recover the vector of control points on a new image, we first express the new image as a linear combination of the eigenvectors by projecting it onto the top few eigenvectors.

/  =    

where  / '    and  is eigenvector.The same linear combination is applied to the vectors of control points on the eigenvectors to recover the new control points.

 /  =   !  

Eyebrow Template

Eye Template

I, ! 2  +ILIJ!2 "5?A# ".9;E-N+>68*4H!

This strategy is a simplification of the approach used by Covell et al[2, 3]. The non-iterative nature of the approach makes it ideal to be used in a real-time system.

In our system, 150 images of eyes and eyebrows from ten different individuals with different facial expressions and different lighting conditions were used in the training set.

These images were hand marked to fit the feature templates.

Eigenvectors and control points on those eigenvectors were computed as described. During the real-time tracking the extracted images of the features are projected on the first forty eigenvectors. These projections are used to recover the control points using the approach explained above.

4. Results

The system was implemented and worked in real time at 30 fps on a Pentium-III 933 MHz Linux machine. The sys- tem was tested on many different subjects in different light- ing conditions. The system worked particularly well on the subjects who had their images in the training database.The system is very efficient and is robust to occlusions and re- covers very quickly when the feature reappears.

Figure 6 and 7 show tracking results of some sequences.

Both the subjects appearing in figure 6 were in the training database. The system is able to track the features very well.

Note that in the first sequence of figure 6 the left eyebrow is not tracked in frames 67, 70 and 75 as it is not present in the image. Similarly all the templates are lost in the frame 29 in the second sequence of figure 6 when the pupils are absent, as the subject blinks. The templates are recovered as soon as the features reappear. To evaluate the system per- formance, eye and eyebrow corners were handmarked in the 93 frames of the first sequence appearing in figure 6. These points were compared with the points tracked automatically by the system. Figure 8 shows the frame by frame RMS difference per control point between the points manually marked and points tracked by the system for each frame.

The mean RMS difference for the whole sequence is 0.65 pixels per control point location.

Figure 7 shows the tracking results for the subjects not

(5)

Frame 57 Frame 67 Frame 70 Frame 75 Frame 88

Frame 27 Frame 28 Frame 29 Frame 30 Frame 31

3I3!1"?+>C>D5L%*4 -684 A% 0 " 5C6 4XLK"?+%LIL.K03*6 !

Frame 59 Frame 60 Frame 61 Frame 68 Frame 87

Frame 25 Frame 28 Frame 38 Frame 45 Frame 55

I,3!1" ,+%C5D>L% 54 I-684 ,A>$0 " >CH684XLIA.6 L'" ,+.LL%K0356 !

(6)

in the training set. Again, note that the second frame in the first sequence does not show any eyes or eyebrows, due to the fact that the subject blinked and hence no pupils were detected. The tracking is recovered in the very next frame when the pupils are visible again. The first sequence appear- ing in figure 7, which is 100 frames long, was hand marked to locate the positions of inner and outer corners of eyes and eyebrows . Figure 9 shows framewise RMS difference per control point between the hand marked and the tracked points. The mean RMS difference for the whole sequence is 0.78 pixels per control point.

The results show that the system is very efficient, runs in real time at 30 fps and is able to track upper facial features robustly in presence of large head motions and occlusions.

One limitation of our implementation is that it is not invari- ant to large zooming in or out as our training set did not have samples with scale changes. Also in few cases with some new subjects, the system did not work well, as the training images were not able to span the whole range of variations in appearance of the individuals. A training set which captures the variations in appearance should be able to overcome these problems.

Frame Count

RMS Difference in Pixels

I,3! 0)%%LICH L *%-4 ,A% 47



%.L.CH54 #)6 #'0 " 5C6 4 L "?+%LIL. 03*6 !

Frame Count

RMS Difference in Pixels

I, 3! 0)%%LICH L *%-4 ,A% 47



%.L.CH54 #)6 #'0 " 5C6 4 LIA.6 L " ,+.LIL%K0356 !

5. Conclusions and Future Work

We have described a real-time system that tracks up- per facial features robustly without any manual alignment or calibration. An infrared camera equipped with infrared LEDs is used to track the pupils. The pupil positions are used to extract the images of eyes and eyebrows. Principal component analysis on these images is used to recover the template parameters. The system is shown to work well on sequences with considerable head motions and occlusions.

The work presented in this paper is work in progress to build a fully automatic facial expression analysis tool. Fu- ture work includes extending the system to track lower fa- cial features and using it as an interface to build socially and emotionally intelligent systems.

6. Acknowledgements

We thank Dave Koons and Ted Selker for their help to build the IBM Blue Eyes camera. This research was sup- ported by NSF ROLE grant 0087768.

References

[1] T. F. Cootes, G. J. Edwards, and C. J. Taylor. Active appear- ance models. Pattern Analysis and Machine Intelligence, 23(6), June 2001.

[2] M. Covell. Eigen-points. In Proceedings of International Conference Image Processing, September 1996.

[3] M. Covell. Eigen-points: control-point location using prin- cipal component analyses. In Proceedings of Conference on Automatic Face and Gesture Recognition, October 1996.

[4] I. Essa, S. Basu, T. Darrell, and A. Pentland. Modeling, tracking and interactive animation of faces and heads using input from video. In Proceedings of Computer Animation Conference, 1996.

[5] A. Haro, I. Essa, and M. Flickner. Detecting and tracking eyes by using their physiological properties. In Proceedings of Conference on Computer Vision and Pattern Recognition, June 2000.

[6] M. J. Jones and T. Poggio. Multidimensional morphable models. In Proceedings of International Conference on Computer Vision, 1998.

[7] C. Morimoto, D. Koons, A. Amir, and M. Flickner. Pupil de- tection and tracking using multiple light sources. Technical report, IBM Almaden Research Center, 1998.

[8] Y. Tian, T. Kanade, and J. F. Cohn. Dual-state parametric eye tracking. In Proceedings of Conference on Automatic Face and Gesture Recognition, 2000.

[9] Y. Tian, T. Kanade, and J. F. Cohn. Recognizing upper face action units for facial expression analysis. In Proceedings of Conference on Computer Vision and Pattern Recognition, June 2000.

[10] A. Yuille, P. Haallinan, and D. S. Cohen. Feature extrac- tion from faces using deformable templates. International Journal of Computer Vision, 1992.

參考文獻

相關文件

For each time-step an image view (in which all feature estimates are projected into the estimated camera frame) is shown next to an external 3D view in which the camera can also

This paper introduces a dynamic color restora- tion method for real time image systems with digital image sensors. Our method restores image colors by a two-step procedure and it

The algorithm used is averaging algorithm, i.e., when shrinking a 512 by 512 image into a 64 by 64 image, which is 1/8 of the original size, we take 8 by 8 blocks and use the

With an infrared sensor device installed in the cushion, the acquired signals from the cushion would determine the sitting position and the time and further,

The latter one is really dangerous, since the error of f (x) may be even smaller then the error of approximating, leading to an inaccurate result.. Page 11

(18%) Determine whether the given series converges or diverges... For what values of x does the series

Consider S 1 as the unit circle in the complex plane.. Prove that F is an injective immersion with

The tree of my title is not a green tree with leaves. It is the name of an abstract structure. In order to relate these abstract structures to the nature of the city , 1 must first