• 沒有找到結果。

A computer interface for the disabled by using real-time face recognition

N/A
N/A
Protected

Academic year: 2021

Share "A computer interface for the disabled by using real-time face recognition"

Copied!
3
0
0

加載中.... (立即查看全文)

全文

(1)

Proceedings of the 25" Annual international Conference of the IEEE EMBS Cancun, Mexico * September 17-21,2003

A

Computer Interface for the Disabled

by

Using Real-Tiime Face Recognition

Cheng-Yao Chen, and Jyh-Homg Chen

Dept. of Electrical Engineering, National Taiwan University, Taipei, 'Taiwan, R.O.C.

Abstract-In this paper we present a computer interface for the disabled by using real-time face recognition algorithm. Based on an adaptive color model, the system allows complex backgrounds (including objects and onlookers) and varying illuminations. Our refined control technique is also robust when suffering the user position shifts. Experiment shows that our system can significantly increase the efficiency for the disabled to deal with internet information and multimedia entertainment.

Keywords- Computer interface, the disabled, real-time, face recognltion

I. INTRODUCTiON

An efficient computer interface for the disabled has long been thought

as

a crucial topic in the field of medical electronics. Professionals around the world have proposed many solutions. However, most of them require wearing extra instruments, such as infrared appliance, headset with camera, artificial feature attachment, and etc.

Thanks

to the progresses in the field of face recognition

[I]

and computer hardware, we build up a man-machine interface by simply using a US9 digital PC camera (Fig. 1 .). The disabled can not only be trained to use it easily and comfortably, but also enrich their lives by the information exchanged through the intemet and multimedia entertainment.

Our

system can be functionally divided into two stages: face recognition and mousdaction control (Fig. 2.). We discuss each block in details in the following paragraph.

Fig. i

.

System Overview

I

I

Image input

n

-

-1 ,

.

Face recognition

Skin-color Eyes and lips

detection detection

Mouselaction control

L

a

b

Relative motion

of

the reference point selection

Fig. 2. System block dia- 11. METHODOLOGY

RGB model is natural1 to human visual perception, but because of its intensity inherited property it is not suitable for computer vision. When the illumination changes, the skin color will cause

a

non-linear effect

in

the RGB representation.

As

a result, we use the normalized

RGB

(range form 0 to I), and transfer it to the HSI (hue, saturation and intensity) color system

[2,3]

which observes the following equations:

where

and

and

(2)

(4)

R + G + B

3

HSI model can significantly remove the white illuminant effect on skin color. However, in order to remove the non-white illuminant effects, we add an adaptive shifting color factor in the detection. Thus,

I =

where F ( x , y ) represents the possible face point candidate, and

RH,s

represents the skin color cluster with non-white illumination compensation updating with time.

After defining the face point candidates, we use

a

linking and searching algorithm to determine the boundaries of the face candidates in the window. Here we assume that the user sits nearest to the camera, and would be the largest face region in the window. Also by using the motion estimation, we can block out the other face candidate.

Within the face, we again use the color information to find the lips and check its ellipse shape by pattern comparison and record the last lips location to filter out impossible movement. Since the color of black and white are easily confused by the illumination, we use another technique to find the eyes.

Eyes are surely the most complicated region in the face region especially in the horizon line [4]. Thus, we first apply the edge detector to find all the edges within the face, and then use morphological operation to separate all the candidate clusters. Aftefwards, we use the shape matching to find the eye candidates, and check some knowledge-based information (ex. typical eye-to-lip distance, eye separation distance, containing both dark and bright points, and etc.). Also, the motion estimation is used here to decrease the system loading.

In

order to avoid user position shifts, we propose

a

rclative motion vector between the moving reference and the facc reference to control the computer mouse (Fig. 3). Here, the face refercncc is defined as the mass center of the face. We observe that with less than 20 degree head rotation, the face mass center would only suffer 5

?4

variation. Thus, we use the temporal low pass filtcr to avoid thc vibration of thc face reference. That will also make thc mouse cursor more stable. Then we use a weighted centcr of eyes and lips to scrve as a moving rcfcrcncc

M ( . x , > J )

as f b ~ ~ o w s :

. l / ( c , , ~ ' l = ~t',E~~,~,,,(.t,j'J+ I L ' , ~ '

,,,,

! ( . t . ! ' J + ~ c ~ ~ L ( . v . j v l (6)

where Erigh, ( x , y ) , E,efl ( x , y )

,

and L ( x , y ) represent the iris locations and lips mass center respectively. Moreover, we set a subtle strategy to let the eyes achieve small movement of the mouse and lips achieve the large movement. The mouse speed

S

can he defined as follows: where

Fr+,ence

represents the face reference and a, b are velocity factors changing with the user location. By such linear modification, we can avoid abrupt mouse velocity change when suffering user position shifts.

We set our mouse with single-click as the default action, and we also add double-click, scroll-up, scroll-down, and pauseirestart commands into OUT action menu to facilitate

the user. If the mouse pointer holds still for more than 1.5 sec, the action menu will pop up. If no action selected within 1.5 sec, the mouse will single click the target automatically. The above proposed waiting time is changeable to fit different user tolerances.

111. RESULTS

We use a laptop equipped with Intel Pentium4TM 1.0 CHz CPU, Logitech QuickCam Pro 3000, and Microsoft Windows XPTM to test our system for the disabled. Our system only take up roughly 4 0 % of the windows resource, and thus allow the user to operate more than two applications simultaneously. Thc results show that the disabled can smoothly use the cotnmon applications such as Microsoft Internet Explorer, Windows Media Player, Word, and Excel in the 1024 x 768 screen resolution. When the user wants to type, we adopt the "Screen Keyboard" to serve as the input device.

We also test our svstem bv the other obvious facc

(3)

:andidales

The result shows that our system maintains its robustness and precision despite the presence of other faces, but the system performance is decrease due to the computation complexities. Finally, in order to test the partial varying illumination, we put more than two light sources in different directions, and OUT system still

successfully work with roughly 10% wrong decisions (Fig. 5 . ) .

IV. DISCUSSION

When the user suffers from the partial varying illumination, OUT system is found some failure operations. Such wrong decisions are resulted form the non-linear shifts of skin color and abrupt artificial edges in the face region, we may adopt a high-order adaptive color model to solve this problem.

Every PC CCD camera has its own pre-process procedures, and may somehow deteriorate the performance of our system. If more precision required, we should include the camera characteristic (such as compression loss, white balance, auto gain control, and etc.) into OUT color model.

!

I

I

I

Fid. 5. Failure when partially varying i1lumi"ation

V. CONCLUSION

A reliable computer interface system for the disabled by using real-time face rec:ognition is presented. Although some failure operations are found because of the partial varying illumination, the robustness in the normal oftice light condition with cornplex background is significant. Furthermore, if we incorporate the camera pre-process characteristic into our color model, we can achieve an even higher precision and make it more suitable for the computer control.

REFERENCE

[ I ] Erik Hjelmas and Boa" Kee Low, "Face Detection: A Survey",

Compuler Vision and lmoge Understanding 83,2001. pp. 236-274. 121 C.H. Lee, J.S. Kim, and K.H. Park, "Automatic human face location in a complex background", Polls?rRecog. 29, 1996, pp. 1877.1889. [3] J. C. Temllon, M. Shimzi, H. Fukamachi, and S. Akamatsu, "Comparative performance of different skin chrominance models and

chrominance spaces for automatic detection of human face in color

images", Proc. I* IEEE Inf. Conf on Automatic Face and Gesrure Recognilion, 2000, pp. 54-61.

141 L.C. De Silva, K. Aizawa, and M. Hato", "Detection and mcking of facial feahlres by using a facial feature model and deformable circular template", IElCE Trons. Info-. Sysfems E78-D(9), 1995, pp.lI95-1207.

數據

Fig. 2.  System  block  dia-  11.  METHODOLOGY

參考文獻

相關文件

She has to face all these life choices by herself, makes her own decisions and face the consequences...I was told the other day that I could not even make an appointment for

Tekalp, “Frontal-View Face Detection and Facial Feature Extraction Using Color, Shape and Symmetry Based Cost Functions,” Pattern Recognition Letters, vol.. Fujibayashi,

Experiment a little with the Hello program. It will say that it has no clue what you mean by ouch. The exact wording of the error message is dependent on the compiler, but it might

Secondly, the key frame and several visual features (soil and grass color percentage, object number, motion vector, skin detection, player’s location) for each shot are extracted and

In the development of data acquisition interface, matlab, a scientific computing software, was applied to acquire ECG data with real-time signal processing.. The developed

A digital color image which contains guide-tile and non-guide-tile areas is used as the input of the proposed system.. In RGB model, color images are very sensitive

[16] Goto, M., Muraoka, Y., “A real-time beat tracking system for audio signals,” In Proceedings of the International Computer Music Conference, Computer Music.. and Muraoka, Y.,

Classifier which uses OpenCV (Open Source Computer Vision Library) was a detector that has been training for face detecting.. The result of the experiment is 2 frames per second, and