• 沒有找到結果。

A Street Scene Surveillance System for Moving Object Detection,Tracking and Classification

N/A
N/A
Protected

Academic year: 2021

Share "A Street Scene Surveillance System for Moving Object Detection,Tracking and Classification"

Copied!
7
0
0

加載中.... (立即查看全文)

全文

(1)

Tracking and Classification

Huei-Yung Lin* and Juang-Yu Wei

Department of Electrical Engineering National Chung Cheng University

Chia-Yi 621, Taiwan, R.O.C.

lin@ee.ccu.edu.tw, wei@image.ee.ccu.edu.tw

Received 9 September 2008; Revised 28 September 2008; Accepted 3 October 2008

Abstract. This paper presents a vision system for street scene surveillance. In addition to the capabilities of

detection and tracking of moving objects, it is also able to recognize and classify the targets based on the walking rhythm. The classification results are further used for event analysis and video retrieval of interested scenes. The proposed technique is computational efficient and can be used for embedded real-time applica-tions. The experimental results are presented for several image sequences of the real scenes.

Keywords: object tracking, street surveillance, classification

1 Introduction

Video surveillance for environment monitoring has gained significant research interests due to the availability of low-cost imaging hardware and high demand of commercial applications. It is traditionally used in the areas of industrial inspection, automated manufacturing, office or building security, and scientific experiments [1]. Re-cently, outdoor surveillance of public spaces for safety and security purposes has become one of the most impor-tant subjects in the computer vision research community. These issues are also the essential problems for the development of intelligent transportation systems (ITS). The ultimate goal is usually to build an intelligent sur-veillance system to reduce the usage of manpower and make the tasks more reliable. Some examples of using video information for intelligent transportation systems include highway traffic control, vehicle parking access control, driver assistance systems for intelligent vehicles, and pedestrian detection and tracking.

For highway traffic control, surveillance systems are used to estimate the vehicle flow and classify the types of observed vehicles based on the detected moving objects in the video sequences [2-3]. Parking access control acquires the images of the vehicle's license plate for pattern recognition and identification using the existing database [4-5]. Video surveillance for driver assistance is aimed to detect the changes of the environment, or monitor the driver's distraction level and used for intelligent safety deployment [6-7]. Similar work dealing with the vehicles or drivers have been extensively investigated for the past few decades. As for the street scene sur-veillance related intelligent transportation, most of the previous research focused on the recognition and tracking of walking pedestrians [8-10]. There are, however, relatively few work concerning the street surveillance in-volving both the pedestrians and vehicles for safety and security purposes.

In this paper we proposed a video surveillance system for the street environment. In addition to the real-time detection and tracking of moving objects in the scene, another important objective is to classify the types of interested targets and then used for event analysis and video retrieval. For example, some surveillance applica-tions might be interested in the pedestrian or vehicle counts for a given video stream, yet others might want to separate the image sequences with different classes of moving objects for efficient data storage or fast informa-tion access. The results can be used for traffic modeling or management, and searching of special events (such as car accidents) with least human visual investigation, respectively.

Under the normal street environment, there are generally three different types of moving objects, namely the vehicles (including cars and trucks), motorcycles (including bicycles), and pedestrians. The major work for our purpose is to first detect and track the moving objects, and then recognize the tracked targets based on the above three possibilities. Thus, the proposed street surveillance system mainly consists of the following three modules:

1. Detection and Tracking: Detect the foreground changes in the video stream and identify the moving tar-get for visual tracking.

(2)

2. Recognition and Classification: Recognize and classify the identified target based on the predefined ob-ject models.

3. Event Summarization: Index the video sequence for event summarization using the moving target clas-sification results.

Due to the mixture of non-rigid human body motion and rigid vehicle motion, conventional object tracking algorithm such as Kalman filter is not suitable for this system [11]. In the object detection and tracking module, a prediction and matching approach is adopted, provided that the uniform object motion is assumed. Recognition of the identified objects is achieved by checking two criteria: the height/width ratio and walking rhythm of the interested target (vehicle, motorcycle, or pedestrian). The surveillance video sequences are finally classified and indexed based on the target recognition results.

This paper is organized as follows. Section 2 describes the fundamentals of image processing techniques used in this work as well as the object detection and tracking module. In Section 3, we present the proposed recogni-tion module for moving target classificarecogni-tion. Experimental results performance analysis of the system are pre-sented in Section 4, followed by event summarization in Section 5. Section 6 concludes the paper and discusses possible directions of future work.

2 Foreground Detection and Object Tracking

The processing pipeline for the proposed object detection and tracking module is shown in Fig. 1. Given a se-quence of images captured by a video camera, the first step of visual surveillance is to segment the moving ob-jects or foreground regions from the static background scene. The simplest method is the direct image subtrac-tion from the video sequence, which attempts to locate the changes in two consecutive image frames. Although this technique is easy to implement, it cannot be used to detect an object without continuous motion in the scene (e.g., the object remains static for a short period of time and then moves away). More sophisticated tracking techniques, such as optical flow, use the image brightness constancy to detect the object motion [12]. In addition to the high computation cost of the algorithms, they are also sensitive to the illumination changes and thus not suitable for the outdoor environment.

Fig. 1. Processing pipeline of the detection and tracking module.

In this work, the background image subtraction method is used to segment the dynamic object in the scene. A background model of the scene is first generated based on the statistics of several input image frames. An inten-sity threshold is then used to segment the moving object from the difference between the background model and the current image frame. Since the background region of the scene will also change slightly due to the non-uniform outdoor illumination condition, it should be continuously updated after a period of time. To model the background scene caused by the illumination changes exclusively, the image is first converted to the HSV color space and only the brightness component is updated [13]. This process ensures that the pixel intensity values do not change significantly and the background scene is more robust for object segmentation. By thresholding the derived difference image, segmented and detected objects are represented by a binary image. After the morpho-logical erosion and dilation operations for noise reduction, the bounding boxes of the targets are given by the nonzero horizontal and vertical projections of the blobs. Fig. 2 shows the result of difference image and the detected object location.

To make the tracking algorithm more robust under different illumination conditions, a prediction and match-ing approach is applied on the detected target region to simultaneously track multiple objects in the image se-quence [1]. The position of a target at time t + 1 is predicted based on its positions at times t and t – 1, and this prediction is then used to match with the detected objects. Since the moving speed of the object is assumed to be uniform, the displacement of the detected targets between two consecutive frames is used to approximate the same object's position in the next image frame. The new target position is updated in the image sequence if an object match is found at time t + 1 by the following criteria:

(dx, dy) = (xt – xt-1, yt – yt-1)

(3)

Fig. 2. The difference image and detected object location.

where (dx, dy) is the displacement vector of the moving object at time t, and (xi, yi) is the object location at image

frame i.

The capability of multiple object tracking is essential for most surveillance systems. One major challenge of this requirement is the occlusion detection of multiple moving targets. However, if the object trajectory can be modeled by a uniform linear motion, then it is possible to use the area size of the tracked target to detect the occlusion -- the target region usually becomes fairly large instantaneously if the occlusion happens between two moving objects. Thus, the positions of two individual moving objects can be recorded prior to the occlusion, and then used to track the subsequent targets after the occlusion ends. To determine the termination of the occlusion for a street surveillance scene, the following three rules are adopted:

1. The blob size of the connected occluded targets is much larger than the size of the individual targets. 2. The positions of newly detected objects are close to the positions predicted prior to the occlusion. 3. The vertical positions of the detected objects remain constant for horizontal object motion.

In our implementation the objects move nearly horizontally along the image scanlines, which generally yields robust occlusion detection. Figs. 3 - 5 illustrate the multiple object tracking based on the above rules. Although the object tracking positions might not be accurate immediately after the termination of the occlusion, they will be modified automatically in the subsequent image frames.

Fig. 3. Multiple object tracking with occlusion (two vehicles).

(4)

Fig. 5. Multiple object tracking with occlusion (one vehicle and one motorcycle).

3 Recognition and Classification

To recognize the identified and tracked objects in the image sequence, we are primarily interested in three classes of objects: vehicle, motorcycle (or bicycle) and pedestrian. For general cases, the vehicles and motorcy-cles can be modeled as rigid bodies, but the pedestrians should be considered as deformable objects. Thus, the basic recognition and classification criteria are based on the height/width ratio and “walking” rhythm of the target. The flowchart of the proposed recognition and classification module is illustrated in Fig. 6.

Fig. 6. Recognition and classification flowchart.

Given an identified target object, the first step is to check the ratio of its height to width. For general cases the height to width ratios of the pedestrians are greater than 1, and the ratios of vehicles and motorcycles are less than 1. Moreover, the ratios of vehicles are usually smaller than those of motorcycles. Thus, these two types of objects can be further distinguished by a threshold of about 0.7 from experience. Checking the target's width/height ratio for recognition is a simple technique, but it might cause some potential problems. First, it is possible that tall motorcycle or bicycle riders make the object's height larger than the object's width. In this case, the target will be classified as pedestrian even if a motorcycle is actually detected. Second, if there is occlusion between two target objects, then the detection and tracking process will stop until the occlusion is terminated. Consequently, the targets can only be recognized and classified based on their appearances near the left and right borders of the image. In these cases the identified target usually has greater height/width ratio than the ground truth since it is partially truncated due to the camera’s limited field of view.

To overcome the above problems and ensure the correct identification of pedestrians, walking rhythm is fur-ther verified for the moving object with the height/width ratio greater 1. The idea is based on the fact that the target width is approximately a constant for vehicles and motorcycles but periodically changed for pedestrians, especially on their lower halves, due to the swing motion of the legs. Thus, pedestrians can be distinguished from vehicles or motorcycles by checking if there exist significant width changes. Since the time domain object width detection might not be accurate enough due to noise and shadow, and our goal is only to distinguish the periodic function from a constant, frequency domain approach will generally provide more robust pattern classi-fication. Thus, Fourier transform is applied on the function of object width (as a function of time) to find the corresponding power spectrum, and then used to distinguish the vehicles and motorcycles from the pedestrians.

(5)

Fig. 7 shows the object width functions of a vehicle and a pedestrian (Figs. 7(a) and 7(c)), and their power spectra (Figs. 7(b) and 7(d)). The response in frequency domain is used for target classification. Since the low frequency components of both cases are large, the second lobe of the power spectrum is used to distinguish between the vehicles and pedestrians. In the implementation, a threshold of 40 is selected to separate these two classes based on the training sample video sequence.

(a) Vehicle width change. (b) Power spectrum of Fig. 7(a).

(c) Pedestrian walking rhythm. (d) Power spectrum of Fig. 7 (c).

Fig. 7. Walking rhythms and the corresponding power spectra.

4 Event Summarization

The objective of event summarization for the street scene surveillance is to index the video sequence for effi-cient storage and fast information retrieval. It is commonly a major issue for a large data set of video recording. For example, searching for a specific event might be tedious work if one has to go through the whole video sequence manually. Thus, in this work we use a simple scheme to separate and index the video stream based on the target classification. The image frames corresponding to static scenes are first removed from the video se-quence. The starting and ending frame numbers of pedestrians, motorcycles and vehicles are indexed for the remaining video stream. As illustrated in Fig. 8, searching for the scenes consisting of any combination of pe-destrians (green), motorcycles (red), and vehicles (blue) based on the time stamps (or indices) of the target ap-pearance interval can be achieved very efficiently. Furthermore, new video sequences can be generated by the event summarization for other specific applications.

5 Experimental Results

The proposed street surveillance system has been tested on several real scenes. Fig. 9 shows the background images of four data sets used for performance evaluation. Each data set contains 27, 18, 20, and 20 minutes of video sequence at 15 fps, respectively. Fig. 10 illustrates the graphical user interface of the surveillance system. The moving object detection and recognition results are shown in Table 1. For the first three video sequences the illumination conditions do not change significantly over time and the recognition rate is about 96.8%. If the environment is under strong sunshine, as appeared in the last data set, the reflection of the target surface usually makes the preprocessing more difficult. Consequently, the recognition rate is slightly lower mainly due to the misclassification of the cases with a single vehicle and two adjacent pedestrians. The overall recognition rates for pedestrian; motorcycle and vehicle are 87.5%, 98.4% and 94.7%, respectively, for the video sequences shown in Table 1.

(6)

Fig. 8. Event summarization schematic.

Fig. 9. Experimental environment background scenes.

Fig. 10. Graphical user interface of the surveillance system.

Table 1. Moving object detection and recognition results

(GT: Ground Truth, TP: True Positive, FP: False Positive, FN: False Negative).

Video #1 Video #2 Video #3 Video #4

GT TP FP FN GT TP FP FN GT TP FP FN GT TP FP FN

Pedestrian 5 4 2 1 1 1 0 0 7 6 0 1 3 3 1 0

Motorcycle 75 72 1 3 65 65 0 0 99 98 2 1 73 72 2 1

Vehicle 39 38 0 1 25 22 0 3 23 22 0 1 27 26 0 1

(7)

6 Conclusion

The application of image analysis techniques to intelligent transportation systems has become an important re-search topic in the past few years. In this work, we have presented a vision system for street surveillance pur-poses. The moving objects in the street scenes are classified as pedestrians, motorcycles or vehicles based on the detection and tracking results. Furthermore, an event summarization scheme is used for quick identification of interested scenes. The proposed technique is computational efficient and can be implemented on embedded devices for real-time applications.

Acknowledgements

The support of this work in part by the National Science Council of Taiwan under Grant NSC-95-2221-E-194-075 is gratefully acknowledged. The authors would like to thank Mr. Li-Wei Kao for preparation of some test videos.

References

[1] R. Jain, R. Kasturi, and B. Schunck, Machine Vision. McGraw-Hill, 1995.

[2] F. C. D.J. Dailey and S. Pumrin, “An algorithm to estimate mean traffic speed using uncalibrated cameras,” IEEE

Trans-action on. Intelligent Transportation Systems, Vol. 1, No. 2, pp.98-107, June 2000.

[3] S. Gupte, O. Masoud, R. F. K. Martin, and N. P. Papanikolopoulos, “Detection and classification of vehicles,” IEEE

Transaction on Intelligent Transportation Systems, Vol. 3, No. 1, pp.37-47, March 2002.

[4] T. Naito, T. Tsukada, K. Yamada, K. Kozuka, and S. Yamamoto, “Robust license-plate recognition method for passing vehicles underoutside environment,” IEEE Transaction on Vehicular Technology, Vol. 49, No. 6, pp.2309-2319, Nov. 2000.

[5] S.-L. Chang, L.-S. Chen, Y.-C. Chung, and S.-W. Chen, “Automatic license plate recognition,” IEEE Transaction on

Intelligent Transportation Systems , Vol. 5, No. 1, pp.42-53, March 2004.

[6] M. Trivedi, S. Y. Cheng, E. Childers, and S. Krotosky, “Occupant posture analysis with stereo and thermal infrared video: algorithms and experimental evaluation,” IEEE Transaction on Vehicular Technology, Vol. 53, No. 6, pp.1698-1712, Nov. 2004.

[7] G. P. Stein, O. Mano, and S. A., “Vision-based acc with a single camera: Bounds on range and range rate accuracy,”

Proceedings of IEEE Intelligent Vehicles Symposium, pp.120-125, 2003.

[8] B. Leibe, E. Seemann, and B. Schiele, “Pedestrian detection in crowded scenes,” in CVPR ’05: Proceedings of the 2005

IEEE Computer Society Conference on Computer Vision and Pattern Recognition (CVPR’05) - Volume 1. Washington,

DC, USA, 2005, pp.878-885.

[9] A. Broggi, M. Bertozzi, F. A., and M. Sechi, “Shape-based pedestrian detection,” Proceedings of IEEE Intelligent

Vehi-cles Symposium, pp.215-218, 2000.

[10] C. Curio, J. Edelbrunner, T. Kalinke, C. Tzomakas, and W. von Seelen, “Walking pedestrian recognition,” IEEE

Trans-actions on Intelligent Transportation Systems, Vol. 1, No. 3, September 2000.

[11] F. Bremond and M. Thonnat, “Tracking multiple non-rigid objects in video sequences,” IEEE Transaction on Circuits

and Systems for Video Technology, Vol. 8, No. 5, 1998.

[12] B. Horn, Robot Vision. MIT Press, 1986.

數據

Fig. 1. Processing pipeline of the detection and tracking module.
Fig. 2. The difference image and detected object location.
Fig. 5. Multiple object tracking with occlusion (one vehicle and one motorcycle).
Fig. 7 shows the object width functions of a vehicle and a pedestrian (Figs. 7(a) and 7(c)), and their power  spectra (Figs
+2

參考文獻

相關文件

The hashCode method for a given class can be used to test for object equality and object inequality for that class. The hashCode method is used by the java.util.SortedSet

Based on [BL], by checking the strong pseudoconvexity and the transmission conditions in a neighborhood of a fixed point at the interface, we can derive a Car- leman estimate for

 Promote project learning, mathematical modeling, and problem-based learning to strengthen the ability to integrate and apply knowledge and skills, and make. calculated

This kind of algorithm has also been a powerful tool for solving many other optimization problems, including symmetric cone complementarity problems [15, 16, 20–22], symmetric

Microphone and 600 ohm line conduits shall be mechanically and electrically connected to receptacle boxes and electrically grounded to the audio system ground point.. Lines in

Because simultaneous localization, mapping and moving object tracking is a more general process based on the integration of SLAM and moving object tracking, it inherits the

– evolve the algorithm into an end-to-end system for ball detection and tracking of broadcast tennis video g. – analyze the tactics of players and winning-patterns, and hence

n Media Gateway Control Protocol Architecture and Requirements.