• 沒有找到結果。

A dynamic vision system for tracking and localization of underwater objects

N/A
N/A
Protected

Academic year: 2021

Share "A dynamic vision system for tracking and localization of underwater objects"

Copied!
8
0
0

加載中.... (立即查看全文)

全文

(1)

A Dynamic Vision System for Tracking and

Localization of Underwater Objects

Sheng-Wen Cheng

*,

Huai-Wen Hsu

**

and Shiao-Chug Cheng**

*Professor, **Graduate Student

Department of Engineering Science and Ocean Engineering

National Taiwan University, Taipei, Taiwan 106

(phone: +886-2-33665760; fax: +886-2-23929885; e-mail: niccheng@ntu.edu.tw).

ABSTRACT-We describe the development of an underwater cross-range high-accuracy 3D metrology by utilizing a pair of video cameras for underwater vehicle applications. The novelty of the approach lies in (1) a polynomial representation of image non-linear distortion induced by lens system and surrounding water; (2) the transfer of photogrammetric bundle adjustment techniques to underwater ranging; and (3) a new representation of underwater stereo vision including camera self-calibration, match points searching, and computation of the target coordinates.

We demonstrate that our method achieves better flexibility and seeurscy compared to traditional stereographic methods. hder Term-underwater vehicle, stereo vision, dynamic vision, photogrammetry, bundle adjustment, camera calibration..

I. INTRODUCTION

Stereo vision is a common and popular method for underwater optical ranging.[l-5] In general practice, one assumes that the two optical axes of cameras are parallel to each other. Base on this assumption, we have developed a dynamic stereo vision system with compensation of image distortion and optical flow analysis.[6, 71 However, we found tuning of the camera poison and orientation, to ensure the parallelism, was much more difficult than performing stereo vision itself.

To solve this problem, the most relevant technique is the bundle adjustment [8-91, which is a photogrammetric technique to combine multiple images of the same scene into an accurate 3D reconstruction. Then bundle adjustment applies an iterative algorithm to compute optimal values for the 3D reconstruction of the scene, camera positions and orientations, by using a least-squares algorithm

This paper presents an underwater dynamic vision system for underwater vehicles to perform tasks of observation, searching, ranging and tracking of underwater objects.

The vision system is composed of three modules for camera calibration, stereo imaging and video image processing.

In the camera calibration module, a polynomial model is proposed to compensate for the image distortion induced by lens and surrounding water. For the stereo imaging module, an epipolar line based algorithm is adopted for matching points searching. Object ranging is carried out using a bundle system model, originated in photogrammehy community, in which parameters of camera calibration and stereo imaging are coupled. The video image processing module is responsible for object movement estimation through optical flow analysis. The optical flow can be dissolved into translation, rotation and distance change components of objects.

Two experiments are conducted in this paper: 1. Localization and tracking of a moving object under diffuse illumination, and 2. 3D mapping of underwater surfaces by using a scanning paint beam. As a result, the dynamic vision system demonstrates feasibility for A W applications.

11. CAMERA CALIBRATION A. Compensation oflmage Distortion

Figure 1 shows an example image of a calibration board on which dots are equally spaced. We can identify image distortion through displaced dots. Source of mage distortion consists oftwo major parts: lens and surround media. Effect of surround media is depicted as figure 2 . This figure demonstrates the fact that 0-7803-8541 -1/04/520.00 82004 IEEE.

(2)

underwater is far-sighted

am

810 a01 bm bin bo, Figure I Example of image distortion: original image

CR

4

4 4

0.0007 -0.4602 5.8875 0.0012 0.9016 13.0059 -0.0008 0.1185 -6.5152 -0.0018 0.0272 -11.7107 0.0015 -0.1910 5.6253 0.0010 1.0548 9.0964 ID1 m . m - . 0 XIJ ,.I "' ,U 0 1% * -i )m xo m so

Figure 2 Effect of surrounding medis on image distortion

am

ao,

Tu compcnrare for the image distortion. w e adopt rhe dual plane m.,dr.l proposed h) (ircmbsn [ I I ]

4

c,"

4

0.0006 -0.3626 4.4153

0.0005 1.0132 5.8813

-0.0010 0.1229 -3.0734

where (x, y , z) is coordinates of a arbitrary point target, (xi, yj) is coordinates ofthe corresponding image point, Mand N are the order of polynomial for xi and yi. respectively. In this paper, we select M=N=3.

Owing to the perspective effect, calibration coefficients ..a and b , must be functions ofrlf Here, we use polynomial forms.

bin bo1

where K=2 in this paper.

0.0010 -0.1242 4.7871

0.0010 1.0635 7.1424

By taking images of calibration board from more than three different distances, as shown in figure 3, non-linear algebraic equations will be derived fmm equation ( I ) and equation (2). Then solving the equations by using least-square algorithm, we obtain calibration coefficients for right and leff cameras as shown in table I and table 2, in case ofundenvater and dry conditions.

an0 a01 bca bio bor CR

4

4

4

0.0023 -0.7706 8.8642

-

0.0 0 8 5 2.5059 -23.2657 0.0002 -0.0208 -0.6509 0.0014 -0.4257 0.0880 -0.0004 0.0370 -1.7612 -0.0068 2.3952 -16.6810

1

bm

I

-0.0018) 0.02361 -11.06361 -216

(3)

0.0024 -0.6478 9.2164 -0.0075 -0.0010 0.1308 -2.9388 bw blo bo, 0.0007 -0.3473 -1.7444

.

-0.0017 0.2112 -5.7911 -0.0020 1.8347 -2.5880

Althdugh calibration coefficients for underwater and dry conditions are different, the image of control points on calibration board aRer compensation will be almost the same as shown in figure 4.

:F1

. *

. .

*

. . . - .

-.. " ~ . . . .

Figure 4 Len and right images of control points afler compensation

B.

Camera Position and Orientation (figure 5)

Considering a target point P whose coordinates are (X, Y, 2 , ) and (X, Y , Z,) in the world coordinates system and camera coordinates system, respectively. If the coordinates of camera center in the world coordinates system is &Yo, ZJ, then

following relation.

Figure 5 Calculation of Camera position and Orientation Equation (3) and equation ( 5 ) are fundamentals ofbundle

where (Xw, U,, &) denotes the position of camera center in the world coordinates system.

By using iterative least square algorithm, the camera position (Xw, Y,, &) and orientation (a,

p,

y) can be solved from

equation (7).

For verification, a rotation with respect to y-axis

(p)

test in five degree increment was conducted. Results are shown as table 3. The accuracy of the proposed algorithm is acceptable.

On the other hand, from equation (I) and equation (2). we obtain the

(4)

I

z d

calibration

b-)

Figure

6

Configuration

for

camera

111. STEREO VISION

USING

CAMERAS OF

ARBITRARY ORIENTATION

Following equation (3) and equation (S), the coordinates of target point Pin figure7 will satisfy following equations.

Xw-XRo

[:!

I=

RR

[:

]

[

1::

‘RI2 ‘Rt3] RR = ‘E21 ‘R22 ‘Rl3 (I2) ‘R32 ‘R33 ‘L32 ‘L33

By eliminating (X, Y,

L)

from above equations, non-linear equations in terms of unknowns ( zC,, Z , ) will be obtained.

Figure 7 General Arrangement of Binocular Visual System

IV. DYNAMIC VISION SYSTEM AND EXPERIMENTS

A. Experimental System Configuration (figure 8) 1.

2. Water tank and xyz Table : 2.4 m X I .2 m x 0.8m. 3. Water-tight CCD cameras : 640 x 480 0 Pixels.

4. Binocular frame : shown as figure 9. 5 . Calibration board.

6. Undenvater Lamps.

7. Video frame grabber : Coreco CobraiC6 card. 8. PC-based controller.

9. GPlB interface card,

Back Snrm for Calbrr

mcmra

fi&tcam=a

Figure 8 System Configureuration of Sensing and Perception Module Development Platform

(5)

Figure 9 Binocular frame

B. Tracking

.

System setup is shown as figure 10. A fish-shaped object was moved along a predetermined 3D path. According to the position and speed of object, the binocular vision system adaptively followed. The degree of freedom was three: bacwforth, upldown and horizontal rotation.

Image sequence is shown as figure 11. During tracking, the system identified the object based on pattern matching technique.

We summarize the results of object tracking and motion trajectory estimation in figure 12. The estimated path matches well with the actual one.

F i m e

IO

Sketch of Underwater Active Stereo Vision Svstem

(a) image pair at position 0

@) image pair at position 1

(c) image pair at positim 2

(6)

@image pair at position 7 Figure 11 Image Sequence

BW -.3w

2 "I XI")

Figure 12 Object Tracking and Motion Trajectory Estimation

C.

Mapping Using Active Laser Beam

In

stereo vision, match points searching task requires high computing power To ease hardware and software burden, we utilized active illumination method [IO] and tbresholding of image grayscale level. As a result, the range data acquisition nearly synchronized with frame acquisition.

Example of surface mapping is shown in figure 13. The white spots of laser beam can be easily located.

By mechanical scanning of laser point beam, we get the surface contour in 3D format shown as figure 14 and figure 15. Figure 16 and figure 17 are

of

2Dcontour format.

Comparing figure14 with figure 15, and figure 16 with 17, it is clear that compensation of image distortion is essential to get accurate results.

(7)

(a)

(b)

Figure 13 (a)left image; (b) right image.

... ... ... ... . .. . bo ... :... ... . . .. . . . . . . . . . . . . . . . .. . . . . . . . . . ... .. .. . ... ~ : , . . .

Figure 14 3D contour

of

underwater object

(without compensation of distortion)

.:I

,

;

, , ,

r

,I

4

bo -50 4 3(1 -2c -IO 0 10 m Jo 40

X("1

Figure

16 2D

contour of underwater object

(without compensation of distortion)

I

.

.

, , , I , ,

.

, ,

-70 bo €0 4 0 30 -20 -10 0 10 20 30

X("l

Figure

17

2D contour of underwater object

(distortion compensated)

Figure

15

3D contour

of

underwater object

(distortion compensated)

V. CONCLUSION

We demonstrated that the proposed method achieves better flexibility and accuracy compared to traditional underwater stereo vision methods.

ACKNOWLEDGMENT

The authors would like to thank the support from National Science Council of Republic of China (Taiwan).

(8)

[I21 C. Stiller and J. Konrad ,”Estimating Motion in Image Sequences,” IEEE Signal Processing Magazine, pp. 1053-5888, 1998.

REFERENCES

[I] Gmsso E. and Tistarelli M., “Active/Dynamic Stereo Vision,” Pattem Analysis and Machine Intelligence, IEEE Transactions on, Vo1.17, No. 9, pp. 868 -879, 1995. Grosso E. and Tislarelli M., “ActiveiDynamic Stereo: a General Framework,” Computer Viiion and Pottern Recognition Proceedings CVPR ‘9% 1993 IEEE Computer Sociery Conference on, pp. 732 -734, 1993.

Inaba, M. and Hara, T., “A Stereo Viewer Based on a Single Camera with View-Conlml Mechanisms,” Proceedings of the 1993 IEEERSJ lntemational Conference on Intelligent Robots and Systems, IROS’93, Vol. 3, pp. 1857-1865, 1993. Kougeaux S. and Kita N., “Binocular Tracking Based on Virtual Horopten,” Proceedings of the IEEEIRSJIGI International Conference on Intelligent Robots and Systems, IROS ‘94, Vol. 3, pp. 2052 -2057, 1994.

A. Goshtasby and W. A. Gruver, “Design of a Single-Lens Stereo Camera System,” Panern Recognition, Vol. 26, No. 6, pp. 923-937, 1993.

[2]

[3]

(41

[SI

[6] Sheng-Wen Cheng, Huai-Wen Hsu and Ling-Hao Yu, “Research on A Real-time Robotic Visual System for Ranging and Motion Estimation of Underwater Objects,” Proceedings of the 4Ih Conference on Underwater Science and Technology (CUST’OZ), pp.CI-19-27, Taipei, Taiwan, April 2002. Sheng-Wen Cheng, Huai-Wen Hsu and Chun-Fu Chang, “Development of a Dynamic Vision System for Tracking and Localization of Underwater Moving Objects,” Proceedings of the 5Ih Conference on Underwater Science and Technology (CUST’O3), pp.68-75, Kao-Hsiung, Taiwan, April 2003. S.I. Granshaw, “Bundle adjushnent methods in engineering photogrammetry,” Photogrametric Record, Vol. IO, No, 56, pp. 181-207, 1980.

KB.

Atkinson, Close Range Photogrammetry and Machine

Vision, Whittles Publishing, Caithness, Scotland, 1996.

[IO] R. Harlley and P. Sturm, ‘Triangulation,” Computer Vision and Image Understanding, Va1.68, No.2. pp. 146-157, 1997. [ I l l Gremban, K. D. and C. E. Thowe, “Geometric Camera

Calibration Using System of Linear Equation”, Proceedings o/ IEEE Internotionol Conference on Robotics and Automotion, pp. 562-567, 1988 [7] [8] [9] ’ -222.

數據

Figure 2 Effect of surrounding medis  on  image distortion
Figure  5  Calculation of Camera position and Orientation  Equation (3) and equation  ( 5 )   are fundamentals ofbundle
Figure  7  General Arrangement  of  Binocular Visual System
Figure 9 Binocular frame
+3

參考文獻

相關文件

Students are asked to collect information (including materials from books, pamphlet from Environmental Protection Department...etc.) of the possible effects of pollution on our

Wang, Solving pseudomonotone variational inequalities and pseudocon- vex optimization problems using the projection neural network, IEEE Transactions on Neural Networks 17

We explicitly saw the dimensional reason for the occurrence of the magnetic catalysis on the basis of the scaling argument. However, the precise form of gap depends

Define instead the imaginary.. potential, magnetic field, lattice…) Dirac-BdG Hamiltonian:. with small, and matrix

• CEPC design has to maintain the possibility for SPPC, but there is no need now to firmly prove the feasibility of SPPC,.. scientifically

incapable to extract any quantities from QCD, nor to tackle the most interesting physics, namely, the spontaneously chiral symmetry breaking and the color confinement.. 

Miroslav Fiedler, Praha, Algebraic connectivity of graphs, Czechoslovak Mathematical Journal 23 (98) 1973,

Microphone and 600 ohm line conduits shall be mechanically and electrically connected to receptacle boxes and electrically grounded to the audio system ground point.. Lines in