Professional Documents
Culture Documents
I. I NTRODUCTION
LIN et al.: EMBEDDED VISION SYSTEM ON AN UNMANNED ROTORCRAFT FOR GROUND TARGET FOLLOWING
Fig. 1.
1039
1040
Fig. 2.
xs + ys2 + zs2
rsp
tan1 xzss
(1)
psp = sp =
.
ys
1
sp
sin
rsp
LIN et al.: EMBEDDED VISION SYSTEM ON AN UNMANNED ROTORCRAFT FOR GROUND TARGET FOLLOWING
Fig. 3.
1041
satr = [s1 , s2 ]
valr = [v1 , v2 ].
(2)
(3)
m
m 2
m 2
02
) + 4 (11
)
2 = (20
(4)
m
m 2
m
m 2
312
) + (03
321
)
3 = (30
(5)
4 =
m
(30
m 2
12
)
m
(03
m 2
21
)
(6)
m
, for p + q = 2, 3, . . ., is the improved normalized
where pq
central moment defined as
m
pq
=
cpq
A(p+q+1)/2
(7)
where A is the interior area of the shape and cpq is the central
moment defined as
c
pq = (x x
)p (y y)q ds,
p, q = 0, 1, . . . .
(8)
C
Note that, in (8), C is the boundary curve of the shape, C is a
line integral along C, ds = (dx)2 + (dy)2 , and [
x, y] is the
coordinate of the centroid of the shape in the image plane.
In addition, compactness is another useful feature descriptor
for recognition. Compactness of a shape is measured by the
ratio of the square root of the area and the perimeter, which
is given by
A
.
(9)
Compactness : c =
C
It can be easily proven that compactness is invariant with
respect to translation, scaling, and rotation.
Color Feature Extraction: To make the target detection and
tracking more robust, we also employ color histogram to represent the color distribution of image area of the target, which
is not only independent of the target orientation, position and
size but also robust to partial occlusion of the target and easy to
implement. Due to the stability in outdoor environments, only
hue and val are employed to construct the color histogram for
object recognition, which is defined as
H = {hist(i, j)} ,
where
hist(i, j) =
(x,y)
i = 1, . . . , Nhue ;
j = 1, . . . , Nval
(10)
hue(x, y)
val(x, y)
i,
j,
Nhue
Nval
and where Nhue and Nval are the partition numbers of hue and
val channels, respectively, is the region of the target, [] is
the nearest integer operator, and (a, b) is the Kronecker delta
function.
1042
Fig. 4. Illustration of segmentation. (a) Input image. (b) Image in HSV color space. (c) Image after thresholding. (d) Image after morphological operations.
(e) Image after contour detection. (f) Regions of interest.
k = 1, . . . , d
wk
k=1
(11)
k,i k,j
k,j
2
+w6
dc (Hi , Gj ) 6,j
6,j
2
where
N
h N
v
dc (Hi , Gj ) =
p=1 q=1
(12)
D=
LIN et al.: EMBEDDED VISION SYSTEM ON AN UNMANNED ROTORCRAFT FOR GROUND TARGET FOLLOWING
1043
(13)
T2
s
1 Ts 0 0
0
2
Ts
0
0 1 0 0
=
=
2
s
0 0 1 Ts
0
2
0 0 0 1
0 Ts
1 0 0 0
H=
0 0 1 0
where Ts is the sampling period of the vision-based tracking
system. A Kalman filter can then be designed based on the
1044
5
k=1
+ w6
wk
i (k) j (k)
j (k)
2
2
2
dc (Hi , Gj )j (6)
zi j (7)
+w7
.
j (6)
j (7)
(14)
Most of the time, the model-based tracker can lock the target
in the image sequence, but sometimes, it may fail due to the
noise or disturbance, such as partial occlusion. Thus, a scheme
is required to check whether the target is still in the image and
then activate other trackers.
2) Switching Mechanism: The purpose of the switching
mechanism is to check whether the target is still in the image
when the target is lost by the model-based tracker. If yes,
the mean-shift tracker will be activated. The lost of the target
can be attributed to the poor match of features due to noise,
distortion, or occlusion in the image. An alternative reason
may be the maneuvering motion of the target, and the target is
out of the image. Therefore, in order to know the reason and
take the special way to find the target again, it is necessary
to formulate the decision making as the following hypothesis
testing problem:
H0 : The target is still in the image;
H1 : The target is not in the image due to maneuvers.
The estimation error is considered as a random variable, which
is defined by
Based on the color model of the target, the back projection algorithm is employed to convert the color
image to the color probability distribution image.
The probability of each pixel Ip (x, y) in the region
of interest r is calculated based on the model of the
target, which is used to map the histogram results
and given by
Ihue (x, y)
Ip (x, y) = histtg
(15)
Nhue
where Ihue is the pixel values of the image in the
hue channel.
Step 2) Mean-shift algorithm: Based on the obtained color
density image, a robust nonparametric method, the
mean-shift algorithm, is used to search the dominated peak in the feature space. The mean-shift
algorithm is an elegant way of identifying these
locations without estimating the underlying probability density function [12].
Recalling the discrete 2-D image probability distributions in (15), the mean location (the centroid) of
the search window is computed by
xc (k) =
= (H
xk1 zk1 )T 1 (H
xk1 zk1 )
where H
xk1 zk1 is assumed to be N (0, ) distributed.
is Chi-square distributed with two degrees of freedom (x and y
directions) under H0
if H0 is true
< = 22 (),
if H1 is true
= 22 (),
M00
M10
M00
yc (k) =
M01
M00
M10 =
(x,y)w
Ip (x, y)x
M01 =
(x,y)w
Ip (x, y)y
LIN et al.: EMBEDDED VISION SYSTEM ON AN UNMANNED ROTORCRAFT FOR GROUND TARGET FOLLOWING
1045
f
0 0
1 x
pi
(16)
0 fy 0 pc
=
1
0 0 1
with
pc = Rc/n pn + tc/n
(17)
(18)
ys
1
sin
p
rsp
pe =
(19)
= N (pi , v) =
p
tan1 xs
zs
where
1
x
s
fx
0
ys = Rs/c (v) 0
fy1
zs
0
0
rsp = x
2s + ys2 + zs2 .
xi
0
0 yi
1
1
(20)
1046
The pan/tilt servomechanism can be approximately considered as two decoupled servomotors, which regulate the visual
sensor for horizontal and vertical rotations, respectively. The
dynamic model of the servomotor can be described by using
a standard second-order system. Before proceeding to design
the control law for the pan/tilt servomechanism, we define the
tracking error function as
e(k) = pe
pe
(pi , v(k))
(21)
(22)
(23)
and p
i in (21), we then can obtain
Replacing v and pi with v
the modified error function as
(k)) N (pi , v
(k)) .
e(k) = N (
pi (k + 1), v
(24)
Kp T s
e(k)
Ti
where the proportional gain and the integral time are chosen
as Kp = 0.65 and Ti = 0.8, respectively. We note that two
identical controllers are respectively used for the pan and tilt
servos, since the dynamics of these two servos are very close.
2) Following Control of the UAV: As shown in Fig. 6, to
estimate the relative distance between the target and the UAV,
we combine the camera model (16) with the transformation in
(17) and generate the overall geometric model from an ideal
image to the NED frame
1
fx
xi
0
0
(25)
fy1 0 yi + tn/c .
pn = Rn/c 0
1
0
0
1
We assume that the ground is flat and the height of the UAV to
the ground h is known. We have
r1 r2 r3
xn/c
tn/c = yn/c
(26)
Rn/c = r4 r5 r6
r7 r8 r9
zn/c
which can be calculated by using the measurements of the
onboard navigation sensors. Based on the assumption that the
target is on the ground, zn is equal to zero. We then can derive
as
=
zn/c
r7 fxxi + r8 fyyi + r9
(27)
xuav
xn
1 0 0
Rn/b cy
0 1 0
yn
yuav
0
=
zuav
h0
uav ref
0
where cx and cy are the desired relative distances between the
target and the UAV in the Xb - and Yb -axes, respectively, h0 is
the predefined height of the UAV above the ground, 0 is the
predefined heading angle of the UAV, and Rn/b is the rotation
matrix from the body frame to the local NED frame, which can
be calculated in terms of the output of the onboard navigation
sensors.
VI. E XPERIMENTAL R ESULTS
To verify the proposed vision system, multiple tests of the
complete system were conducted. During these tests, the proposed vision-based unmanned helicopter SheLion was hovering
autonomously at a certain position. If the moving target entered
into the view of the onboard camera, the target would be
identified and tracked in the video sequence by the vision
system automatically. Based on the vision information, the
pan/tilt servomechanism was controlled to keep the target in
a certain position in the image, as described in Section V-C1.
Then, the operator can command the UAV to enter into the
following mode, in which the UAV followed the motion of the
target autonomously based on the estimated relative distance,
using the algorithm proposed in Section V-C2.
The experimental results of the vision-based target detection
and tracking in flight are shown in Table I, which indicate
that the proposed vision algorithm could effectively identify
and track the target in the video sequence in the presence of
LIN et al.: EMBEDDED VISION SYSTEM ON AN UNMANNED ROTORCRAFT FOR GROUND TARGET FOLLOWING
1047
TABLE I
E XPERIMENTAL R ESULTS OF TARGET D ETECTION AND
T RACKING IN F LIGHT
Fig. 7.
VII. C ONCLUSION
In this paper, we have presented the comprehensive design
and implementation of the vision system for the UAV, including
hardware construction, software development, and an advanced
ground target seeking and following scheme. Multiple real
flight tests were conducted to verify the presented vision system. The experimental results show that this vision system is
not only able to automatically detect and track the predefined
ground target in the video sequence but also able to guide the
UAV to follow the motion of the target in flight. The robustness
and efficiency of the developed vision system for UAVs could
be achieved by the current system. Our future research focus is
to utilize the system for implementing vision-based automatic
landing of the UAV on a moving platform in an environment
without GPS signals.
R EFERENCES
[1] M. J. Adriaans, R. C. Cribbs, B. T. Ingram, and B. P. Gupta, A160
Hummingbird unmanned air vehicle development program, in Proc.
AHS Int. Specialists MeetingUnmanned Rotorcraft: Des., Control Test.,
Chandler, AZ, 2007.
[2] O. Amidi, T. Kanade, and R. Miller, Vision-based autonomous helicopter
research at Carnegie Mellon Robotics Institute 19911997, in Proc.
Amer. Helicopter Soc. Int. Conf., Gifu, Japan, 1998, pp. 112.
[3] J. L. Barron, D. J. Fleet, and S. S. Beauchemin, Performance of optical
flow techniques, Int. J. Comput. Vis., vol. 12, no. 1, pp. 4377, Feb. 1994.
1048
[30] F. Porikli, Achieving real-time object detection and tracking under extreme conditions, J. Real-Time Image Process., vol. 1, no. 1, pp. 3340,
Oct. 2006.
[31] F. Sadjadi, Theory of invariant algebra and its use in automatic target recognition, in Physics of Automatic Target Recognition, vol. 3.
New York: Springer-Verlag, 2007, pp. 2340.
[32] Z. Sarris and S. Atlas, Survey of UAV applications in civil markets, in
Proc. 9th Mediterranean Conf. Control Autom., Dubrovnik, Croatia, 2001,
vol. WA2-A, pp. 111.
[33] A. R. Smith, Color gamut transform pairs, in Proc. 5th Annu. Conf.
Comput. Graph. Interactive Tech., New York, 1978, pp. 1219.
[34] O. Spinka, O. Holub, and Z. Hanzalek, Low-cost reconfigurable control
system for small UAVs, IEEE Trans. Ind. Electron., vol. 58, no. 3,
pp. 880889, Mar. 2011.
[35] T. Y. Tian, C. Tomasi, and D. J. Heeger, Comparison of approaches
to egomotion computation, in Proc. IEEE Conf. Comput. Vis. Pattern
Recognit., San Francisco, CA, 1996, pp. 315320.
[36] S. Tsugawa, Vision-based vehicles in Japan: Machine vision systems
and driving control systems, IEEE Trans. Ind. Electron., vol. 41, no. 4,
pp. 398405, Aug. 1994.
[37] H. Veeraraghavan, P. Schrater, and N. Papanikolopoulos, Robust target
detection and tracking through integration of motion, color and geometry, Comput. Vis. Image Understand., vol. 103, no. 2, pp. 121138, Aug.
2006.
[38] C. R. Wren, A. Azarbayejani, T. Darrell, and A. P. Pentland, Pfinder:
Real-time tracking of the human body, IEEE Trans. Pattern Anal. Mach.
Intell., vol. 19, no. 7, pp. 780785, Jul. 1997.
[39] D. Xu, L. Han, M. Tan, and Y. F. Li, Ceiling-based visual positioning
for an indoor mobile robot with monocular vision, IEEE Trans. Ind.
Electron., vol. 56, no. 5, pp. 16171628, May 2009.
[40] Q. M. Zhou and J. K. Aggarwalb, Object tracking in an outdoor environment using fusion of features and cameras, Image Vis.Comput., vol. 24,
no. 11, pp. 12441255, Nov. 2006.
LIN et al.: EMBEDDED VISION SYSTEM ON AN UNMANNED ROTORCRAFT FOR GROUND TARGET FOLLOWING
1049