Person Follower
Person Follower
I. I NTRODUCTION
Following a specific person is an important task for service
robots. Visual person following in public spaces entails
tracking of multiple persons by a moving camera.
There have been a lot of works on person detection
and tracking using various image features and classification
methods [1], [2], [3], [4], [5]. Many of them, however, use
a fixed camera. In the case of using a moving camera,
foreground/background separation is an important problem.
This paper deals with detection and tracking of multiple
persons for a mobile robot. Laser range finders are widely
used for person detection and tracking by mobile robots
[6], [7], [8]. Image information such as color and texture
is, however, sometimes necessary for person segmentation
and/or identification. Omnidirectional cameras are also used
[9], [10], but their limited resolutions are sometimes inappropriate for analyzing complex scenes.
Stereo is also popular in moving object detection and
tracking. Beymer and Konolige [11] developed a method of
tracking people by continuously detecting people using distance information obtained from a stationary stereo camera.
Howard et al. [12] proposed a person detection method
which first converts a depth map into a polar-perspective
map on the ground and then extracts regions with largelyaccumulated pixels. Calisi et al. [13] developed a robot system that can follow a moving person. It makes an appearance
model for each person using stereo in advance. In tracking,
the robot extracts candidate regions using the model and
PC
Camera
robot motion
Fig. 1.
Robot
Left
Front
Fig. 2.
Right
Depth templates.
W/2
H/2
Fig. 3.
[T (p, q) ID (x + p, y + q)] .
(1)
p=W/2 q=H/2
Fig. 4.
X
Z Y
TABLE I
results
exist
not exist
judged to exist
0
exist
not exist
judged to exist
414
0
exist
not exist
judged to exist
391
0
Fig. 5.
target
turning
center
t+1
Fig. 6.
t+1
vR
vL
left
wheel
2d
right
wheel
0
0
F t xt + Gt wt
0 t
0
0
1
0
0
0
0
1
0
0
t
0
0 , Gt = 0
0
1
t
0
(3)
0
0
0
0
t
1 0
2
Qt = Cov(w t ) = E w t wTt = w
0 1 .
We then consider the case where the robot moves. Figure 6
shows how a wheeled mobile robot moves. The distance of
two wheels is denoted as 2d. When each wheel rotates with
f t (xt , ut ) + Gt wt ,
(4)
where
f t (xt , ut ) =
Zt
Xt cos + Yt sin v
X t sin + Y t cos
= H t xt + v t ,
1
0
0
0
1
0
(5)
0 0 0
0 0 0 ,
yt =
, Ht =
1 0 0
1 0 0
Rt = Cov(v t ) = E v t v Tt = v2 0 1 0 .
0 0 1
input images
Fig. 7.
depth images
(X 2 + Y 2 )/2Y.
person C
(X, Y)
target
robot
person A
direction of camera
7m
ceiling
camera
person B
left
wheel
Fig. 8.
right
wheel
Fig. 10.
Total
Time
40 [ms]
20 [ms]
10 [ms]
20 [ms]
90 [ms]
person B
4m
robot
4m
Fig. 11.
#038
#093
#131
#069
#113
#139
#079
#123
#156
robot view
robot view
robot view
Experimental result with one person to follow and the other two.