CASTER: A Robot for Urban Search and Rescue
Mohammed Waleed Kadous, Raymond Ka-Man Sheh, Claude Sammut
ARC Centre of Excellence for Autonomous Systems, School of Computer Science & Engineering
University of New South Wales
Sydney, NSW 2052, Australia
E-mail: {waleed,rsheh,claude}@cse.unsw.edu.au
Abstract
Designing robots for urban search and rescue
(USAR) is an extremely challenging task, because it violates many of the assumptions made
in the average robotics lab. Robocup Rescue
[Jacoff et al., 2004] is an international competition that provides an opportunity to measure
performance in such environments objectively.
The competition allows for the evaluation of
new and evolving sensing, hardware and software technologies in an environment that is as
close to realistic as possible.
This paper discusses the design and implementation of CASTER, the robot that helped Team
CASualty come 3rd in the 2005 competition.
The strengths of CASTER compared to other
teams include high quality victim information,
excellent maps and an intuitive user interface.
The mapping techniques in particular are novel
because they take advantage of a 3D time-offlight range imager, a technology that has only
recently become widely available.
The results are all the more impressive when
one considers that this was the team’s first entry into the Robocup Rescue Robot League and
that the robot base was delivered seven months
before the competition.
1
1.1
Introduction
Robocup Rescue
One commonly cited example where robots can be of
benefit to humanity is urban search and rescue (USAR).
Robots are deployed at an earthquake site or other disaster and autonomously search the area, co-ordinate with
each other, deliver assistance to those in need and assist
in rescuing survivors. Whilst such dreams are obviously
a long way off, much research is already being undertaken into technologies that may fulfill this ambitious
goal. The RoboCup Rescue Robot League (RRL) aims
Figure 1: Examples of RoboCup RRL Standard Arenas. Left is the USAR test facility at the University of
Technology, Sydney and right is the competition arena
at Osaka
to provide both the motivation, in the form of a competition; and the testing environment, in the form of
standard arenas, for research related to USAR. Doing
so means that very early in the development cycle, new
technologies can be practically evaluated
RRL Standard Arenas are intended to be different, but
comparable, real-world environments where robots may
be tested, much like how various golf courses are different but comparable environments [Jacoff et al., 2004].
Examples of RRL Standard Arenas are shown in Figure 1. Whilst each is different, they are designed based
on the same rules thus results of testing in the different arenas may be compared. They also intend to be
practical approximations to “real world” disaster sites
thus debris, loose material, low clearances, victims with
varying states of consciousness, multiple levels, variable
lighting and radio blackspots are replicated. For practical reasons, other factors such as water and mud are not
included and unstable collapses, dust and fire are simulated by lightweight debris, curtains and “fake” heat
sources.
Every year, teams from around the world bring robots
and support equipment to the RoboCup RRL Competition which, in 2005, was held in Osaka, Japan. Teams deploy their robots in a RRL Standard Arena and are able
to compare mobility, sensing, autonomy, mapping and
human-robot interface (HRI) developments with other
teams. During competitive runs, operators are isolated
from the arena and have no prior knowledge of the layout. The goal of these time-limited competitive runs is
to find as many victims as possible, diagnose their state
of consciousness and generate a map of the disaster area
that a human rescuer can use to locate the victims. Secondary goals include locating landmarks and structurally
interesting objects that may affect rescuer safety.
1.2
Prior Work
Successful entries in the RRL have traditionally been focused on one or two of four research areas – mobility, victim identification, mapping and user interfaces. Considerable work has also been done in a fifth area – autonomy
– but autonomous solutions have thus far had limited, albeit increasing, success. 2003 world champions, Robrno
[Zalud, 2003] had a simple, robust four-wheeled platform
that relied on high ground clearance and strong motors
for its mobility. Their standout characteristic was their
intuitive robot interface that included virtual reality goggles. A head tracker monitored the operator’s head, allowing the onboard camera to mirror the operator’s head
movements. Their exceptional situational awareness allowed them to incur very few penalties [NIST, 2003] and
resulted in their high score.
In 2004, mobility and sensing came to the fore [NIST,
2004]. Toin Pelican [Koyanagi et al., 2004], winners of
the international competition, fielded a highly mobile reconfigurable tracked vehicle that was able to reach further into the rescue arena than any other robot. This
was combined with a raised camera that provided exceptional situational awareness, allowing them to score
well.
In contrast, second place winners in the international
competition, Kurt3D [Surmann et al., 2004b] fielded a
comparitively simple six-wheeled robot. Their distinguishing feature was their ability to generate 3D maps
using a 2D laser scanner mounted on a tilting mechanism, combined with software to automatically match
the resulting datasets [Surmann et al., 2004a]. Despite
having significantly poorer mobility than many other
teams, their exceptional mapping abilities enabled them
to score highly.
1.3
Objectives of the project
In order to provide a platform for a wide variety of research projects, as well as to do well in the RRL competition, it is necessary to address the main areas of USAR
robotics – mobility, victim identification, mapping and
user interfaces. Whilst autonomy is a growing area of
USAR robotics, this project will focus on teleoperation.
The ambitious goals for the project were to produce a
robot system that had the ability to:
Figure 2: Top-down view of CASTER (roll cages removed)
Mobility Safely traverse stairs, ramps and rubble
whilst carrying a full sensor payload
Victim identification Detect at least four signs of life,
including heat images
Mapping Generate dense, textured 3D maps, merge
them in an efficient manner and present them in
an intuitive way
HRI Provide the operator with as much information
about the robots’ immediate surroundings as they
would have with eyes-on control and in an intuitive
way
2
Design and Implementation
Figures 2 and 3 show annotated views of the robot,
dubbed CASTER (Centre for Autonomous Systems
Tracked Exploratory Robot), including sensors.
2.1
Locomotion
CASTER is built on a Yujin Robotics Robhaz DT3 base
[Yujin, 2005]. This robot base is weatherproof and extremely robust, indeed it is designed to survive the detonation of small explosives.
The robot has two pairs of rubber tracks for locomotion. The front pair of tracks follow a triangular path, allowing the DT3 to climb over obstacles and stairs whilst
a conventional flat rear pair provide additional traction.
Two motors provide enough torque to allow the robot to
carry payloads of up to 40 kilograms up inclines in excess
of 45◦ . An internal battery pack provides a runtime of
approximately 90 minutes. No mechanical modifications
were made to the DT3 other than the addition of two
Figure 3: Oblique view of CASTER
1cm thick polycarbonate plastic roll cages to protect the
additional equipment.
2.2
Sensors
CASTER’s sensor fitout was performed with three goals
in mind: high quality 3D mapping, reliable victim identification and good situational awareness.
To this end, the following sensors mounts were installed:
• TracLabs Biclops pan-tilt unit with armoured camera mount
• Logitech QuickCam Pro 4000 colour digital main
camera plus high brightness white LEDs (on armoured camera mount)
• CSEM SwissRanger 2 time-of-flight range imager
(on armoured camera mount)
• FLIR ThermoVision A10 thermal camera (on armoured camera mount)
• Sony ECM-MS907 stereo microphone (on armoured
camera mount)
• IP9100A quad ethernet video capture device
• 3 auxiliary cameras (connected to IP9100A video
capture device)
• ADXL311 accelerometer
Figures 2 and 3 illustrate the placement of these sensors on the robot. The following subsections describe
the sensing fitout in greater detail from the perspectives
of situational awareness, victim identification and mapping.
Situational awareness
The DT3 robot base can move in highly unpredictable
ways on unstructured terrain due to its length, articulation and the fact that it skid-steers. Given the difficulty
of the environment itself and the very real prospect of
becoming stuck or unstable, situational awareness is critically important.
Four colour cameras provide visual feedback to the operator. A high resolution webcam on the pan-tilt unit
forms the main driving camera and is able to observe
the area immediately in front of the robot as well as
to the sides and rear. Large sensing shadows near the
robot’s left and right shoulders are covered by two auxiliary cameras mounted at the rear, which also assist in
lining up the robot with openings and obstacles.
A wide angle rear-view camera covers the sides and
rear of the robot and assists in avoiding collisions with
obstacles that may not be visible from the main camera.
It is mounted on a vibration damped flexible boom to
prevent damage.
CASTER also carries high powered lighting, positioned
so as to not shine directly into any of the cameras. Combined with the main camera’s ability to cope with highly
varied lighting conditions, CASTER is able to operate
in lighting conditions ranging from floodlit to complete
darkness.
The microphone provides the operator with feedback
on the robot’s behaviour. By listening, it is possible
to determine if tracks are slipping, if motors are being
loaded, if the robot is high-centered or if it is brushing
up against an obstacle. Finally, an accelerometer provides information on the robot’s attitude and allows the
operator to make decisions as to CASTER’s stability.
Victim identification and diagnosis
Three groups of sensors were used for victim identification – the four colour cameras, the thermal camera and
the stereo microphone. In addition, the range imager
was used to accurately determine the position of victims
once identified (see section 2.2). To aid victim identification, the thermal image data was superimposed on the
main camera image. Occasionally, the auxiliary cameras
were also used for victim identification, especially near
ground level. These two groups of sensors together provide 3 signs of life – heat, form and motion.
Finally, the stereo microphone was intended for use
in locating victims, identifying their state and providing
an additional sign of life. Unfortunately technical issues
resulted in the output being mono-only, albeit with a
high degree of directionality.
Mapping
The core of CASTER’s mapping capabilities lies in the
range imager [Oggier et al., 2004], which provides 3
channels of information per pixel – distance, reflected
Pan-Tilt Unit
Microphone
Main
Camera
Range
Imager
Thermal
Camera
Motor
Motor
Controllers
Controllers
Embedded PC
Network
switch
Onboard notebook computer
Accelerometer
Wireless LAN access point
Network
switch
Base Station
Driving
notebook
Mapping
notebook
Video capture
device
Auxiliary
Auxiliary
Auxiliary
cameras
cameras
cameras
CASTER Robot
Figure 5: Screen shot of the user interface
Figure 4: Schematic diagram of CASTER’s communications system
near-infrared intensity and ambient near-infrared intensity. Mounted on the pan-tilt unit along with the colour
main camera and thermal imager, 7 channels of information (distance, near IR reflectance, near IR ambient,
red, green, blue and temperature) are available for each
pixel and allow for the construction of very informative
3D maps, as well as provide a potentially very rich data
set for further image processing.
The pan-tilt unit assists in this process by providing
very accurate positioning of the sensors, thus allowing
the accurate integration of sensor measurements taken
from different directions. An accelerometer measures the
robot’s pitch and roll, allowing 3D data to be pre-rotated
to the horizontal.
2.3
Computing and Communications
Figure 4 shows the computers and commuication systems employed on CASTER. An embedded computer in
the DT3 base controls the motor interface boards and the
pan-tilt unit whilst a small notebook computer mounted
at the rear handles most of the sensors through firewire
and USB. The exceptions are the auxiliary cameras at
the side and rear which are accessed through the video
capture device. These cameras have a low refresh rate
of approximately 2Hz which was found to be sufficient
due to the robot’s relatively low speed. An onboard fast
Ethernet switch links the video capture device, embedded computer and notebook.
The on-board notebook communicates with the base
station over 802.11a wireless ethernet via its internal antenna. Through this link, information from all sensors
is transmitted back to the base station approximately 5
times per second.
The base station, embedded computer and onboard
notebook run Linux and use TCP/IP as the communication protocol. In order to simplify software development,
the Player robot server is used [Gerkey et al., 2001] on
all three systems. The embedded PC provides a position interface for locomotion and a PTZ interface for the
pan-tilt unit, both of which are mirrored through passthrough drivers on the notebook. The notebook also
provides several camera interfaces including reprocessing
of data from the video capture device, a position interface for the accelerometer and a wireless signal strength
interface. The base station connects to the notebook as
a client and can retrieve data and send commands as
appropriate.
2.4
Software design
In order to gain experience with the robot and sensors,
this year’s entry focused on teleoperation. Elements of
autonomy would then be incorporated when and where
possible. As in section 2.2, the three goals of high quality 3D mapping, reliable victim identification and good
situational awareness drove the design and development
of this software.
Figure 5 shows the teleoperation interface used to control CASTER.
In order to maintain situational awareness and ensure that the robot was successfully operated, steps were
taken to make the robot as easy to operate as possible.
The following principles guided the design of the user
interface.
• Design the robot and the camera placement in order
to enhance situational awareness.
• Try to present the situation in a manner that reduced cognitive load. For instance, where possible
information is displayed graphically rather than textually or numerically.
• Reduce context switching; where the user must
switch from one mode of working to another. Each
context switch can introduce a delay as the user becomes reoriented with the new context.
• Use metaphors and familiar controls from existing
computer applications and the real world in order
to maximise positive transfer.
The display is designed to minimise the amount
of “context switching” and take advantage of existing
analogies wherever possible. The the top left hand corner
is an artificial horizon that displays roll and pitch from
the accelerometer in a manner familiar to anyone who
has experimented with a flight simulator. On the top
right hand right side is a display is a network strength
indicator that is analogous to that found on a mobile
phone, and a speed indicator that shows the fraction of
maximum speed; similar to the bar graphs on some vehicles. All of these are rendered translucently so that the
user can still see through them and thus maintain some
awareness of what is behind them.
A translucent 3D arrow in the middle of the screen
provides feedback on the angle of the pan-tilt unit. It
faces the direction CASTER is pointing, hence indicating the direction in which the robot would travel if it
were driven forward. It was found to be highly intuitive.
As well as the main camera, three other cameras are
displayed along the edges of the screen. The rear view
camera is displayed at the top of the screen but as the
robot skid-steers (and so turns the same way regardless
of travel direction), it is not mirrored. The side auxiliary
cameras are displayed in the lower corners of the screen
and roughly correspond to the lower door-windows in a
minibus.
To reduce sensory overload, a single command allows
the user to hide all additional telemetery and view only
the main camera.
Images from the thermal camera may be overlaid onto
the main camera image. To reduce clutter, areas of
the image below a threshold temperature are unaffected
whilst those above are tinted with a colour between red
and yellow depending on temperature. Experimentally,
it was found that tinting areas between 30◦ and 35◦ was
effective. The user may also optionally observe the raw
or autoexposed full-frame thermal image, which appears
above the right side camera view.
Intuitive control is obtained by adopting controls similar to those used in first-person-shooter computer games
such as Quake, Half-Life and Unreal. The left hand operates the keyboard and controls robot movement via the
keys W, A, S and D, which are arranged in an invertedT. For instance, the W key moves the robot forward as
long as it is held down. The right hand operates a mouse
and controls the pan-tilt unit by holding down the left
mouse button and dragging. The mouse is not captured
to allow the operator to, optionally, run the driving interface in a window in conjunction with other programs
such as a notepad or messaging system. The scrollwheel
on the mouse controls the robot’s forward, reverse and
turning speeds and operates as a throttle preset.
Due to control lag, moving the pan-tilt unit long distances with the mouse became excessively time consuming, thus preset positions were added. These were activated through hotkeys, rapidly moving the pan-tilt unit
to pre-defined positions such as all-the-way-back or justin-front-of-left-track. The operator could then refine the
position with the mouse. The number keys on top of the
keyboard were chosen for the hotkeys. Other keyboard
controls are available for initiating a scan macro action,
hiding additional telemetry information and indicating
the presence of victims and landmarks.
No additional interface or context switch is used to
indicate the position of a landmark or victim. Instead,
from within the driving interface the operator positions
the mouse cursor over the victim or landmark and, instead of dragging with the left mouse button, clicks the
middle mouse button. Based on the cursor’s image position, the range imager locates the corresponding point
in 3D space and automatically annotates the map. The
cursor appears as a blue square to indicate the area over
which the range imager’s measurement will be averaged.
The interface then prompts the user for details of the
landmark or victim via a text entry window. This window is small and does not obstruct the main driving
interface so as to reduce the disorientating effects of the
context switch. Typing text into this entry window is the
only occasion that requires the operator to deviate from
the left-hand-on-keyboard, right-hand-on-mouse configuration; the use of voice recognition is a possible extension for this purpose.
2.5
Mapping and recording data
One purpose of CASTER’s sensors is to generate textured 3D maps of the environments and present them,
along with other collected information, in an intuitive
manner. This process may be broken down into the following components.
• Collecting range, colour and thermal images in a
particular direction, along with annotations such as
landmark and victim details – a “snap” such as in
figure 6
• Merging multiple “snaps” and corresponding annotations from the one location into a local map – a
“scan” such as in figure 7
• Combining multiple “scans” taken in different locations into one global map such as in figure 11
Figure 6: An example of a snap. Red arrow denotes the
robot’s orientation.
• Postprocessing the map and including photographs
and other data to generate intuitive displays and
reports
Figure 7: An example of a scan. Red arrow denotes the
robot’s orientation. Note the annotation of the RoboCup
sign.
Figure 8 also shows a photograph from an elevated
perspective, showing how the 3D reconstruction corresponds to the actual location.
Snaps
The co-location of the range imager, main camera and
thermal camera allows data taken from all three devices
to be merged into one image with 7 channels per pixel
(see 2.2). The range imager was chosen as the reference
sensor, thus the resulting composite image has a resolution of 160x124 pixels and an effective field of view of
45◦ x41◦ . By reading the location of the precision pantilt unit, the location of each pixel in 3D space relative
to the robot may be determined.
A major advantage of co-locating the three sensors is
that the operator can perform annotations in a highly
intuitive fashion (see 2.4 and automatically record location, visual and thermal data, which is stored in the
“snap” data structure. By anchoring annotations to a
particular set of 3D data, the highly local “map” generated by a single snap or scan may be useful even if the
snap or scan itself is mislocalised.
Scans
While the robot is stationary, the precision pan-tilt unit
allows CASTER to take many “snaps” and directly combine them into a local area map around the robot’s
current location, called a “scan”. No correspondance
matching is required as the relative directions of each
“snap” is known to within fractions of a degree. Measurements from the accelerometer are then used to rotate
this local area map relative to the horizon. This enables
Figure 8: A picture (featuring UTS’ Homer robot in
the same position as CASTER was for figures 6 and 7)
but from an elevated position. Note the excellent resemblance.
mapping to be performed even when the robot is at a
steep angle.
To assist in the production of scans, macro actions
were developed to automate the collection of scans.
These actions involve moving the pan-tilt unit around
in such a way that all locations around the robot may
be covered by “snaps”. Generally this involves taking
10 “snaps” at intervals of 36◦ , stopping at each location
for long enough to ensure no motion blur, a process that
takes approximately 20 seconds overall.
If the operator desires, other “snaps” may also be
recorded should objects of interest be above or below
the ring of “snaps” taken by the macro action, so long
as the robot base does not move. As these snaps are also
located precisely relative to other snaps taken from the
same position, landmark and victim annotations may be
located relative to all the snaps in the scan, producing a
very accurate and informative local area map.
Figure 9: The user interface used for manual scan registration
Global Map
Each “scan” is pre-rotated based on accelerometer data
so it is only necessary to move “scans” in 4 dimensions –
X,Y,Z and yaw – to register them with the global map.
Odometry is too poor on CASTER – especially given the
uneven surface – for use in matching since the robot
skid steers and often operates on surfaces that can shift.
In fact, it was found that an odometric motion model
would often be a poorer estimate of position than a zeromotion model, primarily due to large intermediate errors
in heading.
Variations on the Iterative Closest Point (ICP) [Surmann et al., 2004a] algorithm were investigated in order
to register scans. This algorithm finds correspondences
between points of an incoming data set and an existing
model set by assuming that a given point in one set corresponds to the closest point in the other. The two sets
are then transformed to minimise the least-squared error between corresponding points. The process iterates
until some stopping condition, generally based on the reduction in error between iterations, is reached. Various
heuristics may be applied to cull points that cannot be
matched, break correspondances that can be eliminated
based on distance or other factors, and weight correspondances.
Unfortunately, being a gradient descent algorithm,
ICP suffers from local minima. Each pair of scans involves around 400,000 to 500,000 3D points, of which
around 10% correspond to depth edges and other “significant” features. Thus techniques that attempt to address local minima issues, such as simulated annealing,
become very expensive. Also, there must be significant
overlapping regions between scans, otherwise ICP tends
to match globally consistent features, which are invariably the ring of floor pixels around each scan. Thus
whilst two scans may be taken at a reasonable distance,
the local (and indeed, global) minimum might in fact be
the two scans on top of each other.
As the operator takes scans, landmarks are highlighted. The range imager allows the system to locate
the 3D position of these landmarks relative to the robot.
If two or more landmarks in one scan match landmarks
in the existing model, in theory it becomes possible to
exactly register the new scan with the existing model
(the extra degrees of freedom are accounted for since the
robot’s tilt and roll are known from the accelerometer).
This system was implemented and was effective as long
as the operator marked landmarks accurately.
Unfortunately, this process is time consuming from the
operator’s perspective as they must point the pan-tilt
unit and input the landmark label. It was found that
it was faster for the operator to manually register the
scans using a user interface than it was for the operator
to specify enough landmarks for automated matching.
Therefore, for competition runs, where the 10-minute
time limit resulted in the operator only being able to take
3 or 4 scans per run, manual scan registration was preferred. The interface used for performing manual scan
registration is shown in figure 9.
Synthesis and presentation to incident
commander
Once the scans are registered, another software tool
called RescueVis allows the operator to view the report
and data on victims. It also allows the printing of a report that shows where the victims are and images of the
victims. This map and victim report is then handed to
the Incident Commander for scoring. The user interface
for RescueVis is shown in Figure 10.
Figure 11 shows a typical map generated for the Inci-
Figure 10: User interface showing victim browsing in
RescueVis
dent Commander using the 3D capabilities of RescueVis.
Subsequent screens also show pictures of the victims as
well as nearby landmarks.
3
Evaluation
Initial tests were conducted at the USAR Test Facility at
UTS. Performance at the competition was also carefully
monitored to see how the overall system design functioned and how the individual components fared.
3.1
Mechanical Platform
CASTER’s powerful drivetrain played an important role
in the success of the robot, allowing it to overcome very
difficult obstacles and terrain. Its robustness was also
a great asset as the robot was often high-centered or
pushed up against obstacles. This robustness was graphically demonstrated when the robot flipped and landed
heavily on the armoured camera mount – the only damage sustained was minor damage to a suspension linkage
and the rearview camera boom.
The CASTER’s articulated design helped to keep the
robot’s center of gravity low and allowed the tracks to
follow the terrain. This was especially the case when
climbing stairs and step obstacles. However, the robot’s
weight was a mixed blessing. The robot was less likely
to slide on loose material but was more likely to become
stuck when high-centered. Its weight also caused it to
unintentionally push obstacles and stairs around.
3.2
Situational Awareness
The four cameras placed around the robot were effective
in driving the robot around but the narrow field of view
of the main camera, combined with the slow speed of
the pan-tilt unit, slowed down surveys of the robot’s surroundings. The inability to directly obtain a top-down
Figure 11: Generated 3D map for of Round 1 of the
Preliminaries of Robocup Rescue 2005
view of the robot was a limitation on the step field and
other tight areas. This could be resolved using an omnidirectional or wide angle camera mounted high up. The
side auxiliary cameras provided valuable perspectives for
situational awareness, especially since they were close to
the ground. This location left them vulnerable to damage however – one camera was almost destroyed when it
became snagged on debris.
3.3
Sensor Performance
Main Camera
The main camera was a compromise between a high resolution, narrow angle camera for the purpose of reading
victim labels and identifying victims, and a wide angle
camera for situational awareness and surveying. While
it was acceptable in performing both tasks, it was really good at neither. The use of a camera with optical
zoom or a combination of a wide and narrow angle camera might have been more suited to this task. However,
the camera’s exceptional colour rendition and ability to
automatically deal with highly varied lighting conditions
was quite valuable.
Thermal Camera
All of the victims in the arena emit heat – in fact some
victims are hidden and only visible from their heat signatures. The thermal camera was vital in Team CASualty’s high map scoring as it allowed the victims’ temperatures to be determined unambiguously. The relatively
simple method by which the thermal image was matched
with vision – a linear shift – proved to be acceptable for
the purpose of the competition. More accurate matching, either by using range information or by mounting
the thermal camera closer to the main camera, would
be necessary if automated victim detection were to be
attempted.
Range Imager
The ability to gather a dense depth image in one shot
was extremely valuable, both for generating maps as well
as for automatically locating landmarks and victims relative to the robot. After calibration, this sensor proved
to be very accurate in determining the distance to relatively smooth objects up to a distance of approximately
5 meters. It was also surprisingly robust to lighting, surface type or other objects in the scene. However, being
a very new sensor, the CSEM SwissRanger 2 had some
characteristics that caused major problems. The most
important was the aliasing of measurements for objects
at distances greater than 7.5m. Such distances could often be seen above the partitions used to construct the
robot arena. Almost as severe a problem was focal blurring, which resulted in the loss of sharp discontinuities in
the depth image. These two factors played a large part in
the difficulties encountered with automated registration.
Stereo Microphone
Despite being operated in mono due to technical difficulties, the microphone was useful in detecting sounds
from victims. This could well have been the difference
between 3rd and 5th place. Surprisingly however, the
microphone was also found to be useful for situational
awareness. With practice, the operator could determine
the characteristics of the surface on which the robot was
moving, which tracks were under load, if the robot was
pushing up against an unseen obstacle or if there was
debris caught under the robot.
3.4
Mapping and Victim Scoring
Due to the tight time constraints imposed during the
competitive runs, it was only possible to take 2-4 scans
per run. It was also impractical to take too many landmarks and therefore all the scan registration performed
in scored runs was done manually. Despite this, the 3D
textured and annotated maps produced were received
very well and consistently scored full points.
A large part of this can be attributed to the ability
of the operator to locate a victim in the map by simply pointing and clicking – the presence of the range
imager allows the 3D position of the victim to be determined very easily. By co-locating the thermal camera,
scoring of victims based on thermal images was trivial
and effectively instant, in contrast to methods involving
spot non-contact thermometers where it was necessary
to demonstrate an elevated temperature on a victim.
The victim tags, which contribute greatly to scoring,
were significantly smaller than anticipated and, for the
most part, it was very difficult for the main camera to
observe them due to its height and lack of optical zoom.
This was the case even when run at its highest resolution.
At times, it was necessary to resort to manouvering the
rear of the robot close to a victim tag and using the
side auxiliary cameras (which are lower to the ground)
in order to read the tags.
4
Conclusions and Future Work
Team CASualty came third in the finals of Robocup Rescue 2005. It came second in the preliminary round and
second in the semi-finals. On average, Team CASualty
scored more per victim found than most other teams.
The main factors leading to this were that we were able
to present the information coming back from the different sensors in an intuitive way, allowing the operator to
observe victims easily and produce high quality maps
that were not just top-down but also three dimensional.
There are still many issues that remain to be improved
and resolved. While landmark-based scan registration
seemed to work some of the time, it was not sufficiently
reliable nor fast enough. The characteristics of the range
imager caused problems for automated scan registration,
further work in this area may improve performance.
While Player was a good platform for control, it
was not designed for bandwidth-intensive sensors such
as the high-resolution camera, thermal camera and
range imager. Therefore it is our intention to separate bandwidth-intensive devices from less bandwidth
intensive devices and have an additional server for highbandwidth data.
A number of approaches are also being explored to
address situational awareness issues; such as mounting a
camera on a robotic arm that can be raised in order to
better characterise the surroundings of the robot.
In addition, mobility remains an issue. The DT3 appears to be one of the better platforms available (DT3based robots also took second and fifth place). Despite
this, there were still a number of times when CASTER
ran into difficulties – the robot flipped once and became
terminally stuck in two rounds of the competition. This
may in part be because our operator only had one month
of practice and the video refresh rate was too slow; but it
is also in part a limitation of using a single robot. Other
teams with the same robot base sometimes became stuck
in similar locations. One way to increase robsutness is
therefore to employ multiple robots, but since the replacement cost of CASTER is approximately A$110,000,
multiple CASTERs are not an option.
For future competitions, multiple robots are being
considered. It is likely that one will be autonomous,
in co-operation with the University of Technology (also
part of Team CASualty). But it may also possible to
use multiple small, inexpensive yet highly mobile semiautonomous robots.
5
Acknowledgement
This research was supported by the Australian Research
Council Centre of Excellence for Autonomous Systems.
References
[Gerkey et al., 2001] B. Gerkey, R. Vaughan, K. Sty,
A. Howard, G. Sukhatme, and M. Mataric. Most valuable player: A robot device server for distributed control. In Proceedings of the IEEE/RSJ International
Conference on Intelligent Robots and Systems (IROS),
Wailea, Hawaii, October 2001.
[Jacoff et al., 2004] Adam Jacoff, Elena Messina, and
John Evans. A standard test course for urban search
and rescue robots. In Proceedings of the Performance Metrics for Intelligent Systems Workshop, August 2004.
[Koyanagi et al., 2004] Eiji Koyanagi, Yoshiyuki Ooba,
Shuuhei Yoshida, and Yasuo Hayashibara. Toin pelican. In Proceedings of the 2004 RoboCup Symposium,
2004.
[NIST, 2003] NIST. 2003 robot rescue competitions.
http://robotarenas.nist.gov/2003_competitions/,
2003. Viewed 23 August 2005.
[NIST, 2004] NIST. 2004 robot rescue competitions.
http://robotarenas.nist.gov/2004_competitions/,
2004. Viewed 23 August 2005.
[Oggier et al., 2004] T. Oggier, M. Lehmann, R. Kaufmann, M. Schweizer, M. Richter, P. Metzler, G. Lang,
F. Lustenberger, and N. Blanc. An all-solid-state optical range camera for 3D real-time imaging with subcentimeter depth resolution (SwissRanger). In Optical
Design and Engineering. Edited by Mazuray, Laurent;
Rogers, Philip J.; Wartmann, Rolf. Proceedings of the
SPIE, volume 5249, pages 534–545, February 2004.
[Surmann et al., 2004a] Hartmut Surmann, Andreas
Nuchter, Kai Lingemann, and Joachim Hertzberg. 6d
slam – preliminary report on closing the loop in six dimensions. In Proceedings of the 5th IFAC Symposium
on Intelligent Autonomous Vehicles, 2004.
[Surmann et al., 2004b] Hartmut Surmann, Rainer
Worst, Matthias Hennig, Kai Lingemann, Andreas
Nuechter, Kai Pervoelz, Kiran Raj Tiruchinapalli, Thomas Christaller, and Joachim Hertzberg.
Robocuprescue - robot league team kurt3d, germany.
In Proceedings of the 2004 RoboCup Symposium,
2004.
[Yujin, 2005] Yujin.
Robhaz dt3 web site.
http://www.robhaz.com/about_dt3_main.asp,
2005. Viewed 18 August 2005.
[Zalud, 2003] Ludek
Zalud.
Robocup
rescue robot league competition, awardee paper,
robrno,
czech replublic,
1st place.
http://robotarenas.nist.gov/2003_competitions/,
July 2003.