Sie sind auf Seite 1von 6

Study of the SR4000 camera

Miguel Cazorla, Diego Viejo, and Cristina Pomares


I.U.I.I.
Universidad de Alicante
miguel.cazorla@ua.es, dviejo@dccia.ua.es, cpomares@dccia.ua.es

Abstract space, without being confined to the ground,


watching and avoiding 3D shape and volume
3D data has been used for robotics tasks in the objects.
last years. This data provides valuable infor- In this paper, we are going to study a new
mation about the robot environment. Tradi- 3D camera, the SR4000 from Mesa Imaging.
tionally, stereo cameras has been used to ob- This camera is not a stereo camera, but a
tain 3D data, but these kind of cameras do infrared one. One advantage of this kind of
not provide information in the lack of texture. cameras is that we get 3D data in areas when
There is a new camera, SR4000, which uses no texture is present, like walls. Our goal is
infrared in order to get richer information. In to know how good is this camera for image
this paper, we explore this camera in order to processing and robotics tasks. In a previous
know its main characteristics and how good is work, we have developed an algorithm (see
the data from this camera in robotics tasks. [3]) which extract 2D features (like SIFT and
SURF) from stereo images and gives 3D coor-
dinates to them. Then, an ICP-like algorithm
1 Introduction is applied to get the egomotion of the camera.
We will try to adapt the previous algorithm
During the last years, new 3D sensor devices
to the data from this camera. But, first, we
have been developed, and computing capa-
explain the characteristics of the SR4000 cam-
bilities have been improved. These improve-
era: image resolution, the data obtained from
ments can be used to obtain and process a
it, and some other interesting properties.
better robot environment information in the
The paper is organized as follows: First, in
field of mobile robotics [15], [11], [8]. By now,
section 2 we present an overview of the camera
methods for achieving tasks such as localiza-
SR4000 and the data obtained from it. Then,
tion [2], [5], [13], navigation [14], [10] or au-
Section 3 gives a brief explanation about the
tomatic map building [12], were restricted to
algorithm we are trying to apply to the cam-
the two dimensional world which could be cap-
era in order to test it. The results and con-
tured by the robot sensors. Nevertheless, us-
clusions obtained during this paper and the
ing the new 3D sensors, it is possible to im-
future work in section 4.
prove the representation of observed objects
in order to use them into applications such
as augmented reality, architecture, manufac- 2 Time-of-Flight 3D Camera
turing process, SLAM problem, etc. Further-
more, this new dimension can be used to im- In recent years, Time-of-Flight (ToF) cam-
prove the methods and behaviors used by a eras are being developed as a new technology
robot in order to accomplish its objectives. In that delivers range (distance) and amplitude
this way, the robots equipped with this new maps by the use of a modulated light source.
3D sensors are able to move freely into a 3D The main advantages with respect to other
3D devices are the possibility to acquire data
at video frame rates and to obtain 3D point Table 1: SR4000 specifications
clouds without scanning and from just one
point of view. The basic principle of ToF cam- Pixel array size 176 (h) x 144 (v)
eras consists of an amplitude-modulated in- Field of View 43.6◦ (h) x 34.6◦ (v)
frared light source and a sensor field that mea- Pixel Pitch 40µm
sures the intensity of backscattered infrared Angular Resolution 0.23◦
light. The infrared source is constantly emit- Illumination wavelength 850nm
ting light that varies sinusoidally. Objects that Modulation Frequency 30Mhz
have different distances are reached by differ- Operating range 0.3 to 5.0 meters
ent parts of the sinusoidal wave. The reflected Distance accuracy +/-1cm
light is then compare to the original one, calcu- Maximun frame rate 54 fps
lating the phase shift, by means of measuring Dimensions 65x65x68mm
the intensity of the incoming light since the Weight 470g
phase shift is proportional to the time of flight
of the light reflected by a distant object. A
detailed description of the time-of-flight prin-
ciple can be found in [7]. measurements or to reject low quality ones.
The device used in this work is the Swis- The data output are arrays of 16 bit words,
sRanger SR4000 ToF camera, shown in Fig- of length (176 x 144 = 25344). By default,
ure 1. the amplitude data is converted into a value
which is independent of distance and position
in the image array. The amplitude data has
a range of 0-0x7FFF. In our experiments the
amplitude data has low contrast so they have
been equalized. Figure 2 shows an overview
for the typical data obtained with the SR4000.
The recorded 3D points cloud can be observed
at the figure top center, corresponding am-
plitud by the left side and confidence by the
Figure 1: SwissRanger SR4000 Time-of-Flight right. Reference camera coordinate system is
camera. also shown.

In order to give and idea of the characteris-


tics of the camera used in this work, the prin-
cipal specifications of the SR4000 camera are
reported in Table 1. For more details about
camera specifications see [1].
In our tests all the data were acquired
with the software provided with the camera,
which delivers point coordinates XYZ, ampli-
tude data of the scene and a confidence map of
the distance measurements. In particular, the
confidence map is obtained using a combina-
tion of distance and amplitude measurements
and their temporal variations: it represents a
measure of probability that the distance mea- Figure 2: Camera data overview. The SR4000 cap-
surement of each pixel is correct, so it can be tures both a 3D point set and two maps: amplitude
useful to select regions containing high quality (left) and confidence (right).
The Field-of-View (the size of the area il-
luminated and measured by the SR4000) is a
function of the distance from the camera. The
lateral resolution of the camera is determined
by its Field-of-View in combination with the
number of pixels in the sensor field. Each of
the 176 (horizontal) and 144 (vertical) pixels
corresponds to approximately 0.23◦ of angular
resolution.
ToF cameras allow to generate point clouds
of real time acquisition. The accuracy of ToF
cameras varies with internal components and
the characteristics of the observed scene, such
us objects reflectivity and ambient lighting
conditions. These errors cannot be fully elimi- Figure 3: Effects are shown for different integra-
nated, but they can be reduced and optimized tion time values. First column shows the distance
maps, the second one shows the amplitude maps
thanks to filtering or several techniques, such
and the third one the confidence map. In the first
us averaging techniques or calibration proce-
row the integration time is set to 25ms (noise cor-
dures [4] where a distance error model was rupts the measure), in the second row it is set to
proposed which provided a reduction of dis- 102ms (automatic, better tuning) and in the third
tance errors in the 1.5-4m distance measure- row it is set to 250ms, which causes saturation ef-
ment range. fects.
Integration Time is one of the most impor-
tant camera control. Adjusting this value con-
trols how long each sensor pixel collects light.
For lowest noise measurements the integration
time should be adjusted so that all (or at least integration time and amplifying the illumina-
most) pixels collect as much light as possible tion. The second error, due to interreflections
without saturating. On the other hand if a in the scene, is the multiple ways reflection,
high frame rate is more important then the in- and appears hollows and corners round off and
tegration time may be reduced to achieve the occluding shapes. Third, light scattering oc-
desired frame rate. The camera software al- curs in the lenses of the ToF camera. This,
lows to automatically adjust the Integration near bright objets may superpose the measure-
Time depending on the maximum amplitudes ments from the background objects. Further-
present in the current image. This setting more, there are three systematic errors. First,
can be used to avoid pixel saturation and to there is the distance-related error, based on
achieve a good balance between noise and high the assumption that the emitted light is si-
frame rate. We placed the camera in a fixed nusoidal. Second, an amplitude-error, caused
position and we took several images varying by non-linearities of the pixel’s electronic com-
the integration time. There was a skylight ponents. Third, a fixed pattern phase noise,
ahead the camera so the ambient light affected where the triggering of each pixel depends on
the performance. In Figure 3 three tests show the on chip. These three errors are manage-
the variation of noise and saturation with dif- able by calibration [6] [4]. Other important is-
ferent integration time values . sue for Tof cameras is the aliasing effect due to
The following explanations relate to the er- the periodicity of the modulated signal, where
rors of the ToF camera. There are three non- the distances to objects differing in phase by
systematic errors. A bad signal-to-noise ra- 360◦ of phase shift are not distinguishable. Us-
tio distorts the measurement and cannot be ing multiple frequencies is a common way of
suppressed. A solution can be increasing the dealiasing this data [7].
3 Applying an Feature Extraction conclusions can be drawn from them:
Algorithm to Amplitude Images
• There are only a few number of features.
In this section we are going to test how good Due to the low image resolution, only 150
are images from this camera in order to apply or 200 features are obtained. This is a
a feature extraction algorithm. To do that, we problem if our goal is to match features
compute 6DoF egomotion using a ToF camera. between two images, if there is no so much
This is an adaptation of the method proposed overlapping in those images.
in [3]. In this paper a new approach for com-
puting egomotion from stereo images is pro-
posed. Basically, the process to obtain the
transformation between two consecutive im-
ages consists of extracting features from each
input image, translating the features to the
3D space using the sensor geometry, matching
those 3D features using their descriptors from
each input image and, finally, computing the Figure 4: Example of two consecutive images with
transformation that best aligns the matched the Sift results.
features.
The main difference from a stereo vision
• The quality of the images is low. There
camera and our SR4000 is that ours does not
are a lot of noise. This implies that al-
captures an image in the visible spectra but
though we could get a good feature loca-
in the infrared. Therefore, we use the infrared
tion, descriptors are very different, so a
image for extracting visual features. For the
bad matching is obtained.
experiments presented in this paper we are us-
ing SIFT [9] features and descriptors. We ex- • Another problem with this camera is the
tract features for the amplitude infrared im- one shown in Figure 5 (see [1]). In this
age. Figure 4 shows the SIFT features ex- case, the camera has bad behavior when
tracted from two amplitude images. In the the surfaces are not perpendicular to it.
next step, features are translated to the 3D im- In Figure 4, the one of the left shows a
age. Then, using the confidence map provided distortion on the sides (for example, the
by our camera, we can remove those features column at the right). This makes again
we can not trust. This is an important dif- different descriptors for the same feature.
ference from the stereo systems since it makes
enable an important accuracy improvement as
we remove erroneous points. The rest of the • The last identified problem is the one
method remains unaltered. The features from shown in Figure 6. There, when a bright
two consecutive images are matched looking surface (like a door) is present and close
for the nearest feature in the descriptors do- to the camera, some artifacts are present
main. Finally, the egomotion is computed and bad results are obtained from it. The
from the matched features. left images in Figure 6 are the amplitude
and the right ones are the confidence.

4 Experiments and Conclusions One of the advantages of this camera, as we


have exposed above, is the ability to get data
We have applied the algorithm explained in in the lack of texture.
the previous section to a sequence of images. Finally, we have tried to apply the egomo-
In Figure 4 we can see the result of applying tion algorithm and the result (for a right tra-
the Sift algorithm (Surf had similiar bad re- jectory) is shown in Figure 7. We were not
sults) to two images of the sequence. Some able to get a good result due to the problems
Figure 5: Problems due to distorsions.
Figure 7: Trajectory obtained applying a ICP-like
algorithm.

[2] Wolfram Burgard, Dieter Fox, and Sebas-


tian Thrun. Active mobile robot localiza-
tion. Technical Report IAI-TR-97-3, 25,
1997.
[3] Miguel Cazorla, Diego Viejo, Andrs Hern-
ndez, Juan Nieto, and Eduardo Nebot.
Large scale egomotion and error analysis
with visual features. Journal of Physical
Agents, 4:19–24, 2010.
[4] Filiberto Chiabrando, Roberto
Chiabrando, Dario Piatti, and Ful-
Figure 6: Problems due to reflections. vio Rinaudo. Sensors for 3d imaging:
Metric evaluation and calibration of a
ccd/cmos time-of-flight camera. Sensors,
described above. As future work, we plan to 9(12):10080–10096, 2009.
work with images in order to improve their
[5] Frank Dellaert, Dieter Fox, Wolfram Bur-
quality. We also plan to continue exploring
gard, and Sebastian Thrun. Monte carlo
the response of the camera with computer vi-
localization for mobile robots. In IEEE
sion algorithms.
International Conference on Robotics and
Automation (ICRA99), May 1999.
Acknowledgment [6] Stefan Fuchs and Gerd Hirzinger. Extrin-
sic and depth calibration of tof-cameras.
This work has been supported by grant In CVPR, 2008.
DPI2009-07144 from Ministerio de Ciencia e
Innovacion of the Spanish Government. [7] S. Burak Gokturk, Hakan Yalcin, and
Cyrus Bamji. A time-of-flight depth sen-
sor - system description, issues and so-
References lutions. In CVPRW ’04: Proceedings of
the 2004 Conference on Computer Vi-
[1] SR4000 User Manual. sion and Pattern Recognition Workshop
(CVPRW’04) Volume 3, page 35, Wash- C. Baker, Z. Omohundro, S. Thayer, and
ington, DC, USA, 2004. IEEE Computer W. Whittaker. A system for volumet-
Society. ric robotic mapping of abandoned mines.
In Proc. IEEE International Conference
[8] Chen-Tsung Kuo and Shyi-Chyi Cheng. on Robotics and Automation ICRA ’03,
3d model retrieval using principal plane volume 3, pages 4270–4275, 14–19 Sept.
analysis and dynamic programming. Pat- 2003.
tern Recogn., 40(2):742–755, 2007.
[13] Sebastian Thrun, Dieter Fox, Wolfram
[9] David G. Lowe. Distinctive image fea-
Burgard, and Frank Dellaert. Ro-
tures from scale-invariant keypoints. In-
bust monte carlo localization for mo-
ternational Journal of Computer Vision,
bile robots. Artificial Intelligence, 128(1-
60:91–110, 2004.
2):99–141, 2001.
[10] Gian Luca Mariottini and Domenico
Prattichizzo. Image-based visual servoing [14] A. Tsalatsanis, K. Valavanis, and
with central catadioptric cameras. Int. J. N. Tsourveloudis. Mobile robot naviga-
Rob. Res., 27(1):41–56, 2008. tion using sonar and range measurements
from uncalibrated cameras. In Proc. 14th
[11] D. Schroter and M. Beetz. Acquir- Mediterranean Conference on Control
ing models of rectangular 3d objects for and Automation MED ’06, pages 1–7,
robot maps. In Proc. IEEE International June 2006.
Conference on Robotics and Automation
ICRA ’04, volume 4, pages 3759–3764, [15] J. Weingarten, G. Gruener, and R. Sieg-
Apr 26–May 1, 2004. wart. A Fast and Robust 3D Feature Ex-
traction Algorithm for Structured Envi-
[12] S. Thrun, D. Hahnel, D. Ferguson, ronment Reconstruction. In None, 2003.
M. Montemerlo, R. Triebel, W. Burgard,

Das könnte Ihnen auch gefallen