0 Bewertungen0% fanden dieses Dokument nützlich (0 Abstimmungen)
28 Ansichten6 Seiten
3D data has been used for robotics tasks in the last years. A new camera, SR4000, uses infrared in order to get richer information. Our goal is to know how good is this camera for image processing and robotics tasks.
3D data has been used for robotics tasks in the last years. A new camera, SR4000, uses infrared in order to get richer information. Our goal is to know how good is this camera for image processing and robotics tasks.
Copyright:
Attribution Non-Commercial (BY-NC)
Verfügbare Formate
Als PDF, TXT herunterladen oder online auf Scribd lesen
3D data has been used for robotics tasks in the last years. A new camera, SR4000, uses infrared in order to get richer information. Our goal is to know how good is this camera for image processing and robotics tasks.
Copyright:
Attribution Non-Commercial (BY-NC)
Verfügbare Formate
Als PDF, TXT herunterladen oder online auf Scribd lesen
I.U.I.I. Universidad de Alicante miguel.cazorla@ua.es, dviejo@dccia.ua.es, cpomares@dccia.ua.es
Abstract space, without being confined to the ground,
watching and avoiding 3D shape and volume 3D data has been used for robotics tasks in the objects. last years. This data provides valuable infor- In this paper, we are going to study a new mation about the robot environment. Tradi- 3D camera, the SR4000 from Mesa Imaging. tionally, stereo cameras has been used to ob- This camera is not a stereo camera, but a tain 3D data, but these kind of cameras do infrared one. One advantage of this kind of not provide information in the lack of texture. cameras is that we get 3D data in areas when There is a new camera, SR4000, which uses no texture is present, like walls. Our goal is infrared in order to get richer information. In to know how good is this camera for image this paper, we explore this camera in order to processing and robotics tasks. In a previous know its main characteristics and how good is work, we have developed an algorithm (see the data from this camera in robotics tasks. [3]) which extract 2D features (like SIFT and SURF) from stereo images and gives 3D coor- dinates to them. Then, an ICP-like algorithm 1 Introduction is applied to get the egomotion of the camera. We will try to adapt the previous algorithm During the last years, new 3D sensor devices to the data from this camera. But, first, we have been developed, and computing capa- explain the characteristics of the SR4000 cam- bilities have been improved. These improve- era: image resolution, the data obtained from ments can be used to obtain and process a it, and some other interesting properties. better robot environment information in the The paper is organized as follows: First, in field of mobile robotics [15], [11], [8]. By now, section 2 we present an overview of the camera methods for achieving tasks such as localiza- SR4000 and the data obtained from it. Then, tion [2], [5], [13], navigation [14], [10] or au- Section 3 gives a brief explanation about the tomatic map building [12], were restricted to algorithm we are trying to apply to the cam- the two dimensional world which could be cap- era in order to test it. The results and con- tured by the robot sensors. Nevertheless, us- clusions obtained during this paper and the ing the new 3D sensors, it is possible to im- future work in section 4. prove the representation of observed objects in order to use them into applications such as augmented reality, architecture, manufac- 2 Time-of-Flight 3D Camera turing process, SLAM problem, etc. Further- more, this new dimension can be used to im- In recent years, Time-of-Flight (ToF) cam- prove the methods and behaviors used by a eras are being developed as a new technology robot in order to accomplish its objectives. In that delivers range (distance) and amplitude this way, the robots equipped with this new maps by the use of a modulated light source. 3D sensors are able to move freely into a 3D The main advantages with respect to other 3D devices are the possibility to acquire data at video frame rates and to obtain 3D point Table 1: SR4000 specifications clouds without scanning and from just one point of view. The basic principle of ToF cam- Pixel array size 176 (h) x 144 (v) eras consists of an amplitude-modulated in- Field of View 43.6◦ (h) x 34.6◦ (v) frared light source and a sensor field that mea- Pixel Pitch 40µm sures the intensity of backscattered infrared Angular Resolution 0.23◦ light. The infrared source is constantly emit- Illumination wavelength 850nm ting light that varies sinusoidally. Objects that Modulation Frequency 30Mhz have different distances are reached by differ- Operating range 0.3 to 5.0 meters ent parts of the sinusoidal wave. The reflected Distance accuracy +/-1cm light is then compare to the original one, calcu- Maximun frame rate 54 fps lating the phase shift, by means of measuring Dimensions 65x65x68mm the intensity of the incoming light since the Weight 470g phase shift is proportional to the time of flight of the light reflected by a distant object. A detailed description of the time-of-flight prin- ciple can be found in [7]. measurements or to reject low quality ones. The device used in this work is the Swis- The data output are arrays of 16 bit words, sRanger SR4000 ToF camera, shown in Fig- of length (176 x 144 = 25344). By default, ure 1. the amplitude data is converted into a value which is independent of distance and position in the image array. The amplitude data has a range of 0-0x7FFF. In our experiments the amplitude data has low contrast so they have been equalized. Figure 2 shows an overview for the typical data obtained with the SR4000. The recorded 3D points cloud can be observed at the figure top center, corresponding am- plitud by the left side and confidence by the Figure 1: SwissRanger SR4000 Time-of-Flight right. Reference camera coordinate system is camera. also shown.
In order to give and idea of the characteris-
tics of the camera used in this work, the prin- cipal specifications of the SR4000 camera are reported in Table 1. For more details about camera specifications see [1]. In our tests all the data were acquired with the software provided with the camera, which delivers point coordinates XYZ, ampli- tude data of the scene and a confidence map of the distance measurements. In particular, the confidence map is obtained using a combina- tion of distance and amplitude measurements and their temporal variations: it represents a measure of probability that the distance mea- Figure 2: Camera data overview. The SR4000 cap- surement of each pixel is correct, so it can be tures both a 3D point set and two maps: amplitude useful to select regions containing high quality (left) and confidence (right). The Field-of-View (the size of the area il- luminated and measured by the SR4000) is a function of the distance from the camera. The lateral resolution of the camera is determined by its Field-of-View in combination with the number of pixels in the sensor field. Each of the 176 (horizontal) and 144 (vertical) pixels corresponds to approximately 0.23◦ of angular resolution. ToF cameras allow to generate point clouds of real time acquisition. The accuracy of ToF cameras varies with internal components and the characteristics of the observed scene, such us objects reflectivity and ambient lighting conditions. These errors cannot be fully elimi- Figure 3: Effects are shown for different integra- nated, but they can be reduced and optimized tion time values. First column shows the distance maps, the second one shows the amplitude maps thanks to filtering or several techniques, such and the third one the confidence map. In the first us averaging techniques or calibration proce- row the integration time is set to 25ms (noise cor- dures [4] where a distance error model was rupts the measure), in the second row it is set to proposed which provided a reduction of dis- 102ms (automatic, better tuning) and in the third tance errors in the 1.5-4m distance measure- row it is set to 250ms, which causes saturation ef- ment range. fects. Integration Time is one of the most impor- tant camera control. Adjusting this value con- trols how long each sensor pixel collects light. For lowest noise measurements the integration time should be adjusted so that all (or at least integration time and amplifying the illumina- most) pixels collect as much light as possible tion. The second error, due to interreflections without saturating. On the other hand if a in the scene, is the multiple ways reflection, high frame rate is more important then the in- and appears hollows and corners round off and tegration time may be reduced to achieve the occluding shapes. Third, light scattering oc- desired frame rate. The camera software al- curs in the lenses of the ToF camera. This, lows to automatically adjust the Integration near bright objets may superpose the measure- Time depending on the maximum amplitudes ments from the background objects. Further- present in the current image. This setting more, there are three systematic errors. First, can be used to avoid pixel saturation and to there is the distance-related error, based on achieve a good balance between noise and high the assumption that the emitted light is si- frame rate. We placed the camera in a fixed nusoidal. Second, an amplitude-error, caused position and we took several images varying by non-linearities of the pixel’s electronic com- the integration time. There was a skylight ponents. Third, a fixed pattern phase noise, ahead the camera so the ambient light affected where the triggering of each pixel depends on the performance. In Figure 3 three tests show the on chip. These three errors are manage- the variation of noise and saturation with dif- able by calibration [6] [4]. Other important is- ferent integration time values . sue for Tof cameras is the aliasing effect due to The following explanations relate to the er- the periodicity of the modulated signal, where rors of the ToF camera. There are three non- the distances to objects differing in phase by systematic errors. A bad signal-to-noise ra- 360◦ of phase shift are not distinguishable. Us- tio distorts the measurement and cannot be ing multiple frequencies is a common way of suppressed. A solution can be increasing the dealiasing this data [7]. 3 Applying an Feature Extraction conclusions can be drawn from them: Algorithm to Amplitude Images • There are only a few number of features. In this section we are going to test how good Due to the low image resolution, only 150 are images from this camera in order to apply or 200 features are obtained. This is a a feature extraction algorithm. To do that, we problem if our goal is to match features compute 6DoF egomotion using a ToF camera. between two images, if there is no so much This is an adaptation of the method proposed overlapping in those images. in [3]. In this paper a new approach for com- puting egomotion from stereo images is pro- posed. Basically, the process to obtain the transformation between two consecutive im- ages consists of extracting features from each input image, translating the features to the 3D space using the sensor geometry, matching those 3D features using their descriptors from each input image and, finally, computing the Figure 4: Example of two consecutive images with transformation that best aligns the matched the Sift results. features. The main difference from a stereo vision • The quality of the images is low. There camera and our SR4000 is that ours does not are a lot of noise. This implies that al- captures an image in the visible spectra but though we could get a good feature loca- in the infrared. Therefore, we use the infrared tion, descriptors are very different, so a image for extracting visual features. For the bad matching is obtained. experiments presented in this paper we are us- ing SIFT [9] features and descriptors. We ex- • Another problem with this camera is the tract features for the amplitude infrared im- one shown in Figure 5 (see [1]). In this age. Figure 4 shows the SIFT features ex- case, the camera has bad behavior when tracted from two amplitude images. In the the surfaces are not perpendicular to it. next step, features are translated to the 3D im- In Figure 4, the one of the left shows a age. Then, using the confidence map provided distortion on the sides (for example, the by our camera, we can remove those features column at the right). This makes again we can not trust. This is an important dif- different descriptors for the same feature. ference from the stereo systems since it makes enable an important accuracy improvement as we remove erroneous points. The rest of the • The last identified problem is the one method remains unaltered. The features from shown in Figure 6. There, when a bright two consecutive images are matched looking surface (like a door) is present and close for the nearest feature in the descriptors do- to the camera, some artifacts are present main. Finally, the egomotion is computed and bad results are obtained from it. The from the matched features. left images in Figure 6 are the amplitude and the right ones are the confidence.
4 Experiments and Conclusions One of the advantages of this camera, as we
have exposed above, is the ability to get data We have applied the algorithm explained in in the lack of texture. the previous section to a sequence of images. Finally, we have tried to apply the egomo- In Figure 4 we can see the result of applying tion algorithm and the result (for a right tra- the Sift algorithm (Surf had similiar bad re- jectory) is shown in Figure 7. We were not sults) to two images of the sequence. Some able to get a good result due to the problems Figure 5: Problems due to distorsions. Figure 7: Trajectory obtained applying a ICP-like algorithm.
[2] Wolfram Burgard, Dieter Fox, and Sebas-
tian Thrun. Active mobile robot localiza- tion. Technical Report IAI-TR-97-3, 25, 1997. [3] Miguel Cazorla, Diego Viejo, Andrs Hern- ndez, Juan Nieto, and Eduardo Nebot. Large scale egomotion and error analysis with visual features. Journal of Physical Agents, 4:19–24, 2010. [4] Filiberto Chiabrando, Roberto Chiabrando, Dario Piatti, and Ful- Figure 6: Problems due to reflections. vio Rinaudo. Sensors for 3d imaging: Metric evaluation and calibration of a ccd/cmos time-of-flight camera. Sensors, described above. As future work, we plan to 9(12):10080–10096, 2009. work with images in order to improve their [5] Frank Dellaert, Dieter Fox, Wolfram Bur- quality. We also plan to continue exploring gard, and Sebastian Thrun. Monte carlo the response of the camera with computer vi- localization for mobile robots. In IEEE sion algorithms. International Conference on Robotics and Automation (ICRA99), May 1999. Acknowledgment [6] Stefan Fuchs and Gerd Hirzinger. Extrin- sic and depth calibration of tof-cameras. This work has been supported by grant In CVPR, 2008. DPI2009-07144 from Ministerio de Ciencia e Innovacion of the Spanish Government. [7] S. Burak Gokturk, Hakan Yalcin, and Cyrus Bamji. A time-of-flight depth sen- sor - system description, issues and so- References lutions. In CVPRW ’04: Proceedings of the 2004 Conference on Computer Vi- [1] SR4000 User Manual. sion and Pattern Recognition Workshop (CVPRW’04) Volume 3, page 35, Wash- C. Baker, Z. Omohundro, S. Thayer, and ington, DC, USA, 2004. IEEE Computer W. Whittaker. A system for volumet- Society. ric robotic mapping of abandoned mines. In Proc. IEEE International Conference [8] Chen-Tsung Kuo and Shyi-Chyi Cheng. on Robotics and Automation ICRA ’03, 3d model retrieval using principal plane volume 3, pages 4270–4275, 14–19 Sept. analysis and dynamic programming. Pat- 2003. tern Recogn., 40(2):742–755, 2007. [13] Sebastian Thrun, Dieter Fox, Wolfram [9] David G. Lowe. Distinctive image fea- Burgard, and Frank Dellaert. Ro- tures from scale-invariant keypoints. In- bust monte carlo localization for mo- ternational Journal of Computer Vision, bile robots. Artificial Intelligence, 128(1- 60:91–110, 2004. 2):99–141, 2001. [10] Gian Luca Mariottini and Domenico Prattichizzo. Image-based visual servoing [14] A. Tsalatsanis, K. Valavanis, and with central catadioptric cameras. Int. J. N. Tsourveloudis. Mobile robot naviga- Rob. Res., 27(1):41–56, 2008. tion using sonar and range measurements from uncalibrated cameras. In Proc. 14th [11] D. Schroter and M. Beetz. Acquir- Mediterranean Conference on Control ing models of rectangular 3d objects for and Automation MED ’06, pages 1–7, robot maps. In Proc. IEEE International June 2006. Conference on Robotics and Automation ICRA ’04, volume 4, pages 3759–3764, [15] J. Weingarten, G. Gruener, and R. Sieg- Apr 26–May 1, 2004. wart. A Fast and Robust 3D Feature Ex- traction Algorithm for Structured Envi- [12] S. Thrun, D. Hahnel, D. Ferguson, ronment Reconstruction. In None, 2003. M. Montemerlo, R. Triebel, W. Burgard,
J. William Helton, Orlando Merino (Author) - Classical Control Using H-Infinity Methods - Theory, Optimization and Design (Society For Industrial Mathematics, 1998,0898714192)