Autonomous Exploration of Unknown Environment by Fusion of Stereo Vision and Inertial Sensors
Autonomous Exploration of Unknown Environment by Fusion of Stereo Vision and Inertial Sensors
Disciplines
Electrical Engineering, Electronics, Information Engineering (20%); Computer Sciences (80%)
Keywords
-
Tracking,
Vision,
Augmented Reality,
Robotics
Tracking is required in many important applications, e.g. navigation, 3D human-computer-interaction, and surveillance. Most often, information has to be obtained in "real-time". Many kinds of sensors have been used for this purpose in the past: sonars, radar, laser range finders, GPS, electronic compass, so-called "magnetic trackers" in virtual reality, computer vision, transponders, just to mention the most common techniques. This project investigates a new, generic approach to real-time inside-out tracking (a system which is mounted to a mobile object with the main purpose to track the position and orientation of the object itself) using a combination of "smart sensors". We plan to use a sensor suite consisting of a fixed, calibrated stereo rig (with directly addressable multi-window CMOS cameras) together with an "inertio-tracker" based on accelerometers and gyroscopes. These two sensor types provide complimentary characteristics: visual sensing is very accurate at low velocities while inertial sensors can track fast motions but suffer from a drift particularly at low velocities. This inside-out tracking system will be subjected to arbitrary motion (on a mobile robot, carried by a person, ...) and shall operate in cluttered multi-object scenes with multiple and independent motion and some static, but 3D background. The primary goal is a reliable reconstruction of the trajectory of the system itself, as well as the recovery of 3D structure required for a successful tracking (salient landmarks). A further goal is to trace the trajectories of all detected other moving objects. Similar to the perceptive capabilities of a human, the system shall operate autonomously, without requiring additional information about its localisation and pose (like GPS or artificial landmarks). New CMOS camera technology will be used to gain an order of magnitude in optical tracking speed, and fusion of a suite of new accelerometers and gyroscopes with vision will provide robustness in cases of fast motions as well as the superior accuracy of vision-based tracking. The new tracking techniques will be applied to VR/AR applications at TU Graz and to mobile robot navigation at TU Wien. The goal of the proposed project is beyond previous work on "tracking" of 2D motion in images or 3D structure from motion captured by stationary observers. We aim at the continuous tracing of a moving object`s position and orientation in space and time, and we propose a way of achieving this goal based on sensory information captured by the moving platform itself.
The aim of this project was to research into new sensors and algorithms for online structure and motion analysis. More specifically, camera(s) are mounted on a sensor head and are used to measure visual landmarks in a scene. These vision-based measurements are complemented by inertial sensors. Finally, visual and inertial information is fused to obtain the current pose and the trajectory of the sensor head over time. To be applicable for online analysis, all algorithms should work in real-time, sensors should be sufficiently fast, and calculations should be incremental (as compared to off-line batch-processing, e.g. in photogrammetric bundle adjustment). The main results of the project can be split into two parts: The development of a new high-speed camera, and the development of new, incremental structure and motion algorithms. Once a landmark has been detected in an image, any motion of the camera will lead to a 2D displacement of the landmark in the image plane. The basic idea behind our new high-speed CMOS camera is to read only small windows from the image sensor, instead of whole frames. We have developed the necessary hard- and firmware to address arbitrary sub-windows of a CMOS imaging sensor in a fast and efficient manner. The new camera system works either in a monocular or in a stereo setup and has a standard USB-2 interface. We achieve "frame" rates of up to 2.500 windows per second (as compared to only 5 full frames per second), which allows us to track visual landmarks very fast and reliably. Our new structure and motion algorithm fuses the data from the CMOS sensor with the inertial data from the IMU (Inertial Measurement Unit), which typically operate at different and possibly varying rates. The system can then move into a new environment from a single known starting pose and simultaneously estimate the egomotion of the sensor rig and the unknown structure of the scene. Data fusion achieves the estimation of the egomotion of the sensor rig and takes into account cases such as the dynamics in the scene (occlusions, foreground motion), loss of correspondence of visual features, convergence (in case of non-linear models), fast camera motion (motion blur and especially rotations), and additionally estimates the 3D position of newly detected landmarks in the environment. Such sensors and algorithms have a high potential and might be applied in a number of relevant future systems, including mobile computing, location and pose aware communications, vehicle driver`s assistance, and autonomous navigation.
- Technische Universität Graz - 100%
Research Output
- 406 Citations
- 7 Publications
-
2006
Title Robust Pose Estimation from a Planar Target DOI 10.1109/tpami.2006.252 Type Journal Article Author Schweighofer G Journal IEEE Transactions on Pattern Analysis and Machine Intelligence Pages 2024-2030 -
2005
Title Finding Tables for Home Service Tasks and Safe Mobile Robot Navigation**This work is partially supported by the Austrian Science Foundation under grants P15748 “SmartTracking” and grants S9101-N04, S9103-N04 and S9107-N04 “Cognitive Vision”, and by t DOI 10.1109/robot.2005.1570576 Type Conference Proceeding Abstract Author Vogl R Pages 3035-3040 -
2005
Title Motion and Structure Estimation from Vision and Inertial Sensor Data with High Speed CMOS Camera**The project SmartTracking is funded by the Austrian Science Foundation #P15748. DOI 10.1109/robot.2005.1570383 Type Conference Proceeding Abstract Author Gemeiner P Pages 1853-1858 -
2008
Title On multi-rate fusion for non-linear sampled-data systems: Application to a 6D tracking system DOI 10.1016/j.robot.2007.11.009 Type Journal Article Author Armesto L Journal Robotics and Autonomous Systems Pages 706-715 -
2004
Title On the Design and Structure of Artificial Eyes for Tracking Tasks DOI 10.1109/icccyb.2004.1437689 Type Conference Proceeding Abstract Author Vincze M Pages 135-142 -
2004
Title A New High Speed CMOS Camera for Real-Time Tracking Applications DOI 10.1109/robot.2004.1302542 Type Conference Proceeding Abstract Author Muehlmann U Pages 5195-5200 -
2004
Title Multi-Rate Fusion with Vision and Inertial Sensors DOI 10.1109/robot.2004.1307150 Type Conference Proceeding Abstract Author Armesto L Pages 193-199