Cyrill Stachniss, John J. Leonard and Sebastian Thrun
This chapter provides a comprehensive introduction in to the simultaneous localization and mapping problem, better known in its abbreviated form as SLAM. SLAM addresses the main perception problem of a robot navigating an unknown environment. While navigating the environment, the robot seeks to acquire a map thereof, and at the same time it wishes to localize itself using its map. The use of SLAM problems can be motivated in two different ways: one might be interested in detailed environment models, or one might seek to maintain an accurate sense of a mobile robot’s location. SLAM serves both of these purposes.
We review the three major paradigms from which many published methods for SLAM are derived: (1) the extended Kalman filter (EKF); (2) particle filtering; and (3) graph optimization. We also review recent work in three-dimensional (3-D) SLAM using visual and red green blue distance-sensors (RGB-D), and close with a discussion of open research problems in robotic mapping.
Deformation-based loop closure for dense RGB-D SLAM
Author Thomas Whelan
Video ID : 439
This video shows the integration of SLAM-pose-graph optimization, spatially extended KinectFusion, and deformation-based loop closure in dense RGB-D mapping - integrating several of the capabilities discussed in Chap. 46.3.3 and Chap. 46.4, Springer Handbook of Robotics, 2nd edn (2016). Reference:
T. Whelan, M. Kaess, H. Johannsson, M. Fallon, J.J. Leonard, J. McDonald: Real-time large scale dense RGB-D SLAM with volumetric fusion, Int. J. Robot. Res. 34(4-5), 598-626 (2014).