ENCODE: a dEep poiNt Cloud ODometry nEtwork
Abstract
Ego-motion estimation is a key requirement for the simultaneous localization and mapping (SLAM) problem. The traditional pipeline goes through feature extraction, feature matching and pose estimation, whose performance depends on the manually designed features. In this paper, we are motivated by the strong performance of deep learning methods in other computer vision and robotics tasks. We replace hand-crafted features with a neural network and directly estimate the relative pose between two adjacent scans from a LiDAR sensor using ENCODE: a dEep poiNt Cloud ODometry nEtwork. Firstly, a spherical projection of the input point cloud is performed to acquire a multi-channel vertex map. Then a multi-layer network backbone is applied to learn the abstracted features and a fully connected layer is adopted to estimate the 6-DoF ego-motion. Additionally, a map-to-map optimization module is applied to update the local poses and output a smooth map. Experiments on multiple datasets demonstrate that the proposed method achieves the best performance in comparison to state-of-the-art methods and is capable of providing accurate poses with low drift in various kinds of scenarios.
BibTeX
@conference{Zhang-2021-133750,author = {Yihuan Zhang and Liang Wang and Chen Fu and Yifan Dai and John M. Dolan},
title = {ENCODE: a dEep poiNt Cloud ODometry nEtwork},
booktitle = {Proceedings of (ICRA) International Conference on Robotics and Automation},
year = {2021},
month = {June},
pages = {14375 - 14381},
keywords = {SLAM, egomotion estimation, LIDAR, mapping},
}