Abstract
Point cloud-based place recognition can be used for global localization in large-scale scenes and loop-closure detection in simultaneous localization and mapping (SLAM) systems in the absence of GPS. Current learning-based approaches aim to extract global and local features from 3D point clouds to encode them as descriptors for point cloud retrieval. The key problems are that the occlusion of point clouds by dynamic objects in the scene affects the point cloud structure, a single perceptual field of the network cannot adequately extract point cloud features, and the correlation between features is not fully utilized. To overcome this, we propose a novel network called CapsLoc3D. We first use the static point cloud generation module to remove the occlusion effects of dynamic objects, and then obtain the point cloud descriptors by processing with the CapsLoc3D network which contains the point spatial transformation module, multi-scale feature fusion module, Capsnet module and a GeM Pooling layer. After validation using the Oxford RobotCar, KITTI, and NEU datasets, experiments show that our method performs better and also has good generalization performance and computational efficiency compared with current state-of-the-art algorithms.
Original language | English |
---|---|
Pages (from-to) | 6811-6823 |
Number of pages | 13 |
Journal | IEEE Transactions on Intelligent Transportation Systems |
Volume | 25 |
Issue number | 7 |
Early online date | 16 Jan 2024 |
DOIs | |
Publication status | Published (in print/issue) - Jul 2024 |
Bibliographical note
Publisher Copyright:IEEE
Keywords
- Lidar place recognition
- moving object segmentation
- capsule network
- global localization
- place feature learning
- Point cloud compression
- Laser radar
- Three-dimensional displays
- Semantics
- Feature extraction
- Task analysis
- Vehicle dynamics