LiDAR-based 3D object detection is crucial for perception systems in autonomous driving. However, current methods perform poorly in detecting occluded and small objects due to the sparse and uneven distribution of point clouds in 3D scenes. To address this problem, we propose a long-range and multi-scale information interaction network (LMIINet) for 3D object detection. First, a feature extraction backbone with a spatial feature pyramid block is designed to effectively capture long-range dependencies between features. Then, multi-scale spatial features from the 3D backbone are adaptively fused at the neck to aggregate local and global contextual information. Finally, by fully utilizing the correlations between bounding box parameters, the proposed rotation-decoupled corrected intersection-over-union (RCIoU) loss is employed for the classification and regression supervision of bounding boxes to improve the one-stage point cloud detectors. LMIINet achieves competitive performance on small objects (pedestrians and cyclists) detection in the KITTI dataset. Compared with the benchmark network SECOND, LMIINet increases |
ACCESS THE FULL ARTICLE
No SPIE Account? Create one