Content area
Point cloud registration is a fundamental problem in computer vision and 3D computing, aiming to align point cloud data from different sensors or viewpoints into a unified coordinate system. In recent years, the rapid development of RGB-D sensor technology has greatly facilitated the acquisition of RGB-D data. In previous unsupervised point cloud registration methods based on RGB-D data, there has often been an overemphasis on matching local features, while the potential value of global information has been overlooked, thus limiting the improvement in registration performance. To address this issue, this paper proposes a self-attention-based global information attention module, which learns the global context of fused RGB-D features and effectively integrates global information into each individual feature. Furthermore, this paper introduces alternating self-attention and cross-attention layers, enabling the final feature fusion to achieve a broader global receptive field, thereby facilitating more precise matching relationships. We conduct extensive experiments on the ScanNet and 3DMatch datasets, and the results show that, compared to the previous state-of-the-art methods, our approach reduces the average rotation error by 26.9% and 32% on the ScanNet and 3DMatch datasets, respectively. Our method also achieves state-of-the-art performance on other key metrics.
Details
; Han, Zhenqi 2
; Liu, Lizhaung 2
; Zhang, Jialu 2 1 Shanghai Advanced Research Institute, Chinese Academy of Sciences, Shanghai 201210, China;
2 Shanghai Advanced Research Institute, Chinese Academy of Sciences, Shanghai 201210, China;