Content area
Cross-modal semantic segmentation of individual tree LiDAR point clouds is critical for accurately characterizing tree attributes, quantifying ecological interactions, and estimating carbon storage. However, in forest environments, this task faces key challenges such as high annotation costs and poor cross-domain generalization. To address these issues, this study proposes a cross-modal semantic transfer framework tailored for individual tree point cloud segmentation in forested scenes. Leveraging co-registered UAV-acquired RGB imagery and LiDAR data, we construct a technical pipeline of “2D semantic inference—3D spatial mapping—cross-modal fusion” to enable annotation-free semantic parsing of 3D individual trees. Specifically, we first introduce a novel Multi-Source Feature Fusion Network (MSFFNet) to achieve accurate instance-level segmentation of individual trees in the 2D image domain. Subsequently, we develop a hierarchical two-stage registration strategy to effectively align dense matched point clouds (MPC) generated from UAV imagery with LiDAR point clouds. On this basis, we propose a probabilistic cross-modal semantic transfer model that builds a semantic probability field through multi-view projection and the expectation–maximization algorithm. By integrating geometric features and semantic confidence, the model establishes semantic correspondences between 2D pixels and 3D points, thereby achieving spatially consistent semantic label mapping. This facilitates the transfer of semantic annotations from the 2D image domain to the 3D point cloud domain. The proposed method is evaluated on two forest datasets. The results demonstrate that the proposed individual tree instance segmentation approach achieves the highest performance, with an IoU of 87.60%, compared to state-of-the-art methods such as Mask R-CNN, SOLOV2, and Mask2Former. Furthermore, the cross-modal semantic label transfer framework significantly outperforms existing mainstream methods in individual tree point cloud semantic segmentation across complex forest scenarios.
Details
Datasets;
Deep learning;
Forestry;
Lidar;
Artificial neural networks;
Image annotation;
Mapping;
Data processing;
Unmanned aerial vehicles;
Semantic segmentation;
Trees;
Performance evaluation;
Statistical analysis;
Forests;
Image segmentation;
Carbon sequestration;
Three dimensional models;
Instance segmentation;
Image acquisition;
Methods;
Algorithms;
Semantics
; Chen, Ting 2 ; Zhang, Tao 1 ; Yang Minglu 1 ; Ye, Yuan 3 ; Liu Jiahao 4 1 School of Surveying and Geoinformation Engineering, East China University of Technology, Nanchang 330013, China; [email protected] (F.Z.); [email protected] (T.Z.); [email protected] (M.Y.), Jiangxi Key Laboratory of Watershed Ecological Process and Information, East China University of Technology, Nanchang 330013, China
2 School of Water Resources and Environmental Engineering, East China University of Technology, Nanchang 330013, China; [email protected]
3 Shenzhen DJI Innovations Technology Co., Ltd., Shenzhen 518057, China; [email protected]
4 Jiangxi Helicopter Co., Ltd., Jingdezhen 333036, China; [email protected]