Content area
Abstract
We propose GPRNet, a geometry-aware semantic segmentation framework that integrates a Geometric Prior-Refined Block (GPRB) and a Mutual Calibrated Fusion Module (MCFM) to enhance boundary sensitivity and cross-stage semantic consistency. GPRB leverages learnable directional derivatives to construct structure-aware strength and orientation maps, enabling more accurate spatial localization in complex scenes. MCFM introduces geometric alignment and semantic enhancement mechanisms that effectively reduce the encoder–decoder feature gap. GPRNet achieves consistent performance gains on ISPRS Potsdam and LoveDA, improving mIoU by up to 1.7% and 1.3% respectively over strong CNN-, attention-, and transformer-based baselines.
Incorporating geometric priors through learnable gradient-based features improves the model’s ability to capture structural patterns and preserve fine boundaries in high-resolution remote sensing imagery. The mutual calibration mechanism demonstrates an effective design for encoder–decoder interaction, showing potential for broader applicability across segmentation architectures and modalities. The empirical evidence indicates that geometry-informed representation learning can serve as a general principle for enhancing land-cover mapping in diverse and structurally complex environments. Semantic segmentation of high-resolution remote sensing images remains a challenging task due to the intricate spatial structures, scale variability, and semantic ambiguity among ground objects. Moreover, the reliable delineation of fine-grained boundaries continues to impose difficulties on existing CNN- and transformer-based models, particularly in heterogeneous urban and rural environments. In this study, we propose GPRNet, a novel geometry-aware segmentation framework that leverages geometric priors and cross-stage semantic alignment for more precise land-cover classification. Central to our approach is the Geometric Prior-Refined Block (GPRB), which learns directional derivative filters, initialized with Sobel-like operators, to generate edge-aware strength and orientation maps that explicitly encode structural cues. These maps are used to guide structure-aware attention modulation, enabling refined spatial localization. Additionally, we introduce the Mutual Calibrated Fusion Module (MCFM) to mitigate the semantic gap between encoder and decoder features by incorporating cross-stage geometric alignment and semantic enhancement mechanisms. Extensive experiments conducted on the ISPRS Potsdam and LoveDA datasets validate the effectiveness of the proposed method, with GPRNet achieving improvements of up to 1.7% mIoU on Potsdam and 1.3% mIoU on LoveDA over strong recent baselines. Furthermore, the model maintains competitive inference efficiency, suggesting a favorable balance between accuracy and computational cost. These results demonstrate the promising potential of geometric-prior integration and mutual calibration in advancing semantic segmentation in complex environments.
Details
Deep learning;
Image resolution;
Calibration;
Artificial neural networks;
Boundaries;
Biodiversity;
Remote sensing;
Mapping;
Attention;
Architecture;
Image processing;
Spatial discrimination;
Land use;
Semantic segmentation;
Modules;
Localization;
Rural environments;
Maps;
Machine learning;
Land cover;
Coders;
Alignment;
Image segmentation;
Operators (mathematics);
High resolution;
Effectiveness;
Urban environments;
Semantics
1 College of Hydrology and Water Resources, Hohai University, Nanjing 210098, China; [email protected]
2 College of Computer Science and Software Engineering, Hohai University, Nanjing 211100, China; [email protected], Key Laboratory of Water Big Data Technology of Ministry of Water Resources, Hohai University, Nanjing 211100, China
3 Information Center, Ministry of Water Resources, Beijing 100053, China; [email protected]
4 School of Design and Art, Changsha University of Science and Technology, Changsha 410114, China; [email protected]
5 College of Computer and Information Engineering, Xinxiang University, Xinxiang 453000, China; [email protected]
6 Information Center, Yellow River Conservancy Commission (YRCC), Zhengzhou 450000, China; [email protected] (L.C.); [email protected] (D.L.)