1. Introduction
Land cover classification provides technical support for land planning and management, land change mechanism analysis, and environmental protection. With its macroscopic, dynamic, and rapid characteristics, remote sensing technology has become the most effective means of obtaining land use information [1, 2]. The automatic classification of land cover and thematic information extraction using satellite remote sensing data has long been at the forefront of remote sensing technology applications [3]. In recent years, more studies have utilized high-resolution remote sensing data to achieve automatic land cover classification, yielding significant results [4–6]. Traditional unsupervised remote sensing image classification methods are relatively easy to implement and have lower computational complexity. These methods perform well when pixel clusters exhibit a simple probability distribution in spectral space and point clusters in remote sensing images have convex geometric shapes [7]. Consequently, traditional unsupervised methods often rely on high-resolution data for remote sensing image classification. However, these methods are primarily limited to extracting and analyzing shape and texture features, resulting in constrained classification accuracy. When there are significant differences in pixel counts between clusters or when the pixel sets comprising these clusters do not follow a Gaussian distribution, classification performance deteriorates noticeably [8–10]. In such cases, shape features perform poorly in complex terrains, and relying solely on texture features fails to capture surface objects’ spectral information adequately [11–13]. As a result, multi-feature fusion methods that integrate shape, texture, and spectral information have increasingly become a research focus [14]. These methods effectively utilize shape features to capture the geometric structure of land objects, texture features to represent surface roughness, and spectral features to provide information about physical and chemical composition [15–17]. By leveraging these combined features, classification accuracy and robustness can be significantly enhanced, addressing the limitations of single-feature classification methods [18].
In remote sensing image classification, unsupervised classification methods offer significant advantages over supervised methods. Unsupervised classification does not require pre-labeled training data, which is particularly important when labeled data is expensive or difficult to obtain [19]. Additionally, unsupervised methods can adaptively discover underlying structures and patterns in the data, making them suitable for complex and unknown data distributions. Among various unsupervised classification methods, spectral clustering models demonstrate unique advantages. Spectral clustering constructs a data similarity matrix and applies spectral graph theory to perform classification. It can identify non-convex clusters and does not require assumptions about the global structure of the data [20]. Compared to traditional clustering algorithms such as the Iterative Self-Organizing Data Analysis Technique (ISODATA) and k-means, spectral clustering exhibits greater adaptability to data distribution. It has proven effective in high-resolution remote sensing image processing. Traditional methods like k-means are generally more effective for spherical clusters, whereas spectral clustering excels at handling non-convex clusters. By relying on the global similarity structure rather than the distribution of locally adjacent points, spectral clustering demonstrates high robustness to noise and outliers. Consequently, spectral clustering has become a significant research topic in the field of machine learning [21].
In recent years, numerous scholars have conducted extensive research on the theory and application of spectral clustering, yielding a series of significant results. Theoretically, researchers have proposed several improved spectral clustering algorithms to address the limitations of traditional spectral clustering. Notable contributions include Recursive Spectral Clustering [22], Class-Specific Spectral Clustering [23], Fuzzy Spectral Clustering [24, 25], Mean Shift Spectral Clustering [26], Efficient Evolutionary Spectral Clustering [27], Sparse Kernel Spectral Clustering [28], Fast Kernel Spectral Clustering [29], Non-negative Sparse Spectral Clustering [30], Vector Quantization-based Approximate Spectral Clustering [31], and Compressed Constraint Spectral Clustering [32]. In terms of applications, spectral clustering has been successfully applied to various fields, including face recognition [33], image segmentation [34], big data analysis and processing [35], medical image analysis [36], information retrieval [37], power system modeling [38], protein data analysis [39], and disaster warning systems [40].
Although significant progress has been made in the theory and application of spectral clustering, there has been relatively little research on the unsupervised classification of remote sensing images using spectral clustering. Moreover, most improved Spectral Clustering (SC) algorithms rely on distance metrics such as Euclidean distance, cosine distance, or Gaussian kernel distance, which often fail to capture the topological structure of the data. This study introduces the adaptive spectral clustering algorithm combining the shared proximity and flow distance (SNN-MSC) approach to the field of remote sensing image processing, using high-resolution remote sensing images as the data source. By integrating spectral, edge shape, and texture information, and replacing traditional distance metrics used in spectral clustering with manifold distance, the density factor is incorporated to impose additional constraints on the similarity matrix. This optimization enhances the spectral clustering algorithm for unsupervised remote sensing image classification, enabling effective land cover classification of high-resolution remote sensing imagery. Additionally, through research on automatic land cover classification, this study tests the applicability of spectral clustering algorithms to large-scale data clustering problems, aiming to address the limitations of current algorithms and further develop the theoretical foundations of spectral clustering.
2. Materials and methods
2.1 Study area and data resource
The study area is located within Songyuan City, in the central-western region of Jilin Province, China. It lies within the Harbin-Changchun-Daqing triangle, at the core of the Ha-Chang urban agglomeration, and serves as a crucial transportation hub and logistics distribution center in Northeast China (Fig 1). Situated in the mid-latitude region of the Northern Hemisphere, the study area experiences a temperate continental monsoon climate, characterized primarily by plains. The average annual temperature is approximately 4.5°C, with total annual precipitation ranging from 400 mm to 500 mm. The study area in Antu County is located in southwest of Yanbian Prefecture, Jilin Province, China. It is at the northern foot of Changbai Mountain, where landform is mainly mountains, with warm climate, abundant rainfall and resources. It was a famous hometown of Chinese mineral water and a breeding base for Chinese herbal medicines. In this study, Gaofen-2 (GF-2) satellite images obtained on May 21, 2017 are used (https://www.cresda.com, Authorized). GF-2 is the first satellite launched by Chinese high resolution Earth Observation System. It is equipped with 1 m panchromatic and 4 m multispectral cameras to achieve imaging, and has the meter-level spatial resolution, high radiation and positioning accuracy, fast revisiting cycle. GF-2 images have been widely applied in land resource survey and monitoring. A typical area with rich land-cover types in the study area is selected, corresponding to the 400×400 pixel range of the image. The land-cover types in the area mainly include cropland, forest, grassland, building and water, which can effectively test the classification method. This study utilized ZY1-02D satellite images acquired on May 10, 2023. The ZY1-02D satellite, also known as Resource One 02D, is equipped with both visible-near infrared and hyperspectral cameras, enabling the simultaneous acquisition of detailed texture and rich spectral information of land features (https://www.sasclouds.com, Authorized). Compared to other satellites, the ZY1-02D offers a more comprehensive description of land characteristics and a broader single-pass coverage, making it particularly advantageous for urban and rural planning as well as resource monitoring. A typical area within the study region, characterized by diverse land cover types, was selected, corresponding to an image range of 400 × 400 pixels. The land cover types in this area include water bodies, croplands, forests, buildings, and transportation infrastructure, which are well-suited for evaluating the classification methods.
[Figure omitted. See PDF.]
2.2 SNN-MSC algorithm
Spectral clustering is a prominent deep-learning algorithm based on graph theory, a branch of mathematics focused on the study of graphs. Graph partitioning theory treats different sample points as vertices of a graph, connects each pair of vertices with an edge, and constructs a spectral graph according to specific rules. Similarity between samples is used to assign weights to the edges, resulting in a weighted undirected graph based on sample similarity, thus transforming the clustering problem into a graph partitioning problem [41, 42]. The construction of the similarity matrix directly impacts the accuracy of spectral clustering (SC) algorithms [43]. Traditional SC algorithms often use distance metrics such as Euclidean distance, cosine distance, or Gaussian kernel distance. These metrics typically fail to capture the topological structure of the data, leading to neglect of global consistency and insufficient capture of the data’s intrinsic structure, which results in suboptimal clustering outcomes [44]. The SNN-MSC algorithm introduces a novel manifold distance with exponential terms and proportional factors. By adjusting these terms, the algorithm modifies the similarity ratio between data points within the same manifold and those across different manifolds, thus preserving global and local data distribution consistency. Additionally, the algorithm incorporates a density factor to mitigate noise effects and computes similarity based on the sparsity and density of data neighborhoods, enhancing the neighborhood information between points. Rank constraints are applied to the Laplacian to ensure that the number of connected components in the similarity matrix equals the number of clusters.
First, by integrating the k-nearest neighbors of the samples with the mean level of the dataset’s neighborhood information, the density factor is defined as follows:(1)
Where ρi denotes the density factor of sample xi, with larger values indicating higher local density. ω represents the weight, which is set to 1 in this study. k signifies the number of nearest neighbors, and davg(xi) is the average Euclidean distance between sample xi and its k nearest neighbors. A smaller davg(xi) value implies that sample xi is closer to its k nearest neighbors, indicating a higher local density. ξ represents the mean level of overall neighborhood information within the dataset. KNN(xi) is the set of k nearest neighbors of sample xi, and θ denotes the distance to the k-th nearest neighbor from sample xi.
The distance between samples is calculated by incorporating the k-nearest neighbor information from the original space to compute the density factor, which is then integrated into the manifold distance metric. This redefines the distance between samples, as shown in Eq (2):(2)
Where denotes the distance between samples xi and xj; θ>0 represents the scaling factor, where a smaller θ value makes the corrected distance metric more inclined towards local consistency, while a larger θ value favors global consistency. p denotes a path of length l=|p|−1, |p| represents the length of the path p connecting samples xi and xj, and (pa,pa+1)∈E represents the short edge formed by two adjacent points pa and pa+1 on the path p. Pij indicates the set of all paths connecting samples xi and xj, d(pa,pa+1) denotes the Euclidean distance between any two adjacent nodes on the path, and is the minimum path distance between samples xi and xj on graph G. Finally, ρa>0 and pa+1>0 represent the density factors of two adjacent nodes pa and Pa+1, respectively.
In the same manifold, sample points are typically connected by multiple short edges, whereas sample points from different manifolds require longer edges for connection. In this study, the manifold distance is used in place of the traditional Euclidean distance d(pa,pa+1) to measure the distance between adjacent points. Additionally, the density factor is calculated by incorporating the neighborhood information of the samples, which helps to mitigate the impact of noise.
The regularization parameter is calculated using the following equation:(3)
The mean value of r1,r2,⋯rn is selected:(4)
Using the number of shared neighbors and distances, combined with local scale information, the similarity simij between sample xi and sample xj is redefined in the form of an exponential kernel, as follows:(5)
Where SNN(xi,xj) represents the intersection of the k-nearest neighbor sets of samples xi and xj, i.e., SNN(xi,xj) = KNN′(xi)∩KNN′(xj). Additionally, denotes the distance from sample xi to its k-th nearest neighbor. The quantity |SNN(xi,xj)| indicates the number of shared neighbors between samples xi and xj; a higher value reflects a greater number of shared neighbors and thus a higher similarity. σi and σj represent the local scales of samples xi and xj, respectively, with values corresponding to the distance to the k-th nearest neighbor. Smaller values indicate higher local density, and the local scale can be adaptively adjusted based on neighborhood information, increasing the similarity between samples in sparse clusters to facilitate their aggregation and mitigate the limitations of global scale. The term represents the sum of the squared distances of samples xi and xj to the points in the shared neighbor set. Smaller values suggest that the two samples are relatively close, indicating higher similarity. In summary, a higher value of simij signifies a higher similarity between samples xi and xj.
The calculated similarity matrix is not a normalized matrix; it needs to be normalized to obtain S. The formula is as follows:(6)
It is commonly assumed that samples that are closer to each other have greater similarity. At the same time, to avoid trivial solutions, where only the closest samples are considered as neighbor points, a regularization term is introduced. If the similarity matrix ss is non-negative, the number of eigenvalues equal to zero in the corresponding Laplacian matrix corresponds to the number of connected components in the undirected graph G. Therefore, this paper imposes a rank constraint on the Laplacian matrix, with the objective function presented in Eq (7):(7)
2.3 Feature extraction
2.2.1 Image spectral feature extraction.
Due to the presence of striping artifacts in the shortwave infrared (SWIR) bands of the ZY1-02D satellite hyperspectral data, a "global de-striping" method was employed to correct these artifacts. Additionally, atmospheric water vapor absorption affects wavelengths in the 1350–1420 nm and 1820–1920 nm ranges, leading to insufficient imaging quality of the spectral data. Therefore, spectral channels within these ranges were excluded, and the remaining 153 spectral bands were combined and stored. Radiometric calibration was performed on the remaining bands to correct for errors caused by sensor noise, solar position, and angle variations. The FLASH model was then used for atmospheric correction to obtain surface reflectance. Considering the issues of hyperspectral data mixing and the difference in spatial resolution compared to panchromatic data, the extracted feature data were resampled to a uniform resolution of 10 meters. Subsequently, geometric correction was applied to register the spatial positions of the data, ensuring classification accuracy (Fig 2).
[Figure omitted. See PDF.]
During the electromagnetic radiation transmission and acquisition process, the total radiance measured by the sensor is affected by atmospheric molecules, aerosol scattering, and water vapor absorption, which impairs the ability of the sensor to reflect the true spectral characteristics of surface objects. This significantly impacts the clarity and contrast of the images, severely hindering the application of remote sensing data. Therefore, reducing or eliminating atmospheric noise and enhancing image features to recover and obtain accurate surface spectral reflectance data has been a key focus for researchers domestically and internationally. To assess the usability of ZY1-02D hyperspectral images, the corrected spectral image was overlaid with the current land use map. Endmember spectra for forested areas and water bodies, with 50 samples each, were extracted and compared with field-measured spectral data. The correlation was analyzed using Pearson correlation coefficients to evaluate the calibration effectiveness of the hyperspectral images. As illustrated, the spectral reflectance curves of the pixels and the field-measured spectra generally show similar spectral shapes and characteristic absorption features. Most samples had Pearson correlation coefficients between 0.8 and 0.9. Waterbody correlations were generally higher, while vegetation correlation curves showed more variability, though the average still exceeded 0.8. This indicates a high degree of match, suggesting that the image pixels retain most of the surface’s spectral features and are suitable for rapid urban land use classification. Urban construction land and water bodies exhibit significant differences in spectral reflectance, facilitating effective differentiation of land cover types through spectral information. The SNN-MSC algorithm improves distance metrics by considering both global and local consistency. It achieves this by adjusting the exponent and scaling factor to simultaneously satisfy global and local consistency, thereby better uncovering the intrinsic structure of the data and ensuring the reliability of the final clustering results.
2.2.2 Texture and edge feature extraction.
Texture is related to the spatial distribution of intensity values in an image, reflecting surface variation, structure, and organization properties. Texture features are categorized into four types: statistical, model-based, signal processing, and structural. Among these, the Gray-Level Co-occurrence Matrix (GLCM) is widely used in texture feature classification [45]. Haralick et al. proposed 14 texture features using GLCM, which primarily involve statistical analysis of the spatial correlation of image gray levels to compute texture. This study selects eight commonly used features as texture feature data sources: Mean, Variance, Homogeneity, Contrast, Dissimilarity, Entropy, Angular Second Moment, and Correlation.
2.2.3 Spectral feature selection.
The Successive Projections Algorithm (SPA), introduced by Bregman in 1965, is a pre-variable selection technique that utilizes vector projection analysis to select the most significant vectors and subsequently extract a few characteristic wavelengths through model calibration [46]. The advantages of SPA lie in its ability to select variable combinations with minimal collinearity from the spectral matrix, thereby reducing model redundancy and enhancing both the stability and accuracy of the model. The specific steps are as follows:
xk(0) denotes the initial iteration vector; N denotes the number of variables to be extracted; J denotes the number of columns of the spectral matrix. Randomly pick the j-th column in the spectral matrix and assign it to xj denoted as xk(0); the set of the remaining column vectors is denoted as s. Calculate the projection Pxj of xj on the remaining column vectors, respectively. Extract the spectral wavelength k(n) of the largest projection vector, such that xj = px,j∈s, n = n+1, if n<N, is calculated cyclically according to the following equation.
(8)(9)(10)
Finally, the extracted variables are {xk(n) = 0,⋯,N−1}, respectively, k(0) and N in each cycle of the multiple linear regression model, through the interactive validation of a root-mean-square error value, according to its corresponding candidate subset, select the smallest RMSE value corresponding to the k(0) and N determined as the final optimal value.
2.4 Algorithm modeling
After the initial modeling and evaluation, parameter optimization of the model is usually required to improve the classification accuracy and model stability. The bandwidth parameters of the Gaussian kernel function are first adjusted to optimize the similarity matrix construction. Moreover, feature selection is carried out to improve the classification effect of the model through the best combination of features. Finally, methods such as the Silhouette Coefficient or Elbow Method are used to determine the optimal number of clusters (k) [47, 48].
In addition, k-fold cross-validation is used to assess the robustness and generalization ability of the model [49]. The dataset is divided into k subsets, and one of the subsets is used as the validation set each time, and the remaining k-1 subsets are used as the training set, and repeated k times. Calculate the evaluation metrics for each validation and take the average as the final performance of the model (Fig 3).
[Figure omitted. See PDF.]
2.5 Evaluation index
After performing land cover classification of high-resolution remote sensing images using the spectral clustering algorithm, it is crucial to evaluate the quality of the classification results. The following are two commonly used evaluation metrics:
Overall accuracy measures the overall correctness of the classification model and is calculated as follows [50]:(11)
Where, nii represents the diagonal elements of the confusion matrix, indicating the number of pixels of the i-th class that are correctly classified. N is the total number of pixels. Overall accuracy reflects the classification accuracy of all categories, but it does not reflect the specific performance of each category.
The Kappa coefficient is used to measure the consistency of the classification results relative to the results of random assignment [51], taking into account the correct classifications that occur by chance. The formula for calculation is:(12)
Where, po represents the actual agreement, which is the proportion of correct classifications; pe represents the expected agreement, which is the expected proportion of random classifications. The value of the Kappa coefficient ranges from -1 to 1, where 1 indicates perfect agreement, 0 indicates random agreement, and negative values indicate that the classification results are worse.
(13)(14)(15)
Where, SE(Kappa) represents the standard error, Φ(|Z|) denotes the cumulative distribution function (CDF) of the standard normal distribution for a two-tailed test, and Z refers to the test statistic.
3. Results and discussion
3.1 Feature extraction
Due to the complexity of land cover types in the area, in addition to extracting the spectral information from high-resolution remote sensing imagery, classification is also performed using its shape and texture information, which can address the phenomenon of "same object, different spectra". In remote sensing imagery, there are clear boundaries between different types of land features. Utilizing edge shape information can enhance the classification accuracy of remote sensing images. In this study, Laplacian high-pass filtering is used for filtering treatment, generating images with grayscale ranging from 0 to 255, with filter sizes of 3x3 and 7x7, respectively. The images processed by the Laplacian algorithm are shown in Fig 4. Edge information is extracted by setting thresholds through histogram analysis, generating binary images of edge density.
[Figure omitted. See PDF.]
Laplacian filtered images: (a) the size of filter is 3*3 (b) the size of filter is 7*7.
The Gray-Level Co-occurrence Matrix (GLCM) proposed by Haralick is a classic in statistical methods (Haralick et al., 1973). The GLCM method calculates the occurrence frequency of each gray level at a specified direction and distance, generating the corresponding GLCM. Then, second-order statistical feature values are computed as texture measures to describe the image. The GLCM defines the calculation methods for different texture features. By combining spectral features, texture features are extracted using variance, uniformity, entropy, and second-order moments (Fig 5).
[Figure omitted. See PDF.]
Image texture features: (a) variance (b) homogeneity (c) second moment (d) entropy.
3.2 Classification and accuracy verification
To validate the applicability of the spectral clustering optimized by the Lanczos algorithm in the classification of high-resolution remote sensing imagery, the images of the study area were classified using KNN, SC, and SNN-MSC algorithms, and a land cover map of the study area was produced. In this study, the parameter c in the K-means algorithm represents the actual number of clusters, while the parameter c in the SC algorithm denotes the number of clusters. Extensive experiments have demonstrated that the SNN-MSC algorithm performs optimally when the parameter k is set between 3 and 20, and the parameter θ is set between 1 and 10.
Additionally, classification experiments were conducted considering only spectral information, edge shape, and texture information, which were randomly divided into training groups (model establishment and parameter optimization) and validation groups (model accuracy and generalization ability) in a 3:1 ratio, testing the effect of multi-source information on high-resolution image classification. The results are shown in Fig 6. As shown in Fig 6, both the KNN and SC algorithms, when considering only spectral information, exhibit some missing and incorrect classifications, particularly in areas with sharp grayscale changes (such as the boundary between vegetation and water bodies). However, when edge shape and texture information are incorporated, the model’s classification performance shows a significant improvement. According to the recognition results in Fig 6(E) and 6(F), SNN-MSC outperforms the other two algorithms when considering only spectral information and when using multi-source information. The performance improvement is especially noticeable in the case of multi-source information-based quantitative classification. This indicates that the SNN-MSC algorithm, when considering multi-source information, can effectively identify boundaries between linear objects and different textures, providing a more precise delineation of the boundary between vegetation and water bodies compared to the other two methods. Additionally, for land cover types with similar spectral and texture characteristics, such as construction land and transportation land, SNN-MSC achieves higher accuracy, significantly reducing classification errors and omissions. Therefore, SNN-MSC exhibits more robust adaptability to data distribution and better clustering performance.
[Figure omitted. See PDF.]
Land cover classification results: (a) ISODATA, (b) mutli-source ISODATA, (c) k-means, (d) multi-source k-means, (e) spectral clustering, and (f) multi-source spectral clustering.
3.3 Performance evaluation
This study utilizes a confusion matrix to analyze the accuracy of classification results. Using ZY1-02D imagery as validation data, a spatially balanced sampling method was applied to randomly select 300–500 sample points for each land cover type. The accuracy of the aforementioned methods is assessed by calculating the classification confusion matrix to obtain overall accuracy, user’s accuracy, producer’s accuracy (PA), and the Kappa coefficient. The classification results are statistically presented in Table 1, and a comparison of classification accuracy indicators is shown in Table 2. As illustrated in Table 1, the overall classification accuracy for water bodies and residential areas is relatively high, while the classification accuracy for forests is generally not high. When classifying using pure spectral information, shape information is utilized to delineate the boundaries of land features. After incorporating shape information, the classification accuracy of the three algorithms has significantly improved, with an average increase of 13.89%. As shown in Table 2, the SNN-MSC algorithm exhibits higher classification accuracy for different land cover types compared to the previous KNN and SC algorithms. The overall classification accuracy is 84.63%, and the Kappa coefficient is 0.846, validating the efficiency and superiority of spectral clustering optimized by the Lanczos algorithm in the classification of high-resolution remote sensing imagery. The classification accuracy of multi-source information combined with shape and texture is significantly higher than that of spectral information alone. The Kappa coefficient has improved by 7.7%, 19.11%, and 14.93%, respectively.
[Figure omitted. See PDF.]
[Figure omitted. See PDF.]
The comparison of Producer’s Accuracy (PA) among several experiments is depicted in Fig 7. The figure indicates that the SNN-MSC with multi-source information achieved a higher Kappa coefficient than other methods. Notably, when using pure spectral classification, the classification accuracy for transportation land is relatively low. However, after incorporating shape and texture features, the accuracy for transportation land has significantly improved compared to other land use types, reaching an increase of 14.42%. The significance test results indicate that the p-values of SNN-MSC under both single data source and multi-source information conditions are less than 0.05, demonstrating that it has passed the significance test. This finding suggests that the Kappa coefficient is statistically significant, allowing for rejecting the null hypothesis and indicating that SNN-MSC exhibits a high level of consistency under both conditions.
[Figure omitted. See PDF.]
To visually demonstrate the superiority of SNN-MSC in clustering tasks, the t-SNE technique was employed for feature visualization. Fig 8 presents the visualization results of the features learned in the last layer mapped to a two-dimensional space. From Fig 8, it can be observed that there are clear boundaries in the feature space of the five land classes, with a small number of feature points being confused between different classes, and a phenomenon of feature homology is exhibited in the forest and transportation land. This is consistent with the analysis results in Table 1 and Fig 6, providing corroboration for the high performance of SNN-MSC.
[Figure omitted. See PDF.]
3.4 Complexity analysis of the SNN-MSC algorithm
This section provides a complexity analysis of the SNN-MSC algorithm, where n denotes the number of samples, d represents the sample dimensionality, and k is the number of nearest neighbors. The detailed analysis process is as follows:
1. The time complexity for calculating Euclidean distance in distance measurement is O(dn2), the time complexity for computing the exponential term is O(n), and the time complexity for the density factor is O(kn).
2. In this study, the shortest path distance calculation is implemented using Dijkstra’s algorithm, with a time complexity of O(n2) [52].
3. The time complexity for calculating similarity between samples is O(n2).
4. The time complexity for computing the normalized similarity matrix is O(n2).
5. The time complexity for calculating r1,r2,⋯rn and r is O(kn).
6. The eigen decomposition of the Laplacian matrix is performed using Singular Value Decomposition (SVD), with a time complexity of O(n3).
7. The time complexity for updating the similarity matrix S is O(n2).
In summary, the overall time complexity of the SNN-MSC algorithm is O(n3), which is of the same order of magnitude as that of the SC algorithm.
4. Conclusions
This study applies the SNN-MSC algorithm to remote sensing image processing, providing a novel approach for unsupervised classification of remote sensing images. Compared to previous spectral clustering optimization methods, this approach integrates spectral clustering with a density factor and replaces traditional distance metrics, such as Euclidean distance, with manifold distance. Adjusting the exponential term and scaling factor simultaneously satisfies both global and local consistency, better uncovering the intrinsic structure of the data. Furthermore, by utilizing shared neighbor information between samples and redefining the similarity measure through an exponential kernel, the method, combined with a rank constraint applied to the Laplacian matrix, enables adaptive graph learning, leading to a more accurate capture of the true data structure.
Based on ZY1-02D remote sensing imagery, the classification performance of the SNN-MSC algorithm was validated for land cover classification. The results indicate that, compared to using only spectral information, the SNN-MSC algorithm combined with multi-source information achieves an overall classification accuracy of 84.63%, with a 14.93% increase in the Kappa coefficient. Additionally, the SNN-MSC algorithm with multi-source information achieves producer accuracies (PA) for water bodies, transportation land, construction land, forest, and farmland of 88.03%, 85.29%, 86.44%, 80.96%, and 82.43%, respectively, significantly outperforming other comparative algorithms. The feature visualization results also provide favorable evidence supporting these findings. This demonstrates that the SNN-MSC algorithm, when combined with high-resolution remote sensing data, can quickly perform land cover classification, offering a feasible solution that enhances the accuracy and efficiency of land cover type identification. Furthermore, it shows a notable improvement in accuracy, particularly when handling complex-shaped data, such as transportation land, compared to other widely used methods.
References
1. 1. Li D, Wang S, He Q, Yang Y. Cost-effective land cover classification for remote sensing images. J Cloud Comp. 2022;11: 62.
* View Article
* Google Scholar
2. 2. Xie H, Huang H. Classification of Land Cover Remote-Sensing Images Based on Pattern Recognition. Hernandez JVC, editor. Scientific Programming. 2022;2022: 1–15.
* View Article
* Google Scholar
3. 3. Li R, Gao X, Shi F, Zhang H. Scale Effect of Land Cover Classification from Multi-Resolution Satellite Remote Sensing Data. Sensors. 2023;23: 6136. pmid:37447985
* View Article
* PubMed/NCBI
* Google Scholar
4. 4. Khatami R, Mountrakis G, Stehman SV. A meta-analysis of remote sensing research on supervised pixel-based land-cover image classification processes: General guidelines for practitioners and future research. Remote Sensing of Environment. 2016;177: 89–100.
* View Article
* Google Scholar
5. 5. Tong X, Xie H, Weng Q. Urban Land Cover Classification With Airborne Hyperspectral Data: What Features to Use? IEEE J Sel Top Appl Earth Observations Remote Sensing. 2014;7: 3998–4009.
* View Article
* Google Scholar
6. 6. Wu Q, Zhong R, Zhao W, Song K, Du L. Land-cover classification using GF-2 images and airborne lidar data based on Random Forest. International Journal of Remote Sensing. 2019;40: 2410–2426.
* View Article
* Google Scholar
7. 7. Eyster HN, Beckage B. Applying a deep learning pipeline to classify land cover from low-quality historical RGB imagery. PeerJ Computer Science. 2024;10: e2003. pmid:38855218
* View Article
* PubMed/NCBI
* Google Scholar
8. 8. Gong X, Hou Z, Wan Y, Zhong Y. Multispectral and SAR Image Fusion for Multiscale Decomposition Based on Least Squares Optimization Rolling Guidance Filtering. IEEE TRANSACTIONS ON GEOSCIENCE AND REMOTE SENSING. 2024;62.
* View Article
* Google Scholar
9. 9. Sun L, Wang X, Zheng Y, Wu Z. Multiscale 3-D–2-D Mixed CNN and Lightweight Attention-Free Transformer for Hyperspectral and LiDAR Classification. IEEE TRANSACTIONS ON GEOSCIENCE AND REMOTE SENSING. 2024;62.
* View Article
* Google Scholar
10. 10. Fan X, Hu Z, Zhao Y, Chen J, Wei T, Huang Z. A Small-Ship Object Detection Method for Satellite Remote Sensing Data. IEEE J Sel Top Appl Earth Observations Remote Sensing. 2024;17: 11886–11898.
* View Article
* Google Scholar
11. 11. Yin H, Zhang G, Wu Q, Cui F, Yan B, Yin S, et al. Unraveling Overlying Rock Fracturing Evolvement for Mining Water Inflow Channel Prediction: A Spatiotemporal Analysis Using ConvLSTM Image Reconstruction. IEEE TRANSACTIONS ON GEOSCIENCE AND REMOTE SENSING. 2024;62.
* View Article
* Google Scholar
12. 12. Yan P. Clustered remote sensing target distribution detection aided by density-based spatial analysis. International Journal of Applied Earth Observation and Geoinformation. 2024.
* View Article
* Google Scholar
13. 13. Chen J, Song Y, Li D, Lin X, Zhou S, Xu W. Specular Removal of Industrial Metal Objects Without Changing Lighting Configuration. IEEE Trans Ind Inf. 2024;20: 3144–3153.
* View Article
* Google Scholar
14. 14. Gu Y. MFGTN: A multi-modal fast gated transformer for identifying single trawl marine fishing vessel. Ocean Engineering. 2024.
* View Article
* Google Scholar
15. 15. Planet Craters Detection Based on Unsupervised Domain Adaptation. 2023;59.
* View Article
* Google Scholar
16. 16. Xu X, Fu X, Zhao H, Liu M, Xu A, Ma Y. Three-Dimensional Reconstruction and Geometric Morphology Analysis of Lunar Small Craters within the Patrol Range of the Yutu-2 Rover. Remote Sensing. 2023;15: 4251.
* View Article
* Google Scholar
17. 17. Cai G, Zheng X, Guo J, Gao W. Real-time identification of borehole rescue environment situation in underground disaster areas based on multi-source heterogeneous data fusion. Safety Science. 2025;181: 106690.
* View Article
* Google Scholar
18. 18. Huang J, Ma H, Sedano F, Lewis P, Liang S, Wu Q, et al. Evaluation of regional estimates of winter wheat yield by assimilating three remotely sensed reflectance datasets into the coupled WOFOST–PROSAIL model. European Journal of Agronomy. 2019;102: 1–13.
* View Article
* Google Scholar
19. 19. Proietti C, De Beni E, Cantarero M. One hundred lava flows of Mt. Etna, Italy: July 2019–December 2023 update. Journal of Maps. 2024;20: 2380899.
* View Article
* Google Scholar
20. 20. Carraha J, García J-L, Nussbaumer SU, Fernández-Navarro H, Gärtner-Roer I. Late Pleistocene to Holocene glacial, periglacial, and paraglacial geomorphology of the upper Río Limarí basin (30–31° S) in the Andes of central Chile. Journal of Maps. 2024;20: 2329179.
* View Article
* Google Scholar
21. 21. Wu C, Zhang J. One-Step Joint Learning of Self-Supervised Spectral Clustering With Anchor Graph and Fuzzy Clustering for Land Cover Classification. IEEE J Sel Top Appl Earth Observations Remote Sensing. 2024;17: 11178–11193.
* View Article
* Google Scholar
22. 22. Dietlmeier J, Ghita O, Duessmann H, Prehn JHM, Whelan PF. Unsupervised mitochondria segmentation using recursive spectral clustering and adaptive similarity models. Journal of Structural Biology. 2013;184: 401–408. pmid:24184470
* View Article
* PubMed/NCBI
* Google Scholar
23. 23. David G, Averbuch A. SpectralCAT: Categorical spectral clustering of numerical and nominal data. Pattern Recognition. 2012;45: 416–433.
* View Article
* Google Scholar
24. 24. Liu H, Zhao F, Jiao L. Fuzzy spectral clustering with robust spatial information for image segmentation. Applied Soft Computing. 2012;12: 3636–3647.
* View Article
* Google Scholar
25. 25. Röblitz S, Weber M. Fuzzy spectral clustering by PCCA+: application to Markov state models and data classification. Adv Data Anal Classif. 2013;7: 147–179.
* View Article
* Google Scholar
26. 26. Ozertem U, Erdogmus D, Jenssen R. Mean shift spectral clustering. Pattern Recognition. 2008;41: 1924–1938.
* View Article
* Google Scholar
27. 27. Langone R, Van Barel M, Suykens JAK. Efficient evolutionary spectral clustering. Pattern Recognition Letters. 2016;84: 78–84.
* View Article
* Google Scholar
28. 28. Alzate C, Suykens JAK. Sparse kernel spectral clustering models for large-scale data analysis. Neurocomputing. 2011;74: 1382–1390.
* View Article
* Google Scholar
29. 29. Langone R, Suykens JAK. Fast kernel spectral clustering. Neurocomputing. 2017;268: 27–33.
* View Article
* Google Scholar
30. 30. Lu H, Fu Z, Shu X. Non-negative and sparse spectral clustering. Pattern Recognition. 2014;47: 418–426.
* View Article
* Google Scholar
31. 31. Taşdemir K. Vector quantization based approximate spectral clustering of large datasets. Pattern Recognition. 2012;45: 3034–3044.
* View Article
* Google Scholar
32. 32. Liu W, Ye M, Wei J, Hu X. Compressed constrained spectral clustering framework for large-scale data sets. Knowledge-Based Systems. 2017;135: 77–88.
* View Article
* Google Scholar
33. 33. Orfanidis G, Tefas A, Nikolaidis N, Pitas I. Facial image clustering in stereoscopic videos using double spectral analysis. Signal Processing: Image Communication. 2015;33: 86–105.
* View Article
* Google Scholar
34. 34. Lin J, Xiao Z, Wei X, Duan P, He X, Dian R, et al. Click-Pixel Cognition Fusion Network With Balanced Cut for Interactive Image Segmentation. IEEE Trans on Image Process. 2024;33: 177–190. pmid:38055358
* View Article
* PubMed/NCBI
* Google Scholar
35. 35. Semertzidis T, Rafailidis D, Strintzis MG, Daras P. Large-scale spectral clustering based on pairwise constraints. Information Processing & Management. 2015;51: 616–624.
* View Article
* Google Scholar
36. 36. Higham DJ, Kalna G, Kibble M. Spectral clustering and its use in bioinformatics. Journal of Computational and Applied Mathematics. 2007;204: 25–37.
* View Article
* Google Scholar
37. 37. Chifu A-G, Hristea F, Mothe J, Popescu M. Word sense discrimination in information retrieval: A spectral clustering-based approach. Information Processing & Management. 2015;51: 16–31.
* View Article
* Google Scholar
38. 38. Quirós-Tortós J, Wall P, Ding L, Terzija V. Determination of sectionalising strategies for parallel power system restoration: A spectral clustering-based methodology. Electric Power Systems Research. 2014;116: 381–390.
* View Article
* Google Scholar
39. 39. Qin G, Gao L. Spectral clustering for detecting protein complexes in protein–protein interaction (PPI) networks. Mathematical and Computer Modelling. 2010;52: 2066–2074.
* View Article
* Google Scholar
40. 40. Bellugi D, Milledge DG, Dietrich WE, McKean JA, Perron JT, Sudderth EB, et al. A spectral clustering search algorithm for predicting shallow landslide size and location. JGR Earth Surface. 2015;120: 300–324.
* View Article
* Google Scholar
41. 41. Ding L, Li C, Jin D, Ding S. Survey of spectral clustering based on graph theory. Pattern Recognition. 2024;151: 110366.
* View Article
* Google Scholar
42. 42. Wang N, Ye X, Zhao J, Wang Q. Semantic Spectral Clustering with Contrastive Learning and Neighbor Mining. Neural Process Lett. 2024;56: 141.
* View Article
* Google Scholar
43. 43. Di Nuzzo C. Advancing Spectral Clustering for Categorical and Mixed-Type Data: Insights and Applications. Mathematics. 2024;12: 508.
* View Article
* Google Scholar
44. 44. Nie F, Liu C, Wang R, Li X. A Novel and Effective Method to Directly Solve Spectral Clustering. IEEE Trans Pattern Anal Mach Intell. 2024; 1–12. pmid:39167506
* View Article
* PubMed/NCBI
* Google Scholar
45. 45. Pospíšil L, Frič M, Čermák M. The texture roughness measures for engineering problems. Heraklion, Greece; 2024. p. 300004. https://doi.org/10.1063/5.0212126
46. 46. Qu B, Chang H. Remark on the Successive Projection Algorithm for the Multiple-Sets Split Feasibility Problem. Numerical Functional Analysis and Optimization. 2017;38: 1614–1623.
* View Article
* Google Scholar
47. 47. Elkhouly A, Andrew AM, Rahim HA, Abdulaziz N, Abdulmalek M, Mohd Yasin MN, et al. A Novel Unsupervised Spectral Clustering for Pure-Tone Audiograms towards Hearing Aid Filter Bank Design and Initial Configurations. Applied Sciences. 2021;12: 298.
* View Article
* Google Scholar
48. 48. Liu Y, Han Q, Li C, Xiao D. Numerical modeling of wave propagation for damped elbow pipes using Fourier–Legendre spectral element method in polar coordinates. Arch Appl Mech. 2016;86: 1995–2008.
* View Article
* Google Scholar
49. 49. Wong T-T, Yeh P-Y. Reliable Accuracy Estimates from k -Fold Cross Validation. IEEE Trans Knowl Data Eng. 2020;32: 1586–1594.
* View Article
* Google Scholar
50. 50. Warrens MJ. Relative quantity and allocation disagreement measures for category-level accuracy assessment. International Journal of Remote Sensing. 2015;36: 5959–5969.
* View Article
* Google Scholar
51. 51. Martín Andrés A, Álvarez Hernández M. Estimators of various kappa coefficients based on the unbiased estimator of the expected index of agreements. Adv Data Anal Classif. 2024 [cited 28 Aug 2024].
* View Article
* Google Scholar
52. 52. Wu T-F, Tsai P-S, Hu N-T, Chen J-Y. Combining turning point detection and Dijkstra’s algorithm to search the shortest path. Advances in Mechanical Engineering. 2017;9: 1687814016683353.
* View Article
* Google Scholar
Citation: Wu S, Cao J-M, Zhao X-Y (2025) Land cover classification of high-resolution remote sensing images based on improved spectral clustering. PLoS ONE 20(2): e0316830. https://doi.org/10.1371/journal.pone.0316830
About the Authors:
Song Wu
Roles: Data curation, Resources, Software, Visualization
Affiliation: Jilin Agricultural University, Changchun, China
Jian-Min Cao
Roles: Methodology, Project administration, Resources, Software
E-mail: [email protected]
Affiliation: Jilin Agricultural University, Changchun, China
ORICD: https://orcid.org/0009-0006-5109-4061
Xin-Yu Zhao
Roles: Data curation, Formal analysis, Methodology
Affiliation: Jilin Agricultural University, Changchun, China
[/RAW_REF_TEXT]
1. Li D, Wang S, He Q, Yang Y. Cost-effective land cover classification for remote sensing images. J Cloud Comp. 2022;11: 62.
2. Xie H, Huang H. Classification of Land Cover Remote-Sensing Images Based on Pattern Recognition. Hernandez JVC, editor. Scientific Programming. 2022;2022: 1–15.
3. Li R, Gao X, Shi F, Zhang H. Scale Effect of Land Cover Classification from Multi-Resolution Satellite Remote Sensing Data. Sensors. 2023;23: 6136. pmid:37447985
4. Khatami R, Mountrakis G, Stehman SV. A meta-analysis of remote sensing research on supervised pixel-based land-cover image classification processes: General guidelines for practitioners and future research. Remote Sensing of Environment. 2016;177: 89–100.
5. Tong X, Xie H, Weng Q. Urban Land Cover Classification With Airborne Hyperspectral Data: What Features to Use? IEEE J Sel Top Appl Earth Observations Remote Sensing. 2014;7: 3998–4009.
6. Wu Q, Zhong R, Zhao W, Song K, Du L. Land-cover classification using GF-2 images and airborne lidar data based on Random Forest. International Journal of Remote Sensing. 2019;40: 2410–2426.
7. Eyster HN, Beckage B. Applying a deep learning pipeline to classify land cover from low-quality historical RGB imagery. PeerJ Computer Science. 2024;10: e2003. pmid:38855218
8. Gong X, Hou Z, Wan Y, Zhong Y. Multispectral and SAR Image Fusion for Multiscale Decomposition Based on Least Squares Optimization Rolling Guidance Filtering. IEEE TRANSACTIONS ON GEOSCIENCE AND REMOTE SENSING. 2024;62.
9. Sun L, Wang X, Zheng Y, Wu Z. Multiscale 3-D–2-D Mixed CNN and Lightweight Attention-Free Transformer for Hyperspectral and LiDAR Classification. IEEE TRANSACTIONS ON GEOSCIENCE AND REMOTE SENSING. 2024;62.
10. Fan X, Hu Z, Zhao Y, Chen J, Wei T, Huang Z. A Small-Ship Object Detection Method for Satellite Remote Sensing Data. IEEE J Sel Top Appl Earth Observations Remote Sensing. 2024;17: 11886–11898.
11. Yin H, Zhang G, Wu Q, Cui F, Yan B, Yin S, et al. Unraveling Overlying Rock Fracturing Evolvement for Mining Water Inflow Channel Prediction: A Spatiotemporal Analysis Using ConvLSTM Image Reconstruction. IEEE TRANSACTIONS ON GEOSCIENCE AND REMOTE SENSING. 2024;62.
12. Yan P. Clustered remote sensing target distribution detection aided by density-based spatial analysis. International Journal of Applied Earth Observation and Geoinformation. 2024.
13. Chen J, Song Y, Li D, Lin X, Zhou S, Xu W. Specular Removal of Industrial Metal Objects Without Changing Lighting Configuration. IEEE Trans Ind Inf. 2024;20: 3144–3153.
14. Gu Y. MFGTN: A multi-modal fast gated transformer for identifying single trawl marine fishing vessel. Ocean Engineering. 2024.
15. Planet Craters Detection Based on Unsupervised Domain Adaptation. 2023;59.
16. Xu X, Fu X, Zhao H, Liu M, Xu A, Ma Y. Three-Dimensional Reconstruction and Geometric Morphology Analysis of Lunar Small Craters within the Patrol Range of the Yutu-2 Rover. Remote Sensing. 2023;15: 4251.
17. Cai G, Zheng X, Guo J, Gao W. Real-time identification of borehole rescue environment situation in underground disaster areas based on multi-source heterogeneous data fusion. Safety Science. 2025;181: 106690.
18. Huang J, Ma H, Sedano F, Lewis P, Liang S, Wu Q, et al. Evaluation of regional estimates of winter wheat yield by assimilating three remotely sensed reflectance datasets into the coupled WOFOST–PROSAIL model. European Journal of Agronomy. 2019;102: 1–13.
19. Proietti C, De Beni E, Cantarero M. One hundred lava flows of Mt. Etna, Italy: July 2019–December 2023 update. Journal of Maps. 2024;20: 2380899.
20. Carraha J, García J-L, Nussbaumer SU, Fernández-Navarro H, Gärtner-Roer I. Late Pleistocene to Holocene glacial, periglacial, and paraglacial geomorphology of the upper Río Limarí basin (30–31° S) in the Andes of central Chile. Journal of Maps. 2024;20: 2329179.
21. Wu C, Zhang J. One-Step Joint Learning of Self-Supervised Spectral Clustering With Anchor Graph and Fuzzy Clustering for Land Cover Classification. IEEE J Sel Top Appl Earth Observations Remote Sensing. 2024;17: 11178–11193.
22. Dietlmeier J, Ghita O, Duessmann H, Prehn JHM, Whelan PF. Unsupervised mitochondria segmentation using recursive spectral clustering and adaptive similarity models. Journal of Structural Biology. 2013;184: 401–408. pmid:24184470
23. David G, Averbuch A. SpectralCAT: Categorical spectral clustering of numerical and nominal data. Pattern Recognition. 2012;45: 416–433.
24. Liu H, Zhao F, Jiao L. Fuzzy spectral clustering with robust spatial information for image segmentation. Applied Soft Computing. 2012;12: 3636–3647.
25. Röblitz S, Weber M. Fuzzy spectral clustering by PCCA+: application to Markov state models and data classification. Adv Data Anal Classif. 2013;7: 147–179.
26. Ozertem U, Erdogmus D, Jenssen R. Mean shift spectral clustering. Pattern Recognition. 2008;41: 1924–1938.
27. Langone R, Van Barel M, Suykens JAK. Efficient evolutionary spectral clustering. Pattern Recognition Letters. 2016;84: 78–84.
28. Alzate C, Suykens JAK. Sparse kernel spectral clustering models for large-scale data analysis. Neurocomputing. 2011;74: 1382–1390.
29. Langone R, Suykens JAK. Fast kernel spectral clustering. Neurocomputing. 2017;268: 27–33.
30. Lu H, Fu Z, Shu X. Non-negative and sparse spectral clustering. Pattern Recognition. 2014;47: 418–426.
31. Taşdemir K. Vector quantization based approximate spectral clustering of large datasets. Pattern Recognition. 2012;45: 3034–3044.
32. Liu W, Ye M, Wei J, Hu X. Compressed constrained spectral clustering framework for large-scale data sets. Knowledge-Based Systems. 2017;135: 77–88.
33. Orfanidis G, Tefas A, Nikolaidis N, Pitas I. Facial image clustering in stereoscopic videos using double spectral analysis. Signal Processing: Image Communication. 2015;33: 86–105.
34. Lin J, Xiao Z, Wei X, Duan P, He X, Dian R, et al. Click-Pixel Cognition Fusion Network With Balanced Cut for Interactive Image Segmentation. IEEE Trans on Image Process. 2024;33: 177–190. pmid:38055358
35. Semertzidis T, Rafailidis D, Strintzis MG, Daras P. Large-scale spectral clustering based on pairwise constraints. Information Processing & Management. 2015;51: 616–624.
36. Higham DJ, Kalna G, Kibble M. Spectral clustering and its use in bioinformatics. Journal of Computational and Applied Mathematics. 2007;204: 25–37.
37. Chifu A-G, Hristea F, Mothe J, Popescu M. Word sense discrimination in information retrieval: A spectral clustering-based approach. Information Processing & Management. 2015;51: 16–31.
38. Quirós-Tortós J, Wall P, Ding L, Terzija V. Determination of sectionalising strategies for parallel power system restoration: A spectral clustering-based methodology. Electric Power Systems Research. 2014;116: 381–390.
39. Qin G, Gao L. Spectral clustering for detecting protein complexes in protein–protein interaction (PPI) networks. Mathematical and Computer Modelling. 2010;52: 2066–2074.
40. Bellugi D, Milledge DG, Dietrich WE, McKean JA, Perron JT, Sudderth EB, et al. A spectral clustering search algorithm for predicting shallow landslide size and location. JGR Earth Surface. 2015;120: 300–324.
41. Ding L, Li C, Jin D, Ding S. Survey of spectral clustering based on graph theory. Pattern Recognition. 2024;151: 110366.
42. Wang N, Ye X, Zhao J, Wang Q. Semantic Spectral Clustering with Contrastive Learning and Neighbor Mining. Neural Process Lett. 2024;56: 141.
43. Di Nuzzo C. Advancing Spectral Clustering for Categorical and Mixed-Type Data: Insights and Applications. Mathematics. 2024;12: 508.
44. Nie F, Liu C, Wang R, Li X. A Novel and Effective Method to Directly Solve Spectral Clustering. IEEE Trans Pattern Anal Mach Intell. 2024; 1–12. pmid:39167506
45. Pospíšil L, Frič M, Čermák M. The texture roughness measures for engineering problems. Heraklion, Greece; 2024. p. 300004. https://doi.org/10.1063/5.0212126
46. Qu B, Chang H. Remark on the Successive Projection Algorithm for the Multiple-Sets Split Feasibility Problem. Numerical Functional Analysis and Optimization. 2017;38: 1614–1623.
47. Elkhouly A, Andrew AM, Rahim HA, Abdulaziz N, Abdulmalek M, Mohd Yasin MN, et al. A Novel Unsupervised Spectral Clustering for Pure-Tone Audiograms towards Hearing Aid Filter Bank Design and Initial Configurations. Applied Sciences. 2021;12: 298.
48. Liu Y, Han Q, Li C, Xiao D. Numerical modeling of wave propagation for damped elbow pipes using Fourier–Legendre spectral element method in polar coordinates. Arch Appl Mech. 2016;86: 1995–2008.
49. Wong T-T, Yeh P-Y. Reliable Accuracy Estimates from k -Fold Cross Validation. IEEE Trans Knowl Data Eng. 2020;32: 1586–1594.
50. Warrens MJ. Relative quantity and allocation disagreement measures for category-level accuracy assessment. International Journal of Remote Sensing. 2015;36: 5959–5969.
51. Martín Andrés A, Álvarez Hernández M. Estimators of various kappa coefficients based on the unbiased estimator of the expected index of agreements. Adv Data Anal Classif. 2024 [cited 28 Aug 2024].
52. Wu T-F, Tsai P-S, Hu N-T, Chen J-Y. Combining turning point detection and Dijkstra’s algorithm to search the shortest path. Advances in Mechanical Engineering. 2017;9: 1687814016683353.
You have requested "on-the-fly" machine translation of selected content from our databases. This functionality is provided solely for your convenience and is in no way intended to replace human translation. Show full disclaimer
Neither ProQuest nor its licensors make any representations or warranties with respect to the translations. The translations are automatically generated "AS IS" and "AS AVAILABLE" and are not retained in our systems. PROQUEST AND ITS LICENSORS SPECIFICALLY DISCLAIM ANY AND ALL EXPRESS OR IMPLIED WARRANTIES, INCLUDING WITHOUT LIMITATION, ANY WARRANTIES FOR AVAILABILITY, ACCURACY, TIMELINESS, COMPLETENESS, NON-INFRINGMENT, MERCHANTABILITY OR FITNESS FOR A PARTICULAR PURPOSE. Your use of the translations is subject to all use restrictions contained in your Electronic Products License Agreement and by using the translation functionality you agree to forgo any and all claims against ProQuest or its licensors for your use of the translation functionality and any output derived there from. Hide full disclaimer
© 2025 Wu et al. This is an open access article distributed under the terms of the Creative Commons Attribution License: http://creativecommons.org/licenses/by/4.0/ (the “License”), which permits unrestricted use, distribution, and reproduction in any medium, provided the original author and source are credited. Notwithstanding the ProQuest Terms and Conditions, you may use this content in accordance with the terms of the License.
Abstract
Applying unsupervised classification techniques on remote sensing images enables rapid land cover classification. Using remote sensing imagery from the ZY1-02D satellite’s VNIC and AHSI cameras as the basis, multi-source feature information encompassing spectral, edge shape, and texture features was extracted as the data source. The Lanczos algorithm, which determines the largest eigenpairs of a high-order matrix, was integrated with the spectral clustering algorithm to solve for eigenvalues and eigenvectors. The results indicate that this method can quickly and effectively classify land cover. The classification accuracy was significantly improved by incorporating multi-source feature information, with a kappa coefficient reaching 0.846. Compared to traditional classification methods, the improved spectral clustering algorithm demonstrated better adaptability to data distribution and superior clustering performance. This suggests that the method has strong recognition capabilities for pixels with complex spatial shapes, making it a high-performance, unsupervised classification approach.
You have requested "on-the-fly" machine translation of selected content from our databases. This functionality is provided solely for your convenience and is in no way intended to replace human translation. Show full disclaimer
Neither ProQuest nor its licensors make any representations or warranties with respect to the translations. The translations are automatically generated "AS IS" and "AS AVAILABLE" and are not retained in our systems. PROQUEST AND ITS LICENSORS SPECIFICALLY DISCLAIM ANY AND ALL EXPRESS OR IMPLIED WARRANTIES, INCLUDING WITHOUT LIMITATION, ANY WARRANTIES FOR AVAILABILITY, ACCURACY, TIMELINESS, COMPLETENESS, NON-INFRINGMENT, MERCHANTABILITY OR FITNESS FOR A PARTICULAR PURPOSE. Your use of the translations is subject to all use restrictions contained in your Electronic Products License Agreement and by using the translation functionality you agree to forgo any and all claims against ProQuest or its licensors for your use of the translation functionality and any output derived there from. Hide full disclaimer