Content area
The existing methods in salient object detection (SOD) in hyperspectral images (HSI) have used different priors like center prior, boundary prior to procure cues to find the salient object. These methods fail, if the salient object is slightly touching the boundary. So, we extrapolate boundary connectivity, a measure to check if the object touches the boundary. The salient object is obtained by using background and foreground cues, which are calculated using boundary connectivity and contrast map, respectively. Also, to reduce the information redundancy and hence time complexity, we select top three most informative bands using different feature selection and feature extraction algorithms. The proposed algorithm is tested on HS-SOD dataset. It is observed that the proposed algorithm performs better than the state-of-the-art techniques in almost all the metrics, such as Precision (0.57), Recall (0.46),
Introduction
Object detection recognizes the semantic category and then estimates the spatial position of the object. Salient object detection (SOD) is the detection of the most noticeable part in the image. Mathematically, it is defined as the sparse entries in certain transformed feature space, where background is a low-rank matrix. In Hyperspectral Images (HSI), saliency detection has been achieved using different priors like center prior, boundary prior, etc. They are used to procure cues to detect the object. The individual saliency maps are merged into a single saliency map. The saliency map is further processed to get the most salient object. The salient object detection is demonstrated in Fig. 1. The first row shows HSI rendered sRGB images, and second row shows the corresponding salient object or ground truth. SOD has been explored heavily in low-dimensional images. Approaches like [1] use a guided smoothing technique to enhance the regions with most saliency. By utilizing a deep saliency model with guided non-local blocks, the approach accurately detects salient regions in input images [1]. However, these approaches have not been extrapolated to HSI. These approaches primarily focus on de-noising and may inadvertently blur spectral details critical for SOD in HSI. SOD has a wide range of applications like autonomous driving, robotics, medical image analysis, surveillance, remote sensing, etc.
Fig. 1 [Images not available. See PDF.]
Salient object detection in HSI, the first row shows the HSI rendered sRGB images and second row are the corresponding salient objects in the image
Hyperspectral imaging sensors accumulate information from a broad portion of electromagnetic spectrum. The hyperspectral cameras capture visible (0.4–0.7 m) to short-, mid-, and long-wave infrared region (2.4 m). The wavelength captured is divided into large number of spectral bands [2]. HSI is visualized as a three-dimensional cube with spatial and spectral dimensions along the axis as demonstrated in Fig. 2. The traditional sensors capture information in only three bands in visual spectrum. Hence miss out on the additional non-visual but useful information. The additional information helps us to identify the objects more precisely and exquisitely. The advances in sensor technologies are making data more readily available, building a promise for the future uses of this technology in image analysis tasks. HSI provides valuable auxiliary information enabling fine detection of objects. The low-dimensional images like RGB [1, 3, 4, 5, 6, 7–8] grayscale, etc do not differentiate between two different objects with same color. HSI have been used in the field of biomedical imaging [9], remote sensing [10], quality analysis [11], defense and security-based applications, etc.
Fig. 2 [Images not available. See PDF.]
Hyperspectral image with spectral band frames and reflectance curves [12]
HSI is very expensive to store and process. In addition, the higher spectral resolution induces redundancy across bands. So, it becomes imperative to reduce the dimensionality of HSI. Dimension reduction can be achieved by either feature extraction or feature selection. Feature extraction methods like PCA [13, 14–15], MNF [16] etc. extract new feature set from original features using some functional mapping. Since these methods extract new features, they lose the original physical information. However, in feature selection, subset of original features is selected and indigenous information is preserved. There have been very few studies of the above two categories in the context of SOD in HSI. This paper attempts to compare the two categories by taking two algorithms from each category and perform salient object detection using the features generated.
The spectral information-based approaches use spectral information only. They are effective in detecting spectral saliency. The inclusion of spatial information will improve the results. So, we use Simple Linear Iterative Clustering (SLIC) [17] to make superpixels.
Some researchers have shown that different priors like center prior, boundary prior [18, 19–20] can be useful to detect the salient object. The center prior-based methods [18, 19] assume that the salient object lies near the center. So, the pixels toward center get higher saliency values. These methods have difficulty in detecting objects that lie further from center. This prior gets an added advantage in the images taken under human supervision, where main focus, i.e., the salient part generally lies near center. On the contrary, Boundary prior considers the boundary pixels as non-salient. This method presumes the salient object would not lie on the boundary. This method fails to detect the object if it slightly touches the boundary. In remote sensing images the probability of an object lying near the boundary is higher. So we used boundary connectivity, a measure to evaluate the object connectivity with boundary to generate saliency cues [21].
The scarcity of labeled data and sparsity of high-dimensional data has hindered the growth of Deep Learning (DL) and HSI until now [22]. The high dimensionality and correlation of hyperspectral data complicate parameter estimation and information extraction. Additionally, limitations in ground truth datasets and dataset shift hinder the development and evaluation of models [23]. There have been attempts to use self-supervised Convolution Neural Networks (CNN) [24, 25] to extract features. The CNN model is updated based on cross-entropy loss, and saliency is computed by performing manifold ranking on the features learned by the task. The extracted features are used to detect the final object by applying manifold ranking [24]. These approaches need lots of data to train properly. In the absence of large standard object detection datasets, the object detection research is skewed toward the non-learning-based approach. Moreover, the heuristic integration of individual saliency maps in general saliency detection methods may sometimes give sub-optimal results. So, an optimization-based approach [21] is used to integrate the saliency maps. Subsequently, this paper attempts to achieve a following set of objectives:
The research in salient object detection in HSI has been mostly concerned with improving the results. There has been no attempt to include the processing time in the results. This paper is the first attempt to take the processing time into account and attempt to reduce it. To reduce the processing time, we extrapolated the saliency optimization framework to HSI as a deterministic approach. Henceforth, reducing the processing time.
Since HSI is high dimensional, so the feature reduction has a very important role to play as the research in the field advances. So, this paper attempts to analyze different feature reduction algorithms in the context of salient object detection.
This paper also attempts to compare different feature selection and feature extraction algorithms computational efficiency.
Literature review
Saliency detection was formulated by Itti and koch [26]. They introduced the saliency detection based on the visual attention model of early primates. It was designed for RGB images originally but has been extrapolated to HSI as well [27]. The direct extrapolation of Itti’s model to HSI [28] failed to detect the spectral saliency. Yan cao et al. used PCA to calculate top three components to replace the color components used by Itti. These methods use center surround difference to calculate saliency and hence may get stuck in local minima. Also, these methods use only spectral saliency and hence they would detect spectral saliency only.
To detect the spectrally similar but spatially different objects, some methods have used both spectral and spatial information [18, 19, 29]. It improves the detection but increases the computational complexity of the algorithm. Also, to avoid the effects of spectral variation gradient-based methods have been proposed [29]. The gradient helps to reduce the spectral variation leading to the improved detection performance especially in scenes with uneven illumination and shadows. Super-pixels are obtained by using SLIC algorithm [17]. The saliency score of each region is calculated using central prior and local contrast. The region with most distinctive reflectance property is produced as salient [18]. Zhang et al. [19] uses spatial and spectral information to reduce the effect of weaker cues. The inclusion of regional information helps us to eradicate the weak cues and smooth out noises. This method is insensitive to high contrast edges. In comparison, Liang et al. [27] performs better as it calculates angle similarity, which is independant of the spectral variation.
Imamoglu et al. [24] uses self-supervised CNN to extract the features and applied manifold ranking on the extracted features to detect the final salient object. The CNN extracts 64 features and applies manifold ranking on the extracted feature set. The CNN-based approaches need lots of data to train properly, which is scarce in case of HSI. The data scarcity has also affected the usage of deep learning for hyperspectral object detection tasks. So, we used a non-training-based approach to perform salient object detection.
Proposed method
The proposed method operates in two stages I). Feature reduction II). Saliency detection. The full architecture of the proposed method is given in Fig. 3. The detailed explanation of the various stages is given below:
Feature reduction In feature reduction, we select top most informative bands using different feature selection and feature extraction algorithms. It helps us to reduce the computational complexity and produces better results as-well. We drafted four different feature reduction algorithms with two from each feature selection (MEV-SFS [30], OPBS [31]) and feature extraction (PCA [13], MNF [16]) categories. Different feature reduction algorithms retain the original features or extract the new ones based on certain criteria. MEV-SFS selects the features based on their contribution to ellipsoid volume. The features that maximize the determinant of the selected bands covariance matrix i.e. ellipsoid volume are selected as optimal/final features. OPBS selects features based on their orthogonal projections. The largest orthogonal projection designates the feature to be selected. PCA retains the features which enhances the variance in the transformed feature space. MNF performs noise whitening before calculating the principal components.
Fig. 3 [Images not available. See PDF.]
The overall architecture of the proposed method
The resultant top three features obtained are given as input to the saliency detection algorithm and the results corresponding to all four algorithms were taken. The experimental evaluation indicates that MNF [16] produces better results in terms of precision, but MEV-SFS [32] is less complex computationally and gives almost equal score. The working of MEV-SFS [30] is given in Algorithm 1. The internal comparison and a detailed discussion of the algorithms is given in Sect. 4.2.1.
Saliency detection SOD is performed using background and foreground cues calculated using Eqs. (7) and (8), respectively [21]. These cues are integrated using an optimization framework as given in Eq. (9). To reduce the computational complexity, the pixels are initially grouped by performing segmentation using SLIC [17] on the reduced feature set. The SLIC [17] assists in reducing the complexity by reducing the number of individual units to be processed further. An undirected weighted graph is constructed with super pixels as nodes. The background and foreground cues are calculated using the undirected weighted graph. The euclidean distance between average pixel values of each super pixels is taken as an edge weight. Also, the minimum distance between two super pixels is defined as sum total of weights aggregated along the shortest path between the two superpixels as shown in Eq. (3).
3
where k is the total number of superpixels between i and j.The boundary connectivity is calculated as shown in Eq. (4). The boundary connectivity assists us in calculating the background cue () as shown in Eq. (7). The background cue value is close to 1, if boundary connectivity is large indicating that the super pixel belongs to background. If the boundary connectivity is small, the super pixel could be salient.
4
5
6
where B is the set of superpixels containing boundary pixels. The optimal value for was found empirically and equals to 10.7
The parameter with optimal value for our experimentation equal to 0.75. ; ’1’ indicates that the pixel belongs to background and ’0’ indicates that it belongs to foreground. The effectiveness of boundary connectivity is demonstrated in Sect. 4.2.The foreground cue is calculated using contrast map and background weight as shown in Eq. (8). The contrast map is defined as the addition of color distance to all other super pixels, weighted by position distance. is location distance and is calculated as euclidean distance between average value of pixel coordinates of each superpixel.
8
Finally, the foreground and background cues are integrated using an optimization framework as shown in Eq. (9):9
where is the saliency value of ith segment10
where i and j are two adjacent superpixels.Equation (9) converges by taking smaller value for high probable background pixels and higher value for high probable foreground pixels.
Results and discussion
Dataset and metrics
Dataset
One of the biggest challenges in HSI has been the lack of standard dataset for SOD. Different classification datasets have been used by researchers to evaluate SOD algorithms. However, HS-SOD [24] was made available recently and is now considered as the benchmark dataset for saliency detection in HSI. We evaluate our model using the same dataset. It contains a total of 60 hyperspectral images along with the corresponding ground truths and representative sRGB images. The images have been taken using the wavelength range 380–780 nm. The dataset contains images of different contrasts, sizes and locations. The camera used to capture the images is NH-AIK.
Table 1. A quantitative comparison of different feature reduction algorithm w.r.t salient object detection
Algorithm | Precision | Recall | score | NSS | CC | MAE |
|---|---|---|---|---|---|---|
[30] | 0.57 | 0.46 | 0.51 | 2.13 | 0.43 | 0.09 |
OPBS [30] | 0.56 | 0.48 | 0.50 | 2.10 | 0.43 | 0.08 |
PCA [13] | 0.52 | 0.41 | 0.42 | 1.67 | 0.31 | 0.09 |
MNF [16] | 0.60 | 0.45 | 0.53 | 2.13 | 0.43 | 0.09 |
Metrics
Different metrics evaluate different aspects of saliency, so we used six different metrics to evaluate the performance of our algorithm. We have used both location based (Normalized Scanpath Saliency (NSS)) and distribution-based (Correlation Coefficient (CC)) metrics to evaluate the effects of false positives and false negatives [33]. In addition, we calculate Precision, Recall and score as-well. We also calculated Mean Absolute Error (MAE) to illustrate our models insensitivity to non-salient pixels. Considering and as the predicted saliency and ground truth values, respectively, the metrics used are as follows:
Precision: It is the proportion of correctly identified salient pixels to the total number of pixels predicted as salient. Mathematically, it can be expressed as follows:
11
Recall: It is the proportion of correctly detected salient pixels to the total number of salient pixels. It can be expressed as follows:
12
score: It is the weighted harmonic mean of precision and recall. The weight was set to 0.3 to increase the priority for precision and is defined as below:
13
Normalized Scanpath Saliency (NSS): It calculates the resemblance between ground truth and saliency maps. It becomes sensitive to false positives if there are more false positives. The positive NSS scores show the correspondence and negative shows anti-correspondence as defined below:
14
Correlation Coefficient: It measures the linear correlation between saliency map and ground truth and is defined as:
15
Mean absolute Error (MAE): MAE is sensitive to pixels that are being correctly detected as non-salient, i.e true negatives and is defined as below:
16
Table 2. Time taken by each feature reduction algorithm in seconds
Algorithm | Feature reduction time (In seconds) | Saliency detection time (In seconds) | Total time (In seconds) |
|---|---|---|---|
MNF [16] | 15.15 | 4.33 | 19.48 |
PCA [13] | 13.08 | 4.51 | 17.59 |
OPBS [30] | 7.98 | 4.75 | 12.73 |
MEV-SFS [30] | 8.34 | 4.61 | 12.95 |
Discussion
This section discusses the results in two subsections i.e. feature reduction (Sect. 4.2.1) and saliency detection (Sect. 4.2.2). The feature reduction subsection includes the comparison of different feature selection and feature extraction algorithms with respect to SOD. The saliency detection section compares the SOD algorithm with state-of-the-art methods.
Table 3. Comparison of saliency detection-based algorithm in HSI performed on HS-SOD dataset
Model | Precision | Recall | score | NSS | CC | MAE |
|---|---|---|---|---|---|---|
Itti [26] | 0.19 | 0.23 | 0.35 | 1.36 | 0.35 | 0.25 |
SED [28] | 0.34 | 0.12 | 0.34 | 1.35 | 0.28 | 0.18 |
SAD [28] | 0.15 | 0.23 | 0.26 | 1.18 | 0.30 | 0.23 |
GS [28] | 0.27 | 0.23 | 0.30 | 1.56 | 0.34 | 0.22 |
SED-OCM-GS [28] | 0.27 | 0.22 | 0.33 | 1.59 | 0.37 | 0.18 |
SED-OCM-SAD [28] | 0.15 | 0.23 | 0.26 | 1.53 | 0.39 | 0.21 |
SGC [18] | 0.23 | 0.28 | 0.19 | 1.47 | 0.50 | 0.21 |
HS-MR [34] | 0.43 | 0.36 | 0.36 | 1.27 | 0.35 | 0.18 |
Proposed | 0.57 | 0.46 | 0.51 | 2.13 | 0.43 | 0.09 |
Feature reduction
In feature reduction, we used two feature selection such as MEV-SFS [30], OPBS [30] and two feature extraction algorithms such as MNF [16], PCA [13] to reduce the feature space. We included both the feature selection algorithms, where the final features preserve the original information and feature extraction algorithms that use different components (PCA [13], MNF [16]) and hence does not necessarily preserve the original information. In our experimental evaluation, we found that in feature extraction methods MNF produces better results with respect to the saliency detection. Whereas, PCA produced the poorest results. The MEV-SFS and MNF produce compatible results in terms of score (0.51, 0.53), NSS (2.13, 2.13), CC (0.43, 0.43) and MAE (0.09, 0.09). The quantitative comparison of saliency detection using four selected algorithms is given in Table 1. In addition, MEV-SFS is temporally efficient. It takes an average of 12.95 s to detect an object, which is about 33% less as compared to MNF which takes an average of 19.48 s. The time comparison of different feature reduction algorithm is given in Table 2. We also observed that feature selection algorithms require lower processing time than feature extraction algorithms. Since, feature extraction algorithms requires the data to be transformed to another feature space inducing an additional computational cost.
Fig. 4 [Images not available. See PDF.]
Time comparison (in seconds) between different feature reduction algorithms w.r.t salient object detection
The variation in the results of the feature extraction algorithms maybe explained by the arbitrary nature of transformed feature space. The feature space obtained after applying the transformation function may not be optimal always. MNF performs the noise whitening first, leading to reduction of noise in the final features selected. While as PCA equates variance and information, which is not true for images like HSI, which are highly sensitive to noise. OPBS band selection algorithms performed well with respect to PCA. We found out feature selection algorithms are stable as compared to feature extraction algorithms. This can be explained as feature extraction is non arbitrary with respect to the transformed feature space. The final components may not always contain enough relevant information. A graphical representation of time taken by different feature selection and feature extraction algorithms is shown in Fig. 4. The experiments were performed on Intel Xeon(R) [email protected] with OS type: Ubuntu (64 bit) and 32 GB RAM.
Fig. 5 [Images not available. See PDF.]
Visual sample of salient detection in HSI using different methods. (I) sRGB (II) GT (III). Itti [26] (IV). SED [28] (V). SAD [28] (VI). SG [28] (VII). SED-SAD [28] (VIII). SED-SG [28] (IX). HSISO [25] (X). [25] (XI). Proposed Method
Fig. 6 [Images not available. See PDF.]
A visual sample illustrating the effectiveness of boundary connectivity
Saliency detection
In terms of saliency detection, we observed that our method performs well with respect to state-of-the-art as can be seen from Table 3. Since, we have discussed and compared the feature reduction methods in Sect. 4.2.1.
This section compares results by taking MEV-SFS-based saliency optimization. OPBS is computationally more efficient as compared to MEV-SFS. It takes about 7.98 s on average to select the top three features, while it takes about 8.3 s in case of MEV-SFS. However, we finalized MEV-SFS as our choice for feature selection for its slight improvement in the results. It gave better results on all metrics in comparison with the state-of-the-art. Our model gave higher precision indicating that our model does not produce much false positives. It also produces higher CC and NSS indicating that our model is able to suppress both false positive and false negative fairly well. Moreover, the lower MAE confirms that the model makes very less individual prediction errors.
The proposed method being deterministic doesn’t need training. It also has an added advantage of low temporal complexity with whole method taking 12.11 s to complete on average. In addition, our method performs better than deep learning-based feature extraction methods [24, 25]. Moreover, deep learning-based algorithms require lots of data to train and test, which is not available in case of HSI. A sample visual representation of proposed method and other state-of-the-art algorithms is given in Fig. 5.
Our method generates noticeable saliency maps compared to other already existing methods. It produces better results in almost all metrics used to analyze SOD in HSI. Among the drafted feature reduction algorithms, we found out MEV-SFS is the most efficient algorithm in terms of score, NSS, CC and MAE. While PCA produces the least precision, recall and score. Since MEV-SFS is a feature selection algorithm, it preserves the original features. In addition, the feature selection algorithms do not transform the original feature space, that makes them computationally less complex.
Fig. 7 [Images not available. See PDF.]
Visual representation of final score obtained corresponding to different feature selection and feature extraction algorithms and other state-of-the-art
Fig. 8 [Images not available. See PDF.]
Visual representation of final NSS obtained corresponding to different feature selection and feature extraction algorithms and other state-of-the-art
Fig. 9 [Images not available. See PDF.]
Visual representation of final MAE obtained corresponding to different feature selection and feature extraction algorithms and other state-of-the-art
Moreover, MNF also produces good results but it is computationally more complex than MEV-SFS. The graphical representation of score, NSS and MAE in SOD are given in Figs. 7, 8, and 9. In addition, the proposed algorithm is effective in detecting the object even if it slightly touches the boundary. The effectiveness of the boundary connectivity is demonstrated in Fig. 6.
Conclusion
In this paper, we present a technique for salient object detection using feature reduction and saliency optimization. To improve accuracy of SOD, we used boundary connectivity and illustrated its effectiveness. Moreover, we analyzed different feature selection and feature extraction algorithms. We observed that feature selection methods were computationally less complex. MNF and MEV-SFS gave comparable results but, MNF is computationally less efficient. The experimental results on HS-SOD dataset confirm that the proposed method outperforms other state-of-the-art techniques by achieving better Precision, Recall, score, NSS, CC and lower MAE. The different metrics indicate that our method produces better results and is also somewhat resistant to background noise as-well. For future work, we can use other cues under the optimization framework to evaluate if the results can be improved.
Data availability
The dataset used in this paper is available in [HS-SOD] repository and can be accessed at: [https://github.com/gistairc/HS-SOD]. The code will be made available based on a reasonable request to corresponding author ([email protected]).
Declarations
Conflict of interest
The authors declare they have no conflict of interest.
Publisher's Note
Springer Nature remains neutral with regard to jurisdictional claims in published maps and institutional affiliations.
References
1. Zhu, L; Hu, X; Fu, C-W; Qin, J; Heng, P-A. Saliency-aware texture smoothing. IEEE Trans. Vis. Comput. Graph.; 2018; 26,
2. Fowler, JE; Du, Q. Anomaly detection and reconstruction from random projections. IEEE Trans. Image Process.; 2011; 21,
3. Liu, Z; Liu, J. Hypergraph attentional convolutional neural network for salient object detection. Vis. Comput.; 2022; 2022, pp. 1-27.
4. Liu, Z; Xiang, Q; Tang, J; Wang, Y; Zhao, P. Robust salient object detection for rgb images. Vis. Comput.; 2020; 36, pp. 1823-1835. [DOI: https://dx.doi.org/10.1007/s00371-019-01778-4]
5. Bayoudh, K; Knani, R; Hamdaoui, F; Mtibaa, A. A survey on deep multimodal learning for computer vision: advances, trends, applications, and datasets. Vis. Comput.; 2022; 38,
6. Huang, Z; Chen, B; Zhu, D. Imgeo-votenet: image and geometry co-supported votenet for rgb-d object detection. Vis. Comput.; 2023; 2023, pp. 1-13.
7. Wu, Y; Chang, X; Chen, D; Chen, L; Jia, T. Two-stage salient object detection based on prior distribution learning and saliency consistency optimization. Vis. Comput.; 2023; 39,
8. Das, DK; Shit, S; Ray, DN; Majumder, S. Cgan: closure-guided attention network for salient object detection. Vis. Comput.; 2022; 38,
9. Piqueras, S; Duponchel, L; Tauler, R; De Juan, A. Resolution and segmentation of hyperspectral biomedical images by multivariate curve resolution-alternating least squares. Anal. Chim. Acta; 2011; 705,
10. Manolakis, D; Truslow, E; Pieper, M; Cooley, T; Brueggeman, M. Detection algorithms in hyperspectral imaging systems: an overview of practical algorithms. IEEE Signal Process. Mag.; 2013; 31,
11. Barbin, DF; ElMasry, G; Sun, D-W; Allen, P. Predicting quality and sensory attributes of pork using near-infrared hyperspectral imaging. Anal. Chim. Acta; 2012; 719, pp. 30-42. [DOI: https://dx.doi.org/10.1016/j.aca.2012.01.004]
12. Polder, G., Pekkeriet, E., Snikkers, M.: A spectral imaging system for detection of botrytis in greenhouses (2013)
13. Abdi, H; Williams, LJ. Principal component analysis. Wiley Interdiscip. Rev. Comput. Stat.; 2010; 2,
14. Ali, M; Jones, MW; Xie, X; Williams, M. Timecluster: dimension reduction applied to temporal data for visual analytics. Vis. Comput.; 2019; 35,
15. Devi, RB; Chanu, YJ; Singh, KM. Discriminative object tracking with subspace representation. Vis. Comput.; 2021; 37,
16. Green, AA; Berman, M; Switzer, P; Craig, MD. A transformation for ordering multispectral data in terms of image quality with implications for noise removal. IEEE Trans. Geosci. Remote Sens.; 1988; 26,
17. Achanta, R; Shaji, A; Smith, K; Lucchi, A; Fua, P; Süsstrunk, S. Slic superpixels compared to state-of-the-art superpixel methods. IEEE Trans. Pattern Anal. Mach. Intell.; 2012; 34,
18. Yan, H., Zhang, Y., Wei, W., Zhang, L., Li, Y.: Salient object detection in hyperspectral imagery using spectral gradient contrast. In: 2016 IEEE International Geoscience and Remote Sensing Symposium (IGARSS), pp. 1560–1563 (2016). IEEE
19. Zhang, L; Zhang, Y; Yan, H; Gao, Y; Wei, W. Salient object detection in hyperspectral imagery using multi-scale spectral-spatial gradient. Neurocomputing; 2018; 291, pp. 215-225. [DOI: https://dx.doi.org/10.1016/j.neucom.2018.02.070]
20. Zhao, Z-Q; Zheng, P; Xu, S-T; Wu, X. Object detection with deep learning: a review. IEEE Trans. Neural Netw. Learn. Syst.; 2019; 30,
21. Zhu, W., Liang, S., Wei, Y., Sun, J.: Saliency optimization from robust background detection. In: Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, pp. 2814–2821 (2014)
22. Paoletti, M; Haut, J; Plaza, J; Plaza, A. Deep learning classifiers for hyperspectral imaging: a review. ISPRS J. Photogramm. Remote. Sens.; 2019; 158, pp. 279-317. [DOI: https://dx.doi.org/10.1016/j.isprsjprs.2019.09.006]
23. Wambugu, N; Chen, Y; Xiao, Z; Tan, K; Wei, M; Liu, X; Li, J. Hyperspectral image classification on insufficient-sample and feature learning using deep neural networks: a review. Int. J. Appl. Earth Obs. Geoinf.; 2021; 105, 102603.
24. Imamoglu, N., Oishi, Y., Zhang, X., Ding, G., Fang, Y., Kouyama, T., Nakamura, R.: Hyperspectral image dataset for benchmarking on salient object detection. In: 2018 Tenth International Conference on Quality of Multimedia Experience (QoMEX), pp. 1–3 (2018). IEEE
25. Huang, C; Xu, T; Zhang, Y; Pan, C; Hao, J; Li, X. Salient object detection on hyperspectral images in wireless network using cnn and saliency optimization. Ad Hoc Netw.; 2021; 112, 102369. [DOI: https://dx.doi.org/10.1016/j.adhoc.2020.102369]
26. Itti, L; Koch, C; Niebur, E. A model of saliency-based visual attention for rapid scene analysis. IEEE Trans. Pattern Anal. Mach. Intell.; 1998; 20,
27. Liang, J., Zhou, J., Bai, X., Qian, Y.: Salient object detection in hyperspectral imagery. In: 2013 IEEE International Conference on Image Processing, pp. 2393–2397 (2013). IEEE
28. Le Moan, S; Mansouri, A; Hardeberg, JY; Voisin, Y. Saliency for spectral image analysis. IEEE J. Sel. Top. Appl. Earth Observ. Remote Sens.; 2013; 6,
29. Liang, J; Zhou, J; Tong, L; Bai, X; Wang, B. Material based salient object detection from hyperspectral images. Pattern Recognit.; 2018; 76, pp. 476-490. [DOI: https://dx.doi.org/10.1016/j.patcog.2017.11.024]
30. Zhang, W; Li, X; Dou, Y; Zhao, L. A geometry-based band selection approach for hyperspectral image analysis. IEEE Trans. Geosci. Remote Sens.; 2018; 56,
31. Srujana, O.S., Mhala, N.C., Pais, A.R.: Secure transmission of hyperspectral images. In: 2020 Third ISEA Conference on Security and Privacy (ISEA-ISAP), pp. 94–99 (2020). IEEE
32. Zhang, L; Huang, X. Object-oriented subspace analysis for airborne hyperspectral remote sensing imagery. Neurocomputing; 2010; 73,
33. Bylinskii, Z; Judd, T; Oliva, A; Torralba, A; Durand, F. What do different evaluation metrics tell us about saliency models?. IEEE Trans. Pattern Anal. Mach. Intell.; 2018; 41,
34. Yang, C., Zhang, L., Lu, H., Ruan, X., Yang, M.-H.: Saliency detection via graph-based manifold ranking. In: Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, pp. 3166–3173 (2013)
Copyright Springer Nature B.V. Jan 2025