Linyi Li 1 and Tingbao Xu 2 and Yun Chen 3
Academic Editor:Silvia Conforto
1, School of Remote Sensing and Information Engineering, Wuhan University, Wuhan 430079, China
2, Fenner School of Environment and Society, The Australian National University, Canberra, ACT 2601, Australia
3, Commonwealth Scientific and Industrial Research Organization (CSIRO) Land and Water, Canberra, ACT 2601, Australia
Received 13 April 2017; Accepted 8 June 2017; 6 July 2017
This is an open access article distributed under the Creative Commons Attribution License, which permits unrestricted use, distribution, and reproduction in any medium, provided the original work is properly cited.
1. Introduction
With the rapid development of satellite and sensor technologies, remote sensing has become an important and efficient means to collect earth spatial information in recent years [1-4]. Remote sensing images with high spatial resolutions can be acquired from satellites, such as IKONOS, QuickBird, and WorldView [5, 6]. High resolution remote sensing images provide us with a great deal of information on texture structures and spatial details. The improvement in spatial resolution also increases the intraclass variability of land-cover classes and reduces the interclass variability between different classes [7], which increase the fuzziness of classification and pose a big challenge for automatic classification of remote sensing scenes. Remote sensing scenes are the separated subareas extracted from remote sensing images and possess specific semantic meanings, such as farmlands and residential areas. Remote sensing scene classification is a process to classify specific scenes in remote sensing images, which is essential to many remote sensing applications and has attracted much attention in recent years [8-10]. Various classification methods have been developed, which can be applied to remote sensing scene classification, such as minimum distance method [11], maximum likelihood method [11], neural network methods [12-16], fuzzy methods [17-21], support vector machine methods [21-23], particle swarm optimization methods [19, 24], artificial immune methods [25, 26], and Markov model methods [27-29]. However, due to the complex texture structures and spatial details in high resolution remote sensing scenes, scene classification is still a difficult task. Remote sensing scene classification methods based on visual attention may provide potential solutions to resolve this issue.
Visual attention is an important characteristic of the human visual system [30]. The human visual system can be easily attracted by salient details of an image and recognize objects or scenes in the image. Visual saliency measures to what extent details in an image attract human attention [31]. In the past twenty years, visual attention has become one of the hot spots in the relevant research and applications of artificial intelligence [32-36]. In 1998, Itti et al. proposed a visual attention model [32], which was based on the attention mechanism of the human visual system. The Itti visual attention model can be used to extract a variety of features from input images, such as brightness and color. Then these features were analyzed and consolidated to generate saliency maps. Walther and Koch further developed the saliency model proposed by Itti et al. They introduced a feedback mechanism in generating saliency maps for object recognition [33]. Achantay et al. proposed a frequency-tuned method to compute pixel saliency directly and detect salient regions [34]. Hou and Zhang designed a fast method to detect image saliency by exploring spectral components in an image [35]. Tian et al. proposed a color saliency model to detect salient objects in natural scenes [36]. In their color saliency model, different color features were extracted and analyzed. For different color features, two efficient saliency measurements were employed to compute different saliency maps. And a feature combination strategy was presented to combine multiple saliency maps into one integrated saliency map. Scene feature extraction is a key step in scene classification, which affects the classification accuracy. When the human visual system observes and classifies scenes, it is usually through a multiscale process. However, attempts to extract visual attention features through a multiscale process for scene classification are relatively rare in literatures.
The assumption of this study is that visual attention features could be extracted through a multiscale process for high resolution remote sensing scene classification. Fuzzy theory is an effective mathematical tool to process fuzzy and complex information [17-21], which could be suitable for high resolution remote sensing scene classification. Therefore, the fuzzy classification method [17-19] is preferred in this study. The main goals of this study are (1) to propose a novel visual attention feature extraction algorithm based on wavelet transform, which extracts visual attention features through a multiscale process; (2) to apply a fuzzy classification method (FC) using visual attention features (VAF) to achieve an improved accuracy in the scene classification; (3) to compare and evaluate the effects of FC-VAF with four traditional classification methods using IKONOS, QuickBird, and ZY-3 remote sensing scenes; and (4) to discuss the parameter sensitivity of FC-VAF.
2. Methodology
2.1. Wavelet Transform-Based Visual Attention Feature Extraction
2.1.1. Basic Principle of Wavelet Transform
The wavelet analysis is a powerful mathematical tool to obtain decomposition, reconstruction, and a multiscale representation of signals [37-39]. It introduces inherent scaling and good identification of signals, which is relevant to the human perception. A digital image is regarded as a two-dimensional discrete signal and can be decomposed and reconstructed by the two-dimensional discrete wavelet transform. The two-dimensional discrete wavelet transform allows good localization in both the frequency and spatial domain. The image can be decomposed into multiple levels using wavelet basis functions. It can be considered as a chain of successive levels of decomposition of the image by applying the one-dimensional discrete wavelet transform in the horizontal and vertical directions [37-39]. Two-level two-dimensional discrete wavelet decomposition of an image is illustrated in Figure 1. There are several popular wavelets in the field of the wavelet analysis, such as Daubechies wavelets, Symlets wavelets, and Discrete Meyer wavelet [39]. Different wavelets lead to different wavelet decomposition effects and application results.
Figure 1: Two-level two-dimensional discrete wavelet decomposition of an image.
[figure omitted; refer to PDF]
2.1.2. Visual Attention Feature Extraction through a Multiscale Process
Wavelet transform can obtain the multiscale representation of images. Therefore, a novel visual attention feature extraction algorithm based on wavelet transform is proposed, which extracts visual attention features from the saliency maps of remote sensing scenes through a multiscale process.
Visual saliency in an image measures to what extent details attract human attention [31]. Tian et al. proposed a color saliency model to detect salient objects in natural scenes [36]. In their color saliency model, different color features were extracted and analyzed. For different color features, two efficient saliency measurements were proposed to compute different saliency maps. And a feature combination strategy was presented to combine multiple saliency maps into one integrated saliency map. We adopt the color saliency model above to obtain the integrated saliency map s(x,y) for an image f(x,y) as follows [36]: [figure omitted; refer to PDF] where M=3; sm (x,y) (m=1,2,3) represent the saliency maps of the intensity, hue, and saturation components of the image, respectively; wm (m=1,2,3) represent the weight values of sm (x,y) (m=1,2,3), respectively; fm (x,y) (m=1,2,3) represent the intensity, hue, and saturation components of the image, respectively; fm ¯ is the average value of fm (x,y); dm ¯ is the average value of dm (x,y).
The visual attention features are extracted from an integrated saliency map as follows.
(a) The integrated saliency map is decomposed by N -level two-dimensional discrete wavelet transform. The multiscale representation of the integrated saliency map is obtained and composed by LL1 ,LL2 ,...,LLN . The multiscale representation of an integrated saliency map for visual attention feature extraction is illustrated in Figure 2, where N=2.
Figure 2: The multiscale representation of an integrated saliency map for visual attention feature extraction (N=2).
[figure omitted; refer to PDF]
(b) Visual attention focuses are extracted in the top level of the multiscale representation. The salient points in the top level are extracted based on the saliency values of the points. The human visual system can be easily attracted by the most salient point. Therefore, the most salient point is selected as the first and current visual attention focus. Then visual attention is shifted among the salient points in the top level. The next visual attention focus is the unselected salient point which is closest to the current visual attention focus. For example, there are three salient points in Figure 2. The most salient point A2 is selected as the first and current visual attention focus. Then select the salient point B2 as the second visual attention focus because it is closer to A2 than C2 .
(c) Visual attention is shifted from the top level to the low level of the multiscale representation. Take the visual attention focus A2 in Figure 2, for example. According to the position relation between two adjacent levels of the multiscale representation, A2 in LL2 corresponds to a small region in LL1 . Select the point with maximal value in the region as the corresponding visual attention focus A1 . In the same way, we can obtain the visual attention focus A0 in the visual saliency map.
(d) The saliency values of the visual attention focuses in the visual saliency map are used as the visual attention features for scene classification. In Figure 2, the saliency values of A0 , B0 , and C0 are used as the visual attention features.
2.2. Fuzzy Classification of Remote Sensing Scenes
We apply the fuzzy classification method [17-19] using visual attention features to achieve an improved accuracy of scene classification. The classification procedure is described as follows.
(a) Multiple original features are extracted from the samples of remote sensing scenes, including gray level cooccurrence matrix features [40], Laws texture energy features [40], and visual attention features. These features consist of feature vectors, which represent the corresponding scenes in the recognition process.
(b) The features are transformed into fuzzy features using the standard S-function as follows: [figure omitted; refer to PDF] where a, b, and c are the fuzzy parameters; b=(a+c)/2.
(c) Fuzzy class centers are obtained by using the mean value method. Suppose cij is the jth component of the class center of the ith class, N is the number of the training samples of the ith class, and tnj is the jth component of the feature vector of the training sample n; then cij is computed as follows: [figure omitted; refer to PDF] where j=1,2,...,M; M is the dimension of the feature vectors of the samples.
(d) Test samples are classified using Euclidean fuzzy closeness degree on the basis of the fuzzy closeness principle [18].
(e) Fuzzy classification results are assessed using overall accuracy (OA), Kappa coefficient (KC), average producer's accuracy (APA), and average user's accuracy (AUA) based on confusion matrices [41, 42].
A flowchart of the fuzzy classification process is shown in Figure 3.
Figure 3: Flowchart of the fuzzy classification process.
[figure omitted; refer to PDF]
3. Case Study
3.1. Materials
In order to validate the effectiveness of FC-VAF, 80 samples of remote sensing scenes were selected as the experimental data from widely used high spatial resolution remote sensing images, including IKONOS, QuickBird, and ZY-3 images. The samples consist of four classes, which are residential areas, farmlands, woodlands, and water areas, respectively. Each class has 20 samples where 10 samples are used as the training samples and all are used as the test samples. The size of the samples is 100×100 pixels. Representative samples of remote sensing scenes are shown in Figure 4.
Figure 4: Representative samples of remote sensing scenes. (a) Residential areas; (b) farmlands; (c) woodlands; (d) water areas.
(a) [figure omitted; refer to PDF]
(b) [figure omitted; refer to PDF]
(c) [figure omitted; refer to PDF]
(d) [figure omitted; refer to PDF]
3.2. Methods and Results
To demonstrate the effectiveness of FC-VAF, comparisons were carried out between FC-VAF and scene classification based on four traditional algorithms. The four methods for comparison are standard backpropagation neural network classification (SBPC), adaptive learning rule backpropagation neural network classification (ALRBPC), general regression neural network classification (GRNNC), and fuzzy classification (FC). Four gray level cooccurrence matrix features and four Laws texture energy features were extracted from these samples for all scene classification methods. The Euclidian closeness degree measurement was adopted in both FC and FC-VAF. Symlets wavelet was adopted in FC-VAF. The main parameters of different methods are shown in Table 1.
Table 1: Main parameters of different methods.
Method | Parameter description | Parameter value |
SBPC | Number of hidden layers | 1 |
Number of neurons in hidden layers | 15 | |
Learning rate | 0.01 | |
Maximum number of epochs to train | 5000 | |
| ||
ALRBPC | Number of hidden layers | 1 |
Number of neurons in hidden layers | 15 | |
Learning rate | 0.01 | |
Ratio to increase learning rate | 1.05 | |
Ratio to decrease learning rate | 0.7 | |
Maximum number of epochs to train | 1000 | |
| ||
GRNNC | Spread parameter | 0.5 |
| ||
FC | Fuzzy parameter a | 0.2 |
Fuzzy parameter c | 0.8 | |
| ||
FC-VAF | Fuzzy parameter a | 0.2 |
Fuzzy parameter c | 0.8 | |
Level of wavelet decomposition | 2 | |
Number of VAF | 4 |
We compared the results of different scene classification methods using the measures of OA, KC, APA, and AUA. Table 2 shows the performances in terms of the classification accuracy derived by SBPC, ALRBPC, GRNNC, FC, and FC-VAF. From Table 2, we can see that GRNNC outperformed FC and ALRBPC using OA, KC, APA, and AUA, while SBPC was the worst performer. FC-VAF obtained the best classification results among the five methods according to the values of OA, KC, APA, and AUA. For example, the OA values of SBPC, ALRBPC, GRNNC, FC, and FC-VAF are 76.3%, 78.8%, 82.5%, 80.0%, and 85.0%, respectively. The KC values of SBPC, ALRBPC, GRNNC, FC, and FC-VAF are 0.683, 0.717, 0.767, 0.733, and 0.800, respectively. FC-VAF can obtain satisfactory classification results in such images, because FC-VAF is on the basis of fuzzy theory and utilizes visual attention features in the process of classification.
Table 2: Comparisons of different scene classification methods.
Methods | Scene classification accuracy indicators | |||
Overall accuracy (OA) (%) | Kappa coefficient (KC) | Average producer's accuracy (APA) (%) | Average user's accuracy (AUA) (%) | |
SBPC | 76.3 | 0.683 | 76.3 | 78.4 |
ALRBPC | 78.8 | 0.717 | 78.8 | 81.5 |
GRNNC | 82.5 | 0.767 | 82.5 | 86.8 |
FC | 80.0 | 0.733 | 80.0 | 82.4 |
FC-VAF | 85.0 | 0.800 | 85.0 | 89.1 |
4. Discussion
4.1. Discussion of the Effects of Wavelet Decomposition Levels
The decomposition level (DL) of wavelets is the key parameter of FC-VAF, which affects the accuracy of scene classification. The scene classification accuracy of FC-VAF related to DL was analyzed and discussed. The 80 samples of scenes in the case study were used with different DL values (DL=1,2,3). Other parameters of FC-VAF were kept the same as those in the case study. The classification accuracy of FC-VAF for each DL value is shown in Figure 5. It shows that, with the increase of the DL value, the OA value increases to the maximum 85.0% when DL is 2 and then decreases. KC, APA, and AUA have similar trends as that of OA. Therefore, the optimal value of DL is 2 among the three values for FC-VAF in this application.
Figure 5: The effects of different wavelet decomposition levels (DL) on the classification accuracy. OA represents overall accuracy, KC represents Kappa coefficient, APA represents average producer's accuracy, and AUA represents average user's accuracy.
[figure omitted; refer to PDF]
4.2. Discussion of the Effects of Different Wavelets
Different wavelets lead to different wavelet decomposition effects, which affect the classification accuracy of FC-VAF. The scene classification accuracy of FC-VAF related to wavelets was analyzed and discussed. The 80 samples of scenes in the case study were used with different wavelets. Other parameters of FC-VAF were kept the same as those in the case study. The classification accuracy of FC-VAF using different wavelets is shown in Figure 6. It shows that DMeyer wavelet outperformed Daubechies wavelet using the measures of OA, KC, APA, and AUA, while Symlets wavelet was the best performer. For example, the OA values of Daubechies wavelet, DMeyer wavelet, and Symlets wavelet are 81.3%, 82.5%, and 85.0%, respectively. Therefore, Symlets wavelet is optimal among the three wavelets for FC-VAF in this application.
Figure 6: The effects of different wavelets on the classification accuracy. OA represents overall accuracy, KC represents Kappa coefficient, APA represents average producer's accuracy, and AUA represents average user's accuracy.
[figure omitted; refer to PDF]
5. Conclusions
In this study, a novel visual attention feature extraction algorithm was proposed, which extracted visual attention features through a multiscale process. And a fuzzy classification method using visual attention features (FC-VAF) was developed to perform high resolution remote sensing scene classification. FC-VAF was evaluated by using 80 samples of remote sensing scenes, which were selected from widely used high resolution remote sensing images, including IKONOS, QuickBird, and ZY-3 images. FC-VAF achieved more accurate classification results than four traditional classification methods according to the measures of OA, KC, APA, and AUA. The OA values of SBPC, ALRBPC, GRNNC, FC, and FC-VAF are 76.3%, 78.8%, 82.5%, 80.0%, and 85.0%, respectively. The KC values of SBPC, ALRBPC, GRNNC, FC, and FC-VAF are 0.683, 0.717, 0.767, 0.733, and 0.800, respectively. The classification accuracy of FC-VAF related to the decomposition level and to the wavelets was discussed.
FC-VAF can extract visual attention features through a multiscale process and improve the accuracy of scene classification in high resolution remote sensing images. Therefore, FC-VAF not only advances the research of visual attention models and digital image analysis methods, but also promotes the applications of high resolution remote sensing images. Possible further development of the study will focus on the integration of FC-VAF and other intelligent algorithms to further improve the accuracy of high resolution remote sensing scene classification.
Acknowledgments
This paper was supported by the National Natural Science Foundation of China (Grant no. 41371343). The authors also wish to thank Susan Cuddy at CSIRO for her helpful comments and suggestions.
[1] L. Li, Y. Chen, T. Xu, R. Liu, K. Shi, C. Huang, "Super-resolution mapping of wetland inundation from remote sensing imagery based on integration of back-propagation neural network and genetic algorithm,", Remote Sensing of Environment , vol. 164, pp. 142-154, 2015.
[2] Y. Chen, B. Wang, C. A. Pollino, S. M. Cuddy, L. E. Merrin, C. Huang, "Estimate of flood inundation and retention on wetlands using remote sensing and GIS,", Ecohydrology , vol. 7, no. 5, pp. 1412-1420, 2014.
[3] L. Li, Y. Chen, X. Yu, R. Liu, C. Huang, "Sub-pixel flood inundation mapping from multispectral remotely sensed images based on discrete particle swarm optimization,", ISPRS Journal of Photogrammetry and Remote Sensing , vol. 101, pp. 10-21, 2015.
[4] R. Liu, Y. Chen, J. Wu, L. Gao, D. Barrett, T. Xu, L. Li, C. Huang, J. Yu, "Assessing spatial likelihood of flooding hazard using naïve Bayes and GIS: a case study in Bowen Basin, Australia,", Stochastic Environmental Research and Risk Assessment , vol. 30, no. 6, pp. 1575-1590, 2016.
[5] J. Jeong, C. Yang, T. Kim, "Geo-positioning accuracy using multiple-satellite images: IKONOS, QuickBird, and KOMPSAT-2 stereo images,", Remote Sensing , vol. 7, no. 4, pp. 4549-4564, 2015.
[6] C. Tarantino, M. Adamo, R. Lucas, P. Blonda, "Detection of changes in semi-natural grasslands by cross correlation analysis with WorldView-2 images and new Landsat 8 data,", Remote Sensing of Environment , vol. 175, pp. 65-72, 2016.
[7] B. Demir, L. Bruzzone, "Histogram-Based Attribute Profiles for Classification of Very High Resolution Remote Sensing Images,", IEEE Transactions on Geoscience and Remote Sensing , vol. 54, no. 4, pp. 2096-2107, 2016.
[8] B. Zhao, Y. Zhong, G.-S. Xia, L. Zhang, "Dirichlet-derived multiple topic scene classification model for high spatial resolution remote sensing imagery,", IEEE Transactions on Geoscience and Remote Sensing , vol. 54, no. 4, pp. 2108-2123, 2016.
[9] Y. Zhong, Q. Zhu, L. Zhang, "Scene classification based on the multifeature fusion probabilistic topic model for high spatial resolution remote sensing imagery,", IEEE Transactions on Geoscience and Remote Sensing , vol. 53, no. 11, pp. 6207-6222, 2015.
[10] D. Ghosh, N. Kaabouch, "A survey on remote sensing scene classification algorithms,", WSEAS Transactions on Signal Processing , vol. 10, no. 1, pp. 504-519, 2014.
[11] ERDAS IMAGINE 2016, http://www.hexagongeospatial.com/support/documentation
[12] R. Ressel, A. Frost, S. Lehner, "A Neural Network-Based Classification for Sea Ice Types on X-Band SAR Images,", IEEE Journal of Selected Topics in Applied Earth Observations and Remote Sensing , vol. 8, no. 7, pp. 3672-3680, 2015.
[13] L. Li, T. Xu, Y. Chen, "Improved urban flooding mapping from remote sensing images using generalized regression neural network-based super-resolution algorithm,", Remote Sensing , vol. 8, no. 8, article no. 625, 2016.
[14] A. Taravat, F. Del Frate, C. Cornaro, S. Vergari, "Neural networks and support vector machine algorithms for automatic cloud classification of whole-sky ground-based images,", IEEE Geoscience and Remote Sensing Letters , vol. 12, no. 3, pp. 666-670, 2015.
[15] P. Vasuki, S. M. M. Roomi, "Automatic target classification of man-made objects in synthetic aperture radar images using Gabor wavelet and neural network,", Journal of Applied Remote Sensing , vol. 7, no. 1, 2013.
[16] A. Goltsev, V. Gritsenko, "Investigation of efficient features for image recognition by neural networks,", Neural Networks , vol. 28, pp. 15-23, 2012.
[17] A.-M. Yang, X.-G. Li, Y.-M. Zhou, Y.-F. Hu, "Fuzzy classifier based on support vector machine,", Journal of System Simulation , vol. 20, no. 13, pp. 3414-3419, 2008.
[18] S. Liu, H. Hou, H. Zhang, "Research of pattern recognition classification based on fuzzy theory for stored producted insects,", Computer Engineering and Applications , vol. 40, pp. 227-231, 2004.
[19] L. Li, D. Li, "Fuzzy classification of remote sensing images based on particle swarm optimization," in Proceedings of the International Conference on Electrical and Control Engineering, ICECE 2010, pp. 1039-1042, chn, June 2010.
[20] X. Si, Z. Peng, H. Yuan, G. Chen, "Research on cucumber downy mildew images classification based on fuzzy pattern recognition,", Sensor Letters , vol. 10, no. 1-2, pp. 643-648, 2012.
[21] S. Jenicka, A. Suruliandi, "Fuzzy texture model and support vector machine hybridization for land cover classification of remotely sensed images,", Journal of Applied Remote Sensing , vol. 8, no. 1, pp. 083540, 2014.
[22] H. Liu, S. Li, "Decision fusion of sparse representation and support vector machine for SAR image target recognition,", Neurocomputing , vol. 113, pp. 97-104, 2013.
[23] L. Li, Y. Chen, H. Gao, D. Li, "Automatic recognition of village in remote sensing images by support vector machine using co-occurrence matrices,", Sensor Letters , vol. 10, no. 1-2, pp. 523-528, 2012.
[24] X. Ding, H. Qiu, "Recognition model of oil reservoir based on chaotic particle swarm optimization,", International Journal of Earth Sciences and Engineering , vol. 8, no. 1, pp. 498-503, 2015.
[25] Y. Zhong, L. Zhang, "An Adaptive Artificial Immune Network for Supervised Classification of Multi-/Hyperspectral Remote Sensing Imagery,", IEEE Transactions on Geoscience and Remote Sensing , 2011.
[26] G. Dudek, "An artificial immune system for classification with local feature selection,", IEEE Transactions on Evolutionary Computation , vol. 16, no. 6, pp. 847-860, 2012.
[27] Y. Yuan, Y. Meng, L. Lin, H. Sahli, A. Yue, J. Chen, Z. Zhao, Y. Kong, D. He, "Continuous change detection and classification using hidden Markov model: A case study for monitoring urban encroachment onto farmland in Beijing,", Remote Sensing , vol. 7, no. 11, pp. 15318-15339, 2015.
[28] Y. Qiao, L. Weng, "Hidden markov model based dynamic texture classification,", IEEE Signal Processing Letters , vol. 22, no. 4, pp. 509-512, 2015.
[29] G. A. O. P. Costa, R. Q. Feitosa, "A generalized fuzzy Markov chain-based model for classification of remote-sensing multitemporal images,", International Journal of Remote Sensing , vol. 35, no. 1, pp. 341-364, 2014.
[30] L. Li, J. Ren, X. Wang, "Fast cat-eye effect target recognition based on saliency extraction,", Optics Communications , vol. 350, pp. 33-39, 2015.
[31] Z. Ren, S. Gao, L.-T. Chia, I. Tsang, "Region-based saliency detection and its application in object recognition,", IEEE Transactions on Circuits and Systems for Video Technology , vol. 24, no. 5, pp. 769-779, 2014.
[32] L. Itti, C. Koch, E. Niebur, "A model of saliency-based visual attention for rapid scene analysis,", IEEE Transactions on Pattern Analysis and Machine Intelligence , vol. 20, no. 11, pp. 1254-1259, 1998.
[33] D. B. Walther, C. Koch, "Attention in hierarchical models of object recognition,", Progress in Brain Research , vol. 165, pp. 57-78, 2007.
[34] R. Achantay, S. Hemamiz, F. Estraday, S. Süsstrunky, "Frequency-tuned salient region detection," in Proceedings of the IEEE International Conference on Computer Vision and Pattern Recognition (CVPR '09), pp. 1597-1604, June 2009.
[35] X. D. Hou, L. Q. Zhang, "Saliency detection: a spectral residual approach," in Proceedings of the IEEE Computer Society Conference on Computer Vision and Pattern Recognition (CVPR '07), pp. 1-8, June 2007.
[36] M. Tian, S. Wan, L. Yue, "A color saliency model for salient objects detection in natural scenes,", Lecture Notes in Computer Science (including subseries Lecture Notes in Artificial Intelligence and Lecture Notes in Bioinformatics) , vol. 5916, pp. 240-250, 2009.
[37] S. G. Mallat, "Theory for multiresolution signal decomposition: the wavelet representation,", IEEE Transactions on Pattern Analysis and Machine Intelligence , vol. 11, no. 7, pp. 674-693, 1989.
[38] S. G. Mallat, "Multifrequency Channel Decompositions of Images and Wavelet Models,", IEEE Transactions on Acoustics, Speech, and Signal Processing , vol. 37, no. 12, pp. 2091-2110, 1989.
[39] The MathWorks Inc. Wavelet Toolbox Documentation, 2016, http://cn.mathworks.com/help/wavelet/index.html
[40] Y. Jia, Digital Image Processing , Wuhan University Press, China, 2015. 3rd.
[41] J. Feng, L. Jiao, F. Liu, T. Sun, X. Zhang, "Unsupervised feature selection based on maximum information and minimum redundancy for hyperspectral images,", Pattern Recognition , vol. 51, pp. 295-309, 2016.
[42] P. Garcia-Llamas, L. Calvo, J. M. Álvarez-Martínez, S. Suárez-Seoane, "Using remote sensing products to classify landscape. A multi-spatial resolution approach,", International Journal of Applied Earth Observation and Geoinformation , vol. 50, pp. 95-105, 2016.
You have requested "on-the-fly" machine translation of selected content from our databases. This functionality is provided solely for your convenience and is in no way intended to replace human translation. Show full disclaimer
Neither ProQuest nor its licensors make any representations or warranties with respect to the translations. The translations are automatically generated "AS IS" and "AS AVAILABLE" and are not retained in our systems. PROQUEST AND ITS LICENSORS SPECIFICALLY DISCLAIM ANY AND ALL EXPRESS OR IMPLIED WARRANTIES, INCLUDING WITHOUT LIMITATION, ANY WARRANTIES FOR AVAILABILITY, ACCURACY, TIMELINESS, COMPLETENESS, NON-INFRINGMENT, MERCHANTABILITY OR FITNESS FOR A PARTICULAR PURPOSE. Your use of the translations is subject to all use restrictions contained in your Electronic Products License Agreement and by using the translation functionality you agree to forgo any and all claims against ProQuest or its licensors for your use of the translation functionality and any output derived there from. Hide full disclaimer
Copyright © 2017 Linyi Li et al. This is an open access article distributed under the Creative Commons Attribution License, which permits unrestricted use, distribution, and reproduction in any medium, provided the original work is properly cited.
Abstract
In recent years the spatial resolutions of remote sensing images have been improved greatly. However, a higher spatial resolution image does not always lead to a better result of automatic scene classification. Visual attention is an important characteristic of the human visual system, which can effectively help to classify remote sensing scenes. In this study, a novel visual attention feature extraction algorithm was proposed, which extracted visual attention features through a multiscale process. And a fuzzy classification method using visual attention features (FC-VAF) was developed to perform high resolution remote sensing scene classification. FC-VAF was evaluated by using remote sensing scenes from widely used high resolution remote sensing images, including IKONOS, QuickBird, and ZY-3 images. FC-VAF achieved more accurate classification results than the others according to the quantitative accuracy evaluation indices. We also discussed the role and impacts of different decomposition levels and different wavelets on the classification accuracy. FC-VAF improves the accuracy of high resolution scene classification and therefore advances the research of digital image analysis and the applications of high resolution remote sensing images.
You have requested "on-the-fly" machine translation of selected content from our databases. This functionality is provided solely for your convenience and is in no way intended to replace human translation. Show full disclaimer
Neither ProQuest nor its licensors make any representations or warranties with respect to the translations. The translations are automatically generated "AS IS" and "AS AVAILABLE" and are not retained in our systems. PROQUEST AND ITS LICENSORS SPECIFICALLY DISCLAIM ANY AND ALL EXPRESS OR IMPLIED WARRANTIES, INCLUDING WITHOUT LIMITATION, ANY WARRANTIES FOR AVAILABILITY, ACCURACY, TIMELINESS, COMPLETENESS, NON-INFRINGMENT, MERCHANTABILITY OR FITNESS FOR A PARTICULAR PURPOSE. Your use of the translations is subject to all use restrictions contained in your Electronic Products License Agreement and by using the translation functionality you agree to forgo any and all claims against ProQuest or its licensors for your use of the translation functionality and any output derived there from. Hide full disclaimer