Next Article in Journal
Biovalorization of Grape Stalks as Animal Feed by Solid State Fermentation Using White-Rot Fungi
Previous Article in Journal
Acceleration of an Algorithm Based on the Maximum Likelihood Bolometric Tomography for the Determination of Uncertainties in the Radiation Emission on JET Using Heterogeneous Platforms
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

Spatial Perception Correntropy Matrix for Hyperspectral Image Classification

1
Faculty of Artificial Intelligence in Education, Central China Normal University, Wuhan 430079, China
2
Aerospace Information Research Institute, Chinese Academy of Sciences, Beijing 100094, China
3
Department of Computer and Information Science, University of Macau, Macau 999078, China
4
Yangtze Three Gorges Technology and Economy Development Co., Ltd., Beijing 101100, China
*
Authors to whom correspondence should be addressed.
These authors contributed equally to this work.
Appl. Sci. 2022, 12(13), 6797; https://doi.org/10.3390/app12136797
Submission received: 29 May 2022 / Revised: 1 July 2022 / Accepted: 2 July 2022 / Published: 5 July 2022
(This article belongs to the Section Aerospace Science and Engineering)

Abstract

:
With the development of the hyperspectral imaging technique, hyperspectral image (HSI) classification is receiving more and more attention. However, due to high dimensionality, limited or unbalanced training samples, spectral variability, and mixing pixels, it is challenging to achieve satisfactory performance for HSI classification. In order to overcome these challenges, this paper proposes a feature extraction method called spatial perception correntropy matrix (SPCM), which makes use of spatial and spectral correlation simultaneously to improve the classification accuracy and robustness. Specifically, the dimension reduction is carried out firstly. Then, the spatial perception method is designed to select the local neighbour pixels. Thus, local spectral-spatial correlation is characterized by the correntropy matrix constructed using the selected neighbourhoods. Finally, SPCM representations are fed into the support vector machine for classification. The extensive experiments carried out on three widely used data sets have revealed that the proposed SPCM performs better than several state-of-the-art methods, especially when the training set is small.

1. Introduction

Hyperspectral image (HSI) captures an enormous amount of spectral and spatial information with hundreds of subdivided spectral bands [1,2], which provides an effective tool to discriminate various land covers. Therefore, HSI classification, which predicts the label of each pixel, can be widely used in precision agriculture [3,4], environmental management [5], surveillance [6], and so forth. However, due to high dimensionality, limited or unbalanced training samples, spectral variability, and mixing pixels, it is challenging to achieve satisfactory performance for HSI classification methods [7,8,9].
To mitigate the issues mentioned above, a considerable amount of HSI classification methods have been designed in the past few years. According to the way of feature extraction, existing HSI classification methods can be split into two categories, i.e., spectral-based methods and spectral-spatial-based methods [2]. In the early stage of research, researchers usually focused on making use of spectral information for classification, where the raw pixel vectors are fed into the classifier without any feature extraction [10,11,12]. However, the original spectral information contains large-scale redundancy, and the relationship between spectrum and land cover is nonlinear, which increases the difficulty of classification. Therefore, many researchers began to pay attention to dimension reduction methods for HSI classification. Some dimension reduction methods are widely used for HSI classification, and typical methods include principal component analysis (PCA) [13], linear discriminant analysis(LDA) [14], local linear embedding (LLE) [15], independent component analysis (ICA) [16], and maximum noise fraction (MNF) [17]. Since only global statistical data are considered, the performance of PCA in extracting local features of HSI classification is not prominent [18]. In addition, existing researches also show that MNF is usually superior to PCA in HSI classification [19]. Nevertheless, the spectral-based approach usually leads to unsatisfactory results in the absence of the spatial feature. In HSI classification, the spatial information [20], which provides abundant shape, context and layout features, is very important for high precision HSI classification. It is commonly believed that the joint extraction of spatial and spectral features can improve the classification accuracy and robustness [21].
To incorporate the spatial information, many spectral-spatial-based classification methods have been developed in recent years [22,23,24]. Fang et al. introduced superpixel-based classification via multiple kernels (SCMK) [25,26] to obtain uniform regions for spatial-spectral feature extraction. Huang et al. [27] proposed a method based on local linear spatial-spectral probability distribution to make full use of HSI local geometry and spatial correlation. In the past few years, with the popularity of deep learning and its success in natural image processing, more and more deep learning methods have been designed to exploit the discriminative information for HSI classification [2,28,29], such as convolutional neural network (CNN) [30], deep brief network [31], and graph convolution network [32]. Recently, 3D CNN is also used to preserve the joint spatial and spectral information in the feature learning process, which can learn important discrimination information for HSI classification [33]. Furthermore, 3D generative adversarial network (3D-GAN) has also been applied to HSI classification [34]. Recurrent Neural Network (RNN) is used to characterize the sequence characteristics of hyperspectral pixel vectors for classification tasks. Mou et al. proposed a novel RNN with a specially designed activation function and modified gated recurrent unit (GRU) to solve the multiclass classification for HSI [35]. Although these methods based on deep learning have made great breakthroughs in HSI classification, there are still some important problems to be solved [2]. For example, deep learning-based methods usually need a large number of training data to produce better classification performance. However, labelling HSI data is always expensive and time-consuming.
It is well known that HSIs usually consist of a large number of correlated bands. More importantly, the current research indicates that correlation is important in many computer vision task, especially when the training set is small [36]. However, most of the existing HSI classification methods only focus on dimension reduction to remove redundancy. In order to make full use of the spectral correlation, Fang et al. proposed local covariance matrix representation (LCMR) for HSI classification, which represents each pixel by a covariance matrix [37]. This method uses the covariance matrix to characterize different pixels and achieves good classification performance. However, the high nonlinearity of the HSI decreases the performance of the covariance matrix-based method. Fortunately, correntropy has good advantages in tackling nonlinear and non-Gaussian data [38]. Consequently, Zhang et al. proposed local correntropy matrix (LCEM) and achieved good classification performance [39]. This method represents the pixel of HSI by a local correntropy matrix, which contains rich spatial and spectral information. However, the local correntropy matrix is constructed by a local window, where the centre of the window is aligned with the pixels being processed. In this way, when the pixel being processed is on an edge, its local correntropy matrix will be constructed by pixels belonging to different classes. In this case, this representation can not characterize the class being processed due to the mixing.
To reduce the impact of pixels belonging to different classes, a spatial perception method has been proposed for neighbourhood construction in this paper. This strategy aims to construct a neighbourhood by a suitably selected window rather than a fixed window. So, more pixels belonging to the same class can be obtained. In this way, a more discriminative local correntropy matrix can be constructed to improve the classification accuracy. The main contributions of this paper are as follows:
  • This paper presents a neighbour selection method for sensing the spatial structure of local pixels. In this way, more pixels belonging to the same class can be selected. Therefore, the obtained representation is expected to well characterize the given class to which the pixel being processed belongs;
  • This paper proposes a spatial perception correntropy matrix (SPCM) for feature representation. A larger proportion of similar pixels can improve the feature representation effect of the correntropy matrix. The experimental results obtained using three publicly available hyperspectral data sets indicate that the proposed HSI classification method can extract more discriminative spectral-spatial features without using a large amount of training data.
The rest of this article is structured as follows: Section 2 presents the proposed spatial perception correntropy matrix, which jointly learns spectral and spatial correlation. Section 3 introduces the HSI classification method based on SPCM. Section 4 shows the comparison results between the proposed method and five state-of-the-art methods, which are carried out on three hyperspectral data sets. Finally, conclusions and possible future research directions are given in Section 5.

2. Spatial Perception Correntropy Matrix

Spectral and spatial correlation plays an important role in HSI classification. It is a common strategy to extract spatial correlation through the local structure of HSI. However, the spectral correlation, which can offer important discriminative information, is not fully exploited in the field of HSI classification [37]. In this paper, we propose a new spectral-spatial feature extraction method based on spatial perception correntropy matrix to jointly exploit the spatial and spectral correlation. The focus of the proposed method is how to determine the spatial correlation in the process of the correntropy matrix construction. The correntropy matrix of a pixel is expected to be constructed by the pixels belonging to the same class. In this way, the correntropy matrix can discriminate different classes. Figure 1 describes the process of SPCM method in detail.

2.1. Spatial Perception of Window

In this paper, we propose a spatial perception method to select the neighbouring pixels for correntropy matrix construction. Let A R W × H × N be an HSI, where W, H, and N are the width, height, and spectral bands, respectively. The pixel p x , y of A can be defined as a spectral vector p x , y R N . Figure 1 shows the definition of a window, where ( x , y ) is the position of the pixel to be processed, and ( X , Y ) is used to represent the centre of the side window. The region shown in Figure 2 can be defined as W t R L × L × N (with t = 1 , 2 , , 9 ). The range is from X r to X + r in the horizontal direction and from Y r to Y + r in the vertical direction, so that all the pixels in the selected rectangular shape form a side window. r is the window radius, which is determined by the window width L, that is, L = 2 r + 1 . The choice of hyperparameters will be discussed in Section 4.2. Different windows are obtained by calculating the positions of the two coordinates, namely, changing ( X , Y ) . In this paper, a total of three hyperparameters are contained, which are represented by italics: L is the width of window; S is the first s pixel vectors involved in comparison when selecting the neighbouring pixels, and K is the first k pixels involved in the construction of correntropy matrix.
In this paper, we chose nine different windows by changing the value of ( X , Y ) . Figure 3 shows nine different windows obtained by changing the value of ( X , Y ) . The window is set to square with a fixed length 2 r + 1 . Thus, nine windows of a pixel being processed are taken as a candidate.
Next, the average cosine distance between the pixel being processed and the adjacent pixels within each window can be obtained. Then the window with maximum cosine distance is selected as the optimal window. Finally, the first k most similar pixels of the selected window are used for the construction of the correntropy matrix.
Let us define that the pixel in the window is p 1 and the surrounding pixel is p i , i = 2 , 3 , , L 2 . The cosine distance between the target pixel and the surrounding pixel can be expressed as:
c o s ( p 1 , p i ) = < p 1 , p i > p 1 2 · p i 2 , i = 1 , 2 , , L 2
where 〈·〉 and · 2 represent the inner product and L 2 -norm. Through this formula, the window W t with the largest average value of the first S surrounding pixels is obtained. The pixel-averaged cosine distance CW t of the window W t is expressed as:
CW t = 1 S i = 1 S c o s ( p 1 , p i ) , t = 1 , 2 , , 9
After selecting the optimal window W t , the correntropy matrix is constructed by the first K − 1 most similar pixel vectors in W t .

2.2. Construction of Correntropy Matrix

The optimal window is calculated by Equation (2), and then this window is used to construct the correntropy matrix. K − 1 similar pixel vectors selected by spatial perception are used to construct a correntropy matrix with a target pixel vector. Define q i and q j as two different spectral bands, q i n and q j n as the n-th spectral value on the q i th and q j th spectral bands, then the correntropy between q i and q j can be expressed as:
C ( q i , q j ) = 1 K n = 1 K k σ ( q i n , q j n )
where
k σ ( q i n , q j n ) = 1 2 π σ e x p ( q i n q j n 2 2 2 σ 2 )
where σ is the width parameter of the function, which controls the radial scope of the function. Set σ to σ = 0.05 in the experimental section. Then the following correntropy matrix can be obtained:
M c = { C ( q i , q j ) } i , j = 1 N
where N denotes the number of spectral channels. The nondiagonal element of the correntropy matrix is the correntropy between different bands, reflecting the relationship between different bands. Through the above process, each pixel in the HSI can be represented by a correntropy matrix. The obtained correntropy matrix not only contains spectral information, but also reflects the relationship between similar spectra. So SPCM can provide more distinguishing features for classification and improve classification performance.

3. HSI Classification Based on SPCM

Based on the proposed SPCM, a spectral-spatial HSI classification method is proposed for HSI classification in this paper. Figure 4 describes the general process of SPCM-based classification method in detail. Algorithm 1 details the pseudocode of SPCM. Firstly, the dimension reduction method MNF is used to reduce noise and computational complexity. This paper reduces the dimension to 20. Next, the window closest to the target pixel is selected according to spatial perception and similar pixels are obtained. The size of the window and the number of selected pixels are two important hyperparameters. Then, the correntropy matrix is constructed according to each pixel and its similar pixels. So, each pixel can be represented by a matrix. Finally, a set of matrices are fed into SVM for classification. It is easy to find that the matrix constructed in this paper is a symmetric positive definite matrix, but SVM cannot accept it as input. Therefore, a logm function mentioned in Refs. [40,41] can be used to transform points into Euclidean space for the input of SVM. The logarithm-Euclidean kernel can be defined by:
k l o g m ( M 1 , M 2 ) = t r a c e [ l o g m ( M 1 ) · l o g m ( M 2 ) ] ,
where M 1 and M 2 are two different correntropy matrices. Given a symmetric positive definite matrix C = U U T , its logm can be defined as:
l o g m ( C ) = U l o g ( ) U T .
This way, the proposed HSI classification method, making use of spectral and spatial correlation simultaneously, can achieve a high classification accuracy when the training sample set is small, which solves the problem of poor classification performance in the case of small training samples.
Algorithm 1: SPCM.
Applsci 12 06797 i001

4. Experimental Results

This section demonstrates the effectiveness of our newly proposed method using three well-known HSI data sets. The proposed method is compared with five state-of-the-art methods, including LCMR [37], SCMK [25], CNN-enhanced GCN (CEGCN) [32], random patch network (RPNet) [42], and LCEM [39]. This paper uses three commonly used indicators, including overall accuracy (OA), average accuracy (AA), and k coefficient [43]. In this paper, the average OAs, AAs, and k coefficients of 10 random runs are reported. In addition, full classification maps of different methods are presented for visual analysis. On the Pavia University dataset, the training samples are set to 0.1% of each class; On KSC and Botswana datasets, the training samples are set to 1% per class, while the others were used for testing.
In the experiment, algorithms LCMR, SCMK, RPNet, LCEM, and SPCM are implemented by MATLAB R2020b, and are conducted using the Intel Core E5-2620 2.10 GHz CPU with 64 GB memory. However, algorithm CEGCN is conducted on a device with a 2.6 GHz CPU, 128 GB of RAM, and an A6000 GPU, and is performed on PyCharm 2021.

4.1. HSI Data Sets

In this section, three real public data sets, including Pavia University, Kennedy Space Center (KSC), and Botswana, were used to evaluate the performance of the proposed method. Figure 5 gives a brief overview of three HSI data sets.
  • Pavia University: The first image is Pavia University, often used in HSI classification. The data were obtained by ROSIS sensors in Pavia, Italy. After processing, the size of the dataset is 610 × 340 × 103. The dataset contains 9 categories with 42,776 labelled pixels. The detailed information is tabulated in Figure 5.
  • Kennedy Space Center: The second set of data comes from the Kennedy Space Center. It was collected on 23 March 1996, NASA AVIRIS (AVIRIS) at the Kennedy Space Center (KSC). A total of 176 bands were used in this experiment because of their absorptive capacity and low signal-to-noise ratio. There are 13 kinds of land covers, containing 5211 labelled pixels. The detailed information is tabulated in Figure 5.
  • Botswana: The last dataset is Botswana. Between 2001 and 2004, NASA’s EO-1 satellite collected a set of data in the Okavango Delta in Botswana. On the EO-1, the Hyperion sensor can get 30-m pixels over a 7.7-km strip, and within a 10-nm window, 242 wavelengths in the 400–2500 nm band. The bands without calibration and noise are removed, and the remaining 145 bands are used for HSI classification. The detailed information is tabulated in Figure 5.

4.2. Parameters Setting

The SPCM proposed in this paper can effectively extract the spectral-spatial features. In the stage of spatial structure perception, parameter setting is particularly important. Figure 6 shows the effects of different L, S and K on OA, AA, and Kappa on the KSC dataset. S and K were set to 0.8 L 2 , and then the size of L was changed. It can be found in the figure that the with increase of L, OA, AA and Kappa increased first and then decreased slightly. This is because when the window size is very small, the same kind of pixels that can be provided will be very few, which is not enough with the feature representation; when the window is too large, there will be too many heterogeneous pixels, which will affect the accuracy of feature representation. After determining the value of L, we fine-tuned S and K to achieve the best classification results. Finally, the three parameters that can produce the best classification result can be obtained.

4.3. Performance Evaluation

In this section, we show a comparison between SPCM and other advanced algorithms on three real datasets using small sample training sets.

4.3.1. Experiments with the Pavia University Data Set

In this section, the experimental results on the Pavia University data set were reported firstly, where the number of training samples per class was set to 0.1%, while the others were used for testing. Table 1 lists the average quantification results for each class. Table 1 shows that RPNet achieves the lowest accuracy. The reason for this may be that it is difficult to choose appropriate random patches when the training sample set is small. In addition, we can also find that CEGCN achieves low accuracy. This may be because the algorithm based on deep learning needs a large number of training samples, and 0.1% of the training samples are not enough. Thus, it does not provide satisfactory results. It is easy to find that the proposed SPCM obtains the best performance. Compared with the advanced method LCEM, the improved SPCM is about 2% higher in OA, AA, and Kappa. Both LCEM and SPCM can extract the spectral correlation features, but the performance of SPCM method is higher than that of LCEM. This is because SPCM adopts the spatial perception technique, which can better reflect the nonlinear relationship between spectral bands in HSI compared with the LCEM method. Additionally, SPCM emphasizes intraclass spectral correlation more than LCEM. Consequently, the local correntropy matrices obtained by the spatial perception method can extract more discriminative features, which facilitates the final classification.
In addition to the objective experimental results comparison, we visualize the results to verify the effectiveness of the method. The full classification maps of different methods are shown in Figure 7, and they are generated by one of the random experiments. As we can observe in Figure 7, the proposed method has achieved the best performance using only 0.1% samples for training, and some methods produced maps with more noise and lost some structural details. For example, LCMR and SCMK generate salt-and-pepper noise with a small training set. For PRNet, its classification map loses many edge structures. CEGCN produces an over-smooth classification map, which may be caused by its larger filter size. From Figure 7, it can be seen that the borders between classes are generally well defined and without salt and pepper noise.

4.3.2. Experiments with the KSC Data Set

The second experiment was based on the KSC data set. It was again repeated 10 times over the randomly split training and test data, where 1% per class were selected as training samples. Table 2 reports the average classification results of each method. Table 2 shows that SPCM and LCEM obtain the best and second results of 94.51% and 93.83%, respectively. These results prove again that correntropy matrix-based methods are fit for HSI classification. We can also find that LCMR, LCEM, and SPCM perform better than other compared methods. This phenomenon also verified the effectiveness of the spectral-spatial correlation learning.
Figure 8 shows the full classification maps, where only 1% of samples in each class are used for comparison. Figure 8 shows that in the classification results of SVM, the SCMK and CEGCN methods lose too much image structure. PRNet and LCEM produce severe salt-and-pepper noise. Although LCMR is very obvious in suppressing salt-and-pepper noise, it also has a oversmooth phenomenon. The proposed SPCM can effectively solve this problem. SPCM emphasizes the correlation of the spectrum and also makes use of the spectral and spatial information. As can be seen from Figure 8 and Table 2, compared with other methods, a good trade-off is achieved with SPCM.

4.3.3. Experiments with the Botswana Data Set

The last experiment was carried out on the Botswana data set, setting the number of training samples per class to 1% of the total samples, while the others were used for testing. This data set is collected by satellite, with low spatial resolution, complex shape, variety and irregular distribution of ground objects, so it poses a great challenge to many methods. On this data set, the hyperparameters were set to L = 9 , S = 35 , K = 45 . Table 3 lists the accuracy of different methods. As we can observe, RPNet obtains the worst accuracy. The reason for this may be that it is difficult to learn effective features for HSI with many details when the training sample is small. Compared with other methods, the proposed SPCM has significant improvement with respect to OA, AA, and Kappa. This also demonstrates the advantage of the joint learning of spectral-spatial correlation. Finally, the full classification maps of different methods are also shown in Figure 9. We can find that the result achieved by the proposed method is the best, as the details of HSI are well preserved.
In summary, as shown by the experimental results, the SPCM method proposed in this paper performs better than other state-of-the-art methods on three typical data sets. This section demonstrates that the spectral and spatial correlation is very important for HSI classification.

5. Conclusions

In this paper, an effective spectral-spatial feature extraction method called SPCM is proposed for HSI classification. It can not only characterize the non-linear relationship between different bands based on the kernel method, but also extract spatial structure features using the local neighbourhood construction method. The experimental results on three publicly available data sets have verified that the proposed SPCM-based classification method performs better than other base line methods, especially when the training set is small. The proposed method has achieved good results in respect to different indicators and visual effect. However, how to better combine spatial information and spectral information is still an open problem. In the future, we will further explore how to better learn spectral-spatial features to achieve better classification results.

Author Contributions

Conceptualization, W.C. and Y.W.; methodology, Y.W. and G.Z.; software, G.Z.; validation, G.Z., W.C., and Y.W.; formal analysis, G.Z., W.C. and Y.W.; investigation, G.Z., W.C., and Y.W.; resources, W.C. and Y.W.; data curation, G.Z., W.C., and Y.W.; writing—original draft preparation, G.Z.; writing—review and editing, W.C. and Y.W.; visualization, W.C.; supervision, W.C. and Y.W.; project administration, Y.W.; funding acquisition, Y.W. All authors have read and agreed to the published version of the manuscript.

Funding

This work was supported in part by the Fundamental Research Funds for the Central Universities under Grants CCNU20TD005 and CCNU22QN013, and Knowledge Innovation Project of Wuhan under Grant 2022010801010274.

Institutional Review Board Statement

Not applicable.

Informed Consent Statement

Not applicable.

Data Availability Statement

The data that support the findings of this study are available from the corresponding author upon reasonable request.

Conflicts of Interest

The authors declare no conflict of interest.

References

  1. Wan, S.; Gong, C.; Zhong, P.; Du, B.; Zhang, L.; Yang, J. Multi-scale Dynamic Graph Convolutional Network for Hyperspectral Image Classification. IEEE Trans. Geosci. Remote Sens. 2019, 58, 3162–3177. [Google Scholar] [CrossRef] [Green Version]
  2. Li, S.; Song, W.; Fang, L.; Chen, Y.; Benediktsson, J.A. Deep Learning for Hyperspectral Image Classification: An Overview. IEEE Trans. Geosci. Remote Sens. 2019, 57, 6690–6709. [Google Scholar] [CrossRef] [Green Version]
  3. Jia, S.; Jiang, S.; Lin, Z.; Li, N.; Yu, S. A Survey: Deep Learning for Hyperspectral Image Classification with Few Labeled Samples. Neurocomputing 2021, 448, 179–204. [Google Scholar] [CrossRef]
  4. Teke, M.; Deveci, H.S.; Haliloğlu, O.; Gürbüz, S.; Sakarya, U. A short survey of hyperspectral remote sensing applications in agriculture. In Proceedings of the International Conference on Recent Advances in Space Technologies, Istanbul, Turkey, 12–14 June 2013. [Google Scholar]
  5. Bioucas-Dias, J.M.; Plaza, A.; Camps-Valls, G.; Scheunders, P.; Nasrabadi, N.; Chanussot, J. Hyperspectral Remote Sensing Data Analysis and Future Challenges. IEEE Geosci. Remote Sens. Mag. 2013, 1, 6–36. [Google Scholar] [CrossRef] [Green Version]
  6. Uzkent, B.; Rangnekar, A.; Hoffman, M.J. Aerial Vehicle Tracking by Adaptive Fusion of Hyperspectral Likelihood Maps; IEEE: Piscataway, NJ, USA, 2017. [Google Scholar]
  7. Pan, B.; Shi, Z.; Xu, X. MugNet: Deep learning for hyperspectral image classification using limited samples. ISPRS J. Photogramm. Remote Sens. 2017, 145, 108–119. [Google Scholar] [CrossRef]
  8. Jiang, J.; Ma, J.; Wang, Z.; Chen, C.; Liu, X. Hyperspectral Image Classification in the Presence of Noisy Labels. IEEE Trans. Geosci. Remote Sens. 2019, 57, 851–865. [Google Scholar] [CrossRef] [Green Version]
  9. Paoletti, M.E.; Haut, J.M.; Plaza, J.; Plaza, A. Deep learning classifiers for hyperspectral imaging: A review. ISPRS J. Photogramm. Remote Sens. 2019, 158, 279–317. [Google Scholar] [CrossRef]
  10. Melgani, F.; Bruzzone, L. Classification of hyperspectral remote sensing images with support vector machines. IEEE Trans. Geosci. Remote Sens. 2004, 42, 1778–1790. [Google Scholar] [CrossRef] [Green Version]
  11. Zhong, Y.; Zhang, L. An Adaptive Artificial Immune Network for Supervised Classification of Multi-/Hyperspectral Remote Sensing Imagery. IEEE Trans. Geosci. Remote Sens. 2012, 50, 894–909. [Google Scholar] [CrossRef]
  12. Li, J.; Bioucas-Dias, J.M.; Plaza, A. Semisupervised Hyperspectral Image Classification Using Soft Sparse Multinomial Logistic Regression. IEEE Geosci. Remote Sens. Lett. 2013, 10, 318–322. [Google Scholar]
  13. Licciardi, G.; Marpu, P.R.; Chanussot, J.; Benediktsson, J.A. Linear Versus Nonlinear PCA for the Classification of Hyperspectral Data Based on the Extended Morphological Profiles. IEEE Geosci. Remote Sens. Lett. 2012, 9, 447–451. [Google Scholar] [CrossRef] [Green Version]
  14. Zhang, C.; Zheng, Y. Hyperspectral remote sensing image classification based on combined SVM and LDA. Proc. SPIE Int. Soc. Opt. Eng. 2014, 9263, 462–468. [Google Scholar]
  15. Hsieh, P.F.; Yang, M.H.; Yi-Jay, G.U.; Liang, Y.C. Classification-Oriented Locally Linear Embedding. Int. J. Pattern Recognit. Artif. Intell. 2010, 24, 737–762. [Google Scholar] [CrossRef]
  16. Villa, A.; Chanussot, J.; Jutten, C.; Benediktsson, J.A. On the use of ICA for hyperspectral image analysis. In Proceedings of the IEEE International Geoscience & Remote Sensing Symposium, Cape Town, South Africa, 12–17 July 2009. [Google Scholar]
  17. Green, A.A.; Berman, M.; Switzer, P.; Craig, M. A transformation for ordering multispectral data in terms of image quality with implications for noise removal. IEEE Trans. Geosci. Remote Sens. 1988, 26, 65–74. [Google Scholar] [CrossRef] [Green Version]
  18. Uddin, M.P.; Mamun, M.A.; Afjal, M.I.; Hossain, M.A. Information-theoretic feature selection with segmentation-based folded principal component analysis (PCA) for hyperspectral image classification. Int. J. Remote Sens. 2021, 42, 286–321. [Google Scholar] [CrossRef]
  19. Uddin, M.P.; Mamun, M.A.; Hossain, M.A. PCA-based Feature Reduction for Hyperspectral Remote Sensing Image Classification. IETE Tech. Rev. 2020, 38, 377–396. [Google Scholar] [CrossRef]
  20. Plaza, A.; Plaza, J.; Martin, G. Incorporation of spatial constraints into spectral mixture analysis of remotely sensed hyperspectral data. In Proceedings of the IEEE International Workshop on Machine Learning for Signal Processing, Grenoble, France, 2–4 September 2009. [Google Scholar]
  21. Shenming, Q.; Xiang, L.; Zhihua, G. A new hyperspectral image classification method based on spatial-spectral features. Sci. Rep. 2022, 12, 1541. [Google Scholar] [CrossRef]
  22. Falco, N.; Benediktsson, J.A.; Bruzzone, L. Spectral and Spatial Classification of Hyperspectral Images Based on ICA and Reduced Morphological Attribute Profiles. IEEE Trans. Geosci. Remote Sens. 2015, 53, 6223–6240. [Google Scholar] [CrossRef] [Green Version]
  23. Jia, S.; Shen, L.; Li, Q. Gabor Feature-Based Collaborative Representation for Hyperspectral Imagery Classification. IEEE Trans. Geosci. Remote Sens. 2015, 53, 1118–1129. [Google Scholar]
  24. Li, W.; Chen, C.; Su, H.; Du, Q. Local Binary Patterns and Extreme Learning Machine for Hyperspectral Imagery Classification. IEEE Trans. Geosci. Remote Sens. 2015, 53, 3681–3693. [Google Scholar] [CrossRef]
  25. Fang, L.; Li, S.; Duan, W.; Ren, J.; Benediktsson, J.A. Classification of hyperspectral images by exploiting spectral-spatial information of superpixel via multiple kernels. IEEE Trans. Geosci. Remote Sens. 2015, 53, 6663–6674. [Google Scholar] [CrossRef] [Green Version]
  26. Zhan, T.; Lu, Z.; Wan, M.; Yang, G. Multiscale Superpixel Kernel-Based Low-Rank Representation for Hyperspectral Image Classification. IEEE Geosci. Remote Sens. Lett. 2019, 17, 1642–1646. [Google Scholar] [CrossRef]
  27. Huang, H.; Duan, Y.; He, H.; Shi, G. Local Linear Spatial–Spectral Probabilistic Distribution for Hyperspectral Image Classification. IEEE Trans. Geosci. Remote Sens. 2020, 58, 1259–1272. [Google Scholar] [CrossRef]
  28. Kang, X.; Li, C.; Li, S.; Lin, H. Classification of Hyperspectral Images by Gabor Filtering Based Deep Network. IEEE J. Sel. Top. Appl. Earth Obs. Remote Sens. 2017, 11, 1166–1178. [Google Scholar] [CrossRef]
  29. Hang, R.; Liu, Q.; Hong, D.; Ghamisi, P. Cascaded Recurrent Neural Networks for Hyperspectral Image Classification. IEEE Trans. Geosci. Remote Sens. 2019, 57, 5384–5394. [Google Scholar] [CrossRef] [Green Version]
  30. Chen, Y.; Jiang, H.; Li, C.; Jia, X.; Ghamisi, P. Deep Feature Extraction and Classification of Hyperspectral Images Based on Convolutional Neural Networks. IEEE Trans. Geosci. Remote Sens. 2016, 54, 6232–6251. [Google Scholar] [CrossRef] [Green Version]
  31. Zhong, P.; Gong, Z.; Li, S.; Schonlieb, C.B. Learning to Diversify Deep Belief Networks for Hyperspectral Image Classification. IEEE Trans. Geoence Remote Sens. 2017, 55, 3516–3530. [Google Scholar] [CrossRef]
  32. Liu, Q.; Xiao, L.; Yang, J.; Wei, Z. CNN-Enhanced Graph Convolutional Network With Pixel- and Superpixel-Level Feature Fusion for Hyperspectral Image Classification. IEEE Trans. Geosci. Remote Sens. 2020, 59, 8657–8671. [Google Scholar] [CrossRef]
  33. Yang, X.; Ye, Y.; Li, X.; Lau, R.; Huang, X. Hyperspectral Image Classification With Deep Learning Models. IEEE Trans. Geosci. Remote Sens. 2018, 56, 5408–5423. [Google Scholar] [CrossRef]
  34. Lin, Z.; Chen, Y.; Ghamisi, P.; Benediktsson, J.A. Generative Adversarial Networks for Hyperspectral Image Classification. IEEE Trans. Geosci. Remote Sens. 2018, 56, 5046–5063. [Google Scholar]
  35. Mou, L.; Ghamisi, P.; Zhu, X.X. Deep Recurrent Neural Networks for Hyperspectral Image Classification. IEEE Trans. Geosci. Remote Sens. 2017, 55, 3639–3655. [Google Scholar] [CrossRef] [Green Version]
  36. Yu, L.; Xie, J.; Chen, S.; Zhu, L. Generating labeled samples for hyperspectral image classification using correlation of spectral bands. Front. Comput. Sci. 2016, 10, 292–301. [Google Scholar] [CrossRef]
  37. Fang, L.; He, N.; Li, S.; Plaza, A.J.; Plaza, J. A New Spatial–Spectral Feature Extraction Method for Hyperspectral Images Using Local Covariance Matrix Representation. IEEE Trans. Geosci. Remote Sens. 2018, 56, 3534–3546. [Google Scholar] [CrossRef]
  38. Liu, W.; Pokharel, P.P.; Principe, J.C. Correntropy: Properties and Applications in Non-Gaussian Signal Processing. IEEE Trans. Signal Process. 2007, 55, 5286–5298. [Google Scholar] [CrossRef]
  39. Zhang, X.; Wei, Y.; Cao, W.; Yao, H.; Peng, J.; Zhou, Y. Local Correntropy Matrix Representation for Hyperspectral Image Classification. IEEE Trans. Geosci. Remote Sens. 2022, 60, 1–13. [Google Scholar] [CrossRef]
  40. Arsigny, V.; Fillard, P.; Pennec, X.; Ayache, N. Geometric Means in a Novel Vector Space Structure on Symmetric Positive-Definite Matrices. Siam J. Matrix Anal. Appl. 2011, 29, 328–347. [Google Scholar] [CrossRef] [Green Version]
  41. Davis, L.S. Covariance discriminative learning: A natural and efficient approach to image set classification. In Proceedings of the Proceedings of the 2012 IEEE Conference on Computer Vision and Pattern Recognition (CVPR), Providence, RI, USA, 16–21 June 2012. [Google Scholar]
  42. Xu, Y.; Du, B.; Zhang, F.; Zhang, L. Hyperspectral image classification via a random patches network. ISPRS J. Photogramm. Remote Sens. 2018, 142, 344–357. [Google Scholar] [CrossRef]
  43. Pu, H.; Chen, Z.; Wang, B.; Jiang, G. A Novel Spatial–Spectral Similarity Measure for Dimensionality Reduction and Classification of Hyperspectral Imagery. IEEE Trans. Geosci. Remote Sens. 2014, 52, 7008–7022. [Google Scholar]
Figure 1. Process of the SPCM method. Each pixel of HSI generates nine windows, and then the optimal one is selected. The red point is the pixel to be processed, and is collectively referred to as the target pixel in the following text. A correntropy matrix is constructed from all pixels in this optimal window. The input to the SPCM method is a single pixel and the output is a correntropy matrix.
Figure 1. Process of the SPCM method. Each pixel of HSI generates nine windows, and then the optimal one is selected. The red point is the pixel to be processed, and is collectively referred to as the target pixel in the following text. A correntropy matrix is constructed from all pixels in this optimal window. The input to the SPCM method is a single pixel and the output is a correntropy matrix.
Applsci 12 06797 g001
Figure 2. Window definition for a single pixel. Specify ( x , y ) as the coordinates of the target pixel (red point), and ( X , Y ) as the centre coordinates of the side window. First, the coordinates of the centre pixel are calculated from the coordinates of the target pixels. The illustrated example is ( X , Y ) = ( x + r , y + r ) . The calculation formula of the centre position ( X , Y ) coordinates of different side windows is different. Then, the shape of the side window is delineated by the central pixel, the horizontal range is from X r to X + r , and the vertical direction is from Y r to Y + r . This results in a rectangular side window.
Figure 2. Window definition for a single pixel. Specify ( x , y ) as the coordinates of the target pixel (red point), and ( X , Y ) as the centre coordinates of the side window. First, the coordinates of the centre pixel are calculated from the coordinates of the target pixels. The illustrated example is ( X , Y ) = ( x + r , y + r ) . The calculation formula of the centre position ( X , Y ) coordinates of different side windows is different. Then, the shape of the side window is delineated by the central pixel, the horizontal range is from X r to X + r , and the vertical direction is from Y r to Y + r . This results in a rectangular side window.
Applsci 12 06797 g002
Figure 3. Let W t ( t = 1 , 2 , , 9 ) be one of the windows, and the figure shows a total of nine windows. Different black centre pixels ( X , Y ) are calculated based on the red target pixels, and then the window is determined by the centre pixels. The range of the window is horizontal reverse X r to X + r , vertical direction Y r to Y + r . (a) The target pixel is in the center of the window; (b) The target pixels are in the lower middle and upper middle of the window; (c) The target pixels are in the right middle and left middle of the window; (d) The target pixels are in the lower right corner, lower left corner, upper right corner, and upper left corner of the window respectively.
Figure 3. Let W t ( t = 1 , 2 , , 9 ) be one of the windows, and the figure shows a total of nine windows. Different black centre pixels ( X , Y ) are calculated based on the red target pixels, and then the window is determined by the centre pixels. The range of the window is horizontal reverse X r to X + r , vertical direction Y r to Y + r . (a) The target pixel is in the center of the window; (b) The target pixels are in the lower middle and upper middle of the window; (c) The target pixels are in the right middle and left middle of the window; (d) The target pixels are in the lower right corner, lower left corner, upper right corner, and upper left corner of the window respectively.
Applsci 12 06797 g003
Figure 4. Image display is an HSI classification method based on SPCM. First, we perform dimensionality reduction of a hyperspectral image based on MNF. Then we convert each pixel to a correntropy matrix in turn by SPCM. Finally, all the matrices are input to SVM for classification.
Figure 4. Image display is an HSI classification method based on SPCM. First, we perform dimensionality reduction of a hyperspectral image based on MNF. Then we convert each pixel to a correntropy matrix in turn by SPCM. Finally, all the matrices are input to SVM for classification.
Applsci 12 06797 g004
Figure 5. Details of Pavia University, KSC, and Botswana.
Figure 5. Details of Pavia University, KSC, and Botswana.
Applsci 12 06797 g005
Figure 6. Effects of different L, S and K on the OA, AA, and Kappa on KSC dataset. (a) Setting S and K to 0.8 L 2 , the influence of the change of L on OA, AA, Kappa; (b) Setting L to 9, K equals S, the effect of S on OA, AA, Kappa; (c) After the values of L and S were determined, the effect of K change on OA, AA, and Kappa was determined.
Figure 6. Effects of different L, S and K on the OA, AA, and Kappa on KSC dataset. (a) Setting S and K to 0.8 L 2 , the influence of the change of L on OA, AA, Kappa; (b) Setting L to 9, K equals S, the effect of S on OA, AA, Kappa; (c) After the values of L and S were determined, the effect of K change on OA, AA, and Kappa was determined.
Applsci 12 06797 g006
Figure 7. Classification maps of different methods on the Pavia University data set. (a) False Colour Image; (b) Ground Truth; (c) Training set; (d) Test set; (e) LCMR; (f) SCMK; (g) CEGCN; (h) PRNet; (i) LCEM; (j) SPCM.
Figure 7. Classification maps of different methods on the Pavia University data set. (a) False Colour Image; (b) Ground Truth; (c) Training set; (d) Test set; (e) LCMR; (f) SCMK; (g) CEGCN; (h) PRNet; (i) LCEM; (j) SPCM.
Applsci 12 06797 g007
Figure 8. Classification maps of different methods on the KSC data set. (a) False Colour Image; (b) Ground Truth; (c) Training set (d) Test set; (e) LCMR; (f) SCMK; (g) CEGCN; (h) PRNet; (i) LCEM; (j) SPCM.
Figure 8. Classification maps of different methods on the KSC data set. (a) False Colour Image; (b) Ground Truth; (c) Training set (d) Test set; (e) LCMR; (f) SCMK; (g) CEGCN; (h) PRNet; (i) LCEM; (j) SPCM.
Applsci 12 06797 g008
Figure 9. Classification maps of different methods on the Botswana data set. (a) False Colour Image; (b) Ground Truth; (c) Training set; (d) Test set; (e) LCMR; (f) SCMK; (g) CEGCN; (h) PRNet; (i) LCEM; (j) SPCM.
Figure 9. Classification maps of different methods on the Botswana data set. (a) False Colour Image; (b) Ground Truth; (c) Training set; (d) Test set; (e) LCMR; (f) SCMK; (g) CEGCN; (h) PRNet; (i) LCEM; (j) SPCM.
Applsci 12 06797 g009
Table 1. Classification results obtained by LCMR, SCMK, CEGCN, RPNet, LCEM, and SPCM using 0.1% of the available labelled data for training with the Pavia University data set. The best results are expressed in bold.
Table 1. Classification results obtained by LCMR, SCMK, CEGCN, RPNet, LCEM, and SPCM using 0.1% of the available labelled data for training with the Pavia University data set. The best results are expressed in bold.
No.LCMRSCMKCEGCNRPNetLCEMSPCM
175.5488.0996.9385.5887.6988.41
275.4883.9197.0695.7995.1393.15
364.4164.5085.7259.9971.8267.72
495.5882.9667.1245.2585.9887.12
592.4796.53100.0012.0591.7499.78
677.5063.0265.6142.8593.3295.53
788.8384.6259.597.9092.8197.45
871.7777.1653.6475.8663.6275.75
991.7070.7365.0924.9548.1749.70
OA77.6180.6385.1373.9688.0490.97
AA81.4779.0676.7550.0281.1483.82
Kappa71.4874.6379.8962.7884.2086.97
Table 2. Classification results obtained by LCMR, SCMK, CEGCN, RPNet, LCEM, and SPCM using 1% of the available labelled data for training with KSC data set. The best results are expressed in bold.
Table 2. Classification results obtained by LCMR, SCMK, CEGCN, RPNet, LCEM, and SPCM using 1% of the available labelled data for training with KSC data set. The best results are expressed in bold.
No.LCMRSCMKCEGCNRPNetLCEMSPCM
192.6986.9799.5385.1397.3394.65
291.5173.7970.4663.0780.8888.46
392.3984.3587.0894.0782.5387.55
485.9559.0827.0538.1578.3182.97
590.1953.3357.0768.5581.7678.49
691.3867.8879.8343.1778.1492.35
795.2088.6475.5029.8192.2398.35
887.7281.2480.4582.6795.2691.46
992.8990.4990.9589.3298.0097.39
1099.7083.3392.9964.0098.2597.02
1192.0394.2595.0881.6994.2591.88
1294.7487.7768.8198.8098.7797.85
13100.0099.60100.0099.46100.00100.00
OA93.1385.5485.2580.9793.8394.51
AA91.7980.8279.1172.1490.4492.34
Kappa92.1283.9183.4778.7893.1294.89
Table 3. Classification results obtained by LCMR, SCMK, CEGCN, RPNet, LCEM, and SPCM using 1% of the available labelled data for training with the Botswana data set. The best results are expressed in bold.
Table 3. Classification results obtained by LCMR, SCMK, CEGCN, RPNet, LCEM, and SPCM using 1% of the available labelled data for training with the Botswana data set. The best results are expressed in bold.
No.LCMRSCMKCEGCNRPNetLCEMSPCM
195.0286.39100.0061.8094.0199.79
284.5555.2976.2821.0087.47100.00
328.6088.55100.0061.6973.1579.26
496.2977.5575.2329.5874.2598.17
572.6683.3829.0584.5976.6282.79
671.5767.7856.6562.0357.3769.30
784.9089.73100.00100.0077.3088.80
861.9984.8571.5070.6577.2580.17
995.9589.8494.3887.4688.3579.18
1099.5985.8856.5599.1975.2778.96
1188.3186.88100.0072.1986.0187.45
1232.8987.2185.8793.3053.6959.96
1385.4978.0489.2783.7789.1392.33
1431.4975.1189.2412.7744.4760.52
OA76.6280.5680.4772.3477.2083.53
AA73.5281.1880.2967.1475.3182.68
Kappa74.6379.4578.8269.9575.2482.13
Publisher’s Note: MDPI stays neutral with regard to jurisdictional claims in published maps and institutional affiliations.

Share and Cite

MDPI and ACS Style

Zhang, G.; Cao, W.; Wei, Y. Spatial Perception Correntropy Matrix for Hyperspectral Image Classification. Appl. Sci. 2022, 12, 6797. https://doi.org/10.3390/app12136797

AMA Style

Zhang G, Cao W, Wei Y. Spatial Perception Correntropy Matrix for Hyperspectral Image Classification. Applied Sciences. 2022; 12(13):6797. https://doi.org/10.3390/app12136797

Chicago/Turabian Style

Zhang, Guochao, Weijia Cao, and Yantao Wei. 2022. "Spatial Perception Correntropy Matrix for Hyperspectral Image Classification" Applied Sciences 12, no. 13: 6797. https://doi.org/10.3390/app12136797

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop