**Journal of Computer and Communications**

Vol.03 No.11(2015), Article ID:61276,7 pages

10.4236/jcc.2015.311006

Semi-Supervised Dimensionality Reduction of Hyperspectral Image Based on Sparse Multi-Manifold Learning

Hong Huang^{1}, Fulin Luo^{1}, Zezhong Ma^{2}, Hailiang Feng^{1}

^{1}Key Laboratory of Optoelectronic Technique and Systems, Ministry of Education, Chongqing University, Chongqing, China

^{2}Chongqing Institute of Surveying and Planning for Land, Chongqing, China

Received October 2015

ABSTRACT

In this paper, we proposed a new semi-supervised multi-manifold learning method, called semi- supervised sparse multi-manifold embedding (S^{3}MME), for dimensionality reduction of hyperspectral image data. S^{3}MME exploits both the labeled and unlabeled data to adaptively find neighbors of each sample from the same manifold by using an optimization program based on sparse representation, and naturally gives relative importance to the labeled ones through a graph-based methodology. Then it tries to extract discriminative features on each manifold such that the data points in the same manifold become closer. The effectiveness of the proposed multi-manifold learning algorithm is demonstrated and compared through experiments on a real hyperspectral images.

**Keywords:**

Hyperspectral Image Classification; Dimensionality Reduction, Multiple Manifolds Structure, Sparse Representation, Semi-Supervised Learning

1. Introduction

Hyperspectral image (HSI) contains dozens or even hundreds of contiguous spectral bands, which has been widely used in land cover investigation [1]. However, the high dimensional characteristic of HSI will cause the curse of dimensionality [2]. Consequently, dimensionality reduction (DR) plays a critical role in HSI analysis, especially for the classification task when the number of available labeled training samples is limited.

In recent years, sparse representation (SR) has been successfully applied in HSI dimensionality reduction. SR aims at the sparse reconstructive weight which is associated with sample size. The representative methods include sparse preserving projection (SPP) [3], discriminative learning by sparse representation (DLSP) [4] and discriminant sparse neighborhood preserving embedding (DSNPE) [5].

However, these mentioned works implicitly assume that data points uniformly lie on a single manifold. In real applications, data points may lie on multiple manifolds. In view of this, Ehsan et al. [6] proposed an algorithm called sparse manifold clustering and embedding (SMCE) for simultaneous clustering and DR of data lying in multiple non-linear manifolds. While SMCE is also suffered from the out-of-sample problem, and they do not use the class information provided by training samples, which restricts their discriminating capability. Lu et al. [7] introduced a discriminative multi-manifold analysis method by learning discriminative features from image patches, which can perform well when label information is sufficient.

In the real world, the labeled samples are often very difficult and expensive to obtain. The supervised methods cannot work well when lack of training examples, in contrast, unlabeled examples can be easily obtained [8]. In such situations, it can be beneficial to incorporate the information which is contained in unlabeled samples into a learning problem, i.e., semi-supervised learning should be applied instead of supervised learning. At the same time, most of multi-manifold learning methods that have been applied to the processing of HSI rely either on supervised or unsupervised models, and only a few are focused on the semi-supervised setting.

To overcome the above drawbacks, we propose a new DR algorithm named semi-supervised sparse multi- manifold embedding (S^{3}MME) in this paper. S^{3}MME utilizes the merits of both sparsity property and multi- manifold learning to better characterize the discriminant property of the data. It exploits both the labeled and unlabeled pixels to adaptively find the local neighborhood of each data point by using an optimization program based on SR, and the selected neighbors are from the same manifold other than other manifolds. The weights associated to the chosen neighbors are automatically obtained simultaneously. It also exploits the wealth of labeled samples in HSI data, and naturally gives relative importance to the labeled ones following a semi-super- vised approach. Then, an objective function pushes the homogeneous samples closer to each other is proposed, and the classification performance is further improved.

2. Semi-Supervised Sparse Multi-Manifold Embedding

Suppose that a HSI data set sampled from different manifolds, the first n points are labeled and the rest N−n points are unlabeled. Let denote the class label of x_{i}. The goal of DR is to map by, where. For supervised methods, only labeled points are used for DR. While labeled and unlabeled samples are used for semi-supervised methods.

We assume that naturally occurring data have possibly much fewer degrees of freedom than what the ambient dimension would suggest. Thus, we consider the case where the data lies on or close to multiple low-dimensional manifolds reside in a high dimensional space. To model the manifold structure of data, a similarity graph should be constructed, where nodes represent the data points and edges represent the similarity between data points. Then, a key issue for the similarity graph is to decide which nodes should be connected and how.

To achieve optimal discriminant features, each point will be connected to the points from the same manifold with appropriate weights, while the data pairs from different manifolds are disconnected. At first, for labeled points, we select the data points which have the same class label. Then, we formulate an optimization algorithm as in SMCE to find the unlabeled neighbors from the same manifold. Based on SR techniques, it selects a few data points that are close to x_{i} and span a low-dimensional affine subspace passing near x_{i}. For labeled points, the points with the same class label and the unlabeled neighbors from the same manifold are used for the similarity graph. For unlabeled points, we try to find the neighbors which may come from the same manifold.

The sparse solutions can be obtained by SMCE [6], where both labeled and unlabeled points are used. This motivates that for each data point x_{i} to solve the following weighted sparse optimization program

(1)

where the ℓ_{1}-norm promotes sparsity of the solution, the proximity inducing matrix Q_{i} is a positive-definite diagonal matrix, and X_{i} denotes the matrix of normalized vectors {x_{j}− x_{i}}as follows:

(2)

The elements of Q_{i} should be chosen such that the points that are closer to x_{i} have smaller weights, allowing the assignment of non-zero coefficients to them. The diagonal elements of Q_{i} can be defined as

(3)

The SR of each data point can be used for the construction of graph. Since the non-zero elements of {c_{i}} are expected to come from the same manifold as of x_{i}. We first construct a sparse graph Gs (V, E, W_{s}) with vertex set V = {x_{i}, x_{2}, …, x_{N}}, edge set E, and symmetric weight matrix W_{s}. Then, we put an edge between nodes i and j if x_{i} and x_{j} are from the same class, or x_{i} or x_{j} is unlabeled but c_{ij} is a non-zero element.

Once the graph G_{s} is constructed, the affinity weight of x_{i} can be defined as

(4)

where β is a trade-off parameter to adjust the contribution of labeled and unlabeled data, and, t is the number of non-zero elements in c_{i}.

The obtained sparse graph built in this way has ideally several connected components, where points in the same manifold are connected to each other and there is no connection between two points from different manifolds. In other words, the weight matrix of this graph has the following form

(5)

where is the weight matrix of data in, which includes the weight of labeled and some unlabeled data.

The objective of S^{3}MME is embodied as that it minimizes the sum of distances between data pairs which are expected to come from the same manifold. Then, a reasonable criterion for choosing an optimal projection vector with stronger intra-manifold compactness on manifold is formulated as

(6)

where and are related to the points in the same manifold which are connected to each other. is a diagonal matrix with, and is the Laplacian matrix.

To remove an arbitrary scaling factor in embedding space, we impose a constraint to vectors on, and the objective function on can be recast as

(7)

Then we apply the Lagrange multipliers to Eq. 7, we can get

(8)

where is the generalized eigenvector of and.

Let the column vectors be the solutions of Eq. (8) ordered according to their eigenvalues. Thus, the embedding is given as follows:

(9)

Thus, we can use different weight matrix of the i-th manifold as a similarity between points in the corresponding manifold, and obtain a low-dimensional embedding of the data points.

3. Experiments and Discussion

3.1. Experimental Design

The goal of the experiments is to investigate the effectiveness of the proposed algorithms for classification of PaviaU hyperspectral data set. In each experiment, the data set was divided into training and test sets, and we randomly split the training set into the labeled and unlabeled set. The number of labeled samples l is varied as {10, 20, 40, 80} per class, while the number of unlabeled samples u is chosen in {100, 500, 1000, 2000, 3000}. For supervised DR methods, only the labeled set is used for training, while semi-supervised DR methods can utilize both labeled and unlabeled data. Then, all testing samples are projected into embedding space.

After that, reconstruction err classifier is used for multiple manifold classification [9], and nearest neighborhood (1-NN) is employed for other methods classification in all experiments. The classifier was evaluated against the test set, and we use overall classification accuracies (OAs) and the kappa coefficients () to evaluate the classification results. We repeat the classification process 10 times in each condition.

We compare S^{3}MME with several representative DR algorithms such as PCA, LDA, LPP, LFDA, SPP, DLSP, DSNPE, semi-supervised sub-manifold discriminant analysis (S^{3}MPE) [10] and semi-supervised discriminant analysis (SDA) [11]. The parameter β is set as 40 with cross-validation in S^{3}MME. Note that for SPP, DLSP and DSNPE is generally fixed across various instances of the problem, and we empirically set it to 0.1 in our experiments. For manifold learning methods, the number of nearest neighbors k is set to be 7. For all methods, the dimension of embedding features is set as 30.

3.2. Classification of PaviaU Data Set

The PaviaU dataset was acquired by the reflective optics system imaging spectrometer (ROSIS) sensor during flight campaigns in 2003 over the Pavia University, northern Italy. It consists of 610 × 340 pixels and 115 spectral reflectance bands in the wavelength range 430-860 nm with a high spatial resolution of 1.3 m per pixel. After removing noisy and water absorption bands, it reduced to 103 channels. Figure 1(a) shows a false color composite of the image, and Figure 1(b) shows the nine ground truth classes of interest.

In the first experiment, we evaluated the classification performance of S^{3}MME. Table 1 reports the average classification performance achieved by different DR algorithms, where the OAs and the κ coefficients are displayed for a varying number of labeled samples (10, 20, 40, 80 per class) and 3000 unlabeled samples. For illustrative purposes, Figure 1 shows the classification maps obtained for different methods for the case of l = 40 per class and u = 3000 as in Table 1.

Table 1. OAs (in percent) and κ coefficients (in the brackets) with different numbers of labeled samples per class.

Figure 1. Classification maps with different algorithms (l = 40, u = 3000). (a) False-color composition (bands 5, 40 and 15 for RGB); (b) Ground truth; (c) Labeled data; (d) Original (0.7411); (e) PCA (0.7407); (f) LDA (0.7527); (g) LPP (0.7746); (h) LFDA (0.6969); (i) SPP (0.7589); (j) DLSP (0.7244); (k) DSNPE (0.7410); (l) S3MPE (0.7453); (m) SDA (0.7862); (n) S3MME (0.8491). Note that OAs are given in parentheses.

As can be seen from Table 1, the classification performance improves for all methods as more data samples are used for training. The supervised methods, i.e. LDA and LFDA, degrade the performance of classification accuracy when the labeled sample size of training set is small, mainly due to the overfitting or overtraining. Our S^{3}MME method produces better classification results than other methods in all situations, and the improvement is particularly significant when low number of labeled samples is used. The reason is that S^{3}MME exploits the wealth of labeled and unlabeled samples to adaptively select neighbors from the same manifold and discovers the multi-manifold structure in HSI data, which respects both sparsity property and semi-supervised learning to better characterize the discriminant property of data.

By observing the classification results shown in Figure 1, the numerical results are confirmed by visual inspection of the classification maps. The S^{3} MME method produces more homogenous areas and better classification maps than other methods, especially in Meadows, Trees, and Bare Soil.

To investigate the influence of the numbers of unlabeled data on the performance of S3MME, we evaluate the classification accuracy using a small number of labeled samples (l = 10 per class) and different numbers of unlabeled samples (u = 100, 500, 1000, 2000, 3000), which are randomly selected for training. The classification results averaged over ten runs are shown in Figure 2.

An effective semi-supervised learning method can improve the performance when the number of the available unlabeled data increases. As expected, the OAs and κ coefficients of S^{3}MME is significantly improved when the number of the available unlabeled data increases.

4. Experiments and Discussion

In this paper, we proposed a novel multi-manifold learning method for DR and classification of hyperspectral image. S^{3}MME exploits both the labeled and unlabeled samples to selected neighbors from the same manifold based on sparse representation, and naturally gives relative importance to the labeled ones through a semi-su- pervised neighborhood graph. Then it tries to extract discriminative features on each manifold such that the

Figure 2. The classification results of S^{3}MME using different numbers of unlabeled samples.

data points in the same manifold become closer. The S^{3}MME method has been applied to a real hyperspectral image data set, and extensive experiments and comparisons have been conducted. Promising results have been obtained demonstrating the superiority of the proposed multi-manifold learning methods in hyperspectral image classification. The multi-manifold learning method proposed here are not utilize any spatial information in hyperspectral image. In our future work, we are going to consider the multi-manifold learning methods which make full use of both the spectral and spatial information provided by hyperspectral image.

Acknowledgements

This work is supported by National Science Foundation of China (41371338, 61101168), the Basic and Frontier Research Programs of Chongqing (cstc2013jcyjA40005), the Fundamental Research Funds for the Central Universities of China (106112013CDJZR125501, 1061120131204), and Chongqing University Postgraduates’ Innovation Project (CYB15052). The authors would like to thank the anonymous reviewers for their constructive advice.

Cite this paper

Hong Huang,Fulin Luo,Zezhong Ma,Hailiang Feng, (2015) Semi-Supervised Dimensionality Reduction of Hyperspectral Image Based on Sparse Multi-Manifold Learning. *Journal of Computer and Communications*,**03**,33-39. doi: 10.4236/jcc.2015.311006

References

- 1. Chen, Y., Nasrabadi, N. and Tran, T. (2013) Hyperspectral Image Classification via Kernel Sparse Representation. IEEE Transactions on Geoscience Remote Sensing, 51, 217-231. http://dx.doi.org/10.1109/TGRS.2012.2201730
- 2. Shao, Z. and Zhang, L. (2014) Sparse Dimensionality Reduction of Hyperspectral Image Based on Semi-Supervised Local Fisher Discriminant Analysis. International Journal of Applied Earth Observation and Geoinformation, 31, 122-129. http://dx.doi.org/10.1016/j.jag.2014.03.015
- 3. Qiao, L., Chen, S. and Tan, X. (2010) Sparsity Preserving Projections with Applications to Face Recognition. Pattern Recognition, 43, 331-341. http://dx.doi.org/10.1016/j.patcog.2009.05.005
- 4. Zang, F. and Zhang, J.S. (2011) Discriminative Learning by Sparse Representation for Classification. Neurocomputing, 74, 2176-2183. http://dx.doi.org/10.1016/j.neucom.2011.02.012
- 5. Lu, G.F., Jin, Z. and Zou, J. (2012) Face Recognition Using Discriminant Sparsity Neighborhood Preserving Embedding. Knowledge-Based Systems, 31, 119-127. http://dx.doi.org/10.1016/j.knosys.2012.02.014
- 6. Elhamifar, E. and Vidal, R. (2006) Sparse Manifold Clustering and Embedding. In: Advances in Neural Information Processing Systems, 609-616.
- 7. Lu, J.W., Peng, Y. and Wang, G. (2011) Discriminative Multi-Manifold Analysis for Face Recognition from a Single Training Sample per Person. IEEE International Conference on Computer Vision, 1943-1950. http://dx.doi.org/10.1109/iccv.2011.6126464
- 8. Shi, Q., Zhang, L.P. and Du, B. (2013) Semisupervised Discrimina-tive Locally Enhanced Alignment for Hyperspetral Image Classification. IEEE Transactions on Geoscience Remote Sensing, 51, 4800-4815. http://dx.doi.org/10.1109/TGRS.2012.2230445
- 9. Wang, L.Z., Huang, H. and Feng, H.L. (2012) Multi-Linear Local and Global Preserving Embedding and Its Application in Hyperspectral Remote Sensing Image Classification. Journal of Computer-Aided Design & Computer Graphics, 24, 780-786.
- 10. Song, Y.Q., Nie, F.P. and Zhang, C.S. (2008) Semi-Supervised Sub-Manifold Discriminant Analysis. Pattern Recognition, 29, 1806-1813. http://dx.doi.org/10.1016/j.patrec.2008.05.024
- 11. Cai, D., He, X.F. and Han, J.W. (2007) Semi-Supervised Discriminant Analysis. In: Proceedings of the 11th IEEE International Conference on Computer Vision, 609-616. http://dx.doi.org/10.1109/iccv.2007.4408856