ArticlePDF Available

Unsupervised Dimensionality Reduction for Hyperspectral Imagery via Local Geometric Structure Feature Learning

Authors:

Abstract

Hyperspectral images (HSIs) possess a large number of spectral bands, which easily lead to the curse of dimensionality. To improve the classification performance, a huge challenge is how to reduce the number of spectral bands and preserve the valuable intrinsic information in the HSI. In this letter, we propose a novel unsupervised dimensionality reduction method called local neighborhood structure preserving embedding (LNSPE) for HSI classification. At first, LNSPE reconstructs each sample with its spectral neighbors and obtains the optimal weights for constructing the adjacency graph by modifying its loss function. Then, to discover the scatter information of the training samples, LNSPE minimizes the scatter between the pixels and the corresponding neighbors and maximizes the total scatter of the HSI data. Finally, it incorporates the scatter information and the dual graph structure to enhance the aggregation of the HSI. As a result, LNSPE can effectively reveal the intrinsic structure and improve the classification performance of the HSI data. The experimental results on two real hyperspectral data sets exhibit the efficiency and superiority of LNSPE to some state-of-the-art methods.
This article has been accepted for inclusion in a future issue of this journal. Content is final as presented, with the exception of pagination.
IEEE GEOSCIENCE AND REMOTE SENSING LETTERS 1
Unsupervised Dimensionality Reduction
for Hyperspectral Imagery via Local
Geometric Structure Feature Learning
Guangyao Shi, Hong Huang ,Member, IEEE, and Lihua Wang
Abstract Hyperspectral images (HSIs) possess a large num-
ber of spectral bands, which easily lead to the curse of dimen-
sionality. To improve the classification performance, a huge
challenge is how to reduce the number of spectral bands and
preserve the valuable intrinsic information in the HSI. In this
letter, we propose a novel unsupervised dimensionality reduction
method called local neighborhood structure preserving embed-
ding (LNSPE) for HSI classification. At first, LNSPE reconstructs
each sample with its spectral neighbors and obtains the optimal
weights for constructing the adjacency graph by modifying its loss
function. Then, to discover the scatter information of the training
samples, LNSPE minimizes the scatter between the pixels and
the corresponding neighbors and maximizes the total scatter of
the HSI data. Finally, it incorporates the scatter information and
the dual graph structure to enhance the aggregation of the HSI.
As a result, LNSPE can effectively reveal the intrinsic structure
and improve the classification performance of the HSI data. The
experimental results on two real hyperspectral data sets exhibit
the efficiency and superiority of LNSPE to some state-of-the-art
methods.
Index Terms—Dimension ality reduction (DR), adjacency
graph, hyperspectral images (HSIs), intrinsic structure, scatter
information.
I. INTRODUCTION
HYPERSPECTRAL images (HSIs) are captured by satel-
liteborne and airborne sensors in hundreds of spectral
bands, and each pixel can be represented as a high-dimensional
vector [1], [2]. Although abundant spectral information is
beneficial to improve the classification performance, it will
cause the increase in computational complexity, which requires
huge computational resources and storage capacity [3], [4].
Furthermore, the high-dimensional characteristic of the HSI
data often cause the Hughes phenomena, especially when there
are only few training samples available [5]. Therefore, it is an
urgent task to reduce the dimensionality of the HSI data while
preserving the useful intrinsic information.
Manuscript received April 24, 2019; revised July 16, 2019; accepted
September 15, 2019. This work was supported in part by the
Basic and Frontier Research Programmes of Chongqing under Grant
cstc2018jcyjAX0093 and Grant cstc2018jcyjAX0633, in part by the
Chongqing University Postgraduates Innovation Project under Grant
CYB18048 and Grant CYS18035, and in part by the National
Science Foundation of China under Grant 41371338. (Corresponding author:
Hong Huang.)
The authors are with the Key Laboratory of Optoelectronic Tech-
nology and Systems of the Education Ministry of China, Chongqing
University, Chongqing 400044, China (e-mail: shiguangyao@cqu.edu.cn;
hhuang@cqu.edu.cn; 20170802019t@cqu.edu.cn).
Color versions of one or more of the figures in this letter are available
online at http://ieeexplore.ieee.org.
Digital Object Identifier 10.1109/LGRS.2019.2944970
Dimensionality reduction (DR) serves as an effective tech-
nique to address the aforementioned issues, and it aims to
find a low-dimensional space where some desired properties
can be preserved. In recent years, many effective DR meth-
ods have been developed for classification [6], [7]. Principal
component analysis (PCA) and linear discriminant analy-
sis (LDA) are two well-known methods based on subspace
learning, PCA aims at finding a projection direction along
which the data have the maximum variance [8], while LDA
seeks for the optimal projection matrix by maximizing the
between-class variance and minimizing the within-class vari-
ance in the low-dimensional space [9]. Although these DR
methods proved to be effective for classification, they do not
consider the manifold structure on which high-dimensional
data may possibly reside [10], [11].
Recently, various manifold learning methods have been
developed for DR, including local linear embedding (LLE)
[12], Laplacian eigenmaps (LEs) [13], and local tangent
space alignment (LTSA) [14]. LLE assumes that the global
manifold can be reconstructed by several small overlapped
regions, and it performs linearization to reconstruct the local
properties of data point by its neighbors. LE builds a graph
incorporating neighborhood information of data and computes
a low-dimensional representation by optimally preserving local
neighborhood information. LTSA represents the local geome-
try of the manifold using tangent spaces learned by fitting an
affine subspace in a neighborhood of each data point. Owing
to their nonlinear characteristic, these methods suffer from the
problem of out-of-sample, for without project vector during the
DR process.
To tackle this problem, many linear manifold learning
methods were proposed, such as neighborhood preserving
embedding (NPE) [15], locality preserving projections (LPP)
[16], and linear LTSA (LLTSA) [17]. However, they cannot
effectively reveal the structural relationships of pairwise neigh-
bors, which limits their discriminating ability for land use clas-
sification. To unify the above DR methods, a graph-embedding
(GE) framework was proposed to describe many existing DR
techniques [18]. Sparse manifold embedding (SME) [19] uses
the sparse coefficients to construct a similarity graph and pre-
serves this sparse similarity in embedding space, which further
improves the classification performance of the HSI. Graph-
based discriminate analysis via spectral similarity (GDA-SS)
[20] uses the spectral difference of pairwise pixels and sets a
threshold to evaluate the similarity, and it can effectively reveal
the discriminant manifold structure of the data. However,
it requires the prior label information of samples, which limits
their application in certain scenes.
1545-598X © 2019 IEEE. Personal use is permitted, but republication/redistribution requires IEEE permission.
See http://www.ieee.org/publications_standards/publications/rights/index.html for more information.
Authorized licensed use limited to: Hong Huang. Downloaded on February 15,2020 at 15:54:38 UTC from IEEE Xplore. Restrictions apply.
This article has been accepted for inclusion in a future issue of this journal. Content is final as presented, with the exception of pagination.
2IEEE GEOSCIENCE AND REMOTE SENSING LETTERS
To address the aforementioned problems, a new unsuper-
vised DR method called the local neighborhood structure
preserving embedding (LNSPE) algorithm was proposed for
HSI classification. At first, LNSPE reconstructs each sample
with its spectral neighbors, and it designs a novel loss function
to obtain the optimal weights for constructing an adjacency
graph. Then, it reveals the scatter information of training
samples by minimizing the local neighborhood scatter matrix
and maximizing the total scatter matrix of training samples
simultaneously. Finally, the scatter information and the dual
graph structure are integrated to enhance the aggregation
of the HSI data. As a result, LNSPE can extract effective
discriminant features and subsequently improve the classifi-
cation performance of the HSI. Experimental results show
that LNSPE achieved better performance on the PaviaU and
Botswana HSI data sets than some of the state-of-the-art DR
methods.
This letter is structured as follows. Section II gives
a detailed description of the proposed LNSPE method.
Section III presents the experimental results on the PaviaU
and Botswana HSI data sets to evaluate the effectiveness of
the LNSPE. Finally, Section IV concludes this letter and gives
some suggestions for future work.
II. PROPOSED METHOD
Suppose a data set is composed of npoints, ith point can
be denoted as xi∈
D,whereDis the band number. Let
(xi)∈{1,2,...,c}be the class label of xi,andcindicates
the number of land cover types in the HSI. The goal of the
linear DR methods is to obtain a projection vector V∈
D×d,
which can map X=[x1,x2,...,xn]∈
D×nto Y=
[y1,y2,...,yn]∈
d×n,whereYis the low-dimensional
representation of Xand dDis the embedding dimension in
low-dimensional space. With the projection vector V, we can
compute Yas Y=VTX.
A. Dual Structure Preserving Model
In the HSI, the similarity between different pixels is usually
measured by the spectral-domain Euclidean distance. Two
pixels with a small distance may have a large probability
belonging to the same class. To effectively use the spectral
neighborhood information in the HSI, we reconstruct each
sample with its knearest neighbors (NNs). Denote the kNNs
of xias S(xi)=[xi1,xi2,...,xik ], and the reconstructed
pixel x
ican be given as follows:
x
i=xjS(xi)vjxj
xjS(xi)vj
=k
m=1 vmxim
k
m=1 vm
(1)
where vmis the weight of xim, and it can be calculated as
vm=exp−||xixim||2
2t2
i(2)
where ti=(1/k)k
h=1 ||xixih|| is a kernel parameter.
After obtaining the reconstructed pixels, we have origin
pixels Xand reconstructed pixels X. Then, we can construct
two graphs G(X,W)and G(X,W),andXand Xserve
as the vertexes of graphs Gand G, respectively. Wis the
corresponding weight matrix in graphs Gand G. Denote wij
as the weight of the edge from node ito node j, and it can be
calculated by the following redefined loss function:
min
n
i=1
xi
k
j=1
wij xj+x
i
k
j=1
wij x
j
2
s.t.
k
j=1
wij =1 xjS(xi).
(3)
Denote the kth spectral neighbors of xias xk
i,andhk
i=
xixk
i+x
ixk
imeasures the spectral similarity between
xiand its kth spectral neighbor, and the objective function can
be simplified as
J(W)=min
n
i=1
xi
k
j=1
wij xj+x
i
k
j=1
wij x
j
2
=min
n
i=1
k
j=1
wij xixj+x
ix
j
2
=min
n
i=1
wT
iziwi(4)
where zi=[h1
i,h
2
i,h
3
i,...,h
k
i]T[h1
i,h
2
i,h
3
i,...,h
k
i]and
wi=[wi1,w
i2,w
i3,...,w
ik]. Then, (4) can be given as
follows:
min
n
i=1
wT
iziwi
s.t.
k
j=1
wij =1.
(5)
With the Lagrange multiplier method, wij can be reduced as
wij =k
m=1 zjm
i1
k
p=1 k
q=1 zpq
i1(6)
where zjm
i=(hj
i)Thm
iand zpq
i=(hp
i)Thq
i.
With the weight matrix W, the projection vector Vfor
lower dimensional embedding can be obtained by solving the
following optimization problem:
min J=min
n
i=1
yi
k
j=1
wij yij
2
=minVTXMXTV
s.t.
n
i=1
yi=0,1
nVVT=I
(7)
in which M=(IW)(IW)Tand I=diag(1,1,...,1).
(7) can be solved by the Lagrange multiplier, and it can be
transformed into the following form:
XMXTV=λXXTV.(8)
Authorized licensed use limited to: Hong Huang. Downloaded on February 15,2020 at 15:54:38 UTC from IEEE Xplore. Restrictions apply.
This article has been accepted for inclusion in a future issue of this journal. Content is final as presented, with the exception of pagination.
SHI et al.: UNSUPERVISED DR FOR HYPERSPECTRAL IMAGERY 3
B. Neighborhood Scatter Extraction Model
Considering that the pixels with similar spectrum are most
likely to belong to the same class, we use the scatter informa-
tion of each sample to learn the projection relationships from
high-dimensional space to lower-dimensional space.
Suppose the rNNs of xican be represented as N(xi)=
[xi1,xi2,…, xir], and then the scatter matrix between xiand
N(xi) can be expressed as follows:
hi=r
p=1 vp(xixip)(xixip )T
r
q=1 vq
(9)
where weight vpdefined in (2) measures the spectral similarity
between the neighboring pixels to the central pixel.
For all the pixels in the HSI, the scatter can be represented
as
H=
n
i=1
hi=n
i=1 r
p=1 vp(xixip)(xixip )T
r
q=1 vq
.(10)
In addition, the total scatter of training samples can be
defined as follows:
S=
n
i=1
(xi¯x)(xi¯x)T(11)
where ¯x is the mean of the training samples.
We try to seek a linear projection matrix such that the local
neighborhood preserving scatter is minimized, whereas the
total scatter is maximized in the embedding space. Therefore,
the optimal projection Vcan be obtained by solving the
generalized eigenvalue problem as
HV =λSV (12)
where λis the eigenvalue of (12).
C. Neighborhood Structure Joint Feature Learning
To learn a more effective projection, we propose an LNSPE
algorithm for HSI data, which preserves the scatter information
and the dual graph structure to obtain discriminant features.
Combining (8) and (12), the projection matrix Vcan be
obtained by solving the following eigenvalue problem:
[(1 a)XMXT+aH]V=λ[(1 a)XXT+aS]V(13)
where ais a nonnegative tradeoff parameter between the local
neighborhood structure and the dual graph structure. With
the eigenvectors v1,v
2,...,v
dcorresponding to the first d
eigenvalues, the optimal projection matrix can be represented
as V=[v1v2... v
d]. Then, the low-dimensional embedding
of HSI data can be obtained by Y=VTX.
III. EXPERIMENTAL RESULTS AND ANALYSIS
In this section, two public HSI data sets are adopted to
demonstrate the effectiveness of LNSPE by comparing it with
some state-of-the-art DR algorithms.
Fig. 1. OAs for (a) PaviaU and (b) Botswana data sets under different values
of a.
Fig. 2. OAs for the PaviaU and Botswana data sets with different dimensions.
Fig. 3. OAs for (a) PaviaU and (b) Botswana data sets with kand r.
TAB L E I
CLASSIFICATION RESULTS [PERCENTAGE (%)] USING DIFFERENT
ALGORITHMS FOR PaviaU DATA SET
A. Data Description
1) Pavia University: It was collected by the Reflective
Optics System Imaging Spectrometer (ROSIS) sensor over
Pavia University in 2002. The full scene consists of 610 ×340
pixels with 115 spectral bands, and all pixels belong to nine
classes. Considering that 12 bands suffered from serious water
absorption, the remaining 103 bands are used for experiments.
2) Botswana: It was collected by the NASA EO-1 satellite
over Okavango Delta, Botswana in 2001. The full scene
consists of 1476 ×256 pixels with 242 spectral bands, and
all pixels belong to 14 classes. After removing 97 bands due
Authorized licensed use limited to: Hong Huang. Downloaded on February 15,2020 at 15:54:38 UTC from IEEE Xplore. Restrictions apply.
This article has been accepted for inclusion in a future issue of this journal. Content is final as presented, with the exception of pagination.
4IEEE GEOSCIENCE AND REMOTE SENSING LETTERS
Fig. 4. Classification maps for different methods with NN classifier on the PaviaU data set. (a) Ground truth. (b) Training samples. (c) RAW. (d) PCA.
(e) NPE. (f) LPP. (g) LDA. (h) LFDA. (i) MMC. (j) MFA. (k) LGSFA. (l) GDA-SS. (m) SME. (n) LNSPE.
TAB L E I I
CLASSIFICATION RESULTS [PERCENTAGE (%)] USING DIFFERENT
ALGORITHMS FOR BOT SWANA DATA SET
to serious noise affection, the remaining 145 bands are used
for experiments.
B. Experimental Setup
In each experiment, we randomly divided the HSI data
into the training and test sets, and employed the NN for
classification. Then, overall classification accuracy (OA) was
used to evaluate the effectiveness of each algorithm. For
robustness, all the experiments were repeated for ten times.
To verify the classification performance of LNSPE, we com-
pared it with RAW, PCA, NPE, LPP, LDA, local Fisher
discriminant analysis (LFDA), maximum margin criterion
(MMC), marginal Fisher analysis (MFA), local geometric
structure Fisher analysis (LGSFA) [6], GDA-SS, and SME,
and the RAW method indicates that the OAs are obtained
by the NN classifier without DR. Among all DR algorithms,
PCA, NPE, LPP, and SME are unsupervised algorithms that
do not use prior label information during the process of DR,
while LDA, LFDA, MMC, MFA, LGSFA, and GDA-SS are
supervised methods, which use the prior label information to
improve the classification performance. To better compare the
classification results, we choose optimal parameters for each
algorithm. For NPE, LPP, and LFDA, we set the number
of neighbors as 7. For MFA and LGSFA, they have two
important parameters of intraclass neighbor k1and interclass
neighbor k2=βk1,andwesetk1=9and β=20,
respectively. As for LNSPE, parameter ais tuned in the range
of {0,0.1,0.2,...,1}, and the results are shown in Fig. 1,
where 30 labeled samples per class were selected for training,
and the remaining samples were for testing.
As we can see from Fig. 1, the optimal parameter is a=0.3
for PaviaU and a=0.5for Botswana. Fig. 2 shows the OAs
of each method versus different embedding dimensions. It can
be seen that the OAs of all DR algorithms increase gradually
with the increase in d, and then maintain a stable value when
the dimension exceeds 20. Therefore, we set the embedding
dimension dto 30, and the dimension of the LDA algorithm
is set as c1,wherecis the class number of the HSI.
Fig. 3 shows the OAs with different values of kand r.
As we can see, the OAs first improve with the increase in
kand then maintain a stable value. The reason is that a
larger number of spectral neighbors are helpful to extract
discriminant features for HSI classification. However, if kis
too large, the discriminant information in the neighborhood
structure will be redundant for DR of HSI. Furthermore,
the OAs quickly ascend and then decrease with the increase
in r, which indicates that a larger ris not conducive to the
further improvement of OAs. Based on the above analysis,
we choose r=2, k=25 for the PaviaU data set, and r=3,
k=25 for the Botswana data set in the following experiments.
C. Classification Results
To demonstrate the classification performance of various DR
algorithms on the PaviaU and Botswana data sets, ni(ni=20,
30, 40, 50, 60) samples were randomly selected from each
class as train samples, and the remaining samples were used
for test samples. Tables I and II are the classification results for
the PaviaU and Botswana data sets, respectively. Fig. 4 shows
the classification maps for different methods on the PaviaU
Authorized licensed use limited to: Hong Huang. Downloaded on February 15,2020 at 15:54:38 UTC from IEEE Xplore. Restrictions apply.
This article has been accepted for inclusion in a future issue of this journal. Content is final as presented, with the exception of pagination.
SHI et al.: UNSUPERVISED DR FOR HYPERSPECTRAL IMAGERY 5
TABLE III
COMPUTATIONAL TIME (IN SECONDS)OF DIFFERENT ALGORITHMS ON THE PaviaU AND BOTSWANA DATA SETS
data set, in which 1% samples were randomly selected from
each class for training, and the remaining were used for testing.
As can be seen from Tables I and II, the OAs of all
algorithms improved with the increase of ni. The reason is that
more training samples can lead to abundant information for
feature learning, which is helpful to HSI classification. Among
all the DR methods, LNSPE achieves better classification per-
formance than other state-of-the-art DR algorithms. As shown
in Fig. 4, the classification map of LNSPE is smoother
than other algorithms, especially in the regions of Asphalt,
Meadows,Bitumen,andShadows. Because it considers the
relationships of pairwise neighbors in adjacency graph and
scatter information of samples, which is helpful to discover the
intrinsic structure of the HSI data and extract more effective
discriminant features.
D. Computational Complexity
As for the proposed LNSPE method, the reconstructed
pixels Xis calculated with O(nk). The reconstruction weight
matrix Wcosts O(nk3). The cost of Mis O(n2). The cal-
culation of XMXTtakes O(Dn2). The local neighborhood
preserving scatter matrix Hand the total scatter matrix S
take O(n2r2)and O(n), respectively. The generalized eigen-
value problem of (13) is computed with the cost of O(D3).
Therefore, the total computational complexity of LNSPE is
O(nk3+Dn2+n2r2+D3), and it mainly depends on size of
training samples, band number, and the neighbors number.
To quantitatively compare the complexity of each algorithm,
we show the computational time of each algorithm in Table III.
All the results were obtained on a personal computer, which
has the i3-7100 CPU and 12-G memory. The version of
Windows system and MATLAB are 64-bit Windows 10 and
2017a, respectively. As shown in Table III, the proposed
LNSPE method is slower than NPE. However, the slight
increase in computational time is acceptable relative to the
improvement on classification performance.
IV. CONCLUSION
This letter proposed a novel unsupervised DR method called
LNSPE algorithm for HSI classification. At first, LNSPE
considers the relationships of pairwise neighbors by con-
structing a dual adjacency graph. Then, LNSPE reveals the
scatter information of training samples by minimizing the
local neighborhood scatter and maximizing the total scatter of
training samples simultaneously. Finally, an optimal project
vector is learned by exploring the dual graph structure and
scatter information. LNSPE can effectively reveal the intrinsic
structure of the HSI and extract effective discriminant features,
which bring benefits for enhancing the classification perfor-
mance of the HSI. Experimental results on two real data sets
demonstrated the superiority of LNSPE for HSI classification.
Our future work will focus on how to combine the priori label
information to further improve the classification performance
of LNSPE.
REFERENCES
[1] H. Luo, C. Liu, C. Wu, and X. Guo, “Urban change detection based
on Dempster–Shafer theory for multitemporal very high-resolution
imagery, Remote Sens., vol. 10, no. 7, pp. 980, Jun. 2018.
[2] H. Huang, G. Shi, H. He, Y. Duan, and F. Luo, “Dimension-
ality reduction of hyperspectral imagery based on spatial-spectral
manifold learning,” IEEE Trans. Cybern., to be published. doi:
10.1109/TCYB.2019.2905793.
[3] F. Luo, B. Du, L. Zhang, L. Zhang, and D. Tao, “Feature learning
using spatial-spectral Hypergraph discriminant analysis for hyperspectral
image,” IEEE Trans. Cybern., vol. 49, no. 7, pp. 2406–2419, Jul. 2019.
[4] J. Peng, W. Sun, and Q. Du, “Self-paced joint sparse representation for
the classification of hyperspectral images,” IEEE Trans. Geosci. Remote
Sens., vol. 57, no. 2, pp. 1183–1194, Feb. 2019.
[5] Z. Wang, B. Du, L. Zhang, L. Zhang, and X. Jia, “A novel semisuper-
vised active-learning algorithm for hyperspectral image classification,”
IEEE Trans. Geosci. Remote Sens., vol. 55, no. 6, pp. 3071–3083,
Jun. 2017.
[6] F. Luo, H. Huang, Y. Duan, J. Liu, and Y. Liao, “Local geometric
structure feature for dimensionality reduction of hyperspectral imagery,
Remote Sens., vol. 9, no. 8, pp. 790, Aug. 2017.
[7] W. He, H. Zhang, L. Zhang, W. Philips, and W. Liao, “Weighted sparse
graph based dimensionality reduction for hyperspectral images,” IEEE
Geosci. Remote Sens. Lett., vol. 13, no. 5, pp. 686–690, May 2016.
[8] R. Hang and Q. Liu, “Dimensionality reduction of hyperspectral image
using spatial regularized local graph discriminant embedding,” IEEE
J. Sel. Topics Appl. Earth Observ. Remote Sens., vol. 11, no. 9,
pp. 3262–3271, Sep. 2018.
[9] H. Xu, H. Zhang, W. He, and L. Zhang, “Superpixel-based spatial-
spectral dimension reduction for hyperspectral imagery classification,”
Neurocomputing, vol. 360, pp. 138–150, Sep. 2019.
[10] Y. Zhou, J. Peng, and C. L. P. Chen, “Dimension reduction using spatial
and spectral regularized local discriminant embedding for hyperspectral
image classification,” IEEE Trans. Geosci. Remote Sens., vol. 53, no. 2,
pp. 1082–1095, Feb. 2015.
[11] H. Yu, L. Gao, W. Li, Q. Du, and B. Zhang, “Locality sensitive dis-
criminant analysis for group sparse representation-based hyperspectral
imagery classification,” IEEE Geosci. Remote Sens. Lett., vol. 14, no. 8,
pp. 1358–1362, Aug. 2017.
[12] Y. Chen, Z. Lai, W. Wong, L. Shen, and Q. Hu, “Low-rank linear
embedding for image recognition,” IEEE Trans. Multimedia, vol. 20,
no. 12, pp. 3212–3222, Dec. 2018.
[13] W. Sun, G. Yang, B. Du, L. Zhang, and L. Zhang, “A sparse and low-
rank near-isometric linear embedding method for feature extraction in
hyperspectral imagery classification,” IEEE Trans. Geosci. Remote Sens.,
vol. 55, no. 7, pp. 4032–4046, Jul. 2017.
[14] J. Wang, X. Sun, and J. Du, “Local tangent space alignment via nuclear
norm regularization for incomplete data,” Neurocomputing, vol. 273,
pp. 141–151, Jan. 2018.
[15] S. Wang and W. Zhu, “Sparse graph embedding unsupervised feature
selection,” IEEE Trans. Syst., Man, Cybern., Syst., vol. 48, no. 3,
pp. 329–341, Mar. 2018.
[16] F. Zhong, J. Zhang, and D. Li, “Discriminant locality preserving
projections based on L1-norm maximization,” IEEE Trans. Neural Netw.
Learn. Syst., vol. 25, no. 11, pp. 2065–2074, Nov. 2014.
[17] Y. Lu, Z. Lai, Z. Fan, J. Cui, and Q. Zhu, “Manifold discriminant
regression learning for image classification,” Neurocomputing, vol. 166,
pp. 475–486, Oct. 2015.
[18] Y. Wei, Y. Zhou, and H. Li, “Spectral-spatial response for hyperspectral
image classification,” Remote Sens., vol. 9, no. 3, pp. 203, Feb. 2017.
[19] H. Huang, F. Luo, J. Liu, and Y. Yang, “Dimensionality reduc-
tion of hyperspectral images based on sparse discriminant manifold
embedding,” ISPRS J. Photogramm. Remote Sens., vol. 106, pp. 42–54,
Aug. 2015.
[20] F. Feng, W. Li, Q. Du, and B. Zhang, “Dimensionality reduction
of hyperspectral image with graph-based discriminant analysis con-
sidering spectral similarity,” Remote Sens., vol. 9, no. 4, pp. 323,
Mar. 2017.
Authorized licensed use limited to: Hong Huang. Downloaded on February 15,2020 at 15:54:38 UTC from IEEE Xplore. Restrictions apply.
... A series of algorithms designed under this framework were subsequently proposed. For example, marginal fisher analysis (MFA) [15], collaborative representation projection (CRP) [16], local neighborhood preserving embedding (LNSPE) [17], etc. ...
... Equation (18) is the calculation of the dimensionality reduction projection matrix of the MFSCDP algorithm, which is selected by selecting the top d eigenvectors CA rank[1∼d] according to equation (17) to construct the matrix A. ...
... Ctraindata of low-dimensional training samples in the multiple feature space according to equation (16). (e) Filter the collaborative combination of features according to equation (17) and construct the target projection matrix according to equation (18). ...
Article
Full-text available
At present, the trend of complex and intelligent rotating machinery and equipment is becoming more and more obvious, which generates a large amount of high-dimensional and nonlinear fault monitoring data that is difficult to handle. This makes the traditional dimensionality reduction algorithms based on point-to-point metrics or a small number of graph embedding structures lose their utility. To solve this problem, a multiple feature-spaces collaborative discriminative projection (MFSCDP) algorithm for rotor fault dataset dimensionality reduction is proposed. The algorithm first improves the projection metric from sample point to feature space into the median metric in order to achieve the effect of weakening the extrapolation error of the algorithm, and based on this, we propose a sample point-to-point guided nearest-neighbor feature space selection method to improve the construction efficiency of the feature space embedding graph. Then, by using Relief F to indirectly construct the reduced dimensional projection matrix with multiple feature spaces of collaboration. Finally, the proposed MFSCDP algorithm is used for the dimensionality reduction process of the rotor fault dataset. The algorithm's performance was verified using experimental information from rotor failure simulations of two different structural types. The result shows that the algorithm can reduce the difficulty of fault classification and improve the accuracy of identification.
... Due to the large spatial coverage, the memory overflow occurred on HSI. Shi et al. (2019) developed an Local Neighborhood Structure preserving Embedding (LNSPE) model to classify the HSI. Each sample was reconstructed with its corresponding spectral neighbors and attained an optimal weight by modifying the loss function. ...
Article
Full-text available
Hyperspectral images (HSI) are contiguous band images commonly used in remote sensing applications. Over the past decades, significant advancements have been made in HSI enhancement and classification. However, challenges such as computational complexity, low spatial resolution, overfitting, noisy images, misclassification, and slow convergence speed persist. So the novel HSI enhancement and classification model is proposed to address and overcome these issues. The data are collected from the different HSI datasets namely Pavia University (PU), Indian Pines (IP), Salinas Valley (SV), Houston University (HU) and Kennedy Space Center (KSC) for preprocessing the collected data. The collected data undergoes preprocessing, which includes noise reduction, geometric correction, and radiometric calibration operations to improve image quality. To reduce the high dimensionality of the preprocessed image, Principal Component Analysis (PCA) is employed. Additionally, the Mutation Boosted Aquila Optimization (MBAO) algorithm is applied to enhance the visual quality of the image. The image enhancement process relies on three types of enhancement parameters: histogram equalization, contrast stretching, and adaptive filtering. Finally, the enhanced image is classified using a Convolutional Neural Network (CNN) architecture, with weight updates performed iteratively using the Stochastic Gradient Descent (SGD) model to minimize the loss function. A comparative analysis is conducted to evaluate the superiority of the proposed model. The experiments of the proposed model achieves an F1-score of 98.2%, kappa statistics of 98.3%, SSIM of 0.99, PSNR of 38.81 dB, RMSE of 1.97, and UQI of 0.85. These findings indicate that the proposed model outperforms other existing HSI enhancement and classification models.
... Dimensionality reduction technology has been widely used in face recognition [1], image fusion [2], moving object detection [3], bearing fault diagnosis [4], hyperspectral imagery [5,6], and so on. The aim of dimensionality reduction is to find an effective projection matrix that can extract effective low dimensional features from the original high dimensional samples, which not only improves the recognition performance of the algorithm, but also saves memory for storage. ...
Article
Full-text available
Marginal Fisher analysis (MFA) maintains the nearest neighbor structure according to the class information of samples, so it achieves good recognition results for image recognition task. However, MFA needs to manually determine the number of nearest neighbor samples and simply sets the weight value of nearest neighbor samples to one. Furthermore, MFA only considers the local information and often encounters the small sample size problems while dealing with image recognition. Therefore, based on MFA and collaborative representation (CR), we introduce a new method, called a complete joint global and local collaborative marginal fisher analysis (CJGLCMFA). CJGLCMFA defines inter-class collaborative weighted matrix and intra-class collaborative weighted matrix based on CR and class label information, which is able to automatically construct the weighted matrices and avoid manually choosing nearest neighbors. In order to further enhance the algorithm, the objective function considers both local and global information of samples and then the complete solution for CJGLCMFA is given to solve the small sample size problem. Extensive experiments on some benchmark datasets show that CJGLCMFA is feasible and practicable, and the best recognition result of CJGLCMFA is 98.62%. Compared with other algorithms, the best improvement is 2.44% higher than the other best recognition results.
Article
Full-text available
Hyperspectral Image Processing (HSIP) is an essential technique in remote sensing. Currently, extensive research is carried out in hyperspectral image processing, involving many applications, including land cover classification, anomaly detection, plant classification, etc., Hyperspectral image processing is a powerful tool that enables us to capture and analyze an object's spectral information with greater accuracy and precision. Hyperspectral images are made up of hundreds of spectral bands, capturing an immense amount of information about the earth's surface. Accurately classifying and predicting land cover in these images is critical to understanding our planet's ecosystem and the impact of human activities on it. With the advent of deep learning techniques, the process of analyzing hyperspectral images has become more efficient and accurate than ever before. These techniques enable us to categorize land cover and predict Land Use/Land Cover (LULC) with exceptional precision, providing valuable insights into the state of our planet's environment. Image classification is difficult in hyperspectral image processing because of the large number of data samples but with a limited label. By selecting the appropriate bands from the image, we can get the finest classification results and predicted values. To our knowledge, the previous review papers concentrated only on the classification method. Here, we have presented an extensive review of various components of hyperspectral image processing, hyperspectral image analysis, pre-processing of an image, feature extraction and feature selection methods to select the number of features (bands), classification methods, and prediction methods. In addition, we also elaborated on the datasets used for classification, evaluation metrics used, various issues, and challenges. Thus, this review article will benefit new researchers in the hyperspectral image classification domain.
Article
Dimensionality reduction (DR) is important for feature extraction and classification of hyperspectral images (HSIs). Recently proposed superpixel-based DR models have shown promising performance, where superpixel segmentation techniques were applied to segment an HSI and then DR models like principal component analysis (PCA) or linear discriminant analysis (LDA) were employed to extract the local and/or global features. However, superpixelwise PCA (SuperPCA)-based local features are unsatisfactory because PCA aims to extract features with high variance, which could be inefficient in superpixels with mixed objects or strong noise/outliers. In addition, superpixelwise unsupervised LDA (SuperULDA) based global features may neglect local (spatial-contextual) information. To address these issues, we propose a new spectral–spatial and superpixelwise unsupervised LDA (S3-ULDA) model for unsupervised feature extraction from HSIs. Specifically, the HSI is first segmented into various superpixels with pseudo labels. Then, superpixel-based local reconstruction for HSI denoising is conducted. Next, SuperULDA is performed on both the original HSI and locally reconstructed data to extract global features. Then, superpixelwise unsupervised local Fisher discriminant analysis (SuperULFDA) is developed for local feature extraction, where each superpixel and its adjacent superpixels (along with their pseudo-labels) are fed into local Fisher discriminant analysis (LFDA) to extract local features. The superpixel-level local manifold structures can be effectively modeled by the proposed SuperULFDA. Finally, by fusing the extracted global and local features, novel global–local and spectral–spatial features can be obtained. Our experimental results on several benchmark HSIs demonstrate the superiority of the proposed method over state-of-the-art methods. The code of the proposed model is available at https://github.com/XinweiJiang/S3-ULDA .
Article
Hyperspectral anomaly detection (HAD) is regarded as an indispensable, pivotal technology in remote sensing and earth science domains. Nevertheless, most existing detection approaches for anomaly targets flatten 3-D hyperspectral images (HSIs) with spatial and spectral information into 2-D spectral vector data, which virtually breaks up the internal spatial structure in HSIs and degenerates the detection performance. To this end, we directly consider the HSI data cube as a 3-D tensor and develop a novel tensor low-rank approximation (TLRA) detection algorithm to separate the sparse anomalous component from the background with low-rank characteristics. Then, in light of the multi-subspace structure in heterogeneous backgrounds, we utilize multiple subspace learning (MSL) theory to encode the background tensor with a coefficient tensor and corresponding dictionary tensor. In addition, considering that different singular values indicate different information quantities and should be penalized to different extents, we introduce a tighter tensor rank surrogate named the ϵ-shrinkage tensor nuclear norm (ϵ-TNN) to recover the low-rank component more accurately. Meanwhile, concerning the sparse anomaly target, the l <sub xmlns:mml="http://www.w3.org/1998/Math/MathML" xmlns:xlink="http://www.w3.org/1999/xlink">2,1</sub> constraint is incorporated to represent the group sparsity of the abnormal component. Finally, an effective iterative optimization algorithm based on the alternating direction method of multipliers (ADMM) is devised to solve the proposed TLRA-MSL model. We conduct extensive experiments on six hyperspectral datasets to prove the effectiveness and robustness of our method. The experimental results illustrate that better detection performance is obtained using the proposed model compared with other state-of-the-art algorithms.
Article
Graph embedding provides an effective way to reveal intrinsic feature of high-dimensional data on the foundation of preserving topological properties. Under the framework of graph embedding, the hyperspectral image can be represented by a weighted graph, where pixels and similarities among them are treated as vertices and edge weights, respectively. In this article, an adaptive reference-related graph embedding (ARGE) method is proposed to efficaciously obtain the low-dimensional feature and improve computational efficiency. The ARGE method is composed of two primary processes. The key to connecting these two processes is reference vertices set, which is the abstraction of graph topological features. Firstly, the reference vertices are adaptively selected through a three-step adaptive reference set selection (ARSS) algorithm. Secondly, the original high-dimensional graph is embedded as a low-dimensional graph through preserving the reference-related structure. Specifically, the pairwise similarities between vertices and reference vertices are preserved in embedding space. In addition, a new hybrid dissimilarity measure of Rao distance and spectral information divergence (RD-SID) is designed to depict the spectral difference between pixels. To evaluate the effectiveness of the proposed method, the obtained low-dimensional feature is fed into the anomaly detector to detect anomalous pixels. The experimental results on five real and one synthetic hyperspectral datasets demonstrate the superiority of the proposed ARGE method over the compared feature extraction methods.
Article
Unsupervised feature extraction methods have been widely applied to remove the huge amount of redundancy in hyperspectral images due to their effectiveness when the label information of samples is unreachable. However, because of the lack of label information, unsupervised feature extraction methods are deficient in the discriminant ability compared to supervised methods. When the number of samples is small, the effect of dimension reduction is usually not good enough. To address the problems, an unsupervised structure preserving projection method named contrastive learning based sparsity preserving projection (CL-SPP) is proposed in this paper. Firstly, CL-SPP increases the discriminant ability of samples by introducing the concept of positive and negative pairs, and adjusts the number of positive and negative pairs in the training set through a parameter. Then, by minimizing the contrastive loss function, CL-SPP makes the positive pairs more similar and the negative pairs less similar after projection. Moreover, the proposed contrastive learning-based method is also extended to the supervised case, as well as a general graph embedding model framework based on comparative learning. Experiments on three hyperspectral images demonstrate that the proposed methods have a better performance than related approaches. More impressively, the effect of CL-SPP is comparable to its supervised version.
Article
Full-text available
Fusing multiple change detection results has great potentials in dealing with the spectral variability in multitemporal very high-resolution (VHR) remote sensing images. However, it is difficult to solve the problem of uncertainty, which mainly includes the inaccuracy of each candidate change map and the conflicts between different results. Dempster-Shafer theory (D-S) is an effective method to model uncertainties and combine multiple evidences. Therefore, in this paper, we proposed an urban change detection method for VHR images by fusing multiple change detection methods with D-S evidence theory. Change vector analysis (CVA), iteratively reweighted multivariate alteration detection (IRMAD), and iterative slow feature analysis (ISFA) were utilized to obtain the candidate change maps. The final change detection result is generated by fusing the three evidences with D-S evidence theory and a segmentation object map. The experiment indicates that the proposed method can obtain the best performance in detection rate, false alarm rate, and comprehensive indicators.
Article
Full-text available
Hyperspectral image (HSI) contains a large number of spatial-spectral information, which will make the traditional classification methods face an enormous challenge to discriminate the types of land-cover. Feature learning is very effective to improve the classification performances. However, the current feature learning approaches are most based on a simple intrinsic structure. To represent the complex intrinsic spatial-spectral of HSI, a novel feature learning algorithm, termed spatial-spectral hypergraph discriminant analysis (SSHGDA), has been proposed on the basis of spatial-spectral information, discriminant information, and hypergraph learning. SSHGDA constructs a reconstruction between-class scatter matrix, a weighted within-class scatter matrix, an intraclass spatial-spectral hypergraph, and an interclass spatial-spectral hypergraph to represent the intrinsic properties of HSI. Then, in low-dimensional space, a feature learning model is designed to compact the intraclass information and separate the interclass information. With this model, an optimal projection matrix can be obtained to extract the spatial-spectral features of HSI. SSHGDA can effectively reveal the complex spatial-spectral structures of HSI and enhance the discriminating power of features for land-cover classification. Experimental results on the Indian Pines and PaviaU HSI data sets show that SSHGDA can achieve better classification accuracies in comparison with some state-of-the-art methods.
Article
Full-text available
Marginal Fisher analysis (MFA) exploits the margin criterion to compact the intraclass data and separate the interclass data, and it is very useful to analyze the high-dimensional data. However, MFA just considers the structure relationships of neighbor points, and it cannot effectively represent the intrinsic structure of hyperspectral imagery (HSI) that possesses many homogenous areas. In this paper, we propose a new dimensionality reduction (DR) method, termed local geometric structure Fisher analysis (LGSFA), for HSI classification. Firstly, LGSFA uses the intraclass neighbor points of each point to compute its reconstruction point. Then, an intrinsic graph and a penalty graph are constructed to reveal the intraclass and interclass properties of hyperspectral data. Finally, the neighbor points and corresponding intraclass reconstruction points are used to enhance the intraclass-manifold compactness and the interclass-manifold separability. LGSFA can effectively reveal the intrinsic manifold structure and obtain the discriminating features of HSI data for classification. Experiments on the Salinas, Indian Pines, and Urban data sets show that the proposed LGSFA algorithm achieves the best classification results than other state-of-the-art methods.
Article
Dimension reduction (DR) is a useful preprocessing technology for hyperspectral image (HSI) classification. This paper presents an HSI DR method named superpixel-based spatial-spectral dimension reduction (SSDR), which integrates the spatial and spectral similarity. The HSI is first segmented into non-overlapping superpixels, where pixels belonging to the same superpixel have strong correlations, and should be preserved after DR. We then apply the superpixel-based linear discriminant analysis (SPLDA) method, which learns a superpixel-guided graph to capture the spatial similarity. Pixels from the same label also have strong spectral correlations; thereby, we also construct a label-guided graph to explore the spectral similarity. These two graphs are finally integrated to learn the discriminant projection. The classification results on two widely used HSIs demonstrate the advantage of the proposed algorithms compared to the other state-of-the-art DR methods.
Article
The graph embedding (GE) methods have been widely applied for dimensionality reduction of hyperspectral imagery (HSI). However, a major challenge of GE is how to choose the proper neighbors for graph construction and explore the spatial information of HSI data. In this paper, we proposed an unsupervised dimensionality reduction algorithm called spatial-spectral manifold reconstruction preserving embedding (SSMRPE) for HSI classification. At first, a weighted mean filter (WMF) is employed to preprocess the image, which aims to reduce the influence of background noise. According to the spatial consistency property of HSI, SSMRPE utilizes a new spatial-spectral combined distance (SSCD) to fuse the spatial structure and spectral information for selecting effective spatial-spectral neighbors of HSI pixels. Then, it explores the spatial relationship between each point and its neighbors to adjust the reconstruction weights to improve the efficiency of manifold reconstruction. As a result, the proposed method can extract the discriminant features and subsequently improve the classification performance of HSI. The experimental results on the PaviaU and Salinas hyperspectral data sets indicate that SSMRPE can achieve better classification results in comparison with some state-of-the-art methods.
Article
In this paper, a self-paced joint sparse representation (SPJSR) model is proposed for the classification of hyperspectral images (HSIs). It replaces the least-squares (LS) loss in the standard joint sparse representation (JSR) model with a weighted LS loss and adopts a self-paced learning (SPL) strategy to learn the weights for neighboring pixels. Rather than predefining a weight vector in the existing weighted JSR methods, both the weight and sparse representation (SR) coefficient associated with neighboring pixels are optimized by an alternating iterative strategy. According to the nature of SPL, in each iteration, neighboring pixels with nonzero weights (i.e., easy pixels) are included for the joint SR of a testing pixel. With the increase of iterations, the model size (i.e., the number of selected neighboring pixels) is enlarged and more neighboring pixels from easy to complex are gradually added into the JSR learning process. After several iterations, the algorithm can be terminated to produce a desirable model that includes easy homogeneous pixels and excludes complex inhomogeneous pixels. Experimental results on two benchmark hyperspectral data sets demonstrate that our proposed SPJSR is more accurate and robust than existing JSR methods, especially in the case of heavy noise.
Article
Dimensionality reduction (DR) is an important preprocessing step for hyperspectral image (HSI) classification. Recently, graph-based DR methods have been widely used. Among various graph-based models, the local graph discriminant embedding (LGDE) model has shown its effectiveness due to the complete use of label information. Besides spectral information, an HSI also contains rich spatial information. In this paper, we propose a regularization method to incorporate the spatial information into the LGDE model. Specifically, an oversegmentation method is first employed to divide the original HSI into nonoverlapping superpixels. Then, based on the observation that pixels in a superpixel often belong to the same class, intraclass graphs are constructed to describe such spatial information. Finally, the constructed superpixel-level intraclass graphs are used as a regularization term, which can be naturally incorporated into the LGDE model. Besides, to sufficiently capture the nonlinear property of an HSI, the linear LGDE model is further extended into its kernel counterpart. To demonstrate the effectiveness of the proposed method, experiments have been established on three widely used HSIs acquired by different hyperspectral sensors. The obtained results show that the proposed method can achieve higher classification performance than many state-of-the-art graph embedding models, and the kernel extension model can further improve the classification performance.
Article
Locality preserving projections (LPP) has been widely studied and extended in recent years, because of its promising performance in feature extraction. In this paper, we propose a modified version of the LPP by constructing a novel regression model. To improve the performance of the model, we impose a low-rank constraint on the regression matrix to discover the latent relations between different neighbors. By using the L <sub xmlns:mml="http://www.w3.org/1998/Math/MathML" xmlns:xlink="http://www.w3.org/1999/xlink">2,1</sub> -norm as a metric for the loss function, we can further minimize the reconstruction error and derive a robust model. Furthermore, the L <sub xmlns:mml="http://www.w3.org/1998/Math/MathML" xmlns:xlink="http://www.w3.org/1999/xlink">2,1</sub> -norm regularization term is added to obtain a jointly sparse regression matrix for feature selection. An iterative algorithm with guaranteed convergence is designed to solve the optimization problem. To validate the recognition efficiency, we apply the algorithm to a series of benchmark datasets containing face and character images for feature extraction. The experimental results show that the proposed method is better than some existing methods. The code of this paper can be downloaded from http://www.scholat.com/laizhihui.
Article
Manifold learning approaches seek to find the low-dimensional features of high-dimensional data. When some values of the data are missing, the effectiveness of manifold learning methods may be greatly limited since they have difficulty in determining the local neighborhoods and discovering the local structures of neighborhoods. In this paper, a novel manifold learning approach called local tangent space alignment via nuclear norm regularization (LTSA-NNR) is proposed to discover the nonlinear features of the incomplete data. The neighbors of each sample point are selected using the cosine similarity measurement. A new nuclear norm regularization model is then proposed to discover the local coordinate systems of the determined neighborhoods. Different with the traditional manifold learning approaches, the dimensions of local coordinate systems are various in a reasonable range. The global coordinates of the incomplete data are finally obtained by aligning the local coordinates together. We demonstrate the effectiveness of our method on real-world data sets.
Article
This letter proposes to integrate the locality sensitive discriminant analysis (LSDA) with the group sparse representation (GSR) for a hyperspectral imagery classification. The LSDA is to project the data set to a lower-dimensional subspace to preserve local manifold structure and discriminant information, while the GSR is to encode the projected testing set as a sparse linear combination of group-structured training samples for classification. The proposed approach, denoted as LSDA-GSR classifier (GSRC), is evaluated using two real hyperspectral data sets. Experimental results demonstrate that it can provide considerable improvement to the original counterparts, i.e., SRC and GSRC, with a relatively low computational cost.