This paper was converted on www.awesomepapers.org from LaTeX by an anonymous user.
Want to know more? Visit the Converter page.

\cftpagenumbersoff

figure \cftpagenumbersofftable

Spectral Unmixing of Hyperspectral Images Based on Block Sparse Structure

Seyed Hossein Mosavi Azarang Qom University of Technology, Faculty of Electrical and Computer Engineering, Department of Communications and Electronics, Qom, Iran, 3718146645 Roozbeh Rajabi Qom University of Technology, Faculty of Electrical and Computer Engineering, Department of Communications and Electronics, Qom, Iran, 3718146645 Hadi Zayyani Qom University of Technology, Faculty of Electrical and Computer Engineering, Department of Communications and Electronics, Qom, Iran, 3718146645 Amin Zehtabian Freie Universität Berlin, Institut für Biochemie, Berlin, Germany, 14195
Abstract

Spectral unmixing (SU) of hyperspectral images (HSIs) is one of the important areas in remote sensing (RS) that needs to be carefully addressed in different RS applications. Despite the high spectral resolution of the hyperspectral data, the relatively low spatial resolution of the sensors may lead to mixture of different pure materials within the image pixels. In this case, the spectrum of a given pixel recorded by the sensor can be a combination of multiple spectra each belonging to a unique material in that pixel. Spectral unmixing is then used as a technique to extract the spectral characteristics of the different materials within the mixed pixels and to recover the spectrum of each pure spectral signature, called endmember. Block-sparsity exists in hyperspectral images as a result of spectral similarity between neighboring pixels. In block-sparse signals, the nonzero samples occur in clusters and the pattern of the clusters is often supposed to be unavailable as prior information. This paper presents an innovative spectral unmixing approach for HSIs based on block-sparse structure. Hyperspectral unmixing problem is solved using pattern coupled sparse Bayesian learning strategy (PCSBL). To evaluate the performance of the proposed SU algorithm, it is tested on both synthetic and real hyperspectral data and the quantitative results are compared to those of other state-of-the-art methods in terms of abundance angle distance and mean squared error. The achieved results show the superiority of the proposed algorithm over the other competing methods by a significant margin.

keywords:
Spectral Unmixing, Hyperspectral Images, Block Sparse Structure, Sparse Bayesian Learning

*Roozbeh Rajabi, \linkable[email protected]

1 Introduction

Remote sensing is the science of collecting and interpreting information from objects without physical contact with the scene. One of the modern technologies in remote sensing is hyperspectral imaging. Hyperspectral sensors have the capability of recording electromagnetic energy reflected on the surface of objects in very narrow spectral bands [1].

Today, hyperspectral imaging has been transformed from sparse research into a promising tool. The spectral information produced by advanced hyperspectral imaging instruments has given rise to new insights in a variety of applications such as: monitoring urban and environmental processes, preventing and observing destructive factors including weather detection, environmental hazard detection, oil spill monitoring and other types of chemical contamination [2]. Advanced hyperspectral instruments such as NASA’s Airborne Visible Infrared Imaging Spectrometer (AVIRIS) are able to cover the wavelengths from 400 to 2500 nm containing more than 200 spectral channels with spectral resolution of less than 10 nm [3]. Each pixel vector in the resulting stack of images represents a spectral signature that characterizes the underlying materials within that limited area [4]. Figure 1.(a) shows RGB image of AVIRIS Cuprite dataset using three bands (40, 20, 10) and Figure 1.(b) illustrates 3D hypercube of this dataset using 188 bands after removing noisy and water absorption channels. The figures are generated using Spectral Python (SPy)111www.spectralpython.net module.

Refer to caption
Figure 1: Hyperspectral data illustration of AVIRIS Cuprite dataset (a) RGB image using (40, 20, 10) bands (b) Hypercube using 188 bands.

There are different assumptions regarding the type of spectral mixing. Based on these assumptions, different mixing models have been proposed, which are divided into linear and nonlinear models:

Linear mixing model (LMM): The measured spectra are expressed as linear (convex) combination of all the spectral signatures of materials present in the mixed pixel.

Nonlinear mixing model: In this model, the spectrum recorded in each pixel is a nonlinear function of the spectrum of the principal elements, that the reflections recorded by the sensor is the result of multiple reflections and multiple interactions between materials [5].

Since linear mixing model is rather simple and gives prominent results, it is widely used in the literature. First category of spectral unmixing methods based on LMM assumes that pure pixels exist in the image for each of the principal elements called endmembers, so that the recorded signal of that pixel contains only the frequency spectrum of that element. Vertex component analysis (VCA) method works on this basis [6]. Nonnegative matrix factorization (NMF) methods are another category of spectral unmixing algorithms. These methods have been used extensively to solve the problem of Blind source separation (BSS). To improve the performance of nonnegative matrix factorization methods, the sparsity constraint has been added to their cost function, since the number of nonzero elements of the abundance matrix is very low and this assumption is close to reality [7, 8].

To estimate fractional abundance and endmembers matrices, that are blind problems, the nonnegative matrix fractional NMF and its generalized methods are particularly useful. In multilayer nonnegative matrix factorization (MLNMF) method, deep multilayer decomposition is used and fractional abundance matrices are modeled as a sparse matrix in each layer [9]. Graph regularized NMF (GNMF) is an unmixing method in which sparseness constraint is considered while the geometrical structure of the hyperspectral data is preserved [10]. Another NMF method is based on adding a sparse constraint which in terms of lql_{q} norm. In this method, a network consisting of single node clusters is used, so that each hyperspectral pixel is considered as a node where the sparsity constraint expressed with diffusion least mean p-power (LMP) strategy is optimized [11, 12]. In the clustered multitask NMF algorithm the nodes in the network are clustered by fuzzy c-means method and diffusion LMP strategy has been used to define the cost function [13].

Spectral-spatial constrained sparse unmixing (SSCSUn) describes a method in which a total variation (TV) regulator is used as a spatial weight factor to more effectively utilize spatial information [14]. An innovative linear method that is based on l1l2l_{1}-l_{2} sparsity and TV regularization has been developed to ameliorate the accuracy of hyperspectral unmixing. First, the enhancement of the sparsity based on the l1l2l_{1}-l_{2} norm is investigated to show the sparsity properties of the relative fractional abundance in a sparse regression model, because the l1l2l_{1}-l_{2} norm has stronger sparsity over the l1l_{1} norm. Then, considering the spatial correlation between the neighboring pixels, total variation minimizes the spatial smoothness constraint. Finally, the alternating direction method of multipliers (ADMM) is used to solve the model [15]. To further improve the accuracy of spectral unmixing in the high-coherence spectral library, an algorithm based on a kernel sparse representation model with a total variation constraint is designed. To increase the effect of similarity between measurements, library atoms and hyperspectral data were added to a kernel space in which sparse regression algorithms are used [16]. In semi-supervised spectral unmixing model, it is assumed that the spectral signature of the elements can be obtained from the hyperspectral library of the observed image. This indicates the problem of sparse regression. Sparse unmixing via variable splitting augmented Lagrangian (SUnSAL) method can be used to solve this problem [17]. The goal of iterative spectral mixture analysis (ISMA) algorithm is to find the optimal endmember set to estimate abundance fractions [18]. On the other hand, subspace matching pursuit (SMP) is one of the sparse methods used to solve the problem of blind sources separation [19]. Robust spectral unmixing (RSU) algorithm uses the mean squares error to solve the problem of spectral unmixing [20]. In contrast with the above-mentioned methods where the average value of Gaussian noise used for all spectral bands, the sparse unmixing method with the bandwise unmixing model (SUBM) uses the idea of Gaussian noise difference in each band for spectral separation of hyperspectral images [21]. In the spatial discontinuity-weighted (SDW) sparse unmixing method, the objective is to use spatial information on the edges of the hyperspectral images. This information, extracted using Sobel filter, is then weighted for sparse unmixing [22].

Recently, deep learning based methods are introduces to solve hyperspectral unmixing problem. A general deep learning framework for self-supervised hyperspectral unmixing called EGU-Net (Endmember-Guided Unmixing Network) is introduced by Hong et. al. [23]. EGU-Net uses a two-stream Siamese network to learn a network from pure or nearly-pure endmembers for the purpose of correcting the weights of another network by adding spectral unmixing constraints like abundance sum to one and positivity. Another deep learning method for hyperspectral unmixing is proposed by Vijayashekhar et. al. [24] based on two-stage fully connected self-supervised network. In this method the main goal is reconstructing the hyperspectral data using a jointly optimized network based on two-stage loss function.

In this paper, sparse Bayesian strategy and block sparse structure [25, 26] are used for spectral unmixing of hyperspectral images. The basis of the Bayesian learning strategy is used on hyperparameter information attributed to each pixel. In the proposed method, block sparse structure of the hyperspectral images is exploited. In this way, in addition to using the hyperparameter information attributed to each pixel, hyperparameters of the neighbours of that pixel are also used, which improves Bayesian learning performance.

The rest of this paper is organized as follows. In the second section, mathematical expression of the problem is discussed. In the third section, the sparse signal structure which is the basis of our proposed method, is explained. In the fourth section, we described the proposed algorithm using the sparse Bayesian learning (SBL) strategy. In the fifth section, for the performance evaluation, the proposed algorithm is evaluated using synthetic and real dataset. Proposed method is compared with three competing methods and the results are presented. Finally in the conclusion section, our concluding remarks are presented.

2 Problem Formulation

In the linear mixture model, it is assumed that the spectral response of each pixel in each spectral band is a linear combination of all the end members present in the pixel. For each pixel, the linear mixing model equation is as follows:

yi=j=1qaijxj+niy_{i}=\sum_{j=1}^{q}{a_{ij}x_{j}+n_{i}} (1)

Where yiy_{i} is the reflectance value at iith spectral band, aija_{ij} is the reflectance value of the jjth endmember at the iith spectral band, xjx_{j} is the fractional abundance of the jjth endmember, nin_{i} represents the measurement noise for the spectral band ii, and qq is the total number of endmembers. If it is assumed that the hyperspectral sensor collects data in L-band, (1) can be written in matrix form:

𝐲=𝐀𝐱+𝐧\mathbf{y}=\mathbf{Ax}+\mathbf{n} (2)

Where 𝐲\mathbf{y} is an L×1L\times 1 vector, corresponding to the measured spectrum of the pixel, 𝐀\mathbf{A} is an L×qL\times q matrix comprising of qq pure spectral signatures (endmembers), 𝐱\mathbf{x} is a q×1q\times 1 vector containing the fractional abundances of the endmembers, and 𝐧\mathbf{n} is an L×1L\times 1 vector representing the measurement noise affecting each spectral band. The number of endmembers qq is assumed to be known in our experiments. To estimate qq, virtual dimensionality estimation methods for hyperspectral data can be used [27, 28]. The abundance vector of the pure elements in the combination of each pixel is usually associated with the following two constraint:

𝐱0\mathbf{x}\geq 0 (3)
𝟏T𝐱=1\mathbf{1}^{T}\mathbf{x}=1 (4)

where 1T1^{T} is a row vector of 1’s. These equations are abundance nonnegativity constraint (ANC) and abundance sum to one constraint (ASC) respectively.

3 Proposed Methodology

In this section, our proposed methodology based on block sparse Bayesian learning is presented.

3.1 BLOCK SPARSE SIGNAL REPRESENTATION

The Shannon-Nyquist theory states that for a complete reconstruction of a bounded signal, the sampling rate must be at least twice the maximum frequency in the signal; whereas, in compressed sensing, less number of samples or measurements is required [29]. Compressed sensing or sampling is based on the principle that if a signal in a base or dictionary has a sparse representation, the signal can then be recovered with a far less number of measurements compared with the signal length [30].

In many practical issues, zero coefficients appear as cluster, which is called a block sparse structure. A number of sparse signal recovery methods utilize this structure like the pattern-coupled sparse Bayesian learning (PCSBL) method [31] that is described in Section 4. Hyperspectral images have a block sparse structure due to the proximity of similar materials. In this paper, a new algorithm for spectral unmixing based on block sparse recovery methods is presented in Section 4.

3.2 SPARSE BEYESIAN LEARNING

Here, we address the problem of block sparse signal recovery of vector measurement views that is expressed in Equation 1. The abundance fractions vector has a block sparse structure, but the exact properties of the block such as the location and size of each block are not known. In the sparse Bayesian learning framework, it is modeled as a Gaussian prior distribution:

p(x|α)=i=1np(xi|αi)p(x|\alpha)=\prod_{i=1}^{n}{p(x_{i}|\alpha_{i})} (5)

Where α=Δ{αi}\alpha\overset{\Delta}{=}{\{\alpha_{i}\}} are non-negative hyperparameters controlling the sparsity of the signal x. From Equation 5, it is understood that when αi\alpha_{i} approaches infinity, the corresponding coefficient xix_{i} becomes zero. The hyperparameters αi{\alpha_{i}} can be learned by maximizing their posterior probability. This property of the coefficients provides the possibility of retrieving block sparse signals in a more reliable manner. In sparse Bayesian learning model each parameter is dependent on its neighbors in addition to the coefficient itself. Therefore, if one of the coefficients is identified, this identification will affect the improvement of its two neighboring coefficients [32]. In this manner, by considering the neighbors, a framework for describing block sparse signals will be provided. The sparse Bayesian learning use Gamma distributions as hyperpriors over the hyperparameters.

3.3 Proposed Bayesian algorithm with known noise variance

The sparse Bayesian learning method can be used in recovery of block sparse signals. For ease of exposition, we first assume that the noise variance σ2\sigma^{2} is known a priori. Therefore, based on sparse Bayesian learning, the posterior distribution of x is expressed as follows:

p(x|α,y)p(x|α)p(y|α)p(x|\alpha,y)\propto p(x|\alpha)p(y|\alpha) (6)

It can be verified that the posterior probability p(x|α,y)p(x|\alpha,y) has a Gaussian distribution with the following mean and covariance

μ=σ2ϕATy\mu=\sigma^{2}\phi A^{T}y (7)
ϕ=(σ2ATA+D)1\phi=(\sigma^{2}A^{T}A+D)^{-1} (8)

Where D is a diagonal matrix with its iith diagonal element equal to [31]:

Di=Δ(αi+βαi1+βαi+1)D_{i}\overset{\Delta}{=}(\alpha_{i}+\beta\alpha_{i-1}+\beta\alpha_{i+1}) (9)

β\beta is a parameter indicating the pattern relevance between the coefficient and its neighboring coefficients xi+1,xi1x_{i+1},x_{i-1}. When β=0\beta=0, the prior distribution reduces to the prior for the conventional sparse Bayesian learning. When β>0\beta>0, we see that the sparsity of xix_{i} is not only controlled by the hyperparameter αi\alpha_{i}, but also by the neighboring hyperparameters αi+1,αi1\alpha_{i+1},\alpha_{i-1}. Given a set of estimated hyperparameters αi\alpha_{i}, the maximum a posterior (MAP) estimate of x is the mean of its posterior distribution,

xMAP=μ=(σ2D+ATA)1ATx_{MAP}=\mu=(\sigma^{2}D+A^{T}A)^{-1}A^{T} (10)

with respect to xx, the αi{\alpha_{i}} hyperparameters using the Maximum Expectation (EM) algorithm are estimated as follows [31]:

αi^=k/(0.5ωi+104)i=1,,n\hat{\alpha_{i}}=k/(0.5\omega_{i}+10^{-4})\;\;\;\;\forall i=1,…,n (11)

where k>0k>0 and ωi\omega_{i} represents the mean and weighted covariance as follows:

ωi=(μ^i2+ϕ^i,i)+(μ^i+12+ϕ^i+1,i+1)+(μ^i12+ϕ^i1,i1)\omega_{i}=(\hat{\mu}_{i}^{2}+\hat{\phi}_{i,i})+(\hat{\mu}_{i+1}^{2}+\hat{\phi}_{i+1,i+1})+(\hat{\mu}_{i-1}^{2}+\hat{\phi}_{i-1,i-1}) (12)

Updating hyperparamter αi\alpha_{i} as well as x continues until the stopping criteria in Equation 13 is achieved [33].

R2=μnewμold2ϵ||R||_{2}=||\mu_{new}-\mu_{old}||_{2}\leq\epsilon (13)

where μnew\mu_{new} and μold\mu_{old} are estimated values of abundance vectors xx in two consecutive iterations. In this paper ϵ\epsilon is set to 10810^{-8}.

3.4 Proposed Bayesian algorithm with unknown noise variance

In the previous section it was assumed that the noise variance is known. In this section, as is generally the case in reality, it is assumed that the noise variance is unknown. In this case, a new hyperparameter named γ\gamma is introduced that represents the variance noise:

p(γ)=Γ(c)1dcγcedγp(\gamma)=\Gamma(c)^{-1}d^{c}\gamma^{c}e^{-d\gamma} (14)

where c=d=104c=d=10^{-4}. The only difference between this method and the one in the previous section is the estimation of the hyperparameters α\alpha and the variance of the noise (or equivalent γ\gamma) that is again used in the EM algorithm. After applying the EM algorithm, the α\alpha and γ\gamma estimations are performed, where the estimate of α\alpha is the same as in the previous section. The only difference in this method is the noise equivalent estimation, γ\gamma, that is calculated as follows:

1γt+1=yAx22+(γ(t))1i=1nρi+2dm+2c\frac{1}{\gamma^{t+1}}=\frac{||y-Ax||_{2}^{2}+(\gamma(t))^{-1}\sum_{i=1}^{n}\rho_{i}+2d}{m+2c} (15)

Where ρi=Δ1ϕ^i,i(αi(t)+βαi1(t)+βαi+1(t))\rho_{i}\overset{\Delta}{=}1-\hat{\phi}_{i,i}(\alpha_{i}^{(t)}+\beta\alpha_{i-1}^{(t)}+\beta\alpha_{i+1}^{(t)}) and mm indicates the number of spectral bands. The update of γ\gamma is similar to the update procedure described in the previous section, until the stopping criterion stated in Equation 13 is met.

However, the assumption of known noise variance is not critical and the noise variance can be estimated from the measurements suggested by [34] in the form of:

σ^2=yAx^2/N\hat{\sigma}^{2}=||y-A\hat{x}||^{2}/N (16)

The overall flowchart of the proposed algorithm is depicted in Figure 2. Details and equations in this flowchart are discussed in the previous subsections.

Refer to caption
Figure 2: Flow chart of the proposed algorithm.

4 Experiments and Results

Two types of data have been used to conduct experiments for the performance evaluation of the proposed algorithm: synthetic and real datasets. The real dataset is recorded using hyperspectral sensors and is publicly available, and the synthetic dataset is generated using spectral libraries.

4.1 Synthetic Data

Synthetic images are generated using a set of spectral signatures selected from publicly available spectral libraries, such as the USGS Digital Spectral Library [35]. The data consists of 224 spectral bands, covering wavelengths ranging from 0.38 to 2.5 μm\mu m with a spectral resolution of 10 nm. To make linear combinations, the whole image is divided into small 5x5 blocks, the pixels within each block are pure and so have the same spectral signature, randomly selected from a set of endmembers. To simulate an image with mixed pixels, the resulting image passes through a low pass filter. In this way a hyperspectral scene in which mixed pixels exist were generated. The used low pass filter is a simple averaging filter, k×kk\times k, where kk is controlled by the degree of mixing. To remove pure pixels and generate highly mixed data, all pixels with fractional abundance greater than eighty percent were replaced with a mixture of all pixels with the same fractional abundance. That means a mixture in which each endmember has abundance of 1/c1/c, where cc is the number of endmembers. To simulate sensor errors and noises, Gaussian noise with zero mean is added to the combination, and it is assumed that the noise is spatially and spectrally uncorrelated. Therefore, the covariance matrix of noise will be equal to σ2I\sigma^{2}I. The signal-to-noise ratio (SNR) can be calculated using SNR=10log10(𝔼[xTx]/𝔼[nTn])\text{SNR}=10log10(\mathbb{E}[x^{T}x]/\mathbb{E}[n^{T}n]), where 𝔼[.]\mathbb{E}[.] is an expectation operator.

To compare the results quantitatively, abundance angle distance (AAD) and mean square error (MSE) evaluation criteria are used. These are common criteria in hyperspectral unmixing studies that are used to measure the similarity of results and reference values. The AAD criterion is defined as follows:

AAD=cos1(xTx^xx^)AAD=cos^{-1}(\frac{x^{T}\hat{x}}{||x||||\hat{x}||}) (17)

Where xx is the actual abundance vector of a pixel and x^\hat{x} is the estimated abundance vector. Another evaluation criterion is the MSE, which is calculated based on difference between the recorded and estimated results.

MSE=1ni=1n(yiy^i)2MSE=\frac{1}{n}\sum_{i=1}^{n}(y_{i}-\hat{y}_{i})^{2} (18)

where y^i\hat{y}_{i} and yiy_{i} are the estimated spectrum and the measured spectrum received by sensors for iith pixel, respectively.

The proposed algorithm is first implemented on synthetic data. The parameters to generate this dataset are set as follows. Twelve spectral signatures from the USGS library were randomly selected. A 5×55\times 5 low pass filter is used. Then Gaussian noise with zero mean at 6 different SNR values, i.e. SNR=[15,20,25,30,35,40] is added to the generated data. For the different values of the parameter β=[0.1,0.5,1]\beta=[0.1,0.5,1] the proposed algorithm is implemented.

First, based on the β\beta parameter, the evaluation charts are plotted according to AAD and MSE criteria. As it can be seen from both Figures 3 and 4, when the value of β\beta increases, both the MSE and AAD criteria decrease, indicating the convergence of the proposed algorithm to the least possible error. With increasing SNR, MSE and AAD criteria decrease, which is due to the high level of SNR value. The results is not sensitive to the value of β\beta as it can be seen in Figure 3. Selection of a nonzero value of 0<β<10<\beta<1 is sufficient for recovery improvement.

Refer to caption
Figure 3: Mean Square Error (MSE) relative to different β\betas in different SNR ranges in the proposed algorithm.
Refer to caption
Figure 4: Abundance Angle Distance (AAD) relative to different β\betas in different SNR ranges in the proposed algorithm.

In Figures 5 and 6, the results of the proposed algorithm is compared with SUnSAL [17] and RSU [20] algorithms in different SNRs and based on AAD and MSE evaluation criterion.

Refer to caption
Figure 5: Comparison of Abundance Angle Distance(AAD) of Proposed Algorithm with SUnSAL and RSU Algorithms.
Refer to caption
Figure 6: Comparison of Mean Squared Error (MSE) of Proposed Algorithm with SUnSAL and RSU Algorithms.

Tables 1 and  2 show the MSE and AAD metrics in different SNRs, respectively. In Figure 7, the estimated fractional abundance map of the two synthetic endmembers at SNR=25dB resulted from the proposed algorithm is compared with SUnSAL and RSU algorithms’ results. The comparison indicates that the proposed method more accurately estimates the fractional abundance map in comparison to the two other competing methods.

Table 1: Comparison of different methods in terms of MSE for different SNRs
SNR 15 20 25 30 35 40
SUnSAL [17] 0.01491 0.00789 0.00400 0.00194 0.00091 0.00043
CLSUnSAL [36] 0.01489 0.00785 0.00396 0.00192 0.00089 0.00044
RSU [20] 0.01497 0.00797 0.00404 0.00198 0.00095 0.00043
Proposed PCSBL 0.01398 0.00793 0.00396 0.00190 0.00080 0.00040
Table 2: Comparison of different methods in terms of AAD for different SNRs
SNR 15 20 25 30 35 40
SUnSAL [17] 1.701 1.528 1.405 1.405 1.313 1.244
CLSUnSAL [36] 1.702 1.485 1.463 1.389 1.297 1.240
RSU [20] 1.695 1.533 1.413 1.324 1.260 1.216
Proposed PCSBL 1.470 1.379 1.298 1.240 1.208 1.194

4.2 Real Data

To date, many hyperspectral sensors have recorded images in different missions that are used to test hyperspectral data processing algorithms. One of these hyperspectral sensors is NASA’s AVIRIS sensor that is introduced in Section 1. The data we have used in this research is from Cuprite region in Nevada [37]. There are many benefits using this data, including that this place has been used for remote sensing experiments since 1980 and a lot of high precision research is available on this area. To improve the separation performance, 36 low-SNR bands, that are significantly distorted with noise due to atmospheric effects, have been excluded from the 224-bands data cube. The remaining 188 bands have been then used in the experiments. In Figure 8, the estimated fractional abundance maps of the proposed algorithm are compared with thos of SUnSAL and RSU methods for two endmembers of the Cuprite data. Results show that the proposed method can effectively estimate the abundance fractions maps.

Refer to caption
Figure 7: Fractional abundance map of endmembers for synthetic dataset at SNR=15dB.
Refer to caption
Figure 8: Fractional abundance map of all 12 endmembers of Cuprite real dataset using proposed PCSBL method.

5 Conclusion

In this paper, a new method for spectral unmixing of hyperspectral images is presented that uses sparse Bayesian learning strategy for recovery of block sparse signals whose structure is unknown. In the Sparse Bayesian learning model, the sparsity of each coefficient depends only on the relevant hyperparameter. But in the proposed method, the sparsity also depends on its neighborhood hyperparameters. By maximizing posterior probability, hyperparameters in sparse signals can be estimated based on an iterative algorithm using expectation maximization method. To evaluate the performance of the proposed algorithm, both synthetic data and real AVIRIS Cuprite dataset are used. Results based on quantitative criteria MSE and AAD show that the proposed algorithm using the block sparse structure has better performance compared to other methods, even without knowing the exact location and size of each block. For future research, block sparse methods such as Block-IBA that do not need to know the block structure of the signal can be used for hyperspectral unmixing.

References

  • [1] D. A. Landgrebe, Signal theory methods in multispectral remote sensing, John Wiley & Sons, New Jersey (2003).
  • [2] X. Xu and Z. Shi, “Multi-objective based spectral unmixing for hyperspectral images,” ISPRS Journal of Photogrammetry and Remote Sensing 124, 54–69 (2017).
  • [3] S. Henrot, J. Chanussot, and C. Jutten, “Dynamical spectral unmixing of multitemporal hyperspectral images,” IEEE Transactions on Image Processing 25(7), 3219–3232 (2016).
  • [4] M.-D. Iordache, J. M. Bioucas-Dias, and A. Plaza, “Sparse unmixing of hyperspectral data,” IEEE Transactions on Geoscience and Remote Sensing 49(6), 2014–2039 (2011).
  • [5] Y. Altmann, S. McLaughlin, and A. Hero, “Robust linear spectral unmixing using anomaly detection,” IEEE Transactions on Computational Imaging 1(2), 74–85 (2015).
  • [6] L. Miao and H. Qi, “Endmember extraction from highly mixed data using minimum volume constrained nonnegative matrix factorization,” IEEE Transactions on Geoscience and Remote Sensing 45(3), 765–777 (2007).
  • [7] T. Li and C.-c. Ding, “Nonnegative matrix factorizations for clustering: A survey,” in Data Clustering, 149–176, Chapman and Hall/CRC, London (2018).
  • [8] J. Wodecki, P. Kruczek, A. Bartkowiak, et al., “Novel method of informative frequency band selection for vibration signal using nonnegative matrix factorization of spectrogram matrix,” Mechanical Systems and Signal Processing 130, 585–596 (2019).
  • [9] R. Rajabi and H. Ghassemian, “Spectral unmixing of hyperspectral imagery using multilayer nmf,” IEEE Geoscience and Remote Sensing Letters 12(1), 38–42 (2014).
  • [10] R. Rajabi and H. Ghassemian, “Sparsity constrained graph regularized nmf for spectral unmixing of hyperspectral data,” Journal of the Indian Society of Remote Sensing 43(2), 269–278 (2015).
  • [11] S. Khoshsokhan, R. Rajabi, and H. Zayyani, “Sparsity-constrained distributed unmixing of hyperspectral data,” IEEE Journal of Selected Topics in Applied Earth Observations and Remote Sensing 12(4), 1279–1288 (2019).
  • [12] F. Wen, “Diffusion least-mean p-power algorithms for distributed estimation in alpha-stable noise environments,” Electronics letters 49(21), 1355–1356 (2013).
  • [13] S. Khoshsokhan, R. Rajabi, and H. Zayyani, “Clustered multitask non-negative matrix factorization for spectral unmixing of hyperspectral data,” Journal of Applied Remote Sensing 13(2), 026509 (2019).
  • [14] N. Xu, X. Xiao, X. Geng, et al., “Spectral-spatial constrained sparse unmixing of hyperspectral imagery using a hybrid spectral library,” Remote Sensing Letters 7(7), 641–650 (2016).
  • [15] L. Sun, W. Ge, Y. Chen, et al., “Hyperspectral unmixing employing l 1- l 2 sparsity and total variation regularization,” International Journal of Remote Sensing 39(19), 6037–6060 (2018).
  • [16] X. Weng, W. Lei, and X. Ren, “Kernel sparse representation for hyperspectral unmixing based on high mutual coherence spectral library,” International Journal of Remote Sensing 41(4), 1286–1301 (2020).
  • [17] J. M. Bioucas-Dias and M. A. Figueiredo, “Alternating direction algorithms for constrained sparse regression: Application to hyperspectral unmixing,” in 2010 2nd Workshop on Hyperspectral Image and Signal Processing: Evolution in Remote Sensing, 1–4, IEEE (2010).
  • [18] D. M. Rogge, B. Rivard, J. Zhang, et al., “Iterative spectral unmixing for optimizing per-pixel endmember sets,” IEEE Transactions on Geoscience and Remote Sensing 44(12), 3725–3736 (2006).
  • [19] Z. Shi, W. Tang, Z. Duren, et al., “Subspace matching pursuit for sparse unmixing of hyperspectral data,” IEEE Transactions on Geoscience and Remote Sensing 52(6), 3256–3274 (2013).
  • [20] Y. Ma, C. Li, X. Mei, et al., “Robust sparse hyperspectral unmixing with 2,1\ell_{2,1} norm,” IEEE Transactions on Geoscience and Remote Sensing 55(3), 1227–1239 (2016).
  • [21] C. Li, Y. Liu, J. Cheng, et al., “Sparse unmixing of hyperspectral data with bandwise model,” Information sciences 512, 1424–1441 (2020).
  • [22] S. Zhang, J. Li, Z. Wu, et al., “Spatial discontinuity-weighted sparse unmixing of hyperspectral images,” IEEE Transactions on Geoscience and Remote Sensing 56(10), 5767–5779 (2018).
  • [23] D. Hong, L. Gao, J. Yao, et al., “Endmember-guided unmixing network (egu-net): A general deep learning framework for self-supervised hyperspectral unmixing,” IEEE Transactions on Neural Networks and Learning Systems 33(11), 6518–6531 (2022).
  • [24] V. S. S, V. S. Deshpande, and J. S. Bhatt, “A practical approach for hyperspectral unmixing using deep learning,” IEEE Geoscience and Remote Sensing Letters 19, 1–5 (2022).
  • [25] M. Korki, J. Zhang, C. Zhang, et al., “Iterative bayesian reconstruction of non-iid block-sparse signals,” IEEE Transactions on Signal Processing 64(13), 3297–3307 (2016).
  • [26] M. Korki, H. Zayyani, and J. Zhang, “Bayesian hypothesis testing for block sparse signal recovery,” IEEE Communications Letters 20(3), 494–497 (2016).
  • [27] C.-I. Chang, “A review of virtual dimensionality for hyperspectral imagery,” IEEE Journal of Selected Topics in Applied Earth Observations and Remote Sensing 11(4), 1285–1305 (2018).
  • [28] V. S S, J. S. Bhatt, and B. Chattopadhyay, “Virtual dimensionality of hyperspectral data: Use of multiple hypothesis testing for controlling type-i error,” IEEE Journal of Selected Topics in Applied Earth Observations and Remote Sensing 13, 2974–2985 (2020).
  • [29] C. L. Farrow, M. Shaw, H. Kim, et al., “Nyquist-shannon sampling theorem applied to refinements of the atomic pair distribution function,” Physical Review B 84(13), 134105 (2011).
  • [30] J. Wen, Z. Zhou, Z. Liu, et al., “Sharp sufficient conditions for stable recovery of block sparse signals by block orthogonal matching pursuit,” Applied and Computational Harmonic Analysis 47(3), 948–974 (2019).
  • [31] J. Fang, Y. Shen, H. Li, et al., “Pattern-coupled sparse bayesian learning for recovery of block-sparse signals,” IEEE Transactions on Signal Processing 63(2), 360–372 (2015).
  • [32] L. Wang, L. Zhao, S. Rahardja, et al., “Alternative to extended block sparse bayesian learning and its relation to pattern-coupled sparse bayesian learning,” IEEE Transactions on Signal Processing 66(10), 2759–2771 (2018).
  • [33] J. A. Fessler and A. O. Hero, “Space-alternating generalized expectation-maximization algorithm,” IEEE Transactions on signal processing 42(10), 2664–2677 (1994).
  • [34] H. Zayyani, M. Babaie-Zadeh, and C. Jutten, “An iterative bayesian algorithm for sparse component analysis in presence of noise,” IEEE Transactions on Signal Processing 57(11), 4378–4390 (2009).
  • [35] R. Clark, G. Swayze, R. Wise, et al., “USGS digital spectral library splib06a http://speclab.cr.usgs.gov/spectral.lib06,” (2007).
  • [36] Y. Altmann, M. Pereyra, and J. Bioucas-Dias, “Collaborative sparse regression using spatially correlated supports - application to hyperspectral unmixing,” IEEE Transactions on Image Processing 24(12), 5800–5811 (2015).
  • [37] “Aviris cuprite nevada dataset.” url :http://aviris.jpl.nasa.gov/data/free_data.html.