This paper was converted on www.awesomepapers.org from LaTeX by an anonymous user.
Want to know more? Visit the Converter page.

\UseRawInputEncoding\useunder

\ul

Spectral Variability Augmented Sparse Unmixing of Hyperspectral Images

Ge Zhang, Shaohui Mei, Mingyang Ma, Yan Feng, and Qian Du  This work is supported by National Natural Science Foundation of China (62171381) and the Fundamental Research Funds for the Central Universities. Corresponding author: Shaohui Mei ([email protected]).Ge Zhang, Shaohui Mei, Mingyang Ma, and Yan Feng are with the School of Electronics and Information, Northwestern Polytechnical University, Xi’an 710129, China.Qian Du is with the Department of Electrical and Computer Engineering, Mississippi State University, Starkville, MS 39762, USA.
Abstract

Spectral unmixing (SU) expresses the mixed pixels existed in hyperspectral images as the product of endmember and abundance, which has been widely used in hyperspectral imagery analysis. However, the influence of light, acquisition conditions and the inherent properties of materials, results in that the identified endmembers can vary spectrally within a given image (construed as spectral variability). To address this issue, recent methods usually use a priori obtained spectral library to represent multiple characteristic spectra of the same object, but few of them extracted the spectral variability explicitly. In this paper, a spectral variability augmented sparse unmixing model (SVASU) is proposed, in which the spectral variability is extracted for the first time. The variable spectra are divided into two parts of intrinsic spectrum and spectral variability for spectral reconstruction, and modeled synchronously in the SU model adding the regular terms restricting the sparsity of abundance and the generalization of the variability coefficient. It is noted that the spectral variability library and the intrinsic spectral library are all constructed from the In-situ observed image. Experimental results over both synthetic and real-world data sets demonstrate that the augmented decomposition by spectral variability significantly improves the unmixing performance than the decomposition only by spectral library, as well as compared to state-of-the-art algorithms.

Index Terms:
Hyperspectral Unmixing, Spectral Variability, Sparse Unmixing, Spectral Library.

I Introduction

With the rapid development of remote sensing technology, hyperspectral image (HSI) collecting abundant spectral information with hundreds of contiguous bands, has been widely used in different kinds of applications, such as geoinformation science, space research, material mapping, and geological exploration [1]. However, the phenomenon of mixed pixels is widely existed due to insufficient spatial resolution of imaging system, changeable atmospheric environment, and complex distribution of ground objects. The decomposition of a mixed pixel into spectral signatures (endmembers) with corresponding proportions (abundances) is known as hyperspectral unmixing [2].

Linear mixing model (LMM) has been widely used for spectral mixture analysis as its computational tractability and flexibility and it supposes that the spectra collected by the imaging spectrometer are a linear combination of endmembers weighted by their corresponding abundance fractions [3]. However, nonlinearity and spectral variability affect the performance of spectral unmixing [4]. Techniques based on geometry, statistics, and nonnegative matrix factorization extract endmembers directly from the HSI. In general, an endmember extraction step is applied, and then, abundance value for each pixel is estimated. There are many algorithms for endmember extraction, such as N-FINDR [5], pixel purity index (PPI) [6], and vertex component analysis (VCA) [7]. Nevertheless, these algorithms require pure pixel assumption, and it is not always satisfied due to the spatial resolution. Others of them extract virtual endmembers without physical meaning in a given HSI [8].

Recently, due to the wide availability of spectral libraries, like the U.S. Geological Survey (USGS) digital spectral library, a semi-supervised spectral unmixing approach, called sparse regression-based method, has been shown to circumvent the drawbacks introduced by such virtual endmembers and the unavailability of pure pixels [9, 10]. It assumes that the mixed pixels can be expressed in the form of linear combinations of a small number of pure spectral signatures from a (potentially very large) spectral library known in advance. Another significant advantage is that it does not need to extract endmembers from the hyperspectral data or estimate the number of the endmembers. Generally, the number of endmembers in the scene is small compared with the number of endmembers in the spectral library, which means that only a small number of endmembers contribute to the mixed pixel. Therefore, the abundance vector of the mixed pixel estimated by sparse unmixing is expected to be sparse. These new perspectives introduced by sparse unmixing fostered advanced developments in the field [11, 12, 13].

In the past few years, several algorithms have been developed to enforce the sparsity on the solution of SU [14, 15]. The sparse unmixing algorithm via variable splitting and augmented Lagrangian (SUnSAL) [9] adopts the L1L_{1} regularizer on the abundance matrix to measure the sparsity of the abundance vector in each pixel. The collaborative SUnSAL (CLSUnSAL) [16] introduces the L2,1L_{2,1} regularizer to constrain the pixels in HSIs to share the same active set of endmembers and ensure that all the abundance vectors exhibit global row sparsity. Based on the alternating direction method of multipliers (ADMM) [17], the corresponding sparse regression problems are all convex and can be settled efficiently. Moreover, the spatial information acts a pivotal part in sparse unmixing as the corresponding problem is usually made easier on a local scale [18]. Total variation (TV) [19, 20, 21] regularization exploits the spatial information via a first-order pixel neighborhood system to significantly improve the unmixing results, but it may yield oversmoothness and blurred boundaries. Zhang et al. [22] proposed a spectral Cspatial weighted sparse unmixing (S2WSU) framework constrained simultaneously from the spectral and spatial domains to further imposing sparsity on the solution. To obtain spatial-contextual information, a fast multiscale spatial regularization unmixing algorithm (MUA) was proposed by Borsoi et al. [23]. In this algorithm, superpixel-based segmentation is performed before unmixing to form a coarse domain, and the unmixing result for the coarse domain forms a multiscale spatial regularization for unmixing of the original domain. Another solution is to focus on the local spectral similarity of the HSIs rather than the simple proximity of the location. The local spectral similarity preserving (LSSP) constraint [24] was proposed to preserve spectral similarity in a local region given that adjacent pixels share not only the same endmembers with high probability but also approximated fractional abundances. Meanwhile, the low-rank constraint of the abundance matrix has been increasingly adopted for sparse unmixing, providing a new perspective for spatial correlation [25]. Giampouras et al. [26] simultaneously imposed single sparsity and low rankness on abundance matrices, taking into account both sparsity and spatial correlation information in HSIs. The joint-sparse-blocks and low-rank unmixing (JSpBLRU) algorithm [27] further imposed the joint-sparsity-blocks structure and low rankness on abundance matrices for pixels in a sliding window. Recently, the sparse unmixing method named superpixel-based reweighted low-rank and total variation (SUSRLR-TV) [28] was proposed by using superpixel segmentation to extract the homogeneous regions, and imposing a low-rank constraint to promote the correlation of each superpixel s abundance matrix.

However, the spectral signatures of the materials contained in hyperspectral images can be significantly affected by spectral variability, which results from different imaging conditions including atmospheric effects, illumination, topographic changes, and the intrinsic variation of the spectral signatures of the materials (i.e., due to physicochemical differences), especially in a hyperspectral image with a higher spatial resolution [29, 30, 31]. Therefore, the LMM hardly makes an accurate unmixing in reality, due to such ubiquitous error that passively transfer the unpredicted errors into LMM [32]. The existing methods could be basically divided into two groups: 1) dictionary or bundle-based approaches [33, 34], which try to model endmembers by a certain number of instances of each material, 2) model-based approaches, which describes the endmember variability by a specific statistical distribution [35, 36, 37] or by incorporating the variability in the mixing model based on physically motivated concepts [38, 39, 40]. This work is developed following the first approach.

The key idea in dictionary or bundle-based approaches consists in extracting several instances of each endmember in order to build a dictionary, which is then used for spectral unmixing, offering more than one representative spectral signature per material [41, 42]. Representing each endmember class with large spectral libraries of spectra as a priori does not require the assumption of any particular spectral distribution within a single endmember class. Methods for this category select the appropriate spectra, or bands, from different spectral sets according to pre-defined criteria for weakening the effect of endmember variab ility, which can be roughly divided into four groups: multiple endmember spectral mixture analysis (MESMA), sparse unmixing, spectral transformations, and machine learning.

The basic principle behind MESMA [43] is to iteratively search for the combination of endmember signatures in the library that, among all possibilities, enables the closest reconstruction of each observed pixel under the LMM. The MESMA algorithm and its variants formulate spectral unmixing as a computationally demanding optimization problem and achieve good quality [44]. Chen et al. [45] proposed a sparse multiple-endmember spectral mixture model (SMESMM) by using a block sparse algorithm to obtain an initial block sparse solution and resolving the mixed pixel using the selected land cover materials. Zhang et al. [46] incorporated spatial information in MESMA by using segmentation algorithms to divide the image into different homogeneous objects, which are then unmixed individually by using a library that is also constructed from object based spectra. Borsoi et al. [47] leveraged the power of deep generative models to learn the statistical distribution of the endmembers available in the existing libraries, and then drew new samples to augment the spectral libraries, improving the overall quality of the unmixing process.

Compared to MESMA, sparse unmixing formulations use mathematical relaxations that are computationally easier to solve. A typical approach [48] was to modify the LMM for unmixing mineral spectra in mining applications by including an additional term representing the mixture of the background spectrum of the endmembers. This background spectrum was defined as the low-frequency part of the spectral signatures and estimated a priori from the library as a parametric function of smooth splines. The performance of an L1L_{1} norm-based sparse unmixing framework under this model was reported to be similar to MESMA, albeit at a much smaller computational cost.

In the spectral transformations group, instead of the original reflectance data, transformed spectral information with less spectral variability are used as input to the SMA, like derivative spectra and wavelet transformed spectra [49]. However, spectral transformations are empirically oriented techniques and require a significant degree of expert knowledge about the underlying application [50]. Other types of methods are more machine learning-oriented, which allow to learn the function linking the endmembers available in the dictionary (or in a supervised fashion using an a priori available dictionary). The training can be performed for instance by simulating mixed pixels with endmembers from the dictionary in different proportions [51], or models the latent function between spectra and abundances in a training set and predicts abundances from a given pixel spectrum under Gaussian process framework [52]. Machine learning algorithms provide more flexible ways but these methods have the drawbacks of not explicitly modeling the variability, and at a large computational complexity.

Traditional spectral unmixing (SU) algorithms neglect the spectral variability of the endmembers, which propagates significant modeling errors throughout the whole unmixing process and compromises the quality of the results. The recent library-based SU methods address spectral variability by using libraries of spectra that had to be acquired a priori (e.g., through in situ measurements) without introducing explicit spectral variability term, which limited the applicability of these approaches. Therefore, a spectral variability augmented sparse unmixing model (SVASU) is proposed, which provides a new way to construct spectral variability explicitly from the spectral library for sparse unmixing. Finally, extensive experiments over both synthetic and real-world datasets are conducted to validate the effectiveness and robustness of the proposed SVASU method for hyperspectral sparse unmixing.

The remainder of this paper is organized as follows. Section II introduces the basic idea of sparse unmixing. Section III presents the proposed SVASU model for SU of hyperspectral images. Section IV reports and discusses experimental results on both synthetic and real datasets. Finally, Section V draws a conclusion.

II Sparse Unmixing

Refer to caption
Figure 1: Basic flowchart of sparse unmixing through SVASU.

II-A Linear Mixture Model

The Linear Mixture Model (LMM), which is well-known for its simplicity and explicit physical meaning, has been widely utilized to model the relationship between mixtures and endmembers. In the LMM, the photons reflected from different ground objects within one pixel are assumed not to interfere with each other. As a result, the observed hyperspectral image matrix 𝐑b×n\mathbf{R}\in\mathcal{R}^{b\times n} is assumed to be the product of endmember matrix 𝐄b×p\mathbf{E}\in\mathcal{R}^{b\times p} and its corresponding abundance matrix 𝐀p×n\mathbf{A}\in\mathcal{R}^{p\times n} plus a noise matrix 𝐍b×n\mathbf{N}\in\mathcal{R}^{b\times n}. Their relationship can be formulated as:

𝐑=𝐄𝐀+𝐍,\mathbf{R}=\mathbf{EA}+\mathbf{N}, (1)

in which bb is the number of bands, pp is the number of endmembers, and nn is the number of pixels in the image. Here, the ii-th column vectors of 𝐄\mathbf{E} and 𝐀\mathbf{A}, denoted by 𝐞i\mathbf{e}_{i} and 𝐚i\mathbf{a}_{i}, correspond to the ii-th endmember and the abundance of the ii-th spectral pixel 𝐫i\mathbf{r}_{i}, respectively.

In order to make the LMM physically meaningful, the abundance sum-to-one constraint (ASC) and the abundance nonnegative constraint (ANC) must be imposed, which can be expressed as:

i=1paij=1,\sum_{i=1}^{p}{a_{ij}}=1, (2)
aij0i=1,2,,p,j=1,2,,n.a_{ij}\geq 0\qquad i=1,2,\ldots,p,j=1,2,\ldots,n. (3)

Generally, the ASC can be embedded into this model by adding an additional pseudo band to the data matrix and the endmember matrix [53]. Therefore, in the following analysis, only the ANC is explicitly considered.

II-B Sparse Unmixing

Due to the wide availability of spectral libraries, sparse unmixing models use mathematical relaxations that are computationally easier to solve. Let 𝐌b×m\mathbf{M}\in\mathbb{R}^{b\times m} be a large spectral library containing mm spectral signatures, and 𝐀m×n\mathbf{A}\in\mathbb{R}^{m\times n} denotes the abundance maps corresponding to library 𝐌\mathbf{M} for the observed data 𝐑\mathbf{R}. As the number of endmembers involved in a mixed pixel is usually very small when compared with the size of the spectral library, the abundance matrix 𝐀\mathbf{A} contains many zero values, which means 𝐀\mathbf{A} is sparse. Thus, the unmixing problem can be formulated as an l2l0l_{2}-l_{0} optimization problem,

min𝐀12𝐑𝐌𝐀F2+λ𝐀0s.t.𝐀0,\mathop{\min}\limits_{\mathbf{A}}\frac{1}{2}\|\mathbf{R}-\mathbf{MA}\|_{F}^{2}+\lambda\|\mathbf{A}\|_{0}\ \ \ \mathrm{s.t.}\ \mathbf{A}\geq 0, (4)

where F\|\cdot\|_{F} is the Frobenius norm, 0\|\cdot\|_{0} is the l0l_{0}-norm which denotes the number of nonzero components of a vector, and λ\lambda is a regularization parameter used to adjust the weight of the sparsity in this model.

Problem (4) is nonconvex and difficult to be solved. However, the l0l_{0}-norm constrained sparsity can be relaxed using l1l_{1} norm, by which the sparse unmixing algorithm via variable splitting and augmented Lagrangian (SUnSAL) [9] using the l2l1l_{2}-l_{1} norm to solve the unmixing problem as follows,

min𝐀12𝐑𝐌𝐀F2+λ𝐀1,1s.t.𝐀0,\mathop{\min}\limits_{\mathbf{A}}\frac{1}{2}\|\mathbf{R}-\mathbf{MA}\|_{F}^{2}+\lambda\|\mathbf{A}\|_{1,1}\ \ \ \mathrm{s.t.}\ \mathbf{A}\geq 0, (5)

where 𝐀1,1=j=1n𝐚j\|\mathbf{A}\|_{1,1}=\sum_{j=1}^{n}\|\mathbf{a}_{j}\| with 𝐚j\mathbf{a}_{j} being the jj-th column of 𝐀\mathbf{A}. However, the Problem (5) is heavily influenced by the high correlation of spectral libraries due to the underdetermined nature. To admit a sufficiently sparse solution for sparse unmixing will guarantee a more accurate abundance estimation. In [16], the Collaborative Sparse Unmixing via variable Splitting and Augmented Lagrangian (CLSUnSAL) algorithm imposes the joint sparsity with an l2,1l_{2,1} mixed norm among the endmembers simultaneously for all of the pixels, whose objective function can be defined as follows,

min𝐀12𝐑𝐌𝐀F2+λi=1m𝐚i2s.t.𝐀0,\mathop{\min}\limits_{\mathbf{A}}\frac{1}{2}\|\mathbf{R}-\mathbf{MA}\|_{F}^{2}+\lambda\sum_{i=1}^{m}\|\mathbf{a}^{i}\|_{2}\ \ \ \mathrm{s.t.}\ \mathbf{A}\geq 0, (6)

where 𝐚i\mathbf{a}^{i} denotes the ii-th line of 𝐀\mathbf{A} and i=1m𝐚i2\sum_{i=1}^{m}\|\mathbf{a}^{i}\|_{2} is the so-called l2,1l_{2,1} mixed norm.

III Sparse Unmixing Via SVASU

A considerable number of such sparse unmixing methods has been developed using libraries of spectra that originally had to be acquired a priori (e.g., through in situ measurements), which used to limit the applicability of these approaches. An important recent development concerns methods that can extract spectral libraries directly from the observed images or generate them using physics-based mathematical models of material spectra. This supports the widespread applicability of library-based sparse unmixing techniques in situations where spectral libraries are not available or cannot be built.

However, the spectral signature variability affects the performance of sparse unmixing task since that the inaccurate endmember representation undoubtedly results in the propagation of errors in abundance estimates derived using SMA. Thus, to address spectral variability problem, the spectral variability library is introduced to assist the spectral libraries for better signature approximation in sparse unmixing. Specifically, the introduction of spectral variability library makes a spectrum signature divided into two independent representation parts, which allows the inherent spectral variability into materials to be represented separately and explicitly.

III-A Extraction of In-situ Spectral Library by SPEE

In a hyperspectral image, pixels from a homogeneous ground object often present in adjacent areas, which result in lots of pure spatial neighborhoods in the image. If all these pure spatial neighborhoods can be detected, their representative spectral signatures can be selected as endmember candidates. Different endmember candidates can be considered as coming from homogeneous ground objects provided that the spatial neighborhoods they represent cover pixels in common [54]. Based on this spatial refinement scheme, the remaining spatially independent endmember candidates become more representative since they represent larger areas. The number of endmember candidates to constrcut the in-situ spectral library will also be significantly reduced. Note that the in-situ spectral library 𝐗\mathbf{X} is extracted from the observed image 𝐑\mathbf{R} (refer to [54]), containing multiple spectra of each possible material (a total of pp materials), written as

𝐗=[𝐗(1),𝐗(2),,𝐗(p)]\displaystyle\mathbf{X}=[\mathbf{X}^{(1)},\mathbf{X}^{(2)},\cdot\cdot\cdot,\mathbf{X}^{(p)}] (7)
𝐗(i)=[𝐞1(i),𝐞2(i),,𝐞m(i)],i=1,2,,p\displaystyle\mathbf{X}^{(i)}=[\mathbf{e}^{(i)}_{1},\mathbf{e}^{(i)}_{2},\cdot\cdot\cdot,\mathbf{e}^{(i)}_{m}],\ i=1,2,\cdot\cdot\cdot,p

in which 𝐗(i)\mathbf{X}^{(i)} represents the library subset corresponding to the ii-th possible material and is composed of mm spectral signatures of the current material.

III-B Segmentation of Endmember Library and Spectral Variability Library

Through the previous section, we have obtained the image spectral library including the spatially independent and spectrally pure pixels in the homogenous region according to the pixel purity. Principal Component Analysis (PCA) is a classic method widely used in signal feature extraction, data dimensionality reduction and compression. It retains the main information of the original image data by retaining the first kk principal components corresponding to the larger feature value, and the remaining principal components correspond to noise, spectral variability, and/or outliers.

In this section, PCA is performed on its covariance matrix after standardizing the pixel data XX of the in-situ spectral library as

𝐗¯\displaystyle\bar{\mathbf{X}} =1ni=1n𝐗i,𝐗^=𝐗𝐗¯,\displaystyle=\frac{1}{n}\sum_{i=1}^{n}\mathbf{X}_{i},\ \ \ \ \ \ \ \ \ \ \ \hat{\mathbf{X}}=\mathbf{X}-\bar{\mathbf{X}}, (8)
𝐂\displaystyle\mathbf{C} =1n𝐗^𝐗^T,[𝐖,𝐃]=PCA(𝐂),\displaystyle=\frac{1}{n}\hat{\mathbf{X}}\hat{\mathbf{X}}^{T},\ \ \ \ \ \ \ \ \ \ \ \ [\mathbf{W},\mathbf{D}]=PCA(\mathbf{C}),
𝐖\displaystyle\mathbf{W} =[𝐰1,𝐰2,,𝐰n],𝐃=diag(λ1,λ2,,λn),\displaystyle=[\mathbf{w}_{1},\mathbf{w}_{2},\cdot\cdot\cdot,\mathbf{w}_{n}],\ \ \mathbf{D}=diag(\lambda_{1},\lambda_{2},\cdot\cdot\cdot,\lambda_{n}),\ \

where λ1>λ2>>λn\lambda_{1}>\lambda_{2}>\cdot\cdot\cdot>\lambda_{n} are the eigenvalues of its covariance matrix 𝐂\mathbf{C}, and 𝐖\mathbf{W} is composed of eigenvectors of the corresponding eigenvalue matrix 𝐃\mathbf{D}. Among all the principal components, the first kk correspond to the unique feature types of various pure pixels in the homogeneous area, while the other principal components correspond to the local spectral variability of the pixels in the homogeneous area.

The determination of kk can be achieved by comparing the size of the characteristic value. Set a threshold ζ\zeta, if the cumulative eigenvalues corresponding to the first kk pivotal elements account for the percentage of the total eigenvalues to reach the threshold:

i=1kλi/i=1nλiζ.\displaystyle{\raise 3.01385pt\hbox{${\sum\limits_{i=1}^{k}{{\lambda_{i}}}}$}\!\mathord{\left/{\vphantom{{\sum\limits_{i=1}^{k}{{\lambda_{i}}}}{\sum\limits_{i=1}^{n}{{\lambda_{i}}}}}}\right.\kern-1.2pt}\!\lower 3.01385pt\hbox{${\sum\limits_{i=1}^{n}{{\lambda_{i}}}}$}}\geq\zeta. (9)

Therefore, the first kk principal components can be calculated by the first kk eigenvectors,

𝐅1\displaystyle\mathbf{F}_{1} =𝐖kT𝐗^,\displaystyle=\mathbf{W}_{k}^{T}\hat{\mathbf{X}}, (10)
𝐖k\displaystyle\mathbf{W}_{k} =[𝐰1,𝐰2,,𝐰k],kn,\displaystyle=[\mathbf{w}_{1},\mathbf{w}_{2},\cdot\cdot\cdot,\mathbf{w}_{k}],k\leq n,

Then, the reconstructed spectral library by dominant characteristic components is termed as the endmember library,

Libraryendmember\displaystyle Library_{\rm{endmember}} =𝐖𝐅1+𝐗¯=𝐖k𝐖kT𝐗+𝐗¯.\displaystyle=\mathbf{W}\mathbf{F}_{1}+\bar{\mathbf{X}}=\mathbf{W}_{k}\mathbf{W}_{k}^{T}\mathbf{X}+\bar{\mathbf{X}}. (11)

Similarly, the next nkn-k principal components are constructed as the spectral variability library as

𝐅2=𝐖nkT𝐗^,𝐖nk=[𝐰k+1,𝐰k+2,,𝐰n],\displaystyle\mathbf{F}_{2}=\mathbf{W}_{n-k}^{T}\hat{\mathbf{X}},\ \ \ \mathbf{W}_{n-k}=[\mathbf{w}_{k+1},\mathbf{w}_{k+2},\cdot\cdot\cdot,\mathbf{w}_{n}], (12)
Libraryvariability=𝐖𝐅2+𝐗¯=𝐖nk𝐖nkT𝐗^+𝐗¯.\displaystyle Library_{\rm{variability}}=\mathbf{W}\mathbf{F}_{2}+\bar{\mathbf{X}}=\mathbf{W}_{n-k}\mathbf{W}_{n-k}^{T}\hat{\mathbf{X}}+\bar{\mathbf{X}}.

III-C Spectral Variability Augmented Sparse Unmixing (SVASU)

Fig. 1 displays the flowchart of proposed SVASU where the well-known Jasper data set is taken as an example. The proposed SVASU adopts a two-order decomposition structure, in which the first order is to decompose the hyperspectral image data into the product of endmember library and abundance fractions roughly, the second order is to further represent the reconstruction error of pixels from first order as the influence of spectral variability library.

Let 𝐕b×l\mathbf{V}\in\mathbb{R}^{b\times l} be the spectral variability library containing ll spectral variability signatures, and 𝐁l×n\mathbf{B}\in\mathbb{R}^{l\times n} denotes the coefficients corresponding to library 𝐕\mathbf{V}. Based on the objective function in Eq. (6), an extra spectral variability augmented data fitting term is introduced in the proposed SVASU model, along with the coefficient generalization regularized term, which is expressed as follows,

Θ=min𝐀,𝐁𝐑𝐌𝐀F2+α𝐑𝐌𝐀𝐕𝐁F2\displaystyle\Theta=\mathop{\min}\limits_{\mathbf{A},\mathbf{B}}{\left\|{\mathbf{R}-\mathbf{MA}}\right\|_{F}^{2}}+\alpha{\left\|{\mathbf{R}-\mathbf{MA}-\mathbf{VB}}\right\|_{F}^{2}} (13)
+β𝐀2,1+γ𝐁F2s.t.𝐀0,\displaystyle+\beta{\left\|\mathbf{A}\right\|_{2,1}}+\gamma{\left\|\mathbf{B}\right\|_{F}^{2}}\ \ \ \ \ \ \ \ \ \mathrm{s.t.}\ \mathbf{A}\geq 0,

where α\alpha balances the contribution of data fitting from both spectral spectra and spectral variability, β\beta and γ\gamma controls the sparsity of abundance and the generalization of spectral variability coefficient.

Following [55], we relax the term 𝐀2,1\|\mathbf{A}\|_{2,1} by Tr(𝐀T𝐃𝐀)\rm{Tr}(\mathbf{A}^{T}\mathbf{D}\mathbf{A}). Thus the function in (13) can be written as

Θ=min𝐀,𝐁𝐑𝐌𝐀F2+α𝐑𝐌𝐀𝐕𝐁F2\displaystyle\Theta=\mathop{\min}\limits_{\mathbf{A},\mathbf{B}}{\left\|{\mathbf{R}-\mathbf{MA}}\right\|_{F}^{2}}+\alpha{\left\|{\mathbf{R}-\mathbf{MA}-\mathbf{VB}}\right\|_{F}^{2}} (14)
+βTr(𝐀T𝐃𝐀)+γ𝐁F2s.t.𝐀0,\displaystyle+\beta Tr(\mathbf{A}^{T}\mathbf{D}\mathbf{A})+\gamma{\left\|\mathbf{B}\right\|_{F}^{2}}\ \ \mathrm{s.t.}\ \mathbf{A}\geq 0,

where 𝐃m×m\mathbf{D}\in\mathbb{R}^{m\times m} is a diagonal matrix with the ii-th diagonal element formulated as follows, in which ϵ>0\epsilon>0 is a stabilization parameter,

𝐃ii=12𝐚iT𝐚i+ϵ.\displaystyle{\mathbf{D}}_{ii}=\frac{1}{2\sqrt{\mathbf{a}_{i}^{T}\mathbf{a}_{i}+\epsilon}}. (15)

In order to integrate non-negative constraint conditions into the objective function 𝚯\bf{\Theta}, we introduce Lagrangian multipliers δ+m×n\delta\in\mathbb{R}_{+}^{m\times n} to restrict 𝐀0\mathbf{A}\geq 0.Therefore, the function (14) is equivalent to the following function:

𝚯=min𝐀,𝐁𝐑𝐌𝐀F𝟐+α𝐑𝐌𝐀𝐕𝐁F𝟐\displaystyle\bf{\Theta}=\mathop{\min}\limits_{\mathbf{A},\mathbf{B}}{\left\|{\mathbf{R}-\mathbf{MA}}\right\|_{\emph{F}}^{2}}+\alpha{\left\|{\mathbf{R}-\mathbf{MA}-\mathbf{VB}}\right\|_{\emph{F}}^{2}} (16)
+βTr(𝐀T𝐃𝐀)+γ𝐁F2Tr(δ𝐀T).\displaystyle+\beta{Tr(\mathbf{A}^{T}\mathbf{D}\mathbf{A})}+\gamma{\left\|\mathbf{B}\right\|_{\emph{F}}^{2}}-Tr(\delta\mathbf{A}^{T}).

When 𝐃\mathbf{D} is fixed, the original problem can be turned into a general convex optimization problem. To update all matrixes of the objective function, we use an alternating projected gradient method. By setting the partial derivatives of with respect to 𝐀\mathbf{A} and 𝐁\mathbf{B} respectively, we obtain the following function:

{𝚯𝐀=2((1+α)𝐌T𝐌𝐀(1+α)𝐌T𝐑+α𝐌T𝐕𝐁+β𝐃𝐀)δ𝚯𝐁=2(α𝐕T𝐌𝐀α𝐕T𝐑+α𝐕T𝐕𝐁+γ𝐁)\left\{\begin{aligned} \frac{{\partial\bf{\Theta}}}{{\partial\mathbf{A}}}&=2((1+\alpha){\mathbf{M}^{T}}\mathbf{MA}-(1+\alpha){\mathbf{M}^{T}}\mathbf{R}\\ &\ \ \ \ \ \ \ \ +\alpha{\mathbf{M}^{T}}\mathbf{VB}+\beta\mathbf{D}\mathbf{A})-\delta\\ \frac{{\partial\bf{\Theta}}}{{\partial\mathbf{B}}}&=2(\alpha{\mathbf{V}^{T}}\mathbf{MA}-\alpha{\mathbf{V}^{T}}\mathbf{R}+\alpha{\mathbf{V}^{T}}\mathbf{VB}+\gamma\mathbf{B})\end{aligned}\right. (17)

We take advantage of KKT conditions in this mathematical optimization problem. Therefore, the following condition should be satisfied:

{2((1+α)𝐌T𝐌𝐀(1+α)𝐌T𝐑+α𝐌T𝐕𝐁+β(𝐃+𝐃T)𝐀)ik𝐀ik=02(α𝐕T𝐌𝐀α𝐕T𝐑+α𝐕T𝐕𝐁+γ𝐁)ij𝐁ij=0\left\{\begin{aligned} &2((1+\alpha){\mathbf{M}^{T}}\mathbf{MA}-(1+\alpha){\mathbf{M}^{T}}\mathbf{R}\\ &\ \ \ \ \ \ \ \ \ \ +\alpha{\mathbf{M}^{T}}\mathbf{VB}+\beta(\mathbf{D}+{\mathbf{D}^{T}})\mathbf{A})_{ik}\cdot\mathbf{A}_{ik}=0\\ &2(\alpha{\mathbf{V}^{T}}\mathbf{MA}-\alpha{\mathbf{V}^{T}}\mathbf{R}+\alpha{\mathbf{V}^{T}}\mathbf{VB}+\gamma\mathbf{B})_{ij}\cdot\mathbf{B}_{ij}=0\end{aligned}\right. (18)

Then, we design the following update rules. The convergence of two update rules will be analyzed in the APPENDIX.

{𝐀ik=𝐀~ik((𝐌T𝐑+α𝐌T𝐑α𝐌T𝐕𝐁)ik((1+α)𝐌T𝐌𝐀~+β𝐃𝐀~)ik)14𝐁ij=𝐁~ij((α𝐕T𝐑α𝐕T𝐌𝐀)ij(α𝐕T𝐕𝐁~+γ𝐁~)ij)14\left\{\begin{aligned} &\mathbf{A}_{ik}^{\prime}=\mathbf{\tilde{A}}_{ik}(\frac{{({\mathbf{M}^{T}}\mathbf{R}+\alpha\mathbf{M}^{T}\mathbf{R}-\alpha\mathbf{M}^{T}\mathbf{VB})_{ik}}}{{((1+\alpha){\mathbf{M}^{T}}\mathbf{M}\mathbf{\tilde{A}}+\beta\mathbf{D}\mathbf{\tilde{A}})_{ik}}})^{\frac{1}{4}}\\ &\mathbf{B}_{ij}^{\prime}=\mathbf{\tilde{B}}_{ij}(\frac{{(\alpha{\mathbf{V}^{T}}\mathbf{R}-\alpha{\mathbf{V}^{T}}\mathbf{MA})_{ij}}}{{(\alpha{\mathbf{V}^{T}}\mathbf{V}\mathbf{\tilde{B}}+\gamma\mathbf{\tilde{B}})_{ij}}})^{\frac{1}{4}}\end{aligned}\right. (19)

Therefore, an iterative procedure is adopted. In each iteration, 𝐀\mathbf{A} is calculated with the current 𝐃\mathbf{D} and 𝐁\mathbf{B}, and then 𝐁\mathbf{B} and 𝐃\mathbf{D} are updated based on the current calculated 𝐀\mathbf{A}. The iterative procedure is repeated until the algorithm converges. The pseudo code of SVASU is illustrated in Algorithm 1.

Algorithm 1 SVASU.
1:Hyperspectral data matrix 𝐑b×n\mathbf{R}\in\mathcal{R}^{b\times n}; Observed image spectral library 𝐌b×m\mathbf{M}\in\mathbb{R}^{b\times m}; Spectral variability library 𝐕b×l\mathbf{V}\in\mathbb{R}^{b\times l}.
2:Abundance fraction 𝐀\mathbf{A} and variability coefficient 𝐁\mathbf{B}.
3:Initialize 𝐀m×n\mathbf{A}\in\mathbb{R}^{m\times n} and 𝐁l×n\mathbf{B}\in\mathbb{R}^{l\times n} randomly; t=0t=0;
4:Repeat:
5:Update 𝐀t+1𝐀t(𝐌T𝐑+α𝐌T𝐑α𝐌T𝐕𝐁(1+α)𝐌T𝐌𝐀t+β𝐃𝐀t)14\mathbf{A}_{t+1}\leftarrow\mathbf{A}_{t}\circ(\frac{{{\mathbf{M}^{T}}\mathbf{R}+\alpha\mathbf{M}^{T}\mathbf{R}-\alpha\mathbf{M}^{T}\mathbf{VB}}}{{(1+\alpha){\mathbf{M}^{T}}\mathbf{M}\mathbf{A}_{t}+\beta\mathbf{D}\mathbf{A}_{t}}})^{\frac{1}{4}};
6:Update 𝐁t+1𝐁t(α𝐕T𝐑α𝐕T𝐌𝐀t+1α𝐕T𝐕𝐁t+γBt)14\mathbf{B}_{t+1}\leftarrow\mathbf{B}_{t}\circ(\frac{{\alpha{\mathbf{V}^{T}}\mathbf{R}-\alpha{\mathbf{V}^{T}}\mathbf{MA}_{t+1}}}{{\alpha{\mathbf{V}^{T}}\mathbf{V}\mathbf{B}_{t}+\gamma{B}_{t}}})^{\frac{1}{4}};
7:Update 𝐃\mathbf{D};
8:t=t+1t=t+1;
9:Until Convergence;
10:Return 𝐀\mathbf{A} and 𝐁\mathbf{B};

IV Experiments

In this section, a series of experiments on both the simulated data set and two real-world data sets (Jasper and Cuprite data sets) is designed to demonstrate the effectiveness of SVASU algorithm for HSIs. Several well-known sparse unmixing algorithms are adopted for comparison, including SUnSAL [9], CLSUnSAL [16], ADSpLRU [26], JSpBLRU [27], MUASLIC [23]. For all these methods, the coefficient matrices corresponding to each spectral library are all randomly initialized and the values of parameters in each algorithm refer to the reported optimal values.

IV-A Evaluation Metrics

To quantitatively evaluate the performance of our proposed SVASU algorithm, the two types of pixel reconstruction error are employed as in the experiments. Firstly, the root mean square error (RMSE) is used to measure the distance between the true signal and its estimated value, like that of image pixel 𝐫\rm{\mathbf{r}} and its reconstructed version 𝐫^\rm{\hat{\mathbf{r}}} as RMSER\rm{RMSE_{R}}, and true abundance 𝐚\rm{\mathbf{a}} and its estimated version 𝐚^\rm{\hat{\mathbf{a}}} as RMSEA\rm{RMSE_{A}}, which are defined as,

RMSER=1ni=1n(𝐫i𝐫^i)2.\displaystyle\rm{RMSE_{R}}=\sqrt{\frac{1}{n}\sum\limits_{i=1}^{n}{{{({\mathbf{r}_{i}}-{{\hat{\mathbf{r}}}_{i}})}^{2}}}}. (20)
RMSEA=1ni=1n(𝐚i𝐚^i)2.\displaystyle\rm{RMSE_{A}}=\sqrt{\frac{1}{n}\sum\limits_{i=1}^{n}{{{({\mathbf{a}_{i}}-{{\hat{\mathbf{a}}}_{i}})}^{2}}}}. (21)

Furthermore, the signal-to-reconstruction-error (SRE) is also adopted to measure the quality of the reconstruction of signals by different algorithms, which can demonstrate more information regarding the power of the error in relation to the power of the signal. Generally, the higher the SRE (dB) value, the better the performance of the algorithm. Similarly, we definite SRER\rm{SRE_{R}} and SREA\rm{SRE_{A}} as follows,

SRER=10log10E[𝐫22]E[𝐫𝐫^22],\displaystyle{\rm{SRE_{R}}}=10\rm{log}_{10}\frac{E[\|\mathbf{r}\|_{2}^{2}]}{E[\|{\mathbf{r}}-{{\hat{\mathbf{r}}}}\|_{2}^{2}]}, (22)
SREA=10log10E[𝐚22]E[𝐚𝐚^22].\displaystyle{\rm{SRE_{A}}}=10\rm{log}_{10}\frac{E[\|\mathbf{a}\|_{2}^{2}]}{E[\|{\mathbf{a}}-{{\hat{\mathbf{a}}}}\|_{2}^{2}]}. (23)
Refer to caption
(a) Endmember Library
Refer to caption
(b) Spectral Variability Library
Figure 2: The Endmember Library and Spectral Variability Library obtained of the Synthetic data set through SVASU.
Refer to caption
Refer to caption
Figure 3: The performances of SVASU with respect to parameters (a) β\beta, γ\gamma and (b) α\alpha in terms of SRE and RMSE for both abundance and pixels on the synthetic data set.
TABLE I: Average SRE value and RMSE value obtained by comparison algorithms on the Synthetic data set. Note that the best results are in bolded.
SUnSAL CLSUnSAL JSpBLRU ADSpLRU MUASLIC SVASU
SRER 8.0104 8.0026 7.1616 8.0026 18.9487 34.3370
RMSER 1.2232 1.2241 1.3436 0.0608 0.5680 0.0573
SREA 0.2409 0.2220 1.0789 0.1460 0.2334 1.7884
RMSEA 0.1136 0.1133 0.0901 0.1129 0.1136 0.0732

IV-B Experiments Over Synthetic Data set

In this section, the simulated data are generated in the way similar to [56] and [57] based on the LMM in (1). The synthetic data set has n=10000n=10000 pixels. The endmember spectral signatures are using five spectra of minerals from the U.S. Geological Survey (USGS) containing as many as 420 bands covering from 400 to 2500 nm. Then, 30 dB Gaussian noise is added to each signature to simulate spectral variability. Considering that each pixel is unlikely to have a large number of endmembers in real scenarios (typically, less than 5), the maximum number of active endmembers per pixel is set to 4 to ensure abundance sparsity. Furthermore, the abundance vectors corresponding to active endmembers are generated using a mixture of uniform Dirichlet distributions [58]. Finally, 40 dB Gaussian noise is added to the whole image. In this experiment, the In-situ spectral library is simply composed of the selected five spectra from USGS and their corresponding spectra owning spectral variability. After the library segmentation by PCA, we acquire the (a) endmember library and (b) spectral variability library as shown in Fig. 2.

First of all, the impact of parameters, including α\alpha, β\beta, and γ\gamma, are discussed through several sets of simulated experiments. The curves of these parameters of our method are shown in Fig. 3(a) and (b). The evaluation metrics adopted here are the average SRE and RMSE calculated by (20)-(23). On one hand, in Fig. 3(a), α\alpha is preset to 10 and β\beta and γ\gamma vary in [1e0,1e3][1e0,1e3] and [1e2,1e5][1e2,1e5] exponentially. The proposed SVASU algorithm achieves a desired result when β=10\beta=10 and γ=1e4\gamma=1e4, with lower RMSE value and higher SRE value both for abundance and pixels. On the other hand, Fig. 3(b) displays the performances of SRE and RMSE under different values of parameter α\alpha tend to prominent when α\alpha reaches 9. Thus, the parameter α\alpha, β\beta, and γ\gamma are respectively set to be 9, 10, and 1e41e4 in the experiments on synthetic data set.

Table I shows the average SRE and RMSE value between the ground-truth pixels or abundance and corresponding matrixes acquired by different algorithms over synthetic data set. The performance of the proposed SVASU algorithm outperforms other methods on the estimation of either fractional abundance or pixel reconstruction.

IV-C Experiments Over Jasper Data set

The real-world Jasper data set is collected by an airborne visible/infrared imaging spectrometer (AVIRIS) sensor111Data available online at https://rslab.ut.ac.ir/data., which consists of 224 spectral bands ranging from 380 to 2500 nm and the spectral resolution is up to 9.46 nm. The size of this scene is 512×\times614. In the experiments, a subimage of 100×\times100 pixels is used. Due to dense water vapor and atmospheric effects, there exists some noisy bands, including the bands 1-3, 108-112, 154-166, and 220-224. After removing these bands, a total of 198 reflectance bands are finally adopted. There are four endmembers in the selected region, including road, soil, water, and tree, and their corresponding ground-truth abundance are shown in Fig. 4.

Refer to caption
Figure 4: The reference four endmembers and their corresponding abundance of the Jasper data set.
Refer to caption
(a) Endmember Library
Refer to caption
(b) Spectral Variability Library
Figure 5: The Endmember Library and Spectral Variability Library obtained of the Jasper data set through SVASU.
TABLE II: Average SRE value and RMSE value obtained by comparison algorithms on the Jasper data set. Note that the best results are in bolded.
SUnSAL CLSUnSAL JSpBLRU ADSpLRU MUASLIC SVASU
SRER 15.8183 16.9954 18.7103 17.2011 18.6439 19.5641
RMSER 0.0986 0.1083 0.1024 0.1055 0.0789 0.0753
SREA 3.9046 4.1053 2.9957 2.6846 6.7230 8.1455
RMSEA 0.1150 0.0968 0.1161 0.1224 0.0796 0.0627
Refer to caption
Figure 6: Visual analysis of the proposed SVASU model conducted on Jasper data set.

As shown in Fig. 1, the In-situ spectral library obtained by SPEE is shown in the first column, and then after the segmentation step described in Section III. B, we get the endmember library and spectral variability library shown in Fig. 5. To analyze the impact of three regularization parameters, including α\alpha, β\beta, and γ\gamma, we first fix α\alpha at 10 to discuss the performance of SVASU under different values of β\beta and γ\gamma. As shown in Fig. X(a), with the increase in β\beta value, the performances of abundance estimation and pixel reconstruction demonstrate different trends of development. Since the gap between the maximum and minimum values of evaluation matrixes for pixel reconstruction is small, the optimal parameter β\beta is set to 1e31e3 and γ\gamma is set to 1e61e6 corresponds to good capability of abundance estimation. After that, we further discuss the range of α\alpha as shown in Fig. X(b). Through observation, it can be found that when the value of α\alpha is lower than 3, the unmixing results of both abundance accuracy and data fidelity are superior to other cases. Thus, the optimal value of α\alpha is set to 1.

Table II shows the average SRE and RMSE values between the ground-truth pixels or abundance and corresponding matrixes acquired by different algorithms over synthetic data set. The performance of the proposed SVASU algorithm outperforms other methods on the estimation of fractional abundance. However, due to the second level decomposition by our extracted spectral variability, the performance of pixel reconstruction is not as well as other sparse unmixing methods as expected.

Furthermore, to carefully analyze the results of two-order reconstruction, Fig. 6 depicts the visual results of the important items. Each visual result is averaged over each band, including the image data 𝐑\mathbf{R}, the first order reconstruction item 𝐌𝐀\mathbf{MA}, the second order reconstruction item 𝐕𝐁\mathbf{VB}, the first order reconstruction error item 𝐑𝐌𝐀\mathbf{R-MA}, and the second order reconstruction error item 𝐑𝐌𝐀𝐕𝐁\mathbf{R-MA-VB}. It can be observed that on the Jasper data set, the amount of information occupied by 𝐌𝐀\mathbf{MA} is probably twice as much information as that represents by 𝐕𝐁\mathbf{VB}. Moreover, through the second order reconstruction by spectral variability library, the image reconstruction error reduced to a tenth on Jasper data set, which completely proves the validity and rationality of the proposed SVASU method.

IV-D Experiments Over Cuprite Data set

Refer to caption
Figure 7: Visual analysis of the proposed SVASU model conducted on Cuprite data set.

The hyperspectral data set named Cuprite data set, acquired by AVIRIS222http://aviris.jpl.nasa.gov/html/aviris.freedata.html., is also adopted in real experiments, which contains 224 bands ranging from 370 to 2510 nm with a ground instantaneous field of view of 20 m. The cropped image corresponds to a 350×350350\times 350 pixel subset of the sector labeled as f970619t01p02r02sc03.a.rflf970619t01p02_{r}02_{s}c03.a.rfl in the online data. After removing noisy bands and water-absorption bands (including bands 141-4, 105115105-115, 150170150-170, and 223224223-224), a total of 186 reflectance bands are finally adopted. The optimal parameters of proposed SVASU method are α=3\alpha=3, β=1e4\beta=1e4, and γ=1e6\gamma=1e6.

Given that there is no available ground-truth in terms of abundances and endmembers of Cuprite data set, only visual results of two-order reconstruction in SVASU are shown as Fig. 7. Similarly, each visual result is averaged over each band, including the two-order reconstruction item 𝐌𝐀\mathbf{MA}, 𝐕𝐁\mathbf{VB}, and the two-order reconstruction error item 𝐑𝐌𝐀\mathbf{R-MA}, 𝐑𝐌𝐀𝐕𝐁\mathbf{R-MA-VB}. It can be observed that the amount of information occupied by 𝐌𝐀\mathbf{MA} is more than twice as much information as that by 𝐕𝐁\mathbf{VB}. This also confirms that in the processing of PCAPCA, the main components are constructed by endmember library while the other components are constructed as the spectral variability library. Moreover, through the second order reconstruction by spectral variability library, the image reconstruction error reduced to less than a tenth on Cuprite data set. Therefore, the effectiveness of the proposed SVASU method are confirmed on the real-world data set.

V Conclusion

In this paper, a spectral variability augmented sparse unmixing model (SVASU) is proposed, which provides a new way to construct spectral variability explicitly from the spectral library for sparse unmixing. It is noted that a in-situ spectral library acquired under spatial purity is separated into the endmember library and the variability library by PCA. Afterwards, the proposed SVASU adopts a two-order decomposition structure to perform SU, in which the first order is to decompose the hyperspectral image data into the product of endmember library and abundance fractions roughly, the second order is to further represent the reconstruction error of pixels from first order as the influence of spectral variability library. Experimental results over both synthetic and real-world datasets demonstrate that the proposed SVASU model can certainly improve the performance of spectral linearly unmixing and through the introducing of second order reconstructed loss, the ability of the proposed framework to model the spectral variability is improved. Meanwhile, the use of those smoothness terms guarantees a moderate variation of the spectral variability.

Appendix A Proof of Convergence of SVASU

The problem in (14) contains two unknown matrices, and we design different rules to alternately update one by fixing the other. Following [59, 60], auxiliary function approach is used to prove the convergence of the proposed update rules. Function Q(𝐗,𝐗~)Q(\mathbf{X},\mathbf{\tilde{X}}) is an auxiliary function for L(𝐗)L(\mathbf{X}) if the conditions

Q(𝐗,𝐗~)L(𝐗),Q(𝐗,𝐗)=L(𝐗)\displaystyle Q(\mathbf{X},\mathbf{\tilde{X}})\geq L(\mathbf{X}),\ Q(\mathbf{X},\mathbf{X})=L(\mathbf{X}) (24)

are satisfied for any 𝐗,𝐗~\mathbf{X},\mathbf{\tilde{X}}. Then, if

𝐗(t+1)=argmin𝐗Q(𝐗,𝐗(t)),\displaystyle\mathbf{X}^{(t+1)}=\mathop{arg\min\limits_{\mathbf{X}}}Q(\mathbf{X},\mathbf{X}^{(t)}), (25)

it has been proven that the following inequalities held:

L(𝐗(t+1))Q(𝐗(t+1),𝐗(t))Q(𝐗(t),𝐗(t))=L(𝐗(t)).\displaystyle L(\mathbf{X}^{(t+1)})\leq Q(\mathbf{X}^{(t+1)},\mathbf{X}^{(t)})\leq Q(\mathbf{X}^{(t)},\mathbf{X}^{(t)})=L(\mathbf{X}^{(t)}). (26)

Thus, the function L(𝐗)L(\mathbf{X}) is monotonically decreasing.

A-1 Fixing 𝐁\mathbf{B}, minimizing 𝐀\mathbf{A}

First, when 𝐁\mathbf{B} is fixed, the minimization problem in (14) can be written as

min𝐀𝐑𝐌𝐀F2+α𝐓𝐌𝐀F2+βTr(𝐀T𝐃𝐀)\displaystyle\mathop{\min}\limits_{\mathbf{A}}{\left\|{\mathbf{R-MA}}\right\|_{F}^{2}}+\alpha{\left\|\mathbf{T-MA}\right\|_{F}^{2}}+\beta Tr(\mathbf{A}^{T}\mathbf{DA}) (27)
s.t.𝐀>0,\displaystyle{s.t.\mathbf{A}>0,}

in which 𝐓=𝐑𝐕𝐁\mathbf{T}=\mathbf{R}-\mathbf{VB}. For the first two terms in (27), the following inequality holds [61].

L1(𝐀)\displaystyle{L_{1}}(\mathbf{A}) =𝐑𝐌𝐀F2\displaystyle=\left\|{\mathbf{R}-\mathbf{M}\mathbf{A}}\right\|_{F}^{2} (28)
=Tr(𝐑T𝐑)2Tr(𝐑T𝐌𝐀)+Tr(𝐀T𝐌T𝐌𝐀)\displaystyle=Tr({\mathbf{R}^{T}}\mathbf{R})-2Tr({\mathbf{R}^{T}}\mathbf{M}\mathbf{A})+Tr({\mathbf{A}^{T}}{\mathbf{M}^{T}}\mathbf{M}\mathbf{A})
Tr(𝐑T𝐑)2Tr(𝐑T𝐌𝐀~)2Tr(𝐑T𝐌𝐙)\displaystyle\leq Tr({\mathbf{R}^{T}}\mathbf{R})-2Tr({\mathbf{R}^{T}}\mathbf{M}\mathbf{\tilde{A}})-2Tr({\mathbf{R}^{T}}\mathbf{MZ})
+12Tr(𝐏T𝐌T𝐌𝐀~)+12Tr(𝐀~T𝐌T𝐌𝐀~)\displaystyle+\frac{1}{2}Tr({\mathbf{P}^{T}}{\mathbf{M}^{T}}\mathbf{M}\mathbf{\tilde{A}})+\frac{1}{2}Tr({\mathbf{\tilde{A}}^{T}}{\mathbf{M}^{T}}\mathbf{M}\mathbf{\tilde{A}})
=Q1(𝐀,𝐀~)\displaystyle={Q_{1}}(\mathbf{A},\mathbf{\tilde{A}})
L2(𝐀)\displaystyle{L_{2}}(\mathbf{A}) =𝐓𝐌𝐀F2\displaystyle=\left\|{\mathbf{T}-\mathbf{M}\mathbf{A}}\right\|_{F}^{2} (29)
=Tr(𝐓T𝐓)2Tr(𝐓T𝐌𝐀)+Tr(𝐀T𝐌T𝐌𝐀)\displaystyle=Tr({\mathbf{T}^{T}}\mathbf{T})-2Tr({\mathbf{T}^{T}}\mathbf{M}\mathbf{A})+Tr({\mathbf{A}^{T}}{\mathbf{M}^{T}}\mathbf{M}\mathbf{A})
Tr(𝐓T𝐓)2Tr(𝐓T𝐌𝐀~)2Tr(𝐓T𝐌𝐙)\displaystyle\leq Tr({\mathbf{T}^{T}}\mathbf{T})-2Tr({\mathbf{T}^{T}}\mathbf{M}\mathbf{\tilde{A}})-2Tr({\mathbf{T}^{T}}\mathbf{MZ})
+12Tr(𝐏T𝐌T𝐌𝐀~)+12Tr(𝐀~T𝐌T𝐌𝐀~)\displaystyle+\frac{1}{2}Tr({\mathbf{P}^{T}}{\mathbf{M}^{T}}\mathbf{M}\mathbf{\tilde{A}})+\frac{1}{2}Tr({\mathbf{\tilde{A}}^{T}}{\mathbf{M}^{T}}\mathbf{M}\mathbf{\tilde{A}})
=Q2(𝐀,𝐀~)\displaystyle={Q_{2}}(\mathbf{A},\mathbf{\tilde{A}})

where 𝐏ik=(𝐀ik)4/(𝐀~ik)3\mathbf{P}_{ik}=(\mathbf{A}_{ik})^{4}/(\mathbf{\tilde{A}}_{ik})^{3}, and 𝐙ik=𝐀~ikln(𝐀ik/𝐀~ik)\mathbf{Z}_{ik}=\mathbf{\tilde{A}}_{ik}\ln(\mathbf{A}_{ik}/\mathbf{\tilde{A}}_{ik}). It is obvious that Q1(𝐀,𝐀)=L1(𝐀),Q2(𝐀,𝐀)=L2(𝐀){Q_{1}}(\mathbf{A},\mathbf{A})={L_{1}}(\mathbf{A}),{Q_{2}}(\mathbf{A},\mathbf{A})={L_{2}}(\mathbf{A}). Taking Q1(𝐀,𝐀~)Q_{1}(\mathbf{A},\mathbf{\tilde{A}}) as an example, Q1(𝐀,𝐀~)L1(𝐀)Q_{1}(\mathbf{A},\mathbf{\tilde{A}})\geq L_{1}(\mathbf{A}) can be proved in detail as follows,

Q1(𝐀,𝐀~)L1(𝐀)\displaystyle Q_{1}(\mathbf{A},\mathbf{\tilde{A}})-L_{1}(\mathbf{A}) (30)
=2Tr(𝐑T𝐌𝐀)2Tr(𝐑T𝐌𝐀~)2Tr(𝐑T𝐌𝐙)\displaystyle=2Tr({\mathbf{R}^{T}}\mathbf{M}\mathbf{A})-2Tr({\mathbf{R}^{T}}\mathbf{M}\mathbf{\tilde{A}})-2Tr({\mathbf{R}^{T}}\mathbf{MZ})
+12Tr(𝐏T𝐌T𝐌𝐀~)+12Tr(𝐀~T𝐌T𝐌𝐀~)\displaystyle+\frac{1}{2}Tr({\mathbf{P}^{T}}{\mathbf{M}^{T}}\mathbf{M}\mathbf{\tilde{A}})+\frac{1}{2}Tr({\mathbf{\tilde{A}}^{T}}{\mathbf{M}^{T}}\mathbf{M}\mathbf{\tilde{A}})
Tr(𝐀T𝐌T𝐌𝐀)\displaystyle-Tr({\mathbf{A}^{T}}{\mathbf{M}^{T}}\mathbf{M}\mathbf{A})
=ik2𝐑ikT𝐌ik(𝐀ik𝐀~ik𝐀~ikln𝐀ik𝐀~ik)\displaystyle=\sum\limits_{ik}2{\mathbf{R}^{T}_{ik}}{\mathbf{M}_{ik}}({\mathbf{A}_{ik}}-{\mathbf{\tilde{A}}}_{ik}-{\mathbf{\tilde{A}}}_{ik}\ln\frac{\mathbf{A}_{ik}}{\mathbf{\tilde{A}}_{ik}})
+ik12(𝐌ik)2((𝐀ik)2𝐀~ik𝐀~ik)20.\displaystyle+\sum\limits_{ik}\frac{1}{2}({\mathbf{M}_{ik}})^{2}(\frac{({\mathbf{A}_{ik}})^{2}}{{{\mathbf{\tilde{A}}}_{ik}}}-{\mathbf{\tilde{A}}_{ik}})^{2}\geq 0.

Thus, Q1(𝐀,𝐀~)Q_{1}(\mathbf{A},\mathbf{\tilde{A}}) and Q2(𝐀,𝐀~)Q_{2}(\mathbf{A},\mathbf{\tilde{A}}) can be selected as the auxiliary functions of L1(𝐀)L_{1}(\mathbf{A}) and L2(𝐀)L_{2}(\mathbf{A}) correspondingly.

Then, for the last term L3(𝐀)=Tr(𝐀T𝐃𝐀)L_{3}(\mathbf{A})=Tr({\mathbf{A}^{T}}\mathbf{D}\mathbf{A}), the following auxiliary function Q3(𝐀,𝐀~)Q_{3}(\mathbf{A},\mathbf{\tilde{A}}) is constructed as,

Q3(𝐀,𝐀~)=12Tr(𝐏T𝐃𝐀~)+12Tr(𝐀~T𝐃𝐀~),\displaystyle Q_{3}(\mathbf{A},\mathbf{\tilde{A}})=\frac{1}{2}Tr({\mathbf{P}^{T}}\mathbf{D}\mathbf{\tilde{A}})+\frac{1}{2}Tr({{\mathbf{\tilde{A}}}^{T}}\mathbf{D}\mathbf{\tilde{A}}), (31)

It is straightforward to verify that Q3(𝐀,𝐀)=L3(𝐀)Q_{3}(\mathbf{A},\mathbf{A})=L_{3}(\mathbf{A}). To show that Q3(𝐀,𝐀~)L3(𝐀)Q_{3}(\mathbf{A},\mathbf{\tilde{A}})\geq L_{3}(\mathbf{A}), we have the following calculation,

Q3(𝐀,𝐀~)L3(𝐀)\displaystyle Q_{3}(\mathbf{A},\mathbf{\tilde{A}})-L_{3}(\mathbf{A}) (32)
=ik𝐃ii𝐀~ik(𝐀ik)42(𝐀~ik)3+ik𝐃ii2(𝐀~ik)2ik𝐃ii(𝐀ik)2\displaystyle=\sum\limits_{ik}{\frac{{{\mathbf{D}_{ii}}{{\mathbf{\tilde{A}}}_{ik}}{{({\mathbf{A}_{ik}})}^{4}}}}{{{{2(\mathbf{\tilde{A}}}_{ik})^{3}}}}}+\sum\limits_{ik}{\frac{{{\mathbf{D}_{ii}}}}{2}({{\mathbf{\tilde{A}}}_{ik}})^{2}-\sum\limits_{ik}{{\mathbf{D}_{ii}}({\mathbf{A}_{ik}})}^{2}}
=ik𝐃ii2((𝐀ik)4(𝐀~ik)22(𝐀ik)2+(𝐀~ik)2)\displaystyle=\sum\limits_{ik}\frac{{{\mathbf{D}_{ii}}}}{2}(\frac{({\mathbf{A}_{ik}})^{4}}{({{\mathbf{\tilde{A}}}_{ik}})^{2}}-2({\mathbf{A}_{ik}})^{2}+({{\mathbf{\tilde{A}}}_{ik}})^{2})
=ik𝐃ii2((𝐀ik)2𝐀~ik𝐀~ik)20.\displaystyle=\sum\limits_{ik}\frac{{\mathbf{D}_{ii}}}{2}(\frac{({\mathbf{A}_{ik}})^{2}}{\mathbf{\tilde{A}}_{ik}}-{\mathbf{\tilde{A}}_{ik}})^{2}\geq 0.

Thus, Q3(𝐀,𝐀~)Q_{3}(\mathbf{A},\mathbf{\tilde{A}}) can be an auxiliary function of L3(𝐀)L_{3}(\mathbf{A}).

Finally, Q(𝐀,𝐀~)=Q1(𝐀,𝐀~)+αQ2(𝐀,𝐀~)+βQ3(𝐀,𝐀~)Q(\mathbf{A},\mathbf{\tilde{A}})=Q_{1}(\mathbf{A},\mathbf{\tilde{A}})+\alpha Q_{2}(\mathbf{A},\mathbf{\tilde{A}})+\beta Q_{3}(\mathbf{A},\mathbf{\tilde{A}}) is an auxiliary function of L(𝐀)=L1(𝐀)+αL2(𝐀)+βL3(𝐀)L(\mathbf{A})=L_{1}(\mathbf{A})+\alpha L_{2}(\mathbf{A})+\beta L_{3}(\mathbf{A}) in (14). Currently, according to Eq. (25), we find the minimum of minQ(𝐀,𝐀~)\min Q(\mathbf{A},\mathbf{\tilde{A}}) by fixing 𝐀~\mathbf{\tilde{A}}, which is given by

0\displaystyle 0 =Q(𝐀,𝐀~)𝐀ik\displaystyle=\frac{{\partial Q(\mathbf{A},\mathbf{\tilde{A}})}}{{\partial{\mathbf{A}_{ik}}}} (33)
=Q1(𝐀,𝐀~)𝐀ik+αQ2(𝐀,𝐀~)𝐀ik+βQ3(𝐀,𝐀~)𝐀ik\displaystyle=\frac{{\partial{Q_{1}}(\mathbf{A},\mathbf{\tilde{A}})}}{{\partial{\mathbf{A}_{ik}}}}+\alpha\frac{{\partial{Q_{2}}(\mathbf{A},\mathbf{\tilde{A}})}}{{\partial{\mathbf{A}_{ik}}}}+\beta\frac{{\partial{Q_{3}}(\mathbf{A},\mathbf{\tilde{A}})}}{{\partial{\mathbf{A}_{ik}}}}
=2(1+α)(𝐀ik)3(𝐀~ik)3(𝐌T𝐌𝐀~)ik2𝐀~ik𝐀ik(𝐌T𝐑)ik\displaystyle=2(1+\alpha)\frac{({\mathbf{A}_{ik}})^{3}}{({{\mathbf{\tilde{A}}}_{ik}})^{3}}({\mathbf{M}^{T}}\mathbf{M\tilde{A}})_{ik}-2\frac{{{{\mathbf{\tilde{A}}}_{ik}}}}{{{\mathbf{A}_{ik}}}}{({\mathbf{M}^{T}}\mathbf{R})_{ik}}
2α𝐀~ik𝐀ik(𝐌T𝐓)ik+2β(𝐀ik)3(𝐀~ik)3(𝐃𝐀~)ik.\displaystyle-2\alpha\frac{{{{\mathbf{\tilde{A}}}_{ik}}}}{{{\mathbf{A}_{ik}}}}({\mathbf{M}^{T}}\mathbf{T})_{ik}+2\beta\frac{{{{({\mathbf{A}_{ik}})}^{3}}}}{({{\mathbf{\tilde{A}}}_{ik}})^{3}}{(\mathbf{D\tilde{A}})_{ik}}.

Then, we can get

𝐀ik=𝐀~ik((𝐌T𝐑+α𝐌T𝐑α𝐌T𝐕𝐁)ik((1+α)𝐌T𝐌𝐀~+β𝐃𝐀~)ik)14.\displaystyle\mathbf{A}_{ik}=\mathbf{\tilde{A}}_{ik}(\frac{{({\mathbf{M}^{T}}\mathbf{R}+\alpha\mathbf{M}^{T}\mathbf{R}-\alpha\mathbf{M}^{T}\mathbf{VB})_{ik}}}{{((1+\alpha){\mathbf{M}^{T}}\mathbf{M}\mathbf{\tilde{A}}+\beta\mathbf{D}\mathbf{\tilde{A}})_{ik}}})^{\frac{1}{4}}. (34)

A-2 Fixing 𝐀\mathbf{A}, minimizing 𝐁\mathbf{B}

When 𝐀\mathbf{A} is fixed, the minimization problem in (14) can be written as

min𝐁α𝐓𝐕𝐁F2+γ𝐁F2s.t.𝐁>0,\displaystyle\mathop{\min}\limits_{\mathbf{B}}\alpha{\left\|{\mathbf{T^{\prime}-VB}}\right\|_{F}^{2}}+\gamma{\left\|{\mathbf{B}}\right\|_{F}^{2}}\ \ \ {s.t.\mathbf{B}>0,} (35)

in which 𝐓=𝐑𝐌𝐀\mathbf{T^{\prime}}=\mathbf{R}-\mathbf{MA}.

Similarly, the following two inequalities hold:

L1(𝐁)\displaystyle{L_{1}}(\mathbf{B}) =𝐓𝐕𝐁F2\displaystyle=\left\|{\mathbf{T^{\prime}}-\mathbf{V}\mathbf{B}}\right\|_{F}^{2} (36)
=Tr(𝐓T𝐓)2Tr(𝐓T𝐕𝐁)+Tr(𝐁T𝐕T𝐕𝐁)\displaystyle=Tr({\mathbf{T^{\prime}}^{T}}\mathbf{T^{\prime}})-2Tr({\mathbf{T^{\prime}}^{T}}\mathbf{V}\mathbf{B})+Tr({\mathbf{B}^{T}}{\mathbf{V}^{T}}\mathbf{V}\mathbf{B})
Tr(𝐓T𝐓)2Tr(𝐓T𝐕𝐁~)2Tr(𝐓T𝐕𝐙)\displaystyle\leq Tr({\mathbf{T^{\prime}}^{T}}\mathbf{T^{\prime}})-2Tr({\mathbf{T^{\prime}}^{T}}\mathbf{V}\mathbf{\tilde{B}})-2Tr({\mathbf{T^{\prime}}^{T}}\mathbf{VZ^{\prime}})
+12Tr(𝐏T𝐕T𝐕𝐁~)+12Tr(𝐁~T𝐕T𝐕𝐁~)\displaystyle+\frac{1}{2}Tr({\mathbf{P^{\prime}}^{T}}{\mathbf{V}^{T}}\mathbf{V}\mathbf{\tilde{B}})+\frac{1}{2}Tr({\mathbf{\tilde{B}}^{T}}{\mathbf{V}^{T}}\mathbf{V}\mathbf{\tilde{B}})
=Q1(𝐁,𝐁~)\displaystyle={Q_{1}}(\mathbf{B},\mathbf{\tilde{B}})
L2(𝐁)\displaystyle L_{2}(\mathbf{B}) =𝐁F2=Tr(𝐁T𝐁)\displaystyle=\left\|{\mathbf{B}}\right\|_{F}^{2}=Tr({\mathbf{B}^{T}}\mathbf{B}) (37)
12Tr(𝐏T𝐁~)+12Tr(𝐁~T𝐁~)=Q2(𝐁,𝐁~)\displaystyle\leq\frac{1}{2}Tr(\mathbf{P^{\prime}}^{T}\mathbf{\tilde{B}})+\frac{1}{2}Tr(\mathbf{\tilde{B}}^{T}\mathbf{\tilde{B}})={Q_{2}}(\mathbf{B},\mathbf{\tilde{B}})

in which 𝐏ij=(𝐁ij)4/(𝐁~ij)3\mathbf{P^{\prime}}_{ij}=(\mathbf{B}_{ij})^{4}/(\mathbf{\tilde{B}}_{ij})^{3}, and 𝐙ij=𝐁~ijln(𝐁ij/𝐁~ij)\mathbf{Z^{\prime}}_{ij}=\mathbf{\tilde{B}}_{ij}\ln(\mathbf{B}_{ij}/\mathbf{\tilde{B}}_{ij}).

Thus, Q(𝐁,𝐁~)=αQ1(𝐁,𝐁~)+γQ2(𝐁,𝐁~)Q(\mathbf{B},\mathbf{\tilde{B}})=\alpha Q_{1}(\mathbf{B},\mathbf{\tilde{B}})+\gamma Q_{2}(\mathbf{B},\mathbf{\tilde{B}}) is an auxiliary function of L(𝐁)=αL1(𝐁)+γL2(𝐁)L(\mathbf{B})=\alpha L_{1}(\mathbf{B})+\gamma L_{2}(\mathbf{B}) in Eq. (35). Currently, according to Eq. (25), we find the minimum of minQ(𝐁,𝐁~)\min Q(\mathbf{B},\mathbf{\tilde{B}}) by fixing 𝐁~\mathbf{\tilde{B}}, which is given by

0\displaystyle 0 =Q(𝐁,𝐁~)𝐁ij\displaystyle=\frac{{\partial Q(\mathbf{B},\mathbf{\tilde{B}})}}{{\partial{\mathbf{B}_{ij}}}} (38)
=αQ1(𝐁,𝐁~)𝐁ij+γQ2(𝐁,𝐁~)𝐁ij\displaystyle=\alpha\frac{{\partial{Q_{1}}(\mathbf{B},\mathbf{\tilde{B}})}}{{\partial{\mathbf{B}_{ij}}}}+\gamma\frac{{\partial{Q_{2}}(\mathbf{B},\mathbf{\tilde{B}})}}{{\partial{\mathbf{B}_{ij}}}}
=2α(𝐁ij)3(𝐁~ij)3(𝐕T𝐕𝐁~)ij2α𝐁~ij𝐁ij(𝐕T𝐓)ij\displaystyle=2\alpha\frac{({\mathbf{B}_{ij}})^{3}}{(\mathbf{\tilde{B}}_{ij})^{3}}({\mathbf{V}^{T}}\mathbf{V}\mathbf{\tilde{B}})_{ij}-2\alpha\frac{{\mathbf{\tilde{B}}_{ij}}}{\mathbf{B}_{ij}}({\mathbf{V}^{T}}\mathbf{T^{\prime}})_{ij}
+2γ(𝐁ij)3(𝐁~ij)2\displaystyle+2\gamma\frac{({\mathbf{B}_{ij}})^{3}}{(\mathbf{\tilde{B}}_{ij})^{2}}

Solving for 𝐁ij\mathbf{B}_{ij}, the minimum is

𝐁ij=𝐁~ij((α𝐕T𝐑α𝐕T𝐌𝐀)ij(α𝐕T𝐕𝐁~+γ𝐁~)ij)14.\displaystyle{\mathbf{B}_{ij}}=\mathbf{\tilde{B}}_{ij}(\frac{{(\alpha{\mathbf{V}^{T}}\mathbf{R}-\alpha{\mathbf{V}^{T}}\mathbf{MA})_{ij}}}{{(\alpha{\mathbf{V}^{T}}\mathbf{V}\mathbf{\tilde{B}}+\gamma\mathbf{\tilde{B}})_{ij}}})^{\frac{1}{4}}. (39)

Thus, updating 𝐀\mathbf{A} and 𝐁\mathbf{B} alternatively by using the rules in (LABEL:ruleA) and (39) can guarantee that the objective function in (14) is non-increasing.

References

  • [1] P. Ghamisi, N. Yokoya, J. Li, W. Liao, S. Liu, J. Plaza, B. Rasti, and A. Plaza, “Advances in hyperspectral image and signal processing: A comprehensive overview of the state of the art,” IEEE Geoscience and Remote Sensing Magazine, vol. 5, no. 4, pp. 37–78, 2017.
  • [2] J. M. Bioucas-Dias, A. Plaza, N. Dobigeon, M. Parente, and J. Chanussot, “Hyperspectral unmixing overview: Geometrical, statistical, and sparse regression-based approaches,” IEEE Journal of Selected Topics in Applied Earth Observations and Remote Sensing, vol. 5, no. 2, 2012.
  • [3] D. Heinz and Chein-I-Chang, “Fully constrained least squares linear spectral mixture analysis method for material quantification in hyperspectral imagery,” IEEE Transactions on Geoscience and Remote Sensing, vol. 39, no. 3, pp. 529–545, 2001.
  • [4] R. Borsoi, T. Imbiriba, J. C. Bermudez, C. Richard, J. Chanussot, L. Drumetz, J.-Y. Tourneret, A. Zare, and C. Jutten, “Spectral variability in hyperspectral data unmixing: A comprehensive review,” IEEE Geoscience and Remote Sensing Magazine, pp. 2–49, 2021.
  • [5] M. Winter, “N-FINDR: an algorithm for fast autonomous spectral end-member determination in hyperspectral data,” in Proceedings of SPIE, Image Spectrometry V, vol. 3753, 1999, pp. 266–277.
  • [6] J. Boardman, F. Kruse, and R. Green, “Mapping target signatures via partial unmixing of aviris data,” in Summaries of JPL Airborne Earth Science Workshop, 1995, pp. 23–26.
  • [7] J. Nascimento and J. Dias, “Vertex component analysis: a fast algorithm to unmix hyperspectral data,” IEEE Transactions on Geoscience and Remote Sensing, vol. 43, no. 4, pp. 898–910, Apr. 2005.
  • [8] J. Li, A. Agathos, D. Zaharie, J. M. Bioucas-Dias, and X. Li, “Minimum volume simplex analysis: A fast algorithm for linear hyperspectral unmixing,” IEEE Transactions on Geoscience and Remote Sensing, vol. 53, no. 9, pp. 5067–5082, 2015.
  • [9] M. D. Iordache, J. M. Bioucas-Dias, and A. Plaza, “Sparse unmixing of hyperspectral data,” IEEE Transactions on Geoscience and Remote Sensing, vol. 49, no. 6, pp. 2014–2039, 2011.
  • [10] S. Mei, K. Yan, M. Ma, X. Chen, S. Zhang, and Q. Du, “Remote sensing scene classification using sparse representation-based framework with deep feature fusion,” IEEE Journal of Selected Topics in Applied Earth Observations and Remote Sensing, vol. 14, pp. 5867–5878, 2021.
  • [11] X. Xu, B. Pan, Z. Chen, Z. Shi, and T. Li, “Simultaneously multiobjective sparse unmixing and library pruning for hyperspectral imagery,” IEEE Transactions on Geoscience and Remote Sensing, vol. 59, no. 4, pp. 3383–3395, 2021.
  • [12] M. Ma, S. Mei, S. Wan, Z. Wang, Z. Ge, V. Lam, and D. Feng, “Keyframe extraction from laparoscopic videos via diverse and weighted dictionary selection,” IEEE Journal of Biomedical and Health Informatics, vol. 25, no. 5, pp. 1686–1698, 2021.
  • [13] S. Mei, M. Ma, S. Wan, J. Hou, Z. Wang, and D. D. Feng, “Patch based video summarization with block sparse representation,” IEEE Transactions on Multimedia, vol. 23, pp. 732–747, 2021.
  • [14] L. Qi, J. Li, Y. Wang, Y. Huang, and X. Gao, “Spectral cspatial-weighted multiview collaborative sparse unmixing for hyperspectral images,” IEEE Transactions on Geoscience and Remote Sensing, vol. 58, no. 12, pp. 8766–8779, 2020.
  • [15] Z. Zhang, S. Liao, H. Fang, H. Zhang, and S. Wang, “Hyperspectral unmixing using spectral library sparse scaling and guided filter,” IEEE Geoscience and Remote Sensing Letters, pp. 1–5, 2020.
  • [16] M.-D. Iordache, J. M. Bioucas-Dias, and A. Plaza, “Collaborative sparse regression for hyperspectral unmixing,” IEEE Transactions on Geoscience and Remote Sensing, vol. 52, no. 1, pp. 341–354, 2014.
  • [17] J. M. Bioucas-Dias and M. A. T. Figueiredo, “Alternating direction algorithms for constrained sparse regression: Application to hyperspectral unmixing,” in 2010 2nd Workshop on Hyperspectral Image and Signal Processing: Evolution in Remote Sensing, 2010, pp. 1–4.
  • [18] T. Ince, “Double spatial graph laplacian regularization for sparse unmixing,” IEEE Geoscience and Remote Sensing Letters, pp. 1–5, 2021.
  • [19] M.-D. Iordache, J. M. Bioucas-Dias, and A. Plaza, “Total variation spatial regularization for sparse hyperspectral unmixing,” IEEE Transactions on Geoscience and Remote Sensing, vol. 50, no. 11, pp. 4484–4502, 2012.
  • [20] J. Qin, H. Lee, J. T. Chi, L. Drumetz, J. Chanussot, Y. Lou, and A. L. Bertozzi, “Blind hyperspectral unmixing based on graph total variation regularization,” IEEE Transactions on Geoscience and Remote Sensing, vol. 59, no. 4, pp. 3338–3351, 2021.
  • [21] X.-R. Feng, H.-C. Li, S. Liu, and H. Zhang, “Correntropy-based autoencoder-like nmf with total variation for hyperspectral unmixing,” IEEE Geoscience and Remote Sensing Letters, pp. 1–5, 2020.
  • [22] S. Zhang, J. Li, H.-C. Li, C. Deng, and A. Plaza, “Spectral cspatial weighted sparse regression for hyperspectral image unmixing,” IEEE Transactions on Geoscience and Remote Sensing, vol. 56, no. 6, pp. 3265–3276, 2018.
  • [23] R. A. Borsoi, T. Imbiriba, J. C. M. Bermudez, and C. Richard, “A fast multiscale spatial regularization for sparse hyperspectral unmixing,” IEEE Geoscience and Remote Sensing Letters, vol. 16, no. 4, pp. 598–602, 2019.
  • [24] J. Li, Y. Li, R. Song, S. Mei, and Q. Du, “Local spectral similarity preserving regularized robust sparse hyperspectral unmixing,” IEEE Transactions on Geoscience and Remote Sensing, vol. 57, no. 10, pp. 7756–7769, 2019.
  • [25] L. Sun, F. Wu, C. He, T. Zhan, W. Liu, and D. Zhang, “Weighted collaborative sparse and l1/2 low-rank regularizations with superpixel segmentation for hyperspectral unmixing,” IEEE Geoscience and Remote Sensing Letters, pp. 1–5, 2020.
  • [26] P. V. Giampouras, K. E. Themelis, A. A. Rontogiannis, and K. D. Koutroumbas, “Simultaneously sparse and low-rank abundance matrix estimation for hyperspectral image unmixing,” IEEE Transactions on Geoscience and Remote Sensing, vol. 54, no. 8, pp. 4775–4789, 2016.
  • [27] J. Huang, T.-Z. Huang, L.-J. Deng, and X.-L. Zhao, “Joint-sparse-blocks and low-rank representation for hyperspectral unmixing,” IEEE Transactions on Geoscience and Remote Sensing, vol. 57, no. 4, pp. 2419–2438, 2019.
  • [28] H. Li, R. Feng, L. Wang, Y. Zhong, and L. Zhang, “Superpixel-based reweighted low-rank and total variation sparse unmixing for hyperspectral remote sensing imagery,” IEEE Transactions on Geoscience and Remote Sensing, vol. 59, no. 1, pp. 629–647, 2021.
  • [29] L. Drumetz, J. Chanussot, and C. Jutten, “Spectral unmixing: A derivation of the extended linear mixing model from the hapke model,” IEEE Geoscience and Remote Sensing Letters, vol. 17, no. 11, pp. 1866–1870, 2020.
  • [30] L. Drumetz, J. Chanussot, C. Jutten, W.-K. Ma, and A. Iwasaki, “Spectral variability aware blind hyperspectral image unmixing based on convex geometry,” IEEE Transactions on Image Processing, vol. 29, pp. 4568–4582, 2020.
  • [31] J. Theiler, A. Ziemann, S. Matteoli, and M. Diani, “Spectral variability of remotely sensed target materials: Causes, models, and strategies for mitigation and robust exploitation,” IEEE Geoscience and Remote Sensing Magazine, vol. 7, no. 2, pp. 8–30, 2019.
  • [32] R. A. Borsoi, T. Imbiriba, and J. C. M. Bermudez, “A data dependent multiscale model for hyperspectral unmixing with spectral variability,” IEEE Transactions on Image Processing, vol. 29, pp. 3638–3651, 2020.
  • [33] T. Uezato, M. Fauvel, and N. Dobigeon, “Hyperspectral unmixing with spectral variability using adaptive bundles and double sparsity,” IEEE Transactions on Geoscience and Remote Sensing, vol. 57, no. 6, pp. 3980–3992, 2019.
  • [34] L. Drumetz, T. R. Meyer, J. Chanussot, A. L. Bertozzi, and C. Jutten, “Hyperspectral image unmixing with endmember bundles and group sparsity inducing mixed norms,” IEEE Transactions on Image Processing, vol. 28, no. 7, pp. 3435–3450, 2019.
  • [35] A. Halimi, N. Dobigeon, and J.-Y. Tourneret, “Unsupervised unmixing of hyperspectral images accounting for endmember variability,” IEEE Transactions on Image Processing, vol. 24, no. 12, pp. 4904–4917, 2015.
  • [36] H. Liu, Y. Lu, Z. Wu, Q. Du, J. Chanussot, and Z. Wei, “Bayesian unmixing of hyperspectral image sequence with composite priors for abundance and endmember variability,” IEEE Transactions on Geoscience and Remote Sensing, pp. 1–15, 2021.
  • [37] Y. Zhou, E. B. Wetherley, and P. D. Gader, “Unmixing urban hyperspectral imagery using probability distributions to represent endmember variability,” Remote Sensing of Environment, vol. 246, p. 111857, 2020.
  • [38] P.-A. Thouvenin, N. Dobigeon, and J.-Y. Tourneret, “Hyperspectral unmixing with spectral variability using a perturbed linear mixing model,” IEEE Transactions on Signal Processing, vol. 64, no. 2, pp. 525–538, 2015.
  • [39] L. Drumetz, M.-A. Veganzones, S. Henrot, R. Phlypo, J. Chanussot, and C. Jutten, “Blind hyperspectral unmixing using an extended linear mixing model to address spectral variability,” IEEE Transactions on Image Processing, vol. 25, no. 8, pp. 3890–3905, 2016.
  • [40] D. Hong, N. Yokoya, J. Chanussot, and X. X. Zhu, “An augmented linear mixing model to address spectral variability for hyperspectral unmixing,” IEEE Transactions on Image Processing, vol. 28, no. 4, pp. 1923–1938, 2019.
  • [41] C. Zhang, L. Ma, J. Chen, Y. Rao, Y. Zhou, and X. Chen, “Assessing the impact of endmember variability on linear spectral mixture analysis (lsma): A theoretical and simulation analysis,” Remote Sensing of Environment, vol. 235, pp. 1–21, 11 2019.
  • [42] L. Drumetz, J. Chanussot, and C. Jutten, “Variability of the endmembers in spectral unmixing: Recent advances,” in 2016 8th Workshop on Hyperspectral Image and Signal Processing: Evolution in Remote Sensing (WHISPERS), 2016, pp. 1–5.
  • [43] D. A. Roberts, M. Gardner, R. Church, S. Ustin, G. Scheer, and R. Green, “Mapping chaparral in the santa monica mountains using multiple endmember spectral mixture models,” Remote sensing of environment, vol. 65, no. 3, pp. 267–279, 1998.
  • [44] M. Xu, L. Zhang, B. Du, and L. Zhang, “An image-based endmember bundle extraction algorithm using reconstruction error for hyperspectral imagery,” Neurocomputing, vol. 173, pp. 397–405, 2016.
  • [45] F. Chen, K. Wang, and T. F. Tang, “Spectral unmixing using a sparse multiple-endmember spectral mixture model,” IEEE Transactions on Geoscience and Remote Sensing, vol. 54, no. 10, pp. 5846–5861, 2016.
  • [46] C. Zhang, “Multiscale quantification of urban composition from eo-1/hyperion data using object-based spectral unmixing,” International journal of applied earth observation and geoinformation, vol. 47, pp. 153–162, 2016.
  • [47] R. A. Borsoi, T. Imbiriba, J. C. M. Bermudez, and C. Richard, “Deep generative models for library augmentation in multiple endmember spectral mixture analysis,” IEEE Geoscience and Remote Sensing Letters, pp. 1–5, 2020.
  • [48] M. Berman, L. Bischof, R. Lagerstrom, Y. Guo, J. Huntington, P. Mason, and A. A. Green, “A comparison between three sparse unmixing algorithms using a large library of shortwave infrared mineral spectra,” IEEE Transactions on Geoscience and Remote Sensing, vol. 55, no. 6, pp. 3588–3610, 2017.
  • [49] K. D. Singh, D. Ramakrishnan, and L. Mansinha, “Relevance of transformation techniques in rapid endmember identification and spectral unmixing: A hypespectral remote sensing perspective,” in 2012 IEEE International Geoscience and Remote Sensing Symposium, 2012, pp. 4066–4069.
  • [50] Y. Shao and J. Lan, “A spectral unmixing method by maximum margin criterion and derivative weights to address spectral variability in hyperspectral imagery,” Remote Sensing, vol. 11, no. 9, p. 1045, 2019.
  • [51] F. A. Mianji and Y. Zhang, “Svm-based unmixing-to-classification conversion for hyperspectral abundance quantification,” IEEE Transactions on Geoscience and Remote Sensing, vol. 49, no. 11, pp. 4318–4327, 2011.
  • [52] T. Uezato, R. J. Murphy, A. Melkumyan, and A. Chlingaryan, “A novel spectral unmixing method incorporating spectral variability within endmember classes,” IEEE Transactions on Geoscience and Remote Sensing, vol. 54, no. 5, pp. 2812–2831, 2016.
  • [53] S. Mei, M. He, Z. Wang, and D. Feng, “Mixture Analysis by Multichanel Hopfield Neural Network,” IEEE Geoscience and Remote Sensing Letters, vol. 7, pp. 455–459, 2010.
  • [54] S. Mei, M. He, Z. Wang, and D. Feng, “Spatial purity based endmember extraction for spectral mixture analysis,” IEEE Transactions on Geoscience and Remote Sensing, vol. 48, no. 9, pp. 3434–3445, 2010.
  • [55] F. Nie, H. Huang, X. Cai, and C. Ding, “Efficient and robust feature selection via joint l2,1l_{2,1}-norms minimization,” Advances in Neural Information Processing Systems 23: 24th Annual Conference on Neural Information Processing Systems (NIPS), vol. 23, pp. 1813–1821, 2010.
  • [56] L. Zhuang, C.-H. Lin, M. A. T. Figueiredo, and J. M. Bioucas-Dias, “Regularization parameter selection in minimum volume hyperspectral unmixing,” IEEE Transactions on Geoscience and Remote Sensing, vol. 57, no. 12, pp. 9858–9877, 2019.
  • [57] J. M. Bioucas-Dias, A. Plaza, N. Dobigeon, M. Parente, Q. Du, P. Gader, and J. Chanussot, “Hyperspectral unmixing overview: Geometrical, statistical, and sparse regression-based approaches,” IEEE Journal of Selected Topics in Applied Earth Observations and Remote Sensing, vol. 5, no. 2, pp. 354–379, 2012.
  • [58] J. M. P. Nascimento and J. M. Bioucas-Dias, “Hyperspectral unmixing based on mixtures of dirichlet components,” IEEE Transactions on Geoscience and Remote Sensing, vol. 50, no. 3, pp. 863–878, 2012.
  • [59] D. Lee and H. Seung, “Algorithms for non-negative matrix factorization,” International Conference on Neural Information Processing Systems, pp. 535–541, 2001.
  • [60] R. Huang, X. Li, and L. Zhao, “Spectral cspatial robust nonnegative matrix factorization for hyperspectral unmixing,” IEEE Transactions on Geoscience and Remote Sensing, vol. 57, no. 10, pp. 8235–8254, 2019.
  • [61] F. Wang, T. Li, W. Xin, S. Zhu, and C. Ding, “Community discovery using nonnegative matrix factorization,” Data Mining and Knowledge Discovery, vol. 22, no. 3, pp. 493–521, 2011.