Buckets:
Robust Hyperspectral Unmixing with Correntropy based Metric
Ying Wang, Chunhong Pan, Shiming Xiang, and Feiyun Zhu
Abstract
Hyperspectral unmixing is one of the crucial steps for many hyperspectral applications. The problem of hyperspectral unmixing has proven to be a difficult task in unsupervised work settings where the endmembers and abundances are both unknown. What is more, this task becomes more challenging in the case that the spectral bands are degraded with noise. This paper presents a robust model for unsupervised hyperspectral unmixing. Specifically, our model is developed with the correntropy based metric where the non-negative constraints on both endmembers and abundances are imposed to keep physical significance. In addition, a sparsity prior is explicitly formulated to constrain the distribution of the abundances of each endmember. To solve our model, a half-quadratic optimization technique is developed to convert the original complex optimization problem into an iteratively re-weighted NMF with sparsity constraints. As a result, the optimization of our model can adaptively assign small weights to noisy bands and give more emphasis on noise-free bands. In addition, with sparsity constraints, our model can naturally generate sparse abundances. Experiments on synthetic and real data demonstrate the effectiveness of our model in comparison to the related state-of-the-art unmixing models.
Index Terms
Hyperspectral unmixing, linear mixture model, non-negative matrix factorization, robust estimation, correntropy based metric
Ying Wang, Chunhong Pan, Shiming Xiang and Feiyun Zhu are with the National Laboratory of Pattern Recognition, Institute of Automation, Chinese Academy of Sciences (e-mail: {ywang, chpan, smxiang, fyzhu}@nlpr.ia.ac.cn).## I. INTRODUCTION
Hyperspectral remote sensing has wide applications since it provides digital images in many contiguous and very narrow spectral bands. However, due to the low spatial resolution of hyperspectral sensor, microscopic material mixing, and multiple scattering, spectra measured by hyperspectral sensor are likely to be mixed by several materials [1], [2], [3]. Accordingly, mixture spectra present a challenge to hyperspectral image analysis because their spectral signatures do not correspond to any single well-defined material [1]. Thus, hyperspectral unmixing is a significant step for hyperspectral image analysis. The task of hyperspectral unmixing is to separate the mixed pixel spectra from a hyperspectral image into a collection of constituent spectra, called endmembers and a set of corresponding fractions, called abundances [1].
In general, there are two kinds of spectral unmixing model: linear and nonlinear. Linear mixing model (LMM) is the most popular model for hyperspectral unmixing. It assumes that the spectrum of a given pixel is a linear combination of the endmembers [1], [3]. Due to its practical advantages, there are a lot of unmixing models proposed based on LMM [4], [5], [6], [7], [8], [9], [10], [11]. Conversely, nonlinear unmixing model [3], [12], [13] describes mixed spectra by assuming that the observed pixel is generated from a nonlinear function of the abundance associated with the endmembers.
In literature, conventional algorithms for hyperspectral unmixing involve two steps: endmember extraction and mixed-pixel decomposition. For endmember extraction, a majority of algorithms have been designed under the pure pixel assumption [14], [15], [16], [17], [18]. N-FINDR [14] and Vertex component analysis (VCA) [17] are two typical this kind of algorithms. N-FINDR [14] aims at looking for the pure pixels in the hyperspectral data that determine the maximum volume. VCA [17] considers that the endmembers are the vertices of a simplex and extracts the endmembers iteratively by projecting the data onto the direction which is orthogonal to the subspace spanned by the already extracted endmembers. All these algorithms assume the presence of pure pixels for each endmember in the original hyperspectral data. Once the endmembers have been identified, the corresponding abundances can be estimated by solving a constrained least square problem. The constraints include the non-negativity constraint, the sum-to-one constraint,and so on [4], [19].
Since the assumption of pure pixel presence is usually unreliable, a number of blind source separation algorithms have been proposed to obtain the endmembers and the abundances simultaneously [20], [5], [21], [9], [11]. Among these algorithms, non-negative matrix factorization (NMF) shows significant potential for hyperspectral unmixing [8], [5], [11], [6]. NMF decomposes a non-negative hyperspectral data matrix into the product of two non-negative matrices, which are the endmembers and the abundances respectively. NMF has two main advantages for hyperspectral unmixing. First, the non-negative constraint on both endmembers and abundances is physically consistent with the mixing pixels in real world. Second, NMF usually provides a parts-based representation of the data, making the decomposition matrices more intuitive and interpretable [22], [23]. There are several extension models based on NMF proposed in hyperspectral unmixing literature [8], [5], [9], [11], [7]. Specifically, in [5], the authors introduce the piecewise smoothness of spectral data and sparseness of abundance fraction to NMF. In [7], the authors extend the NMF method by incorporating the $\ell_{1/2}$ sparsity constraint. These algorithms, however, take no account of the noise in each band in the hyperspectral data. Actually, in hyperspectral images, due to the distinctive properties of the hyperspectral sensor in each wavelengths, the noise intensity in different bands is different. Thus, these algorithms may fail to obtain the accurate endmembers and abundances.
Practically, due to the variations in atmospheric conditions and the properties of the hyperspectral sensor in each wavelengths, the noisy bands commonly exist in hyperspectral data [24]. Furthermore, the noise level differs from band to band (as can be seen from Fig. 4). Naturally, the intuitive motivation of this paper is that the bands which have high-noise-intensity should have small contributions for hyperspectral unmixing. This observation motivates us to propose a robust unmixing algorithm based on correntropy based metric [25], [26] to take into account the noise of each band. It can adaptively and automatically assign small weights to noisy bands. Thus, the proposed algorithm is much more insensitive to noisy bands.
The main contributions of this paper are highlighted as follows:
- A robust hyperspectral unmixing model is proposed based on the correntropy along withimposing non-negative constraints and the use of a sparse penalty. Beyond directly using the traditional square of the Euclidean distance, the objective function of the proposed model is developed based on correntropy based metric, which can adaptively assign small weights to noisy bands and give more emphasis on noise-free bands. As a result, our model can effectively cope with the noisy bands in hyperspectral data. Experiments on synthetic and real data illustrate the effectiveness of our model.
- An optimization algorithm of our model is proposed based on half-quadratic optimization technique. It converts the original nonlinear and nonconvex optimization problem into an iteratively re-weighted NMF with sparsity constraints, which can be easily solved by multiplicative update algorithm. Besides, the convergence of this optimization algorithm is proven theoretically.
The remainder of this paper is structured as follows. We first give a brief introduction to the LMM and its extensions in Section II. In Section III, we propose our robust unmixing model based on correntropy based metric and develop a convergent algorithm for its optimization. Section IV reports the experimental results and performs algorithmic comparisons with other related state-of-the-art techniques. Conclusions are drawn in Section V.
II. RELATED WORKS
A. Linear Mixture Model
Due to its physical effectiveness and mathematical simplicity [2], [3], Linear Mixture Model (LMM) is commonly used for hyperspectral unmixing. LMM assumes that the observed pixel is approximated by a linear combination of a small number of endmembers. The mathematical formulation of the LMM is given by
where $\mathbf{y} \in \mathbb{R}^D$ denotes an observed pixel, $\mathbf{x}_k \in \mathbb{R}^D$ is the $k$ -th endmember, $w_k$ is the abundance fraction for the $k$ -th endmember and $\mathbf{e} \in \mathbb{R}^D$ is an additive noise.Considering a hyperspectral image of $N$ pixels, its matrix notation is $\mathbf{Y} = [\mathbf{y}_1, \dots, \mathbf{y}_N] \in \mathbb{R}^{D \times N}$ , then LMM can be rewritten as
where $\mathbf{X} = [\mathbf{x}_1, \dots, \mathbf{x}_K] \in \mathbb{R}^{D \times K}$ is the endmember matrix, $\mathbf{W} = [\mathbf{w}_1, \dots, \mathbf{w}_N] \in \mathbb{R}^{K \times N}$ is the abundance matrix, and $\mathbf{E} \in \mathbb{R}^{D \times N}$ represents the additive noise. The number of endmembers $K$ is often much smaller than the number of band $N$ . As mentioned in [1], for spectral unmixing problem, LMM should satisfy two physical constraints, referred to as the non-negative constraint and the abundance sum-to-one constraint, which can be formalized as follows:
B. Non-negative Matrix Factorization
Recently, NMF has received much attention [8], [5], [11], [6] for hyperspectral unmixing because of its non-negative nature. Essentially, NMF is a LMM model with non-negative constraints. It aims at decomposing the given non-negative matrix $\mathbf{Y}$ into two non-negative matrices $\mathbf{X}$ and $\mathbf{W}$ . The objective function of NMF is defined as follows:
where $|\cdot|_F$ denotes the frobenius norm.
There are numerous optimization algorithms to solve the above problem (3). The most famous one is the multiplicative update algorithm which was proposed by Lee and Seung [22]. It has been proved that the multiplicative update algorithm can find local minima of the objective function (3). There are also several other optimization algorithms, such as the alternation non-negative least least squares [27], the projected gradient descent [28] and the active-set method [29], proposed to solve the NMF problem.
The non-negative constraints on $\mathbf{X}$ and $\mathbf{W}$ of NMF only allow additive combinations among different bases to represent the sample, which makes NMF learn a parts-based representation ofdata [22], [23]. The advantage of this parts-based representation has been successfully applied in many real-world problems such as document clustering [30], [31], hyperspectral unmixing [32], [5], [7], [11] and face recognition [33], [34].
However, due to the nonconvexity of the objective function of NMF, there are a lot of local minima of the objective function (3). In order to reduce the feasible solution set, several constraints are introduced to NMF framework [35], [32]. In summary, these extensions can be expressed as the following objective function
where $J_{\mathbf{X}}(\mathbf{X})$ and $J_{\mathbf{W}}(\mathbf{W})$ denote the regularization terms for $\mathbf{X}$ and $\mathbf{W}$ , respectively, and $\lambda \geq 0, \beta \geq 0$ are the regularization parameters.
Due to the physical consistency with unmixing problem, NMF and its extension models have been widely applied to hyperspectral unmixing. However, most of these methods do not take into account the difference of noise intensity in each band. These algorithms consider that all of the bands in the hypersepctral data have the same signal-to-noise ratio. Actually, the noise intensity in different bands is different in hyperspectral data. Accordingly, along with the line of correntropy based metric, in this paper we propose a robust unmixing method to suppress the contributions of the bands with high-noise-intensity.
III. ROBUST UNMIXING BASED ON CORRENTROPY BASED METRIC
In this section, we first describe the correntropy. Then, we propose our robust unmixing model based on correntropy based metric. Finally, we present the optimization algorithm of our model and summarize its implementation details.
A. Correntropy
In hyperspectral images, the noise levels of different bands are usually different (as can be seen from Fig. 4). Meanwhile, noise is usually difficult for computer to predict in advance. Thesenoisy bands will severely affect the unmixing results. Recently, in information theoretical learning [36], correntropy was proposed to obtain robust analysis [37]. It has shown many advantages in robust learning and pattern recognition [26], [38]. Correntropy is a nonlinear similarity measure between two random variables $\mathbf{x}$ and $\mathbf{y}$ , which is defined by
where $\kappa_\sigma(\cdot)$ denotes a kernel function that satisfies Mercer's Theorem [39] and $\mathbb{E}$ is the expectation operator.
Actually, the joint probability density function of $\mathbf{x}$ and $\mathbf{y}$ is practically unknown. Only a finite number of data ${(\mathbf{x}_n, \mathbf{y}n)}{n=1}^N$ are available. Thus, the sample estimator of correntropy is estimate by
In this work, we adopt the Gaussian kernel function $\kappa_\sigma(\mathbf{x} - \mathbf{y}) = \exp\left(-\frac{|\mathbf{x}-\mathbf{y}|_2^2}{\sigma^2}\right)$ .
Actually, correntropy has a close relationship with Welsch M-estimator [40]. In [25], the authors have proved that correntropy is a robust function for linear and nonlinear regression.
B. Proposed model
For convenience, a hyperspectral image with $N$ pixels and $D$ bands is denoted as $\mathbf{Y} = [\mathbf{y}_1, \dots, \mathbf{y}_N] = [\mathbf{y}^1, \dots, \mathbf{y}^D]^T \in \mathbb{R}^{D \times N}$ , where the pixel $\mathbf{y}_n \in \mathbb{R}^D$ is the $n$ -th column of $\mathbf{Y}$ and the band $\mathbf{y}^d \in \mathbb{R}^N$ is represented by the $d$ -th row of $\mathbf{Y}$ . Suppose the observed hyperspectral image $\mathbf{Y}$ is linearly mixed by the endmembers $\mathbf{X} = [\mathbf{x}_1, \dots, \mathbf{x}_K] \in \mathbb{R}^{D \times K}$ and the abundance $\mathbf{W} = [\mathbf{w}_1, \dots, \mathbf{w}_N] \in \mathbb{R}^{K \times N}$ . The commonly used cost function for LMM is the square of the Euclidean distance between $\mathbf{Y}$ and $\mathbf{XW}$ , which is defined as
However, the solution based on square of the Euclidean distance has been recognized to be sensitive to noise and outliers [41]. The existence of noise in a hyperspectral image will limitthe precision of spectral unmixing. Since hyperspectral images have hundreds of bands and the noise level in different bands is different, we define the following correntropy based model.
where $(\mathbf{XW})^d$ denotes the $d$ -th row of matrix $\mathbf{XW}$ . Different from the NMF based methods where all bands will contribute equally to the objective function, our model treats individual band differently and gives more emphasis on those bands that have more exactly linear reconstruction. If the $d$ -th band $\mathbf{y}^d$ is degraded by serious noise, then this band will yield a large reconstruction error. According to the properties of Welsch M-estimator [40], the $d$ -th band will provide small contribution to the objective function in Eq. (8). This implies that if there are noise-corrupted bands, those bands will have small contributions to the objective function.
It is worth to point out that the above model defined in (8) is different from the traditional robust estimation methods [41], which consider that there are outliers in the samples (pixels). In practical situations, there are usually noisy bands in the hyperspectral images. Consequently, our model considers that there are “outlier” bands, which is more physically consistent with the real situation in hyperspectral data.
In addition, we take into account the sparseness constraint of the abundance. In most cases, the abundance distribution of any endmember does not apply to the whole scene. This implies that the observed pixel is usually mixed by only a few endmembers. Therefore, the abundance fractions of each pixel have a sparsity prior. In order to enforce sparsity, an $\ell_0$ -norm penalty term $\sum_{n=1}^N |\mathbf{w}_n|_0$ can be incorporated into our model. However, the $\ell_0$ minimization problem is NP hard, which means that it is very hard to solve.
Recent development in the theory of sparse representation and compressed sensing [42], [43], [44] shows that if the solution is sparse enough, then the solution of $\ell_0$ minimization problem is equal to the solution of the $\ell_1$ problem. Besides, since its convexity property, $\ell_1$ minimization problem is much easier to be solved. Thus, we introduce the following $\ell_1$ -norm penalty terminto our model.
Combining the correntropy based loss term (8) and the $\ell_1$ sparse penalty term (9), the cost function of our robust unmixing model is defined as follows:
where $\lambda \geq 0$ is the regularization parameter. It is worth noting that the minimization problem (10) is subjected to $\mathbf{X}{dk} \geq 0$ and $\mathbf{W}{kn} \geq 0$ , which are the non-negative constraints. Here, $\mathbf{X}{dk}$ denotes the spectrum of the $k$ -th endmember in $d$ -th band and $\mathbf{W}{kn}$ represents the abundance fraction of the $n$ -th pixel corresponding to $k$ -th endmember. It is obvious that our model is essentially a correntropy based NMF (CENMF) along with an $\ell_1$ sparse penalty term. We call this model $\ell_1$ -CENMF.
C. Optimization
Obviously the objective function in (10) is nonconvex and nonlinear with respect to $\mathbf{X}$ and $\mathbf{W}$ together, and it is difficult to optimize directly. Here, we propose an alternative iterative algorithm for its optimization.
Our algorithm utilizes the half-quadratic technique [45], [46] to solve the optimization problem (10). According to the property of the convex conjugated function, we have [47]:
Proposition 1. There exists a convex conjugated function $\psi(\cdot)$ of $g(x) = \exp(-x)$ , such that
and for a fixed $x$ , the maximum is reached at $u = -g(x)$ .
According to the proposition 1, substituting $x$ with $\frac{|\mathbf{x}|_2^2}{\sigma^2}$ , we get the following propositionProposition 2. There exists a convex conjugated function $\psi(\cdot)$ of $\kappa_\sigma(\mathbf{x}) = \exp(-\frac{|\mathbf{x}|_2^2}{\sigma^2})$ such that
and for a fixed $\mathbf{x}$ , the minimum is reached at $u = \kappa_\sigma(\mathbf{x})$ 1.
Substituting (12) into (10), we have the following augmented objective function in an enlarged parameter space
where $\mathbf{u} = [u_1, u_2, \dots, u_D]^T$ is the auxiliary vector introduced by half-quadratic optimization.
According to the proposition 2, for a fixed $\mathbf{X}, \mathbf{W}$ , we obtain the following equation,
Note that the minimization is reached at $u_d = \exp\left(-\frac{|\mathbf{y}^d - (\mathbf{XW})^d|_2^2}{\sigma^2}\right)$ .
Then, we can conclude that minimizing $\mathcal{G}(\mathbf{X}, \mathbf{W})$ is identical to minimizing the augmented function $\widehat{\mathcal{G}}(\mathbf{X}, \mathbf{W}, \mathbf{u})$ . The augmented function $\widehat{\mathcal{G}}(\mathbf{X}, \mathbf{W}, \mathbf{u})$ can be optimized by the following alternate minimization:
- Fixed $\mathbf{W}$ and $\mathbf{X}$ , minimizing $\widehat{\mathcal{G}}(\mathbf{X}, \mathbf{W}, \mathbf{u})$ with respect to $\mathbf{u}$ , which obtains
This conclusion can be easily obtained according to proposition 2.
1Here, we replace $u = -u$ in Eq. (12).2) Fixed $\mathbf{u}$ , minimizing $\widehat{\mathcal{G}}(\mathbf{X}, \mathbf{W}, \mathbf{u})$ with respect to $\mathbf{W}$ and $\mathbf{X}$ , which is equivalent to
where $\text{Tr}(\cdot)$ denotes the trace of a matrix and $\mathbf{U}$ is a diagonal matrix whose diagonal element is $\mathbf{U}_{dd} = \frac{u_d}{\sigma^2}$ . It is clear that the optimization problem in (16) is a weighted NMF with an $\ell_1$ sparse constraint.
It is easy to prove that the objective function $\mathcal{G}$ in (10) is nonincreasing under the above iterative procedure.
Proof: Let $\mathbf{W}^{(t)}$ , $\mathbf{X}^{(t)}$ and $\mathbf{u}^{(t)}$ denote the solutions at $t$ -th iteration. According to Eq. (14), it is obvious that
According to the step 1 of the above alternate iteration minimization, because $\mathbf{u}^{(t+1)}$ is the solution of minimizing $\widehat{\mathcal{G}}(\mathbf{X}^{(t)}, \mathbf{W}^{(t)}, \mathbf{u})$ , then we obtain
Similarly, according the step 2, because $\mathbf{X}^{(t+1)}$ and $\mathbf{W}^{(t+1)}$ are the solutions of minimizing $\widehat{\mathcal{G}}(\mathbf{X}, \mathbf{W}, \mathbf{u}^{(t+1)})$ , we have
Summarizing Eqs. (17), (18) and (19), we have
In this way, the objective function $\mathcal{G}$ in Eq. (10) is non-increasing under the above alternative iterative procedure. ■#### D. Implementation details
Note that the key task of minimizing the objective function $\mathcal{G}$ in Eq. (10) is to solve the problem (16) in the step 2 of the above alternative iterative procedure. It is clear that this optimization problem (16) is a weighted $\ell_1$ -NMF, which can be further formulated as follows:
where $\hat{\mathbf{Y}} = \mathbf{U}^{\frac{1}{2}}\mathbf{Y}$ and $\hat{\mathbf{X}} = \mathbf{U}^{\frac{1}{2}}\mathbf{X}$ .
The above problem (21) is a typical $\ell_1$ sparse constrained NMF, which can be solved easily by multiplicative update algorithm [48].
Let $\Phi_{dk}$ and $\Psi_{kn}$ be the Lagrange multiplier for constraint $\hat{\mathbf{X}}{dk} \geq 0$ and $\mathbf{W}{kn} \geq 0$ respectively, and $\Phi = [\Phi_{dk}] \in \mathbb{R}^{D \times K}$ , $\Psi = [\Psi_{kn}] \in \mathbb{R}^{K \times N}$ . Then the Lagrange is
Then, the partial derivatives of $\mathcal{L}$ with respect to $\hat{\mathbf{X}}$ and $\mathbf{W}$ are
Using the Karush-Kuhn-Tucker conditions $\Phi_{dk}\hat{\mathbf{X}}{dk} = 0$ and $\Psi{kn}\mathbf{W}_{kn} = 0$ , we can obtain the following equations
Thus, the optimization problem (21) can be solved by the following multiplicative update rules
The detail steps of optimizing the problem (21) are presented in Algorithm 1.
Algorithm 1 Algorithm for optimizing the problem (21).
Input: Input data matrix $\hat{\mathbf{Y}}$ , the initial $\hat{\mathbf{X}}$ and $\mathbf{W}$ .
Output: Two matrices $\hat{\mathbf{X}}$ and $\mathbf{W}$ .
- 1: repeat
- 2: update $\hat{\mathbf{X}}$ by Eq. (27).
- 3: update $\mathbf{W}$ by Eq. (28).
- 4: until convergence
In summary, the optimization procedure of our model is outlined in Algorithm 2. As mentioned previously, the optimization procedure of our model is an adaptive weighted $\ell_1$ -NMF problem, which can automatically assign small weights to noisy bands. Specifically, the entries of the auxiliary vector $\mathbf{u} = [u_1, \dots, u_D]^T$ are viewed as the weights. Here, $u_d$ is the weight of the $d$ -th band. If the $d$ -th band of a hyperspectral image contains serious noise, then this band will yield a large reconstruction error to the LMM model. Consequently, the $d$ -th band will be assigned a small weight according to Eq. (15).
Finally, we analyze the computation complexity of the proposed algorithm. It is clear that the computation complexity of optimizing the $\ell_1$ -NMF problem (21) in Algorithm 1 is $\mathcal{O}(DNK)$ . According to the description in Section III-B, the optimization procedure of our model is equivalent to the iteratively re-weighted NMF with an $\ell_1$ sparse constraint. This implies that if the iteration of Algorithm 2 ends after $T$ steps, the complexity of our $\ell_1$ -CENMF is $\mathcal{O}(TDNK)$ .---
Algorithm 2 Algorithm of our $\ell_1$ -CENMF unmixing method.
Input: Hyperspectral data $\mathbf{Y} \in \mathbb{R}_+^{D \times N}$ , the number of endmembers $K$ , the sparse regularization parameter $\lambda$ and the stopping criteria $\epsilon$ .
Output: The endmember matrix $\mathbf{X} \in \mathbb{R}+^{D \times K}$ and the abundance matrix $\mathbf{W} \in \mathbb{R}+^{K \times N}$ .
- 1: Initialization. Initialize $\mathbf{X}$ with VCA [17], and then initialize $\mathbf{W}$ with the method in [49]. Initialize $u_d = 1, d = {1, \dots, D}$ .
- 2: while $|\mathcal{G}^{(t+1)} - \mathcal{G}^{(t)}| \geq \epsilon$ do
- 3: ${\mathcal{G}^{(t)} \text{ denotes the value of function } \mathcal{G}(\mathbf{X}, \mathbf{W}) \text{ defined in Eq. (10) at the } t\text{-th iteration.}}$
- 4: compute $\hat{\mathbf{Y}} = \mathbf{U}^{\frac{1}{2}} \mathbf{Y}, \hat{\mathbf{X}} = \mathbf{U}^{\frac{1}{2}} \mathbf{X}$ .
- 5: solve the optimization problem (21) according to Algorithm 1, and get the solution $\hat{\mathbf{X}}$ and $\mathbf{W}$ .
- 6: set $\mathbf{X} = \mathbf{U}^{-\frac{1}{2}} \hat{\mathbf{X}}$ .
- 7: update vector $\mathbf{u}$ and Gaussian kernel parameter $\sigma$ according to (15) and (29) respectively.
- 8: end while
IV. EXPERIMENTS
In this section, we employed both synthetic and real-world data to evaluate the performance of the proposed algorithm. To evaluate the proposed algorithm, we compared it with three related state-of-the-art methods: NMF [22], $\ell_1$ -NMF [48], and $\ell_{1/2}$ -NMF [7]. It is worth to notice that all of the methods need to initialize the endmember matrix $\mathbf{X}$ and the abundance matrix $\mathbf{W}$ first. Here, we adopt VCA [17] to provide the initial endmember matrix $\mathbf{X}$ and then utilize the method in [49] to obtain the initial abundance matrix $\mathbf{W}$ .
A. Parameter Settings
Notice that the parameter $\sigma$ of the Gaussian kernel in Eq. (15) will affect the performance of the proposed method, and it is often determined empirically. In this study, we adopt the method in [38] to determine the parameter $\sigma$ , which is computed by
In this paper, the default value of $\alpha$ is taken as $\alpha = 1$ .
There is another important parameter in our model: the regularization parameter $\lambda$ of the sparse penalty term. The value of parameter $\lambda$ is related to the sparseness of the material abundances.However, the material abundances are the unknown variables. Here, we adopt the sparseness criteria in [11], [7], [50] to estimate $\lambda$ from hyperspectral data,
where $\mathbf{y}^d$ denotes the $d$ -th band of the hyperspectral data.
B. Evaluation Metrics
To evaluate the performance of the proposed method quantitatively, we adopt spectral angle distance (SAD) and root-mean-square error (RMSE) to measure the accuracy of the extracted endmembers and the corresponding abundances, respectively. These two metrics are widely adopted to evaluate the performance of the unmixing algorithms [5], [51], [7], [11]. The SAD is used to measure the distance between the ground true endmember and its estimated endmember, which is defined as follows:
where $\bar{\mathbf{x}}_k$ denotes the $k$ -th true endmember and $\mathbf{x}_k$ is its estimated endmember ( $\mathbf{x}_k$ is the $k$ -th column of the endmember matrix $\mathbf{X}$ ).
Additionally, the RMSE is used to evaluate the distance between the ground true abundance map and its estimated abundance map, which is defined as follows:
where $\bar{\mathbf{w}}^k$ denotes the $k$ -th true abundance map and $\mathbf{w}^k$ is its estimated abundance map ( $\mathbf{w}^k$ is the $k$ -th row of the abundance matrix $\mathbf{W}$ ).
C. Experiments on Synthetic Data
In this subsection, we performed simulative experiments on synthetic data. The synthetic data is generated from a random selection of six true spectra from the United States Geological Survey (USGS) digital spectral library [52]. These six true signatures are shown in Fig. 1.Figure 1. The selected endmember spectra used to generate synthetic data.
We adopt the method in [53] to generate the synthetic data. The generation process includes following steps:
- An image of size $z^2 \times z^2$ is divided into units of $z \times z$ small blocks ( $z = 8$ in this work).
- Each block is filled up with the same type of ground cover, which is randomly selected from the six true spectra.
- Utilize a $(z + 1) \times (z + 1)$ spatial low-pass filter to smooth the image, which generate mixed pixels.
- Remove pure pixels and generate sparse abundances. Replace all the pixels whose largest abundance is larger than 0.8 with a mixture made up of only two endmembers with equal abundances.
- Simulate noise. Add zero-mean Gaussian noise to the generate image data. The Signal-to-Noise Ratio (SNR) is defined as follows:
where $\mathbf{y}$ denotes the true spectra and $\mathbf{n}$ is the noise and $\mathbb{E}[\cdot]$ denotes the expectation operator. Here, we assume the noise is both spatially and spectrally uncorrelated. Given a particular value of SNR, the variance of the zero-mean Gaussian noise is
Notice that, to simulate the hyperspectral data in accordance with real situation, we add different levels of noise for different bands. Specifically, the SNR of each band is sampled from a Gaussian distribution $\text{SNR} \sim \mathcal{N}(\overline{\text{SNR}}, \epsilon^2)$ . Here, we fix $\epsilon = 5$ in this work.
Here, we evaluate the performance of the proposed method in different noise levels. The proposed algorithm is compared with NMF, $\ell_1$ -NMF and $\ell_{1/2}$ -NMF. Figs. 2 and 3 show the performance results with different noise levels ( $\overline{\text{SNR}} = {50, 40, 30, 20, 10}$ ). Fig. 2 presents the SAD curves of the four methods at different noise levels. Each sub-figure in Fig. 2 stands for the SAD of one of the endmembers. Fig. 3 presents the RMSE of the four methods at different noise levels. Each sub-figure in Fig. 3 denotes the RMSE of one of the abundance maps. It is obvious that our method ( $\ell_1$ -CENMF) achieves much lower values than other methods both in terms of SAD and RMSE. The main reason lies in that our method takes into account the difference of noise levels in different bands and models the noise in each band accurately. Therefore, our model performs significantly better than other methods. Meanwhile, as can be seen from Figs. 2 and 3, $\ell_1$ -NMF and $\ell_{1/2}$ -NMF outperform the original NMF method. This is mainly because the sparsity constraint is consistent with the sparseness of the abundance maps in the data.
D. Experiments on Real Data
In this subsection, we applied our unmixing model to two real hyperspectral images: the HYDICE Urban data set and the Jasper Ridge data set. Both of the data sets contain noisy bands due to dense water vapor, atmospheric effects, and sensor noise.
Urban data set is one of the most widely used hyperspectral data for hyperspectral unmixing researches [5], [51], [7]. There are $307 \times 307$ pixels in this data set. It is composed of 210 spectral channels with a spectral resolution of 10nm from 400nm to 2500nm. As can be seen from Fig.Figure 2. Comparison of the algorithms at different noise levels in terms of SAD. (a) asphalt-gds366. (b) board-gds362. (c) plastic-gds369. (d) sheetmetal-gds352. (e) cedar-gds357. (f) fiberglass-gds355.
4, the noise intensity of each band in this data set is different. Meanwhile, we can obviously observe that the SNR of each band in this figure satisfies $\text{SNR}{205} > \text{SNR}{206} > \text{SNR}{207} > \text{SNR}{208}$ . Actually, the bands 1 – 4, 76, 87, 101 – 111, 136 – 153 and 198 – 210 are corrupted by serious noise. For most of the existing unmixing methods, these bands with low SNR should be removed before performing unmixing algorithms. Here, we take all of bands to perform the unmixing algorithms.
According to the ground truths [54] of the data, there are four interesting endmembers: Tree, Roof, Asphalt and Grass. Fig. 5 presents the intuitive compassion of the abundance maps obtained by all of the methods. As can be seen from this figure, the abundance maps obtained by NMF, $\ell_1$ -NMF and $\ell_{1/2}$ -NMF contain more noise than the results of our method. The reason is that NMF, $\ell_1$ -NMF and $\ell_{1/2}$ -NMF consider that all bands in the data contribute equally to the unmixing procedure, thus the noisy bands will significantly affect the unmixing results. By comparison, our model, automatically and adaptively assign small weights to those bands with low SNR, isFigure 3. Comparison of the algorithms at different noise levels in terms of RMSE. (a) asphalt-gds366. (b) board-gds362. (c) plastic-gds369. (d) sheetmetal-gds352. (e) cedar-gds357. (f) fiberglass-gds355.
Figure 4. Some noisy bands in Urban data set. The noise level in this data set differs from band to band.
more insensitive to noisy bands. As can be seen from Fig. 5, our model obtains the most similar abundance maps compared with the ground truths. The quantitative comparison of $SAD^2$ and RMSE on the Urban data is shown in Table I. From this table, we see that our $\ell_1$ -CENMF model obtains much better results than other methods.
2Here, we removed the spectra of the noise corrupted bands (bands 1-4, 76, 87, 101-111, 136-153 and 198) before computing the SAD.Figure 5. Abundance maps on the Urban data set. From the first row to the fourth row denote the abundance maps obtained by NMF, $\ell_1$ -NMF, $\ell_{1/2}$ -NMF and our $\ell_1$ -CENMF, respectively. The last row illustrates the Ground Truths.
After the optimization algorithm of our model converged, we can get the weight of each band according to the auxiliary vector $\mathbf{u}$ . Fig. 6 plots the weights of all bands in Urban data set. It is clearly seen that the bands with small weights correspond to the bands with serious noiseTable I
THE SAD AND RMSE OF THE FOUR METHODS ON URBAN DATA SET.
| Endmember | SAD | RMSE | ||||||
|---|---|---|---|---|---|---|---|---|
| NMF | -NMF | -NMF | -CENMF | NMF | -NMF | -NMF | -CENMF | |
| Tree | 0.1654 | 0.1403 | 0.0923 | 0.0882 | 0.2930 | 0.2924 | 0.2572 | 0.1496 |
| Roof | 0.5609 | 0.4924 | 0.3019 | 0.2576 | 0.2017 | 0.2014 | 0.1914 | 0.1137 |
| Asphalt | 0.3291 | 0.3165 | 0.2570 | 0.1090 | 0.2639 | 0.2242 | 0.2239 | 0.1590 |
| Grass | 0.4509 | 0.3861 | 0.3627 | 0.2209 | 0.4015 | 0.3557 | 0.3546 | 0.1928 |
| Mean | 0.3766 | 0.3338 | 0.2535 | 0.1689 | 0.2901 | 0.2684 | 0.2568 | 0.1538 |
(bands 1 – 4, 76, 87, 101 – 111, 136 – 153 and 198 – 210). Here, we show the image of the 105-th band of Urban data set in Fig. 7. As can be seen from this figure, this band is corrupted by serious noise. Meanwhile, this band is assigned a small weight as illustrated in Fig. 6. This demonstrates that our correntropy based model can effectively cope with the noise bands in the hyperspectral image.
Figure 6. The weights of all bands in Urban data.Figure 7. The image of the 105-th band in Urban data set.
The other real data is the Jasper Ridge data set. It is another popular hyperspectral data set introduced in ENVI Tutorials [55] for hyperspectral analysis. It is composed of 224 spectral channels covering the wavelength from 380nm to 2500nm, with a spectral resolution of 10nm. After the black bands (bands 1, 108-112 and 154-166) are removed, only 205 bands remain. The selected subregion from the Jasper Ridge data set is shown in Fig. 8, which contains $100 \times 100$ pixels. There are four interesting endmembers: Water, Tree, Soil and Road. It can be observed from Fig. 8 that this data set contains noisy bands and the noise level in each band is different from those of the others.
Figure 8. Some bands in Jasper data set.
Fig. 9 presents the abundance maps obtained by all the methods. This figure shows that our$\ell_1$ -CENMF method yields the most similar abundance maps compared with the ground truths. Table II shows the quantitative comparison in terms of SAD and RMSE on the Jasper Ridge
Figure 9. Abundance maps on Jasper Ridge data set. From the first row to the fourth row denote the abundance maps obtained by NMF, $\ell_1$ -NMF, $\ell_{1/2}$ -NMF and our $\ell_1$ -CENMF, respectively. The last row illustrates the Ground Truths.
data. It is shown that our $\ell_1$ -NMF obtains the best performance for most of the endmembers and their corresponding abundance maps.Table II
THE SAD AND RMSE OF THE FOUR METHODS ON JASPER DATA SET.
| Endmember | SAD | RMSE | ||||||
|---|---|---|---|---|---|---|---|---|
| NMF | -NMF | -NMF | -CENMF | NMF | -NMF | -NMF | -CENMF | |
| Water | 0.1081 | 0.0935 | 0.0915 | 0.0602 | 0.1791 | 0.1764 | 0.1958 | 0.0707 |
| Tree | 0.1336 | 0.1213 | 0.1285 | 0.1155 | 0.1214 | 0.1041 | 0.1088 | 0.1121 |
| Soil | 0.2194 | 0.2110 | 0.1713 | 0.1472 | 0.1312 | 0.1018 | 0.1034 | 0.0904 |
| Road | 0.2046 | 0.1713 | 0.1618 | 0.1015 | 0.1837 | 0.1782 | 0.2092 | 0.0949 |
| Mean | 0.1643 | 0.1493 | 0.1382 | 0.1061 | 0.1539 | 0.1401 | 0.1543 | 0.0920 |
We also present the weights of all bands in Jasper Ridge data in Fig. 10. It can be observed that the first band is assigned a small weight. We show the first band of the Jasper Ridge data in Fig. 11. As can be seen from this figure, this band contains high-level noise. This illustrates that our model can adaptively assign small weights to the bands with serious noise, which is the main reason that our $\ell_1$ -CENMF outperforms the other methods.
In order to further demonstrate the effectiveness of the proposed model, we perform another experiment on Urban data set. In this experiment, we first remove the noisy bands (bands 1 – 4, 76, 87, 101 – 111, 136 – 153 and 198 – 210), thus only 162 bands remain. Then, $\ell_{1/2}$ -NMF and our $\ell_1$ -CENMF are applied to unmix this data. Table III reports the performances of the two methods in terms of SAD and RMSE. For comparison, Table III also shows the quantitative results of Urban data set with all of the bands have been used to perform spectral unmixing. Here, the table heading (162) denotes the Urban data with 162 bands, and the table heading (all) denotes the Urban data with all bands. From the results listed in Table III, we can conclude the following three points. First, $\ell_{1/2}$ -NMF achieves better results on Urban data with 162 bands than that with all bands. The reason is that $\ell_{1/2}$ -NMF considers that all bands contribute equally to spectral unmixing, thus the noisy bands will significantly affect the unmixing results. This implies that $\ell_{1/2}$ -NMF is very sensitive to noisy bands. Second, the results of our $\ell_1$ -CENMF on the Urban data with 162 bands and all bands are very close. The reason is that, the noisy bands, which are adaptively assigned to small weights, have little effect to the objective function. Accordingly, our method is very robust to noisy bands. Third, as demonstrated in Table III, ourFigure 10. The weights of all bands in Jasper Ridge data.
method has better performance both on the data of 162 bands and all bands than $\ell_1$ -NMF does on the data of 162 bands. The reasons are twofold. On the one hand, the removed bands are manually selected, which leads to discarding available spectral information. As illustrated in Fig. 4, although there is some noise in the 205-th band, it contains useful spectral information. Our method can effectively and adaptively exploit the useful spectral information from all of the bands. On the other hand, due to the fact that the noise level in different bands is different and manually removing noisy bands is primarily dependent on subjective observation, the rest of the bands may still contains noise.
In summary, our method can adaptively assign small weights to bands with low SNR and give more emphasis on bands with high SNR. The adaptive capacity of our model can utilize the data more effectively for hyperspectral unmixing, and thus facilitates its real-world applications.Figure 11. The image of the first band in Jasper Ridge data.
Table III
THE SAD AND RMSE OF $\ell_{1/2}$ -NMF AND $\ell_1$ -CENMF METHODS ON URBAN DATA SET. THE COLUMN HEADING (162) DENOTES THE URBAN DATA WITH 162 BANDS, AND THE COLUMN HEADING (ALL) DENOTES THE URBAN DATA WITH ALL BANDS.
| Endmember | SAD | RMSE | ||||||
|---|---|---|---|---|---|---|---|---|
| -NMF (all) |
-NMF (162) |
-CENMF (all) |
-CENMF (162) |
-NMF (all) |
-NMF (162) |
-CENMF (all) |
-CENMF (162) |
|
| Tree | 0.0923 | 0.0751 | 0.0882 | 0.0699 | 0.2572 | 0.1488 | 0.1496 | 0.1407 |
| Roof | 0.3019 | 0.2513 | 0.2576 | 0.2457 | 0.1914 | 0.1350 | 0.1137 | 0.1250 |
| Asphalt | 0.2570 | 0.1409 | 0.1090 | 0.1258 | 0.2239 | 0.1586 | 0.1590 | 0.1391 |
| Grass | 0.3627 | 0.2285 | 0.2209 | 0.2143 | 0.3546 | 0.2296 | 0.1928 | 0.2028 |
| Mean | 0.2535 | 0.1739 | 0.1689 | 0.1640 | 0.2568 | 0.1581 | 0.1538 | 0.1519 |
V. CONCLUSIONS
In this paper, we have proposed a robust hyperspectral unmixing algorithm based on correntropy based metric. The main contributions of this work were as follows: (1) the proposed model utilizes correntropy based metric to define the objective function, which can adaptively assign small weights to noisy bands and give more emphasis on noise-free bands. Thus, our model is much more insensitive to noise; (2) a convergent optimization algorithm is proposed based on half-quadratic optimization technique. Additionally, theoretical analyses about the derived modelfor its robustness were given. Comparative experiments on both synthetic and real-world data show that the proposed model achieves much better results than the state-of-the-art methods.
Although the optimization algorithm of the proposed method is convergent, the convergence is slow. The main reason is that the multiplicative update algorithm was adopted to solve the subproblem (16) in step 2 of the alternate optimization procedure, which suffers from drawback of slow convergence. In the future, we would like to speed up our algorithm by adopting more efficient optimization techniques.
REFERENCES
- [1] N. Keshava and J. Mustard, "Spectral unmixing," IEEE Signal Processing Magazine, vol. 19, no. 1, pp. 44–57, jan 2002.
- [2] A. Plaza, Q. Du, J. M. Bioucas-Dias, X. Jia, and F. A. Kruse, "Foreword to the special issue on spectral unmixing of remotely sensed data," IEEE Transactions on Geoscience and Remote Sensing, vol. 49, no. 11, pp. 1–8, nov. 2011.
- [3] J. Bioucas-Dias, A. Plaza, N. Dobigeon, M. Parente, Q. Du, P. Gader, and J. Chanussot, "Hyperspectral unmixing overview: Geometrical, statistical, and sparse regression-based approaches," IEEE Journal of Selected Topics in Applied Earth Observations and Remote Sensing, vol. 5, no. 2, pp. 354–379, april 2012.
- [4] D. Gillis, J. Bowles, E. J. Ientilucci, and D. W. Messinger, "A generalized linear mixing model for hyperspectral imagery," in Algorithms and Technologies for Multispectral, Hyperspectral, and Ultraspectral iMage XIV, SPIE, 2008, pp. 69661B–69661B–11.
- [5] S. Jia and Y. Qian, "Constrained nonnegative matrix factorization for hyperspectral unmixing," IEEE Transactions on Geoscience and Remote Sensing, vol. 47, no. 1, pp. 161–173, jan. 2009.
- [6] A. Huck, M. Guillaume, and J. Blanc-Talon, "Minimum dispersion constrained nonnegative matrix factorization to unmix hyperspectral data," IEEE Transactions on Geoscience and Remote Sensing, vol. 48, no. 6, pp. 2590–2602, june 2010.
- [7] Y. Qian, S. Jia, J. Zhou, and A. Robles-Kelly, "Hyperspectral unmixing via $l_{1/2}$ sparsity-constrained nonnegative matrix factorization," IEEE Transactions on Geoscience and Remote Sensing, vol. 49, no. 11, pp. 4282–4297, nov. 2011.
- [8] M.-D. Iordache, J. Bioucas-Dias, and A. Plaza, "Sparse unmixing of hyperspectral data," IEEE Transactions on Geoscience and Remote Sensing, vol. 49, no. 6, pp. 2014–2039, june 2011.
- [9] Z. Yang, G. Zhou, S. Xie, S. Ding, J.-M. Yang, and J. Zhang, "Blind spectral unmixing based on sparse nonnegative matrix factorization," IEEE Transactions on Image Processing, vol. 20, no. 4, pp. 1112–1125, april 2011.
- [10] M.-D. Iordache, J. Bioucas-Dias, and A. Plaza, "Total variation spatial regularization for sparse hyperspectral unmixing," IEEE Transactions on Geoscience and Remote Sensing, vol. 50, no. 11, pp. 4484–4502, nov. 2012.
- [11] X. Lu, H. Wu, Y. Yuan, P. Yan, and X. Li, "Manifold regularized sparse nmf for hyperspectral unmixing," IEEE Transactions on Geoscience and Remote Sensing, vol. PP, no. 99, pp. 1–12, Oct 2012.
- [12] A. Halimi, Y. Altmann, N. Dobigeon, and J. Tourneret, "Nonlinear unmixing of hyperspectral images using a generalized bilinear model," IEEE Transactions on Geoscience and Remote Sensing, vol. 49, no. 11, pp. 4153–4162, nov. 2011.- [13] Y. Altmann, N. Dobigeon, S. McLaughlin, and J. Tourneret, “Unsupervised nonlinear unmixing of hyperspectral images using gaussian processes,” in IEEE International Conference on Acoustics, Speech and Signal Processing (ICASSP), march 2012, pp. 1249–1252.
- [14] M. E. Winter, “N-findr: an algorithm for fast autonomous spectral end-member determination in hyperspectral data.” Proc. SPIE Conf. Imaging Spectrometry V, vol. 3753, 1999.
- [15] A. Plaza, P. Martinez, R. Perez, and J. Plaza, “Spatial/spectral endmember extraction by multidimensional morphological operations,” IEEE Transactions on Geoscience and Remote Sensing, vol. 40, no. 9, pp. 2025–2041, sep 2002.
- [16] ———, “A quantitative and comparative analysis of endmember extraction algorithms from hyperspectral data,” IEEE Transactions on Geoscience and Remote Sensing, vol. 42, no. 3, pp. 650–663, march 2004.
- [17] J. M. P. Nascimento and J. M. B. Dias, “Vertex component analysis: a fast algorithm to unmix hyperspectral data.” IEEE Transactions on Geoscience and Remote Sensing, vol. 43, no. 4, pp. 898–910, 2005.
- [18] T.-H. Chan, W.-K. Ma, A. Ambikapathi, and C.-Y. Chi, “A simplex volume maximization framework for hyperspectral endmember extraction,” IEEE Transactions on Geoscience and Remote Sensing, vol. 49, no. 11, pp. 4177–4193, nov. 2011.
- [19] Y. Zhang, X. Fan, Y. Zhang, and R. Wei, “Linear spectral unmixing with generalized constraint for hyperspectral imagery,” in IEEE International Geoscience and Remote Sensing Symposium (IGARSS), july 2012, pp. 4106–4109.
- [20] J. Bayliss, J. A. Gualtieri, and R. F. Crompton, “Analyzing hyperspectral data with independent component analysis,” in Proc. SPIE AIPR Workshop, vol. 3240. SPIE, 1997, pp. 133–143.
- [21] N. Dobigeon, S. Moussaoui, M. Coulon, J.-Y. Tourneret, and A. Hero, “Joint bayesian endmember extraction and linear unmixing for hyperspectral imagery,” IEEE Transactions on Signal Processing, vol. 57, no. 11, pp. 4355–4368, nov. 2009.
- [22] D. D. Lee and H. S. Seung, “Algorithms for non-negative matrix factorization.” in NIPS, 2000, pp. 556–562.
- [23] S. Z. Li, X. Hou, H. Zhang, and Q. Cheng, “Learning spatially localized, parts-based representation.” in CVPR, 2001, pp. 207–212.
- [24] D. Manolakis and G. Shaw, “Detection algorithms for hyperspectral imaging applications,” IEEE Signal Processing Magazine, vol. 19, no. 1, pp. 29–43, jan 2002.
- [25] W. Liu, P. Pokharel, and J. Principe, “Correntropy: Properties and applications in non-gaussian signal processing,” IEEE Transactions on Signal Processing, vol. 55, no. 11, pp. 5286–5298, nov. 2007.
- [26] X. Yuan and H. Gang, “Robust feature extraction via information theoretic learning,” in Proceedings of the 26th Annual International Conference on Machine Learning. ACM, 2009, pp. 1193–1200.
- [27] M. W. Berry, M. Browne, A. N. Langville, V. P. Pauca, and R. J. Plemmons, “Algorithms and applications for approximate nonnegative matrix factorization,” Computational Statistics and Data Analysis, vol. 52, pp. 155–173, 2007.
- [28] C.-J. Lin, “Projected gradient methods for nonnegative matrix factorization,” Neural Comput., vol. 19, no. 10, pp. 2756–2779, Oct. 2007.
- [29] H. Kim and H. Park, “Nonnegative matrix factorization based on alternating nonnegativity constrained least squares and active set method,” SIAM Journal on Matrix Analysis and Applications, vol. 30, no. 2, pp. 713–730, 2008.
- [30] D. Cai, X. He, J. Han, and T. Huang, “Graph regularized nonnegative matrix factorization for data representation,” IEEE Transactions on Pattern Analysis and Machine Intelligence, vol. 33, no. 8, pp. 1548–1560, aug. 2011.- [31] H. Liu, Z. Wu, X. Li, D. Cai, and T. Huang, "Constrained nonnegative matrix factorization for image representation," IEEE Transactions on Pattern Analysis and Machine Intelligence, vol. 34, no. 7, pp. 1299–1311, july 2012.
- [32] M. Arngren, M. Schmidt, and J. Larsen, "Bayesian nonnegative matrix factorization with volume prior for unmixing of hyperspectral images," in IEEE International Workshop on Machine Learning for Signal Processing, sept. 2009, pp. 1–6.
- [33] T. Zhang, B. Fang, Y. Y. Tang, G. He, and J. Wen, "Topology preserving non-negative matrix factorization for face recognition," IEEE Transactions on Image Processing, vol. 17, no. 4, pp. 574–584, april 2008.
- [34] R. Sandler and M. Lindenbaum, "Nonnegative matrix factorization with earth mover's distance metric for image analysis," IEEE Transactions on Pattern Analysis and Machine Intelligence, vol. 33, no. 8, pp. 1590–1602, aug. 2011.
- [35] V. P. Pauca, J. Piper, and R. J. Plemmons, "Nonnegative matrix factorization for spectral data analysis," Linear Algebra and its Applications, vol. 416, no. 1, pp. 29–47, 2006.
- [36] J. V. Davis, B. Kulis, P. Jain, S. Sra, and I. S. Dhillon, "Information-theoretic metric learning," in Proceedings of the 24th international conference on Machine learning, ser. ICML '07. New York, NY, USA: ACM, 2007, pp. 209–216.
- [37] I. Santamaria, P. Pokharel, and J. Principe, "Generalized correlation function: definition, properties, and application to blind equalization," IEEE Transactions on Signal Processing, vol. 54, no. 6, pp. 2187–2197, june 2006.
- [38] R. He, W.-S. Zheng, and B.-G. Hu, "Maximum correntropy criterion for robust face recognition," IEEE Transactions on Pattern Analysis and Machine Intelligence, vol. 33, no. 8, pp. 1561–1576, aug. 2011.
- [39] V. N. Vapnik, The nature of statistical learning theory. New York, NY, USA: Springer-Verlag New York, Inc., 1995.
- [40] P. J. Huber, Robust Statistics. John Wiley, 1981.
- [41] C. V. Stewart, "Robust parameter estimation in computer vision," SIAM Review, vol. 41, no. 3, pp. 513–537, Sep. 1999.
- [42] E. Candès and T. Tao, "Decoding by linear programming," IEEE Transactions on Information Theory, vol. 51, no. 12, pp. 4203–4215, dec. 2005.
- [43] D. L. Donoho, "Compressed sensing," IEEE Transactions on Information Theory, vol. 52, no. 4, pp. 1289–1306, april 2006.
- [44] J. Wright, A. Yang, A. Ganesh, S. Sastry, and Y. Ma, "Robust face recognition via sparse representation," IEEE Transactions on Pattern Analysis and Machine Intelligence, vol. 31, no. 2, pp. 210–227, feb. 2009.
- [45] M. Nikolova and M. K. Ng, "Analysis of half-quadratic minimization methods for signal and image recovery," SIAM Journal on Scientific Computing, vol. 27, no. 3, pp. 937–966, Oct. 2005.
- [46] M. Nikolova and R. Chan, "The equivalence of half-quadratic minimization and the gradient linearization iteration," IEEE Transactions on Image Processing, vol. 16, no. 6, pp. 1623–1627, june 2007.
- [47] S. Boyd and L. Vandenberghe, Convex Optimization. New York, NY, USA: Cambridge University Press, 2004.
- [48] P. O. Hoyer, "Non-negative sparse coding," in NEURAL NETWORKS FOR SIGNAL PROCESSING XII (PROC. IEEE WORKSHOP ON NEURAL NETWORKS FOR SIGNAL PROCESSING), 2002, pp. 557–565.
- [49] D. Heinz and Chein-I-Chang, "Fully constrained least squares linear spectral mixture analysis method for material quantification in hyperspectral imagery," IEEE Transactions on Geoscience and Remote Sensing, vol. 39, no. 3, pp. 529–545, mar 2001.
- [50] P. O. Hoyer, "Non-negative matrix factorization with sparseness constraints," Journal of Machine Learning Research, vol. 5, pp. 1457–1469, Dec. 2004.- [51] X. Liu, W. Xia, B. Wang, and L. Zhang, "An approach based on constrained nonnegative matrix factorization to unmix hyperspectral data," IEEE Transactions on Geoscience and Remote Sensing, vol. 49, no. 2, pp. 757–772, feb. 2011.
- [52] R. Clark, G. Swayze, R. Wise, E. Livo, T. Hoefen, R. Kokaly, and S. Sutley, "Usgs digital spectral library splib06a;," U.S. Geological Survey, Digital Data Series 231, 2007.
- [53] L. Miao and H. Qi, "Endmember extraction from highly mixed data using minimum volume constrained nonnegative matrix factorization," IEEE Transactions on Geoscience and Remote Sensing, vol. 45, no. 3, pp. 765–777, march 2007.
- [54] S. Jia and Y. Qian, "Spectral and spatial complexity-based hyperspectral unmixing," IEEE Transactions on Geoscience and Remote Sensing, vol. 45, no. 12, pp. 3867–3879, dec. 2007.
- [55] Envi Tutorials, 2006. [Online]. Available: http://www.exelisvis.com/language/en-us/productsservices/envi/tutorials.aspx
Xet Storage Details
- Size:
- 61.5 kB
- Xet hash:
- a9a6151e616bb94b9396eed3ccfdf0295d87354446f8b97357557bbbcc8aad2f
Xet efficiently stores files, intelligently splitting them into unique chunks and accelerating uploads and downloads. More info.