Skip to main content
Advertisement
Browse Subject Areas
?

Click through the PLOS taxonomy to find articles in your field.

For more information about PLOS Subject Areas, click here.

  • Loading metrics

Collaborative representation based on enhanced tensor robust PCA for hyperspectral anomaly detection

  • Ruhan A,

    Roles Data curation, Formal analysis, Investigation, Software, Validation, Visualization, Writing – original draft, Writing – review & editing

    Affiliations Xianyang Normal University, Xianyang, Shaanxi, China, Xi’an Peihua University, Xi’an, Shaanxi, China

  • Cheng Liang Zhong ,

    Roles Conceptualization, Formal analysis, Investigation

    zhongcl19@mails.tsinghua.edu.cn

    Affiliation Tsinghua University, Beijing, China

  • Senjia Wang,

    Roles Visualization, Writing – review & editing

    Affiliation Zhengzhou University of Industrial Technology, Xinyang, Henan, China

  • Quanxue Gao

    Roles Conceptualization, Methodology, Validation

    Affiliation Xidian University, Xi’an, Shaanxi, China

Abstract

This paper presents a novel hyperspectral anomaly detection (HAD) method, ETRPCA-CRD, which integrates enhanced tensor robust principal component analysis (ETRPCA) with collaborative representation detection (CRD) to effectively separate anomalous targets from background data. The key novelty lies in the use of weighted tensor Schatten-p norm minimization (WTSNM) within the ETRPCA framework, which assigns distinct weights to different singular values to preserve important information while eliminating noise. The ETRPCA problem is efficiently solved by Fourier transform, generalized soft-thresholding (GST), and T-singular value decomposition (SVD) methods. This approach significantly improves detection accuracy by fully utilizing the spectral-spatial information of hyperspectral images (HSIs) represented as tensors. The low-rank tensor obtained from ETRPCA serves as the background data for CRD, further enhancing detection performance. Experiments on three real hyperspectral datasets and one simulated dataset demonstrate that ETRPCA-CRD outperforms several state-of-the-art algorithms, achieving superior detection accuracy and robustness. The proposed method’s ability to effectively distinguish anomalies from background data while preserving salient signals makes it a powerful tool for hyperspectral anomaly detection.

Introduction

In hyperspectral images (HSIs), accurately identifying anomalous targets can be challenging due to contamination by abnormal data and background noise. To address this problem, various algorithms have been developed to separate clean background data from anomalous target data, thereby improving detection performance.

Based on the principle that anomalous pixels cannot be linearly represented by their surrounding data while background pixels can, Li and Du presented a classic hyperspectral anomaly detection (HAD) method, known as collaborative representation detection (CRD) [1], in which background data are collected in a sliding dual window. CRD effectively detects anomalous pixels by removing the approximated background from the original HSIs. Building upon CRD, kernel collaborative representation detection was introduced to further improve the anomaly detection performance. To address the problem of potential outliers in the sliding dual window in CRD, Su et al. proposed CRD with Principal Component Analysis (PCA) remove outlier (PCAroCRD) [2], which uses principal component analysis to purify the background data. In this approach, the top principal components contain the most background information, while the latter contain anomaly-related information. This purification process improves the anomaly detection accuracy. Yin et al. proposed a method called the selective search collaborative representation detector for HAD [3]. Their method utilizes global information and spectral similarity to fuse similar pixels. This selective search technique purifies the local background estimation and reduces the reliance on the dual window size in CRD. Another approach, known as the copula-based CRD, was introduced to address challenges related to limited spatial information use and anomalous target contamination [4]. In this approach, background purification is performed using copula-based outlier detection, and spatial difference enhancement is achieved using guided filtering, effectively distinguishing anomalous targets from the background. Additionally, Ma et al. proposed a novel representation-theory-based CRD approach that allows the detector to use nonlinear features beyond the Gaussian distribution [5]. Furthermore, they proposed an anomaly contamination detection system to address potential false positives due to neighboring anomalous pixels. Hu et al. introduced an algorithm for target detection based on binary-class collaborative representation (BCRD) [6], where background data are linearly represented by surrounding pixels, and anomalous targets are approximated by prior information. Anomaly detection is achieved by estimating the residuals of both representations. BCRD enhances the detection accuracy through background dictionary purification; however, it requires prior knowledge of anomalous targets, which is often challenging in practical HAD applications.

Additionally, low-rank and sparse matrix decomposition-based methods can precisely distinguish between abnormal target and background data. The low-rank and sparse matrix decomposition detector (LRaSMD) [7] was developed based on the assumption that noise can be simulated by Gaussian random variables, that abnormal targets follow a sparse distribution, and that the background image is low-rank. In addition, the low-rank and sparse matrix decomposition-based Mahalanobis distance (LSMAD) [8] suppresses abnormal target contamination in the background.

Robust principal component analysis(RPCA) [9] decomposes a matrix into low-rank and sparse components. However, it may not effectively approximate rank in practice. Additionally, nuclear norm regularization in RPCA shrinks all singular values equally, although different singular values contain different information. To address this limitation, methods such as partial sum minimization of singular values in Robust PCA(PSSV) [10], truncated nuclear norm [11], and weighted nuclear norm minimization [12] have been proposed to preserve important information.

On the basis of the above research, we propose the enhanced tensor RPCA-CRD (ETRPCA-CRD) method for HAD, in which tensor data are reconstructed into a low-rank tensor and sparse tensor using ETRPCA. Weighted tensor Schatten-p norm minimization (WTSNM) retains important information by treating different singular values differently, and an efficient method is employed to solve ETRPCA. In addition, the use of as the background data for CRD effectively enhances the detection performance and accuracy. Our proposed method is compared with several state-of-the-art methods on three real datasets and one simulated hyperspectral dataset.

  • In HAD, there are both spectral and spatial differences between abnormal targets and surrounding background data. By ignoring the spatial correlation of the data, previous HAD approaches expanded three-dimensional (3D) hyperspectral data into two-dimensional matrices. In this paper, HSIs are represented by tensors to retain the original information and structure of the data and fully utilize the spectral and spatial information of the HSIs in detection.
  • The tensor data are reconstructed into a low-rank tensor and sparse tensor using the ETRPCA method. The background information in is retained, whereas anomalous targets and noisy information are removed. Therefore, using as the background data for the CRD algorithm effectively enhances the detection performance and accuracy.
  • Our proposed method employs WTSNM, in which each singular value is treated separately. Therefore, the important information of large singular values is retained due to less shrinkage. By assigning a specific weight to each singular value, the salient signals and filters can be preserved out of noise, greatly improving the detection accuracy.
  • The ETRPCA problem is efficiently solved by Fourier transform, generalized soft-thresholding (GST), and T-singular value decomposition (SVD) methods.

Proposed algorithm

Tensor decomposition has been selected as the core technique for processing hyperspectral data in this study, owing to its unique advantages in handling the high dimensionality and complex structure of such data. By representing hyperspectral images as multi-dimensional tensors, tensor decomposition preserves their inherent spectral-spatial structure, thereby avoiding the information loss often associated with traditional matrix flattening approaches. This method is particularly effective in capturing the multilinear relationships between different dimensions, which is crucial for accurately identifying anomalies that exhibit distinct spectral signatures and spatial patterns. The enhanced tensor robust principal component analysis (ETRPCA) employed in this study is designed to be robust against noise and outliers, enabling more reliable separation of background and anomaly signals. The computational complexity associated with ETRPCA is effectively managed through the use of Fourier transforms, generalized soft-thresholding (GST), and T-singular value decomposition (SVD) methods. These techniques collectively ensure that the ETRPCA problem can be solved efficiently, thereby addressing the inherent computational challenges and maintaining the practicality of the method for large-scale datasets.

This paper adopts a notation convention for different mathematical entities. Matrices (e.g., ), vectors (e.g., ), and scalars (e.g., a) are represented by uppercase bold letters, lowercase bold letters, and lowercase letters, respectively. An identity matrix is represented by , while a third-order tensor is denoted by bold calligraphic letters (e.g., ). Real and complex fields are denoted by and , respectively. The entries of a tensor are denoted by or simply aijk, where . Slices of are represented as (horizontal slice), (lateral slice), and (frontal slice). A matrix representing the i-th frontal slice is denoted by . A tube of a third-order tensor is defined by fixing the first two indices and varying the third, such as , representing the ij-th tube of . The inner product between matrices and in is defined as , where denotes the conjugate transpose. For tensors and in , their inner product is defined as . In the context of real tensors , the result of the discrete fast Fourier transform (FFT) along the third dimension is represented by , denoted by . The inverse FFT is denoted by . can be expressed as a block diagonal matrix, where each block on the diagonal is the frontal slice of [13]. Several norms of matrix A employed in this paper are the -norm , Frobenius norm , nuclear norm , and -norm . The where is a low-rank tensor (Background information) and is a sparse tensor (noise and anomalous target). To accurately recover low-rank tensors and sparse tensors, in previous studies, tensor nuclear norm minimization[14] was used as a convex relaxation. However, in solving the tensor nuclear norm minimization problem, all singular values are regularized and shrinked equally. Different singular values contain different information, especially the first several large singular values, which contain most of the prominent information. To preserve the prominent information, first several large singular values should shrink less. Therefore, the weighted tensor Schatten-p norm [15] is used in this paper.

To preserve the original spectral and spatial information of the HSIs, we represent the hyperspectral data by a 3D tensor, denoted by . We then decompose this tensor, , into a low-rank tensor and sparse tensor :

(1)

where represents background information, and contains noise and anomalous targets. Tensor nuclear norm minimization (TNNM) was previously employed as a convex relaxation for recovering low-rank and sparse tensors. However, TNNM tends to regularize and shrink all singular values equally. Since different singular values contain different information, it is crucial to shrink the first several large singular values less because they contain most of the important information. Therefore, this paper adopts the weighted tensor Schatten-p norm (WTSN) [1517]. For and , the WTSN of is defined as

(2)

where denotes a nonnegative weighted vector. The weights are set to the inverse ratio of the singular values to ensure less shrinkage of the first several large singular values: , where C is a compromising constant, denotes the j-th largest singular value of a matrix, ε is a small positive number to avoid division by zero, and p (0 < p < 1) is the power of the matrix singular value.

The objective function is given by

(3)

where is the Lagrange Multiplier and μ is a positive scaler. From this problem, update with other fixed variables is given by By solving Eq 3, we can recover and . The optimization problem is solved using the augmented Lagrange multiplier method. The constrained problem is converted into an unconstrained one with additional penalty terms, as follows:

(4)

Here, μ is a positive scalar, and is the Lagrange multiplier. Letting other variables be fixed, is updated as follows:

(5)

Here, . Inspired by the soft-thresholding operator, we obtain

(6)

Here, the (i,j,k)-th element of is .

With other variables fixed, is updated as follows:

(7)

Here, , which is referred to as WTSNM [1517] and can be efficiently solved using GST.

The other variables are updated as follows:

(8)(9)

GST is employed to effectively solve the WTSNM problem, as briefly described below.

Theorem 1. For matrices A and B, with singular values denoted by and , where , it holds that . Equality occurs if and only if there exist unitaries U and V such that and , where and are ordered eigenvalue matrices with singular values and along the diagonal with the same order, respectively [18,19].

Lemma 1. Let the SVD of Y be with , assuming a nonascending order of all singular values [18,19]. For a given , , and , the following model,

(10)

has an optimal solution with . Here, by solving r independent subproblems, we can obtain , where

(11)

The subproblems in Eq 11 can be effectively solved using GST. Given p and δ, a specific threshold exists:

(12)

For the subproblem defined in Eq 11, the optimal solution depends on the relationship between and the threshold . If , then the optimal solution of Eq 11 is 0. In contrast, if , the optimal solution of Eq 11 is and involves solving the following equation:

(13)

This condition captures a soft-thresholding operation on the singular values. If the singular value is above the threshold, the soft-thresholding function is applied, and the solution is obtained by solving the indicated equation. This process facilitates the efficient handling of cases where the singular values should be thresholded to achieve the optimal solution for the given problem. where is the i-frontal slice of , is the i-frontal slice of . Eq 14 can be decomposed into n3 () independent subproblems, and the i-th subproblem is as follows:

As expressed in Eq 7, the WTSNM problem is first transformed into the Fourier domain. The transformed problem can be represented as follows:

(14)

where and represent the i-th frontal slice of and , respectively ().

Eq 14 can be decomposed into n3 independent subproblems using the following formulation:

(15)

where . This decomposition enables the independent solution of each subproblem, simplifying the overall optimization process.

According to Lemma 1, , , the optimal solution of Eq 15 is , where , , and denote the i-th frontal slice of , , and , respectively. By traversing 1 to n3, we obtain the optimal solution of each subproblem. Then, according to T-SVD [13], we obtain the optimal solution of Eq 7 (WTSNM).

(16)

where and

Solving the optimization problem Eq 4 using this approach yields a low-rank tensor and sparse tensor . The former contains the global background data, while the latter contains anomaly data. The separation of anomalies from background is thereby achieved. Therefore, can be applied to the CRD algorithm to achieve background purification and improve detection performance.

is the spectral vector of a pixel in . serves as clean background data by eliminating noise and anomalous targets, making it suitable for collecting background data in the CRD algorithm. By using a local dual-window approach, we collect background data from , resulting in a set of selected samples denoted by , where and s represents the number of selected samples. The CRD algorithm formula is as follows:

(17)

and its solution is

(18)

where is

(19)

In CRD, anomalies can be determined from the following residual image:

(20)

Here, y is an anomalous pixel if r1 is above the threshold. We refer to this HAD method as ETRPCA-CRD.

Experiments

Datasets

The detection performance of ETRPCA-CRD was experimentally evaluated on three real hyperspectral datasets and one simulated dataset. All experiments were conducted on a system featuring an E5-2680v3 CPU, 32 GB of DDR4 memory, a 1-TB NVMe SSD, and an RTX 3060 graphics card with 12 GB of VRAM [20].

  • AVIRIS-I: The San Diego Naval Airport HSI, originally taken by the AVIRIS sensor, has an image size of pixels, a spatial resolution of 3.5 m per pixel, and 224 spectral bands covering the wavelength range of 370–2510 nm. We eliminated low-SNR, sensor fault, and water vapor absorption bands (1–6, 33–35, 97, 107–113, 153–166, 221–224), obtaining an image with 189 spectral bands. Finally, we prepared the AVIRIS-I dataset by capturing a subimage from the top left corner of the obtained image, containing numerous ground objects and exhibiting a complex distribution. Three airplanes of 58 pixels were used as anomalies for experimental purposes.
  • AVIRIS-II: A pixel dataset was selected from the central area of the San Diego Naval Airport image, containing various types of ground objects. Three airplanes of 134 pixels were selected as anomalous targets in the experiment. The target morphology was more complete, and the spectral difference between the target and background was distinct.
  • abu-urban-3: This dataset was derived from the urban scene of the ABU dataset [21] taken at Gainesville by the AVIRIS sensor. The image has a size of 100 × 100 pixels, a spatial resolution of 3.5 m per pixel, and 191 spectral bands. Various vehicles and other artificial targets were selected as anomalous targets.
  • Salinas-simulated: This simulated dataset was generated by embedding anomalous targets in the real Salinas dataset. The anomalies were six square regions duplicated in reverse order and arranged in another line at a close distance, and their spectral bands were generated using a linear mixing model.

Evaluation indicators

In this study, a color detection map (CDM) was employed as an intuitive indicator to qualitatively evaluate the detection performance of different HAD algorithms. The commonly used ROC curve and area under the curve (AUC) were selected as accuracy measures to compare the HAD algorithms. The more the ROC curve hugs the top left corner, the closer the AUC value is to 1, indicating a superior HAD algorithm [20].

Detection results

We compared ETRPCA-CRD with CRD [1], LRaSMD [7], LSMAD [8], RPCA [9], SSRX [22], TenB [23], and MFIFD [24]. In ETRPCA-CRD, the regularization parameter λ equals , and μ is a positive scalar. The compromising constant is C = 101/p when solving the WTSNM problem. The P value ranges from 0.5 to 0.95 with an interval of 0.5, and the optimal P is determined when achieving the maximum AUC in the experiment. ρ is set to 1.2 when updating μ [13]. The optimal window size for CRD was determined via grid search with ranges from to , by maximizing the AUC value. Table 1 presents the AUC values of ETRPCA-CRD and the seven other algorithms.

On the AVIRIS-I dataset, the optimal window size for CRD is determined to be (25, 23). The principal component numbers (K1, K2, and K3) in TenB are all set to 1. ETRPCA-CRD yields the best result at p = 0.2. Table 1 and Fig 1 present the AUC values. ETRPCA-CRD exhibits a significantly higher AUC value compared to that of the other algorithms, and the ranking of algorithms in terms of decreasing AUC is as follows: ETRPCA-CRD, CRD, SSRX, LSMAD, MFIFD, TenB, RPCA, and LRaSMD. Fig 1 presents the ROC curves of the eight algorithms, revealing that the ROC curves of the other algorithms are below that of ETRPCA-CRD. Fig 2 presents the CDMs of the compared methods, indicating that LRaSMD fails to detect anomalies in the experiment. The positions of the three airplanes are detected by LSMAD, RPCA, TenB, and SSRX, with blurred airplane shapes. CRD, RPCA, MFIFD, and TenB do not effectively separate anomalous and background pixels, resulting in poor anomaly detection. LSMAD and SSRX effectively suppress background pixels; however, they display dim abnormal targets. Compared with the other algorithms, ETRPCA-CRD accurately reconstructs the positions and shapes of the three airplanes and effectively distinguishes them from the background, exhibiting superior detection performance.

thumbnail
Fig 1. (a) AUC values and (b) ROC curves for the AVIRIS-I dataset.

https://doi.org/10.1371/journal.pone.0331894.g001

On the AVIRIS-II dataset, the optimal window size for CRD is determined to be (19, 17). K1, K2, and K3 in TenB are all set to 3. ETRPCA-CRD achieves the optimal result at p = 0.15. Fig 3 presents the ROC curves of the eight algorithms. Of all ROC curves, that of ETRPCA-CRD is closest to the top left corner. Fig 3 and Table 1 present the AUC values. The ranking of the algorithms in terms of decreasing AUC is as follows: ETRPCA-CRD, SSRX, CRD, LSMAD, MFIFD, RPCA, TenB, and LRaSMD. In comparison with the other methods, ETRPCA-CRD provides a significantly higher AUC value of 0.9834. This is because each singular value is assigned an individual weight in ETRPCA-CRD, which effectively preserves salient signals and eliminates noise. Fig 4 presents the CDMs of the different methods. It can be seen that MFIFD, RPCA, and TenB perform poorly in separating anomalous and background pixels. They correctly detect the positions of the three airplanes; however, the airplane shapes are unclear. LRaSMD and LSMAD accurately detect the positions of the three airplanes; however, one of them has an incomplete shape. CRD and SSRX perform well in background suppression; however, there are two airplanes with incomplete shapes. In contrast, ETRPCA-CRD determines the accurate positions and complete shapes of all three airplanes, thereby achieving the best performance.

thumbnail
Fig 3. (a) AUC values and (b) ROC curves for the AVIRIS-II dataset.

https://doi.org/10.1371/journal.pone.0331894.g003

On the abu-urban-3 dataset, the optimal window size for CRD is determined to be (23, 21). K1, K2, and K3 in TenB are set to 4, 1, and 1, respectively. ETRPCA-CRD achieves the optimal result at p = 0.15. As illustrated in Table 1 and Fig 5, ETRPCA-CRD, which has an AUC value of 0.9899, outperforms the other algorithms. The algorithms are ranked in terms of decreasing AUC as follows: ETRPCA-CRD, CRD, LSMAD, MFIFD, SSRX, RPCA, TenB, and LRaSMD. Fig 5 presents the ROC curves of the eight algorithms. Of all ROC curves, that of ETRPCA-CRD is closest to the top left corner. Fig 6 presents the CDMs of the different algorithms. It can be seen that LRaSMD, LSMAD, RPCA, TenB, and SSRX mistakenly classify numerous background pixels as anomalous, while CRD and MFIFD fail to effectively separate the background from the anomalous targets. Overall, the proposed algorithm exhibits effective background suppression, achieving the best detection performance among the compared algorithms.

thumbnail
Fig 5. (a) AUC values and (b) ROC curves for the abu-urban-3 dataset.

https://doi.org/10.1371/journal.pone.0331894.g005

On the Salinas-simulated dataset, the optimal window size for CRD is determined to be (19, 17). K1, K2, and K3 in TenB are all set to 3. ETRPCA-CRD achieves the best results at p = 0.15. As illustrated in Table 1 and Fig 7, ETRPCA-CRD surpasses the other algorithms in terms of its AUC value of 0.9999, which is almost 1. The ranking of the algorithms in terms of decreasing AUC is as follows: ETRPCA-CRD, MFIFD, LRaSMD, LSMAD, SSRX, RPCA, TenB, and CRD. Fig 7 presents the ROC curves of the eight algorithms. The ROC curve of ETRPCA-CRD overlaps with the top left corner, while the curves of the other algorithms are below it. Fig 8 displays the CDMs of the different algorithms. It can be seen that CRD fails to detect anomalies. LRaSMD, LSMAD, RPCA, TenB, MFIFD, and SSRX can detect the positions and shapes of all abnormal targets; however, they perform poorly in separating anomalous pixels from background pixels. In contrast, ETRPCA-CRD successfully detects the positions and shapes of all anomalous targets, which appear prominently bright.

thumbnail
Fig 7. (a) AUC values and (b) ROC curves for the Salinas-simulated dataset.

https://doi.org/10.1371/journal.pone.0331894.g007

thumbnail
Fig 8. Color detection maps for Salinas-simulated dataset.

https://doi.org/10.1371/journal.pone.0331894.g008

thumbnail
Fig 9. Grid search results for p value on AVIRIS-I dataset.

https://doi.org/10.1371/journal.pone.0331894.g009

Parameter analysis

  1. 1) Analysis of p: The efficacy of the proposed method exhibits significant dependence on two critical hyperparameters: the dual-window configuration in contextual region detection (CRD) and the Schatten p-norm coefficient in weighted tensor Schatten p-norm minimization (WTSNM). To determine optimal parameterization, a systematic grid search protocol was executed. The p-value space spanning [0.05, 0.95] was discretized into 19 equidistant points (step size 0.05). The parameter analysis for the p value across four datasets is shown in Figs 9 through 12, revealing distinct trends in model performance as measured by AUC.
  2. The AVIRIS-I dataset presents the most stable performance profile. After reaching a maximum AUC of 0.9990 at p = 0.15, the model maintains AUC values above 0.9950 across nearly the entire p value spectrum from 0.10 to 0.95, with only minimal variation observed.
    On the AVIRIS-II dataset, the AUC exhibits a near-linear increase from 0.7026 at p = 0.05 to a peak of 0.9846 at p = 0.10. Beyond this point, the AUC stabilizes, maintaining values above 0.9820 across p values from 0.15 to 0.95, indicating robust performance across this range. For the abu-urban-3 dataset, the AUC rises dramatically from 0.7663 at p = 0.05 to 0.9979 at p = 0.15. Performance then exhibits slight fluctuation around 0.9850 between p = 0.20 and p = 0.95, with minor deviations of ±0.0050 from this mean value.
    The Salinas-simulated dataset demonstrates a similar pattern but with higher absolute AUC values. Here, the maximum AUC of 0.99993 is achieved at p = 0.15. However, unlike AVIRIS-II, this dataset shows a plateau of near-maximal performance spanning p = 0.10 to p = 0.25, with AUC values exceeding 0.9950 in this interval.
    Collectively, these results indicate that optimal p values typically fall between 0.10 and 0.20 across datasets, with performance plateaus suggesting flexibility in parameter selection while maintaining high predictive accuracy. The parameter tuning process confirms the importance of dataset-specific optimization within this critical p value range.
  3. 2) Analysis of dual-window size: For the dual-window size, configurations ranging from 11×13 to 33×35 were tested, with the configuration yielding the highest AUC selected as optimal. The grid search results are presented in Tables 2, 3, 4, and 5.
  4. Table 2 illustrates the grid search results for the AVIRIS-I dataset. The optimal dual-window configuration is identified as 23×25, achieving an AUC of 0.9990. For the AVIRIS-II dataset (Table 3), the 17×19 configuration provides the highest AUC of 0.9846. The abu-urban-3 dataset (Table 4) shows that the best performance is obtained with a dual-window size of 21×23, resulting in an AUC of 0.9979. For the Salinas-simulated dataset (Table 5), the optimal dual-window size is 17×19, with an AUC of 0.9993.
    These results indicate that optimal dual-window configurations vary across different datasets, underscoring the importance of dataset-specific parameter tuning. The grid search approach ensures that the selected dual-window sizes maximize the model’s performance, thereby contributing to the overall effectiveness of the proposed method.
thumbnail
Table 2. Grid search results of dual-window parameters on AVIRIS-I dataset.

https://doi.org/10.1371/journal.pone.0331894.t002

thumbnail
Table 3. Grid search results of dual-window parameters on AVIRIS-II dataset.

https://doi.org/10.1371/journal.pone.0331894.t003

thumbnail
Table 4. Grid search results of dual-window parameters on abu-urban-3 dataset.

https://doi.org/10.1371/journal.pone.0331894.t004

thumbnail
Table 5. Grid search results of dual-window parameters on Salinas-simulated dataset.

https://doi.org/10.1371/journal.pone.0331894.t005

thumbnail
Fig 10. Grid search results for p value on AVIRIS-II dataset.

https://doi.org/10.1371/journal.pone.0331894.g010

thumbnail
Fig 11. Grid search results for p value on abu-urban-3 dataset.

https://doi.org/10.1371/journal.pone.0331894.g011

thumbnail
Fig 12. Grid search results for p value on Salinas-simulated dataset.

https://doi.org/10.1371/journal.pone.0331894.g012

Computational complexity

The computational complexity of the proposed ETRPCA-CRD algorithm is a critical aspect for practical implementation. The algorithm consists of two main components: the enhanced tensor robust principal component analysis (ETRPCA) and the collaborative representation detection (CRD).

In ETRPCA, the tensor is transformed into the frequency domain via FFT, decomposing the problem into n3 independent frontal slices. This allows parallel processing of each slice, reducing the complexity from coupled multi-dimensional operations to independent matrix problems. The FFT/IFFT operations exhibit a complexity of: . This linear scaling with n3 (spectral bands) is achieved through the efficient FFT algorithm, which avoids the exponential complexity of direct tensor operations.

T-SVD leverages the block-diagonal structure of the frequency-domain tensor to decompose each frontal slice into orthogonal matrices and a diagonal singular value matrix. For each slice, the SVD complexity is: . With n3 slices processed independently, the total SVD complexity becomes: . This avoids the cubic complexity of traditional tensor decompositions by converting the problem into parallel matrix operations.

GST applies adaptive thresholds to singular values, prioritizing large values and suppressing small ones. For each slice, the GST operation involves element-wise soft-thresholding with complexity: . Across all slices, this sums to: By avoiding uniform regularization of all singular values, GST reduces unnecessary computations and accelerates convergence.

the CRD step, the algorithm involves solving a least squares problem for each pixel in the image. For an image of size , the complexity of solving the least squares problem for each pixel is , assuming the number of selected background samples is n3. Therefore, the total complexity of CRD is . Combining these, the overall computational complexity of the ETRPCA-CRD algorithm is dominated by the tensor decomposition step, which is .

Notably, the CRD component in the ETRPCA-CRD algorithm poses a limitation. Its double-window mechanism for solving least squares problems at each pixel results in a high computational complexity of . This can slow down processing, especially for high-dimensional data or large images, potentially limiting the algorithm’s performance in time-sensitive or resource-constrained scenarios.

Conclusions

In conclusion, this paper proposes the ETRPCA-CRD method for hyperspectral anomaly detection, which integrates enhanced tensor robust principal component analysis and collaborative representation to effectively leverage spectral-spatial information. By decomposing hyperspectral data into low-rank and sparse tensors via weighted tensor Schatten-p norm minimization, the approach achieves robust background purification and improved anomaly detection accuracy. Experimental results on multiple datasets validate its superiority over state-of-the-art methods in terms of detection performance and reliability.

Notably, the current implementation involves two areas for potential refinement. The dual-window size in the CRD component and the p value in weighted tensor Schatten-p norm minimization are currently determined via rid search to optimize detection results, which may require dataset-specific adjustments . Additionally, while the CRD mechanism ensures precise anomaly identification, its computational complexity—stemming from the dual-window least squares solving—presents opportunities for efficiency improvements, especially in high-dimensional data scenarios. Future research will focus on developing automated parameter selection frameworks to reduce dependency on rid search, such as adaptive algorithms that dynamically tune dual-window sizes and p values based on data characteristics. Furthermore, exploring parallel computing architectures and approximate solution strategies for the CRD component will be prioritized to enhance computational efficiency while maintaining detection accuracy, enabling broader applications in real-time hyperspectral monitoring systems.

References

  1. 1. Li W, Du Q. Collaborative representation for hyperspectral anomaly detection. IEEE Trans Geosci Remote Sensing. 2015;53(3):1463–74.
  2. 2. Su H, Wu Z, Du Q, Du P. Hyperspectral anomaly detection using collaborative representation with outlier removal. IEEE J Sel Top Appl Earth Observations Remote Sensing. 2018;11(12):5029–38.
  3. 3. Yin C, Gao L, Wang M, Liu A. Selective search collaborative representation for hyperspectral anomaly detection. Applied Sciences. 2022;12(23):12015.
  4. 4. Wang X, Wang L, Wang J, Sun K, Wang Q. Hyperspectral anomaly detection via background purification and spatial difference enhancement. IEEE Geoscience and Remote Sensing Letters. 2022;19:1–5.
  5. 5. Ma N, Liu Q, Wu W, He Q, Liu L, Peng Y. A weighted collaborated representation anomaly detector for hyperspectral image. In: 2021 IEEE 15th International Conference on Electronic Measurement & Instruments (ICEMI). 2021. p. 43–8.
  6. 6. Zhu D, Du B, Zhang L. Binary-class collaborative representation for target detection in hyperspectral images. IEEE Geosci Remote Sensing Lett. 2019;16(7):1100–4.
  7. 7. Sun W, Liu C, Li J, Lai YM, Li W. Low-rank and sparse matrix decomposition- based anomaly detection for hyperspectral imagery. Journal of Applied Remote Sensing. 2014;8(1):083641.
  8. 8. Zhang Y, Du B, Zhang L, Wang S. A low-rank and sparse matrix decomposition-based mahalanobis distance method for hyperspectral anomaly detection. IEEE Trans Geosci Remote Sensing. 2016;54(3):1376–89.
  9. 9. Candès EJ, Li X, Ma Y, Wright J. Robust principal component analysis?. J ACM. 2011;58(3):1–37.
  10. 10. Oh T-H, Tai Y-W, Bazin J-C, Kim H, Kweon IS. Partial sum minimization of singular values in robust PCA: algorithm and applications. IEEE Trans Pattern Anal Mach Intell. 2016;38(4):744–58. pmid:26353362
  11. 11. Hu Y, Zhang D, Ye J, Li X, He X. Fast and accurate matrix completion via truncated nuclear norm regularization. IEEE Trans Pattern Anal Mach Intell. 2013;35(9):2117–30. pmid:23868774
  12. 12. Gu S, Xie Q, Meng D, Zuo W, Feng X, Zhang L. Weighted nuclear norm minimization and its applications to low level vision. International Journal of Computer Vision. 2017;121(2):183–208.
  13. 13. Kilmer ME, Martin CD. Factorization strategies for third-order tensors. Linear Algebra and its Applications. 2011;435(3):641–58.
  14. 14. Guo W, Che H, Leung M-F. Tensor-based adaptive consensus graph learning for multi-view clustering. IEEE Trans Consumer Electron. 2024;70(2):4767–84.
  15. 15. Gao Q, Zhang P, Xia W, Xie D, Gao X, Tao D. Enhanced tensor RPCA and its application. IEEE Trans Pattern Anal Mach Intell. 2021;43(6):2133–40. pmid:32809937
  16. 16. A R, Mu X, He J. Enhance tensor RPCA-based mahalanobis distance method for hyperspectral anomaly detection. IEEE Geosci Remote Sensing Lett. 2022;19:1–5.
  17. 17. A R, Mu X, He J, Zhang J, et al. Enhance tensor RPCA-LRX anomaly detection algorithm for hyperspectral image. Geocarto International. 2022;37(26):11976–97.
  18. 18. Xie Y, Gu S, Liu Y, Zuo W, Zhang W, Zhang L. Weighted schatten p-norm minimization for image denoising and background subtraction. IEEE Trans on Image Process. 2016;25(10):4842–57.
  19. 19. Yang X, Che H, Leung M-F. Tensor-based unsupervised feature selection for error-robust handling of unbalanced incomplete multi-view data. Information Fusion. 2025;114:102693.
  20. 20. A R, Mu X, Feng L, He J. A fast recursive LRX algorithm with extended morphology profile for hyperspectral anomaly detection. Canadian Journal of Remote Sensing. 2021;47(5):731–48.
  21. 21. Kang X, Zhang X, Li S, Li K, Li J, Benediktsson JA. Hyperspectral anomaly detection with attribute and edge-preserving filters. IEEE Trans Geosci Remote Sensing. 2017;55(10):5600–11.
  22. 22. Schaum A. Joint subspace detection of hyperspectral targets. In: 2004 IEEE Aerospace Conference Proceedings, 2004. p. 1824.
  23. 23. Xing Z, Wen G, Wei D. A tensor decomposition-based anomaly detection algorithm for hyperspectral image. IEEE Transactions on Geoscience & Remote Sensing. 2016;54(10):5801–20.
  24. 24. Wang R, Nie F, Wang Z, He F, Li X. Multiple features and isolation forest-based fast anomaly detector for hyperspectral imagery. IEEE Trans Geosci Remote Sensing. 2020;58(9):6664–76.