Figures
Abstract
Multi-scale image decomposition is crucial for image fusion, extracting prominent feature textures from infrared and visible light images to obtain clear fused images with more textures. This paper proposes a fusion method of infrared and visible light images based on spatial domain and image features to obtain high-resolution and texture-rich images. First, an efficient hierarchical image clustering algorithm based on superpixel fast pixel clustering directly performs multi-scale decomposition of each source image in the spatial domain and obtains high-frequency, medium-frequency, and low-frequency layers to extract the maximum and minimum values of each source image combined images. Then, using the attribute parameters of each layer as fusion weights, high-definition fusion images are through adaptive feature fusion. Besides, the proposed algorithm performs multi-scale decomposition of the image in the spatial frequency domain to solve the information loss problem caused by the conversion process between the spatial frequency and frequency domains in the traditional extraction of image features in the frequency domain. Eight image quality indicators are compared with other fusion algorithms. Experimental results show that this method outperforms other comparative methods in both subjective and objective measures. Furthermore, the algorithm has high definition and rich textures.
Citation: Zhao L, Zhang Y, Dong L, Zheng F (2022) Infrared and visible image fusion algorithm based on spatial domain and image features. PLoS ONE 17(12): e0278055. https://doi.org/10.1371/journal.pone.0278055
Editor: Ashwani Kumar, Sant Longowal Institute of Engineering and Technology, INDIA
Received: August 2, 2022; Accepted: September 30, 2022; Published: December 30, 2022
Copyright: © 2022 Zhao et al. This is an open access article distributed under the terms of the Creative Commons Attribution License, which permits unrestricted use, distribution, and reproduction in any medium, provided the original author and source are credited.
Data Availability: All image files are from TNO_Image_Fusion_Datasete and can be obtained from Figshare (DOI: 10.6084/m9.figshare.1008029.v2).
Funding: Our study is supported by: 1.Development of an intelligent management platform for grassland resources, XJXMKXY_2021 2.Grassland Resource Degradation Monitoring and Evaluation, XJXMKXY_202102 3.Major Research plan of the National Natural Science Foundation of China(Grant No.31860679) 4. Wuliangye Fund (No. CXY2020R001) The funders had no role in study design, data collection and analysis, decision to publish, or preparation of the manuscript.
Competing interests: The authors have declared that no competing interests exist.
Introduction
Although a huge amount of sensor data has been collected with the advancement of sensor technology, single sensor data can provide limited information about the scene. Thus, multiple sensors have recently been combined to obtain more comprehensive and accurate scene information. However, the different characteristics of different sensors cause some challenges in subsequent processes. Also, multiple sensors provide redundancy in the collection of information, resulting in low utilization of transmission bandwidth and storage space. In order to overcome these problems, image fusion technology has been researched in many fields. Especially, the fusion of the visible-light image and infrared image has attracted attention due to its practical value in the military [1, 2] and security applications [3]. Infrared images are sensitive to the heat source targets in the scene, enabling users to quickly grasp the target location information. However, the infrared sensor cannot detect cold or heat-equilibrium objects in practical applications, losing much useful information [4]. In contrast, the visible-light sensor can provide an image with rich scene texture, but the collected image is susceptible to the influence of light intensity and smoke occlusion. It often makes the target information difficult to be quickly recognized by the user [5]. Therefore, the fusion of infrared image and visible-light image allows combining the advantages of each source image into one image.
Image fusion algorithms are divided into three groups according to the fusion level: pixel level, feature level, and decision level fusions [6]. Further, a hybrid level fusion has been explored to combine different levels of fusion methods. In the past few decades, many classic image fusion algorithms have emerged, such as deep learning-based methods [7, 8], neural network-based methods [1], sparse representation-based methods [2], and subspace-based methods [3]. However, those methods provided insufficient image anti-interference ability with a large amount of information loss in the infrared images. Multi-scale decomposition [6] was proposed to zoom the image and superimpose multiple layers of spatial feature information for the effective fusion of infrared images and visible-light images, based on multi-scale transform, including pyramid transform [7], Wavelet transform [8], curvelet transform [9], contourlet transform [10], and non-subsampled contourlet transform [11]. However, regardless of the effectiveness of multi-scale decomposition, the fused images often suffered from halos artifacts and other interference factors [12]. In addition, determining and decomposing the number of layers in the image space is a challenge. The larger the number of decomposition layers, the richer the texture details of the fused image. However, the middle- and low-frequency layer coefficients affect most fused pixel values in the fused image as the number of image decomposition layers increases. The fusion effect and the robustness of the algorithm are a trade-off. Accordingly, a single-level fusion approach is difficult to deal with complex and diverse image fusion problems.
With the development of machine learning technology methods such as neural networks, scholars introduced support vector machines and genetic algorithms into image processing [13–21] to solve the defects of traditional frequency-domain feature extraction technology. Among them, the neural network method is more common in image fusion, which includes three types. The first is based on pre-training the network. This network model allows the machine to extract images adaptively through many pre-training models. However, this method requires large-scale training samples. The extracted features cannot meet the application requirements under insufficient samples. At the same time, the neural network models are designed for classifiers and may not be suitable for image fusion tasks. In the second type proposed by Li et al., the self-encoding network was introduced into image fusion. The network comprises an encoder and a decoder. The encoder can extract the features of the image to be fused, and the decoder generates the fusion image, which solves the tuning requirement problem. However, both the encoder and the decoder should be designed for the corresponding fusion task, and the algorithm has weak robustness. In the third type, many scholars proposed neural networks to solve the defects of fused images. Jiayi Ma proposed FusinGAN [22]. With a well-designed network structure, good fusion results can be achieved without hand-crafted strategies. However, a vast amount of data specified to the target fusion tasks are required, which often cannot perform well for the different data. For example, the fusion model trained with bright images cannot work well for low illumination images. Hui Li [23] proposed a novel image fusion framework based on MDLatLRR, where the source image was decomposed into detail parts (salient features) and base parts. Then, those detail and base parts were fused by the nuclear-norm-based fusion strategy. This method effectively solved the artifacts and halo that happened in the traditional multi-scale fusion method (Fig 1). Still, the integral definition of the fused image remains to be optimized. The MDLatLRR is similar to the anisotropic diffusion-based method. Both methods decomposed the infrared and visible light images into the basic and detail layers, although the decomposition algorithm and fusion rules are different. They can prevent artifacts generated in the fusion process, but both suffer poor image definition.
(a) IR, (b) Vis, (c) GTF, (d) FusionGAN, (e) MDLatLRR, (f) Our.
In order to overcome the above-mentioned issues, this paper proposes an infrared and visible image fusion method based on spatial domain and image features. In the method, an efficient multi-scale decomposition is adopted to smooth the source image, which can effectively extract the intermediate frequency layers of the infrared image and the visible-light image. Unlike strong decomposition algorithms such as the anisotropic diffusion algorithm [24], the proposed multi-scale decomposition rather causes large edge gradients between different pixel groups of the intermediate frequency layer. Thus, the mean filter is applied to obtain a better intermediate frequency layer. Also, the high-frequency and low-frequency layers are extracted, and then the large layer and the small layer of each source image are extracted. The final fused image is obtained through the adaptive weight fusion according to the feature parameters of each source image and layer pixel.
To highlight the major advantage of our method, we illustrate a representative example in Fig 1. The infrared and visible images are fused, where the visible image contains detailed background and the infrared image highlights the target, i.e., the pedestrian. Fig 1C and 1D show the fused images by the traditional method (GTF) and advanced method (FusionGAN), respectively. Although it highlights the target location information, the clarity of the environmental texture display is not significantly improved. The fused images by MDLatLRR (Fig 1E) can reflect the information of the infrared and visible images, but it is still blurred. In contrast, our method (Fig 1F) can provide a fused image where environment texture and target information can be easily recognized.
The experimental results show that the proposed fusion method provides a higher definition and more comprehensive information than the other fusion algorithms. The contributions of this paper are as follows. First, we propose an image clustering algorithm for image fusion, greatly improving the quality of fused images. Second, different from the traditional multi-scale decomposition strategy, we extract one more dark detail layer on the basis of predecessors, which makes the texture gradient of the fusion result easier to reflect the information carried by the image. As a result, the clarity of the fused image is improved. Third, we propose a fusion method based on the principle of expanding texture and the layer eigenvalue as the weight. This makes the weight-average strategy more effective, addressing the poor texture definition of existing fused images.
The remainder of this paper is structured as follows. Section 2 describes the proposed method in detail. Section 3 introduces the parameter settings of the proposed fusion theory. Section 4 provides experimental results, and Section 5 concludes this paper.
Proposed method
This section describes the proposed method in detail, followed by the analysis of the parameter setting. The proposed method first decomposes images in multi-scale by using the superpixel-based fast fuzzy C-means. Then, the high-frequency layer and low-frequency layer of the source image are extracted. Lastly, those layers are combined with a fusion strategy to obtain the final fused image.
A. Mathematical model
The multi-scale decomposition for image fusion requires the following two conditions. 1) it can reasonably and effectively decompose the source image information in multi-scales, and 2) the computational time should be fast. Accordingly, the multi-scale decomposition method is designed to achieve multi-scale decomposition with a low computational time. We propose the superpixel-based fast fuzzy C-means image hierarchical multi-scale decomposition method. The decomposition effect of the source image is shown in Fig 2.
Example of the multi-scale decomposition (a) source image, (b) the proposed multi-scale decomposition.
Fig 2 shows the schematic diagram of the proposed multi-scale decomposition method to extract the intermediate frequency layer. In the intermediate frequency layer, the clustered areas of the source image are effectively retained, and the texture gradient between the segments is weakened. However, the clustering results (Fig 2B) still have a lot of texture information. Therefore, we need to improve the algorithm to fully extract texture information.
In order to evaluate the image clustering ability of the proposed clustering algorithm, the image clustering algorithm in [24] is employed for comparison, and the resulting surf graph is shown in Fig 3.
The first row includes, from left to right: infrared image, ADF clustering results, and the proposed algorithm clustering results; the second row includes, from left to right: infrared image partial area surf map, ADF clustering partial area surf map, and the proposed clustering algorithm partial area surf map. The abscissa of the Surf map represents the spatial position of the pixel, and the ordinate represents the size of the pixel value.
As shown in Fig 3, the proposed clustering algorithm has a higher degree of image blurring than ADF. Combined with the surf map observation, the surf map corresponding to the infrared image contains many texture gradients. Although the surf map of ADF smooths a lot of details texture, there is still a lot of nonsmoothed information compared with our surf map, which will lead to multi-scale decomposition. Accordingly, our fusion results are more precise since part of the information in the feature layer cannot be extracted.
1) Mathematical model of the proposed multi-scale decomposition.
According to the above multi-scale decomposition tool conditions, the existing image clustering algorithms, such as the C-means clustering algorithm and K-means clustering algorithm, are analyzed while developing the multi-scale decomposition tool. The iterative calculation of distances between pixels increases the computational cost of the clustering algorithm. At the same time, the local features are destroyed due to the fixed dimension of the clustering window, making the clustering result unfavorable for subsequent multi-scale decomposition. Therefore, image clustering is implemented by adopting a filter window based on c-means with better adaptive and irregular local space provided by superpixels, which can effectively solve the problems of high computational time cost and local structure damage during clustering.
The objective function of the proposed multi-scale decomposition is defined as follows. The determination of the multipliers is given by [25]:
(1)
where l represents the gray intensity levels, q is the number of superpixels, and 1≤l≤q.
Sl represents the number of pixels in the lth region Rl, and c indicates the number of image clusters. ukl is the membership matrix of the area of the lth superpixel, and the uk represents the center of the kth cluster. xp represents a pixel value in the image. c is the number of clusters.
The optimization problem of Formula (1) can be transformed into an unconstrained problem using a Lagrangian multiplier, thereby minimizing the objective function and satisfying:
(2)
where λ is the Lagrangian multiplier. The ukl and uk partial differential equation in Eq (2) satisfies:
(3)
(4)
Combining the Formulas (3) and (4), ukl and uk are computed as follows:
(5)
(6)
The process of decomposition is conducted as follows:
- (A) The same pixels of the source image are combined according to the function Jm, denoted as I (x).
- (B) Morphological expansion processing is performed. I (x) and b represent the morphological window dimension and the window dimension value, respectively. The result after the expansion is denoted as I1 (x), where I1(x) = I(x)⊕b (Fig 6).
- (C) Morphological corrosion treatment is performed. The result after the expansion is denoted by I2 (x), where I2 (x) = I1(x) ⊝b. Therefore, we have:
(7)
- (D) The processed layers are composed into an image I2 (x), followed by the mean filtering with a m×m window to obtain the final rsults, denoted as IM.
2) Extraction of high-frequency and low-frequency layers.
The multi-scale decomposed image IM is obtained with the window Wn×n where n = 150 as the intermediate frequency layer. The high-frequency layer IH is obtained by subtracting the intermediate frequency layer IM from the source image IS:
(8)
The low-frequency layer image IL is obtained by subtracting the source image IE from the intermediate frequency layer IM:
(9)
The multi-scale decomposition results of the source image are shown in Fig 4.
I(ir) represents the infrared image, I(vis) represents the visible image, and the yellow cube represents the proposed smoothing method. The processed infrared and visible images are denoted as IM(ir) and IM(vis), respectively. Through Eq (8), the visible image is decomposed into bright detail layer IH(vis) and dark detail layer IL(vis), while, through Eq (9), the infrared image is decomposed into bright detail layer IH(ir) and dark detail layer IL(ir).
B. Fusion strategy
The fusion process consists of two parts:
1) intermediate-frequency layer fusion and 2) high-frequency and low-frequency layers fusion. First, in the fusion process of the intermediate-frequency layer, the pixels with the larger value between visible-light and infrared images are selected for the maximum layer. The pixel with the smaller value is selected at the corresponding coordinates of the two images to form the minimum layer.
Then, the fused image is computed as follows:
(10)
where IZ indicates the fusion result of the frequency images in each source image. IM1 and IM2 represent the intermediate frequency layers of the visible-light and infrared images. max(IM1,IM2) represents the pixel standard deviation corresponding to the extremely large layer of the intermediate frequency image expressed by σ1, and min(IM1,IM2) represents the pixel standard deviation corresponding to the extremely small layer of the intermediate frequency image expressed by σ2.
The existing multi-scale algorithm converts the image from the spatial domain into the frequency domain and extracts the image features in the frequency domain using the corresponding fusion strategy, such as orthogonal change, sparse representation, and other methods. In order to avoid information loss and increase the amount of computation caused by the transformation between the frequency and spatial domains, the image features in the spatial domain are directly extracted, and the multi-scale decomposition method is applied to linear addition and subtraction. The fusion strategy also linearly reconstructs the fused image. The target is not outlined, although IZ retained the basic texture of the source image. In order to compensate for this limitation, the maximum mixed image max(IS1,IS2) and the minimum mixed image min(IS1,IS2) of the source image are used. IS1 and IS2 represent the visible-light and infrared source images, respectively.
max(IE, II) and min(IE, II) are respectively linearly fused with the intermediate frequency layer IM with a weight of 0.1. The fused image IZF of the new intermediate frequency layer after the texture enhancement is formulated as follows:
(11)
The formulation process of IZ at the intermediate frequency layer is shown in Fig 5.
In the second part of the fusion process, the high-frequency and low-frequency layers are fused based on the fused intermediate layer IZF under the principle of texture gradient expansion. It is defined as follows:
(12)
Where IF represents the fused image. IH1 and IL1 are the high-frequency layer and low-frequency layer of the visible-light image, respectively. IH2 and IL2 are the infrared image’s high-frequency and low-frequency layers, respectively. Inspired by the existing linear fusion algorithm, the total weight of linear fusion is 1, and the fusion weight of each base layer is 0.5, such as the basic fusion strategy of the ADF algorithm. We believe that the fixed fusion weight cannot provide strong robustness to the algorithm. Therefore, since the total weight is 1 and the fusion weight of each layer is 0.5, the weights are adjusted appropriately according to the feature ratio of the layers to be fused to improve the robustness of the fusion algorithm. ω1 and ω2 respectively represent the fusion weights controlling the relative significance of the visible-light and infrared high-frequency layers, defined as follows:
(13)
(14)
(15)
where W indicates the sum of the weights of the high-frequency detail layer fusion of each source image. σ3, σ4, σ5, and σ6 represent the standard deviations of visible-light image IS1, infrared image IS2, the visible-light intermediate frequency layer IM1, and the visible-light intermediate frequency layer IM2, respectively.
In combination with intermediate frequency layer IZ, high- and low-frequency layers of infrared ad visible light layers are fused into the final fusion result IF, as shown in Fig 6.
Parameter settings
The proposed algorithm includes several parameters required to be set. In this Section, the influence of the parameters on the results is analyzed to determine the optimal settings.
A. The dimension of the clustering window
The clustering results IC are varied according to the clustering window dimension n, affecting the frequency layers extraction and, consequently, the image fusion results. Fig 7 depicts the fusion results according to the clustering window dimension n. As shown in Fig 7, when n is 0, many scattered artifacts appear in the fusion result. As the dimension n increases, such artifacts gradually gather to form a concentrated artifact area. Also, the overall brightness and sharpness of the image gradually increase until n = 150, while most artifacts are eliminated. When n = 200, the effect of the "smoke" image is almost the same as that of n = 150, and the artifacts of the "military car" become more severe. Accordingly, the window dimension n is set to 150.
(a) n = 0, (b) n = 50, (c) n = 100, (d) n = 150, (e) n = 200.
B. The dimension of the mean filter window
In this Section, the mean filtering result, applied to remove the artifacts in IC, is analyzed according to the window dimension m. The fusion results according to a varied dimension m are shown in Fig 8 as the dimension of the mean filter window Wm×m increases, the artifacts are removed more. When m = 11, the obtained fusion result is almost the same as in the ’smoke’ image when m = 11. There is a slight difference in the ’Military Vehicles’, and the fusion effect when m = 11 is better than when m = 9. Thus, the window dimension m is set to 11.
(a)m = 3, (b) m = 5, (c) m = 7, (d) m = 9, (e) m = 11.
The flow chart of the proposed method is given in Table 1.
It is worth noting that the most time-consuming part of the proposed algorithm is step 2: the image multi-scale decomposition. The diagram of the proposed algorithm is depicted in Fig 9.
The computational complexity of the proposed method mainly includes the following categories:
- Multi-scale decomposition includes image clustering and image smoothing. Therefore, its computational complexity is O(n2).
- The computational complexity of image reconstruction for each layer is O(n)
- The final complexity T(n) of the proposed method is expressed as follows:
Simulation results
The performance of the proposed method is validated in terms of both objective and subjective evaluations, compared with the comparison algorithms, including Nestfuse [26], FusionGAN [27], MDLatLRR [28], SEDRFuse [29], STDFusionNet, GANMcC [22], and ResNetFusion [30]. All experiments were conducted on Windows 10, 2.60GHz CPU, 8GB RAM, with MATLAB2016a. The experimental data were obtained from: TNO_Image_Fusion_Dataset(https://figshare.com/articles/dataset/TNO_Image_Fusion_Dataset/1008029).
A. Evaluation metrics
The objective evaluation indexes include average gradient (AG) [31], information entropy (H) [32], standard deviation (SD) [33], spatial frequency (SF) [34], edge strength (EI) [35], fusion function (Qab/f) [36], the amount of artifact (Nab/f) [37], and the fusion loss function (Lab/f) [33]. The larger the evaluation value of AG, H, SD, SF, and EI, the better the image quality, the larger the Qab/f value indicates that the fusion image contains more information of the source image, and the smaller the Nab/f value indicates that the fusion image is produced, the fewer artifacts. The smaller the Lab/f value, the smaller the loss of source image information during the fusion process.
B. Evaluation of fusion performance
1) Dataset.
The performance of the compared methods was evaluated on the surveillance images from TNO Human Factors. The dataset includes registered multispectral night-time imagery of different military-relevant scenarios. We selected seven typical pairs for qualitative illustration: Two men in front of the house, Soldierbehindsmoke_3, Soldierintrench_1, Houseswith3men, Kaptein_1123, Marne_04, and Sandpath. In addition, we tested our method on the INO database, which is provided by the National Optics Institute of Canada and contains several pairs of visible and infrared videos representing different scenarios captured under various weather conditions. Specifically, we grabbed 21 visible and infrared image pairs from the video named Trees and runner for qualitative and quantitative comparisons.
2) Results of the TNO dataset.
Eight typical image pairs from the TNO dataset were used to qualitatively evaluate the performance of the proposed method and the compared seven methods, as shown in Fig 10.
(The first row: Infrared images; The second row: Visible images; The third row: FusionGAN; The fourth row: GANMcC; The fifth row: MDLatLRR; The sixth row: NestFuse; The seventh row: RESNetFusion; The eighth row: SEDRFuse; The ninth row: STDFusionNet; The tenth row: OUR).
As shown in Figs 10–12, all the methods provide comparable fusion results with respective advantages. In overall quality, FusionGAN, GANMcC, and ResNetFusion generate more infrared-like images, taking advantage of a significant target retaining heat source but losing the background texture information. Although the MDLatLRR algorithm can keep the heat source target and background texture information at the same time, the overall fusion result is fuzzy, the significance of the heat source target is lost, and the target is not easily found. NestFuse, SEDRFuse, STDFusionNet, and our method can clearly reflect the critical information in infrared and visible images. In combination with the visible light evaluation analysis, our results of AG, SD, and EI evaluation indexes ranked first in all the images except for the fifth image (third rank). In terms of the H and SF indices, our method ranked second, reflecting the change degree between image pixels and the sharpness of the image. From the perspective of the evaluation indicators, the sharpness of this paper is superior to the other compared methods in most of the evaluation indicators. To reflect the fusion image consisting of infrared and visible light image information of each evaluation function Qab\f, the mean of the evaluation results of our algorithm is 0.3505, which is at the median level, showing that our results fusion of infrared and visible light information is not the most. The reason is that the infrared image contains a lot of interference information, such as heat-source target (e.g., humans), which makes the background saturated. Also, over-exposed backgrounds in visible light images influence the observation of the background texture. The proposed method can adaptively compensate for this part of the interference, obtaining a clearly fused image. From the loss function Lab/f and artifact function Nab/f, it can be proved that although some information is lost in our algorithm, both of them are loss interference information. Therefore, in Lab/f evaluation, our results can provide the best performance, and Nab/f evaluation value is the largest. It also proves that our algorithm considers the average of infrared and visible image information. When some parts where heat-source target highlights, our fusion results can also be observed that the texture of the goals of the heat source form, Nab/f large source of value, i.e., the cause of the picture without the part information. Our algorithm uses adaptive weights to jointly display the information of infrared and visible images in this area in order to make full use of the fusion advantages; thus, the average Nab/f is too large.
The compared methods were also evaluated on different weather conditions of images from the INO dataset. The videos, Runner and Tree, were split into 21 image frames, which were used for quantitative and qualitative evaluations. Figs 13 and 14 present the qualitative and quantitative comparisons, respectively, showing that all the eight compared algorithms can preserve texture information well. However, FusionGAN and ResNetFusion generated infrared-lie results, while GANMcC generated a relatively fuzzy result. NestFuse, SEDRFUSE, and STDFusionNet can retain visible light texture and infrared heat source targets. Our algorithm can retain not only infrared information but also background texture information with better sharpness. As quantitative evaluations show, our algorithm also can provide high robustness, clarity, and balanced information inclusion between infrared and visible images.
(a) the infrared and visible images.(b) the fusion results of FusionGAN, GANMcC, MDLatLRR, and NestFuse. (c) the fusion results of ResNetFusion, SEDRFuse, STDFusionNet, and OUR.
It should be noted that in practical engineering, the operation time of the algorithm is critical. Since the proposed algorithm operates pixel-wise in the spatial domain, the computational complexity is far lower than other frequency-domain algorithms. Table 2 compares the computational time for each dataset.
All the experiments were performed on the CPU. (unit: s).
Conclusion
This study employs the clustering algorithm to decompose the source image into the spatial domain directly. The results indicate that the multi-scale image decomposition into the spatial domain can extract more image features, eliminating the requirements while converting from the frequency domain to the spatial domain. As a result, the proposed algorithm consumes less time than the contrasting algorithm, and generates fusion results with higher clarity. Although the image features of the layers are employed as a reference for the fusion weights in the fusion process of the extracted layers, from a global perspective, these weights are still relatively rough. At the same time, feature extraction relies on the internal information of the source image, and there is a lack of correlation between the source image and the fused one, resulting in a poor fit of the proposed feature layer. When extracting image features, each source image can reference each other so that the obtained layer features have a high degree of fit. In future research, a new multi-scale tool will be applied to solve this problem. At the same time, the fusion weights with better performance are further explored.
References
- 1. Liu Y, Chen X, Cheng J, et al. Infrared and visible image fusion with convolutional neural networks[J]. International Journal of Wavelets, Multiresolution and Information Processing, 2018, 16(3):231–243.
- 2. Zhang Q, Liu Y, Blum R S, et al. Sparse Representation based Multi-sensor Image Fusion: A Review[J]. Information Fusion, 2018, 40:57–75.
- 3. Ba Virisetti D P. Multi-sensor image fusion based on fourth order partial differential equations[C]// 20th International Conference on Information Fusion (Fusion), 2017. IEEE, 2017.
- 4. Wang J, Peng J, Feng X, et al. Fusion method for infrared and visible images by using non-negative sparse representation[J]. Infrared Physics & Technology, 2014, 67:477–489.
- 5. Zhao J, Zhou Q, Chen Y, et al. Fusion of visible and infrared images using saliency analysis and detail preserving based image decomposition[J]. Infrared Physics & Technology, 2013, 56:93–99.
- 6. Ma J, Chen C, Li C, et al. Infrared and visible image fusion via gradient transfer and total variation minimization[J]. Information Fusion, 2016, 31:100–109.
- 7. Burt P J, Adelson E H. The Laplacian Pyramid as a Compact Image Code[J]. Readings in Computer Vision, 1987, 31(4):671–679.
- 8. Lewis J J RJ O’Callaghan , Nikolov S G, et al. Pixel- and region-based image fusion with complex wavelets[J]. Information Fusion, 2007, 8(2):119–130.
- 9. Nencini F, Garzelli A, Baronti S, et al. Remote sensing image fusion using the curvelet transform[J]. Information Fusion, 2007, 8(2):143–156.
- 10. da Cunha, Arthur, et al. The Nonsubsampled Contourlet Transform: Theory, Design, and Applications.[J]. IEEE Transactions on Image Processing, 2006, 15(10):3089–3101.
- 11. Kong W, Zhang L, Lei Y. Novel fusion method for visible light and infrared images based on NSST-SF-PCNN[J]. Infrared Physics & Technology, 2014, 65:103–112.
- 12.
Xiyu Han. A Study on the Fusion Algorithm of Infrared and Visible Image Based on Multi-Scale and Significant Region Analysis [D]. Beijing: Graduate University of the Chinese Academy of Sciences, 2020.(in Chinese)
- 13. Mohd Anul Haq Abhijit Ghosh, Rahaman Gazi, Baral Prashant. Artificial neural network‐based modeling of snow properties using field data and hyperspectral imagery[J]. Natural Resource Modeling, 2019, 32(4).
- 14. Mohd Anul Haq. SMOTEDNN: A Novel Model for Air Pollution Forecasting and AQI Classification[J]. Computers, Materials & Continua, 2022, 71(1).
- 15. Mohd Anul Haq. CDLSTM: A Novel Model for Climate Change Forecasting[J]. Computers, Materials & Continua, 2022, 71(2).
- 16. Mohd Anul Haq. CNN Based Automated Weed Detection System Using UAV Imagery[J]. COMPUTER SYSTEMS SCIENCE AND ENGINEERING, 2022, 42(2).
- 17. Mohd Anul Haq Prashant Baral, Yaragal Shivaprakash, Rahaman Gazi. Assessment of trends of land surface vegetation distribution, snow cover and temperature over entire Himachal Pradesh using MODIS datasets[J]. Natural Resource Modeling, 2020, 33(2).
- 18. Haq Mohd Anul Ahmed Ahsan, Ilyas Khan, Jayadev Gyani, Abdullah Mohamed, Attia El Awady, et al. Analysis of environmental factors using AI and ML methods[J]. Scientific Reports, 2022, 12(1).
- 19. Kriti Mohd Anul Haq, Garg Urvashi, Mohd Abdul Rahim Khan, Rajinikanth V. Fusion-Based Deep Learning Model for Hyperspectral Images Classification[J]. Computers, Materials & Continua, 2022, 72(1).
- 20. Haq Mohd Anul Alshehri Mohammed, Gazi Rahaman, Abhijit Ghosh, Prashant Baral, Chander Shekhar. Snow and glacial feature identification using Hyperion dataset and machine learning algorithms[J]. Arabian Journal of Geosciences, 2021, 14(15).
- 21. Mohd Anul Haq, Abdul Khadar Jilani, Prabu P. Deep Learning Based Modeling of Groundwater Storage Change[J]. Computers, Materials & Continua, 2022, 70(3).
- 22. Ma J, Zhang H, Shao Z, et al.GANMcC:A Generative Adversarial Network With Multiclassification Constraints for Infrared and Visible Image Fusion[J]. IEEE Transactions on Instrumentation and Measurement, 2020, PP(99):1–14.
- 23. Hui Li,Xiao-Jun Wu,Kittler Josef. MDLatLRR: A novel decomposition method for infrared and visible image fusion.[J]. IEEE transactions on image processing: a publication of the IEEE Signal Processing Society,2020,29: 4733–4746.
- 24. Bavirisetti D P, Dhuli R. Fusion of infrared and visible sensor images based on anisotropic diffusion and Karhunen-Loeve transform[J]. IEEE Sensors Journal, 2015, 16(1): 203–209.
- 25. Lei T, Jia X, Zhang Y, et al. Superpixel-based Fast Fuzzy C-Means Clustering for Color Image Segmentation[J]. IEEE Transactions on Fuzzy Systems, 2018:1–1.
- 26. Hui Huang, Linlu Dong, Xiaofang Liu, et al. Low Light Image Enhancement [J] Opt. Precision Eng. 2020, 28(08):1835–1849.(in Chinese)
- 27. Dianwei Wang, Qibin Xing, Pengfei Han, et al. Enhancement of low illumination panoramic images based on simulated multi-exposure fusion [J]Opt. Precision Eng., 2021, 29(02):349–362.(in Chinese)
- 28. Hongkui Xu, Xiao Han, qu Huaijing. Image grayscale transformation [J]Opt. Precision Eng., 2017, 25(04):538–544.(in Chinese)
- 29. Jian L, Yang X, Liu Z, et al. SEDRFuse: A Symmetric Encoder–Decoder With Residual Block Network for Infrared and Visible Image Fusion[J]. IEEE Transactions on Instrumentation and Measurement, 2020, 70: 1–15.
- 30. Chen J, Li X, Luo L, et al. Infrared and visible image fusion based on target-enhanced multiscale transform decomposition[J]. Information Sciences, 2020, 508: 64–78.
- 31. Cui G., Feng H., Xu Z., Li Q., Chen Y., Detail preserved fusion of visible and infrared images using regional saliency extraction and multi-scale image decomposition, Optics Communications, 2015, 341 (15): 199–209.
- 32. Tsai D Y, Lee Y, Matsuyama E. Information entropy measure for evaluation of image quality.[J]. Journal of Digital Imaging, 2008, 21(3):338–347.
- 33. Rao Y.-J., In-fifibre bragg grating sensors[J], Measurement Science and Technology, 1997, 8 (4):355.
- 34. Eskicioglu A. M., Fisher P. S., Image quality measures and their performance[J], IEEE Transactions on Communications, 1995 43 (12):2959–2965.
- 35. Zhang X, Feng X, Wang W, et al. Edge Strength Similarity for Image Quality Assessment[J]. IEEE Signal Processing Letters, 2013, 20(4):319–322.
- 36. Qu G., Zhang D., Yan P., Information measure for performance of image fusion[J], Electronics Letters, 2002, 38 (7):313–315.
- 37.
Petrovic V S, Xydeas C S. Objective Image Fusion Performance Characterisation[C]// 10th IEEE International Conference on Computer Vision (ICCV 2005), 17–20 October 2005, Beijing, China. IEEE, 2005.