Skip to main content
Advertisement
Browse Subject Areas
?

Click through the PLOS taxonomy to find articles in your field.

For more information about PLOS Subject Areas, click here.

  • Loading metrics

Adaptive multi-channel dehazing for enhanced visibility in underground coal mine images

  • Yingbo Fan,

    Roles Formal analysis, Project administration, Software, Validation, Writing – original draft

    Affiliation Institute of Remote Sensing and Geographic Information Systems, Peking University, Beijing, China

  • Shanjun Mao ,

    Roles Conceptualization

    sjmao_pku@163.com

    Affiliation Institute of Remote Sensing and Geographic Information Systems, Peking University, Beijing, China

  • Mei Li ,

    Contributed equally to this work with: Mei Li, Boxiang Yang, Yinglu Yang

    Roles Supervision, Validation

    Affiliation Institute of Remote Sensing and Geographic Information Systems, Peking University, Beijing, China

  • Boxiang Yang ,

    Contributed equally to this work with: Mei Li, Boxiang Yang, Yinglu Yang

    Roles Methodology

    Affiliation Institute of Remote Sensing and Geographic Information Systems, Peking University, Beijing, China

  • Yinglu Yang

    Contributed equally to this work with: Mei Li, Boxiang Yang, Yinglu Yang

    Roles Visualization

    Affiliation China University of Mining & Technology, Beijing, China

Abstract

Image dehazing has gained significant attention due to its importance in enhancing image clarity in various applications. However, existing algorithms often struggle with suboptimal performance in underground coal mine environments, characterized by dim lighting and atmospheric interference. This paper presents an adaptive multi-channel dehazing algorithm tailored for enhancing images from underground coal mines. By utilizing an improved color attenuation prior method, the algorithm effectively detects fog density, incorporating texture information and illumination invariance features from the HSV space for enhanced adaptability and robustness. The algorithm segregates foggy and fog-free image regions, applying image enhancement in clear areas and threshold multi-channel inspection dehazing in foggy regions. A multi-scale pyramid and guided filtering approach are employed to refine the estimation of image transmittance, mitigating blocky artifacts. For video dehazing, a parameter reuse mechanism leveraging inter-frame similarity significantly improves real-time performance. Experimental results on coal mine datasets and public benchmarks demonstrate that the proposed algorithm outperforms existing methods in defogging effectiveness, computational efficiency, and stability, rendering it suitable for real-time applications such as safety monitoring in underground coal mines.

Introduction

As security surveillance becomes more widespread and computer vision rapidly advances, the demand for acquiring clear, high-quality images is increasing [1,2]. Consequently, in the current field of digital image processing, image dehazing has gained extensive research attention as a critical method for enhancing image clarity [3,4]. Because atmospheric disturbances, such as haze, smoke, and dust, are prevalent in most scenes and often severely affect image quality and visual performance, they can result in loss of image details, color distortion, and reduced contrast [5,6].

To overcome the adverse effects of atmospheric interference on image quality, numerous advanced dehazing techniques based on traditional methods and deep learning have emerged. These techniques aim to restore clear visual information from haze-affected images, making them closer to real-world scenes [7]. However, in practical applications, current image dehazing algorithms still face several challenges and issues [8,9]. Firstly, traditional dehazing methods based on physical models often rely on accurate estimation of environmental lighting and atmospheric scattering models. These parameters are usually difficult to obtain directly, resulting in insufficient stability and robustness of the algorithms [10,11]. Secondly, although deep learning-based image dehazing algorithms have achieved remarkable results, their performance in handling complex scenes and varying weather conditions remains suboptimal [12,13]. Deep neural networks typically require a large amount of labeled data for training, and obtaining large-scale labeled data in real-world scenarios poses significant difficulties and challenges [14]. Furthermore, there is often a trade-off between detail restoration and maintaining vivid color in image dehazing algorithms [15]. Most current algorithms struggle to simultaneously achieve both high detail restoration and enhanced color fidelity in the processed images.

In production environments like coal mines, which are typically dimly lit, the quality of underground images is often affected by atmospheric interferences such as haze and coal dust [16]. These factors lead to significant blurring and degradation of images, severely impacting their visibility and the ability to extract information [17]. Therefore, the research and application of image dehazing technology are crucial for enhancing underground monitoring and ensuring safe production in coal mines [18,19]. However, most current image dehazing algorithms are only suitable for specific types of haze or atmospheric interference, and their effectiveness in the complex and variable conditions of underground coal mines is not guaranteed [20]. Additionally, existing algorithms often require substantial computational resources, resulting in long processing times which are not conducive to real-time monitoring and emergency response. Moreover, some algorithms may introduce distortions or artifacts during the dehazing process, compromising the accuracy and reliability of the images [21]. These issues present significant challenges to the safety monitoring and production efficiency in coal mining environments [22].

To better address the urgent demand for image dehazing in industrial production environments such as coal mines, this paper proposes an adaptive multi-channel dehazing for enhanced visibility in underground coal mine images (AMCD). The algorithm first detects haze concentration using an improved color attenuation prior method. It then enhances the image in haze-free regions and performs threshold multi-channel verification dehazing in hazy regions, ultimately outputting a complete dehazed image. The brief process is illustrated in Fig 1. The proposed algorithm is designed to be adaptive and robust in complex industrial production environments. It ensures the accuracy and reliability of the dehazing effect, meeting the practical needs of safety monitoring and production management in environments with uneven lighting, such as coal mines. This approach provides a reference for the further development of image dehazing technology.

thumbnail
Fig 1. Flowchart of image defogging algorithm based on thresholding multi-channel inspection.

https://doi.org/10.1371/journal.pone.0334251.g001

Related work

Currently, there is a significant amount of research focused on image dehazing, which can be primarily categorized into two types: prior-based dehazing algorithms and deep learning-based dehazing algorithms [23]. Prior-based image dehazing algorithms can be further divided into image enhancement-based and image restoration-based dehazing algorithms, based on different principles. Image enhancement-based dehazing algorithms typically aim to restore clear, haze-free images by removing noise and improving image contrast. These methods focus on enhancing the visual quality of the image without necessarily modeling the physical properties of the haze [24].

For example, Chen proposed a single-image dehazing method based on adaptive histogram equalization and gamma correction linearization [25]. This method is extremely fast and suitable for video sequence data. Li proposed a method that utilizes Retinex theory and Taylor series expansion to denoise nighttime images, referred to as “RDT”. During the image fusion and color conversion process, atmospheric light images and potential haze-free images are used to obtain the final haze-free image [26]. Khan calculated atmospheric light for a given hazy image by performing wavelet domain decomposition and retaining high-frequency sub-bands [27]. Dense haze removal is then carried out on the approximate low-frequency sub-bands of the given hazy image. Image restoration-based methods mainly include the dark channel prior algorithm proposed by He, which restores clear images by inversely solving the formation process of hazy images, laying the foundation for subsequent methods [28]. For instance, Li proposed a method to estimate the parameters in the physical model based on an improved bright channel prior and dark channel prior [29]. This method effectively addresses the issue where the dark channel prior fails in sky regions. Generally, dehazing effects based on the atmospheric degradation model are superior to those of image enhancement-based dehazing algorithms [30].

Advantages of deep learning neural networks in image task processing have also led to a significant number of neural network-based approaches in dehazing algorithms. The first category of methods is based on the atmospheric degradation model, utilizing neural networks to estimate parameters within the model [31]. For instance, Cai proposed a trainable end-to-end system (DehazeNet) for estimating the transmission rate [32]. DehazeNet takes a hazy image as input and outputs the transmission rate, which is then used to invert the atmospheric scattering model to recover a haze-free image. Chi Yoon designed a scaled convolution group to handle computationally intensive operations at a low resolution, thereby reducing the dehazing processing time of the end-to-end network model without compromising performance [33]. The second category of CNN methods utilizes the input hazy image to directly output a dehazed image. For example, Li proposed an All-in-One Dehazing Network (AOD-Net), which generates a clear image directly through a lightweight CNN [34]. This novel end-to-end design facilitates the integration of AOD-Net into other deep models (e.g., faster R-CNN) to enhance the performance of high-level tasks on blurry images. Manu suggested using a Generative Adversarial Network (GAN) for dehazing given blurry input images [35]. The proposed GAN architecture employs a Feature Residual Dense Network (FRDN) as the generator and a Markov discriminator with additional layers (PatchGAN) as the discriminator. This approach enhances the visibility of the scene and the authenticity of the dehazed images.

Method

Fog concentration detection based on an improved color decay prior

Traditional color attenuation prior suggests that the luminance and saturation values of a hazy image significantly change with the increase of fog density, which is used for fog density detection and removal [36]. However, this method heavily relies on the color attenuation features of the scene, which may lead to poor performance under different scenes or lighting conditions [37]. Moreover, some scenes may have complex color distributions, and traditional algorithms may not effectively handle this diversity, leading to increased errors [38]. Therefore, this paper proposes an improved method for fog density detection based on the color attenuation prior. The algorithm introduces texture information of the image in the HSV space to enhance the adaptability of the algorithm to different scenes. Additionally, the introduction of illumination invariance features makes the algorithm more robust to changes in lighting.

Fig 2 illustrates the comparison and differences in RGB and HSV color values for pixel blocks with varying fog concentrations in a dimly lit scene throughout the day. As shown in Fig 3, for fog-free areas, the luminance (V channel) and saturation (S channel) exhibit relatively higher average values (V: 120–150, S: 80–110), with minimal differences between them (<10). As fog density increases (Fig 4 and Fig 5), luminance values further increase (V: 160–220), while saturation sharply decreases (S: 20–60), causing larger differences (>100) between luminance and saturation. This paper extracts frames from video within a coal mine scene, obtaining 20,000 foggy images under various production scenarios. Based on these images, the aforementioned test was validated, and the conclusion confirms that in dimly lit scenes like coal mines, the RGB and HSV color values generally conform to the observed patterns.

thumbnail
Fig 3. RGB and HSV differences of pixel blocks under different fog concentrations (heavy fog).

https://doi.org/10.1371/journal.pone.0334251.g003

thumbnail
Fig 4. RGB and HSV differences of pixel blocks under different fog concentrations (moderately dense fog).

https://doi.org/10.1371/journal.pone.0334251.g004

thumbnail
Fig 5. RGB and HSV differences of pixel blocks under different fog concentrations (light fog).

https://doi.org/10.1371/journal.pone.0334251.g005

Based on this prior knowledge, this paper proposes an improved fog density detection method that takes into account color attenuation, converting images from the RGB color space to the HSV space, and calculating the fog density using both color spaces. The image is divided into multiple pixel blocks of a certain size, and for each pixel block B, the initial fog concentration in the HSV space is calculated using Equation (1) as follows:

(1)

The formula fhsv(x) represents the initial fog density for the current pixel block in the HSV space, where N is the total number of pixels within block B, and h(x), s(x), and v(x) are the channels of the image in the HSV space. The initial fog density in the RGB space is calculated as shown in Equation (2):

(2)

The formula frgb(x) represents the initial fog density for the current pixel block in the RGB space, with r(x), g(x), and b(x) being the channels of the image in the RGB space. Finally, the maximum values of fhsv(x) and frgb(x) are normalized and then added together to obtain the final estimate of the initial fog density, as shown in the following Equation (3):

(3)

After obtaining the initial fog density estimate, an adaptive threshold is used to perform a binarization segmentation operation on the fog density. The connected region with the maximum fog content is extracted, and the area of this region is calculated. The area of the fog-containing connected region is compared with the area threshold. If it is greater than the area threshold, the image is classified as foggy; otherwise, it is classified as non-foggy.

De-fogging algorithm based on multi-channel threshold inspection

In computer vision, the model is usually defined as Equation (4) for haze images:

(4)

In the model, I(x) represents the foggy image, J(x) represents the clear image without fog, t(x) denotes the transmission rate, which primarily describes the portion of light that is not scattered as it passes through the medium and reaches the camera [39]. A represents the global atmospheric light. The first term of the formula is called the direct attenuation term, which describes the attenuation of light as it propagates through the medium in the scene captured by the image. The second term is called the atmospheric light, which is mainly caused by scattering in front of the image acquisition device and can easily lead to a shift in the scene’s hue [40]. When the atmosphere is uniformly distributed in the scene, the transmission rate t(x) can be expressed as shown in Equation (5):

(5)

In Equation (6), represents the scattering coefficient of the atmosphere. This formula indicates that as the depth of the scene increases, the brightness of the scene decreases exponentially. Therefore, if the transmission rate can be estimated with reasonable accuracy, it is possible to recover the clear image without fog effectively [41].

Scene atmospheric light value estimation.

As stated by the atmospheric scattering model, accurately estimating the global atmospheric light value is crucial for the subsequent accurate estimation of the transmission rate [42]. Furthermore, the accuracy of the transmission rate estimation directly impacts the subsequent image dehazing process. The atmospheric scattering model explains that as the fog concentration increases, the scattering effect of the reflected light from scene objects intensifies, leading to brighter colors in the image scene [43]. Therefore, a common practice is to consider the brightest color in the image as the global atmospheric light value. However, in certain scenes, objects may exhibit brightness levels that exceed the atmospheric light value, such as light sources in environments like underground mines [22]. Failing to account for this scenario may lead to incorrect estimations, which in turn affects the accuracy of the transmission rate estimation. Therefore, it is necessary to adopt methods that can correctly identify and handle such special cases to ensure the accuracy of the global atmospheric light value estimation, and subsequently, the accuracy of the transmission rate estimation [44].

Therefore, this paper adopts the approach of constructing a multiscale pyramid followed by using a threshold multi-channel prior method to estimate the local minimum atmospheric light value. Specifically, the first step is to construct the multiscale pyramid of the image. This can be achieved by downsampling the original image I(x) at various levels to obtain a sequence of images with different resolutions, denoted as Is(x)Ss=1, where S represents the number of scales. Given that the overall illumination in a coal mine environment is dark and the light distribution is extremely uneven, using the traditional dark channel check cannot accurately estimate the atmospheric light value within the scene. Hence, the threshold multi-channel prior method is employed to calculate the corresponding image Js(x) for each scale s is computed as shown in Equation (6):

(6)

Where Is,r(y) denotes the pixel value of image Is(x) on channel r, Is,g(y) denotes the pixel value of image Is(x) on channel g, Is,b(y) denotes the pixel value of image Is(x) on channel b, B(x) denotes a region of a block of pixels centered on pixel x, and wr, wg, and wb denote the proportion of weighting values of the individual channels of the RGB, respectively.

The local minimum atmospheric light values estimated on different scales are integrated by weighted averaging, where different image resolutions correspond to different weight values and higher resolution images are given higher weights. Finally, the local minimum atmospheric light values estimated at different scales are weighted and averaged together to obtain the final estimated global atmospheric light value A. The final estimated global atmospheric light value A is obtained by integrating the local minimum atmospheric light values estimated at different scales with the weighted average as shown in Equation (7):

(7)

The wr in Equation (7). denotes the extent to which the corresponding image resolution of each scale contributes to the integration results.

Image transmittance estimation.

Assuming that the scene depth is the same within a pixel block B, the transmission rate t(x) is the same for all pixels in block B. For each pixel block, its contrast is as shown in the Equation (8):

(8)

In the Equation (8), Jc is a single-channel fog-free image and is the mean value of Jc (x) within the pixel block D. It can be obtained by bringing Equation (1) into the above equation:

(9)

In Equation (9), Ic(x) is a single-channel fogged image, and is the mean value of Ic(x) within the pixel block D. As indicated by Equation (1), there exists a linear relationship between the output value J(x) and the input value I(x). When the input value falls within an intermediate range, the mapped output value will reside within the interval [0, 255]. The input value is directly dependent on the transmittance value t. For input values outside this range, the output values will be truncated, leading to a loss of color information in the image [45]. CMSE is a monotonically decreasing function of the transmittance t; a transmittance t that is too small will result in excessive truncation of the recovered pixel values, causing severe loss of color information. Therefore, while maximizing contrast, it is essential to minimize the loss of color information to the greatest extent possible. According to the constraint condition that minimizes color loss, the transmittance t should satisfy the following condition formula:

(10)

And the contrast cost function can be expressed by EC with the formula shown below Equation (11):

(11)

Eventually the problem of transmittance estimation is converted to solving for the minimum of the cost function E, as shown in Equation (12):

(12)

In practical applications within industrial scenarios, such as coal mines, one of the challenges in the image restoration and defogging task is to handle the complex scenes present in the image. In these scenes, each pixel block may correspond to multiple different scene targets, leading to variations in their scene depth and transmittance [46]. In such cases, traditional pixel block-based transmittance estimation methods may result in noticeable blockiness after image recovery, which diminishes the visual quality and realism of the image.

To overcome this issue, a guided filtering approach can be employed to transform the transmittance estimation from a pixel block-based method to a pixel-based method, thereby alleviating or eliminating the occurrence of blockiness while recovering image details. This method can smooth the transmittance within local regions based on the structure and features of the image. Unlike traditional mean or Gaussian filtering, guided filtering takes into account the similarity and correlation between pixels in the image, better preserving the detailed information. Moreover, guided filtering elevates the estimation of transmittance from the pixel block level to the pixel level. By estimating the transmittance within a local window surrounding each pixel point in the image, rather than considering only the average transmittance of the entire pixel block, the transmittance can more finely reflect the variations in optical properties across different regions of the image. The guided filtering process applied to the image I(x) can be represented by the following Equation (13):

(13)

In the equation, t(x) represents the filtered transmittance; S is the scaling coefficient; and b is the bias value. By using the conjugate gradient method to iteratively minimize the error between the initial transmittance and the filtered transmittance, the optimal solutions for the scaling coefficient and bias can be obtained.

Time continuity factor.

Considering the continuity of video frames, this paper proposes a method to determine whether recalculation of parameters such as transmittance is necessary by detecting the similarity between video frames, thereby reducing redundant computations and enhancing the real-time performance of video dehazing. The perceptual hashing method can be employed to compare the similarity between the current frame and the previous one, determining whether the current dehazing parameters can be retained. Furthermore, the perceptual hashing method converts images or video frames into grayscale to simplify computational load and accelerate algorithm execution [47]. The implementation principle of the perceptual hashing algorithm is illustrated in the Fig 6.

thumbnail
Fig 6. Schematic diagram of the principle of perceptual hashing.

https://doi.org/10.1371/journal.pone.0334251.g006

The perceptual hashing algorithm, when assessing the similarity between the original image and the image under test, first extracts the hash features of both images. After obtaining the unique hash feature values for each image, they are compared [48]. If the Hamming distance between the hash feature sequences of the original frame and the next frame is less than a predetermined threshold, the two images are deemed similar. Conversely, if the Hamming distance exceeds the threshold, the images are considered dissimilar. Since the hash features of an image are unique, it ensures that identical images will not yield different feature sequences, thereby offering a high level of accuracy. The selection of the distance threshold is determined empirically through experiments, and it is used to judge the similarity of images based on this threshold.

To enhance image details and improve the accuracy of the algorithm’s detection, this paper utilizes the Discrete Cosine Transform (DCT) to convert the images undergoing similarity detection into the frequency domain. Subsequently, appropriate elements from the obtained frequency coefficient matrix are selected to calculate the image hash feature sequence. The formula for the Discrete Cosine Transform (DCT) is expressed as Equation (14):

(14)

In the equation, i and j represent the coordinates of elements in the pixel domain of the image, with f(i, j) being the corresponding value of the element. The variable n denotes the order of the pixel matrix. The variables x and y represent the coordinates of elements in the frequency domain of the image, with F(x, y) being the elements of the coefficient matrix in the frequency domain after transformation. The coefficient matrix is denoted as , where represents the top-left matrix. The steps of the perceptual hash algorithm are as shown in Algorithm 1.

  • Step 1: Initially, transform two adjacent frames of the image into a k × k matrix, resulting in a total of pixels.
  • Step 2: Convert the transformed image into a grayscale image.
  • Step 3: Perform a DCT on the image, transitioning from the pixel domain to the frequency domain. Calculate the frequency coefficient matrix and denote the top-left m × m matrix of this coefficient matrix as .
  • Step 4: Compute the average value of the frequency coefficient matrix , denoted as .
  • Step 5: Traverse each element in and compare with . If is greater than , assign a value of 1; otherwise, assign a value of 0. This generates the perceptual hash value of the image.
  • Step 6: After obtaining the hash feature sequences of the two images, compare the Hamming distance of the feature sequences with a predetermined threshold to determine the similarity of the images.

Algorithm 1. Sense Hash Algorithm

Input Two adjacent frames: IN, IN+1; threshold θ

Output: Similarity of two neighboring image frames

1: Transform two neighboring frames to k × k for a total of k2 pixels

2: Converting the transformed image to a grayscale map

3: Discrete cosine transform applied to the image to obtain the frequency coefficient matrix Nk×k and the m × m matrix in the upper left corner of the matrix is denoted as nm×m

4: Calculate the average of the frequency coefficient matrix nm×m, denoted as navg

5: for ni in nm×m do:

6:    if ni >navg do:

7:       Denoted as 1

8:    else do:

9:       Denoted as 0

10: Get the hash feature sequence of the two images HN, HN+1

11: if d (HN, HN+1)> θ do:

12:    Two adjacent images are not similar

13: else do:

14:    Two adjacent images are similar

Experiments

The image dehazing experiments in this paper were conducted on a workstation with the following specifications: CPU: Intel Core i9-13900HX (2.2 GHz, 6 cores, 12 threads), RAM: 16GB DDR4 2400MHz, GPU: NVIDIA GeForce GTX 4070 Ti (8GB GDDR5X). The algorithm was implemented using Python 3.6 and OpenCV 3.4.3. Due to the limited availability of public datasets for image dehazing research in dark environments and other industrial production settings, this study, in addition to validation on a self-collected dataset of coal mine faces with and without fog, also conducted validations on publicly available datasets NH-Haze2 [49], Dense-Haze [50], and SOTS-indoor [51]. The experimental parameters were set consistently with those described in C2PNet [52].

Fog concentration detection segmentation effect

During the process of image dehazing, applying dehazing algorithms to clear, fog-free images can not only lead to inefficient use of computational resources but can also backfire, potentially causing problems such as overexposure or uneven contrast which further degrade image quality [53]. To allocate computational resources more effectively, this paper proposes a fog density detection method based on an improved color attenuation prior. This method can detect the presence of fog in an image and decide whether to apply the subsequent dehazing algorithm based on the detection result. To validate the effectiveness of this approach, images of coal mining machines at the darker coal mine faces were selected as test samples. In the experiments, a fog area threshold of 25% was set, meaning that if the area of fog pixels in an image exceeds 25% of the total number of pixels, the image is determined to be foggy.

The experimental results, as shown in Fig 7 and Fig 8, demonstrate that the fog density detection method based on the improved color attenuation prior can accurately identify the distribution of fog in images. It is capable of identifying the largest connected foggy region in the fog concentration distribution map, where red areas represent foggy regions and green areas represent non-foggy regions. For instance, in the first set of images in Fig 7, the detected largest connected foggy region accounts for 76.35% of the total pixel count, which exceeds the set threshold of 25%. Therefore, this image is classified as foggy and requires dehazing treatment. Conversely, in the image of Fig 8, the largest connected foggy region constitutes only 16.87% of the total pixels, which is below the threshold, thus classifying the image as non-foggy and exempting it from dehazing operations. This fog density detection method intelligently identifies foggy images and automatically skips non-foggy ones, thereby targeting dehazing operations only where necessary, enhancing processing efficiency and the utilization of computational resources. Additionally, prior detection of foggy regions in images can optimize the video image processing workflow, especially in resource-constrained environments like coal mine faces, where it holds significant practical importance.

thumbnail
Fig 7. Schematic diagram of the effect of coal mine belt fog concentration detection and area segmentation of fogged images.

https://doi.org/10.1371/journal.pone.0334251.g007

thumbnail
Fig 8. Schematic diagram of the effect of coal mine belt fog concentration detection and area segmentation of non-fogged images.

https://doi.org/10.1371/journal.pone.0334251.g008

De-fogging effect based on thresholded multi-channel test

In order to evaluate the practical effectiveness of the threshold-based multi-channel inspection defogging method proposed in this study in foggy video scenes of underground coal mine working face, an exhaustive experimental comparative analysis is conducted in this paper. The image de-fogging algorithms such as AODNet [34], GDN [54], DeHamer [55] and C2PNet [52] are selected as the control group for comparative analysis. In terms of dataset this paper uses the image dataset constructed by extracting frames from some of the coal mine underground videos, and also adopts the coal mine underground dataset, public datasets such as NH-Haze2, Dense-Haze, and SOTS-indoor for comparative testing.

Qualitative comparison results.

In this study, qualitative experiments are designed to measure the performance of the de-fogging algorithm proposed in this paper in foggy videos from coal mine underground through intuitive visual evaluation. These experiments fully consider the subjective perception of the defogging effect and the adaptability in different scenes. In this paper, we first construct a dataset of underground coal mine images containing a variety of fog concentrations and scene complexities based on sampling frames from selected surveillance videos from a coal mine underground. These images were selected to cover a wide range of conditions from light to dense fog, as well as different lighting and environmental settings, thus ensuring the comprehensiveness of the test. One of the images of this coal mine dataset with different fog concentrations is shown in Fig 9.

thumbnail
Fig 9. Schematic diagram of different fog concentrations in a coal mine.

https://doi.org/10.1371/journal.pone.0334251.g009

Threshold-based multi-channel test defogging method is applied to these images and compared with the intuitive visualization of multiple existing defogging algorithms as shown in Fig 10.

thumbnail
Fig 10. Comparison of coal mine gangplank image and other defogging algorithms’ defogging effect.

https://doi.org/10.1371/journal.pone.0334251.g010

As can be seen from the figure, compared with other defogging algorithms, the algorithm in this paper performs relatively well in terms of image clarity, completeness of detail retention, and restoration of color fidelity. Considering the diversity of working environments in underground coal mines, this paper tests the defogging algorithm in different underground scenarios (including different lighting conditions, various fog concentrations, etc.). The purpose of these tests is to evaluate the adaptability and robustness of the algorithm to different environmental conditions. When dealing with the complex scenes in the underground, the defogging algorithm in this paper also has good results in displaying the details of mechanical equipment, the degree of fog dissipation in the far and near areas, and the clarity of the edges of people and objects. In addition, this paper also observes the comprehensive visual effect of the image after defogging, including the improvement of contrast, saturation and brightness, as shown in Fig 11.

thumbnail
Fig 11. De-fogging effect of AMCD and other de-fogging algorithms on coal mine images with different concentrations.

https://doi.org/10.1371/journal.pone.0334251.g011

Quantitative comparison results.

Quantitative experimental comparisons of the defogging algorithms are a key part of this study used to quantitatively assess the performance of the defogging algorithms proposed in this paper. These experiments aim to accurately measure the defogging effect, computational efficiency, and stability of the algorithms through a series of objective metrics. In order to quantitatively assess the defogging effect, we use a variety of objective metrics, including peak signal-to-noise ratio (PSNR) and structural similarity index (SSIM). Together, these metrics assess the similarity and difference between the defogged image and the original fog-free image. Among them, PSNR mainly measures the quality of image reconstruction, with higher values indicating better image quality, while SSIM evaluates the structural similarity between the two images, with values close to 1 indicating that the defogged image is visually closer to the original. The comparisons of AMCD with other methods on the above metrics are shown in Table 1.

thumbnail
Table 1. Comparison of quantitative evaluation of different defogging algorithms.

https://doi.org/10.1371/journal.pone.0334251.t001

This paper compares the proposed dehazing algorithm with several existing algorithms, including traditional dehazing algorithms and the latest deep learning methods. It is evident that, under the premise of a very small number of parameters, the performance of the algorithm proposed in this paper surpasses that of traditional algorithms and some deep learning algorithms on most public datasets. By running all algorithms under the same hardware and software conditions, the fairness of the comparison is ensured. Comparative experiments not only assess the relative advantages of our algorithm in terms of dehazing effect but also reveal differences in processing speed and resource consumption among the various algorithms.

In addition to the quantitative assessment of dehazing effects, computational efficiency is another important aspect considered in this paper for evaluating dehazing algorithms. This paper records the average time required by the algorithm to process a single frame image, thereby assessing the algorithm’s real-time processing capability. This metric is particularly important for real-time video monitoring systems in coal mines, as high latency could affect the system’s responsiveness and practicality. Similarly, using a coal mine dataset as the test set, this paper tests the differences in dehazing effects and time consumption among different dehazing algorithms, as shown in Table 2.

thumbnail
Table 2. Comparison of quantitative evaluation of different defogging algorithms.

https://doi.org/10.1371/journal.pone.0334251.t002

The results of the quantitative experiments provide a comprehensive set of data for the effects of various dehazing algorithms, enabling this paper to evaluate the performance of the proposed algorithm from multiple dimensions. By analyzing metrics such as PSNR and SSIM, this paper obtains an objective measure of the dehazing effect of the algorithm. Additionally, through computational efficiency analysis, this paper gains a deeper understanding of the feasibility of the threshold-based multi-channel inspection dehazing algorithm in practical applications. The comprehensive results of the quantitative experiments reveal that the proposed dehazing algorithm not only ensures high-quality dehazing effects but also demonstrates good computational efficiency and stability, making it a strong candidate for real-time applications such as video surveillance in coal mines. Furthermore, comparative experiments with other algorithms further confirm the relative advantages and application potential of the algorithm in this paper.

Conclusion and future work

In response to the urgent need for image dehazing in industrial environments such as coal mines, this paper presents an image dehazing algorithm based on threshold multi-channel prior-based defogging algorithm for underground coal mine images. The algorithm employs an improved color attenuation prior method for fog density detection, enhanced by incorporating texture information in the HSV space and illumination invariance characteristics, thereby improving the algorithm’s adaptability and robustness across varying lighting conditions. During the dehazing process, the algorithm enhances fog-free areas and applies threshold multi-channel inspection to foggy regions, effectively improving the accuracy and reliability of the dehazing results. Furthermore, by constructing a multi-scale pyramid and utilizing a guided filtering approach, the algorithm enables a more refined estimation of image transmittance, alleviating the blocky artifacts common in traditional methods while preserving image details. For video dehazing, a parameter reuse mechanism based on inter-frame similarity is designed, employing perceptual hashing to reduce redundant computations and enhance the real-time performance of video dehazing. In summary, the proposed algorithm is not only applicable to complex industrial settings but also provides new insights and methods for the further development of image dehazing technology.

Although the proposed image dehazing algorithm demonstrates considerable improvements, there remains scope for further enhancement. Specifically, optimizing the algorithm’s performance under extremely low visibility conditions could be achieved by refining the fog density detection models to ensure effective operation in environments with nearly zero visibility. Moreover, the algorithm could benefit from enhancements in processing dynamic scenes, which would entail the development of dynamic scene recognition and adaptive parameter adjustment techniques. Additionally, employing advanced color correction methods could significantly improve color fidelity and accuracy, thereby enhancing the visual quality of dehazed images. Such improvements would not only extend the algorithm’s applicability beyond industrial settings but also enable its generalization across a variety of environments, including urban and natural landscapes.

Acknowledgments

We are grateful to the reviewers for a thorough reading and constructive comments on this paper. Reprints and permissions information is available at www.nature.com/reprints.

References

  1. 1. Hameed Abdulkareem K, Arbaiy N, Hussein Arif Z, Nasser Al-Mhiqani M, Abed Mohammed M, Kadry S, et al. Mapping and Deep Analysis of Image Dehazing: Coherent Taxonomy, Datasets, Open Challenges, Motivations, and Recommendations. IJIMAI. 2021;7(2):172–98.
  2. 2. Cui Y, Zhi S, Liu W, Deng J, Ren J. An improved dark channel defogging algorithm based on the HSI colour space. IET Image Processing. 2021;16(3):823–38.
  3. 3. He Y, Liu Z. A Feature Fusion Method to Improve the Driving Obstacle Detection Under Foggy Weather. IEEE Trans Transp Electrific. 2021;7(4):2505–15.
  4. 4. Arif ZH, Mahmoud MA, Abdulkareem KH, Mohammed MA, Al‐Mhiqani MN, Mutlag AA, et al. Comprehensive Review of Machine Learning (ML) in Image Defogging: Taxonomy of Concepts, Scenes, Feature Extraction, and Classification techniques. IET Image Processing. 2021;16(2):289–310.
  5. 5. Guo J, Zhan W, Jiang Y, Ge W, Chen Y, Xu X, et al. MFHOD: Multi-modal image fusion method based on the higher-order degradation model. Expert Systems with Applications. 2024;249:123731.
  6. 6. Tran L-A, Park D-C. Lightweight image dehazing networks based on soft knowledge distillation. Vis Comput. 2024;41(6):4047–66.
  7. 7. Sharma N, Kumar V, Singla SK. Single Image Defogging using Deep Learning Techniques: Past, Present and Future. Arch Computat Methods Eng. 2021;28(7):4449–69.
  8. 8. Luo X, Cao S, Wang Z, Chen Y. LCDA-Net: Efficient Image Dehazing with Contrast-Regularized and Dilated Attention. Neural Process Lett. 2023;55(8):11467–88.
  9. 9. Tang Y, Xiang Y, Chen G. A Nighttime and Daytime Single-Image Dehazing Method. Applied Sciences. 2022;13(1):255.
  10. 10. Ma D, Fang H, Wang N, Zheng H, Dong J, Hu H. Automatic defogging, deblurring, and real-time segmentation system for sewer pipeline defects. Automation in Construction. 2022;144:104595.
  11. 11. Wu XP, Gao QQ, Huang SW, Wang K. Adaptive Retinex Image Defogging Algorithm Based on Depth-of-Field Information. Laser Optoelectron Prog. 2023;60(12):1210013.
  12. 12. Chen DD. In 19th IEEE Winter Conference on Applications of Computer Vision (WACV). 1375–383 (2019).
  13. 13. Liu Z, Zhao S, Wang X. Research on Driving Obstacle Detection Technology in Foggy Weather Based on GCANet and Feature Fusion Training. Sensors (Basel). 2023;23(5):2822. pmid:36905026
  14. 14. Guo Y, Liang R, Cui Y, Zhao X, Meng Q. A domain‐adaptive method with cycle perceptual consistency adversarial networks for vehicle target detection in foggy weather. IET Intelligent Trans Sys. 2022;16(7):971–81.
  15. 15. Ye M, Yang S, He Y, Peng Z. A grayscale image enhancement algorithm based on dense residual and attention mechanism. Vis Comput. 2023;40(3):1983–95.
  16. 16. Mao Q, Li S, Hu X, Xue X. Coal Mine Belt Conveyor Foreign Objects Recognition Method of Improved YOLOv5 Algorithm with Defogging and Deblurring. Energies. 2022;15(24):9504.
  17. 17. Chen X, Feng H-Y, Wang J-L. Robust coal granularity estimation via deep neural network with an image enhancement layer. Connection Science. 2021;34(1):472–91.
  18. 18. Xu P, Zhou Z, Geng Z. Safety monitoring method of moving target in underground coal mine based on computer vision processing. Sci Rep. 2022;12(1):17899. pmid:36284147
  19. 19. Gao J, Li B, Wang X, Zhang J, Wang L. A coal and gangue detection method for low light and dusty environments. Meas Sci Technol. 2023;35(3):035402.
  20. 20. Mao Q, Wang Y, Zhang X, Zhao X, Zhang G, Mushayi K. Clarity method of fog and dust image in fully mechanized mining face. Machine Vision and Applications. 2022;33(2).
  21. 21. Zhang X, et al. Research on low illumination video enhancement technology in coal mine heading face. Coal Geology & Exploration. 2023;51:309–16.
  22. 22. Dai L, Qi P, Lu H, Liu X, Hua D, Guo X. Image Enhancement Method in Underground Coal Mines Based on an Improved Particle Swarm Optimization Algorithm. Applied Sciences. 2023;13(5):3254.
  23. 23. Li W, Fan G, Gan M. Progressive encoding-decoding image dehazing network. Multimed Tools Appl. 2023;83(3):7657–79.
  24. 24. Guo Y, Zhu Y, Liu L, Huang Q. Research review of space-frequency domain image enhancement methods. Computer Engineering and Application. 2022;58:23–32.
  25. 25. Chen MC, Pan YW. In 14th International Conference on Graphics and Image Processing (ICGIP). (2022).
  26. 26. Li P, Tian J, Tang Y, Wang G, Wu C. Deep Retinex Network for Single Image Dehazing. IEEE Trans Image Process. 2021;30:1100–15. pmid:33259301
  27. 27. Khan H, Sharif M, Bibi N, Usman M, Haider SA, Zainab S, et al. Localization of radiance transformation for image dehazing in wavelet domain. Neurocomputing. 2020;381:141–51.
  28. 28. He K, Sun J, Tang X. Single Image Haze Removal Using Dark Channel Prior. IEEE Trans Pattern Anal Mach Intell. 2011;33(12):2341–53. pmid:20820075
  29. 29. Li C, Yuan C, Pan H, Yang Y, Wang Z, Zhou H, et al. Single-Image Dehazing Based on Improved Bright Channel Prior and Dark Channel Prior. Electronics. 2023;12(2):299.
  30. 30. Dong H, Chen J. Infrared image defogging algorithm based on atmospheric scattering model and retinex. Infrared Technology. 2019;41:347–56.
  31. 31. Fan D, Lu X, Liu X, Chi W, Liu S. An iterative defogging algorithm based on pixel-level atmospheric light map. IJMIC. 2020;35(4):287.
  32. 32. Bolun C, Xiangmin X, Kui J, Chunmei Q, Dacheng T. DehazeNet: An End-to-End System for Single Image Haze Removal. IEEE Trans Image Process. 2016;25(11):5187–98. pmid:28873058
  33. 33. Jeong CY, Moon K, Kim M. An end-to-end deep learning approach for real-time single image dehazing. J Real-Time Image Proc. 2023;20(1).
  34. 34. Li B, Peng X, Wang Z, Xu J, Feng D. In 2017 IEEE International Conference on Computer Vision (ICCV).
  35. 35. Manu CM, Sreeni KG. GANID: a novel generative adversarial network for image dehazing. Visual Computer. 2022.
  36. 36. Ngo D, Lee G-D, Kang B. Improved Color Attenuation Prior for Single-Image Haze Removal. Applied Sciences. 2019;9(19):4011.
  37. 37. Liang Z, Wang Y, Ding X, Mi Z, Fu X. Single underwater image enhancement by attenuation map guided color correction and detail preserved dehazing. Neurocomputing. 2021;425:160–72.
  38. 38. Liu K, Liang Y. Underwater image enhancement method based on adaptive attenuation-curve prior. Opt Express. 2021;29(7):10321–45. pmid:33820170
  39. 39. Zhuang Y, Zheng Z, Zhang Y, Lyu L, Jia X, Lyu C. Dimensional Transformation Mixer for Ultra-High-Definition Industrial Camera Dehazing. IEEE Trans Ind Inf. 2024;20(3):5022–32.
  40. 40. Li C, He T, Wang Y, Zhang L, Liu R, Zheng J. Pipeline image haze removal system using dark channel prior on cloud processing platform. IJCSE. 2020;22(1):84.
  41. 41. Huang S, Li D, Zhao W, Liu Y. Haze Removal Algorithm for Optical Remote Sensing Image Based on Multi-Scale Model and Histogram Characteristic. IEEE Access. 2019;7:104179–96.
  42. 42. Mahajan P, Jakhetiya V, Abrol P, Lehana PK, Subudhi BN, Guntuku SC. Perceptual Quality Evaluation of Hazy Natural Images. IEEE Trans Ind Inf. 2021;17(12):8046–56.
  43. 43. Fu H, Liu W, Chen H, Wang Z. An Anisotropic Gaussian Filtering Model for Image De-Hazing. IEEE Access. 2020;8:175140–9.
  44. 44. Wu X, Chen X, Wang X, Zhang X, Yuan S, Sun B, et al. A real-time framework for HD video defogging using modified dark channel prior. J Real-Time Image Proc. 2024;21(2).
  45. 45. Guo Z, Nan B, Wang K. Research and application of video fog concentration detection and real-time fog removal method in underground coal mine. Industry and Mine Automation. 2023;49:31–8.
  46. 46. Si L, Wang Z, Xu R, Tan C, Liu X, Xu J. Image Enhancement for Surveillance Video of Coal Mining Face Based on Single-Scale Retinex Algorithm Combined with Bilateral Filtering. Symmetry. 2017;9(6):93.
  47. 47. Xu D, Ren N, Zhu C. Integrity Authentication Based on Blockchain and Perceptual Hash for Remote-Sensing Imagery. Remote Sensing. 2023;15(19):4860.
  48. 48. Wang R, Zheng J, Zhou W, Wang Q, Lu Y, Tao Y. Adaptive hash code balancing for remote sensing image retrieval. International Journal of Remote Sensing. 2023;44(2):690–712.
  49. 49. Ancuti CO, Ancuti C, Timofte R. NH-HAZE: An Image Dehazing Benchmark with Non-Homogeneous Hazy and Haze-Free Images. IEEE. 2020.
  50. 50. Ancuti CO, Ancuti C, Sbert M, Timofte R. Dense haze: A benchmark for image dehazing with dense-haze and haze-free images. arXiv. 2019.
  51. 51. Ge W, Lin Y, Wang Z, Wang G, Tan S. An Improved U-Net Architecture for Image Dehazing. IEICE Trans Inf & Syst. 2021;E104.D(12):2218–25.
  52. 52. Zheng Y, Zhan J, He S, Dong J, Du Y. In: Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition. 5785–94.
  53. 53. Liu B, Wei Q, Ding K. ZYNQ-Based Visible Light Defogging System Design Realization. Sensors (Basel). 2024;24(7):2276. pmid:38610491
  54. 54. Liu X, Ma Y, Shi Z, Chen J. In 2019 IEEE/CVF International Conference on Computer Vision (ICCV).
  55. 55. Guo CL. In: Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition. 5812–20.
  56. 56. Yu Z, et al. FSAD-Net: Feedback Spatial Attention Dehazing Network. IEEE Transactions on Neural Networks and Learning Systems. 2022;(PP).