Figures
Abstract
As one of the most important ways for humans to perceive the world, images contain a wealth of visual information. Digital image processing is a technology that uses computer methods to process and enhance photographs in order to extract meaningful information and improve image quality. However, current image processing techniques have poor performance in processing complex images. To improve the quality of complex images, research proposes an image information optimization processing method based on fractional order differentiation and WT algorithm. Image edge detection and image fusion are important technologies in the field of image processing, with wide application value. Therefore, the study is based on wavelet transform algorithm and fractional order differentiation to perform edge detection and image fusion. The results revealed that when the study used the four evaluation metrics of information entropy, recall, mean square error, and precision to evaluate the effectiveness of image edge detection, the Sobel operator had the highest precision of detection recall, and the smallest information entropy and mean square error. The method achieved an 80% recall rate, a minimum information entropy of 3.13, a highest detection precision of 78.9%, and a minimum mean square error of 152. The average gradient, information entropy, spatial frequency, mutual information of the method adopted by the study for image fusion was compared with other methods in case of different groups of images. The method adopted by the study for image fusion provided the best results. The precision of the proposed method edge detection by the study was higher and the performance of image fusion was better and effective in improving the quality of the image.
Citation: Long Q (2025) Image information optimization processing based on fractional order differentiation and WT algorithm. PLoS One 20(5): e0324392. https://doi.org/10.1371/journal.pone.0324392
Editor: Fatih Uysal, Kafkas University: Kafkas Universitesi, TÜRKIYE
Received: November 28, 2024; Accepted: April 25, 2025; Published: May 23, 2025
Copyright: © 2025 Qiong Long. This is an open access article distributed under the terms of the Creative Commons Attribution License, which permits unrestricted use, distribution, and reproduction in any medium, provided the original author and source are credited.
Data Availability: All relevant data are within the article and its supporting information files.
Funding: The author(s) received no specific funding for this work.
Competing interests: The authors have declared that no competing interests exist.
List of Abbreviations: FRWT, Fractional Wavelet Transform; WT, Wavelet Transform; FOD, Fractional order derivative; FO, Fractional order; SMC, Sliding mode controller; LF, Low frequency; TF, Temporal frequency; DWT, Discrete wavelet transform; RL, Riemann-Liouville; G-L, Grimwald-Letnikov; ED, Edge detection; RO, Roberts Operators; PO, Prewitt Operators; SO, Sobel Operators; LOG-O, LOG Operators; CA, Canny algorithm; GF, Gaussian filtering; GAV, Gradient amplitude value; PP, Pixel point; CO, Canny Operators; HF, High frequency; CWT, Continuous wavelet transform; OS, Original signal; MV, Mean value; var, variance; cov, Covariance; SA, Sobel algorithm; RA, Roberts algorithm; LOG-A, LOG algorithm; PA, Prewitt algorithm; LFW, Labeled Faces in the Wild
1. Introduction
As the visual basis of human perception of the world, images play a crucial role in the information-rich society of the 21st century by enabling people to express, receive, and transmit information [1–2]. However, people can perceive the world more accurately and objectively by using digital image processing technologies. It is possible to restore clarity and brightness to blurred or even invisible photographs using image enhancement techniques. There are two important steps in image processing, namely image edge detection and image fusion [3]. Many scholars have pointed out that current image processing techniques may distort or lose details when processing complex images. Moreover, the noise and complex background of images make it difficult for current image processing techniques to accurately segment and recognize targets [4–5]. The use of wavelet transform multi-scale processing technology for image denoising and feature extraction has demonstrated efficient performance. With the increase of data volume, how to effectively remove noise and interference from images while maintaining image details remains a challenge. When using fractional differentiation for edge detection, not only can it effectively extract the edge information of the image, but it can also preserve weak edge information in smooth regions. The image fusion method based on fractional wavelet transform (FRWT) can remove noise and interference in the image while maintaining image details, improving the quality and clarity of the image [6]. The image quality requirements in fields such as medical diagnosis and geological exploration are relatively high. There are problems in multi-focus image fusion, such as easy detail loss and artifacts at image edges. In view of this, in order to improve the quality of multi focus image fusion, avoid the loss of local details and edge artifacts, a method based on wavelet transform (WT) algorithm and fractional order differentiation (FOD) is proposed to optimize the information of medical and satellite images.
When studying the optimization processing of image information based on fractional order differentiation and WT algorithms, the innovation of this study lies in the combination of fractional order differentiation and WT algorithm. Compared with traditional integer order differentiation, fractional order differentiation can more flexibly adjust the order of differentiation, thereby finely controlling the local feature extraction of images. The main contribution of this study is to provide a high-performance solution for edge detection and fusion of digital images, which greatly promotes the development of digital image processing research and the intelligence of image fusion. It provides an effective method for the application of image edge detection in displacement monitoring equipment, medical images and remote sensing images.
The research will investigate the image information optimization processing method based on FOD and WT algorithm from four aspects. First, an overview of the state of FOD and WT algorithm research is presented. Second is the research on image information optimization processing methods for FOD and WT algorithms. Next comes the experimental verification of the suggested research design. Lastly, there is a summary of the research.
2. Related works
Fractional calculus generalizes differential and integral operations by unifying them into a single fractional-order (FO) derivative of arbitrary order, and is commonly used in finance, engineering, and science [7]. To address mismatched disturbances in DC/DC buck converters, X. Lin et al. presented a FO sliding mode management approach based on a higher-order nonlinear disturbance observer. The findings showed that the technique suggested by the study enhanced the system’s transient performance while lessening the system’s sensitivity to mismatch disturbances [8]. B. Long et al. presented a fuzzy FO non-singular terminal sliding mode controller (SMC) for lower control limit-grid connected converter (LCL-GCC) grid current management in an effort to solve the issue that chattering in SMC tended to excite low-frequency (LF) unmodeled dynamics and may deteriorate tracking precision. The study was based on state estimation of minimally sampled sensors by Kalman filtering. The results indicated that the proposed controller of the study could converge quickly with high tracking precision and robustness [9]. A novel finite-time FO command filtering implementation approach was proposed by X. You et al. [10]. The stabilizing function’s FO derivatives did not need to be analytically calculated in the command filtering implementation technique. The outcomes showed that the suggested strategy could guarantee that the tracking error converges in a finite amount of time to a tiny neighborhood around the origin [11]. B. Babes et al. suggested an adaptive fuzzy FO non-singular terminal SMC for synthetic robustness in an effort to enhance a DC-DC buck converter’s output voltage tracking control performance. The study combined the fractional calculus with the hybrid control method of non-singular terminal sliding mode controller (NTSMC). The study’s suggested method enhanced the conventional SMCs’ resilience to disturbances and parameter fluctuations [12]. To prevent repeated resonances that could impair system stability and power quality, M. A. Azghandi et al. presented a delay based lossless FO virtual capacitor analysis and design. The capacitor used as a resonance damper for a multiple parallel paralleling current-source inverters (CSI) system. The suggested capacitor, according to the data, offered more degrees of freedom to improve the control’s robustness and frequency behavior [13].
For time-frequency (TF) analysis and signal processing, the WT is a perfect instrument since it offers a frequency-varying “TF” window [14]. A WT real-time defect detection technique based on machine learning was presented by Y. Ma et al. for naval DC pulsating loads. The study’s suggested method could locate further flaws that could cause anomalous disruptions in the load current profile [15]. To create four distinct types of insulation failures in power cables, M. H. Wang et al. suggested a convolutional neural network power cable fault detection method based on discrete wavelet transform (DWT) and chaotic system. The findings showed that the suggested approach was capable of accurately identifying fault status changes in power cables and fault signal variations in power cables in real time [16]. R. P. Medeiros et al. found some delays in phase-based differential protection during fault occurrence due to phase convergence and proposed a new time-domain power transformer differential protection based on Clarke and WT that could be used for any power transformer. The suggested approach was effective, quick, easy to use, and independent of the differential current’s fundamental and harmonic components, according to the results [17]. Empirical WT and differential fault energy were used by J. Gao and colleagues to decompose the differential fault energy using empirical WT, which allowed for the distinction between high impedance faults and normal disturbances in distribution systems. The findings showed that the study’s suggested method could accurately identify high impedance defects from normal disturbances [18]. Xu et al. proposed an interference suppression technique for vehicle-mounted millimeter-wave radar in the adjustable Q-factor WT domain. The technique involved optimizing the model using a split-enhanced Lagrangian contraction algorithm, which addressed the issue of increased noise level caused by the distortion of radar echoes in the presence of mutual interference. As evidenced by the data, the suggested strategy lessens radar echo interference [19].
In summary, the above literature demonstrates the powerful application potential and advantages of fractional calculus and WT techniques in multiple fields. These studies have achieved significant results in improving system performance, enhancing robustness, and solving specific problems. However, there are still some challenges and limitations in terms of implementation complexity, computational cost, model dependency, dynamic response performance, long-term stability, and adaptability to complex environments, which need to be addressed and optimized in further research in the future. The proposed method combines fractional order differentiation and wavelet transform algorithms, which can effectively extract image edge information while preserving weak edge information in smooth areas. This method introduces fractional order differentiation, allowing the algorithm to adjust the differentiation order more flexibly and thus more finely control the extraction of local features in the image.
3. Image information optimization processing
3.1. Image edge detection based on fod
There are many ways of defining fractional calculus and the commonly used forms of definition are Riemann-Liouville (RL) definition, Grümwald-Letnikov (G-L) definition, and Caputo definition [20–21]. The Grumwald-Letnikov fractional calculus definition is obtained by higher-order differential generalization, which is a fundamental concept in fractional calculus, providing an extension of the traditional notion of integer-order derivatives. The G-L formula is important in fractional calculus because it provides a way to compute non-integer order derivatives. With FO derivatives, the derivative’s order might be a real number as opposed to only an integer [22]. The G-L formula defines the FO derivative of the function
at the point
. The FO derivative
is illustrated in Equation (1).
In Equation (1), denotes a binomial coefficient and
denotes a real number.
denotes the integration time step.
denotes a non-negative integer. When
is equal to 0, the binomial coefficient is defined as 1. The RL integral is a FO promotion of the successive indefinite integrals [23]. Equation (2) expresses the RL integral of the function
.
In Equation (2), denotes the gamma function and
denotes the integral variable. t denotes the integration time,
denotes the initial point of integration taken.
is the derivatives of the original function and
is the bounded linear operator. Caputo fractional derivative is one of the generalizations of integer order derivatives, which actually means any order derivative greater than or equal to 0 [24]. Caputo fractional derivative is shown in Equation (3).
In Equation (3), denotes the smallest integer greater than
.
,
is the
order derivative of
. In image processing, difference of Gaussian (DOG) is a very useful technique for detecting edges and corner points in an image. The difference between these two results is obtained by convolving the image with Gaussian kernels of different standard deviations [25]. DOG calculation is shown in Equation (4).
In Equation (4), denotes the variance and
denotes the coordinates. The smaller
is the higher the peak of the Gaussian function. The larger
is, the smoother the peaks are and the blurrier the image convolution effect is. There is a large change in the gray scale (GS) of the image edge, and the gradient value can reflect the drastic degree of GS change, so it is necessary to find the gradient in each direction of the image, and thus determine the edge region of the image. Currently the commonly used operators for edge detection (ED) and extraction are Roberts operator (RO), Prewitt operator (PO), Sobel operator (SO), LOG operator (LOG-O) and Canny algorithm (CA). The fractional gradient operator can finely control the local feature extraction of signals or images by adjusting the differential order, thereby achieving optimal performance in different application scenarios. Compared with integer order differentiation, fractional gradient operators can better capture subtle changes in signals, especially when dealing with complex nonlinear systems. Secondly, the fractional gradient operator also has memory, that is, it not only considers the value of the current point, but also considers the values of all past points. In addition, the fractional gradient operator can more effectively suppress noise while extracting signal features. This is because the fractional gradient operator can balance the smoothness of the signal and the accuracy of feature extraction by adjusting the order, thereby preserving important features of the signal while removing noise. The following is the specific ED process. First, the acquired image is transformed into a GS image, and then noise is removed from the image using Gaussian filtering (GF). Subsequently, an appropriate ED method is chosen to identify the image, and the edges that are identified are binaryized to produce an image with black and white edges. Finally, the detected edges are connected and detailed to output the processed edge image. The steps of image ED are shown in Fig 1.
The RO is a gradient algorithm that is based on cross-differentials and is also referred to as the cross-differential algorithm [26]. The approach is often used to process steep, low-noise images and discovers edge lines through the computation of local differences. When the image edges are near plus or minus 45 degrees, the algorithm performs better. When RO performs ED on an image, first the selected operator template is used to operate on the image to be detected and calculate the gradient magnitude value (GAV) of the image. Secondly, a suitable threshold is selected and the GAV of the image is judged point by point. Points that meet the criteria are assigned a value of 1, while the remaining points are assigned a value of 0. Finally, the edge image of the image to be detected is obtained. The bias derivation of RO for ,
direction is shown in Equation (5).
The LOG-O smooths the image by GF function to suppress the effect of noise [27]. The bias derivation of the LOG-O for the ,
directions is shown in Equation (6).
In Equation (6), denotes the variance and
denotes the exponential function. Fig 2 displays the LOG-O’s ED step.
PO is used to achieve ED by using the difference generated by the gray values of the pixels in a particular region. The image is first subjected to a convolution operation in order to extract the gradient values of each pixel point (PP) in both the horizontal and vertical directions. Secondly, the binarized edge image is obtained by thresholding the gradient intensity for segmentation [28]. The ED effects of the PO are more noticeable in the horizontal and vertical dimensions since it employs a 3 × 3 template for the pixel values in the region. It works well for identifying photographs with a higher level of noise and GS gradient. The bias derivation of PO for ,
direction is shown in Equation (7).
The goal of CA is to find an optimal ED solution or to find the location in an image where the gray intensity varies the most [29]. Three factors are mostly used to determine optimal ED: lowest response, high localizability, and low error rate. The specific steps of ED by Canny operator (CO) are as follows. First, the image is smoothed and its gradient is found using GF. Second, double thresholding is employed to determine the possible boundaries and non-maximum suppression technique is utilized to filter out the non-edge pixels. Finally, hysteresis technique is utilized to track the boundaries. Fig 3 depicts the CO ED process.
Combining differential derivation and GF, the SO is a discrete differential operator [30]. By labeling certain points in the region that exceed a certain number as edges, depending on how bright or dark they are adjacent to the edge of the image, the SO is used to approximate the brightness and darkness of an image. The SO adds the concept of weight to the PO and considers that the proximity of neighboring points has a different effect on the current PP. The image is sharpened and the edge contours are emphasized when the PPs are closer together, since they have a greater effect on the current pixel. The weighted difference of the gray levels of a PP’s upper and lower, left and right neighbors reaches an extreme value near the edge, which is how SO determines the edge. The expression of FOD of SO is shown in Equation (8).
In Equation (8), .
indicates that the derivative with respect to
is of order
. The specific steps for ED of SO are as follows. The original image is first subjected to Gaussian blurring before being transformed into a GS image. Second, the Sobel function is used to get the derivatives in both the x and y axes. The derivatives in the x and y directions are finally overlaid. The rate of change of the image’s GS is reflected by spatial frequency (SF), and a larger SF suggests a clearer image and higher-quality fused image. The FO SF is shown in Equation (9).
In Equation (9), denotes FO SF.
denotes FO row frequency and
denotes FO column frequency. To accurately refine the wide edges of the gradient amplitude image, it is necessary to perform non-maximum suppression on the gradient image and remove some non-edge points. The direction of the gradient is discriminated into four directions: 1, 2, 3, and 4, and replace them with directions. Map the gradient direction of the pixel to the nearest direction. Map the gradient direction of the pixel to the nearest direction. If the gradient direction of the central pixel belongs to the second direction and is greater than the average gray value and the maximum gray value of the gradient image, then the pixel is determined to be a local maximum point. In addition, to determine the appropriate threshold, the Otsu algorithm is used to calculate the threshold. This method traverses the gray values of the entire image and selects the gray value corresponding to the maximum inter-class variance as the threshold. The formula for calculating the inter-class variance is shown in Equation (10).
In Equation (10), represents inter class variance.
represents the probability that the grayscale is less than the segmentation threshold.
represents the average grayscale of images with grayscale less than the segmentation threshold.
represents the average grayscale of images with grayscale greater than the segmentation threshold.
represents the probability that the grayscale is greater than the segmentation threshold.
represents the total average grayscale of the image.
3.2. FRWT-based image fusion
The study of image ED based on FOD followed by IF based on FRWT is able to accurately extract edge information. WT is a brand-new transform analysis technique that is perfect for processing and analyzing signal TF. While FRWT is a multi-resolution analysis of signal TF domain. Performing FRWT on a signal is equivalent to passing the signal through an FO filter bank consisting of a scale function and wavelet functions with different scales to obtain the LF portion of the signal and the high-frequency (HF) portion of the signal in different frequency bands [31–32]. Its primary capabilities include the ability to fully highlight the features of specific problem aspects through transformation and the ability to localize time and SF analysis. The method eventually accomplishes time subdivision at HF and frequency subdivision at LF, which can automatically adjust to the requirements of TF signal analysis, thereby focusing on arbitrary features of the signal. The signal is gradually refined at numerous scales by telescopic translation operations. Compared with the traditional WT, it has better resolution, stronger localization and better anti-noise performance, especially suitable for processing signals with complex spatio-temporal features. The image processing based on FRWT is shown in Fig 4.
In WT theory, DWT and continuous wavelet transform (CWT) are the two main transformations [33]. The signal is transformed in the discrete time or space domain by DWT, whereas the signal is transformed in the continuous time or space domain by CWT. The discretized WT performs and is more stable in real time [34–35]. The basic principle of DWT is to discretize the CWT, which can decompose the signal or image into components of different frequencies [36]. DWT has the ability to analyze both time and frequency domains. Compared with general pyramid decomposition, DWT image decomposition can obtain LF information in horizontal, vertical and diagonal directions while extracting LF information of the image. By choosing the mother wavelet reasonably, DWT can make DWT more effective in extracting significant information such as texture and edge while compressing noise. Pyramid decomposition has information correlation between scales, while DWT has higher independence in different scales. For IF based on DWT, firstly, WT is performed on the source image. Secondly, the transformed coefficients are merged according to certain rules. Ultimately, the fused image is obtained by applying the wavelet inverse transform to the joined coefficients. In Fig 5, the IF based on DWT is displayed.
WT is a representation of the original signal (OS) using a scale function. The scale grows larger and larger as the scale level drops, representing the OS in an increasingly erratic and hazy manner while also increasing the gap between the OS. It is used as a scale function to meet the requirements that the scale function is orthogonal to its integer translation, and that the function space grown at low scales is nested within high scales. The scale function expression is shown in Equation (11).
In Equation (11), ,
determines the position of
.
denotes the scale function and
denotes the expansion coefficient. The relationship between scale space and wavelet space is shown in Equation (12).
In Equation (12), denotes the wavelet function and
denotes the wavelet function coefficients. In the discrete case, the approximation coefficients are calculated in Equation (13).
In Equation (13), represents the approximation coefficient.
,
determines the width and height, and
. The fine coefficients are calculated in Equation (14).
In Equation (14), denotes the fine coefficient. Bootstrap filtering is an image filtering technique that can achieve edge smoothing with double filtering by filtering the initial image with one bootstrap image. The bootstrap filtering uses one bootstrap image to generate weights, which are processed on the input image. This process expression is shown in Equation (15).
In Equation (15), is the output image.
is the bootstrap image.
is the input image.
and
are the index of the PP in the image.
denotes the weights. The bootstrap filtering implementation steps are shown in Fig 6.
As shown in Fig 6, firstly, the boxFilter filter is used to complete the correlation coefficient parameter (CCP). The mean value (MV) includes autocorrelation MV, cross-correlation MV, bootstrap image MV, and the original image MV that needs to be filtered. CCP calculates based on MV, including autocorrelation variance (var) and cross-correlation covariance (cov). Secondly, calculate the coefficients of the window linear transformation parameters a and b. The MV of parameters a and b is calculated according to the formula. Finally, use these parameters to obtain the output image of the bootstrap filter. The regions of the image where the brightness or gray value varies slowly are represented by the LF component. This expansive, level region of the picture characterizes the majority of the image and provides a thorough gauge of the overall intensity of the image. Equation (16) illustrates the fusion of the LF coefficient.
In Equation (16), ,
denote the weighting coefficients.
,
denote the LF coefficients.
is the fused LF coefficient. The areas of the image that fluctuate significantly, such as the borders, noise, and detail, are corresponding to the HF components. HF coefficient fusion first calculates the absolute value of the HF coefficients, and second selects the HF coefficients with the largest absolute value at each position as the fusion result. The specific steps of FRWT based IF are as follows. Firstly, the input images are pre-processed, such as gray-scaling and normalization. Then DFRWT decomposition is performed on the two input images to get the LF coefficients and LF coefficients of the images. Secondly, the LF coefficients are weighted and fused. The weights can be determined according to the image quality, information content and other factors. For example, for better quality images, higher weights can be assigned. Maximum value fusion is performed for LF coefficients. That is, the HF coefficients with the largest absolute value at each position are selected as the fusion result. Finally, the fused LF coefficients and HF coefficients are inverted by DFRWT to obtain the fused image. The specific steps of FRWT-based IF are shown in Fig 7.
4. Quality analysis of image edge detection and fusion
4.1. Experimental parameter setting
The study employs Sobel algorithm (SA), Roberts algorithm (RA), LOG algorithm (LOG-A), Prewitt algorithm (PA) and CA for three sets of image ED simulation experiments. Considering the diversity and representativeness of the dataset, as well as the comprehensiveness of performance evaluation, the selected datasets for the experiment include Labeled Faces in the Wild (LFW) dataset, Kaggle PCB defect dataset, and Rice image dataset. Compared to the datasets used in other literature, experiments conducted on the aforementioned three datasets can verify the applicability of the algorithm in various practical applications. For example, edge detection of facial images can be applied to facial recognition and expression analysis; edge detection of circuit board images can be used for industrial quality control; edge detection of rice images can be utilized for image analysis in the agricultural field. To test the generalization performance of the proposed algorithm, the LFW dataset and Kaggle PCB Defects dataset will be used for training, while the Rice Image dataset will be used as an independent dataset to test the algorithm’s performance. The LFW dataset is used to test the edge detection performance of the proposed method, the Kaggle PCB dataset is used to verify the edge detection effect of the research algorithm on images with rich edges, and the Rice Image dataset is used to verify the algorithm’s noise resistance performance. Among them, Experiment 1 performs ED on face images. Experiment 2 performs ED on complex circuit board images. Experiment 3 performs ED on rice grain images with 0.05 pretzel noise added. Fifteen sets of images of size 512*512 are selected for fusion in the study. The three methods NSCT method, PNCC method and L_P method are also selected for comparison and analysis with the FRWT method used in the study. The operating system chosen for the experiment is Windows 11X64, RAM is 16GB, CPU is 12th Gen Intel(R) Core(TM) i7-12700H @ 2.30GHz, the graphics card model is NVIDIA GeForce RTX 3060, and the graphics card memory is 16G, and the software chosen is MATLAB R2018a. FO P takes the values of 0.4 and 0.5, respectively.
4.2. Quality analysis of image edge detection based on fod
The study employs RA, SA, PA, LOG-A and CA to perform ED on three images with different levels of complexity. Among them, the image structure of Experiment 1 is more complex, the image of Experiment 2 has more complex textures at the edges, and the image structure of Experiment 3 is simpler. B is the eligible edge points and A is the total detected points. The results of ED are shown in Table 1. The B/A value of SA is the smallest in all the three sets of experiments and ED has the highest precision rate. When the detected image structure is more complex, the B/A value of SA is the smallest and the B/A value of PA is the highest, which are 0.0143, 0.0915 (p < 0.05), respectively. When the image edges have more complex textures, the B/A value of SA is the smallest and the B/A value of RA is the highest, which are 0.0219, 0.3039 (p < 0.05), respectively. When the detected image structure is simple, SA has the smallest B/A value and PA has the highest B/A value, which is 0.0146, 0.0270 (p < 0.05), respectively.
The execution efficiencies of RA, SA, PA, LOG-A and CA are shown in Table 2 when the complexity of the detected images is different. When the structure of the detected image is more complex, the RA has the lowest execution efficiency and the CA has the highest execution efficiency of 0.0497 and 0.0518 (p < 0.05), respectively. When the image edges have more complex textures, the PA is executed with the lowest efficiency and the CA is executed with the highest efficiency with 0.0598 and 0.0801 (p < 0.05), respectively. When the structure of the detected image is relatively simple, the PA is executed with the lowest efficiency and the CA is executed with the highest efficiency with 0.0363 and 0.0492 (p < 0.05), respectively. Overall, the performance of the algorithm proposed in the study on independent datasets is not significantly different from that of the LFW dataset and Kaggle PCB Defects dataset, indicating that the model has good generalization performance.
The study adopts two evaluation metrics, mean square error (MSE) and precision rate to evaluate the effect of image ED. In this, the study detects different representative regions of added noise image. The precision rate denotes the probability that the boundary pixels generated in ED are the real boundary pixels and its statistical results are shown in Fig 8. In Fig 8(a), with the increase of representative regions, the precision rate of image ED using various techniques exhibits an increasing and subsequently declining trend. When 1–5 regions are selected as representative regions, the precision rate of image ED by different methods shows a gradual increasing trend. Compared with other methods, SO has the highest detection precision rate and PO has the lowest detection precision rate. The highest detection precision rate is 78.9% for SO and 66.3% (p < 0.05) for PO. In Fig 8(b), the MSE curves of different operators for image ED fluctuate little as the representative region increases. SA has the smallest MSE and PA has the largest MSE, with SA’s MSE range being [775,850] and PA’s MSE range being [152,200] (p < 0.05).
The study adopts two evaluation metrics, information entropy and recall, to evaluate the image ED effect. Among them, the recall rate indicates the probability of detecting a real boundary pixel over all real boundary pixels. The statistical results are shown in Fig 9. In Fig 9(a), the recall of different methods for image ED increases with the increase of representative regions. When 5–7 regions are selected as representative regions, the fluctuation of the recall curve of different methods for image ED is small. Compared with other methods, SO has the highest detection recall and RO has the lowest detection recall. The highest detection recall is 80% for SO and 67.9% (p < 0.05) for PO. In Fig 9(b), the information entropy curves of different operators for image ED fluctuate less as the representative region increases. SA has the minimum information entropy and PA has the maximum information entropy. The minimum information entropy of SA and PA is 3.13 and 5.61 (p < 0.05), respectively. The maximum information entropy of SA and PA is 3.99 and 5.98 (p < 0.05), respectively.
To further analyze the performance of the proposed edge detection algorithm, it is compared with the Triple B-spline wavelet transform and Franklin moment (TBSWT-FM) and the improved canny operator (ICO). The test results are shown in Table 3.
According to Table 3, compared to other algorithms, the proposed edge detection algorithm has a smaller B/A ratio, which does not exceed 0.022 (p < 0.05). It can be concluded that the proposed edge detection algorithm in this study has higher accuracy and higher edge integrity. The effects of edge detection for the different datasets are shown in Figure Fig 10.
As shown in Fig 10, compared to other algorithms, the proposed method not only achieves accurate edge detection of ordinary images, but also improves the edge detection effect of images with rich edges. In addition, this method has good resistance to salt and pepper noise, and its noise profile is significantly reduced during the detection process.
4.3. FRWT-based quality analysis of image fusion
To verify the performance of the algorithm proposed in the study, five typical methods are selected for comparative analysis with the proposed method. Three sets of multi focus images with a size of 512*512 are selected for fusion in the experiment, with fractional order p values of 0.4 and 0.5. The quantitative evaluation indicators based on WT image fusion are compared with existing methods, and the statistical results are shown in Table 4. Among them, MI represents the mutual information between two random variables, SF represents SF, and QAB/F represents the degree of image edge preservation. Observing Table 4, it can be seen that when the fractional order p = 0.4, the MI, QAB/F, SF, and values of the WT based image fusion method are the best, which are 5.0183, 0.6981, and 12.6585, respectively. The image fusion methods based on L_P have the lowest values of MI, QAB/F, SF, and are 4.6235, 0.5817, and 7.1638, respectively. When the fractional order p = 0.5, the MI, QAB/F, SF, and values of the WT based image fusion method are optimal, with values of 5.1205, 0.7405, and 9.0975, respectively. The image fusion methods based on L_P have the lowest values of MI, QAB/F, SF, and are 4.2372, 0.4987, and 5.1348, respectively.
The average gradient, information entropy of the method used in the study for IF is compared with other methods in case of different groups of images. In Fig 10, the statistical data are displayed. In Fig 11(a), the average gradient of the method adopted by the study for IF is higher than the other methods in the case of different image groups. The higher the average gradient, the more image layers and the clearer the image. FRWT, PNCC, NSCT and L_P methods have the highest average gradient of 3.75, 3.19, 3.28 and 2.91 (p < 0.05), respectively. In Fig 11(b), the information entropy of the method used in the study for IF is lower than the other methods in the case of different image groups. The lesser the information entropy, the lesser the IF. FRWT, PNCC, NSCT and L_P methods have the least information entropy of 5.61, 6.59, 7.50 and 8.10 (p < 0.05), respectively.
In the case of different groups of images, the study used the method for IF of SF, mutual information is compared with other methods. In Fig 11, the statistical data are displayed. In Fig 12(a), the space of the method adopted by the study for IF is higher than the other methods in the case of different image groups. The image’s GS’s rate of change is reflected in the SF. Greater clarity and improved fused image quality are indicated by the larger SF. The highest SF is 6.820, 6.7836.805 and 6.748 (p < 0.05) for FRWT, PNCC, NSCT and L_P methods, respectively. In Fig 12(b), the mutual information of the methods used in the study for IF is lower than the other methods in case of different image groups. FRWT, PNCC, NSCT and L_P methods have minimum mutual information of 0.630.78, 0.94 and 1.14 (p < 0.05), respectively.
The bootstrap filtering determines the size of the filtering window. With different filter radii, the study used the metric SF to evaluate the effect of fusion of different five sets of images. The statistical results are shown in Fig 13. SF decreases with increasing filter radius. When the filter radius is greater than 6, the SF converges gradually. When the filter radius is 2, the SF values of the first, second, third, fourth and fifth group of images are 6.417, 6.376, 6.369, 6.310, and 6.275 respectively. When the filter radius is 6, the SF values of the first, second, third, fourth and fifth group of images are 6.404, 6.367, 6.335, 6.296, and 6.234 respectively. When the filter radius is 10, the SF values of the first, two, three, four and five groups of images have SF values of 6.378, 6.336, 6.324, 6.273, and 6.206, respectively.
Discussion.
The study proposed a method for optimizing image information processing that is based on FOD and WT algorithms. This method has demonstrated effectiveness in the domains of image edge detection and image fusion. Through experimental verification, this method performs well in image edge detection and image fusion, especially in key indicators such as detection accuracy, information entropy, recall rate, and mean square error (MSE), which are superior to traditional methods. X. Luo et al. proposed a fusion method based on quaternion WT and feature level copula model for image fusion problems. This method was developed to avoid the visual quality degradation of images caused by false information, as compared with other methods [40]. D. Agrawal et al. proposed an image fusion method based on DWT. This approach was developed to improve the visibility and quantitative performance of fused images, as compared with other methods [41]. However, despite significant achievements, there are still some potential optimization directions and application extensions worth further exploration. To further improve the performance of this method and apply it to a wider range of fields, such as video processing and real-time applications. In real-time applications, such as surveillance systems and autonomous driving, the real-time and accuracy of image processing are crucial. To achieve real-time processing, the methods proposed in the research can be simplified and optimized. For example, reducing the number of layers in wavelet decomposition or using fast fractional order differentiation algorithms. In addition, the combination of hardware acceleration technology can further improve real-time processing capabilities. Meanwhile, to adapt to the constantly changing environment in real-time applications, online learning mechanisms can be introduced to enable algorithms to dynamically adjust parameters to adapt to new image features. Different image information processing optimization methods are shown in Fig 14.
As can be seen from Fig 14, compared with literature [40] and literature [41], the image information optimization quality of the proposed method is better, which not only retains the new solution intact, but also basically has no noise interference.
Conclusion
To facilitate image analysis, image storage, image transmission and improve image quality, the study proposed an image information optimization processing method based on FOD and WT algorithm. The study used FOD based approach for detecting the edges of the image. Among them, RO, SO, PO, LOG-O, and CA were used to extract the edges of the image. Second, FRWT based method was used for fusion of images. The results revealed that when the study used the four evaluation metrics of information entropy, recall rate, MSE, and precision rate to evaluate the effectiveness of image ED, the SO had the highest precision rate for detection recall rate, and the lowest information entropy and MSE. The method achieved an 80% recall rate, a minimum information entropy of 3.13, a highest detection precision rate of 78.9%, and a minimum MSE of 152. The WT based IF methods had the best MI, QAB/F, and SF values when the FO took the values of 0.4 and 0.5. When the structure of the detected image was complex, RA had the lowest execution efficiency and CA had the highest execution efficiency of 0.0497 and 0.0518, respectively. PA, RA, and SA had execution efficiencies of 0.0505, 0.0513, and 0.0500, respectively. Compared with other methods, the precision rate and execution efficiency of ED of the research proposed method are improved, and the IF is more effective. Due to the fact that only 2D and grayscale images were considered when constructing the proposed algorithm, its ability to handle higher dimensional or more complex image data is not yet clear. In view of this, it is considered to introduce directional fractional wavelet transform in the future to improve the algorithm’s ability to process high-dimensional images.
References
- 1. Xu J, Ai B, Chen W, Yang A, Sun P, Rodrigues M. Wireless Image Transmission Using Deep Source Channel Coding With Attention Modules. IEEE Trans Circuits Syst Video Technol. 2022;32(4):2315–28.
- 2. Kurka DB, Gunduz D. Bandwidth-Agile Image Transmission With Deep Joint Source-Channel Coding. IEEE Trans Wireless Commun. 2021;20(12):8081–95.
- 3. Hasanvand M, Nooshyar M, Moharamkhani E, Selyari A. Machine Learning Methodology for Identifying Vehicles Using Image Processing. AIA. 2023;1(3):154–62.
- 4. He K, Gan C, Li Z, Rekik I, Yin Z, Ji W, et al. Transformers in medical image analysis. Intelligent Medicine. 2023;3(1):59–78.
- 5. Windhager J, Zanotelli VRT, Schulz D, Meyer L, Daniel M, Bodenmiller B, et al. An end-to-end workflow for multiplexed image processing and analysis. Nat Protoc. 2023;18(11):3565–613. pmid:37816904
- 6. Khushaba RN, Hill AJ. Radar-Based Materials Classification Using Deep Wavelet Scattering Transform: A Comparison of Centimeter vs. Millimeter Wave Units. IEEE Robot Autom Lett. 2022;7(2):2016–22.
- 7. Cao J, Udhayakumar K, Rakkiyappan R, Li X, Lu J. A Comprehensive Review of Continuous-/Discontinuous-Time Fractional-Order Multidimensional Neural Networks. IEEE Trans Neural Netw Learn Syst. 2023;34(9):5476–96. pmid:34962883
- 8. Lin X, Liu J, Liu F, Liu Z, Gao Y, Sun G. Fractional-Order Sliding Mode Approach of Buck Converters With Mismatched Disturbances. IEEE Trans Circuits Syst I. 2021;68(9):3890–900.
- 9. Long B, Lu PJ, Chong KT, Rodriguez J, Guerrero J. Robust Fuzzy-Fractional-Order Nonsingular Terminal Sliding-Mode Control of LCL-Type Grid-Connected Converters. IEEE Trans Ind Electron. 2022;69(6):5854–66.
- 10. Shi J, Cao J, Liu H, Zhang X. Compound Adaptive Fuzzy Output Feedback Control for Uncertain Fractional-Order Nonlinear Systems with Fuzzy Dead-Zone Input. Int J Fuzzy Syst. 2023;25(6):2439–52.
- 11. You X, Dian S, Liu K, Guo B, Xiang G, Zhu Y. Command Filter-Based Adaptive Fuzzy Finite-Time Tracking Control for Uncertain Fractional-Order Nonlinear Systems. IEEE Trans Fuzzy Syst. 2023;31(1):226–40.
- 12. Babes B, Mekhilef S, Boutaghane A, Rahmani L. Fuzzy Approximation-Based Fractional-Order Nonsingular Terminal Sliding Mode Controller for DC–DC Buck Converters. IEEE Trans Power Electron. 2022;37(3):2749–60.
- 13. Azghandi MA, Barakati SM, Yazdani A. Passivity-Based Design of a Fractional-Order Virtual Capacitor for Active Damping of Multiparalleled Grid-Connected Current-Source Inverters. IEEE Trans Power Electron. 2022;37(7):7809–18.
- 14. Li C, Yu Y, Yang Z, Liu Q, Peng X. ESR Estimation for Aluminum Electrolytic Capacitor of Power Electronic Converter Based on Compressed Sensing and Wavelet Transform. IEEE Trans Ind Electron. 2022;69(2):1948–57.
- 15. Ma Y, Maqsood A, Oslebo D, Corzine K. Wavelet Transform Data-Driven Machine Learning-Based Real-Time Fault Detection for Naval DC Pulsating Loads. IEEE Trans Transp Electrific. 2022;8(2):1956–65.
- 16. Wang M-H, Lu S-D, Liao R-M. Fault Diagnosis for Power Cables Based on Convolutional Neural Network With Chaotic System and Discrete Wavelet Transform. IEEE Trans Power Delivery. 2022;37(1):582–90.
- 17. Medeiros RP, Bezerra Costa F, Melo Silva K, Muro J de JC, Junior JRL, Popov M. A Clarke-Wavelet-Based Time-Domain Power Transformer Differential Protection. IEEE Trans Power Delivery. 2022;37(1):317–28.
- 18. Gao J, Wang X, Wang X, Yang A, Yuan H, Wei X. A High-Impedance Fault Detection Method for Distribution Systems Based on Empirical Wavelet Transform and Differential Faulty Energy. IEEE Trans Smart Grid. 2022;13(2):900–12.
- 19. Xu Z, Yuan M. An Interference Mitigation Technique for Automotive Millimeter Wave Radars in the Tunable Q-Factor Wavelet Transform Domain. IEEE Trans Microwave Theory Techn. 2021;69(12):5270–83.
- 20. Bezerra FDM, Santos LA. Chebyshev polynomials for higher order differential equations and fractional powers. Math Ann. 2022;388(1):675–702.
- 21. Yu Z, Zhang Y, Jiang B, Su C-Y, Fu J, Jin Y, et al. Enhanced Recurrent Fuzzy Neural Fault-Tolerant Synchronization Tracking Control of Multiple Unmanned Airships via Fractional Calculus and Fixed-Time Prescribed Performance Function. IEEE Trans Fuzzy Syst. 2022;30(10):4515–29.
- 22. Ragb O, Wazwaz A, Mohamed M, Matbuly MS, Salah M. Fractional differential quadrature techniques for fractional order Cauchy reaction‐diffusion equations. Math Methods in App Sciences. 2023;46(9):10216–33.
- 23. Ma Z, Liu Z, Huang P, Kuang Z. Adaptive Fractional-Order Sliding Mode Control for Admittance-Based Telerobotic System With Optimized Order and Force Estimation. IEEE Trans Ind Electron. 2022;69(5):5165–74.
- 24. Plouraboue F, Uszes P, Guibert R. Source Identification of Propagating Waves Inside a Network. IEEE Trans Netw Sci Eng. 2022;9(3):1437–50.
- 25. Liu H, Ong Y-S, Yu Z, Cai J, Shen X. Scalable Gaussian Process Classification With Additive Noise for Non-Gaussian Likelihoods. IEEE Trans Cybern. 2022;52(7):5842–54.
- 26. Wang W, Li L, Zhang F. Crack image recognition on fracture mechanics cross valley edge detection by fractional differential with multi-scale analysis. SIViP. 2022;17(1):47–55.
- 27. Ramasamy MD, Periasamy K, Periasamy S, Muthusamy S, Ramamoorthi P, Thangavel G, et al. A novel Adaptive Neural Network-Based Laplacian of Gaussian (AnLoG) classification algorithm for detecting diabetic retinopathy with colour retinal fundus images. Neural Comput & Applic. 2023;36(7):3513–24.
- 28. Budiman M, Sunardi S. Deteksi Tepi Plat Nomor Kendaraan Menggunakan Metode Operator Prewitt. ARL. 2023;7(2):157–65.
- 29. Liu L, Liu Z, Hou A, Qian X, Wang H. Adaptive edge detection of rebar thread head image based on improved Canny operator. IET Image Processing. 2023;18(5):1145–60.
- 30. S. Supiyandi T, Panggabean N, Ramadhan S, Dewi R, Yusra S. “Deteksi Tepi Sederhana Pada Citra Menggunakan Operator Sobel. Repeater. 2024;2(3):43–56.
- 31. Pramanik S. An adaptive image steganography approach depending on integer wavelet transform and genetic algorithm. Multimed Tools Appl. 2023;82(22):34287–319.
- 32. Dai TKV, Yu Y, Theilmann P, Fathy AE, Kilic O. Remote Vital Sign Monitoring With Reduced Random Body Swaying Motion Using Heartbeat Template and Wavelet Transform Based on Constellation Diagrams. IEEE J Electromagn RF Microw Med Biol. 2022;6(3):429–36.
- 33. Ahmadian R, Ghatee M, Wahlstrom J. Discrete Wavelet Transform for Generative Adversarial Network to Identify Drivers Using Gyroscope and Accelerometer Sensors. IEEE Sensors J. 2022;22(7):6879–86.
- 34. Xavier GVR, Coelho R de A, Silva HS, Serres AJR, da Costa EG, Oliveira ASR. Partial Discharge Location Through Application of Stationary Discrete Wavelet Transform on UHF Signals. IEEE Sensors J. 2021;21(21):24644–52.
- 35. Liu D, Dysko A, Hong Q, Tzelepis D, Booth CD. Transient Wavelet Energy-Based Protection Scheme for Inverter-Dominated Microgrid. IEEE Trans Smart Grid. 2022;13(4):2533–46.
- 36. Singh G, Chiluveru SR, Raman B, Tripathy M, Kaushik BK. Novel Architecture for Lifting Discrete Wavelet Packet Transform With Arbitrary Tree Structure. IEEE Trans VLSI Syst. 2021;29(7):1490–4.
- 37. A. Lawrance N, S. Shiny Angel T. Image Fusion Based on NSCT and Sparse Representation for Remote Sensing Data. Computer Systems Science and Engineering. 2023;46(3):3439–55.
- 38. K P B, M RK. ELM speaker identification for limited dataset using multitaper based MFCC and PNCC features with fusion score. Multimed Tools Appl. 2020;79(39–40):28859–83.
- 39. Song M, Lim S, Kim W. Monocular Depth Estimation Using Laplacian Pyramid-Based Depth Residuals. IEEE Trans Circuits Syst Video Technol. 2021;31(11):4381–93.
- 40. Luo X, Li K, Wang A, Zhang Z, Wu X. Infrared and visible image fusion based on quaternion wavelets transform and feature-level Copula model. Multimed Tools Appl. 2023;83(10):28549–77.
- 41. Agrawal D, Yadav AC, Tyagi PK. Low-light and hazy image enhancement using retinex theory and wavelet transform fusion. Multimed Tools Appl. 2024;83(29):72519–36.