Figures
Abstract
This work aims to promote early and accurate diagnosis of Temporal Lobe Epilepsy (TLE) by developing state-of-the-art deep learning techniques, with the goal of minimizing the consequences of epilepsy on individuals and society. Current approaches for TLE detection have drawbacks, including applicability to particular MRI sequences, moderate ability to determine the side of the onset zones, and weak cross-validation with different patient groups, which hampers their practical use. To overcome these difficulties, a new Hybrid Attention-Enhanced Transformer Network (HAETN) is introduced for early TLE diagnosis. This approach uses newly developed Fuzzy-AAL Segmentation Framework (FASF) which is a combination of Fuzzy Possibilistic C-Means (FPCM) algorithm for segmentation of tissue and AAL labelling for labelling of tissues. Furthermore, an effective feature selection method is proposed using the Dipper- grey wolf optimization (DGWO) algorithm to improve the performance of the proposed model. The performance of the proposed method is thoroughly assessed by accuracy, sensitivity, and F1-score. The performance of the suggested approach is evaluated on the Temporal Lobe Epilepsy-UNAM MRI Dataset, where it attains an accuracy of 98.61%, a sensitivity of 99.83%, and F1-score of 99.82%, indicating its efficiency and applicability in clinical practice.
Citation: Khan H, Alutaibi AI, Tejani GG, Sharma SK, Khan AR, Ahmad F, et al. (2025) MRI based early Temporal Lobe Epilepsy detection using DGWO based optimized HAETN and Fuzzy-AAL Segmentation Framework (FASF). PLoS One 20(7): e0325126. https://doi.org/10.1371/journal.pone.0325126
Editor: Xiaohui Zhang, Bayer Crop Science United States: Bayer CropScience LP, UNITED STATES OF AMERICA
Received: December 29, 2024; Accepted: May 7, 2025; Published: July 2, 2025
Copyright: © 2025 Khan et al. This is an open access article distributed under the terms of the Creative Commons Attribution License, which permits unrestricted use, distribution, and reproduction in any medium, provided the original author and source are credited.
Data Availability: Temporal Lobe Epilepsy-UNAM MRI Dataset (https://openneuro.org/datasets/ds004469/versions/1.1.3/download). This dataset contains resting-state fMRI and task-based fMRI data specifically designed for evaluating working memory. It includes imaging data from both patients diagnosed with Temporal Lobe Epilepsy (TLE) and healthy control participants.
Funding: The author extends appreciation to the Deanship of Postgraduate Studies and Scientific Research at Majmaah University for funding this research through project number R-2025-1769.
Competing interests: The authors have declared that no competing interests exist.
1. Introduction
TLE is one of the most common types of epilepsy, occurring in approximately 1 in 2,000 individuals globally. It is responsible for roughly 60% of the cases of drug-resistant epilepsies [1]. TLE often results in major neurological and cognitive impairments that devastate the quality of life of the patient [2,3]. TLE is mostly attributed to predispositions inherited through genetics, brain infections, and traumatic injuries to the head, among other neurological disorders, such as hippocampal sclerosis. This could cause distortions to normal temporal lobe activities leading to unprovoked and recurrent seizures associated with TLE [4,5]. Tables 1 and 2 list the abbreviations and symbols utilized in the current result work, respectively.
TLE is best treated and managed if diagnosed early and accurately. Previous approaches of TLE diagnosis involve clinical evaluation, followed by a qualitative analysis of MRI scans where time is greatly consumed and the results are prone to human errors [6,7]. Moreover, the minute anomalies signifying TLE often slip through routine assessment procedures. Therefore, it becomes critical to design more efficient, reliable, and automated detection tools that will aid physicians in the early identification of TLE [8]. Early diagnosis shall thus drastically improve the prognosis for the patients. In this regard, thorough interventions can be made to control the progression of the disorder, thus improving quality of life [9,10].
In the last few years, DL has been extended to automatically improve the diagnostic capabilities of medical imaging, especially in the neuroimaging domain. CNNs are the most applicable deep learning models in image analysis since they allow the detection of many patterns and features within medical images [11,12]. These models have shown great effectiveness in diagnosing different neurological disorders including Alzheimer’s disease, multiple sclerosis, and brain tumours [13]. However, their application in the context of TLE detection remains underexplored. Current MRI based approaches for TLE detection include CNN models for distinguishing between TLE, Alzheimer’s and normal control subjects based on T1w images, 3DCNNs trained on RS-fMRI for seizure prognosis and automated classification of hippocampal pathology lateralization based on multimodal MRI data [14,15]. These methods have been found useful but they have some drawbacks such as low specificity, high computational costs, data type dependency and low transferability across different patients. This work aims at addressing this gap by developing a new DL model for TLE identification using structural MRI scans as a quantitative and objective diagnostic approach.
The major contribution of the work includes:
- To detect TLE based on the Temporal Lobe Epilepsy-UNAM MRI Dataset.
- To introduce novel FASF which is an integration of the FPCM algorithm and AAL labelling to enhance tissue segmentation, achieving more accurate delineation of brain regions for TLE detection.
- To implement the advanced pre-processing techniques, including skull stripping, bias field correction, min-max normalization, and median filtering, to ensure high-quality input MRI images.
- To develop an advanced novel DGWO algorithm for selecting the most relevant features, reducing computational complexity, and enhancing model performance.
- To implement HAETN deep learning model that integrates attention mechanisms and transformer-based architectures to achieve superior accuracy, sensitivity, and robustness in TLE detection
The organization of the work: Section 2 contains review of literature, section 3 presents the proposed methodology in detail followed by simulation result and discussion in section 4, and finally the suggested work concludes in section 5.
2. Related work
In 2023, Chang et al. [16] suggested the use of CNN algorithm for the classification of TLE, AD and healthy control subjects using T1-weighted MRI scans. Further, they suggest using feature visualization methodologies to determine the parts of the brain that the CNN uses to distinguish between these diseases.
In 2022, Luckett et al. [17] proposed to use a 3DCNN that is trained with RS-fMRI data from the healthy controls with synthetic changes to the regions of interest for predicting the side of the TLE patient’s seizure onset. Further, they suggest applying Grad-CAM to determine the areas of the brain that provide the most discriminative information regarding the seizure onset zones.
In 2021, Caldairou et al. [18] designed an MRI based fully automated classifier for lateralizing covert hippocampal pathology in TLE patients based on T1, T2 and FLAIR/T1 features. The classifier is established on MRI data of patients with histologically confirmed HS and is expected to yield a higher lateralization than electroclinical data, including the side of surgery. The model’s performance is also evaluated in TLE cohorts outside the model, that is, in independent TLE cohorts.
In 2021, Beheshti et al. [19] suggested the analysis of the DIR data in combination with machine learning to differentiate between normal controls and epileptic patients as well as to determine the focus side in MRI-negative PET-positive TLE patients. They use whole-brain DIR data from participants who were scanned with high-resolution structural MRI and DIR to train a linear support-vector machine model.
In 2022, Aslam et al. [20] utilized volumetric MRI and 18F FDG PET to non-invasively diagnose or rule out TLE based on statistically generated threshold of asymmetry in these imaging studies. They plan to do so using PVL from amygdalohippocampal volumetry and PML from PET in order to distinguish TLE patients from extra TLE patients.
In 2021, Qu et al. [21] suggested the use of 3D-CNN framework derived from ResNet to detect mesial TLE in T2-FLAIR MRI images. The framework is to find the symmetrical differences of the corresponding brain areas, where the inputs are the symmetrical cubes. The proposed 3D-CNN is then compared with radiomics algorithms and visual assessment to demonstrate its potential for accurate and efficient diagnosis of MTLE, and could be used as a CAD system for epilepsy patients.
In 2021, Sherman et al. [22] proposed a clinical automated quantitative MRI measurement within statistical models to identify surgery outcomes for TLE patients. This entails employing pre-surgical T1-weighted MRI volumetric measurements derived from NeuroQuant to estimate the probability of seizure freedom and an Engel score of I at the time of surgery. The proposed study will also improve the prediction model and the role of volumetric data to include cortical volume loss, both focal and diffuse changes outside the surgical area affecting the seizure outcomes.
In 2021, Fu et al. [23] suggested that resting state fMRI and network-based connectivity analysis should be employed to compare the functional connectivity between MTLE and BECT. They argue that reduced functional connectivity in MTLE particularly between cortical networks and subcortical structures such as the hippocampus means the network is low efficiency and associated with poor prognosis. On the other hand, hyperconnectivity in BECT might be a compensatory process, which could be the reason for the better outcome. The findings of this analysis will seek to explain the differences in the brain network patterns of different types of epilepsy.
In 2021, Shi et al. [24] suggested that one should explore alterations in functional homotopy and FC in the whole brain in TLE. It also intends to determine which brain regions are important for classification through fMRI. The study applies VMHC and MVPA to determine areas in the brain affected by TLE and their correlation with the neuropsychological tests.
In 2021, Hadar et al. [25] proposed a three-dimensional GluCEST imaging for analysing brain glutamate networks in patients with no lesional TLE. This method is an advancement from previous single-slice glutamate imaging, allowing for more comprehensive spatial analysis. It aims to lateralize seizure onset in MRI-negative, no lesional TLE patients by detecting increased ipsilateral GluCEST signal in the hippocampus. Table 3 represents the comparison of existing techniques.
3. Proposed methodology
This work presents a new method for diagnosing TLE, a type of epilepsy that originates in the temporal region of the brain. The detection process starts from obtaining raw MRI images from the Temporal Lobe Epilepsy-UNAM MRI Dataset after which the images undergo rigorous pre-processing stages including skull stripping, bias field correction, min-max normalization and median filtering to provide improved input image for the subsequent steps. For tissue segmentation, the newly developed FASF which integrates FPCM algorithm and AAL labelling to improve the segmentation of different brain regions. After that, texture, shape and colour feature techniques are used for feature extraction from the segmented images, which offer a complete description of essential features that can be used to achieve accurate TLE identification. To select optimal features, the DGWO is used to improve feature selection, minimize computational burden, and improve model performance. Last, the detection of TLE is done through the newly developed HAETN, a deep learning model that combines attention and transformer-based structures to provide higher accuracy, sensitivity, and robustness. This integrated approach is a major step forward in improving the detection of TLE and can overcome the existing methods to provide better diagnostic and clinical results. The architecture of the proposed approach is represented in Fig 1.
Data collection and ethical consideration: The suggested approach is evaluated based on the Temporal Lobe Epilepsy-UNAM MRI Dataset (https://openneuro.org/datasets/ds004469/versions/1.1.3/download) This dataset contains resting-state fMRI and task-based fMRI data specifically designed for evaluating working memory. It includes imaging data from both patients diagnosed with TLE and healthy control participants. This comprehensive dataset supports studies on memory function and neural dynamics in TLE and healthy populations.
Ethical considerations: Secondary analysis was performed on this publicly available data set. The Temporal Lobe Epilepsy-UNAM MRI Dataset is completely anonymized and had been made available within the bounds of ethics for data sharing. Thus, the dataset was used only for scientific research and had no information revealing the identities of subjects. There is no further need for ethics approval or informed consent because it is a retrospective analysis of anonymous data drawn from an open-access source. This research conformed to the code of ethics, as laid down by the institutional review boards and other worthy professional organizations.
3.1. Pre-processing
Normally, pre-processing is applied to attain noise-free data from the raw images. In case of brain MRI, the complex structures need to be pre-processed to enhance image quality and highlight the significant areas. It includes:
3.1.1. Skull stripping.
Skull stripping is a process of eradicating all the features other than the brain tissue from the MRI or CT scan image. In TLE detection, skull stripping is required because it separates the brain from other structures. This enables deep learning models to concentrate on the regions of the brain that are most useful in identifying the abnormalities linked with epilepsy, including the temporal lobe lesions while excluding the non-brain tissues. When skull stripping is done, the image only has the brain region that is void of the skull and soft tissue artefacts [26]. This cleaned-up image benefits the detection algorithms as it minimizes the noise, increases the accuracy of brain region segmentation, and makes it easier to identify areas related to epilepsy.
3.1.2. Bias Field Correction (BFC).
Bias field correction serves to promote image quality. Most images have intensity in-homogeneity problems due to bias fields. Bias field signals are low-frequency signals that attenuate high-frequency information and thus degrade image quality [27]. Bias field correction is, therefore, an attempt to remedy this by applying energy-minimization operations. Bias field correction proceeds in two steps in which images are decomposed into two multiplicative components. The first component is the estimation of the bias field, followed by the correction of the bias field. These two components are optimized using energy minimization methods.
BFC is required in TLE identification since MRI images are susceptible to intensity inhomogeneity resulting from variations in the magnetic field. These inhomogeneities can also affect the tissue contrast which in turn affects the ability of deep learning models to accurately identify abnormalities such as lesion or structural change in the brain [28]. After BFC, the image intensity is more uniform throughout the brain because the intensity variations due to the MRI scanner’s magnetic field are minimized. This results in improved definition and less ambiguity in the depiction of the brain tissues, which in turn helps the model to identify corresponding features for TLE diagnosis with more precision.
3.1.3. Min-Max-Score Normalization.
Min-Max-Score Normalization is a preprocessing technique used to scale image pixel intensities to a fixed range, typically between 0 and 1. This is achieved using the formula:
where Min and Max are the minimum and maximum intensity values in the image.
Min-Max-Score Normalization is important in the detection of TLE as it ensures that all the pixel intensities are scaled consistently, so all the features contribute proportionately to the learning process and prevents dominance by higher intensity ranges. This improves the accuracy of the model, accelerates the convergence during training, and also enhances the contrast uniformity of the image. Rescaling pixel values to a defined range (e.g., 0–1) after normalization in order to optimize the image for effective feature extraction and analysis by the deep learning model [29].
3.1.4. Median Filter.
The Median Filter is a non-linear technique in image processing and it eliminates noises in MRI images by replacing the central pixel of a given window with the median of the neighbouring pixels. This effectively eliminates salt-and-pepper noise and small variations of the intensity while keeping edges and other structures. In the detection of TLE, the Median Filter is essential in enhancing the quality of the image, since it removes noise without blurring the features that are critical for proper analysis. After filtering, the MRI image becomes smoother with minimized noise and preserved key anatomical details, allowing the deep learning model to extract more reliable features. Fig 2 presents the MRI images at various pre-processing stages. Fig 2 shows some sample MRI images and the different stages of pre-processing. Each row is for a different MRI subject, highlighting improvements at every step.
(a) Median filtered images used to enhance tissue contrast along with suppressing noise. (b) Fuzzy Possibilistic C-Means (FPCM) clustering outcomes with the segmented brain regions and corresponding distinct tissue classes. (c) Anatomical labeling by AAL-or Automatic Anatomical Labeling-atlas with matching segmented regions to known brain structures for its anatomical localization and interpretation.
- Column one contains the input images in their original form, including noise, skull artifacts, and intensity inhomogeneities (Fig 2a).
- Column two (Fig 2b) applies skull stripping to remove non-brain tissues, retaining the brain area for critical analysis.
- Column three (Fig 2c) shows bias field-corrected images in which intensity non-uniformity due to scanner flaws or inconsistencies in the magnetic field is corrected, creating a more homogeneous appearance across brain tissues.
- Then Column four (Fig 2d) displays a min-max scaling of the images whereby pixel intensities are normalized within a specified range that allows for faster as well as more stable convergence of the model during training.
- Then Column five (Fig 2e) represents the median-filtered images smoothed further to suppress any residual noise but maintain important brain structures and edges.
This pre-processing step sequence is very important in making sure that the input data is clean, consistent, and optimally adapted for reliable segmentation and classification within the next stages of the proposed framework.
3.2. Segmentation phase via Fuzzy-AAL Segmentation Framework (FASF)
One of the best algorithms for segmenting brain tissues is the extension of the original FCM known as the FPCM, adding the concept of possibilistic membership to solve some of the constraints of the original FCM, in coping with noise and outliers often found in medical imagery like MRI. FPCM combines the merits of fuzzy clustering (soft membership) with probabilistic clustering (robustness against outliers), which makes it suitable for the complicated nature of MRI data. FPCM is the technique used in medical image segmentation, which is especially relevant for MRI scans, where fuzzy and possibilistic clustering combine to improve medical image segmentation. The fuzzy component allows every voxel to belong to different tissue types with different membership degrees and supports overlapping or gradually transitioning tissues, and the possibilistic component improves robustness by dealing with noisy data and outliers. The fuzzy membership represents the degree of association of each voxel with its tissue type, based on similarity in intensity, and the possibilistic membership lowers the influence of the outliers. The objective function is iterated until it converges, and the mean intensity values for each tissue type (GM, WM, CSF) are indicated by the cluster centres. This results in high-quality segmentation so that ROI are delineated accurately, which is crucial for the detection of diseases like TLE from MRI data. Fuzzy-AAL Segmentation Framework (FASF) represents segmentation results on brain MRI scans in Fig 3. Each row signifies a patient MRI, thereby indicating the resilience of the pre-processing and segmentation pipeline against different sets of anatomical structures and intensity variations.
(a) Original MRI slices as feeds into the deep learning model. (b) Grad-CAM heatmaps depicting the most influential zones contributing to model’s decision making, where warmer colors mean stronger activations. (c) Superimposed pictures of the Grad-CAM heatmap with the original MRIs, allowing the interpretation of the model’s predictions in relation to the anatomy.
- First column entries (Median-filtered images, Fig 3a) show considerable noise reduction in MRI scans while retaining the fine anatomical details. The Median Filter is applied to buff off salt-and-pepper noise and random noise artifacts that do exhibit major areas of interest like the hippocampus or the temporal lobe structures, which become critical in the case of temporal lobe epilepsy (TLE) analysis. The cleaner images guarantee that the later segmentation processes get carried out on high-quality data, preserving the structural detail necessary for accurately localizing and delineating regions that harbor anomalies.
- The second column (FPCM segmentation results, Fig 3b) clearly displays the working of the Fuzzy Possibilistic C-Means Algorithm in segmenting the brain tissues into various regions that are based on intensities similarity. Here, Cyan pertains to intermediate brain tissues (gray matter); while the yellow indicates very intense tissues (such as white matter or perhaps lesion domains), Brown refers to probable lesions or specific anatomy variations, and dark blue highlights background or non-brain regions. Of particular note is the fact that the FPCM results evidenced clearer demarcation of brain compartments from non-brain elements thus giving an excellent intermediate representation more appropriate to clinical interpretation and further functional mapping.
- Third is the column showing the results for AAL Labelling, which applied the AAL atlas onto the pre-segmented images (Fig 3c). Mapping as such marks those areas that are very highly activated or very intense with White and possibly above functional significance. These are themselves colored Blue according to the various areas that have been anatomically labeled and which show the structural-functional mapping of the brain, whereas Red indicate particular anatomically or functionally relevant regions of interest (e.g., hippocampus, amygdala) and which are predominantly critical in studies of epilepsy. This segmentation provides the standard anatomical references to compare results across subjects as well as to enhance the interpretability of potential lesion localization by following up the tissue-level clustering achieved through FPCM.
- In general, the figure depicts a three-tiered segmentation capability of FASF: (1) suppressed noise imaging, (2) intensity-based clustering for tissue, and (3) anatomical labeling, all of which combine for accurate and interpretable detection of pathological features in TLE-related studies.
3.2.1. Fuzzy possibilistic membership update.
During segmentation, membership functions and
are updated so that the degree of membership of a pixel
in a cluster
is established. The update of fuzzy membership
is defined by computing the distance between a pixel
and the centre of the cluster
, weighted with a fuzziness value
and it can be deliberated using the following Eq. (2),
where, is the fuzzy membership of pixel
in a cluster
,
is the Euclidean distance between pixel
and the cluster centre
,
is the fuzziness parameter (usually
>1),
is the number of clusters (e.g., GM, WM, CSF). Outliers are accounted for by updating the possibilistic membership function
and it can be arithmetically deliberated using the following Eq. (3),
It’s designated as , which is the probability of a pixel
belonging to cluster
,
is the Euclidean distance between pixel
and cluster centre
,
is a tuning parameter that controls the fuzziness of the cluster in probabilistic clustering,
is a tuning parameter that controls the sensitivity of the membership.
The objective function combines fuzzy and possibilistic terms to reduce the distance between picture pixels and cluster centres by considering both types of memberships and it can be arithmetically deliberated using the following Eq. (4),
Here, the parameter is the objective function to be minimized,
is the total number of pixels or voxels in the MRI image,
is the number of tissue types or clusters (e.g., GM, WM, CSF),
is the fuzzy membership of voxel,
in cluster
,
is the possible membership of Voxel
in cluster
,
is the distance between the voxel
and cluster centre
,
is the possibilistic distance (a measure of how far the voxel
is from the cluster centre).
Cluster Centre Update: is the average intensity level of the corresponding tissue types. The centres are updated iteratively by the weighted membership of the pixels. Moreover, the cluster centre updating has been mathematically deliberated in the following Eq. (5),
Here, is the updated cluster centre of tissue type
,
is the intensity value of voxel
,
and
are fuzzy and possibilistic membership values of voxel
in cluster
.
Stopping Criteria: The iteration is reiterated till convergence, and usually, it is measured by the variation of objective function between two successive iterations, thus it can be mathematically given in Eq. (6),
Here, the following parameters and
were the objective functions over the iterations
, as well as
correspondingly and
is the lowest threshold rate which indicates the level of convergence.
Segmentation Results: After convergence, each voxel is assigned to the tissue type with the highest membership value, which may be fuzzy or possibilistic. Segmented tissue maps such as GM, WM, and CSF can be obtained by using Eq. (7),
Here, the following parameter is the output segmented label for the voxel
. Moreover, the following parameters such as
and
were the fuzzy and possibilistic memberships for voxel
in cluster
.
3.2.2. AAL labelling.
The AAL procedure maps segmented brain areas to preset anatomical ROIs using the AAL template. The transformation process can be described as follows in a numerical form by using the following Eq. (8),
Here,
Transformation Function : Affine or non-linear transformation is needed to align the segmented image coordinates
to the standardized space of the AAL atlas. This ensures that an accurate spatial correspondence between a patient’s MRI scan is achieved with predefined anatomical regions of interest in ROIs of the AAL template. This method, which is usually carried out using tools such as SPM or ANTs, standardizes the orientation and size of MRI data to allow precise mapping of segmented brain tissues into their correct regions, such as the temporal lobe, for further analysis and diagnosis.
Region Mapping : This is an anatomically designated region in the AAL atlas, indexed by
. Moreover, the labelling function assigns each voxel to the ROI
that has the highest overlap or intensity correspondence after alignment.
Final Mapping : Returns the anatomical label
for each voxel in the transformed space. This will indicate the correctness of labelling regional divisions like gray matter, white matter, and CSF segmented into AAL-defined ROIs including the temporal lobe.
3.3. Feature extraction
The mathematical expressions for texture, color and shape features are manifested in Table 4.
Features from texture: Local Binary Patterns (LBPs): LBP is critical in retrieving the textural features within the MRI images, thereby very useful in identifying the structural abnormalities that characterize TLE. LBP compares the intensity of a center pixel with the intensity of its neighbors, which it then encodes as a binary pattern. The patterns that result from this are histograms that portray local textures in grey matter, white matter, and cerebrospinal fluid. This strong representation is extremely helpful for the detection of minute textural changes that other approaches may miss.
Color Features: It summarizes the statistical features of the intensities of colors of an image, which is very crucial for MRI scans, especially when they have pseudo-color representations. The first moment corresponds to mean intensity, the second moment to the spread of intensities, and the third moment to the asymmetry in the distribution of intensity. These parameters characterize the tissue properties with a precise description of the abnormality in intensity patterns, suggesting epileptogenic areas in the temporal lobe.
Shape features: Shape features analyse the morphology of the segment regions in MRI images, depicting structural abnormalities in the temporal lobe. Area, perimeter, eccentricity, and compactness are the measures by which the size and shape and the border as well as the regularity of a region are quantified. An abnormal shape or abnormally large eccentricity of the segmented grey matter can point to degenerative changes associated with epilepsy. Form descriptors thus enhance the localization and identification capabilities of epileptic tissue by the DL model.
3.4. Feature selection using the proposed Dipper-Grey Wolf Optimization (DGWO)
The proposed DGWO algorithm is designed to enhance feature selection for temporal lobe epilepsy detection by selecting the most relevant features from texture, color, and shape attributes. By leveraging the combined strengths of DTO and GWO; DGWO achieves optimal feature selection while reducing computational complexity.
- 1). Dipper Throated Optimization
The proposed DGWO takes into account the extracted texture, color and shape features. The DTO simulates the real process of identifying the locations and velocities of swimming and flying birds in order to locate food. Eq. (15) is used to update the position and speed of the swimming birds. The features that have the highest discriminating power (among texture, color as well as shape features) are prioritized based on the relevance to detection performance.
where represents the iteration number,
and
denotes the normal location and best location of the bird,
and
are the adaptive values these values are changed based on the random values and iteration number during the optimization process. On the other side, perform the update of the flying bird’s location by the given Eqs (10) and (11).
where, represents the updated speed of each bird,
denotes the global best location,
is the weight value,
indicates the random number in [0,1],
and
are constants. Based on Eq. (17), the dynamic feature selection is accomplished by means of reducing the classification loss.
- 2). Grey Wolf optimization
An old-fashioned metaheuristic algorithm called the Grey Wolf optimization (GWO) replicates the hunting habits of four grey wolves such as wolf in a pack. These wolves work together to find, locate, and encircle their prey. The algorithm continuously raises the goal of the solution space by using a mathematical model that imitates the foraging habits of a pack of grey wolves. The GWO algorithm’s main model is shown below.
There are two stages in the GWO algorithm: the siege stage and the seek-for-prey stage. The following Eqs (12) and (13) determines the siege phase:
where represents the wolf position in iteration
, the position vector of prey is
,
and
denote the coefficient vectors, which are determined in Eqs (14) and (15):
Assume that throughout the hunting phase, alpha, beta, and delta will hunt and will know the potential for each position based on their experience. This can be expressed using Eq. (16) to Eq. (18), respectively.
During the searching and attacking stage, wolves attack a victim if and use a casually created vector
within the range of [-2a, 2a]. To prevent local optimization, a random variable called
influences the decoy search. Thus, features are extracted based on the fitness function of the DGWO algorithm. Algorithm 1 presents the pseudocode of the hybrid DGWO algorithm. The adaptive tuning performance of DTO is utilized to explore the feature space, while the balance distribution of GWO is utilized to refine and exploit the selected features for enhanced classification performance.
Algorithm 1: DGWO algorithm
1. Initialize the location of the bird with size
2.Fitness function ,
,
,
,
,
,
,
,
,
and max iterations
Here, is the classification error.
3. Evaluate fitness function for each
4. Find the best bird
5.While do
6. for do
7. If (R < 0.5) then
8. Update the location of the grey wolf agents using:
9.
10. else
11. Update the speed of the flying bird using:
12.
13. Update the location of the swimming bird using:
14.
15. end for
16. Evaluate fitness function for each
17. Update ,
,
,
,
,
,
18. Find the best bird
19. Set
20. Set
21. end while
22. return
3.5. Deep learning-based detection via HAETN
A new HAETN is introduced in this research work for accurate as well as early detection of the TLE. This HAETN model encapsulates the BiLSTM with Attention, Transformer-based models, Lightweight MobileNets, and CBAM, respectively. The sequence relationships existing in the images’ feature representations are taken up by the BiLSTM component, while the attention focuses on crucial areas regarding epileptic activity. The Transformer architectures extend capacity for training the model on ultimately more global dependencies and contextual information across feature maps via multi-head self-attention and positional encoding. The lightweight MobileNets contribute to efficient extraction of features, thus allowing the model to deploy in a scenario of constraint resources. The CBAM applies channel and spatial attention to only attend to relevant features that improve the detection’s accuracy. Through this combination of components, the HAETN has effect of being able to have a very high sensitivity and specificity when processing neuroimaging data for detecting subclinical markers of temporal epilepsy as well as being computationally efficient for real-time clinical or scale-up applications.
3.5.1. BiLSTM with attention.
BiLSTM networks improve context capture by processing the sequence in both forward and backward. Then, attention is used to highlight crucial temporal phases. In BiLSTM, the forward hidden state and backward hidden state
are computed for each time step
Then, attention weights are calculated to focus on important time steps.
The strength of this attention-weighted output , which puts weights on the most relevant chunks of the sequence, thereby increasing the robustness of this model towards spoofing is expressed. Architecture of the proposed HAETN is graphically depicted in Fig 4.
3.5.2. Transformer-based model.
Based on the transformer-based models, Self-Attention Transformers with multi-head attention to seize both local and global dependencies for mimicking fine-grained patterns towards TLE detection. Conformer provides both local and long-range contexts for higher accuracy by combining transformer and convolution layers. They are very useful where ASV cases are subjected to complex patterns of TLE.
- i). Self-Attention Transformers
Self-Attention Transformers can spot complex spoofing patterns that involve both local and global dependencies by using multi-head attention to make a grab for global dependencies by focusing on different segments of the input sequence. In multi-head attention, several attention heads operate in parallel, each capturing various aspects of the relationships between the sequence’s tokens through distinct input projections.
where denotes the number of attention heads. In order to provide information on the relative positions of tokens in the sequence, positional encodings are added to the input embeddings because transformers are inherently disorganized.
where denotes the positional encodings. The final multi-head attention output is passed via feed-forward layers and residual connections are determined in Eq. (25).
where indicates the position-wise feedforward network and
is applied for normalization.
- ii). Conformer
The Conformer, a hybrid deep learning architecture, brings together the best parts of transformers and CNN with the purpose of appropriate local and global dependency modelling over sequential data. For typical applications of time-series data, like spoofing detection, the conventional transformer models could be further improved by inclusion of convolutional layers to enable better local temporal or spatial pattern capturing. That’s how the Conformer learns long-range dependencies while the sequence, with the combination of convolutional layers along with self-attention methods, effectively captures fine-grained information, or it can say that the model can manage contextual information – both short-term and long term – owing to this combination, which makes it quite useful when both local patterns such as speech phoneme structures and global relationship, such as sentence context, are important.
3.5.3. Lightweight MobileNets.
MobileNet are effective for real-time applications because both employ depthwise separable convolutions to reduce computation costs. Here, the separate convolutional kernels are used in the depth-wise convolution to filter each input channel.
where is the depth-wise convolution filter for each channel
In pointwise convolution,
convolution is applied for combining the depth-wise convolution output across channels.
Denoted as is the pointwise filter that connects channel k with the output channel
CNNs have proven to be more effective and efficient in the task of spoof detection, especially on devices with scarce resources.
3.5.4. Convolutional Block Attention Module (CBAM).
Improving upon the initial emphasis on a CNN by using channel-wise and spatial attention methods, CBAM improves concentration of a CNN onto relevant spatial and channel features. Channel attention from a channel attention module depends upon shared multi-layer perceptron-MLP and is calculated by adding global average and max pooling along with the spatial dimensions.
where is the sigmoid function. The spatial attention
is calculated in the spatial attention module by convolving the channel dimension and applying average and max pooling.
where represents the concatenation. The final attention-weighted output is determined in Eq. (30).
The output from the Transformer Encoder, is passed through Fully Connected (FC) layers for classification. The FC layer computes the class probabilities using Eq. (31), where
and
address weight matrix and bias for the FC layer, and
ensures the output represents a probability distribution over the possible classes.
Table 5 delivers the hyperparameter settings of the proposed HAETN. Fig 5 displays the architecture of feature extraction phase in proposed HAETN. Fig 4 depicts the overall architecture of proposed HAETN.
(a) Input MRI images showing brain anatomy in different subjects. (b) Grad-CAM visualizations depicting areas of intensity in the model attention with red and yellow showing high-importance areas, while blue suffices minimally. (c) Grad-CAM heatmaps could overlay on original MRI images for a clearer understanding of how the model’s focus corresponds with the underlying anatomical structures.
4. Result and discussion
4.1. Experimenatal setup
The proposed TLE detection model using the suggested HAETN approach has been implemented in Python on Intel core® core i3 processor 7020U@2.3 Ghz, 8 GB RAM, 64-bit operating system. Here, Temporal Lobe Epilepsy-UNAM MRI Dataset is employed for detection which is available in the link https://openneuro.org/datasets/ds004469/versions/1.1.3/download. The simulation results demonstrated the effectiveness and significance of the developed model. The evaluation was carried out utilizing a variety of performance metrics including sensitivity, specificity, and accuracy. Additionally, a comparative study is performed to analyse the competence of proposed HAETN with baseline models such as 3D CNN, ST-LSTM, and ResNet [21], and recent DL models like CNN-LSTM, CNN [16].
Gradient-weighted Class Activation Mapping (Grad-CAM) has been demonstrated to be influential in highlighting regions that discriminate between classes affecting the decision of a deep learning model aimed at MRI-based analysis. In this figure, there are three columns: (a) original input MRI images; (b) Grad-CAM heatmaps; and (c) Grad-CAM outputs are overlaid on the original MRIs. Each row from top to bottom represents a different subject, indicating that the model performs robustly across brain anatomies and MRI acquisition parameters.
In Fig 5, in the first column (Input MRI images), we have rendered in pure form preprocessed brain MRI slices into the deep learning model. A good appreciation of the characteristic structural variabilities and subtle textural patterns captured in these images is fundamental to any accurate interpretation regarding the diagnosis of complex neurological circumstances such as epilepsy or brain tumors.
The second column (Grad-CAM heatmaps, Fig 5b) highlights regions of interest (ROIs) that the model focuses on during its predictions. The heatmaps are color-coded; red and yellow indicate regions of high activation and importance while blue indicates lower activation.
In all cases, the model identifies the central areas of the brain, suggesting that the pathological characteristics affecting classification mainly reside in and around these regions. This central focus is important in diagnosing cases regarding lesion detection, localization of seizure foci, and segmenting brain tumors since these abnormalities would usually be located in and around critical structures.
The third column (Superimposed Grad-CAM maps, Fig 5c) provides a fused visualization of the original MRI and the Grad-CAM heatmap. Superimposition is an intuitive way for medically minded experts to interpret the attention of the model in relation to the anatomical landscape.
Regions indicated by the model correspond to anatomically plausible regions of abnormality or areas of interest with respect to function. Thus, it is evident that the model is not merely relying on data but also develops clinically interpretable rationale.
Furthermore, heterogeneous patterns of superimposed maps from patient to patient show that the model can dynamically adjust where it attends depending on individual variations of anatomy and pathology and strengthens the model’s generalization potential.
Essentially, Grad-CAM visualizations demonstrate that the deep learning model is not reliant on spurious correlations or irrelevant image artifacts but on internally consistent, meaningful, and medically relevant patterns. This further reinforces model trustworthiness and viability for an actual clinical decision support system.
4.2. Confusion matrix
The confusion matrix shown in Fig 6 summarizes the prediction results over the test data. It signifies the performance of the classification model in terms of true positives, true negatives, false positives, and false negatives:
- True Positives (TP): The model makes a correct prediction for the positive class.
- True Negatives (TN): The model correctly predicted the negative class.
- False Positives (FP): The model made a false call in predicting a positive for a negative sample.
- False Negatives (FN): The positive sample wrongly predicted negative by the model.
From the confusion matrix (values are simulated), the proposed model has better classification accuracy minimizing both false positives (FP) and false negatives (FN). This directly impacts the three critical performance metrics-precision, sensitivity, and specificity:
4.3. ROC analysis
Fig 7 presents the ROC curve of the proposed TLE detection model with a notable AUC score nearing 1.0 (i.e., AUC = 0.98), which implies its excellent competency in distinguishing the positive and negative classes: • Proposed Model: The true positive rate (TPR) of the proposed model remains high while the false positive rate (FPR) remains quite low even towards extreme thresholds. • Baseline Models: The models such as ResNet or 3D-CNN tend to produce AUC values in the range of 0.94–0.96, pegging their discriminatory capacity from slightly lesser to mild. This proves that the developed model tends to be more biased towards balancing sensitivity and specificity.
4.4. Training and validation analysis: Accuracy-loss vs epoch
Training and Validation loss: Based on the simulation graph shown in Fig 8, the proposed model has a significant and smooth decrease of gradient from training and validation losses, which indicates effective learning without overfitting. Example: After 20 epochs, training loss = 0.05 and validation loss = 0.07 They incline to have higher gaps between the training and validation losses, as they suffer from overfitting. The proposed model minimizes both losses better and ensures a larger generalization on unseen data.
Training and Validation Accuracy: The proposed model outperforms baseline models and is therefore considered superior in terms of training and validation accuracies: Proposed Model: Training accuracy = 98%, Validation accuracy = 97.5% (after 20 epochs). Baseline Models: Training accuracy = 93–95%, Validation accuracy = 92–94%. Accuracy converges consistently without a large gap between the two types, thus indicating that the proposed model has been properly optimized to prevent overfitting. Others like ResNet or CNN-LSTM might take long to converge or may have larger differences in accuracy, points toward less generalization. All three models outperform all baselines in Accuracy, in F1-Score, and in Sensitivity to ensure balanced and trustworthy performances. Models like 3D-CNN and ResNet have delivered impressive results; however, they lack in precision and specificity and hence create more misclassifications. The proposed model, as such, dominates the baseline models, being the best performer in all critical metrics but specifically in F1-Score (0.986) and Accuracy (0.977). This improvement happens because of the following abilities of the given model:
- Cut down inaccuracies in classification as shown in the confusion matrix.
- Strike an optimal balance between sensitivity and specificity as manifested with ROC and AUC scores.
- Ensure good generalization robustness as illustrated by the graphs of loss and accuracy in training and validation.
Thus, the proposed model works more appropriately for applications in identifying TLEs where robust predictions are essential.
4.5. Performance analysis of proposed HAETN model over existing models for early TLE detection
The proposed HAETN model was extensively evaluated to detect TLE from MRI data against a number of existing models and more contemporary deep learning architectures, including as 3D-CNN, ResNet [21], and CNN [16]. Various metrics for performance, including F1-score, MCC, NPV, FPR, FNR, sensitivity, specificity, accuracy, and precision, have been employed in the analysis. Two different data splits, 70% for training and 30% and 80% for training and 20%, respectively, were employed for training and testing in the evaluation. The findings, displayed as a graphical representation in Fig 9, clearly demonstrates that HAETN performs better than the other metrics on all criteria that was analysed. This outstanding result demonstrates the model’s effectiveness in accurate and frequently determining TLE circumstances. The representation highlights the proposed method’s adaptability, which makes it an appropriate strategy for medical imaging diagnostics where reliability and precision are essential.
The HAETN model’s performance in comparison to other existing models is presented in Table 6. The dataset is split into 70% for training and 30% for testing, and the metrics comprise accuracy, precision, sensitivity, specificity, F1-score, MCC, NPV, FPR, and FNR. With an accuracy of 97.79%, the proposed approach scores significantly higher than other models like 3D-CNN [17] (95.39%) and ResNet [21] (94.53%). Its sensitivity of 98.39% and accuracy of 97.56% indicate its efficacy in accurately determining TLE patients. Furthermore, the model has a high sensitivity of 97.53%, demonstrating the ability to reliably identify non-TLE cases with a low probability of false positives. metrics which include the F1-score of 98.64%, MCC of 97.98%, and NPV of 97.81% highlight its reliability and consistent classification performance. Among all models, it has the lowest FPR of 2.04% and FNR of 1.23%, showing its adaptability to minimizing error.
The performance of the HAETN model in comparison to various existing models is illustrated in Table 7. the metrics comprises Accuracy, precision, sensitivity, specificity, F1-score, MCC, NPV, FPR, and FNR; the data split is divided to 80% for training and 20% for testing. With a larger training dataset, the HAETN model shows even better results, achieving an impressive accuracy of 98.61% and sensitivity of 99.83%, indicating nearly perfect detection of TLE cases. Its specificity of 98.37% and precision of 98.01% are consistent with its strong performance across other metrics. Its high classification quality is demonstrated by its 99.82% F1-score and 98.52% MCC. In comparison with CNN-LSTM or ST-LSTM, which achieved lower performance, the model’s very low False Positive Rate of 1.53% and False Negative Rate of 0.94% display it’s reliability in decreasing misclassifications.
4.6. Performance evaluation for varying learning data
The efficiency of the Hybrid Attention-Enhanced Transformer Network (HAETN) was evidenced against other deep learning models through the performance evaluation of varying learning data ratios: 70:30 and 80:20. The results acquired are manifested in Table 8. The model achieved the shortest training time of 5.46 seconds for a 70:30 split and 4.52 seconds for an 80:20 split, demonstrating computational efficacy. In fact, other conventional deep learning architectures, like 3D-CNN, ResNet, CNN-LSTM, standard CNN, and ST-LSTM, showed greatly increased training times above HAETN. ResNet, for example, ran the longest-respective times of 8.57s for 70:30 and 8.05s for 80:20, most likely due to its more complex architecture and deep layers. In addition, CNN and ST-LSTM also reveal considerable time spent in training exceeding 8 seconds on the 70:30 split. Classifying training time reduction for the model can be attributed to effective feature selection by means of Dipper-Grey Wolf Optimization (DGWO) and the Fuzzy-AAL Segmentation Framework (FASF), which minimize redundant computations allowing easier feature extraction. Furthermore, improvement in training time when moving from a 70:30 to an 80:20 learning split denotes better generalization as the model requires far fewer epochs to learn from a larger training set. These results also prove that HAETN is an efficient way of increasing classification accuracy and can be good for early detection of Temporal Lobe Epilepsy (TLE) in clinical practice.
4.7. Statistical analysis of proposed model over SOTA
Statistical and comparative analysis of the HAETN model against best state-of-the-art (SOTA) deep learning models (shown in Table 9) indicates its performance superiority through different performance indices. It produces maximum precision (97.8% ± 0.6), accuracy (98.0% ± 0.5), sensitivity (98.3% ± 0.5), specificity (98.1% ± 0.4), and F1-score (98.0% ± 0.5)-a robust classifier in the detection of Temporal Lobe Epilepsy (TLE). Moreover, Matthews Correlation Coefficient (MCC) of 97.9% ± 0.5 indicates a strong match between the predicted and actual classifications, thus signifying the reliable performance of the model. Competing models-like 3D-CNN, ResNet, CNN-LSTM, CNN, and ST-LSTM-show very inferior performance concerning all metrics. The next-best model among these is 3D-CNN which achieves accuracy the maximum with 96.0% ± 0.7. ResNet brings up the rear at 95.5% ± 0.8. However, both traditional CNN and ST-LSTM models do not, scoring below 94.5%, indicating their lack of ability in capturing elaborate spatial-temporal patterns in MRI data. Furthermore, HAETN boasts an FPR of 1.9% ± 0.3 along with an FNR of 1.7% ± 0.3, showing that there are lesser chances of misclassification. These improvements are owing to hybrid transformer-based attention mechanism, the Fuzzy-AAL Segmentation Framework (FASF), and Dipper-Grey Wolf Optimization (DGWO), feature selection coming together to enhance feature extraction, segmentation accuracy, and classification efficiency. In summary, it offers an understanding of the potentials of HAETN-a highly accurate, reliable, and efficient model for early TLE diagnosis-with performance superior to many conventional deep learning models both in diagnostic accuracy and computational efficiency.
4.8. K-Fold validation of proposed model
The Hybrid Attention-Enhanced Transformer Network (HAETN) generalization and robustness were tested through 5 folds cross-validation. As confirmed by the results shown in Table 10, the model can consistently perform quite well in classifying TLE MRI data across all five folds. It attains an impressive average accuracy of ~98% revealing the ability to classify highly. Precision, sensitivity, specificity, and F1-score consistently score extremely high across all folds with minimum diversions, assuring the model is perfect while balancing false positives and false negatives. The Matthews Correlation Coefficient (MCC), which is a very strong performance indicator for imbalanced datasets, continues to stay above 97.4% in all folds, thus validating the predictive strength of the model. Both False Positive Rate (FPR) and False Negative Rate (FNR) remain very low but with slight discrepancies across folds, proving their reliability by distinguishing TLE from non-TLE cases. K1 achieved the highest accuracy (98.5%), precision (98.4%), and sensitivity (98.8%), which reflects exceptional classification performance for this data subset. K3 drops slightly on the accuracy side (97.5%), but precision (97.3%) and sensitivity (97.8%) remain very high, indicating the model’s robustness even when data splits are not as favorable. K4 and K5 showed very good performance (~98% accuracy), thus demonstrating that the model stays stable across distributions of data. From the above, one can conclude that HAETN does not overfit at specific training data partitions since there is a valid performance on unseen data. This further enhances reliability for real-life TLE diagnosis applications.
4.9. Computational efficiency and inference performance analysis of HAETN
From the performance evaluation of the Hybrid Attention-Enhanced Transformer Network (HAETN) at varying epochs shown in Table 11, one can conclude that the architecture can scale well and be efficient for Temporal Lobe Epilepsy (TLE)-based MRI classification. Different epoch numbers, ranging from 20 to 100, yield trade-offs between the computational costs (GFLOPs) and inference time. For example, under the 20-epoch case, the computational cost could amount to 12.5 GFLOPs with an inference time of 45 ms. However, in this case, if one trains with more epochs, computational costs increase linearly from 20 to 100 epochs, while the inference time decreases from 45 to 38 ms, with values of 12.5 and 62.5 GFLOPs, respectively, demonstrating better optimization and speed in decision-making. The increase in the momentum parameter from 0.9 to 0.99 helps stabilize convergence, ensuring that gradient updates are applied efficiently. The constant batch size of 32 provides a visible balance between computational efficiency and model generalization. Hence, these results emphasize that extended training with optimized momentum facilitates fast inference and improved model performance, thus rendering the proposed model a computationally viable solution for TLE diagnosis in real-time.
4.10. Ablation study
The study examines the effect of different deep-learning architectures on MRI-based early Temporal Lobe Epilepsy (TLE) detection, and the results acquired are manifested in Table 12. The proposed Hybrid Attention Enhanced Transformer Network (HAETN) gives the highest accuracy of 98.0 ± 0.5%, in comparison with baseline models, such as 3D-CNN (96.8 ± 0.6%), ResNet (96.3 ± 0.7%), CNN-LSTM (95.9 ± 0.8%), and ST-LSTM (96.8 ± 0.6%). Also, the proposed model provides enhanced sensitivity (98.3 ± 0.5%) and greater specificity (98.1 ± 0.4%), which enable robust classifications of epileptic and non-epileptic cases: F1-score (98.0 ± 0.5%) and MCC (97.9 ± 0.5%) demonstrate strong generalization ability of the model, while both False Positive Rate (FPR: 1.9 ± 0.3%) and False Negative Rate (FNR: 1.7 ± 0.3%) are still lower than the corresponding levels for competing models. The data validate the efficacy of the CNN model [16] feature extraction, showing comparable performance. But HAETN’s hybrid attention mechanism has particular advantages that can optimally integrate spatial and temporal dependence to substantially improve the diagnostic accuThe performance of the proposed method is thoroughly assessed by accuracyracy and reliability.
5. Conclusion
This work proposed a novel method for diagnosing TLE, starting with raw MRI images from the Temporal Lobe Epilepsy-UNAM MRI Dataset that was pre-processed through techniques like skull stripping, bias field correction, min-max normalization, and median filtering to ensure high-quality inputs. The novel Fuzzy-AAL Segmentation Framework (FASF) algorithm was adopted for segmenting the brain tissues and the tissues are segmented accurately. The texture, shape, and colour of the images are extracted and the best features was selected by using the newly introduced DGWO for increased performance and to minimize the computational cost. Detection was done using the novel DL based HAETN. The proposed approach was implemented in Python on Intel core® core i3 processor 7020U@2.3 Ghz, 8 GB RAM, 64-bit operating system. The result obtained based on the stated dataset indicate the overwhelming performance of the suggested detection approach with accuracy of 98.61%, specificity of 98.37%, and FNR of 0.0094. These result has proven the greater detection accuracy than current techniques. The proposed method uses the best deep learning technologies to enhance the accuracy and reliability of the method. In doing so it eliminates the drawbacks of current techniques and opens the door to more effective diagnosis and treatment of the disease at an earlier stage. In the future, it can be applied to identify other kinds of epilepsy and other neurological diseases and incorporated into real-time clinical applications. Additional improvement could be done in the area of computational efficiency for use in the low-resource environment.
References
- 1. Yu Z, Kachenoura A, Jeannès RLB, Shu H, Berraute P, Nica A, et al. Electrophysiological brain imaging based on simulation-driven deep learning in the context of epilepsy. Neuroimage. 2024;285:120490. pmid:38103624
- 2. Wang B, Xu Y, Peng S, Wang H, Li F. Detection Method of Epileptic Seizures Using a Neural Network Model Based on Multimodal Dual-Stream Networks. Sensors (Basel). 2024;24(11):3360. pmid:38894151
- 3. Zhao W, Wang W-F, Patnaik LM, Zhang B-C, Weng S-J, Xiao S-X, et al. Residual and bidirectional LSTM for epileptic seizure detection. Front Comput Neurosci. 2024;18:1415967. pmid:38952709
- 4. Huang L, Zhou K, Chen S, Chen Y, Zhang J. Automatic detection of epilepsy from EEGs using a temporal convolutional network with a self-attention layer. Biomed Eng Online. 2024;23(1):50. pmid:38824547
- 5. Zhu L, Wang W, Huang A, Ying N, Xu P, Zhang J. An efficient channel recurrent Criss-cross attention network for epileptic seizure prediction. Med Eng Phys. 2024;130:104213. pmid:39160021
- 6. Nandan D, Kanungo J, Mahajan A. An error-efficient Gaussian filter for image processing by using the expanded operand decomposition logarithm multiplication. J Ambient Intell Human Comput. 2024:1–8.
- 7. Shajahan S, Pathmanaban S, Tiruvenkadam K. RIBM3DU‐Net: Glioma tumour substructures segmentation in magnetic resonance images using residual‐inception block with modified 3D U‐Net architecture. Int J Imag Syst Technol. 2024;34(2):e23056.
- 8. Hedayati R, Khedmati M, Taghipour-Gorjikolaie M. Deep feature extraction method based on ensemble of convolutional auto encoders: Application to Alzheimer’s disease diagnosis. Biomed Signal Process Control. 2021;66:102397.
- 9. Zheng X, Liu W, Huang Y. A novel feature extraction method based on Legendre multi-wavelet transform and auto-encoder for steel surface defect classification. IEEE Access. 2024.
- 10. Al-Khuzaie MIM, Al-Jawher WAM. Enhancing brain tumor classification with a novel three-dimensional convolutional neural network (3D-CNN) fusion model. J Port Sci Res. 2024;7(3):254–67.
- 11. Yadav S, Dhage S. TE-CapsNet: time efficient capsule network for automatic disease classification from medical images. Multimedia Tool Appl. 2024;83(16):49389–418.
- 12. Pandey SK, Janghel RR, Mishra PK, Ahirwal MK. Automated epilepsy seizure detection from EEG signal based on hybrid CNN and LSTM model. Signal Image Video Process. 2023;17(4):1113–22.
- 13. Qu R, Ji X, Wang S, Wang Z, Wang L, Yang X, et al. An Integrated Multi-Channel Deep Neural Network for Mesial Temporal Lobe Epilepsy Identification Using Multi-Modal Medical Data. Bioengineering (Basel). 2023;10(10):1234. pmid:37892964
- 14. Zhu Q, Yang J, Xu B, Hou Z, Sun L, Zhang D. Multimodal Brain Network Jointly Construction and Fusion for Diagnosis of Epilepsy. Front Neurosci. 2021;15:734711. pmid:34658773
- 15. Lucas A, Cornblath EJ, Sinha N, Caciagli L, Hadar P, Tranquille A, Davis KA. Improved seizure onset-zone lateralization in temporal lobe epilepsy using 7T resting-state fMRI: A direct comparison with 3T. medRxiv; 2023.
- 16. Chang AJ, Roth R, Bougioukli E, Ruber T, Keller SS, Drane DL, et al. MRI-based deep learning can discriminate between temporal lobe epilepsy, Alzheimer’s disease, and healthy controls. Commun Med. 2023;3(1):33.
- 17. Luckett PH, Maccotta L, Lee JJ, Park KY, U F Dosenbach N, Ances BM, et al. Deep learning resting state functional magnetic resonance imaging lateralization of temporal lobe epilepsy. Epilepsia. 2022;63(6):1542–52. pmid:35320587
- 18. Caldairou B, Foit NA, Mutti C, Fadaie F, Gill R, Lee HM, et al. MRI-Based Machine Learning Prediction Framework to Lateralize Hippocampal Sclerosis in Patients With Temporal Lobe Epilepsy. Neurology. 2021;97(16):e1583–93. pmid:34475125
- 19. Beheshti I, Sone D, Maikusa N, Kimura Y, Shigemoto Y, Sato N, et al. Accurate lateralization and classification of MRI-negative 18F-FDG-PET-positive temporal lobe epilepsy using double inversion recovery and machine-learning. Comput Biol Med. 2021;137:104805. pmid:34464851
- 20. Aslam S, Rajeshkannan R, Sandya CJ, Sarma M, Gopinath S, Pillai A. Statistical asymmetry analysis of volumetric MRI and FDG PET in temporal lobe epilepsy. Epilepsy Behav. 2022;134:108810. pmid:35802989
- 21. Ruowei Q, Shifen W, Zhengfang L, Junhua G, Guizhi X. 3D-CNN frameworks for mesial temporal lobe epilepsy diagnosis in MRI images. Int J Appl Electromagnet Mech. 2022;70(4):515–23.
- 22. Morita-Sherman M, Li M, Joseph B, Yasuda C, Vegh D, De Campos BM, et al. Incorporation of quantitative MRI in a model to predict temporal lobe epilepsy surgery outcome. Brain Commun. 2021;3(3):fcab164. pmid:34396113
- 23. Fu C, Aisikaer A, Chen Z, Yu Q, Yin J, Yang W. Different Functional Network Connectivity Patterns in Epilepsy: A Rest-State fMRI Study on Mesial Temporal Lobe Epilepsy and Benign Epilepsy With Centrotemporal Spike. Front Neurol. 2021;12:668856. pmid:34122313
- 24. Shi K, Pang X, Wang Y, Li C, Long Q, Zheng J. Altered interhemispheric functional homotopy and connectivity in temporal lobe epilepsy based on fMRI and multivariate pattern analysis. Neuroradiology. 2021;63(11):1873–82. pmid:33938990
- 25. Hadar PN, Kini LG, Nanga RPR, Shinohara RT, Chen SH, Shah P, et al. Volumetric glutamate imaging (GluCEST) using 7T MRI can lateralize nonlesional temporal lobe epilepsy: A preliminary study. Brain Behav. 2021;11(8):e02134. pmid:34255437
- 26. Mittal M, Goyal LM, Kaur S, Kaur I, Verma A, Jude Hemanth D. Deep learning based enhanced tumor segmentation approach for MR brain images. Applied Soft Computing. 2019;78:346–54.
- 27. Guillemaud R, Brady M. Estimating the bias field of MR images. IEEE Trans Med Imaging. 1997;16(3):238–51. pmid:9184886
- 28. Saboor A, Li JP, Ul Haq A, Shehzad U, Khan S, Aotaibi RM, et al. DDFC: deep learning approach for deep feature extraction and classification of brain tumors using magnetic resonance imaging in E-healthcare system. Sci Rep. 2024;14(1):6425. pmid:38494517
- 29. Geem D, Hercules D, Pelia RS, Venkateswaran S, Griffiths A, Noe JD, et al. Progression of Pediatric Crohn’s Disease Is Associated With Anti-Tumor Necrosis Factor Timing and Body Mass Index Z-Score Normalization. Clin Gastroenterol Hepatol. 2024;22(2):368–76.e4. pmid:37802268