Figures
Abstract
Machine learning (ML) offers great potential in healthcare, especially in the analysis of complex physiological signals like electroencephalography (EEG). EEG recordings hold valuable insights into neurological function and can aid in diagnosing various conditions. In this work, we explore the use of a Conditional Variational Autoencoder (CVAE) that injects a binary health label (healthy or orthopedic impairment) into both the encoder input and the latent space coupled with the extracted features, leveraging the conditional input vector to learn representations specific to different health conditions. Our study involved using two public OpenNeuro datasets [1,2]. From the healthy dataset we randomly selected seven subjects to match the seven impaired participants; in both sets we retained the same 11 scalp channels (C3, Cz, C4, FC3, FCz, FC4, CP3, CPz, CP4, F3, F4). Six descriptors-Short time Fourier Transform (STFT), Hurst Exponent (HE), Detrended Fluctuation Analysis (DFA), Correlation Dimension (CD), Kolmogorov-Sinai Entropy (permutation entropy; KS-proxy), and the Largest Lyapunov Exponent (LLE)-are extracted channel-wise and concatenated to form the input feature vector, which distills distinct characteristics from the EEG signals. We rigorously evaluated the performance of our CVAE model in combination with each feature extraction technique. The conditional supply of class labels to both encoder and decoder enabled the CVAE to achieve 93% accuracy on the unseen test split of the dataset with precision of 93%, a recall of 93%, and an F1-score of 0.93 outperforming re-trained CNN baselines. These results highlight the promise of CVAEs and the significance of well-suited feature extraction for robust EEG classification. This work could contribute to the development of automated healthcare diagnostic tools.
Citation: Tibermacine IE, Russo S, Scarano G, Tedesco G, Rabehi A, Alhussan AA, et al. (2025) Conditional VAE for personalized neurofeedback in cognitive training. PLoS One 20(10): e0335364. https://doi.org/10.1371/journal.pone.0335364
Editor: Zeheng Wang, Commonwealth Scientific and Industrial Research Organisation, AUSTRALIA
Received: August 15, 2025; Accepted: October 9, 2025; Published: October 31, 2025
Copyright: © 2025 Tibermacine et al. This is an open access article distributed under the terms of the Creative Commons Attribution License, which permits unrestricted use, distribution, and reproduction in any medium, provided the original author and source are credited.
Data Availability: All raw EEG recordings analyzed in this study are publicly available on OpenNeuro. The orthopedic-impairment dataset is “Multimodal EEG and fNIRS Biosignal Acquisition during Motor Imagery Tasks in Patients with Orthopedic Impairment” (OpenNeuro accession ds004022, DOI 10.18112/openneuro.ds004022.v1.0.0). The healthy cohort is “EEG Motor Movement/Imagery Dataset” (OpenNeuro accession ds004362, DOI 10.18112/openneuro.ds004362.v1.0.0). No new datasets were generated. Derived per-window feature matrices and train/test splits can be recreated from the raw datasets using the preprocessing and feature-extraction parameters reported in the Materials and Methods; these derived files are not deposited separately.
Funding: Princess Nourah bint Abdulrahman University Researchers Supporting Project number (PNURSP2024R754), Princess Nourah bint Abdulrahman University, Riyadh, Saudi Arabia.
Competing interests: The authors have declared that no competing interests exist.
1 Introduction
EEG provides a non-invasive method to observe the complex electrical activity of the brain, establishing itself as a fundamental tool in diagnosing and studying various neurological conditions such as epilepsy, sleep disorders, and cognitive impairments [3]. Traditional EEG analysis relies heavily on the subjective interpretation by clinicians, but the advent of machine learning techniques, particularly through brain-computer interfaces (BCI), has transformed this process [4,5]. BCIs leverage EEG signals to facilitate interaction between humans and digital or robotic devices, enhancing not only medical diagnostics but also the quality of life for individuals with cognitive impairments through technologies such as gesture control and eye tracking [6].
The primary functionality of BCIs is predicated on the accurate detection of specific brain activities that translate into computer commands [6,7]. This application is prominently executed using EEG due to its non-invasive advantages, where users perform Motor Imagery (MI) tasks [8,9] to generate recognizable EEG patterns [7,10]. MI involves imagining motor movements without actual execution, activating specific cerebral cortex regions and leading to changes in EEG rhythm signals [11], such as event-related desynchronization (ERD) and synchronization (ERS) across mu and beta wave spectra [12,13].
However, EEG signals are inherently noisy and non-stationary, with significant variability between individuals due to physiological differences [14–16]. This variability presents a considerable challenge in processing EEG signals, requiring effective pre-processing to improve signal quality and robust feature extraction methods [17,18]. The transition from traditional BCI methodologies to more advanced implementations integrates not only handling the inherent noise and variability in EEG data, but also improving the interpretability and usability of EEG-based commands in real-world applications [19–21].
Deep learning models, particularly CVAEs, have emerged as powerful tools in this context, especially in the realm of EEG classification within BCIs [22,23]. CVAEs excel in learning complex high-dimensional representations from EEG data, facilitating both the classification and reconstruction of EEG signals [9,24]. Using CVAEs, it is possible to improve the accuracy and efficiency of BCIs, allowing a more precise interpretation of user intentions from EEG signals [25,26]. The ability of CVAEs to handle both classification and reconstruction tasks provides valuable insight into the underlying patterns of brain activity, pushing the limits of what can be achieved with EEG data in both medical and technological applications [27–29].
EEG signals have been extensively studied using a wide array of analytical methods rooted in information theory [30], nonlinear dynamics [31], fractal theory [32], and complex systems techniques [33–36]. The inherently non-linear nature of EEG signals makes them particularly suitable for analysis using nonlinear dynamics methods, including techniques such as Hurst exponents [37], Lyapunov exponent [38], entropy measures [39], and fractal dimensions [40]. These methods have been instrumental in distinguishing between EEG signals from individuals with neurological conditions such as epileptic seizures and those from healthy individuals [41,42].
The potential of machine learning in healthcare care is transformative, especially in the analysis of complex physiological signals such as EEG, which holds significant promise for the diagnosis and understanding of neurological conditions. Traditional machine learning techniques, such as Support Vector Machines (SVMs) [43,44] and decision trees [45], have shown early success in EEG classification, demonstrating the capabilities of automated analysis. However, the complex and high-dimensional nature of EEG signals poses challenges that often exceed the capabilities of traditional ML approaches, driving the adoption of more powerful deep learning models [28].
Feature extraction is critical in EEG analysis as it distills complex and noisy signals into representations that highlight salient characteristics suitable for machine learning models [46]. Various techniques contribute to this process: The Hurst Exponent quantifies long-range dependencies in time series, providing information on the persistence of EEG signal, while the detrended fluctuation analysis measures scaling properties and fractal-like patterns, revealing the underlying dynamics [47,48]. The correlation dimension assesses the complexity and non-linear dynamics of brain signals, and the Kolmogorov-Sinai entropy estimates the complexity and irregularity of the signal, reflecting potential differences in the brain activity states. The Short-time Fourier Transform analysis decomposes a signal into overlapping segments analyzed using the Fourier transform, providing a time-frequency distribution that elucidates the evolving spectral characteristics [49,50].
Deep learning, particularly through the use of Convolutional Neural Networks (CNNs) [51] and Recurrent Neural Networks (RNNs), has outperformed traditional machine learning techniques in various EEG classification tasks [52,53]. The impressive capabilities of Deep Convolutional Neural Networks (DCNNs) [54] in EEG-based attention decoding tasks demonstrate their ability to analyze complex EEG signals and extract meaningful patterns related to attention, speech perception, and noise suppression. These networks are specially designed to process the complex spatial-temporal patterns characteristic of multi-channel EEG recordings [55,56]. To address the challenges of limited datasets, data augmentation using CVAEs has been shown to improve model performance significantly [57].
Variational Autoencoders (VAEs) are potent in feature extraction from EEG data, capable of modeling complex distributions and uncovering underlying neurological patterns that may indicate subtle differences related to conditions such as obesity [58,59]. The integration of CVAEs in our EEG analysis framework is based on established encoder-decoder principles. The encoder component of a CVAE transforms the complex, high-dimensional EEG input into a compact latent representation, which captures essential patterns relevant to the classification task [60]. By conditioning the CVAE on health status labels (‘Healthy’ or ‘Unhealthy’), the model learns to discriminate between these conditions effectively. The decoder then attempts to reconstruct the original EEG signal from the latent representation, allowing the model to address dual objectives: minimizing reconstruction loss to ensure accurate signal regeneration and enhancing classification accuracy by producing features that distinguish between different EEG patterns [61,62]. The integration of EEG with advanced deep learning techniques through BCIs represents a significant advancement in neurotechnology [63,64]. This synergy not only enhances diagnostic capabilities and interaction modalities in medical and technological fields but also paves the way for future innovations in understanding and interfacing with human brain dynamics [65,66].
However, the success of Deep Learning models hinges on the quality of the input data. Raw EEG signals often contain noise and complex dynamics, making direct classification challenging. Feature extraction techniques play a pivotal role in transforming raw EEG signals into meaningful representations that highlight relevant characteristics, thus improving model performance.
This paper explores the synergy between various feature extraction methods and a CVAE model for robust EEG classification. Our primary research questions include:
- Feature Optimization: Which feature extraction method, among the Hurst Exponent, DFA, CD, STFT, Kolmogorov-Sinai Entropy, and their combinations, yields the most informative representations for EEG classification?
- CVAE Reconstruction: To what extent can this novel model accurately reconstruct EEG signals, and how does this ability relate to its classification performance?
- CVAE Performances: How does the CVAE’s performance compare to existing state-of-the-art approaches for EEG classification?
Methodologically, the present investigation utilizes EEG datasets that concentrate on motor imagery tasks, employing sophisticated feature extraction techniques such as the Hurst Exponent, Detrended Fluctuation Analysis, Short-time Fourier Transform, Correlation Dimension, and Kolmogorov-Sinai Entropy. Subsequently, we train a Conditional Variational Autoencoder, which is predicated on both the extracted EEG features and health condition labels, to proficiently model and classify the underlying neural patterns. A comprehensive preprocessing pipeline guarantees data consistency and reliability, encompassing noise elimination, signal filtration, and data normalization to improve the quality of the subsequent analysis.
This study constitutes the first motor-imagery EEG framework that embeds a binary health-condition label simultaneously in the encoder input and in the latent code of a conditional variational auto-encoder trained on a physiologically grounded, hand-crafted feature stack. This dual conditioning unifies interpretable signal descriptors with modern generative modeling, producing a compact representation that surpasses current discriminative pipelines and establishes a new benchmark for diagnosis-oriented EEG analysis.
The principal contribution of our research resides in illustrating the CVAE’s capacity to acquire meaningful and discriminative representations from EEG signals, in addition to its potential for precise signal reconstruction. These outcomes not only enhance our comprehension of how CVAEs can be optimally employed for EEG analysis but also establish a new standard for the automated diagnosis of neurological disorders utilizing EEG data. Our endeavor creates a significant reference framework that may inform the future advancement of EEG-based diagnostic tools, thereby improving their accuracy and applicability within clinical settings.
2 Materials and methods
2.1 Dataset
This study leverages two public EEG-MI datasets with distinct cohorts and acquisition setups. We analyze only EEG signals (ignoring auxiliary modalities), convert all recordings to a common montage, and harmonize sampling and preprocessing across datasets.
- Healthy dataset [2]. EEG Motor Movement/Imagery: 109 volunteers, 64-channel 10–10 montage, original sampling 160 Hz (EDF+). Each subject performed multiple MI runs comprising left/right hand, both fists, and feet tasks (executed and imagined). We used only MI runs (baselines discarded).
- Orthopedic-impairment dataset [1]. Multimodal MI in patients with orthopedic impairment: 7 participants, EEG montage with 18 channels, original sampling 500 Hz, MI trials recorded in multiple sessions. We use only the EEG stream for analysis.
Subject selection and class definition: To achieve parity in participant counts, we selected 7 subjects from the healthy dataset to match the impaired cohort; where age/sex metadata were available, we preferentially chose healthy subjects to approximate the clinical cohort’s demographics. We define a binary label for healthy (0) vs. orthopedic impairment (1). All per-window labels inherit the subject’s group label (task identity is used only as an auxiliary input where stated elsewhere).
Common montage: According to the channel-description files, both datasets share an identical subset of 11 electrodes: C3, Cz, C4, FC3, FCz, FC4, CP3, CPz, CP4, F3, F4. All subsequent analyses are confined to this montage; no interpolation or channel dropping is required.
Signal harmonization: All recordings are re-referenced as in their source datasets, notch-filtered, and band-pass filtered from 8–50 Hz using a zero-phase, fourth-order Butterworth design (filtfilt). Signals are then resampled to fs = 125 Hz with zero-phase FIR decimation so that each 1 s epoch contains exactly 125 samples. To ensure consistency across datasets, only channels in the common montage are retained before resampling.
Segmentation and event handling: Following filtering, continuous recordings are segmented into non-overlapping 1 s windows (125 samples). For the healthy dataset we exclude baseline (eyes-open/closed) runs and retain only MI runs; for the impaired dataset we retain all MI trials. Trial/event markers are used solely to isolate MI segments; we do not stratify by MI task for the main binary classification unless otherwise stated.
Quality control and exclusion: We remove windows with missing event annotations and windows with non-physiological amplitudes (V after re-referencing). Remaining windows are z-scored per recording (mean and standard deviation computed from training data only to avoid leakage).
Balancing windows across groups: After selecting subjects, residual imbalances in the number of MI windows across groups and subjects are addressed in two steps: (i) per-subject stratified undersampling to equalize window counts across subjects within each group; and (ii) if a residual class imbalance
remains, we apply inverse-frequency weights in the classification loss,
where N is the total number of training windows and Nc is the number from class c. In practice, after step (i) the classes are near-balanced and weights remain close to 1; we do not use extreme weights (the earlier placeholder is removed as a typographical artifact).
The final analysis set contains an equal number of participants per group (7 healthy, 7 impaired), a standardized 11-channel montage, and uniformly preprocessed EEG at 125 Hz segmented into 1 s windows. Dataset-specific acquisition details and accession information are given in [1,2]; our evaluation protocol (stratified 80/20 split and metrics) is described later.
Covariates sensitivity: EEG differs between individuals for reasons that are not related to the disease status, including age- and sex-related effects on rhythm power, peak frequencies, and long-range dependence. The two public datasets used here were curated independently and provide incomplete demographic metadata; consequently, a formal covariate-controlled analysis (e.g., regression or matched subsampling) cannot be reported without introducing unverifiable assumptions or discarding large portions of data.
Several design choices in the present study reduce—but do not eliminate—covariate influence: (i) a fixed stratified 80/20 split (subject proportions preserved) maintains class balance but does not enforce subject exclusivity; (ii) per-recording z-scoring removes absolute amplitude differences that can correlate with age/sex; (iii) a fixed 11-channel common montage avoids spatial-coverage disparities; and (iv) participant counts are balanced across cohorts. These steps improve comparability at the signal and sampling levels, yet unmeasured covariates (age, sex, medication, sleep, cap fit) may still contribute residual variance. Reported group differences should therefore be interpreted as condition-linked patterns within the recorded cohorts, not as causal effects independent of demographics. A covariate-adjusted evaluation requires complete metadata and is outside the scope of the current re-analysis of public datasets.
2.2 Preprocessing and enhancements
2.2.1 Filtering:.
Raw EEG was notch-filtered at the dataset-appropriate mains frequency (50 or 60 Hz), then band-pass filtered between 8 and 50 Hz using a zero-phase, fourth-order Butterworth design (filtfilt). To suppress edge transients, signals were mirrored by 3 s at both ends before zero-phase application. The upper cut-off preserves γ-band activity (30–50 Hz), which indexes motor preparation and sensorimotor integration during motor-imagery tasks [67,68], while attenuating higher-frequency myogenic artifacts. (Resampling to 125 Hz follows the dataset harmonization in Sect 2.1; anti-aliasing low-pass was applied prior to decimation.)
2.2.2 Segmentation:.
Following filtering, recordings were partitioned into non-overlapping 1 s windows (125 samples at Hz). Windows were drawn only from MI intervals indicated by event markers; baseline/rest periods were excluded. Any window that straddled an event boundary (onset/offset) was discarded to avoid label contamination. The 1 s duration is a canonical compromise for MI decoding and for estimating long-range dependency measures (Hurst, DFA) on short segments; preliminary tests with 0.5 s and 2 s windows yielded no consistent gains, while 1 s aligns with the temporal resolution of the γ-band STFT features used downstream.
2.2.3 Channel montage:.
All recordings were restricted to the 11 common electrodes listed earlier (C3, Cz, C4, FC3, FCz, FC4, CP3, CPz, CP4, F3, F4); no interpolation or channel exclusion was necessary, ensuring an identical spatial representation across datasets. Re-referencing followed the convention of each source dataset prior to reduction to the common montage.
2.2.4 Normalization:.
Each channel was -scored on a per-recording basis (mean and standard deviation computed from the training split only and then applied to validation/test windows of the same recording) to remove inter-subject amplitude variability and prevent train–test leakage. Normalization precedes all feature computations.
2.3 Feature extraction
All features are computed per 1 s window and per channel on z-scored data (mean and standard deviation estimated on the training split only to avoid leakage). Unless stated otherwise, logarithms are natural and constants are fixed across all experiments.
Short-Time Fourier Transform (STFT) band power: Each 1 s window has N = 125 samples at fs = 125 Hz. We compute the magnitude STFT with a Hann window of length Nw = 64 samples and 75% overlap (hop h = 16), yielding frames per window. The DFT size is
(zero-padding as needed). Let X(k,m) be the STFT at bin k and frame m, with bin frequency
. Define band index sets
Log-power per band uses a stabilizer :
Hurst exponent (, rescaled range): For scales
(50% overlap), we compute the classical rescaled-range statistic. Given a demeaned segment
, form the cumulative deviate series
(
); the range is
and the standard deviation
. The segment statistic is
. Averaging over the segments for each s, we fit
via OLS; the shortest scale is omitted when fewer than two cycles of the dominant rhythm fall within s.
Detrended Fluctuation Analysis (DFA): For each 1 s window, build the integrated profile , partition into
non-overlapping segments of length
, fit and subtract a least-squares linear trend Ys(k) within each segment, then compute
The DFA exponent is the slope of
vs.
(OLS on valid scales).
KolmEn (permutation/ordinal entropy; KS proxy): We estimate complexity via permutation (ordinal) entropy as a finite-sample proxy for the KS entropy rate. Using embedding dimension m = 3 and delay , let
be the
ordinal patterns and
the empirical frequency of pattern
within the 1 s window. The KolmEn feature is
computed without normalization.
Correlation dimension (CD): With time-delay embedding and Theiler window
to exclude temporal neighbours, we compute the Grassberger–Procaccia correlation sum
where are embedded vectors. Radii r are log-spaced over
, with
and
set to the 5th and 95th percentiles of pairwise distances. The CD is the slope of
vs.
over the largest contiguous scaling region selected by maximal R2 (plateau of local slopes).
Largest Lyapunov exponent (LLE): Using Rosenstein’s method with the same embedding and Theiler window W, for each state choose its nearest neighbor
with
. Track separations
and average the log-divergence
The LLE is the slope of over a short linear regime
; we use
,
samples at
Hz, chosen to maximize linear fit R2 while avoiding saturation.
Feature vector: Per channel, we concatenate
(8 scalars). Stacking across the 11-channel montage yields features per 1 s window. All logs use the natural base; features are computed on the z-scored signals using training-set statistics.
2.4 Feature–selection rationale
We retain six descriptors because they (i) capture complementary neurophysiological phenomena across spectral, statistical, and nonlinear dynamics and (ii) are fast enough for real-time use on 1 s windows ( ms per channel in our setup). Long-range dependency measures—the Hurst exponent and detrended-fluctuation analysis (DFA)—quantify temporal persistence modulated by motor preparation [69]. KolmEn (permutation/ordinal entropy; Eq 2.3) and the largest Lyapunov exponent capture irregularity and local instability, both implicated in pathological sensorimotor reorganization [70]. Correlation dimension characterizes the fractal geometry of the reconstructed EEG attractor, providing a compact index of dynamical complexity that varies with cortical pathology during motor tasks. Finally, the Short-Time Fourier Transform (STFT) summarizes power in the μ (8–13 Hz), β (13–30 Hz), and task-relevant γ (30–50 Hz) rhythms; under low-latency constraints it offers a robust time–frequency portrait for MI decoding [67,68,71].
Together, these six descriptors yield 8 scalars per channel (3 STFT band-powers Hurst
DFA
KolmEn
Correlation dimension
Lyapunov exponent), balancing interpretability and computational tractability for on-device BCI deployment. Exact parameter choices are fixed a priori and reported in Table 1 and Sect 2.3.
2.4.1 Embedding choices for nonlinear features.
Following standard guidance [72], we adopt distinct embeddings for ordinal-pattern and state-space methods: (i) KolmEn uses embedding dimension and delay
(short-window setting with a favourable bias–variance trade-off at
Hz; Eq 2.3); (ii) Correlation dimension (Grassberger–Procaccia) and Lyapunov exponent (Rosenstein) use
with a Theiler window
to exclude temporal neighbors; scaling/linear regions are selected by the plateau of local slopes [73]. These settings yield stable estimates on 1 s MI windows without exceeding a phase-space dimension of 10.
2.5 Evaluation protocol
Splits and leakage control: All results are reported on a stratified 80/20 train–test split at the window level (stratified by class; subject proportions preserved). This preserves class balance but does not enforce subject exclusivity; we therefore interpret generalization as within-cohort and note that a subject-exclusive split would be more conservative.
Metrics: We report accuracy, macro–F1, precision, and recall. Where two models are compared, we report their scores on the identical test split.
Baseline fairness: All baselines are retrained under the identical preprocessing pipeline (11-channel common montage, 1 s non-overlapping windows, per-recording z-scoring) and evaluated on the same stratified 80/20 split. CNN baselines use raw EEG windows; the CVAE uses the feature vector from Sect 2.3. Hyperparameter grids and best settings are listed in Table 2.
2.6 Additional experiments
We systematically evaluated the performance of our CVAE model in combination with various feature extraction techniques mentioned previously.
However, upon closer examination, we observe nuanced variations in the loss values across different feature extraction techniques and VAE settings. Particularly intriguing is the performance of the model with the Hurst feature extraction in a β-VAE setting. This configuration exhibited slightly lower losses compared to other feature extraction methods. This is also confirmed by other metrics such as Precision, Recall and F1-Score.
The Hurst feature extraction method captures the long-range dependence structure inherent in EEG signals, providing rich information about their fractal nature. When combined with the β-VAE framework, which encourages disentangled and interpretable latent representations, the Hurst feature appears to enhance the model’s ability to learn discriminative features while minimizing reconstruction and classification losses.
To further validate the robustness of our findings and explore the model’s performance across different feature domains, we conducted a final experiment utilizing features derived from STFT analysis.
For the STFT analysis, we specifically focused on three frequency bands of particular relevance to motor tasks:
- Alpha Band (8-13 Hz): This band is often associated with relaxed wakefulness and is known to be modulated during movement preparation and execution.
- Beta Band (13-30 Hz): Beta activity is linked to active movement, motor control, and sensorimotor integration.
- Gamma Band (30-50 Hz): Gamma oscillations are thought to play a role in higher-order cognitive processes and complex motor coordination.
We retrained our CVAE model on the STFT features computed, using the same pre-processing and enhancement techniques applied to the other feature sets. This experiment provides further evidence of the model’s ability to generalize across diverse feature representations and reinforces the consistency of our findings across different analysis approaches. The results of this STFT-based analysis, including model performance metrics and comparative evaluations with other feature extraction pipelines, are detailed later in the results section.
2.7 Motor task and channel filtering
In addition to the main 11-channel analysis, we performed a compact ablation study that isolates the sensorimotor hotspot C3 and retains only left- and right-hand motor-imagery epochs. After the usual pre-processing steps, the resulting class distribution was balanced with a k-nearest–neighbor synthetic minority over-sampling (KNN-SMOTE via imbalanced-learn). The same CVAE backbone, with a reduced latent size of 8, attained 93.0 % accuracy while maintaining a low reconstruction error (MSE < 0.03). The confusion matrices confirm high true-positive and true-negative rates, indicating that even a single well-placed electrode captures sufficient discriminatory information. However, we stress that all the headline results reported in Sects 4 and 5 rely on the full 11-channel montage and inverse-frequency class weighting; the single-channel study is included only to illustrate the robustness of the model under extreme reduction in dimensionality.
3 Proposed models
This section introduces a Conditional Variational Auto-Encoder (CVAE) tailored to EEG data; the health label (healthy or unhealthy) is supplied to both encoder and decoder, whereas a standard VAE would remain label-agnostic. The detailed architecture is illustrated in Figs 1, 2, and 3.
The encoder maps to
, the decoder reconstructs
from
, and the classifier operates on z.
Reparameterization yields z as in Eq (9).
The proposed model has three objectives. (i) Accurately reconstruct and, when required, synthesize realistic EEG epochs, enabling data augmentation and simulation in BCI pipelines. (ii) Leverage the conditional branch to classify each input epoch as healthy or unhealthy, providing a diagnostic read-out. (iii) Concatenate the binary label to the latent representation, thereby forcing the CVAE to encode class information explicitly, which promotes a more separable latent space and facilitates condition-specific signal generation.
3.1 Conditional variational autoencoder
We model the per-window EEG feature vector (
, see Sect 2.3) conditioned on the observed health label
. The conditional generative model is
with an amortized variational posterior
Decoder conditioning is implemented by concatenating y to z, i.e., .
3.1.1 Conditional ELBO and training objective.
Maximizing admits the conditional ELBO
Because x are z-scored continuous features, we use a homoscedastic Gaussian likelihood
i.e., the mean-squared error (MSE) per feature dimension.
The decoder reconstructs the 88-dimensional feature vector rather than raw EEG time series. All reconstruction losses are computed as feature-wise MSE (Eq 4). Feature-domain reconstruction quality is directly relevant to classification because (i) the classifier operates on the latent z inferred from these features, and (ii) lower feature-space reconstruction error typically coincides with more class-separable latent embeddings under the conditional objective (Eq 5).
In addition, we learn an auxiliary discriminative head trained with cross-entropy. The final objective is the negative conditional ELBO with a β-weight on the divergence term and a classification regularizer:
where is either KL or MMD (see below),
,
, and
unless otherwise stated.
Closed-form KL for diagonal Gaussians: For and
,
Parameterizing yields convenient gradients:
Maximum Mean Discrepancy (MMD): When with a Gaussian kernel
, we use the unbiased finite-sample estimator on a minibatch
and
:
with bandwidth set by the median heuristic. (A multi-kernel sum
is also supported but not used in the main results.)
3.1.2 Architecture and parameterization.
We implement conditional distributions with MLPs and dual concatenation of y.
Inputs (feature space): with
(11 channels
8 features/channel).
Encoder . We form
and apply two linear layers with ReLU:
followed by linear heads to with
. Sampling uses the reparameterization trick
Decoder : We concatenate again in latent space:
, then apply
producing in (4).
Auxiliary classifier : A linear head on z outputs logits in
:
Shapes:
3.1.3 Concatenation strategy and generation.
We follow a dual concatenation protocol (cf. [77]): for inference and
for generation. At test time, label-controlled synthesis uses
3.2 Back-propagation and optimization
We minimize (5) with Adam (,
,
), batch size 128, for 50 epochs. Gradients flow through the stochastic path via (9). With KL, the total gradient decomposes as
When using MMD, in (11) is replaced by the gradient of (8), which has closed-form derivatives through the kernel.
(i) We do not use KL-annealing; is fixed (see Tables 3 and 4). (ii) The reconstruction term is mean MSE per feature to harmonize scale with the divergence and classification terms. (iii) The prior is class-agnostic
in all main results; a class-conditional prior
is possible but not used here.
4 Results
We report implementation-oriented metrics on a workstation with an RTX 3060 GPU (forward pass) and CPU-based feature extraction. The CVAE is deliberately compact: with the specified MLP layers (encoder with two 32-d heads; decoder
; classifier
), the total parameter count is 18,938, corresponding to ≈0.076 MB in FP32. Feature extraction dominates latency: the previously profiled per-channel cost (≤2 ms) yields an upper bound of ≤22 ms per 11-channel window. The CVAE forward pass adds a sub-millisecond cost on commodity GPUs, so end-to-end per-window latency is ≈23–25 ms for the 1 s window configuration. This corresponds conservatively to ∼40–45 windows/s sustained throughput, with negligible VRAM demand for the model itself and batch sizes used here.
The CVAE model achieved high performance on the EEG classification task across both training and testing sets. After approximately 50 epochs, the model reached impressive accuracies, with final accuracies consistently between 90-94%. A stratified 80/20 train–test partition (window-level; stratified by class, subject proportions preserved) is used here, following EEGNet [74].
The seven healthy volunteers completed twelve motor-imagery runs of 2 min each (24 min total), providing non-overlapping 1-s epochs per subject—10 080 in aggregate. Each orthopedic-impairment participant contributed 8 min of motor-imagery data, yielding
epochs per subject and 3 360 epochs overall. To prevent class imbalance, an equal number of healthy epochs (3 360) was selected by random sampling, resulting in 6 720 artifact-free epochs (3 360 healthy + 3 360 unhealthy). A stratified 80/20 partition allocated 5 376 epochs to training and 1 344 to testing, with subject and class proportions preserved in both splits.
This demonstrates the effectiveness of the CVAE in learning meaningful representations from EEG signals and its ability to accurately classify neurological conditions. Injecting the health condition label directly into the model’s input and latent representation allows the model to learn features that are specifically tuned for health status classification, leading to more accurate and robust predictions.
As shown in Tables 5 and 6, this project achieved impressive total objective losses, with mean classification losses and mean losses varying slightly but consistently performing well across different feature extraction methods. Notably, the results were strong for both standard KL divergence and Beta-VAE settings, demonstrating the robustness and flexibility of the CVAE model in handling different types of EEG feature representations.
To further illustrate the CVAE’s performance, Fig 4 present the training and testing loss/accuracy plots. These plots visually demonstrate the model’s rapid convergence and stability in achieving exceptionally high classification accuracy. Additionally, Figs 5 and 6 provide illustrative time-domain visualizations derived post hoc from the feature-space reconstructions (the model’s training target is the 88-D feature vector). These are shown for qualitative context only.
True-positive and true-negative rates remain high despite the single-channel reduction.
The generated sample (orange) is obtained by decoding with
. Note: the model reconstructs 88-D feature vectors; time-domain traces are illustrative only.
These findings are highly promising and suggest that the developed CVAE has the potential to be a valuable tool for the automated diagnosis of neurological conditions using EEG data.
5 Discussion and analysis
5.1 Overview of results
The results of our study strongly demonstrate the effectiveness of the CVAE in accurately classifying EEG signals for distinguishing between healthy and unhealthy brain activity. The CVAE model, when integrated with various advanced feature extraction techniques (results in Table 7), consistently exhibited high classification accuracy. Notably, the optimal performance was achieved when the Hurst Exponent was employed as the feature extraction method in conjunction with the β-VAE configuration, yielding a classification accuracy of 93.1% on the test set. These findings underscore the potential of the CVAE framework to enhance the accuracy of EEG-based diagnostic systems.
5.2 Impact of feature extraction techniques
The choice of feature extraction method significantly influenced the performance of the CVAE model, as evidenced by the comparative analysis presented in Tables 5 and 6. Among the methods evaluated, the Hurst Exponent emerged as the most informative feature for EEG classification. This exponent, which captures the long-range dependencies and fractal characteristics of EEG signals, provided the most robust input for the CVAE, resulting in the highest observed classification accuracy.
The superior performance of the Hurst Exponent can be attributed to its ability to effectively characterize the temporal structure of EEG signals, which is critical in distinguishing between healthy and pathological states. In contrast, other feature extraction methods such as DFA and CD, while still valuable, did not achieve the same level of accuracy, likely due to their focus on different aspects of EEG signal dynamics. These findings suggest that the Hurst Exponent may capture more relevant features for the specific task of binary health classification in EEG data.
Previous CVAE applications to EEG fall into two categories: (i) unconditioned VAEs that model raw time series for artifact removal [78], and (ii) task-conditioned CVAEs that rely exclusively on deep convolutional encoders without explicit feature engineering [79]. To our knowledge, the present work is the first to (a) inject a binary health label directly into both the encoder input and the latent code, (b) operate on a compact, physiologically motivated feature stack (Hurst, DFA, KS-H entropy, Lyapunov exponent, STFT), and (c) demonstrate that this hybrid, hand-crafted–plus–generative approach yields state-of-the-art accuracy on motor-imagery data while retaining interpretability of individual feature contributions.
5.3 Efficacy of the CVAE architecture
The architecture of CVAE, as illustrated in Figs 1, 2, and 3, played a pivotal role in the success of the model. The inclusion of health condition labels directly into the latent space via concatenation allowed the CVAE to learn more discriminative features, which was reflected in the model’s high performance across various feature extraction techniques. This innovative approach of embedding condition-specific information into both the input and latent representation enabled the CVAE to effectively differentiate between healthy and unhealthy EEG signals.
The training curves depicted in Fig 4 demonstrate the model’s rapid convergence and stable training process, with a clear reduction in loss and consistent improvement in accuracy over approximately 50 epochs. The minimal difference between training and testing losses indicates that the model generalizes well to unseen data, thereby reducing the risk of overfitting.
5.4 Statistical analysis
Because all methods are evaluated on the same test set, we quantify the robustness of the observed performance gap using uncertainty estimates and a conservative significance test. We report 95% Wilson confidence intervals for accuracy (test size windows) and, where available, macro–F1. As paired per-window predictions are not retained in the manuscript, we provide a conservative two-proportion z-test (unpaired, identical n) comparing each baseline to the CVAE; this test understates the power relative to a paired analysis (e.g., McNemar) and thus yields a cautious inference. As summarized in Table 8, the CVAE’s improvements remain highly significant under this conservative analysis.
Accuracy shown with 95% Wilson CIs. Two-proportion z-test compares each baseline to the CVAE (two-sided).
5.5 Reconstruction capabilities and signal generation
The CVAE’s ability to accurately reconstruct EEG signals, as demonstrated in Figs 5 and 6, is indicative of its potential not only for classification but also for generating realistic synthetic EEG data. The reconstructed signals closely match the true EEG signals, and the generated signals from Gaussian noise exhibit similar characteristics to unhealthy signals. This suggests that the CVAE has effectively learned a meaningful latent representation of the EEG data, which could be leveraged for data augmentation or for simulating EEG signals in the development of advanced BCI systems.
5.6 Analysis of confusion matrices
The confusion matrices shown in Fig 7 for both the training and test sets provide further insights into the robustness of the CVAE model. The high number of true positives and true negatives across both datasets underscores the model’s effectiveness in accurately classifying EEG signals. The low incidence of false positives and false negatives is particularly noteworthy, as it highlights the model’s reliability in real-world applications where mis-classification could have serious implications.
The training target is the 88-D feature vector; time-domain plots are for qualitative context.
However, despite the high overall accuracy, the presence of a small number of misclassifications suggests that there is still room for improvement. These errors could arise from the inherent variability in the EEG signals or from subtle differences that are not fully captured by current feature extraction methods. Future work could explore the integration of more sophisticated or hybrid feature extraction techniques to further enhance the accuracy of the model.
5.7 Comparative analysis
When compared to existing state-of-the-art approaches, the CVAE model demonstrates competitive performance, particularly in the context of EEG signal classification. The model’s integration of condition-specific information into the latent space is a key innovation that enhances its ability to learn and generalize from EEG data. This positions the CVAE as a valuable tool for advancing the field of automated EEG analysis.
However, it is important to recognize that further optimization of the model is possible. For instance, investigating more complex architectures, such as those utilizing attention mechanisms or graph-based neural networks, could potentially offer further improvements in both classification accuracy and model interpretability. Such advancements would not only enhance the CVAE’s performance but also broaden its applicability to more complex and varied EEG datasets.
Because the present study prioritizes generative utility, we nonetheless retrained the canonical discriminative baselines on the harmonized data set using exactly the same pre-processing pipeline (11 common electrodes, 1 s non-overlapping epochs, per-recording z scoring) and the stratified 80-20 train–test split adopted for CVAE. As shown in Table 9, the retraining yielded accuracies of 85.4% for EEGNet [74], 86.1% for DeepConvNet [74], 82.6% for the Riemannian geometry SVM [76], and 87.0% for the TCN transformer [75]. Under the identical protocol, the class-conditioned CVAE attains 93.1%, exceeding the strongest convolutional baseline by roughly 6% and the remaining methods by 7–10%. This controlled comparison confirms that embedding the health label in both the encoder input and the latent representation—coupled with a compact, physiologically grounded feature stack—can outperform deeper, parameter- intensive architectures while simultaneously providing label-conditioned generative capability.
The class-conditioned CVAE not only eliminates the performance gap with state-of-the-art, end-to-end CNN architectures but in fact surpasses their best re-trained scores on the identical dataset—despite relying on a markedly lighter network and an explicitly interpretable, hand-crafted feature stack. These results substantiate two central claims: (i) supplying the health label to both the encoder input and the latent code decisively improves class separability, and (ii) a hybrid hand-crafted–feature + generative paradigm can equal or out-perform substantially deeper, parameter-heavy discriminative pipelines.
Accordingly, the proposed CVAE establishes a rigorous baseline for diagnosis-oriented EEG analysis while simultaneously offering the unique advantage of label-conditioned signal synthesis for data augmentation, uncertainty estimation, and latent-space probing.
Our principal evaluation adopts a stratified window-level split that preserves class balance and subject proportions but does not enforce subject exclusivity. While standard and informative for within-cohort generalization, such splits can be optimistic if subject-specific structure is shared across sets. For deployment-oriented scenarios, a subject-independent protocol—e.g., leave-one-subject-out (LOSO)—offers a stricter and more conservative bound on across-subject generalization and is a natural extension of the present work.
6 Limitations and future research directions
Although the results of this study are promising, it is crucial to acknowledge certain limitations. The computational demands of the CVAE, particularly in terms of GPU memory usage and training time, were substantial. As a result, there may be trade-offs between model complexity and computational feasibility, especially when scaling to larger datasets or more complex architectures.
Furthermore, the data sets used in this study were relatively clean and well pre-processed, which may not fully reflect the challenges posed by real-world EEG data, which often contain significant noise and artifacts. Future research should focus on improving the robustness of the CVAE to noisy data, possibly through the adoption of more sophisticated preprocessing techniques or noise-resilient architectures.
Furthermore, while this study focused on a binary classification task (healthy vs. unhealthy), expanding the model to handle multi-class classification tasks or more nuanced health conditions would be a valuable next step. Moreover, the use of motor task labels as conditions for the CVAE represents an exciting avenue for future research, particularly in the context of developing personalized BCIs.
7 Conclusion
This research has demonstrated the exceptional potential of the Conditional Variational Autoencoder coupled with advanced feature extraction techniques for accurate EEG signal classification. Our systematic evaluation of various feature extraction methods, including Hurst Exponent, DFA, CD, and Kolmogorov-Sinai Entropy, showed that all tested methods enabled the CVAE to achieve remarkable classification accuracies on the unseen test set. These findings underscore the CVAE’s ability to learn meaningful and discriminative representations from EEG data when combined with well-suited feature extraction techniques.
Despite these promising results, it is essential to acknowledge the limitations and potential challenges that can arise in more complex scenarios. Our experiments highlighted the significant computational demands of the CVAE model, particularly in terms of GPU VRAM usage and training time, especially when using very large models or complex architectures. These computational challenges necessitated a careful balance between model complexity and hardware capabilities.
Furthermore, the remarkable efficacy of the CVAE model can be partially ascribed to the superior quality of the datasets used, which were comparatively unblemished and meticulously preprocessed. In contrast, authentic EEG data frequently encompass a multitude of noise sources and artifacts that may introduce additional complexities for precise signal analysis and classification. Moreover, the present implementation concentrated on a relatively straightforward binary classification task. As the number of categories or the complexity of the classification tasks escalates, further enhancement and refinement of the model may be warranted.
The CVAE’s reconstruction capability also suggests its potential for EEG signal augmentation and simulation, which could have valuable applications in research and the development of BCIs. Looking ahead, an innovative direction for this research could involve assembling a larger and more diverse EEG data set with a wider range of neurological conditions to further validate the generalizability of our CVAE model. This would allow us to explore how to optimize feature extraction methods to enhance performance while preserving dataset size to the greatest extent possible.
Moreover, transitioning from the binary healthy/unhealthy condition to utilizing motor task labels as conditions for the CVAE Encoder/Latent Variable presents novel avenues for exploration. We have already structured our model framework to easily integrate the motor task condition, which would enable us to generate signals conditioned on specific motor tasks, thereby opening up new, unexplored insights into EEG-based classification and simulation.
References
- 1.
Lee S, Jung HR, Wang IN, Jung MK, Kim H, Kim DJ. Multimodal EEG and fNIRS biosignal acquisition during motor imagery tasks in patients with orthopedic impairment. 2022.
- 2.
Schalk G, McFarland DJ, Hinterberger T, Birbaumer N, Wolpaw JR. EEG motor movement/imagery dataset. 2022.
- 3. Yang YJ, Jeon EJ, Kim JS, Chung CK. Characterization of kinesthetic motor imagery compared with visual motor imageries. Sci Rep. 2021;11(1):3751. pmid:33580093
- 4.
Tibermacine IE, Tibermacine A, Guettala W, NAPOLI C, Russo S. Enhancing sentiment analysis on SEED-IV dataset with vision transformers: a comparative study. In: Proceedings of the 2023 11th International Conference on Information Technology: IoT and Smart City. 2023. p. 238–46. https://doi.org/10.1145/3638985.3639024
- 5. Djemal R, Bazyed AG, Belwafi K, Gannouni S, Kaaniche W. Three-class EEG-based motor imagery classification using phase-space reconstruction technique. Brain Sci. 2016;6(3):36. pmid:27563927
- 6. Grimm F, Walter A, Spüler M, Naros G, Rosenstiel W, Gharabaghi A. Hybrid neuroprosthesis for the upper limb: combining brain-controlled neuromuscular stimulation with a multi-joint arm exoskeleton. Front Neurosci. 2016;10:367. pmid:27555805
- 7. Ferrero L, Quiles V, Ortiz M, Iáñez E, Azorín JM. A BMI based on motor imagery and attention for commanding a lower-limb robotic exoskeleton: a case study. Applied Sciences. 2021;11(9):4106.
- 8. Pfurtscheller G, Neuper C. Motor imagery and direct brain-computer communication. Proc IEEE. 2001;89(7):1123–34.
- 9. Russo S, Tibermacine IE, Tibermacine A, Chebana D, Nahili A, Starczewscki J, et al. Analyzing EEG patterns in young adults exposed to different acrophobia levels: a VR study. Front Hum Neurosci. 2024;18:1348154. pmid:38770396
- 10. Hwang J, Park S, Chi J. Improving multi-class motor imagery EEG classification using overlapping sliding window and deep learning model. Electronics. 2023;12(5):1186.
- 11.
Tibermacine IE, Russo S, Tibermacine A, Rabehi A, Nail B, Kadri K. Riemannian geometry-based eeg approaches: a literature review. 2024. https://doi.org/arXiv:240720250
- 12. Guerrero-Mendez CD, Blanco-Diaz CF, Ruiz-Olaya AF, López-Delis A, Jaramillo-Isaza S, Milanezi Andrade R, et al. EEG motor imagery classification using deep learning approaches in naïve BCI users. Biomed Phys Eng Express. 2023;9(4):10.1088/2057-1976/acde82. pmid:37321179
- 13. Pfurtscheller G, Neuper C, Flotzinger D, Pregenzer M. EEG-based discrimination between imagination of right and left hand movement. Electroencephalogr Clin Neurophysiol. 1997;103(6):642–51. pmid:9546492
- 14. Tibermacine IE, Russo S, Citeroni F, Mancini G, Rabehi A, Alharbi AH, et al. Adversarial denoising of EEG signals: a comparative analysis of standard GAN and WGAN-GP approaches. Front Hum Neurosci. 2025;19:1583342. pmid:40395688
- 15. Wang X, Lu H, Shen X, Ma L, Wang Y. Prosthetic control system based on motor imagery. Comput Methods Biomech Biomed Engin. 2022;25(7):764–71. pmid:34533381
- 16. Tibermacine A, Guettala W, Tibermacine IE. Efficient one-stage deep learning for text detection in scene images. Electrotehnica, Electronica, Automatica. 2024;72(4).
- 17. Chaudhary U, Birbaumer N, Ramos-Murguialday A. Brain-computer interfaces for communication and rehabilitation. Nat Rev Neurol. 2016;12(9):513–25. pmid:27539560
- 18. Rimbert S, Gayraud N, Bougrain L, Clerc M, Fleck S. Can a subjective questionnaire be used as brain-computer interface performance predictor?. Front Hum Neurosci. 2019;12:529. pmid:30728772
- 19. Padfield N, Zabalza J, Zhao H, Masero V, Ren J. EEG-based brain-computer interfaces using motor-imagery: techniques and challenges. Sensors (Basel). 2019;19(6):1423. pmid:30909489
- 20. Qian L, Ge X, Feng Z, Wang S, Yuan J, Pan Y, et al. Brain network reorganization during visual search task revealed by a network analysis of fixation-related potential. IEEE Trans Neural Syst Rehabil Eng. 2023;31:1219–29. pmid:37022804
- 21.
Tibermacine A, Tibermacine IE, Zouai M, Rabehi A. EEG classification using contrastive learning and riemannian tangent space representations. In: 2024 International Conference on Telecommunications and Intelligent Systems (ICTIS). 2024. p. 1–7. https://doi.org/10.1109/ictis62692.2024.10894645
- 22. Boutarfaia N, Russo S, Tibermacine A, Tibermacine IE. Deep learning for EEG-based motor imagery classification: towards enhanced human-machine interaction and assistive robotics. Life. 2023;2(3):4.
- 23.
Izsó L. The significance of cognitive infocommunications in developing assistive technologies for people with non-standard cognitive characteristics: CogInfoCom for people with non-standard cognitive characteristics. In: 2015 6th IEEE International Conference on Cognitive Infocommunications (CogInfoCom). 2015. p. 77–82.
- 24. Naidji I, Tibermacine A, Tibermacine IE, Russo S, Napoli C. EGDN-KL: dynamic graph–deviation network for EEG anomaly detection. Biomedical Signal Processing and Control. 2026;112:108597.
- 25.
Stephygraph LR, Arunkumar N. Brain-actuated wireless mobile robot control through an adaptive human–machine interface. In: Proceedings of the International Conference on Soft Computing Systems: ICSCS 2015. 2016. p. 537–49.
- 26.
Tibermacine A, Akrour D, Khamar R, Tibermacine IE, Rabehi A. Comparative analysis of SVM and CNN classifiers for EEG signal classification in response to different auditory stimuli. In: 2024 International Conference on Telecommunications and Intelligent Systems (ICTIS). 2024. p. 1–8. https://doi.org/10.1109/ictis62692.2024.10894292
- 27.
Naidji I, Tibermacine A, Guettala W, Tibermacine IE. Semi-mind controlled robots based on reinforcement learning for indoor application. In: ICYRIME. 2023. p. 51–9.
- 28. Ladjal B, Tibermacine IE, Bechouat M, Sedraoui M, Napoli C, Rabehi A, et al. Hybrid models for direct normal irradiance forecasting: a case study of Ghardaia zone (Algeria). Nat Hazards. 2024;120(15):14703–25.
- 29. Pfurtscheller G, Neuper C. Motor imagery activates primary sensorimotor area in humans. Neurosci Lett. 1997;239(2–3):65–8. pmid:9469657
- 30. Gao J, Hu J, Tung W-W, Cao Y, Sarshar N, Roychowdhury VP. Assessment of long-range correlation in time series: how to avoid pitfalls. Phys Rev E Stat Nonlin Soft Matter Phys. 2006;73(1 Pt 2):016117. pmid:16486226
- 31. Hwa RC, Ferree TC. Scaling properties of fluctuations in the human electroencephalogram. Phys Rev E Stat Nonlin Soft Matter Phys. 2002;66(2 Pt 1):021901. pmid:12241208
- 32. Gao J, Hu J, Tung W-W. Complexity measures of brain wave dynamics. Cogn Neurodyn. 2011;5(2):171–82. pmid:22654989
- 33. Gierałtowski J, Żebrowski JJ, Baranowski R. Multiscale multifractal analysis of heart rate variability recordings with a large number of occurrences of arrhythmia. Phys Rev E Stat Nonlin Soft Matter Phys. 2012;85(2 Pt 1):021915. pmid:22463252
- 34. Berthouze L, Farmer SF. Adaptive time-varying detrended fluctuation analysis. J Neurosci Methods. 2012;209(1):178–88. pmid:22677174
- 35. Karkare S, Saha G, Bhattacharya J. Investigating long-range correlation properties in EEG during complex cognitive tasks. Chaos, Solitons & Fractals. 2009;42(4):2067–73.
- 36. Hu K, Ivanov PC, Chen Z, Carpena P, Stanley HE. Effect of trends on detrended fluctuation analysis. Phys Rev E Stat Nonlin Soft Matter Phys. 2001;64(1 Pt 1):011114. pmid:11461232
- 37. Preissl H, Lutzenberger W, Pulvermüller F, Birbaumer N. Fractal dimensions of short EEG time series in humans. Neurosci Lett. 1997;225(2):77–80. pmid:9147378
- 38. Jeong J, Kim SY, Han SH. Non-linear dynamical analysis of the EEG in Alzheimer’s disease with optimal embedding dimension. Electroencephalogr Clin Neurophysiol. 1998;106(3):220–8. pmid:9743280
- 39. Bein B. Entropy. Best Pract Res Clin Anaesthesiol. 2006;20(1):101–9. pmid:16634417
- 40. Weiss B, Clemens Z, Bódizs R, Vágó Z, Halász P. Spatio-temporal analysis of monofractal and multifractal properties of the human sleep EEG. J Neurosci Methods. 2009;185(1):116–24. pmid:19646476
- 41.
Russo S, Ahmed S, Tibermacine IE, Napoli C. Enhancing EEG signal reconstruction in cross-domain adaptation using CycleGAN. In: 2024 International Conference on Telecommunications and Intelligent Systems (ICTIS). 2024. p. 1–8. https://doi.org/10.1109/ictis62692.2024.10894543
- 42. Plummer C, Harvey AS, Cook M. EEG source localization in focal epilepsy: where are we now?. Epilepsia. 2008;49(2):201–18. pmid:17941844
- 43. Li S, Zhou W, Yuan Q, Geng S, Cai D. Feature extraction and recognition of ictal EEG using EMD and SVM. Comput Biol Med. 2013;43(7):807–16. pmid:23746721
- 44.
Tibermacine IE, Tibermacine A, Zouai M, Russo S, Bouchelaghem S, Napoli C. Enhanced EEG classification via Riemannian normalizing flows and deep neural networks. In: 2025 International Symposium on iNnovative Informatics of Biskra (ISNIB). 2025. p. 1–6. https://doi.org/10.1109/isnib64820.2025.10982792
- 45. Duan L, Ge H, Ma W, Miao J. EEG feature selection method based on decision tree. Biomed Mater Eng. 2015;26 Suppl 1:S1019-25. pmid:26405856
- 46.
Abdalsalam M. E, Yusoff MZuki, Kamel N, Malik A, Meselhy M. Mental task motor imagery classifications for noninvasive brain computer interface. In: 2014 5th International Conference on Intelligent and Advanced Systems (ICIAS). 2014. p. 1–5. https://doi.org/10.1109/icias.2014.6869531
- 47. Madan S, Srivastava K, Sharmila A, Mahalakshmi P. A case study on discrete wavelet transform based hurst exponent for epilepsy detection. J Med Eng Technol. 2018;42(1):9–17. pmid:29188743
- 48.
Napoli C, Ponzi V, Puglisi A, Russo S, Tibermacine I. Exploiting robots as healthcare resources for epidemics management and support caregivers. In: CEUR Workshop Proceedings. 2024. p. 1–10.
- 49. Ting W, Guo-zheng Y, Bang-hua Y, Hong S. EEG feature extraction based on wavelet packet decomposition for brain computer interface. Measurement. 2008;41(6):618–25.
- 50. Wolpaw JR, Birbaumer N, McFarland DJ, Pfurtscheller G, Vaughan TM. Brain-computer interfaces for communication and control. Clin Neurophysiol. 2002;113(6):767–91. pmid:12048038
- 51.
Bouchelaghem S, Tibermacine IE, Balsi M, Moroni M, Napoli C. Cross-domain machine learning approaches using hyperspectral imaging for plastics litter detection. In: 2024 IEEE Mediterranean and Middle-East Geoscience and Remote Sensing Symposium (M2GARSS). 2024. p. 36–40. https://doi.org/10.1109/m2garss57310.2024.10537535
- 52. Roy Y, Banville H, Albuquerque I, Gramfort A, Falk TH, Faubert J. Deep learning-based electroencephalography analysis: a systematic review. J Neural Eng. 2019;16(5):051001. pmid:31151119
- 53. Shenoy P, Krauledat M, Blankertz B, Rao RP, Muller KR. Towards adaptive classification for BCI. Journal of Neural Engineering. 2006;3(1):R13.
- 54. Kaiser V, Bauernfeind G, Kreilinger A, Kaufmann T, Kübler A, Neuper C, et al. Cortical effects of user training in a motor imagery based brain-computer interface measured by fNIRS and EEG. Neuroimage. 2014;85 Pt 1:432–44. pmid:23651839
- 55.
Katona J, Ujbanyi T, Sziladi G, Kovari A. Speed control of festo robotino mobile robot using NeuroSky MindWave EEG headset based brain-computer interface. In: 2016 7th IEEE International Conference on Cognitive Infocommunications (CogInfoCom). 2016. p. 251–6. https://doi.org/10.1109/coginfocom.2016.7804557
- 56. Nail B, Benrabeh D, Tibermacne IE, Napoli C, Nabil H, Abdelaziz R. Gas turbine vibration monitoring based on real data using neuro-fuzzy system. Diagnostyka. 2024;25(1):1–9.
- 57.
TanveerMA. Deep convolution neural network for attention decoding in multi-channel EEG with conditional variational autoencoder for data augmentation. 2023.
- 58. de Vries IEJ, van Driel J, Olivers CNL. Posterior α EEG dynamics dissociate current from future goals in working memory-guided visual search. J Neurosci. 2017;37(6):1591–603. pmid:28069918
- 59. Amprimo G, Rechichi I, Ferraris C, Olmo G. Measuring brain activation patterns from raw single-channel EEG during exergaming: a pilot study. Electronics. 2023;12(3):623.
- 60. Wolpaw JR, Birbaumer N, Heetderks WJ, McFarland DJ, Peckham PH, Schalk G, et al. Brain-computer interface technology: a review of the first international meeting. IEEE Trans Rehabil Eng. 2000;8(2):164–73. pmid:10896178
- 61.
Yue Y, Deng JD, Ridder DD, Manning P, Adhia D. Variational autoencoder learns better feature representations for EEG-based obesity classification. 2023.
- 62. Eisapour M, Cao S, Domenicucci L, Boger J. Virtual reality exergames for people living with dementia based on exercise therapy best practices. Proceedings of the Human Factors and Ergonomics Society Annual Meeting. 2018;62(1):528–32.
- 63. Dai G, Zhou J, Huang J, Wang N. HS-CNN: a CNN with hybrid convolution scale for EEG motor imagery classification. J Neural Eng. 2020;17(1):016025. pmid:31476743
- 64. Katona J. A review of human–computer interaction and virtual reality research fields in cognitive InfoCommunications. Applied Sciences. 2021;11(6):2646.
- 65. Liu Y, Yu Y, Ye Z, Li M, Zhang Y, Zhou Z, et al. Fusion of spatial, temporal, and spectral EEG signatures improves multilevel cognitive load prediction. IEEE Trans Human-Mach Syst. 2023;53(2):357–66.
- 66.
Markopoulos E, Lauronen J, Luimula M, Lehto P, Laukkanen S. Maritime Safety Education with VR Technology (MarSEVR). In: 2019 10th IEEE International Conference on Cognitive Infocommunications (CogInfoCom). 2019. p. 283–8. https://doi.org/10.1109/coginfocom47531.2019.9089997
- 67. Muthukumaraswamy SD. Functional properties of human primary motor cortex gamma oscillations. J Neurophysiol. 2010;104(5):2873–85. pmid:20884762
- 68. Donoghue T, Voytek B. Automated meta-analysis of the event-related potential (ERP) literature. Sci Rep. 2022;12(1):1867. pmid:35115622
- 69. Finnigan S, O’Connell RG, Cummins TDR, Broughton M, Robertson IH. ERP measures indicate both attention and working memory encoding decrements in aging. Psychophysiology. 2011;48(5):601–11. pmid:21039584
- 70. Bandt C, Pompe B. Permutation entropy: a natural complexity measure for time series. Physical Review Letters. 2002;88(17):174102.
- 71. Park H-D, Piton T, Kannape OA, Duncan NW, Lee K-Y, Lane TJ, et al. Breathing is coupled with voluntary initiation of mental imagery. Neuroimage. 2022;264:119685. pmid:36252914
- 72. Kantz H, Ragwitz M. Phase space reconstruction and nonlinear predictions for stationary and nonstationary markovian processes. Int J Bifurcation Chaos. 2004;14(06):1935–45.
- 73. Grassberger P. Grassberger-Procaccia algorithm. Scholarpedia. 2007;2(5):3043.
- 74. Lawhern VJ, Solon AJ, Waytowich NR, Gordon SM, Hung CP, Lance BJ. EEGNet: a compact convolutional neural network for EEG-based brain-computer interfaces. J Neural Eng. 2018;15(5):056013. pmid:29932424
- 75.
Ingolfsson TM, Hersche M, Wang X, Kobayashi N, Cavigelli L, Benini L. EEG-TCNet: an accurate temporal convolutional network for embedded motor-imagery brain–machine interfaces. In: 2020 IEEE International Conference on Systems, Man, and Cybernetics (SMC). 2020. https://doi.org/10.1109/smc42975.2020.9283028
- 76. Barachant A, Bonnet S, Congedo M, Jutten C. Multiclass brain-computer interface classification by Riemannian geometry. IEEE Trans Biomed Eng. 2012;59(4):920–8. pmid:22010143
- 77.
Ribeiro TFR, Silva F, de C Costa RL. Reconstructing spatiotemporal data with C-VAEs. Springer; 2023. p. 59–73. http://dx.doi.org/10.1007/978-3-031-42914-9_5
- 78. Li Y, Liu A, Yin J, Li C, Chen X. A segmentation-denoising network for artifact removal from single-channel EEG. IEEE Sensors J. 2023;23(13):15115–27.
- 79. Rong G, Xu Y, Sawan M. Machine learning techniques for effective pathogen detection based on resonant biosensors. Biosensors (Basel). 2023;13(9):860. pmid:37754094