Figures
Abstract
The diagnosis of Parkinson’s disease relies heavily on the subjective assessment of physicians, which depends on their individual experience and training, potentially leading to inconsistent diagnostic results. Therefore, developing an objective and efficient diagnostic method is essential to improve the accuracy and timeliness of Parkinson’s disease diagnosis. In this study, we utilized the PhysioNet dataset, a time-series dataset comprising data from 93 Parkinson’s patients and 73 healthy individuals. The dataset contains vertical ground reaction forces recorded from 16 sensors (8 per foot) during a 2-minute test at a sampling rate of 100 Hz. To address challenges such as limited dataset size, high labeling noise, and high intra-class variability, we performed data preprocessing and applied various data augmentation techniques, including jittering, scaling, rotation, permutation, magnitude warping, time warping, cropping, and linear residuals. These methods were evaluated using one-dimensional-convolutional neural network (1D-ConvNet) and one-dimensional Transformer networks. By conducting 10-fold cross-validation, we observed significant improvements in classification performance. The best data augmentation strategy achieved 90.8% accuracy, 92.0% precision, 91.0% recall, and a 91.0% F1 score in assessing disease severity. These results highlight the importance of selecting appropriate data augmentation techniques for time-series data to improve model generalization and diagnostic reliability, while also offering new insights for researchers working with sensor device data. Our results demonstrate that data-enhanced methods can significantly boost the performance of machine-learning models in the field of Parkinson’s disease diagnosis.
Citation: Huo H, Jiao S, Li D, Ma L, Xu N (2025) Efficient quantification of Parkinson’s disease severity using augmented time-series data. PLoS ONE 20(4): e0319826. https://doi.org/10.1371/journal.pone.0319826
Editor: Luigi Borzì, Politecnico di Torino, ITALY
Received: October 29, 2024; Accepted: February 7, 2025; Published: April 2, 2025
Copyright: © 2025 Huo et al. This is an open access article distributed under the terms of the Creative Commons Attribution License, which permits unrestricted use, distribution, and reproduction in any medium, provided the original author and source are credited.
Data Availability: All the data can be accessed at https://physionet.org/content/ gaitpdb/1.0.0/.
Funding: The author(s) received no specific funding for this work.
Competing interests: The authors have declared that no competing interests exist.
1 Introduction
Wearables are smart devices worn on the human body that utilize sensors and communication technologies to monitor and collect data on the user’s physiology, movement, environment, and other related aspects. Parkinson’s disease (PD) is one of the most common neurodegenerative diseases, typically diagnosed through clinical evaluation and neuroimaging techniques [1]. However, traditional diagnostic methods often suffer from subjectivity and delays, making early diagnosis challenging. Wearable sensors [2] such as accelerometers, gyroscopes, pressure sensors, magnetometers, and electrophysiological sensors can capture large amounts of motion data and continuously monitor the patient’s condition without real-time intervention from doctors. These devices can detect subtle changes in movement without the patient’s direct awareness, [3,4] providing strong support for the early diagnosis of PD. With the advancement of science and technology, the development of high-sensitivity wearable devices has further promoted research into PD [5]. Traditional PD diagnosis relies on subjective assessment, whereas the combination of deep learning (DL) and wearable sensors offers new avenues for objective and accurate diagnostic and monitoring tools [3]. Digital biomarkers have shown significant potential in assessing disease severity and monitoring treatment outcomes, with sensor technologies greatly improving the precision of disease monitoring [6]. Wearable devices have enhanced the timeliness and accuracy of monitoring, while digital technologies show great promise in disease tracking, personalized treatment, and remote management [7]. Additionally, digital biomarkers hold broad prospects in quantifying symptom severity and evaluating treatment effectiveness [8]. Since PD is a progressive condition, its symptoms are often subtle in the early stages. Automated severity estimation can help detect changes in the condition early, especially when symptoms worsen. Inertial data, such as accelerometer and gyroscope readings, is typically used to measure overall movement changes, but it cannot provide detailed information about the specific contact between the foot and the ground. Therefore, pressure data from the feet offers more direct and detailed feedback [9] in gait analysis.
However, despite the progress brought by digital biomarkers and wearable technologies, challenges remain, such as sample and methodological heterogeneity and the lack of public datasets [10]. Current major challenges include insufficient data standardization, small sample sizes, algorithm interpretability, and device consistency [3]. Another challenge is to design fast and accurate modeling algorithms to fine-tune the analysis of data generated by wearable devices, especially when dealing with the correlation between different sensors and timing considerations. Therefore, our research focuses on how to utilize the limited resources of plantar pressure time-series data acquired by the sensor for Data Augmentation (DA) to obtain more usable data for gait analysis, especially in the context of grading the severity of PD. Augmentation methods for image data [11] (e.g., brightness adjustment, panning, flipping, etc.) have been widely used in computer vision, and these methods improve the training effect and generalization ability of the model by increasing the diversity and quantity of data. However, similar DA methods have been applied less in sensor-acquired time series data, even though these methods could theoretically improve the performance of time series models by introducing diversity as well. This imbalance may be due to the unique continuity and dependency of time series data, making traditional image augmentation methods difficult to apply directly. Consequently, it is important to explore and develop augmentation methods applicable to time series data, especially in areas that require high accuracy and robustness, such as health monitoring and disease diagnosis.
We aim to explore DA methods for footwear time series data to provide more reliable input for the model. Another challenge is creating simplified, effective network models. Current models are complex, and we seek ways to reduce their complexity while maintaining accuracy. Simplifying the model reduces computation and training time, improving efficiency in real-world applications. By optimizing the model structure and reducing parameters, we hope to maintain good performance with plantar pressure data, which is important for early Parkinson’s diagnosis.CNNs [12] capture local features but struggle with time series data. In contrast, Transformers [13] handle long-range dependencies well and perform better with long sequences. The multi-head attention mechanism enhances efficiency, allowing the model to handle larger datasets. Therefore, we chose Deep CNN and Transformer models to process plantar pressure data and validated the DA method through experiments.
Our contributions in this paper are mainly as follows:
- DA: For the time series data acquired by wearable devices, we used multiple DA methods based on a limited dataset to verify its validity in a single or combined form.
- Feature Selection: We performed feature selection for plantar data sensors to facilitate later feature selection, thereby reducing unnecessary data information and minimizing model redundancy.
- Excellent Performance: By introducing a new DA method to process the plantar data of Parkinson’s patients and normal people, we achieve efficient PD severity grading. Experiments have demonstrated that our proposed method performs excellently in terms of performance and achieves significant results.
- Simplified Model: We verified that a single model can also achieve optimal results through DA techniques. This finding not only reduces the complexity of the model, but also simplifies the overall architecture, improving operational efficiency and practicality.
The structure of our paper is as follows: Sect 2 reviews the latest advancements and related work in DL for PD research. Sect 3 introduces the DA methods and experimental models used in our study. Sect 4 presents and analyzes the experimental results. Finally, Sect 5 summarizes the strengths and weaknesses of our work and outlines future research directions.
2 Related works
2.1 Wearable sensors in health monitoring
In recent years, the impact of load measurement, training adaptation, and injury prevention strategies in sports has been the subject of extensive research, particularly the impact of monitoring and analyzing training loads in athletes on optimizing performance, reducing injury risk, and aiding in the recovery process [14,15]. Many studies have focused on using wearable devices for real-time monitoring and assessment of gait and tremor in PD patients. For example, study [16] developed a smartphone-based measurement method to assess the severity of PD, symptom fluctuations, and the response to dopaminergic treatment. Using a smartphone application, patients remotely completed five tasks: voice, finger tapping, gait, balance, and reaction time. Bamberg et al. [17] developed an long-lasting wearable system “Gaitshoe” and used it to provide quantitative gait analysis beyond the scope of a traditional sports laboratory. GaitShoe was analyzed and validated to have a strong ability to detect heel strike and toe-off, as well as estimate foot orientation and position. In [18], wearable devices (such as accelerometers, gyroscopes, and other sensors) were used to collect patients’ motion data, particularly gait and tremor symptoms. DL methods, such as CNN and Recurrent Neural Networks (RNN), were applied to analyze the collected sensor data. The DL models are capable of extracting important features from the raw data, and subsequently predicting the severity of the patient’s condition. By using DL models to process motion data in real-time, this approach enables the monitoring and assessment of PD severity in patients. The study [19] used 16 biomedical speech features as the input dataset, with the total UPDRS (Unified Parkinson’s Disease Rating Scale is a tool used to assess the severity of PD symptoms. It helps doctors evaluate the patient’s motor skills, daily functions, and other symptoms to monitor the disease’s progression) score as the output variable. The classification accuracy for the training and test sets were 94.44% and 62.73%, respectively.
Although these methods show potential in assessing the severity of PD, they face various challenges, including reliance on data quality, individual differences, and computational resources. Therefore, expanding the dataset has become a key factor in improving the effectiveness of these methods.
2.2 Foot pressure analysis in Parkinson’s disease
The feasibility and applicability of objective gait measurements based on wearable sensors in PD have been demonstrated to reach a relatively mature processing technique in both large-scale clinical studies and individual patient care. Schlachetzki et al. [20] analyzed a cross-sectional study of 190 patients with PD and 101 age-matched patients and revealed significant spatiotemporal gait parameter differences and corresponding their results to physician scores successfully demonstrated the clinical applicability of wearable sensor-based gait analysis, providing high biomechanical resolution for gait disorders in PD.
With the development of machine learning (ML), researchers have also designed a variety of different models to process Parkinson’s plantar time series data, which are designed to improve the accuracy of detection and prediction of the disease, thus helping patients to get better treatment and management. Among these models, CNN, Transformer, and LSTM have been widely used and studied due to their powerful ability to process time series data and capture complex patterns. Imanne El Maachi et al. [21] used a 1D-ConvNet to construct a DNN classifier. The model processes 18 one-dimensional signals of vertical ground reaction forces (VGRF) measured from foot sensors. Their algorithm validated the model on the UPDRS grading scale, which is the first model to utilize the UPDRS scale to classify the severity of PD. Due to the need for large datasets in DL algorithms, the study divides each walking session into segments of 100 time steps, with a 50% overlap between segments. Each segment is labeled with the corresponding class for training. The segmentation process is carried out within each fold, and segments from the same subject are not shared between the training and validation sets. The final classification accuracy for disease severity is 85.3%, with a total of 64,468 segments in the dataset. Duc Minh Dimitri Nguyen et al. [22] devised a Transformer-based spatial-temporal network structure to process Parkinson’s plantar gait data, and proved them to be effective in extracting relevant features from one-dimensional signals extracting relevant features is effective. Structurally they decoupled temporal and spatial information to minimize the complexity of the model. Their architecture uses a temporal converter, a dimensionality reduction layer to reduce the dimensionality of the data, as well as a spatial converter, two fully connected layers, and an output layer for final prediction.
Moreover, the study [23] introduces a spatio-temporal capsule network to classify Parkinson’s disease severity using gait data. The experiment used a Ga dataset with dimensions 19×100, where 19 represents features and 100 represents time steps. The model achieved 100% accuracy for healthy individuals and 82% for patients with severity level 3. The model proposed by Aşuroğlu, Tunç, et al. [24] combines CNN and locally weighted random forests, using GRF sensors with a 100 Hz sampling rate. Data preprocessing removed the initial and final portions of the data and outliers were eliminated using a median filter, improving the accuracy of gait feature representation. The final model achieved a correlation coefficient of 0.897, root mean square error (RMSE) of 3.009, and root mean square (RMS) of 4.556. Shcherbak et al. [25] proposed a method based on wearable sensors and ML to distinguish between healthy controls and patients with PD stages 1 and 2. The study designed 11 common movements and collected data using commercial accelerometer, gyroscope, and magnetometer sensors from 113 participants. By applying ML techniques (feature extraction, dimensionality reduction, and classification) to analyze the data, the accuracy of early PD diagnosis was significantly improved. The best results showed F1 scores of 0.78 and 0.88 for PD stages 1 and 2, respectively. This study [26] combines ML methods to classify PD and essential tremor (ET) based on balance and gait features collected through wearable sensors. The ML model achieved an F1 score of 0.48 for PD classification, demonstrating the practicality of this approach in distinguishing different motor disorders. In study [27], 74 patients visited the laboratory every three months, completing a total of seven visits. Data on walking (2 minutes) and postural sway (30 seconds, eyes closed) were collected using six inertial measurement unit sensors. Based on this data, the MDS-UPDRS-III scale was estimated. The performance of each model was evaluated using the average RMSE on the validation set through five-fold cross-validation analysis. The results showed that the best-performing model achieved an average RMSE of 10.02. However, due to the small size of the Ga dataset, we believe the results of this study may be influenced by the limited data volume, thus lacking broader applicability to real-world scenarios.
Although existing studies have made progress in assessing PD severity using gait and motion data, challenges such as limited sample size, feature extraction constraints, model performance limitations, and individual variability remain. To address these issues, we experimented with eight DA methods to capture complex spatiotemporal features and enhance model generalization. Our study not only overcomes the limitations posed by small datasets but also provides new insights into time-series DA, laying a solid foundation for the practical application of PD severity assessment.
2.3 Time series data angmentation techniques
DA is an essential preprocessing step for achieving optimal performance in DL methods [12,28]. In image processing, common techniques include rotation, flipping, cropping, scaling, and adding noise [29]. In speech recognition, frequently used methods involve time warping, frequency masking, adding background noise, and pitch shifting [30]. These approaches effectively increase data diversity, enhancing the model’s generalization and robustness. However, the application of DA methods for time series is still relatively limited, primarily focusing on basic techniques such as time warping, jittering, and cropping. While these methods improve model performance to some extent, there remains significant room for development compared to image [11] and speech [19] processing. Therefore, exploring richer and more effective augmentation techniques for time series data is crucial for improving model performance. In order to better utilize sensor data, there is an urgent need to develop DA techniques that can handle complex temporal patterns and multivariate features to improve the generalization ability and robustness of models.
With the increasing application of generative adversarial networks in the field of image generation, some researchers have also focused the method on time series data. For example, Li et al. [31] proposed a Transformer-Based Time-Series Generative Adversarial Network (TTS-GAN), a generative model, a Transformer-based Generative Adversarial Network (GAN) model for synthesizing virtual data with the same length as the original time series. Subsequently, Yoon et al. [32] developed TimeGAN, a generative method for multivariate time series data. Through experimental validation, TimeGAN shows better results in maintaining data diversity and accuracy compared to other GAN. However, despite the advantages demonstrated by GAN networks in certain time series tasks, the core challenge in generating data is to ensure that the generator and discriminator reach a Nash equilibrium [33]. This equilibrium is sometimes difficult to achieve consistently in the presence of insufficient data, leading to inconsistent quality of data generated by the model. In addition, due to the complexity of the GAN training process, its stability is often difficult to ensure, making the generated virtual data may deviate from the desired results.
Window cropping or slicing is a time series DA technique. [34] proposed a method to generate new samples by extracting continuous subsequences from the original series. [35] compared it to cropping in computer vision, where adjustable-length slices are randomly extracted from the original series for classification. The labels of the slices are the same as those of the original series, and majority voting is used for classification during testing. A perturbation and ensemble-based method is proposed in [36], which generates new time series using Dynamic Time Warping (DTW) and performs ensemble via a weighted version of Barycentric Averaging (DBA) algorithm. This method improves classification performance on some UCR datasets. Alle et al. [37] designed LPGNet, which pre-processes time-series data with down-sampling, normalization, and linear predictive analysis, after which linear predictive residuals (LPRs) are used to extract discriminative patterns from gait recordings, and a 1D-ConvNet and depth-separable CNN is used to diagnose class recognition. However, these processes may lead to the loss of valuable information, especially when the data is noisy or the sampling frequency is low, and the performance of the model may be degraded. Terry T. Um et al. [38] studied wearable sensor data and proposed DA methods based on signal transformation. They used these augmentation methods in conjunction with CNN models to classify the motor states of PD patients. Experimental results show that appropriate DA improves the classification performance from 77.54% to 86.88%, validating the effectiveness of the DA methods. In study [39], which evaluated the impact of using HOAs (Hand Optimization Algorithms) to expand the training dataset and transferring weights from unsupervised pretraining on classification performance. The results showed that after expanding the dataset and pretraining on hand movement tasks, the model’s accuracy improved by 12.2%. This indicates that DA and pretraining effectively enhanced the model’s classification performance.
These methods have achieved certain successes in time series DA, but there are still some limitations. Techniques like DTW and DBA perform well on some datasets, but their applicability is narrow and cannot be widely used for complex or multivariate time series data. These methods are computationally expensive, require long training times, and may not be suitable for real-time or large-scale data applications. Meanwhile, many augmentation techniques have limited capabilities in modeling spatiotemporal features, failing to effectively capture the complex dynamic changes in the data. Finally, some methods are optimized for specific models and lack universality, resulting in significant performance variations across different models. Therefore, we aim to explore more flexible and diverse augmentation methods to better adapt to complex data structures and noisy environments, enhancing the model’s generalization ability and stability.
3 Methods
In our study, we focused on how to accurately assess the severity of PD in patients based on plantar sensor data. This study utilized a publicly available dataset that recorded VGRF data from subjects walking on a flat surface at their natural gait for approximately 2 minutes. These data were obtained from 16 sensors on the soles of the feet, and the output signals from each sensor were digitized and recorded at a sampling rate of 100 times per second. To fully characterize gait, two composite signals were recorded, each representing the sum of outputs from eight sensors on each foot. This approach ensures precise motion capture while providing a rich dataset for analysis. Using various DA methods, we significantly enhanced the model’s performance on a limited time series dataset. Our DA process is illustrated in Fig 1.The methods included Jittering, Scaling, Rotation, Permutation, Mag_Warping, Time_Warping, and Cropping. Additionally, we examined the impact of linear prediction residuals (LPR) at different scales. By comparing residuals across scales, we better evaluated their influence on model training and predictions. This study not only highlights the role of residuals in DA but also refines the analytical framework, ensuring the model’s robustness and accuracy. The code can be found in here.
3.1 DA methods
This paper introduces eight DA methods for wearable sensor data, laying a foundation for future research. These methods are based on seven independent techniques that can be flexibly combined to achieve DA. As shown in Fig 1, we employ seven independent methods for time series DA, which can be used in any combination. Below is a brief description of our specific implementation of DA:
- Jittering: Using a normal distribution to generate random noise, with a mean of 0 and a standard deviation controlled by the parameter sigma, the noise is added to the original data to produce the augmented data. Then, the function is applied to the data for augmentation, with sigma set to 2 to control the noise intensity.
- Scaling: By generating a random scaling factor that follows a normal distribution and applying it to the original data, each row of data is multiplied by this scaling factor. The mean of the scaling factor is 1, and the standard deviation is controlled by sigma, which defaults to 0.1, thus controlling the degree of variation when augmenting the data.
- Rotation: The average value of each feature group is calculated as the rotation axis vector, which is then normalized to a unit vector. A random rotation angle is generated, and based on this axis vector and angle, a rotation matrix is created. The rotation matrix is applied to each feature group to obtain the rotated data.
- Magnitude Warping: The function generates random control points in the feature dimension of the data (based on a normal distribution), and then uses these control points to generate random curves. These curves apply deformation to the feature data through interpolation. The parameter sigma controls the magnitude of the curve’s variation, while knot determines the number of control points for the curve.
- Permutation: The function first randomly divides the input data into multiple segments, with the number of segments determined by the nPerm parameter, and the minimum length of each segment controlled by the minSegLength parameter to ensure each segment is sufficiently long. Then, these segments are randomly shuffled, and the rearranged data is returned.
- Time Warping: The time steps of the data are distorted by generating random curves and applying cubic spline interpolation. First, random control points are generated for each feature, and cubic spline interpolation is used to construct curves based on these control points, with the variation of the control points controlled by the sigma parameter. Then, the cumulative sum of these curves is used to convert them into cumulative time steps, simulating different time deformations.
- Random Sampling: The function generates random time indices for each column of the data and ensures that the starting and ending time steps are valid. Then, the function uses these randomly generated time steps to resample the data, applying linear interpolation to realign the data to the new time steps.
Among these, each DA method has its unique advantages. Jittering, by adding random noise, improves the model’s generalization ability and reduces the risk of overfitting. Scaling helps the model adapt to data changes of different magnitudes, enhancing its ability to handle various data scales. Rotation increases data diversity by altering the time order of the data, enabling the model to better cope with sequence changes in time series. Magnitude warping enhances the model’s robustness by nonlinearly adjusting the data’s magnitude, allowing it to adapt to fluctuations of different amplitudes. Permutation increases data diversity, helping the model handle changes in the order of the data. Time warping improves the model’s ability to adapt to time variations by simulating changes in time scales. Random sampling enhances data diversity by selecting different segments from the time series, helping the model recognize and process different parts of the data.
The use of these methods can significantly enhance the effectiveness of model training, enabling it to handle more complex and variable time series data, and improving the generalization ability and robustness of the model. In addition, the DA method we have taken is LPR. This is a concept in signal processing and time series analysis where the signal is predicted by using a linear model and then analyzing the difference between the original signal and the predicted value. This method helps in capturing and understanding the fine structure in the signal. The LPR is the difference between the original signal and its estimated value using a linear prediction model. The linear prediction model estimates the future values of a signal based on its past values using the following formula:
where is the predicted value of the signal at time n, akis the linear prediction coefficient,
is the past value of the signal, and p is the order of the model. The LPR is the difference between the original signal and its predicted value. The residual formula is as follows:
where is the linear prediction residual at time,
is the value of the original signal at time. And
is the predicted value of the linear prediction model at time. In summary, the linear prediction residual formula can be expressed as:
shows that the LPR is the difference between the original signal and its estimated value by a linear prediction model. By analyzing the residuals, nonlinear components and anomalies in the signal can be revealed, which is important for signal processing, time series analysis, and various application areas. As shown in Fig 2, this is the visualization of our data after LPR processing. We first loaded the original gait data file and extracted the data using an extraction factor (decimationRate) of 4, 2 and 1, respectively.
Specifically, for each dataset, we first apply mean filtering and smoothing to the data from each sensor, then perform normalization and decimation. After that, we compute the predicted values using linear prediction coefficients and generate the prediction residual matrix.
3.2 Model based on CNN
Next, we verify the effectiveness of our DA using two classical network models respectively. Firstly, the detailed structural design of our 1D-ConvNet is shown in Fig 3. This 1D-ConvNet accepts input data of shape (100, 1), and passes sequentially through two convolutional layers (with 8 and 16 filters, respectively, and kernel size of 3, using the SELU activation function, as ), one max pooling layer (with a window size of 2), and then another two convolutional layers (with 16 and 1 filters, respectively, and kernel size of 3, using the SELU activation function) and one max pooling layer (with a window size of 2).
Subsequently, the multidimensional feature maps are flattened into one-dimensional vectors through a flattened layer, and finally, a 50-dimensional feature vector is output through a fully connected layer with 50 neurons (using the SELU activation function). The whole design aims to extract important features in the input data through convolution and pooling layers to simplify the dimensionality of the data and reduce the computational complexity, and integrate the features through a fully connected layer to generate the final feature representation. During the training process, to obtain more data, we segmented the data at the individual level, and each walk was divided into 100 time-step segments with 50% overlap, which also allowed us to adopt a majority voting mechanism through segment classification to ultimately determine the severity level of PD.
3.3 Model based on transformer
Model two (shown in Fig 4) begins with a variety of DAs and preprocessing of the data. Fig 5 shows the visualization results for one channel (the left foot total force channel as an example) that was processed with LPR. Next, the data was split into 18 (16 sensor channels + 2 additional channels) parallel inputs into the network. Each sensor channel is first processed through a time encoder to capture temporal dependencies. The specific design of our encoder is shown in Fig 6. First, the input data is processed through layer normalization, which helps ensure that each feature maintains consistent statistical properties during processing. Next, using the MultiHeadAttention mechanism, the encoder is able to simultaneously attend to information at different locations in the input data and perform a dropout operation after calculating the attention weights to enhance the generalization ability of the model. The attention output is then summed with the original input through residual connections, a step that helps mitigate the gradient vanishing problem and allows for a smoother transfer of information between different layers. After completing the attention mechanism, layer normalization is again performed to prepare the data for subsequent Dense layer processing. This feedforward network usually consists of fully connected layers containing activation functions for nonlinear transformation and feature extraction of the data. Eventually, the results of the feedforward processing are added to the input of the residual connection to obtain the output of the encoder. This design not only improves the performance and efficiency of the model, but also helps to capture complex relationships and features in the input data, thus laying the foundation for accurate execution of subsequent tasks.
This split-channel design maintains the independence of each channel and ensures that the data inputs obtained from different sensors are theoretically sound. After the data processing is completed, considering the importance of positional information in the plantar pressure sensors, we employ a spatial encoder design to further process and fuse the extracted feature data. This helps to capture and understand the deep information and complex patterns in the plantar pressure data, thus reflecting the dynamic changes of the plantar more accurately. Ultimately, by connecting two linear layers, we obtain the output of the multiclassification hierarchy.
It is worth mentioning that the DA processes (including data impact factors and sampling rates) in Model 1 and Model 2 are identical, ensuring that both models receive the same input data. However, in terms of preprocessing, Model 2 differs from Model 1 in that the encoder of Model 2 incorporates position encoding. By introducing position encoding, Model 2 is better able to capture the temporal features of the time series, thereby enhancing its ability to model temporal dependencies. This improvement helps the model to more accurately capture the relationships between different time steps when processing time-series data with temporal patterns. During the validation process, a specific approach was adopted to ensure the independence of subjects. When dividing the training set and the validation set, the data was distinguished into a control group and a Parkinson’s disease group. Based on the specified number of folds, a portion of subjects was selected, and then the range of corresponding segments of these subjects was determined. All the records of the selected subjects were either entirely included in the validation set or the training set, avoiding the dispersion of records from the same subject into different sets. Finally, the cumulative number of segments in the validation set and the training set was calculated. In this way, in each fold of cross - validation, all the gait records of the same subject could be ensured to be assigned to the same set, effectively avoiding result biases and poor model generalization ability caused by randomly allocating different records of the same subject to the training set and the validation set. For the data of the control group and the Parkinson’s disease group, a portion of patients was selected at a ratio of 1/10 respectively. All the data segments of these selected patients were integrally included in the validation set, while the data segments of the remaining patients were included in the training set. This division method effectively ensured that all the data records of the same subject would only appear in either the training set or the validation set, avoiding potential result biases caused by the dispersion of different records of the same subject in two sets, thus ensuring the independence of subjects.
4 Experiments and results
In the next sections, we will describe our experimental procedure in detail and analyze and interpret its results. First, we will introduce the dataset used and analyze the gait feature information of the dataset through visualization to show the differences between Parkinson’s patients and healthy subjects more clearly. Second, we will try multiple DA methods on different models to validate their effectiveness. Subsequently, we will explore the detailed debugging work performed during the training of the models. Finally, to reduce the complexity of the model, we try to perform feature selection based on the importance of the channels in order to simplify the model. This step helps to improve the interpretability and generalization of the model and provides a valuable reference for future model design.
4.1 Datasets and metrics
In this study, we used the publicly available dataset PhysioNet [40]. This dataset was contributed by several researchers including Yogev et al. [41], Hausdorff et al. [42] and Frenkel-Toledo et al. [43].The PhysioNet dataset covers data from three different gait patterns including walking on flat ground, rhythmic auditory stimulation (RAS) walking and treadmill walking. Detailed descriptions of these datasets are shown in Table 1. Divided by the number of steps walked, as in Table 2, a total of 306 gait records were collected, of which 214 (70%) were from PD patients and 92 (30%) were from healthy individuals. Table 3 shows the total number of healthy individuals and PD patients in each dataset and their corresponding severity levels, which were determined according to the H&Y (Hoehn-Yahr) scale.
In Fig 7, we visualize the magnitude of the VGRF over time in the left and right foot for PD patients versus controls. With these visualizations, we can clearly see that the VGRF signals of PD patients and control subjects differed significantly in peak value, amplitude, shape, and timing. Specifically, the VGRF signals of the control subjects showed a more organized and consistent pattern. These signals had stable amplitudes and regular shapes with predictable timing. This consistency suggests that control subjects have good gait symmetry and stability during gait. However, the situation is different in PD patients. As the severity of PD increased, the deviations in their VGRF signals became more pronounced, as shown in Fig 8. These signals show greater variability in peak, amplitude, and timing, and tend to exhibit greater asymmetry. For example, there is a significant difference in the VGRF signal between the patient’s left and right feet, and the force exerted by the right foot may be significantly less. This asymmetry and variation not only reflects gait abnormalities, but may also result in significant impairment of the patient’s mobility in daily life.
This increase in VGRF signal deviations may be due to the lack of stability and coordination during gait in patients with PD, causing their gait patterns to become more irregular. As the disease worsens, these deviations may be further exacerbated, resulting in a greater negative impact on the patient’s daily life and functional abilities. To cope with these changes, patients with PD may require more intensive treatment and management to help them maintain a better quality of life and mobility.
We used Precision, Recall, F1, and Accuracy to evaluate our model, where the control group was identified as the negative (N) group and the Parkinson’s group was identified as the positive (P) group, and the following are the formulas for the four metrics that we used:
Where,
- TP: The number of true positives
- TN: The number of true negatives
- FP: The number of false positives
- FN: The number of false negatives
4.2 Parameter analysis experiment
We conducted experiments on various parameter configurations for 1D-ConvNet, see Table 4. The results indicate that the configuration of Model 1 performs best: Conv1D settings are (8,3), (16,3), (16,3), (1,3), with SELU as the activation function, dropout set to 0.1, learning rate at 0.01, and batch size of 200. Under this setup, the model achieves Precision, Recall, and F1 scores of 0.85, 0.84, and 0.84, respectively, demonstrating excellent classification ability and generalization performance.
In comparison, Model 5, which uses larger convolution kernels, expands the range of feature extraction but shows slightly lower performance. This suggests that excessively large kernels may hinder effective capture of local features. Additionally, Model 6, which uses Tanh as the activation function, experiences a significant drop in F1 score to 0.74, further supporting SELU’s suitability for this task. The batch size analysis also confirms that a setting of 200 outperforms smaller batch sizes, leading to more stable gradient updates and improved model performance. In summary, the parameter configuration of Model 1 outperforms other combinations, highlighting its strong adaptability and advantages for the given task.
We also conducted experiments on various parameter configurations for temporal-spatial transformer. The experimental results (see Table 5) clearly indicate that the hyperparameter combination we selected (Num_blocks = 1, Key_dim = 100, Num_heads = 2, Drop_rate = 0.3, LR = 0.01, Batchsize = 100) significantly outperforms other configurations. In terms of performance metrics, this combination achieved a precision of 0.82, recall of 0.79, and F1 score of 0.79, all of which are higher than those of other setups. Compared to multi-layer Transformers (e.g., Num_blocks = 2), the single-layer Transformer is more efficient at capturing key features, avoiding overfitting, and saving training time. The choice of feature dimension (Key_dim = 100) and the number of attention heads (Num_heads = 2) strikes a good balance between model complexity and data suitability. Higher feature dimensions (e.g., Key_dim = 256) may lead to performance degradation. The selected dropout rate (Drop_rate = 0.3) effectively reduces overfitting while maintaining the model’s learning capacity. A lower learning rate (LR = 0.01) ensures smoother optimization, while a larger batch size (Batchsize = 100) enhances training efficiency and model stability.
Compared to other parameter combinations, our choice simplifies the model structure and optimizes hyperparameter settings, significantly improving the model’s generalization ability and stability. This results in the best overall performance with a limited sample size, further demonstrating the importance of appropriate parameter selection for model optimization.
4.3 Results on different models
Results on CNN: To evaluate the effectiveness of the DA methods, we used eight different augmentation techniques and performed a detailed comparison of the performance of these methods with the designed 1D-ConvNet model based on four core metrics, namely Precision, Recall, F1 value and Accuracy. The results of the DA experiments are shown in Table 6, which underlines the best results for each label (e.g., health status and different disease severity) on each metric.
Overall, the augmented model significantly outperforms the unaugmented data on all labels and assessment metrics, with the Jittering method performing the best with an accuracy of 89.6%. This technique adds random noise to the data to increase the perturbation, which improves the model’s generalization ability and makes it more stable when dealing with unknown data.Scaling performs very close to the same performance, achieving an accuracy of 86.3%. By scaling the data, Scaling helps the model to better adapt to changes in the data at different scales Magnitude Warping, another effective augmentation strategy, achieves an accuracy of 85.8%. This method improves the diversity by adjusting the magnitude of the data to enhance the robustness of the model. Permutation and Time Warping methods perform less well, permutation increases the diversity of the data by disrupting the temporal order of the data, but with a slightly lower accuracy. Time Warping achieves a similar effect by nonlinearly distorting the timeline, which improves the robustness of the model, but its overall performance is still not as good as that of Jittering and Scaling. Jittering and Scaling. Rotation method performs the worst, with an accuracy of 69%. Rotation, as a time-series DA method, performed poorly, possibly because it disrupted the temporal relationships in the data. Additionally, when applied to multi-axis signals, rotation might not accurately reflect the changes in the data, potentially introducing distortion and reducing the model’s performance. In addition, the table shows the results of Random Sampling and three different strengths of regularization (lp_25, lp_50, and lp_100). The regularization treatment provides limited improvement in model performance, while Random Sampling fails to significantly outperform the other augmentation methods.
In summary, there are significant differences in the effectiveness of different DA strategies in improving model performance. Among them, Jittering, Scaling and Magnitude Warping are the best performers in several indicators, especially in accuracy and F1 value. In contrast, Rotation and other methods fail to improve the model performance effectively because they change the feature pattern of the data. This suggests that the choice of different DA strategies is crucial for model performance on time series data.
Results on Transformer: As shown in Table 7 are the experimental results of the encoder based model. It can be observed from the table that the baseline model without using DA methods achieves 79% accuracy in Accuracy. In contrast, most of the DA methods we used outperform the baseline model in Accuracy, Precision, Recall, and F1. Specifically:
- Jittering The model is 89.6% accurate after processing. Jittering is a technique that adds small amounts of noise to the raw data, which helps to increase the robustness of the model and makes it better able to handle noisy data in a real-world environment.
- Scaling processing resulted in 82% accuracy of the model. The scaling technique improves the generalization ability of the model by adjusting the magnitude of the data so that it is trained on different scales.
- After Permutation processing, the model has the highest accuracy of 90.83%. Permutation techniques can improve model accuracy by increasing data diversity through changing the order of the data, particularly in multi-axis signals. By modifying the sequence of the data, permutation allows the model to learn different combinations, helping it identify new patterns. In time-series data, permutation not only enhances the data diversity but also helps the model discover dependencies in different orders, thus improving generalization and accuracy.
- In the Random Sampling processing stage, the model’s accuracy was 81.5%. Random Sampling allows the model to better cope with data imbalances by randomly selecting a subset of the original data for training.
However, not all DA methods significantly improve model performance. For example:
- After Rotation and Time Warping processing, the model fails to process the data efficiently and performs poorly. Rotation increases the diversity of the data by rotating it, but in some cases may destroy the original structure of the data. Time Warping distorts the data on the time axis by making it change in the time dimension, but this change may result in the loss of the temporal characteristics of the data.
- In Magnitude Warping processing, the model performance is the same as the baseline model, and there is no significant advantage. Magnitude Warping increases the diversity of the data by changing its magnitude, but this method retains less of the data features, and therefore fails to improve the model performance significantly.
To summarize, through different DA methods, we can significantly improve the performance of the model, but we need to pay attention to choosing the appropriate method to avoid performance degradation in some cases. This suggests that in practical applications, the selection of DA techniques should be optimized with specific tasks and data features and different models to achieve the best results.
Results with Other Methods: This Table 8 demonstrates the performance comparison of different models in terms of prediction accuracy. The model we designed with DA and Spatio-Temporal Transformer techniques achieves an accuracy of 90.8%, which is a significant improvement compared to other models.
Specifically, Caramia et al.’s SVM-RBF model achieved 75.6% accuracy on the H&Y scale, which is the lowest result in the table. The Artificial Neural Network (ANN) model used by Veeraragavan et al. achieved 87.1% accuracy, while El Maachi et al.’s 1D-ConvNet under the UPDRS rating system achieved 85.3%, and 85.3% F1-score. Safwen Naimi et al. achieved 87.89% accuracy by combining CNN with Transformer architecture and performed well on H&Y grading, and 88.0% F1-score.
Compared to these models, our DA + spatio-temporal Transformer model shows an extremely high accuracy of 90.8% on the H&Y grading system, and 91.0% F1-score. This suggests that our approach not only fully utilizes DA strategies (e.g., Jittering, Scaling, etc.), but also better captures temporal and spatial features in the data by introducing the spatio–temporal Transformer model. This performance enhancement illustrates the value of combining DA with advanced modeling structures and provides strong support for further exploration of similar applications.
We present the confusion matrix results for two models in Figs 9 and 10. In a confusion matrix, the diagonal elements represent the number of samples correctly predicted by the model for each class, while off-diagonal elements represent misclassifications where one class is predicted as another. It can be observed that our models demonstrate high prediction accuracy across various classes.
Moreover, to verify the stability of the model’s performance, we conducted an independent samples t-test on the results before and after DA. The p-value we obtained from the test was 0.0326, which is less than the commonly used significance level of 0.05, indicating that there is a significant difference in the predicted values between augmented and non-augmented data. Specifically, the DA method had a significant impact on the model’s prediction results. This means that, after using DA, the model’s prediction performance significantly differs from when DA was not used. The result suggests that DA may have improved the model’s robustness and accuracy when handling different variations in the data, thereby enhancing the model’s generalization ability.
To ensure the reliability of our conclusions, we further visualized the distribution of the predicted values for augmented and non-augmented data, as shown in Fig 11. The blue bars represent non-augmented data, the red bars represent augmented data, and the purple area represents the overlap between the two. From the graph, it can be seen that, for the mild Parkinson’s prediction values (i.e., early and middle stages), the red bars (augmented data) are generally higher than the blue bars. This indicates that, after DA, our model performs better at identifying misclassifications in the non-augmented data, particularly in cases where healthy data is misclassified as disease data. DA likely improved the model’s prediction frequency and coverage, especially for low-frequency categories, further optimizing the overall performance of the model.
4.4 Implementation details
The parameters of each factor of the DA method we used are shown in the Table 9, and in linear residual prediction we used decimationRate extraction factors of 1, 2, and 4 for the experiments, respectively. The parameter settings of the 1D-ConvNet are shown in Fig 3.
We set the learning rate to 0.0002 during training and enabled the EarlyStopping mechanism. This callback function checks the monitored metric (val_loss) at the end of each training cycle. Training is stopped if the metric does not improve significantly (i.e., the improvement is less than min_delta) within the specified PATIENCE period. This helps prevent the model’s performance from deteriorating on the validation set and avoids overfitting. We chose the Nadam (Nesterov-accelerated Adaptive Moment Estimation) optimizer, a variant based on the Adam optimizer that combines the advantages of the Nesterov-accelerated gradient and Adam. The specific parameters are β1 = 0 . 9, β2 = 0 . 999, and β2 = 0 . 999. a
To ensure the independence of subjects, we guarantee that each subject’s data appears only in either the training set or the testing set, avoiding information leakage. During cross-validation, the data of each subject is appropriately assigned to different subsets, ensuring that the subjects in the training and validation sets are completely independent, thereby improving the reliability of model evaluation.
4.5 Feature selection experiment
We observed some interesting phenomena when we conducted ablation experiments on the data from eighteen sensors. Specifically, we compared sensors symmetrically located on the left and right feet. As can be seen from Table 10, the experimental results show that L2&R2 have the lowest accuracy, which suggests that the second sensors on both the left and right feet are the most important, likely containing critical discriminative feature information. Additionally, the accuracy of L5&R5 and L8&R8 also decreased, indicating that these sensors play a crucial role in the model, and thus these features should be given special attention during modeling. In contrast, L3&R3 have the highest accuracy, and their absence has minimal impact on overall performance, suggesting that the third sensors on both feet are of lower importance. Considering the computational resource constraints of the model, it will be worth considering removing these two features or reducing their weight in the model to optimize both computational efficiency and performance.
5 Conclusion and future work
In this study, we experimented with various DA methods, such as permutation, jittering, scaling, and other techniques for plantar time series data, and the experimental results showed that the use of appropriate DA can significantly improve the robustness and generalization ability of the model. We chose 1D-ConvNet and spatio-temporal combined one-dimensional encoder as the experimental test models, and the results show that our data preprocessing design not only reduces the cost of computational resources, but also shortens the training time, which makes the model efficiently transferable to real-time applications. Through the experiments with the above methods, our model achieves an accuracy of 90.8% on the H&Y grading scale, and realizes a precision of 92.0%, a recall of 91.0%, and an F1 score of 91.0%. Clearly superior to existing state-of-the-art techniques, this demonstrates the validity and strength of our method. Unlike in the past, we achieve high precision while still maintaining the simplicity of the model design, which means that the model is easier to deploy and maintain in real-world applications.
In addition, we conducted a feature selection study for different sensors, aiming at effective feature screening. This study not only provides important data support for our current work, but also provides valuable inspiration for future directions. By delving deeper into these data, we will be able to identify and eliminate redundant information, enabling more accurate extraction of useful features. This will significantly enhance the model’s performance and efficiency, making it more reliable and effective in practical applications. Additionally, by optimizing the feature selection process, we will build more concise and efficient models, providing clinicians and researchers with more valuable insights and support.
Although the proposed method achieves remarkable results in feature selection and model optimization, it still has some limitations. For instance, the dataset has limited coverage, suggesting the need for broader data collection in the future. Moreover, since the dataset is primarily derived from controlled environments, the model’s generalizability is restricted. Additionally, its adaptability to dynamic changes in environments or data collection conditions requires further improvement. Compared to previous studies, this work incorporates DA techniques, which not only enhance the processing speed but also maintain overall efficiency. Unlike traditional approaches that directly use full-scale features or build complex models, this study employs a systematic feature selection process, providing valuable references and insights for future research on feature selection. In terms of computational cost, the simplified single-model approach demonstrates significant advantages: by reducing model complexity and parameters, it markedly lowers computational resource requirements and runtime while remaining resource-friendly and easy to maintain. This enables the model to operate reliably on low-performance devices, offering a more efficient and dependable solution for primary healthcare and resource-constrained settings.
References
- 1. Bargman J, Varughese S. Actinomyces neuii PD peritonitis-resolution of infection without catheter removal. Peritoneal dialysis international. J ISPD. 2014.
- 2. Sena J, Mostafiz T, Zhang J, Davidson A, Bandyopadhyay S, Ren Y, et al. The potential of wearable sensors for assessing patient acuity in intensive care unit (ICU). 2023. arXiv Preprint. arXiv:abs/1405.3531 https://api.semanticscholar.org/CorpusID:265033794
- 3. Sigcha L, Borzì L, Amato F, Rechichi I, Ramos-Romero C, Cárdenas A, et al. Deep learning and wearable sensors for the diagnosis and monitoring of Parkinson’s disease: A systematic review. Expert Syst Appl. 2023;229:120541.
- 4. Burtscher J, Bourdillon N, Janssen Daalen JM, Patoz A, Bally JF, Kopp M, et al. Movement analysis in the diagnosis and management of Parkinson’s disease. Neural Regen Res 2025;20(2):485–6. pmid:38819059
- 5. Lu R, Xu Y, Li X, Fan Y, Zeng W, Tan Y, et al. Evaluation of wearable sensor devices in Parkinson’s disease: a review of current status and future prospects. Parkinson’s Dis 2020;2020(2):1–8. pmid:33029343
- 6. Moreau C, Rouaud T, Grabli D, Benatru I, Remy P, Marques A-R, et al. Overview on wearable sensors for the management of Parkinson’s disease. NPJ Parkinsons Dis 2023;9(1):153. pmid:37919332
- 7. Sun Y-M, Wang Z-Y, Liang Y-Y, Hao C-W, Shi C-H. Digital biomarkers for precision diagnosis and monitoring in Parkinson’s disease. NPJ Digit Med 2024;7(1):218. pmid:39169258
- 8. Mughal H, Javed AR, Rizwan M, Almadhor AS, Kryvinska N. Parkinsons’s disease management via wearable sensors: A systematic review. IEEE Access. 2022;10:35219–37.
- 9. Shalin G, Pardoel S, Lemaire ED, Nantel J, Kofman J. Prediction and detection of freezing of gait in Parkinson’s disease from plantar pressure data using long short-term memory neural-networks. J Neuroeng Rehabil 2021;18(1):167. pmid:34838066
- 10. Polvorinos-Fernández C, Sigcha L, Borzì L, Olmo G, Asensio C, López JM, et al. Evaluating motor symptoms in Parkinson’s disease through wearable sensors: a systematic review of digital biomarkers. Appl Sci 2024;14(22):10189.
- 11. Xu M, Yoon S, Fuentes A, Park DS. A comprehensive survey of image augmentation techniques for deep learning. Pattern Recognit. 2023;137:109347.
- 12. Krizhevsky A, Sutskever I, Hinton G. ImageNet classification with deep convolutional neural networks. In: Advances in neural information processing systems. Vol. 25:2012. p. 1097–1105.
- 13. Vaswani A, Shazeer N, Parmar N, Uszkoreit J, Jones L, Gomez AN, et al. Attention is all you need. In: Proceedings of the 31st international conference on neural information processing systems. California: Curran Associates Inc. 2017. p. 6000–6010. ISBN: 9781510860964
- 14. Impellizzeri FM, Menaspà P, Coutts AJ, Kalkhoven J, Menaspà MJ. training load and its role in injury prevention, part I: back to the future. J Athl Train 2020;55(9):885–92. pmid:32991701
- 15. Coles PA. An injury prevention pyramid for elite sports teams. Br J Sports Med 2018;52(15):1008–10. pmid:28343160
- 16. Zhan A, Mohan S, Tarolli C, Schneider RB, Adams JL, Sharma S, et al. Using smartphones and machine learning to quantify Parkinson disease severity: the mobile Parkinson disease score. JAMA Neurol. 2018;75(7):876–80. pmid:29582075
- 17. Morris SJ, Paradiso JA. Shoe-integrated sensor system for wireless gait analysis and real-time feedback. In: Engineering in medicine and biology, 2002. 24th annual conference and the annual fall meeting of the biomedical engineering society EMBS/BMES conference, 2002. Proceedings of the second joint. 2002.
- 18. Goschenhofer J, Pfister F, Yuksel K, Bischl B, Fietzek U, Thomas J. Wearable-based Parkinson’s disease severity monitoring using deep learning. In: Proceedings of the European conference on machine learning and principles and practice of knowledge discovery in databases (ECML/PKDD). 2019. Available from: https://api.semanticscholar.org/CorpusID:129945865
- 19. Grover S, Bhartia S, Akshama, Yadav A, Seeja KR. Predicting severity of Parkinson’s disease using deep learning. In: Procedia computer science, Vol. 132:2018. p. 1788–1794. International conference on computational intelligence and data science. Available from: https://www.sciencedirect.com/science/article/pii/S1877050918308883 doi: doi: https://doi.org/10.1016/j.procs.2018.05.154
- 20. Schlachetzki JCM, Barth J, Marxreiter F, Gossler J, Kohl Z, Reinfelder S, et al. Wearable sensors objectively measure gait parameters in Parkinson’s disease. PLoS One 2017;12(10):e0183989. pmid:29020012
- 21. El Maachi I, Bilodeau G-A, Bouachir W. Deep 1D-Convnet for accurate Parkinson disease detection and severity prediction from gait. Expert Syst Appl. 2020;143:113075.
- 22. Nguyen DMD, Miah M, Bilodeau G-A, Bouachir W. Transformers for 1D signals in Parkinson’s disease detection from gait. In: 2022 26th International conference on pattern recognition (ICPR). 2022. p. 5089–5095.
- 23. Zhao A, Dong J, Li J, Qi L, Zhou H. Associated spatio-temporal capsule network for gait recognition. IEEE Trans Multimedia. 2022;24:846–60.
- 24. Aşuroğlu T, Oğul H. A deep learning approach for parkinson’s disease severity assessment. Health Technol 2022;12(5):943–53.
- 25. Shcherbak A, Kovalenko E, Somov A. Detection and classification of early stages of parkinson’s disease through wearable sensors and machine learning. IEEE Trans Instrum Meas. 2023;72:1–9.
- 26. Moon S, Song H-J, Sharma VD, Lyons KE, Pahwa R, Akinwuntan AE, et al. Classification of Parkinson’s disease and essential tremor based on balance and gait characteristics from wearable motion sensors via machine learning techniques: a data-driven approach. J Neuroeng Rehabil 2020;17(1):125. pmid:32917244
- 27. Sotirakis C, Su Z, Brzezicki MA, Conway N, Tarassenko L, FitzGerald JJ, et al. Identification of motor progression in Parkinson’s disease using wearable sensors and machine learning. NPJ Parkinsons Dis 2023;9(1):142. pmid:37805655
- 28. He K, Zhang X, Ren S, Sun J. Deep residual learning for image recognition. In: 2016 IEEE conference on computer vision and pattern recognition (CVPR). 2016. p. 770–778.
- 29. Chatfield K, Simonyan K, Vedaldi A, Zisserman A. Return of the devil in the details: Delving deep into convolutional nets. 2014. arXiv Preprint.
- 30. Cui X, Goel V, Kingsbury B. Data augmentation for deep neural network acoustic modeling. IEEE/ACM Trans Audio Speech Lang Process 2015;23(9):1469–77.
- 31. Li X, Metsis V, Wang H. TTS-GAN: A Transformer-Based Time-Series Generative Adversarial Network. In: Proceedings of the international conference on artificial intelligence in medicine. 2022.
- 32. Yoon J, Jarrett D, van der Schaar M. Time-series generative adversarial networks. In: Proceedings of the 33rd international conference on neural information processing systems. New York: Curran Associates Inc. 2019. p. 494.
- 33. Goodfellow IJ, Pouget-Abadie J, Mirza M, Xu B, Warde-Farley D, Ozair S, et al. Generative adversarial nets. In: Proceedings of the 27th international conference on neural information processing systems. Vol. 2:2014. p. 2672–2680.
- 34. Le Guennec A, Malinowski S, Tavenard R. Data augmentation for time series classification using convolutional neural networks. In: ECML/PKDD workshop on AALTD. 2016.
- 35. Chen W, Shi K. Multi-scale attention convolutional neural network for time series classification. Neural Netw. 2021;136:126–40. pmid:33485098
- 36. Ismail Fawaz H, Forestier G, Weber J, Idoumghar L, Muller P-A. Data augmentation using synthetic data for time series classification with deep residual networks. 2018. arXiv Preprint.
- 37. Alle S, Priyakumar UD. Linear prediction residual for efficient diagnosis of Parkinson’s disease from gait. In: Medical image computing and computer assisted intervention – MICCAI. Cham: Springer. 2021. p. 614–623.
- 38. Um TT, Pfister FMJ, Pichler DC, Endo S, Lang M, Hirche S, et al. Data augmentation of wearable sensor data for Parkinson’s disease monitoring using convolutional neural networks. In: Proceedings of the 19th ACM international conference on multimodal interaction. 2017. https://api.semanticscholar.org/CorpusID:21056258
- 39. Singh M, Prakash P, Kaur R, Sowers R, Brašić JR, Hernandez ME. A deep learning approach for automatic and objective grading of the motor impairment severity in Parkinson’s disease for use in tele-assessments. Sensors (Basel) 2023;23(21):9004. pmid:37960703
- 40. Goldberger AL, Amaral LA, Glass L, Hausdorff JM, Ivanov PC, Mark RG, et al. PhysioBank, PhysioToolkit, and PhysioNet: components of a new research resource for complex physiologic signals. Circulation 2000;101(23):E215–20. pmid:10851218
- 41. Yogev G, Giladi N, Peretz C, Springer S, Simon ES, Hausdorff JM. Dual tasking, gait rhythmicity, and Parkinson’s disease: which aspects of gait are attention demanding? Eur J Neurosci. 2005;22(5):1248–56. pmid:16176368
- 42. Hausdorff JM, Lowenthal J, Herman T, Gruendlinger L, Peretz C, Giladi N. Rhythmic auditory stimulation modulates gait variability in Parkinson’s disease. Eur J Neurosci 2007;26(8):2369–75. pmid:17953624
- 43. Frenkel-Toledo P, Giladi N, Peretz C, Herman T, Hausdorff J. Treadmill walking as an external pacemaker to improve gait rhythm and stability in Parkinson’s disease. Movement Disorders 2010;20(9):1177–83.
- 44. Caramia C, Torricelli D, Schmid M, Munoz-Gonzalez A, Gonzalez-Vargas J, Grandas F, et al. IMU-based classification of Parkinson’s disease from gait: a sensitivity analysis on sensor location and feature selection. IEEE J Biomed Health Inform 2018;22(6):1765–74. pmid:30106745
- 45. Veeraragavan S, Gopalai A, Gouwanda D, Ahmad S. Parkinson’s disease diagnosis and severity assessment using ground reaction forces and neural networks. Front Physiol. 2020;11.
- 46. Naimi S, Bouachir W, Bilodeau G-A. 1D-convolutional transformer for Parkinson disease diagnosis from gait. Neural Comput & Applic 2023;36(4):1947–57.