Skip to main content
Advertisement
Browse Subject Areas
?

Click through the PLOS taxonomy to find articles in your field.

For more information about PLOS Subject Areas, click here.

  • Loading metrics

Deep learning-based gait phase detection using shank-mounted IMU data: Classification approach

  • Wonseok Choi,

    Roles Data curation, Formal analysis, Investigation, Methodology, Software, Validation, Visualization, Writing – original draft

    Affiliation Department of Intelligent Robotics, Sungkyunkwan University, Suwon, South Korea

  • Mun-Taek Choi

    Roles Conceptualization, Data curation, Formal analysis, Funding acquisition, Investigation, Methodology, Project administration, Resources, Supervision, Validation, Writing – review & editing

    mtchoi@skku.edu

    Affiliation Department of Intelligent Robotics, Sungkyunkwan University, Suwon, South Korea

Abstract

Gait is a key indicator for assessing an individual’s mobility and overall health, and accurate detection of gait cycle phases is essential for precise gait analysis. Most existing gait phase detection approaches rely on multiple wearable sensors, increasing system complexity and limiting real-world applicability. Moreover, rule-based methods, traditional machine learning models, and convolutional neural network (CNN)-based approaches often fail to capture complex temporal dependencies, resulting in limited robustness and real-time performance. And approaching gait phase detection as a classification method enables automatic phase recognition directly from sequential signals, simplifying the processing pipeline and supporting stable performance. In this study, we utilized an end-to-end supervised classification learning approach using a state-of-the-art Transformer model. We utilized the publicly available NONAN GaitPrint dataset, time-series gait recordings from 35 healthy young adults who walked along a 200-meter indoor level-ground track, to train deep learning models for gait phase detection. Using the three-axis acceleration and three-axis angular velocity signals from the shank, we trained and evaluated a variety of classification models, including a one-dimensional CNN, a hybrid long short-term memory and gated recurrent unit (Hybrid LSTM+GRU) model, and a Transformer. The results showed that the Transformer model achieved the F1-score of approximately 92.99%, while the CNN and Hybrid LSTM+GRU models demonstrated comparable performance, indicating no substantial differences among the models. These findings demonstrate that clinically and practically feasible gait phase detection can be achieved using shank-mounted IMUs with an end-to-end learning approach.

Introduction

Human gait is a primary mode of locomotion that involves a complex interplay of biomechanical and neuromuscular components [1]. To quantitatively analyze gait, it is essential to divide the gait cycle into sequential intervals and define the functional characteristics of each segment—commonly referred to as gait phases [2]. A typical gait cycle consists of the following phases: Loading Response (LR), where the body absorbs impact following heel contact; Mid Stance (MS) and Terminal Stance (TS), which correspond to periods of single-limb support; Pre-Swing (PSw), the transition phase prior to toe-off; and Swing (Sw), during which the limb is airborne and progresses in the forward direction. These gait phases serve as fundamental units that describe the temporal progression of the gait cycle, and are widely used to quantify gait continuity, stability, and functional changes [3,4]. Accurate classification of gait phases is critical in various applications, including clinical diagnostics, rehabilitation therapy, and real-time monitoring systems using wearable sensors [57].

Accurate classification of gait phases requires detailed analysis of the gait cycle, and a variety of measurement and analytical techniques have been proposed to achieve this. Common methods for collecting gait data include camera-based 3D motion capture systems, instrumented treadmills, and wearable sensor systems [8,9]. However, these approaches often involve high equipment costs, complex setup procedures, and limitations to controlled laboratory environments [10]. In contrast, inertial measurement units (IMUs) are increasingly utilized for gait analysis, given their low cost, portability, and practicality in everyday settings [1113]. Nevertheless, many previous studies have relied on attaching multiple IMUs to various body segments in order to accurately segment gait phases. This approach increases the burden on the user, raises the overall equipment cost, and reduces feasibility for field applications [1416]. To overcome these limitations, recent research has focused on classifying gait phases with a minimal number of sensors, particularly with single-IMU configurations, which aim to reduce user discomfort while maintaining high classification performance [1719].

The shank segment is known to exhibit the most distinctive movement patterns during walking, and changes in acceleration and angular velocity are clearly observable at this location [2022]. Signals measured by IMU sensors—specifically acceleration and angular velocity—sensitively reflect both linear and rotational movements during gait, with notable fluctuations, directional shifts, and periodic differences occurring especially at phase transition points [23,24]. For this reason, shank-mounted IMUs are often sufficient to capture the essential discriminative features required for gait phase classification [25]. This makes it particularly suitable for deep learning-based time-series classification models, which can effectively learn the structural characteristics of each gait phase from raw sensor signals.

Previous studies on gait phase analysis have primarily employed regression-based approaches [26] or classification-based approaches [27], utilizing various machine learning methods such as k-nearest neighbors, support vector machines, and hidden Markov models [2832]. While these traditional methods offer advantages such as relatively low computational cost and interpretable structures, they often struggle to capture the temporal dynamics of gait phases when the input data are nonlinear and feature boundaries are ambiguous due to the complexity of human movement and inter-individual variability.

Recent studies have increasingly applied deep learning techniques to IMU signals for gait phase classification, demonstrating that deep models can effectively capture temporal patterns and salient features across various walking conditions [33,34]. Building on this trend, deep learning models can autonomously extract high-level representations—such as temporal patterns, periodicity, and phase transition characteristics—from raw time-series inputs, achieving high classification accuracy without explicit feature engineering [35]. This capability enables effective learning of the temporal structure inherent in gait signals and has been shown to yield robust performance even under real-world measurement environments.

In this study, we use gait data from healthy adults collected via IMU sensors attached to both shanks, employing three-axis acceleration and angular velocity signals to evaluate deep learning–based gait phase classification. Gait phases are labeled by extracting gait events from foot-contact data and assigning the corresponding phase to each time frame. We adopt an end-to-end learning framework without hand-crafted features, allowing the models to directly learn temporal dynamics and phase-transition characteristics from raw time-series inputs. To examine how well gait phases can be identified under a minimal-sensor configuration, we evaluate three representative deep learning architectures—a convolutional neural network (CNN), a hybrid long short-term memory and gated recurrent unit (Hybrid LSTM+GRU), and a Transformer encoder. The results show that all three models achieve consistent and reliable performance using only shank-mounted IMU data, indicating that shank-mounted IMUs provides sufficient temporal information for multi-class gait phase classification in level-ground walking. By comparing these architectures under identical experimental conditions, this study clarifies the strengths of lightweight models for simple, highly periodic gait patterns and highlights the advantages of more advanced architectures for scenarios involving greater temporal complexity. Overall, this work demonstrates that minimal-sensor deep learning approaches can substantially reduce system complexity while maintaining clinically and practically meaningful gait phase recognition performance.

Materials and methods

Dataset

In this study, we utilized the publicly available NONAN GaitPrint dataset [36] to train and evaluate deep learning models for gait phase detection. The dataset comprises time-series gait recordings from 35 healthy young adults who walked along a 200-meter indoor track. Each participant completed 18 trials (4 minutes per trial) over two separate days, with data that had been collected from multiple joints using Noraxon Ultium Motion IMU sensors, sampled at 200 Hz. Table 1 provides the key features and structure of the NONAN GaitPrint dataset. For this study, the IMU data were resampled to 100 Hz to improve computational efficiency and reduce overfitting. We used acceleration and angular velocity along three axes from IMU sensors mounted on both shanks. Fig 1 illustrates the placement of the sensors and visualizes one representative axis from both acceleration and angular velocity signals.

thumbnail
Table 1. Overview of NONAN GaitPrint dataset characteristics.

https://doi.org/10.1371/journal.pone.0344002.t001

thumbnail
Fig 1. Depiction of IMU sensor positioning on the shank and foot, along with sensor axes and representative signals.

The foot-mounted sensor was used to obtain contact data.

https://doi.org/10.1371/journal.pone.0344002.g001

Data preprocessing

In consideration of the extensive size of the NONAN dataset, and to ensure computational efficiency and practical feasibility of the analysis, we utilized four randomly selected trials per participant out of the total 18 trials. Based on this, prior to performing gait phase labeling, four gait phases were defined as illustrated in Fig 2. Among them, Late Stance (LS) was formed by combining MS and TS into a continuous support phase, since the functional difference between them was regarded as minimal in real-time applications [37]. This integration was intended to simplify the model and enhance its practical applicability.

thumbnail
Fig 2. Gait phase transition diagram for a gait cycle.

Black indicates the right leg and gray indicates the left leg.

https://doi.org/10.1371/journal.pone.0344002.g002

Gait phase labeling was performed by first automatically detecting Heel Strike (HS) and Toe Off (TO) events from the bilateral foot contact signals. Each phase was then defined based on the temporal intervals between these gait events [38]. Figs 3 and 4 illustrate the detected HS and TO event timings from the left and right foot contact signals, providing a clear visualization of the temporal structure in which these events occur within the gait cycle. The gait phases were defined according to the following rules: the interval from right HS to left TO was labeled as the LR, and the interval from left HS to left TO was labeled as the LS. Subsequently, the period from left TO to the next left HS was labeled as the PSw, and the interval from left TO to the next right HS was labeled as the Sw. Based on these rules, gait phase labels were automatically assigned to the entire time-series sequence using both sides’ contact data.

thumbnail
Fig 3. Input signals of left shank: acceleration, angular-velocity, and contact data.

https://doi.org/10.1371/journal.pone.0344002.g003

thumbnail
Fig 4. Input signals of right shank: acceleration, angular-velocity, and contact data.

https://doi.org/10.1371/journal.pone.0344002.g004

To train and evaluate the deep learning models, we constructed input–output sequences consisting of time-series inputs x together with their associated gait phase labels y. The input sequence x comprises IMU time-series data recorded from both shanks, while y represents the corresponding gait phase label at each time step. The dataset contains n time steps, with each time point associated with synchronized shank IMU signals and their respective phase labels. To better model the temporal dynamics of gait phases, we designed a sequence-based input structure using a sliding window approach [39]. Table 2 illustrates how the dataset was organized using this method. The window length w and the prediction offset are both tunable hyperparameters; in this study, they were fixed to 20 and +1, respectively, to ensure consistent model training and stable sequence representation. Based on the defined window length, each input sequence consists of w consecutive IMU measurements , and the corresponding output label is taken from the subsequent time step yw+1. This one-step-ahead prediction structure is intended to mitigate potential label latency in real-time time-series applications by predicting the phase at the next time step rather than the final point of the input window. The window then slides forward by one time step to generate the next input–output pair, resulting in a total of training samples. Specifically, at the final time step n, the input sequence is , and the corresponding output label is yn. This structure enables the model to estimate the upcoming gait phase based on preceding temporal information, thereby facilitating the learning of transition dynamics and temporal dependencies within the gait sequence.

thumbnail
Table 2. Sliding window representation of training data. Each sample consists of w consecutive IMU inputs and the label at the next time step.

https://doi.org/10.1371/journal.pone.0344002.t002

Deep Learning Models

To effectively classify gait phases, several deep learning models were utilized, including CNN, Hybrid LSTM+GRU, and Transformer architectures. The structures of these models are illustrated in Fig 5, and their objective is to capture spatial and temporal features from sequential IMU data so as to accurately classify each gait phase within the gait cycle.

thumbnail
Fig 5. Architectural diagrams for deep learning models employed in gait phase classification.

The figure illustrates (a) CNN model, (b) Hybrid LSTM+GRU model, and (c) Transformer encoder model.

https://doi.org/10.1371/journal.pone.0344002.g005

The CNN model is based on a one-dimensional CNN, which is well-suited for extracting local spatial patterns from time-series sensor data. The architecture consists of two convolutional layers, followed by pooling layers and fully connected dense layers [40]. The convolution operation can be expressed as:

where xt+i represents the i-th input vector at time step t + i, wi is the filter weight, b is the bias term, and is a nonlinear activation function.

The hybrid LSTM+GRU model is constructed to leverage both short- and long-term temporal dependencies in the gait signal. It consists of two LSTM layers followed by a GRU layer. The LSTM layers capture long-range temporal features, while the GRU layer refines the sequential representation with improved computational efficiency [41]. The forward computations for the hybrid structure are summarized as follows:

where xt is the input at time t, and denote the hidden states of the LSTM and GRU layers, respectively, and is the previous cell state in the LSTM. Finally, the output of the GRU layer is passed through fully connected layers, and a sigmoid activation function is applied to perform multi-class gait phase classification.

The Transformer encoder is employed in this study to detect gait phases, leveraging its ability to model long-range dependencies and process sequences in parallel [42]. As the task involves predicting a single class label from an input sequence, only the encoder component is used, omitting the decoder. The core mechanism of the encoder is self-attention, which enables each time step to attend to others in the sequence. It is defined as:

Here, Q, K, and V represent the query, key, and value matrices, and dk is the dimension of the key vectors. This operation allows the model to dynamically weight the influence of each time step when forming the output representation.

Each encoder layer consists of three main components. First, multi-head attention allows the model to jointly attend to information from multiple representation subspaces. Second, a position-wise feed-forward network enhances nonlinearity by applying two dense layers to each time step. Third, sinusoidal positional encoding is added to input embeddings to convey order information:

Through this encoder-only architecture, the Transformer effectively captures global temporal patterns in gait sequences and offers improved computational efficiency over recurrent models, which process data sequentially.

Training configuration

Data partitioning was performed using a subject-wise split to evaluate how well the models generalize to unseen participants. The training dataset consisted of data from 24 subjects, with three sessions per subject allocated to training and the remaining session split for validation. The test dataset included all four sessions from 11 subjects who were not involved in either training or validation. To normalize the statistical distribution of the time-series data, scikit-learn’s [43] StandardScaler was applied so that each sensor channel had zero mean and unit variance. In addition, to prevent biased learning toward particular classes, the training data were resampled to equalize the sample count of the smallest class [44].

In the model training stage, all models were trained for 80 epochs to ensure a fair comparison. However, certain hyperparameters were applied differently across models to achieve optimal performance. Specifically, the CNN and Hybrid LSTM+GRU models applied a dropout rate of 0.3 and the Adam optimizer with a batch size of 1024, whereas the Transformer model was trained with a dropout rate of 0.4, the AdamW optimizer, and a batch size of 128. Other key hyperparameters were optimized for each model using Optuna prior to evaluation. This experimental design was intended to promote stable convergence under limited training data conditions while enabling a fair and quantitative comparison of performance across models.

Results

Hyperparameter

Optuna [45] was employed to perform hyperparameter tuning for each deep learning model, and the results are provided in Table 3. This table presents the explored tuning ranges along with the best hyperparameters, which were selected based on validation accuracy. Using these best parameters, the final test dataset was used to evaluate each deep learning model after training. Each deep learning model, trained with the selected best parameters, is compared using accuracy (ACC) and macro-averaged precision (PR), recall (RC), and F1-score (F1).

thumbnail
Table 3. Optuna-based hyperparameter tuning: ranges and best parameters.

https://doi.org/10.1371/journal.pone.0344002.t003

Multi-class classification results

The analysis of the training and validation loss curves is presented in Figs 68. All three models converged stably in the final stages of training, although their early convergence patterns differed. The CNN exhibited a rapid and smooth reduction in loss from the initial epochs, attributed to its strong inductive bias from locality and weight sharing, as well as its relatively small number of parameters, and ultimately converged to a train loss of 0.14 and a validation loss of 0.11. The Hybrid LSTM+GRU showed greater fluctuations in validation loss during the early epochs, reflecting the time required for the initialization and stabilization of recurrent states and gating mechanisms, but eventually stabilized with a train loss of 0.09 and a validation loss of 0.13. The Transformer was more sensitive to parameter scaling and learning rate settings during the early training phase, resulting in relatively larger fluctuations, but gradually stabilized as training progressed, converging to a train loss of 0.158 and a validation loss of 0.155. Since the differences between training and validation losses remained small and the final loss values converged at low levels across all models, no severe overfitting was observed, indicating that the models were trained stably on the given dataset. These patterns illustrate typical learning dynamics, where the initial instability observed in Transformer and Hybrid LSTM+GRU models is eventually resolved into stable convergence, while the CNN, owing to its structural simplicity and strong inductive bias, converged stably from the outset.

thumbnail
Fig 7. Learning curve of the Hybrid LSTM+GRU.

https://doi.org/10.1371/journal.pone.0344002.g007

The confusion matrices for the test dataset are shown in Tables 4, 5, and 6. Consistent with the stable convergence observed in the learning curves, the models achieved overall balanced classification performance. The confusion matrix was reported using raw counts, and due to the dataset distribution, the number of samples in LS and Sw was about three times larger than in LR and PSw, leading to apparent differences in absolute misclassification counts. However, when interpreted in terms of relative proportions, LR and PSw were more frequently misclassified as Sw, resulting in lower performance for these phases. This misclassification tendency is attributable to the transitional nature of these phases, where LR and PSw are adjacent to Sw and exhibit overlapping acceleration and angular velocity patterns. In contrast, LS and Sw maintained relatively stable classification accuracy despite their larger sample sizes, suggesting that they were less affected by class imbalance. These findings highlight the need to interpret confusion matrices by considering both absolute counts and class distribution, and they reaffirm the common tendency in time-series gait phase classification for misclassifications to concentrate around boundary regions between adjacent phases.

thumbnail
Table 4. Test dataset confusion matrix of the CNN.

https://doi.org/10.1371/journal.pone.0344002.t004

thumbnail
Table 5. Test dataset confusion matrix of the Hybrid LSTM+GRU.

https://doi.org/10.1371/journal.pone.0344002.t005

thumbnail
Table 6. Test dataset confusion matrix of the transformer.

https://doi.org/10.1371/journal.pone.0344002.t006

Table 7 presents the performance comparison of the CNN, Hybrid LSTM+GRU, and Transformer models using hyperparameters optimized with Optuna. Although the Transformer model achieved the highest Macro F1-score of approximately 92.99%, it did not demonstrate a clear performance advantage over the other models, and all three models showed comparable levels of performance. These findings indicate that gait phase classification using an IMUs mounted on the shank can achieve relatively stable and consistent performance across different deep learning architectures.

thumbnail
Table 7. Comparison of classification performance across models.

https://doi.org/10.1371/journal.pone.0344002.t007

Discussion

This study investigated gait phase classification using shank-mounted IMUs, and the performance of CNN, Hybrid LSTM+GRU, and Transformer models was compared. The results showed only minor differences among the models, confirming that reliable classification can be achieved even with a single-sensor setup. In contrast, Su et al. [33] attached seven IMUs to the pelvis, thighs, shanks, and feet, and used foot switches to label five gait phases (LR, MS, TS, Psw, Sw). They reported high accuracies with a frame-based deep convolutional neural network, achieving approximately 97% in intra-subject and over 95% in inter-subject evaluations. However, their approach depended on multiple sensors and foot switches, and did not explicitly account for temporal dependencies in the data. By comparison, the present study utilized contact data for labeling without requiring additional devices, and achieved comparable performance with only shank sensors. Although the overall accuracy was slightly lower than that reported by Su et al. [33], the reduced sensor burden and simplified configuration underscore the practical value of our results. These findings indicate that the proposed minimal-sensor approach can reduce system complexity and cost while still maintaining clinically and practically meaningful performance. Furthermore, by evaluating not only CNN but also Hybrid LSTM+GRU and Transformer architectures, this study provides practical guidance on model selection for real-world wearable applications.

A notable finding of this study is that the Transformer model did not outperform the simpler CNN or Hybrid LSTM+GRU models. This can be explained by several factors. First, level-ground gait exhibits highly repetitive and short-range temporal dynamics, and most discriminative information occurs within short local segments. Prior research has shown that healthy gait patterns demonstrate stable and cyclic structure at short temporal scales [46], limiting the potential benefit of long-range dependency modeling. Second, the sliding-window length used in this study was only 20 frames, which restricts the sequence context available to the Transformer. The literature indicates that Transformers show clear advantages primarily when sufficiently long sequences are provided [47]. Third, Transformer architectures typically require large datasets to fully utilize their representational capacity, whereas smaller datasets often favor simpler models. Prior work has demonstrated that complex deep learning models do not necessarily outperform simpler architectures when data are limited [48]. Finally, IMU-based gait signals are dominated by local temporal patterns—such as short-term accelerometer and gyroscope fluctuations—making convolutional and recurrent architectures inherently well suited for this task. Prior studies have noted that Transformer models excel at capturing global dependencies, while CNNs and RNNs can more effectively learn local temporal structures [49]. These factors collectively explain why the Transformer did not exhibit superior performance under the minimal-sensor, short-window setting used in this study.

Gait phases were defined as four categories: LR, LS, Psw, and Sw. Unlike previous studies that separated MS and TS, we integrated the two into a single LS phase, considering both clinical interpretability and research objectives. This simplified four-phase structure was sufficient to capture essential gait-cycle characteristics. In the results, LS and Sw phases were classified with consistently high performance, whereas LR and Psw phases showed greater classification difficulty. Both LR and Psw represent short transitional phases within the gait cycle, where sensor signals are highly variable and boundaries between phases are less distinct. This tendency is consistent with findings in Su et al. [33], where TS exhibited lower detection accuracy, suggesting that transitional phases remain the most challenging regions for classification models.

In terms of training stability, notable differences were observed among the models. The CNN exhibited stable convergence from the early stages of training, suggesting that its relatively simple architecture facilitated easier optimization. In contrast, both the Hybrid LSTM+GRU and Transformer models initially showed considerable fluctuations in loss, which is typical for architectures that directly model temporal dependencies. However, with sufficient training, both models converged stably, and the Transformer demonstrated the ability to capture complex temporal patterns when appropriate training conditions and optimization strategies were applied. These results indicate that CNNs offer strengths in efficiency and practicality, whereas Transformers demonstrate superior capacity in learning more complex patterns—highlighting the importance of selecting models that match the complexity of the application scenario. For simple and highly periodic environments such as level-ground walking, lightweight models like CNNs may be more suitable than excessively complex architectures.

Regarding phase-specific classification tendencies, the LS and Sw phases consistently demonstrated robust performance, whereas LR and PSw were more frequently misclassified as Sw. This pattern likely arises because LR and PSw are short transitional segments in which signal characteristics gradually shift toward those of the adjacent Sw phase. These results reaffirm that transitional phases remain the most challenging segments for time-series–based gait phase classification and suggest the need for more advanced feature extraction strategies capable of enhancing temporally localized transitional characteristics.

This study demonstrated that deep learning–based gait phase classification can be effectively applied even in a simplified phase framework using a minimal sensor configuration, and that clinically and practically meaningful performance can be achieved with only shank-mounted IMUs. The shank provides a stable and consistent attachment site in wearable and robotic applications, making it more reliable than alternatives such as pelvis-mounted sensors, which often suffer from placement variability and motion artifacts. Although deep learning models require higher computational resources during training, this cost is justified by their ability to learn complex temporal patterns and subtle phase-transition dynamics that simpler methods cannot capture, a benefit that has been consistently demonstrated across time-series learning research [50]. In the inference stage, however, techniques such as model quantization can substantially reduce computation time and memory usage, enabling the proposed models to run efficiently on mobile embedded platforms for real-time gait phase detection [51]. Future research should extend the current work beyond level-ground walking in healthy individuals by incorporating diverse environments—such as stairs, irregular terrains, and pathological gait patterns—and by further validating sensor fusion and real-time deployment strategies. Such efforts will enhance the generalization capability and reinforce the utility of the proposed approach as a practical and reliable gait phase classification method under a wide range of real-world conditions.

Conclusion

In this study, we applied an end-to-end learning approach using time-series data from shank-mounted IMUs to directly predict gait phases, and compared the performance of CNN, Hybrid LSTM+GRU, and Transformer models. The results showed that the overall performance differences among the models were not substantial, and that reliable classification can be achieved even with single-sensor data. In simple environments, lightweight models such as CNNs provided sufficient accuracy, underscoring the importance of selecting model complexity based on the specific requirements of the task. This study was limited to level-ground walking data from healthy individuals. Future work will incorporate more diverse gait environments, such as stairs and irregular terrains, as well as patient populations, to further improve the generalization capability of the proposed models. Through these extensions, the proposed approach is expected to serve as a reliable method for gait phase classification under a broader range of real-world and clinically challenging conditions.

References

  1. 1. Wan C, Wang L, Phoha VV. A survey on gait recognition. ACM Comput Surv. 2018;51(5):1–35.
  2. 2. Agostini V, Balestra G, Knaflitz M. Segmentation and classification of gait cycles. IEEE Trans Neural Syst Rehabil Eng. 2014;22(5):946–52. pmid:24760911
  3. 3. Kharb A, Saini V, Jain YK, Dhiman S. A review of gait cycle and its parameters. Int J Comput Eng Manag. 2011;13:78–83.
  4. 4. Birch I, Vernon W, Walker J, Young M. Terminology and forensic gait analysis. Sci Justice. 2015;55(4):279–84.
  5. 5. Zhang Y, Cao G, Ling Z, Li W, Cheng H, He B, et al. A multi-information fusion method for gait phase classification in lower limb rehabilitation exoskeleton. Front Neurorobot. 2021;15:692539. pmid:34795571
  6. 6. Tao W, Liu T, Zheng R, Feng H. Gait analysis using wearable sensors. Sensors (Basel). 2012;12(2):2255–83. pmid:22438763
  7. 7. Hulleck AA, Menoth Mohan D, Abdallah N, El Rich M, Khalaf K. Present and future of gait assessment in clinical practice: towards the application of novel trends and technologies. Front Med Technol. 2022;4:901331. pmid:36590154
  8. 8. Vafadar S, Skalli W, Bonnet-Lebrun A, Assi A, Gajny L. Assessment of a novel deep learning-based marker-less motion capture system for gait study. Gait Posture. 2022;94:138–43. pmid:35306382
  9. 9. Roerdink M, Coolen BH, Clairbois BHE, Lamoth CJC, Beek PJ. Online gait event detection using a large force platform embedded in a treadmill. J Biomech. 2008;41(12):2628–32. pmid:18657816
  10. 10. Trautmann J, Zhou L, Brahms CM, Tunca C, Ersoy C, Granacher U. TRIPOD—a treadmill walking dataset with IMU, pressure-distribution and photoelectric data for gait analysis. Data. 2021;6(9):95.
  11. 11. Perez-Ibarra JC, Siqueira AAG, Krebs HI. Real-time identification of gait events in impaired subjects using a single-IMU foot-mounted device. IEEE Sensors J. 2020;20(5):2616–24.
  12. 12. Lin F, Wang A, Zhuang Y, Tomita MR, Xu W. Smart insole: a wearable sensor device for unobtrusive gait monitoring in daily life. IEEE Trans Ind Inf. 2016;12(6):2281–91.
  13. 13. Anwary AR, Yu H, Callaway A, Vassallo M. Validity and consistency of concurrent extraction of gait features using inertial measurement units and motion capture system. IEEE Sensors J. 2021;21(2):1625–34.
  14. 14. Benson LC, Clermont CA, Watari R, Exley T, Ferber R. Automated accelerometer-based gait event detection during multiple running conditions. Sensors. 2019;19(7):1483.
  15. 15. González RC, López AM, Rodriguez-Uría J, Alvarez D, Alvarez JC. Real-time gait event detection for normal subjects from lower trunk accelerations. Gait Posture. 2010;31(3):322–5. pmid:20034797
  16. 16. Alahakone AU, Senanayake SMNA, Senanayake CM. Smart wearable device for real time gait event detection during running. In: 2010 IEEE Asia Pacific Conference on Circuits and Systems. 2010. p. 612–5. https://doi.org/10.1109/apccas.2010.5774975
  17. 17. Ben Mansour K, Rezzoug N, Gorce P. Analysis of several methods and inertial sensors locations to assess gait parameters in able-bodied subjects. Gait Posture. 2015;42(4):409–14. pmid:26341531
  18. 18. Gurchiek RD, Garabed CP, McGinnis RS. Gait event detection using a thigh-worn accelerometer. Gait Posture. 2020;80:214–6. pmid:32535399
  19. 19. Tomc M, Matjačić Z. Real-time gait event detection with adaptive frequency oscillators from a single head-mounted IMU. Sensors (Basel). 2023;23(12):5500. pmid:37420666
  20. 20. Romijnders R, Warmerdam E, Hansen C, Schmidt G, Maetzler W. A deep learning approach for gait event detection from a single shank-worn IMU: validation in healthy and neurological cohorts. Sensors (Basel). 2022;22(10):3859. pmid:35632266
  21. 21. Panebianco GP, Bisi MC, Stagni R, Fantozzi S. Analysis of the performance of 17 algorithms from a systematic review. Gait Posture. 2018;66:76–82.
  22. 22. Storm FA, Buckley CJ, Mazzà C. Gait event detection in laboratory and real life settings. Gait Posture. 2016;50:42–6.
  23. 23. Yang M, Zheng H, Wang H, McClean S, Newell D. iGAIT: an interactive accelerometer based gait analysis system. Comput Methods Programs Biomed. 2012;108(2):715–23. pmid:22575802
  24. 24. Saboor A, Kask T, Kuusik A, Alam MM, Le Moullec Y, Niazi IK, et al. Latest research trends in gait analysis using wearable sensors and machine learning: a systematic review. IEEE Access. 2020;8:167830–64.
  25. 25. Derawi MO, Bours P, Holien K. Improved cycle detection for accelerometer based gait authentication. In: 2010 Sixth International Conference on Intelligent Information Hiding and Multimedia Signal Processing, 2010. p. 312–7. https://doi.org/10.1109/iihmsp.2010.84
  26. 26. Soliman A, Ribeiro GA, Torres A, Wu L-F, Rastgaar M. Gait Phase Estimation of Unsupervised Outdoors Walking Using IMUs and a Linear Regression Model. IEEE Access. 2022;10:128090–100.
  27. 27. Tan HX, Aung NN, Tian J, Chua MCH, Yang YO. Time series classification using a modified LSTM approach from accelerometer-based data: A comparative study for gait cycle detection. Gait Posture. 2019;74:128–34. pmid:31518859
  28. 28. Wu Z, Begg RK, Lai DTH, Wang L. Ensemble kNN classifiers for human gait recognition based on ground reaction forces. In: Lecture Notes in Computer Science, 2007. p. 196–207. https://doi.org/10.1109/HSI.2015.7170648
  29. 29. Begg RK, Palaniswami M, Owen B. Support vector machines for automated gait classification. IEEE Trans Biomed Eng. 2005;52(5):828–38. pmid:15887532
  30. 30. Martinez-Hernandez U, Mahmood I, Dehghani-Sanij AA. Simultaneous bayesian recognition of locomotion and gait phases with wearable sensors. IEEE Sensors J. 2018;18(3):1282–90.
  31. 31. Lee M, Roan M, Smith B, Lockhart TE. Gait analysis to classify external load conditions using linear discriminant analysis. Hum Mov Sci. 2009;28(2):226–35. pmid:19162355
  32. 32. Bae J, Tomizuka M. Gait phase analysis based on a Hidden Markov Model. Mechatronics. 2011;21(6):961–70.
  33. 33. Sarshar M, Polturi S, Schega L. Gait phase estimation by using LSTM in IMU-based gait analysis-proof of concept. Sensors (Basel). 2021;21(17):5749. pmid:34502640
  34. 34. Xiang X, Tang S, Yang P, Liu M, Tran M. ED-FNN: a new deep learning algorithm to detect percentage of the gait cycle for powered prostheses. Sensors. 2023;23(18):7723.
  35. 35. Su B, Smith C, Gutierrez Farewik E. Gait phase recognition using deep convolutional neural network with inertial measurement units. Biosensors (Basel). 2020;10(9):109. pmid:32867277
  36. 36. Wiles R, Maharatna K, Conway R, Holden K, Biswas D, Thiyagarajan C. NONAN GaitPrint: benchmarking gait phase segmentation using wearable sensors. Front Digit Health. 2023;5:1124977.
  37. 37. Wang Z, Tang X, Liu W, Li Z. A novel gait phase recognition method based on DPF-LSTM-CNN using wearable inertial sensors. Sensors. 2022;22(7):2595.
  38. 38. Ye J, Wu H, Wu L, Long J, Zhang Y, Chen G, et al. An adaptive method for gait event detection of gait rehabilitation robots. Front Neurorobot. 2020;14:38. pmid:32903323
  39. 39. Arshad MZ, Jamsrandorj A, Kim J, Mun K-R. Gait events prediction using hybrid CNN-RNN-based deep learning models through a single waist-worn wearable sensor. Sensors (Basel). 2022;22(21):8226. pmid:36365930
  40. 40. Tian J. Adversarial vulnerability of deep neural network-based gait event detection. Biomed Signal Process Control. 2022;73:103429.
  41. 41. Rehman A, Saba T, Mujahid M, Alamri FS, ElHakim N. Parkinson’s disease detection using hybrid LSTM-GRU deep learning model. Electronics. 2023;12(13):2856.
  42. 42. Vaswani A, Shazeer N, Parmar N, Uszkoreit J, Jones L, Gomez AN. Attention is all you need. In: Advances in Neural Information Processing Systems, 2017. p. 6000–10.
  43. 43. Kramer O. Machine learning for evolution strategies. Cham, Switzerland: Springer; 2016.
  44. 44. Carvalho M, Pinho AJ, Brás S. Resampling approaches to handle class imbalance. J Big Data. 2025;12:71.
  45. 45. Akiba T, Sano S, Yanase T, Ohta T, Koyama M. Optuna: a next-generation hyperparameter optimization framework. In: 2019. p. 2623–31.
  46. 46. Hausdorff JM. Gait variability: methods, modeling and meaning. J Neuroeng Rehabil. 2005;2:19. pmid:16033650
  47. 47. Grinsztajn L, Oyallon E, Varoquaux G. Why do tree-based models stilloutperform deep learning on typical tabular data? In: NeurIPS Datasets and Benchmarks Track; 2022.
  48. 48. Ahmed S, Shams R, Islam MT. Transformers in time-series analysis: a tutorial. arXiv preprint. 2023.
  49. 49. Thundiyil S, Shalamzari SS, Picone J, McKenzie S. Transformers for modeling long-term dependencies in time series data: a review. In: 2023 IEEE Signal Processing in Medicine and Biology Symposium (SPMB). 2023. p. 1–5. https://doi.org/10.1109/spmb59478.2023.10372632
  50. 50. Bengio Y, Courville A, Vincent P. Representation learning: a review and new perspectives. IEEE Transactions on Pattern Analysis and Machine Intelligence. 2013;35(8):1798–828.
  51. 51. Han S, Mao H, Dally WJ. Deep compression: compressing deep neural networks with pruning, trained quantization and Huffman coding. In: International Conference on Learning Representations (ICLR); 2016.