Skip to main content
Advertisement
Browse Subject Areas
?

Click through the PLOS taxonomy to find articles in your field.

For more information about PLOS Subject Areas, click here.

  • Loading metrics

Objective classification and scoring of movement deficiencies in patients with anterior cruciate ligament reconstruction

  • Chris Richter ,

    Roles Conceptualization, Data curation, Investigation, Methodology, Visualization, Writing – original draft

    mr.chris.richter@gmail.com

    Affiliations Sports Medicine, Sports Surgery Clinic, Dublin, Ireland, Department of Life Sciences, University of Roehampton, London, United Kingdom

  • Enda King,

    Roles Conceptualization, Methodology, Writing – review & editing

    Affiliations Sports Medicine, Sports Surgery Clinic, Dublin, Ireland, Department of Life Sciences, University of Roehampton, London, United Kingdom

  • Siobhan Strike,

    Roles Conceptualization, Methodology, Writing – review & editing

    Affiliation Department of Life Sciences, University of Roehampton, London, United Kingdom

  • Andrew Franklyn-Miller

    Roles Conceptualization, Methodology, Supervision, Writing – review & editing

    Affiliations Sports Medicine, Sports Surgery Clinic, Dublin, Ireland, Centre for Health, Exercise and Sports Medicine, University of Melbourne, Melbourne, Australia

Abstract

Motion analysis systems are widely employed to identify movement deficiencies—e.g. patterns that potentially increase the risk of injury or inhibit performance. However, findings across studies are often conflicting in respect to what a movement deficiency is or the magnitude of association to a specific injury. This study tests the information content within movement data using a data driven framework that was taught to classify movement data into the classes: NORM, ACLOP and ACLNO OP, without the input of expert knowledge. The NORM class was presented by 62 subjects (124 NORM limbs), while 156 subjects with ACL reconstruction represented the ACLOP and ACLNO OP class (156 limbs each class). Movement data from jumping, hopping and change of direction exercises were examined, using a variety of machine learning techniques. A stratified shuffle split cross-validation was used to obtain a measure of expected accuracy for each step within the analysis. Classification accuracies (from best performing classifiers) ranged from 52 to 81%, using up to 5 features. The exercise with the highest classification accuracy was the double leg drop jump (DLDJ; 81%), the highest classification accuracy when considering only the NORM class was observed in the single leg hop (81%), while the DLDJ demonstrated the highest classification accuracy when considering only for the ACLOP and ACLNO OP class (84%). These classification accuracies demonstrate that biomechanical data contains valuable information and that it is possible to differentiate normal from rehabilitating movement patterns. Further, findings highlight that a few features contain most of the information, that it is important to seek to understand what a classification model has learned, that symmetry measures are important, that exercises capture different qualities and that not all subjects within a normative cohort utilise ‘true’ normative movement patterns (only 27 to 71%).

Introduction

Motion analysis systems are widely employed within both universities and clinical facilities, to explore the association between movement patterns and athletic performance or risk of injury, which is of great interest to coaches, physiotherapists and other medical professionals. However, while there are many studies that have examined movements using kinematic and kinetic measurements seeking to identify features (e.g. maximum knee flexion) that might be related to injury, there are no well established evidence-based guidelines that state what movement deficiencies are, what normative ranges of variability are or what a “normal” movement looks like. Additionally, there is little agreement as to which, if any, exercises or movement can expose patterns that could lead to injury or what properties an exercise should hold (single or double leg, movement in one, two or three planes, ecological validity and so on). For example, both Hewett et al., [1] and Krosshaug et al., [2] examined features extracted from a double leg drop jump (DLDJ) to assess their ability to predict the risk of sustaining an anterior cruciate ligament (ACL) injury. Both studies performed a prospective analysis on a large population ([1] = 205; [2] = 782) of female athletes participating in field sports. While Hewett et al., [1] concluded that specific features (describing knee motion and knee loading) were very sensitive in predicting ACL injury, Krosshaug et al., [2] reported poor prediction abilities. The conflict in findings between movement analysis studies, provoked Bahr [3] to state in a recent review: “To date, there is no screening test available to predict sports injuries with adequate test properties and no intervention study providing evidence in support for screening for injury risk”.

A possible source of conflicting conclusions is the way features are extracted. When describing a movement, studies often extract features based on prior knowledge (previous research and / or personal and clinical experience) or post hoc analysis to reduce dimensionality within the highly multivariate datasets, across joints and time. These features are then used to compare the magnitude or timing between groups [3] assuming that the extracted features capture the underlying function of a signal. While discrete points can be helpful in understanding movements, the selection of discrete points has the potential to: discard important information [4, 5], to compare features that present unrelated neuromuscular capacities [6] and to encourage fishing for significance—e.g. non-trivially biased non-directed hypothesis testing [7]. Due to the apparent limitations in discrete point selection, other analyses have been introduced in recent years—statistical parametric mapping [7], (functional) principal component analysis [810], analysis of characterising phases [6], point by point manner testing [11] and other techniques [3, 10, 1214] to improve the analysis of movement patterns.

Another possible source for conflicting conclusions is the way extracted features are compared across groups. Comparisons are often made using statistical significance, and conclusions are developed by inferring properties about a population by testing hypotheses and deriving estimates by probability values (p-values). While, p values are useful they are frequently criticised [1520] due to the possibility of committing type 1 and 2 errors, that any difference can be statistically significant—with large enough sample size [18]—that p values do not provide statistical precision and that conclusions do not account for multiple movement strategies (subgroups) within a dataset. As such, injury related features can be masked during an analysis. Differences in movement strategies within a dataset may be caused by differences in anthropometric measures as well as training, sporting and injury history and there is growing evidence for different movement strategies across individuals [2131].

When examining a human movement, the method of data analysis chosen needs to contend with a complex and multivariate system and any analysis using an inference test may not help to progress the understanding of movement as it does not account for differences in movement strategy or the interrelationship of segments. More suitable methods for movement analysis might be a data driven machine learning techniques that do not require expert knowledge, which have gained popularity in biomechanics and other fields as they have demonstrated an enhanced ability to understand complex and multivariate systems—see Barton et al., [14], Halilaj et al., [32], Rajpurkar et al., [33] or Figueiredo et al. [34]. While studies do not frequently use machine learning techniques in sport-biomechanics, machine learning techniques have been used in gait studies to classify movement data into pathological and able-bodied [35], to detect / judge severity of diseases [36] or to predict surgical / therapy outcomes [37, 38]. For the interested reader Figueiredo et al. [34] reviewed experiments that used machine learning in gait-biomechanics, while Halilaj et al. [32] reviewed the usage of machine learning technique in human movement biomechanics stating best practices and common pitfalls. One of the few studies within sports-biomechanics that used a data-driven supervised learning approach to classify the skill level of athletes (novice and elite) using biomechanical data is Ross et al., [10] and reported an accuracy of 70 to 80%. However, machine learning techniques have not been applied to differentiate movement patterns between non-injured and rehabilitating athletes (e.g. after ACL reconstruction). As such it is not known a) what the most appropriate machine learning technique for such a classification task is, b) what classification accuracies one could expect or c) if movement patterns differ on a group level (e.g. non-injured vs. rehabilitating) or limb level (normal limb vs. operated limb and limb contralateral to operated limb) as movement pattern difference might by driven by anatomical/morphological differences, which could affect both legs or just one limb in rehabilitating cohorts.

The aim of this study was to develop and test a data driven framework (feature generation based on no expert or prior knowledge) to classify movement patterns of normal and rehabilitating athletes using only biomechanical data.

Materials and methods

Subjects

This study used a cohort of males recovering from ACL reconstruction as the rehabilitating cohort (ACL reconstruction should have altered kinematics and neuromuscular properties within the operated knee [39]) and a cohort of healthy males as control / normal group. The ACL group (n = 156) was recruited from the caseload of two orthopaedic surgeons who specialise in knee surgery between January 2014 and December 2016. Inclusion criteria were: biomechanical assessment approximately 9 months after ACL reconstruction, intention of returning to full participation in multi-directional sport after surgery, bone patellar tendon bone graft or hamstring graft from the ipsilateral side, gender (male) and an age between 18 and 35. Exclusion criteria were: multiple or previous ACL reconstructions and any meniscal repairs. The control group (n = 62) contained males that participated in multi-directional sport (i.e. Gaelic Football, Soccer, Hurling, Rugby Union) that were free of injury in the 3 months prior to testing, had no previous knee surgery and were between 18 and 35 years of age. The study received ethical approval from Sports Surgery Clinic Hospital Ethics committee (25AFM010) and was registered on clinicaltrials.gov (NCT02771548). The examined data were fully anonymised and all subjects provided informed written consent to have their data used for research.

The ACL group had an average age of 24.8 ± 4.8 years was 180 ± 8 cm tall and had a body mass of 84 ± 15.2 kg. The control group had an average age of 24.8 ± 4.2 years was 183 ± 6 cm tall and had a body mass of 82 ± 8.9 kg.

Data capture and pre-processing

The testing took place in the motion analysis laboratory using an eight-camera motion analysis system (200Hz; Bonita-B10, Vicon, UK), synchronised with two force platforms (1000Hz BP400600, AMTI, USA). Before data collection, all subjects undertook a standardised warm-up and wore their own athletic footwear with 24 reflective markers secured to the shoe or to the skin using tape, at bony landmarks according to the Plug-in-Gait marker set. Three trials of each limb for the following seven exercises were captured: double leg (DL) countermovement jump (CMJ; [40]), single leg (SL) CMJ [40], DL drop jump (DJ; [40]), SLDJ [40], Hurdle Hop (HuHo; [40]), SL Hop (SLHop; [40]), planned change of direction (CoDP [41]) and unplanned change of direction (CoDU [41]). Marker and force data were low-pass filtered using a fourth-order Butterworth filter [42]—before computing kinematic and kinetic measures using Nexus (1.8.5; Vicon, UK). Data pre-processing (gap filling and waveform screening) was performed on examined biomechanical measures using a custom developed MATLAB program (R2015a, MathWorks Inc., USA). All kinetic variables were normalised to body mass. The start and end point of an exercise was defined using the force trace or a combination of center of mass (CoM) power and force trace [40, 41] and all measures were landmark registered using a dynamic time warping process [43] to align the end of the eccentric phase across the all curves [40, 41]. Only the maximum trial of the captured (based in maximal jump height and shortest contact time) was chosen for data analysis.

The framework

The steps taken during data analysis can be described as follows: feature engineering, selection of a supervised learning technique, feature selection and final validation of minimal feature model. The performance of any generated classification model described in this study is the average classification accuracy of a 100 split stratified shuffle split cross-validation. A shuffle spilt cross-validation was used in preference to a k-fold cross validation because, although information leaking can occur between the splits, it allows the re-use of subjects that have been assigned previously to the trainings group and hence a number of splits and greater sample size during the training / fitting of the prediction model. The description of the generation of the framework is based on a single exercise and was done for every exercise separately.

During the performed splits, false and correct classifications were recorded to allow the generation of a confusion matrix for every model generated on a group (e.g. ACL and NORM), limb (class: ACLOP, ACLNO OP and NORM) and subject level. This information is essential when examining the prediction model, in respect to errors (confusion or misclassifications) made and can also help in understanding the examined data—e.g. do features differ on a group or limb level (misclassifications of class but not group level), do some subjects behave differently than others within a class (misclassification of specific subjects) or if one model made misclassifications another did not (could to models complemented each other). In addition to recording false and correct classifications, during every split a “guess prediction” was made with each trial having a 1 in 3 chance of belonging to the class: ACLOP, ACLNO OP or NORM. The best guess, the highest guess classification accuracy observed within splits, was then used as a threshold to judge the meaningfulness of the generated models. A classification model was defined meaningful if the lowest observed accuracy during the splits was greater than the best guess.

This study evaluated the created models on a number of levels: the number of features selected for inclusion (3 levels: all features, 20 features, minimal feature), the machine learning technique (7 widely used techniques were applied) and movement type (8 movement frequently used to assess the rehabilitation status of athletes with ACL and to predict future injury). Except of the minimal feature model, the evaluation was performed by assessing the average classification accuracy. Evaluation of the minimal feature model was based on average classification accuracy, classification errors and sensitivity. Other classification performance measures e.g. area under the curve, precision, recall, f1-score or specificity would very likely have resulted in different models and performance rankings.

Feature engineering.

Traditional analyses, which use discrete measures, are dependent on the selection of features. To enable a data-driven feature engineering that does not require expert knowledge, this study reduced dimensionality of commonly examined motion capture time-series by examining variability within the time-series across examined in 100 random shuffle splits using analysis of characterising phases [6]. This information was then used to reduce a time-series to just a few features that describe the variability of the signal across all participants. This process was performed on each examined time series separately. The following 82 time series were examined separately: ground reaction force (GRF; x, y, z), GRF impulse (x, y, z), center of mass (CoM) velocity (xy, xyz, z), CoM power (x, y, z), CoM in pelvis, hip, knee and ankle (x, y, z) as well as joint angles of the ankle, knee, hip, pelvis, thorax and thorax on pelvis in sagittal, frontal and transversal planes, joint angular velocities of the ankle, knee, hip, pelvis, thorax and thorax on pelvis in sagittal, frontal and transversal plane, joint powers, moments, work and impulse of ankle, knee, hip and pelvis in sagittal, frontal and transversal plane, time and the rotation foot angle to pelvis.

Before performing the stratified shuffle splits, a matrix (memory matrix) was defined for every examined signal consisting of 101 rows (corresponding to length of the time-series) and 100 columns (corresponding to the number of splits) storing “0” values. The first step within every split was the random selection of 50 subjects from the ACL and NORM cohort to create a dataset of 200 trials—containing 50 ACLOP limbs, ACLNO OP limbs and 2x50 NORM limbs (Fig 1—step 1). Analysis of characterising phases (described in [6]) was then used to identify phases of variation within the created dataset. For every time point that was within a detected phase of variation, the default “0” value was changed to “1” for the corresponding row and column in the memory matrix. For example, if the described process identified the time points from 5 to 15 within a time series to be a phase of variation in the 7th split, then the value of row 5 to 15 in column 7 in the memory matrix was changed to “1” (Fig 1—step 2).

thumbnail
Fig 1. Illustration of the work-flow during the features engineering.

https://doi.org/10.1371/journal.pone.0206024.g001

After completing the 100 splits, the sum of the values in the memory matrix was computed along the split axis generating a 101x1 vector that stored the accumulated appearance as a phase of variation for each frame. A value of “0” indicates that this time point was never within a phase of variation, while the value 84 indicates that a time point was 84 times within a phase of variation (Fig 1—step 4). Every collection of time points that occurred in 75% of the splits and spanned at least 5% of the movement cycle was considered robust phase of variation and used for the feature engineering. The investigators used a 75% threshold, instead of 95% [44], to maintain more information—i.e. allowing more phases to be classified as robust phases of variation. Features describing a phase were calculated as the average value of a phase and stored into a feature matrix that contained i rows / features (i is the number of detected phases of variation) and n = 436 columns / observations (156 operated limbs [class: ACLOP], 156 limb contralateral to the operated limb [class: ACLNO OP] and 124 control limbs [class: NORM]). For every double leg exercise, symmetry was also calculated as limbcontralateral limb—e.g. leftright for left leg trials or rightleft for right leg trials.

During the feature engineering the data examined for each exercise generated between 97 and 186 features for each of the 436 observations (Table 1). The detailed description of results of this step is given in S1 Appendix—Detected Phases. These features were used within the following steps as input features during the fitting (training) and validation (testing) of every generated model.

thumbnail
Table 1. Number of extracted features that were used as input features during fitting (training) and validation (testing) of the supervised learning techniques.

https://doi.org/10.1371/journal.pone.0206024.t001

Selection of supervised learning technique.

The second step in the model generation was the identification of the most appropriate machine learning technique. This process was done because the selection of the best learning technique can be challenging as there are multiple techniques and each technique has different abilities to learn relationships between the to be predicted classes and input features [45, 46]. This study examined only the ability of techniques included within the statistical toolbox of Matlab (R2015a, MathWorks Inc., USA): a decision tree (fitctree), an ensemble of decision trees (n trees = 50; TreeBagger), a discriminant analysis model (fitcdiscr), a naive Bayesian classifier (fitcnb), a k-nearest neighbour model (k = 5; fitcknn), a multi class model for support vector machines (fitcecoc), a linear regression model (mnrfit; in stepwise forward) and a neural network (patternnet). No grid search was performed at any time and is it very likely that accuracies in better-tuned models will outperform the presented findings. The random assignment of observations into training and testing dataset, during a split, was performed on athlete level to prevent data leakage from training into testing data—e.g. if a subject was selected to belong to the training dataset both the right and left limb were used within the training. Each learning technique was tested separately using raw values and normalised values (z-scores) of the feature matrix.

During a split, the feature matrix built during the feature engineering step was split into training, testing and hold out dataset (Fig 2—step 1). The sample size of the datasets was chosen based on the NORM sample size (training ≈ 70%, testing ≈ 25% and hold out ≈ 5%). Hence, the training dataset contained around 88 trials from each class (≈ 56% of ACLOP and ACLNO OP) resulting in a training matrix containing i features and ≈ 264 trials. The testing dataset contained around 30 trials from each class (≈ 20% of ACLOP and ACLNO OP) resulting in matrix containing i features and ≈ 90, while the hold out dataset contained the remaining trials. The training dataset was used to teach each machine learning technique to predict the three classes using the previously extracted features. After the training had been completed, class membership of the testing dataset was predicted and compared to the actual class to assess the performance of each learning technique (Fig 2—step 3). When using normalised scores, the testing dataset was normalised using the mean and standard deviation of the training data. The purpose of the hold out dataset was increase the “interchange” of trials within splits, increasing the variation across datasets within the splits and facilitated balanced class distributions within the datasets.

thumbnail
Fig 2. Illustration of the workflow to identify the most suitable learning technique.

https://doi.org/10.1371/journal.pone.0206024.g002

After the stratified shuffle split cross-validation was completed, the learning technique with the highest mean accuracy across the splits was judged most appropriate and used for subsequent analysis (Fig 2—step 5).

Findings during this step demonstrated small differences between classifications accuracies from raw and z-scores (1%). Based on the normalised scores (z-scores), this process identified that the neural network performed best on the DLDJ features (80%), the logistic regression performed best on the SLCMJ (57%), DLCMJ (72%), HuHo (69%) and SLHop (72%) features, while the discriminant analysis performed best on the SLDJ (60%), CoDP (68%) and CoDU (65%) features. The results of this step are described in Table 2 and are visualised in detail in S2 Appendix—Best Learning Technique.

thumbnail
Table 2. Description of the performance of the supervised learning techniques using all extracted features within an exercise.

https://doi.org/10.1371/journal.pone.0206024.t002

Feature selection.

The third step was the identification of the features that are most important in the classification. This process used only the best performing machine-learning technique and sought to reduce the effect of over-fitted models by identifying the minimal number of features required to capture most of the information within an exercise. To identify the most important features, a wrapper method (evaluate performance of subsets of features in a forward selection matter) in combination with a stratified shuffle split cross-validation with 100 splits was used (Fig 3). Again, no grid search was performed.

thumbnail
Fig 3. Illustration of the workflow to identify key features toward classification.

https://doi.org/10.1371/journal.pone.0206024.g003

Before starting the selection of important features, an empty matrix (model base) was defined. After the creating of this model base matrix a stratified shuffle split cross-validation with 100 splits was started assessing the classification accuracy of every feature within the feature matrix on its own. In every split the data were divided into training, testing and hold out datasets (Fig 3—step 1). The training dataset was used to teach the previously selected machine-learning technique to predict the three classes (ACLOP, ACLNO OP and NORM). After the training phase had been completed, comparing the predicted class of the testing dataset to the actual class assessed the performance of the generated model. After the 100 splits, the feature with the highest mean accuracy was identified, added to the model base matrix and removed from the feature matrix (generated during the feature engineering). All features that correlated with the identified feature (greater than 0.70) were also removed from the feature matrix, to increase interpretability. The value of.70 was a subjective choice as this magnitude is often considered as strong. Subsequently, a second stratified shuffle split cross-validation with 100 splits was started assessing the classification accuracy of the model base matrix in combination with every remaining feature within the feature matrix. The feature that performed best was added to the model base matrix. The just added and any collinear feature where then removed from the feature matrix (Fig 3—step 5). This process was repeated until the model base matrix grew to 20 features (this number was chosen to keep computing time to a reasonable duration; Fig 3—step 6).

The last step during the feature selection was determining how many features should be within a model that presents a good trade-off between efficiency and effectiveness (minimal feature model—e.g. the model with the smallest number of features that accounts for a large part of the contained information). To do this the Elbow method was used—described in Hastie and Tibshirani [45] or Vapnik and Vapnik [46], with the “elbow” being defined as the point n where the differentiation of accuracy f improved less than 10% of its range (Eq 1). (1)

The detailed description of results of this step is given in S3 Appendix—Best Features.

Final validation of minimal feature model.

The number of parameters and models that were tested in this study is large and information could have leaked during the shuffle spilt cross-validation (e.g. possibly over fitting the presented models). To examine if a model was over fitted a 10 fold cross-validation was used to re-evaluate the minimal feature models. During the cross-validation the minimal feature model was trained using 156 subjects (n = 52 per class; ≈90% of Norm and ≈33% of ACLOP and ACLNO OP) and validated using 18 subjects (n = 6 per class; ≈9% of Norm and ≈4% of ACLOP and ACLNO OP) as testing data. Ten folds were chosen to balance the number of subjects in testing and training with an acceptable accuracy resolution (1 in 18 = 5.56%). The mean accuracy of the 10 fold were then compared to the mean accuracy of the previous used shuffle spilt cross-validation.

Matlab (R2015a, MathWorks Inc., USA) was used for data processing and analysis.

Results

Findings reported are based on normalised scores (z-scores), for two reasons: the difference between classifications accuracies for the raw and z-scores was small at less than 1% at most and z-scores remove any possible magnitude effect during the selection of features within the optimal model. The highest guess observed during the performed splits was 42%.

Model performance evaluation

Generated models were generally judged successful in classifying the classes (ACLOP, ACLNO OP and NORM). All 3 models generated within an exercise (all features, 20 features, minimal features) were judged meaningful except for the minimal feature for the SLCMJ and CoDU movements. The classification accuracies increased on average 6.1% when reducing the number of features from every extracted feature to 20 features (Table 3). The classification accuracies of the minimal feature models were comparable to the full models (average difference -4.4%; Table 3). When comparing the full model to the minimal feature model: the performance of the DLCMJ, DLDJ, SLHop and HuHo stagnated or improved slightly (0 to 4%), while a slight decrease in performance (-4 to -5%) was observed in the SLCMJ, SLDJ and CoDP. The CoDU minimal feature model decreased its performance by 14% compared to the full model and 15% to the 20 feature model.

thumbnail
Table 3. Description of performance observed during the model generation process (using all features, 20 most important features and optimal features.

https://doi.org/10.1371/journal.pone.0206024.t003

Minimal feature models: Accuracies, errors and sensitivity

The minimal feature models used between 2 and 5 features (Table 4) and the classification accuracies observed were (in decreasing order): DLDJ (81%), SLHop (74%), DLCMJ (70%), HuHo (69%), CoDP (63%), SLDJ (57%), SLCMJ (53%) and CoDU (52%). The detailed description of results of this step is given in S4 Appendix—Minimal Feature Models.

thumbnail
Table 4. Description of minimal feature models: Used machine learning techniques, identified features and corresponding accuracies.

https://doi.org/10.1371/journal.pone.0206024.t004

Results for the 10-fold cross-validation of the minimal feature model demonstrated differences in mean accuracy to the shuffle-split cross-validation (Table 4; in decreasing order): DLCMJ (-11%), CoDU (+11%), SLHop (5%), DLDJ (-4%), CoDP (-3%), HuHo (+3%), SLDJ (0%) and SLCMJ (0%).

Classification errors originated from misclassifications of the class ACLOP and ACLNO OP in the optimal SLHop and HuHo model (Table 5). In these models the percentage of misclassification within the ACL group was higher (≈ 2 times) than misclassification of either ACL class with NORM, while the SLCMJ model tended to misclassify both ACL classes with NORM class. The DLCMJ and DLDJ results displayed increased misclassifications between NORM and either ACL class compared to between the ACL classes (≈ 1.8 and 4 times; Table 5). The SLDJ displayed increased misclassifications (≈ 2 times) between the ACLNO OP and NORM class compared to ACLOP with ACLNO OP or NORM (Table 5). In the CoDP and CoDU models, misclassifications were equally distributed between the classes (Table 5).

thumbnail
Table 5. Description of the confusion matrix for the minimal feature model of every examined exercise: Reporting true positive, true negative, false positive and false negative as count (percentage across all classes).

https://doi.org/10.1371/journal.pone.0206024.t005

When assessing the performance of minimal feature models on a group level (ACL [ACLOP and ACLNO OP] and NORM) the single leg exercises demonstrate large improvements (12 to 24%), while the double leg exercises did not (2 to 8%; Table 6).

thumbnail
Table 6. Description of class specific sensitivity, performance of binary classification (ACL and NORM) and percentage of the athletes within the NORM group that were never confused.

https://doi.org/10.1371/journal.pone.0206024.t006

The sensitivity for the classification of the NORM class was highest in the SLHop (81%), while for the sensitivity for ACLOP and ACLNO OP trials was highest for the DLDJ (84 and 85%; see Table 6).

The percentage of the athletes within the NORM group that were never confused (’true’ NORMs) ranged from 27 to 71% depending on the exercise (Table 6). The percentage of the athletes within the ACLNO OP and ACLOP group that were never confused ranged from 52 to 85% depending on the exercise (Table 6).

Discussion

This study examined the classification accuracy of a framework that combines a data-driven feature extraction procedure that did not require expert knowledge and a supervised machine learning technique to differentiate movement patterns from ACL operated class (ACLOP), from a limb contralateral to an ACL operated limb (ACLNO OP) and a non-injured control limbs (NORM) using only biomechanical data. Findings demonstrate that biomechanical data, regardless of the exercise, contained sufficient information to outperform guessing and suggest that it is possible to differentiate normal movement patterns from movement patterns after an ACL reconstruction on a group and limb level.

Model evaluation

Classification accuracy.

This study developed a minimal feature model in 3 steps that generated three prediction models: one that utilised a large number of features, one model utilised only 20 (most important) features and a minimal feature model. The minimal feature model contained 2 to 5 features and classification accuracies were similar between the three generated models. For the DLCMJ, DLDJ, HuHo and SLHop, using the minimal feature models reduced the information utilised by 94 to 98% (160 features vs. ≈ 4 features) but achieved performances similar (-2 to 2%) to models using all generated features. In contrast, the performance of the minimal feature model decreased for the SLCMJ, SLDJ, CoDP and CoDU in comparisons to the model using all generated features (-5 to -12%). The decrease in accuracy in the minimal feature model SLCMJ, SLDJ, CoDP and CoDU indicates an increased complexity in these exercises, compared to the DLCMJ, DLDJ, HuHo and SLHop, and suggests that these models need more information (features) to perform as well as they could. However, all exercises improved performances when reducing the number of features from 97 to 186 to only 20, which is a reduction of used features of ≈ 84% (138 features vs. 20 features), by ≈ 7%. This suggests that there was an effect of over-fitting to the training data in the models that used all extracted features and highlights that the importance of some kind of feature selection before model building using biomechanical analysis using comparable sample sizes. However, this finding also highlights that a relatively small number of biomechanical features (between 2 and 20) captures a large amount of the information content from a biomechanical assessment. Regarding the information contained within the examined movement data when assessing classification accuracies in a multi-class classifier, classification models achieved accuracies of (in decreasing order): DLDJ (81%), SLHop (75%), DLCMJ (73%), HuHo (69%), CoDP (58%), SLDJ (56%), SLCMJ (53%) and CoDU (52%). Little research exists currently that used biomechanical data to classify group / cohort membership. The closest research is that of Ross et al., [10] who analysed the movements in elite and non-elite performers and found that a data driven model could classify an individual into novice and elite with an classification accuracies between 71 to 80% across 13 exercises. While the classification accuracies of the SLCMJ, SLDJ, CoDP and CoDU seem to be low in comparison to other exercises or the findings of Ross et al., [10], it should be considered that their performance was assessed in an multi-class classifier. The classification accuracy of the SLCMJ, SLDJ, CoDP and CoDU in a binary or group classifier (ACL or NORM) was 67, 69, 77 and 70%, making them comparable to reported classification accuracies observed by Ross et al., [10].

Classification errors.

Classification errors in the tested multi-class classifiers could have occurred due to a reduced / smaller difference in movement within the ACL classes compared to the ACL and NORM classes—e.g. because of anatomical / morphological differences between the groups (ACL and NORM). When assessing the performances on a group level (ACL and NORM) performances improved significantly, compared to the limb class level, for every single leg exercise (≈ 16%) but not for double leg exercises (≈ 5%), suggesting differences between the single and double leg exercises. These differences might be explained by the additional symmetry information that was contained in the double exercises but not in the single leg exercises. This assumption is supported by the fact that in the double leg exercises, symmetry features were selected frequently (4 out of 7). The magnitude of symmetry or asymmetry seems to be a useful feature within the generated models, which supports findings of Myer et al., [47] and others [4850]. While symmetry features could have been included within the single leg models, it requires the intervention of the investigators—e.g. symmetry calculation as mean symmetry, symmetry between trial x and trial y and so on [51]. Symmetry was not included in this study because it cannot be calculated without setting subjective rules (e.g. expert knowledge), as every execution presents different external and internal conditions. Nevertheless, findings suggest that symmetry measures are important and hence they should have been considered.

Regarding errors that might have been caused by anatomical or morphological differences between the groups, a large part of error in the SLHop and HuHo classification errors originated from confusing ACLOP and ACLNO OP (see Fig 4). In these models the percentage of misclassification within the ACL group was higher (≈ 2 times) than confusing either ACL class with NORM. This suggests that the movement of both limbs in the ACL group is affected by an ACL reconstruction—as both limbs can be differentiated from the NORM pattern but less from each other. As such, the ACLNO OP may not be ideal reference when judging the rehabilitation status and readiness to return to play. Another reason could be that the differences between ACLOP and ACLNO OP are not well described in the examined features.

thumbnail
Fig 4. Illustration of the results of the SLHop model: The model accuracy is displayed on the top right, while the confusion matrix of a 2-feature model is displayed below (right button).

The three graphs on the left display the confusion pattern (button) and selection frequency of each trail within the ACLOP, ACLNO OP and NORM (top).

https://doi.org/10.1371/journal.pone.0206024.g004

The SLDJ, DLDJ, SLCMJ, CoDP and CoDU models demonstrated different error or misclassification patterns. The SLDJ would have benefited most from a binary-classifier that classified for ACL reconstructed limb (ACLOP) and not ACL reconstructed limb (ACLNO OP and NORM), as a large part of misclassifications were made by confusing ACLNO OP and NORM (confusions within the ACL group were less common). This would suggest that the SLDJ is able to detect differences between a limb with ACL reconstruction and limb without ACL reconstruction—implying that the ACLNO OP limb could be used as reference when judging the progress of rehabilitation. The interpretation of the selected features in the minimal feature model, however, suggests that the execution of the exercise was modified for within the ACL classes and between the groups. The vertical CoM velocity at impact was selected yet it should not hold any meaningful information, as it should be nearly the same (impact velocity should theoretically be 1.981 m/sec.) across all trails because of the controlled drop height. Inspection the model visually reveals: a) that the ACLNO OP and NORM class demonstrate similar CoM resultant velocity at take off values, while values in ACLOP are reduced, and b) that the vertical CoM velocity at impact values are different between ACLNO OP and NORM and spanned over the range of classes for the ACLOP class (Fig 5). Based on a post hoc analysis, the ACLNO OP and some of the ACLOP class have ‘changed’ the drop height by adapting a stepping down pattern (Fig 5), which was not detected and resolved during the biomechanical assessment, in spite of careful instruction and frequent lab quality assessments. This highlights is the importance of interrogating what the model has learned during a data driven process as other features should be been normalised to the ‘changed’ the drop height and highlights possible psychological differences between the group [52].

thumbnail
Fig 5. Illustration of the 2 most important SLDJ model features (CoM resultant velocity at take off values and vertical CoM velocity at impact) plotted against each other on the left and the CoM velocity at impact for the logic cases on the right.

Fully coloured bubbles represent trials with ‘true’ group pattern, while faded coloured bubbles represent trails that did not.

https://doi.org/10.1371/journal.pone.0206024.g005

In the DLCMJ and DLDJ, errors originated from misclassifying NORM with either ACL class or both ACL classes with NORM, with the DLDJ demonstrating a larger inequality in distribution (≈ 1.8 vs. 4 times). This might be explained by an increase of the physical demand of the DLDJ that poses an additional challenge to symmetry and reactivity. While the minimal feature SLCMJ model demonstrated classification errors that originated equally from confusion between the classes, the 20 feature or all feature SLCMJ model displayed that confusion between ACLNO OP and NORM were ≈ 2 times higher compared to other classification errors. This suggests that models that utilised more features possibly detected some patterns that differ between ACL operated and not operated limbs (when being provided with information from many features). Similar to the SLCMJ the minimal feature CoDP and CoDU models demonstrated classification errors that originated equally from confusion between classes, while the 20-feature model displayed an increased error from misclassification within the ACL group (≈ 2 times). This would suggest that the ACLOP and ACLNO OP class share more similarity with each other than to the NORM class.

Classification sensitivities.

Another source of error for the models is the misclassification of specific athletes within a class that does not present the ‘true’ class movement patterns. When examining the prediction errors on an athlete level, it was noted that some athletes are always classified correctly; some are occasionally classified correct, while some are always misclassified. This was true for all classes and exercises. The detection of members within a group that are not always classified correctly could help to understand group differences [30] as they could be excluded from subsequent analysis (as done in Fig 5 to better understand the class pattern the model has learned). The existence of such athletes or the proportion of such athletes within a group may also explain why there are so many conflicting findings in inference based studies. However, knowing that such athletes exist also enables the development of our understanding of injuries as the probability of belonging to a desired class computed during a classification could be used when judging movement analysis and injury risk. Consequently, the probability of membership to a class (in this case the NORM class) could also allow the objective judgment of a movement or the rehabilitation progress. The probability of membership can give an objective measure of how close a trial is to a desired class and presents a clear criterion if an athlete has returned to normal.

Practical implications

Return to play after ACL reconstruction as well as the prevention of subsequent re-injury is not always guaranteed [53, 54] and this might be in part due to absence of clear criteria identifying if an athlete has returned to pre-injury levels or completed rehabilitation. Current clinical testing batteries often utilise biomechanics to assess a movement quality. However, there is little consensus on the appropriateness of biomechanical analysis and / or specific exercise tests and measures when differentiating between two specific groups. This study demonstrates that biomechanical data holds enough information to differentiate between ACLOP, ACLNO OP and NORM with classification accuracies above 70%, even if the ACL group was 9.4 ± 0.7 months post-op and most had completed their rehabilitation.

Findings suggest that classification models from different exercises capture different pieces of information and hence a variety of exercises should be used. A biomechanical assessment should contain measures of symmetry and chosen exercises might be altered throughout the rehabilitation process to increase complexity, while familiarising the athlete to the demands of sports. The selection of a single exercise test cannot be recommended as all examined exercises demonstrated that they contain valuable information and we did not adjust parameters within the models that could lead to improvements in classification accuracy.

Findings highlight problems with the assumption that the majority of a control group demonstrate healthy movement patterns, which could be a reason for conflicting findings in studies. Within the examined exercises only 27 to 71% of the limbs within NORM did present a true NORM pattern.

Limitations

Examined data.

Like most movement analysis studies, this study involved the recording of multiple trials but examined only the best trial. Examining the best trial is likely to be more valid than averaging multiple trials, which creates and examines an artificial movement where local peaks are altered in magnitude and temporal appearance [55] and the intersegmental link (coordination) between joints might be lost. However, the approach used selects a unique instance (maximal performance) and could bias an analysis towards a non-realistic situation. No athlete will perform a task over and over with a maximal effort and consequently the sub-maximal efforts should not be discarded. An alternative approach is to utilise repeated random sampling, where the captured trials are selected at random and the analysis is run multiple times [5658]. This can overcome the ‘maximal effort bias’ and can also provide a measure of expected differences or accuracy within future studies. Further, such an approach can also overcome discrepancies between findings that are caused by the selection of a reference limb when comparing an abnormal (injured) to a normal or uninjured group.

Examined features.

The use of waveform analysis and total exclusion of discrete points ignores the magnitudes and location of peak values that might be important. For example, angular velocity and GRF time series can be multimodal (multiple local maxima’s) and maximal or minimal might not be located in the same phases and the information of these features is hence lost. Such features and their temporal properties (position) might carry important information that has been discarded in this study. Further, the manner in which the data driven framework approached the engineering feature was based on no input of expert knowledge (e.g. complexity [59], stiffness [60], variability [61]) or any advanced feature engineering techniques (principal component scores, non negative matrix factorisation or interactions between features). No additional engineering feature was used to keep the interpretable and meaningful for rehabilitation, while additional more complex features are likely to increase the reported classification accuracies.

Model building.

This study tested machine learning techniques using only their default values and did not perform a grid search (optimise adjustable parameters within a technique; e.g. changing L1 and L2 regularisation in logistic regression or changing k in the k-nearest neighbour technique), which would very likely have resulted in higher classification accuracies and might impact the selected features in the optimal model. Additionally, for the DLCMJ and CoDU findings should be interpreted with care as findings from a 10-fold cross-validation possibly indicate over-fitting for the DLCMJ model and an under-fitting for the CoDU model.

Conclusion

This study tested a data-driven framework that required no expert or prior knowledge and demonstrated that biomechanical data can predict with high accuracy (≈ 70%) if a movement was performed by a limb following ACL reconstruction, the contralateral limb and a limb of a healthy control group. Findings highlight that a few features can contain most of the information content, that symmetry measures are important, that it is important to seek to understand what a classification / prediction model has learned, that different exercises capture different movement characteristics and that not all subjects within a normative cohort utilise a ‘true’ normative movement pattern.

Acknowledgments

We are thankful to our colleagues who provided expertise that greatly assisted the data processing or though process of this study.

References

  1. 1. Hewett TE, Myer GD, Ford KR, Heidt RS, Colosimo AJ, McLean SG, et al. Biomechanical measures of neuromuscular control and valgus loading of the knee predict anterior cruciate ligament injury risk in female athletes. The American journal of sports medicine. 2005;33(4):492–501. pmid:15722287
  2. 2. Krosshaug T, Steffen K, Kristianslund E, Nilstad A, Mok KM, Myklebust G, et al. The vertical drop jump is a poor screening test for ACL injuries in female elite soccer and handball players: a prospective cohort study of 710 athletes. The American journal of sports medicine. 2016;44(4):874–883. pmid:26867936
  3. 3. van Emmerik RE, Ducharme SW, Amado AC, Hamill J. Comparing dynamical systems concepts and techniques for biomechanical analysis. Journal of Sport and Health Science. 2016;5(1):3–13. pmid:30356938
  4. 4. Dona G, Preatoni E, Cobelli C. Application of functional principal component analysis in race walking: an emerging methodology. Sports Biomechanics. 2009;8(4):284–301. pmid:20169759
  5. 5. Donoghue OA, Harrison AJ, Coffey N, Hayes K. Functional data analysis of running kinematics in chronic Achilles tendon injury. Medicine and Science in Sport and Exercise. 2008;40(7):1323–35. pmid:18580414
  6. 6. Richter C, Marshall B, Moran K, et al. Comparison of discrete-point vs. dimensionality-reduction techniques for describing performance-related aspects of maximal vertical jumping. Journal of biomechanics. 2014;47(12):3012–3017. pmid:25059895
  7. 7. Pataky TC, Robinson MA, Vanrenterghem J. Vector field statistical analysis of kinematic and force trajectories. Journal of Biomechanics. 2013;46(14):2394–2401. pmid:23948374
  8. 8. Jolliffe I. Principal component analysis. John Wiley & Sons; 2005.
  9. 9. Warmenhoven J, Cobley S, Draper C, Harrison A, Bargary N, Smith R. Considerations for the use of functional principal components analysis in sports biomechanics: examples from on-water rowing. Sports Biomechanics. 2017;0(0):1–25.
  10. 10. Ross GB, Dowling B, Troje NF, Fischer SL, Graham RB. Objectively Differentiating Movement Patterns between Elite and Novice Athletes. Medicine and science in sports and exercise. 2018;.
  11. 11. Dixon PC, Stebbins J, Theologis T, Zavatsky AB. Spatio-temporal parameters and lower-limb kinematics of turning gait in typically developing children. Gait & Posture. 2013;xxx(xxx):xxx.
  12. 12. Chau T. A review of analytical techniques for gait data. Part 1: fuzzy, statistical and fractal methods. Gait & Posture. 2001;13(1):49–66.
  13. 13. Chau T. A review of analytical techniques for gait data. Part 2: neural network and wavelet methods. Gait & Posture. 2001;13(2):102–120.
  14. 14. Barton GJ, Hawken MB, Scott MA, Schwartz MH. Movement Deviation Profile: A measure of distance from normality using a self-organizing neural network. Human movement science. 2012;31(2):284–294. pmid:20728953
  15. 15. Neyman J, Pearson E. On the problem of the most efficient tests of statistical inference. Biometrika A. 1933;20:175–240.
  16. 16. Goodman S. A dirty dozen: twelve p-value misconceptions. In: Seminars in hematology. vol. 45. Elsevier; 2008. p. 135–140.
  17. 17. Nester MR. An applied statistician’s creed. Applied Statistics. 1996; p. 401–410. https://doi.org/10.2307/2986064
  18. 18. Ranstam J. Why the P-value culture is bad and confidence intervals a better alternative; 2012.
  19. 19. Leek J, McShane BB, Gelman A, Colquhoun D, Nuijten MB, Goodman SN. Five ways to fix statistics. Nature. 2017;551(7682):557–559. pmid:29189798
  20. 20. Amrhein V, Greenland S, McShane B. Scientists rise up against statistical significance; 2019.
  21. 21. Richter C, Marshall B, Moran K, et al. Clustering vertical ground reaction force curves produced during countermovement jumps. Journal of biomechanics. 2014;47(10):2385–2390. pmid:24845694
  22. 22. Carriero A, Zavatsky A, Stebbins J, Theologis T, Shefelbine SJ. Determination of gait patterns in children with spastic diplegic cerebral palsy using principal components. Gait & Posture. 2009;29(1):71–75.
  23. 23. Kienast G, Bachmann D, Steinwender G, Zwick EB, Saraph V. Determination of gait patterns in children with cerebral palsy using cluster analysis. Gait & Posture. 1999;10(1):57.
  24. 24. O’Byrne JM, Jenkinson A, O’Brien TM. Quantitative analysis and classification of gait patterns in cerebral palsy using a three-dimensional motion analyzer. Journal of Child Neurology. 1998;13(3):101–108. pmid:9535234
  25. 25. O’Malley MJ, Abel MF, Damiano DL, Vaughan CL. Fuzzy clustering of children with cerebral palsy based on temporal-distance gait parameters. Rehabilitation Engineering, IEEE Transactions on,. 1997;5(4):300–309.
  26. 26. Stout JL, Bruce B, Gage JR, Schutte L. Joint kinetic patterns in children with spastic hemiplegia cerebral palsy. Gait & Posture. 1995;3(4):274.
  27. 27. Toro B, Nester CJ, Farren PC. Cluster analysis for the extraction of sagittal gait patterns in children with cerebral palsy. Gait & Posture. 2007;25(2):157–165.
  28. 28. Franklyn-Miller A, Richter C, King E, Gore S, Moran K, Strike S, et al. Athletic groin pain (part 2): a prospective cohort study on the biomechanical evaluation of change of direction identifies three clusters of movement patterns. British Journal of Sports Medicine. 2016; p. bjsports–2016.
  29. 29. Phinyomark A, Osis S, Hettinga BA, Ferber R. Kinematic gait patterns in healthy runners: A hierarchical cluster analysis. Journal of biomechanics. 2015;48(14):3897–3904. pmid:26456422
  30. 30. Richter C, King E, Falvey E, Franklyn-Miller A. Supervised learning techniques and their ability to classify a change of direction task strategy using kinematic and kinetic features. Journal of biomechanics. 2018;66:1–9. pmid:29146284
  31. 31. Kobsar D, Ferber R. Wearable Sensor Data to Track Subject-Specific Movement Patterns Related to Clinical Outcomes Using a Machine Learning Approach. Sensors. 2018;18(9):2828.
  32. 32. Halilaj E, Rajagopal A, Fiterau M, Hicks JL, Hastie TJ, Delp SL. Machine learning in human movement biomechanics: Best practices, common pitfalls, and new opportunities. Journal of biomechanics. 2018;. pmid:30279002
  33. 33. Rajpurkar P, Hannun AY, Haghpanahi M, Bourn C, Ng AY. Cardiologist-level arrhythmia detection with convolutional neural networks. arXiv preprint arXiv:170701836. 2017;.
  34. 34. Figueiredo J, Santos CP, Moreno JC. Automatic recognition of gait patterns in human motor disorders using machine learning: A review. Medical engineering & physics. 2018;53:1–12.
  35. 35. Kamruzzaman J, Begg RK. Support vector machines and other pattern recognition approaches to the diagnosis of cerebral palsy gait. IEEE Transactions on Biomedical Engineering. 2006;53(12):2479–2490. pmid:17153205
  36. 36. Rozumalski A, Schwartz MH. Crouch gait patterns defined using k-means cluster analysis are related to underlying clinical pathology. Gait & posture. 2009;30(2):155–160.
  37. 37. Schwartz MH, Rozumalski A, Novacheck TF. Femoral derotational osteotomy: surgical indications and outcomes in children with cerebral palsy. Gait & posture. 2014;39(2):778–783.
  38. 38. Ardestani MM, Chen Z, Wang L, Lian Q, Liu Y, He J, et al. A neural network approach for determining gait modifications to reduce the contact force in knee joint implant. Medical engineering & physics. 2014;36(10):1253–1265.
  39. 39. Hart HF, Culvenor AG, Collins NJ, Ackland DC, Cowan SM, Machotka Z, et al. Knee kinematics and joint moments during gait following anterior cruciate ligament reconstruction: a systematic review and meta-analysis. Br J Sports Med. 2016;50(10):597–612. pmid:26265562
  40. 40. King E, Richter C, Franklyn-Miller A, Daniels K, Wadey R, Moran R, et al. Whole-body biomechanical differences between limbs exist 9 months after ACL reconstruction across jump/landing tasks. Scandinavian journal of medicine & science in sports. 2018;28(12):2567–2578.
  41. 41. King E, Richter C, Franklyn-Miller A, Daniels K, Wadey R, Jackson M, et al. Biomechanical but not timed performance asymmetries persist between limbs 9 months after ACL reconstruction during planned and unplanned change of direction. Journal of biomechanics. 2018;81:93–103. pmid:30322642
  42. 42. Kristianslund E, Krosshaug T, van den Bogert AJ. Effect of low pass filtering on joint moments from inverse dynamics: implications for injury prevention. Journal of biomechanics. 2012;45(4):666–671. pmid:22227316
  43. 43. Moudy S, Richter C, Strike S. Landmark registering waveform data improves the ability to predict performance measures. Journal of biomechanics. 2018;78:109–117. pmid:30126719
  44. 44. Richter C, King E, Falvey E, Franklyn-Miller A. Supervised learning techniques and their ability to classify a change of direction task strategy using kinematic and kinetic features. Journal of Biomechanics. 2017;. pmid:29146284
  45. 45. Hastie T, Tibshirani R. Elements of Statistical Learning; 2002.
  46. 46. Vapnik VN, Vapnik V. Statistical learning theory. vol. 1. Wiley New York; 1998.
  47. 47. Myer GD, Paterno MV, Ford KR, Quatman CE, Hewett TE. Rehabilitation after anterior cruciate ligament reconstruction: criteria-based progression through the return-to-sport phase. Journal of Orthopaedic & Sports Physical Therapy. 2006;36(6):385–402.
  48. 48. Markolf KL, Burchfield DM, Shapiro MM, Shepard MF, Finerman GA, Slauterbeck JL. Combined knee loading states that generate high anterior cruciate ligament forces. Journal of Orthopaedic Research. 1995;13(6):930–935. pmid:8544031
  49. 49. Fleming BC, Renstrom PA, Beynnon BD, Engstrom B, Peura GD, Badger GJ, et al. The effect of weightbearing and external loading on anterior cruciate ligament strain. Journal of biomechanics. 2001;34(2):163–170. pmid:11165279
  50. 50. Hame SL, Oakes DA, Markolf KL. Injury to the anterior cruciate ligament during alpine skiing: a biomechanical analysis of tibial torque and knee flexion angle. The American journal of sports medicine. 2002;30(4):537–540. pmid:12130408
  51. 51. Bishop C, Read P, Chavda S, Turner A. Asymmetries of the lower limb: The calculation conundrum in strength training and conditioning. Strength & Conditioning Journal. 2016;38(6):27–32.
  52. 52. Webster KE, Nagelli CV, Hewett TE, Feller JA. Factors associated with psychological readiness to return to sport after anterior cruciate ligament reconstruction surgery. The American journal of sports medicine. 2018;46(7):1545–1550. pmid:29718684
  53. 53. Ardern CL, Taylor NF, Feller JA, Webster KE. Fifty-five per cent return to competitive sport following anterior cruciate ligament reconstruction surgery: an updated systematic review and meta-analysis including aspects of physical functioning and contextual factors. Br J Sports Med. 2014;48(21):1543–1552. pmid:25157180
  54. 54. Webster KE, Feller JA. Exploring the high reinjury rate in younger patients undergoing anterior cruciate ligament reconstruction. The American journal of sports medicine. 2016;44(11):2827–2832. pmid:27390346
  55. 55. Dames KD, Smith JD, Heise GD. Averaging Trials Versus Averaging Trial Peaks: Impact on Study Outcomes. Journal of applied biomechanics. 2017;33(3):233–236. pmid:27992244
  56. 56. Ohtani K. Bootstrapping R2 and adjusted R2 in regression analysis. Economic Modelling. 2000;17(4):473–483.
  57. 57. O’Malley E, Richter C, King E, Strike S, Moran K, Franklyn-Miller A, et al. Countermovement Jump and Isokinetic Dynamometry as Measures of Rehabilitation Status After Anterior Cruciate Ligament Reconstruction. Journal of athletic training. 2018;53(7):687–695. pmid:30109947
  58. 58. Welch N, Richter C, Franklyn-Miller A, Moran K. Principal Component Analysis of the Biomechanical Factors Associated With Performance During Cutting. Strength & Conditioning Journal. 2019;XX(XX):XX.
  59. 59. Gore S, Franklyn-Miller A, Richter C, Falvey E, King E, Moran K. Biomechanical complexity: a measure to delineate between athletic groin pain patients and uninjured controls. ISBS Proceedings Archive. 2017;35(1):156.
  60. 60. Gore S, Franklyn-Miller A, Richter C, Falvey E, King E, Moran K. Is stiffness related to athletic groin pain? Scandinavian journal of medicine & science in sports. 2018;28(6):1681–1690.
  61. 61. Baida SR, Gore SJ, Franklyn-Miller AD, Moran KA. Does the amount of lower extremity movement variability differ between injured and uninjured populations? A systematic review. Scandinavian journal of medicine & science in sports. 2018;28(4):1320–1338.