Skip to main content
Advertisement
Browse Subject Areas
?

Click through the PLOS taxonomy to find articles in your field.

For more information about PLOS Subject Areas, click here.

  • Loading metrics

Comparison of sequential data analysis and functional data analysis for locomotor adaptation

  • Torin Quinlivan ,

    Roles Formal analysis, Investigation, Methodology, Writing – original draft

    torin.quinlivan@gmail.com

    Affiliation Department of Mathematics, Knox College, Galesburg, Illinois, United States of America

  • Kacy Kane,

    Roles Formal analysis, Investigation, Methodology

    Affiliation Department of Statistics and Actuarial Science, Northern Illinois University, DeKalb, Illinois, United States of America

  • Christopher M. Hill,

    Roles Data curation, Funding acquisition, Investigation

    Affiliation School of Kinesiology, Louisiana State University, Baton Rouge, Louisiana, United States of America

  • Duchwan Ryu

    Roles Formal analysis, Investigation, Methodology, Writing – original draft

    Affiliation Department of Statistics and Actuarial Science, Northern Illinois University, DeKalb, Illinois, United States of America

Abstract

Learning rates for skills such as walking may depend on circumstances or time, while incentivization with punishments or rewards may affect human skill learning. We consider a state space model for dynamically changed learning rates and figure out the effect of incentivization on the learning rates by utilizing a dynamically weighted particle filter. However, estimations of model parameters, including the learning rate, require a demanding computational burden, especially when the data are collected over a long period. To overcome computational difficulty, we utilize an efficient sequential Monte Carlo method, dynamically weighted particle filter, in the estimations of model parameters. Alternatively, we consider a functional data analysis for the learning rates and the effect of the incentivization. Two approaches have led to reasonable estimations of learning rates. We present the estimated learning rates and the effect of incentivization on the learning rates from two approaches, as well as the comparisons of their results.

1 Introduction

Walking is one of the many skills that we learn through a trial-and-error practice. To generate a new walking pattern, the brain must learn to correct movement errors based on feedback received from the body and environment. This adaptive process results in the learning and retention of a diverse repertoire of locomotor patterns that are maintained throughout our lifespan [1]. The plastic nature of walking adaptation serves as the basis for walking rehabilitation for neurological disease and injury, for example stroke [2,3]. Walking adaptation stems from the improved correction of errors by the central nervous system. There are two types of errors. The sensory prediction errors (SPE) result from discrepancies between expected sensory feedback such as visual, vestibular and proprioception and actual sensory feedback derived from our movement [4]. The reward prediction errors (RPE) is similar, but instead of sensory feedback, it results from differences in the reinforcement expected and the reinforcement received after a movement. Minimizing the discrepancies between one or both of these error types develops an efficient motor plan to perform the new walking pattern [2,3]. Interestingly, the type of feedback received during learning can differently influence learning rates. For instance, provided visual feedback of performance has been widely shown to enhance the rate of visuomotor learning. Others have utilized reinforcement feedback such as reward and punishment [5]. These types of feedback elicit different influences on skill learning. Most researchers have found that punishment feedback increases the rate of visuomotor adaptation and reward enhances motor memory of task [510]. Conversely, other tasks have noted increased learning with reward and visual feedback compared to punishment feedback. Interestingly, in the area of locomotor adaptation, studies have found mixed effects of reinforcement. For instance, one study found that both reward and punishment impaired locomotor adaptation [11], while another found that punishment increased memory and reward impaired adaptation [12]. Thus, the effect of reinforcement on locomotor learning are complex, multifaceted, and in need of a consensus and more robust analysis to understand the optimal conditions for reinforcement feedback usage.

Previous investigations have leveraged standard state space models to investigate retention factors and the learning rates in motor adaptation [1315]. Using the movement kinematics collected from individual participants over time, a state space model can be used to estimate the retention and learning factors over the course of skill learning. Such analysis has provided interesting insights into how reward and punishment alter skill learning and change motor memory[9,1315]. In this model, learning rate (i.e., parameter ) refers to how fast the person adapts their movement to meet the environmental demand. Retention (parameter ) refers to the ability to carry over or remember previous performed movement. Each of these constructs tell us something different about how the brain adapts and recalls our movement. More specifically, parameter reflects the rapid adjustment of cerebellar circuits based on the large amount of sensory prediction errors present early in adaptation. Parameter reflects how the movement pattern is stored in the neural connections between the motor cortex and the cerebellum. In summation, the parameters provide us with a picture of the underlying mechanisms used by the nervous system to facilitate adaptation across a short timescale [9,1315]. Recent work has outlined the challenges associated with state space modeling in the context of skill learning. For instance, estimating retention factors is increasingly difficult with state space models, which may be subject to hidden factors that underlie motor memory [16]. Moreover, processes of learning are sources of stochasticity that are typically not taken into consideration by these models. Other practical factors also play a role. Missing values or unexpected and dynamic changes of movement patterns over time may limit the robust nature of state space modeling. Taken together, there is great need to develop new approaches to modeling robust motor learning behavior. Such models could pave the way for more accurate prediction of rehabilitation outcomes.

In this paper, we consider two approaches to overcome these challenges and to more accurately estimate the retention factors and learning rate of participants during a locomotor adaptation paradigm, where participants learn a new knee flexion pattern with visual feedback of their angular kinematics, and, moreover, examine the effect of reinforcement (i.e. reward and punishment) on these rates. The first approach constructs a state space model for the knee angles with group means and variances. Using a dynamically weighted particle filter [17,18], the parameters involved in the model are efficiently estimated. Based on the sequences of the estimated means and variances and the observed means and variances for each group of participants, we solve the structural equation model to estimate the retention factors and the learning rate. The second approach utilizes a Bayesian functional data analysis. The functional data analysis [19,20] handles a sequence of time-varying measurements as the realizations of a smooth function. We consider a more flexible nonlinear state space model and fit Bayesian cubic regression splines [21,22] to each of the group means of knee angles over time. The derivative of fitted curves can be used for the estimation of the retention rates, learning rates, and their confidence intervals for inference. However, it should be noted that the confidence intervals of the derivative of estimated curves are not readily achievable. By the virtue of the Markov Chain Monte Carlo samples of the fitted curves, we derive the non-constant retention factors and learning rates and their confidence intervals. Furthermore, we investigate the effect of reward-punishment condition on the retention factors and learning rates and compare two approaches.

The rest of the paper is structured as follows. Sect 2 describes how the data was collected. Sect 3 describes the two methods in detail, with explanations of the parameters and full prior and posterior distributions, with Sect 3.1 detailing the state space model and Sect 3.2 detailing the functional data analysis. Sect 4 outlines the results of fitting the data using these methods and the parameter estimates. Finally, Sect 5 gives our conclusions.

2 Data collection methods

All procedures were approved by the Northern Illinois University Institutional Review Board [protocol number HS21-0399]. Thirty-three young healthy adults participated in this study [age range: 19–34 years, mean age ± standard deviation (SD): years, body height: cm, body weight: kg, males: 16, females: 17)]. Participants were classified as right-handed using the Edinburgh Handedness Inventory (EHI) (>50 implies right-handedness, mean handedness score ± SD: ) and were free of major physiological (musculoskeletal, neurological, cardiovascular) and psychological (drug abuse, depression, generalized anxiety) disorders. Participants were recruited from the local population of the University and the surrounding communities using word of mouth, electronic announcements, and posted flyers. Each participant was randomly allocated to one of three feedback groups: Reward (n = 11), Punishment (n = 11), Control (n = 11). All participants were informed of the task goals by being read aloud a script before the start of the experiment. Each participant adapted to a new walking pattern on a treadmill (Woodway, Waukesha, WI) with a different type of visual feedback in two distinct task conditions. Particpants were read aloud task a script describing different conditions, the task goal of changing their knee flexion angle, and the meaning of the visual feedback. Afterwards, participants were asked to recall the task goals to ensure understanding of the procedures. Questions from the participants concerning procedures and goals were addressed by the experimenters. To begin the protocol, participants began walking in a dimly lit room, to minimize distractions, at a calculated speed based on each participant’s step length (two-thirds of the leg length (m)) multiplied by a cadence constant of 1.33 (90 steps/60 seconds) [11,12]. Participants were asked to walk on a treadmill at a speed dictated by their stride length in a controlled environment. The flexion knee angle was measured by using two XSENS IMUs affixed to the participant’s thigh and shank on their right leg and derived from the difference between the two angular positions in reference to a world coordinate system. After taking 250 steps to establish a baseline flexion knee angle (Baseline), each participant was asked to increase their walking flexion knee angle by 30 degrees and take 500 steps (Adaptation). The 30-degree increase was chosen to allow for a wider exploration of the task space. During this period participants were provided one of three different kinds of feedback depending on the group each participant were assigned to (Control, Reward, Punishment).

Visual feedback was displayed on a 60 cm screen, 85 cm in front of the participant, after every two steps with the right leg, completing a stride cycle, as shown in Fig 1A. Previous work has leveraged visual error feedback of lower extremity kinematics to alter locomotor patterns and induce learning [1,23,24]. To this end, we are utilizing the methods employed by these previous studies to compare with reinforcement feedback, to determine the effectiveness each type of feedback in altering locomotor adaptation and memory, which aligns with other studies in this domain [3,11,12]. As in Fig 1B, the Control group was provided with a vertical scale with their current and target line representing the desired maximum knee flexion angle. This target line remained in the same location throughout the experiment. The Reward and Punishment groups used a monetary scoring feedback system based on the difference between current and desired knee flexion angle. Participants were shown a number corresponding to a monetary gain or loss. The magnitude of scoring feedback was dependent on the amount of error in the previous stride cycle and followed these criteria:

thumbnail
Fig 1. A. Visual feedback presented on a screen, concurrently to walking.

Progression of task conditions. Duration was dictated by number steps. B. Example feedback based on group assignment from each task condition.

https://doi.org/10.1371/journal.pone.0329940.g001

  • Reward: +4 points: meets desired angle; +3 points: within ; +2 points: within ; +1 point: within ; 0 points: exceeds or fewer than .
  • Punishment: 0 points: meets desired angle; -1 point: within ; -2 points: within ; -3 points: within ; -4 points: exceeds or fewer than .

All groups started with a total of zero points. Those in the Reward group earned positive points, while those in the Punishment group accrued negative points. The Reward group was instructed that they begin with 0.00 USD and earn money based on their performance. The Punishment group was instructed that they began with 30.00 USD and lost money based on their performance. In addition, participants in the Control group were given the instructions of the reward or punishment groups, so as to control the effects of the script. To ensure equity in compensation, all participants were compensated, despite group assignments, with the full 30.00 USD at the end of the study, regardless of performance. Feedback presentation followed the same timing latency, ten milliseconds after completing a stride cycle. The task goal was to minimize error by matching the current knee flexion angle with the desired knee flexion angle. In this paper we chose to focus our attention on the first 500 steps; the 250 in the Baseline round and the first 250 of the Adaptation round, and tried to estimate the learning rate and retention factor of these time points. Further study may look at the additional time points.

To ensure homogeneity in the data collection, the groups were allocated in such a way as to have a similar number of males and females in each feedback group, and walked with similar treadmill speeds. Further, in a preliminary data analysis for the homogeneity of participants in age, height and weight, one-way analyses of variance revealed there was no significant differences among the three feedback groups. For EHI, the F-statistic with 2 and 3 degrees of freedom was 2.857. Subsequently, F = 0.116 and p-value = 0.891 for Age, F = 0.082 and p-value = 0.921 for Height, F = 0.320 and p-value = 0.729 for Body Weight, and F = 0.144 and p-value = 0.866 for Step Length, respectively.

The data we collected can be found in the supporting information S4 File, the raw flexion angles of the participants walking; and S5 File, the flexion angles adjusted as the difference of from the participants’ baseline angles.

3 Statistical methodologies

We consider the standard state space model for the knee flexion angle during walking with the retention factor and the learning rate as in [13]. To examine the effect of feedback, we separated the participants into groups by the different reward-punishment conditions. Let yijt denote the knee angle in walking collected from the participant in the group at time , . At each time, we establish a response model to characterize a group effect as the mean of knee angles of participants in the group with random noise. In terms of the progression of the knee angles of each participant over time, the angle of the next step is affected by a combination of the angle of the current step and the group effect shared at the current time. In other words, we may construct a state model to project the next knee angle based on the current knee angle for the retention portion and the group effect or the common behavior for the learning portion of the group. Let denote the retention factor and denote the learning rate from the participants in group j at time . The standard state space model can then be described by the response model and the state model such that, for and ,

(1)

where mjt is the mean knee angle at time across participants in group j involving perturbation .

In the model (1), we may utilize two approaches to estimate the mean angles mjt. Using sequential analysis, the mean knee angles are time-varying parameters and a dynamic model for parameters can be estimated for each time. On the other hand, using a functional data analysis, the mean knee angles are considered as functional values at each time. Once the mean functions are estimated, their functional values evaluated at each time can be used. Based on the estimated mean angles, we may derive the time-varying retention factors and learning rates. The two approaches are described in the following subsections.

3.1 Sequential analysis with state space model

For the knee angle yijt from the participant i in group j at time , , and , we assume the perturbation to be Normally distributed with mean zero and variance for the response model in (1). Since the response is measured sequentially along with time , the perturbations at two adjacent time points from the same participant, and for , need not be statistically independent, while the perturbations from two participants at the same time, and for and , can be independent. To address the dependency of and , we utilize a transition model for the variance of random perturbations for each group such that

(2)

where is an initial variance, aj is a transition coefficient and ujt is a Normal disturbance term with mean zero and fixed variance . Thus, we model the mean and variance of knee angles from the participants in each group over all time points with the state space model (1) and an additional autoregressive transition model (2) for state parameters mjt and respectively. In this paper, we set aj = 1 and and initialize the variance as .

Under the Bayesian framework, we assign a conjugate inverse Gamma prior distribution for the state parameter at the initial time point , where we set Asj = 2 and Bsj = 1 for this paper, and use the transition model specified in (2) after the initial time point instead of assigning a prior. For the other parameters, we allocate conventional priors such that , for and . In this paper, we set the hyperparameters as Amjt = 0 and Bmjt = 10 to pursue vague priors.

We collect the Markov Chain Monte Carlo (MCMC) samples of parameters in the model for the estimation by using a Gibbs sampler [25] and dynamically weighted particle filter methods (DWPF) with dynamically weighted importance sampling [26]. Let denote a vector of responses for each participant at time t, denote a vector of model parameters and denote a vector of state parameters that are varying by time. Further, let denote a sequence of responses from time to , and denote the sequence of state parameters up to time . Then, we use the full conditional distributions to generate MCMC samples for and use a dynamically weighted particle filter to estimate . The full conditional distributions are given by, for ,

where , , and . A full description of the dynamically weighted particle filter algorithm can be found in Supplementary File 1 as in [27].

Based on the population of particles, we may estimate the sequence of mean knee angles by taking the weighted average of particles at each time point. Because we have such a large weight-controlled population, on the order of thousands of particles, the weighted mean and the simple mean of particles are quite similar to each other. Thus for simplicity, in this paper, we use the simple mean of particles as the Bayesian estimate of parameters. Let denote the sequence of estimated mean knee angles. Further, let denote the observed mean knee angles for group at time , . Then, using the R package lavaan [28], we can estimate the retention factor and the learning rate by solving the following structural equation model such that

(3)

All code used to run the dynamically weighted particle filter can be found in the supporting information S2 File.

3.2 Functional data analysis

As an alternative approach we consider a functional data analysis (FDA) by taking the sequentially measured knee angles yijt from the participant in group over time , as functional responses. In the state space model (1) the response model is expanded with a smooth function fj such that

(4)

where the mean knee angle mjt is treated as a random effect, is the random discrepancy which explains the variation of group means, is the random error of individual participants and and are mutually independent. The discrepancy and random error are assumed to follow the Normal distribution with zero means and constant variances , , and , respectively.

In this paper, specifically, we consider the cubic P-splines [29] as a smoothing function fj. The K knot points are selected from time points , . The basis function has the value of if and 0 otherwise, for knot . The cubic P-splines can be expressed by

where are regression coefficients. Specifically, we take K = 19 equally spaced knot points within the range of such that .

As a Bayesian approach we assign the prior distributions for unknown parameters in the model (4). Let denote a vector of regression coefficients for the mean of group j and assign independent conjugate Normal prior distributions

where is a smoothing parameter scaled by the discrepancy and is a generalized inverse matrix of diagonal matrix , which is a diagonal matrix with zeros for the first 4 diagonal elements and ones for the rest of diagonal elements. Regarding group-specific smoothing parameters and discrepancies, we assign independent and identical conjugate Gamma prior distributions and inverse Gamma prior distributions, and , , with hyperparameters At, Bt, Ad and Bd, respectively. Likewise, for the variances of random error, we assign conjugate inverse Gamma distribution such that and with hyperparameters Ae and Be.

We estimate the model using Markov Chain Monte Carlo (MCMC) samples of parameters pulled from the joint posterior distribution. Utilizing a Gibbs sampler, we can collect the MCMC samples based on the full conditional distribution. Let denote the vector of responses, denote the vector of mean angles and denote the design matrix with the row vectors for . Then, the full conditional distributions of parameters can be derived such that

for . In this paper, we collect 1000 MCMC samples of parameters after 1000 burning iterations. From each set of MCMC samples of , denoted , we can derive the MCMC samples of the vector of functional responses such that , . By taking the differences of functional responses we can also approximate the derivatives of functional responses at time such that , . The Bayesian estimates of mean function and its derivative at time t can be obtained by taking the averages of MCMC samples, that is and . The credible intervals of the mean function and its derivative can be estimated by taking the quantile values of and , respectively. It should be noted that the credible intervals of derivatives of functions are generally not easy to calculate. However, in this paper, we can readily use the quantiles of MCMC samples in the calculation of credible intervals under a Bayesian framework.

Regarding the estimation of the retention factors and the learning rates , the FDA provides time-varying functions of the retention factors and the learning rates. From the state space model (1), the group mean function mjt is estimated by in FDA. Regarding the state model as a regression model for the knee angles at time on the knee angles at time t with disturbance term , the retention faction is a slope that represents the change rate of knee angles from current time to the next. Hence, for the estimated mean functions, the retention factor or the change of responses by the time is estimated by in FDA. Hence, denoting and as the function of retention factors and the function of learning rates respectively, we apply the results from FDA to the transition model in (1) and estimate the functions with the observed mean knee angles such that

(5)

The credible region of is equivalent to that of and the credible region of can be calculated by taking the quantile values of , . It should be noted that in this instance actually shows the change of the retention factor over time, and it can be interpreted as at each time to compare with the results from the state space model in Sect 3.1.

All code used to run the functional data analysis can be found in the supporting information S3 File.

4 Results

4.1 Sequential analysis with state space model

The dynamically weighted particle filter (DWPF) estimated the group means and variances in the response model at (1) with 95% credible intervals as shown in Fig 2. The estimated mean angles have shown consistently below the observed mean angles. However, it is not surprising because the observed knee angles from the same participant are not independent and their sample means are well known to be biased. The estimated means have shown a jump when the time is greater than 250 in all three groups. The estimated variances have stayed at a low level with narrow credible interval when the time is less than 250. After that they increased with wider credible intervals and have become stabilized with reduced width of credible intervals in all groups. Fig 3 shows the three groups laid over each other to show where they deviate from one another.

thumbnail
Fig 2. Estimated Means and Variances.

In (a), (b) and (c) light blue dots are observed angles, blue dotted lines are 95% credible regions, blue solid lines are the estimated mean functions and red lines are the smoothed mean function, respectively. In (d), (e) and (f) light blue dots are observed mean angles, blue dotted lines are 95% credible regions, blue solid lines are the estimated variance functions.

https://doi.org/10.1371/journal.pone.0329940.g002

thumbnail
Fig 3. Estimated Means and Variances.

In (a), the estimated means, the smoothed mean functions and the 95% credible regions are shown for each group. In (b), the estimated variance functions and the 95% credible regions are shown for each group. Blue lines denote the control group, green lines the punishment group and red lines the reward group.

https://doi.org/10.1371/journal.pone.0329940.g003

Using the estimated mean functions in the Sect 3.1, for and , and the observed mean angles we can estimate the retention factor and the learning rate for each group. Table 1 presents the estimated ’s and ’s and the results of their chi-squared difference tests to examine if three groups produce different ’s and ’s. Structural equation modeling was used to estimate the parameters. It has been shown that chi-squared differences between nested SEM models are equivalent to the standard analysis of variance (ANOVA) tests in [30]. The retention factors ’s have been estimated to be close to 1 in all groups with a statistically significant difference; whereas the learning rates ’s have been estimated to around 4% in the Control group, 5% in the Punishment group and 12% in the Reward group, a statistically significant difference.

thumbnail
Table 1. Estimates and Inferences for Retention Factor and Learning Rate for Model using Particle Filter Means.

https://doi.org/10.1371/journal.pone.0329940.t001

Using the individual sequence of the observed knee angles instead of the group mean angles, we have also estimated and for each individual participant. The distributions of the individual estimated ’s and ’s within group are shown in the boxplots in Fig 4. Interestingly, the distribution of estimated ’ in the control group have shown significantly lower levels than those in the punishment group and the reward group; whereas the distributions of the individual estimated ’s of three group show similar levels. This seems to contradict the results in Table 1, which shows the results for the group learning and retention rates. However, that can be explained that some of participants especially in the control group have quite unusual patterns of knee angles compared to the other participants.

thumbnail
Fig 4. Estimates for the Retention Factor and Learning Rate for the Three Treatments.

Boxplots are based on the estimated values for each case.

https://doi.org/10.1371/journal.pone.0329940.g004

To evaluate the performance of the particle filter, we considered a simpler model using a structural equation model without the particle filter mean at time t in the term containing , and replacing it with the sample mean at time t. The results of the models are given in Table 2. This simpler model finds very similar results to the model containing the particle filter mean, slightly over-estimating and slightly under-estimating compared with the model in Table 1. Both models find that the coefficients and are significantly different between the treatments. This suggests that the particle filter does efficiently and effectively model the mean of the data.

thumbnail
Table 2. Estimates and Inferences for Retention Factor and Learning Rate for Model without Particle Filter Means.

https://doi.org/10.1371/journal.pone.0329940.t002

4.2 Functional data analysis

The flexion knee angles data used in the previous subsection have been fitted by the Bayesian cubic P-splines. The estimated group mean functions, in the Sect 3.2, for three groups are presented in Fig 5. It should be noted that is a smooth version of in the Sect 3.1. In all three groups, the mean angles have been increased in the average with a jump at time 250, while the control has a stabilized trend, the punishment has an increasing trend and the reward has a slightly decreasing trend after the jump.

thumbnail
Fig 5. Estimated Mean Functions.

The dots are observed means of knee angles, the lines are fitted curves and the shaded areas are 95% credible intervals.

https://doi.org/10.1371/journal.pone.0329940.g005

Using the estimated mean functions and the related derivatives in Sect 3.2, the time-varying retention factors and the learning rates have been obtained as shown in Fig 6. While the three group take similar shapes in the function and , the jump of knee angles at time 250 have brought a peak in the retention factors . The learning rates take a drop just before time 250 and a jump afterward in all groups. Regarding the 95% credible intervals the estimated retention factor functions have wider intervals than the learning rate functions have. While take similar width except for the both ends, take an increasing width as the time increased.

thumbnail
Fig 6. Estimated Retention Factor and Learning Rate .

The lines are estimated and , and the shade areas are 95% credible intervals.

https://doi.org/10.1371/journal.pone.0329940.g006

Based on the 95% credible intervals the difference of and can be examined. As shown in Fig 7, comparing three groups from Fig 6 against each other, the credible intervals of overlap in all time, indicating no significant difference between groups. The credible intervals of shows separated intervals in the beginning and after time 250, which indicates significantly different effects of the groups.

thumbnail
Fig 7. Estimates for the Retention Factor and Learning Rate.

Blue lines and light blue area are estimated functions and 95% credible regions for the control group, green lines and light green area are those for the punishment group and red lines and light red areas are those for the reward group, respectively.

https://doi.org/10.1371/journal.pone.0329940.g007

5 Conclusion

We have utilized two approaches to examine the retention factor and the learning rate in the changes of knee angles during striding under three different feedback conditions, the state space model and the functional data analysis. Both approaches have led to compatible results, while the method based on the functional data analysis has allowed more flexible inference.

In the state space model approach, the retention factor that shows the relationship between two consecutive observations has been turned out about without significant difference by the types of feedback. That is, the knee angles in the following steps have been about 97% of the knee angles of preceding step in the average no matter what type of feedback has been given to the participant. Meanwhile, in the functional data analysis approach, the retention factor has been varying while it mostly remains zero, , and has a peak in the middle at time 250. It again has shown that the angles of following steps have been similar to the angles in the preceding steps except for the steps around time 250, when the phase shifts from Baseline to Adaptation, without any significantly different effect of the feedback.

Regarding the learning rate that indicates the contribution of projected knee angle in the preceding step, the state space model approach has shown less than 5%. However, different rewards and punishments have resulted in significantly different contributions. Similarly, the functional data analysis approach has also indicated that the projected angles of the preceding step has little contribution to the angle of the next step and the effect of feedback has been significant. This implies that the use of reinforcement methods, reward-based or punishment-based, does seem to have an impact on participants learning to increase the flexion angles of their knee while walking. In addition, the contribution has been substantially increased after the peak of retention rate .

Past work in the field concerning learning and retention rates [9,13,15] have dealt with short timescales and have relied on simpler methods of analysis, namely ANOVA. These methods are limited in properly considering large datasets collected over larger timescales and can require large computational loads.

The functional data analysis method allows for elegant analysis of the changing values of the retention factor and the learning rate over time. Also, by overlaying the estimates and the credible intervals makes inference simple and easily interpretable, by looking for the areas where the credible intervals do not overlap. However, one drawback of FDA is that if you want to update after adding new data, everything needs to be rerun on the full set of the data, both the original data and the new additions. The dynamically weighted particle filter approach lets us consider only the new data point by carrying the particles forward based on the results of the prior analysis, meaning we can quickly and efficiently incorporate new data.

The models used in this paper provide more efficient and flexible methods for estimating the parameters and the confidence intervals. The confidence intervals in Figs 6 and 7 are examples that can be difficult to estimate cleanly, but these models offer easy solutions for finding credible intervals. Applying these methods to further data sets will provide an opportunity to further develop the methods.

Supporting information

S1 File. Dynamically Weighted Particle Filter Model Algorithm.pdf

A full description of the dynamically weighted particle filter algorithm.

https://doi.org/10.1371/journal.pone.0329940.s001

(PDF)

S2 File. Full DWPF Code.R

All code used to run the dynamically weighted particle filter and use those particles to estimate the parameters of the state space model.

https://doi.org/10.1371/journal.pone.0329940.s002

(R)

S3 File. Full FDA Code.R

All code used to run the functional data analysis and estimate the means, the learning rate and retention factor and credible intervals for each.

https://doi.org/10.1371/journal.pone.0329940.s003

(R)

S4 File. Adaptation Peak Data.csv

Dataset of flexion knee angles, of which we considered the first 500 time points.

https://doi.org/10.1371/journal.pone.0329940.s004

(CSV)

S5 File. Adaptation Peak Data Baseline.csv

Dataset of baseline flexion knee angles for each participant.

https://doi.org/10.1371/journal.pone.0329940.s005

(CSV)

References

  1. 1. Roemmich RT, Long AW, Bastian AJ. Seeing the errors you feel enhances locomotor performance but not learning. Curr Biol. 2016;26(20):2707–16. pmid:27666970
  2. 2. Roemmich RT, Bastian AJ. Two ways to save a newly learned motor pattern. J Neurophysiol. 2015;113(10):3519–30. pmid:25855699
  3. 3. Hasson CJ, Manczurowsky J, Yen S-C. A reinforcement learning approach to gait training improves retention. Front Hum Neurosci. 2015;9:459. pmid:26379524
  4. 4. Torrecillos F, Albouy P, Brochier T, Malfait N. Does the processing of sensory and reward-prediction errors involve common neural resources? Evidence from a frontocentral negative potential modulated by movement execution errors. J Neurosci. 2014;34(14):4845–56. pmid:24695704
  5. 5. Krakauer JW, Hadjiosif AM, Xu J, Wong AL, Haith AM. Motor learning. Compr Physiol. 2019;9(2):613–63. pmid:30873583
  6. 6. Hill CM, Stringer M, Waddell DE, Del Arco A. Punishment feedback impairs memory and changes cortical feedback-related potentials during motor learning. Front Hum Neurosci. 2020;14:294. pmid:32848669
  7. 7. Hill CM, Waddell DE, Del Arco A. Cortical preparatory activity during motor learning reflects visuomotor retention deficits after punishment feedback. Exp Brain Res. 2021;239(11):3243–54. pmid:34453554
  8. 8. Quattrocchi G, Greenwood R, Rothwell JC, Galea JM, Bestmann S. Reward and punishment enhance motor adaptation in stroke. J Neurol Neurosurg Psychiatry. 2017;88(9):730–6. pmid:28377451
  9. 9. Song Y, Smiley-Oyen AL. Probability differently modulating the effects of reward and punishment on visuomotor adaptation. Exp Brain Res. 2017;235(12):3605–18. pmid:28887626
  10. 10. Song Y, Lu S, Smiley-Oyen AL. Differential motor learning via reward and punishment. Q J Exp Psychol (Hove). 2020;73(2):249–59. pmid:31382855
  11. 11. Hill CM, Sebastião E, Barzi L, Wilson M, Wood T. Reinforcement feedback impairs locomotor adaptation and retention. Front Behav Neurosci. 2024;18:1388495. pmid:38720784
  12. 12. Sato S, Cui A, Choi JT. Visuomotor errors drive step length and step time adaptation during “virtual” split-belt walking: the effects of reinforcement feedback. Exp Brain Res. 2022;240(2):511–23. pmid:34816293
  13. 13. McDougle SD, Bond KM, Taylor JA. Explicit and implicit processes constitute the fast and slow processes of sensorimotor learning. J Neurosci. 2015;35(26):9568–79. pmid:26134640
  14. 14. Smith MA, Ghazizadeh A, Shadmehr R. Interacting adaptive processes with different timescales underlie short-term motor learning. PLoS Biol. 2006;4(6):e179. pmid:16700627
  15. 15. Galea JM, Mallia E, Rothwell J, Diedrichsen J. The dissociable effects of punishment and reward on motor learning. Nat Neurosci. 2015;18(4):597–602. pmid:25706473
  16. 16. Albert ST, Shadmehr R. Estimating properties of the fast and slow adaptive processes during sensorimotor adaptation. J Neurophysiol. 2018;119(4):1367–93. pmid:29187548
  17. 17. Liang F. Dynamically weighted importance sampling in Monte Carlo computation. Journal of the American Statistical Association. 2002;97(459):807–21.
  18. 18. Ryu D, Liang F, Mallick BK. Sea surface temperature modeling using radial basis function networks with a dynamically weighted particle filter. Journal of the American Statistical Association. 2013;108(501):111–23.
  19. 19. Ramsay JO, Silverman BW. Functional data analysis. International encyclopedia of the social & behavioral sciences. Elsevier; 2001. p. 5822–8. https://doi.org/10.1016/b0-08-043076-7/00434-4
  20. 20. Ramsay J, Silverman BW. Functional data analysis. New York: Springer; 2005.
  21. 21. Ryu D, Xu H, George V, Su S, Wang X, Shi H, et al. Differential methylation tests of regulatory regions. Stat Appl Genet Mol Biol. 2016;15(3):237–51. pmid:26982617
  22. 22. Chatterjee S, Chowdhury S, Ryu D, Basu S. Bayesian functional data analysis over dependent regions and its application for identification of differentially methylated regions. Biometrics. 2023;79(4):3294–306. pmid:37479677
  23. 23. Cherry-Allen KM, Statton MA, Celnik PA, Bastian AJ. A dual-learning paradigm simultaneously improves multiple features of gait post-Stroke. Neurorehabil Neural Repair. 2018;32(9):810–20. pmid:30086670
  24. 24. Statton MA, Toliver A, Bastian AJ. A dual-learning paradigm can simultaneously train multiple characteristics of walking. J Neurophysiol. 2016;115(5):2692–700. pmid:26961100
  25. 25. Geman S, Geman D. Stochastic relaxation, gibbs distributions, and the bayesian restoration of images. IEEE Trans Pattern Anal Mach Intell. 1984;6(6):721–41. pmid:22499653
  26. 26. Liang F. Dynamically weighted importance sampling in Monte Carlo computation. Journal of the American Statistical Association. 2002;97(459):807–21.
  27. 27. Ryu D, Liang F, Mallick BK. Sea surface temperature modeling using radial basis function networks with a dynamically weighted particle filter. Journal of the American Statistical Association. 2013;108(501):111–23.
  28. 28. Rosseel Y. lavaan: AnRPackage for structural equation modeling. J Stat Soft. 2012;48(2).
  29. 29. Berry SM, Carroll RJ, Ruppert D. Bayesian smoothing and regression splines for measurement error problems. Journal of the American Statistical Association. 2002;97(457):160–9.
  30. 30. Lee Helm J, Langenberg B, Grossman E, Poulin J, Mayer A. Using structural equation modeling in place of between-subjects analysis of variance. Structural Equation Modeling: A Multidisciplinary Journal. 2022;30(1):123–31.