Skip to main content
Advertisement
Browse Subject Areas
?

Click through the PLOS taxonomy to find articles in your field.

For more information about PLOS Subject Areas, click here.

  • Loading metrics

Advancing training effectiveness prediction in mass sport through longitudinal data: A mathematical model approach based on the Fitness-Fatigue Model

  • Wenxing Wang ,

    Contributed equally to this work with: Wenxing Wang, Yuanhui Zhao

    Roles Conceptualization, Data curation, Funding acquisition, Investigation, Methodology, Writing – original draft, Writing – review & editing

    Affiliation School of Sport Science, Beijing Sport University, Beijing, China

  • Yuanhui Zhao ,

    Contributed equally to this work with: Wenxing Wang, Yuanhui Zhao

    Roles Investigation, Methodology, Writing – review & editing

    Affiliation School of Sport Science, Beijing Sport University, Beijing, China

  • Xiao Hou,

    Roles Methodology, Writing – review & editing

    Affiliation School of Sport Science, Beijing Sport University, Beijing, China

  • Wenlang Yu,

    Roles Methodology, Writing – review & editing

    Affiliation School of Sport Science, Beijing Sport University, Beijing, China

  • Hong Ren

    Roles Conceptualization, Funding acquisition, Methodology, Supervision, Writing – review & editing

    renhong@bsu.edu.cn

    Affiliation School of Sport Science, Beijing Sport University, Beijing, China

Abstract

Despite the critical need for scientific training load assessment in mass sports, the Fitness-Fatigue Model (FFM) requires further mathematical optimization and practical output indicators. The aim of this study was to optimize the mathematical relationship between “adaptation” and “fatigue” in the FFM, identify generalizable model output indicators, and evaluate its performance in predicting training effectiveness in mass sport. To account for the nonlinear and time-varying characteristics of training effectiveness, this study proposed new mathematical assumptions and optimized parameters against individual longitudinal data. The external load (speed and wattage) and internal load (wearable-compatible heart rate variability [HRV] and heart rate recovery [HRR] related indicators) of each training day were collected for 28–42 days per person (420 paired data from 13 subjects during 12 weeks of medium-intensity continuous cycling). The longitudinal data were used to perform parameter estimation and model evaluation for each individual separately. When the optimal model output indicator was selected, the R2 values of the optimized model ranged from 0.61–0.95, with fitting root mean square error (RMSE) at 0.07–0.37, and mean absolute percentage error (MAPE) in predictive ability assessment at 3.99%−31.99%. However, a few individuals had larger fitting errors (minimum R2 of 0.32, maximum RMSE of 0.90) and predictive errors (maximum MAPE of 86.57%) when the output indicator was inappropriate. The original model generally has lower R2 and higher RMSE and MAPE. This shows the optimization of functional relationships and the application of individual longitudinal data have resulted in better performance of the model, but optimal indicator selection varies by individual. Furthermore, HRV and HRR related indicators are generalizable model output indicators that can be used to predict training effectiveness in mass sports through wearable devices and machine learning technology. However, the study has limitations including the homogeneous sample and single training type. Future research should validate the model across different sports and populations, investigating the factors affecting model fitting and prediction.

1 Introduction

Training load and its effectiveness evaluation play a crucial role in competitive and recreational sports. A reasonable training load is essential in athletic training [15], but studies have shown that coaches often struggle to develop optimal training plans without precise data feedback, leading to inconsistencies between planned and actual training loads [6]. In mass training, a well-designed training plan can enhance physical fitness and prevent chronic diseases [79] while improper training load arrangements can result in injuries [10]. Therefore, establishing a scientific and predictive method for assessing training effectiveness has high practical significance for the improvement of sports performance and health benefits.

The Fitness-Fatigue Model (FFM), proposed by Banister and his team, simplifies the training process and quantifies the training load and its effectiveness [11]. However, the mathematical relationship in the model needs optimization. Physiologically, “adaptation” represents the positive functional and structural changes induced by training stimuli, such as an increase in muscle protein synthesis [12], mitochondrial-associated indices [13], and neuromuscular transmission efficiency [14]. Conversely, “fatigue” represents a temporary decline in bodily function that occurs after exercise, which may be related to factors such as muscle glycogen depletion [15], accumulation of metabolites [16,17], and nervous system fatigue [18]. FFM posits that external loads cause adaptive and fatigue responses, changing over time [19,20]. These factors jointly determine the overall internal response. By mathematically relating “adaptation” and “fatigue”, researchers develop training load prediction models [21]. These models have proven effective in predicting training effects and influenced competitive sports [2228], such as their application in pre-competition tapering strategy. Researchers can compare the potential effects of different tapering plans based on FFM, reducing reliance on coach experience [20,26,29,30]. However, traditional FFM uses first-order linear equations to describe the relationship between training response and external loads [11,21]. This doesn’t reflect the nonlinear exercise training response and fading of adaptation and fatigue over time, inconsistent with reality. Due to the decisive role of the mathematical relationship between “adaptation” and “fatigue” in the FFM fitting effect [25], it is necessary to optimize the mathematical relationship between models to increase its interpretability and model fitting effectiveness.

Furthermore, while FFM is predominantly used in competitive sports, some studies have shown good model fitting in the general population [24]. This suggests that FFM has potential in recreational sports, but there is insufficient research evidence to support its application in the general population. With the rapid development of wearable devices in recent years, ordinary individuals can now access a wealth of training data (such as exercise heart rate, average running speed, etc.) through sports bands, sports watches, and other devices. However, there is a lack of means to process and utilize this data, making it difficult for ordinary individuals to receive suitable training plan adjustments. As people increasingly focus on daily fitness activities, those without professional coaching guidance may be at risk of sports injuries and a decline in their quality of life. Therefore, the use of data-based models to optimize the effects of exercise interventions has gained attention [3133]. The mathematical structure of FFM allows its parameters to be estimated by machine learning algorithms [21]. With sufficient data accumulation, the model can provide personalized guidance based on individual characteristics, thereby enhancing sports performance and preventing sports injuries. However, research on the application of FFM in the general population is limited, and the model output indicators selected in existing studies are also not easily applicable to mass sports (using invasive indicators to evaluate training effectiveness in the daily exercise of ordinary individuals is unrealistic). Additionally, considering the varying physiological responses of individuals to the same external load level, the focus of training load assessment should be on the “personalized” evaluation of training effects and the subsequent adjustment of external loads [34,35]. FFM can obtain personalized parameters based on individual longitudinal data and machine learning techniques for personalized evaluation. But most studies have only established universal model parameters and have not explored the “personalized” advantages of this model.

Therefore, the objective of this study is to optimize the mathematical relationship between “adaptation” and “fatigue” of FFM, identify generalizable model output indicators, and evaluate the model’s performance in predicting training effectiveness in mass sport. To validate the model, we chose moderate-intensity continuous training, a common training mode in recreational sports. Moreover, training load comprises both external and internal components [36]. External load refers to the predetermined physical stress imposed on an individual during training, such as exercise intensity and volume. Internal load, on the other hand, reflects the physiological responses of the individual to this external stress [37]. Accordingly, for model input and output indicators, we utilized the power bicycle to collect external load and selected heart rate variability (HRV) and heart rate recovery (HRR) related indicators that reflect autonomic nervous activity as internal load indicators. Paired data collected on each training day were used for parameter estimation and mathematical modeling for each individual.

We hypothesize that (1) the mathematical modeling in this study will optimize the relationship between adaptation and fatigue, improving model interpretability through physiologically meaningful parameters; (2) HRV and HRR will serve as applicable and generalizable model output indicators for mass sports; (3) The model will demonstrate superior performance to the original model fitting with the higher fitting effect and lower prediction error through longitudinal data. It will also be capable of offering personalized evaluation and prediction of training outcomes, thereby offering effective guidance for improving sports performance and physical fitness in the general population.

2 Methods

2.1 Optimization of mathematical models

Based on the nonlinear and time-varying characteristics of training effectiveness, this study proposed new mathematical assumptions and formulated them into an optimized mathematical model. The specific modeling assumption are as follows:

  1. (1) During training, fatigue (F) and adaptation (A) changes are primarily driven by external loads, and there is no occurrence of overtraining. This study retains the core concept of the FFM originally proposed by Banister et al., which describes performance as a function of adaptation and fatigue resulting from external loads [11,20,19]. However, to avoid the complexity associated with overtraining responses, this assumption is added to ensure the model focuses on the typical training response.
  2. (2) External loads will have a beneficial effect on adaptation and fatigue, while adaptation and fatigue will fade over time, and the rate of decline is related to the individual’s adaptation and fatigue level. To address the limitations of the traditional FFM, this study incorporates nonlinear dynamics to optimize the mathematical relationship between adaptation and fatigue, differing from the first-order linear equations used in the original model [11,21]. This approach better reflects the real-world training response, accounting for the nonlinear nature of exercise training and the time-dependent fading of adaptation and fatigue.
  3. (3) The change in performance level is reflected by the difference between adaptation and fatigue, a principle retained from the original FFM [11,20,19]. Additionally, this study introduces specificity coefficients to better capture individual differences in adaptation and fatigue dynamics, enabling the model to provide more personalized guidance based on individual characteristics.

Based on these modeling assumptions, the process of establishing an optimized mathematical model is as follows.

After training, the regression coefficients of A and F are τa and τf, respectively, namely:

(1)(2)

The general solutions of equation (1) and (2) are the regression function of A and F:

(3)(4)

To further reflect individual differences, increasing the specificity coefficients a and f:

(5)(6)

The integration constants C1 and C2 were determined by estimating them alongside the model coefficients, based on the initial conditions (see Section 2.3.1 for estimation details). The new training load (Wn) will cause gains in A and F (Ka, Kf is the gain coefficient), so on the nth day (n ≥ 1):

(7)(8)

The performance level Pn is:

(9)(10)

Pn represents the performance level on the nth day. Wn represents the external load level on the nth day. t represents time within the training period. a, τa, f, and τf are parameters related to adaptation and fatigue regression. Ka and Kf are gain coefficients for adaptation and fatigue, respectively. C1 and C2 are constant parameters in the general solution of the function. The physiological interpretation of each parameter in Equations (10) is provided in S1 Table.

2.2 Data collection methods for model databases

2.2.1 Subjects.

The subjects of this study are full-time university students in Beijing, China. The inclusion criteria are: ① Subjects have sufficient time; ② No other training arrangements during the experiment; ③ Volunteer to participate in the experiment and sign the informed consent form. Exclusion criteria are: ① Subjects have exercise contraindications; ② Have a history of physical or mental illness; ③ Simultaneously conducting high-dose training other than this experiment. Subjects’ characteristics such as age, gender, BMI, and pre-experiment physical activity level (using the International Physical Activity Questionnaire long roll, a reliable and valid tool for physical activity assessment [3840]) were collected before formal training.

The studies involving human participants were reviewed and approved by the Sport Science Experiment Ethics Committee of Beijing Sport University (No. 2022178H). The recruitment period is from September 1, 2022 to December 31, 2022. All methods were performed in accordance with relevant guidelines and regulations, and the research has obtained all subjects’ written informed consent. This study did not include minors.

2.2.2 Experimental design.

This study implemented 12 weeks of medium-intensity continuous cycling, with 3 sessions per week and 30 minutes per session. Prior to formal training, each participant underwent an incremental load cycling test to determine their individual HRmax. Based on this, the target heart rate range for training was set at 60–85% of each participant’s HRmax [41]. Before and after each training session, 5–10 minutes of warm-up or stretching activities were conducted, and each participant’s training process was strictly monitored. Data were collected for each training session.

2.2.3 Data collection methods.

In this study, the model input indicator is the external load collected during the each training session using an ergoselect100 power bicycle, a classic ergometer for exercise tests [42]. The model output indicator is the ratio of external load to internal load on the same day [43], which reflects the performance of the subject. To evaluate the internal load, HRV [35,4448] and HRR related indicators [4954] are selected for their ability to reflect autonomic nervous function, wearable compatibility, and potential for generalization in mass sports.

Data acquisition followed a specific timeline, as illustrated in Fig 1, divided into pre-training, training, and post-training phases. The data collection and indicators calculation methods are as follows.

  1. (1) Pre-training

POLAR H10 heart rate band, a reliable and valid device for heart rate monitoring [55,56], was used to obtain the root mean square of the successive differences (RMSSD) of the difference between adjacent R-R intervals. The RMSSD was measured for 5 minutes while participants remained seated and inactive. This pre-training RMSSD data (Pre5) served as a baseline for HRV analysis. After the RMSSD test, participants were required to warm up before training.

  1. (2) Training

During training sessions conducted on an ergoselect100 power bicycle [42], the bicycle speed and resistance level were continuously recorded (sampled at a frequency of 1 Hz). The external load was calculated by integrating the product of the speed and resistance level over time. Participants were required to reach the maximum value of the specified heart rate interval (85% HRmax) in the last minute of training to collect HRR indicators under submaximal intensity, denoted as HRpeak in Fig 1.

  1. (3) Post-training

Immediately after training, participants underwent a 5 minute recovery period while seated. Heart rate data was collected during this time to calculate post-training RMSSD (Post5) and HRR indicators.

The HRR indicators included the difference in HRR after 1 minute of training (ΔHRR1) and HRR percentage (HRr%), calculated using the formulas shown in equations (11) and (12):

(11)(12)

HRpeak: Peak heart rate in the last minute of training; HRR1: Average heart rate 1 minute after training; HRstage1: Average heart rate 15 seconds after training.

Thirty minutes post-training, the RMSSD of participants was measured again while seated and resting for 5 minutes (Post30). This Post30 RMSSD data provided insights into longer-term autonomic nervous system recovery. Subsequently, the subjects were asked to perform muscle stretching.

The HRV index (TLHRV) was calculated using Pre5, Post5, and Post30 RMSSD values for each training day to evaluate internal load, as shown in equation (13):

(13)

Pre5: 5 minutes RMSSD before formal training; Post5: 5 minutes RMSSD immediately after training; Post30: 30–35 minutes RMSSD after training.

2.3 Statistical analysis

Each indicator (external load and internal load) is normalized with an interval of 0.1–1 using MATLAB 2021a (MathWorks, USA). This specific normalization range prevents potential division by zero errors in the model output indicator calculation. The transformation maintained relative scale differences while ensuring mathematical validity across all observations.

2.3.1 Parameter estimation and fitting effect evaluation method.

Using the first 80% of training day data from each subject as the model learning database, model fitting and parameter estimation were performed using MATLAB 2021a (MathWorks, USA). Three internal load indicators (ΔHRR1, HRr%, TLHRV) were used to calculate the output indicators of the model for fitting separately, as detailed in Section 2.2.3. These output indicators were then used in the cost function. The parameters were estimated by minimizing the sum of squared errors (SSE) between model-predicted values and observed values, using nonlinear least-squares minimization via the Trust-Region algorithm within the Curve Fitting Toolbox. The optimization process utilized the toolbox’s default configuration, including automatic initial parameter generation and standard convergence criteria. Initial conditions were handled by treating the integration constants C1 and C2 as adjustable parameters, which were estimated together with the other model coefficients (a,τa, Ka, f, τf, Kf). The fitted model surfaces were visualized as three-dimensional plots. The goodness of fit was evaluated using the coefficient of determination (R2), SSE, and root mean square error (RMSE).

2.3.2 Model prediction ability evaluation method.

Use the fitted parameters to predict the performance of each subject in the last 20% training day data (the model test database), compare it with the actual value, calculate the RMSE and mean absolute percentage error (MAPE) in the model test database, and evaluate the prediction ability of the model. Moreover, temporal dependency analysis using Spearman rank correlation analysis was used to investigate the relationship between prediction horizon (days since last training day in the model learning database) and absolute percentage error, assessing whether prediction errors systematically vary with temporal distance. Statistical significance was determined at p < 0.05.

To further evaluate the performance of the fitted model, the above indicators were compared with the same indicators’ values obtained from the FFM original model before optimization [20,19]. The normality of differences was assessed using the Shapiro-Wilk test. Paired t-tests (for normally distributed differences) or Wilcoxon tests (for non-normally distributed differences) were then performed, with Bonferroni correction for multiple comparisons. All analyses were conducted using MATLAB 2021a (MathWorks, USA) at a significance level of p < 0.05. The original model is shown in equation (14):

(14)

Pn represents the performance level on the nth day. Wn represents the external load level on the nth day. A and F represents the adaptation and fatigue levels generated after training. Ka and Kf are gain coefficients for adaptation and fatigue, respectively.

3 Results

3.1 Subjects’ characteristics and training completion status

Training data comprised paired external load and internal load measurements from each session, with 28–42 days of records per participant and a total of 420 paired data from 13 subjects. Due to the separate training of participants and the use of longitudinal data from each participant for personalized parameter estimation and model testing, the training completion status was different (Table 1). The Training Attendance Rate, defined as the ratio of actual training sessions attended to planned sessions and averaging 88.5 ± 5.9% across all subjects, indicates excellent compliance among the participants. Most deviations from 100% attendance were attributed to menstrual cycles and the overlap of the final training week with the school winter break. Three overweight subjects (BMI ≥ 24 kg/m2) and two subjects with low physical activity levels were given additional sports risk monitoring to prevent injury. None of the subjects withdrew from the study, further underscoring the good compliance throughout the research period. The detailed training data for each training session of all participants can be found in S1 File.

thumbnail
Table 1. Subjects’ characteristics and training situation.

https://doi.org/10.1371/journal.pone.0337824.t001

3.2 Fitting effect evaluation result

Using ΔHRR1, HRr%, and TLHRV to calculate the output indicators for model fitting and parameter estimation separately (the ratio of external load to internal load on the same day), each individual obtained their own three fitting models parameters using their longitudinal data (S2-S4 Tables).

When utilizing ΔHRR1 as a factor in calculating the output indicators of the model, the evaluation of the model fitting effect showed that all 13 subjects obtained small SSE and RMSE, with R2 values ranging from 0.54 to 0.84 (54% of subjects exhibited R2 values between 0.50 and 0.70, and 46% of subjects exhibited R2 values between 0.70 and 0.90). As for HRr%, the majority of subjects achieved good model fitting effect and low error, but there were also four subjects with R2 values less than 0.50. When it comes to TLHRV, most subjects achieved good fitting results, with R2 values ranging from 0.40 to 0.95 and RMSE being low. However, three individuals had slightly high SSE. The fitting effect of the original model is significantly worse than that of the optimized fitting model in this study. Statistical analysis revealed significant increases in R2 for ΔHRR1 (Bonferroni-corrected p = 0.030) and HRr% (Bonferroni-corrected p = 0.021). Moreover, most fitting indicators showed consistent positive trends in the optimized model, as presented in Fig 2. The SSE, RMSE, and R2 values obtained through fitting with different output indicators are also shown in Fig 2. Specific values for each subject are provided in S5S7 Tables.

thumbnail
Fig 2. R2 (a), RMSE (b), and SSE (c) values obtained for each subject after fitting with different output indicators in the optimized and original model.

* indicates p < 0.05 (Bonferroni-corrected) where the optimized model significantly better. † indicates a non-significant trend favoring the optimized model (mean difference >0). †? indicates inconsistent direction.

https://doi.org/10.1371/journal.pone.0337824.g002

3.3 Model prediction ability evaluation results

Although the prediction error of some subjects’ fitting models is slightly higher than that of the original model, the overall predictive ability of the fitting model is still better. Temporal dependency analysis showed there were no significant relationships between prediction horizon and absolute percentage error for most subjects, indicating no population-wide consistency in temporal dependency of prediction accuracy. Significant but directionally inconsistent correlations were observed in a few cases, suggesting a highly individualized impact of temporal factors on prediction accuracy. The evaluation results of predictive ability are shown in Fig 3. Specific values for each subject, including those from the temporal dependency analysis, are provided in S8S13 Tables.

thumbnail
Fig 3. MAPE (a) and RMSE (b) values obtained for each subject through model test database in the optimized and original model.

† indicates a non-significant trend favoring the optimized model (mean difference >0). †? indicates inconsistent direction.

https://doi.org/10.1371/journal.pone.0337824.g003

Taking subject 4 as an example, the 3D images of the fitted model obtained using three output indicators are shown in Fig 4.

thumbnail
Fig 4. Left side view(a, d, g), frontal view(b, e, h), and right side view(c, f, i) of the fitting models’ 3D images of subject 4 using three databases separately.

The black dots represent the model learning data, the green dots represent the model testing data. P1, P2, and P3 represent model databases where the output indicators are calculated by HRR1, HRr%, and TLHRV respectively. W is the external load. P is the performance. T is the nth day within the training period.

https://doi.org/10.1371/journal.pone.0337824.g004

4 Discussion

The model optimized mathematical relationships in FFM, reflecting the time-varying characteristics of adaptation and fatigue. The fitting effect and predictive ability of the optimized fitting model are generally better than those of the original model, though individual variation exist. What’s more, the use of longitudinal data also enables the model to have good fitting performance and predictive ability. HRV and HRR related indicators were used to verify the mathematical relationship in the model, and all indicators can bring good model fitting effects, indicating the generalizability of these indicators in mass sports. However, there are individual differences in the selection of the optimal output indicators of the model, and it is necessary to expand the collection of confounding factors, to explore and eliminate relevant factors that may introduce “noise” into the analysis.

The foundation for achieving good model performance is the selection of sensitive output indicators. Immediately after training, faster autonomic neuromodulation can lead to greater HRR in a shorter time, resulting in improved performance [49]. Because of that, the post-exercise HRR index which has previously served as a health monitoring indicator, has found widespread application in the training field [5054] and is widely regarded as a robust indicator reflecting individual health and training status [49]. As for TLHRV, HRV is sensitive to physiological and psychological stress, as well as steady-state disturbances, making it capable of immediately reflecting the training stimulus response after daily training, particularly moderate-intensity training [57]. Therefore, utilizing HRV as a guide for training is a reliable and promising method [35,4448]. In summary, the three indicators selected in this study can effectively reflect the changes in response and performance after training, laying the foundation for the fitting of the model. What’s more, these indicators can be conveniently collected through wearable devices, which are generalizable model output indicators that can be promoted in mass sports.

Moreover, the key to achieving good model performance in this study also lies in the application of longitudinal data to establish parameters for different individuals separately. To enhance the effectiveness of training, many scholars have pointed out the importance of using HRV and HRR indicators for longitudinal analysis. Some studies have indicated that when comparing different athletes horizontally, there is no consistent correlation between individuals with faster HRr% and better competition performance. However, in the longitudinal comparison of athletes themselves, changes in HRr% can better reflect post-exercise response and fatigue levels, thereby aiding in improved performance [58]. Similarly, HRV also has substantial individual differences, and many researchers recommend establishing longitudinal “HRV fingerprints” for individuals to monitor their HRV changes [58,59], which is consistent with this study. Marcelle investigated the role of HRV changes during the warm-up phase in training monitoring. They believe that HRV can effectively evaluate athletes’ physiological status and has great potential in guiding training. However, individual differences exist, and further data is needed to validate the results of this monitoring approach in a larger and more homogeneous population [59]. Plews specifically emphasized the need to explore the HRV characteristics of athletes and implement longitudinal monitoring within each individual to track the unique “HRV fingerprints” of different athletes. Training should then be guided based on these fingerprints to enhance athletic performance. Multiple studies have also corroborated this perspective [34,45,58]. In collecting subjects’ data, our study observed distinct personalized characteristics in the internal load of subjects, particularly in HRV indicators, where individuals exhibited significant variations in quiet HRV values. However, the HRV changes before and after training were remarkably similar, aligning with previous research findings [35,58]. Therefore, a significant breakthrough in training monitoring lies in identifying a method that can capture pre- and post-training indicator changes while also reflecting individual characteristics and identifying their unique “fingerprints” to enhance training guidance effectiveness. The model developed in this study successfully fulfills this requirement. By leveraging the characteristics of FFM, the model can mathematically depict the fundamental patterns before and after training. Additionally, the integration of machine learning empowers the model with the capability to identify an individual’s “fingerprint” based on their training data, thereby facilitating personalized evaluation. Founded on real-time training data acquisition through wearable devices, the model incorporates machine learning to continuously learn and capture individual physiological characteristics. This approach establishes a foundation for developing adaptive training monitoring software [60]. Once integrated into smartphone applications, such software can provide users with an intuitive interface, offering personalized training adjustment recommendations, such as intensity modulation and rest advice. Moreover, temporal dependency analysis in this study further revealed that the temporal dependency of prediction errors lacks population-wide consistency, with only a few individuals showing significant but directionally inconsistent correlations. This suggests that the sensitivity of prediction accuracy to temporal distance is highly individualized. Future research could explore adding a “time-adaptive module” to exercise monitoring platforms. This module can determine whether an individual’s model predictions exhibit temporal dependence based on their training data. For individuals with negative correlations, longer training cycles could be adopted, while those with positive correlations might benefit from high-frequency short cycles. This is crucial for enhancing the scientific nature of training in mass sports. Meanwhile, this application model also holds significant promise in both clinical and competitive sports fields. In clinical sports rehabilitation, customizing training loads according to individual recovery situations in a safer and more precise manner can maximize rehabilitation efficacy and minimize injury risks [61], which is consistent with the application outlook of optimizing the model in this study. As for competitive sports, given that the FFM originated in this domain, its combination with machine learning can further provide a basis for precise training.

Although three indicators have a good fitting effect, there is a “personalized” characteristic in the selection of the optimal output indicators of the model. When using HRr% to calculate the output indicators, subjects 2, 11, and 13 had slightly high fitting errors in their models, but they all achieved good fitting effects and predictive ability when using ΔHRR1 and TLHRV as the output indicators. Similarly, the predictive ability test revealed that the majority of subjects exhibited good model fits with TLHRV, while subjects 3 and 6 had slightly higher prediction errors. The above analysis suggests that there is a “personalized” characteristic in the selection of the optimal output indicators of the model. Individuals with varying levels of baseline training exhibit different degrees of changes in their autonomic nerve activities when subjected to the same intensity of training. Therefore, studying the patterns of HRR, HRV, and other indicators reflecting autonomic nerve activities should be conducted on a population level. Several researchers have highlighted that individuals with different training levels display varying HRR responses [62], and factors such as training level and gender can influence the effectiveness of HRV-based endurance training guidance [45]. Consequently, the pre-training physical activity level and daily physical activity during training may lead to different autonomic nervous responses. Although this study required subjects to refrain from engaging in other training activities outside of the experiment, the pre-experiment physical activity levels were not screened, and daily physical activity was not restricted during the experiment. This could explain the slight variations in the optimal model output indicators observed among the participants in this study. Furthermore, even individuals with the same training level may exhibit subtle differences in their autonomic nervous responses, leading to specific changes in their HRR and HRV indicators. Some studies have highlighted that while both HRV and HRR can reflect changes in autonomic nerve activity, these two indicators have distinct physiological determinants. As a result, they may reflect different aspects of cardiac parasympathetic nervous system function [63]. Additionally, differences in the practical application of these two indicators have also been noted. Moreover, the specific factors that influence these indicators (such as training intensity, peripheral stimuli, genetic factors, etc.) are not yet fully understood, necessitating further exploration [35,45]. Therefore, the complexity of the physiological mechanisms underlying HRR and HRV may contribute to the varying optimal model output indicators observed among the participants in this study.

What’s more, another possibility of the reasons that cause the “personalized” characteristic in the selection of the optimal output indicators is the “noise” in the data. Some experiments examining the autonomic nervous activity of subjects experiencing overtraining (particularly long-term overtraining) have reported different HRV responses among individuals, including increases, decreases, and no change [34]. Therefore, despite numerous practical studies demonstrating the effectiveness of HRV indicators in training monitoring [44,45,58,59], it is still crucial to exercise caution when interpreting their values [58]. Although the model developed in this study effectively captures changes in performance, it does not account for the occurrence of overtraining. Considering the complex physiological and biochemical reactions that can take place in the body after overtraining, the assumption of “no overtraining during the training process” was made when establishing the model in this study. Since this study solely utilizes collected data to test the model’s fitting and predictive abilities, without employing the model for dynamic adjustment of the training plan, there may be deviations from this assumption during actual data collection, resulting in some “noise” that impacts the model’s predictive performance. However, this is merely the author’s speculation, and some researchers have also raised concerns regarding studies on overtraining individuals mentioned earlier. They argue that inconsistent results in such studies may be attributed to methodological differences, as the experimental design of these studies employs extreme training plans to induce an overtraining state, which does not align with real-world training scenarios [34]. Therefore, further investigation should be conducted to delve deeper into the changes and physiological mechanisms of output indicators under different conditions [35], to better understand the factors influencing model fitting and predictive effectiveness. Additionally, the differences in training period between participants may have influenced the model performance. Although the overall training attendance rate in our study was high and the missing training sessions did not show a regular impact on the model fitting effect, the variation in training period should not be overlooked. Longer training periods could provide more data points for parameter estimation, potentially enhancing model accuracy. However, due to individual differences in training responses, changes in training period do not directly correspond to proportional changes in model performance. Future studies could investigate this relationship further by controlling for training period or including it as a covariate in the analysis. To better capture the comprehensive impact of training period differences on model performance, it is also recommended that future research employ larger sample sizes and collect data over longer timeframes.

Overall, this study has several limitations that should be acknowledged. While the model demonstrated good fitting and prediction effects, it lacks an in-depth exploration of the personalized differences observed in various output indicators. Future research should focus on identifying optimal output indicators for individuals. This would provide a more robust foundation for selecting model indicators and achieving more stable fitting results. Additionally, the factors affecting model fitting and prediction need to be studied further. Future studies should collect more confounding factors to explore and eliminate those that may introduce “noise” into the analysis. This will serve as a basis for further refining the model and enhancing its practical application. Moreover, the sample used in this study is homogeneous (average age 22 ± 3 years) and only involves cycling as the training type. This limits the generalizability of the findings to other age groups and training types. Future research should validate the model across different sports, age groups, and genders to ensure its reliability and effectiveness in diverse populations. To address the risk of overfitting, future studies should also employ rigorous cross-validation techniques and consider incorporating additional variables that may influence model outcomes. Multicenter and randomized studies are also recommended to further establish the model’s validity and applicability across different contexts. Finally, the current model does not account for overtraining. Future research could investigate extreme autonomic responses, such as paradoxical changes in HRV, to enhance the model’s ability to detect and interpret specific training states.

5 Conclusion

The good performance of the model is attributed to the optimization of functional relationships and the application of individual longitudinal data. HRV and HRR related indicators are generalizable model output indicators that can be used to predict training effectiveness in mass sports, though individual differences in optimal indicators selection exist. This study provides a foundation for applying the model in wearable technology, offering a basis for developing adaptive training monitoring software. Such software, when integrated into smartphones, can provide users with personalized training recommendations, offering potential applications for general population, coaches, and rehabilitation physicians.

Supporting information

S1 Table. Physiological interpretation of optimized FFM parameters.

https://doi.org/10.1371/journal.pone.0337824.s001

(DOCX)

S2 Table. Model parameter estimation results (using ΔHRR1 to calculate the output indicators).

https://doi.org/10.1371/journal.pone.0337824.s002

(DOCX)

S3 Table. Model parameter estimation results (using HRr% to calculate the output indicators).

https://doi.org/10.1371/journal.pone.0337824.s003

(DOCX)

S4 Table. Model parameter estimation results (using TLHRV to calculate the output indicators).

https://doi.org/10.1371/journal.pone.0337824.s004

(DOCX)

S5 Table. Evaluation results of model fitting effect (using ΔHRR1 to calculate the output indicators).

https://doi.org/10.1371/journal.pone.0337824.s005

(DOCX)

S6 Table. Evaluation results of model fitting effect (using HRr% to calculate the output indicators).

https://doi.org/10.1371/journal.pone.0337824.s006

(DOCX)

S7 Table. Evaluation results of model fitting effect (using TLHRV to calculate the output indicators).

https://doi.org/10.1371/journal.pone.0337824.s007

(DOCX)

S8 Table. Evaluation results of model prediction ability and temporal dependency analysis (using ΔHRR1 to calculate the output indicators).

https://doi.org/10.1371/journal.pone.0337824.s008

(DOCX)

S9 Table. The predicted values and actual values obtained from the model (using ΔHRR1 to calculate the output indicators and taking Subject 4 as an example).

https://doi.org/10.1371/journal.pone.0337824.s009

(DOCX)

S10 Table. Evaluation results of model prediction ability and temporal dependency analysis (using HRr% to calculate the output indicators).

https://doi.org/10.1371/journal.pone.0337824.s010

(DOCX)

S11 Table. The predicted values and actual values obtained from the model (using HRr% to calculate the output indicators and taking Subject 4 as an example).

https://doi.org/10.1371/journal.pone.0337824.s011

(DOCX)

S12 Table. Evaluation results of model prediction ability and temporal dependency analysis (using TLHRV to calculate the output indicators).

https://doi.org/10.1371/journal.pone.0337824.s012

(DOCX)

S13 Table. The predicted values and actual values obtained from the model (using TLHRV to calculate the output indicators and taking Subject 4 as an example).

https://doi.org/10.1371/journal.pone.0337824.s013

(DOCX)

Acknowledgments

We thank all the subjects who participated in this study. And we also acknowledge assistance from proof-readers and editors.

References

  1. 1. Fernandez-Fernandez J, Sanz D, Sarabia JM, Moya M. The Effects of Sport-Specific Drills Training or High-Intensity Interval Training in Young Tennis Players. Int J Sports Physiol Perform. 2017;12(1):90–8. pmid:27140481
  2. 2. Fox JL, Stanton R, Sargent C, Wintour S-A, Scanlan AT. The Association Between Training Load and Performance in Team Sports: A Systematic Review. Sports Med. 2018;48(12):2743–74. pmid:30225537
  3. 3. Lima RF, Silva A, Afonso J, Castro H, Clemente FM. External and internal Load and their Effects on Professional Volleyball Training. Int J Sports Med. 2020;41(7):468–74. pmid:32059245
  4. 4. McGuigan MR, Wright GA, Fleck SJ. Strength training for athletes: does it really help sports performance?. Int J Sports Physiol Perform. 2012;7(1):2–5. pmid:22461461
  5. 5. Osipov AY, Slizik M, Bartik P, Kudryavtsev MD, Iermakov SS. The impact of various strength training protocols on the strength and sporting performance of junior male judokas. Archives of Budo. 2021;17:153-U9.
  6. 6. Foster C, Rodriguez-Marroyo JA, de Koning JJ. Monitoring Training Loads: The Past, the Present, and the Future. Int J Sports Physiol Perform. 2017;12(Suppl 2):S22–8. pmid:28253038
  7. 7. Saunders DH, Sanderson M, Hayes S, Johnson L, Kramer S, Carter DD, et al. Physical fitness training for stroke patients. Cochrane Database of Systematic Reviews. 2020;2020(3).
  8. 8. Platta ME, Ensari I, Motl RW, Pilutti LA. Effect of Exercise Training on Fitness in Multiple Sclerosis: A Meta-Analysis. Arch Phys Med Rehabil. 2016;97(9):1564–72. pmid:26896750
  9. 9. Mendonça FR, Ferreira de Faria W, Marcio da Silva J, Massuto RB, Castilho Dos Santos G, Correa RC, et al. Effects of aerobic exercise combined with resistance training on health-related physical fitness in adolescents: A randomized controlled trial. J Exerc Sci Fit. 2022;20(2):182–9. pmid:35401769
  10. 10. Drew MK, Finch CF. The Relationship Between Training Load and Injury, Illness and Soreness: A Systematic and Literature Review. Sports Med. 2016;46(6):861–83. pmid:26822969
  11. 11. Banister EW, Calvert TW, Savage MV, Bach TM. A systems model of training for athletic performance. Australian Journal of Sports Medicine. 1975;7(3):57–61.
  12. 12. Witard OC, Bannock L, Tipton KD. Making Sense of Muscle Protein Synthesis: A Focus on Muscle Growth During Resistance Training. Int J Sport Nutr Exerc Metab. 2022;32(1):49–61. pmid:34697259
  13. 13. Hadjispyrou S, Dinas PC, Delitheos SM, Koumprentziotis I-A, Chryssanthopoulos C, Philippou A. The Effect of High-Intensity Interval Training on Mitochondrial-Associated Indices in Overweight and Obese Adults: A Systematic Review and Meta-Analysis. Front Biosci (Landmark Ed). 2023;28(11):281. pmid:38062841
  14. 14. Tumkur Anil Kumar N, Oliver JL, Lloyd RS, Pedley JS, Radnor JM. The Influence of Growth, Maturation and Resistance Training on Muscle-Tendon and Neuromuscular Adaptations: A Narrative Review. Sports. 2021;9(5):59.
  15. 15. Vigh-Larsen JF, Ørtenblad N, Spriet LL, Overgaard K, Mohr M. Muscle Glycogen Metabolism and High-Intensity Exercise Performance: A Narrative Review. Sports Med. 2021;51(9):1855–74. pmid:33900579
  16. 16. Allen DG, Lamb GD, Westerblad H. Skeletal muscle fatigue: cellular mechanisms. Physiol Rev. 2008;88(1):287–332. pmid:18195089
  17. 17. Fiorenza M, Hostrup M, Gunnarsson TP, Shirai Y, Schena F, Iaia FM, et al. Neuromuscular Fatigue and Metabolism during High-Intensity Intermittent Exercise. Med Sci Sports Exerc. 2019;51(8):1642–52. pmid:30817710
  18. 18. Becker KM, Goethel M, Parolini F, de Paula Silva V, Vilas-Boas JP, Ervilha UF. The Price of Force: Motor Unit Adaptation Mechanisms in Pain Versus Fatigue. Physiother Res Int. 2025;30(3):e70081. pmid:40514775
  19. 19. Banister EW, Hamilton CL. Variations in iron status with fatigue modelled from training in female distance runners. Eur J Appl Physiol Occup Physiol. 1985;54(1):16–23. pmid:4018049
  20. 20. Fitz-Clarke JR, Morton RH, Banister EW. Optimizing athletic performance by influence curves. J Appl Physiol (1985). 1991;71(3):1151–8. pmid:1757312
  21. 21. Imbach F, Sutton-Charani N, Montmain J, Candau R, Perrey S. The Use of Fitness-Fatigue Models for Sport Performance Modelling: Conceptual Issues and Contributions from Machine-Learning. Sports Med Open. 2022;8(1):29. pmid:35239054
  22. 22. Calvert TW, Banister EW, Savage MV, Bach T. A Systems Model of the Effects of Training on Physical Performance. IEEE Trans Syst, Man, Cybern. 1976;SMC-6(2):94–102.
  23. 23. Scott AL, Spaniol FJ, Ocker L, Bonnette R, Melrose D. The Application of a Systems Model to the Training and Performance of Adolescent Swimmers. Journal of Strength and Conditioning Research. 2011;25:S27–8.
  24. 24. Busso T, Carasso C, Lacour JR. Adequacy of a systems structure in the modeling of training effects on performance. J Appl Physiol (1985). 1991;71(5):2044–9. pmid:1761506
  25. 25. Busso T, Candau R, Lacour JR. Fatigue and fitness modelled from the effects of training on performance. Eur J Appl Physiol Occup Physiol. 1994;69(1):50–4. pmid:7957156
  26. 26. Banister EW, Carter JB, Zarkadas PC. Training theory and taper: validation in triathlon athletes. Eur J Appl Physiol Occup Physiol. 1999;79(2):182–91. pmid:10029340
  27. 27. Wallace LK, Slattery KM, Coutts AJ. A comparison of methods for quantifying training load: relationships between modelled and actual training responses. Eur J Appl Physiol. 2014;114(1):11–20. pmid:24104194
  28. 28. Stephens Hemingway BH, Burgess KE, Elyan E, Swinton PA. The effects of measurement error and testing frequency on the fitness-fatigue model applied to resistance training: A simulation approach. International Journal of Sports Science & Coaching. 2020;15(1):60–71.
  29. 29. Morton RH. The quantitative periodization of athletic training: A model study. Sports Medicine, Training and Rehabilitation. 1991;3(1):19–28.
  30. 30. Bellinger P. Functional Overreaching in Endurance Athletes: A Necessity or Cause for Concern?. Sports Med. 2020;50(6):1059–73. pmid:32064575
  31. 31. Mistiri ME, Rivera DE, Klasnja P, Park J, Hekler E. Model Predictive Control Strategies for Optimized mHealth Interventions for Physical Activity. Proc Am Control Conf. 2022;2022:1392–7. pmid:36238385
  32. 32. Vandelanotte C, Trost S, Hodgetts D, Imam T, Rashid MDM, To QG, et al. MoveMentor-examining the effectiveness of a machine learning and app-based digital assistant to increase physical activity in adults: protocol for a randomised controlled trial. Trials. 2025;26(1):233. pmid:40597442
  33. 33. Kozan Cikirikci EH, Esin MN. The impact of machine learning on physical activity-related health outcomes: A systematic review and meta-analysis. Int Nurs Rev. 2025;72(2):e70019. pmid:40353307
  34. 34. Plews DJ, Laursen PB, Stanley J, Kilding AE, Buchheit M. Training adaptation and heart rate variability in elite endurance athletes: opening the door to effective monitoring. Sports Med. 2013;43(9):773–81. pmid:23852425
  35. 35. Mishica C, Kyröläinen H, Hynynen E, Nummela A, Holmberg H-C, Linnamo V. Relationships between Heart Rate Variability, Sleep Duration, Cortisol and Physical Training in Young Athletes. J Sports Sci Med. 2021;20(4):778–88. pmid:35321140
  36. 36. Seshadri DR, Li RT, Voos JE, Rowbottom JR, Alfes CM, Zorman CA, et al. Wearable sensors for monitoring the internal and external workload of the athlete. NPJ Digit Med. 2019;2:71. pmid:31372506
  37. 37. Schwellnus M, Soligard T, Alonso J-M, Bahr R, Clarsen B, Dijkstra HP, et al. How much is too much? (Part 2) International Olympic Committee consensus statement on load in sport and risk of illness. Br J Sports Med. 2016;50(17):1043–52. pmid:27535991
  38. 38. Craig CL, Marshall AL, Sjöström M, Bauman AE, Booth ML, Ainsworth BE, et al. International physical activity questionnaire: 12-country reliability and validity. Med Sci Sports Exerc. 2003;35(8):1381–95. pmid:12900694
  39. 39. Hagströmer M, Oja P, Sjöström M. The International Physical Activity Questionnaire (IPAQ): a study of concurrent and construct validity. Public Health Nutr. 2006;9(6):755–62. pmid:16925881
  40. 40. Kim Y, Park I, Kang M. Convergent validity of the international physical activity questionnaire (IPAQ): meta-analysis. Public Health Nutr. 2013;16(3):440–52. pmid:22874087
  41. 41. Sport GA. National Fitness Guide. Chinese General Administration of Sport. 2017.
  42. 42. GmbH E, ergoselect. Bicycle ergometer. Product technical page. 2025. https://www.ergoline.com/en/__productdetails/artn-ergoselect-100.html
  43. 43. Yong H. Modeling of the training load - performance relationship in sanda competitors. Shanghai University of Sport. 2011.
  44. 44. da Silva DF, Ferraro ZM, Adamo KB, Machado FA. Endurance Running Training Individually Guided by HRV in Untrained Women. J Strength Cond Res. 2019;33(3):736–46. pmid:28570494
  45. 45. Granero-Gallegos A, González-Quílez A, Plews D, Carrasco-Poyatos M. HRV-Based Training for Improving VO2max in Endurance Athletes. A Systematic Review with Meta-Analysis. Int J Environ Res Public Health. 2020;17(21):7999. pmid:33143175
  46. 46. Roy S, Singh N, Gupta R. Heart rate variability in different intensities of static and dynamic exercises graded using similar heart rate. Sport Sci Health. 2020;16(4):669–75.
  47. 47. Grad C. Heart rate variability and heart rate recovery as prognostic factors. Clujul Med. 2015;88(3):304–9. pmid:26609261
  48. 48. Hwang I-S, Noh Y-S, Jeong I-C, Yoon H-R. Optimized exercise load control system based on Heart Rate Variability. Biomed Eng Lett. 2011;1(4):232–8.
  49. 49. Daanen HAM, Lamberts RP, Kallen VL, Jin A, Van Meeteren NLU. A systematic review on heart-rate recovery to monitor changes in training status in athletes. Int J Sports Physiol Perform. 2012;7(3):251–60. pmid:22357753
  50. 50. Lamberts RP, Maskell S, Borresen J, Lambert MI. Adapting workload improves the measurement of heart rate recovery. Int J Sports Med. 2011;32(9):698–702. pmid:21618159
  51. 51. Bilsel T, Terzi S, Akbulut T, Sayar N, Hobikoglu G, Yesilcimen K. Abnormal heart rate recovery immediately after cardiopulmonary exercise testing in heart failure patients. Int Heart J. 2006;47(3):431–40. pmid:16823249
  52. 52. Ganzevles SPM, de Haan A, Beek PJ, Daanen HAM, Truijens MJ. Heart-Rate Recovery After Warm-up in Swimming: A Useful Predictor of Training Heart-Rate Response?. Int J Sports Physiol Perform. 2017;12(6):742–8. pmid:27736252
  53. 53. Le Meur Y, Buchheit M, Aubry A, Coutts AJ, Hausswirth C. Assessing Overreaching With Heart-Rate Recovery: What Is the Minimal Exercise Intensity Required?. Int J Sports Physiol Perform. 2017;12(4):569–73. pmid:27617566
  54. 54. Villelabeitia-Jaureguizar K, Vicente-Campos D, Senen AB, Jiménez VH, Garrido-Lestache MEB, Chicharro JL. Effects of high-intensity interval versus continuous exercise training on post-exercise heart rate recovery in coronary heart-disease patients. International Journal of Cardiology. 2017;244:17–23.
  55. 55. Oy PE. H10 heart rate sensor - technical specifications. 2025. https://www.polar.com/en/sensors/h10-heart-rate-sensor
  56. 56. Schaffarczyk M, Rogers B, Reer R, Gronwald T. Validity of the Polar H10 Sensor for Heart Rate Variability Analysis during Resting State and Incremental Exercise in Recreational Men and Women. Sensors (Basel). 2022;22(17):6536. pmid:36081005
  57. 57. Saboul D, Balducci P, Millet G, Pialoux V, Hautier C. A pilot study on quantification of training load: The use of HRV in training practice. Eur J Sport Sci. 2016;16(2):172–81. pmid:25657120
  58. 58. Schüttler D, Hamm W, Bauer A, Brunner S. Routine heart rate-based and novel ECG-based biomarkers of autonomicnervous system in sports medicine. Dtsch Z Sportmed. 2020;71(6):141–50.
  59. 59. Schaffarczyk M, Rogers B, Reer R, Gronwald T. Fractal correlation properties of HRV as a noninvasive biomarker to assess the physiological status of triathletes during simulated warm-up sessions at low exercise intensity: a pilot study. BMC Sports Sci Med Rehabil. 2022;14(1):203. pmid:36457040
  60. 60. Weyns D, Schmerl B, Kishida M, Leva A, Litoiu M, Ozay N, et al. Towards Better Adaptive Systems by Combining MAPE, Control Theory, and Machine Learning. In: 2021 International Symposium on Software Engineering for Adaptive and Self-Managing Systems (SEAMS), 2021. 217–23.
  61. 61. Taylor JB, Ford KR, Queen RM, Owen EC, Gisselman AS. Incorporating Internal and External Training Load Measurements in Clinical Decision Making After ACL Reconstruction: A Clinical Commentary. Int J Sports Phys Ther. 2021;16(2):565–78. pmid:33842052
  62. 62. Bentley RF, Vecchiarelli E, Banks L, Gonçalves PEO, Thomas SG, Goodman JM. Heart rate variability and recovery following maximal exercise in endurance athletes and physically active individuals. Appl Physiol Nutr Metab. 2020;45(10):1138–44. pmid:32294393
  63. 63. Casanova-Lizón A, Manresa-Rocamora A, Flatt AA, Sarabia JM, Moya-Ramón M. Does Exercise Training Improve Cardiac-Parasympathetic Nervous System Activity in Sedentary People? A Systematic Review with Meta-Analysis. Int J Environ Res Public Health. 2022;19(21):13899. pmid:36360777