Figures
Abstract
Background
A national clinical registry was established in the Netherlands containing data directly sampled from electronic health record systems of physical therapists (PTs). This registry aims to evaluate the potential of patient reported outcome measures (PROMs) to develop quality indicators (QIs) in physical therapy care.
Purpose
To test to what extent the collected PROM data are reliable, valid and discriminatory between practices in measuring outcomes of patients with non-specific low back pain (NSLBP).
Methods
In this retrospective cohort study 865 PT practices with 6,560 PTs voluntarily collected PROM data of patients with NSLBP, using the Quebec Back Pain Disability Scale (QBPDS), the Numeric Pain Rating Scale (NPRS) and the Patient Specific Functioning Scale (PSFS). Reliability was determined by analysing the completeness of the dataset, the comparability by using national reference data, and through checking selection bias in the included patients. Validity was tested using the known-groups contrast between patients with (sub)acute vs. chronic NSLBP. To determine discriminative ability of outcomes between PT practices, case-mix corrected hierarchical multilevel analyses were performed.
Results
Reliability was sufficient by confirming fifteen of the sixteen hypotheses: 59% of all patients opted in for data analysis, 42% of these included patients showed repeated measurement, comparing with reference data and potential selection bias showed < 5% between group differences, while differences between (sub)acute and chronic NSLB-groups were significantly larger than 5% (less treatment sessions, lager differences in outcomes in (sub)acute NSLB patients). In addition, all nine adjusted hierarchical multilevel models confirm that the collected dataset on outcomes in PT care is able to discriminate between practices using PROM results of patients with NSLBP (ICC-scores range 0.11–0.21).
Limitations
Although we have shown the reliability, validity and discriminative ability of the dataset in the quest to develop QIs, we are aware that reducing missing values in patient records and the selective participation of PTs that belong to the innovators needs attention in the next stages of implementation to avoid bias in the results.
Citation: Meerhoff GA, Verburg AC, Schapendonk RM, Cruijsberg J, Nijhuis-van der Sanden MWG, van Dulmen SA, et al. (2021) Reliability, validity and discriminability of patient reported outcomes for non-specific low back pain in a nationwide physical therapy registry: A retrospective observational cohort study. PLoS ONE 16(6): e0251892. https://doi.org/10.1371/journal.pone.0251892
Editor: Alison Rushton, Western University, CANADA
Received: January 30, 2021; Accepted: May 4, 2021; Published: June 3, 2021
Copyright: © 2021 Meerhoff et al. This is an open access article distributed under the terms of the Creative Commons Attribution License, which permits unrestricted use, distribution, and reproduction in any medium, provided the original author and source are credited.
Data Availability: We are happy to be able to share the anonymized dataset with PLOS ONE. This enables us to comply with the important academic FAIR principles. We have added all necessary datasets and syntaxes as Supporting Information files. In uploading the datafiles (SPSS) we had to compress (.ZIP) the files.
Funding: The Quality in Motion program was funded by the association KNGF (Koninklijk Nederlands Genootschap voor Fysiotherapie). The funders had no role in study design, data collection and analysis, decision to publish, or preparation of the manuscript.
Competing interests: The authors have declared that no competing interests exist.
Introduction
Within healthcare, the use of health outcomes as quality indicators (QIs) to enable transparency of the service delivery is gaining momentum [1–5]. Although seemingly promising, there is still limited experience with the development and implementation of QIs based on health outcomes [6, 7]. Presumably, data collected in clinical registries can be used in developing such QIs, ultimately assisting in achieving more transparency of service delivery [8–10].
QIs are “measurement tools, screens, or flags that are used as guides to monitor, evaluate, and improve the quality of patient care, clinical support services, and organizational functions that affect patient outcomes” [11] (p 524). QIs consist of explicitly defined and measurable items referring to the structures, processes or outcomes of care [12–14], and have the potential to be used to support quality improvement, accountability and to provide transparency of service delivery in healthcare [15]. A prerequisite of QIs to be of added value, is that they are defined using data, for example from KNGF’s clinical registry, that are valid, reliable and able to discriminate between groups of interest [2].
The Royal Dutch Society for Physical Therapy (KNGF) aims to develop and implement QIs using data from a clinical registry. To do so they initiated a quality program called Quality in Motion [16]. In this program a national clinical registry has been developed that enables the anonymous collection of data on: patient characteristics, structure, processes and (patient reported) health outcomes. Data are only collected after a one-off informed consent was provided by the patient and their therapist and recorded in the patient file. The data is collected directly from the electronic health record (EHR) systems used in primary care physical therapy (PT) practices [16].
Clinical practice guidelines (CPGs) often include recommendations on healthcare delivery and outcomes that can be transformed into QIs [12, 17–20]. An example of such a recommendation is the use of (patient reported) outcomes that measure health-related aspects such as physical functioning or perceived pain and can be used to evaluate treatment [21–26]. Since the KNGF has a long history in developing high quality CPGs [22, 27], they decided to use the recommendations in their CPGs for the selection of patient reported outcome measures (PROMs) as a basis for developing QIs. PROMs were chosen for the development of QIs, since they enable the measurement of health outcomes based on the direct perception of patients and these instruments are recommended in the high quality CPGs of KNGF.
The purpose of this study was to test to what extent health outcomes collected with PROMs in the clinical registry of KNGF are reliable, valid and able to discriminate between PT practices. These psychometric properties have been tested on data collected in patients with non-specific low back pain (NSLBP), a patient category with a large prevalence in physical therapy care. The following research questions were formulated:
- What is the reliability and validity of health outcomes in the clinical registry measured with PROMs?
- To what extent do health outcomes of physical therapy care collected in the clinical registry discriminate between practices using PROM results of patients with NSLBP?
Materials and methods
Study design and setting
This is a retrospective observational cohort study based on data collected in KNGF’s clinical registry by PT practices in Dutch primary care physical therapy. All data collected between 01-01-2013 until 28-11-2018 were used.
Participants
A total of 865 PT practices with 6,560 physical therapists (PTs) voluntarily collected data of patients with NSLBP.
Data collection
Data on all patient categories that visited the PT were collected in the registry. This was done by uploading anonymous data from patient records in the clinically used EHR-systems to KNGF’s clinical registry. To ensure uniformity of the data collection [28], all data in the registry are collected based on predefined technical specifications [16]. This procedure of data collection has shown to be feasible [16].
Data were only uploaded to the clinical registry from the EHR-systems if: 1) the patient provided their informed consent for anonymous use of the clinical data from their patient record, and 2) the therapist provided their informed consent on the use of all patient records for which they received an informed consent of the patient. For the initial reliability analysis that focussed on the completeness and comparability, all data from the clinical registry were used. For the other analyses, that specifically aimed on patients with NSLBP a selection was made and data of patients were only included if: patients visited their PT due to NSLBP, the treatment episode was finished, patients were 18 years or older, if gender and the level of chronicity of their complaints was recorded (i.e. (sub)acute vs. chronic, with (sub)acute being 0–12 weeks since onset, and chronic >12 weeks [29]), and if a PT practice had collected data from at least 30 patients with NSLBP.
Selected outcome measures
We used the PROMs that were recommended in KNGF‘s clinical practice guideline (CPG) for low back pain [25]. These PROMs are: the Quebec Back Pain Disability Scale (QBPDS), the Numeric Pain Rating Scale (NPRS) and the Patient Specific Functioning Scale (PSFS).
The QBPDS is a 20-item PROM which measures the domain physical functioning using a 6-point Likert scale, ranging from 0 “not difficult at all” to 5 “unable to do”. The overall score on the QBPDS is the sum score of the 20 items. A minimum score (0) represents “not being disabled at all” and the maximum score (100) represents “being maximally disabled” [30, 31]. The QBPDS is a feasible PROM which takes patients approximately 10 minutes to administer. The QBPDS scores limited to moderate evidence for good reliability, validity and usability [32]. In addition, it has been identified as a responsive PROM with a minimal important change (MIC) of 20 points [32, 33].
The NPRS is an instrument that assesses the domain perceived pain intensity [34]. It is a 1-item questionnaire with an 11-point scale ranging from 0 “no pain” to 10 “extreme pain” [35, 36]. The NPRS is identified to be feasible and easy to administer [35, 36]. The NPRS scores moderate to high on the psychometric properties reliability and validity [34, 37] and is responsive with a MIC of 2 points [37, 38].
The PSFS, which is equivalent to the Patient Specific Complaints (PSC) instrument [39], measures the domain physical functioning and “involves four steps, in which the patient’s main activity problems are identified, prioritized, scored and evaluated” [40] (p. 2). Each activity is scored on a 11-point scale ranging from 0 “Able to perform activity at pre-injury level” to 10 “Unable to perform activity” [41]. The PSFS is a feasible instrument with good to excellent measurement properties (reliability, validity and sensitivity to change) [38, 41]. The MIC for the PSFS was set at 2 points [42].
Sample size
In deciding if the selected data exceeds the sample size threshold, a general rule of thumb concerning the ability to discriminate outcomes between practices states that a minimum of 30 PT practices are required which should include a minimum of 30 patients each equalling 900 patients [28, 43]. The included PROMs in this study measure the domains physical functioning, and pain intensity. Perreault and Dionne (2005) estimated the agreement between patients and PTs on the domain physical functioning (ICC = 0.56) and pain intensity (ICC = 0.55) [44]. This enabled us to calculate a more specific sample size, using the equation of Twisk (2013), presented in S1 Appendix [45]. For this study this resulted in a minimum sample size of 1963 patients, treated in 66 PT practices.
Data analysis
Reliability of the data.
The reliability of the included data was determined in four ways. First, by determining the completeness of the complete dataset of the clinical registry. This was done using the mean percentage of patients that were included (opted-in) in the database from the EHR-systems. Since data collection in the clinical registry is innovative and involves the anonymous processing of personal data on health status, we assumed that not all patients were willing or invited by their therapist to cooperate, which may result in selection bias if the number of patients that did not opt-in is too high. Given the early stage of implementation of the registry, we hypothesized it to be realistic to aim for a percentage of opted-in patients that lies above 50%. In addition, using the same dataset, completeness was determined by calculating the mean percentage of the (opted-in) patient records where a repeated pre- and post-treatment measurement with one of the selected PROMs was executed. Based on previous studies in the Netherlands and Israel, we hypothesized it to be realistic to set this percentage at a minimum of 40% [16, 46].
The second evaluation of reliability was estimating comparability of patient characteristics in the opted-in population in the complete clinical registry with national reference data. We compared the characteristics age and gender of all patient records of the clinical registry with a dataset that is considered to be the national reference data [47]. This analysis enabled us to check if the data in the registry is comparable to a dataset which is seen as a national benchmark. The age categories chosen for this study were aligned with those used in the national reference dataset.
The third evaluation estimated the potential selection bias in the dataset of patients with NSLBP selected from the clinical registry based on inclusion rates per practice. This analysis was executed by dividing the records of patients with NSLBP in the clinical registry into two groups: practices with low vs. practices with high inclusion rates. The two groups were created using the median percentage of patients per PT practice that are included (opted-in) in the registry, based on the patient’s informed-consent. The group with low inclusion rates was compared to the group with high inclusion rates on the characteristics age, gender and number of treatment sessions. This analysis enabled us to test again if ‘the percentage of opted-in patients’ created selection bias in the included population, which in turn would decrease reliability.
For the fourth reliability analysis, we further estimated potential selection bias in the dataset on patients with NSLBP through a different approach. The presence or absence of a repeated measurement with one of the selected PROMs (QBPDS, NPRS or PSFS) was used to create two groups. The group with a repeated measurement was compared to the group without a repeated measurement on the characteristics age, gender and number of treatment sessions. This enabled us to check if ‘the presence/absence of a repeated PROM measurement’ created selection bias in the included population, which in turn would decrease reliability.
Due to the large sample sizes it was expected that, using a t-test, the differences between the groups for the second, third and fourth reliability analyses would be statistically significant on all items. Nevertheless, we hypothesized that the analyses would not result in relevant differences between the created groups. We set an a priori threshold of >5% difference as relevant, i.e. groups were considered equal if the differences were ≤5%. Differences between the created categories were established by calculating the relative differences on the outcomes of the created groups.
Validity of the data.
Regarding the validity of the collected data on patients with NSLBP the known-groups validity was determined, which is a component of construct validity frequently used to determine psychometric properties of measurement instruments [48]. With known-groups validity, the ability to distinguish or discriminate among distinct groups is defined [49]. We expected that distinct groups in the clinical registry would be present, and we defined the known-groups to be patients with chronic NSLBP versus those with (sub)acute NSLBP. Based on the findings of Costa et al. (2012) we hypothesized that patients with chronic NSLBP would need a higher number of treatment sessions and would achieve a lower change score on the three PROMs of interest (QBPDS, NPRS and PSFS) [29]. Due to the large sample sizes it was expected that, using a t-test, the differences between the groups would be statistically significant on all items, therefore we set an a priori threshold of >5% difference between the groups as being relevant. Differences between the created categories were established by calculating the relative differences on the outcomes of the created groups. For this analysis only the records of patients with NSLBP were included in which data of at least one of the selected PROMs was available at baseline (T0) and endpoint (Tend) of treatment.
The ability to discriminate between practices.
To determine the ability of the collected dataset on patients with NSLBP in the clinical registry to discriminate outcomes between PT practices using the PROM results, several hierarchical multilevel analyses were performed. Initially, for each of the outcome measures -QBPDS, NPRS and PSFS- three intercept-only hierarchical multilevel models were estimated. Each model resulted in an intraclass correlation coefficient (ICC), representing the ability of the collected data to discriminate outcomes between PT practices. If an ICC is >0.10 it can be interpreted as adequate, indicating that the model is able to discriminate outcomes between PT practices [50]. The ICC values in discriminating outcomes of healthcare typically range between 0.05 and 0.20 [51]. The three different models were estimated since they represent the possible outcomes of the collected PROMs. The first model estimated the mean pre-posttreatment (T0—Tend) change scores with 95% confidence intervals (CI) of individual practices. The second model estimated the mean percentage and 95% CI of patients in which the MIC was achieved of individual practices. The third model estimated the mean post-treatment score (Tend) with 95% CI on the selected PROMs of individual practices. For each of the three models overall mean scores for all practices combined were also estimated.
Next, it was tested if stronger models were created when adjusted models were estimated, applying case-mix correction for the independent variables: age, gender, chronicity of the complaints (i.e. (sub)acute vs. chronic, with (sub)acute being 0–12 weeks since onset and chronic >12 weeks) and severity of complaints at the start of the treatment (using the results on the PROMs). These independent variables were seen as contextual variables that may influence the results on the outcome measures. As defined by Twisk (2019) [51], all such variables can be included in the model at once [51]. The ICC presents “a good gauge of whether a contextual variable has a significant effect on the outcome” [52] (p 818). The model with the highest ICC value represents the strongest model and was selected to estimate the ability of the collected data in the clinical registry to discriminate outcomes between PT practices.
Differences between groups for reliability analyses were statistically tested using independent-sampled t-tests or Fisher’s Exact test if the assumption on normality was met. IBM SPSS Statistics for Windows, version 23 (IBM Corp., Armonk, NY) was used for all analyses.
Ethical considerations
This study was conducted according to the principles of the Declaration of Helsinki (version October 2013) and in accordance with the Medical Research Involving Human Subjects Act (WMO). The study protocol was approved by the regional Medical Ethical Committee of Radboud university medical center (registration #2014/260).
Results
Overall the clinical registry contained 213,245 records of patients with NSLBP, collected from 865 PT practices, see S1 Fig for a flowchart of the patient inclusion. The mean age of the patients was 52.4 years (SD = 17.2), 55.0% of the patients was female and 77.5% of the patients had (sub)acute complaints. The registry contained 21,758, 54,904 and 73,554 patient episodes with repeated measurements of the QBPDS, NPRS and PSFS, respectively (See Table 1). See Table 2 for a summary on all executed analyses on the different psychometric properties which are explained below.
Reliability of the collected data
The analysis on the completeness, using the complete dataset of the clinical registry, showed that 59.2% of the patients provided permission for the use of their data, thus are opted-in. In total 41.7% of the complete dataset of the clinical registry has executed a repeated measurement with one of the selected PROMs. Both percentages exceed the benchmark that we have defined, confirming our hypotheses on completeness.
Comparing data from the complete dataset of the clinical registry with the national reference data showed a statistically non-significant difference in gender of 0.2%, and significant differences in all age groups with percentages between -2.3% and 2.5% (see Table 3). None of the comparisons exceeded the a priori defined threshold of 5% for being relevant differences, confirming our hypothesis on comparability.
Estimating the potential selection bias in the included patients with NSLBP, the analysis comparing practices with high versus low inclusion rates of patients, resulted in non-significant differences of -0.2%, 1.1% and -2.8% on age, gender and number of treatment sessions, respectively (see Table 4). The analysis comparing presence or absence of a repeated measurement with one of the selected PROMs, using the same dataset, resulted in statistically significant differences, of 2.7%, -0.2% and 35.0% on age, gender and number of treatment sessions, respectively (see Table 5). Five of the six analyses confirmed our hypotheses on selection bias since they did not exceed the a priori defined threshold of 5%.
Known-groups validity of the collected data
Patients with (sub)acute NSLBP needed 25.7% less treatment sessions than patients with chronic NSLBP. The three other analyses showed that patients with (sub)acute NSLBP compared to patients with chronic NSLBP achieved a 40.1%, 29.4% and 23.8% higher change score on the QBPDS, NPRS and the PSFS, respectively. The results of these four analyses for the known-groups validity all exceeded the set threshold of 5%, confirming our hypothesis on the known-groups validity of the data (see Table 6).
The ability to discriminate between practices based on the collected data
Seven of the nine ICC-scores obtained from the intercept-only hierarchical multilevel analyses exceed the a priori formulated threshold of 0.10 (see Table 7). In the adjusted-models all nine ICC-scores exceeded this threshold with scores ranging from 0.11–0.21. These results indicate that the adjusted-models are able to discriminate outcomes between PT practices, confirming the discriminative ability of the data [50].
Discussion
Overall the results of this study show that the analyses regarding the reliability of the data in the clinical registry were in line with our a priori formulated hypotheses: the data was sufficiently complete, comparable and we did not identify selection bias in the patients that were included in the registry. All analyses on the known-groups validity met the a priori formulated hypotheses. Based on these results it can be concluded that both reliability and known-groups validity are confirmed. In addition, the hierarchical multilevel analyses confirm that the collected dataset on outcomes in PT care is able to discriminate between practices using PROM results of patients with NSLBP.
Only one of the six analyses aiming to check for selection bias in the included patients, as part of the reliability analyses, did not meet the a priori formulated hypothesis. This analysis represented the number of treatment sessions required in patients with a repeated PROM measurement versus those without. The results showed that patients who required fewer treatments more often did not complete a repeated PROM measurement. This might be explained by the fact that PT’s and/or patient’s belief that the completions of a repeated PROM measurement have no added value for this patient category with a short treatment period and fast recovery.
To the best of our knowledge, within the field of physical therapy similar studies, which execute a so-called practice test, aiming to investigate the reliability, validity and discriminative ability of the collected data in a clinical registry, have not been published. In several countries and healthcare settings similar initiatives started the collection of PROMs scores in clinical registries, as a prerequisite for the development of QIs based on clinical data. A first example of such registry is initiated by the National Health Service in the United Kingdom, which introduced their national PROMs programme in 2009. A second example is initiated in 2010 by the Dutch Institute for Clinical Auditing (DICA). DICA manages 22 clinical registries, belonging to different scientific associations of medical specialists (e.g. neurologists & oncologists) [53]. Despite this experience in building clinical registries, several publications suggest that the integrated implementation of PROMs, using them to stimulate shared decision-making on clinical level and as performance information on managerial level has not yet been successfully achieved [8, 54].
In the quest for developing QIs using data from clinical registries, the execution of a practice test to evaluate the psychometric properties of the collected data is crucial. There are a limited number of studies in other fields than PT that have conducted such a practice test. Examples of such studies are analysed in the review on the validity of QI’s on the readmission rate of Fischer et al. (2012). This study showed that only a small proportion 21 of the 486 included papers test the actual validity of the data in some sort of practice test [55]. In addition, a systematic review of Langendam et al. (2020) showed that performing a practice test to validate the formulated QI which are based on recommendations of CPGs is relatively rare [20].
There are some publications providing a framework on the development of QIs [18, 56]. These frameworks also emphasize the importance of the execution of a practice test, since it enables formulating a benchmark for a QI that meets all psychometric criteria [2, 28, 56]. Unfortunately, several studies state that there is no standard definition of what a practice test has to contain [18, 20]. Often only the Delphi methodology is used to reach consensus in formulating QIs and practice tests are not executed [17, 19, 42, 57].
In short, there is no guidance available for applying a practice test as we have done in this study. The results of this study are therefore a good starting point for the development of standards for a practice test to be carried out in the development of benchmarks for QIs based on data from a clinical registry that meet all relevant psychometric properties.
Strengths and limitations
The strength of this paper is the large number of PTs that voluntarily participated and collected data on a very large number of participating patients.
Nevertheless, this paper is subject to several limitations. Despite the fact that the infrastructure of the clinical registry has been successfully built up, we identified that input fields from the patient files in the EHR-systems are still registered as missing values, and thus could not be included in this study. These missing values sometimes occurred due to technical omissions (e.g. there are several QBPDS questionnaires defined in the EHR-software but only one of the questionnaires is eligible for extraction to the registry), but also due the submission of incompletely registered patient files from the EHR-system to the national registry. In the further implementation of the clinical registry, we make a continuous effort in trying to decrease the number of missing values. A specific challenge is the large number of different EHR-suppliers that is involved. We have improved the collaboration with the different EHR-system suppliers, which assists in overcoming the technical omissions. In addition, we have executed different projects trying to influence the behaviour of PTs in completing all fields of a patient file before uploading. Examples of such projects are: the organization of several regional symposia to teach PTs on how to use the registry and the development of an online (free for use) feedback dashboard. Such implementation activities are important in the further exploitation of a clinical registry. After all, having a registry with complete patient files is crucial in the development of valid benchmarks for QIs. A second limitation is the chance of selection bias based on the PT practices that participate in the data collection. This is likely to have happened since -in this early stage of implementation- the current users of the clinical registry, the PTs who voluntarily provided all data, are mainly the innovators, early adopters and early majority [58]. These participants are probably PTs who favour the use of PROMs. A third limitation might be the fact that we have used somewhat older data from the registry. Despite this, we assume that the data are still relevant as a good representation of the clinical practice and that they shed light on the potential that a clinical registry has in formulating QIs and their benchmarks. Moreover, the outcomes can be used as baseline measurement to evaluate changes over time.
Implications for future research and clinical practice
Future research should focus on several topics. First, the further development of a standardised practice test, which is an important step before defining QIs and their benchmarks. Such a test should finally include all psychometric properties that need to be tested in defining QIs. These aspects are: reliability, validity, discriminative ability, responsiveness to change, feasibility and usability [2, 28]. The development of such a practice test will likely help with the implementation of QIs in clinical practice and will minimize the resistance in the use of QI and their benchmark. Second, research should focus on the development of actual benchmark based on the collected data thus far. This means that an extensive developmental process must be completed before a QI, including a psychometrically sound benchmark is finished. Given the investment needed it does not seem feasible to develop QIs for all conditions in the field of PT. Therefore it is relevant to investigate what solution can be found for this problem. A possible solution might be the development of several generic QIs for the field of PT and only develop specific QIs for conditions that are seen on a very regular basis by all PTs, as is the case in patients with NSLBP.
In addition, from a clinical practice perspective, efforts should be made to assist PTs in completing their patient files (e.g. by using modern technology such as applications for mobile devices which enable patients to fill in PROMs on their mobile phones) and to increase the number of PTs who will provide their data to the clinical registry. Finally, continuous investments must be made in improving the infrastructure of the clinical registry, on the one hand to detect errors (e.g. missing values) and on the other to improve efficiency and ease of use.
Conclusions
This study showed that the health outcomes of patients with NSLBP collected with PROMs in the national clinical registry of KNGF are reliable, valid and are able to discriminate between primary care PT practices.
Supporting information
S1 Appendix. The calculation used to define a more specific sample size, using the equation of Twisk et al. (2013).
https://doi.org/10.1371/journal.pone.0251892.s001
(DOCX)
S1 Fig. A flowchart of the patient inclusion.
https://doi.org/10.1371/journal.pone.0251892.s002
(DOCX)
S1 Syntax. 20180309_cleaning_definitief_GUUS_JC.sps.
https://doi.org/10.1371/journal.pone.0251892.s003
(SPS)
S2 Syntax. 20201603 descriptives and provisionary analysis and QI analysis.sps.
https://doi.org/10.1371/journal.pone.0251892.s004
(SPS)
S1 Dataset. 20180116_Totaalbestand_aanlever.sav.
https://doi.org/10.1371/journal.pone.0251892.s005
(ZIP)
S3 Dataset. werkbestand_20180308_DCSPH_leeftijd.sav.
https://doi.org/10.1371/journal.pone.0251892.s007
(ZIP)
S4 Dataset. werkbestand_20180308_DCSPH_leeftijd_geslacht.sav.
https://doi.org/10.1371/journal.pone.0251892.s008
(ZIP)
S5 Dataset. werkbestand_20180308_DCSPH_leeftijd_geslacht_afgesloten.sav.
https://doi.org/10.1371/journal.pone.0251892.s009
(ZIP)
S6 Dataset. werkbestand_20180308_DCSPH_leeftijd_geslacht_afgesloten_duurklachten.sav.
https://doi.org/10.1371/journal.pone.0251892.s010
(ZIP)
S8 Dataset. One sample T-test Nivel data vs LDF data.sav.
https://doi.org/10.1371/journal.pone.0251892.s012
(ZIP)
Acknowledgments
We would like to acknowledge all patients, physical therapists and practices who voluntarily collected the valuable data into KNGF’s clinical registry over the past years. In addition, we would like to thank our research colleagues Marjo Maas, Janine Liefers and Annick Bakker-Jacobs for all the time they have invested over the past years in training PTs on how to work with the registry and in their role in analysing the data which results in valuable information for the academic, the clinicians and ultimately for the patient.
References
- 1. Karthaus EG, Vahl A, Kuhrij LS, Elsman BHP, Geelkerken RH, Wouters MWJM, et al. The Dutch Audit of Carotid Interventions: Transparency in Quality of Carotid Endarterectomy in Symptomatic Patients in the Netherlands. Eur J Vasc Endovasc Surg. 2018;56(4):476–485. pmid:30077438
- 2.
Fischer C. Quality Indicators for Hospital Care. (Theunisse I, ed.). Ede: GVO drukkers & vormgevers B.V; 2015.
- 3. Brand CA, Martin-Khan M, Wright O, Jones RN, Morris JN, Travers CM, et al. Development of quality indicators for monitoring outcomes of frail elderly hospitalised in acute care health settings: Study Protocol. BMC Health Serv Res. 2011;11(1):281. pmid:22014061
- 4. Kuhrij LS, Wouters MWJM, van den Berg-Vos RM, de Leeuw FE, Nederkoorn PJ. The Dutch Acute Stroke Audit: Benchmarking acute stroke care in the Netherlands. Eur Stroke J. 2018;3(4):361–368. pmid:31236484
- 5. Petersson IF, Strömbeck B, Andersen L, Cimmino M, Greiff R, Loza E, et al. Development of healthcare quality indicators for rheumatoid arthritis in Europe: The eumusc.net project. Ann Rheum Dis. 2014;73(5):906–908. pmid:23960093
- 6. Govaert JA, Dijk WA Van, Fiocco M, Scheffer AC. Nationwide Outcomes Measurement in Colorectal Cancer Surgery: Improving Quality and Reducing Costs. J Am Coll Surg. 2016;222(1):19–29.e2. pmid:26721750
- 7. Overveld van L. Quality indicators of integrated care for patients with head and neck cancer. Clin Otolaryngol. 2017;42:322–329. pmid:27537106
- 8. van der Wees P, Nijhuis-van der Sanden M, Ayanian J, Black N, Westert G, Schneider E. Integrating the Use of Patient-Reported Outcomes for Both Clinical Practice and Performance Measurement: Views of Experts from 3 Countries. Milbank Q. 2014;92(4):754–775. pmid:25492603
- 9. Black N. Patient reported outcome measures could help transform healthcare. Bmj. 2013;346(January):2–5. pmid:23358487
- 10. Wu AW, Kharrazi H, Boulware LE, Snyder CF. Measure once, cut twice—Adding patient-reported outcome measures to the electronic health record for comparative effectiveness research. J Clin Epidemiol. 2013;66(SUPPL.8):S12–S20. pmid:23849145
- 11. Mainz J. Defining and classifying clinical indicators for quality improvement. Int J Qual Heal Care. 2003;15(6):523–530. pmid:14660535
- 12. Westby MD, Klemm A, Li LC, Jones CA. Emerging Role of Quality Indicators in Physical Therapist Practice and Health Service Delivery. Phys Ther. 2016;96(1):90–100. pmid:26089040
- 13. McGlynn EA, Asch SM. Developing a clinical performance measure. Am J Prev Med. 1998;14(3 SUPPL.):14–21. pmid:9566932
- 14.
Centers for Medicare and Medicaid Services. https://www.cms.gov/Medicare/Quality-Initiatives-Patient-Assessment-Instruments/QualityMeasures/index.html. Published 2019. Accessed January 16th 2021
- 15. Mainz JAN. Developing evidence-based clinical indicators: a state of the art methods primer. Int J Qual Heal Care. 2003;15(supplement I):5–12. pmid:14660518
- 16. Meerhoff G, van Dulmen S, Maas M, Heijblom K, Nijhuis-van der Sanden M, van der Wees P. Development and Evaluation of an Implementation Strategy for Collecting Data in a National Registry and the Use of Patient Reported Outcome Measure (PROMs) in Physical Therapist Practice: Quality Improvement Study. Phys Ther J Am Phys Ther Assoc. 2017;97(8):837–851. pmid:28789466
- 17. Kötter T, Blozik E, Scherer M. Methods for the guideline-based development of quality indicators-a systematic review. Implement Sci. 2012;7(1):21. pmid:22436067
- 18. Nothacker M, Stokes T, Shaw B, Lindsay P, Sipilä R, Follmann M, et al. Reporting standards for guideline-based performance measures. Implement Sci. 2016;11(6):1–11. pmid:26772173
- 19. Peter WF, Hurkmans EJ, van der Wees PJ, Hendriks EJM, van Bodegom-Vos L, Vliet Vlieland TPM. Healthcare Quality Indicators for Physiotherapy Management in Hip and Knee Osteoarthritis and Rheumatoid Arthritis: A Delphi Study. Musculoskeletal Care. 2016;14(4):219–232. pmid:26799718
- 20. Langendam MW, Piggott T, Nothacker M, Agarwal A, Armstrong D, Baldeh T, et al. Approaches of integrating the development of guidelines and quality indicators: a systematic review. 2020:1–11. pmid:32938461
- 21. van Doormaal MCM, Meerhoff GA, Vliet Vlieland TPM, Peter WF.clinical practice guideline for physical therapy in patients with hip or knee osteoarthritis. Musculoskeletal Care. 2020;18(4):575–595. pmid:32643252
- 22. Bier J, Scolten-Peeters W, Staal J, Pool J, van Tulder M, Beekman E, et al. Clinical practice guideline for physical therapy assessment and treatment in patients with nonspecific neck pain. Phys Ther J Am Phys Ther Assoc. 2018;98(3):162–171. pmid:22954069
- 23. Delitto A, George SZ, Dillen LVAN, Whitman JM, Sowa G, Shekelle P, et al. Clinical Practice Guidelines Linked to the International Classification of Functioning, Disability, and Health from the Orthopaedic Section. J Orthop Sport Phys Ther. 2012;42(4):A1–A57. pmid:22466247
- 24. Heemskerk M, Staal J, Bierma-Zeinstra S, de Haan G, Hagenaars L, Lanser K, et al. KNGF guideline Complaints of Arm Neck and Shoulder. Ned Tijdschr voor Fysiother. 2010;120(1 (supplement)).
- 25. Staal J, Hendriks E, Heijmans M, Kiers H, Lutgers-Boomsma AM. Rutten G, et al. KNGF guideline Low Back Pain. Ned Tijdschr voor Fysiother. 2013;123(1 (supplement)).
- 26. Berwick D, Black N, Cullen D, Deerberg-Wittram J, Degos L, Diverty B, et al. Recommendations to OECD Ministers of Health from the High Level Reflection Group on the Future of Healh Statistics.; 2017. https://www.oecd.org/els/health-systems/Recommendations-from-high-level-reflection-group-on-the-future-of-health-statistics.pdf. Accessed January 16th 2021.
- 27. Hendriks H, Bekkering G, van Ettekoven H, Brandsma J, van der Wees P, de Bie R. Development and Implementation of National Practice Guidelines: A Prospect for Continuous Quality Improvement in Physiotherapy. Physiotherapy. 2000;86(10):535–547.
- 28. van Kessel P, Triemstra M, de Boer D. Handreiking Voor Het Meten van Kwaliteit van Zorg Met Patient Reported Outcome Measures.; 2014. https://www.nivel.nl/nl/publicatie/handreiking-voor-het-meten-van-kwaliteit-van-zorg-met-patient-reported-outcome-measures. Accessed January 16th 2021.
- 29. Costa L da CM, Maher CG, Hancock MJ, McAuley JH, Herbert RD, Costa LOP. The prognosis of acute and persistent low-back pain: a meta-analysis. Can Med Assoc J. 2012;184(11):1229–1230. pmid:22586338
- 30. Kopec J, Esdaile J, Abrahamowicz M, Abenhaim L, Wood-Dauphinee S, Lamping D, et al. The Quebec Back Pain Disability Scale Measurement properties. Spine (Phila Pa 1976). 1995;20(3):341–352. pmid:7732471
- 31. Kopec J, Esdaile J, Abrahamowicz M, Abenhaim L, Wood-Dauphinee S, Lamping D, et al. The quebec back pain disability scale measurement properties. J Clin Epidemiol. 1996;49(2):151–161. pmid:8606316
- 32. Speksnijder C, Koppenaal T, Knottnerus J, Spigt M, Staal J, Terwee C. Measurement Properties of the Quebec Back Pain Disability Scale in Patients With Nonspecific Low Back Pain: Systematic Review. Phys Ther J Am Phys Ther Assoc. 2016;96(11):1816–1831. pmid:27231271
- 33. Ostelo R, Deyo R, Strattford P. Interpreting change scores in pain and functional status in low back pain. Spine (Phila Pa 1976). 2008;33(1):90–94. http://scholar.google.com/scholar?hl=en&btnG=Search&q=intitle:Interpreting+Change+Scores+for+Pain+and+Functional+Status+in+Low+Back+Pain#5. Accessed January 16th 2021.
- 34. Kahl C, Cleland J. Visual analogue scale, numeric pain rating scale and the McGill pain Questionnaire: an overview of psychometric properties. Phys Ther Rev. 2005;10(2):123–128.
- 35. Williams A, Davies H, Chadury Y. Simple pain rating scales hide complex idiosyncratic meanings. Pain. 2000;85(3):457–463. pmid:10781919
- 36. Jensen M, Karoly P, Braver S. The measurement of clinical pain intensity: a comparison of six methods. Pain. 1986;27(1):117–126. pmid:3785962
- 37. Childs JD, Piva SR, Fritz JM. Responsiveness of the numeric pain rating scale in patients with low back pain. Spine (Phila Pa 1976). 2005;30(11):1331–1334. pmid:15928561
- 38. Pengel L, Refshauge K, Haher C. Responsiveness of pain, disability, and physical impairment outcomes in patients with low back pain. Spine (Phila Pa 1976). 2004;29(8):879–883. pmid:15082988
- 39. Stratford P, Gill C, Westaway M, Binkley J. Assessing disability and change on individual patients: a report of a patient specific measure. Physiother Canada. 1995;47(4):258–263.
- 40. Stevens A, Köke A, Van Der Weijden T, Beurskens A. Ready for goal setting? Process evaluation of a patient-specific goal-setting method in physiotherapy. BMC Health Serv Res. 2017;17(1):1–10. pmid:28049468
- 41. Westaway MD, Stratford PW, Binkley JM. The Patient-Specific Functional Scale: Validation of’lts Use in Persons With Neck Dysfunction. J Orthop Sport Phys Ther. 1998;27(5):331–338. pmid:9580892
- 42. Oostendorp RA, Rutten GM, Dommerholt J, Nijhuis-van der Sanden M, Harting J. Guideline-based development and practice test of quality indicators for physiotherapy care in patients with neck pain. J Eval Clin Pract. 2013;19:1044–1053. pmid:23510397
- 43. Maas CJM, Hox JJ. Robustness issues in multilevel regression analysis. Stat Neerl. 2004;58(2):127–137.
- 44. Perreault K, Dionne CE. Patient-physiotherapist agreement in low back pain. J Pain. 2005;6(12):817–828. pmid:16326370
- 45.
Twisk J. Applied Multilevel Analysis. 6th ed. Cambridge: Cambridge University Press; 2013.
- 46. Deutscher D, Horn SD, Dickstein R, Hart DL, Smout RJ, Gutvirtz M, et al. Associations Between Treatment Processes, Patient Characteristics, and Outcomes in Outpatient Physical Therapy Practice. Arch Phys Med Rehabil. 2009;90(8):1349–1363. pmid:19651269
- 47. van den Dool J, Schermer T. Zorg Door de Fysiotherapeut Jaarcijfers 2018 En Trendcijfers 2014–2018.; 2019. https://nivel.nl/sites/default/files/bestanden/1003595.pdf. Accessed January 16th 2021.
- 48. Rodrigues IB, Adachi JD, Beattie KA, Lau A, MacDermid JC. Determining known-group validity and test-retest reliability in the PEQ (personalized exercise questionnaire). BMC Musculoskelet Disord. 2019;20(1):1–10. pmid:30611236
- 49. McConnell S, Kolopack P, Davis AM. The Western Ontario and McMaster Universities Osteoarthritis Index (WOMAC): a review of its utility and measurement properties. Arthritis Rheum. 2001;45(5):453–461. pmid:11642645
- 50. Boer de D, Triemstra M, Koopman L, Zuidgeest M. Tool Analyses: Voor Het Vergelijken van Patiëntervaringen Tussen Zorgaanbieders | Versie 1.0 2016. https://www.nivel.nl/nl/publicatie/tool-analyses-voor-het-vergelijken-van-patientervaringen-tussen-zorgaanbieders-versie-10. Accessed January 16th 2021.
- 51. Twisk JWR. Applied Mixed Model Analysis: A Practical Guide.; 2019.
- 52.
Field A. Discovering Statistics Using IBM SPSS Statistics. 4th ed. Sage Publications Ltd.; 2013.
- 53.
https://dica.nl/. https://www.dica.nl. Published 2020. Accessed November 8, 2020.
- 54. Greenhalgh J, Dalkin S, Gooding K, et al. Functionality and Feedback: A Realist Synthesis of the Collation, Interpretation and Utilisation of Patient-Reported Outcome Measures Data to Improve Patient Care. Vol 5.; 2017. pmid:28121094
- 55. Fischer C, Anema HA, Klazinga NS. The validity of indicators for assessing quality of care: A review of the European literature on hospital readmission rate. Eur J Public Health. 2012;22(4):484–491. pmid:22140251
- 56. van der Wees PJ, Verkerk EW, Verbiest MEA, Zuidgeest M, Bakker C, Braspenning J, et al. Development of a framework with tools to support the selection and implementation of patient-reported outcome measures. J Patient-Reported Outcomes. 2019;3(1). pmid:31889232
- 57. Boulkedid R, Abdoul H, Loustau M, Sibony O, Alberti C. Using and reporting the Delphi method for selecting healthcare quality indicators: A systematic review. PLoS One. 2011;6(6). pmid:21694759
- 58.
Rogers E. Diffusion of Innovations Fifth Edition. New York: Free Press; 2003.