Skip to main content
Advertisement
Browse Subject Areas
?

Click through the PLOS taxonomy to find articles in your field.

For more information about PLOS Subject Areas, click here.

  • Loading metrics

A time-adjusted control chart for monitoring surgical outcome variations

  • Quentin Cordier ,

    Roles Conceptualization, Formal analysis, Investigation, Methodology, Software, Validation, Visualization, Writing – original draft, Writing – review & editing

    quentin.cordier@chu-lyon.fr

    Affiliations Research on Healthcare Performance RESHAPE, INSERM U1290, Université Claude Bernard Lyon 1, Lyon, France, Health Data Department, Hospices Civils de Lyon, Lyon, France

  • My-Anh Le Thien,

    Roles Conceptualization, Investigation, Methodology, Writing – review & editing

    Affiliation Health Data Department, Hospices Civils de Lyon, Lyon, France

  • Stéphanie Polazzi,

    Roles Conceptualization, Data curation, Methodology, Validation, Writing – review & editing

    Affiliations Research on Healthcare Performance RESHAPE, INSERM U1290, Université Claude Bernard Lyon 1, Lyon, France, Health Data Department, Hospices Civils de Lyon, Lyon, France

  • François Chollet,

    Roles Validation, Writing – review & editing

    Affiliation Health Data Department, Hospices Civils de Lyon, Lyon, France

  • Matthew J. Carty,

    Roles Conceptualization, Writing – review & editing

    Affiliation Center for Surgery and Public Health, Brigham and Women’s Hospital, Harvard Medical School, Boston, Massachusetts, United States of America

  • Jean-Christophe Lifante,

    Roles Methodology, Supervision, Writing – review & editing

    Affiliations Research on Healthcare Performance RESHAPE, INSERM U1290, Université Claude Bernard Lyon 1, Lyon, France, Hospices Civils de Lyon, Centre Hospitalier Lyon Sud, Service de Chirurgie Générale et Endocrinienne, Pierre Bénite, France

  • Antoine Duclos

    Roles Conceptualization, Data curation, Formal analysis, Investigation, Methodology, Project administration, Software, Supervision, Validation, Visualization, Writing – original draft, Writing – review & editing

    Affiliations Research on Healthcare Performance RESHAPE, INSERM U1290, Université Claude Bernard Lyon 1, Lyon, France, Health Data Department, Hospices Civils de Lyon, Lyon, France, Center for Surgery and Public Health, Brigham and Women’s Hospital, Harvard Medical School, Boston, Massachusetts, United States of America

Abstract

Background

Statistical Process Control (SPC) tools providing feedback to surgical teams can improve patient outcomes over time. However, the quality of routinely available hospital data used to build these tools does not permit full capture of the influence of patient case-mix. We aimed to demonstrate the value of considering time-related variables in addition to patient case-mix for detection of special cause variations when monitoring surgical outcomes with control charts.

Methods

A retrospective analysis from the French nationwide hospital database of 151,588 patients aged 18 and older admitted for colorectal surgery between January 1st, 2014, and December 31st, 2018. GEE multilevel logistic regression models were fitted from the training dataset to predict surgical outcomes (in-patient mortality, intensive care stay and reoperation within 30-day of procedure) and applied on the testing dataset to build control charts. Surgical outcomes were adjusted on patient case-mix only for the classical chart, and additionally on secular (yearly) and seasonal (quarterly) trends for the enhanced control chart. The detection of special cause variations was compared between those charts using the Cohen’s Kappa agreement statistic, as well as sensitivity and positive predictive value with the enhanced chart as the reference.

Results

Within the 5-years monitoring period, 18.9% (28/148) of hospitals detected at least one special cause variation using the classical chart and 19.6% (29/148) using the enhanced chart. 59 special cause variations were detected overall, among which 19 (32.2%) discordances were observed between classical and enhanced charts. The observed Kappa agreement between those charts was 0.89 (95% Confidence Interval [95% CI], 0.78 to 1.00) for detecting mortality variations, 0.83 (95% CI, 0.70 to 0.96) for intensive care stay and 0.67 (95% CI, 0.46 to 0.87) for reoperation. Depending on surgical outcomes, the sensitivity of classical versus enhanced charts in detecting special causes variations ranged from 0.75 to 0.89 and the positive predictive value from 0.60 to 0.89.

Conclusion

Seasonal and secular trends can be controlled as potential confounders to improve signal detection in surgical outcomes monitoring over time.

Background

Adverse events following surgery remain frequent, resulting in one out of ten patients being exposed to severe and preventable harm [1]. As a solution, surgical outcomes monitoring over time has proven impactful for reducing patient mortality and morbidity [2]. To avoid misinterpretation in assessing outcomes variation prospectively, indicators need to be adjusted for patient case-mix. However, data granularity extracted from available data sources does not fully capture every patient nuance or surgical procedure complexity. Hospital data warehouse accuracy is also uncertain because information is not primarily collected for quality improvement purposes [3]. Data validity can be heterogeneous across hospitals, resulting in flawed interpretation of their performance [4, 5].

To deal with these methodological challenges, considering time-related proxies such as secular and seasonal trends could be of particular interest when tracking surgical safety. In past control chart literature, these variables were not identified as critical for surgical outcomes adjustment [6]. Secular trends reflect enhancing or deteriorating outcomes over years nationwide under the influence of a myriad of drivers [7]. Over seasons, patient case-mix also fluctuates with a marked volume reduction in elective procedures and more complex cases performed during the summer holidays [8, 9]. Therefore, integrating annual secular trends and quarterly seasonal variations in the adjustment scheme may indirectly control variations related to both patient severity and data validity that may influence outcome irrespective of the delivered surgical quality [10].

The Shewhart control chart is increasingly used for healthcare improvement, providing a visual presentation of data easily interpretable by healthcare professionals. This decision support tool plots successive indicator measurements in chronological order, with control limits demarcating their expected variation [11]. Validated through a century of usage in industry, previous experience suggested its transferability to improve healthcare quality [12]. We assumed that adjusting control charts for time-related variables would change the interpretation of indicator variations and refine the detection of outliers in surgical outcomes. The present study aimed to compare the detection of special cause variations in colorectal surgery using an enhanced time-adjusted control chart adjusted for secular and seasonal trends in addition to the patient case-mix versus a classical control chart accounting for patient case-mix only.

Methods

Study population, outcomes and design

We included all stays for patients aged 18 and older admitted to French public or private hospitals for colorectal surgery between January 1st, 2014, and December 31st, 2018. Inpatient stays for palliative care or organ retrieval were excluded, as well as hospitals providing colorectal surgery discontinuously (less than 1 stay per quarter over 5 years), with low volume of stays for colorectal surgery (less than 100 stays per year over 5 years), with important variations in their annual volume of stays (variation greater than 50%), or without at least one occurrence of each monitored outcome per year, allowing us to build charts with sufficient volume of procedures per quarter in each hospitals along the 5 years study. Considering their low frequency, patients with missing data were also excluded (Fig 1).

The three monitored outcomes on control charts were major adverse events occurring within 30 days from the colorectal surgical procedure, categorized according to the Dindo-Clavien classification, which ranks a surgical complication in an objective and reproducible manner [13]. Those outcomes included inpatient death, intensive care stay (at least two nights in intensive care unit or five nights in intermediate care unit), and reoperation (open or laparoscopic digestive tract procedure).

To fit models for outcome adjustment independently from the data to be monitored on control charts, we randomly divided our dataset into a training set and a testing set with stratification by hospital and the period for every quarter per year [14, 15]. The training set accounted for 20% of the data (30,630 stays) and the testing set for the 80% remaining (120,958 stays). The choice of 20/80 splitting implied that the study sample was large enough to build reliable models using the training set first, then allowing us to use a majority share of the data to build the charts from the testing set. Classical control charts were adjusted for patient case-mix only, while enhanced control charts were adjusted with the same set of variables and the additional input of secular (yearly) and seasonal (quarterly) trends. Finally, the detection of quarterly special cause variation signals between those charts was compared.

Data source

Anonymised data were obtained from the French Medical Information System (Programme de Médicalisation des Systèmes d’Information (PMSI), source: ATIH [16]). This nationwide database is used routinely for care reimbursement and is updated weekly with data collected prospectively from all hospitals in France. Due to the accuracy and exhaustive data collection of the PMSI database, no patients were lost to follow-up during the study period. Inpatient stays were recorded as standard discharge abstracts containing compulsory information about patients and their primary or secondary diagnoses using ICD-10 (International Classification of Diseases, 10th revision) codes as well as detailed procedural codes associated with the care provided. Patients’ socioeconomic status (median household income in the city of residence) and medical accessibility (mean number of family medicine consultations per year and per inhabitant in the city of residence) were retrieved from their city of residence postcode. The study was conducted in accordance with the guidelines of the Declaration of Helsinki and recorded by the French National Data Protection Commission (CNIL, MR005 N° MR0612180919). It was strictly observational, relying on anonymous data available to investigators through the Secured Data Access Platform of the French Technical Agency of Hospital Information under formal authorization and a secure process. Consequently, according to French law, the need for written consent was waived by the CNIL, and no institutional review board approval was necessary.

Statistics and charts

Multilevel logistic regression models were fitted from the training set using Generalized Estimating Equations (GEEs, see S1 Supporting information in S1 File) [17] for parameters associated with the three outcomes, with an exchangeable working correlation structure to account for clustering of patients within each hospital [18]. Classical control charts models were adjusted for case-mix variables only including: age, gender, patients’ socioeconomic status, medical accessibility, emergency admission, hospital status, surgical procedure complexity (segmental colectomy, total colectomy, coloproctectomy, rectopexy, rectal surgery, ostomy and Hartmann procedure reversal), primary diagnosis (cancer, diverticulosis, inflammatory bowel disease, bowel occlusion, ostomy surveillance, other surgical indication) and comorbidities (in dummy variables) as determined by the Elixhauser score [19]. Enhanced case-mix and time-adjusted control charts models were adjusted with the same set of variables, in addition to the year as a proxy for secular trends and the quarter as a proxy of seasonal variations. Model calibration was assessed by plotting the observed frequency of events per tenths of predicted risk [20], and model discrimination was evaluated using c-statistics (or area under the curve AUC).

We retained the regression estimates and the intraclass correlation coefficient (ICC) to determine the design effect of each hospital and reflect the inflation in variance due to the clustering of patients within the same hospital. Then, using regression estimates obtained from the training dataset, we computed for each operated patient of the testing dataset the expected probability of each outcome.

Shewhart p-control charts were designed to monitor outcomes within each hospital using the testing dataset. A control chart is a graphical tool used to monitor the evolution of a statistical process compared to a central line, upper and lower control limits. It enables its user to determine whether the observed value falls within the control limits, indicating that the process is under control. Conversely, if the observed value is beyond the control limits, the process is considered out of control, signaling the presence of a special cause variation [21]. As our outcomes were binary, we have opted for an adapted form of control charts specifically designed for binary indicators, the Shewhart p-control chart. Each data point on the chart represented the crude proportion of the observed outcome per quarter, accounting for a total of 20 points over the five years monitoring period. The central line value of the charts was not constant and represented the expected proportion of complications per quarter calculated from the models initially fitted on the training dataset and then applied to the testing dataset. Exact control and warning limits were set at 3 SD (99.73%) and 2 SD (95.45%) from the central line respectively using the exact binomial distribution [11].

We additionally corrected these limits through the intra-class correlation coefficient to account for the inflation of variance due to the pooling of patients within a same hospital [18]. Finally, in order to avoid inter-hospital hospital benchmarking and allow the accurate detection of special cause variations within the same hospital, we recalibrated the central line and limits depending on the overall observed rates per hospital. In cases where calculated values were beyond 1 or below 0 (values that are conceptually impossible), such limits were reset to 1 and 0, respectively. Formulas used to build the charts are available in S2 Supporting information in S1 File). The detection of special cause variation signals was defined as a single point outside the 3 SD control limits or two out of three successive points between the 2 SD warning limit and the 3 SD control limit on the same side of the central line. Since the risk of detecting false-positive special cause variation increases as the number of detection increases, we restricted our detection methodology to these two rules, corresponding to an overall 1/225 (0.444%) risk of false positive per plotted value [22], and an 1-(1–0.444)20 = 8.5% risk of detecting at least one false-positive signal in a 20-points control chart [23].

Following the chart construction, an agreement table for the detection of special cause variations was set up between the classical control charts and the enhanced control charts for each monitored outcome among all hospitals (148 hospitals x 20 quarters = 2,960 points in total). Agreement measure was computed using Cohen’s Kappa statistic [24] with corresponding 95% CI. Agreement in signals detection of special cause variations was considered acceptable above 0.60 and good above 0.80 [25].

Finally, considering the enhanced control chart as the reference, we computed sensitivity and positive predictive values of the classical control chart to detect valid special cause variation.

Data manipulation and analyses were performed using SAS software (version 9.4; SAS Institute Inc., Cary, NC).

Results

A total of 151,588 inpatient stays for colorectal surgery performed in 148 hospitals were considered in our analyses. Inpatient stay characteristics per quarter are described in Table 1, revealing higher risk and poorer outcomes among patients who underwent surgery during the third quarter. Patients treated during this period were significantly older (mean age of 65.4 versus 64.9 for the whole year) with a greater share of males (50.3% versus 48.8%), higher Elixhauser comorbidity scores (mean score of 1.6 versus 1.5) than patients treated in other quarters. They were also more often admitted emergently (14.4% versus 13.6%) with cancer diagnoses (55.9% versus 52.3%). Among the monitored adverse events, inpatient mortality was significantly higher during the third quarter (4.2% versus 3.9% for the whole year), as were rates of intensive care stay (9.9% versus 9.2%) and reoperation (9.4% versus 8.9%).

Models used to build the classical and the enhanced charts had discrimination (using C-statistics) ranging from 0.904 (95% Confidence Interval [95% CI], 0.896 to 0.911) for mortality to 0.812 (95% CI, 0.804 to 0.820) for intensive care stay and 0.715 (95% CI, 0.705 to 0.725) for reoperation. Model calibration was excellent, with a calibration line close to the diagonal for all models (see S3 Supporting information in S1 File). Fig 2 presents nationwide variations of surgical outcomes with iterative peak of expected deaths, intensive care stays and reoperations over the third quarter every year. The enhanced control charts considered those peaks in adjustment, providing a different interpretation of special cause variations compared to the classical control charts adjusted for patient case-mix only. Examples of classical and enhanced control charts discrepancies are displayed for two hospitals in Fig 3. In both cases the classical control charts did not detect special cause variations related to deterioration or improvement in surgical outcomes compared to the enhanced control charts.

thumbnail
Fig 2. Histograms of observed outcomes and the related curves of expected trends with classical or enhanced adjustments.

Expected rates of complications per quarter were calculated from the GEEs models initially fitted on the training dataset and then applied to the testing dataset. Models used for the construction of classical control charts were adjusted for case-mix variables only (age, gender, socioeconomic status, medical accessibility, emergency admission, hospital status, surgical procedure complexity, primary diagnosis, and comorbidities in dummies from the Elixhauser score). Models used for the construction of enhanced case-mix and time adjusted control charts were adjusted with the same set of variables, in addition to the year as a proxy for secular trends and the quarter as a proxy of seasonal variations.

https://doi.org/10.1371/journal.pone.0303543.g002

thumbnail
Fig 3. Example of classical and enhanced control charts for two hospitals.

Crude observed rates (dotted black line) were monitored over 20 quarters for all three surgical outcomes. 2-SD warning limits (light green/red lines) and 3-SD control limits (bold green/red lines) were based on the central line (blue line) computed through the GEE models. A special cause variation related to a deterioration of surgical outcomes was detected in case of one single point beyond the 3-SD upper control limit (3-SD UCL), or 2 out of 3 consecutive points beyond the 2-SD upper warning limit (2-SD UWL). Conversely, a special cause variation related to an improvement of surgical outcomes was detected in case of one single point below the 3-SD lower control limit (3-SD LCL), or 2 out of 3 consecutive points below the 2-SD lower warning limit (2-SD LWL). The signal detection was considered at the first point beyond the limit when using the 2 out of 3 consecutive points rule. The two selected hospitals demonstrated discordances (encircled in pink) in interpretation of surgical outcome variations between classical and enhanced charts. Hospital A detected a special cause variation of increased mortality during the second quarter of 2018 using the time-adjusted chart (1 point above the 3-SD upper control limit, chart A2) but not the case-mix only adjusted chart (chart A1). Similarly, hospital B detected a special cause variation of decreased reoperation rate during the second quarter of 2015 using the time-adjusted chart (2 points out of 3 below the 2-SD lower warning limit, chart B6) but not the classical one (only 1 point out of 3 below the 2-SD warning limit, chart B5).

https://doi.org/10.1371/journal.pone.0303543.g003

Overall, among 59 signals detection of special cause variations, 19 discordances (32.2%) were observed between classical and enhanced charts, including 4 discordances for mortality, 6 for intensive care stay and 9 for reoperation. These discordances were mostly observed in the second and third quarters (6 discordances each), followed by the fourth (4) and the first quarters (3). There were 2 discordances in 2014, 6 in 2015, 2 in 2016, 4 in 2017, and 5 in 2018.

Using the classical chart, 9.5% (14/148) of hospitals detected at least one special cause variation for mortality along the 5-years monitoring period, 4.7% (7/148) for intensive care stay and 7.4% (11/148) for reoperation. Corresponding numbers with the enhanced chart were 9.5% (14/148), 4.7% (7/148) and 8.1% (12/148), respectively. Observed kappa agreement between classical and enhanced charts for detecting special cause variation in Table 2 was 0.89 for mortality (95% CI, 0.78 to 1.00), 0.83 for intensive care stay (95% CI, 0.70 to 0.96), and 0.67 for reoperation (95% CI, 0.46 to 0.87). Considering the enhanced time-adjusted chart as the reference, sensitivity and positive predictive values of the classical chart to detect special causes variations were both 0.89 (95% CI, 0.65 to 0.99) for inpatient mortality and 0.83 (95% CI, 0.59 to 0.96) for intensive care stay, and respectively 0.75 (95% CI, 0.43 to 0.95) and 0.60 (95% CI, 0.32 to 0.84) for reoperations.

thumbnail
Table 2. Comparison of special cause variations detection between classical and enhanced control charts.

https://doi.org/10.1371/journal.pone.0303543.t002

Discussion

Main findings

Because variations in patient case-mix and data validity over time may bias interpretation of continuous surgical outcomes monitoring, we developed a control chart adjusted for secular and seasonal trends. Signal detection of special cause variations based on this enhanced chart was compared against a classical control chart that only considered patient case-mix. Although the number of special cause variations detecting better or worse surgical outcomes were similar for the classical and the enhanced control charts, a third of special cause variations were discordant. Indeed, a significant number of special cause variations were detected, whether rightly or wrongly, by one chart but not the other one. Under the assumption that the enhanced chart might have outperformed the classical chart since it considered more parameters, it would have detected not only more true positive signals but also less false positives. Omitting time-related proxies in adjustment schemes could lead to consider common cause as special cause variations, hence potentially undertaking inadequate interventions. Conversely, it could also lead to consider special cause as common cause variations, hence missing opportunities to improve care.

Comparison with other studies

Statistical process control (SPC) tools such as control charts have already been applied to surgery and have proven to be useful in enhancing patient safety [2]. Because the mix of every patient and surgical procedure is different, adjusting SPC tools has been proposed to allow a more accurate interpretation of surgical outcome variations [26, 27]. Numerous studies have shown clear advantages in adjusting models comparatively to non-risk-adjusted tools [28] using validation methodologies similar to ours [25, 26]. A systematic review found that half of studies using control charts in surgery considered outcomes adjustment [6]. However, adjustment is complex and requires caution with the employed methodology and data [2931]. Controversy exists over whether administrative data granularity and validity are sufficient to control potential confounders accurately [32]. Key variables might be unavailable, missing or inaccurate to measure case-mix variations directly.

Using proxies built from other available variables represents a solution to better capture case-mix indirectly, however. Seasonality and secular trends are frequently considered in the study of interrupted time series, but have not yet been used to build control charts [33]. Seasonality based on quarter as time period measurement can be considered as a useful proxy reflecting unknown case-mix parameters. Past studies have highlighted a greater concentration of severe cases during the summer [34], and deterioration in surgical outcomes due to the postponement of elective care and differences in staffing composition [8, 9, 35]. Secular trends in surgical outcomes over years have also been frequently described [15, 36]. Their adjustment can significantly change results interpretation and revealed crucial to avoid erroneous signal detection not associated with special cause variation within a particular hospital but rather with long-term improvement or deterioration in care delivery nationwide [37, 38].

Strengths and limitations

In the present study, we designed an enhanced control chart accounting for both secular trends and seasonal variations through the integration of the year and quarter of performed surgical procedure in the adjustment scheme. We extracted data from a nationwide inpatient sample, an approach that can be easily reproduced with similar diagnosis related-like data sources available in other countries [39]. By focusing on colorectal surgery, we selected a high-volume surgery with frequent adverse events identifiable from medico-administrative data. We also intended to reduce heterogeneity in patient case-mix and to characterize accurately patient severity and procedure complexity by using specific algorithms. Reproducing our methodology would be relevant for assessing time-adjusted control charts in other healthcare contexts and countries experiencing various patterns of seasonal variations and secular trends. We also ensured that changes measured in care safety reflected only intra-hospital variations and not inter-hospital variations by modelling outcomes using multi-level GEE logistic regression models to capture risks specific to each hospital, considering the hospital status, and recalibrating our control charts.

Several limitations should also be acknowledged. First, we compared signal detection between the two control charts, assuming the time-adjusted chart as the most valid one. However, the true gold standard would have been to investigate systematically every signal detected from both charts within each hospital. The retrospective and nationwide nature of the pursued work did not allow us to pursue this approach. Second, adjusting for seasonal variations categorizes all resulting fluctuations as common cause variations, potentially masking recurrent quality deteriorations during suboptimal conditions such as in the summer. Third, even if the hospitals and patients sample size was reasonably large, signals occurrence was scarce and the related number small to compute statistics. Proposed metrics such as Cohen’s Kappa coefficients, sensitivity or positive predictive values might have been influenced by such small numbers resulting in a limited accuracy. Cohen’s Kappa might also be too optimistic due to the very high numbers in non-detections. We were also exposed to traditional limitations inherent to the use of a hospital administrative database. This may have resulted in case-mix heterogeneity issues due to differences in coding practices between hospitals [40, 41]. We attempted to bypass such potential pitfalls by monitoring surgical outcomes within each hospital instead of benchmarking across hospitals, and by recalibrating our charts based on each hospital’s average performance over the whole study period. In addition, changes in national data coding rules or coding variations that may exist within individual hospitals over time was considered by integrating the year and quarter in adjustment scheme. Finally, the broad diversity in diagnoses present in the study led us to adopt a classification that may have lacked finesse, especially in regards to procedures that involved other surgical specialties.

Policy implications

We built a novel system to monitor surgical outcomes nationwide within every French hospital over time using time-adjusted control charts. This approach goes beyond traditional approaches such as benchmarking, avoiding potentially misleading conclusions [4] and enabling a dynamic follow-up over time. The use of hospital administrative databases that were not initially intended for healthcare quality and safety assessment is controversial. Using a clinical registry such as the ACS-NSQIP would have provided more reliable information on patient comorbidities and outcomes [42, 43]. However, administrative claims data provides nationwide exhaustibility and are easily replicable worldwide at low cost, as clinical registries are more expensive to maintain and only provide a sample of the whole population. Registries also have their limitations since there will always exist unmeasured patient-related variables to influence outcomes independently of the quality of care.

Our study underlined significant discordances in signal detection between the classical case-mix adjusted control chart and the enhanced time-adjusted chart. Not considering seasonal variations and secular trends in outcome adjustment is a concern to allow a fair interpretation of hospital performance. Indeed, biased interpretation of control chart signals can lead to inadequate investigation of special cause variations. Wrongly ignoring a signal (i.e. a false negative) deprives surgical teams of resolving significant safety issue and represents a missing opportunity to improve healthcare delivery. Conversely, wrongly detecting a signal that only reflects random variation (i.e. false positive) will waste the time and energy of teams investigating an event that never happened, and potentially introduce irrelevant changes with undesirable variations in surgical practices. Seasonal variations and secular trends are easily available in medico-administrative data and should be systematically considered in risk adjustment when monitoring surgical outcomes over time to enhance control charts interpretation.

References

  1. 1. Panagioti M, Khan K, Keers RN, Abuzour A, Phipps D, Kontopantelis E, et al. Prevalence, severity, and nature of preventable patient harm across medical care settings: systematic review and meta-analysis. BMJ. 2019 17;366:l4185. pmid:31315828
  2. 2. Duclos A, Chollet F, Pascal L, Ormando H, Carty MJ, Polazzi S, et al. Effect of monitoring surgical outcomes using control charts to reduce major adverse events in patients: cluster randomised trial. BMJ. 2020 Nov 4;371:m3840. pmid:33148601
  3. 3. Lindenauer PK, Lagu T, Shieh MS, Pekow PS, Rothberg MB. Association of diagnostic coding with trends in hospitalizations and mortality of patients with pneumonia, 2003–2009. JAMA. 2012 Apr 4;307(13):1405–13. pmid:22474204
  4. 4. Mohammed MA, Deeks JJ, Girling A, Rudge G, Carmalt M, Stevens AJ, et al. Evidence of methodological bias in hospital standardised mortality ratios: retrospective database study of English hospitals. BMJ. 2009 Mar 18;338:b780. pmid:19297447
  5. 5. Haviari S, Chollet F, Polazzi S, Payet C, Beauveil A, Colin C, et al. Effect of data validation audit on hospital mortality ranking and pay for performance. BMJ Qual Saf. 2019 Jun;28(6):459–67. pmid:30366969
  6. 6. Le Thien MA, Cordier Q, Lifante JC, Carty MJ, Duclos A. Control Charts Usage for Monitoring Performance in Surgery: A Systematic Review. J Patient Saf. 2023 Mar 1;19(2):110–6. pmid:36603595
  7. 7. Vincent C, Aylin P, Franklin BD, Holmes A, Iskander S, Jacklin A, et al. Is health care getting safer? BMJ. 2008 Nov 13;337:a2426. pmid:19008272
  8. 8. Caillet P, Payet C, Polazzi S, Carty MJ, Lifante JC, Duclos A. Increased Mortality for Elective Surgery during Summer Vacation: A Longitudinal Analysis of Nationwide Data. Abe T, editor. PLOS ONE. 2015 Sep 25;10(9):e0137754. pmid:26407191
  9. 9. Spencer E, Berry M, Martin P, Rojas-Garcia A, Moonesinghe SR. Seasonality in surgical outcome data: a systematic review and narrative synthesis. Br J Anaesth. 2022 Feb;128(2):321–32. pmid:34872715
  10. 10. Mant J. Process versus outcome indicators in the assessment of quality of health care. Int J Qual Health Care J Int Soc Qual Health Care. 2001 Dec;13(6):475–80. pmid:11769750
  11. 11. Duclos A, Voirin N. The p-control chart: a tool for care improvement. Int J Qual Health Care J Int Soc Qual Health Care. 2010 Oct;22(5):402–7.
  12. 12. Thor J, Lundberg J, Ask J, Olsson J, Carli C, Härenstam KP, et al. Application of statistical process control in healthcare improvement: systematic review. Qual Saf Health Care. 2007 Oct;16(5):387–99. pmid:17913782
  13. 13. Dindo D, Demartines N, Clavien PA. Classification of surgical complications: a new proposal with evaluation in a cohort of 6336 patients and results of a survey. Ann Surg. 2004 Aug;240(2):205–13. pmid:15273542
  14. 14. Hastie T, Tibshirani R, Friedman J. The Elements of Statistical Learning: Data Mining, Inference, and Prediction, Second Edition [Internet]. 2nd ed. New York: Springer-Verlag; 2009 [cited 2021 Apr 13]. (Springer Series in Statistics). https://www.springer.com/gp/book/9780387848570
  15. 15. Duclos A, Polazzi S, Lipsitz SR, Couray-Targe S, Gawande AA, Colin C, et al. Temporal variation in surgical mortality within French hospitals. Med Care. 2013 Dec;51(12):1085–93. pmid:24231784
  16. 16. Access to ATIH databases [Internet]. https://www.atih.sante.fr/bases-de-donnees/commande-de-bases
  17. 17. Hubbard AE, Ahern J, Fleischer NL, Van der Laan M, Satariano SA, Jewell N, et al. To GEE or Not to GEE: Comparing Population Average and Mixed Models for Estimating the Associations Between Neighborhood Risk Factors and Health. Epidemiology. 2010;21(4):467–74. pmid:20220526
  18. 18. Kish L. Survey Sampling (New York: John Wiley & Sons, Inc., 1965). Am Polit Sci Rev. 1965 Dec;59(4):1025–1025.
  19. 19. Quan H, Sundararajan V, Halfon P, Fong A, Burnand B, Luthi JC, et al. Coding algorithms for defining comorbidities in ICD-9-CM and ICD-10 administrative data. Med Care. 2005 Nov;43(11):1130–9. pmid:16224307
  20. 20. Altman DG, Vergouwe Y, Royston P, Moons KGM. Prognosis and prognostic research: validating a prognostic model. BMJ. 2009 May 28;338:b605. pmid:19477892
  21. 21. Liu R, Lai X, Wang J, Zhu X, Liu Y. SCR-CUSUM: An illness-death semi-Markov model-based risk-adjusted CUSUM for semi-competing risk data monitoring. Comput Ind Eng. 2023 Oct 1;184:109530.
  22. 22. Champ CW, Woodall WH. Exact Results for Shewhart Control Charts With Supplementary Runs Rules. Technometrics. 1987 Nov 1;29(4):393–9.
  23. 23. Benneyan JC, Lloyd RC, Plsek PE. Statistical process control as a tool for research and healthcare improvement. Qual Saf Health Care. 2003 Dec;12(6):458–64. pmid:14645763
  24. 24. Cohen J. A Coefficient of Agreement for Nominal Scales. Educ Psychol Meas. 1960 Apr 1;20(1):37–46.
  25. 25. Altman DG. Practical Statistics for Medical Research [Internet]. [cited 2022 Jun 13]. (Chapman & Hall/CRC). https://www.routledge.com/Practical-Statistics-for-Medical-Research/Altman/p/book/9780412276309
  26. 26. Alemi F, Oliver DW. Tutorial on risk-adjusted P-charts. Qual Manag Health Care. 2001;10(1):1–9. pmid:11702466
  27. 27. Steiner SH, Cook RJ, Farewell VT, Treasure T. Monitoring surgical performance using risk-adjusted cumulative sum charts. Biostat Oxf Engl. 2000 Dec;1(4):441–52. pmid:12933566
  28. 28. Novick RJ, Fox SA, Stitt LW, Forbes TL, Steiner S. Direct comparison of risk-adjusted and non-risk-adjusted CUSUM analyses of coronary artery bypass surgery outcomes. J Thorac Cardiovasc Surg. 2006 Aug;132(2):386–91. pmid:16872967
  29. 29. Marshall T, Mohammed MA. Case-mix and the use of control charts in monitoring mortality rates after coronary artery bypass. BMC Health Serv Res. 2007 Apr 30;7:63. pmid:17470276
  30. 30. Duclos A, Voirin N, Touzet S, Soardo P, Schott AM, Colin C, et al. Crude versus case-mix-adjusted control charts for safety monitoring in thyroid surgery. Qual Saf Health Care. 2010 Dec;19(6):e17. pmid:20378622
  31. 31. Iezzoni LI. The risks of risk adjustment. JAMA. 1997 Nov 19;278(19):1600–7. pmid:9370507
  32. 32. Hannan EL, Kilburn H, Lindsey ML, Lewis R. Clinical versus administrative data bases for CABG surgery. Does it matter? Med Care. 1992 Oct;30(10):892–907. pmid:1405795
  33. 33. Hategeka C, Ruton H, Karamouzian M, Lynd LD, Law MR. Use of interrupted time series methods in the evaluation of health system quality improvement interventions: a methodological systematic review. BMJ Glob Health. 2020 Oct;5(10):e003567. pmid:33055094
  34. 34. Shah PS, Yoon W, Kalapesi Z, Bassil K, Dunn M, Lee SK. Seasonal variations in healthcare-associated infection in neonates in Canada. Arch Dis Child Fetal Neonatal Ed. 2013 Jan;98(1):F65–69. pmid:22556205
  35. 35. Marchegiani G, Andrianello S, Nessi C, Giuliani T, Malleo G, Paiella S, et al. Seasonal variations in pancreatic surgery outcome A retrospective time-trend analysis of 2748 Whipple procedures. Updat Surg. 2020 Sep;72(3):693–700. pmid:32816284
  36. 36. Ko CY, Hall BL, Hart AJ, Cohen ME, Hoyt DB. The American College of Surgeons National Surgical Quality Improvement Program: achieving better and safer surgery. Jt Comm J Qual Patient Saf. 2015 May;41(5):199–204. pmid:25977246
  37. 37. Chen YF, Hemming K, Stevens AJ, Lilford RJ. Secular trends and evaluation of complex interventions: the rising tide phenomenon. BMJ Qual Saf. 2016 May;25(5):303–10. pmid:26442789
  38. 38. Pate A, van Staa T, Emsley R. An assessment of the potential miscalibration of cardiovascular disease risk predictions caused by a secular trend in cardiovascular disease in England. BMC Med Res Methodol. 2020 Nov 30;20(1):289. pmid:33256644
  39. 39. OECD. Health at a Glance 2019: OECD Indicators [Internet]. Paris: Organisation for Economic Co-operation and Development; 2019 [cited 2022 Jun 13]. https://www.oecd-ilibrary.org/social-issues-migration-health/health-at-a-glance-2019_4dd50c09-en
  40. 40. Lilford R, Mohammed MA, Spiegelhalter D, Thomson R. Use and misuse of process and outcome data in managing performance of acute medical care: avoiding institutional stigma. Lancet Lond Engl. 2004 Apr 3;363(9415):1147–54. pmid:15064036
  41. 41. Nicholl J. Case-mix adjustment in non-randomised observational evaluations: the constant risk fallacy. J Epidemiol Community Health. 2007 Nov;61(11):1010–3. pmid:17933961
  42. 42. Shiloach M, Frencher SK, Steeger JE, Rowell KS, Bartzokis K, Tomeh MG, et al. Toward robust information: data quality and inter-rater reliability in the American College of Surgeons National Surgical Quality Improvement Program. J Am Coll Surg. 2010 Jan;210(1):6–16. pmid:20123325
  43. 43. Lawson EH, Louie R, Zingmond DS, Brook RH, Hall BL, Han L, et al. A comparison of clinical registry versus administrative claims data for reporting of 30-day surgical complications. Ann Surg. 2012 Dec;256(6):973–81. pmid:23095667