Skip to main content
Advertisement
Browse Subject Areas
?

Click through the PLOS taxonomy to find articles in your field.

For more information about PLOS Subject Areas, click here.

  • Loading metrics

Performance of cardiopulmonary exercise testing for the prediction of post-operative complications in non cardiopulmonary surgery: A systematic review

  • Daniel J. Stubbs,

    Roles Conceptualization, Data curation, Formal analysis, Methodology, Project administration, Validation, Writing – original draft, Writing – review & editing

    Affiliation University Division of Anaesthesia, Department of Medicine, Addenbrooke’s Hospital, Hills Road, Cambridge, CB2 0QQ, Cambridge, United Kingdom

  • Lisa A. Grimes,

    Roles Data curation, Formal analysis, Project administration, Writing – review & editing

    Affiliation University Division of Anaesthesia, Department of Medicine, Addenbrooke’s Hospital, Hills Road, Cambridge, CB2 0QQ, Cambridge, United Kingdom

  • Ari Ercole

    Roles Conceptualization, Project administration, Supervision, Validation, Writing – review & editing

    ae105@cam.ac.uk

    Affiliation University Division of Anaesthesia, Department of Medicine, Addenbrooke’s Hospital, Hills Road, Cambridge, CB2 0QQ, Cambridge, United Kingdom

Abstract

Introduction

Cardiopulmonary exercise testing (CPET) is widely used within the United Kingdom for preoperative risk stratification. Despite this, CPET’s performance in predicting adverse events has not been systematically evaluated within the framework of classifier performance.

Methods

After prospective registration on PROSPERO (CRD42018095508) we systematically identified studies where CPET was used to aid in the prognostication of mortality, cardiorespiratory complications, and unplanned intensive care unit (ICU) admission in individuals undergoing non-cardiopulmonary surgery. For all included studies we extracted or calculated measures of predictive performance whilst identifying and critiquing predictive models encompassing CPET derived variables.

Results

We identified 36 studies for qualitative review, from 27 of which measures of classifier performance could be calculated. We found studies to be highly heterogeneous in methodology and quality with high potential for bias and confounding. We found seven studies that presented risk prediction models for outcomes of interest. Of these, only four studies outlined a clear process of model development; assessment of discrimination and calibration were performed in only two and only one study undertook internal validation. No scores were externally validated. Systematically identified and calculated measures of test performance for CPET demonstrated mixed performance. Data was most complete for anaerobic threshold (AT) based predictions: calculated sensitivities ranged from 20-100% when used for predicting risk of mortality with high negative predictive values (96-100%). In contrast, positive predictive value (PPV) was poor (2.9-42.1%). PPV appeared to be generally higher for cardiorespiratory complications, with similar sensitivities. Similar patterns were seen for the association of Peak VO2 (sensitivity 85.7-100%, PPV 2.7-5.9%) and VE/VCO2 (Sensitivity 27.8%-100%, PPV 3.4-7.1%) with mortality.

Conclusions

In general CPET’s ‘rule-out’ capability appears better than its ability to ‘rule-in’ complications. Poor PPV may reflect the frequency of complications in studied populations. Our calculated estimates of classifier performance suggest the need for a balanced interpretation of the pros and cons of CPET guided pre-operative risk stratification.

Introduction

There is increasing focus on the identification of the ‘high-risk’ surgical patient. Large scale national projects such as the ‘National Emergency Laparotomy Audit’ (NELA) incorporate this pre-operative assessment of risk into pathways to stratify perioperative care [1]. In the elective setting techniques including risk scales, patient function, organ biomarkers, and levels of comorbidity have all been used with the aim of risk stratifying patients [2]. Risk stratification may facilitate precision interventions to improve outcome in this patient cohort, as well as assist in the process of shared decision making. Cardiopulmonary exercise testing (CPET) is attractive as a basis for patient stratification as it represents an objective and holistic physiological assessment of cardiopulmonary reserve. Briefly, an individual exercises (typically on a cycle) in a graded manner whilst simultaneous measurements of inspired, and expired gases, heart rate, and electrocardiogram are made [3]. Numerous objective summaries of cardiopulmonary function can be derived from CPET including the oxygen consumption at the anaerobic threshold (AT), maximal oxygen consumption (VO2Max, a plateau in oxygen uptake at maximal exercise levels), ‘Peak VO2’,(or the highest oxygen uptake recorded in a sub-maximal test), as well as ventilatory equivalents for both oxygen (VE/ VO2) and carbon dioxide (VE/ VCO2) [4].

In 1993 it was reported that an AT of <11ml/kg/min accurately identified patients at risk of cardiovascular mortality undergoing intra-abdominal surgery [5]. Subsequent work by the same group utilised this finding to guide perioperative care, with high risk patients undergoing major surgery allocated to high dependency (HDU) or intensive care unit (ICU) beds [6]. Importantly, this work also demonstrated that a single metric alone may not be adequately discriminating. Patients with AT of >11mL/Kg/min but with evidence of myocardial ischaemia during CPET testing, or with severe pulmonary disease also suffered cardiopulmonary mortality and the overall numbers of deaths was low. Patients deemed to be ‘fit’ on CPET with an AT of >11mL/Kg/min and with no myocardial ischaemia had no instances of cardiac mortality. The authors themselves suggest that CPET is “even more reliable at detecting those not at risk…” [6].

More recently, CPET has been used to characterise individuals with chronic cardiac or respiratory failure [7] and, especially in the United Kingdom (UK), is being increasingly used for the stratification of patients prior to high risk procedures [8]. With nearly 15,000 individual tests a year from 2011 [9], this represents a significant logistical, organisational, and financial undertaking with the National Institute for Health and Care Excellence (NICE) quoting an average unit cost of £183 [10].

CPET is often used to gate access to more advanced care (such as intensive care unit admission) [6]. This strategy, of referring those identified as high risk using CPET to elective critical care stay, could have serious consequences if, in fact, the positive predictive value of CPET is poor. Critical care bed capacity is under incredible pressure [11] and unnecessary utilisation could add further strain to an already stretched system. Furthermore lack of critical care beds is already a significant risk factor for day of surgery cancellation [12], if this admission is based on the result of information with poor predictive performance this could lead to an unnecessary delay in an individual’s care. The issue of predictive performance is also important when decisions not to operate. Treatment options could be limited based on an individual being judged high risk but this result may in fact poorly predict the occurrence of complications. In a survey of practice the majority of respondents did not recommend cancellations based on CPET results but a significant minority (33%) did or, based on free-text comments, used these tests to inform subsequent discussions between clinical staff and the patient [9]. In their 2016 update on preoperative investigation NICE failed to make specific recommendations regarding the use of CPET but noted that other simpler, and cheaper tests may also be used for risk prediction [10]. For these reasons, the performance of CPET as a prognostic test should be carefully evaluated, as with other tests, in terms of its ability to correctly (or incorrectly) identify patients with and without the outcome of interest [13]. If CPET is used to identify an adverse outcome of interest within some time frame then the statistical interpretation of test performance can be performed using methods familiar from diagnostic testing [14]. Many of these (such as sensitivity, or specificity) are well known. However, no single metric can fully encapsulate a tests performance [15] and it is vital to recognise the potential consequences of false positives and negatives within the specific clinical context.

Numerous systematic reviews of the perioperative utility of CPET have been performed [1618]. These have highlighted the breadth of current work and concluded that CPET appeared to have promise as a marker of outcome amongst certain specialties [16, 17] but not all [18]. Formal meta-analysis has not been performed due to the heterogeneity of summary data and outcomes, making proper comparison difficult. The use of prediction models in medicine for diagnostic, and prognostic purposes to aid in patient decision making, and targeting of resources to high risk individuals is now widespread [19]. Despite CPET’s potential utility in these roles we are unaware of a systematic appraisal of generated prediction models that incorporate information from CPET testing.

In this work we aimed to identify newly published studies assessing CPET’s prognostic capabilities for cardiorespiratory complications, 30 day or in-hospital death, and unplanned intensive care unit admission for non-cardiopulmonary surgery. These outcomes were chosen a priori based on their relevance to the underlying physiology under test (Cardiorespiratory complications), their importance and examination in initial CPET studies (Mortality) [5], and that elective critical care admission is often advocated for CPET determined high risk cases [6] (unplanned ICU admission). As an extension over previously published systematic reviews we set out to generate comparable metrics of CPET test performance from the published data to aid clinicians in interpreting its performance as a prognostic test. Finally, for studies presenting a risk model using regression techniques, we aimed to provide the first evaluation of the methodological quality of these models using a validated approach [20].

Materials and methods

Search strategy

A systematic search strategy was developed in consultation with a medical librarian. We combined keywords for cardiorespiratory complications featured in the European Perioperative Clinical Outcome (EPCO) definitions [21] by the European Society of Anesthesiology (ESA) and the European Society of Intensive Care Medicine (ESICM), death and unplanned ICU admission. Further keywords to identify the perioperative period and cardiopulmonary exercise testing were also used. Medical Subject Headings (MeSH) terms were used as appropriate. Both Medline and Embase databases were searched on March 22nd 2017 using the Ovid interface. Due to the length of time required for study identification and data extraction the search was rerun on July 24th 2018 with new, applicable studies included. After initial submission the search was re-run on 2nd December 2019 for calendar years 2018 and 2019 to ensure our findings were current. PROSPERO registration was obtained (CRD42018095508). Full details of the search strategy is given in S1 Appendix. Titles and abstracts were screened for full-text inclusion by DJS and LAG. In cases of dispute inclusion was determined by arbitration by AE. References of included studies were then searched to identify further studies. Full numbers of studies identified and analysed at each stage are shown in the PRISMA diagram in Fig 1.

thumbnail
Fig 1. PRISMA diagram demonstrating identification of included studies.

Panel A demonstrates those included from initial search, Panel B those from repeat search. Meta-analysis refers to those included in calculation of confusion matrix metrics.

https://doi.org/10.1371/journal.pone.0226480.g001

Outcomes

We included studies where CPET was used to identify the risk of the following outcomes. Mortality (in-hospital/peri-operative and within 30 days of surgery), unplanned ICU admission, and complications affecting the cardiorespiratory systems. Studies looking at cardiorespiratory complications must have looked at the occurrence of at least one of the complications featured in the EPCO list of perioperative outcomes [21]. Our initial plan was that each study must follow the strict definitions incorporated within this document. However, it became apparent that this would lead to us including very few studies. Therefore studies were included as long as they looked at the occurrence of a complication featured in the list of cardiorespiratory complications or used one of the cited composite outcome measures (e.g. the Postoperative Morbidity Score—POMS). We included studies where cardiorespiratory complications were presented as a pooled outcome but not if they were included with other organ system complications (e.g. cardiorespiratory complications as judged using POMS was accepted, ‘any POMS positive’ was not).

Inclusion and exclusion criteria

Our pre-specified inclusion criteria for analysis were: Any primary research paper using CPET derived variables to aid in identification of adult (aged 18 or over) patients at risk of one of our pre-defined outcomes of interest. Studies could be prospective or retrospective in their design. Patients must have been undergoing non-cardiopulmonary surgery. Our exclusion criteria were: papers where complications were analysed as a pooled outcome, those that were not published in a peer-reviewed journal or available only as an abstract, review articles or correspondence. Only studies written in English were included. Previous meta-analyses on the subject were identified and their references, along with those of included studies, were searched for suitable studies for inclusion.

Data extraction

Risk of bias was assessed using the ‘Quality in Prognostic Studies’ (QUIPS) tool [22]. This was performed independently by DJS and LAG with arbitration of any conflicts undertaken by discussion between all three authors. Papers where a multivariable logistic regression model containing CPET derived variables was generated for one of our declared outcomes were also critiqued using the validated prognostic model checklist ‘CHARMS’ [20]. For all studies, we recorded whether a significant association was found for CPET variables and chosen outcomes, we also noted the statistical methodology used. We also extracted reported sensitivities and specificities for CPET variables. To supplement these published values we also data mined included papers to generate our own confusion matrices. In this process a ‘true positive’ was defined as a patient who had a CPET value categorising them as high-risk and suffered the outcome in question. A ‘true negative’ was a patient whose CPET variables categorised them as low-risk and did not suffer the specified adverse outcome. Our generated confusion matrices are available in supplementary material (S1 Table. We also recorded any AUC (area under the receiver operator curve) values for CPET and other prognostic variables reported in the paper. As a secondary aim we undertook the above process for other examined prognostic variables (e.g. American Society of Anesthesiologists- ASA score) within each paper. All data extraction was performed independently by DJS and LAG.

Prediction performance metrics

After generating confusion matrices for each study, classifier performance was evaluated on the basis of sensitivity, specificity, positive predictive value (PPV), negative predictive value (NPV), accuracy, F1 score, and Matthews Correlation Coefficient (MCC). The F1 score is the harmonic mean between a tests PPV and sensitivity and ranges between 0 (worst performance) and 1 (perfect performance). To score highly a test must have both a high PPV and should miss very few cases (high sensitivity). The MCC is a correlation coefficient between predictions and outcomes and can be interpreted analogously to the Pearson’s product moment correlation coefficient [23]. It incorporates all domains of the contingency table and therefore all classifier outcomes contribute equally to its calculation (i.e. misclassification of both positive and negative cases are important to its calculation). It ranges between -1 (test is always wrong) to +1 (test is always correct) with 0 indicating a performance no better than chance. Although our initial protocol dictated that we would attempt to generate summary receiver operator characteristic curves (SROC) their calculation demonstrated significant data heterogeneity that would render their findings invalid and, as such, were not included in this article.

Results

Study characteristics

Our initial search identified 534 records for screening with a further 24 identified through citation searching: These constituted 490 unique records. In total, 28 studies were included for qualitative review, from which classifier values could be extracted from 25. Our search was updated on two occasions. Initially, four new studies were included for qualitative analysis, including one paper where its inclusion proceeded after contacting the authors for unpublished data [24]. One of these studies was included in classifier analysis. A repeat search prior to publication identified four further studies from one of which measures of classifier performance could be calculated [25]. In total we therefore present results from 36 studies, with classifier values calculated from 27. Full inclusion and exclusion information is demonstrated in Fig 1. Studies covered multiple surgical specialities. Nine studies were performed in colorectal patients, six in vascular, six in upper gastrointestinal, six in hepatobiliary, six in mixed ‘major surgical populations’, two in urological, and one each in transplant and bariatric populations. Full details of all studies is shown in Table 1. The median number of study participants was 105 [Interquartile range 72-199]. 10 (28%) of studies were blinded meaning in the majority of studies (26/36–72%) information from CPET was either definitely or potentially used to guide the care of study participants (e.g. by stratifying admission to intensive care). In one study [26], despite not being blinded it was stated that CPET did not directly influence care. 23 (64%) of studies were performed prospectively. 28 studies looked at mortality as an outcome with 14 studies reporting on risk of combined cardiorespiratory morbidity. six studies reported purely cardiovascular complications, four respiratory complications, and six studies looked at rates of unplanned ICU admission. Risk of bias assessment for included studies, performed using the QUIPS checklist is shown in S1 Fig.

thumbnail
Table 1. Summary of included studies in systematic review of the performance of cardiopulmonary exercise testing (CPET) for identification of post-operative outcomes of interest n = 33.

https://doi.org/10.1371/journal.pone.0226480.t001

Selection of patients for CPET testing varied between studies. In two cases patients were selected for CPET on the basis of low levels of estimated functional capacity on a validated questionnaire. 12 studies used an age threshold for gating access to CPET, in seven of these a lower age threshold was used if the individual suffered from specific comorbidities (normally cardiovascular or diabetes). In five cases [2731] we felt it was impossible to determine, due to phrasing of the papers or study design, exactly who was selected to undergo CPET and whether eligible participants might have been missed. In one study the inclusion criteria changed throughout the inclusion period from being selected on the basis of comorbidity to being consecutive. For details see Table 1.

CPET variables and statistical significance

Studies looked at a variety of CPET derived variables. 33 studies (92%) looked at AT, 11 (31%) VE/VCO2, 11 (31%) the Peak VO2, and five (15.6%) VO2 Max. Numerous other variables including cardiac ischaemia, oxygen pulse, work rates, chronotropic incompetence, and combined metrics were looked at by smaller numbers of studies. Details of all CPET variables studied in individual papers is shown in Table 1. No variable was found to be consistently statistically significant between studies. 17 of 33 studies (51.5%) found a significant association between AT and a pre-specified outcome of interest, 5/11 (45.4%) found associations between Peak VO2, 4/11 (36.4%) VE/VCO2, and 2/5 (40%) VO2 Max. Of note, despite all studies utilising either a retrospective or prospective cohort design only one study utilised measures of relative risk [32]. None utilised Poisson or time-to-event statistical methods.

Prediction performance metrics

27 studies either presented sensitivity and specificity or presented data in a format that allowed us to calculate these and other performance measures by generating confusion matrices. These studies, values, outcomes, and CPET cut-offs are presented in Table 2. Due to the multitude of surgical specialities, outcome definitions, and differences in defined populations who underwent CPET we did not generate ‘average’ values of each metric for each outcome. Instead this table summarises ranges based on the outcomes of: mortality (all timescales up-to 30 days), cardiac complications alone, respiratory complications alone, combined morbidity, and unplanned ICU admission.

thumbnail
Table 2. Confusion matrix metrics for CPET derived cardiorespiratory variables.

https://doi.org/10.1371/journal.pone.0226480.t002

The most complete information was available for AT. For mortality, sensitivity values for AT ranged from 0-100%, specificity ranged from 41.5% to 92%. NPV of AT for mortality had a high estimated range (94.4-100%) which was higher than that for PPV (0–44.4%). These values are presented graphically in the boxplot in Fig 2 along with corresponding ranges for cardiorespiratory complications, and unplanned ICU admission. The range for MCC values was -0.18-0.47. F1 values ranged between 0 and 0.57.

thumbnail
Fig 2. Boxplots of calculated and published measures of test performance of anaerobic threshold in identification of individuals who suffer specific post-operative complications (n = 17 Studies).

Results are across heterogeneous surgical populations and use a variety of cut-offs and outcome definitions. A: Performance for identification of Mortality (up-to 30 day postoperatively), B: Cardiorespiratory complications (alone and/or in combination), C: Unplanned ICU admission. For source data see Table 2.

https://doi.org/10.1371/journal.pone.0226480.g002

For Peak VO2 as a predictor of mortality the sensitivity was consistently high ((85.7% to 100% from three studies) with correspondingly high NPV figures (98.9%–100%). Specificity ranged from 44.8% to 66.7%. F1 scores ranged between 0.05 and 0.11 (reflecting PPV values of 2.7% to 5.9%), MCC was between 0.12 and 0.16.

VE/VCO2 demonstrated a wide range of sensitivities for predicting mortality ((27.8% to 100%)) with specificity of 48.5% to 93.0%. NPV was consistently above 97% and PPV between 3.4% and 7.1%. All MCC, and F1 values were 0.12 or lower. Data from more than one study was available for Peak VO2, and VE/VCO2 as predictors of mortality, these are graphically demonstrated in S2 Fig. Values for other CPET variables are shown in Table 2.

Logistic regression models and CHARMS checklist

Seven studies generated a multivariable model for one of our pre-specified outcomes using, or including, CPET derived variables (Table 3). We critiqued these studies using the CHARMS checklist. All were derived from Cohort studies, four of which were prospective [27, 3335], two were retrospective [36, 37], and in one case it was unclear [29]. Two studies generated a model to predict mortality, four cardiorespiratory complications, and one unplanned ICU admission. The populations from which the models were generated varied by surgical speciality (Upper GI, vascular, hepatobiliary, renal transplant) and, in one paper [27], an unclear percentage of the eligible population were sent for CPET. Included covariates included CPET variables (AT, VE/VCO2, VO2 peak, work parameters) as well as scoring systems (ASA score, revised cardiac risk index), comorbidites, and specific medications. In four cases a clear process of model development was followed with the inclusion of variables significant on univariable analysis (normally p<0.1) or for pre-specified biological reasons, being included in initial multivariable models [29, 3436]. At the end of model generation one study [35] simplified to using AT alone as a predictor variable, and in one case [27] the process of model generation or simplification was unclear. Median sample size was 123 [IQR: 93-168]. Three of the studies had event to variable ratios (EVR) of above 10. An EVR of less than 10 is associated with a risk of ‘overfitting’ and generating a model with poor external validity [19]. Only two of the papers reported an assessment of model performance. One reported tests of discrimination (AUC) and calibration (Hosmer-Lemeshow) [34] with one reporting only a measure of discrimination (AUC) [35]. Only one study reported to have undertaken a process of internal evaluation of model performance (unspecified cross-validation approach) [34], no model was verified in an external dataset.

thumbnail
Table 3. Assessment of prediction models in identified studies using the CHARMS checklist (CHecklist for critical appraisal and data extraction for systematic reviews of prediction modelling studies.

https://doi.org/10.1371/journal.pone.0226480.t003

AUC and other predictors

Nine studies reported AUC values for CPET variables. For mortality these ranged from 0.75 (AT) to 0.95 (VE/VCO2). Presented values for cardiorespiratory complications appeared lower and ranged from 0.54 (Peak VO2) to 0.83 (AT). Full values are available in S1 Table. For seven studies classifier measures for other predictors (e.g. ASA score) were presented, or calculated (S2 Table). Potential predictors included blood tests, other scoring systems, or clinician determined high and low risk groups. For those studies looking at mortality [26, 28, 32, 33, 38], sensitivity ranged from 44.4% (presence of a single cardiac risk factor) to 100% (study delineated ‘high risk group’ based on CPET referral criteria). In the paper by James et al [39]. AUCs for a variety of other predictors were generated but no tests of significance between these and the AUCs calculated for CPET variables was performed.

Discussion

To our knowledge, this is the first systematic review to extract and calculate comparable measures of CPET’s performance as a prognostic test. We feel this framing of CPETs performance in familiar numerical terms is a key strength of our study as it provides a clear structure for considering the implications of CPET’s performance in routine practice.

A key paper in the field was discovered during our repeat search [40]. This was a prospective trial of over 1300 patients in which CPET was compared to both subjective clinician assessment and other objective methods of physical performance. Due to its importance we reviewed it for inclusion in our systematic review but it failed to meet our inclusion criteria due to its use of a pooled outcome of 30 day death and myocardial infarction. We will however highlight key messages from this paper in the context of our other findings. We did include a secondary analysis of this cohort where chronotropic incompetence during CPET was assessed as a predictor of postoperative myocardial injury assessed using rises in serum troponin [41].

Classifier performance of CPET

The differences in calculated NPV and PPV across studies examining AT is particularly striking (Fig 2). Both NPV and PPV depend on the frequency of the complication within the population, [42] being at their most extreme with a rare outcome (such as mortality). A gradual reduction in NPV, and increase in PPV as complications become more frequent (e.g. cardiorespiratory complications) can be seen. We also generated two lesser known measure of classifier performance, the MCC and the F1 score [15]. Its interpretation is relatively intuitive. It is a correlation coefficient, ranging from -1 to 1 with the two bounds representing a test that is always incorrect, and one that is always correct. Importantly, given the paucity of events in some studies, the measure is relatively resistant to class imbalance (i.e. low frequency of specific outcomes). We report a broad range (from -0.18 to 0.92) but with median values for our differing complications all being less than 0.17 indicating, across all domains of patient classification that CPET’s performance is relatively poor. Although no single measure of a test’s performance can fully encompass its clinical utility, in the context of CPET the F1 score arguably comes close. This is a score that equally weights the PPV and the sensitivity both of which, ideally, should be high (i.e. a positive CPET test indicates a high likelihood of developing a complication and that there are few false negatives). Across all outcomes and studies our estimates ranged from 0.04 to 0.64 with 22/48 values being 0.20 or lower. As shown in Table 2 studies with higher F1 scores demonstrated higher PPV. A deeper understanding of why these studies exhibited better PPV could be vital in determining optimal use of CPET (e.g. was it due to specific procedures, recruitment measures, or derived indices). The F1 score weights both false positives and false negatives equally but, depending on context this weighting can be adjusted (the Fβ) score) [15]. We utilised equal weighting in the CPET context as although we would like a sensitive measure, potential interventions (such as critical care beds), have a finite capacity and thus could not accommodate unlimited numbers of people who will actually not suffer a complication.

Our results suggest that most individuals with a ‘high-risk’ CPET result actually have a greater chance of not suffering a complication than they are to suffer one. This poor rule-in performance is important if decisions to not proceed to surgery are taken on the basis of such a result. It is also highly relevant if a scarce resource (such as a critical care bed) is being allocated preemptively based on a low likelihood of a complication occurring. Worse still, we have no systematic information to suggest whether the allocation of a critical care resource in the immediate post-operative period would be effective in preventing adverse events. Not only is this impossible from such observational work, but this may also be determined by timings of adverse events which may occur after a preemptive critical care admission is over.

Conversely, rule out performance is better, with a ‘low-risk’ CPET patient being unlikely to suffer complications meaning, in the correct context, this could reassure clinicians about proceeding with surgery. Acceptable thresholds for NPV, and PPV in terms of gating further care are open for discussion but should be considered in light of their potential benefits on the shared decision-making process in the run up-to surgery. Overall our results would suggest that CPET is better at ruling out complications (as indicated by its relatively high NPV) but this is also likely influenced by the frequency of complications and the sample size within the study in question. A nationwide survey in 2011 suggested that 33% of CPET centres may recommend cancellation based on CPET results, although in the majority of cases the results form part of the overall shared decision making process [9]. From our results we cannot comment specifically on CPET’s performance in this context although each result that we have calculated will be dependent on the preceding stratification process which was highly variable between studies (Table 1). Given its complexity in terms of organisation, staffing, and patient attendance the justification of CPET over other stratification tools relies on a demonstration that it offers additional benefit to patients and clinicians. An argument could be made, that in a sufficiently ‘high-risk’ population the frequency of complications would increase to a point that the NPV, and PPV of CPET would become more favourable. The correct method of doing this is not apparent from our included studies. A cautionary note for such a strategy is the fact that, in a population selected on the basis of cardiac risk factors, CPET offered no additional discriminating capabilities on risk of cardiac complications or death [40].

These, and other gaps in the literature have only recently been filled by the prospective, multi-centre ‘METS’ (Measurement of Exercise Tolerance before Surgery) trial [40]. This followed nearly 1,400 patients for one year after major, non-cardiac surgery, with daily examination for cardiac complications whilst an inpatient. They utilised the ‘net reclassification index’ to identify whether CPET, when added to baseline models of age, sex, cardiac risk factors, and surgery type was able to correctly reclassify individuals at risk of complications. For their primary outcome of 30-day mortality or myocardial infarction they showed no additional prognostic benefit of CPET derived variables. The only significant CPET variable was the addition of peak oxygen consumption for the identification of those at risk of a composite of all moderate to severe in-hospital complications. The study also assessed both the Duke Activity Score Index (DASI), and NT-pro-BNP for improved identification of those at risk of their primary outcome. Neither were significant for this outcome but, both demonstrated benefit at identifying sub-clinical myocardial injury (as assessed by troponin assay). The METS study suggests that CPET adds little to easily assessed baseline factors for predicting risk. An important caveat might be the frequency of outcomes, with the study only having 28 individuals suffering their composite primary outcome [40].

In addition, our study is the first to apply the CHARMS [20] criteria to regression models generated using CPET variables. We critiqued models generated against any one of our pre-specified outcomes. In general, generation, and assessment of regression models was incomplete. Only one study assessed internal validity (using cross-validation) and three of the models were at risk of overfitting due to low event to variable (EVR) ratios. No model has (from our results) ever been validated in a different patient cohort or used as a starting point for refinement by subsequent authors. As such, generalising the findings of these analyses is unfounded. The lack of clear outputs of model performance (such as discrimination, and calibration) means that even internal validity may be difficult to determine in certain cases.

The results of our systematic review indicate a broad range of estimates for CPET performance as a prognostic test in non-cardiopulmonary surgery. The single centre, retrospective nature of many of the studies limits the ability to generalize findings to everyday practice and it is telling that the original paper identifying an AT of 11ml/kg/O2 as an important threshold is now 25 years old and, arguably, may not be reflective of today’s surgical practice. It is telling that, as highlighted in Table 1, many studies did not blind clinicians to the results of CPET, introducing the potential for bias and confounding into their (and, subsequently, our) findings. Nationally, CPET has been used to determine various components of perioperative care including ICU admission, monitoring, or type of surgical procedure [9]. Some of these interventions (e.g. routine critical care for high risk individuals) could, arguably, bias findings towards the null by having averted potential complications in the high-risk cohort. The fact that 7 of 11 studies where CPET was explicitly or potentially stated to have not altered care, failed to demonstrate significant associations with complications suggests this is unlikely to be the whole truth.

Whilst our work illustrates that CPET alone is not a good predictor of adverse events our results may not fully encompass the subtleties of its use within clinical practice. Certainly it is possible that there exists a spectrum of use and it may be that CPET is of predictive value in a subset of very high risk patients. Unfortunately, a lack of methodological standardization makes this impossible to assess. Indeed, in our included studies there has been little or no systematic attempt to assess the incremental discrimination of CPET over and above other, perhaps simpler assessments.

Strengths and weaknesses of our approach

We acknowledge that due to the potential for bias and confounding within the primary studies that our extracted and calculated estimates are, themselves, likely to be biased. We defend this by arguing that the widespread adoption of CPET is arguably grounded in the interpretation of these studies and that we are merely quantifying metrics which should be implicit in the extrapolation of these studies to clinical practice. We judged that calculating classifier metrics based on these studies was a valid approach as we focused on a single time period (the in-hospital period) and that the majority of studies did not delineate the exact time a complication was screened for or occurred. If raw data was available then use of Poisson or time to event (e.g. Cox) regression techniques would be an alternative to demonstrate an association between CPET variables and outcomes of interest. We feel that our use of classifier measures in this prognostic context (which is supported in the literature [14]) offers a framework for interpretation that can be directly related to clinical practice.

Our results are heterogeneous due to the plethora of settings, surgical sub-specialities, selection criteria, outcomes, and outcome definitions used. It is for this reason that we have taken no steps to present a pooled or aggregate measure of CPET’s performance. This is an important finding in itself and demonstrates a troubling lack of methodological standardisation in the literature and presumably therefore perhaps also in how CPET is used clinically. Our initial aim had been to only include studies utilising rigorously defined definitions of specific postoperative complications [21] but this had to be broadened to any named complication within this document to allow for a suitable number of studies to be included. We justify this by saying that any concerns we had with the chosen method of outcome definition is reflected in the QUIPS score given to the study for that domain (S1 Fig). Our selection of outcomes for inclusion was underpinned by physiological rationale and the utilisation of CPET in routine practice. We tried to minimise the use of composite outcomes to reduce heterogeneity in our generated classifier metrics. We did extract pooled cardiorespiratory outcomes if so presented due to the potential for crossover between these measures depending on the scoring system used. For instance the postoperative morbidity score requires an individual to be on oxygen to class as having a respiratory complication, which is highly likely to have occurred in the context of an individual with heart failure (who would be classed as having a cardiovascular complication) [21]. As a caveat to this, if multi domain scores like this were presented we calculated the occurrence of cardiac and respiratory complications separately to minimise the risk of double counting individuals. The use of composite outcomes is widespread especially as they can lead to an increase in statistical power [43]. Issues with this approach are that any effect may be interpreted by the reader to apply equally across all components of the composite, and that there should be an equivalent effect on each component. Furthermore, an interesting area of further work would be to explore how valid the use of ‘all complications’ is from a patient perspective, the presentation of information pertaining to such a composite is arguably only at its most valid if the patient weighs the impact of each component equally.

We accept that in many of the included studies our outcome was not the author’s primary outcome and this may introduce bias into our results by skewing indices such as sensitivity and specificity (See calculations in S1 File). We hope that by providing full access to our calculations, including the location of generated numbers within the corresponding paper, that we have provided sufficient information for clinicians to interpret our findings in context.

Conclusion

Our systematic review highlights potential limitations of CPET by generating common and novel measures of test performance. Specifically, given the frequency of complications in studies we highlight the low values of PPV, although far from perfect for gating access to surgery or critical care they may help inform a process of shared decision making prior to surgery. The performance of CPET in day to day practice may be improved if a sufficiently high risk population could be identified for it to be employed in. The high NPV offered by CPET may be testament to accuracy of its underlying physiological rationale. However, based on our results, and those from a recent high-quality prospective study [40] the precision of CPET for reliably identifying individuals who will suffer postoperative cardiorespiratory complications appears poor. Certainly, given the values of PPV we report it would be difficult to justify not proceeding to surgery on the basis of a high-risk CPET test alone. Due to its role within the wider system of perioperative assessment, optimisation, and postoperative care a nuanced interpretation of its ability to rule-in and rule-out potential complications is justified.

Supporting information

S1 Appendix. Literature search strategy.

Example given for repeat search—same terms used for initial search, see Methods.

https://doi.org/10.1371/journal.pone.0226480.s001

(PDF)

S1 Fig. Risk of bias assessment for included studies.

Scoring performed using the QUIPS (Quality in Prognosis Studies) tool. Reported values are those reached after individual assessment by two authors and arbitration amongst all three. Red = High risk of bias, Orange = Moderate risk of bias, Green = Low risk of bias (for associated domain).

https://doi.org/10.1371/journal.pone.0226480.s004

(TIF)

S2 Fig. Boxplots of calculated and published measures of test performance, and positive and negative predictive values for peak VO2(A), and VE/VCO (B) as predictors of mortality.

Source data can be seen in table of calculated confusion matrix metric (S1 File).

https://doi.org/10.1371/journal.pone.0226480.s005

(TIF)

S1 Table. Published area under the receiver operator characteristic (ROC) curve values from identified studies.

Brackets indicate published 95% confidence interval. Asterisk indicates that published confidence interval crosses 1 (likely calculated as a normal appproximation).

https://doi.org/10.1371/journal.pone.0226480.s007

(PDF)

S2 Table. Performance of non cardiopulmonary exercise testing (CPET) variables for prediction of specific outcomes.

MET = Metabolic equivalents, BNP = B type natriuretic peptide, NT-proBNP = N-terminal pro b-type natriuretic peptide, eGFR = estimated glomerular filtration rate, CRP = C Reactive Protein, ASA = American Society of Anesthesiologists score, RF = Risk Factors, IHD = Ischaemic Heart Disease, AT = Anaerobic Threshold, RCRI = Revised Cardiac Risk Index. POSSUM = Physiological and Operative Severity Score for the Enumeration of Morbidity and Mortality. AUC = Area under the receiver operator curve, OR = Odds ratio, * calculations exclude 14 high risk patients who didn’t proceed to surgery based ‘partly’ on CPET findings.

https://doi.org/10.1371/journal.pone.0226480.s008

(PDF)

Acknowledgments

DJS is supported by a Wellcome Trust Clinician PhD Fellowship [Overarching grant: 204017/Z/16/Z].

The authors would like to thank Prof Simon Griffin and Dr Juliet Usher-Smith for numerous helpful discussions as well as Ms Isla Kuhn for assistance with the design of search strategy.

References

  1. 1. NELA Project Team The fourth patient report of the National Emergency Laparotomy Audit. RCoA London, 2018. Available at: https://www.nela.org.uk/reports.
  2. 2. Ajitsaria P, Eissa SZ, Kerridge RK Risk Assessment Curr Anesthesiol Rep. 2018; 8: 1–8. pmid:29527132
  3. 3. Agnew N Preoperative cardiopulmonary exercise testing Contin Educ Anaesthesia, Crit Care Pain. 2010; 10: 33–37.
  4. 4. Balady GJ, Arena R, Sietsema K, Myers J, Coke L, Fletcher GF, et al Clinician’s guide to cardiopulmonary exercise testing in adults: A scientific statement from the American heart association Circulation. 2010; 122: 191–225. pmid:20585013
  5. 5. Older P, Smith R, Courtney P, Hone R Preoperative evaluation of cardiac failure and ischemia in elderly patients by cardiopulmonary exercise testing Chest. 1993; 104: 701–704. pmid:8365279
  6. 6. Older P, Hall A, Hader R Cardiopulmonary exercise testing as a screening test for perioperative management of major surgery in the elderly. Chest. 1999; 116: 355–362. pmid:10453862
  7. 7. Paolillo S, Agostoni P Prognostic role of cardiopulmonary exercise testing in clinical practice. Ann Am Thorac Soc. 2017; 14:S53–S58. pmid:28362512
  8. 8. Biccard BM Improving the evidence-base for preoperative cardiopulmonary exercise testing. Br J Anaesth. 2018; 120: 419–421. pmid:29452793
  9. 9. Huddart S, Young El, Smith R-L, Holt PJ, Prabhu PK Preoperative cardiopulmonary exercise testing in England—a national survey Perioper Med. 2013; 2: 4.
  10. 10. National Institute fort Health and Care Excellence. Preoperative tests(update): Routine preoperative tests for elective surgery 2016. Available from: https://www.nice.org.uk/guidance/ng45/evidence.
  11. 11. The Faculty of Intensive Care Medicine Critical Capacity 2018. Available from: https://www.ficm.ac.uk/sites/default/files/ficm_critical_capacity_-_a_short_research_survey_on_critical_care_bed_capacity.pdf.
  12. 12. Wong DJN, Harris SK, Moonesinghe SR, et al. Cancelled operations: a 7-day cohort study of planned adult inpatient surgery in 245 UK National Health Service hospitals. Br J Anaesth. 2018; 121: 730–738. pmid:30236235
  13. 13. Maxim LD, Niebo R, Utell MJ. Screening tests: a review with examples. Inhal Toxicol. 2014; 26: 811–828. pmid:25264934
  14. 14. Agency for Healthcare Research and Quality Methods guide for medical test reviews. AHRQ Publication No. 12-EC017 2012. Available from: https://effectivehealthcare.ahrq.gov/topics/methods-guidance-tests/overview-2012/.
  15. 15. Lever J, Krzywinski M, Altman N. Classification evaluation. Nat Methods. 2016; 13: 603–605.
  16. 16. Moran J, Wilson F, Guinan E, McCormick P, Hussey J, Moriarty J Role of cardiopulmonary exercise testing as a risk-assessment method in patients undergoing intra-abdominal surgery: a systematic review. Br J Anaesth. 2016; 116: 177–191. pmid:26787788
  17. 17. Ney M, Haykowsky MJ, Vandermeer B, Shah A, Ow M, Tandon P. Systematic review: pre- and post-operative prognostic value of cardiopulmonary exercise testing in liver transplant cadidates. Aliment Pharmacol Ther. 2016; 44: 796–806. pmid:27539029
  18. 18. Young El, Karthikesalingam A, Huddart S, Pearse RM, Hinchliffe RJ, Loftus IM, et al. A systematic review of the role of cardiopulmonary exercise testing in vascular surgery. Eur J Vasc Endovasc Surg. 2012; 44:64–71. pmid:22521839
  19. 19. Steyerberg EW, Vickers AJ, Cook NR, Gerds T, Gonen M, Obuchowski N, et al. Assessing the performance of prediction models: a framework for traditional and novel measures. Epidemiology. 2010; 21: 128–138. pmid:20010215
  20. 20. Moons KGM, de Groot JAH, Bouwmeester W, Vergouewe Y, Mallett S, Altman DG, et al. Critical Appraisal and Data Extraction for Systematic Reviews of Prediction Modelling Studies: The CHARMS Checklist. PLoS Med. 2014; 11:e1001744. pmid:25314315
  21. 21. Jammer I, Wickboldt N, Sander M, Smith A, Schultz MJ, Pelosi P, et al. Standards for definitions and use of outcome measures for clinical effectiveness research in perioperative medicine: European Perioperative Clinical Outcome (EPCO) definitions. Eur J Anaesthesiol. 2015; 32: 88–105. pmid:25058504
  22. 22. Hayden JA, van der Windt DA, Cartwright JL, Cote P, Bombardier C. Assessing bias in studies of prognostic factors. Ann Intern Med. 2013; 158: 280–286. pmid:23420236
  23. 23. Powers DMW Evaluation: From precision, recall and F-Measure to ROC, Informedness, Markedness and Correlation J Mach Learn Technol. 2011; 2: 37–63.
  24. 24. Dunne DFJ, Jones RP, Lythgoe DT, Pilkington FJ, Palmer DH, Malik HZ, et al. Cardiopulmonary exercise testing before liver surgery J Surg Oncol. 2014; 110: 439–444. pmid:24894657
  25. 25. Drummond RJ, Vass D, Wadhawan H, Craig CF, Fullarton GM, Forshaw MJ. Routine pre- and post-neoadjuvant chemotherapy fitness testing is not indicated for oesophagogastric cancer surgery Ann R Coll Surg Engl. 2018; 100: 515–519. pmid:29692190
  26. 26. Ausania F, Vallance AE, Manas DM, Holmes LR, Jaques BC, White SA, et al. Effects of low cardiopulmonary reserve on pancreatic leak following pancreaticoduodenectomy Br J Surg. 2012; 999: 1290–1294.
  27. 27. Barakat HM, Shahin Y, McCollum PT, Chetter IC. Prediction of organ-specific complications following abdominal aortic aneurysm repair using cardiopulmonary exercise testing Anaesthesia. 2015; 70: 679–685. pmid:25656939
  28. 28. Bowles TA, Sanders KM, Colson M, Watters DA. Simplified risk stratification in elective colorectal surgery ANZ J Surg. 2008; 78: 24–27. pmid:18199201
  29. 29. Hartley RA, Pichel AC, Grant SW, Hickey GL, Lancaster PS, Wisely NA, et al. Preoperative cardiopulmonary exercise testing and risk of early mortality following abdominal aortic aneurysm repair. Br J Surg. 2012; 99: 1539–1546. pmid:23001820
  30. 30. Goodyear SJ, Yow H, Saedon M, Shakespeare J, Hill CE, Watson D, et al. Risk stratification by pre-operative cardiopulmonary exercise testing improves outcomes following elective abdominal arotic aneurysm surgery: a cohort surgery Perioper Med. 2013;2:10.
  31. 31. Whibley J, Peters CJ, halliday LJ, Chaudry AM, Allum WH. Poor performance in incremental shuttle walk and cardiopulmonary exercise testing predicts poor overall survival for patients undergoing esophago-gastric resection Eur J Surg Oncol. 2018;44:594–599g.
  32. 32. Wilson RJT, Davies S, Yates D, Redman J, Stone M. Impaired functional capacity is associated with all-cause mortality after major elective intra-abdominal surgery. Br J Anaesth. 2010;105:297–303. pmid:20573634
  33. 33. Junejo MA, Mason JM, Sheen AJ, Moore J, Foster P, Atkinson D, et al. Cardiopulmonary exercise testing for preoperative risk assessment before hepatic resection Br J Surg 2012; 99: 1097–10104. pmid:22696424
  34. 34. Ting SMS, Iqbal H, Hamborg T, Imray CHE, Hewins S, Banerjee P, et al. Reduced functional measure of cardiovascular reserve predicts admission to critical care unit following kidney transplantation PLoS One. 2013; 8: e64335. pmid:23724043
  35. 35. Snowden CP, Prentis J, Jacques B, Anderson H, Manas D, Jones D, et al. Cardiorespiratory Fitness predicts mortality and hospital length of stay after major elective surgery in older people Ann Surg. 2013; 257: 999–1004. pmid:23665968
  36. 36. Lam S, Alexandre L, Hardwick G, Hart AR. The association between preoperative cardiopulmonary exercise-test variables and short-term morbidity after esophagectomy: A hospital-based cohort study Surgery. 2019;166:28–33. pmid:30981415
  37. 37. Nagamatsu Y, Shima I, Yamana H, Fujita H, Shirouzu K, Ishitake T. Preoperative evaluation of cardiopulmonary reserve with the use of expired gas analysis during exercise testing in patients with squamous cell carcinoma of the thoracic esophagus. J Thorac Cardiovasc Surg. 2001; 121: 1064–1068. pmid:11385372
  38. 38. Junejo M, Mason J, Close H, Sheen AJK, Bryan A, Moore J, et al. Cardiopulmonary exercise testing for preoperative risk assessment before pancreticoduodenectomy for cancer Ann Surg Oncol. 2014;21:1929–1936. pmid:24477709
  39. 39. James S, Jhanji S, Smith A, Brien GO, Fitzgibbon M, Pearse RM. Comparison of the prognostic accuracy of scoring systems, cardiopulmonary exercise testing, and plasma biomarkers: a single-centre observational pilot study. Br J Anaesth. 2014; 112: 491–497. pmid:24148323
  40. 40. Wijeysundera DN, Pearse RM, Shulman MA, Abbott TEF, Torres E, Ambosta A, et al. Assessment of functional capacity before major non-cardiac surgery: an international, prospective cohort study Lancet. 2018;391: 2631–2640. pmid:30070222
  41. 41. Abbott TEF, Pearse RM, Beattie WS, Phull M, Beilstein C, Raj A, Grocott MPW, Cuthbertson BH, Wijeysundera D, Ackland GL. Chronotropic incompetence and myocardial injury after noncardiac surgery: planned secondary analysis of a prospective observational international cohort study Br J Anaesth. 2019; 123: 17–26. pmid:31029407
  42. 42. Molinaro AM How to estimate the positive predictive value. Neuro-Oncology Pract. 2014; 2: 161–165.
  43. 43. Freemantle N, Calvert M, Wood J, Eastaugh J, Griffin C Composite outcomes in randomised trials. Greater precision but with greater uncertainty? JAMA. 2003;289:2554–2559. pmid:12759327
  44. 44. Nugent AM, Riley M, Megarry J, Reilly MJGO, Macmahon J, Lowry R. Cardiopulmonary exercise testing in the pre-operative assessment of patients for repair of abdominal aortic aneurysm. Ir J Med Sc. 1998;167: 238–241.
  45. 45. McCullough P, Gallagher MJ, DeJong AT, Sandberg KR. Cardiorespiratory fitness and short-term complications after bariatric surgery. Chest. 2006; 130: 517–525. pmid:16899853
  46. 46. Forshaw MJ, Strauss DC, Davies AR, Wilson D, Lams B, Pearce A, et al. Is cardiopulmonary exercise testing a useful test before esophagectomy Ann Thorac Surg. 2008; 85: 294–299. pmid:18154826
  47. 47. Struthers R, Erasmus P, Holmes K, Warman P, Collingwood A, Sneyd JR. Assessing fitness for surgery: A comparison of questionnaire, incremental shuttle walk, and cardiopulmonary exercise testing in general surgical patients. Br J Anaesth. 2008; 101: 774–780. pmid:18953057
  48. 48. Snowden CP, Prentis JM, Anderson HL, Roberts DR, Randles D, Renton M, et al. Submaximal cardiopulmonary exercise testing predicts complications and hospital length of stay in patients undergoing major elective surgery. Ann Surg. 2010; 251: 535–541. pmid:20134313
  49. 49. Thompson A, Peters N, Lovegrove R, Ledwidge S, Kitching A, MAgee TR, et al. Cardiopulmonary exercise testing provides a predictive tool for early and late outcomes in abdominal aortic aneurysm patients Ann R Coll Surg Engl. 2011; 93: 474–481. pmid:21929919
  50. 50. Chandrabalan VV, McMillan DC, Carter R, Kinsella J, McKay CJ, Carter CR, et al. Pre-operative cardiopulmonary exercise testing predicxts adverse post-operative events and non-progression to adjuvant therapy after major pancreatic surgery HPB. 2013; 15: 899–907. pmid:23458160
  51. 51. Lai CW, Minto G, Challand CP, Hosie KB, Sneyd JR, Creanor S, et al. Patients’ inability to perform a preoperative cardiopulmonary exercise test or demonstrate an anaesrobic threshold is associated with inferior outcomes after major colorectal surgery Br J Anaesth. 2013; 111: 607–611. pmid:23744818
  52. 52. Moyes LH, McCaffer CJ, Carter RC, Fullarton GM, Mackay CK, Forshaw MJ. Cardiopulmonary exercise testing as a predictor of complications in oesophagogastric cancer surgery Ann R Coll Surg Engl. 2013; 95: 125–130. pmid:23484995
  53. 53. Prentis JM, Trenell MI, Vasdev N, French R, Dines G, Thorpe A, et al. Impaired cardiopulmonary reserve in an elderly population is related to postoperative morbidity and length of hospital stay after radical cystectomy BJU Int. 2013;112: E13–E19. pmid:23795790
  54. 54. West MA, Lythgoe D, Barben CP, Noble L, Kemp GJ, Jack S, et al. Cardiopulmonary exercise variables are associated with postoperative morbidity after major colonic surgery: a prospective blinded observational study Br J Anaesth. 2014; 112: 665–671. pmid:24322573
  55. 55. West MA, Loughney L, Barben CP, Sripadam R. The effects of neoadjuvant chemoradiotherapy on physical fitness and morbidity in rectal cancer surgery patients Eur J Surg Oncol. 2014; 40: 1421–1428. pmid:24784775
  56. 56. Chan KE, Pathak S, Smart NJ, Batchelor N, Daniels IR. The impact of cardiopulmonary exercise testing on patients over the age of 80 undergoing elective colorectal cancer surgery Colorectal Dis. 2015; 18: 578–585.
  57. 57. Nikolopoulos I, Ellwood M, George M, Carapeti E, Williams A. Cardiopulmonary exercise testing versus spirometry as predictors of cardiopulmonary complications after colorectal surgery. Eur Surg. 2015;47: 324–330.
  58. 58. West MA, Asher R, Browning M, Minto G, Swart M, Richardson K, et al. Validation of preoperative cardiopulmonary exercise testing-derived variables to predict in-hospital morbidity after major colorectal surgery Br J Surg. 2016;103: 744–752. pmid:26914526
  59. 59. Kanakaraj M, Yates Dr, Wilson RJT, Baroni Ml, Davies SJ. Prognostic markers of outcome in patients undergoing infra-inguinal revascularisation: A prospective observational pilot study. Eur J Vasc Endovasc Surg. 2017;54:212–219. pmid:28625356
  60. 60. Wilson RJT, Yates DRA, Walkington JP, Davies SJ. Ventilatory inefficiency adversely affects outcomes and longer-term survival after planned colorectal cancer surgery Br J Anaesth. 2019; 123: 238–245. pmid:30916023