Skip to main content
Browse Subject Areas

Click through the PLOS taxonomy to find articles in your field.

For more information about PLOS Subject Areas, click here.

  • Loading metrics

Development and psychometric testing of the clinical networks engagement tool

  • Jill M. Norris ,

    Affiliation Faculty of Nursing, University of Calgary, Calgary, Alberta, Canada

  • Kent G. Hecker,

    Affiliations Department of Community Health Sciences, Cumming School of Medicine, University of Calgary, Calgary, Alberta, Canada, Department of Veterinary Clinical and Diagnostic Sciences, Faculty of Veterinary Medicine, University of Calgary, Calgary, Alberta, Canada

  • Leora Rabatach,

    Affiliation Department of Community Health Sciences, Cumming School of Medicine, University of Calgary, Calgary, Alberta, Canada

  • Tom W. Noseworthy,

    Affiliation Department of Community Health Sciences, Cumming School of Medicine, University of Calgary, Calgary, Alberta, Canada

  • Deborah E. White

    Affiliation Faculty of Nursing, University of Calgary, Calgary, Alberta, Canada



Clinical networks are being used widely to facilitate large system transformation in healthcare, by engagement of stakeholders throughout the health system. However, there are no available instruments that measure engagement in these networks.


The study purpose was to develop and assess the measurement properties of a multiprofessional tool to measure engagement in clinical network initiatives. Based on components of the International Association of Public Participation Spectrum and expert panel review, we developed 40 items for testing. The draft instrument was distributed to 1,668 network stakeholders across different governance levels (leaders, members, support, frontline stakeholders) in 9 strategic clinical networks in Alberta (January to July 2014). With data from 424 completed surveys (25.4% response rate), descriptive statistics, exploratory and confirmatory factor analysis, Pearson correlations, linear regression, multivariate analysis, and Cronbach alpha were conducted to assess reliability and validity of the scores.


Sixteen items were retained in the instrument. Exploratory factor analysis indicated a four-factor solution and accounted for 85.7% of the total variance in engagement with clinical network initiatives: global engagement, inform (provided with information), involve (worked together to address concerns), and empower (given final decision-making authority). All subscales demonstrated acceptable reliability (Cronbach alpha 0.87 to 0.99). Both the confirmatory factor analysis and regression analysis confirmed that inform, involve, and empower were all significant predictors of global engagement, with involve as the strongest predictor. Leaders had higher mean scores than frontline stakeholders, while members and support staff did not differ in mean scores.


This study provided foundational evidence for the use of this tool for assessing engagement in clinical networks. Further work is necessary to evaluate engagement in broader network functions and activities; to assess barriers and facilitators of engagement; and, to elucidate how the maturity of networks and other factors influence engagement.


Large-scale transformation in healthcare requires engaging stakeholders across the health system.[14]. Engagement has been described as the active involvement of stakeholders in maintaining and enhancing the performance of their organisation.[2,3] Evidence suggests that when healthcare professionals are engaged in their health system, organizations benefit from reductions in mortality, adverse drug events, errors, and infection rates,[58] as well as enhanced quality of care and patient experience.[7,9,10] Findings from a large-scale study in the NHS indicated that trusts with higher staff engagement exhibit better financial performance.[7] Alongside this growing evidence for the link between engagement and performance, there have been a number of advances in physician[3,6,11,12] and patient engagement in healthcare,[1315] Efforts to engage the spectrum professionals and stakeholders who design and carry out quality improvement initiatives, however—a process outlined within numerous healthcare improvement models.[1622]—have often been met with limited success.[2325]

Arguably, one of the major gaps towards improving suboptimal engagement in healthcare improvement is the lack of a tool to measure the process. Appraisals about who is sufficiently engaged and levels that create meaningful improvements have depended upon subjective judgements to date. We are aware of only one recent cross-sectional study that included a measure of work engagement in a large improvement program,[26] and two studies that assessed physician engagement with minority patients[27] and with service quality.[3] Within the organizational literature, engagement has been measured as a multidimensional concept comprised of psychological states (e.g., vigour), enduring traits (e.g., personality) and behaviours (e.g., involvement).[2835] In contrast, the healthcare improvement literature predominantly describes engagement as a process conducted by the organization,[28] in which stakeholders are encouraged to participate in a range of improvement activities and phases, including priority-setting and sustaining projects long term.[16,1921]

One model that is potentially applicable to engagement in healthcare improvement is the International Association of Public Participation (IAP2) Spectrum of Public Participation.[36] Multiple healthcare organizations have applied the spectrum with multiprofessional clinician groups in Australia[37] and Canada, as well as to patient groups.[38] Building upon Arnstein’s[39] ladder of participation, later modified by Conner[40] and Pretty[41], the IAP2 spectrum of engagement[36] posits that organizational outcomes improve when stakeholders are included in increasing levels of engagement. This includes setting priorities and decision-making through participatory, transformative and democratic processes.[42] Five levels of engagement processes are outlined within the IAP2 spectrum: inform, consult, involve, collaborate, and empower. At one end of the spectrum, stakeholders are informed about an initiative; on the other end, they are empowered with final decision-making authority. While promising, the IAP2 has not thus far undergone empirical evaluation. The purpose of this study was to (1) develop a multiprofessional tool based on the IAP2 spectrum to measure engagement processes in clinical network initiatives, and (2) assess the measurement properties of this tool, including an assessment of reliability and validity of the scores.



Over the past decade, health systems globally have implemented clinical networks, and one variant, strategic clinical networks (SCNs), as a means of improving quality of care through a systems-level approach to change.[4346] SCNs were established in 2012 by Alberta Health Services (AHS), the provincial health system in Alberta, Canada. Health systems in Europe (especially in the UK) along with those in Australia and Canada were early adopters of these networks; more recently, we have seen these networks established in New Zealand,[47] Malawi,[48] and Kuwait.[49] Clinical networks are thought to benefit from communication channels and relationships across multidisciplinary and hierarchical levels in the organization, and intend to achieve meaningful active involvement of various stakeholders.[45,46,50] While clinical networks vary greatly in their focus—whether it be a clinical area, condition, population, operational area, or an intervention—they are all structured to enable better coordination between essential healthcare stakeholders.

We studied engagement in the newly formed SCNs because part of their mandate was to engage partners across the health system. This could provide valuable baseline information to decision makers during the formative and developing stages of these structures. Moreover, AHS had adapted the IAP2 spectrum as their model of clinician engagement. We previously established relationships with the senior decision-makers in AHS, which provided an excellent opportunity to co-design this study and propose a co-developed program of research to evaluate SCNs. This resulted in the adoption of an integrated knowledge approach in which researchers and knowledge-users worked together to craft research questions, refined the methodology, and remained involved in interpreting and using the findings.

Item construction

The purpose of the tool was to measure perceptions of engagement over various phases of the implementation of projects undertaken by the SCNs. The descriptions and labels within the IAP2 Spectrum[36] first informed the construction of 25 original items for each of the five levels of engagement. Next, we then tailored the item wording to ensure relevance and refer to the activities of the networks, and further worked with our knowledge users to generate an additional 15 items that aimed to capture engagement more broadly than the Spectrum. The 40 items then underwent independent review by a panel of six purposively sampled content experts who had direct experience working with the SCNs, including health services and clinical network researchers, employees of AHS who were working with the SCNs on their engagement strategies, and SCN decision-makers. They were asked to give an assessment of item contents, item style and comprehensiveness of the instrument.[51] In addition, a convenience sample of six clinicians (nurses, physicians) independently reviewed the items before testing the instrument with a larger sample.

Based on the feedback, the research team revised and compiled a final set of 40 items: 4 items that intended to measure the overall (i.e., global) engagement construct (dependent variable) and 36 items that characterized five potential engagement level constructs (inform, consult, involve, collaborate, empower; independent variables). As a response to requests during the item construction stage, 6 = not applicable option was added for engagement level items, as were items related to evaluating projects. Within each construct grouping, items assessed engagement across four phases of implementation: setting priorities, planning, implementing, and evaluating projects. Overall engagement items were anchored with 1 = not at all engaged to 5 = extremely engaged. Engagement-level items were anchored with 1 = strongly disagree and 5 = strongly agree. Table A in S1 File lists the original items.

Instrument testing

Participants and procedures.

Between January and July 2014, participants were recruited from the nine SCNs purposively selected for participation in this study (Addiction & Mental Health; Bone & Joint Health; Cancer; Critical Care; Cardiovascular Health & Stroke; Diabetes Obesity & Nutrition; Emergency; Senior’s Health; Surgery). Voluntary participation was sought across four levels of the SCN governance structure: (1) leaders; (2) core and working group members, among other stakeholders outside of formalized SCN membership; including, (3) support personnel; and, (4) frontline healthcare stakeholders. Members were identified from SCN membership lists, which included contact information, their position, role on the SCN (e.g., clinician, co-executive leader, zone lead, executive director, working group member), and other organizational contact information. Recruitment occurred first through presentations by the research lead (DW) during SCN meetings. Next, a personalized email describing the study was sent to SCN members. To facilitate response rate, we attended to web survey principles that are effective in achieving higher response rates (e.g., online format, option to complete paper-based survey, simple and usable design, follow-up reminder emails and phone calls at 2- and 4-week intervals).[52]. Eligibility for study participation included the following: a) SCN member or SCN-identified stakeholder outside of the SCN membership; b) English-speaking; and, c) access to a computer.

Sample size.

We estimated that each SCN would have maximally 100 formal members in the core and working groups from whom to sample, for an estimated pool of 900 SCN members. Using a conservative response rate of 25% for online surveys, a sample of 225 participants would surpass a recommended sample size of 200 participants for both exploratory factor analysis [53,54] and confirmatory factor analysis.[55,56]

Statistical analysis

Data were first compiled and anonymized, then cleaned prior to analysis. Descriptive statistics were calculated for each item and subscales in SPSS v22 (IBM, Chicago, IL, USA). No values were imputed for missing data.

Validation evidence.

To examine evidence of construct validity—whether the tool measures the construct(s) intended to assess[57]—we used exploratory factor analysis (EFA), confirmatory factor analysis (CFA), linear regression and one-way multivariate analysis of variance (MANOVA). To test dimensionality of the instrument, an EFA was conducted in SPSS v22 (IBM, Chicago, IL, USA). Unweighted least squares extraction with oblimin rotation and Kaiser normalization was chosen to maximize the variance extracted, given that we assumed that factors would be correlated. To identify the number of factors, eigenvalues > 1 and the scree plot of eigenvalues plotted against factors were examined. Table B in S1 File details the results of the preliminary EFA with all items, which loaded onto six factors, or subscales.

Items that had more than 10% of missing data, cross-loaded (i.e., loads at .32 or greater on two or more factors[58]), or did not load to a factor were explored to determine the conceptual importance of the item, or if inadequately written prior to making a decision whether or not to drop items from further analysis. No items were removed as a result of these inspections. Next, subscale internal consistency and alpha-if-item-deleted were examined, as well as correlations between subscale means for potential multicollinearity (i.e., high correlations between independent variables; r > .75). An analysis of internal consistency (Cronbach’s alpha) was completed for the total scale and each of the resulting subscale scores. A Cronbach’s alpha of .70 or greater was considered acceptable. Items that increased the subscale alpha were removed one-by-one from scales that were highly correlated, keeping a minimum of three items that loaded on a factor. Before removal, each item was reviewed (JN, KH, DW) for its potential practical and theoretical implications. In total, 24 items were removed and a final set of 16 items were retained for the final analysis.

To confirm the EFA, a CFA was conducted using EQS 6.1 (Multivariate Software, Encino, CA, USA)[59], in which data were fit to a covariance matrix, using the robust estimation parameter. Fit indices used were the root mean squared error of approximation (RMSEA), and Bentler’s comparative fit index (CFI). Good model fit was indicated by CFI > .95 and RMSEA < .10. A criterion of .40 was set as an acceptable standardized factor loading.

In addition, backward linear regression was employed to determine which factor or combination of factors were predictive of engagement. The dependent variable was global engagement and the independent variables were the other resulting subscales from the EFA. One-way multivariate analysis of variance (MANOVA) was used to examine between-group differences of the resulting subscales. We wanted to determine if those who were higher in the governance structure of the SCNs (leaders, formal SCN members) would differ in their levels of engagement compared to those who were more removed from the work of the SCNs (support staff, frontline healthcare stakeholders). We used Games-Howell post-hoc tests as group sizes differed and equal variances were not assumed.[60] Effect sizes were classified by Cohen’s criterion.[61]

Ethical considerations

Ethical approval was granted by the University of Calgary Conjoint Health Research Ethics Board. Submitting the online survey implied informed consent by respondents. Participants created their own unique identifier for anonymity, and participants’ characteristics have been aggregated to ensure that individuals are not identifiable.


Descriptive characteristics of sample

From a total of 1,668 eligible participants within the networks, 424 individuals agreed to participate (25.4% response rate). Participant characteristics are detailed in Table 1. Most respondents were SCN members (55.6%), female (66.7%), older than 50 years (53.3%), and had 25 or more years of professional experience (50.7%). Nearly three quarters of the sample were professional healthcare providers (72.3%), and respondents worked in a wide variety of areas and positions.

Data handling–addressing not applicable responses

Not applicable responses displayed a pattern across items, whereby the proportion of responses increased with items that intended to measure higher levels of engagement (i.e., empower; see Table A in S1 File) and varied significantly by governance level (see Table C in S1 File). Post hoc z tests indicated that the proportion of not applicable responses were significantly higher in support staff and stakeholders in comparison to leaders and members (p < .025). We viewed these as important patterns, and subsequently assessed the not applicable data through various methods: (a) removal, (b) recoded to bottom of scale, (c) recoded to middle of scale, (d) imputed with the expectation-maximization (EM) algorithm, and (e) none (treated as ordinal-level scale). We compared results of these various methods and decided to conduct an EFA appropriate for non-parametric data with the full sample (n = 424), without modifying or removing the not applicable response option. Successive analysis used data from which not applicable responses were removed (n = 310, due to missing data) and data were then treated as parametric, including the calculation of means (SD), the CFA to confirm the underlying factor structure from the EFA, correlation, and multiple regression.

Item distribution

Item descriptive statistics are in Table 2 (for the original 40 items, see S2 File). Means for inform items (M = 3.35–3.91) were greater than involve items (M = 3.24–3.32) and empower items (M = 2.19–2.27). The full range of responses were used for each item and less than 5% had missing data.

Exploratory factor analysis

The EFA resulted in a 4-factor solution that accounted for 85.7% of the total variance. Table 3 details the rotated factor loadings, eigenvalues, and percent of variance explained for each factor. The final 16 items clustered on four factors consistent with four of six proposed constructs: involve (4 items), empower (4 items), global engagement (4 items), and inform (4 items).

Table 3. Exploratory factor analysis of final engagement items.

Confirmatory factor analysis

Using a CFA, we tested whether the EFA four-factor solution could be replicated after removing the not applicable responses. Fig 1 illustrates the results of the CFA, which confirmed the 4-factor structure with good model fit (CFI = .96; RMSEA = .09). Standardized factor loadings were all statistically significant (p < .001) and ranged from .71 to .98. Involve was the best predictor of global engagement, with a factor loading of .56; inform and empower were less predictive, with factor loadings of .10 and .20, respectively.

Scale internal consistency

Analysis of scale internal consistency using Cronbach’s alpha revealed acceptable reliability for all four subscales, which ranged from .87 to .99 (see Table 4). The overall tool achieved a Cronbach’s alpha of .93. Correlations between subscale means were statistically significant (ps < .001) and positive.

Table 4. Cronbach’s alpha, means, standard deviations, and correlations between subscale scores.

Linear regression

Results from the CFA were further confirmed by the regression analysis. Scores from inform, involve, and empower were regressed on global engagement. Overall, the level of engagement variables explained 54% of the variance (R = .73) in global engagement, F (3,306) = 109.44, p < .001. Inform (B = .17, SE B = .06, β = .17; 95% CI [.04, .29]; p = .008), involve (B = .41, SE B = .07, β = .41; 95% CI [.27, .54]; p < .001), and empower were significant predictors (B = .26, SE B = .05, β = .425; 95% CI [.16, .37]; p < .001).

Group comparisons

Governance level had a statistically significant medium effect on engagement scores (F (12,802) = 8.61, p < .001; Wilk’s Λ = 0.726, η2 = 0.10; see Tables 5 and 6). Significant group differences were found in post-hoc comparisons for each engagement subscale, with mostly medium to large effect sizes. Across all scales, leaders exhibited significantly higher scores than stakeholders (d = 1.1 to 1.6, p < .001), while members and support staff scores did not significantly differ. Leaders had higher scores than both members and support staff for global engagement (d = 1.33, 0.90; p < .001), inform (d = 0.64, 0.64, p < .01), and involve (d = 0.73, 0.55; p < .05), respectively; leaders also had higher scores than members for empower (d = 0.63; p < .001). In the opposite direction, stakeholders had lower scores than both members and support staff for global engagement (d = 0.84, 1.02; p < .001), and inform (d = 0.81, 0.70, p < .001), involve (d = 0.71, 0.98; p < .001), and empower (d = 0.41, 1.01; p < .05), respectively.

Table 5. Means and standard deviations of engagement scores by governance level.

Table 6. Multivariate and univariate analysis of variance for engagement scores.


The objective of this study was to create a brief, multiprofessional tool to measure engagement in SCN initiatives, and establish evidence for reliability and construct validity of the tool. We created items based on the IAP2 spectrum of engagement[36] and included the input of experts and clinicians. From 16 items, four distinct subscales were established through the EFA and CFA: (1) global engagement, (2) inform, (3) involve, and (4) empower; all of the subscales demonstrated acceptable reliability. Inform, involve, and empower were all significant predictors of global engagement, but both the CFA and regression analysis demonstrated that involve was the strongest predictor. Leaders exhibited significantly higher scores across all scales than stakeholders, while members and support staff did not differ in their scores. In sum, we have established preliminary psychometric evidence of this engagement tool for use with SCNs.

The EFA supported three of the five levels within the IAP2 spectrum.[35] Similarly, other models of community engagement have three levels, spanning from concepts that represent one-way information sharing, two-way communication, to shared decision making.[6264] Although we used direct language from the IAP2 spectrum in developing the items, several members from the expert and clinician panel expressed that there was little differentiation between involve and collaborate items. This view has been further supported by a critic of the IAP2 who contested that the three middle levels of the spectrum (consult, involve, collaborate) are all “an invitation to selected stakeholders to participate in joint decision making, including the design of the process itself.”[p. 1; 65] As factors loaded by IAP2 levels, our analysis did not support distinctions between four implementation phases: setting priorities, planning, implementing, and evaluating. In the overall sample and governance level groups, there were minimal differences between phase-specific item totals, indicating that participants viewed engagement as interactions with the SCNs instead of involvement in discrete phases. A number of healthcare improvement models would support this, by suggesting that stakeholders should be involved right from the beginning of any innovation project through to sustaining and disseminating project findings.[16,1921]

While participants had higher inform scores than involve and empower scores, both the CFA and regression analysis identified that participants felt more engaged when they were involved. Used alone, informing stakeholders is an insufficient strategy if engagement is the desired outcome. One-way dissemination of information has been regarded as passive participation in other models of participation, as there are no assurances that stakeholders’ views on an issue will be considered without the venue for those views to be expressed.[42] One could theorize that being informed is a prerequisite to being involved or participating in final decision making; however, further research needs to be conducted to determine if inform is an empirical precursor to involve.

Based on both the not applicable responses and group comparisons, healthcare professionals viewed engagement differently depending on the governance level within which they work. Not surprisingly, SCN leaders at the top of the hierarchy and accountability structures, and often the initiators of engagement, were the most engaged group. This was in direct comparison to the frontline stakeholders who felt much less engaged, with mean scores ranging from 1.74 to 2.69. Engaging frontline healthcare professionals in quality improvement has been an issue that remains concerning for health systems globally,[2,23,6668] Interestingly, SCN members and support staff who work more peripherally with the SCNs did not differ in their engagement scores, although support staff more often reported that items were not applicable, particularly empower items.

This tool can provide a mechanism to quickly assess the dimensions of engagement in clinical networks, and to help networks evaluate the intended results of engagement efforts. Before use, however, organizations are encouraged to pilot test the questionnaire, as context between locations may differ thus yielding different results due to missing questions (referred to as construct underrepresentation). This would allow for context specific questions to be developed in order to best capture engagement. Results of this study suggest that we need to further explore engagement in targeted stakeholder groups and over time. Moreover, further study is required to determine the organizational and individual barriers facing particular groups. This could lead to mapping more effective strategies to ameliorate low engagement in targeted groups (e.g., opinion leaders, champions, targeted messaging, organizational interventions, education, financial incentives). Based on the moderate-to-low engagement scores of frontline stakeholders, the direction and strategies to enhance practitioner engagement should be reconsidered within this health system context. Hess and colleagues[69] suggest that engagement strategies that are embedded in the cultural context of the organization, that enhance interactivity among team members, and that build social learning spaces and processes (audit and feedback) offer opportunities for teams to meet their goals and have ownership of improvements. For SCN leaders, this will require examining existing communication and knowledge pathways between the SCN members (clinicians, patient representatives, leaders, operational leaders), support staff, and frontline end-users to formally and informally exchange ideas for improvement of implementation of existing SCN initiatives, as well as inform development of other initiatives where practice gaps exist.


This study has several limitations. First, we acknowledge the potential for response bias from the self-report survey approach; however, we believe that social desirability was unlikely given the range of item scores and means exhibited. Second, we used a self-selected convenience sample and were not able to compare responders to non-responders, which may have led to significant selection bias. Our intention was to sample diverse healthcare stakeholders—from frontline providers to the executive teams. We know from a comprehensive scoping review (in progress) that these networks globally connect many relevant stakeholder groups; indeed, this is the a key aim of these networks globally. However, our study included stakeholdersconnected to nine SCNs undertaking specific project in Alberta, Canada. This limits the generalizability of our findings to other SCNs, healthcare organizations,, activities, and low- and middle-income countries. We also recognize that further work needs to be done to confirm our findings outside of clinical networks, and with larger samples, to replicate the pattern of engagement across governance levels and professional groups. Moreover, measurement of engagement has to take into consideration the degree of maturation of the network. In the SCNs studied, there were variations in their length of operation and degree of maturation. Lastly, the not applicable response option may have been interpreted in various ways by the participants, including expectations around engagement (“I did not expect to be engaged, so rated not applicable”) or very recent/new involvement with the networks (“I do not know enough about what has been going on, so rated not applicable).


To conclude, this clinical networks engagement tool demonstrates preliminary evidence of construct validity and reliability. In further work, we propose to assess engagement in broader network activities beyond that of discrete projects, as well as evaluating the factors that influence engagement and how the maturity of networks factors into engagement.

Supporting information

S1 File. Original Item Descriptive Statistics and Exploratory Factor Analysis, and Item Not Applicable Responses by Governance Level.


S2 File. Final Clinical Networks Engagement Tool.



We thank Tracy Wasylak (Alberta Health Services) for her guidance and involvement as key knowledge user throughout the research project. We are also grateful to the expert and clinician panels who helped develop the items, as well as members of the Patient Safety and Quality of Care Research Group in the Faculty of Nursing, University of Calgary.

Author Contributions

  1. Conceptualization: JMN KGH DEW TWN.
  2. Formal analysis: JMN KGH LR.
  3. Funding acquisition: JMN KGH DEW TWN.
  4. Investigation: JMN KGH LR DEW.
  5. Methodology: JMN KGH DEW TWN.
  6. Project administration: DEW LR.
  7. Resources: DEW KGH.
  8. Supervision: JMN KGH DEW.
  9. Writing – original draft: JMN.
  10. Writing – review & editing: JMN KGH DEW LR TWN.


  1. 1. Chaudoir SR, Dugan AG, Barr CH. Measuring factors affecting implementation of health innovations: a systematic review of structural, organizational, provider, patient, and innovation level measures. Implement Sci. 2013;8: 22. pmid:23414420
  2. 2. Greenhalgh T, Robert G, Bate P, Kyriakidou O, Macfarlane F, Peacock R. How to spread good ideas: A systematic review of the literature on diffusion, dissemination and sustainability of innovations in health service delivery and organisation. Report for the National Co-ordinating Centre for NHS Service Delivery and Organisation R & D (NCCSDO). 2004. Available:
  3. 3. Spurgeon P, Barwell F, Mazelan P. Developing a medical engagement scale (MES). Int J Clin Leadersh. 2008;16: 213–223.
  4. 4. Best A, Greenhalgh T, Lewis S, Saul JE, Carroll S, Bitz J. Large-system transformation in health care: a realist review. Milbank Q. 2012;90(3): 421–456. pmid:22985277
  5. 5. Dickinson H, Ham C. Engaging doctors in leadership: review of the literature. 2008. Available:
  6. 6. Reinertsen JL, Gosfield AG, Rupp W, Whittington JW. Engaging physicians in a shared quality agenda: IHI Innovation Series white paper. 2007. Available:
  7. 7. West M, Dawson J, Admasachew L, Topakas A. NHS staff management and health service quality: results from the NHS staff survey and related data. 2011. Available:
  8. 8. Spence Laschinger HK, Leiter MP. The impact of nursing work environments on patient safety outcomes: the mediating role of burnout/engagement. J Nurs Adm. 2006;36(5): 259–267. pmid:16705307
  9. 9. Sikorska-Simmons E. Linking resident satisfaction to staff perceptions of the work environment in assisted living: a multilevel analysis. Gerontologist. 2006;46(5): 590–598. pmid:17050750
  10. 10. Collins KS, Collins SK, McKinnies R, Jensen S. Employee satisfaction and employee retention: catalysts to patient satisfaction. Health Care Manag. 2008;27(3): 245–251.
  11. 11. Taitz JM, Lee TH, Sequist TD. A framework for engaging physicians in quality and safety. BMJ Qual Saf. 2012;21(9): 722–728. pmid:21757468
  12. 12. Kaissi A. Enhancing physician engagement: an international perspective. Int J Health Serv. 2014;44(3): 567–592. pmid:25618990
  13. 13. Ocloo J, Matthews R. From tokenism to empowerment: progressing patient and public involvement in healthcare improvement. BMJ Qual Saf. 2016;25(8): 626–632. pmid:26993640
  14. 14. Sarrami-Foroushani P, Travaglia J, Debono D, Braithwaite J. Implementing strategies in consumer and community engagement in health care: results of a large-scale, scoping meta-review. BMC Health Serv Res. 2014;14(1): 402.
  15. 15. Snyder H, Engstrom J. The antecedents, forms and consequences of patient involvement: a narrative review of the literature. Int J Nurs Stud. 2016;53: 351–378. pmid:26602069
  16. 16. Damschroder LJ, Aron DC, Keith RE, Kirsh SR, Alexander JA, Lowery JC. Fostering implementation of health services research findings into practice: a consolidated framework for advancing implementation science. Implement Sci. 2009;4: 50. pmid:19664226
  17. 17. Harvey G, Kitson A. PARIHS revisited: from heuristic to integrated framework for the successful implementation of knowledge into practice. Implement Sci. 2016;11(1): 1–13.
  18. 18. Haines M, Brown B, Craig J, D'Este C, Elliott E, Klineberg E, et al. Determinants of successful clinical networks: the conceptual framework and study protocol. Implement Sci. 2012;7(16): 16.
  19. 19. Meyers DC, Durlak JA, Wandersman A. The quality implementation framework: a synthesis of critical steps in the implementation process. Am J Community Psychol. 2012;50(3–4): 462–480. pmid:22644083
  20. 20. Saldana L, Chamberlain P, Wang W, Hendricks Brown C. Predicting program start-up using the stages of implementation measure. Adm Policy Ment Health. 2012;39(6): 419–425. pmid:21710257
  21. 21. Canadian Institutes of Health Research. About knowledge translation. 2015. Available:
  22. 22. Stetler CB, Mittman BS, Francis J. Overview of the VA Quality Enhancement Research Initiative (QUERI) and QUERI theme articles: QUERI Series. Implement Sci. 2008;3: 8. pmid:18279503
  23. 23. Davies H, Powell A, Rushmer R. Healthcare professionals’ views on clinician engagement in quality improvemen: A literature review. 2007. Available:
  24. 24. Dixon-Woods M, McNicol S, Martin G. Ten challenges in improving quality in healthcare: lessons from the Health Foundation's programme evaluations and relevant literature. BMJ Quality & Safety. 2012;21(10): 876–884.
  25. 25. Ferlie EB, Shortell SM. Improving the quality of health care in the United Kingdom and the United States: a framework for change. Milbank Q. 2001;79(2): 281–315. pmid:11439467
  26. 26. White M, Wells JS, Butterworth T. The impact of a large-scale quality improvement programme on work engagement: preliminary results from a national cross-sectional-survey of the 'Productive Ward'. Int J Nurs Stud. 2014;51(12): 1634–1643. pmid:24890896
  27. 27. Alexander GC, Lin S, Sayla MA, Wynia MK. Development of a measure of physician engagement in addressing racial and ethnic health care disparities. Health Serv Res. 2008;43(2): 773–784. pmid:18370978
  28. 28. Bailey C, Madden A, Alfes K, Fletcher L. The meaning, antecedents and outcomes of employee engagement: a narrative synthesis. Int J Manage Rev. 2015;19: 1–23.
  29. 29. Shuck B, Wollard K. Employee engagement and HRD: a seminal review of the foundations. Hum Resource Dev Rev. 2010;9(1): 89–110.
  30. 30. Saks A, Gruman J. What do we really know about employee engagement? Human Resource Dev Q. 2014;25(2): 155–182.
  31. 31. Macey WH, Schneider B. The meaning of employee engagement. Ind Org Psy. 2008;1: 3–30.
  32. 32. Demerouti E, Mostert K, Bakker AB. Burnout and work engagement: a thorough investigation of the independency of both constructs. J Occup Health Psychol. 2010;15(3): 209. pmid:20604629
  33. 33. Harter JK, Schmidt FL, Keyes CLM. Well-being in the workplace and its relationship to business outcomes: a review of the Gallup studies. In: Keyes CL, Haidt J, editors. Flourishing: the positive person and the good life. Washington, DC: American Psychological Association 2003. pp. 205–224.
  34. 34. Maslach C, Jackson SE, Leiter MP. Maslach burnout inventory: Manual. Palo Alto, CA: Consulting Psychologists Press;
  35. 35. Schaufeli WB, Bakker AB. Utrecht work engagement scale: preliminary manual. Utrecht, NL: Occupational Health Psychology Unit, Utrecht University; 2003.
  36. 36. International Association for Public Participation. IAP2 Spectrum of Public Participation. 2007. Available:
  37. 37. Metro South Health. Metro South Health Clinician Engagement Strategy 2012–2015. 2012. Available:
  38. 38. Kovacs Burns K, Bellows M, Eigenseher C, Gallivan J. 'Practical' resources to support patient and family engagement in healthcare decisions: a scoping review. BMC Health Serv Res. 2014;14(1): 175.
  39. 39. Arnstein SR. A ladder of citizen participation. J Am Inst Plann. 1969;35(4): 216–224.
  40. 40. Connor DM. A new ladder of citizen participation. National Civic Review. 1988;77(3): 249–257.
  41. 41. Pretty JN. Participatory learning for sustainable agriculture. World Devel. 1995;23(8): 1247–1263.
  42. 42. Cornwall A. Unpacking 'Participation': models, meanings and practices. Community Dev J. 2008;43(3): 269–283.
  43. 43. Hamilton K, Sullivan F, Donnan P, Taylor R, Ikenwilo D, Scott A, et al. A managed clinical network for cardiac services: set-up, operation and impact on patient care. Int J Integr Care. 2005;5(3).
  44. 44. Spencer A, Ewing C, Cropper S. Making sense of strategic clinical networks. Arch Dis Child. 2013;98(11): 843–845. pmid:23887390
  45. 45. McInnes E, Middleton S, Gardner G, Haines M, Haertsch M, Paul CL, et al. A qualitative study of stakeholder views of the conditions for and outcomes of successful clinical networks. BMC Health Serv Res. 2012;12(1): 49.
  46. 46. Goodwin N, Peck E, Freeman T, Posaner R. Managing across diverse network of care: lessons from other sectors. 2004. Available:
  47. 47. Williams MJ, Harding SA, Devlin G, Nunn C, El-Jack S, Scott T, et al. National variation in coronary angiography rates and timing after an acute coronary syndrome in New Zealand (ANZACS-QI 6). N Z Med J. 2016;129(1428): 66–78. pmid:26914194
  48. 48. O’Hare B, Phiri A, Lang H-J, Friesen H, Kennedy N, Kawaza K, et al. Task sharing within a managed clinical network to improve child health in Malawi. Human Resources for Health. 2015;13(1): 1–5.
  49. 49. Conway NT, Ai Wotayan R, Alkuzam A, Al-Refaei FF, Badawi D, Barake R, et al. The Kuwait-Scotland eHealth Innovation Network (KSeHIN): a sustainable approach to quality improvement in healthcare. Qual Prim Care. 2014;22(1): 43–51. pmid:24589150
  50. 50. Ferlie E, Fitzgerald L, McGivern G, Dopson S, Exworthy M. Networks in health care: a comparative study of their management, impact and performance. 2010. Available:
  51. 51. Grant JS, Davis LL. Selection and use of content experts for instrument development. Res Nurs Health. 1997;20(3): 271–274.
  52. 52. Dillman DA, Smyth JD, Christian LM. Internet, mail, and mixed-mode surveys: the tailored design method. Hoboken, NJ: John Wiley & Sons; 2008.
  53. 53. Bryant FB, Yarnold PR. Principal-components analysis and exploratory and confirmatory factor analysis. In: Grimm LG, Yarnold PR, editors. Reading and understanding multivariate statistics. Washington, DC: American Psychological Association 1995. pp. 99–136.
  54. 54. Gorsuch RL. Factor analysis. Hillsdale, NJ: Erlbaum; 1983.
  55. 55. Bentler PM, Yuan KH. Structural equation modeling with small samples: test statistics. Multivariate behavioral research. 1999;34(2): 181–197. pmid:26753935
  56. 56. Bollen KA. Structural equations with latent variables. New York, NY: Wiley; 1989.
  57. 57. American Educational Research Association, American Psychological Association, National Council on Measurement in Education. Standards for educational and psychological testing. Washington, DC: American Educational Research Association; 2002.
  58. 58. Tabachnick BG, Fidell LS. Using multivariate statistics. Boston, MA: Allyn and Bacon; 2007.
  59. 59. Bentler PM. EQS 6 structural equations program manual. Encino, CA: Multivariate Software; 2004.
  60. 60. Field AP. Discovering statistics using IBM SPSS Statistics. London, UK: Sage; 2013.
  61. 61. Cohen J. Quantitative methods in psychology: a power primer. Psychol Bull. 1992;112(1): 115–119.
  62. 62. Gregory J. Engaging consumers in discussion about Australian health policy: key themes emerging from the AIHPS study. 2008. Available:
  63. 63. Carman KL, Dardess P, Maurer M, Sofaer S, Adams K, Bechtel C, et al. Patient and family engagement: a framework for understanding the elements and developing interventions and policies. Health Aff (Millwood). 2013;32(2): 223–231.
  64. 64. Bowen F, Newenham-Kahindi A, Herremans I. When suits meets roots: the antecedents and consequences of community engagement strategy. J Bus Ethics. 2010;95: 297–318.
  65. 65. Carson L. The IAP2 Spectrum: Larry Susskind in conversation with IAP2 members. Int J Public Particip. 2008;2(2): 67–84.
  66. 66. Clarke AL, Shearer W, McMillan AJ, Ireland PD. Investigating apparent variation in quality of care: the critical role of clinician engagement. Med J Aust. 2010;193(8 Suppl): S111–113. pmid:20955138
  67. 67. Detwiller M, Petillion W. Change management and clinical engagement: critical elements for a successful clinical information system implementation. Comput Inform Nurs. 2014;32(6): 267–273. pmid:24651254
  68. 68. Sawka C, Ross J, Srigley J, Irish J. The crucial role of clinician engagement in system-wide quality improvement: the Cancer Care Ontario experience. Healthc Q. 2012;15: 38–41.
  69. 69. Hess DW, Reed VA, Turco MG, Parboosingh JT, Bernstein HH. Enhancing provider engagement in practice improvement: a conceptual framework. J Contin Educ Health Prof. 2015;35(1): 71–79. pmid:25799975