Skip to main content
Advertisement
Browse Subject Areas
?

Click through the PLOS taxonomy to find articles in your field.

For more information about PLOS Subject Areas, click here.

  • Loading metrics

Measuring partnership synergy and functioning: Multi-stakeholder collaboration in primary health care

  • Ekaterina Loban ,

    Contributed equally to this work with: Ekaterina Loban, Cathie Scott, Virginia Lewis, Jeannie Haggerty

    Roles Conceptualization, Data curation, Formal analysis, Investigation, Methodology, Project administration, Writing – original draft, Writing – review & editing

    ekaterina.loban@mail.mcgill.ca

    Affiliations St. Mary’s Research Centre, Montreal, Quebec, Canada, Department of Family Medicine, McGill University, Montreal, Quebec, Canada

  • Cathie Scott ,

    Contributed equally to this work with: Ekaterina Loban, Cathie Scott, Virginia Lewis, Jeannie Haggerty

    Roles Conceptualization, Formal analysis, Supervision, Writing – review & editing

    Affiliation Department of Community Health Sciences, University of Calgary, Alberta, Canada

  • Virginia Lewis ,

    Contributed equally to this work with: Ekaterina Loban, Cathie Scott, Virginia Lewis, Jeannie Haggerty

    Roles Conceptualization, Supervision, Writing – review & editing

    Affiliation Australian Institute for Primary Care & Ageing, La Trobe University, Melbourne, Australia

  • Jeannie Haggerty

    Contributed equally to this work with: Ekaterina Loban, Cathie Scott, Virginia Lewis, Jeannie Haggerty

    Roles Conceptualization, Formal analysis, Methodology, Supervision, Writing – review & editing

    Affiliations St. Mary’s Research Centre, Montreal, Quebec, Canada, Department of Family Medicine, McGill University, Montreal, Quebec, Canada

Abstract

In primary health care, multi-stakeholder partnerships between clinicians, policy makers, academic representatives and other stakeholders to improve service delivery are becoming more common. Literature on processes and approaches that enhance partnership effectiveness is growing. However, evidence on the performance of the measures of partnership functioning and the achievement of desired outcomes is still limited, due to the field’s definitional ambiguity and the challenges inherent in measuring complex and evolving collaborative processes. Reliable measures are needed for external or self-assessment of partnership functioning, as intermediate steps in the achievement of desired outcomes. We adapted the Partnership Self-Assessment Tool (PSAT) and distributed it to multiple stakeholders within five partnerships in Canada and Australia. The instrument contained a number of partnership functioning sub-scales. New sub-scales were developed for the domains of communication and external environment. Partnership synergy was assessed using modified Partnership Synergy Processes and Partnership Synergy Outcomes sub-scales, and a combined Partnership Synergy scale. Ranking by partnership scores was compared with independent ranks based on a qualitative evaluation of the partnerships’ development. 55 (90%) questionnaires were returned. Our results indicate that the instrument was capable of discriminating between different levels of dimensions of partnership functioning and partnership synergy even in a limited sample. The sub-scales were sufficiently reliable to have the capacity to discriminate between individuals, and between partnerships. There was negligible difference in the correlations between different partnership functioning dimensions and Partnership Synergy sub-scales. The Communication and External Environment sub-scales did not perform well metrically. The adapted partnership assessment tool is suitable for assessing the achievement of partnership synergy and specific indicators of partnership functioning. Further development of Communication and External Environment sub-scales is warranted. The instrument could be applied to assess internal partnership performance on key indicators across settings, in order to determine if the collaborative process is working well.

Introduction

Multi-stakeholder partnerships have been defined as a “voluntary and collaborative relationship … between various [stakeholders from different organizations] … in which all participants agree to work together to achieve a common purpose or undertake a specific task and to share risks, responsibilities, resources, competencies and benefits” [1]. Partnerships can help people and organizations generate outcomes that are greater than those that can be achieved working independently [24]. Partnerships continue to be widely embraced with governments and funding bodies mandating partnerships as an essential element of the programs and initiatives that they support [3,5,6].

In practice, however, partnerships can generate a great deal of frustration, as different organizations and individuals frequently struggle to achieve measurable and mutually beneficial outcomes [3,5,79]. These problems are not surprising given that partnerships are resource-intensive, take time and require governance, procedures and processes that are very different from how independent organizations are run [3]. Measures of partnership functioning and partnership self-assessment tools can help organizations assess how the partnership is evolving, benchmark against external criteria, stay accountable to the partnership’s stakeholders and funders, and monitor and maximize the effectiveness of the collaborative effort [6,1012].”

This article presents findings from a survey of five multi-stakeholder partnerships involving decision makers, academic representatives, clinicians and organizational representatives with an interest in improving primary health care (PHC) accessibility. This quantitative study was part of a larger mixed methods sequential exploratory study to understand the processes that enhance multi-stakeholder partnerships in PHC service delivery. The qualitative phase entailed a comparative qualitative case study exploring, in a smaller sample of partnerships, the different manifestations of partnership synergy, different types of partnership resources and broad categories of partnership processes relevant to multi-stakeholder partnerships in PHC. Based on qualitative findings, approximately three-to-four years into the partnerships, we adapted an existing quantitative tool to measure different functions of partnerships and specifically to measure the concept of partnership synergy as an intermediate measure of partnership effectiveness. This article reports on the metric performance of the adapted measures of partnership functioning and partnership synergy.

Partnership evaluation and synergy

Some scholars and practitioners have argued that various dimensions of partnership functioning—including partner relationships, governance, leadership, management, use of resources, and the external environment–influence the achievement of the planned partnership outcomes [3,13,14]. The literature offers different definitions and conceptualizations of partnership effectiveness, and complex processes and dynamics are inherently difficult to evaluate [9,1518].

The notion of “partnership synergy” has recently been proposed as an intermediate outcome between partnership functioning and the achievement of the planned ultimate outcomes [3]. Synergy is viewed as the key mechanism on the pathway by which partnerships gain advantage over the independent work of individual stakeholders working towards the same goals [3]. Partnerships are said to be synergistic when they leverage resources successfully and mobilize the complementary knowledge and expertise of all partners, with the effect that the whole becomes greater than the sum of the parts [2,3]. The assumption is that if a partnership achieves a high level of synergy, it is better placed to achieve desired partnership goals. The advantage of measuring synergy is that it is manifested (or not) earlier and is more directly attributed to partnership functioning than the ultimate outcomes [3].

Two validated scales of partnership synergy have been developed. Both scales purport to assess the degree to which the partnership has maximized its synergy but each measures partnership synergy differently. The Weiss et al. scale [19] conceptualizes synergy as a product of good quality partnership processes [3], whereas the Jones synergy scale measures synergy as both a partnership process and a partnership product [20]. Both synergy scales are embedded in instruments that include other measures of different partnership functioning dimensions [7,19,21].

Partnership Self-Assessment Tool (PSAT)

We selected as the study instrument the Partnership Self-Assessment Tool (PSAT) [22] which contains the Weiss et al. partnership synergy scale. The PSAT also measures the various dimensions of partnership functioning including leadership, efficiency, administration and management, non-financial resources, financial and other capital resources, decision-making, benefits and drawbacks of participation and overall satisfaction with participation [22]. The PSAT was developed by public health scholars for use by groups working to promote health and well-being in their communities [19,23,24]. Although very similar to another partnership instrument that integrated the Weiss et al. and the Jones partnership synergy scales (Partner Questionnaire) [20], we primarily used the PSAT, but integrated some of the nuances from the Partner Questionnaire. The PSAT is in the public domain and no conditions are specified for its use [25]. The PSAT was “provided to partnerships at no charge by the Center for the Advancement of Collaborative Strategies in Health at The New York Academy of Medicine with funding from the W.K. Kellogg Foundation” [23].

We made some adaptations to the PSAT. First, we modified the Partnership Synergy scale, incorporating items from the Weiss et al. and the Jones partnership synergy scales [20,22]. Second, we developed new measures of communication and external environment and we adapted the content of other sub-scales based on the findings of the comparative case study of two of the partnerships. The objective of this study was to assess the reliability of the adapted PSAT instrument and sub-scales, specifically the measures of partnership synergy, to assess its capacity to discriminate between partnerships, and to provide conclusions that were coherent with the independent qualitative developmental evaluation that ran in parallel to the mixed methods study.

Methods

We used a cross-sectional survey design, distributing a self-administered questionnaire to multiple stakeholders within five collaborative local partnerships in three Canadian provinces (Quebec, Ontario and Alberta) and two Australian states (Victoria and New South Wales). The survey took place between September 2018 and March 2019. Ethics approval for the study was obtained from the St. Mary’s Hospital Center Research Ethics Committee (No. SMHC-13-30C). All participants provided consent prior to data collection.

Study population and sampling

The five partnerships were established in 2014 within a Canada-Australia research program entitled “Innovative Models Promoting Access-to-Care Transformation” (IMPACT). The program aimed at implementing organizational interventions to improve access to appropriate PHC for vulnerable populations [26]. Using a participatory action research approach, each partnership identified local PHC access need, and selected, adapted, implemented and evaluated an organizational intervention to address the need. The ultimate outcome was to meet the health care needs of the specific population through PHC and avoid unnecessary emergency room use and hospitalizations.

Each partnership involved a combination of the following stakeholders: policy makers, clinicians and other health and community service providers, health system administrators, university researchers and research coordinators, and, in some cases, members and representatives of vulnerable populations [26]. The premise of the program was that a strong partnership was key to mobilizing the resources for the pilot intervention and ultimately sustaining it, if proved potentially effective. Each IMPACT local partnership was unique, based upon the specific context within which it unfolded, the local access need that the partnership tried to address, tailored processes and requirements that were required to meet this need, and the relationships that formed to move the work forward. As a result, the interventions in each partnership were different both in focus and in the ways in which the partnerships evolved. See Table 1 for an overview of interventions in five partnerships.

thumbnail
Table 1. Overview of interventions in five partnerships [adapted from 26,27].

https://doi.org/10.1371/journal.pone.0252299.t001

The partnerships met the following criteria for the PSAT use: the partnerships had been in existence for at least six months, had at least five active partners, had continually worked together to develop and modify strategies in order to achieve their goals, and had begun to implement their plans [23]. We used a census sample of all members of the five local partnerships at the time of data collection. All 61 stakeholders who were active in the partnerships at the time of administration were considered eligible and were invited to participate.

Instrument adaptation and administration

We adapted the validated Partnership Self-Assessment Tool (PSAT) [22,24] as outlined in supplementary material S1 Appendix. First, we made small linguistic adaptations to fit the PHC study context, by reflecting the roles and organizational expertise of the stakeholders in PHC partnerships. Second, we adapted the PSAT based on analysis of the case studies in two partnerships with the objective of understanding and describing how partnership processes and approaches to enhance partnerships were perceived and experienced by local partnership stakeholders. We selected PSAT sub-scales and items that corresponded to qualitative codes from the case study and excluded items that were deemed not relevant. In addition, we supplemented synergy items from the PSAT with synergy items from the Partnership Questionnaire [20] to capture more information on partnership processes. The adapted instrument retained the following dimensions of partnership functioning from the PSAT: Decision-making (four items), Leadership (11 items), Administration & Management (11 items), Non-financial Resources (six items); Financial & Other Resources (three items); and Resource Utilization (three items, referred to as “Efficiency” in the PSAT).

We added elements that emerged in qualitative data but were not part of the PSAT: new proposed sub-scales for Communication (three items) and External Environment (two items). The case studies spoke to the importance of communication as an integral dimension of partnership functioning and the critical role of external influences.

Partnership Synergy was assessed using two sub-scales: 1) the adapted Partnership Synergy Processes sub-scale incorporating five items from the eight-item synergy scale developed by Jones and Barry [20]; and 2) the adapted Partnership Synergy Outcomes sub-scale retaining two items from the nine-item synergy scale by Weiss et al [19]. The items from these original scales were selected based upon the relevance to the types of partnerships highlighted throughout the qualitative inquiry, and with a view to reduce participant burden. More items were retained from the Jones than the Weiss et al. synergy scale as they were deemed more relevant in light of our overall aim to understand the processes and approaches of successful partnerships. In addition, they were more appropriate in our sample of partnerships that were underway and that exhibited varied levels of engagement of community-based stakeholders representing the target population. The partnerships were at a stage in their development when it was premature or impossible to evaluate elements of the Weiss et al. scale, such as the ability “to respond to the needs and problems of the community” or “to implement strategies that are most likely to work in the community” [22]. Respondents were also asked to assess the extent to which the goal of developing a meaningful partnership had been achieved (1 item).

The majority of questions offered five-point Likert response options. We changed the order of response options to present the most negative response first, in order to reduce the commonly observed positive bias in responding [28]. The use of the “don’t know” response option was minimized, to reduce the occurrence of missing data and on the grounds that there was little reason to suppose that respondents would not know. The instrument captured the stakeholders’ assessments of the partnerships: some questions elicited the individual stakeholders’ own perceptions; others, the perspectives of the stakeholders’ respective organizations. The respondents were asked to provide an assessment of the partnership at the point in time when the tool was administered.

Finally, the questionnaire included questions about perceived Benefits (11 items) and Drawbacks (5 items) of participation to their respective organizations, and an overall assessment of benefits compared to drawbacks (1 item). It also elicited descriptive information about the stakeholder (type, of role in the organization, length of time in the partnerships, frequency of engagement).

The questionnaire was tested through cognitive interviewing [29] with three professionals familiar with PHC terminology, to verify that the content of the questions was understood as intended, to select the most appropriate response scales, and to ascertain whether the language, the length and format were appropriate. Modifications were mostly confined to amending response scales and choosing the appropriate terminology in French. The translation into French was performed by the first author (EL) and verified by a member of the research team familiar with the terminology and the IMPACT program context.

The questionnaire was self-administered according to existing guidelines in the survey literature [28]. Both paper and web-based questionnaires were offered, and in either English or French. The electronic questionnaire was hosted on the Qualtrics [30] online survey platform.

Data analysis

The unit of analysis for analysing the metric properties of the instrument was the individual respondent (n = 54), whereas the unit of analysis for partnership functioning and synergy was the aggregate of individual responses within a given partnership (n = 5). We preferred the median and IQR as the most appropriate way of representing central tendency and spread given the categorical nature of the data.

We used descriptive statistics to characterize the sample and the distribution of the items and sub-scale scores. The scores for each sub-scale were derived through unweighted mean and medians of all the component items; scores for a respondent were calculated only if at least 50% of the items in the scale had been completed. Another method for handling the missing data was pairwise deletion. The small sample size precluded classical factor analysis. Instead, the analysis focused on examining some of the metric properties that may compromise precision and reliability. We looked for floor and ceiling effects, using a threshold ≥ 45% respondents in the lowest or highest response option. These items were tagged for exclusion as they indicated reduced capacity to discriminate between respondents. We used Spearman correlation coefficients to estimate the correlation between items within purported sub-scales, and between sub-scales, specifically dimensions of partnership functioning and partnership synergy. We examined internal consistency of the purported sub-scales by calculating the Cronbach’s Alpha.

At the partnership level, the score for each scale was calculated by aggregating the scale scores of the component member respondents within each partnership. For partnership synergy, the two sub-scales of Partnership Synergy Processes and Partnership Synergy Outcomes were analysed separately, at the level of individual respondents. However, given that they were highly correlated with each other and demonstrated similar correlations with dimensions of partnership functioning, both sub-scales were combined at the level of the partnerships to provide a single score of Partnership Synergy. We used the non-parametric Kruskal–Wallis test to determine if we were able to detect statistically significant differences in sub-scale scores between any of the partnerships. The ranks of the scores were also derived. SPSS 23 for Windows [31] was used for data analyses.

In lieu of statistical testing of predictive and convergent validity, we used consensus ranking from two external content reviewers who had conducted the independent longitudinal developmental evaluation of the five IMPACT partnerships as part of the IMPACT program’s effort to support the partnerships. They ranked the five partnerships based on the operational definitions and item content of the Partnership Synergy sub-scales and of the following partnership functioning sub-scales: Communication, Decision-making, Problem-solving, Resource Utilization and External Environment. The qualitative and quantitative ranks were compared. Finally, we correlated the score of Partnership Synergy with the partnership aggregate assessment of the extent to which developing a meaningful partnership had been achieved.

Results

Sample characteristics

The response rate was 90% (55/61), of which 62% (34) were online and 38% (21) were in paper format; 82% (45) in English and 18% (10) in French. One respondent was subsequently excluded from analysis due to comments on the questionnaire admitting that they picked answers at random. Our final sample was therefore 54 participants (see supplementary material S1 Dataset for dataset). Of note, two partnerships, namely Primary Care Connection and Community Health Resources, participated in both phases of this mixed methods study, with 15 partnership stakeholders participating in both the qualitative phase (involving interviews) and this quantitative phase (entailing a survey).

The partnerships varied in their size and composition. Table 2 outlines the descriptive characteristics of each of the partnerships. Academic representatives constituted the largest single group of stakeholders in each partnership. Of the five partnerships, the Community Health Resources partnership had the largest representation of community stakeholders (representing community organizations and patients).

thumbnail
Table 2. Study sample characteristics per partnership (N = 5).

https://doi.org/10.1371/journal.pone.0252299.t002

Five participants did not finish completing the questionnaire for a variety of reasons (stopping at questions 14 (health care manager), 14 (academic representative), 15 (academic representative), 21 (health care manager) and 23 (patient representative) of a total of 76 questions). The missing data rates per question were calculated only on the sample of respondents who completed the questionnaire to that point. Missing data varied from 0% to 18% and the median was 2%, which was under the conventional missing rate of 10% [32]. The missing data rates higher than 10% occurred in the sub-scales of Financial Resources, Resource Utilization and External Environment, suggesting these may be problematic metrically. Ceiling effects were observed in two of the 65 Likert-scale questions in the sub-scales of Decision-making and Leadership. No floor effects were present.

Partnership functioning and synergy

Table 3 provides descriptive statistics for the dimensions of partnership functioning. We observed that the responses on all dimensions were skewed towards positive evaluations.

thumbnail
Table 3. Score distribution for the partnership functioning and partnership synergy sub-scales (N = 54).

Responses to various five-point Likert scales response options have a negative (1) to positive (5) valence.

https://doi.org/10.1371/journal.pone.0252299.t003

Table 4 presents summary information on the Spearman rank correlations between the items within various dimensions of partnership functioning and Partnership Synergy (see supplementary material S2 Appendix for table of individual correlations). We used the following cut-offs for interpreting the scores: .10 –weak correlation; .30 –moderate; .50 –strong [33]. Despite our limited sample size, we examined the correlations to shed light on a number of assumptions that relate to the validity and reliability of measures. We observed that items within the following sub-scales met the assumption of item-convergent validity by being strongly correlated (>0.50) within the sub-scale: Leadership, Administration & Management, Financial Resources and both Partnership Synergy sub-scales. The assumption of item-discriminant validity states that the correlation of items within a hypothesized sub-scale will be higher than with items in other sub-scales. We observed that correlations between items within the following sub-scales were higher than with other sub-scales: Leadership, Administration & Management, Financial Resources and both Partnership Synergy sub-scales.

thumbnail
Table 4. Spearman rank correlations between items within each of the sub-scales and between sub-scales of dimensions of partnership functioning and partnership synergy, showing the average, minimum and maximum correlation coefficients, and a measure of internal consistency (N = 54).

https://doi.org/10.1371/journal.pone.0252299.t004

Finally, reliability of a scale is assumed if each item has an approximately equivalent correlation within the hypothesized sub-scale and contributes approximately the same proportion of information to the total score. The low and variable correlations in Communication and External Environment translated into a low internal consistency that failed to meet the criterion of ≥0.70 [34]. Cronbach’s α values demonstrated good internal consistency and reliability for most other sub-scales, although the very high within-scale correlation and high internal consistencies in the Leadership, Administration & Management suggest redundancy in the items.

Table 5 displays correlations between the various dimensions of partnership functioning and Partnership Synergy, as assessed using Spearman’s rank correlation coefficient. Together, Tables 4 and 5 demonstrate that the new sub-scales of Communication and External Environment did not perform well metrically, on assumptions of item-convergent validity, item-discriminant validity, and internal consistency. We subsequently excluded these sub-scales from our partnership-level analysis.

thumbnail
Table 5. Spearman rank correlations between partnership synergy outcomes, partnership synergy processes, and the different dimensions of partnership functioning (N = 54).

Values shaded in light grey denote weak correlations, in dark grey–strong correlations.

https://doi.org/10.1371/journal.pone.0252299.t005

In addition, we observed that the shortened Partnership Synergy Outcomes and Partnership Synergy Processes sub-scales were strongly correlated with each other (r = 0.61), and there were few differences in how each of them correlated with the different dimensions of partnership functioning. We consequently chose to collapse the two synergy sub-scales together to a single Partnership Synergy scale (Partnership Synergy) for the analysis of associations between Partnership Synergy and functioning dimensions at the partnership level.

Partnership functioning and synergy at the partnership level

We aggregated the scores at the partnership level for Partnership Synergy Outcomes, Partnership Synergy Processes, Partnership Synergy and the following dimensions of partnership functioning: decision-making, leadership, administration & management, non-financial resources, financial resources and resource utilization. We then ranked the five partnerships on the basis of their aggregate scores. Fig 1 displays the comparison of the ranking based on quantitative scores with the independent, qualitative rankings by the content reviewers. The quantitative and qualitative rankings were completely coherent for Partnership Synergy Outcomes, showing that partnership synergy was highest in the Community Outreach partnership and lowest in the Service Linkage partnership. Our one-way analysis of variance was suggestive of statistically significant differences between at least two of the medians of Partnership Synergy Processes (P = .09) and Partnership Synergy (P = .07). The differences in medians were statistically significant between the following partnerships: Primary Care Connection-Service Linkage, Community Outreach-Community Health Resources, Community Outreach-Diabetes Self-management and Community Outreach-Service Linkage (in decreasing order of magnitude of difference). To further explore the extent to which differences in aggregate scores were meaningful, we calculated and found low to moderate variation for partnership synergy in all partnerships, suggesting that the statistically significant Kruskal-Wallis test of difference between at least two partnerships was not spurious. There was some variation in the content reviewers’ ranking of partnership synergy processes and partnership synergy outcomes, that was not reflected in the quantitative scores (see supplementary material S3 Appendix), suggesting that the items may not have captured nuances in the construct.

thumbnail
Fig 1. Partnership synergy outcomes (partnership-level median scores) and rankings of partnerships based on assessment by content reviewers (highest = 1 to lowest = 5) (N = 5).

https://doi.org/10.1371/journal.pone.0252299.g001

Fig 2 demonstrates that the partnership ranking was generally reflected in the scores of the partnership functioning dimensions, which was largely coherent with the results in Table 5 that show moderate to strong correlations between certain dimensions of partnership functioning and Partnership Synergy sub-scales. However, Fig 2 also displays a variation in ranking of the dimensions of partnership functioning, suggesting that partnership synergy is not simply another measure of partnership functioning.

thumbnail
Fig 2. Partnership-level scores on dimensions of partnership functioning.

Partnerships are ordered by increasing score on partnership synergy (and rank).

https://doi.org/10.1371/journal.pone.0252299.g002

Irrespective of the variations in ranking of the different dimensions of partnership functioning, the partnership-level scores for achieving the overall goal of developing a meaningful partnership were largely consistent with partnership-level partnership synergy scores. The majority of respondents in all partnerships reported that a meaningful partnership had been achieved leading to an uninformative median of 4 across all partnerships. However, as shown in Table 6, there was considerable variation within the partnerships that was reflected in the means and standard deviations. The metric that corresponds best to both partnership synergy scores and the qualitative ranking is the percentage endorsing that the partnership goal was achieved very well or extremely well.

thumbnail
Table 6. Partnership-level assessment of the extent to which the goal of developing a meaningful partnership has been achieved.

https://doi.org/10.1371/journal.pone.0252299.t006

Discussion

This study demonstrated that an adapted version of the PSAT was capable of discriminating between different levels of dimensions of partnership functioning and partnership synergy even in a limited sample of ongoing partnerships working towards improving PHC. The measures are sufficiently reliable to have the capacity to discriminate between individuals’ perceptions of the partnerships, and between partnerships based on an aggregated score. The quantitative measures could not distinguish meaningfully between Partnership Synergy Processes and Partnership Synergy Outcomes, and we proposed a single measure of Partnership Synergy. Our new sub-scales on Communication and External Environment did not perform well metrically, but could be further developed. Our findings demonstrate that the PSAT can be a valuable evaluation tool for partnerships that are underway, in that it supports identification of the areas of strength and potential improvement to achieve the collaborative advantage over organizations working on their own, even before achievement of the planned ultimate outcomes can be assessed. The tool requires further refinement and possible elimination of redundancies in order to be administered at several points in the life of partnerships in PHC transformation.

We adapted the PSAT [22] on the basis of findings from a qualitative investigation that preceded this study. Some of the items were adapted based on qualitative codes, and new items were developed for Communication and External Environment. In the 2002 version of the PSAT, item content related to communication was found in the following scales: Synergy, Leadership, and Administration & Management [22]. While we retained these items in the Leadership and Administration & Management sub-scales, we also developed a separate sub-scale for the domain of communication, for our qualitative data demonstrated the importance of communication as a stand-alone dimension of partnership functioning [35]. Clear, targeted and timely communication via a variety of communication mechanisms was deemed by qualitative study participants to be one of the key drivers in ensuring transparency and enhancing stakeholder engagement. However, the three-item sub-scale that we designed did not perform well metrically based on assessment of item-convergent validity, item-discriminant validity, or internal consistency. This may be because we used a different response approach for each item: a descriptive range for the first question on ways stakeholders were informed about what is happening in the partnership; rating of usefulness of information received; and an assessment of adequacy of quantity of information. Despite various attempts to express the nominal descriptive item in ordinal categories of increasing magnitude, the exercise did not yield satisfactory results, and even after eliminating this poorly functioning item and retaining only two items the internal consistency remained low (0.49). However, the Spearman rank correlations between all three communication items and partnership synergy were statistically significant, different from zero and strong in the Primary Care Connection partnership (0.71) and the Community Health Resources partnership (0.51), upholding our qualitative conclusion that communication was an important dimension of partnership functioning. Future research should develop a more appropriate sub-scale to measure communication as an integral component of partnership functioning.

Similarly, for external environment, our qualitative findings demonstrated that external context had an important impact on the work of the partnerships in the case study [35]. The interventions in both partnerships under qualitative investigation unfolded within the context of significant changes in the respective health care systems. This contextual volatility necessitated ongoing adaptations to interventions to respond to evolving environmental opportunities and threats. However, our newly developed sub-scale on External Environment was not correlated with partnership synergy, nor any other dimension of partnership functioning. This finding is consistent with prior research. A study by Weiss, Anderson and Lasker (2002) revealed that community-related challenges, as measured by a three-item scale that the authors developed, did not have any relationship to synergy. The lack of correlation in our study might, however, have been due to the limitations of the scale used: the internal consistency of the scale that we developed was unacceptable (-0.74). This negative result might be due to the fact that it is easier to adapt to a smaller number of environmental changes, which would explain the reverse scoring of the two questions. Alternatively, the two items might not be measuring a single construct. The lack of correlation in our study may also be explained by the fact that external environment is more likely to influence how well the partnership achieves its specific objectives but less likely to have an impact on how the partnership achieves the goal of developing a meaningful partnership. In addition, external environment is highly variable and would require tailored adaptation strategies in each case.

Financial Resources was another sub-scale where we observed a low correlation with most other dimensions of partnership functioning and Partnership Synergy Outcomes. The three-item sub-scale assessed the degree to which the following resources were either present or absent: financial support, space, and equipment and goods. This finding was surprising given the results of prior research demonstrating the critical role of financial resources in partnership functioning [36]. In addition, the partnership synergy framework posits that money and other material resources are among the core determinants of synergy [3]. However, a later study by a team that included the authors behind the partnership synergy framework [19] looked at the relationship between partnership synergy and different dimensions of partnership functioning, and omitted financial resources as a dimension of interest. The authors stated that they had not expected money to be related to the ability of partners to achieve partnership synergy and that financial resources would more likely affect synergy indirectly, by enabling the presence of staff and resources to support the work of the partnership [19]. The weak correlation between Partnership Synergy Outcomes and Financial Resources in our study may be due to the limitations of our Partnership Synergy Outcomes sub-scale (Cronbach’s α < 0.70). The projects undertaken by partnerships in our sample were funded under the envelope of the larger IMPACT program of research. This funding covered the partnerships’ coordinating infrastructure/research support, including the partnership coordinator position in each site, as well as the evaluation of the interventions, but not intervention implementation. Partnership stakeholders, other than research coordinators, were not remunerated for the time spent on partnership activities. Each partnership was required to mobilize adequate local resources to respond to regional access needs and to maintain interventions beyond the life of the IMPACT research funding. Our qualitative study revealed the critical importance of the funded coordinating infrastructure, but that the efforts to secure funding for the interventions had variable success. However, our Financial Resources sub-scale did not make a distinction between different types of funding. This lack of granularity might have affected the uninformative median of 4 for Financial Resources across all five partnerships. In addition, most partnership stakeholders participated on a continuous voluntary basis which suggests that they were driven by considerations other than financial benefit. Future research should focus on further assessing the relationship between financial resources, partnership synergy and different partnership functioning dimensions.

Synergy outcomes and processes

We assessed partnership synergy with two constructs with a limited number of items (two items from the Weiss et al. scale and five items from the Jones synergy scale). We were unable to detect a significant distinction between the two Partnership Synergy sub-scales in terms of how strongly each one of them correlated with other dimensions of partnership functioning. This finding is consistent with prior research that suggested that the Weiss et al. and Jones synergy scales could be used interchangeably [20]. On the other hand, our inability to distinguish between these sub-scales might be due to the limited number of items that we had retained. However, the qualitative ranking by independent content reviewers highlighted differences between the two constructs, suggesting the need for further development of separate partnership synergy processes and partnership synergy outcomes measures. Further assessment of the applicability of both scales is warranted. The scales could be administered longitudinally, at various stages in the partnership evolution, to determine if one is more applicable within the scope of partnership process improvements and the other–as part of outcome evaluation, and whether both scales yielding similar results only occurs at a certain point in partnership development.

At the partnership level, the partnership with the highest total partnership synergy score (Community Outreach) achieved on average the highest scores for partnership functioning on most dimensions. However, the results in the case of partnerships with lower scores on the Partnership Synergy and partnership functioning sub-scales were mixed. One partnership (Primary Care Connection) achieved a high synergy score despite a number of lower scores on the partnership functioning sub-scales, the most noticeable one of which was Resource Utilization. The sub-scale Resource Utilization was referred to as “Efficiency” in the 2002 version of the PSAT [22]. Efficiency was conceptualized as the degree to which a partnership optimized the involvement of its partners [3]. The three-item sub-scale assessed how well the partnership used partner’s time, other non-financial resources, and financial resources [22]. Our qualitative investigation revealed that the Primary Care Connection partnership (which scored lower than other partnerships on this sub-scale) experienced difficulty in meaningfully engaging community-based stakeholders from or representing the target population. In addition, the partnership witnessed reduced engagement of some stakeholders, as they reported that the opportunity cost of conducting activities on behalf of the partnership beyond the face-to-face meetings was too great. Moreover, the extent of environmental impact and adaptation in the case of the Primary Care Connection partnership was particularly profound. The partnership had to redesign the intervention several times to respond to evolving environmental opportunities and threats, and lost some key partners as their organizational positions had changed. These developments may have affected the perceived degree to which the use of financial resources and partners’ time were optimized. The relatively high synergy score for this partnership suggests that there might be certain dimensions that are still lacking from our adapted version of the PSAT and that partnership synergy is an important intermediate outcome worthy of more research. Taking into consideration the emphasis that the stakeholders in all five partnerships put on the quality of relationships, we recommend that future research focus on incorporating into a revised iteration of the PSAT an assessment of the quality of stakeholder relationships, including such aspects as trust, as a stand-alone scale or sub-scale.

Practice implications

The PSAT is a generic tool that has demonstrated robustness in assessing the quality of partnership processes in multi-stakeholder partnerships in several areas, including public health [24]. The PSAT has also been identified as a valid tool for measuring group processes in interprofessional health and social service partnerships at the front-line service provider group level [24]. The original creators of the tool argue that it could have an even broader appeal, including partnerships outside of health care [23]. However, the tool was designed for internal evaluation and not for use by external evaluators [6].

The PSAT may need to be tailored to different partnership types and local conditions. For example, community-driven partnerships with fewer resources, a decentralized partnership structure and more distributed decision-making may need to adapt the Leadership and Administration & Management sub-scales. Alternatively, partnerships may prioritise different aspects for evaluation altogether. There is a variety of other partnership assessment tools that may be more or less suitable depending on the purpose of evaluation and the intended users of its results. For example, the Human Services Integration Measure [37] will be more useful in evaluating the structural components of partnerships, whereas the Team Climate Inventory [38]–for assessing the partnership’s climate for innovation.

The utility of the PSAT as a longitudinal measure of change has not yet been established [23]. We contend that once the PSAT tool is refined and validated, it can reasonably be administered at different times in the evolution of the partnership. The original creators of the PSAT caution that the tool is not intended to be administered at the partnership inception stage [23]. The partnerships should have reached a certain degree of maturity: have been in existence for at least six months, have taken steps to implement plans and be comprised of at least five stakeholders who have continually worked together [23]. For newly formed partnerships, other tools (e.g., the Partnership Analysis Tool developed by VicHealth [11] or the Verona Benchmark [12]) would be more appropriate, as an introduction to the basic concepts of partnership and a guide to structured partnership development.

Once the partnership composition is established, and the partners have had the chance to work together (including face-to-face meetings), the revised PSAT can be administered in order to assess how the partnership is evolving and identify any areas for improvement. It may take some partnerships longer than six months to create momentum and therefore the timing of initial administration needs to be determined based on the specific circumstances of each partnership. At the end of the project the tool would benefit from additional sub-scales on effectiveness and sustainability to assess the likelihood of sustainability and impact on the target populations. In longer projects, if time and resources permit, we recommend that the tool be also administered at the mid-project point, to quantify progress related to how the collaboration is working, and identify strengths and areas that would require more work. In multi-partnership projects the tool would have to be administered at relatively similar intervals to ensure comparability across sites.

Partnership self-assessment, including the administration of the tool, its analysis and sharing of results can be time-consuming, costly in terms of resources, and may contribute to participant burden [15,39]. These factors will affect decisions regarding the tool’s recurrent use. However, we are aware of at least one attempt to develop a shorter version of the PSAT [40], which may be more appropriate for a collaboration “health check” [10] at shorter intervals.

Limitations

Our results should be interpreted in light of the study’s limitations. The first limitation concerns the generalizability of the study’s results. The sample of partnerships is small and the scope is narrow. All participants were sampled from a single international program of research with a targeted scope of implementing organizational interventions to improve the performance of accessibility to PHC, for a particular vulnerable population. Academic representatives constituted the largest single group of respondents in each partnership, which might have influenced our results. Despite the fact that most survey questions were fairly generic, they may have elicited different meaning for different participant groups [41]. For example, different stakeholders may have conceptualized success differently. In addition, it was revealed in the course of our qualitative investigation that academic stakeholders were more likely to have experienced a multi-stakeholder collaborative approach to problem solving and were more familiar with evaluation processes. This may have influenced their willingness to participate in this survey. Secondly, due to the cross-sectional study design, the causality and temporality of associations between the dimensions of partnership functioning and the level of partnership synergy achieved could not be inferred. Taking into consideration that we focused on evolving partnerships, we did not use a separate tool to measure partnership effectiveness. Therefore, we could not demonstrate definitively if partnership synergy was indeed a determinant of partnership effectiveness. In addition, this snapshot assessment might not have captured the holistic nature of partnerships as complex, constantly changing entities. Lastly, while being considered a strength in certain respects, developing the quantitative instrument in both English and French might have resulted in certain inconsistencies when it comes to translating specific terms.

We considered and employed a number of mitigation strategies to overcome some of these limitations. In order to address the issue of small sample size, we used recommended statistical procedures and supplemented our analyses with qualitative information by independent content reviewers. In terms of the scope of partnership activities, we supplemented our survey data with information generated throughout the qualitative phase that preceded this quantitative study. Despite the specificity of partnerships under investigation, there was enough heterogeneity among them, in terms of their contexts, their composition and how the partnerships evolved, to assess the utility of the tool in comparing partnership performance across various settings. In terms of possible translation limitations, we ensured the accuracy of terminology through cognitive testing. Enriching our analyses with qualitative data from the first phase of this larger study has also allowed us to counter some of the challenges involved in using standardized self-assessment tools to assess partnerships. With regards to the causal relationships between partnership functioning and synergy, the administration of a scale to assess the effectiveness of both processes and outcomes would be warranted as all partnerships have now completed their work. This would allow us to determine whether higher synergy actually leads to better sustainability of interventions and better outcomes for the target populations.

Conclusions

We adapted an existing partnership self-assessment tool and administered it to a small sample of developing partnerships. Our findings demonstrate that the tool is suitable for assessing the achievement of partnership synergy and the performance of each partnership along specific indicators of partnership functioning. The tool allowed us to detect meaningful differences between partnerships in how they performed on various dimensions, and corresponded overall to independent qualitative evaluations. The instrument demonstrates good capacity to discriminate between partnerships, and could be applied to assess internal partnership performance on these key indicators across settings, in order to determine if their collaborative process is working well. Therefore, it may have broader applicability to other partnerships in PHC, and arguably beyond PHC. We also propose that additional constructs of quality of partner relationships, communication and external environment could be developed and validated as part of partnership assessment.

In conclusion, this study adds to the growing body of literature on assessing partnership synergy as a predictor of effectiveness of multi-stakeholder partnerships in health care, with a specific focus on PHC. In general, this work has contributed to the evolving body of knowledge on partnership synergy as a useful framework for studying collaborative ventures and identifying the main ingredients or requirements for synergistic partnerships. These requirements should be considered in order to determine if working in partnership should and can reasonably be pursued.

Supporting information

S2 Appendix. Spearman’s rank individual correlations.

https://doi.org/10.1371/journal.pone.0252299.s002

(XLSX)

S3 Appendix. Comparison of rankings of partnerships.

https://doi.org/10.1371/journal.pone.0252299.s003

(DOCX)

Acknowledgments

This research would not have been possible without the support of the IMPACT research program team and all stakeholders who were involved in this study. A special thank you to Fatima Bouharaoui, Biostatistician at the St. Mary’s Hospital Centre, for sharing her statistical expertise and Christine Beaulieu, former IMPACT Research Coordinator, for her support throughout this study.

References

  1. 1. United Nations General Assembly. Enhanced cooperation between the United Nations and all relevant partners, in particular the private sector: Report of the Secretary-General. 2003.
  2. 2. Gray B. Collaborating: finding common ground for multiparty problems. 1st ed. ed. San Francisco: Jossey-Bass; 1989.
  3. 3. Lasker RD, Weiss ES, Miller R. Partnership Synergy: A Practical Framework for Studying and Strengthening the Collaborative Advantage. Milbank Q. 2001;79(2):179–205. pmid:11439464
  4. 4. Zuckerman HS, Kaluzny AD, Ricketts TC 3rd. Alliances in health care: what we know, what we think we know, and what we should know. Health care management review. 1995;20(1):54–64. pmid:7744606
  5. 5. Drahota A, Meza RD, Brikho B, Naaf M, Estabillo JA, Gomez ED, et al. Community-Academic Partnerships: A Systematic Review of the State of the Literature and Recommendations for Future Research. Milbank Q. 2016;94(1):163–214. pmid:26994713
  6. 6. Joss N, Keleher H. Partnership tools for health promotion: are they worth the effort? Glob Health Promot. 2011;18(3):08–14. pmid:24803555
  7. 7. Jones J, Barry MM. Exploring the relationship between synergy and partnership functioning factors in health promotion partnerships. Health Promot Internation. 2011a;26(4):408–20.
  8. 8. Kreuter MW, Lezin NA, Young LA. Evaluating Community-Based Collaborative Mechanisms: Implications for Practitioners. Health Promot Pract. 2000;1(1):49–63.
  9. 9. Mitchell SM, Shortell SM. The Governance and Management of Effective Community Health Partnerships: A Typology for Research, Policy, and Practice. Milbank Q. 2000;78(2):241–89. pmid:10934994
  10. 10. Halliday J, Asthana SNM, Richardson S. Evaluating Partnership: The Role of Formal Assessment Tools. Evaluation. 2004;10(3):285–303.
  11. 11. Victorian Health Promotion Foundation. The partnerships analysis tool: A resource for establishing, developing and maintaining partnerships for health promotion 2016 [Available from: https://www.vichealth.vic.gov.au/media-and-resources/publications/the-partnerships-analysis-tool.
  12. 12. Watson J, Speller V, Markwell S, Platt S. The Verona Benchmark: applying evidence to improve the quality of partnership. Promotion & education. 2000;7(2):16–23. pmid:11924343
  13. 13. Butterfoss FD, Goodman RM, Wandersman A. Community coalitions for prevention and health promotion: factors predicting satisfaction, participation, and planning. Health education quarterly. 1996;23(1):65–79. pmid:8822402
  14. 14. Roussos ST, Fawcett SB. A Review of Collaborative Partnerships as a Strategy for Improving Community Health. Annual Review of Public Health. 2000;21(1):369–402. pmid:10884958
  15. 15. Butterfoss FD, Francisco VT. Evaluating Community Partnerships and Coalitions with Practitioners in Mind. Health Promot Pract. 2004;5(2):108–14. pmid:15090164
  16. 16. Kegler MC, Steckler A, McLeroy K, Malek SH. Factors That Contribute to Effective Community Health Promotion Coalitions: A Study of 10 Project ASSIST Coalitions in North Carolina. Health Educ Behav. 1998;25(3):338–53. pmid:9615243
  17. 17. Larkan F, Uduma O, Lawal SA, van Bavel B. Developing a framework for successful research partnerships in global health. Globalization and Health. 2016;12(1). pmid:27154550
  18. 18. Ramaswamy R, Kallam B, Kopic D, Pujic B, Owen MD. Global health partnerships: building multi-national collaborations to achieve lasting improvements in maternal and neonatal health. Globalization and Health. 2016;12(1). pmid:27206731
  19. 19. Weiss ES, Anderson RM, Lasker RD. Making the Most of Collaboration: Exploring the Relationship Between Partnership Synergy and Partnership Functioning. Health Educ Behav. 2002;29(6):683–98. pmid:12456129
  20. 20. Jones J, Barry MM. Developing a scale to measure synergy in health promotion partnerships. Glob Health Promot. 2011b;18(2):36–44.
  21. 21. Cramm JM, Nieboer AP. Disease-management partnership functioning, synergy and effectiveness in delivering chronic-illness care. International Journal for Quality in Health Care. 2012;24(3):279–85. pmid:22302071
  22. 22. Center for the Advancement of Collaborative Strategies in Health. Partnership Self-Assessment Tool Questionnaire. 2002. Available from: https://atrium.lib.uoguelph.ca/xmlui/bitstream/handle/10214/3129/Partnership_Self-Assessment_Tool-Questionnaire_complete.pdf?sequence=1&isAllowed=y.
  23. 23. Center for the Advancement of Collaborative Strategies in Health. Partnership Self-Assessment Tool—Coordinator Guide [September 13, 2020]. Available from: https://atrium.lib.uoguelph.ca/xmlui/bitstream/handle/10214/3132/Partnership_Self-Assessment_Tool-_Coordinators_Guide_complete.pdf?sequence=1&isAllowed=y.
  24. 24. Butt G, Markle-Reid M, Browne G. Interprofessional partnerships in chronic illness care: a conceptual model for measuring partnership effectiveness. International Journal of Integrated Care [Internet]. 2008; 8(2). pmid:18493591
  25. 25. McMaster University National Collaborating Centre for Methods and Tools. Partnership evaluation: The Partnership Self-Assessment Tool [September 13, 2020]. Available from: https://www.nccmt.ca/knowledge-repositories/search/10.
  26. 26. Russell G, Kunin M, Harris M, Levesque JF, Descôteaux S, Scott C, et al. Improving access to primary healthcare for vulnerable populations in Australia and Canada: protocol for a mixed-method evaluation of six complex interventions. BMJ Open. 2019;9(7):e027869. pmid:31352414
  27. 27. Dahrouge S, Gauthier A, Chiocchio F, Presseau J, Kendall C, Lemonde M, et al. Access to Resources in the Community Through Navigation: Protocol for a Mixed-Methods Feasibility Study. JMIR Research Protocols. 2019;8(1). pmid:30679151
  28. 28. Dillman DA, Smyth JD, Christian LM. Internet, Phone, Mail, and Mixed-Mode Surveys: The Tailored Design Method. 4th ed. New Jersey: John Wiley & Sons, Inc; 2014.
  29. 29. Willis GB. Cognitive Interviewing: A Tool for Improving Questionnaire Design. Thousand Oaks, Calif.: Sage Publications; 2005.
  30. 30. Qualtrics. September 2018 edn. [Internet]. Provo, Utah, USA. 2005.
  31. 31. IBM Corp. IBM SPSS Statistics for Windows. 23 ed. Armonk, NY2015.
  32. 32. Bennett DA. How can I deal with missing data in my study? Australian and New Zealand Journal of Public Health. 2001;25(5):464–9. pmid:11688629
  33. 33. Cohen J. Statistical power analysis for the behavioral sciences. 2nd ed. ed. Hillsdale, N.J.: L. Erlbaum Associates; 1988.
  34. 34. Nunnally JC. Psychometric theory. 2d ed. ed. New York: McGraw-Hill; 1978.
  35. 35. Loban E, Scott C, Lewis V, Law S, Haggerty J. Improving primary health care through partnerships: Key insights from a cross-case analysis of multi-stakeholder partnerships in two Canadian provinces. Submitted.
  36. 36. Corbin JH, Mittelmark MB. Partnership lessons from the Global Programme for Health Promotion Effectiveness: a case study. Health Promot Internation. 2008;23(4):365–71. pmid:18835888
  37. 37. Browne G, Roberts J, Gafni A, Byrne C, Kertyzia J, Loney P. Conceptualizing and validating the human services integration measure. International journal of integrated care. 2004;4:e03. pmid:16773145
  38. 38. Anderson NR, West MA. Measuring Climate for Work Group Innovation: Development and Validation of the Team Climate Inventory. Journal of Organizational Behavior. 1998;19(3):235–58.
  39. 39. Centers for Disease Control and Prevention. Evaluation Guide: Fundamentals of Evaluating Partnerships. U.S. Department of Health and Human Services; 2008.
  40. 40. Cramm J, Nieboer A, Strating M. Development and validation of a short version of the Partnership Self-Assessment Tool (PSAT) among professionals in Dutch disease-management partnerships. BMC Research Notes. 2011;4(1):1–8.
  41. 41. El Ansari W, Phillips CJ, Hammick M. Collaboration and partnerships: developing the evidence base. Health Soc Care Community. 2001;9(4):215–27. pmid:11560737