Browse Subject Areas

Click through the PLOS taxonomy to find articles in your field.

For more information about PLOS Subject Areas, click here.

  • Loading metrics

Innovation in Evaluating the Impact of Integrated Service-Delivery: The Integra Indexes of HIV and Reproductive Health Integration

  • Susannah H. Mayhew ,

    Affiliation Department of Global Health & Development, London School of Hygiene & Tropical Medicine, 15–17 Tavistock Place, London, WC1H 9SH, United Kingdom

  • George B. Ploubidis,

    Affiliation Centre for Longitudinal Studies, Department of Social Science, UCL Institute of Education, University College London, 55–59 Gordon Square, London, WC1H 0NU, United Kingdom

  • Andy Sloggett,

    Affiliation Department of Population Health, London School of Hygiene & Tropical Medicine, Keppel St, London, WC1E 7HT, United Kingdom

  • Kathryn Church,

    Affiliation Department of Population Health, London School of Hygiene & Tropical Medicine, Keppel St, London, WC1E 7HT, United Kingdom

  • Carol D. Obure,

    Affiliation Department of Epidemiology and Public Health, Swiss Tropical and Public Health Institute, P.O. Box 4002, Basel, Switzerland

  • Isolde Birdthistle,

    Affiliation Department of Population Health, London School of Hygiene & Tropical Medicine, Keppel St, London, WC1E 7HT, United Kingdom

  • Sedona Sweeney,

    Affiliation Department of Global Health & Development, London School of Hygiene & Tropical Medicine, 15–17 Tavistock Place, London, WC1H 9SH, United Kingdom

  • Charlotte E. Warren,

    Affiliation Reproductive Health Services and Research, Population Council, Suite 280, 4301 Connecticut Ave NW, Washington, District of Columbia, 20008, United States of America

  • Charlotte Watts,

    Affiliation Department of Global Health & Development, London School of Hygiene & Tropical Medicine, 15–17 Tavistock Place, London, WC1H 9SH, United Kingdom

  • Anna Vassall,

    Affiliation Department of Global Health & Development, London School of Hygiene & Tropical Medicine, 15–17 Tavistock Place, London, WC1H 9SH, United Kingdom

  • Integra Initiative

    Members of the Integra Initiative can be found in the Acknowledgments.

Innovation in Evaluating the Impact of Integrated Service-Delivery: The Integra Indexes of HIV and Reproductive Health Integration

  • Susannah H. Mayhew, 
  • George B. Ploubidis, 
  • Andy Sloggett, 
  • Kathryn Church, 
  • Carol D. Obure, 
  • Isolde Birdthistle, 
  • Sedona Sweeney, 
  • Charlotte E. Warren, 
  • Charlotte Watts, 
  • Anna Vassall



The body of knowledge on evaluating complex interventions for integrated healthcare lacks both common definitions of ‘integrated service delivery’ and standard measures of impact. Using multiple data sources in combination with statistical modelling the aim of this study is to develop a measure of HIV-reproductive health (HIV-RH) service integration that can be used to assess the degree of service integration, and the degree to which integration may have health benefits to clients, or reduce service costs.

Methods and Findings

Data were drawn from the Integra Initiative’s client flow (8,263 clients in Swaziland and 25,539 in Kenya) and costing tools implemented between 2008–2012 in 40 clinics providing RH services in Kenya and Swaziland. We used latent variable measurement models to derive dimensions of HIV-RH integration using these data, which quantified the extent and type of integration between HIV and RH services in Kenya and Swaziland. The modelling produced two clear and uncorrelated dimensions of integration at facility level leading to the development of two sub-indexes: a Structural Integration Index (integrated physical and human resource infrastructure) and a Functional Integration Index (integrated delivery of services to clients). The findings highlight the importance of multi-dimensional assessments of integration, suggesting that structural integration is not sufficient to achieve the integrated delivery of care to clients—i.e. “functional integration”.


These Indexes are an important methodological contribution for evaluating complex multi-service interventions. They help address the need to broaden traditional evaluations of integrated HIV-RH care through the incorporation of a functional integration measure, to avoid misleading conclusions on its ‘impact’ on health outcomes. This is particularly important for decision-makers seeking to promote integration in resource constrained environments.


Since the 1990s there has been an assumption that integration of HIV-related services (e.g. HIV testing, condom provision, and HIV treatment) with family planning (FP), antenatal care (ANC) and postnatal care (PNC) services, would streamline service delivery. In practice, countries take myriad different approaches to the organisation of care, and ‘integrated’ primary care services remain poorly defined.[13] In lower- and middle-income contexts it is usually understood as the amalgamation of previously separate components of care, or the addition of a new intervention into an existing service (e.g. adding HIV testing to FP services).[4] In industrialised country settings it is often interpreted as a mechanism to improve the coordination of care between different organisations and professional bodies at different levels of the health system.[5] As a result, there is no standard definition of “integration” which has been variously categorised by different authors and studies.[4, 610]

The multi-dimensional nature of integration raises the fundamental question of how ‘integration’ or improvements in integration should be measured. Without clarity on this, it is difficult to assess the extent to which services are integrated, and whether the delivery of integrated services leads to cost savings, greater client satisfaction, and/or improved patient outcomes. To date, only a few studies have attempted to measure, rank, or assess different levels or types of HIV/reproductive health (RH) service integration.[11] One study in the US attempted to rank clinics by their degree of integration of HIV with primary care services, determined by interviews with facility managers, then applied this to HIV client visit data to give an ‘index of integrated care utilisation’.[12] Another in South Africa attempted to quantify different levels of integration, again based on questionnaires with staff.[13] A third sought to describe the relationship between integration and ‘teen-friendly’ service outcomes, and measured integration through staff and client interviews.[14]

The Integra Initiative is the largest complex evaluation of its kind seeking to determine the impact of service integration on service and health outcomes in Kenya and Swaziland. Integra’s focus is the integration of HIV/sexually transmitted infection (HIV/STI) services (including HIV/STI counselling, testing and treatment) and reproductive health (RH) services (including: FP, ANC and PNC), for which multiple benefits have been claimed yet the evidence is at best inconsistent.[15, 16] Integra’s initial intent was to conduct a pre/post study with pair-matched intervention (integrated) and comparison (non-integrated) sites. However, as with many other pragmatic trials of this kind, the study had no direct control over clinic organisation, but instead supplied inputs to facilitate integration such as training and supplies. In such a ‘real world setting’ this lack of control, high staff turnover and migration between sites, external donor activities and evolving policy meant that over time some control sites also integrated services during the study. It therefore quickly became clear that any longitudinal analysis would be confounded by the varying levels of integration already existing prior to the start of the intervention and then developing over time, independently, in both intervention and comparison sites. To address this, a need was identified to develop a tool to capture and measure the extent of integration at the study sites, and then use this to explore the relationship between integration and a range of study outcomes.

This situation is not unique: when evaluating complex multi-service health interventions within ‘real world’ settings, study designs that capture the process and extent of the implementation of the intervention, both at baseline and throughout the course of the study are becoming increasingly necessary, particularly where organisational change is involved.[1719] This paper thus aims to contribute to the field of complex intervention evaluation, as well as the broader policy debate on health service integration, by describing the development of a tool to measure the degree of HIV-RH integration achieved in the health facilities studied: the Integra Index.


The Integra Initiative

The Integra Initiative is a non-randomised, pre/post intervention trial using household and facility-based data. Integra uses mixed methods to analyse a hypothesised causal pathway between integration and its theorised outcomes, which include costs, quality of care, service utilisation, stigma and sexual and reproductive behaviours. Integra is embedded research, working in public-sector and NGO facilities in Kenya and Swaziland. It aims to evaluate the impact of different models of delivering integrated HIV-RH services in Kenya and Swaziland on a range of health and service outcomes.[20] Research was conducted in both high- and moderate-HIV prevalence settings (Swaziland, 10 clinics; and Kenya, 30 clinics). An intervention was developed in collaboration with the national Ministries of Health to support service integration in study clinics (see Warren et al. for details[20]). The 40 study facilities consisted of: primary care clinics including dispensaries, health centres, public health units (n = 30); and secondary out-patient clinics at district or provincial hospitals (n = 10). A total of 32 clinics were primarily managed by the public sector, while eight were run by an NGO (6 in Kenya, 2 in Swaziland).


Ethical clearance was granted by the Kenya Medical Research Institute (KEMRI) Ethical Review Board (approval # KEMRI/RES/7/3/1, protocol #SCC/113 and #SCC/114), the Ethics Review Committee of the London School of Hygiene & Tropical Medicine (LSHTM) (#5426 and #5436) and the Population Council’s Institutional Review Board (#443 and #444). The Integra Initiative is a registered non-randomised trial: Identifier: NCT01694862.

Data collected are described below, but involved either routine aggregated clinic data (on human and physical resource-use) or anonymised (at point of data collection) data from clients using the facility during the five-day client flow data collection period (about services received). No personal or client-identifying information was collected and data were only used in aggregate form per clinic. No individuals’ clinical records were used and no individual provider details were recorded. For providers who were interviewed, individual written informed consent was obtained, recorded on an informed-consent sheet. In addition, verbal informed consent was obtained from facility managers for all facilities involved in the Integra study. These forms of consent were approved by the above ethics boards.

Measurement & ranking of integrated care: the Integra Index

The Integra Index was developed to measure the level of integration at baseline and follow-up in all the 40 study facilities. The construction of the Index involved four steps, detailed below.

1) Identification of integration attributes and indicators.

We reviewed the key aspects of integration identified in the literature [410, 2123] to help identify the dimensions of integration that needed to be measured. The dimensions emerged as: 1) physical integration—the most common way for integration to be described, in terms of multiple services being available at the same facility (with or without referrals between units/rooms) or in the same room;[4, 6, 7, 22] 2) Temporal integration—multiple services being available throughout the week vs. on specific days;[8] 3) Provider-level integration—one provider gives multiple services in a consultation (either actively or in response to client demands) or over one day.[9, 22, 23] In addition, we identified a fourth, higher level dimension: receipt of integrated care by clients, or what we termed “Functional integration” within a single visit or consultation. We validated these dimensions by consulting with service providers and researchers from Kenya and Swaziland (described in 3) below).

As the index was developed post initial study design, we then assessed the study data we had to identify the best and most feasible measure of each integration dimension. The multi-disciplinary research team (including evaluation researchers, epidemiologists, health systems researchers, economists and statisticians) identified eight ‘attributes’ from our study data that best reflected the above dimensions of integration: physical (what rooms/buildings different services are delivered in); temporal (on what days/times); provider (by whom); and functional (defined as “actual services received by client”) (Table 1). Given the study’s focus, the indicators measured the provision/receipt of any RH service (FP, ANC, PNC) AND any HIV/STI service (HIV counselling and testing, HIV anti-retroviral therapy (ART) treatment, CD4 count services, STI treatment, cervical cancer screening).

2) Use of clinic data to generate attribute scores.

We drew on two Integra datasets to capture the eight selected attributes, with data from each collected from the 40 clinics at baseline (2008–9) and endline (late 2010-early 2012): (i) a dataset used for costing ‘the economics dataset’ (that included service statistics) and (ii) a client flow dataset. Summary descriptions are provided in Table 2.

The economics dataset was derived from costing and periodic activity review tools, completed by researchers in collaboration with facility managers and staff.[24] The tools collected data on expenditures, facility characteristics, staffing and services. They included data from facility registers on services provided, as well as observations of services offered and resource use. The latter involved researchers observing staff members and facility practice over a one week period. In addition, interviews were conducted with facility staff, including completion of timesheets, to better understand how both physical infrastructure and human resources were used to provide services. The economics dataset was used to confirm the range of services available in the relevant department and facility, to estimate the average number of different HIV-RH services provided in each consultation room per day, and measure the range of services provided per staff member, and as such, provides information on the physical and human resource infrastructure that is in place and being used (structural integration).

The client flow dataset was derived from a five-day assessment of service utilisation patterns in each study clinic.[25] A client flow form was used to record all services received by/referred to for each client in every consultation over one day’s visit. Forms were completed by each provider seen (providers were, therefore, aware of the data being recorded). A total of 25,539 visits were tracked across 24 facilities in Kenya (10,266 in Eastern Province and 15,270 in Central Province) and 8263 visits across 8 facilities were tracked in Swaziland. Baseline and endline data only are reported in this paper: sample sizes are shown in Table 2. The dataset was used to measure which services were being offered on site; the range of services provided across days of the week; the range of services provided in single consultations; and the range provided in single visits. In this way, this data provides information on each clinic’s ability to deliver integrated services to clients (functional integration).

From these data sources, tables containing eight data points (for each attribute in Table 1) for each study clinic were constructed (S1 Table).

3) Expert opinions of integration attribute weights.

Given that the initial development of the index was by an internationally led study team we identified a need to further add to the development of measurement tools by using a formal process to gather local expert opinion to inform our index model. We sought the views of 23 service providers, managers and researchers from Kenya and Swaziland on the selected integration attributes in order to weigh the relative importance of attributes, allowing sensitivity within the model to different attributes of care. Participants were purposively sampled: those with knowledge of the country contexts, services being investigated and a range of the study clinics.

All participants were asked to rank the eight attributes in order of their importance to defining integrated care, using a modified Delphi technique involving ranking, discussion and re-ranking to reach consensus.[26, 27] The results of the final ranking round were used to inform the modelling process, however these findings did not change the overall substance of the models. They did not change the need for a two factor solution (see next section) but in fact worsened the fit of the models. For these reasons the expert opinion data was not used in the final modelling process for generating the Integra indexes, leaving the basis of the models objective data only.

4) Generating the Integra Indexes.

“Integration” is accepted as a complex phenomenon embracing multiple concepts and definitions. Integration is thus viewed as a metric whose true values cannot be directly observed [28] and the assumption is that our attributes are manifestations of this latent construct of integration.

We used latent variable models to develop models that combine information from the eight attributes of integration listed in Table 1. These models allow the combination of information from these attributes, without making any assumptions about their measurement unit, and also allow the empirical assessment of the reliability and validity of these. Results for these models are shown in Tables 3, 4 and 5.

Table 3. Measures of model fit for baseline models showing χ2 confidence interval, associated p value and Proportional Scale Reduction Criterion.

Table 4. Standardised factor loading scores for attributes of a single-factor model, at baseline and endline.

Table 5. Standardised factor loading scores for the two-factor model at baseline and endline.

We modelled the relationship between observed attributes of integration and latent integration using 2-parameter probit link functions for the binary and ordinal nature of the indicators.[28] In this framework the integration latent variable is influenced by all attributes, the relative contribution of each attribute to the latent summary of integration being expressed by the factor loading. For binary or ordinal attributes probit thresholds represent the level of latent integration that needs to be reached for each particular category. Given the small sample of 40 clinics, the Bayesian estimation framework was used. The Bayesian framework offers an attractive alternative to maximum likelihood estimation which may produce biased estimates in small-sample studies because of its reliance on large sample (asymptotic) theory. We employed normally distributed “non-informative priors” in an attempt to approximate maximum likelihood in our small sample.

All models were estimated using the Markov Chain Monte Carlo algorithm (four chains, 50,000 Bayes iterations) based on the Gibbs sampler, firstly on baseline data (2009) in Mplus 6.[29] Model convergence was assessed with the Proportional Scale Reduction (PSR) criterion (values close to 1 indicate model convergence) and model fit with the 95% confidence interval for the difference between the observed and replicated chi square values (the inclusion of zero in the interval and a non-significant p value indicate good fit).


Integration measurement models

Table 3 shows criteria of model fit for the one and two-factor models using a latent variable model with non-informative priors for all parameters as an alternative to maximum likelihood estimation. Indexes of fit are shown for the baseline data only as endline data models returned very similar fit (results available from corresponding author) as well as a similar pattern of factor loadings. For the single factor model the PSR value is close to indicating convergence. However the χ2 95% confidence interval for the difference between the observed and replicated χ2 excludes zero and the associated p value is significant, indicating poor fit. However the 2-factor model shows good fit by the same criteria.

Table 4 shows the standardised factor loadings for the model at baseline and endline and indicates a single latent dimension of integration. These loadings simply describe the relative weight or association between the attribute and the latent summary of “integration”, i.e. the strength of association. Conventionally, an absolute factor loading of 0.7 is “very satisfactory” and above 0.4 “acceptable” when assessing such a model.[30]

The factor loadings in Table 4 suggest that structural integration attributes and functional integration attributes (the ability to deliver integrated HIV-RH services) are not correlated. In other words, structural integration does not necessarily result in integrated delivery of HIV-RH services to the client. While one might expect structural integration characteristics to be correlated with (or be a pre-requisite for) integrated service delivery, in fact the findings suggest that an inverse relationship may exist. We ran further tests to assess whether the results were driven by facility size (not shown): there were no significant differences although confidence intervals were too wide to interpret.

As a result of the differences in the loadings of the attributes of structural and functional integration, as well as the poor fit of the unidimensional integration models, we estimated a two-factor data-derived (non-informative priors) model where the index was separated to create two sub-indexes. Table 5 shows the strong data-driven loadings of the attributes on each latent factor implying that these most likely reflect valid variance and not systematic error due to different data collection techniques.

Our analysis of data suggests that the structural and the functional attributes constitute two distinct and uncorrelated (orthogonal) dimensions of integration and therefore each needs to be considered separately.

The two factor model had a good fit to the data as indicated by the inclusion of zero in the χ2 95% confidence interval for the difference between the observed and replicated χ2 and the associated non-significant p value (Table 3). Furthermore, Table 4 shows that the factor loadings remain quite consistent over the three years, indicating that the two-factor model is well replicated over time—i.e. the stability of the measurement model is good. This does not imply that the latent factors do not change over time, but that any change can be attributed to true change in integration rather than inconsistency in measuring integration.

Finally, all attributes show sizeable loadings and are directionally sensible. This situation is normally accepted as evidence of construct or internal validity and we felt confident that this was so, particularly given the better model-fit of the two-factor model and its stable factor loadings over time. As expected from the stability of factor loadings over time, the measurement equivalence of the two factor model between 2009 and 2012 was empirically confirmed (results available from corresponding author). Thus we retained two distinct latent factors: a Structural Integration Index and a Functional Integration Index.

Clinic rankings

To illustrate the application of the index we present Fig 1 that then applies these indexes to clinics in the study sample. Based on the two factor model latent integration scores and their corresponding standard errors, values were calculated as Bayesian plausible values [31, 32] for both factors. The stability of the factor loadings (measurement equivalence) over the three years allowed us to calculate latent scores for both waves using baseline measurement model parameters in order to obtain a fair comparison between the two time points (2009 vs 2012). Latent scores can theoretically range from minus infinity to infinity, but in practice the range is usually from -3 to 3 (transformed to 1 to 7 for graphing purposes). For the purpose of analysis, a model based latent score (Bayesian plausible value) was assigned to each clinic. The clinics were then ranked (from 1 = high, to 42 = low) with respect to the two latent dimensions of integration.

Fig 1 shows model-derived index scores for all clinics, at baseline and endline, by factor (functional/structural) and by country. In the upper panel (functional integration scores) the clinics have been ordered by descending baseline index score, from left to right for Kenya and for Swaziland. In the lower panel the clinic order is consistent with the upper panel.

Correlations between structural and functional scores were tiny in both waves (r = -0.12 (2009), -0.04 (2012)) which is to be expected as the scores are model-based from a model with two orthogonal factors.

The pattern in the upper panel (functional integration) is striking: high baseline scores have mostly diminished by endline whereas lower baseline scores have increased moderately (with two particularly notable increases in Kenya). The stability of the 2012 scores in Kenya, apart from three high scoring clinics, is quite remarkable and is likely to be a classic case of regression to the mean. The pattern is similar for Swaziland although 2012 improvements are more marked.

This paper is concerned with measurement issues rather than any program effect but, bearing in mind that latent scores were very similar for the two periods, and that we chose the baseline model to construct scores for both periods so that they were comparable, the pattern suggests an improving situation for initially low scoring clinics.

By comparison the lower panel (structural integration), when viewed in a clinic-by-clinic comparison shows a less coherent pattern (because it is ordered consistently with the upper panel, rather than by factor scores for the structural Index). If ordered in the same fashion (not shown) the differences between baseline and endline are far less marked and the pattern is less distinctive. There is still some suggestion of high scoring clinics losing score and vice versa, but this is less marked and restricted to the extremes. It seems much more likely that these smaller effects are indeed regression to the mean.

The sum of absolute differences between baseline and endline are 31.6 for the upper panel combined and 15.3 for the lower panel. A higher correlation between baseline and endline scores in the combined lower panel than in the upper panel (r = 0.83 vs 0.41 respectively) is consistent with the absolute differences.

Smaller differences for the structural integration component is perhaps not surprising since structural integration is often dictated by infrastructure, which is difficult to change. However what is of interest here is that the structural integration component has shown relatively little change between periods but the functional integration component has revealed distinctly patterned differences over time.

Whether these changes are related to program effects, external effects, or entirely regression to the mean will be subject to further analysis but the use of an objectively derived index has uncovered and quantified a pattern worthy of investigation, and has facilitated further research in a way that subjective judgements of experts, or indeed researchers, may not have done.


This paper addresses an important deficiency in the literature on measurement of integrated HIV-RH care by developing a tool that can be used to assess the degree of service integration. The analysis uses multiple data from a moderate number of clinics, in combination with statistical modelling, to develop a measure of service integration. This is an important advance on previous research: many previous reports simply categorize clinics as integrated or not depending on whether they received an intervention (or as self-reported by staff), but do not formally assess the extent of integration as a multi-dimensional continuum. Even where formal measurements are made previous studies have rarely included any measurement of whether clients receive integrated care (i.e. multiple HIV-RH services received by a client from one provider or in one visit).[3, 15, 16]

Two findings from our study are noteworthy. First, the two uncorrelated factors that emerged from our analysis strongly suggest that simply putting infrastructure and multi-tasking staff in place (“structural integration”) may not be sufficient to achieve integrated HIV-RH service receipt by the client (“functional integration”). This is not necessarily counter-intuitive, as it is plausible that some sites, particularly smaller ones, may have high levels of structural integration, but in practice may not be able to deliver integrated services because of barriers like vertical, duplicate reporting/recording systems, time constraints and poor staff motivation.

The emergence of two distinct dimensions illustrates the difference between the structural integration of a facility offering potential for integrated HIV-RH delivery and integrated HIV-RH services actually received by the client. This distinction is important for future outcomes assessments since it suggests that measures of physical integration and staff multi-tasking should not be used on their own to measure HIV-RH integration or relate integration to HIV and RH outcomes, but that a more nuanced causal pathway analysis may provide a better interpretation of results. Yet many studies do equate structural integration with integrated delivery of care and fail to explore the extent to which structural integration leads to integrated delivery, which then leads to outcomes. Our findings underline the need for a broader approach to measuring HIV-RH integration, combining a functional integration measure to better understand the link between an intervention in the area of integration and observed health outcomes.

Second, our findings show the high degree of heterogeneity across clinics in both countries, illustrating that integration is highly complex and was implemented and achieved differently in every clinic. This reinforces the need to adopt a measurement approach that allows for assessment along a continuum and is able to quantify the extent of service integration. This approach is also critical for embedded or programme-science research designs, where ensuring a standardised dose of an intervention may be influenced by a wide variety of factors beyond the control of the research study.

A number of important limitations need to be taken into account when reflecting on our findings. Costing data based on observation, self-reporting and routine service records is susceptible to a number of biases, including provider reporting bias. Client flow data, collected over one week, may not be representative of typical monthly/annual client flow (e.g. there may have been staff on training that week, national holidays, seasonal use etc.). Coordinating the same five days across facilities was logistically challenging and not always achieved. Nevertheless, clinic register data could not be used as an alternative since registers could not record how many different services each individual client received in a single consultation or visit. However, the use of multiple methods allowed for complementarity between different data sources for the indicators used, rather than relying on a single source.

Despite these limitations, the Structural and Functional Integration Indexes have useful future applications. First, they can be used to assess how clinics are changing over time relative to other clinics in terms of service integration—useful for policy or programme decision-makers interested in knowing how clinics are progressing. The general decline in clinic scores at endline (Fig 1) suggests sustaining integration over time can be challenging and further analysis is ongoing to understand the processes by which integration is successfully sustained. More work is underway to select a single or small number of indicators that could be routinely (or more easily) measured, while still providing an accurate assessment of HIV-RH integration. These indicators can be used by programme and policy decision-makers to monitor programme achievements on HIV-RH service integration in both high- and low-income settings.

Second, our Integration Indexes help identify what attributes are most closely associated with integration, or in other words what drives integration, in different contexts. This is important for policy makers and funders who wish to know where to channel resources. Many impact studies of integration focus on the impact of an intervention on observed outcomes but fail to account for the drivers of this impact. [33,34] For example, observed improvements in uptake of services following integration could be the result of better supported, more motivated staff at intervention sites, rather than the mere presence of an additional service. Consequences of this for funding are important since it may be that large-scale investment in structural components (e.g. roll-out of drugs or test-kits on site; clinical training of staff) may not result in sustained improvements in health outcomes unless equally large investments are made in supporting the workforce (e.g. through mentorship, quality supervision). A growing literature on the importance of supporting the “software” (people) within health systems and services, beyond narrow clinical training, is important here. [35, 36] To further our understanding of the drivers of integration, we are conducting further analysis of the individual attributes to determine whether a sequencing of inputs can be identified (e.g. do you need physical and human resource integration in place before you get availability of services within an MCH unit and what enables this to lead to delivery of integrated care?) [37,38]. Third, the Indexes can enable researchers to improve the assessment of the attribution of a particular health or service outcome to service integration.[3] For example, within the Integra Initiative we are using the Indexes to provide a sophisticated measure to assess a dose-response relationship between women’s cumulative exposure to integrated services and study outcomes, including HIV-testing, condom use and technical quality of care.[39] Where positive impacts are correlated with functional integration we are conducting further analysis to determine drivers of functional integration.

Furthermore, beyond SRH/HIV integration, the principles of the need to measure both a dimension of “structural” service integration and “functional” delivery of integrated care may be applicable for other service-integration packages, though further application is needed to test this.

To conclude, the Integra Indexes strongly suggest that “integration” exists in two forms that are distinct but easily confused. Functional integration is linked to actual receipt of multiple services at one time and place and may be unrelated to structural integration where different services are “available”, but not necessarily provided in an integrated form. Furthermore, our findings have important implications for research on integrated HIV-RH services since they underline the importance of 1) having a measure to quantify the degree of service integration; 2) assessing both structural integration (physical and human resources) and functional integration (delivery of integrated services to a client) in order to determine the ‘attainment’ of integrated services. We conclude, that the Integra Indexes are a useful conceptual and methodological contribution to measuring HIV-RH integration and enabling the attribution of particular health/service outcomes to integration—achievements which have proven elusive to date.

Supporting Information

S1 Table. Descriptive Data for Eight Variables in the Integra Index Models, by clinic at baseline and endline.


S2 Table. Supporting data for Fig 1: baseline and endline clinic scores, by country.



The authors would like to thank other members of the Integra Initiative research team who contributed to the design of the study: Timothy Abuya, Ian Askew, Manuela Colombini, Natalie Friend du-Preez, Joshua Kikuvi, James Kimani, Jackline Kivunaga, Jon Hopkins, Joelle Mak, Christine Michaels-Igbokwe, Richard Mutemwa, Charity Ndigwa and Weiwei Zhou.

The Integra Initiative lead author is Susannah Mayhew: The Integra Initiative team members are: At the London School of Hygiene & Tropical Medicine: Susannah Mayhew (PI), Anna Vassall (co-PI), Isolde Birdthistle, Kathryn Church, Manuela Colombini, Martine Collumbien, Natalie Friend-DuPreez, Natasha Howard, Joelle Mak, Richard Mutemwa, Dayo Obure, Sedona Sweeney, Charlotte Watts. At the Population Council: Charlotte Warren (PI), Timothy Abuya, Ian Askew, Joshua Kikuvi, James Kimani, Jackline Kivunaga, Brian Mdawida, Charity Ndwiga, Erick Oweya. At the International Planned Parenthood Federation: Jonathan Hopkins (PI), Lawrence Oteba, Lucy Stackpool-Moore, Ale Trossero; at FLAS: Zelda Nhlabatsi, Dudu Simelane; at FHOK: Esther Muketo; at FPAM: Mathias Chatuluka

We would also like to thank: the fieldwork teams; the research participants who gave their time to be interviewed; the managers of the clinics who facilitated data collection; and the Ministries of Health in Kenya and Swaziland who supported the research and are key partners in the Integra Initiative. Finally, we would like to thank other colleagues who gave feedback on aspects of research design and/or analysis.

Author Contributions

Conceived and designed the experiments: SM GP AS AV. Performed the experiments: GP SS CDO. Analyzed the data: SM GP KC SS CDO IB. Contributed reagents/materials/analysis tools: GP SM AV. Wrote the paper: SM AS GP AV. Contributed to the conceptualisation of the index, selection of the indicators, and interpretation of the results: SM KC CDO IB SS CEW CW AS AV.


  1. 1. Atun R., de Jongh T, Secci F, Ohiri K and Adjeyi O., A systematic review of the evidence on integration of targeted health interventions into health systems. Health Policy Plan, 2010. 25(1): p. 1–14. pmid:19959485
  2. 2. Atun R., de Jongh T, Secci F, Ohiri K and Adjeyi O., Integration of targeted health interventions into health systems: a conceptual framework for analysis. Health Policy Plan, 2010. 25(2): p. 104–11. pmid:19917651
  3. 3. Dudley L. and Garner P., Strategies for integrating primary health services in low- and middle-income countries at the point of delivery. Cochrane Database Syst Rev, 2011. 7.
  4. 4. Ekman B., Pathmanathan I., and Liljestrand J., Integrating health interventions for women, newborn babies, and children: a framework for action. Lancet, 2008. 372(9642): p. 990–1000. pmid:18790321
  5. 5. Curry N. and Ham C., Clinical and service integration: the route to improved outcomes. 2010, The King's Fund: London.
  6. 6. Fleischman Foreit, K. G., Hardee K., and Agarwal K., When does it make sense to consider integrating STI and HIV services with family planning services? International Family Planning Perspectives, 2002. 28(2).
  7. 7. Askew I., Achieving synergies in prevention through linking sexual and reproductive health and HIV services, in International Conference on Actions to Strengthen Linkages between Sexual and Reproductive Health and HIV/AIDS. 2007: Mumbai, India.
  8. 8. Criel B., De Brouwere V., and Dugas S., Integration of vertical programmes in multi-function health services, in Studies in Health Services Organisation and Policy. 1997, ITG Press: Antwerp.
  9. 9. Maharaj P. and Cleland J., Integration of sexual and reproductive health services in KwaZulu-Natal, South Africa. Health Policy Plan, 2005. 20(5): p. 310–8. pmid:16113402
  10. 10. Mitchell, M., S.H. Mayhew, and I. Haivas, Integration revisited. Background paper to the report "Public choices, private decisions: sexual and reproductive health and the Millennium Development Goals". 2004, United Nations Millennium Project.
  11. 11. Sweeney S., Obure C.D., Maier C., Greener R., Dehne K., Vassall A. Costs and efficiency of integrating HIV/AIDS services with other health services: a systematic review of evidence and experience. Sex Transm Infect, 2012. 88(2): p. 85–99. pmid:22158934
  12. 12. Hoang T., Goetz M.B., Yano E.M., Rossman B., Anaya H.D., Knapp H. et al., The impact of integrated HIV care on patient health outcomes. Med Care, 2009. 47(5): p. 560–7. pmid:19318998
  13. 13. Uebel K.E., Joubert G., Wouters E., Mollentze W.F. and van Rensburg D., Integrating HIV care into primary care services: quantifying progress of an intervention in South Africa. PLoS One, 2013. 8(1): p. e54266. pmid:23349843
  14. 14. Brindis C.D., Loo V.S., Adlet N.E., Bolan G.A., Wasserheit J.N.., Service integration and teen friendliness in practice: a program assessment of sexual and reproductive health services for adolescents. J Adolesc Health, 2005. 37(2): p. 155–62. pmid:16026725
  15. 15. Kennedy C.E., Spaulding A.B. and Brickley D.B., Linking sexual and reproductive health and HIV interventions: a systematic review. J Int AIDS Soc, 2010. 13: p. 26. pmid:20642843
  16. 16. Church K. and Mayhew S.H., Integration of STI and HIV prevention, care, and treatment into family planning services: a review of the literature. Studies in Family Planning, 2009. 40(3): p. 171–186. pmid:19852408
  17. 17. Victora C.G., Habicht J.P., and Bryce J., Evidence-based public health: moving beyond randomized trials. Am J Public Health, 2004. 94(3): p. 400–5. pmid:14998803
  18. 18. Habicht J.P., Victora C.G., and Vaughan J.P., Evaluation designs for adequacy, plausibility and probability of public health programme performance and impact. Int J Epidemiol, 1999. 28(1): p. 10–8. pmid:10195658
  19. 19. Cousens S., Hargreaves J., Bonell C., Armstrong B., Thomas J. Kirkwood B.R.and Hayes R., Alternatives to randomisation in the evaluation of public-health interventions: statistical analysis and causal inference. J Epidemiol Community Health, 2011. 65(7): p. 576–81. pmid:19666633
  20. 20. Warren C.E., Mayhew S.H., Vassall A., Kimani J.K., Church K., Obure C.D et al., Study protocol for the Integra Initiative to assess the benefits and costs of integrating sexual and reproductive health and HIV services in Kenya and Swaziland. BMC Public Health, 2012. 12: p. 973. pmid:23148456
  21. 21. WHO, UNFPA and Unicef, Sexual and reproductive health & HIV/AIDS: a framework for priority linkages. 2005, World Health Organization: Geneva.
  22. 22. Bradley H., Bedada A., Tsui A., Barhmbhatt H., Gillespie D. and Kidanu A., HIV and family planning service integration and voluntary HIV counselling and testing client composition in Ethiopia. AIDS Care, 2008. 20(1): p. 61–71. pmid:18278616
  23. 23. Zwarenstein M., Fairall L.R., Lombard C., Mayars P., Bheekie A., English R.G. et al., Outreach education for integration of HIV/AIDS care, antiretroviral treatment, and tuberculosis care in primary care clinics in South Africa: PALSA PLUS pragmatic cluster randomised trial. BMJ, 2011. 342: p. d2022. pmid:21511783
  24. 24. Obure C.D., Sweeney S., Darsamo V., Michaels-Igbokwe C., Guinnes L., Terris-Prestholt F. et al., The costs of delivering integrated HIV and sexual reproductive health services inlimited resource settings. PLOS One 10(5): e0124476. pmid:25933414
  25. 25. Birdthistle I.J., Mayhew S.H., Kikuvi J., Zhou W., Church K., Warren C.,et al., Integration of HIV and maternal healthcare in a high HIV-prevalence setting: analysis of client flow data over time in Swaziland. BMJ Open, 2014. 4(3): p. e003715. pmid:24607560
  26. 26. Hasson F., Keeney S., and McKenna H., Research guidelines for the Delphi survey technique. J Adv Nurs, 2000. 32(4): p. 1008–15. pmid:11095242
  27. 27. Okoli C. and Pawlowski S.D., The Delphi method as a research tool: an example, design considerations and applications. Information & Management, 2004. 42(1): p. 15–29.
  28. 28. Rabe-Hesketh S. and Skrondal A., Classical latent variable models for medical research. Statistical Methods in Medical Research, 2008. 17(1): p. 5–32. pmid:17855748
  29. 29. Muthen L.K. and Muthen B.O., Mplus User's Guide. Sixth Edition, ed. Muthen M.. 1998–2010, Los Angeles, CA.
  30. 30. Gorsuch R.L., Factor Analysis. 2nd ed.. 1983, Hillsdale, N.J.: Lawrence Erlbaum.
  31. 31. Asparouhov, T. and B. Muthen, Plausible Values Using Mplus. Technical Report. 2010, Mplus: Los Angeles.
  32. 32. Mislevy R., Johnson E., and Muraki E., Scaling Procedures in NAEP. Journal of Educational Statistics, Special Issue: National Assessment of Educational Progress, 1992. 17(2): p. 131–154.
  33. 33. Turan J, Onono M., Steinfeld R., Shade S., Washington S, Bukusi E, et al. (2015) Effects of Antenatal Care and HIV Treatment Integration on Elements of the PMTCT Cascade: Results From the SHAIP Cluster-Randomized Controlled Trial in Kenya J Acquir Immune Defic Syndr Vol 69 No. 5: e172–e181
  34. 34. Grossman D, Ononoc M., Newmann S.J, Blat C., Bukusi E.A., Shade S.B. et al. (2013) Integration of family planning services into HIV care and treatment in Kenya: a cluster-randomized trial AIDS, 27 (Suppl 1):S77–S85 pmid:24088687
  35. 35. Sheikh K, George A. and Gilson L (2014) People-centred science: strengthening the practice of health policy and systems research. Health Research Policy and Systems 2014(12):19.
  36. 36. Okello D.R.O. and Gilson L. (2015) Exploring the influence of trust relationships on motivation in the health sector: a systematic review Human Resources for Health 13:16 pmid:25889952
  37. 37. Sweeney S, Obure C.D., Terris-Prestholt F., Darsamo V., Michaels-Igbokwe C., Muketo E., et al. (2014) How does HIV/SRH service integration impact workload? A descriptive analysis from the Integra Initiative in two African settings. Human Resources for Health 12:42.
  38. 38. Mayhew S.H., Collumbien M., Sweeney S., Colombini M., Mutemwa R., Lut I. et al. Numbers, people and multiple truths: Insights from a mixed methods non-randomised evaluation of integrated HIV and reproductive health services delivery in Kenya and Swaziland. Health Policy and Planning (Supplement) Forthcoming 2016
  39. 39. Mutemwa R., Mayhew S.H., Warren C.E., Abuya T., Ndwiga C., Kivunaga J. et al., Does service integration improve technical quality of care in low-resource settings? An evaluation of a model integrating HIV care into family planning services in Kenya Health Policy and Planning (Supplement) Forthcoming 2016.