Citation: Bates I, Phillips R, Martin-Peprah R, Kibiki G, Gaye O, Phiri K, et al. (2011) Assessing and Strengthening African Universities' Capacity for Doctoral Programmes. PLoS Med 8(9): e1001068. https://doi.org/10.1371/journal.pmed.1001068
Published: September 13, 2011
Copyright: © 2011 Bates et al. This is an open-access article distributed under the terms of the Creative Commons Attribution License, which permits unrestricted use, distribution, and reproduction in any medium, provided the original author and source are credited.
Funding: This study was funded by a grant from the Wellcome Trust, UK, to the London School of Hygiene and Tropical Medicine MCDC project (http://www.mcdconsortium.org/phd-programme.php). The funders had no role in study design, data collection and analysis, decision to publish, or preparation of the manuscript.
Competing interests: The authors have declared that no competing interests exist.
Abbreviations: MCDC, Malaria Capacity Development Consortium
Provenance: Not commissioned; externally peer reviewed
- Universities can make a major contribution to good policy-making by generating nationally relevant evidence, but little is known about how to strategically support universities in poorer countries to train and nurture sufficient internationally competitive researchers.
- It is difficult for universities to develop a coherent strategy to identify and remedy deficiencies in their doctoral training programmes because there is currently no single process that can be used to evaluate all the components needed to make these programmes successful.
- We have developed an evidence-based process for evaluating doctoral programmes from multiple perspectives that comprises an interview guide and a list of corroborating documents and facilities; we refined and validated this process by testing it in five diverse African universities.
- The strategy and priority list that emerged from the evaluation process facilitated “buy-in” from internal and external agencies and enabled each university to lead the development, implementation, and monitoring of their own strategy for remedying doctoral programme deficiencies.
Relationship between Indigenous Doctoral Programmes in Africa and Better Health Outcomes
The generation of local research, and the ability to innovate and to use research results, are essential for good policy making and ultimately for better health outcomes . Research for policy should be led by a country's own scientists ,, but very few universities in low-income African countries are able to “home grow” sufficient world class researchers . Traditionally, doctoral students from low-income countries have been trained overseas, where they often learn skills they cannot use when they return home. Consequently there has been a recent shift in funding emphasis towards supporting students to remain in their home institution.
Implicit in this new approach is the need to strengthen African universities' capacity to deliver doctoral programmes, and to focus efforts not only on the training itself but also on creating an enabling environment for research, in particular, leadership, career development, infrastructure, and access to information . Currently, the international academic community has insufficient understanding of the policies and processes required to develop research capacity in African universities, and this makes it difficult to target resources strategically towards priority capacity gaps .
Lack of Methodologies to Evaluate Gaps in Universities' Policies and Processes for Doctoral Programmes
To enhance the capacity of African universities to run doctoral programmes, the Malaria Capacity Development Consortium (MCDC)  has funded 19 African researchers to undertake doctoral training in universities in Ghana, Malawi, Senegal, Tanzania, and Uganda. The doctoral programmes in each of the African universities were at different stages of maturity at the start of the programme, with the number of registered doctoral students varying from none to 147. The doctoral programme coordinators in each institution recognised that having a cohort of new doctoral students provided an opportunity for their institutions to strengthen their systems for doctoral programmes or, for those universities just starting doctoral programmes, to develop the necessary structures and processes. The African project coordinators therefore asked the MCDC secretariat for support in identifying ways in which their doctoral programmes could be improved. The project was advertised, and following a selection process, a contract to evaluate the African doctoral programmes was awarded to a team of researchers from the United Kingdom and Africa (IB, RP, RM-P, and SP).
A process was needed to identify gaps in the African institutions' existing doctoral programmes, so discussions between representatives from the universities and our research team defined the critieria that should be met by such a process (Box 1). However, published information about evaluating capacity development is very scarce , and a search of the literature failed to identify any single process that could be used to evaluate all the policies and processes needed to run successful doctoral programmes. The purpose of our study was therefore to develop and test an evidence-based process that could be used to evaluate all the components of doctoral programmes and to standardise the process so that it is transferable across universities.
Box 1. Criteria to Be Fulfilled by the Evaluation Process
The process should enable the university authorities to
- review all aspects of their doctoral programmes
- identify gaps in their capacity to manage these programmes
- develop strategies to remedy these gaps
- generate indicators that can be used to evaluate progress in filling these gaps
Development of the Process for Evaluating Policies and Systems for Doctoral Programmes in Africa
To ensure that the final evaluation process was robust, it was derived from published evidence. We scanned published literature, Web sites and documents from universities, educational agencies, and regulatory bodies to identify and synthesize existing methods for evaluating any aspect of doctoral programmes (Box 2). Through this process we produced a list of all the policies, processes, and facilities needed to run doctoral programmes (summarised in Figure 1). We amalgamated all the information obtained into a draft evaluation process, which consisted of a list of stakeholders to be interviewed, an interview guide for each of the different cadres of stakeholders, a list of documents to be reviewed, and a list of facilities to be visited.
Box 2. Examples of Sources of Information about Doctoral Policies and Processes
- Code of practice from the UK Quality Assurance Agency  (used as a platform for incorporating other pieces of information)
- Institutional quality standards for the contents of doctoral programmes and research skills needed by doctoral students –
- Handbook and checklist for managing quality assurance in education programmes 
- Framework for conducting an assessment of institutional health research capacity 
- Personal development plan for African doctoral students 
Testing and Finalising the Evaluation Process
The evaluation took place during site visits of 2–3 days to each of the five African universities that are partners in MCDC. These visits were preceded by pro-active engagement of key individuals in the universities who would facilitate the evaluations. The research team conducting the evaluation was independent of the MCDC managers, and the team's four members had expertise in health care delivery, research and doctoral supervision, academic and health care systems, and educational development. Interview bias was reduced by using different combinations of research team members to conduct the interviews.
Initially no assumptions were made about which questions could be answered by which interviewee, and all interviewees were asked every question. Interviewees were enthusiastic about being interviewed, but they were able to provide reliable information only about the aspects of doctoral programmes in which they were directly involved, and so there were many instances where inaccurate or incomplete information was provided. We were therefore meticulous about double-checking all the information we were given by asking the same question of more than one individual. Responses from the interviewees were corroborated by referring to institutional documents and directly observing facilities as appropriate. Any discrepancies were resolved through discussions between the researchers and the doctoral programme coordinators in each institution.
As the evaluations progressed we were able to match questions to specific types of interviewees, thereby improving the efficiency of the interviews. By honing the questions, removing duplications and adding new interviewees or observations to the process, as suggested by interviewees, we refined and streamlined the process and reduced the interview times from over 60 minutes to around 20 minutes. No new interviews or observations were added to the process after the third site visit, indicating the we had reached saturation for the interview guide. In total across the five medical schools we interviewed 83 individuals involved in all aspects of doctoral programmes, reviewed 40 documents, and visited laboratories, libraries, computer centres, and field sites.
The final evaluation process consisted of interviews using the developed guide (i.e., a grid with questions mapped against specific interviewees) (see Text S1), and a review of documents (e.g., policies, regulations, course handbooks) and facilities (e.g., laboratories, libraries, computer centres) to corroborate information from the interviews. The list of interviewees comprised university policy makers (e.g., principals, provosts, deans), researchers (e.g., doctoral students, post-doctoral scientists, research supervisors, research centre staff), and support staff (e.g., ethicists, administrators, accountants, librarians, laboratory scientists). The questions in the final interview guide were compared to the initial list of doctoral programme components we had extracted from the literature to ensure that the iterative adaptations of the process had not resulted in any major omissions.
Outcomes of the Evaluation Process
At the end of the site visits the doctoral programme coordinator at each university was provided with a confidential report about their own institution and also an anonymised overview report that amalgamated and summarised the key findings from the individual institutional reports. The institutional reports provided a narrative account of corroborated interviewee responses, a list of gaps in doctoral programme provision, and potential remedies proposed by the interviewees. It was possible to identify gaps in provision because our evaluation process was based on an extensive literature review and was therefore a “benchmark” that contained all the elements needed to run a successful doctoral programme.
Following the evaluation process the institutions used the evaluations to develop and implement their own plans to remedy the gaps in capacity and to derive indicators that could be used to monitor progress. Evidence suggests that these indicators will need to be revised regularly as the universities' doctoral programmes mature and become more sophisticated . The capacity gaps that occurred in more than one institution, and potential remedies to address these gaps, have been amalgamated in Table 1. Interestingly, over half of these gaps could potentially be addressed by the universities themselves without any additional external resources (e.g., see Box 3). To address other capacity gaps, such as lack of access to suitable electronic resources and the inexperience of many of the supervisors, additional external inputs would be required (e.g., see Box 4). Using the recommendations in the report, each university coordinator prioritised the areas identified as needing support, identified the steps that would be needed to fill these gaps, and estimated the costs of doing so. The MCDC programme will be able to finance some of these action plans. By applying this standardised evaluation process to several institutions it was also possible to identify common capacity gaps (Box 5), which could become the focus of cross-institutional efforts by external agencies.
Box 3. Case Study 1
The doctoral programme evaluation was conducted in an African university that had just started to run its own doctoral programmes. The evaluation revealed that information for students about what was expected of them, how the programme was organised, and what resources were available to them was lacking. Some documents were available in the university, but they were incomplete, not specifically for students in the medical college, and difficult for students to access. The programme coordinator reviewed doctoral students' handbooks from several sources and also used our doctoral programme evaluation interview guide to develop a comprehensive student handbook for doctoral students, which was made available on their intranet.
Box 4. Case Study 2
One of the universities had a particular problem with slow and unreliable Internet access. Although this had long been recognised as a problem in the institution, the doctoral evaluation process revealed that every cadre of interviewee mentioned it as a major problem, not only for research students but also for undergraduates and staff. This weight of evidence collected by an external team enabled the university to make a case for several funders to join forces to provide a fast broadband connection for the university.
Box 5. Common Gaps in Capacity for Doctoral Programmes in African Universities
- Incomplete supporting documentation such as policies, regulations, and handbooks; documents sometimes inappropriately combined with handbooks for Master's courses and often not easily available to students and/or staff
- Lack of suitably qualified academic faculty with experience of supervising doctoral students
- Inadequate resources (such as books, journals, computers, and Internet access) to support doctoral programmes; staff responsible for providing these resources were generally not involved in planning for the doctoral programme, so they were not able to cater to the needs of these students
- Lack of dedicated desk space for doctoral students and very little formal opportunity for mutual support
- Lack of a formal induction programme to make students aware of, for example, the requirements of the programme or the availability of resources to support their studies
- Lack of a systematic skills development programme within the institution for either supervisors or students
- Unclear mechanisms for identifying and managing students who were failing to progress or who had missed their completion dates
- Excessive time taken to complete the final examination process (often exceeded 12 months)
- Students being unaware of appeal processes that were independent of their supervisor(s)
- Lack of mechanisms for soliciting feedback from students and staff or for using this routinely to enhance the programme
Limitations of the Evaluation Process
The quantitative indicators we identified for monitoring progress in strengthening institutional capacity (e.g., number of students; time to complete course; number of publications, grants, or presentations) can be measured relatively easily, but they do not adequately capture factors that contribute to developing an enabling environment for research . Our evaluation process therefore also included some qualitative indicators (e.g., student satisfaction, quality of learning spaces), although we recognise that these may be more difficult to measure than quantitative indicators. This strategy of combining qualitative and quantitative indicators is similar to the approach taken in the few other published studies that have evaluated research capacity development .
The evaluation process was developed with and for health faculties in universities in developing African countries, and it has not been evaluated beyond this context. Nevertheless, because the key components outlined in the overview (Figure 1) were derived from the global literature, the process is likely to be applicable to doctoral programmes in faculties and universities outside Africa. However, the specific types of capacity gaps may vary between countries with different levels of socio-economic development (e.g., slow Internet access and lack of doctoral research supervisors were common gaps in our five African institutions, but this may not be the case in other regions).
We have developed a comprehensive evidence-based process for evaluating all the policies and systems required for doctoral programmes. Unlike previously published methods for evaluating doctoral programmes, our process incorporated the perspectives of students, staff, the local research community, and the universities' policy makers and was applicable across different countries and programmes of differing maturity. Our standardised evaluation process not only enabled the universities to develop and monitor strategies to address their own capacity gaps but also provided them with a mechanism for justifying, planning, commissioning, and monitoring inputs by external funders while retaining leadership of the process. Unlike the traditional, externally controlled accountability imposed by international donors' agendas, there is evidence that this “endogenous accountability” type of monitoring is likely to promote better ownership and performance .
We are grateful to the staff and students from the five universities, including Moses Kamya of Makerere University in Uganda, who provided information for this study or who provided logistic or administrative support.
Conceived and designed the project: IB SP. Performed the project: IB SP RP RMP. Analyzed the data: IB SP RP RMP. Contributed reagents/materials/analysis tools: IB SP RP RMP GK OG KP HT. Wrote the first draft of the manuscript: IB. Contributed to the writing of the manuscript: IB SP RP RMP. ICMJE criteria for authorship read and met: IB SP RP RMP GK OG KP HT. Agree with manuscript′s results and conclusions: IB SP RP RMP GK OG KP HT.
- 1. Ofori-Adjei D, Gyapong J, Box HML (2009) Capacity building for relevant health research in developing countries. In: Engelhard R, editor. Knowledge on the move: emerging agendas for development-oriented research. The Hague: International Development Publications. pp. 178–184. Available: http://www.nwo.nl/nwohome.nsf/pages/NWOP_7YCJVK_Eng. Accessed 3 July 2011.
- 2. Nuyens Y (2005) No development without research: a challenge for research capacity strengthening. Geneva: Global Forum for Health Research.
- 3. Horton R, Pang T (2008) Health, development, and equity-call for papers. Lancet 371: 101–102.
- 4. Fonn S (2006) African PhD research capacity in public health: raison d'etre and how to build it. In: Matlin S, editor. Global forum update on research for health volume 3: combating disease and promoting health. Geneva: Global Forum for Health Research. pp. 80–83.
- 5. Land T, Hauck V, Baser H (2009) Capacity development: between planned interventions and emergent processes. Implications for development cooperation. Policy management brief No. 22. Maastricht (The Netherlands): European Centre for Development Policy Management. Available: http://www.ecdpm.org/Web_ECDPM/Web/Content/Download.nsf/0/5E619EA3431DE022C12575990029E824/$FILE/PMB22_e_CDapproaches-capacitystudy.pdf. Accessed 25 January 2011.
- 6. Malaria Capacity Development Consortium (2011) PhD programme. London: Malaria Capacity Development Consortium. Available: http://www.mcdconsortium.org/phd-programme.php. Accessed 3 July 2011.
- 7. Kakuma R, Cole DC, Bates I, Ezeh A, Fonn S, et al. (2010) Evaluating capacity development in global health research—where is the evidence? [abstract]. Global Health Forum for Health Research First Global Symposium for Health Systems Research; 16–19 November 2010; Montreux, Switzerland.
- 8. Baser H, Morgan P (2008) Capacity, change and performance: study report. Maastricht (The Netherlands): European Centre for Development Policy Management. Available: http://www.ausaid.gov.au/hottopics/pdf/capacity_change_performance_final_report.pdf. Accessed 25 January 2011.
- 9. Cooke J (2005) A framework to evaluate research capacity building in health care. BMC Fam Pract 6: 44. Available: http://www.ncbi.nlm.nih.gov/pmc/articles/PMC1289281/pdf/1471-2296-6-44.pdf. Accessed 25 January 2011.
- 10. Watson D (2006) Monitoring of capacity and capacity development. Discussion paper 58B. Maastricht (The Netherlands): European Centre for Development Policy Management. Available: http://www.ecdpm.org/Web_ECDPM/Web/Content/Content.nsf/vwDocID/59833D39F5B7DBB2C12570B5004DCB92?OpenDocument. Accessed 25 January 2011.
- 11. Quality Assurance Agency (2004) Code of practice for the assurance of academic quality and standards in higher education. Section 1: postgraduate research programmes. Gloucester: Quality Assurance Agency. Available: http://www.qaa.ac.uk/Publications/InformationAndGuidance/Pages/Code-of-practice-section-1.aspx. Accessed 7 September 2011.
- 12. Watson D (2010) Embracing innovative practice. Capacity.org. Available: http://www.capacity.org/capacity/opencms/en/topics/monitoring-and-evaluation/embracing-innovative-practice.html. Accessed 3 July 2011.
- 13. UK Research Councils, Arts and Humanities Research BoardJoint statement of the Research Councils'/AHRB's skills training requirements for research students.
- 14. European Network of Excellence on Aspect-Oriented Software Development (2006) Guidelines for AOSD-certified PhD Programme. Available: http://www.aosd-europe.net/deliverables/d30.pdf. Accessed 25 January 2011.
- 15. Nabwera H, Purnell S, Bates I (2008) Development of a quality assurance handbook to improve educational courses in Africa. Hum Resour Health 6: 28.
- 16. Bates I, Akoto AY, Ansong D, Karikari P, Bedu-Addo G, et al. (2006) Evaluating health research capacity building: an evidence-based tool. PLoS Med 3: e299.
- 17. McCullough H (2011) Using personal development planning for career development with research scientists based in sub-Saharan Africa [PhD dissertation]. Liverpool (United Kingdom): Disease Control Strategy Group of the Liverpool School of Tropical Medicine, University of Liverpool.