Skip to main content
  • Loading metrics

Testing approaches to sharing trial results with participants: The Show RESPECT cluster randomised, factorial, mixed methods trial



Sharing trial results with participants is an ethical imperative but often does not happen. We tested an Enhanced Webpage versus a Basic Webpage, Mailed Printed Summary versus no Mailed Printed Summary, and Email List Invitation versus no Email List Invitation to see which approach resulted in the highest patient satisfaction with how the results were communicated.

Methods and findings

We carried out a cluster randomised, 2 by 2 by 2 factorial, nonblinded study within a trial, with semistructured qualitative interviews with some patients (ISRCTN96189403). Each cluster was a UK hospital participating in the ICON8 ovarian cancer trial. Interventions were shared with 384 ICON8 participants who were alive and considered well enough to be contacted, at 43 hospitals. Hospitals were allocated to share results with participants through one of the 8 intervention combinations based on random permutation within blocks of 8, stratified by number of participants. All interventions contained a written plain English summary of the results. The Enhanced Webpage also contained a short video. Both the Enhanced Webpage and Email contained links to further information and support. The Mailed Printed Summary was opt-out.

Follow-up questionnaires were sent 1 month after patients had been offered the interventions. Patients’ reported satisfaction was measured using a 5-point scale, analysed by ordinal logistic regression estimating main effects for all 3 interventions, with random effects for site, restricted to those who reported receiving the results and assuming no interaction. Data collection took place in 2018 to 2019.

Questionnaires were sent to 275/384 randomly selected participants and returned by 180: 90/142 allocated Basic Webpage, 90/133 Enhanced Webpage; 91/141 no Mailed Printed Summary, 89/134 Mailed Printed Summary; 82/129 no Email List Invitation, 98/146 Email List Invitation. Only 3 patients opted out of receiving the Mailed Printed Summary; no patients signed up to the email list. Patients’ satisfaction was greater at sites allocated the Mailed Printed Summary, where 65/81 (80%) were quite or very satisfied compared to sites with no Mailed Printed Summary 39/64 (61%), ordinal odds ratio (OR) = 3.15 (1.66 to 5.98, p < 0.001). We found no effect on patient satisfaction from the Enhanced Webpage, OR = 1.47 (0.78 to 2.76, p = 0.235) or Email List Invitation, OR = 1.38 (0.72 to 2.63, p = 0.327). Interviewees described the results as interesting, important, and disappointing (the ICON8 trial found no benefit). Finding out the results made some feel their trial participation had been more worthwhile. Regardless of allocated group, patients who received results generally reported that the information was easy to understand and find, were glad and did not regret finding out the results. The main limitation of our study is the 65% response rate.


Nearly all respondents wanted to know the results and were glad to receive them. Adding an opt-out Mailed Printed Summary alongside a webpage yielded the highest reported satisfaction. This study provides evidence on how to share results with other similar trial populations. Further research is needed to look at different results scenarios and patient populations.

Trial registration


Author summary

Why was this study done?

  • Previous research has shown that most people who take part in clinical trials want to be told the results of those trials, but many participants never get to find them out.
  • There is little evidence to guide researchers on how best to share results with the people taking part in their trials.

What did the researchers do and find?

  • We carried out a study to test different ways of sharing trial results with participants in an ovarian cancer trial.
  • We randomly assigned hospitals that were part of the ovarian cancer trial to share results with the women taking part in different ways: a basic webpage or an enhanced webpage; a printed summary of the results by mail; and an email list to receive the results.
  • Nine in 10 women wanted to be told the results of the trial they had taken part in.
  • Women at hospitals which sent out the printed summary by mail, were more likely to be satisfied with how the results were shared and were more likely to find out the results than those at other hospitals.
  • Women who received the results said that the information was easy to understand and find, were glad and did not regret finding out the results.

What do these findings mean?

  • These findings suggest that trials with similar participants to our ovarian cancer trial (mainly women aged 50 or older), where webpages are used to share results with people taking part, should also share results through opt-out mailed printed summaries.
  • This will enable more people who want to know the results to find them out, and improve satisfaction.


Sharing results with people who have taken part in trials is an ethical imperative [1], with the Declaration of Helsinki saying “All medical research subjects should be given the option of being informed about the general outcome and results of the study” [2]. Doing this demonstrates respect for their contribution, with some suggestion that it may increase the likelihood of participants taking part in future medical research, or recommending taking part in trials to others [37]. Studies have repeatedly shown that, while most participants want to receive results [3,812], many are not offered the opportunity to receive them [1315].

Trial teams may be uncertain about which method to use for sharing results with participants. Most of the current evidence is based on surveys of participants or the public, prospectively asking how they would prefer to be informed, or retrospectively asking whether an approach that was used was acceptable or understandable, rather than systematically comparing outcomes from different approaches [36,8,11,12,14,1621]. Most of the published evidence to date relates to sharing results with participants via mailed letters or leaflets; these studies generally report high acceptability of this approach [3,5,6,8,9,12,14,17,18,22,23]. However, sending out results by mail has resource implications. Sharing results via webpages has a number of potential advantages, including the ability to offer links to further information or support, and include audio and visual content alongside written summaries, and being discoverable by participants who have been lost to follow-up. There are also potential drawbacks in terms of accessibility for populations with low computer literacy. Fewer studies have reported sharing results via webpages. Mancini and colleagues randomised participants in a breast cancer trial to receive a letter containing a link to a website with the trial results, or no letter. They found that participants who received the letter had better understanding of the results but were not significantly more likely to have received the trial results than participants who did not receive the letter [24]. Other studies have reported low uptake of results shared via webpages [14,18], or lower levels of satisfaction with how the results were shared [12]. There is less evidence around the use of email to share results with participants; however, some studies have found that potential research participants would be happy to receive results that way [16]. Other approaches to sharing results that have been reported include face-to-face meetings [12,18,21], teleconferences [20], and individual telephone calls or helpline services [3,20,22], although the resource requirements for these approaches may be prohibitive, particularly to large trials, and uptake of these services may be low, with Dixon-Woods and colleagues reporting no calls to a telephone helpline [22].

ICON8 (ISRCTN10356387) was a Phase III randomised controlled trial looking at 3 chemotherapy regimens for up-front treatment of ovarian cancer. Results from the earlier of the 2 co-primary endpoints, progression-free survival, were published in 2019 [25], showing no difference in progression-free survival between the 3 regimens. The Show Results to Participants Engaged in Clinical Trials (Show RESPECT) study (ISRCTN96189403) sought to generate evidence to inform trial teams on how to share results with trial participants through a mixed methods cluster randomised factorial study within the ICON8 trial. Show RESPECT tested the following 3 hypotheses, in terms of participant satisfaction with how the results were communicated:

  1. An Enhanced Webpage will be superior to a Basic Webpage;
  2. A Mailed Printed Summary sent by post will be superior to no Mailed Printed Summary; and
  3. An invitation to join an Email List to receive updates about the trial results will be superior to no invitation to join an Email List.


Show RESPECT was a mixed methods study, comprised of a factorial cluster randomised controlled trial within a trial to assess multiple approaches to communicating trial results, and embedded explanatory qualitative study. The data collection period for the quantitative and qualitative components was concurrent. Patients were identified for the semistructured interviews from their questionnaire responses and contact form returned alongside the questionnaire, so interviews took place after quantitative data collection for those individuals (while quantitative data collection continued for others). This paper reports both qualitative and quantitative results from data collected from trial participants. We consider the qualitative and quantitative data to have equal weight in their contribution to addressing the research aims. The full protocol for the study is available online [26] and as S4 Appendix.

We also collected data from site staff, but results from that part of the study are beyond the scope of this paper.

Ethics approval

The study obtained ethics approval from the London-Chelsea Research Ethics Committee, MREC number 18/LO/1011.

Patient and public involvement

Substantial patient and public involvement (PPI) was carried out to design and conduct this study, including focus groups, a PPI survey, patient representation on the study steering group, and input from patient groups and individuals on the design and content of the interventions.

Quantitative methods

Trial design.

Show RESPECT was a cluster randomised 2 by 2 by 2 factorial trial within a host trial, ICON8, an RCT evaluating chemotherapy schedules in ovarian cancer. We randomised each United Kingdom trial site (secondary or tertiary hospital) in ICON8 that agreed to take part in the Show RESPECT study to a combination of interventions to feedback ICON8 trial results to participants, as shown in Fig 1. A cluster design was chosen for this study, as it was felt that implementing individual randomisation for sharing results would be impractical for sites. Each site was a cluster. Allocation to each intervention was on a 1:1 ratio.

Fig 1. Show RESPECT trial schema.

Diagram showing the 3 randomisations within Show RESPECT: (1) Link to Basic Webpage or Enhanced Webpage; (2) No Mailed Printed Summary or Mailed Printed Summary; and (3) No invitation to join Email List or Invitation to join Email List.


Participating sites sent all ICON8 patients at those sites a printed Patient Update Information Sheet thanking them for taking part in ICON8, reminding them of the aims of the ICON8 trial, informing them that trial results were now available, and how they could access them. This included the URL of their randomised webpage (Basic or Enhanced). The Patient Update Information Sheet told patients at sites randomised to the Mailed Printed Summary that they would be sent a Mailed Printed Summary of the results after 3 weeks and that they should let their ICON8 site team know if they did not want to be sent this. Patients at sites randomised to the Email List Invitation were given a URL to sign up to the email list. The Patient Update Information Sheet was based on guidance from the Health Research Authority on End of Study Information Sheets [27]. S1 Table contains a detailed description of the study interventions, and links to the Basic and Enhanced Webpages. The Patient Update Information Sheet (S1 Appendix), Mailed Printed Summary (S2 Appendix), and results emails (S3 Appendix) can be found in the Supporting information.

Randomisation 1—All participants in Show RESPECT received a link to either the Basic or Enhanced Webpage. The Basic Webpage contained a plain English summary of results, using the structure and headings recommended for lay summaries in the EU database of clinical trials [28]. The Enhanced Webpage used a structure adapted from the Multi-Regional Clinical Trials Center guidance on feedback of results [29], and included a short video of a doctor explaining the results, links to further information and support, 2 graphics showing the trial treatment schedules and main side effects, and the opportunity to submit questions to be answered in the page’s Frequently Asked Questions (FAQ) section.

Randomisation 2—The Mailed Printed Summary followed the same structure as the Enhanced Webpage, without the video or FAQ section, and was sent by post to participants’ homes. The Patient Update Information Sheet for sites randomised to no Mailed Printed Summary told patients that if they were unable to access the webpage or email list, and wanted to find out the results, they should contact their research nurse.

Randomisation 3—The first email sent to those who signed up to the email list followed the same structure as the Enhanced Webpage, without the video. Participants were invited to submit questions about the results, which were answered in subsequent emails.


Show RESPECT collected data from women with ovarian cancer who had taken part in the ICON8 trial, were currently still alive, and in follow-up at a site participating in Show RESPECT. ICON8 participants were not invited to join Show RESPECT if they were considered by site staff to be too unwell to be contacted about this study.


Our primary outcome measure was participants’ reported satisfaction with how the results were communicated to them, measured using a 5-point Likert-type scale (1 = very unsatisfied; 5 = very satisfied). Secondary outcome measures collected from participants were: the proportion of participants wanting to know the results that did find out; ease of finding out the results; whether the information about the trial results told participants everything they wanted to know; ease of understanding the results; how upsetting participants found the results; willingness to take part in a future trial; likelihood of recommending taking part in a clinical trial to friends and family; whether participants felt glad to have found out the results; and whether participants regretted finding out the results. Apart from proportion of those who wanted to know the results who found them out, these were measured using separate 5-point Likert-type scales. Quantitative data were collected from ICON8 participants by a questionnaire sent by site staff to their home address.

Data were collected between December 2018 to September 2019. Data collection finished 4 months after the final randomisation as it was felt that longer follow-up would run the risk of participants being unable to recall their experience of receiving results accurately.

Sample size.

At trial sites, the allocated Show RESPECT intervention was offered to all eligible ICON8 participants (through the Patient Update Information Sheet). However, we did not approach all eligible participants for data collection, so as to reduce the burden on participants and staff, and because in cluster randomised trials the marginal information value of each participant declines as cluster size increases [30]. Specifically, at small sites (≤5 eligible participants), all eligible participants were invited to provide outcome data, but at medium sites (6 to 12), we aimed to collect outcome data from 6 participants, and from large sites (≥13), we aimed to collect data from 12. For medium and large sites, the individuals invited to participate were selected at random centrally. At medium and large sites, if a participant who was invited to take part chose not to, we invited the next participant from a randomly ordered, centrally held list to take part to replace the original participant, until the target number of participants at that site was reached, or no eligible participants remained.

The primary outcome measure was ordinal but for simplicity, because of lack of knowledge of its likely distribution, and to be scientifically conservative, we considered it as a binary outcome for our power calculations. We anticipated that the proportion of respondents “satisfied” without any of the research interventions would be between 20% and 80%, and in the absence of specific prior information considered values of the Intracluster Correlation Coefficient (ICC) between 0.01 and 0.05. We considered power to detect an effect for any of the 3 interventions, for simplicity considering each in turn, i.e., effectively conducting a power calculation for each intervention assuming the other two would have no effect. We also assumed no appreciable interactions between the 3 interventions. We calculated that, based on 21 sites with and without an intervention, and an average of 4 respondents per site (172 in total), at an ICC of 0.01, we would have 80% power to detect an increase from 20% to 40%, from 50% to 71%, or from 80% to 95% in the satisfied group. Should the ICC be 0.05, then this sample size would have provided 80% power to detect an increase from 20% to 42%, 50% to 73%, or 80% to 95%. Calculations were conducted in Stata using the “power two proportions” command and assumed a coefficient of variation in cluster size of 0.6. No power calculations were made for the secondary outcomes.


Sites were randomised in blocks of 8 (the number of allocation arms available) once sites had obtained the necessary approvals. In the first phase, we randomised 3 blocks each of one site size (small, medium, and large), but, subsequently, blocks were of mixed sizes. Randomisation was conducted through random permutation within blocks.

To ensure allocation blinding, although the Show RESPECT trial statistician generated the allocations for the blocks and was aware of which clinics featured in each block, a second statistician unaware of these allocations randomly permuted the clinic names within blocks. The allocations and clinic names for each block were then matched together by a third party and revealed to the trial team. Sites were informed of their randomised allocation and sent the matching Patient Update Information Sheet.

Clusters were recruited between September 2018 to May 2019. Sites were randomised between November 2018 to May 2019. The trial was registered in February 2019, which was after some sites had been randomised, due to human resource constraints.


Once randomisation had been performed, it was not possible to blind site staff to the allocation of their site. ICON8 participants were not informed that the way they were being offered the results was determined by randomisation and were not aware of the interventions being offered to participants at other sites. The questionnaire contained an embedded informed consent element, in line with the UK Health Research Authority’s guidance on proportionate approaches to informed consent for self-administered questionnaire-based research [31], with completion and return of the questionnaire taken to indicate consent to use the data has been given.

Statistical analysis.

The full statistical analysis plan can be found in S5 Appendix. The primary outcome measure was defined only for participants who received the ICON8 trial results, and hence analysis for this outcome was restricted to participants who reported receiving the ICON8 results. For this reason, we describe the primary analysis as following modified intention to treat (mITT). All other secondary outcomes are similarly only defined for participants who received the ICON8 results, with the exception of “report finding out the ICON8 results,” which we present separately among participants who report they wanted to find the results out, and among participants who report they did not. To assess the overall effect of the intervention, it is important to interpret the results of the primary analysis alongside results concerning the possible effect of the interventions on whether participants actually found out the ICON8 results.

In the ICON8 setting, participants’ health may be poor and may deteriorate before the Show RESPECT interventions were received or between intervention exposure and follow-up by questionnaire. Participants who died or became too sick to complete a questionnaire were not considered “eligible” for data collection or analysis and were not considered as missing data.

There was no specific prior evidence to suggest whether or not there would be interactions between the interventions. We were unable to think of a mechanism for potential interactions, so designed the trial based on the assumption that the effect of each intervention (e.g., enhanced versus basic webpage) would not be substantially affected by whether or not the participant was allocated to the other interventions. Hence, the primary analysis was of the main effects of each intervention adjusting for the others. However, for the primary outcome measure, we also tested each of the 3 two-way interactions and report the effect of each of 7 intervention combinations relative to control (Basic Webpage only). Adjustments were not made for multiple testing as we view our 3 study hypotheses as distinct, so all confidence intervals (CIs) presented are at the standard 5% significance level.

To reflect the study design, we adjusted for site size stratum, and also first phase versus later randomisation phases. All models included random effects for site. Estimates were also adjusted for age (continuous–linear), education (graduate versus not), and internet use (daily versus less).

Effect measures for the interventions are estimated and presented based on regression models. Ordinal random effects logistic regression was used for the primary and other Likert-type scale outcomes unless the proportional odds assumption was clearly violated. Consider odds ratios (ORs) in relation to each way the outcome could be dichotomised, e.g., quite unsatisfied or better versus very unsatisfied, quite or very satisfied versus neither or worse. Under the proportional odds assumption, these ORs are all equal, and their common value is estimated through ordinal logistic regression. The response categories were merged for the regression analysis in the event of very low reporting of one or more categories (<5% of responses). All decisions about merging response categories were taken based on an initial dataset without cluster or allocation identifiers.

For the primary outcome measure only, we conducted prespecified subgroup analyses by age group (≤70 versus ≥71 years), allocated arm of the ICON8 trial, education category (graduate versus not), and reported internet use (daily versus not). For each subgroup analysis, the effect of each intervention within subgroups were presented, and an interaction test was conducted. All interactions were with binary subgroups, with the exception of age, which was used as a continuous variable. These subgroup analyses were conducted for each of the 3 interventions separately.

Statistical analyses were performed using Stata version 16.1 (Stata Corp, Texas).

Qualitative methods

Qualitative data collection.

Semistructured interviews were carried out with participants either face-to-face (at the participant’s home or other location chosen by them) or by telephone by the lead qualitative researcher, AS, who holds an MPhil and MSc, is a research communication specialist, is female, and has been trained in qualitative research methods. The interviews were informed by a topic guide (S1 Text), which was informed by PPI. The interviewer is a research communicator by profession and was involved in developing the interventions tested in Show RESPECT. The topic guide was amended as interviews proceeded to follow-up on issues that emerged in early interviews and to improve clarity [32]. Only the participant and interviewer were present during the interviews. No repeat interviews were carried out. Interviews lasted between 32 minutes to 102 minutes. The interviews were audio recorded, and field notes were made immediately after the interviews. The interviews were transcribed verbatim. The transcriptions were checked back against the recordings for accuracy, and any identifying data were redacted. Transcripts were not returned to participants. Free-text questions within the questionnaire were also used to collect qualitative data.


Invitations to take part in interviews were sent out with the Show RESPECT questionnaire, with participants asked to complete a contact details form if they wanted to find out more about the interviews and return it alongside their questionnaire. Purposive sampling was carried out, based on their questionnaire responses, to include people offered the range of Show RESPECT interventions, different levels of satisfaction with how the results were communicated, education level, internet usage, and age. Respondents who completed the contact details form and filled one or more gaps in the sampling frame were contacted by telephone with more information about the study and, if they were willing to take part, a time and date was arranged for the interview. Participants gave written informed consent. Interviews were carried out until all the gaps in the sampling frame were filled, or until no more volunteers were available who would fill a gap in the sampling frame. Using the Information Power model [33] to assess the necessary sample size, the study aim was reasonably narrow, focusing on just one aspect of trial experience (receiving results), although it did look at several approaches to results communication. The sample specificity was dense, with all interviewees having highly relevant experiences. As described in the analysis section below, an established model was applied during the analysis. The quality of dialogue in most interviews was good, resulting in a rich dataset. The analysis strategy was cross-case. Taken together, these factors suggest that a moderate sample size should provide sufficient information power to meet the aims of the study.

Qualitative analysis.

The first step of analysis was familiarisation with the data, by listening to the recordings and reading the transcripts a number of times, recording ideas for initial codes. A thematic analysis approach was employed [34]. Both inductive and deductive approaches were for coding the data, which was carried out by AS. Initial codes were then grouped into potential themes. Emerging themes were discussed with staff from the ICON8 and Show RESPECT trial management teams. As analysis proceeded, it was found that the Information-seeking and Communication Model (ISCM) [35] fitted the codes well, so codes were categorised using concepts from that model. The ISCM is a model of information behaviour that covers both information users and information providers, their contexts, the activities of information seeking, information use and communication, and the factors that affect them [36]. Network diagrams were produced to visualise links between codes within themes, and themes were reviewed in relation to the coded extracts. As themes were generated, we searched for cases which did not fit the existing structure. Inductive thematic saturation was reached at the 13th interview, as was data saturation. Participant checking did not take place, but a PPI discussion group was held to reflect on the emerging findings and interpretation. Analysis was conducted in Atlas.ti version 8.4 (ATLAS.ti Scientific Software Development GmbH).

A “following the thread” approach was used to triangulate the results of the qualitative and quantitative components of the study [37]. This was done at the analysis stage. Each data set was initially analysed using approaches applicable to the type of data to identify key themes and questions. The qualitative data were then interrogated to explore issues raised in the quantitative data (following the “thread” from one dataset to the other).


Participation in Show RESPECT

Fig 2 shows the CONSORT diagram for the study. The 83 ICON8 sites in the UK were assessed for eligibility. Approximately 40 sites were excluded for reasons including lack of ICON8 participants eligible for Show RESPECT (5), lack of capacity (6), declining to take part (4), failing to obtain site approvals in time (12), or nonresponse to the invitation (13). About 43 (52%) ICON8 UK sites took part in Show RESPECT. Table 1 shows the number of sites randomised to the interventions by site size strata, number of eligible participants who were offered the interventions, sent the questionnaire, and returned the questionnaire. Data collection took place between December 2018 and September 2019. In total, 384 ICON8 participants were offered the Show RESPECT interventions; 275 were sent the questionnaire of which 182 questionnaires were returned from 180 participants (65%) (2 of the 182 returned questionnaires were duplicates so not analysed).

Fig 2. CONSORT diagram for Show RESPECT.

CONSORT diagram showing flow of sites and participants through the Show RESPECT study.

Table 1. Recruitment of sites and participants to Show RESPECT, by site size1.

Delivery of the interventions.

Logs kept by sites showed that Patient Update Information Sheets went to 100% of eligible ICON8 participants at participating sites. Three ICON8 participants opted out of receiving the Mailed Printed Summary. According to site logs, all other eligible ICON8 participants at sites randomised to Mailed Printed Summaries were sent them.

Baseline characteristics of participants.

Table 2 shows the baseline characteristics of those who returned the questionnaire, and S2 Table shows the baseline characteristics of all eligible ICON8 participants at Show RESPECT sites. The mean age of participants who returned the questionnaire was 67, with approximately one-third from each of the 3 ICON8 arms. There was a wide range of reported highest level of educational attainment, with 38 (21%) reporting no qualifications, and 41 (23%) holding a degree or higher qualification. Nearly all participants who returned the questionnaire reported English being their first language. About 61 (40%) respondents reported using the internet or email less frequently than every day, with 26 (15%) never using internet or email.

Table 2. Baseline characteristics of participants who returned the questionnaire.

About 94 participants were invited to take part in the qualitative study, and 13 (14%) were interviewed. Table 3 shows the characteristics of the interviewed participants. The parts of the sampling frame we were unable to recruit participants for were “opted out of Mailed Printed Summary,” “had used the email list,” and “aged 50 or younger.”

Primary outcome: Did the interventions improve satisfaction with how the results were shared?

Quantitative findings on satisfaction with how the results were shared.

Tables 46 shows the patient-reported outcomes relating to the experience of receiving the results, by randomised intervention. For the primary outcome of participant satisfaction with how the results were communicated, among the 3 interventions, only the Mailed Printed Summary led to a significant improvement (adjusted OR 3.15, 95% CI 1.66 to 5.98, p < 0.001). The effect sizes for the Enhanced versus Basic Webpages (adjusted OR 1.47, 95% CI 0.78 to 2.76, p = 0.235) and Email List Invitation (adjusted OR 1.38, 95% CI 0.72 to 2.63, p = 0.327) were much smaller. Furthermore, there was no evidence of interaction between any pair of interventions (interaction between webpage and printed summary p = 0.161, webpage and email p = 0.624, printed summary and email p = 0.995).

Table 4. Reported outcomes relating to experience of receiving the results by randomised intervention: Enhanced versus Basic Webpage.

Table 5. Reported outcomes relating to experience of receiving the results by randomised intervention: Mailed Printed Summary versus no Mailed Printed Summary.

Table 6. Reported outcomes relating to experience of receiving the results by randomised intervention: Email List Invitation versus no Email List Invitation.

There was no evidence of heterogeneity of the effects of the interventions on the primary outcome by age, arm in ICON8, education, or reported frequency of internet or email use (S3 Table and S1 Fig). When the 8 possible combinations of interventions were looked at individually, only those that contained the Mailed Printed Summary significantly improved the odds of participants reporting being satisfied with how the results were communicated (S4 Table).

Qualitative findings on the reasons for satisfaction.

S5 Table contains a description of the categories from the qualitative data. Participants cited many reasons for their reported satisfaction, including characteristics related to the information products (clear and understandable); ease of accessing the results; receiving results in their preferred way; the process by which they received results (the Patient Update Information Sheet being sent out first to give them options); and their reflections on the emotional impact of the results and perceived impact for others. Participants who were unsatisfied with how the results were shared (16% of questionnaire respondents) cited a number of reasons, including the following: not knowing how to find out the results; problems accessing the webpage; finding the results difficult to understand; preferring to have found out the results in a different way (for example, wanting a more personal approach, such as being told the results in person or by telephone); perceived lack of timeliness in receiving the results; and the information not giving enough detail.

Did patients want the results, and did they find them out?

Nearly all participants (164/177 (93%)) reported wanting to know the results, and 145 (88%) of these 164 reported finding out the results. None of the participants who reported not wanting to know the results reported having found them out. These 13 participants were spread across the Show RESPECT interventions.

Tables 46 shows the OR and CIs by intervention, and S6 Table gives details of those who reported finding out the results by randomised intervention and subgroup. Of the 3 interventions, only the Mailed Printed Summary significantly increased the odds of finding out the results among those who wanted to know the results, with 78/83 (94%) reporting finding out the results, compared to 67/81 (83%) of those in the no Mailed Printed Summary arms, an OR of 3.57 (95% CI 1.18 to 10.77, p = 0.024), adjusted for age, education level, internet use, strata, randomisation phase, and clustering. No participants subscribed to the email list. Further information on the uptake of the interventions can be found in S2 Text, and qualitative findings around participants’ desire for the results can be found in S3 Text.

Did the information tell participants everything they wanted to know?

Most participants agreed that the information told them everything they wanted to know (Tables 46). Participants at sites allocated to the Enhanced Webpage were more likely to agree that the information told them everything they wanted to know (adjusted OR 2.15, 95% CI 1.13 to 4.07, p = 0.019) than those allocated to the Basic Webpage. There were no significant differences between the Mailed Printed Summary versus No Mailed Printed Summary, or Email List Invitation versus No Email List Invitation. See S3 Text for qualitative findings relating to this outcome.

Was the information understandable?

Approximately 80% of participants reported that they found the results easy to understand. There was no statistically significant difference in any of the randomised comparisons for this outcome (Tables 46). See S3 Text for qualitative findings relating to this outcome.

Was the information easy to find?

Quantitative results on whether the information was easy to find.

Almost three-quarters of participants reported easily finding the results, with no significant differences between any of the Show RESPECT interventions for this outcome (Tables 46).

Qualitative findings on whether the information was easy to find.

The Mailed Printed Summaries were seen as accessible to everyone, as they were not reliant on people’s computer literacy or access to the internet.

“Like my mum, for instance, in her 80s, she wouldn’t have access to this [webpage], so she would only want … She would only be able to have posted results, really. GMI02

When asked whether there were other ways in which they would have liked to have received the results, 22/91 (24%) questionnaire respondents from hospitals not randomised to the Mailed Printed Summary said they would have liked to receive the results by mail, with mail being seen as convenient and easier to access.

Rarely, questionnaire respondents reported not having been told how to access the results. It is unclear whether or not they received the Patient Update Information Sheet (which site logs record as having been sent). Others (from sites not randomised to the Mailed Printed Summary) reported receiving the Patient Update Information Sheet but missed the information on how to obtain the results that the sheet contained. About 11/180 questionnaire respondents reported difficulties accessing the webpage, either not having access to computers, or finding it hard to get onto the webpage, with some participants eventually gaining access, alone or with the help of family members, and others not succeeding. One woman decided to not try to access the results if it meant going online. Other participants, who had been able to access the results themselves, were concerned that sharing results via webpages/email alone would be inaccessible to other participants, either because of lack of computer literacy or lack of access to the internet.

“We live in quite a small community here in [County] but there’s several people that aren’t computer literate. And I think to presume that everybody has got access to web pages and what have you would be a mistake. And also, even things like the bandwidth or whatever you call it here is dire. Sometimes our connection is awful and I still know people in [County] who can’t get a connection so if they’re going to have to go to Costa Coffee to get connected to find out the results of a trial, that doesn’t feel very comfortable. DLI01

One patient commented that the process of having to type in a URL from the Patient Update Information Sheet to get to the webpage was a barrier to accessing the results, and she would have preferred to be sent them by email without having to visit a webpage to sign up for the email list.

How did patients react to finding out the results?

While 127/145 (88%) of participants reported being glad they had found out results, only 4/138 (3%) reported regretting finding the results. About 23/140 (16%) of participants strongly or slightly agreed that they found the results upsetting, which is higher than the proportion regretting finding out the results, suggesting that while some participants were upset by the results, they did not regret having received them. There was no statistically significant difference between the arms on any of these outcomes (Tables 46). See S3 Text for qualitative findings around participants’ emotional responses to the results.

What did patients think about the communication interventions?

S7 Table summarises the qualitative feedback from questionnaires and interviews on the interventions tested within Show RESPECT.

What were patients’ attitudes to trial participation and the ICON8 results?

With no evidence of difference between the Show RESPECT interventions (Table 7), 131/146 (90%) of respondents reported being willing to take part in future research, and 132/147 (90%) said they were likely to recommend taking part in research to others. See S3 Text for qualitative findings around participants’ attitudes to the research.

Table 7. Reported outcomes relating to take part in or recommend taking part in research.


The Show RESPECT study demonstrated that sharing results with trial participants via Mailed Printed Summaries in addition to webpages increased participant satisfaction with how the results were communicated compared to webpages alone, and also enabled more participants who wanted to know the results to find them out. This satisfaction was due to the clear and understandable nature of the results summaries, ease of access, using their preferred approach, the two-stage process used, and the perceived impact of the trial (despite its “negative” results). Among women taking part in an ovarian cancer treatment trial, nearly all wanted to know the overall trial results. None of the participants who did not want to know the results found them out. It is important to look at these outcomes (satisfaction among those who received results, and proportions of people who wanted or did not want the results who received them) together, as they may not necessarily have pointed in the same direction. The two-stage process, informing participants that the results are available and how to access them, rather than automatically sending results out to all participants, was important to ensure that the wishes of the 7% of participants who did not want to find out the results were respected. This may be especially important in trials where the participant population may be vulnerable, or the results may be emotionally challenging for some participants. The additional features of the Enhanced Webpage did not increase satisfaction with how the results were communicated compared to the Basic Webpage but did lead to a higher proportion of participants reporting that it told them everything they wanted to know. The lack of uptake of the Email List Invitation suggests that for similar trial populations, it is not worth creating email lists at the end of the trial.

The qualitative findings show that participants liked the Mailed Printed Summary as an approach, as it was seen as more accessible for patients with limited access to the internet or computer literacy, and it also facilitated keeping the results for future reference, or showing to friends and family. The results sparked a range of responses, including both positive emotions and disappointment and upset, but there was no evidence to suggest that they were experienced as harmful by these individuals, and nearly all participants were glad to have received the results, even if some had found them upsetting.

Our trial employed a cluster randomised factorial design to assess 3 methods of sharing results with participants, allowing us to be confident that the differences observed were due to the interventions. Extensive PPI was carried out to ensure that the study was asking a question that was important to patients and that the interventions tested were appropriate. The interventions selected were designed to be easily replicable in other studies. The mixed methods approach allowed us to explore the reasons behind the quantitative results, while gaining an overall picture across the study population.

Budget constraints meant that we were unable to send questionnaires to all ICON8 participants at the participating sites. However, we used random selection of participants to avoid selection bias, and the characteristics of respondents in terms of age and ICON8 arm are similar to that of all eligible participants at trial sites. Our response rate of those invited to complete the questionnaire was 65%. This introduces uncertainty into the interpretation of our results, as we do not know how outcomes would vary between responders and nonresponders. However, our response rate is similar to that seen in other studies looking at communication of results to trial participants [8,38]. We cannot discern if there are differences between respondents and nonrespondents in other potentially relevant characteristics (e.g., education level, computer literacy); however, respondents cover the range of these characteristics, and the subgroup analysis showed no evidence of heterogeneity in effect by these subgroups. We do not have data on the ethnicity of participants in either the host ICON8 trial or the embedded Show RESPECT trial, meaning that we cannot assess whether ethnicity influences desire to know trial results, or how these results should be shared. We are also unable to explore the impact of what respondents’ first language was on their experience of receiving results. Future clinical trials of treatment may wish to systematically collect this information up front to assess inclusivity [39] and facilitate embedded trials, like Show RESPECT.

This randomised controlled trial contributes to the, as yet, scant evidence base on how to communicate study results to trial participants, providing high-quality evidence to a field that is dominated by observational data, surveys asking about hypothetical scenarios, and expert opinion. Our study adds comparative data around the effectiveness of different communication approaches in practice. Our results around participants’ desire for results are consistent with findings from previous studies [4]. Our participants’ positive reaction to receiving trial results is also consistent with that reported by previous studies [17,18,20,22], even in the context of potentially upsetting results [23]. Box 1 lists our recommendations on points for trialists to consider around sharing results with trial participants.

Box 1. Recommendations based on this research.

  • Trial teams should consider when planning their study how to share results with participants, taking into account:
    • the characteristics of the study population, including, but not limited to, health literacy, computer literacy, access to the internet, age, and, likely, health status;
    • the need to offer choice to participants, allowing those who want to find out the results easy access but not forcing them on those who do not want to receive them; and
    • how to make it possible for participants to keep the results, so they can refer to them in the future.
  • Trial teams should adequately budget the necessary resources to fulfil their obligation to offer the results to study participants in a way that is appropriate to the study population—the lowest cost approaches (e.g., a basic webpage) may not be optimal for every study population.
  • Patient and public involvement is essential for planning how to share results with participants, identifying the outcomes and study results that are important and relevant to participants, and developing the content of results summaries.
  • Care is needed to ensure that the wording of results summaries is both clear to participants and sensitively written.

Show RESPECT was conducted within the context of an ovarian cancer treatment trial, where the population was women with an average age of 67 years and living in the UK. It is unclear how generalizable these results are to trials with different patient populations (e.g., all male or mixed, younger participants and participants likely more familiar with technology, trials studying non-life-threatening conditions, or where results are available soon after receiving trial treatment). Webpages are a low-cost communication approach and may be useful alongside printed summaries, giving opportunities to provide links to further information and support, and audiovisual content that Printed Summaries cannot provide. However, 4 in 10 of our respondents reported using the internet or email less than daily, with 15% never using them. Data from the UK Office for National Statistics in 2019 show that 10% of the UK population are classed as internet nonusers, having either never used the internet or not used it in the last 3 months [40]. Internet nonusers in the UK are more likely to be women, over the age of 65, have a disability, or be economically inactive (particularly those on long-term sick leave) [40]. Households with lower incomes are also less likely to have an internet connection [40]. While the number of internet nonusers has been declining in recent years, trials should be careful about relying on the internet or email to share results with participants if their trial population overlaps with some of the groups most likely to be internet nonusers. Failure to take this into account could exclude a significant proportion of participants from accessing results. If trials are to meet the ethical obligation to offer results to participants, they need to plan and budget for this in a way that is accessible to the trial population. Trials with similar patient populations should budget for Mailed Printed Summaries or ensure that alternative approaches to webpages are easily available and known to those participants for whom webpages are inaccessible. Problems accessing the internet were not the only reasons participants preferred the printed summaries; even those who use the internet daily were more likely to be satisfied with the printed summary, which made it easy for participants to file along with other trial information and to share with others.

Other trial settings may pose different challenges around sharing results; however, qualitative research from the BRACELET study, which focused on neonatal intensive care trials, found similar responses to the receipt of results among bereaved parents [41].

The ICON8 results that we were communicating (no difference between the trial regimens) may have influenced participants’ reported satisfaction, interacting with the communication interventions. The results of the ICON8 trial did not come out of the qualitative research as a major reason for satisfaction or dissatisfaction with how they were shared, and most participants in the qualitative interviews understood the importance of “negative” trial results. However, our qualitative findings do raise the question of whether modified or tailored approaches would be needed to communicate results to participants in each of the trial’s randomised groups if the trial had found a strong clinical difference. People on the poorer performing arm may need additional support or more personalised approaches to receiving results. All eventualities need to be anticipated if a feedback strategy is built into a trial protocol. Research is needed to explore whether our results are reproducible in trials that do find significant benefit or harm.

Future analysis of data collected within Show RESPECT will focus on the perspective of site staff involved in sharing the results with participants, the process used for this, and resource implications of the communication approaches used. Further research is needed to explore the issue of sharing results with the relatives of trial participants who die during a trial, to see if this is something that relatives want, and if so, how it can be done without causing unnecessary distress.


There is a lot of evidence that trial participants want to be offered the overall results of their trial. A common criticism is that there is not enough guidance as to how this might happen. By testing a number of approaches in a sensitive area, and finding out what is acceptable to participants, Show RESPECT moves the field forward. For the patient population of the ICON8 ovarian cancer trial, adding Mailed Printed Summaries to web-based approaches improved patient satisfaction and was better at ensuring those who wanted to know the results were able to find them out. Box 1 contains recommendations based on this research.

Supporting information

S1 Text. Topic guide for qualitative interviews with patients.


S2 Text. Further information on uptake of interventions.


S1 Fig. Forest plot of satisfaction with how the results were shared, by subgroup.


S3 Appendix. Results email sent to those who signed up to the email list.


S1 Table. Description of the interventions tested in Show RESPECT.


S2 Table. Baseline characteristics of all eligible participants at trial sites.


S3 Table. Subgroup analyses (primary outcome only).


S4 Table. Effect of combinations of interventions on satisfaction with how the results were shared.


S6 Table. Proportion of participants who wanted to know the results reporting finding them out, by randomisation and subgroup.


S7 Table. Qualitative feedback on the interventions within Show RESPECT.


S8 Table. CONSORT 2010 checklist of information to include when reporting a cluster randomised trial.


S9 Table. COREQ (COnsolidated criteria for REporting Qualitative research) Checklist.


S10 Table. List of staff involved at Show RESPECT sites.



We are very thankful to the patients who participated in this study and the people who contributed to our Patient and Public Involvement activities. We are very grateful for the work of the Show RESPECT site teams in carrying out this study. A list of team members from sites can be found in S10 Table. We thank Eva Burnett, who is a patient representative on the Study Steering Group for Show RESPECT and ICON8, for her input on the design and conduct of this study and for her comments on this manuscript, and Amanda Hunn for her contributions to the steering group. We also acknowledge the hard work and diligence of Sierra Santana and Ania Spurdens, who were data managers for Show RESPECT. We also wish to thank the ICON8 trial team for their support of this project, particularly Andrew Clamp, Babasola Popoola, Francesca Schiavone, Jonathan Badrock, and Rick Kaplan. We also thank Julia Bailey for her advice and guidance on the qualitative aspect of this study.


  1. 1. Taylor J. Reporting research findings to participants is an ethical imperative. BMJ. 2019;367:l6324. Epub 2019/11/07. pmid:31685487.
  2. 2. World Medical A. World Medical Association Declaration of Helsinki: ethical principles for medical research involving human subjects. JAMA. 2013;310(20):2191–4. pmid:24141714.
  3. 3. Partridge AH, Wong JS, Knudsen K, Gelman R, Sampson E, Gadd M, et al. Offering participants results of a clinical trial: sharing results of a negative study. Lancet. 2005;365(9463):963–4. Epub 2005/03/16. pmid:15766998.
  4. 4. Shalowitz DI, Miller FG. Communicating the results of clinical research to participants: attitudes, practices, and future directions. PLoS Med. 2008;5(5):e91. Epub 2008/05/16. pmid:18479180; PubMed Central PMCID: PMC2375946.
  5. 5. Partridge AH, Wolff AC, Marcom PK, Kaufman PA, Zhang L, Gelman R, et al. The impact of sharing results of a randomized breast cancer clinical trial with study participants. Breast Cancer Res Treat. 2009;115(1):123–9. Epub 2008/06/11. pmid:18543100.
  6. 6. Williams SL, Ferrigno L, Maraini G, Rosmini F, Sperduto RD. A post-trial survey to assess the impact of dissemination of results and unmasking on participants in a 13-year randomised controlled trial on age-related cataract. Trials. 2011;12:148. Epub 2011/06/16. pmid:21672204; PubMed Central PMCID: PMC3136405.
  7. 7. Buchwald H, Fitch LL, Matts JP, Johnson JW, Hansen BJ, Stuenkel MR, et al. Perception of quality of life before and after disclosure of trial results: a report from the Program on the Surgical Control of the Hyperlipidemias (POSCH). Control Clin Trials. 1993;14(6):500–10. pmid:8119065.
  8. 8. Donaldson S, Khetani N, Maniatis G, Stephens D, Wright JG. Sharing clinical trial results with adolescent idiopathic scoliosis patients. J Pediatr Orthop. 2009;29(5):467–75. pmid:19568019.
  9. 9. Brealey S, Andronis L, Dennis L, Atwell C, Bryan S, Coulton S, et al. Participants’ preference for type of leaflet used to feed back the results of a randomised trial: a survey. Trials. 2010;11:116. Epub 2010/12/03. pmid:21122094; PubMed Central PMCID: PMC3002349.
  10. 10. CISCRP. Perceptions & Insights Study: Report on Study Participant Experiences. Boston, MA: CISCRP; 2013.
  11. 11. Elzinga KE, Khan OF, Tang AR, Fernandez CV, Elzinga CL, Heng DY, et al. Adult patient perspectives on clinical trial result reporting: A survey of cancer patients. Clin Trials. 2016;13(6):574–81. pmid:27559022.
  12. 12. Long CR, Stewart MK, Cunningham TV, Warmack TS, McElfish PA. Health research participants’ preferences for receiving research results. Clin Trials. 2016;13(6):582–91. pmid:27562368.
  13. 13. Partridge AH, Hackett N, Blood E, Gelman R, Joffe S, Bauer-Wu S, et al. Oncology physician and nurse practices and attitudes regarding offering clinical trial results to study participants. J Natl Cancer Inst. 2004;96(8):629–32. Epub 2004/04/22. pmid:15100341.
  14. 14. Getz K, Hallinan Z, Simmons D, Brickman MJ, Jumadilova Z, Pauer L, et al. Meeting the obligation to communicate clinical trial results to study volunteers. Expert Rev Clin Pharmacol. 2012;5(2):149–56. pmid:22390557.
  15. 15. Schroter S, Price A, Malički M, Richards T, Clarke M. Frequency and format of clinical trial results dissemination to patients: a survey of authors of trials indexed in PubMed. BMJ Open. 2019;9(10):e032701. pmid:31636111
  16. 16. Fernandez CV, Gao J, Strahlendorf C, Moghrabi A, Pentz RD, Barfield RC, et al. Providing research results to participants: attitudes and needs of adolescents and parents of children with cancer. J Clin Oncol. 2009;27(6):878–83. Epub 2009/01/24. pmid:19164211; PubMed Central PMCID: PMC2668636.
  17. 17. Dixon-Woods M, Tarrant C, Jackson CJ, Jones DR, Kenyon S. Providing the results of research to participants: a mixed-method study of the benefits and challenges of a consultative approach. Clin Trials. 2011;8(3):330–41. Epub 2011/07/07. pmid:21730081.
  18. 18. Dalal H, Wingham J, Pritchard C, Northey S, Evans P, Taylor RS, et al. Communicating the results of research: how do participants of a cardiac rehabilitation RCT prefer to be informed? Health Expect. 2010;13(3):323–30. Epub 2009/11/13. pmid:19906214.
  19. 19. Cox K, Moghaddam N, Bird L, Elkan R. Feedback of trial results to participants: a survey of clinicians’ and patients’ attitudes and experiences. Eur J Oncol Nurs. 2011;15(2):124–9. Epub 2010/08/03. pmid:20674498; PubMed Central PMCID: PMC3078333.
  20. 20. Dorsey ER, Beck CA, Adams M, Chadwick G, de Blieck EA, McCallum C, et al. Communicating clinical trial results to research participants. Arch Neurol. 2008;65(12):1590–5. Epub 2008/12/10. pmid:19064746.
  21. 21. Darbyshire JL, Holman RR, Price HC. Presenting the results of clinical trials to participants. Clin Med (Lond). 2009;9(5):415–6. Epub 2009/11/05. pmid:19886097.
  22. 22. Dixon-Woods M, Jackson C, Windridge KC, Kenyon S. Receiving a summary of the results of a trial: qualitative study of participants’ views. BMJ. 2006;332(7535):206–10. Epub 2006/01/13. pmid:16401631; PubMed Central PMCID: PMC1352050.
  23. 23. Snowdon C, Garcia J, Elbourne D. Reactions of participants to the results of a randomised controlled trial: exploratory study. BMJ. 1998;317(7150):21–6. Epub 1998/07/04. pmid:9651262; PubMed Central PMCID: PMC28597.
  24. 24. Mancini J, Genre D, Dalenc F, Ferrero JM, Kerbrat P, Martin AL, et al. Participants’ uptake of clinical trial results: a randomised experiment. Br J Cancer. 2010;102(7):1081–4. Epub 2010/03/04. pmid:20197767; PubMed Central PMCID: PMC2853089.
  25. 25. Clamp AR, James EC, McNeish IA, Dean A, Kim JW, O’Donnell DM, et al. Weekly dose-dense chemotherapy in first-line epithelial ovarian, fallopian tube, or primary peritoneal carcinoma treatment (ICON8): primary progression free survival analysis results from a GCIG phase 3 randomised controlled trial. Lancet. 2019;394(10214):2084–95. Epub 2019/12/04. pmid:31791688; PubMed Central PMCID: PMC6902268.
  26. 26. Show RESPECT team. Show RESPECT: Show RESults to Participants Engaged in Clinical Trials: A cluster randomised factorial trial of different modes of communicating results to participants of the ICON8 phase III ovarian cancer trial London. 2018 [cited 2021 Jun 8]. Available from:
  27. 27. Health Research Authority. Information for participants at the end of a study: Guidance for Researchers/Sponsors/ Chief Investigators/Principal Investigators London: Health Research Authority; 2015 [cited 2021 Mar 16]. Available from:
  28. 28. Expert group on clinical trials for the implementation of regulation (EU) No 536/2014 on clinical trials on medicinal products for human use. Summaries of Clinical Trial Results for Laypersons. Brussels, Belgium: European Commission; 2018.
  29. 29. Aldinger C, Bierer B, Collyar D, Li R, Myers L. MRCT Return of Results Guidance Document. The Multi-regional Clinical Trials Center of Brigham and Women’s Hospital and Harvard; 2016 Jul 13. Report No.
  30. 30. Hemming K, Eldridge S, Forbes G, Weijer C, Taljaard M. How to design efficient cluster randomised trials. BMJ. 2017;358:j3064. Epub 2017/07/16. pmid:28710062; PubMed Central PMCID: PMC5508848.
  31. 31. Health Research Authority. Applying a proportionate approach to the process of seeking consent: Health Research Authority; 2017 [cited 2021 Jul 29]. v1.01 17.01.2017. Available from:
  32. 32. Guest G, Namey E, Mitchell M. Chapter 1: Qualitative Research: Defining and Designing. 2013 2021/03/24. Collecting Qualitative Data: A Field Manual for Applied Research [Internet]. 55 City Road, London: SAGE Publications, Ltd. Available from:
  33. 33. Malterud K, Siersma VD, Guassora AD. Sample Size in Qualitative Interview Studies: Guided by Information Power. Qual Health Res. 2015;26(13):1753–60. pmid:26613970
  34. 34. Braun V, Clarke V. Using thematic analysis in psychology. Qual Res Psychol. 2006;3(2):77–101. pmid:223135521.
  35. 35. Robson A, Robinson L. Building on models of information behaviour: linking information seeking and communication. J Doc. 2013;69(2):169–93. WOS:000317150200002.
  36. 36. Robson A, Robinson L. The Information Seeking and Communication Model. J Doc. 2015;71(5):1043–69. pmid:109264960.
  37. 37. Moran-Ellis J, Alexander VD, Cronin A, Dickinson M, Fielding J, Sleney J, et al. Triangulation and integration: processes, claims and implications. Qual Res. 2006;6(1):45–59.
  38. 38. Mancini J, Genre D, Dalenc F, Ferrero JM, Kerbrat P, Martin AL, et al. Participants uptake of clinical trial results: A randomised experiment. Br J Cancer. 2010;102(7):1081–4. pmid:20197767
  39. 39. Witham MD, Anderson E, Carroll C, Dark PM, Down K, Hall AS, et al. Developing a roadmap to improve trial delivery for under-served groups: results from a UK multi-stakeholder process. Trials. 2020;21(1):694. pmid:32738919
  40. 40. Office for National Statistics. Exploring the UK’s digital divide. Office for National Statistics; 2019 Mar 4. Report No.
  41. 41. Snowdon C, Brocklehurst P, Tasker R, Ward Platt M, Harvey S, Elbourne D. Death, bereavement and randomised controlled trials (BRACELET): a methodological study of policy and practice in neonatal and paediatric intensive care trials. Health Technol Assess. 2014;18(42):1–410. Epub 2014/07/06. pmid:24987820; PubMed Central PMCID: PMC4781192.