Skip to main content
Advertisement
Browse Subject Areas
?

Click through the PLOS taxonomy to find articles in your field.

For more information about PLOS Subject Areas, click here.

  • Loading metrics

Exploring the application of machine learning to expert evaluation of research impact

  • Kate Williams ,

    Roles Conceptualization, Funding acquisition, Investigation, Methodology, Project administration, Resources, Supervision, Writing – original draft, Writing – review & editing

    kate.williams@unimelb.edu.au

    Affiliation School of Social and Political Sciences, University of Melbourne, Melbourne, Victoria, Australia

  • Sandra Michalska,

    Roles Conceptualization, Formal analysis, Methodology, Validation, Visualization, Writing – original draft, Writing – review & editing

    Affiliation Policy Institute, King’s College London, London, Greater London, United Kingdom

  • Eliel Cohen,

    Roles Conceptualization, Writing – original draft, Writing – review & editing

    Affiliation Policy Institute, King’s College London, London, Greater London, United Kingdom

  • Martin Szomszor,

    Roles Conceptualization, Data curation, Formal analysis, Methodology, Software, Supervision, Validation, Visualization

    Affiliation Electric Data Solutions, London, Greater London, United Kingdom

  • Jonathan Grant

    Roles Conceptualization, Funding acquisition, Writing – review & editing

    Affiliation Different Angles, Cambridge, Cambridgeshire, United Kingdom

Abstract

The objective of this study is to investigate the application of machine learning techniques to the large-scale human expert evaluation of the impact of academic research. Using publicly available impact case study data from the UK’s Research Excellence Framework (2014), we trained five machine learning models on a range of qualitative and quantitative features, including institution, discipline, narrative style (explicit and implicit), and bibliometric and policy indicators. Our work makes two key contributions. Based on the accuracy metric in predicting high- and low-scoring impact case studies, it shows that machine learning models are able to process information to make decisions that resemble those of expert evaluators. It also provides insights into the characteristics of impact case studies that would be favoured if a machine learning approach was applied for their automated assessment. The results of the experiments showed strong influence of institutional context, selected metrics of narrative style, as well as the uptake of research by policy and academic audiences. Overall, the study demonstrates promise for a shift from descriptive to predictive analysis, but suggests caution around the use of machine learning for the assessment of impact case studies.

Introduction

While the term artificial intelligence (AI) often conjures up notions of entities with a general intelligence that resembles (and surpasses) human intelligence, much more common is the more modest and specific development and application of computational systems able to process information and make decisions that have traditionally required human intelligence. In the past decade, machine learning (ML)—one aspect of AI—has greatly advanced the capability of such systems, notably in visual detection of images [1, 2], language processing [3, 4], and speech and audio recognition [5, 6]. While major technical advances in capability are being driven mainly by big technology companies, some domains have begun exploring the practical implementation of these new technologies outside of technology-oriented environments. In particular, law [7], health [8], and cybersecurity [1] have begun to build capability towards using AI to improve accuracy and efficiency.

There is an increasing academic interest in more streamlined, less time-consuming alternatives to current peer review processes, which are used to determine the results of journal submissions, funding applications, and the outcomes of research quality assessments [911]. Prior research on peer review processes has identified often lengthy delays [12] as well as pervasive issues of bias [1315] and superficiality [16]. Similarly, research has highlighted differences in the criteria used between disciplines [17] and between reviewers and journal editors [18]. Thus, a key question is whether emerging technologies can contribute to a leaner and fairer form of evaluation. Yet, the application of ML in research assessment is still rare, despite recent calls for improving current evaluation policy and practice [19]. At least in theory, the (semi)automated nature of ML can add to the overall objectivity, transparency and reproducibility of the process, potentially making it an attractive complement to peer review. However, there is limited published work that explores what this might look like in practice, particularly for large scale evaluation programs.

Expert peer review in the UK’s Research Excellence Framework

The United Kingdom’s Research Excellence Framework (REF) is a national program that assesses the quality of research in UK higher education institutions. Its results determine the distribution of approximately £2 billion annual funding [20], as well as indirect outcomes that arise from improved reputation and increased student numbers. The UK has conducted research evaluations approximately every 5–7 years since 1986. Of these, REF2014 represented a substantial shift in the priorities and practices of the evaluation through the introduction of the explicit assessment of the wider societal and economic impact of academic research, which was absent from the previous iterations. The REF2014 submissions were evaluated along three main components—outputs, impact and environment—which accounted for 65%, 20% and 15% of the total score, respectively. Impact was assessed by ‘impact case studies’, which are structured narrative documents of four pages, designed to demonstrate the impact of a body of work conducted by academics within the submitting department. Impact was defined as “an effect on, change in or benefit to the economy, society, culture, public policy or services, health, the environment or quality of life, beyond academia” [21]. For REF2014, 154 institutions submitted 6,975 impact case studies, which were assessed by 36 sub-panels guided by four Main Panels [22].

Exceeding the previous rounds, REF2014 was estimated to cost almost £250 million [23], with the impact assessment component alone estimated at £55M [24]. Although the assessment was relatively cheap when compared to the amount spent on grant funding in the UK (estimated at 3% versus 13%) [24]), it nevertheless represents a sizeable financial burden for the sector. Despite the increasing cost of such a comprehensive and labour-intensive national evaluation, the latest round of assessment, REF2021, was conducted along largely similar lines of 2014 [10, 25]. As a specific mode of assessment, the UK’s system has been subject to criticism from some quarters while receiving support elsewhere. This occurs within the context of a broader body of literature on the challenges and flaws of the peer review process. Yet, despite the ongoing debates about the value of the exercise, one outcome of the REF has been the generation of a great deal of freely available information on the wider social and economic effects of the country’s publicly funded research [23].

Emerging machine learning techniques for research evaluation

Researchers have begun to explore the potential of ML techniques in the area of research evaluation. One study by Yuan et al. [26] built a natural language processing (NLP) system to automate the peer review process, which they compared with expert evaluation. The results show that the NLP models were not as good as experts at overall summarising, but better at producing more comprehensive reviews. Researchers have also experimented with advanced multi-layered deep learning (DL) architectures for the prediction of ‘impactful’ research in medical science. DL is an engineering science based on multiple-layered artificial neurons propelled by large amounts of data and advances in computing power [27]. For instance, to explore impact, Nelson et al. [28] fed their DL model with publications’ metadata, title and abstract-level features to determine inclusion in patents, guidelines or policy documents, while Weis et al. [29] used paper, author, affiliation, journal and citation-based features in their DL model to predict its inclusion in the top 5% of the node centrality network after 5 years of publication. Both works favoured ML/DL approaches with many dimensions, which allow for sufficient expressivity and flexibility to capture complexity of established bibliographic and innovative semantic predictors of impact. One practical benefit of this approach is that DL is empirically driven, rather than guided by theory and prior assumptions. However, it is important to note that these types of technologies should not replace human judgment, but rather provide tools for producing evidence to support these judgements. If uncritically used, ML could have severe implications for research evaluation (for example, favouring the features of past successful cases at the expense of novel and unconventional cases).

Given the wealth of data arising from the evaluation, the existing literature on the REF2014 impact component provides some evidence of the potential for various ML tools to provide meaningful insights on large scale evaluations. An overview of the literature that analyses REF2014 impact submissions is summarised in Table 1. Much of this work primarily focuses on the qualitative data that arises from the universities’ narrative impact case studies. The approaches include topic modelling [25, 30], social network analysis [31, 32], and comparative linguistic analysis of high-scoring versus low-scoring submissions [33, 34]. These provide a wealth of information on the narrative features, such as the work of Reichard et al. [33], which used text mining to evidence statistically significant variation in writing style between the high-scoring and low-scoring impact case study submissions. In general, these analyses tend to focus mainly on particular disciplinary groupings or domains of interest, such as Business & Management [35] or Library & Information Studies [36], which limit their generalisability and potential for cross-domain insights. The review of the literature also shows that the application of more advanced ML techniques in research assessment is limited. That is, there is little work that uses these techniques in a predictive rather than descriptive manner. One notable exception is the work of Balbuena [9], who considered the quantitative metadata of REF2014 submissions. The study sought to replicate overall REF2014 Grade Point Average (GPA) results from data from Times Higher Education Report, the UK Higher Education Statistics Agency (HESA), the Web of Science, and the Guardian league tables using Bayesian Additive Regression Trees (BART) model [37]. A number of strong GPA predictors were identified, including how selective the university’s admissions standards were, the percentage of faculty with PhDs, and the average h-index of the faculty. The study concluded with a recommendation to complement peer review with ML in UK’s research assessment practice. Thus, so far, the use of advanced ML techniques has been limited to quantitative features. This highlights a key opportunity to combine the qualitative and quantitative aspects of the entire dataset using advanced ML techniques.

thumbnail
Table 1. Existing literature analysing REF2014 submissions.

https://doi.org/10.1371/journal.pone.0288469.t001

By integrating qualitative and quantitative information from all REF2014 impact case studies, this article seeks to build upon this literature by exploring how ML techniques classify case studies that have been evaluated as high or low-scoring by expert panels. It provides an opportunity to consider the predictive potential (rather than descriptive potential) of such methods in relation to research assessment. The study addresses the following research questions:

  1. RQ1 - Can we predict high-scoring impact case studies using ML?
  2. RQ2 - What are the characteristics of high-scoring impact case studies?

Methodology

The primary task was the binary classification of each individual impact case study as either high-scoring or low-scoring. The secondary task was the extraction of the specific features that were predictive of either high-scoring or low-scoring case studies. The following subsections specify the data sources used, the feature extraction process and the model development and interpretation stages.

Data sources

REF submission data.

6,637 REF2014 submissions (including impact scores) were downloaded from the official REF2014 website (https://ref.ac.uk/2014/). The impact case studies were structured in five sections: A) summary of the impact, B) underpinning research, C) references to the research, D) details of the impact, and E) sources to corroborate the impact. The case studies were scored by expert panels on a scale ranging from 4* (world-leading) down to 1* (recognised but modest), plus an extra ‘unclassified’ category. The published REF2014 impact scores were aggregated and scored at the submission level for each Institution and Unit of Assessment (UoA). The 154 submitting institutions made submissions in 36 UoAs that correspond to disciplinary groupings, so that submissions were primarily at a departmental level. Classifying the case studies as either ‘high’ or ‘low’ scoring for the purposes of this study was not immediately possible, given the scores of individual case studies were not published. To address this, following Balbuena [9], we distinguished between the highest and lowest-scoring submissions and assigned a GPA score to all studies in a given submission using the Eq 1, where 4*, 3*, etc. is the percentage of the relevant scores. There were 2 special cases in GPA estimation: 1) one case study associated with multiple institutions, and 2) one associated with multiple UoAs within the institution—in both cases we took the mean GPA for that case study. (1)

To allow for further subsetting, we also collected other information that was used by REF2014 (https://impact.ref.ac.uk/casestudies/FAQ.aspx) to provide context to the submissions. Specifically, we recorded the overarching Main Panels that the UoAs fell under (S1 Table) and the institutions’ income categories from HESA (S2 Table), which had assigned universities to economic Peer Groups on the basis of income data in 2004–05.

Once a GPA was calculated for each case study, we ranked them in percentiles according to Main Panel and labelled 20% of the top case studies as high-scoring (1284) and 20% of the bottom case studies as low-scoring (1319). Normalisation by Main Panel (instead of top/bottom 20% on a global dataset) allowed a relatively equal distribution across departments for the extraction of more generic features. The distribution of case studies based on the estimated GPA scores across the panels (and the 20% top/bottom cut-off) is shown in S1 Fig. Based on the co-authors’ expertise in research evaluation, we opted for a 20% cut-off. This allowed investigation of the characteristics of very high and very low scoring case studies (whereas widening this threshold would make those characteristics less prominent).

Bibliometric data.

Using the OpenAlex [43] API (api.openalex.org) we extracted the bibliometric data based on the DOIs (Digital Object Identifier) referenced in the case studies. To mirror the survey period used in REF2014, only citations from papers published in 2014 or earlier were considered. 23,967 DOIs were identified, and 21,263 were found in OpenAlex.

Policy citation data.

The Overton database (https://www.overton.io/) was used for the extraction of policy citation data. Also in keeping with the REF survey period, a filter was applied to omit policy citations received after 2014. 3,959 of the 23,967 DOIs were cited more than once in Overton. A total of 16,992 citations came from the policy documents.

Feature extraction

Using the whole dataset of case studies, we extracted six overarching groups of features: discipline, institution, explicit text, implicit text, bibliometric indicators and policy indicators. Each feature is set out in the context of its group, metric, and brief description in Table 2.

thumbnail
Table 2. Features extracted from REF2014 impact narratives.

https://doi.org/10.1371/journal.pone.0288469.t002

The discipline feature group is concerned with the UoA of the submitting institution. This feature was constructed as a vector length of 36 (total number of units) where 1 represents the UoA responsible for submission while 0 represents the remaining 35 UoAs. The second group—institution—relates to the UK Provider Reference Number (UKPRN). This feature was constructed as a vector length of 198, where 1 represents the submitting institution.

For the impact case study text, we distinguished between implicit and explicit. The narrative style feature represents the explicit text included by the submitting institutions to demonstrate impact. We applied the TF-IDF (Term Frequency-Inverse Document Frequency) weighting scheme, which is often used in text summarisation and classification. TF-IDF allows extraction of the relevant and informative terms in the prediction task while excluding commonly occurring stop-words. It does this by assigning a higher weight to meaningful words and a lower weight assigned to common words. We also followed standard pre-processing steps, removing punctuation, diacritics, English stop words, terms that appear in either <3 or >50% of the case studies, URLs (keeping just domain), and converting to lower-case. We optimised for the ‘n-gram range’ (i.e. the number of the consecutive words used in feature set) and vocabulary size, which we detailed in S2 Fig. Here, we used all sections of the case studies (i.e. A-E) as the TF-IDF scheme naturally filters out words without meaningful contribution to the classification.

The readability and sentiment features represent the implicit aspects of case study text. The readability feature was given by five metrics that assign numerical scores to generate a ranking of books or other texts in order of ease of comprehension [44, 45]. These include the Flesch Reading Ease (FRE) metric ([46], the Automated Readability Index (ARI) [47], and Dale Chall Readability Score [48], which are now predominantly used in marketing, research and policy communication to measure the degree to which a piece of text is understood and engaged with [49]. For the sentiment feature, we calculated the percentiles of polarity scores in each sentence. Readability and sentiment were limited to sections A,B and D (as C and E include mainly references). This was to investigate whether there is an increased tendency to highlight the positive implications or results of research (e.g. case studies focusing on solutions as reflected in more positive sentiment), or rather highlight current gaps and limitations (e.g. case studies focusing on the problem as reflected in more negative sentiment).

The bibliometric indicators group contains characteristics derived from the publications that were included by submitting institutions to evidence their impact. This included three features: publication, which gives bibliometric counts; author, which classifies the location of authorship; and author affiliation, which provides categories of institution type from the Global Research Identifier Database (GRID) (https://www.grid.ac/). For this study, Open Access status included the labels ‘gold’, ‘bronze’, ‘green’, and ‘hybrid’. This set of features extends the work of Weis and Jacobson [29] by adding author affiliation metrics, designed to offer greater insight into inter-sector collaborations.

The final feature group is policy indicators. This represents the use of the submitted publications by policy contexts. Raw counts of policy citations are provided, as well as a category normalised metric to account for disciplinary differences.

Model evaluation

The study used binary text classification (high-scoring/low-scoring) belonging to the supervised subcategory of ML and more broadly AI. Text classification is a “supervised learning task that is defined as the identification of categories of new documents based on the probability suggested by a specified training corpus of already labelled (identified) documents” [52]. The ML models used for such tasks range from well-established Support Vector Machines (SVM) and Random Forest (RF) to more recent Neural Networks (NN)-based classifiers. The conventional classifiers of SVM and RF were developed in the reverse order to the development of NNs. SVMs evolved from theory to implementation and experiments, while RF was based on idea that the performance of a set of many weak classifiers is usually better than a single classifier given the same quantity of trained information [53]. RF constructs many decision trees that are used to classify a new instance by the majority vote, and the suggested number of trees in a forest range between 64 and 128 [54] (we selected 100). NNs, on the other hand, followed a more heuristic path, from applications and extensive experimentation to theory [55].

Given the empirical nature of ML (and NN in particular), the comparison between the traditional approaches can be evaluated against NN as well as between the various architectures of NN themselves. For training the models, we used a proxy of a) the top 20% panel-normalised case studies to assign the high-scoring label and b) the bottom 20% panel-normalised to assign the low-scoring label. Training, in the context of ML, describes an automatic search process for better representations of the pre-specified classes/labels. We used two conventional classifiers (SVM and RF) and three NN-based architectures of increasing complexity. SVM and RF are the most straightforward, followed by net1, which has zero hidden layers, then net2, which has one hidden layer and 128 nodes, and finally, net3, which is an undercomplete autoencoder with four hidden layers and 250, 5, 5, 125 nodes [56]. For the SVM, we used the Linear Support Vector Classification. For all NN-based architecture, a ‘sigmoid’ activation function was used on the output. All hidden layers were densely connected and used the ‘relu’ activation function, except for net3 which uses ‘selu’, and were followed by a drop-out layer with a rate set to 0.3.

We conducted a comprehensive evaluation to: 1) compare the classification accuracy between models, 2) measure the relative contribution of different feature groups, 3) identify differences between the main panels and income categories. In total, we ran 5 × 23 × 12 experiments. This involved the five models (SVM, RF, net1, net2, net3), 23 feature group combinations (e.g. 1-explicit text & implicit text, 2-explicit text & implicit text & bibliometric indicators, etc.) and 12 (sub)sets (global dataset, four Main Panels, seven Income Categories). The application of multiple methods assures that variances are related to the features rather than the method [57].

The design used a stratified shuffle split, which involved randomly shuffling and dividing the dataset into ten parts, with nine parts used for training and one reserved for testing. This process was duplicated ten times, with each iteration reserving a different set for testing. This allowed us to measure the accuracy across the ten splits. The average accuracy was calculated for each model-feature-set scenario as explained above. The decision on which class each case study belongs to was based on the prediction score (PS), where PS = {0, 1} and a cut-off threshold was set at PS > 0.5 for class assignment. The evaluation was then based on the proportion of the correct predictions using the standard accuracy metric, where T-True, F-False, P-Positive, N-Negative (i.e. TP indicates case studies correctly identified as high-scoring, and TN indicates those correctly identified as low-scoring and so on). The standard accuracy metric uses the equation (Eq 2): (2)

Model interpretability

Prior work has confirmed that the performance of highly non-linear NN models—or ‘black box’ approaches—surpass the performance of the conventional classifiers such as SVM and RF (known as the performance versus explainability trade-off [58]). However, apart from simply providing a satisfactory classification system, the interpretability of the models’ internal workings is also a crucial element of our study. The overall idea behind NN is to determine which are the most useful representations of the input data in reflecting the expected output, in our case, the high-scoring/low-scoring case studies. This input-output transformation takes place in the hidden layers in the process of training on the labelled input data. The learning essentially occurs via finding the weights for all the nodes in the hidden layers that minimise the loss function (objective function) between the predictions and the targets. In the case of NN, the exact features that contribute towards the predictions are regarded as more opaque due to the highly non-linear and empirical character of the approach.

Although net1 and net2 have relatively basic architectures with zero and one hidden layers respectively, the main idea behind net3—the undercomplete autoencoder—is to learn a highly compressed input representation via the so called ‘bottleneck’, which is a hidden layer with a limited number of neurons. That is, the complete architecture in our case consists of four hidden layers with 250 neurons (1st), five neurons (2nd and 3rd—the compressed representations), and 125 neurons (4th). Autoencoders aim to desensitise the irrelevant features during encoding and compress the most meaningful ones during decoding while reconstructing an initial input (in our case, the impact case studies and the assigned label). Here, although it is difficult to elucidate precisely how a model works, autoencoders can nonetheless provide useful insights for practitioners and end users of ML [59]. The ML/DL approach can be used efficiently as a first pass assessment to identify areas for further exploration. This is a particularly useful approach when dealing with high-volume and high-variety contexts for which traditional methods can prove limited [60].

For the purposes of interpretability, the advantage of the traditional SVM and RF classifiers lies in their ease of extracting the predictive coefficients from the model. These techniques are highly popular and long-standing supervised ML algorithms used for both classification and regression problems, and detailed information can be found in [54, 61]. Thus, we followed standard procedure in extracting the SVM and RF coefficients to better understand which particular features contributed towards classification as high- or low-scoring (i.e. a direction indicative of the predicted class and a magnitude indicative of the relative importance in terms of SVM, and a magnitude indicative of the relative importance in terms of RF).

Results & discussion

Classification accuracy

This section considers how accurately the ML models classify high-scoring versus low-scoring REF impact submissions (RQ1). Here, we explore the models’ ability to predict GPA score from REF impact case studies given our set of six feature groups.

The experiments demonstrated the effects of the feature groups on the classification outcome. We compared the classification of our ML models on various combinations of groups to measure the predictive contribution of each. Across all models, there was a strong influence of institutional and disciplinary information (Table 3). Accordingly, we split the analysis into inclusive and exclusive of institution and discipline. By excluding these, the effect of the remaining feature groups became more prominent (Table 4). On average across the models, the classification ranged from 58.5% (just discipline) up to 90.4% (institution, discipline and policy indicators). After exclusion of institutional information, the maximum classification was 77.8% (the average across the models) for the explicit, implicit and bibliometric groups. Model training included explicit features, which were then filtered out when ranking important features. This finding indicates high fidelity of prediction from the qualitative submission narrative (both the words used directly and the latent metrics derived secondarily from the text) as well as quantitative information on publication, author and author affiliation.

thumbnail
Table 3. Accuracy across feature groups (incl. discipline & institution).

https://doi.org/10.1371/journal.pone.0288469.t003

thumbnail
Table 4. Accuracy across feature groups (excl. discipline & institution).

https://doi.org/10.1371/journal.pone.0288469.t004

To explore these findings further, we trained separate models on subsets of the data representing each of the REF’s four Main Panels and the income category Peer Group of the submitting institutions. For each, we used the same combinations of feature categories as the global dataset above and also split the groups into inclusive and exclusive of discipline and institution information.

In terms of the Main Panels, the comparative accuracy of the classification for each scenario is shown in Fig 1. Overall, predictions were stronger among Panels A (medicine, health and life sciences), B (physical sciences, engineering and mathematics) and C (social sciences) in comparison with Panel D (arts and humanities), particularly when institution and discipline were included. In Panels B, C and D, the implicit text features (readability, sentiment) were most predictive in comparison with Panel A. There was also a stronger impact of bibliometric and policy indicator groups in Panel A than in Panel D. The addition of more than three feature groups did not improve the classification accuracy on average.

thumbnail
Fig 1. Classification accuracy by Main Panel: I. With institution and discipline, II. Without institution and discipline (column 1—svm, column 2—rf, column 3—net1, column 4—net2, column 5—net3).

https://doi.org/10.1371/journal.pone.0288469.g001

For the income category Peer Groups, the comparative accuracy is shown in Fig 2. Greater predictive capacity was observed among the lower income groups D (research income of between 5% and 8% of total income), E (teaching institutions) and F (smaller teaching institutions) on average, compared to the higher income groups A (Russell Group, which are traditionally considered the most research-intensive and prestigious universities in the UK), B (research income of 22% or more of total income) and C (research income of between 8% and 21% of total income). As above, this was particularly true when institution and discipline were included. The models can more easily predict the high- or low-scoring case studies from the top Peer Group (A) or lower Peer Groups (D, E, F), whereas the middle income categories (B, C) are less clear. The exception was Peer Group G (specialised music/arts teaching institutions), due to a much smaller sample size (37). In addition, more than three feature groups did not improve accuracy. Future work would need a larger sample size to validate this, especially for Peer Groups C, E and F.

thumbnail
Fig 2. Classification accuracy by income category Peer Group: I. With institution and discipline, II. Without institution and discipline (column 1—svm, column 2—rf, column 3—net1, column 4—net2, column 5—net3).

https://doi.org/10.1371/journal.pone.0288469.g002

Previous analyses of REF impact results have identified that the assessment process seems to privilege certain disciplines and, typically, more prestigious, better-resourced universities [62, 63]. The novel contribution that our analysis makes to this evidence is an estimate (or more specifically, a number of estimates) of the extent to which discipline and institution predict the outcomes of REF impact assessments. As an illustration, when extracting the most important features for the classification of high scoring case studies (including institutional information), between 60% (svm model) and 77% (rf model) of the top 30 institutions belonged to the Russell Group.

Predictive features

This section considers the insights that can be gained from ML models in terms of the characteristics of high-scoring REF submissions (RQ2).

We narrowed this section of the analysis down to the following feature groups: implicit text (readability, sentiment), bibliometric indicators (publication, author, author affiliation) and policy indicators (policy citations). Here, the explicit text feature group was excluded because of the amount and format of the data (i.e. around 7,787 n-grams, which would obscure the other findings). Given the prevalence of institutions’ names and locations which would reveal specific institutions, we did not include the top K words by feature weights in the analysis. To affirm the findings derived from standard procedures around the use of SVM coefficients, the feature weights were taken for both SVM and RF and each of the 10 folds (final average was taken). The weights were then ranked and the average of those ranks for both models (final rank) is shown in Table 5.

thumbnail
Table 5. Features important for high scoring case studies classification (svm and rf models coefficients).

https://doi.org/10.1371/journal.pone.0288469.t005

Interestingly, particularly prominent across high-scoring case studies were the readability features. The average of FRE, SI, ARI, DCRS metrics for high-scoring case studies indicated a more straightforward writing style. The strong finding around writing style builds on the work of Rechaird et al. [33], which emphasised the influence of latent characteristics derived from the submitted impact narratives. Thus, as well as the significance, reach and attribution of impact, as given by the REF’s published criteria, the presentation of the narratives seems also to have influenced impact assessment.

Overall, policy indicators were strongly predictive of high-scoring case studies. The study provided more granular insight into the relevance of specific policy-related features. In order of relevance, these were: 1) the maximum number of Overton citations for referenced DOIs (e.g. 1 highly-cited paper in policy documents), 2) the number of DOIs that have at least one citation in policy documents, and 3) the maximum citation percentile using category normalisation as in the work of Szomszor [51]. This reinforces the finding that there is a link between expert evaluation of impact and policy citation rank [51], and highlights the importance of policy influence to wider notions of research impact. Still, Smith et al. [64] (and based on wider discussion in Boswell et al. [65]) caution against rewarding individual researchers for ‘achieving’ research impact based on narrow indicators, such as citations in policy documents.

Publication-related features extracted from the OpenAlex database were also highly predictive of the high-scoring submissions. In order of importance, these were: 1) the mean rank of citations and 2) the maximum rank of citations. In terms of the affiliated sectors (the count of institutional affiliations with the selected sector according to GRID type), the strongest association with high-scoring case studies was found for company affiliation, followed by the healthcare and government affiliations. On the other hand, those related to archive, non-profit, education and facility affiliations were more predictive in low-scoring cases, possibly indicating that assessors hold assumptions about the kinds of collaborators that are most valuable for impact. As for the countries of the authors, ‘the number of DOIs with at least one non-UK author’ feature was more predictive of low-scoring examples (although the strength of the association was relatively small). Sentiment or open access (OA)-level features (count of DOIs with OA status) were not strongly predictive of either high or low-scoring classifications.

Conclusions

We ran a series of experiments to investigate the ability of ML techniques to predict and elucidate the results of a large-scale expert evaluation system. By training five models on a range of qualitative and quantitative feature groups of REF2014 impact submissions, we showed that ML techniques are able to predict high-scoring and low-scoring case studies. We then used trained model weights to identify the key characteristics of high-scoring submissions.

The approach we used is novel because of the inclusion of both quantitative and qualitative feature groups (including both explicit and implicit qualitative features). Previous literature has focused on the selection and evaluation of either qualitative or quantitative features, such as the narrative components of the assessment [33] or the external metadata [9]. Our approach thus sought to build on this literature by further drawing out the complex holistic nature of the expert evaluation. It represents a step towards a predictive approach to the analysis of REF impact case studies, in contrast with the currently dominant descriptive approaches (in line with big data analytics [66]).

The results provide insight into the characteristics of impact submissions that mattered in REF2014. As in the global dataset, discipline and institution were shown to be strongly predictive of high scores when the models were trained on the Main Panel and Peer Group subsets. This supports the notion that the rules and traditions within research fields shape what counts as meaningful impact, and that some disciplines may find it easier to evidence more recognisable forms of impact. It also suggests that the evaluation captures something about the universities’ material resources or prestige. Thus, it may be that REF2014 was not entirely able to avoid the social conditions that surround specific institutions. That is, there may have been implicit pre-conceptions that were captured in the evaluation.

We chose to focus this analysis on the REF2014 evaluation in order to compare results to existing work on the same publicly available data. We sought to explore the potential of this line of inquiry and develop a proof-of-concept approach. Particularly as the number of studies on REF2021 grow, future research could fruitfully apply this framework to the most recent REF data. Key limitations of this study include that the assumed ‘gold standard’ of high/low-scoring labels were only estimates, given the that the exact scores at the individual case study-level were not made publicly available. In addition, the findings directly reflect the REF2014 data, and for wider generalisation (in terms of both accuracy and predictive features) more data is needed. For example, this may take the form of data from several REF iterations or integrating similar case studies from outside of the UK. This is especially true for NN architectures that learn from the examples rather than rules set a priori.

This study illuminates the key question of whether ML techniques can contribute to a leaner and fairer form of evaluation. Our analysis suggests that while these techniques may offer the potential for leaner evaluation, a high degree of caution is required around expecting ‘fairness’. This is because of the influence of past characteristics of high-scoring case studies at the expense of emerging and infrequent ones. Using the wealth of data provided by REF2014, we have demonstrated how an ML approach to evaluation might have looked in practice and what characteristics may have been emphasised and with what accuracy.

Overall, our work makes two key contributions. First, it shows that ML models are able to accurately classify case studies that have been evaluated as high or low-scoring by expert panels. It thus demonstrates that ML architectures are able to process information to make decisions that resemble the conclusions of expert evaluators. In doing so, it highlights the possible role of emerging technologies that can learn, and help us learn, about the wealth of valuable data generated by assessments such as the REF. Second, it demonstrates the value of ML in illuminating the features of case studies that seem to influence peer reviewers’ assessments. The features extraction elicited the types of characteristics that may be favoured if an ML approach was applied for their automated assessment. Most significantly, this included the institution and disciplinary context in which the researchers being assessed are embedded, but also strongly highlighted the way that the narratives are written, the uptake by policy and academic audiences, and the sector affiliation of collaborators.

Supporting information

S1 Table. REF2014 Units of Assessment (UoA) and overarching Main Panels.

The Main Panels can be broadly grouped as follows: A is medicine, health and life sciences, B is physical sciences, engineering and mathematics, C is social sciences and D is arts and humanities.

https://doi.org/10.1371/journal.pone.0288469.s001

(TEX)

S2 Table. REF2014 UKPRN and institutional Peer Groups based on HESA income categories for 2020–21.

References to income are to 2012–13 HESA data, where Peer group A: Institutions with a medical school and research income* of 20% or more of total income; Peer group B: All other institutions with research income* of 15% or more of total income; Peer group C: Institutions with a research income* of between 5% and 15% of total income; Peer group D: Institutions with a research income* less than 5% of total income and total income greater than £150M; Peer group E: Institutions with a research income* less than 5% of total income and total income less than or equal to £150M; Peer group F: Specialist music/arts teaching institutions (*research income is defined as the funding council recurrent research grant plus the total research grants and contracts returned in the 2012–13 HESA Finance Statistics Return (FSR)).

https://doi.org/10.1371/journal.pone.0288469.s002

(TEX)

S1 Fig. Distribution of calculated GPA scores of impact case studies by Main Panels.

https://doi.org/10.1371/journal.pone.0288469.s003

(TIFF)

S2 Fig. Optimal vocabulary size and ngram range that returns maximum accuracy (using tf-idf scheme).

https://doi.org/10.1371/journal.pone.0288469.s004

(TIFF)

References

  1. 1. Cui Z and Xue F and Cai X and Cao Y and Wang G and Chen J. Detection of malicious code variants based on deep learning. IEEE Transactions on Industrial Informatics. 2018; 14(7): 3187–3196.
  2. 2. Zielinski B and Sroka-Oleksiak A and Rymarczyk D and Piekarczyk A and Brzychczy-Wloch M. Deep learning approach to describe and classify fungi microscopic images. PLoS one. 2020. 15(6). pmid:32603329
  3. 3. Arras L and Horn F and Montavon G and Muller KR and Samek W. What is relevant in a text document?”: An interpretable machine learning approach. PLoS one. 2017. 12(8).
  4. 4. Venkataraman GR and Pineda AL and Bear Don’t Walk OJ IV and Zehnder AM and Ayyar S and Page RL, et al. FasTag: Automatic text classification of unstructured medical narratives. PLoS one. 2020. 15(6). pmid:32569327
  5. 5. Coutinho E and Schuller B. Shared acoustic codes underlie emotional communication in music and speech—Evidence from deep transfer learning. PLoS one. 2017. 12(6). pmid:28658285
  6. 6. Yoo TK and Oh EK, Hong K and Ryu IH and Lee IS and Kim JS, et al. Deep learning-based smart speaker to confirm surgical sites for cataract surgeries: A pilot study. PLoS one. 2020. 15(4). pmid:32271836
  7. 7. Medvedeva M and Wieling M and Vols M. Rethinking the field of automatic prediction of court decisions. Artificial Intelligence and Law. 2022. 1–18.
  8. 8. Varghese A and Agyeman-Badu G and Cawley M. Deep learning in automated text classification: a case study using toxicological abstracts. Environment Systems and Decisions. 2020. 40(4): 465:479.
  9. 9. Balbuena LD. The UK Research Excellence Framework and the Matthew effect: Insights from machine learning. PLoS one. 2018. 13(11).
  10. 10. Curry, S and Gadd, E and Wilsdon, J. REF 2021: The metric tide rises again. 2022. Available from: https://www.researchprofessionalnews.com/rr-news-political-science-blog-2022-5-the-metric-tide-rises-again/
  11. 11. Wilsdon J and Allen L and Belfiore E and Campbell P and Curry S and Hill S, et al. The Metric Tide: Report of the Independent Review of the Role of Metrics in Research Assessment and Management. Higher Education Funding Council for England. 2015.
  12. 12. Cornelius JL. Reviewing the review process: Identifying sources of delay. The Australasian medical journal. 2012. 5(1). pmid:22905052
  13. 13. McGarity, TO. Peer review in awarding federal grants in the arts and sciences. High Tech. LJ. 1994. HeinOnline; 9: 1.
  14. 14. Tennant JP and Dugan JM and Graziotin D and Jacques DC and Waldner F and Mietchen D, et al. A multi-disciplinary perspective on emergent and future innovations in peer review. F1000Research. 2017. 6. pmid:29188015
  15. 15. Tomkins A and Zhang M and Heavlin WD. Reviewer bias in single-versus double-blind peer review. Proceedings of the National Academy of Sciences. 2017. 114 (48): 12708–12713. pmid:29138317
  16. 16. Lasker SP. Peer Review system: A Golden standard for publications process. Bangladesh Journal of Bioethics. 2018. 9(1).
  17. 17. Casnici N and Grimaldo F and Gilbert N and Squazzoni F. Journal of the Association for Information Science and Technology. 2017. Wiley Online Library; 68(7): 1763–1771.
  18. 18. Chauvin A and Ravaud P and Baron G and Barnes C and Boutron I. The most important tasks for peer reviewers evaluating a randomized controlled trial are not congruent with the tasks most often requested by journal editors. BMC medicine. 2015. 13(1): 1–10.
  19. 19. Tickell, A. Review of research bureaucracy. 2021. Available from: https://www.gov.uk/government/publications/review-of-research-bureaucracy/
  20. 20. Research England. REF2021. REF 2021 (July). Available from: https://www.ref.ac.uk/
  21. 21. REF impact. UKRI (March). 2022. Available from: https://www.ukri.org/about-us/research-england/research-excellence/ref-impact/
  22. 22. Williams K and Grant J. A Comparative review of how the policy and procedures to assess research impact evolved in Australia and the UK. Research Evaluation. 2018. 27(2): 93–105.
  23. 23. Farla, K and Simmonds, P. REF Accountability Review: Costs, benefits and burden. 2015. Available from: https://www.technopolis-group.com/report/ref-accountability-review-costs-benefits-and-burden/
  24. 24. Manville, C and Morgan Jones, M and Frearson, M and Castle-Clarke, S and Henham, ML and Gunashekar, S, et al. Preparing impact submissions for REF 2014: An evaluation. RAND Europe: Cambridge, UK. 2015.
  25. 25. Grant, J and Hinrichs, S. The nature, scale and beneficiaries of research impact: An initial analysis of the Research Excellence Framework (REF) 2014 impact case studies. HEFCE-Higher Education Funding Council for England; 2015.
  26. 26. Yuan, W and Liu, P and Neubig, G. Can we automate scientific reviewing? 2021. arXiv preprint arXiv:2102.00176
  27. 27. Kim, S. Deep learning with R, Francois Chollet, Joseph J. Allaire, Shelter Island, NY: Manning. The International Biometric Society. 2020. 123–345.
  28. 28. Nelson, APK and Gray, RJ and Ruffle, JK and Watkins, HC and Herron, D and Sorros, N, et al. Patterns. Elsevier; 2022.
  29. 29. Weis JW and Jacobson JM. Learning on knowledge graph dynamics provides an early warning of impactful research. Nature Biotechnology. 2021; 39(10): 1300–1307. pmid:34002098
  30. 30. Chowdhury G and Koya K and Philipson P. Measuring the impact of research: Lessons from the UK’s Research Excellence Framework 2014. PLoS one. 2016. 11(6). pmid:27276219
  31. 31. Bonaccorsi A and Chiarello F and Fantoni G. Impact for whom? Mapping the users of public research with lexicon-based text mining. Scientometrics. 2021. 126(2): 1745–1774.
  32. 32. Adams, J and Loach, T and Szomszor, M. The diversity of UK research and knowledge. Analyses from the REF impact case studies. London, Digital research reports. 2015.
  33. 33. Reichard B and Reed MS and Chubb J and Hall G and Jowett L and Peart A, et al. Writing impact case studies: a comparative study of high-scoring and low-scoring case studies from REF2014. Palgrave Communications. 2020. Palgrave; 6(1): 1–17.
  34. 34. Wroblewska, MN. Impact evaluation in Norway and in the UK: A comparative study, based on REF 2014 and Humeval 2015-2017. ENRESSH working paper (01).2019.
  35. 35. Hughes T and Webber D and O’Regan N. Achieving wider impact in business and management: Analysing the case studies from REF 2014. Studies in Higher Education. 2019. 44(4): 628–642.
  36. 36. Marcella R and Lockerbie H and Bloice L Beyond REF 2014: The impact of impact assessment on the future of information research. Journal of Information Science. SAGE; 2016. 42(3): 369–385.
  37. 37. Chipman HA and George EI and McCulloch RE. BART: Bayesian additive regression trees. The Annals of Applied Statistics. 2010. 4(1): 266–298.
  38. 38. Reddick G and Malkov D and Sherbon B. Understanding the funding characteristics of research impact: A proof-of-concept study linking REF 2014 impact case studies with Researchfish grant agreements [version 2]. F1000Research. 2022.
  39. 39. Jordan, K and Carrigan, M. How was social media cited in 2014 REF Impact Case Studies? Impact of Social Sciences blog. London School of Economics and Political Science; 2018.
  40. 40. Brook L. Evidencing impact from art research: analysis of impact case studies from the REF 2014. The Journal of Arts Management, Law, and Society. 2018. 48(1): 57–69.
  41. 41. Kelly D and Kent B and McMahon A and Taylor J and Traynor M. Impact case studies submitted to REF 2014: The hidden impact of nursing research. Journal of Research in Nursing. 2016. 21(4): 256–268.
  42. 42. Terämä E and Smallman M and Lock SJ and Johnson C and Austwic MZ. Beyond Academia—Interrogating Research Impact in the Research Excellence Framework. PLoS one. 2016. 12(11). pmid:27997599
  43. 43. Priem, J and Piwowar, H and Orr, R. OpenAlex: A fully-open index of scholarly works, authors, venues, institutions, and concepts. 2022. arXiv preprint arXiv:2205.01833.
  44. 44. Fry E. Readability versus leveling. The reading teacher. 2002. 56(3): 286–291.
  45. 45. Harris, TL and Hodges, RE. The literacy dictionary: The vocabulary of reading and writing. 1995.
  46. 46. Flesch R. A new readability yardstick. Journal of Applied Psychology. 1948. 32(3): 221–233. pmid:18867058
  47. 47. Smith EA and Senter RJ. Automated readability index. Aerospace Medical Research Laboratories. 1967. 66(220). pmid:5302480
  48. 48. Dale, E and Chall, JS. A formula for predicting readability: Instructions. Educational research bulletin. 1948. 37–54.
  49. 49. Kincaid, J Peter and Fishburne Jr, Robert P and Rogers, Richard L and Chissom, Brad S. Derivation of new readability formulas (automated readability index, fog count and flesch reading ease formula) for navy enlisted personnel. Naval Technical Training Command Millington TN Research Branch. 1975.
  50. 50. Mc Laughlin GH. SMOG grading-a new readability formula. Journal of reading. JSTOR; 1969. Journal of reading. 12(8): 639–646.
  51. 51. Szomszor, M and Adie, E Overton–A bibliometric database of policy document citations. arXiv preprint arXiv:2201.07643. 2022.
  52. 52. Lilleberg, J and Zhu, Y and Zhang, Y. Support vector machines and word2vec for text classification with semantic features. 2015 IEEE 14th International Conference on Cognitive Informatics & Cognitive Computing (ICCI* CC). IEEE; 2015. 136–140.
  53. 53. Sirikulviriya N and Sinthupinyo S. Integration of rules from a random forest. International Conference on Information and Electronics Engineering. 2011. 6: 194–198.
  54. 54. Oshiro, TM and Perez, PS and Baranauskas, JA. How many trees in a random forest? In: International workshop on machine learning and data mining in pattern recognition. Springer. 2012. 154–168.
  55. 55. Kecman, V. Support vector machines–an introduction. In: Support vector machines: theory and applications, p.2. Springer. 2005.
  56. 56. Kinne Jan and Lenz David. Predicting innovative firms using web mining and deep learning. PLoS one. 2021. 16(4). pmid:33793626
  57. 57. Gimenez, C. Case studies and surveys in supply chain management research—Two Complementary Methodologies. Research methodologies in supply chain management. 2005. Springer; 315–330.
  58. 58. Gunning D and Stefik M and Choi J and Miller T and Stumpf S and Yang GZ. XAI—Explainable artificial intelligence. Science robotics. 2019. 4(37). pmid:33137719
  59. 59. Lipton ZC. The mythos of model interpretability: In machine learning, the concept of interpretability is both important and slippery. Queue. 2018. 16(3): 31–57.
  60. 60. Nelson APK. Reframing research impact. Patterns. Elsevier; 2022. 3(5). pmid:35607627
  61. 61. Joachims, T. Learning to classify text using support vector machines. Springer Science & Business Media. 2002. 668.
  62. 62. Oancea A. Interpretations of Research Impact in Seven Disciplines. European Educational Research Journal. 2013. 2(12): 242–250.
  63. 63. Slyder JB and Stein BR and Sams BS and Walker DM and Beale BJ and Feldhaus JJ, et al. Citation pattern and lifespan: a comparison of discipline, institution, and individual. Scientometrics. 2011. 3(89): 955.
  64. 64. Smith KE and Bandola-Gill J and Meer N and Stewart E and Watermeyer R. The impact agenda: Controversies, consequences and challenges. Policy Press. 2020.
  65. 65. Boswell C and Smith K. Rethinking policy ‘impact’: four models of research-policy relations. Palgrave Communications. 2017. 3(1): 1–10.
  66. 66. Gandomi A and Haider M. Beyond the hype: Big data concepts, methods, and analytics. International journal of information management. 2015. 35(2): 137–144.