Figures
Abstract
Artificial intelligence (AI) algorithms are transforming several areas of the digital world and are increasingly being applied in healthcare. Mobile apps based on predictive machine learning models have the potential to improve health outcomes, but there is still no consensus on how to inform doctors about their results. The aim of this study was to investigate how healthcare professionals prefer to receive predictions generated by machine learning algorithms. A systematic search in MEDLINE, via PubMed, EMBASE and Web of Science was first performed. We developed a mobile app, RandomIA, to predict the occurrence of clinical outcomes, initially for COVID-19 and later expected to be expanded to other diseases. A questionnaire called System Usability Scale (SUS) was selected to assess the usability of the mobile app. A total of 69 doctors from the five regions of Brazil tested RandomIA and evaluated three different ways to visualize the predictions. For prognostic outcomes (mechanical ventilation, admission to an intensive care unit, and death), most doctors (62.9%) preferred a more complex visualization, represented by a bar graph with three categories (low, medium, and high probability) and a probability density graph for each outcome. For the diagnostic prediction of COVID-19, there was also a majority preference (65.4%) for the same option. Our results indicate that doctors could be more inclined to prefer receiving detailed results from predictive machine learning algorithms.
Citation: Wichmann RM, Fagundes TP, de Oliveira TA, Batista AFdM, Chiavegatto Filho ADP (2022) Physician preference for receiving machine learning predictive results: A cross-sectional multicentric study. PLoS ONE 17(12): e0278397. https://doi.org/10.1371/journal.pone.0278397
Editor: Avid Roman-Gonzalez, Business on Engineering and Technology S.A.C (BE Tech), PERU
Received: May 22, 2022; Accepted: November 15, 2022; Published: December 14, 2022
Copyright: © 2022 Wichmann et al. This is an open access article distributed under the terms of the Creative Commons Attribution License, which permits unrestricted use, distribution, and reproduction in any medium, provided the original author and source are credited.
Data Availability: Data provided as a spreadsheet in .xls format with the unidentified dataset described as supporting files (Dataset.xls).
Funding: This work was supported by grant number 206/2020, Paraíba State Research Foundation (FAPESQ), the National Council for Scientific and Technological Development (CNPq) under Grant Number 402626/2020-6, and Microsoft (Microsoft AI for Health COVID-19 Grant). The funders had no role in study design, data collection and analysis, decision to publish, or preparation of the manuscript.
Competing interests: The authors have declared that no competing interests exist.
Introduction
Doctors, nurses, physiotherapists, psychologists, among other healthcare professionals, face a massive amount of health information, and traditional ways of managing and evaluating information may not be enough for an efficient use of these resources [1]. A large number of information technologies (e-health applications) are already being used by organizations and individuals to make healthcare data applicable to patients [2].
Applications based on Artificial Intelligence (AI), specifically on Machine Learning (ML), have been on the rise in the last decade [3]. For example, machine learning-based diagnostic models have been developed to identify individuals with influenza [4]. A recent study carried out by US researchers validated a prediction model for Ebola patients, later deployed into a mobile app [5]. During the COVID-19 pandemic, geographic tracking applications were also developed to identify clusters of infected individuals and to improve collective decisions based on these results [6].
Some of the resistance from physicians in using e-health solutions may be related to the fear of new responsibilities [7]. Thus, interventions that require too many adaptations by the professionals may not be very effective. In addition, it has been suggested that presenting the result by asking the professional to make a specific decision (e.g. to intubate) or showing scores that are difficult to understand should be avoided [8]. In order to work directly with predictive algorithms, physicians will need skills usually not taught during medical training, such as performing and interpreting advanced calculations, database management, and programming [9]. For AI to have its potential fully implemented in healthcare, it is important that users are able to interpret the outputs of these predictions. Doctor-patient relationships can be negatively impacted if the professional does not know how to communicate uncertainties from a decision that was taken with the help of AI-based applications [10].
The predictive performance of a model may be insufficient to determine the success and popularity of a ML-based medical application among health professionals [11]. The presentation of the results, its usability and the amount of information provided can directly impact the use of the application. Equipping physicians with easy-to-use and accessible algorithms can help improve decisions especially in the face of a new disease scenario [12]. Regarding medical mobile applications, developers should also consider the usability, which is defined by The International Organization for Standardization as "the extent to which a product can be used by a defined group of users to achieve specific goals with effectiveness, efficiency and satisfaction in a given context of use” [13]. There are validated questionnaires for assessing the usability of applications in general, not necessarily medical, such as the System Usability Scale Task Completion (SUS) and the Suitability Assessment of Materials (SAM), which assess the complexity, integration, need for previous knowledge to use it, and presence of inconsistency of the applications, among other factors [14].
To our knowledge, there are still no tools available for usability assessment of ML-based eHealth applications. There are also no studies that investigate the level of technical details of these results that should be shown to healthcare professionals. The aim of this study is therefore to test how health professionals prefer to receive the results of ML-based prediction models and to contribute to the improvement of future technologies that use AI algorithms.
Methods
A mobile application, RandomIA, was developed by the IACOV-BR (Artificial Intelligence for COVID-19 in Brazil) network to provide doctors with predictions of negative health outcomes. Patient data such as sex, age, vital signs, and complete blood count were used to develop the predictive algorithm (S1 Table). The study was approved by the Research Ethics Committee of the School of Public Health of the University of São Paulo (CAAE: 44298521.1.0000.5421).
Study design
A search in EMBASE and MEDLINE via PubMed was conducted using the terms “mobile application”, “artificial intelligence” and “COVID-19” to identify studies that developed questionnaires for validating e-health applications, specifically regarding those employing machine learning algorithms. Then, a cross-sectional multicentric study was conducted from November 2021 to March 2022 by applying a questionnaire to physicians from the IACOV-BR network. Each professional had access to RandomIA, where they were asked to input clinical data from at least three patients affected by COVID-19. Three distinct types of results were presented with varying degrees of complexity. Finally, the physicians answered a questionnaire that asked to order the type of results they preferred. A group of 20 questions were selected (S2 Table) and incorporated into RandomIA. The first ten questions are part of a widely validated form, the System User’s Usability scale (SUS), which quantifies the interaction between users and an application. All participants accepted the informed consent form to evaluate the application.
Population
The main inclusion criterion for the study was to be a physician in a hospital from the IACOV-BR network associated with the project. After selecting 18 hospitals to participate in the study, a convenience sampling of medical users was conducted and a minimum number of doctors was established to compose the group. This was determined based on a finite number of physicians from the IACOV-BR institutions (30,000) and the calculation of the sample size from a simple random sample [15]. The minimum sample size of physicians to represent this specific population was estimated at 68 sample units (n = 68), when considering a confidence level of 90% [16].
RandomIA
RandomIA is a digital application that aims to use artificial intelligence algorithms to provide diagnostic and prognostic predictions for COVID-19 (Fig 1) The testing process by the physician began with the login to RandomIA via the random-ia.com URL and the filling in of the mandatory login and password fields, previously created and made available to each user by the research group. After logging into the application, the following options appeared on the screen: new prediction, history, learn more and evaluation survey. The user was then able to select “new prediction” to start filling the available fields with data from patients for which a RT-PCR exam was performed, regardless of the result (positive or negative). This process was performed three different times by each physician, using data from three patients to then visualize three different ways of receiving the prediction results (Fig 2).
Statistical analysis
Descriptive statistics of the results were performed, in which numerical variables were presented as means and standard deviations, and categorical variables as absolute and relative frequencies. The variables originally distributed on a Likert scale are those which the response required participants to classify their preferences according to the following possibilities: “strongly disagree”, “disagree”, “neither agree nor disagree’’, “agree” and “strongly agree” on an equal distance scale [17]. For interpreting the results of the analysis, we considered that “strongly disagree” and “disagree” were low grades, that is, unfavorable to the statement of the item. On the other hand, the options “strongly agree” and “agree’’ represented the high grades of the Likert scale. Column graphs on the Likert scale were created using the R Likert package [18].
An ordinal multinomial regression analysis was also performed. Here we used the Akaike information criterion-based approach to determine if the data fitted as likely the model as a model that suggests no difference between the groups (nested models). At the same time, likelihood ratio tests were applied to the nested models to assess the hypothesis that the answers were parallel. This verification was performed via a chi-square test statistic χ2 for the likelihood ratio [19]. The probabilities were plotted in a line chart for comparison, using the net package from R, version 4.0.3. Factor Analysis and Principal Component Analysis aim to preserve the original variability of the data by summarizing variables correlated with each other into a smaller set of uncorrelated variables that gather the most information from the original set [20]. A scree-plot graphic approach was used [21], in order to assess whether there were differences between the groups in terms of sex, age, medical specialty, and region of the country.
Results and discussion
A total of 69 physicians were recruited and asked to evaluate the options to visualize the result after testing the application at least three times.
Table 1 presents the descriptive results of the questionnaire. Variable Q1 (“I would use the app frequently”) had an average response of 3.36, slightly higher than the median of 3.0, indicating a higher proportion of physicians likely to use the app. On the other hand, variable Q2 (“I considered the application hard to use”), had an average of 1.87, demonstrating the ease of use of the application and corroborating the result of variable Q1. For variable Q3, (“I found the application easy to use”), 25% of the participants rated it up to 4 on the Likert scale, while the remaining 75% responded 4 or 5 on the same scale, suggesting an easy use of RandomIA. For variable Q4 (“I would need help from a person with technical knowledge to use the application”), an average of 1.38 was found, reinforcing the result obtained for variable Q3.
The even number questions were unfavorable to RandomIA’s usability, while odd questions were favorable. The means and medians for questions 1, 3, 5, 7 and 9 were of higher values than the means and medians for questions 2, 4, 6, 8 and 10 (Table 1). The visualization of the results from the Likert scale from Table 1 can be found in S1 Fig.
For Q11 (“I would change my medical management based on the results provided by RandomIA after its final and validated version, such as intubating, hospitalizing, or starting intensive care early”), the mean was 2.71. The value, when compared to the median of 3 on the Likert scale, suggests that physicians would not possibly change their medical conduct based solely on the result provided. In variable Q13 (“How confident are you in the prediction information available in the application”), the proportion of positive scores (4 and 5) was higher than the proportion of negative scores (1 and 2), indicating confidence in the information contained in the application.
Fig 3 presents the correlations between the items in the RandomIA questionnaire. In general, correlations were weak, ranging from -0.5 to 0.5. Questions Q19 with Q1, Q9, Q11, Q13 and Q14; and Q3 with Q2, Q8 and Q12 had strong correlations, but below 0.8. For detailed information on the definition of each item, see S2 Table.
Fig 4a highlights the favorable responses to the usability of the application. The visualization of high grades in blue scale shows a greater proportion of these answers. On the other hand, Fig 4b shows a higher proportion of lower grades. However, this result corroborates the favorable usability of the application, since these questions, visualized in shades of red, report against the usability of the application.
Table 2 presents the proportion of responses according to each variable and their categories. In Q12 (“How do you consider using the application for customer service”), 79.7% of the participants found the application easy to use and quick to obtain the answer, while only 1.4% considered it difficult and slow to obtain. In variable Q14 ("The application’s prediction results were contrary to the diagnostic and/or prognostic impressions in their daily clinical practice"), for 42% of the participating physicians there was no divergence between the prediction and the clinical practice impressions, but 17.39% responded that there was a discrepancy with the diagnostic impression. In variable Q15 (“What do you think about the number of patient information/variables currently available in the app?”), 39.1% of professionals reported that the number of information requested was adequate, while 43.5% responded that there was a lack of relevant information.
In Q17, 62.3% of participants preferred equal views of prediction results for diagnostic and prognostic outcomes, while the other 37.7% opted for different options. Among those who opted for similar visualizations, 65.9% preferred the third, i.e., the most complex. Still in this group, 13.04% indicated that their choice was motivated by: (i) the presentation was simpler and more intuitive, (ii) the presentation was visually easier to understand, (iii) the subtitles were more explanatory, and (iv) the colors helped in the interpretation of the results. Among the 37.7% who opted for visualization of different results, 65.4% preferred the more complex visualization for the diagnostic outcome results and 62.9% opted for a more complex visualization for the prognostic outcome.
In variable Q18 (“The explanation of the options was sufficient to understand the results of the predicted outcomes”), 81.2% of physicians understood the explanation provided by the description of the outcomes, while 17.4% partially understood it. Finally, for variable Q19 (“I would recommend this prediction diagnostic app for use as clinical decision support”), 69.6% of participants would recommend the prediction app for clinical decision support.
Table 3 presents the differences in proportions of the response options distributed on a Likert Scale, according to sex. There were differences between positive and negative proportions in almost all System Use Scale (SUS) questions. There was no statistically significant difference between the proportions of questions Q5 and Q8 for females. The responses were statistically favorable to the usability of the application, when using the Wilcoxon Test (Table 3).
Regarding question Q5 (“I found the various functions in this system were well integrated”), despite the superiority of positive responses over the negative, there was no statistically significant difference between these proportions for females. The same occurred with Q8 (“I found the system very cumbersome to use”), in which the answers were negative when considering the female gender, but with no statistically significant difference. Overall, the profile of responses between the sexes was different. The graphical visualizations of the proportions of the Likert scale for biological sex can be found in the S2 and S3 Figs.
Table 4 presents the difference in proportions for the Likert scale options according to the age group of physicians. The age variable of physicians was separated into two groups: younger (18 to 39 years of age) and seniors (40 years old or more). The answers according to the age groups were uniform, since in the odd questions, which is about the positive usability of the application, the answers had a higher proportion of higher (positive) scores. This indicates that there was a significant difference when using the Wilcoxon test in relation to responses with a low score (negative) in both age groups (young and senior). The same behavior was observed for questions that had statements against the good usability of the application (even questions). In these, the scores were predominantly low and, therefore, contrary to the statements of low usability. In this type of questions, the Wilcoxon test detected a significant difference in all responses.
When the results were analyzed along the Likert scale, we found that they were parallel, i.e., that they did not present significant differences. This indicates that age did not influence the type of visualization and the response that clinicians expect from an e-health application based on a machine learning solution. The graphical visualizations of the proportions of the Likert scale for age can be found in the S4 and S5 Figs.
Table 5 presents the results according to having a medical specialty and the S7 and S8 Figs show the graphical visualizations of the proportions of their Likert scale. When comparing physicians with some type of specialty with general practitioners, we found that the overall response pattern was the same. That is, there was no statistical difference when considering the professional profiles of physicians (p-value of χ2 for the likelihood ratio). Regarding the difference in positive and negative answers for the SUS questions, we found that for the odd-numbered questions, there were differences in the proportion of high scores in relation to low scores. The proportion of high grades was higher in the odd (Q1, Q3, Q5, Q7 and Q9) and even (Q2, Q4, Q6, Q8 and Q10) items, reiterating the favorable evaluation of the application. For differences in Likert scale proportions by geographic region of Brazil, see S3A and S3B Table and S6–S11 Figs.
To identify the factors associated with the usability of the application, we performed Factor Analyses. Initially, applicability was tested using the Kaiser-Meyer-Olkin (KMO) test and a value of 0.64 was obtained (p-value >0.5). This indicates the suitability of applying the technique. Then the Bartlett sphericity test was performed to test the null hypothesis that the correlation matrix is an identity matrix, and a significance p-value < 0.01 was obtained. Then the scree-plot was performed to determine the number of values necessary to retain the greatest amount of variability (S12 Fig).
The commonalities are amounts of variances (correlations) of each variable that are explained by the factors. The greater the commonality, the greater is the explanatory power of that variable by the factor. Specificity or error is the portion of data variance that cannot be explained by the factor and is characterized by the unique proportion of the variable not shared with the others. This value is obtained by subtracting one from the commonality. The higher the specificity, the lower the relevance of a given variable in the factor model. In Table 6, the variables Q23, Q22, Q20, Q17, Q16, Q14 had less weight in the factor analysis.
The variables that make up each factor are shown in Fig 5. Factor 3 represented by PA1 is composed of positive loadings in Q12, Q8, Q2 and negative in Q3 and Q7. This factor involves variables regarding the use of the application, i.e., the usability factor. Factor 4 (PA2) is composed of the items (Q13, Q1, Q11, Q9 and Q14) in a positive way and Q19 and Q15 in a negative way. This factor can be considered a measure of confidence in the use of the application. Factor 5 (PA3) is composed of variables Q20 and Q23 with positive loadings, and Q22 with negative loading. This combination is related to the characteristics of the physicians participating in the study. Factor 4 has questions Q10, Q4, Q18 and Q6, a factor related to difficulties in using the application. Factor 5 is linked to the amount of application outcomes. Factor 6 refers to the integration of application functions. Factor 7, of great relevance to the outcome of the present study, refers to the type of preference for viewing outcomes. Finally, Factor 8 is related to age.
The principal component analyzes and biplots to verify the association between the questions and the physicians, based on the RandomIA questionnaire, can be found in S4 and S5 Tables, S13 and S14 Figs.
Biplot analysis by subgroups helped in the evaluation of the robustness of the results found for the RandomIA questionnaire. Such representations usually allow the visualization of vectors represented by the questions, in which the size of these vectors is associated with the importance of these items, while ellipses are representations of the area covered by these questions within each group.
An analysis by the presence or absence of a medical specialty is represented by the number 1 for general practitioners and 2 for specialist doctors (Fig 6). There was no difference between the groups since the ellipses are superimposed. Thus, it was not possible to analyze which specialty has greater acceptance of the application.
Regarding the sex of the participants, the association represented by the biplot in Fig 7 indicates that there was no difference between the groups, which is both men (group 1) and women (group 2) are superimposed for the questions.
The differences between the regions of the country are represented by S15 Fig. Through the techniques shown, it was possible to reduce a 23-question questionnaire to eight latent variables with factor analysis. Such a tool is useful for summarizing information obtained from questionnaires applied to the most diverse subjects. It was also possible to identify the factors giving them a practical meaning. These results corroborated the findings of the Likert scale.
Discussion
The study found a high physician adherence to a ML-based application for predicting health outcomes. More importantly, the questions designed to test the usability of the application proved useful for this purpose. We also found that most physicians preferred a more complex form of visualization for receiving the final predictions generated by the machine learning algorithm.
There have been some previous proposals for mobile applications based on ML, all developed from 2016 onwards [5, 22–26]. The limitations listed by these studies include insufficient observations from physicians, lack of publicly available data, unavailability of internet connection in remote areas and lack of reliability of sources that provided the data sets. However, the interaction between users, in this case the medical community, and the results provided by the algorithms are rarely discussed as a potential limitation of the use of such ML models. The difficulties of implementing a machine learning model in clinical practice are not only due to technical difficulties in developing the algorithms, but may be also due to ethical difficulties [27] and the diversity of users and developers involved [28], among other challenges.
This study had some limitations. As this is a new area of research, previous questionnaires capable of evaluating the usability of ML-based health applications were not identified in the literature, so we had to adapt SUS for this specific purpose. Another limitation is that a convenience sampling was the only technique available to collect data among participants of the network. Lastly, the high proportion of negative results in Q11 (“I would change my medical conduct based on the results provided by RandomIA after its final and validated version, e.g., intubation or not hospitalized or not starting intensive care early, etc.”), may be since due to the fact that the actual algorithm used to provide the prediction has not yet been properly calibrated for accurate assertiveness regarding COVID-19 prognostic results.
Conclusions
Our study was the first to assess how healthcare professionals prefer to receive the results of ML-based applications. The main results from the study can be extended to the development of other e-health technologies and the improvement of already existing ones, as it may help in the understanding of discouraging factors for users and in the better knowledge of the target audience of new AI applications.
Supporting information
S1 Table. Variables used in predictive models.
https://doi.org/10.1371/journal.pone.0278397.s001
(DOCX)
S3 Table.
A. Difference in the proportions of the Likert Scale options by Brazil region for the first five questions. B. Difference in the proportions of the Likert Scale by Brazil region for the last five questions.
https://doi.org/10.1371/journal.pone.0278397.s003
(ZIP)
S4 Table. Details of factors obtained through Principal Component Analysis.
https://doi.org/10.1371/journal.pone.0278397.s004
(DOCX)
S5 Table. Principal Component Analysis results.
https://doi.org/10.1371/journal.pone.0278397.s005
(DOCX)
S1 Fig. Choropleth representation of proportions on the Likert scale.
https://doi.org/10.1371/journal.pone.0278397.s006
(DOCX)
S2 Fig. Barplot proportions of the Likert Scale options by biological sex for odds questions.
https://doi.org/10.1371/journal.pone.0278397.s007
(DOCX)
S3 Fig. Barplot proportions of the Likert Scale options by biological sex for even questions.
https://doi.org/10.1371/journal.pone.0278397.s008
(DOCX)
S4 Fig. Barplot proportions of the Likert Scale options by age group for odds questions.
https://doi.org/10.1371/journal.pone.0278397.s009
(DOCX)
S5 Fig. Barplot proportions of the Likert Scale options by age group for even questions.
https://doi.org/10.1371/journal.pone.0278397.s010
(DOCX)
S6 Fig. Barplot proportions of the Likert Scale options by medical speciality for odds questions.
https://doi.org/10.1371/journal.pone.0278397.s011
(DOCX)
S7 Fig. Barplot proportions of the Likert Scale options by medical speciality for even questions.
https://doi.org/10.1371/journal.pone.0278397.s012
(DOCX)
S8 Fig. Barplot proportions of the Likert Scale options by Brazil regions for odds questions (Q1, Q3, Q5).
https://doi.org/10.1371/journal.pone.0278397.s013
(DOCX)
S9 Fig. Barplot proportions of the Likert Scale options by Brazil regions for odds questions (Q7, Q9).
https://doi.org/10.1371/journal.pone.0278397.s014
(DOCX)
S10 Fig. Barplot proportions of the Likert Scale options by Brazil regions for even questions (Q2, Q4, Q6).
https://doi.org/10.1371/journal.pone.0278397.s015
(DOCX)
S11 Fig. Barplot proportions of the Likert Scale options by Brazil regions for even questions (Q8, Q10).
https://doi.org/10.1371/journal.pone.0278397.s016
(DOCX)
S12 Fig. Scree-plot of the eigenvalues sorted in descending order for the RandomIA questionnaire data.
https://doi.org/10.1371/journal.pone.0278397.s017
(DOCX)
S13 Fig. Biplot showing the association between items plotted in the first two dimensions from the RandomIA questionnaire.
https://doi.org/10.1371/journal.pone.0278397.s018
(DOCX)
S14 Fig. Biplot showing the dispersion of doctors in relation to the questions in the first two dimensions, based on the RandomIA questionnaire.
https://doi.org/10.1371/journal.pone.0278397.s019
(DOCX)
S15 Fig. Biplot showing association of the questions in the first two dimensions from the RandomIA questionnaire by Brazilian regions.
https://doi.org/10.1371/journal.pone.0278397.s020
(DOCX)
Acknowledgments
†In the name of the IACOV-BR network, in alphabetic order: Ana Maria Espírito Santo de Brito (Instituto de Medicina, Estudos e Desenvolvimento—IMED, São Paulo, São Paulo); Anamaria Mello Miranda Paniago (School of Medicine, Federal University of Mato Grosso do Sul); Danielle Saad Nemer (Hospital Municipal Vila Santa Catarina); Fernando Anschau (Setor de Pesquisa da Gerência de Ensino e Pesquisa do Grupo Hospitalar Conceição, RS—Brasil e Programa de Pós-Graduação em Neurociências da Universidade Federal do Rio Grande do Sul); Gabriel Ferreira dos Santos Silva (HCOR Innovation Laboratory, SP, Brazil); Gustavo Wenzel Sainatto (Instituto do Câncer do Estado de São Paulo–ICESP); João Conrado Bueno dos Reis (Hospital São Francisco); Liane de Oliveira Cavalcante (Hospital Santa Júlia); Luiz Fernando Nogueira Simvoulidis (Hospital Unimed-Rio); Maria Elizete de Almeida Araujo (Federal University of Amazonas, University Hospital Getúlio Vargas, Manaus, Am Brazil); Renan Magalhães Montenegro Junior (Complexo Hospitalar da Universidade Federal do Ceará–EBSERH); Renan Martello Cristófalo (LABDAPS da Faculdade de Saúde Pública da USP); Renata Vicente da Penha (Hospital Evangélico de Vila Velha); Rogério Nadin Vicente (Hospital Santa Catarina de Blumenau), Ruchelli França de Lima (Hospital Israelita Albert Einstein); Sandro Rodrigues Batista (Faculdade de Medicina, Universidade Federal de Goiás, Goiânia, Goiás e Secretaria de Estado da Saúde de Goiás, Goiânia, Goiás).
We thank all the 69 medical doctors who have tested the RandomIA app. They are in alphabetical order: Alex Fantinatti Teixeira; Alexandre Bertucci; Amanda Ellen de Morais; Ana Maria Trufelli; Anamaria Mello Miranda Paniago; André Filipe Marcondes Vieira; Angélica Lanza Vieira de Souza; Anthony Magalhães Morais Santiago; Arlene dos Santos Pinto; Bárbara Seabra Carneiro; Clara Elisa Frare de Avelar Teixeira; Claudia Elizabeth Volpe Chaves; Claudia Maria Costa de Oliveira; Danielle Saad Nemer; Dany Jasinowodolinski; Douglas Nunes Cavalcante; Eduardo Menezes Lopes; Ettore Mendes Azenha; Fabiana Finisgalli Romanello Campos; Fernanda Arns de Castro; Fernando Da Silveira; Gabriela Studart Galdino; Gedealvares Francisco de Souza Junior; Giulia Maria Ximenes Verdi; Guilherme Pina do Carmo; Guilherme Quireza Silva; Gustavo Shikanai Kerr; Henrique Souza Santos; Humberto Bolognini Tripadelli; Ítalo Gustavo Lima Monteiro; Jamell Cristina Malta Ohev Zion; João Luiz Miraglia; Kevin Yun Kim; Laís Cerqueira de Moraes; Laura de Freitas Faveri; Leonardo Marques; Lílian Machado Contente Nogueira; Lucas Gondim Briand Vieira; Lucas Messias Ribeiro da Cunha; Luiz Fernando Nogueira Simvoulidis; Márcia Barbosa de Freitas; Márcio Emanuel Gonoring França; Mateus Gustavo Favaro; Matheus Lacerda da Silva Ferro Costa; Matheus Merlin Felizola; Milena Martello Cristófalo; Naiane Moreira Barbosa; Natalia Martins Bonassi; Otilio da Silva Canuto; Paloma Farina de Lima; Peter Gonçalves; Phamela da Silva Feitosa; Rafael Machado Goncalves; Rainardo Antonio Puster; Renata Jordão Guimarães; Rogério Nadin Vicente; Suzana Alves da Silva; Taiani Vargas; Tainá Veras de Sandes Freitas; Tamiris Silva de Oliveira; Thaís de Melo Costa; Thalita Torres Sales; Thalles Augusto dos Santos Porfirio; Thiago Rodrigues Sequeira; Thyago Gregório Mota Ribeiro; Vanessa Pinheiro de Queiroz; Vitor Negreiro Leão; Viviane Soares Damascena dos Santos; Wilands Patricio Procopio Gomes.
References
- 1. Klerings I, Weinhandl AS, Thaler KJ. Information overload in healthcare: too much of a good thing? Zeitschrift fur Evidenz, Fortbildung und Qualitat im Gesundheitswesen. 2015;109(4–5):285–90. pmid:26354128
- 2. Viitanen J, Hyppönen H, Lääveri T, Vänskä J, Reponen J, Winblad I. National questionnaire study on clinical ICT systems proofs: physicians suffer from poor usability. International journal of medical informatics. 2011;80(10):708–25. pmid:21784701
- 3. Swapnarekha H, Behera HS, Nayak J, Naik B. Role of intelligent computing in COVID-19 prognosis: A state-of-the-art review. Chaos, solitons, and fractals. 2020;138:109947. pmid:32836916
- 4. Yao Y, Sun G, Matsui T, Hakozaki Y, van Waasen S, Schiek M. Multiple Vital-Sign-Based Infection Screening Outperforms Thermography Independent of the Classification Algorithm. IEEE transactions on bio-medical engineering. 2016;63(5):1025–33. pmid:26394412
- 5. Colubri A, Hartley M-A, Siakor M, Wolfman V, Felix A, Sesay T, et al. Machine-learning Prognostic Models from the 2014–16 Ebola Outbreak: Data-harmonization Challenges, Validation Strategies, and mHealth Applications. eClinicalMedicine. 2019;11:54–64. pmid:31312805
- 6. Kamel Boulos MN, Geraghty EM. Geographical tracking and mapping of coronavirus disease COVID-19/severe acute respiratory syndrome coronavirus 2 (SARS-CoV-2) epidemic and associated events around the world: how 21st century GIS technologies are supporting the global fight against outbreaks and epidemics. International Journal of Health Geographics. 2020;19(1):8. pmid:32160889
- 7. Boonstra A, Broekhuis M. Barriers to the acceptance of electronic medical records by physicians from systematic review to taxonomy and interventions. BMC health services research. 2010;10:231. pmid:20691097
- 8. Diprose WK, Buist N, Hua N, Thurier Q, Shand G, Robinson R. Physician understanding, explainability, and trust in a hypothetical machine learning risk calculator. Journal of the American Medical Informatics Association: JAMIA. 2020;27(4):592–600.
- 9. Paranjape K, Schinkel M, Nannan Panday R, Car J, Nanayakkara P. Introducing Artificial Intelligence Training in Medical Education. JMIR Med Educ. 2019;5(2):e16048–e. pmid:31793895
- 10. Price W. N. II. Regulating Black-Box Medicine. Mich L Rev 421. 2017. pmid:29240330
- 11. Cutillo CM, Sharma KR, Foschini L, Kundu S, Mackintosh M, Mandl KD, et al. Machine intelligence in healthcare—perspectives on trustworthiness, explainability, usability, and transparency. NPJ digital medicine. 2020;3(1):47. pmid:32258429
- 12. Syeda HB, Syed M, Sexton KW, Syed S, Begum S, Syed F, et al. Role of Machine Learning Techniques to Tackle the COVID-19 Crisis: Systematic Review. JMIR Med Inform. 2021;9(1):e23811–e. pmid:33326405
- 13.
International Organization for Standardization. ISO 9241–11:2018(en). Ergonomics of human-system interaction—Part 11: Usability: Definitions and concepts. Geneva: ISO, 2018.
- 14. Maramba I, Chatterjee A, Newman C. Methods of usability testing in the development of eHealth applications: A scoping review. International journal of medical informatics. 2019;126:95–104. pmid:31029270
- 15.
Levy PSL, S. Sampling of Populations: Methods and Applications. 4 ed. Hoboken: John Wiley & Sons; 2008.
- 16. Shanmugam R. Modern survey sampling. Journal of Statistical Computation and Simulation. 2019;89(5):948-.
- 17.
Robbins NB, Heiberger RM, editors. Plotting Likert and Other Rating Scales 2011.
- 18.
Jason Bryer KS. likert: Analysis and Visualization Likert Items https://cran.r-project.org/2016 [An approach to analyzing Likert response items, with an emphasis on visualizations. The stacked bar plot is the preferred method for presenting Likert results. Tabular results are also implemented along with density plots to assist researchers in determining whether Likert responses can be used quantitatively instead of qualitatively. See the likert(), summary.likert(), and plot.likert() functions to get started.]. https://CRAN.R-project.org/package=likert.
- 19.
Rao JNK, Scott AJ. Chi-squared Tests For Multiway Contigency Tables with Proportions Estimated From Survey Data. 1984(12):46–60.
- 20.
JOHNSON RA, WICHERN, D. W. Applied multivariate statistical analysis: Pearson; 2018. 808 p.
- 21. Cattell RB. The Scree Test For The Number Of Factors. Multivariate behavioral research. 1966;1(2):245–76. pmid:26828106
- 22. Jiménez-Serrano S, Tortajada S, García-Gómez JM. A Mobile Health Application to Predict Postpartum Depression Based on Machine Learning. Telemedicine journal and e-health: the official journal of the American Telemedicine Association. 2015;21(7):567–74. pmid:25734829
- 23. De Souza MLM, Lopes GA, Branco AC, Fairley JK, Fraga LAO. Leprosy Screening Based on Artificial Intelligence: Development of a Cross-Platform App. JMIR mHealth and uHealth. 2021;9(4):e23718. pmid:33825685
- 24.
Zia S, Khan AN, Mukhtar M, Ali SE, Shahid J, Sohail M, editors. Detection of Motor Seizures and Falls in Mobile Application using Machine Learning Classifiers. 2020 IEEE International Conference on Industry 40, Artificial Intelligence, and Communications Technology (IAICT); 2020 7–8 July 2020.
- 25. Oyebode O, Alqahtani F, Orji R. Using Machine Learning and Thematic Analysis Methods to Evaluate Mental Health Apps Based on User Reviews. IEEE Access. 2020;8:111141–58.
- 26. Pascucci M, Royer G, Adamek J, Asmar MA, Aristizabal D, Blanche L, et al. AI-based mobile application to fight antibiotic resistance. Nature communications. 2021;12(1):1173. pmid:33608509
- 27. Heyen NB, Salloch S. The ethics of machine learning-based clinical decision support: an analysis through the lens of professionalisation theory. BMC Medical Ethics. 2021;22(1):112. pmid:34412649
- 28.
Doshi-Velez F, Kim BJapa. Towards a rigorous science of interpretable machine learning. 2017.