Skip to main content
Advertisement
Browse Subject Areas
?

Click through the PLOS taxonomy to find articles in your field.

For more information about PLOS Subject Areas, click here.

  • Loading metrics

Assessing the feasibility of large language models to identify top research priorities in enhanced external counterpulsation

  • Shengkun Gai,

    Roles Methodology, Writing – original draft, Writing – review & editing

    Affiliations Department of Cardiology, The First Affiliated Hospital, Sun Yat-sen University, Guangzhou, China, Department of Cardiology, Linfen People’s Hospital, Linfen, Shanxi, China

  • Fangwan Huang,

    Roles Data curation, Formal analysis

    Affiliation College of Computer and Data Science, Fuzhou University, Fuzhou, China

  • Xuanyun Liu,

    Roles Data curation, Formal analysis

    Affiliation College of Computer and Data Science, Fuzhou University, Fuzhou, China

  • Ryan G. Benton,

    Roles Supervision

    Affiliation School of Computing, University of South Alabama, Mobile, Alabama, United States of America

  • Glen M. Borchert,

    Roles Supervision

    Affiliation College of Medicine, University of South Alabama, Mobile, Alabama, United States of America

  • Jingshan Huang ,

    Roles Writing – review & editing

    huang@southalabama.edu (JH); lengxy@mail.sysu.edu.cn (XL)

    Affiliation School of Computing and College of Medicine, University of South Alabama, Mobile, Alabama, United States of America

  • Xiuyu Leng

    Roles Supervision

    huang@southalabama.edu (JH); lengxy@mail.sysu.edu.cn (XL)

    Affiliations Department of Cardiology, The First Affiliated Hospital, Sun Yat-sen University, Guangzhou, China, NHC Key Laboratory of Assisted Circulation and Vascular Diseases (Sun Yat-sen University), Guangzhou, China

Correction

3 Nov 2025: Gai S, Huang F, Liu X, Benton RG, Borchert GM, et al. (2025) Correction: Assessing the feasibility of large language models to identify top research priorities in enhanced external counterpulsation. PLOS ONE 20(11): e0336055. https://doi.org/10.1371/journal.pone.0336055 View correction

Abstract

Enhanced External Counterpulsation (EECP), as a non-invasive, cost-effective, and efficient adjunctive circulatory technique, has been widely applied in in the cardiovascular field. Numerous studies and clinical observations have confirmed the obvious advantages of EECP in promoting blood flow perfusion to vital organs such as the heart, brain, and kidneys. However, many potential mechanisms of EECP remain insufficiently validated, necessitating researchers to dedicate substantial time and effort to in-depth investigations. In this work, large language models (such as ChatGPT and Ernie Bot) were used to identify top research priorities in five key topics in the field of EECP: mechanisms, device improvements, cardiovascular applications, neurological applications, and other applications. After generating specific research priorities in each domain through language models, a panel of nine experienced EECP experts was invited to independently evaluate and score them based on four parameters: relevance, originality, clarity, and specificity. Notably, high average and median scores for these evaluation parameters were obtained, indicating a strong endorsement from experts in the EECP field. This study preliminarily suggests that large language models like ChatGPT and Ernie Bot could serve as powerful tools for identifying and prioritizing research priorities in the EECP domain.

1. Introduction

Enhanced External Counterpulsation (EECP) is a non-invasive adjunctive circulatory technique that inflates and deflates cuffs wrapped around the limbs and buttocks in sync with the cardiac cycle under electrocardiographic gating control. EECP has been clinically demonstrated to significantly improve organ perfusion, regulate endothelial function, combat coronary artery atherosclerosis, treat complications of diabetes and sudden sensorineural hearing loss, among other benefits [13]. Although many evidences suggest there is a great deal of untapped potential for external counterpulsation, traditional approaches to identifying research priorities for EECP mainly rely on expert opinion and consensus building which are often labor-intensive and biased. In recent years, natural language processing (NLP) technology [4] has been increasingly recognized as a new means of identifying research priorities. Large language models (LLMs) such as ChatGPT [5] and Ernie Bot [6], which are trained on extensive text data, possess the ability to understand human-like language and have demonstrated significant potential in proposing and prioritizing research priorities [7]. In the medical domain, LLMs have shown promising results in various tasks, including disease diagnosis, medical record automation, literature retrieval, and patient education [8]. Adi Lahat et al. assessed the effectiveness of ChatGPT in generating research questions within gastroenterology and concluded that ChatGPT could be used to produce high-quality research inquiries [9]. Building on the recognition of NLP technology and the potential of large language models like ChatGPT and Ernie Bot to identify research priorities, their effectiveness in determining primary research priorities related to EECP technology were specifically evaluated in this work. Five key areas were examined: mechanisms, device enhancements, cardiovascular applications, neurological applications, and other applications. Utilizing ChatGPT and Ernie Bot, specific research priorities in these domains were generated, after which they were reviewed by experienced EECP experts and then rated to assess their relevance and importance.

2. Related work

Large language models have shown broad applicability in entertainment, education, and customer service, but their potential in the medical field remains largely untapped. Given the high standards for information quality and communication reliability in medicine, the application of large language models requires careful consideration. In recent years, scholars have begun to explore the use of large language models in medicine, yielding promising results. In the field of cardiology, Gala et al. [10] believed that LLMs can be utilized to analyze a large number of academy papers and medical record resources to help clinicians keep up with the latest advances in cardiology. Nevertheless, they also pointed to the limitations of LLMs in explaining cultural or emotional factors that may influence medical practice. Cascella et al. [11] explored the reasoning abilities of ChatGPT on public health topics. Through a question-and-answer session, ChatGPT listed four possible research topics. While some of the responses of ChatGPT may be stereotyped and depend on the prompts, it can be used to summarize the scientific literature and generate new research hypotheses. Additionally, George et al. [12] proposed that large language models could serve as a supplementary resource to traditional medical tools, improving the efficiency and productivity of medical practices. Unfortunately, these studies do not provide a quantitative assessment of the ability of LLMs to identify medical research priorities.

Importantly, in order to assess the effectiveness of LLMs in the medical domain, it is essential to conduct statistical analyses on numerical results obtained from experiments and/or surveys. In evaluating the pertinent literature on LLMs, Tang et al. [13] invited field experts to assess the summary quality of LLMs by using a five-point Likert scale along four dimensions: coherence, factual consistency, comprehensiveness, and harmfulness. Man-Whitney U test was used to assess the differences in response between GPT-3.5 and ChatGPT. Michael et al. [14] employed average scoring and fixed-effects consistency to calculate the Intraclass Correlation Coefficient (ICC), investigating the potential application of artificial intelligence-based LLMs in the realm of medical ethics. Similarly, Dave et al. [15] utilized Pearson and Spearman coefficients to juxtapose the assessment outcomes of large language models against the evaluations of medical professionals, thereby further substantiating their dependability. Furthermore, besides correlation analysis, similarity metrics are frequently utilized to gauge the efficacy of LLMs. For example, in 2024, Sebastian et al. [16] evaluated the pairwise accuracy between LLMs and human assessments by analyzing the cosine similarity matrix. In measuring factual knowledge within LLMs, Pezeshkpour [17] successfully utilized Kullback-Leibler (KL) divergence to analyze the predictive probability distributions of the model before and after instilling target knowledge. In investigating bias issues within large pre-trained language models, Guo et al. [18] used the Jensen-Shannon (JS) divergence to measure the consistency between different demographic distributions, offering a robust tool for reducing human-like biases and unwanted societal stereotypes. JS divergence is an improved version of KL divergence, whereas the KL scatter is asymmetric, making the JS scatter more accurate in identifying similarities.

3. Methods

3.1. Research priorities

ChatGPT (based on GPT-3.5) and Ernie Bot 3.5 to generate research priorities in five key topics (Tables 1 and 2, respectively) pertaining were leveraged to EECP mechanisms [1,19], structural enhancements, applications in cardiovascular domains [3,20,21], neurological applications [22, 23], and other applications [3,24,25].

thumbnail
Table 1. ChatGPT-generated research priorities on five key topics in the field of EECP research.

https://doi.org/10.1371/journal.pone.0305442.t001

thumbnail
Table 2. ERNIE bot-generated research priorities on five key topics in the field of EECP research.

https://doi.org/10.1371/journal.pone.0305442.t002

3.2. Expert evaluation

The expert evaluation panel was comprised of nine highly experienced EECP specialists as evidenced by panelists having authored an average of twenty relevant research publications in the field. They gained their expertise through clinical practice and made significant contributions to academic research, and experts have published at least five scholarly articles related to EECP. Furthermore, they have actively contributed to the development of guidelines in the EECP field. Panelists reviewed and assessed the inquiries presented by ChatGPT and Ernie Bot independently. Experts rated five priorities on four parameters (relevance, originality, clarity, and specificity) using a 1–5 scale with 5 representing the highest score. The a priori relationships generated by ChatGPT and Ernie Bot were then compared to current EECP research queries identified through a manual literature review. Importantly, in order to ensure the objectivity and relevance of responses, ChatGPT and Ernie Bot were instructed to treat each key topic as an independent query, thereby eliminating potential biases that may have existed in previous conversations.

4. Statistical analysis

Data were collected and analyzed using standard statistical methods, and all statistical analyses were conducted using IBM SPSS Statistics version 25 and Python 3.10. Initially, descriptive statistical methods were employed to provide a summary of the data, including measures such as mean, standard deviation (SD), and median. Afterwards, “divergence” was adopted to assess the similarity between ratings provided by experts in EECP and queries generated by two large language models. In the realm of data mining, JS divergence was computed to evaluate the similarity of ratings among evaluators using a rating table structured with evaluators as column attributes. JS divergence values from 0 to 1, with smaller values indicating greater similarity between ratings. Additionally, Spearman’s rank correlation coefficient and Kendall’s τ coefficient were also used to evaluate pairwise correlations between parameters. Positive coefficients indicate a positive correlation, while negative coefficients imply a negative correlation. The closer the coefficient is to 1 the stronger the correlation.

5. Results

The statistical analysis shows high reliability for the questionnaires assessing ChatGPT and Ernie Bot, with Cronbach’s alpha coefficients of 0.978 and 0.971, respectively. Both coefficients exceed the 0.8 threshold, indicating strong survey reliability. This suggests that the questionnaires effectively reflect the proficiency of ChatGPT and Ernie Bot in determining research priorities for EECP.

Based on this, the study conducted data analysis on the ratings provided by the 9 evaluators from three perspectives: (1) descriptive statistics; (2) similarity of ratings among evaluators; and (3) rank correlation of evaluation metrics. The data analysis tools utilized were IBM SPSS Statistics Version 25 and Python 3.10.

5.1 Descriptive statistics

Three score tables for each large language model were constructed, featuring evaluation metrics, evaluators, and topics as column attributes. For example, in the score table with five topics as column attributes, each column represents the scores from nine evaluators on four evaluation indicators for five research priorities within a specific topic. As shown in Tables 35, the results were derived from descriptive statistics applied to these three score tables. Since the mean and standard deviation have been commonly used to describe normal or approximately normal distributions, the quartiles in Tables 35 were considered to accurately reflect potential non-normal distributions. It is believed that the combination of mean/standard deviation and quartiles effectively reduces the impact of extreme values that may not fully represent the actual situation. From Table 3, it is clear that the two large language models excel in relevance, with originality following closely behind. In-depth descriptive statistical analyses of evaluation metrics are presented in Tables 35. The major models performed best in relevance, with originality close behind. Although originality exhibited the largest standard deviation, suggesting significant variation in expert opinions regarding originality, clarity demonstrated the smallest standard deviation, indicating minimal fluctuations in scores for each question. Additionally, variations in performance between the two models (ChatGPT and Ernie Bot) across different evaluation metrics and topics can be observed. Concerning relevance, Ernie Bot’s average score slightly exceeds ChatGPT’s, suggesting a slight advantage in addressing user-related questions, although this was not statistically significant. In terms of originality, ChatGPT’s score was slightly less than Ernie Bot’s, with a higher fluctuation in scoring standard deviation, indicating some disagreement among experts regarding the originality of ChatGPT’s queries. Both models demonstrate similar performance in clarity and specificity, indicating their similarity in providing clear and specific answers. Results of scores from EECP experts for all priorities are visually presented in Fig 1 with the outermost rings corresponding to the highest score of 5 and inner rings indicating lower scores.

thumbnail
Table 3. Descriptive statistics of evaluation metrics.

https://doi.org/10.1371/journal.pone.0305442.t003

thumbnail
Fig 1. Ratings of 25 research focal points by nine evaluators based on four criteria.

https://doi.org/10.1371/journal.pone.0305442.g001

thumbnail
Table 4. Descriptive statistics of evaluator.

https://doi.org/10.1371/journal.pone.0305442.t004

Table 4 presents the scores given by different raters for the ChatGPT and Ernie Bot models. The analysis shows that in the evaluations of most raters, ChatGPT and Ernie Bot have similar average scores indicating a certain level of competitiveness in overall performance. However, it is worth noting that in the ratings of Rater3 and Rater4, Ernie Bot’s average score was clearly higher than ChatGPT’s, reflecting a more outstanding performance of Ernie Bot from the perspectives of these two raters. In terms of score stability, there were differences between the two models among different raters. Specifically, in the evaluations of Rater3 and Rater4, Ernie Bot had a lower standard deviation, indicating more stable scores and consistent performance. Conversely, Rater8’s Ernie Bot scores demonstrated significantly higher standard deviation. In contrast, although overall score stability was slightly inferior to Ernie Bot’s performance for a subset of raters, ChatGPT’s standard deviation among multiple raters was relatively more consistent. These differences in evaluation may stem from personal preferences, evaluation criteria, and model performance across different topics.

In all topics (Table 5), Ernie Bot consistently received higher average scores than ChatGPT, suggesting a relative advantage in overall performance. Although their performances in terms of median scores were similar, Ernie Bot achieved an upper quartile score of 5.00 in specific topics such as mechanisms, device improvements and applications in neurology, indicating higher recognition in these areas. Meanwhile ChatGPT’s standard deviation across multiple topics was slightly lower than Ernie Bot’s, suggesting relatively better score stability. However, this difference was not significant. Notably, clear domain-specific differences were observed, while Ernie Bot’s average score significantly surpassed ChatGPT’s in structural improvements and applications in neurology domains, ChatGPT demonstrated superior performance in other domains.

5.2 Similarity of raters’ scores

Regarding the similarity of raters’ scores, the JS divergence of scores between each pair of raters for ChatGPT and Ernie Bot was calculated (Fig 2). The results indicate that the JS divergence range of scores for ChatGPT is [0, 0.102], while for Ernie Bot, it is [0, 0.148]. Since a smaller JS divergence value indicates higher similarity, it can be concluded that the evaluations of these two large language models by raters exhibit relatively high consistency. It is worth noting that, for both ChatGPT and Ernie Bot, the similarity of scores between rater 8 and other raters is the lowest. From Fig 1, it is evident that the scores given by rater 8 are significantly lower than those given by other raters. Further analysis of the data in Table 4 reveals that the average scores given by rater 8 for both ChatGPT and Ernie Bot are the lowest (2.20 and 2.44 respectively). Besides, they have the highest standard deviations (0.80 and 1.21 respectively). Excluding the influence of rater 8’s scores, the upper limit of the JS divergence of scores for ChatGPT would decrease from 0.102 to 0.052, and from 0.148 to 0.063 for Ernie Bot.

thumbnail
Fig 2. JS divergence heat map depicting the similarity of ratings between pairs of evaluators.

https://doi.org/10.1371/journal.pone.0305442.g002

5.3 Correlation of evaluation metrics

In terms of the correlation of evaluation metrics, we calculated both the Spearman [26] and Kendall [27] coefficients between pairs of evaluation metrics in the scoring results for ChatGPT and Ernie Bot (see Tables 6 and 7). These analyses passed significance tests, with all p-values below 0.01 indicating a significant positive correlation between relevance, originality, clarity, and specificity. This implies that when evaluating these two models, the score trends among these metrics were consistent, demonstrating high consistency and reliability. That said, ChatGPT exhibited a lower correlation between originality and relevance, while Ernie Bot showed a lower correlation in the analysis of specificity and relevance. The clarity of both models was highly correlated with relevance and/or specificity.

thumbnail
Table 6. Rank correlation coefficients between evaluation metrics (ChatGPT).

https://doi.org/10.1371/journal.pone.0305442.t006

thumbnail
Table 7. Rank correlation coefficients between evaluation metrics (Ernie Bot).

https://doi.org/10.1371/journal.pone.0305442.t007

6. Discussion

Here, the ability of ChatGPT and Ernie Bot was evaluated to generate research priorities in the field of EECP, covering mechanisms, structural improvements, applications in cardiology, applications in neurology, and applications in other fields. Both models demonstrated significant potential in consistently generating relevant and clear research priorities, which could offer valuable new tools for EECP research. Both scored relatively low in specificity, possibly due to limitations in handling domain-specific knowledge, indicating a need for improvement in accuracy and precision. To enhance their performance, fine-tuning with domain-specific data and expert knowledge will likely be required. While both models lacked originality in their responses, relying heavily on learned information and language patterns, future research should focus on enhancing their creativity to generate more unique research questions in the EECP field.

Notably, the performances of Ernie Bot and ChatGPT, two prominent language systems were compared. Ernie Bot demonstrated a slight but definitive advantage in terms of relevance, possibly due to its more precise semantic understanding and higher matching with user needs. In terms of originality, ChatGPT scored slightly lower with more fluctuation, indicating some disagreement among evaluators regarding its ability to offer novel and unique perspectives. This variance might stem from differences in the models’ performance across different contexts or from evaluators’ subjective criteria, such as their acceptance of research priorities that challenge existing cognitive frameworks or their willingness to explore unknown areas of study. In contrast, Ernie Bot received more consistent recognition for its originality, likely due to its more flexible and innovative thinking patterns. Regarding clarity and specificity, both models performed equally well, demonstrating high levels of proficiency. This suggests that they excel in providing clear, understandable responses and specific, detailed explanations, which are equally important for large language models as users often expect answers that are both clear and specific to better understand and apply the provided information.

From the evaluators’ perspective, most evaluators held similar views on the performance of the two models. However, in certain specific cases, such as Rater3 and Rater4, Ernie Bot received higher scores. Additionally, as compared to other raters, Rater8’s scores were significantly lower and deviated more substantially, and exclusion of Rater8 increased the performance of both models.

In certain specific topics such as mechanisms, applications in neurology, and cardiovascular applications, Ernie Bot performed better whereas ChatGPT’s performance slightly surpassed that of Ernie Bot in others, indicating that each model has its strengths and weaknesses in different domains and application scenarios.

Consequently, future research should be performed explore how to effectively integrate the strengths of both models to improve the performance and efficacy of large language models in real-world applications.

Our study applied ChatGPT and ERNIE Bot in the field of EECP to identify high-quality research priorities for the first time. It also offers a cross-disciplinary examination of the potential applications of EECP in neurology, metabolism, orthopedics, nephrology, and other areas. Furthermore, combining expert evaluations with statistical analysis enhances the scientific rigor and accuracy of our findings. This novel approach not only advances the development and refinement of EECP technology but also opens up new possibilities for patient treatment.

7. Limitations

Although this study presents promising outcomes, there are also some limitations in this study. Firstly, the expert panels involved may not fully represent the broader research community, which could have influenced the evaluation results. Secondly, the use of subjective ratings may introduce bias and variability in assessing the performance of ChatGPT and Ernie Bot. Lastly, the models may not have access to the latest biomedical literature, which could affect the quality of question generation. If this is the case, integrating domain-specific APIs with up-to-date information could enhance research quality. For future work, key directions include improving expert panel representation, optimizing large language models with more domain-specific training data, enhancing data transparency, applying more robust statistical methods, and fostering interdisciplinary collaboration. These efforts aim to address the identified limitations and promote innovation and advancement in EECP research.

8. Conclusion

Overall, this assessment of ChatGPT and Ernie Bot as generators of research priorities for Enhanced External Counterpulsation (mechanisms, device improvements, applications in cardiovascular medicine, applications in neurology, and applications in other non-cardiovascular and non-neurological fields) produced some promising results. Both models have demonstrated the capacity to generate high-quality research priorities in these areas, indicating their potential value as tools to drive research not only in EECP but also in broader medical fields through streamlining the process of identifying crucial research priorities and thereby save considerable time and effort. While there is room for improvement in terms of specificity and originality, both models have shown a capability to produce diverse, relevant, and coherent research priorities, likely aiding advancements in EECP research. Each model has its strengths in various domains and application scenarios, and further exploration could focus on leveraging these strengths to enhance the overall effectiveness of large language models in practical settings. In conclusion, our findings suggest that ChatGPT and Ernie Bot are poised to become valuable assistants for researchers in the EECP field and likely other medical domains, offering new momentum for scientific progress.

Supporting information

S1 File. Raw data and results - ERNIE Botts 2 and ChatGPT.

https://doi.org/10.1371/journal.pone.0305442.s001

(ZIP)

References

  1. 1. Zhang Y, He X, Chen X, Ma H, Liu D, Luo J, et al. Enhanced external counterpulsation inhibits intimal hyperplasia by modifying shear stress responsive gene expression in hypercholesterolemic pigs. Circulation. 2007;116(5):526–34. pmid:17620513
  2. 2. Lin S, Xiao-Ming W, Gui-Fu W. Expert consensus on the clinical application of enhanced external counterpulsation in elderly people (2019). Aging Med (Milton). 2020;3(1):16–24. pmid:32232188
  3. 3. Xu L, Chen X, Cui M, Ren C, Yu H, Gao W, et al. The improvement of the shear stress and oscillatory shear index of coronary arteries during Enhanced External Counterpulsation in patients with coronary heart disease. PLoS One. 2020;15(3):e0230144. pmid:32191730
  4. 4. Kang Y, Cai Z, Tan C-W, Huang Q, Liu H. Natural language processing (NLP) in management research: a literature review. J Manag Anal. 2020;7(2):139–72.
  5. 5. Open AI. ChatGPT: optimizing language models for dialogue; 2022. [Cited 2024 Mar 15]. Available from: https://openai.com/blog/chatgpt/
  6. 6. Ernie Bot. Baidu’s knowledge-enhanced large language model built on full AI stack technology; [Cited 2024 Mar 15]. Available from: https://yiyan.baidu.com/
  7. 7. Lahat A, Shachar E, Avidan B, Glicksberg B, Klang E. Evaluating the utility of a large language model in answering common patients’ gastrointestinal health-related questions: are we there yet? Diagnostics (Basel). 2023;13(11):1950. pmid:37296802
  8. 8. Ao G, Wang G, Chen Y, Xia J, Su S. Application research of large language models in medicine: status, problems, and future. In: 2023 13th International Conference on Information Technology in Medicine and Education. IEEE; 2023. p. 734–739. Available from:
  9. 9. Lahat A, Shachar E, Avidan B, Shatz Z, Glicksberg BS, Klang E. Evaluating the use of large language model in identifying top research questions in gastroenterology. Sci Rep. 2023;13(1):4164. pmid:36914821
  10. 10. Gala D, Makaryus AN. The utility of language models in cardiology: a narrative review of the benefits and concerns of ChatGPT-4. Int J Environ Res Public Health. 2023;20(15):6438. pmid:37568980
  11. 11. Cascella M, Montomoli J, Bellini V, Bignami E. Evaluating the feasibility of ChatGPT in healthcare: an analysis of multiple clinical and research scenarios. J Med Syst. 2023;47(1):33. pmid:36869927
  12. 12. George Pallivathukal R, Kyaw Soe HH, Donald PM, Samson RS, Hj Ismail AR. ChatGPT for academic purposes: survey among undergraduate healthcare students in Malaysia. Cureus. 2024;16(1):e53032. pmid:38410331
  13. 13. Tang L, Sun Z, Idnay B, Nestor JG, Soroush A, Elias PA, et al. Evaluating large language models on medical evidence summarization. NPJ Digit Med. 2023;6(1):158. pmid:37620423
  14. 14. Balas M, Wadden JJ, Hébert PC, Mathison E, Warren MD, Seavilleklein V, et al. Exploring the potential utility of AI large language models for medical ethics: an expert panel evaluation of GPT-4. J Med Ethics. 2024;50(2):90–6. pmid:37945336
  15. 15. Van Veen D, Van Uden C, Blankemeier L, Delbrouck J-B, Aali A, Bluethgen C, et al. Adapted large language models can outperform medical experts in clinical text summarization. Nat Med. 2024;30(4):1134–42. pmid:38413730
  16. 16. Joseph SA, Chen L, Trienes J, Göke HL, Coers M, Xu W, et al. FactPICO: factuality evaluation for plain language summarization of medical evidence. In: 62nd Annual Meeting of the Association for Computational Linguistics. Association for Computational Linguistics; 2024. p. 8437–8464. Available from:
  17. 17. Pezeshkpour P. Measuring and modifying factual knowledge in large language models. In: 2023 International Conference on Machine Learning and Applications. IEEE; 2023. p. 831–838. Available from:
  18. 18. Guo Y, Yang Y, Abbasi A. Auto-debias: debiasing masked language models with automated biased prompts. In: 60th Annual Meeting of the Association for Computational Linguistics. Association for Computational Linguistics; 2022. p. 1012–1023. Available from:
  19. 19. Yang D, Wu G. Vasculoprotective properties of enhanced external counterpulsation for coronary artery disease: beyond the hemodynamics. Int J Cardiol. 2013;166(1):38–43. pmid:22560950
  20. 20. Shea ML, Conti CR, Arora RR. An update on enhanced external counterpulsation. Clin Cardiol. 2005;28(3):115–8. pmid:15813616
  21. 21. Gorcsan J, Crawford L, Soran O, Wang H, Severyn D, de Lame P. Improvement in left ventricular performance by enhanced external counterpulsation in patients with heart failure. J Am Coll Cardiol. 2000;35(2):655.
  22. 22. Éneeva MA, Kostenko EV, Razumov AN, Petrova LV, Bobyreva SN, Nesuk OM. The enhanced external counterpulsation as a method of non-invasive auxiliary blood circulation used for the combined rehabilitative treatment of the patients surviving after ischemic stroke (a review). Vopr Kurortol Fizioter Lech Fiz Kult. 2015;92(3):45–52. pmid:26285334
  23. 23. Xiong L, Lin W, Han J, Chen X, Leung T, Soo Y, et al. Enhancing cerebral perfusion with external counterpulsation after ischaemic stroke: how long does it last?. J Neurol Neurosurg Psychiatry. 2016;87(5):531–6. pmid:25934015
  24. 24. Sardina PD, Martin JS, Avery JC, Braith RW. Enhanced external counterpulsation (EECP) improves biomarkers of glycemic control inpatients with non-insulin-dependent type II diabetes mellitus for up to 3 months following treatment. Acta Diabetol. 2016; 53(5): 745–752. pmid:27179825.
  25. 25. Froschermaier SE, Werner D, Leike S, Schneider M, Waltenberger J, Daniel WG, et al. Enhanced external counterpulsation as a new treatment modality for patients with erectile dysfunction. Urol Int. 1998;61(3):168–71. pmid:9933838
  26. 26. Spearman C. The proof and measurement of association between two things. Int J Epidemiol. 2010;39(5):1137–50. pmid:21051364
  27. 27. Kendall MG. A new measure of rank correlation. Biometrika. 1938;30(1–2):81–93.