Skip to main content
Advertisement
Browse Subject Areas
?

Click through the PLOS taxonomy to find articles in your field.

For more information about PLOS Subject Areas, click here.

  • Loading metrics

Racial bias in clinician assessment of patient credibility: Evidence from electronic health records

  • Mary Catherine Beach ,

    Roles Conceptualization, Data curation, Formal analysis, Funding acquisition, Supervision, Writing – original draft, Writing – review & editing

    mcbeach@jhmi.edu

    Affiliations Department of Medicine, School of Medicine, Johns Hopkins University, Baltimore, Maryland, United States of America, Center for Health Equity, Johns Hopkins University, Baltimore, Maryland, United States of America, Berman Institute of Bioethics, Johns Hopkins University, Baltimore, Maryland, United States of America

  • Keith Harrigian,

    Roles Formal analysis, Methodology, Software, Validation, Writing – review & editing

    Affiliation Department of Computer Science, Whiting School of Engineering, Johns Hopkins University, Baltimore, Maryland, United States of America

  • Brant Chee,

    Roles Data curation, Formal analysis, Software, Writing – review & editing

    Affiliation Applied Physics Lab, Johns Hopkins University, Baltimore, Maryland, United States of America

  • Alya Ahmad,

    Roles Data curation, Formal analysis, Methodology, Writing – review & editing

    Affiliation Department of Medicine, School of Medicine, Johns Hopkins University, Baltimore, Maryland, United States of America

  • Anne R. Links,

    Roles Formal analysis, Investigation, Writing – review & editing

    Affiliation Department of Medicine, School of Medicine, Johns Hopkins University, Baltimore, Maryland, United States of America

  • Ayah Zirikly,

    Roles Data curation, Investigation, Writing – review & editing

    Affiliation Department of Computer Science, Whiting School of Engineering, Johns Hopkins University, Baltimore, Maryland, United States of America

  • Dingfen Han,

    Roles Formal analysis, Validation, Visualization, Writing – review & editing

    Affiliation Department of Medicine, School of Medicine, Johns Hopkins University, Baltimore, Maryland, United States of America

  • Emily Boss,

    Roles Conceptualization, Investigation, Methodology, Writing – review & editing

    Affiliation Department of Otolaryngology, School of Medicine, Johns Hopkins University, Baltimore, Maryland, United States of America

  • Shari Lawson,

    Roles Conceptualization, Investigation, Supervision, Writing – review & editing

    Affiliation Department of Obstetrics and Gynecology, School of Medicine, Howard University, Washington, District of Columbia, United States of America

  • Mustapha Saheed,

    Roles Conceptualization, Investigation, Methodology, Writing – review & editing

    Affiliation Department of Emergency Medicine, School of Medicine, Johns Hopkins University, Baltimore, Maryland, United States of America

  • Yahan Li,

    Roles Investigation, Writing – review & editing

    Affiliation Department of Computer Science, Whiting School of Engineering, Johns Hopkins University, Baltimore, Maryland, United States of America

  • Mark Dredze,

    Roles Conceptualization, Methodology, Software, Supervision, Writing – review & editing

    Affiliation Department of Computer Science, Whiting School of Engineering, Johns Hopkins University, Baltimore, Maryland, United States of America

  • Somnath Saha

    Roles Funding acquisition, Methodology, Supervision, Writing – original draft, Writing – review & editing, Conceptualization

    Affiliations Department of Medicine, School of Medicine, Johns Hopkins University, Baltimore, Maryland, United States of America, Center for Health Equity, Johns Hopkins University, Baltimore, Maryland, United States of America

Abstract

Objective

Black patients disproportionately report feeling disbelieved or having concerns dismissed in medical encounters, suggesting potential racial bias in clinicians’ assessment of patient credibility. Because this bias may be evident in the language used by clinicians when writing notes about patients, we sought to assess racial differences in use of language either undermining or supporting patient credibility within the electronic health record (EHR).

Methods

We analyzed 13,065,081 notes written between 2016–2023 about 1,537,587 patients by 12,027 clinicians at a large health system with 5 hospitals and an extensive network of ambulatory practices in the mid-Atlantic region of the United States. We developed and applied natural language processing models to identify whether or not a note contained terms undermining or supporting patient credibility, and used logistic regression with generalized estimating equations to estimate the association of credibility language with patient race/ethnicity.

Results

The mean patient age was 43.3 years and 55.9% were female; 57.6% were non-Hispanic White, 28.0% non-Hispanic Black, 8.3% Hispanic/Latino, and 6.1% Asian. Clinician-authors were attending physicians (44.9%), physicians-in-training (40.1%) and advanced practice providers (15.0%). Terms specifically related to patient credibility were relatively uncommon, with 106,523 (0.82%) notes containing terms undermining patient credibility, and 33,706 (0.26%) supporting credibility. In adjusted analyses, notes written about non-Hispanic Black vs. White patients had higher odds of containing terms undermining credibility (aOR 1.29, 95% CI 1.27–1.32), and lower odds of supporting credibility (aOR 0.82; 95% CI 0.79–0.85). Notes written about Hispanic/Latino vs. White patients had similar odds of language undermining (aOR 0.99, 95% CI 0.95–1.03) and supporting credibility (aOR 0.95, 95% CI 0.89–1.02). Notes written about Asian vs. White patients had lower odds of language undermining credibility (aOR 0.85, 95% CI 0.81–0.89), and higher odds of supporting credibility (aOR 1.30, 95% CI 1.23–1.38).

Conclusions

Clinician documentation undermining patient credibility may disproportionately stigmatize Black individuals and favor Asian individuals. As stigmatizing language in medical records has been shown to negatively influence clinician attitudes and decision making, these racial differences in documentation may influence patient care and outcomes and exacerbate health inequities.

Introduction

There is mounting evidence that electronic health records (EHR) contain language reflecting the unconscious biases of clinicians, [17] and that this language may undermine the quality of care that patients receive.[8,9] Most of these studies have focused on stigmatizing language broadly, characterizing numerous ways that patients are negatively portrayed [1,2,10] and documenting racial disparities in the prevalence of negative or stigmatizing language [37].

One category of negative language that receives limited attention relates to assessments of patient credibility. Credibility assessments reflect a very specific type of bias, termed testimonial injustice, wherein a person (e.g., patient) may be less likely to be believed due to a prejudice on the part of the listener (e.g., clinician) [11]. Several qualitative studies suggest that Black Americans are more likely to have their concerns dismissed in healthcare settings [12,13], and two recent studies have documented this phenomenon by examining medical records. In our own earlier work [14], we found that notes from an academic ambulatory internal medicine clinic for patients identified as Black compared to White contained more markers of doubt in each of three linguistic categories: evidentials reporting the patient’s history as hearsay, patient quotes, and specific ‘doubt’ words. A recent study by Lee and colleagues [15] in an inpatient internal medicine setting similarly found greater use of evidentials in the notes of Black compared to White patients. While these early studies provide suggestive evidence of testimonial injustice in the EHR, they have been limited to the field of internal medicine and have used methods with limited specificity in identifying credibility-related language.

In the current study we sought to expand and strengthen the evidence related to bias in credibility assessments documented in the EHR in three important ways. First, we focused specifically on words used to undermine patient credibility using highly reliable natural language processing (NLP) models. Second, using these NLP models, we further distinguished different categories of terms undermining credibility, including those that cast doubt on patients’ sincerity as well as terms that undermine or support patients’ competence in providing accurate information (poor/good historian). Finally, we expanded beyond the field of internal medicine to include a wider range of clinical disciplines. Our overall goal was to assess possible racial differences in clinician use of language either supporting or undermining patient credibility within the EHR of patients from a large health system.

Methods

Study design, subjects, and setting

We conducted a cross-sectional analysis of notes written between 1/1/2016 and 10/30/2023 in four clinical fields – emergency medicine, internal medicine, obstetrics & gynecology (OB/GYN), and surgery – including subspecialties within each discipline – at a large academic health system, including a diverse group of 5 hospitals and hospital-based ambulatory practices as well as a network of 15 community-based practices across the mid-Atlantic region of the United States. Notes were included if they were: 1) written about a patient >=18 years of age identified in the EHR as non-Hispanic Black, Hispanic/Latino, non-Hispanic Asian, or non-Hispanic White; 2) written by an attending physician, a physician-in-training (medical students, residents, fellows), or an advanced practice provider (APP, including nurse practitioners and physician assistants) in one of the four clinical departments across the health system; and 3) in a note type containing free (vs. only templated) text (e.g., History & Physicals, Progress Notes, Discharge Summaries, and Consults). The study was reviewed and approved with a waiver of informed consent by the Johns Hopkins Institutional Review Board (IRB00363939). Because this retrospective review of medical records contained protected health information, all data were stored on an approved, secure, and compliant platform; only study team members had access. The data for this analysis were first accessed on February 14, 2024.

Identification of credibility language

We developed a conceptual framework for credibility language based on previous analyses of medical records [1,14] informed by the theory of epistemic injustice [11], wherein assessments of credibility are made based on either competence (whether or not the person is capable of relaying valuable information) or sincerity (whether or not a person is likely to be lying). We conceptualized assessments of patient competence to be made based on documentation of how the patient is characterized as a historian, with modifiers such as poor, unreliable, challenging, limited, inconsistent, etc. undermining competence, and modifiers such as excellent, very good, good, reliable, detailed, etc. supporting patient competence. Based on earlier work [14], language undermining credibility on the basis of sincerity included all forms of the word claims (e.g., “claiming that Tylenol does not work for her”), insists (e.g., “she insisted she took her medication”), adamant (e.g., “he adamantly denied smoking”), as well as any documentation of drug or narcotic seeking, malingering, or secondary gain. Although our previous work included two additional methods of casting doubt on patients’ testimonies (use of quotes and evidentials), we did not include those in this analysis because those linguistic features have lower specificity, meaning that they are often used in ways that do not explicitly express doubt.

Development of natural language processing models

We developed natural language processing (NLP) models to detect credibility language within clinical notes. Our NLP development process is described in detail elsewhere [16]. Briefly, we extracted over 10,000 excerpts from clinical notes from 2 different health systems that included stigmatizing terms, 2,337 of which included one or more of the credibility-related terms described above. The excerpts were reviewed by at least 2 reviewers and coded for inclusion or exclusion. For instance, “he insists he has stopped drinking” would be coded as related to credibility, whereas “he insists on being seen by a cardiologist” would be coded as unrelated, because it casts the patient as demanding rather than lacking credibility. Likewise, excerpts including “not a good historian” were coded as undermining credibility despite using the words “good historian.” Prior to full-scale annotation, we used a small sample of notes to resolve discrepancies and develop an annotation guide for each term in our dictionary. Inter-annotator agreement was high (Krippendorff’s α 0.87), and disagreements were resolved by consensus with a 3rd reviewer [16]. We then used this annotated dataset, employing a 70/20/10% split of the excerpts to train, refine, and validate, respectively, a machine learning classifier to correctly identify credibility-related statements. We used a Clinical BERT NLP model, [17] which had 92% overall accuracy and macro-F1 scores from 0.70 to 0.89.[16] Descriptions and code for our NLP models can be accessed at https://github.com/kharrigian/ehr-stigma.

Statistical analysis

We created a dichotomous variable indicating whether there was one or more instances of language within each category in each note. We used descriptive analyses to characterize our study sample, and then t-tests and χ2 tests to examine the association of patient race/ethnicity with other patient demographic characteristics (age, gender, insurance, interpreter use) and variables related to clinical context (diagnosis of substance use disorder [SUD] [18] or severe mental illness [SMI] [19,20], clinical department, and clinician role/training level). We adjusted for SUD and SMI to account for potential confounding by these conditions, as both were associated with patient race and with use of terms questioning credibility. ICD codes defining SUD and SMI are included in the S1 Appendix. We conducted logistic regression analyses to estimate the odds of each type of language based on patient race/ethnicity, using generalized estimating equations to account for clustering of notes within patient-clinician dyads. We used patient-clinician dyads to account for clustering of notes within patients and clinicians due to the non-hierarchical nature of our data; notes for a given patient could have been written by multiple clinicians, making hierarchical analytic methods (notes within patients and patients within clinicians) inappropriate. We first executed unadjusted models, then added patient demographic characteristics to the model, and finally added clinical context variables. Data were complete for all variables except insurance (0.5% missing) and interpreter use (0.002% missing); multivariate analyses excluded these few notes with missing data. Finally, after examining patterns uncovered by the primary analysis, we conducted separate exploratory models examining the independent association of each word group used to undermine credibility (adamant, claims, insists, malingering, narcotic/drug seeking, secondary gain, poor historian), comparing the notes of Black vs. non-Black patients. All analyses were conducted using Stata (Version 18.0, StataCorp. 2023. College Station, TX, 77845 USA).

Results

Study sample

There were 13,065,081 notes written about 1,537,587 patients by 12,027 clinicians. Clinician-authors were attending physicians (44.9%), physicians in training (40.1%) and advanced practice providers (15.0%). Characteristics of patients and notes are shown in Table 1. The mean patient age was 43.3 years and 56% of patients were female. More than half (57.6%) were non-Hispanic White, 28.0% were non-Hispanic Black, 8.3% Hispanic/Latino, and 6.1% Asian. Most patients had commercial insurance (58.7%); 19.2% had Medicare, 14.9% had Medicaid and 7.2% were uninsured. Most notes were written for encounters in the Internal Medicine or Surgery departments (46.4 and 30.1% respectively), and most were written by attending physicians (65.7%). There were significant differences by patient race/ethnicity for all demographic and clinical variables, as outlined in Table 1.

Language undermining patient credibility

Fewer than 1% of notes contained language undermining patient credibility (n = 106,523; 0.82%), with 62,480 (0.48%) containing language undermining patients’ sincerity and 52,243 (0.40%) containing language undermining competence.

In adjusted analyses (Table 2), notes written about non-Hispanic Black vs. White patients had overall higher odds of containing terms undermining credibility (aOR 1.29, 95% CI 1.27–1.32). This disparity was observed for terms undermining both patients’ sincerity (aOR 1.16; 95% CI 1.14–1.19) and patients’ competence (aOR 1.50; 95% 1.47–1.54). In terms of differences in specific words used in notes of Black vs. non-Black patients (Fig 1), there was a higher odds of each individual word undermining patient credibility, except for the terms related to narcotic or drug-seeking, which had a lower odds of appearing in the notes of Black vs. non-Black patients.

thumbnail
Table 2. Racial/ethnic differences in language undermining and supporting credibility1 in 13,681,829 notes across 4 clinical departments.

https://doi.org/10.1371/journal.pone.0328134.t002

thumbnail
Fig 1. Differences in specific terms undermining credibility for black vs. non-black patients.

https://doi.org/10.1371/journal.pone.0328134.g001

In fully adjusted models, there was no difference between Hispanic/Latino vs. White patients in language undermining patient credibility overall (aOR 0.99, 95%CI 0.95–1.03). When examining terms related to sincerity and competence separately, our findings were bidirectional. Notes written about Hispanic/Latino vs. White patients had lower odds of undermining sincerity (aOR 0.88, 95%CI 0.83–0.93). In unadjusted analyses, notes written about Hispanic/Latino vs. White patients also had lower odds of undermining competence, but this relationship was reversed in the fully adjusted model (aOR 1.20; 95% CI 1.13–1.27).

Notes written about Asian vs. White patients had lower odds of having language undermining credibility overall (aOR 0.85, 95% CI 0.81–0.89). When examining terms related to sincerity and competence separately, in the fully adjusted models notes written about Asian vs. White patients had lower odds of language undermining sincerity (aOR 0.74, 95% CI 0.69–0.79) but did not differ in the prevalence of language undermining competence (aOR 0.95, 95% CI 0.89–1.01).

Language supporting patient credibility

Language supporting patients’ credibility was found in 33,706 notes (0.26%). Notes written about Black compared to White patients had lower odds of language supporting credibility (aOR 0.82; 95% CI 0.79–0.85), whereas notes written about Asian vs. White patients had higher odds of supporting credibility (aOR 1.30, 95% CI 1.23–1.38). There was no difference in language supporting patient credibility in notes written about Hispanic/Latino vs. White patients.

Discussion

We found evidence of racial bias in clinician assessments of patient credibility, based on the language they used in documenting clinical encounters. Specifically, after accounting for multiple potential confounders, we found that clinicians were more likely to insinuate in the EHR that Black compared to White patients were not truthful and not competent in reporting their own experiences. This testimonial injustice can harm Black individuals in numerous ways. Having one’s concerns dismissed has an immediate negative impact on the individual’s experience of respect, which erodes trust and decreases likelihood of engaging in future care or following treatment recommendations [1214]. Further, failure to attend seriously to patients’ own reported symptoms may lead to physical harms such as untreated pain, delayed diagnoses, incorrect treatments, serious medical errors, and even death [2125]. Because notes with biased language can affect the attitudes and decision-making of subsequent clinicians reading those notes [8], language undermining Black patients’ credibility may contribute to healthcare inequities.

We might question, in light of the low prevalence of credibility language detected by our models, whether the racial differences we observed, even if statistically significant as a result of our large sample size, are clinically important. We believe the language detected by our NLP models, which relied upon a limited set of specific words that explicitly cast doubt on the patient’s testimony, represents the tip of an iceberg in terms of how words are used to undermine patients’ credibility, and that our findings accordingly signal larger underlying disparities in credibility assessments. In our own qualitative work examining medical records [1], we have found doubt conveyed by clinicians in the medical record in more nuanced and implicit ways, such as placing the patient’s words in quotes (e.g., he states his pain is “still a 10”), juxtaposing the patient’s subjective report with their own objective observation (e.g., reports adherence to meds but doesn’t know Lasix dosing), or using a greater number of “evidential” terms that treat patient-reported information as “hearsay,” allowing the clinician to be noncommittal about believing that information (e.g., she says she has a headache vs. she has a headache). Our own and another prior study found higher use of quotes [14] and evidentials in the notes of Black compared to White patients [14,15], adding to evidence of a credibility deficit experienced by Black patients [26].

It has been suggested that the term ‘poor historian’ (or any of its variants) may be used by clinicians who themselves failed to obtain a good history [2731]. Thus, disproportionate use of this term in the notes of Black patients may reflect, in addition to credibility bias, ineffectiveness in communicating with Black patients [12,13,32,33]. Importantly, use of the term ‘poor historian’ has the potential to bias future clinicians, who may discount the patient’s symptoms or be less likely to solicit the patient’s story. There are times when a clinician may reasonably be concerned that they weren’t able to gather the information they needed because the patient was not able to communicate it clearly. In those instances, we suggest specifying that to be the case – “patient unable to provide a complete history” or “patient is uncertain of some details” – that would convey less blame or avoid a global discrediting label.

One could argue that it may also be dangerous to not question the accuracy of a patient’s history when that patient is not being truthful, is misremembering, or has a recognized mental health disorder such as malingering. In these cases, it is not our intent to prohibit or criticize the conveying of such doubt in a medical record. We cannot and do not intend to assert that any individual clinician is biased against any individual patient simply because they use one of these words. Our primary concern here is that the observed racial differences in the use of these words suggest an implicit racial bias, on a population-level, in assessments of credibility. It is this bias to which we encourage clinicians to become attentive, both in terms of their own assessments of patient credibility as well as in interpretation of what has been written in the record by other clinicians. When examining specific credibility-related terms, we found that, after adjusting for demographic and clinical characteristics including substance use disorder and mental illness, all terms were used more often among Black vs non-Black patients, except for those related to narcotic or drug seeking. While we cannot discern from our data the reasons underlying this finding, we suspect clinicians may avoid using those terms for Black patients out of awareness of the racial stereotypes they perpetuate. If so, raising clinicians’ awareness of the biases and stereotypes that likely underlie the disproportionate use of other terms undermining Black patients’ credibility may help to remedy racial inequities in credibility assessments and credibility-related documentation.

Although we found consistent evidence of credibility bias against Black compared to White patients, we did not find similar evidence for Hispanic/Latino or Asian patients. Indeed, we found the opposite, with fewer terms questioning sincerity among Hispanic/Latino or Asian patients compared to Whites. While our data do not provide clear explanations for these findings, it may be that the sociocultural distance between clinicians and patients perceived as “foreign” [34] result in less in-depth, more superficial interactions and therefore fewer opportunities to suspect patients of insincerity. Similarly, Asian patients’ being more likely than White patients to be explicitly called out as “good historians” may reflect a lower expectation for accurate history provision among patients perceived as foreign. This represents a form of “reverse linguistic stereotyping,” in which a person’s race, ethnicity, or appearance influences expectations and perceptions of that person’s communication and may lead a clinician to praise articulate histories in the face of anticipated language or cultural barriers.[35] “Poor historian” terminology had a lower prevalence in unadjusted analyses for Hispanic/Latino patients; however this association was inverted after adjustment for demographics (primarily patient age), which may reflect greater language barriers among older Hispanic patients.[36] Further studies exploring clinician perceptions of patient credibility by race and ethnicity could test these hypothetical explanations. Based on our findings, it appears that credibility bias is worst for Black Americans.

Our study has several limitations. We studied records from one health system, although the system does expand across the mid-Atlantic region and includes a broad array of clinical settings (5 different hospitals, both community-based and academic, as well as a network of ambulatory practices) and disciplines. Our health system includes a higher proportion of Black patients than the national average, and we do not know whether disparities might be larger or smaller in other settings. We did not have data on clinicians beyond their role/training and clinical department and could not examine the influence of other clinician characteristics (e.g., age, race, gender). We also did not analyze nursing notes because, in our health system, they are mostly templated. These are important areas for future study. We selected a limited set of terms representing credibility language; as noted above, our findings may therefore underestimate its true prevalence [14]. We used NLP models with high, but not perfect, accuracy in detecting credibility-related language [16]; these models may have therefore misclassified some notes and thereby under- or overestimated the prevalence of credibility-related language. Any resulting misclassification, however, was likely to be non-differential; i.e., it is unlikely that the NLP models were less accurate in classifying credibility-related terms for one racial/ethnic group vs another. The impact of such misclassification, therefore, would be to bias toward more conservative (less pronounced) estimates of racial disparities. In addition, we did not capture evidence for supporting credibility in the ‘sincerity’ category because, while clinicians do often document that a patient is a good historian, there is no parallel documentation practice for complimenting the patient on being truthful. Finally, we had a relatively small proportion of notes from patients who were not identified as Black or White, and documentation of demographic data in the medical record is not always accurate.

In conclusion, clinician documentation undermining patient credibility may disproportionately stigmatize Black individuals and favor Asian individuals. As stigmatizing language in medical records has been shown to negatively influence clinician attitudes and decision-making [8,9], these racial differences in documentation may negatively influence quality of care and outcomes and further exacerbate health inequities. This specific type of bias related to credibility assessment has been relatively overlooked and has therefore remained unchecked. Medical educators have taken up efforts to train future physicians to engage in more respectful, patient-centered documentation, [37,38] and to learn about the structural inequities and interpersonal attitudes [39] that underlie biased language use in the EHR. We recommend that these curricula include content on testimonial injustice and suggest eliminating phrases such as ‘poor historian,’ as well as veiled insinuations about patient insincerity. As artificial intelligence applications are increasingly adopted to aid in note writing, consideration should be given to programming those applications to also avoid such language. Finally, we recommend that health professionals continue to confront the racial biases underlying differential assessments of patient credibility in the first place.

Supporting information

S1 Appendix. ICD-10 Codes for Substance Use Disorders and Severe Mental Illness.

https://doi.org/10.1371/journal.pone.0328134.s001

(DOCX)

References

  1. 1. Park J, Saha S, Chee B, Taylor J, Beach MC. Physician Use of Stigmatizing Language in Patient Medical Records. JAMA Netw Open. 2021;4(7):e2117052. pmid:34259849
  2. 2. Barcelona V, Scharp D, Idnay BR, Moen H, Goffman D, Cato K, et al. A qualitative analysis of stigmatizing language in birth admission clinical notes. Nurs Inq. 2023;30(3):e12557. pmid:37073504
  3. 3. Topaz M, Song J, Davoudi A, McDonald M, Taylor J, Sittig S, et al. Home Health Care Clinicians’ Use of Judgment Language for Black and Hispanic Patients: Natural Language Processing Study. JMIR Nurs. 2023;6:e42552. pmid:37067893
  4. 4. Sun M, Oliwa T, Peek ME, Tung EL. Negative Patient Descriptors: Documenting Racial Bias In The Electronic Health Record. Health Affairs. 2022;41(2):203–11.
  5. 5. Himmelstein G, Bates D, Zhou L. Examination of Stigmatizing Language in the Electronic Health Record. JAMA Netw Open. 2022;5(1):e2144967. pmid:35084481
  6. 6. Cobert J, Mills H, Lee A, Gologorskaya O, Espejo E, Jeon SY, et al. Measuring Implicit Bias in ICU Notes Using Word-Embedding Neural Network Models. Chest. 2024;165(6):1481–90. pmid:38199323
  7. 7. Wesevich A, Patel-Nguyen S, Fridman I, Langan E, Parente V. Patient Factors Associated With Biased Language in Nightly Resident Verbal Handoff. JAMA Pediatr. 2023;177(10):1098–100. pmid:37578802
  8. 8. P Goddu A, O’Conor KJ, Lanzkron S, Saheed MO, Saha S, Peek ME, et al. Do Words Matter? Stigmatizing Language and the Transmission of Bias in the Medical Record. J Gen Intern Med. 2018;33(5):685–91. pmid:29374357
  9. 9. Carpenter JE, Catalanotti J, Notis M, Brokus C, Moran TP, Akselrod H, et al. Use of nonstigmatizing language is associated with improved outcomes in hospitalized people who inject drugs. J Hosp Med. 2023;18(8):670–6. pmid:37286190
  10. 10. Cox C, Fritz Z. Presenting complaint: use of language that disempowers patients. BMJ. 2022;377:e066720. pmid:35477529
  11. 11. Fricker M. Epistemic injustice: power and the ethics of knowing. Oxford; New York: Oxford University Press. 2007.
  12. 12. Beach MC, Branyon E, Saha S. Diverse patient perspectives on respect in healthcare: A qualitative study. Patient Educ Couns. 2017;100(11):2076–80. pmid:28602565
  13. 13. Agarwal AK, Gonzales RE, Sagan C, Nijim S, Asch DA, Merchant RM, et al. Perspectives of Black Patients on Racism Within Emergency Care. JAMA Health Forum. 2024;5(3):e240046. pmid:38457129
  14. 14. Beach MC, Saha S, Park J, Taylor J, Drew P, Plank E, et al. Testimonial Injustice: Linguistic Bias in the Medical Records of Black Patients and Women. J Gen Intern Med. 2021;36(6):1708–14. pmid:33754318
  15. 15. Lee CR, Aysola J, Chen X, Addisu E, Klein A, Weissenbacher D, et al. Race and Ethnicity and Clinician Linguistic Expressions of Doubt in Hospital Admission Notes. JAMA Netw Open. 2024;7(10):e2438550. pmid:39401039
  16. 16. Harrigian K, et al. Characterization of stigmatizing language in medical records. PhysioNet. 2023.
  17. 17. Alsentzer E, Murphy J, Boag W, Weng W-H, Jindi D, Naumann T, et al. Publicly Available Clinical. In: Proceedings of the 2nd Clinical Natural Language Processing Workshop. Minneapolis, Minnesota, USA: Association for Computational Linguistics. 2019.
  18. 18. Wang L, Homayra F, Pearce LA, Panagiotoglou D, McKendry R, Barrios R, et al. Identifying mental health and substance use disorders using emergency department and hospital records: a population-based retrospective cohort study of diagnostic concordance and disease attribution. BMJ Open. 2019;9(7):e030530. pmid:31300509
  19. 19. Mansour H, Mueller C, Davis KAS, Burton A, Shetty H, Hotopf M, et al. Severe mental illness diagnosis in English general hospitals 2006-2017: A registry linkage study. PLoS Med. 2020;17(9):e1003306. pmid:32941435
  20. 20. Warren N, Leske S, Arnautovska U, Northwood K, Kisely S, Siskind D. Prevalence of frailty in severe mental illness: findings from the UK Biobank. BJPsych Open. 2023;9(6):e185. pmid:37821357
  21. 21. Villarosa L. Why America’s Black mothers and babies are in a life-or death crisis. New York Times. 2018.
  22. 22. N. Kirtner, “A mother speaks for her daughter: Doctors, listen to your patients,” Seattle Times. Mar. 10, 2017.
  23. 23. Cumming E. ‘Our child died because doctors didn’t listen’ – Martha’s rule will ensure patients are heard. The Guardian. 2023.
  24. 24. Eligon J. Black doctor dies of covid-19 after complaining of racist treatment. New York Times. 2020.
  25. 25. Haywood C Jr, Beach MC, Lanzkron S, Strouse JJ, Wilson R, Park H, et al. A systematic review of barriers and interventions to improve appropriate use of therapies for sickle cell disease. J Natl Med Assoc. 2009;101(10):1022–33. pmid:19860302
  26. 26. Beach MC, Saha S. The Credibility Deficit Experienced by Black Patients. JAMA Netw Open. 2024;7(10):e2438521. pmid:39401043
  27. 27. Green AR, Nze C. Language-Based Inequity in Health Care: Who Is the “Poor Historian”?. AMA J Ethics. 2017;19(3):263–71. pmid:28323607
  28. 28. Tiemstra J. The poor historian. Acad Med. 2009;84(6):723. pmid:19474545
  29. 29. Lajeunesse M. Language that disempowers patients: the doctor is the poor historian. BMJ. 2022;377:o1296. pmid:35618269
  30. 30. Block MR, Coulehan JL. Teaching the difficult interview in a required course on medical interviewing. J Med Educ. 1987;62(1):35–40. pmid:3795241
  31. 31. Coulehan JL. Who Is the Poor Historian?. JAMA. 1984;252(2):221.
  32. 32. Beach MC, Saha S, Korthuis PT, Sharp V, Cohn J, Wilson IB, et al. Patient-provider communication differs for black compared to white HIV-infected patients. AIDS Behav. 2011;15(4):805–11. pmid:20066486
  33. 33. Park J, Beach MC, Han D, Moore RD, Korthuis PT, Saha S. Racial disparities in clinician responses to patient emotions. Patient Educ Couns. 2020;103(9):1736–44. pmid:32253063
  34. 34. Zou LX, Cheryan S. Two axes of subordination: A new model of racial position. J Pers Soc Psychol. 2017;112(5):696–717. pmid:28240941
  35. 35. Kang O, Rubin DL. Reverse Linguistic Stereotyping: Measuring the Effect of Listener Expectations on Speech Evaluation. J Language Social Psychol. 2009;28(4):441–56.
  36. 36. Velasco PFT, Lopez MH, Martínez J, Gabriel. When labels don’t fit: Hispanics and their views of identity. https://www.pewresearch.org/race-and-ethnicity/2012/04/04/when-labels-dont-fit-hispanics-and-their-views-of-identity/. Accessed 2025 March 21.
  37. 37. Eng K, Johnston K, Cerda I, Kadakia K, Mosier-Mills A, Vanka A. A Patient-Centered Documentation Skills Curriculum for Preclerkship Medical Students in an Open Notes Era. MedEdPORTAL. 2024;20:11392. pmid:38533390
  38. 38. Lam BD, Dupee D, Gerard M, Bell SK. A patient-centered approach to writing ambulatory visit notes in the cures act era. Appl Clin Inform. 2023;14(1):199–204. pmid:36889340
  39. 39. Weiner S. Medical schools overhaul curricula to fight inequities. AAMC News. 2021. https://www.aamc.org/news/medical-schools-overhaul-curricula-fight-inequities