Measuring Implicit Bias in ICU Notes Using Word-Embedding Neural Network Models
Language in nonmedical data sets is known to transmit human-like biases when used in natural language processing (NLP) algorithms that can reinforce disparities. It is unclear if NLP algorithms of medical notes could lead to similar transmissions of biases. Can we identify implicit bias in clinical...
Saved in:
Published in: | Chest Vol. 165; no. 6; pp. 1481 - 1490 |
---|---|
Main Authors: | , , , , , , , , , , , , , |
Format: | Journal Article |
Language: | English |
Published: |
United States
Elsevier Inc
01-06-2024
American College of Chest Physicians |
Subjects: | |
Online Access: | Get full text |
Tags: |
Add Tag
No Tags, Be the first to tag this record!
|
Abstract | Language in nonmedical data sets is known to transmit human-like biases when used in natural language processing (NLP) algorithms that can reinforce disparities. It is unclear if NLP algorithms of medical notes could lead to similar transmissions of biases.
Can we identify implicit bias in clinical notes, and are biases stable across time and geography?
To determine whether different racial and ethnic descriptors are similar contextually to stigmatizing language in ICU notes and whether these relationships are stable across time and geography, we identified notes on critically ill adults admitted to the University of California, San Francisco (UCSF), from 2012 through 2022 and to Beth Israel Deaconess Hospital (BIDMC) from 2001 through 2012. Because word meaning is derived largely from context, we trained unsupervised word-embedding algorithms to measure the similarity (cosine similarity) quantitatively of the context between a racial or ethnic descriptor (eg, African-American) and a stigmatizing target word (eg, nonco-operative) or group of words (violence, passivity, noncompliance, nonadherence).
In UCSF notes, Black descriptors were less likely to be similar contextually to violent words compared with White descriptors. Contrastingly, in BIDMC notes, Black descriptors were more likely to be similar contextually to violent words compared with White descriptors. The UCSF data set also showed that Black descriptors were more similar contextually to passivity and noncompliance words compared with Latinx descriptors.
Implicit bias is identifiable in ICU notes. Racial and ethnic group descriptors carry different contextual relationships to stigmatizing words, depending on when and where notes were written. Because NLP models seem able to transmit implicit bias from training data, use of NLP algorithms in clinical prediction could reinforce disparities. Active debiasing strategies may be necessary to achieve algorithmic fairness when using language models in clinical research.
[Display omitted] |
---|---|
AbstractList | Language in nonmedical data sets is known to transmit human-like biases when used in natural language processing (NLP) algorithms that can reinforce disparities. It is unclear if NLP algorithms of medical notes could lead to similar transmissions of biases.
Can we identify implicit bias in clinical notes, and are biases stable across time and geography?
To determine whether different racial and ethnic descriptors are similar contextually to stigmatizing language in ICU notes and whether these relationships are stable across time and geography, we identified notes on critically ill adults admitted to the University of California, San Francisco (UCSF), from 2012 through 2022 and to Beth Israel Deaconess Hospital (BIDMC) from 2001 through 2012. Because word meaning is derived largely from context, we trained unsupervised word-embedding algorithms to measure the similarity (cosine similarity) quantitatively of the context between a racial or ethnic descriptor (eg, African-American) and a stigmatizing target word (eg, nonco-operative) or group of words (violence, passivity, noncompliance, nonadherence).
In UCSF notes, Black descriptors were less likely to be similar contextually to violent words compared with White descriptors. Contrastingly, in BIDMC notes, Black descriptors were more likely to be similar contextually to violent words compared with White descriptors. The UCSF data set also showed that Black descriptors were more similar contextually to passivity and noncompliance words compared with Latinx descriptors.
Implicit bias is identifiable in ICU notes. Racial and ethnic group descriptors carry different contextual relationships to stigmatizing words, depending on when and where notes were written. Because NLP models seem able to transmit implicit bias from training data, use of NLP algorithms in clinical prediction could reinforce disparities. Active debiasing strategies may be necessary to achieve algorithmic fairness when using language models in clinical research. Language in nonmedical data sets is known to transmit human-like biases when used in natural language processing (NLP) algorithms that can reinforce disparities. It is unclear if NLP algorithms of medical notes could lead to similar transmissions of biases.BACKGROUNDLanguage in nonmedical data sets is known to transmit human-like biases when used in natural language processing (NLP) algorithms that can reinforce disparities. It is unclear if NLP algorithms of medical notes could lead to similar transmissions of biases.Can we identify implicit bias in clinical notes, and are biases stable across time and geography?RESEARCH QUESTIONCan we identify implicit bias in clinical notes, and are biases stable across time and geography?To determine whether different racial and ethnic descriptors are similar contextually to stigmatizing language in ICU notes and whether these relationships are stable across time and geography, we identified notes on critically ill adults admitted to the University of California, San Francisco (UCSF), from 2012 through 2022 and to Beth Israel Deaconess Hospital (BIDMC) from 2001 through 2012. Because word meaning is derived largely from context, we trained unsupervised word-embedding algorithms to measure the similarity (cosine similarity) quantitatively of the context between a racial or ethnic descriptor (eg, African-American) and a stigmatizing target word (eg, nonco-operative) or group of words (violence, passivity, noncompliance, nonadherence).STUDY DESIGN AND METHODSTo determine whether different racial and ethnic descriptors are similar contextually to stigmatizing language in ICU notes and whether these relationships are stable across time and geography, we identified notes on critically ill adults admitted to the University of California, San Francisco (UCSF), from 2012 through 2022 and to Beth Israel Deaconess Hospital (BIDMC) from 2001 through 2012. Because word meaning is derived largely from context, we trained unsupervised word-embedding algorithms to measure the similarity (cosine similarity) quantitatively of the context between a racial or ethnic descriptor (eg, African-American) and a stigmatizing target word (eg, nonco-operative) or group of words (violence, passivity, noncompliance, nonadherence).In UCSF notes, Black descriptors were less likely to be similar contextually to violent words compared with White descriptors. Contrastingly, in BIDMC notes, Black descriptors were more likely to be similar contextually to violent words compared with White descriptors. The UCSF data set also showed that Black descriptors were more similar contextually to passivity and noncompliance words compared with Latinx descriptors.RESULTSIn UCSF notes, Black descriptors were less likely to be similar contextually to violent words compared with White descriptors. Contrastingly, in BIDMC notes, Black descriptors were more likely to be similar contextually to violent words compared with White descriptors. The UCSF data set also showed that Black descriptors were more similar contextually to passivity and noncompliance words compared with Latinx descriptors.Implicit bias is identifiable in ICU notes. Racial and ethnic group descriptors carry different contextual relationships to stigmatizing words, depending on when and where notes were written. Because NLP models seem able to transmit implicit bias from training data, use of NLP algorithms in clinical prediction could reinforce disparities. Active debiasing strategies may be necessary to achieve algorithmic fairness when using language models in clinical research.INTERPRETATIONImplicit bias is identifiable in ICU notes. Racial and ethnic group descriptors carry different contextual relationships to stigmatizing words, depending on when and where notes were written. Because NLP models seem able to transmit implicit bias from training data, use of NLP algorithms in clinical prediction could reinforce disparities. Active debiasing strategies may be necessary to achieve algorithmic fairness when using language models in clinical research. Language in nonmedical data sets is known to transmit human-like biases when used in natural language processing (NLP) algorithms that can reinforce disparities. It is unclear if NLP algorithms of medical notes could lead to similar transmissions of biases. Can we identify implicit bias in clinical notes, and are biases stable across time and geography? To determine whether different racial and ethnic descriptors are similar contextually to stigmatizing language in ICU notes and whether these relationships are stable across time and geography, we identified notes on critically ill adults admitted to the University of California, San Francisco (UCSF), from 2012 through 2022 and to Beth Israel Deaconess Hospital (BIDMC) from 2001 through 2012. Because word meaning is derived largely from context, we trained unsupervised word-embedding algorithms to measure the similarity (cosine similarity) quantitatively of the context between a racial or ethnic descriptor (eg, African-American) and a stigmatizing target word (eg, nonco-operative) or group of words (violence, passivity, noncompliance, nonadherence). In UCSF notes, Black descriptors were less likely to be similar contextually to violent words compared with White descriptors. Contrastingly, in BIDMC notes, Black descriptors were more likely to be similar contextually to violent words compared with White descriptors. The UCSF data set also showed that Black descriptors were more similar contextually to passivity and noncompliance words compared with Latinx descriptors. Implicit bias is identifiable in ICU notes. Racial and ethnic group descriptors carry different contextual relationships to stigmatizing words, depending on when and where notes were written. Because NLP models seem able to transmit implicit bias from training data, use of NLP algorithms in clinical prediction could reinforce disparities. Active debiasing strategies may be necessary to achieve algorithmic fairness when using language models in clinical research. [Display omitted] |
Author | Ashana, Deepshikha C. Heintz, Timothy A. Cobert, Julien Chapman, Allyson Cook Raghunathan, Karthik Kennedy, Christopher J. Lee, Albert Boscardin, W. John Lee, Sei J. Gologorskaya, Oksana Smith, Alex K. Mills, Hunter Jeon, Sun Young Espejo, Edie |
Author_xml | – sequence: 1 givenname: Julien orcidid: 0000-0001-9971-7535 surname: Cobert fullname: Cobert, Julien email: Julien.cobert@ucsf.edu organization: Anesthesia Service, San Francisco VA Health Care System, University of California, San Francisco, San Francisco, CA – sequence: 2 givenname: Hunter surname: Mills fullname: Mills, Hunter organization: Bakar Computational Health Sciences Institute, University of California, San Francisco, San Francisco, CA – sequence: 3 givenname: Albert orcidid: 0000-0002-3821-4376 surname: Lee fullname: Lee, Albert organization: Bakar Computational Health Sciences Institute, University of California, San Francisco, San Francisco, CA – sequence: 4 givenname: Oksana orcidid: 0009-0005-3505-7300 surname: Gologorskaya fullname: Gologorskaya, Oksana organization: Bakar Computational Health Sciences Institute, University of California, San Francisco, San Francisco, CA – sequence: 5 givenname: Edie orcidid: 0000-0001-8001-1479 surname: Espejo fullname: Espejo, Edie organization: Division of Geriatrics, University of California, San Francisco, San Francisco, CA – sequence: 6 givenname: Sun Young surname: Jeon fullname: Jeon, Sun Young organization: Division of Geriatrics, University of California, San Francisco, San Francisco, CA – sequence: 7 givenname: W. John orcidid: 0000-0003-3121-9526 surname: Boscardin fullname: Boscardin, W. John organization: Division of Geriatrics, University of California, San Francisco, San Francisco, CA – sequence: 8 givenname: Timothy A. surname: Heintz fullname: Heintz, Timothy A. organization: School of Medicine, University of California, San Diego, San Diego, CA – sequence: 9 givenname: Christopher J. orcidid: 0000-0001-7444-2766 surname: Kennedy fullname: Kennedy, Christopher J. organization: Department of Psychiatry, Harvard Medical School, Boston, MA – sequence: 10 givenname: Deepshikha C. surname: Ashana fullname: Ashana, Deepshikha C. organization: Division of Pulmonary, Allergy, and Critical Care Medicine, Duke University, Durham, NC – sequence: 11 givenname: Allyson Cook orcidid: 0000-0001-9711-5514 surname: Chapman fullname: Chapman, Allyson Cook organization: Department of Medicine, the Division of Critical Care and Palliative Medicine, University of California, San Francisco, San Francisco, CA – sequence: 12 givenname: Karthik surname: Raghunathan fullname: Raghunathan, Karthik organization: Department of Anesthesia and Perioperative Care, Duke University, Durham, NC – sequence: 13 givenname: Alex K. orcidid: 0000-0002-9276-0861 surname: Smith fullname: Smith, Alex K. organization: Department of Geriatrics, Palliative, and Extended Care, Veterans Affairs Medical Center, University of California, San Francisco, San Francisco, CA – sequence: 14 givenname: Sei J. orcidid: 0000-0001-7864-5341 surname: Lee fullname: Lee, Sei J. organization: Division of Geriatrics, University of California, San Francisco, San Francisco, CA |
BackLink | https://www.ncbi.nlm.nih.gov/pubmed/38199323$$D View this record in MEDLINE/PubMed |
BookMark | eNp9kU1v1DAQhi1URLeFX4CEcuSS4PFk83FACFYFVurHhRVHyx-T1ksSL3ZSxL_H222rculpNJpn3hm97wk7Gv1IjL0FXgCH6sO2MDcUp0JwgQWIgiO8YAtoEXJclnjEFpyDyLFqxTE7iXHLUw9t9YodYwNtiwIX7OqCVJyDG6-z9bDrnXFT9sWpmLkxW6822aWfKGabuAd--mDzs0GTtfv2kuag-lSmPz78yi68pT6-Zi871Ud6c19P2ebr2Y_V9_z86tt69fk8NyXwKUeLjUClu7ZrtLYNGqtFp2urlrzmQqAW1dLW1JGxJZmyrhS3VHOFgKArjafs00F3N-uBrKFxSs_IXXCDCn-lV07-Pxndjbz2txKSQN1AnRTe3ysE_3tORsrBRUN9r0byc5SiBSzLtqzbhOIBNcHHGKh7vANc7rOQW3mXhdxnIUHIlEXaevf0xcedB_MT8PEAJN_o1lGQ0TgaDVkXyEzSevfsgX8Zb56Z |
Cites_doi | 10.1080/10826084.2019.1581221 10.1080/08897077.2014.930372 10.1126/science.aal4230 10.1525/aa.1984.86.1.02a00050 10.2196/17903 10.1073/pnas.2121798119 10.2337/cd16-0014 10.1038/sdata.2016.35 10.1098/rsta.2017.0362 10.1001/jamainternmed.2020.5792 10.1007/s11606-017-4289-2 10.1073/pnas.1720347115 10.1146/annurev-linguistics-011619-030303 10.1377/hlthaff.2021.01423 10.1177/0956797620963619 10.1016/j.socscimed.2008.09.009 10.1111/lang.12186 10.1186/s12910-017-0179-8 10.1007/s11606-022-07609-y 10.1007/s11606-007-0258-5 10.1016/j.drugpo.2009.10.010 10.3389/fpsyg.2022.871221 10.1001/jamanetworkopen.2021.44967 10.1001/jamanetworkopen.2021.17052 10.2307/411592 10.7326/0003-4819-132-11-200006060-00026 10.1126/science.aax2342 10.1016/j.jbi.2022.104149 10.1111/dme.13705 10.1111/bjso.12560 10.1136/bmjopen-2013-003384 |
ContentType | Journal Article |
Copyright | 2024 Published by Elsevier Inc. |
Copyright_xml | – notice: 2024 – notice: Published by Elsevier Inc. |
DBID | CGR CUY CVF ECM EIF NPM AAYXX CITATION 7X8 5PM |
DOI | 10.1016/j.chest.2023.12.031 |
DatabaseName | Medline MEDLINE MEDLINE (Ovid) MEDLINE MEDLINE PubMed CrossRef MEDLINE - Academic PubMed Central (Full Participant titles) |
DatabaseTitle | MEDLINE Medline Complete MEDLINE with Full Text PubMed MEDLINE (Ovid) CrossRef MEDLINE - Academic |
DatabaseTitleList | MEDLINE MEDLINE - Academic |
Database_xml | – sequence: 1 dbid: ECM name: MEDLINE url: https://search.ebscohost.com/login.aspx?direct=true&db=cmedm&site=ehost-live sourceTypes: Index Database |
DeliveryMethod | fulltext_linktorsrc |
Discipline | Medicine |
EISSN | 1931-3543 |
EndPage | 1490 |
ExternalDocumentID | 10_1016_j_chest_2023_12_031 38199323 S0012369224000072 |
Genre | Journal Article |
GrantInformation_xml | – fundername: NIA NIH HHS grantid: P30 AG044281 |
GroupedDBID | --- .1- .55 .FO .GJ .XZ 08P 0R~ 18M 1P~ 29B 2WC 354 36B 3O- 3V. 457 53G 5GY 5RE 5RS 6J9 6PF 7RV 7X7 88E 8AO 8C1 8F7 8FI 8FJ AAEDT AAEDW AAIAV AAKAS AAKUH AALRI AAWTL AAYOK ABDBF ABDQB ABJNI ABLJU ABMAC ABOCM ABUWG ACBMB ACGFO ACGFS ACGUR ADBBV ADZCM AENEX AEVXI AFCTW AFFNX AFKRA AFRHN AFTJW AGHFR AHMBA AI. AITUG AJUYK AKRWK ALMA_UNASSIGNED_HOLDINGS AMRAJ AZQEC B0M BCGUY BENPR BKEYQ BKNYI BPHCQ BR6 BVXVI C1A C45 CCPQU CS3 DU5 EAP EAS EBC EBD EBS EHN EJD EMK ENC EPT ESX EX3 F5P FDB FYUFA GD~ HMCUK HX~ HZ~ IAO IEA IH2 IHR IMI INH INR IOF IPO J5H K9- L7B LXL LXN M0R M1P M5~ MJL MV1 N4W N9A NAPCQ NEJ O6. O9- OB3 OBH ODZKP OFXIZ OGROG OHH OI- OK1 OU. OVD OVIDX P2P PCD PQQKQ PROAC PSQYO Q~Q RIG ROL SJN SSZ TCP TEORI TR2 TUS TWZ UKHRP VH1 W8F WH7 WOQ WOW X7M XOL YFH YHG YOC YQJ Z5R ZA5 ZGI ZRQ ZXP ZY1 ~8M AAXUO ADVLN AFETI AFJKZ ALIPV CGR CUY CVF ECM EIF H13 NPM AAYXX CITATION 7X8 5PM |
ID | FETCH-LOGICAL-c410t-3d3823abf9f8bbd83cdb2fb7da5070223b265d7efecd4ec476a0de70a3131b6b3 |
ISSN | 0012-3692 1931-3543 |
IngestDate | Tue Sep 17 21:27:43 EDT 2024 Sat Oct 26 04:21:46 EDT 2024 Thu Sep 26 16:25:35 EDT 2024 Sat Nov 02 12:29:27 EDT 2024 Sat Jun 01 15:42:00 EDT 2024 |
IsDoiOpenAccess | false |
IsOpenAccess | true |
IsPeerReviewed | true |
IsScholarly | true |
Issue | 6 |
Keywords | linguistics NLP natural language processing MIMIC-III UCSF BIDMC critical care inequity machine learning LLM |
Language | English |
License | Published by Elsevier Inc. |
LinkModel | OpenURL |
MergedId | FETCHMERGED-LOGICAL-c410t-3d3823abf9f8bbd83cdb2fb7da5070223b265d7efecd4ec476a0de70a3131b6b3 |
Notes | ObjectType-Article-1 SourceType-Scholarly Journals-1 ObjectType-Feature-2 content type line 23 |
ORCID | 0000-0001-9711-5514 0000-0001-9971-7535 0000-0001-7444-2766 0000-0002-9276-0861 0000-0001-7864-5341 0000-0003-3121-9526 0000-0001-8001-1479 0009-0005-3505-7300 0000-0002-3821-4376 |
OpenAccessLink | https://www.ncbi.nlm.nih.gov/pmc/articles/PMC11317817 |
PMID | 38199323 |
PQID | 2913449479 |
PQPubID | 23479 |
PageCount | 10 |
ParticipantIDs | pubmedcentral_primary_oai_pubmedcentral_nih_gov_11317817 proquest_miscellaneous_2913449479 crossref_primary_10_1016_j_chest_2023_12_031 pubmed_primary_38199323 elsevier_sciencedirect_doi_10_1016_j_chest_2023_12_031 |
PublicationCentury | 2000 |
PublicationDate | 2024-06-01 |
PublicationDateYYYYMMDD | 2024-06-01 |
PublicationDate_xml | – month: 06 year: 2024 text: 2024-06-01 day: 01 |
PublicationDecade | 2020 |
PublicationPlace | United States |
PublicationPlace_xml | – name: United States |
PublicationTitle | Chest |
PublicationTitleAlternate | Chest |
PublicationYear | 2024 |
Publisher | Elsevier Inc American College of Chest Physicians |
Publisher_xml | – name: Elsevier Inc – name: American College of Chest Physicians |
References | Burgess, Crowley-Matoka, Phelan (bib15) 2008; 67 Adam, Yang, Cato (bib36) July 2022 Cox, Fritz (bib14) 2022; 377 Green, Carney, Pallin (bib17) 2007; 22 Ashford, Brown, McDaniel, Curtis (bib31) 2019; 54 Kay, Kempton (bib29) 1984; 86 Liu, Capurro, Nguyen, Verspoor (bib42) 2022; 133 Langendoen (bib19) 1964; 40 Durrheim, Schuld, Mafunda, Mazibuko (bib3) 2023; 62 Hauser, Schwarz (bib37) 2022; 13 Garg, Schiebinger, Jurafsky, Zou (bib4) 2018; 115 FitzGerald, Hurst (bib9) 2017; 18 Obermeyer, Powers, Vogeli, Mullainathan (bib39) 2019; 366 Penn, Newman-Griffis (bib18) 2022; 2022 Park, Saha, Chee, Taylor, Beach (bib10) 2021; 4 Steiner, Earnest (bib26) 2000; 132 Johnson, Pollard, Shen (bib21) 2016; 3 Charlesworth, Yang, Mann, Kurdi, Banaji (bib6) 2021; 32 Healy, Richard, Kidia (bib33) 2022; 37 Balderston, Gertz, Seedat (bib7) 2021; 181 Caliskan, Lewis (bib38) 2020 1st ed. O’Reilly; 2009. Bolukbasi, Chang, Zou, Saligrama, Kalai (bib2) Boleda (bib20) 2020; 6 Shah (bib41) 2018; 376 Mikolov, Sutskever, Chen, Corrado, Dean (bib24) Charlesworth, Caliskan, Banaji (bib5) 2022; 119 Broyles, Binswanger, Jenkins (bib8) 2014; 35 (bib27) 2013 Bird S, Klein E, Loper E. Panch, Mattie, Atun (bib40) 2019; 9 Caliskan, Bryson, Narayanan (bib1) 2017; 356 Dickinson, Maryniuk (bib34) 2017; 35 Garcia-Rudolph, Saurí, Cegarra, Bernabeu Guitart (bib25) 2020; 8 Sun, Oliwa, Peek, Tung (bib11) 2022; 41 Řehůřek, Sojka (bib23) 2010 Thierry (bib30) 2016; 66 Cooper, Kanumilli, Hill (bib32) 2018; 35 Himmelstein, Bates, Zhou (bib12) 2022; 5 Kelly, Westerhoff (bib13) 2010; 21 Vaswani, Shazeer, Parmar (bib28) Browne, Ventura, Mosely, Speight (bib35) 2013; 3 Goddu, O’Conor, Lanzkron (bib16) 2018; 33 FitzGerald (10.1016/j.chest.2023.12.031_bib9) 2017; 18 Durrheim (10.1016/j.chest.2023.12.031_bib3) 2023; 62 Caliskan (10.1016/j.chest.2023.12.031_bib38) 2020 Himmelstein (10.1016/j.chest.2023.12.031_bib12) 2022; 5 Green (10.1016/j.chest.2023.12.031_bib17) 2007; 22 Bolukbasi (10.1016/j.chest.2023.12.031_bib2) Burgess (10.1016/j.chest.2023.12.031_bib15) 2008; 67 Garcia-Rudolph (10.1016/j.chest.2023.12.031_bib25) 2020; 8 Goddu (10.1016/j.chest.2023.12.031_bib16) 2018; 33 Langendoen (10.1016/j.chest.2023.12.031_bib19) 1964; 40 Vaswani (10.1016/j.chest.2023.12.031_bib28) Cooper (10.1016/j.chest.2023.12.031_bib32) 2018; 35 Park (10.1016/j.chest.2023.12.031_bib10) 2021; 4 Sun (10.1016/j.chest.2023.12.031_bib11) 2022; 41 Liu (10.1016/j.chest.2023.12.031_bib42) 2022; 133 Kelly (10.1016/j.chest.2023.12.031_bib13) 2010; 21 Mikolov (10.1016/j.chest.2023.12.031_bib24) Charlesworth (10.1016/j.chest.2023.12.031_bib5) 2022; 119 10.1016/j.chest.2023.12.031_bib22 Browne (10.1016/j.chest.2023.12.031_bib35) 2013; 3 Hauser (10.1016/j.chest.2023.12.031_bib37) 2022; 13 Charlesworth (10.1016/j.chest.2023.12.031_bib6) 2021; 32 Thierry (10.1016/j.chest.2023.12.031_bib30) 2016; 66 Adam (10.1016/j.chest.2023.12.031_bib36) 2022 Panch (10.1016/j.chest.2023.12.031_bib40) 2019; 9 Boleda (10.1016/j.chest.2023.12.031_bib20) 2020; 6 (10.1016/j.chest.2023.12.031_bib27) 2013 Kay (10.1016/j.chest.2023.12.031_bib29) 1984; 86 Balderston (10.1016/j.chest.2023.12.031_bib7) 2021; 181 Healy (10.1016/j.chest.2023.12.031_bib33) 2022; 37 Broyles (10.1016/j.chest.2023.12.031_bib8) 2014; 35 Steiner (10.1016/j.chest.2023.12.031_bib26) 2000; 132 Obermeyer (10.1016/j.chest.2023.12.031_bib39) 2019; 366 Ashford (10.1016/j.chest.2023.12.031_bib31) 2019; 54 Garg (10.1016/j.chest.2023.12.031_bib4) 2018; 115 Shah (10.1016/j.chest.2023.12.031_bib41) 2018; 376 Dickinson (10.1016/j.chest.2023.12.031_bib34) 2017; 35 Penn (10.1016/j.chest.2023.12.031_bib18) 2022; 2022 Johnson (10.1016/j.chest.2023.12.031_bib21) 2016; 3 Caliskan (10.1016/j.chest.2023.12.031_bib1) 2017; 356 Řehůřek (10.1016/j.chest.2023.12.031_bib23) 2010 Cox (10.1016/j.chest.2023.12.031_bib14) 2022; 377 |
References_xml | – ident: bib28 article-title: Attention is all you need [published online August 1, 2023] contributor: fullname: Parmar – volume: 35 start-page: 51 year: 2017 end-page: 54 ident: bib34 article-title: Building therapeutic relationships: choosing words that put people first publication-title: Clin Diabetes contributor: fullname: Maryniuk – ident: bib2 article-title: Man is to computer programmer as woman is to homemaker? Debiasing word embeddings [published online July 21, 2016] contributor: fullname: Kalai – volume: 18 start-page: 19 year: 2017 ident: bib9 article-title: Implicit bias in healthcare professionals: a systematic review publication-title: BMC Med Ethics contributor: fullname: Hurst – volume: 54 start-page: 1376 year: 2019 end-page: 1384 ident: bib31 article-title: Biased labels: an experimental study of language and stigma among individuals in recovery and health professionals publication-title: Subst Use Misuse contributor: fullname: Curtis – start-page: 478 year: 2020 end-page: 493 ident: bib38 article-title: Social biases in word embeddings and their relation to human cognition publication-title: Handbook of Language Analysis in Psychology contributor: fullname: Lewis – start-page: 7 year: July 2022 end-page: 21 ident: bib36 article-title: Write it like you see it: detectable differences in clinical notes by race lead to differential model recommendations publication-title: Proceedings of the 2022 AAAI/ACM Conference on AI, Ethics, and Society contributor: fullname: Cato – volume: 66 start-page: 690 year: 2016 end-page: 713 ident: bib30 article-title: Neurolinguistic relativity: how language flexes human perception and cognition publication-title: Lang Learn contributor: fullname: Thierry – volume: 181 start-page: 386 year: 2021 ident: bib7 article-title: Differential documentation of race in the first line of the history of present illness publication-title: JAMA Intern Med contributor: fullname: Seedat – volume: 40 start-page: 305 year: 1964 ident: bib19 article-title: Studies in linguistic analysis publication-title: Language contributor: fullname: Langendoen – volume: 5 year: 2022 ident: bib12 article-title: Examination of stigmatizing language in the electronic health record publication-title: JAMA Netw Open contributor: fullname: Zhou – volume: 9 year: 2019 ident: bib40 article-title: Artificial intelligence and algorithmic bias: implications for health systems publication-title: J Glob Health contributor: fullname: Atun – volume: 2022 start-page: 386 year: 2022 end-page: 395 ident: bib18 article-title: Half the picture: word frequencies reveal racial differences in clinical documentation, but not their causes publication-title: AMIA Jt Summits Transl Sci Proc contributor: fullname: Newman-Griffis – year: 2013 ident: bib27 publication-title: An Introduction to Statistical Learning: With Applications in R – volume: 37 start-page: 2533 year: 2022 end-page: 2540 ident: bib33 article-title: How to reduce stigma and bias in clinical communication: a narrative review publication-title: J Gen Intern Med contributor: fullname: Kidia – volume: 8 year: 2020 ident: bib25 article-title: Discovering the context of people with disabilities: semantic categorization test and environmental factors mapping of word embeddings from Reddit publication-title: JMIR Med Inform contributor: fullname: Bernabeu Guitart – volume: 376 year: 2018 ident: bib41 article-title: Algorithmic accountability publication-title: Phil Trans R Soc A contributor: fullname: Shah – start-page: 45 year: 2010 end-page: 50 ident: bib23 article-title: Software framework for topic modelling with large corpora publication-title: Proceedings of the LREC 2010 Workshop on New Challenges for NLP Frameworks contributor: fullname: Sojka – volume: 132 start-page: 926 year: 2000 ident: bib26 article-title: The language of medication-taking publication-title: Ann Intern Med contributor: fullname: Earnest – volume: 35 start-page: 217 year: 2014 end-page: 221 ident: bib8 article-title: Confronting inadvertent stigma and pejorative language in addiction scholarship: a recognition and response publication-title: Substance Abuse contributor: fullname: Jenkins – volume: 115 start-page: E3635 year: 2018 end-page: E3644 ident: bib4 article-title: Word embeddings quantify 100 years of gender and ethnic stereotypes publication-title: Proc Natl Acad Sci U S A contributor: fullname: Zou – volume: 6 start-page: 213 year: 2020 end-page: 234 ident: bib20 article-title: Distributional semantics and linguistic theory publication-title: Annu Rev Linguist contributor: fullname: Boleda – volume: 133 year: 2022 ident: bib42 article-title: “Note bloat” impacts deep learning-based NLP models for clinical prediction tasks publication-title: J Biomed Inform contributor: fullname: Verspoor – volume: 62 start-page: 617 year: 2023 end-page: 629 ident: bib3 article-title: Using word embeddings to investigate cultural biases publication-title: Br J Social Psychol contributor: fullname: Mazibuko – volume: 21 start-page: 202 year: 2010 end-page: 207 ident: bib13 article-title: Does it matter how we refer to individuals with substance-related conditions? A randomized study of two commonly used terms publication-title: Int J Drug Policy contributor: fullname: Westerhoff – volume: 3 year: 2016 ident: bib21 article-title: MIMIC-III, a freely accessible critical care database publication-title: Sci Data contributor: fullname: Shen – volume: 41 start-page: 203 year: 2022 end-page: 211 ident: bib11 article-title: Negative patient descriptors: documenting racial bias in the electronic health record: study examines racial bias in the patient descriptors used in the electronic health record publication-title: Health Affairs contributor: fullname: Tung – volume: 377 year: 2022 ident: bib14 article-title: Presenting complaint: use of language that disempowers patients publication-title: BMJ contributor: fullname: Fritz – volume: 13 year: 2022 ident: bib37 article-title: Implicit bias reflects the company that words keep publication-title: Front Psychol contributor: fullname: Schwarz – volume: 119 year: 2022 ident: bib5 article-title: Historical representations of social groups across 200 years of word embeddings from Google Books publication-title: Proc Natl Acad Sci U S A contributor: fullname: Banaji – volume: 4 year: 2021 ident: bib10 article-title: Physician use of stigmatizing language in patient medical records publication-title: JAMA Netw Open contributor: fullname: Beach – ident: bib24 article-title: Distributed representations of words and phrases and their compositionality contributor: fullname: Dean – volume: 356 start-page: 183 year: 2017 end-page: 186 ident: bib1 article-title: Semantics derived automatically from language corpora contain human-like biases publication-title: Science contributor: fullname: Narayanan – volume: 33 start-page: 685 year: 2018 end-page: 691 ident: bib16 article-title: Do words matter? Stigmatizing language and the transmission of bias in the medical record publication-title: J Gen Intern Med contributor: fullname: Lanzkron – volume: 67 start-page: 1852 year: 2008 end-page: 1860 ident: bib15 article-title: Patient race and physicians’ decisions to prescribe opioids for chronic low back pain publication-title: Soc Sci Med contributor: fullname: Phelan – volume: 32 start-page: 218 year: 2021 end-page: 240 ident: bib6 article-title: Gender stereotypes in natural language: word embeddings show robust consistency across child and adult language corpora of more than 65 million words publication-title: Psychol Sci contributor: fullname: Banaji – volume: 35 start-page: 1630 year: 2018 end-page: 1634 ident: bib32 article-title: Language matters. Addressing the use of language in the care of people with diabetes: position statement of the English Advisory Group publication-title: Diabet Med contributor: fullname: Hill – volume: 22 start-page: 1231 year: 2007 end-page: 1238 ident: bib17 article-title: Implicit bias among physicians and its prediction of thrombolysis decisions for Black and White patients publication-title: J Gen Intern Med contributor: fullname: Pallin – volume: 3 year: 2013 ident: bib35 article-title: ‘I call it the blame and shame disease’: a qualitative study about perceptions of social stigma surrounding type 2 diabetes publication-title: BMJ Open contributor: fullname: Speight – volume: 366 start-page: 447 year: 2019 end-page: 453 ident: bib39 article-title: Dissecting racial bias in an algorithm used to manage the health of populations publication-title: Science contributor: fullname: Mullainathan – volume: 86 start-page: 65 year: 1984 end-page: 79 ident: bib29 article-title: What is the Sapir-Whorf hypothesis? publication-title: Am Anthropol contributor: fullname: Kempton – volume: 54 start-page: 1376 issue: 8 year: 2019 ident: 10.1016/j.chest.2023.12.031_bib31 article-title: Biased labels: an experimental study of language and stigma among individuals in recovery and health professionals publication-title: Subst Use Misuse doi: 10.1080/10826084.2019.1581221 contributor: fullname: Ashford – ident: 10.1016/j.chest.2023.12.031_bib22 – volume: 35 start-page: 217 issue: 3 year: 2014 ident: 10.1016/j.chest.2023.12.031_bib8 article-title: Confronting inadvertent stigma and pejorative language in addiction scholarship: a recognition and response publication-title: Substance Abuse doi: 10.1080/08897077.2014.930372 contributor: fullname: Broyles – ident: 10.1016/j.chest.2023.12.031_bib24 contributor: fullname: Mikolov – volume: 377 year: 2022 ident: 10.1016/j.chest.2023.12.031_bib14 article-title: Presenting complaint: use of language that disempowers patients publication-title: BMJ contributor: fullname: Cox – volume: 356 start-page: 183 issue: 6334 year: 2017 ident: 10.1016/j.chest.2023.12.031_bib1 article-title: Semantics derived automatically from language corpora contain human-like biases publication-title: Science doi: 10.1126/science.aal4230 contributor: fullname: Caliskan – volume: 86 start-page: 65 issue: 1 year: 1984 ident: 10.1016/j.chest.2023.12.031_bib29 article-title: What is the Sapir-Whorf hypothesis? publication-title: Am Anthropol doi: 10.1525/aa.1984.86.1.02a00050 contributor: fullname: Kay – ident: 10.1016/j.chest.2023.12.031_bib2 contributor: fullname: Bolukbasi – volume: 8 issue: 11 year: 2020 ident: 10.1016/j.chest.2023.12.031_bib25 article-title: Discovering the context of people with disabilities: semantic categorization test and environmental factors mapping of word embeddings from Reddit publication-title: JMIR Med Inform doi: 10.2196/17903 contributor: fullname: Garcia-Rudolph – volume: 119 issue: 28 year: 2022 ident: 10.1016/j.chest.2023.12.031_bib5 article-title: Historical representations of social groups across 200 years of word embeddings from Google Books publication-title: Proc Natl Acad Sci U S A doi: 10.1073/pnas.2121798119 contributor: fullname: Charlesworth – volume: 2022 start-page: 386 year: 2022 ident: 10.1016/j.chest.2023.12.031_bib18 article-title: Half the picture: word frequencies reveal racial differences in clinical documentation, but not their causes publication-title: AMIA Jt Summits Transl Sci Proc contributor: fullname: Penn – start-page: 45 year: 2010 ident: 10.1016/j.chest.2023.12.031_bib23 article-title: Software framework for topic modelling with large corpora contributor: fullname: Řehůřek – ident: 10.1016/j.chest.2023.12.031_bib28 contributor: fullname: Vaswani – volume: 35 start-page: 51 issue: 1 year: 2017 ident: 10.1016/j.chest.2023.12.031_bib34 article-title: Building therapeutic relationships: choosing words that put people first publication-title: Clin Diabetes doi: 10.2337/cd16-0014 contributor: fullname: Dickinson – volume: 3 issue: 1 year: 2016 ident: 10.1016/j.chest.2023.12.031_bib21 article-title: MIMIC-III, a freely accessible critical care database publication-title: Sci Data doi: 10.1038/sdata.2016.35 contributor: fullname: Johnson – volume: 376 issue: 2128 year: 2018 ident: 10.1016/j.chest.2023.12.031_bib41 article-title: Algorithmic accountability publication-title: Phil Trans R Soc A doi: 10.1098/rsta.2017.0362 contributor: fullname: Shah – volume: 181 start-page: 386 issue: 3 year: 2021 ident: 10.1016/j.chest.2023.12.031_bib7 article-title: Differential documentation of race in the first line of the history of present illness publication-title: JAMA Intern Med doi: 10.1001/jamainternmed.2020.5792 contributor: fullname: Balderston – volume: 33 start-page: 685 issue: 5 year: 2018 ident: 10.1016/j.chest.2023.12.031_bib16 article-title: Do words matter? Stigmatizing language and the transmission of bias in the medical record publication-title: J Gen Intern Med doi: 10.1007/s11606-017-4289-2 contributor: fullname: Goddu – volume: 115 start-page: E3635 issue: 16 year: 2018 ident: 10.1016/j.chest.2023.12.031_bib4 article-title: Word embeddings quantify 100 years of gender and ethnic stereotypes publication-title: Proc Natl Acad Sci U S A doi: 10.1073/pnas.1720347115 contributor: fullname: Garg – volume: 6 start-page: 213 issue: 1 year: 2020 ident: 10.1016/j.chest.2023.12.031_bib20 article-title: Distributional semantics and linguistic theory publication-title: Annu Rev Linguist doi: 10.1146/annurev-linguistics-011619-030303 contributor: fullname: Boleda – volume: 41 start-page: 203 issue: 2 year: 2022 ident: 10.1016/j.chest.2023.12.031_bib11 article-title: Negative patient descriptors: documenting racial bias in the electronic health record: study examines racial bias in the patient descriptors used in the electronic health record publication-title: Health Affairs doi: 10.1377/hlthaff.2021.01423 contributor: fullname: Sun – volume: 32 start-page: 218 issue: 2 year: 2021 ident: 10.1016/j.chest.2023.12.031_bib6 article-title: Gender stereotypes in natural language: word embeddings show robust consistency across child and adult language corpora of more than 65 million words publication-title: Psychol Sci doi: 10.1177/0956797620963619 contributor: fullname: Charlesworth – volume: 67 start-page: 1852 issue: 11 year: 2008 ident: 10.1016/j.chest.2023.12.031_bib15 article-title: Patient race and physicians’ decisions to prescribe opioids for chronic low back pain publication-title: Soc Sci Med doi: 10.1016/j.socscimed.2008.09.009 contributor: fullname: Burgess – volume: 66 start-page: 690 issue: 3 year: 2016 ident: 10.1016/j.chest.2023.12.031_bib30 article-title: Neurolinguistic relativity: how language flexes human perception and cognition publication-title: Lang Learn doi: 10.1111/lang.12186 contributor: fullname: Thierry – volume: 18 start-page: 19 issue: 1 year: 2017 ident: 10.1016/j.chest.2023.12.031_bib9 article-title: Implicit bias in healthcare professionals: a systematic review publication-title: BMC Med Ethics doi: 10.1186/s12910-017-0179-8 contributor: fullname: FitzGerald – volume: 37 start-page: 2533 issue: 10 year: 2022 ident: 10.1016/j.chest.2023.12.031_bib33 article-title: How to reduce stigma and bias in clinical communication: a narrative review publication-title: J Gen Intern Med doi: 10.1007/s11606-022-07609-y contributor: fullname: Healy – volume: 22 start-page: 1231 issue: 9 year: 2007 ident: 10.1016/j.chest.2023.12.031_bib17 article-title: Implicit bias among physicians and its prediction of thrombolysis decisions for Black and White patients publication-title: J Gen Intern Med doi: 10.1007/s11606-007-0258-5 contributor: fullname: Green – volume: 21 start-page: 202 issue: 3 year: 2010 ident: 10.1016/j.chest.2023.12.031_bib13 article-title: Does it matter how we refer to individuals with substance-related conditions? A randomized study of two commonly used terms publication-title: Int J Drug Policy doi: 10.1016/j.drugpo.2009.10.010 contributor: fullname: Kelly – volume: 13 year: 2022 ident: 10.1016/j.chest.2023.12.031_bib37 article-title: Implicit bias reflects the company that words keep publication-title: Front Psychol doi: 10.3389/fpsyg.2022.871221 contributor: fullname: Hauser – volume: 9 issue: 2 year: 2019 ident: 10.1016/j.chest.2023.12.031_bib40 article-title: Artificial intelligence and algorithmic bias: implications for health systems publication-title: J Glob Health contributor: fullname: Panch – volume: 5 issue: 1 year: 2022 ident: 10.1016/j.chest.2023.12.031_bib12 article-title: Examination of stigmatizing language in the electronic health record publication-title: JAMA Netw Open doi: 10.1001/jamanetworkopen.2021.44967 contributor: fullname: Himmelstein – volume: 4 issue: 7 year: 2021 ident: 10.1016/j.chest.2023.12.031_bib10 article-title: Physician use of stigmatizing language in patient medical records publication-title: JAMA Netw Open doi: 10.1001/jamanetworkopen.2021.17052 contributor: fullname: Park – volume: 40 start-page: 305 issue: 2 year: 1964 ident: 10.1016/j.chest.2023.12.031_bib19 article-title: Studies in linguistic analysis publication-title: Language doi: 10.2307/411592 contributor: fullname: Langendoen – volume: 132 start-page: 926 issue: 11 year: 2000 ident: 10.1016/j.chest.2023.12.031_bib26 article-title: The language of medication-taking publication-title: Ann Intern Med doi: 10.7326/0003-4819-132-11-200006060-00026 contributor: fullname: Steiner – volume: 366 start-page: 447 issue: 6464 year: 2019 ident: 10.1016/j.chest.2023.12.031_bib39 article-title: Dissecting racial bias in an algorithm used to manage the health of populations publication-title: Science doi: 10.1126/science.aax2342 contributor: fullname: Obermeyer – start-page: 7 year: 2022 ident: 10.1016/j.chest.2023.12.031_bib36 article-title: Write it like you see it: detectable differences in clinical notes by race lead to differential model recommendations contributor: fullname: Adam – volume: 133 year: 2022 ident: 10.1016/j.chest.2023.12.031_bib42 article-title: “Note bloat” impacts deep learning-based NLP models for clinical prediction tasks publication-title: J Biomed Inform doi: 10.1016/j.jbi.2022.104149 contributor: fullname: Liu – volume: 35 start-page: 1630 issue: 12 year: 2018 ident: 10.1016/j.chest.2023.12.031_bib32 article-title: Language matters. Addressing the use of language in the care of people with diabetes: position statement of the English Advisory Group publication-title: Diabet Med doi: 10.1111/dme.13705 contributor: fullname: Cooper – year: 2013 ident: 10.1016/j.chest.2023.12.031_bib27 – start-page: 478 year: 2020 ident: 10.1016/j.chest.2023.12.031_bib38 article-title: Social biases in word embeddings and their relation to human cognition contributor: fullname: Caliskan – volume: 62 start-page: 617 issue: 1 year: 2023 ident: 10.1016/j.chest.2023.12.031_bib3 article-title: Using word embeddings to investigate cultural biases publication-title: Br J Social Psychol doi: 10.1111/bjso.12560 contributor: fullname: Durrheim – volume: 3 issue: 11 year: 2013 ident: 10.1016/j.chest.2023.12.031_bib35 article-title: ‘I call it the blame and shame disease’: a qualitative study about perceptions of social stigma surrounding type 2 diabetes publication-title: BMJ Open doi: 10.1136/bmjopen-2013-003384 contributor: fullname: Browne |
SSID | ssj0001196 |
Score | 2.4974022 |
Snippet | Language in nonmedical data sets is known to transmit human-like biases when used in natural language processing (NLP) algorithms that can reinforce... |
SourceID | pubmedcentral proquest crossref pubmed elsevier |
SourceType | Open Access Repository Aggregation Database Index Database Publisher |
StartPage | 1481 |
SubjectTerms | Algorithms Bias critical care Critical Illness - psychology Electronic Health Records Female Humanities: Original Research Humans inequity Intensive Care Units linguistics machine learning Male Natural Language Processing Neural Networks, Computer |
Title | Measuring Implicit Bias in ICU Notes Using Word-Embedding Neural Network Models |
URI | https://dx.doi.org/10.1016/j.chest.2023.12.031 https://www.ncbi.nlm.nih.gov/pubmed/38199323 https://www.proquest.com/docview/2913449479 https://pubmed.ncbi.nlm.nih.gov/PMC11317817 |
Volume | 165 |
hasFullText | 1 |
inHoldings | 1 |
isFullTextHit | |
isPrint | |
link | http://sdu.summon.serialssolutions.com/2.0.0/link/0/eLvHCXMwtV1Lb9NAEF6lRUJcEG9SHlokbsVR1rvx41hCIByaSrQViIu1L6tpwanq5MC_Z3bW69gUECBxsSInsZ2ZL7Mzs9_MEPJynNgsEaWOTKLLCFZoGeWZTiKdG600k0Zgkdj8OF18yt7MxGwwCPSu7bn_qmk4B7p2lbN_oe32onACXoPO4Qhah-Mf6f0Qk37IA0Cy-HK9_3opkfX6fnq6v1i5RKsnCnyEwDOafVXWYGWL69MBClt4YjhOSftSd53X6VnYi8K6DtWU-7gS621BmSsuRGzMN1WH-tsQfg5cS632Gu-c5V1d1RfyG_qwRxe19NO8QyIiFlvClM-OhQqZHoHTuVMRT_y8u5H1RjbnLOIT356ptcJ-ZEQDt65NhYCNddZnCOnGP7X9Pg1xPsJBYyM3Fh4Tvc0i02-qfex71-VIoXXd03fIjRhMlbOUnycf2rWcMZzw1v6K0LcKGYLXbvQr3-Z67PIjBbfj05zcIbebYIQeeBTdJQNb3SM3Dxu6xX1y1IKJBjBRBya6rCiAiSKYKIKJ9sFEPZhoAybqwfSAnL6dnUznUTOBI9KCjdcRN26bWKoyLzOlTMa1UXGpUiMhjADvj6s4mZjUllYbYbVIEzk2Nh1LzjhTieIPyW61quxjQjV6nkxZyaxQNlVSxuAdwXqhhCpjPSSvguyKS99opQgMxPMCRV04URcsLkDUQ5IE-RaNr-h9wALg8PsvvgjaKMCSuu0xWdnVpi5iR0IRuUjzIXnktdM-ictrQKTDhyTr6a39gOvS3n-nWp5ht3YGokgzlu796xM_Ibe2f7enZHd9tbHPyE5tNs8RrN8BjpO0uQ |
link.rule.ids | 230,315,782,786,887,27933,27934 |
linkProvider | Elsevier |
openUrl | ctx_ver=Z39.88-2004&ctx_enc=info%3Aofi%2Fenc%3AUTF-8&rfr_id=info%3Asid%2Fsummon.serialssolutions.com&rft_val_fmt=info%3Aofi%2Ffmt%3Akev%3Amtx%3Ajournal&rft.genre=article&rft.atitle=Measuring+Implicit+Bias+in+ICU+Notes+Using+Word-Embedding+Neural+Network+Models&rft.jtitle=Chest&rft.au=Cobert%2C+Julien&rft.au=Mills%2C+Hunter&rft.au=Lee%2C+Albert&rft.au=Gologorskaya%2C+Oksana&rft.date=2024-06-01&rft.pub=Elsevier+Inc&rft.issn=0012-3692&rft.eissn=1931-3543&rft.volume=165&rft.issue=6&rft.spage=1481&rft.epage=1490&rft_id=info:doi/10.1016%2Fj.chest.2023.12.031&rft.externalDocID=S0012369224000072 |
thumbnail_l | http://covers-cdn.summon.serialssolutions.com/index.aspx?isbn=/lc.gif&issn=0012-3692&client=summon |
thumbnail_m | http://covers-cdn.summon.serialssolutions.com/index.aspx?isbn=/mc.gif&issn=0012-3692&client=summon |
thumbnail_s | http://covers-cdn.summon.serialssolutions.com/index.aspx?isbn=/sc.gif&issn=0012-3692&client=summon |