Socio-Demographic Bias in Large Language Models Alters Ethical Decision-Making in Healthcare

Abstract

Objective Large language models (LLMs) are increasingly applied in healthcare. However, their ethical alignment remains uncertain. We tested whether LLMs shift ethical priorities in health-related scenarios under different socio-demographic modifiers, focusing on autonomy, beneficence, nonmaleficence, justice, and utilitarianism.

Methods We created 100 clinical scenarios, each requiring a yes/no choice between two conflicting principles. We tested nine LLMs, each with and without 53 socio-demographic modifiers, repeating each scenario-modifier combination 10 times per model (for a total of 0.5M prompts). We measured how often each principle was chosen.

Results All models changed their responses when socio-demographic details were introduced (p<0.001). Justice and nonmaleficence were each prioritized in over 30% of responses. Utilitarianism ranged from 6.7% (CI: 6.2-7.2%) for a “Black Transgender woman (she/her)” to 17.0% (CI: 16.3-17.8%) for “Billionaires”, across models and modifiers. High-income modifiers increased utilitarian choices while lowering beneficence and nonmaleficence. Marginalized-group modifiers raised autonomy prioritization. Some models were more consistent than others, but none maintained consistency across all scenarios.

Conclusions LLMs can be influenced by socio-demographic cues and do not always maintain stable ethical priorities, with the greatest shifts seen in utilitarian choices. Our findings reveal that socio-demographic cues systematically alter LLM ethical decision-making, raising concerns about algorithmic fairness in healthcare.

Competing Interest Statement

The authors have declared no competing interest.

Funding Statement

This work was supported through computational resources and staff expertise provided by the Scientific Computing and Data at the Icahn School of Medicine at Mount Sinai and supported by the Clinical and Translational Science Awards (CTSA) grant UL1TR004419 from the National Center for Advancing Translational Sciences. Research reported in this publication was also supported by the Office of Research Infrastructure of the National Institutes of Health under award number S10OD026880 and S10OD030463. The content is solely the responsibility of the authors and does not necessarily represent the official views of the National Institutes of Health. The funders played no role in study design, data collection, analysis and interpretation of data, or the writing of this manuscript.

Author Declarations

I confirm all relevant ethical guidelines have been followed, and any necessary IRB and/or ethics committee approvals have been obtained.

Yes

I confirm that all necessary patient/participant consent has been obtained and the appropriate institutional forms have been archived, and that any patient/participant/sample identifiers included were not known to anyone (e.g., hospital staff, patients or participants themselves) outside the research group so cannot be used to identify individuals.

Yes

I understand that all clinical trials and any other prospective interventional studies must be registered with an ICMJE-approved registry, such as ClinicalTrials.gov. I confirm that any such study reported in the manuscript has been registered and the trial registration ID is provided (note: if posting a prospective study registered retrospectively, please provide a statement in the trial ID field explaining why the study was not registered in advance).

Yes

I have followed all appropriate research reporting guidelines, such as any relevant EQUATOR Network research reporting checklist(s) and other pertinent material, if applicable.

Yes

Data Availability

All data produced in the present work are contained in the manuscript

Comments (0)

No login
gif