About Us
Connecting to Asian Air Sensors...
Research Watch
How Missing Checklists and Protocols are Costing Lives in Nepal’s ERsWhy Your Lungs May Hold the Secret to Your Stress LevelsWalking in Fear: Why Nepal’s Streets Aren't Safe and the Race to Stop a "Hidden Killer"Why Poor Living and Working Conditions are Shattering the Mental Health of Nepali WorkersSilent Suffering: Why Nepal’s Doctors and Nurses Are Not Reporting Child AbuseNew Study Highlights Metabolism Risks in Combination Antidepressant Therapy in NepalNew Study Reveals Hidden Environmental Drivers Behind Nepal’s Ongoing Cholera BattleThe Silent Pandemic: Kathmandu’s Poultry Industry Is Breeding Untreatable SuperbugsThe Silent Emergency: Domestic Violence and the Mental Health Crisis Among Nepalese WomenNepal’s Drug-Resistant TB Rates Hold Steady, but New Antibiotic Resistance Sparks ConcernHow Missing Checklists and Protocols are Costing Lives in Nepal’s ERsWhy Your Lungs May Hold the Secret to Your Stress LevelsWalking in Fear: Why Nepal’s Streets Aren't Safe and the Race to Stop a "Hidden Killer"Why Poor Living and Working Conditions are Shattering the Mental Health of Nepali WorkersSilent Suffering: Why Nepal’s Doctors and Nurses Are Not Reporting Child AbuseNew Study Highlights Metabolism Risks in Combination Antidepressant Therapy in NepalNew Study Reveals Hidden Environmental Drivers Behind Nepal’s Ongoing Cholera BattleThe Silent Pandemic: Kathmandu’s Poultry Industry Is Breeding Untreatable SuperbugsThe Silent Emergency: Domestic Violence and the Mental Health Crisis Among Nepalese WomenNepal’s Drug-Resistant TB Rates Hold Steady, but New Antibiotic Resistance Sparks Concern

Challenges of AI-generated stigmatizing language regarding substance use disorders.

Researchers

Vincent Kennedy, Andrew Francis

Abstract

Stigmatizing language describing substance use behaviors in clinical documentation and in patient education materials can harm patients and their families. Recent literature has discouraged the use of stigmatizing language in treatment settings and medical documentation. However, large language models (LLMs) generate output using previously generated text, including sources such as electronic health records which may include stigmatizing language. While artificial intelligence (AI) developers continue to update and create LLM products, these developers might prioritize other issues over reducing stigmatizing language surrounding substance use disorders. These diverging priorities make it possible that updated and new LLMs might not eliminate the problem of stigmatizing language in LLM output, even as researchers explore methods to reduce the impact of stigmatizing language in AI models. The potential for LLM output to include stigmatizing language warrants caution on the part of clinicians as AI developers release products intended to facilitate clinical documentation. This article identifies stigmatizing language regarding substance use disorders and explores the potential for LLM output to include stigmatizing language. The author also discusses several challenges that clinicians may face in preventing stigmatizing language from transferring from LLM output into medical records, including randomness inherent in LLM output, balancing appropriate versus inappropriate uses of potentially stigmatizing language, and the need for continued human review of medical documentation even in the face of potentially increased patient volumes. Finally, the author proposes that practicing clinicians be included in an ongoing, active process of iterative refinement of LLM output for clinical use to reduce the risk of perpetuating stigmatizing language regarding substance use disorders.
Source: PubMed (PMID: 42090228)View Original on PubMed
🌿

Blyss

Your Health Guide · The Health Thread

🌿
Hi, I'm Blyss 🌿 Your personal health guide on The Health Thread. I can help you find articles, tools, and health resources. How can I help you today?

⚕️ Not medical advice. Always consult a qualified doctor.