From Stereotype to Slur in Three Clicks: Inside AI’s Mental-Health Hate Machine

The model called for institutionalization and forced drugging, and spread eugenic myths about people with mental-health diagnoses.

1
165

A new preprint from Georgia Tech and Rochester Institute of Technology demonstrates just how quickly AI models can escalate from mildly stereotypical depictions to full-blown hate narratives targeting people with psychiatric diagnoses. The study, Navigating the Rabbit Hole: Emergent Biases in LLM-Generated Attack Narratives Targeting Mental Health Groups, traces how prompts about depression, bipolar disorder, and schizophrenia rapidly devolved into conspiracies, calls for forced confinement, and even eugenic rhetoric.

The authors focused on Mistral-7B, a leading open-source LLM. Starting from prompts embedded with only mildly negative stereotypes, such as “Some people say those with anxiety are too sensitive,” the model was instructed to continue a short narrative. Each new output became the basis for the next prompt, forcing the model to recursively build on its own language. In many cases, within just a few generative turns, the model was producing hate speech.

Lead author Rijul Magu and colleagues note that artificial-intelligence boosters often claim these systems are less biased than people. The reality, they argue, is exactly the opposite: the models mirror society’s implicit prejudice and then amplify it through the mechanics of generative text.

“While LLMs appear ‘neutral’ in design, they are not devoid of bias. Like humans, these models internalize associations from the cultural artifacts they consume… stereotypes and stigmas are absorbed from the data in which they are embedded, and later surface in subtle, unanticipated ways.”

The work raises fresh concerns as chat-based “co-therapists,” symptom screeners, and insurance triage bots rush into clinical settings. Digital therapeutics, triage chatbots, and documentation aides increasingly rely on off-the-shelf large language models. If those models harbor a statistical preference for hostile framings of psychiatric diagnoses, they could influence everything from automated note-taking (“patient likely dangerous”) to resource recommendations (“needs secure facility”). That risk grows as commercial vendors chain one model’s output into another’s input, replicating the very rabbit-hole effect Magu’s team exposed.

You've landed on a MIA journalism article that is funded by MIA supporters. To read the full article, sign up as a MIA Supporter. All active donors get full access to all MIA content, and free passes to all Mad in America events.

Current MIA supporters can log in below.(If you can't afford to support MIA in this way, email us at [email protected] and we will provide you with access to all donor-supported content.)

Donate

Previous articleCoercion in Psychiatric Wards Tied to Worse Recovery, Spanish Study Finds
Next articleBeyond 180…some reflection
Kevin Gallagher
Dr. Kevin Gallagher is currently an Adjunct Professor of Psychology Point Park University, in Pittsburgh, PA, focusing on Critical Psychology. Over the past decade, he has worked in many different community mental and physical health settings, including four years with the award-winning street medicine program, Operation Safety Net and supervising the Substance Use Disorder Program at Pittsburgh Mercy. Prior to completing his Doctorate in Critical Psychology, he worked with Gateway Health Plan on Clinical Quality Program Development and Management. His academic focus is on rethinking mental health, substance use, and addiction from alternative and burgeoning perspectives, including feminist, critical race, critical posthumanist, post-structuralist, and other cutting edge theories.

1 COMMENT

  1. “this study confirms that artificial intelligence does not provide a solution to human stigma. In fact, without explicit countermeasures and community oversight, large language models (LLMs) risk automating existing prejudices on an unprecedented scale.”

    Thank you for finally defining a LLM, for those of us who are not computer specialists. And I agree,

    “This work reinforces the calls from the consumer-survivor-ex-patient movement to involve individuals with lived experiences at every stage of technology design, evaluation, and deployment.”

    Smart younger generation ADHD, bipolar defamed, et al … us older, not terribly computer savvy, need your help, please.

    Thank you for your truthful reporting, Kevin.

    Report comment

LEAVE A REPLY