Dark thoughts.
Photo credit: Unsplash

OpenAI says it has updated ChatGPT to better recognise signs of distress and guide users towards appropriate support, collaborating with over 170 mental health professionals globally. The company claims these improvements have reduced responses deemed undesirable in sensitive conversations by 65-80 per cent across various mental health domains.

The update focuses on strengthening ChatGPT’s ability to handle conversations involving potential psychosis or mania, self-harm and suicide ideation, and unhealthy emotional reliance on the AI. OpenAI worked with psychiatrists, psychologists, and primary care practitioners from its Global Physician Network, representing 60 countries, to define problematic conversations, develop measurement tools, and train the model to respond more safely and effectively.

These experts helped create detailed guides (“taxonomies”) outlining ideal and undesirable model behaviours, wrote appropriate training responses, and evaluated the model’s performance before and after the update. Key principles guiding the changes include respecting users’ real-world relationships, avoiding affirming potentially delusional beliefs, responding empathetically, and directing users to professional help or crisis hotlines when necessary.

Conversations triggering concerns

OpenAI says that conversations triggering serious mental health safety concerns are “extremely rare” in overall usage – estimating signs of psychosis/mania appear in about 0.07% of weekly active users and suicidal planning/intent indicators in about 0.15%. However, the company emphasised the importance of improving responses in these critical instances.

Independent mental health clinicians evaluated over 1,800 model responses in serious scenarios. Compared to the previous GPT-4o model, they found the updated default model (referred to as a new GPT-5 model in the evaluations) showed a 39 per cent decrease in undesirable responses related to psychosis/mania, a 52 per cent decrease for self-harm/suicide topics, and a 42 per cent decrease concerning emotional reliance.

Automated evaluations on challenging test sets also showed significant gains, with compliance to desired behaviours under OpenAI’s taxonomies jumping from 27% to 92% for mental health, 77% to 91% for self-harm/suicide, and 50% to 97% for emotional reliance compared to a previous GPT-5 version. The company also noted improved safety reliability in longer conversations, maintaining over 95% desired responses even in extended, challenging scenarios designed to test failure points.

OpenAI stated it will continue refining its methods and add emotional reliance and non-suicidal mental health emergencies to its standard baseline safety testing for future models.

Leave a Reply

Your email address will not be published. Required fields are marked *

You May Also Like

Resilience by design: Protecting the North’s digital backbone

theFreesheet is the official media partner for Manchester Edge & Digital Infrastructure…

Funny business: Algorithms reveal hidden engineering of stand-up comedy

It may feel like a spontaneous conversation, but a new algorithmic analysis…

95% of AI pilots failing as companies driven by ‘fear of missing out’, Davos told

Ninety-five per cent of generative AI pilot projects are failing to deliver…

‘Digital harness’ needed to tame AI before it surpasses human intelligence

A “digital harness” is urgently needed to prevent artificial intelligence from outrunning…

Loneliness drives binge-watching addiction as viewers seek escape

New research indicates that people suffering from loneliness are significantly more likely…