Ptechhub
  • News
  • Industries
    • Enterprise IT
    • AI & ML
    • Cybersecurity
    • Finance
    • Telco
  • Brand Hub
    • Lifesight
  • Blogs
No Result
View All Result
  • News
  • Industries
    • Enterprise IT
    • AI & ML
    • Cybersecurity
    • Finance
    • Telco
  • Brand Hub
    • Lifesight
  • Blogs
No Result
View All Result
PtechHub
No Result
View All Result

A Research Leader Behind ChatGPT’s Mental Health Work Is Leaving OpenAI

By Wired by By Wired
November 24, 2025
Home AI & ML
Share on FacebookShare on Twitter


An OpenAI safety research leader who helped shape ChatGPT’s responses to users experiencing mental health crises announced her departure from the company internally last month, WIRED has learned. Andrea Vallone, the head of a safety research team known as model policy, is slated to leave OpenAI at the end of the year.

OpenAI spokesperson Kayla Wood confirmed Vallone’s departure. Wood said OpenAI is actively looking for a replacement and that, in the interim, Vallone’s team will report directly to Johannes Heidecke, the company’s head of safety systems.

Vallone’s departure comes as OpenAI faces growing scrutiny over how its flagship product responds to users in distress. In recent months, several lawsuits have been filed against OpenAI alleging that users formed unhealthy attachments to ChatGPT. Some of the lawsuits claim ChatGPT contributed to mental health breakdowns or encouraged suicidal ideations.

Amid that pressure, OpenAI has been working to understand how ChatGPT should handle distressed users and improve the chatbot’s responses. Model policy is one of the teams leading that work, spearheading an October report detailing the company’s progress and consultations with more than 170 mental health experts.

In the report, OpenAI said hundreds of thousands of ChatGPT users may show signs of experiencing a manic or psychotic crisis every week, and that more than a million people “have conversations that include explicit indicators of potential suicidal planning or intent.” Through an update to GPT-5, OpenAI said in the report it was able to reduce undesirable responses in these conversations by 65 to 80 percent.

“Over the past year, I led OpenAI’s research on a question with almost no established precedents: how should models respond when confronted with signs of emotional over-reliance or early indications of mental health distress?” wrote Vallone in a post on LinkedIn.

Vallone did not respond to WIRED’s request for comment.

Making ChatGPT enjoyable to chat with, but not overly flattering, is a core tension at OpenAI. The company is aggressively trying to expand ChatGPT’s user base, which now includes more than 800 million people a week, to compete with AI chatbots from Google, Anthropic, and Meta.

After OpenAI released GPT-5 in August, users pushed back, arguing that the new model was surprisingly cold. In the latest update to ChatGPT, the company said it had significantly reduced sycophancy while maintaining the chatbot’s “warmth.”

Vallone’s exit follows an August reorganization of another group focused on ChatGPT’s responses to distressed users, model behavior. Its former leader, Joanne Jang, left that role to start a new team exploring novel human–AI interaction methods. The remaining model behavior staff were moved under post-training lead Max Schwarzer.



Source link

Tags: Artificial Intelligencechatgptopenairesearchsafety
By Wired

By Wired

Next Post
Chief data officers hire for tech innovation | Computer Weekly

Chief data officers hire for tech innovation | Computer Weekly

Recommended.

The Kraft Group CIO Michael Israel On The ‘Enormous’ NWN Deal And How AI Is Changing Decision-Making

The Kraft Group CIO Michael Israel On The ‘Enormous’ NWN Deal And How AI Is Changing Decision-Making

April 16, 2025
Here’s why ‘dead’ investors outperform the living

Here’s why ‘dead’ investors outperform the living

April 5, 2025

Trending.

Chai AI Announces Upcoming Rollout of Apple and Google Age Verification APIs to Enhance Platform Safety

Chai AI Announces Upcoming Rollout of Apple and Google Age Verification APIs to Enhance Platform Safety

March 10, 2026
Huawei lanceert Next Generation FAN-oplossing

Huawei lanceert Next Generation FAN-oplossing

March 7, 2026
Baidu Announces Fourth Quarter and Fiscal Year 2025 Results

Baidu Announces Fourth Quarter and Fiscal Year 2025 Results

February 26, 2026
Half of Google’s software development now AI-generated | Computer Weekly

Half of Google’s software development now AI-generated | Computer Weekly

February 5, 2026
Ghost Campaign Uses 7 npm Packages to Steal Crypto Wallets and Credentials

Ghost Campaign Uses 7 npm Packages to Steal Crypto Wallets and Credentials

March 24, 2026

PTechHub

A tech news platform delivering fresh perspectives, critical insights, and in-depth reporting — beyond the buzz. We cover innovation, policy, and digital culture with clarity, independence, and a sharp editorial edge.

Follow Us

Industries

  • AI & ML
  • Cybersecurity
  • Enterprise IT
  • Finance
  • Telco

Navigation

  • About
  • Advertise
  • Privacy & Policy
  • Contact

Subscribe to Our Newsletter

  • About
  • Advertise
  • Privacy & Policy
  • Contact

Copyright © 2025 | Powered By Porpholio

No Result
View All Result
  • News
  • Industries
    • Enterprise IT
    • AI & ML
    • Cybersecurity
    • Finance
    • Telco
  • Brand Hub
    • Lifesight
  • Blogs

Copyright © 2025 | Powered By Porpholio