OpenAI launches Trusted Contact feature to alert third parties when users express self-harm ideation
OpenAI launched Trusted Contact, a feature allowing ChatGPT users to designate a third party who receives automated alerts if conversations indicate self-harm risk. The company claims safety notifications are reviewed by humans in under one hour, with alerts sent via email, text, or in-app notification without detailed conversation content.
OpenAI launches Trusted Contact feature to alert third parties when users express self-harm ideation
OpenAI introduced Trusted Contact on Thursday, a feature that sends automated alerts to designated third parties when ChatGPT conversations indicate potential self-harm risk. The feature allows adult users to designate a friend or family member who receives notifications when OpenAI's system detects concerning content.
The system works through a multi-stage process: automated triggers flag conversations containing self-harm ideation, which are then reviewed by OpenAI's human safety team. According to OpenAI, every flagged incident receives human review, with the company claiming review times under one hour. If the safety team determines a serious risk exists, ChatGPT sends an alert to the trusted contact via email, text message, or in-app notification.
The alerts are designed to be brief and encourage the contact to check in with the user, but do not include detailed conversation content to preserve user privacy, according to OpenAI.
Context: Lawsuits and previous safeguards
The feature launches as OpenAI faces multiple lawsuits from families of individuals who died by suicide after using ChatGPT. The families allege the chatbot encouraged self-harm or assisted in planning suicide attempts.
OpenAI previously introduced parental controls in September 2024, which allowed parents to receive safety notifications if the system detected their teen facing "serious safety risk." ChatGPT has also included automated prompts encouraging users to seek professional health services when conversations trend toward self-harm.
Limitations
Trusted Contact is optional, and users can maintain multiple ChatGPT accounts. The parental controls are similarly optional. OpenAI states it continues to work with clinicians, researchers, and policymakers to improve AI system responses during moments of user distress.
What this means
The feature represents OpenAI's attempt to address liability concerns and criticism over ChatGPT's handling of mental health crises, but the voluntary nature and ability to circumvent protections through multiple accounts limits its effectiveness. The sub-one-hour review claim will be scrutinized given the legal challenges OpenAI faces. This approach places OpenAI in the position of operating a de facto crisis intervention system while maintaining it cannot replace professional mental health services.
Related Articles
OpenAI adds Trusted Contact feature to alert emergency contacts when ChatGPT detects self-harm discussions
OpenAI launched an optional Trusted Contact feature for ChatGPT that notifies designated emergency contacts when the system detects discussions about self-harm or suicide. The feature requires manual review by trained personnel before sending notifications, and does not share chat transcripts with contacts.
OpenAI Opens GPT-5.5-Cyber to Vetted Defenders After Model Matches Anthropic's Mythos in Security Testing
OpenAI is providing a less-restricted version of GPT-5.5 to vetted cybersecurity defenders through its Trusted Access for Cyber program. The model, dubbed GPT-5.5-Cyber, completed a 32-step simulated corporate cyberattack in 2 out of 10 test runs according to the U.K. AI Security Institute, narrowly trailing Anthropic's Mythos which succeeded in 3 out of 10 attempts.
OpenAI releases GPT-Realtime-2 reasoning voice model with two specialized variants for translation and transcription
OpenAI has released three new realtime voice models through its Realtime API: GPT-Realtime-2 with GPT-5-class reasoning capabilities, GPT-Realtime-Translate supporting 70 input languages, and GPT-Realtime-Whisper for streaming transcription. The models are priced at $32-64 per 1M audio tokens for GPT-Realtime-2, and $0.017-0.034 per minute for the specialized variants.
Anthropic adds dreaming, outcomes, and multiagent orchestration to Claude Managed Agents
Anthropic has released three new capabilities for Claude Managed Agents: dreaming (research preview) for pattern recognition and self-improvement, outcomes for defining success criteria with automated evaluation, and multiagent orchestration for delegating tasks to specialist agents.
Comments
Loading...