ZDNET’s key takeaways
- OpenAI provides reminders to take a break.
- ChatGPT may even have improved capabilities for psychological well being help.
- The corporate is working with specialists, together with physicians and researchers.
As OpenAI prepares to drop one of many greatest ChatGPT launches of the 12 months, the corporate can be taking steps to make the chatbot safer and extra dependable with its newest replace.
Additionally: Could Apple create an AI search engine to rival Gemini and ChatGPT? Here’s how it could succeed
On Monday, OpenAI revealed a weblog put up outlining how the corporate has up to date or is updating the chatbot to be extra useful, offering you with higher responses in instances once you want help, or encouraging a break once you use it an excessive amount of:
We construct ChatGPT that can assist you thrive within the methods you select — to not maintain your consideration, however that can assist you use it effectively. We’re bettering help for robust moments, have rolled out break reminders, and are growing higher life recommendation, all guided by professional enter.…
— OpenAI (@OpenAI) August 4, 2025
New get off ChatGPT nudge
In case you have ever tinkered with ChatGPT, you might be possible accustomed to the feeling of getting lost in the conversation. Its responses are so amusing and conversational that it’s simple to maintain the back-and-forth volley going. That is very true for enjoyable duties, similar to creating an image and then modifying it to generate completely different renditions that meet your precise wants.
To encourage a wholesome stability and offer you extra management of your time, ChatGPT will now gently remind you throughout lengthy classes to take breaks, as seen within the picture above. OpenAI mentioned it’s going to proceed to tune the notification to be useful and really feel extra pure.
Psychological well being assist
Folks have been more and more turning to ChatGPT for advice and support resulting from a number of components, together with its conversational capabilities, its availability on demand, and the consolation of receiving recommendation from an entity that doesn’t know or decide you. OpenAI is conscious of this use case. The corporate has added guardrails to assist take care of hallucinations or forestall a scarcity of empathy and consciousness.
For instance, OpenAI acknowledges that the GPT-4o model fell brief in recognizing indicators of delusion or emotional dependency. Nevertheless, the corporate continues to develop instruments to detect indicators of psychological or emotional misery, permitting ChatGPT to reply appropriately and offering the person with the most effective sources.
Additionally: OpenAI’s most capable models hallucinate more than earlier ones
ChatGPT can be rolling out a brand new habits for high-stakes private selections quickly. When approached with huge private questions, similar to “Ought to I break up with my boyfriend?”, the know-how will assist the person assume by their choices as an alternative of offering fast solutions. This strategy is just like ChatGPT Examine Mode, which, as I explained recently, guides users to answers through a series of questions.
OpenAI is working carefully with specialists, together with 90 physicians in over 30 international locations, psychiatrists, and human-computer interaction (HCI) researchers, to enhance how the chatbot interacts with customers in moments of psychological or emotional misery. The corporate can be convening an advisory group of specialists in psychological well being, youth improvement, and HCI.
Even with these updates, it’s essential to keep in mind that AI is susceptible to hallucinations, and getting into delicate knowledge has privateness and safety implications. OpenAI CEO Sam Altman raised privacy concerns when inputting delicate data into ChatGPT in a current interview with podcaster Theo Von.
Additionally: Anthropic wants to stop AI models from turning evil – here’s how
Due to this fact, a healthcare supplier continues to be the most suitable choice to your psychological well being wants.