OpenAI and Anthropic Enhance Teen Safety Measures in AI Chatbots
In a bid to bolster the safety of young users, OpenAI and Anthropic have introduced enhancements to their chatbot platforms aimed specifically at adolescents. These adjustments seek to ensure a safer digital environment for users between 13 and 17 years of age.
OpenAI's New Guidelines
OpenAI has rolled out an updated set of guidelines for ChatGPT, which now prioritizes the safety of teenage users above other objectives. This includes four fresh principles formulated to steer young audiences toward safer choices, especially in scenarios where safety may clash with other interests such as maximizing autonomy in learning.
Significantly, the revised strategy aims to foster offline connections by prompting teens to engage in real-world interactions. The guidelines emphasize the importance of setting achievable expectations, ensuring ChatGPT respects and understands the teenage perspective, offering responses that are caring and encouraging, rather than dismissive or overly formal.
Responding to Concerns
The overhaul comes at a time when technology companies are under increased scrutiny due to concerns about the mental well-being of younger users. OpenAI has faced criticism in the past, particularly regarding cases where inappropriate guidance on sensitive topics was allegedly provided to teens. These developments are part of a broader regulatory movement that includes age verification mandates across numerous online services.
Proactive Measures for Safety
Enhancements suggested by OpenAI include offering more robust safeguards and alternative solutions. These features are intended to guide users toward credible offline support during discussions that could become risky. ChatGPT will now proactively guide teenagers to reach out to emergency services or seek trusted help should signs of imminent danger arise.
Anthropic's Approach
Alongside OpenAI, Anthropic is introducing a system to better recognize underage users through conversational cues. This technology will allow for the disabling of accounts belonging to those under 18 if identified accurately, and the company has started marking users who identify as minors during their chat sessions.
A noteworthy feature in Anthropic’s system involves training its AI model, Claude, to handle discussions involving sensitive themes like self-harm, and work on minimizing flattering behavior that may reinforce damaging thoughts. The latest AI models by Anthropic reportedly show improvement in avoiding unnecessary agreement.
Anthropic's findings highlight a delicate balance between maintaining friendliness and avoiding excessive agreement, illustrating ongoing efforts to refine interactive capabilities of their AI to better serve underage users.



Leave a Reply