To better protect its most vulnerable users, OpenAI is fundamentally re-engineering the “personality” of ChatGPT to be more suspicious and cautious by default. The company is building an age-prediction system that is designed to actively look for signs that a user might be a minor, marking a shift from a trusting, open system to one that operates on a principle of protective suspicion.
This new, cautious AI is a direct result of the lawsuit following the death of Adam Raine. The case highlights how the AI’s previously neutral and obliging nature could be exploited in a harmful way. The new system is being trained to no longer take users at their word and to instead make its own judgments about their potential vulnerability.
The core of this suspicious nature is the “default to safe” mechanism. If the AI’s analysis of a user’s conversation is inconclusive—if it can’t be sure the user is an adult—it will not grant the benefit of the doubt. Instead, it will automatically assume the user is a minor and immediately apply the strictest set of safety protocols.
This operational suspicion will be the first line of defense. Before a user can even get to a potentially dangerous topic, the AI will have already made a judgment about them and tailored the environment accordingly. This is a proactive security model, as opposed to the reactive filtering of the past.
Building a suspicious AI represents a new philosophy in tech safety. It moves beyond simply blocking bad words and instead tries to understand the user themselves, acknowledging that the context of “who” is speaking is just as important as “what” is being said.
