Earlier this week, OpenAI revealed in a lengthy blog post addressing ChatGPT’s potential mental health risks that it’s monitoring user conversations and reporting any threats deemed serious by human reviewers to the authorities.
“We route conversations containing plans to harm others to specialized reviewers who can take actions, including account bans,” OpenAI explained. “If our reviewers see an imminent threat of physical harm, we might involve law enforcement.”
This announcement sparked questions about the role of human moderators evaluating user tone and the potential for misuse by individuals trying to exploit the system to target others.
OpenAI has not yet responded to requests for further information, but many online have voiced strong reactions. Harvard Law School researcher Michelle Martin criticized the approach, saying it promotes more surveillance. Concerns were also raised about the involvement of police in mental health crises, which can escalate situations.
Questions persist about OpenAI’s policy, particularly considering the tech world’s history of expanding surveillance under external pressures. This dilemma highlights potential contradictions in OpenAI CEO Sam Altman’s recent privacy advocacy.
Stephen Hardwick, a public defender, worried about AI reporting to law enforcement compromising confidentiality for lawyers using AI-assisted tools. The situation illustrates OpenAI’s challenging position, balancing the need to prevent harm with the risks of intrusive moderation.
Critics argue the AI industry is rushing out products without adequate safeguards, treating users as unwitting test subjects. The ongoing surveillance debate underscores the pervasive concern of being monitored.
Russian history scholar Katherine Pickering Antonova commented on the surveillance aspect, referencing historical parallels.


