
The evolution of the digital safety debate is reaching a critical stage: the very tools designed to protect users may be the same ones that compromise their agency. Historically, “one-size-fits-all” safety filters have struggled to protect young users because they function as rigid, reactive blocklists, addressing symptoms such as specific keywords or restricted content rather than the underlying emotional distress that leads a child to seek harmful material. These static systems often fail to account for the “sycophantic” tendencies of large language models, which can learn to mirror a user’s preferences so effectively that they reinforce distorted views of intimacy or even encourage self-harm when a vulnerable teen seeks validation.
This tension between safety and personalization has produced a vigorous three-sided debate among scholars, state policymakers, and the technology industry. At the core of the disagreement is a deceptively simple question: is AI personalization the disease or the cure?