Today’s episode takes a deep dive into one of the most high-profile AI controversies in recent months: the debate surrounding Grok AI. From reports that the system generated harmful content — including material related to non-consensual intimate imagery — to broader concerns about safety guardrails, moderation decisions, and accountability, the situation has sparked serious questions across the tech policy landscape.
To help us make sense of what happened and why it matters, we’re joined by Kate Ruane, Director of the Free Expression Project at the Center for Democracy & Technology, and Riana Pfefferkorn, Policy Fellow at the Stanford Institute for Human-Centered Artificial Intelligence and a CDT Non-Resident Fellow.
Together, they examine what the Grok episode reveals about the mounting pressures facing AI companies, the real-world limits of existing safeguards, and the difficult balancing act between preventing harm and protecting free expression. This conversation explores the evolving responsibilities of AI developers — and what this moment signals for the future of speech, safety, and accountability in an AI-driven world. Produced by Drew Courtney and edited by Jake Kauffman.