Big Tech’s Quiet Retreat from AI Regulation
From Urgency to Evasion
Just months ago, executives from AI giants were loudly urging governments to regulate artificial intelligence, warning of existential risks and ethical crises. But that tune is changing. According to a Washington Post investigation, many of the same leaders now appear far less enthusiastic about enforceable guardrails, opting instead for voluntary commitments and closed-door conversations. The shift reflects an evolving calculus: while appearing responsible may have public relations benefits, legally binding regulation could slow innovation and yield competitive disadvantages. With billions at stake in a rapidly advancing field, talking about the dangers of AI is easy—but doing something about it is proving politically and commercially inconvenient.
Strategic Silence and Soft Commitments
As momentum builds in Washington and Brussels for harder rules around AI, tech firms are adjusting their strategies. Public testimony has softened, and lobbying efforts focus on vague principles rather than concrete obligations. The Biden administration’s executive order on AI safety is a rare step forward, but lacks the enforceability that some experts say is urgently needed. Lawmakers report that behind closed doors, many AI leaders now push back against substantive regulation. Meanwhile, voluntary pacts like the one signed by OpenAI, Google, and Anthropic offer optics but little accountability. Alarmingly, even as technology grows more powerful, corporate appetite for real oversight appears to be shrinking.