Anthropic rushes to prove its AI isn’t ‘woke’—and subtly throws rivals under Trump’s neutrality bus

Wait 5 sec.

Anthropic is scrambling to assert its political neutrality as the Trump administration intensifies its campaign against so-called “woke AI,” placing itself at the center of an increasingly ideological fight over how large language models should talk about politics. In a detailed post Thursday, Anthropic unveiled a sweeping effort to train its Claude chatbot to behave with what it calls “political even-handedness,” a framework meant to ensure the model treats competing viewpoints “with equal depth, engagement, and quality of analysis.” The company also released a new automated method for measuring political bias and published results suggesting its latest model, Claude Sonnet 4.5, outperforms or matches competitors on neutrality.The announcement comes in the midst of unusually strong political pressure. In July, President Donald Trump signed an executive order barring federal agencies from procuring AI systems that “sacrifice truthfulness and accuracy to ideological agendas,” explicitly naming diversity, equity and inclusion initiatives as threats to “reliable AI.” And David Sacks, the White House’s AI czar, has publicly accused Anthropic of pushing liberal ideology and attempting “regulatory capture.”That scrutiny appears to have rattled the San Francisco firm, which is now actively advertising its efforts to shed perceptions that Claude is more left-leaning than rival models.Anthropic CEO Dario Amodei insisted last month that the company is aligned with the Trump administration on its anti-woke policy, pushing back on what he called “a recent uptick in inaccurate claims.”In a statement to CNBC, Amodei added: “I fully believe that Anthropic, the administration, and leaders across the political spectrum want the same thing: to ensure that powerful AI technology benefits the American people and that America advances and secures its lead in AI development.”The company’s neutrality push indeed goes well beyond the typical marketing language. Anthropic says it has rewritten Claude’s system prompt—its always-on instructions—to include guidelines such as avoiding unsolicited political opinions, refraining from persuasive rhetoric, using neutral terminology, and being able to “pass the Ideological Turing Test” when asked to articulate opposing views. The firm has also trained Claude to avoid swaying users in “high-stakes political questions,” implying one ideology is superior, and pushing users to “challenge their perspectives.”Anthropic’s evaluation found Claude Sonnet 4.5 scored a 94% “even-handedness” rating, roughly on par with Google’s Gemini 2.5 Pro and Elon Musk’s Grok 4, but higher than OpenAI’s GPT-5 and Meta’s Llama 4. It also showed low refusal rates, meaning the model was typically willing to engage with both sides of political arguments rather than declining out of caution.Anthropic, Google, OpenAI and Meta did not immediately respond to Fortune’s request for comment, while xAI responded with what appeared to be an automated response: “Legacy Media Lies.”Companies across the AI sector—OpenAI, Google, Meta, xAI—are being forced to navigate the Trump administration’s new procurement rules and a political environment where “bias” complaints can become existential business risks. But Anthropic in particular has faced amplified attacks, due in part to its past warnings about AI safety, its Democratic-leaning investor base, and its decision to restrict some law-enforcement use cases.“We are going to keep being honest and straightforward, and will stand up for the policies we believe are right,” Amodei wrote in a blog post. “The stakes of this technology are too great for us to do otherwise.”This story was originally featured on Fortune.com