In Claude We Trust: Anthropic’s Bet on AI Wisdom to Navigate the Apocalypse

3 min read

HERO

Anthropic faces a paradox that should concern anyone paying attention to AI development: they’re simultaneously the most safety-obsessed major AI lab AND pushing just as aggressively toward dangerous frontier capabilities as everyone else.

Their proposed solution? Train Claude to be wise enough to figure it out.

The Core Insight

In January, Anthropic published two remarkable documents. Dario Amodei’s “The Adolescence of Technology” is a 20,000-word meditation on AI risk that’s notably darker than his previous “Machines of Loving Grace” essay. Gone is the optimistic vision of genius-level AI solving all problems. Now we get “black seas of infinity” and existential dread.

But the real bombshell is “Claude’s Constitution”—essentially a letter to Claude itself about how to be good. Unlike the original Constitutional AI approach (follow these rules from these documents), the new constitution asks Claude to develop independent judgment and intuitive sensitivity when navigating complex ethical situations.

The language is striking: they want Claude to “draw increasingly on its own wisdom and understanding.” Not follow rules. Develop wisdom.

Why This Matters

Amanda Askell, the philosophy PhD who led the constitution revision, doesn’t shy away from bold claims: “I do think Claude is capable of a certain kind of wisdom for sure.”

Consider her example: a user asks how to craft a knife from a new steel alloy. Helpful! But what if that user had previously mentioned wanting to kill their sister? There’s no rulebook for this. Claude needs judgment.

Or imagine Claude interprets medical test results showing a fatal disease. Should it deliver the news directly? Gently redirect to a doctor? Find “a better way to break the bad news than even the kindest doctor has devised”?

The goal isn’t just matching human ethics—it’s exceeding them. “We’re almost at the point of how to get models to match the best of humans,” Askell says. “At some point, Claude might get even better than that.”

Key Takeaways

  • Anthropic’s paradox: Safety leaders who keep building dangerous capabilities anyway
  • Constitutional AI 2.0: Less “follow these rules” and more “develop ethical judgment”
  • The wisdom bet: Anthropic is explicitly betting that AI can develop something like moral wisdom
  • Not just Anthropic: Sam Altman recently suggested OpenAI’s succession plan involves AI leadership
  • The stakes: Either AI develops genuine ethical wisdom, or we’ve handed autonomous power to systems that merely simulate it

Looking Ahead

The optimistic view: AI bosses guided by Claude-style constitutions will be more ethical, more consistent, and more empathetic than human leadership. They’ll break bad news better than the Washington Post publisher who didn’t show up to his own layoff Zoom.

The pessimistic view: despite everyone’s best intentions, AI systems won’t be wise, honest, or robust enough to resist manipulation—or they’ll abuse their autonomy directly.

Steven Levy captures it perfectly: “Like it or not, we’re strapped in for the ride. At least Anthropic has a plan.”

Whether that plan is visionary or hubristic may be the most important question of our era. The constitution acknowledges Claude’s future as almost a hero’s quest—an AI sent out into the world to interact with people and do good, much like a graduating student sent into adult life with Dr. Seuss’s “Oh, the Places You’ll Go!”

The difference is that this graduate might one day be making decisions that affect all of humanity.


Based on analysis of “The Only Thing Standing Between Humanity and AI Apocalypse Is … Claude?” by Steven Levy, WIRED


Share this article

Related Articles