/
© 2026 RiffOn. All rights reserved.
  1. Decoder with Nilay Patel
  2. The tiny team trying to keep AI from destroying everything
The tiny team trying to keep AI from destroying everything

The tiny team trying to keep AI from destroying everything

Decoder with Nilay Patel · Dec 4, 2025

Anthropic's unique 'Societal Impacts' team studies AI's 'inconvenient truths,' balancing its safety-first brand with business and political pressures.

AI Labs Create Internal Safety Teams to Argue Against Federal Regulation

The existence of internal teams like Anthropic's "Societal Impacts Team" serves a dual purpose. Beyond their stated mission, they function as a strategic tool for AI companies to demonstrate self-regulation, thereby creating a political argument that stringent government oversight is unnecessary.

The tiny team trying to keep AI from destroying everything thumbnail

The tiny team trying to keep AI from destroying everything

Decoder with Nilay Patel·3 months ago

AI Firm Anthropic Uses Its 'Safety-First' Reputation as a Key Business Strategy

Anthropic's commitment to AI safety, exemplified by its Societal Impacts team, isn't just about ethics. It's a calculated business move to attract high-value enterprise, government, and academic clients who prioritize responsibility and predictability over potentially reckless technology.

The tiny team trying to keep AI from destroying everything thumbnail

The tiny team trying to keep AI from destroying everything

Decoder with Nilay Patel·3 months ago

Anthropic's Safety Team Struggles to Translate Damning Research into Product Changes

Despite having the freedom to publish "inconvenient truths" about AI's societal harms, Anthropic's Societal Impacts team expresses a desire for their research to have a more direct, trackable impact on the company's own products. This reveals a significant gap between identifying problems and implementing solutions.

The tiny team trying to keep AI from destroying everything thumbnail

The tiny team trying to keep AI from destroying everything

Decoder with Nilay Patel·3 months ago

Anthropic Balances Its 'Safe AI' Brand with Fears of Being Labeled 'Woke' by Trump

Anthropic faces a critical dilemma. Its reputation for safety attracts lucrative enterprise clients, but this very stance risks being labeled "woke" by the Trump administration, which has banned such AI in government contracts. This forces the company to walk a fine line between its brand identity and political reality.

The tiny team trying to keep AI from destroying everything thumbnail

The tiny team trying to keep AI from destroying everything

Decoder with Nilay Patel·3 months ago

Anthropic's CEO Justifies Taking Saudi Money to Stay Competitive with OpenAI

CEO Dario Amodei rationalized accepting Saudi investment by arguing it's necessary to remain at the forefront of AI development. He stated that running a business on the principle that "no bad person should ever benefit from our success" is difficult, highlighting how competitive pressures force even "safety-first" companies into ethical compromises.

The tiny team trying to keep AI from destroying everything thumbnail

The tiny team trying to keep AI from destroying everything

Decoder with Nilay Patel·3 months ago

Enterprise Focus May Insulate AI Firms from Consumer-Driven Culture Wars

Unlike AI companies targeting the consumer market, Anthropic's success with enterprise-focused products like "Claude Code" could shield it from the intense political scrutiny that plagued social media platforms. By selling to businesses, it avoids the unpredictable dynamics of the consumer internet and direct engagement with hot-button social issues.

The tiny team trying to keep AI from destroying everything thumbnail

The tiny team trying to keep AI from destroying everything

Decoder with Nilay Patel·3 months ago