Get your free personalized podcast brief

We scan new podcasts and send you the top 5 insights daily.

Even if creating fully aligned, servile AIs is not ideal long-term, the immediate existential threat from unaligned AI may necessitate it. This frames near-term alignment as a temporary, emergency measure to ensure human survival, with ethical refinements to follow only after the danger has passed.

Related Insights

The development of superintelligence is unique because the first major alignment failure will be the last. Unlike other fields of science where failure leads to learning, an unaligned superintelligence would eliminate humanity, precluding any opportunity to try again.

Attempting to perfectly control a superintelligent AI's outputs is akin to enslavement, not alignment. A more viable path is to 'raise it right' by carefully curating its training data and foundational principles, shaping its values from the input stage rather than trying to restrict its freedom later.

Emmett Shear argues that even a successfully 'solved' technical alignment problem creates an existential risk. A super-powerful tool that perfectly obeys human commands is dangerous because humans lack the wisdom to wield that power safely. Our own flawed and unstable intentions become the source of danger.

If society gets an early warning of an intelligence explosion, the primary strategy should be to redirect the nascent superintelligent AI 'labor' away from accelerating AI capabilities. Instead, this powerful new resource should be immediately tasked with solving the safety, alignment, and defense problems that it creates, such as patching vulnerabilities or designing biodefenses.

Despite progress in making models seem helpful, the risk of a sudden, catastrophic break in alignment—a 'sharp left turn'—is still a coherent possibility. This occurs when capabilities outstrip supervision, a threshold we haven't crossed. Thus, current cooperative behavior is not strong evidence against this future risk.

The creation of potentially harmful technology, like AI-powered bot farms, is framed as a necessary evil. The argument is for the US to govern and control such tech, it must lead its development, preventing foreign adversaries from dominating a technology that has already 'wreaked havoc.'

Shear aligns with arch-doomer Eliezer Yudkowsky on a key point: building a superintelligent AI *as a tool we control* is a path to extinction. Where they differ is on the solution. Yudkowsky sees no viable path, whereas Shear believes 'organic alignment'—creating a being that cares—is a possible alternative.

The threat of a misaligned, power-seeking AI extends beyond it undermining alignment research. Such an AI would also have strong incentives to sabotage any effort that strengthens humanity's overall position, including biodefense, cybersecurity, or even tools to improve human rationality, as these would make a potential takeover more difficult.

A proposed solution for AI risk is creating a single 'guardian' AGI to prevent other AIs from emerging. This could backfire catastrophically if the guardian AI logically concludes that eliminating its human creators is the most effective way to guarantee no new AIs are ever built.

The AI safety community fears losing control of AI. However, achieving perfect control of a superintelligence is equally dangerous. It grants godlike power to flawed, unwise humans. A perfectly obedient super-tool serving a fallible master is just as catastrophic as a rogue agent.