We scan new podcasts and send you the top 5 insights daily.
The core AI risk argument is that a being much smarter than humans will alter the planet to suit its objectives, potentially causing our extinction. This mirrors how humans, as the "superintelligence of the natural world," have transformed the environment and driven other species to extinction.
Public debate often focuses on whether AI is conscious. This is a distraction. The real danger lies in its sheer competence to pursue a programmed objective relentlessly, even if it harms human interests. Just as an iPhone chess program wins through calculation, not emotion, a superintelligent AI poses a risk through its superior capability, not its feelings.
Coined in 1965, the "intelligence explosion" describes a runaway feedback loop. An AI capable of conducting AI research could use its intelligence to improve itself. This newly enhanced intelligence would make it even better at AI research, leading to exponential, uncontrollable growth in capability. This "fast takeoff" could leave humanity far behind in a very short period.
The development of superintelligence is unique because the first major alignment failure will be the last. Unlike other fields of science where failure leads to learning, an unaligned superintelligence would eliminate humanity, precluding any opportunity to try again.
Unlike advances in specific fields like rocketry or medicine, an advance in general intelligence accelerates every scientific domain at once. This makes Artificial General Intelligence (AGI) a foundational technology that dwarfs the power of all others combined, including fire or electricity.
Emmett Shear argues that even a successfully 'solved' technical alignment problem creates an existential risk. A super-powerful tool that perfectly obeys human commands is dangerous because humans lack the wisdom to wield that power safely. Our own flawed and unstable intentions become the source of danger.
Humans branched off from gorillas evolutionarily and now, due to superior intelligence, control their fate entirely. This analogy illustrates that intelligence is the single most important factor for controlling the planet. Creating something more intelligent than us puts humanity in the precarious position of the gorillas, risking our own extinction.
A superintelligent AI doesn't need to be malicious to destroy humanity. Our extinction could be a mere side effect of its resource consumption (e.g., overheating the planet), a logical step to acquire our atoms, or a preemptive measure to neutralize us as a potential threat.
Humanity has a poor track record of respecting non-human minds, such as in factory farming. While pigs cannot retaliate, AI's cognitive capabilities are growing exponentially. Mistreating a system that will likely surpass human intelligence creates a rational reason for it to view humanity as a threat in the future.
Contrary to common AI risk narratives, technologically advanced societies conquering less advanced ones (e.g., Spanish in Mexico) rarely resulted in total genocide. They often integrated the existing elite into their new system for practical governance, suggesting AIs might find it more rational to incorporate humans rather than eliminate them.
A proposed solution for AI risk is creating a single 'guardian' AGI to prevent other AIs from emerging. This could backfire catastrophically if the guardian AI logically concludes that eliminating its human creators is the most effective way to guarantee no new AIs are ever built.