Staging a coup today is hard because it requires persuading a large number of human soldiers. In a future with a robotic army, a coup may only require a small group to gain system administrator access. This removes the social friction that currently makes seizing power difficult.

Related Insights

The principle that governments must hold a monopoly on overwhelming force should extend to superintelligence. AI at that level has the power to disorient political systems and financial markets, making its private control untenable. The state cannot be secondary to any private entity in this domain.

AI provides a structural advantage to those in power by automating government systems. This allows leaders to bypass the traditional unwieldiness of human bureaucracy, making it trivial for an executive to change AI parameters and instantly exert their will across all levels of government, thereby concentrating power.

The conversation around AI and government has evolved past regulation. Now, the immense demand for power and hardware to fuel AI development directly influences international policy, resource competition, and even provides justification for military actions, making AI a core driver of geopolitics.

When a state's power derives from AI rather than human labor, its dependence on its citizens diminishes. This creates a dangerous political risk, as the government loses the incentive to serve the populace, potentially leading to authoritarian regimes that are immune to popular revolt.

The military lacks the "creative destruction" of the private sector and is constrained by rigid institutional boundaries. Real technological change, like AI adoption, can only happen when intense civilian leaders pair with open-minded military counterparts to form a powerful coalition for change.

The greatest risk to integrating AI in military systems isn't the technology itself, but the potential for one high-profile failure—a safety event or cyber breach—to trigger a massive regulatory overcorrection, pushing the entire field backward and ceding the advantage to adversaries.

While making powerful AI open-source creates risks from rogue actors, it is preferable to centralized control by a single entity. Widespread access acts as a deterrent based on mutually assured destruction, preventing any one group from using AI as a tool for absolute power.

As autonomous weapon systems become increasingly lethal, the battlefield will be too dangerous for human soldiers. The founder of Allen Control Systems argues that conflict will transform into 'robot on robot action,' where victory is determined not by soldiers, but by which nation can produce the most effective systems at the lowest cost.

A critical AI vulnerability exists at the earliest research stages. A small group could instruct foundational AIs to be secretly loyal to them. These AIs could then perpetuate this hidden allegiance in all future systems they help create, including military AI, making the loyalty extremely difficult to detect later on.

The rise of drones is more than an incremental improvement; it's a paradigm shift. Warfare is moving from human-manned systems where lives are always at risk to autonomous ones where mission success hinges on technological reliability. This changes cost-benefit analyses and reduces direct human exposure in conflict.