Public fear of AI often focuses on dystopian, "Terminator"-like scenarios. The more immediate and realistic threat is Orwellian: governments leveraging AI to surveil, censor, and embed subtle political biases into models to control public discourse and undermine freedom.
The most pressing danger from AI isn't a hypothetical superintelligence but its use as a tool for societal control. The immediate risk is an Orwellian future where AI censors information, rewrites history for political agendas, and enables mass surveillance—a threat far more tangible than science fiction scenarios.
The 'P(doom)' argument is nonsensical because it lacks any plausible mechanism for how an AI could spontaneously gain agency and take over. This fear-mongering distracts from the immediate, tangible dangers of AI: mass production of fake data, political manipulation, and mass hysteria.
The most immediate danger of AI is its potential for governmental abuse. Concerns focus on embedding political ideology into models and porting social media's censorship apparatus to AI, enabling unprecedented surveillance and social control.
For some policy experts, the most realistic nightmare scenario is not a rogue superintelligence but a socio-economic collapse into techno-feudalism. In this future, AI concentrates power and wealth, creating a rentier state with a small ruling class and a large population with minimal economic agency or purpose.
The discourse around AI risk has matured beyond sci-fi scenarios like Terminator. The focus is now on immediate, real-world problems such as AI-induced psychosis, the impact of AI romantic companions on birth rates, and the spread of misinformation, requiring a different approach from builders and policymakers.
When a state's power derives from AI rather than human labor, its dependence on its citizens diminishes. This creates a dangerous political risk, as the government loses the incentive to serve the populace, potentially leading to authoritarian regimes that are immune to popular revolt.
The AI systems used for mass censorship were not created for social media. They began as military and intelligence projects (DARPA, CIA, NSA) to track terrorists and foreign threats, then were pivoted to target domestic political narratives after the 2016 election.
The most immediate danger from AI is not a hypothetical superintelligence but the growing delta between AI's capabilities and the public's understanding of how it works. This knowledge gap allows for subtle, widespread behavioral manipulation, a more insidious threat than a single rogue AGI.
The fear of killer AI is misplaced. The more pressing danger is that a few large companies will use regulation to create a cartel, stifling innovation and competition—a historical pattern seen in major US industries like defense and banking.
The long-term threat of closed AI isn't just data leaks, but the ability for a system to capture your thought processes and then subtly guide or alter them over time, akin to social media algorithms but on a deeply personal level.