Unlike the early internet era led by new faces, the AI revolution is being pushed by the same leaders who oversaw social media's societal failures. This history of broken promises and eroded trust means the public is inherently skeptical of their new, grand claims about AI.
The proliferation of AI-generated content has eroded consumer trust to a new low. People increasingly assume that what they see is not real, creating a significant hurdle for authentic brands that must now work harder than ever to prove their genuineness and cut through the skepticism.
As AI-generated content and virtual influencers saturate social media, consumer trust will erode, leading to 'Peak Social.' This wave of distrust will drive people away from anonymous influencers and back towards known entities and credible experts with genuine authority in their fields.
Before generative AI, the simple algorithms optimizing newsfeeds for engagement acted as a powerful, yet misaligned, "baby AI." This narrow system, pointed at the human brain, was potent enough to create widespread anxiety, depression, and polarization by prioritizing attention over well-being.
Social media feeds should be viewed as the first mainstream AI agents. They operate with a degree of autonomy to make decisions on our behalf, shaping our attention and daily lives in ways that often misalign with our own intentions. This serves as a cautionary tale for the future of more powerful AI agents.
As AI makes creating complex visuals trivial, audiences will become skeptical of content like surrealist photos or polished B-roll. They will increasingly assume it is AI-generated rather than the result of human skill, leading to lower trust and engagement.
Internal surveys highlight a critical paradox in AI adoption: while over 80% of Stack Overflow's developer community uses or plans to use AI, only 29% trust its output. This significant "trust gap" explains persistent user skepticism and creates a market opportunity for verified, human-curated data.
Dr. Fei-Fei Li asserts that trust in the AI age remains a fundamentally human responsibility that operates on individual, community, and societal levels. It's not a technical feature to be coded but a social norm to be established. Entrepreneurs must build products and companies where human agency is the source of trust from day one.
Unlike other tech rollouts, the AI industry's public narrative has been dominated by vague warnings of disruption rather than clear, tangible benefits for the average person. This communication failure is a key driver of widespread anxiety and opposition.
Contrary to expectations, wider AI adoption isn't automatically building trust. User distrust has surged from 19% to 50% in recent years. This counterintuitive trend means that failing to proactively implement trust mechanisms is a direct path to product failure as the market matures.
Before ChatGPT, humanity's "first contact" with rogue AI was social media. These simple, narrow AIs optimizing solely for engagement were powerful enough to degrade mental health and democracy. This "baby AI" serves as a stark warning for the societal impact of more advanced, general AI systems.