Platforms designed for frictionless speed prevent users from taking a "trust pause"—a moment to critically assess if a person, product, or piece of information is worthy of trust. By removing this reflective step in the name of efficiency, technology accelerates poor decision-making and makes users more vulnerable to misinformation.
Reducing the number of clicks is a misguided metric. A process with eight trivially easy clicks is better than one with two fraught, confusing decisions. Each decision burns cognitive energy and risks making the user feel stupid. The ultimate design goal should be to prevent users from having to think.
The primary problem for AI creators isn't convincing people to trust their product, but stopping them from trusting it too much in areas where it's not yet reliable. This "low trustworthiness, high trust" scenario is a danger zone that can lead to catastrophic failures. The strategic challenge is managing and containing trust, not just building it.
The obsession with removing friction is often wrong. When users have low intent or understanding, the goal isn't to speed them up but to build their comprehension of your product's value. If software asks you to make a decision you don't understand, it makes you feel stupid, which is the ultimate failure.
As AI-generated content and virtual influencers saturate social media, consumer trust will erode, leading to 'Peak Social.' This wave of distrust will drive people away from anonymous influencers and back towards known entities and credible experts with genuine authority in their fields.
Kara Swisher argues that friction is critical for moving forward. The tech industry's obsession with creating seamless, easy experiences is misguided. Hardship and challenges are what lead to growth, cognitive health, and true innovation, whereas frictionless AI can lead to mental atrophy.
Dr. Fei-Fei Li asserts that trust in the AI age remains a fundamentally human responsibility that operates on individual, community, and societal levels. It's not a technical feature to be coded but a social norm to be established. Entrepreneurs must build products and companies where human agency is the source of trust from day one.
We are cognitively wired with a "truth bias," causing us to automatically assume that what we see and hear is true. We only engage in skeptical checking later, if at all. Scammers exploit this default state, ensnaring us before our slower, more deliberate thinking can kick in.
Advanced AI tools like "deep research" models can produce vast amounts of information, like 30-page reports, in minutes. This creates a new productivity paradox: the AI's output capacity far exceeds a human's finite ability to verify sources, apply critical thought, and transform the raw output into authentic, usable insights.
Historically, trust was local (proximity-based) then institutional (in brands, contracts). Technology has enabled a new "distributed trust" era, where we trust strangers through platforms like Airbnb and Uber. This fundamentally alters how reputation is built and where authority lies, moving it from top-down hierarchies to sideways networks.
The era of limited information sources allowed for a controlled, shared narrative. The current media landscape, with its volume and velocity of information, fractures consensus and erodes trust, making it nearly impossible for society to move forward in lockstep.