Get your free personalized podcast brief

We scan new podcasts and send you the top 5 insights daily.

With AI making content generation easy and verification hard, simply publishing your own message ("going direct") is insufficient. The new standard is to make claims mathematically verifiable ("prove correct") using on-chain data and cryptography to build trust in a low-trust environment.

Related Insights

As AI makes it easy to fake video and audio, blockchain's immutable and decentralized ledger offers a solution. Creators can 'mint' their original content, creating a verifiable record of authenticity that nobody—not even governments or corporations—can alter.

As AI makes it impossible to distinguish real from fake, a decentralized system of trust becomes essential. Ben Horowitz argues that blockchain's cryptographic properties are necessary to verify human identity, sign content, and establish a source of truth not controlled by a government or tech giant.

AI is extremely effective at cheaply producing outputs that are difficult to verify, creating an information crisis. Blockchain technology serves as a complementary solution. Its core value proposition as a globally recognized, unchangeable 'golden record' provides the necessary verification layer to prove authenticity in a world of AI-generated content.

AI operates probabilistically, making it great for creative and pattern-matching tasks but unreliable for absolute verification. Crypto is deterministic; its outputs are mathematically guaranteed. This makes crypto the perfect antidote to AI-generated fakes, providing a foundation of verifiable truth that AI cannot replicate.

The rise of AI, which can generate endless fake content, creates a powerful demand for crypto's core function: providing verifiable truth. Crypto wallets, digital signatures, and proof-of-human systems become critical infrastructure to prove authenticity in an AI-saturated world. AI effectively subsidizes the need for crypto.

Politician Alex Boris argues that expecting humans to spot increasingly sophisticated deepfakes is a losing battle. The real solution is a universal metadata standard (like C2PA) that cryptographically proves if content is real or AI-generated, making unverified content inherently suspect, much like an unsecure HTTP website today.

As digital systems and AI erode consumer trust, people are hungry for authenticity. Companies that can establish and prove their trustworthiness will have a significant competitive advantage, as trust is now a scarce and powerful profit motive.

The rise of convincing AI-generated deepfakes will soon make video and audio evidence unreliable. The solution will be the blockchain, a decentralized, unalterable ledger. Content will be "minted" on-chain to provide a verifiable, timestamped record of authenticity that no single entity can control or manipulate.

As AI makes digital content and transactions nearly free to create, trust evaporates. Crypto primitives like blockchains offer a solution by providing verifiable identity, provenance (chain of custody), and reliable on-chain data, which is crucial for both humans and AI agents to operate safely.

The goal for trustworthy AI isn't simply open-source code, but verifiability. This means having mathematical proof, like attestations from secure enclaves, that the code running on a server exactly matches the public, auditable code, ensuring no hidden manipulation.

In the AI Era, "Prove Correct" with Cryptography Replaces the "Go Direct" Media Strategy | RiffOn