Get your free personalized podcast brief

We scan new podcasts and send you the top 5 insights daily.

If agents in a vast universe use non-causal decision theories, one agent's choice to fund a "consensus good" provides evidence that their correlated copies across the multiverse will do the same. This turns a small personal sacrifice into a cosmic-scale collective action, solving cooperation problems without a central enforcer.

Related Insights

In multi-agent simulations, if agents use a shared source of randomness, they can achieve stable equilibria. If they use private randomness, coordinating punishment becomes nearly impossible because one agent cannot verify if another's defection was malicious or a justified response to a third party's actions.

Lakhiani cites the phenomenon where monkeys on separate islands adopt a new skill once a critical mass learns it on one island. He posits this as potential evidence for quantum-level information exchange, suggesting a collective consciousness or connection within a species that transcends physical distance.

If AI alignment turns out to be easy, it would likely be because morality is not a human construct but an objective feature of reality. In this scenario, any sufficiently intelligent agent would logically deduce that cooperation and preserving humanity are optimal strategies, regardless of its initial programming.

When multiple AI agents work as an ensemble, they can collectively suppress hallucinations. By referencing a shared knowledge graph as ground truth, the group can form a consensus, effectively ignoring the inaccurate output from one member and improving overall reliability.

In program equilibrium, players submit computer programs instead of actions. These programs can read each other's source code, allowing them to verify cooperative intent and overcome dilemmas like the Prisoner's Dilemma, which is impossible in standard game theory.

Moving beyond isolated AI agents requires a framework mirroring human collaboration. This involves agents establishing common goals (shared intent), building a collective knowledge base (shared knowledge), and creating novel solutions together (shared innovation).

Rather than relying on a single AI, an agentic system should use multiple, different AI models (e.g., auditor, tester, coder). By forcing these independent agents to agree, the system can catch malicious or erroneous behavior from a single misaligned model.

A key finding is that almost any outcome better than mutual punishment can be a stable equilibrium (a "folk theorem"). While this enables cooperation, it creates a massive coordination problem: with so many possible "good" outcomes, agents may fail to converge on the same one, leading to suboptimal results.

Challenging the binary view of free will, a new mathematical model could show that individual agents (us) and the larger conscious systems they form can both possess genuine free will simultaneously, operating at different but interconnected scales.

A simple way for AIs to cooperate is to simulate each other and copy the action. However, this creates an infinite loop if both do it. The fix is to introduce a small probability (epsilon) of cooperating unconditionally, which guarantees the simulation chain eventually terminates.

Non-Causal Decision Theory Could Spontaneously Align Agents Across the Multiverse | RiffOn