We scan new podcasts and send you the top 5 insights daily.
Instead of competing on speed and energy alone, Normal Computing is designing ASICs that introduce noise as a third optimization vector. These chips are ideal for probabilistic workloads like diffusion models, which are inherently noisy and approximate, mapping the software's physics to the hardware's.
The brain's hardware limitations, like slow and stochastic neurons, may actually be advantages. These properties seem perfectly suited for probabilistic inference algorithms that rely on sampling—a task that requires explicit, computationally-intensive random number generation in digital systems. Hardware and algorithm are likely co-designed.
To achieve 1000x efficiency, Unconventional AI is abandoning the digital abstraction (bits representing numbers) that has defined computing for 80 years. Instead, they are co-designing hardware and algorithms where the physics of the substrate itself defines the neural network, much like a biological brain.
Digital computing, the standard for 80 years, is too power-hungry for scalable AI. Unconventional AI's Naveen Rao is betting on analog computing, which uses physics to perform calculations, as a more energy-efficient substrate for the unique demands of intelligent, stochastic workloads.
Designing custom AI hardware is a long-term bet. Google's TPU team co-designs chips with ML researchers to anticipate future needs. They aim to build hardware for the models that will be prominent 2-6 years from now, sometimes embedding speculative features that could provide massive speedups if research trends evolve as predicted.
We are building AI, a fundamentally stochastic and fuzzy system, on top of highly precise and deterministic digital computers. Unconventional AI founder Naveen Rao argues this is a profound mismatch. The goal is to build a new computing substrate—analog circuits—that is isomorphic to the nature of intelligence itself.
Model architecture decisions directly impact inference performance. AI company Zyphra pre-selects target hardware and then chooses model parameters—such as a hidden dimension with many powers of two—to align with how GPUs split up workloads, maximizing efficiency from day one.
A key technical risk for space compute is chip failure from radiation. However, Starcloud found that the stochastic nature of GPU inference workloads means a radiation-induced bit flip alters the specific output (e.g., a slightly different poem) but doesn't degrade its overall quality, significantly de-risking the hardware.
OpenAI is designing its custom chip for flexibility, not just raw performance on current models. The team learned that major 100x efficiency gains come from evolving algorithms (e.g., dense to sparse transformers), so the hardware must be adaptable to these future architectural changes.
The rise of agent orchestration using specialized, open-source models will drive demand for custom ASICs. Jerry Murdock argues that putting a model on a dedicated chip will be far cheaper and more tunable for specific workloads than using expensive, general-purpose GPUs like Nvidia's, spurring a hardware shift.
The current 2-3 year chip design cycle is a major bottleneck for AI progress, as hardware is always chasing outdated software needs. By using AI to slash this timeline, companies can enable a massive expansion of custom chips, optimizing performance for many at-scale software workloads.