Get your free personalized podcast brief

We scan new podcasts and send you the top 5 insights daily.

Unlike the past where Cisco could build general-purpose silicon for all customers, the immense and specific demands of AI workloads from hyperscalers require custom chip designs. Each major cloud provider effectively becomes a unique market demanding bespoke technology, fundamentally changing the hardware design process.

Related Insights

Cisco's SVP Vijoy Pandey reframes the company's core identity as enabling horizontal 'scale-out' through distributed systems. This directly contrasts with the dominant AI trend of 'scaling up' by creating ever-larger, monolithic models, positioning Cisco to power a future of collaborative, distributed AI.

With AI infrastructure spend topping $100B annually, hyperscalers like Amazon and Google are vertically integrating. They now manage everything from data center construction and micro-nuclear power to designing their own custom chips. For them, custom silicon has become a 'rounding error' in their budget and a key strategy to optimize costs.

Tech giants often initiate custom chip projects not with the primary goal of mass deployment, but to create negotiating power against incumbents like NVIDIA. The threat of a viable alternative is enough to secure better pricing and allocation, making the R&D cost a strategic investment.

Google is abandoning its single-line TPU strategy, now working with both Broadcom and MediaTek on different, specialized TPU designs. This reflects an industry-wide realization that no single chip can be optimal for the diverse and rapidly evolving landscape of AI tasks.

CoreWeave argues that large tech companies aren't just using them to de-risk massive capital outlays. Instead, they are buying a superior, purpose-built product. CoreWeave’s infrastructure is optimized from the ground up for parallelized AI workloads, a fundamental shift from traditional cloud architecture.

For a hyperscaler, the main benefit of designing a custom AI chip isn't necessarily superior performance, but gaining control. It allows them to escape the supply allocations dictated by NVIDIA and chart their own course, even if their chip is slightly less performant or more expensive to deploy.

Specialized AI cloud providers like Nebius don't aim to push alternative chips like AMD or TPUs. Instead, they are "market catchers," responding directly to overwhelming customer demand, which is currently focused entirely on NVIDIA. This demand-driven approach dictates their hardware strategy.

OpenAI's compute deal with Cerebras, alongside deals with AMD and Nvidia, shows that hyperscalers are aggressively diversifying their AI chip supply. This creates a massive opportunity for smaller, specialized silicon teams, heralding a new competitive era reminiscent of the PC wars.

The current 2-3 year chip design cycle is a major bottleneck for AI progress, as hardware is always chasing outdated software needs. By using AI to slash this timeline, companies can enable a massive expansion of custom chips, optimizing performance for many at-scale software workloads.

CEO Chuck Robbins credits the 2016 acquisition of Israeli silicon company Leaba as the critical move that allows Cisco to compete for hyperscaler and AI business. This in-house capability to design high-performance networking silicon differentiates them from competitors reliant on generic merchant silicon, giving them a key strategic advantage.