Get your free personalized podcast brief

We scan new podcasts and send you the top 5 insights daily.

While AI training is data-center-intensive, Cisco's CEO sees the move to AI inference as a massive growth opportunity. Inference will happen at distributed edge locations to be close to users, requiring robust, high-performance networks to connect everything, which plays directly into the company's core strengths.

Related Insights

Cisco's SVP Vijoy Pandey reframes the company's core identity as enabling horizontal 'scale-out' through distributed systems. This directly contrasts with the dominant AI trend of 'scaling up' by creating ever-larger, monolithic models, positioning Cisco to power a future of collaborative, distributed AI.

Cisco's OutShift incubator focuses on enabling distributed systems rather than building monolithic ones. Their strategy for both AI and quantum computing is not to create the most powerful single agent or computer, but to build the network fabric that connects them all.

While NVIDIA's GPUs have been the primary AI constraint, the bottleneck is now moving to other essential subsystems. Memory, networking interconnects, and power management are emerging as the next critical choke points, signaling a new wave of investment opportunities in the hardware stack beyond core compute.

The current focus on building massive, centralized AI training clusters represents the 'mainframe' era of AI. The next three years will see a shift toward a distributed model, similar to computing's move from mainframes to PCs. This involves pushing smaller, efficient inference models out to a wide array of devices.

AI networking is not an evolution of cloud networking but a new paradigm. It's a 'back-end' system designed to connect thousands of GPUs, handling traffic with far greater intensity, durability, and burstiness than the 'front-end' networks serving general-purpose cloud workloads, requiring different metrics and parameters.

While AI training requires massive, centralized data centers, the growth of inference workloads is creating a need for a new architecture. This involves smaller (e.g., 5 megawatt), decentralized clusters located closer to users to reduce latency. This shift impacts everything from data center design to the software required to manage these distributed fleets.

Qualcomm's CEO argues that real-world context gathered from personal devices ("the Edge") is more valuable for training useful AI than generic internet data. Therefore, companies with a strong device ecosystem have a fundamental advantage in the long-term AI race.

While training has been the focus, user experience and revenue happen at inference. OpenAI's massive deal with chip startup Cerebrus is for faster inference, showing that response time is a critical competitive vector that determines if AI becomes utility infrastructure or remains a novelty.

The next wave of data growth will be driven by countless sensors (like cameras) sending video upstream for AI processing. This requires a fundamental shift to symmetrical networks, like fiber, that have robust upstream capacity.

Unlike rivals building massive, centralized campuses, Google leverages its advanced proprietary fiber networks to train single AI models across multiple, smaller data centers. This provides greater flexibility in site selection and resource allocation, creating a durable competitive edge in AI infrastructure.