A trend called "tokenmaxxing" is emerging in Silicon Valley, where companies like Meta use leaderboards to track employee AI token usage. This reflects a corporate bet that higher token consumption correlates with increased productivity, turning AI usage into a new, albeit gameable, performance metric for engineers.
Broadcom is solidifying its position as the key alternative to NVIDIA's locked-in ecosystem by becoming the preferred design partner for custom AI chips (ASICs). Its deep partnerships with major players like Anthropic and OpenAI to develop specialized hardware highlight a growing demand for tailored, cost-efficient silicon.
Amazon is taking a hard line with its wholesale suppliers, refusing to absorb rising tariff and oil costs. This signals a strategic shift away from its historical focus on vast product selection, even at a loss, towards prioritizing profitable items, which could squeeze smaller brands and ultimately reduce its product variety.
Anthropic is forcing developers using tools like OpenClaw to pay for API access separately from consumer subscriptions. This move, driven by compute constraints and pre-IPO financial discipline, indicates the era of venture-subsidized, low-cost AI usage is ending as model providers must cover massive compute expenses.
Google's cloud division (GCP), incentivized to sell compute, is allocating scarce TPU chips to external customer Anthropic. This directly constrains Google's own AI lab, Gemini, hindering its progress in the hyper-competitive AI race and revealing significant internal friction between business units with conflicting goals.
Anthropic's strategy of running workloads on diverse chips (NVIDIA, Google TPU, AWS Trainium) is less about long-term diversification and more about immediate survival. In a market where compute is severely constrained, the ability to utilize any available chip becomes a critical competitive advantage, forcing deep technical competence across architectures.
