Amazon and Google’s accelerating push into in-house AI silicon marks the most serious challenge yet to NVIDIA’s long-standing dominance in AI compute. With AWS scaling Trainium chips across its global data centres now powering partners like Anthropic and Google expanding TPU deployments through similar alliances, the AI infrastructure market is entering a new, more competitive phase.
For Amazon, Trainium is no longer an experimental bet. CEO Andy Jassy’s reference to multi-billion-dollar revenues signals that custom chips are becoming a meaningful economic lever, helping AWS lower dependency on external vendors, improve margins, and offer customers more cost-efficient AI training and inference at scale. In a market where compute costs can define profitability, this vertical integration is strategically critical.
Google’s TPUs tell a similar story, but at even greater scale. Broadcom’s disclosure that TPU revenues now run into the tens of billions highlights how deeply Google has embedded custom silicon into its cloud and AI stack. By pairing TPUs with its software, models, and cloud services, Google is reinforcing a closed-loop AI ecosystem that competes not just on performance, but on economics and optimisation.
This shift doesn’t immediately dethrone NVIDIA, whose GPUs remain the gold standard for flexibility and ecosystem depth. However, it does signal a structural change: hyperscalers are no longer content to be price-takers in the AI compute supply chain. Instead, they are becoming chip designers, infrastructure owners, and platform orchestrators.
For the broader AI market, the implications are significant. Customers may benefit from lower costs, diversified supply, and faster innovation, while NVIDIA faces increased pressure to defend its moat through software, networking, and next-gen architectures.
Overall, Amazon and Google’s silicon push underscores a defining reality of the AI era: control over compute is now control over the future of AI economics.

