Amazon Web Services just launched its biggest challenge yet to Nvidia’s dominance in AI computing. The Trainium3 UltraServer, powered by Amazon’s new 3-nanometer Trainium3 chip, promises 4x faster performance with 4x more memory than its predecessor for AI training and inference workloads.
The announcement came at AWS re:Invent 2025, where CEO Matt Garman emphasized that “AI assistants are starting to give way to AI agents that can perform tasks and automate on your behalf.” Translation: the AI infrastructure race is accelerating, and Amazon isn’t content to let Nvidia set the pace.
What Trainium3 Actually Delivers
The numbers are impressive on paper. Amazon claims Trainium3 offers a 4x improvement in training performance compared to Trainium2, while also quadrupling available memory per chip. For AI workloads, memory is often the bottleneck, as large language models need to hold billions of parameters in fast-access storage during training and inference.
The 3-nanometer process node is significant. Smaller transistors mean more computing power in the same physical space, better energy efficiency, and the ability to handle more complex workloads. Amazon is using TSMC’s advanced manufacturing for these chips, the same foundry that builds Apple’s latest processors and Nvidia’s cutting-edge GPUs.
AWS is positioning Trainium3 as purpose-built for the specific demands of AI workloads rather than general-purpose computing. That specialization allows architectural choices that wouldn’t make sense for a broader chip but deliver major advantages for training and running AI models.
The Nvidia Problem
Every major cloud provider faces the same challenge: they’re increasingly dependent on Nvidia for AI computing power, and Nvidia knows it.
Nvidia’s GPUs, particularly the H100 and newer Blackwell series, dominate AI training. The company’s CUDA software ecosystem has created massive switching costs, with most AI frameworks and tools optimized for Nvidia hardware. This gives Nvidia pricing power that makes cloud providers uncomfortable and raises costs for their customers.
Amazon, Google, and Microsoft have all invested billions in developing their own AI chips. The race for AI infrastructure is reshaping how tech giants approach computing. Google has TPUs, Microsoft is working on Maia, and Amazon has Trainium and Inferentia. But so far, none have seriously threatened Nvidia’s market position.
The challenge isn’t just building good chips. It’s building the software ecosystem that makes developers want to use them. Nvidia spent years creating CUDA, cuDNN, and countless other tools that AI researchers rely on daily. Replicating that ecosystem is harder than replicating the hardware.
Why This Matters for AI Costs
The economics of AI training are brutal. Training a frontier model like GPT-4 or Claude costs tens of millions of dollars in compute alone. Inference, running trained models to generate responses, costs less per query but adds up quickly at scale.
If Amazon can deliver comparable performance to Nvidia at lower prices, and they control the entire stack from chip to cloud service, they can offer AI computing at costs competitors can’t match. Even a 20-30% cost advantage could be decisive for startups choosing where to build.
The memory improvement may matter more than raw speed for many use cases. Large language models are getting larger, and fitting them into available memory without expensive workarounds determines what’s practically possible. More memory per chip means running bigger models without complex distributed computing setups.
The Bigger Picture
Amazon’s chip push is part of a broader trend of vertical integration in tech. Companies that once relied on commodity hardware are increasingly building custom silicon to gain advantages competitors can’t easily copy.
Apple proved the model works with its M-series chips, achieving performance and efficiency that Intel and AMD struggled to match. Now every major tech company is exploring custom silicon for their specific needs.
For AI specifically, the stakes are enormous. Whoever controls AI computing infrastructure controls a chokepoint in the most important technology race in a generation. Amazon isn’t trying to replace Nvidia entirely. They’re trying to reduce dependence enough that they have options.
The massive investments flowing into AI infrastructure make sense only if computing costs eventually come down. Custom chips from cloud providers are how that happens.
Whether Trainium3 delivers on its promises remains to be seen. Previous generations showed potential but didn’t fundamentally shift the market. Amazon is betting this generation is different. Given what’s at stake, they’re probably not done trying even if it isn’t.
Sources: AWS re:Invent 2025, TechCrunch, The Verge, Reuters.





