By Paulina Likos
Publication Date: 2025-12-02 21:20:00
Amazon Web Services’ two-track approach to artificial intelligence came into better focus Tuesday as the world’s biggest cloud pushed forward with its own custom chips and got closer to Nvidia . During Amazon ‘s annual AWS Re:Invent 2025 conference in Las Vegas, Amazon Web Services CEO Matt Garman unveiled Trainium3 — the latest version of the company’s in-house custom chip. It has four times more compute performance, energy efficiency, and memory bandwidth than previous generations. AWS said that early results of customers testing Trainium3 are reducing AI training and inference costs by up to 50%. Custom chips, like Trainium, are becoming more and more popular for the big tech companies that can afford to make them. And, their use cases are broadening. For example, Google’s tensor processing units (TPUs), co-designed by Broadcom , have also been getting a lot of attention since last month’s launch of the well-received Gemini 3 artificial intelligence model. It is powered by TPUs. There was even a report that Meta Platforms was considering TPUs in addition to Nvidia ‘s graphics processing units (GPUs), which are the gold standard for all-purpose AI workloads. At the same time, Amazon also announced that it’s deepening its work with Nvidia. In Tuesday’s keynote, Garman introduced AWS Factories, which provides on-premise AI infrastructure for customers to use in their own data centers. The service combines Trainium accelerators and Nvidia graphics processing units, which…