Amazon Web Services Launches Trainium2 as Nvidia and AMD Reveal New AI Chips
AI chip rollouts accelerate as Amazon Web Services moves Trainium2 to general availability, Nvidia expands data center GPU supply, and AMD unveils new accelerators at CES. Export controls and HBM supply shape near-term availability, with executives signaling rapid infrastructure buildouts.
Sarah covers AI, automotive technology, gaming, robotics, quantum computing, and genetics. Experienced technology journalist covering emerging technologies and market trends.
- Amazon Web Services moves Trainium2 to general availability, targeting faster model training and lower cost per token according to AWS.
- Nvidia signals expanded shipments of data center GPUs and packaging capacity alignment with suppliers to address backlog per Reuters.
- AMD debuts new Instinct accelerators and AI PC platforms during CES week, focusing on enterprise deployments and software ecosystem support in AMD announcements.
- HBM supply updates from SK hynix and Samsung underpin near-term AI accelerator availability, with pilot and ramp timelines outlined by SK hynix and Samsung.
| Company | Product or Update | Announcement Date | Status and Source |
|---|---|---|---|
| Amazon Web Services | Trainium2 EC2 instances | December 2025 | General availability AWS News Blog |
| Nvidia | Expanded data center GPU supply | December 2025 | Shipment ramp reported Reuters |
| AMD | Instinct accelerator updates at CES | January 2026 | Product briefing AMD Newsroom |
| Intel | Gaudi platform availability and pricing | January 2026 | Channel update Intel Newsroom |
| SK hynix | HBM roadmap and pilot timing | January 2026 | Roadmap note SK hynix Newsroom |
| Samsung | HBM3E capacity additions | December 2025 | Supplier update Samsung Newsroom |
- AWS News Blog - Amazon Web Services, December 2025
- AI chip supply and demand coverage - Reuters, December 2025–January 2026
- AI hardware infrastructure reporting - Bloomberg, December 2025–January 2026
- AMD Newsroom - AMD, January 2026
- Intel Newsroom - Intel, January 2026
- SK hynix Newsroom - SK hynix, January 2026
- Samsung Newsroom - Samsung Electronics, December 2025
- Gartner Research Newsroom - Gartner, December 2025–January 2026
- IDC Research - IDC, December 2025–January 2026
- Technology coverage on export controls - Financial Times, December 2025–January 2026
About the Author
Sarah Chen
AI & Automotive Technology Editor
Sarah covers AI, automotive technology, gaming, robotics, quantum computing, and genetics. Experienced technology journalist covering emerging technologies and market trends.
Frequently Asked Questions
What did AWS announce about Trainium2 and how does it impact training costs?
Amazon Web Services announced general availability for Trainium2-based EC2 instances in December 2025, emphasizing higher throughput and improved price-performance for large model training. AWS highlights that customers can scale to thousands of accelerators with predictable capacity and integrate with mainstream frameworks. Early user feedback cited by AWS indicates lower cost per token and shorter training times. The move underscores AWS’s strategy to combine custom silicon with a broad software ecosystem across its AI stack to manage both performance and cost.
How are Nvidia and AMD addressing AI accelerator supply challenges right now?
Nvidia says it is working with foundry and HBM partners to expand shipments of data center GPUs, aligning advanced packaging capacity to reduce backlogs. AMD used the CES window to showcase Instinct updates and ROCm software maturation, aiming to ease workload portability and deployment. Both vendors emphasize multi-quarter supply visibility with hyperscalers through reservations and take-or-pay contracts. Industry reporting points to incremental relief as HBM and packaging capacity ramps through early 2026, though demand remains elevated.
What is the significance of HBM supply from SK hynix and Samsung for AI chips?
HBM is a key bottleneck for top-tier accelerators, determining memory capacity and bandwidth per device. SK hynix updated its HBM roadmap and pilot timing, while Samsung outlined HBM3E capacity additions and steps toward HBM4. Increases in HBM throughput directly raise the number of shippable AI accelerators each quarter. As a result, even modest gains in HBM output can translate into materially higher cloud capacity, influencing time-to-train metrics and the scheduling of enterprise AI deployments.
How are export controls affecting AI chip products and delivery schedules?
U.S. export controls are shaping the specifications of accelerators destined for restricted markets, impacting interconnect capabilities, memory bandwidth, and performance limits. Vendors have responded with region-specific variants to maintain compliance while serving global demand. These rules can shift deliveries between quarters and alter product mix, prompting hyperscalers to diversify across merchant GPUs and custom silicon. Ongoing regulatory updates require suppliers and buyers to adapt procurement plans and maintain flexibility across software and hardware stacks.
What should enterprises monitor to plan AI infrastructure purchases in early 2026?
Enterprises should track HBM3E and HBM4 disclosures, advanced packaging capacity updates, and hyperscaler earnings for signals on accelerator mix, utilization, and price-performance. Software portability—across CUDA, ROCm, and open compilers—will influence workload placement and vendor lock-in risk. Watch for long-term supply agreements and capacity reservations that stabilize availability and pricing. These indicators, combined with published time-to-train and cost-per-token metrics, can guide procurement timing and architecture choices for training and inference fleets.