Amazon Web Services Launches Trainium2 as Nvidia and AMD Reveal New AI Chips

AI chip rollouts accelerate as Amazon Web Services moves Trainium2 to general availability, Nvidia expands data center GPU supply, and AMD unveils new accelerators at CES. Export controls and HBM supply shape near-term availability, with executives signaling rapid infrastructure buildouts.

Published: January 15, 2026 By Sarah Chen, AI & Automotive Technology Editor Category: AI Chips

Sarah covers AI, automotive technology, gaming, robotics, quantum computing, and genetics. Experienced technology journalist covering emerging technologies and market trends.

Amazon Web Services Launches Trainium2 as Nvidia and AMD Reveal New AI Chips
Executive Summary
  • Amazon Web Services moves Trainium2 to general availability, targeting faster model training and lower cost per token according to AWS.
  • Nvidia signals expanded shipments of data center GPUs and packaging capacity alignment with suppliers to address backlog per Reuters.
  • AMD debuts new Instinct accelerators and AI PC platforms during CES week, focusing on enterprise deployments and software ecosystem support in AMD announcements.
  • HBM supply updates from SK hynix and Samsung underpin near-term AI accelerator availability, with pilot and ramp timelines outlined by SK hynix and Samsung.
AI Accelerators Hit the Gas Amazon Web Services moved its second-generation AI training silicon into customer hands, announcing general availability for Trainium2-based EC2 instances in December, with stated gains in throughput and efficiency for large-scale training workloads via the AWS News Blog. Dave Brown, VP of Amazon EC2 at Amazon Web Services, said, "Customers want price-performance and predictable capacity for frontier and enterprise models, and Trainium2 is designed to deliver both at scale," in remarks published alongside the launch by AWS. Nvidia signaled continued momentum in data center GPU supply, with industry reports pointing to expanded shipments of high-bandwidth-memory-equipped accelerators into cloud and enterprise channels as packaging capacity tightens per Reuters. Jensen Huang, CEO of Nvidia, said recently that "every layer of the stack, from networking to software, is scaling to meet accelerated computing demand," emphasizing collaboration with foundry and memory partners to reduce lead times as reported by Bloomberg. AMD and Intel Press the Datacenter Case During the early January CES news cycle, AMD outlined updates to its Instinct portfolio aimed at enterprise AI training and inference deployments, alongside software optimizations in ROCm intended to streamline model portability from CUDA according to AMD newsroom materials. "Open software, consistent performance, and reliable supply are the priorities our customers raise most often, and we are executing against all three," AMD Chair and CEO Lisa Su said during the company’s CES week briefings as posted by AMD. On the networking and price-performance front, Intel emphasized expanded availability of Gaudi-based systems for AI training clusters and outlined updated pricing designed to broaden adoption for mid-scale deployments in Intel newsroom updates. Sandra Rivera, Executive Vice President and General Manager of Intel’s Data Center and AI Group, said, "Choice and predictable TCO are critical for customers ramping from pilots to production, and our Gaudi platform is tuned for that inflection," in remarks released this month by Intel. Supply Chains Pivot Around HBM and Advanced Packaging High-bandwidth memory remains the gating factor for many accelerator deliveries, with SK hynix updating customers on its HBM roadmap and pilot timelines and Samsung Electronics highlighting capacity additions for HBM3E as it readies HBM4 process transitions per company updates and Samsung statements. Analysts note that advanced packaging throughput—particularly CoWoS—continues to shape quarterly shipment profiles for leading accelerators according to Gartner commentary. Foundry partners have reiterated plans to scale advanced packaging capacity serving AI GPUs and custom accelerators, aligning with cloud providers’ multi-year capex plans as reported by Reuters. These dynamics are feeding into procurement strategies by hyperscalers including Google Cloud, Microsoft Azure, and Oracle Cloud, which are simultaneously deploying in-house and third-party accelerators to diversify supply per Bloomberg reporting. For more on related AI Chips developments. Regulatory and Geopolitical Pressure Intensifies Export-control scrutiny continues to influence shipment timing and product variants destined for restricted markets, with U.S. rules shaping chip-to-chip interconnect thresholds, memory bandwidth, and performance-per-watt profiles according to Reuters. Industry sources indicate near-term adjustments to China-specific accelerators and supply allocations as vendors seek compliance without disrupting broader global deliveries per Financial Times coverage. Cloud and enterprise buyers are responding by dual-sourcing across custom silicon and merchant GPUs, while leaning on software portability layers to hedge supply risk according to Forrester analysis. "We continue to invest across merchant and custom silicon to ensure capacity for both training and inference," said a senior infrastructure executive at Microsoft, highlighting ongoing deployment of the in-house Maia accelerator alongside merchant GPUs in Azure datacenters as reflected in Microsoft Azure engineering posts. This builds on broader AI Chips trends. Key Product and Funding Signals Investors remain active across the AI silicon stack, with startups developing inference-optimized architectures and software-defined sparsity engines reporting fresh funding and customer pilots in December and January per TechCrunch. For more on [related voice ai developments](/ftc-finalizes-impersonation-ban-and-fcc-targets-ai-robocalls-in-voice-ai-crackdown-11-01-2026). Larger ecosystem plays include multi-year supply commitments between hyperscalers and chipmakers, often paired with capacity reservation fees and take-or-pay clauses designed to smooth quarterly variability according to Bloomberg. Cereals of new SKUs from the major vendors in the past six weeks—spanning data center training, inference, and edge deployments—point to a competitive cadence that is likely to persist through the first half of 2026 as tooling and memory supply catch up per IDC. "Customers want to accelerate time-to-train while managing power envelopes and sustainability targets—those goals are now inseparable," Nvidia’s Huang added, underscoring the alignment between architecture, interconnect, and software orchestration per Reuters. Company and Product Milestones Announced Since December
CompanyProduct or UpdateAnnouncement DateStatus and Source
Amazon Web ServicesTrainium2 EC2 instancesDecember 2025General availability AWS News Blog
NvidiaExpanded data center GPU supplyDecember 2025Shipment ramp reported Reuters
AMDInstinct accelerator updates at CESJanuary 2026Product briefing AMD Newsroom
IntelGaudi platform availability and pricingJanuary 2026Channel update Intel Newsroom
SK hynixHBM roadmap and pilot timingJanuary 2026Roadmap note SK hynix Newsroom
SamsungHBM3E capacity additionsDecember 2025Supplier update Samsung Newsroom
What to Watch Next Key near-term markers include the pace of HBM3E availability, clarity on HBM4 pilot and qualification timing, and any new export-control guidance that could alter delivery schedules into restricted markets per FT reporting. Cloud providers are likely to share expanded detail on custom accelerator deployments and mix with merchant GPUs in upcoming earnings cycles, offering signals on cost per token, time-to-train, and utilization trends according to Bloomberg. Analysts also point to software portability and orchestration as differentiators as enterprises seek to avoid lock-in, with ROCm, CUDA, and growing support for open compilers shaping workload placement across heterogeneous fleets per Gartner. Expect more disclosures on capacity reservations, long-term supply agreements, and cross-licensing that collectively stabilize supply and pricing as demand remains elevated into the first half of 2026 according to IDC. FAQs { "question": "What did AWS announce about Trainium2 and how does it impact training costs?", "answer": "Amazon Web Services announced the general availability of Trainium2-based EC2 instances in December 2025, positioning the chip for higher throughput at lower cost per token for large model training. For more on [related investments developments](/sp-500-index-fund-uk-forecast-2026-growth-projections-investment-strategies-20-12-2025). AWS says customers can scale clusters to thousands of accelerators with predictable capacity, aiming to cut training times and operational costs for frontier and enterprise models. Early user feedback cited by AWS highlights improvements in price-performance and integration with popular frameworks. The announcement reflects AWS’s strategy to balance custom silicon with broad software support across its AI stack." } { "question": "How are Nvidia and AMD addressing current AI accelerator supply constraints?", "answer": "Nvidia indicates it is working with foundry and memory partners to expand shipments of data center GPUs, aligning packaging capacity and HBM supply to reduce lead times. AMD, meanwhile, used CES week to outline Instinct updates and ROCm software improvements designed to ease deployment and portability for enterprise customers. Both vendors emphasize multi-quarter visibility with cloud providers through supply reservations. Industry reports suggest packaging throughput and HBM remain the primary bottlenecks, with incremental relief expected as capacity ramps." } { "question": "What role do HBM suppliers SK hynix and Samsung play in near-term availability?", "answer": "High-bandwidth memory suppliers are central to the pace of AI accelerator deliveries. SK hynix updated its HBM roadmap and pilot timelines, while Samsung highlighted capacity additions for HBM3E and transitions toward HBM4. These steps aim to increase bits shipped and support larger GPU memory configurations. Since most top-end training accelerators rely on stacked HBM, even modest throughput gains can materially improve quarterly shipment numbers, directly affecting cloud capacity rollouts and enterprise project timelines." } { "question": "How are export controls influencing product variants and delivery schedules?", "answer": "U.S. export controls continue to shape the specifications and timing of accelerators shipped to restricted markets, affecting interconnect capabilities, memory bandwidth, and overall performance thresholds. Vendors have developed market-specific variants to maintain compliance while serving global demand. Reports indicate that these constraints can shift deliveries between quarters, prompting hyperscalers to diversify across merchants and custom silicon. As regulators refine guidance, suppliers adjust product roadmaps and supply allocations to minimize disruption to broader shipment plans." } { "question": "What indicators should enterprises track to plan AI infrastructure in early 2026?", "answer": "Enterprises should monitor HBM3E and HBM4 timeline disclosures, foundry packaging capacity updates, and hyperscaler earnings for insights on accelerator mix, utilization, and cost per token. Software portability developments—across CUDA, ROCm, and open compilers—will influence workload placement and vendor lock-in risk. Additionally, watch for new long-term supply agreements, capacity reservations, and pricing clarity from Nvidia, AMD, and Intel. These signals collectively inform project timing, procurement strategies, and total cost of ownership for training and inference fleets." } References

About the Author

SC

Sarah Chen

AI & Automotive Technology Editor

Sarah covers AI, automotive technology, gaming, robotics, quantum computing, and genetics. Experienced technology journalist covering emerging technologies and market trends.

About Our Mission Editorial Guidelines Corrections Policy Contact

Frequently Asked Questions

What did AWS announce about Trainium2 and how does it impact training costs?

Amazon Web Services announced general availability for Trainium2-based EC2 instances in December 2025, emphasizing higher throughput and improved price-performance for large model training. AWS highlights that customers can scale to thousands of accelerators with predictable capacity and integrate with mainstream frameworks. Early user feedback cited by AWS indicates lower cost per token and shorter training times. The move underscores AWS’s strategy to combine custom silicon with a broad software ecosystem across its AI stack to manage both performance and cost.

How are Nvidia and AMD addressing AI accelerator supply challenges right now?

Nvidia says it is working with foundry and HBM partners to expand shipments of data center GPUs, aligning advanced packaging capacity to reduce backlogs. AMD used the CES window to showcase Instinct updates and ROCm software maturation, aiming to ease workload portability and deployment. Both vendors emphasize multi-quarter supply visibility with hyperscalers through reservations and take-or-pay contracts. Industry reporting points to incremental relief as HBM and packaging capacity ramps through early 2026, though demand remains elevated.

What is the significance of HBM supply from SK hynix and Samsung for AI chips?

HBM is a key bottleneck for top-tier accelerators, determining memory capacity and bandwidth per device. SK hynix updated its HBM roadmap and pilot timing, while Samsung outlined HBM3E capacity additions and steps toward HBM4. Increases in HBM throughput directly raise the number of shippable AI accelerators each quarter. As a result, even modest gains in HBM output can translate into materially higher cloud capacity, influencing time-to-train metrics and the scheduling of enterprise AI deployments.

How are export controls affecting AI chip products and delivery schedules?

U.S. export controls are shaping the specifications of accelerators destined for restricted markets, impacting interconnect capabilities, memory bandwidth, and performance limits. Vendors have responded with region-specific variants to maintain compliance while serving global demand. These rules can shift deliveries between quarters and alter product mix, prompting hyperscalers to diversify across merchant GPUs and custom silicon. Ongoing regulatory updates require suppliers and buyers to adapt procurement plans and maintain flexibility across software and hardware stacks.

What should enterprises monitor to plan AI infrastructure purchases in early 2026?

Enterprises should track HBM3E and HBM4 disclosures, advanced packaging capacity updates, and hyperscaler earnings for signals on accelerator mix, utilization, and price-performance. Software portability—across CUDA, ROCm, and open compilers—will influence workload placement and vendor lock-in risk. Watch for long-term supply agreements and capacity reservations that stabilize availability and pricing. These indicators, combined with published time-to-train and cost-per-token metrics, can guide procurement timing and architecture choices for training and inference fleets.