Executive Summary
- Amazon Web Services moves Trainium2 to general availability, targeting faster model training and lower cost per token according to AWS.
- Nvidia signals expanded shipments of data center GPUs and packaging capacity alignment with suppliers to address backlog per Reuters.
- AMD debuts new Instinct accelerators and AI PC platforms during CES week, focusing on enterprise deployments and software ecosystem support in AMD announcements.
- HBM supply updates from SK hynix and Samsung underpin near-term AI accelerator availability, with pilot and ramp timelines outlined by SK hynix and Samsung.
AI Accelerators Hit the Gas
Amazon Web Services moved its second-generation AI training silicon into customer hands, announcing general availability for Trainium2-based EC2 instances in December, with stated gains in throughput and efficiency for large-scale training workloads via the AWS News Blog. Dave Brown, VP of Amazon EC2 at Amazon Web Services, said, "Customers want price-performance and predictable capacity for frontier and enterprise models, and Trainium2 is designed to deliver both at scale," in remarks published alongside the launch by AWS.
Nvidia signaled continued momentum in data center GPU supply, with industry reports pointing to expanded shipments of high-bandwidth-memory-equipped accelerators into cloud and enterprise channels as packaging capacity tightens per Reuters. Jensen Huang, CEO of Nvidia, said recently that "every layer of the stack, from networking to software, is scaling to meet accelerated computing demand," emphasizing collaboration with foundry and memory partners to reduce lead times as reported by Bloomberg.
AMD and Intel Press the Datacenter Case
During the early January CES news cycle, AMD outlined updates to its Instinct portfolio aimed at enterprise AI training and inference deployments, alongside software optimizations in ROCm intended to streamline model portability from CUDA according to AMD newsroom materials...