Amazon Rolls Out New AI Chips, UltraServers and AI Factories

Amazon’s Trainium3 accelerator chip — the company’s first AI silicon built on 3nm technology — is now in general release. It comes to market with the Trainium3 UltraServer, a high-density integrated system purpose-built for large-scale GenAI model training. Trn3 UltraServers can scale to 144 Trainium3 chips, delivering up to 362 FP8 PFLOPs. The Trn3 chips are viewed as challengers to Nvidia’s AI GPUs and Google’s Tensor TPUs. But Amazon also announced it will provide enterprises with turnkey AWS AI Factories that utilize key Nvidia components and support its AI chips, allowing customers to choose or combine brands. Continue reading Amazon Rolls Out New AI Chips, UltraServers and AI Factories

DGX Cloud Lepton: Nvidia’s New GPU Compute Marketplace

Nvidia is rolling out DGX Cloud Lepton, a platform that connects AI developers with GPU access available through various cloud providers. Nvidia calls it “a compute marketplace” that offers tens of thousands of GPUs through a global network that features Nvidia Cloud Partners (NCPs). Among them: CoreWeave, Crusoe, Firmus, Foxconn, GMI Cloud, Lambda, Nebius, Nscale, Softbank Corp. and Yotta Data Services — offering Nvidia Blackwell and other architecture GPUs. Developers can tap into GPU compute capacity in specific regions for both on-demand and long-term computing, Nvidia says, adding that it expects leading cloud computing providers to eventually sign on. Continue reading DGX Cloud Lepton: Nvidia’s New GPU Compute Marketplace