Google Unveils New AI Chips, Announces Deal with Anthropic

Google Cloud is rolling out its seventh-generation Tensor Processing Unit (TPU), Ironwood, and new Arm-based computing options that aim to meet exploding demand for AI model deployment in what the Alphabet company describes as a business shift from training models to serving end users. “Constantly shifting model architectures, the rise of agentic workflows, plus near-exponential growth in demand for compute, define this new age of inference,” explains Google Cloud. The company said that Anthropic — known for its Claude family of large language models — “plans to access up to 1 million” of the new TPUs. The deal is reportedly “worth billions.” Continue reading Google Unveils New AI Chips, Announces Deal with Anthropic

Meta Deploys Gen 2 MTIA AI Accelerator Chip in Data Centers

Meta’s next generation AI silicon is a 5nm chip designed to power the models that provide recommendations to those who use its social network platforms. The new MTIA inference accelerator is part of a “broader full-stack development program for custom, domain-specific silicon that addresses our unique workloads and systems,” Meta says. The next-gen MTIA more than doubles the compute and memory bandwidth of its predecessor, the 7nm MTIA v1 chip introduced in May 2023, resulting in 3x the performance, according to Meta, which says the new silicon is already live in 16 data centers. Continue reading Meta Deploys Gen 2 MTIA AI Accelerator Chip in Data Centers