Google Unveils New AI Chips, Announces Deal with Anthropic

Google Cloud is rolling out its seventh-generation Tensor Processing Unit (TPU), Ironwood, and new Arm-based computing options that aim to meet exploding demand for AI model deployment in what the Alphabet company describes as a business shift from training models to serving end users. “Constantly shifting model architectures, the rise of agentic workflows, plus near-exponential growth in demand for compute, define this new age of inference,” explains Google Cloud. The company said that Anthropic — known for its Claude family of large language models — “plans to access up to 1 million” of the new TPUs. The deal is reportedly “worth billions.” Continue reading Google Unveils New AI Chips, Announces Deal with Anthropic

Google Ironwood TPU is Made for Inference and ‘Thinking’ AI

Google has debuted a new accelerator chip, Ironwood, a tensor processing unit designed specifically for inference — the ability of AI to predict things. Ironwood will power Google Cloud’s AI Hypercomputer, which runs the company’s Gemini models and is gearing up for the next generation of artificial intelligence workloads. Google’s TPUs are similar to the accelerator GPUs sold by Nvidia, but unlike the GPUs they’re designed for AI and geared toward speeding neural network tasks and mathematical operations. Google says when deployed at scale Ironwood is more than 24 times more powerful than the world’s fastest supercomputer. Continue reading Google Ironwood TPU is Made for Inference and ‘Thinking’ AI