Google released TPU v8 to accelerate large-scale model training and inference. This update focuses on energy efficiency and interconnect speeds to reduce latency. Meanwhile, Tesla is building a dedicated research fab for AI chips. These infrastructure plays aim to lower the soaring cost of compute for next-generation frontier models.