TPU Ironwood is Google’s 7th-generation custom AI chip, and unlike its predecessors, it was built for inference first. Here’s what that means and why it matters.
GPU Inference Without the Cluster. Cloud Run Finally Makes That Real.
Cloud Run now supports GPUs with scale-to-zero billing. For AI inference workloads that are bursty, sporadic, or just getting started, that changes the math entirely.

