Brandon Seppa Navigation
  • Home
  • About
  • Search
  • Home
  • About
  • Search
Home Posts AI Infrastructure

Tag Archive

Below you'll find a list of all posts that have been tagged as “AI Infrastructure”

Your Load Balancer Has No Idea What an LLM Is (Google Fixed That)

Standard load balancers treat LLM inference like any other HTTP traffic. That is expensive and slow. GKE Inference Gateway knows the difference.

AI InfrastructureGKEGoogle CloudKubernetesLLM Inference

Your AI Inference Bill Is About to Become a Strategy Problem

Google Research just published a way to cut AI serving costs by 50% with zero accuracy loss. The interesting part is what happens to the ISVs who figure this out first.

AI InfrastructureCost OptimizationGoogle CloudLLM InferenceTPU
LinkedIn
BRANDONSEPPA.COM © 2026