Discussion about this post

User's avatar
Neural Foundry's avatar

The economics breakdown here is revealing, particularly the point about Google controlling the full stack from Broadcom TPU dies through optical fabric. What's especially compelling is how Anthropic's effective FLOP utilization approach reframes the entire comparison, moving beyond marketing peak FLOPs to real training workloads at ~40% utilization versus typical GPU deployments at ~30%. The ICI 3D torus architecture you describe is fundamentally different from InfiniBand or Ethernet approaches, and the ability to connect 9,216 chips on one fabric versus NVL72's limitation of72 GPUs per system creates a structural cost advantage that compounds at scale. Your observation about the "Nvidia tax" is spot on: even for labs that remain primarily on GPUs, having credible TPU economics in the negotiation fundamentally changes pricing dynamics. The debt structure around OpenAI's infrastructure partners is particularly interesting when viewed against this backdrop, the fact that $100B+ is being loaded onto Oracle, SoftBank, and others while OpenAI stays nearly debt-free suggests they're hedging against being locked into any single compute architecture.

Expand full comment

No posts

Ready for more?