For the past three years, the AI industry has operated under a simple assumption: more centralized compute solves everything. Bigger clusters. Bigger data centers. Bigger power contracts. The logic was intuitive and, for training workloads, largely correct. Training a frontier model requires tightly coupled GPU clusters with high-bandwidth interconnects, and