<?xml version="1.0" encoding="UTF-8"?><rss xmlns:dc="http://purl.org/dc/elements/1.1/" xmlns:content="http://purl.org/rss/1.0/modules/content/" xmlns:atom="http://www.w3.org/2005/Atom" version="2.0" xmlns:itunes="http://www.itunes.com/dtds/podcast-1.0.dtd" xmlns:googleplay="http://www.google.com/schemas/play-podcasts/1.0"><channel><title><![CDATA[The Gigawatt Machine]]></title><description><![CDATA[Massive-Scale AI infrastructure]]></description><link>https://www.gigawattmachine.com</link><image><url>https://substackcdn.com/image/fetch/$s_!vqus!,w_256,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F59867208-fae5-40b9-a570-54b2271d292a_1280x1280.png</url><title>The Gigawatt Machine</title><link>https://www.gigawattmachine.com</link></image><generator>Substack</generator><lastBuildDate>Fri, 15 May 2026 14:27:53 GMT</lastBuildDate><atom:link href="https://www.gigawattmachine.com/feed" rel="self" type="application/rss+xml"/><copyright><![CDATA[Tony Wan]]></copyright><language><![CDATA[en]]></language><webMaster><![CDATA[gigawattmachine@substack.com]]></webMaster><itunes:owner><itunes:email><![CDATA[gigawattmachine@substack.com]]></itunes:email><itunes:name><![CDATA[Tony Wan]]></itunes:name></itunes:owner><itunes:author><![CDATA[Tony Wan]]></itunes:author><googleplay:owner><![CDATA[gigawattmachine@substack.com]]></googleplay:owner><googleplay:email><![CDATA[gigawattmachine@substack.com]]></googleplay:email><googleplay:author><![CDATA[Tony Wan]]></googleplay:author><itunes:block><![CDATA[Yes]]></itunes:block><item><title><![CDATA[Two Network Topologies: Hierarchical vs. Flat ]]></title><description><![CDATA[The Gigawatt Machine: NVIDIA, Google, and the Engineering of Scale 6/12]]></description><link>https://www.gigawattmachine.com/p/two-network-topologies-hierarchical</link><guid isPermaLink="false">https://www.gigawattmachine.com/p/two-network-topologies-hierarchical</guid><dc:creator><![CDATA[Tony Wan]]></dc:creator><pubDate>Thu, 11 Dec 2025 00:50:18 GMT</pubDate><enclosure url="https://substackcdn.com/image/fetch/$s_!vqus!,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F59867208-fae5-40b9-a570-54b2271d292a_1280x1280.png" length="0" type="image/jpeg"/><content:encoded><![CDATA[<h3>1. Introduction: Beyond the Node</h3><p>In the previous articles, we explored how NVIDIA and Google define the &#8220;compute node&#8221;&#8212;NVIDIA through progressive rack-scale integration (72 to 572 GPUs) and Google through configurable Pods (up to 8,960 TPUs). Now we must confront what happens when you leave the node.</p><p>A trillion-parameter model cannot fit within any single node, no matter how large. Training GPT-5 or Gemini requires tens of thousands of chips working in perfect synchrony across hundreds of racks distributed throughout a data center. The challenge is not computational&#8212;each chip knows how to multiply matrices&#8212;but communications: how do you move petabytes of data per second across a building-scale fabric without turning the network into a bottleneck?</p><p>The physics of this problem forced both companies to abandon the traditional &#8220;fat tree&#8221; network topology used in cloud data centers. Instead, they engineered fundamentally different solutions:</p><p><strong>NVIDIA&#8217;s approach: The Hierarchical Federation.</strong> A cluster is a federation of ultra-dense rack-scale supercomputers (NVL72, and in the future NVL144 and NVL572 nodes) connected by layers of electrical switches arranged in a Rail-Optimized topology. Data moves through a three-tier hierarchy: copper NVLink inside the rack, InfiniBand between racks within a pod, and inter-pod networking at the data center scale.</p><p><strong>Google&#8217;s approach: The Flat Optical Mesh.</strong> A cluster is a single, massive optical fabric where every TPU connects to its six neighbors through dedicated circuit-switched paths. There is no hierarchy&#8212;the &#8220;network&#8221; is just the 3D Torus mesh of ICI links, with Optical Circuit Switches (OCS) dynamically reconfiguring the topology as needed.</p><p>This article examines how these two topologies handle the traffic patterns of AI training, and the traffic storms that would collapse a conventional network&#8212;and what happens when hardware fails in a supercomputer.</p><h3>2. The Two-Domain Architecture: Super-Node vs. Scale-Out</h3><p>Before diving into each company&#8217;s specific implementation, we must understand a fundamental pattern that has emerged in both architectures: the bifurcation of networking into two distinct domains.</p><p>As models scale beyond 100,000 chips, physics forces a structural reality that neither company explicitly planned but both discovered independently. A trillion-parameter cluster requires two domains with fundamentally different characteristics.</p><h3>Domain 1: The &#8220;Super-Node&#8221; Fabric (Proprietary)</h3><p>This is the domain of maximum bandwidth and tightest coupling. Here, the &#8220;network&#8221; acts less like a cable and more like a motherboard bus. Chips in this domain communicate as if they were all on the same die.</p><p><strong>Defining characteristics:</strong></p><ul><li><p><strong>Proprietary technology:</strong> Custom interconnects optimized for minimum latency</p></li><li><p><strong>Tight coupling:</strong> Sub-microsecond latency, often with hardware-coherent memory</p></li><li><p><strong>Physical constraints:</strong> Limited by the physics of the interconnect medium (copper range or optical circuit-switch scale)</p></li><li><p><strong>Traffic type:</strong> Handles the most bandwidth-intensive operations that require instant synchronization</p></li></ul><p><strong>The strategic purpose:</strong> Keep the highest-traffic communication patterns&#8212;specifically operations like Tensor Parallelism, where a single matrix multiply is split across multiple chips&#8212;entirely within this domain. By doing so, these operations never touch the slower external network.</p><p><strong>NVIDIA&#8217;s implementation:</strong> The GB200 NVL72 rack, where 72 GPUs connect via a copper backplane using NVLink 5.0, creating a 130 TB/s shared-memory domain. The rack&#8217;s physical boundaries (roughly 1 meter) are defined by copper signal integrity limits.</p><p><strong>Google&#8217;s implementation:</strong> The TPU Pod, where thousands of chips (up to 9,216 in TPU v7) connect via the ICI optical mesh with Optical Circuit Switches. Unlike NVIDIA&#8217;s copper limitation, optical fibers allow the Pod to span multiple physical racks&#8212;though Google can provision domains of various sizes from the Pod. This flexibility comes at the cost of requiring compile-time scheduling rather than hardware coherence.</p><h3>Domain 2: The Scale-Out Fabric (Standard)</h3><p>When a workload exceeds the boundaries of Domain 1, it must cross into the external network. Scaling the proprietary super-node fabric across an entire building is impractical due to signal degradation, cabling complexity, and the operational risks of creating a single massive fault domain.</p><p><strong>Defining characteristics:</strong></p><ul><li><p><strong>Standard protocols:</strong> Based on industry standards (InfiniBand, Ethernet) for interoperability</p></li><li><p><strong>Packet-switched:</strong> Dynamic routing, buffering, and congestion control</p></li><li><p><strong>Building-scale:</strong> Can span hundreds of meters using optical transceivers</p></li><li><p><strong>Multi-purpose:</strong> Must also connect to storage, CPUs, and external networks</p></li></ul><p><strong>The strategic purpose:</strong> Connect the super-node islands together, bridge to storage systems, and provide access to the outside world. Performance here matters, but flexibility and interoperability matter more.</p><p><strong>NVIDIA&#8217;s implementation:</strong> InfiniBand (Quantum-X800) or Converged Ethernet (Spectrum-X), using Rail-Optimized topologies and in-network computing (SHARP) to maintain high performance while providing resilience through dynamic routing.</p><p><strong>Google&#8217;s implementation:</strong> Jupiter Data Center Network, using standard Ethernet frames with customized protocols (Swift) for precise timing. This layer connects TPU Slices in Multislice configurations and bridges to Google&#8217;s massive storage infrastructure.</p><h3>The Convergence Nobody Planned</h3><p>Both companies have arrived at a similar two-domain structure through completely different paths:</p><p><strong>Google started with a flat mesh</strong> (the ICI fabric) within a single Pod. To scale beyond the Pod maximum (9,216 chips for TPU v7), they developed Multislice&#8212;technology that connects multiple Pods via packet-switched Ethernet (Jupiter DCN). This created Domain 2.</p><p><strong>NVIDIA started with a hierarchy</strong> (discrete servers connected by networks) and is aggressively expanding Domain 1 to encompass more compute. The NVL72 rack (and future NVL144, NVL572) turns what used to require networking into a single tightly-coupled node.</p><p>Both topologies now have two domains. Google&#8217;s flat mesh uses scale-out networking to grow beyond the Pod. NVIDIA&#8217;s hierarchy uses massive &#8220;flat&#8221; nodes to reduce networking. Zoom out, and we see huge islands of determinism connected by oceans of dynamic networking.</p><p>The philosophical difference remains: NVIDIA believes the network should adapt to hardware imperfections. Google believes the hardware should be perfectly configured by software. But the two-domain structure&#8212;proprietary super-nodes plus standard scale-out&#8212;has become the inevitable architecture of the Gigawatt era.</p><p>With this framework established, we can now examine how each company implements these two domains in practice.</p><h3>3. NVIDIA&#8217;s Implementation: The Hierarchical Federation</h3><p>NVIDIA&#8217;s architecture treats the supercomputer as a hierarchy of networks, each layer optimized for different bandwidth requirements and traffic patterns. Understanding this hierarchy requires examining each layer and how they interconnect.</p><h4>Domain 1: The Rack as Super-Node (NVLink)</h4><p>As established in Article 5, the GB200 NVL72 rack contains 72 GPUs connected by a massive copper backplane. This creates a 130 TB/s all-to-all fabric using NVLink 5.0, where every GPU can talk to every other GPU in the rack at full speed through passive copper traces.</p><p>NVIDIA&#8217;s roadmap expands this rack-scale integration: the Vera Rubin NVL144 (2026) will integrate 144 GPUs, and the Vera Rubin Ultra NVL572 (2027) will reach 572 GPUs&#8212;all within a single tightly-coupled super-node. This progressive expansion of Domain 1 is NVIDIA&#8217;s strategy for reducing dependency on the external network: the larger the rack-scale unit, the more computation stays within the ultra-low-latency domain. </p><p><strong>Why copper:</strong> At these bandwidths (1.8 TB/s per GPU), electrical signals degrade rapidly. The NVL72 rack is engineered to keep all 72 GPUs within the maximum distance that passive copper can handle&#8212;roughly 1 meter. This avoids the latency and power tax of optical transceivers. (Saves ~200 nanoseconds per hop converting between light and electricity. Saves ~20 kW per rack in power consumption.)</p><p><strong>The strategic implication:</strong> NVIDIA architected the rack-scale super-node to handle the most bandwidth-intensive operations&#8212;specifically Tensor Parallelism, where a single matrix operation is split across multiple GPUs&#8212;entirely within the copper domain. As the roadmap progresses from 72 &#8594; 144 &#8594; 572 GPUs, it&#8217;s possible to keep the highest-traffic operations for even trillion-parameter models  within Domain 1, avoiding the external network entirely for the most latency-sensitive work.</p><p><strong>The result:</strong> To the software, the entire rack looks like one giant GPU. NCCL (NVIDIA&#8217;s communication library) sees the rack as a single shared-memory domain where communication is instantaneous and lossless.</p><h4>Domain 2: Connecting the Super-Nodes (InfiniBand)</h4><p>When workloads require more than 72 GPUs, they must connect multiple NVL72 racks. Each rack maintains its own isolated NVLink domain&#8212;the copper backplane cannot extend beyond the physical cabinet. These separate NVLink domains must be bridged by a scale-out network: InfiniBand (or Ethernet).</p><p>As NVIDIA&#8217;s roadmap expands the rack-scale node (144 GPUs with Vera Rubin NVL144 in 2026, 572 GPUs with Vera Rubin Ultra NVL572 in 2027), the fundamental architecture remains the same. Domain 1 grows larger&#8212;keeping more computation within the low-latency copper or proprietary interconnect. Domain 2 (InfiniBand or Ethernet) provides the scale-out fabric connecting these progressively larger super-nodes.</p><p>For this scale-out layer, NVIDIA uses the Quantum-X800 InfiniBand switch&#8212;their fastest network switch for AI workloads.</p><p>The specifications:</p><ul><li><p>Bandwidth: 800 Gbps per port (InfiniBand XDR speed)</p></li><li><p>Latency: Sub-130 nanoseconds port-to-port</p></li><li><p>Radix: 144 ports per switch</p></li><li><p>Scale: To build a non-blocking network for 576 GPUs (8 NVL72 racks), approximately 60-80 switches are required</p></li></ul><h4>The Rail-Optimized Topology</h4><p>This is where NVIDIA&#8217;s architecture becomes radically different from conventional data center networking. Instead of building a single large network where any server can talk to any server (a &#8220;fat tree&#8221;), NVIDIA segregates the network into 72 parallel, independent networks called &#8220;Rails.&#8221;</p><p>To understand Rails, start with the scaling challenge: each NVL72 rack is an isolated 72-GPU NVLink domain. To build a cluster larger than 72 GPUs, you must deploy multiple racks and connect them via InfiniBand. Each GPU in a rack has a dedicated external network connection, allowing it to communicate with corresponding GPUs in other racks.</p><p><strong>How Rails work:</strong></p><ul><li><p>Rail 0 connects GPU #0 from every NVL72 rack in the cluster</p></li><li><p>Rail 1 connects GPU #1 from every NVL72 rack</p></li><li><p>Rail 2 connects GPU #2 from every NVL72 rack</p></li><li><p>...and so on through Rail 71</p></li></ul><p>GPU #0 never competes for bandwidth with GPU #1 (which uses Rail 1), GPU #2 (Rail 2), or any other GPU. Each GPU position within the rack has its own physically isolated network path to corresponding GPUs in all other racks.</p><p><strong>The result: Elimination of contention.</strong> By segregating traffic based on GPU position within the rack, the Rail-Optimized topology ensures that the massive, synchronized flows of AI training never cross paths. It effectively creates 72 separate, non-blocking supercomputers operating in parallel across all racks.</p><p><strong>As the rack-scale node grows</strong> (144 GPUs in NVL144, 572 GPUs in NVL572), the Rail principle scales proportionally. An NVL144 cluster would use 144 Rails; an NVL572 cluster would use 572 Rails. The larger the Domain 1 unit, the more Rails required&#8212;but the architecture remains the same: one Rail per GPU position, connecting corresponding GPUs across all super-nodes in the cluster.</p><h4>In-Network Computing (SHARP)</h4><p>The Quantum-X800 switches include a critical InfiniBand feature: SHARP v4 (Scalable Hierarchical Aggregation and Reduction Protocol). Traditional All-Reduce requires each GPU to send data to a designated &#8220;reducer&#8221; GPU, which adds the values and sends the result back. SHARP moves this arithmetic into the switch itself.</p><p>When packets arrive at the Quantum-X800, an ALU (Arithmetic Logic Unit) inside the switch performs the addition as packets pass through. Instead of requiring N hops to aggregate data from N GPUs, SHARP reduces the operation to log(N) hops. For a cluster with 8 NVL72 racks (576 GPUs total), this cuts network traffic by roughly 50% and latency by 2-3x. The switch is no longer just routing packets&#8212;it&#8217;s computing.</p><h4>Scaling Across the Data Center</h4><p>When clusters scale to tens of thousands of GPUs&#8212;requiring hundreds of rack-scale super-nodes&#8212;they enter additional layers of switching. Here, NVIDIA uses additional Quantum-X800 switches configured as &#8220;Spine&#8221; switches to interconnect groups of racks distributed across the data center.</p><p><strong>The physical challenge:</strong> Racks may be 50-100 meters apart, far beyond copper cable range. The solution is optical fiber with active transceivers in the switches. Each Quantum-X800 Spine switch converts electrical signals to light, routes the optical packets, and converts back to electricity at the destination.</p><p><strong>The latency penalty:</strong> Each optical hop adds approximately 200 nanoseconds&#8212;not much in human terms, but significant when GPUs operate at 4 GHz (one clock cycle every 0.25 nanoseconds). For operations that must traverse multiple spine switches, this latency accumulates. This is why NVIDIA&#8217;s roadmap focuses on expanding the rack-scale super-node&#8212;by keeping more GPUs within Domain 1 (72 &#8594; 144 &#8594; 572 GPUs), fewer operations need to traverse the higher-latency Domain 2 fabric.</p><p><strong>The rail structure extends:</strong> Even at the spine level, the Rail-Optimized topology persists. Whether connecting 72-GPU racks, 144-GPU racks, or future 572-GPU super-nodes, the Rails remain physically separate all the way to the top of the hierarchy, ensuring zero contention even in 100,000-GPU clusters.</p><h4>The Ethernet Alternative: Spectrum-X</h4><p>While InfiniBand remains NVIDIA&#8217;s highest-performance option for dedicated AI supercomputers, they also offer Spectrum-X Ethernet for customers integrating AI into existing cloud infrastructures.</p><p>Spectrum-X uses standard Ethernet cabling but modifies protocol behavior to mimic InfiniBand&#8217;s lossless characteristics. It employs:</p><ul><li><p>RoCE v2 (RDMA over Converged Ethernet) for direct memory access</p></li><li><p>Adaptive Routing that dynamically sprays packets across all available paths</p></li></ul><p>This allows AI traffic to coexist with traditional cloud workloads while maintaining 95% effective bandwidth utilization&#8212;approaching InfiniBand performance while preserving Ethernet&#8217;s ubiquity and interoperability.</p><h4>Software Orchestration: NCCL (Dynamic Discovery)</h4><p>NVIDIA&#8217;s communication library, NCCL (NVIDIA Collective Communications Library), is designed for the flexibility that hierarchical networks require.</p><p>When an NVIDIA cluster powers on, NCCL performs dynamic discovery: it &#8220;looks around&#8221; to see which GPUs are available, measures the network topology, and builds an optimal communication tree on the fly. If one of the thousands of switches in a massive cluster fails, NCCL detects the failure and dynamically reroutes traffic around it. The cluster continues operating, potentially at slightly reduced performance, rather than halting entirely.</p><p>This flexibility comes at a cost: NCCL&#8217;s runtime overhead. Every communication operation requires negotiation&#8212;checking which path is optimal, managing buffers, handling unexpected congestion. For stable, predictable workloads, this overhead is pure inefficiency. But for real-world deployments where hardware fails, cables get unplugged, and maintenance windows require partial cluster shutdowns, NCCL&#8217;s resilience is essential.</p><h3>4. Google&#8217;s Implementation: The Flat Optical Mesh</h3><p>Google&#8217;s network architecture rejects hierarchy entirely&#8212;at least within a Pod. Instead of building layers of progressively slower networks, they group thousands of chips into a single, unified supercomputer called a &#8220;Pod.&#8221; Inside this Pod, a massive optical fabric ensures that every TPU has identical bandwidth and latency to every other TPU. By flattening the topology, Google creates a grid where thousands of chips can communicate as if they were all next to each other.</p><h4>Domain 1: The Pod as Super-Node (ICI)</h4><p>As introduced in Article 5, Google&#8217;s Inter-Chip Interconnect (ICI) creates a 3D Torus topology. Each TPU connects directly to six neighbors&#8212;north, south, east, west, up, and down&#8212;via dedicated 600 GB/s optical links.</p><p><strong>The key property: Uniform distance.</strong> In a 3D Torus, every chip is equidistant from every other chip in terms of network hops. A TPU in position (0,0,0) reaches a TPU in position (32,32,16) through exactly the same number of hops as it would reach (16,16,32). There are no &#8220;fast&#8221; connections near the &#8220;center&#8221; and no &#8220;slow&#8221; connections at the &#8220;edge&#8221;&#8212;because there is no center or edge.</p><p><strong>Circuit-switched, not packet-switched:</strong> Unlike NVIDIA&#8217;s InfiniBand network, which routes packets dynamically based on congestion, Google&#8217;s ICI uses optical circuit switching. At the start of a training run, the Optical Circuit Switches (OCS) physically rotate thousands of tiny MEMS mirrors to create dedicated optical paths between specific TPU pairs. Once these paths are established, data flows as continuous light beams without packet headers, routing decisions, or buffering.</p><p><strong>The performance implication:</strong> Google&#8217;s 600 GB/s per-chip bandwidth is lower than NVIDIA&#8217;s 1.8 TB/s NVLink. However, the circuit-switched nature means there is zero packet overhead&#8212;no headers, no routing lookups, no buffering delays. The full 600 GB/s is available bandwidth, and latency is deterministic: exactly the same number of hops from any chip to any other.</p><h4>The Optical Circuit Switch (OCS): Enabling the Flat Mesh</h4><p>The physical device enabling Google&#8217;s flat mesh is the <strong>O</strong>ptical Circuit Switch, internally codenamed &#8220;Palomar.&#8221; This is one of Google&#8217;s most closely guarded hardware innovations.</p><p>The technology: A Palomar OCS is a 136&#215;136 port switch containing thousands of tiny MEMS (Micro-Electro-Mechanical Systems) mirrors. Each mirror is roughly the width of a human hair and can rotate on a microscopic gimbal. When a beam of light enters the OCS, a mirror physically redirects that beam to one of 136 output ports, creating a direct optical connection.</p><p><strong>No electrical conversion:</strong> In the OCS, photons enter as light, bounce off mirrors, and exit as light. This eliminates the ~200 nanosecond penalty of optical-electrical-optical conversion and saves power (no lasers needed in the switch itself).</p><p><strong>Reconfigurability:</strong> The MEMS mirrors can physically rotate to new positions in milliseconds. This means Google can reconfigure the network topology between training runs. If one experiment requires a Torus topology and another requires a Dragonfly topology, the mirrors rotate, and the network physically becomes that topology. The hardware is programmable at the optical layer.</p><p><strong>The Apollo fabric layer:</strong> Multiple Palomar OCS units are arranged into the &#8220;Apollo&#8221; optical switching platform. Apollo acts as a building-scale reconfigurable patch panel, connecting thousands of TPU server trays into the desired mesh topology. For a 9,216-chip TPU v7 Pod, hundreds of OCS units work in concert to create the 3D Torus mesh.</p><h4>The Pod vs. The Rack: Optical Freedom vs. Copper Constraints</h4><p>To understand Google&#8217;s architecture, one must distinguish between NVIDIA&#8217;s physically-based Domain 1 and Google&#8217;s optically-based Domain 1.</p><p><strong>NVIDIA&#8217;s Domain 1: The Physical Rack</strong></p><p>For NVIDIA, the fundamental building block of Domain 1 is the copper-based rack. This is a hard physical boundary defined by copper physics. The 72 GPUs inside are bound together by a copper backplane with a maximum range of ~1 meter. You cannot dynamically decide to make a &#8220;100-GPU Rack&#8221; or a &#8220;50-GPU Rack&#8221; without physically rewiring the hardware.</p><p>The rack is fixed. Once manufactured, its boundaries are immutable.</p><p><strong>Google&#8217;s Domain 1: The Optical Pod</strong></p><p>For Google, the fundamental building block of Domain 1 is the optically-based Pod. A Pod is the maximum ICI-connected configuration&#8212;up to 9,216 chips for TPU v7 (also called a &#8220;SuperPod&#8221; in some Google documentation). </p><p>This maximum is determined by the physical limits of the Optical Circuit Switch fabric: the reach of optical fibers and the port count of the OCS units. Because the cabling is optical (fiber) rather than electrical (copper), distance is not the limiting factor that constrains NVIDIA&#8212;but there is still a practical maximum to how many chips can be woven into a single deterministic mesh.</p><p><strong>Physical flexibility:</strong> The Pod can span multiple rows of physical racks. The OCS mirrors define which TPUs connect to which, regardless of their physical location in the data center&#8212;as long as they&#8217;re within fiber range.</p><p><strong>Configurable domains:</strong> Google can provision compute domains of various sizes from the Pod (64 chips, 512 chips, 4,096 chips, etc.) by programming the OCS mirrors. The same physical hardware can be reconfigured to serve different topologies.</p><p><strong>Isolation:</strong> Each provisioned domain is optically isolated. Traffic within one domain never contends with traffic from other domains, even if they share the same physical racks.</p><p>Where NVIDIA scales by deploying more physical racks, Google scales by connecting more TPUs into larger optical Pods, then provisioning domains of appropriate sizes from that hardware pool.</p><h4>Domain 2: Scaling Beyond the Pod (Multislice)</h4><p>While the OCS allows for massive Pods, there is a physical limit to how many chips can be connected in a single low-latency ICI mesh&#8212;currently 9,216 chips for TPU v7. This limit is determined by the optical fiber reach and the port count of the Optical Circuit Switches. Within this boundary, every chip communicates via the ultra-fast, proprietary ICI mesh.</p><p>To scale beyond 9,216 chips, Google employs <strong>Multislice</strong>&#8212;a technology that connects multiple full Pods together. For example, connecting two TPU v7 Pods via Multislice creates an 18,432-chip system. The largest Multislice deployments (reportedly used for training Gemini Ultra) have spanned tens of thousands of chips, requiring multiple fully-populated Pods connected together.</p><p>This architecture bifurcates traffic into two domains:</p><p>Intra-Pod (Domain 1): Traffic remains on the deterministic ICI fabric (the 3D Torus). Here, XLA&#8217;s compile-time scheduling ensures perfect synchronization.</p><p>Inter-Pod (Domain 2): Traffic traverses the Jupiter Data Center Network (DCN). This is standard Ethernet with Google&#8217;s Swift protocol for enhanced congestion control.</p><p>This functionally mirrors NVIDIA&#8217;s spine layer. While local traffic within a Pod enjoys the &#8220;clockwork&#8221; precision of optical circuits, traffic between Pods enters a standard packet-switched network that behaves more like the traditional internet&#8212;dynamic and subject to minor jitter. This introduces the two-domain structure we saw earlier: absolute determinism within the Pod, and managed dynamism between Pods.</p><p><strong>Jupiter&#8217;s role:</strong> Jupiter is Google&#8217;s unified data center network architecture. It connects not just TPU Pods to each other (in Multislice configurations), but also TPUs to Google&#8217;s massive storage infrastructure (Colossus), CPU fleets, and external internet gateways. By using standard Ethernet frames (with customized protocols), Jupiter enables interoperability across Google&#8217;s entire infrastructure.</p><h4>Software Orchestration: XLA (Deterministic Clockwork)</h4><p>Google&#8217;s compiler, XLA (Accelerated Linear Algebra), takes a radically different approach from NVIDIA&#8217;s NCCL. XLA assumes perfect knowledge of the network topology and perfect reliability.</p><p>Compile-time scheduling: When a JAX program is compiled, XLA analyzes the computation graph and the physical TPU mesh topology. It calculates the exact nanosecond when each piece of data will leave Chip A and arrive at Chip B. There is no runtime negotiation, no dynamic routing, no &#8220;checking if the path is clear.&#8221; The schedule is computed once, at compile time, and executed blindly.</p><p>No handshakes: In NCCL, a sender waits for acknowledgment before transmitting. In XLA, senders transmit without waiting. They know&#8212;mathematically&#8212;that the receiver will be ready because XLA scheduled the receiver&#8217;s computation to complete exactly when the data arrives. This &#8220;clockwork execution&#8221; eliminates all handshake overhead but requires absolute predictability.</p><p>The trade-off: If a TPU fails mid-computation, XLA cannot dynamically reroute. The entire training job typically must restart from the last checkpoint. Google accepts this trade-off because their infrastructure is designed for extremely high reliability (more on this in Section 6), and the performance gains from eliminating all runtime overhead outweigh the cost of occasional restarts.</p><h3>5. The Physics of Data Movement: Massive Flows and Synchronization</h3><p>To understand why these radically different topologies exist, we must examine the traffic patterns they&#8217;re designed to handle. AI training doesn&#8217;t generate random, bursty traffic like a web server. It generates synchronized, massive data transfers.</p><h3>The All-Reduce Storm</h3><p>The fundamental operation in data-parallel training is All-Reduce. At the end of each training step, every GPU has computed gradients (updates to the model weights) based on its batch of data. To calculate the true average gradient, every GPU must share its results with every other GPU.</p><p><strong>The traffic pattern:</strong></p><ul><li><p>Volume: For a trillion-parameter model, each All-Reduce operation moves terabytes of data</p></li><li><p>Timing: It happens simultaneously across all chips&#8212;every GPU sends data at the exact same millisecond</p></li><li><p>Synchronization: Training cannot proceed to the next step until every GPU has received every other GPU&#8217;s gradients</p></li></ul><p>This is fundamentally different from traditional networking workloads. There&#8217;s no &#8220;bursty&#8221; traffic&#8212;it&#8217;s a continuous drumbeat of synchronized massive flows. Every few milliseconds, the entire supercomputer pauses to perform All-Reduce, then resumes computation.</p><h3>Tail Latency: The Straggler Problem</h3><p>In a synchronized system, performance is determined by the <strong>slowest operation</strong>, not the average operation. If 99,999 packets arrive in 10 microseconds but a single packet takes 10 milliseconds because it got stuck in a switch buffer, the entire 100,000-GPU cluster halts for 10 milliseconds waiting for that straggler.</p><h3>Congestion Control: Different Philosophies</h3><p>Both companies have engineered solutions to eliminate tail latency, but through opposite approaches.</p><p><strong>NVIDIA (Reactive):</strong> InfiniBand uses credit-based flow control. A sender cannot transmit until the receiver has explicitly signaled that buffer space is available. This makes InfiniBand &#8220;lossless by design&#8221;&#8212;packet drops are physically impossible. When congestion occurs, the network pushes back against senders, forcing them to slow down.</p><p>The Quantum-X800 adds adaptive routing on top of this: when one path gets congested, traffic automatically sprays across alternative paths. Individual packets may take different routes, but they all arrive reliably, and the receiving NIC reassembles them in order.</p><p>Result: NVIDIA&#8217;s approach handles congestion reactively. The network continuously monitors its own state and adjusts routing dynamically to avoid bottlenecks. This works even when traffic patterns are unpredictable or when hardware components are operating at different speeds due to thermal throttling or partial failures.</p><p><strong>Google (Preventative):</strong> The ICI fabric doesn&#8217;t have &#8220;congestion&#8221; in the traditional sense because communication is scheduled at compile time. XLA knows exactly which chips will communicate when, and it schedules operations such that no two transfers ever collide on the same optical circuit.</p><p>If the compiler cannot find a valid schedule (because the requested communication pattern would cause congestion), the program fails to compile rather than failing at runtime. The developer must redesign the algorithm to fit the hardware&#8217;s communication capacity.</p><p>Result: Google&#8217;s approach prevents congestion through perfect planning. The network never encounters unexpected traffic because every data movement was calculated in advance. This requires predictable workloads and stable hardware but delivers maximum efficiency when those conditions are met.</p><h3>6. Two Network Operations Doctrines</h3><p>The architectural differences between NVIDIA and Google&#8217;s networks create fundamentally different operational philosophies. How do you keep a supercomputer running when components inevitably fail?</p><h3>NVIDIA: &#8220;Detect &amp; Adapt&#8221;</h3><p>NVIDIA&#8217;s approach is built on the assumption that hardware will fail, and the system must gracefully adapt.</p><p><strong>The architecture enables flexibility:</strong> Because InfiniBand uses dynamic routing and NCCL performs runtime discovery, the system can route around failures. When a switch fails, NCCL detects it (usually within seconds) and rebuilds the communication tree using alternative paths.</p><p><strong>Switch failure:</strong> If a Quantum-X800 switch fails, NCCL dynamically reroutes traffic through alternative switches. The cluster continues operating, potentially with reduced bandwidth on certain paths, but without halting.</p><p><strong>Rack failure:</strong> If an entire NVL72 rack fails (power outage, cooling failure, etc.), the cluster can isolate that rack and continue training with the remaining racks. For a data-parallel workload trained across 1,000 racks, losing one rack means restarting from the last checkpoint with 999 racks&#8212;annoying but not catastrophic.</p><p><strong>Cable failure:</strong> Individual cable failures are detected automatically. NCCL marks the failed path as unavailable and routes around it. Cables can be replaced during maintenance windows without shutting down the entire cluster.</p><p><strong>The cost: Performance variability.</strong> With dynamic routing and rerouting, job-to-job performance varies based on which specific hardware components are currently operational. A training run might take 10% longer this week than last week because two switches are down for maintenance. NVIDIA accepts this variability in exchange for continuous operation.</p><p><strong>Monitoring and telemetry:</strong> NVIDIA&#8217;s infrastructure relies heavily on runtime monitoring. Every switch, cable, and NIC continuously reports health metrics. When anomalies are detected (increased error rates, higher-than-expected latency), the system can proactively isolate potentially failing components before they cause job failures.</p><h3>Google: &#8220;Predict &amp; Purge&#8221;</h3><p>Google&#8217;s approach assumes that with sufficient care, hardware won&#8217;t fail&#8212;and when it does, you remove it before it causes problems.</p><p><strong>The architecture requires perfection:</strong> Because XLA schedules communication down to the nanosecond, a single &#8220;slow&#8221; chip (not even broken, just lagging due to thermal issues) breaks the global clockwork. All chips must operate in perfect synchrony, or the deterministic schedule collapses.</p><p><strong>Aggressive telemetry:</strong> Google&#8217;s management software (Borg/GKE) constantly monitors error rates, thermal variance, and performance metrics. If a chip shows pre-failure symptoms&#8212;slightly elevated error rates, minor thermal throttling, inconsistent latency&#8212;the system proactively evicts the workload from that Pod or migrates it to healthy hardware.</p><p><strong>Proactive replacement:</strong> Rather than waiting for components to fail, Google uses telemetry to predict failures. A TPU showing signs of degradation is removed from service during scheduled maintenance and replaced before it impacts production workloads.</p><p><strong>Frequent checkpointing:</strong> Training jobs checkpoint every few minutes. When a failure occurs (or when a component is proactively removed), the job restarts from the most recent checkpoint, losing only minutes of work. The cost of restarting is low enough that dynamic rerouting is unnecessary.</p><p><strong>The benefit: Predictable performance.</strong> Every training run on a given Pod configuration achieves identical performance because there&#8217;s no dynamic routing introducing variability. This makes capacity planning straightforward and performance debugging easier&#8212;if a job is slower than expected, it&#8217;s a software problem, not a hardware configuration issue.</p><h3>7. Which Network Topology Is Better?</h3><p>Both architectures successfully train trillion-parameter models, but they optimize for different values and constraints.</p><p><strong>NVIDIA&#8217;s hierarchical, resilient approach is ideal for:</strong></p><ul><li><p>Multi-tenant cloud environments where diverse customers run varied workloads with different scaling requirements</p></li><li><p>Organizations with varying operational capabilities that may not maintain Google-level infrastructure discipline</p></li><li><p>Workloads requiring flexibility where jobs must adapt to partial cluster availability or hardware heterogeneity</p></li><li><p>Incremental scaling where infrastructure grows gradually (72 &#8594; 144 &#8594; 572 GPUs) rather than in massive Pods</p></li></ul><p><strong>Google&#8217;s flat mesh, deterministic approach is ideal for:</strong></p><ul><li><p>Single-tenant research environments training frontier models where the entire cluster serves one purpose</p></li><li><p>Organizations with infrastructure maturity capable of maintaining ultra-high reliability through proactive management</p></li><li><p>Workloads demanding performance predictability where consistent iteration time accelerates research progress</p></li><li><p>Large-scale deployments where provisioning entire 9,216-chip Pods makes economic sense</p></li></ul><p>Each architecture reflects different engineering values:</p><ul><li><p>NVIDIA values resilience and flexibility&#8212;the network must work in messy, real-world conditions with imperfect hardware</p></li><li><p>Google values efficiency and predictability&#8212;the network operates as deterministic clockwork, assuming infrastructure excellence</p></li></ul><p>Both approaches have successfully trained the world&#8217;s largest models. The choice depends not on technical superiority but on organizational fit.</p><h3>8. Conclusion: The Two-Tier Reality</h3><p>The network topologies engineered by NVIDIA and Google represent the two dominant philosophies of the AI era: NVIDIA&#8217;s resilient hierarchy versus Google&#8217;s deterministic mesh. Yet, as models scale beyond 100,000 chips, physics is forcing a structural similarity that neither company explicitly planned. Both companies have discovered that a trillion-parameter cluster requires two distinct domains:</p><p><strong>Domain 1: The &#8220;Super-Node&#8221; Fabric</strong> - A massive, proprietary, ultra-low-latency island where compute is tightly coupled.</p><p>For Google, this is the Pod (up to 9,216 TPUs in TPU v7). Inside this boundary, the optical mesh creates a deterministic, flat &#8220;bubble&#8221; of perfect synchronization. Every chip is equidistant from every other, and XLA schedules communication with nanosecond precision.</p><p>For NVIDIA, this is the Rack (72 to 572 GPUs connected via NVLink). By moving to a copper backplane, NVIDIA has essentially turned the rack into a single giant GPU, mimicking the tight coupling of a Google Pod&#8212;just at smaller scale with different trade-offs.</p><p><strong>Domain 2: The Scale-Out Fabric</strong> - A standard, packet-switched network to connect these islands.</p><p>For Google, this is Multislice (Jupiter). They have conceded that the flat mesh cannot scale infinitely. To grow beyond 9,216 chips, they must introduce hierarchy, connecting Pods via standard data center networking. Traffic between Pods uses packet-switched Ethernet, entering a world of dynamic routing and managed congestion&#8212;exactly what they avoided within the Pod.</p><p>For NVIDIA, this is the InfiniBand/Ethernet fabric with Rail-Optimized topology. They use this to bridge their massive-scale racks, employing SHARP in-network computing and adaptive routing to maintain high performance across building-scale distances.</p><p>Both have accepted the two-domain structure (super-node + scale-out). Google&#8217;s flat mesh uses a hierarchy to scale beyond the Pod (Multislice). NVIDIA&#8217;s hierarchy uses massive &#8220;flat&#8221; nodes to scale (NVL72/144/572). Zoom out, and we see huge islands of determinism connected by oceans of dynamic networking.</p><p><strong>A fundamental philosophical divide persists</strong>: hierarchical versus flat, dynamic versus deterministic, resilient versus efficient. </p><p>NVIDIA believes the network should adapt to the hardware&#8217;s imperfections. Google believes the hardware should be perfectly configured by software.</p><p>These are not merely technical choices&#8212;they reflect different assumptions about how a supercomputer should be built and operated.</p><p>In the next article, <strong>&#8220;Parallelism: The Blueprint of Training&#8221;, </strong> we will leave the physical layer and move up the stack. We will examine how these topological choices dictate the specific parallelism strategies for training trillion-parameter models.</p>]]></content:encoded></item><item><title><![CDATA[Two Compute Nodes: Physical vs. Virtual]]></title><description><![CDATA[The Gigawatt Machine: NVIDIA, Google, and the Engineering of Scale 5/12]]></description><link>https://www.gigawattmachine.com/p/two-compute-nodes-physical-vs-virtual</link><guid isPermaLink="false">https://www.gigawattmachine.com/p/two-compute-nodes-physical-vs-virtual</guid><dc:creator><![CDATA[Tony Wan]]></dc:creator><pubDate>Thu, 11 Dec 2025 00:49:23 GMT</pubDate><enclosure url="https://substackcdn.com/image/fetch/$s_!vqus!,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F59867208-fae5-40b9-a570-54b2271d292a_1280x1280.png" length="0" type="image/jpeg"/><content:encoded><![CDATA[<h3>1. Introduction</h3><p>In the previous articles, we examined the silicon engines (Article 2), the density strategies (Article 3), and the software ecosystems (Article 4) that power AI supercomputers. Now we turn to the fundamental question: What is the atomic unit of computation?</p><p>For the last decade, the industry agreed on a definition: a &#8220;node&#8221; was a discrete, physical server chassis housing 8 GPUs. This &#8220;supercomputer-in-a-box&#8221; (like the NVIDIA DGX) was the standard building block. You connected thousands of these nodes with high-speed networking to build a cluster.</p><p>However, as we enter the Gigawatt era,the definition of the &#8220;compute node&#8221; has bifurcated into two contrasting philosophies, each driven by the architectural choices we explored in Article 1:</p><p><strong>Path A (NVIDIA): The Physical Evolution.</strong> The node is getting bigger. NVIDIA has evolved from a server-scale node (8 GPUs) to a rack-scale node (72+ GPUs), turning the entire cabinet into a single, liquid-cooled monolith. The strategy is to compress maximum compute into a defined physical volume.</p><p><strong>Path B (Google): The Virtual Evolution.</strong> The node is disappearing. Google has abstracted away the physical hardware, treating the &#8220;node&#8221; as a virtual entry point into a massive, flat mesh of TPUs. The strategy is to make the physical packaging irrelevant so the fabric can grow without bounds.</p><p>This article deconstructs how the physics of copper, cooling, and fabric have driven NVIDIA to build bigger boxes, while Google has made the box disappear entirely.</p><h3>2. The NVIDIA Path: Physical Scale-Up</h3><p>NVIDIA&#8217;s philosophy is rooted in density. To make models run faster, compress the compute into the smallest possible physical space. This has driven a two-phase evolution: from the server to the rack.</p><h4>Phase 1: The Server-Scale Node (DGX B200)</h4><p>For enterprise deployments and standard clusters, the atomic unit is the 8-GPU server. The DGX B200 represents the current instantiation of this design philosophy&#8212;the most powerful air-cooled system in a server footprint.</p><p><strong>Why 8 GPUs? The Engineering Limits</strong></p><p>The number 8 results from two hard physical constraints:</p><ul><li><p>The Copper Limit: At NVLink speeds (1.8 TB/s bidirectional), electrical signals degrade rapidly. The 8-GPU chassis is the maximum size where you can connect all chips via passive copper traces on a motherboard without signal integrity issues.</p></li><li><p>Switch Radix: Eight GPUs saturate the port count of the internal NVSwitch chips. Adding a 9th GPU would break the non-blocking fabric architecture.</p></li></ul><p><strong>The Internal Fabric: A Single NVLink Domain</strong></p><p>Inside the DGX B200, the 8 GPUs function as one unified system:</p><ul><li><p>GPU-to-GPU (East-West): 18 NVLink 5.0 connections per GPU provide 1.8 TB/s of bidirectional bandwidth. This switching happens entirely inside the box.</p></li><li><p>CPU-to-GPU (North-South): The system uses the GB200 Superchip, replacing legacy PCIe with NVLink-C2C. This fuses the Grace CPU and Blackwell GPU into a single memory space, allowing GPUs to access the 960GB of LPDDR5X system memory as their own.</p></li></ul><p><strong>The Thermal Wall: Pushing Air Cooling to Its Limit</strong></p><ul><li><p>Form Factor: The DGX B200 expanded from 8U (DGX H100) to 10U to accommodate larger heatsinks.</p></li><li><p>Power Draw: 14.3 kW&#8212;a 40% increase over the 10.2 kW DGX H100.</p></li><li><p>Deployment Challenge: At this power density, operators must leave empty rack slots to prevent &#8220;hot spots,&#8221; and fans scream at 10,000+ RPM.</p></li></ul><p>The DGX B200 delivers 144 PetaFLOPS (FP8) per node, but it represents the absolute limit of what air cooling can achieve.</p><h4>Phase 2: The Rack-Scale Node (GB200 NVL72)</h4><p>To break the 8-GPU limit, NVIDIA redefined the physical boundaries of the node. The node is no longer a server; the node is the entire rack.</p><p>The Monolith: 72 GPUs as One Computer</p><p>The GB200 NVL72 is a 120 kW, liquid-cooled, 42U rack. It is not a collection of servers&#8212;it&#8217;s a single computer:</p><ul><li><p>The Architecture: 72 B200 GPUs and 36 Grace CPUs are hard-wired together with a massive copper backplane (the &#8220;spine&#8221;) that runs vertically through the cabinet.</p></li><li><p>The Copper Advantage: Because the GPUs are stacked vertically in one cabinet, NVIDIA can connect all 72 using passive copper cabling. This avoids power-hungry optical transceivers, saving approximately 20kW per rack.</p></li><li><p>Unified Memory: The entire system functions as one giant accelerator with 31 TB of unified memory (13.8 TB HBM3e + 17.3 TB LPDDR5X).</p></li></ul><p><strong>Performance: The Generational Leap</strong></p><p>The evolution from Hopper through Blackwell and into the roadmapped Vera Rubin platforms reveals the trajectory of rack-scale computing:</p><div class="captioned-image-container"><figure><a class="image-link image2 is-viewable-img" target="_blank" href="https://substackcdn.com/image/fetch/$s_!sKHF!,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F8e37bd83-ddcb-487c-bcbb-a92924f61c92_1510x764.png" data-component-name="Image2ToDOM"><div class="image2-inset"><picture><source type="image/webp" srcset="https://substackcdn.com/image/fetch/$s_!sKHF!,w_424,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F8e37bd83-ddcb-487c-bcbb-a92924f61c92_1510x764.png 424w, https://substackcdn.com/image/fetch/$s_!sKHF!,w_848,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F8e37bd83-ddcb-487c-bcbb-a92924f61c92_1510x764.png 848w, https://substackcdn.com/image/fetch/$s_!sKHF!,w_1272,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F8e37bd83-ddcb-487c-bcbb-a92924f61c92_1510x764.png 1272w, https://substackcdn.com/image/fetch/$s_!sKHF!,w_1456,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F8e37bd83-ddcb-487c-bcbb-a92924f61c92_1510x764.png 1456w" sizes="100vw"><img src="https://substackcdn.com/image/fetch/$s_!sKHF!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F8e37bd83-ddcb-487c-bcbb-a92924f61c92_1510x764.png" width="1456" height="737" data-attrs="{&quot;src&quot;:&quot;https://substack-post-media.s3.amazonaws.com/public/images/8e37bd83-ddcb-487c-bcbb-a92924f61c92_1510x764.png&quot;,&quot;srcNoWatermark&quot;:null,&quot;fullscreen&quot;:null,&quot;imageSize&quot;:null,&quot;height&quot;:737,&quot;width&quot;:1456,&quot;resizeWidth&quot;:null,&quot;bytes&quot;:274295,&quot;alt&quot;:null,&quot;title&quot;:null,&quot;type&quot;:&quot;image/png&quot;,&quot;href&quot;:null,&quot;belowTheFold&quot;:true,&quot;topImage&quot;:false,&quot;internalRedirect&quot;:&quot;https://www.gigawattmachine.com/i/181195531?img=https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F8e37bd83-ddcb-487c-bcbb-a92924f61c92_1510x764.png&quot;,&quot;isProcessing&quot;:false,&quot;align&quot;:null,&quot;offset&quot;:false}" class="sizing-normal" alt="" srcset="https://substackcdn.com/image/fetch/$s_!sKHF!,w_424,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F8e37bd83-ddcb-487c-bcbb-a92924f61c92_1510x764.png 424w, https://substackcdn.com/image/fetch/$s_!sKHF!,w_848,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F8e37bd83-ddcb-487c-bcbb-a92924f61c92_1510x764.png 848w, https://substackcdn.com/image/fetch/$s_!sKHF!,w_1272,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F8e37bd83-ddcb-487c-bcbb-a92924f61c92_1510x764.png 1272w, https://substackcdn.com/image/fetch/$s_!sKHF!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F8e37bd83-ddcb-487c-bcbb-a92924f61c92_1510x764.png 1456w" sizes="100vw" loading="lazy"></picture><div class="image-link-expand"><div class="pencraft pc-display-flex pc-gap-8 pc-reset"><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container restack-image"><svg role="img" width="20" height="20" viewBox="0 0 20 20" fill="none" stroke-width="1.5" stroke="var(--color-fg-primary)" stroke-linecap="round" stroke-linejoin="round" xmlns="http://www.w3.org/2000/svg"><g><title></title><path d="M2.53001 7.81595C3.49179 4.73911 6.43281 2.5 9.91173 2.5C13.1684 2.5 15.9537 4.46214 17.0852 7.23684L17.6179 8.67647M17.6179 8.67647L18.5002 4.26471M17.6179 8.67647L13.6473 6.91176M17.4995 12.1841C16.5378 15.2609 13.5967 17.5 10.1178 17.5C6.86118 17.5 4.07589 15.5379 2.94432 12.7632L2.41165 11.3235M2.41165 11.3235L1.5293 15.7353M2.41165 11.3235L6.38224 13.0882"></path></g></svg></button><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container view-image"><svg xmlns="http://www.w3.org/2000/svg" width="20" height="20" viewBox="0 0 24 24" fill="none" stroke="currentColor" stroke-width="2" stroke-linecap="round" stroke-linejoin="round" class="lucide lucide-maximize2 lucide-maximize-2"><polyline points="15 3 21 3 21 9"></polyline><polyline points="9 21 3 21 3 15"></polyline><line x1="21" x2="14" y1="3" y2="10"></line><line x1="3" x2="10" y1="21" y2="14"></line></svg></button></div></div></div></a></figure></div><p>The GB200 NVL72 achieves 81x the raw compute of the DGX H100, while simultaneously improving power efficiency by 6.4x (from ~1.57 to ~10 PetaFLOPS per kilowatt). The roadmap shows NVIDIA&#8217;s commitment to continuing this rack-scale expansion, with the Vera Rubin Ultra NVL572 targeting nearly 4x the performance of today&#8217;s GB200 NVL72.</p><p><strong>Liquid Cooling: The Infrastructure Mandate</strong></p><p>At 120 kW per rack, air cooling is physically impossible. The GB200 NVL72 uses direct-to-chip (DTC) liquid cooling, where cold plates are mounted on every GPU and CPU:</p><ul><li><p>Water Requirement: Standard facility water (inlet temperature up to 25&#176;C / 77&#176;F).</p></li><li><p>Infrastructure Impact: Data centers must install industrial-scale hydraulic systems, circulating millions of gallons of coolant.</p></li></ul><p>As rack-scale nodes continue to grow (144 GPUs, then 572 GPUs), the power density and cooling requirements will intensify further, driving data centers toward becoming industrial power plants with integrated hydraulic infrastructure.</p><p><strong>The Network: Pushing the Bottleneck</strong></p><p>The critical innovation is where the &#8220;cluster&#8221; begins:</p><ul><li><p>DGX B200 (Server-Scale): The cluster begins immediately outside the 8-GPU chassis. Even communication between servers in the same rack must traverse the external InfiniBand network (800 Gb/s).</p></li><li><p>GB200 NVL72 (Rack-Scale): The cluster begins only when connecting one rack to another. All 72 GPUs communicate over the internal NVLink fabric, avoiding the external network entirely. This pushes the network bottleneck from every 8 GPUs to every 72 GPUs&#8212;a 9x reduction in inter-node traffic.</p></li></ul><p><strong>Scaling Further: The Multi-Generation Roadmap</strong></p><p>The GB200 NVL72 is not the endpoint of NVIDIA&#8217;s rack-scale vision. Using the 5th-generation and future NVLink Switch Systems, NVIDIA is systematically expanding the maximum size of a single NVLink Domain:</p><p>Today - Blackwell (2024-2025):</p><ul><li><p>GB200 NVL72: 72 GPUs in a single rack</p></li><li><p>GB200 NVL72 Pods: Up to 8 racks = 576 GPUs in one continuous NVLink Domain</p></li></ul><p>Near Future - Vera Rubin (2026):</p><ul><li><p>Rubin NVL144: 144 GPUs in a multi-rack configuration</p></li><li><p>This effectively doubles the size of the rack-scale node, allowing massive models to remain within a single, high-speed memory space without crossing into slower inter-rack networking.</p></li></ul><p>Far Future - Vera Rubin Ultra (2027):</p><ul><li><p>Rubin Ultra NVL572: 572 GPUs in a single NVLink Domain</p></li><li><p>At this scale, the &#8220;Super-Node&#8221; effectively becomes a small supercomputer, with nearly 600 GPUs addressable as one giant, coherently-cached brain.</p></li><li><p>The 5,000 PetaFLOPS of compute in a single domain approaches the entire computational capacity of early supercomputing facilities.</p></li></ul><p>This roadmap reveals NVIDIA&#8217;s long-term strategy: expand the physical node to encompass what was once an entire cluster. By 2027, the rack-scale node will be 8x larger than today (572 vs 72 GPUs), effectively turning multiple aisles of racks into one massive, unified logic gate.</p><p>Implication: The &#8220;network bottleneck&#8221; that limits distributed training will be pushed from every 72 GPUs (today) to every 572 GPUs (2027)&#8212;nearly eliminating it for most trillion-parameter model training workloads.</p><h3>3. The Google Path: Virtual Scale-Out</h3><p>Google&#8217;s philosophy is rooted in fabric scale. They prioritize the size of the network over the density of the box. Consequently, their &#8220;node&#8221; story is not about building a bigger box, but about making the box irrelevant so the matrix can grow.</p><h4>The &#8220;Invisible&#8221; Physical Node</h4><p>If you walked into a Google data center and pulled out a TPU server tray, you would hold a relatively modest piece of hardware: a standard-sized board with 4 or 8 TPU chips.</p><p>Unlike the NVIDIA DGX, this physical node is architecturally transparent:</p><ul><li><p>No Local Switch: There is no equivalent to NVSwitch inside the tray.</p></li><li><p>Passthrough Design: Every single TPU chip has optical ports that bypass the board and connect directly to the massive data center fabric (the Optical Circuit Switches).</p></li></ul><p>The philosophy: The physical server tray is just a holder for the silicon. The software treats the tray as a simple socket.</p><h4>The Virtual Node: The TPU Pod</h4><p>Because the physical node is abstracted, the &#8220;Atomic Unit&#8221; for a Google engineer becomes the TPU Pod&#8212;a massive, flat mesh of thousands of chips.</p><p>Scaling the Matrix</p><p>Instead of building denser racks (Scale-Up), Google scales by adding more TPUs to the flat fabric (Scale-Out):</p><ul><li><p>TPU v4 Pod: 4,096 chips in a 3D Torus mesh</p></li><li><p>TPU v5p Pod: 8,960 chips in a 3D Torus mesh</p></li><li><p>TPU v7 &#8220;Ironwood&#8221;: 9,216 chips</p></li></ul><p>This represents a Pod size 16x larger than NVIDIA&#8217;s current maximum NVLink domain (576 GPUs in a GB200 NVL72 Pod). </p><p>However, with NVIDIA&#8217;s roadmap targeting 572 GPUs in a single Rubin Ultra NVL572 domain, the gap is narrowing&#8212;NVIDIA is converging on Google&#8217;s Pod-scale integration through the inverse path of extreme rack-scale density.</p><h4><strong>The Interconnect: ICI (Inter-Chip Interconnect)</strong></h4><p>This virtual node relies on Google&#8217;s proprietary ICI fabric:</p><ul><li><p>Bandwidth: Each TPU v5p has 600 GB/s of ICI bandwidth (compared to NVIDIA&#8217;s 1.8 TB/s).</p></li><li><p>Topology: These links connect each chip directly to its 6 neighbors in the mesh (north, south, east, west, up, down).</p></li><li><p>The Trade-off: Google accepts lower per-chip bandwidth in exchange for a massive, flat fabric that requires no hierarchy.</p></li></ul><h4><strong>The Optical Circuit Switch (OCS): The Physical Fabric</strong></h4><p>Unlike NVIDIA&#8217;s electrical NVLink switches, Google&#8217;s ICI relies on Optical Circuit Switching (OCS)&#8212;a fundamentally different networking technology:</p><ul><li><p>Physical Mechanism: The OCS uses arrays of tiny, steerable mirrors (MEMS mirrors) to redirect beams of light. Each mirror can physically rotate to point a laser beam from any input fiber to any output fiber, creating a direct optical path between two TPU chips.</p></li><li><p>Circuit-Switched Architecture: Unlike traditional packet-switched networks (like InfiniBand), the OCS does not forward packets. Instead, it establishes dedicated optical circuits&#8212;direct &#8220;light pipes&#8221; between TPU pairs. Once a circuit is established, data flows at full speed with zero switching overhead.</p></li><li><p>Static Topology: The 3D Torus topology is programmed into the OCS at the start of a training job. The mirrors are set to specific angles, creating the mesh pattern, and they remain in that configuration for the duration of the workload. This is why Google&#8217;s approach is called &#8220;deterministic&#8221;&#8212;the network topology is fixed and known in advance.</p></li><li><p>The Reconfiguration Trade-off: Physically rotating MEMS mirrors takes milliseconds&#8212;far too slow for dynamic packet routing. However, because AI training workloads are predictable and long-running (days to weeks), this static topology is ideal. The XLA compiler knows the exact mesh structure and schedules all data movement accordingly.</p></li></ul><p>This is the critical innovation that enables Google&#8217;s &#8220;virtual node&#8221; concept: the physical server tray is irrelevant because every TPU is directly connected to every other TPU through pre-configured optical paths. There is no hierarchy, no switch hops, no routing decisions&#8212;just dedicated optical circuits forming a massive, flat mesh.</p><h4><strong>Why Lower Bandwidth Works</strong></h4><p>At first glance, 600 GB/s (TPU) versus 1.8 TB/s (GPU) seems like a massive disadvantage. However, the OCS architecture delivers advantages that offset the lower per-chip bandwidth:</p><ol><li><p>Uniform Distance: In a 3D Torus, every chip is equidistant from every other chip (in terms of network hops). There are no &#8220;fast&#8221; chips and &#8220;slow&#8221; chips. Every TPU is exactly 6 optical hops from any other TPU, creating predictable latency.</p></li><li><p>Zero Switch Overhead: Because the OCS establishes dedicated optical circuits, there is no packet switching, no routing lookups, no buffer management, and no congestion control. The full 600 GB/s bandwidth is available without protocol overhead.</p></li><li><p>Deterministic Routing: Because the topology is fixed and known at compile time, XLA can schedule data movement with perfect precision, eliminating congestion. The compiler calculates the exact nanosecond when each data transfer will occur, ensuring no two transfers ever collide on the same optical path.</p></li><li><p>Scale Advantage: The ability to keep 9,216 chips in a single domain means that massive workloads never leave the optical fabric&#8212;they don&#8217;t hit the slower Tier-2 network that even NVIDIA&#8217;s largest future clusters will require beyond 572 GPUs.</p></li></ol><p>The result: Google&#8217;s lower per-chip bandwidth delivers higher effective utilization because the OCS eliminates the inefficiencies of packet-switched networks.</p><h3>4. Liquid Cooling the Chip</h3><p>While NVIDIA positioned liquid cooling as a recent breakthrough necessary for the Blackwell generation, Google has been using liquid cooling in production data centers since 2017. However, their approach differs significantly from NVIDIA&#8217;s integrated rack-scale design.</p><p><strong>TPU v2/v3 (2017-2018): The Liquid Cooling Transition</strong></p><p>With TPU v2, Google adopted liquid cooling at the board level&#8212;each TPU server tray includes integrated cold plates, but unlike NVIDIA&#8217;s GB200 NVL72, the cooling infrastructure is not monolithically integrated into the rack. The TPU trays remain modular with quick-disconnect fittings for coolant lines, while cooling distribution units (CDUs) are rack-mounted but separate from compute trays.</p><p><strong>TPU v4/v5p/v7 (2021-Present): Advanced Liquid Cooling</strong></p><p>The current generation continues with liquid cooling at higher power densities. TPU v5p chips consume approximately 275W, while TPU v7 &#8220;Ironwood&#8221; is estimated at 400-600W per chip given its 10x performance leap.</p><p><strong>The Key Difference: Modular vs. Monolithic</strong></p><p>The fundamental distinction reflects their broader architectural strategies:</p><p><strong>Google&#8217;s Approach (Modular):</strong></p><ul><li><p>Each TPU tray is an independent, liquid-cooled unit with quick-disconnect fittings</p></li><li><p>Rack is just a physical shelf; cooling infrastructure is separate from compute hardware</p></li><li><p>Failed trays can be hot-swapped without draining the rack</p></li><li><p>Philosophy: Liquid cooling is a utility provided by data center infrastructure</p></li></ul><p><strong>NVIDIA&#8217;s Approach (Integrated):</strong></p><ul><li><p>GB200 NVL72: Cooling system built directly into the 42U rack</p></li><li><p>Entire rack ships as one pre-integrated unit with coolant distribution, pumps, and manifolds</p></li><li><p>Philosophy: Liquid cooling is part of the product</p></li></ul><h3><strong>5. The Engineering Trade-offs</strong></h3><h4><strong>NVIDIA: Granularity and Modularity (The Rack-Scale Node)</strong></h4><p>The NVL72 / NVL144 / NVL576 architectures prioritize Composable Scale. Even though the &#8220;node&#8221; has grown from a server to a rack (or row), it remains a flexible building block.</p><ul><li><p>Elastic Deployment: A data center can deploy a single GB200 NVL72 (72 GPUs) for inference, or connect eight of them into a Rubin Ultra NVL576 (576 GPUs) for training. You are not forced to deploy a warehouse-sized mesh to get started.</p></li><li><p>Physical Segmentation: The system respects physical boundaries. A 72-GPU rack is a discrete thermal and power domain. If one rack fails or needs maintenance, it can be isolated without destabilizing a 10,000-chip cluster.</p></li><li><p>Multi-Tenancy: This architecture is ideal for cloud providers who need to serve diverse workloads&#8212;allocating one rack to Team A for Llama 3 training, and another to Team B for Claude inference.</p></li></ul><p>Verdict: Ideal for Enterprise AI and Cloud, where flexibility and incremental scaling are critical.</p><h4><strong>Google: Integration and Unity (The TPU Pod)</strong></h4><p>The TPU Pod architecture prioritizes Maximum Integration. It treats the entire warehouse as a single device.</p><ul><li><p>The &#8220;Zero-Boundary&#8221; Fabric: In a 9,216-chip Ironwood Pod, the workload stays entirely within the optical ICI fabric. There is no &#8220;performance cliff&#8221; every 72 or 576 chips; the mesh is continuous.</p></li><li><p>Simplified Programming: To the XLA compiler, the Pod looks like one giant 9,000-core processor. Developers don&#8217;t need to manually partition the model across separate racks; the compiler handles the data flow across the entire mesh.</p></li><li><p>Stability via Rigidity: Because the topology is fixed and uniform, the system is less flexible but more predictable. A broken chip is treated as a &#8220;bad sector&#8221; on a hard drive&#8212;mapped out by software without disrupting the massive job.</p></li></ul><p>Verdict: Ideal for Frontier Model Training (Gemini, Claude), where a single massive workload monopolizes the entire cluster for months.</p><h4><strong>The Convergence</strong></h4><p>Interestingly, both approaches are converging toward similar goals through different paths, now that both companies are doing their version of both scaling up and scaling out:</p><ul><li><p>NVIDIA: The GB200 NVL72 effectively creates a &#8220;mini-Pod&#8221; of 72 GPUs. By 2027, the Rubin Ultra NVL576 will expand this to 576 GPUs&#8212;closing the gap with Google&#8217;s integration. It mimics the TPU Pod&#8217;s unity but retains the rack-based modularity.</p></li><li><p>Google: The TPU v7 &#8220;Ironwood&#8221; represents a 10x leap in per-chip performance. By chasing NVIDIA-class density, Google is reducing the <em>number</em> of chips needed to do the same work, effectively making their massive mesh more potent per square meter.</p></li></ul><p>The Key Difference: From a networking perspective, NVIDIA&#8217;s approach remains Modular (the Rack is the unit), while Google&#8217;s approach remains Monolithic (the Pod is the unit).</p><h3>5. Conclusion: Two Definitions of the Node</h3><p>The &#8220;compute node&#8221; has bifurcated to meet different strategic goals:</p><p><strong>NVIDIA defines the node by physics.</strong> They compress more power into a defined physical volume (Server &#8594; Rack &#8594; Multi-Rack), using copper and liquid cooling to maximize density. The result is the Super-Node:</p><ul><li><p>Today: GB200 NVL72 (72 GPUs, 31 TB unified memory)</p></li><li><p>2026: Vera Rubin NVL144 (144 GPUs)</p></li><li><p>2027: Vera Rubin Ultra NVL572 (572 GPUs, 5,000 PetaFLOPS)</p></li></ul><p><strong>Google defines the node by fabric.</strong> They expand the network to encompass more chips, rendering the physical packaging architecturally irrelevant. The result is the Super-Matrix (TPU Pod)&#8212;a 9,216-chip optical mesh where every chip is equidistant from every other.</p><p>Both philosophies solve the same problem&#8212;training trillion-parameter models&#8212;through inverse strategies:</p><ul><li><p>NVIDIA says: &#8220;Build ultra-dense islands, then connect them.&#8221;</p></li><li><p>Google says: &#8220;Build a massive, seamless fabric that eliminates the islands.&#8221;</p></li></ul><p>By 2027, NVIDIA&#8217;s largest &#8220;island&#8221; (Vera Rubin Ultra NVL572 with 572 GPUs) will approach the scale of a small supercomputer, while Google&#8217;s &#8220;fabric&#8221; will continue to seamlessly connect over 9,000 chips. The two approaches will still differ in granularity and deployment philosophy, but both will deliver multi-exaFLOP capability within a single coherent memory domain.</p><p>Regardless of whether the node is a physical monolith or a virtual matrix, these massive systems share a common vulnerability: they are composed of millions of parts, and parts break.</p><p>In the next article, <strong>&#8220;Article 6: Two Network Topologies,&#8221;</strong> we will examine how NVIDIA&#8217;s hierarchical NVLink fabric and Google&#8217;s flat ICI mesh handle the physics of data movement at gigawatt scale.</p>]]></content:encoded></item><item><title><![CDATA[Two Software Foundations for Scale]]></title><description><![CDATA[The Gigawatt Machine: NVIDIA, Google, and the Engineering of Scale 4/12]]></description><link>https://www.gigawattmachine.com/p/two-software-foundations-for-scale</link><guid isPermaLink="false">https://www.gigawattmachine.com/p/two-software-foundations-for-scale</guid><dc:creator><![CDATA[Tony Wan]]></dc:creator><pubDate>Tue, 09 Dec 2025 13:38:20 GMT</pubDate><enclosure url="https://substackcdn.com/image/fetch/$s_!vqus!,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F59867208-fae5-40b9-a570-54b2271d292a_1280x1280.png" length="0" type="image/jpeg"/><content:encoded><![CDATA[<h3>1. Introduction</h3><p>The B200 and TPU v7 we examined in Article 3 represent the pinnacle of silicon density. But to unlock their trillion-FLOP potential, they require fundamentally different software philosophies: Kernel-Centric (NVIDIA) and Compiler-Centric (Google).</p><p>To understand the difference, we must look at the hierarchy of how software talks to hardware.</p><p>In any computer, there is a gap between the high-level code a human writes (Python) and the low-level electrical signals a chip understands. There are two competing software foundations to bridge this gap.</p><h4>The Kernel-Centric Approach (NVIDIA) </h4><p>In this model, the software controls the hardware step-by-step.</p><ul><li><p>How it works: The high-level software breaks the AI model down into a sequence of individual operations (e.g., &#8220;Multiply Matrix A by Matrix B,&#8221; then &#8220;Save to Memory&#8221;).</p></li><li><p>The &#8220;Kernel&#8221;: For each operation, the system calls a specific, pre-written mini-program called a &#8220;Kernel&#8221; (from libraries like cuDNN). The GPU executes this kernel, reports back, and waits for the next command.</p></li><li><p>The Result: This provides maximum flexibility. Because the software manages every step individually, researchers can easily change the model architecture on the fly without breaking the system.</p></li></ul><h4>The Compiler-Centric Approach (Google) </h4><p>In this model, the software controls the hardware by creating a comprehensive plan upfront.</p><ul><li><p>How it works: Instead of sending commands one by one, the high-level software sends the entire mathematical graph of the AI model to a Compiler (XLA).</p></li><li><p>The Optimization: The compiler analyzes the whole program at once. It looks for efficiencies&#8212;like combining three separate math steps into a single hardware action&#8212;and generates a single, monolithic binary file.</p></li><li><p>The Result: This provides maximum efficiency. By planning the entire route before the data moves, the compiler eliminates the pause between steps, but it makes the system more rigid and harder to change during runtime.</p></li></ul><p>Understanding these two stacks is critical because they dictate how an AI Supercomputer is architected, how models are coded, and how performance is unlocked at the silicon level.</p><h3>2. The Foundation: CUDA vs. XLA</h3><p>The deepest point of divergence lies in how the developer interacts with the silicon.</p><h4>NVIDIA: The CUDA &#8220;Kernel&#8221; Approach</h4><p>The foundation of the NVIDIA stack is CUDA (Compute Unified Device Architecture). It allows developers to write &#8220;kernels&#8221;&#8212;C++ functions that execute directly on the GPU&#8217;s parallel cores. This model is imperative: the developer (or library author) explicitly manages memory allocation, thread synchronization, and data movement.</p><ul><li><p>The Mechanism: Most researchers never write raw CUDA. Instead, they leverage libraries like cuDNN (CUDA Deep Neural Network library). These contain hand-written, assembly-optimized kernels for specific operations (e.g., FlashAttention-3 or FP8 Matrix Multiplication). When a framework like PyTorch executes <code>torch.matmul</code>, it dispatches these pre-compiled binaries to the GPU.</p></li><li><p>The Trade-off: This requires significant engineering effort to hand-tune kernels for each new GPU architecture (e.g., Hopper vs. Blackwell), but it offers maximum flexibility for dynamic workloads.</p></li></ul><h4>Google: The XLA &#8220;Compiler&#8221; Approach</h4><p>Google&#8217;s TPUs (Tensor Processing Units) leverage XLA (Accelerated Linear Algebra). Unlike CUDA, XLA is not a programming language but a domain-specific compiler. The developer does not write kernel code for a TPU.</p><ul><li><p>The Mechanism: Frameworks like JAX emit a computational graph. XLA analyzes the entire graph at once. It performs whole-program optimization, fusing multiple operations (e.g., &#8220;Add + Multiply + Activation&#8221;) into a single hardware instruction to minimize memory access.</p></li><li><p>The Trade-off: This approach is deterministic. The compiler calculates the exact clock cycle for every memory movement <em>before</em> the program runs. This eliminates the runtime overhead of managing threads but makes the system rigid; dynamic shapes or conditional branching can trigger costly re-compilations.</p></li></ul><h4>Verdict: When is each approach superior?</h4><ul><li><p>CUDA Wins: When Algorithm Velocity is the priority. If you are inventing new layer types daily (e.g., sparse mixture-of-experts, state-space models), CUDA allows you to write a custom kernel and run it immediately without waiting for a compiler to understand the new math.</p></li><li><p>XLA Wins: When Production Efficiency is the priority. If your model architecture is stable (e.g., a standard Transformer), XLA can squeeze more performance out of the silicon by fusing operations that a human developer might miss.</p></li></ul><h4><strong>Practical Implication for Trillion-Parameter Models</strong> </h4><p>For frontier model training, the &#8220;Kernel&#8221; approach (NVIDIA) currently dominates because research velocity is paramount. Debugging a compilation error on a 10,000-chip XLA cluster can take days. </p><p>However, once a model architecture is frozen, the XLA approach offers a theoretical path to lower training costs by maximizing Model Flops Utilization (MFU).</p><h3>3. How Close to the Metal is CUDA vs XLA?</h3><p>To understand the difference between these ecosystems, we must visualize the layers of abstraction between the programmer&#8217;s Python code and the silicon transistors.</p><h4><strong>The NVIDIA Stack (The Layered Hierarchy)</strong> </h4><p>CUDA is not the absolute lowest level of access to an NVIDIA GPU, but it is the lowest level accessible to software developers. It relies on a stack of layers to translate code into action.</p><ul><li><p>Bare Metal: The physical hardware (Silicon, Transistors, Memory).</p></li><li><p>Microcode/Firmware: Internal low-level instructions that control voltage and basic operations. (Inaccessible to humans).</p></li><li><p>PTX &amp; SASS: The assembly languages the GPU actually reads. The driver translates CUDA into these machine instructions.</p></li><li><p>CUDA: The programming layer. It sits on top of the driver, allowing developers to write C++ code that explicitly manages the GPU&#8217;s memory and processor threads.</p></li><li><p>Libraries (cuDNN): Collections of pre-written, highly optimized CUDA code for specific math tasks.</p></li><li><p>Frameworks (PyTorch): The user-friendly layer that calls the libraries.</p></li></ul><h4><strong>The Google Stack (The Direct Compilation)</strong> </h4><p>The XLA approach is flatter. It removes the intermediate &#8220;programming layer&#8221; (CUDA) entirely.</p><ul><li><p>Bare Metal: The physical TPU hardware.</p></li><li><p>VLIW Machine Code: The TPU requires massive, complex instructions called &#8220;Very Long Instruction Words&#8221; that control hundreds of chip components simultaneously. This code is too complex for humans to write manually.</p></li><li><p>XLA (The Compiler): Instead of a human writing code to manage the chip, the XLA compiler translates the high-level math from the framework directly into the machine code required by the hardware.</p></li></ul><p><strong>The Comparison:</strong></p><ul><li><p>CUDA is &#8220;Close to the Metal&#8221; via Control: It allows the developer to manually manage the hardware resources. You explicitly define how to split the work across the chip&#8217;s cores and which memory banks to use.</p></li><li><p>XLA is &#8220;Close to the Metal&#8221; via Optimization: It allows the software to generate a monolithic machine-code program that perfectly matches the hardware&#8217;s physical layout and timing, without human intervention.</p></li></ul><h3>4. Multi-Chip Communication: NCCL vs. ICI</h3><p>Training large models requires distributing the workload across thousands of chips. The software that orchestrates this data movement differs fundamentally between the two ecosystems.</p><h4>NVIDIA: NCCL (Hierarchical &amp; Dynamic)</h4><p>NCCL (NVIDIA Collective Communication Library), pronounced &#8220;nickel&#8221;, is a topology-aware library designed to navigate the complex, two-tier hierarchy of GPU clusters. It performs runtime discovery of the network topology to route data.</p><ol><li><p>Tier 1 (Intra-Node/Rack): Inside a GB200 NVL72 rack, NCCL utilizes NVLink 5.0 to move data at 1.8 TB/s (bidirectional).</p></li><li><p>Tier 2 (Inter-Node): Between racks, NCCL switches protocols to use InfiniBand or Ethernet (RoCEv2) via ConnectX-7/8 NICs.</p></li></ol><p>In-Network Reduction (SHARP): For the Blackwell generation, NCCL leverages SHARP. This offloads collective operations (like <code>AllReduce</code>) to the NVSwitch silicon itself. Instead of GPUs exchanging data and performing summation locally, the switch performs the math as the data passes through it, dramatically reducing latency.</p><h4>Google: ICI (Flat &amp; Static)</h4><p>Google TPUs are connected via a proprietary ICI (Inter-Chip Interconnect). Unlike NVIDIA&#8217;s hierarchical tree, ICI connects TPUs in a fixed, flat 3D Torus mesh.</p><ul><li><p>Compiler-Managed Routing: Because the hardware topology is static and known at compile time, XLA manages the communication. There is no dynamic routing protocol or packet headers. The compiler schedules data to move from Chip A to Chip B at specific clock cycles.</p></li><li><p>Performance: While ICI often has lower peak bandwidth than NVLink, the elimination of networking overhead results in extremely high utilization for predictable workloads.</p></li></ul><h4>When is each approach superior?</h4><ul><li><p>In Cloud/Rental Environments, if you are renting GPUs from Azure or CoreWeave, you may not get a physically contiguous block of racks. NCCL can dynamically detect the topology and route around fragmented allocation.</p></li><li><p>If you own the data center (like Google) and can guarantee a perfect 64x64x64 cube of chips, ICI eliminates the massive cost and power overhead of Ethernet/InfiniBand switches.</p></li></ul><h4><strong>Practical Implication for Trillion-Parameter Models</strong></h4><p>The &#8220;Flat Mesh&#8221; (ICI) is technically superior for 3D Parallelism (splitting a model across Data, Tensor, and Pipeline dimensions) because neighbors are always equidistant. </p><p>However, NVIDIA has closed this gap with the NVL72, which effectively creates a &#8220;mini-mesh&#8221; of 72 GPUs that mimics the TPU&#8217;s advantage while retaining the flexibility of NCCL for the broader cluster.</p><h3>5. NVIDIA&#8217;s NCCL Performance from Hopper to Blackwell</h3><p>NCCL&#8217;s performance is directly tied to the hardware it runs on. The advancements from the H100 to the B200 and the new GB200 NVL72 architecture showcase a massive leap in communication speed and efficiency.</p><h4>DGX H100 (Hopper)</h4><p>The H100 generation set the baseline for modern AI clusters:</p><ul><li><p>Intra-Node (within the 8-GPU server): NCCL utilizes 900 GB/s of total bidirectional bandwidth from 4th-generation NVLink.</p></li><li><p>Inter-Node (server-to-server): In multi-node scenarios using NDR InfiniBand networking, NCCL sustains communication speeds up to 400 Gb/s across nodes.</p></li></ul><h4>DGX B200 (Blackwell &#8220;Scale-Out&#8221;)</h4><p>The DGX B200 is the 8-GPU &#8220;scale-out&#8221; server successor. It doubles the performance on both network tiers:</p><ul><li><p>Intra-Node: NCCL leverages the 5th-generation NVLink, doubling the total bidirectional bandwidth to 1.8 TB/s (1800 GB/s) within the server.</p></li><li><p>Inter-Node: The DGX B200 uses 800 Gb/s networking (via InfiniBand Quantum-X800 or Ethernet Spectrum-X800). NCCL can sustain communication speeds up to 800 Gb/s across nodes, again doubling the previous capability.</p></li></ul><h4>GB200 NVL72 (Blackwell &#8220;Scale-Up&#8221;)</h4><p>The GB200 NVL72 rack-scale system represents a fundamental architectural shift, and NCCL has been redesigned to exploit it.</p><ul><li><p>A Single, Massive Fabric: NCCL no longer sees a small 8-GPU node. It sees a massive 72-GPU NVLink domain (which can scale to a 576-GPU pod) where <em>all</em> communication runs on 5th-gen NVLink, providing 1.8 TB/s of bandwidth to every GPU in the entire system.</p></li><li><p>In-Network Reduction (INR): This is the most significant performance innovation. The 5th-generation NVSwitch itself has compute engines. NCCL is able to offload the &#8220;reduce&#8221; (summation) part of the all-reduce operation directly into the network switch. Instead of GPUs waiting to receive data, perform math, and send it on, the switch performs the math as the data is in transit.</p></li></ul><p>This INR feature dramatically accelerates gradient synchronization, freeing up the B200 GPUs&#8217; compute cores to continue working on the next training step, which is a crucial factor in shortening training times for trillion-parameter models.</p><p>These hardware advancements, unlocked by NCCL, directly impact the most critical metric: total training time.</p><p>Note: Although NCCL is open-source, it is primarily developed and optimized by NVIDIA for their GPUs and networking hardware. Optimal functionality, especially for advanced features like 5th-Gen NVLink, In-Network Reduction, or GPUDirect RDMA, relies on proprietary NVIDIA drivers and libraries.</p><h3>6. The High-Level Interface: PyTorch vs. JAX vs. TensorFlow</h3><p>While CUDA and XLA handle the heavy lifting, researchers work in Python. The choice of framework dictates not just the syntax, but how the software interacts with the underlying &#8220;Kernel&#8221; or &#8220;Compiler&#8221; philosophy.</p><h4>PyTorch (The NVIDIA Standard)</h4><p>PyTorch is the dominant framework for AI research and the training of most large language models (LLMs) today (including GPT-4 and Llama 3).</p><ul><li><p>Philosophy: &#8220;Eager Execution.&#8221; PyTorch runs code line-by-line, immediately executing the math on the GPU. This makes it feel like standard Python&#8212;easy to debug and flexible.</p></li><li><p>The Hardware Connection: It is deeply integrated with the Kernel-Centric (NVIDIA) stack. When you write <code>torch.matmul</code>, it immediately calls a specific pre-compiled CUDA kernel.</p></li><li><p>Evolution: To compete with compiler efficiencies, PyTorch 2.0 introduced <code>torch.compile</code>. This captures the model graph and uses TorchInductor to generate optimized kernels, blending its dynamic flexibility with compiler-like speed.</p></li></ul><h4>TensorFlow (The Hybrid)</h4><p>TensorFlow (TF), developed by Google, is the mature incumbent. It sits awkwardly between the two philosophies.</p><ul><li><p>Philosophy: Originally rigid and graph-based (TF 1.x), it pivoted to &#8220;Eager Execution&#8221; (TF 2.x) to match PyTorch&#8217;s ease of use. It now uses Keras as its high-level API, prioritizing user-friendliness.</p></li><li><p>The Hardware Connection: TensorFlow is a hybrid. It works excellently on NVIDIA GPUs (dispatching kernels), but because it is a Google product, it also integrates tightly with XLA. This allows it to run on TPUs, though with more overhead than JAX.</p></li><li><p>The Role: While less popular for <em>training</em> new frontier models today, TensorFlow remains massive in <em>production</em> environments due to its robust serving ecosystem (TFLite, TFServing).</p></li></ul><h4>JAX (The Compiler Native)</h4><p>JAX is the modern successor to TensorFlow&#8217;s original vision, stripped of the bloat. It is the &#8220;purest&#8221; expression of the Compiler-Centric philosophy.</p><ul><li><p>Philosophy: &#8220;Function Transformations.&#8221; JAX is not a neural network library; it is a math library that supports hardware acceleration. It forces a functional programming style (stateless, pure math).</p></li><li><p>The Hardware Connection: JAX is designed specifically to feed the XLA compiler. It uses Just-In-Time (JIT) compilation: it traces your Python function, compiles it into a single binary, and executes it on the TPU (or GPU).</p></li><li><p>The Superpower: Because it is pure math, JAX makes parallelization trivial. APIs like <code>pmap</code> (parallel map) allow researchers to split a model across thousands of chips with a single line of code&#8212;a task that is notoriously difficult in PyTorch and TensorFlow.</p></li></ul><h4>How do they relate?</h4><ul><li><p>PyTorch vs. TensorFlow: These are direct competitors. Both offer a flexible, &#8220;Python-first&#8221; experience. PyTorch won the research war because its debugging is superior; TensorFlow holds the enterprise ground because of its deployment tools.</p></li><li><p>TensorFlow vs. JAX: This is a generational shift. JAX is effectively &#8220;TensorFlow done right&#8221; for high-performance computing. It discards the baggage of Keras and data loaders to focus entirely on generating the fastest possible XLA graph.</p></li></ul><h4>When is each approach superior?</h4><ul><li><p>PyTorch Wins: For Researcher Velocity. If a training run crashes, PyTorch points you to the exact line of Python code that failed. This is invaluable when debugging complex, billion-dollar training runs.</p></li><li><p>JAX Wins: For Scaling Elegance. JAX&#8217;s <code>pmap</code> (parallel map) and <code>shard_map</code> APIs allow you to describe how to split a model across 10,000 chips in just a few lines of code, whereas PyTorch often requires complex add-ons (like Megatron-LM) to handle distributed training state.</p></li></ul><h4><strong>Practical Implication for Trillion-Parameter Models</strong></h4><ul><li><p>Use PyTorch: If you want access to the largest ecosystem of open-source models, tutorials, and engineers. It is the safe, flexible choice for NVIDIA hardware.</p></li><li><p>Use JAX: If you are building a custom supercomputer (using TPUs) or if you need absolute maximum mathematical efficiency at extreme scale (50,000+ chips).</p></li><li><p>Use TensorFlow: Generally avoided for <em>new</em> frontier model training, but critical if you are integrating into a legacy enterprise pipeline that requires robust mobile or edge deployment.</p></li></ul><p>The industry has largely voted for PyTorch (used by OpenAI, Meta, xAI) simply because the talent pool of engineers who know it is vastly larger. </p><p>However, teams that are willing to pay the &#8220;JAX Tax&#8221; (learning a difficult new language) often report superior stability and efficiency at the extreme scale of 50,000+ chips.</p><h3>7. Summary: The Engineering Trade-Off</h3><div class="captioned-image-container"><figure><a class="image-link image2 is-viewable-img" target="_blank" href="https://substackcdn.com/image/fetch/$s_!r_Wk!,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F17e96720-dd33-4ee8-a66d-14bb2e4f1d06_928x872.png" data-component-name="Image2ToDOM"><div class="image2-inset"><picture><source type="image/webp" srcset="https://substackcdn.com/image/fetch/$s_!r_Wk!,w_424,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F17e96720-dd33-4ee8-a66d-14bb2e4f1d06_928x872.png 424w, https://substackcdn.com/image/fetch/$s_!r_Wk!,w_848,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F17e96720-dd33-4ee8-a66d-14bb2e4f1d06_928x872.png 848w, https://substackcdn.com/image/fetch/$s_!r_Wk!,w_1272,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F17e96720-dd33-4ee8-a66d-14bb2e4f1d06_928x872.png 1272w, https://substackcdn.com/image/fetch/$s_!r_Wk!,w_1456,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F17e96720-dd33-4ee8-a66d-14bb2e4f1d06_928x872.png 1456w" sizes="100vw"><img src="https://substackcdn.com/image/fetch/$s_!r_Wk!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F17e96720-dd33-4ee8-a66d-14bb2e4f1d06_928x872.png" width="928" height="872" data-attrs="{&quot;src&quot;:&quot;https://substack-post-media.s3.amazonaws.com/public/images/17e96720-dd33-4ee8-a66d-14bb2e4f1d06_928x872.png&quot;,&quot;srcNoWatermark&quot;:null,&quot;fullscreen&quot;:null,&quot;imageSize&quot;:null,&quot;height&quot;:872,&quot;width&quot;:928,&quot;resizeWidth&quot;:null,&quot;bytes&quot;:245243,&quot;alt&quot;:null,&quot;title&quot;:null,&quot;type&quot;:&quot;image/png&quot;,&quot;href&quot;:null,&quot;belowTheFold&quot;:true,&quot;topImage&quot;:false,&quot;internalRedirect&quot;:&quot;https://bigcompute.substack.com/i/148544210?img=https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F17e96720-dd33-4ee8-a66d-14bb2e4f1d06_928x872.png&quot;,&quot;isProcessing&quot;:false,&quot;align&quot;:null,&quot;offset&quot;:false}" class="sizing-normal" alt="" srcset="https://substackcdn.com/image/fetch/$s_!r_Wk!,w_424,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F17e96720-dd33-4ee8-a66d-14bb2e4f1d06_928x872.png 424w, https://substackcdn.com/image/fetch/$s_!r_Wk!,w_848,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F17e96720-dd33-4ee8-a66d-14bb2e4f1d06_928x872.png 848w, https://substackcdn.com/image/fetch/$s_!r_Wk!,w_1272,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F17e96720-dd33-4ee8-a66d-14bb2e4f1d06_928x872.png 1272w, https://substackcdn.com/image/fetch/$s_!r_Wk!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F17e96720-dd33-4ee8-a66d-14bb2e4f1d06_928x872.png 1456w" sizes="100vw" loading="lazy"></picture><div class="image-link-expand"><div class="pencraft pc-display-flex pc-gap-8 pc-reset"><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container restack-image"><svg role="img" width="20" height="20" viewBox="0 0 20 20" fill="none" stroke-width="1.5" stroke="var(--color-fg-primary)" stroke-linecap="round" stroke-linejoin="round" xmlns="http://www.w3.org/2000/svg"><g><title></title><path d="M2.53001 7.81595C3.49179 4.73911 6.43281 2.5 9.91173 2.5C13.1684 2.5 15.9537 4.46214 17.0852 7.23684L17.6179 8.67647M17.6179 8.67647L18.5002 4.26471M17.6179 8.67647L13.6473 6.91176M17.4995 12.1841C16.5378 15.2609 13.5967 17.5 10.1178 17.5C6.86118 17.5 4.07589 15.5379 2.94432 12.7632L2.41165 11.3235M2.41165 11.3235L1.5293 15.7353M2.41165 11.3235L6.38224 13.0882"></path></g></svg></button><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container view-image"><svg xmlns="http://www.w3.org/2000/svg" width="20" height="20" viewBox="0 0 24 24" fill="none" stroke="currentColor" stroke-width="2" stroke-linecap="round" stroke-linejoin="round" class="lucide lucide-maximize2 lucide-maximize-2"><polyline points="15 3 21 3 21 9"></polyline><polyline points="9 21 3 21 3 15"></polyline><line x1="21" x2="14" y1="3" y2="10"></line><line x1="3" x2="10" y1="21" y2="14"></line></svg></button></div></div></div></a></figure></div><h3>8. Conclusion</h3><p>Ultimately, the choice between these ecosystems represents a fundamental architectural decision rather than a binary right-or-wrong answer. It is a choice of strategic priority.</p><p>The NVIDIA path prioritizes agility. It allows research teams to iterate rapidly, debug easily, and leverage the massive community support of PyTorch. It relies on raw hardware power to mask software inefficiencies, making it the ideal choice for teams exploring the unknown frontiers of model architecture.</p><p>The Google path prioritizes efficiency. It demands a more rigorous, structured approach to coding upfront, but rewards engineers with a system that executes with deterministic precision at scale. It transforms the data center into a predictable &#8220;math factory,&#8221; ideal for stable, massive-scale production workloads.</p><p>As we move forward, the lines are blurring&#8212;PyTorch is adopting compiler techniques, and XLA is becoming more flexible&#8212;but understanding the distinct DNA of these two stacks remains essential for any infrastructure engineer.</p><p>In our next article, <strong>Article 5, "The Compute Node",</strong> we will examine the physical manifestation of the NVIDIA and Google ecosystem: the compute node.</p>]]></content:encoded></item><item><title><![CDATA[Two Strategies for Maximum Density]]></title><description><![CDATA[The Gigawatt Machine: NVIDIA, Google, and the Engineering of Scale 3/12]]></description><link>https://www.gigawattmachine.com/p/two-strategies-for-maximum-density</link><guid isPermaLink="false">https://www.gigawattmachine.com/p/two-strategies-for-maximum-density</guid><dc:creator><![CDATA[Tony Wan]]></dc:creator><pubDate>Tue, 09 Dec 2025 13:34:50 GMT</pubDate><enclosure url="https://substackcdn.com/image/fetch/$s_!vqus!,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F59867208-fae5-40b9-a570-54b2271d292a_1280x1280.png" length="0" type="image/jpeg"/><content:encoded><![CDATA[<h3><strong>1. Introduction</strong></h3><p>In Article 1, we established that the &#8220;Gigawatt Machine&#8221; is defined by two competing architectural philosophies: NVIDIA&#8217;s hierarchical &#8220;Scale-Up&#8221; approach and Google&#8217;s flat &#8220;Scale-Out&#8221; approach.</p><p>These macroscopic architectures are not arbitrary; they are the direct expression of the microscopic silicon engines that power them. To understand why a data center looks the way it does, one must understand the limitations and capabilities of the processor at its core.</p><p>This article deconstructs the two primary engines of the AI era: the NVIDIA Blackwell B200 GPU and the Google TPU v7 &#8220;Ironwood.&#8221;</p><p>While both chips are Application-Specific Integrated Circuits (ASICs) designed for matrix multiplication, they represent divergent engineering paths.</p><h3><strong>2. The Physics of the Problem: Three Levers of Density</strong></h3><p>For decades, Moore&#8217;s Law allowed engineers to double performance simply by shrinking transistors. That era has ended. Modern photolithography has hit a hard physical barrier known as the &#8220;reticle limit&#8221;, the maximum size of a chip that can be etched in a single exposure (roughly 858 mm&#178;).</p><p>Since engineers are nearing the limit of how many transistors they can squeeze on a chip, they have three remaining levers to increase performance:</p><ol><li><p>The Process Lever (Transistor Density): The manufacturing process technology, measured in nanometers (nm). The &#8220;nm&#8221; label in a process node is shorthand for the generation of chipmaking technology. While historically it measured the physical size of a transistor, today it signifies a technology tier. Each step down the &#8220;node ladder&#8221; (e.g., 5nm &#8212;&gt; 4nm &#8212;&gt; 3nm) means increased compute density and higher performance.</p></li><li><p>The Packaging Lever (Silicon Area): Stitching multiple dies together to create a &#8220;Superchip&#8221; that physically exceeds the reticle limit.TSMC&#8217;s CoWoS (Chip-on-Wafer-on-Substrate) places multiple compute dies (GPU/TPU) and memory stacks (HBM) side-by-side on a massive silicon interposer, containing thousands of ultra-dense wires (far denser than standard circuit board). </p></li><li><p>The Precision Lever (Computational Density): Reducing the number of bits required for each calculation (e.g., 8-bit &#8212;&gt; 4-bit). Traditional High-Performance Computing (like weather simulation) requires 64-bit precision. However, neural nets are more noise-resilient; distinguishing if next word is &#8220;the&#8221; or &#8220;cat&#8221; doesn&#8217;t require seven decimal places. The industry is moving to FP8 for training frontier models, while FP4 is being used for inference.</p></li></ol><p>The Strategic Split:</p><p>In this generation, the two giants have pulled different levers.</p><ul><li><p>NVIDIA paused the Process lever to go &#8220;all-in&#8221; on packaging and precision.</p></li><li><p>Google pulled the process lever.</p></li></ul><h3><strong>3. The NVIDIA Strategy</strong></h3><p>NVIDIA&#8217;s Blackwell B200 represents a strategy of &#8220;architectural density.&#8221; Rather than moving to the 3nm process node, NVIDIA optimized the mature 4NP (4-nanometer) process to ensure high yields and reliability for a mass-market product.</p><h4><strong>The Dual-Die Design </strong></h4><p>The B200 is not a single chip.They manufactured two massive dies and stitched them together into a single GPU using advanced CoWoS-L packaging. This &#8220;chiplet&#8221; approach allowed them to pack 208 billion  (vs. 80 Billion in Hopper)transistors into one unit. NVIDIA to effectively doubled the silicon area available for compute without the yield risks of a larger monolithic die.</p><h4><strong>The GB200 Superchip</strong></h4><p>NVIDIA&#8217;s packaging innovation extends beyond just making the GPU bigger. With the GB200 Grace Blackwell Superchip, they have fundamentally altered the definition of a &#8220;processor.&#8221;</p><ul><li><p>The Innovation: Instead of plugging a GPU into a motherboard slot miles away (electronically speaking) from the CPU, NVIDIA packages two B200 GPUs and one Grace CPU onto a single board, fused by NVLink-C2C.</p></li><li><p>The Result: This is not just a &#8220;CPU next to a GPU&#8221;; it is a unified engine where the boundary between the two dissolves. By packaging them this tightly, NVIDIA eliminates the PCIe bottleneck entirely, allowing the GPUs to access the CPU&#8217;s massive LPDDR5X memory pool as their own.</p></li></ul><h4><strong>The Precision Shift: FP4</strong></h4><p>The most significant leap in Blackwell is the sheer density of math operations it can perform.</p><ul><li><p>H100 (Hopper): Delivered 4,000 TFLOPS of FP8 compute.</p></li><li><p>B200 (Blackwell): Doubles this to 8,000 TFLOPS of FP8 compute per GPU.</p></li><li><p>GB200 (Superchip): By fusing two B200s, the superchip delivers a staggering 16,000 TFLOPS of FP8 performance.</p></li></ul><p>Crucially, the B200 architecture introduces native support for FP4 (4-bit Floating Point) precision. By processing data with half the bits of the previous generation, the chip can process twice as much data per clock cycle, doubling throughput again for inference workloads.</p><h4><strong>The Future of Density: Rubin and Rubin Ultra</strong></h4><p>NVIDIA&#8217;s ambition doesn&#8217;t stop at the chip. The goal is to make the &#8220;GPU&#8221; synonymous with the &#8220;Data Center.&#8221;</p><ul><li><p>Today (Blackwell): The maximum size of a single NVLink domain is 72 GPUs (the GB200 NVL72). Beyond this, you must use a slower Tier-2 network.</p></li><li><p>Tomorrow (Rubin): With the upcoming Rubin (R100) architecture in 2026, NVIDIA plans to expand this domain to 144 GPUs.</p></li><li><p>The Horizon (Rubin Ultra): By 2027, the Rubin Ultra platform aims to connect 576 GPUs into a single, coherently addressed NVLink domain. This would effectively turn an entire aisle of racks into one massive, unified logic gate.</p></li></ul><h3><strong>4. The Google Strategy</strong></h3><p>With the TPU v7 &#8220;Ironwood,&#8221; instead of relying solely on fabric scale, Google aggressively chased per-chip density by moving to TSMC&#8217;s cutting-edge 3nm (N3P) process. This allowed them to pack significantly more transistors into a single compute die (~700mm&#178;) than was possible with 4nm technology.</p><ul><li><p>The Economics of Yield: Why use a more expensive, lower-yield process than NVIDIA? Vertical Integration. Because Google does not sell the TPU chip (they sell the <em>service</em>), they can tolerate higher manufacturing costs and lower yields per wafer. They don&#8217;t need to protect hardware margins in the same way a merchant supplier like NVIDIA does.</p></li></ul><p>Like Blackwell, the TPU v7 has moved beyond the monolithic die. It employs a dual-chiplet design, stitching together two massive compute dies and HBM memory stacks using advanced CoWoS packaging.</p><h3><strong>5. Head-to-Head: B200 vs. TPU v7</strong></h3><p>In an apples-to-apples comparison at FP8 precision, the two engines are remarkably close.</p><div class="captioned-image-container"><figure><a class="image-link image2 is-viewable-img" target="_blank" href="https://substackcdn.com/image/fetch/$s_!HEgZ!,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fb392d52c-b784-4846-9d0a-b0d6eeeb37c1_978x500.png" data-component-name="Image2ToDOM"><div class="image2-inset"><picture><source type="image/webp" srcset="https://substackcdn.com/image/fetch/$s_!HEgZ!,w_424,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fb392d52c-b784-4846-9d0a-b0d6eeeb37c1_978x500.png 424w, https://substackcdn.com/image/fetch/$s_!HEgZ!,w_848,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fb392d52c-b784-4846-9d0a-b0d6eeeb37c1_978x500.png 848w, https://substackcdn.com/image/fetch/$s_!HEgZ!,w_1272,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fb392d52c-b784-4846-9d0a-b0d6eeeb37c1_978x500.png 1272w, https://substackcdn.com/image/fetch/$s_!HEgZ!,w_1456,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fb392d52c-b784-4846-9d0a-b0d6eeeb37c1_978x500.png 1456w" sizes="100vw"><img src="https://substackcdn.com/image/fetch/$s_!HEgZ!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fb392d52c-b784-4846-9d0a-b0d6eeeb37c1_978x500.png" width="978" height="500" data-attrs="{&quot;src&quot;:&quot;https://substack-post-media.s3.amazonaws.com/public/images/b392d52c-b784-4846-9d0a-b0d6eeeb37c1_978x500.png&quot;,&quot;srcNoWatermark&quot;:null,&quot;fullscreen&quot;:null,&quot;imageSize&quot;:null,&quot;height&quot;:500,&quot;width&quot;:978,&quot;resizeWidth&quot;:null,&quot;bytes&quot;:133534,&quot;alt&quot;:null,&quot;title&quot;:null,&quot;type&quot;:&quot;image/png&quot;,&quot;href&quot;:null,&quot;belowTheFold&quot;:true,&quot;topImage&quot;:false,&quot;internalRedirect&quot;:&quot;https://www.gigawattmachine.com/i/181026984?img=https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fb392d52c-b784-4846-9d0a-b0d6eeeb37c1_978x500.png&quot;,&quot;isProcessing&quot;:false,&quot;align&quot;:null,&quot;offset&quot;:false}" class="sizing-normal" alt="" srcset="https://substackcdn.com/image/fetch/$s_!HEgZ!,w_424,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fb392d52c-b784-4846-9d0a-b0d6eeeb37c1_978x500.png 424w, https://substackcdn.com/image/fetch/$s_!HEgZ!,w_848,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fb392d52c-b784-4846-9d0a-b0d6eeeb37c1_978x500.png 848w, https://substackcdn.com/image/fetch/$s_!HEgZ!,w_1272,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fb392d52c-b784-4846-9d0a-b0d6eeeb37c1_978x500.png 1272w, https://substackcdn.com/image/fetch/$s_!HEgZ!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fb392d52c-b784-4846-9d0a-b0d6eeeb37c1_978x500.png 1456w" sizes="100vw" loading="lazy"></picture><div class="image-link-expand"><div class="pencraft pc-display-flex pc-gap-8 pc-reset"><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container restack-image"><svg role="img" width="20" height="20" viewBox="0 0 20 20" fill="none" stroke-width="1.5" stroke="var(--color-fg-primary)" stroke-linecap="round" stroke-linejoin="round" xmlns="http://www.w3.org/2000/svg"><g><title></title><path d="M2.53001 7.81595C3.49179 4.73911 6.43281 2.5 9.91173 2.5C13.1684 2.5 15.9537 4.46214 17.0852 7.23684L17.6179 8.67647M17.6179 8.67647L18.5002 4.26471M17.6179 8.67647L13.6473 6.91176M17.4995 12.1841C16.5378 15.2609 13.5967 17.5 10.1178 17.5C6.86118 17.5 4.07589 15.5379 2.94432 12.7632L2.41165 11.3235M2.41165 11.3235L1.5293 15.7353M2.41165 11.3235L6.38224 13.0882"></path></g></svg></button><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container view-image"><svg xmlns="http://www.w3.org/2000/svg" width="20" height="20" viewBox="0 0 24 24" fill="none" stroke="currentColor" stroke-width="2" stroke-linecap="round" stroke-linejoin="round" class="lucide lucide-maximize2 lucide-maximize-2"><polyline points="15 3 21 3 21 9"></polyline><polyline points="9 21 3 21 3 15"></polyline><line x1="21" x2="14" y1="3" y2="10"></line><line x1="3" x2="10" y1="21" y2="14"></line></svg></button></div></div></div></a></figure></div><h3><strong>6. Conclusion</strong></h3><p>The B200 and TPU v7 represent two valid solutions to the end of Moore&#8217;s Law. While their per-chip performance is now effectively neck-and-neck (~4.6 PFLOPS), they differ radically in how they cluster that power.</p><p>At first glance, Google&#8217;s ability to connect 9,216 chips into a single domain seems vastly superior to NVIDIA&#8217;s limit of 72 GPUs (Blackwell) or 576 GPUs (Rubin Ultra). But this is not a case of &#8220;more is better&#8221;; it is a trade-off between Bandwidth Intensity and Fabric Extent.</p><h4>NVIDIA: Scale Up</h4><p>NVIDIA builds and connects ultra-dense islands.</p><ul><li><p>The Domain: A single NVLink domain (72 or 576 GPUs) is a &#8220;tight&#8221; cluster. Inside this island, bandwidth is massive (1.8 TB/s per chip) and memory is fully unified. It is effectively one giant, coherently caching brain.</p></li><li><p>The Scale: To build a 100,000-chip supercomputer, NVIDIA connects ~1,400 of these islands together using InfiniBand.</p></li><li><p>The Advantage: This modularity allows for extreme flexibility. You can build a supercomputer of <em>any</em> size by simply adding more racks.</p></li></ul><h4>Google: Scale Out</h4><p>Google builds massive, seamless meshes.</p><ul><li><p>The Domain: A single TPU Pod (9,216 chips) is a &#8220;wide&#8221; cluster. The bandwidth per chip is lower (~1.2 TB/s), but the fabric extends much further without hitting a switch.</p></li><li><p>The Scale: To build a 100,000-chip supercomputer, Google connects ~11 of these massive pods together using Optical Circuit Switching.</p></li><li><p>The Advantage: This minimizes the &#8220;performance cliff&#8221; for massive models, as nearly the entire training run can stay within the optical fabric.</p></li></ul><p>The Verdict:</p><p>They are neck-and-neck because they arrive at the same destination&#8212;the Gigawatt Machine&#8212;via different granularities.</p><ul><li><p>NVIDIA stacks thousands of ultra-dense blocks.</p></li><li><p>Google weaves a few massive sheets.</p></li></ul><p>Both architectures have successfully continued to scale even though the reticle limit has been reached.</p><p>But raw silicon power is useless without software to orchestrate it. In <strong>Article 4, &#8220;Two Software Foundations for Scale&#8221;</strong>, we examine how NVIDIA&#8217;s CUDA and Google&#8217;s XLA translate human intent into hardware action.</p><p></p><p></p>]]></content:encoded></item><item><title><![CDATA[Two Silicon Foundations for Scale]]></title><description><![CDATA[The Gigawatt Machine: NVIDIA, Google, and the Engineering of Scale 2/12]]></description><link>https://www.gigawattmachine.com/p/two-silicon-foundations-for-scale</link><guid isPermaLink="false">https://www.gigawattmachine.com/p/two-silicon-foundations-for-scale</guid><dc:creator><![CDATA[Tony Wan]]></dc:creator><pubDate>Tue, 09 Dec 2025 13:33:38 GMT</pubDate><enclosure url="https://substack-post-media.s3.amazonaws.com/public/images/d7d63a5d-6e10-4fb9-a096-620d41842ee1_2048x2048.png" length="0" type="image/jpeg"/><content:encoded><![CDATA[<h3>1. Introduction</h3><p>To build a machine capable of training a trillion-parameter models, you need silicon designed for one specific purpose: massive, parallel matrix multiplication, collectively known as AI Accelerators.</p><p>While the industry often uses &#8220;GPU&#8221; as a catch-all term for these chips, the reality is a tale of two distinct lineages. To understand the Gigawatt Machine, we must trace the evolution of the two dominant families that power it:</p><ol><li><p>The GPU (NVIDIA): A flexible, general-purpose parallel processor that evolved from rendering video games to training AI.</p></li><li><p>The TPU (Google): A specialized, domain-specific ASIC (Application-Specific Integrated Circuit) built from a blank sheet of paper to optimize matrix math.</p></li></ol><p>This article traces the architectural history of these two families, defining the vocabulary&#8212;Tensor Cores, HBM, and Systolic Arrays&#8212;that you will need to understand the engineering deep dives in the rest of this series.</p><h3><strong>2. Key Performance Definitions</strong></h3><p>Regardless of the architecture (GPU or TPU), three metrics define the power of an accelerator. We will reference these throughout the series:</p><p><strong>1. FLOPS (Compute)</strong> How much math the chip can do per second.</p><ul><li><p>Example (NVIDIA B200): Delivers ~4,500 TeraFLOPS (FP8).</p></li><li><p>Example (Google TPU v7): Delivers ~4,614 TeraFLOPS (FP8).</p></li><li><p>Note how closely matched these two engines are when running the same 8-bit math.</p></li></ul><p><strong>2. HBM (Memory Capacity)</strong> How big a model fits on the chip.</p><ul><li><p>Example (NVIDIA B200): Packs 192 GB of HBM3e.</p></li><li><p>Example (Google TPU v7): Matches this with 192 GB of HBM3e.</p></li><li><p>Why it matters: If the model doesn&#8217;t fit in HBM, it must be split across chips. Since both chips have the same capacity, they can hold the same size model slices, making the <em>network</em> the deciding factor for performance.</p></li></ul><p><strong>3. Memory Bandwidth (Speed)</strong> How fast data moves from memory to the compute cores.</p><ul><li><p>Example (NVIDIA B200): 8 TB/s.</p></li><li><p>Example (Google TPU v7): ~7.4 TB/s.</p></li><li><p>Why it matters: This is often the true bottleneck in AI training. A fast chip with slow memory spends most of its time idling.</p></li></ul><h3>3. Evolution to <em>Lower</em> Precision</h3><p>In traditional High-Performance Computing (like weather simulation), precision is everything (64-bit). In AI, the rules are inverted. Neural networks are surprisingly resilient to noise; they do not need 7 decimal places to decide if an image is a &#8220;Cat.&#8221; They just need &#8220;directional accuracy.&#8221;</p><p>By using fewer bits to represent a number (Quantization), we gain two massive advantages:</p><ol><li><p>Memory Bandwidth: Sending a 4-bit number moves 8x faster over the wire than a 32-bit number.</p></li><li><p>Compute Density: You can pack 4x as many 4-bit calculators into the same silicon area.</p></li></ol><h3><strong>4. The NVIDIA GPU: The &#8220;Simultaneous&#8221; Machine</strong></h3><p>The NVIDIA GPU architecture is defined by SIMT (Single Instruction, Multiple Threads).</p><p>In a traditional CPU (Sequential Processing), one core executes one instruction on one piece of data at a time. In a GPU, a single instruction controller drives thousands of cores simultaneously.</p><p>The Mechanism: The GPU groups threads into bundles that execute in lockstep. A single instruction drives thousands of active data lanes simultaneously, allowing the vast majority of the silicon budget to be spent on raw math.</p><p>The Legacy: This architecture was originally engineered for graphics to calculate the color of millions of independent pixels on a screen. However, this massive data parallelism proved mathematically indistinguishable from the needs of Deep Learning: performing the same matrix operation on millions of floating-point numbers simultaneously.</p><p>Over the last decade, NVIDIA has evolved this architecture through three defining eras, shifting the focus from &#8220;Graphics&#8221; to &#8220;AI First.&#8221;</p><h4><strong>The Ampere Era (A100)</strong></h4><p>The A100 was the chip that industrialized deep learning. It introduced the third-generation Tensor Core&#8212;a specialized sub-unit inside the GPU designed specifically to accelerate dense matrix math.</p><ul><li><p>It established HBM2e as the standard for memory, delivering 1.5 TB/s of bandwidth.</p></li><li><p>The A100 remains the backbone of many inference fleets today, but its lack of native FP8 support limits its efficiency for modern Large Language Models (LLMs).</p></li></ul><h4><strong>The Hopper Era (H100 &amp; H200)</strong></h4><p>With the H100, NVIDIA realized that AI models were becoming resilient enough to run on lower precision. They introduced the Transformer Engine, which dynamically adjusts calculations to 8-bit (FP8) precision.</p><ul><li><p>Lower precision effectively doubled the throughput for LLMs without increasing the chip size. The H100 delivered 4,000 TFLOPS of FP8 compute (vs. ~600 TFLOPS of FP16 on the A100).</p></li><li><p>The Memory Upgrade: The H100 utilized HBM3 (3 TB/s bandwidth). Its mid-cycle refresh, the H200, upgraded this to 141 GB of HBM3e running at 4.8 TB/s, allowing larger models to fit on a single chip.</p></li></ul><h4><strong>The Blackwell Era (B200)</strong></h4><p>Blackwell represents the current frontier. It is not just a bigger chip; it is a platform designed to be stitched together.</p><ul><li><p>Precision: It introduces native FP4 (4-bit) support, doubling the raw throughput again to 9,000+ TFLOPS (FP4).</p></li><li><p>Bandwidth: It features 8 TB/s of memory bandwidth and 1.8 TB/s of NVLink interconnect speed, essential for the rack-scale architectures we will discuss later.</p></li></ul><div class="captioned-image-container"><figure><a class="image-link image2 is-viewable-img" target="_blank" href="https://substackcdn.com/image/fetch/$s_!S_72!,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F62f00fc9-8f83-4387-99c6-6ed9fb44c741_1288x824.png" data-component-name="Image2ToDOM"><div class="image2-inset"><picture><source type="image/webp" srcset="https://substackcdn.com/image/fetch/$s_!S_72!,w_424,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F62f00fc9-8f83-4387-99c6-6ed9fb44c741_1288x824.png 424w, https://substackcdn.com/image/fetch/$s_!S_72!,w_848,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F62f00fc9-8f83-4387-99c6-6ed9fb44c741_1288x824.png 848w, https://substackcdn.com/image/fetch/$s_!S_72!,w_1272,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F62f00fc9-8f83-4387-99c6-6ed9fb44c741_1288x824.png 1272w, https://substackcdn.com/image/fetch/$s_!S_72!,w_1456,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F62f00fc9-8f83-4387-99c6-6ed9fb44c741_1288x824.png 1456w" sizes="100vw"><img src="https://substackcdn.com/image/fetch/$s_!S_72!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F62f00fc9-8f83-4387-99c6-6ed9fb44c741_1288x824.png" width="1288" height="824" data-attrs="{&quot;src&quot;:&quot;https://substack-post-media.s3.amazonaws.com/public/images/62f00fc9-8f83-4387-99c6-6ed9fb44c741_1288x824.png&quot;,&quot;srcNoWatermark&quot;:null,&quot;fullscreen&quot;:null,&quot;imageSize&quot;:null,&quot;height&quot;:824,&quot;width&quot;:1288,&quot;resizeWidth&quot;:null,&quot;bytes&quot;:297694,&quot;alt&quot;:null,&quot;title&quot;:null,&quot;type&quot;:&quot;image/png&quot;,&quot;href&quot;:null,&quot;belowTheFold&quot;:true,&quot;topImage&quot;:false,&quot;internalRedirect&quot;:&quot;https://bigcompute.substack.com/i/148373366?img=https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F62f00fc9-8f83-4387-99c6-6ed9fb44c741_1288x824.png&quot;,&quot;isProcessing&quot;:false,&quot;align&quot;:null,&quot;offset&quot;:false}" class="sizing-normal" alt="" srcset="https://substackcdn.com/image/fetch/$s_!S_72!,w_424,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F62f00fc9-8f83-4387-99c6-6ed9fb44c741_1288x824.png 424w, https://substackcdn.com/image/fetch/$s_!S_72!,w_848,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F62f00fc9-8f83-4387-99c6-6ed9fb44c741_1288x824.png 848w, https://substackcdn.com/image/fetch/$s_!S_72!,w_1272,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F62f00fc9-8f83-4387-99c6-6ed9fb44c741_1288x824.png 1272w, https://substackcdn.com/image/fetch/$s_!S_72!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F62f00fc9-8f83-4387-99c6-6ed9fb44c741_1288x824.png 1456w" sizes="100vw" loading="lazy"></picture><div class="image-link-expand"><div class="pencraft pc-display-flex pc-gap-8 pc-reset"><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container restack-image"><svg role="img" width="20" height="20" viewBox="0 0 20 20" fill="none" stroke-width="1.5" stroke="var(--color-fg-primary)" stroke-linecap="round" stroke-linejoin="round" xmlns="http://www.w3.org/2000/svg"><g><title></title><path d="M2.53001 7.81595C3.49179 4.73911 6.43281 2.5 9.91173 2.5C13.1684 2.5 15.9537 4.46214 17.0852 7.23684L17.6179 8.67647M17.6179 8.67647L18.5002 4.26471M17.6179 8.67647L13.6473 6.91176M17.4995 12.1841C16.5378 15.2609 13.5967 17.5 10.1178 17.5C6.86118 17.5 4.07589 15.5379 2.94432 12.7632L2.41165 11.3235M2.41165 11.3235L1.5293 15.7353M2.41165 11.3235L6.38224 13.0882"></path></g></svg></button><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container view-image"><svg xmlns="http://www.w3.org/2000/svg" width="20" height="20" viewBox="0 0 24 24" fill="none" stroke="currentColor" stroke-width="2" stroke-linecap="round" stroke-linejoin="round" class="lucide lucide-maximize2 lucide-maximize-2"><polyline points="15 3 21 3 21 9"></polyline><polyline points="9 21 3 21 3 15"></polyline><line x1="21" x2="14" y1="3" y2="10"></line><line x1="3" x2="10" y1="21" y2="14"></line></svg></button></div></div></div></a></figure></div><h3><strong>5. The Google TPU: The &#8220;Assembly Line&#8221; Machine</strong></h3><p>While NVIDIA GPUs dominate the merchant market, Google charted a different course starting in 2015. They realized that standard GPUs, carrying the legacy baggage of graphics rendering, were too inefficient for their scale. </p><p>The result is the Tensor Processing Unit (TPU)&#8212;an ASIC designed from the ground up for one specific workload: matrix multiplication.</p><h4><strong>The Heart of the TPU: The Systolic Array</strong></h4><p>The defining feature of the TPU is the <strong>Systolic Array</strong>.</p><ul><li><p>How it works: In a standard GPU, data is constantly moved from memory to registers for every calculation. In a Systolic Array, data flows through a massive grid of processing units like a &#8220;heartbeat&#8221; (systole). The output of one unit flows directly into the input of the next without writing back to memory.</p></li><li><p>The Advantage: This drastically reduces register access and power consumption, making the TPU inherently more power-efficient per operation than a general-purpose GPU.</p></li></ul><h4><strong>The Lineage of the TPU</strong></h4><p>Google&#8217;s silicon has evolved through three distinct eras, each solving a different bottleneck:</p><ul><li><p>v1 (2015): The Inference Engine. A simple, air-cooled chip designed solely to run search queries and AlphaGo. It could <em>run</em> models, but it could not <em>train</em> them.</p></li><li><p>v2/v3 (2017-2018): The Training Pivot. Google added High Bandwidth Memory (HBM) and floating-point capability, allowing TPUs to train models. TPU v3 introduced liquid cooling to the data center years before it was common in the merchant market.</p></li><li><p>v4 (2021): The Optical Era. This generation introduced Optical Circuit Switches (OCS), allowing 4,096 chips to be connected in a reconfigurable 3D Torus mesh. This architecture defined the modern &#8220;Pod&#8221; structure that challenges NVIDIA&#8217;s clusters today.</p></li><li><p>TPU v5p (2023): Expanded the pod size to 8,960 chips, doubling the per-chip bandwidth to support larger models.</p></li></ul><p>Google realized that &#8220;one size fits all&#8221; was inefficient, so they split their silicon strategy.</p><ul><li><p>TPU v5e / Trillium (v6e): Designed for high-volume inference (Search, YouTube) where performance-per-dollar matters most.</p></li><li><p>TPU v7 &#8220;Ironwood&#8221;: A massive, high-density chip designed to for training and inference of frontier models like Gemini.</p></li></ul><p>Here is the revised <strong>&#8220;Key Performance Definitions&#8221;</strong> section for Article 2. It now consistently uses the <strong>B200</strong> and <strong>TPU v7</strong> for every example to allow for a direct head-to-head comparison.</p><h3>6. Conclusion</h3><p>Both NVIDIA and Google have pushed their respective architectures&#8212;the flexible GPU and the efficient TPU&#8212;to the absolute limit of what a single piece of silicon can do.</p><p>In the next article, <strong>&#8220;Article 3: Two Strategies for Maximum Density,&#8221;</strong> we will look at how engineers are breaking this physical barrier&#8212;using 3nm process nodes, Chiplet packaging, and 4-bit precision to build the monsters that power the Gigawatt Machine.</p>]]></content:encoded></item><item><title><![CDATA[Two Paths to Gigawatt Machines]]></title><description><![CDATA[The Gigawatt Machine: NVIDIA, Google, and the Engineering of Scale 1/12]]></description><link>https://www.gigawattmachine.com/p/two-anatomies-of-scale</link><guid isPermaLink="false">https://www.gigawattmachine.com/p/two-anatomies-of-scale</guid><dc:creator><![CDATA[Tony Wan]]></dc:creator><pubDate>Wed, 03 Dec 2025 03:45:20 GMT</pubDate><enclosure url="https://substackcdn.com/image/fetch/$s_!vqus!,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F59867208-fae5-40b9-a570-54b2271d292a_1280x1280.png" length="0" type="image/jpeg"/><content:encoded><![CDATA[<h3>1. Introduction</h3><p>When AI Labs like OpenAI, Microsoft, Meta, and xAI announce gigawatt-scale data centers such as &#8220;Stargate,&#8221; &#8220;Fairwater,&#8221; &#8220;Prometheus,&#8221; and &#8220;Colossus,&#8221; they are describing a new class of computing: the AI Supercomputer. </p><p>These are not just large data centers; they are city-sized machines, purpose-built with hundreds of thousands of accelerators to train a single, massive AI model.</p><p>But what does an AI Supercomputer look like? There is no single answer. The industry has split into two competing philosophies on how to address the physics of scaling, driving two completely different physical anatomies.</p><ol><li><p><strong>Strategy A (NVIDIA):</strong> <strong>&#8220;Scale-Up the Node.&#8221;</strong> Increase the density of the chip and compress the computer into a dense monolith to make the wires shorter.</p></li><li><p><strong>Strategy B (Google):</strong> <strong>&#8220;Scale-Out the Fabric.&#8221;</strong> Connect thousands of simpler chips into a massive, uniform optical mesh.</p></li></ol><p>This 12-part series is a comprehensive guide to deconstructing these systems. </p><p>We will primarily examine the Google (TPU) and NVIDIA (GPU) ecosystem. This article explores the fundamental engineering trade-offs that define these two anatomies. </p><h3><strong>2. The Physics of the Problem: Density vs. Fabric</strong></h3><p>To train a trillion-parameter model, you must distribute the workload across thousands of chips. However, sending data between chips takes time (latency). To train faster, you must minimize this latency. Engineers have two levers to pull:</p><h4><strong>Lever 1: Extreme Density (The NVIDIA Strategy)</strong></h4><p>If you can pack more compute power into a smaller physical space, the electrons don&#8217;t have to travel as far.</p><ul><li><p>The Tactic: NVIDIA pushed the limits of physics to increase per-chip density. By moving to 4-bit precision (FP4), the Blackwell B200 GPU delivers a massive 9x generational leap in performance.</p></li><li><p>The Consequence: A chip this powerful becomes a &#8220;gravity well&#8221; that demands massive amounts of data instantly. It requires a hierarchical, ultra-fast network to feed it.</p></li></ul><h4><strong>Lever 2: Massive Fabric (The Google Strategy)</strong></h4><p>If you cannot compress the computer, you must build a faster network.</p><ul><li><p>The Tactic: Historically, Google accepted lower per-chip density (using standard precision), choosing instead to spread the workload across a much larger physical footprint.</p></li><li><p>The Consequence: To make this &#8220;power sprawl&#8221; work, they built a massive, flat, optical network (ICI) that connects thousands of chips directly, making the distance between them virtually irrelevant.</p></li></ul><h3><strong>3. Anatomy A: The &#8220;Super-Node&#8221; (NVIDIA)</strong></h3><p>NVIDIA&#8217;s pursuit of extreme density created the Hierarchical Architecture. The defining characteristic is the &#8220;Scale-Up Node&#8221;&#8212;a system where the atomic unit of the data center is no longer a server, but an entire rack.</p><h4><strong>The Physical Node: The GB200 NVL72</strong></h4><p>This is the new building block. It is not a server; it is a 120 kW, liquid-cooled, 72-GPU rack.</p><ul><li><p>The Density: By compressing 72 GPUs into a single cabinet, NVIDIA keeps them close enough to connect via Copper Cables. This passive copper backplane saves ~20kW of power per rack compared to optical transceivers.</p></li><li><p>The Topology: The rack functions as one single, massive accelerator with a 31 TB unified memory pool.</p></li></ul><h4><strong>The Network: Two-Tier Hierarchy</strong></h4><p>Because the node is so dense, the network must be hierarchical.</p><ul><li><p>Tier 1 (Intra-Pod): The NVLink Fabric. A proprietary, copper-based, packet-switched network that connects up to 576 GPUs (8 racks) into a single &#8220;Pod&#8221; (NVLink Domain). Inside this Pod, bandwidth is massive (1.8 TB/s), allowing for efficient Tensor Parallelism (splitting a model across chips).</p></li><li><p>Tier 2 (Inter-Pod): The Scale-Out Fabric. To build a supercomputer (like &#8220;Stargate&#8221;), you connect hundreds of these Pods using a standard InfiniBand or Ethernet network.</p></li></ul><p>The Look: A dense forest of &#8220;monoliths.&#8221; Extremely tall, heavy (3,700 lbs), liquid-cooled cabinets that require reinforced concrete floors and industrial-scale plumbing.</p><h3><strong>4. Anatomy B: The &#8220;Flat-Mesh&#8221; (Google)</strong></h3><p>Google and Amazon&#8217;s pursuit of fabric scale created the Flat-Fabric Architecture. The defining characteristic is the &#8220;Mesh&#8221;&#8212;a massive, continuous grid of accelerators.</p><h4><strong>The Physical Node: Abstracted</strong></h4><p>The physical node (a server tray with 4-8 chips) is small and architecturally irrelevant. The true unit of scale is the Pod.</p><h4><strong>The Network: Single-Tier Mesh</strong></h4><p>Instead of a hierarchy, the system is designed as one massive, uniform web.</p><ul><li><p>Tier 1 (Intra-Pod): The ICI Fabric. Google uses Optical Circuit Switches (OCS) to connect 8,960 TPUs (in a v5p Pod) into a 3D Torus Mesh.</p></li><li><p>The Difference: In this mesh, every chip connects directly to its neighbors using optical fibers. There is no central switch. The &#8220;Pod&#8221; is 15x larger than NVIDIA&#8217;s, meaning massive workloads can run without ever hitting a slower Tier-2 network.</p></li></ul><p>The Look: A sprawling &#8220;field.&#8221; Rows and rows of standard-height racks connected by a visible canopy of yellow optical fibers (the OCS fabric).</p><h3><strong>5. The &#8220;Brains&#8221; of the Anatomy: Software Stacks</strong></h3><p>The software must match the shape of the hardware. The divergence in network topology (Hierarchy vs. Mesh) forces a divergence in software strategy.</p><h4><strong>NVIDIA: Dynamic Runtime Orchestration (Library-Based)</strong></h4><p>NVIDIA&#8217;s network is packet-Switched and dynamic. Data traffic is unpredictable and bursty.</p><ul><li><p>The Software: CUDA / cuDNN.</p></li><li><p>The Strategy: Runtime Flexibility. NVIDIA uses a &#8220;toolbox&#8221; of pre-compiled libraries. When congestion happens, smart switches and software (NCCL) adapt in real-time, routing packets around traffic jams. This &#8220;eager execution&#8221; model offers maximum flexibility for researchers.</p></li></ul><h4><strong>Google: Deterministic Static Scheduling (Compiler-Based)</strong></h4><p>Google&#8217;s network is Circuit-Switched and static. The OCS mirrors must be physically pointed to the right destination.</p><ul><li><p>The Software: XLA (Compiler).</p></li><li><p>The Strategy: Compile-Time Scheduling. The XLA compiler analyzes the <em>entire</em> AI model before it runs. It pre-calculates the exact path of every data packet and orchestrates a perfect, collision-free flow. It doesn&#8217;t react to traffic; it prevents it. This offers maximum efficiency for known, massive workloads.</p></li></ul><h3><strong>6. The Industrial Reality: Gigawatt Infrastructure</strong></h3><p>Regardless of the anatomy chosen, the sheer scale of these systems has forced a transition from &#8220;Data Center&#8221; to &#8220;Industrial Plant.&#8221;</p><ul><li><p>Power: A 100,000-chip cluster consumes Gigawatts of power&#8212;roughly the output of a nuclear reactor powering a city. The primary engineering challenge shifts from IT administration to grid-scale energy logistics.</p></li><li><p>Cooling: Managing 120kW per rack (NVIDIA) or massive mesh density (Google) makes air cooling physically impossible. The facility becomes a massive hydraulic system, circulating millions of gallons of coolant to manage thermal loads.</p></li></ul><h3><strong>7. Conclusion: The Asymmetric Shift</strong></h3><p>The story of the next generation is one of divergence: NVIDIA stays the course, while Google pivots.</p><p>NVIDIA has remained consistent: build the most powerful, dense node possible, and then arrange those nodes in a hierarchy.</p><ul><li><p>The Evolution: They haven&#8217;t changed their philosophy; they&#8217;ve just scaled the physics. They went from an 8-GPU node (DGX) to a 72-GPU node (GB200 NVL72), creating a &#8220;Super-Node&#8221; that is 9x more powerful.<sup>2</sup> They accept the complexity of a two-tier network (NVLink + InfiniBand) as the necessary cost of this extreme density.</p></li></ul><p>Google, however, has altered its silicon strategy.</p><ul><li><p>The Old Way: For years, Google relied on a &#8220;Fabric-First&#8221; approach&#8212;using massive meshes of moderately powerful chips (TPU v4/v5).</p></li><li><p>The Pivot: With the TPU v7 (&#8221;Ironwood&#8221;), Google effectively admitted that fabric scale alone is no longer sufficient. By driving a 10x leap in per-chip performance, they are also chasing density now, attempting to combine both strategies: NVIDIA-class per-chip density deployed on a Google-class flat optical mesh.</p></li></ul><p>As we enter the Gigawatt era, the architectural battle lines are drawn.</p><ul><li><p>NVIDIA bets that the hierarchical super-node (the 120kW Rack) is the ultimate building block.</p></li><li><p>Google bets that a dense flat mesh (9,000+ high-power chips) can eliminate the hierarchy entirely.</p></li></ul><p>In the next article, <strong>&#8220;Article 2: Two Silicon Foundations for Scale,&#8221;</strong> we will zoom into the silicon die itself, exploring the engines that power these competing visions.</p><div class="subscription-widget-wrap-editor" data-attrs="{&quot;url&quot;:&quot;https://www.gigawattmachine.com/subscribe?&quot;,&quot;text&quot;:&quot;Subscribe&quot;,&quot;language&quot;:&quot;en&quot;}" data-component-name="SubscribeWidgetToDOM"><div class="subscription-widget show-subscribe"><div class="preamble"><p class="cta-caption">Thanks for reading The Gigawatt Machine! Subscribe for free to receive new posts and support my work.</p></div><form class="subscription-widget-subscribe"><input type="email" class="email-input" name="email" placeholder="Type your email&#8230;" tabindex="-1"><input type="submit" class="button primary" value="Subscribe"><div class="fake-input-wrapper"><div class="fake-input"></div><div class="fake-button"></div></div></form></div></div><p></p>]]></content:encoded></item><item><title><![CDATA[Welcome to The Gigawatt Machine Series]]></title><description><![CDATA[A 12-Part Guide to NVIDIA, Google, and the Engineering of AI Infrastructure Scale]]></description><link>https://www.gigawattmachine.com/p/welcome-to-the-gigawatt-machine-series</link><guid isPermaLink="false">https://www.gigawattmachine.com/p/welcome-to-the-gigawatt-machine-series</guid><dc:creator><![CDATA[Tony Wan]]></dc:creator><pubDate>Wed, 03 Dec 2025 03:16:51 GMT</pubDate><enclosure url="https://substack-post-media.s3.amazonaws.com/public/images/c47a4ae7-edea-4183-b9b3-8db4ccbb0a1f_2048x2048.png" length="0" type="image/jpeg"/><content:encoded><![CDATA[<h3><strong>The Era of the Gigawatt Machine</strong></h3><p>When companies like OpenAI, Microsoft, Meta, and xAI announce their next-generation infrastructure, they are no longer describing data centers. They are describing AI Supercomputers&#8212;city-scale industrial machines that function as a single, unified brain.</p><p>We are witnessing a profound shift in industrial engineering. The &#8220;Gigawatt-scale&#8221; machine is not just a larger version of a traditional server farm; it is a fundamentally different class of system. To build one, engineers must solve immutable laws of physics&#8212;latency, heat, and yield.</p><p>We are witnessing an industrial arms race to build AI Supercomputers. But there is no single blueprint for how to build one. The industry has split into two competing philosophies, largely driven by two dominant companies in AI hardware: NVIDIA and Google.</p><h3><strong>Why We Study the Rivalry</strong></h3><p>We are using the rivalry between NVIDIA and Google not just to compare specs, but as a lens to better understand the AI Supercomputer at a system level.</p><p>Engineers at both companies are solving for the same things&#8212;scale, latency, heat, and yield&#8212;but they have contrasting business needs and different starting points.</p><ul><li><p>NVIDIA: Solving for Flexibility. They must build a general-purpose platform that works for everyone (OpenAI, Meta, Microsoft).</p></li><li><p>Google: Solving for Efficiency. They build for themselves (Search, Gemini) and select strategic partners. They can prioritize cost-efficiency over flexibility.</p></li></ul><p>This produces two very different architectures.</p><ul><li><p>The NVIDIA Way: A hierarchical architecture designed for flexibility. It scales up by building massive, powerful nodes that act as &#8220;super-chips,&#8221; designed to serve the diverse needs of the global AI market.</p></li><li><p>The Google Way: A flat architecture designed for efficiency. It scales out by building a massive, uniform mesh of specialized chips, optimized to run specific internal workloads with ruthless cost-efficiency.</p></li></ul><p>Understanding these two distinct approaches elucidates the fundamental concepts of the system itself.</p><h3><strong>What is This Series?</strong></h3><p>Our mission is to deconstruct these machines at a systems level. We treat the training of trillion-parameter models not as a software problem, but as a systems engineering challenge that spans from the silicon transistor to the facility&#8217;s cooling towers.</p><p>This is a 12-article master class that builds an AI Supercomputer from first principles. Rather than getting lost in the battle of raw specifications, we focus on the interplay of hardware and software:</p><ul><li><p>The Building Blocks: How individual accelerators are fused into Nodes, how Nodes are networked into Pods, and how Pods are interconnected into Supercomputers.</p></li><li><p>The Software Synergy: How the choice of software&#8212;whether a flexible library or a predictive compiler dictates the physical design of the network fabric itself.</p></li></ul><h3><strong>Who Is This For?</strong></h3><p>In the current landscape, technical information is predominantly found at two extremes, creating a critical knowledge gap:</p><ul><li><p>Too High-Level: Marketing decks that use buzzwords without explaining the mechanics.</p></li><li><p>Too Low-Level: Dense vendor documentation or academic papers lost in minutiae.</p></li></ul><p>This series bridges that gap. It is written for the professionals such as Executives, Marketing and Sales Leaders, Investors, Architects, Program Managers, Supply Chain Managers, and Engineers. People who need to understand the &#8220;Why&#8221; and  &#8220;How&#8221; of system architecture.</p><h3><strong>The Syllabus: A 12-Part Comparative Journey</strong></h3><p>We have consolidated our curriculum into four logical phases. Each phase explores a layer of the stack, highlighting how NVIDIA and Google diverged to solve the same problem.</p><h4><strong>Phase I: The Hardware (Building the Machine)</strong></h4><p><em>Focus: Deconstructing the physical systems from the chip to the rack.</em></p><ul><li><p><strong>Article 1: The Two Anatomies of Scale</strong></p><ul><li><p>NVIDIA&#8217;s Hierarchical &#8220;Super-Node&#8221; vs. Google&#8217;s Flat &#8220;Optical Mesh.&#8221;</p></li></ul></li><li><p><strong>Article 2: The Silicon Engine</strong></p><ul><li><p>NVIDIA Blackwell vs. Google TPU v7.</p></li></ul></li><li><p><strong>Article 3: The Compute Node</strong></p><ul><li><p>The shift from the 8-GPU Server (DGX) to the 72-GPU Rack (NVL72) vs. the Virtual Pod (TPU).</p></li></ul></li><li><p><strong>Article 4: The Laws of Physics</strong></p><ul><li><p>Why both architectures&#8212;despite their differences&#8212;have been forced to adopt liquid cooling at the Gigawatt scale.</p></li></ul></li></ul><h4><strong>Phase II: The Fabric (Connecting the Machine)</strong></h4><p><em>Focus: The network topologies that turn isolated racks into a supercomputer.</em></p><ul><li><p><strong>Article 5: The Tier-1 Fabric (Inside the Pod)</strong></p><ul><li><p>NVIDIA&#8217;s NVLink (Copper/Electrical) vs. Google&#8217;s ICI (Optical/Circuit-Switched).</p></li></ul></li><li><p><strong>Article 6: The Tier-2 Fabric (The Scale-Out Layer)</strong></p><ul><li><p>NVIDIA&#8217;s InfiniBand/Ethernet vs. Google&#8217;s Jupiter Data Center Network.</p></li></ul></li><li><p><strong>Article 7: Traversing the Fabric</strong></p><ul><li><p>Following data as it traverses the hierarchical NVIDIA network vs. the flat Google mesh.</p></li></ul></li></ul><h4><strong>Phase III: The Workload (Animating the Machine)</strong></h4><p><em>Focus: How software bridges the gap between math and silicon.</em></p><ul><li><p><strong>Article 8: The Software Ecosystem</strong></p><ul><li><p>NVIDIA&#8217;s Library-Based stack (CUDA/cuDNN) for flexibility vs. Google&#8217;s Compiler-Based stack (XLA/JAX) for efficiency.</p></li></ul></li><li><p><strong>Article 9: Parallelism</strong></p><ul><li><p>How Tensor, Pipeline, and Data Parallelism map differently to hierarchical vs. flat hardware.</p></li></ul></li><li><p><strong>Article 10: Orchestration &amp; Storage</strong></p><ul><li><p>Managing data loading at the exabyte scale.</p></li></ul></li></ul><h4><strong>Phase IV: The Facility (Housing the Machine)</strong></h4><p><em>Focus: The industrial reality of the Gigawatt era.</em></p><ul><li><p><strong>Article 11: The Gigawatt Facility</strong></p><ul><li><p>Power, piping, and concrete. How the data center building itself must change to support 120kW racks.</p></li></ul></li><li><p><strong>Article 12: Conclusion: The Road to Zettaflops</strong></p><ul><li><p>Where do we go when we hit the limits of copper, optics, and the power grid?</p></li></ul></li></ul><h2><strong>Ready to Build?</strong></h2><p>Each article is designed to be a concise, system-level deep dive. We avoid getting bogged down in minutiae to keep our eyes on the big picture: the systems engineering of intelligence.</p><p>Let&#8217;s begin with <a href="https://gigawattmachine.substack.com/p/two-anatomies-of-scale">Article 1, &#8220;The Two Anatomies of Scale.&#8221;</a></p><div class="subscription-widget-wrap-editor" data-attrs="{&quot;url&quot;:&quot;https://www.gigawattmachine.com/subscribe?&quot;,&quot;text&quot;:&quot;Subscribe&quot;,&quot;language&quot;:&quot;en&quot;}" data-component-name="SubscribeWidgetToDOM"><div class="subscription-widget show-subscribe"><div class="preamble"><p class="cta-caption">Thanks for reading The Gigawatt Machine! Subscribe for free to receive new posts and support my work.</p></div><form class="subscription-widget-subscribe"><input type="email" class="email-input" name="email" placeholder="Type your email&#8230;" tabindex="-1"><input type="submit" class="button primary" value="Subscribe"><div class="fake-input-wrapper"><div class="fake-input"></div><div class="fake-button"></div></div></form></div></div>]]></content:encoded></item></channel></rss>