The most expensive component of a modern AI supercomputer is not the GPU itself, but the silence between them. In the high-stakes environment of large-scale model training, millions of data packets traverse the network every second, yet a single delayed packet can trigger a catastrophic ripple effect. When one GPU stalls while waiting for a piece of data, the entire synchronous training process grinds to a halt, leaving thousands of H100s or B200s idling in a state of costly inactivity. For years, the developer community has treated this network congestion as an inevitable tax on scale, debating whether the solution lay in proprietary fabrics or more rigid Ethernet configurations. OpenAI has now stepped in to rewrite the rules of this communication with the introduction of the Multipath Reliable Connection, or MRC, a protocol designed to ensure that the network is no longer the ceiling for AI performance.
The Architecture of MRC and the Push for Open Standards
OpenAI did not build MRC in a vacuum. The protocol is the result of a two-year collaborative effort involving a powerhouse consortium of hardware and infrastructure giants, including AMD, Broadcom, Intel, Microsoft, and NVIDIA. Rather than keeping this as a proprietary secret, OpenAI has released the specifications through the Open Compute Project (OCP), ensuring that the industry can standardize the way AI clusters communicate. At its core, MRC is an evolution of RoCEv2 (RDMA over Converged Ethernet), the existing standard that allows hardware to access memory directly across a network without involving the CPU. However, RoCEv2 was not originally designed for the sheer scale of 100,000-GPU clusters.
To bridge this gap, MRC integrates technologies from the Ultra Ethernet Consortium (UEC) and utilizes source routing based on SRv6 (Segment Routing over IPv6). By embedding routing information directly into the packet headers, MRC allows the network to handle traffic with far greater flexibility. A critical component of this architecture is the Network-based Congestion Control (NSCC) algorithm contributed by AMD. This addition allows MRC to maintain full compatibility with existing RDMA programming models while introducing the ability to utilize multiple paths simultaneously. The result is a communication layer that treats the network not as a series of fixed pipes, but as a dynamic fabric capable of rerouting traffic in real-time to avoid hotspots.
From Single-Lane Traffic to Intelligent Packet-Spray
To understand why MRC is a paradigm shift, one must look at how traditional AI networks handle data. Historically, data transmission relied on a single network path between point A and point B. If a specific switch or link became congested, every packet following that path suffered, leading to the dreaded tail latency that kills training efficiency. MRC replaces this rigid linear approach with a technique called Intelligent Packet-Spray. Instead of committing to one path, MRC distributes packets across hundreds of available paths simultaneously, effectively spraying the data across the fabric to ensure no single link becomes a bottleneck.
The true technical twist, however, is where the intelligence resides. In traditional networks, the switches are the brains; they calculate the best path and manage the traffic. This creates a vulnerability: when a link fails, the switches must communicate and recalculate routes, a process that can take several seconds. In the world of AI training, a few seconds of downtime can lead to a massive loss in compute efficiency. MRC shifts the routing intelligence from the switch to the NIC (Network Interface Card). By moving the decision-making to the edge of the network, MRC can detect a failure and reroute traffic in microseconds. The switches are relegated to a simpler role, following pre-configured paths without performing complex calculations. This architectural reversal eliminates the risk of adaptive mechanisms between switches conflicting with one another, creating a more stable and predictable environment.
This shift in intelligence fundamentally alters the physical layout of the data center. By utilizing an 800Gb/s network interface and splitting it into multiple smaller links, MRC allows for unprecedented scaling. A standard 64-port switch can now be operated at a 512-port scale. This efficiency allows OpenAI to connect more than 130,000 GPUs using only two layers of switches. In contrast, traditional architectures typically require three or four layers to achieve similar scale. The physical implications are staggering: the number of optical components is reduced by two-thirds, and the total number of switches is reduced to three-fifths of what was previously required.
Currently, MRC is operational on the industry's most advanced hardware. It is supported by high-performance NICs including the NVIDIA ConnectX-8, AMD Pollara, AMD Vulcano, and Broadcom Thor Ultra. On the switching side, it has been deployed within NVIDIA Spectrum-4/5 and Broadcom Tomahawk 5 environments. This is not a theoretical exercise; the protocol is already powering the most ambitious AI projects on the planet, including OpenAI's massive NVIDIA GB200 supercomputer and Microsoft's Fairwater supercomputer.
Network optimization has evolved from a background infrastructure concern into a primary algorithmic challenge that directly dictates the cost and speed of intelligence.




