Close Menu
    Trending
    • Scaling Feature Engineering Pipelines with Feast and Ray
    • Mixing generative AI with physics to create personal items that work in the real world | MIT News
    • Breaking the Host Memory Bottleneck: How Peer Direct Transformed Gaudi’s Cloud Performance
    • Aliasing in Audio, Easily Explained: From Wagon Wheels to Waveforms
    • How to Define the Modeling Scope of an Internal Credit Risk Model
    • AI to help researchers see the bigger picture in cell biology | MIT News
    • Enhancing maritime cybersecurity with technology and policy | MIT News
    • Optimizing Token Generation in PyTorch Decoder Models
    ProfitlyAI
    • Home
    • Latest News
    • AI Technology
    • Latest AI Innovations
    • AI Tools & Technologies
    • Artificial Intelligence
    ProfitlyAI
    Home » Breaking the Host Memory Bottleneck: How Peer Direct Transformed Gaudi’s Cloud Performance
    Artificial Intelligence

    Breaking the Host Memory Bottleneck: How Peer Direct Transformed Gaudi’s Cloud Performance

    ProfitlyAIBy ProfitlyAIFebruary 25, 2026No Comments9 Mins Read
    Share Facebook Twitter Pinterest LinkedIn Tumblr Reddit Telegram Email
    Share
    Facebook Twitter LinkedIn Pinterest Email


    launched Gaudi accelerators to Amazon’s EC2 DL1 situations, we confronted a problem that threatened your complete deployment. The efficiency numbers weren’t simply disappointing; they had been disastrous. Fashions that required coaching successfully had been seeing as much as 50% of their efficiency degradation when scaling throughout a number of nodes. The issue? A community topology that routed all bytes of information via host reminiscence, inflicting a bottleneck that undermined the whole lot Gaudi was designed to do.

    I led the engineering effort to deal with this situation, which in the end resulted within the improvement of what we now name Peer Direct. It’s a characteristic that remodeled the best way Gaudi accelerators talk in cloud environments, and its historical past has some helpful classes on distributed AI coaching at scale.

    The Drawback with Host NICs

    Gaudi was designed with the NIC (Community Interface Card) being embedded straight within the silicon. Every chip has ten community interfaces that may deal with 100 Gbps and help RDMA with RoCE v2, permitting gadgets to entry one another’s reminiscence straight without having the CPU or This structure is extremely environment friendly for AI coaching workloads, the place collective operations like AllReduce have to accumulate gradients from dozens or tons of of gadgets per coaching iteration.

    However cloud deployments should not all the time compliant with good architectures. When Amazon examined Gaudi for DL1 situations, they needed to utilise extraordinary host NICs fairly than Gaudi’s built-in networking. The explanations had been pragmatic: price financial savings and the logistics of working round current knowledge centre infrastructure to accommodate a brand new community topology. From their enterprise perspective, leveraging established community infrastructure made good sense.

    From the efficiency standpoint, it was a catastrophe. As a substitute of peer-to-peer RDMA transfers between Gaudi playing cards, all communication went the great distance round. Information needed to be duplicated out of Gaudi’s high-bandwidth reminiscence into host DRAM, processed by the host CPU, despatched out the host NIC over TCP/IP, acquired by the far host, and duplicated again into the far Gaudi’s reminiscence. All of the added hops triggered latency, stole CPU cycles, and added bandwidth restrictions that fully ruined the scalability of distributed coaching.

    The efficiency shortfall was so dangerous that one questioned whether or not deployment would ever be value it in any respect. This wasn’t a matter of some trivial optimisation; it was an existential risk to your complete association with AWS.

    Why Efficiency Issues This A lot

    It’s value understanding why a 50% lack of efficiency is so disastrous within the life of coaching fashions, and particularly giant fashions equivalent to GPT-5. It now takes weeks or months to coach big language fashions even on humongous clusters. In case you are messing round with fashions which have billions or trillions of parameters, each proportion level of efficiency interprets straight into time and {dollars}.

    Take into account the economics. If it takes 30 days to coach a mannequin versus 15, you’re not solely ready longer; you’re paying for double the compute time. At cloud scale, with tons of or 1000’s of accelerators in steady use, this provides as much as hundreds of thousands of {dollars}. Worse, it halves your iteration velocity. In an aggressive AI world the place firms are racing to develop improved fashions, doubling the variety of assessments inside the similar timeframe may be the excellence between being in entrance and being behind.

    Environmental price can be essential. Massive fashions require loads of electrical energy to show. Higher efficiency means much less compute time, which halves power consumption and carbon emissions. As extra strain is mounted on the AI business to chop its carbon footprint, good points in effectivity are now not a luxurious however fairly a necessity.

    The answer we designed, Peer Direct, delivered RDMA-like efficiency when the bodily community structure wasn’t appropriate for regular RDMA. We would have liked direct reminiscence entry between Gaudi gadgets on totally different programs with out traversing host reminiscence, however on host NICs that weren’t designed for this within the first place.

    The enabler was AWS Elastic Material Adapter, a high-performance community interface for HPC and AI workloads on EC2. EFA offers low-latency OS-bypass communications, sometimes sub-10 microsecond latency. EFA offers RDMA-like semantics utilizing libfabric, an in-user-space communication library offering a standard interface throughout a number of networking applied sciences.

    The duty was to mix libfabric with Habana’s Collective Communication Library, HCCL, which handles all distributed coaching workloads. HCCL was constructed on the belief of native RDMA utilizing Gaudi’s on-chip NICs. We would have liked to create a bridge enabling HCCL to leverage libfabric transparently for communications with out compromising its efficiency ensures and communication semantics.

    The answer wanted a number of technical advances. First, we launched a reminiscence registration system that allowed libfabric to straight entry Gaudi’s high-bandwidth reminiscence. We utilised the Linux kernel DMA-BUF framework, which offers a shared mechanism for sharing machine driver buffers. When HCCL must switch knowledge, the Gaudi driver offers a DMA-BUF file descriptor for the reminiscence area, which libfabric can utilise to create RDMA transfers straight from machine reminiscence.

    Second, we included an LRU cache for reminiscence registrations. Reminiscence registration is pricey; it includes kernel calls and setup operations that may trigger important overhead. By caching the mapping of reminiscence addresses to their libfabric handles, we might reuse registrations in hot-access areas, eliminating most registration overhead from precise coaching.

    Picture by writer

    The outcome was a communication pipeline that appeared one thing like this: HCCL calls the OFI wrapper, which calls the cached libfabric deal with to carry out an RDMA switch straight from supply Gaudi reminiscence to vacation spot Gaudi reminiscence, with neither CPU ever being referred to as. The OFI wrapper was launched to maintain the codebase clear and keep away from direct header inclusions — it’s a light-weight library that dynamically hyperlinks to HCCL and permits using libfabric with out requiring direct integration

    After the switch is full, libfabric stories via a completion queue, and HCCL continues computation with the lately acquired knowledge.

    The Improvement Expertise

    Constructing Peer Direct concerned venturing into new territory on tight schedules. Libfabric wasn’t but mainstream within the discipline of AI accelerators but. There wasn’t loads of public documentation out there, and dialogue was meagre. There was extra of an emphasis on diving into libfabric supply code and reverse-engineering primarily based on experimentation.

    The communication with AWS engineers was paramount however time-zone constrained. Working with a group twelve hours forward meant that debug iterations had 24-hour turnarounds. Each situation wanted cautious documentation and correct communication, as real-time collaboration was not potential.

    The stakes had been excessive because the total DL1 deployment was driving on this performance working. Delays would have thwarted a significant product launch. No person on our group had deep background information of libfabric internals, so we had been studying a fancy codebase whereas designing a crucial integration concurrently.

    The Outcomes

    After we really deployed Peer Direct, the velocity enhancements had been all the trouble was value. We noticed a 1.5 to 2x throughput enhance for collective operations on a 32MB message dimension. On bigger messages, the efficiency was much more astounding, with as much as 1.76x higher throughput at a 256MB message dimension. CPU overhead created a bottleneck that fully disappeared.

    Most importantly, these microbenchmark enhancements straight translated into actual mannequin coaching efficiency. Coaching Habana’s DeepSpeed BERT mannequin with 5 billion parameters throughout 128 Gaudi gadgets, we noticed substantial throughput acquire. Fashions utilizing extra aggressive reminiscence optimisation strategies, like ZeRO-2, that are extra collective operation dependent, benefited disproportionately from Peer Direct.

    PeerDirect was one of many foremost enablers for Gaudi efficiency on AWS DL1 situations, permitting high-scale distributed coaching to run effortlessly on the launch day. Past this preliminary influence, the trouble set the groundwork for future high-performance communication options and proved that cloud-native AI accelerators might stay aggressive regardless of the constraints of cloud infrastructure.

    The expertise jogged my memory of an vital lesson in programs engineering: usually an important efficiency enhancements don’t outcome from optimising the quick path, however from sidestepping unjustified detours altogether. Throughout distributed AI coaching, having knowledge journey straight throughout accelerators with no pointless copies and no CPU intervention is what makes a working system versus one which scales.

    Key takeaways? One vital “takeaway” from this mission is that assumptions about community topology ought to be examined on the earliest level within the distributed coaching course of. As lots of the accelerator stacks had been constructed primarily based on an idealised setting, they don’t have in mind the extra hops, translation layers, and/or cost-driven components that exist within the cloud environments. Due to this fact, earlier than specializing in optimising both mannequin degree or kernel degree, engineers ought to carry out easy collective microbenchmarking throughout the specified topology. If scaling effectivity dramatically decreases with rising node counts or message sizes, the seemingly purpose is the information path, not the kernel. By figuring out the host-memory detour early on, engineers can focus their efforts the place they’ll have the best influence.

    One other vital lesson discovered was the necessity to deal with each reminiscence registration and knowledge switch as first-class efficiency issues. Reminiscence registration overhead can enormously exceed the time spent speaking if every knowledge switch requires a brand new registration. The LRU cache for registered recollections was a non-glamorous addition to HCCL; nonetheless, it successfully eradicated a systemic supply of latency and made the RDMA path viable for real-world workloads. When creating distributed programs, engineers ought to profile not solely the out there community bandwidth but additionally the lifecycle prices related to allocating buffers, registering them, and tearing down these registrations. Small adjustments to those management paths may end up in giant will increase in end-to-end throughputs.

    Lastly, the mixing methodology used on this mission offers a sample for integration. As a substitute of rewriting HCCL to make use of libfabric straight, we created a skinny abstraction layer that maintained current semantics whereas changing the underlying transport layer. This offered a number of advantages, together with minimising threat, lowering code churn, and permitting incremental testing. Groups going through the same problem (i.e., adapting accelerator-native communication libraries to cloud-native materials) ought to try to isolate the transport layer, preserve collective semantics, and create small, testable interfaces between the 2. This not solely permits for quicker improvement but additionally permits for less complicated help of future transport backends.

    Disclosure: I work as an AI Runtime Crew Supervisor at Intel. The views shared on this article are my very own.



    Source link

    Share. Facebook Twitter Pinterest LinkedIn Tumblr Email
    Previous ArticleAliasing in Audio, Easily Explained: From Wagon Wheels to Waveforms
    Next Article Mixing generative AI with physics to create personal items that work in the real world | MIT News
    ProfitlyAI
    • Website

    Related Posts

    Artificial Intelligence

    Scaling Feature Engineering Pipelines with Feast and Ray

    February 25, 2026
    Artificial Intelligence

    Mixing generative AI with physics to create personal items that work in the real world | MIT News

    February 25, 2026
    Artificial Intelligence

    Aliasing in Audio, Easily Explained: From Wagon Wheels to Waveforms

    February 25, 2026
    Add A Comment
    Leave A Reply Cancel Reply

    Top Posts

    The ascent of the AI therapist

    December 30, 2025

    These four charts show where AI companies could go next in the US

    July 16, 2025

    AI’s giants want to take over the classroom

    July 15, 2025

    The Machine Learning “Advent Calendar” Day 17: Neural Network Regressor in Excel

    December 17, 2025

    When (Not) to Use Vector DB

    December 16, 2025
    Categories
    • AI Technology
    • AI Tools & Technologies
    • Artificial Intelligence
    • Latest AI Innovations
    • Latest News
    Most Popular

    DHS is using Google and Adobe AI to make videos

    January 29, 2026

    Using Local LLMs to Discover High-Performance Algorithms

    January 19, 2026

    Researchers glimpse the inner workings of protein language models | MIT News

    August 18, 2025
    Our Picks

    Scaling Feature Engineering Pipelines with Feast and Ray

    February 25, 2026

    Mixing generative AI with physics to create personal items that work in the real world | MIT News

    February 25, 2026

    Breaking the Host Memory Bottleneck: How Peer Direct Transformed Gaudi’s Cloud Performance

    February 25, 2026
    Categories
    • AI Technology
    • AI Tools & Technologies
    • Artificial Intelligence
    • Latest AI Innovations
    • Latest News
    • Privacy Policy
    • Disclaimer
    • Terms and Conditions
    • About us
    • Contact us
    Copyright © 2025 ProfitlyAI All Rights Reserved.

    Type above and press Enter to search. Press Esc to cancel.