October 25, 2023

Director of Product Marketing

AI Networking Moves Ahead with DriveNets and Broadcom Jericho3-AI

We have all witnessed the significant growth of the AI market since the end of 2022. Hyperscale companies like Microsoft, Google, Meta, Amazon, and others have led to increased investments in AI/ML infrastructure. In this dynamic landscape, AI networking takes a crucial role as the foundational fabric for AI clusters, where high performance and robust internal connectivity play a pivotal role in ensuring the success of AI processes.

AI Networking Moves Ahead with DriveNets and Broadcom Jericho3-AI

At DriveNets, we remain committed to precisely delivering that through our Ethernet-based solution. Our Network Cloud-AI, introduced in May, is well-equipped to handle the ever-expanding AI workloads demand and deliver high performance at scale. We achieved this through our Distributed Disaggregated Chassis (DDC) architecture and white boxes based on Broadcom’s Jericho 2C+. 

Today, we are strengthening our commitment to high performance at scale as the previously introduced Broadcom’s Jericho 3-AI and the Ramon 3 based white boxes are available for order. 

Different requirements mean a different AI networking solution  

DriveNets’ recently available whiteboxes, based on Broadcom’s latest silicons, raises the standards for AI clusters. AI workloads and especially generative AI workloads have several unique networking requirements that differ from other workloads. These differences are due to the nature of AI workloads, which are often highly compute-intensive and require the rapid exchange of large, long-lived flows, between nodes. 

The new white boxes, with their unique properties, are purpose-built to meet these requirements, enabling AI infrastructure builders to harness top-notch performance and exclusive features tailored for AI workloads 

The main features of network cloud fabric (NCF) for AI: 

  • Based Accton’s Ramon3 System 
  • 51.2 Tbps with High-radix 128 x 800G OSFP fabric ports
  • Open Ethernet ecosystem 
  • Efficient cell-based switching 
  • Low power, low latency 

The main features of network cloud packet forwarder (NCP) for AI: 

  • Based on Accton’s Jericho3 System 
  • 14.4Tbps – 18 x 800G OSFP network interface ports & 20X 800G OSFP fabric interface ports 
  • Support 400G GPU clusters and future 800G GPU clusters by software upgrade 
  • Open Ethernet ecosystem 
  • Efficient cell-based switching 
  • Low power, low latency 


Accton’s new Jericho3 and Ramon3 VoQ Fabric Systems, presented at the DriveNets booth at the OCP summit, 2023 

DriveNets’ Network Cloud-AI software can harness the cutting-edge whiteboxes’s power, enabling hyperscale companies and other AI infrastructure builders to establish a more open and versatile AI infrastructure. 

This innovative solution offers: 

  • Scalability: The Jericho3-AI fabric can scale up to 32K GPUs (800Gbps) per cluster. 
  • Optimal Network Utilization: distributes traffic across all fabric links, ensuring efficient use at all times. 
  • Fully Scheduled Fabric: This guarantees congestion-free operation with end-to-end traffic scheduling, eliminating flow collisions and jitter. 
  • Zero-Impact Failover: Achieves sub-10ns automatic path convergence, enhancing Job Completion Time (JCT). 

Benefits of DriveNets AI Infrastructure 

DriveNets position themselves as leaders in innovative networking solutions, and as such, we are committed to delivering the best networking solutions. DriveNets already provides a DDC networking infrastructure solutions for all network use cases – from the core to edge and we power the world’s largest software-based network. Leveraging our extensive DDC experience, we’ve introduced a new, open, multi-vendor AI infrastructure that effectively addresses the primary challenges and trade-offs associated with AI workloads. 

Proprietary options like Nvidia’s InfiniBand lack interoperability and result in vendor lock-in, while traditional Ethernet architecture struggles to scale high-performance AI workloads. DriveNets Network Cloud-AI presents the optimal solution. It increases JCT performance, enhances resource utilization, enables interoperability via standard Ethernet, and offers vendor choice. 

DriveNets Network Cloud-AI - The best of both worlds 2

DriveNets Network Cloud-AI – The best of both worlds 

With the new hardware, AI infrastructure builders can utilize DriveNets Network Cloud-AI to fully leverage the benefits of the DDC architecture in AI networking and the capabilities of Jericho3-AI. 

DriveNets Network Cloud-AI - The best of both worlds 1

DriveNets Network Cloud-AI

Here are the key advantages of DriveNets Network Cloud-AI for hyperscale companies and other AI infrastructure builders: 

  • Superior Performance and Efficiency:
    DriveNets Network Cloud-AI software can now scale up to 32K GPUs in a single cluster, providing reliable connectivity, low latency, and minimal jitter. 
  • Open and flexible AI networking solution:
    DriveNets’ Network Cloud-AI, based on Distributed and Disaggregated Chassis, is hardware-agnostic, enabling AI infrastructure builders to use diverse GPUs, ODMs and ASICs. 
  • Reduce cost and simplify operations:
    DriveNets Network Cloud cuts AI networking costs, improving operations and reducing cluster expenses by up to 10%. Moreover, DriveNets provides a globally recognized Ethernet-based solution that demands no specialized skills or expensive resources, unlike InfiniBand 
  • Proven AI networking solution 
    High-scale live network deployments and successful PoCs with hyperscalers demonstrate the reliability and performance of DriveNets’ solution. Network Cloud-AI can provide a risk-free path for adopting a high-performance, standardized Ethernet AI networking solution 

A Networking infrastructure solution that eliminates trade-offs 

This recent joint effort with the availability of Accton whiteboxes powered with Broadcom’s Jericho 3-AI-and Ramon 3 based white boxes, represents another significant milestone in DriveNets’ ongoing commitment to providing hyperscalers and other AI infrastructure builders with a networking solution that eliminates trade-offs.  

DriveNets Network Cloud-AI software leverages the cutting-edge capabilities of the new hardware to provide the ultimate AI performance at scale. It boosts JCT at scales up to 32K GPUs in one cluster, enhances resource utilization, facilitates standard Ethernet for seamless interoperability, and provides freedom of choice among vendors. 

Read more about Network Cloud-AI and see how we can maximize your AI infrastructure.  

Download White Paper

Distributed Disaggregated Chassis (DDC) as an Effective Interconnect for Large AI Compute Clusters

Read more