09/09/2025 | News release | Distributed by Public on 09/09/2025 22:59
AI and Generative AI (Gen AI) require substantial infrastructure, and tasks like fine-tuning, customization, deployment, and querying can strain resources. Scaling up these operations becomes problematic without adequate infrastructure. Additionally, diverse compliance and legal requirements must be met across various industries and countries. Gen AI solutions must ensure access control, proper workload placement, and audit readiness to comply with these standards. To address these challenges, Broadcom introduced VMware Private AI to help customers run models next to their proprietary data. By combining innovations from both companies, Broadcom and NVIDIA aim to unlock the power of AI and unleash productivity with lower total cost of ownership (TCO).
Our technical white paper, "Deploy VMware Private AI on HGX servers with Broadcom Ethernet Networking," details the end-to-end deployment and configuration, focusing on DirectPath I/O (passthrough) GPUs and Thor 2 NICs with Tomahawk 5 Ethernet switch. This guide is essential for infrastructure architects, VCF administrators, and data scientists aiming to achieve optimal performance for their AI models in VCF.
The white paper provides in-depth guidance on:
The solution detailed in the white paper focuses on VMware Private AI certified HGX systems, which typically feature 4x or 8x H100/H200 GPUs with NVSwitch and NVLink interconnects. The target environment is a VCF-based private cloud utilizing Broadcom 400G BCM957608 NICs and clustered NVIDIA H100 GPUs connected through Ethernet.
A crucial aspect of this deployment is the emphasis on DirectPath I/O for GPUs and Thor2 NICs, allowing for dedicated access to hardware resources and maximizing performance. The guide also covers vital aspects such as:
This comprehensive guide serves as an invaluable resource for anyone looking to deploy and optimize AI inference workloads on a robust, virtualized infrastructure using NVIDIA HGX servers and Broadcom Ethernet. By following the best practices outlined, organizations can build scalable and high-performing AI platforms that meet the demands of modern deep learning applications.
For a deeper dive into the technical specifics and deployment procedures, we encourage you to read the full white paper: https://www.vmware.com/docs/paif-hgx-brcm-eth
Ready to get started on your AI and ML journey? Check out these helpful resources: