Meta and NVIDIA Deepen AI Infrastructure Collaboration, Deploying Millions of GPUs and CPUs
2026-02-18 14:42
Favorite

Meta and NVIDIA announced a deepening of their multi-year, multi-generation partnership to jointly expand AI infrastructure, planning to deploy millions of Blackwell and Rubin GPUs in data centers globally, while also adopting NVIDIA CPUs and Spectrum-X networking. The agreement covers both on-premises and cloud environments, supporting Meta's long-term development in large-scale AI training and inference. The two companies will co-design CPUs, GPUs, networking, and software to optimize Meta's personalized, recommendation, and generative AI workloads.

Meta will build hyperscale data centers optimized for model training and production inference. Deployments will include systems based on NVIDIA's GB300 and integrate NVIDIA Spectrum-X Ethernet switches into Meta's Facebook Open Switching System platform. Meta has also adopted NVIDIA's confidential computing technology for private processing in WhatsApp, enabling AI features while protecting user data, with plans to expand to more services.

The infrastructure expansion also includes NVIDIA's Arm-based Grace CPUs, marking the first large-scale, pure Grace deployment. Meta reports improved performance per watt for its data center applications through hardware-software joint optimization. The partners are developing the next-generation Vera CPU, targeting large-scale deployment in 2027, with engineering teams conducting deep co-design to accelerate Meta's next-generation AI models.

Jensen Huang, founder and CEO of NVIDIA, stated: "No one deploys AI at Meta's scale—combining cutting-edge research with industrial-scale infrastructure to deliver the world's largest personalized and recommendation systems to billions of users."

Analysis points out that the Meta-NVIDIA collaboration reflects a full-stack standardization strategy. By integrating GB300 systems, Grace CPUs, and Spectrum-X Ethernet, it reduces the complexity of training and inference clusters, improving software optimization and network efficiency. In the context of AI infrastructure expansion, this co-design helps optimize performance per watt, lower the total cost of ownership for data centers, and support a vertically integrated AI factory model from GPUs to CPUs.

This bulletin is compiled and reposted from information of global Internet and strategic partners, aiming to provide communication for readers. If there is any infringement or other issues, please inform us in time. We will make modifications or deletions accordingly. Unauthorized reproduction of this article is strictly prohibited. Email: news@wedoany.com