Meta Develops A Duo of “AI-Focused” Data Centers, Equipped With 24,000 NVIDIA H100 GPUs

Meta's Advancements In Bumping Up AI Compute Power Continues As The Firm Plans For Massive Developments Moving Ahead Using NVIDIA H100 AI GPUs

Mark Zuckerberg's Meta AI venture has taken a new high as the firm proceeds towards rapid development on AGI, which is the next big thing after generative AI in the field. To achieve optimal computing power, Meta has developed two new data center clusters, as reported by Datacenter Dynamics, with the sole aim of AI research and LLM development in consumer-specific applications such as voice and image recognition. The firm has decided to integrate none other than NVIDIA's H100 AI GPUs, with both clusters containing 24,576 units.

Expanding more upon what the clusters offer, both of them come with a 400Gbps interconnect capability, with one having Meta's self-developed fabric solution based on the Arista 7800, while the other cluster features NVIDIA's Quantum2 InfiniBand fabric to ensure a seamless inter-connectivity experience. Moreover, the clusters are built upon Meta's very own open-GPU Grand Teton AI platform, which is built to leverage the capabilities present in modern-day accelerators through having a bumped-up host-to-GPU bandwidth and compute capacity.

Meta's two new clusters are a part of the company's plans to be a leading force in having a larger AI computing capacity than its competitors. The integration of NVIDIA's H100s comes under the plan of employing 350,000 of these AI GPUs by the end of this year, amounting to a total of 600,000 H100s in their bag. This is undoubtedly a "hardcore" commitment by the firm to progress in the era of AI, where the technology has seen a massive adoption in recent times. And, by the looks of it, Meta has plans to grow its AI infrastructure rapidly, evolving it with next-generation components.