麻豆精品视频Innovation for Next-Generation AI Factories Featured by NVIDIA
Arslan Munir, Ph.D., co-author and associate professor in the 麻豆精品视频Department of Electrical Engineering and Computer Science. (Photo credit: New York University, Abu Dhabi)
听
Research Snapshot: As artificial intelligence data centers expand in size and power demand, traditional air-cooling systems are struggling to keep high-density GPU clusters cool and efficient. Rising temperatures limit performance, increase energy use, and increase operating costs 鈥 creating a major barrier to scaling the next generation of AI infrastructure.
Research from the College of Engineering and Computer Science, highlighted by NVIDIA, shows that direct-to-chip liquid cooling solves these bottlenecks by keeping GPUs significantly cooler while using less power. The result is up to 17% higher performance, lower energy consumption, and millions in potential annual savings 鈥 offering a practical, sustainable path for building more efficient AI factories.
The research, now featured on NVIDIA鈥檚 , demonstrates that direct-to-chip liquid-cooled GPU (Graphics Processing Unit) systems deliver up to 17% higher computational throughput while reducing node-level power consumption by 16% compared to traditional air-cooled systems. These results translate to potential annual facility-scale savings of $2.25 million to $11.8 million for AI data centers operating between 2,000 to 5,000 GPU nodes.
The study titled, 鈥淐omparison of Air-Cooled Versus Liquid-Cooled NVIDIA GPU Systems,鈥 was co-authored by Arslan Munir, Ph.D., associate professor in the 麻豆精品视频Department of Electrical Engineering and Computer Science, alongside collaborators from Johnson Controls and Lawrence Berkeley National Laboratory, and was conducted using NVIDIA HGX鈩 H100 GPU systems.
鈥淥ur research shows that liquid cooling fundamentally unlocks higher sustained performance, superior energy efficiency, and the thermal headroom needed to push AI workloads to their full potential,鈥 said Munir. 鈥淭hese results provide the technical foundation for designing AI factories capable of handling extreme rack densities while maintaining optimal performance, sustainability and cost efficiency.鈥
The findings from this research arrive at a pivotal time. According to recent federal and industry reports, U.S. investment in AI and data-center infrastructure is expected to exceed $400 billion over the next several years, as the government and private sector build out new AI-training campuses and cloud facilities nationwide. These AI factories are the backbone of modern innovation, powering breakthroughs in health care, national security, transportation and climate research. However, they also consume vast amounts of energy.
鈥淏y directly confronting the thermal and power-efficiency bottlenecks that define today鈥檚 large-scale GPU clusters, this research is charting a clear path toward truly sustainable, high-performance computing,鈥 said Stella Batalma, Ph.D., dean of the College of Engineering and Computer Science. 鈥淭his work demonstrates how innovative cooling technologies can lower operational costs, reduce environmental impact, and significantly increase compute density within existing data-center footprints 鈥 advancing both the science and the scalability of AI.鈥
Key findings from the FAU-led study reveal:
- Thermal Advantage: Liquid cooling maintained GPU temperatures between 46 C to 54 C compared to 55 C 71 C for air cooling.
- Performance Gain: Up to 17% higher computational throughput and 1.4% faster training times for large AI workloads.
- Energy Efficiency: Average 1 kilowatt reduction per server node, equating to 15% to 20% lower facility-level energy use.
- Operational Impact: Up to $11.8 million in annual energy-cost savings for large AI data-center deployments.
- Sustainability: Liquid cooling shifts thermal load from inefficient node-level fans to centralized systems, enabling more accurate Power Usage Effectiveness metrics and reduced carbon footprint.
The work underscores FAU鈥檚 growing leadership in intelligent systems, data-center innovation, and quantum- and AI-driven computing. Munir鈥檚 in the College of Engineering and Computer Science is pioneering research that intersects AI hardware acceleration, hybrid quantum-classical computing, and sustainable HPC system design.
鈥淓nergy-efficient AI infrastructure isn鈥檛 just an engineering optimization 鈥 it鈥檚 a national imperative,鈥 said Munir. 鈥淏y rethinking how we cool and power AI factories, we can dramatically increase performance while aligning with global sustainability and cost-reduction goals.鈥
Collaborators on the white paper include Imran Latif, vice president of global technology and innovation for data centers at Johnson Controls; Alex Newkirk, energy technology researcher at Lawrence Berkeley National Laboratory; 麻豆精品视频doctoral researcher Hayat Ullah; and Kansas State University doctoral researcher Ali Shafique, both supervised by Munir.
The white paper, Comparison of Air-Cooled Versus Liquid-Cooled NVIDIA GPU Systems, can be accessed .
-FAU-
Tags: AI | technology | engineering | research | faculty and staff