Blog

Where Does AI Data Center & Server Cooling Go From Here?

ai data center - Liquid Cooling

Server racks are sagging under the weight of an imaginary force.

The exponential growth of artificial intelligence (AI) platforms is putting data center operations through their paces, particularly when it comes to cooling. AI applications, especially those involving machine learning or media generation, require big-brain processing power, leading to increased strain on data center infrastructure.

The world is relying heavily on AI and cloud storage to keep businesses operating smoothly. If anything goes wrong with electronic storage, it’s *gone*. 

Fortunately, help is on the way. New cooling technology, methods, and products are in development to keep up with the worldwide sprawl of AI. 

Now’s a pivotal time to review how AI data center cooling is reaching a breaking point … and the solutions coming our way.

Rise of the Machines = Data Center Cooling Demand

AI may “talk” and act like it’s sentient, but that doesn’t mean it’s self-sufficient.

You, a human, are independently capable of waking up on a Wednesday morning, checking off mental tasks, and moving from room to room. Your warm-blooded nature generally keeps you at the right temperature, and in the dog days of summer, you can turn up the A/C or take a dip in a pool. And you can do all of those things without being a nuisance to the world.

AI, on the other hand? It’s more like an obnoxious, entitled, and needy neighbor (who also happens to be quite handy).

REASON #1: Punishing CPU Demands

In 2022-23, data centers consumed 2% of the world’s total residential, commercial, and industrial energy. Experts believe that by 2030, the number will jump to 10%. Data center electricity demand will at least double from 2022 to 2026 alone, according to the International Energy Agency

AI is almost entirely to blame.

In the history of electronics, there’s no precedent for AI’s energy appetite. As Time Magazine so eloquently put it:

“Generative AI does not simply float upon ephemeral intuition. Rather, it gathers strength via thousands of computers in data centers across the world, which operate constantly on full blast.”

Generative AI is the area of machine learning that allows users to create new images, text and so on. ChatGPT and other large language models (LLMs) use 10x more energy to answer a query than Google Search does.

To be able to “converse” with human users, LLMs must be able to effortlessly crunch vast amounts of data and perform complex calculations in seconds. Because of that, AI platforms use intense processing power in the form of:

  • High-performance CPUs
  • Advanced GPUs
  • Accelerator chips, an emerging tech for speeding up AI workloads

At home, this would be like running a huge oven range, high-performance hair-dryer, and 3,000-watt space heater on the same circuit. 

It’s not just the AI chipsets themselves that are straining data centers. Cooling systems were already one of the largest consumers of power in data centers. The addition of advanced chipsets is compounding the energy that cooling devices need to keep up with AI processing.

REASON #2: Thermal Management

Why do server rooms need more cooling power than ever before?

You’ve probably already made the connection: Energy equals heat.

AI applications often ask data centers to maintain high-density server configurations to handle these extreme computational loads. A high-density server can maximize computing power and storage capacity in a small space, but this creates so much excess heat that traditional data center cooling systems can have trouble keeping up.

Of course, that’s not an option here. Excessive heat can lead to:

  • Hardware malfunctions
  • System instability
  • Reduced chip efficiency
  • Higher chip failure rates
  • Complete system failure

All of which can lead to catastrophic results. And none of which are acceptable in a world that now expects AI to process data and provide instant insight 24/7/365. Anyone involved in budgeting at a data center won’t appreciate having to replace these expensive processors, either.

A data center’s temperature should hover between 64-81°F to work at peak efficiency, according to ASHRAE guidelines. This range feels further out of reach with each passing generation of NVIDIA chip. The tech giant’s new Blackwell GPU runs at 1,200 W. That hair dryer in your bathroom? About the same.

Designing cooling architecture is a bit like a pyramid. You start at the top, identifying the most difficult area of the assembly to cool. That, without a doubt, is the AI chipset.

REASON #3: The Noise Factor

Data centers are key to the growth of U.S. tech infrastructure. Unfortunately, just like the growth of a city, server room growth comes with a bothersome byproduct: pollution.

Noise pollution is a real problem with computer room air conditioning – and not only inside the facility. Just ask citizens in northern Virginia, who have logged formal complaints comparing a nearby facility’s whirring fans to a leaf blower running 24/7.

As power density rises, cooling a data center requires more air handling units with more cooling potency. Inevitably, more air movers plus faster speeds equals a higher decibel level. Even a 2-minute drive away, this humming can be loud enough to upset a neighborhood.

6 Solutions Heating up in AI Data Center Cooling

ChatGPT and friends have forever changed how much cooling is needed for a server room. If data centers are struggling to keep up now, how will they cope with the next generation of AI chips?

Turning up the A/C won’t cut it. Not only does it make for an inefficient server cooling system, it also doesn’t address sustainability.

To soothe these sore spots, more data centers are turning to advanced cooling solutions:

1. Liquid Cooling

The explosion of heat-intensive AI processes is a threat to space and efficiency issues inside server farms. In enterprise-level setups, some technicians are starting to use liquid coolant distribution units (CDUs). They’re not for all facility sizes, but liquid-based techniques are great at efficiently absorbing and dissipating heat in extremely dense networks. 

Liquid cooling takes a highly targeted approach to carrying away thermal energy, sometimes in conjunction with fans. Liquid-based systems remove heat from components by circulating a specialized fluid close to or directly over them. The liquid is either deionized water, glycol-based fluid, or low-boiling-point dielectric fluid

There are three main methods of liquid cooling for data centers:

    • Direct-to-chip/cold plate – Coolant flows through a cold plate attached to the chip, absorbing excess heat
    • Immersion – Submerges entire components or servers in a nonconductive liquid
    • Rear-door – Draws ambient air into the cabinet, where radiator-like doors remove heat and reintroduce the air into the room

Liquid Cooling

Users in most markets prefer a door-mounted heat exchanger (which also includes a fan ~90% of the time), or a separate cabinet that sits next to the AI setup. 

Liquid cooling isn’t a new discussion; however, it’s a much bigger one today thanks to AI and other high-performance computing tasks. Several companies are already commercializing liquid cooling solutions: 

So why isn’t everyone sprinting to RFQ a liquid coolant system? There are a couple of main reasons:

  • Cost: While they may offer more performance and durability in the long run, fluid-based systems are an investment that’s not practical for some IT hubs.
  • Maintenance: These systems need maintenance and close monitoring to prevent leaks (which could cause a host of other issues).

2. Advanced Compact Fans

Computing infrastructure may be getting denser, but that doesn’t mean designers and installers are getting more room to play around with.

Since electronic equipment is becoming smaller and sleeker, so must the devices used to cool them.

Luckily, there are a variety of high-performance fans that are small enough to fit into compact spaces and provide the advanced cooling technology necessary. (Consider a compact fan to be any product that’s 40-200 mm in diameter.)

 

Fan 🌀 Cooling

Liquid 💧 Cooling

Cost

▪️Low initial cost
▪️Low-maintenance

▪️Higher initial cost
▪️High-maintenance

Power

▪️Many types & performance levels available

▪️Can manage intense heat loads from CPUs & GPUs

Size

▪️Low profile suitable for small spaces 

▪️May not fit in smaller systems

Usability

▪️Simple design
▪️Easy to install & maintain

▪️Complex setup with pumps & reservoirs

Reliability

▪️Fewer failure points for leaks
▪️Durable in dusty, humid locations

▪️Potential for pump failure, leaks
▪️May need more maintenance in poor conditions

A wide range of fan concepts – many of them available today – are working to address AI data center demand:

For High Pressure

High static pressure is a major indicator of resistance against airflow in a cooling system. This resistance can come from dense server racks, long ductwork, or clogged filters. 

Fans with steep power curves can efficiently push air through rack spaces with low-to-medium component density.AI-data-center-cooling-AxiForce-Group-NB-fan-1Example: The AxiForce series is the latest innovation in axial fans. This tubeaxial fan features an aerodynamic housing and impeller. Combine that with the latest powerful motor technology and electronics, and AxiForce is a viable solution for targeted, high-intensity cooling.

AI-data-center-cooling-AxiEco-fan-1For Efficiency

In data center airflow management, all cooling products should add to your solution, not to your problems.

New fans are coming out with clever aerodynamics and motor technology, minimizing the need for energy. In racks with low-to-medium congestion, these advanced designs are cooling effectively while remaining sustainable and relatively quiet.

Example: The AxiEco series is ideal for applications with high back pressure. It’s a great candidate for door-mounted heat exchangers, cabinet cooling, and in-row cooling. (More on that last one soon.)

For Noise Reduction

When you need a high-force fan capable of cooling dense server racks, it’s hard to beat the power and airflow of a counterrotating fan.

Just one problem: They’re horrendous to the human ear.

Optimizing airflow with an alternative design or newer technology can go a long way toward lowering data center decibels. The latest diagonal compact fans can accomplish that and more, offering a way to futureproof a heat-management system without resorting to liquid cooling.

AI-data-center-cooling-DiaForce-120-Angle-fanExample: DiaForce was recently developed for specifically this reason. A customer was upgrading its IT fans to meet the new demands of AI, but needed 50% more air performance and 6 decibels less noise than any product available at the time. While counterrotating fans would often be the go-to choice to meet the airflow demands, their noise level wasn’t acceptable.

DiaForce combines the high airflow of axial fans (preferred for easy installation) with the high back pressure of centrifugal fans (to push through racks of dense components). The result outperforms counterrotating fans with significant energy savings, noise reduction, and psychoacoustic tonality improvements.

3. Smart Fans

Smartphones. Smart TVs. Smart refrigerators. Smart fans?

Electric fans haven’t evolved a ton since their 1882 invention, but they’ve picked up the pace lately – sometimes literally – with smart technology.

These advanced commercial fans allow you to plan ahead for maintenance and avoid costly downtime. The benefits tie together nicely:

  • Real-time monitoring – IoT platforms enable easy viewing & control of fan performance
  • Variable speed control – Can adjust force based on temperature readings
  • Energy efficiency – Can reduce power use during lower-demand periods
  • AI learning – Can “learn” the most optimal way to operate within its environment

Today’s data center fans outpace the old misconceptions that fans: A) are dirty, and B) break too easily. Choosing an enclosure fan with smart technology allows your system to adapt in the moment, adding durability to key areas of the data center.

4. Self-Cooling Enclosures

Want to knock out a server rack’s thermal issues in one shot?

Self-cooling enclosures do just what their name says. They combine a cooling system with a sealed unit to regulate its temperature. Rather than targeting specific parts of an assembly, cooling enclosures control the environment for the entire device. 

Common features include:

  • Integrated cooling – Built-in fans, liquid coolers, or heat exchangers that operate independently
  • Smart technology – Often equipped with sensors that adjust cooling parameters based on real-time temperature data
  • Modular design – Allow for scalable and flexible use based on computational load

Self-cooling enclosures, together with high-performance fans, can serve as a complete solution that’s easy to set up and modify within your existing infrastructure. It’s ideal for situations where traditional data center cooling is inadequate or impractical, but that’s not all. Self-cooling enclosures are also a solution for environments where liquid cooling might be too complex or risky, such as in remote or small-scale installations.

5. Targeted & In-Row Cooling

When heat generation in a server center is fairly consistent from rack to rack, it’s generally safe to rely on room-wide cooling. In AI data center design, heavy-duty chips need attention that’s a little more personal.

Targeted cooling is the use of anti-heat measures specifically in areas with the highest thermal loads, such as AI processing clusters. This approach prevents hotspots and ensures that all equipment operates within safe temperature thresholds.

AI-data-center-cooling-In-Row-Cooling-Products

In-row cooling, for example, is a method in which the technician puts fan units or CDUs directly among or adjacent to the server racks. This reduces how far the air or coolant must travel to the servers, minimizing the mixing of hot and cold air. As the server area grows, it’s easy to add more in-row units without overhauling the entire system.

If your data center handles a mix of high- and lower-intensity processing, localized cooling will probably be more efficient than room-based systems.

6. Containment Systems

Containment systems are even more relentless in physically separating hot exhaust air from cold supply air

Their goal is to provide a stable, controlled airflow to high-density IT equipment, while also returning dry and warm air to the cooling unit.

There are two main types of data center containment:

  • Hot aisle containment – This system encloses the backs of server racks where heat exits. Fans guide the aisle’s hot air into the conditioning unit before it can mix with other outside air.

  • Cold aisle containment – This system encloses the front intake areas of server racks. It prevents servers from intaking hot air from other equipment. The technician can put fans in raised floor tiles to improve delivery to the cold aisle.

Both methods help maintain more consistent temperatures and increase cooling capacity. In an AI server rack, containment systems can be a cost-effective way to maximize your existing cooling setup.

Beat the Heat in Your AI Data Center Infrastructure

Are your electronic components adding up? It’s time for your cooling capabilities to level up, too.

Today, heat flux is orders of magnitude beyond the data centers of last generation. As AI applications drive the need for more powerful and densely packed servers, the demand for energy-efficient cooling solutions has gone from “add-on” to “absolute necessity.” 

Efficient server rack cooling fans and techniques not only save energy; they also support uptime in big-data analytics operations. Where traditional cooling methods fail, the new solutions listed above are picking up the slack. Maybe AI isn’t so threatening after all.

For more tips on meeting requirements for server rooms and other projects, grab our free guide to electronic parts sourcing:

 

(This article is in collaboration with ebm-papst, the world’s top maker of high-performance fans for data centers. The company also serves commercial refrigeration, medical technology, industrial, HVAC, auto, and other markets. Energy efficiency, service life, noise minimization, and sustainability are at the heart of ebm-papst’s innovations.)

 

No Comments Yet

Let us know what you think