Data centres are mushrooming in various parts of Southeast Asia, as the demand for AI workloads and the region’s growing digital economy drive the need for more computing, storage and networking performance.
One big challenge for the region, of course, is the tropical climate that makes it difficult to cool down the more densely packed and hot-running servers in new data centres.
As a result, many data centres have resorted to liquid cooling. Some even immerse an entire server rack in liquid to transfer the heat away more efficiently than by using the air pushed out by fans.
Traditional cooling methods will not be enough to handle the heat generated by more demanding AI workloads in future, says Chee Hoe Ling, vice-president of product management at Vertiv Asia, which makes equipment such as cooling units used in data centres.
However, even if liquid cooling becomes the norm, the answer lies not just in removing heat from one specific part of the data centre, he explains.
“Data centres in this region will require extensive capacity increases across the entire power train, from the grid to chips in each rack,” he adds, in this month’s Q&A.
NOTE: Responses have been edited for brevity and house style.
Q: Briefly, how can we overcome the challenge of cooling down a data centre in tropical countries in Southeast Asia to support the region’s growing digital economy?
A: Cooling methods in tropical climates can’t follow the rulebook of regions with cooler environments. For instance, in warm and humid climates like Singapore, cooling IT equipment accounts for 37 per cent of total energy consumption in data centres.
Different approach to cooling can be taken. The Vertiv Liebert AHU chilled water system is one example. This cooling unit uses water- or air-side “economisations” and high-efficiency fans, working with hot-aisle containment to avoid the mixture of supply and return air.
The system also operates at higher return air temperature and with higher chilled water temperature, increasing the system efficiency.
It offers a higher-capacity yet compact cooling solution as kilowatt-per-rack demand continues to increase and data centre space is limited. It operates efficiently to combat rising energy costs.
Q: Many technologies have been tested over the years, which now include more exotic ones like liquid immersion. Which ones have proven to be both cost effective and sustainable for the long term?
A: There’s a tangible reason behind the growing profile of liquid immersion, and that is the higher thermal transfer properties of water or other fluids. In fact, liquid cooling can be up to 3,000 times more effective than using air, which is especially pertinent in the face of high-density racks.
Liquid immersion solutions have also demonstrated the potential to reduce cooling energy costs by as much as 95 per cent. These efficiencies were long recognised in mainframe and gaming applications, and now we’re seeing this expand into other applications. I think, in the long run, this will emphatically prove to not only be better for data centre operators’ bottomlines, but for the environment as well.
However, unlike air cooling, liquid cooling affects the numerator (total data centre power) and the denominator (IT equipment power) in the power usage effectiveness (PUE) calculation. This makes PUE an ineffective means to compare the efficiency of liquid with air-cooling systems.
Instead, total usage effectiveness (TUE) is more helpful – when this is done we notice liquid cooling reduces total data centre power by 10.2 per cent, while improving TUE by over 15 per cent.
Q: AI workloads are driving up denser data centres to deliver the performance needed. Are data centres in Southeast Asia able to deliver this, considering the challenges inherent with the tropical climate?
Q: Data centres in this region will require extensive capacity increases across the entire power train, from the grid to chips in each rack. Introducing liquid-cooling technologies into the data centre white space and eventually enterprise server rooms, will be a requirement for most deployments.
Traditional cooling methods will not be able to handle the heat generated by graphics processing units (GPUs) running AI calculations. Investments to upgrade the infrastructure needed to power and cool AI hardware are substantial and navigating these new design challenges is critical.
The transition to accelerated computing will not happen overnight. Data centre and server room designers across Southeast Asia must look for ways to make power and cooling infrastructure future-ready, with considerations for the future growth of their workloads.
Getting enough power to each rack requires upgrades from the grid to the rack. In the white space specifically, this likely means high amperage busway and high-density rack power distribution units (PDUs).
To reject the massive amount of heat generated by hardware running AI workloads, three liquid cooling technologies are emerging as primary options – direct-to-chip, rear-door heat exchange, and immersion cooling.
For direct-to-chip liquid cooling, cold plates sit atop the heat-generating components – usually chips such as CPUs and GPUs – to draw off heat.
Rear-door heat exchangers, however, replace the rear door of the IT rack with heat exchanging coils, through which fluid absorbs heat produced in the rack.
Finally, immersion cooling submerges servers and other components in the rack in a thermally conductive dielectric liquid or fluid, eliminating the need for air cooling altogether.
In addition, with the recently launched Sustainable Tropical Data Centre Testbed (STDCT) in Singapore, there is more reason to be buoyant about unlocking newer, more efficient thermal management technologies that are tailored to the specificities of tropical climates.
Q: Greenfield data centres may come with the latest efficient cooling but older ones need work to improve their PUE. How challenging is this work to redesign or retrofit such facilities in the region?
A: Retrofitting’s energy efficiency benefits overwhelmingly come to the fore when operators pivot from PUE to TUE.
On the issue of modernising infrastructure, there are several chokepoints that data centre managers may be wary of. These include compatibility, space constraints, downtime during upgrades, as well as potential disruptions to ongoing operations.
We understand that data centre managers have highly specific requirements about introducing liquids into the rack. When retrofitting or repurposing existing environments, we can help with our optimised designs that minimize disruption to existing workloads by leveraging available cooling infrastructure and heat rejection where possible.
For example, we can integrate direct-to-chip liquid cooling with a rear-door heat exchanger to maintain a room-neutral cooling solution. In this case, the rear-door heat exchanger prevents excess heat from escaping into the room.
For an air-cooled facility looking to add liquid cooling equipment without any modifications to the site itself, we have liquid-to-air design options available. This same strategy can be deployed in a single rack, in a row, or at scale in a large high-performance computing (HPC) deployment. For multi-rack designs, we have also included high amperage busway and high-density rack PDUs to distribute power to each rack.
These options are compatible with a range of different heat rejection options that can be paired with liquid cooling. This establishes a clean and cost-effective transition path to high-density liquid cooling without disrupting other workloads in the data room.