AI Infrastructure relies on heat innovation for viability

By Richard Clifford, VP of sales & Solutions EMEA Salute.

  • Sunday, 3rd May 2026 Posted 2 hours ago in by Phil Alsop

AI growth and demand is transforming the physical architecture of data centres, increasing the rack density and thermal output of facilities. This presents a pressing challenge: traditional infrastructure was designed to sustain basic air-cooling systems and lower-density compute environments. But now, these same cooling systems are not up to the challenge presented by high-density environments, and the industry is now at a point where heat innovation directly contributes to infrastructure viability.  

AI workloads are straining traditional cooling

Existing cooling architectures are failing to meet demand under AI workloads due to the immense density shift. Traditional racks are designed to withhold 5-15kW workloads, whereas new AI clusters are generating between 50-100kW+ per rack, with GPUs and AI accelerators packing substantial compute into small footprints. Heat in AI workloads are being produced faster than traditional airflows can discard. At higher rack densities, traditional cooling infrastructure comes under strain because air has limited thermal transfer capacity, which not only creates hotspots, but also increases energy usage. 

Data centres are facing increasing sustainability and regulatory pressures as ineffective cooling infrastructure increases the carbon footprint. Operators must invest in infrastructure that can sustain growth while upholding sustainability and compliance. However, traditional cooling systems are struggling to accommodate the increasing strain from AI. So, the industry is undergoing an immense transformation as it strives to create and adopt new cooling systems to accommodate the new compute workload. 

To resolve this density challenge, data centres are shifting from room cooling to component cooling, moving cooling infrastructure closer to the heat source. In the past, integrate room cooling was the standard, which is designed to keep the data centre cool by circulating cold air across every server. More recently, rack cooling was adopted to cool individual racks directly and remove heat more efficiently. However, liquid cooling, particularly direct-to-chip, has presented a reliable means to handle higher density deployments. 

Everything you should know about direct-to-chip cooling 

The capabilities of direct-to-chip cooling are transforming modern data centres: liquid circulates through cold plates attached to CPUs and GPUs, removing heat from the source. Liquids can remove heat much more efficiently than air, making this method optimal for cooling high-density servers. Direct-to-chip supports extreme rack densities and simultaneously presents a more reliable thermal management technique. 

When data centres adopt direct-to-chip cooling, they can operate at a higher compute density, enable power GPU clusters and generate power more efficiently. Since fast AI deployment has become a top priority in digital infrastructure, direct-to-chip cooling can provide centres with a competitive advantage and opportunities for sustainable growth. 

However, while there are clear benefits to this form of cooling, it also introduces a few operational considerations, including infrastructure redesign. Direct-to-chip cooling typically requires new piping and heat exchangers, indicating that many data centres may need retrofitting to implement it into their existing systems. 

From a risk management perspective, integrating liquid cooling near electronics presents potential leak risks, so operators must update their safety procedures to help prepare for and reduce the risk of significant facility damages and operational downtime. These risks and considerations highlight the need for technical data centre cycle expertise from design to operations, which means ongoing workforce training to stay on top of the skills and expertise required by engineers. 

Cooling innovation and the evolution of data centre design

Innovations in cooling do not only impact equipment design, but it also affects infrastructure strategy, site selection and how data centres are built. Direct-to-chip cooling supports higher-density racks, but also significantly increases heat and power requirements. Facilities now depend on high power availability to support the increasing AI workloads and infrastructural stability. Locations that offer large areas of space and can accommodate advanced thermal management infrastructure and large heat rejection capacities are the only viable options for AI-dense workloads. Site selection may be further influenced by local environmental regulations and concerns surrounding energy availability, which contributes to the complexity of AI integration in digital infrastructure.

Cooling decisions are now impacting every phase of infrastructure planning. Decisions no longer merely impact the space, power and performance of facilities, but they now need be considered throughout every phase of the data centre lifecycle. Zoning, environmental and building legislations can impact the construction of centres with high-density liquid facilities. For example, planning restrictions can determine where a data centre is built, its dimensions and what mechanical infrastructure it is allowed to house, which can reduce the possibility of implementing direct-to-chip cooling systems. 

Advanced cooling systems have different requirements when it comes to site feasibility to accommodate the necessary power density and workload. Cooling requirements directly shape rack layouts, airflow paths and other mechanical systems, indefinitely changing facility design, to ensure facilities can safely manage high-density compute output. Integrating advanced cooling systems like direct-to-chip requires specialised pumps and containment measures, which must be taken into consideration throughout construction. 

During the commissioning stage, cooling systems are regularly tested in conjunction with IT deployment to guarantee thermal goals are met. Cooling strategies play a central role in monitoring performance, as efficient cooling promotes hardware reliability, guides maintenance and reduces operational risks. Inevitably, increasing workloads and rack-densities call for enhancements and replacements, so it is essential to make cooling decisions early to ensure facilities can scale over time. Ultimately, lifecycle expertise and support are vital for data centres wanting to seamlessly adapt to rapid technological changes. 

The sustainability and regulatory pressures

Thermal engineering is emerging as a strategic component in global infrastructure planning, and against a backdrop of regulatory and regional pressures, operators must now also consider legislations such as the EU Energy Efficiency Directive, and other water usage and environmental regulations. Advanced cooling systems require vast amounts of energy and water to operate, and jurisdictions often have niche regulations on energy efficiency, emissions and environmental impact. Because of this, data centres must ensure their cooling strategies align with sustainable standards to uphold ESG goals and avoid incurring penalties. 

Grid constraints are a key consideration; ignoring them may prevent facilities from operating at full capacity and puts centres at risk of overloads. Liquid cooling systems commonly use water as a heat transfer medium, which can be hard to implement in regions with water restrictions and limitations, which calls for additional treatment systems or alternative approaches. 

AI workloads are reinventing the infrastructure of data centres, as heat management becomes a central technical limitation on scaling AI compute. Direct-to-chip cooling has provided the necessary conditions to manage the high-density environments driven by AI. As this technology continues to develop, cooling innovations will inevitably shape the design, construction and geographic expansion of data centres. Lifestyle expertise can help centres to thrive, scaling AI infrastructure in a sustainable and reliable manner. However, adopting innovative cooling systems like direct-to-chip cooling and sustainable regulation-compliant practices will help to ensure the long-term viability and operational efficiency of data centres.  The future of digital infrastructure is not merely defined by compute power, but how the industry manages the thermal output that power generates. 

First Australian Data Centre Forecast Report provides new insight and context on data centre...
In data centres, direct-to-chip liquid cooling relies on fast-acting valves to regulate coolant...
By Fabrizio Landini, Global Data Centre Segment Leader, Hitachi Group.
By Michael Poto - Product Manager - Global Chilled Water Systems at Vertiv.
Cooling is one of the most expensive operational outlays in data centre operations: reducing its...
By JP Buzzell, chief data centre architect at Eaton.
Subzero Engineering discusses why Asia-Pacific is fast becoming the proving ground for...