Intel

Analysis, market context, and editorial on the business of data center cooling.

Eaton Says AI Data Centers Need Aerospace-Grade Engineering. At 600 kW Per Rack, the Argument Is Hard to Dismiss.

Peter de Bock, VP of Data Center Energy and Cooling at Eaton, argues that NVIDIA Vera Rubin racks at 600 kW demand aerospace-grade reliability, not telecom engineering. PUE masks real efficiency by 0.3 points or more. His proposed replacement metric: tokens per watt. Hot-water cooling at 45°C in and 60°C out exploits a cubic relationship in heat rejection that drops fan power by a factor of eight.

Read more →

Google's Data Center Steam Is Grounding Flights in The Dalles, Oregon. The Cooling Industry Has an Externalities Problem.

Cooling towers at Google's $2.4 billion campus in The Dalles produce steam plumes that settle into the Columbia River corridor, forming fog thick enough to divert flights from the nearby airport. Google consumed 550 million gallons of water in a single year for evaporative cooling, roughly 40% of the city's total. Locals call the phenomenon the "Google Cloud." The cooling industry needs to reckon with the externalities of open-loop evaporative systems.

Read more →

EPA Launches WRAP 2.0 to Push Recycled Water Into Data Center Cooling. The Federal Government Just Picked a Side in the Water Fight.

EPA Administrator Lee Zeldin launched the Water Reuse Action Plan 2.0, naming data center cooling as an explicit federal priority for recycled water. Microsoft's $31 million reuse facility in Quincy, WA reclaims 138 million gallons per year. Loudoun County and Google's Georgia campus already run on reclaimed water. WRAP 2.0 is voluntary, but 200 partners are signed on. Within five years, recycled water becomes the default cooling supply for new evaporative builds.

Read more →

Belden and OptiCool Partnered on a 120 kW Rear-Door Heat Exchanger System. Brownfield Operators Should Be Watching.

Belden and OptiCool deliver an integrated rack-level system with two-phase rear-door heat exchangers rated at 120 kW per rack, 85% cooling energy reduction, and a claimed PUE of 1.02. The refrigerant-based system leaves servers untouched and ships through Belden's existing channel. For brownfield operators who cannot rip out raised floors, the rear-door heat exchanger just became harder to ignore.

Read more →

Chinese Researchers Built a Diamond-Copper Composite That Nearly Doubles Cooling Efficiency. The TIM Supply Chain Just Got a New Front.

A team at the Chinese Academy of Sciences deployed a diamond-copper composite thermal interface material at an AI computing node in Zhengzhou, achieving 80% cooling efficiency improvement and thermal conductivity exceeding 1,000 W/mK. The material also delivered a 10% chip performance gain. China's domestic TIM production is a strategic play to reduce dependence on Western thermal materials suppliers.

Read more →

Schneider Electric Maps the Full Liquid Cooling Stack From 50kW to 1MW Per Rack. The Transition Is Already Underway.

Sujit Kamble of Schneider Electric laid out the full cooling progression: air tops out at 50kW, direct-to-chip handles 80% of heat at source, and two-phase immersion supports up to 1MW per rack. CDU specs include 25-micron filtration, 2.5MW capacity, and 20-year lifespan. With individual servers exceeding $1 million in value, cooling reliability becomes an asset protection problem.

Read more →

Data Centers Burned 683 TWh in 2024 and Cooling Ate Half of It. Most Operators Still Cannot Measure Where the Energy Goes.

The global data center fleet consumed 683 TWh of electricity in 2024. Cooling accounts for up to 50% of that total. The IEA projects 1,479 TWh by 2030. Every degree Celsius raised at server inlet saves roughly 4% in cooling energy. Most facilities still overcool because they lack the instrumentation to know what raising the temperature two degrees will do to row 14.

Read more →

Castrol's Partnership Strategy Is How the Immersion Fluid Market Actually Gets Built

Castrol is assembling the testing infrastructure immersion cooling has been missing. The Submer collaboration puts SmartPod and MicroPod tanks inside Castrol's own labs. Iceotope covers the precision liquid side. RISE and the Open Compute Project are where standards get written. The fluid is the easy part. The ecosystem that proves the fluid works is the hard part.

Read more →

Data Center Power Hits 1,050 TWh This Year. Cooling Is the Constraint That Decides What Gets Built.

Global data center power consumption crosses 1,050 TWh by end of 2026, roughly the annual electricity use of Japan. Direct-to-chip, immersion, and two-phase cooling cut related power draw 50-60% versus air. The cooling decision has become a capacity-planning decision. Operators with the right thermal architecture fit more compute inside smaller utility allocations.

Read more →

DataCool Launched Three Air-Handler Lines at 100,000 CFM and 300 Tons. The Question Is Where Air Still Fits.

DataCool's new Alpine, Glacier, and Kodiak product lines target AI-adjacent air cooling workloads. 300 tons equals roughly 1.05 MW thermal, enough for 6-8 AI racks or 20-25 enterprise racks. Matt Polizzi frames it around AI demand. The harder read is where air still fits in a hybrid build where direct-to-chip takes the GPU load and AHUs handle the storage and network tiers.

Read more →

India Jumped From 10 kW Racks to 150 kW GPU Clusters. The Partner Ecosystem Has Two Years to Catch Up.

Indian data centers are jumping from 5-10 kW per rack to 150 kW GPU clusters in the same buildings. Yotta's Nitin Jadhav, Colt's Arif Khan, and Rackbank's Narendra Sen all confirm cooling architecture is now the primary deal qualifier. Traditional colocation is becoming commoditized. The margin is moving to partners who can lead thermal planning and GPU workload optimization.

Read more →

Iceotope Crossed 200 Patents. That Is the Moat Precision Liquid Cooling Has Been Building for Twenty Years.

Iceotope's patent portfolio hit 200 filings, 109 granted. Chassis-level precision liquid cooling cuts energy use 40% and water use 96% versus traditional methods. Neil Edmunds, Chief Innovation Officer, called it years of engineering work on practical cooling problems. Competing vendors now have to engineer around the IP wall or have a licensing conversation. That is how categories get locked.

Read more →

Orbital Is Launching an AI Data Center Into Low Earth Orbit in April 2027. The Cooling Problem Changes Completely.

Orbital, backed by a16z Speedrun, will launch NVIDIA-powered AI servers on a SpaceX Falcon 9 in April 2027. Continuous solar power. Radiative cooling into deep space. CEO Euwyn Poon says AI progress is being constrained by the grid. If inference workloads move off-planet, terrestrial capacity is freed for training. The bet is long. The constraint is real.

Read more →

Schneider Published the Liquid Cooling Deployment Playbook. Most Operators Still Don't Have One.

The liquid cooling market is growing at 30%+ annually. Schneider Electric, through its Motivair acquisition, just introduced the MCDU-70, a 2.5 MW coolant distribution unit scalable beyond 10 MW. The real substance is operational: constant differential pressure control, redundant pumps, UPS on every CDU. Most facilities teams are commissioning liquid loops for the first time. Schneider is trying to write the rulebook before someone learns these lessons the expensive way.

Read more →

Digital Realty Is Spending €2 Billion to Turn Italy Into a Mediterranean Data Hub

Two new campuses. Rome gets 62 MW, opening early 2027. Milan gets 84 MW, first phase 2028. The Mediterranean corridor now accounts for 56% of Digital Realty's planned EMEA capex through 2030. Data traffic from Asia, the Middle East, and Africa flows into Europe through this region. Digital Realty is also in discussions on an EU-backed AI gigafactory requiring around 200 MW.

Read more →

EdgeConneX Rewrote Every Operational Procedure It Had. Then Built a Training Center Around the New Ones.

EdgeConneX launched a Global Training Center to prepare operations teams for AI-era infrastructure. Every existing procedure rewritten. New ones covering areas never previously managed, including direct-to-chip liquid cooling. Their Ingenuity platform supports 600+ kW per rack. The workforce gap in liquid cooling is real. EdgeConneX decided to solve it internally.

Read more →

Modine's Data Center Cooling Revenue Jumped 78%. The Stock Followed.

Modine shares rose 7.5% after fiscal Q3 showed net sales up 31% to $805 million. The driver: data center sales within Climate Solutions surged 78% year-over-year. Management raised guidance. Analysts are unanimously bullish. One thing to watch: all 11 insider trades over the past six months have been sales. The business is performing. The people running it are taking chips off the table.

Read more →

Air-Cooled Data Centers Hit 80 Decibels. Communities Are Starting to Push Back.

Air-cooled facilities produce 60-80 decibels, comparable to a gas lawnmower running continuously. Water-cooled facilities sit at 40-60 dB. In water-scarce regions where air cooling is the only option, noise is joining water usage, energy costs, and light pollution on the list of community grievances. Cooling method selection is becoming a community relations decision as much as a thermal engineering one.

Read more →

Utilities Weren't Built for What AI Data Centers Are Asking of Them. The Reckoning Is Arriving.

US AI data center power demand: 4 GW in 2024, projected 123 GW by 2035 (Deloitte). Grid interconnection takes 4–5 years minimum; 10 years in congested markets. Hyperscalers are now co-investing in grid infrastructure to jump the queue. Operators who are not hyperscalers are competing for what's left. Liquid cooling's 25–30 MW efficiency gain per 100 MW facility is now a grid capacity argument, not just a density argument.

Read more →

California's Governor Vetoed a Data Center Water Disclosure Law. The Data Gaps Remain.

California has at least 270 data centers and no requirement to report water use. The governor vetoed the disclosure bill while signing one that only studies electricity. Some facilities under construction are allocated up to 8 million gallons per day. On hot summer days, a large data center can exceed 1 million gallons of withdrawal. The numbers are real. They're just not being counted.

Read more →

AI Data Centers Are Expanding Into the World's Most Water-Stressed Region. The Cooling Math Doesn't Work.

MENA holds 6.3% of the global population and 1.4% of global renewable freshwater. 11 of the 17 most water-stressed countries on earth are in the region. Kuwait runs 90% of drinking water through desalination. Oman runs 86%. Data center investment is accelerating in anyway. Closed-loop liquid cooling with dry heat rejection is the only architecture that doesn't compound the problem.

Read more →

Seventy Percent of Global Data Center Capacity Is Already Built. Liquid Cooling Has to Work Around It.

More than 70% of global data center capacity sits in existing buildings designed for 5–15 kW per rack. Brownfield retrofit is not the edge case — it is the mainstream condition for liquid cooling adoption. Rack-by-rack deployment is the approach that works. The real bottleneck is not the technology. It is the workforce gap between what the hardware requires and what facilities teams currently know how to commission.

Read more →

PIMCO Is Weighing a $14 Billion Oracle Data Center Debt Deal. The SPV Structure Is the Story.

PIMCO is in discussions to underwrite up to $14 billion in debt financing for Oracle data center infrastructure through a special purpose vehicle structure. Oracle is targeting $45–50 billion in 2026 infrastructure financing, bypassing its own balance sheet through off-balance-sheet project finance. The deal structure defines what data center project finance looks like at hyperscale.

Read more →

Data Centers Are Warming Their Neighborhoods by 2°C on Average. One Site Hit 9.1°C.

A new study quantifies the local heat island effect of data center waste heat rejection: approximately 2°C average warming within nearby areas of large facilities, with one worst-case measurement of 9.1°C above baseline. No before/after measurements from actual operational sites exist yet. The thermal externality is real, growing, and increasingly a regulatory flashpoint.

Read more →

Liquid Cooling Is No Longer Optional at 100+ kW Rack Densities. The Math Is Simple.

Air cooling tops out at 20–25 kW per rack. The GB200 NVL72 draws 120–130 kW per rack. The delta is not bridgeable with fans. Retrofit costs run 40–60% of original construction. The liquid cooling market hit $3 billion in 2025 and is projected to reach $7 billion by 2029. The decision has already been made for anyone buying current-generation hardware.

Read more →

GPU Platforms Iterate Every 18 Months. Traditional Data Center Construction Takes 36. Modular Closes the Gap.

The mismatch between GPU iteration cycles (12–18 months) and traditional data center design-build timelines (24–36 months) is the core argument for modular data center deployment. Factory-built modules ship in 12–16 weeks, with factory testing replacing on-site commissioning. Operators building traditional facilities for Blackwell are already behind the next architecture cycle.

Read more →

Penn State Built an AI That Cuts Data Center Cooling Costs 25% by Reading the Weather.

A Penn State research team developed physics-informed reinforcement learning software that can cut data center cooling costs by up to 25% without hardware changes. The system analyzes real-time climate conditions and electricity pricing, dynamically adjusting cooling intensity. Results: up to 25% reduction in cooling electricity costs. Presentation at IEEE ITherm Conference, May 2026.

Read more →

Neocloud Deals Are Stalling Because Colocation Providers Won't Sign Without Investment-Grade Credit.

A well-funded neocloud operator recently failed to close a colocation deal despite 15-year terms, 6 months prepaid, millions in liquid cooling coverage, and pricing at $155–160/kW — above market. The reason: no investment-grade credit. ABI Research pegs the neocloud market at $250 billion by 2030. Cooling vendors who built 2026 pipelines around neocloud deal assumptions should be updating their models.

Read more →

AI Workloads Create Millisecond Power Spikes That Standard UPS Systems Can't Handle. Nickel-Zinc Has an Answer.

AI GPU clusters create millisecond-scale transient power spikes that traditional lead-acid and lithium-ion UPS batteries were never engineered to absorb. Nickel-zinc chemistry offers a water-based electrolyte with no thermal runaway risk, 25–50% lower lifecycle greenhouse gas emissions, and over 90% recyclability — with no active thermal management requirement.

Read more →

IBM and Arm Are Merging Enterprise AI Architectures. The Target Is Regulated Industries That Can't Move Data Off-Premise.

IBM and Arm are enabling Arm-native applications to run inside IBM Z and LinuxOne mainframe systems through virtualization compatibility — without hardware modifications. Regulated industries with data sovereignty requirements can now run Arm-native AI inference inside the same security boundary as their transaction data, without ETL pipelines or cloud exposure.

Read more →

Vietnam at 46%. China at 145%. The Tariff Math on Cooling Hardware.

Trump's reciprocal tariffs land hardest on the countries where CDUs, cold plates, and cooling components are manufactured. CoolIT makes hardware in China and Vietnam. CDU lead times are already 16–24 weeks. The vendors who close 2026 without a clear USMCA-eligible production percentage are quoting prices they cannot hold.

Read more →

How Data Center Cooling Works in 2026: The Complete Guide

The definitive resource on data center cooling. Air cooling fundamentals, direct-to-chip liquid cooling, immersion cooling, CDUs, heat rejection, the PFAS crisis, Nvidia Vera Rubin's 45C warm-water architecture, hybrid retrofit economics, water regulation, and every market number that matters. 13 sections. 31 sources. If you build, operate, invest in, or sell into data centers, this is the map.

Read more →

Vertiv Bought a Heat Exchanger Company. Because the Real Bottleneck Was Never at the Chip.

Vertiv announced the acquisition of ThermoKey, an Italian dry cooler and microchannel heat exchanger manufacturer founded in 1991. CEO Giordano Albertazzi: "Heat rejection is becoming increasingly critical." Liquid cooling concentrates heat at the rack. The outdoor rejection equipment at most sites was never designed for what comes next. Vertiv is building ownership of the full thermal chain.

Read more →

Cooling Is the Single Greatest Risk in Commissioning a 100MW AI Data Center

Applied Digital COO Laura Laltrello and VP of Engineering Stephen Lattimer: mechanical cooling systems are the #1 operational vulnerability at hyperscale. Commissioning starts 30-45 days after groundbreaking. Every assumption the design engineers made gets stress-tested when actual GPUs generate actual heat. The assumptions are always partially wrong.

Read more →

Liquid Cooling Fixed the GPU Problem. Now Storage Is Overheating.

GPUs moved to liquid cooling. The fans disappeared. But every NVMe SSD ever designed assumed airflow. In NVIDIA's fanless GB200 NVL72 racks, storage drives sit in still air with no thermal path. Solidigm shipped the first cold-plate cooled enterprise SSD. Wiwynn built a 100% fanless 96-drive chassis. The bottleneck nobody planned for.

Read more →

Huawei's 600-Watt AI Chip Just Got Orders From ByteDance and Alibaba. The Cooling TAM Just Forked.

The Ascend 950PR draws 600W and delivers 1.56 PFLOPS of FP4 compute. Huawei plans to ship 750,000 units this year. ByteDance and Alibaba are ordering. At 600W per card, these chips cannot be air-cooled at scale. A parallel cooling ecosystem is forming in China with domestic vendors Sugon and Envicool. The thermal engineering is identical. The vendor ecosystems are diverging.

Read more →

70% of Wisconsin Voters Say Data Centers Cost More Than They're Worth. The Cooling Industry Should Listen.

Marquette Law School Poll: 70% opposition, up from 55% four months ago. Bipartisan. State and federal moratorium bills filed. The three flashpoints driving the backlash: cooling noise, water consumption, and grid strain. Operators who show up with immersion cooling and zero evaporative loss will have a different conversation than those proposing rooftop air handlers.

Read more →

Tokyo Wants to Put Data Centers Under Railway Overpasses. The Cooling Problem Is the Whole Experiment.

Four Tokyu Group companies will install a modular data center beneath Tokyo's Oimachi Line starting June 2026. Trains overhead introduce variable thermal conditions and severe vibration. The target workload: generative AI inference. Immersion cooling may be the only viable path. If it works, Tokyu's eight-line railway network becomes a distributed edge computing corridor.

Read more →

From 2 kW to 130 kW Per Rack in 40 Years. Liquid Cooling Is No Longer Optional.

Nvidia's GB200 NVL72 draws 120-130 kW per rack. Cold plate liquid cooling achieves 300x faster heat removal than air. The liquid cooling market hit $3B in 2025, projected $7B by 2029. 59% of operators plan liquid cooling within 5 years. Google, Microsoft, and Meta have already rebuilt their stacks around liquid as the default. Air cooling is the legacy system now.

Read more →

Ecolab Says Cooling Demand Will Outlast Any AI Bubble. They Just Bet $4.75 Billion on It.

Josh Magnuson, Ecolab's general manager for global water solutions, told Bloomberg that cooling demand will grow 20% or more per year "for the foreseeable future," regardless of whether AI is a bubble. "If there's an AI bubble or not, I don't think it matters," he said, days after Ecolab agreed to pay $4.75 billion for CoolIT Systems. KKR is walking away with a 15x return in three years. The cooling industry just got its clearest signal yet that the investment case has decoupled from the AI hype cycle.

Read more →

Trane Just Bought an Immersion Cooling Company. The HVAC Giants Are Done Watching.

Trane Technologies completed its acquisition of LiquidStack on March 3, giving the $21 billion HVAC conglomerate a proprietary two-phase immersion cooling platform. CEO Dave Regnery made three acquisitions in rapid succession: modular cooling plants, direct-to-chip cold plates, and now immersion tanks. Combined with Eaton's $9.5B Boyd deal and Ecolab's $4.75B CoolIT deal, three legacy industrials spent north of $14 billion on liquid cooling companies in March alone.

Read more →

Blackstone Just Bought the Company That Builds Cooling for Fighter Jets. Now It Wants Data Centers.

Blackstone Energy Transition Partners announced a majority stake acquisition in Advanced Cooling Technologies, a Lancaster, Pennsylvania manufacturer of two-phase liquid cooling, heat pipes, and cold plates. ACT, founded by Jon Zuo and Scott Garner in 2003, built its reputation on defense and aerospace thermal engineering. Blackstone's BETP IV fund closed at its $5.6 billion hard cap. ACT slots into a portfolio that already includes power generation, grid infrastructure, and backup power.

Read more →

Data Center Cooling Could Require $58 Billion in Water Infrastructure. The Bill Is Coming Due.

A UC Riverside and Caltech study led by Shaolei Ren found that data center peak water demand could require $10 billion to $58 billion in new water infrastructure by 2030. Peak daily demand from evaporative cooling can spike 6 to 10 times above average, with some facilities exceeding 30x. The aggregate peak need rivals New York City's daily water supply. Google already consumes 33% of The Dalles, Oregon's water. The people paying the infrastructure bill are not the ones training the models.

Read more →

A 141-Year-Old Heat Exchanger Company Just Entered the Data Center Cooling Market. Pay Attention.

Alfa Laval launched FreeWaterLoop at Data Center World in London, its first purpose-built data center product. The $6.6 billion Swedish industrial giant combines 88 years of plate heat exchanger R&D with Framo pump engineering to create a facility-loop cooling system that uses natural water sources and returns them with near-zero net consumption. When a company founded in 1883 redirects R&D toward your industry, the opportunity is real.

Read more →

Munters Booked $200 Million in Cooling Orders and Built a Factory Next to Its Biggest Customers

Munters recorded 2.1 billion SEK ($200M+) in data center cooling orders in 2025, anchored by an $82 million Geoclima Circlemiser chiller deal for an AI facility. The company is opening a 200,000 sq ft Virginia factory in Q2 2026, planting chiller production within driving distance of the largest data center market on the planet. The Circlemiser's cylindrical condenser design increases heat exchange surface by 45% in the same footprint. Proximity and integration are Munters' bet.

Read more →

Nvidia Wants to Put Data Centers in Orbit. The Cooling Equation Is the Whole Point.

Nvidia unveiled the Vera Rubin Space-1 module for AI data centers in orbit, developed with Axiom Space. The initiative treats space as the next compute frontier, where solar power is five times stronger and the vacuum of space serves as an infinite heat sink. For terrestrial cooling vendors, the signal matters more than the timeline: Nvidia is telling the market that ground-based thermal constraints are a binding limitation on AI scaling.

Read more →

TSMC Is Spending $45 Billion This Year. Every Dollar of It Creates Cooling Demand.

TSMC reported $33.7 billion in quarterly revenue, up 35.9% year over year. The HPC segment grew 48% and now accounts for 58% of total revenue. The company has committed $45 billion in 2026 capex and commenced 2nm mass production. For the cooling industry, TSMC's production volume is a leading indicator for thermal management demand 12 to 18 months out. The chip volumes are coming. Whether the cooling supply chain is ready is a different question.

Read more →

Musk's $20 Billion Chip Factory Will Need Cooling Infrastructure Before It Needs Chips

Elon Musk launched the Terafab on Saturday in Austin, a joint venture between Tesla, SpaceX, and xAI to build a 2-nanometer chip fab with a $20 to $25 billion price tag. The facility targets 100,000 wafer starts per month, scaling to one million. The cooling implications are stacked three deep: the fab itself needs industrial-scale process cooling in central Texas, the chips it produces will fill data center racks that need liquid cooling, and the orbital data center ambition signals that the people writing the largest checks view terrestrial cooling as a bottleneck worth $20 billion to circumvent.

Read more →

The Immersion Cooling Market Has a $14 Billion Price Tag. Most of It Hasn't Been Earned Yet.

A new market forecast projects data center immersion cooling will grow from $2.49 billion in 2024 to $14.06 billion by 2034 at an 18.9% CAGR. The numbers look impressive in a slide deck. On the ground, immersion still faces the same obstacles that have kept it in pilot programs for years: operational complexity, workforce readiness, PFAS regulatory risk, and direct-to-chip cold plates eating its lunch at the volume tier.

Read more →

17 Billion Gallons a Year and Counting. The AI Water Crisis Just Got Its Own News Cycle.

U.S. data centers consumed 17 billion gallons of water for cooling in 2023. Projections show 33 billion by 2028. Internal Microsoft forecasts obtained by the New York Times showed the company's own water use tripling by 2030. The Great Lakes region has 738 new data centers announced or under construction. Sam Altman called the concerns "completely untrue." The data says otherwise. Seven major publications ran investigations in the same quarter.

Read more →

Supermicro's Cofounder Allegedly Smuggled $2.5 Billion in Nvidia GPUs to China. The Cooling Supply Chain Should Care.

Supermicro cofounder Wally Liaw was arrested for conspiring to route banned Nvidia H200 and B200 GPUs to China through a Southeast Asian intermediary. The scheme involved dummy servers staged for compliance audits and servers repacked in unmarked boxes. Supermicro stock dropped 33%. For the cooling industry, $2.5 billion in diverted GPU servers means thousands of high-density racks generating thermal loads that need liquid cooling outside legitimate supply chains.

Read more →

Google Sent a Procurement Team to China. They Were Shopping for Liquid Cooling.

A procurement team from Google's Taiwan operations traveled to China this month to meet with Envicool and at least one other manufacturer about purchasing liquid cooling equipment for AI data centers. The trip reveals a supply problem that the entire hyperscale sector is now contending with. The global market for AI server liquid cooling systems is projected to reach more than $17 billion in 2026, up from $8.9 billion last year.

Read more →

Illinois Gave Data Centers $983 Million in Tax Breaks. Now It Wants a Two-Year Pause.

Governor Pritzker has proposed a two-year pause on new tax credits for data center projects. The POWER Act, working its way through committee, would require data centers to source renewable energy, disclose water use, and obtain water permits from the Illinois EPA. Communities across the state have already rejected proposals, and nearly 5,000 petition signatures were filed against one Joliet project.

Read more →

An E-Bike Motor Company in Santa Rosa Thinks It Can Cool 200kW Racks. The Math Is Interesting.

Orbis Electric, a 21-person startup in Sonoma County that started building e-bike motors, has unveiled a rare-earth-free cooling engine for AI data centers. The HaloDrive motor-pump claims 96% efficiency and supports rack densities exceeding 200 kW. Founder Marcus Hays projects $500 million in annual revenue by 2028. The gap between that projection and today's headcount is considerable.

Read more →

NYU Researchers Want to Cool Data Centers with Truckloads of Dried Rocks. The Energy Savings Are Hard to Ignore.

A team at NYU Tandon has proposed using zeolite thermal batteries, charged with factory waste heat and trucked to data centers, to replace compression chillers. The modeled result: an 86% reduction in cooling electricity consumption. The system is still on paper, but the median distance between U.S. data centers and the nearest industrial waste heat sources is just 57 kilometers.

Read more →

Microsoft Built Cooling Channels Directly Into the Chip. It Works Three Times Better Than Cold Plates.

Microsoft has developed an in-chip microfluidics cooling system that removes heat up to three times more efficiently than current cold plate technology. Microscopic channels etched into the back surface of a silicon die circulate coolant at the exact locations where heat is generated. Lab testing showed a 65% reduction in maximum GPU temperature rise. The approach, developed with Swiss startup Corintis, could shift the cooling function from the data center supply chain into semiconductor packaging itself.

Read more →

Nvidia Just Made Liquid Cooling Standard on Rubin. The Optional Era Is Over.

Nvidia's Vera Rubin architecture will ship with liquid cooling as a standard component in H2 2026. At GTC, Nvidia named four cold plate suppliers, including Asia Vital Components and Cooler Master, and released standardized specifications. Cooling infrastructure costs exceed $57,000 per system. Jensen Huang stated that power delivery and liquid cooling now require co-design. The GPU company just decided the cooling question for every operator buying its hardware.

Read more →

Seawater Can Cool Data Centers. Most Operators Haven't Figured Out How to Use It.

Evaporative cooling is incompatible with seawater due to salt fouling — but three viable approaches exist. Google uses Wet Surface Air Cooling at its Finland facility. Microsoft's Project Natick used closed-loop cold seawater from the ocean floor. The fresh water crisis makes the coastal cooling conversation urgent for the industry's next buildout cycle.

Read more →

The Water-Power Tradeoff That Data Center Operators Keep Getting Wrong

A growing number of data center operators have started swapping water-cooled systems for air-cooled alternatives, claiming sustainability wins. The math tells a different story. Air cooling eliminates on-site water use, sure. But it doubles or triples electricity consumption, pushing the water burden upstream to power plants that need their own cooling loops to generate that extra juice.

Read more →

A growing number of data center operators have started swapping water-cooled systems for air-cooled alternatives, claiming sustainability wins. The math tells a different story. Air cooling eliminates on-site water use, sure. But it doubles or triples electricity consumption, pushing the water burden upstream to power plants that need their own cooling loops to generate that extra juice. The problem doesn't vanish. It moves.

Colocation vacancy rates have cratered to 2.3%, down from 9.8% in 2020. The construction pipeline grew tenfold over the same period. Every new facility that comes online has to make a fundamental call on how it manages heat, and that decision ripples through local water tables and power grids for decades. A large data center drinks roughly what a town of 50,000 people does in a day. Regulators in multiple states have started blocking projects over that kind of draw.

The operators getting this right tend to match their cooling architecture to their actual scale. Hyperscalers running 100+ MW loads are exploring on-site power generation, including hydrogen fuel cells that produce water as a byproduct. Facilities in central Ohio are already piloting private microgrids built around this concept. Mid-tier and edge deployments, meanwhile, are finding that modern evaporative cooling towers can hit the efficiency marks without the electricity penalty. And micro data centers, anything from a large closet to a shipping container, remain firmly in air-cooling territory, where even the smallest cooling tower would be ten times more capacity than needed.

True sustainability means refusing to solve one problem by creating another. The operators who claim green credentials while tripling their grid draw are playing an accounting trick, not running an efficient facility.

Adaptive Cooling, Immersion Bets, and the Vendors Shaping What Comes Next

Schneider Electric ships cooling units packed with IoT sensors that run predictive maintenance cycles before failures happen. Iceotope has built immersion cooling platforms that work across traditional, hyperscale, and edge environments, pushing PUE numbers into territory that air-cooled facilities cannot touch. Three very different approaches from three companies that agree on one thing: the old way of blowing cold air through server rows has a ceiling, and the industry is about to hit it.

Read more →

Schneider Electric ships cooling units packed with IoT sensors that run predictive maintenance cycles before failures happen. Iceotope has built immersion cooling platforms that work across traditional, hyperscale, and edge environments, pushing PUE numbers into territory that air-cooled facilities cannot touch. Stulz leans on free cooling and precise humidity control to shave CO2 output. Three very different approaches from three companies that agree on one thing: the old way of blowing cold air through server rows has a ceiling, and the industry is about to hit it.

The next shift is adaptive cooling, systems that use AI to learn a facility's thermal behavior in real time and adjust output to match actual load. Most data centers today over-cool by 20 to 40% because their control systems react to worst-case thresholds rather than live conditions. Adaptive systems eliminate that cushion, and the energy savings compound across thousands of racks.

Digital Realty introduced direct liquid cooling across 170 data centers worldwide in 2024, signaling that the colocation giants see liquid as table stakes rather than a premium add-on. Edge computing adds another dimension. Smaller facilities in distributed locations create opportunities for cooling designs that would never make sense at hyperscale, from geothermal loops to ambient-air setups in northern climates.

The vendors winning contracts right now are the ones who can deliver across all three tiers: hyperscale, colo, and edge. Single-product companies are getting boxed out.

500 Megawatts in an Indiana Cornfield: The Physical Cost of the AI Buildout

Pull up satellite imagery of New Carlisle, Indiana, from 2023 and you see farmland. Pull it up today and you see seven rectangular data centers with 23 more permitted. A single campus there already draws over 500 megawatts, enough to power several hundred thousand homes. When the full build finishes, the load will exceed what two cities the size of Atlanta consume.

Read more →

Pull up satellite imagery of New Carlisle, Indiana, from 2023 and you see farmland. Pull it up today and you see seven rectangular data centers with 23 more permitted. A single campus there already draws over 500 megawatts, enough to power several hundred thousand homes. When the full build finishes, the load will exceed what two cities the size of Atlanta consume.

The Atlantic's Matteo Wong reported from these sites, including Memphis, where a new data center megaproject sits downwind from an active natural-gas plant in a neighborhood already dealing with pollution from decades of industrial use. KeShaun Pearson, who runs the nonprofit Memphis Community Against Pollution, told Wong the area's air already tastes like soot and asphalt. Another facility won't improve things.

The numbers at a national level tell the same story. U.S. data centers consumed 176 terawatt-hours in 2023, roughly 4.4% of total national electricity. Globally, the figure hit 415 TWh in 2024 and is projected to double to 945 TWh by 2030. AI-related capital spending now accounts for 92% of GDP growth in the first half of 2025, and the tech sector has ballooned from 22% to a third of the S&P 500 since ChatGPT launched. That concentration of economic activity in a single sector, built on a single resource constraint, should make anyone in infrastructure planning pay attention.

Cooling is the bottleneck inside the bottleneck. Forty percent of a data center's electricity goes to thermal management. At the densities AI training requires, the cooling problem scales faster than the compute problem.

67% Energy Savings Are on the Table. Most Data Centers Leave Them There.

A comprehensive review published in the International Journal of Refrigeration examined every major cooling optimization technology available to data centers today. The headline finding: advanced cooling architectures can cut energy consumption by up to 67.2% compared to conventional setups. The industry average PUE, according to Uptime Institute's 2024 survey, sits at 1.56. State-of-the-art facilities report 1.06. That gap represents billions of kilowatt-hours left on the table every year.

Read more →

A comprehensive review published in the International Journal of Refrigeration examined every major cooling optimization technology available to data centers today. The headline finding: advanced cooling architectures can cut energy consumption by up to 67.2% compared to conventional setups. The industry average PUE, according to Uptime Institute's 2024 survey, sits at 1.56. State-of-the-art facilities report 1.06. That gap represents billions of kilowatt-hours left on the table every year.

The research breaks down where the waste lives. In a typical data center, only 30% of electricity actually reaches the servers doing useful work. The thermal management stack, air conditioning, chillers, humidifiers, consumes 45%. The rest goes to power distribution and overhead. Those proportions have been roughly stable for years, which means the industry has been building new capacity without fixing the fundamental inefficiency of how it cools existing capacity.

Liquid cold plates, immersion tanks, heat pipes, and thermosiphon-based systems all showed measurable PUE improvements in the review. AWS reported a 46% drop in mechanical cooling energy after deploying a custom liquid solution, bringing its global PUE to 1.15. Vertiv's data shows that moving to 75% liquid cooling in a hybrid facility cuts total site power consumption by 15.5%.

Microprocessor thermal design power is expected to blow past 700 watts this year. Air cooling tops out around 280 watts. The arithmetic on when liquid becomes mandatory has already been done. The only question is how many operators will wait until they have no other option.

Lonestar Data Holdings Raised $66 Million to Build a Data Center on the Moon.

Lonestar Data Holdings has raised $66 million to develop data center infrastructure on the lunar surface. The moon's far side offers electromagnetic shielding and falls outside any national jurisdiction. Lunar thermal cycling runs from 127°C at noon to -173°C at night — the core engineering challenge the program must solve before the first hardware ships.

Read more →

$64 Billion in Data Center Projects Blocked or Delayed by the People Next Door

Community opposition has stalled or killed $64 billion worth of U.S. data center projects. That figure comes from Good Jobs First, which has been tracking the growing collision between hyperscale ambitions and local resistance since the buildout accelerated in 2024.

Read more →

Community opposition has stalled or killed $64 billion worth of U.S. data center projects. That figure comes from Good Jobs First, which has been tracking the growing collision between hyperscale ambitions and local resistance since the buildout accelerated in 2024.

The opposition is not coming from environmentalists alone. Homeowners worried about property values. Farmers who do not want to sell to a land agent working for an unnamed tech company. Municipal leaders watching their water tables drop. School boards wondering why a $2 billion facility pays almost nothing in property taxes thanks to abatement deals negotiated behind closed doors.

New York's S.9144, introduced in early 2026, would impose a three-year statewide pause on permits for data centers drawing 20 megawatts or more. The bill has not passed its originating chamber, but the fact that it was introduced in one of the country's most important data center markets says something about where the political winds are blowing.

Moratorium bills have been introduced in 11 states across 14 separate pieces of legislation in 2026. None have passed yet. But the pattern is consistent: proposals are getting more specific, the sponsors are getting more serious, and the public comment periods are getting louder.

Over 300 data center bills were filed across more than 30 states in the first six weeks of 2026 legislative sessions. The industry spent years operating in an incentive-friendly regulatory environment. That environment is shifting. Operators who plan multi-year construction timelines without accounting for community opposition are building schedule risk into every project.

Why Immersion Cooling Keeps Losing to Cold Plates

Immersion cooling can hit a PUE of 1.02. Direct-to-chip liquid cooling lands around 1.15 to 1.20. On raw thermal efficiency, immersion wins. It has won that comparison for years. And yet direct-to-chip holds 47% of the liquid cooling market while immersion sits at roughly $270 million.

Read more →

Immersion cooling can hit a PUE of 1.02. Direct-to-chip liquid cooling lands around 1.15 to 1.20. On raw thermal efficiency, immersion wins. It has won that comparison for years. And yet direct-to-chip holds 47% of the liquid cooling market while immersion sits at roughly $270 million, growing at 25% CAGR toward a projected $2.54 billion by 2032.

The gap between what works in a lab and what ships in volume comes down to three things that have nothing to do with thermodynamics.

Server compatibility is the first. Direct-to-chip cold plates mount onto existing CPU and GPU packages inside standard server chassis. Dell, HPE, and Lenovo all offer factory-integrated DTC options. Immersion requires purpose-built or heavily modified servers. Standard components with standard connectors and standard cable routing do not survive submersion in dielectric fluid.

Workforce readiness is the second. Most data center operations teams have spent their careers managing air-cooled environments. DTC adds manifolds, hoses, and coolant distribution units. Immersion asks a maintenance technician to pull a server out of a tank of fluid, let it drain, service it, and resubmerge it.

Retrofit economics is the third. DTC fits into existing rack infrastructure with manageable modifications. Immersion requires tanks, fluid inventory, specialized containment, and a fundamentally different floor layout. For the majority of operators adding liquid cooling to facilities that were built for air, DTC is the path that does not require gutting the room.

The Liquid Cooling Supply Chain Race Has Three Frontrunners and a Hundred Chasers

The data center liquid cooling market hit $5.52 billion in December 2025. It is projected to reach $15.75 billion by 2030. That kind of growth rate attracts everyone. The question is who can actually manufacture at the scale the buildout demands.

Read more →

The data center liquid cooling market hit $5.52 billion in December 2025. It is projected to reach $15.75 billion by 2030. That kind of growth rate attracts everyone. The question is who can actually manufacture at the scale the buildout demands.

Schneider Electric moved first among the industrial conglomerates. Their acquisition of Motivair in February 2025 gave them a dedicated liquid cooling portfolio: coolant distribution units, ChilledDoor rear-door heat exchangers, dynamic cold plates, and chillers. Since the acquisition, Motivair has opened a fourth production facility and is tripling global manufacturing capacity across plants in Buffalo, Italy, and India.

Vertiv has been in the cooling business longer than most of its competitors have existed. Their rear-door heat exchangers and CDU product lines are specified by default at several major colocation providers.

Eaton's $9.5 billion acquisition of Boyd Thermal was the largest pure-play cooling deal in data center history. Boyd brings manufacturing depth in heat exchangers, cold plates, and thermal interface materials. The combination mirrors Schneider's strategy: own enough of the cooling and power stack to sell integrated solutions.

Demand is running ahead of supply across CDUs, cold plates, and rear-door heat exchangers. Lead times have stretched. The vendors who can ship on schedule will capture market share regardless of whose product benchmarks better on a spec sheet.

NVIDIA's Watt Roadmap Is Writing the Cooling Industry's Business Plan

Every cooling technology decision being made in data centers right now traces back to a single forcing function: how many watts NVIDIA's next GPU generates. The H100 runs at 700 watts. The Blackwell B200 pushes 1,000 watts. Rubin, the next generation, is expected to climb higher.

Read more →

Every cooling technology decision being made in data centers right now traces back to a single forcing function: how many watts NVIDIA's next GPU generates. The H100 runs at 700 watts. The Blackwell B200 pushes 1,000 watts. Rubin, the next generation, is expected to climb higher. Each step up the power ladder compounds the thermal load per rack, per row, per facility.

Air cooling tops out around 25 to 30 kW per rack. That ceiling has not moved meaningfully in years and it will not move meaningfully in the future. The physics of convective heat transfer through air set a hard limit.

Non-AI workloads across the global data center fleet total approximately 38 gigawatts. AI workloads are expected to hit 44 GW in 2026. The crossover point, where AI thermal load exceeds everything else combined, is arriving this year.

The cooling industry is, in effect, building to NVIDIA's spec. When Jensen Huang announces a new chip architecture, the thermal management implications ripple through CDU manufacturers, cold plate suppliers, and facility designers within weeks. The vendors who can design, qualify, and ship cooling hardware matched to the next GPU generation before that generation reaches volume production will own the upgrade cycle.

SpaceX Filed for Up to 1 Million Satellites. The Orbital Data Center Megaconstellation Is the Goal.

SpaceX's FCC filing for up to 1 million satellites frames orbital data center infrastructure at a scale that dwarfs any terrestrial buildout. Current Starlink sits at roughly 6,000 satellites. Getting from there to a viable orbital compute constellation requires hitting a $10/kg launch cost target. The economics, not the physics, are the binding constraint.

Read more →

Frore Systems Just Raised $143 Million at a $1.64 Billion Valuation for Chip-Level Cooling

Frore Systems closed a funding round that valued the company at $1.64 billion. The $143 million raise was led by MVP Ventures, with participation from Fidelity and Qualcomm Ventures. The company makes solid-state cooling devices. No fans. No moving fluids.

Read more →

Frore Systems closed a funding round that valued the company at $1.64 billion. The $143 million raise was led by MVP Ventures, with participation from Fidelity and Qualcomm Ventures.

The company makes solid-state cooling devices. No fans. No moving fluids. Frore's AirJet technology uses piezoelectric membranes that vibrate at ultrasonic frequencies to create localized airflow directly over a chip surface. The data center application extends the same principle to GPU and CPU packages where targeted, high-velocity airflow can supplement or replace broader cooling architectures.

A $1.64 billion valuation for a cooling component company is extraordinary for this category. For context, the entire immersion cooling segment is projected at $2.54 billion by 2032. Frore is valued at 65% of that projected market before its data center product is widely deployed.

Whether AirJet technology scales to the thermal loads of a 1,000-watt Blackwell GPU remains to be demonstrated at production volume. Frore does not need to replace liquid cooling. It needs to prove that a chip-level supplement adds enough thermal headroom to justify the per-unit cost. At $1.64 billion, the market is betting it can.

Zero-Water Cooling Pilots Are Launching in Phoenix and Mt. Pleasant. The Results Will Set the Standard.

Two facilities opening in 2026 will answer the question that the entire cooling industry has been arguing about: can you cool a high-density data center without consuming water, in a climate where you actually need cooling?

Read more →

Two facilities opening in 2026 will answer the question that the entire cooling industry has been arguing about: can you cool a high-density data center without consuming water, in a climate where you actually need cooling?

Phoenix, Arizona is the first test. A zero-water pilot project launching this year in a market where summer temperatures regularly exceed 110 degrees Fahrenheit and the municipal water supply depends on a Colorado River system that has been in sustained decline for two decades. If zero-water cooling works in Phoenix, it works everywhere in the continental United States.

Mt. Pleasant, Wisconsin is the second. A facility designed to operate without consumptive water use in a humid Midwestern climate where the thermal challenge is different. Humidity limits the effectiveness of certain dry cooling approaches.

The technology stack for zero-water cooling is well understood. Dry coolers, closed-loop liquid systems, and heat rejection without evaporation. The tradeoff is capacity and cost. Dry coolers sized for peak ambient temperatures in Phoenix require two to three times the radiator surface area and fan power than an equivalent evaporative system.

The results from these pilots will influence permitting decisions, facility design standards, and vendor selection for the next generation of builds. If zero-water performance holds at commercially viable cost points, the argument for new evaporative installations in water-stressed regions collapses.

The Biggest Barrier to Liquid Cooling Adoption Has Nothing to Do With Technology

Ask a cooling vendor what slows down liquid cooling deployment and you will hear about PUE, capex, and fluid compatibility. Ask the operator who just signed a purchase order for 500 racks of direct-to-chip cooling and the answer is different. They cannot find people who know how to install it.

Read more →

Ask a cooling vendor what slows down liquid cooling deployment and you will hear about PUE, capex, and fluid compatibility. Ask the operator who just signed a purchase order for 500 racks of direct-to-chip cooling and the answer is different. They cannot find people who know how to install it, maintain it, and troubleshoot it when something goes wrong.

The workforce skills gap is the single most cited operational barrier to liquid cooling adoption in 2026. The data center industry spent two decades building a labor force trained on air-cooled infrastructure. CRAC units, raised floors, hot-aisle/cold-aisle containment. Those skills are mature and widely distributed.

Liquid cooling introduces an entirely different set of competencies. Plumbing and fluid dynamics replace airflow management. Leak detection systems require different monitoring protocols. Cold plate connections at the server level demand mechanical precision that a technician accustomed to swapping fans and filters has never been asked to deliver.

Some operators are solving this by partnering directly with cooling vendors for managed maintenance. Others are investing in internal training programs, pulling from adjacent trades like HVAC, plumbing, and industrial process cooling where the mechanical skills overlap.

The vendors who bundle training and certification into their sales process will have an advantage that does not show up on a spec sheet. The cooling hardware market has multiple credible suppliers. The cooling labor market does not have enough credible technicians. That imbalance will shape purchasing decisions as much as price or performance.

Bezos Says Gigawatt-Scale Data Centers in Space Are Coming in 10 to 20 Years. He's Building the Rocket to Prove It.

At Italian Tech Week, Jeff Bezos laid out the case for gigawatt-scale data centers in orbit: no clouds, no rain, no weather, effectively unlimited solar power. He named AI training as the target workload. Blue Origin's New Glenn is the infrastructure bet underlying the claim. The terrestrial cooling industry has 10 to 20 years to make the economics of ground-based facilities look better than what he's describing.

Read more →