By Aaron Foyer
VP, Orennia
In the summer of 1958, a man named Jack Kilby working at Texas Instruments was enduring one of the worst droughts in Texas history. Stuck inside and with time to spare, the electrical engineer took to trying to solve an emerging problem in electronics: how to shrink the electronic circuit.
To amplify electrical signals at the time, circuits relied on transistors, bulky devices that resembled the alien tripods from War of the Worlds. A circuit with 1,000 transistors that could require 10,000 hand-soldered connections is not exactly scalable or cheap.
That September, Kilby showed off a prototype of his miniature electric circuit made of germanium, aluminum and gold, all on a single chip. Despite looking more like a high science lab gone wrong, it worked and might have been the most important invention of the 20th century. The National Academy of the Sciences attributed Kilby’s integrated circuit to launching Second Industrial Revolution.
The importance of the invention took even Kilby by surprise. “What we didn’t realize (when we developed the integrated circuit) was that it would reduce the cost of electronic functions by a factor of a million to one,” said Kilby.
Fast forward to 2024: It’s safe to assume Kilby also didn’t anticipate his integrated circuit would eventually send the entire American power sector into a furor from warehouses containing millions of integrated circuits, but here we are. Data centers are, to borrow from the critically acclaimed movie, everything everywhere all at once.
If you’re worried that we’re too far into the data center boom now to ask questions like what exactly hyperscalers are and why servers use so much electricity, you’re not alone. Here’s an overview of data centers.
Background
To start, data centers have been around for a while. The first were built during World War II, well before Jack Kilby’s integrated circuit. At the time, the US military used data centers to run the world’s first programable computer which it applied to calculate artillery fire and to help build the atomic bomb. Their development was supercharged by the Cold War and companies like IBM helped commercialize data centers for industry.
Enter AI: Skipping over several key chapters like the development of the microprocessor, the rise of the personal computer, the launch of the internet, the dot-com boom, Counter-Strike LAN parties, smartphones and cloud computing, we arrive in the early 2020s. OpenAI’s launch of ChatGPT, an artificial intelligence (AI) chatbot blew open the public’s imagination about the technology’s potential. Shortly thereafter, attention was drawn to the huge energy needs of the data centers needed to train and operate these programs.
Electricity Use of AI by Comparison
Many users have already substituted Google for ChatGPT, which is many times more energy intensive. A single request on OpenAI’s chatbot requires 10x more power than a Google search. Asking an AI image creator like DALL-E to imagine what Harry Potter would have looked like in the 1970s uses energy more comparable to driving a Tesla than running an internet search. And using software like Sora to make a single AI video rivals running an air conditioner for an hour. It’s no surprise the heavy use of these technologies has people thinking about how they’ll satisfy the energy needs.
The rise of industrial data centers
This may come as a shock to some GenZ readers, but there was a time not too long ago when most companies hosted their own servers in data spaces in their own buildings, on premises or “on prem”. As the world digitized and the use of the cloud grew, so did demand for larger spaces designed specifically for housing and networking servers. This eventually led to today’s modern data centers—large, 100,000-square-foot, shoe-box-shaped buildings that can house tens of thousands of servers. Backup infrastructure, data management and cooling systems in the buildings all ensure servers run efficiently and reliably.
Data Centers Across the US
Scale also helps with cost: Instead of hundreds of companies, with hundreds of cooling systems and hundreds of IT professions, shared resources make data more affordable to own and operate.
The various types
An Apple iPhone user looking to store 900 pictures of their cockapoo on the cloud has very different data needs from a bank needing to securely protect the banking details of two hundred thousand clients. This creates natural segmentation in the industry.
The key segment types:
- Hyperscalers: The largest data centers around, used for cloud computing and data management. They usually contain more than 5,000 servers with power demands that can exceed 100 megawatts. Hyperscalers are often owners and operated by household names like Amazon, Google and Microsoft.
- Enterprise: More modest in size compared to hyperscalers, the fully company-owned facilities are used to process internal data and host applications critical to company operations. Think JPMorgan Chase, ExxonMobil, AT&T, Pfizer or General Motors.
- Colocation providers: These data centers allow companies or individuals to rent equipment, space, and bandwidth from the owner. Colocation providers can technically be hyperscalers but serve multiple customers. Users might include small- and medium-sized enterprises, startups, healthcare providers, and government agencies.
- Edge data centers: Small, localized facilities for processing data closer to the source or end users. For example, AT&T to support 5G networks, Netflix to reduce buffering or Sony to minimize latency for gaming.
- Real Estate Investment Trusts (REITS): These operators invest in data center properties and lease space to various tenants similar to real estate management.
- Other types: There are many, including micro data centers for remote or mobile environments, high-performance computing data centers for complex modeling in science and research and carrier hotels that serve as hubs for network interconnections.
The various data center companies operate in different segments. In aggregate, the companies with the largest total data center capacities in the US are colocation providers, according to our data.
The 20 Largest Operating Data Center Owners in the US
With different user types come different needs and preferences. Some data customers are more price sensitive but lenient on things like reliability and downtime. On the flip side, others will pay hefty premiums to avoid downtime altogether. This introduces the concept of the four tiers.
Data center tiers:
- Tier 1: The lowest tier, characterized by basic infrastructure, zero redundancy and a single distribution path. Annual downtime is estimated to be 28.8 hours.
- Tier 2: Offers improved redundancy over Tier 1, with some backup power and cooling, but still maintains a single distribution path. Annual downtime is estimated to be 22 hours.
- Tier 3: Features multiple power and cooling paths, and each component can be removed or replaced without affecting operations. Annual downtime is estimated to be 1.6 hours.
- Tier 4: The highest tier, with multiple independent and physically isolated systems for power and cooling. Annual downtime is estimated to be 0.4 hours.
Local retailers or small consultancies might be just fine with the small downtimes that come with the lower costs of a Tier 1 data center, but major banks and large e-commerce websites like Amazon require the reliability of Tier 4.
Power consumption
What has really transformed the power sector is how much electricity data centers currently use superimposed on their expected growth. According to the International Energy Agency, data centers accounted for 1% to 1.3% of global electricity demand in 2022, that’s between 230 and 340 terawatt hours. Those numbers don’t include crypto mining, which accounts for another 0.4% of global power demand.
Breakdown of the Energy Consumption of a Data Center
The surge in applications of AI is driving the need for more data centers, which is expected to contribute meaningfully to load growth over the rest of the decade. Goldman Sachs estimates global power demand from data centers will increase 160% by 2030 and will be the single largest driver of new electricity demand over the period, beating out electric vehicles and home electrification.
What the power is used for: First and foremost, for the basic function of the microchip. Electricity is used to switch on and off the millions of transistors on a chip to process information and then transmit signals, this accounts for 40% of data center energy use. Cooling plays an equal roll. Servers and IT equipment generate heat as they operate, which degrades their performance. Air and liquid cooling systems improve equipment function by removing waste heat, a process that accounts for another 40% of energy consumption. The remaining 20% is balance of plant—power supply, equipment and storage.
Looking ahead: Will improved chip efficiencies save us?
Billion-dollar problems often spur billion-dollar solutions. The idea of access to power being the limiting factor in data center growth paired with challenges in siting new facilities have prompted companies to make chips more efficient. Earlier this year, chipmaker Nvidia announced that its next-generation Blackwell graphics processing unit (GPU) used for AI workloads had an energy efficiency 25 times better than its previous model. While markedly better, it’s unlikely to tame load growth.
Why not: With access to more efficient chips, developers will almost certainly elect to improve data center performance over making a smaller facility that uses less power. This would be in keeping with history.
Since the debut of Jack Kilby’s integrated circuit in the mid-1950s, the energy efficiency of chips has improved by ~15 billion times. Improved efficiencies did not reduce the total system energy use of computing, but instead created opportunity for new use cases and industries to emerge. Improved GPU chips are more likely to lower costs, making it more affordable for industries where computation is currently too expensive.
Alternative endings
The future of AI and data centers is not set in stone. There has been a growing chorus of views that the pace of investment into AI is unsustainable. In late June, Goldman Sachs published a white paper that went viral highlighting how leading tech giants look geared to spend $1 trillion on AI capex, but spending to date has little economic benefit to show for it.
On the other hand, AI also shows flashes of incredible potential. In 2021, the AI program AlphaFold led by Google accurately predicted protein structures, something that has challenged biologists for 50 years and paves the way for drug discovery and better understanding diseases. And earlier this year, researchers at Princeton used AI to solve plasma instability in nuclear fusion reactors, a key hurdle to overcome on the road to their commercialization.
Sparked thoughts
- Jevons Paradox at work? All the claims that improved efficiencies will solve our data center power woes just don’t jive with human behavior. Better chips will be a benefit to humanity, but that will almost certainly come from new industries emerging, the result of cheaper computation costs rather than energy savings.
- AI looking bubbly? The Goldman Sachs white paper suggested that nearly $1 trillion will be spent on AI capex over the coming years, but that US GDP would grow just 0.9% as a result. Those two numbers don’t jive… maybe that’s why Nvidia’s share price has fallen 15% since June.
- Economics, the ultimate hammer? In the end, economics trump engineering. Companies will eventually be forced to justify spending hundreds of billions on capex for AI data centers. If the economics don’t go around, the taps will shut.
Aaron Foyer is Vice President, Research and Analytics at Orennia. Prior to Orennia, he leveraged his technical background in management consulting and finance roles. He has experience across the energy landscape including clean hydrogen, renewables, biofuels, oil and gas, petrochemicals and carbon capture.