Skip to content

Exploring the future of tech and economics.

Go back

Hosting the world

By Leo Traven

Edit page

TL;DR:

Living at the foundation of the digital economy

Over the last decade, hyperscalers like Amazon Web Services (AWS), Microsoft Azure and Google Cloud have become the backbone of the IT infrastructure of the biggest companies in the world. They have managed to be at the very foundation of the global digital economy. By doing so, Amazon, Microsoft and Google got into the top 5 of the world’s most valuable companies. “Hyperscaling” is derived from the kind of IT architecture these companies provide. It is designed to scale seamlessly to meet massive fluctuations in demand. This way, their customers can provision cloud resources depending on their needs without worrying about data centers. In the past, the excessive amounts of capital they put into expanding their services led to great financial returns.

Betting the company on AI

During the AI hype, the infrastructure spending of the hyperscalers has reached another dimension. Nowadays, the financial toll is so significant that they begin to face significant headwinds from investors. The money is being used to buy Nvidia GPUs, design custom chips, secure real estate and lock down power grids to feed massive data centers. The bills are not purely paid for by cashflow, but increasingly with debt. From the perspective of the hyperscalers, they feel like they have no other choice: the promised potential of AI is so huge that not taking the risk could mean that the company gets surpassed in the future.

The amount of actual intelligence per compute output generated inside these new data centers is accelerating beyond historical norms. Instead of doubling every two years, computing power per dollar nowadays doubles roughly every 10 months due to highly specialized AI accelerators. Meanwhile, the relative growth rate of data center capacity with respect to power usage has stayed more or less the same: the total power capacity has historically doubled roughly every 4 years, which is still the case. The only problem is that the absolute numbers tend to become very large as they continue to grow exponentially even when the growth rate stays constant.

Will the infrastructure spending break the cloud business model?

Given the scale of their spending, the question is how fast enterprise adoption of AI will be. If the global economy adopts AI too slowly, the utilization of data centers won’t be high enough to pay the bills. In this case, the excess data centers would just blow up the balance sheets without really bringing business value. They would decrease the profitability of the hyperscalers.

Aligned short term incentives improve long term data center utilization

The organizations that want to be successful in the long term will have to use AI to improve their offers and processes. Because it’s not at the core of their competence, most organizations are not willing to build their own data centers for services that are not absolutely critical. As the people taking investment decisions at companies are usually not incentivised for decades, they tend to accept dependencies that materialize only when they have most probably left the company. This is not to be understood as an accusation: it’s almost impossible to show short term results and think decades into the future at the same time. Especially in these dynamic times, it’s very hard to balance future and present results. This will lead to many organizations willingly buying the ever-increasing pay-per-use AI services as they are pushed into the market by the hyperscalers with full force.

This way, hyperscalers will continue to build the ultimate moat: competing with these multi-billion dollar companies is almost impossible as no one else has the financial capacity. Additionally, companies that once accepted to build on a cloud platform are usually stuck with precisely this platform. As the cloud services are oftentimes somewhat specialized and tightly integrated into the company, customers realistically cannot get out without transforming their entire organization. Therefore, it can be expected that price increases will not lead to happy customers, but will generally be accepted as there is no other realistic choice. It’s a fact that the dependency on the big data center companies is already so huge that letting them fail would pose a systemic risk for the whole world economy.

Agility at scale

Let’s not forget that the hyperscalers are doing an incredible job. Constantly living in the future, they are working on vertically integrating their AI data center supply chain. This means that, at this point in time, they have already been building their own AI models and AI chips for years and will continue to do so. By doing so, they will increase their resilience in the long term. They have shown an incredible amount of agility for their size again and again. One of the newest examples is Google that has managed to become one of the AI leaders after being humiliated by young AI labs only 2.5 years ago.

In the AI goldrush, the hyperscalers are lending the shovels to their customers. This way, they enable them to drive the transformation process of their company. By absorbing massive costs now, they manifest their monopoly-like status over artificial intelligence compute, one of the most important utilities of the 21st century.


Edit page
Share this post on:

Previous Post
The power of system design
Next Post
The infrastructure of AI