The surge of interest in AI is creating a massive demand for computing power. Around the world, companies are trying to keep up with the vast amount of GPUs needed to power more and more advanced AI models. While GPUs are not the only option for running an AI model, they have become the hardware of choice due to their ability to efficiently handle multiple operations simultaneously—a critical feature when developing deep learning models.
But not every AI startup has the capital to invest in the huge numbers of GPUs now required to run a cutting-edge model. For some, it’s a better deal to outsource it. This has led to the rise of a new business: GPU-as-a-Service (GPUaaS). In recent years, companies like Hyperbolic, Kinesis, Runpod, and Vast.ai have sprouted up to remotely offer their clients the needed processing power.
While tech giants like Amazon or Microsoft offering cloud computing services own their infrastructure, smaller startups like Kinesis have created techniques to make the best out of the existing idle compute.
“Businesses need compute. They need the model to be trained or their applications to be run; they don’t necessarily need to own or manage servers,” says Bina Khimani, co-founder of Kinesis.
Studies have shown that more than half of the existing GPUs are not in use at any given time. Whether we’re talking personal computers or colossal server farms, a lot of processing capacity is under-utilized. What Kinesis does is identify idle compute—both for GPUs and CPUs—in servers worldwide and compile them into a single computing source for companies to use. Kinesis partners with universities, data centers, companies, and individuals who are willing to sell their unused computing power. Through a special software installed on their servers, Kinesis detects idle processing units, preps them, and offers them to their clients for temporary use.
“At Kinesis, we have developed technology to pool together fragmented, idle compute power and repurpose it into a server-less, auto-managed computing platform,” says Khimani. Kinesis customers even have the possibility to choose from where they want their GPUs or CPUs to come.
AI Is Growing Faster Than Servers Can Keep Up
GPUaaS is filling a growing gap in the AI industry. As learning models get more sophisticated, they need more power and an infrastructure that can process information faster and faster. In other words, without a sufficient number of GPUs, big AI models cannot operate—let alone improve. In October, OpenAI’s CEO, Sam Altman, admitted that the company was not releasing products as often as they had wished because they were facing “a lot of limitations” with their computing capacity.
Also in October, Microsoft’s CFO, Amy Woods, told the company’s investors in a conference call that demand for AI “continues to be higher” than their “available capacity.”
The biggest advantage of GPUaaS is economical. By removing the need to purchase and maintain the physical infrastructure, it allows companies to avoid investing in servers and IT management, and to instead put their resources toward improving their own deep learning, large language, and large vision models. It also lets customers pay for the exact amount of GPUs they use, saving the costs of the inevitable idle compute that would come with their own servers.
Server-less startups like Kinesis also claim to be friendlier to the environment than traditional cloud computing companies. By leveraging existing, unused processing units instead of powering additional servers, they say they significantly reduce energy consumption. In the last five years, big tech companies like Google and Microsoft have seen their carbon emissions soar due to the amount of energy consumed by AI. In response, some have turned their eyes to nuclear energy to sustainably power their servers. Kinesis and other new startups offer a third route in which no further servers need to be plugged in.
“Industry leaders are deeply committed to sustainability,” Khimani says. “With the focus on innovation and efficiency, they can optimize existing computing power that is already active and consuming energy, rather than continually adding more servers for every new application they run.”
The growing demand for machine learning and colossal data consumption is turning GPUaaS into a very profitable tech sector. In 2023, the industry’s market size was valued at US $3.23 billion; in 2024, it grew to $4.31 billion. It’s expected to rise to $49.84 billion by 2032.
“The AI industry is rapidly advancing to a stage where the focus is shifting from merely building and training models to optimizing efficiency,” Khimani says. “Customers are increasingly asking questions like, ‘When training a new model, how can we do it extremely targeted and not consume an ocean of data that requires an enormous amount of compute and energy?’”
From Your Site Articles
Related Articles Around the Web