HomeNewsOpinionLarge AI Models: Their huge costs mean there’s going to be no free AI

Large AI Models: Their huge costs mean there’s going to be no free AI

Large Language Models like ChatGPT and Bard burn a lot of cash in chip acquisition and various operational costs. These have implications like the latest AI offerings not being made publicly available because every free user costs money for the company at the backend, and not-for-profit business models becoming rendered unviable. How AI companies monetise LLMs will be keenly watched

July 07, 2023 / 14:44 IST
Story continues below Advertisement
Artificial intelligence
A very significant aspect of why it took so long to publicly deploy powerful generative AI systems is the humongous cost associated with training and deploying them.

As the ubiquity of AI increases in our daily lives and parlance, an aspect of this transformation which has received scant attention is the cost associated with it all. While AI took the computing world decades to develop, its widespread deployment was not impeded by technological obstacles alone. Instead, a very significant aspect of why it took so long to publicly deploy powerful generative AI systems is the humongous cost associated with training and deploying them.

OpenAI’s GPT-4, Google’s PaLM (Pathways Language Model), and even the relatively smaller LLaMA model by Meta are expensive to train and deploy. And these costs are multi-dimensional: No easy fixes can bring it down dramatically in the short-term. Given the role AI is going to play in our lives it is important to explore the nature of these costs and the direction of future AI development.

Story continues below Advertisement

Expensive To Train

The first major cost component to consider is hardware. This, most significantly, includes GPUs (Graphics Cards), essential for training and deploying AI models. Just one of Nvidia’s A100 cards, which are some of the most common GPUs for such tasks, is priced north of $10,000. And given the scale of computation required, firms need clusters of tens of thousands of these GPUs to train and deploy their biggest models.