December 3, 2024
Editors' notes
This article has been reviewed according to Science X's editorial process and policies. Editors have highlighted the following attributes while ensuring the content's credibility:
fact-checked
reputable news agency
proofread
Why tech firms are aiming for smaller, leaner AI models
AI firms have long boasted about the enormous size and capabilities of their products, but they are increasingly looking at leaner, smaller models that they say will save on energy and cost.
Programs like ChatGPT are underpinned by algorithms known as "large language models", and the chatbot's creator bragged last year that its GPT-4 model had nearly two trillion "parameters"—the building blocks of the models.
The vast size of GPT-4 allows ChatGPT to handle queries about anything from astrophysics to zoology.
But if a company needs a program with knowledge only of, say, tigers, the algorithm can be much smaller.
"You don't need to know the terms of the Treaty of Versailles to answer a question about a particular element of engineering," said Laurent Felix of Ekimetrics, a firm that advises companies on AI and sustainability.
Google, Microsoft, Meta and OpenAI have all started offering smaller models.
Amazon too allows for all sizes of models on its cloud platform.
Kara Hurst, Amazon's chief sustainability officer, said at a recent event in Paris that it showed the tech industry was moving towards "sobriety and frugality".
Energy needs
Smaller models are better for simple tasks like summarizing and indexing documents or searching an internal database.
US pharmaceutical company Merck, for example, is developing a model with Boston Consulting Group (BCG) to understand the impact of certain diseases on genes.
"It will be a very small model, between a few hundred million and a few billion parameters," said Nicolas de Bellefonds, head of AI at BCG.
Laurent Daudet, head of French AI startup LightOn, which specializes in smaller models, said they had several advantages over their larger siblings.
They were often faster and able to "respond to more queries and more users simultaneously", he said.
He also pointed out that they were less energy hungry—the potential climate impact being one of the major concerns over AI.
Huge arrays of servers are needed to "train" the AI programs and then to process queries.
These servers—made up of highly advanced chips—require vast amounts of electricity both to fuel their operation and to cool them down.
Daudet explained that the smaller models needed far fewer chips, making them cheaper and more energy efficient.
Multi-model future
Other proponents point out that they can run without using data centers altogether by being installed directly on devices.
"This is one of the ways to reduce the carbon footprint of our models," Arthur Mensch, head of French start-up Mistral AI, told the Liberation newspaper in October.
Laurent Felix pointed out that direct use on a device also meant more "security and confidentiality of data".
The programs could potentially be trained on proprietary data without fear of it being compromised.
The larger programs, though, still have the edge for solving complex problems and accessing wide ranges of data.
De Bellefonds said the future was likely to involve both kinds of models talking to each other.
"There will be a small model that will understand the question and send this information to several models of different sizes depending on the complexity of the question," he said.
"Otherwise, we will have solutions that are either too expensive, too slow, or both."
© 2024 AFP
Citation: Why tech firms are aiming for smaller, leaner AI models (2024, December 3) retrieved 3 December 2024 from https://techxplore.com/news/2024-12-tech-firms-aiming-smaller-leaner.html This document is subject to copyright. Apart from any fair dealing for the purpose of private study or research, no part may be reproduced without the written permission. The content is provided for information purposes only.
Explore further
French startup Mistral AI vows to maintain open source 0 shares
Feedback to editors