That has led tech groups like Meta and Google to pitch small language models with just a few billion parameters as cheaper, energy-efficient, customisable alternatives that require less power to train and run, which can also ringfence sensitive data.
“By having this much quality at a lower cost point, you actually enable so many more applications for customers to go in and do things that . . . there wasn’t enough return on that investment for them to justify really doing it,” said Eric Boyd, corporate vice-president of Microsoft’s Azure AI Platform, which sells AI models to businesses.
Google, Meta, Microsoft and French start-up Mistral have also released small language models that show advancing capabilities and can be better focused on specific applications.
Nick Clegg, Meta’s president of global affairs, said Llama 3′s new 8bn parameter model was comparable to GPT-4. “I think on pretty much every measurement you could think of, you see superior performance,” he said. Microsoft said its Phi-3-small model, with 7bn parameters, outperformed GPT-3.5, an earlier version of OpenAI’s model.
The small models can process tasks locally on a device, rather than send information to the cloud, which could appeal to privacy-conscious customers who want to ensure information is kept within internal networks.
Charlotte Marshall, a managing associate at Addleshaw Goddard, a law firm that advises banks, said “one of the challenges I think a lot of our clients have had” in adopting generative AI products was adhering to regulatory requirements over handling and transferring data. She said smaller models provided “an opportunity for businesses to overcome” legal and cost concerns.
Smaller models also allow AI features to run on devices such as mobile phones. Google’s “Gemini Nano” model is embedded inside its latest Pixel phone and Samsung’s latest S24 smartphone.
Apple has hinted that it is also developing AI models to run on its bestselling iPhone. Last month, the Silicon Valley giant released its OpenELM model, a small model designed to perform text-based tasks.
Microsoft’s Boyd said smaller models would lead to “interesting applications, all the way down into phones and into laptops”.
OpenAI chief Sam Altman said in November that the San Francisco-based start-up offered different-sized AI models to customers that “serve separate purposes”, and it would continue to build and sell these options.
“There are some things where smaller models will work really well,” he added. “I’m excited for that.”
However, Altman added OpenAI would remain focused on building larger AI models with scaled-up capabilities, including the ability to reason, plan and execute tasks and eventually achieve human-level intelligence.
“There are a lot of times where I think people just want the best model,” he said.
Written by: Cristina Criddle and Madhumita Murgia in London. Additional reporting by George Hammond in San Francisco
© Financial Times