Who needs 50 billion parameters in a large language model (LLM), when 3 billion or so will do just fine?
Market intelligence firm Gartner sees security benefits in “small language models”—computational machine learners with fewer than 10 billion parameters, or training variables.
“You don’t need your language model to write a poem about cats and dogs eating spaghetti under a bridge. You need it to answer an HR-related question,” Birgi Tamersoy, Gartner’s senior director analyst for AI technologies, said in a live presentation on September 12.
Known, domain-specific data—that HR-related info, for example—can be embedded into a small language model to solve a specific task, Tamersoy said.
A July 2024 survey from another market intel firm, IDC, found that 20% of IT pro respondents said they “don’t expect to use small models”; 25% have deployed them; and 26% characterize their current use of small models as “learning.” (A further 17% said “evaluating,” and 13% said “testing.”)
Read the rest here.—BH
|