By 2027, organisations will use small, task-specific AI models 3x more than general-purpose LLMs: Gartner
Gartner, Inc. predicts that by 2027, organisations will implement small, task-specific AI models, with usage volume at least three times more than those of general-purpose large language models (LLMs).
While general-purpose LLMs provide robust language capabilities, their response accuracy declines for tasks requiring specific business domain context.
“The variety of tasks in business workflows and the need for greater accuracy are driving the shift towards specialised models fine-tuned on specific functions or domain data,” said Sumit Agarwal, VP Analyst at Gartner. “These smaller, task-specific models provide quicker responses and use less computational power, reducing operational and maintenance costs.”
Enterprises can customise LLMs for specific tasks by employing retrieval-augmented generation (RAG) or fine-tuning techniques to create specialised models. In this process, enterprise data becomes a key differentiator, necessitating data preparation, quality checks, versioning and overall management to ensure relevant data is structured to meet the fine-tuning requirements.
“As enterprises increasingly recognise the value of their private data and insights derived from their specialised processes, they are likely to begin monetising their models and offering access to these resources to a broader audience, including their customers and even competitors,” said Agarwal. “This marks a shift from a protective approach to a more open and collaborative use of data and knowledge.”
By commercialising their proprietary models, enterprises can create new revenue streams while simultaneously fostering a more interconnected ecosystem.
Implementing small task-specific AI models
Enterprises looking to implement small task-specific AI models must consider the following recommendations:
- Pilot contextualised models: Implement small, contextualised models in areas where business context is crucial or where LLMs have not met response quality or speed expectations.
- Adopt composite approaches: Identify use cases where single model orchestration falls short, and instead, employ a composite approach involving multiple models and workflow steps.
- Strengthen data and skills: Prioritise data preparation efforts to collect, curate and organise the data necessary for fine-tuning language models. Simultaneously, invest in upskilling personnel across technical and functional groups such as AI and data architects, data scientists, AI and data engineers, risk and compliance teams, procurement teams and business SMEs, to effectively drive these initiatives.