Over the past 18-months, large language models (LLMs) have evolved rapidly, offering businesses powerful tools for tasks like content generation, natural language understanding and automation. However, as organisations continue to explore the use of LLMs, they will need to navigate critical decisions about whether to train their own or instead use pre-trained models ‘out of the box’. This article delves into the strategic choices that organisations must make when implementing LLMs, in a process balancing flexibility, cost, performance and governance.
As LLMs evolve, so too do their capabilities, as well as their performance in different use case scenarios. This rapid development requires businesses to adopt a flexible, iterative approach to LLM deployment, one that emphasises agility and continual improvement. The key here is to embrace flexibility, adopting a test, deploy, iterate approach:
In these ways, by implementing a dynamic framework for LLM testing and deployment, institutions will be better able to stay at the forefront of AI advancement, allowing room for agile iteration and refinement where needed.
As organisations adopt LLMs, they will each face a crucial decision around whether to build and train their own custom models or use pre-trained models from providers. The decision itself hinges on the specific needs of the organisation, their available resources and their appetite for risk, since both approaches have their ‘pros and cons’ (see Fig 1 below).
Training Custom Models |
Using Pre-Trained Models |
---|---|
Offers highly specialised, domain-specific performance tailored to the organisation's unique requirements. |
Pre-trained models from a range of providers can be fine-tuned with minimal resources to suit specific tasks, offering instituions a cost effective and scalable solution. |
Training a model from scratch requires significant investment in data, computational resources and expertise. |
May lack the depth of customisation that specialised tasks demand. |
Source: Deloitte Experience
Assessing an organisation’s specific requirements, budget and domain expertise can help them decide whether to invest in training their own custom models or leverage pre-trained LLMs.
Once an organisation has fine-tuned an LLM with proprietary data, maintaining that model’s performance over time can present fresh challenges, especially as new versions of base models are released. Implementing a clear upgrade strategy for fine-tuned models can ensure they stay up-to-date and continue to deliver value to the organisation. As part of developing a sustainable Gen-AI upgrade path, consider the following:
As LLM adoption increases, so too will regulatory scrutiny. Ensuring transparency, accountability and explainability in how these models operate is essential to maintaining compliance with evolving regulation. In developing a compliance strategy for their LLMs, organisations should consider the following:
By implementing robust logging and monitoring systems for LLM usage, and by staying informed about regulatory developments, firms will be best equipped to ensure their LLMs maintain continuous compliance.
Evaluating LLMs requires more than accuracy. Given the complexity and variability of language tasks, organisations need to adopt comprehensive evaluation frameworks that also address robustness, fairness and reliability. This type of extended framework should include:
Adopting a multi-faceted approach to LLM evaluation, using both automated and human-centric methods, can help ensure models are robust, fair and aligned with the needs of the business.
While LLMs are powerful, they also have inherent limitations that can impact their reliability and effectiveness. Recognising and addressing these challenges is key to responsible LLM deployment. In particular, organisations need to be ready to tackle the following issues:
Similarly, even the most advanced LLMs have other limitations, such as context length constraints, high resource requirements for fine-tuning and challenges around interpretability. Understanding these limitations is essential for setting realistic expectations and ensuring responsible AI usage. For example:
Actively addressing the limitations of LLMs through prompt engineering, regular data updates, and bias mitigation can all help to support the reliability, fairness and currency of their outputs. By acknowledging and proactively addressing the limitations of LLMs, firms that implement regular audits and optimised processes will be best placed to maintain model quality and compliance.
In conclusion, choosing whether to train custom LLMs or use pre-trained models instead requires careful consideration of a number of context-specific factors. These include cost, performance, regulatory compliance and domain-specific needs. However, by adopting a flexible, iterative approach to LLM strategy, and by embracing the inherent challenges and limitations of these models, organisations can unlock the full potential of LLMs while mitigating the risks.
In our next and final article in this series we pull together the insights and recommendations from this and our other pieces to provide a simplified summary of the challenges and mitigations available to financial institutions as the proceed along the Gen-AI path.