This report explores current trends and considerations regarding how banks can effectively manage new and evolving operational risks arising from Artificial Intelligence (AI)I. Using the popular ORX taxonomy as a foundation, we assess how AI reshapes existing risk landscapes by introducing novel operational risk categories and complexities to integrate within existing risk frameworks. Our report emphasises the need for updated and enhanced governance and controls, informed and enforced by specialised expertise to effectively manage AI-driven operational risks while ensuring regulatory compliance.
Key Takeaways
Over the last few years, AI in various forms, has taken the world by storm. Banks are exploring and implementing different AI use cases - from “traditional” machine learning to the latest Generative and Agentic AI solutions to drive cost efficiencies and innovation within their organisations.
However, balancing the innovation agenda, whilst effectively managing risk is a key challenge for the management teams in banks. Despite the potential for creating undeniable benefits, we see that AI has already transformed the operational risk landscape, cutting across multiple risk types and dimensions including the areas of information security (including cyber), technology & data, model, conduct, compliance and third-party risk. AI both amplifies and transforms these well-known operational risks, as well as introducing new “AI native” risks that are novel to the technology. At the same time, AI provides new opportunities to enhance operational risk processes and reduce risk exposures such as through advanced analytics and greater precision in process execution.
To better understand the impact of AI on operational risk in banking, we have taken a structured approach for our assessment and leveraged the ORX (Operational Risk eXchange) Reference Taxonomy, an industry-standard for classifying operational and non-financial risksV. Exhibit 1 provides a non-exhaustive assessment of how AI could impact risk exposures across several of the ORX Level 1 (L1) Event Types.
With this changing risk landscape, it is not surprising that existing controls are not equipped to address the new challenges posed by the deployment and use of AI systems. This leads to control failures that look nothing like those risks which organisations have planned to mitigate for. In a Deloitte 2024 survey, 26% of executives revealed that their organisations had experienced deepfake incidents, while 50% of respondents expected a rise in attacks over the following 12 months [4].
Exhibit 2 provides a non-exhaustive gap assessment of how AI could stress traditional controls across several of the ORX Level 1 (L1) risk types.
As organisations become increasingly AI-driven, root causes of operational risks will shift from understanding discrete human or system failures, to deciphering complex, non-transparent algorithmic logic. Accountability will become more challenging due to the complexity of AI systems and intertwined causes of failure, especially considering how regulatory obligations are shared between providers and deployers of (high-risk) specific purpose AI systems. Furthermore, traditional governing processes will be stressed, as democratises powerful development capabilities to all employees. Exhibit 3 provides some examples of shifting root causes will demand a change in approach in operational risk management.
Effectively incorporating AI within operational risk frameworks will be an important step to managing new threats and maximising the benefits from new opportunities in a structured way. Based on current industry trends, we see that most banks are integrating AI risks as part of their existing operational risk taxonomies rather than treating them AI as a new, standalone risk type [10]. Regulators and industry bodies (e.g., ORX, Bank for International Settlements, European Banking Authority) have all stated that AI amplifies existing risks, rather than viewing AI as an entirely new risk category [11].
While the ORX taxonomy provides comprehensive coverage, AI introduces unique challenges that blur traditional risk category boundaries. This makes it difficult to map an AI risk scenario to a single ORX category, as identifying the root cause is not so straightforward.
There are also several 'AI native' risks that we believe do not directly map to traditional ORX Level 1 risk categories.
The ORX Reference Taxonomy has introduced an “AI/ML” flag for classifying AI-related risks [12] to address this challenge. The flag applies across both the Event Type and the Cause & Impact dimensions of the Taxonomy. This enables institutions to flag incidents involving AI, regardless of the primary event type classification. ORX’s flagging approach is a useful practical workaround that preserves the integrity of the existing taxonomy structure while introducing a way to track AI-specific risks. However, it's not a substitute for proactive governance adjustments. Exhibit 6 illustrates an example scenario where an ML solution that incorrectly classifies transactions used for regulatory reporting leads to financial penalties and reputational damage.
Operational AI risk management is a critical component of a larger AI Governance puzzle. The scale, impact, and complexity of AI-driven transformations, coupled with increasing regulatory pressures, demand organisations take a broad view to AI Governance. There are differing views on what constitutes AI Governance within financial services. In addition to risk management, AI Governance can include organisational structures, training and awareness, internal control systems (including model risk management), policies and guidelines, setting the tone at the top, and defining the organisation’s strategic approach to AI. In recent years, several new standards, frameworks, and a wealth of literature have been developed for risk management of AI. For example, ISO/IEC 42001 (AI Management System) and the NIST AI Risk Management Framework (RMF). Combined with existing frameworks and standards such as COSO (ERM and Internal Control) and ISO 31000 (Risk Management), these provide a comprehensive foundation for managing AI risks in a structured and consistent manner.
Exhibit 7 illustrates a conceptual framework for AI Governance, integrating key components from these standards and frameworks, while highlighting the broad-ranging themes and areas to consider when designing a detailed operating model. Human Oversight, AI Quality Management, and AI Risk Management complement one another to form a holistic AI Governance process, collectively addressing the underlying domains of People, Process and Technology.
To address the changing risk landscape driven by AI, organisations and management teams must evolve and adapt their approach to risk management.
We have identified the following practical “no-regret” recommendations organisations should consider as priority areas of investment. We have also mapped these recommendations to Exhibit 7. It is important to note that measuring ROI for AI initiatives can be challenging due to the pace of change, broad ranging business impacts, and longer-term benefit time horizon [13]. Leveraging traditional operational risk loss methodologies can provides a practical, quantitative approach to support investment decisions and strengthen business cases, especially within the risk and compliance domain.
Each organisation will have its own existing structures, its own technical and process debt, which will call for the right Operational Risk Management tailored to its individual needs. Just as the costs of quality rise, the later defects are discovered in the manufacturing process, even a small investment in any or all of these categories can yield great returns in terms of cost avoidance: costs of quality, costs of compliance, costs to reputation / client loyalty. The important point is not to leave things to chance, to apply the sound principles of enterprise risk management also to the field of AI. Doing so will strengthen the prospects of sustainable benefits from this transformational technology.
Special thanks to Sanjay Patel for his invaluable contributions to developing this article, building on his extensive expertise in AI Governance. His contribution, insight and leadership played a key role in completing the underlying analysis and making it relevant for businesses and other institutions.
Footnotes:
I. AI includes classic machine learning techniques, generative AI (particularly those based on large language models), and agentic AI.
II. Hallucination refers to the generation of outputs that are factually incorrect, nonsensical, or entirely fabricated.
III. Prompt injection is an attack that manipulates an AI model's input prompt to elicit unintended or malicious outputs by bypassing its safety mechanisms.
IV. Such losses can be calculated as the sum of risk severity multiplied by frequency across all relevant risk scenarios. In a manner similar to insurance loss modelling, which takes into account factors such as rework costs, client attrition, and regulatory fines, organisations can estimate plausible magnitudes and probabilities for each risk “peril” to conduct a robust cost-benefit analysis. This approach enables a clear connection between compliance-related investments and tangible outcomes.
V. The ORX Event Type Taxonomy consists of 16 “Type 1” category risk events, expanding on Basel operational risk event types, as well as adding conduct and compliance events.
VI. Adversarial inputs are carefully crafted data samples (often imperceptibly altered) that are designed to mislead a machine learning model into making incorrect predictions or classifications.
VII. Deepfake impersonation is the use of AI-generated synthetic media to mimic a person’s likness, voice, or behaviour in order to deceive others into believing it is real.
VIII. Data poisoning is the deliberate manipulation of training data to corrupt an AI model’s learning and cause it to behave incorrectly or maliciously.
Sources: