Start receiving our RegBrief straight to your inbox!


Subscribe

Back
Article

ECB’s Revised Guide to Internal Models: The Use of Machine Learning Techniques in Internal Models

Hanwen Yang
Senior Consultant • Risk Advisory for Banking

Introduction

The ECB has published an updated version of its "Guide to internal models" with guidance for financial institutions on the use of internal models for calculating regulatory capital requirements across various risk types: credit risk, market risk, and counterparty credit risk.

The revised guide also includes a new section on machine learning which specifies expectations for using machine learning (ML) techniques in internal models. In the past, machine learning  models were often seen as “black boxes”  which left regulators and banks uncertain about how to approach them. Recent ECB guidelines have changed that, providing much clearer guidance and expectations around the use of ML in internal models. With this added clarity, banks can move forward with more confidence in using ML methods - especially given that these models usually offer higher accuracy and stronger performance than traditional methods.

The scope of the rules extends to all internal models used for Pillar 1 purposes, namely those for credit risk, market risk, and counterparty credit risk. The ECB explicitly states that the use of ML is considered an overarching topic applicable across all these model types, regardless of the underlying risk category. While the guidance was primarily developed from the standpoint of credit risk models, it has been supplemented and adapted where possible to include market and counterparty credit risk models. Therefore, institutions using ML within any Pillar 1 internal model - whether as a main modelling approach, for data preparation, or as a supporting component - are subject to the same overarching principles and expectations.

However, due to the complexity of ML techniques, institutions are also facing some challenges to ensure the compliance. This blogpost summarizes the key requirements from this new section on machine learning, discussing the challenges and providing recommendations to the institutions.

The table below summarizes the key takeaways we’d like to highlight from the revised guide on the ML topics:

TopicGuidanceRequired Action
General PrincipalClear regulatory requirements and expectations for models using ML techniquesRefine governance framework and process to comply with regulatory guidance
GovernanceAll stakeholders need to possess necessary skills, capabilities and expertiseArrange regular training to the stakeholders and ensure the required knowledge on ML techniques
Model Complexity and ExplainabilityEnsure the model explainability, appropriate complexity, transparent and sufficient documentation, avoiding black-box approachesSelect the appropriate model approach, provide sufficient model justification and documentation, ensure the model explainability
Internal Validation and AuditClear and strict requirements on internal validation and audit in terms of ML techniquesRefine the internal validation and audit procedure to cater for the requirements on ML-based model validation and audit

ECB Guide Key Update

Definition

The ECB defines ML techniques as:

  • Models which rely on a large number of parameters, can represent highly non-linear functions and/or may require large amounts of data for their calibration (training). 
  • Techniques that are well adapted to process unstructured data, such as text. 
  • Techniques that are characterised as being highly complex and presenting limited explainability / interpretability.
     

Governance

  • Governance frameworks and processes. Institutions should reflect the risks stemming from the use of ML-based internal models in their governance frameworks and processes. 
  • Skills, capabilities and expertise.All the stakeholders (i.e. all three lines of defence, including compliance and the risk control unit as the second line and internal audit as the third line) should possess the necessary skills, capabilities and expertise.
  • Model change. The initial switch from a largely traditional approach to a mostly ML-based approach should be classified as a material change. Where using dynamic ML-based internal models or components, institutions should justify the reason and effectively monitor the evolution at an appropriate frequency. 
  • Model assessment. The use of ML techniques should be viewed as a driver in their complexity assessment and materiality assessment. ML-based internal models are more likely to be classified as material, with subsequently higher expectations towards management reporting and internal validation.

Data Governance and Maintenance

  • Data assessment standards. Institutions should define standards for assessing the adequacy of the data types, amounts and sources used as inputs  for ML-based internal models. 
  • Additional data analysis. Input data for ML techniques should undergo an additional exploratory analysis to ensure a clear understanding of data formats, the treatment of missing values, and potential sources of bias.

Model Development

  • Methodology justification. The methodology should be adequately justified and, where appropriate, based on statistical/optimality criteria.
  • Potential bias consideration. Due to the ability of ML algorithms to fit the training data sample beyond their generalisation capabilities, the distribution of the prediction of an ML technique on the training data sample will usually differ from the distribution based on the overall population. Institutions should consider the potential bias on the estimates stemming from over- and under-fitting and choose an appropriate level of model complexity.
  • Hyperparameters identification. Institutions should identify all relevant hyperparameters of their ML components. Unless otherwise justified, the determination of their values should be based on the model’s ability to generalise. 
  • Documentation for replicability. The documentation of ML components should allow for their replication, including a determination of the parameters and hyperparameters. 

Complexity and Explainability

  • Appropriate complexity. Institutions need to consider any trade-off between the complexity of the modelling techniques and the performance increase of the ML-based internal models.
  • Justification of risk drivers. The inclusion of a large number of risk drivers needs proper justification. Institutions need to ensure the economic rationale of ML-based internal models and their risk drivers.
  • Use of explainability techniques and tools. 
    • Functionality. Institutions are expected to rely on a set of explainability techniques and tools to support the plausibility and intuitiveness of their estimates. It should enable: 
      • a) an assessment and quantification of the contribution made by the individual risk drivers to the model prediction on a global level (i.e. across all predictions), and the relationship between the output of the model and each risk driver; 
      • b) the contribution of the individual key risk drivers to each given prediction.
    • Documentation. Institutions should define and document their use of the explainability techniques and tools, including the dimensions and criteria, their weaknesses and limits
    • Results. The explanation results should be robust, accurate and actionable, so as to permit informed human judgement and well-founded human oversight of the pre-overrides model outputs.

Model Use

  • Scope and purpose. Institutions should explicitly define the scope and purpose of the use of ML-based internal models in their risk management and decision-making processes and in their credit approval, internal capital allocation and corporate governance functions. 
  • Documentation on exceptions and discrepancies. Where institutions use internal ratings or risk estimates resulting from ML-based internal models, if (part of) ML-based model is not used in specific areas, they should provide adequate reasons in the documentation for this choice, thus ensuring that any discrepancies are explained in a sound and understandable manner.
  • Use in ICAAP. When an institution uses ML-based internal models in ICAAP, it should ensure that model outcomes in stress tests and scenario analyses are explainable, plausible and not overly optimistic.

Human Judgement

  • Indicator of model appropriateness. When observing excessive numbers/extents of overrides and/or an accumulation of inappropriate justifications, it should be considered as a strong indicator questioning the design and assumptions of the ML-based internal model, especially with regard to its complexity.
  • Override monitoring. Institutions should implement the monitoring of overrides, including an assessment of the impact of the overrides on the rating model performance. This monitoring should distinguish between input and output overrides.
  • Override documentation. Institutions must document and store sufficient information on all the overrides they conduct.
  • Use of explainability techniques and tools. Institutions should consider the relevant explainability techniques and tools before conducting an override. They need to document the explainability techniques and tools considered for each rating override.

Internal Validation

  • Requirements. The internal validation should:
    • challenge the modelling decisions taken regarding ML-based internal models; 
    • assess whether the model complexity is justified; 
    • evaluate the stability and robustness of the outcomes, the hyperparameters selected and the randomness of the training process;
    • assess if the ML techniques fit the training dataset beyond its generalisation capacities (including overfitting); and 
    • assess the model performance using out-of-sample and out-of-time data. 
  • Frequency. The validation frequency should be commensurate with the ML-based internal model’s complexity and model risk, to ensure that any material deficiency is detected within a reasonable time period. 
  • Use of explainability techniques and tools. The internal checks are expected to rely on a set of explainability techniques and tools, to reliably identify:
    • circumstances (including instances) where an ML-based internal model is not working effectively;
    • the causes for a deterioration in the model’s performance and stability;
    • the reasons for the deviation in the risk estimates from their realised values. 
  • Assessment of explainability techniques and tools. The explainability techniques and tools used in conjunction with the ML-based internal models should also be assessed. These analyses should be performed at least yearly, addressing in particular the adequacy and appropriateness of the chosen explainability techniques and tools.

Internal Audit

  • Higher model risk consideration. When performing the general risk assessment and drawing up the work plan, the institution should consider whether the use of ML-based internal model leads to higher model risk. 
  • Audit engagements. Internal audit should set an appropriate intensity and frequency for the audit engagements of ML-based internal models. 

Challenges and Recommendations to the Institutions

Institutions are now facing a strict and comprehensive regulation for ML-based models. To comply with the regulatory expectations, here are some key challenges and the corresponding recommended actions:

TopicChallengesRecommendations
Skills, capabilities and expertise of all stakeholdersConsidering the complexity of ML techniques, it’s hard to ensure all the stakeholders have sufficient knowledge and experience to handle ML-based models.Provide regular trainings to different stakeholders, depending on their roles and responsibilities.
Data availability, quality and representativenessAs ML models usually involve large amounts of data from various sources, when assessing data quality, it might not be sufficient to apply the same criteria used for traditional data. Data may not always be available or representative for its application, as sources, formats and quality vary over time.Enhance the data governance and assessment framework to ensure data quality and applicability.
Complexity, reliability and interpretability of ML model resultsThe complexity of ML models leads to challenges in explaining the results. Complex models such as deep learning may achieve high accuracy but their “black box” nature makes it difficult to interpret outcomes and replicate logic.Define clear assessment criteria in model development, approval, monitoring, validation and audit. Ensure model explainability and replicability.
Overfitting and instability of model performanceOverfitting is a common issue for ML models. When training data is biased or limited, model performance becomes unstable and less generalizable.Implement robust model validation procedures, including stress testing, cross-validation, and regular performance monitoring to mitigate overfitting risks.
Documentation on the underlying assumptions and theoriesML-based models often rely on complex mathematical methodologies and several key assumptions. Due to the complexity, documenting all assumptions and theories in a way that allows replication can be challenging.Establish a comprehensive documentation framework that clearly explains model logic, data assumptions, and methodological choices.
Development and use of explainability techniques and toolsRegulators expect institutions to use explainability techniques and tools across model development, validation, and human judgment. However, developing and adopting these tools takes time, and ensuring proper use by all stakeholders is challenging.Adopt a dual-layer approach combining global feature importance and local decision explanations (e.g., SHAP, LIME). Provide training to ensure proper usage.
Model risk and assessment of model changeML models tend to have higher model risk, requiring more robust governance and monitoring. They also bring challenges in classifying and assessing the materiality of model changes.Conduct frequent validation and audits. If significant issues arise, such as performance deterioration or lack of explainability, initiate remedial actions like redevelopment or reversion to traditional models.

Conclusion

To summarize, the ECB’s revised guide to internal models provides a comprehensive guidance for internal models using ML techniques. Given the complexity of machine learning, the institutions are facing challenges in terms of model governance, documentation, explainability, and model use. To comply with the regulatory expectations, institutions need to take immediate actions to refine and enhance the model management framework when using ML techniques in internal models.

How Finalyse can help?

At Finalyse, our goal is to support banks and financial institutions in managing the challenges around ML models and regulatory expectations. Our team has significant expertise in both regulatory and non-regulatory models with deep knowledge of supervisory requirements.

What we offer to you:

  • Experienced professionals – our consultants have years of practical and theoretical experience in applying ML models in different environments.
  • Clear guidance on governance and transparency – we help you meet supervisory requirements with confidence.
  • Training and workshops – we equip your teams with the skills to understand, interpret, and apply ML models responsibly in line with regulatory expectations.

Our expertise includes:

  • Optimising model development through our in-house ML framework.
  • Addressing interpretability challenges to maintain transparent and explainable models.
  • Establish clear assessment criteria for model approach selection, providing the support through the full model cycle (e.g. model development, monitoring, validation, etc.)
  • Sharing know-how through practical methods and real-life case studies

If you would like to talk about these points, we would be glad to discuss how we can support your institution.
 

AI Summary Prompt: The ECB’s updated *Guide to Internal Models* provides comprehensive guidance for financial institutions on using internal models to calculate regulatory capital for credit, market, and counterparty credit risks, now including a new section dedicated to machine learning (ML). This update clarifies supervisory expectations for ML use, moving beyond the “black box” perception and allowing banks to adopt these techniques with greater confidence due to their enhanced accuracy and performance. The guidance applies across all Pillar 1 internal models, whether ML is used as the primary modelling method, for data preparation, or as a supporting tool. While the framework primarily draws on credit risk experience, it extends to market and counterparty credit risk models. Given ML’s complexity, institutions face new compliance challenges — the ECB’s new guidelines and the accompanying analysis aim to help address these through practical recommendations.

Frequently Asked Questions

The ECB’s revised guide helps banks navigate this challenge by setting clear expectations for transparency, governance, and compliance. It shows how ML can be applied responsibly in credit, market, and counterparty credit risk models, turning complex algorithms into tools that decision-makers can trust.

ML is powerful but very complex. Banks need to make sure their teams have the right skills, datasets, and that model outputs are understandable. In addition, challenges such as overfitting, rigorous documentation, and stringent audit requirements introduce further layers of complexity that institutions need to manage carefully.

A structured approach is essential. Banks should strengthen governance, provide their teams with the right expertise and support, ensure data quality, and use suitable explainability tools to interpret model outputs effectively.

Share this article:

Finalyse Insurance

Finalyse offers specialized consulting for insurance and pension sectors, focusing on risk management, actuarial modeling, and regulatory compliance. Their services include Solvency II support, IFRS 17 implementation, and climate risk assessments, ensuring robust frameworks and regulatory alignment for institutions.

Finalyse Banking

Finalyse leverages 35+ years of banking expertise to guide you through regulatory challenges with tailored risk solutions.
Our multidisciplinary team masters both technical and regulatory aspects, helping you adopt best practices and boost performance.

Finalyse Valuation

Valuing complex products is both costly and demanding, requiring quality data, advanced models, and expert support. Finalyse Valuation Services are tailored to client needs, ensuring transparency and ongoing collaboration. Our experts analyse and reconcile counterparty prices to explain and document any differences.

Finalyse Publications

Discover Finalyse writings, written for you by our experienced consultants, read whitepapers, our RegBrief and blog articles to stay ahead of the trends in the Banking, Insurance and Managed Services world

About Finalyse

Our aim is to support our clients incorporating changes and innovations in valuation, risk and compliance. We share the ambition to contribute to a sustainable and resilient financial system. Facing these extraordinary challenges is what drives us every day.

Finalyse Careers

Unlock your potential with Finalyse: as risk management pioneers with over 35 years of experience, we provide advisory services and empower clients in making informed decisions. Our mission is to support them in adapting to changes and innovations, contributing to a sustainable and resilient financial system.