A fresh take on risk and valuation

Start receiving our RegBrief straight to your inbox!


Subscribe

Back
Article

Machine learning for credit risk IRB models

Written by Barend Janse Van Rensburg, Consultant 

Abstract

Machine learning has shown a capacity for considerable improvements in several business areas, and its usage in the financial industry will continue to be explored. However, In the credit risk sector, financial institutions have been hesitant to implement ML into banks’ IRB models to forecast regulatory capital requirements. 

This article examines the challenges and potential opportunities presented by machine learning when used to compute regulatory capital for credit risk using internal ratings-based (IRB) models. It also gives an overview of machine learning techniques that can be utilised for IRB modelling, as well as the results that can be expected. 

Regulatory background

In 1988 the Basel Committee on Banking Supervision (BCBS) published the so-called Basel I covering capital requirements for credit risk and appropriate risk-weighting of assets. Basel II, the new capital framework, superseded the Basel I framework in 2004, introducing three main pillars:

-       Minimum capital requirements, which concern the maintenance of regulatory capital estimated for credit risk, operational risk, and market risk;

-       Internal and supervisory assessments of an institution's capital sufficiency; and

-       Market discipline and sensible banking practices can both benefit from increased transparency.

Basel II provided banks with three options for determining regulatory capital requirements for credit risk, namely (i) the standardised approach, (ii) the Foundation Internal Rating-Based Approach (F-IRB), and (iii) Advanced IRB (A-IRB). These options remained essentially the same with some minor adjustments in 2010s Basel III and 2017s Basel IV. However, since then, banks have not used materially different IRB modelling methodologies to estimate credit risk. Regulators and supervisors concentrated on strengthening the definition of core concepts to ensure the comparability of the estimates given by different models. This meant that the significance and challenges of advanced technologies such as machine learning (ML) and artificial intelligence (AI) were underappreciated

Machine learning in "Big Data era"

The Report on BD&AA, released by the EBA in January 2020, stated that advanced analytics include ‘predictive and prescriptive analytical techniques, often using AI and ML in particular, and are used to understand and recommend actions based on the analysis of high volumes of data from multiple sources, internal or external to the institution’.

Machine Learning (ML) can be defined as a method of teaching computers to parse, learn from, and then make a decision or prediction regarding new data. Machine Learning overlaps with statistical learning as both attempt to find and learn from patterns and trends within large datasets to make predictions.

 

Machine learning encompasses a broad array of models of varying degrees of complexity. Linear regressions, straightforward models with a minimal number of parameters, are an example of simple ML models. Deep neural networks are an example of the opposite extreme of the complexity scale. The number of factors in these models might reach millions, rendering understanding and application difficult. However, the term Machine Learning is primarily used to denote the more complicated models. This has long been the case in the financial sector, where linear and logistic regressions have been frequently applied, but the term machine learning is reserved for the more innovative models.

Depending on the purpose of the model and the type of data necessary, a variety of learning paradigms can be employed to train ML models. The three most prevalent learning paradigms are currently:

-       Supervised learning: the algorithm learns model-building rules from a labelled dataset and applies these rules to predict labels on incoming input data.

-       Unsupervised learning: the algorithm learns from an unlabelled input training dataset with the purpose of understanding the data distribution and/or finding a more efficient representation of it.

-       Reinforcement learning: instead of learning from a training dataset, the algorithm learns by interacting with the environment. In addition, unlike supervised learning, reinforcement learning does not need the use of labelled input/output pairs. By trial and error, the algorithm learns to do a certain task.

The adoption of Machine Learning driven IRB models

To date, the adoption of ML-driven IRB models has been unenthusiastic – or non-existent in the case of many European banks. According to the IIF 2019 report, credit decisions/pricing is the most popular application of machine learning in credit risk, followed by credit monitoring and collections, restructuring and recovery. Machine Learning is all but avoided in regulatory domains like credit risk capital requirements, stress testing, and provisioning. Regulatory restrictions are seen as a hurdle for the deployment of ML models in this area, as they are more difficult to analyse and explain.

The usage of ML in IRB modelling has been limited to supplementing the traditional models for calculating capital requirements. Examples of where machine learning techniques are currently applied in IRB models to meet CRR requirements are:

-       Model validation: Machine learning models are used to create challenger models that serve as a comparison to the standard model for calculating capital requirements.

-       Data quality: Machine learning techniques can be used to improve data quality and can be used to analyse rich datasets.

-       Variable selection: ML could be used to find explanatory factors and combinations in a large dataset with good prediction capacities.

-       Risk differentiation: ML models can be used as a module for risk differentiation in the probability of default (PD) model.

However, the following section outlines the challenges for the ML models to be utilised as primary IRB models for prediction (e.g., risk differentiation).

Challenges of using Machine Learning models

The major difficulty of ML models is their complexity, which leads to challenges in:

  1. Interpreting their results.
  2. Assuring proper knowledge by management functions.
  3. Proving their results to supervisors.

ML models may provide issues or offer solutions unique to the context in which they are deployed. The following are the key areas of application: 

- Risk differentiation;

- Risk quantification;

- Model validation; 

- Other fields, such as data preparation or the application of machine learning models for credit risk mitigation.

The intricacy and interpretability of some ML models may represent extra hurdles for institutions developing compliant IRB models, depending on the context of their use. However, this does not rule out the possibility of using these models.

Machine learning models for risk differentiation

ML models for risk differentiation face several unique issues, including those linked to the following CRR requirements:

Requirement

Challenges

The definition and assignment criteria to grades or pools

  • If complex ML models are used as the primary risk differentiation models, it may be difficult to analyse.
  • This may limit the application of models where there is no apparent economic link between the input and output variables.
  • It may be difficult to find the requisite clear economic theory and assumptions for the model.
  • Institutions should seek out appropriate tools for interpreting these sophisticated machine learning models.

Complementing human judgement

  • It necessitates model comprehension, which may limit the use of difficult-to-understand machine learning models.
  • The intricacy of machine learning may provide unique obstacles in terms of human judgment. When an institution employs statistical models to assign grades or pools, human judgment should be used.
  • The complexity of ML models may make evaluating the modelling assumptions and determining relevant risk drivers more difficult.

Documentation

  • Because some ML models are sophisticated and complex, providing a clear overview of the theory, assumptions, and mathematical basis for the final assignment of estimates might be difficult.
  • In order to document the model's flaws, the institution's relevant staff must have a thorough understanding of the model's capabilities and limitations.

 

Machine learning models for risk quantification

Using the ML model for risk quantification has its own set of issues, including those linked to the following CRR requirements:

Requirement

Challenges

Concerning the estimation process, the plausibility and intuitiveness of the estimates

  • Machine learning models can produce unexpected results, especially when the model's structure is difficult to understand.
  • It can also be difficult to make correct judgement assessments when combining results from diverse approaches and adjusting for various types of restrictions.

Length of the underlying historical observation

  • This could be a problem when using huge data or unstructured data, which may not be available for a long enough period of time.
  • Furthermore, data retention restrictions connected to the General Data Protection Regulation (GDPR) may make it more difficult for natural persons to achieve the required five-year period of the underlying historical observation.

Validation of Machine Learning models

Validation of ML models may provide some unique issues. There are two categories of challenges:

  1. Difficulties comprehending / addressing the validation findings.
  2. Challenges with the validation tasks themselves.

Requirement

Challenges

Interpreting/resolving the findings

  • Machine learning models may make resolving detected flaws more difficult.
  • Internal estimations may be more difficult to validate especially when institutions has to explain any significant disparity between actual default rates and the projected range of variability of PD estimations for each grade.
  • Due to difficulty determining the effect of the economic cycle on the logic of the model, comparing the projections of the models to observed default rates can be more complicated.

Validation of the core model performance

  • Assessing representativeness and meeting more operational data requirements may be more complex (e.g. data quality or data storage and maintenance).
  • Furthermore, extra caution should be exercised while assessing the quality of data supplied.
  • Due to the high danger of overfitting, special attention must be paid to the use of out-of-sample and out-of-time samples in the evaluation of model results.
  • The validation function should examine and test the model's design, assumptions, and approach. As a result, a more complicated model will be more difficult to dispute effectively.
  • Validation may require more statistical understanding, thus institutions should guarantee that the validation function's personnel are properly trained.

Other challenges related to the use of a Machine Learning models

Requirement

Challenges

Corporate Governance

  • The institution's management body is still in charge of adhering to legal obligations. The adoption of machine learning models makes this more difficult, but it does not absolve these parties of their responsibilities.
  • These individuals must have a broad awareness of the institution's grading systems as well as a thorough understanding of the institution's management reports.

The implementation process

  • The complicated requirements of ML models can jeopardize soundness and integrity of implementation process.
  • The complexities of ML models, in particular, may make verifying the correct implementation of internal ratings and risk parameters in IT systems more difficult.

Categorisation of model changes

  • This may be difficult for models with time-varying weights connected with variables that are updated frequently.
  • Validation of models may also be problematic.
  • It may be more difficult to detect biases.

The use of big and unstructured data

  • Implementing a method for evaluating data inputs into the model that assures the data's accuracy, completeness, and appropriateness may pose a challenge.
  • It may be difficult to ensure that the data used to develop the model is representative of the application portfolio.

Benefits from the use of Machine Learning models

With some effort, machine learning models could improve the IRB models whilst meeting all prudential standards. The following are some examples of areas where machine learning models could be useful:

- Improving risk differentiation through improving model discriminatory power and offering valuable tools for identifying all key risk drivers and their relationships. ML models could also be used to improve portfolio segmentation, construct robust models across geographical and industrial sectors/products, and make data-driven decisions that balance data availability with model granularity requirements. Furthermore, machine learning models may aid in the confirmation of data features chosen by expert judgment in 'conventional' model creation, providing a data-driven viewpoint on the feature selection process.

- Improving risk quantification by increasing model predictive power and recognising material biases, as well as offering tools for identifying recovery patterns in LGD models. ML models could potentially aid in the calculation of the necessary modifications.

- Improving data collecting and preparation processes, such as input data cleansing, data treatment and data quality checks. Through unsupervised learning approaches, ML models could be effective for analysing representativeness. Furthermore, machine learning models could be used to discover outliers and correct errors. In addition, unstructured data (e.g. qualitative data like business reports) could be employed with ML models, expanding the data sets that can be used for parameter estimation.

- Improving credit risk mitigation approaches, including the use of machine learning algorithms for collateral valuation through haircut models.

- Providing reliable systems for model validation and monitoring where model challengers or a supporting analysis for alternative assumptions or techniques could be generated using machine learning models.

How Can Machine Learning be Properly Leveraged?

Since machine learning models are more complex than traditional techniques like regression analysis or simple decision trees, current risk management procedures and governance frameworks may require further enhancements.

Most concerns relate to the ML models' complexity and reliability, with the main decisive challenges appearing to be the interpretability of the results, governance, with a particular focus on increased staff training needs, and the difficulty in evaluating a model's generalisation capacity (i.e. avoiding overfitting).

To overcome these challenges, several techniques have been developed that provide insight into a model's fundamental logic. The following are some of the most commonly utilised techniques:

  1. Visualisation tools that show how an explanatory variable affects the model. 
  2. Feature importance metrics reflect the overall model's relevance of each explanatory variable.
  3. Shapley values measure the influence of each explanatory variable on a specific model prediction.
  4. Simple approximations of the model in the vicinity of an observation are provided by local explanations such as LIME and anchors.
  5. Counterfactual explanations show how a specific model prediction could be changed with as few changes to the explanatory variables as possible. 

ML models may be useful if they provide adequate monitoring, validation, and explanation of the process and model results. Strong institutional awareness of their IRB models is critical, and this is especially true when machine learning models are utilised for regulatory reasons. ML can be applied to a variety of tasks and levels, including data preparation, risk differentiation, risk quantification, and internal validation. 

All of the following criteria apply when ML models are to be used for risk differentiation and risk quantification.

Appropriate level of knowledge of the model’s functioning

If an institution seeks to employ machine learning models for regulatory capital, all relevant stakeholders should have a sufficient understanding of how the model works. The EBA recommends that institutions make certain that:

  1. Staff in the model development unit, credit risk control unit (CRCU), and validation unit are sufficiently skilled to design and evaluate ML models.
  2. By providing suitable high-level documentation so that the management body and senior management are in a position to have a thorough knowledge of the model.

Avoid unnecessary complexity

Furthermore, banks should avoid unnecessary complexity in the modelling approach. Institutions should refrain from:

  1. incorporating an excessive number of explanatory drivers or drivers with little predictive value;
  2. utilizing unstructured data if more traditional data with equivalent predictive capabilities is available; and
  3. excessively complicated modelling decisions when simpler alternatives with equivalent findings are available.

Understanding and interpreting the model

Institutions should also do the following to ensure that the model is accurately translated and understood:

  1. Analyse the relationship of each individual risk driver with the output variable, and analyse the total weight of each risk driver to determine what risk drivers impact model prediction most.
  2. Examine each risk driver's economic relationship to the output variable to confirm that the model's estimates are reasonable and intuitive.
  3. Create a summary document explaining the model in simple terms using the analysis results stated in point a. All key stakeholders, including workers who utilise the model for internal purposes, should be addressed in the document. The paper should at the very least describe: 
  4. The model's major risk drivers 
  5. The key connections between risk factors and model projections.
  6. Check for any model biases (for example, overfitting to the training sample).

Human judgement

When machine learning techniques are combined with human judgment, a thorough grasp of the model is essential. Particularly the Staff in charge should be able to assess the modelling assumptions when human judgment is applied in the development of the model.

Regular updates

When Machine Learning models are updated frequently, the institution must thoroughly investigate and monitor the rationale for the frequent updates. Furthermore, institutions should always compare modifications to the most recent approved model to ensure that a high frequency of minor changes does not result in an unnoticed material change within a given period of time. As a result, institutions must assess the benefits of automatic and regular updates to their IRB model against the hazards.

Machine Learning model validation

A reliable validation is especially critical for complicated ML models with limited explainability or for often updated models. It is suggested that institutions pay attention to:

          i.         Overfitting issues: Overfitting the development sample is a common problem with machine learning models. As a result, financial institutions should pay close attention to comparing model performance evaluated within the development sample to model performance measured using out-of-sample and out-of-time samples. 

         ii.         Challenging the model design: The hyperparameters used to describe the model's structure and to tailor the learning algorithm are frequently based on human judgment. As a result, the validation unit should pay special attention to the rationale for selecting these hyperparameters.

       iii.         Representativeness and data quality issues: When large datasets are fed into machine learning techniques for risk differentiation and quantification, data quality must be ensured. When these data are external data, institutions should pay special attention to determining the external data's representativeness in relation to the application portfolio. When employing unstructured data, institutions should take extra precautions to ensure that the data is accurate, full, and suitable.

       iv.         Analysis of the stability of the estimates: It is useful to analyse both the stability: 

    • In the assignment process of each debtor/exposure to grades or pools. Indeed, machine learning algorithms may bring point-in-time (PIT) aspects into models, which may compromise the stability of the rating assignment process when compared to more through-the-cycle (TTC) models, potentially resulting in quick changes in capital requirements.
    • Of the relationship between the output variable and the drivers in subsequent releases of the model based on ML techniques to provide an assessment of whether changes between inputs and outputs require regulatory approval notification. 

Data preparation

It is advised that institutions ensure that there are clear rules and documentation. Institutions should use a comprehensive set of checks and controls to guarantee that the methodology used to analyse data is appropriate.

Parting Thoughts

Machine learning has a well-established place in the future of the banking and finance business, and it's believed that risk management will use machine learning approaches to improve their capabilities as well. Despite being criticised for acting like a black box, machine learning approaches have the potential to analyse large amounts of data without being bound by distribution assumptions and provide significant value in exploratory analysis, classification, and predictive analytics.

This may change the risk management landscape. Machine learning, which has been regarded as one of the technologies with significant risk management implications, can help risk managers construct more accurate risk models by recognising complicated, nonlinear patterns in vast datasets.
Given its size and scope, machine learning requires greater collaboration between the industry and regulators to guarantee that it protects clients without limiting innovation in the financial sector. Because of the risk-averse nature of the financial industry, pilot programs that investigate and test novel technologies should be encouraged by legislators and supervisors.

Share this article: