Operational Risk Management: Added Value of Advanced Methodologies



Similar documents
Contents. List of Figures. List of Tables. List of Examples. Preface to Volume IV

Tail-Dependence an Essential Factor for Correctly Measuring the Benefits of Diversification

Institute of Actuaries of India Subject CT3 Probability and Mathematical Statistics

STAT2400 STAT2400 STAT2400 STAT2400 STAT2400 STAT2400 STAT2400 STAT2400&3400 STAT2400&3400 STAT2400&3400 STAT2400&3400 STAT3400 STAT3400

Exam P - Total 23/

An Internal Model for Operational Risk Computation

Basel II: Operational Risk Implementation based on Risk Framework

Applying Generalized Pareto Distribution to the Risk Management of Commerce Fire Insurance

How to Model Operational Risk, if You Must

A Simple Formula for Operational Risk Capital: A Proposal Based on the Similarity of Loss Severity Distributions Observed among 18 Japanese Banks

SENSITIVITY ANALYSIS AND INFERENCE. Lecture 12

Validation of Internal Rating and Scoring Models

Modelling operational risk in the insurance industry

Imputing Values to Missing Data

Operational Risk - Scenario Analysis

Basics of Statistical Machine Learning

Service courses for graduate students in degree programs other than the MS or PhD programs in Biostatistics.

An analysis of the dependence between crude oil price and ethanol price using bivariate extreme value copulas

LDA at Work: Deutsche Bank s Approach to Quantifying Operational Risk

Operational Value-at-Risk in Case of Zero-inflated Frequency

Statistics Graduate Courses

An Introduction to Extreme Value Theory

Package CoImp. February 19, 2015

Practical Calculation of Expected and Unexpected Losses in Operational Risk by Simulation Methods

Statistics for Retail Finance. Chapter 8: Regulation and Capital Requirements

Monte Carlo Simulation

Managing Capital Adequacy with the Internal Capital Adequacy Assessment Process (ICAAP) - Challenges and Best Practices

Quantitative Operational Risk Management

THE USE OF STATISTICAL DISTRIBUTIONS TO MODEL CLAIMS IN MOTOR INSURANCE

The Variability of P-Values. Summary

Non Linear Dependence Structures: a Copula Opinion Approach in Portfolio Optimization

The advanced measurement approach for banks

Bayesian Phylogeny and Measures of Branch Support

Pair-copula constructions of multiple dependence

Copula Simulation in Portfolio Allocation Decisions

Regulatory and Economic Capital

STATISTICA Formula Guide: Logistic Regression. Table of Contents

Java Modules for Time Series Analysis

itesla Project Innovative Tools for Electrical System Security within Large Areas

Extreme Value Theory with Applications in Quantitative Risk Management

Least Squares Estimation

Third Edition. Philippe Jorion GARP. WILEY John Wiley & Sons, Inc.

Data Preparation and Statistical Displays

Modeling Operational Risk: Estimation and Effects of Dependencies

Uncertainty quantification for the family-wise error rate in multivariate copula models

Pricing of a worst of option using a Copula method M AXIME MALGRAT

STA 4273H: Statistical Machine Learning

Gamma Distribution Fitting

Generating Random Samples from the Generalized Pareto Mixture Model

Analysis of Financial Time Series

Bank Capital Adequacy under Basel III

BAYESIAN ANALYSIS OF AN AGGREGATE CLAIM MODEL USING VARIOUS LOSS DISTRIBUTIONS. by Claire Dudley

APPROACHES TO COMPUTING VALUE- AT-RISK FOR EQUITY PORTFOLIOS

CONTENTS. List of Figures List of Tables. List of Abbreviations

Operational risk capital modelling. John Jarratt National Australia Bank

Contributions to extreme-value analysis

Modeling Individual Claims for Motor Third Party Liability of Insurance Companies in Albania

Actuarial and Financial Mathematics Conference Interplay between finance and insurance

How To Test For Significance On A Data Set

MATHEMATICAL METHODS OF STATISTICS

SAS Certificate Applied Statistics and SAS Programming

Statistical Rules of Thumb

A reserve risk model for a non-life insurance company

Working Paper: Extreme Value Theory and mixed Canonical vine Copulas on modelling energy price risks

Operational Risk. Modeling Analytics. Harry H. Panjer A JOHN WILEY & SONS, INC., PUBLICATION

The Assumption(s) of Normality

Statistical methods to expect extreme values: Application of POT approach to CAC40 return index

Quantitative Methods for Finance

More details on the inputs, functionality, and output can be found below.

Online Appendices to the Corporate Propensity to Save

Inference on the parameters of the Weibull distribution using records

How To Understand The Theory Of Probability

Probability and Statistics

(Part.1) FOUNDATIONS OF RISK MANAGEMENT

CATASTROPHIC RISK MANAGEMENT IN NON-LIFE INSURANCE

Implementing an AMA for Operational Risk

Basel Committee on Banking Supervision. Results from the 2008 Loss Data Collection Exercise for Operational Risk

Curriculum Map Statistics and Probability Honors (348) Saugus High School Saugus Public Schools

**BEGINNING OF EXAMINATION** The annual number of claims for an insured has probability function: , 0 < q < 1.

Loss Distribution Approach for operational risk

Supplement to Call Centers with Delay Information: Models and Insights

Operational Risk Modeling *

Bayesian Analysis for the Social Sciences

Inference on Phase-type Models via MCMC

Statistics I for QBIC. Contents and Objectives. Chapters 1 7. Revised: August 2013

GN47: Stochastic Modelling of Economic Risks in Life Insurance

Tutorial 5: Hypothesis Testing

SOCIETY OF ACTUARIES/CASUALTY ACTUARIAL SOCIETY EXAM C CONSTRUCTION AND EVALUATION OF ACTUARIAL MODELS EXAM C SAMPLE QUESTIONS

Fairfield Public Schools

ATV - Lifetime Data Analysis

Transcription:

Operational Risk Management: Added Value of Advanced Methodologies Paris, September 2013 Bertrand HASSANI Head of Major Risks Management & Scenario Analysis Disclaimer: The opinions, ideas and approaches expressed or presented are those of the author and do not necessarily reflect Santander s position. As a result, Santander cannot be held responsible for them. The values presented are just illustrations and do not represent Santander losses. Copyright: ALL RIGHTS RESERVED. This presentation contains material protected under International Copyright Laws and Treaties. Any unauthorized reprint or use of this material is prohibited. No part of this presentation may be reproduced or transmitted in any form or by any means, electronic or mechanical, including photocopying, recording, or by any information storage and retrieval system without express written permission from the author. Santander UK

Key statements 2 1. Risk management moto: Si Vis Pacem Para Belum 1. Awareness 2. Prevention 3. Control 4. Mitigation 2. A risk is not a loss. 1. It may never crystallise 2. The caracteristics are assumptions 3. Modelling is not the truth 1. Exactly replicating the univers with math is Utopia 2. A model defines itself by its limitations.

Basel Definition 3 Operational Risks (BIS definition): It s «the risk of loss resulting from inadequate or failed internal processes, people and systems or from external events.» However, the Basel Commitee recognizes that operational risk is a term that has a variety of meanings and therefore banks are allowed to implement specific and internal definition of operational risk. The Basel II definition of operational risk clearly excludes strategic risk and reputational risk.

Basel Implications 4 Basel II-III / Solvency II uses a "three pillars" definition. The first pillar minimum capital requirements (addressing risk) basic indicator approach or BIA, standardized approach or TSA, advanced measurement approach or AMA. The second pillar (supervisory review ) The second pillar deals with the regulatory response to the first pillar : the tools, including Backtesting etc. And the framework for dealing with all the other risks a bank may face. The third pillar (market discipline ) The third pillar greatly increases the disclosures that the bank must make.

Framework 5

6 Part 1: Data Analysis

Analysing the data: Internal Taxonomy 7 1. Basel Matrix Vs Internal Taxonomy 2. Cross Distributions

Analysing the data: Statistical Analysis 8 1. Statistical Moments 2. Stationarity (Ergodicity)

Analysing the data: Internal Vs External Data 9

Analysing the data: Autocorrelation 10

Scenario Development Process 11 Scenario Process Scenario Development Operational Risk Control workshop to select initial scenarios Submission to the business owners who are entitled to amend them if they are not representative. Values 1:10 1:40 Evaluation by the business owners Scenario process & output Challenge of the Values Required output Description Sign-offs The type of severe event / scenario that could occur a few times in your career. Expected about once every 10 years in the bank (Juglar Cycle) The type of severe event / scenario that could occur in your career. Expected about once every 40 years in the bank (Kondratiev Cycle) A set of scenarios have been developed for each of the 17 Model Risk Categories. The scenario development process involves the following steps: Pre-interview preparation - briefing notes and calls to scenario participants, internal and external loss exhibits, risk assessment information, etc. Selection of workshop participants Business owners invite relevant business and risk experts for each of the interviews Scenario Assessment Workshop Key stakeholders meeting to finalise scenario descriptions & values across all categories. Challenge the values On three basis Consistency check Conservativeness Integrity Finalisation & Sign-off Agree and document the final descriptions and values 11

12 Part 2: Capital Modelling

The Loss Distribution Approach 13 DATA : Frequency : - Incident Dates Calibration Frequency distribution Probability Discret distribution: Poisson, Binomial, etc. Quantity Monte Carlo Simulation, FFT, Panjer Probability Loss Distribution CR Probability Severity : - Incident Amounts Severity distribution Amount 99,9% Quantile Continous distribution: Lognormal, GPD, Gumbel etc. Amounts VaR

Scenario Based Approach: EVT 14 All the participants are required to provide their quotation on these extreme incidents with regard to their business area/unit, without any cooperation. Then based on the Fisher-Tippet theorem, a generalized extreme value distribution is fitted on these values and the theoretical statistics obtained (mean, median, etc.) are compared to those obtained from the consensus. Depending on the outcome a more or less large confidence interval will be built around the consensus. Mathematical Assumptions The Fisher-Tippet theorem (Fisher and Tippett (1928), Gnedenko (1943)) states that under regular conditions the distribution ofha ε sequence of maxima converges asymptotically to a GEV distribution, whose density is equal to, x u 1 + ε > 0 1 x u h( x; u, β, ε ) = 1 + ε β β 1 x u ε 1+ ε β +* For β, where u R is the location parameter β R is the scale parameter and ε R is the shape parameter. This distribution contains the Fréchet ( ε > 0), the Weibull ( ε = 0) and the Gumbel Distributions ( ε < 0). 1 1 ε e

Scenario Based Approach: Bayesian Networks (Ben-Gal I., Bayesian Networks, in Ruggeri F., Faltin F. & Kenett R., Encyclopedia of Statistics in Quality & Reliability, Wiley & Sons (2007).) 15 Based on Bayes Theorem for Conditional Probabilities DPA (C) EF (S) EDPM (W) BDSF (B) Loss (A)

Data Based Approach: How can we find the appropriate distribution? (1/2) 16 We focus on the severities and try to fit some Distributions. More than 200 simulations 3 goodness-of-fit tests (Kolmogorov-Smirnov, Anderson-Darling, Cramer-Von-Mises): The Lognormal distribution is rejected in 82% of the cases. The Weibull distribution is rejected in 86.5% of the cases. The Exponential distribution is rejected in 100% of the cases One could try Gamma, Gumbel, Beta, Pareto, Burr, Alpha-Stabe etc. until he finds a good one, if there is.

Data Based Approach: How can we find the appropriate distribution? (2/2) 17 Historical Quantiles > Lognormal Quantiles

Data Based Approach: Left Truncated Distributions 18 Expectation-Maximization algorithm f ( x X u) Constrained Maximum Likelihood (BFGS, Nelder-Mead etc.) ( x; X < u) ( X < u) f < = I, where I < P { x< u} { x u} 1if x < u = 0 if x u

Data Based Approach: Right Truncated Distributions 19 Hypothesis: ( ) ( ) > = n p n f * x x 0 g x ( ) = pn f g x n = 1 n = 1 The small losses and the large ones follow different distributions p 0 x = 0 Lognormal distribution to model the central part (below the threshold). ( ) ( ) > = n p n f * x x 0 g x Extreme Value Theory: Pickands n = 1 theorem (1973), above a sufficiently high threshold «u», the data follows a p0 x = 0 Generalized Pareto Distribution: ( x) P( X u x X > u ) The severity distribution: The Peak-over-Threshold method (POT) ( u + x) F ( u ) 1 F ( u ) F F u = = F ~ > GPD ( u, β = scale, ξ = p 0 * n ( x) x > 0 x = 0 shape ) f Global ( x)= w f body + (1 w) f tail Where, w = P( x < u)

Data Based Approach: POT - Threshold Estimation 20 1. Shape parameter estimation based on order statistics (Hill plot and its turning point) Thre sho ld 7 0 0 0.0 0 2 0 3 0.0 0 8 5 0.0 0 3 9 7.0 0 2 9 8.0 0 2 5 0.0 0 1 9 6.0 0 1 9 0.0 0 1 5 1.0 0 1 3 6.0 0 1 1 9.0 0 9 7.5 0 9 7.5 0 9 7.5 0 8 7.8 0 7 7.7 0 6 8.1 0 6 6.0 0 xi (CI, p =0.95) 0.5 1.0 1.5 2.0 2.5 3.0 3.5 1 5 2 7 3 9 5 1 6 3 7 5 8 7 9 9 1 1 2 1 2 7 1 4 2 1 5 7 1 7 2 1 8 7 2 0 2 2 1 7 2 3 2 2 4 7 2 6 2 2 7 7 2 9 2 3 0 7 3 2 2 3 3 7 3 5 2 3 6 7 3 8 2 3 9 7 4 1 2 Order S tatistics

Data Based Approach: POT - Illustration 21

22 Combining Data and Scenarios: Credibility Theory (P.V. Shevchenko (2011). Modelling Operational Risk using Bayesian Inference, Springer, Berlin. ) The strategy involves the following: Selection of the distribution characterizing the risk category Estimation of the parameters using the scenarios on a standalone basis Estimation of the parameters using the Internal Loss data only Valuation of the weights building a non-parametric distribution built implementing a bootstrap/ jackknife approach. Loss Distribution built using a Monte Carlo Simulation Non Parametric Distribution of the Parameters Frequency 0 100 200 300 400 8.40 8.45 8.50 8.55 Parameters 22

Combining Data and Scenarios: Penalised Maximum Likelihood 23 Two steps approach: Maximum Likelihood estimation on the internal data Ordinary Least Square considering the quantile generated using the distribution parameterised and those obtained from the scenarios Internal data percentage obtained using External Data Q-Q plot Sample Quantiles -2 0 2 4-3 -2-1 0 1 2 3 Theoretical Quantiles 23

24 Combining Data and Scenarios: Cascade Bayesian Approach (The Cascade Bayesian Approach for a controlled integration of internal data, external data and scenarios Hassani B., Renaudin A., 2013) Procedure Prior Fitted on Scenario Updated with External Data First Posterior distribution used as Prior Final Posterior distribution Updated with Internal Data Θ* We use the scenarios as a starting point to inform the prior (i.e. distribution of the parameters) and the joint prior. Then we update the joint prior by using the external data (to be consistent, we only use the losses above the 1:10 scenario) which provides us the posterior distribution. This first posterior distribution is then used a prior for a second Bayesian Inference Approach. These are updated using the internal data to obtain a second posterior distribution, from which we are able to derive the optimal parameters of the statistical distribution assumed. Depending on the assumption: we have conjugate priors, i.e. no numerical algorithm is necessary to obtain the posterior distribution (lognormal approach), we do not have conjugate priors, we need a MCMC method to obtain the posterior distribution. 24

Toward a multivariate distribution to simulate the contagion 25 Elliptic Copula f 1 f 2 f 3 f 4 f 5 f 6 f 7 f 8 f 9 f 10 f 1 1 0.26-0.37 0.24 0 0.64-0.41-0.11 0.52 0.41 f 2 0.26 1 0.22 0.21-0.34-0.15 0.71 0.24 0.1 0.05 f 3-0.37 0.22 1 0.21 0.19 0.03-0.41-0.04 0.52 0.36 f 4 0.24 0.21 0.21 1-0.68 0.11 0.36-0.41 0.01-0.47 f 5 0-0.34 0.19-0.68 1 0.24 0.21-0.21 0.36-0.41 f 6 0.64-0.15 0.03 0.11 0.24 1-0.05-0.21-0.34 0.02 f 7-0.41 0.71-0.41 0.36 0.21-0.05 1 0.14 0.52-0.16 f 8-0.11 0.24-0.04-0.41-0.21-0.21 0.14 1-0.03-0.25 f 9 0.52 0.1 0.52 0.01 0.36-0.34 0.52-0.03 1 0.14 f 10 0.41 0.05 0.36-0.47-0.41 0.02-0.16-0.25 0.14 1 Vine Archtiecture A copula function enables creating a multivariate distribution including the various marginal risk distributions. However, elliptic copulas allowing an easier calibration fail capturing asymmetric shocks while extreme values or archimedean copulas permitting the capture of upper tail dependences may be quite difficult to parameterize.

Nested Structures 26 Fully nested copulas in a 4-dimensional case: the initial level is composed by univariate distributions, the second level presents the copula considering one-dimensional marginals, the third level is built with a copula considering bivariate marginals and a univariate distribution, and so on. Partially nested copula in a 4-dimensional case: the initial level is composed of univariate distributions representing each Basel category for instance, the second level represents the bivariate copulas linking these margins and the third level corresponds to the copula linking the two previous ones.

Vines Structures 27 Pair-copula architectures: suggested by Joe (1996), developed successively by Bedford and Cooke (2001; 2002), Kurowicka and Cooke (2004) and Guégan and Maugis (2010), are more flexible: any class of copula can be used (elliptical, Archimedean or extreme-value for instance), and no restriction is required on the parameters. Nevertheless, compared to nested structures, vine copulas require testing and estimating a large quantity of copulas. Pair-copula architectures: Let X = [X 1,X 2,...,X n ] be a vector of random variables, with joint distribution F and marginal distributions F 1, F 2,..., F n : Sklar (1959) theorem insures the existence of a function mapping the individual distributions to the joint one. f denotes the density function associated to the distribution F, then a vine decomposition may be written:

A Risk Management & Measurement Framework 28 Univariate Approach Expected Loss Budget Tolerance Regulatory Capital Economic Capital Reverse Stress-Testing f 1 Multivariate Approach ( f ) M C f, θ 1, K n Bank Resilience f n

29 Part 3: Next Steps

Next Steps 30 Applying

Next Steps 31 The table presents the parameters of the distributions fitted on either the i.i.d. losses or the residuals characterising the CPBP / Retail Banking weekly aggregated. Traditional LDF denotes Frequency Severity, Time Series LDF characterises the second approach considering i.i.d. weekly losses, AR denotes the autoregressive process and both ARFI and Gegenbauer denote their related processes. The standard deviations are provided in brackets. The Goodness-of-Fit (GoF) is considered to be satisfactory if the value is > 5%. If it is not, then the distribution associated to the largest p-value is retained. The best fit per column are presented in bold characters. Note: NA denotes a model "Not Applicable", and OD denotes results which may be provided "On Demand".

Next Steps: Characterisation of the crisis 32 High Correlations Defaults No Liquidity Problem of Capital Contagion Effects Systemic Risk: What Defaults Are Telling Us Kay Giesecke and Baeho Kim Stanford University March 7, 2010