Estimation of Truncated Data Samples in Operational Risk Modeling

Similar documents
Estimation of Truncated Data Samples in Operational Risk Modeling

Estimation of Operational Risk Capital Charge under Parameter Uncertainty

Estimating Operational Risk Capital with Greater Accuracy, Precision, and Robustness

Statistical Distributions and Uncertainty Analysis. QMRA Institute Patrick Gurian

Modelling Operational Risk Using Bayesian Inference

Using R in Undergraduate Probability and Mathematical Statistics Courses. Amy G. Froelich Department of Statistics Iowa State University

Statistical Contributions to Operational Risk Modeling

Robust Statistics vs. MLE for OpRisk Severity Distribution Parameter Estimation

ASTIN Colloquium 1-4 October 2012, Mexico City

Robust Statistics vs. MLE for OpRisk Severity Distribution Parameter Estimation (with and without truncation)

Chapter 9. Non-Parametric Density Function Estimation

Test Problems for Probability Theory ,

Exam C Solutions Spring 2005

Time-varying failure rate for system reliability analysis in large-scale railway risk assessment simulation

Three hours. To be supplied by the Examinations Office: Mathematical Formula Tables and Statistical Tables THE UNIVERSITY OF MANCHESTER.

Multivariate Operational Risk: Dependence Modelling with Lévy Copulas

Two hours. To be supplied by the Examinations Office: Mathematical Formula Tables and Statistical Tables THE UNIVERSITY OF MANCHESTER.

Expected Shortfall is not elicitable so what?

SPRING 2007 EXAM C SOLUTIONS

Using copulas to model time dependence in stochastic frontier models

VaR vs. Expected Shortfall

Asymptotic distribution of the sample average value-at-risk

Estimating VaR in credit risk: Aggregate vs single loss distribution

Inferring from data. Theory of estimators

1 Degree distributions and data

Parameter Estimation

SYSM 6303: Quantitative Introduction to Risk and Uncertainty in Business Lecture 4: Fitting Data to Distributions

Expected Shortfall is not elicitable so what?

Chapter 9. Non-Parametric Density Function Estimation

Overview of Extreme Value Theory. Dr. Sawsan Hilal space

ACTEX CAS EXAM 3 STUDY GUIDE FOR MATHEMATICAL STATISTICS

Loss Distribution Approach for Operational Risk Capital Modelling under Basel II: Combining Different Data Sources for Risk Estimation

Ruin, Operational Risk and How Fast Stochastic Processes Mix

Model Fitting. Jean Yves Le Boudec

DELTA METHOD and RESERVING

Reformulation of chance constrained problems using penalty functions

ON THE FAILURE RATE ESTIMATION OF THE INVERSE GAUSSIAN DISTRIBUTION

A Study of Five Parameter Type I Generalized Half Logistic Distribution

Step-Stress Models and Associated Inference

IE 303 Discrete-Event Simulation

Consideration of prior information in the inference for the upper bound earthquake magnitude - submitted major revision

ABC methods for phase-type distributions with applications in insurance risk problems

Standard Error of Technical Cost Incorporating Parameter Uncertainty

Operational risk modeled analytically II: the consequences of classification invariance. Vivien BRUNEL

Introduction to Rare Event Simulation

P Values and Nuisance Parameters

Point and Interval Estimation for Gaussian Distribution, Based on Progressively Type-II Censored Samples

Fall 2017 STAT 532 Homework Peter Hoff. 1. Let P be a probability measure on a collection of sets A.

Practice Problems Section Problems

Sharp bounds on the VaR for sums of dependent risks

Extreme Value Theory.

Mathematical statistics

Exact Inference for the Two-Parameter Exponential Distribution Under Type-II Hybrid Censoring

How to make projections of daily-scale temperature variability in the future: a cross-validation study using the ENSEMBLES regional climate models

Estimation MLE-Pandemic data MLE-Financial crisis data Evaluating estimators. Estimation. September 24, STAT 151 Class 6 Slide 1

Statistical Tools and Techniques for Solar Astronomers

Risk Aggregation. Paul Embrechts. Department of Mathematics, ETH Zurich Senior SFI Professor.

Empirical Examination of Operational Loss Distributions

Probability and Estimation. Alan Moses

Simulating Exchangeable Multivariate Archimedean Copulas and its Applications. Authors: Florence Wu Emiliano A. Valdez Michael Sherris

Efficient Estimation of Distributional Tail Shape and the Extremal Index with Applications to Risk Management

Bayesian Inference and MCMC

AN IMPORTANCE SAMPLING METHOD FOR PORTFOLIO CVaR ESTIMATION WITH GAUSSIAN COPULA MODELS

Why Correlation Matters in Cost Estimating

Maximum Likelihood Estimation of the Flow Size Distribution Tail Index from Sampled Packet Data

Estimating Operational Risk Capital: the Challenges of Truncation, the Hazards of MLE, and the Promise of Robust Statistics *

Logistic Regression Review Fall 2012 Recitation. September 25, 2012 TA: Selen Uguroglu

Variations. ECE 6540, Lecture 10 Maximum Likelihood Estimation

Truncation and Censoring

Fitting the generalized Pareto distribution to data using maximum goodness-of-fit estimators

Statistical Methods. Missing Data snijders/sm.htm. Tom A.B. Snijders. November, University of Oxford 1 / 23

MODULE 6 LECTURE NOTES 1 REVIEW OF PROBABILITY THEORY. Most water resources decision problems face the risk of uncertainty mainly because of the

The Correlation Problem in Operational Risk

4. Distributions of Functions of Random Variables

AGEC 661 Note Fourteen

Department of Statistical Science FIRST YEAR EXAM - SPRING 2017

Modern Methods of Data Analysis - WS 07/08

Introduction and Overview STAT 421, SP Course Instructor

Reparametrization of COM-Poisson Regression Models with Applications in the Analysis of Experimental Count Data

Introduction to Maximum Likelihood Estimation

Recall the Basics of Hypothesis Testing

Financial Econometrics and Volatility Models Extreme Value Theory

Primer on statistics:

Bayes Decision Theory

Ninth ARTNeT Capacity Building Workshop for Trade Research "Trade Flows and Trade Policy Analysis"

Sequential Importance Sampling for Rare Event Estimation with Computer Experiments

On the Comparison of Fisher Information of the Weibull and GE Distributions

Estimators as Random Variables

Research Article Strong Convergence Bound of the Pareto Index Estimator under Right Censoring

Statistics: Learning models from data

Robustness and bootstrap techniques in portfolio efficiency tests

R Package glmm: Likelihood-Based Inference for Generalized Linear Mixed Models

Frequency Analysis & Probability Plots

Problem 1 (20) Log-normal. f(x) Cauchy

Asymptotic distribution of the sample average value-at-risk in the case of heavy-tailed returns

Statistícal Methods for Spatial Data Analysis

Estimating and Accounting for the Output Gap with Large Bayesian Vector Autoregressions

CPSC 340: Machine Learning and Data Mining. MLE and MAP Fall 2017

ECON Introductory Econometrics. Lecture 11: Binary dependent variables

Bootstrapping the Confidence Intervals of R 2 MAD for Samples from Contaminated Standard Logistic Distribution

Transcription:

Estimation of Truncated Data Samples in Operational Risk Modeling Bakhodir Ergashev FRB of Richmond Prepared for OpRisk North America Conference New York, NY March 20 2013

Caveats The presentation is based on the following working paper (so I am not the sole author): Ergashev, B., Pavlikov, K., Uryasev, S., Sekeris, E. (2012) Estimation of truncated data samples in operational risk modeling The views expressed in this presentation are those of the authors and do not necessarily reflect the position of the Federal Reserve Bank of Richmond or the Federal Reserve System. 2

Introduction Basel II AMA Modeling Regulatory capital is the 99.9 percentile of the distribution of annual aggregate losses. A common approach to modeling operational risk exposure is LDA: Annual loss severity and frequency distributions are estimated. The distribution of annual losses is found by convoluting the two distributions. Available samples of loss event data are usually truncated from below ($10K is a common data collection threshold). 3

Introduction How to treat truncated data samples while modeling severity and frequency distributions? The literature offers the following three approaches to fitting truncated samples: The naïve approach: treat the data as if it is not truncated. The shifting approach: shift the data to the origin, fit it, and add the threshold value to the fitted distribution. The truncation approach: fit the conditional distribution by explicitly acknowledging the existence of censored loses falling below the threshold. In this research, we focus on the last two approaches. 4

Truncation vs. Shifting Shifting approach Numerical algorithms converge fast. The results are stable. Convolution algorithms are more efficient. For heavy tailed distributions the difference is often negligible. The right tail heavily influences the behavior of the left tail. Truncation approach The fitting is not always successful. Frequently, the likelihood surface is ascending with no global maximum. Sometimes frequency estimates are too high making the convolution computationally challenging. The right tail does not influence the behavior of the left tail. 5

Findings by Recent Literature Luo, Shevchenko, Donnelly (2007): The shifting approach overstates or understates VaR depending on the chosen combination of the parameters and threshold. Shevchenko (2009): Naïve and shifted models are easy to fit but induced bias can be very large. Opdyke (2011) emphasizes that truncation exacerbates the non-robustness of MLE and proposes robust alternatives to MLE. For truncated lognormal distribution, large contaminations lead to arbitrarily small mean and arbitrarily large standard deviation parameters. Cope (2011) proposes penalized likelihood estimators for estimating distributional parameters in the presence of data truncation. The cost of the reduction in the variance is a small bias in the parameter estimates. Cavallo et al (2012): Overstatement or understatement depends on the characteristics of the data sample. They propose the use of Vuong s test to choose the best fit model; it does not always work, though. Rozenfeld (2010) proposes to calculate capital from truncated data using the shifting approach by utilizing the information on losses falling below threshold. Caveat: some technical derivations of the paper are not clear to me. 6

Our Contributions In the case of Poisson frequency and lognormal severity, we derive a necessary and sufficient condition for the existence of the global solution to the likelihood maximization problem under the truncation approach. If this regularity condition is violated, the MLE suggests that lognormal is not an appropriate severity distribution for a given data sample. An alternative severity distribution should be used. Violations of the regularity condition are the main reason for obtaining unstable parameter estimates under the truncating approach. When the regularity condition is satisfied, we offer a simple procedure with explicit formulae for calculating the mean and st.dev. parameters of the truncated lognormal severity. The procedure can be implemented in Excel. With the use of the proposed procedure practitioners avoid the challenges associated with maximizing the likelihood function. 7

Our Findings on Capital Bias Using the SLA and the explicit formulae, we approximate capital bias under the truncation and shifting approaches. The approximations are not great, but they reveal the tendencies that are supported by recent literature. To obtain more accurate estimates of capital bias, we conduct a simulation study. The results are as follows: For small samples, both approaches induce significant bias. As the sample size increases, the bias induced by the truncation approach becomes negligible, while the bias induced by the shifting approach does not. Caveat: we explored only the following range: 8 μ 15 and 0.25 σ 4. The main reason our findings are different is that we filter out random samples that violate the regularity condition. 8

The Regularity Condition Notations Severity: X is distributed according to truncated lognormal(μ,σ). Frequency is Poisson(λ). L is the data collection threshold. Truncated sample: X 1,, X n, X > L and Y 1,, Y n = (ln X 1,, ln X n ) Sample moments: M = 1 n Statistic A: A = Function G: G t = V M 2 (M ln L) 2. n n i=1 Y i, i=1 Y 2 i. V = 1 n 1 ( 1 t) and F t F t t F t t = φ(t) 1 Ф(t), where φ and Ф are pdf and cdf of the standard normal distribution. 9

The Regularity Condition 10

The Severity Estimation Procedure 11

Derivation of the Estimation Procedure We use the method of moment estimation. The MM estimation is equivalent to the ML estimation in this setting. Let Y be a lognormal r.v. truncated from below at lnl. There exist an explicit system of equations for E(Y) and E(Y 2 ) expressed in terms of cdf and pdf of the standard normal distribution. 12

Derivation of the Estimation Procedure We rewrite the system of equations for E(Y) and E(Y 2 ) as: where t = σ = G t 1 = ln L μ σ. E Y ln L F t t var Y (E(Y) ln L) 2 Theorem (Barrow and Cohen, 1954) 1. G t 1, when t, 2. G t 2, when t +, 3. G t is a monotonically increasing function on, +. 13

Derivation of the Estimation Procedure The MM system: σ = E Y ln L F t t and G t 1 = var Y (E(Y) ln L) 2. Due to Barrow and Cohen theorem 0 < Statistic A is the sample equivalent of A = var Y (E(Y ln L)) 2 < 1. var Y (E(Y) ln L) 2 V M^2 (M ln L) 2 > 0. A unique solution to the MM system exits when A < 1 and there is no solution when A > 1. We also prove that there is no global maximum of the likelihood function when A >1. 14

Capital Bias Quantifying capital bias using the Single Loss Approximation The explicit formulae are presented in the paper for the SLA approximation to capital bias under both approaches. The accuracy of the approximations is not great. They still reveal the tendencies that are supported by our simulation study. Quantifying capital bias by simulation (next few slides) 15

8.0 8.5 9.0 10.0 9.5 11.0 10.5 11.5 12.0 12.5 13.0 13.5 14.0 14.5 15.0 Bias Under Shifting Approach Sample Size = 500 100% 80% 60% 40% 20% 0% -20% -40% -60% -80% -100% 16

0.75 8.0 8.5 1.00 1.25 9.0 1.50 9.5 10.5 10.0 11.5 11.0 4.00 3.75 3.50 3.25 3.00 2.75 2.50 2.25 2.00 1.75 12.5 12.0 13.5 13.0 14.5 15.0 14.0 Bias Under Shifting Approach Sample Size = 2000 100% 80% 60% 40% 20% 0% -20% -40% -60% -80% -100% 17

8.0 8.5 1.00 9.0 1.50 2.00 10.0 9.5 2.50 3.00 3.50 11.0 10.5 4.00 12.0 11.5 13.0 12.5 14.0 14.5 15.0 13.5 Bias Under Truncation Approach Sample Size = 500 100% 80% 60% 40% 20% 0% -20% -40% -60% -80% -100% 18

8.0 8.5 9.0 10.0 9.5 10.5 11.0 11.5 12.0 12.5 13.0 14.0 13.5 15.0 14.5 Bias Under Truncation Approach Sample Size = 2000 100% 80% 60% 40% 20% 0% -20% -40% -60% -80% -100% 1.00 1.50 2.00 2.50 4.00 3.50 3.00 19

References Barrow, D., Cohen A. (1954) On some functions involving Mill s ratio, The Annals of Mathematical Statistics, 25(2), 405-408. Bӧcker, K., and Klüppelberg, C. (2005) Operational VaR: Closed Form approximation. Risk, 90-93. Cavallo A., Rosenthal, B., Wang, X., Yan, J. (2012) Treatment of the Data Collection Threshold in Operational Risk: A Case Study with the Lognormal Distribution, Journal of Operational Risk Vol. 7, 3-38. Cope, E.W. (2011) Penalized likelihood estimators for truncated data Journal of Statistical Planning and Inference, Vol. 141, 345 358. Luo, X., Shevchenko, P., and Donnelly, J. (2007) Addressing Impact of Truncation and Parameter Uncertainty on Operational Risk Estimates, Journal of Operational Risk, Vol. 2, pp. 3-26. Opdyke, J.D., (2011) Robust Statistics Vs. MLE for Operational Risk Severity distribution Parameter Estimation, Working Paper. Shevchenko, P. (2009) Implementing loss distribution approach for operational risk, Applied Stochastic Models in Business and Industry, Vol. 26, 227-307. Rozenfeld, I. (2010) Using Shifted Distributions in Computing Operational Risk Capital. Available at SSRN: http://ssrn.com/abstract=1596268 or http://dx.doi.org/10.2139/ssrn.1596268 20