05.12.2012 Views

Quantifying operational risk guided by kernel ... - Cass Knowledge

Quantifying operational risk guided by kernel ... - Cass Knowledge

Quantifying operational risk guided by kernel ... - Cass Knowledge

SHOW MORE
SHOW LESS

Create successful ePaper yourself

Turn your PDF publications into a flip-book with our unique Google optimized e-Paper software.

<strong>Quantifying</strong> <strong>operational</strong> <strong>risk</strong> <strong>guided</strong> <strong>by</strong><br />

<strong>kernel</strong> smoothing and continuous<br />

credibility<br />

Jim Gustafsson<br />

Codan Insurance<br />

Jens P. Nielsen, Paul Pritchard and Dix Roberts<br />

Royal&SunAlliance<br />

Codan, Gammel Kongevej 60<br />

DK 1790 Copenhagen V,Denmark<br />

e-mail: jgu@codan.dk,e-mail: npj@codan.dk<br />

One Plantation Place 9th Floor<br />

30 Fenchurch Street, London, EC3M 3BD, UK<br />

e-mail: paul.pritchard@gcc.royalsun.com, e-mail: dix.roberts@gcc.royalsun.com<br />

Abstract: The challenge of how much capital is necessary to protect an<br />

organisation against exposure to <strong>operational</strong> <strong>risk</strong> losses underpins this pa-<br />

per (<strong>operational</strong> <strong>risk</strong> itself is defined as the <strong>risk</strong> of loss arising from inad-<br />

equate or failed internal processes, people and systems or from external<br />

events). The evolutionary nature of <strong>operational</strong> <strong>risk</strong> modelling to establish<br />

capital charges is recognised emphasizing the importance of capturing tail<br />

behaviour. Challenges surrounding the quantification of <strong>operational</strong> <strong>risk</strong><br />

particularly those associated with sparse data are addressed with mod-<br />

ern statistical methodology including nonparametric smoothing techniques<br />

with a particular view to comparison with extreme value theory (EVT). The<br />

credibility approach employed supports analysis from pooled data across<br />

business lines on a dataset from an internationally active insurance com-<br />

pany. The approach has the potential to be applied more generally, for<br />

example where data might be pooled across <strong>risk</strong> types or where a combi-<br />

nation of internal company losses and publicly reported (external) data is<br />

1<br />

imsart ver. 2005/05/19 file: fest_gustafsson.tex date: December 20, 2005


Gustafsson, Nielsen, Pritchard and Roberts/<strong>Quantifying</strong> <strong>operational</strong> <strong>risk</strong> 2<br />

used.<br />

AMS 2000 subject classifications: Primary 65D10; secondary 91B30.<br />

Keywords and phrases: Operational <strong>risk</strong>, <strong>kernel</strong> smoothing, credibility<br />

theory, extreme value theory.<br />

imsart ver. 2005/05/19 file: fest_gustafsson.tex date: December 20, 2005


Gustafsson, Nielsen, Pritchard and Roberts/<strong>Quantifying</strong> <strong>operational</strong> <strong>risk</strong> 3<br />

1. Introduction<br />

There is increasing interest in financial services companies in identifying loss dis-<br />

tributions associated with <strong>operational</strong> <strong>risk</strong>s, driven <strong>by</strong> both regulatory consid-<br />

erations and also in recognition of the greater importance placed on <strong>operational</strong><br />

<strong>risk</strong> management. Building on the 1988 Basel Capital Accord, the Basel Com-<br />

mittee on Banking Supervision published; International Convergence of Capital<br />

Measurement and Capital Standards 1 , in June 2004. This document addressed<br />

the challenge of how much capital is necessary to protect an organization against<br />

unexpected losses, and established the need for an explicit charge for the expo-<br />

sure to <strong>operational</strong> <strong>risk</strong> losses. Operational <strong>risk</strong> itself is defined as the <strong>risk</strong> of<br />

loss arising from inadequate or failed internal processes, people and systems or<br />

from external events. The framework also recognized the evolutionary nature of<br />

<strong>operational</strong> <strong>risk</strong> modelling emphasizing the importance of capturing tail events.<br />

Challenges surrounding the quantification of <strong>operational</strong> <strong>risk</strong> are the sub-<br />

ject of this paper and include lack of suitable data and the need to focus on<br />

tail behaviour. These challenges are addressed with modern statistical method-<br />

ologies including nonparametric smoothing techniques. One of the commonly<br />

applied techniques in <strong>operational</strong> <strong>risk</strong> is Extreme Value Theory (EVT) which<br />

offers a broadly accepted methodology for estimating the tail of a distribution,<br />

see Embrecht, Klüppelberg, and Mikosch [9] for a detailed mathematical treat-<br />

ment, also Embrecht [10], Reiss and Thomas [23] and Coles [7]. It is our view<br />

that EVT shows conceptual similarities to other established density estimation<br />

techniques including the approach we consider in this paper. In essence these<br />

techniques seek to estimate a density with particular emphasis on the tail. In<br />

our approach we transform losses into the interval [0,1].<br />

The introduction of the concept of local constant and local linear density<br />

estimators <strong>by</strong> Jones [16] was a very important contribution since both automat-<br />

1 Details at http//www.bis.org/publ/bcbs107.htm.<br />

imsart ver. 2005/05/19 file: fest_gustafsson.tex date: December 20, 2005


Gustafsson, Nielsen, Pritchard and Roberts/<strong>Quantifying</strong> <strong>operational</strong> <strong>risk</strong> 4<br />

ically adjust at boundaries. Jones [16] also pointed out the importance of dirac<br />

functions while dealing with local projections of data. When dealing with re-<br />

gression, local constant and local linear estimators projections can be obtained<br />

relatively easily, see Fan and Gijbels [11] for an overview. However, in other<br />

areas of mathematical statistics such as density estimation and hazard estima-<br />

tion, the dirac function approach becomes important. See Nielsen [18] for the<br />

multivariate density case and Nielsen [19] and Nielsen and Tanggaard [21] for<br />

the hazard case.<br />

The understanding of density and hazard estimation has recently been taken<br />

to a deeper level through the papers of Jiang and Doksum [14] and Jiang and<br />

Doksum [15]. These papers give an overview on how local polynomial densities<br />

and hazards are to be understood as projections even when complicated trun-<br />

cation and censoring is present. In general terms, these papers introduce local<br />

polynomial hazards and densities as simple plug-in estimators with the empir-<br />

ical cumulative densities and hazards being utilized. The asymptotic theory of<br />

the estimators and their derivatives follow as an immediate consequence.<br />

The parametric transformation approach to <strong>kernel</strong> smoothing of Wand, Mar-<br />

ron and Ruppert [25] has recently been considered with a particular interest in<br />

heavy tailed distributions <strong>by</strong> Bolance, Guillen and Nielsen [1], Clements, Hurn<br />

and Lindsay [6] and Buch-Larsen, Nielsen, Bolance and Guillen [2]. We use<br />

the methodology of Buch-Larsen et al. [2] in this paper. They utilize the three<br />

parameter modified Champernowne distribution which not only demonstrates<br />

desired tail behavior, but unlike EVT, is also informed <strong>by</strong> data from the full<br />

distribution.<br />

Application of a non-parametric local constant <strong>kernel</strong> estimator allows an im-<br />

proved fitting to the sample data. Consideration is given here to the issues raised<br />

<strong>by</strong> Diebold, Schuermann and Stroughair [8], who noted that the bias/variance<br />

trade-off in selection of a tail cut-off for EVT applications is analogous to that<br />

relating to the choice of bandwidth in non-parametric density estimation. In<br />

imsart ver. 2005/05/19 file: fest_gustafsson.tex date: December 20, 2005


Gustafsson, Nielsen, Pritchard and Roberts/<strong>Quantifying</strong> <strong>operational</strong> <strong>risk</strong> 5<br />

our approach the established approach to selection of bandwidth in density es-<br />

timation replaces the complex question of selecting the optimal cut off in EVT.<br />

Diebold et al. [8] highlighted the ‘unfortunate’ use of automatic bandwidth se-<br />

lection to support their statement that <strong>kernel</strong> smoothing performs poorly in<br />

the tail (away from the region where most of the data falls), Buch-Larsen et<br />

al. [2] point out however that the transformation approach to <strong>kernel</strong> density<br />

estimation effectively resolves this issue.<br />

In summary this paper is principally concerned with tools with potentially<br />

broad application to <strong>operational</strong> <strong>risk</strong> quantification. These are applied to an op-<br />

erational loss data set obtained from an internationally active insurer. Following<br />

application of non-parametric <strong>kernel</strong> smoothing we apply continuous credibility<br />

theory to density estimation facilitating the appropriate weighting of pooled<br />

(portfolio) losses as compared to data from individual business lines. This theo-<br />

retical approach is inspired <strong>by</strong> the continuous credibility technique that Hardy<br />

and Panjer [12] and Nielsen and Sandqvist [20] introduced to hazard estimation.<br />

The techniques considered in this paper complement and add to existing knowl-<br />

edge demonstrating potential utility in the significant computational challenges<br />

posed <strong>by</strong> <strong>operational</strong> <strong>risk</strong> quantification. Not only is this potentially useful in the<br />

context described here i.e. pooling data across business lines but also in con-<br />

sidering pooling of internal (company) data with that from other companies,<br />

reported publicly or shared through consortia.<br />

Section 2 describes the structure and general assumptions underpinning the<br />

proposed credibility model on densities in the interval [0, 1]. Section 3 is devoted<br />

to the nonparametric <strong>kernel</strong> density estimator, and section 4 presents the esti-<br />

mation of the credibility model. Section 5 is an application to <strong>operational</strong> <strong>risk</strong><br />

data. Here we lay out the transformation process also a plug-in bandwidth is<br />

discussed. The final section evaluates the <strong>operational</strong> <strong>risk</strong> exposure based on a<br />

one year simulation using the loss data.<br />

imsart ver. 2005/05/19 file: fest_gustafsson.tex date: December 20, 2005


Gustafsson, Nielsen, Pritchard and Roberts/<strong>Quantifying</strong> <strong>operational</strong> <strong>risk</strong> 6<br />

2. The credibility model on densities in interval [0,1]<br />

We assume that we have k lines of <strong>operational</strong> <strong>risk</strong> and consider an i belonging<br />

to the set {1, ..., k}. For such an i assume that Xi1, ..., Xini are independent<br />

identically distributed stochastic variables with density fi on [0, 1]. These den-<br />

sities are assumed to be stochastically varying around a common base density<br />

function f. More precisely, we assume that<br />

fi(x) = θi(x)f(x), (2.1)<br />

where x ∈ [0, 1] and θ1(x), ..., θk(x) are identically distributed stochastic pro-<br />

cesses defined on the interval [0, 1]. This model assumption assumes a relation-<br />

ship between the k business lines. Also we assume that � θi(x)f(x)dx = 1,<br />

E(θi(x)) = 1 and V (θi(x)) = ν 2 i<br />

for every i between 1 and k. The assumption<br />

that E{θi(x)} = 1 for all i is crucial for the estimators developed. This assump-<br />

tion implies that for every single x, the business lines varies around the common<br />

base density function f. This assumption allows us to construct our estimators<br />

from a Hilbert space projection of the fi(x) ′ s down at their estimators for each<br />

x individually. The model proposes that each line has <strong>risk</strong> stemming from a<br />

common source as well as its own line of <strong>risk</strong>. The multiplicative construction<br />

above is known from the bias reduction literature, see Hjort and Glad [13] and<br />

Jones, Linton and Nielsen [17]. The difference is, however, that there the pur-<br />

pose is to eliminate bias while here we use the element in the structure as a<br />

stochastic process. The multiplicative structure of our approach is an important<br />

part of our model building. In the bias reduction literature the multiplicative<br />

structure is just used as a trick to reduce bias. However, similarities remain:<br />

we also estimate the multiplicative error nonparametrically and use it for our<br />

estimation purposes - just like one does in the bias reduction literature.<br />

imsart ver. 2005/05/19 file: fest_gustafsson.tex date: December 20, 2005


Gustafsson, Nielsen, Pritchard and Roberts/<strong>Quantifying</strong> <strong>operational</strong> <strong>risk</strong> 7<br />

3. The <strong>kernel</strong> density estimator<br />

Let K denote a probability density function symmetric about zero with support<br />

[−1, 1] and let Kh(x) = K(x/h)/h for any x in [0, 1] and any positive h. We<br />

also define the following functions for the asymptotic properties of the <strong>kernel</strong><br />

density estimator around boundaries.<br />

akl(x, h) =<br />

min{1,x/h} �<br />

max{−1,(x−1)/h}<br />

u k K(u) l du, for x ∈ [0, 1].<br />

Note that a01(x, h) = 1 and a11(x, h) = 0 for the interior points x in the<br />

interval [h, 1 − h]. In the boundary points belonging to the intervals [0, h) and<br />

(1−h, 1], a01(x, h) and a11(x, h) take nontrivial values. The local constant <strong>kernel</strong><br />

estimator 2 of fi is<br />

�fi(x) = (a01(x, hi)ni) −1<br />

ni �<br />

Khi (x − Xij) . (3.1)<br />

The local constant <strong>kernel</strong> estimator of the entire data set can be taken as an<br />

estimator of the common base density f. We assume that the estimation error<br />

of the global <strong>kernel</strong> density estimator<br />

⎛<br />

k�<br />

�f(x) = ⎝(a01(x, hi)ni) −1<br />

⎞<br />

ni �<br />

Khi (x − Xij) ⎠ (3.2)<br />

i=1<br />

is of lower order of magnitude than the estimation error of the individual densi-<br />

ties f1, ..., fk. This is important for our continuous credibility approach defined<br />

in the next section.<br />

4. Estimation of the credibility model<br />

Credibility theory as known from actuarial science is simply a methodology to<br />

find out how much information a common model carries onto a specific line of<br />

2 See Jones [16] for a definition of local costant and local linear density estimators and for<br />

j=1<br />

j=1<br />

a precision of the automatic boundary corrections of these estimators.<br />

imsart ver. 2005/05/19 file: fest_gustafsson.tex date: December 20, 2005


Gustafsson, Nielsen, Pritchard and Roberts/<strong>Quantifying</strong> <strong>operational</strong> <strong>risk</strong> 8<br />

business. Credibility theory tells us that we can rely more on data from one<br />

specific line when it is plentiful, when data for one line is sparse more weight<br />

should be put on a common source. In our case we fix x ∈ [0, 1] and define a<br />

Hilbert space projection, � fi(x), of fi(x) defined <strong>by</strong> (2.1), onto the linear space<br />

{ax + bx � fi(x) | ax, bx ∈ [0, 1]}. We get<br />

�fi(x) = (1 − zi,x) E( � fi(x)) + zi,x � fi(x), (4.1)<br />

where the credibility factor zi,x = COV (fi(x), � fi(x))/V ( � fi(x)) is between zero<br />

and one. This projection gives us the optimal linear credibility estimator � fi(x)<br />

minimizing E((fi(x) − � fi(x)) 2 ), see also Hardy and Panjer [12] and Nielsen and<br />

Sandqvist [20].<br />

The credibility factor quantifies the amount of weight the common mean<br />

and the individual mean should have for each of the lines of <strong>operational</strong> <strong>risk</strong>.<br />

For the original approach to credibility theory, see Bühlmann and Straub [3].<br />

For a recent overview of the origins of credibility theory and a Hilbert space<br />

interpretation of the Bühlmann and Straub model, see Norberg [22]. To calculate<br />

the three moment quantities of (4.1), we prove in the Appendix that<br />

and<br />

V ( � fi(x)) =<br />

E( � fi(x)) = f(x) (1 + op(1)) ,<br />

COV (fi(x), � fi(x)) = f(x) 2 ν 2 i (1 + op(1))<br />

�<br />

f(x) 2 ν 2 i + a02(x, hi)f(x)<br />

a01(x, hi) 2 �<br />

(1 + op(1)) .<br />

hini<br />

From these moment expressions we can approximate the optimal credibility<br />

estimator of (4.1) to<br />

where<br />

zi,x =<br />

�fi(x) = (1 − zi,x) f(x) + zi,x � fi(x),<br />

f(x) 2 ν 2 i a01(x, hi) 2 hini<br />

f(x) 2 ν 2 i a01(x, hi) 2 hini + f(x)a02(x, hi) (1 + op(1)) .<br />

imsart ver. 2005/05/19 file: fest_gustafsson.tex date: December 20, 2005


Gustafsson, Nielsen, Pritchard and Roberts/<strong>Quantifying</strong> <strong>operational</strong> <strong>risk</strong> 9<br />

To estimate the local variance ν 2 i of the stochastic process we note that � fi(x) is<br />

an estimator of fi(x) = θi(x)f(x) and<br />

�θi(x) = � fi(x)<br />

�f(x)<br />

(4.2)<br />

is an estimator of θi(x). Therefore, for a given x, we estimate the variance of<br />

θi(x) <strong>by</strong><br />

�ν 2 i = 1<br />

ni �<br />

ni<br />

j=1<br />

� �2 �θi(xj) − 1 =<br />

� 1<br />

0<br />

� �2 �θi(x) − 1 d � Fi(x), (4.3)<br />

where � Fi is the empirical distribution function of the i’th business line.<br />

We can here<strong>by</strong> estimate zi,x <strong>by</strong><br />

�zi,x =<br />

�f(x) 2�ν 2 i a01(x, hi) 2hini �f(x) 2�ν 2 i a01(x, hi) 2hini + � . (4.4)<br />

f(x)a02(x, hi)<br />

Combining (3.1), (3.2) and (4.4) the final expression for the optimal credibility<br />

estimator is approximately equal to<br />

� �f i(x) = (1 − �zi,x) � f(x) + �zi,x � fi(x). (4.5)<br />

To arrive at a credibility based estimator of θi(x) we use the above equality<br />

(4.5) divided <strong>by</strong> � f(x)<br />

� �θi(x) = (1 − �zi,x) + �zi,x<br />

�fi(x)<br />

�f(x) = 1 − �zi,x(1 − � θi(x)). (4.6)<br />

We propose here that each line has <strong>risk</strong> stemming from a common source as well<br />

as its own individual source. This property of our basic model is clearly reflected<br />

in the expression of our final estimators � � f i(x) and � � θi(x); both are constructed<br />

as a sum of one common and one global element.<br />

5. Application to <strong>operational</strong> <strong>risk</strong> data<br />

We utilize data from seven lines of <strong>operational</strong> <strong>risk</strong> with observed number of<br />

losses N1, ..., N7. The data periods in these seven groups measured in years<br />

are T1, ..., T7. We assume that we have seven independent homogenous Poisson<br />

imsart ver. 2005/05/19 file: fest_gustafsson.tex date: December 20, 2005


Gustafsson, Nielsen, Pritchard and Roberts/<strong>Quantifying</strong> <strong>operational</strong> <strong>risk</strong> 10<br />

processes describing the number of losses in each line and therefore obtain the<br />

following maximum likelihood estimator of the intensity of losses in the i th<br />

business line: � λi = Ni/Ti. Conditional on N1 = n1, ..., N7 = n7, we now have<br />

our <strong>operational</strong> <strong>risk</strong> data for each group, Yi1, ..., Yini , defined on the positive<br />

real axis. We now transform our data from the positive real axis to the interval<br />

[0, 1] using the modified Champernowne distribution defined in Buch-Larsen et<br />

al. [2]:<br />

Fp(y) =<br />

(y + c) α − cα (y + c) α + (M + c) α , y ∈ R+,<br />

− 2cα where p = {α, M, c} is a parameter vector. For c = 0 this distribution is a<br />

special case of the parametric distribution suggested <strong>by</strong> Champernowne [4] and<br />

Champernowne [5] and the cumulative density function equals the transforma-<br />

tion used <strong>by</strong> Clements et al. [6] in their approach to density estimation based<br />

on transformed data. However, an extensive simulation study <strong>by</strong> Buch-Larsen<br />

et al. [2] show that the flexibility of the modified Champernowne distribution<br />

outweigh the advantages of the stability obtained in the simple case of c = 0.<br />

Buch-Larsen et al. [2] estimate M, the median, from the empirical median and<br />

the parameters (α, c) <strong>by</strong> the maximum likelihood method. We use the modified<br />

Champernowne distribution for the i th line of business as follows<br />

Fpi(y) =<br />

(y + c) α − c α<br />

(y + c) α + (Mi + c) α − 2c α , y ∈ R+, (5.1)<br />

where pi = {α, Mi, c} is the extended parameter vector for the individual line<br />

of business. We use the same values of the variables (α, c) for the seven lines of<br />

<strong>operational</strong> <strong>risk</strong> while the median, a scaling parameter, is estimated individually<br />

for each line. This parametric model is the common source of information for<br />

our <strong>operational</strong> <strong>risk</strong> modelling. Firstly we transform the data from each line into<br />

the interval [0, 1] using this parametric model. That is, we define<br />

Xij = F�pi (Yij), (5.2)<br />

imsart ver. 2005/05/19 file: fest_gustafsson.tex date: December 20, 2005


Gustafsson, Nielsen, Pritchard and Roberts/<strong>Quantifying</strong> <strong>operational</strong> <strong>risk</strong> 11<br />

as the transformed losses obtained <strong>by</strong> the cumulative modified Champernowne<br />

distribution, where �pi = {�α, � Mi, �c} are the estimated parameter vector for<br />

i ∈ {1, . . . , k} on observations j ∈ {1, . . . , ni}. Now we are ready to apply<br />

the methodology outlined in sections 2 and 3 using the transformed data set<br />

Xij. Note that if we had only one line of business, then the credibility approach<br />

would of course be superfluous and we have the exact same one-dimensional<br />

estimation method as the one suggested in Buch-Larsen et al. [2]. When esti-<br />

mating the initial individual <strong>kernel</strong> estimator of the i th line of business, defined<br />

<strong>by</strong> (3.1), we use the Epanechnikov <strong>kernel</strong> function:<br />

K(u) = 3 � 2<br />

1 − u<br />

4<br />

� 1 {|u|


Gustafsson, Nielsen, Pritchard and Roberts/<strong>Quantifying</strong> <strong>operational</strong> <strong>risk</strong> 12<br />

Consequently, if we differentiate this with respect to the bandwidth hi, we obtain<br />

the theoretical optimal choice<br />

�<br />

a02(x, hi)<br />

hi =<br />

a21(x, hi) 2 �1/5 � .<br />

ni f ′′ (x) 2dx Here everything is known except for the density f. Assuming the unknown f<br />

follows a normal distribution, we obtain Silverman’s rule of thumb (Silverman<br />

[24]):<br />

� √ �1/5 40 π<br />

hi =<br />

�σi. (5.3)<br />

ni<br />

The following section shows a summary of the <strong>operational</strong> loss data set fol-<br />

lowed <strong>by</strong> a visual comparison applied to the data.<br />

Table 1: Summary Statistics- <strong>operational</strong> loss data set.<br />

Line of<br />

Business ni Ti hi mean(Yij) median(Yij) sd(Yij) max(Yij)<br />

1 250 3 0.24 8981 4059.5 16490 163233<br />

2 46 3 0.36 13450 2829.5 58512 394969<br />

3 924 3 0.16 7835 2990 36164 874400<br />

4 34 3 0.47 13963 2277 18467 52200<br />

5 7 3 0.48 19633 14610 14621 38205<br />

6 7 3 0.58 959382 197600 2123445 5765217<br />

7 23 2 0.44 1127077 242220 2333035 8477700<br />

The second column shows the number of observations ni for each line of<br />

business. There is considerable variation in the number of recorded losses, line<br />

of business 3 has 925 losses compared to line 5 and 6 which only have been<br />

exposed to 7 losses each. The third column gives the time over which data<br />

was collected, in years. The fourth column shows the estimated bandwidths,<br />

obtained <strong>by</strong> (5.3). Note that the estimated bandwidths decrease with sample<br />

size as expected. Columns five to eight show some empirical results on each line<br />

imsart ver. 2005/05/19 file: fest_gustafsson.tex date: December 20, 2005


Gustafsson, Nielsen, Pritchard and Roberts/<strong>Quantifying</strong> <strong>operational</strong> <strong>risk</strong> 13<br />

of business. Note that the mean is significantly larger than the median in all<br />

cases, consistent with right skewed distributions.<br />

Figure 1 shows histograms of both the transformed global data set and the<br />

seven individual data sets. In each graph the respective semiparametric estima-<br />

tors are estimated. If we examine the global estimator � f(x) more closely (top-left<br />

graph), we observe that in the interval [0, 0.3) the estimator is less than one.<br />

This means that the <strong>kernel</strong> estimator corrects the parametric density where it<br />

has diverged from the <strong>operational</strong> <strong>risk</strong> data. In [0.3, 0.7) the <strong>kernel</strong> function<br />

corrects the parametric density where it was too low. In [0.7, 1] the estimator is<br />

once again below one and therefore adjustment is made, equivalent to a lighter<br />

tail. The other individual estimators � fi(x), i = 1, ...7, should be interpreted<br />

analogously.<br />

Insert Figure 1 about here<br />

In Figure 2 we show the credibility approach applied to each line. Here we<br />

omit the histograms to allow easier visual interpretation. We compare the stan-<br />

dard semiparametric estimator (3.1), show with a dotted curve, with an es-<br />

timator where we have applied the credibility approach (solid curve), defined<br />

through (4.5). Note that for lines of <strong>operational</strong> <strong>risk</strong> with sparse information the<br />

appearance of the credibility estimator is similar to the global source. This is be-<br />

cause the global source provides more weight for lines with sparse information,<br />

particular lines 5, 6 and 7.<br />

Insert Figure 2 about here<br />

Figure 3 shows the estimated stochastic processes with and without the appli-<br />

cation of a credibility approach. The left-hand graph shows estimated stochastic<br />

processes for all lines estimated through formula (4.2). Here we can see that the<br />

processes take values between 0.5 and 1.5. This means that when a line of<br />

business deviates from the horizontal we know that there is a large difference<br />

between the specific line and the global source. The right-hand graph shows<br />

imsart ver. 2005/05/19 file: fest_gustafsson.tex date: December 20, 2005


Gustafsson, Nielsen, Pritchard and Roberts/<strong>Quantifying</strong> <strong>operational</strong> <strong>risk</strong> 14<br />

the estimated stochastic processes applying the credibility approach, calculated<br />

using (4.6). The most obvious difference is that the lines are much closer to the<br />

horizontal. This is, of course, to be expected since, to a greater or lesser degree,<br />

the individual lines will use information from the global source.<br />

Insert Figure 3 about here<br />

6. Evaluating an <strong>operational</strong> loss distribution<br />

In this section we simulate an <strong>operational</strong> <strong>risk</strong> loss distribution for the company.<br />

This is done <strong>by</strong> using the severity distribution obtained from (4.5), together with<br />

a Poisson based frequency distribution. The frequency and severity distributions<br />

are used to create simulated one year loss distribution through Monte Carlo<br />

analysis. We sample from a Poisson process of event times through all lines of<br />

<strong>operational</strong> <strong>risk</strong>, and combine with loss sizes taken form the relevant severity<br />

distribution. To obtain our original scale, we transform the estimator (4.5) to<br />

its original axis, The relevant estimators takes the form<br />

ni �<br />

�fi(y)<br />

−1<br />

= (a01(F�pi (y), hi)ni)<br />

�f(y) =<br />

j=1<br />

⎛<br />

k�<br />

ni �<br />

⎝(a01(F�pi<br />

−1<br />

(y), hi)ni)<br />

i=1<br />

Khi (F�pi (y) − F�pi (Yij)) f�pi (y) (6.1)<br />

j=1<br />

Khi (F�pi (y) − F�pi (Yij)) f�pi (y)<br />

where f�pi (y) is the modified Champernowne density defined <strong>by</strong><br />

f�pi (y) = α(y + c)α−1 ((Mi + c) α − cα )<br />

((y + c) α + (Mi + c) α − 2cα 2 , y ∈ R+.<br />

)<br />

The loss sizes are then taken from the severity distribution<br />

� ∞<br />

�<br />

Hi(y) = �f i(ξ)dξ,<br />

0<br />

where � � f i are defined through the estimators in (6.1) with structure as (4.6).<br />

Summation of amounts from all lines provides a single estimate, this process is<br />

then repeated 20000 times to create a simulated loss distribution as shown in<br />

imsart ver. 2005/05/19 file: fest_gustafsson.tex date: December 20, 2005<br />

⎞<br />


Gustafsson, Nielsen, Pritchard and Roberts/<strong>Quantifying</strong> <strong>operational</strong> <strong>risk</strong> 15<br />

Figure 4. From this we can identify the loss amount associated with the relevant<br />

quantile and thus the capital to be held.<br />

Insert Figure 4 about here<br />

In table 2 we present summary statistics for the individual loss distribution<br />

and for the global source.<br />

Table 2: Summary statistics for simulated loss distribution<br />

<strong>by</strong> business line and global source.<br />

Line of<br />

Business x0.5 OpVaR0.95 OpVaR0.99 OpVaR0.999<br />

1 1840605 372988 528891 747380<br />

2 238045 138215 212161 294414<br />

3 5657744 1124957 1690317 1897271<br />

4 156984 118278 180286 258575<br />

5 154260 342791 594155 843894<br />

6 2088010 4865225 10405343 15955968<br />

7 9856090 8621910 13031243 19455952<br />

Global 20706695 9675355 14890803 22462507<br />

The Operational Value-at-Risk (OpVaR) from table 2 is a measure of the<br />

unexpected losses to a specific quantile level. To obtain this value we subtract<br />

the median (presented as column 1) from three different upper quantiles with<br />

confidence levels 95%, 99% and 99.9%. Table 2 shows that the appearance in<br />

the tail region varies between the lines of business. More precisely, line 1 and 3<br />

generate a light tailed distribution while line of business 2, 4 and 7 are charac-<br />

terized <strong>by</strong> moderate tails. The remaining two lines of business 5 and 6 are heavy<br />

tailed distributed. This is not extraordinary since both lines are embodied with<br />

extremely small samples.<br />

Table 3 shows measures of the length of the tail. This is done <strong>by</strong> dividing<br />

the upper quantile (with one of the three confidence levels) <strong>by</strong> the median, i.e.<br />

imsart ver. 2005/05/19 file: fest_gustafsson.tex date: December 20, 2005


Gustafsson, Nielsen, Pritchard and Roberts/<strong>Quantifying</strong> <strong>operational</strong> <strong>risk</strong> 16<br />

ϕ (q) = xq/x0.5 , q = 0.95, 0.99 or 0.999. Here, as in table 2 lines of business 5<br />

and 6 stand out being most heavy tailed.<br />

Table 3: Tail length characteristics.<br />

Line of<br />

Business ϕ (0.95) ϕ (0.99) ϕ (0.999)<br />

1 1.203 1.287 1.406<br />

2 1.581 1.891 2.237<br />

3 1.199 1.299 1.335<br />

4 1.753 2.148 2.647<br />

5 3.222 4.852 6.471<br />

6 3.330 5.983 8.642<br />

7 1.875 2.320 2.974<br />

Global 1.467 1.719 2.085<br />

In this paper we have demonstrated that a number of statistical approaches<br />

(complementary to those commonly applied such as EVT) can provide signif-<br />

icant benefit in the quantification of <strong>operational</strong> <strong>risk</strong> for estimation of capital<br />

requirements in organizations. The fundamental challenge, that of sparse data,<br />

is addressed <strong>by</strong> transforming available data and then applying non-parametric<br />

<strong>kernel</strong> smoothing. This approach (even without further application of credibility<br />

analysis) should assist in maximizing useful information from limited data sets,<br />

particularly as the direct application of non parametric techniques would not<br />

yield useful information. The subsequent credibility element could also be ap-<br />

plied in isolation when considering data that is pooled (e.g. across businesses or<br />

<strong>risk</strong> types) or for the mixing of internal and external data (e.g. from a subscrip-<br />

tion database). We acknowledge the limitations inherent in this work, specifically<br />

that the data set used does not give full coverage across all <strong>operational</strong> <strong>risk</strong> cat-<br />

egories and that it is based on a limited collection period. Nevertheless for the<br />

purposes of illustration it does demonstrate that the techniques can be applied<br />

imsart ver. 2005/05/19 file: fest_gustafsson.tex date: December 20, 2005


Gustafsson, Nielsen, Pritchard and Roberts/<strong>Quantifying</strong> <strong>operational</strong> <strong>risk</strong> 17<br />

to real life company data, furthermore that considerable potential exists for its<br />

further application in this area.<br />

7. Appendix<br />

Through standard <strong>kernel</strong> smoothing theory we obtain<br />

and<br />

� �<br />

E �fi(x) | θi(x)<br />

� �<br />

E �fi(x)<br />

= θi(x)f(x) (1 + op(1))<br />

� � ��<br />

= E E �fi(x) | θi(x)<br />

= E (θi(x)f(x)) (1 + op(1))<br />

= f(x) (1 + o(1)) .<br />

To derive the variance expressions we first note that<br />

V<br />

This implies that<br />

and<br />

V<br />

� �<br />

�fi(x)<br />

� �<br />

�fi(x) | θi(x)<br />

= V<br />

=<br />

= θi(x)a02(x,hi)f(x)<br />

a01(x,hi) 2 hini<br />

� � ��<br />

E �fi(x) | θi(x)<br />

�<br />

f(x) 2 ν 2 i<br />

+ a02(x,hi)f(x)<br />

a01(x,hi) 2 hini<br />

�<br />

+ E<br />

(1 + op(1)) .<br />

V<br />

� ��<br />

�fi(x) | θi(x)<br />

�<br />

(1 + op(1)) .<br />

�<br />

COV fi(x), � �<br />

fi(x)<br />

�<br />

� ��<br />

= COV E (fi(x) | θi(x)) , E �fi(x) | θi(x)<br />

� �<br />

+E COV fi(x), � ��<br />

fi(x) | θi(x)<br />

� �<br />

= E COV fi(x), � ��<br />

fi(x) | θi(x)<br />

= � E � f(x) 2 θi(x) 2� − f(x) 2� (1 + op(1))<br />

= � f(x) 2 E � θi(x) 2 − 1 �� (1 + op(1))<br />

= f(x) 2 ν 2 i (1 + op(1)) .<br />

imsart ver. 2005/05/19 file: fest_gustafsson.tex date: December 20, 2005


References<br />

Gustafsson, Nielsen, Pritchard and Roberts/<strong>Quantifying</strong> <strong>operational</strong> <strong>risk</strong> 18<br />

[1] Bolance, C., Guillen, M. and Nielsen, J.P. (2003). Kernel density<br />

estimation of actuarial loss functions. Insurance: Mathematics and Eco-<br />

nomics, 32, 19-36.<br />

[2] Buch-Larsen, T., Nielsen, J.P., Guillen, M. and Bolance, C.<br />

(2005). Kernel density estimation for heavy tailed distributions using the<br />

Champernowne distribution. Manuscript Universidad de Barcelona (to ap-<br />

pear).<br />

[3] Bühlmann, H and Straub, E. (1970). Glabwürdigkeit für schadensätze.<br />

Bulletin of the Association of Swiss Actuaries 70, 111-133.<br />

[4] Champernowne D.G. (1936). The Oxford meeting, September 25-29,<br />

1936, Econometrica, Vol. 5, No. 4, October 1937.<br />

[5] Champernowne D.G. (1952). The graduation of income distributions.<br />

Econometrica, 20, 591-615.<br />

[6] Clements, A.E., Hurn, A.S. and Lindsay, K.A. (2003). Möbius-like<br />

mappings and their use in <strong>kernel</strong> density estimation. Journal of the Amer-<br />

ical Statistical Association 98, 993-1000.<br />

[7] Coles, S. (2001). An introduction to statistical modeling of Extreme Val-<br />

ues. Springer.<br />

[8] Diebold, F.X., Schuermann, T. and Stroughair,J. (2000). Pitfalls<br />

and Opportunities in the Use of Extreme Value Theory in Risk Manage-<br />

ment. London: Extremes and Integrated Risk Management London: Risk<br />

Books.<br />

[9] Embrechts, P., Klüppelberg, C. and Mikosch, T. (1999). Modeling<br />

Extremal Events for Insurance and Finance. Springer.<br />

[10] Embrechts, P. (2000). Extremes and Integrated Risk Management. Lon-<br />

don: Risk Books, Risk Waters Group.<br />

[11] Fan, J. and Gijbels, I. (1996). Local Polynomial Modelling and its Ap-<br />

imsart ver. 2005/05/19 file: fest_gustafsson.tex date: December 20, 2005


Gustafsson, Nielsen, Pritchard and Roberts/<strong>Quantifying</strong> <strong>operational</strong> <strong>risk</strong> 19<br />

plications. Chapman and Hall, London.<br />

[12] Hardy, M.R. and Panjer, H. (1998). A credibility approach to mor-<br />

tality <strong>risk</strong>. ASTIN Bulletin 28, 2, 269-283.<br />

[13] Hjort, N.L. and Jones, M.C. (1995). Nonparametric density estima-<br />

tion with a parametric start. The Annals of Statistics, Vol. 23, No. 3,<br />

882-904.<br />

[14] Jiang, J. and Doksum, K. (2003a). On local polynomial estimation of<br />

hazard rates and their derivatives under random censoring, Constance van<br />

Eeden Volume, IMS, 463-481.<br />

[15] Jiang, J. and Doksum, K. (2003b). Empirical plug-in curve and surface<br />

estimates. In: Mathematical and Statistical Methods in Reliability. Series<br />

on Quality, Reliability and Engineering Statistics, Vol. 7, World Scientific<br />

Publishing, Singapore, 433-453.<br />

[16] Jones, M.C. (1993). Simple boundary correction for <strong>kernel</strong> density esti-<br />

mation. Statistics and Computing 3, 135 - 146.<br />

[17] Jones, M.C., Linton O. and Nielsen, J.P. (1995). A simple bias<br />

reduction method for density estimation. Biometrica, Vol. 82, No. 2, 327-<br />

338.<br />

[18] Nielsen, J.P. (1998). Marker dependent <strong>kernel</strong> hazard estimation from<br />

local linear estimation, Scand. Actuar. J. 2 , 113-124<br />

[19] Nielsen, J.P. (1999). Multivariate <strong>kernel</strong>s from local linear estimation,<br />

Scand. Actuar. J. 1 , 93-95.<br />

[20] Nielsen, J.P. and Sandqvist, B.L. (2000). Credibility weighted hazard<br />

estimation. Astin Bulletin 30, 405-417.<br />

[21] Nielsen, J.P. and Tanggaard, C. (2001). Boundary and bias correc-<br />

tion in <strong>kernel</strong> hazard estimation. Scandinavian Journal of Statistics, 28,<br />

675-698.<br />

[22] Norberg, R (2004). Credibility theory. In: Encyclopedia of Actuarial<br />

Science, Wiley.<br />

imsart ver. 2005/05/19 file: fest_gustafsson.tex date: December 20, 2005


Gustafsson, Nielsen, Pritchard and Roberts/<strong>Quantifying</strong> <strong>operational</strong> <strong>risk</strong> 20<br />

[23] Reiss, R.D. and Thomas J.A. (2001). Statistical Analysis of Extreme<br />

Values. Basel.<br />

[24] Silverman, B.W. (1986). Density estimation for statistics and data anal-<br />

ysis. Chapman and Hall.<br />

[25] Wand, P., Marron, J.S. and Ruppert, D. (1991). Transformations in<br />

Density Estimation. Journal of the Americal Statistical Association, 30,<br />

405-417.<br />

imsart ver. 2005/05/19 file: fest_gustafsson.tex date: December 20, 2005


0.0 0.2 0.4 0.6 0.8 1.0 1.2<br />

0.0 0.5 1.0 1.5 2.0 2.5 3.0 3.5<br />

Gustafsson, Nielsen, Pritchard and Roberts/<strong>Quantifying</strong> <strong>operational</strong> <strong>risk</strong> 21<br />

Global<br />

0.0 0.4 0.8<br />

Line 4<br />

0.0 0.4 0.8<br />

0.0 0.5 1.0 1.5<br />

0.0 0.5 1.0 1.5 2.0 2.5<br />

Line 1<br />

0.0 0.4 0.8<br />

Line 5<br />

0.0 0.4 0.8<br />

0.0 0.5 1.0 1.5 2.0<br />

0.0 0.5 1.0 1.5 2.0 2.5<br />

Line 2<br />

0.0 0.4 0.8<br />

Line 6<br />

0.0 0.4 0.8<br />

0.0 0.2 0.4 0.6 0.8 1.0 1.2<br />

0.0 0.5 1.0 1.5 2.0<br />

Line 3<br />

0.0 0.4 0.8<br />

Line 7<br />

0.0 0.4 0.8<br />

Fig 1: Histogram of data,including the global semiparametric estimator (3.2) in the<br />

top-left graph and the individual semiparametric estimators defined <strong>by</strong> (3.1)<br />

in the remaining graphs.<br />

0.6 0.8 1.0 1.2 1.4 1.6<br />

0.6 0.8 1.0 1.2 1.4 1.6<br />

Global<br />

0.0 0.4 0.8<br />

Line 4<br />

0.0 0.4 0.8<br />

0.6 0.8 1.0 1.2 1.4 1.6<br />

0.6 0.8 1.0 1.2 1.4 1.6<br />

Line 1<br />

0.0 0.4 0.8<br />

Line 5<br />

0.0 0.4 0.8<br />

0.6 0.8 1.0 1.2 1.4 1.6<br />

0.6 0.8 1.0 1.2 1.4 1.6<br />

Line 2<br />

0.0 0.4 0.8<br />

Line 6<br />

0.0 0.4 0.8<br />

0.6 0.8 1.0 1.2 1.4 1.6<br />

0.6 0.8 1.0 1.2 1.4 1.6<br />

Line 3<br />

0.0 0.4 0.8<br />

Line 7<br />

0.0 0.4 0.8<br />

Fig 2: The severity densities for all lines of <strong>operational</strong> <strong>risk</strong> including also the global<br />

source (3.2) <strong>by</strong> top-left graph. Dotted curve represent standard<br />

semiparametric approach (3.1), while the solid curve is estimators with an<br />

credibility approach (4.5)<br />

imsart ver. 2005/05/19 file: fest_gustafsson.tex date: December 20, 2005


0.5 1.0 1.5<br />

Gustafsson, Nielsen, Pritchard and Roberts/<strong>Quantifying</strong> <strong>operational</strong> <strong>risk</strong> 22<br />

Line 1<br />

Line 2<br />

Line 3<br />

Line 4<br />

Line 5<br />

Line 6<br />

Line 7<br />

0.0 0.2 0.4 0.6 0.8 1.0<br />

0.5 1.0 1.5<br />

Line 1<br />

Line 2<br />

Line 3<br />

Line 4<br />

Line 5<br />

Line 6<br />

Line 7<br />

0.0 0.2 0.4 0.6 0.8 1.0<br />

Fig 3: The stochastic processes for all lines of <strong>operational</strong> <strong>risk</strong>. The left-hand graph<br />

presents processes without credibility defined <strong>by</strong> (4.2), and the right-hand<br />

graph with the application of the credibility approach calculated in (4.6).<br />

Total Number<br />

0 200 400 600 800 1000<br />

Expected Losses Unexpected Losses<br />

Mean Value 99% quantile<br />

10.000.000 20.000.000 30.000.000 40.000.000 50.000.000<br />

Annual Aggregated Loss Amount (DKK)<br />

Fig 4: The simulated loss distribution based on 20000 simulation summation over all<br />

lines of <strong>operational</strong> <strong>risk</strong>.<br />

imsart ver. 2005/05/19 file: fest_gustafsson.tex date: December 20, 2005

Hooray! Your file is uploaded and ready to be published.

Saved successfully!

Ooh no, something went wrong!