Quantifying operational risk guided by kernel ... - Cass Knowledge
Quantifying operational risk guided by kernel ... - Cass Knowledge
Quantifying operational risk guided by kernel ... - Cass Knowledge
Create successful ePaper yourself
Turn your PDF publications into a flip-book with our unique Google optimized e-Paper software.
<strong>Quantifying</strong> <strong>operational</strong> <strong>risk</strong> <strong>guided</strong> <strong>by</strong><br />
<strong>kernel</strong> smoothing and continuous<br />
credibility<br />
Jim Gustafsson<br />
Codan Insurance<br />
Jens P. Nielsen, Paul Pritchard and Dix Roberts<br />
Royal&SunAlliance<br />
Codan, Gammel Kongevej 60<br />
DK 1790 Copenhagen V,Denmark<br />
e-mail: jgu@codan.dk,e-mail: npj@codan.dk<br />
One Plantation Place 9th Floor<br />
30 Fenchurch Street, London, EC3M 3BD, UK<br />
e-mail: paul.pritchard@gcc.royalsun.com, e-mail: dix.roberts@gcc.royalsun.com<br />
Abstract: The challenge of how much capital is necessary to protect an<br />
organisation against exposure to <strong>operational</strong> <strong>risk</strong> losses underpins this pa-<br />
per (<strong>operational</strong> <strong>risk</strong> itself is defined as the <strong>risk</strong> of loss arising from inad-<br />
equate or failed internal processes, people and systems or from external<br />
events). The evolutionary nature of <strong>operational</strong> <strong>risk</strong> modelling to establish<br />
capital charges is recognised emphasizing the importance of capturing tail<br />
behaviour. Challenges surrounding the quantification of <strong>operational</strong> <strong>risk</strong><br />
particularly those associated with sparse data are addressed with mod-<br />
ern statistical methodology including nonparametric smoothing techniques<br />
with a particular view to comparison with extreme value theory (EVT). The<br />
credibility approach employed supports analysis from pooled data across<br />
business lines on a dataset from an internationally active insurance com-<br />
pany. The approach has the potential to be applied more generally, for<br />
example where data might be pooled across <strong>risk</strong> types or where a combi-<br />
nation of internal company losses and publicly reported (external) data is<br />
1<br />
imsart ver. 2005/05/19 file: fest_gustafsson.tex date: December 20, 2005
Gustafsson, Nielsen, Pritchard and Roberts/<strong>Quantifying</strong> <strong>operational</strong> <strong>risk</strong> 2<br />
used.<br />
AMS 2000 subject classifications: Primary 65D10; secondary 91B30.<br />
Keywords and phrases: Operational <strong>risk</strong>, <strong>kernel</strong> smoothing, credibility<br />
theory, extreme value theory.<br />
imsart ver. 2005/05/19 file: fest_gustafsson.tex date: December 20, 2005
Gustafsson, Nielsen, Pritchard and Roberts/<strong>Quantifying</strong> <strong>operational</strong> <strong>risk</strong> 3<br />
1. Introduction<br />
There is increasing interest in financial services companies in identifying loss dis-<br />
tributions associated with <strong>operational</strong> <strong>risk</strong>s, driven <strong>by</strong> both regulatory consid-<br />
erations and also in recognition of the greater importance placed on <strong>operational</strong><br />
<strong>risk</strong> management. Building on the 1988 Basel Capital Accord, the Basel Com-<br />
mittee on Banking Supervision published; International Convergence of Capital<br />
Measurement and Capital Standards 1 , in June 2004. This document addressed<br />
the challenge of how much capital is necessary to protect an organization against<br />
unexpected losses, and established the need for an explicit charge for the expo-<br />
sure to <strong>operational</strong> <strong>risk</strong> losses. Operational <strong>risk</strong> itself is defined as the <strong>risk</strong> of<br />
loss arising from inadequate or failed internal processes, people and systems or<br />
from external events. The framework also recognized the evolutionary nature of<br />
<strong>operational</strong> <strong>risk</strong> modelling emphasizing the importance of capturing tail events.<br />
Challenges surrounding the quantification of <strong>operational</strong> <strong>risk</strong> are the sub-<br />
ject of this paper and include lack of suitable data and the need to focus on<br />
tail behaviour. These challenges are addressed with modern statistical method-<br />
ologies including nonparametric smoothing techniques. One of the commonly<br />
applied techniques in <strong>operational</strong> <strong>risk</strong> is Extreme Value Theory (EVT) which<br />
offers a broadly accepted methodology for estimating the tail of a distribution,<br />
see Embrecht, Klüppelberg, and Mikosch [9] for a detailed mathematical treat-<br />
ment, also Embrecht [10], Reiss and Thomas [23] and Coles [7]. It is our view<br />
that EVT shows conceptual similarities to other established density estimation<br />
techniques including the approach we consider in this paper. In essence these<br />
techniques seek to estimate a density with particular emphasis on the tail. In<br />
our approach we transform losses into the interval [0,1].<br />
The introduction of the concept of local constant and local linear density<br />
estimators <strong>by</strong> Jones [16] was a very important contribution since both automat-<br />
1 Details at http//www.bis.org/publ/bcbs107.htm.<br />
imsart ver. 2005/05/19 file: fest_gustafsson.tex date: December 20, 2005
Gustafsson, Nielsen, Pritchard and Roberts/<strong>Quantifying</strong> <strong>operational</strong> <strong>risk</strong> 4<br />
ically adjust at boundaries. Jones [16] also pointed out the importance of dirac<br />
functions while dealing with local projections of data. When dealing with re-<br />
gression, local constant and local linear estimators projections can be obtained<br />
relatively easily, see Fan and Gijbels [11] for an overview. However, in other<br />
areas of mathematical statistics such as density estimation and hazard estima-<br />
tion, the dirac function approach becomes important. See Nielsen [18] for the<br />
multivariate density case and Nielsen [19] and Nielsen and Tanggaard [21] for<br />
the hazard case.<br />
The understanding of density and hazard estimation has recently been taken<br />
to a deeper level through the papers of Jiang and Doksum [14] and Jiang and<br />
Doksum [15]. These papers give an overview on how local polynomial densities<br />
and hazards are to be understood as projections even when complicated trun-<br />
cation and censoring is present. In general terms, these papers introduce local<br />
polynomial hazards and densities as simple plug-in estimators with the empir-<br />
ical cumulative densities and hazards being utilized. The asymptotic theory of<br />
the estimators and their derivatives follow as an immediate consequence.<br />
The parametric transformation approach to <strong>kernel</strong> smoothing of Wand, Mar-<br />
ron and Ruppert [25] has recently been considered with a particular interest in<br />
heavy tailed distributions <strong>by</strong> Bolance, Guillen and Nielsen [1], Clements, Hurn<br />
and Lindsay [6] and Buch-Larsen, Nielsen, Bolance and Guillen [2]. We use<br />
the methodology of Buch-Larsen et al. [2] in this paper. They utilize the three<br />
parameter modified Champernowne distribution which not only demonstrates<br />
desired tail behavior, but unlike EVT, is also informed <strong>by</strong> data from the full<br />
distribution.<br />
Application of a non-parametric local constant <strong>kernel</strong> estimator allows an im-<br />
proved fitting to the sample data. Consideration is given here to the issues raised<br />
<strong>by</strong> Diebold, Schuermann and Stroughair [8], who noted that the bias/variance<br />
trade-off in selection of a tail cut-off for EVT applications is analogous to that<br />
relating to the choice of bandwidth in non-parametric density estimation. In<br />
imsart ver. 2005/05/19 file: fest_gustafsson.tex date: December 20, 2005
Gustafsson, Nielsen, Pritchard and Roberts/<strong>Quantifying</strong> <strong>operational</strong> <strong>risk</strong> 5<br />
our approach the established approach to selection of bandwidth in density es-<br />
timation replaces the complex question of selecting the optimal cut off in EVT.<br />
Diebold et al. [8] highlighted the ‘unfortunate’ use of automatic bandwidth se-<br />
lection to support their statement that <strong>kernel</strong> smoothing performs poorly in<br />
the tail (away from the region where most of the data falls), Buch-Larsen et<br />
al. [2] point out however that the transformation approach to <strong>kernel</strong> density<br />
estimation effectively resolves this issue.<br />
In summary this paper is principally concerned with tools with potentially<br />
broad application to <strong>operational</strong> <strong>risk</strong> quantification. These are applied to an op-<br />
erational loss data set obtained from an internationally active insurer. Following<br />
application of non-parametric <strong>kernel</strong> smoothing we apply continuous credibility<br />
theory to density estimation facilitating the appropriate weighting of pooled<br />
(portfolio) losses as compared to data from individual business lines. This theo-<br />
retical approach is inspired <strong>by</strong> the continuous credibility technique that Hardy<br />
and Panjer [12] and Nielsen and Sandqvist [20] introduced to hazard estimation.<br />
The techniques considered in this paper complement and add to existing knowl-<br />
edge demonstrating potential utility in the significant computational challenges<br />
posed <strong>by</strong> <strong>operational</strong> <strong>risk</strong> quantification. Not only is this potentially useful in the<br />
context described here i.e. pooling data across business lines but also in con-<br />
sidering pooling of internal (company) data with that from other companies,<br />
reported publicly or shared through consortia.<br />
Section 2 describes the structure and general assumptions underpinning the<br />
proposed credibility model on densities in the interval [0, 1]. Section 3 is devoted<br />
to the nonparametric <strong>kernel</strong> density estimator, and section 4 presents the esti-<br />
mation of the credibility model. Section 5 is an application to <strong>operational</strong> <strong>risk</strong><br />
data. Here we lay out the transformation process also a plug-in bandwidth is<br />
discussed. The final section evaluates the <strong>operational</strong> <strong>risk</strong> exposure based on a<br />
one year simulation using the loss data.<br />
imsart ver. 2005/05/19 file: fest_gustafsson.tex date: December 20, 2005
Gustafsson, Nielsen, Pritchard and Roberts/<strong>Quantifying</strong> <strong>operational</strong> <strong>risk</strong> 6<br />
2. The credibility model on densities in interval [0,1]<br />
We assume that we have k lines of <strong>operational</strong> <strong>risk</strong> and consider an i belonging<br />
to the set {1, ..., k}. For such an i assume that Xi1, ..., Xini are independent<br />
identically distributed stochastic variables with density fi on [0, 1]. These den-<br />
sities are assumed to be stochastically varying around a common base density<br />
function f. More precisely, we assume that<br />
fi(x) = θi(x)f(x), (2.1)<br />
where x ∈ [0, 1] and θ1(x), ..., θk(x) are identically distributed stochastic pro-<br />
cesses defined on the interval [0, 1]. This model assumption assumes a relation-<br />
ship between the k business lines. Also we assume that � θi(x)f(x)dx = 1,<br />
E(θi(x)) = 1 and V (θi(x)) = ν 2 i<br />
for every i between 1 and k. The assumption<br />
that E{θi(x)} = 1 for all i is crucial for the estimators developed. This assump-<br />
tion implies that for every single x, the business lines varies around the common<br />
base density function f. This assumption allows us to construct our estimators<br />
from a Hilbert space projection of the fi(x) ′ s down at their estimators for each<br />
x individually. The model proposes that each line has <strong>risk</strong> stemming from a<br />
common source as well as its own line of <strong>risk</strong>. The multiplicative construction<br />
above is known from the bias reduction literature, see Hjort and Glad [13] and<br />
Jones, Linton and Nielsen [17]. The difference is, however, that there the pur-<br />
pose is to eliminate bias while here we use the element in the structure as a<br />
stochastic process. The multiplicative structure of our approach is an important<br />
part of our model building. In the bias reduction literature the multiplicative<br />
structure is just used as a trick to reduce bias. However, similarities remain:<br />
we also estimate the multiplicative error nonparametrically and use it for our<br />
estimation purposes - just like one does in the bias reduction literature.<br />
imsart ver. 2005/05/19 file: fest_gustafsson.tex date: December 20, 2005
Gustafsson, Nielsen, Pritchard and Roberts/<strong>Quantifying</strong> <strong>operational</strong> <strong>risk</strong> 7<br />
3. The <strong>kernel</strong> density estimator<br />
Let K denote a probability density function symmetric about zero with support<br />
[−1, 1] and let Kh(x) = K(x/h)/h for any x in [0, 1] and any positive h. We<br />
also define the following functions for the asymptotic properties of the <strong>kernel</strong><br />
density estimator around boundaries.<br />
akl(x, h) =<br />
min{1,x/h} �<br />
max{−1,(x−1)/h}<br />
u k K(u) l du, for x ∈ [0, 1].<br />
Note that a01(x, h) = 1 and a11(x, h) = 0 for the interior points x in the<br />
interval [h, 1 − h]. In the boundary points belonging to the intervals [0, h) and<br />
(1−h, 1], a01(x, h) and a11(x, h) take nontrivial values. The local constant <strong>kernel</strong><br />
estimator 2 of fi is<br />
�fi(x) = (a01(x, hi)ni) −1<br />
ni �<br />
Khi (x − Xij) . (3.1)<br />
The local constant <strong>kernel</strong> estimator of the entire data set can be taken as an<br />
estimator of the common base density f. We assume that the estimation error<br />
of the global <strong>kernel</strong> density estimator<br />
⎛<br />
k�<br />
�f(x) = ⎝(a01(x, hi)ni) −1<br />
⎞<br />
ni �<br />
Khi (x − Xij) ⎠ (3.2)<br />
i=1<br />
is of lower order of magnitude than the estimation error of the individual densi-<br />
ties f1, ..., fk. This is important for our continuous credibility approach defined<br />
in the next section.<br />
4. Estimation of the credibility model<br />
Credibility theory as known from actuarial science is simply a methodology to<br />
find out how much information a common model carries onto a specific line of<br />
2 See Jones [16] for a definition of local costant and local linear density estimators and for<br />
j=1<br />
j=1<br />
a precision of the automatic boundary corrections of these estimators.<br />
imsart ver. 2005/05/19 file: fest_gustafsson.tex date: December 20, 2005
Gustafsson, Nielsen, Pritchard and Roberts/<strong>Quantifying</strong> <strong>operational</strong> <strong>risk</strong> 8<br />
business. Credibility theory tells us that we can rely more on data from one<br />
specific line when it is plentiful, when data for one line is sparse more weight<br />
should be put on a common source. In our case we fix x ∈ [0, 1] and define a<br />
Hilbert space projection, � fi(x), of fi(x) defined <strong>by</strong> (2.1), onto the linear space<br />
{ax + bx � fi(x) | ax, bx ∈ [0, 1]}. We get<br />
�fi(x) = (1 − zi,x) E( � fi(x)) + zi,x � fi(x), (4.1)<br />
where the credibility factor zi,x = COV (fi(x), � fi(x))/V ( � fi(x)) is between zero<br />
and one. This projection gives us the optimal linear credibility estimator � fi(x)<br />
minimizing E((fi(x) − � fi(x)) 2 ), see also Hardy and Panjer [12] and Nielsen and<br />
Sandqvist [20].<br />
The credibility factor quantifies the amount of weight the common mean<br />
and the individual mean should have for each of the lines of <strong>operational</strong> <strong>risk</strong>.<br />
For the original approach to credibility theory, see Bühlmann and Straub [3].<br />
For a recent overview of the origins of credibility theory and a Hilbert space<br />
interpretation of the Bühlmann and Straub model, see Norberg [22]. To calculate<br />
the three moment quantities of (4.1), we prove in the Appendix that<br />
and<br />
V ( � fi(x)) =<br />
E( � fi(x)) = f(x) (1 + op(1)) ,<br />
COV (fi(x), � fi(x)) = f(x) 2 ν 2 i (1 + op(1))<br />
�<br />
f(x) 2 ν 2 i + a02(x, hi)f(x)<br />
a01(x, hi) 2 �<br />
(1 + op(1)) .<br />
hini<br />
From these moment expressions we can approximate the optimal credibility<br />
estimator of (4.1) to<br />
where<br />
zi,x =<br />
�fi(x) = (1 − zi,x) f(x) + zi,x � fi(x),<br />
f(x) 2 ν 2 i a01(x, hi) 2 hini<br />
f(x) 2 ν 2 i a01(x, hi) 2 hini + f(x)a02(x, hi) (1 + op(1)) .<br />
imsart ver. 2005/05/19 file: fest_gustafsson.tex date: December 20, 2005
Gustafsson, Nielsen, Pritchard and Roberts/<strong>Quantifying</strong> <strong>operational</strong> <strong>risk</strong> 9<br />
To estimate the local variance ν 2 i of the stochastic process we note that � fi(x) is<br />
an estimator of fi(x) = θi(x)f(x) and<br />
�θi(x) = � fi(x)<br />
�f(x)<br />
(4.2)<br />
is an estimator of θi(x). Therefore, for a given x, we estimate the variance of<br />
θi(x) <strong>by</strong><br />
�ν 2 i = 1<br />
ni �<br />
ni<br />
j=1<br />
� �2 �θi(xj) − 1 =<br />
� 1<br />
0<br />
� �2 �θi(x) − 1 d � Fi(x), (4.3)<br />
where � Fi is the empirical distribution function of the i’th business line.<br />
We can here<strong>by</strong> estimate zi,x <strong>by</strong><br />
�zi,x =<br />
�f(x) 2�ν 2 i a01(x, hi) 2hini �f(x) 2�ν 2 i a01(x, hi) 2hini + � . (4.4)<br />
f(x)a02(x, hi)<br />
Combining (3.1), (3.2) and (4.4) the final expression for the optimal credibility<br />
estimator is approximately equal to<br />
� �f i(x) = (1 − �zi,x) � f(x) + �zi,x � fi(x). (4.5)<br />
To arrive at a credibility based estimator of θi(x) we use the above equality<br />
(4.5) divided <strong>by</strong> � f(x)<br />
� �θi(x) = (1 − �zi,x) + �zi,x<br />
�fi(x)<br />
�f(x) = 1 − �zi,x(1 − � θi(x)). (4.6)<br />
We propose here that each line has <strong>risk</strong> stemming from a common source as well<br />
as its own individual source. This property of our basic model is clearly reflected<br />
in the expression of our final estimators � � f i(x) and � � θi(x); both are constructed<br />
as a sum of one common and one global element.<br />
5. Application to <strong>operational</strong> <strong>risk</strong> data<br />
We utilize data from seven lines of <strong>operational</strong> <strong>risk</strong> with observed number of<br />
losses N1, ..., N7. The data periods in these seven groups measured in years<br />
are T1, ..., T7. We assume that we have seven independent homogenous Poisson<br />
imsart ver. 2005/05/19 file: fest_gustafsson.tex date: December 20, 2005
Gustafsson, Nielsen, Pritchard and Roberts/<strong>Quantifying</strong> <strong>operational</strong> <strong>risk</strong> 10<br />
processes describing the number of losses in each line and therefore obtain the<br />
following maximum likelihood estimator of the intensity of losses in the i th<br />
business line: � λi = Ni/Ti. Conditional on N1 = n1, ..., N7 = n7, we now have<br />
our <strong>operational</strong> <strong>risk</strong> data for each group, Yi1, ..., Yini , defined on the positive<br />
real axis. We now transform our data from the positive real axis to the interval<br />
[0, 1] using the modified Champernowne distribution defined in Buch-Larsen et<br />
al. [2]:<br />
Fp(y) =<br />
(y + c) α − cα (y + c) α + (M + c) α , y ∈ R+,<br />
− 2cα where p = {α, M, c} is a parameter vector. For c = 0 this distribution is a<br />
special case of the parametric distribution suggested <strong>by</strong> Champernowne [4] and<br />
Champernowne [5] and the cumulative density function equals the transforma-<br />
tion used <strong>by</strong> Clements et al. [6] in their approach to density estimation based<br />
on transformed data. However, an extensive simulation study <strong>by</strong> Buch-Larsen<br />
et al. [2] show that the flexibility of the modified Champernowne distribution<br />
outweigh the advantages of the stability obtained in the simple case of c = 0.<br />
Buch-Larsen et al. [2] estimate M, the median, from the empirical median and<br />
the parameters (α, c) <strong>by</strong> the maximum likelihood method. We use the modified<br />
Champernowne distribution for the i th line of business as follows<br />
Fpi(y) =<br />
(y + c) α − c α<br />
(y + c) α + (Mi + c) α − 2c α , y ∈ R+, (5.1)<br />
where pi = {α, Mi, c} is the extended parameter vector for the individual line<br />
of business. We use the same values of the variables (α, c) for the seven lines of<br />
<strong>operational</strong> <strong>risk</strong> while the median, a scaling parameter, is estimated individually<br />
for each line. This parametric model is the common source of information for<br />
our <strong>operational</strong> <strong>risk</strong> modelling. Firstly we transform the data from each line into<br />
the interval [0, 1] using this parametric model. That is, we define<br />
Xij = F�pi (Yij), (5.2)<br />
imsart ver. 2005/05/19 file: fest_gustafsson.tex date: December 20, 2005
Gustafsson, Nielsen, Pritchard and Roberts/<strong>Quantifying</strong> <strong>operational</strong> <strong>risk</strong> 11<br />
as the transformed losses obtained <strong>by</strong> the cumulative modified Champernowne<br />
distribution, where �pi = {�α, � Mi, �c} are the estimated parameter vector for<br />
i ∈ {1, . . . , k} on observations j ∈ {1, . . . , ni}. Now we are ready to apply<br />
the methodology outlined in sections 2 and 3 using the transformed data set<br />
Xij. Note that if we had only one line of business, then the credibility approach<br />
would of course be superfluous and we have the exact same one-dimensional<br />
estimation method as the one suggested in Buch-Larsen et al. [2]. When esti-<br />
mating the initial individual <strong>kernel</strong> estimator of the i th line of business, defined<br />
<strong>by</strong> (3.1), we use the Epanechnikov <strong>kernel</strong> function:<br />
K(u) = 3 � 2<br />
1 − u<br />
4<br />
� 1 {|u|
Gustafsson, Nielsen, Pritchard and Roberts/<strong>Quantifying</strong> <strong>operational</strong> <strong>risk</strong> 12<br />
Consequently, if we differentiate this with respect to the bandwidth hi, we obtain<br />
the theoretical optimal choice<br />
�<br />
a02(x, hi)<br />
hi =<br />
a21(x, hi) 2 �1/5 � .<br />
ni f ′′ (x) 2dx Here everything is known except for the density f. Assuming the unknown f<br />
follows a normal distribution, we obtain Silverman’s rule of thumb (Silverman<br />
[24]):<br />
� √ �1/5 40 π<br />
hi =<br />
�σi. (5.3)<br />
ni<br />
The following section shows a summary of the <strong>operational</strong> loss data set fol-<br />
lowed <strong>by</strong> a visual comparison applied to the data.<br />
Table 1: Summary Statistics- <strong>operational</strong> loss data set.<br />
Line of<br />
Business ni Ti hi mean(Yij) median(Yij) sd(Yij) max(Yij)<br />
1 250 3 0.24 8981 4059.5 16490 163233<br />
2 46 3 0.36 13450 2829.5 58512 394969<br />
3 924 3 0.16 7835 2990 36164 874400<br />
4 34 3 0.47 13963 2277 18467 52200<br />
5 7 3 0.48 19633 14610 14621 38205<br />
6 7 3 0.58 959382 197600 2123445 5765217<br />
7 23 2 0.44 1127077 242220 2333035 8477700<br />
The second column shows the number of observations ni for each line of<br />
business. There is considerable variation in the number of recorded losses, line<br />
of business 3 has 925 losses compared to line 5 and 6 which only have been<br />
exposed to 7 losses each. The third column gives the time over which data<br />
was collected, in years. The fourth column shows the estimated bandwidths,<br />
obtained <strong>by</strong> (5.3). Note that the estimated bandwidths decrease with sample<br />
size as expected. Columns five to eight show some empirical results on each line<br />
imsart ver. 2005/05/19 file: fest_gustafsson.tex date: December 20, 2005
Gustafsson, Nielsen, Pritchard and Roberts/<strong>Quantifying</strong> <strong>operational</strong> <strong>risk</strong> 13<br />
of business. Note that the mean is significantly larger than the median in all<br />
cases, consistent with right skewed distributions.<br />
Figure 1 shows histograms of both the transformed global data set and the<br />
seven individual data sets. In each graph the respective semiparametric estima-<br />
tors are estimated. If we examine the global estimator � f(x) more closely (top-left<br />
graph), we observe that in the interval [0, 0.3) the estimator is less than one.<br />
This means that the <strong>kernel</strong> estimator corrects the parametric density where it<br />
has diverged from the <strong>operational</strong> <strong>risk</strong> data. In [0.3, 0.7) the <strong>kernel</strong> function<br />
corrects the parametric density where it was too low. In [0.7, 1] the estimator is<br />
once again below one and therefore adjustment is made, equivalent to a lighter<br />
tail. The other individual estimators � fi(x), i = 1, ...7, should be interpreted<br />
analogously.<br />
Insert Figure 1 about here<br />
In Figure 2 we show the credibility approach applied to each line. Here we<br />
omit the histograms to allow easier visual interpretation. We compare the stan-<br />
dard semiparametric estimator (3.1), show with a dotted curve, with an es-<br />
timator where we have applied the credibility approach (solid curve), defined<br />
through (4.5). Note that for lines of <strong>operational</strong> <strong>risk</strong> with sparse information the<br />
appearance of the credibility estimator is similar to the global source. This is be-<br />
cause the global source provides more weight for lines with sparse information,<br />
particular lines 5, 6 and 7.<br />
Insert Figure 2 about here<br />
Figure 3 shows the estimated stochastic processes with and without the appli-<br />
cation of a credibility approach. The left-hand graph shows estimated stochastic<br />
processes for all lines estimated through formula (4.2). Here we can see that the<br />
processes take values between 0.5 and 1.5. This means that when a line of<br />
business deviates from the horizontal we know that there is a large difference<br />
between the specific line and the global source. The right-hand graph shows<br />
imsart ver. 2005/05/19 file: fest_gustafsson.tex date: December 20, 2005
Gustafsson, Nielsen, Pritchard and Roberts/<strong>Quantifying</strong> <strong>operational</strong> <strong>risk</strong> 14<br />
the estimated stochastic processes applying the credibility approach, calculated<br />
using (4.6). The most obvious difference is that the lines are much closer to the<br />
horizontal. This is, of course, to be expected since, to a greater or lesser degree,<br />
the individual lines will use information from the global source.<br />
Insert Figure 3 about here<br />
6. Evaluating an <strong>operational</strong> loss distribution<br />
In this section we simulate an <strong>operational</strong> <strong>risk</strong> loss distribution for the company.<br />
This is done <strong>by</strong> using the severity distribution obtained from (4.5), together with<br />
a Poisson based frequency distribution. The frequency and severity distributions<br />
are used to create simulated one year loss distribution through Monte Carlo<br />
analysis. We sample from a Poisson process of event times through all lines of<br />
<strong>operational</strong> <strong>risk</strong>, and combine with loss sizes taken form the relevant severity<br />
distribution. To obtain our original scale, we transform the estimator (4.5) to<br />
its original axis, The relevant estimators takes the form<br />
ni �<br />
�fi(y)<br />
−1<br />
= (a01(F�pi (y), hi)ni)<br />
�f(y) =<br />
j=1<br />
⎛<br />
k�<br />
ni �<br />
⎝(a01(F�pi<br />
−1<br />
(y), hi)ni)<br />
i=1<br />
Khi (F�pi (y) − F�pi (Yij)) f�pi (y) (6.1)<br />
j=1<br />
Khi (F�pi (y) − F�pi (Yij)) f�pi (y)<br />
where f�pi (y) is the modified Champernowne density defined <strong>by</strong><br />
f�pi (y) = α(y + c)α−1 ((Mi + c) α − cα )<br />
((y + c) α + (Mi + c) α − 2cα 2 , y ∈ R+.<br />
)<br />
The loss sizes are then taken from the severity distribution<br />
� ∞<br />
�<br />
Hi(y) = �f i(ξ)dξ,<br />
0<br />
where � � f i are defined through the estimators in (6.1) with structure as (4.6).<br />
Summation of amounts from all lines provides a single estimate, this process is<br />
then repeated 20000 times to create a simulated loss distribution as shown in<br />
imsart ver. 2005/05/19 file: fest_gustafsson.tex date: December 20, 2005<br />
⎞<br />
⎠
Gustafsson, Nielsen, Pritchard and Roberts/<strong>Quantifying</strong> <strong>operational</strong> <strong>risk</strong> 15<br />
Figure 4. From this we can identify the loss amount associated with the relevant<br />
quantile and thus the capital to be held.<br />
Insert Figure 4 about here<br />
In table 2 we present summary statistics for the individual loss distribution<br />
and for the global source.<br />
Table 2: Summary statistics for simulated loss distribution<br />
<strong>by</strong> business line and global source.<br />
Line of<br />
Business x0.5 OpVaR0.95 OpVaR0.99 OpVaR0.999<br />
1 1840605 372988 528891 747380<br />
2 238045 138215 212161 294414<br />
3 5657744 1124957 1690317 1897271<br />
4 156984 118278 180286 258575<br />
5 154260 342791 594155 843894<br />
6 2088010 4865225 10405343 15955968<br />
7 9856090 8621910 13031243 19455952<br />
Global 20706695 9675355 14890803 22462507<br />
The Operational Value-at-Risk (OpVaR) from table 2 is a measure of the<br />
unexpected losses to a specific quantile level. To obtain this value we subtract<br />
the median (presented as column 1) from three different upper quantiles with<br />
confidence levels 95%, 99% and 99.9%. Table 2 shows that the appearance in<br />
the tail region varies between the lines of business. More precisely, line 1 and 3<br />
generate a light tailed distribution while line of business 2, 4 and 7 are charac-<br />
terized <strong>by</strong> moderate tails. The remaining two lines of business 5 and 6 are heavy<br />
tailed distributed. This is not extraordinary since both lines are embodied with<br />
extremely small samples.<br />
Table 3 shows measures of the length of the tail. This is done <strong>by</strong> dividing<br />
the upper quantile (with one of the three confidence levels) <strong>by</strong> the median, i.e.<br />
imsart ver. 2005/05/19 file: fest_gustafsson.tex date: December 20, 2005
Gustafsson, Nielsen, Pritchard and Roberts/<strong>Quantifying</strong> <strong>operational</strong> <strong>risk</strong> 16<br />
ϕ (q) = xq/x0.5 , q = 0.95, 0.99 or 0.999. Here, as in table 2 lines of business 5<br />
and 6 stand out being most heavy tailed.<br />
Table 3: Tail length characteristics.<br />
Line of<br />
Business ϕ (0.95) ϕ (0.99) ϕ (0.999)<br />
1 1.203 1.287 1.406<br />
2 1.581 1.891 2.237<br />
3 1.199 1.299 1.335<br />
4 1.753 2.148 2.647<br />
5 3.222 4.852 6.471<br />
6 3.330 5.983 8.642<br />
7 1.875 2.320 2.974<br />
Global 1.467 1.719 2.085<br />
In this paper we have demonstrated that a number of statistical approaches<br />
(complementary to those commonly applied such as EVT) can provide signif-<br />
icant benefit in the quantification of <strong>operational</strong> <strong>risk</strong> for estimation of capital<br />
requirements in organizations. The fundamental challenge, that of sparse data,<br />
is addressed <strong>by</strong> transforming available data and then applying non-parametric<br />
<strong>kernel</strong> smoothing. This approach (even without further application of credibility<br />
analysis) should assist in maximizing useful information from limited data sets,<br />
particularly as the direct application of non parametric techniques would not<br />
yield useful information. The subsequent credibility element could also be ap-<br />
plied in isolation when considering data that is pooled (e.g. across businesses or<br />
<strong>risk</strong> types) or for the mixing of internal and external data (e.g. from a subscrip-<br />
tion database). We acknowledge the limitations inherent in this work, specifically<br />
that the data set used does not give full coverage across all <strong>operational</strong> <strong>risk</strong> cat-<br />
egories and that it is based on a limited collection period. Nevertheless for the<br />
purposes of illustration it does demonstrate that the techniques can be applied<br />
imsart ver. 2005/05/19 file: fest_gustafsson.tex date: December 20, 2005
Gustafsson, Nielsen, Pritchard and Roberts/<strong>Quantifying</strong> <strong>operational</strong> <strong>risk</strong> 17<br />
to real life company data, furthermore that considerable potential exists for its<br />
further application in this area.<br />
7. Appendix<br />
Through standard <strong>kernel</strong> smoothing theory we obtain<br />
and<br />
� �<br />
E �fi(x) | θi(x)<br />
� �<br />
E �fi(x)<br />
= θi(x)f(x) (1 + op(1))<br />
� � ��<br />
= E E �fi(x) | θi(x)<br />
= E (θi(x)f(x)) (1 + op(1))<br />
= f(x) (1 + o(1)) .<br />
To derive the variance expressions we first note that<br />
V<br />
This implies that<br />
and<br />
V<br />
� �<br />
�fi(x)<br />
� �<br />
�fi(x) | θi(x)<br />
= V<br />
=<br />
= θi(x)a02(x,hi)f(x)<br />
a01(x,hi) 2 hini<br />
� � ��<br />
E �fi(x) | θi(x)<br />
�<br />
f(x) 2 ν 2 i<br />
+ a02(x,hi)f(x)<br />
a01(x,hi) 2 hini<br />
�<br />
+ E<br />
(1 + op(1)) .<br />
V<br />
� ��<br />
�fi(x) | θi(x)<br />
�<br />
(1 + op(1)) .<br />
�<br />
COV fi(x), � �<br />
fi(x)<br />
�<br />
� ��<br />
= COV E (fi(x) | θi(x)) , E �fi(x) | θi(x)<br />
� �<br />
+E COV fi(x), � ��<br />
fi(x) | θi(x)<br />
� �<br />
= E COV fi(x), � ��<br />
fi(x) | θi(x)<br />
= � E � f(x) 2 θi(x) 2� − f(x) 2� (1 + op(1))<br />
= � f(x) 2 E � θi(x) 2 − 1 �� (1 + op(1))<br />
= f(x) 2 ν 2 i (1 + op(1)) .<br />
imsart ver. 2005/05/19 file: fest_gustafsson.tex date: December 20, 2005
References<br />
Gustafsson, Nielsen, Pritchard and Roberts/<strong>Quantifying</strong> <strong>operational</strong> <strong>risk</strong> 18<br />
[1] Bolance, C., Guillen, M. and Nielsen, J.P. (2003). Kernel density<br />
estimation of actuarial loss functions. Insurance: Mathematics and Eco-<br />
nomics, 32, 19-36.<br />
[2] Buch-Larsen, T., Nielsen, J.P., Guillen, M. and Bolance, C.<br />
(2005). Kernel density estimation for heavy tailed distributions using the<br />
Champernowne distribution. Manuscript Universidad de Barcelona (to ap-<br />
pear).<br />
[3] Bühlmann, H and Straub, E. (1970). Glabwürdigkeit für schadensätze.<br />
Bulletin of the Association of Swiss Actuaries 70, 111-133.<br />
[4] Champernowne D.G. (1936). The Oxford meeting, September 25-29,<br />
1936, Econometrica, Vol. 5, No. 4, October 1937.<br />
[5] Champernowne D.G. (1952). The graduation of income distributions.<br />
Econometrica, 20, 591-615.<br />
[6] Clements, A.E., Hurn, A.S. and Lindsay, K.A. (2003). Möbius-like<br />
mappings and their use in <strong>kernel</strong> density estimation. Journal of the Amer-<br />
ical Statistical Association 98, 993-1000.<br />
[7] Coles, S. (2001). An introduction to statistical modeling of Extreme Val-<br />
ues. Springer.<br />
[8] Diebold, F.X., Schuermann, T. and Stroughair,J. (2000). Pitfalls<br />
and Opportunities in the Use of Extreme Value Theory in Risk Manage-<br />
ment. London: Extremes and Integrated Risk Management London: Risk<br />
Books.<br />
[9] Embrechts, P., Klüppelberg, C. and Mikosch, T. (1999). Modeling<br />
Extremal Events for Insurance and Finance. Springer.<br />
[10] Embrechts, P. (2000). Extremes and Integrated Risk Management. Lon-<br />
don: Risk Books, Risk Waters Group.<br />
[11] Fan, J. and Gijbels, I. (1996). Local Polynomial Modelling and its Ap-<br />
imsart ver. 2005/05/19 file: fest_gustafsson.tex date: December 20, 2005
Gustafsson, Nielsen, Pritchard and Roberts/<strong>Quantifying</strong> <strong>operational</strong> <strong>risk</strong> 19<br />
plications. Chapman and Hall, London.<br />
[12] Hardy, M.R. and Panjer, H. (1998). A credibility approach to mor-<br />
tality <strong>risk</strong>. ASTIN Bulletin 28, 2, 269-283.<br />
[13] Hjort, N.L. and Jones, M.C. (1995). Nonparametric density estima-<br />
tion with a parametric start. The Annals of Statistics, Vol. 23, No. 3,<br />
882-904.<br />
[14] Jiang, J. and Doksum, K. (2003a). On local polynomial estimation of<br />
hazard rates and their derivatives under random censoring, Constance van<br />
Eeden Volume, IMS, 463-481.<br />
[15] Jiang, J. and Doksum, K. (2003b). Empirical plug-in curve and surface<br />
estimates. In: Mathematical and Statistical Methods in Reliability. Series<br />
on Quality, Reliability and Engineering Statistics, Vol. 7, World Scientific<br />
Publishing, Singapore, 433-453.<br />
[16] Jones, M.C. (1993). Simple boundary correction for <strong>kernel</strong> density esti-<br />
mation. Statistics and Computing 3, 135 - 146.<br />
[17] Jones, M.C., Linton O. and Nielsen, J.P. (1995). A simple bias<br />
reduction method for density estimation. Biometrica, Vol. 82, No. 2, 327-<br />
338.<br />
[18] Nielsen, J.P. (1998). Marker dependent <strong>kernel</strong> hazard estimation from<br />
local linear estimation, Scand. Actuar. J. 2 , 113-124<br />
[19] Nielsen, J.P. (1999). Multivariate <strong>kernel</strong>s from local linear estimation,<br />
Scand. Actuar. J. 1 , 93-95.<br />
[20] Nielsen, J.P. and Sandqvist, B.L. (2000). Credibility weighted hazard<br />
estimation. Astin Bulletin 30, 405-417.<br />
[21] Nielsen, J.P. and Tanggaard, C. (2001). Boundary and bias correc-<br />
tion in <strong>kernel</strong> hazard estimation. Scandinavian Journal of Statistics, 28,<br />
675-698.<br />
[22] Norberg, R (2004). Credibility theory. In: Encyclopedia of Actuarial<br />
Science, Wiley.<br />
imsart ver. 2005/05/19 file: fest_gustafsson.tex date: December 20, 2005
Gustafsson, Nielsen, Pritchard and Roberts/<strong>Quantifying</strong> <strong>operational</strong> <strong>risk</strong> 20<br />
[23] Reiss, R.D. and Thomas J.A. (2001). Statistical Analysis of Extreme<br />
Values. Basel.<br />
[24] Silverman, B.W. (1986). Density estimation for statistics and data anal-<br />
ysis. Chapman and Hall.<br />
[25] Wand, P., Marron, J.S. and Ruppert, D. (1991). Transformations in<br />
Density Estimation. Journal of the Americal Statistical Association, 30,<br />
405-417.<br />
imsart ver. 2005/05/19 file: fest_gustafsson.tex date: December 20, 2005
0.0 0.2 0.4 0.6 0.8 1.0 1.2<br />
0.0 0.5 1.0 1.5 2.0 2.5 3.0 3.5<br />
Gustafsson, Nielsen, Pritchard and Roberts/<strong>Quantifying</strong> <strong>operational</strong> <strong>risk</strong> 21<br />
Global<br />
0.0 0.4 0.8<br />
Line 4<br />
0.0 0.4 0.8<br />
0.0 0.5 1.0 1.5<br />
0.0 0.5 1.0 1.5 2.0 2.5<br />
Line 1<br />
0.0 0.4 0.8<br />
Line 5<br />
0.0 0.4 0.8<br />
0.0 0.5 1.0 1.5 2.0<br />
0.0 0.5 1.0 1.5 2.0 2.5<br />
Line 2<br />
0.0 0.4 0.8<br />
Line 6<br />
0.0 0.4 0.8<br />
0.0 0.2 0.4 0.6 0.8 1.0 1.2<br />
0.0 0.5 1.0 1.5 2.0<br />
Line 3<br />
0.0 0.4 0.8<br />
Line 7<br />
0.0 0.4 0.8<br />
Fig 1: Histogram of data,including the global semiparametric estimator (3.2) in the<br />
top-left graph and the individual semiparametric estimators defined <strong>by</strong> (3.1)<br />
in the remaining graphs.<br />
0.6 0.8 1.0 1.2 1.4 1.6<br />
0.6 0.8 1.0 1.2 1.4 1.6<br />
Global<br />
0.0 0.4 0.8<br />
Line 4<br />
0.0 0.4 0.8<br />
0.6 0.8 1.0 1.2 1.4 1.6<br />
0.6 0.8 1.0 1.2 1.4 1.6<br />
Line 1<br />
0.0 0.4 0.8<br />
Line 5<br />
0.0 0.4 0.8<br />
0.6 0.8 1.0 1.2 1.4 1.6<br />
0.6 0.8 1.0 1.2 1.4 1.6<br />
Line 2<br />
0.0 0.4 0.8<br />
Line 6<br />
0.0 0.4 0.8<br />
0.6 0.8 1.0 1.2 1.4 1.6<br />
0.6 0.8 1.0 1.2 1.4 1.6<br />
Line 3<br />
0.0 0.4 0.8<br />
Line 7<br />
0.0 0.4 0.8<br />
Fig 2: The severity densities for all lines of <strong>operational</strong> <strong>risk</strong> including also the global<br />
source (3.2) <strong>by</strong> top-left graph. Dotted curve represent standard<br />
semiparametric approach (3.1), while the solid curve is estimators with an<br />
credibility approach (4.5)<br />
imsart ver. 2005/05/19 file: fest_gustafsson.tex date: December 20, 2005
0.5 1.0 1.5<br />
Gustafsson, Nielsen, Pritchard and Roberts/<strong>Quantifying</strong> <strong>operational</strong> <strong>risk</strong> 22<br />
Line 1<br />
Line 2<br />
Line 3<br />
Line 4<br />
Line 5<br />
Line 6<br />
Line 7<br />
0.0 0.2 0.4 0.6 0.8 1.0<br />
0.5 1.0 1.5<br />
Line 1<br />
Line 2<br />
Line 3<br />
Line 4<br />
Line 5<br />
Line 6<br />
Line 7<br />
0.0 0.2 0.4 0.6 0.8 1.0<br />
Fig 3: The stochastic processes for all lines of <strong>operational</strong> <strong>risk</strong>. The left-hand graph<br />
presents processes without credibility defined <strong>by</strong> (4.2), and the right-hand<br />
graph with the application of the credibility approach calculated in (4.6).<br />
Total Number<br />
0 200 400 600 800 1000<br />
Expected Losses Unexpected Losses<br />
Mean Value 99% quantile<br />
10.000.000 20.000.000 30.000.000 40.000.000 50.000.000<br />
Annual Aggregated Loss Amount (DKK)<br />
Fig 4: The simulated loss distribution based on 20000 simulation summation over all<br />
lines of <strong>operational</strong> <strong>risk</strong>.<br />
imsart ver. 2005/05/19 file: fest_gustafsson.tex date: December 20, 2005