Estimation, Evaluation, and Selection of Actuarial Models

Estimation, Evaluation, and Selection of Actuarial Models Estimation, Evaluation, and Selection of Actuarial Models

01.08.2014 Views

58 CHAPTER 3. SAMPLING PROPERTIES OF ESTIMATORS Changing the signs and inverting produces the same values as in (3.4). This is a feature of the lognormal distribution that need not hold for other models. ¤ Sometimes it is not even possible to take the derivative. In that case an approximate second derivative can be used. A reasonable approximation is ∂ 2 f(θ) . 1 = [f(θ + h i e i /2+h j e j /2) − f(θ + h i e i /2 − h j e j /2) ∂θ i ∂θ j h i h j −f(θ − h i e i /2+h j e j /2) + f(θ − h i e i /2 − h j e j /2)] where e i is a vector with all zeroes except for a one in the ith position and h i = θ i /10 v where v is one-third the number of significant digits used in calculations. Example 3.32 Repeat the previous example, using approximate derivatives. Assume that there are 15 significant digits being used. Then h 1 = 6.1379/10 5 and h 2 = 1.3894/10 5 . Reasonably close values are 0.00006 and 0.00001. The first approximation is ∂ 2 l ∂µ 2 . = l(6.13796, 1.3894) − 2l(6.1379, 1.3894) + l(6.13784, 1.3894) (0.00006) 2 = −157.713893081982 − 2(−157.713893049683) + (−157.713893054682) (0.00006) 2 = −10.3604. The other two approximations are ∂ 2 l/∂σ∂µ = . 0.0003 ∂ 2 l/∂σ 2 . = −20.7208. We see that here the approximation works very well. ¤ For grouped data, the process works the same way. In Loss Models on pages 59 and 66 it is noted that the information can be obtained using only first partial derivatives. These still may need to be numerically evaluated. Exercise 67 Determine 95% confidence intervals for the parameters of exponential and gamma models for Data Set B. The likelihood function and maximum likelihood estimates were determined in Example 2.32. Exercise 68 Let X have a uniform distribution on the interval from 0 to θ. Show that the maximum likelihood estimator is ˆθ =max(X 1 ,...,X n ). Use Examples 3.7 and 3.10 to show that this estimator is asymptotically unbiased and to obtain its variance. Show that Theorem 3.29 yields a negative estimate of the variance and that item (ii) in the conditions does not hold. The information matrix provides a method for assessing the quality of the maximum likelihood estimators of a distribution’s parameters. However, we are often more interested in a quantity that

3.3. VARIANCE AND CONFIDENCE INTERVALS 59 is a function of the parameters. For example, we might be interested in the lognormal mean as an estimate of the population mean. That is, we want to use exp(ˆµ +ˆσ 2 /2) as an estimate of the population mean, where the maximum likelihood estimates of the parameters are used. It is very difficult to evaluate the mean and variance of this random variable because it is a complex function of two variables that already have complex distributions. The following theorem (from Linear Statistical Inference and its Applications by C. R. Rao (Wiley, 1965)) can help. The method is often called the delta method. Theorem 3.33 Let X n =(X 1n ,...,X kn ) T be a multivariate random variable of dimension k based on a sample of size n. Assume that X n is asymptotically normal with mean θ and covariance matrix Σ/n, where neither θ nor Σ depend on n. Letg be a function of k variables that is totally differentiable. Let G n = g(X 1n ,...,X kn ).ThenG n is asymptotically normal with mean g(θ) and variance (∂g) T Σ(∂g)/n, where∂g is the vector of first derivatives, that is, ∂g =(∂g/∂θ 1 ,...,∂g/∂θ k ) T and it is to be evaluated at θ, the true parameters of the original random variable. ¤ The statement of the Theorem is hard to decipher. The Xs are the estimators and g is the function of the parameters that are being estimated. For a model with one parameter, the Theorem reduces to the following statement: Let ˆθ be an estimator of θ that has an asymptotic normal distribution with mean θ and variance σ 2 /n. Then g(ˆθ) has an asymptotic normal distribution with mean g(θ) and asymptotic variance [g 0 (θ)](σ 2 /n)[g 0 (θ)] = g 0 (θ) 2 σ 2 /n. Example 3.34 Use the delta method to approximate the variance of the maximum likelihood estimator of the probability that an observation from an exponential distribution exceeds 200. Apply this result to Data Set B. From Example 2.32 we know that the maximum likelihood estimate of the exponential parameter is the sample mean. We are asked to estimate p =Pr(X>200) = exp(−200/θ). The maximum likelihood estimate is ˆp =exp(−200/ˆθ) =exp(−200/¯x). Determining the mean and variance of this quantity is not easy. But we do know that that Var( ¯X) =Var(X)/n = θ 2 /n. Furthermore, and therefore the delta method gives For Data Set B, g(θ) = e −200/θ g 0 (θ) = 200θ −2 e −200/θ Var(ˆp) . = [200θ −2 e −200/θ ] 2 θ 2 /n =40, 000θ −2 e −400/θ /n. ¯x = 1, 424.4 ˆp = exp(−200/1, 424.4) = 0.86900 dVar(ˆp) = 40, 000(1, 424.4) −2 exp(−400/1, 424.4)/20 = 0.0007444. A 95% confidence interval for p is 0.869 ± 1.96 √ 0.0007444 or 0.869 ± 0.053. ¤ Example 3.35 Construct a 95% confidence interval for the mean of a lognormal population using Data Set B. Compare this to the more traditional confidence interval based on the sample mean.

3.3. VARIANCE AND CONFIDENCE INTERVALS 59<br />

is a function <strong>of</strong> the parameters. For example, we might be interested in the lognormal mean as<br />

an estimate <strong>of</strong> the population mean. That is, we want to use exp(ˆµ +ˆσ 2 /2) as an estimate <strong>of</strong><br />

the population mean, where the maximum likelihood estimates <strong>of</strong> the parameters are used. It is<br />

very difficult to evaluate the mean <strong>and</strong> variance <strong>of</strong> this r<strong>and</strong>om variable because it is a complex<br />

function <strong>of</strong> two variables that already have complex distributions. The following theorem (from<br />

Linear Statistical Inference <strong>and</strong> its Applications by C. R. Rao (Wiley, 1965)) can help. The method<br />

is <strong>of</strong>ten called the delta method.<br />

Theorem 3.33 Let X n =(X 1n ,...,X kn ) T be a multivariate r<strong>and</strong>om variable <strong>of</strong> dimension k based<br />

on a sample <strong>of</strong> size n. Assume that X n is asymptotically normal with mean θ <strong>and</strong> covariance matrix<br />

Σ/n, where neither θ nor Σ depend on n. Letg be a function <strong>of</strong> k variables that is totally differentiable.<br />

Let G n = g(X 1n ,...,X kn ).ThenG n is asymptotically normal with mean g(θ) <strong>and</strong> variance<br />

(∂g) T Σ(∂g)/n, where∂g is the vector <strong>of</strong> first derivatives, that is, ∂g =(∂g/∂θ 1 ,...,∂g/∂θ k ) T <strong>and</strong><br />

it is to be evaluated at θ, the true parameters <strong>of</strong> the original r<strong>and</strong>om variable. ¤<br />

The statement <strong>of</strong> the Theorem is hard to decipher. The Xs are the estimators <strong>and</strong> g is the<br />

function <strong>of</strong> the parameters that are being estimated. For a model with one parameter, the Theorem<br />

reduces to the following statement: Let ˆθ be an estimator <strong>of</strong> θ that has an asymptotic normal<br />

distribution with mean θ <strong>and</strong> variance σ 2 /n. Then g(ˆθ) has an asymptotic normal distribution<br />

with mean g(θ) <strong>and</strong> asymptotic variance [g 0 (θ)](σ 2 /n)[g 0 (θ)] = g 0 (θ) 2 σ 2 /n.<br />

Example 3.34 Use the delta method to approximate the variance <strong>of</strong> the maximum likelihood estimator<br />

<strong>of</strong> the probability that an observation from an exponential distribution exceeds 200. Apply<br />

this result to Data Set B.<br />

From Example 2.32 we know that the maximum likelihood estimate <strong>of</strong> the exponential parameter<br />

is the sample mean. We are asked to estimate p =Pr(X>200) = exp(−200/θ). The maximum<br />

likelihood estimate is ˆp =exp(−200/ˆθ) =exp(−200/¯x). Determining the mean <strong>and</strong> variance <strong>of</strong><br />

this quantity is not easy. But we do know that that Var( ¯X) =Var(X)/n = θ 2 /n. Furthermore,<br />

<strong>and</strong> therefore the delta method gives<br />

For Data Set B,<br />

g(θ) = e −200/θ<br />

g 0 (θ) = 200θ −2 e −200/θ<br />

Var(ˆp) . = [200θ −2 e −200/θ ] 2 θ 2 /n =40, 000θ −2 e −400/θ /n.<br />

¯x = 1, 424.4<br />

ˆp = exp(−200/1, 424.4) = 0.86900<br />

dVar(ˆp) = 40, 000(1, 424.4) −2 exp(−400/1, 424.4)/20 = 0.0007444.<br />

A 95% confidence interval for p is 0.869 ± 1.96 √ 0.0007444 or 0.869 ± 0.053. ¤<br />

Example 3.35 Construct a 95% confidence interval for the mean <strong>of</strong> a lognormal population using<br />

Data Set B. Compare this to the more traditional confidence interval based on the sample mean.

Hooray! Your file is uploaded and ready to be published.

Saved successfully!

Ooh no, something went wrong!