(5 pts) Assume that the claims process $S_t$ is represented by a compound Poisson process. The mean time between adjacent claims is 2 hours and the random value $X$ of a particular claim is uniformly distributed on $[0,10]$. The insurance company sets premiums according to $c_t = 1.2 \E[S_t]$.
Solution: may be found here
Note that for the second part, I changed the ruin probability to be less that $0.1$ instead of $0.05$ in the solutions. So set
(5 pts) Assume that $S_t$ is a compound Poisson process with rate $\lambda$ and $c_t = (1 + \theta) \E[S_t]$.
Find the adjustment coefficient when the process has jumps that are $\Gamma(2,a)$ distributed. Hint I want you to compute this exactly like we did in class for the exponential case. Recall that $\Gamma(2,a)$ has MGF
If you flipped the roles of $2$ and $a$, that’s fine too. But you probably won’t be able to solve it explicitly.
Solution: Again found in the solutions above for problem 3.
Customers arrive at a service facility according to a Poisson process with an average rate of 5 per hour. Find
Solutions Can be found here (problem 1).
The number of claims coming into a company is a Poisson process that has mean time between claims equal to 1/2 hours. With probability $p$, a claim is made by a woman over 60 years old. What kind of process is the ${N_t}_{t \in \R^+}$, the number of claims coming from women over 60 years old? Prove your answer.
Solution The time between claims is of course $\tau \sim \Exponential(\lambda)$ is a $\Poisson(\lambda)$ process. This means $\E[\tau] = \lambda^{-1}$. Let
where $X_i$ are $\Bernoulli(p)$. $X_i = 1$ if the $i^{th}$ customer is over 60 years old. $N_t$ is the $\Poisson$ process. $S_t$ represents the total number of senior customers who come in by time $t$.
We know that $S_t$ is compound Poisson and has MGF This is the MGF of a Poisson process for all $t$.
Customers arrive at a service facility according to a non-homogeneous Poisson process with a rate of 3 customers per hour between 9 and 11 am. After 11 am the rate is decreasing linearly from 3 at 11 am to 0 at 5pm. What is the probability that no more than 15 customers arrive between 10am and 4pm?
Solution Recall that $\overline{\lambda}(s,t) = \int_s^t \lambda(r) dr$. In the 10am to 4pm(=16:00) period, we have to look at
Let $N_t$ be the number of customers. Then
Assume the portfolio of a company consists of two homogeneous groups of risks. For the first group, the number of clients is $n_1 = 2000$ and the probability of loss event for each client is $q_1 = 0.1$. The payment, if a loss event occurs, is a non-random amount of $z_1 = 10$ units of money. For the second group, the corresponding quantities are $n_2 = 400$, $q_2 = 0.05$ and $z_2 = 30$. Let $n = n_1 + n_2$, and $S_n = S_{n_1} + S_{n_2}$.
Find
Solutions
Since $S_{n_2}$ and $S_{n_1}$ are approximately normal and independent, a good approximation to their sum is the normal distribution! Clearly,
Then, we’re looking for the probability
How do you solve this?
Assume that the occurrence of traffic accidents is well described by a Poisson process with rate $\lambda = 30$ per day. The probability that an accident causes serious injuries is $p = 0.1$, and the outcomes of different accidents are independent. Write down a formula for the probability that the number of accidents with serious injuries during a month (of $30$ days) exceeds $100$. Estimate this probability using normal approximation.
Solution: may be found here
You should also find the exact formula for the probability of having more than 100 serious accidents. Let $S_N = \sum_{i=1}^N X_i$ be the number of accidents with serious injuries. $N$ is the total number of accidents in a month. Also
Suppose that an insurance company has $d$ groups of clients, and that the clients within each group are homogeneous. That is, in the $i$th group there is a probability $r_i$ that any given client makes a claim. Also assume that the $i$th group has $n_i$ clients. Let $N$ be the total number of clients that make claims over all of the different groups, by Poisson approximation we can assume that $N$ will be Poisson distributed. What will be the parameter $\lambda$ that describes this random variable? Give mathematical justification.
Solution: May be found here
Take a sequence of random variables $X_n$ that are $\Binomial(n,\lambda/n)$ distributed. Do they have a limit in distribution: that is, is there a random variable $Y$ such that
Solution We proved this in class:
using taking the logarithm and using L’hospital’s rule to evaluate the limits. That is, you ought to show
to get full credit.
Clearly this means $Y = \Poisson(\lambda)$.
Solution: The same calculation works
To show that the above limit is $0$, we see that and Putting this and the previous calculation together shows that $\Prob(Y = k) = 0$ for all $k$. This means that there is no random variable that is the limit of $X_n$.
Suppose that 10,000 cars are traveling across a city. One out of 4 cards is gray. Suppose that the probability that any single car has an accident on a particular day is 0.002.
Why can you approximate a Binomial by a Poisson in this case?
Solution
Well-Traveled insurance company sells a travel insurance policy that reimburses travelers for any expenses incurred for a planned vacation that is canceled because of airline bankruptcies. Individual claims follow Pareto with $\theta = 500$ and $\alpha = 2$.
Because of financial difficulties in the airline industry, Well-Traveled imposes a limit of $$1000$ on each claim. If a policyholder’s planned vacation is canceled due to airline bankruptcies and he or she has incurred more than $$1000$ in expenses, what is the expected non-reimbursed amount of the claim? Recall that $\operatorname{Pareto}(\theta,\alpha)$ in its alternate form has the following tail
Solution: This is on page 153, example 3 (2nd edition). Let’s choose $$500$ (since this the Pareto parameter is $$500$) as a unit of money. We consider the distribution of the loss $\xi$ given that $\xi > 2$. The tail of the conditional is
Clearly, $\Prob( \xi > x | \xi > 2) = 1$ if $x < 2$. Then, |
So what we’ve computed is the expected loss given that the loss is greater than $$1000$. This value is $5 \times 500 = $2500$.
This is the total amount of the claim, given that the traveler occurred more than $$1000$ in expenses. The amount that the company reimburses is $$1000$. So the non-reimbursed amount is $$1500$.
Convolutions. Let $X$ and $Y$ be independent random variables. Compute the convolution of $X$ and $Y$ when
Solution:
Let $S = X + Y$.
$X \sim \operatorname{Bernouilli}(p)$ and $Y \sim \operatorname{Bernouilli}(q)$.
Then,
$X \sim N(0,1)$ and $Y \sim \operatorname{Bernouilli}(p)$.
$X \sim \operatorname{Exp}(\alpha)$ and $Y \sim \operatorname{Poisson}(\beta)$. For $t > 0$
where $\lfloor t \rfloor$ is the largest integer smaller than $t$.
Let $X \sim \operatorname{Gamma}(\alpha,\beta)$ so that its pdf is
Solution:
Recall
The MGF of the gamma function may be found as follows. Let $t < \beta$. Then, where we’ve used the change of variables $(\beta - t)x = u$.
Using the theorem about moment generating functions and convolutions
Similarly let $S = \sum_{i=1}^{\infty} X_i$. Recall that In particular, this gives
Then
The earlier definition of the density of $\operatorname{Gamma}(1,1)$ gives with MGF $1/(1-t)$. Hence, $M_S(t)$ is the MGF of $\operatorname{Exp}(1)$.
Fun with MGFs. Compute the moment generating functions of
A Gaussian random variable $X \sim N(0,\sigma^2)$.
Hint:
Can you make a change of variable and complete the proof? IMPORTANT. The computation I’ve shown is you for the $N(0,1)$ case; i.e., the variance of the Gaussian random variable was assumed to be $1$. I want you do it for $N(0,\sigma^2)$, which is slightly different.
Let $X_1 \sim \operatorname{Poisson}(\lambda_1)$ and $X_2 \sim \operatorname{Poisson}(\lambda_2)$ be independent. Show that $X_1 + X_2 \sim \operatorname{Poisson}(\lambda_1 + \lambda_2)$ in two distinct ways:
By using moment generating functions.
Hint: When you have two independent random variables, $X_1$ and $X_2$, then Using this formula, show that the $X_1 + X_2$ has the moment generating function of $\operatorname{Poisson}(\lambda_1 + \lambda_2)$.
Find the moment generating function of a Pareto random variable with tail ($\alpha > 1$)
Is the Pareto random variable heavy or light-tailed? State precisely why or why not.
Hint: It turns out that Pareto is heavy tailed (it’s merely Polynomial). Start your proof like this. Suppose there are constants $B,c > 0$ such that
However, we can prove (you must prove this) that
This is a contradiction! This means that the Pareto distribution cannot be light-tailed. Therefore the moment generating function of the Pareto distribution must be $M(z) = +\infty$ for all $z > 0$. (Why? Recall the theorem we proved in class)
Alternate approach: You may directly show that $M(z) = +\infty$, of course. It follows that the Pareto random variable is heavy-tailed.
Solutions: This week’s solutions may be found here
An EUM customer of an insurance company has a total wealth of 100 and is facing a arandom loss $\xi$ distributed as follows.
*Hint:* Recall that for the insurance company, if they take premium $p$, their final wealth is
$$
X = w + p - \xi
$$
Use the EUM criterion for the company to determine the minimum premium. 1. Repeat the previous question when $\xi$ is uniformly distributed on $[0,100]$ instead.
Let $u(x) = x$ for $x \in [0,1]$ and $u(x) = 1/2 + x/2$ for $x \geq 1$. Is an EU maximizer with this utility function a risk averter (according to condition $Z$)? Why?
Let $d > 0$ and $r_d(x)$ be the function
(8 pts) Let us construct an example where the mean-variance criterion is non monotone. Let $X$ be Pareto distributed. That is, let
The Pareto distribution has $\alpha > 1$.
Let $Y$ be uniformly distributed on $[0,1]$.
Solution: Clearly the pdf is
This gives
Which gives
Solution: Since $Y \in [0,1]$ and $X \in [1,\infty)$. Clearly $X \geq Y$, and it makes no difference whether or not they’re independent.
The mean-variance criterion says that
where
Solution: We need to compute the range of $\tau$ for which
This results in the solution
You will find that we can show that the condition can be satisfied by choosing $\tau$ large enough when $\alpha \in (2,1/2(5 + \sqrt{33})]$.
Solution: We have found an interval where $V(Y) \geq V(X)$ even though $Y \leq X$ with probability $1$! This means the condition cannot be monotone.
(5 pts) Which of the following functions are convex on their entire domain of definition? Justify your answer.
$u \colon \R \to \R$, $u(x) = | x | $ |
(7 pts) Another gambler type: Let a gambler’s utility function be $u(x) = e^x$. Consider a game where the gambler wins $a$ dollars with probability $p$ and loses the same amount with probability $q=1-p$.
Based on the EUM criterion, determine the ranges of values for $a$ and $p$ for which the gambler should play the game.
Is the EUM criterion monotone for this particular utility function? Why?
Would you say the gambler likes to takes risks, or is he very cautious? Why or why not?
Interest rates: There are two banks in the town of Bankopia, UT. J.P Young offers offers you a nominal interest rate of $4.5\%$ compounded every quarter on your savings. Goldman-Woodruff offers an effective rate of $5.3 \%$ per annum.
a. What is the effective interest rate that J.P Young offers?
b. If you started with $1000 in your account, how much would you have at the end of one year? Compare both banks.
Solution: Let $e$ be effective rate of interest offered by J.P. Young.
If you started with $
1000 in your account, you’d get $(1+0.045/4)^4 1000$ with J.P Young and $(1+0.053)1000$ with Goldman-Woodruff.
Double integral: Let $X$ and $Y$ independent random variables. Suppose $X$ and $Y$ are standard normal. We will find the probability distribution of the following random variable $M$.
NOTE You must follow the steps I’ve written down here. You may not state the distribution of $M$ by citing a theorem.
All I want you to do here is to write $F_{M | Y=y}~(t)$ as an integral from $-\infty$ to $\infty$. |
Hint: The Gaussian pdf is given by $f_Y(y) = (1/\sqrt{2\pi}) e^{-y^2/2}$.
Solution: We have
Hint: Make the following change of variables
Remember double integrals from Calc 3? First compute the limits of integration in the $u$ and $v$ variables. You will need to compute a Jacobian (a determinant) and plug it into the integral.
Solution: First we make the change of variables to $(u,v)$. Notice that that Another miracle of mathematics. Then compute the Jacobian of the transformation
Also notice that the condition on $(x + y)/\sqrt{2} \leq t$ translates to $u \leq t$. I suggest you draw a picture of the area to verify this.
Therefore
Solution: From part $2$, it follows that $M$ is normally distributed with mean zero and variance $1$.
1. Conditional distributions. Let $Y$ be any random variable and let $X$ be discrete. The conditional distribution of $Y$ given $X$ is defined to be
The conditional expectation is a function of $X$, and is defined to be
Let $X$ be the number that shows up on the roll of a standard fair die. Let $Y$ be an exponential random variable with parameter $X$. That is $Y \overset{d}{=} \exp(X)$.
Find $F_{Y | X}(y)$. |
Hint: The definition may also be found on page 49 of your textbook under the section called conditional expectations. The notation used in your textbook is $F_Y(y | X = x)$, which is the same thing as $F_{Y | X}(y)$. Once the value of $X$ is fixed, the distribution of $Y$ is completely determined. |
To make things even more explicit, let’s think of the problem in the following way. Let there be 6 variables $Y_1,\ldots, Y_6$ all independent of each other and $X$. Let $Y_i$ be exponentially distributed with parameter $i$. Then,
For example, then
Solution: Simply redo the above calculation to get
Find $\E[Y | X]$. |
Hint: This object is again a function of the value of $X$, but the dependence on $Y$ has been integrated out so to speak. Using integration by parts, we get
Now repeat this caclulation for general $X = i$.
Question to ask yourself: Why did the integral change from $(-\infty,\infty)$ to $(0,\infty)$?
Solution:
The integral changed from $(-\infty,\infty)$ to $(0,\infty)$ because $Y$ is exponentially distributed; its density is zero on the negative reals.
When $y,x \in \R$, find the joint distribution
Hint: Use $\Prob(A , B) = \Prob (A \cap B) = \Prob(A | B) \Prob(B)$. Notice that the joint depends on $x$ and $y$. The events here are $A = { Y \leq y }$ and $B = { X \leq x }$. |
Solution:
We’ve used the result of an earlier computation in the second step of the equation.
Find the marginal distribution $F_{Y}$.
Hint: The marginal distribution of $Y$ can be found the following way.
Solution: We may simply set $x = 6$ in the joint marginal distribution (Why?!). That is,
Given that $3 \leq Y \leq 4$, calculate the probability that a three was shown on the dice.
Hint: Bayes rule. So the two events here are $A = { 3 \leq Y \leq 4 }$ and $B = { X = 3}$. The problem asks you to compute $P(B | A)$. It ought to be easy to compute $\Prob( A | B)$. Can you use Bayes rule to compute $\Prob( B | A)$? |
Solution: We have to compute
The second line in the above followed from the definition of the conditional marginal in the earlier part of the problem. It’s left to you to complete this computation.
2. Tower property
The conditional expectation satisfies a very important property called the tower property.
Consider the previously defined random variables $Y$ and $X$ again. Verify the tower property by
Calculating $\E[Y]$ using the joint distribution $F_{Y,X}$.
Hint: It’s better to compute this using the marginal distribution $F_Y$.
Solution:
Calculating $\E[Y]$ using the tower property. That is, calculate
Hint: We’ve computed $\E[ Y | X = i]$ in a previous step. Just plug it in. |
Solution:
3. Independence and random variables We say that $Y$ does not depend on an event $A$ if $Y$ and $1_A$, the indicator of the event $A$ are independent. The indicator function $1_A = 1$ if $A$ occurs and is $0$ if $A$ does not occur. Show that when $Y$ is any random variable, and $A$ is any event independent of $Y$, we have
Hint: Use the definition of conditional expectation given above. We’ve written down the conditional distribution $F_{Y | X}\,(y)$ in the first problem. Here $X = 1_A$. It can take two values, $0$ or $1$. Can you repeat the argument of the first problem to show that $\E[Y | 1_A = 0] = \E[Y]$ and $\E[Y | 1_A = 1] = \E[Y]$. |
Solution:
We have used independence in the third step. This can be repeated on $1_A = 0$, and this proves the result.