Natural Hedging Using Multi-population Mortality...
Transcript of Natural Hedging Using Multi-population Mortality...
Natural Hedging Using
Multi-population Mortality Forecasting Models
by
Shuang Chen
B.Sc., Nankai University, 2011
Thesis Submitted in Partial Fulfillment of the Requirements for the Degree of
Master of Science
in the
Department of Statistics and Actuarial Science
Faculty of Science
© Shuang Chen 2014 SIMON FRASER UNIVERSITY
Fall 2014
All rights reserved.
However, in accordance with the Copyright Act of Canada, this work may be
reproduced without authorization under the conditions for “Fair Dealing.”
Therefore, limited reproduction of this work for the purposes of private study,
research, criticism, review and news reporting is likely to be in accordance
with the law, particularly if cited appropriately.
APPROVAL
Name: Shuang Chen
Degree: Master of Science
Title: Natural Hedging Using Multi-population Mortality Fore-
casting Models
Examining Committee: Chair: Dr. Gary Parker
Associate Professor
Dr. Cary Chi-Liang Tsai
Senior Supervisor
Associate Professor
Simon Fraser University
Dr. Yi Lu
Supervisor
Associate Professor
Simon Fraser University
Dr. David Campbell
Internal Examiner
Associate Professor
Simon Fraser University
Date Approved: December 11th, 2014
ii
Abstract
No mortality projection model can capture future mortality changes accurately so
that the actual mortality rates are different from the projected ones. The movement
of mortality rates has oppositive impacts on the values of life insurance and annuity
products, which creates a chance of nature hedge for both life insurer and annuity
provider. A life insurer and an annuity provider can swap their life insurance and
annuity business for each other to form their own portfolios for natural hedge. This
project is mainly focused on determining the weights of a portfolio of life insurance
and annuity products by minimizing the variance of the loss function of the portfolio
to reduce mortality and longevity risks for each of the life insurer and the annuity
provider. Four Lee-Carter-based models are applied to model the co-movement
of two populations of life insurance and annuity insureds, and then determine the
weights for comparisons. The block bootstrap method, a model-/parameter-free
approach, is also adopted with numerical illustrations to compare the hedging per-
formances among the four models.
iv
Acknowledgments
I would like to express my deepest appreciation to all those who supported and
guided me to complete this project. A special gratitude is given to my supervisor,
Dr. Cary Tsai, who provided me with patient guidance and valuable suggestions
that inspired me and helped me realize self-improvements throughout my two years
of studies in SFU. I will always be grateful to have the opportunity to learn new
ideas and conduct research under his supervision. Without his devoting time to
this project, I wouldn’t have completed the project. I hereby thank you for all your
time spent on guiding me.
Furthermore, I would also like to acknowledge with my great appreciation the
members of my project committee, Dr. Gary Parker, Dr. Yi Lu and Dr. David
Campbell. Thank you for reading my project with your valuable time and giving me
insightful comments and suggestions.
A special thank goes to the whole statistics and actuarial science department,
where I mastered more advanced knowledge that will benefit me forever. Thank
you to all the professors who have ever given me lectures and answered my ques-
tions. You helped me to fulfill a high level academic achievement.
Last but not least, I would like to take this opportunity to thank my family and my
friends who supported me and encouraged me whenever I felt confused and lost.
Especially, I would like to express my appreciation to my fellow graduate students,
Annie, Biljana, Elena, Fei, Huijing, Sabrina, Vicky and Yi, who brought me fresh
ideas and memorable moments during my study.
vi
Contents
Approval ii
Partial Copyright License iii
Abstract iv
Dedication v
Acknowledgments vi
Contents vii
List of Tables ix
List of Figures x
1 Introduction 1
1.1 Motivations . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 1
1.2 Outline . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 2
2 Literature Review 4
2.1 Mortality forecasting models . . . . . . . . . . . . . . . . . . . . . . . 4
2.2 Hedging of mortality and longevity risks . . . . . . . . . . . . . . . . 6
3 Introduction of Multi-population Models 9
3.1 Concepts and notations . . . . . . . . . . . . . . . . . . . . . . . . . 9
3.2 Multi-population models and constrains . . . . . . . . . . . . . . . . 11
3.2.1 Independent Lee-Carter model . . . . . . . . . . . . . . . . . 11
3.2.2 Joint-k Lee-Carter model . . . . . . . . . . . . . . . . . . . . 16
3.2.3 Co-integrated model . . . . . . . . . . . . . . . . . . . . . . . 20
vii
3.2.4 Augmented common factor model . . . . . . . . . . . . . . . 23
4 Application in Mortality Swap 28
4.1 Natural hedging and mortality swap . . . . . . . . . . . . . . . . . . 28
4.2 Numerical illustrations . . . . . . . . . . . . . . . . . . . . . . . . . . 33
4.2.1 Assumptions and portfolios . . . . . . . . . . . . . . . . . . . 33
4.2.2 Robustness testing . . . . . . . . . . . . . . . . . . . . . . . . 36
5 Block Bootstrap Method 45
6 Conclusion 56
Bibliography 57
viii
List of Tables
4.1 the median of 50 optimal weights . . . . . . . . . . . . . . . . . . . . 37
4.2 comparisons of sample variances (×1023) and HE’s . . . . . . . . . 38
5.1 p-values for testing stationarity of {dx,t,i} . . . . . . . . . . . . . . . . 48
5.2 comparisons of sample variances (×1023) and HE’s (block bootstrap) 55
ix
List of Figures
3.1 cohort mortality sequence and period mortality sequence . . . . . . 10
3.2 kt,i against t . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 13
3.3 95% predictive intervals on qx,2009+1,i for the Independent model . . . 15
3.4 95% predictive intervals on qx,2009+1,i for the joint-k model . . . . . . 19
3.5 95% predictive intervals on qx,2009+1,2 for the co-integrated and inde-
pendent models . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 22
3.6 95% predictive intervals on qx,2009+1,i for the augmented common fac-
tor model . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 26
3.7 mortality curve comparisons among models and actual rates . . . . 27
4.1 Mortality swap: minimizing V ar(LL) + V ar(LA) . . . . . . . . . . . . 30
4.2 Mortality swap: minimizing V ar(LL) . . . . . . . . . . . . . . . . . . 30
4.3 Mortality swap: minimizing V ar(LA) . . . . . . . . . . . . . . . . . . 30
4.4 age span [25, 100] and year span [1981, 2010] . . . . . . . . . . . . . 35
4.5 optimal weights for the independent and joint-k models . . . . . . . . 39
4.6 optimal weights for the co-integrated and augmented common factor
models . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 40
4.7 variances after swap . . . . . . . . . . . . . . . . . . . . . . . . . . . 41
4.8 simulated loss distributions using (wL+Al , wL+A
a ) . . . . . . . . . . . . 42
4.9 simulated loss distributions using (wLl , w
La ) . . . . . . . . . . . . . . . 43
4.10 simulated loss distributions using (wAl , w
Aa ) . . . . . . . . . . . . . . . 44
5.1 {dx,t,i} for x = 35, 45, 55, 65 . . . . . . . . . . . . . . . . . . . . . . . . 47
5.2 a circle diagram of dt,i’s . . . . . . . . . . . . . . . . . . . . . . . . . 49
5.3 simulated loss distributions using (wL+Al , wL+A
a ) . . . . . . . . . . . . 52
5.4 simulated loss distributions using (wLl , w
La ) . . . . . . . . . . . . . . . 53
5.5 simulated loss distributions using (wAl , w
Aa ) . . . . . . . . . . . . . . . 54
x
Chapter 1
Introduction
1.1 Motivations
Natural hedging is a strategy of hedging two risks responding oppositely to a
change in a common factor. Since life insurers and annuity providers face mor-
tality risk (the actual death probabilities are higher than expected) and longevity
risk (the actual survival probabilities are larger than expected), respectively, both
of them can adopt natural hedging by swapping a portion of their business between
each other to reduce mortality and longevity risks.
In this project, we would like to find the optimal swapped weights of life and
annuity business for a life insurer and an annuity provider, respectively, which min-
imize the variance of a loss function of the life insurer or the annuity provider, or
the sum of the variances of the two loss functions. To reach the goal, we need
some mortality models to forecast mortality rates for life insurance policyholders
and annuitants.
Since life insurance and annuities tend to be issued to different populations,
using the same mortality table to determine the premiums is inappropriate. Fur-
thermore, even though we use two different mortality tables for life insurance pol-
icyholders and annuitants, we cannot ignore the dependence between these two
populations. Generally speaking, two populations within a territory or country are
exposed to the same medical and environmental conditions. Therefore, the mortal-
ity rates for two populations might not be independent. In this project, two different
1
CHAPTER 1. INTRODUCTION 2
mortality tables and four multi-population mortality models are applied to forecast-
ing deterministic and stochastic mortality rates.
The deterministic mortality rates are used to determine the premiums of life and
annuity products, and the stochastic mortality ones are used to simulate the losses
of the life insurer and annuity provider. Then we calculate the sample variances
and sample covariance of the losses of the life insurer and annuity provider for ob-
taining the formulas of the optimal weights. A variance reduction ratio, called hedge
effectiveness, is proposed to compare the performances of hedging mortality and
longevity risks.
This project also carries out the robustness testing to study the sensitivity of
the weights and variances obtained from the simulated mortality rates. Finally, be-
cause all the results are derived based on the parametric mortality models, we are
not sure whether the assumed mortality model is the true one. To avoid the pa-
rameter and model risks, the model-free block bootstrap method is used to project
mortality rates. Then we redo the numerical results which are compared with those
produced by the parametric mortality models.
1.2 Outline
This project consists 5 chapters. Chapter 2 is a review of the previous research on
both mortality projection models and hedging of mortality and longevity risks.
In Chapter 3, some actuarial notations used in this project are provided, fol-
lowed by the introduction of four mortality projection models, the independent Lee-
Carter model, joint-k Lee-Carter model, co-integrated Lee-Carter model and aug-
mented common factor model, among which the last three can model the depen-
dent structure between the mortality rates for two populations. These models can
be used to forecast deterministic and stochastic mortality rates for both life insur-
ance policyholders and annuitants.
CHAPTER 1. INTRODUCTION 3
Chapter 4 focuses on natural hedging with mortality swap. In this chapter, three
pairs of optimal weights for swapping life insurance and annuity business are stud-
ied. Numerical illustrations using the U.S. mortality tables are exhibited. Also, the
robustness testing is conducted to test the sensitivity of the numerical results.
In Chapter 5, the model free block bootstrap method is introduced in details,
which is then applied to testing the results produced by the multi-population models
in Chapter 4.
Chapter 2
Literature Review
The two main parts of this project are mortality projection and natural hedge. The
first section will go through the articles about the mortality forecasting models and
the second section will review those about hedging of mortality and longevity risks.
2.1 Mortality forecasting models
Mortality projection is one of the keys for pricing insurance and annuity products.
The most widely used mortality projection model so far is the Lee-Carter model
(Lee and Carter (1992)). Between 1900 and 1988, the life expectancy in the United
States increased from 47 to 75. Based on the assumption that the life expectancy
would continue to increase, Lee and Carter (1992) applied the time series model to
forecasting long term mortality rates. This model considers two age-varying factors
and one time-varying factor to capture the downward trend in mortality rates. By
empirical mortality data, it is assumed that the time-varying factor is fitted by an
ARIMA (autoregressive integrated moving average) model.
One of the most significant features of the Lee-Carter model is that it is easy to
interpret and its parameters are estimable. Compared to other mortality projection
models, the Lee-Carter model bases its forecast on long term historical data, and
no explicit assumptions or a life span limit is attached to the model. Moreover, the
Lee-Carter model provides confidence regions, which was further proved that the
model is efficient in forecasting mortality rates.
4
CHAPTER 2. LITERATURE REVIEW 5
Afterwards, Lee and Carter (1992) pointed out that in order to forecast the mor-
tality rates for two populations within a territory or country, the same time-varying
index, called joint-k index, should be used so that the same trend of mortality
growth in response to time for the two populations can be reflected. This assump-
tion makes sense because usually two populations in a country have the same
medical service and environmental condition, etc. Thus, when these aspects im-
prove, the life expectancy of each population will increase simultaneously to the
same level.
Lee and Li (2005) extended the original Lee-Carter model by adding an extra
term to the model. They thought that the world is becoming more closely con-
nected by modern communication technology. Thus, in a long term, the forecasted
mortality rates for different populations will converge. They called it the augmented
common factor model, which was developed in two steps. Firstly, Lee and Li (2005)
used a common factor to capture the overall mortality growth; then the mortality
rates of that population are specified by an augmented term which reflects the fea-
ture of that population. This method was applied to forecasting the mortality rates
of Canadian populations. In different provinces, the common factor was used to
realize the long term convergence, and the augmented term was used to separate
the mortality rates for each province.
With the same concern about the long-term convergence problem, Li and Hardy
(2011) suggested the so-called co-integrated Lee-Carter model. By investigating
the time-varying indices of two populations, they noticed that there exists a co-
integrated relationship between the two time indices. In the co-integrated model,
two time-varying indices are modeled by a linear function.
The last three mentioned models based on the Lee-Carter model all take into
consideration the dependent structures among multiple populations; we call them
the multi-population mortality projection models. In this project, because the natu-
ral hedging strategy is adopted by swapping portions of life and annuity business
based on two populations of life insurance policyholders and annuitants, the multi-
population mortality projection models are needed.
CHAPTER 2. LITERATURE REVIEW 6
2.2 Hedging of mortality and longevity risks
Hedging of mortality and longevity risks is one of the most popular research areas
recently in Actuarial Science. Researchers have conducted large amount of study
on how to reduce mortality and longevity risks. Some of the financial instruments
have been applied to hedging these two risks, such as mortality-linked securities,
q-forward, etc. The practice of using mortality-linked securities was first suggested
by Blake and Burrows (2001) by pouring longevity risk into the capital market.
They introduced a survivor bond for which the future coupons are based on the
percentage of the alive retirees at the future coupon payments dates.
In 2003, Swiss Re issued the first mortality-linked security which is to protect
insurers from the loss of catastrophes. The bonds were issued via SPV (special
purpose vehicle). The payments depend on a mortality index. This bond has been
successfully operated for years.
BNP-Paribas and the European Investment Bank in 2004 issued a 25-year
longevity bond to hedge longevity risks. This is the first longevity bond that came
into real business. The coupons were linked to the survivor index based on the
actual mortality rates of England and Welsh males aged 65 in 2002.
In 2006, a pension buyout market attracted a lot of attention in UK. In this market
both the assets and liabilities of a pension plan were transferred to a life insurer.
To realize the transfer, the pension plan has to pay the life insurer the amount of
deficit if the assets are less than the liabilities. On the contrary, the insurer will pay
the surplus to the pension plan if assets exceeds liabilities. Thus, the pension plan
can be secured. This is an efficient method of transferring longevity risk; however,
the pension plan will possibly experience loss since the life insurer tends to re-
evaluate the pension plan by measuring the assets and liability in a way of more
risk aversion.
In 2007, J.P. Morgan introduced q-forwards, another financial derivatives, to
transfer mortality rates to the capital market. The q-forwards are described as a
zero coupon swap between a pre-fixed mortality rate at inception date and the
realized mortality rate at maturity date of the q-forward contract.
CHAPTER 2. LITERATURE REVIEW 7
Not only the pricing and structure of the mortality-linked securities have been
systematically studied but also how to evaluate the hedging methodology. Li and
Ng (2011) pointed out that because of lack of longevity trading index from which to
evolve the market price of the mortality-linked securities, it still remains a question
of how to evaluate the hedging technics. Thus, they proposed a pricing framework
for evaluating the mortality-linked securities based on canonical valuation. To con-
struct the framework, they suggested a nonparametric model, which helps to avoid
the risks from model itself and its parameters.
Gaillardetz et al. (2012) proposed a method of evaluating the hedging error
under a stochastic mortality projection. It is very likely that when catastrophes
occur insurers would experience huge amount of losses. Thus, evaluating the
distribution of hedging error is very important. They applied the regime-switching
model (Milidonis et al. (2011)) to extracting the hedging error and conducted the
error distribution that made the evaluation feasible.
In recent years, another kind of hedging method called natural hedge has be-
come appealing for researchers. Because instead of transferring the risk into fi-
nancial market, natural hedge can reduce mortality and longevity risks simply by
swapping proportions of life and annuity business between a life insurer and an an-
nuity provider. Cox and Lin (2007) are the first to thoroughly introduce this concept
by noticing the empirical evidence that companies adopting natural hedge within
their business usually offer lower premiums than the other companies.
After the proposal of natural hedge, tons of studies have been done in this field.
Wang et al. (2010) suggested that by mortality duration and convexity matching,
the optimal weights for conducting an immunization portfolio can be obtained. Lin
and Tsai (2013) gave more elaborate formulas for mortality duration and convexity,
which can be applied to a two-product or three-product portfolio.
However, some researchers also threw doubt on natural hedging that all the
studies are based on some models, but in reality the future mortality rates might be
totally model free, which may cause the method inefficient. Zhu and Bauer (2014)
used a non-parametric model to test the natural hedge method and came to the
conclusion that higher order variations in mortality rates would affect the efficiency
of natural hedging. Another problem of natural hedging pointed out by Cox and Lin
CHAPTER 2. LITERATURE REVIEW 8
(2007) is that for estimating the future mortality rates, most of the studies use the
same life table for both annuity and life insurance, which is not practical.
This project uses multi-population mortality projection models to forecast future
mortality rates for life insurance policyholders and annuitants separately based
on two life tables. The method of Langrage multipliers is used to help obtain the
optimal weights. Moreover, robust testing, together with model testing based on
the model-free block bootstrap method, is carried out.
Chapter 3
Introduction of Multi-populationModels
In this chapter, we first introduce four mortality models for multi-populations, which
are based on the well known Lee-Carter model. Then we fit the models with mor-
tality data from the Human Mortality Database to get the estimated parameters
which can be applied to forecasting future mortality rates. The deterministic mor-
tality rates can be used to determine the prices of life insurance and annuity prod-
ucts, and the stochastic ones can be used to simulate the realized/actual prices in
Chapter 4.
3.1 Concepts and notations
Let qx,t,i denote the probability that an individual aged x in year t for population
i will die within one year. We add an extra subscript i for studying the multi-
population mortality projection models. When forecasting the future mortality rates,
there are mainly two types of mortality sequences. One is the cohort mortality se-
quence {qx0+j,t0+j,i : j = 0, 1, 2, ...}, and the other is the period mortality sequence
{qx0+j,t0,i : j = 0, 1, 2, ...}. In this project, the cohort mortality sequence is used
for pricing. Figure 3.1 shows the difference between the cohort mortality sequence
and the period one.
Next, denote μx,t,i(s) the function of force of mortality, which is an instantaneous
death rate between age x and x+s in year t for population i aged x. It can be shown
9
CHAPTER 3. INTRODUCTION OF MULTI-POPULATION MODELS 10
Figure 3.1: cohort mortality sequence and period mortality sequence
(see Bowers et al. (1997)) that
1− qx,t,i = e−∫ 10 μx,t,i(s)ds.
Under the assumption of constant force of mortality within one year, that is, μx,t,i(s) =
μx,t,i(0)�=μx,t,i for s ∈ [0, 1), we have
qx,t,i = 1− e−μx,t,i .
Another frequently used mortality rate is the central death rate mx,t,i which is
defined as the ratio of the number of deaths aged x in year t to the average number
of people aged x in year t. Again, under the assumption of constant force of
mortality within one year, it can be shown that μx,t,i = mx,t,i, and thus,
qx,t,i = 1− e−mx,t,i .
The equation above provides data transformation between qx,t,i and mx,t,i.
CHAPTER 3. INTRODUCTION OF MULTI-POPULATION MODELS 11
3.2 Multi-population models and constrains
3.2.1 Independent Lee-Carter model
Lee and Carter (1992) proposed a model that the natural logarithm of central death
rates can be expressed by two age specific factors and one time specific factor as
follows:
ln(mx,t,i) = ax,i + bx,i × kt,i + εx,t,i,
subject to two constraints,
• ∑x bx,i = 1,
and
• ∑t kt,i = 0,
where
• ax,i is the average age-specific mortality factor at age x for population i,
• kt,i is the general mortality level in year t for population i,
• bx,i is the age-specific reaction to kt,i at age x for population i, and
• εx,t,i is the error term which is assumed independent and identically normally
distributed with mean 0 and variance σ2εx,i
for all t′s.
To fit the model with a given matrix of ln(mx,t,i)′s for population i and estimate
all parameters, we may use the singular value decomposition (SVD) method. How-
ever, there is a close approximation to SVD.
First, ax,i can be derived by the average of the sum of ln(mx,t,i) over a given year
span [t0, t0+n−1] by the second constraint, and kt,i equals the sum of [ln(mx,t,i)−ax,i] over a given age span [x0, x0 +m− 1] by the first constraint. That is,
t0+n−1∑
t=t0
ln(mx,t,i) = n× ax,i + bx,i ×t0+n−1∑
t=t0
kt,i = n× ax,i,
CHAPTER 3. INTRODUCTION OF MULTI-POPULATION MODELS 12
implying
ax,i =
∑t0+n−1t=t0
ln(mx,t,i)
n, x = x0, x0 + 1, ...., x0 +m− 1.
Similarly,
x0+m−1∑
x=x0
[ln(mx,t,i)− ax,i] = kt,i ×x0+m−1∑
x=x0
bx,i,
yielding
kt,i =
x0+m−1∑
x=x0
[ln(mx,t,i)− ax,i], t = t0, t0 + 1, ..., t0 + n− 1.
Finally, bx,i can be obtained by regressing [ln(mx,t,i) − ax,i] on kt,i without the
constant term involved for each age x. For projecting the future time-varying index
kt,i, a key to project the mortality rates, an ARIMA(0,1,0) time series (a random
walk with drift θi) is used, which is given by
kt,i = kt−1,i + θi + εt,i,
where the error term εt,i follows an independent and identical normal distribution
with mean 0 and variance σ2ε,i for all t′s. The parameter θi can be estimated by
θi =1
n− 1
t0+n−1∑
t=t0+1
(kt,i − kt−1,i) =kt0+n−1,i − kt0,i
n− 1.
Figure 3.2 displays the plots of {kt,i} from year 1981 to year 2009 (t0 = 1981
and n = 29) based on the U.S. male (i = 1) and female (i = 2) mortality rates for
an age span [25, 100] from the Human Mortality Database (www.mortality.org). The
data set is used throughout this chapter. From Figure 3.2 we can see that the {kt,i}for each of male and female populations has a decreasing trend over time.
CHAPTER 3. INTRODUCTION OF MULTI-POPULATION MODELS 13
Figure 3.2: kt,i against t
The logarithm of the stochastic central death rate for age x in year t0+n− 1+ τ
denoted by mx,t0+n−1+τ,i is
ln(mx,t0+n−1+τ,i) =ax,i + bx,i × (kt0+n−1,i + τ × θi +√τ × εt,i) + εx,t,i
=ln(mx,t0+n−1+τ,i) +√τ × bx,i × εt,i + εx,t,i,
where mx,t0+n−1+τ,i is the deterministic central death rate. Note that ln(mx,t0+n−1+τ,i)
is normally distributed with mean
ln(mx,t0+n−1+τ,i) = ax,i + bx,i × (kt0+n−1,i + τ × θi), τ = 1, 2, ....
and variance σ2(ln(mx,t0+n−1+τ,i)) = τ × b2x,i × σ2ε,i + σ2
εx,i, where {εx,t,i} and {εt,i}
are assumed independent.
Then
mx,t0+n−1+τ,i = exp[ax,i + bx,i × (kt0+n−1,i + τ × θi)],
or
qx,t0+n−1+τ,i = 1− exp[−exp(ax,i + bx,i × (kt0+n−1,i + τ × θi))].
CHAPTER 3. INTRODUCTION OF MULTI-POPULATION MODELS 14
The estimate of the variance of εx,t,i is given by
σ2εx,i
=1
n− 2
t0+n−1∑
t=t0
ε2x,t,i =
∑t0+n−1t=t0
[ln(mx,t,i)− ax,i − bx,i × kt,i]2
n− 2,
where x = x0, x0 + 1, ..., x0 +m− 1. The estimate of the variance of εt,i is given as
σ2ε,i =
1
n− 2
t0+n−1∑
t=t0+1
ε2t,i =
∑t0+n−1t=t0+1 (kt,i − kt−1,i − θi)
2
n− 2.
Therefore, the estimate of the variance of ln(mx,t0+n−1+τ,i) is obtained by
σ2(ln(mx,t0+n−1+τ,i))�= σ
2x,t0+n−1+τ,i = τ × b2x,i × σ2
ε,i + σ2εx,i
,
and a 100(1− γ)% predictive interval on qx,t0+n−1+τ,i can be constructed as
1− exp[−exp(ln(mx,t0+n−1+τ,i)± z γ2× σx,t0+n−1+τ,i)].
Figure 3.3 shows the projected period mortality rates in year 2010 based on the
central death rates from year 1981 to 2009. It is obvious that the estimated mor-
tality rates are quite close to the actual ones, and a narrower predictive interval on
the female mortality rates than that on the male ones is produced.
CHAPTER 3. INTRODUCTION OF MULTI-POPULATION MODELS 15
Figure 3.3: 95% predictive intervals on qx,2009+1,i for the Independent model
CHAPTER 3. INTRODUCTION OF MULTI-POPULATION MODELS 16
3.2.2 Joint-k Lee-Carter model
Apparently, two independent models ignore the co-movements between the mor-
tality rates for two populations. The males and females in a country live in the
same environment, and their mortality rates are affected by some common fac-
tors. To relate one population’s mortality rates to the other’s ones, Carter and
Lee (1992) introduced the joint-k model where the time-varying index kt,i, which
shows the general mortality level over time, is the same for both populations, that
is, kt,1 = kt,2 = Kt. The model can be expressed as
ln(mx,t,i) = ax,i + bx,i ×Kt + εx,t,i, i = 1, 2,
subject to two new constraints:
• ∑2i=1
∑x bx,i = 1
and
• ∑t Kt = 0.
Similarly, ax,i can be derived by the average of the sum of ln(mx,t,i) over a year
span [t0, t0+n−1], and Kt is equal to the sum of [ln(mx,t,i)− ax,i] over an age span
[x0, x0 +m− 1] and the population index. That is,
t0+n−1∑
t=t0
ln(mx,t,i) = n× ax,i + bx,i ×t0+n−1∑
t=t0
Kt,
implying
ax,i =
∑t0+n−1t=t0
ln(mx,t,i)
n, x = x0, x0 + 1, ..., x0 + n− 1;
and2∑
i=1
x0+m−1∑
x=x0
ln(mx,t,i) =2∑
i=1
x0+m−1∑
x=x0
ax,i +2∑
i=1
x0+m−1∑
x=x0
bx,i ×Kt,
giving
Kt =2∑
i=1
x0+m−1∑
x=x0
[ln(mx,t,i)− ax,i], t = t0, t0 + 1, ..., t0 + n− 1.
As for bx,i, it is obtained by regressing [ln(mx,t,i) − ax,i] on Kt without the constant
term involved for each age x. The common time-varying index Kt is also assumed
CHAPTER 3. INTRODUCTION OF MULTI-POPULATION MODELS 17
to follow a random walk with drift θ. That is
Kt = Kt−1 + θ + εt,
where the error term εt follows an independent and identical normal distribution
with mean 0 and variance σ2ε , and εt is independent of εx,t,i. The parameters θ is
estimated by
θ =Kt0+n−1 − Kt0
n− 1.
The logarithm of the forecasted central death rate for age x in year t0 + n− 1+ τ is
ln(mx,t0+n−1+τ,i) = ax,i + bx,i × (Kt0+n−1 + τ × θ +√τ × εt,i) + εx,t,i
= ln(mx,t0+n−1+τ,i) +√τ × bx,i × εt,i + εx,t,i,
where
ln(mx,t0+n−1+τ,i) = ax,i + bx,i × (Kt0+n−1 + τ × θ), τ = 1, 2, ....
The estimate of the variance of εx,t,i is
σ2εx,i
=
∑t0+n−1t=t0
[ln(mx,t,i)− ax,i − bx,i × Kt]2
n− 2,
and the estimate of the variance of εt is
σ2ε =
∑t0+n−1t=t0+1 (Kt − Kt−1 − θ)2
n− 2.
Thus, the variance of ln(mx,t0+n−1+τ,i) is
σ2x,t0+n−1+τ,i = τ × b2x,i × σ2
ε + σ2εx,i
,
which can be used to construct a 100(1− γ)% predictive interval on qx,t0+n−1+τ,i as
1− exp[−exp(ln(mx,t0+n−1+τ,i)± z γ2× σx,t0+n−1+τ,i)].
Using the same data set as that for Figure 3.3, we display corresponding forecasted
mortality rates and associated predictive intervals based on the joint-k model in
CHAPTER 3. INTRODUCTION OF MULTI-POPULATION MODELS 18
Figure 3.4. The projected mortality rates and the predictive interval are almost the
same with those for the independent model, where the estimated mortality rates
are close to actual ones, and the 95% predictive interval on the male mortality rates
is wider than that on the female ones.
CHAPTER 3. INTRODUCTION OF MULTI-POPULATION MODELS 19
Figure 3.4: 95% predictive intervals on qx,2009+1,i for the joint-k model
CHAPTER 3. INTRODUCTION OF MULTI-POPULATION MODELS 20
3.2.3 Co-integrated model
In the joint-k model, two populations share the same time-varying index Kt. In the
co-integrated model, the time-varying index for population 2 is a linear function of
the time-varying index for population 1. Specifically, assume the mortality rates for
populations 1 and 2 are modeled by
ln(mx,t,1) = ax,1 + bx,1 × kt,1 + εx,t,1
and
ln(mx,t,2) = ax,2 + bx,2 × kt,2 + εx,t,2.
By fitting the model with given data, kt,1 and kt,2 can be estimated separately in
the same way as those for two independent Lee-Carter models. Under the co-
integrated model, we assume there is a linear relationship plus an error term et
between kt,1 and kt,2, that is,
kt,2 = α + β × kt,1 + et. (3.1)
Since kt,1 and kt,2 are known, the estimates of parameters α and β can be obtained
by simple linear regression method. To build the link between kt,1 and kt,2, the co-
integrated model suggests re-estimating kt,2 to get ˆkt,2 with kt,1 unchanged using
(3.1) by plugging in the values of α and β. That is,
ˆkt,2 = α + β × kt,1.
In the co-integrated model, the estimates of the two variances for population 1,
σ2εx,1
and σ2ε,1, are the same as those for the original Lee-Carter model. However,
for the second population, the drift of the time-varying index is estimated by
θ2 =ˆkt0+n−1,2 − ˆ
kt0,2n− 1
= β × kt0+n−1,1 − kt0,1n− 1
= β × θ1,
CHAPTER 3. INTRODUCTION OF MULTI-POPULATION MODELS 21
and the estimate of the variance of the error term for the time-varying index is
σ2ε,2 =
∑t0+n−1t=t0+1 (
ˆkt,2 − ˆ
kt−1,2 − θ2)2
n− 2
= β2 ×t0+n−1∑
t=t0+1
(kt,1 − kt−1,1 − θ1)2
n− 2
= β2 × σ2ε,1.
The expression of the predictive interval on qx,t0+n−1+τ,i is still the same as that
for independent Lee-Carter model. Figure 3.5 is the forecasted mortality rates
qx,2009+1,2 (female, τ = 1) and the associated 95% predictive intervals for the co-
integrated model as well as the corresponding one for the independent model for
comparison. We can see that both the forecasted mortality rates and predictive
intervals for both models are quite close to each other. Starting from age 40, the
forecasted mortality rates for the co-integrated model become slightly higher than
those for the independent model. The predictive intervals for both models almost
overlap.
CHAPTER 3. INTRODUCTION OF MULTI-POPULATION MODELS 22
Figure 3.5: 95% predictive intervals on qx,2009+1,2 for the co-integrated and indepen-dent models
CHAPTER 3. INTRODUCTION OF MULTI-POPULATION MODELS 23
3.2.4 Augmented common factor model
To avoid divergence in life expectancy in a long-run, Lee and Li (2005) suggested
adding a common factor to the original model. For the common term, it has the
following features:
• bx,1 = bx,2 = Bx for all x, and
• for the time-varying index, kt,1 = kt,2 = Kt for all t.
Two similar constrains apply, which are
• ∑2i=1
∑x wi × Bx = 1
and
• ∑t Kt = 0,
where w1 and w2 are the weights for populations 1 and 2, respectively, and w1+w2 =
1. Thus, the common factor model can be expressed as
ln(mx,t,i) = ax,i +Bx ×Kt + εx,t,i.
According to Li and Lee (2005), ax,i can still be derived by averaging ln(mx,t,i)
over a year span [t0, t0 + n− 1], that is,
t0+n−1∑
t=t0
ln(mx,t,i) =
t0+n−1∑
t=t0
ax,i +Bx ×t0+n−1∑
t=t0
Kt,
implying
ax,i =
∑t0+n−1t=t0
ln(mx,t,i)
n,
and Kt is obtained by
2∑
i=1
x0+m−1∑
x=x0
wi × [ln(mx,t,i)− ax,i] = Kt ×2∑
i=1
x0+m−1∑
x=x0
wi ×Bx,
yielding
Kt =2∑
i=1
x0+m−1∑
x=x0
wi × [ln(mx,t,i)− ax,i].
CHAPTER 3. INTRODUCTION OF MULTI-POPULATION MODELS 24
We set w1 = w2 = 0.5 for the U.S. male and female life tables. Since
2∑
i=1
wi × [ln(mx,t,i)− ax,i] = Bx × Kt ×2∑
i=1
wi = Bx × Kt,
we regress∑2
i=1 wi × [ln(mx,t,i) − ax,i] on Kt to get Bx for each age x. To fit the
data better, Li and Lee (2005) added a factor b′x,i × k′t,i for each population to the
common factor model to form the so-called augmented common factor model as
ln(mx,t,i) = ax,i +Bx ×Kt + b′x,i × k
′t,i + εx,t,i, i = 1, 2,
with an extra constrain∑
x b′x,i = 1. Notice that b′x,i and k
′t,i here are different from
bx,i and kt,i in the independent model. The constrain∑
x bx,i = 1 implies that
k′t,i =
x0+m−1∑
x=x0
[ln(mx,t,i)− ax,i − Bx × Kt].
Finally, b′x,i is derived by regressing [ln(mx,t,i) − ax,i − Bx × Kt] on k′t,i without the
constant term involved for each age x.
The logarithm of the forecasted central death rate for age x in year t0+n−1+ τ
is
ln(mx,t0+n−1+τ,i) =ax,i + Bx × (Kt0+n−1 + τ × θ +√τ × εt)
+ b′x,i × (k
′t0+n−1,i + τ × θi +
√τ × εt,i) + εx,t,i
=ln(mx,t0+n−1+τ,i) +√τ × (Bx × εt + b
′x,i × εt,i) + εx,t,i
which is normal with mean
ln(mx,t0+n−1+τ,i) = ax,i + Bx × (Kt0+n−1 + τ × θ) + b′x,i × (k
′t0+n−1,i + τ × θi),
and variance
σ2(ln(mx,t0+n−1+τ,i)) = τ × (Bx × σ2ε + b
′2x,i × σ2
ε,i) + σ2εx,i
CHAPTER 3. INTRODUCTION OF MULTI-POPULATION MODELS 25
where the three error terms, {εx,t,i}, {εt} and {εt,i}, are assumed independent,
θ =Kt0+n−1 − Kt0
n− 1
and
θ′i =
k′t0+n−1 − k
′t0
n− 1, i = 1, 2.
To construct the predictive intervals and simulate mortality rates, we need the esti-
mates of σ2εx,i
, σ2ε and σ2
ε,i, which are
σ2εx,i
=
∑t0+n−1t=t0
[ln(mx,t,i)− ax,i − Bx × Kt − b′x,i × k
′t,i]
2
n− 3, i = 1, 2,
σ2ε =
∑t0+n−1t=t0+1 (Kt − Kt−1 − θ)2
n− 2,
and
σ2ε,i =
∑t0+n−1t=t0+1 (k
′t,i − k
′t−1,i − θ
′i)
2
n− 2, i = 1, 2.
Figure 3.6 displays the forecasted qx,2009+1,i and associated predictive intervals
as well as the actual qx,2009+1,i for the augmented common factor model. When x
is small, the forecasted values are quite close to the true values. However, when
x goes up, the projected mortality rates tend to be higher than the actual ones.
Again, the predicted interval for the females is narrower than that for the males.
CHAPTER 3. INTRODUCTION OF MULTI-POPULATION MODELS 26
Figure 3.6: 95% predictive intervals on qx,2009+1,i for the augmented common factormodel
CHAPTER 3. INTRODUCTION OF MULTI-POPULATION MODELS 27
Figure 3.7 compares the predicted mortality rates among four models with the
actual mortality rates for year 2010 (τ = 1). It is obvious that all forecasted values
are close to actual ones for small x. We notice that the forecasted values from the
joint-k model are closer to the true values, while the mortality curve constructed
from the augmented common factor model is not as adjacent to the actual mortality
curve as the other models.
Figure 3.7: mortality curve comparisons among models and actual rates
Chapter 4
Application in Mortality Swap
4.1 Natural hedging and mortality swap
Mortality (longevity) risk is the risk that the number of deaths (survivors) is higher
than expected. When there is an unexpected change in mortality rates, either
life insurers or annuity providers will experience a loss. If the mortality rates in-
crease unexpectedly in a year, the number of deaths during the year is higher
than expected so that life insurance companies need to pay more death benefits.
However, in this case, annuity providers gain from the mortality increase. If the
mortality rates decline unexpectedly, the impacts on financial situation of life in-
surers and annuity providers reverse. Natural hedge is a strategy of hedging two
risks responding oppositely to a change in a common factor. Since life insurers
and annuity providers face mortality and longevity risks, respectively, both of them
can adopt natural hedge by swapping a portion of their business each other. When
a life insurer (an annuity provider) owns both life and annuity business at the same
time, the mortality and longevity risks of the portfolio can be offset to a lower level
no matter how the mortality rates change. Now, a question rises: what are the opti-
mal portions of business swapped between the life insurer and the annuity provider
such that the risk of the portfolio is minimized?
Let L denote the loss function at time zero, which is the present value of future
liabilities less the present value of future premium incomes. The values of both
future liabilities and premium incomes depend on the future mortality rates. Before
swapping the business, a life insurer (an annuity provider) has a portfolio of life
28
CHAPTER 4. APPLICATION IN MORTALITY SWAP 29
(annuity) business, and the loss function for the portfolio is denoted by Ll (La). To
hedge mortality (longevity) risk, the life insurer (annuity provider) would like to swap
wl (wa) of life (annuity) business to the annuity provider (life insurer); the resulting
loss functions of the life insurer and annuity provider become
LL = (1− wl)× Ll + wa × La
and
LA = (1− wa)× La + wl × Ll,
respectively, with variances
V ar(LL) =(1− wl)2 × V ar(Ll) + w2
a × V ar(La)
+ 2× (1− wl)× wa × Cov(Ll, La)(4.1)
andV ar(LA) =(1− wa)
2 × V ar(La) + w2l × V ar(Ll)
+ 2× (1− wa)× wl × Cov(Ll, La).(4.2)
There are three aspects to approach the optimal pair of weights which min-
imizes the variance of a loss function or the sum of the variances of two loss
functions. The first pair of weights, (wL+Al , wL+A
a ), is used to minimize V ar(LL) +
V ar(LA) (see Figure 4.1); the second pair of weights, (wLl , w
La ), is used to mini-
mized V ar(LL) (see Figure 4.2), and the third pair of weights, (wAl , w
Aa ), is used to
minimize V ar(LA) (see Figure 4.3), where we place superscripts L+A, L and A on
wl and wa to denote the weights that minimize V ar(LL) + V ar(LA), V ar(LL) and
V ar(LA), respectively.
CHAPTER 4. APPLICATION IN MORTALITY SWAP 30
Figure 4.1: Mortality swap: minimizing V ar(LL) + V ar(LA)
Figure 4.2: Mortality swap: minimizing V ar(LL)
Figure 4.3: Mortality swap: minimizing V ar(LA)
CHAPTER 4. APPLICATION IN MORTALITY SWAP 31
In mathematical optimization, the method of Lagrange multipliers is a strat-
egy for finding the local maximum (minimum) of a function subject to some con-
straint(s). Let Pl (Pa) stand for the present value of the future premiums of all
life (annuity) policies in the portfolio before swap. When the life insurer (annuity
provider) swaps wl (wa) of life (annuity) policies to the annuity provider (life insurer),
the life insurer (annuity provider) loses premium wl×Pl (wa×Pa) and gets premium
wa × Pa (wl × Pl). We set a swap condition wl × Pl = wa × Pa which will be applied
as the constraint in the three optimization problems mentioned above using the
method of Lagrange multipliers. Specifically, we would like to find (wl, wa) which
minimizes f(wl, wa) subject to wl × Pl = wa × Pa where f is V ar(LL) + V ar(LA),
V ar(LL) or V ar(LA).
To obtain (wL+Al , wL+A
a ), define f(wL+Al , wL+A
a ) = V ar(LL) + V ar(LA). By (4.1)
and (4.2),
f(wL+Al , wL+A
a ) =(1− wL+Al )2 × V ar(Ll) + (wL+A
a )2 × V ar(La)
+ 2× (1− wL+Al )× wL+A
a × Cov(Ll, La)
+ (1− wL+Aa )2 × V ar(La) + (wL+A
l )2 × V ar(Ll)
+ 2× (1− wL+Aa )× wL+A
l × Cov(Ll, La).
According to the method of Lagrange multipliers with a constraint wL+Al × Pl =
wL+Aa × Pa, the Lagrange function is defined by
ϕ(wL+Al , wL+A
a , λ) = f(wL+Al , wL+A
a ) + λ(wL+Al × Pl − wL+A
a × Pa),
where λ is called a Lagrange multiplier. To obtain the optimal solution, we differen-
tiate ϕ with respect to wL+Al , wL+A
a , and λ, respectively, and set all results to zero.
That is,
∂ϕ
∂wL+Al
= 4wL+Al Vl − 4wL+A
a σ2 + 2σ2 − 2Vl + λPl = 0,
∂ϕ
∂wL+Aa
= 4wL+Aa Va − 4wL+A
l σ2 + 2σ2 − 2Va − λPa = 0,
CHAPTER 4. APPLICATION IN MORTALITY SWAP 32
and
∂ϕ
∂λ= wL+A
l Pl − wL+Aa Pa = 0,
where Vl = V ar(Ll), Va = V ar(La) and σ2 = Cov(Ll, La). Then (wL+Al , wL+A
a ) can
be solved as
wL+Al =
1
2× Pa × PaVl + PlVa − (Pl + Pa)σ
2
P 2aVl − 2PlPaσ2 + P 2
l Va
(4.3)
and
wL+Aa =
1
2× Pl × PaVl + PlVa − (Pl + Pa)σ
2
P 2aVl − 2PlPaσ2 + P 2
l Va
. (4.4)
Similarly, when f(wLl , w
La ) = V ar(LL) with a constraint wL
l × Pl = wLa × Pa, the
optimal weights, wLl and wL
a , become
wLl =Pa × PaVl − Plσ
2
P 2aVl − 2PlPaσ2 + P 2
l Va
(4.5)
and
wLa =Pl × PaVl − Plσ
2
P 2aVl − 2PlPaσ2 + P 2
l Va
; (4.6)
when f(wAl , w
Aa ) = V ar(LA) with a constraint wA
l ×Pl = wAa ×Pa, the optimal weights
wAl and wA
a are
wAl =Pa × PlVa − Paσ
2
P 2aVl − 2PlPaσ2 + P 2
l Va
(4.7)
and
wAa =Pl × PlVa − Paσ
2
P 2aVl − 2PlPaσ2 + P 2
l Va
. (4.8)
Note that wL+Al = 1
2(wL
l + wAl ) and wL+A
a = 12(wL
a + wAa ). It is very difficult to obtain
the theoretical expressions of Vl, Va and σ2. Instead, we would like to compute the
corresponding sample variances and covariance by simulating thousands of Vl, Va
and σ2.
CHAPTER 4. APPLICATION IN MORTALITY SWAP 33
4.2 Numerical illustrations
4.2.1 Assumptions and portfolios
In practice, a life (an annuity) portfolio consists of a variety of life (annuity) products.
For simplicity, we assume that the portfolio for the life insurer consists of (65 − x)-
payment whole life insurance issued to the insureds aged x = 25 ∼ 64 and the
death benefits are paid at the end of the year of death, and that the portfolio for the
annuity provider is composed of (65−x)-payment and (65−x) years deferred whole
life annuity due issued to the insureds aged x = 25 ∼ 64. Since life insurance
(annuities) are more often purchased by those who have poorer (better) health
conditions, and we have no life (annuity) tables for consecutive years for forecasting
mortality rates with the models proposed in the preceding chapter, we use the U.S.
male (female) mortality table for the life (annuity) insureds. Because the year span
[t0, t0 + n − 1] is used for estimating the parameters of the mortality models, and
forecasting the mortality rates for years t0 + n − 1 + τ , τ = 1, 2, ...., we set the
beginning of year t1 = t0 + n as time 0. Denote lx,t1,i the initial number of insureds
aged x at time 0 for population i (i = 1 for life and i = 2 for annuity) and we
set l25,t1,1 = l25,t1,2 = 107. By lx+1,t1,i = lx,t1,i × px,t1−1,i, x = 25, ..., 63, the initial
numbers of insureds for the entire portfolio can be obtained. The death benefit for
life insurance is Bl = 100, 000 and the annual survival benefit is Ba = 10, 000. Based
on the assumptions, the loss functions of the life insurer and annuity provider at
time zero are
Ll =64∑
x=25
lx,t1,1 × (Ax,1 × Bl − ax:65−x|,1 × Px,l), (4.9)
La =64∑
x=25
lx,t1,2 × (65−x|ax,2 × Ba − ax:65−x|,2 × Px,a), (4.10)
Ax,1 =100−x∑
k=0
kpx,t1,1 · qx+k,t1+k,1 · vk+1, (4.11)
ax:65−x|,i =65−x−1∑
k=0
kpx,t1,i · vk, i = 1, 2, (4.12)
CHAPTER 4. APPLICATION IN MORTALITY SWAP 34
65−x|ax,2 =100∑
k=65
(k−x)px,t1,2 · vk−x, (4.13)
Px,l =Ax,1 × Bl
ax:65−x|,1, (4.14)
andPx,a =
65−x|ax,2 × Ba
ax:65−x|,2, (4.15)
where kpx,t1,i =∏k−1
j=0 px+j,t1+j,i with 0px,t1,i = 1, v = (1 + i)−1 and i = 6% is the
interest rate. Note that we set the limiting age equal to 100 for both populations, that
is, q100,t,i = 1, i = 1, 2. Moreover, Pl and Pa, the total present values of the future
premiums for the life insurer and annuity provider, respectively, in (4.3) ∼ (4.8) are
(see (4.9) and (4.10))
Pl =64∑
x=25
lx,t1,1 × ax:65−x|,1 × Px,l
and
Pa =64∑
x=25
lx,t1,2 × ax:65−x|,2 × Px,a.
The premiums Px,l and Px,a in (4.14) and (4.15), and the total premiums Pl and Pa
are pre-determined and do not respond to a change in mortality rates, whereas
Ax,1, ax:65−x|,i and 65−x|ax,2 in (4.11), (4.12) and (4.13) vary in response to the
realized mortality rates. Therefore, the deterministic mortality rates, qx,t0+n−1+τ ,
τ = 1, 2, ..., are used to calculate the premiums Px,l and Pa,x, and the stochastic
ones, qx,t0+n−1+τ , τ = 1, 2, ..., are used for simulations to compute Ax,1, ax:65−x|,i and
65−x|ax,2. When the realized mortality rates are different from the expected ones,
each of the loss functions Ll and La is either positive or negative. To forecast de-
terministic and stochastic mortality rates for ages 25 ∼ 64 and years 2011 ∼ 2086
using the four models given in Chapter 3, we set the year span [1981, 2010] and age
span [25, 100] (see Figure 4.4) to estimate the parameters with the male and female
mortality data from the Human Mortality Database for the life and annuity policies,
respectively. Specifically, for the independent model, we give the following steps to
forecast deterministic mortality rates for computing the premiums Px,l and Px,a:
CHAPTER 4. APPLICATION IN MORTALITY SWAP 35
Figure 4.4: age span [25, 100] and year span [1981, 2010]
1. compute ln(mx,2010+τ,i) = ax,i + bx,i × (k2010,i + τ × θi), i = 1, 2, τ = 1, 2, ..., 76,
x = 25, ..., 64;
2. transfer ln(mx,2010+τ,i) to qx,2010+τ,i; and
3. take the diagonal entries qx+τ−1,2010+τ,i, i = 1, 2, x = 25, ..., 64, τ = 1, 2, ..., (101−x).
To generate stochastic mortality rates for simulating Ax,1, ax:65−x|,i and 65−x|ax,2, a
similar procedure is given as follows:
1. generate sε,τ,i from N(0, 1) and sε,τ,i from N(0, 1), i = 1, 2, τ = 1, 2, ..., 76;
2. multiply sε,τ,i by σε,i, and sε,τ,i by σεx,i such that sε,τ,i × σε,i ∼ N(0, σ2ε,i) and
sε,τ,i × σεx,i ∼ N(0, σ2εx,i);
3. get simulated ln(mx,2010+τ,i) = ln(mx,2010+τ,i)+√τ×bx,i×sε,τ,i×σε,i+sε,τ,i×σεx,i,
i = 1, 2, τ = 1, 2, ..., 76, x = 25, ..., 64;
4. transfer ln(mx,2010+τ,i) to qx,2010+τ,i;
5. take the diagonal entries qx+τ−1,2010+τ,i, i = 1, 2, x = 25, ..., 64, τ = 1, 2, ..., (101−x); and
6. repeat steps (1) ∼ (5) for N times (N = 1, 000 in this project).
CHAPTER 4. APPLICATION IN MORTALITY SWAP 36
The procedures of forecasting the deterministic and stochastic mortality rates
for the joint-k, the co-integrated and the augmented common factor models are
similar to those above. With {qx+τ−1,2010+τ,i : i = 1, 2, x = 25, ..., 64, τ = 1, ..., (101−x)} and N {qx+τ−1,2010+τ,i : i = 1, 2, x = 25, ..., 64, τ = 1, ..., (101 − x)}’s, we can
calculate N realized values of Ll and La, get the sample variances Vl and Va and
the sample covariance σ2, and obtain the optimal weights with (4.3) ∼ (4.8).
4.2.2 Robustness testing
In the preceding subsection, for each of three optimization problems, a pair of opti-
mal weights for life and annuity portfolios is produced by some stochastic mortality
model which generates N cohort mortality rates from age x to age 100 at time zero
for each of x = 25, ..., 64. If we re-run the procedure and generate another set of
N cohort mortality rates, can we still produce a pair of optimal weights of close
values? In this subsection, we will perform robustness testing. Robustness testing
is originally used in computer science whether a computer system can continue
to work well in case of invalid inputs. In our case, robustness testing is a way to
investigate whether the optimal weights produced by a model is insensitive to the
simulated mortality rates.
To complete the robustness testing, we repeat the simulation procedure M (M
is set to 50) times, and yield M pairs of optimal weights for each model. Figures
4.5 and 4.6 show scatter plots for the optimal weights (wl, wa) generated from each
model, from which we can see that the 50 pairs of weights obtained through 50
simulation procedures for each model are quite close to each other. That means
the four models are robust to simulations. Within each model, the first two pairs
of optimal weights (wL+Al , wL+A
a ) and (wLl , w
La ) seem to be more consistent than
(wAl , w
Aa ).
Table 4.1 summarizes the median values of 50 optimal weights for each type
obtained from the four models. The weight wl is more than twice as big as wa for all
types because wl/wa = Pa/Pl ≈ 2.3. All the optimal weights are within (0, 1) except
for wAl for the independent model. The pairs of optimal weights (wl, wa) from the
joint-k and co-integrated models are quite close to each other. The independent
model produces the largest (wL+Al , wL+A
a ) and (wAl , w
Aa ) but the smallest (wL
l , wLa ),
CHAPTER 4. APPLICATION IN MORTALITY SWAP 37
whereas the augmented common factor model yields the lowest (wL+Al , wL+A
a ) and
(wAl , w
Aa ) but the highest (wL
l , wLa ) among the four models.
IND JK Co-Int ACFwl wa wl wa wl wa wl wa
L+A 0.8505 0.3623 0.7385 0.3146 0.7431 0.3169 0.7110 0.3029L 0.4797 0.2044 0.6459 0.2752 0.6384 0.2723 0.6868 0.2926A 1.2213 0.5203 0.8311 0.3541 0.8479 0.3616 0.7352 0.3132
Table 4.1: the median of 50 optimal weights
Figure 4.7 displays V ar(LL)’s and V ar(LA)’s for four models based on 50 cor-
responding (wL+Al , wL+A
a )’s, (wLl , w
La )’s and (wA
l , wAa )’s, respectively. These figures
further confirm the comments on Table 4.1 above, and show the variability of 50
variances. Under the joint-k and co-integrated models, the variances from 50 runs
of simulations are quite close and smaller, whereas for the independent and aug-
mented common factor models, the sample variances are higher and not as stable
as those based on the joint-k and co-integrated models.
Figures 4.8, 4.9 and 4.10 exhibit the simulated loss distributions before and af-
ter swap using the median optimal weights for all four models. It is obvious that
the loss distributions after swap for the life insurer and annuity provider are al-
most narrowed, which implies that the variance of the loss distribution is reduced
significantly. No matter using (wL+Al , wL+A
a ), (wLl , w
La ) or (wA
l , wAa ), the loss distri-
butions for the joint-k and co-integrated models after swap for the annuity provider
are much narrower than those for the independent and augmented common factor
models, and for the life insurer, there is not much difference in the loss distribu-
tions after swap among the joint-k, co-integrated and augmented common factor
models.
To future quantify and compare the performances of hedging mortality and
longevity risks after swap for the life insurer and annuity provider, respectively,
we give a measure called hedge effectiveness (HE; see Li and Hardy (2011)) as
follows:
HE(L+ A) = 1− V ar(LL) + V ar(LA)
V ar(Ll) + V ar(La),
CHAPTER 4. APPLICATION IN MORTALITY SWAP 38
HE(L) = 1− V ar(LL)
V ar(Ll),
and
HE(A) = 1− V ar(LA)
V ar(La).
The HE measure is a variance reduction (variance of a loss function before hedge
less the variance of the loss function after hedge) ratio. Clearly, the larger the HE
is, the more effective the hedge is. Table 4.2 shows the comparisons of HEs, which
are consistent with the results from Figures 4.8, 4.9 and 4.10. The independent
model overall performs the worst among all models. The HE(L), HE(A) and
HE(L + A) for the joint-k model are the largest among the four models, which
implies that the joint-k model is the most effective in hedging mortality and longevity
risks. However, the co-integrated model produces the smallest variances of the
losses before and after swap for both the life insurer and annuity provider.
Independent V ar(Ll) V ar(LL) V ar(La) V ar(LA) HE(L+ A) HE(L) HE(A)
(wL+Al , wL+A
a ) 1.8904 1.5384 11.2661 6.1193 0.4180 0.1862 0.4568(wL
l , wLa ) 1.8904 1.0158 11.2661 7.6872 0.3385 0.4627 0.3177
(wAl , w
Aa ) 1.8904 23.1063 11.2661 5.5967 0.3385 -0.6433 0.5032
Joint-K V ar(Ll) V ar(LL) V ar(La) V ar(LA) HE(L+ A) HE(L) HE(A)
(wL+Al , wL+A
a ) 2.6716 0.3320 5.5132 1.6081 0.7630 0.8757 0.7083(wL
l , wLa ) 2.6716 0.2829 5.5132 1.7553 0.7510 0.8941 0.6816
(wAl , w
Aa ) 2.6716 0.4792 5.5132 1.5590 0.7510 0.8206 0.7172
Co-Integrated V ar(Ll) V ar(LL) V ar(La) V ar(LA) HE(L+ A) HE(L) HE(A)
(wL+Al , wL+A
a ) 1.8272 0.3185 4.2565 1.5635 0.6906 0.8257 0.6327(wL
l , wLa ) 1.8272 0.2768 4.2565 1.6887 0.6769 0.8485 0.6033
(wAl , w
Aa ) 1.8272 0.4437 4.2565 1.5218 0.6769 0.7572 0.6425
ACF V ar(Ll) V ar(LL) V ar(La) V ar(LA) HE(L+ A) HE(L) HE(A)
(wL+Al , wL+A
a ) 6.3736 1.8738 15.4550 10.2970 0.4424 0.7060 0.3337(wL
l , wLa ) 6.3736 1.8682 15.4550 10.3138 0.4419 0.7069 0.3320
(wAl , w
Aa ) 6.3736 1.8906 15.4550 10.2914 0.4419 0.7034 0.3341
Table 4.2: comparisons of sample variances (×1023) and HE’s
CHAPTER 4. APPLICATION IN MORTALITY SWAP 39
wL+Al , wL+A
a wL+Al , wL+A
a
wLl , w
La wL
l , wLa
wAl , w
Aa wA
l , wAa
Figure 4.5: optimal weights for the independent and joint-k models
CHAPTER 4. APPLICATION IN MORTALITY SWAP 40
wL+Al , wL+A
a wL+Al , wL+A
a
wLl , w
La wL
l , wLa
wAl , w
Aa wA
l , wAa
Figure 4.6: optimal weights for the co-integrated and augmented common factormodels
CHAPTER 4. APPLICATION IN MORTALITY SWAP 41
V ar(LL)
variances using (wL+Al , wL+A
a )
V ar(LA)
variances using (wL+Al , wL+A
a )
variances using (wLl , w
La ) variances using (wL
l , wLa )
variances using (wAl , w
Aa ) variances using (wA
l , wAa )
Figure 4.7: variances after swap
CHAPTER 4. APPLICATION IN MORTALITY SWAP 42
Figure 4.8: simulated loss distributions using (wL+Al , wL+A
a )
CHAPTER 4. APPLICATION IN MORTALITY SWAP 43
Figure 4.9: simulated loss distributions using (wLl , w
La )
CHAPTER 4. APPLICATION IN MORTALITY SWAP 44
Figure 4.10: simulated loss distributions using (wAl , w
Aa )
Chapter 5
Block Bootstrap Method
In the preceding chapter, we forecast deterministic and stochastic mortality rates
with some two-population mortality models to determine the premiums of life and
annuity products and simulate the loss functions of portfolios of life and annuity
business. The weights of business for swap are calculated by the sample variances
and the sample covariance of the loss functions of the life insurer and annuity
provider before swap, which can minimize the risk of the portfolio and produce
high hedge effectiveness. All the great results are based on that the assumed
mortality model is the actual one, which, however, might not be true. Therefore,
both the life insurer and annuity provider face model risk and parameter risk that
will potentially affect the results. In this chapter, a bootstrap method which is model
and parameter free will be applied to generating samples for the future mortality
rates to calculate the weighted loss functions and their variances with the weights
obtained by each of the four models in Chapter 3.
Bootstrap is usually used to resample data with replacement to estimate some
statistic of a population from the sampled data. The procedure of the bootstrap
(naive bootstrap) is as follows:
1. Draw values from the original data set with replacement to form a new data
set of size n∗.
2. Repeat the first step N∗ times to obtain N∗ new data sets.
3. Compute the test statistic using the new data sets.
45
CHAPTER 5. BLOCK BOOTSTRAP METHOD 46
However, the naive bootstrap fails when applied to the mortality rates. First of
all, mortality rates over years can be treated as a time series displaying a decreas-
ing trend because of the improvement of medical and environmental conditions,
which shows mortality rates are not stationary. Secondly, the naive bootstrap is
likely to destroy the dependency of mortality rates on both age and time dimen-
sions. Alternatively, the block bootstrap can solve the problems above. First, re-
garding the non-stationary problem, differencing is a popular and effective method
of removing trend from a time series and making the time series weakly stationary.
The procedure is given below.
• Convert the empirical mortality rates qx,t,i to ln(mx,t,i), i = 1, 2, x = 25, ..., 100,
t = 1981, ..., 2009.
• Devide ln(mx,t+1,i) by ln(mx,t,i) to get the ratio, denoted by rx,t,i, that is, rx,t,i =ln(mx,t+1,i)
ln(mx,t,i), t = 1981, ..., 2009.
• Subtract rx,t,i from rx,t+1,i to get the difference, denoted by dx,t,i, that is, dx,t,i =
rx,t+1,i − rx,t,i, t = 1981, ..., 2008.
Figure 5.1 shows the time series {dx,t,i}, i = 1, 2, for the U.S. males and females
aged 35, 45, 55 and 65, which look stationary.
CHAPTER 5. BLOCK BOOTSTRAP METHOD 48
To further ensure the time series {dx,t,i} is stationary, the Phillips-Perron (PP)
and the Kwiatkowski-Phillips-Schmidt-Shin (KPSS) hypothesis tests are conducted.
For the PP test where an AR(1) model is assumed, the null hypothesis is that the
time series has a unit root so that the time series is non-stationary, which implies
that a small p-value suggests a stationary time series. For the KPSS test, the null
hypothesis is that the time series is stationary; thus, a big p-value indicates a sta-
tionary time series. Table 5.1 exhibits the results of the hypothesis tests. All the
p-values for the PP test are below 0.05, and the ones for the KPSS test are larger
than 0.1, which all suggest that the time series {dx,t,i} is stationary.
35 45 55 65
Male PP 0.01 0.01 0.01 0.01KPSS >0.1 >0.1 >0.1 >0.1
Female PP 0.01 0.01 0.01 0.01KPSS >0.1 >0.1 >0.1 >0.1
Table 5.1: p-values for testing stationarity of {dx,t,i}
For the second problem regarding dependency, according to Li and Ng (2011),
the two-dimensional mortality rates matrix Mi for population i is converted to a
series of column vectors,
mt,i = (mx0,t,i,mx0+1,t,i, ...,mx0+m−1,t,i)′, t = t0, t0 + 1, ..., t0 + n− 1.
Thus, the matrix can be expressed as Mi = {mt0,i,mt0+1,i, ...,mt0+n−1,i}, which
contains n elements. In this way, the age dependency can be retained. Before
proceeding to the next step, as discussed in the first problem above, the vector
mt,i needs to be converted to ln(mt,i), get the ratio vector rt,i and then form the
difference vector
dt,i = (dx0,t0+t−1,i, dx0+1,t0+t−1,i, ..., dx0+m−1,t0+t−1,i)′, t = 1, 2, ..., n− 2, i = 1, 2.
For the time dependency problem, it’s assumed that k consecutive time series,
dt,i, ...,dt+k−1,i, are dependent. While the dependency between dt,i and dt+k−1,i
gradually becomes weaker as k increases, and it’s going to vanish thoroughly for a
large k. Thus, the block bootstrapping suggests splitting d into n overlapped blocks
CHAPTER 5. BLOCK BOOTSTRAP METHOD 49
Figure 5.2: a circle diagram of dt,i’s
of size k as follows:
D ={(d1,i,d2,i, ...,dk,i),
(d2,i,d3,i, ...,dk+1,i),
(d3,i,d4,i, ...,dk+2,i),
...,
(dn−2,i,d1,i ...,dk−1,i)}
Figure 5.2 gives a circle diagram of dt,i’s. We make all dt,i’s a circle and every con-
secutive k dt,i’s form a block. After the (n−2) blocks of size k, which are numbered
1, 2, ..., (n − 2), have been created, the block number will be drawn with replace-
ment to form a sample. The values within each block don’t change at all so that
the dependent structure among k dt,i’s is maintained. In Chapter 4, we simulate 76
time series of {q2010+τ,i}’s, τ = 1, 2, ..., 76, for computing one value of loss functions
LL and LA. To simulate the same number of time series of {q2010+τ,i}’s with the
block bootstrap method, we need to draw Z = Wk
block numbers with replacement
as a bootstrap sample, where W = 76 and k is the block size. If Wk
is not an integer,
then (z + 1) blocks need to be drawn and take the first s elements of the (z + 1)th
block to make the sample size equal to W where z and s are the integer and the
remainder of Wk
, respectively.
CHAPTER 5. BLOCK BOOTSTRAP METHOD 50
Finally, the block size is determined by the size of observations V . As sug-
gested by Hall et al. (1995), the optimal block size, k, can be V13 , V
14 or V
15 . Since
there are 28 {dt,i}’s (V = n− 2 = 28), a block size of 2 is chosen.
The following is the procedure of the block bootstrap method.
1. Take the logarithm on the real central death rates, ln(mx,t,i), for i = 1, 2,
x = x0, ..., x0 +m− 1, and t = t0, ..., t0 + n− 1.
2. Get rx,t,i =ln(mx,t+1,i)
ln(mx,t,i)and dx,t,i = rx,t+1,i − rx,t,i.
3. Form column vector dt,i = (dx0,t0+t−1,i, dx0+1,t0+t−1,i, ..., dx0+m−1,t0+t−1,i)′, t =
1, ..., n− 2.
4. Form the jth block of size 2, (dj,i,dj+1,i), j = 1, ..., n − 3, and the (n − 2)th
block, (dn−2,i,d1,i).
5. Draw a bootstrap sample of size Z = W2
if W is even, or Z = (W+1)2
if W is
odd to get the block numbers b1, b2,..., bZ (all the indices fall between 1 and
n− 2).
6. Get the indices, b1, b1 +1, b2, b2 +1,..., bZ , bZ +1 if W is even, or b1, b1 +1, b2,
b2 +1,..., bZ if W is odd; if bj = n− 2 for some j, then assign index 1 to bj +1.
Denote the jth index as cj, j = 1, ...,W .
7. Obtain the column vector dcj ,i = (dx0,t0+cj−1,i, ..., dx0+m−1,t0+cj−1,i)′, j = 1, ...,W .
8. Obtain simulated mx,t0+n−1+τ,i by
mx,t0+n−1+τ,i = exp{ln(mx,t0+n−1+τ−1,i)× [ln(mx,t0+n−1,i)
ln(mx,t0+n−2,i)+
τ∑
j=1
dx,t0+cj−1,i]},
i = 1, 2, τ = 1, 2, ..., 76, x = x0, x0 + 1, ..., x0 +m− 1.
9. Convert simulated mx,t0+n−1+τ,i matrix to qx,t0+n−1+τ,i matrix and take the di-
agonal entries to form cohort mortality sequences.
10. Repeat (3)-(9) for N times (N = 1000).
CHAPTER 5. BLOCK BOOTSTRAP METHOD 51
The median weights and premiums calculated in the preceding chapter are
used in simulating the loss functions before and after swap using the block boot-
strap method. By applying the pre-determined premiums and 1000 simulated mor-
tality paths to (4.9) and (4.10), 1000 La’s and Ll’s can be obtained, from which
sample V ar(Ll), V ar(La) and Cov(Ll, La) can be calculated. Plugging V ar(Ll),
V ar(La) and Cov(Ll, La) into (4.1) and (4.2) with the pre-determined weights, the
variances of the loss functions LL and LA after swap, V ar(LL) and V ar(LA), are
obtained.
Figures 5.3, 5.4 and 5.5 display the simulated loss distributions before and after
swap. Compared with the loss distributions in Figures 4.8, 4.9 and 4.10, the overlap-
ping between the simulated loss distributions before and after swap becomes far
smaller. The loss distributions before and after swap for the life insurer mainly fall
in the negative (gain) and positive (loss) territories, respectively, whereas those for
the annuity provider largely spread in the negative (gain) territory. Under the block
bootstrap method, the life insurer will benefit more gains from swap and the annuity
provider will suffer more losses (or less gains) from swap than under the parametric
mortality models. Moreover, even though the simulated loss functions generated
from the block bootstrap method exhibit big differences from those based on the
four mortality models, the pre-determined weights still lower down the risks for
both insurance and annuity portfolios, which can be seen from the narrowed loss
distribution curves after swap.
CHAPTER 5. BLOCK BOOTSTRAP METHOD 52
Figure 5.3: simulated loss distributions using (wL+Al , wL+A
a )
CHAPTER 5. BLOCK BOOTSTRAP METHOD 55
Similar to Table 4.2, Table 5.2 shows the sample variances and hedge effec-
tiveness (HE) under the block bootstrap method. In this case, the differences
in HE and V ar(L) among four models are not as obvious as those using multi-
population mortality models because the same simulated mortality rates from the
block bootstrap method are applied to the loss functions before and after swap with
different (wl, wa)’s from four mortality models. Compared with the values in Table
4.2, the variances of the loss functions before and after swap for both life insurer
and annuity provider are far enlarged except for V ar(LL) based on (wAl , w
Aa ) for
the independent model. The HEs generally increase for the independent model,
and decrease for the joint-k and co-integrated models. Among the four models,
the co-integrated and independent models achieve the highest and the lowest
hedge effectiveness, respectively, for HE(L + A), HE(L) and HE(A) based on
(wL+Al , wL+A
a ), (wLl , w
La ) and (wA
l , wAa ), respectively. The joint-k model can also pro-
duce high hedge effectiveness. Thus, under model-free block bootstrap method,
the joint-k and co-integrated models still outperform the other two models.
Independent V ar(Ll) V ar(LL) V ar(La) V ar(LA) HE(L+ A) HE(L) HE(A)
(wL+Al , wL+A
a ) 12.1508 5.6147 43.0437 23.2084 0.4778 0.5379 0.4608(wL
l , wLa ) 12.1508 4.4826 43.0437 27.8747 0.4138 0.6311 0.3524
(wAl , w
Aa ) 12.1508 12.9029 43.0437 24.6982 0.3188 -0.0619 0.4262
Joint-K V ar(Ll) V ar(LL) V ar(La) V ar(LA) HE(L+ A) HE(L) HE(A)
(wL+Al , wL+A
a ) 12.1509 4.6247 43.0495 23.9732 0.4819 0.6194 0.4431(wL
l , wLa ) 12.1509 4.2296 43.0495 25.0259 0.4700 0.6519 0.4187
(wAl , w
Aa ) 12.1509 5.4035 43.0495 23.3042 0.4799 0.5553 0.4587
Co-Integrated V ar(Ll) V ar(LL) V ar(La) V ar(LA) HE(L+ A) HE(L) HE(A)
(wL+Al , wL+A
a ) 12.1508 4.6638 43.0511 23.9150 0.4823 0.6162 0.4445(wL
l , wLa ) 12.1508 4.2211 43.0511 25.1140 0.4686 0.6526 0.4166
(wAl , w
Aa ) 12.1508 5.5980 43.0511 23.2074 0.4782 0.5393 0.4782
ACF V ar(Ll) V ar(LL) V ar(La) V ar(LA) HE(L+ A) HE(L) HE(A)
(wL+Al , wL+A
a ) 12.1231 4.4638 43.0398 24.2210 0.4800 0.6318 0.4372(wL
l , wLa ) 12.1231 4.3527 43.0398 24.4902 0.4771 0.6410 0.4310
(wAl , w
Aa ) 12.1231 4.6011 43.0398 23.9780 0.4819 0.6205 0.4429
Table 5.2: comparisons of sample variances (×1023) and HE’s (block bootstrap)
Chapter 6
Conclusion
The strategy of natural hedging in this project shows a significant effect on reducing
risks of insurance and annuity portfolios by swapping business. The robustness
testing exhibits that the optimal weights and the variances of the loss functions after
swap for both life insurer and annuity provider obtained from four multi-population
models are robust to simulations.
The performances of hedging mortality and longevity risks for each model are
compared in two ways; one is based on parametric multi-population mortality mod-
els, and the other is based on non-parametric block bootstrap method. Both ways
suggest that the joint-k and co-integrated models outperform the independent and
augmented common factor models, which implies that the improvement on mortal-
ity rates between two populations tend to become more related, and assuming a
stronger bond between two time-varying factors for two populations seems to be
more reasonable when forecasting future mortality rates.
Although natural hedging can achieve the goal of reducing the variance of the
loss, future research can still be carried out so that it can be further applied to more
practical situations. For example, both a life insurer and an annuity provider won’t
swap their business directly in practice. Generally, companies prefer to transfer
their business to a financial intermediary called SPV (special purpose vehicle) who
is in charge of business swapping. Another problem is that the portfolio of life
(annuity) business in this project consists of only one type of life insurance (annuity)
product, the (65− x)-payment whole life insurance ((65− x)-payment and (65− x)
deferred whole life annuity due), which is too simple to be practical.
56
Bibliography
[1] D. Blake and W. Burrows. Survivor bonds: Helping to hedge mortality risk.Journal of Risk and Insurance, 68:339–348, 2001.
[2] N.L. Bowers, H.U. Gerber, J.C. Hickman, D.A. Jones, and C.J. Nesbitt. Actu-arial mathematics. 2nd ed. The Society of Actuaries, 1997.
[3] S.H. Cox and Y. Lin. Natural hedging of life and annuity mortality risks. NorthAmerican Actuarial Journal, 11:1–15, 2007.
[4] P. Gaillardetz, H.Y. Li, and A. MacKay. Equity-linked products: Evaluation ofthe dynamic hedging errors under stochastic mortality. European ActuarialJournal, 2:243–258, 2012.
[5] P. Hall, J.L. Horowitz, and B.Y. Jing. On blocking rules for the bootstrap withdependent data. Biometrika, 82:561–574, 1995.
[6] R.D. Lee and L.R. Carter. Modeling and forecasting U.S. mortality. Journal ofthe American Statistical Association, 87:659–671, 1992.
[7] J.S. Li and M.R. Hardy. Measuring basis risk in longevity hedges. NorthAmerican Actuarial Journal, 15:177–200, 2011.
[8] J.S. Li and A.C. Ng. Canonical valuation of mortality-linked securities. TheJournal of Risk and Insurance, 78:853–884, 2011.
[9] N. Li and R. Lee. Coherent mortality forecasts for a group of populations: Anextension of the Lee-Carter method. Demography, 42:575–594, 2005.
[10] T. Lin and C.C.L. Tsai. On the mortality/longevity risk hedging with mortalityimmunization. Insurance: Mathematics and Economics, 53:580–596, 2013.
[11] A. Milidonis, Y. Lin, and S.H. Cox. Mortality regimes and pricing. North Amer-ican Actuarial Journal, 15:266–289, 2011.
[12] J.L. Wang, H.C. Huang, S.S. Yang, and J.T. Tsai. An optimal product mix forhedging longevity risk in life insurance companies: The immunization theoryapproach. The Journal of Risk and Insurance, 77:473–497, 2010.
[13] N. Zhu and D. Bauer. A cautionary note on natural hedging of longevity risk.North American Actuarial Journal, 18:104–115, 2014.
57