Mean-variance portfolio optimization when means and...

49
Introduction and review A high dimensional plug-in covariance matrix estimator A modified Markowitz framework Conclusion Mean-variance portfolio optimization when means and covariances are estimated Zehao Chen June 1, 2007 Joint work with Tze Leung Lai (Stanford Univ.) and Haipeng Xing (Columbia Univ.) Zehao Chen M.V. optimization when means and covariances are estimated

Transcript of Mean-variance portfolio optimization when means and...

Introduction and reviewA high dimensional plug-in covariance matrix estimator

A modified Markowitz frameworkConclusion

Mean-variance portfolio optimization when means

and covariances are estimated

Zehao Chen

June 1, 2007

Joint work with Tze Leung Lai (Stanford Univ.) and Haipeng Xing (ColumbiaUniv.)

Zehao Chen M.V. optimization when means and covariances are estimated

Introduction and reviewA high dimensional plug-in covariance matrix estimator

A modified Markowitz frameworkConclusion

Outline

1 Introduction and reviewThe Markowitz frameworkDifferent efficient frontier definitions under stochastic setting

2 A high dimensional plug-in covariance matrix estimatorThe L2 boosting estimatorSimulation and empirical study

3 A modified Markowitz frameworkThe frameworkAn example and simulation

4 Conclusion

Zehao Chen M.V. optimization when means and covariances are estimated

Introduction and reviewA high dimensional plug-in covariance matrix estimator

A modified Markowitz frameworkConclusion

The Markowitz frameworkDifferent efficient frontier definitions under stochastic setting

The basic formulation

We denote the returns of p risky assets (e.g. stock returns) by ap × 1 vector R , and an unobserved future return by r ,

E(R) = µ, Cov(R) = Σ

The mean-variance optimization solves for an asset allocation w ,which minimizes the portfolio risk σ2

w , while achieving a certaintarget return µ∗, i.e,

minw

wTΣw = minw

σ2w

subject to

wTµ ≥ µ∗, w1 + w2 + ... + wp = 1

Zehao Chen M.V. optimization when means and covariances are estimated

Introduction and reviewA high dimensional plug-in covariance matrix estimator

A modified Markowitz frameworkConclusion

The Markowitz frameworkDifferent efficient frontier definitions under stochastic setting

The basic formulation

This formulation has a closed form solution for w ,

w∗ = f (µ, µ∗,Σ−1)

The weights sometimes have additional constraints, i.e.

li ≤ wi ≤ ui , i = 1, ..., p

If li ≥ 0, the constraint is also called no short selling constraint.The solution under this additional constraint requires quadraticprogramming.

Zehao Chen M.V. optimization when means and covariances are estimated

Introduction and reviewA high dimensional plug-in covariance matrix estimator

A modified Markowitz frameworkConclusion

The Markowitz frameworkDifferent efficient frontier definitions under stochastic setting

Efficient frontier

Zehao Chen M.V. optimization when means and covariances are estimated

Introduction and reviewA high dimensional plug-in covariance matrix estimator

A modified Markowitz frameworkConclusion

The Markowitz frameworkDifferent efficient frontier definitions under stochastic setting

”Plug-in” efficient frontier

For w = w(X , µ∗) and reasonable µ and Σ estimates µ(X ) andΣ(X ),

minw

wT Σw

subject towT µ ≥ µ∗

Define the ”plug-in” efficient frontier as parametrized by µ∗

C (µ∗) = (wT Σw ,wTµ)

Zehao Chen M.V. optimization when means and covariances are estimated

Introduction and reviewA high dimensional plug-in covariance matrix estimator

A modified Markowitz frameworkConclusion

The Markowitz frameworkDifferent efficient frontier definitions under stochastic setting

”Plug-in” covariance estimates

Factor model (with domain knowledge)

R = α + BF + ǫ

Σ = W1 + W2 = BΩBT + Cov(ǫ)

Zehao Chen M.V. optimization when means and covariances are estimated

Introduction and reviewA high dimensional plug-in covariance matrix estimator

A modified Markowitz frameworkConclusion

The Markowitz frameworkDifferent efficient frontier definitions under stochastic setting

”Plug-in” covariance estimates

Factor model (with domain knowledge)

R = α + BF + ǫ

Σ = W1 + W2 = BΩBT + Cov(ǫ)

Shrinkage estimator

αF + (1 − α)Σ

Zehao Chen M.V. optimization when means and covariances are estimated

Introduction and reviewA high dimensional plug-in covariance matrix estimator

A modified Markowitz frameworkConclusion

The Markowitz frameworkDifferent efficient frontier definitions under stochastic setting

Random efficient frontier?

0.04 0.06 0.08 0.1 0.12 0.14 0.160

1

2

3

4

5

6

7

8

9x 10

−3

0.04 0.06 0.08 0.1 0.12 0.14 0.160

1

2

3

4

5

6

7

8

9x 10

−3

0.04 0.06 0.08 0.1 0.12 0.14 0.160

1

2

3

4

5

6

7

8

9x 10

−3

0.04 0.06 0.08 0.1 0.12 0.14 0.163

3.5

4

4.5

5

5.5

6

6.5

7

7.5

8x 10

−3

Figure: n = 50, p = 5 i.i.d multivariate normal

Zehao Chen M.V. optimization when means and covariances are estimated

Introduction and reviewA high dimensional plug-in covariance matrix estimator

A modified Markowitz frameworkConclusion

The Markowitz frameworkDifferent efficient frontier definitions under stochastic setting

Random efficient frontier?

The data dependent efficient frontier is not well defined, andis a random curve.

Zehao Chen M.V. optimization when means and covariances are estimated

Introduction and reviewA high dimensional plug-in covariance matrix estimator

A modified Markowitz frameworkConclusion

The Markowitz frameworkDifferent efficient frontier definitions under stochastic setting

Random efficient frontier?

The data dependent efficient frontier is not well defined, andis a random curve.

By plugging the estimates of mean µ, it’s essentiallyconstraining on

wT µ ≥ µ∗

Zehao Chen M.V. optimization when means and covariances are estimated

Introduction and reviewA high dimensional plug-in covariance matrix estimator

A modified Markowitz frameworkConclusion

The Markowitz frameworkDifferent efficient frontier definitions under stochastic setting

Random efficient frontier?

The data dependent efficient frontier is not well defined, andis a random curve.

By plugging the estimates of mean µ, it’s essentiallyconstraining on

wT µ ≥ µ∗

Conceptually, one should constrain on

E(wT r) ≥ µ∗

Zehao Chen M.V. optimization when means and covariances are estimated

Introduction and reviewA high dimensional plug-in covariance matrix estimator

A modified Markowitz frameworkConclusion

The Markowitz frameworkDifferent efficient frontier definitions under stochastic setting

Resampled efficient frontier

For bootstrapped samples X ∗

1 ,...,X ∗

B of X ,

wM(X , µ∗) =1

B

B

Σi=1

w(X ∗

i , µ∗)

Define the resampled efficient frontier as

CM(µ∗) = (wTMΣwM ,wT

Mµ)

Zehao Chen M.V. optimization when means and covariances are estimated

Introduction and reviewA high dimensional plug-in covariance matrix estimator

A modified Markowitz frameworkConclusion

The Markowitz frameworkDifferent efficient frontier definitions under stochastic setting

Previous definitions of efficient frontiers

”Plug-in” efficient frontier

minw

wT Σw , wT µ ≥ µ∗

C (µ∗) = (wT Σw ,wTµ)

Resampled efficient frontier

wM(X , µ∗) =1

B

B

Σi=1

w(X ∗

i , µ∗)

CM(µ∗) = (wTMΣwM ,wT

Mµ)

Zehao Chen M.V. optimization when means and covariances are estimated

Introduction and reviewA high dimensional plug-in covariance matrix estimator

A modified Markowitz frameworkConclusion

The L2 boosting estimatorSimulation and empirical study

If one really wants to use the plug-in

Estimate Σ or Σ−1?

Zehao Chen M.V. optimization when means and covariances are estimated

Introduction and reviewA high dimensional plug-in covariance matrix estimator

A modified Markowitz frameworkConclusion

The L2 boosting estimatorSimulation and empirical study

If one really wants to use the plug-in

Estimate Σ or Σ−1?

Employ a sparsity assumption (in practice, residuals from somefactor model): reduce the number of parameters to estimate.

Zehao Chen M.V. optimization when means and covariances are estimated

Introduction and reviewA high dimensional plug-in covariance matrix estimator

A modified Markowitz frameworkConclusion

The L2 boosting estimatorSimulation and empirical study

If one really wants to use the plug-in

Estimate Σ or Σ−1?

Employ a sparsity assumption (in practice, residuals from somefactor model): reduce the number of parameters to estimate.

Impose proper weight constraints.

Zehao Chen M.V. optimization when means and covariances are estimated

Introduction and reviewA high dimensional plug-in covariance matrix estimator

A modified Markowitz frameworkConclusion

The L2 boosting estimatorSimulation and empirical study

A high dimensional covariance estimator

Use Modified Cholesky decomposition Σ−1 = T ′D−1T toreparametrize covariance matrix, and enforce sparsity on T .

Zehao Chen M.V. optimization when means and covariances are estimated

Introduction and reviewA high dimensional plug-in covariance matrix estimator

A modified Markowitz frameworkConclusion

The L2 boosting estimatorSimulation and empirical study

A high dimensional covariance estimator

Use Modified Cholesky decomposition Σ−1 = T ′D−1T toreparametrize covariance matrix, and enforce sparsity on T .

Use coordinate-wise greedy (component-wise L2 boosting)with a modified BIC stopping criterion. Can be shown to be aconsistent estimator.

Zehao Chen M.V. optimization when means and covariances are estimated

Introduction and reviewA high dimensional plug-in covariance matrix estimator

A modified Markowitz frameworkConclusion

The L2 boosting estimatorSimulation and empirical study

A high dimensional covariance estimator

Use Modified Cholesky decomposition Σ−1 = T ′D−1T toreparametrize covariance matrix, and enforce sparsity on T .

Use coordinate-wise greedy (component-wise L2 boosting)with a modified BIC stopping criterion. Can be shown to be aconsistent estimator.

Spectral norm convergence for both Σ and Σ−1.

Zehao Chen M.V. optimization when means and covariances are estimated

Introduction and reviewA high dimensional plug-in covariance matrix estimator

A modified Markowitz frameworkConclusion

The L2 boosting estimatorSimulation and empirical study

Modified Cholesky decomposition

For a random vector Y = (y1, y2, ..., yp), one can write them in the”auto-regressive” form, for k ≥ 2

yk = µk + φk,1y1 + φk,2y2 + ... + φk,k−1yk−1 + ǫk

Let T be a p × p unit lower triangular matrix, with −φi ,j (i < j) onthe lower triangular. And let D be the p × p diagonal matrix

D = diag(var(y1), var(ǫ2), var(ǫ3), ..., var(ǫp))

The Modified Cholesky decomposition is

Σ−1 = T ′D−1T

Zehao Chen M.V. optimization when means and covariances are estimated

Introduction and reviewA high dimensional plug-in covariance matrix estimator

A modified Markowitz frameworkConclusion

The L2 boosting estimatorSimulation and empirical study

The coordinate-wise greedy algorithm

The main ideas is to iteratively look for the predictor which ismostly correlated with the current model residual. Include thispredictor and re-calculate the residual.

Zehao Chen M.V. optimization when means and covariances are estimated

Introduction and reviewA high dimensional plug-in covariance matrix estimator

A modified Markowitz frameworkConclusion

The L2 boosting estimatorSimulation and empirical study

The coordinate-wise greedy algorithm

The main ideas is to iteratively look for the predictor which ismostly correlated with the current model residual. Include thispredictor and re-calculate the residual.

This process will generate a sequence of nested modelsM1 ⊆ M2 ⊆ ... ⊆ Mmn .

Zehao Chen M.V. optimization when means and covariances are estimated

Introduction and reviewA high dimensional plug-in covariance matrix estimator

A modified Markowitz frameworkConclusion

The L2 boosting estimatorSimulation and empirical study

The coordinate-wise greedy algorithm

The main ideas is to iteratively look for the predictor which ismostly correlated with the current model residual. Include thispredictor and re-calculate the residual.

This process will generate a sequence of nested modelsM1 ⊆ M2 ⊆ ... ⊆ Mmn .

BIC: log σ2 + log nn

· #Mk

Zehao Chen M.V. optimization when means and covariances are estimated

Introduction and reviewA high dimensional plug-in covariance matrix estimator

A modified Markowitz frameworkConclusion

The L2 boosting estimatorSimulation and empirical study

The coordinate-wise greedy algorithm

The main ideas is to iteratively look for the predictor which ismostly correlated with the current model residual. Include thispredictor and re-calculate the residual.

This process will generate a sequence of nested modelsM1 ⊆ M2 ⊆ ... ⊆ Mmn .

BIC: log σ2 + log nn

· #Mk

Proposed Modified BIC: nσ2 + σ2 log p log n · #Mk

Zehao Chen M.V. optimization when means and covariances are estimated

Introduction and reviewA high dimensional plug-in covariance matrix estimator

A modified Markowitz frameworkConclusion

The L2 boosting estimatorSimulation and empirical study

Convergence results

Definition of spectral norm:

||A||2 =√

λmax(A′A)

Convergence:||Σ − Σ||2 → 0

||Σ−1 − Σ−1||2 → 0

Zehao Chen M.V. optimization when means and covariances are estimated

Introduction and reviewA high dimensional plug-in covariance matrix estimator

A modified Markowitz frameworkConclusion

The L2 boosting estimatorSimulation and empirical study

An example

Setting: n = 300, p = 600, number of nonzero parameters in T is900 (i.e. 3 per row), and T elements uniformly distributed in[0.5, 1].

Zehao Chen M.V. optimization when means and covariances are estimated

Introduction and reviewA high dimensional plug-in covariance matrix estimator

A modified Markowitz frameworkConclusion

The L2 boosting estimatorSimulation and empirical study

MSE to true inverse covariance

50 100 150 200 250 3000

10

20

30

QE

1

50 100 150 200 250 3000

500

1000

1500

2000

p (p = 2n)

QE

1

BoostingPLM

Shrinkage

Zehao Chen M.V. optimization when means and covariances are estimated

Introduction and reviewA high dimensional plug-in covariance matrix estimator

A modified Markowitz frameworkConclusion

The L2 boosting estimatorSimulation and empirical study

MSE to true covariance

50 100 150 200 250 3000

2

4

6

8Q

E2

/ 100

0

BoostingShrinkage

50 100 150 200 250 3000

200

400

600

p (p = 2n)

QE

2 / 1

000

PLM

Around 8

Zehao Chen M.V. optimization when means and covariances are estimated

Introduction and reviewA high dimensional plug-in covariance matrix estimator

A modified Markowitz frameworkConclusion

The L2 boosting estimatorSimulation and empirical study

An empirical example: NASDAQ-100, 1990-2006

0.015 0.02 0.025 0.03 0.035 0.04 0.0450.002

0.004

0.006

0.008

0.01

0.012

0.014

0.016

0.018

0.02

BoostingShrinkageSample

Figure: Empirical performance curve with S&P500 index as market factor

Zehao Chen M.V. optimization when means and covariances are estimated

Introduction and reviewA high dimensional plug-in covariance matrix estimator

A modified Markowitz frameworkConclusion

The frameworkAn example and simulation

Efficient frontier when means and covariances are unknown

For w(X , µ∗), solveminw

var(wT r)

E(wT r) ≥ µ∗

Define the expected efficient frontier as

C (µ∗) = EX(wT Σw ,wTµ)

Zehao Chen M.V. optimization when means and covariances are estimated

Introduction and reviewA high dimensional plug-in covariance matrix estimator

A modified Markowitz frameworkConclusion

The frameworkAn example and simulation

A comparison of efficient frontiers

”Plug-in” efficient frontier

minw

wT Σw , wT µ ≥ µ∗, C (µ∗) = (wTΣw ,wT µ)

Resampled efficient frontier

wM(X , µ∗) =1

B

B

Σi=1

w(X ∗

i , µ∗), CM(µ∗) = (wTMΣwM ,wT

Mµ)

Expected efficient frontier

minw

var(wT r), E(wT r) ≥ µ∗, C (µ∗) = EX(wT Σw ,wTµ)

Zehao Chen M.V. optimization when means and covariances are estimated

Introduction and reviewA high dimensional plug-in covariance matrix estimator

A modified Markowitz frameworkConclusion

The frameworkAn example and simulation

Solving for the expected efficient frontier

Introduce a risk-aversion parameter λ and reformulate the problemas

minw

−E(wT r) + λvar(wT r)

Zehao Chen M.V. optimization when means and covariances are estimated

Introduction and reviewA high dimensional plug-in covariance matrix estimator

A modified Markowitz frameworkConclusion

The frameworkAn example and simulation

A not-so-bayesian stochastic control approach

Solve for a portfolio weight w(X , λ)

−E(wT r) + λvar(wT r) = −EwT r + λE(wT r)2 − λE2(wT r)

Zehao Chen M.V. optimization when means and covariances are estimated

Introduction and reviewA high dimensional plug-in covariance matrix estimator

A modified Markowitz frameworkConclusion

The frameworkAn example and simulation

A not-so-bayesian stochastic control approach

Solve for a portfolio weight w(X , λ)

−E(wT r) + λvar(wT r) = −EwT r + λE(wT r)2 − λE2(wT r)

However, one needs the following conditional form,

EX [−wTE(r |X ) + λwTE(rrT |X )w − λwTE(??|X )w ]

Zehao Chen M.V. optimization when means and covariances are estimated

Introduction and reviewA high dimensional plug-in covariance matrix estimator

A modified Markowitz frameworkConclusion

The frameworkAn example and simulation

A not-so-bayesian stochastic control approach

This conditional form entitles us to plug in reasonable guessesfor E (r |X ),E (rrT |X ) and E (??|X ).

Zehao Chen M.V. optimization when means and covariances are estimated

Introduction and reviewA high dimensional plug-in covariance matrix estimator

A modified Markowitz frameworkConclusion

The frameworkAn example and simulation

A not-so-bayesian stochastic control approach

This conditional form entitles us to plug in reasonable guessesfor E (r |X ),E (rrT |X ) and E (??|X ).

Use bayesian way to develop a procedure and approximate itin frequentist setting.

Zehao Chen M.V. optimization when means and covariances are estimated

Introduction and reviewA high dimensional plug-in covariance matrix estimator

A modified Markowitz frameworkConclusion

The frameworkAn example and simulation

Embedding technique

Embedding technique1:This bayesian formulation can be shown to be equivalent to

EX [−ηwT E(r |X ) + λwTE(rrT |X )w ]

where η = 1 + 2λE((w∗

λ)T r) and w∗

λ is the solution to the originaloptimization problem with risk aversion λ.

1X. Y. Zhou and D. Li (2000)Zehao Chen M.V. optimization when means and covariances are estimated

Introduction and reviewA high dimensional plug-in covariance matrix estimator

A modified Markowitz frameworkConclusion

The frameworkAn example and simulation

How to proceed? A simple example

Take prior Σ ∼ IW (Φ, v), so

E(Σ|X ) = αΦ + (1 − α)Σ

E(r |X ) = E(µ|X ) ≈ X

E(rrT |X ) = E(Σ|X ) + var(µ|X ,Σ) + E(r |X )E(r |X )T

≈n + 1

n[αΦ + (1 − α)Σ] + X XT

”Semi-Empirical Bayes”: estimate Φ by diag(Σ)

Zehao Chen M.V. optimization when means and covariances are estimated

Introduction and reviewA high dimensional plug-in covariance matrix estimator

A modified Markowitz frameworkConclusion

The frameworkAn example and simulation

How to proceed? A simple example

Define a parametrized function w(X , λ, α, η) that solves

−ηwT E(r |X ) + λwTE(rrT |X )w

with E(r |X ) and E(rrT |X ) replaced by approximations.

Zehao Chen M.V. optimization when means and covariances are estimated

Introduction and reviewA high dimensional plug-in covariance matrix estimator

A modified Markowitz frameworkConclusion

The frameworkAn example and simulation

How to proceed? A simple example

Having a parametrized procedure w(X , λ, α, η), we wish toevaluate

R(λ, α, η) = −EwT r + λvar(wT r)

=EX [−E(wT r |X ) + λvar(wT r |X )] + λvar[E(wT r |X )]

Zehao Chen M.V. optimization when means and covariances are estimated

Introduction and reviewA high dimensional plug-in covariance matrix estimator

A modified Markowitz frameworkConclusion

The frameworkAn example and simulation

How to proceed? A simple example

Having a parametrized procedure w(X , λ, α, η), we wish toevaluate

R(λ, α, η) = −EwT r + λvar(wT r)

=EX [−E(wT r |X ) + λvar(wT r |X )] + λvar[E(wT r |X )]

Bootstrap

Bootstrap: X ∗

1 ,X ∗

2 , ...,X ∗

B

Zehao Chen M.V. optimization when means and covariances are estimated

Introduction and reviewA high dimensional plug-in covariance matrix estimator

A modified Markowitz frameworkConclusion

The frameworkAn example and simulation

How to proceed? A simple example

The above can be approximated by the bootstrapped empirical risk,

R(λ, α, η) = EX [−E(wT r |X )] + λEX [var(wT r |X )] + λvar[E(wT r |X )]

EX [−E(wT r |X )] =1

B

B

Σi=1

[−wT (X ∗

i , λ, α, η)µ(X ∗

i )]

EX [var(wT r |X )] =1

B

B

Σi=1

[wT (X ∗

i , λ, α, η)Σ(X ∗

i )w(X ∗

i , λ, α, η)]

var[E(wT r |X )] = var[wT (X ∗

i , λ, α, η)µ(X ∗

i )]

where µ(X ∗

i ) and Σ(X ∗

i ) are reasonable estimates for mean andcovariance matrix of X ∗

i .

Zehao Chen M.V. optimization when means and covariances are estimated

Introduction and reviewA high dimensional plug-in covariance matrix estimator

A modified Markowitz frameworkConclusion

The frameworkAn example and simulation

How to proceed? A simple example

Solve forminα,η

R(λ, α, η)

with reasonable initial values.

Zehao Chen M.V. optimization when means and covariances are estimated

Introduction and reviewA high dimensional plug-in covariance matrix estimator

A modified Markowitz frameworkConclusion

The frameworkAn example and simulation

Example: n = 50, p = 5

0.03 0.035 0.04 0.045 0.05 0.055 0.06 0.065 0.07 0.0752.5

3

3.5

4

4.5

5

5.5x 10

−3

Std.

Ret

urn

SBFShrinkageResamplingSampleTrue

Zehao Chen M.V. optimization when means and covariances are estimated

Introduction and reviewA high dimensional plug-in covariance matrix estimator

A modified Markowitz frameworkConclusion

The frameworkAn example and simulation

Example: n = 50, p = 30

0.03 0.035 0.04 0.045 0.05 0.055 0.064.5

5

5.5

6

6.5

7

7.5

8

8.5

9x 10

−3

Std.

Ret

urn

SBFShrinkageResamplingSample

Zehao Chen M.V. optimization when means and covariances are estimated

Introduction and reviewA high dimensional plug-in covariance matrix estimator

A modified Markowitz frameworkConclusion

The frameworkAn example and simulation

Example: n = 50, p = 50

4 4.5 5 5.5 6 6.5 7

x 10−3

0.01

0.011

0.012

0.013

0.014

0.015

0.016

Std.

Ret

urn

SBFShrinkageHCD

Zehao Chen M.V. optimization when means and covariances are estimated

Introduction and reviewA high dimensional plug-in covariance matrix estimator

A modified Markowitz frameworkConclusion

Conclusion

The proposed high dimensional covariance estimator isappropriate in sparse setting, and has better performance.

Zehao Chen M.V. optimization when means and covariances are estimated

Introduction and reviewA high dimensional plug-in covariance matrix estimator

A modified Markowitz frameworkConclusion

Conclusion

The proposed high dimensional covariance estimator isappropriate in sparse setting, and has better performance.

The modified Markowitz framework tries to optimize on theexpected efficient frontier. It’s a general framework and hasbetter performance in many practical scenarios.

Zehao Chen M.V. optimization when means and covariances are estimated