© September 21, 2020, Christopher D. Carroll Equiprobable-Returns

Equiprobable Approximation to Bivariate Lognormal Returns

This handout starts by defining a convenient notation to represent two rate-of-return shocks that are distributed according to a multivariate lognormal that allows for nonzero covariances. It next presents a computationally simple method for constructing a numerical approximation to that joint distribution. The final section uses both the numerical approximation, and more accurate (but enormously slower) standard numerical integration tools to assess the accuracy of the Campbell-Vicera analytical approximation to the solution to the optimal portfolio choice problem described in Portfolio-Multi-CRRA.

1 Statistical Theory

Consider a set of two normally distributed risks

                     2   2
𝜃2,t+1 ∼  𝒩  (− 0.5Σ  , Σ  )
𝜃      ∼  𝒩  (− 0.5(x Σ )2,(x Σ )2)
  1,t+1

which are statistically independent (𝜃1,t+1  ⊥ 𝜃2,t+1   ) even though the scale of shock 1’s standard deviation is determined by the proportionality factor x  multiplied by the value of shock 2’s standard deviation, Σ  . (This permits us to increase or decrease the size of both risks by changing Σ  and to adjust the relative size of risk 1 vs risk 2 by adjusting x  .)

The 𝜃  are interpreted as the logarithms of level variables Θ  , and the means of the log variables have been chosen such that the expectations of the levels are independent of the size of the risk (cf. MathFacts). That is, defining Θ   =  𝔼 [e𝜃i,t+1]
  i      t  for i ∈ {1, 2 } :

Θ   ≡  𝔼 [Θ      ] = 1 ∀ i
  i     t   i,t+1
   log  𝔼t[Θi,t+1 ] = 0 ∀ i.
(1)

From the first shock we can construct a log rate-of-return variable that can be represented equivalently in either of two ways:

                ≡𝔯𝔯𝔯
          ◜-----◞◟1-----◝
r      =  r +  0.5 (xΣ )2 + 𝜃
 1,t+1      1                 1,t+1
       = 𝔯𝔯𝔯1 − 0.5(x Σ )2 + 0.5(x Σ )2 + 𝜃1,t+1
          ◟-----◝◜ -----◞   ◟-------◝ ◜-------◞
                =r1                ≡/𝜃1,t+1
(2)

where /𝜃      ∼   𝒩  (0.,(x Σ )2)
/1,t+1  ; the notation for /𝜃
/  is motivated by the fact that the addition of the extra term cancels the nonzero mean of the original 𝜃  . Then

            R      ≡  er1,t+1
              1,t+1
R    ≡ 𝔼  [R      ] = er1+ (x Σ)2∕2
   1     t   1,t+1
    log 𝔼t [R1,t+1 ] = r1 +  (xΣ )2∕2  =  𝔯𝔯𝔯1
(3)

where note to avoid confusion that 𝔯𝔯𝔯1 ⁄=  𝔼t [r1,t+1]  while log  𝔼t[R1,t+1 ] ⁄= r1 =  𝔼t [r1,t+1]  .

Using 𝔼t [𝜃2,t+1 ] = − 0.5Σ2   and 𝔼t [(ω ∕x )𝜃1,t+1 ] = − 0.5( (ω∕x )Σ )2   , we can by analogy define a second return

r2,t+1 ≡  𝔯𝔯𝔯2 + ζ +  (ω ∕x )𝜃1,t+1 +  𝜃2,t+1
                                              2                    2
       =  𝔯𝔯𝔯2 + ζ +  (ω ∕x )(//𝜃1,t+1 −  0.5(x Σ ) ) + (/𝜃/2,t+1 −  0.5Σ  )
       =  𝔯𝔯𝔯  + ζ −  0.5(ω ∕x )(x Σ )2 − 0.5Σ2  + (ω ∕x )//𝜃     +  //𝜃
          ◟2---------------◝◜----------------◞           t+1,1    2,t+1
                           r2
(4)

for some constants ω  and ζ  .

Since (ω ∕x )𝜃1,t+1   is the only component of r2,t+1   that is correlated with r
  1,t+1   ,

cov (r     ,r     ) =  cov (r     ,(ω ∕x )r     )
      1,t+1   2,t+1           1,t+1          1,t+1
                    =  (ω ∕x )cov (r1,t+1, r1,t+1)
                              ◟-------◝ ◜-------◞
                                     =x2Σ2
                            2
                    =  ωx Σ  .

Thus, the parameter ω  controls the covariance between the risky returns. If we set ω  =  0  then r1,t+1 ⊥  r2,t+1   (the returns are independent).

Next we want to find the value of ζ  such that the expected level of the return is unaffected by Σ  (so that we will be able to explore independently the distinct effects of the components of each shock and their covariance):

R2  ≡  𝔼t [R2,t+1 ] = e𝔯𝔯𝔯2
(5)

regardless of the values of x  and Σ  . Using (4), we therefore need:

         ζ+(ω∕x)𝜃1,t+1+ 𝜃2,t+1
    𝔼t [e                   ] = 1
log 𝔼  [e ζ+(ω∕x)𝜃1,t+1+ 𝜃2,t+1] = 0.
      t
(6)

Using standard facts about lognormals (cf. MathFacts), and for convenience defining ˆω  = ( Σ∕x )ω  , we have

0.=  ζ −  0.5 ˆωx2  − 0.5 Σ2 +  0.5ωˆ2x2  +  0.5 Σ2
               2
   = ζ −  0.5x  ˆω (1 −  ˆω)
                 2  2                 2   2
ζ  = 0.5 (ˆω −  ˆω  )x  =  0.5 (ωx −  ω  )Σ
(7)

which means that we can rewrite (4) directly as

r      = 𝔯𝔯𝔯 +  0.5(ωx  −  ω2 )Σ2  + (ω ∕x )𝜃      + 𝜃
 2,t+1      2                                1,t+1     2,t+1
       = 𝔯𝔯𝔯2 + 0.5(ωx  −  ω2 )Σ2  − 0.5 (ω ∕x)(x Σ )2 − 0.5 Σ2 +  (ω ∕x )//𝜃t+1,1 + //𝜃2,t+1
                        2            2        2  2         2
       = 𝔯𝔯𝔯2 + 0.5ωx  Σ  −  0.5ωx  Σ  −  0.5 ω Σ   −  0.5Σ   + (ω ∕x )//𝜃t+1,1 + 𝜃//2,t+1
                      2       2
       = 𝔯𝔯𝔯◟2 −-(1-+◝◜ω--)0.5Σ-◞+  (ω∕x )/𝜃/t+1,1 +  //𝜃2,t+1.
                   r
                    2

Hence, from the independent mean-one lognormally distributed shocks Θ1   and Θ2   , we have constructed a pair of jointly lognormally distributed shocks whose covariance is controlled by the parameter ω  , whose relative and absolute variances are controlled by the parameters x  and Σ  , and whose means are r1   and r2   .

To sum up, the process can be described in either of two ways:

(         )     (                )     (                                 )
   r               𝒩 (r  ,x2Σ2 )            𝒩 (𝔯𝔯𝔯  −  0.5(x Σ )2,x2 Σ2 )
    1,t+1     ∼          1     2     =           1          2      2   2
   r2,t+1           𝒩 (r2,   Σ  )          𝒩  (𝔯𝔯𝔯2 − (1 +  ω  )0.5Σ  ,Σ  )

with covariance matrix

(  σ2   σ    )    (  (x Σ )2     x ωΣ2     )
     1    12    ≡
   σ12   σ22           xω Σ2   (ω2 +  1)Σ2
(8)

where a final useful result that follows from (8) and (1) is

corr (r1,t+1, r2,t+1 ) ≡  cov(r1,t+1, r2,t+1 )∕σ1 σ2
                                  2 0.5
                     = ω ∕ (1 + ω  )   .
(9)

2 Computational Theory

To reduce clutter, define ˜Ri  =  Ri,t+1   and interpret 𝔼  as 𝔼t []  , so that we can write the expectation of some function h  that depends on the realization of the return shocks as:

                   ∫  ¯R1 ∫ R¯2
𝔼 [h( ˜R  ,R˜ )] =             h (˜R  ,R˜  )dF (R˜ , ˜R  )
        1   2                      1    2       1    2
                    R1    R2
(10)

where F (˜R1, R˜2  )  is the cumulative distribution function for a multivariate lognormal with the covariance matrix defined by (8).1 Standard numerical computation software can compute this double integral, but at such a slow speed as to be unusable for many purposes. Computation of the expectation can be massively speeded up by advance construction of a numerical approximation to F (˜R  ,R˜  )
     1   2  .

Such approximations often take the approach of replacing the distribution function with a discretized approximation to it; appropriate weights w [i,j ]  are attached to each of a finite set of points indexed by i  and j  , and the approximation to the integral is given by:

                  ∑ n  ∑m
      ˜   ˜                   ˆ         ˆ
𝔼 [h(R1,  R2 )] ≈          h (R1 [i,j],R2  [i,j ])w  [i,j ]
                   i=1 j=1
(11)

where various methods are used for constructing the weights w [i,j]  and the nodes (corresponding to the {i,j } pairs). The matrices Rˆ1   and Rˆ2   contain the conditional mean values of R˜
  1   and R˜
  2   associated with each of the regions.

Perhaps the most popular such method is Gauss-Hermite interpolation (see Judd (1998) for an exposition, or Kopecky and Suen (2010) for a recent candidate for a better choice). Here, we will pursue a particularly simple and intuitive alternative: Equiprobable discretization. In this method, m  =  n  and boundaries on the joint CDF are determined in such a way as to divide up the total probability mass into submasses of equal size (each of which therefore has a mass of n− 2   ). This is conceptually easier if we represent the underlying shocks as statistically independent, as with Θ1   and Θ2   above; in that case, each submass is a square region in the Θ
  1   and Θ
  2   grid. We then compute the average value of Θ1   and Θ2   conditional on their being located in each of the subdivisions of the range of the CDF. Since Θ
  1   and Θ
  2   are IID, the representation of the approximating summation is even simpler than (11):

  [  (          ) ]        ∑ n  ∑n    (                                       )
𝔼  h   ˜R  , ˜R  )    ≈  n− 2         h   R  (ˆΘ  [i], ˆΘ  [j]),R  (Θˆ [i], ˆΘ [j]))
         1    2                           1   1      2       2   1      2
                            i=1  j=1

where Ri (Θ1, Θ2 )  are the (linear) functions relating the return shocks to the IID shocks.

3 Computational Results

PIC

Figure 1:Portfolio Share By Method as a Function of ω

Figure 1 compares the computed optimal portfolio share for a numerical solution using the built-in numerical optimizer and maximization functions (the lowest, black, locus), the Campbell-Viceira solution (the highest, red locus) and an equiprobable approximation using 20 approximation points (green, middle) as well as the solution using the equiprobable approximation at an evenly-spaced grid of points (blue dots).

Careful examination indicates that the numerical approximation is quite close to the full numerical solution, while the CV approximation diverges substantially from the numerical answer. The tradeoff is that the equiprobable solution is about 2000 times slower than the CV approximation, while the direct solution is more than 100 times slower than the equiprobable solution. Depending on the requirements of the problem being examined, these differences in efficiency can make a tremendous difference in the feasibility of a research project.2

References

   Judd, Kenneth L. (1998): Numerical Methods in Economics. The MIT Press, Cambridge, Massachusetts.

   Kopecky, Karen A., and Richard M.H. Suen (2010): “Finite State Markov-Chain Approximations To Highly Persistent Processes,” Review of Economic Dynamics, 13(3), 701–714, http://www.karenkopecky.net/RouwenhorstPaper.pdf.