Sunteți pe pagina 1din 14

ECONOMICS

The Equity Premium Puzzle: A Model for


Its Behavior
Timothy P. Lavelle

I n the United States, there exist independent markets for risk free government bond securities
and risky equity securities. A rational/sophisticated investor will build a portfolio out of these
two components in such a way as to optimize his portfolio in terms of the ratio of riskiness (or,
more specifically, idiosyncratic variability) and expected return (Black and Litterman 1992).
Once this equilibrium condition is met (an investor-specific optimal allocation between risky and
risk free securities) the rules of arbitrage pricing should keep the return spread between these two
securities at an equilibrium level (Brennan, Schwartz and Lagnado 1997). From this notion of an
arbitrage pricing model, one can look at historical returns in the United States and appreciate that
the aggregate return spread between the market index for equity securities and government debt
instruments should be characterized by a function of expected returns, risk, and investor
preference for risk. This premium has been roughly 6% over the long run (Kocherlakota 1996).
Given the observed level of risk for each asset class, one can then derive the implied risk
preference profile of the investing community in aggregate. This stream of analysis has been
applied to what has come to be known as the equity premium puzzle. While the fact that a
premium exists is not a puzzle to even the most naïve of market observers, it is the magnitude of
this premium that is interesting and perplexing. No theory to date has been able to explain it
fully, and thus, the focus of this article will be to demystify its enigmatic behavior using a
dynamical systems approach.
As stated previously, the premium return of equity securities over government debt
securities is inevitable due to the greater systematic and idiosyncratic risk of equity returns in the
market. The problem, however, is determining whether this premium is appropriate given the
level of risk and investors’ risk aversion, or if it is, in fact, extreme. Siegel and Thaler use a
utility preference model to describe the level of risk aversion implied by the long run 6% equity
premium (Siegel and Thaler 1997). The following anecdote demonstrates this level of risk
aversion. Given a fair bet with 50% probability of winning and exactly doubling one’s entire
wealth and a 50% probability of losing and exactly halving one’s entire wealth, an individual
with the implied level of risk aversion derived from the equity premium would be willing to pay
49% of his or her entire wealth to avoid the gamble (Siegel and Thaler 1997). Thus, if one has
wealth of one million dollars and is given a 50% chance to either be worth two million dollars or
500 thousand dollars, that individual would rather pay 490 thousand dollars to avoid this bet.
This person would rather sacrifice 490 thousand dollars with 100% certainty than have a 50%
chance of winning more money and a 50% chance of losing only ten thousand dollars more.
While there is no way to prove mathematically that this implies the equity premium is
implausible and ludicrous, one can see from this demonstration that it is self-evident. It would be
difficult, if not impossible, to find an individual who would behave under these conditions in a
similar manner. Thus, if it is hard to find one person, it is certainly impossible to assume that the
average market participant would behave such as this.
From this brief (and relatively rough) anecdote, one can innately see that the equity risk
premium is too large to be described using a simple rational investor model. In fact, Mehra and
Prescott resolve that the only way to explain or make sense of the equity premium is if the
typical investor is implausibly, and thus unnaturally, averse to risk (Mehra and Prescott 1985).
As a result, many economists have attempted to explain and describe the causes of this premium,
but as of yet, none have fully succeeded (Kocherlakota 1996). Accordingly, the question of this
study is not what accounts for the equity risk premium, or even why it is so large. Rather, this
study will attempt to describe the dynamic nature of its level through time.
Ever since the advent of Modern Portfolio Theory, it has been widely accepted that the
best way to examine an investment in a financial security is not to analyze its historical returns
and deviations, but rather to make estimates of the future behavior of these securities (Fabozzi,
Gupta, and Markowitz 2002). That being said, often the best proxy for the future is the
performance of the past. Many of the attempts at describing the equity premium puzzle have
concentrated on predicting the future level of the premium based on forecasted future levels of
consumption, return, and risk aversion (Kocherlakota 1996). In contrast, this study will attempt
to explain the oscillations of the equity premium as an entity in and of itself rather than as a
spread between two unrelated (or related) securities. The central hypothesis of this study is that
the equity premium has a mean, which is non-stationary as a result of market shocks, and
behaves in a constant, predictable manner around that mean after accounting for noise.

Method

Subjects. The subjects used in this analysis are the New York Stock Exchange (NYSE)
aggregate index and the 90-day U.S. Treasury bill. The NYSE aggregate index was selected as a
broad market index to track the general movement of the U.S. security prices for an extended
period of time. This index has been available for over 100 years and has always represented a
wide range of diversified industries present and operating within the United States. Unlike some
other market indices, it has exhibited somewhat more damped market cyclicality as a result of its
broad industry base. The constituents of the index include all equity securities actively traded on
the New York Stock Exchange as of the date of each data measurement. Therefore, a
survivorship bias is not introduced and thus the index is constantly evolving. The value-weighted
version of this market index was chosen so that securities that are larger on the market are thus
larger within the market index and the opposite for small securities.
The 90-day U.S. Treasury bill was selected as the representative risk-free rate for several
reasons. Since it is a U.S. Treasury debt instrument, it is backed by the full faith and trust of the
United States Government. In the financial community, this is seen as the equivalent of default
risk-free. Given that this is a fixed maturity, zero coupon 90-day bill, inflation risk is virtually
nonexistent as well. For the purposes of this study, it is assumed that U.S. inflation was relatively
constant for any 90-day period over the last 100 years. For the few time periods that may have
exhibited more turbulent inflation, it is assumed that these effects are damped in the analysis by a
corresponding effect on the equity index levels observed. While using a shorter maturity bill
would reduce the inflation risk just slightly more, the more volatile nature of these securities
outweighs the risk-reduction benefit and may in fact add unwanted noise to the analysis.
Apparatus. The apparatus used to collect the price information for the study’s time period
(1934 to 2004) for each security is the Center for Research in Security Prices (CRSP). The CRSP
is operated and maintained by the University of Chicago Graduate School of Business and is
widely considered the industry standard in historical security prices. The data was downloaded
on October 2, 2005 and should be considered current and accurate as of the most recent CRSP
updates to that point. The data will be analyzed using R statistical software in an effort to isolate
the non-stationary mean and fit the resulting trends to an appropriate model.
Procedure. For each security, the monthly nominal price levels from 1934 to 2004 were
used to calculate monthly nominal returns. Monthly data points were used as they provide many
data points while also not introducing an undue level of noise to the analysis from weekly or
daily returns. The period 1934 to 2004 is used as it is the longest period available for which both
securities have a complete data set on record in the CRSP database. Once the monthly returns are
calculated, the risk free return is subtracted from the equity return to calculate the equity risk
premium. A plot of this time series is available in appendix A.

Results

Descriptive statistics. Given that this data set represents a time series of returns, the most
appealing measure of central tendency is the geometric average return. This return (GAR) is
calculated as follows (1):

Where n is equivalent to the number of discrete periods (852 months) and r is equivalent
to the period return. Thus, for this time series, the resultant geometric average return is 0.57%
per month. In order to get an annual estimation of the equity premium, the rate can be
compounded for 12 months (2):

Consequently, the annual geometric average equity risk premium from 1934 to 2004 is
7.10% (SD = 4.49%). This is consistent with Mehra and Prescott’s calculation of the market risk
premium as 6.90% (notice the discrepancy as a result of calculations based on different years).
Given that this is a time series that incorporates regular compounding, the geometric
average return is the most accurate and appropriate measure of central tendency. Though, for the
sake of comparison, the arithmetic monthly return annualized is 8.41% and the median monthly
return annualized is 11.90%. Thus, the geometric return is a considerably more conservative
description of the central tendency of the time series than either the arithmetic average or
median.
In order to determine whether this time series phenomenon can be described as a series of
overlapping sinusoidal waves, a Fourier analysis was completed. This Fourier analysis should be
able to indicate whether there is some sort of a recurring cycle underneath the string of
seemingly white noise. In order to perform the analysis, the Fast Fourier Transform function in
the R program was used. This function has as input the time series in question (appendix A) and
outputs the full frequency power spectrum. This spectrum can be observed on the periodogram in
appendix B. Notice that very little power of the overall function can be described by a sinusoidal
wave at any one frequency. In fact, what little power that is described may be merely a function
of random number sequences within the data. Based upon this evidence, it appears that there is
no underlying recurring pattern that is being obscured by noise.
In order to further describe the time series at hand, it may be appropriate to determine the
autocorrelation of the discrete returns over time. While it may be customary to calculate the
autocorrelation by calculating the correlation between one unit of return with the next sequential
unit of return repeatedly until each period has been factored into the analysis, this may not
represent the most appropriate method. Given a time lag (tau), the previously described method
would calculate the autocorrelation with tau of 1. This tau may not be the appropriate tau as it
may pick up too much noise in the calculation, thus obscuring the true autocorrelation. On the
other hand, there is no concrete algorithm for determining the optimal tau. As such, the
autocorrelation vector calculated for this analysis displays a continuum of autocorrelations for
the time series as a whole, but for each tau from 1 to 851. Observe that with a tau of 1, the
autocorrelation can be calculated based on 851 values; however, with a tau of 851, there is only
one correlation that can be made as this time series includes 852 measurements. Thus, the lower
the tau, the more measurements can be factored into the equation, but the greater the chance for
picking up noise – and vice versa for a large tau. This autocorrelation data vector is displayed as
appendix C. The graph shows the correlations for both positive and negative taus as the
correlation calculation is commutative. Consequently, the correlations from 0 to 851 are
symmetrical but flipped on the vertical axis (at tau=0) to the correlations from 0 to -851. This is
done purely for illustrative purposes.
At tau equal to zero, the auto correlation is equal to 1.00 by definition. As soon as tau
begins to grow, the correlations fall to roughly 0.00 and then oscillate around this value until
they increase to approaching 1.00 and -1.00 as tau approaches 851. This correlation expansion is
a function of the decreasing number of observations when tau is large. At first glance, appendix
C may appear to indicate that there is essentially no autocorrelation and that the values are
categorized by pure white noise. In fact, appendix D is the same autocorrelation representation,
but it was created using uniformly distributed random values between zero and one. Notice that
appendix C and appendix D look very similar.
Though in order to be confident that the market premium data in this analysis are in fact
not autocorrelated to any power greater than mere random values, a surrogate data test is
necessary. Since autocorrelation is based on the temporal nature of the data, truly un-
autocorrelated data should not depend on the order in which they are aligned. Thus, perhaps the
small fluctuations away from zero in appendix C are simply a function of the numbers in the
time series as opposed to some true autocorrelation (though minimal) at certain levels of tau.
Thus, the surrogate data test will remove the time-dependency of this series and run the
autocorrelation again to develop a 95% confidence interval. This is done simply by randomly
assigning the values in the time series to different time positions and calculating the
autocorrelation for each level of tau and then running multiple iterations of this . Then, the 95%
1

confidence interval is determined by associating the autocorrelation for each level of tau that is
2.5% from the maximum value and 2.5% from the minimum value. This 95% confidence
interval is then plotted on top of the time-dependent autocorrelation. This new merged graph is
illustrated in appendix E. The dotted, red lines represent the bounds of the confidence interval
and the solid, black line is the actual autocorrelation. Notice that at tau equal to 348, we can
reject the null hypothesis that the autocorrelation is equal to zero.
For tau equal to 348, an embedded state space plot was derived. This embedded state
space plots the value of the return along the x-axis and the value of the return tau months later
along the y-axis (appendix F). In order to improve visibility of the clusters of data, all of the data
are randomly jittered a small amount. Thus, this state space plot can illustrate the most probable
future outcomes in tau periods given a current measurement based on the empirical results.
Notice that for this level of tau, the data appear to make a pennant (as denoted by the two red
lines) with a band of outliers both before and after the pennant. The largest clusters of data are at
(0.00, 0.04), (0.00, 0.00), and (0.08, 0.01). Interestingly, there appear to be few clusters
considerably below 0.00 on the y-axis.
Given the significance level at tau equal to 348, the geometric average equity premium
can be recalculated with this in mind. A long-term calculated equity premium given historical
data is only relevant for the current period in time. Thus, the 75-year long-term average equity
risk premium in 1985 may not be equal to the 75-year long-term average equity risk premium in
any other year. Thus, the statistical program R can be used to calculate the moving equity risk
premium given a certain lag period. The autocorrelation analysis above suggests that 348 months
is the only lag period that is statistically significant. The moving equity risk premium average for
this level of tau is illustrated in appendix G. For the sake of comparison, the moving averages for
both 10 and 20 years are described in appendixes H and I, respectively . Notice that for each time
2

period, the current long-term average equity risk premium is roughly 6.00% to 7.00%.
Inferential statistics. Based upon the central hypothesis of this study – that the U.S.
equity risk premium has a non-stationary mean around which the premium fluctuates – a damped
linear oscillator model seems the most appropriate method by which to attempt to fit the data. In
a damped linear oscillator model, a system is assumed to be regulated by a certain set of
parameters that dictate both the frequency of oscillation and the rate of damping. The greater the
level of damping, the more quickly the function reaches its equilibrium level. The greater the
frequency, the more times the function crosses its stable equilibrium level in a given time period.
The central concept is that forces act to bring the system to its equilibrium (or perhaps, forces act
to move the system away from its equilibrium). In the case of the equity risk premium, the
market forces of assumed symmetrical information, rational investors, and endogenous market
entry allow arbitrage pricing to regulate the system to presumably its stable equilibrium over
time.
The parameters for this model are zeta and eta as described by the following equation (3):

Here, zeta is the damping parameter and eta is the frequency. The length of time
necessary for the system to reach equilibrium is sensitive to both the parameters described above
as well as the initial value of x, though since the system will ultimately reach equilibrium no
matter what the initial value of x, the system is said to have insensitive dependence on initial
conditions.
In order to estimate the parameters eta and zeta, two methods will be used: local linear
approximation and latent differential equation modeling. In the local linear approximation
method, a three dimensional state space is created for each level of tau between 1 and 850. At
each level of tau, estimates of the first derivative of the function (4) and the second derivative (5)
are calculated given the following formulae:

Here, x1, x2, x3 are equal to the x value of each dimension of the state space embedding.
These estimates of the derivative can then be used to develop eta and zeta estimates for each
level of tau. The level of tau that provided the estimates with the greatest goodness of fit
parameter (R2 = .700) is eight. At tau equal to eight, zeta is approximated as -0.004 and eta is
approximated as -0.033. Given these parameters, one can create a model of the behavior of the
system.

a. b.

Figure 1: (a) Damped linear oscillator model conforming to equation 3 given LLA
parameters and (b) vector field representation of this model for the change in x and .

This representation has initial condition of x at time one equal to seven as a proxy for the
equity market premium. The model damps very slowly and begins to approach zero as time
approaches 2000 periods (months). Relatively shorter periods of time, such as 120 months, result
in an oscillating premium that appears much more sinusoidal and regular.

a.

Figure 2: (a) Damped linear oscillator model conforming to equation 3 given LLA
parameters.
However, the null hypothesis of local linear approximation is that ητ2 = -2. Thus, for this
tau and all other meaningful taus observed in this analysis, we must fail to reject the null
hypothesis.
The data may also be fit to a latent differential equation model using the Mx software
package. The structure of the model used to fit the data can be described by figure 3:

a.

Figure 3: (a) Structure of latent differential equation model used in eta and zeta
parameter estimation in Mx.

This analysis provides somewhat different estimates for the parameters eta and zeta. This
model predicts that the best estimation of eta is 0.211 and zeta is -0.098 (measure of dynamic
error = 4.71x10-5). These parameters are once again used to simulate the behavior of the
function.

a. b.

Figure 4: (a) Damped linear oscillator model conforming to equation 3 given LDE
parameters and (b) vector field representation of this model for the change in x and .

Notice here that in just five months, the level of the premium will quadruple according to
the model. In Figure 3b, one can clearly see the point repellent at the origin. This point repellent
is what causes the system to move violently from its equilibrium value. According to the latent
differential equation model, it appears that there is no steady state equilibrium for this system.

Discussion

The one definitive statement that can be made about the level of the equity risk premium
is that its behavior is characterized and masked by a high amount of noise. This is evidenced by
the low descriptive power of the Fourier analysis as well as by the autoregressive plot that
exhibits only one statistically significant level of tau given a 95% confidence interval. This is not
to say that the level of the equity premium is entirely a random walk, however. One must use the
two inferential models to analyze the random or deterministic nature of the system. According to
the local linear approximation model, one cannot reject the null hypothesis, and thus, one cannot
claim that the model parameters are statistically significant. According to the latent differential
model, it appears that the system does not actually behave as the hypothesis predicts. In fact, the
latent differential model predicts that the system is easily thrown out of equilibrium and
continues towards infinity. Thus, the implication is that market shocks must bring the system
back in order, as opposed to the hypothesis that market shocks cause the system to depart from a
steady state equilibrium.
Given the very slow nature of the damping in the linear approximation method, an
examination of the premium’s fluctuations for a more practical time period (10 to 20 years) may
appear to behave, to the casual observer, as a sinusoidal wave. This is consistent with the popular
security valuation method of comparable multiples analysis (Reilly and Brown 2003). This
method assumes that market security prices systematically fall out of sync with the market
reference but will eventually return and surpass their intrinsic value. The primary concept is that
the market value oscillates around its intrinsic value.
The data and analyses also suggest that the popular method of estimating the current
long-term equity risk premium as simply the geometric average of the historical equity risk
premium is flawed and virtually useless. There are two commonly accepted practices for
estimating the equity risk premium necessary for asset pricing: forward looking models and
historical averages (Arnott and Bernstein 2002). The estimation of the equity risk premium is
vitally important for asset pricing. Given the current paradigm for asset pricing, the Capital Asset
Pricing Method (CAPM), one can observe that the expected return for any capital asset is
extremely sensitive to the equity risk premium. The CAPM can be described as the following
(6):

Here, rfis the risk-free rate (the 90-day US Treasury rate in this study), β is the sensitivity
to systemic market risk, and rm-rfrepresents the equity risk premium. Thus, for a β of 1.00,
which is the market average, each 1% change in the estimation of the equity risk premium will
alter the estimation of the asset return by 1% as well. This is economically significant.
Currently, it is widely held that the long-term geometric average of the equity risk
premium is a good proxy for its future value. This average, as described earlier, is roughly
7.10%. Given that that the autoregressive analysis in this study shows that there is almost no
statistically significant lag for autocorrelation, the suggestion is that historical returns are an
invalid source for determining a future proxy for the premium. Thus, one must turn to forward-
looking models. According to Arnott and Bernstein, the most appropriate estimation from
forward-looking models of the equity risk premium is between 2.00% and 4.00% (Arnott and
Bernstein 2002). That is a very large discrepancy from the more widely used 7.10%.

Conclusion
The results of this study are inconclusive. While the local linear approximation method to
estimate parameters for the damped linear oscillator model appears to suggest that the equity risk
premium in the United States is self-regulating about a certain mean level and damps itself
slowly as a result of endogenous market forces, this is not statistically significant. The results
also suggest that this mean reversion is periodically interrupted by exogenous market shocks. On
the other hand, the latent differential model appears to suggest the opposite – that the equity
premium is not stable and is easily sent spiraling to unreasonable levels. Thus, these data tend to
suggest that perhaps the linear oscillator model is not an appropriate paradigm for this function.
This study also finds evidence to conclude that historical levels of the equity risk
premium are not a good proxy for future estimations of its value. Nevertheless, this is currently
far and away the most prevalent method even among industry professionals. Welch finds that of
236 economists surveyed, the average expected future equity risk premium is 7% – clearly a
function of using historical averages rather than forward-looking models (Welch 2000). Areas
for further study include finding a model for the equity risk premium and a psychological study
of why economists continue to use historical averages as a proxy for the future premium.

Notes

1
This analysis uses 40 iterations.
2
10 and 20-year historical equity risk premium averages are commonly used within the financial
community for practical applications such as valuation and capital budgeting.

References

Arnott, R. D., and P.L. Bernstein (2002). “What risk premium is ‘normal’?” Financial Analysts
Journal, 58, 64-85.
Black, F., and R. Litterman (1992). “Global portfolio optimization.” Financial Analysts Journal,
48, 28-44.
Brennan, M. J., E.S. Schwartz, and R. Lagnado (1997). “Strategic asset allocation.” Journal of
Economic Dynamics and Control, 21, 1377-1403.
Fabozzi, F. J., F. Gupta, and H.M. Markowitz (2002). “The legacy of modern portfolio theory.”
The Journal of Investing, Fall, 7-22.
Kocherlakota, N. R. (1996). “The equity premium: It’s still a puzzle.” Journal of Economic
Literature, 34, 42-71.
Mehra, R. and E.C. Prescott (1985). “The equity premium: A puzzle.” Journal of Monetary
Economics, 15, 145-161.
Reilly, F. K., and K.C. Brown (2003). Investment analysis & portfolio analysis (7e). New York,
New York: Thomson.
Shiller, R. J. (2003). “From efficient markets theory to behavioral finance.” Journal of Economic
Perspectives, 17, 83-104.
Siegel, J. J and R.H. Thaler (1997). “The equity premium puzzle.” Journal of Economic
Perspectives, 11, 191-200.
Welch, I. (2000). “Views of financial economists on the equity premium and on financial
controversies.” Journal of Business, 73, 501-537.
Appendices
 

S-ar putea să vă placă și