Long memory and nonlinearities in realized volatility: A Markov switching approach

https://doi.org/10.1016/j.csda.2010.12.008Get rights and content

Abstract

Realized volatility is studied using nonlinear and highly persistent dynamics. In particular, a model is proposed that simultaneously captures long memory and nonlinearities in which level and persistence shift through a Markov switching dynamics. Inference is based on an efficient Markov chain Monte Carlo (MCMC) algorithm that is used to estimate parameters, latent process and predictive densities. The in-sample results show that both long memory and nonlinearities are significant and improve the description of the data. The out-sample results at several forecast horizons show that introducing these nonlinearities produces superior forecasts over those obtained using nested models.

Introduction

It is well-known that accurately measuring and forecasting financial volatility plays a central role in many pricing and risk management problems. With high frequency intra-daily data sets becoming widely available, more accurate estimates of volatility can be obtained. Realized volatility (RV), i.e. the sum of intra-day squared returns, reduces the noise in the volatility estimate considerably when compared to other volatility measures, such as squared or absolute daily returns. Thus, volatility becomes observable and can be modeled directly, rather than being treated as a latent process both in a GARCH or in a stochastic volatility setup.

Many empirical regularities on RV have been well-documented in recent literature and a detailed review has been provided by McAleer and Medeiros (2008b). One of the most relevant empirical regularities is that RV exhibits high persistence, as is evidenced in the research of Andersen et al. (2003), Koopman et al. (2005) and Corsi (2009). For this reason, linear fractionally integrated models (ARFIMA) are generally used to capture this feature. A flexible strategy to model serial dependencies for RV has been proposed in Barndorff-Nielsen and Shephard (2002) through a superposition of ARMA(1, 1) processes, whereas Corsi (2009) has introduced the Heterogeneous Autoregressive model (HAR) given by a combination of volatilities measured over different time horizons. However, Granger and Ding (1996) found out that persistence in volatility tends to be non-constant over time and Longin (1997) provided evidence of an usually high level of persistence when volatility is low, thus suggesting the presence of nonlinearities.

On the other hand, it is well-known that a long memory can be overestimated when a regime shifts or structural breaks are not taken into account. In fact, these are confounding factors and distinguishing between them can be rather troublesome (Diebold and Inoue, 2001). However, recent statistical tests aimed at disentangling the effects of long memory and level shifts have been proposed in Baillie and Kapetanios (2007) and in Ohanissian et al. (2008). Baillie and Kapetanios (2007) found the presence of nonlinearity together with a long memory in realized volatilities for currencies, whereas Ohanissian et al. (2008), in their empirical application, provided evidence that realized volatility of exchange rates, such as DM/$ and Yen/$ is properly described by a true long memory process. This latter result seems in contrast with Perron and Qu (2010), who claim that short memory models with level shifts are appropriate to describe volatility. Similarly, Carvalho and Lopes (2007) model volatility with a short memory switching regime dynamics. Chen et al. (2008) propose a range-based threshold heteroskedastic model, whereas He and Maheu (2010) propose a GARCH model subject to an unknown number of structural breaks.

The question of whether long memory is spurious is still debateable. It thus appears of obvious interest to join both models’ features into a single time series model. In this way, it is also possible to check whether the benefits of combining long memory and nonlinearities represent an improvement in forecasting accuracy. Recently, some time series models have suggested combining long memory and nonlinearities to describe conditional variances. Contributions in this direction are, respectively, Martens et al. (2004) and Hillebrand and Medeiros (2008) who built from an ARFIMA model by allowing for smooth level shifts, day of the week effects and leverage. McAleer and Medeiros (2008a) propose a different strategy by introducing a multiple regime smooth transition extension of HAR. Finally, Lux and Morales-Arias (2010) introduce a regime-switching multifractal model.

In this paper, we consider a different strategy to model abrupt changes in the conditional mean and time varying long range dependency. We base our analysis on a Markov switching model that is in line with the seminal work of Hamilton (1989) in which level shifts are modeled through a binary non-observable Markov process and in which the parameters, including the degree of persistence, are state dependent. Persistence is introduced through a standard ARFIMA model.

Furthermore, we think it is important to include exogenous regressors in the model’s specification. Following Bandi and Perron (2006), we consider the implied volatility as a predictor, since it is proven to be an unbiased long run forecast of future RV, once controlling for a fractionally cointegration relation. An interesting study on the relation between realized volatility, long range persistence and exogenous regressors has recently been completed by Corsi and Renò (2010).

We use Bayesian estimation techniques and goodness-of-fit indicators to assess the in-sample performance of our model. We also pay particular attention to the forecasting ability of the proposed models.

We base our empirical analysis on the 5 min intra-daily series of Standard & Poor’s 500 (S&P500) stock index over the period 1 January 2000 to 28 February 2005. Our results showed that implied volatility is important for predicting RV and also that, in the short run, long memory, together with nonlinearities improve forecasting performance. In the long run, the ARFIMA effect seems to be dominant with respect to the switching regime mechanism.

The remainder of the paper is organized as follows. Our data set is described in Section 2. Section 3 contains the definitions of the Markov switching models that are considered in this research. Our inferential solution is outlined in Section 4. The forecasting methodology is explained in Section 5 and empirical results based on simulated and real data are illustrated in Section 6 and in Section 7, respectively.

Section snippets

Realized and implied volatility

Realized volatility is an efficient and unbiased measure of the actual volatility based on the quadratic variation of a stochastic process. Theoretical and empirical features on this subject have been investigated in Barndorff-Nielsen and Shephard (2002) and in Andersen et al. (2003). Consider, for instance, a simple continuous time model for the log-price of a financial security p(t+τ)=μ(t+τ)dτ+σ(t+τ)dW(t+τ)0τ1,t=1,2, in which W(t) is a standard Brownian motion. Using well-known results on

A switching regime model for long memory realized volatility

We propose a long memory switching regime process based on a mixture of two ARFIMA(0,di,0),i=0,1 dynamics and hypothesize difference persistence levels according to two volatility regimes. We label the regime with low volatility and small conditional variance, which can be interpreted as the stable and normal volatility state of the market, as 0, whereas label 1 indicates high and fluctuating volatility. Two regimes models have been proposed in Cai (1994), Dueker (1997), Susmel (2000), Klaassen

MCMC methodology

Our goal is to jointly estimate long memory dynamics and the latent process. Inference for this model is not obvious, since the latent process S=(S1,,ST) is not observable and also because inference for long range parameters is not standard. Inference for regime switching models can be performed in the classical, as well as in the Bayesian framework, and many inferential techniques have been proposed to numerically evaluate the likelihood function. A detailed review on this topic can be found

In- and out-of-sample comparisons

To compare the out-of-sample predictive accuracy of the models, we use the Weighted Likelihood Ratio test (WLR) of Amisano and Giacomini (2007), which is based on density forecasts computed through classic or Bayesian techniques. WLR tests require a rolling window estimation procedure, that is, the recursion of the h-steps-ahead, h1, density forecasts have to be obtained using a fixed sample of size N. In practice, the first h-steps-ahead forecast is based on (y1,,yN), the second on (y2,,yN+1

Simulation results

In this section, we provide some illustrative examples to show the performance of the algorithm. The goal of this section is threefold. First, we will verify the ability of the algorithm to estimate the parameters and the latent process. Second, we will aim at determining which representation of the ARFIMA dynamics, namely AR and MA, provides a better approximation of the true model. Finally, we would like to estimate the truncation parameter that is fundamental in the determination of the

Empirical application: S&P500 realized volatility

The empirical analysis is based on 5 min returns of the Standard & Poor’s 500 index (S&P500) observed from 1 January 2000 to 28 February 2005. Our MCMC procedure has been written using the Ox©5.0 language of Doornik (2001), combined with the state space library ssfpack of Koopman et al. (1999). We ran the algorithm for 10,000 iterations with a burn-in of 2500. In our experience, that is an adequate choice to remove the dependence on the initial conditions. Results are reported in Table 3. An

Conclusions

In this paper we have modeled realized volatility through regime switching models and long range persistence. To perform inference we proposed an MCMC algorithm based on the MA representation of the ARFIMA dynamics. We also included in our specification exogenous regressors, i.e., the implied volatility, to improve the forecasting performance of the models proposed. Analyses based on simulated and real data, evidence that realized volatility is characterized by long memory and changes in

Acknowledgements

Financial support from the MIUR under grant PRIN 2008 Prot. N. 2008MRFM2H_003 is gratefully acknowledged. We wish to thank two anonymous referees and the Associate Editor for helpful comments, David Ardia, Roberto Casarin, the participants at the Second International Workshop on Computational and Financial Econometrics held in Neuchâtel on June 2008 and the participants at the 14th International Conference on Computing in Economics and Finance held in Paris on June 2008. All errors are our own.

References (65)

  • G. Koop et al.

    Bayesian analysis of long memory and persistence using ARFIMA models

    Journal of Econometrics

    (1997)
  • S. Koopman et al.

    Forecasting daily variability of the S&P100 stock index using historical, realised and implied volatility mesurements

    Journal of Empirical Finance

    (2005)
  • M. McAleer et al.

    A multiple regime smooth transition heterogeneous autoregressive model for long memory and asymmetries

    Journal of Econometrics

    (2008)
  • J.H. Albert et al.

    Bayes inference via Gibbs sampling of autoregressive time series subject to Markov mean and variance shifts

    Journal of Business & Economic Statistics

    (1993)
  • G. Amisano et al.

    Comparing density forecasts via weighted likelihood ratio tests

    Journal of Business & Economic Statistics

    (2007)
  • T. Andersen et al.

    Answering the skeptics: yes, standard volatility models do provide accurate forecasts

    International Economic Review

    (1998)
  • T.G. Andersen et al.

    The distribution of exchange rate volatility

    Journal of the American Statistical Association

    (2001)
  • T.G. Andersen et al.

    Modeling and forecasting realized volatility

    Econometrica

    (2003)
  • D. Ardia

    Bayesian estimation of a Markov-switching threshold asymmetric GARCH model with student-t innovations

    Econometrics Journal

    (2009)
  • R. Baillie et al.

    Testing for neglected nonlinearity in long memory models

    Journal of Business & Economic Statistics

    (2007)
  • F.M. Bandi et al.

    Long memory and the relation between implied and realized volatility

    Journal of Financial Econometrics

    (2006)
  • F.M. Bandi et al.

    Microstructure noise, realized volatility, and optimal sampling

    Review of Economic Studies

    (2008)
  • O. Barndorff-Nielsen et al.

    Econometric analysis of realized volatility and its use in estimating stochastic volatility models

    Journal of the Royal Statistical Society. Series B

    (2002)
  • A. Brockwell

    Likelihood based analysis of a class of generalized long memory models

    Journal of Time Series Analysis

    (2006)
  • J. Cai

    A Markov model of unconditional variance in ARCH

    Journal of Business & Economic Statistics

    (1994)
  • C.K. Carter et al.

    On Gibbs sampling for state space models

    Biometrika

    (1994)
  • G. Celeux et al.

    Deviance information criteria for missing data models

    Bayesian Analysis

    (2006)
  • N.H. Chan et al.

    State space modeling of long-memory processes

    Annals of Statistics

    (1998)
  • F. Corsi

    A simple approximate long-memory model of realized volatility

    Journal of Financial Econometrics

    (2009)
  • Corsi, F., Renò, R., 2010. Discrete-time volatility forecasting with persistent leverage effect and the link with...
  • J. Doornik

    Ox: An Object-Oriented Matrix Programming Language

    (2001)
  • Doornik, J.A., Ooms, M., 2006. A package for estimating, forecasting and simulating ARFIMA models: ARFIMA package 1.04...
  • Cited by (51)

    View all citing articles on Scopus
    View full text