In statistics, stochastic volatility models are those in which the variance of a stochastic process is itself randomly distributed.[1] They are used in the field of mathematical finance to evaluate derivative securities, such as options. The name derives from the models' treatment of the underlying security's volatility as a random process, governed by state variables such as the price level of the underlying security, the tendency of volatility to revert to some long-run mean value, and the variance of the volatility process itself, among others.
Stochastic volatility models are one approach to resolve a shortcoming of the Black–Scholes model. In particular, models based on Black-Scholes assume that the underlying volatility is constant over the life of the derivative, and unaffected by the changes in the price level of the underlying security. However, these models cannot explain long-observed features of the implied volatility surface such as volatility smile and skew, which indicate that implied volatility does tend to vary with respect to strike price and expiry. By assuming that the volatility of the underlying price is a stochastic process rather than a constant, it becomes possible to model derivatives more accurately.
A middle ground between the bare Black-Scholes model and stochastic volatility models is covered by local volatility models. In these models the underlying volatility does not feature any new randomness but it isn't a constant either. In local volatility models the volatility is a non-trivial function of the underlying asset, without any extra randomness. According to this definition, models like constant elasticity of variance would be local volatility models, although they are sometimes classified as stochastic volatility models. The classification can be a little ambiguous in some cases.
The early history of stochastic volatility has multiple roots (i.e. stochastic process, option pricing and econometrics), it is reviewed in Chapter 1 of Neil Shephard (2005) "Stochastic Volatility," Oxford University Press.
Starting from a constant volatility approach, assume that the derivative's underlying asset price follows a standard model for geometric Brownian motion:
where is the constant drift (i.e. expected return) of the security price , is the constant volatility, and is a standard Wiener process with zero mean and unit rate of variance. The explicit solution of this stochastic differential equation is
The maximum likelihood estimator to estimate the constant volatility for given stock prices at different times is
its expected value is
This basic model with constant volatility is the starting point for non-stochastic volatility models such as Black–Scholes model and Cox–Ross–Rubinstein model.
For a stochastic volatility model, replace the constant volatility with a function that models the variance of . This variance function is also modeled as Brownian motion, and the form of depends on the particular SV model under study.
where and are some functions of , and is another standard gaussian that is correlated with with constant correlation factor .
The popular Heston model is a commonly used SV model, in which the randomness of the variance process varies as the square root of variance. In this case, the differential equation for variance takes the form:
where is the mean long-term variance, is the rate at which the variance reverts toward its long-term mean, is the volatility of the variance process, and is, like , a gaussian with zero mean and variance. However, and are correlated with the constant correlation value .
In other words, the Heston SV model assumes that the variance is a random process that
Some parametrisation of the volatility surface, such as 'SVI',[2] are based on the Heston model.
The CEV model describes the relationship between volatility and price, introducing stochastic volatility:
Conceptually, in some markets volatility rises when prices rise (e.g. commodities), so . In other markets, volatility tends to rise as prices fall, modelled with .
Some argue that because the CEV model does not incorporate its own stochastic process for volatility, it is not truly a stochastic volatility model. Instead, they call it a local volatility model.
The SABR model (Stochastic Alpha, Beta, Rho), introduced by Hagan et al.[3] describes a single forward (related to any asset e.g. an index, interest rate, bond, currency or equity) under stochastic volatility :
The initial values and are the current forward price and volatility, whereas and are two correlated Wiener processes (i.e. Brownian motions) with correlation coefficient . The constant parameters are such that .
The main feature of the SABR model is to be able to reproduce the smile effect of the volatility smile.
The Generalized Autoregressive Conditional Heteroskedasticity (GARCH) model is another popular model for estimating stochastic volatility. It assumes that the randomness of the variance process varies with the variance, as opposed to the square root of the variance as in the Heston model. The standard GARCH(1,1) model has the following form for the continuous variance differential:[4]
The GARCH model has been extended via numerous variants, including the NGARCH, TGARCH, IGARCH, LGARCH, EGARCH, GJR-GARCH, Power GARCH, Component GARCH, etc. Strictly, however, the conditional volatilities from GARCH models are not stochastic since at time t the volatility is completely pre-determined (deterministic) given previous values.[5]
The 3/2 model is similar to the Heston model, but assumes that the randomness of the variance process varies with . The form of the variance differential is:
However the meaning of the parameters is different from Heston model. In this model, both mean reverting and volatility of variance parameters are stochastic quantities given by and respectively.
Using estimation of volatility from high frequency data, smoothness of the volatility process has been questioned.[6] It has been found that log-volatility behaves as a fractional Brownian motion with Hurst exponent of order , at any reasonable timescale. This led to adopting a fractional stochastic volatility (FSV) model,[7] leading to an overall Rough FSV (RFSV) where "rough" is to highlight that . The RFSV model is consistent with time series data, allowing for improved forecasts of realized volatility.[6][8]
Once a particular SV model is chosen, it must be calibrated against existing market data. Calibration is the process of identifying the set of model parameters that are most likely given the observed data. One popular technique is to use maximum likelihood estimation (MLE). For instance, in the Heston model, the set of model parameters can be estimated applying an MLE algorithm such as the Powell Directed Set method [1] to observations of historic underlying security prices.
In this case, you start with an estimate for , compute the residual errors when applying the historic price data to the resulting model, and then adjust to try to minimize these errors. Once the calibration has been performed, it is standard practice to re-calibrate the model periodically.
An alternative to calibration is statistical estimation, thereby accounting for parameter uncertainty. Many frequentist and Bayesian methods have been proposed and implemented, typically for a subset of the abovementioned models. The following list contains extension packages for the open source statistical software R that have been specifically designed for heteroskedasticity estimation. The first three cater for GARCH-type models with deterministic volatilities; the fourth deals with stochastic volatility estimation.
Many numerical methods have been developed over time and have solved pricing financial assets such as options with stochastic volatility models. A recent developed application is the local stochastic volatility model.[13] This local stochastic volatility model gives better results in pricing new financial assets such as forex options.
There are also alternate statistical estimation libraries in other languages such as Python:
{{cite journal}}
: Cite journal requires |journal=
(help)