Final Thesis: Master's Degree in Economics and Finance
Final Thesis: Master's Degree in Economics and Finance
Economic’s Department
Final Thesis
Graduand:
Karim Hasouna,
N° 859207
"We all know that art is not truth. Art is a lie that makes us realize truth, at
least the truth that is given us to understand. The artist must know the manner
whereby to convince others of the truthfulness of his lies". Pablo Picasso. 1923
Contents
4 Concluding remarks 85
4
5 Bibliography 88
A Appendix 92
A.1 A Appendix . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 92
A.2 B Appendix . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 94
5
Abstract
In this thesis I shall examine the characteristics of the volatility in the finan-
cial markets. In particular, the volatility is extrapolated both from the historical
volatility by time series of past market prices and from derivative instruments pro-
viding an implied volatility. The first part explores the causes of volatility, espe-
cially volatility clustering, and explain the behavioural reactions of the stockhold-
ers. It is a well-known fact that there are GARCH models and many others that are
accurate and useful to estimate the conditional variance. Anyway, looking the his-
torical returns could be not be enough to fit the model on the data. Our purpose is
to create a non-linear Univariate model to evaluate the financial markets using the
Generalised Autoregressive Conditional Heteroskedasticity (GARCH) model with
the CBOE Volatility Index (VIX) as a exogenous variable. The exogenous variable
VIX is independent of the GARCH model but is included in the new model we
want to realize. Using the daily rates of return of 10 major indices, we want to de-
termine if the new model created, adding an exogenous variable, is better than
the single GARCH model. Therefore, the empirical analysis analyses the volatility
implementing the GARCH with the exogenous implied volatility, determined look
forward in time, being derived from the market price of a market-traded deriva-
tive. It is using the Variance Swaps, based on the S&P 500 Index, the core index for
U.S. equities, and estimates expected volatility by aggregating the weighted prices
of S&P 500 puts and call plain vanilla options over a wide range of strike prices.
By empirically examining the time series of different world indices we hope to
produce a more complete understanding of the utility of the VIX into the GARCH
models.
[LaTeX typeset]
6
1 Introduction and brief history
series over time and is essentially a measure of deviation from the ex-
and negative deviation. We can calculate the volatility using the histori-
cal volatility, which measures a time series of past market prices, or with
ity is that is not directly observable, but can be estimated from historic
data, or using an option pricing model. From the returns is not possible
trading day but is possible to estimate the daily volatility using the in-
contain only very limited information about the overnight volatility. The
returns can be analysed using different time orders like years, months,
days, hours or minutes but each one have some problems and choose
the accurate one depends of the model. There are various character-
istics that are commonly seen in asset returns. The most important is
the volatility clustering where the volatility may be high for certain time
periods and low for other periods; but the volatility follows also other
properties such as the volatility evolves over time with a continuous be-
havior, but does not diverge to infinity. We notice that there is also a
7
leverage effect, meaning the volatility react differently to a big price in-
Scholes formula, then we can use the price to obtain the implied volatil-
ity. The implied volatility is derived under the assumption that the price
might be different from the actual volatility. The volatility index (VIX)
plied volatility, started to trade in futures on March 26, 2004.1 One char-
ticity, but we can solve this problem using the ARCH (Autoregressive
model, more complex but also more parsimonious. The GARCH model
is a non linear statistical model for time series that describes the vari-
tion available on the time series itself and not on any exogenous infor-
1
See, Ruey S. Tsay. (2010).[1]
8
mation.2 . The idea of the new model is to use the GARCH model with
the implied volatility as a exogenous variable. The use of the VIX with
the GARCH is because the volatility index, seen as a general fear index
the relationship between the volatility of prices and their returns cause
The uncertainty of the future price is hard to forecast and the better fore-
cast of the future is analysing the past time series. Anyway, only an anal-
ysis of the past time series is not enough to forecast the future price be-
cause there are many others variables that influence the market like the
2
See, Vega Ezpeleta. (2015).[2]
9
2 Chapter 1: The structure of the model
..., yn where i is a time index. This kind of data are usually referred as a
time series.3 The analysis of time series is essential when we fit a model
that explain how the observed data works. The time series is defined
observe a time series, the fluctuations appear random, but often with
the same type of stochastic behaviour from one time period to the next.
One of the most useful methods for obtaining parsimony in a time se-
variance, and covariance are unchanged by time shifts. Thus, the mean
and variance do not change with time and the correlation between two
observations depends only on the lag, the time distance between them.
of (yt 1 , ....., yt k ) is identical to that of (yt 1+t , ....., yt k +t ) for all t, where k is
integers. In other words, strict stationarity requires that the joint distri-
bution of (yt 1 , ...., yt k ) is invariant under time shift. This is a very strong
a weakly stationary time series when the both mean of yt and the covari-
3
See, Pastore (2018).[3]
10
ance between yt and yt −ι are time invariant, where ι is an arbitrary in-
that it is stationary but there are many different tests to evaluate if the
time series data either to better understand the data or to predict fu-
ture points in the series. Usually, a good approach is to use the ARMA
and/or with other variants. The stationarity aids to evaluate the time se-
ries because the statistical properties of the process do not change over
Many statistical models, analysing the time series, assume that a ran-
dom sample comes from a normal distribution. It well known that from
that the time series randomly display outliers) therefore to use the nor-
investigate how the distribution of the data differs from a normal distri-
4
See, Ruppert D., Matteson D.(2015).[4]
11
2.2 Volatility
and the stock prices may appear too volatile to be justified by changes in
volatility can form the basis for efficient price discovery, as the relation-
ship between risk and return, while volatility dependence implies pre-
Equilibrium prices, obtained from asset pricing models (CAPM), are af-
tional volatility and early evidences uveils that bad news in the Futures
market increase volatility in the cash markets more than good news. The
sis is not the only force behind asymmetries but many others may well
12
5
the returns between daily close prices of instruments. In the math-
the variance. The square root of the variance is easier to interpret be-
the same unit of measurement as the original data. Therefore, the stan-
a set of values. A high standard deviation indicates than the values are
spread out over a wide range increasing the risk, while a low standard
deviation means that the values tend to be close to the mean of the set.
the fluctuations of the price of a given asset or, more globally, the risk of
basis for investment decisions. As the capital asset pricing model, if the
well, an increase known as the risk premium. So, if the investment car-
13
Let X be a random variable with mean value µ :
E [X ] = µ (1)
Æ
σ= E [(X − µ)2 ] (2)
Æ
= E [X 2 ] + E [−2µX ] + E [µ2 ] (3)
Æ
= E [X 2 ] − 2µE [X ] + µ2 (4)
Æ
= E [X 2 ] − 2µ2 + µ2 (5)
Æ
= E [X 2 ] − µ2 (6)
p
= E [X 2 ] − (E [X ])2 (7)
14
since these expected values need not exist.
from most of many studies is that price changes and rates of return are
symmetric distribution with fatter tails than the normal, However, even
though the time series are serially uncorrelated, they are not indepen-
dent increments process for daily stock returns in Hinich and Patterson.
The model of this thesis works when there is volatility clustering there-
fore when there is period of higher, and of lower, variation within each
ance of each series. As such, the variance of daily returns can be high
one month and show low variance the next. This occurs to such a de-
15
to new information with large price movements, these high-volatility
environments tend to ensure for a while after that first shock. In other
ing that there is not a request that the data being analyzed meet cer-
variation of volatility, but does not go into depth on why volatility varies
factory.
16
respond to stockholders dynamically solving for new asset prices.
• Leverage: When firms are financed using both debt and equity, only
the equity will reflect the volatility of the firms cash flows. However,
as the price of equity falls, a smaller quantity must reflect the same
volatility of the firm’s cash flows and so negative returns should lead
are few participants willing to buy (sell), this shock has a large ef-
will cause a small decrease in demand but also here there will be a
• State Uncertainty: When the state is uncertain, like the actual situa-
17
tion due to Covid-19, slight changes in beliefs may cause large shifts
state.
structural breaks.
18
cient to use only the most recent observations to forecast volatility, or
4.Stochastic volatility.
5. Implied volatility.
6. Realized volatility.
usually assume the mean return is zero. While this is obviously not cor-
rect, the daily mean is orders of magnitude smaller than volatility and
therefore can usually be safely ignored for the purpose of volatility fore-
12
See, Danielsson(2011).[11]
19
2.6 Moving Average models
The most obvious and easy way to forecast volatility is simply to cal-
culate the sample standard error from a sample of returns. Over time,
we would keep the sample size constant, and every day add the newest
return to the sample and drop the oldest. This method is called the Mov-
ing Average (MA) model. The observations are equally weighted, which
the most recent data are more indicative of whether we are in a high-
so that the most recent returns have the biggest weight in forecasting
where 0 < λ < 1 is the decay factor, σ̂2t the conditional volatility fore-
cast on day t. This model si not optimal because use only a λ that is
constant and identical for all assets. Obviously is not realistic that λ is
13
See, Ruppert D., Matteson D.(2015).[4]
20
2.8 The ARCH/GARCH Model
explain what is the GARCH model and why it is used. The model is useful
for the volatility clustering, therefore when the market goes through pe-
riods of high volatility and other periods when volatility is low. The vari-
ance of this model is not constant and use the conditional volatility, de-
dence to measure the impact of last period’s forecast error and volatility
turns on previous days, where older returns have a lower weight than
more recent returns. The parameters of the model are typically esti-
14
see, Engle(1982).[13]
21
dom variables Yt and is useful separate estimation of the mean from the
volatility estimation to use a more efficient model. So, the E(Yt ) is equal
ble change the distribution that fit the model, like the Student-t.
p
X
σ2t =ω+ αi Yt 2−i (10)
i =1
where p is the number of lags. Setting the lag to one in the formula
will result in the ARCH(1) model which states that the conditional vari-
that is:
22
The unconditional volatility of the ARCH(1) model is given by:
ω
σ2t = (12)
1−α
E (Y 4 )
K ur t osis = (13)
(E (Y 2 ))2
3(1 − α2 )
K ur t osis = >3 i f 3α2 < 1. (14)
1 − 3α2
There are two main restrictions that are often imposed on the param-
defines, impose:
p
X
αi < 1. (16)
j =1
23
It is only the nonnegativity constraint that always has to be imposed
One of the biggest problems with the ARCH model concerns the long
p
X q
X
σ2t =ω+ αi Yt 2−i + β σ2t − j (17)
i =1 j =1
where p,q are the numbers of lags. Setting the lag to one means that
the model is the GARCH(1,1). The most common version of the GARCH
24
where
So,
ω
σ2t = (20)
1−α−β
There are two main restrictions that are often imposed on the param-
defined when α + β > 1. We should not impose the constraint when all
run mean value, and the variance of the volatility process itself, among
25
term that is not known at time t. However, these models cannot explain
smile and skew, which indicate that implied volatility does tend to vary
16
with respect to strike price and expiry.
Volatility Index (VIX), which was originally designed to measure the mar-
index option prices. in 2003, CBOE together with Goldman Sachs, up-
dated the VIX index to reflect a new way to measure expected volatility,
agers and volatility traders alike. The new VIX index is based on the
S&P 500 Index, the core index for U.S. equities, and estimates expected
volatility by aggregating the weighted prices of S&P 500 puts and calls
over a wide range of strike prices. In 2014, Cboe enhanced the VIX in-
dex to include series of S&P 500 Weekly’s. It allows the VIX index to be
calculated with S&P 500 index option series that most precisely match
the 30-day target timeframe for expected volatility that the VIX index is
26
The VIX is a volatility index comprised of options rather than stocks,
with the price of each option reflecting the market’s expectation of fu-
ture volatility. Like conventional indices, the VIX Index calculation em-
index values. Some different rules and procedures apply when calculat-
ing the VIX index value to be used for the final settlement value of VIX
measure of how much the market expects the S&P 500 Index will fluctu-
ate in the 30 days from the time of each tick of the VIX Index.
Intraday VIX Index values are based on snapshots of SPX option bid/ask
is the most popular indicator of volatility and has been regarded as the
its market name as "the fear gauge". In general, VIX, starts to rise dur-
Thus, the greater the fear, the higher the volatility index would be.
17
see CBOE VIX (2019).[15]
27
2.10.1 VXO
The first VIX was renamed VXO, the old VIX index based on the Black-
old VIX, two puts and two calls for strikes immediately above and be-
low the current index are chosen. Near maturities (greater than eight
set of eight options. By inverting the BSM pricing formula using cur-
rent market prices, an implied volatility is found for each of the eight
σ. These volatilities are then averaged, first the puts and the calls, then
Because the BSM model assumes the index follow a geometric Brownian
motion with constant volatility, when in fact it does not, the old VIX will
only approximate the true risk-neutral implied volatility over the com-
ing month. In reality the price process is likely more complicated than
tail and a less heavy right tail than the lognormal distribution. Traders
18
See, McAleer (2007).[24]
28
use volatility smiles to allow for non-lognormality. The volatility smile
X X X X
X l (< S ) σc ,1l σp ,1
l
σc ,2l σp ,1
l
X X X X
X u (> S ) σc ,1u σp ,1u σc ,2u σp ,2u
where the X l is the strike below the price of the current index, S is
the share price of the index at maturity (higher than 8 days) and the X u
is the strike above the price of the current index. In the table there are
The first step is to average the put and call implied volatilitIes for each
X X X X X X X X
X
(σc ,1l + σp ,1
l
) X
(σc ,1u + σp ,1u ) X
(σc ,2l + σp ,2
l
) X
(σc ,2u + σp ,2u )
σ1 l = ; σ1 u = ; σ2 l = ; σ2 u =
2 2 2 2
(21)
Now average the implied volatility above and below the index level as
19
Hull (2018).[16]
29
follows:
Xu − S S − Xl Xu − S S − Xl
X Xu X Xu
σ1 = σ1 l +σ1 ; σ2 = σ2 l +σ2
Xu − Xl Xu − Xl Xu − Xl Xu − Xl
(22)
The final step is calculating the VXO is to interpolate between the two
index.
Nt 2 − 22 22 − Nt 2
V X O = V I X o l d = σ1 + σ2 (23)
Nt 2 − Nt 1 Nt 2 − Nt 1
changes are being made to update and improve VIX. The new VIX is cal-
mation from the volatility skew. The VXO used only at-the money op-
tions. The new VIX uses a newly developed formula to derive expected
volatility directly from the prices of a weighted strip of options. The VXO
The new VIX uses options on the S&P 500 Index, which is the primary
U.S. stock market benchmark. So, the new VIX provide a more precise
20
see, HaoNewVix, Hao Zhou and Matthew Chesnes. (2003).[19]
30
and robust measure of expected market volatility and to create a viable
underlying index for tradable volatility products. The VIX is more practi-
cal and simpler because it uses a formula that derives the market expec-
tation of volatility directly from index option prices rather than an algo-
model.
cess is a particular type of stochastic process where only the actual value
variable and the way that the present has emerged from the past is irrel-
evant. Predictions for the future are uncertain and must be expressed
not dependent on the particular path followed by the price in the past.
process, with a mean change of zero and a variance rate of 1 per year
It follows from the first property that ∆z itself has a normal distribution
31
with
mean = 0
p
standard deviation = ∆t
T
N=
∆t
Thus,
N
X p
z (T ) − z (0) = εi ∆t (24)
i =1
where the ε (i= 1,2,....N) are distributed φ(0,1). We know from the sec-
other. It follows that z(T) -z(0) is normally distributed, with mean equal
p
to zero and standard deviation T .
The mean change per unit time for a stochastic process is known as
the drift rate and the variance per unit time is known as the variance
32
terms on d z as
d x = ad t + bd z (25)
The a d t term implies that x has an expected drift rate of a per unit of
Wiener process. A Wiener process has a variance rate per unit time of
1.0. It follows that b times a Wiener process has a variance rate per unit
time of b 2 .21
p
∆x = a ∆t + b ε ∆t (26)
Similar arguments to those given for a Wiener process show that the
Now, we discuss the stochastic process usually assumed for the price
33
Wiener process but the assumption of constant expected drift rate is in-
pected return is constant. If S is the stock price at time t, then the ex-
rameter µ. This means that in a short interval of time, ∆t, the expected
the stock. The uncertainty of the process, the standard deviation, should
dS
= µd t + σd z (27)
S
is the volatility of the stock price. The model represents the stock price
σ2
d (l nSt ) = (µ − )d t + σd Z t (28)
2
d St σ2
− d (l nSt ) = dt (29)
St 2
34
rate, V¯ , between time 0 and time T is given by
T T
Z Z
1 2 d St St
V¯ = σ2 d t = −ln (30)
T 0
T 0
St S0
2 F0 2 St
Eˆ (V¯ ) = l n − Eˆ l n (31)
T S0 T S0
time T.
Consider
Z S∗
1
ma x (K − ST , 0)d K (32)
K =0
K 2
for some value S ∗ of S. When S ∗ < ST this integral is zero. When S ∗ >
ST it is
Z S∗
1 S ∗ ST
(K − ST , 0)d K = l n + −1 (33)
K =ST
K2 ST S ∗
Consider next
Z ∞
1
ma x (ST − K , 0)d K (34)
K =S ∗
K2
Z ST
1 S ∗ ST
(ST − K , 0)d K = l n + −1 (35)
K =S ∗
K2 ST S ∗
35
From these results it follows that
Z S∗ Z ∞
1 1 S ∗ ST
ma x (K −ST , 0)d K + ma x (ST −K , 0)d K = l n + −1
K =0
K 2
K =S ∗
K2 ST S ∗
(36)
Z S∗ Z ∞
ST ST 1 1
ln = −1− ma x (K −ST , 0)d K − ma x (ST −K , 0)d K
S∗ S∗ K =0
K 2
K =S ∗
K2
(37)
This shows that a variable that pays off in ST can be replicated using
Z S∗ Z∞
ST F0 1 RT 1 RT
Eˆ l n = −1− e P (K )d K − e C (K )d K (38)
S∗ S∗ K =0
K 2
K =S ∗ K 2
where C(K) and P(K) are the prices of European call and put options
with strike price K and maturity T and R is the risk-free interest rate for
ST S∗ ST
Eˆ l n = + Eˆ l n (39)
S0 S0 S∗
36
Z S ∗
2 F0 2 S ∗ 2 F0 2 1 RT
Eˆ (V¯ ) = l n − l n − −1 + e P (K )d K +
T S0 T S0 T S ∗ T K =0 K 2
Z ∞
1 RT
+ e C (K )d K (40)
K =S ∗
K2
which reduces to
Z S ∗ Z∞
2 F0 2 F0 2 1 RT 1 RT
Eˆ (V¯ ) = l n − −1 + e P (K )d K + e C (K )d K
T S∗ T S∗ T K =0 K 2 K =S ∗
K 2
(41)
This result is the Variance Swaps where we find the VIX formula. The
where the parties agree to exchange a pre-agreed variance level for the
option, represents the level of volatility bought of sold and is set at trade
the swap initially has zero value. If the subsequent realised volatility is
above the level set by the strike, the buyer of a variance swap will be
a variance swap is short volatility and profits if the level of variance sold
37
(the variance swap strike) exceeds that realised.
expansion:
2
F0 F0 1 F0
= −1 − −1 (42)
S∗ S∗ 2 S∗
culated as
2 X ∆K 1
F
2
i
Eˆ (V¯ )T = σ2 = V I X 2 = 2
e RT Qi (K i ) − −1 (43)
T i Ki T K0
Where:
σ *100 = V I X
T =Time to expiration
P0 )
Q (K i ) The midpoint of the bid-ask spread for each option with strike K i .
38
So, the constant 30-day volatility index VIX:
v
u§ NT2 − N30 N30 − NT1 N365
ª
V I X = 100 ∗ T1 σ1
2
+ T2 σ2
2
∗ ; (44)
t
NT2 − NT1 NT2 − NT1 N30
range of movement in the S&P 500 index over the next month, at a 68%
tribution). For example, if the VIX is 30, this represents an expected an-
The expected volatility range for a single month can be calculated by di-
viding the VIX by (12) which would imply a range of +/- 8.67% over
p
the next 30-day period. Similarly, expected volatility for a week would
be 30 divided by (52), or +/- 4.16%. The VIX uses calendar day annu-
p
per day. The calendar day approach does not account for the number of
when the financial markets are open in a calendar year. In the financial
markets trading days typically amount to 252 days out of a given calen-
dar year.
39
2.10.4 Alternatives of the VIX
The VIX is not the only volatility index but there are many different in-
for S&P 500, VXN for the Nasdaq, VSTOXX for the Euro Stoxx 50, VDAX
for DAX and VSMI for the SMI). They represent the theoretical level of
traded option prices. In fact, theoretical variance swap levels are first
calculated for listed option maturities, and then the 30-day index level
ery 10 minutes), and using the data to obtain the covariance matrix.
The main advantage is that it is purely data driven and there is no re-
to be available; such data are often difficult to obtain, hard to use, not
to deal with diurnal patterns in volume and volatility when using real-
ume and volatility throughout the day). Moreover, the particular trading
platform in use is likely to impose its own patterns on the data. All these
40
issues complicate the implementation of realized volatility models.
that use the historical volatility and the implied volatility using their re-
clustering and use the time series. The VIX is useful during periods of
high uncertainty and the value of the Index increase rapidly. The for-
mula of the VIX use the S&P 500 option prices with, on average, a 30
days maturity.
the VIX and we evaluate if this model increase the performance in the
is given by
GARCH with the implementation of the parameter δ > 0 that capture the
41
Therefore,
α ≥ 0; (46)
β ≥ 0; (47)
α + β < 1; (48)
So, like in the GARCH model, all variables must be positive and the
evaluate the model with the EWMA and the GARCH model to determine
if the new model created is more accurate than the others already avail-
model.
42
3 Chapter 2: Empirical Research
3.1 Data
daily log returns series of each closing price for the 10 major global in-
DAX 30 (Germany)
CAC 40 (France)
TOPIX (Japan)
43
DAX30
CAC40
FTSE100
80
FTSEMIB
SP500
NASDAQ
HANGSENG
CHINA50
TOPIX
60
SPI100
40
% benefits
20
0
−20
dates
world in order to test the model with different time series, looking to
diversify and reduce the correlation among them. Every index is com-
posed with different sector and Iights and the comparison among them
tures of the indices because they are highly liquid ensuring smaller trans-
action costs due to bid/ask spreads and an efficient asset pricing. The
44
period analyzed starts the 10, November 2014 and ends the 26, June 2020
daily returns of every time series and fitted data among them with the
manage and analyze the empirical research. So, for a better interpre-
tation of data graphs, the x axis show values starting from 0, concern-
ing the 10 October 2014, the 500 concerning the 10 October 2016, 1000
concerning the 29 October 2018 and to the ends the 1418 observation
45
Figure 2: Logarithmic daily returns
As we can see from figure 2, the returns of the indices are plotted over
time and the volatility tends to vary. The differenced daily log returns
series show higher degree of volatility in the last 6 Months than relative
data permits to remove the trend of the time series and make a time se-
46
ries stationary. The graphs show a different range of daily returns ex-
hibit different volatility among them, but the volatility is different also
Usually the returns of time series are assumed to follow a normal dis-
tribution, but the empirical analysis show how assume the normality is
wrong and is almost impossible find data that roughly fits a bell curve
from leptokurtosis.
to fit the model in the returns. We computed the Jarque-Bera test that
sample data have the skewness and kurtosis matching a normal distri-
the test statistics is far from zero, it signals the data do not have a normal
the hypothesis that the data are from a normal distribution. the null hy-
pothesis is a joint hypothesis of the skewness being zero and the excess
47
kurtosis being zero. To avoid the risk to take to falsely reject H0 or do
alpha of the statistics. For example, if the p-value of a test statistic result
X-squared df p-value
DAX 30 5313.1 2 0
CAC 40 10652 2 0
FTSE 100 8339.6 2 0
FTSE MIB 19397 2 0
S&P 500 21717 2 0
NASDAQ 8910.4 2 0
HANG SENG 1291.3 2 0
CHINA 50 10884 2 0
TOPIX 3381.6 2 0
SPI 100 12911 2 0
As we can see, the test exhibit high signals that the distribution is
not normal. The p-value is statistically significant to reject the null hy-
how differs from a normal distribution. First of all, I compute the his-
togram of the time series, then I calculate the normal density function
with the sample mean and the sample standard deviation of the given
series (dashed line) and, last but not least, I compute the estimate of the
48
50
50
40
40
30
Density
Density
30
20
20
10
10
0
0
−0.10 −0.05 0.00 0.05 0.10 −0.10 −0.05 0.00 0.05 0.10
40
10 20 30 40 50
30
Density
Density
20
10
0
49
50
60
40
Density
Density
30
40
20
20
10
0
0
−0.10 −0.05 0.00 0.05 0.10 −0.10 −0.05 0.00 0.05 0.10
SP500 NASDAQ
10 20 30 40 50 60
40
30
Density
Density
20
10
0
HANGSENG CHINA50
50
10 20 30 40 50 60
50
40
Density
Density
30
20
10
0
TOPIX SPI100
So, we have the histograms that present the returns of the indices
showing outliers. Then, we have a dashed line that represent the nor-
mal distribution. Finally we have a solid line that represent the empir-
figures shows that a normal distribution does not fits well the sample es-
timate of the density of returns for the Major Indices. With the Shapiro
test every p-value was less than a given confidence level (0.05) then the
null hypothesis was rejected for every time series. The t-distribution
of their use in testing and confidence intervals when the data are mod-
51
gained added importance as models for the distribution of heavy-tailed
and kurtosis. Moreover, I provided the Ljung-Box test with the p-value
and in autoror the five lags of the AutoCorrelation Function (ACF). The
pothesis of the Ljung-Box test is not rejected when the data are indepen-
dently distributed, and the critical region for rejection of the hypothesis
The following table 1 shows the output for each time series.
52
Table 2: Descriptive statistics and Ljung-box test
DAX 30 CAC 40 FTSE 100 FTSE MIB S&P 500 NASDAQ HANG SENG CHINA 50 TOPIX SPI 100
mean 2.21E-04 1.30E-04 -4.69E-05 5.54E-05 2.78E-04 6.11E-04 1.43E-04 4.22E-04 1.69E-04 -2.18E-05
std 0.0131 0.013 0.0111 0.0152 0.0117 0.0132 0.012 0.0181 0.0125 0.0114
min -0.1175 -0.1322 -0.1008 -0.169 -0.1095 -0.1148 -0.0768 -0.1598 -0.0845 -0.1027
max 0.1006 0.0824 0.0829 0.0817 0.0935 0.0927 0.0557 0.1611 0.0883 0.0725
skewness -0.5952 -1.2946 -0.9968 -1.7525 -0.9001 -0.8143 -0.5607 -0.4358 -0.4123 -1.1113
kurtosis 12.4079 16.1753 14.7122 20.777 22.0872 15.1721 7.5384 16.5448 10.5202 17.6142
Ljung Box 0.0000 0.0000 0.0000 0.0000 1.0000 1.0000 1.0000 1.0000 0.0000 1.0000
P-value 0.2273 0.4217 0.8822 0.3527 1.41E-11 2.01E-14 0.0028 8.32E-05 0.122 1.44E-08
Stats 6.9114 4.9526 1.7524 5.5482 59.688 73.4015 18.0867 26.1569 8.6912 45.0106
autoror : : : : : : : : : :
lag1 0.0060 -0.0006 -0.0248 -0.0299 -0.1584 -0.1927 -0.0821 -0.0932 -0.0031 -0.1593
lag2 0.0679 0.0529 -0.0024 0.0462 0.0782 0.0653 0.0446 -0.0502 0.0187 0.0457
lag3 0.0137 0.0189 0.0205 0.0233 0.0232 0.0367 0.0265 -0.0005 -0.0583 0.0208
lag4 0.0056 -0.0003 -0.0137 -0.0067 -0.085 -0.0737 -0.0565 0.0724 -0.0341 -0.0268
lag5 -0.0001 -0.0183 0.0023 0.017 0.0561 0.0599 0.0105 -0.0446 -0.0347 0.0558
As we can see in the table, every time series shows negative skewness,
The kurtosis is high thus the data seems to have a leptokurtosis distri-
bution (fat-tailed), due to extreme outliers. The time series are simi-
lar about the skewness and the kurtosis following the fact that they are
similar in terms of structure of data. The Ljung-Box test (lbq test) has
Ljung-Box test show that there is no serial dependence in the DAX 30,
CAC 40, FTSE 100, FTSE MIB indices and in the TOPIX. It is determined
because the test do not reject the null hypothesis, the p-value is higher
53
than alpha (the significance level) 0.05 and the critical region for rejec-
ble. Otherwise, we reject the null hypothesis in the S&P 500, NASDAQ,
HANG SENG, CHINA 50 AND SPI 100 indices, showing how the observed
than the pre-specified significance level alpha. The t-statistics show that
the values are higher than the 11.07 and the ACF of lags 5 are autocorre-
create a model.
3.4 Correlation
It has been frequently observed that USA markets leads other devel-
oped markets in Europe or Asia, and that at times the leader becomes
the follower. With different markets, some assets’ returns are observed
culated with the correlation of their returns. The correlation is just a par-
54
they present also high correlation with the USA markets.
As we can see, the correlation increase with the status of the devel-
55
Then, we can show the correlation between the indices and the VIX
to analyse how the Volatility index used show negative correlation with
VIX
DAX 30 -0.4246994
CAC 40 -0.4430257
NASDAQ -0.6443102
CHINA 50 -0.1266067
TOPIX -0.1659127
but is highly relevant in the Standard & Poor index and in the Nasdaq.
56
First of all, I plot a graph of the squared time series to analyze if exhibit
volatility clustering, then I evaluate the squared residuals using the Au-
in exam.
1.0
0.012
0.8
0.010
0.6
0.008
y1^2
ACF
0.006
0.4
0.004
0.2
0.002
0.0
0.000
Time Lag
0.8
0.6
0.010
y2^2
ACF
0.4
0.005
0.2
0.0
0.000
Time Lag
57
FTSE 100 Squared returns ACF
1.0
0.010
0.8
0.008
0.6
0.006
y3^2
ACF
0.4
0.004
0.2
0.002
0.0
0.000
Time Lag
0.8
0.020
0.6
0.015
y4^2
ACF
0.4
0.010
0.2
0.005
0.0
0.000
Time Lag
58
SP500 Squared returns ACF
0.012
1.0
0.010
0.8
0.008
0.6
0.006
y5^2
ACF
0.4
0.004
0.2
0.002
0.0
0.000
Time Lag
0.8
0.010
0.008
0.6
y6^2
ACF
0.006
0.4
0.004
0.2
0.002
0.0
0.000
Time Lag
59
HANG SENG Squared returns ACF
0.006
1.0
0.005
0.8
0.004
0.6
0.003
y7^2
ACF
0.4
0.002
0.2
0.001
0.0
0.000
Time Lag
0.8
0.020
0.6
0.015
y8^2
ACF
0.4
0.010
0.2
0.005
0.0
0.000
Time Lag
60
TOPIX Squared returns ACF
0.008
1.0
0.8
0.006
0.6
0.004
y9^2
ACF
0.4
0.002
0.2
0.0
0.000
Time Lag
0.8
0.008
0.6
0.006
y10^2
ACF
0.4
0.004
0.2
0.002
0.0
0.000
Time Lag
As we can show, the last period is influenced by the Covid-19 and the
reaction of the time series were increase the volatility and then create a
61
volatility clustering. The period analysed show highest evidence in the
Covid period for many indices, excluding the HANG SENG index and the
and political events like Brexit, USA elections and so on. The volatility
of the HANG SENG is the opposite of the other indices showing almost
always high volatility with a brief time of low volatility, given graphical
evidence of volatility clustering. The TOPIX index has almost the same
situation. Instead the CHINA A50 index shows high volatility during the
first period and now, with the Covid crisis, is lower than before. The
FTSE MIB has reached highest values of volatility during the last period
due to Covid-19. In the right side, there are their respective correlogram
showing that the lags, in all cases, decrease obtaining the autocorrela-
tion of the squared residuals less than 0.05, the significance level. This is
positive for the model because all time series shows volatility clustering
ing the Engle’s ARCH test statistics. The test assesses the null hypoth-
the no ARCH effects null hypothesis in favour of the alternative. h=0 in-
62
dicates failure to reject the no ARCH effects null hypothesis. We have
used one lag. Then the output are pvalue, useful to consider the signif-
icance level, the test statistics, using a Lagrange multiplier test statistic
We immediately note all the time series rejects the null hypothesis in
that use the conditional variance. The time series statistics far exceeds
the critical values. The p-value is 0, offering an further confirm of the re-
63
3.7 The empirical EWMA
that the intercept ω is equal to zero, with the smoothing parameter (λ)
We can analyse every time series making a backtest and evaluate, using
the VaR, how many times the historical returns exceed the confidence
model and I can consider the model with the GARCH(1,1) and with the
EWMA model with the confidence interval of 95%, therefore the model
fits on the time series when the outliers are exceed the confidence in-
tervals for 5% of times. To compute the graph I used the package Stats
by R. The black line is the log daily time series, the dashed lines are the
calculated with the EWMA formula, using lambda euqal to 0.94 and the
constant equal to zero. The red points are the points beyond the confi-
dence intervals.
64
EWMA DAX30
0.10
0.05
Group Summary Statistics
UCL
0.00
LCL
−0.05
−0.10
1 42 90 145 207 269 331 393 455 517 579 641 703 765 827 889 951 1019 1094 1169 1244 1319 1394
Group
Number of groups = 1418 Smoothing parameter = 0.94
Center = 0.0002214613 Control limits at 2*sigma
StdDev = 0.01148198 No. of points beyond limits = 107
EWMA CAC40
0.05
UCL
Group Summary Statistics
0.00
LCL
−0.05
−0.10
1 42 90 145 207 269 331 393 455 517 579 641 703 765 827 889 951 1019 1094 1169 1244 1319 1394
Group
Number of groups = 1418 Smoothing parameter = 0.94
Center = 0.0001301072 Control limits at 2*sigma
StdDev = 0.01082107 No. of points beyond limits = 113
65
EWMA FTSE100
0.05
Group Summary Statistics
UCL
0.00
LCL
−0.05
−0.10
1 42 90 145 207 269 331 393 455 517 579 641 703 765 827 889 951 1019 1094 1169 1244 1319 1394
Group
Number of groups = 1418 Smoothing parameter = 0.94
Center = −4.685586e−05 Control limits at 2*sigma
StdDev = 0.009267366 No. of points beyond limits = 108
EWMA FTSEMIB
0.05
UCL
0.00
Group Summary Statistics
LCL
−0.05
−0.10
−0.15
1 42 90 145 207 269 331 393 455 517 579 641 703 765 827 889 951 1019 1094 1169 1244 1319 1394
Group
Number of groups = 1418 Smoothing parameter = 0.94
Center = 5.543939e−05 Control limits at 2*sigma
StdDev = 0.01317496 No. of points beyond limits = 97
66
EWMA SP500
0.10
0.05
Group Summary Statistics
UCL
0.00
LCL
−0.05
−0.10
1 42 90 145 207 269 331 393 455 517 579 641 703 765 827 889 951 1019 1094 1169 1244 1319 1394
Group
Number of groups = 1418 Smoothing parameter = 0.94
Center = 0.0002784529 Control limits at 2*sigma
StdDev = 0.009163892 No. of points beyond limits = 102
EWMA NASDAQ
0.10
0.05
Group Summary Statistics
UCL
0.00
LCL
−0.05
−0.10
1 42 90 145 207 269 331 393 455 517 579 641 703 765 827 889 951 1019 1094 1169 1244 1319 1394
Group
Number of groups = 1418 Smoothing parameter = 0.94
Center = 0.0006105464 Control limits at 2*sigma
StdDev = 0.01120885 No. of points beyond limits = 95
67
EWMA HANGSENG
0.06
0.04
0.02
UCL
Group Summary Statistics
0.00
−0.02
LCL
−0.04
−0.06
−0.08
1 42 90 145 207 269 331 393 455 517 579 641 703 765 827 889 951 1019 1094 1169 1244 1319 1394
Group
Number of groups = 1418 Smoothing parameter = 0.94
Center = 0.0001429119 Control limits at 2*sigma
StdDev = 0.01104393 No. of points beyond limits = 92
EWMA CHINA50
0.15
0.10
0.05
Group Summary Statistics
UCL
0.00
LCL
−0.05
−0.10
−0.15
1 42 90 145 207 269 331 393 455 517 579 641 703 765 827 889 951 1019 1094 1169 1244 1319 1394
Group
Number of groups = 1418 Smoothing parameter = 0.94
Center = 0.000421779 Control limits at 2*sigma
StdDev = 0.01547745 No. of points beyond limits = 104
68
EWMA TOPIX
0.05
Group Summary Statistics
UCL
0.00
LCL
−0.05
1 42 90 145 207 269 331 393 455 517 579 641 703 765 827 889 951 1019 1094 1169 1244 1319 1394
Group
Number of groups = 1418 Smoothing parameter = 0.94
Center = 0.0001687 Control limits at 2*sigma
StdDev = 0.01047247 No. of points beyond limits = 108
EWMA SPI100
0.05
UCL
Group Summary Statistics
0.00
LCL
−0.05
−0.10
1 42 90 145 207 269 331 393 455 517 579 641 703 765 827 889 951 1019 1094 1169 1244 1319 1394
Group
Number of groups = 1418 Smoothing parameter = 0.94
Center = −2.176164e−05 Control limits at 2*sigma
StdDev = 0.00935759 No. of points beyond limits = 90
69
As I already explained, the EWMA model is a simple model because
daily returns. As we can see, in the graphs there are many points that go
beyond the confidence intervals. For example, just the DAX 30 shows 18
points on the range of the last year. Every graph shows the unconditional
standard deviation and the number of points beyond limits. in the DAX
30 index, the number of points beyond limits are 107, having 1418 ob-
servations means that the time series exceeds the 7.5% of times under
evaluating the risk. The others indices shows the same thing, therefore
ity they give us the information of the wrong model. The China 50, for
and identical for all assets. This implies that it is not optimal for any
asset in the sense that the GARCH models discussed above are optimal.
GARCH models, even though the difference can be very small in many
cases.
70
3.8 The structure of the GARCH VIX model
the variance swaps, explained above. Our approach is based on the fact
that VIX approximates the 30-day variance swap rate on the S&P 500.
lihood Estimation (MLE) method using return time series. The goal of
the model is use information of the VIX index to estimate GARCH mod-
the parameters ω, α, β and δ, useful to evaluate how the model fit. First
of all, I explain the MLE and I estimate the parameters considering the
22
See, Kanniainen, Binghuan & Yang (2014).[21]
71
3.9 The Maximum Likelihood
nary least squared (OLS), useful in linear models. Bollerslev and Wooldridge
estimator. However, QML is not efficient unless the true density actually
is normal.
following way:
n
Y
f (Yn |θ ) = f (Yi |θ )
i =1
likelihood function.
The theta of the model GARCHVIX are omega, alpha, beta and delta
72
θ = [ω; α; β ; δ]
normally distributed:
Yt = σt εt
εt ∼ N (0, 1).
p
X q
X s
X
σ2t =ω+ αi Yt 2−i + β j σ2t − j + δs V I X t2−s
i =1 j =1 s =1
T T T
Y T −1 1X 2 1 X Yt 2
f Yi (Yi , θ ) = − l o g (2π) − l o g (σt ) −
i =1
2 2 t =2 2 t =2 σ2t
T T p q s
Y T −1 1X X
2
X
2
X
f Yi (Yi , θ ) = − l o g (2π)− l o g (ω+ αi Yt −i + β j σt − j + δs V I X t2−s )...
i =1
2 2 t =2 i =1 j =1 s =1
73
T
1X Yt 2
... −
2 t =2 ω + pi=1 αi Yt 2−i + qj=1 β j σ2t − j + ss =1 δs V I X t2−s
P P P
tion using the MLE. The parameter estimates are obtained by numer-
3.10 Estimation
variance. The output of the function fmincon give me the value of the
the inverse of the hessian and I calculated the variance. Then I extrap-
olated the standard error of every parameter using the square root and
74
T
Y T −1 1X
f Yi (yi , θ ) = − l o g (2π)− l o g (ω+α1 Yt 2−1 +β1 σ2t −1 +δ1 V I X t2−1 )−...
i =1
2 2 t
1X Yt 2
... −
2 t ω + α1 Yt 2−1 + β1 σ2t −1 + δ1 V I X t2−1
of the model, the management of the data and the constraints I esti-
ance with time t using the MLE to estimate the parameters of the histori-
adding the historical returns of the VIX. The parameters are estimated
for every index and below the estimate of every parameter there is the
75
The output by MATLAB of the fmincon function are:
the parameter there is the test statistics. A test statistics contains infor-
mation about the data that is relevant for deciding whether to reject the
null hypothesis. As we can see we have the parameters of the S&P 500
and the NASDAQ where there is strong significance levels to reject the
two indices is very high and the result follows the criteria that to analyze
76
one of the other give me the same result. The others indices does not
reject the δ parameter give us the information that the model fit better
with the δ parameter. This is not true in the case of β parameter where
standard methods such as the t-test to see whether the parameters are
Likelihood Ratio test (LR test) that is used when one model nests inside
criterion (AIC) and the Bayesan information criterion (BIC), the infor-
mation criteria test to select the model with the lowest value.
77
Table 7: Diagnostics table
(1,1,1)
The table 7 shows the LR test and the information criteria. We can
see that the LR test confirm a better model when we have a GARCHVIX
model rather than the GARCH model in every index with two exception,
the S%P 500 and the NASDAQ. This is what I determined before with
the test statistics using the maximum likelihood. The results does not
change and the AIC and BIC give us the same result.
78
3.12 Backtesting and VaR
occurring within a specified time and with a give probability. The valid-
various risk models. It aims to take ex ante Value at risk (VaR) forecasts
from a particular model and compare them with ex post realized return
is said to have occurred. Models that do not perform well during back-
tios, therefore every value that go beyond limits of 2∗σt for a GARCH and
for a GARCHVIX model. We show the table 8 where the percentile should
the graphs on the GARCH model on the left and the GARCHVIX model
on the right.
79
Table 8: Backtesting
80
81
Figure 6: Conditional variance
82
Unfortunately, the GARCHVIX model is better only in some cases an-
alyzing the time series. I have painted in yellow the best results and we
do not have a single solution for every time series but depends on the
time series. The results here, therefore, suggest that including VIX in the
ket risk and therefore the violation ratio we can consider the Expected
Shortfall, more sensitive to the shape of the tail of the loss distribution.
mine the best model. In most of the cases the GARCHVIX model is better
83
than the GARCH model, considering that sometimes underperform or
84
4 Concluding remarks
ologies focusing mainly in the GARCH model implementing the VIX In-
that explain the time series. Within on this research the author works
certainty. In the first chapter we studied the evaluation of the VIX, how it
is computed and how is changed from the creation of the index to now.
the structure of the VIX as a variance swaps. We confirmed that the new
model model is more accurate than the previous one using the forward-
looking plain vanilla options of the Standard & Poor’s 500 index. The
useful for the accuracy of the model. Not only with the GARCH model we
compared our model but also with the EWMA model, considering a 5%
85
the author examined the major indices using the descriptive statistics
and some relative tests useful to determine and fix the time series. After
being ensured that the volatility clustering on time series and therefore
that the constructed model is useful in the real world, the author esti-
model was created and the conditional variance estimated. Then, the
model compared with the GARCH and the EWMA investigating the ac-
useful to determine, with the Value at Risk and the expected shortfall,
which one is the best model and be the more accurate one. However,
not all assets that we considered are characterized by the same trend and
thus the model perform differently for every time series, Hence, apply-
ing the model, we notice these differences and we cannot confirm if one
model is better than the other one. Summarizing, both the GARCH (1,1)
and the GARCH VIX (1,1,1) model are useful and accurate to the time
interesting citation by George Box that said “All models are wrong but
some are useful” explains how we should fit the model on the time series
86
and analyse the forecasts of the conditional variance using ARIMA mod-
87
5 Bibliography
References
[1] Ruey S. Tsay (2010). Analysis of Financial Time Series (Third Edi-
tion). The University of Chicago Booth School of Business Chicago,
IL. 109-140
[2] Vega Ezpeleta. (2015). Modeling volatility for the Swedish stock
market. Uppsala University, Department of Statistics, Advisor: Lars
Forsberg. 4-20
[3] Andrea Pastore (2018). Statistical methods for Risk Analysis. Ca’
Foscari University of Venice. 20-25
[4] Ruppert D., Matteson D.(2015). Statistics and Data Analysis for Fi-
nancial Engineering, Second Edition 2015. 45-70. 307-451
88
[12] Jim Gatheral (2006). The Volatility Surface: A Practitioner’s Guide.
Wiley
[15] CBOE VIX (2019). White Paper. Cboe Volatility Index. Cboe Ex-
change Inc. 1-20
[16] John C. Hull. (2018). Options, Futures and other Derivatives. cap.15
cap. 26 ; Technical Note No. 22* Valuation of a Variance Swap
[17] Hao Zhou and Matthew Chesnes. (2003). Vix Index Becomes Model
Free and Based on S&P 500, Board of Governors of the Federal Re-
serve System, division of research and statistics
[19] Hao Zhou and Matthew Chesnes. (2003). Vix Index Becomes Model
Free and Based on S&P 500, Board of Governors of the Federal Re-
serve System, division of research and statistics
89
[24] McAleer (2007). Volatility Index (VIX) and S&P 100 Volatility Index
(VXO) School of Economics and Commerce University of Western
Australia and Faculty of Economics Chiang Mai University
90
List of Figures
List of Tables
1 Jarque-Bera test . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 48
2 Descriptive statistics and Ljung-box test . . . . . . . . . . . . . 53
3 Correlation Matrix . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 55
4 Correlation with the VIX . . . . . . . . . . . . . . . . . . . . . . . . . 56
5 Engle’s ARCH test . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 63
6 Estimated parameters with the maximum likelihood . . . . 76
7 Diagnostics table . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 78
8 Backtesting . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 80
9 Expected Shortfall . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 83
91
A Appendix
A.1 A Appendix
A Appendix: tests
Jarque-Bera test
In statistics the Jarque-Bera test is a test where the sample data have
the skewness and kurtosis that match a normal distribution. The test
statistics give always a positive result. The skewness of a normal distri-
bution is 0 and the kurtosis is 3, when is far from zero, it means that the
data do not have a normal distribution.
The test statistic JB is defined as:
n 2 1
JB = (S + (K − 3)2 ) (50)
6 4
where n is the number of observations (or degrees of freedom in gen-
eral); S is the sample skewness and K is the sample kurtosis. If the data
comes from a normal distribution, the Jarque Bera statistics has a chi-
squared distribution with 2 degrees of freedom, so the statistics can be
used to test the hypothesis that the data are from a normal distribution.
The null hypothesis is a joint hypothesis of the skewness being zero and
the excess kurtosis being zero. Samples from a normal distribution have
an expected skewness of 0 and an expected excess kurtosis of 0 (which is
the same as a kurtosis of 3). As the definition of JB shows, any deviation
from this increases the JB statistic.
Ljung-Box test
The Ljung-Box test is a type of statistical test of whether any of a group of
autocorrelations of a time series are different from zero. Instead of test-
ing randomness at each distinct lag, it tests the "overall" randomness
based on a number of lags, and is therefore a portmanteau test. The test
statistic is:
h
X ρk2
Q = n(n + 2) (51)
k =1
n − k
where n is the sample size, ρk is the sample autocorrelation at lag
k, and h is the number of lags being tested. Under H0 the statistic Q
asymptotically follows a χh2 . For significance level a, the critical region
92
for rejection of the hypothesis of randomness is:
2
Q > χ1−α,h (52)
where χ1−α,h
2
is the 1-a-quantile of the chi-squared distribution with h
degrees of freedom.
93
A.2 B Appendix
B Appendix: table
94
95
96