Professional Documents
Culture Documents
Co-supervisor :
Prof. Dr. GREGOR WEIß
Faculty of Economics and Management
Chair in Sustainable Finance and Banking
The dissertation entitled “Model risk of Copula-GARCH financial time series mod-
els ”, by Supriya Murdia is approved for the degree of Master of Science in Mathematics.
Prof. D. C. Joshi
(External Examiner)
Prof. V. H. Pradhan
(Internal Examiner)
Date :
Place : Surat
3
DECLARATION
I hereby declare that the dissertation titled “Model risk of Copula-GARCH finan-
cial time series models” is a genuine record of research work carried out by me and no
part of this dissertation has been submitted to any university or institution for the award of
any degree or diploma.
Supriya Murdia
(I14MA013)
Applied Mathematics and Humanities Department
Sardar Vallabhbhai National Institute of Technology
Surat - 395007
Date:25-04-2019
Place: Surat
4
CERTIFICATE
This is to certify that the dissertation titled, “Model risk of Copula-GARCH financial
time series models ” is based on a part of research work carried out by Miss. Supriya
Murdia under my guidance and supervision at Sardar Vallabhbhai National Institute of
Technology, Surat and Faculty of Economics and Management, Universitat Leipzig, Ger-
many.
Supervisor:
Dr.Jayesh Dhodiya
Associate Professor
Department of Applied Mathematics and Humanities
Sardar Vallabhbhai National Institute of Technology
Surat, India.
5
6
ACKNOWLEDGEMENT
I would like to extend my heartfelt gratitude to Dr. Jayesh Dhodiya, for understanding my
fondness towards this topic, for supporting me and mentoring me, for making any resource
that I would require available at the speed of light and for bringing me out of gloom when I
thought I would fail to implement the factor copula. I would thank him for having us at the
lab throughout the semester, the only reason why we were able to complete our work and
make memories to last for a lifetime.
This dissertation has been one stupendous journey and I am grateful for every part of
it. I would like to thank Prof. Dr. Gregor Weiß for his relentless support and guidance
throughout the work, and for putting up with me for my doubts with the directions to the
study, with the theory even when they turned out to be naive, often. I would specially thank
him for taking out the time and the efforts, even against the unsteady network connection
to still motivate and guide me on the project. I hope I am much less trouble in future with
exploring this research question in full, and effectuating it into a worthwhile publication.
Most importantly, I would like to thank my mother for the selfless, undying attention,
support and love she fills me up with, to the brim and more. I would also like to thank
Surbhi Ma’am, Ankita Ma’am, Ankit Sir for helping me with software installations and
latex specifications and my batch mates Nagesh, Deevanshu and Saloni for lending me their
laptops for my simulations and keeping me updated on the semester proceedings.
7
PREFACE
2 Literature Review 18
2.1 Univariate time series modelling . . . . . . . . . . . . . . . . . . . . . . . . . 18
2.1.1 Mean modelling . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 18
2.1.2 Volatility modelling . . . . . . . . . . . . . . . . . . . . . . . . . . . . 19
2.1.3 ARMA-GARCH modelling . . . . . . . . . . . . . . . . . . . . . . . . 19
2.2 Multivariate time series modelling . . . . . . . . . . . . . . . . . . . . . . . . 19
2.2.1 Copulas . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 20
2.3 Model Averaging . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 21
2.3.1 Loss Functions . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 22
2.4 Similar Work . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 22
2.5 Objective of research . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 22
1
TABLE OF CONTENTS 2
7 Appendix 83
List of Figures
3
LIST OF FIGURES 4
5.1 Backtest results of models with rolling window for the entire duration: 2003-2013. 42
5.2 MCS results for models with rolling window for the entire duration: 2003-2013. . . 42
5.3 MCS results for the period of financial crisis: December,2008-March, 2012 for mod-
els with rolling window. . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 43
5.4 MCS results for the entire period : 2003-2013 for models with rolling window. . . 43
5.5 MCS results for the entire period: 2003-2013 for models with rolling window. . . . 44
5.6 MCS results for the period of financial crisis: Dec, 2008-March, 2012 for models
with rolling window. . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 44
5.7 Sener’s penalties for 1% VaRs of copula-GARCH models with rolling window through
2003-2013. . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 45
5.8 MCS results for the entire duration: 2003-2006 including DCC-GARCH(1,1) for
models with rolling window. . . . . . . . . . . . . . . . . . . . . . . . . . . . . 46
5.9 MCS results for the entire duration: 2003-2006 including the non copula DCC-
GARCH(1,1) model for models with rolling window. . . . . . . . . . . . . . . . . 46
5.10 Exceedances of models without rolling window for the entire duration: 2003-2013. 58
5.11 Backtest results of models without rolling window for the entire duration: 2003-2013. 59
5.12 MCS results for the entire duration: 2003-2013 for models without rolling window. 59
5.13 MCS results for the entire time for models without rolling window. . . . . . . . . 60
5.14 MCS results for the entire period : 2003-2013 for models without rolling window. . 60
5.15 MCS results for the entire duration: 2003-2013 for models without rolling window. 61
5.16 MCS results for the period of financial crisis: Dec, 2008-March, 2012 for models
without rolling window. . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 61
5.17 Sener’s penalties for 1% VaRs of copula-GARCH models with rolling window through
2003-2013. . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 62
5.18 Sener’s penalties for 1% VaRs of copula-eGARCH models with rolling window dur-
ing 2003. . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 70
5
List of abbreviations
Abbreviation Denotes
MA Moving Average
AR Auto-Regression
ARMA Auto-Regressive Moving Average
ARIMA Auto-Regressive Integrated Moving Average
ARFIMA Auto-Regressive Fractionally Integrated
Moving Average
ARCH Auto-Regressive Conditional Heteroskedasticity
GARCH Generalised Auto-Regressive Conditional
Heteroskedasticity
N (0, 1) Standard Normal Distribution
VARMA Vector Auto-Regressive Moving Average
CCC Constant Conditional Correlation
DCC Dynamic Conditional Correlation
BEKK Baba, Engle, Kraft and Kroner model
VaR Value-at-Risk
MCS Model Confidence Set
Chapter 1
• Models in which the present value of time series data is related to the prior values of
time series data (mean and volatility) and past prediction errors which contribute to
the white noise in prediction. For instance, the ARIFMA model.
7
1.2. TIME SERIES MODELS 8
When modelling time series we have to take care of the following in the data:
• Trends
• Seasonal variation
• Heteroskedasticity, if present
• Outliers
AR(p) model
Here AR stands for auto-regression. Auto-regression is nothing but regression on the
dependent variable’s own past values. [9] It is a form of multi-variable regression where the
independent variables are the series’ own past values, p determining the number of such,
lagged terms and the dependent variable is the present day value of the series. A general
AR(p) process is of the following nature:
p
X
Xt = c + ψi Xt−i + t . (1.2.1)
i=0
Here, the error terms t ∼ N (0, σ 2 ), where t is independent and identically distributed,
ψi are the target parameters that need to be estimated, c is a constant and Xt are the data
values to be modelled/ estimated.
MA(q) model
A q th order moving average model, denoted by M A(q) is of the form:
q
X
Xt = µ + θi wt−i , (1.2.2)
i=0
where wt ∼ independent and identically distributed N (0, 1), , θi are the parameters that
need to be estimated, and t−i are the white noise error terms and µ is the mean of the
series. Here q is the order of the moving average model, i.e., the number of lagged terms
considered in the model. Xt is the return or the time series value being modelled.
an AR(auto regression) process of order p, and the latter is for an MA (moving average)
process part of order q
p q
X X
Xt = c + ψi Xt−i + θi t−i , (1.2.3)
i=0 i=0
where, the notations retain their former meaning. It was first described in 1951 in the thesis
of Peter Whittle.
ARIMA(p,d,q) model
ARIMA, as in, auto-regressive integrated moving average model is nothing but a
generalisation of ARMA(p,q) that has the capability to concern with non-stationarity. The
model is determined as given below:
p q
X X
i d i
1− φi L (1 − L) Xt = 1 + θi L t , (1.2.4)
i=1 i=1
where, θi s are the moving average parameters, L is lag operator, Xt is time-series data and
t is the corresponding time-index, αi s are the autoregressive parameters used to derive φi s
as gievn below:
p+d p
X X
i
1− αi L = 1 − i
φi L (1 − L)d . (1.2.5)
i=1 i=1
Here d denotes multiplicity of a unit root (L = 1) of the polynomial of on the left hand side
of the equation above. Hence, an ARIMA process with d > 0 is not wide-sense stationary.
Here α is the smoothing constant and t is the time instant being observed.
ARCH(q) model
Heteroskedasticity: Consider a first sequence of random variables, {Yt }nt=1 , and another
sequence of vectors of random variables, {Ŷt }nt=1 which is a regression fitting estimate of
the original sequence of random variables. If Yt varies unequally compared to Ŷt over the
range of time, even for some t, then such a circumstance is called heteroskedasticity. This
is why ordinary least squares method and other regression tools lack efficiency when dealing
heteroskedastic data.
1.2. TIME SERIES MODELS 10
t = σt zt , (1.2.7)
q
X
σt2 = α0 + α1 2t−1 + ... + αq 2t−q = α0 + αi 2t−i , (1.2.8)
i=1
where α0 > 0 and αi ≥ 0, i > 0, q(≥ 0) denotes the number of the lagged innovations in the
model and the zt are independent and identically distributed N (0, 1). An ARCH(q) model
can be estimated via ordinary least squares method.
GARCH model
A GARCH(p,q) model is enhanced into the following by Bollerslev [16]:
yt = x0t b + t , (1.2.9)
where, t denotes a real-valued stochastic discrete time process, ψt denotes the σ-field of
all information throughout the time range, p(≥ 0) is the number of lagged residual terms,
and βj (≥ 0, j = 1, . . . , p), the corresponding coefficients to be estimated. Evidently, it is a
generalisation of ARCH(q) model with an additional lag on the conditional variances being
estimated themselves.
where γj ∈ (−1, 1), j = 1, . . . , p, σt2 denotes the conditional variance at instant t, and at
are the white noise terms.
1.2. TIME SERIES MODELS 11
where the conditional variance is given by σt2 , g(Zt ) = θZt + λ(|Zt | − E(|ZT |)), ω, β, α, θ, γ
are coefficients, Zt follows the generalised error distribution.
Threshold GARCH also known as tGARCH(p,q) is given by
− −
σt = K + δσt−1 + α1+ +
t−1 + α1 t−1 , (1.2.12)
−
where + +
1 = t−1 if t−1 > 0, and 1 = 0 if t−1 ≤ 0 and likewise for t−1
σt2 = K + δσt−1
2
+ α2t−1 + φ2t−1 It−1 , (1.2.13)
where It−1 = 0, if t−1 ≥ 0 and It−1 = 1, if t−1 < 0 , which helps measure asymmetry in
ARCH models.
The symmetric GARCH(p,q) or the sGARCH(p,q) is useful and relatively time saving
in modelling symmetric univariate GARCH marginals.
VARMA-GARCH
The VARMA-GARCH is simply the multivariate extension of the GARCH model in which
instead of auto-regressing on data values, it is performed on past data vectors (multi-auto-
regression).
BEKK model
Engle and Kroner(1995) discuss the following multivariate GARCH specification :
q
K X p
K X
X X
Ht = C0 C0T + ATki t−i Tt−i Aki + T
Bkj Ht−j Bkj , (1.2.14)
k=1 i=1 k=1 j=1
1.2. TIME SERIES MODELS 12
where, Aki and Bki are d × d-dimensional parameter matrices, C0 is lower-triangular matrix.
To simplify understanding, we can first take a look at the bivariate case, taking K = p = 1
and q = 0. The conditional variance ε1t can be given by the following:
h12,t = c211 + a211 ε21t + a212 ε22t + 2a11 a12 ε1t ε2t , (1.2.15)
where as the conditional variance is given by
h11,t = c11 c21 + a11 a21 ε22t + (a12 a21 + a11 + a22 )ε1t ε2t . (1.2.16)
BEKK stands for Baba, Engle, Kraft and Kroner (1990) on the names of those who first
proposed it. The BEKK specification consolidates that Ht is positive definite under weak
assumptions. However, the sufficient condition for positivity requires that at least one of C0
or Bki have full rank matrices as well as H9 , . . . , H1−p should be positive definite matrices.
A beneficial point of BEKK is that it permits casualties in variances to be modelled, as it
can handle the dependency of conditional variances on the lagged values of the other [12].
where p and q can be estimated by information criterion methods. Qi,j t is the correlation
between rti and rtj , νt are the standardised residuals given by νt = Dt−1 (rt − µ), µ i the vector
of expected returns, µ i the vector of expected returns, rt is the vector of returns,PDt , the
diagonal matrix of conditional volatilities, and the matrix R is defined as R = T1 Tt=1 νt νt0
[30].
1.2. TIME SERIES MODELS 13
Copula-GARCH models
The univariate GARCH marginals are coupled with copulas to obtain a multivariate model
of the diversified portfolio.
A Copula, derived from the theory of t-norms from metric spaces, is a function that
possesses the following properties:
where u, v ∈ [0, 1]
It is a subclass of t-norms, as in metric spaces, constricted to the unit hypercube. This
function gives the cumulative joint probability distribution of the random variables involved.
Archimedean Copulas
Let ψ : [0, 1] → [0, ∞) be a strictly decreasing, continuous and convex function such that
ψ(1) = 0, ψ(0) ≤ ∞ and let ψ [−1] be the dynamics in the observed correlation. pseudo-
inverse of the above defined by: Nelson, (2006) showed that the following defines a family of
d-dimensional Archimedean copulas:
for every d ≥ 2. ψ is known as the generator of the copula, additive for Archimedean copulas.
The copula is Archimedean only when the pseudo-inverse ψ [−1] of the generator, ψ of the
copula is absolutely monotonic on R+ . In other words, we can say ψ [−1] ∈ L∞ with
The corresponding cumulative densities for the Gumbel and Clayton copulas are given by:
1
θ +(−log(u ))θ +···+(−log(u ))θ ) θ
CθGumbel (u) = e−((−log(u1 )) 2 d
, (1.2.24)
for θ ∈ [1, ∞)
1.2. TIME SERIES MODELS 14
1
CθClayton (u) = [max{u−θ −θ −θ
1 + u2 + . . . + ud − 1; 0}] ,
θ (1.2.25)
for θ ∈ [−1, ∞)\{0}, respectively. (refer [10])
The Frank copula is also an Archimedean copula. It’s density function is given by :
(e−θu1 − 1)(e−θu2 − 1) . . . (e−θud − 1)
Fr 1
Cθ (u) = − 1 + , (1.2.26)
θ e−θ − 1
with θ ∈ R\{0} and generator:
e−θt − 1
ψ F rank (t) = −log( ). (1.2.27)
e−θ − 1
Joe copula density is given by :
d
1
X
CθJoe (1 − ( (1 − ui )θ ) θ ), (1.2.28)
i=1
Copulas can also be elliptical in nature. Gaussian copula and Student t-copula
are two most popular and commonly used elliptical copulas. They are distributions over
the unit hypercube [0, 1]d with multiplicative associativity. Given a parameter correlation
matrix A ∈ [−1, 1]d×d , an elliptical copula has the following structure:
Vine Copula
The marginal density can only be estimated unless the true underlying data generat-
ing process is known. The process of estimating the multivariate density is by step-wise
decomposing the joint probability distribution into a factorisation of conditional and uni-
variate marginal. The conditional densities at the corresponding steps can be substituted
with suitable copulas. This method of obtaining the multivariate density was introduced by
Joe [50], who called it a Vine Copula. Let f denote the joint probability density function,
F denote the cumulative probability density function of the original vector. Let the joint
density function of X = (X1 , X2 , . . . , Xd )0 be given by:
Here c12...d determines the joint density function obtained from the copula description (i.e.
the f (xd |xd−1 )) which is nothing but the pair copula for the first two variables is given by:
c12 (F1 (x1 ), F2 (x2 )) · f1 (x1 )
Continuing in this fashion:
f (xd−2 |xd−1 ) = c(d−2)(d−1) (Fd−1 (xd−2 ), Fd−1 (xd−1 )) · fd−2 (xd−2 ), (1.2.33)
=⇒ f (xd−2 |xd−1 , xd ) = c(d−2)d|d−1 (Fd−2|d−1 (xd−2 |xd−1 ), Fd|(d−1) (xd , xd−1 )),
·c(d−2)(d−1) (Fd−2 (xd−2 ), Fd−1 (xd−1 )) · fd−2 (xd−2 ). (1.2.34)
However, this is not unique as the condition can be put on xd instead of xd−1 . The general
formula can be hence, given by:
where v denotes a d-dimensional vector. vj denote the components v and v−j denotes the
components of v without including vj .
In this way, a d-dimensional conditional density can be decomposed into a conditional
d − 1-dimensional density and a pair copula. We can hence obtain decomposition constructs.
They can be sorted by regular vines resulting in a graphical model presenting the entire de-
composition structure. A D-Vine decomposition for the joint density function, f (x1 , . . . , xd )
is given as follows:
d d−j
d−1 Y
Y Y
f (xk ) ci,i+j|i+1,...,i+j−1 (F (xi |xi+1 , . . . , xi+j−1 ), F (xi+j |xi+1 , . . . xi+j−1 )). (1.2.36)
k=1 j=1 i=1
[X1 , . . . , Xn ]0 ≡ X = BZ + ε, (1.2.39)
where,
0
QT,S ≡ gT,S (θ)Ŵ gT,S (θ),
gT,S (θ) ≡ m̂T − m̃S (θ).
ŴT represents a positive definite matrix of coefficient weights. Oh and Patton [63] proved
that as T → ∞, if S/T → ∞ , under regulatory conditions, the simulation method of mo-
ments’ estimator is asymptotically normal, consistent. The simulation method of moments
is abbreviated as the SMM estimator.
√ d
T (θ̂T,S − θ0 ) →
− N (0, Ω0 ) as T, S → ∞. (1.2.42)
Here,
Mixture copula
Very often, multiple copulas can encompass the unit space of probabilities better than a
parametric class of one copula. For this purpose, we use a weighted mixture of few common,
flexible copulas in our study. They first appeared in literature in the works of NUMBER
[62], (1992). A sophisticated extension to this would be to use the popular Expectation-
Maximisation algorithm to estimate the weights corresponding to the mixture of copulas
according to the data.
Literature Review
We will now literature review that motivated our work. To approximate the distribution of
a single time series panel, we need to opt univariate time series. We begin with discussing
the vast literature on univariate time series modelling.
18
2.2. MULTIVARIATE TIME SERIES MODELLING 19
distributions and describe a dependence structure appropriately. Patton [65], (2009) showed
that the amount of parameters modelled is manageable and the univariate conditional distri-
bution need not describe the multivariate distribution. However, it can be one, if that’s the
best fit anyway. This flexibility about copulas is undoubtedly the best thing about them.
Liu and Luger [57], (2009) determined how these copula-based models allow maximum like-
lihood to be estimated using the entire data instead of just the overlapping part between
elements relating only to marginals and elements relating to multivariate distribution . A
study by Weiß [71], (2013) illustrates that the copula-GARCH models do have a prediction
performance superiority over the DCC-GARCH model.
Copula models have been gaining immense popularity since the past decade, especially
in field of financial forecasting. Studies, for example, Minović [60] have shown copula models
to outperform several less complex (non-copula) models, and ones that do not take the
inter-dependency of variables into account.
2.2.1 Copulas
The term copula was coined by Sklar (1959), extending its primary definition from the the-
ory of metric spaces, a subclass of t-norms. The first works on this were more rudimentary,
based on the consolidation of the structure of copulas, for instance by Wassily Hoeffding
and Maurice Fretchet. Application of copulas to joint-life mortality was indirectly initiated
by the works of Clayton (1978). Oakes (1982), Cook and Johnson (1981) further extended
their study in this field. Copula literature in probability has been studied since 40 years
(Schweizer, 1991). The books by Joe (1997) and Nelson (1999) were exceptionally promi-
nent in imparting fundamental structural knowledge of copulas as a tool. The latter was in
fact my first book on copulas itself! The baby steps in finance and Actuarial science (see
Bowers et al, 1997) with this tool were laid in the books by Frees and Valdez (1998) and
Embrechts, McNeil and Straumann (1999) focussing on modeling using copula structures.
Copulas have also been used in biostatistics, for instance David Oakes (1989), and Houg-
gard (2001). Works in survival analysis by Jason Fine, Phillipe Lambertt, Joana Shih still
stand out. Profound contributions regarding the study of copulas in statistics have been
made by the Quebec group (Bruno Remillard, Louis-Pal Rivest, Philippe Caperaa, Belka-
cem Abdous, etc). Drauet-Mari and Kotz (2001) focused on correlation and dependence,
and the use of copulas as the corresponding measure. Research literature in copulas dates
back to the early 2000s, the first few pedagological works published during 2002-2004 by
Thierry Roncalli and his associates. Embrechts, McNeil and Straumann (1999) and Li(2000)
served as effective catalysts in developments in this field. As under risk measurement, an
early example of VaR studies applying copulas is observed in Embrechts, Hoing and Juri
(2003) , an area that’s received widespread recognition in the following years. Patton (2004)
proposed bivariate equity management using copulas stimulating its application in portfolio
management. Cherubini, Luciano and Vecciato (2004) have co-authored a remarkable book -
Copula Methods in Finance. Copulas in finance find their application in several areas, a chief
one being risk management. The book on extreme financial risks authored by Malevergne
and Sornette (2006) is a commendable contribution to the literature of copulas in finance.
As in Mikosch (2006), the amounts of Google hits for copulas struck 650,000 from an
initial figure of 10,000, between the years 2003 to 2005. And were in 2019 now! The
popularity of copulas is self-explanatory. C. Genest et al (2009) presented an elaborate
2.3. MODEL AVERAGING 21
methods that give certain absolute outcomes. This is model confidence set approach. Ab-
breviated as MCS, it was introduced by Hansen et al. [46], (2010) As we actually implement
this mechanism, we would discuss the relevant theory and algorithm in Chapter 3.
For the MCS procedure to be able to compare models, we need to feed it with a relative
performance of the two models. This is obtained by obtaining a loss distribution of forecasts
from the model.
1. There have been several studies on comparing multivariate models for financial time-
series data. There have also been studies comparing multivariate Copula-GARCH
methods to model series. Our aim is to use Copula-GARCH models for a diverse
portfolio and compare them using Hansen’s model confidence set approach.
2. The Generalised autoregressive score factor copula, proposed by Patton and Oh (2010)
is a breakthrough model when it comes to large dimension. Our objective is to im-
plement Factor copula in R.
3. We apply Christoffersen’s backtest for conditional coverage, Kupeic’s backtest for un-
conditional coverage and Engle and Manganelli’s backtest. We compare these results
2.5. OBJECTIVE OF RESEARCH 23
with the MCS results and qualitatively analyse them aganist the time consumed in
forecasting from the models.
4. With this study we intend to arrive at economically significant models, which are less
parsimonious and time consuming and almost as efficient as the intricately sophisti-
cated ones. For this reason and to also capture effective profitability from a model,
we apply Sener’s violation space penalty to better rank the models based on their
predictive efficiency.
5. Our study also extends to study a less explored dimension of financial time series
modelling: the effect of marginal choice on the performance of the model. We
first answer which marginal models are superior based on the MCS comparison.
6. We also intend to answer the question: How harmful is an inferior marginal for
the overall model?
Chapter 3
The methodology we implement is the crux of our study as similar studies have been per-
formed but the application of copula-GARCH models to study the nature of the financial
time series and their application to figure out the statistically best and the economically
best models using Hansen’s model confidence set procedure Hansen et al. [45] and also apply
Sener’s penalty measure. The structure of the methodology implemented in our work is as
follows:
1. Obtain the dataset to model.
2. Fit the marginals of the dataset. We will fit multiple marginals to an index in order
to make comparisons on the basis of marginals as well.
3. Identify and fit the dependence structure using copula and non-copula models
4. Simulate from the fitted structure. We follow two processes of simulation.
The first mechanism employs simple simulation from the fitted structure to forecast
‘k’days in future (traditional out-of-sample forecasts).
We employ a second mechanism in which we take an estimation window (of say,
500 days) and predict one day in future. The estimation window is then updated to
contain the latest true value of the day for which the forecast was performed, and
one oldest value is removed, so that the window size remains constant and there is no
baseline-wandering like effect (or the long term memory effect).
5. We obtain the 5%, 1% and 0.1% quantiles or technically, the Value-at-Risk forecasts.
6. The loss associated with each model is obtained by applying the Asymmetric VaR
Loss function [40] and Continuous Rank Probability Score function [39] on
our VaR forecasts. These serve as measures of the predictive performance of a model.
7. On these loss distributions, we apply Hansen’s Model Confidence Set procedure [45] to
construct a superior set of models, for a given confidence interval.
8. In the end, we construct tables for the time taken by each model for the computations,
and its rank as obtained by the MCS algorithm to qualitatively analyse our results.
We extract the data, with the following description.
24
3.1. STOCK INDICES DATA 25
1. Estimate AR(q) model that fits the underlying data the best:
q
X
yt = α0 + αi yt−i + t , (3.2.1)
i=1
where, the notations retain their meaning from eqn. (1.2.9). This can be done using
ordinary least squares method or by method of moments (using Yule-Walker equations)
[8].
√
3. ρ(i)’s asymptotic standard deviation is given by 1/ T Individual values greater than
this are GARCH errors. The total number of lags can be estimated by the Ljung-
Box test [21], (1970) until say, 10% significance is achieved.If the squared residuals, 2t
are uncorrelated, the Ljung-Box Q-statistic follows χ2 distribution with n degrees of
freedom.
3.3. COPULA ESTIMATION 26
T
X 1 1 1 u2t
= (− ln(2π) − ln(σt2 ) − 2
). (3.2.4)
t=1
2 2 2 σt
1. Let the copula density be c(u1 , . . . , ud ; θ ) = ∂u∂ 1 · · · ∂u∂ d C(u1 , . . . , ud ; θ ). Here θ denotes
the parameters that need to estimated corresponding to copula, C. The log-likelihood
function can then be constructed as follows:
n
X
l(θ; Û1 , . . . , Ûn ) = log(c(Ûi,1 , . . . , Ûi,d ; θ)). (3.3.2)
i=1
∂C(u1 , . . . , ud |θ)
c(u1 , . . . , ud |θ) = , u1 , . . . , ud ∈ [0, 1]. 3.3.3
∂u1 · · · ∂ud
2. The Maximum likelihood (ML) estimator was constructed by Genest et al. [36] as
follows:
θ̂nM L (U) ≡ arg max lU (θ). (3.3.4)
θ∈Θ
3. In this manner, likelihood of all candidate copula models can be constructed. Following
the log-likelihood estimation method to fit the copula, an optimal copula, i.e. the best
fit to the underlying data is the one with the lowest log-likelihood value.
We summarise the above methodology in the following flowcharts: Let rt be the realised
returns, LL be a vector of log-likelihoods to be computed, µ, φ be parameters of the ARMA
process, αi , βi are parameters that need to be estimated for the GARCH process and B, the
number of simulations to be performed.Let εt denote the residuals (errors), σt , the standard
deviation and T be the entire range (time window) for which the estimation is being done.
Figure 3.1 summarises the GARCH estimation process.
Once we have the VaR forecasts, we can proceed with the copula estimation. Let X be
the original vector of return indices, then the copula fitting procedure is given as in figure
3.2. In this manner, taking different copula densities or marginals, we will have a set of
models with different copulas and different marginals. Our objective is to compare these
models. We do this using Hansen’s model confidence set (MCS) procedure.
3.3. COPULA ESTIMATION 28
Figure 3.2: The process of estimating a copula and simulating from it.
Chapter 4
In order to work out the MCS procedure for rankings, we will have to obtain a loss distri-
butions from the set of models.
We now have the loss distributions from the different models obtained. We will use the
MCS procedure to compare the models.
The expected loss of model i relative to model j throughout time, T is given as follows:
Li < Lj ,
=⇒ di,j < 0,
=⇒ µij < 0. (4.2.3)
Therefore, the model confidence set, i.e. the set of surviving models, given confidence level
(1 − α) is defined as
M∗ ≡ {i ∈ M0 : µij ≤ 0, ∀j ∈ M0 }. (4.2.4)
The model confidence set (MCS) is constructed by a process of sequential hypothesis testing
where competing models are tested for equal predictive ability (EPA) by an equivalence test,
δM . If this EPA hypothesis is rejected, inferior models are eliminated using an elimination
rule, eM . The null and alternative hypotheses are, respectively:
sequential testing is performed until H0,M is finally accepted under the given confidence
level, (1 − α).
The MCS p-values, analogous to those defined in classical statistical inference, are defined
as: PH0,Mi corresponding to the hypothesis, H0,Mi
where, Fi (ti ) is the c.d.f. i.e, the cumulative density function of the ith test statistic, ti .
Using these, the MCS p-values corresponding to each model can be evaluated as
This translates to the fact that, model i belongs to the MCS, M̂ ∗1−α if p̂eMi ≥ α. Hence, it
is unlikely that models with low p-values will make it to MCS.
To compare the performances of the candidate models i and j, [45] proposed a multiple
t-statistic approach using the following statistics:
T
where ti = √ di
, di = (Li − L) , Li ≡ T −1 Li,t and L = m−1
P P
0 Li for i, j ∈ M.
var(d
ˆ i) t=1 i∈M
The hypotheses for this alternative test statistics are a little different, given by:
H0,M : µi = 0, ∀i ∈ M, (4.2.10)
HA,M : µi 6= 0, f or some i ∈ M,
where µi = E(di ). Inferring from the equivalence between H0,M {Hij , ∀i, j ∈ M}and{Hi , ∀i, j ∈
M} where M = 1, . . . , m0 , we arrive at
An extension of these results helps us arrive at the test statistics TR,M and Tmax,M Since these
statistics depend on nuisance parameters, their asymptotic distributions are non-standard.
However, with estimation using the bootstrap algorithm, this becomes less of an issue.
The elimination rule should satisfy the assumption, limn→∞ P(eM ∈ M∗ |HA,M ) = 0 and
it should be coherent with the equivalence test, δM , i.e. P(δM = 1, eM ∈ M∗ ) ≤ P0 (δM = 1).
The elimination rule for the test statistic Tmax,M is
This identity shows that the diagonal elements of Ω̂ are given as follows:
B B B
T X ∗ 2 T X ∗ 2 T X ∗ ∗ 2
√
(Z b,i − Z i ) = (db,i − di ) = (ζb,i − ζb,· ) = var(
ˆ T (di )). (4.2.15)
B b=1 B b=1 B b=1
Therefore, the distribution of Tmax can be approximated under the null hypothesis as
follows:
p −1 √
∗ ∗
max( D̂ T (Z b − Z))i = max(diag(var(d ˆ 1 ))−1/2 (Z b − Z))i ,
ˆ 1 ), . . . , var(d
i i
∗ ∗ ∗
db,i − di ζb,i − ζb,·
= max q = max q = max t∗b,i ,
i i
var(d
ˆ i) var(d
ˆ i)
∗
= Tb,max . (4.2.16)
A detailed description of the bootstrap algorithm and the complete MCS algorithm can
be found in the working paper of Hansen et al (2005). [7] Let M0 be the set of all models
1, . . . , m0 and [Li,t ]Tt=1 , the loss distribution corresponding to each model. Then figure 3.3
presents a flowchart summarising the MCS mechanism we just discussed.
4.3. SENER’S PENALTY MEASURE 33
α−1 X
α−i zi
Y Zi+m
X 1 Y
= (1 + b,i ) (1 + b,i+m ) − 1
i=1 m=1
ki,i+m b=1 i=1
Figure 4.1: Hansen’s model confidence set procedure for comparing models.
Chapter 5
The initial objective of our study was to apply Copula-GARCH models to financial time-
series data and compare them using Hansen’s model confidence set procedure to be able to
make statements on significance valid in the traditional sense, as in classical statistical theory.
The results however, motivated us to study another aspect of this modelling process: the
marginal models. Copulas are so very famous, because of the added advantage they produce
by allowing to model the marginals and the dependence structure separately. As stated in
2.1, as an extension of our study, we will also address as to which marginals are superior over
the others and statistically determine how harmful could the choice of an inferior marginal
be.
Also, for all our models, we have performed out-of-sample forecasts (March 2003 to
May 2013). In-sample-forecasts are good to analyse the performance of a model, that test a
model’s efficiency at data points used to estimate the model itself. This technique however,
is not a true meter to analyse how the model adjusts to exogenous shocks. For this and for
actual economic utility, we employ out-of-sample forecasting method.
Our final results, i.e. the plots and the MCS results are based on an equi-weighted
portfolio of the four log indices we consider: DAXINDEX, FTSEINDEX, SnPIN-
DEX, REXINDEX. It economically makes sense only if can analyse the co-movement of
the indices together in order to construct a diverse portfolio with minimum loss on invest-
ment, the entire point of multivariate modelling.
Our computations have been performed using R 3.5.2, R Project for Statistical Com-
puting on a 32-bit processor, 2012 MB RAM and a 64-bit processor, 4GB RAM.
We will specify what models have been computed with what specifications, as we proceed.
Two processors have been used due to computational restraints, because of which, we will
not consider time as an absolute quantitative parameter to capture the complexity of the
models. However, these simulations have been performed a couple times over, using three
different processors, so a qualitative safe judgement would be made.
35
5.1. PHASE I: MODELS WITH ROLLING WINDOW 36
VaR forecasts for 3000 days in the future, updated daily with 10,000 simulations per day using a rolling
window of 500 days. GARCH(1,1) is fitted on demeaned data. Returns indicate log stock indices of an
equi-weighted portfolio of the four indices.
Note:
• * The GAS Factor copula (figure 5.8) is enormously time taking for one update, hence,
this is an exceptional model that does not follow the rolling window mechanism. Also,
all simulations need to be stored for it’s processing, we are only able to take 500 sim-
ulations instead of the usual 10,000 because of R array size and our RAM limitations.
5.1. PHASE I: MODELS WITH ROLLING WINDOW 37
VaR forecasts for 3000 days in the future, updated daily with 10,000 simulations per day using a rolling
window of 500 days. GARCH(1,1) is fitted on demeaned data. Returns indicate log stock indices of an
equi-weighted portfolio of the four indices.
VaR forecasts for 3000 days in the future, updated daily with 10,000 simulations per day using a rolling
window of 500 days. GARCH(1,1) is fitted on demeaned data. Returns indicate log stock indices of an
equi-weighted portfolio of the four indices.
5.1. PHASE I: MODELS WITH ROLLING WINDOW 38
VaR forecasts for 3000 days in the future, updated daily with 10,000 simulations per day using a rolling
window of 500 days. GARCH(1,1) is fitted on demeaned data.Returns indicate log stock indices of an
equi-weighted portfolio of the four indices.
VaR forecasts for 3000 days in the future, updated daily with 10,000 simulations per day using a rolling
window of 500 days. GARCH(1,1) is fitted on demeaned data.Returns indicate log stock indices of an
equi-weighted portfolio of the four indices.
5.1. PHASE I: MODELS WITH ROLLING WINDOW 39
VaR forecasts for 3000 days in the future, updated daily with 10,000 simulations per day using a rolling
window of 500 days. GARCH(1,1) is fitted on demeaned data.Returns indicate log stock indices of an
equi-weighted portfolio of the four indices.
VaR forecasts for 3000 days in the future, updated daily with 10,000 simulations per day using a rolling
window of 500 days. GARCH(1,1) is fitted on demeaned data.Returns indicate log stock indices of an
equi-weighted portfolio of the four indices.
5.1. PHASE I: MODELS WITH ROLLING WINDOW 40
VaR forecasts for 3000 days in the future with 500 simulations per day. GARCH(1,1) is fitted on demeaned
data.Returns indicate log stock indices of an equi-weighted portfolio of the four indices.
VaR forecasts for 3000 days in the future, updated daily with 10,000 simulations per day. GARCH(1,1) is
fitted on demeaned data. Returns indicate log stock indices of an equi-weighted portfolio of the four indices.
• The mixture copula (figure 5.10) is a 1 : 2 weighted combination of the Gaussian and
Clayton copulas. This combination has been chosen randomly via hit and trial and
5.1. PHASE I: MODELS WITH ROLLING WINDOW 41
VaR forecasts for 3000 days in the future, updated daily with 10,000 simulations per day using a rolling
window of 500 days. GARCH(1,1) is fitted on demeaned data.Returns indicate log stock indices of an
equi-weighted portfolio of the four indices.
We focus on the negative tail of the VaR forecasts in our analysis since negative sentiments
breed deeper in the financial community. Backtests: This is a hindcasting technique in
which the credibility of a model is tested taking some points in the estimation window
space, and quantifying the model’s predictive ability. We use the backtestVaR() method in
R [6] to perform the unconditional coverage test of Kupeic (Lruc), conditional coverage test
of Christofferson’s (Lrcc) and the Dynamic Quantile test of Engel and Manganelli’s (DQ)
backtests. AD shows the mean absolute deviation and maximum absolute deviation in the
test determined by Da Veiga and McAleer(2008) measured from the actual observations and
the quantiles. AE stands for the actual over exceedance ratio for a model. It is counted as
one exceedance when the realised return falls below the Value-at-Risk (VaR) estimate.
Concluding remarks:
• Needless to say, the Vine copula-GARCH performs very well, given it’s superior flexi-
bility and relative ease of computation.
• Mixture copula-GARCH comes well as statistically the best, on average. This is also
predictable accrediting to the doubled flexibility (at least) compared to the rest of the
copulas. However, it’s better performance, as evident from the graph comes at the cost
of reduced profitability as is also reflected in Sener’s penalty meaasure, which is why
we would not weigh its god performance as an economically significant result.
5.1. PHASE I: MODELS WITH ROLLING WINDOW 42
Table 5.1: Backtest results of models with rolling window for the entire duration: 2003-2013.
Table 5.2: MCS results for models with rolling window for the entire duration: 2003-2013.
Table 5.3: MCS results for the period of financial crisis: December,2008-March, 2012 for models
with rolling window.
Table 5.4: MCS results for the entire period : 2003-2013 for models with rolling window.
Table 5.5: MCS results for the entire period: 2003-2013 for models with rolling window.
Table 5.6: MCS results for the period of financial crisis: Dec, 2008-March, 2012 for models with
rolling window.
Table 5.7: Sener’s penalties for 1% VaRs of copula-GARCH models with rolling window through
2003-2013.
• The GAS Factor copula had our hopes set quite high, but it’s under-performance is,
but obvious, since it is not updated at all where as the others are updated regularly.
• We update them so frequently to infer how well the models can adjust to exogenous
shocks and trends. Evidently, from the plots, we see that of all models, the exceedances
increase manifold at the beginning of the financial crisis, about in 2007-2008 and reduce
gradually over time.
• An important result is that, the models in the superior set remain consistent even
during an extremely high period of volatility (during the financial crisis) i.e. to say,
that the models that perform well, do so in stable as well as volatile trends. This
is incredible as it directly proves that the models that survive are also capable to
adjusting to exogenous shocks.
• A striking result here is presented in the table 5.8. We rely on the MCS procedure
to help us find the best set of models, given an α confidence but it largely focusses
on the informativeness of the data, i.e. the loss distributions penalising losses only.
It fails to capture if a model is significantly reducing the profitability of the portfolio.
Sener’s penalty shows that Mixture copula performs best in violation space but has
the highest penalty in the safe space, making it the worst penalised model of all.
• The Vine copula, a fair competitor since the beginning has an adequate penalty and
is statistically superior as well, as shown by the MCS results.
We next consider those days of forecast until the DCC doesn’t converge.
Concluding remarks:
Table 5.8: MCS results for the entire duration: 2003-2006 including DCC-GARCH(1,1) for models
with rolling window.
Table 5.9: MCS results for the entire duration: 2003-2006 including the non copula DCC-
GARCH(1,1) model for models with rolling window.
performs competitively with regard to the other copula models and even statistically,
outperforms the simple Arcchimedean copulas.
• Having performed these simulations, we can though state that working with Vine or
GAS factor or Mixture copulas is relatively easier as they encounter less convergence
issues compared to he DCC-GARCH(1,1).
• It is not a surprise that the GAS Factor copula and the Gaussian copulas perform
adequately well, even though they do not make it to the model confidence set, judging
by the fact that the survive quite a few rounds of the sequential hypothesis testing
procedure we employ.
5.2. PHASE II: MODELS WITHOUT ROLLING WINDOW 48
VaR forecasts for 3000 days in the future, with 10,000 simulations per day. The data is not demeaned to fit
the ARMA(1,0)-GARCH(1,1) process for all models in this phase. Returns indicate log stock indices of an
equi-weighted portfolio of the four indices.
5.2. PHASE II: MODELS WITHOUT ROLLING WINDOW 49
VaR forecasts for 3000 days in the future, with 10,000 simulations per day. The data is not demeaned to fit
the ARMA(1,0)-GARCH(1,1) process for all models in this phase. Returns indicate log stock indices of an
equi-weighted portfolio of the four indices.
5.2. PHASE II: MODELS WITHOUT ROLLING WINDOW 50
VaR forecasts for 3000 days in the future, with 10,000 simulations per day. The data is not demeaned to fit
the ARMA(1,0)-GARCH(1,1) process for all models in this phase. Returns indicate log stock indices of an
equi-weighted portfolio of the four indices.
5.2. PHASE II: MODELS WITHOUT ROLLING WINDOW 51
VaR forecasts for 3000 days in the future, with 10,000 simulations per day. The data is not demeaned to fit
the ARMA(1,0)-GARCH(1,1) process for all models in this phase. Returns indicate log stock indices of an
equi-weighted portfolio of the four indices.
5.2. PHASE II: MODELS WITHOUT ROLLING WINDOW 52
VaR forecasts for 3000 days in the future, with 10,000 simulations per day. The data is not demeaned to fit
the ARMA(1,0)-GARCH(1,1) process for all models in this phase. Returns indicate log stock indices of an
equi-weighted portfolio of the four indices.
5.2. PHASE II: MODELS WITHOUT ROLLING WINDOW 53
VaR forecasts for 3000 days in the future, with 10,000 simulations per day. The data is not demeaned to fit
the ARMA(1,0)-GARCH(1,1) process for all models in this phase. Returns indicate log stock indices of an
equi-weighted portfolio of the four indices.
5.2. PHASE II: MODELS WITHOUT ROLLING WINDOW 54
VaR forecasts for 3000 days in the future, with 10,000 simulations per day. The data is not demeaned to fit
the ARMA(1,0)-GARCH(1,1) process for all models in this phase. Returns indicate log stock indices of an
equi-weighted portfolio of the four indices.
5.2. PHASE II: MODELS WITHOUT ROLLING WINDOW 55
VaR forecasts for 3000 days in the future, with 10,000 simulations per day. The data is not demeaned to fit
the ARMA(1,0)-GARCH(1,1) process for all models in this phase. Returns indicate log stock indices of an
equi-weighted portfolio of the four indices.
5.2. PHASE II: MODELS WITHOUT ROLLING WINDOW 56
VaR forecasts for 3000 days in the future, with 500 simulations per day. The data is demeaned to fit the
ARMA(1,1)-GARCH(1,1) process for all models in this phase. Returns indicate log stock indices of an
equi-weighted portfolio of the four indices.
5.2. PHASE II: MODELS WITHOUT ROLLING WINDOW 57
VaR forecasts for 3000 days in the future, with 10,000 simulations per day. The data is not demeaned to fit
the ARMA(1,0)-GARCH(1,1) process for all models in this phase. Returns indicate log stock indices of an
equi-weighted portfolio of the four indices.
5.2. PHASE II: MODELS WITHOUT ROLLING WINDOW 58
Table 5.10: Exceedances of models without rolling window for the entire duration: 2003-2013.
Model Exceedances
(Copula) DAXINDX FTSEINDX SnPCOMPINDX REXINDX
-ARMA(1,0) 0.1% 1% 5% 0.1% 1% 5% 0.1% 1% 5% 0.1% 1% 5%
-GARCH(1,1)
Vine 37 119 239 41 135 281 27 85 189 5 65 197
Clayton 35 122 238 38 134 279 25 87 189 6 64 196
Normal 32 122 237 43 132 277 28 86 89 5 63 198
Student t 33 123 237 39 138 281 25 88 189 6 64 199
Gumbel 35 123 238 43 131 280 26 87 187 5 60 199
Frank 33 120 237 42 132 276 27 87 189 7 62 197
Joe 33 123 239 39 134 277 27 85 189 6 59 199
Joe-ARMA(1,1) 32 118 236 42 134 282 25 85 188 6 64 196
GAS* 15 114 289 28 132 301 10 71 197 9 57 185
Mixture 0 14 157 0 21 181 0 11 132 0 2 108
Note:
Concluding remarks: We get to conclude quite a few things from this part, where we
do not update the models throughout our time range.
• The VaR forecasts are not economically significant, as is evident from the graphs since
they cannot accustom themselves with exogenous shocks, which is but obvious since
we are not updating our realisations.
• The Vine copula and the mixture copula still perform adequately well but it is inter-
esting to notice how sensitive these results are to the choice of the loss distribution.
• As expected the performance of the GAS Factor considerably improves. However, this
leads us to certain drawbacks of the factor copula model. It is parsimonious and almost
equivalent in predictive performance to standard elliptical copulas for small number
of variables (in our case, 4). It is difficult to update the GAS factor because of its
computational complexity (at least in R).
5.2. PHASE II: MODELS WITHOUT ROLLING WINDOW 59
Table 5.11: Backtest results of models without rolling window for the entire duration: 2003-2013.
Table 5.12: MCS results for the entire duration: 2003-2013 for models without rolling window.
Table 5.13: MCS results for the entire time for models without rolling window.
Table 5.14: MCS results for the entire period : 2003-2013 for models without rolling window.
Table 5.15: MCS results for the entire duration: 2003-2013 for models without rolling window.
Table 5.16: MCS results for the period of financial crisis: Dec, 2008-March, 2012 for models
without rolling window.
Table 5.17: Sener’s penalties for 1% VaRs of copula-GARCH models with rolling window through
2003-2013.
• It is however, interesting to note than the GAS still performs better in high volatility
periods compared to the rest of the models which establishes its predictive superiority
over the others, even if it is not via a huge margin.
• The most remarkable result of this part is what the even MCS fails to capture. The
MCS constantly results in the Mixture copula as one of the best models where it
is one whose composition is chosen randomly. It is evidently not profitable because
forecasts from it are far below the realised returns, so although there are no losses
occurring statistically, from an economic perspective, we are losing far more as indirect
opportunity cost. As from table 5.18, for the mixture copula, the violation space
penalty is negligible but the safe space penalty is manifold, enough to allow the net
penalty for a Mixture copula three times that of the other models.
VaR forecasts for 100 days in the future, updated daily with 10,000 simulations per day using a rolling
window of 500 days. GARCH(1,1) is fitted on demeaned data. Returns indicate log stock indices of an
equi-weighted portfolio of the 4 indices.
5.3. PHASE III: MODELS WITH FIXED COPULA AND VARYING MARGINALS 64
We take a confidence level 80% using the Asymmetric VaR loss distribution and test statistic TR. The order
of the models correspond to their order in Fig.5.21
We take a confidence level 80% using the Asymmetric VaR loss distribution and test statistic Tmax. The
order of the models correspond to their order in Fig.5.21
5.3. PHASE III: MODELS WITH FIXED COPULA AND VARYING MARGINALS 65
We take a confidence level 80% using the Continuous ranked probability score loss distribution and test
statistic Tmax. The order of the models correspond to their order in Fig.5.21
We take a confidence level 80% using the Continuous ranked probability score loss distribution and test
statistic TR. The order of the models correspond to their order in Fig.5.21
5.3. PHASE III: MODELS WITH FIXED COPULA AND VARYING MARGINALS 66
VaR forecasts for 100 days in the future, updated daily with 10,000 simulations per day using a rolling
window of 500 days. GARCH(1,1) is fitted on demeaned data. Returns indicate log stock indices of an
equi-weighted portfolio of the 4 indices.
5.3. PHASE III: MODELS WITH FIXED COPULA AND VARYING MARGINALS 67
We take a confidence level 80% using the Asymmetric VaR loss distribution and test statistic TR. The order
of the models correspond to their order in Fig.5.26
We take a confidence level 80% using the Asymmetric VaR loss distribution and test statistic Tmax. The
order of the models correspond to their order in Fig.5.26
5.3. PHASE III: MODELS WITH FIXED COPULA AND VARYING MARGINALS 68
We take a confidence level 80% using the Continuous ranked probability score loss distribution and test
statistic Tmax. The order of the models correspond to their order in Fig.5.26
We take a confidence level 80% using the Continuous ranked probability score loss distribution and test
statistic TR. The order of the models correspond to their order in Fig.5.26
5.3. PHASE III: MODELS WITH FIXED COPULA AND VARYING MARGINALS 69
Concluding remarks:
• Hansen and Lunde [43] showed that for univariate models, almost nothing beats a
GARCH(1,1) process. Our simple experiment here reinstates this for multivariate
copula-GARCH models as well, that the GARCH(1,1) is indeed best when coupled
with copulas in multi-dimensions.
• The sGARCH(1,1) is consistently the second best whereas the eGARCH marginal has
a considerably poor fitting. As we see in the MCS results, it almost always fails to
belong to the superior set of models.
• Based on the fact that quite a few marginals out of the 7 we employ get eliminated from
the superior set of models, indicates that not all the marginals are equally as good and
they do have some power in changing the quality of the model. The Vine-GARCH(1,1)
performs best where as Vine-eGARCH doesn’t. This answers our question that the
marginal choice does play a role in the overall quality of prediction form a multivariate
model.
5.4. PHASE IV: MODELS WITH FIXED MARGINAL AND VARYING COPULAS 70
Table 5.18: Sener’s penalties for 1% VaRs of copula-eGARCH models with rolling window during
2003.
(g) Joe-ARMA(1,1)-eGARCH(1,1)
VaR forecasts for 150 days in the future, updated daily with 10,000 simulations per day using a rolling
window of 500 days. GARCH(1,1) is fitted on demeaned data. Returns indicate the log stock indices of an
equi-weighted portfolio of the 4 indices.
5.4. PHASE IV: MODELS WITH FIXED MARGINAL AND VARYING COPULAS 72
We take a confidence level 80% using the Asymmetric VaR loss distribution and test statistic TR. The order
of the models correspond to their order in Fig.5.31
We take a confidence level 80% using the Asymmetric VaR loss distribution and test statistic Tmax. The
order of the models correspond to their order in Fig.5.31
5.4. PHASE IV: MODELS WITH FIXED MARGINAL AND VARYING COPULAS 73
We take a confidence level 80% using the continuous ranked probability score loss distribution and test
statistic Tmax. The order of the models correspond to their order in Fig.5.31
We take a confidence level 80% using the continuous ranked probability score loss distribution and test
statistic TR. The order of the models correspond to their order in Fig.5.31
5.5. DRAWBACKS OF HANSEN’S MODEL CONFIDENCE SET PROCEDURE 74
• The MCS takes only the losses incurred from a model as input. It does not consider
the time complexity of a model. In fact there exists no quantifier, to the best of
our knowledge, that would take into account, both computation time and prediction
efficiency of a model and optimize the criteria to converge to the best set of models.
• It is highly dependent on the nature of the loss function. This is not exactly a bane,
because loss functions are meant to focus on the aspect one is trying to control. How-
ever, when we reach at conclusions like, an X copula is best, this is all subject to the
loss criteria applied.
• A potential drawback, particular to our application could be than we only consider the
negative tail of VaR forecasts and the results might change to some extent when we
also consider the positive tail VaR forecasts, specially when modelling with asymmetric
copulas.
These drawbacks are well overcome by Sener’s penalisation measure and our results
show the precise pitfall they cover. Together with the MCS results and Sener’s penalisation
measure, we can safely conclude about the model risk posed by the copula-GARCH models.
Chapter 6
• Capturing the dependence between variables via copula models is a better approach
than multivariate con-copula models on the basis of time complexity, optimisation
simplicity, and predictive efficiency. (refer results from Phase I)
• We have been able to implement the GAS factor copula in R. This model is relatively
tedious compared to the others and it is not easy or economical to update this model
very frequently. Phase II shows that when models do not need to be updated frequently,
the GAS performs adequately well. However, the Vine copula-GARCH(1,1) is nearly
as good in small dimensions like ours.
• The choice of a wrong, so to say, an inferior marginal can do enough wrong to the overall
predictive ability of the model, for example our case study with the eGARCH(1,1) and
several elliptical and Archimedean copulas. (refer Phase IV)
• Sener’s penalisation measure gives remarkably fair conclusions on what models are
best, from the perspective of economic significance.
75
6.3. CONCLUSION ON MODEL RISK 76
• Our study qualitatively suggests that the model risk imposed by the marginals models
in multivariate models is significant.
• The mixture copula can be improvised to estimate the weights of copula constituents
via the Expectation-Maximization algorithm.
• The Dynamic copula are a recent addition to the literature of copulas and can also
be implemented and compared with.
• A very significant study could be performed on the comparison of model risk im-
posed via univariate modelling versus that by multivariate modelling(diverse
portfolio). In univariate modelling, the only source of model risk is the choices of the
marginals. However, for a diverse portfolio, model risk could emanate from the choice
of the marginal, as well as from the additional dependency structure, i.e. the copula.
Risks imposed from these two factors could mutually cancel out each other since, co-
movement of market trends will be better capture in this case, or it could compound
from the level of a univariate model. Statistically, it is possible to compare the two,
as the univariate model allows to forecast for an individual return, and even with a
multivariate model, one can forecast for individual constituents after the model fitting.
Bibliography
[14] Albanese, C., Chen, O., Dalessandro, A., and Vidler, A. (2005). Dynamic credit corre-
lation modeling. preprint.
[15] Albanese, C. and Vidler, A. (2008). Dynamic conditioning and credit correlation baskets.
Available at SSRN 1113108.
[16] Bollerslev, T. (1986). Generalized autoregressive conditional heteroskedasticity. Journal
of econometrics, 31(3):307–327.
[17] Bollerslev, T., Engle, R. F., and Nelson, D. B. (1994). Arch models. Handbook of
econometrics, 4:2959–3038.
[18] Bollerslev, T. et al. (1987). A conditionally heteroskedastic time series model for spec-
ulative prices and rates of return. Review of economics and statistics, 69(3):542–547.
[19] Bollerslev, T. et al. (1990). Modelling the coherence in short-run nominal exchange rates:
a multivariate generalized arch model. Review of Economics and statistics, 72(3):498–505.
[20] Bougerol, P. and Picard, N. (1992). Stationarity of garch processes and of some non-
negative time series. Journal of econometrics, 52(1-2):115–127.
[21] Box, G. E. and Pierce, D. A. (1970). Distribution of residual autocorrelations in
autoregressive-integrated moving average time series models. Journal of the American
statistical Association, 65(332):1509–1526.
[22] Brockwell, P. J., Davis, R. A., and Calder, M. V. (2002). Introduction to time series
and forecasting, volume 2. Springer.
[23] Christoffersen, P. (2010). Backtesting. Encyclopedia of Quantitative Finance.
[24] Christoffersen, P. and Pelletier, D. (2004). Backtesting value-at-risk: A duration-based
approach. Journal of Financial Econometrics, 2(1):84–108.
[25] David, M., Ramahatana, F., Trombe, P.-J., and Lauret, P. (2016). Probabilistic fore-
casting of the solar irradiance with recursive arma and garch models. Solar Energy,
133:55–72.
[26] Drost, F. C. and Nijman, T. E. (1993). Temporal aggregation of garch processes.
Econometrica: Journal of the Econometric Society, pages 909–927.
[27] Embrechts, P., Frey, R., and McNeil, A. (2005). Quantitative risk management (prince-
ton series in finance, vol. 10).
[28] Embrechts, P., Lindskog, F., and McNeil, A. (2001). Modelling dependence with copulas,
rapport technique, dép. de math. Inst. Féd. de Technol. de Zurich, Zurich, Switzerland.
[29] Engle, R. (2000). Dynamic conditional correlation% a simple class of multivariate garch
models. university of california at san diego. Economics Working Paper Series, 9.
[30] Engle, R. (2002). Dynamic conditional correlation: A simple class of multivariate gen-
eralized autoregressive conditional heteroskedasticity models. Journal of Business & Eco-
nomic Statistics, 20(3):339–350.
BIBLIOGRAPHY 80
[31] Engle, R. and McFadden, D. (1994). Handbook of econometrics, volume iv. Bollerslev
T, Engle RF and Nelson D B. ARCH models. Amsterdam: Elsevier Science BV, pages
1959–3038.
[32] Engle, R. F. and Bollerslev, T. (1986). Modelling the persistence of conditional vari-
ances. Econometric reviews, 5(1):1–50.
[34] Engle, R. F. and Sheppard, K. (2001). Theoretical and empirical properties of dynamic
conditional correlation multivariate garch. Technical report, National Bureau of Economic
Research.
[35] Francq, C. and Zakoian, J.-M. (2019). GARCH models: structure, statistical inference
and financial applications. Wiley.
[36] Genest, C., Ghoudi, K., and Rivest, L.-P. (1995). A semiparametric estimation pro-
cedure of dependence parameters in multivariate families of distributions. Biometrika,
82(3):543–552.
[37] Geweke, J. and Porter-Hudak, S. (1983). The estimation and application of long memory
time series models. Journal of time series analysis, 4(4):221–238.
[38] Giacomini, R. and White, H. (2006). Tests of conditional predictive ability. Economet-
rica, 74(6):1545–1578.
[39] Gneiting, T. and Ranjan, R. (2011). Comparing density forecasts using threshold-and
quantile-weighted scoring rules. Journal of Business & Economic Statistics, 29(3):411–422.
[40] González-Rivera, G., Lee, T.-H., and Mishra, S. (2004). Forecasting volatility: A reality
check based on option pricing, utility function, value-at-risk, and predictive likelihood.
International Journal of forecasting, 20(4):629–645.
[41] Granger, C. W. and Morris, M. J. (1976). Time series modelling and interpretation.
Journal of the Royal Statistical Society: Series A (General), 139(2):246–257.
[42] Hansen, P. R. et al. (2001). An unbiased and powerful test for superior predictive ability.
Technical report.
[43] Hansen, P. R. and Lunde, A. (2005). A forecast comparison of volatility models: does
anything beat a garch (1, 1)? Journal of applied econometrics, 20(7):873–889.
[44] Hansen, P. R., Lunde, A., and Nason, J. M. (2003). Choosing the best volatility models:
the model confidence set approach. Oxford Bulletin of Economics and Statistics, 65:839–
861.
[45] Hansen, P. R., Lunde, A., and Nason, J. M. (2011a). The model confidence set. Econo-
metrica, 79(2):453–497.
BIBLIOGRAPHY 81
[46] Hansen, P. R., Lunde, A., and Nason, J. M. (2011b). The model confidence set. Econo-
metrica, 79(2):453–497.
[47] Hartz, C., Mittnik, S., and Paolella, M. (2006). Accurate value-at-risk forecasting based
on the normal-garch model. Computational Statistics & Data Analysis, 51(4):2295–2312.
[49] Holt, C. C. (1957). Forecasting trends and seasonal by exponentially weighted averages,
office of naval research memorandum 52. reprinted in holt, charles c.(january–march 2004),
forecasting trends and seasonal by exponentially weighted averages. International Journal
of Forecasting, 20(1):510.
[51] Kaeck, A., Rodrigues, P., and Seeger, N. J. (2018). Model complexity and out-of-sample
performance: evidence from s&p 500 index returns. Journal of Economic Dynamics and
Control, 90:1–29.
[52] Kupiec, P. (1995). Techniques for verifying the accuracy of risk measurement models.
The J. of Derivatives, 3(2).
[54] Laurent, S., Rombouts, J. V., and Violante, F. (2012). On the forecasting accuracy of
multivariate garch models. Journal of Applied Econometrics, 27(6):934–955.
[55] Li, W., Ling, S., and McAleer, M. (2002). Recent theoretical results for time series
models with garch errors. Journal of Economic Surveys, 16(3):245–269.
[57] Liu, Y. and Luger, R. (2009). Efficient estimation of copula-garch models. Computa-
tional Statistics & Data Analysis, 53(6):2284–2297.
[58] Lojowska, A., Kurowicka, D., Papaefthymiou, G., and van der Sluis, L. (2010). Advan-
tages of arma-garch wind speed time series modeling. In 2010 IEEE 11th International
Conference on Probabilistic Methods Applied to Power Systems, pages 83–88. IEEE.
[59] McWilliam, N., Loh, K.-W., and Huang, H. (2011). Incorporating multi-dimensional
tail dependencies in the valuation of credit derivatives. Quantitative Finance, 11(12):1803–
1814.
[60] Minović, J. Z. (2009). Modeling multivariate volatility processes: Theory and evidence.
Theoretical & Applied Economics, 16(5).
[61] Nelsen, R. B. (1997). Dependence and order in families of archimedean copulas. Journal
of Multivariate Analysis, 60(1):111–122.
BIBLIOGRAPHY 82
[62] NUMBER, P. (1992). Working experience and research projects. Statistics, 8:10.
[64] Oh, D. H. and Patton, A. J. (2018). Time-varying systemic risk: Evidence from a dy-
namic copula model of cds spreads. Journal of Business & Economic Statistics, 36(2):181–
195.
[65] Patton, A. J. (2009). Copula–based models for financial time series. In Handbook of
financial time series, pages 767–785. Springer.
[68] Şener, E., Baronyan, S., and Mengütürk, L. A. (2012). Ranking the predictive per-
formances of value-at-risk estimation methods. International Journal of Forecasting,
28(4):849–873.
[69] Sentman, L. (1949). Some observations on the role of the faculty representative in field
work. Social Service Review, 23(3):337–346.
[70] Weiß, G. N. (2011). Are copula-gof-tests of any practical use? empirical evidence for
stocks, commodities and fx futures. The Quarterly Review of Economics and Finance,
51(2):173–188.
[72] White, H. (2000). A reality check for data snooping. Econometrica, 68(5):1097–1126.
Chapter 7
Appendix
I
Here is a snippet of the data we use, with specifications described in section 3.1
83
CHAPTER 7. APPENDIX 84
II
This section contains abstract codes of the various models we have implemented
Figure 7.6: Code for auxiliary functions and vine copula-GARCH(1,1) model fitting.
CHAPTER 7. APPENDIX 87
Figure 7.7: Code for simulating and forecasting from a vine copula-GARCH(1,1) model.
Figure 7.9: Abstract code for estimation of GAS Factor copula-GARCH(1,1) model parameters.
CHAPTER 7. APPENDIX 89
Figure 7.10: Abstract code for estimation of GAS Factor copula-GARCH(1,1) model density.
CHAPTER 7. APPENDIX 90
Figure 7.11: Abstract code for simulation from GAS Factor copula-GARCH(1,1) model.
CHAPTER 7. APPENDIX 91