Professional Documents
Culture Documents
Existing studies on covariance estimation generally assume that the future covariance matrix must be
estimated based only on the limited history of a given set of portfolio constituents. In this study, we propose
a new perspective on how to estimate the covariance matrix. We present a purely data-driven approach
that selects the estimation parameters using cross validation to be historically optimal on a disjoint transfer
set of assets according to the given objective. The proposed approach additionally uses a second shrinkage
target that is determined based on how much the sample eigenvalues are imbalanced according to their Gini
coefficient. Our empirical evaluation based on a total of six stock market indices shows that the proposed
approach outperforms established estimators in minimizing variance and maximizing risk-adjusted return.
The second shrinkage target is particularly relevant for high-dimensional covariance matrices where the
number of assets is greater than the number of historic datapoints. To the best of our knowledge, this study
is the first to apply the concept of transfer learning to the problem of covariance estimation.
Key words : Covariance estimation, transfer learning, non-linear shrinkage, second shrinkage target,
1. Introduction
Covariance estimation is a well studied problem in the finance and operations research literature. An
accurate estimate of the future covariance is required for standard portfolio optimization procedures
* Views expressed in this paper are those of the author and do not necessarily reflect those of Deka Investment or its
employees.
like minimizing variance and maximizing risk-adjusted return (Markowitz 1952). The task of the
problem is to estimate the future covariance of a given portfolio of N assets based on the history of
T datapoints. However, the sample covariance is generally a poor estimate of the future covariance,
which leads to deteriorated out-of-sample performance (Frost and Savarino 1986, Jorion 1986,
Michaud 1989, Olivares-Nadal and DeMiguel 2018). As a remedy, researchers have proposed several
approaches to improve the sample covariance estimate. The proposed methods can be grouped into
(i) explicit (non-) linear shrinkage (e.g., Ledoit and Wolf 2003, 2017b, Engle et al. 2019), (ii) implicit
shrinkage by combining sample and reference weights (e.g., Frahm and Memmel 2010, Bodnar
et al. 2018), (iii) thresholding (e.g., Bickel and Levina 2008, Fan et al. 2013), and signal/noise
separation (e.g., Laloux et al. 1999, Plerou et al. 2002, Zhao et al. 2019), (iv) factor models (e.g.,
Fan et al. 2011, 2018, De Nard et al. 2019), or (v) regularization of the inverse covariance matrix
(e.g., Friedman et al. 2008, Lam and Fan 2009, Cai et al. 2011, Nguyen et al. 2021). Other studies
have proposed estimation error reduction through constrained portfolio optimization in contrast to
covariance adjustments (e.g., Jagannathan and Ma 2003, DeMiguel et al. 2009a, Fan et al. 2012).
Extensive literature reviews and empirical studies of covariance estimation can be found in Fan
et al. (2016), Ledoit and Wolf (2017a) and Ledoit and Wolf (2021a).
However, all of the aforementioned studies considered the problem of covariance estimation from
a rather restrictive perspective, where the only available real-world data for parameter selection is
given by the history of the actual portfolio constituents. To determine the estimation parameters
such as shrinkage intensities, prior studies applied loss functions with statistically derived priors for
an unknown oracle covariance, e.g., by using random matrix theory (most recently Ledoit and Wolf
2021b). In a similar fashion, using sample data only, others calculated in-sample portfolio volatility
to directly determine the shrinkage intensity (e.g., Frahm and Memmel 2010, Bodnar et al. 2018)
Following prior studies, we also assume that the future covariance must be estimated for N assets
based on only T datapoints. However, we assume that we are additionally provided with other
stock market time-series data from a disjoint set of assets to select the estimation parameters.
From a real-world perspective, the history of joint datapoints may be reduced due to company
merges and acquisitions, splits, or one of the portfolio constituents being recently listed at the stock
market. In this case, the future covariance matrix must be estimated based on the comparably short
history of the portfolio constituents. We then argue that financial investors can rely on additional
time-series data from a disjoint set of assets that are not part of the portfolio. More precisely, we
argue that we can select the estimation parameters using cross validation based on the history of
other assets. A major advantage of applying cross validation for parameter selection is that we can
apply this method to estimate the future covariance according to different optimization objectives,
such as minimum variance or maximum risk-adjusted return. The historically optimal parameters
are subsequently used to estimate the actual covariance based on the given assets and datapoints.
The concept of “transfer learning” is well known in the context of machine learning (e.g., Bastani
2021, Pan and Yang 2009) or the pricing literature (e.g., Bastani et al. 2021). Yet, we are not aware
of any approach that applies the same idea to the problem of parameter selection in covariance
We present a novel covariance estimator based on non-linear shrinkage, which we call “Cross
Validation based Transfer Learning” (CVTL). The approach can be attributed to the group of
rotation equivariant non-linear shrinkage estimators. This means that the sample eigenvectors are
left unchanged, while the sample eigenvalues are shrunk, in our case, against the mean eigenvalue
and a second non-linear shrinkage target. The second shrinkage target is determined by a two-step
process. First, we scale the sample eigenvalue imbalance measured by their Gini coefficient to a
desired quantity. Second, we search for a Beta distribution, for which the corresponding cumulative
distribution function achieves the desired Gini coefficient. All parameters are selected by cross
validation to be historically optimal on a disjoint history with respect to the given objective.
We recommend the disjoint history to contain at least one additional year of trading data. This
allows for performing at least twelve train-evaluate iterations during cross validation with monthly
rebalancing to achieve a stable parameter configuration. The objective can be chosen freely and it
can be completely tailored towards the overarching portfolio optimization goals, e.g., minimizing
variance or maximizing risk-adjusted return. Therefore, our approach is purely data-driven and
agnostic with respect to the desired eigenvalue distribution and optimization objective. In case
of maximizing risk-adjusted return, our approach operates also on the covariance estimate only,
without having to estimate future returns. Instead, the weights are always calculated according to
the Markowitz global minimum variance portfolio (Markowitz 1952) but the eigenvalues are set to
be historically optimal under the desired objective function on the disjoint dataset.
Our numerical evaluation based on several global stock indices as well as aggregated industry
portfolios shows that CVTL is superior to established covariance estimators in minimizing variance
and maximizing risk-adjusted return. The latter holds in particular when additionally accounting
for transaction costs. The second shrinkage target is assigned a non-zero shrinkage intensity when
estimating high-dimensional covariance matrices where the number of assets is greater than the
number of datapoints. However, the intensity of the second shrinkage target decreases if the number
of datapoints increases relatively to the number of assets. Besides, we empirically compare the
resulting eigenspectra of CVTL with the analytically derived spectra from the non-linear shrinkage
approaches QuEST and BN. The results indicate a similar eigenspectrum, while CVTL tends to
increase instead of decrease the largest eigenvalue for high-dimensional covariance matrices.
1.2. Contributions
CVTL that combines cross validation and transfer learning. We show that transfer learning allows us
to replace the traditional loss functions based on theoretically derived priors of the oracle covariance
with a flexible objective function tailored to the overarching optimization problem. In contrast
to existing approaches, CVTL is agnostic towards the desired eigenspectrum as all estimation
parameters are selected based on a disjoint dataset. We further outline a novel non-linear shrinkage
target based on the cumulative distribution function of a Beta distribution to achieve a desired
Gini coefficient. The second shrinkage target is particularly useful for high-dimensional estimation
problems with N > T . In addition, we demonstrate the flexibility of CVTL by setting the objective
function towards maximizing risk-adjusted return. However, we do not explicitly estimate future
returns as we calculate the portfolio weights solely based on the formula of the GMV portfolio,
1.3. Outline
The remainder of this paper is structured as follows. Section 2 introduces the preliminaries on
methods. Section 3 presents our covariance estimator that applies cross validation based transfer
learning. In particular, it presents a detailed decription of how to select the shrinkage parameters.
Section 4 describes our empirical evaluation including the competing approaches and the datasets
used for parameter estimation and out-of-sample evaluation. Section 5 presents the results. We
first present the main results for the objectives minimum variance and maximum risk-adjusted
return. Second, we present several sensitivity analyses, where we show how varying the number of
cross validations and the parameter search space influences out-of-sample performance. Third, we
show how our approach is linked to established covariance estimators from an empirical perspective.
2. Preliminaries
Portfolio optimization aims at finding the optimal allocation of weights w = (w1 , . . . , wN ) for a
portfolio consisting of N assets. The global minimum variance (GMV) portfolio allocation minimizes
the portfolio risk (Markowitz 1952). Let Σ ∈ RN ×N denote the future covariance matrix and 1 the
vector consisting of N ones. The weights of the GMV portfolio are given by
Σ−1 1
w= . (1)
10 Σ−1 1
However, the future covariance matrix Σ is an unknown parameter that needs to be estimated
based on historic data. Let X T,N = x1 , . . . , xT with xt ∈ RN denote the sequence of historic returns
1
PT
used for covariance estimation. The average return vector is given as x̄ = T t=1 xt . The sample
The dimension D = N/T specifies the fraction between the number of assets and datapoints used for
covariance estimation. If D < 1 (more datapoints than assets), the estimation problem is considered
low-dimensional. Here, Σ
b is symmetric and positive semi-definite by construction. Conversely, if
2.1. Eigendecomposition
Eigendecomposition can be used to write the sample covariance matrix as the product of an
orthonormal matrix Λ ∈ RN ×N , where column i contains the eigenvector associated with eigenvalue
λi , and a diagonal matrix diag(λ1 , . . . , λN ) containing the eigenvalues in descending order λi ≥ λi+1
b = Λ diag(λ1 , . . . , λN ) Λ0 .
Σ (3)
The eigenvalues of the covariance matrix reflect the risk contribution of the portfolio constituents
(Roncalli and Weisang 2016). If eigenvalue λi is greater than λj , then the eigenvector Λi captures
more variance than the eigenvector Λj . The GMV weights (1) are calculated based on the inverse
For instance, it can easily be shown that setting all eigenvalues to the same value c ∈ R>0 results in
Shrinkage estimators present a common approach to improve the sample covariance estimate (e.g.,
Ledoit and Wolf 2003, 2004a, Frahm and Memmel 2010, Bodnar et al. 2018). Shrinkage approaches
Linear Shrinkage The idea of linear shrinkage (LS) is to combine the sample covariance matrix
b LS = (1 − δ ∗ ) Σ
Σ b + δ ∗ R. (6)
A frequent choice for R is the covariance matrix implied by the 1/N portfolio, which is given
by a diagonal matrix that contains the mean sample eigenvalue (Ledoit and Wolf 2004b, Frahm
and Memmel 2010). Such approaches are usually seen an application of “Stein-type” covariance
estimators (see a discussion in Ledoit and Wolf 2021a). The fundamental idea of “Stein-type”
estimators was originally derived to reduce mean squared error of multivariate mean estimation
by shrinking the estimate to a target vector (Stein 1956, James and Stein 1961). The shrinkage
intensity δ ∗ specifies how close the shrunk covariance matrix will be to the reference matrix R.
Setting δ ∗ to 0 yields the sample covariance, while setting δ ∗ to 1 yields the reference matrix R.
In the related literature, δ ∗ is often expressed as a function of D (e.g., Frahm and Memmel 2010,
Bodnar et al. 2018). As an alternative, δ ∗ can be derived by implicitly optimizing a loss function
(see Ledoit and Wolf (2021c) for a review on different loss functions), e.g., the Frobenius loss
P P
P P
2
LF c, =
c −
measured in units of risk between the sample and the reference portfolio
F
(e.g., Ledoit and Wolf 2003). Since LF cannot be explicitly calculated, researchers make assumptions
P P
about the characteristics of to minimize LF implicitly. Assumptions about are based, among
others, on an uninformed identity matrix (e.g., Ledoit and Wolf 2004b, Frahm and Memmel 2010,
Bodnar et al. 2018), an identity matrix with the average correlation on its off-diagonals (e.g., Ledoit
and Wolf 2004a) or an arbitrary factor structure (e.g., Ledoit and Wolf 2003).
based on the sample eigenvalues to obtain the shrunk eigenvalues λ∗1 , . . . , λ∗N . The N LS covariance
b N LS = Λ diag(λ∗1 , . . . , λ∗N ) Λ0 .
Σ (8)
f N LS (λ1 , . . . , λN ) = (1 − δ ∗ )λ + δ ∗ λθ , (9)
with non-linear shrinkage target λθ . The difference to linear shrinkage is that λθ is not a vector that
simply repeats a given value like the mean eigenvalue. Instead, the sample eigenvalues are shrunk
Recent approaches set the f N LS to the Stieltjes transform of the rescaled Marcenko Pastur
(MP) density (e.g., De Nard et al. 2019, Bun et al. 2017, Ledoit and Wolf 2012). Intuitively, this
distribution presents the limit eigenvalue distribution of a random covariance matrix with N → ∞.
Recently, Ledoit and Wolf (2021b) proposed another N LS approach that combines traditional
“Stein-type” shrinkage with MP-based estimators. The approach applies quadratic shrinkage with
two shrinkage targets, where the respective shrinkage intensity mainly depends on the dimension D.
The more data points are available for covariance estimation, the higher the intensity of “Stein-type”
Another common approach is to divide the sample eigenvalues into signal and noise regions
(Laloux et al. 1999, Plerou et al. 2002, Zhao et al. 2019). The function f N LS then rescales only
the noisy eigenvalues while preserving the signal eigenvalues. Similar approaches are known as
“thresholding” (Fan et al. 2013) where the noise eigenvalues are erased from the covariance estimate.
The concept of cross validation is well known in the machine learning literature (e.g., Pan and Yang
2009, Ban et al. 2018). The underlying idea is to evaluate multiple train/test splits of the available
data to select the optimal model parameters. By evaluating multiple train/test splits, the resulting
parameters are more robust than the parameters obtained by using a single train/test split (e.g.,
In covariance estimation, the shrinkage parameters like intensity and target could, in theory, also
be selected based on cross validation. This requires a long history of datapoints to be available to
perform a sufficient number of cross train-evaluate iterations. However, if there was a large amount
of data available for cross validation, the same data could also directly be used to estimate the
sample covariance matrix (2). The central limit theorem states that the sample covariance estimate
generally becomes better if the number of datapoints T increases. In particular for high-dimensional
estimation problems, the history of datapoints is smaller than the number of assets, so that there is
We propose an approach that applies cross validation based transfer learning to identify the
optimal shrinkage parameters. We solve the problem of limited data availability by using a sufficiently
long history of disjoint assets for parameter selection. The optimal parameters identified on the
disjoint dataset are then transferred to the actual covariance estimation problem.
the class of rotation equivariant non-linear shrinkage estimators. That is, we shrink the sample
eigenvalues, while not changing the eigenvectors. The main idea of our approach is to select the
shrinkage parameters based on a comparatively longer but disjoint history of other assets. The
approach is fully data-driven and agnostic towards achieving a particular shrinkage target or
intensity.
Let U = {1, . . . , U } denote the index set of all assets in the universe. Given a portfolio of N
assets with stock indices X ⊂ U , we need to estimate the future covariance matrix based on the
joint history X T,N . We now assume that there is at least one other index set V ⊂ U of size N with
V ∩ X = ∅ but longer joint history. Specifically, we assume that the history V contains at least
one additional year of trading data so that we can perform at least twelve cross validations for
parameter selection given monthly rebalancing (e.g., Ackermann et al. 2017, De Nard et al. 2019,
Zhao et al. 2019). The optimal shrinkage parameters are then transferred to the actual estimation
The main challenge of non-linear shrinkage is to find an optimal parameter setting which specifies
how to adjust the sample eigenvalues. Specifically, this requires one or more shrinkage targets and
We shrink the sample eigenvalues λ = λ1 , . . . , λN against two shrinkage targets (λ̄ and λθ )
λ∗ = δ1 λ + δ2 λ + (1 − δ1 − δ2 )λθ . (10)
The first shrinkage target λ is given by a vector of length N that simply repeats the average sample
1
PN
eigenvalue N i=1 λi as suggested by Ledoit and Wolf (2003). The shrinkage parameters δ1 , δ2 , and
λθ are selected using cross validation based transfer learning on a disjoint history V K,N .
The second shrinkage target λθ is calculated based on the imbalance of the sample eigenvalues.
We measure imbalance according to the Gini coefficient (Dorfman 1979). The values of the Gini
coefficient range between 0 and 1, where a value of 0 indicates perfect balance (λi = c ∈ R>0 ,
∀i = 1, . . . , N ), while a value of 1 indicates perfect imbalance (λ = (c, 0, . . . , 0)). We prefer the Gini
over other measures of imbalance such as entropy as the Gini coefficient is limited to the interval
[0, 1]. Since the Gini coefficient requires λi ≥ 0 for all eigenvalues, we set all negative eigenvalues to
We define G Σ
b as the Gini coefficient of the eigenvalues of a sample covariance estimate Σ.b
Given G Σ b , we first determine the desired Gini coefficient Gθ for the second shrinkage target.
Subsequently, we search for a cumulative Beta distribution that achieves the desired Gini coefficient.
Thereby, we are flexible in achieving almost any possible eigenspectrum in the non-linear shrinkage
target.
Given the desired Gini coefficient Gθ , we generate artificial eigenvalues λθ so that G(λθ ) ≈ Gθ .
For this purpose, we follow (e.g., Ledoit and Wolf 2012, 2017a) in using the cumulative distribution
function (CDF) of a Beta distribution with parameters α, β. The main argument for a Beta
distribution is its bounded support and easily adjustable shape for which Ledoit and Wolf (2012)
calls it the ”best suited family” of distribution for such purpose. The eigenvalues are then calculated
as the values of CDFα,β over an equally spaced grid over the interval [0, 1]. This yields the following
optimization problem
We optimize α over the set {1, 2, 3, 4, 5, 10, 20, . . . , 100, 150} while setting β = 1. Thereby, we can
generate a large variety of eigenvalue distributions with different Gini coefficients as suggested by
the optimal parameter setting found through cross validation. To provide an intuition, Figure 1
shows the CDFs of three Beta distributions and the corresponding Gini coefficients for N = 100.
Figure 1 Cumulative Distribution Functions of Different Beta Distributions and Corresponding Gini Coefficients
(a) α = 2, β = 1, G(λ2,1 ) = 0.502 (b) α = 5, β = 1, G(λ5,1 ) = 0.716 (c) α = 10, β = 1, G(λ10,1 ) = 0.834
We normalize the resulting eigenvalues λα,β from (13) in order to keep the trace of the sample
1
PN
θ α,β N i=1 λi
λ =λ . (15)
||λ ||1
α,β
This step is necessary in order to ensure that the second shrinkage target has a reasonable influence
∗
,β ∗
on the resulting shrunk eigenvalues. If the artificial eigenvalues λα are all much larger than their
pendants in the sample eigenvalues λ, the shrinkage operation is strongly dominated by λα,β . Note
that normalization does not alter the Gini coefficient (i.e., G(λα,β ) = G(λθ )) as the Gini coefficient
We need to select a total of three parameters, namely the adjustment factor γ for the Gini coefficient
and the shrinkage intensities δ1 , δ2 . The parameters are chosen such that the resulting weights
w(γ, δ1 , δ2 ) are historically optimal on a disjoint history V K,N according to the given objective
φ(w) (e.g., minimize variance or maximize risk-adjusted return). The disjoint set of asset V can be
9: λ∗ = δ1 λ + δ2 λ̄ + (1 − δ1 − δ2 )λθ
∗ ∗ 0
10: Σ = Λ λ Λ
11: output: Σ∗ covariance estimate
selected by randomly sampling indices from U \ X , as done in this study. As an alternative, one
could attempt to specifically select similar portfolio constituents, e.g., from a similar stock market
We assume monthly rebalancing so that each portfolio allocation is held for ρ = 21 trading days
(e.g., Ackermann et al. 2017, De Nard et al. 2019, Zhao et al. 2019). Our empirical results imply
that the parameter setting becomes stable after approximately twelve train-evaluate iterations for
cross validation (see EC.2.2). Hence, we recommend to use a disjoint history V K,N containing at
least another trading year (i.e., 12 · 21 = 252 days) of data so that K ≥ T + 252 and applied the
The process of cross validation is described by Algorithm 2. Without loss of generality, we assume
that φ(w) needs to be minimized. For each parameter configuration γ, δ1 , δ2 , we first check if the
not invertible, the parameter configuration is discarded. Otherwise, we perform time-series cross
validation with monthly rebalancing to calculate the historical performance of γ, δ1 , δ2 under the
given objective. Let the disjoint history V K,N = v1 , . . . , vK be sorted from the oldest (v1 ) to the
most recent datapoint (vK ). We first estimate a covariance matrix with the given parameters based
on v1 , . . . , vT (see Algorithm 1). The resulting covariance matrix is then evaluated based on the
periods by ρ days towards the future. Accordingly, the second covariance matrix is estimated
based on vρ+1 , . . . , vρ+1+T and its performance is measured on the second out-of-sample period
vρ+T +2 , . . . , v2ρ+T +2 . This continues until the most recent datapoint vK becomes part of the last
out-of-sample period. We use n to denote the number of train-evaluate iterations that are performed
during cross validation (it is ensured that n ≥ 12). The performance of the current parameter
configuration is then aggregated as the mean performance over all evaluated out-of-sample periods.
For the purpose of this study, we select the estimation parameters from the following search
space P . We later provide a sensitivity analysis to assess the influence of a more fine-grained grid.
We include an additional filter ψ(γ, δ1 , δ2 ) → {true, f alse}, which allows us to limit the search
3.3. Objectives
The objective function φ(w) is evaluated on all out-of-sample periods that occur during cross
validation. Let Σi and v̄i denote the covariance matrix and mean return vector of the out-of-sample
period in the ith train-evaluate iteration during cross validation. To minimize variance, the objective
φ(w) = w0 Σi w. (18)
v̄i0 w
φ(w) = . (19)
w0 Σi w
Note that the weights are always calculated according to the formula of the GMV portfolio (1)
for any given objective. Existence of the inverse covariance matrix is guaranteed by filtering the
estimation parameters accordingly. In particular, our approach does not require us to estimate future
returns. Instead, we select an eigenvalue distribution and shrinkage intensities to explicitly optimize
the given objective φ(w) for the disjoint history. The same parameters should then implicitly
optimize φ(w) in the future investment period. This approach has – to the best of our knowledge –
When optimizing for risk-adjusted return (19), we observed overfitting in terms of heavily
dispersed portfolio weights. Specifically, the resulting portfolios exhibit high out-of-sample volatility
combined with higher out-of-sample returns, or, the opposite; low out-of-sample returns paired
with low out-of-sample volatility, while not achieving a stable configuration. Therefore, we limit
the search space to improve the stability of the parameter configurations. For this purpose, we
calculate the optimal η ∗ ∈ [0, 1] that maximizes risk-adjusted return as a linear combination between
the weights of the sample covariance wGM V and the 1/N portfolio w1/N using the disjoint transfer
the parameter ζ. We set ζ = 0.30, however, the performance is not sensitive to small changes in ζ
as we later show in our sensitivity analyses (see table 9). We then calculate the acceptance interval
Ψσ in terms of the out-of-sample variance of the portfolios wa and wb achieved by combining the
We define σ(γ, δ1 , δ2 ) as the mean variance of the given parameter configuration over all periods
Finally, we define the acceptance function to filter for parameters γ, δ1 , δ2 so that the resulting
Thereby, we avoid the aforementioned problem of achieving portfolios with overly low or high
out-of-sample variance, which leads to more stable parameter estimates. Note that ψ can be defined
in regard to arbitrary objectives. For instance, one could control transaction cost by only accepting
portfolios that do not exceed a given turnover during cross validation or incorporate specific goals
and behavioral theory of individual investors directly into the covariance matrix as suggested in
Altogether, we consider the following approaches for the remainder of this study
• CVTLLS. CVTL with objective minimum variance (18). We require δ1 + δ2 = 1 to exclude the
CVTL is related to several existing estimators like linear-shrinkage (e.g., Ledoit and Wolf 2003,
2004b,a). When setting δ1 + δ2 = 1, our approach (CVTLLS) has the same shrinkage target as Ledoit
and Wolf (2004b). Similarly, setting δ2 = 1 yields the 1/N portfolio. The intuition behind CVTL is to
adjust the sample eigenvalues to be historically optimal under a given objective φ(w) on the disjoint
history. Thereby, CVTL implicitly performs the following steps: (i) estimating the covariance matrix,
(ii) inverting the covariance matrix, (iii) optimizing the portfolio towards the given objective. This
can be interpreted as a flexible extension of the minimum variance loss suggested by (Engle et al.
2019), where the oracle estimator is derived by using cross validation. However, the difference is that
we replace the statistically derived prior for the oracle eigenvalues by Engle et al. (2019) with the
objective function based on the disjoint history. Furthermore, CVTL is related to the bounded-noise
estimator by Zhao et al. (2019) in the sense that the covariance adjustment is entirely driven by
the given objective. However, CVTL differs from the bounded-noise approach as it does not rely on
Figure 2 Conceptual Difference between Established Non-Linear Shrinkage Methods (left, dark gray) and
Cross validation
Σ∗ = Λ diag(λ∗ ) Λ0
The conceptual difference between established non-linear shrinkage estimators and CVTL is
illustrated in Figure 2. While existing non-linear shrinkage methods (e.g., Ledoit and Wolf 2015, Bun
et al. 2017, Ledoit and Wolf 2020, 2021b), select the shrinkage parameters to implicitly minimize a
history according to a given objective φ(w). CVTL uses cumulative Beta distribution functions to
generate the shrinkage target. However, setting δ1 + δ2 = 0 does not lead to the same non-linear
shrinkage as those of Ledoit and Wolf (e.g., 2015, 2021b). Instead, we could achieve a comparable
estimator by replacing the second shrinkage target of CVTL with the non-linear function presented
in Ledoit and Wolf (2015). Using the Gini coefficient as a measure of imbalance to generate the
second (non-linear) shrinkage target is related to linear shrinkage to constant correlation (Ledoit
and Wolf 2004a). The higher the average correlation in the sample covariance, the higher the Gini
4. Evaluation Method
We evaluate CVTL according to the objectives minimum variance and maximum risk-adjusted
return against established covariance estimators from the literature. An overview of all considered
BN VAR BN estimator with objective maximum risk-adjusted return (Zhao et al. 2019)
NC2R Generalized unconstrained partial min-var portfolio trained on return (DeMiguel et al. 2009a)
of prior period
CT Combined Talmud estimator: Combination of 50% sample covariance (Tu and Zhou 2011)
GMV and 50% 1/N portfolio
1/N 1/N portfolio with equal weights (DeMiguel et al. 2009b)
Dataset Our evaluation is based on six different datasets. First, we use three stock datasets
from Bloomberg: (i) all US listed stocks on the NYSE, AMEX, and NASDAQ (US ), (ii) European
stocks listed in the Stoxx 600 Europe index (EU ), (iii) a geographic mixture of stocks listed in the
MSCI World index (WO). Second, we use three publicly available aggregated industry portfolios
with (iv) 10, (v) 30, and (vi) 49 industries from (French 2021) (FFI ). All datasets span the time
period from January 3, 2000 – December 31, 2020. We delete all common NaN data points, e.g.,
weekends, national holidays and other non-trading days. For each dataset, the respective disjoint
history is randomly sampled from the union of the US, EU and WO datasets.
predictions, where the covariance matrix is estimated based on daily returns. We apply monthly
portfolio rebalancing to keep the turnover low (e.g., De Nard et al. 2019). We analyze low- and
N
high-dimensional covariance estimation problems with the dimensions D = T
∈ {2, 32 = 1.50, 34 ≈
Each result is reported as the mean over 50 evaluations. In each evaluation, we randomly select
N ∈ {100, 200, 300} portfolio constituents from the respective dataset. Industry portfolio results
are based on the full dataset with N ∈ {10, 30, 49} industry constituents. For both datasets, we
calculate the following performance metrics as the mean over all occurring out-of-sample periods
Performance Metrics We consider two objectives, namely, minimizing variance and maximizing
return as the annualized return divided by the annualized volatility. In addition, we provide the
turnover and risk-adjusted return after transaction costs assuming 0.25 % per trade (Thapa and
Poshakwale 2010). Additional performance metrics are provided in EC.4. We perform two-sided
t-tests with α = 0.05 to check whether the performance of CVTL is significantly different from
existing estimators.
5. Results
5.1. Minimum Variance
We first consider the results for the objective minimum variance as shown in Table 2. We present
the out-of-sample volatility for estimation problems with D = 43 ≈ 1.33 and D = 0.40. We also show
the results of the estimators with objective maximum risk-adjusted return as a high risk-adjusted
return may also imply low volatility. The volatility of the best approach per setting is highlighted
in bold. Underlined values indicate that the respective approach is significantly outperformed with
p < 0.05 by CVTL. The results show that CVTL and CVTLLS consistently achieve the lowest
volatility among competing estimators independent from the dataset and portfolio size. CVTL
outperforms CVTLLS in most settings. The relative advantage of the second shrinkage target in
CVTL is stronger in the high-dimensional estimation problems. Furthermore, we find that QuEST
leads consistently to the lowest volatility among the competing covariance estimators.
Note: Results of all evaluated covariance estimators for US, EU and WO data with N ∈ {100, 200, 300}. Each value is
given in percent as the average over 50 evaluations of random portfolio constituents. Out-of-sample volatility greater
than 50 percent due to estimation errors is denoted by “–”. The best estimator per problem setting is highlighted in
bold. Underlined values indicate significant differences from CVTL with p < 0.05.
We further assess the results for out-of-sample volatility with respect to different dimensions
D = N/T ∈ {2, 1.50, 1.33, 0.40, 0.20, 0.13} of the sample data X. The results for the US dataset are
shown in Table 3. The results for all other datasets can be found in EC.4. Again, CVTL generally
outperforms the competing estimators. The difference between CVTL and CVTLLS is significantly
larger in the high-dimensional problems with D > 1. However, for the low-dimensional problems
with D < 1, the benefit of the second shrinkage target diminishes and both CVTL models achieve
Note: Results of all evaluated covariance estimators for US data with N ∈ {100, 200, 300} and D ∈
{2, 1.50, 1.33, 0.40, 0.20, 0.13}. Each value is given in percent as the average over 50 evaluations of random portfolio
constituents. Out-of-sample volatility greater than 50 percent due to estimation errors is denoted by “–”. The best
estimator per problem setting is highlighted bold. Underlined values indicate significant differences from CVTL with
p < 0.05.
Regarding the competing estimators, we observe a general pattern that the achieved volalities
are less diverged for D < 1. We find that QuEST (Ledoit and Wolf 2015) even presents the overall
dominant approach for a total of three low-dimensional problems with N = 200, D = 0.13 and
N = 300, D ∈ {0.20, 0.13}. Other estimators that rely on the weights given by the sample covariance
matrix, namely FMEst and CT, show significantly worse performance for high D up to the point
where the resulting portfolio yields an out-of-sample volatility of above 50 percent, which is denoted
by “–”.
Interestingly, we find that CVTL, QuEST, and QIS (Ledoit and Wolf 2021b) achieve comparably
lower variances for higher than for lower dimensions D. This seems surprising, as greater D implies
a more deteriorated covariance matrix with greater imbalance in eigenvalues, and, thus, a more
challenging estimation problem. However, this difference in volatility can be explained as the
number and timeframe of the individual out-of-sample periods in the rolling window evaluations
are different. For instance, the setting D = 0.13 and N = 300 requires 2250 datapoints, or ten years
of return history to estimate the first covariance matrix. All of these datapoints are hence not used
as out-of-sample periods. Conversely, the setting D = 2 and N = 300 requires only 150 datapoints
to estimate the first covariance matrix. Accordingly, the high-dimensional evaluation is based on
(a) Annualized Volatility vs. Annualized Return (b) Annualized Volatility vs. Risk-Adjusted Return
Note: Visualization of out-of-sample annualized volatility, annualized return, and risk-adjusted for all covariance
estimators based on US data with N = 100 and D = 0.20. Each value is given in percent as the average over 50
simulations with random portfolio constituents. The area under the frontier created by CVTL provides additional
diversification benefits.
Figure 3a illustrates the out-of-sample performance of each estimator along the efficient frontier
in a return versus volatility diagram for N = 100 and D = 0.20. In addition to our prior findings,
we observe that CVTL leads to comparable or better returns than most competing estimators. As
a result, it shifts the efficient frontier , which traditionally starts at the sample portfolio, to the
We also consider the results for out-of-sample volatility based on small aggregated industry
portfolios as done by Zhao et al. (2019). Note that these datasets are publicly available (see French
2021), which facilitates reproducibility of our results. Table 4 presents the results for volatility for
N ∈ {10, 30, 49} industry portfolios. CVTL now achieves only average results if the portfolio is very
small, with N = 10. However, both CVTL approaches become superior to the competing approaches
for larger potfolios with N > 10. For N = 10, POET I achieves the lowest volatility. Besides, we
find similar patterns for the otherwise superior QuEST model, which is outperformed by linear
shrinkage for N < 49. This was already suggested by Ledoit and Wolf (2021a) and is mainly due to
Table 4 Out-Of-Sample Annualized Volatility for FFI Data and Different Dimensions.
N 10 30 49
D = N/T 2 1.50 1.33 0.40 0.20 0.13 2 1.50 1.33 0.40 0.20 0.13 2 1.50 1.33 0.40 0.20 0.13
Note: Results of all evaluated covariance estimators for FFI industry portfolios with N ∈ {10, 30, 49} and
D ∈ {2, 1.50, 1.33, 0.40, 0.20, 0.13}. Out-of-sample volatility greater than 50 percent due to estimation errors is denoted
by “–”. The best estimator per problem setting is highlighted in bold. Underlined values indicate significant differences
from CVTL with p < 0.05.
Next, we consider all covariance estimators based on the objective maximum risk-adjusted return.
Table 5 presents the out-of-sample risk-adjusted return as annual return divided by annual volatility.
Again, we consider all estimators, including the estimators that were specifically designed to
minimize volatility as a high risk-adjusted return may also be achieved through low volatility. In
this case, underlined values indicate that the respective approach is significantly outperformed with
Note: Results of all evaluated covariance estimators for US, EU and WO data with N ∈ {100, 200, 300}. Each value is
given in percent as the average over 50 evaluations of random portfolio constituents. Out-of-sample volatility greater
than 50 percent due to estimation errors is denoted by “–”. The best estimator per problem setting is highlighted in
bold. Underlined values indicate significant differences from CVTL µ with p < 0.05.
σ
The results do not suggest a dominating estimator, which is somewhat expected as the estimation
of returns is usually more unstable than the estimation of risk (Michaud 1989). However, we observe
that CVTL σµ and CVTLLS σµ usually perform among the best models in the considered low- and high-
dimensional settings. The highest risk-adjusted return in a few settings (D = 1, 33, N = 200, EU and
WO dataset) is achieved by CVTL and CVTLLS, albeit these approaches are primarily minimizing
problem with N = 100 on the MSCI World (WO) dataset. In addition, we find that CVTL and
CVTL σµ achieve similar risk-adjusted returns even though they were designed to optimize different
objectives. There is only one setting where CVTL or CVTL σµ do not outperform the approaches
BN VAR and NC2R, which are both designed to maximize risk-adjusted return.
Following our prior analyses, we also assess the results for risk-adjusted return based on the US
dataset and dimensions D = N/T ∈ {2, 1.50, 1.33, 0.40, 0.20, 0.13}. The results are shown in Table 6.
Here, we observe that CVTL σµ and CVTLLS σµ achieve the highest risk-adjusted return in most
settings.
Note: Results of all evaluated covariance estimators for US data with N ∈ {100, 200, 300} and D ∈
{2, 1.50, 1.33, 0.40, 0.20, 0.13}. Each value is given in percent as the average over 50 evaluations of random portfolio
constituents. Out-of-sample volatility greater than 50 percent due to estimation errors is denoted by “–”. The best
estimator per problem setting is highlighted in bold. Underlined values indicate significant differences from CVTL µ
σ
with p < 0.05.
Figure 3b illustrates the out-of-sample performance of each estimator along the efficient frontier
in a risk-adjusted return vs volatility diagram for N = 100 and D = 0.20. We observe that, by
introducing CVTL σµ , we obtain an improved efficient frontier between CVTL, CVTL σµ and the 1/N
portfolio that dominates existing estimators. Importantly, as shown by Figure 3a, the increase in
out-of-sample risk-adjusted return in CVTL σµ is not caused by strategic scaling, e.g., decreasing
variance towards zero to improve risk-adjusted return. Instead, CVTL σµ exhibits comparable, but
improved, levels of absolute risk and return than the traditional efficient frontier.
The results for maximum risk-adjusted return on the FFI industry portfolios are shown in Table 7.
The results differ from those of our prior analyses on risk-adjusted return based on single stock data.
For small portfolios with N = 10, POET II leads to the highest risk-adjusted return. For N = 49,
linear shrinkage to constant correlation (LinShriCC) dominates all other estimators. The CVTL
estimators perform weakly, which seems to be caused by the small portfolio sizes, in particular for
N < 49.
Table 7 Out-Of-Sample Risk-Adjusted Return for FFI Data and Different Dimensions.
N 10 30 49
D = N/T 2 1.50 1.33 0.40 0.20 0.13 2 1.50 1.33 0.40 0.20 0.13 2 1.50 1.33 0.40 0.20 0.13
Estimators with Objective Minimum Variance
CVTL 0.63 0.77 0.62 0.68 0.89 0.96 0.78 0.82 0.83 0.84 0.94 0.92 0.75 0.81 0.88 0.89 0.73 0.95
CVTLLS 0.66 0.78 0.64 0.70 0.94 0.98 0.78 0.85 0.84 0.86 0.93 0.93 0.77 0.86 0.88 0.90 0.72 0.93
QIS 0.85 0.92 0.98 0.88 1.07 1.04 0.59 0.86 0.81 0.80 0.87 0.90 0.82 0.81 0.71 0.73 0.64 0.85
QuEST 0.92 0.94 0.92 0.91 1.11 1.05 0.62 0.82 0.76 0.79 0.86 0.91 0.79 0.78 0.71 0.78 0.65 0.85
LShriCC 0.79 0.84 0.90 0.87 0.95 0.87 0.71 0.65 0.65 0.79 0.90 0.93 0.85 0.99 0.96 0.85 0.71 0.91
LShri 0.68 0.82 0.76 0.82 1.04 1.04 0.73 0.83 0.77 0.83 0.89 0.91 0.76 0.72 0.75 0.81 0.67 0.88
BPSEst 0.62 0.72 0.69 0.81 1.05 1.02 0.50 0.82 0.72 0.78 0.85 0.87 0.79 0.74 0.80 0.59 0.60 0.81
FMEst – 0.68 – 0.84 1.08 1.04 – – – 0.74 0.85 0.88 – – – 0.55 0.60 0.81
POET I 0.67 0.89 0.89 0.85 1.02 1.01 0.78 0.80 0.72 0.60 0.83 0.86 0.36 0.65 0.16 0.72 0.62 0.85
POET II 1.16 1.30 1.27 0.99 1.11 0.96 0.61 0.84 0.78 0.72 0.79 0.81 0.82 0.75 0.79 0.87 0.63 0.74
BN 0.78 0.84 0.71 0.86 1.09 1.06 0.70 0.84 0.84 0.78 0.85 0.89 0.67 0.84 0.80 0.79 0.63 0.91
Sample – – – 0.83 1.11 1.06 – – – 0.68 0.86 0.89 – – – 0.50 0.60 0.79
Note: Results of all evaluated covariance estimators for FFI industry portfolios with N ∈ {10, 30, 49} and
D ∈ {2, 1.50, 1.33, 0.40, 0.20, 0.13}. Out-of-sample volatility greater than 50 percent due to estimation errors is denoted
by “–”. The best estimator per problem setting is highlighted in bold. Underlined values indicate significant differences
from CVTL µ with p < 0.05.
σ
We also consider out-of-sample risk-adjusted return after transaction costs (0.25 percent per
trade (Thapa and Poshakwale 2010)), for which we briefly summarize the findings. Detailed results
are provided in EC.1. We first consider the turnover rates (i.e., the sum of weight changes during
the evaluation period) as they are directly influencing transaction costs. Here, we find that CVTL
usually achieves the lowest turnover of all estimators with the objective minimizing risk. Among the
estimators with objective maximum risk-adjusted return, CVTL σµ most often achieves the lowest
turnover rates. The only estimator that shows lower turnover for N = 300 is N2CR. This can be
expected as N2CR employs weight constraints. Regarding risk-adjusted return after transaction
costs, we can confirm the results of Kourtis (2015), which suggest that it is hard to outperform
the 1/N portfolio on a risk-adjusted basis after accounting for transaction costs. However, CVTL σµ
and CVTLLS σµ are the only estimators that outperform the 1/N portfolio in several settings. The
results about risk-adjusted return after costs for other datasets are provided in EC.4.
We now assess the sensitivity of CVTL in regard to the cross validation parameters. Specifically,
we consider a higher number of disjoint histories used in cross validation and different sizes of
the search grid for the covariance estimation parameters γ, δ1 , δ2 . We analyze the sensitivity for
D ∈ {0.40, 1.33} and portfolio size N = 100 based on the US dataset. In addition, we analyze the
influence of the parameter ζ that is used to define the function (24) that reduces the parameter
In our main analyses, we used only one transfer dataset (i.e., one disjoint history) for cross valida-
tion and the parameter grid from (17). In this analysis, we evaluate the difference in performance
for 1, . . . , 5 transfer sets and two additional parameter grids. We consider a smaller grid Psmall
1
γ ∈ {−1, − , . . . , 1} (25)
3
δ1 , δ2 ∈ {0, 0.1, . . . , 1} with δ1 + δ2 ≤ 1 (26)
Table 8 shows the difference in basis points (bps, 1 basis point = 0.01 %) for the particular
performance metrics between the modified parameter setting and the reference setting (i.e., one
transfer set and the medium-sized grid (17)). The scores are calculated in a way that positive values
indicate an improvement, while negative values indicate a decline in the respective performance
metric. For the objective minimum variance, increasing the number of transfer sets has a small
positive effect as it reduces out-of-sample volatility by one bp. Increasing the grid size from medium
to large reduces out-of-sample volatility by 5 bps in the high-dimensional setting and by 8 bps in
the low-dimensional setting. Conversely, reducing the grid size from medium to small increases
low-dimensional setting. Interestingly, we find that increasing the grid and the number of transfer
sets simultaneously leads to higher out-of-sample volatility than using only one transfer set in
Table 8 Sensitivity of CVTL Performance Towards Different Search Grids and Number of Transfer Sets.
High-Dimensional Setting (D = 1.33) Low-Dimensional Setting (D = 0.40)
Search Small Medium Large Small Medium Large
grid
Transfer
Objective Minimum Variance (CVTL)
sets
1 −5 – 5 −8 – 8
2 −5 1 0 −6 2 4
3 −5 1 0 −6 2 4
4 −4 1 0 −6 2 4
5 −5 1 0 −6 2 4
Transfer
Objective Risk-Adjusted Return (CVTL µ )
sets σ
1 2 – −3 8 – −8
2 4 1 −2 7 2 −5
3 4 1 0 9 2 −6
4 3 0 0 11 2 −5
5 4 2 0 13 3 −5
Note: Scores are given in basis points. Positive values indicate improvements, while negative values indicate declines in
the respective performance metric. The reference setting is given by a medium-sized grid and one transfer set for cross
validation.
For CVTL σµ with the objective maximum risk-adjusted return, increasing the number of transfer
sets has a greater positive effect of up to 2 bps in the low-dimensional and 3 bps in the high-
dimensional setting. However, increasing the grid size from medium to large decreases the risk-
adjusted return by up to 8 bps in the high-dimensional setting. In particular, we find that it is best
to decrease the grid from medium to small as this increases out-of-sample risk-adjusted return by up
to 4 bps in the low-dimensional and 13 bps in the high-dimensional setting. These findings suggest
that using a very fine-grained grid leads to overfitting when optimizing the covariance estimate to
maximize risk-adjusted return. Instead, it is beneficial to use a more coarse grid to achieve better
We also analyze the influence of the ζ parameter that is used to decide whether or not a given
main analysis, we set ζ = 0.30. We now analyze how out-of-sample risk-adjusted return changes for
different values ζ ∈ {0.10, 0.20, 0.30, 0.40, 0.50} and if we remove the search space reduction. The
results are shown in Table 9. Evidently, it is generally beneficial to filter the parameter configurations
so that the resulting portfolios exhibit non-extreme volatility in cross validation. Note that not
filtering the parameter configurations can also lead to impressively good results, e.g.for N = 200 and
D = 0.13, however, we believe that such scores only appeared by chance and will not be reproducible
in the future. In addition, we find that risk-adjustred return changes only marginally for different
values of ζ. The difference in risk-adjusted return across different values of ζ is a maximum of 6 bps
ζ = 0.10 1.08 1.12 1.00 0.86 1.15 0.82 1.28 1.44 1.34 1.26 0.79 0.81 1.50 1.30 1.20 0.97 0.79 1.44
ζ = 0.20 1.08 1.12 1.01 0.88 1.13 0.84 1.27 1.44 1.38 1.25 0.80 0.82 1.51 1.34 1.25 1.05 0.84 1.50
ζ = 0.30 1.08 1.10 1.00 0.89 1.11 0.84 1.27 1.46 1.39 1.24 0.82 0.83 1.51 1.34 1.26 1.09 0.87 1.52
ζ = 0.40 1.09 1.10 0.99 0.89 1.08 0.84 1.30 1.45 1.36 1.24 0.83 0.85 1.51 1.35 1.25 1.11 0.85 1.54
ζ = 0.50 1.06 1.10 0.99 0.88 1.06 0.84 1.29 1.45 1.36 1.22 0.83 0.86 1.51 1.34 1.23 1.11 0.86 1.54
0.71 1.09 0.73 0.87 1.02 0.84 1.30 1.44 1.32 1.15 0.79 2.74 1.47 1.34 1.28 1.12 0.73 0.77
Note: Results of all evaluated covariance estimators for US data with N ∈ {100, 200, 300} and D ∈
{2, 1.50, 1.33, 0.40, 0.20, 0.13}. Each value is given in percent as the average over 50 evaluations of random portfolio
constituents. The best estimator per problem setting is highlighted in bold.
Finally, we analyze the resulting estimation parameters and eigenvalue distributions from an
empirical point of view. In addition, we compare the resulting eigenvalue distributions with the
established non-linear shrinkage estimator QuEST (Ledoit and Wolf 2015) and BN (Zhao et al.
2019).
γ, δ1 , δ2 that are used to estimate the future covariance based on sample data. Table 10 provides
the average parameters of CVTL and CVTL σµ for US data with N = 100 and different dimensions.
We observe that δ3 is considerably higher in the high-dimensional settings with D > 1, which is
consistent with the recent results by Ledoit and Wolf (2021b) who also found greater usage of
the non-linear shrinkage target for D > 1. For the low-dimensional problems, δ3 decreases and δ1
increases, which implies a greater weight of the sample eigenvalues. We further observe a decreasing
scaling parameter γ for increasing D. The parameters of CVTL σµ show similar patterns. The weight
of the non-linear shrinkage target increases for higher D, while the weight of the sample eigenvalues
increases for lower D. Additional empirical observations about the most frequent parameters for
larger portfolios, parameter values over time, and parameter interdependence are provided in EC.2.
Table 10 Average Parameter Calibration for CVTL Estimators and Different Sample Dimensions.
CVTL CVTL µ
σ
D δ1 δ2 δ3 γ δ1 δ2 δ3 γ
2.00 20.9 26.6 52.4 88.6 23.2 43.0 33.8 40.8
1.50 30.7 32.9 36.4 90.9 24.0 47.5 28.5 59.0
1.33 36.5 36.7 26.8 91.9 28.6 46.4 25.0 42.1
0.40 60.6 27.5 11.9 84.8 33.5 45.8 20.7 30.0
0.20 68.8 19.5 11.7 81.8 33.7 47.0 19.3 11.4
0.13 69.4 17.7 12.9 69.0 29.3 45.3 25.5 35.8
Note: Average parameter calibration for US data with N = 100 and D ∈ {2, 1.50, 1.33, 0.40, 0.20, 0.13}. Each score is
given in percent as the average over 50 evaluations with random portfolio constituents. δ3 = 1 − δ1 − δ2 .
QuEST (Ledoit and Wolf 2015) and BN (Zhao et al. 2019). Figure 4 plots the shrunk eigenvalues
compared to the sample eigenvalues for CVTL (black cross), QuEST (dark gray circle) and BN
(gray triangles). The plotted eigenvalues denote the average over 50 evaluations with N = 100 based
on the US data.
Note: The plotted eigenvalues are averaged over 50 simulations with random portfolio constituents based on the US
We make several observations regarding the shrunk eigenvalues of CVTL. First, CVTL shrinks the
eigenvalues in a clearly non-linear way. Second, CVTL shrinks small eigenvalues in a similar fashion
to QuEST, especially for lower dimensions. This is surprising given that CVTL does not induce any
statistically derived lower limit for small eigenvalues. Accordingly, this supports the use of cross
validation for parameter selection as it leads to comparable results like statistically derived methods.
Third, CVTL does not reduce the largest eigenvalue for all dimensions. Instead, it increases the
first eigenvalue in the high dimensional problems. As a result, the weight directions of the first
eigenvalue will be suppressed to a stronger extent compared to other estimators. This indicates
that it is not always necessary to decrease the first eigenvalue. Independent of the dimension, the
eigenvalues achieved by BN deviate from QuEST and CVTL, which can be expected due to the
separation of signal and noise. The same analysis for CVTL σµ is provided in EC.2.3.
analysis to estimate the isolated effects of using cross validation for parameter selection and the
benefit of the proposed second shrinkage target. For this purpose, we generate a new covariance
estimator called “QuESTCV” that uses the shrinkage target by QuEST (Ledoit and Wolf 2015)
but cross validation to select the shrinkage intensity. This allows us to calculate the individual
benefits of (i) cross validation for selection of shrinkage intensities as the difference in bps between
QuESTCV and QuEST, and (ii) the proposed second shrinkage target λθ as the difference in bps
between CVTL and QuESTCV. Table 11 presents the results for the objectives minimum variance
(we only consider the objective minimum variance as QuEST specifically minimizes volatility). For
instance, the value 7 in the row for cross validation for N = 100 and D = 0.40 indicates that using
cross validation to select the shrinkage intensity of QuEST instead of applying the original QuEST
Table 11 Individual Benefits of Cross Validation and Second Shrinkage Target for US data.
N 100 200 300
D = N/T 2 1.50 1.33 0.40 0.20 0.13 2 1.50 1.33 0.40 0.20 0.13 2 1.50 1.33 0.40 0.20 0.13
Cross −1 0 1 7 6 3 −2 −1 2 6 3 1 1 2 3 5 3 0
validation
Shrinkage 7 5 3 −1 0 −3 4 5 4 0 −1 −3 2 1 1 −1 −3 −1
target λθ
Note: CV denotes the benefit of cross validation as the difference in basis points when using cross validation to select
the shrinkage intensity of QuEST (Ledoit and Wolf 2015) instead of applying the original QuEST estimator. λθ
denotes benefit of the proposed second shrinkage target as the difference in basis points when using CVTL instead
of QuEST with cross validation to select the shrinkage intensity. Each score denotes the difference in out-of-sample
volatility basis points, e.g., 1 denotes the difference 10.00% vs. 10.01%. Positive values indicate an improvement in the
respective performance metric. Negative values indicate a decline.
We find that using cross validation for selecting shrinkage intensities has the greatest benefit
for low-dimensional problems, while the benefit for high-dimensional problems is smaller or even
negative. Conversely, using the second shrinkage target of CVTL instead of the non-linear shrinkage
target by QuEST reduces out-of-sample volatility in particular for the high-dimensional problems,
while it often has a negative effect for low-dimensional problems. This supports the recent idea of
Ledoit and Wolf (2021b) to shrink the eigenvalues quadratically for low-dimensional problems using
a non-linear and a linear shrinkage target. Additional analyses about the benefits of cross validation
6. Concluding Remarks
We proposed a novel approach for covariance estimation that applies cross validation based transfer
learning called “CVTL”. The proposed approach applies non-linear shrinkage to the eigenvalues of
the sample covariance according to a given objective function. In contrast to existing methods, which
generally rely on analytically derived values for the estimation parameters, CVTL is purely data-
driven and agnostic with respect to the resulting eigenvalue distribution and shrinkage intensities.
All estimation parameters are selected using cross validation and the given objective based on a
disjoint history of assets. The resulting parameters are subsequently used to estimate the actual
covariance matrix of the given portfolio constituents. Thereby, our study presents a novel perspective
Although CVTL is purely data-driven, the resulting shrunk eigenvalues are similar to those
problems with objective minimum variance, where CVTL tends to increase the largest eigenvalue
which presents the exact opposite of existing estimators. This is surprising as, so far, the underlying
idea of non-linear shrinkage estimators was to push small eigenvalues up and large eigenvalues
down, as recently summarized by Ledoit and Wolf (2021a). A major advantage of CVTL over
existing estimators is its flexibility, as the estimation parameters can be selected based on a
predefined objective function. All portfolio weights are directly calculated through the GMV portfolio
optimization procedure that relies only on the covariance matrix. For instance, to maximize risk-
adjusted return, CVTL does not need to explicitly estimate future returns but they are implicitly
estimated via the chosen parameter configuration to shrunk the covariance eigenvalues.
This study opens up several avenues for future research. First, one could improve on the selection
of transfer datasets used during cross validation. In our study, we used random portfolio constituents
for the disjoint history. However, it should be possible to obtain more accurate estimation parameters
by specifically selecting similar stocks, e.g., from the same sectors but in different countries. Second,
the estimation parameters could be selected based on more sophisticated search procedures, like
random search (Bergstra and Bengio 2012), where each parameter value is drawn from a predefined
distribution. Third, it also seems reasonable to evaluate our approach based on other important
criteria like sustainability. Recently, portfolios with a particular focus on ESG (Environmental,
Social and Governance) criteria have received growing attention from investors. A possible objective
function could be defined as the ESG score divided by volatility, which would present a novel way
References
Ackermann F, Pohl W, Schmedders K (2017) Optimal and naive diversification in currency markets. Manage-
Bai J, Ng S (2002) Determining the number of factors in approximate factor models. Econometrica 70(1):191–
221.
Ban GY, El Karoui N, Lim AEB (2018) Machine learning and portfolio optimization. Management Sci.
64(3):1136–1154.
Bastani H (2021) Predicting with proxies: Transfer learning in high dimension. Management Sci. 67(5):2964–
2984.
Bastani H, Simchi-Levi D, Zhu R (2021) Meta dynamic pricing: Transfer learning across experiments.
Bergmeir C, Benı́tez JM (2012) On the use of cross-validation for time series predictor evaluation. Inf. Sci.
191:192–213.
Bergstra J, Bengio Y (2012) Random search for hyper-parameter optimization. J. Mach. Learn. Res. 13(2):281–
305.
Bickel PJ, Levina E (2008) Covariance regularization by thresholding. Ann. Stat. 36(6):2577–2604.
Bodnar T, Parolya N, Schmid W (2018) Estimation of the global minimum variance portfolio in high
Bun J, Bouchaud JP, Potters M (2017) Cleaning large correlation matrices: Tools from random matrix theory.
Cai T, Liu W, Luo X (2011) A constrained l1 minimization approach to sparse precision matrix estimation.
Das SR, Markowitz H, Scheid J, Statman M (2010) Portfolio optimization with mental accounts. J. Financial
Das SR, Ostrov D, Radhakrishnan A, Srivastav D (2018) A new approach to goals-based wealth management.
De Nard G, Ledoit O, Wolf M (2019) Factor models for portfolio selection in large dimensions: The good, the
DeMiguel V, Garlappi L, Uppal R (2009b) Optimal versus naive diversification: How inefficient is the 1/N
Dorfman R (1979) A formula for the gini coefficient. Rev. Econ. Stat . 61(1):146–149.
Engle RF, Ledoit O, Wolf M (2019) Large dynamic covariance matrices. J. Bus. Econ. Stat. 37(2):363–375.
Fan J, Liao Y, , Mincheva M (2011) High-dimensional covariance matrix estimation in approximate factor
Fan J, Liao Y, Liu H (2016) An overview of the estimation of large covariance and precision matrices. Econom.
J. 19(1):C1–C32.
Fan J, Liao Y, Mincheva M (2013) Large covariance estimation by thresholding principal orthogonal comple-
Fan J, Liu H, Wang W (2018) Large covariance estimation through elliptical factor models. Ann. Stat.
46(4):1383–1414.
Fan J, Zhang J, Yu K (2012) Vast portfolio selection with gross-exposure constraints. J. Am. Stat. Assoc.
107(498):592–606.
Frahm G, Memmel C (2010) Dominating estimators for minimum-variance portfolios. J. Econom. 159(2):289–
302.
french/data_library.html.
Friedman J, Hastie T, Tibshirani R (2008) Sparse inverse covariance estimation with the graphical lasso.
Biostatistics 9(3):432–441.
Frost PA, Savarino JE (1986) An empirical Bayes approach to efficient portfolio selection. J. Financial Quant.
Anal. 21(3):293–305.
Jagannathan R, Ma T (2003) Risk reduction in large portfolios: Why imposing the wrong constraints helps.
J. Finance 58(4):1651–1683.
James W, Stein C (1961) Estimation with quadratic loss. Proc. 4th Berkeley Symp. Mathematical Statistics
Probability, 361–380.
Jorion P (1986) Bayes-stein estimation for portfolio analysis. J. Financial Quant. Anal. 21(3):68–74.
Laloux L, Cizeau P, Bouchaud JP, Potters M (1999) Noise dressing of financial correlation matrices. Phys.
Laloux L, Cizeau P, Potters M (2000) Random matrix theory and financial correlations. Int. J. Theor. Appl.
Finance 3(3):391—-397.
Lam C, Fan J (2009) Sparsistency and rates of convergence in large covariancematrix estimation. Ann. Stat.
37(6B):4254–4278.
Ledoit O, Wolf M (2003) Improved estimation of the covariance matrix of stock returns with an application
Ledoit O, Wolf M (2004a) Honey, I shrunk the sample covariance matrix. J. Portf. Manag. 4(30):110–119.
Ledoit O, Wolf M (2004b) A well-conditioned estimator for large-dimensional covariance matrices. J. Multivar.
Anal. 88:365–411.
Ledoit O, Wolf M (2012) Nonlinear shrinkage estimation of large-dimensional covariance matrices. Ann. Stat.
40(2):1024–1060.
Ledoit O, Wolf M (2015) Spectrum estimation: a unified framework for covariance matrix estimation and Pca
Ledoit O, Wolf M (2017a) Nonlinear shrinkage of the covariance matrix for portfolio selection: Markowitz
Ledoit O, Wolf M (2017b) Numerical implementation of the quest function. Comput. Stat. Data. Anal.
115:199–223.
Ledoit O, Wolf M (2020) Analytical nonlinear shrinkage of large-dimensional covariance matrices. Ann. Stat.
48(5):3043–3065.
Ledoit O, Wolf M (2021a) The power of (non-)linear shrinking: A review and guide to covariance matrix
Ledoit O, Wolf M (2021b) Quadratic shrinkage for large covariance matrices. Bernoulli Fortcoming.
Ledoit O, Wolf M (2021c) Shrinkage estimation of large covariance matrices: Keep it simple, statistician? J.
Michaud RO (1989) The Markowitz optimization enigma: Is ‘optimized’ optimal? Financial Anal. J. 45(1):31–
42.
Nguyen VA, Kuhn D, Esfahani PM (2021) Robust inverse covariance estimation: The wasserstein shrinkage
Olivares-Nadal AV, DeMiguel V (2018) Technical note - A robust perspective on transaction costs in portfolio
Pan SJ, Yang Q (2009) A survey on transfer learning. IEEE Trans. Knowl. Data. Eng. 22(10):1345–1359.
Plerou V, Gopikrishnan P, Rosenow B, Amaral LAN, Guhr T, Stanley HE (2002) A random matrix approach
Roncalli T, Weisang G (2016) Risk parity portfolios with risk factors. Quant. Finance 16(3):377–388.
Rossi B, Inoue A (2012) Out-of-sample forecast tests robust to the choice of window size. J. Bus. Econ. Stat.
30(3):432–453.
Shefrin H, Statman M (2000) Behavioral portfolio theory. J. Financial Quant. Anal. 35(2):127–151.
Stein C (1956) Inadmissibility of the usual estimator for the mean of a multivariate normal distribution. Proc.
Thapa C, Poshakwale SS (2010) International equity portfolio allocations and transaction costs. J. Bank.
Financ. 34(11):2627–2638.
Tu J, Zhou G (2011) Markowitz meets Talmud: A combination of sophisticated and naive diversification
Zhao L, Chakrabarti D, Muthuraman K (2019) Portfolio construction by mitigating error amplification: The
adjusted return after transaction costs. EC.2 provides several insights about the observed parameter
configurations of our approach. EC.3 evaluates the possible benefit of our approach over existing
methods. EC.4 presents the results for minimum variance and maximum risk-adjusted return for
per trade (Thapa and Poshakwale 2010)). The turnover rates are shown in Table EC.1. The results
for risk-adjusted return after transaction costs for different datasets are shown in Table EC.3 for
D = 1.33 and Table EC.2 for D = 0.40. The results for all dimensions based on the US dataset are
Table EC.1 Turnover for Different Portfolio Sizes and Dimensions for US Data.
N 100 200 300
D = N/T 2 1.50 1.33 0.40 0.20 0.13 2 1.50 1.33 0.40 0.20 0.13 2 1.50 1.33 0.40 0.20 0.13
Estimators with Objective Minimum Variance
CVTL 21.25 18.57 17.03 9.36 6.61 5.32 17.69 15.53 14.08 8.03 6.43 5.41 15.29 13.39 12.68 7.61 6.18 5.25
CVTLLS 22.78 20.12 18.33 9.90 6.82 5.40 19.59 17.04 15.40 8.56 6.56 5.52 16.66 14.79 13.95 8.12 6.35 5.43
QIS 22.67 21.95 21.77 13.00 8.45 6.55 19.21 18.14 17.63 11.12 7.71 6.18 16.94 16.60 16.33 10.38 7.36 5.53
QuEST 20.48 19.16 18.44 12.09 8.25 6.48 17.25 15.79 15.04 10.73 7.63 6.14 15.10 14.58 14.20 10.14 7.31 5.50
LShriCC 36.50 33.01 31.53 16.52 9.82 7.28 36.30 34.03 33.05 17.00 9.98 7.30 37.01 35.64 34.46 17.68 9.80 6.78
LShri 27.85 27.20 26.99 16.62 9.90 7.31 30.39 30.53 30.58 17.39 10.04 7.34 31.40 32.95 33.20 17.94 9.83 6.76
BPSEst – – – 24.34 11.49 7.94 – – – 22.76 10.92 7.68 – – – 22.22 10.52 7.03
FMEst – – – 25.91 11.66 8.00 – – – 23.65 11.03 7.71 – – – 22.91 10.60 7.05
POET I – – – 18.63 9.52 6.86 – – – 20.89 9.61 6.91 – – – 24.42 9.75 6.52
POET II – 43.65 37.67 15.09 9.55 7.50 – 39.98 34.15 13.85 9.33 7.32 – 38.53 32.73 13.55 9.13 6.59
BN 17.02 16.49 16.00 11.78 7.45 5.87 17.11 16.24 15.71 10.98 7.20 5.98 17.08 16.09 15.45 10.66 7.31 5.78
Sample – – – 28.41 12.26 8.32 – – – 25.12 11.43 7.89 – – – 24.03 10.85 7.13
Note: Empirical results based on the All US data set with N ∈ {100, 200, 300}. Each value is given in percent as the
average over 50 evaluations of random portfolio constituents. Turnover greater than 50 due to estimation errors is
denoted by “–”. The best estimator per problem setting (excluding the 1/N portfolio) is highlighted in bold.
Table EC.2 Out-Of-Sample Annualized Turnover and Risk-Adjusted Return after Cost for Different Datasets.
D Turnover Risk-Adjusted Return After Cost
N 100 200 300 100 200 300
Data US EU WO US EU WO US EU WO US EU WO US EU WO US EU WO
Estimators with Objective Minimum Variance
CVTL 17.03 20.64 19.70 14.08 18.50 17.55 12.68 17.93 16.62 0.49 0.55 0.32 0.88 0.91 0.45 0.76 0.87 0.40
CVTLLS 18.33 22.10 20.96 15.40 19.71 18.85 13.95 18.87 17.80 0.47 0.50 0.32 0.83 0.85 0.45 0.70 0.81 0.39
QIS 21.77 23.21 23.36 17.63 20.04 21.04 16.33 19.42 20.72 0.35 0.42 0.22 0.74 0.82 0.33 0.59 0.78 0.24
QuEST 18.44 20.00 20.06 15.04 17.64 18.59 14.20 17.39 18.68 0.45 0.51 0.30 0.84 0.90 0.40 0.68 0.86 0.31
LShriCC 31.53 32.74 30.87 33.05 33.48 32.73 34.46 34.11 34.66 −0.20 −0.11 −0.19 −0.17 −0.03 −0.25 −0.58 −0.21 −0.47
LShri 26.99 36.12 32.16 30.58 43.38 37.90 33.20 48.24 42.25 0.19 0.09 0.01 0.24 0.09 −0.12 −0.14 −0.26 −0.48
BPSEst – – – – – – – – – – – – – – – – – –
FMEst – – – – – – – – – – – – – – – – – –
POET I – – – – – – – – – – – – – – – – – –
POET II 37.67 45.78 44.45 34.15 42.83 42.53 32.73 41.48 41.32 −0.16 −0.20 −0.33 0.08 0.07 −0.27 −0.13 −0.05 −0.38
BN 16.00 17.71 16.86 15.71 18.01 17.53 15.45 18.84 18.04 0.57 0.60 0.40 0.82 0.89 0.45 0.76 0.89 0.39
Sample – – – – – – – – – – – – – – – – – –
Note: Turnover and risk-adjusted return after costs (0.25% per trade) of all evaluated covariance estimators for
US, EU and WO data with N ∈ {100, 200, 300} and D = 0.4. Each score is given in percent as the average over 50
evaluations with random portfolio constituents. Out-of-sample volatility greater than 50 percent due to estimation
errors is denoted by “–”. The best estimator per problem setting (in terms of turnover excluding 1/N portfolio) is
highlighted in bold.
Table EC.3 Out-Of-Sample Annualized Turnover and Risk-Adjusted Return after Cost for Different Datasets.
D Turnover Risk-Adjusted Return After Cost
N 100 200 300 100 200 300
Data US EU WO US EU WO US EU WO US EU WO US EU WO US EU WO
Estimators with Objective Minimum Variance
CVTL 9.36 11.27 10.41 8.03 10.84 10.00 7.61 10.65 10.03 0.54 0.78 0.44 0.80 1.07 0.74 0.73 1.06 0.65
CVTLLS 9.90 11.84 11.05 8.56 11.25 10.45 8.12 10.82 10.45 0.53 0.75 0.46 0.78 1.04 0.74 0.70 1.03 0.65
QIS 13.00 13.15 13.93 11.12 12.59 13.51 10.38 12.25 13.45 0.39 0.68 0.33 0.66 0.99 0.57 0.59 0.99 0.53
QuEST 12.09 12.36 13.04 10.73 12.12 13.19 10.14 11.97 13.46 0.41 0.71 0.35 0.68 1.01 0.58 0.60 0.99 0.53
LShriCC 16.52 16.48 16.15 17.00 17.50 17.59 17.68 18.62 18.81 0.06 0.35 0.15 0.10 0.32 0.22 −0.11 0.25 0.11
LShri 16.62 20.55 19.04 17.39 22.25 20.89 17.94 23.54 22.34 0.26 0.42 0.20 0.33 0.49 0.28 0.18 0.39 0.19
BPSEst 24.34 26.30 26.31 22.76 27.13 27.31 22.22 27.48 28.10 0.04 0.20 0.06 0.16 0.28 0.08 0.05 0.21 0.01
FMEst 25.91 28.53 28.68 23.65 28.72 29.13 22.91 28.61 29.58 −0.02 0.14 −0.03 0.11 0.22 0.00 0.01 0.18 −0.04
POET I 18.63 22.04 21.42 20.89 24.59 25.45 24.42 26.63 28.55 0.37 0.49 0.31 0.06 0.85 0.03 0.64 0.56 −0.10
POET II 15.09 18.16 18.11 13.85 17.34 17.41 13.55 17.22 16.93 0.26 0.50 0.22 0.42 0.68 0.46 0.34 0.62 0.42
BN 11.78 11.83 11.43 10.98 12.12 11.88 10.66 12.72 12.18 0.47 0.78 0.41 0.70 1.16 0.72 0.62 1.17 0.61
Sample 28.41 32.42 32.86 25.12 31.25 32.09 24.03 30.51 32.00 −0.12 0.05 −0.16 0.03 0.13 −0.11 −0.04 0.12 −0.11
Note: Turnover and risk-adjusted return after costs (0.25% per trade) of all evaluated covariance estimators for US,
EU and WO data with N ∈ {100, 200, 300} and D = 1.33. Each score is given in percent as the average over 50
evaluations with random portfolio constituents. Out-of-sample volatility greater than 50 percent due to estimation
errors is denoted by “–”. The best estimator per problem setting (in terms of turnover excluding 1/N portfolio) is
highlighted in bold.
Table EC.4 Out-Of-Sample Risk-Adjusted Return after Cost for US Data and Different Sample Dimensions.
N 100 200 300
D = N/T 2.0 1.50 1.33 0.40 0.20 0.13 2.0 1.50 1.33 0.40 0.20 0.13 2.0 1.50 1.33 0.40 0.20 0.13
Estimators with Objective Minimum Variance
CVTL 0.46 0.59 0.49 0.54 0.74 0.66 0.63 0.89 0.88 0.80 0.52 0.57 0.93 0.81 0.76 0.73 0.52 1.28
CVTLLS 0.47 0.57 0.47 0.53 0.74 0.66 0.58 0.84 0.83 0.78 0.53 0.56 0.86 0.76 0.70 0.70 0.51 1.23
QIS 0.45 0.51 0.35 0.39 0.65 0.60 0.56 0.81 0.74 0.66 0.46 0.51 0.85 0.67 0.59 0.59 0.44 1.22
QuEST 0.51 0.59 0.45 0.41 0.66 0.60 0.63 0.89 0.84 0.68 0.47 0.51 0.93 0.76 0.68 0.60 0.45 1.22
LShriCC – – – 0.06 0.44 0.46 – – – 0.10 0.17 0.32 – – – – 0.09 0.91
LShri 0.30 0.37 0.19 0.26 0.57 0.56 0.17 0.32 0.24 0.33 0.32 0.44 0.22 – – 0.18 0.28 1.08
BPSEst – – – 0.04 0.50 0.53 – – – 0.16 0.29 0.42 0.28 – – 0.05 0.25 1.05
FMEst – 0.29 0.29 – 0.49 0.52 – – – 0.11 0.28 0.42 0.28 – – 0.01 0.24 1.05
POET I – 0.58 – 0.37 0.71 0.66 – – – 0.06 0.50 0.55 – – – 0.64 0.49 1.23
POET II – – – 0.26 0.56 0.54 – – 0.08 0.42 0.32 0.44 – – – 0.34 0.33 1.08
BN 0.58 0.62 0.57 0.47 0.76 0.66 0.63 0.91 0.82 0.70 0.53 0.57 0.86 0.73 0.76 0.62 0.53 1.33
Sample – 0.29 0.29 – 0.44 0.50 – – – 0.03 0.25 0.41 0.28 – – – 0.22 1.05
Note: Empirical results of all evaluated covariance estimators for US data with N ∈ {100, 200, 300} and
D ∈ {2.0, 1.5, 1.33, 0.4, 0.2, 0.3}. Each score is given in percent as the average over 50 evaluations with random portfolio
constituents. Out-of-sample volatility greater than 50 percent due to estimation errors is denoted by “–”. The best
Table EC.5 Out-Of-Sample Risk-Adjusted Return after Cost for FFI Data and Different Sample Dimensions.
N 10 30 49
D = N/T 2.0 1.50 1.33 0.40 0.20 0.13 2.0 1.50 1.33 0.40 0.20 0.13 2.0 1.50 1.33 0.40 0.20 0.13
Estimators with Objective Minimum Variance
CVTL 0.37 0.49 0.34 0.34 0.59 0.73 0.05 0.11 0.15 0.45 0.66 0.68 – 0.01 0.15 0.47 0.44 0.69
CVTLLS 0.39 0.48 0.35 0.32 0.62 0.75 0.10 0.17 0.18 0.46 0.64 0.67 – 0.10 0.17 0.45 0.42 0.67
QIS 0.05 – 0.00 – 0.45 0.59 – – – 0.03 0.34 0.50 – – – – 0.13 0.45
QuEST 0.13 0.06 – 0.07 0.51 0.58 – – – 0.08 0.34 0.51 – – – 0.07 0.16 0.46
LShriCC – – – 0.18 0.50 0.54 – – – 0.09 0.41 0.56 – – – 0.02 0.20 0.51
LShri 0.10 0.21 0.14 0.17 0.56 0.67 – – – 0.01 0.34 0.52 – – – – 0.12 0.47
BPSEst 0.14 0.37 0.52 – 0.39 0.55 – – – – 0.19 0.42 – – – – – 0.35
FMEst – 0.67 0.68 – 0.37 0.54 0.44 – – – 0.17 0.41 – – 0.08 – – 0.34
POET I – 0.10 0.10 0.17 0.56 0.67 – – – – 0.36 0.52 – – – – 0.15 0.49
POET II 0.28 0.34 0.28 0.07 0.57 0.58 – – – – 0.32 0.47 – – – 0.12 0.18 0.41
BN 0.33 0.31 0.17 0.20 0.56 0.62 0.08 0.18 0.15 0.24 0.42 0.52 0.01 0.17 0.15 0.28 0.22 0.55
Sample – – 0.68 – 0.29 0.50 0.44 – – – 0.12 0.40 – – 0.08 – – 0.31
Note: Empirical results of all evaluated covariance estimators for FFI data with N ∈ {10, 30, 49} and D ∈
{2.0, 1.5, 1.33, 0.4, 0.2, 0.3}. Each score is given in percent as the average over 50 evaluations with random portfolio
constituents. Out-of-sample volatility greater than 50 percent due to estimation errors is denoted by “–”. The best
The mean parameter values on the US dataset for different dimensions are shown in Table EC.6.
D = N/T δ1 δ2 δ3 γ δ1 δ2 δ1 δ2 δ3 γ δ1 δ2
N = 100
2.00 20.90 26.60 52.40 88.60 61.90 38.10 23.20 43.00 33.80 40.80 57.60 42.40
1.50 30.70 32.90 36.40 90.90 63.00 37.00 24.00 47.50 28.50 59.00 56.40 43.60
1.33 36.50 36.70 26.80 91.90 62.60 37.40 28.60 46.40 25.00 42.10 55.10 44.90
0.40 60.60 27.50 11.90 84.80 75.80 24.20 33.50 45.80 20.70 30.00 45.40 54.60
0.20 68.80 19.50 11.70 81.80 83.50 16.50 33.70 47.00 19.30 11.40 44.20 55.80
0.13 69.40 17.70 12.90 69.00 85.40 14.60 29.30 45.30 25.50 35.80 43.30 56.70
N = 200
2.00 25.10 24.90 50.00 91.80 63.60 36.40 27.30 42.60 30.10 48.80 56.60 43.40
1.50 33.30 29.20 37.50 94.60 65.00 35.00 31.20 46.10 22.70 50.30 57.00 43.00
1.33 38.50 33.00 28.50 96.20 64.10 35.90 30.80 48.60 20.60 16.30 52.00 48.00
0.40 59.60 26.30 14.00 96.10 75.90 24.10 34.90 49.30 15.80 14.00 43.20 56.80
0.20 63.40 17.60 19.00 77.50 84.60 15.40 30.20 44.80 25.10 49.10 41.90 58.10
0.13 66.40 15.30 18.30 74.90 87.60 12.40 34.30 38.50 27.20 18.20 53.30 46.70
N = 300
2.00 30.00 29.30 40.60 91.00 61.50 38.50 27.90 55.90 16.20 19.30 37.90 62.10
1.50 36.10 33.00 30.90 96.00 62.50 37.50 28.90 52.00 19.10 46.30 43.90 56.10
1.33 39.60 33.80 26.60 97.30 63.60 36.40 29.70 46.80 23.50 41.30 54.40 45.60
0.40 51.10 23.60 25.30 98.50 75.20 24.80 30.20 51.90 17.90 40.80 39.40 60.60
0.20 62.90 18.40 18.70 82.70 83.60 16.40 30.50 45.80 23.70 31.00 46.10 53.90
0.13 63.20 12.60 24.20 69.70 88.90 11.10 40.10 45.20 14.70 −4.50 43.50 56.50
Note: Each parameter value is given in percent as the average over 50 evaluations with random portfolio constituents.
Next, we present how the parameters of our approach change over time after model inception. We
provide the corresponding plots for N ∈ {100, 200, 300} and D ∈ {2, 1.5, 1.33, 0.40, 0.20, 0.13} for
the objectives minimum variance (Figure EC.1, Figure EC.2, and Figure EC.3), and maximum
Figure EC.1 Parameter Values over Time after Model Inception for Minimum Variance and N=100.
D = 2.00 D = 1.50 D = 1.33
1 1 1
Paramter value
0 0 0
0 0 0
δ1 δ2 γ
Note: Values are aggregated over 50 evaluations with random portfolio constituents for US data and N = 100.
Figure EC.2 Parameter Values over Time after Model Inception for Minimum Variance and N=200.
D = 2.00 D = 1.50 D = 1.33
1 1 1
Paramter value
0 0 0
0 0 0
δ1 δ2 γ
Note: Values are aggregated over 50 evaluations with random portfolio constituents for US data and N = 200.
Figure EC.3 Parameter Values over Time after Model Inception for Minimum Variance and N=300.
D = 2.00 D = 1.50 D = 1.33
1 1 1
Paramter value
0 0 0
0 0 0
δ1 δ2 γ
Note: Values are aggregated over 50 evaluations with random portfolio constituents for US data and N = 300.
Figure EC.4 Parameter Variation Over Time for Minimum Variance and N=100, D=1.33.
δ1 δ2
1 1
Paramter value
0.5 0.5
0 0
−0.5 −0.5
0 10 20 30 40 50 0 10 20 30 40 50
1
Paramter value
−1
0 10 20 30 40 50
δ1 δ2 γ
Note: Parameter mean value and standard deviations. Values are aggregated over 50 evaluations with random
Figure EC.5 Parameter Variation Over Time for Minimum Variance and N=100, D=0.40.
δ1 δ2
1 1
Paramter value
0.5 0.5
0 0
−0.5 −0.5
0 10 20 30 40 50 0 10 20 30 40 50
1
Paramter value
−1
0 10 20 30 40 50
δ1 δ2 γ
Note: Parameter mean value and standard deviations. Values are aggregated over 50 evaluations with random
Figure EC.6 Parameter Values over Time after Model Inception for Maximum Risk-Adjusted Return and
N=100.
D = 2.00 D = 1.50 D = 1.33
1 1 1
Paramter value
0 0 0
0 0 0
δ1 δ2 γ
Note: Values are aggregated over 50 evaluations with random portfolio constituents for US data and N = 100.
Figure EC.7 Parameter Values over Time after Model Inception for Maximum Risk-Adjusted Return and
N=200.
D = 2.00 D = 1.50 D = 1.33
1 1 1
Paramter value
0 0 0
0 0 0
δ1 δ2 γ
Note: Values are aggregated over 50 evaluations with random portfolio constituents for US data and N = 200.
Figure EC.8 Parameter Values over Time after Model Inception for Maximum Risk-Adjusted Return and
N=300.
D = 2.00 D = 1.50 D = 1.33
1 1 1
Paramter value
0 0 0
0 0 0
δ1 δ2 γ
Note: Values are aggregated over 50 evaluations with random portfolio constituents for US data and N = 300.
Figure EC.9 Parameter Variation Over Time for Maximum Risk-Adjusted Return and N=100, D=1.33.
δ1 δ2
1 1
Paramter value
0.5 0.5
0 0
−0.5 −0.5
0 10 20 30 40 50 0 10 20 30 40 50
1
Paramter value
−1
0 10 20 30 40 50
δ1 δ2 γ
Note: Parameter mean value and standard deviations. Values are aggregated over 50 evaluations with random
Figure EC.10 Parameter Variation Over Time for Maximum Risk-Adjusted Return and N=100, D=0.40.
δ1 δ2
1 1
Paramter value
0.5 0.5
0 0
−0.5 −0.5
0 10 20 30 40 50 0 10 20 30 40 50
1
Paramter value
−1
0 10 20 30 40 50
δ1 δ2 γ
Note: Parameter mean value and standard deviations. Values are aggregated over 50 evaluations with random
In the main paper, we compared the eigenvalues of CVTL against those of QuEST (Ledoit and
Wolf 2015) under the objective minimum variance. The same analysis under the objective maximum
Note: The plotted eigenvalues are averaged over 50 simulations with random portfolio constituents based on the US
We also consider the relations between the parameters δ1 , δ2 , γ. For convenience, we define δ3 =
1 − δ1 − δ2 as the intensity of the second shrinkage target. We consider the relations between our
estimation parameters after 12 months following model inception. Figures EC.12 and EC.13 show
the relations and histogramms for the objective minimum variance based on dimensions D = 1.33
and D = 0.40. Figures EC.14 and EC.15 show the same plots but for the objective maximum
risk-adjusted return.
We observe that the parameter pairs (δ1 , δ2 ), (δ1 , δ3 ), and (δ1 , γ) are in opposite relation for low-
Figure EC.12 Parameter Relations for High-Dimensional Problem with D = 1.33 under Objective Minimum
Variance.
Note: The plotted eigenvalues are averagered over 50 simulations with random portfolio constituents based on the US
Figure EC.13 Parameter Relations for Low-Dimensional Problem with D = 0.40 under Objective Minimum
Variance
Note: The plotted eigenvalues are averaged over 50 simulations with random portfolio constituents based on the US
Figure EC.14 Parameter Relations for High-Dimensional Problem with D = 1.33 under Objective Maximum
Risk-Adjusted Return.
Note: The plotted eigenvalues are averaged over 50 simulations with random portfolio constituents based on the US
Figure EC.15 Parameter Relations for Low-Dimensional Problem with D = 0.40 under Objective Maximum
Risk-Adjusted Return.
Note: The plotted eigenvalues are averaged over 50 simulations with random portfolio constituents based on the US
the proposed second shrinkage target of CVTL. In particular, we illustrate how different grid
entries γ, δ1 , δ2 ∈ P relate to traditional combinations between the sample covariance portfolio and
the 1/N portfolio. In addition, we illustrate how grid entries from CVTL relate to possible linear
shrinkage combinations (EC.3.1). We further extend the analysis about individual benefits of cross
validation and the second shrinkage target from the main paper (see Section 5.4) by relating results
from CVTL to the hypothetical combination of CVTL and QuEST (QuESTCV) (EC.3.2). We
also examine the potential benefit of CVTL (over the sample covariance estimate) and the second
shrinkage target by providing the share of grid entries from P that would lead to an outperformance
Figures EC.16 and EC.17 illustrate the historic results of hypothetical grid entries from P for the
US dataset with N = 100 and D ∈ {1.33, 0.40}. Bright gray transparent circles represent the grid
entries of theoretically reachable portfolios, filled gray circles represent possible grid entries in the
class of rotation equivariant estimators resulting in linear shrinkage with λ1 + λ2 = 1. Black crosses
represent linear combinations between the sample and 1/N portfolios. Figure EC.17 zooms into
non-linear grid entries that have the potential to either improve volatility or risk-adjusted return.
Comparing linear shrinkage (filled gray circles) and non-linear shrinkage (bright gray transparent
circles), Figure EC.16 indicates that the hypothetically optimal grid entries converge in terms of
the portfolio that minimizes variance. Along the efficient frontier, however, both types of shrinkage
procedures diverge within the hypothetical grid entries in the sense that our non-linear shrinkage
target has the potential to increase return for the same level of risk. Yet, the non-linear shrinkage
target also leads to hypothetical grid entries that are clearly dominated by others, especially the
linear shrinkage combination. Hence, we attribute greater potential but also greater predictive risk
to CVTL than to CVTLLS. The right side of Figure EC.16 supports the common literature (e.g.
Ledoit and Wolf 2004a) that it is not beneficial to rely solely on the sample covariance matrix when
Figure EC.17 zooms into entries located near the historically optimal hypothetical grid entry that
are optimal in minimizing variance or maximizing risk-adjusted return. We observe two opposing
findings. In case of minimum variance, we observe significantly more non-linear versus linear grid
entries that lead to lower variance in the high-dimensional than in the low-dimensional problem.
This supports our finding from Table 11, where we find greater benefit of the non-linear shrinkage
target for high-dimensional problems. In case of maximum risk-adjusted return, we observe the
opposite, that is, significantly more non-linear versus linear hypothetical grid entries that lead to
Note: Values are averaged over 50 simulations with random portfolio constituents based on the US data with N = 100.
Note: Values are averaged over 50 simulations with random portfolio constituents based on the US data with N = 100.
Figures EC.18 and EC.19 enlarge figs. EC.16 and EC.17 by also showing the grid entries of the
hypothetical estimator QuESTCV described in section 5.4 represented by small dark filled circles.
The traditional QuEST estimator is represented by one large dark filled circle. QuESTCV is achieved
by combining the eigenvalues of QuEST and cross validation to select the shrinkage intensity.
Figure EC.18 indicates that using the QuEST implied eigenvalues restricts the improvement
potential to minimizing variance. Along the efficient frontier their is no hypothetical grid entry that
leads to higher returns for the same level of variance compared to the linear shrinkage entries. Also
this can be expected by construction of the QuEST model, it highlights the flexibility of our second
shrinkage target. Figure EC.19 confirms this finding by indicating that neither in high nor in the low
dimensional case any hypothetical grid entry of QuESTCV leads to a higher risk-adjusted return
than the respective linear shrinkage grid entries. It further supports the recent idea of Ledoit and
Wolf (2021b), that combining QuEST eigenvalues with a linear shrinkage target has the potential
to improve the QuEST estimator, however, depends on finding the optimal intensity as several
QuESTCV grid entries have the potential of outperform QuEST. Finally, Figure EC.19 supports
our second non-linear shrinkage target as it leads to potentially lower variance than any combination
Note: Values are averaged over 50 simulations with random portfolio constituents based on the US data with N = 100.
Note: Values are averaged over 50 simulations with random portfolio constituents based on the US data with N = 100.
Table EC.7 further underlines the possible benefit of CVTL over the sample portfolio when
minimizing variance, and over the 1/N portfolio when maximizing risk-adjusted return. In addition,
Table EC.7 shows the possible benefits of the second shrinkage target of CVTL over CVTLLS
(linear shrinkage). We measure the potential benefit by counting the number of grid entries that
outperform the respective benchmark multiplied by 100 and divide the result by the grid size. As
an example, the possible benefit over the sample portfolio for N = 10 and D = 2 is 87 implying
that 87% of the hypothetical CVTL grid entries would outperform the sample portfolio in terms
of minimizing variance. The comparison using CVTL σµ measures the potential benefit in terms of
risk-adjusted return. Hence, Table EC.7 is the numerical representation of Figures EC.16 and EC.17.
When comparing CVTL to the sample covariance, the possible benefit decreases with an increasing
D. Under the objective maximum risk-adjusted return, the possible benefit of CVTL σµ over the
1/N portfolio increases with increasing D and N . Similar findings can be made when comparing
the volatility achieved by CVTL against CVTLLS. However, when D becomes larger, the possible
benefit is almost zero. This consistent with our findings in the main paper that the relative advantage
N CVTL CVTL µ
σ
D = N/T 2 1.50 1.33 0.40 0.20 0.13 2 1.50 1.33 0.40 0.20 0.13
Possible benefit over sample portfolio Possible benefit over 1/N portfolio
10 87.00 87.60 87.50 75.00 39.10 27.40 7.00 11.50 27.70 35.20 49.10 48.10
30 83.00 84.10 84.40 48.00 20.40 7.50 62.30 68.90 69.40 77.30 83.00 75.80
49 77.00 80.40 79.80 41.80 14.80 7.10 71.90 69.80 77.10 81.20 83.90 92.10
100 71.00 75.20 75.80 40.30 17.40 10.70 90.50 90.50 88.40 94.40 96.60 89.90
200 66.90 69.20 70.10 36.80 12.80 6.10 90.50 90.50 90.50 95.20 95.50 94.50
300 67.70 69.80 68.90 37.70 11.30 3.90 91.00 90.50 90.50 92.70 92.40 97.20
Possible benefit over CVTLLS Possible benefit over CVTLLS µ
σ
10 0.90 0.40 0.90 0.00 0.00 0.10 0.20 4.30 6.30 0.70 3.10 10.90
30 2.40 1.70 1.40 0.00 0.20 0.40 3.00 1.30 5.50 9.20 1.10 2.10
49 7.00 5.90 3.50 0.20 0.40 0.00 2.60 2.20 3.00 2.00 1.00 4.70
100 5.30 3.50 1.80 0.30 0.30 0.20 8.00 3.30 7.50 14.50 11.00 8.90
200 2.90 1.80 1.10 0.30 0.20 0.00 3.80 1.40 3.70 1.40 6.30 8.40
300 1.90 1.20 1.30 0.20 0.00 0.00 3.70 1.10 1.80 0.20 1.80 0.50
Note: Possible benefit of CVTL over different benchmarks based on FFI data for N ∈ 10, 30, 49 and US data for
N ∈ 100, 200, 300. For US data, each value is given in percent as the average over 50 evaluations of random portfolio
constituents.
The analyses in the main paper were primarily based on the US dataset and the metrics volatility
and risk-adjusted return. Here, we present the results on the WO and EU dataset. In addition, we
present of several other performance metrics, namely, the annualized return (Ret), the maximum
drawdown (maximum relative price depreciation over the time series averaged per annum, MDD),
downside volatility (annualized volatility of negative returns only, DwVol), Calmar ratio (we assume
the ratio as Ret/MDD, Calm), Sortino ratio (we assume the ratio as Ret/DwVol, Sort), and portfolio
turnover (sum of actively forced changes in weights averaged per annum, Turn). Altogether, we
• US Dataset (EC.4.1)
• WO Dataset (EC.4.2)
— Volatility (EC.4.2.1)
• EU Dataset (EC.4.3)
— Volatility (EC.4.3.1)
EC.4.1. US Data
Note: Each value is given in percent as the average over 50 evaluations of N = 100 portfolio constituents with
D ∈ (1.33, 0.40). The covariance estimator with the best score per metric is highlighted in bold. Underlined values
indicate significant difference from CVTL with p < 0.05.
Note: Each value is given in percent as the average over 50 evaluations of N = 200 portfolio constituents with
D ∈ (1.33, 0.40). The covariance estimator with best score per metric is highlighted in bold. Underlined values indicate
significant difference from CVTL with p < 0.05.
Note: Each value is given in percent as the average over 50 evaluations of N = 300 portfolio constituents with
D ∈ (1.33, 0.40). The covariance estimator with best score per metric is highlighted in bold. Underlined values indicate
significant difference from CVTL with p < 0.05.
EC.4.2. WO Data
EC.4.2.1. Volatility
Table EC.11 Out-Of-Sample Annualized Volatility for WO Data and Different Dimensions.
N 100 200 300
D = N/T 2 1.50 1.33 0.40 0.20 0.13 2 1.50 1.33 0.40 0.20 0.13 2 1.50 1.33 0.40 0.20 0.13
Note: Results of all evaluated covariance estimators for WO data with N ∈ {100, 200, 300} and D ∈
{2, 1.50, 1.33, 0.40, 0.20, 0.13}. Each value is given in percent as the average over 50 evaluations of random portfolio
constituents. Out-of-sample volatility greater than 50 percent due to estimation errors is denoted by “–”. The best
estimator per problem setting is highlighted in bold. Underlined values indicate significant difference from CVTL with
p < 0.05.
Table EC.12 Out-Of-Sample Risk-Adjusted Return for WO Data and Different Dimensions.
N 100 200 300
D = N/T 2 1.50 1.33 0.40 0.20 0.13 2 1.50 1.33 0.40 0.20 0.13 2 1.50 1.33 0.40 0.20 0.13
Note: Results of all evaluated covariance estimators for WO data with N ∈ {100, 200, 300} and D ∈
{2, 1.50, 1.33, 0.40, 0.20, 0.13}. Each value is given in percent as the average over 50 evaluations of random portfolio
constituents. Out-of-sample volatility greater than 50 percent due to estimation errors is denoted by “–”. The best
estimator per problem setting is highlighted in bold. Underlined values indicate significant difference from CVTL µ
σ
with p < 0.05.
Table EC.13 Out-Of-Sample Risk-Adjusted Return after Cost for WO Data and Different Dimensions.
N 100 200 300
D = N/T 2.0 1.50 1.33 0.40 0.20 0.13 2.0 1.50 1.33 0.40 0.20 0.13 2.0 1.50 1.33 0.40 0.20 0.13
Estimators with Objective Minimum Variance
CVTL 0.01 0.37 0.32 0.44 0.80 0.72 0.18 0.49 0.45 0.74 0.58 0.60 0.38 0.44 0.40 0.65 0.63 1.09
CVTLLS 0.05 0.36 0.32 0.46 0.81 0.72 0.19 0.49 0.45 0.74 0.59 0.58 0.38 0.42 0.39 0.65 0.61 1.12
QIS 0.02 0.31 0.22 0.33 0.70 0.65 0.17 0.43 0.33 0.57 0.50 0.54 0.31 0.30 0.24 0.53 0.56 1.08
QuEST 0.09 0.39 0.30 0.35 0.71 0.66 0.23 0.50 0.40 0.58 0.50 0.54 0.38 0.37 0.31 0.53 0.56 1.08
LShriCC – – – 0.15 0.58 0.57 – – – 0.22 0.36 0.44 – – – 0.11 0.38 0.94
LShri – 0.08 0.01 0.20 0.65 0.63 – – – 0.28 0.42 0.51 – – – 0.19 0.46 1.01
BPSEst – – – 0.06 0.59 0.61 – – – 0.08 0.38 0.49 – – – 0.01 0.42 0.99
FMEst – – – – 0.57 0.60 – – – – 0.37 0.49 – – – – 0.42 0.99
POET I – – – 0.31 0.73 0.70 – – – 0.03 0.49 0.58 – – – – 0.67 1.14
POET II – – – 0.22 0.64 0.62 – – – 0.46 0.46 0.56 – – – 0.42 0.56 1.02
BN 0.29 0.43 0.40 0.41 0.78 0.69 0.28 0.52 0.45 0.72 0.57 0.58 0.40 0.42 0.39 0.61 0.60 1.09
Sample – – – – 0.51 0.58 – – – – 0.34 0.48 – – – – 0.40 0.97
Note: Empirical results of all evaluated covariance estimators for WO data with N ∈ {100, 200, 300} and
D ∈ {2.0, 1.5, 1.33, 0.4, 0.2, 0.3}. Each value is given in percentage and is based on the average over 50 evaluations
of random investment universes. – is set for annualized out-of-sample volatility > 50% and negative out-of-sample
risk-adjusted returns. Best estimator per setting is highlighted in bold.
Note: Each value is given in percent as the average over 50 evaluations of N = 100 portfolio constituents with
D ∈ (1.33, 0.40). The covariance estimator with best score per KPI is marked bold. Underlined values indicate
significant difference from CVTL with p < 0.05.
Note: Each value is given in percent as the average over 50 evaluations of N = 200 portfolio constituents with
D ∈ (1.33, 0.40). The covariance estimator with best score per metric is highlighted in bold. Underlined values indicate
significant difference from CVTL with p < 0.05.
Note: Each value is given in percent as the average over 50 evaluations of N = 300 portfolio constituents with
D ∈ (1.33, 0.40). The covariance estimator with best score per metric is highlighted in bold. Underlined values indicate
significant difference from CVTL with p < 0.05.
EC.4.3. EU Data
EC.4.3.1. Volatility
Table EC.17 Out-Of-Sample Annualized Volatility for EU Data and Different Dimensions.
N 100 200 300
D = N/T 2 1.50 1.33 0.40 0.20 0.13 2 1.50 1.33 0.40 0.20 0.13 2 1.50 1.33 0.40 0.20 0.13
Note: Results of all evaluated covariance estimators for EU data with N ∈ {100, 200, 300} and D ∈
{2, 1.50, 1.33, 0.40, 0.20, 0.13}. Each value is given in percent as the average over 50 evaluations of random portfolio
constituents. Out-of-sample volatility greater than 50 percent due to estimation errors is denoted by “–”. The best
estimator per problem setting is highlighted in bold. Underlined values indicate significant difference from CVTL with
p < 0.05.
Table EC.18 Out-Of-Sample Risk-Adjusted Return for EU Data and Different Dimensions.
N 100 200 300
D = N/T 2 1.50 1.33 0.40 0.20 0.13 2 1.50 1.33 0.40 0.20 0.13 2 1.50 1.33 0.40 0.20 0.13
Note: Results of all evaluated covariance estimators for EU data with N ∈ {100, 200, 300} and D ∈
{2, 1.50, 1.33, 0.40, 0.20, 0.13}. Each value is given in percent as the average over 50 evaluations of random portfolio
constituents. Out-of-sample volatility greater than 50 percent due to estimation errors is denoted by “–”. The best
estimator per problem setting is highlighted in bold. Underlined values indicate significant difference from CVTL µ
σ
with p < 0.05.
Table EC.19 Out-Of-Sample Risk-Adjusted Return after Cost for EU Data and Different Dimensions.
N 100 200 300
D = N/T 2.0 1.50 1.33 0.40 0.20 0.13 2.0 1.50 1.33 0.40 0.20 0.13 2.0 1.50 1.33 0.40 0.20 0.13
Estimators with Objective Minimum Variance
CVTL 0.36 0.55 0.55 0.78 1.05 1.00 0.59 0.84 0.91 1.07 0.98 0.98 0.90 0.99 0.87 1.06 1.06 1.67
CVTLLS 0.34 0.50 0.50 0.75 1.03 0.98 0.54 0.78 0.85 1.04 0.98 0.95 0.82 0.91 0.81 1.03 1.01 1.62
QIS 0.36 0.46 0.42 0.68 0.99 0.95 0.56 0.77 0.82 0.99 0.96 0.96 0.86 0.92 0.78 0.99 1.02 1.63
QuEST 0.42 0.55 0.51 0.71 1.00 0.95 0.63 0.84 0.90 1.01 0.97 0.97 0.93 1.00 0.86 0.99 1.03 1.63
LShriCC – – – 0.35 0.65 0.72 – – – 0.32 0.60 0.71 – – – 0.25 0.62 1.33
LShri 0.06 0.15 0.09 0.42 0.83 0.85 – 0.08 0.09 0.49 0.76 0.85 0.02 – – 0.39 0.78 1.50
BPSEst – – – 0.20 0.75 0.80 – – – 0.28 0.71 0.82 – – – 0.21 0.73 1.48
FMEst – – – 0.14 0.74 0.80 – – – 0.22 0.70 0.82 – – – 0.18 0.73 1.48
POET I – – – 0.49 1.00 0.99 – – – 0.85 1.03 1.06 – – – 0.56 1.13 1.80
POET II – – – 0.50 0.82 0.82 – – 0.07 0.68 0.78 0.88 – – – 0.62 0.98 1.43
BN 0.53 0.62 0.60 0.78 1.09 0.99 0.68 0.87 0.89 1.16 1.03 1.00 0.88 0.93 0.89 1.17 1.14 1.64
Sample – – – 0.05 0.70 0.78 – – – 0.13 0.68 0.82 – – – 0.12 0.73 1.48
Note: Empirical results of all evaluated covariance estimators for EU data with N ∈ {100, 200, 300} and
D ∈ {2.0, 1.5, 1.33, 0.4, 0.2, 0.3}. Each score is given in percentage and is based on the average over 50 evaluations
of random investment universes. – is set for annualized out-of-sample volatility > 50% and negative out-of-sample
risk-adjusted returns. Best estimator per setting is highlighted in bold. Underlined values indicate significant difference
from CVTL µ with p < 0.05.
σ
Note: Each value is given in percent as the average over 50 evaluations of N = 100 portfolio constituents with
D ∈ (1.33, 0.40). The covariance estimator with best score per KPI is marked bold. The lowest turnover of any
optimized portfolio is marked bold. Underlined values indicate significant difference from CVTL with p < 0.05.
Note: Each value is given in percent as the average over 50 evaluations of N = 200 portfolio constituents with
D ∈ (1.33, 0.40). The covariance estimator with best score per KPI is marked bold. Underlined values indicate
significant difference from CVTL with p < 0.05.
Note: Each value is given in percent as the average over 50 evaluations of N = 300 portfolio constituents with
D ∈ (1.33, 0.40). The covariance estimator with best score per KPI is marked bold. Underlined values indicate
significant difference from CVTL with p < 0.05.
Note: Values are given in percent based on N = 10 portfolio constituents (industries) with D ∈ (1.33, 0.40). The
covariance estimator with best score per metric is highlighted in bold. Underlined values indicate significant difference
from CVTL with p < 0.05.
Note: Values are given in percent based on N = 30 portfolio constituents (industries) with D ∈ (1.33, 0.40). The
covariance estimator with best score per metric is hihglighted in bold. Underlined values indicate significant difference
from CVTL with p < 0.05.
Note: Values are given in percent based on N = 49 portfolio constituents (industries) with D ∈ (1.33, 0.40). The
covariance estimator with best score per metric is highlighted in bold. Underlined values indicate significant difference
from CVTL with p < 0.05.