Professional Documents
Culture Documents
Institute for New Economic Thinking at the Oxford Martin School, University of
Oxford, Oxford OX2 6ED, U.K.
2
Mathematical Institute, University of Oxford, Oxford OX1 3LP, U. K.
3
Santa-Fe Institute, Santa Fe, NM 87501, U.S.A
4
London Institute for Mathematical Sciences, London W1K 2XF, U.K.
5
United Nations University - MERIT, 6211TC Maastricht, The Netherlands
February 18, 2015
Abstract
Recently it has become clear that many technologies follow a generalized version of Moores
law, i.e. costs tend to drop exponentially, at different rates that depend on the technology. Here
we formulate Moores law as a time series model
and apply it to historical data on 53 technologies. Under the simple assumption of a correlated geometric random walk we derive a closed
form expression approximating the distribution
of forecast errors as a function of time. Based on
hind-casting experiments we show that it is possible to collapse the forecast errors for many dif-
Acknowledgements: We would like to acknowledge Diana Greenwald and Aimee Bailey for their
help in gathering and selecting data, as well as
Glen Otero for help acquiring data on genomics,
Chris Goodall for bringing us up to date on developments in solar PV, and Christopher Llewellyn
Smith for comments.
This project was supported
by the European Commission project FP7-ICT-2013611272 (GROWTHCOM) and by the U.S. Dept. of
Solar Energy Technologies Office under grant DEEE0006133.
Contacts: doyne.farmer@inet.ox.ac.uk;
francois.lafond@oxfordmartin.ox.ac.uk
Introduction
50.0
solar energy
price in $/Wp
5.0
1e+00
price in $/kWh
1e+02
0.5
1950
2000
0.1
1e02
2050
We show how one can combine the experience from all these different technologies to say,
Based on experience with many other technologies, the trend in photovoltaic modules is unlikely to reverse". Indeed we can assign a probability to different price levels at different points
in the future, as is done later in Figure 10. Of
course technological costs occasionally experience structural breaks where trends change
there are several clear examples in our historical data. The point is that, while such structural breaks happen, they are not so common
as to over-ride our ability to forecast. And of
course, every technology has its own story, its
own specific set of causes and effects that explain why costs went up or down in any given
year. Nonetheless, as we demonstrate here, the
long term trends tend to be consistent, and can
be captured via historical time series methods
with no direct information about the underlying technology-specific stories.
of other benefits, such as increased speed, decreased power consumption, and cheaper manufacture costs per unit of computation, with appropriate alterations in the exponent, it quickly
became clear that Moores law applies more
broadly. For example, when combined with the
predicted increase in transistor density, the scaling of transistor speed as a function of size gives
a doubling of computational performance every
18 months.
Moores law stimulated others to look at related data more carefully, and they discovered
that exponential improvement is a reasonable
approximation for other types of computer hardware as well, such as hard drives. Since the
performance of hard drives depends on physical
factors that are unrelated to transistor density
this is an independent fact (though of course the
fact that mass storage is essential for computation causes a tight coupling between the two
technologies). Lienhard, Koh and Magee, and
others4 have examined data for other products
and postulated that exponential improvement is
a much more general phenomenon that applies
to many different technologies. An important
difference that in part explains why exponential
scaling was discovered later for other technologies is that the rates of improvement vary widely
and computer hardware is an outlier in terms of
rate.
Although Moores law is traditionally applied
as a regression model, we reformulate it here as
a geometric random walk with drift. This allows
Several methods have been defined to obtain prediction intervals, i.e. error bars for the forecasts (Chatfield 1993). The classical Box-Jenkins methodology for
ARIMA processes uses a theoretical formula for the variance of the process, but does not account for uncertainty
due to parameter estimates. Another approach is to use
the empirical forecast errors to estimate the distribution
of forecast errors. In this case, one can use either the
in-sample errors (the residuals, as in e.g. Taylor & Bunn
(1999)), or the out-of-sample forecast errors (Williams
& Goodman 1971, Lee & Scholtes 2014). Several studies
have found that using residuals leads to prediction intervals which are too tight (Makridakis & Winkler 1989).
on other methods.
The general approach of basing technological
forecasts on historical data, which we pursue
here, stands in sharp contrast to the most widely
used method, which is based on expert opinion.
The use of expert opinions is clearly valuable,
and we do not suggest that it should be supplanted, but it has several serious drawbacks.
Expert opinions are subjective and can be biased
for a variety of reasons, including common information, herding, or vested interest (Albright
2002, National Research Council 2009). Forecasts for the costs of nuclear power in the US, for
example, were for several decades consistently
low by roughly a factor of three (Cooper 2009).
A second problem is that it is very hard to assess the accuracy of expert forecasts. In contrast
the method we develop here is objective and the
quality of the forecasts is known. Nonetheless
we believe that both methods are valuable and
that they should be used side-by-side.
The remainder of the paper develops as follows: In Section 2 we derive the error distribution for forecasts based on the geometric random
walk as a function of time horizon and other
variables and show how the data for different
technologies and time horizons should be collapsed. We also introduce the Integrated Moving Average Model and derive similar (approximate) formulas. In Section 3 we describe our
data set and present an empirical relationship
between the variance of the noise and the improvement rate for different technologies. In
Section 4 we describe our method of testing the
models against the data and present the results.
We then apply our method to give a distributional forecast for solar module prices in Section 5 and show how this can be used to forecast
the likelihood that one technology will overtake
another. Finally we give some concluding remarks in Section 6. A variety of technical results
are given in the appendices.
2
2.1
Models
(1)
2.2
(2)
t
X
ni .
(3)
i=1
A K
t+
X
ni
(5)
i=t+1
yt+ = yt +
,
with
where
is the estimated . The forecast error
is defined as
E = yt+ yt+ .
A = + 2 /m.
(7)
t+
X
ni ,
(11)
(8)
i=t+1
m3
m
K
2.3
Integrated
model
Moving
Average
with
2
2(m 1)
2
+
+
A = 2 + 1 +
m
m
(14)
Note that we recover Eq. 11 when = 0. In
the case where the variance is estimated, it is
possible to derive an approximate formula for
the growth and distribution of the forecast errors
and E are independent. The
by assuming that K
expected mean squared normalized error is
" #
2
m 1 A
E
=
, (15)
( ) E
m 3 1 + 2
K
yt yt1 = + vt + vt1 ,
(12)
3.1
E N (0, 2A ),
A /(1 + ) K
(13)
10
Data
11
Data collection
pcdb.santafe.edu
our dataset contains appliances such as television sets, that have dramatically increased in
quality through time12 .
One should therefore regard our results here
as a lower bound on what is possible, in the sense
that performing the analysis with better data in
which all technologies had invariant units would
very likely improve the quality of the forecasts.
We would love to be able to make appropriate
normalizations, but the work involved is prohibitive; if we dropped all questionable examples
we would end with little remaining data. Most
of the data are costs, but in a few cases they are
prices; again, this adds noise but if we were able
to be consistent that should only improve our
results. We have done various tests removing
data and the basic results are not sensitive to
what is included and what is omitted (see Fig.
15 in the appendix).
We have removed some technologies that are
too similar to each other from the Performance
Curve Database. For instance, when we have
two datasets for the same technology, we keep
only one of them. Our choice was based on data
quality and length of the time series. This selection left us with 66 technologies, belonging
to different sectors that we label as chemistry,
genomics, energy, hardware, consumer durables
and food.
log(cost)
1e08
1e05
1e02
1e+01
1e+04
reflects the fact that it comes from a book published by the Boston Consulting Group in 1972.
Table 1 gives a summary of the properties of the
data and more description of the sources can be
found in appendix A.
Chemical
Hardware
Consumer Goods
Energy
Food
Genomics
1940
1960
1980
2000
3.2
700
300
200
# of forecasts
400
500
40
30
20
100
10
# of technologies
25
12
20
15
Frequency
~
K
5
0.6
0.8
0.0
0.4
0.6
0.8
Frequency
10
15
0.2
20
40
60
80
1.0
0.5
0.0
20
30
40
50
0
0.4
20
0.2
10
10
10
8
6
4
0
Frequency
0.0
Frequency
600
# technologies
# of forecasts
50
0.5
1.0
Since our dataset includes technologies of different length (see Table 1 and Figure 4), and
because we are doing exhaustive hindcasting, as
described in the next section13 , the number of
possible forecasts is highest for = 1 and decreases for longer horizons. Figure 4 shows the
number of possible forecasts and the number of
technologies as a function of the forecast horizon . We somewhat arbitrarily impose an upper bound of max = 20, but find this makes
very little difference in the results (see Appendix
13
The number of forecast errors produced by a technology of length Ti is [Ti (m + 1)][Ti m]/2 which
is O(Ti2 ). Hence the total number of forecast errors contributed by a technology is disproportionately dependent
on its length. However, we have checked that aggregating the forecast errors so that each technology has an
equal weight does not qualitatively change the results.
10
Technology
Transistor
Geothermal.Electricity
Milk..US.
DRAM
Hard.Disk.Drive
Automotive..US.
Low.Density.Polyethylene
Polyvinylchloride
Ethanolamine
Concentrating.Solar
AcrylicFiber
Styrene
Titanium.Sponge
VinylChloride
Photovoltaics
PolyethyleneHD
VinylAcetate
Cyclohexane
BisphenolA
Monochrome.Television
PolyethyleneLD
Laser.Diode
PolyesterFiber
Caprolactam
IsopropylAlcohol
Polystyrene
Polypropylene
Pentaerythritol
Ethylene
Wind.Turbine..Denmark.
Paraxylene
DNA.Sequencing
NeopreneRubber
Formaldehyde
SodiumChlorate
Phenol
Acrylonitrile
Beer..Japan.
Primary.Magnesium
Ammonia
Aniline
Benzene
Sodium
Methanol
MaleicAnhydride
Urea
Electric.Range
PhthalicAnhydride
CarbonBlack
Titanium.Dioxide
Primary.Aluminum
Sorbitol
Aluminum
Free.Standing.Gas.Range
CarbonDisulfide
Ethanol..Brazil.
Refined.Cane.Sugar
CCGT.Power
HydrofluoricAcid
SodiumHydrosulfite
Corn..US.
Onshore.Gas.Pipeline
Motor.Gasoline
Magnesium
Crude.Oil
Nuclear.Electricity
Industry
Hardware
Energy
Food
Hardware
Hardware
Consumer Goods
Chemical
Chemical
Chemical
Energy
Chemical
Chemical
Chemical
Chemical
Energy
Chemical
Chemical
Chemical
Chemical
Consumer Goods
Chemical
Hardware
Chemical
Chemical
Chemical
Chemical
Chemical
Chemical
Chemical
Energy
Chemical
Genomics
Chemical
Chemical
Chemical
Chemical
Chemical
Food
Chemical
Chemical
Chemical
Chemical
Chemical
Chemical
Chemical
Chemical
Consumer Goods
Chemical
Chemical
Chemical
Chemical
Chemical
Chemical
Consumer Goods
Chemical
Energy
Food
Energy
Chemical
Chemical
Food
Energy
Energy
Chemical
Energy
Energy
T
38
26
79
37
20
21
17
23
18
26
13
15
19
11
34
15
13
17
14
22
15
13
13
11
9
26
10
21
13
20
12
13
13
11
15
14
14
18
40
13
12
17
16
16
14
12
22
18
9
9
40
8
17
22
10
25
34
10
11
9
34
14
23
19
23
20
p value
-0.50
0.00
-0.05
0.00
-0.02
0.00
-0.45
0.00
-0.58
0.00
-0.08
0.00
-0.10
0.00
-0.07
0.00
-0.06
0.00
-0.07
0.00
-0.10
0.00
-0.07
0.00
-0.10
0.00
-0.08
0.00
-0.10
0.00
-0.09
0.00
-0.08
0.00
-0.05
0.00
-0.06
0.00
-0.07
0.00
-0.08
0.00
-0.36
0.00
-0.12
0.00
-0.10
0.00
-0.04
0.00
-0.06
0.00
-0.10
0.00
-0.05
0.00
-0.06
0.00
-0.04
0.00
-0.10
0.00
-0.84
0.00
-0.02
0.00
-0.07
0.00
-0.03
0.00
-0.08
0.00
-0.08
0.01
-0.03
0.01
-0.04
0.01
-0.07
0.02
-0.07
0.02
-0.05
0.02
-0.01
0.02
-0.08
0.02
-0.07
0.03
-0.06
0.03
-0.02
0.03
-0.08
0.03
-0.01
0.03
-0.04
0.04
-0.02
0.06
-0.03
0.06
-0.02
0.09
-0.01
0.10
-0.03
0.12
-0.05
0.13
-0.01
0.23
-0.04
0.25
-0.01
0.25
-0.01
0.29
-0.02
0.30
-0.02
0.31
-0.00
0.47
-0.00
0.47
0.01
0.66
0.13
0.99
K
0.24
0.02
0.02
0.38
0.32
0.05
0.06
0.06
0.04
0.07
0.06
0.05
0.10
0.05
0.15
0.08
0.06
0.05
0.05
0.08
0.08
0.29
0.10
0.08
0.02
0.09
0.07
0.07
0.06
0.05
0.09
0.83
0.02
0.06
0.04
0.09
0.11
0.05
0.09
0.10
0.10
0.09
0.02
0.14
0.11
0.09
0.04
0.15
0.02
0.05
0.08
0.05
0.04
0.04
0.06
0.22
0.06
0.15
0.04
0.07
0.17
0.14
0.05
0.04
0.07
0.22
0.19
0.15
0.04
0.14
-0.15
1.00
0.46
0.32
0.36
0.91
0.02
0.74
0.61
-0.22
0.05
0.12
0.33
0.38
-0.03
0.02
0.88
0.37
-0.16
0.40
-0.24
-0.04
0.26
0.30
-0.26
0.75
-1.00
0.26
0.83
0.36
0.85
-1.00
1.00
-1.00
0.24
1.00
0.75
-0.10
0.42
0.29
0.73
0.04
-0.14
0.31
-1.00
-0.41
0.39
-1.00
0.73
-0.30
-0.04
-0.62
-1.00
-1.00
0.13
-1.00
-1.00
0.62
0.43
0.58
0.63
-0.13
Table 1: Descriptive statistics and parameter estimates (using the full sample) for all available technologies,
ordered by p-value of a one-sided t-test for
. The improvement of the last 13 technologies is not statistically
significant and so they are dropped from further analysis.
11
C.3).
Empirical results
4.1
3.3
Chemical
Energy
Hardware
Consumer Goods
Food
0.10
0.02
0.05
~
K
0.20
0.50
loglog fit
linear fit
0.02
0.05
0.10
0.20
0.50
for technologies
Figure 5: Scatter plot of
and K
with a significant improvement rate. The linear fit
(which is curved when represented in log scale) is
= 0.020.76
K
, with R2 = 0.87 and a p-value 0.
= 0.51(
The log-log fit is K
)0.72 , with R2 = 0.73
and p-value 0. Technologies with a faster rate of
improvement also have higher uncertainty in their
improvement.
14
12
500
2 =
K
t0
200
10
5
tX
0 1
1
[(yj+1 yj )
t0 ]2 . (18)
m 1 j=t m
4.2
m 1 2
+
m 3 m
20
()
100
50
real data
t0 1
1 X
yt yt0 m
t0 =
(yj+1 yj ) = 0
, (17)
m j=t m
m
10
20
Forecast horizon
averaging over all technologies and for all forecasts with < 20, as shown by the black dots in
Figure 6. The dashed line compares this to the
predicted normalized error under the geometric
random walk hypothesis.
The random walk model does a good job of
predicting the scaling of the forecast errors as a
function of , but the predicted errors are too
small by roughly a factor of two. The fact that
it underestimates the errors is not surprising
the random walk model is an extreme assumption. The correct prediction of the scaling of the
errors as a function of the forecast horizon is not
an obvious result; alternative hypotheses, such
as long-memory, can produce errors that grow in
time faster than the random walk. Given that
long-memory is a natural result of nonstationarity, which is commonly associated with technological change, our prior was that it was a
highly plausible alternative16 . These procedures
15
This is different from the maximum likelihood estimator, which does not make use of Bessels correction
(i.e. dividing by (m 1) instead of m). Our choice is
driven by the fact that in practice we use a very small
m, making the bias of the maximum likelihood estimator
rather large.
16
13
involve some data snooping, i.e. they use knowledge about the future to set , but since this is
only one parameter estimated from a sample of
more than 6, 000 forecasts the resulting estimate
should be reasonably reliable. See the more detailed discussion in Appendix D.
To understand why the errors in the model are
larger than those of the random walk we investigated two alternative hypotheses, heavy tails
and autocorrelated innovations. We found little evidence for heavy tails (appendix C.4) but
much stronger evidence for positive autocorrelations. Testing for autocorrelation is difficult due
to the fact that the series are short, and the sample autocorrelations of the individual samples
are highly variable and in many cases not very
statistically significant (see Figure 3). Nonetheless, on average the individual series have positive autocorrelations, suggesting that this is at
least a reasonable starting point.
We thus selected the IMA(1,1) model discussed in Section 2.3 as an alternative to the
simple random walk with drift. This model
requires an additional parameter , which describes the strength of the autocorrelation. Because we are using a small number m of past
data points to fit the model, a dynamic sample estimate based on a moving average is not
statistically stable.
To counter this problem we use a global value
of , i.e. we use the same value for all technologies and all points in time. We use two different methods for doing this and compare them in
what follows. The first method takes advantage
of the fact that the magnitude of the forecast
errors is an increasing function of (we assume
0.020 0.050
20
0.002 0.005
cumulative distribution
0.200 0.500
10
10
where H is the Hurst exponent. In the absence of longmemory H = 1/2, but for long-memory 1/2 < H < 1.
Long-memory can arise from many causes, including
nonstationarity. It is easy to construct plausible processes with the parameter varying where the mean
squared errors grow faster than 2 .
14
5e01
1e01
5e03 2e02
1e03
2e04
cumulative distribution
7 shows the distribution of rescaled forecast errors at several different values of using the
IMA(1,1) rescaling factor A (eq. 14) and compares it to the predicted distribution. Fig. 8
shows the distribution with all values of pooled
together and compares the empirical results for
the random walk with drift to the empirical results with the IMA(1,1). The predicted distribution is fairly close to the theoretical prediction,
and as expected the IMA(1,1) is better than the
random walk with drift.
Given the complicated overlapping structure
of the exhaustive hindcasting approach that we
use here, the only feasible approach to statistically testing our results is simulation18 . To
give a feeling for the expected size of the fluctuations, in Appendix E we generate an ensemble of results under the null hypothesis of the
IMA(1,1) model, generating an artificial data set
that mimics the real data, as described above.
This gives a feeling for the expected fluctuations in Figure 8 and makes it clear that the
random walk can confidently be rejected. The
estimate based on the IMA(1,1) model with
= w = 0.25 is also rejected, with no more
than 1% of the simulations generating deviations
as large as the real data (this depends on the way
in which deviations from the distribution are
measured). Unsurprisingly the IMA(1,1) model
with = m
= 0.63 is not rejected (see Appendix E for details).
Student t(m1)
IMA(1,1)
RWD
15
10
10
15
4.3
17
i and = w
K =K
= 0.25. We then estimate the parameters just as we did for the real data and compute the
mean squared normalized error in forecasting the artificial data as a function of . The curves corresponding to
the IMA models in Figure 6 are the result of doing this
1000 times and taking the average. The error bars are
also computed using the simulations (95% of the simulated datasets ( = 0.25) have a mean square normalized
forecast error within the grey area)
Dependence on m
The fact that we use a rolling window approach implies overlaps in both the historical sample used to estimate parameters at each time t0 and overlapping intervals in the future for horizons with > 1. This implies
substantial correlations in the empirical forecast errors,
which complicates statistical testing.
15
see whether we are correctly predicting the distributional forecasting accuracy of our method.
We now address the question of the optimal
value of m. In a stationary world, in which the
models are well-specified and there is a constant
true value of , one should always use the largest
possible value of m. In a non-stationary world,
however, it can be advantageous to use a smaller
value of m, or alternatively a weighted average
that decays as it goes into the past. We experimented with increasing m as shown in Figure 9
and as described in more detail in Appendix C.1,
and we find that the errors drop as m increases
roughly as one would expect if the process were
stationary. Note that to check forecast errors for
high m we have used technologies for which at
least m + 2 years were available.
6.69
2.46
0.9
0.33
0.05
10
1980
1986
1992
1998
2004
2010
2016
2022
2028
4
8
12
16
1
10
20
Forecast horizon
1
1.5
2
0.02
50
20
()
100
500
Empirical
IMA( = 0.25)
0.12
5.1
Application to solar PV
modules
In this section we provide a distributional forecast for the price of solar photovoltaic modules.
We then show how this can be used to make a
16
is now19 .
0.2
0.6
0.8
0.4
5.2
0.0
KC = 0
KC = 0.1
KC = 0.2
2013
2017
2021
2025
2029
2033
2037
2041
Forecast horizon
2 A /(1 + 2 )),
yS (t + ) N (yS (t) +
S , K
S
S
Figure 11: Probability that photovoltaic manufacture becomes cheaper than a hypothetical competitor
which starts three times cheaper and does not improve. The curves show Eq. 19 using
= 0.10,
C = S = m
. The probability that yS < yC
is the probability that the random variable Z =
yS yC is positive. Since yS and yC are normal,
assuming they are independent their difference
is normal
Z N Z , Z2 ,
17
electricity requires predicting the balance of system costs, for which we lack consistent historical data, and unlike module costs, the full cost
depends on factors such as insolation, interest
rates and local installation costs. As solar PV
grows to be a significant portion of the energy
supply the cost of storage will become very important. Nonetheless, it is useful to discuss it
in relation to the two competitors mentioned in
the introduction.
5.3
Discussion of PV relative to
coal-fired electricity and nuclear power
5.4
Forecasting production
So far we have focused exclusively on forecasting costs, but in many cases it may be useful
to also forecast production, or related factors
such as cumulative capacity or consumption. As
originally shown by Sahal (1979) the combination of exponentially decreasing costs and exponentially increasing production leads to Wrights
law. Nagy et al. (2013) found that as a rough approximation most of the technologies in our data
a large effect, one should bear in mind that the drop
is tiny in comparison to the factor of about 2,330 by
which solar PV modules have dropped in price. The
small change induced by fracking is only important because it is competing in a narrow price range with other
fossil fuel technologies. In work with other collaborators
we have examined not just oil, coal and gas, but more
than a hundred minerals; all of them show remarkably
flat historical prices, i.e. they all change by less than an
order of magnitude over the course of a century.
21
Levelized costs decrease more slowly than module
costs, but do decrease (Nemet 2006). For instance, installation costs per watt have fallen in Germany and are
now about half what they are in the U.S. (Barbose et al.
2014).
22
See
http://www.renewableenergyworld.com/rea/
news/article/2015/01/dubai-utility-dewa- procures-theworlds-cheapest-solar-energy-ever
20
18
20% primar
2027
y energy
1e+01
Primary Energy
Oil
Gas
Coal
1e01
1e+03
Nuclear
Hydro
Solar
1e03
Wind
Geothermal, Biomass,
and other renewables
1980
2000
2020
2040
2060
23
Conclusion
progress that are stated as distributions of outcomes rather than point forecasts. We assume
that all technologies follow a similar process except for their rates of improvement (the drift
term) and volatility. Under this assumption we
can pool forecast errors of different technologies
to obtain an empirical estimation of the distribution of forecast errors.
One of the essential points of this paper is
that the use of many technologies allows us to
make a better forecast for a given technology,
such as solar PV modules. Although using many
technologies does not affect our point forecast,
it is the essential element that allows us to make
a distributional forecast. The point is that by
treating all technologies as essentially the same
except for their parameters, and collapsing all
the data onto a single distribution, we can pool
data from many technologies to gain confidence
in and calibrate our method for a given technology. It is of course a bold assumption to say that
all technologies follow a random process with
the same form, but the empirical results indicate that this hypothesis is useful for forecasting. Of course it is always possible that other
methods (e.g. more detailed hypotheses) would
reveal more profound differences in the random
processes characterising the improvement of different technologies.
We do not want to suggest in this paper that
we think that Moores law provides an optimal forecasting method. Quite the contrary, we
believe that by gathering more historical data,
and by adding other auxiliary variables, such as
production, R&D, patent activity, there should
be considerable room for improving forecasting
power. Nonetheless, our results provide a benchmark against which others can be measured.
They provide a proof of principle that technologies can be successfully forecast, and that the
errors in the forecasts can be predicted.
From a policy perspective we believe that
these methods can be used to provide objec-
Appendix
A
Data
The data are mostly taken from the SantaFe Performance Curve DataBase, accessible at
pcdb.santafe.edu. The database has been constructed from personal communications and
from Colpier & Cornland (2002), Goldemberg
et al. (2004), Lieberman (1984), Lipman & Sperling (1999), Zhao (1999), McDonald & Schrattenholzer (2001), Neij et al. (2003), Moore
(2006), Nemet (2006), Schilling & Esmundo
(2009). The data on photovoltaics prices has
been collected from public releases of Strategies Unlimited, Navigant and SPV Market Research. The data on nuclear reactors in Table 1
is from Koomey & Hultman (2007). The data
for nuclear power in Figure 1 is from Cooper
(2009). The DNA sequencing data is from Wet20
If Z N (0, 1), U p
2 (r), and Z and U are
independent, then Z/ U/r t(r). Taking Z
from Eq. 23, U from Eq. 24 and assuming independence, we find that the rescaled normalized
forecast errors have a Student t distribution
1 E
t(m 1).
AK
Distribution of forecast
errors
B.1
Note that the t distribution has mean 0 but variance df /(df 2), where df are the degrees of
freedom. Hence the expected squared normalized forecast error is
"
2 #
1 E
m1
1 E
E
= 0 + V ar
=
,
m3
AK
AK
N (, K 2/m).
B.2
(20)
i=t+1
ni N (0, K 2 ).
(21)
As a result, using Eq. 8 we see that the distribution of forecast errors is Gaussian
E = (
) +
t+
X
i=t+1
ni N (0, K 2A),
yt+ = yt + +
(22)
[vi + vi1 ].
E = yt+ yt+ = (
) +
t+
X
i=t+1
N (0, 1).
AK
(25)
t+
X
[vi + vi1 ].
i=t+1
t1
t1
1 X
1 X
(yi+1 yi ) = +
[vi+1 + vi ]
m i=tm
m i=tm
to obtain
E=
m
(24)
21
t1
X
i=tm
[vi+1 + vi ] +
t+
X
i=t+1
[vi + vi1 ].
t1
(1 + ) X
vi
E = vtm
m
m
i=tm+1
t+
1
X
vt + (1 + )
vi + vt+ .
+
m
i=t+1
10000
C
simulation
analytical
asymptotic
Robustness checks
m=5
1000
C.1
=2
10
( )
10
( )
15
100
m=40
20
m=15
10
12
14
16
50
=5
= 15
( )
80
Forecast horizon
50
20
40
250
20
16
150
10
14
( )
100
12
60
10
350
10
m
12
14
16
10
12
14
16
Figure 14: Empirical mean squared normalized forecast errors as a function of the size of learning window, for different forecast horizons. The dots are
the empirical errors, and the plain lines are those
expected if the true model was an IMA(1,1) with
= 0.25.
= w
(1 + )
A
+ (m 1)
m
m
2
+
+ ( 1)(1 + )2 + 1.
m
22
C.3
C.2
()
Increasing max
10 20
50
Data selection
200
500
2000
14 shows that the normalized mean square forecast error consistently decreases as the learning
window increases.
10
20
50
Forecast horizon
Figure 16: Robustness to increasing max . Main results (i.e as in Figures 6 and 8) using max = 73.
We use again = 0.25
50
100
200
500
10
20
()
C.4
All technologies
Half technologies: Mean
Half technologies: 95% C.I.
1
10
To check the effect of non-normal noise increments on ( ), we simulated random walks with
drift with noise increments drawn from a Student distribution with 3 or 7 degrees of freedom.
Figure 17 shows that fat tail noise increments do
not change the long horizon errors very much.
While the IMA(1,1) model produces a parallel
shift of the errors at medium to long horizons,
the Student noise increments generate larger errors mostly at short horizons.
20
Forecast horizon
23
0.0
20
0.2
0.4
0.6
4
1
Empirical
Simulation: mean
Simulation: 95% C.I.
^ ^ ) x 100
ln(IMA
RWD
2
0
1
2
10
^ ^ )x100
ln(IMA
RWD
10
15
20
10
()
50
100
200
IMA(1,1)
RWD, Normal
RWD, t(df=3)
RWD, t(df=7)
Figure 19: Using the IMA model to make better forecasts. The right panels uses = 0.25
10
20
50
Forecast horizon
0.0
0.4
0.8
0.245
w()
1.4
*m
1.0
1.8
0.255
2.2
*w = E(w())
10
15
20
mation problems. After removing these 8 technologies the mean with equal weights for each
technology is 0.27 with standard deviation 0.35.
We can also compute the weighted mean at each
forecast horizon, with the weights being equal to
the share of each technology in the number of
forecast errors available at a given forecast horizon. In this case, the weighted mean w ( ) will
not necessarily be constant over time. Figure 18
(right) shows that w ( ) oscillates between 0.24
and 0.26. Taking the average
P20over the first 20
1
periods, we have w = 20
=1 w ( ) = 0.25,
which have used as our main estimate of in
the previous section, in particular, figures 6, 7
and 8. When doing this, we do not mean to imply that our formulas are valid for a system with
heterogenous i ; we simply propose a best guess
for a universal .
The second approach is to select in order to
match the errors. As before we generate an artificial data set using the IMA(1,1) model. Larger
values of imply that using the simple random
walk model to make the forecasts will result in
higher forecast errors. Denote by ( )empi the
empirical mean squared normalized forecast error as depicted in Figure 6, and by ( )sim,
the expected mean squared normalized forecast
error obtained by simulating an IMA(1,1) process 3,000 times with a particular global value
of . We study the ratio of the these two,
averaged over
1 . . . max = 20 periods, i.e.
P20all(
)empi
1
Z() = 20
=1 ( )sim, . The values shown in
8 10
400
Frequency
200
1500
Frequency
4
0.0
0.2
0.4
0.6
0.8
0.00
0.04
0.08
10
15
400
max
200
Frequency
600
0
Frequency
()2
200
600
0 200
Frequency
1000
max
( )2
200
500
600
400
Frequency
600
We consider the results obtained when pooling the forecast errors at all horizons, up to the
maximum horizon max = 20. We construct the
empirical CDF of this distribution, split its support in 1000 equal intervals, and compare the
empirical value in each interval k with the theoretical value, taken from the CDF of Student
distribution with m 1 degrees of freedom. The
difference in each interval is denoted by k .
|Z 1| is minimum is at m
= 0.63.
We also tried to make forecasts using the
IMA model to check that forecasts are improved:
which value of allows the IMA model to produce better forecasts? We apply the IMA(1,1)
model with different values of to make forecasts (with the usual estimate of the drift term
0.0
0.4
0.8
1.2
0.02
0.06
0.10
Figure 20: Expected deviations from the Student distribution. The histograms show the sampling distribution of a given statistic, and the thick black line
shows the empirical value on real data. The simulations use = 0.25 (3 upper panels) and = 0.63
(3 lower panels)
Three
of deviation are
P different
P measures
2
used:
k |k |,
k (k ) , and max k . These
measures are computed for the real empirical
data and for 10,000 simulated similar datasets,
a larger (m
= 0.63), these numbers are 0.21,
0.16, 0.20, so if we take a large enough we
cannot reject the model based on the expected
Benson, C. L. & Magee, C. L. (2014b), Quantitative determination of technological improvement from patent data, ESD Working Paper
Series 2014-29.
References
Alberth, S. (2008), Forecasting technology costs
via the experience curve: Myth or magic?,
Technological Forecasting and Social Change
75(7), 952983.
Albright, R. E. (2002), What can past technology forecasts tell us about the future?,
Technological Forecasting and Social Change
69(5), 443464.
Benson, C. L. & Magee, C. L. (2014a), On improvement rates for renewable energy technologies: Solar PV, wind turbines, capacitors,
Cooper, M. (2009), The economics of nuclear reactors: Renaissance or relapse, Technical report, Vermont Law School.
26
Extension to energy technology, Technological Forecasting and Social Change 75(6), 735
758.
Dutton, J. M. & Thomas, A. (1984), Treating progress functions as a managerial opportunity, Academy of Management Review
9(2), 235247.
International Energy Agency (2014), Technology roadmap: Solar photovoltaic energy (2014
ed.), Technical report, OECD/IEA, Paris.
Lipman, T. E. & Sperling, D. (1999), Experience curves for policy making: the case of
energy technologies.
Schilling, M. A. & Esmundo, M. (2009), Technology s-curves in renewable energy alternatives: Analysis and implications for industry
and government, Energy Policy 37(5), 1767
1781.
Thompson, P. (2012), The relationship between unit cost and cumulative quantity and
the evidence for organizational learning-bydoing, The Journal of Economic Perspectives
26(3), 203224.
Williams, W. & Goodman, M. (1971), A simple method for the construction of empirical confidence limits for economic forecasts,
Journal of the American Statistical Association 66(336), 752754.
Nordhaus, W. D. (2009), The perils of the learning model for modeling endogenous technological change, Technical report, National Bureau of Economic Research.
29