Professional Documents
Culture Documents
a r t i c l e
i n f o
Keywords:
Similarity measure
Dynamic time warping
Piecewise linear approximation
Time series mining
a b s t r a c t
We propose a new method to calculate the similarity of time series based on piecewise linear approximation (PLA) and derivative dynamic time warping (DDTW). The proposed method includes two phases.
One is the divisive approach of piecewise linear approximation based on the middle curve of original time
series. Apart from the attractive results, it can create line segments to approximate time series faster than
conventional linear approximation. Meanwhile, high dimensional space can be reduced into a lower one
and the line segments approximating the time series are used to calculate the similarity. In the other
phase, we utilize the main idea of DDTW to provide another similarity measure based on the line segments just we got from the rst phase. We empirically compare our new approach to other techniques
and demonstrate its superiority.
Crown Copyright 2011 Published by Elsevier Ltd. All rights reserved.
1. Introduction
Time series is an ubiquitous data form which is relative to the
time and distributed in various elds, such as the stock data (Yang,
Wang, & Philip, 2003), the data of production consumption and
web transaction (Samia & Conrad, 2007). Although some have
nothing to do with the time, it can be transformed into the form
of time series and studied by the models and algorithms of time
series. For example, the data of the shape of the tree leaves can
be treated as time series (Ye & Keogh, 2009). There is much valuable information hiding in time series, including interesting patterns (Ajumobi, Pken, & Preda, 2004; Anthony, Wu, & Lee, 2009),
anomaly points (Keogh, Lin, & Fu, 2005) and motifs (Lin, Keogh,
Lonardi, & Patel, 2002). In most cases, we need to measure the similarity (Chen, Hong, & Tseng, 2009) or dissimilarity (distance) between two time series in advance. However, dimensionality curse
of time series goes against the accurate similarity measure.
There are many ways to reduce the dimensionality, such as the
discrete fourier transform (DFT) (Agrawal, Faloutsos, & Swami,
1993; Agrawal, Psaila, Wimmers, & Zait, 1995), singular value
decomposition (SVD) (Korn, Jagadish, & Faloutsos, 1997), discrete
wavelet transform (DWT) (Chan & Fu, 1999), piecewise linear
approximation (PLA) (Manjula, Morgan, & Layne, 2008), and symbolic aggregate approximation (SAX) (Keogh & Pazzani, 1998;
Keogh et al., 2005) based on the piecewise aggregate approxima Corresponding author. Tel.: +86 41184708007.
E-mail addresses: hailin@mail.dlut.edu.cn (H. Li), guochonghui@tsinghua.org.cn
(C. Guo), qiuone@163.com (W. Qiu).
tion (PAA) (Hung & Duong, 2008). In particularity, SAX and PLA
are widely applied to many elds (Keogh, Chakrabarti, Pazzani, &
Mehrotra, 2001; Keogh, Chu, Hart, & Pazzani, 2001; Lin & Keogh,
2006) and obtain very good results.
After reducing the dimensionality of time series data, Euclidean
distance is useful and simple for similarity measure, but it has
some disadvantages. For example, the abnormal data in time series
affects the whole similarity measure. Moreover, it will abandon the
sequence query too early, which causes the false alarm when
indexing. Another popular method to compare time series in diverse areas is the dynamic time warping (DTW) (Keogh & Pazzani,
1999; Keogh & Ratanamahatana, 2005). It offers a more reasonable
measure for description of the relations between the different time
series by time warping. Its improved version, which is called derivative dynamic time series (DDTW) (Keogh & Pazzani, 2001), can
produce the more intuitive warping and better results by considering the derivative of the time series.
In this paper, we propose a novel approach to measure the similarity of time series. Firstly, a divisive approach of piecewise linear
approximation (DPLA), whose time complexity is lower than the
conventional ones, is given to approximate time series. Secondly,
we propose middle curve piecewise linear approximation (MPLA)
based on DPLA to approximate time series. What we do have two
advantages at least. One is that it is more suitable for the middle
curve to describe the local and whole trends. Because some
conventional methods based on the original time series are difcult to express the trend and easy to fall into local optimization,
it is reasonable for utilizing a middle curve to represent the
original time series for their trends. The other is the lower time
0957-4174/$ - see front matter Crown Copyright 2011 Published by Elsevier Ltd. All rights reserved.
doi:10.1016/j.eswa.2011.05.007
14733
tire dataset (Park et al., 1999). However, time complexity of topdown algorithm is larger than the two previous. Its time-consuming is O(Km2). The reason is that each recursion should calculate
the divisive errors of the two line segments when one point breaks
one sequence into two subsequences. That is, each recursion
should nd a best breakpoint to partition the present sequence
so that the approximation error of the two subsequences located
in the both sides of the breakpoint is minimum.
The PLA algorithms represent time series with some line segments. Most of the algorithms have a low computational complexity which is linear to length of the time series, but some have
higher complexity (Bauman, Dorofeyuk, & Kornilovm, 2004; Zhang
& Wan, 2008) because they are in pursuit of the optimal results. In
Section 3, we will propose another algorithm based on top-down
algorithm, whose time complexity is linear to the length of the
time series. We call it divisive piecewise linear approximation
(DPLA).
2. Related work
2.1. Piecewise linear approximation
There are many kinds of piecewise linear approximation (Keogh
& Chu et al., 2001) to reduce the dimensionality of time series data,
which can be grouped into three classes.
Window-Sliding: Once another next point joins in a window,
which results in the cost of the line segment to approximate the
subsequence larger than a threshold values, then create the line
segment of the subsequence.
Bottom-Up: Merge the adjacent segments until every merged
cost of the segments is larger than a threshold value.
Top-Down: Segment the time series from top to down recursively until some threshold value is met.
The main idea of the window-sliding algorithm is to slide the
window to the point which is not fall into the window. At the same
time, the line segment approximating the subsequence within the
window is formed suitably. The approximation depends on a
threshold value, which will cause several pathologically poor results
under some circumstances. Shatkay notices the problem and gives
some explanation (Shatkay & Zdonik, 1996). Later, a modied version (Park, Lee, & Chu, 1999) is given to improve the algorithm.
The two most important properties of the linear approximation
are the linear time complexity and the online computation.
The bottom-up method is different from the sliding window.
Firstly, it works well by the m/2 segments approximating time series of the length of m. Secondly, after merging the adjacent segments with a minimum merging cost, it must delete one of the
two merged segments and respectively recalculate the merging
cost between the other one and the new one. Redo above steps until every merged cost is larger than threshold value. Similar to the
sliding window, the bottom-up approach is widely used for time
series mining (Hunter & McIntosh, 1999) and its time complexity
is linear to the length of time series, i.e. O(Km), where K is the number of line segments. However, it is not an online algorithm.
As opposed to the above bottom-up algorithm, the top-down
algorithm approximates time series into different line segments
by searching the best location each time. At each recursive step
the algorithm should search a minimum divisive error. If the minimum error at the ith point is smaller than a threshold value e, the
subsequence of time series Q will be partitioned into two parts at
the ith point. Next, let it consider the two subsequences again
and redo the above steps until all the divisive minimum errors
are larger than the threshold value.
The top-down algorithm also has been widely used in diverse
areas, such as a framework for the sequence mining, text mining
and time series mining. Part et al. introduced a modied version
where they rst mark every peak and valley after scanning the en-
0
q0i @
ki
X
1,
qj A k;
i 1; 2; . . . ; w;
jki11
where k = m/w.
From formula (1), we know that time series data is divided into
w equal sized frames whose length is k and the element value of
new vector is the mean of points falling within the frame. The time
complexity of PAA for the measure similarity is linear to the length
of time series, i.e. O(m). The representative value of the original
time series in new space is shown in Fig. 1. SAX is a method to
change the mean values into discrete string representations (Lin
et al., 2003). Fig. 2 shows the result of SAX after PAA procedure.
It is easy to nd that there are some special values in original
space to be generalized, which cannot express the property of time
series value well, such as local trend, whole trend and individual
point distribution. There are several different cases hiding the
2
1.5
1
0.5
0
0.5
1
1.5
2
0
10
20
30
40
50
60
14734
3
2
1
0
1
10
20
30
40
50
60
Fig. 2. Symbolic representation of time series by SAX, but they only reect the
trend of some points.
DTWQ ; C minwl
( s )
k
P
1
wl :
k
l1
Generally, the optimal path can be found by using dynamic programming method
8
>
< ri; j 1;
ri; j di; j min ri 1; j 1;
>
:
ri 1; j:
...
w
w2
w1
30
30
25
25
20
20
15
15
10
10
10
20
30
40
50
60
10
20
(a) DTW
30
(b) DDTW
Fig. 4. The difference of the DTW and DDTW.
40
50
60
14735
Di q
D0 i; j Di q Dj c2 :
3.5
3
2.5
2
2
1.5
1
10
15
Fig. 5. The curve have strong turning. It has three different trend (descending, at
and ascending), but PAA cannot not express the sharp turning and trend well.
Li : j
qi j qj i qj qi
t;
ji
ji
i 6 t 6 j:
8
qi
< a qjji
;
: b qi jqj i :
ji
Dqt ; Li : j
jat qt bj
p :
a2 1
For a time series Q of length m or subsequence Q(i : j) with discrete real values, if we directly link every two adjacent points with
m 1 or j i line segments, the distance of every point to the corresponding line segment is equal to 0. It means the line segment
approximation of time series is the best and the approximation
cost is minimum (equal to 0). Although it is best to approximate
time series, it causes no dimensionality reduction and is meaningless. Therefore, we let the distance between the point and the corresponding line segment close to some value rather than the
minimal distance. In other words, let the distance be larger than
a dened threshold value e.
For subsequence Q(i : j), if the distance of point ql(i 6 l 6 j) to the
line segment L(i : j) is maximum and D(ql, L(i : j)) > e, the point ql is
a breakpoint and subsequence Q(i : j) is divided into two parts, Q(i :
l) and Q(l : j); otherwise, the line segment approximates subsequence Q(i : j) well, it is not necessary to divide the subsequence
Q(i : j) any more.
The algorithm of divisive piecewise linear approximation is
shown as follows.
Step 1: Input time series Q(i : j) and threshold value e. A vector Bp
is used to restore the breakpoints. k records the number of
the present breakpoints. pos denotes the position of the
newest breakpoint. Initially, i = 1, j = m, where m is the
length of time series. Since the rst point and the last point
are the special breakpoints, let k = 2, Bp(1) = q1 and
Bp(2) = qm.
Step 2: For time series Q(i : j), create line segment L(i : j) according
to the formula (6). Set two variables l = i + 1 and
best_so_far = 0.
14736
Step 3: Calculate the distance of point ql to the line segment L(i : j),
that is D(ql, L(i : j)).
Step 4: If D(ql, L(i : j)) > best_so_far, best_so_far = D(ql, L(i : j)) and
pos = l.
Step 5: l = l + 1. If l P j, go to the step 6; otherwise, go back to step
3.
Step 6: If best_so_far P e, k = k + 1, Bp(k) = qpos, go back to the Step
2 and let the two subsequences Q(i : pos) and Q(pos : j)
redo the step 2 to step 6, respectively.
Step 7: Sort the element of vector Bp by an ascending time and
output the sorted result.
the rst line segment L(1 : m) as the upper bound of the permissible divisive condition. That is
e STDDqt ; L1 : m
s
m
m
1 X
1 X
Dqt ; L1 : m
Dqt ; L1 : m2 :
m t1
m t1
1.5
1.5
0.5
0.5
0.5
0.5
1.5
1.5
10
20
30
40
50
60
1.5
1.5
0.5
0.5
0.5
0.5
1.5
1.5
10
20
30
40
50
60
10
20
30
40
50
60
10
20
30
40
50
60
Fig. 6. The number and the trends of line segments are various according to the different threshold values.
14737
5
4
L1
3
L3
2
1
0
1
2
L2
0
10
20
30
40
50
60
Fig. 7. L1 is far away from most of the points in time series, which will cause the
standard deviation large. We should look for the next line segment to replace it, and
nd that the L2 embraces most of points in subsequence Q(1:59). So we should
regard L2 as L1 and apply the formula to compute the e value. Since the distance
deviation of the points in subsequence Q(59:60) to L3 is equal to zero, L3 is not
suitable to be regarded as L1. Actually, it is one of the nal line segments
approximating the subsequence.
Actually, the distance D(qt, L(i : j)) of the point qt to the line segment L(i : j) can be replaced by the length of the right angled side of
a right-angled triangle. According to the right triangle theorem, the
length of the hypotenuse is longer than any other right-angled
sides. Therefore, to calculate the distance faster, we use
Dqt ; Li : j jqt at bj
qi j qj i qj qi
qt
ji
ji
10
instead of the formula (8), which can speed up the whole divisive
piecewise linear approximation algorithm.
3.2. Middle curve-based piecewise linear approximation
In Fig. 8, some line segments with sharp slopes cannot express
the trends of subsequences well. For example, the line segment
marked a directly links the two endpoints in disregard of any
point between them. All the points of the subsequence only appear
on the bottom-left side of the line segment. It is not reasonable for
this approximation. Likewise, for the subsequences in time series
marked b, there are too many line segments to approximate the
trends. Actually, it is easy to nd that the sequences are smooth except for some frequent amplitudes and can be approximated naturally by one or two line segments with slightly at slopes.
Why does it happen in our algorithm? The reason is that we
search the breakpoints and link them directly without considering
1.5
2
1
0.5
0
0.5
1.5
2
10
20
30
40
50
1
0
10
20
30
40
50
60
60
Fig. 9. Three points in time series have large amplitudes and should be saved for
the middle curve with considering the trends of points.
14738
are the upper and lower points. To execute the proposed algorithm
conveniently, as illustrated in Fig. 10, we only interpolate several
new values into the line segments (L12 and L23).
Given a line segment L(i : i + 1), if it is interpolated with n values, the interpolated line segment Ll : l n0 q0l ; q0l1 ; . . . ; q0ln ,
i.e.
0 6 j 6 n:
11
Mi
Ti Ti 1
;
2
i 1; 2; . . . lengthT 1:
2
1.5
0.5
0
0.5
1
1.5
2
10
20
30
40
50
60
Fig. 11. The upper bound and lower bound of time series. The points marked A
represent the lower bound points, the points marked B mean the upper bound
points and the points marked C denote upper bound point and lower bound point
together.
2
1.5
1
0.5
0
0.5
1
1.5
2
10
20
30
40
50
Fig. 12. The middle curve and the original time series.
60
14739
So, we transform the data with formula (5) and calculate the
similarity by DDTW. The warping result is shown in Fig. 14. It is
well known that the complexities of DTW and DDTW are O(n m),
where n and m are the length of Q and C, respectively. However,
the time complexity of our approach is O(N M), where N and M
are the number of endpoints of line segments approximating Q
and C. They are much smaller than m and n, that is, N n and
M m.
4. Experimental evaluation
Many experiments have already demonstrated that the SAX
algorithm based on PAA and its extended versions are useful and
feasible for time series mining by Keogh and Lin (2002). Therefore,
in this section, we mainly compare our method with SAX algorithm.
4.1. Approximation results
Fig. 13. The process of creating the middle piecewise linear approximation.
Ldi; j
12
otherwise;
where
d(i, j) = (qbpi(2) cbpj(2))2,
P(i, j) = ((jqbpi(1) qbpj(1)
2
j r)w) and w is the mean of Euclidean distance of the subsequences with the length of m chosen from two time series. That
q
Pm
2
is, w m1
i1 qi c i ; m < minlengthq; lengthc.
From the formula (12), we know that if the time difference of
two endpoints respectively from two line segments surpasses the
window length r, we need to punish the distance of q(i) and c(j)
so that time series in the different groups have a larger distance
values.
Compress Ratio
13
From Fig. 16, it states that MPLA automatically creates the 73 line
segments to approximate time series with 300 points. The lines express the original time series well. They not only reect the whole
trend but also describe the local trend. Of course, if we want use less
number of line segments to express the original time series, we only
need to use the man-made setting of the threshold e to control the
line segments in the process of MPLA.
Since SAX is based on PAA, we make another experiment illustrate the average errors of the approximations to the original time
series for PAA and MPLA respectively. We still use the Stock data
web page (2005) dataset and choose arbitrarily 30 subsequences
of length 2000. When reducing dimensionality by the two methods, we deliberately let the dimensionality number of PAA larger
than that of MPLA as shown in Fig. 17(a). It means that the approximation error of the SAX should be smaller than that of MPLA.
However, in Fig. 17(b), the result of the experiments demonstrates
that the error of MPLA is smaller than that of the PAA in spite of its
12
0.29
10
0.28
Compress ratio
8
6
4
2
0.27
0.26
0.25
0.24
0.23
0.22
0.21
0.2
1000 2000 3000 4000 5000 6000 7000 8000 9000 10000
2
10
20
30
40
50
60
Fig. 14. The warping result of middle line segments based on DDTW.
14740
0.4
0.6
time series
line segments
0.6
time series
line segments
0.8
0.8
1
1.2
1.2
1.4
1.4
1.6
1.6
1.8
1.8
2
50
100
150
200
250
300
50
100
(a)
150
200
250
300
(b)
Fig. 16. (a) The separated curves of line segments and time series. (b) The overlap of line segments and time series, which means the line segments t the time series well.
1000
0.016
Dimensionality number of PAA
Dimensionality number of MPLA
900
0.014
PAA Error
MPLA Error
700
0.012
Error rate
Dimensionality number
800
600
500
400
0.01
0.008
300
200
0.006
100
0
500
1000
1500
2000
0.004
500
1000
1500
2000
Fig. 17. (a) Dimensionality number of PAA is greater than that of MPLA. (b) The error of MPLA is smaller than that of PAA to approximate the original time series, whose
length is 2000. It means that we can use the lower dimensionality of the MPLA to approximate the time series better than the higher dimensionality of PAA.
4
3
1
12
10
11
7
6
2
5
13
14
9
8
4.2. Clustering
To testify the new similarity measure, we experiment clustering
study on synthetic control chart dataset (Alcock & Manolopoulos,
1999) derived from UCI data. Comparing the hierarchical clustering
is one of the best ways for similarity comparison. The evaluation is
typically objective. We only observe which dissimilarity measure is
close to the Euclidean distance. In view of the results of comparing
the Euclidean, SAX, IMPACTS and SDA in the paper Lin et al. (2003),
we know that the SAX was the best one whose distance measure is
close to Euclidean. So we only compare our method with Euclidean
and SAX and observe that whether our method is better than SAX.
If it is true, our method is at least better than SAX, IMPACTS and
SDA. We arbitrarily choose 14 time series to cluster with the hierarchy clustering. The results are shown in Figs. 1821, which is
used by different methods, including Euclidean, DDTW, SAX and
MPLA. Note that the 14 data objects derive from 6 groups,
normal{1, 2}, cyclic{3, 4, 5}, increasing trend{6, 7}, decreasing
trend{8, 9}, upward shift{10, 11, 12}, downward shift{13, 14}.
From Figs. 1821, we know that the clustering result of DDTW
is the worst, our method is at least better than DDTW. It also
means that our method is an improved DDTW because of the
40
(a)
60
80
100
(b)
14741
1
2
6
11
7
12
9
3
10
14
13
8
5
4
0.2
0.25
(a)
0.3
0.35
(b)
Fig. 19. The best cluster result of the derivative dynamic time warping (DDTW).
13
14
9
8
5
3
2
4
1
12
6
11
10
7
1
(a)
(b)
Fig. 20. The best cluster result of SAX whose parameter is w = 10, word_size = 9.
5
4
3
2
1
12
10
11
7
6
9
8
14
13
2.5
2
5
(a)
Especially, for the 1st and 2nd time series, our method does not
group them into normal group at the prior phase, which is the
same to the Euclidean. But SAX is prior to combine the 2nd with
the 1st and 4th because of considering of the means of the points
within the frames rather than the local trend. For considering local
trend of time series, the cyclic group (3rd, 4th, 5th) should be analyzed because the cyclic property can better to reect the local
trend. We nd that SAX cannot classify the cyclic group into the
same cluster, but our method is able to and prior to do it. It is very
important that time series {6, 7}, {8, 9}, {10, 12}, {13, 14} are
grouped into the same class respectively, but SAX and other methods cannot. Therefore, our method can deal with the local trend
and the whole trend simultaneously.
Apart from the above advantages, we need to point out that
there is none of parameter to be set for the MPLA. However, the
SAX needs to set two parameters (the number of segments w
and alphabet sizes a) which are usually hard to decide. Fig. 20 is
the best clustering result by adjusting the two parameters.
Although the MPLA essentially has one threshold to set, yet it is
easy to be dened by formula (9). According to the different
parameters, SAX produces the rough clustering result as shown
in Fig. 22. It means that we must have the priori information to
set the parameters, otherwise we could not get good clustering
results.
In additional, for the similar shapes of different time series, our
approach can recognize it. For example, the similar shapes in two
time series are shown in Fig. 23. If we use the SAX to measure
the similarity, it cannot get the proper result of shape comparison,
which should be equal to 1. However, the similarity result of our
approach is equal to 1, which reects the trend of the whole sequence and subsequence well in time series. Therefore, it is better
for our method to measure the similarity of the time series in spite
of equal and unequal length. It is benecial to similarity search for
time series mining.
10
15
(b)
1.5
1
0.5
trend and neglects the local trend, it is prior to classify the 3rd and
5th time series into the same group. Since our method considers
the local and whole trend of time series, it not only classies the
4th and 5th time series into the same cluster in advance but also
clusters the 3rd,4th and 5th time series into the same group.
11
12
10
14
13
Fig. 22. The result of SAX used to cluster is rough when the two parameters is set
by w = 10 and a = 4, which demonstrates that it does not consider the local trend at
all.
14742
Fig. 23. (a) Shows the two time series with the very similar shape in the same frame of axes; (b) shows the result of the SAX algorithm to approximate; (c) shows the new
method to approximate.
0.8
0.8
0.6
0.4
0.6
0.4
0.2
0.2
0.2
0.2
c
b
0.4
0.4
0.6
0.6
0.8
0.8
1
0
500
1000
1500
2000
a
5
10
15
(a)
20
25
30
35
40
45
50
(b)
Fig. 24. (a) The result of similarity search. (b) The zoom view of the result of similarity search.
algorithm based on our method not only can search the pattern sequence marked a but also discover the similar subsequences in
time series Q(1 : 2000). They are the subsequences marked a,
b, c and d. Therefore, the similarity research based on our
method also can get a good result.
5. Conclusions
In this paper, we propose a new method to measure the similarity of time series. DPLA adaptively divides time series into unequal
segments, which does not require any parameter to preset. Moreover, its time complexity is O(kn), which is much lower than the
conventional topdown linear approximation. We use MPLA,
which is based on DLPA, to approximate time series with line segments. It is better to reect the trends of the subsequence and the
whole sequence in time series. Moveover, the line segments produced by MPLA to approximate the middle curve of the original
time series make the approximation error rate less. For the particularity of the results of the MPLA, the modied derivative dynamic
time warping is proposed to calculate similarity of time series,
which is benecial to separate the members from different groups.
The empirical results demonstrate that the new similarity measure
is an effective method for time series mining.
The experimental analysis shows that our method is better for
dimensionality reduction and similarity measure, but the time
complexity is a little higher than SAX. The reason is that our method
is based on derivative dynamic time warping to calculate the
similarity.
The symbolic representations to measure time series can be applied in our method. We have found a way (Li et al., 1998) to objectively describe the slopes or angles of line segments as symbolic
representations. Perhaps our method is a good choice for improving the present and available methods of SAX to nd motifs and
discover rules. Of course, we also can use our method to do other
work for time series mining, such as classication, clustering and
other time series mining tasks.
Acknowledgments
This work is supported by the Natural Science Foundation of
China (70871015 and 71031002) and the Fundamental Research
Funds for the Central Universities (DUT11SX04). We also would
like to acknowledge Prof. Eamonn Keogh for the dataset and his
procedure source code.
References
Agrawal, R., Psaila, G., Wimmers, E. L., & Zait, M. (1995). Querying shapes of
histories. In Proceedings of the 21st international conference on very large
databases, Zurich, Switzerland (pp. 502514).
Agrawal, R., Faloutsos, C., & Swami, A. (1993). Efcient similarity search in sequence
databases. In Proceedings of the 4th international conferences on foundations of
data organization and algorithms. Chicago: Springer-Verlag.
Ajumobi, U., Pken, B., & Preda, A. (2004). Discovering all frequent trends in time
series. In Proceedings of the winter international synposium on information and
communication technologies (Vol. 6, pp. 16).
Alcock, R. J., & Manolopoulos, Y. (1999). Time-series similarity queries employing a
feature-based approach. In 7th Hellenic conference on informatics (pp. 19).
14743
Li, D. Y., Han, J. W., Shi, X. M., & Chan, M. C. (1998). Knowledge representation and
discovery based on linguistic atoms. Knowledge-based Systems, 10, 431440.
Lin, J., & Keogh, E. (2006). Group sax: extending the notion of contrast sets to time
series and multimedia data. In Proceedings of the 10th European conference on
principles and practice of knowledge discovery in databases (pp. 284296).
Lin, J., Keogh, E., Lonardi, S., & Patel, P. (2002). Finding motifs in time series. In The
8th ACM international conference on knowledge discovery and data mining (pp.
5368).
Lin, J., Keogh, E., & Lonardi, S. (2003). A symbolic representation of time series, with
implications for streaming algorithms. In Proceedings of the 8th ACM SIGMOD
workshop on research issues in data mining and knowledge discovery (Vol. 7, pp.
211).
Manjula, A. I., Morgan, M. H., & Layne, T. W. (2008). A performance comparison of
piecewise linear estimation methods. In Proceedings of the 2008 Spring
Simulation Multi-Conference (Vol. 4, pp. 273278).
Park, S., Lee, D., & Chu, W. W. (1999). Fast retrieval of similar subsequence in long
sequence databases. In Proceedings of the 3nd IEEE Knowledge and Data
Engineering Exchange Workshop (pp. 6067).
Rabiner, L., & Juang, B. (1993). Fundamentals of speech recognition. Englewood Cliffs,
N.J: Prentice Hall.
Sakeo, H., & Chiba, S. (1990). Dynamic programming algorithm optimization for
spoken word recognition. Readings in Speech Recognition, 159165.
Samia, M., & Conrad, F. (2007). A time-series representation for temporal web
mining using a data band approach. In Proceedings of the 2007 conference on
databases and information systems IV (Vol. 6, pp. 161174).
Shatkay, H., Zdonik, S.B. (1996). Approximate queries and representations for large
data Sequences. In Proceedings of the 12th International Conference on Data
Engineering (pp. 536545).
Stock data web page, 2005. <http://www.cs.ucr.edu/wli/FilteringData/stock.zip>.
Yang, J., Wang, W., & Philip, S. Y. (2003). Mining asynchronous periodic patterns in
time series data. IEEE Transactions on Knowledge and Data Engineering, 15,
613628.
Ye, L. X., & Keogh, E. (2009). Time series shapelets: A new primitive for data mining.
In International conference on knowledge discovery and data mining, Paris, France
(pp. 947956).
Zhang, H., & Wan, S. N. (2008). Linearly constrained global optimization via
piecewise-linear approximation. Journal of Computational and Applied
Mathematics, 214, 111120.