Professional Documents
Culture Documents
Abstract—Many smart data services (e.g., smart energy, than 2.5 micrometers) composition particles contribute to
smart homes) collect and utilize time series data (e.g., energy different types of diabetes [5]; One PM2.5 particle can be
production and consumption, human body movement) to con- considered as one variable in some MTS data. Identifying
duct data analysis. Among such analysis tasks, classification is
a widely utilized technique to provide data-driven solutions. which PM2.5 particles contribute to specific diabetes (e.g.,
Most existing classification methods extract a single set of Type-2 diabetes) is a challenging problem in this domain
features from the data and use this feature set for classification and revealing this information would be a major benefit to
across multiple classes. This often ignores the reality that researchers.
different and class-specific subsets of the initial feature set Compared with other data, MTS data contains time-
may better facilitate classification. In this paper, we propose
two convolutional neural network (CNN) models using class- dependency in each time series. How to model and analyze
specific variables to solve the multi-class classification problem this time-dependency is one major challenge in MTS classifi-
over multivariate time series (MTS) data. A new loss function cation. Many algorithms can successfully classify MTS data,
is introduced for training the CNN models. We compare our but very few of them can directly capture the relationships
proposed methods with 13 baseline approaches using 14 real between classes and specific features.
datasets. The extensive experimental results show that our
new approaches can not only outperform other methods on Convolutional neural networks (CNN) have shown high
classification accuracy, but also successfully identify important accuracy in time-series classification [2], [6], [7], [8]. The
class-specific variables. CNN approach is capable of automatically extracting fea-
Keywords-Smart data-services; Multivariate Time Series; tures from the MTS data and utilizing such features to
Classification recognize different classes. In this paper, we propose two
newly designed convolutional neural network (CNN) based
I. I NTRODUCTION approaches by leveraging class-specific variables. These two
Smart data-services provide data-driven solutions to de- approaches redesign the fully connected layer in a typical
cision makers by utilizing different types of data. Much of CNN structure to reveal and leverage the class-specific
such data is in the form of multivariate time series (MTS) variables. Identifying these class-specific variables is very
and collected through smart infrastructures, such as smart important, especially in neural network-based algorithms [9].
energy, smart homes, and smart healthcare. For example, We show that our CNN approaches can not only achieve
multiple current and frequency waveforms (MTS data) col- better classification performance than state-of-the-art meth-
lected in smart energy systems can be used to identify system ods but also successfully identify the class-specific variables,
faults; the movement information of the different parts of a which gives us a better understanding of the classes in the
human (MTS data) collected by sensors can help identify data. The main contributions of this paper are the following.
the type of movement this person is making; the amount • We propose two novel CNN based approaches with
of the many PM2.5 composition particles in a period of class-specific variables. The class-specific variables are
time (MTS data) can help monitor air quality. Analyzing identified and leveraged during model training.
MTS data has received increasing interest in the past decade • A new loss function is designed to fit the structure of
due to the deployment of smart infrastructures. Extracting the proposed CNN models. The new loss function can
features and classifying MTS are applied in various smart better separate similar classes and address the issue with
data-services, including human activity recognition [1], [2], imbalanced datasets.
healthcare [3], voice processing [4] and many others. In • The proposed approaches can identify important class-
MTS classification problems, besides generating accurate specific variables. Compared with other state-of-the-art
predictions, it is also important to understand the features methods, our approaches can identify the class-specific
or factors that are most critical for prediction interpretation variables more effectively and efficiently.
or decision making. For example, recent research shows • We evaluate the classification performance of the pro-
that different PM2.5 (particulate matter with diameter less posed approaches with four state-of-the-art methods
and nine other baselines. Our experiments on 14 contains two variables representing the height of the sensors
datasets show that the proposed approaches achieve the on someone’s left arm (LA) and left leg (LL). Fig. 1 draws
best averaged accuracy. the LA sequences of the three instances.
• We compare the identified class-specific variables from
Class Variables Time sequences
our approach with the features identified from state- LA 10, 20, 29, 39, 40, 40, 40
Standing in Elevator (SE)
of-the-art methods. Our approach can identify similar LL 5, 15, 25, 33, 35, 35, 35
LA 12, 18, 31, 37, 42, 38, 41
class-specific variables with much better efficiency. Moving around in Elevator (ME)
LL 7, 14, 27, 30, 37, 34, 37
The paper is organized as follows. Section II formally Playing Basketball (PB)
LA 10, 14, 18, 13, 9, 7, 10
LL 4, 6, 9, 8, 7, 3, 5
defines the problem and related terminology. Section III
∗
LA/LL: the y-coordinate of the left arm/leg sensor
presents our proposed approaches. Section IV presents our
experiments and shows the effectiveness and efficiency of Table II: Example dataset
our proposed approaches. Section V discusses the literature.
Finally, Section VI concludes our work. ÇD
ÇĆ
ĊD
II. P ROBLEM F ORMULATION AND T ERMINOLOGY ĊĆ
À BĀǺǼĂÆ
ČD
Multivariate time series (MTS) data records values for ČĆ
ÃA
ÁA
multiple variables in a period of time. (e.g., daily tempera- ĈD
ĈĆ Â
ture and humidity over one month). Each MTS consists of D
ÄÅĄĂ
vi is a numerical value recorded for one variable at the i-th
time point and m is the length of the time series.
Figure 1: The LA sequences for the example dataset
When conducting classifications, it is important to extract
useful features from MTS data. Choosing an optimal selec-
tion of features can impact both accuracy and performance A. Motivation of Proposed Approaches
of the classification, and may reveal underlying associations In classification problems, when an instance is wrongly
between features and classes. For example, an association classified to a class label (say c), a common reason is that
between certain medical readings and the classification of a this instance’s features are similar to the features of class
patient’s disease could improve understanding of how that c. We observe that while some instances in a dataset can
disease develops. However, determining optimal features in be easily classified with the correct class labels (no matter
MTS data can be very challenging. This paper works on which classifiers are used), other instances may be mis-
MTS classification and class-specific variable (CV) identifi- classified to classes with similar features. Let us use fuzzy
cation simultaneously. class to denote the classes that can be easily confused with
Definition 1 (Research Problem): Given an MTS with (and be classified as) other classes and use non-fuzzy class
multiple class labels, the problem is to accurately classify to denote the other classes.
instances of the MTS by extracting and utilizing class- Given the example, a CNN can benefit from class-specific
specific variables. variables that can differentiate the “SE” class and the “ME”
class. One (or multiple) fully-connected layer(s) for all the
Symbol Meaning
classes cannot differentiate instances belonging to fuzzy
C # of distinct classes in a dataset
V # of variables in an MTS dataset classes. With such fuzzy-class instances, many iterations
N # of instances in an MTS dataset used to train a neural network model may not improve
m length of one time series in an MTS dataset
Y the actual class labels of the instances accuracy much. Instead, accuracy may become worse in
later iterations. For example, in one iteration, many instances
Table I: Symbols used in this paper belonging to fuzzy classes are wrongly classified into a
class with similar features. In the next iteration, the adapta-
III. M ETHODOLOGY tions adjusted by the learning algorithm may help correctly
classify instances in one fuzzy class, but may wrongly
This section presents our CNN based approaches to classify instances in another fuzzy class. A desirable model
conduct the multi-class classification of MTS data. Table needs to identify features that help differentiate instances in
1 summarizes the meaning of major parameters. We use a the fuzzy classes. We design class-specific fully-connected
toy dataset throughout this paper to explain the concepts and components to alleviate this issue.
the computations of our algorithms.
Example 1 (MTS data): Table II shows a small dataset B. CNN with Class-Specific Fully-Connected Components
with three classes: Standing in Elevator (SE), Moving around A CNNmts [2], [9] model for MTS consists of several
in Elevator (ME), and Playing Basketball (PB). This dataset convolutional layers, pooling layers, and fully-connected
'=55A .8770.<0/ -58.4
D
JU N JU O JU WQVPN 5,A0:; JC R
8=<9=<
D
G A C
+53/371 G A JU C
FBBA IBA G H D
?37/8? GEAA JNST C
F A I A GEHH JOST
5071<2%4! È ǺĄǼČÅĆĈ C
FEAA IEAA C
$"# % 6,@ $"#" $"$" B" $":!
JUST
ÂĄĊÅÇ À B Ã È ÂĄĊÅÇ À ÐÐBĐÆÊÈEĎÉE Ã È ÇÀ È Ã A
ÂĄĊÅÇ À ÐBĐÆÊÈE Ã
Figure 2: CNN structure for MTS using in [9] (LCP: last convolutional or pooling layer)
EK
A nents. The c-th component is responsible for calculating the
Ĉ A
EGIJ A probability that one instance belongs to class c. Fig. 3 shows
D
C A the structure of the FCC layer in our CNN-CF.
EKIJ
Ĉ
Ĉ
Ĉ
A
ĆÀ Č Â
A 1) Set up the convolutional and pooling layers in the CNNmts model
EK A 2) Let L be the matrices of LCP, which are vectorized to
Ĉ
EGIJ
A [(h11 , · · · , hV 1 ), · · · , (h1F L , · · · , hV F L )]
A 3) Make C copies of L: L1 , · · · , LC
EKIJ 4) Initialize an array Mcf to keep the weights for C fully-connected compo-
nents
5) For each class label c (c = 1 · · · C), construct a Yc vector
Figure 3: The FCC layer in CNN-CF a) For the j-th instance in Y
i) if (Y [j] == c) Yc [j] = 1 else Yc [j] = 0
b) Use the weights in Mcf [c] to connect L and Yc
6) Train a constructed CNN model
layers.
1) Overall Design of CNN-CF: The design of our Figure 4: Algorithm to construct a CNN-CF model
CNN model with Class-specific Fully-connected compo-
nents (CNN-CF) is built upon the CNNmts model [9] , as The pseudo-code of constructing a CNN-CF model is
shown in Fig. 2. Given the MTS input with size N × m × V shown in Algorithm CNN-CF-Construction (Fig. 4). The
(N instances where each instance keeps V sequences with convolutional and pooling layers are set as in CNNmts
length m), different from the regular CNN model for image (Line 1). It flattens the matrices in LCP to vectors and makes
classification, whose convolutional (pooling) filter size is C copies of the vectorized hidden units, L1 , · · · , LC . The
k × k (r × r), the convolutional (pooling) filter in CNNmts c-th fully connected component links the Lc to the output
has size k × 1 (r × 1). The convolutional and pooling layer of predicting an instance to be class c or not c.
filters are applied to each time series variable. The last 2) Loss Function of CNN-CF: To train any neural net-
convolutional/pool (LCP) layer of CNNmts aggregates the work model, a loss function is needed to optimize the
time series output sequences from the previous convolutional parameters (weights) of the neural network. The gradient
layer. The output from the LCP is N × 1 × V × F L (F L descent optimization method trains the parameters (weights)
is the number of filters in the LCP layer). When using of the neural network such that the loss function is mini-
a CNNmts model to conduct multi-class classification, the mized. Cross-Entropy (CE) is the most commonly used loss
fully-connected layer (or the last fully-connected layer if function. In MC problems, CE is formally defined as
there are multiple such layers) is utilized to calculate C N X
X C
probabilities for all C classes. The class with the highest CE = yc [i] × log(pc [i])
probability is selected as the predicted class. i=1 c=1
CNN-CF decomposes a large task (learning C probabil- Here, yc [i] is 1 if the i-th instance’s actual class is c and
ities for C classes) to smaller tasks with binary classifi- is 0 otherwise, pc [i] is the probability of the i-th instance
belonging to class c. Directly applying CE as a loss function C. CNN-CF with Class-Specific Features
in CNN-CF generally produces high CE values because the
Features from LCP do not equally contribute to classifying
CE calculation depends on the prediction of all the instances
various classes. Furthermore, features that can help classify
and the predictions of instances belonging to the fuzzy
instances belonging to one class may not be helpful in
classes are generally poor.
classifying instances belonging to other classes [9]. It is
Example 2: In the dataset shown in Example 1, instances
critical that we utilize different features that are important
from “SE” and “WE” are easily predicted to be one of the
in differentiating separate classes.
fuzzy classes in one iteration. The poor performance of “SE”
and “WE” increases the overall CE value. A high CE value +%%, !(''#!*#" %(!$
may cause a bad optimization of the parameters for “PB” in DFF
DDGF
+%%, !(''#!*#"
!(&)('#'*
the next iteration. B C
DG
F
EF
ÁÆĄÄÃÄÅǺÅBC ĄĂ ĀǺÃǼǼ Č Ĉ
DKF EG
The fully connected components that calculate C binary DDFH
(+*)+*
classifications need to address another issue where the B C
DG
Ĉ
Ĉ
Ĉ
of the dataset. When a dataset has unbalanced instances, B C
DFG
DGG
+%%, !(''#!*#"
!(&)('#'*
Weighted Cross-Entropy (WCE) is generally used. However, BACC
ÁÆĄÄÃÄÅǺÅBC ĄĂ ĀǺÃǼǼ
C EF
B DKG EG
WCE can only alleviate the issue to some degree. This Ĉ
(+*)+*
DFIJJ
paper introduces a new measurement, Binary-class Cross- DGIJJ
A
ĆÀ Č Â B C
Entropy (BCE), to alleviate the effect of unbalanced data in BACC DHIJ
KIJ
binary classification problems. For each binary classification
problem (the prediction is either c or ¬c), we obtain a set Figure 5: FCC layer of CNN-CF2 (the shaded areas denote
Sc with all the instances which either have c as the real the important features (variables and filters)
label or are predicted to belong to c. BCEc calculates the
cross-entropy using only the instances from Sc , as shown in
Eq. (1). Nc in Eq. (1) is the total number of instances in Sc We design a new CNN approach to incorporate class-
and Lc is the flattened vector L. The design of BCEc can specific variables in the model. This new approach is denoted
force the algorithm to learn better weights to separate class as CNN-CF2 (Fig. 5).
c from other classes which have similar features to class c. Building upon the CNN-CF model, it extracts important
1 X features from the LCP layer (L1 , · · · , LC ) and feeds only
BCEc = yc [i]log(pc [i]) + (1 − yc [i])(1 − log(pc [i]) the important features for class c to the corresponding c-
Nc
i∈Sc th fully-connected component for prediction. The features
(1) are chosen from the LCP layer for two reasons. First, it
where pc [i] = ω~ c × Lc . is the last convolutional/pooling layer, thus its features are
Example 3: For the dataset in Example 1, BCE for class at higher-level and have more valuable information than the
“SE” very likely considers the instances from “SE” and features in the previous layers. Second, the LCP hidden units
“ME”. The BCE for “SE” class focuses on separating “SE” absorb the time dimension through the previous layers, thus
and “ME” instances. the time dependency does not need to be considered when
We further define a Class-combined Cross-Entropy (CCE) choosing features in the LCP layer.
as the loss function in the training of the FCC layer with C The top features are related to two factors, the variables,
fully-connected components. CCE adds up the C BCE val- which are represented in the data, and the filters (or kernels),
ues from all the binary classification problems (Equation 2). which are set in the CNN-CF2 model. Selecting the top
C
X features involves identifying (i) the top important variables
CCE = BCEi (2) for each filter without considering the specific classes (using
c=1 Algorithm CNN-CF2 -TVI in Fig. 6) , (ii) the top filters for
The derivative of CCE is the sum of the derivatives of the C each class (using Algorithm CNN-CF2 -TFI in Fig. 7) , and
BCEs based on the sum rule of differentiation operation [10]. (iii). the important variables for each class (Section III-D)
The first step is to identify the important variables for
∂(CCE) ∂BCE1 ∂BCEC ∂BCEc
= + ... + = (3) each filter (kernel) because one filter in the CNN model can
∂ωc ∂ωc ∂ωc ∂ωc capture significant features for some variables, but not for
In the binary classification problem (c vs. ¬c), the weight all the variables. For each filter f and each instance n, the
and bias are trained using BCEc . The weight and bias for algorithm (Fig. 6) retrieves the variables with high feature
class c are only depended on whether c can be separated values because a high feature value indicates that the filter
over ¬c. Equation 3 shows the derivation calculation for ωc . matches the shape of the sequence for that variable.
Algorithm: CNN-CF2 -TVI (L, σV )
selection is conducted after CNN-CF2 -TFI. Please note that
Input: (1) L: N ×V ×F L array from the last convolutional layer, the significant variables selected at this step are class specific
(3) σV the percentage of requested variables for each filter
Output: L" : N ×V ×F L with the non-top-rated variables are zeros
while the important variables selected through CNN-CF2 -
1) Initialize an N ×V ×F L array L" with value zero TVI are for each filter, not class specific.
2) For each filter feature f (f =1 · · ·F L ) For each class c, we get nc and n¬c . They denote the
a) For each instance n (n=1 · · ·N )
indexes of the instances belonging to c and not belonging
i) Lfn = L[n, 1 · · · V, f ]
ii) topfn = the variable indies of the (σV × V ) highest values from to c respectively. For each class c and each variable v, we
Lfn extract the features for all the variables from the top filters
iii) For each variable v in topfn
L[nc , v, T Fc ] and L[n¬c , v, T Fc ], and calculate their mean
A) L" [n, v, f ] = L[n, v, f ]
3) Return L" vectors, meanLc and meanL¬c (whose length is |T Fc |) by
averaging on the first dimension. The score of a variable v is
Figure 6: The top-variable identification algorithm the euclidean distance between meanLc and meanL¬c . The
higher this score is, the higher the differentiation function
Algorithm: CNN-CF2 -TFI (L" , Y , σF , C) that v can play. The variables with the high score values can
Input: (1) L" : N ×V ×F L output array from CNN-CF2 -TVI,
(2) Y : the class-label vector for N instances, better differentiate class c from other classes.
(3) σF the percentage of requested features for each class,
(4) C the total number of classes Dataset N C V m
Output: T Fset ={T F1 , T F2 , · · · , T FC } where T Fc consists of #σF ·F L $ Action 560 20 570 100
top important filters for the class label c Activity 320 16 570 337
1) Initialize an C ×F L array ω with score zero Ara Voice 8800 88 39 91
2) For each class c (c = 1, · · · , C) Auslan 2565 95 22 96
a) nc = indexes of instances belonging to class c Daliy Sport 9120 19 45 125
b) n¬c = indexes of instances not belonging to class c Ges 396 5 18 214
c) For each filter f (f =1 · · ·F L ) Har 10299 6 9 128
i) L" fc = L" [ nc , 1 · · · V , f ] Ht Sensor 100 3 11 5396
ii) L" f¬c = L" [n¬c , 1 · · · V , f ] JapaneseVowels 640 9 12 26
iii) meanLc , meanL¬c = mean vector of L" fc and L" f¬c by averaging OHC 2858 20 30 173
on the first dimension respectively Net 1337 2 4 994
iv) ω[c, f ] = dist(meanLc , meanL¬c )) Eeg 128 2 13 117
3) For each class c Eeg2 1200 2 64 256
Ozone 346 2 72 291
a) T Fc = the top #σF ·F L $ highest values in ω[c, 1· · ·F L ]
4) Return T Fset ={T F1 , T F2 , · · · , T FC } Table III: Datasets
Figure 7: The top-filter identification algorithm
IV. E XPERIMENTS
After selecting the important variables for all the filters, All the methods are implemented using P ython 3.4,
the second step is to choose important filters because not and tested on a server with Intel(R) Xeon(R) CPU E5-
all the filters can generate significant features (Algorithm 2695 v2 @ 2.40GHz and 256 GB RAM. TensorFlow
CNN-CF2 -TFI in Fig. 7). The importance of a filter is (www.tensorflow.org [11]) is used to build the CNN model
evaluated using the top variables chosen for that filter by and Adamoptimizer is used in the training process.
the CNN-CF2 -TVI algorithm. CNN-CF2 -TFI first separates A. Methods for Comparison
the training data into two groups, where the instances in
one group belong to class c while the instances in the other The performance of the proposed approaches are com-
group do not belong to c (Lines 2a-2b). Then, it calculates pared with existing state-of-art methods: (i) Long Short Term
the distances of features from the c-class instances and the Memory Fully Convolutional Networks (LSTM-FCN) [6]
¬c-class instances (Line 2c). The filters with larger feature and (ii) an Attention LSTM-FCN (ALSTM-FCN) [6],
distances between the c-class instances and the ¬c-class which are defined for classifying univariate time series
instances are considered more important because they can and have been adapted for MTS, (iii) Multivariate LSTM-
better differentiate class c and ¬c (Line 3). Different from FCN (MLSTM-FCN) [7], and (iv) Multivariate Attention
CNN-CF, CNN-CF2 only feeds the important feature iden- LSTM-FCN (AMLSTM-FCN) [7]. Other than the previous
tified by CNN-CF2 -TFI to each fully connected component four approaches, more Other Baseline (OB), WMUSE [12],
in model training. ARKernel [13], LPS [14], mv-ARF [13], SMTS [15],
HULM [16], DTW [1], SVM [17], and RF [18] are used
D. Class-Specific Variable Identification in the comparison. Only the best results from the baseline
Besides improving the classification accuracy of MC approaches are presented in this section.
problems on MTS data, we are also interested in reporting B. Experimental settings
the features that are important to differentiate a specific class
c from the other classes. We call this step Class-specific (1) Datasets: We use 14 real datasets, which have at
variable identification (CVI). The class-specific feature (CF) least 100 instances, to test the proposed approaches. The
Methods
Dataset LSTM-FCN MLSTM-FCN ALSTM-FCN MALSTM-FCN Best-of-OB CNNmts CNN-CF CNN-CF2
Action 0.717 0.754 0.727 0.747 0.707 0.747 0.798 [1] 0.795 [2]
Activity 0.531 0.619 0.556 0.588 0.663 0.581 0.606 [4] 0.613 [3]
Ara Voice 0.980 0.980 0.986 0.983 0.946 0.961 0.972 [5] 0.973 [4]
Auslan 0.970 0.970 0.960 0.960 0.980 0.947 0.970 [3] 0.972 [2]
Daily Sport 0.997 0.997 0.997 0.997 0.984 0.993 0.995 [3] 0.996 [2]
Ges 0.505 0.535 0.525 0.531 0.409 0.535 0.545 [2] 0.556 [1]
Har 0.960 0.967 0.955 0.967 0.816 0.946 0.960 [3] 0.961 [2]
HT Sensor 0.680 0.780 0.720 0.800 0.720 0.760 0.860 [1] 0.860 [1]
JapaneseVowels 0.990 1.000 0.990 0.990 0.980 0.980 0.990 [2] 1.000 [1]
OHC 1.000 1.000 1.000 1.000 0.990 0.990 0.999 [2] 1.000 [1]
Averaged 0.833 [7] 0.860 [3] 0.842 [6] 0.856 [4] 0.794 [8] 0.843 [5] 0.870 [2] 0.872 [1]
* Values in [] denote the ranks in each row
* To better display the results, we only show the ranks of the results from CNN-CF and CNN-CF2 for each individual dataset.
Table IV: Multi-class classification (Best-of-OB: the best results from all the other baseline approaches)
Methods
Dataset LSTM-FCN MLSTM-FCN ALSTM-FCN MALSTM-FCN Best-of-OB CNNmts CNN-CF CNN-CF2
Eeg 0.609 0.656 0.641 0.641 0.625 0.578 0.584 [7] 0.615 [5]
Eeg2 0.907 0.910 0.907 0.913 0.775 0.941 0.972 [1] 0.967 [2]
Net 0.940 0.950 0.930 0.950 0.980 0.947 0.963 [3] 0.968 [2]
Ozone 0.676 0.815 0.792 0.798 0.751 0.791 0.815 [1] 0.815 [1]
Averaged 0.783 [7] 0.833 [3] 0.818 [5] 0.826 [4] 0.783 [7] 0.814 [6] 0.834 [2] 0.841 [1]
* Values in [] denote the ranks in each row
* To better display the results, we only show the ranks of the results from CNN-CF and CNN-CF2 for each individual dataset.
detailed statistics for the datasets are shown in Table III. (2) The results in the last two columns show that the CNN-
Evaluation measurements: The most commonly utilized CF2 achieves better or the same performance as CNN-CF
classification metric, accuracy, is reported as there are no im- on most datasets (nine out of ten datasets). Both CNN-CF
balanced datasets. (3) Parameter setting: The convolutional and CNN-CF2 have better performance than the CNNmts
layers of the CNN model contain three layers with filter sizes with the same configuration. The results are consistent with
8∗1, 5∗1, and 3∗1, the corresponding numbers of filters for our expectations and intuition.
the three layers are 128, 256, and 128. The pooling filter in 2) Comparison on Binary-Class Classification: The pro-
the global pooling layer is set to be the same as the length of posed approaches are designed for the multi-class classifi-
the time series output from the previous convolutional layer. cation problem (MC). However, they can also be applied to
These settings in the convolutional and pooling layers are conduct Binary classification (BC). We evaluate the perfor-
the same as those in [7]. Both the percentage for variables mance of the proposed approaches on four datasets with two
(σV ) and features (σF ) are set to be 50%. class labels. Table V shows the comparison results using all
methods. On those four datasets, the CNN-CF and CNN-CF2
C. Effectiveness Analysis still outperformed in terms of averaged accuracy although
In this section, we evaluate the performance of the pro- the improvement is less than that in MC classification.
posed approaches in two aspects: classification performance 3) Effect of the Loss Function: This section compares
and the meaningfulness of the selected features. the performance of using the newly designed loss function,
1) Comparison on Multi-Class Classification: This sec- class-combined cross-entropy (CCE), with the commonly
tion compares the classification accuracy on 10 multi- used weighted cross-entropy (WCE) for multi-class classi-
class MTS datasets of our proposed methods, CNN-CF and fications. Due to space limitation, we report results only
CNN-CF2 , with other state-of-the-art approaches. Table IV on three datasets, “Ht sensor” with the minimum number
presents the accuracy values on classifying those multi-class of classes (3 classes), “Auslan” with the maximum number
datasets. of classes (95 classes), and “Daily Sport” with the median
The ranks for the two proposed approaches and the aver- number of classes (19 classes).
age accuracy for all methods are provided. For all 10 MC
Datasets
datasets, CNN-CF2 achieves the highest average accuracy Loss function Ht Sensor Daily Sport Auslan
(the last row in Table IV) and the best ranking. CNN-CF2 WCE 0.800 0.991 0.948
gets the highest accuracy on four datasets and receives the CCE 0.860 0.995 0.970
[4] S. Fong, K. Lan, and R. Wong, “Classifying human voices [20] A. Quattoni, S. Wang, L. Morency, M. Collins, and T. Darrell,
by using hybrid sfx time-series preprocessing and ensemble “Hidden conditional random fields,” IEEE Transactions on
feature selection,” BioMed research international, vol. 2013, Pattern Analysis and Machine Intelligence, vol. 29, no. 10,
p. 720834, 10 2013. pp. 1848–1852, Oct 2007.
[7] F. Karim, S. Majumdar, H. Darabi, and S. Harford, “Multi- [24] D. S. Raychaudhuri, J. Grabocka, and L. Schmidt-Thieme,
variate lstm-fcns for time series classification,” Neural Net- “Channel masking for multivariate time series shapelets,”
works, vol. 116, pp. 237–245, 2019. CoRR, vol. abs/1711.00812, 2017.