You are on page 1of 10

This article has been accepted for publication in a future issue of this journal, but has not been

fully edited. Content may change prior to final publication. Citation information: DOI 10.1109/ACCESS.2019.2929872, IEEE Access

Dealing with Alarms in Optical Networks Using an


Intelligent System
DANSHI WANG1, LIQI LOU1, MIN ZHANG1, ANTHONY C. BOUCOUVALAS2, (Fellow IEEE), CHUNYU
ZHANG1, and XUETIAN HUANG3
1
State Key Laboratory of Information Photonics and Optical Communications, Beijing University of Posts and Telecommunications, Beijing 100876, China.
2
Department of Informatics and Telecommunications, University of Peloponnese, Terma Karaiskaki, Tripoli, Greece, 22100
3
China Telecom Corporation, Beijing, 100015, China
Corresponding author: Min Zhang (e-mail: mzhang@bupt.edu.cn).
This work was supported by NSFC Project No. 61705016 and the National Key R&D Program of China (2016YFB0901200).

ABSTRACT Millions of alarms in optical layer may appear in optical transport networks every month, which
brings great challenges to network operation, administration and maintenance. In this paper, we deal with this
problem and propose a method of alarm pre-processing and correlation analysis for this network. During the alarm
pre-processing, we use the method of combined Time series segmentation and Time sliding window to extract the
alarm transactions, and then we use the algorithm of combined K-means and Back propagation neural network to
evaluate the alarm importance quantitatively. During the alarm correlation analysis, we modify a classic rule mining
algorithm, i.e. Apriori algorithm, into a Weighted Apriori to find the high-frequency chain alarm sets among the
alarm transactions. Through the actual alarm data from the record in optical layer of a provincial backbone of China
Telecom, we conducted experiments and the results show that our method is able to perform effectively the alarm
compressing, alarm correlating and chain alarm mining. By parameter adjustment, the alarm compression rate is able
to vary from 60% to 90% and the average fidelity of chain alarm mining keeps around 84%. The results show our
approach and method is promising for trivial alarm identifying, chain alarm mining and root fault locating in
existing optical networks.

INDEX TERMS Alarm pre-processing, K-means, back propagation neural network, alarm compression, alarm
correlation analyzing, optical network

I. INTRODUCTION predicting chain alarms, and locating faults. Traditional alarm


As the scale of optical transport network (OTN) expands, the correlation analysis can find the alarms associations to a certain
number of alarms in optical layer may reach over one million extent and depend mainly on the expert systems or even manual
within only one week, which brings great trouble to the network operation of the experienced staff. With the growing number of
operation, administration, and maintenance. For example, if optical links and OTN systems, it is increasingly difficult for
alarms appear due to an overtime service delay or an extra-high experts to keep up with the rapid changes in the network and
bit-error-rate (BER), the network administrator has to judge the then discover truly useful knowledge from the alarms.
fault location promptly and correctly in order to repair the fault Most of the existing studies are based on association rule
in time. However, the situation in actual network is very mining, correlation coefficient, or mutual information to
complicated. A fault is often related to a large number of alarms, analyze the alarm correlation [1–4]. Meanwhile, alarm
including many redundant alarms. Meanwhile, there exist lots association analysis based on association rule mining has been
of false alarms corresponding to no fault. Therefore, it is widely concerned due to its advantages of compressing alarm
difficult for people to find useful information promptly from volume and finding high-frequency chain alarms. Typical rule
such a large number of alarms. If these alarms can be mining methods are Apriori-like algorithms [5–7], which locate
compressed reasonably and automatically, the alarm analysis frequent alarm transactions by repeatedly scanning the database
will be much easier, which will help the network administrator and present them as rules. However, there are few studies or
to find the root cause of the faults and restore the OTN in time. reports on alarm analysis for optical networks and the existing
Therefore, the alarm compression has become an urgent and methods are not very suitable for optical networks. The alarms
also tough problem in optical networking, and the premise of in OTNs rise not only from the optical transport plane but also
alarm compression is alarm correlation analysis. As an the service plane and the control plane, so that the large amount
important part of the network fault management, the alarm of alarms from different planes often interact with each other.
correlation analysis is helpful for deleting redundant alarms, As a result, if the association rules are directly mined from the

This work is licensed under a Creative Commons Attribution 4.0 License. For more information, see https://creativecommons.org/licenses/by/4.0/.
This article has been accepted for publication in a future issue of this journal, but has not been fully edited. Content may change prior to final publication. Citation information: DOI 10.1109/ACCESS.2019.2929872, IEEE Access

original alarms, the performance of the algorithm will be Given its present huge scale in trans-provincial backbones and
degraded significantly. Moreover, the original alarms from metropolitans networks, there are a large number of original
actual optical networks always suffer from several problems alarms in the existing optical network. Moreover, the original
(i.e., information redundant, time asynchronous, and ambiguous alarm data are facing the problems of information redundant,
importance of alarm attributes, etc.), which are detrimental to time asynchronous, and the unclear IAAs. We use the TT
alarm analysis and compression. Therefore, alarm pre- method to divide all the original alarms into alarm sequences.
processing is required. Then the alarm synchronization, redundancy removal, and
The common methods of quantitative evaluation of the alarm alarm transaction extraction are performed for each alarm
importance rely on experienced network experts in determining sequence via a time sliding window. But there are many alarm
the relative importance of alarms [8–10]. Different network attributes and their IAAs are unknown, so that we use the KB
experts or researchers may derive different weights for the method to evaluate the IAAs quantitatively. Firstly, the alarm
alarms in the same network. However, the alarm number rises attributes are quantized, and then the K-means algorithm is used
to tens of thousands and also the network is becoming more and to classify the alarms. Then, the Back propagation neural
more dynamic. It becomes impossible to determine the relative network (BP-NN) algorithm is used to infer each IAA weight.
importance of all alarms accurately by manual operation or According to the network management logs in existing
experts alone. To address the problem of a huge number of provincial OTN, the original alarms usually contain the
alarms with various attributes and of uncertain importance in following attributes: equipment name, equipment type,
the dynamic optical network, an effective and objective method equipment address, network element type, alarm level, alarm
of alarm importance evaluation is necessary. In an optical name, alarm type, alarm time, subnet, and so on. Among these
network, the alarm usually contains many attributes and each attributes, the equipment name can be determined by the
attribute with a wide range of values makes the alarm equipment address. The network element type and its subnet are
importance ambiguous. Therefore, it is necessary to give unified within a certain network segment. Therefore, as shown
weights to these attributes and thus evaluate the alarm in Fig. 1, to start with, we extract the six attributes, the
importance quantitatively. Recently, machine learning is equipment address (Ea), the alarm name (An), the alarm level
playing an increasingly important role in optical communication (Al), the alarm type (At), the equipment type (Et) and the alarm
research, and has been applied in diverse areas such as time (t) to form a new alarm, which is marked as a 6-tuple {Ea,
predicting equipment failure in optical network [11], reducing An, Al, At, Et , t}.
nonlinear phase noise [12,13], compensating physical New Alarm Attributions
Alarm_num
impairments [14], monitoring optical performance [15,16], Equip_name
adaptive nonlinear decision at the receivers [17], adaptive Equip_type Equip_address (Ea)
demodulator[18], and traffic-aware bandwidth assignment [19]. Equip_address Alarm_name (An)
Original Alarm_level
In this paper, we propose the TKBW method for alarm pre- Alarm Alarm_name Alarm_level (Al)
Alarm_time
Extract
processing and alarm association rule mining. The time required Attributions Alarm_type (At)
Alarm_type
for different faults to trigger a series of chain alarms is different Interface_type Equip_type (Et)
and, therefore, we use the time sliding window (TT) to divide Subnet Alarm_time (t)
the original alarms into alarm sequences and thus extract the Alarm_parameter

alarm transactions. Meanwhile, the alarm synchronization and


FIGURE 1. Extraction of original alarm attributes.
redundancy removal are performed in each of alarm sequence.
To address the problem of a huge number of alarms with
various attributes and uncertain importance, in this paper we use 1) TT METHOD OF ALARM TRANSACTION EXTRACTION
KB method to give the weights to the Importance on Alarm In the existing OTN, the number of alarms for different faults is
Attributes (IAAs). After the alarm pre-processing by combined different and the time required to form an alarm sequence is
K-means and Back propagation neural network (TT-KB) also different. Therefore, during time-segmenting of a large
method, we modified the traditional Apriori algorithm as W- number of alarms, we have to ensure that a complete alarm
Apriori algorithm to find out the association rules between sequence is collected. Otherwise it will affect the subsequent
alarms, and facilitate the alarm compression and fault location. alarm correlation analysis. Here, we calculate the time series
The rest of the paper is organized as follows: In Section II, similarity to implement the time segmenting, which makes the
alarm similarity in the same time period maximum while the
we discuss the principles of the alarm analysis scheme,
alarm similarity in different time periods minimal.
including the TT-KB method for alarm pre-processing, the W-
According to a direct calculating method of the time series
Apriori method for alarm correlation, and their implementation. similarity [20], the time interval is used as the criterion for the
Section III shows the experimental results and discussion. similarity measure among alarms, x, y as given by Eq. (1),
Finally, we draw conclusions in Section IV.
dist ( x, y)   ( x  y) 2
(1)
II. PRINCIPLE OF ALARM ANALYSIS SCHEME Thus, the intra-segment similarity function SI(t) is defined as
The proposed TKBW method is mainly divided into two parts: the sum of the squares of the time intervals from each moment
alarm pre-processing and alarm correlation analysis. to the midpoint of the time period, as follows,

A. TT-KB BASED ALARM PRE-PROCESSING

This work is licensed under a Creative Commons Attribution 4.0 License. For more information, see https://creativecommons.org/licenses/by/4.0/.
This article has been accepted for publication in a future issue of this journal, but has not been fully edited. Content may change prior to final publication. Citation information: DOI 10.1109/ACCESS.2019.2929872, IEEE Access

k k {A, B, C}, {B, C, E}, {C, E, A, B}, {E, A, B}, {B, D}, and we
SI (t )   I (ti )  dist (t , ci ) (2) treat the total alarm transactions as an alarm transaction
i 1 i 1 tci
database TD1={{A,B,C}, {B,C,E}, {C,E, A, B}, {E, A, B}, {B,
where ci represents the midpoint of the time period ti . D}}.
TABLE I
The inter-segment similarity function SO(t) is defined as the Example of the alarm transaction database TD1
sum of the time intervals between the midpoint of each time Transaction Name Items
period, as follows T1 A(1), B(3), C(5)
SO (t )   dist(ci , c j ) (3)
T2 B(3), C(5), E(7)
1i  j  k
T3 C(5), E(7), A(8), B(9)
Then the sum of the squared error (SSE) is adopted as the
objective function and also the index to measure the division of T4 E(7), A(8), B(9)
the time window, as given by Eq. (4): T5 B(9), D(10)
k
SSE   | t , ci |
2 2) QUANTITATIVE EVALUATION OF ALARM IMPORTANCE
(4) To evaluate the alarm importance quantitatively, it is necessary
i 1 tci
to assign IAA weights (Fig.4).
where the optimal result is to obtain the minimal SSE. The main We apply the KB method where the alarm attributes are taken
flow of the algorithm is attached as Algorithm I in Appendix. as the input and the alarms with similar attributes are classified
After the time segmenting, we use the time sliding window into the same class by the K-means algorithm. Then both the
method [21] to perform time synchronization, redundancy alarm attributes and the classifications are taken as the input of
removal, and alarm transaction extraction for the alarms in each the BP-NN to train the connection weights between the neurons,
time period. Time synchronization means that the alarms so that the connection weights map the information of the alarm
appearing in the same time window are regarded as concurrent attributes, and thus obtain IAA weights.
alarms and are to be extracted into the same alarm transaction. According to the advice of experienced network
In addition, if an alarm appears several times in a short interval, administrators and by observation of the alarm sets and the fault
it is recorded only once in the same time window in order to sets in the actual network logs, we select three alarm attributes
eliminate redundant alarms. The alarm transaction refers to the that are most closely related to the faults as the initial sample
collection of alarms appearing in the same time period. input, i.e. A={Al, At, Et} in Fig.1. Each attribute in the
Regarding the time sliding window method, we give the collection has several values that indicate the relative
following definitions: importance. Here we use the K-means algorithm to
Definition 1. Time window. The given time window width is automatically classify similar alarm samples, and get the
V, and the time window is used to slide from the beginning of a comprehensive alarm classifications after fully considering
time period until the end of the time period. these three alarm attributes.
Definition 2. The window slide step h, which is the length of We rewrite Eq. (1) as Eq. (5), where xi and xj represent the
each movement of the time window. attribute values, and D represents the number of attributes.
Figure 2 shows an example that the total alarms {A, A, B, D
C,...,C, B} are divided into several alarm sequences (e.g.,
D1 , D2 ,..., Dk ) according to the time series similarity calculation.
dist ( xi , x j )   (x
d 1
i ,d  x j ,d ) 2 (5)

Then given the window width V of 5 and the sliding step size h All alarms are divided into K classes, denoted as C1 ,
of 2, the time sliding window method is used to perform alarm C2,......Ck. Then, we take A={Al, At, Et} as the input and the
transaction extraction for each alarm sequence. If for example
K alarm classes as the labels to train a stable BP-NN model
alarms B(23) and B(27) occur in the same window, the alarm B
owing to the strong self-learning and fitting ability of the neural
is recorded only once. The time when the alarm occurs is network. Thus, via the method in [22], the connection weights
recorded as 23s and 27s, which is expressed as B(23,27):1. of neurons are mapped to the IAA weights, as given by Eq. (6):
Similarly if alarms A(28) and A(28) are reported multiple times
at the same time, alarm A is recorded once. The alarm occurs FeatureX   Hidden Y
(6)
XY

for 28s, which is expressed as A(28):1. where X is the weight matrix between the input layer neurons
and the hidden layer neurons, and Y is the weight matrix
between the output layer neurons and the hidden layer neurons.
Finally, we give the score of the alarm importance as Eq. (7),
where N is the number of inputs.
1
W   Feature1  Al  Feature2  At  Feature3  Et  (7)
N
FIGURE 2. Example of alarm transactions extraction using the time
sliding window. B. W-APRIORI BASED ALARM CORRELATION ANALYSIS
The score of the alarm importance is helpful for setting a
Table I gives an example of the alarm transactions extracted threshold and thus discard those trivial alarms and false alarms.
by the time period alarm sequence. The alarm transaction refers
to the set of alarms collected in a given time window, such as

This work is licensed under a Creative Commons Attribution 4.0 License. For more information, see https://creativecommons.org/licenses/by/4.0/.
This article has been accepted for publication in a future issue of this journal, but has not been fully edited. Content may change prior to final publication. Citation information: DOI 10.1109/ACCESS.2019.2929872, IEEE Access

For the remaining alarms, correlation analysis is still needed to defined as Eq. (8), where n is the number of items in X.
obtain chain-alarms. Meanwhile, minwSup is used to represent the minimum
The existing correlation analysis method is not designed for weighted support threshold, which is used to evaluate the
optical networks and they do not consider the difference in minimum limit of the transaction frequency.
alarm importance. For example, the typical Apriori algorithm
wSupport ( X)  Support ( X)  ( I X, j 1W j )
n
(8)
treats different alarms as equal, and there is no difference in the j

alarm attributes. However, in an actual optical network, an Definition 4. Weighted Confidence. The traditional
alarm usually consists of many attributes, and different attribute confidence of a rule X⇒Y is Confidence (X⇒Y)=
combinations indicates different alarm severities. Support(X∪Y)/Support(X). Then the weighted confidence
When mining the alarm correlation, we improve the Apriori (wConfidence) of a rule X⇒Y is defined as Eq. (9), where n is
algorithm as W-Apriori, in which the score of alarm importance the number of items containing the union of X and Y, and m is
is used as the weight. the number of items in X. Meanwhile, minwConf is used to
Given a database with a set of alarm transactions D={T1, represent the minimum weighted confidence threshold, which is
T2,...,Tm}, where Tj(j=1,2,...,m) is the set of alarms collected used to evaluate the minimum limit of the transaction
within time window. I={i1, i2,..., in} is the set of all alarms in the association.
database, and each alarm transaction set Tj is a subset of I (i.e. Support ( X  Y)  ( I ( XY ), j 1W j )
n

Tj⊂I). The set of alarm scores in the corresponding alarm set I (9)
wConfidence( X  Y)  j

Support ( X)  ( I X,k 1Wk )


m
can be expressed as W={W1, W2,..., Wn} (0 ≤ Wj ≤ 1, j = 1,2, ..., j

n), and Wj is the score of the alarm ij. Definition 5. Weighted Frequent itemsets. Given a database
In traditional Apriori, the association rule is the expression of D and a weighted support threshold minwSup, if a pattern X
the form X⇒Y, where X and Y are disjoint item sets (i.e. satisfies: wSupport(X) ≥ minwSup. Then X is the weighted
X∩Y=Ø). The strength of the association rule is measured by its frequent itemsets.
support and confidence. The support determination rules can be Therefore, the purpose of the W-Apriori algorithm is to find
used for the frequency of the given data set, and the confidence all the association rules where the wSupport and the
determines how frequently Y appears in transactions that wConfidence satisfy the conditions wSupport ≥ minwSup and
contain X. The W-Apriori algorithm redefines the above two. wConfidence ≥ minwConf in the given alarm database D
Definition 3. Weighted Support. The traditional support of respectively. The description of the W-Apriori algorithm is
pattern X(X⊆I) is denoted as Support(X)=Count(X)/|D|, where given as Algorithm II (seen in Appendix).
Count(X)=|Tj|X  Tj,Tj  D| is the number of transactions for
which item set X appears in D, and |D| is the total number of
transactions in D. Then the weighted support (wSupport) of X is

Tid Items Itemset support Itemset support


T1 A(1),B(3),C(5) Scan {A} 60%
Generate Connect
{A} 60%
T2 B(3),C(5),E(7) {B} 100% {B} 100%
T3 C(5),E(7),A(8),B(9) {C} 60% {C} 60%
T4 E(7),A(8),B(9) {D} 20% {E} 60%
T5 B(9),D(10) {E} 60%

Connect Generate Itemset support Scan Itemset


Scan Itemset Itemset support {A,B} 60% {A,B}
{A,B,C} {A,B} 60% {A,C} 40% {A,C}
{A,B,E} {B,C} 60% {A,E} 40% {A,E}
{B,C,E} {B,E} 60% {B,C} 60% {B,C}
{B,E} 60% {B,E}
Itemset support Generate {C,E} 40% {C,E}
{A,B,C} 40%
{A,B,E} 40% Itemset Symbol Description

{B,C,E} 40% support > (minwSup=50%) Candidate itemset


Frequent n-itemset
FIGURE 3. Execution process of W-Apriori algorithm.

Figure 3 is an example to illustrate the implementation of the we filter out high-frequency chain-alarm {A, B}, {B, C}, and
W-Apriori algorithm. We use the alarm transaction database D1 {B, E}.
given in Table I, which contains 5 alarm transaction sets: {A, B,
C. IMPLEMENTATION OF TKBW ALGORITHM
C}, {B, C, E}, {C, E, A, B}, {E, A, B}, {B, D}, where
minwSup = 50% is used as the minimum support threshold. The The application scenario of the alarm analysis scheme is shown
W-Apriori algorithm uses an iterative strategy of layer-by-layer in Fig. 4. In the OTN network, the alarm data in optical layer
search: in the k-th cycle, frequent k-itemsets are generated collected through the network management system or the
through a combination of the transaction database and candidate controller in software defined optical network (SDON), and
k-itemsets, and then a new candidate (k+1)-itemsets are then the TKBW method is used for alarm analysis.
generated based on the k-itemsets. And so on, the algorithm
stops until the maximum itemset of a cycle is empty. Finally,

This work is licensed under a Creative Commons Attribution 4.0 License. For more information, see https://creativecommons.org/licenses/by/4.0/.
This article has been accepted for publication in a future issue of this journal, but has not been fully edited. Content may change prior to final publication. Citation information: DOI 10.1109/ACCESS.2019.2929872, IEEE
Access

FIGURE 4. Application scenario of the alarm analysis scheme.

Alarm Preprocessing Module


TT Method Alarm Correlation
Analysis Module
Alarm
Time Synchronization Processing
Transaction
Redundancy Processing W-Apriori
Alarm Transaction Extraction Database
Method
Collections of
Alarm High-Frequency
Association Chain-Alarms
Original Rules Mining
Alarm KB Method
Alarm Alarm
Database Quantification
Attributes Weights
Scoring

FIGURE 5. Schematic diagram of TKBW method.

Figure 5 illustrate the overall principle of the TKBW Step 3. Scoring the alarm importance by using KB
method. TKBW consists of two modules, i.e. an alarm pre- method;
processing module and an alarm correlation analysis Step 4. Analyzing the alarm correlation with W-Apriori
module. The alarm pre-processing module extracts the algorithm and find out the chain alarms.
alarm transaction and evaluates the alarm weight and
converts the alarm data into alarm transactions suitable for III. EXPERIMENTS AND DISCUSSIONS
correlation analysis. The alarm correlation analysis module Experiments were conducted with the alarm data from the
is used to mine the association rules from the alarm network log of optical layer equipment in a provincial
transactions, and thus finds out the chain alarms. The backbone of China Telecom that contains 441 OTN nodes.
procedure of the TKBW method is shown as follows. We collected 5,100,000 original alarms within 30 days. The
Step 1. Collecting the alarms and select the useful alarm proposed TKBW method was developed via Python and
attributes; implemented on a computer with Windows 7 operating
Step 2. Extracting the alarm transactions by using TT system, Intel(R) Core (TM) processor i5-4345 with 2G
method; main memory.

This work is licensed under a Creative Commons Attribution 4.0 License. For more information, see https://creativecommons.org/licenses/by/4.0/.
This article has been accepted for publication in a future issue of this journal, but has not been fully edited. Content may change prior to final publication. Citation information: DOI 10.1109/ACCESS.2019.2929872, IEEE
Access

A. ALARM TRANSACTION EXTRACTION


First, the original alarm data of the first 20 days were used
as the test data, and Step 1 and Step 2 in TKBW procedure
2.0
were performed to extract the alarm transaction. In the
actual OTN, the operator generally uses the alarm data 1.9

Mean Square Error


extracted every 15 minutes, which is not extracted in real 1.8 K-means
time. Therefore, we divided all the alarms per day by 1.7
intervals of 15 minutes and thus got averagely 90 alarm 1.6
transaction databases per day, namely D1, D2, ... and D90. 1.5
Then, a time sliding window processing was performed for
1.4
each alarm transaction database to extract the alarm
transactions. Here we set the window width V as 5 minutes 1.3
and the sliding step has 2 minutes. 1.2
We compared the time consumption of alarm transaction 1.1
extraction by the TT method and the uniform time (UT) 1.0
window method. It can be seen from Fig. 6 that the more 0 2 4 6 8 10 12 14 16 18 20
complex TT method takes slightly longer time than the UT Iteration Epochs
method does. And as the alarm transaction number FIGURE 7. Mean square error varies with the iteration epoch of K-
increases, the difference fades away. We still prefer to use means.
the TT method because it fully considers the unevenness of
the distribution of alarm data, and also can remove the After the K-means classification, all the alarms were
isolated points and abnormal points as invalid alarms. automatically classified into four classes, recorded as C1,
C2, C3 and C4, respectively. Then, this classification was
used as the input label of BP-NN, where the connection
50 weights of all BP-NN neurons were obtained through BP-
45 NN feedback learning. Here, we considered the effect and
Time Consumption (s)

40 complexity of the algorithms, and set numbers of the input,


output and hidden layers of BP-NN as 3, 4 and 10
35
respectively, and set the learning rate as 0.01. Fig. 8 shows
30 the connection weight between the input neurons and the
25 first hidden layer neurons during BP-NN training, which
20 changes with the number of training iterations. It can be
15 seen that after 10 epochs of iteration, the connection weight
UT method tends to be convergent, which shows that it is feasible to
10 TT method score the alarms and their attributes.
5
0
0 1 2 3 4 5 6 7 8 9 10 0.50
50
45
50
Alarm Transaction Number (103) 0.45
Time Consumption (s)

40
35
Time Consumption (s)

45
40

30
35

25
30

20
25

15
10
UT method
20
15

5
TT method 10
UT method
TT method

0
5

0 1 2 3 4 5 6 7 8 9 10
Alarm Transaction Number (104)
0
0 1 2 3 4 5 6 7 8 9 10
Alarm Transaction Number (104)

FIGURE 6. Comparison of the time consumptions in extracting


Connection Weight

0.40
alarm transactions by TT method and UT method.
0.35
0.30
B. QUANTITATIVE ANALYSIS OF ALARM IMPORTANCE
In step 3, according to the advice of experienced network 0.25
administrators, we selected three alarm attributes that are 0.20
the most closely related to the faults, namely, alarm level Al, 0.15 w1,1
alarm type At, and alarm equipment type Et. These three 0.10 w2,1
attributes were initialized. For example, for alarm level AL,
we initialized emergency as 1, importance as 2, secondary
0.05 w3,1
as 3 and prompt as 4. Then, the alarm transactions with 0.00
0 5 10 15 20 25 30 35
initialized attributes were taken as the input of K-means
Iteration Epochs
algorithm, where the alarms with similar attributes were
divided into the same class. As shown in Fig. 7, after about FIGURE 8. Connection weights vary with the iteration epoch of BP-NN.
15 iterations, the mean squared error of classification
converged to 1.5, indicating that the K-means algorithm is In the above, we marked the output classifications (C1 to
trained. C4) by K-means as four different values, but these values
are in fact only tags of the classes and they are not
necessarily the only representations. Therefore, there are 24

This work is licensed under a Creative Commons Attribution 4.0 License. For more information, see https://creativecommons.org/licenses/by/4.0/.
This article has been accepted for publication in a future issue of this journal, but has not been fully edited. Content may change prior to final publication. Citation information: DOI 10.1109/ACCESS.2019.2929872, IEEE
Access

kinds of combinations when assigning 1~4 to C1~C4


respectively. We repeated the above experiment with
300
different combinations of the initial values of C1~C4 and
then calculated the IAA weights given by Eq. (6). The

Time Consumption (s)


250
results are shown in Fig. 9. It can be seen that, although
with different initial values, via the classification of K- W-Apriori
200
means and BP-NN, the IAA weights are able to be Apriori
stabilized at around 0.55, 0.25 and 0.20 respectively. It 150
indicates that the combined algorithm of K-means and BP-
NN works independent of the initially quantized values. 100
Therefore, the idea of giving quantitative evaluation to the
alarms and the abstract alarm attributes works. Then, the 50
scores of alarms can be calculated according to Eq. (7), and
the wSupport and wConfidence of alarms can be calculated 0
according to Eq. (8) and Eq. (9). Thereafter, as illustrated 0.0 0.1 0.2 0.3 0.4 0.5 0.6 0.7 0.8 0.9 1.0
by the C1 step in Fig.3, we obtained the wSupport for each Minimal wSupport
alarm and then deleted those alarms whose wSupports are FIGURE 10. Time consumption varies with minwSups.
less than the minwSup. Here, the minwSup acts as a
threshold for the preliminary alarm compression and this Next, we tested the effects of alarm number on the
threshold should be assigned according to the requirements performance of both algorithms. Fig. 11 shows the time
of the actual network management. consumptions for mining 10,000 to 100,000 alarm
transactions by both algorithms, with a fixed minwSup of
0.2. As the alarm number increases, the time consumption
0.60 of both algorithms become longer, but the W-Apriori
0.55
outperforms the Apriori gradually. With larger alarm
number, the advantages of W-Apriori becomes more
0.50 obvious.
0.45 Both Fig. 10 and Fig. 11 show that the modified W-
Alarm level
IAA Weight

0.40 Apriori outperforms W-Apriori in terms of time


Alarm type consumption, since the W-Apriori considers the alarm
0.35
Equipment type importance and thus removes a large number of redundant
0.30
alarms (e.g., false alarms).
0.25
0.20
140
0.15
Time Consumption (s)

0.10 120
0 5 10 15 20 25
Combination of original quantized values 100
FIGURE 9. IAA weights for different quantitative combinations. 80

60
C. ALARM CORRELATION ANALYSIS AND ALARM
COMPRESSION 40
After the preliminary alarm compression, we analyzed the W-Apriori
20 Apriori
correlation among the remaining alarms and explored ways
to compress the alarms in terms of alarm compression rate 0
and fidelity. 0 1 2 3 4 5 6 7 8 9 10
First, the time consumptions of the traditional Apriori Alarm Number (103)
and the modified W-Apriori were compared, as shown in FIGURE 11. Time consumption varies with alarm number.
Fig. 10, where the number of alarm transactions sets is
100,000 and minwSup changes from 0.1 to 0.6. As the Figure 12 presents the comparison of the compressed
minwSup increases, the time consumptions of both alarm collection number by both algorithms, with various
algorithms decrease gradually. The main reason is that minwSups and with a fixed alarm transaction number of
when the threshold minwSup is small, more alarm 10,000. To some extent, the same minwSup means that the
transaction sets need to be processed. In addition, the W- same compression fidelity can be obtained. Fig. 12
Apriori takes less time than the Apriori does. indicated with the same compression fidelity, W-Apriori
works more effectively and leads to fewer alarm collections.

This work is licensed under a Creative Commons Attribution 4.0 License. For more information, see https://creativecommons.org/licenses/by/4.0/.
This article has been accepted for publication in a future issue of this journal, but has not been fully edited. Content may change prior to final publication. Citation information: DOI 10.1109/ACCESS.2019.2929872, IEEE
Access

over the number of total original alarm collections. It can


be seen that as minwSup increases from 0.1 to 0.2 and then
Compressed Alarm Collections

1400 to 0.3, the alarm compression rates rises from 67% to 82%
1200 and then to 89%, respectively. That is, larger minwSup
yields higher compression rate. In other words, with larger
1000 minwSup, more alarms will be removed as false alarms.
Moreover, given a fixed minwSup, the compression rate
800 remains almost stable, independent of both the training
600
samples in the first 20 days and the application samples in
W-Apriori the last 10 days, which indicates that the W-Apriori works
400 Apriori stable and has little deviation between the training samples
and the application samples.
200

0 100%

Average Compression Rate


0.1 0.2 0.3 0.4 0.5 0.6 0.7 0.8 0.9 1.0
Minimal wSupport 95%
FIGURE 12. Compressed alarm collections by both algorithm with 90%
various minwSups.
85%
We applied the TKBW trained by the alarms in the first
20 days and then analyzed the alarms in the last 10 days in 80%
Minimal wSupport = 0.1
order to evaluate the alarm compression performance. Here, 75% Minimal wSupport = 0.2
the average number of alarm transactions generated per day Minimal wSupport = 0.3
is approximately 5,000. Fig. 13 shows the variation of 70%
compressed alarm collections with various minwSup, in 65%
which, as minwSup increases from 0.1 to 0.3, the alarm
collections are obviously compressed. In particular, when 60%
0 5 10 15 20 25 30
minwSup is 0.2, the number of compressed alarm
collections is about only half of that when minwSup is 0.1. Time (day)
This indicates that different minwSup means different FIGURE 14. Compression rate with different minwSups.
compression rates. Therefore, the bigger the minwSup is,
the greater the compression rate is, and the fewer alarm The results above also indicate that, by adjusting the
collections are finally obtained. minwSup, we are able to obtain different compression rates
for different purposes. Higher alarm compression rate is not
necessarily desired, for that we must consider the
Compressed Alarm Collections

1800 compression fidelity, which is defined as follows.


1600 After the alarm compression, the collections of high-
frequency chain alarms were obtained. We assume that the
1400
high-frequency chain alarms are the substantial key alarms
1200 that can represent the actual faults, and the location of the
1000 fault nodes are known. Then, we define compression
fidelity as the proportion of the compressed high-frequency
800 chain alarms occurrence location containing the actual fault
600 node location in each time period.
400 We measured the quality of the alarm compression in
Minimal wSupport = 0.1 terms of compression fidelity. Fig. 15 shows the variation
200 Minimal wSupport = 0.2
of compression fidelity with different minwsups when
Minimal wSupport = 0.3
0 minwConf is 0.1. Here, we set minwConf to a smaller value,
1 2 3 4 5 6 7 8 9 10 mainly in order to retain a large number of original alarm
Time (day) collections and facilitate the statistical experiment results. It
FIGURE 13. Number of compressed alarm collections v.s. minwSup. can be seen that as the minwSup is 0.1 or 0.2, the fidelity
keeps relatively high, namely 84% and 80% respectively,
indicating that after alarm compression with such a
In order to show the effects of minwSup on the alarm minwSup, the high-frequency chain alarms can reflect the
compression more directly, we plotted the alarm actual fault situation to a large extent. However, as the
compression rate with different minwSup for total 30 days, minwSup is 0.3, too much alarms are removed during
as shown in Fig. 14. The alarm compression rate here refers compression and thus the fidelity remains even lower than
to the ratio of the number of compressed alarm collections

This work is licensed under a Creative Commons Attribution 4.0 License. For more information, see https://creativecommons.org/licenses/by/4.0/.
This article has been accepted for publication in a future issue of this journal, but has not been fully edited. Content may change prior to final publication. Citation information: DOI 10.1109/ACCESS.2019.2929872, IEEE
Access

60%, which means in this case the alarm compression is not require further testing and analysis. Whatever the outcome,
very believable. the proposed scheme offers a good promising model, which
Moreover, from Figs. 14 and 15, it can be seen that when is interesting and will needs further future research.
minwSup is 0.2, the obtained alarm compression rate is
15% larger than when minwSup is 0.1, which indicates that IV. CONCLUSION
when the minwSup is 0.2, the original alarms can be To deal with the problem of coping with the huge number
compressed to a larger extent while relatively high of Alarms in OTNs, we have proposed the TKBW method,
compression fidelity is kept. Therefore, considering both which offers alarm compression and high-frequency chain
the alarm compression rate and fidelity, it may be alarms mining. By taking actual data of network logs from
appropriate to select the minwSup threshold as 0.2. China Telecom, we have conducted experiments, the results
However, in actual optical network, the minwSup threshold of which show that the TKBW method is able to give a
should be set to a particular optimal range of values score to indicate the importance of each alarm and the score
according to the network scale. If the threshold is set too is independent of the initial quantization values. In addition,
small, many redundant alarms (e.g., false alarms) will be the TKBW method is able to mine the chain alarms and
retained, which is not conducive to the subsequent alarm compress the alarms to a rate on demand by adjusting
correlation analysis and alarm compression; if the threshold parameters, such as minwSup. The results are promising
is set too large, some important alarms (e.g., substantial and helpful for false alarm identifying and root failure
alarms) will be removed. Therefore, in practical operation, locating.
it is necessary to set a threshold according to actual OAM
requirements and operational effects. APPENDIX

ALGORITHM I
100% Time Division Method based on Similarity of Time Series
Input: k: the number of time segments
90% n: database containing n alarm events
Output: k: k time intervals with minimum sum of squared errors
1: Select initial cluster center points for k time segments;
80% 2: for (j = 1; j < = n; j++) {
Fidelity

3: Assign each tj to the time interval which has the closest


mean; }
70% Minimal wSupport = 0.1
Minimal wSupport = 0.2 4: for (i = 1; i < = k; i++) {
Minimal wSupport = 0.3 5: find the cluster center point ci for each time segment; }
60% k

 | t , c
2
6: compute SSE  i | ;
i 1 tci
50%
7: Repeat until SSE conversed.
0 1 2 3 4 5 6 7 8 9 10
Time (day) ALGORITHM II
FIGURE 15. Compression fidelity with different minwSups. W-Apriori Algorithm
Ck: Candidate itemsets of size k
According to the experimental results, the TKBW
Lk: frequent itemsets of size k
method can effectively perform alarm compression, alarm Input: D: transaction database
correlation analysis and chain alarm mining to implement minsup: minimum support threshold
the compression of alarms and obtain high-frequency chain Output: L: frequent itemsets of D
alarms. This means that we can use high frequency chain 1: L1= FindFrequent_1_itemsets(D);
2: //generate the frequent 1-itemsets
alarms to find the root fault nodes location. Moreover, the
3: for (k = 2; Lk-1 ≠ ; k++)
method is not complicated in implementation. Therefore, 4: Ck = Genetate Candidates (Lk-1,minsup);
the proposed method is easy to set up in the controller. 5: //generate the new candidate itemsets
However, since the experimental data have come from the 6: for each transaction t  D
actual alarm monitoring record in optical layer of a 7: Ct= subset(Ck,t);
8: //find out all candidate k-itemsets contained in
provincial backbone of China Telecom, the number of alarm
transaction t
attributes of the fiber layer equipment is small and the 9: end for
attributes are relatively thick. Our method cannot fully 10: Lk ={ c  Ck| support ≥ minwsup}
guarantee that all OTNs are applicable. If there are more data 11: end for
from different OTNs, a more complete analysis method can 12: return L=∪Lk;
be proposed and the universality of the method can be
verified. In addition, we extracted six important and valid ACKNOWLEDGMENT
attributes in the experiment. If we change the number of We would like to acknowledge the China Telecom
alarm attributes (e.g., reduce the number of alarm attributes) Corporation for providing the network data, and the
and re-experiment, weather we get the same results will

This work is licensed under a Creative Commons Attribution 4.0 License. For more information, see https://creativecommons.org/licenses/by/4.0/.
This article has been accepted for publication in a future issue of this journal, but has not been fully edited. Content may change prior to final publication. Citation information: DOI 10.1109/ACCESS.2019.2929872, IEEE
Access

support from the Fundamental Research Funds for the [20] Keogh E, Chu S, Hart D, Pazzani M (2003) Segmenting time series:
Central Universities (2019RC12). A survey and novel approach. Work 57: 1–21.
[21] M. Hauptmann, J. H. Lubin, P. Rosenberg, et al. The use of sliding
time windows for the exploratory analysis of temporal effects of
REFERENCES smoking histories on lung cancer risk. Statistics in Medicine, 2000,
[1] N. Amani, M. Fathi, M. Dehghan, A case-based reasoning method 19(1): 2185-2194
for alarm filtering and correlation in telecommunication networks, [22] J.D.Olden, M.K.Joy, R.G., An accurate comparison of methods for
in: Proceedings of the Canadian Conference on Electrical and quantifying variable importance in artificial neural networks using
Computer Engineering, 2005, pp. 2182–2186. simulated data Eco Model,178 (3) (2004), pp.389-397
[2] M. Klemettinen, H. Mannila, H. Toivonen, Rule discovery in
telecommunication alarm data, Network and Systems Management
Journal 7 (4) (1999) 395–423.
[3] J. Wu , and X. Li, "An Effective Mining Algorithm for Weighted
Association Rules in Communication Networks." Journal of
Computers 3.10(2008).
[4] Yan, L., Li, C.: Incorporating pageview weight into an association-
rule-based web recommendation system. In: Australian Conference
on Artificial Intelligence, pp.577–586 (2006)
[5] R. Agrawal, R. Srikant, Fast algorithm for mining association rules,
in: Proceedings of the 20th VLDB Conference, 1994, pp. 487–499.
[6] G. Grahne, J. Zhu, High performance mining of maximal frequent
itemsets, in: Proceedings of the Sixth SIAM International Workshop
on High Performance Data Mining, 2003, pp. 135–143.
[7] T. Wang, P.L. He, Database encoding and an anti-apriori algorithm
for association rules mining, in: Proceedings of the Fifth
International Conference on Machine Learning and Cybernetics,
2006, pp. 1195–1198.
[8] J. Han, J. Pei, Y. Yin, R. Mao, Mining frequent patterns without
candidate generation: a frequent-pattern tree approach, Data Mining
and Knowledge Discovery 8 (1) (2004) 3–87.
[9] J.W. Han, Y.J. Fu, Discovery of multiple-level association rules
from large databases, in: Proceedings of the 1995 International
Conference on Very Large Data Bases, 2002, pp. 420–431.
[10] C.H. Cai, Ada W.C. Fu, C.H. Cheng, W.W. Kwong, Mining
association rules with weighted items, in: Proceedings of IEEE
International Database Engineering and Applications Symposium,
1998, pp. 68–77.
[11] Z. Wang, M. Zhang, D. Wang, C. Song, M. Liu, J. Li, L. Lou, and Z.
Liu, “Failure prediction using machine learning and time series in
optical network,” Opt. Express 25, 18553–18565 (2017).
[12] D. Wang, M. Zhang, Z. Li, Y. Cui, J. Liu, Y. Yang, and H. Wang,
“Nonlinear decision boundary created by a machine learning-based
classifier to mitigate nonlinear phase noise,” in Proceedings of
Optical Communication (ECOC 2015), pp. 1–3.
[13] D. Wang, M. Zhang, Z. Li, C. Song, M. Fu, J. Li, and X. Chen,
“System impairment compensation in coherent optical
communications by using a bio-inspired detector based on artificial
neural network and genetic algorithm,” Opt. Commun. 399, 1–12
(2017).
[14] D. Wang, M. Zhang, M. Fu, Z. Cai, Z. Li, H. Han, Y. Cui, and B.
Luo, “Nonlinearity Mitigation Using a Machine Learning Detector
Based on k-Nearest Neighbors,” IEEE Photonics Technol. Lett.
28(19), 2102–2105 (2016).
[15] D. Wang, M. Zhang,J. Li, Y. Xin, J. Li, M. Wang and X. Chen,
“Intelligent Optical Spectrum Analyzer Using Support Vector
Machine,” IEEE Photonics Society Summer Topical Meeting Series
(SUM), 239–240 (2018).
[16] D. Wang, M. Zhang, J. Li, Z. Li, J. Li, C. Song, and X. Chen,
“Intelligent constellation diagram analyzer using convolutional
neural network-based deep learning,” Opt. Express 25(15), 17150–
17166 (2017).
[17] Y. Yuan, M. Zhang, P. Luo, Z. Ghassemlooy, D. Wang, X. Tang,
and D. Han, “SVM detection for superposed pulse amplitude
modulation in visible light communications,” in Proceedings of
Communication Systems, Networks and Digital Signal Processing
(CSNDSP 2016), pp. 1–5.
[18] J. Li, M. Zhang, and D. Wang, “Adaptive demodulator using
machine learning for orbital angular momentum shift keying,” IEEE
Photon. Technol. Lett. 25, 1455–1458 (2017).
[19] C. Song, M. Zhang, X. Huang, Y. Zhan, D. Wang, M. Liu and Y.
Rong, “Machine Learning Enabling Traffic-Aware Dynamic Slicing
for 5G Optical Transport Networks,” Conference on Lasers and
Electro-Optics (CLEO), 1 – 2 (2018)

This work is licensed under a Creative Commons Attribution 4.0 License. For more information, see https://creativecommons.org/licenses/by/4.0/.

You might also like