You are on page 1of 8

Computers in Industry 75 (2016) 27–34

Contents lists available at ScienceDirect

Computers in Industry
journal homepage: www.elsevier.com/locate/compind

Multi-sensor data fusion by a hybrid methodology – A comparative


study
Sarvesh Rawat a,*, Surabhi Rawat b
a
School of Electronics and Electrical Engineering (SELECT), VIT University, Vellore 632014, India
b
School of Biotechnology & Biosciences, Lovely Faculty of Technology & Sciences, Punjab, India

A R T I C L E I N F O A B S T R A C T

Article history: Multi-sensor data fusion is considered as an inherent problem in wireless sensor network applications. It
Received 4 May 2015 is widely assumed as a sturdy non linear system in view of the complexities involved in its operation. An
Received in revised form 14 October 2015 accurate and precise methodical solution is therefore a complicated task to accomplish. It is crucial for
Accepted 29 October 2015
the sensory systems that they should not be influenced in terms of accuracy and precision by any means.
Available online 13 November 2015
To address these issues a hybrid model employing rough set (RS) with back-propagation neural network
(BPNN) is used to ameliorate the data fusion capability of the system with an illustrative example.
Keywords:
Experimental results have demonstrated an escalating improvement in the predictive accuracy of the
Multi-sensor data fusion
hybrid model as compared to the traditional BPNN model.
Hybrid architecture
Neural network ß 2015 Elsevier B.V. All rights reserved.
Diagnostic tool
Rough-sets

1. Introduction The greater part of the information related issues emerge from
the restrictions existing in the sensor technologies and unpro-
A multi-sensor system sustains to several abrupt anomalies like cessed raw data that can be categorized as follows.
electronic disturbances and thermal changes which changes the Data imperfection: It exists when data provided by sensors is
state of the system [1]. Therefore, a framework should be primed influenced by some degree of uncertainty in the measurement
that guarantees a stable and smooth function under the influence system or due to impreciseness present in the data. An
of unreliable external or unknown influences. Systems employing imperfection in the data can be caused due to the ambiguity
multi sensors for their functions gain more informative knowledge present in sensory data or due to false information value.
than the one using single sensor system [2]. Complex systems used Additionally, sometimes this imperfection also exists due to
in missile tracking, live targets tracking, employs various kinds of incomplete/missing information. A data fusion algorithm should
sensors for acquiring high accuracy, improved reliability and express such imperfections present in the data and viably reduce
robustness of the system [6]. Therefore, an inconceivable stress is them.
put upon investigating the calculations focusing on multi-sensory Inconsistent data and data correlation: There are several causes
data combinations [3]. These systems have become extremely vital for a dataset to be inconsistent. When a system is encountered
these days owing to their high accuracy and large parameter with outlier/spurious data or encountered by disordered or
strengths. There are several issues involved in a sensory network conflicting data that are primarily caused by inconsistencies and
that make data fusion a complicated task. The real world data ambiguities present in the environment, it fails to distinguish
fusion application tackles with several data related challenges that between the real data and the associated ambiguities with it. The
are shown in Fig. 1. The input data to a multisensory fusion system Sensory Data fusion algorithm should exploit the redundant data
is inconsistent, imperfect, correlated and/or in disparate forms. to alleviate such effects. Data correlation is often experienced in
Each of these can be subcategorized further into distinctive subsets distributed data fusion setting. For example, a WSN has several
and discussed in the following section. nodes that are exposed to the same external noise, hence biasing
the measurements. These data dependencies must be accounted
for the exact operation of the system, in order to refrain them
* Corresponding author.
suffering from over/under confidence in the effect.
E-mail addresses: sss.sarvesh888@gmail.com (S. Rawat), Disparateness: It alludes to the integration of the two or more
surabhirawat40@gmail.com (S. Rawat). disparate sources, that helps in overcoming some of the limitations

http://dx.doi.org/10.1016/j.compind.2015.10.012
0166-3615/ß 2015 Elsevier B.V. All rights reserved.
[(Fig._1)TD$IG]
28 S. Rawat, S. Rawat / Computers in Industry 75 (2016) 27–34

Fig. 1. Taxonomy of inaccuracies present in the multisensory data fusion system.

of the single source based system. While, on the other hand, a sources has become a challenging task [25]. In our example, a
multi-sensor system comprises of different data domains which multisensory system containing four sensors (sensor 1, sensor 2,
add another challenge for the data fusion [26]. For example, a sensor 3 and sensor 4) is used to monitor the Biochemical Oxygen
typical sensory system consists of temp sensor, pressure sensor Demand (BOD) in the water area. The other four sensors (sensor 5,
and moisture sensor. Likewise, in the same manner, some sensors sensor 6, sensor 7 and sensor 8) monitor the Chemical Oxygen
deal with the chemical and physical nature of the material. Each Demand (COD) of the water. The way of installation of the sensor
sensor has its own signature and has its own classification and net also affects the final fusion result that is represented as sensor
description. It becomes very hard for the system’s operator to 9. Another sensor denoted as sensor 10 that describes the status of
assimilate the data and mine effective rules from it [27]. the sewage disposal machine stating if it is running or not. Fig. 2
The traditional sensor fusion methodologies are highly impre- shows the input data aggregated from the sensors.
cise, uncertain and inaccurate since the unverifiable parameters The output value that determines the status of water pollution
and obscure attributes lead to a large error and complications level is shown in Fig. 3. The range of output values lies between
[5,6]. The outside disturbances and uncertainty factors are very 3 and 4.
complex and non linear in nature. To analyze the whole incoming The system operator analyzes all the incoming data collected
sensory data, identify the state of the system and effectively mine from different sensors to estimate the status of the pollution level
results from it become a difficult task for the system’s operator. in the water. The data samples consist of multiple domains of
Hence, there is a huge demand for developing new sensor systems values that make it a typical data fusion problem. There is an acute
which can partition the error gap. To redress these raised issues need of devising a precise and detailed mathematical model which
several hybrid methodologies are employed such as fuzzy logic, can model the whole system. To resolve this problem, a hybrid
artificial neural network, neuro-fuzzy and rough sets [11,12,16]. To method is proposed in this work that integrates data fusion and
identify the state of the system using the training data is a primary data mining together to produce effective rules for the system’s
reason of using machine learning and soft computing techniques. operator.
The aim of this work is to propose an effective and proficient
architecture capable of delivering a high predictive accuracy. To 3. Literature review
meet the objectives, a hybrid architecture employing RS–BPNN is
employed in this work to enhance the reliability and capability of There are several issues associated with data fusion methodol-
the multi sensor fusion system. The proposed methodology is ogies making it a challenging task for the system’s operator as well
tested in a water pollution monitoring system. It utilizes the data, as researchers. The majority of these issues emerged from the
taken from the monitoring system, to train and test the proposed unprocessed imperfect data that is highly uncertain and vague in
hybrid network. RS preprocess and mine the incoming data nature as discussed in previous sections. It is quite troublesome
receiving from the sensors. It identifies the crucial features and task for the system’s operator to process the data from disparate
removes the superfluous information from the database. The sources and mine useful results. Various researchers have done a
important features are fed to the NN classifier and trained with BP significant work to deal with the issues associated with the multi-
algorithm to predict the pollution level of the area. sensor data fusion system.
The rest of the paper is arranged as follows. Section 2 discusses Few researchers utilized a probabilistic approach with Bayesian
about the water pollution monitoring system. Section 3 discusses network in their work to deal with the data fusion problem
about the work done by the various researchers and the prevailed [19,21].This approach relies on the probabilistic density function to
problems of this field. Sections 4 and 5 consist of the basics of RST express the uncertainty present in the data. Although it is a
and NN, while Section 6 shows the experiment and the results. traditional approach to treat the data filled with uncertainty but it
Finally, the conclusion of the work along with discussion is is incapable of addressing the other imperfection aspects of the
presented in Section 7. data, such as dealing with the missing data samples or data
correlation [20]. Few other researchers used fuzzy reasoning
2. Water pollution monitoring system technique to process the vague data and generate the decision
rules for the system’s operator [11]. An integrated system
Water pollution is considered as one of the most critical employing fuzzy with data fusion was employed to minimize
problems of the modern world [25]. Rapid industrialization has led the failure risk in an integrated vehicle systems [4,7]. The major
to a surplus increment in the quantity of toxins and wastes in the limitation while using fuzzy based system lies in the fact that they
water. Now a day, monitoring the water pollution level in water are limited merely to the fusion of the vague data [15]. An
[(Fig._2)TD$IG] S. Rawat, S. Rawat / Computers in Industry 75 (2016) 27–34 29

sensor1 sensor 2 sensor 3 sensor 4 sensor 5


sensor 6 sensor 7 sensor 8 sensor 9 sensor 10
2.5

Sensor values
1.5

0.5

0
1 3 5 7 9 11 13 15 17 19 21 23 25 27 29 31 33 35 37 39 41 43 45 47 49
Total Samples

Fig. 2. Input features showing the status of different sensors.


[(Fig._3)TD$IG]
Output
6
4
2
0
1 3 5 7 9 11 13 15 17 19 21 23 25 27 29 31 33 35 37 39 41 43 45 47 49

Fig. 3. Output decisive attribute.

evidential approach to solve the problem with relying on the system identification and prediction. In this work, a hybrid model
probability measure to categorize the data was proposed in [21]. It is proposed that provides a more comprehensive treatment to the
enables the fusion of ambiguous and uncertain data, but does not inconsistent, uncertain and, imperfect data. An offline hybrid
deal with the other aspects of the data imprecision. Hence it classifier involving RS alongside BPNN is used to process the
renders ineffective for the fusion of highly conflicting data [28]. Set incoming data and identify the state of the system. It is a two step
approximation based techniques employing RS was also used in process. In the first step, RS is used as a set theoretic approach that
processing the multisensory data [12,13]. The major associated effectively deals with the impreciseness, uncertainty and ambi-
advantage with RS is it does not require any prior or additional guity present in the data. Using the notions of upper and lower
information about the database. Yet at the same time it obliges a approximations it finds the crucial features of the dataset. These
fitting level of the data granularity which makes it less productive features are fed to the NN that is trained using BP algorithm and the
[28]. Few researchers used sensor validation techniques and prediction value of the decision is made.
stochastic adaptive sensor modeling techniques that provide a
basic structure for the detection of the bogus data without having 4. Rough sets
prior knowledge [22]. However, this model had an exceptionally
constrained extension as it was confined to a couple of RST is used as a tool to discover the ambiguity from a sample
fundamental and particular prior known failure models [16,18]. data set and remove the redundant information [12,13]. It provides
Another significant work using correlated data fusion methods by a proficient algorithm to discover hidden patterns from the given
implying covariance intersection was proposed in [26]. Although it dataset. It is profoundly used in the reduction of ineffectual data
provides an enhanced efficiency, yet it offers tighter covariance from a large database. It assesses the significance present in the
estimation and has a limited scope in the real time world system. data set and finds a possible set of data that can be employed for
As the system is becoming more complex, the data fusion training the system [13,14]. A given information table can be
methodologies have stumbled upon various new challenges represented as I = (U, A [ D, V, f). Here U represents a non empty
[17]. The system’s operator has to deal with a lot of uncertainties finite set; A [ D denotes the total attributes where A is the total
vagueness and imperfections in the data. It turns into an entangled number of condition attributes and D is the decision attribute. It
errand for the operator to mine the effective rules and estimating should be noted that A \ D = f. Now, the function f is a denoted as
the state of the system. Furthermore, in some data fusion methods, f : U  a ! Va, in a way f(y, a) 2 Va. Here, y 2 U and a 2 A. Va 2 V is a
a system operator must have some prior information, such as domain of the attribute a 2 A. In common the entire information
statistical characteristic or probabilistic measures about the system is represented as I = (U, A).
sensing object so that an efficient decision can be made [23,24]. The basic concept of RS includes the following terms.
Current sensor fusion technologies and methods are inefficient
rendering a large error gap persists in the system. Consequently, to 4.1. Indiscernibility
avoid the complicated systematic approach, several delicate and
more compelling techniques utilizing soft computing are often A binary relation P  X  X, for an object x is called as an
used today. The foremost idea behind using a soft computing equivalence relation if it is reflexive, follows xRx, symmetric (xPy
system is to assimilate the training data and utilize it for the then yPx) and transitive (yPx and yPz then xPz). A fundamental
30 S. Rawat, S. Rawat / Computers in Industry 75 (2016) 27–34

relation that holds this property is termed as indiscernibility step involves the calculation of error signal using the equation
relation. For an information system I and P  A, the indiscernibility ej ¼ ðpj oj Þf ðnj Þ. The derivative of this function is given as f ðnj Þ ¼
¯
relation is given as IND(P) = {(x, z) 2 U2j 8 p 2 P, f(x, p) = a(z, p)}. 1
In a way such that (x, z) 2 IND(P), x and z are called as 1þexpðnj Þ and f¯ðnj Þ ¼ oj ð1oj Þ, where j = 1, 2, . . ., no. The value of

P-indiscernible. The efficacy of this relation is to divide the error signal in hidden neuron depends on the output neuron error
P
universe into several equivalence classes given as U/IND(P). that is given by the equation ek ¼ no j¼1 ðej wkj Þf¯ðnk Þ, where k = 1, 2,
. . ., nh and f ðnk Þ ¼ sk ð1sk Þ.
¯
4.2. Approximations Now, BP algorithm is used to adjust the weights between the
output layer and the hidden layer that is given by
Let us assume the approximate space as (U, P), where U signifies P
Dwkj ¼ ho ej sk . The error function E is given by E ¼ 12 no j¼1 ðpj oj Þ
2

the universe of objects. Let X is an element that belongs to U in a


and oj = f(nj). Putting these values in the original equation and
way that X 2 U, here U = (x1, x2, x3, . . .. . .) and P  U  U represents
further solving, we get Dwkj ¼ ho ej sk . The adjustment equations
an indiscernibility relation. Here, P denotes an equivalence
ðtþ1Þ ðtÞ ðtþ1Þ ðt Þ ðtþ1Þ
relation, and [x]P represents an equivalence class of an element are given a Dwkj ¼ hej sk þ aDwkj and wkj ¼ wkj þ Dwkj .
x 2 U with indiscernibility relation P. Hence, we can define two Here, j = 1, 2, . . ., no, k = 1, 2, . . ., ns, a is momentum rate and
subsets that as DwðkjtÞ represents the difference between the prior weight the
current weight.
 Upper approximations that are denoted by ðPXÞ and it can be
¯ Next step involves adjusting the weights between input layer
represented as P X ¼ fxj½x \ X 6¼ ;g. It contains all the subsets
¯ and the hidden layer that is given by Dvik ¼ ho  @@VE . Here E is the
ðP X ¼ PX þ BP ðXÞÞ which possibly belong to the set. ik
¯ same error function. On further solving we get
 Lower approximations that are denoted by ð PXÞ represented by P
PX ¼ fxj½x  Xg. It contains subsets those are certainly/definite- Dvik ¼ ho ek bi . Here, ek ¼ noj¼1 ðej wkj Þsk ð1sk Þ. The adjustment
ly belongs to the set. equation is given as vik
ðtþ1Þ ðt Þ
¼ vik þ Dvik
ðtþ1Þ
and Dvik
ðtþ1Þ
¼ ho ek bi þ
 Boundary approximations that are denoted by BP(X). It is the ðtÞ ðtÞ
difference between upper and lower approximation that denoted
aDv where Dv represents the difference between the prior
ik
, ik

by BP ðXÞ ¼ ðP X PXÞ. It neither belongs to Xa nor (U  X) and weight and the current weight, k = 1, 2, . . ., nh and i = 1, 2, . . ., ns and
¯ a is the momentum rate.
contains the uncertain regions of the data sets.
After the network is trained, the MSE function is calculated that
P P I 2
is given by Error ¼ 12 Ii¼1 no I
j¼1 ðpj oj Þ and termination conditions
4.3. Core and reducts are verified. If condition fails the whole process is started again.
Fig. 4 shows the flowchart of BPNN for training the network and
The concept of core and reduct is to filter out the unnecessary forecasting the output values.
information from the system. Reducts are the attribute, R  A such
that IN(R) = IN(A). Here IN(R) and IN(A), both represents an 6. Experiment and results
indiscernibility relation. The intersection of all reducts is called
as ‘‘CORE’’. Mathematically, it can be expressed as COR- The first and primary step involves the preprocessing of the
E(R) = \ Red(R). Here reduct is represented as Red(R). Core is an sensory data. It incorporates normalization and refining of the
important feature of the information system and is common for all dataset. The next step includes the utilization of RST for identifying
reducts. the crucial features from the dataset. The pertinent elements are
Let IS = (U, A) represents an information system with m objects. then nourished to the NN that is further trained by the LM-BP
The discernibility matrix of IS will contain m  m elements. Let the algorithm.
universe U = x1, x2, x3, . . ., xn and the matrix will be defined as S = sij,
is a jUj  jUj matrix in which the elements sij for an objects pair of 6.1. Step 1. Data preprocessing
(xi, U/D) is obtained by sij = {a 2 C : a(xi) 6¼ a(xj) ^ (d 2 D),
d(xi) 6¼ d(xj)), i, j = 1, 2 . . . n. The primary step involves processing of the data. It seeks to
Here sij is the set of attributes which classifies objects xi and xj refine the irrelevant and superfluous elements from the data. It
into different decision class labels using the partition of U/D. The usually employs several methodologies such as feature extraction,
physical meaning of the element sij is the objects xi and xj which can normalization, selection and transformation of the features. In our
be distinguished by any attribute in sij. The discernibility matrix case we have discretize the entire dataset to a single numeric value
‘‘S’’ is a symmetric matrix that follows sij = sji. to make the work easier.

6.2. Step 2. Feature reduction using RST


5. Back propagation neural network

6.2.1. (i) Approximations


Back-propagation algorithm is primarily employed to train the
Let X be a classification of U where X = {E1, E2, E3, . . ., En}, called
neural network. It calculates the minimum error function of
as the classes of X. The approximations of X in I (information
the network [8]. The initial weights are chosen randomly for the
system) is represented by PX ¼ f PE1 ; PE2 ; PE3 ; . . .; PEn g and
initialization of the network [9,10]. The gradient of the error function
P X ¼ fP E1 ; P E2 ; P E3 ; . . .; PEn g. The quality of approximation consists
is used to correct the input weights of the network [8]. Training the ¯ ¯ ¯ ¯ ¯
the ratio of all definitely/possibly classified objects to the total
network and use it for prediction is a chronological process.
objects of the system, which is mathematically expressed as
The first step involves setting the default value of network error Pn
ð PE Þ
(MSE) to zero and the training pattern error to one. The hidden g P ðXÞ ¼ j¼1U j . For every X 2 U, and P 2 Q the accuracy of the
neuron activity is calculated by the unipolar sigmoid function approximation is defined by N p ðXÞ ¼ Cardð PXÞ=CardðP XÞ.
P ¯
using the equation Sk ¼ f ð ns i¼1 bi vik Þ where k = 1, 2, . . ., Table 1 shows the values of upper and lower approximations.
nh. Likewise, the output neuron activity is calculated by using The accuracy of the approximation is 0.612903. The quality of
P
the function oj ¼ f ð nh k¼1 sk wkj Þ where j = 1, 2, 3,. . ., no. The next classification is 0.7600 and the entropy is 0.76908.
[(Fig._4)TD$IG] S. Rawat, S. Rawat / Computers in Industry 75 (2016) 27–34 31

Table 2
START Reducts from the dataset.

Reduct no. Reducts Length

Get the paern and Inialize the weights R1 A1,A2,A3,A8,A10 5


R2 A1,A2,A7,A8,A9,A10 6
R3 A1,A3,A6,A7,A8,A10 6
Iniate the training of the network R4 A1,A3,A4,A7,A8,A9,A10 7
R5 A1,A3,A5,A7,A8,A9,A10 7
R6 A1,A2,A3,A5,A8 5
Pass the training paern from input layer to the hidden layer
R7 A1,A2,A4,A5,A7,A8,A9 7
R8 A1,A2,A4,A7,A8,A10 6
Pass the training paern from hidden layer to the output layer R9 A1,A3,A4,A5,A6,A7,A8 7
R10 A1,A3,A4,A5,A7,A8,A10 7

Find the output node error and hidden node error


observed to be common in each pair. Hence, A1 and A8 represent
the core attributes of the information system.
Adjust the weights between the output and hidden layers
The next step involves the determination of the reduct that is
most appropriate for the final training of the network. All
Adjust the weights between the input and hidden layers 10 reducts are independently sustained to the NN and trained
with BP algorithm. The reduct having the best yield and minimum
MSE will be chosen for the final training of the network.
No
Need a different
training paern 6.3. Step 3. Training with BPNN to recognize the best reduct for the
system
Yes
The 10 reducts indicated in Table 2 are fed to the NN
Determine the funcon value
individually and trained using Levenberg–Marquardt back-propa-
gation (LM-BP) algorithm. The database is divided into three sets.
10% of the samples are utilized for validating the network to stop
Terminaon No
the training before over-fitting happens. 20% of the sample is
criteria achieved
utilized for testing and rest 70% is utilized for training the network.
A two-layered feed forward network employing a sigmoid function
Yes
is used. 10 output samples are taken and tested using LM-BP
END algorithm. Table 3 demonstrates the anticipated values of all the
10 samples trained by the BPNN for all the 10 reducts.
Fig. 4. Flowchart of BPNN for the training of network and prediction of outputs.
From Table 3 it is evident that the estimated output values of
reducts R3, R5 and R7 are very close to the original output values of
the sample. To chose the best reduct among these, mean square
Table 1 error for every one of the reducts is investigated and indicated in
Upper and lower approximation.
Fig. 5.
Decision Total objects present Lower Upper Accuracy It is clearly depicted from the Fig. 4 that the MSE error of R5 is
class in the class approximation approximation the least. This step infers that R5 having 7 input features is best
1 21 14 26 0.5385 suited to train the final network. To realize the accuracy of the
2 29 24 36 0.6667 proposed RS–BPNN, technique it is compared with the conven-
tional testing method of BPNN.

6.2.2. (ii) Finding the core and reducts 6.4. Step 4. Comparison of BPNN and RS–BPNN
In our case, there are 10 input attributes and 1 output attribute.
Applying RST to the database, we discovered 10 reducts that are In this step, we have employed the traditional testing approach
shown in Table 2. of NN with LM-BP algorithm without using RS. All the 10 input
Table 2 shows that 10 pairs of reducts extracted from the features from the database are fed to the network to train the
database. Taking the intersection of all the reducts, A1 and A8 are system, without reducing any input features.

Table 3
Predicted output values of the reducts.

Testing sample no. Actual value Predicted values of the reducts

R1 R2 R3 R4 R5 R6 R7 R8 R9 R10

1 3.000 2.987 2.955 3.000 2.999 3.000 3.105 2.999 2.927 2.993 3.129
2 4.000 3.992 4.224 4.000 4.000 4.000 3.949 4.000 4.2661 4.000 4.002
3 4.000 4.006 3.603 4.000 4.000 4.000 4.009 3.999 4.393 3.996 3.865
4 3.000 3.006 3.040 3.000 2.999 3.000 2.958 2.999 3.263 2.998 2.902
5 4.000 3.974 3.905 4.000 4.000 4.000 4.076 4.000 3.513 4.004 3.850
6 3.000 2.995 2.926 3.000 2.999 3.000 2.835 3.000 2.799 3.011 3.039
7 4.000 4.017 4.003 4.000 4.000 4.000 4.068 4.000 4.076 4.000 3.661
8 4.000 4.002 3.957 4.000 4.000 4.000 3.943 3.999 3.343 4.011 3.521
9 4.000 3.983 4.142 4.000 4.000 4.000 3.941 3.999 3.820 3.999 4.233
10 3.000 3.006 2.925 3.000 2.999 3.000 3.025 3.000 2.756 2.991 2.978
[(Fig._5)TD$IG]
32 S. Rawat, S. Rawat / Computers in Industry 75 (2016) 27–34

2.00E+00
1.80E+00
1.60E+00
1.40E+00
1.20E+00

MSE
1.00E+00
8.00E-01
6.00E-01
4.00E-01
2.00E-01
0.00E+00
1 2 3 4 5 6 7 8 9 10 11
Reducts

Fig. 5. MSE error of the reducts.

To evaluate the performance of both the methods, experiments each goal are obtained. Finally, we are able to compare the
were conducted to determine the relative significance of each accuracy of predictions from RS–BPNN and BPNN with the original
independent parameter of the output. The MSE is served to assess output values as depicted in Fig. 6.
the differences between the expected and actual values for both From Table 4 and Fig. 5, it can be concluded that RS–BPNN has
the methods. MSE surveys the quality of the prediction tool by accomplished a high accuracy as compared to the traditional
measuring the average of the squares of the errors, additionally approach of BPNN. The system’s forecasting efficiency and
characterized as the distinction between the estimator and the productivity have enhanced extensively while utilizing RS–BPNN.
assessed worth. It can be further inferred that of 10 input features only 7 features
Table 4 demonstrates the comparison between MSE for both the are required for making a well efficient decision system.
methods. Conventional and routine data fusion technologies are highly
It can be evident from Table 4 that network trained with RS– inefficient, incapable and ineffective in dealing with multisensory
BPNN has comparatively less testing error than the network data. It sometimes leads to a large error gap due to the uncertainty
trained with BPNN alone. Similarly, the training error of the former factors and the external predicaments experienced by the
is lesser than the later. The comparison between both the multisensory data system. Therefore, to avoid the baffling
methodologies recommends that NN with RS has the more systematic methods, several AI based methodologies are frequent-
significant effect than the system trained with NN alone. In ly utilized today. In this work a soft computing based classifier is
general, RS is used for reducing the unnecessary features from the built from the offline signature of the incoming multisensory data
dataset. Here, in this case it reduces the undesirable features from employing a hybrid model of RS and NN. Fig. 7 shows the enhanced
the dataset, making it less demanding for the system to get trained. view of the proposed architecture for predicting the pollution level
Correspondingly, the training error falls to almost null value when in the water sources.
the system is trained with RS. Now, few forecasting samples are The offline network comprises of a RS classifier and BPNN
taken as input vectors in the simulation, and the predictions of predictor. RS is used to expel the superfluous information from
database, thus reducing the data size. The crucial features are
distinguished and the dispensable features are removed from the
Table 4 dataset. In the next phase BPNN is used for training the network
Comparison of mean square error.
and estimating the decision rules. On the basis of the output of the
Mean square error classifier, the decision regarding the level of pollution can be easily
Training error Testing error made by the system’s operator. This permits him to take
proprietary steps and initiate the maintenance schedule in time.
BPNN 1.544e2 1.6928
RS–BPNN 0 1.0650
The problems encounter in the data fusion applications, for
example, multisensory systems drives us to combine data fusion
[(Fig._6)TD$IG]
BP Actual value RS-BP
5
4.5
4
3.5
Output values

3
2.5
2
1.5
1
0.5
0
1 2 3 4 5 6 7 8 9 10
Samples

Fig. 6. Comparison between both the methods and the original values.
[(Fig._7)TD$IG] S. Rawat, S. Rawat / Computers in Industry 75 (2016) 27–34 33

Fig. 7. An enhanced architecture for the pollution monitoring system.

and data mining. This work shows and discusses about how data [5] D. Crisan, A. Doucet, A survey of convergence results on particle
filtering methods for practitioners, IEEE Trans. Signal Process. 50 (3) (2002)
mining and data fusion techniques together can play a mutually 736–746.
supportive role. This works demonstrates a promising future in the [6] G. Shafer, A Mathematical Theory of Evidence, Princeton University Press,
multisensory data processing fusion system. 1976.
[7] A.R. James, Toward reducing failure risk in an integrated vehicle health
maintenance system: a fuzzy multi-sensor data fusion Kalman filter approach
7. Conclusion for IVHMS, Expert Syst. Appl. 39 (2012) 9821–9836.
[8] H. De Garis, Building artificial nervous systems using genetically programmed
neural network modules, in: Machine Learning: Proceedings of the Seventh
Data fusion and data mining in the multisensory system for International Conference, 2014, pp. 132–139.
accomplishing an accurate and precise solution is a challenging [9] L.K. Hansen, P. Salamon, Neural network ensembles, IEEE Trans. Pattern Anal.
Mach. Intell. 12.10 (1990) 993–1001.
errand. To deal with the problem, this paper attempted to propose [10] T.D. Sanger, Optimal unsupervised learning in a single-layer
a hybrid architecture that links RST with NN through an example of linear feed forward neural network’’, Neural Netw. 2 (6) (1989)
pollution monitoring system. RS is used as a preprocessing layer as 459–473.
[11] H. Zhu, O. Basir, A novel fuzzy evidential reasoning paradigm for data fusion
a part of the RS–BP network to filter out the noisy, uncertain and with applications in image processing, Soft Comput. J. – Fusion Found.
superfluous data. It mines the data and identifies the crucial Methodol. Appl. 10 (12) (2006) 1169–1180.
features from the dataset. These attributes are fed to the NN that is [12] L. Yong, X. Congfu, P. Yunhe, A new approach for data fusion: implement
rough set theory in dynamic objects distinguishing and tracing, in: Proc. of the
trained using LM-BP algorithm. On comparing the prediction
IEEE International Conference on SMC, 2004, 3318–3322.
results of BPNN and RS–BPNN, the latter was discovered more [13] J.F. Peters, S. Ramanna, A. Skowron, J. Stepaniuk, Z. Suraj, Sensor fusion:
superior than the former. In the future work, we would compare a rough granular approach, in: Proc. of the IFSA World Congress and 20th
the accuracy and reliability of the proposed architecture with other NAFIPS International Conference, 2001, 1367–1371.
[14] W. Haijun, C. Yimin, Sensor data fusion using rough set for mobile robot
ML and soft computing techniques. Also, we would test the system, in: Proc. of the IEEE/ASME International Conference on Mechatronic
proposed system in a real time scenario. It would help us to get the and Embedded Systems and Applications, 2006, 1–5.
feedback from the system’s control operator so that the system can [15] R.R. Yager, Generalized probabilities of fuzzy events from fuzzy belief
structures, Inf. Sci. 28 (1) (1982) 45–62.
be improved. This is an additional aspect of one of the possibilities [16] R.P.S. Mahler, Random sets: unification and computation for information
of our research in the near future. fusion – a retrospective assessment, in: Proc. of the International Conference
on Information Fusion, 2004, 1–20.
[17] S.P. McLaughlin, R.J. Evans, V. Krishnamurthy, Data incest removal in a
Compliance with ethical standards survivable estimation fusion architecture, in: Proc. of the International
Conference on Information Fusion, 2003, 229–236.
[18] L.Y. Pao, M. Kalandros, Algorithms for a class of distributed
We, the authors, hereby jointly declare that there is no architecture tracking, in: Proc. of the American Control Conference, 1997,
conflict of interest of any kind whatsoever, whether directly, 1434–1438.
indirectly or materially with respect to financial or non-financial [19] S.P. McLaughlin, R.J. Evans, V. Krishnamurthy, A graph theoretic
approach to data incest management in network centric warfare, in:
funding, and use of any humans and animals concerning this Proc. of the International Conference on Information Fusion, 2005,
manuscript. 1162–1169.
[20] T. Brehard, V. Krishnamurthy, Optimal data incest removal in
Bayesian decentralized estimation over a sensor network, in: IEEE
References International Conference on Acoustics, Speech and Signal Processing, 2007,
III173–III176.
[1] A. Makarenko, A. Brooks, T. Kaupp, H.F. Durrant-Whyte, F. Dellaert, [21] J. Dezert, A. Martin, F. Smarandache, Comments on ‘‘A new combination of
Decentralized data fusion: a graphical model approach, in: Proc. of the evidence based on compromise’’ by K. Yamada, Fuzzy Sets Syst. 160 (2009)
International Conference on Information Fusion, 2009, 545–554. 853–855.
[2] W.J. Farrell, C. Ganesh, Generalized chernoff fusion approximation for practical [22] S.J. Wellington, J.K. Atkinson, R.P. Sion, Sensor validation and fusion using the
distributed data fusion, in: Proc. of the International Conference on Nadaraya–Watson statistical estimator, in: Proc. of the International
Information Fusion, 2009, 555–562. Conference on Information Fusion, 2002, 321–326.
[3] H.F. Durrant-Whyte, T.C. Henderson, Multisensor data fusion, in: B Siciliano, O. [23] P.H. Ibarguengoytia, L.E. Sucar, V. Vadera, Real time intelligent sensor
Khatib (Eds.), Handbook of Robotics, Springer, 2008, pp. 585–610. validation, IEEE Trans. Power Syst. 16 (4) (2001) 770–775.
[4] C.Y. Dong, Q. Yuan, Q. Wang, A combined wavelet analysis-fuzzy adaptive [24] J. Frolik, M. Abdelrahman, P. Kandasamy, A confidence-based approach to the
algorithm for radar/infrared data fusion, Expert Syst. Appl. 37 (2010) self-validation, fusion and reconstruction of quasi redundant sensor data, IEEE
2563–2570. Trans. Instrum. Meas. 50 (6) (2001) 1761–1769.
34 S. Rawat, S. Rawat / Computers in Industry 75 (2016) 27–34

[25] B. Han, T.J. Wu, Data mining in multisensor system based on rough set theory, [27] E. Cardarelli, L. Sabattini, C. Secchi, C. Fantuzzi, Multisensor data fusion for
IEEE Proc. Am. Control Conf. 6 (2001) 4427–4431. obstacle detection in automated factory logistics, in: IEEE International
[26] B. Khaleghi, A. Khamis, F. Karray, Random finite set theoretic based soft/hard Conference on Intelligent Computer Communication and Processing (ICCP),
data fusion with application for target tracking, in: Proc. of the IEEE IEEE, (2014), pp. 221–226.
International Conference on Multisensor Fusion and Integration for Intelligent [28] B. Khaleghi, et al., Multisensor data fusion: a review of the state-of-the-art,
Systems, 2010, 50–55. Inf. Fusion (2011), http://dx.doi.org/10.1016/j.inffus.2011.08.001.

You might also like