Professional Documents
Culture Documents
SỐ 04 (CS.01) 2021 TẠP CHÍ KHOA HỌC CÔNG NGHỆ THÔNG TIN VÀ TRUYỀN THÔNG 22
A DISTRIBUTED APPROACH FOR SUPERVISED SOM AND APPLICATION TO FACIES CLASSIFICATION
discussion of the proposed approach. Section 5 concludes The training process goes through many iterations to
what have been accomplished in the research. update the weights of neurons in the network. Each
iteration is also called a competition. In the tth iteration, a
II. SUPERVISED SELF-ORGANIZING MAP data vector 𝑥 is randomly selected from the input data set
𝑋 . The algorithm determines the "winning" neuron 𝑤𝑐,
A. Self-organizing map
which is also known as the best matching unit (BMU), for
The self-organizing map (SOM) composes of a set of the sample 𝑥. Euclidean distance, 𝐷(𝑥, 𝑤𝑖 ), is commonly
nodes (or neurons) connected to each other via the topology used as the measure to determine how close a neuron i is
of rectangle or hexagon. Each neuron contains a vector of to the input sample 𝑥. The BMU is defined as the neuron
weights of the same dimension as the input data. There are having the smallest distance to the sample. The weight
usually to types of SOM representation [14], semantic vector, 𝑤𝑖 (𝑡), of the BMU 𝑖 is then updated by a learning
representation and spatial positioning representation. rule:
Semantic representation includes neurons ordered in a
network having 1 or 2 dimensions. The input data is 𝑤𝑖 (𝑡) = 𝑤𝑖 (𝑡 − 1) +
mapped into a K neuron network. Each neuron can
+𝛼(𝑡 − 1). ℎ𝑖𝑐 (𝑡 − 1). 𝐷(𝑥(𝑡 − 1), 𝑤𝑖 (𝑡 − 1))
represent multiple input data samples. The network of
neurons is painted with different colors representing Where, 𝑡 is the current iteration; 𝛼(𝑡 − 1) is the
different clusters of input data. Each color corresponds to a learning rate at the previous iteration , which normally
set of similar input samples. decreases during the training process, 𝛼(𝑡) =
𝛼0
; ℎ𝑖𝑐 (𝑡 − 1) is the neighborhood function at
1+𝑑𝑒𝑐𝑎𝑦𝑟𝑎𝑡𝑒∗𝑡
the previous iteration. The neighborhood function
determines which nearby neurons are updated along with
the BMU i. There are two popular neighborhood functions
used in the literature, which are bubble and Gaussian
functions [15]. In bubble function algorithm, all neurons
within the neighborhood region are updated with the same
rate, i.e.
1 if neuron c is inside neighborhood
ℎ𝑖𝑐 (𝑡) = {
0 if neuron c is outside neighborhood
In Gaussian function approach, the updating rate of one
neighboring neuron depends on how close it is between
that neuron and the BMU i. The Gaussian neighborhood
a) Semantic representation function is defined as
2
(𝐷(𝑤𝑐 , 𝑤𝑖 ))
−
ℎ𝑖𝑐 (𝑡) = 𝑒 2𝜂𝑖(𝑡)2
SỐ 04 (CS.01) 2021 TẠP CHÍ KHOA HỌC CÔNG NGHỆ THÔNG TIN VÀ TRUYỀN THÔNG 23
Hoa Dinh Nguyen
problems. In order to tackle supervised classification and relationships with other depofacies. Characterization
problems, traditional SOM must be modified to increase of depofacies is important for reservoir modeling and
the classification accuracy. predicting reservoir continuity away from the borehole.
For example, a channel axis depofacies may consist of
B. Supervised SOM-LVQ algorithm conglomerate lithofacies and massive sandstone
Learning vector quantization (LVQ) is developed from lithofacies, and have a predictable range of widths and
SOM algorithm. LVQ is used for supervised learning predictable relationship with channel margin depofacies.
applications. The training rule applied in the LVQ is an Identification of such geologic units from core data and
association method for training the SOM network in a manual interpretation from wire-line logs has been
supervised learning manner [2]. Specifically, each neuron established, but core sampling is typically limited, and
in the competitive network will be assigned the label of the manual interpretation of wire-line logs involves some
data cluster it represents. Many different neurons can have degree of uncertainty and subjectivity [17].
the same data class label. After training, a given the new
Recently, most of the researches on facies classification
input vector will be assigned the class label of the neuron
are based on well-log data. It is desirable to find either the
closest to it.
relationship between well-log measurements and facies
In this research, the two-stage SOM-LVQ model for classes or well-logs patterns corresponding to each class
classification problem introduced in [13] is adopted. In this representation. There have been a lot of methods based on
supervised model, the training data is first clustered by an wire-line log measurements including statistical
SOM algorithm. The label for each neuron is then assigned approaches, fuzzy methods, and artificial neural networks
according to the class of the nearest data sample from the [17].
training set. LVQ is finally applied to train the whole
labeled network. After training process, the labeled neurons B. Facies classification based on distributed supervised
are moved closer to the regions dominated by the data SOM
having the same class label. In case the training data is not In this study, supervised SOM is applied on facies
distributed in specific regions for each class label, the label classification in a distributed manner using well-log data.
neurons are located close to some local data samples. General speaking, instead of using the whole data features
collected during the drilling process to generate a
III. DISTRIBUTED SUPERVISED SOM AND generalized supervised SOM model for predicting the
APPLICATION TO FACIES CLASSIFICATION class label of each well-log sample, multiple local SOMs
In this research, a distributed supervised SOM is generated from many individual local feature subsets.
framework is proposed and applied on facies classification The output of all local SOMs are then fused to provide a
problem. The data-set is assumed to be collected from final decision on the facies label of the log data.
different types of sensors. Each sensor contributes one
piece of information, and is presented by a sub set of general feature set
features. One local supervised SOM is built based on the
sub feature set from each local sensor. The outputs of all
local supervised SOMs are then fused to produce the final
decision on the class of the input data. The distributed Feature Feature Feature
supervised SOM method is depicted in Figure 2. subset 1 subset 2 subset k
A. Facies classification
Facies are the overall characteristics of a rock unit that
reflect its origin and differentiate the unit from others Supervised Supervised Supervised
around it. Each facies class distinguishes itself from other SOM 1 SOM 2 SOM k
classes based on mineralogy and sedimentary source, fossil
content, sedimentary structures and texture. In reservoir
characterization and simulation, the most important facies
properties are the petrol-physical characteristics which
control the fluid behavior in it [16]. Some certain facies Decision fusion
classes exhibit characteristic measurement signatures that
help facilitate the prediction of some important properties
such as permeability, porosity, and liquid content. Hence, Class labels for each data sample
correct presentation of facies classes for well-log data is an
important and challenging task for oil and gas engineers. Figure 2. distributed supervised SOM framework
Deep-water reservoirs are deposited in a wide range of A data set with full attribute values often has many
depositional environments, and exhibit a variety of different distributions in the data space, each of which may
temporal and spatial scales. Detailed core description of have different and complex properties. In addition, one
reservoir intervals allows identification of facies and strati- general SOM-LVQ network can only organize the neurons
graphic units at multiple scales. At the small scale, according to the distribution of general data clusters. This
lithofacies are rocks with similar lithology, sedimentary means that one single SOM-LVQ model can ignore the
structures and rock properties. In many depositional distribution of complex data clusters located in a larger data
environments these can be grouped together into set. Therefore, with complex data, especially geological
depositional facies (or depofacies) that represent data, if only one SOM-LVQ network is used to model the
genetically related deposits with predictable dimensions entire data set, its efficiency in data interpretation is not
SỐ 04 (CS.01) 2021 TẠP CHÍ KHOA HỌC CÔNG NGHỆ THÔNG TIN VÀ TRUYỀN THÔNG 24
A DISTRIBUTED APPROACH FOR SUPERVISED SOM AND APPLICATION TO FACIES CLASSIFICATION
high. The main advantages of a distributed approach for The data-set is divided into two subsets, training data
classification using SOM-LVQ model are two folds. First, subset contains all samples from well 1 and well 2, and
the original data set will be examined in smaller testing data subset includes the data from well 3. It can be
dimensional sub-spaces to find such data distributions that seen that the number of facies labels 1 and 3 is small, which
would be difficult to discern if viewed in the general all- means those two facies rarely present in those three wells.
dimensional space of the original data. Second, there is In this research, all samples of classes 1 and 3 are omitted,
always an overlap between the data of different facies which means only facies labels 0 and 2 are used.
classes. The majority part of missed classified well-log
Table 1. The distribution of data samples in the dataset
samples fall into the overlapping regions of different facies.
By using local information from different feature subsets, Information Data in well Data in well Data in well
the information of each well-log data is put under different 1 2 3
angles and processed individually. This approach can help Number of 438 457 232
exploit all valuable information for the classification samples
process. Each local supervised SOMs provides different Samples of 249 194 146
probabilities that the data sample belongs to each of class 0
possible facies labels. Samples of 5 13 0
class 1
C. Decision fusion Samples of 169 211 80
Multiple supervised SOM models produce multiple class 2
local decisions from different data feature subsets. An Samples of 15 39 6
efficient fusion process, which combines all these local class 3
decisions, may influence the classification performance of
B. Experimental scenarios
the whole system. There are normally two main rules for
decision fusion process, majority voting and confidence The accuracy score is used to evaluate the
score based fusion [18]. classification performance of the system. The accuracy is
calculated as the ratio of correct decision number to total
In majority voting approach, the final class label is number of testing samples.
selected as the one having the most number of local
decisions on it. A modification version of majority voting 𝑁𝑡𝑟𝑢𝑒 𝑝𝑟𝑒𝑑𝑖𝑐𝑡𝑖𝑜𝑛
𝐴𝑐𝑐𝑢𝑟𝑎𝑐𝑦 =
is to select the class label having the highest weights 𝑁𝑡𝑜𝑡𝑎𝑙 𝑝𝑟𝑒𝑑𝑖𝑐𝑡𝑖𝑜𝑛
among all local classifiers. There can be many ways to
define the weight for each local decision. Regarding This evaluate metric works well in case there is a
supervised SOM classifiers, where the label of the test data balance among the sample numbers of all classes, which is
is decided based on the neuron closest to that data, the true for this data-set.
reasonable way to calculate the weight of the output is the The experiments have been conducted to evaluate the
reciprocal of the best matching distance. classification performance of the proposed distributed
In confidence score based approach, the final class classification approach, and compare it with the
label is assigned based on the highest average value of all centralized supervised SOM model.
confidence scores from local classifiers. In other words, The distributed model adopts multiple small sized
each local output is associated with a confidence value. SOM-LVQ models working on multiple feature subsets. A
The confidence score of data sample 𝑥 belonging to class total of 10 randomly selected feature sets are generated. In
𝑐 can be defined based on the distances from neurons and order to randomly create different feature sets, each feature
data sample as follows. is first assigned a weight based on its correlation with the
1 output labels. Higher correlation means higher weights.
∑𝑘𝑖=1,𝑦𝑖=𝑐 The selection process is conducted in 10 iterations, in each
𝐷(𝑤𝑖 , 𝑥)
𝑆𝑐 (𝑥) = of which one feature is selected. The features having
1
∑𝑘𝑖=1 higher weights tend to be selected more times than others.
𝐷(𝑤𝑖 , 𝑥) This means high correlated features are expected to
Where, 𝑦𝑖 is the label of neuron 𝑖; 𝑘 is the number of present in many feature subsets. 10 local supervised SOM
neurons nearest to the data sample 𝑥 ; 𝐷(𝑤𝑖 , 𝑥) is the models are created based on 10 feature subsets. Majority
distance between neuron 𝑖 and sample 𝑥. The average of all voting based on maximum average weights are used in the
confidence scores of local classifiers having the same decision fusion stage.
output label is used to determine the final label of the data
A general traditional supervised SOM is generated
sample at the fusion stage.
from the whole feature set. The training process for single
IV. EXPERIMENTAL RESULTS supervised SOM model is the same as that for multiple
distributed SOM models. The only difference is that the
A. Dataset number of training iterations for multiple distributed SOM
The proposed distributed classification approach is smaller than the single SOM model due to their simpler
based on supervised SOM is evaluated using the well-log structures. The training process is repeated multiple times
data collected from three wells in Cuu Long basin. There and the best models for each scenario are recorded.
are 4 facies classes in the data-set. The distribution of data C. Results and discussion
sample in each facies class in the data-set is presented in
Table 1. Table 2 presents the training parameters of all
supervised SOM models implemented in the experiments
together with their classification accuracy scores.
SỐ 04 (CS.01) 2021 TẠP CHÍ KHOA HỌC CÔNG NGHỆ THÔNG TIN VÀ TRUYỀN THÔNG 25
Hoa Dinh Nguyen
The experimental results in table 2 show that the learning models. Compared with KNN model, the
combination of multiple smaller sized SOM models works distributed SOM-LVQ model requires less memory
better than one single complicated SOM model. This result resource. Specifically, KNN model needs to store all
is more meaningful considering the training time as well training sample to find the nearest data during the testing
as the processing time of the classification models. In each process, while the SOM-LVQ models only need to store
training iteration, the distributed models only have to work the weights of their neurons after training process. The
on 9 neurons compared with 100 neurons of the single number of the neurons is normally much smaller than the
models. Additionally, distributed models can work in number of training data samples.
parallel, which helps speed up the computational process
significantly. The distributed supervised SOM Distributed SOM-LVQ model and random forest have
classification system is more flexible than the single some similar characteristics. Random forest may have an
supervised SOM model since different fusion rules can be advantage over the distributed SOM-LVQ model when it
applied to adjust the classification performance of the can process the training data in multiple small data subsets,
whole system. This is meaningful in real practice where in which some special properties of the training data can be
different real data-sets may require different system recognized. However, this distributed SOM-LVQ model
structure setups. can be further modified to cope with different smaller
subsets of training data, from which more local supervised
Table 2. Model parameters and accuracy scores. SOM classifiers can be created. As a results, the
Multiple
classification performance of distributed SOM-LVQ model
Single SOM- can be adjusted accordingly.
Parameters distributed SOM-
LVQ model
LVQ models
V. CONCLUSIONS
Number of SOM
1 10 In this paper, a new distributed classification approach
models
SOM size 10x10 3x3 based on supervised SOM is introduced. Multiple SOM-
Weight initialize LVQ models are generated from different local feature sets
PCA PCA and their decisions are fused to produce the final class label
method
2000 for the input data. Each local supervised SOM can exploit
Number of training
2000 (200 for each different pieces of information presented in its smaller
iterations
local model) feature subset. Fusion stages combines all local
Initial updating rate 1 1 classification decisions from multiple supervised SOM
Updating rate
1 1
models to output the final class label for the test sample.
declining coefficient Multiple pieces of the information are combined using
Neighborhood different fusion rules, which open up the flexibility of the
Bubble Bubble
function proposed system to cope with different situations in real life
Neighborhood radius 1.5 1.5 practice. This newly distributed classification system is
Accuracy score 0.9181 0.9397 applied on facies classification problems, which is a well-
The experiment is further conducted on different known and difficult issue in geology. The experimental
classification machine learning models, such as decision results show that the distributed classification approach for
tree, Naïve Bayes, K nearest neighbor (K=5) and Random supervised SOM works better than the traditional single
forest (n=10). The results are presented in Figure 3. supervised SOM model as well as many other conventional
machine learning models. In our further research, more
experiments with bigger number of data features will be
conducted to investigate the usefulness and flexibility of
this distributed approach.
REFERENCES
[1] J.-S. R. Jang and C.-T. Sun, "Neuro-Fuzzy Modeling and
Control," Proceedings of The IEEE, pp. 378-406, 1995.
[2] M.T. Hagan, H.B. Demuth, M.H. Beale, O.D. Jesus.
“Neural network design (2nd edition)”, ISBN-10: 0-
9717321-1-6, ISBN-13: 978-0-9717321-1-7, 1996.
[3] Mia Louise Westerlund, "Classification with Kohonen Self-
Organizing Maps," Soft Computing, 2005.
[4] T. Kohonen. "Self-Organized Formation of Topologically
Correct Feature Maps", Biological Cybernetics. 43 (1), pp.
59 - 69, 1982.
Figure 3. Accuracy scores of different classification [5] N. Skific and J. Francis, “Self-Organizing Maps: A
models on the same data-set Powerful Tool for the Atmospheric Sciences”, 2012.
[6] S. Haykin, “Neural Network and Learning Machines”, 3rd
The experimental results show that the distributed ed., Pearson, 2008.
SOM-LVQ model works best in comparison with other [7] A. Rauber and D. Merkl, "Automatic Labeling of Self-
machine learning models. The decision tree and Naïve Organizing Maps: Making a Treasure-Map Reveal Its
Bayes models both have the lowest accuracy scores, while Secrets", Methodologies for Knowledge Discovery and
KNN and random forest have almost the same Data Mining, p. 228 – 237, 1999.
classification performance. [8] P.N. Suganthan. "Hierarchical overlapped SOM's for
pattern classification", IEEE Transactions on Neural
The distributed SOM-LVQ model also has some Networks, vol. 10, pp. 193-196, 1999.
additional advantages over other traditional machine
SỐ 04 (CS.01) 2021 TẠP CHÍ KHOA HỌC CÔNG NGHỆ THÔNG TIN VÀ TRUYỀN THÔNG 26
A DISTRIBUTED APPROACH FOR SUPERVISED SOM AND APPLICATION TO FACIES CLASSIFICATION
[9] O. Kurasova, "Strategies for Big Data Clustering", 2014 tán sẽ hữu ích hơn khi số lượng các tính năng đầu vào là
IEEE 26th International Conference on Tools with Artificial lớn, và là một giải pháp linh hoạt cho nhiều bài toán phân
Intelligence, 2014.
loại trên thực tiễn.
[10] P. Stefanovič, O. Kurasova. "Outlier detection in self-
organizing maps and their quality estimation", Neural Từ khóa: bản đồ tự tổ chức, huấn luyện lượng tử hoá
Network World, 28 (2), pp. 105-117, 2018.
vector, phân tán, phân loại facies
[11] L.A. Silva, E.D.M. Hernandez. "A SOM combined with
KNN for classification task", Proceedings of the Hoa Dinh Nguyen earned
International Joint Conference on Neural Networks, pp. bachelor and master of science
2368-2373, 2011. degrees from Hanoi University of
[12] M. Mishra, H.S. Behera. "Kohonen Self Organizing Map Technology in 2000 and 2002,
with Modified K-means clustering For High Dimensional respectively. He got his PhD.
Data Set", International Journal of Applied Information degree in electrical and
Systems, pp. 34-39, 2012.
computer engineering in 2013
[13] Hoa Dinh Nguyen, “A Boosting Classification Approach from Oklahoma State University.
Based on SOM”. Journal of Science and Technology on
Information and Communications, Vol 1 No 2, 2020. He is now a lecturer in
information technology at PTIT.
[14] M. C. Kind and R. J. Brunner, "SOMz: photometric redshift
PDFs with self-organizing maps and random atlas," 2013. His research fields of interest
include dynamic systems, data
[15] N. Wangsoh, W. Watthayu and D. Sukawat, "Appropriate
Learning Rate and Neighborhood Function of Self- science, and machine learning.
organizing Map (SOM) for Specific Humidity Pattern Email: hoand@ptit.edu.vn
Classification over Southern Thailand," International
Journal of Modeling and Optimization, pp. 61-65, 2016.
[16] O. Serra. “Fundamentals of well-log interpretation. Volume
2: the interpretation of log data”, ENSPM, Etudes et
Productions Schlumberger, Montrouge, France, 632p.,
1985.
[17] C.M. Gifford, A. Agah.“ Collaborative multi-agent rock
facies classification from wireline well-log data”,
Engineering Applications of Artificial Intelligence, 23,
pp.1158–1172, 2010.
[18] A. Mandelbaum and D. Weinshall, "Distance-based
Confidence Score for Neural Network Classifiers,"
2017.W.D.Patrick, Natural sciences citations and
references. Available online at
ftp://ftp.freesoftware.com/pub/
tex/ctan/macros/contrib/supported/natbib, (1997) 2-10.
SỐ 04 (CS.01) 2021 TẠP CHÍ KHOA HỌC CÔNG NGHỆ THÔNG TIN VÀ TRUYỀN THÔNG 27