Professional Documents
Culture Documents
Bab 1 5 Edited
Bab 1 5 Edited
Models for ANN and ANFIS Based on PCA and BGA Feature
Selection
DESMIRA
ELECTRICAL ENGINEERING
2023
Faculty of Electrical Engineering
Desmira
Electrical Engineering
2023
Improved Computation Time in Energy Consumption Prediction
Models for ANN and ANFIS Based on PCA and BGA Feature
Selection
DESMIRA
A thesis submitted
in fulfillment of the requirements for the degree of
Doctor of Philosophy
2023
DECLARATION
I declare that this Choose an item. entitled “Improved Computation Time in Energy
Consumption Prediction Models for ANN and ANFIS Based on PCA and BGA Feature
Selection” is the result of my own research except as cited in the references. The
Choose an item.
has not been accepted for any degree and is not concurrently submitted in
Signature :
Name :
Date :
APPROVAL
Choose an item. and in my opinion, this Choose an item. is adequate in terms of scope
and quality for the award of the degree of Doctor of Philosophy Electrical Enginering.
Signature :
Supervisor Name :
Date :
DEDICATION
This thesis discusses the increase in computation time by combining principal component
analysis (PCA) and binary genetic algorithm (BGA) feature selection. The selection of
PCA and BGA features is implemented using the ANN and ANFIS models to predict
energy use in a building, especially in the UNTIRTA PVTE laboratory
Tesis ini membincangkan peningkatan masa pengiraan daripada gabungan pemilihan ciri
Analisis Komponen Utama (PCA) dan Algoritma Genetik Binari (BGA). Pemilihan ciri
PCA dan BGA dilaksanakan menggunakan model ANN dan ANFIS dalam meramal
penggunaan tenaga dalam sesebuah bangunan terutamanya di makmal UNTIRTA PVTE
Dalam Bab 2, kami menyemak dengan teliti literatur yang berkaitan tentang kaedah
pemilihan ciri. Kaedah pemilihan ciri sedia ada tidak dapat memberikan masa
pengkomputeran yang lebih baik. Kami ambil perhatian bahawa kebanyakan kaedah
pemilihan ciri sedia ada tidak dapat meningkatkan masa pengiraan yang lebih baik untuk
meramalkan penggunaan tenaga. Satu lagi kelemahan kaedah pemilihan ciri sedia ada
ialah kebanyakan literatur sedia ada memilih ciri satu demi satu dan tidak memberikan
ukuran langsung untuk menilai sama ada ciri tambahan akan ditambah atau tidak. Untuk
mengatasi masalah ini, mereka memerlukan kaedah yang digunakan untuk memilih ciri-
ciri yang dipilih supaya pengumpulan data akan berlaku yang boleh meningkatkan masa
pengiraan dalam proses latihan dan ujian. Walau bagaimanapun, dalam aplikasi sebenar,
adalah sukar untuk menganggarkan bilangan ciri berguna sebelum proses pemilihan ciri.
Tesis ini menangani kelemahan ini, dan mengisi jurang dalam literatur pemilihan ciri.
Dalam Bab 3 kami mencadangkan pendekatan pembangunan pemilihan ciri PCA dan
pemilihan ciri BGA. Terdapat tiga bahan baru. Pertama, dengan memasukkan
pengelompokan data dalam PCA, kita hanya menjadikan pengelompokan menjadi 2
pengelompokan data. Kedua, kita dapati pengelompokan data yang setakat ini dinilai
dalam pengelompokan yang dilakukan secara individu menghasilkan pengelompokan data
yang agak banyak. Ketiga, kami mendapati bahawa perbandingan masa dan ketepatan
pengiraan tidak selalu menunjukkan nilai yang lebih baik. Akhir sekali, berdasarkan
gabungan pemilihan ciri PCA dan BGA, kami menilai kepentingan ciri menggunakan
hanya 2 set data. Kelebihan mengumpulkan data ini ialah ia boleh meningkatkan masa
pengiraan apabila proses latihan dan ujian sedang dijalankan. Keputusan eksperimen
menunjukkan bahawa menggabungkan pemilihan ciri PCA dan BGA boleh meningkatkan
masa pengiraan dengan model ANN dan ANFIS. Dalam Bab 4, kami mendapati bahawa
dalam beberapa situasi di mana pelaksanaan pemilihan ciri dalam PCA dengan hanya 2
kumpulan data boleh meningkatkan masa pengiraan menjadi lebih pantas. Inilah yang
mendorong saya dalam masalah penyelidikan bagaimana pengumpulan data berlaku
dalam PCA dengan hanya menggabungkan 2 kumpulan data supaya saya dapat
menggabungkan aplikasi pemilihan ciri BGA dengan gelung proses. Untuk mendapatkan
hasil pengiraan yang lebih baik, kami menggunakan pengumpulan data dalam PCA dan
gelung proses dalam BGA. Keputusan eksperimen menunjukkan bahawa keberkesanan
gabungan pemilihan PCA dan BGA boleh membawa kepada pengiraan data yang lebih
baik. Sebagai tambahan kepada kaedah pemilihan ciri yang menggabungkan PCA dan
BGA, kami mengkaji dua kaedah pemilihan ciri dengan pengelompokan pada PCA dalam
Bab 5, khususnya, kami mencadangkan teknik pengelompokan pada PCA menggunakan
hanya pengelompokan data PC1 dan PC2 di mana nilai ƛ ialah satu. untuk PC1 dan Jika
nilainya di bawah sifar maka ia diletakkan dalam PC2. Pada langkah pertama, kami
menggunakan model yang diuji, iaitu ANN dan ANFIS, dalam langkah ini model diuji
tanpa menggunakan pemilihan ciri. Kedua, kami menguji pemilihan ciri menggunakan
First and foremost, I want to thank and praise Almighty God, my Creator, my Sustainer,
for all that I have received since the beginning of my life. I would like to express my ap-
preciation to Universiti Sultan Ageng Tirtayasa (UNTIRTA) and Universiti Teknikal
Malaysia Melaka (UTeM) for providing a research platform. Thanks also to the Malaysian
Ministry of Higher Education (MOHE) for its financial assistance.
My highest appreciation goes to my main supervisor, Ir.Dr. Norazhar Bin Abu Bakr, for all
his support, advice and inspiration. Don't forget to also thank the accompanying supervi-
sor, Dr. Mohd Ruzaini Hashim and Dr. Romi wiryadinata, S.T,. M.Eng who have provided
guidance and direction during the work of this dissertation.
For my husband who I love the most because of Allah who has given spirit and hope. For
my children, whom I love because of Allah, Ayra's sister, Aisya's sister, Ardam's brother,
Azkian's brother and Arman during this dissertation are often left behind so that they can
complete the dissertation. For those of you who have been called by Allah, hopefully they
will be placed in the back in the sun of Allah. Make it easy for all my brothers and sisters
to always be in the protection of Allah.
PAGE
DECLARATION
APPROVAL
DEDICATION
ABSTRACT..........................................................................................................................ii
ABSTRAK...........................................................................................................................iii
ACKNOWLEDGEMENTS.................................................................................................v
TABLE OF CONTENTS....................................................................................................vi
LIST OF TABLES..............................................................................................................ix
LIST OF FIGURES............................................................................................................xi
LIST OF SYMBOLS AND ABBREVIATIONS............................................................xvi
LIST OF APPENDICES....................................................................................................17
LIST OF PUBLICATIONS...............................................................................................18
CHAPTER 1 INTRODUCTION.....................................................................................19
1.1 Background....................................................................................................................19
1.2 Problem Statement.........................................................................................................20
1.4 Scope of Research..........................................................................................................23
1.5 Contribution of Research................................................................................................25
1.6 Thesis Outline.................................................................................................................25
CHAPTER 2 LITERATURE REVIEW.......................................................................27
2.1 Introduction....................................................................................................................27
2.2 Feature Selection Using PCA ( Principal Component Analysis )..................................33
2.3 Prediction of Energy Consumption Using the Feature Selection Binary Genetic
Algorithm (BGA).............................................................................................................40
2.3.1. General Structure of Binary Genetic Algorithm (BGA)................................44
2.3.2. Chromosomal Coding.....................................................................................46
2.3.3. Genetic Operators...........................................................................................47
2.3.4. Selection Process............................................................................................47
2.3.5. Crossover........................................................................................................49
2.3.6. Mutation.........................................................................................................52
2.4.Energy Consumption Prediction Using Artificial Neural Network (ANN) Model........56
2.4.1 Initialize..........................................................................................................57
2.4.2 Activation........................................................................................................58
Table 5 Eigenvalue..............................................................................................................
Table 6 PC Data..................................................................................................................
Table 21 Actual Data vs Predicted Data PCA and ANFIS Models .................................
Table 23 Data Actual Vs prediction PCA and BGA with Model ANFIS.........................
(House 1).........................................................................................................
Figure 1 Data flow diagram of the PCA and BGA-based Energy consumption
prediction method..............................................................................................
Figure 2 Conceptual model for the effectiveness of using PCA for the feature
Figure 24 ANN model network during training in predicting the use of electrical
Figure 26 The MSE model of ANN in training predicts the use of electrical energy.......
Figure 27 The correlation coefficient of the ANN model in training predicts the use
of electrical energy..........................................................................................
Figure 28 The correlation coefficient of 2 ANN models in training predicts the use
of electrical energy..........................................................................................
Figure 29 The ANN model network at the time of testing in predicting the use of
Figure 30 The MSE model of ANN in testing predicts the use of electrical energy.........
Figure 31 The correlation coefficient of the ANN model in testing predicts the use of
electrical energy...............................................................................................
Figure 32 The correlation coefficient of 2 ANN models in testing predicts the use of
electrical energy...............................................................................................
Figure 33 Training state model ANN in testing in predicting electricity usage ...............
Figure 35 The correlation coefficient of the ANFIS model in training predicts the
Figure 38 The correlation coefficient of the 2 ANFIS models in testing predicts the
Figure 40 ANN model network with PCA during training in predicting the use of
Figure 41 ANN training state model with PCA in training in predicting electricity
consumption ...................................................................................................
Figure 42 The correlation coefficient of the PCA model and the ANN model in
Figure 43 The correlation coefficient of the PCA model and the ANN model in
Figure 44 The PCA correlation coefficient with the ANN model in testing predicts
Figure 45 The correlation coefficient of 2 ANN models in testing predicts the use of
Figure 47 The correlation coefficient of the ANFIS model in training predicts the
Figure 48 The correlation coefficient of the 2 ANFIS models in testing predicts the
electricity consumption...................................................................................
Figure 50 The PCA correlation coefficient with the ANFIS model in testing predicts
electricity consumption...................................................................................
Figure 52 Correlation coefficient of PCA and BGA model of ANN training data
Figure 53 Correlation coefficient of PCA and BGA model of ANN training data
Figure 54 The correlation coefficient of PCA and BGA models of ANN testing data
Figure 55 The correlation coefficient of PCA and BGA models of ANN testing data
Figure 56 The correlation coefficient of combining PCA and BGA feature selection
Figure 57 The correlation coefficient of 2 PCA and BGA ANFIS models in training
Figure 58 Correlation coefficient combining PCA and BGA feature selection of the
Figure 60 Prediction energiy actual Vs prediction PCA and BGA with Model AN-
FIS...................................................................................................................
ANN,ANFIS,PCA+ANN,PCA+ANFIS..........................................................
testing time......................................................................................................
I Identity matrix
X original variable
U eigenvector matrix
BGA - Binary Genetic Algorithm
Np - Population size
i - Number og chormosome
f(x) - Fitness function
vi - Fitness value eval
pi - Perents
Pc - Probability crossover
Pm - Mutation probability
Desmira,D, Bakar, N. A., Hashim, M. R., Wiryadinata, R., & Hamid, M. A. (2022).
Laboratory prediction energy control system based on artificial intelligence network.
Bulletin of Electrical Engineering and Informatics, 11(3), 1280–1288.
https://doi.org/10.11591/eei.v11i3.1821 ISI indexed, Q3, IF = 2.4 (2022)
Desmira, D, Hamid, M. A., Bakar, N. A., Nurtanto, M., & Sunardi, S. (2022). Smart Traffic
Lights Using Microcontroller Based on Fuzzy Logic. IAES International Journal of
Artificial Intelligence, 11(2), 809–818. https://doi.org/10.11591/ijai.v11.i3.pp809-818
ISI indexed, Q2, IF = 2.3 (2022)
Desmira,D,Bakar, N. A., Wiryadinata, R., Hamid, M. A., & Kholifah, N. (2022).
Comparison of principal component analysis and ANFIS to improve EEVE Laboratory
energy use prediction performance. 27(2), 970–979.
https://doi.org/10.11591/ijeecs.v27.i2.pp970-979 ISI indexed, Q3, IF = 2.4 (2022)
INTRODUCTION
1.1 Background
The world’s energy crisis is getting more and more so that it is necessary to hold something
rare to make efficient use of that energy. Various efforts have been made by the Indonesian
government to save energy in all fields. An initiative is to use energy effectively in a build-
strategy is required. In this study, a feature selection technique was used to select significant
input variables for managing energy usage and optimizing processor memory.
2. Reducing the amount of time spent on computing during the training and testing processes.
The features selected in this study were principal component analysis (PCA) and binary ge-
netic algorithm (BGA). In this study, two stages of the feature selection method are intro-
duced, where feature selection using principal component analysis (PCA) is used as the first
stage selection of input variables, where from nine input variables will be reduced to only
four significant input variables. The binary genetic algorithm (BGA) will be used as the sec-
ond stage of the feature selection method to further reduce the number of significant input
variables to only three. The focus of this study is to minimize the computational time re-
When applying machine learning to predict energy consumption, particularly within large
datasets, a complex web of challenges surfaces that significantly impact computation time
and efficiency.
1. Computation Time and Efficiency: The handling of large datasets in machine learning
prolongs computation time. Selecting appropriate features that optimize input variables is
crucial for speeding up this process, especially for predicting electricity consumption (Y.
transfer rates without proper feature selection (Paudel et al., 2017), and the necessity for
selection steps may hinder computational time efficiency in predictions (Philippe Leray
et al, 2021).
room energy use. These include problems in data grouping techniques, particularly in
selecting PCA features (Qu & Chen, 2020); (Shapi et al., 2021), difficulties in achieving
faster computation time without combining relevant feature selection (Y. Zhou et al.,
2020), and the challenges in applying BGA feature selection to increase computation
5. Solutions and Future Prospects: The proposed resolution emphasizes a strategic com-
bination of PCA and BGA selection methods, focusing on improving computational effi-
ciency within energy consumption prediction models for ANN and ANFIS. Despite these
efforts, feature selection still presents accuracy issues that require further exploration and
development.
The formulation of this problem stems from the inherent complexity of machine learning
including computation time, data handling, and feature selection, revealing the intricate na-
ture of balancing efficiency, accuracy, and technological advancement. The above challenges
highlight the necessity for innovative and targeted solutions, particularly those that focus on
feature selection methods that accommodate the unique needs of large-scale data processing
ing further insights and refinements to enhance the speed and accuracy of energy predictions
in various contexts.
This study aims to confront the challenges and limitations associated with machine learning
in predicting energy consumption. This research seeks to explore innovative solutions and
Objective: To improve and enhance existing principal component analysis (PCA) methods
computation time and data grouping, aligning with the first and second problem statements.
Objective: To develop new data grouping strategies and integrate PCA-based feature
earlier, focusing on overcoming the limitations of existing feature selection and the
Expected Outcome: Increased computing speed for predicting indoor energy consumption,
Objective: To evaluate the effectiveness and efficiency of the combined PCA and BGA
of improvement in the combined PCA and BGA methodology, offering insights for future
in summary, these research objectives develop, innovate, and critically evaluate methods that
reduce input variables and accelerate computation time. The target models, ANN and ANFIS,
will be used to predict energy consumption within the UNTIRTA PVTE laboratory. By
meeting these objectives, this research will contribute to the field of machine learning for
energy consumption prediction by enhancing both efficiency and increased compute time and
2. Feature selection will be used to predict the energy consumption of a room, especially in
3.This feature selection development will use 3350 data, with 2450 training data, 833 test
The data will be collected based on real retrieval in the UNTIRTA PVTE laboratory for three
months. The first retrieval was performed from January 6, 2021 to April 10, 2021. Data
retrieval was based on data read by sensors connected to the microcontroller using a data
recording device comprising several supporting tools for data recording. It also requires a
PC/laptop where the data will be recorded and displayed on the Excel PLX-DAQ software.
Parameter variables such as pressure, current, voltage, power, light, humidity, temperature,
and altitude will be recorded. Recording will be conducted on weekdays and operating hours
because the laboratory room is only active on weekdays and operating hours. Starting
Monday– Friday, 09.00 to 17.00 pm. Electrical energy management for energy efficiency in
predicting energy use in the laboratory focuses on selecting the features used when managing
energy in one or more buildings, especially in this research conducted in the UNTIRTA
4. The data was also tested with data on the UCI Machine Learning Repository website with
PERACON data with 10 houses, and then the data was tested with one house totaling 525600
data where the data includes energy consumption from 6/1/2018 12:00:00 AM to 5/31/2019
11:59:00PM
ii) Application of the process looping to PCA and BGA feature selection can increase speed
iii) Training and testing data based on input variables from 9 real inputs taken at the PVTE
UNTIRTA laboratory. The calculation of energy consumption was compared with actual and
iv) Experienced a 15% increase over the BGA and PCA studies.
Based on the objectives previously presented and on the approach proposed before, this thesis
is made up of five (5) chapters, the contents of which are summarized as follows:
research.
laboratory, and how to manage energy in a room. Short formula for energy
used today.
predict energy in a room, especially in the laboratory, using the PCA and
Chapter 4. Case studies. In this chapter, the model is tested using a model
before and after feature selection with real data taken for 3 months in the
LITERATURE REVIEW
2.1 Introduction
Based on International Energy Agency (IEA) projections for the period 2006-2030,
the world’s energy demand mostly comes from non-Organisation for Economic Co-operation
and Development (OECD) countries, which is 87%. [16][17]. China’s energy demand growth
is projected to be the largest among other regions [18][19]. India recently also showed that
growth in energy demand is quite large and just one level below that of China. In Indonesia
itself, various efforts have been made to reduce energy use in rooms or universities. Various
efforts have also been made to achieve efficient and stable energy use. Energy use to achieve
energy efficiency needs to be achieved by developing a smart room with various sensor in-
puts that affect consumption in the room [20] . Energy consumption is the demand used to
indicate the amount of energy that must be supplied to a building at any given time. Exces-
sive use of electrical energy will increase energy waste and have a negative impact on the
ing energy, which benefits both individuals and society. Predicting energy consumption in
buildings is essential for improving energy performance, with the aim of achieving energy
A building that produces the largest energy consumption, usually 40% of the energy
used. Increasing efficiency is a good effort to reduce the negative caused from an economic
viewpoint. Another impact of efficiency is the reduction of fossil fuels and the stabilization of
increase the effectiveness and efficiency of decisions made to reduce energy demand Predic-
tion Energy consumption models are mostly used in the first step in energy saving and effi-
ciency improvement models and reduce costs [23] thermal energy storage adhesive to im-
prove energy efficiency [24]. The input variables in a model are very decisive in the perfor-
mance of the network model that will be used in making predictions that will be faster and
more cost-effective [2]. A technique used to reduce data from larger data to smaller ones is
This study was developed to improve energy management in predicting energy con-
climate[33] Guang Dong University has conducted research with the results that the construc-
was found that it was difficult to measure energy consumption indicators [34]. University
buildings in China develop target energy efficiency strategies according to local conditions,
with multiple campuses and different climates being one of the factors of difficulty in imple-
menting energy efficiency in buildings [11]. Historical data on daily electricity usage in two
London South Bank University buildings were regressed over normalized data from six input
variables to obtain efficient energy use in a building [35]. A case study of three campus
buildings in Tianjin predicts that the average electricity consumption of one occupant varies
depending on the function of the building and the control mode of electrical equipment [36].
The development of statistical regression models is used to understand the relationship be-
tween each variable and energy consumption [37]. In the previous research method, the im-
In a building, among the input parameters that affect energy use in a building is the
behavior of room users, which greatly affects the use of the room, and it is difficult to deter -
mine how much energy consumption is needed. [39]. In the previous energy consumption
prediction model [40][41][42][43][44][45], features are directly used as inputs in the predic-
tion stage. Feature selection is performed with the aim of selecting features that most affect
energy consumption [46]. With feature selection, it is expected that the prediction stage can
be performed more effectively and increase accuracy. The purpose of the feature selection
method is to obtain a reduced set by removing some features that are considered irrelevant for
text sentiment classification to increase classification accuracy and decrease the duration of
the machine learning algorithm [47]. However, the disadvantage of attribute selection is that
it requires training a large dataset to obtain reliable transformations [48]. One technique to
overcome the high dimension of features is feature selection. Various feature selection tech-
niques [49], such as information gain [50], have been used to reduce vector dimensions.
In addition to feature selection techniques, another technique that can be used to over-
come the problem of high feature dimensions is dimensionality reduction. The purpose of the
dimension reduction technique is to obtain a new data representation that is managed into a
lower dimension [51]. The linear dimension reduction algorithm consists of a singular value
decomposition (SVD) algorithm and a principal component analysis algorithm [52]. How-
ever, the weakness of the linear dimensionality reduction algorithm is that it produces a linear
combination of all features that may be contaminated by noise and reduces the performance
of the classification algorithm. The linear dimensionality reduction algorithm will have diffi-
tion include pressure, temperature (DHT22), current, voltage, power, light intensity, tempera-
ture (BMP180), altitude, and humidity. The output of the prediction model is the energy con-
sumption. A flowchart of the energy consumption prediction model using PCA as the feature
selection method is shown in Fig. 1 following. Experiments have shown that both the preci-
sion and durability of the optimized GA-BP ANN are better [54]. The optimal PI controller
tuning program developed on the genetic algorithm was developed and implemented in a
non-linear HVAC system simulated by the HVACSIM + package [55] . Simple and sophisti-
cated controllers were used in this study: ON– OFF, PID, and fuzzy controllers. The opti-
mized ANN validation demonstrated good predictive accuracy, as the regression coefficients
R2 for consumption and discomfort were greater than 0.77 and 0.84 for the three tested con-
trollers, respectively. Various distance predictions and ANN amounts of training data were
tested [56] . Feasibility and application of ANFIS in building energy load forecasting areas
with 96.5 and 83.8% for heating and cooling energy predictions [57]. An ideal module
through Model 4, this model is extended to predict the thermal performance of real homes,
thereby providing a fast and effective prediction tool. [58]. Electricity consumption in Thai -
land from 1986 to 2010 results show that the ANN model reduces the mean absolute percent-
age error (MAPE) to 0.996%, while ARIMA and MLR are 2.80981 and 3.2604527%, respec-
tively.[59]. The electricity demand is based on two soft computing techniques, and the exper-
imental ARIMA model reveals that EFuNN performs better [60] In this study, excellent re-
sults were obtained using the electrical load forecasting method on buildings based on ANN
feedback [61]. The ANN and ANFIS techniques have much better performance than the em-
pirical formula (for the dataset test, ANN R2 = 0.97, ANFIS R2 = 0.92 and Marciano R2 =
0.54). Between ANN and ANFIS, the ANN model is slightly better although the difference is
small [62]. To improve the overall performance of the heating system in terms of energy effi -
[63]. In this study, the GA-ANFIS method also obtained a better modeling time than ANN
[64]. Research data by comparison with actual consumption in the fifth year. The results
show that ANN is 6% better, whereas MR, GP, SVM, and DNN have MAPEs of 8.5%, 8.7%,
9%, and 11%, respectively [65]. For electricity use, MAPE has a set number of validations of
34.0%, a reduction of 31.8% against a theoretical benchmark; to heat the fuel using MAPE
an ANN as a reliable design tool [66] The GA-based ANN model has a mean square error
equal to 0.990 and 0.002 for each test dataset in predicting the bearing capacity of the pile.
[67] . Among several studies related to energy consumption using fuzzy logic [68][69][70]
[71].
For that, how to manage electrical energy for more efficiency in predicting energy in
the electrical engineering vocational education (PVTE) laboratory. With several learning al-
gorithms simulated with MATLAB R2018a software, three types of input parameters (9 in-
tude, and humidity. In another study, Kim, Jung, and Kang [80] used residential energy con-
and user features in South Korea. Their study discusses the nation’s influencing parameters,
namely, the number of exterior walls, direction of housing, housing area, number of years
occupied, number of household members, and occupation of the head of the household on
energy estimates. The results show the high accuracy of the ANN model in predicting energy
consumption. In a similar study by Bui, Nguyen, Ngo, and Nguyen-Xuan [81], building en-
ergy consumption, including heating and cooling loads, was estimated using the hybridization
of the ANN model with EFA. For building energy prediction, selecting inputs and grid struc-
tures that may be relevant is important to ensure generalization performance. Rivals and Per-
sonnaz [82] proposed a systematic approach based on least squares estimation and statistical
tests. This method obtained a simple neural network on the Internet for the application of an
energy prediction contest [83], which was used in this study for performance comparison
Figure 2 Data flow diagram of the PCA and BGA-based Energy consumption
prediction method
Feature selection that will be used in energy management for efficiency in predicting energy
consumption in the Untirta PVTE laboratory room is to use feature selection with PCA. The
input variables in predicting energy consumption in the laboratory room are obtained from
the activities that occur, which are taken using real sensors, which include pressure sensors,
altitude , and humidity. The data are processed using PCA to obtain a linear combination of
the initial variables. Geometrically, this linear combination is a new coordinate system
obtained from the rotation of the original system. The PCA method is very useful if the
existing data has many variables and a correlation between the variables. Principal
representing the distribution of data from a dataset. The purpose of PCA is to select the
existing variables without losing the information contained in the original data. Using PCA,
the previous n variables are selected to become k new variables called principal components,
with the amount of k being less than n. Using only k principal components will produce the
same value using n variables. The variable resulting from the selection is called the principal
component. the new variables formed by PCA analysis later, in addition to having a smaller
number of variables and eliminating the correlation between the variables formed.
transforms an original data matrix into a set of linear combinations that are fewer but absorb
most of the variance from the initial data. The main goal is to explain as much as possible the
number of variants of the original data with as few principal components as possible. The
principal component is a new set of variables that is a linear combination of the observed
variables. The principal component has a decreasing variance; most of the variation/diversity
of information in the observed set of variables tends to be collected in the first few principal
component. This means that the principal component in the last sequence can be ignored
without losing much information. PCA can be used to select the variables being observed.
Principal components are orthogonal, which means that each principal component represents
all the original variables so that the principal component can be used as a substitute for the
original variable. energy management for efficiency in predicting energy consumption in the
Untirta PVTE laboratory room to reduce the dimensions of the data through a dataset
Figure 3 Conceptual model for the effectiveness of using PCA for the feature selection
stage [84]
PCA is used to explain the structure of the variance– covariance matrix of a set of variables
through a linear combination of these variables. In general, principal components are useful
for feature selection and the interpretation of variables. Principal component analysis is
1. Identify the new variable value that underlies the multiple variable data
many correlated variable values by maintaining as much diversity as possible in the dataset
3. Eliminate the value of the original variable which has a relatively small contribution of
information.
Consider p variables consisting of n objects. From the p variables, there are k principal
components (with k <= p), which is a linear combination of these p variables. The k principal
components can replace the p variables without losing much information about the entire
variable. Generally, PCA is an intermediate analysis, which means that the results of the
The mathematical form that Y is a linear combination of variables x1, x2, .., xp can be
P
Zm=∑ ( wi∗xi ) ............................................................................................... (1)
I
Where :
xi : variable to –i
In Principal Component Analysis the method to get the coefficient values or weights from the
each other.
c. Principal components are formed based on the order of variance from the largest to the
smallest. According to Johnson, if we get a vector x '= [x1, x2,..,xp] which has a covariance
follows :
where :
Yp= component to – i
a = eigenvector value
The i component, Yp, is a linear combination of x1, x2,...,xp with weights, namely: ap1x1,…
ap2x2…,apixi, the selection must maximize the ratio of the first component variance (PC1) to
a weighted linear combination of all the original variables, uncorrelated with the first
component (PC1), and must absorb the maximum remaining variance. To obtain the principal
observed variables. In this case, the eigenvalue is the variance of each principal component,
According to Jolliff, the Principal component analysis procedure for dimension reduction of
the random variable x (matrix of size n x n, where the rows contain n observations of the
1. Calculating the covariance matrix from the observation data The covariance matrix is
a matrix in which the covariance values in each cell are obtained from the sample. Let
n
1
Cov ( A 1 , A 2 )= ∑ ( x −μ ) ( y −μ y ).................................................
n−1 i=1 i x i
(3)
With μ x and μ y being the average (mean) of the samples of the variables x and y,
where x i and y are the values of the i observation of the variables x and y. From the value data
Looking for eigenvalues and eigenvectors from the covariance matrix that has been obtained
[87] namely: Eigenvalues (λ) are scalar numbers and A is a matrix with size n x n to obtain
the value of n eigenvalues (λ1, λ2……..λn) then fulfill the following equation: namely:
Determinant (A – I) = 0
A = matrix nxn
λ = eigenvalue
I = identity matrix is a square matrix with the main diagonal element is 1, while the other
The eigenvalues and eigenvector define matrix A. The equation used to calculate the
Ax = λx ......................................................................................................... (4)
Ax – λx = 0
(A – λ) x = 0
(A – λI) x = 0, x ≠ 0
λ = eigenvalue
x = non-zero matrix
Eigenvalue & eigenvector equations are Eigen Value Decomposition (EVD), with the
following equation:
AX = XD ............................................................................................... (5)
A= X D X-1
3. Determine the new variable (principal component) by multiplying the original variable by
Zki = n x n matrix of the principal component with the coordinates of the i object at the k-th
m
c j=∑ |v pj|............................................................................................... (7)
p=1
Choose the largest number of c j values according to the number of features you want
Eigenvalue
Value of Proportionof Principal Component ( % )= x 100 %. (9)
Covariance variance
A number of selected features are then used as inputs in the energy consumption pre -
diction stage. The prediction of energy consumption is performed using two different algo-
rithms, namely the Artificial Neural Network (ANN) algorithm and the Adaptive Neuro
2.3 Prediction of Energy Consumption Using the Feature Selection Binary Genetic
Algorithm (BGA)
In addition to PCA, the feature selection used by BGA is the input and output of the
resulting feature selection in the form of binary numbers. The purpose of feature selection is
to obtain data that are initially large and are reduced to smaller ones [88][89][90].
determining the best solution. Genetic algorithms were first introduced by John H. Holland in
his book "Adaptation in Natural and Artificial Systems" in 1978 and Goldberg David E in
based on the evolutionary mechanism of living organisms using binary numbers. The
principle of evolution is as follows: a problem is determined before the candidate for its
which is a set of genes. The gene is a component of the solution to the problem. This group of
which is known as the attractiveness value. Based on these values, there will be mutual
attraction between individuals (with a certain process, for example the roulette wheel
technique to choose a partner who will marry) so that the marriage process occurs. In this
marriage, there will be a process of gene exchange between the mating individuals. There are
two main operators in this process: crossover and mutation. From the mating process
produced offspring (spring). The children are then mated with the same steps as above, which
The results of the binary genetic algorithm technique are not global optimum (ie
optimum for all search areas) but are not easily trapped in local optimum (ie optimum for
only certain areas). This technique is suitable for complex problems, and there is no
optimization technique. Binary genetic algorithm (BGA) is information that we can change,
which affects the behavior in the system, fitness, and the genes in the system. Genetics of
in a search problem. Genetic algorithms are a special class of evolutionary algorithms that
use techniques inspired by evolutionary biology, such as heredity, mutation, natural selection,
and recombination. Genetic algorithms apply a method for moving from one population
the genetic algorithm proposed by John Holland is used to express alternative solutions to a
problem. Each chromosome consists of a series of bit strings consisting of 0 and 1 bits called
genes, although it is also possible to use different chromosome coding [94] . Each
chromosome can undergo the exchange of genetic material between chromosomes. The
mutation process randomly changes gene values at several chromosome locations . Genetic
algorithms are different from conventional search techniques, in which the condition begins
with the initial setting of a random solution called a population. Each individual in the
is a string of symbols, which is generally a bit string of 0s and 1[95]. Chromosomes evolve
chromosomes, called offspring, are formed either by merging two chromosomes from the
initial generation using the crossover operator or by modifying the chromosomes using the
mutation operator. A new generation is formed through the selection process of several
parents (parents) and offspring (offspring), according to the fitness value, and through the
elimination of other chromosomes so that the population size remains constant. After several
generations, the algorithm can generate the best chromosomes that are expected to represent
the optimal or suboptimal solution to the problem. In general, the start process is assumed to
Binary Genetics Algorithm is almost the same process in doing the process to get the
most optimal value in a problem[95]. The steps are 1. Represent the problem in the form of
bit strings 0 and 1 as chromosomes that have a fixed length, and determine Pc (Crossover
performance, which establishes the basis for selecting chromosomes to mate during
reproduction. Look for the fitness value, which is a measure of whether or not an individual is
good or whether the solution obtained is good or not. 3. Determine the initial population
generation process. This is usually done using random generation.4. Calculate the fitness
Parental chromosomes are selected with a high probability, and very matched chromosomes
have a high probability of being combined with less matched chromosomes. 6. Creating a
Creating offspring chromosomes (offspring) in the new population 8. Repeat step 5 until the
size of the new population of chromosomes is equal to that of the initial population. 9.
Replacing the parent chromosome population with a new offspring population. 10. Continue
step 4 and repeat the process until the desired criteria are reached. The cycle of the genetic
The first stage of the binary genetic algorithm is the start of the population, which
is described by encoding both binary and string chromosomes. The next step is to calculate
the fitness value based on the genes on the chromosomes in each population. Then, the
selection stage selects the chromosomes that are selected as parents. Furthermore, the
selected parent will undergo cross-breeding (Crossover). The crossover process plays a role
in forming offspring chromosomes, which also plays an important role in increasing the
diversity of strings in the population. Furthermore, the chromosomes enter the mutation
stage, which ensures that the diversity of chromosomes in a population is maintained to avoid
premature convergence.
Start
The genetic algorithm has characteristics that need to be known so that it can be
distinguished from other search procedures, namely:[92] 1. Coding scheme for the problem
solution based on the parameter set. Finding solutions from many individuals not just from
one individual 3. Genetic algorithms are based on the objective (fitness)of evaluating
individuals who have the best solution, not derivatives of a function. Binary genetic
The first step in implementing a computer program is the selection of data types. This
is a variation of the initial differences between Holland’s original theory and several
alternative theories of genetic algorithms that later emerged. Holland decodes the
chromosomes into a string of binary digits. A number of binary encoding operations are used
to generate simple genetic algorithms; however, there are many ways to represent an object’s
genes, each of which has its own implicit advantages. To represent a problem in the form of
value encoding, and tree encoding. The encoding process discussed in this study is value
values. Values can be integers, real numbers, characters, or objects. The chromosomal coding
technique is shown in Figure 5 [100]. The process begins by initializing several individuals or
size: defines how many chromosomes and how many genes in one chromosome are involved
during the search process. Crossover Probability: specifies the crossover probability between
two chromosomes. Mutation probability: specifies the probability that a bitwise mutation will
occur. Termination criteria: specify the conditions for ending the search for a solution in the
genetic algorithm.
After the chromosome coding process is performed in the genetic algorithm, the next
stage is the selection of offspring for the next generation. The purpose of the selection
process is to select the best chromosomes (individuals) in the population with the hope that
the new offspring will have a good fitness value.[93] . The selection process used in this
cuts the number of chromosomes in the population so that it can reduce the number of
individuals to be selected. This can speed up the process of selecting prospective parents.
Only the best individuals will be selected as parents. The parameter used is a threshold value
trunc, which indicates the size of the population to be selected as the parent, which ranges
from 50% to 10%. Individuals below this threshold will not produce offspring [102]. The
truncation method is shown in Figure 6. First, we increased the number of individuals in the
chromosomes with the best fitness values. 4. After going through the truncation process, the
individuals with the best fitness were obtained, namely, K4, K2, and K3.
The equation of the fitness function f(x) can be determined according to the equation:
Popsize
F= ∑ f (x ) .......................................................................................... (8)
i=1
To get how much influence Pi has on the fitness function f(x) by using the formula: To get
how much influence Pi has on the fitness function f(x) by using the formula:
f (x )i
Pi= Np
.................................................................................. (9)
∑ f (x)i
i=1
2.3.5. Crossover
Crossover is a genetic operator that combines two chromosomes (parent) to produce a new
chromosome (offspring) with probability crossover (Pc). In this study, the crossover method
used was a one-point crossover [104] [105]. In genetic algorithms, there are several types of
crossover among others. One-point crossover (One Point) This operator determines the
crossover point gene from the first parent chromosome plus genes from the crossover point to
the last gene from the second parent chromosome[67]. The second new chromosome contains
the first gene until the crossover point gene from the second parent is added to the gene from
the crossover point until the gene from the first parent chromosome [106]. The crossover
Parent 2: 4 5 3 6 8 | 9 7 2 1
After the crossover process, the derivatives that can be generated from the two
Parent 1: 1 2 3 4 5 | 9 7 2 1
Parent 2: 4 5 3 6 8 | 6 7 8 9
Two-Point Crossover The process of two-point crossover is almost the same as that
of the one-point crossover procedure, except that in two-point crossover, two crossover points
must be selected and only genes between the two crossover points will be exchanged. [107].
Parent 1: 1 2 3 4 | 5 6 | 7 8 9
Parent 2: 4 5 3 6 | 8 9 | 7 2 1
After the crossover process, the derivatives that can be generated from the two parents above
are:
Parent 1: 1 2 3 4 | 8 9 | 7 8 9
Parent 2: 4 5 3 6 | 5 6 | 7 2 1
Parent 1 : 4 8 7| 3 6 5| 1 10 9 2
Parent 2 : 3 1 4| 2 7 9| 10 8 6 5
In (3,2), (6,7), (5,9) are swap places with mappings Parent 1 and Parent 2 where
swaps are (2,3), (7,6), (9.5). From the crossover results obtained offstring as follows.:
Child 1 : 4 8 6| 2 7 9| 1 10 5 3
Child 2 : 2 1 4| 3 6 5| 10 8 7 9
Uniform crossover operators Single- and two-point crossovers must determine the
crossover point before performing the crossover operation. Uniform crossover extends the
this, binary strings of the same length as the chromosomes are first constructed randomly. If
any bit in this binary string has the value 1, the genes from both parents need to be exchanged
at the same position. After the exchange process, two children were created [108].
Shuffle crossover operator When shuffle crossover is applied, the genes of the two
chromosomes are randomly shuffled. Then, one-point crossover is applied to the randomized
chromosomes after selecting the crossover points at random, and two children are generated.
After recombination, the offspring’s genes are then shuffled in the same way that they were
Reducing the surrogate crossover operator Reducing the surrogate crossover limits the
crossover operation in cases where the parents have the same genes, i.e., the crossover point
can only be the point where the genes of the two parents differ. One position is chosen at
random after all points are found. Finally, a single point crossover is performed at this point
2.3.6. Mutation
mutation probability (Pm) (mutation rate or mutation probability), leading to genetic diversity
that helps escape the trap of local optimal search processes[67]. There are several mutation
methods in the genetic algorithm as follows. Mutation Swap changes 2 (two) positions of n
randomly selected elements from permutations, and other elements remain in their respective
gene with a small random number. 3. Uniform mutation: This mutation works by replacing
the value of the selected gene for the mutation with a random number that has a limit. The
limit is determined such that the resulting solution is legal. The uniform mutation process is
shown in Figure 10. Random mutation is performed by increasing or decreasing the value of
the selected gene with a small random number. Uniform mutation: This mutation works by
replacing the value of the gene selected for the mutation with a random number that has a
limit. The limit is determined such that the resulting solution is legal. The uniform mutation
selection
K. K. W. regression consumption
ANN
Salo,dkk security
(SOM) dan
algoritma k-
Means
PA)
SVR,Regression
tree,Random
forest,KNN
2.4. Energy Consumption Prediction Using Artificial Neural Network (ANN) Model
The ANN model maps the input data (feature selection results) at the input layer
to the output layer through the neurons in the hidden layer. An example of the ANN
The input data are propagated forward and connected by input weights that have previously
been randomly initialized to the neurons in the hidden layer. In the hidden layer, the input
data associated with the weights are then processed using the activation function.
Furthermore, the processed data from the hidden layer are connected by hidden weights to the
The results obtained are then compared with the target data to obtain the error rate. If the
error rate obtained is smaller than the previously set error rate (target error), the propagation
process will stop. However, if the error rate is greater than the constant error rate, a back
The following is the step of the backpropagation neural network algorithm. The input data are
propagated forward and connected by input weights that have previously been randomly
initialized to the neurons in the hidden layer. In the hidden layer, the input data associated
data from the hidden layer are connected by hidden weights to the neurons in the output
layer.
The results obtained are then compared with the target data to obtain the error rate. If the
error rate obtained is smaller than the previously set error rate (target error), the propagation
process will stop. However, if the error rate is greater than the constant error rate, a back
2.4.1 Initialize
Initialize all weights on the hidden and output layers and define the activation function in
each layer. Set the learning rate. Initialization of all weights can use random numbers in a
small range:
2.4.2 Activation
y d 1 ( p ) , y d 2 ( p ) , … , y dn ( p).
a. Calculate the output obtained from the neurons in the hidden layer
n
v j ( p )=∑ x i ( p ) . w ij ( p).............................................................................................. (8)
i=1
1
y j ( p )= − v (p ) ............................................................................................................ (9)
1+e j
b. Calculate the output obtained from the neurons in the output layer:
m
v k ( p ) =∑ x j ( p ) . w jk ( p)............................................................................................ (10)
j =1
1
y k ( p )= −v ( p ) ........................................................................................................ (11)
1+e k
The weights are updated when the error is propagated back in the ANN, and the error is
returned according to the direction of the output signal.a. Calculate the error gradient for the
e k ( p )= y dk ( p )− y k ( p)................................................................................................ ( 12 )
δ k ( p )= y k ( p ) ×[1− y k ( p ) ]× ek ( p).......................................................................... ( 13 )
b. Calculate the error gradient for the neurons in the hidden layer::
2.4.4 Iteration
Increase one for iteration p, return to step 2, and repeat the process until the error
using the best weights and biases from the network training stages. The resulting energy
consumption is compared with the actual energy consumption to obtain the accuracy of the
2.5. Prediction of Energy Consumption Using the Adaptive Neuro Fuzzy Inference
System (ANFIS) Model
Feature selection results are also used as input in the neuro-fuzzy algorithm. This
algorithm is a combination of two systems: a fuzzy logic system and an artificial neural
network. The neuro-fuzzy system is based on a fuzzy inference system that is trained using a
learning algorithm derived from an artificial neural network system. Thus, the neuro-fuzzy
system has all the advantages of fuzzy inference and artificial neural networks. From their
ability to learn, neuro-fuzzy systems are often referred to as adaptive neuro-fuzzy inference
systems. One form of structure that is well known is shown in Figure 11.
In this structure, the fuzzy inference system is the Takagi–Sugeno–Kang fuzzy inference
model
In the neuro-fuzzy system, there are five layers of processes in which the functions and
Let O1 ,i be the output of each node in layer 1. Each node i in this layer is an adaptive node
with the function node O1 ,i=μ A i( x)) for i=1 , 2; or O1 ,i=μ Bi ( y )) for i=1,2, where x is the
input to nodei and Ai is the linguistic label (small, large, etc.) that corresponds to the
function of this node. On the other hand O1 ,i is a membership function of A 1 and the degree
of membership is specific for a given x which is sufficient to quantify Ai . The most widely
used membership functions are the Bell and Gaussian forms. The Bell form membership
(19)
with parameter b usually positive. Parameter c is located in the middle of the curve. The
Gaussian membership function is expressed by, with parameter b usually being positive.
Parameter c is located in the middle of the curve. The Gaussian membership function is
expressed as follows:
2
( x−c )
A ( x )=e 2a
2
.................................................................................................................... (20)
with parameter b usually positive. Parameter c is located in the middle of the curve. The
Each node in this layer consists of the prod t-norm operator as a node function. This layer
synthesizes the transmission of information using layer 1, multiplies all incoming signals, and
sends the product out. The output of the product layer is expressed as follows:
Each node in this layer serves as a measure of the strength of the rule. The output in this layer
Each node in this layer normalizes the weight function obtained from the previous product
wi
O3 ,i = ………………………………….…………………………………….. (22)
w1 +w 2
Nodes in this layer are naturally adaptive. The defuzzification output of this layer is
A single node in this layer synthesizes the information transmitted with layer 4 and returns
Σ wi y i
O5 ,i = ………………………………………………………………………… (24)
Σ
Similar to the calculation of energy consumption using the ANN algorithm, the ANFIS
algorithm performs forward propagation using the best weights and biases from the network
training stages. The resulting energy consumption is compared with the actual energy
From the references that have been discussed in chapter 2, it can be concluded that the
research steps.
The data used as a reference in this study were obtained from research conducted for
approximately three months, namely January– April 2021. Data In this study, we used 3350
with 2450 training data, 833 test data, and 67 predictive data. Data are collected on the basis
of recorded data using a microcontroller with a sensor input with 9 input variables. Pressure,
temperature I, current, voltage, power, light, temperature II, altitude, and humidity.
Parameter Input
PCA feature
selection
BGA feature
selection
Stage 1
Stage 2
Prediction of energy
Prediction of energy
consumption using
consumption using
the ANFIS
the ANN algorithm
algorithm
Before processing the data obtained during the research in the electrical engineering
vocational education laboratory (PVTE). The main purpose of feature selection here is that
the reduction function reduces the number of variables that were initially vast to fewer to
facilitate analysis at a later stage. The transformation function is used to change the variables
Stage 1
After obtaining the looping process, feature selection is processed using artificial intelligence
Stage 2
This stage is almost the same as stage 1, the difference being that the model used is an ANN.
• Application of the looping process to the selection of features that combine PCA and BGA
PCA+BGA+ANFIS models.
A number of selected features are then used as inputs in the energy consumption prediction
stage. The prediction of energy consumption is performed using two different algorithms,
namely the Artificial Neural Network (ANN) algorithm and the Adaptive Neuro Fuzzy
From Figure 14, it can be seen how the steps that will be taken when the research focuses
on grouping data on combining PCA and BGA features selection. The purpose of grouping
room using machine learning (ML). The grouping of data on the PCA feature selections a
but the looping process in BGA with data grouping will determine the value of 1 and if BGA
has a logic 0 then the data will stop when logic 1 is found. After the data have stopped when
the looping process is optimal, the data are tested with the ANN and ANFIS models.
The BGA algorithm is a binary genetics algorithm. The input and output are binary
numbers used to find the most optimal solution to an optimization problem. In the
and BGA features selection by grouping the data before the data is trained in the training and
testing process. The initial stage of this research is to select the input variable by grouping the
data based on the analysis component formed from the input variable as many as nine
variables. The data of the nine variables is reduced to PC1,PC2……..PC9, then performs the
transformation by selecting the PC value that value of one which will be chosen to influence
the prediction results using the ANN and ANFIS models according to the block diagram in
Figure 15.
PCA feature
selection
1. Pressure
2. Temperature
(DHT22) Dimension reduction
3. Electric current, TransformationP
PC1,PC2,PC3,PC3,PC
4. Voltage C1,PC2,PC3,PC
4,PC5,PC6,PC7,PC8,P
5. Power 4
C9
6. Luminous intensity
7. Temperature (BMP
180)
8. Altitude BGA feature
9. Humidity ANN Model and selection
ANFIS Model 1011
model that I will propose by looking at the characteristics of the PCA model which has a lack
of data similarity level by choosing a number that has a value of 1 and a number below 0 is
not continued to the next process with experiments that have been carried out, it can be
concluded that only PC1, PC2, PC3, and PC4 are selected to enter the second feature
selection stage using BGA. `Only 2 components will be formed, namely PC1 and PC2, where
PC1 is the result of data grouping which is worth one while PC2 contains a value of 0.
1. Each principal component (PC) is a linear combination of all variables. In other words,
each PC is a linear combination of all variables with a load assigned to each variable. The
resulting load value is usually not zero. These results in the PC results obtained will be
difficult to interpret.
techniques based on natural genetics. In the genetic algorithm used to produce an optimal
solution, the search process is performed among several alternative optimal points based
Genetic algorithms with binary coding are commonly used in function optimization
problems. The weakness of binary coding is that the solution range is in a continuous
region. [118]. Early convergence often occurs because the solution population is trapped
in the local optimum. Distributed genetic algorithms (DGA) use several subpopulations,
and each subpopulation uses operators. The advantages of binary genetic algorithm
1. This algorithm performs only a few mathematical calculations related to the problem to
be solved. Because of the changing nature of natural evolution, this algorithm seeks a
solution without regard to the processes associated with the problem being solved
directly. This algorithm can also control the objective function and defined constraints in
3. This algorithm has high flexibility and can be combined with other search methods to
be more effective.
The weakness of genetic algorithms when compared with other search algorithms is the
deal with probabilistic random numbers. The results of this data grouping will be
processed using the BGA algorithm with only two variables. The data to be processed are
ANN or ANFIS model. By looking at the characteristics of PCA where the data grouping
is PC1 to PC9, then the input parameters are determined based on the eigenvalue and
eigenvector values where PC1 is the highest value of the formation PC2 to PC9 is the
lowest value, while BGA with its characteristics results obtained by setting 50% of the
PCA feature
selection
Dimension reduction
1. Pressure
PC1 and PC2 values Transformation
2. Temperature <=1 PC1 and PC3
(DHT22) If PC2 value = 0, then values >=1
3. Electric current, PC3<=1
4. Voltage
5. Power
6. Luminous
intensity BGA feature
7. Temperature ANN Model and selection
(BMP 180)
8. Altitude ANFIS Model PC1 and PC3
From the model I propose to reduce the training time for the desired target is:
.
Dimension reduction
PC1 > PC2 PAGE \* MERGEFORMAT 179
If PC2 value = 0, and PC3 value = 1
Transformation
PC1 and PC3 value =1
No
BGA termination
PC1 and PC3
Yes
From Figure 19, it can be seen that the feature selection process uses PCA. Then,
the selection process from the BGA which consists of selection, crossover, mutation, and
regeneration, is replaced by using the PCA stages of dimension reduction and transformation
while testing the validity of the data using the process from BGA if the selection data is
appropriate. with the desired target; if not, the selection conditions will loop until the target
conditions match the energy prediction data to be achieved. The selection condition will stop
when it matches the desired target, and then the next process will continue testing with the
I will focus on the looping process resulting from the merging of PCA and BGA.
This flow looping process improves the computations in the training process without
considering the results of the PCA and BGA merging models. However, previous studies
selection for classification has also been performed with CMIM, and BGA has also been
performed to speed up computing [15]; however, this research does not explain how much
METHODOLOGY
3.1 Introduction
The theoretical review in chapter 2 discussed how the features selection that have existed
in previous studies have been discussed. The application of feature selection in energy
management in a building has not been widely applied, especially in the prediction of
energy use in a room. The application of feature selection in buildings has also been
widely applied in an area or city. In this case, there has been no application of feature
selection that has contributed to this dissertation research. The application of the previous
feature selection is also discussed, and how the weaknesses and shortcomings of previous
research have become the basis for this dissertation. Finally, at the end of chapter 2,
The data used in this study comprised training and testing data. This dataset has several
proportional attributes and a missing value. This dataset is divided into 80% as training
data, 20% as prediction data, and 10% as prediction data. In this study, we used 3283 with
2450 training data, 833 test data, and 67 predictive data. The data were collected based on
real retrieval in the PVTE UNTIRTA laboratory for 3 months. The first retrieval is from
January 06, 2021 to April 10, 2021. Data collection is based on data read by sensors
supporting tools for data recording. In addition, a PC/laptop is required where the data will
Recorded variables such as pressure, current, voltage, power, light, humidity, temperature,
and altitude. Recording is only performed on weekdays and operating hours because the
laboratory room is only active on weekdays and operating hours. Starting Monday– Friday,
09.00 to 17.00 pm. Prediction of electrical energy consumption in the laboratory where this
dissertation only focuses on the features selection used when managing energy in a
building or more, especially in this research, was carried out in the UNTIRTA PVTE
laboratory, with the help of Matlab R2018a. The working procedure of this research can be
Output: Testing
Input : PCA feature BGA feature
with ANN and
Dataset selection selection
ANFIS Models
In this study, several processes were conducted to solve the research problems. The
procedure for solving these problems is shown in Figure 18. In this study, there are 2
stages that will be performed before testing with the tested model, namely ANN and
ANFIS. In stage 1, the dataset is completed using feature selection using PCA Figure 3.2.
in stage 2, completion using feature selection using BGA. In stage 1, nine variables were
and humidity.
Determine the
Eigenvalues
(Eigenvalues)
Calculating
Attribute Weights
using
Eigenvectors
In this study, using Metlab 2018a, the attributes of the preprocessing results are repre-
sented in the form of labels (x1, x2, x3…x9), which represent the sequence of data that
X1 = Pressure (atm)
X2 = Temp@ ( DHT22 )
X3 = Current (A)
X4 = Voltage ( V )
X5 = Power ( W )
X6 = Light ( Lux )
X7 = Temp© (BMP180)
X8 = Altitude (m)
X9 = Humidity(%)
Covariance is used to measure the magnitude of the relationship between the two
attributes. The types of relationships that can occur between two attributes based on their
covariance values are: a. Positive: if the covariance value is positive or > 0 b. Negative: the
Therefore, to obtain the covariance value, equation 3. Next is the covariance matrix
Where cov(x,x), cov(y,y) and cov(z,z) the value is the same as the calculation of the
If the value of the diagonal component (var(x), var(y), and var(z)) is 1, then the covariance
matrix is the same as the correlation matrix because the correlation result from the non-
standardized data is the same as the covariance result from the standardized data. This
means that the covariance matrix is a symmetrical. When decomposed with Eigenvalue,
cov(y,x) = cov(x,y)
cov(z,y) = cov(y,z)
The eigenvalue (λ ) is a scalar number that defines the covariance matrix, which is a square
matrix measuring m m. Then, the eigenvalue (λ) corresponding to the covariance matrix is
obtained through equation 4, so that each scalar (λ1, 2……..m) satisfies equation 5 to form
an eigenvector matrix.
Principal component analysis (PCA) is an analytical technique that transforms the original
attributes that are still correlated with one another into a new set of uncorrelated attributes.
In this study, the selected principal component has the maximum contribution amount with
a value of 99% (variance threshold) based on the proportion of variance of each selected
principal component because this value is sufficient to explain the total covariance
variance of the original attribute. the proportion of variance of each principal component is
obtained through equation 9, where the covariance variance is obtained from the sum of
To determine which attributes are included in the principal component (attributes that
greatly affect the covariance variance of the observation data) with a maximum
9, where each x vector value formed corresponds to one eigenvalue for each attribute.
3.2.1.6 Choose the original attribute based on the weight of the highest eigenvector value.
In this study, the interpretation of the results of feature selection using PCA is obtained by
selecting the attribute that has the highest weighted eigenvector value from several
principal components that have a fairly large correlation with the formation of the original
attribute with a total proportion of the variance covariance of 99%. Thus, many attributes
are obtained from the attribute selection from the dataset of 9 input variables . Other
attributes that have low value weights do not have enough influence on the formation of
the original attributes of each data taken in real time at the PVTE Laboratory. It can be
concluded that this selected attribute is modeled into the BGA feature selection
Population initialization
Crossover
Mutation
Search
Termination?
Optimal fitness
results
Finish
The data used for feature selection use PCA BGA results, which are formed on the basis of
the eigenvalues that have the highest values. The data with the highest value are grouped
into PC1, PC2, PC3, and PC4 on the basis of the latent value obtained from Metlab 2018a.
Steps in the BGA feature selection with the help of the supporting software, namely
a. The initial step of this GA method is to determine the initial population with fitness
values from numerous populations and then represent the genes and chromosomes
b. Determine the fitness probability value of some fitness measures over several
generations and determine the fitness function. Then, the value is evaluated.
c. The next step is to select the population. A good or optimal value is obtained.
a new value. Not all values in a population will undergo this process.
e. The next step is mutation. This mutation technique plays a role in returning the
optimal solution that can be lost due to the previous crossover process.
using machine learning (ML), focuses on how to apply feature selection to speed up data
computing during the training and testing process. The feature selection observed based on
the references used were PCA and BGA. The data used as a reference in this study
wereobtained for approximately three months, namely January– April 2021, from the
PVTE laboratory. 3350 with 2450 training data, 833 test data, and 67 predictive data will
be used. Data sharing is based on the journal Adedeji et al. [33] [120] training data using
70% and 30% testing [121]. Other studies also discuss about 60% and 40% [122]. Other
studies 80% and 10% and predictions 10% [123]. Besides training and testing, there are
also those who use testing with a larger sample than training [124] Data are collected on
PAGE \* MERGEFORMAT 179
the basis of recorded data using a microcontroller with a sensor input with 9 input
variables. Pressure, temperature I, current, voltage, power, light, temperature II, altitude,
and humidity. Before processing the data obtained during the research in the electrical
selection here is that the reduction function reduces the number of variables that were
initially vast to fewer to facilitate analysis at a later stage. The transformation function is
used to change the variables that were initially correlated to become uncorrelated. There
are two stages in data collection. The first stage is to compare the data when the PCA and
BGA features are selected. The second stage involves data that do not use the selection
The BGA algorithm is a binary genetics algorithm. The input and output are binary
numbers used to find the most optimal solution to an optimization problem. In the
and BGA features selection by grouping the data before the data is trained in the training
and testing process. The initial stage of this research is to select the input variable by
grouping the data based on the analysis component formed from the input variable as many
as nine variables. The data of the nine variables is reduced to PC1,PC2……..PC9, then
performs the transformation by selecting the PC value that value of one which will be
chosen to influence the prediction results using the ANN and ANFIS models according to
updated model based on the combination of PCA and BGA feature selection. By looking at
the characteristics of the PCA model, which has a similar level of lack of data by choosing
a number that has a value of 1 and a number below 0 does not proceed to the next process
with the experiments that have been carried out, it can be concluded that only PC1, PC2,
PC3, and PC4 were selected to enter the next stage. second feature selection using BGA.
Only 2 components will be formed, namely PC1 and PC2, where PC1 is the result of
grouping data with a value of one, while PC2 contains a value of 0. However, there are
words, each PC is a linear combination of all variables with a load assigned to each
PAGE \* MERGEFORMAT 179
variable. The resulting load value is usually not zero. These results in the PC results
techniques based on natural genetics. In the genetic algorithm used to produce an optimal
solution, the search process is performed among several alternative optimal points based
Genetic algorithms with binary coding are commonly used in function optimization
problems. The weakness of binary coding is that the solution range is in a continuous
region. [118]. A weakness of the genetic algorithm is that early convergence often occurs
because the solution population is trapped in the local optimum. Advantages of the Binary
Genetic Algorithm:
problem to be solved. Because of the changing nature of natural evolution, this algorithm
seeks a solution without regard to the processes associated with the problem being solved
directly. This algorithm can also control the objective function and defined constraints in
3. This algorithm has high flexibility and can be combined with other search
The weakness of genetic algorithms when compared with other search algorithms
algorithms deal with probabilistic random numbers. The results of this data grouping will
be processed using the BGA algorithm with only two variables. The data to be processed
are two PCs that have the largest eigenvalues after which the data are processed using the
ANN or ANFIS model. By looking at the characteristics of PCA where the data grouping
is PC1 to PC9, then the input parameters are determined based on the eigenvalue and
eigenvector values where PC1 is the highest value of the formation PC2 to PC9 is the
lowest value, while BGA with its characteristics results obtained by setting 50% of the data
PCA
selection
Dimension reduction Transformatio
feature
1. Pressure PC1 and PC2 values n
2. Temperatur <=1 PC1 and PC3
e (DHT22)
3. Electric
If PC2 value = 0, then values >=1
current, PC3<=1
4. Voltage
5. Power
6. Luminous
intensity
BGA feature
7. Temperatur ANN Model and
e (BMP selection
ANFIS Model
180) PC1 and PC3
8. Altitude
9. Humidity
From the model 1 propose to reduce the training time in the desired target is:
Dimension reduction
PC1 > PC2
If PC2 value = 0, and PC3 value = 1
Transformation
PC1 and PC3 value =1
.
BGA termination
PC1 and PC3
No
Yes
ANN Models and Models
ANFIS
From Figure 24, it can be seen that the feature selection process uses PCA. Then,
the selection process from the BGA which consists of selection, crossover, mutation, and
transformation while testing the validity of the data using the process from BGA if the
selection data is appropriate. with the desired target; if not, the selection conditions will
PAGE \* MERGEFORMAT 179
loop until the target conditions match the energy prediction data to be achieved. The
selection condition will stop when it matches the desired target, and then the next process
2-4 5-7 8-10 11-12 1-3 4-6 7-9 10-12 1-3 4-6 7-9 10-12 1-2
1 Study of
literature
Sensor Design
2 AC Voltage,
WCS1700 for
measuring AC
Current Sensor,
DC Voltage
module, ACS712
Sensor for
measuring
current sensor
DC, dan modul
PAGE \* MERGEFORMAT 179
BH1750
3 Research data
collection
4 Research data
processing
Publications,
3 patents and
dissertations
In this research, I will focus on the looping process between combining PCA
and BGA feature selection to reduce input parameters and speed up time when con-
ducting training and testing. PCA feature selection is the grouping of data based on
eigenvalues and eigenfactors that affect data grouping. The function of feature se-
lection in PCA is to reduce the input variables and data computation time during the
training and testing process. As discussed above, the input variable consists of 9
input variables that will be reduced and transformed so that only PC1, PC2, PC3,
and PC4 will be processed according to the data grouping value of 1, while the data
grouping value below 0 will not be processed. After the data are processed with
PCA feature selection , the data are processed using BGA feature selection . The
looping process that will occur in the training and testing process makes the process
only binary numbers 1 and 0, which will be processed to the next stage. based on
4.1 Introduction
In this chapter, we explain PCA – BGA feature selection. Testing using 3350 data
points with 2450 training data, 833 testing data, and 67 prediction data. The distri-
bution of data based on the journal Adedeji et al. [33] [120] training data using 70%
and 30% testing [121][33] other studies also discuss about 60% and 40%[122].
Other studies also discuss 60% and 40%[123] in this study used 70% training, 20%
The main contribution of this dissertation is presented in the previous chapter. Con-
sidering that the combination of PCA and BGA feature selection aims to increase
the speed of data computing when the training and testing process is carried out, the
advantages and limitations that characterize the selection of PCA and BGA features.
This feature extraction process considers the characteristics of the training and test-
ing data. The difference between taking features from the training data and testing
data is the feature taking in the training data and the distribution of data according
to data collection for approximately 3 months. The choice of PCA features relates to
the explanation of the structure of the variance and covariance of variable clusters
tion without losing the original data. If we have some p The variable is then con-
verted into k variable, where is the expected component k smaller than component p
( k<p ) variable . In selecting PCA features, there are several steps that need to be
performed:
matrix [48][125] standardization results for training data 2450. Suppose attributes
A1, A2, A3, A4, A3, A4, A5, A6, A7, A8, A9.. Amount of data n = 2450.
n
2
∑ ( X 1−X )
var ( A 1 )= i=1
( n−1 )
No A1 A2 … A9
0.50521656
2 1 -0.17815 … 1.485819
0.56244173
3 8 -0.17815 … 1.506006
0.56244173
4 8 -0.17815 … 1.506006
0.50521656
5 1 -0.17815 … 1.495913
0.56244173
6 8 -0.17815 … 1.5161
0.56244173
7 8 -0.17815 … 1.485819
0.50521656
8 1 -0.17815 … 1.485819
0.50521656
9 1 -0.17815 … 1.495913
10 8 -0.17815 … 1.485819
… … … … …
2.06102E-
Averat 8.41231E-
e 13 3.06E-14 … -1.8E-15
∑ ( X 1− X ) ( Y 1−Y )
var ( A 1 , A2 ) = i=1
( n−1 )
data collection dataset, equation 2.7 is used. Covariance matrix based on the
coefficient matrix of the input variable [112]. The following is the calculation of
1
( 0.505216561−8.41231E-13 )(−0.178149221−3.05657E-14 )+ …=−0.160532037
2450−1
1
Cov(A1,A1) = ((0.505216561)2 +( 0.505216561)2+….+( -
2450−1
0.124260384)2=1
Next, we enter the results of the calculation of the covariance value in each
measuring 9x9 where the value of Cov(A1,A2) is the same as Cov(A2,A1), the
value of Cov(A1,A3) is the same as that of Cov (A3,A1), and so on in the same
way also applies to each attribute pair. The following is the diagonal form of
Covariance C1 C2 C3 C4 C5 C6 C7 C8 C9
To see the results of the covariance matrix from the complete electrical energy
The covariance matrix formed from the training data has a square matrix
matrix is obtained through equation 2.7, so that each scalar (λ1, λ 2……..λm)
satisfies equation 2.8 to calculate attribute weights using eigenvectors. To obtain the
eigenvalues (λ1, λ 2……λm), the lambda values are obtained according to table 7:
Table 5 Eigenvalue
λ Eigenvalue
λ1 2.397030227
λ2 1.586244249
λ3 1.46894975
λ4 1.079500193
λ5 0.95746891
λ6 0.684931487
λ7 0.380870837
λ8 0.258642745
λ9 0.186361601
The eigenvalues are formed from the diagonal of the covariance matrix
diagonal value is the same as the result of the calculation of the covariance variance
= 1+1+1+1+1+1+1+1+1
=9
Eigenvalue PC 1
1. Value Proportion PC 1 ( % )= x 100 %
Variance Covariance
2.39703022735865
= x 100 %=26.63366919 %
9
Eigenvaluen PC 2
2. Value Proportion PC 2 ( % )= x 100 %
Variance Covariance
PAGE \* MERGEFORMAT 179
1.58624424912761
= x 100 %=17.6249361 %
9
Eigenvalue PC 3
3. Value Proportion PC 3 ( % )= x 100 %
Variance Covariance
1.46894975
= x 100 %=16.32166389 %
9
Eigenvalue PC 4
4. Value Proportion PC 4 ( % )= x 100 %
Variance Covariance
1.079500193
= x 100 %=11.99444659 %
9
Eigenvalue PC 5
5. Value Proportion PC 5 ( % )= x 100 %
Variance Covariance
0.957468910371012
= x 100 %=10.63854345 %
9
Eigenvalue PC 6
6. Value Proportion PC 6 ( % )= x 100 %
Variance Covariance
0.684931486861588
= x 100 %=7.610349854 %
9
Eigenvalue PC 7
7. Value Proportion PC 7 ( % )= x 100 %
Variance Covariance
0.380870837201489
= x 100 %=4.231898191 %
9
Eigenvalue PC 8
8. Value Proportion PC 8 ( % )= x 100 %
Variance Covariance
0.258642745253906
= x 100 %=2.873808281 %
9
Eigenvalue PC 9
9. Value Proportion PC 9 ( % )= x 100 %
Variance Covariance
0.186361601262716
= x 100 %=2.070684458 %
9
Based on the 2018a metlab calculation, PC1 to PC9 is obtained which can
be seen in table 8
Table 6 PC Data
From the results of PC calculations obtained PC1, PC2, PC3 and PC4
To obtain the covariance of the two attributes [15] . Data are taken on the
basis of input variables from 9 sensors that have been selected using PCA feature
selection. The results of feature selection from PCA are presented in Table 5. In this
study, the dataset was used to find the fitness value according to Table 5 using a
linear equation. Feature selection is bypassed by the steps of BGA feature selection.
Results of processing after PCA using BGA and processing with BGA after
PCA. The initial step in processing using BGA first determines the function of the
training data as much as 2450 Feature selection is done after PCA using BGA
PCA.
problem to be solved. The maximum limit of the variable to be searched for is 2450.
Based on the available data, it can be seen that the value of x will be in the range of
0 < x < 2450. Thus, the most optimal data or the best is between 0 and 2450.
is done randomly or randomly. To generate random numbers, you can use exell ( (
processing table with 2450 training data. Then, a gene mutation in the second
iteration on the k3 gene using one point crossover in k3. Max =3 iterations using the
roulette wheel .
data from PC1, PC2, PC3, and PC4 are converted into binary In the third
step of BGA processing, the chromosomes of each data from PC1, PC2,
the population, where the selection is based on the fitness value. Individuals will be
selected on the basis of the smallest fitness value because in predicting energy use
in the room they get the smallest value, to get an estimate of the minimum value
In evaluating using the roulette wheel based on the fitness value f(x) %. The
Table 9 Chromosome % Pi
Total 693,808.15
Average 283.19
MIN (353.54)
MAX 2,997.92
The selection of individuals from the population was based on fitness. Individuals
with the lowest value were selected because the minimum value was used to predict
i
q 1= ∑ P j ......................................................................................... (2)
j=i
range{ 0,1 } if r < qi, then the first chromosome will be selected. q i−1< r <qi the th
chromosome −i detected. From equation 3 it can be seen that the fitness function is
because what is sought is minimal, then what is taken is % Pi which has smaller
than percent 2
1 1.001001011
1.0011010011 0.111010110011 0.0011111011
0
2 1.001001000
1.0011001111 0.111010000100 0.0100000010
1
3 1.001011111
1.0011011100 0.000111000110 0.0100110110
1
4 1.001101001
1.0011100101 0.110100110000 0.0101000101
1
5 1.000101001
1.0111001110 1.010100000100 -0.0011110101
0
6 1.000111110
1.0111110110 1.010010100000 -0.0100010011
1
7 1.001000111
1.0011001001 0.110000000011 0.0011111111
0
8 1.000011110
1.0111000110 1.010011110110 -0.0100011010
0
9 1.000101001
1.0100101110 1.000110111010 -0.0000101000
0
10 1.000111001
1.0110100111 1.000111001001 -0.0011010110
1
11 1.000111110
1.0011011000 0.111001000011 0.0010001101
1
12 1.001000010
1.0010110001 0.110111101111 0.0011101110
1
13 1.0011101011 0.110011101001 0.0010101111 1.001001100
PAGE \* MERGEFORMAT 179
1
14 1.000011001
1.0010100101 0.110000001101 0.0011111011
0
15 1.000111110
1.0010000100 0.100110100010 0.0011111011
1
16 0.111110101
1.0001001110 0.100011100000 0.0001101111
0
17 1.000001100
1.0111011011 1.000110100011 -0.0111001100
1
18 1.000111111
1.0100101101 0.111010010001 -0.0001011100
1
19 1.001001001
1.0101010110 0.111010110011 -0.0001101100
0
… … … … …
245
0.162887706 1.162887706 2.162887706 3.162887706
0
From table 10, it can be seen that binary conversion is used for the
inherit some characteristics of their parents. The crossover used in this study
was a one-point crossover. The procedure for choosing which parent will
the parent chromosome if the ith chromosome is less than or equal to the
crossover probability (Pc). The next process occurs when the number of
selected parent chromosomes is only one; this process is repeated until the
using probability (Pc), and no pair may experience a cross. If this occurs, the
offspring will have the same chromosomes as the parent (the chromosomes of the
k-0
1 0 0 1 ...... 0 0 1 1 0
0 0 1 1 ...... 1 1 1 0 0
number r (float) in the range [0, 1].- if r < pc then the chromosome is
selected for the crossover process- determine crossover point and perform a
crossover.
parent 2 with parent 3, and so on until the 2449 data. It can be seen in table
15 that the crossover that then occurs with the child resulting from the
offspring, or children from marriage. Before performing the cross, the probability of
crossover ( Pc ) is determined first. For each parent pair that is formed, we first
determine a random number between 0 and 1. For the crossover process using Excel
RAND(). For each pair, first note the Pc value. If Pc > is greater than a random
randomly, for example, using the help of exel RANBETWEEN ( 1,9 ). Suppose that
This iteration process is up to 2450 data points after obtaining a new population,
then continued with iteration 2 until the fitness function f(x) is found according to
4.3.4. Mutation
one chromosome. The easiest way to perform a mutation is to change one or more
parts of the chromosome, which depends on the mutation rate. The mutation rate
determines the probability of the number of bits in a population that are expected to
undergo mutations. If the value of the mutation rate is too small, many useful bits
may not appear in the population; however, if it is too high, the resulting offspring
will lose traits that may be superior to their parents. The mutation operator is used
to change some elements in selected individuals with the mutation probability (Pm)
the local optimal search process. The mutation that will be performed in this study
involves replacing a randomly selected gene with a new value obtained at random.
and helps avoid the local optimal search process. The mutation that will be
Steps for solution iteration 2. The repetition of the process will continue
until the most optimal value is found using Metlab 2018a. The most optimal value
performed using probability (Pm). Pm is usually small and ranges from 0.001 to
0.1. This is so that mutations are not carried out too often. The position of the
Parent 3
0001 .011 001. 0101 000. 0100 01.0 0010
1001 1100 0000 0100 1101 1000 1001 0111
0011 0101 1000 1011 0110 0111 0011
Parent 4
001. 0011 000. 1101 000. 0100 0001 .001
0111 0000 1001 1010 1101 1000 0111 0110
1011 1111 0111 0110 0111 1000 0101
Parent 3 Parent 4
001. 000. 000. 000.
0001 .01 01.0 001. 01.0
0101 0100 1101 0100
1 1001 0010 0011 0010
0000 1101 1001 1101
1100 1001 0111 1001
0100 1000 0 1010 1000
0011 0111 0000 0111
1000 110 1111 0110
0101 0011 1011 0011
1011 0111 0111 0111
Child 3 Child 4
0001 .01 001. 000. 01.0 001. 000. 000. 01.0
PAGE \* MERGEFORMAT 179
0101 0100 1101 0100
1 1001 0000 1101 0010 0011 1001 1101 0010
1100 0100 1000 1001 0111 1010 1000 1001
0011 1000 0110 0111 0000 1111 0110 0111
0101 1011 0111 0011 1011 0111 0111 0011
Child Child
M1 M2
111. 111.
0001 .01 01.0 001. 01.0
1011 1011
1 1001 110. 1010 0010 0011 111. 0010 0010
0010 0010
1100 1111 1011 1001 0111 0110 0101 1001
0111 0111
0011 0111 0100 0111 0000 0000 1000 0111
1001 1001
0101 0011 1011 0011
1000 1000
Based on the results of the Child M1 mutation and the results of the Child mutation
of the M2 Child mutation, the new participants had better scores than the previous
therefore, this data was processed using the ANN and ANFIS models according to
the desired target when calculating energy efficiency in the Electrical Engineering
Based on the results of the experiments conducted, how do the comparisons when
the PCA and BGA feature selection are implemented? The models being tested are
Correla-
tion coeffi- PCA+AN- PCA+BGA+ PCA+BGA
ANN ANFIS PCA+ANN
cient train- FIS ANN +ANFIS
ing
Time com- 396.34 1397.40
6.621 5.506 4.640 0.403
putation 9 1
Accuration 99% 100% 98% 98% 93% 86%
MSE 817 0 1,572.40 2,090.20 8,451.30 16,084.90
RSME 28.6 0 39.70 45.7 91.9 126.80
MAPE Inf Inf Inf Inf Inf Inf
Correla-
tion coeffi- PCA+AN- PCA+BGA+ PCA+BGA
ANN ANFIS PCA+ANN
cient test- FIS ANN +ANFIS
ing
Time com-
0.047 0.097 0.237 0.003 0.027 0.002
putation
Accuration 99% 100% 95% 93% 94% 93%
3,227.
MSE
70 537.09 39,563.65 42,852.96 21,772.39 33,933.49
RSME 56.81 23.18 198.91 207.01 147.55 184.21
MAPE 7.4 0.39 80.41 79.51 48.46 42.17
The test results in table 20 use 50% for training and 50% for testing use 10% of the
testing data.
Correlation
PCA+ PCA+ PCA+BGA PCA+BGA+
coefficient ANN ANFIS
ANN ANFIS +ANN ANFIS
training
Time com- 460.20
69394.828 1308.493 61.377 431.35 8.448
putation 8
Accuration 98% 97% 96% 97% 96% 95%
MSE 0.016 0.03 0.035 0.034 0.039 0.044
RSME 0.128 0.174 0.188 0.183 0.198 0.21
MAPE Inf Inf Inf Inf Inf Inf
Correlation
PCA+ PCA+ PCA+BGA PCA+BGA+
coefficient ANN ANFIS
ANN ANFIS +ANN ANFIS
testing
Time com-
0.455 4.064 0.439 0.599 0.433 0.29
putation
Accuration 98% 98% 96% 97% 96% 95%
MSE 0.033 0.016 0.035 0.034 0.039 0.044
RSME 0.18 0.128 0.19 0.183 0.198 0.21
MAPE Inf Inf Inf Inf Inf Inf
In the ANN model, the input parameters are used as the input for all. Table 17
shows the comparison of accuracy and speed of computation time when the data are
trained and tested with input variables Pressure, Temperature (DHT22), Electric
and Humidity when the model is tested with the ANN model. In this experiment,
3332 data points were tested, which were taken every 10 min from Monday. Janu-
ary 6, 2020 to Friday, April 10, 2020. The data are then divided 70% for training,
gradient
100
100
mu
10-5
-1
0 100 200 300 400 500 600 700 800 900 1000
1000 Epochs
105
Mean Squared Error (mse)
104
103
102
101
0 100 200 300 400 500 600 700 800 900 1000
1000 Epochs
Figure 27 The MSE model of ANN in training predicts the use of electrical
energy
2500 Data
Fit
Y=T
2000
Output ~= 0.99*Target + 3.6
1500
1000
500
0
0 500 1000 1500 2000 2500
Target
Figure 28 The correlation coefficient of the ANN model in training predicts the
2000
Output
1500
1000
500
0
0 500 1000 1500 2000 2500
Dataset Index
Testing based on table 16 shows that the data used is 20% of the total data obtained
in the measurements in the PVTE 2333 to 2998 laboratory. The elapsed time was
7.3961.
105
Mean Squared Error (mse)
104
103
102
101
0 100 200 300 400 500 600 700 800 900 1000
1000 Epochs
Figure 31 The MSE model of ANN in testing predicts the use of electrical
energy.
2000
Output ~= 0.99*Target + 5.7
1500
1000
500
Figure 32 The correlation coefficient of the ANN model in testing predicts the
2000
Output
1500
1000
500
0
0 100 200 300 400 500 600 700
Dataset Index
gradient
100
100
mu
10-5
-1
0 100 200 300 400 500 600 700 800 900 1000
1000 Epochs
The prediction data used in data processing is 10% of the data used in experiments
2998 to 3332.
…
10:20
10:30
10:40
10:50
11:00
11:10
11:20
11:30
11:40
11:50
17:00
KAMIS, KAMIS, KAMIS, KAMIS, KAMIS, KAMIS, KAMIS, KAMIS, KAMIS, KAMIS, … JUMAT,
2 APRIL 2 APRIL 2 APRIL 2 APRIL 2 APRIL 2 APRIL 2 APRIL 2 APRIL 2 APRIL 2 APRIL 10
2020 2020 2020 2020 2020 2020 2020 2020 2020 2020 APRIL
2020
1 2 3 4 5 6 7 8 9 10 … 335
ANN
O'cloc Actua
No Day/date/year Predictio
k l Data
n Data
THURSDAY, 2 400.1
1 10:20 427.7216
APRIL 2020 6
THURSDAY, 2 398.6
2 10:30 424.4568
APRIL 2020 4
THURSDAY, 2 401.2
3 10:40 423.8666
APRIL 2020 8
THURSDAY, 2
4 10:50 179.6 156.1327
APRIL 2020
In the ANFIS model, only 6 input parameters are used: pressure, temperature
(DHT22), electric current, voltage, and power. As with the ANN model, the data are
divided into 70% for training, 20% for testing, and 10% for prediction.
In processing the training data used 70%, namely data 1 to 2332 with minimal train-
2500 Data
Fit
Y=T
Output ~= 1*Target + 3.7e-06
2000
1500
1000
500
0
0 500 1000 1500 2000 2500
Target
2000
Output
1500
1000
500
0
0 500 1000 1500 2000 2500
Dataset Index
For the ANFIS model the data used is 20% 2333 to 2998
Data
2500 Fit
Y=T
Output ~= 0.99*Target + 4
2000
1500
1000
500
electricity consumption
2000
Output
1500
1000
500
0
0 100 200 300 400 500 600 700
Dataset Index
…
10:20
10:30
10:40
10:50
11:00
11:10
11:20
11:30
11:40
11:50
17:00
KAMIS, KAMIS, KAMIS, KAMIS, KAMIS, KAMIS, KAMIS, KAMIS, KAMIS, KAMIS, … JUMAT,
2 APRIL 2 APRIL 2 APRIL 2 APRIL 2 APRIL 2 APRIL 2 APRIL 2 APRIL 2 APRIL 2 APRIL 10
2020 2020 2020 2020 2020 2020 2020 2020 2020 2020 APRIL
2020
1 2 3 4 5 6 7 8 9 10 … 335
ANFIS
O'cloc Actua
No Day/date/year Predictio
k l Data
n Data
THURSDAY, 2 400.1
1 10:20 399.1512
APRIL 2020 6
THURSDAY, 2 398.6
2 10:30 402.7014
APRIL 2020 4
THURSDAY, 2 401.2
3 10:40 180.1402
APRIL 2020 8
THURSDAY, 2
4 10:50 179.6 184.5721
APRIL 2020
THURSDAY, 2 184.1
5 11:00 196.293
APRIL 2020 6
THURSDAY, 2 196.1
6 11:10 169.9352
APRIL 2020 6
THURSDAY, 2 169.4
7 11:20 184.6343
APRIL 2020 4
THURSDAY, 2
8 11:30 184.4 169.4262
APRIL 2020
THURSDAY, 2 169.2
9 11:40 256.1654
APRIL 2020 8
It can be seen that the training in Figures 40 and 41 has an R that is almost the
same, close to 1, around 98.7%.
gradient
102
-1
0.04
lr
0.02
0
0 100 200 300 400 500 600 700 800 900 1000
1000 Epochs
electricity consumption
2500 Data
Fit
Y=T
2000
Output ~= 0.97*Target + 8.5
1500
1000
500
0
0 500 1000 1500 2000 2500
Target
Figure 43 The correlation coefficient of the PCA model and the ANN model in
2000
Output
1500
1000
500
0
0 500 1000 1500 2000 2500
Dataset Index
Figure 44 The correlation coefficient of the PCA model and the ANN model in
When the data are processed by PCA with the model being tested, the ANN model
2000
1500
1000
500
Figure 45 The PCA correlation coefficient with the ANN model in testing
2000
Output
1500
1000
500
0
0 100 200 300 400 500 600 700
Dataset Index
…
10:20
10:30
10:40
10:50
11:00
11:10
11:20
11:30
11:40
11:50
17:00
KAMIS, KAMIS, KAMIS, KAMIS, KAMIS, KAMIS, KAMIS, KAMIS, KAMIS, KAMIS, … JUMAT,
2 APRIL 2 APRIL 2 APRIL 2 APRIL 2 APRIL 2 APRIL 2 APRIL 2 APRIL 2 APRIL 2 APRIL 10
2020 2020 2020 2020 2020 2020 2020 2020 2020 2020 APRIL
2020
1 2 3 4 5 6 7 8 9 10 … 335
Data ANN
O'cloc
No Day/date/year Actua Predictio
k
l n Data
THURSDAY, 2 400.1
1 10:20 499.115
APRIL 2020 6
THURSDAY, 2 398.6
2 10:30 495.3655
APRIL 2020 4
THURSDAY, 2 401.2
3 10:40 521.9342
APRIL 2020 8
THURSDAY, 2
4 10:50 179.6 428.5002
APRIL 2020
THURSDAY, 2 184.1
5 11:00 405.5754
APRIL 2020 6
THURSDAY, 2 196.1
6 11:10 369.7468
APRIL 2020 6
THURSDAY, 2 169.4
7 11:20 324.2928
APRIL 2020 4
THURSDAY, 2
8 11:30 184.4 381.2793
APRIL 2020
THURSDAY, 2 169.2
9 11:40 402.9195
APRIL 2020 8
Based on the results of the comparison of feature selection, the computational speed
Pelatihan: R=0.98309
2500 Data
Fit
Y=T
2000
Output ~= 0.97*Target + 9.4
1500
1000
500
0
0 500 1000 1500 2000 2500
Target
Pelatihan (Training)
3000
Target
Output
2500
2000
Output
1500
1000
500
0
0 500 1000 1500 2000 2500
Dataset Index
that data computation speeds are faster. Elapsed time: 0.002887 s, R = 0.92482,
Pengujian: R=0.92482
2500 Data
Fit
Y=T
Output ~= 0.72*Target + 2.1e+02
2000
1500
1000
500
Figure 50 The PCA correlation coefficient with the ANFIS model in testing
2000
Output
1500
1000
500
0
0 100 200 300 400 500 600 700
Dataset Index
Figure 51 The PCA correlation coefficient with the ANFIS model in testing
Figure 52 shows the difference between the prediction of the actual data by apply-
ing the PCA feature selection with the ANFIS model. You can see that the signifi -
cant difference between the actual data and the prediction data using PCA with the
ANFIS model is the average difference of 175.28, and the data closest to the actual
data is on Thursday at 10.40 with a difference that is not too far from the actual data
401.28, while the predicted data using PCA feature selection with the ANFIS model
…
10:20
10:30
10:40
10:50
11:00
11:10
11:20
11:30
11:40
11:50
16:50
KAMIS, KAMIS, KAMIS, KAMIS, KAMIS, KAMIS, KAMIS, KAMIS, KAMIS, KAMIS, … JUMAT,
2 APRIL 2 APRIL 2 APRIL 2 APRIL 2 APRIL 2 APRIL 2 APRIL 2 APRIL 2 APRIL 2 APRIL 10
2020 2020 2020 2020 2020 2020 2020 2020 2020 2020 APRIL
2020
1 2 3 4 5 6 7 8 9 10 … 334
PCA+ANFI
Day/date/ O'cloc Data S
No
year k Actual Prediction
Data
THURSDAY, 549.166462
1 10:20 400.16
2 APRIL 2020 3
THURSDAY, 612.242093
2 10:30 398.64
2 APRIL 2020 3
THURSDAY,
3 10:40 401.28 392.071271
2 APRIL 2020
KAMIS, 2 382.095496
4 10:50 179.6
APRIL 2020 2
THURSDAY, 393.145309
5 11:00 184.16
2 APRIL 2020 2
THURSDAY,
6 11:10 196.16 376.041908
2 APRIL 2020
THURSDAY, 410.265384
7 11:20 169.44
2 APRIL 2020 3
The selection of PCA and BGA features with parameter 9 yields good results,
where it is possible that the results are no better than those of the ANN model
Combining PCA and BGA feature selection, the computational speed of the elapsed
2500 Data
Fit
Y=T
2000
Output ~= 0.85*Target + 39
1500
1000
500
0
0 500 1000 1500 2000 2500
Target
2000
Output
1500
1000
500
0
0 500 1000 1500 2000 2500
Dataset Index
2000
Output ~= 0.84*Target + 50
1500
1000
500
Figure 55 The correlation coefficient of PCA and BGA models of ANN testing
2000
Output
1500
1000
500
0
0 100 200 300 400 500 600 700
Dataset Index
Figure 56 The correlation coefficient of PCA and BGA models of ANN testing
300
200
100
0
…
10:20
10:30
10:40
10:50
11:00
11:10
11:20
11:30
11:40
11:50
16:50
KAMIS, KAMIS, KAMIS, KAMIS, KAMIS, KAMIS, KAMIS, KAMIS, KAMIS, KAMIS, … JUMAT,
2 APRIL 2 APRIL 2 APRIL 2 APRIL 2 APRIL 2 APRIL 2 APRIL 2 APRIL 2 APRIL 2 APRIL 10
2020 2020 2020 2020 2020 2020 2020 2020 2020 2020 APRIL
2020
1 2 3 4 5 6 7 8 9 10 … 334
PCA+BGA+A
Actu
N Day/date/ O'clo NN
al
o year ck Prediction
Data
Data
THURSDAY, 2 400.1
1 10:20 240.4725
APRIL 2020 6
THURSDAY, 2 398.6
2 10:30 269.1279
APRIL 2020 4
THURSDAY, 2 401.2
3 10:40 269.3941
APRIL 2020 8
THURSDAY, 2
4 10:50 179.6 269.4417
APRIL 2020
THURSDAY, 2 184.1
5 11:00 251.2617
APRIL 2020 6
THURSDAY, 2 196.1
6 11:10 226.6766
APRIL 2020 6
THURSDAY, 2 169.4
7 11:20 165.5239
APRIL 2020 4
THURSDAY, 2
8 11:30 184.4 149.6222
APRIL 2020
THURSDAY, 2 169.2
9 11:40 159.3401
APRIL 2020 8
THURSDAY, 2 255.7
10 11:50 163.8037
APRIL 2020 6
… … … … …
33 THURSDAY, 2 328.3
16:50 389.8919
4 APRIL 2020 2
Selection of PCA and BGA features results in calculating the elapsed time is
Pelatihan: R=0.86134
2500 Data
Fit
Y=T
2000
Output ~= 0.74*Target + 67
1500
1000
500
0
0 500 1000 1500 2000 2500
Target
2000
Output
1500
1000
500
0
0 500 1000 1500 2000 2500
Dataset Index
Merging PCA and BGA feature selection. Elapsed time is 0.002002 s. R = 0.92923,
2000
1500
1000
500
2000
Output
1500
1000
500
0
0 100 200 300 400 500 600 700
Dataset Index
400
350
300
250
200
150
100
50
…
10:20
10:30
10:40
10:50
11:00
11:10
11:20
11:30
11:40
11:50
16:50
KAMIS, KAMIS, KAMIS, KAMIS, KAMIS, KAMIS, KAMIS, KAMIS, KAMIS, KAMIS, … JUMAT,
2 APRIL 2 APRIL 2 APRIL 2 APRIL 2 APRIL 2 APRIL 2 APRIL 2 APRIL 2 APRIL 2 APRIL 10
2020 2020 2020 2020 2020 2020 2020 2020 2020 2020 APRIL
2020
1 2 3 4 5 6 7 8 9 10 … 334
Figure 61 Prediction energiy actual Vs prediction PCA and BGA with Model
ANFIS
Table 23 Data Actual Vs prediction PCA and BGA with Model ANFIS
PCA+BGA+A
Data
Day/date/ O'clo NFIS
No Actu
year ck Prediction
al
Data
THURSDA
400.1
1 Y, 2 APRIL 10:20 240.4725
6
2020
THURSDA
398.6
2 Y, 2 APRIL 10:30 269.1279
4
2020
4.7. Comparison 3 of PCA and BGA with ANN and ANFIS models
To show the reliability of this feature selection model, the proposed feature selec-
study [5][6]. Where in this test consisted of 42 houses analyzed, but in this study it
was only tested using 1 house with 462240, with the input parameters being AC in
Usage_kW. Data collection starts on 7/15/2018 0:00 until 5/31/2019 23:59 collec-
ing, and 10% for prediction. Therefore, the results are presented in table 28.
Correla-
tion coeffi- PCA PCA+ PCA+BGA PCA+BGA
ANN ANFIS
cient train- +ANN ANFIS +ANN +ANFIS
ing
Time com- 548.30
14461.238 1135.118 44.067 523.485 7.242
putation 8
Accuration 98% 97% 96% 97% 96% 95%
MSE 0.016 0.03 0.035 0.034 0.043 0.046
RSME 0.128 0.174 0.188 0.183 0.208 0.214
MAPE Inf Inf Inf Inf Inf Inf
Correla-
tion coeffi- PCA PCA+ PCA+BGA PCA+BGA
ANN ANFIS
cient test- +ANN ANFIS +ANN +ANFIS
ing
Time com-
0.455 4.064 0.439 0.599 0.433 0.29
putation
Accuration 90% 89% 86% 86% 86% 85%
MSE 0.023 0.027 0.034 0.033 0.037 0.039
RSME 0.15 0.164 0.185 0.182 0.192 0.197
MAPE Inf Inf Inf Inf Inf Inf
In table 24, it can be seen that the test was carried out by testing data with a total of
70% of the total data usage so that testing data, testing data, and predictions were
obtained 323568, as shown in Table 28. When the ANN model was tested, it
obtained an accuracy value equal to testing and training, which is 0.983 but has not
too much difference between MSE and RSME in training and testing. However, it
still has the most campuses in the ANN and ANFIS models.
Correla-
tion coeffi- PCA+ PCA+ PCA+BGA+ PCA+BGA+
ANN ANFIS
cient train- ANN ANFIS ANN ANFIS
ing
PAGE \* MERGEFORMAT 179
Time com- 13479. 1049.7 486.78
37.998 460.268 6.986
putation 543 84 5
Accuration 98% 97% 96% 97% 96% 95%
MSE 0.016 0.03 0.035 0.034 0.043 0.214
RSME 0.128 0.174 0.188 0.183 0.208 0.046
MAPE Inf Inf Inf Inf Inf Inf
Correla-
tion coeffi- PCA+ PCA+A PCA+BGA+ PCA+BGA+
ANN ANFIS
cient test- ANN NFIS ANN ANFIS
ing
Time com-
1.503 1.872 0.501 0.728 0.454 0.296
putation
Accuration 0.983 0.968 0.963 0.965 0.955 0.952
MSE 0.016 0.03 0.035 0.034 0.043 0.046
RSME 0.128 0.174 0.188 0.183 0.208 0.215
MAPE Inf Inf Inf Inf Inf Inf
The graph of training time and testing computational time can be seen in Figure 55,
where the largest computation time is found in the ANN model when presenting
presented. The computation time is quite long, i.e., 13479,543 s for training and
1,503 s for testing. When the ANN model is added and the PCA feature is selected,
the speed of data computation becomes shorter, i.e., 486.785 s for training and test-
ing becomes 0.501 s. When ANN is added, select PCA and BGA features so that
the data speed is faster by utilizing the looping process from BGA to increase the
data speed for training 460.268 seconds for the testing process 0.454 seconds.
13000
11000
9000
7000
5000
3000
1000
ANN ANFIS PCA+ANN PCA+ANFIS PCA+BGA+A PCA+BGA+A
NN NFIS
Training Time ( NaN NaN 13479.5427 1049.78350 486.784799 37.998437 460.267827 6.986356
seconds) 8 3
Testing Time NaN NaN 1.503112 1.872113 0.501056 0.728083 0.454095 0.295787
( seconds)
The computation speed in table 29 can be seen from the graph in Figure 55, which
shows that the largest computation occurs in the ANN. It can be seen how the sig-
nificant difference from the ANN model then ANFIS shows that the acceleration for
quite a long time is found in the ANN model and while the fastest speed is found in
the model that has added PCA and BGA feature selection with the model being
tested is ANFIS.
1.5
0.5
0
1 2 3 4 5 6 7 8 9 10 11
As seen in Figure 62, a comparison of the differences between feature selection and
the model being tested shows that the prediction of energy use in house 1 is feature
selection with the addition of PCA and BGA, with the models tested being ANN
and ANFIS.
consumption (House 1)
As seen in table 26, it can be seen how close the actual data is to the prediction data
when added to the selection of PCA and BGA features with the models tested,
namely ANN and ANFIS. It can be seen from the first data that the target energy
use in house 1 is 1,731 kW before adding the ANN feature selection to 1.884113
kW. Meanwhile, when the PCA feature selection is added, the data becomes 1,807
kW, and when the PCA feature selection is combined with BGA, the data is closest
to the actual data, namely 1.753058 kW. Likewise, the model tested with the
ANFIS model is shown in Table 26. It can be seen that the ANFIS model when
added with the selection of PCA and BGA data is closer to the actual data, where
PAGE \* MERGEFORMAT 179
the actual data has a goal of 1,731 kW, whereas when the selection of PCA and
4.5. Comparison 1 of PCA and BGA with ANN and ANFIS models
On the basis of the comparison of the time when added to the feature selection by
not using the feature selection in accordance with Figure 85 and Figure 24, the
efficiency of the given time is shown. In Figure 25, it can be seen that the accuracy
ANN,ANFIS,PCA+ANN,PCA+ANFIS
ANFIS, PCA + ANN, and PCA + ANFIS training shows that the difference is not
too far from the linearity of the comparison of each model tested, as shown in table
16. The contribution of this research is how data speed when using PCA and BGA
In Figure 64, a comparison of the correlation with training and testing computation
time shows that the longest computation time is found in the ANFIS model with
s. Computational training and testing time will be faster when PCA feature selection
is added when the model is tested by ANFIS, with a computation time of 5.505738
s for training and 0.002887 s for testing. When the selection of BGA and PCA
features is added to the ANFIS model, the computation time becomes faster at
0.402992 s for training and 0.002002 for testingLikewise, if the ANN model is
tested without using feature selection, the computation time is relatively longer
BGA.
4.6. Comparison 2 of PCA and BGA with ANN and ANFIS models
Table 27 shows a comparison of the six parameters used as input for the ANN.AN-
curacy value was found in the ANFIS model with a correlation R = 1 training, and
the same test had the best accuracy value. Training for the ANFIS model with MSE
0.004 and RSME 0.061, but the computational speed is still relatively long, namely
7.447 s, while for testing, the computation time is 0.009 s. As for the testing compu-
tational speed, MSE was 1.217, RSME was 1.103, and MAPE was 0.038.
Correlation
PCA+ PCA+ PCA+BGA+ PCA+BGA
coefficient ANN ANFIS
ANN ANFIS ANN +ANFIS
training
Time com- 149.059
putation 586 7.447 5.387 0.51 4.841 0.258
Accuration 98% 100% 99% 99% 97% 95%
1,233.0 1,474.0
MSE
11.256 0.004 0 0 3,202.10 5,972.90
RSME 3.355 0.061 35.115 38.393 56.587 77.284
MAPE 25.829 Inf Inf Inf Inf Inf
Correlation
PCA+ PCA+ PCA+BGA+ PCA+BGA
coefficient ANN ANFIS
ANN ANFIS ANN +ANFIS
testing
Time com- 149.059
7.447 5.387 0.51 4.841 0.258
putation 586
Accuration 84% 100% 97% 96% 96% 95%
19187. 22071.
MSE 0.058 1.217 29350.787 57349.905
743 79
148.56
RSME 0.241 1.103 138.52 171.321 239.478
6
MAPE 35.244 0.038 55.894 50.265 75.175 100.431
PAGE \* MERGEFORMAT 179
The ANN model has a correlation value of R1 testing 0.980 with a testing value of
R2 0.836 but has a poor MSE of 11.256, RSME 3.355, and MAPE 25.829.
Figure 65 shows a comparison between the actual data and the tested model. It can
be seen that the models closest to the actual data are the ANN and AN-FIS models
but have the longest data rates from the comparison of the application of the PCA
and BGA selection features based on table 26. Comparison of the energy predic-
tions 2
Predicted data can be seen from table 27. There is a huge difference between the
actual and predicted data using PCA and BGA feature selection with the ANN and
ANFIS models.
where the longest computation time is in the ANN and ANFIS models.
The computational time for the ANN model is 149.059586 s for training and 0.403 s
for testing. A comparison of the predicted and actual data is shown in Table 28.
Comparison of the implementation of PCA and BGA feature selection with the
models tested by ANN and ANFIS. The input parameters used were six parameters
with training, testing, and prediction data equal to 70% of the data used in predict-
(PVTE). Table 29 shows a comparison of the energy use with feature selection. It
can be seen that the ANFIS model shows an accuracy value of R=1 with an MSE
value of 0.00, while the RSME is 3.355. However, the computation time is rela-
tively long, training is 7.925, and testing is 0.015. Selection of PCA and BGA fea -
Correlation
PCA+ PCA+ PCA+BGA PCA+BGA
coefficient ANN ANFIS
ANN ANFIS +ANN +ANFIS
training
Time com-
151.187 7.925 5.746 2.408 4.694 0.601
putation
Accuration 98% 100% 99% 99% 97% 95%
1,233.0 1,473.9
MSE 11.26 0 3,202.08 5,972.86
4 8
RSME 0.128 3.355 35.115 38.393 56.587 77.284
MAPE Inf Inf Inf Inf Inf Inf
Correlation
PCA+ PCA+ PCA+BGA PCA+BGA
coefficient ANN ANFIS
ANN ANFIS +ANN +ANFIS
testing
Time com-
0.082 0.015 0.011 0.007 0.01 0.004
putation
Accuration 98% 100% 99% 99% 97% 95%
1,233.0 1,473.9
MSE 11.26 0 3,202.08 5,972.96
4 8
RSME 0.128 0.061 35.115 38.393 56.587 77.285
MAPE Inf Inf Inf Inf Inf Inf
PAGE \* MERGEFORMAT 179
Comparison of the prediction results based on table 29 of the actual data with the
predicted results of the ANN model. The best accuracy value can be seen in the
ANFIS model with an accuracy value of R = 1 or equal to 100% with less computa-
tion time than the ANN model, which is 7,925 s for training, while the computing
time for testing is smaller, namely 0,015 s. However, if the ANFIS model is added
by selecting the PCA feature, the accuracy value becomes R = 0.988 with a smaller
computation time than the model that is not added by selecting the training feature
(2.408 s) and testing 0.007 seconds. If the data are added to the selection of PCA
and BGA features, the data speed when performing computations becomes faster
compared with models that are not added with feature selection.
225
175
125
75
25
1 2 3 4 5 6 7 8 9
Dat 191.6 196.88 207.6 219.04 62.64 55.68 219.92 0 51.6
a
ac-
tual
AN 176.33431 176.34236 176.34271 176.34281 103.57535 103.56794 176.34282 0 103.56595
N 3267363 1610564 1664145 8312647 0166998 4417556 7130621 8858151
ANFIS model is the closest model to the actual data according to Table 29, where
130
110
90
70
50
30
7.925059 5.746121 4.693943
10 0.082346 0.014511 0.011144 2.4077
0.006839 0.010058 0.600842
0.003721
ANN ANFIS PCA+ANN PCA+ANFI PCA+BGA+ PCA+BGA+
S ANN ANFIS
Training NaN NaN 151.18719 7.925059 5.746121 2.4077 4.693943 0.600842
Time ( sec- 2
onds)
Testing Time NaN NaN 0.082346 0.014511 0.011144 0.006839 0.010058 0.003721
( seconds)
A comparison of training time with testing time is shown in Figure 68, which shows
that the longest computation occurs in the ANN and ANFIS models. However, if
PCA feature selection is added, both the ANN and ANFIS models have relatively
faster computational speeds compared with models that are not selected selection.
This feature is observed when the ANN model has a computing speed of 151,187 s
for training and 0,082 s for testing. When added with the selection of the PCA
feature, the speed of the ANN model becomes 5,746 s for training and 0,011 s for
testing. When data are added, the selection of PCA and BGA features is relatively
faster, with 4,694 s for training and 0,010 s for testing. Similar to the ANFIS model,
added to the PCA feature selection, the speed of the ANFIS model will be relatively
fast, i.e., 2,408 s for training and 0,007 s for testing. When the ANFIS model is
added, the PCA and BGA feature selection is 0.601 s for training and 0.004 s for
speed testing, which is faster than the model that has not been added and has been
4.9. Comparison 4 of PCA and BGA with ANN and ANFIS models and data
normalisasi
Table 31 shows the comparison between the models when normalized. It can be
seen that there is an increase and improvement in the data when normalized. It in-
creases when the ANN model without training and testing normalization shows R1
and R2 99%, but when it is normalized, the accuracy value becomes 100% with
relatively smaller errors. RSME when training and testing shows 0.00, so with MSE
it can be seen that the error is below 0%, namely 4.66E-10 for training and 8.40E-
07 for testing. You can see how the difference is when the data before being nor-
malized the error is too large even though R1 and R2 show a fairly good correlation
Correlation
PCA+ PCA+ PCA+BGA+ PCA+BGA
coefficient ANN ANFIS
ANN ANFIS ANN +ANFIS
training
Time com- 389.93 1329.2
6.38 0.791 5.183 0.636
putation 6 11
Accuration 99% 100% 99% 98% 92% 86%
1,572. 2,090.1
MSE 816.97 0 9,404.61 16,084.90
43 7
RSME 28.583 0.048 39.654 45.718 39.654 126.826
Correlation
PCA+ PCA+ PCA+BGA+ PCA+BGA
coefficient ANN ANFIS
ANN ANFIS ANN +ANFIS
testing
Time com-
0.01 0.244 0.034 0.003 0.014 0.003
putation
Accuration 99% 92% 95% 92% 94% 93%
3,227. 39,563 42,852.
MSE
70 537.09 .65 96 20,721.77 16,084.85
198.90
RSME
56.813 23.175 6 207.01 143.951 126.826
MAPE 7.396 0.39 80.41 79.507 48.985 Inf
Correla-
tion coeffi-
cient train- PCA+ PCA+ PCA+BGA+ PCA+BGA
ANN
ing (nor- ANN ANFIS ANN +ANFIS
maliza-
tion)
Time com-
178.323 6.047 0.698 14.063 13.875
putation
Accuration 100% 92% 98% 96% 96%
MSE 4.66E-10 0 2,090.17 0 0
RSME 0 0.037 45.718 0.036 0.036
MAPE Inf Inf Inf Inf Inf
Correla-
tion coeffi-
cient test- PCA+ PCA+ PCA+BGA+ PCA+BGA
ANN
ing (nor- ANN ANFIS ANN +ANFIS
maliza-
tion)
Time com-
0.093 0.017 0.005 0.005 0.004
putation
Accuration 100% 92% 92% 96% 96%
42,852.9
MSE 8.40E-07 0.01 0 0
6
RSME 0.001 0.1 207.01 0.046 0.046
MAPE 0.081 96.483 79.507 44.578 44.578
building has contributed significantly. It can be seen that there is a decrease in the
computational speed of almost 99% of the tested models, namely the ANN and
selection using BGA and PCA [6][5], However, in this study, how the application
training and testing with the models tested are ANN and ANFIS. However, for
with other algorithms such as SVM, Fuzzy and each of the many more models that
can be tested.
5.1 Conclusion
In the reference that is read as a reference in writing this dissertation, there are
many ways that can be done in managing energy in a building, one of which is to
all connected devices that use a voltage source, especially in a laboratory that uses a
lot of electric voltage, in research on how to efficiently and predict energy use of
energy in a building or room with 9 variables. From the research here, the
(ML) has already been implemented using the BGA and PCA feature selection, but
in this dissertation the process and calculations are also the same, but in this
dissertation I apply the PCA feature selection first and only after it is a new BGA
feature selection .
speed in predicting energy, does not focus on the accuracy results obtained, and is
the laboratory. a building. In the reference that is read as a reference in writing this
dissertation, there are many ways that can be done in managing energy in a
air conditioning, lights used, all connected devices that use a voltage source,
efficiently and predict energy use of energy in a building or room with 9 variables.
From the research here, the application of feature selection in energy management
using a learning machine (ML) has already been implemented using the BGA and
PCA feature selection, but in this dissertation the process and calculations are also
the same, but in this dissertation I apply the PCA feature selection first and only
speed in predicting energy, does not focus on the accuracy results obtained, and is
Combining PCA and BGA feature selection and utilizing the looping process in
this dissertation, the stages of each stage are the same as mathematical calculations
for each of the characteristics of mature selection. However, it is hoped that in the
values while still considering the speed and computation of the data when the data
5.2 Recommendations
For future improvements, it is expected that the accuracy value will also be
values. It is hoped that in testing for subsequent developments, models other than
ii) The feature selection used here is connected in series or each feature selection
works according to the stages possessed by each feature selection used. It is hoped
building can be further simplified by merging in parallel so that you can find the
latest models in the feature selection that will come, but still consider the speed and
iii) The application of the ANFIS model in energy management can only be
simulated with 6 input variables; therefore, all nine variables cannot be used. In the
future, the application of ANFIS can be used using other models or two or more
ANFIS models.
iv) Feature selection using PCA and BGA are two very different characteristics of
Quantity
Date Tem Predic-
Tem Cur- Pow p© Alti- Data
No and Light tion
Pres- p© rent Volt- er (BM tude Hu- Actual
time (lux) Result
sure (DH (A) age (W) P180 (m) midity
(atm) T22) (V) ) (%)
Mon-
day, 0.992 26. 225.2 16.8 26.0 135.11 270.512
1 0.07 38.06 41.15 75.60
6/01/2 8 28 3 9 3 184 162
020
Tues-
day, 0.992 25. 225.9 12.4 26.7 99.660 268.907
2 0.06 42.54 47.32 74.06
7/01/2 1 74 1 6 1 408 36
020
Wedn
esday, 0.991 26. 225.6 14.9 26.1 119.27 269.163
3 0.07 37.75 47.88 76.25
8/01/2 7 34 6 1 9 51 045
020
Thus-
day, 0.992 26. 221.2 39.0 27.4 312.56 275.908
4 0.18 44.63 43.94 63.34
9/01/2 3 76 3 7 1 163 288
020
Friday,
0.991 26. 225.8 16.5 25.9 132.15 269.108
5 10/01/ 0.07 37.31 49.71 76.48
6 04 7 2 1 347 408
2020
Mon-
day, 0.991 26. 226.2 29.7 26.9 237.80 271.659
6 0.13 67.50 48.39 64.03
13/01/ 7 32 6 3 6 898 958
2020
Tues-
day,14 0.992 25. 226.0 17.1 25.5 136.77 269.544
7 0.08 37.76 46.45 71.38
/01/20 2 70 9 0 4 061 788
20
Wedn
esday, 0.991 27. 223.1 32.6 100.4 28.2 261.10 273.060
8 0.15 51.40 56.96
15/01/ 6 36 1 4 0 4 204 935
2020
9 Thus- 0.990 27. 0.11 223.2 24.4 67.16 27.8 49.99 62.17 195.77 269.807
PAGE \* MERGEFORMAT 179
day,
16/01/ 5 03 0 7 4 796 936
2020
Friday,
0.991 26. 170.4 29.1 27.0 233.27 273.882
10 17/01/ 0.16 32.04 47.04 55.62
9 63 4 6 3 184 953
2020
Mon-
day, 0.992 25. 225.2 34.0 25.6 272.54 273.079
11 0.15 66.43 45.04 63.24
20/01/ 3 78 4 7 4 857 341
2020
Tues-
day, 0.991 25. 224.7 38.8 26.9 310.41 273.112
12 0.17 50.74 51.35 64.33
21/01/ 6 99 4 0 3 469 956
2020
Wedn
esday, 0.992 25. 225.7 26.9 25.7 215.62 272.010
13 0.12 37.01 42.51 66.68
22/01/ 6 94 1 5 6 612 646
2020
Thus-
day, 0.992 25. 226.0 25.8 25.3 206.41 271.311
14 0.11 52.76 42.67 66.77
23/01/ 6 29 3 0 0 633 731
2020
Friday,
0.992 21. 227.0 24.9 21.9 199.90 268.753
15 24/01/ 0.11 58.83 44.05 60.93
4 66 8 9 2 204 55
2020
Mon-
day, 0.991 28. 220.0 27.3 100.0 29.1 219.07 272.450
16 0.13 53.00 55.09
27/01/ 5 07 6 8 6 3 102 673
2020
Tues-
day, 0.992 27. 220.5 26.2 104.6 27.7 209.62 273.043
17 0.12 42.35 54.92
28/01/ 6 04 7 0 6 9 612 253
2020
Wedn
esday, 0.992 26. 223.5 27.6 27.1 221.32 272.351
18 0.12 66.53 45.29 56.63
29/01/ 4 52 6 7 0 898 57
2020
Thus-
day, 0.992 26. 226.7 38.0 27.6 304.44 275.374
19 0.17 54.76 43.56 55.34
30/01/ 5 66 3 6 2 571 407
2020
Friday,
0.992 28. 226.3 37.5 28.9 300.68 276.766
20 31/01/ 0.17 48.40 46.10 55.45
2 60 3 9 4 571 017
2020
PAGE \* MERGEFORMAT 179
Mon-
day, 0.990 24. 223.8 25.2 26.3 201.82 269.441
21 0.12 44.14 55.76 66.98
3/02/2 9 90 8 3 4 204 702
020
Tues-
day, 0.991 27. 212.7 29.4 28.5 235.41 271.853
22 0.13 52.15 51.84 60.93
4/02/2 3 15 4 3 2 388 2
020
Wedn
esday, 0.991 25. 222.2 28.5 28.1 228.03 270.602
23 0.13 52.47 53.91 62.14
5/02/2 1 99 2 0 7 592 723
020
Thus-
day, 0.992 26. 226.1 39.5 26.6 316.50 275.735
24 0.17 31.11 41.07 61.32
6/02/2 8 03 7 6 4 449 264
020
Tues-
day, 0.992 26. 221.2 40.2 27.3 322.31 276.047
25 0.18 43.61 44.04 63.11
7/02/2 5 66 1 9 3 51 946
020
Mon-
day, 0.992 26. 225.2 36.7 26.4 294.15 273.845
26 0.16 49.49 44.98 67.68
10/02/ 2 02 3 7 1 51 679
2020
Tues-
day, 0.992 26. 224.3 36.1 104.7 27.0 289.53 274.442
27 0.16 43.41 61.40
11/02/ 5 11 6 9 2 3 959 906
2020
Wedn
esday, 0.992 26. 226.2 40.0 26.9 320.56 274.779
28 0.17 38.13 49.21 63.51
12/02/ 1 50 3 7 7 816 996
2020
Thus-
day, 0.991 26. 224.9 31.8 27.2 254.53 272.673
29 0.14 44.57 49.52 58.78
13/02/ 8 66 9 2 3 224 058
2020
Friday,
0.992 26. 227.5 52.6 26.8 421.01 278.159
30 14/02/ 0.23 16.75 47.32 61.14
1 19 8 3 5 878 088
2020
Mon-
day 0.993 28. 222.8 38.0 28.4 304.42 278.052
31 0.17 68.54 39.45 55.43
17/02/ 0 09 5 5 8 286 701
2020
32 Tues- 0.991 25. 0.54 221.8 116. 53.80 26.5 50.80 61.27 928.88 288.617
PAGE \* MERGEFORMAT 179
day,
18/02/ 7 63 6 11 1 327 293
2020
Wedn
esday, 0.992 25. 211.6 38.1 26.4 305.42 273.646
33 0.18 55.11 47.87 58.40
19/02/ 1 06 6 8 7 367 135
2020
Thus-
day, 0.991 26. 221.8 63.0 26.5 504.52 272.283
34 0.15 25.22 51.98 55.98
20/02/ 5 16 8 7 6 408 657
2020
Friday,
0.993 27. 219.2 43.7 27.7 349.75 275.491
35 21/02/ 0.13 48.61 34.28 59.57
6 46 6 2 8 837 78
2020
Mon-
day, 0.992 25. 155.8 20.0 26.5 160.75 273.082
36 0.15 58.85 44.42 56.83
24/02/ 3 79 2 9 9 265 496
2020
Tues-
day, 0.992 25. 223.4 28.4 25.8 227.52 274.648
37 0.18 57.91 43.73 57.34
25/02/ 5 19 9 4 4 49 942
2020
Wedn
esday, 0.990 27. 222.4 14.9 28.2 119.75 275.473
38 0.21 44.58 60.78 61.09
26/02/ 5 61 5 7 5 51 611
2020
Thus-
day, 0.992 25. 105.3 13.3 26.3 106.72 273.936
39 0.16 63.98 44.62 54.94
27/02/ 4 64 3 4 3 163 294
2020
Friday,
0.991 26. 197.4 26.9 27.4 215.59 272.203
40 28/02/ 0.15 36.91 54.36 58.78
0 89 1 5 6 347 852
2020
Mon-
day, 0.991 27. 223.8 24.0 28.4 192.01 271.321
41 0.10 47.30 47.82 57.97
2/03/2 9 62 1 0 3 796 223
020
Tues-
day, 0.992 26. 225.3 31.1 27.6 249.04 277.419
42 0.22 41.31 45.76 55.36
3/03/2 2 24 6 3 1 163 829
020
43 Wedn 0.992 26. 0.10 225.8 25.6 102.6 28.2 40.98 53.55 205.08 271.959
esday, 7 74 4 4 7 0 735 468
4/03/2
PAGE \* MERGEFORMAT 179
020
Thus-
day, 0.992 27. 225.7 30.8 27.6 246.80 277.011
44 0.19 55.70 44.72 43.30
5/03/2 4 33 4 5 3 327 372
020
Friday,
0.992 26. 225.3 31.1 27.6 249.04 277.419
45 6/03/2 0.22 41.31 45.76 55.36
2 24 6 3 1 163 829
020
Thurs-
day, 0.990 29. 222.5 35.6 30.5 285.02 287.349
46 0.41 67.76 56.35 71.18
12/03/ 9 50 6 3 4 531 502
2020
Friday,
0.989 30. 222.0 63.7 31.3 510.12 286.190
47 13/03/ 0.40 66.46 70.64 72.58
3 26 3 7 3 245 786
2020
Mon-
day, 0.992 26. 221.1 151. 27.2 1210.1 286.351
48 0.38 32.98 46.21 57.00
16/03/ 2 46 4 27 7 976 938
2020
Tues-
day, 0.990 28. 218.7 43.5 29.1 348.72 285.427
49 0.38 46.22 62.31 55.61
17/03/ 3 36 1 9 9 163 97
2020
Wedn
esday, 0.992 26. 223.3 29.4 27.0 235.40 287.491
50 0.42 68.11 44.96 48.49
18/03/ 3 48 8 3 9 082 411
2020
Thus-
day, 0.993 24. 219.2 43.6 25.7 349.33 287.597
51 0.43 53.38 37.56 46.33
19/03/ 2 95 3 7 2 551 534
2020
Friday,
0.991 26. 224.0 26.8 27.2 215.00 287.309
52 20/03/ 0.43 46.53 49.40 50.49
8 61 8 8 3 082 03
2020
Mon-
day, 0.991 28. 223.4 28.2 28.9 225.66 287.294
53 0.41 78.28 52.20 57.66
23/03/ 5 21 3 1 9 857 074
2020
Tues-
day, 0.993 27. 222.7 23.7 28.3 190.08 288.061
54 0.40 79.08 33.92 62.23
24/03/ 6 41 9 6 0 98 576
2020
Wedn 0.990 28. 225.9 40.7 29.2 326.23 285.183
55 0.38 86.43 63.70 56.99
esday, 1 56 2 8 1 673 355
PAGE \* MERGEFORMAT 179
25/03/
2020
Thus-
day, 0.991 27. 225.6 38.6 28.0 309.29 285.416
56 0.37 76.53 56.24 60.79
26/03/ 2 45 7 6 5 469 668
2020
Friday,
0.991 29. 222.3 27.9 30.5 223.44 287.641
57 27/03/ 0.41 65.71 54.08 77.59
3 71 8 3 3 163 872
2020
Mon-
day, 0.991 26. 221.2 27.9 27.1 223.81 282.856
58 0.34 65.39 56.90 45.41
30/03/ 0 11 1 8 8 388 667
2020
Tues-
day, 0.991 27. 223.4 29.6 29.3 237.32 286.801
59 0.40 73.30 52.58 51.91
31/03/ 4 89 9 7 0 735 456
2020
Wedn
esday, 0.990 28. 226.1 28.7 29.4 230.29 283.739
60 0.33 60.48 57.85 47.23
1/04/2 8 14 8 9 9 224 466
020
Thus-
day, 0.991 27. 226.3 39.6 29.0 317.43 286.400
61 0.40 64.27 54.54 45.26
2/04/2 2 16 8 8 6 837 581
020
Friday,
0.990 30. 225.1 45.9 32.4 367.34 286.698
62 3/04/2 0.40 66.34 64.21 74.14
1 19 3 2 8 694 352
020
Mon-
day, 0.991 25. 225.4 36.1 27.9 289.20 285.682
63 0.40 64.63 52.94 44.42
6/04/2 4 84 7 5 4 49 957
020
Tues-
day, 0.992 27. 225.5 48.5 28.5 387.99 287.158
64 0.40 74.63 46.34 53.21
7/04/2 2 31 9 0 1 837 152
020
Wedn
esday, 0.992 26. 225.6 45.4 27.1 363.24 286.953
65 0.42 61.85 47.96 53.69
8/04/2 0 10 3 1 5 898 195
020
Thus-
day, 0.993 28. 227.1 54.0 29.6 432.04 287.680
66 0.38 66.22 39.93 68.84
9/04/2 1 44 5 1 1 082 642
020
PAGE \* MERGEFORMAT 179
Friday,
0.993 27. 224.7 48.9 29.0 391.76 287.549
67 10/04/ 0.38 67.52 37.75 61.26
2 73 5 7 0 653 815
2020
44.64 118 9801. 1446 2424. 1215. 2107. 2747. 11572. 12309.9
Total 25 5.63 6.90 73 .60 22 28 59 41 82 6
0.992 26. 217.8 32.1 27.0
Average 1 35 0.15 2 5 53.87 1 46.84 61.05 257.17 273.55
Quantity
Date Tem Predic-
Lig Data
No and Tem Cur- p© Alti- tion
Power ht Actual
Pres- p© rent Volt- (BM tude Hu- Result
time (W) (lux
sure (DHT (A) age P180 (m) midity
)
(atm) 22) (V) ) (%)
Mon-
day, 26.2 0.0 225.2 38. 26.0 41.1 135.11
1 0.9928 16.89 75.60 155.826
6/01/2 8 7 3 06 3 5 184
020
Tues-
day, 25.7 0.0 225.9 42. 26.7 47.3 99.660 151.950
2 0.9921 12.46 74.06
7/01/2 4 6 1 54 1 2 408 8
020
Wedne
sday, 26.3 0.0 225.6 37. 26.1 47.8 119.27 154.656
3 0.9917 14.91 76.25
8/01/2 4 7 6 75 9 8 51 9
020
Thus-
day, 26.7 0.1 221.2 44. 27.4 43.9 312.56 315.844
4 0.9923 39.07 63.34
9/01/2 6 8 3 63 1 4 163 3
020
Friday,
26.0 0.0 225.8 37. 25.9 49.7 132.15 170.597
5 10/01/ 0.9916 16.52 76.48
4 7 7 31 1 1 347 9
2020
Mon-
day, 26.3 0.1 226.2 67. 26.9 48.3 237.80 240.238
6 0.9917 29.73 64.03
13/01/ 2 3 6 50 6 9 898 4
2020
Tues-
day,14/ 25.7 0.0 226.0 37. 25.5 46.4 136.77 174.910
7 0.9922 17.10 71.38
01/202 0 8 9 76 4 5 061 5
0
8 Wedne 0.9916 27.3 0.1 223.1 32.64 100 28.2 51.4 56.96 261.10 256.254
PAGE \* MERGEFORMAT 179
sday,
15/01/ 6 5 1 .40 4 0 204 4
2020
Thus-
day, 27.0 0.1 223.2 67. 27.8 49.9 195.77 205.570
9 0.9905 24.47 62.17
16/01/ 3 1 0 16 4 9 796 6
2020
Friday,
26.6 0.1 170.4 32. 27.0 47.0 233.27 282.835
10 17/01/ 0.9919 29.16 55.62
3 6 4 04 3 4 184 4
2020
Mon-
day, 25.7 0.1 225.2 66. 25.6 45.0 272.54 261.788
11 0.9923 34.07 63.24
20/01/ 8 5 4 43 4 4 857 7
2020
Tues-
day, 25.9 0.1 224.7 50. 26.9 51.3 310.41 283.976
12 0.9916 38.80 64.33
21/01/ 9 7 4 74 3 5 469 2
2020
Wedne
sday, 25.9 0.1 225.7 37. 25.7 42.5 215.62 221.888
13 0.9926 26.95 66.68
22/01/ 4 2 1 01 6 1 612 9
2020
Thus-
day, 25.2 0.1 226.0 52. 25.3 42.6 206.41 213.978
14 0.9926 25.80 66.77
23/01/ 9 1 3 76 0 7 633 4
2020
Friday,
21.6 0.1 227.0 58. 21.9 44.0 199.90 200.038
15 24/01/ 0.9924 24.99 60.93
6 1 8 83 2 5 204 4
2020
Mon-
day, 28.0 0.1 220.0 100 29.1 53.0 219.07 226.518
16 0.9915 27.38 55.09
27/01/ 7 3 6 .06 3 0 102 5
2020
Tues-
day, 27.0 0.1 220.5 104 27.7 42.3 209.62 225.877
17 0.9926 26.20 54.92
28/01/ 4 2 7 .66 9 5 612 8
2020
Wedne
sday, 26.5 0.1 223.5 66. 27.1 45.2 221.32
18 0.9924 27.67 56.63 229.466
29/01/ 2 2 6 53 0 9 898
2020
19 Thus- 0.9925 26.6 0.1 226.7 38.06 54. 27.6 43.5 55.34 304.44 299.536
day, 6 7 3 76 2 6 571 5
30/01/
PAGE \* MERGEFORMAT 179
2020
Friday,
28.6 0.1 226.3 48. 28.9 46.1 300.68
20 31/01/ 0.9922 37.59 55.45 293.357
0 7 3 40 4 0 571
2020
Mon-
day, 24.9 0.1 223.8 44. 26.3 55.7 201.82 240.154
21 0.9909 25.23 66.98
3/02/2 0 2 8 14 4 6 204 7
020
Tues-
day, 27.1 0.1 212.7 52. 28.5 51.8 235.41 238.161
22 0.9913 29.43 60.93
4/02/2 5 3 4 15 2 4 388 5
020
Wedne
sday, 25.9 0.1 222.2 52. 28.1 53.9 228.03
23 0.9911 28.50 62.14 238.168
5/02/2 9 3 2 47 7 1 592
020
Thus-
day, 26.0 0.1 226.1 31. 26.6 41.0 316.50 301.261
24 0.9928 39.56 61.32
6/02/2 3 7 7 11 4 7 449 2
020
Tues-
day, 26.6 0.1 221.2 43. 27.3 44.0 322.31
25 0.9925 40.29 63.11 315.725
7/02/2 6 8 1 61 3 4 51
020
Mon-
day, 26.0 0.1 225.2 49. 26.4 44.9 294.15 280.125
26 0.9922 36.77 67.68
10/02/ 2 6 3 49 1 8 51 3
2020
Tues-
day, 26.1 0.1 224.3 104 27.0 43.4 289.53 282.877
27 0.9925 36.19 61.40
11/02/ 1 6 6 .72 3 1 959 7
2020
Wedne
sday, 26.5 0.1 226.2 38. 26.9 49.2 320.56
28 0.9921 40.07 63.51 299.138
12/02/ 0 7 3 13 7 1 816
2020
Thus-
day, 26.6 0.1 224.9 44. 27.2 49.5 254.53 256.377
29 0.9918 31.82 58.78
13/02/ 6 4 9 57 3 2 224 4
2020
Friday,
26.1 0.2 227.5 16. 26.8 47.3 421.01 372.520
30 14/02/ 0.9921 52.63 61.14
9 3 8 75 5 2 878 8
2020
31 Mon- 0.9930 28.0 0.1 222.8 38.05 68. 28.4 39.4 55.43 304.42 317.438
PAGE \* MERGEFORMAT 179
day
17/02/ 9 7 5 54 8 5 286
2020
Tues-
day, 25.6 0.5 221.8 53. 26.5 50.8 928.88 736.931
32 0.9917 116.11 61.27
18/02/ 3 4 6 80 1 0 327 6
2020
Wedne
sday, 25.0 0.1 211.6 55. 26.4 47.8 305.42 277.162
33 0.9921 38.18 58.40
19/02/ 6 8 6 11 7 7 367 3
2020
Thus-
day, 26.1 0.1 221.8 25. 26.5 51.9 504.52 263.690
34 0.9915 63.07 55.98
20/02/ 6 5 8 22 6 8 408 9
2020
Friday,
27.4 0.1 219.2 48. 27.7 34.2 349.75 240.449
35 21/02/ 0.9936 43.72 59.57
6 3 6 61 8 8 837 2
2020
Mon-
day, 25.7 0.1 155.8 58. 26.5 44.4 160.75 261.416
36 0.9923 20.09 56.83
24/02/ 9 5 2 85 9 2 265 7
2020
Tues-
day, 25.1 0.1 223.4 57. 25.8 43.7 227.52 282.923
37 0.9925 28.44 57.34
25/02/ 9 8 9 91 4 3 49 5
2020
Wedne
sday, 27.6 0.2 222.4 44. 28.2 60.7 119.75
38 0.9905 14.97 61.09 301.184
26/02/ 1 1 5 58 5 8 51
2020
Thus-
day, 25.6 0.1 105.3 63. 26.3 44.6 106.72 276.206
39 0.9924 13.34 54.94
27/02/ 4 6 3 98 3 2 163 7
2020
Friday,
26.8 0.1 197.4 36. 27.4 54.3 215.59 259.893
40 28/02/ 0.9910 26.95 58.78
9 5 1 91 6 6 347 1
2020
Mon-
day, 27.6 0.1 223.8 47. 28.4 47.8 192.01 186.151
41 0.9919 24.00 57.97
2/03/2 2 0 1 30 3 2 796 2
020
42 Tues- 0.9922 26.2 0.2 225.3 31.13 41. 27.6 45.7 55.36 249.04 355.350
day, 4 2 6 31 1 6 163 6
3/03/2
PAGE \* MERGEFORMAT 179
020
Wedne
sday, 26.7 0.1 225.8 102 28.2 40.9 205.08 192.979
43 0.9927 25.64 53.55
4/03/2 4 0 4 .67 0 8 735 7
020
Thus-
day, 27.3 0.1 225.7 55. 27.6 44.7 246.80 328.752
44 0.9924 30.85 43.30
5/03/2 3 9 4 70 3 2 327 1
020
Friday,
26.2 0.2 225.3 41. 27.6 45.7 249.04 355.350
45 6/03/2 0.9922 31.13 55.36
4 2 6 31 1 6 163 6
020
Thurs-
day, 29.5 0.4 222.5 67. 30.5 56.3 285.02 280.445
46 0.9909 35.63 71.18
12/03/ 0 1 6 76 4 5 531 2
2020
Friday,
30.2 0.4 222.0 66. 31.3 70.6 510.12
47 13/03/ 0.9893 63.77 72.58 394.035
6 0 3 46 3 4 245
2020
Mon-
day, 26.4 0.3 221.1 32. 27.2 46.2 1210.1 581.557
48 0.9922 151.27 57.00
16/03/ 6 8 4 98 7 1 976 6
2020
Tues-
day, 28.3 0.3 218.7 46. 29.1 62.3 348.72 363.252
49 0.9903 43.59 55.61
17/03/ 6 8 1 22 9 1 163 6
2020
Wedne
sday, 26.4 0.4 223.3 68. 27.0 44.9 235.40 647.672
50 0.9923 29.43 48.49
18/03/ 8 2 8 11 9 6 082 7
2020
Thus-
day, 24.9 0.4 219.2 53. 25.7 37.5 349.33 778.715
51 0.9932 43.67 46.33
19/03/ 5 3 3 38 2 6 551 9
2020
Friday,
26.6 0.4 224.0 46. 27.2 49.4 215.00 586.281
52 20/03/ 0.9918 26.88 50.49
1 3 8 53 3 0 082 7
2020
Mon-
day, 28.2 0.4 223.4 78. 28.9 52.2 225.66 409.236
53 0.9915 28.21 57.66
23/03/ 1 1 3 28 9 0 857 4
2020
Tues- 27.4 0.4 222.7 79. 28.3 33.9 190.08
54 0.9936 23.76 62.23 630.428
day, 1 0 9 08 0 2 98
PAGE \* MERGEFORMAT 179
24/03/
2020
Wedne
sday, 28.5 0.3 225.9 86. 29.2 63.7 326.23 352.720
55 0.9901 40.78 56.99
25/03/ 6 8 2 43 1 0 673 9
2020
Thus-
day, 27.4 0.3 225.6 76. 28.0 56.2 309.29 446.408
56 0.9912 38.66 60.79
26/03/ 5 7 7 53 5 4 469 5
2020
Friday,
29.7 0.4 222.3 65. 30.5 54.0 223.44 233.673
57 27/03/ 0.9913 27.93 77.59
1 1 8 71 3 8 163 3
2020
Mon-
day, 26.1 0.3 221.2 65. 27.1 56.9 223.81 428.572
58 0.9910 27.98 45.41
30/03/ 1 4 1 39 8 0 388 1
2020
Tues-
day, 27.8 0.4 223.4 73. 29.3 52.5 237.32 441.053
59 0.9914 29.67 51.91
31/03/ 9 0 9 30 0 8 735 4
2020
Wedne
sday, 28.1 0.3 226.1 60. 29.4 57.8 230.29 380.463
60 0.9908 28.79 47.23
1/04/2 4 3 8 48 9 5 224 3
020
Thus-
day, 27.1 0.4 226.3 64. 29.0 54.5 317.43 478.461
61 0.9912 39.68 45.26
2/04/2 6 0 8 27 6 4 837 6
020
Friday,
30.1 0.4 225.1 66. 32.4 64.2 367.34 311.875
62 3/04/2 0.9901 45.92 74.14
9 0 3 34 8 1 694 3
020
Mon-
day, 25.8 0.4 225.4 64. 27.9 52.9 289.20 512.220
63 0.9914 36.15 44.42
6/04/2 4 0 7 63 4 4 49 8
020
Tues-
day, 27.3 0.4 225.5 74. 28.5 46.3 387.99 541.833
64 0.9922 48.50 53.21
7/04/2 1 0 9 63 1 4 837 6
020
Wedne
sday, 26.1 0.4 225.6 61. 27.1 47.9 363.24 608.622
65 0.9920 45.41 53.69
8/04/2 0 2 3 85 5 6 898 5
020
PAGE \* MERGEFORMAT 179
Thus-
day, 28.4 0.3 227.1 66. 29.6 39.9 432.04 427.570
66 0.9931 54.01 68.84
9/04/2 4 8 5 22 1 3 082 3
020
Friday,
27.7 0.3 224.7 67. 29.0 37.7 391.76
67 10/04/ 0.9932 48.97 61.26 544.482
3 8 5 52 0 5 653
2020
44.642 1185 9801. 1446.6 242 1215 2107 2747. 11572. 12025.5
Total 5 .63 6.90 73 0 4.22 .28 .59 41 82 0
26.3 0.1 217.8 53. 27.0 46.8
Average 0.9921 5 5 2 32.15 87 1 4 61.05 257.17 267.23