You are on page 1of 214

Improved Computation Time in Energy Consumption Prediction

Models for ANN and ANFIS Based on PCA and BGA Feature
Selection

DESMIRA

ELECTRICAL ENGINEERING

2023
Faculty of Electrical Engineering

Improved Computation Time in Energy Consumption Prediction Models


for ANN and ANFIS Based on PCA and BGA Feature Selection

Desmira

Electrical Engineering

2023
Improved Computation Time in Energy Consumption Prediction
Models for ANN and ANFIS Based on PCA and BGA Feature
Selection

DESMIRA

A thesis submitted
in fulfillment of the requirements for the degree of
Doctor of Philosophy

Faculty of Electrical Engineering

UNIVERSITI TEKNIKAL MALAYSIA MELAKA

2023
DECLARATION

I declare that this Choose an item. entitled “Improved Computation Time in Energy

Consumption Prediction Models for ANN and ANFIS Based on PCA and BGA Feature

Selection” is the result of my own research except as cited in the references. The
Choose an item.
has not been accepted for any degree and is not concurrently submitted in

candidature of any other degree.

Signature :

Name :

Date :
APPROVAL

Choose an item. and in my opinion, this Choose an item. is adequate in terms of scope

and quality for the award of the degree of Doctor of Philosophy Electrical Enginering.

I hereby declare that I have checked this

Signature :

Supervisor Name :

Date :
DEDICATION

Start to write your dedication here....

This part you need to key in manually.

Info: Times New Roman, Font Size 12, Double spaced


ABSTRACT

This thesis discusses the increase in computation time by combining principal component
analysis (PCA) and binary genetic algorithm (BGA) feature selection. The selection of
PCA and BGA features is implemented using the ANN and ANFIS models to predict
energy use in a building, especially in the UNTIRTA PVTE laboratory

In Chapter 2, we thoroughly review the relevant literature on feature selection methods.


The existing feature selection method donor provides a better computing time. We note
that most of the existing feature selection methods have not been able to improve the
computational time, which is better at predicting energy consumption. Another weakness
of the existing feature selection methods is that most of the existing literature selects
features one by one and does not provide a direct measure to assess whether additional
features will be added or not. To overcome this problem, they need a method for selecting
the selected features so that data grouping will occur, which can increase the computation
time in the training and testing process. However, in real applications, it is difficult to
estimate the number of useful features before the feature selection process. This thesis
addresses these weaknesses and fills a gap in the feature selection literature. In Chapter 3,
we propose a development approach for PCA and BGA feature selection. There are three
new ingredients. First, by including data grouping in PCA, we only group the data into two
data groups. Second, we find data groupings that have been evaluated in groupings
performed individually, resulting in several data groupings. Third, we found that a
comparison of computational time and accuracy does not always yield a better value.
Finally, based on the combination of PCA and BGA feature selection, we evaluate feature
importance using only two datasets. The advantage of grouping these data is that it can
increase the computation time when the training and testing process is in progress. The
experimental results show that combining PCA and BGA feature selection can increase the
computation time with the ANN and ANFIS models. In Chapter 4, we found that in some
situations, the implementation of feature selection in PCA with only 2 data groupings can
increase the computation time. This motivated me to research how data grouping occurs in
PCA by only combining 2 groups of data so that I can combine the application of BGA
feature selection with process looping. To obtain better computational results, we use data
grouping in PCA and process looping in BGA. The experimental results show that the
combination of PCA and BGA selection can lead to better data computation. In addition to
the feature selection method that combines PCA and BGA, in Chapter 5, we examine two
feature selection methods with clustering on PCA. In particular, we propose a clustering
technique for PCA using only PC1 and PC2 data grouping, where the value ƛ is one for
PC1. If the value is below zero, it is placed in PC2. In the first step, we used the models
tested, namely ANN and ANFIS. In this step, the model was tested without feature
selection. Second, we tested feature selection using PCA with models tested by ANN and
ANFIS. Third, we tested the model with PCA and BGA feature selection where PC1 and
PC2 data grouping was applied to speed up data computation time. We summarize the
contributions of this thesis and analyze those developed. Finally, we provide suggestions
for further work on feature selection.

PAGE \* MERGEFORMAT xviii


ABSTRAK

Tesis ini membincangkan peningkatan masa pengiraan daripada gabungan pemilihan ciri
Analisis Komponen Utama (PCA) dan Algoritma Genetik Binari (BGA). Pemilihan ciri
PCA dan BGA dilaksanakan menggunakan model ANN dan ANFIS dalam meramal
penggunaan tenaga dalam sesebuah bangunan terutamanya di makmal UNTIRTA PVTE
Dalam Bab 2, kami menyemak dengan teliti literatur yang berkaitan tentang kaedah
pemilihan ciri. Kaedah pemilihan ciri sedia ada tidak dapat memberikan masa
pengkomputeran yang lebih baik. Kami ambil perhatian bahawa kebanyakan kaedah
pemilihan ciri sedia ada tidak dapat meningkatkan masa pengiraan yang lebih baik untuk
meramalkan penggunaan tenaga. Satu lagi kelemahan kaedah pemilihan ciri sedia ada
ialah kebanyakan literatur sedia ada memilih ciri satu demi satu dan tidak memberikan
ukuran langsung untuk menilai sama ada ciri tambahan akan ditambah atau tidak. Untuk
mengatasi masalah ini, mereka memerlukan kaedah yang digunakan untuk memilih ciri-
ciri yang dipilih supaya pengumpulan data akan berlaku yang boleh meningkatkan masa
pengiraan dalam proses latihan dan ujian. Walau bagaimanapun, dalam aplikasi sebenar,
adalah sukar untuk menganggarkan bilangan ciri berguna sebelum proses pemilihan ciri.
Tesis ini menangani kelemahan ini, dan mengisi jurang dalam literatur pemilihan ciri.
Dalam Bab 3 kami mencadangkan pendekatan pembangunan pemilihan ciri PCA dan
pemilihan ciri BGA. Terdapat tiga bahan baru. Pertama, dengan memasukkan
pengelompokan data dalam PCA, kita hanya menjadikan pengelompokan menjadi 2
pengelompokan data. Kedua, kita dapati pengelompokan data yang setakat ini dinilai
dalam pengelompokan yang dilakukan secara individu menghasilkan pengelompokan data
yang agak banyak. Ketiga, kami mendapati bahawa perbandingan masa dan ketepatan
pengiraan tidak selalu menunjukkan nilai yang lebih baik. Akhir sekali, berdasarkan
gabungan pemilihan ciri PCA dan BGA, kami menilai kepentingan ciri menggunakan
hanya 2 set data. Kelebihan mengumpulkan data ini ialah ia boleh meningkatkan masa
pengiraan apabila proses latihan dan ujian sedang dijalankan. Keputusan eksperimen
menunjukkan bahawa menggabungkan pemilihan ciri PCA dan BGA boleh meningkatkan
masa pengiraan dengan model ANN dan ANFIS. Dalam Bab 4, kami mendapati bahawa
dalam beberapa situasi di mana pelaksanaan pemilihan ciri dalam PCA dengan hanya 2
kumpulan data boleh meningkatkan masa pengiraan menjadi lebih pantas. Inilah yang
mendorong saya dalam masalah penyelidikan bagaimana pengumpulan data berlaku
dalam PCA dengan hanya menggabungkan 2 kumpulan data supaya saya dapat
menggabungkan aplikasi pemilihan ciri BGA dengan gelung proses. Untuk mendapatkan
hasil pengiraan yang lebih baik, kami menggunakan pengumpulan data dalam PCA dan
gelung proses dalam BGA. Keputusan eksperimen menunjukkan bahawa keberkesanan
gabungan pemilihan PCA dan BGA boleh membawa kepada pengiraan data yang lebih
baik. Sebagai tambahan kepada kaedah pemilihan ciri yang menggabungkan PCA dan
BGA, kami mengkaji dua kaedah pemilihan ciri dengan pengelompokan pada PCA dalam
Bab 5, khususnya, kami mencadangkan teknik pengelompokan pada PCA menggunakan
hanya pengelompokan data PC1 dan PC2 di mana nilai ƛ ialah satu. untuk PC1 dan Jika
nilainya di bawah sifar maka ia diletakkan dalam PC2. Pada langkah pertama, kami
menggunakan model yang diuji, iaitu ANN dan ANFIS, dalam langkah ini model diuji
tanpa menggunakan pemilihan ciri. Kedua, kami menguji pemilihan ciri menggunakan

PAGE \* MERGEFORMAT xviii


PCA dengan model yang diuji oleh ANN dan ANFIS. Ketiga, kami menguji model dengan
pemilihan ciri PCA dan BGA di mana pengelompokan data PC1 dan PC2 digunakan
untuk mempercepatkan masa pengiraan data. Kami meringkaskan sumbangan tesis ini,
dan menganalisis yang dibangunkan. Akhir sekali, kami menyediakan beberapa cadangan
untuk kerja lanjut mengenai pemilihan ciri.

PAGE \* MERGEFORMAT xviii


ACKNOWLEDGEMENTS

In the name of Allah, the Most Gracious, the Most Merciful

First and foremost, I want to thank and praise Almighty God, my Creator, my Sustainer,
for all that I have received since the beginning of my life. I would like to express my ap-
preciation to Universiti Sultan Ageng Tirtayasa (UNTIRTA) and Universiti Teknikal
Malaysia Melaka (UTeM) for providing a research platform. Thanks also to the Malaysian
Ministry of Higher Education (MOHE) for its financial assistance.

My highest appreciation goes to my main supervisor, Ir.Dr. Norazhar Bin Abu Bakr, for all
his support, advice and inspiration. Don't forget to also thank the accompanying supervi-
sor, Dr. Mohd Ruzaini Hashim and Dr. Romi wiryadinata, S.T,. M.Eng who have provided
guidance and direction during the work of this dissertation.

For my husband who I love the most because of Allah who has given spirit and hope. For
my children, whom I love because of Allah, Ayra's sister, Aisya's sister, Ardam's brother,
Azkian's brother and Arman during this dissertation are often left behind so that they can
complete the dissertation. For those of you who have been called by Allah, hopefully they
will be placed in the back in the sun of Allah. Make it easy for all my brothers and sisters
to always be in the protection of Allah.

PAGE \* MERGEFORMAT xviii


TABLE OF CONTENTS

PAGE
DECLARATION
APPROVAL
DEDICATION
ABSTRACT..........................................................................................................................ii
ABSTRAK...........................................................................................................................iii
ACKNOWLEDGEMENTS.................................................................................................v
TABLE OF CONTENTS....................................................................................................vi
LIST OF TABLES..............................................................................................................ix
LIST OF FIGURES............................................................................................................xi
LIST OF SYMBOLS AND ABBREVIATIONS............................................................xvi
LIST OF APPENDICES....................................................................................................17
LIST OF PUBLICATIONS...............................................................................................18
CHAPTER 1 INTRODUCTION.....................................................................................19
1.1 Background....................................................................................................................19
1.2 Problem Statement.........................................................................................................20
1.4 Scope of Research..........................................................................................................23
1.5 Contribution of Research................................................................................................25
1.6 Thesis Outline.................................................................................................................25
CHAPTER 2 LITERATURE REVIEW.......................................................................27
2.1 Introduction....................................................................................................................27
2.2 Feature Selection Using PCA ( Principal Component Analysis )..................................33
2.3 Prediction of Energy Consumption Using the Feature Selection Binary Genetic
Algorithm (BGA).............................................................................................................40
2.3.1. General Structure of Binary Genetic Algorithm (BGA)................................44
2.3.2. Chromosomal Coding.....................................................................................46
2.3.3. Genetic Operators...........................................................................................47
2.3.4. Selection Process............................................................................................47
2.3.5. Crossover........................................................................................................49
2.3.6. Mutation.........................................................................................................52
2.4.Energy Consumption Prediction Using Artificial Neural Network (ANN) Model........56
2.4.1 Initialize..........................................................................................................57
2.4.2 Activation........................................................................................................58

PAGE \* MERGEFORMAT xviii


2.4.3 Update the weights..........................................................................................58
2.4.4 Iteration...........................................................................................................59
2.5. Prediction of Energy Consumption Using the Adaptive Neuro Fuzzy Inference System
(ANFIS) Model................................................................................................................60
2.5.1. Fuzzification Layer.........................................................................................61
2.5.2. Product Layer.................................................................................................61
2.5.3. Normalization Layer......................................................................................62
2.5.4. Defuzzification Layer....................................................................................62
2.5.4. Total Output...................................................................................................62
2.4 Summary........................................................................................................................63
CHAPTER 3 METHODOLOGY..................................................................................72
3.1 Introduction....................................................................................................................72
3.2 Research Design.............................................................................................................73
3.2.1 PCA feature selection procedure.....................................................................74
3.2.2. The data used by BGA feature selection........................................................79
3.3 Proposed Methodology...................................................................................................80
3.4. Limitation of Proposed Methodology...........................................................................86
CHAPTER 4 ......................................................................................................................90
4.1 Introduction....................................................................................................................90
4.2. BGA and PCA feature selection....................................................................................90
4.3 BGA Feature Selection...................................................................................................98
4.3.2 Evaluation of the fitness of each chromosome is based on %Pi, the selected
chromosomes are crossover........................................................................100
4.3.3. Perform a crossover for the selected parent (crossover)..............................103
4.3.4. Mutation.......................................................................................................114
4.4. Comparison 1 of PCA and BGA with ANN and ANFIS models...............................117
4.4.1 ANN model...................................................................................................119
4.4.1.1 ANN model with training data...................................................................119
4.4.1.2 ANN model with testing data.....................................................................124
4.4.1.3 ANN model with prediction data...............................................................129
4.5.1 ANFIS model................................................................................................131
4.5.1.1 ANFIS model training data........................................................................131
4.5.1.2 ANFIS model testing data..........................................................................133
4.5.1.2 ANFIS data prediction model.....................................................................135
4.6.1. PCA dan model ANN...................................................................................137
4.6.1.1 PCA dan model ANN training data............................................................137
4.6.1.2 PCA dan model ANN testing data.............................................................141
4.6.1.3 PCA and model ANN with prediction data................................................143
4.7.1. PCA and ANFIS models..............................................................................145
4.7.1.1 PCA with ANFIS training data model.......................................................145
4.7.1.2 PCA with ANFIS testing data model.........................................................146
4.7.1.3 PCA with ANFIS data prediction model....................................................148
4.8.1 PCA and BGA with ANN models.................................................................150
4.8.1.1 PCA and BGA model ANN training data..................................................150
4.8.1.2 PCA and BGA model ANN testing data....................................................152
4.8.1.3 PCA and BGA ANN models with prediction data.....................................154

PAGE \* MERGEFORMAT xviii


4.9.1. PCA and BGA with ANFIS models.............................................................156
4.9.1.1 PCA and BGA with ANFIS training data models......................................156
4.9.1.2 PCA and BGA with ANFIS testing data models.......................................157
4.8.1.2 PCA and BGA models of ANFIS prediction data......................................159
4.7. Comparison 3 of PCA and BGA with ANN and ANFIS models...............................161
4.5. Comparison 1 of PCA and BGA with ANN and ANFIS models...............................167
4.6. Comparison 2 of PCA and BGA with ANN and ANFIS models...............................169
4.8. Comparison 4 of PCA and BGA with ANN and ANFIS models...............................172
4.9. Comparison 4 of PCA and BGA with ANN and ANFIS models and data normalisasi
........................................................................................................................................177
4.1.0 Summary...................................................................................................................179
CHAPTER 5 ....................................................................................................................180
5.1 Conclusion....................................................................................................................180
5.2 Recommendations........................................................................................................181
REFERENCES.................................................................................................................183

PAGE \* MERGEFORMAT xviii


LIST OF TABLES

TABLE TITLE PAGE

Table 1 . Energy forecast review..........................................................................................

Table 2 BGA Journal Study................................................................................................

Table 3 Data Standardization (calculate correlation)..........................................................

Table 4 Covariance Matrix Results.....................................................................................

Table 5 Eigenvalue..............................................................................................................

Table 6 PC Data..................................................................................................................

Table 7 Data Reduction......................................................................................................

Table 8 Data processed after going through the BGA process...........................................

Table 9 Chromosome % Pi ..............................................................................................

Table 10 Chromosome PC1, PC2, PC3 and PC4..............................................................

Table 11 Cross-marriage of parents..................................................................................

Table 12 New population ................................................................................................

Table 13 First mutation process........................................................................................

Table 14 Second mutation process ..................................................................................

Table 15 The third mutation process ................................................................................

Table 16 Comparison of Energy Predictions 1.................................................................

Table 17 Comparison of Energy Predictions 2.................................................................

Table 18 ANN Actual Vs Predicted Data ........................................................................

Table 19 ANFIS Actual Vs Predicted Data.......................................................................

Table 20 Data Actual Vs Data Prediksi PCA and Model ANN........................................

Table 21 Actual Data vs Predicted Data PCA and ANFIS Models .................................

PAGE \* MERGEFORMAT xviii


Table 22 Data Actual Vs prediction PCA and BGA with Model ANN............................

Table 23 Data Actual Vs prediction PCA and BGA with Model ANFIS.........................

Table 24 Comparison of Energy Predictions....................................................................

Table 25 Comparison of Energy Predictions 3.................................................................

Table 26 Prediction Data Actual and prediction of housing energy consumption

(House 1).........................................................................................................

Table 27 Comparison of Energy Predictions 2.................................................................

Table 28 Prediction Data Actual and prediction model ANN ..........................................

Table 29 Comparison of Energy Predictions 3.................................................................

Table 30 Comparison of Energy Predictions 4.................................................................

Table 31 Comparison of Energy Predictions 5.................................................................

PAGE \* MERGEFORMAT xviii


LIST OF FIGURES

FIGURE TITLE PAGE

Figure 1 .Research overview................................................................................................

Figure 1 Data flow diagram of the PCA and BGA-based Energy consumption

prediction method..............................................................................................

Figure 2 Conceptual model for the effectiveness of using PCA for the feature

selection stage [84]............................................................................................

Figure 3 Genetik Algorithm Flowchart [99]......................................................................

Figure 4 Genes,Chromosomes, dan populatin [100]..........................................................

Figure 5 Truncation Selection Method Source : [103]........................................................

Figure 6 One-point crossover method.................................................................................

Figure 7 Two-point crossover method................................................................................

Figure 8 Metode Swap Mutation [110]...............................................................................

Figure 9 Uniform mutation..................................................................................................

Figure 10 ANN Architecture...............................................................................................

Figure 11 Structure of ANFIS.............................................................................................

Figure 12 . PCA+BGA Research Contribution....................................................................

Figure 13 Research framework...........................................................................................

Figure 14 Block diagram of PCA aplication to the ANFIS model [6][5]..........................

Figure 15 Block Diagram of PCA application to ANN and ANFIS models......................

Figure 16 Flow chart combining PCA and BGA as a feature selection..............................

Figure 17 Stages of Research Image...................................................................................

Figure 18 PCA feature selection procedure........................................................................

PAGE \* MERGEFORMAT xviii


Figure 19 BGA feature selection procedure........................................................................

Figure 20 Matlab R2018a Support Software.......................................................................

Figure 21 Block diagram of PCA aplication to the ANFIS model [6][5]..........................

Figure 22 Block Diagram of PCA application to ANN and ANFIS models......................

Figure 23 Flow chart combining PCA and BGA as a selection feature..............................

Figure 24 ANN model network during training in predicting the use of electrical

energy in the PVTE room................................................................................

Figure 25 ANN training state model in training in predicting electricity usage...............

Figure 26 The MSE model of ANN in training predicts the use of electrical energy.......

Figure 27 The correlation coefficient of the ANN model in training predicts the use

of electrical energy..........................................................................................

Figure 28 The correlation coefficient of 2 ANN models in training predicts the use

of electrical energy..........................................................................................

Figure 29 The ANN model network at the time of testing in predicting the use of

electrical energy in the PVTE room................................................................

Figure 30 The MSE model of ANN in testing predicts the use of electrical energy.........

Figure 31 The correlation coefficient of the ANN model in testing predicts the use of

electrical energy...............................................................................................

Figure 32 The correlation coefficient of 2 ANN models in testing predicts the use of

electrical energy...............................................................................................

Figure 33 Training state model ANN in testing in predicting electricity usage ...............

Figure 34 Prediction energiy actual Vs prediction Model ANN.......................................

Figure 35 The correlation coefficient of the ANFIS model in training predicts the

use of electrical energy....................................................................................

PAGE \* MERGEFORMAT xviii


Figure 36 The correlation coefficient of the 2 ANFIS models in testing predicts the

use of electrical energy....................................................................................

Figure 37 The correlation coefficient of the ANFIS model in testing predicts

electricity consumption ..................................................................................

Figure 38 The correlation coefficient of the 2 ANFIS models in testing predicts the

use of electrical energy ...................................................................................

Figure 39 Prediction energiy actual Vs prediction Model ANFIS....................................

Figure 40 ANN model network with PCA during training in predicting the use of

electrical energy in the PVTE room................................................................

Figure 41 ANN training state model with PCA in training in predicting electricity

consumption ...................................................................................................

Figure 42 The correlation coefficient of the PCA model and the ANN model in

testing predicts electricity consumption .........................................................

Figure 43 The correlation coefficient of the PCA model and the ANN model in

testing predicts electricity consumption..........................................................

Figure 44 The PCA correlation coefficient with the ANN model in testing predicts

the use of electrical energy..............................................................................

Figure 45 The correlation coefficient of 2 ANN models in testing predicts the use of

electrical energy ..............................................................................................

Figure 46 Prediction energiy actual Vs prediction PCA and Model ANN.......................

Figure 47 The correlation coefficient of the ANFIS model in training predicts the

use of electrical energy ...................................................................................

Figure 48 The correlation coefficient of the 2 ANFIS models in testing predicts the

use of electrical energy ...................................................................................

PAGE \* MERGEFORMAT xviii


Figure 49 The PCA correlation coefficient with the ANFIS model in testing predicts

electricity consumption...................................................................................

Figure 50 The PCA correlation coefficient with the ANFIS model in testing predicts

electricity consumption...................................................................................

Figure 51 Prediction energiy actual Vs prediction PCA and Model ANFIS....................

Figure 52 Correlation coefficient of PCA and BGA model of ANN training data

predicts electricity consumption .....................................................................

Figure 53 Correlation coefficient of PCA and BGA model of ANN training data

predicts electricity consumption .....................................................................

Figure 54 The correlation coefficient of PCA and BGA models of ANN testing data

predicts electricity consumption .....................................................................

Figure 55 The correlation coefficient of PCA and BGA models of ANN testing data

predicts electricity consumption .....................................................................

Figure 56 The correlation coefficient of combining PCA and BGA feature selection

in the ANFIS model in training predicts electricity consumption ..................

Figure 57 The correlation coefficient of 2 PCA and BGA ANFIS models in training

predicts the use of electrical energy ...............................................................

Figure 58 Correlation coefficient combining PCA and BGA feature selection of the

ANFIS model in testing predicts electricity consumption .............................

Figure 59 The correlation coefficient of 2 PCA and BGA models of ANFIS in

testing predicts electricity consumption..........................................................

Figure 60 Prediction energiy actual Vs prediction PCA and BGA with Model AN-

FIS...................................................................................................................

Figure 61 Computasi time 3 Training Vs Testing.............................................................

PAGE \* MERGEFORMAT xviii


Figure 62 Comparison of Energy Predictions 4 Vs Usage_KW.......................................

Figure 63 Comparison correlation coefficient R1 and R2

ANN,ANFIS,PCA+ANN,PCA+ANFIS..........................................................

Figure 64 Comparison of corrections R1 and R2 with computational training and

testing time......................................................................................................

Figure 65 Comparison of Energy Predictions 2 Vs data actual........................................

Figure 66 Computasi time 3 Training Vs Testing.............................................................

Figure 67 Comparison of Energy Predictions 4 Vs data actual........................................

Figure 68 Comparison of Energy Predictions 4 Vs data actual........................................

PAGE \* MERGEFORMAT xviii


LIST OF SYMBOLS AND ABBREVIATIONS

PCA - Principal Components Analysis


λ - Eigenvalues
xi - variable to –i
Zm - linear combination of the variables x
-
Yp - component to – i
-
-
a - eigenvector value
-
x - Standard value of variable
-
A - Matrix nxn

I Identity matrix

D Eigenvalue of its eigenvector

X original variable

X-1 inverse of eigenvector X

U eigenvector matrix
BGA - Binary Genetic Algorithm
Np - Population size
i - Number og chormosome
f(x) - Fitness function
vi - Fitness value eval
pi - Perents
Pc - Probability crossover
Pm - Mutation probability

PAGE \* MERGEFORMAT xviii


LIST OF APPENDICES

APPENDIX TITLE PAGE

No table of Figures entries found.


LIST OF PUBLICATIONS

Journal with Impact Factor

Desmira,D, Bakar, N. A., Hashim, M. R., Wiryadinata, R., & Hamid, M. A. (2022).
Laboratory prediction energy control system based on artificial intelligence network.
Bulletin of Electrical Engineering and Informatics, 11(3), 1280–1288.
https://doi.org/10.11591/eei.v11i3.1821 ISI indexed, Q3, IF = 2.4 (2022)
Desmira, D, Hamid, M. A., Bakar, N. A., Nurtanto, M., & Sunardi, S. (2022). Smart Traffic
Lights Using Microcontroller Based on Fuzzy Logic. IAES International Journal of
Artificial Intelligence, 11(2), 809–818. https://doi.org/10.11591/ijai.v11.i3.pp809-818
ISI indexed, Q2, IF = 2.3 (2022)
Desmira,D,Bakar, N. A., Wiryadinata, R., Hamid, M. A., & Kholifah, N. (2022).
Comparison of principal component analysis and ANFIS to improve EEVE Laboratory
energy use prediction performance. 27(2), 970–979.
https://doi.org/10.11591/ijeecs.v27.i2.pp970-979 ISI indexed, Q3, IF = 2.4 (2022)

PAGE \* MERGEFORMAT 179


CHAPTER 1

INTRODUCTION

1.1 Background

The world’s energy crisis is getting more and more so that it is necessary to hold something

rare to make efficient use of that energy. Various efforts have been made by the Indonesian

government to save energy in all fields. An initiative is to use energy effectively in a build-

ing. To create an efficient energy environment in a building, a proper energy management

strategy is required. In this study, a feature selection technique was used to select significant

input variables for managing energy usage and optimizing processor memory.

Figure 1.Research overview

PAGE \* MERGEFORMAT 179


The purpose of the feature selection here is

1. Reduction and transformation of the input variables.

2. Reducing the amount of time spent on computing during the training and testing processes.

The features selected in this study were principal component analysis (PCA) and binary ge-

netic algorithm (BGA). In this study, two stages of the feature selection method are intro-

duced, where feature selection using principal component analysis (PCA) is used as the first

stage selection of input variables, where from nine input variables will be reduced to only

four significant input variables. The binary genetic algorithm (BGA) will be used as the sec-

ond stage of the feature selection method to further reduce the number of significant input

variables to only three. The focus of this study is to minimize the computational time re-

quired during the learning process.

1.2 Problem Statement

When applying machine learning to predict energy consumption, particularly within large

datasets, a complex web of challenges surfaces that significantly impact computation time

and efficiency.

1. Computation Time and Efficiency: The handling of large datasets in machine learning

prolongs computation time. Selecting appropriate features that optimize input variables is

crucial for speeding up this process, especially for predicting electricity consumption (Y.

Zhou et al., 2020)(Jurado et al., 2015)(Guyon., 2011).

PAGE \* MERGEFORMAT 179


2. Data Transfer and Dimensionality: Larger datasets create issues such as elongated data

transfer rates without proper feature selection (Paudel et al., 2017), and the necessity for

dimensional reduction to facilitate training and testing (Shreem et al., 2022).

3. Feature Selection Limitations and Inefficiencies: The lack of development in feature

selection steps may hinder computational time efficiency in predictions (Philippe Leray

et al, 2021).

4. Analyzing current machine learning features reveals substantial constraints in predicting

room energy use. These include problems in data grouping techniques, particularly in

selecting PCA features (Qu & Chen, 2020); (Shapi et al., 2021), difficulties in achieving

faster computation time without combining relevant feature selection (Y. Zhou et al.,

2020), and the challenges in applying BGA feature selection to increase computation

time with minimal looping processes in predictions (Ameri et al., 2013).

5. Solutions and Future Prospects: The proposed resolution emphasizes a strategic com-

bination of PCA and BGA selection methods, focusing on improving computational effi-

ciency within energy consumption prediction models for ANN and ANFIS. Despite these

efforts, feature selection still presents accuracy issues that require further exploration and

development.

The formulation of this problem stems from the inherent complexity of machine learning

applications, specifically in predicting energy consumption. It encompasses multiple layers,

including computation time, data handling, and feature selection, revealing the intricate na-

ture of balancing efficiency, accuracy, and technological advancement. The above challenges

highlight the necessity for innovative and targeted solutions, particularly those that focus on

feature selection methods that accommodate the unique needs of large-scale data processing

PAGE \* MERGEFORMAT 179


in energy consumption predictions. Future studies may expand on these foundations, provid-

ing further insights and refinements to enhance the speed and accuracy of energy predictions

in various contexts.

1.3 Research Objective

This study aims to confront the challenges and limitations associated with machine learning

in predicting energy consumption. This research seeks to explore innovative solutions and

frameworks by focusing on the following three main objectives:

1.Expand Existing Methods in PCA Feature Selection:

Objective: To improve and enhance existing principal component analysis (PCA) methods

for feature selection.

Relevance to Problem Statement: This will alleviate the challenges related to

computation time and data grouping, aligning with the first and second problem statements.

Expected Outcome: Enhanced computational efficiency by optimizing PCA feature

selection, hence streamlining the training and testing processes.

2.Develop Data Grouping and Combine Feature Selection:

Objective: To develop new data grouping strategies and integrate PCA-based feature

selection with BGA process looping.

PAGE \* MERGEFORMAT 179


 Relevance to Problem Statement: This statement addresses the third problem identified

earlier, focusing on overcoming the limitations of existing feature selection and the

challenges of implementing BGA.

 Expected Outcome: Increased computing speed for predicting indoor energy consumption,

such as that in laboratories, using a novel combination of PCA and BGA.

3. Analyze and Evaluate the Combined PCA and BGA Methodology:

Objective: To evaluate the effectiveness and efficiency of the combined PCA and BGA

the feature selection method for predicting energy consumption.

Relevance to Problem Statement: This objective resonates with previously identified

challenges and provides an analytical perspective to validate the proposed solutions.

Expected Outcome: A comprehensive understanding of the strengths and potential areas

of improvement in the combined PCA and BGA methodology, offering insights for future

research and practical implementations.

in summary, these research objectives develop, innovate, and critically evaluate methods that

reduce input variables and accelerate computation time. The target models, ANN and ANFIS,

will be used to predict energy consumption within the UNTIRTA PVTE laboratory. By

meeting these objectives, this research will contribute to the field of machine learning for

energy consumption prediction by enhancing both efficiency and increased compute time and

by providing a rigorous evaluation of the implemented methods.

1.4 Scope of Research

The scope of this research is limited to the following three aspects:


PAGE \* MERGEFORMAT 179
1. In this thesis, one of the feature selection PCA and BGA will be combined.

2. Feature selection will be used to predict the energy consumption of a room, especially in

the laboratory and several different data sets.

3.This feature selection development will use 3350 data, with 2450 training data, 833 test

data, and 67 predictive data.

The data will be collected based on real retrieval in the UNTIRTA PVTE laboratory for three

months. The first retrieval was performed from January 6, 2021 to April 10, 2021. Data

retrieval was based on data read by sensors connected to the microcontroller using a data

recording device comprising several supporting tools for data recording. It also requires a

PC/laptop where the data will be recorded and displayed on the Excel PLX-DAQ software.

Parameter variables such as pressure, current, voltage, power, light, humidity, temperature,

and altitude will be recorded. Recording will be conducted on weekdays and operating hours

because the laboratory room is only active on weekdays and operating hours. Starting

Monday– Friday, 09.00 to 17.00 pm. Electrical energy management for energy efficiency in

predicting energy use in the laboratory focuses on selecting the features used when managing

energy in one or more buildings, especially in this research conducted in the UNTIRTA

PVTE laboratory, with the help of Matlab R2018a.

4. The data was also tested with data on the UCI Machine Learning Repository website with

PERACON data with 10 houses, and then the data was tested with one house totaling 525600

data where the data includes energy consumption from 6/1/2018 12:00:00 AM to 5/31/2019

11:59:00PM

PAGE \* MERGEFORMAT 179


1.5 Contribution of Research

Contributions to this thesis are made in the following related areas:

i) Application of feature selection in energy management in buildings to predict the use of

electrical energy to achieve energy efficiency in PVTE laboratory room.

ii) Application of the process looping to PCA and BGA feature selection can increase speed

and computation when data are trained and tested.

iii) Training and testing data based on input variables from 9 real inputs taken at the PVTE

UNTIRTA laboratory. The calculation of energy consumption was compared with actual and

predictive data using the ANN and ANFIS models.

iv) Experienced a 15% increase over the BGA and PCA studies.

1.6 Thesis Outline

Based on the objectives previously presented and on the approach proposed before, this thesis

is made up of five (5) chapters, the contents of which are summarized as follows:

 Chapter 1. Introduction. This chapter presents the background of the study,

research problems, objectives, scopes, contributions, and significance of the

research.

 Chapter 2. Literature review. Chapter two begins by briefly describing the

importance of efficiency and predicting energy in a room, especially in the

laboratory, and how to manage energy in a room. Short formula for energy

prediction and energy efficiency in an energy management room commonly

used today.

PAGE \* MERGEFORMAT 179


 Chapter 3. Methodology. This chapter presents the methodology developed to

predict energy in a room, especially in the laboratory, using the PCA and

BGA features selection.

 Chapter 4. Case studies. In this chapter, the model is tested using a model

before and after feature selection with real data taken for 3 months in the

UNTIRTA PVTE laboratory.

 Chapter 5. Conclusion and future works. Chapter 5 contains the conclusions

and achievements of future work.

PAGE \* MERGEFORMAT 179


CHAPTER 2

LITERATURE REVIEW

2.1 Introduction

Based on International Energy Agency (IEA) projections for the period 2006-2030,

the world’s energy demand mostly comes from non-Organisation for Economic Co-operation

and Development (OECD) countries, which is 87%. [16][17]. China’s energy demand growth

is projected to be the largest among other regions [18][19]. India recently also showed that

growth in energy demand is quite large and just one level below that of China. In Indonesia

itself, various efforts have been made to reduce energy use in rooms or universities. Various

efforts have also been made to achieve efficient and stable energy use. Energy use to achieve

energy efficiency needs to be achieved by developing a smart room with various sensor in-

puts that affect consumption in the room [20] . Energy consumption is the demand used to

indicate the amount of energy that must be supplied to a building at any given time. Exces-

sive use of electrical energy will increase energy waste and have a negative impact on the

environment. Predicting a building’s energy consumption is an important approach for sav-

ing energy, which benefits both individuals and society. Predicting energy consumption in

buildings is essential for improving energy performance, with the aim of achieving energy

savings and reducing environmental impact.

A building that produces the largest energy consumption, usually 40% of the energy

used. Increasing efficiency is a good effort to reduce the negative caused from an economic

viewpoint. Another impact of efficiency is the reduction of fossil fuels and the stabilization of

PAGE \* MERGEFORMAT 179


energy prices. Prediction of energy consumption [21] and monitoring [22] in a room to help

increase the effectiveness and efficiency of decisions made to reduce energy demand Predic-

tion Energy consumption models are mostly used in the first step in energy saving and effi-

ciency improvement models and reduce costs [23] thermal energy storage adhesive to im-

prove energy efficiency [24]. The input variables in a model are very decisive in the perfor-

mance of the network model that will be used in making predictions that will be faster and

more cost-effective [2]. A technique used to reduce data from larger data to smaller ones is

information gain (IG) [25][26], PCA [27][28],GA[29][30],SVM[31][32].

This study was developed to improve energy management in predicting energy con-

sumption in the Electrical Engineering Vocational Education (PVTE) laboratory at Ageng

Tirtayasa University (UNTIRTA). University energy consumption is strongly influenced by

climate[33] Guang Dong University has conducted research with the results that the construc-

tion of a conservation-oriented campus has a significant effect on energy savings; however, it

was found that it was difficult to measure energy consumption indicators [34]. University

buildings in China develop target energy efficiency strategies according to local conditions,

with multiple campuses and different climates being one of the factors of difficulty in imple-

menting energy efficiency in buildings [11]. Historical data on daily electricity usage in two

London South Bank University buildings were regressed over normalized data from six input

variables to obtain efficient energy use in a building [35]. A case study of three campus

buildings in Tianjin predicts that the average electricity consumption of one occupant varies

depending on the function of the building and the control mode of electrical equipment [36].

The development of statistical regression models is used to understand the relationship be-

tween each variable and energy consumption [37]. In the previous research method, the im-

PAGE \* MERGEFORMAT 179


proved modeling can be used in other types of buildings as long as it has an energy consump -

tion monitoring platform, not limited to campus buildings [38].

In a building, among the input parameters that affect energy use in a building is the

behavior of room users, which greatly affects the use of the room, and it is difficult to deter -

mine how much energy consumption is needed. [39]. In the previous energy consumption

prediction model [40][41][42][43][44][45], features are directly used as inputs in the predic-

tion stage. Feature selection is performed with the aim of selecting features that most affect

energy consumption [46]. With feature selection, it is expected that the prediction stage can

be performed more effectively and increase accuracy. The purpose of the feature selection

method is to obtain a reduced set by removing some features that are considered irrelevant for

text sentiment classification to increase classification accuracy and decrease the duration of

the machine learning algorithm [47]. However, the disadvantage of attribute selection is that

it requires training a large dataset to obtain reliable transformations [48]. One technique to

overcome the high dimension of features is feature selection. Various feature selection tech-

niques [49], such as information gain [50], have been used to reduce vector dimensions.

In addition to feature selection techniques, another technique that can be used to over-

come the problem of high feature dimensions is dimensionality reduction. The purpose of the

dimension reduction technique is to obtain a new data representation that is managed into a

lower dimension [51]. The linear dimension reduction algorithm consists of a singular value

decomposition (SVD) algorithm and a principal component analysis algorithm [52]. How-

ever, the weakness of the linear dimensionality reduction algorithm is that it produces a linear

combination of all features that may be contaminated by noise and reduces the performance

of the classification algorithm. The linear dimensionality reduction algorithm will have diffi-

PAGE \* MERGEFORMAT 179


culties when dealing with non-linear data.[53]. The features used to predict energy consump-

tion include pressure, temperature (DHT22), current, voltage, power, light intensity, tempera-

ture (BMP180), altitude, and humidity. The output of the prediction model is the energy con-

sumption. A flowchart of the energy consumption prediction model using PCA as the feature

selection method is shown in Fig. 1 following. Experiments have shown that both the preci-

sion and durability of the optimized GA-BP ANN are better [54]. The optimal PI controller

tuning program developed on the genetic algorithm was developed and implemented in a

non-linear HVAC system simulated by the HVACSIM + package [55] . Simple and sophisti-

cated controllers were used in this study: ON– OFF, PID, and fuzzy controllers. The opti-

mized ANN validation demonstrated good predictive accuracy, as the regression coefficients

R2 for consumption and discomfort were greater than 0.77 and 0.84 for the three tested con-

trollers, respectively. Various distance predictions and ANN amounts of training data were

tested [56] . Feasibility and application of ANFIS in building energy load forecasting areas

with 96.5 and 83.8% for heating and cooling energy predictions [57]. An ideal module

through Model 4, this model is extended to predict the thermal performance of real homes,

thereby providing a fast and effective prediction tool. [58]. Electricity consumption in Thai -

land from 1986 to 2010 results show that the ANN model reduces the mean absolute percent-

age error (MAPE) to 0.996%, while ARIMA and MLR are 2.80981 and 3.2604527%, respec-

tively.[59]. The electricity demand is based on two soft computing techniques, and the exper-

imental ARIMA model reveals that EFuNN performs better [60] In this study, excellent re-

sults were obtained using the electrical load forecasting method on buildings based on ANN

feedback [61]. The ANN and ANFIS techniques have much better performance than the em-

pirical formula (for the dataset test, ANN R2 = 0.97, ANFIS R2 = 0.92 and Marciano R2 =

0.54). Between ANN and ANFIS, the ANN model is slightly better although the difference is

small [62]. To improve the overall performance of the heating system in terms of energy effi -

PAGE \* MERGEFORMAT 179


ciency and thermal comfort using the ANFIS model, the highest RMSE result is 0.5782 C

[63]. In this study, the GA-ANFIS method also obtained a better modeling time than ANN

[64]. Research data by comparison with actual consumption in the fifth year. The results

show that ANN is 6% better, whereas MR, GP, SVM, and DNN have MAPEs of 8.5%, 8.7%,

9%, and 11%, respectively [65]. For electricity use, MAPE has a set number of validations of

34.0%, a reduction of 31.8% against a theoretical benchmark; to heat the fuel using MAPE

reduced to 25.1%, to 49.8%. Prediction performance needs to be significantly improved using

an ANN as a reliable design tool [66] The GA-based ANN model has a mean square error

equal to 0.990 and 0.002 for each test dataset in predicting the bearing capacity of the pile.

[67] . Among several studies related to energy consumption using fuzzy logic [68][69][70]

[71].

Table 1. Energy forecast review

Number Study Study field Energy Method Scope


market
1 [72] Electricity Greek ARIMA Daily
2 [73] Electricity Taiwan Neural net- Monthly
work
Time serie
3 [74] Electricity Turkey Neural net- Annual
work
4 [75] Electricity Turkey Fuzzy logic Annual
5 [76] Electricity Turkey GA Annual
6 [77] All kind of de- 3 Com- ANFIS,ANN Monthly
mand pany in
Istanbul
7 [78] Transportation Jordon ANFIS Annual
energy
8 [79] Energy model- Finlandia ANFIS-PSO Daily
ing
9 [33] Energy con- African PSO-ANFIS Monthly
sumption

For that, how to manage electrical energy for more efficiency in predicting energy in

the electrical engineering vocational education (PVTE) laboratory. With several learning al-

gorithms simulated with MATLAB R2018a software, three types of input parameters (9 in-

PAGE \* MERGEFORMAT 179


puts)were used: pressure, temperature, current, voltage, power, light, temperature(BMP), alti-

tude, and humidity. In another study, Kim, Jung, and Kang [80] used residential energy con-

sumption based on ANN prediction models. In modeling, residential building information,

and user features in South Korea. Their study discusses the nation’s influencing parameters,

namely, the number of exterior walls, direction of housing, housing area, number of years

occupied, number of household members, and occupation of the head of the household on

energy estimates. The results show the high accuracy of the ANN model in predicting energy

consumption. In a similar study by Bui, Nguyen, Ngo, and Nguyen-Xuan [81], building en-

ergy consumption, including heating and cooling loads, was estimated using the hybridization

of the ANN model with EFA. For building energy prediction, selecting inputs and grid struc-

tures that may be relevant is important to ensure generalization performance. Rivals and Per-

sonnaz [82] proposed a systematic approach based on least squares estimation and statistical

tests. This method obtained a simple neural network on the Internet for the application of an

energy prediction contest [83], which was used in this study for performance comparison

between the ANN and ANFIS models.

Historical Data PCA and  Computing Eigen Value ( PCA)


=1. Pressure
(All Features) BGA Model  Limitations
2. Temperature
 Looping prosess ( BGA)
(DHT22)
3. Electric current,
4. Voltage
Reduced
5. Power Features Data Driven
 ANN
Models
6. Luminous  ANFIS
intensity
7. Temperature
(BMP 180)
8. Altitude energy management for
9. Humidity efficiency in predicting energy
consumption in the Untirta
PVTE laboratory room

Figure 2 Data flow diagram of the PCA and BGA-based Energy consumption

prediction method

PAGE \* MERGEFORMAT 179


2.2 Feature Selection Using PCA ( Principal Component Analysis )

Feature selection that will be used in energy management for efficiency in predicting energy

consumption in the Untirta PVTE laboratory room is to use feature selection with PCA. The

input variables in predicting energy consumption in the laboratory room are obtained from

the activities that occur, which are taken using real sensors, which include pressure sensors,

temperature (DHT22), current, voltage, power, light intensity, temperature (BMP180),

altitude , and humidity. The data are processed using PCA to obtain a linear combination of

the initial variables. Geometrically, this linear combination is a new coordinate system

obtained from the rotation of the original system. The PCA method is very useful if the

existing data has many variables and a correlation between the variables. Principal

component analysis is based on the calculation of the eigenvalues and eigenvectors

representing the distribution of data from a dataset. The purpose of PCA is to select the

existing variables without losing the information contained in the original data. Using PCA,

the previous n variables are selected to become k new variables called principal components,

with the amount of k being less than n. Using only k principal components will produce the

same value using n variables. The variable resulting from the selection is called the principal

component. the new variables formed by PCA analysis later, in addition to having a smaller

number of variables and eliminating the correlation between the variables formed.

Technically, PCA is a multivariate (multivariable) data selection technique that changes or

transforms an original data matrix into a set of linear combinations that are fewer but absorb

most of the variance from the initial data. The main goal is to explain as much as possible the

number of variants of the original data with as few principal components as possible. The

principal component is a new set of variables that is a linear combination of the observed

variables. The principal component has a decreasing variance; most of the variation/diversity

of information in the observed set of variables tends to be collected in the first few principal

PAGE \* MERGEFORMAT 179


components, and less information from the original variable is collected in the last principal

component. This means that the principal component in the last sequence can be ignored

without losing much information. PCA can be used to select the variables being observed.

Principal components are orthogonal, which means that each principal component represents

all the original variables so that the principal component can be used as a substitute for the

original variable. energy management for efficiency in predicting energy consumption in the

Untirta PVTE laboratory room to reduce the dimensions of the data through a dataset

hypothesis of m variables is shown in Figure 3.

Figure 3 Conceptual model for the effectiveness of using PCA for the feature selection
stage [84]

PCA is used to explain the structure of the variance– covariance matrix of a set of variables

through a linear combination of these variables. In general, principal components are useful

for feature selection and the interpretation of variables. Principal component analysis is

usually used to:

1. Identify the new variable value that underlies the multiple variable data

PAGE \* MERGEFORMAT 179


2. Reduce the number of dimensions of the set of variable values which usually consists of

many correlated variable values by maintaining as much diversity as possible in the dataset

3. Eliminate the value of the original variable which has a relatively small contribution of

information.

Consider p variables consisting of n objects. From the p variables, there are k principal

components (with k <= p), which is a linear combination of these p variables. The k principal

components can replace the p variables without losing much information about the entire

variable. Generally, PCA is an intermediate analysis, which means that the results of the

principal component can be used for further analysis.

The mathematical form that Y is a linear combination of variables x1, x2, .., xp can be

expressed as follows:: [84]

P
Zm=∑ ( wi∗xi ) ............................................................................................... (1)
I

Where :

wi : weight or coefficient for variable i

xi : variable to –i

Zm : linear combination of the variables x

In Principal Component Analysis the method to get the coefficient values or weights from the

linear combination of the constituent variables is as follows [84]:

PAGE \* MERGEFORMAT 179


a. There are as many p principal components, that is, as many variables as observed and each

principal component is a linear combination of these variables

b. Each p principal component is mutually orthogonal (perpendicular) and independent of

each other.

c. Principal components are formed based on the order of variance from the largest to the

smallest. According to Johnson, if we get a vector x '= [x1, x2,..,xp] which has a covariance

matrix with a number of eigenvalues: 1 λ2 …≥ p 0, then we get a linear combination [85] as

follows :

Y 1=a1 x=a11 x 1+ a12 x 1 +…+a 1 p x p...............................................................(2)

Y 2=a2 x=a21 x 1 +a22 x1 +…+ a2 p x p

Y P=a P x=a P 1 x 1 +a 22 x 1+ …+a p 2 x p

where :

Yp= component to – i

a = eigenvector value

x = standard value of variable

The i component, Yp, is a linear combination of x1, x2,...,xp with weights, namely: ap1x1,…

ap2x2…,apixi, the selection must maximize the ratio of the first component variance (PC1) to

PAGE \* MERGEFORMAT 179


the number of variants ( total variance) of the original data. The next component, PC2, is also

a weighted linear combination of all the original variables, uncorrelated with the first

component (PC1), and must absorb the maximum remaining variance. To obtain the principal

component coefficients simultaneously, one of the following models can be used:

a. Eigenvalue and eigenvector decomposition of the correlation or covariance matrix of the

observed variables. In this case, the eigenvalue is the variance of each principal component,

and the eigenvector is the coefficient of the principal component.

b. The singular value decomposition of the n x n data matrix.

According to Jolliff, the Principal component analysis procedure for dimension reduction of

the random variable x (matrix of size n x n, where the rows contain n observations of the

random variable x) are as follows: [86]:

1. Calculating the covariance matrix from the observation data The covariance matrix is

a matrix in which the covariance values in each cell are obtained from the sample. Let

and y be random variables [86].

n
1
Cov ( A 1 , A 2 )= ∑ ( x −μ ) ( y −μ y ).................................................
n−1 i=1 i x i
(3)

With μ x and μ y being the average (mean) of the samples of the variables x and y,

where x i and y are the values of the i observation of the variables x and y. From the value data

used, a covariance matrix of size n x n is obtained.

Looking for eigenvalues and eigenvectors from the covariance matrix that has been obtained

[87] namely: Eigenvalues (λ) are scalar numbers and A is a matrix with size n x n to obtain

the value of n eigenvalues (λ1, λ2……..λn) then fulfill the following equation: namely:

PAGE \* MERGEFORMAT 179


Eigenvalues (λ) are scalar numbers and A is a matrix with size n x n to obtain the value of n

eigenvalues (λ1, 2……..λn) then fulfill the following equation:

Determinant (A – I) = 0

A = matrix nxn

λ = eigenvalue

I = identity matrix is a square matrix with the main diagonal element is 1, while the other

elements are zero

The eigenvalues and eigenvector define matrix A. The equation used to calculate the

eigenvector is as follows: eigenvalues and eigenvector can define matrix A. Equation to

calculate the eigenvector is:

Ax = λx ......................................................................................................... (4)

Ax – λx = 0

(A – λ) x = 0

(A – λI) x = 0, x ≠ 0

A = nxn matrix which has n eigenvalues (λn)

λ = eigenvalue

x = non-zero matrix

PAGE \* MERGEFORMAT 179


I = identity matrix

So that we get a linear combination that is:

a. λ1, λ2, λ3… λn is the eigenvalue of the matrix A

b. x1, x2, x3… xn is an eigenvector according to its eigenvalue (λn)

Eigenvalue & eigenvector equations are Eigen Value Decomposition (EVD), with the

following equation:

AX = XD ............................................................................................... (5)

A= X D X-1

A = an nxn matrix that has n eigenvalues (λn)

D = eigenvalue of its eigenvector

X = eigenvector of the matrix A

X-1 = inverse of eigenvector X

3. Determine the new variable (principal component) by multiplying the original variable by

the eigenvector matrix through the equation [86] :

Zki = U1k X1i + U2k X2i + …+ Upk Xpi ...................................................................... (6)

Zki = n x n matrix of the principal component with the coordinates of the i object at the k-th

position of the principal component

PAGE \* MERGEFORMAT 179


U = p x k matrix (eigenvector matrix)

X = matrix n x n (original variable)

4. Calculate the contribution of each feature with the following equation

m
c j=∑ |v pj|............................................................................................... (7)
p=1

Choose the largest number of c j values according to the number of features you want

to maintain, so that the j th feature is a significant feature.

Eigenvalue
Value of Proportionof Principal Component ( % )= x 100 %. (9)
Covariance variance

A number of selected features are then used as inputs in the energy consumption pre -

diction stage. The prediction of energy consumption is performed using two different algo-

rithms, namely the Artificial Neural Network (ANN) algorithm and the Adaptive Neuro

Fuzzy Inference System (ANFIS).

2.3 Prediction of Energy Consumption Using the Feature Selection Binary Genetic
Algorithm (BGA)

In addition to PCA, the feature selection used by BGA is the input and output of the

resulting feature selection in the form of binary numbers. The purpose of feature selection is

to obtain data that are initially large and are reduced to smaller ones [88][89][90].

A binary genetic algorithm is a method for finding solutions to problems and

determining the best solution. Genetic algorithms were first introduced by John H. Holland in

his book "Adaptation in Natural and Artificial Systems" in 1978 and Goldberg David E in

PAGE \* MERGEFORMAT 179


1989, who presented genetic algorithms with the theory of evolution in biology, especially

Darwin’s theory of evolution [91][92] [93].

The binary genetic algorithm optimization method is a heuristic search algorithm

based on the evolutionary mechanism of living organisms using binary numbers. The

principle of evolution is as follows: a problem is determined before the candidate for its

solution, and each candidate is referred to as an individual. An individual has a chromosome,

which is a set of genes. The gene is a component of the solution to the problem. This group of

individuals is known as a population. Each individual is then calculated as attractiveness,

which is known as the attractiveness value. Based on these values, there will be mutual

attraction between individuals (with a certain process, for example the roulette wheel

technique to choose a partner who will marry) so that the marriage process occurs. In this

marriage, there will be a process of gene exchange between the mating individuals. There are

two main operators in this process: crossover and mutation. From the mating process

produced offspring (spring). The children are then mated with the same steps as above, which

will produce quite good individuals. Otherwise, the opposite is produced.

The results of the binary genetic algorithm technique are not global optimum (ie

optimum for all search areas) but are not easily trapped in local optimum (ie optimum for

only certain areas). This technique is suitable for complex problems, and there is no

optimization technique. Binary genetic algorithm (BGA) is information that we can change,

which affects the behavior in the system, fitness, and the genes in the system. Genetics of

Fitness Behavior: What is Genetics's Value?

PAGE \* MERGEFORMAT 179


A genetic algorithm is a technique for finding an approximate optimization solution

in a search problem. Genetic algorithms are a special class of evolutionary algorithms that

use techniques inspired by evolutionary biology, such as heredity, mutation, natural selection,

and recombination. Genetic algorithms apply a method for moving from one population

(chromosome) to a new population using natural selection. The concept of chromosomes in

the genetic algorithm proposed by John Holland is used to express alternative solutions to a

problem. Each chromosome consists of a series of bit strings consisting of 0 and 1 bits called

genes, although it is also possible to use different chromosome coding [94] . Each

chromosome can undergo the exchange of genetic material between chromosomes. The

mutation process randomly changes gene values at several chromosome locations . Genetic

algorithms are different from conventional search techniques, in which the condition begins

with the initial setting of a random solution called a population. Each individual in the

population is called a chromosome, which represents a solution to a problem. A chromosome

is a string of symbols, which is generally a bit string of 0s and 1[95]. Chromosomes evolve

through continuous iterations, called generations. In each generation, chromosomes are

evaluated according to an evaluation function[96]. To produce the next generation, new

chromosomes, called offspring, are formed either by merging two chromosomes from the

initial generation using the crossover operator or by modifying the chromosomes using the

mutation operator. A new generation is formed through the selection process of several

parents (parents) and offspring (offspring), according to the fitness value, and through the

elimination of other chromosomes so that the population size remains constant. After several

generations, the algorithm can generate the best chromosomes that are expected to represent

the optimal or suboptimal solution to the problem. In general, the start process is assumed to

be random. Recombination generally involves crossover and mutation to produce offspring.

PAGE \* MERGEFORMAT 179


Actually, there are only two kinds of operations in genetic algorithms, namely: [97] 1.

Genetic operations: Crossovers and mutations. 2. Evolution operation: selection.

Binary Genetics Algorithm is almost the same process in doing the process to get the

most optimal value in a problem[95]. The steps are 1. Represent the problem in the form of

bit strings 0 and 1 as chromosomes that have a fixed length, and determine Pc (Crossover

Probability) and Pm (Mutation Probability). 2. Define a fitness function to measure

performance, which establishes the basis for selecting chromosomes to mate during

reproduction. Look for the fitness value, which is a measure of whether or not an individual is

good or whether the solution obtained is good or not. 3. Determine the initial population

generation process. This is usually done using random generation.4. Calculate the fitness

value of each chromosome. 5. Select a pair of chromosomes to combine in the population.

Parental chromosomes are selected with a high probability, and very matched chromosomes

have a high probability of being combined with less matched chromosomes. 6. Creating a

pair of offspring chromosomes by applying genetic operators (Crossover and mutation).7.

Creating offspring chromosomes (offspring) in the new population 8. Repeat step 5 until the

size of the new population of chromosomes is equal to that of the initial population. 9.

Replacing the parent chromosome population with a new offspring population. 10. Continue

step 4 and repeat the process until the desired criteria are reached. The cycle of the genetic

algorithm is shown in Figure 6.

PAGE \* MERGEFORMAT 179


Figure 6 Genetic Algorithm Cycle [98]

The first stage of the binary genetic algorithm is the start of the population, which

is described by encoding both binary and string chromosomes. The next step is to calculate

the fitness value based on the genes on the chromosomes in each population. Then, the

selection stage selects the chromosomes that are selected as parents. Furthermore, the

selected parent will undergo cross-breeding (Crossover). The crossover process plays a role

in forming offspring chromosomes, which also plays an important role in increasing the

diversity of strings in the population. Furthermore, the chromosomes enter the mutation

stage, which ensures that the diversity of chromosomes in a population is maintained to avoid

premature convergence.

2.3.1. General Structure of Binary Genetic Algorithm (BGA)

Genetic algorithms in general can be illustrated in the diagram in Figure 4.

Start

population initialization PAGE \* MERGEFORMAT 179


Figure 4 Genetik Algorithm Flowchart [99]

The genetic algorithm has characteristics that need to be known so that it can be

distinguished from other search procedures, namely:[92] 1. Coding scheme for the problem

solution based on the parameter set. Finding solutions from many individuals not just from

one individual 3. Genetic algorithms are based on the objective (fitness)of evaluating

individuals who have the best solution, not derivatives of a function. Binary genetic

algorithms use probability transition rules and not deterministic rules.

PAGE \* MERGEFORMAT 179


2.3.2. Chromosomal Coding

The first step in implementing a computer program is the selection of data types. This

is a variation of the initial differences between Holland’s original theory and several

alternative theories of genetic algorithms that later emerged. Holland decodes the

chromosomes into a string of binary digits. A number of binary encoding operations are used

to generate simple genetic algorithms; however, there are many ways to represent an object’s

genes, each of which has its own implicit advantages. To represent a problem in the form of

genes, the substances in the solution must be represented as a collection of units of

information [91]. Several types of chromosomal encoding techniques (encoding) can be

performed in binary genetic algorithms, including binary encoding, permutation encoding,

value encoding, and tree encoding. The encoding process discussed in this study is value

encoding. In value encoding, each gene on a chromosome is represented as a string of several

values. Values can be integers, real numbers, characters, or objects. The chromosomal coding

technique is shown in Figure 5 [100]. The process begins by initializing several individuals or

a population. Each individual is a solution or fitness value to be sought. Each individual is a

collection of genes called a chromosome.

Figure 5 Genes,Chromosomes, dan populatin [100]

PAGE \* MERGEFORMAT 179


2.3.3. Genetic Operators

Operators in genetic algorithms consist of several control parameters[101]. Population

size: defines how many chromosomes and how many genes in one chromosome are involved

during the search process. Crossover Probability: specifies the crossover probability between

two chromosomes. Mutation probability: specifies the probability that a bitwise mutation will

occur. Termination criteria: specify the conditions for ending the search for a solution in the

genetic algorithm.

2.3.4. Selection Process

After the chromosome coding process is performed in the genetic algorithm, the next

stage is the selection of offspring for the next generation. The purpose of the selection

process is to select the best chromosomes (individuals) in the population with the hope that

the new offspring will have a good fitness value.[93] . The selection process used in this

research is as follows: Truncation Selection Truncation or cutting is a selection method that

cuts the number of chromosomes in the population so that it can reduce the number of

individuals to be selected. This can speed up the process of selecting prospective parents.

Only the best individuals will be selected as parents. The parameter used is a threshold value

trunc, which indicates the size of the population to be selected as the parent, which ranges

from 50% to 10%. Individuals below this threshold will not produce offspring [102]. The

truncation method is shown in Figure 6. First, we increased the number of individuals in the

population. In this example, six are raised.

Initial Population Initial Population Initial Population


Kromosom Fitness Kromosom Fitness Kromosom Fitness
K1 25 K4 33 K4 33
K2 32 K2 32 K2 32
K3 28 K3 28 K3 28

PAGE \* MERGEFORMAT 179


K4 33 K1 25
K5 20 K6 22
K6 22 K5 20

Figure 6 Truncation Selection Method Source : [103]

Fitness Chromosomes A 10 B 5 C 15 D 4 E 5 Fitness Chromosomes Fitness Order C 15 1 A

10 2 B 5 3 E 5 4 D 4 5 2. These individuals are sorted by fitness value. 3. After sorting, the

truncation process is performed by selecting as many as 50% of the individuals in the

chromosomes with the best fitness values. 4. After going through the truncation process, the

individuals with the best fitness were obtained, namely, K4, K2, and K3.

The equation of the fitness function f(x) can be determined according to the equation:

Popsize
F= ∑ f (x ) .......................................................................................... (8)
i=1

To get how much influence Pi has on the fitness function f(x) by using the formula: To get

how much influence Pi has on the fitness function f(x) by using the formula:

f (x )i
Pi= Np
.................................................................................. (9)
∑ f (x)i
i=1

2.3.5. Crossover

Crossover is a genetic operator that combines two chromosomes (parent) to produce a new

chromosome (offspring) with probability crossover (Pc). In this study, the crossover method

used was a one-point crossover [104] [105]. In genetic algorithms, there are several types of

crossover among others. One-point crossover (One Point) This operator determines the

PAGE \* MERGEFORMAT 179


crossover point (certain gene). The first new chromosome contains the first gene to the

crossover point gene from the first parent chromosome plus genes from the crossover point to

the last gene from the second parent chromosome[67]. The second new chromosome contains

the first gene until the crossover point gene from the second parent is added to the gene from

the crossover point until the gene from the first parent chromosome [106]. The crossover

method is illustrated in Figure 7.

Figure 7 One-point crossover method

One point Parent 1: 1 2 3 4 5 | 6 7 8 9

Parent 2: 4 5 3 6 8 | 9 7 2 1

After the crossover process, the derivatives that can be generated from the two

parents above are:

Parent 1: 1 2 3 4 5 | 9 7 2 1

Parent 2: 4 5 3 6 8 | 6 7 8 9

Two-Point Crossover The process of two-point crossover is almost the same as that

of the one-point crossover procedure, except that in two-point crossover, two crossover points

must be selected and only genes between the two crossover points will be exchanged. [107].

The two-point crossover method is shown in Figure 8.

PAGE \* MERGEFORMAT 179


Figure 8 Two-point crossover method

Parent 1: 1 2 3 4 | 5 6 | 7 8 9

Parent 2: 4 5 3 6 | 8 9 | 7 2 1

After the crossover process, the derivatives that can be generated from the two parents above

are:

Parent 1: 1 2 3 4 | 8 9 | 7 8 9

Parent 2: 4 5 3 6 | 5 6 | 7 2 1

PMX (Partial Mapped Crossover) Invented by Goldberg and Lingle. PMX is a

modified formulation of two-point crossover. The highlight of PMX is the two-point

crossover and some additional procedures.

Parent 1 : 4 8 7| 3 6 5| 1 10 9 2

Parent 2 : 3 1 4| 2 7 9| 10 8 6 5

In (3,2), (6,7), (5,9) are swap places with mappings Parent 1 and Parent 2 where

swaps are (2,3), (7,6), (9.5). From the crossover results obtained offstring as follows.:

Child 1 : 4 8 6| 2 7 9| 1 10 5 3

Child 2 : 2 1 4| 3 6 5| 10 8 7 9

Uniform crossover operators Single- and two-point crossovers must determine the

crossover point before performing the crossover operation. Uniform crossover extends the

PAGE \* MERGEFORMAT 179


schema to each locus. In other words, each gene locus can be a crossover point. To achieve

this, binary strings of the same length as the chromosomes are first constructed randomly. If

any bit in this binary string has the value 1, the genes from both parents need to be exchanged

at the same position. After the exchange process, two children were created [108].

Shuffle crossover operator When shuffle crossover is applied, the genes of the two

chromosomes are randomly shuffled. Then, one-point crossover is applied to the randomized

chromosomes after selecting the crossover points at random, and two children are generated.

After recombination, the offspring’s genes are then shuffled in the same way that they were

previously shuffled [109]

Reducing the surrogate crossover operator Reducing the surrogate crossover limits the

crossover operation in cases where the parents have the same genes, i.e., the crossover point

can only be the point where the genes of the two parents differ. One position is chosen at

random after all points are found. Finally, a single point crossover is performed at this point

to create children [106]

2.3.6. Mutation

Mutation operators are used to alter some elements in selected individuals by

mutation probability (Pm) (mutation rate or mutation probability), leading to genetic diversity

that helps escape the trap of local optimal search processes[67]. There are several mutation

methods in the genetic algorithm as follows. Mutation Swap changes 2 (two) positions of n

randomly selected elements from permutations, and other elements remain in their respective

positions [110]. The swap mutation method is shown in Figure 9

PAGE \* MERGEFORMAT 179


Figure 9 Metode Swap Mutation [110]

Random mutation is performed by increasing or decreasing the value of the selected

gene with a small random number. 3. Uniform mutation: This mutation works by replacing

the value of the selected gene for the mutation with a random number that has a limit. The

limit is determined such that the resulting solution is legal. The uniform mutation process is

shown in Figure 10. Random mutation is performed by increasing or decreasing the value of

the selected gene with a small random number. Uniform mutation: This mutation works by

replacing the value of the gene selected for the mutation with a random number that has a

limit. The limit is determined such that the resulting solution is legal. The uniform mutation

process is shown in Figure 11.

Figure 10 Uniform mutation

Table 2 BGA Journal Study

No Study features Scope Methode Information

selection

1 [111] S. No countries ARIMA-ANFIS Prediction

PAGE \* MERGEFORMAT 179


Barak,dkk energy

2 [112] D. PCA Wind Power ANN Prediction

He,dkk Prediction energy

3 [7] J. C. Lam, PCA Building linear multiple Electric

K. K. W. regression consumption

Wan,dkk techniques model.

4 [13] K. Li, C. PCA Building iPSO-ANN Electricity

Hu,dkk campus shorter modeling consumption

time than GA-

ANN

5 [113] F. IG-PCA Building SVM,IBK,MLP Network

Salo,dkk security

6 [114] J. No Building Kohonen Self- Convenience

Yun,dkk Organizing Maps

(SOM) dan

algoritma k-

Means

7 [13] K. Li,dkk PCA Buiding iPSO-ANN lebih -

baik dari GA-

ANN dan ANN

PAGE \* MERGEFORMAT 179


8 [115]Y. CFS-BA Building Ensemble -

Zhou,dkk approach (C4.5,

RF, and Forest

PA)

9 [1] J. Cai,dkk PCA Building Linear regression, -

SVR,Regression

tree,Random

forest,KNN

[111] S. No Countries ARIMA-ANFIS Prediction


10 Barak,dkk energi

[112] D. PCA Wind Power ANN Prediction


11 He,dkk Prediction energy

[7] J. C. Lam, PCA Building linear multiple Electric


12 K. K. W. regression consumption
Wan,dkk techniques model.
[13] K. Li, C. PCA Building iPSO-ANN Electricity
13 Hu,dkk campus modeling time is consumption
shorter than GA-
ANN
[113] F. IG-PCA Building SVM,IBK,MLP Network
14 Salo,dkk security

[114] J. No Building Kohonen Self- Convenience


15 Yun,dkk Organizing Maps
(SOM) dan
algoritma k-
Means
[13] K. Li,dkk PCA Buiding iPSO-ANN lebih -
16

PAGE \* MERGEFORMAT 179


baik dari GA-
ANN dan ANN

[115]Y. CFS-BA Building ensemble -


17 Zhou,dkk approach (C4.5,
RF, and Forest
PA)
[1] J. Cai,dkk PCA Building Linear regression, -
18 SVR,Regression
tree,Random
forest,KNN
[1] J. Cai,dkk PCA Building Linear regression, -
19 SVR,Regression
tree,Random
forest,KNN
[116] Z. Hu,dkk Algoritma Power system MSVR-MA Energy
20 Memetic planning and consumption
Algoritma operation
Firefly
[5] Yousaf,dkk BGA+PCA Residential ANFIS Menejemen
21 energy

[6] A. BGA+PCA Residential ANFIS+SVR Menejemen


22 Yousaf,dkk energy

2.4. Energy Consumption Prediction Using Artificial Neural Network (ANN) Model

The ANN model maps the input data (feature selection results) at the input layer

to the output layer through the neurons in the hidden layer. An example of the ANN

architecture is shown in Fig. 11 as follows.

PAGE \* MERGEFORMAT 179


Figure 11 ANN Architecture

The input data are propagated forward and connected by input weights that have previously

been randomly initialized to the neurons in the hidden layer. In the hidden layer, the input

data associated with the weights are then processed using the activation function.

Furthermore, the processed data from the hidden layer are connected by hidden weights to the

neurons in the output layer.

The results obtained are then compared with the target data to obtain the error rate. If the

error rate obtained is smaller than the previously set error rate (target error), the propagation

process will stop. However, if the error rate is greater than the constant error rate, a back

propagation process is performed by updating the weights.

The following is the step of the backpropagation neural network algorithm. The input data are

propagated forward and connected by input weights that have previously been randomly

initialized to the neurons in the hidden layer. In the hidden layer, the input data associated

PAGE \* MERGEFORMAT 179


with the weights are then processed using the activation function. Furthermore, the processed

data from the hidden layer are connected by hidden weights to the neurons in the output

layer.

The results obtained are then compared with the target data to obtain the error rate. If the

error rate obtained is smaller than the previously set error rate (target error), the propagation

process will stop. However, if the error rate is greater than the constant error rate, a back

propagation process is performed by updating the weights.

The following is the step of the backpropagation neural network algorithm.

2.4.1 Initialize

Initialize all weights on the hidden and output layers and define the activation function in

each layer. Set the learning rate. Initialization of all weights can use random numbers in a

small range:

2.4.2 Activation

Activate network by applying input, x 1 ( p ) , x 2 ( p ) , … , x n ( p), and the expected output,,

y d 1 ( p ) , y d 2 ( p ) , … , y dn ( p).

a. Calculate the output obtained from the neurons in the hidden layer

n
v j ( p )=∑ x i ( p ) . w ij ( p).............................................................................................. (8)
i=1

1
y j ( p )= − v (p ) ............................................................................................................ (9)
1+e j

PAGE \* MERGEFORMAT 179


n is the number of inputs to neuron j in the hidden layer.

b. Calculate the output obtained from the neurons in the output layer:

m
v k ( p ) =∑ x j ( p ) . w jk ( p)............................................................................................ (10)
j =1

1
y k ( p )= −v ( p ) ........................................................................................................ (11)
1+e k

2.4.3 Update the weights

The weights are updated when the error is propagated back in the ANN, and the error is

returned according to the direction of the output signal.a. Calculate the error gradient for the

neurons in the output layer as follows:

e k ( p )= y dk ( p )− y k ( p)................................................................................................ ( 12 )

δ k ( p )= y k ( p ) ×[1− y k ( p ) ]× ek ( p).......................................................................... ( 13 )

Calculate weight correction::

∆ w jk ( p )=η × y j ( p)× δ k (p) ....................................................................................... ( 14 )

Update the weights on the output layer neurons::

w jk ( p+1 )=w jk ( p ) + ∆ w jk ( p)................................................................................... ( 15)

b. Calculate the error gradient for the neurons in the hidden layer::

PAGE \* MERGEFORMAT 179


l
δ j ( p )= y j ( p ) × [ 1− y j ( p ) ] + ∑ δ k ( p ) . w jk ( p)....................................................... ( 16 )
k=1

Calculate weight correction:

∆ wij ( p )=η × x i ( p)× δ j ( p).......................................................................................... ( 17 )

Update weights on hidden layer neurons

w ij ( p+1 )=wij ( p )+ ∆ w ij ( p)....................................................................................... (18)

2.4.4 Iteration

Increase one for iteration p, return to step 2, and repeat the process until the error

criterion is met. In the calculation of energy consumption, forward propagation is performed

using the best weights and biases from the network training stages. The resulting energy

consumption is compared with the actual energy consumption to obtain the accuracy of the

training and testing results.

2.5. Prediction of Energy Consumption Using the Adaptive Neuro Fuzzy Inference
System (ANFIS) Model

Feature selection results are also used as input in the neuro-fuzzy algorithm. This

algorithm is a combination of two systems: a fuzzy logic system and an artificial neural

network. The neuro-fuzzy system is based on a fuzzy inference system that is trained using a

learning algorithm derived from an artificial neural network system. Thus, the neuro-fuzzy

system has all the advantages of fuzzy inference and artificial neural networks. From their

ability to learn, neuro-fuzzy systems are often referred to as adaptive neuro-fuzzy inference

systems. One form of structure that is well known is shown in Figure 11.

PAGE \* MERGEFORMAT 179


Figure 12 Structure of ANFIS

In this structure, the fuzzy inference system is the Takagi–Sugeno–Kang fuzzy inference

model

In the neuro-fuzzy system, there are five layers of processes in which the functions and

equations of each layer are explained as follows:

2.5.1. Fuzzification Layer

Let O1 ,i be the output of each node in layer 1. Each node i in this layer is an adaptive node

with the function node O1 ,i=μ A i( x)) for i=1 , 2; or O1 ,i=μ Bi ( y )) for i=1,2, where x is the

input to nodei and Ai is the linguistic label (small, large, etc.) that corresponds to the

function of this node. On the other hand O1 ,i is a membership function of A 1 and the degree

of membership is specific for a given x which is sufficient to quantify Ai . The most widely

used membership functions are the Bell and Gaussian forms. The Bell form membership

function is expressed by:

PAGE \* MERGEFORMAT 179


1
f ( x , a , b , c )= 2b
x−c ..........................................................................................................
1+
a

(19)

with parameter b usually positive. Parameter c is located in the middle of the curve. The

Gaussian membership function is expressed by, with parameter b usually being positive.

Parameter c is located in the middle of the curve. The Gaussian membership function is

expressed as follows:

2
( x−c )

A ( x )=e 2a
2
.................................................................................................................... (20)

with parameter b usually positive. Parameter c is located in the middle of the curve. The

Gaussian membership function is expressed as follows:

2.5.2. Product Layer

Each node in this layer consists of the prod t-norm operator as a node function. This layer

synthesizes the transmission of information using layer 1, multiplies all incoming signals, and

sends the product out. The output of the product layer is expressed as follows:

O2 ,i=μ A i( x). μ Bi ( y )=w i……………………………………………………… (21)

Each node in this layer serves as a measure of the strength of the rule. The output in this layer

acts as a function of the weight.

PAGE \* MERGEFORMAT 179


2.5.3. Normalization Layer

Each node in this layer normalizes the weight function obtained from the previous product

layer. The normalized output is calculated by:

wi
O3 ,i = ………………………………….…………………………………….. (22)
w1 +w 2

2.5.4. Defuzzification Layer

Nodes in this layer are naturally adaptive. The defuzzification output of this layer is

calculated by the formula:

O4 ,i=O3 ,i (α 4 ,i=O3 , i (α i x)+ β i y+ γ i) .................................................................. (23)

where α i, β i and γ i are linear parameters of the consequent suitability of node i.

2.5.4. Total Output

A single node in this layer synthesizes the information transmitted with layer 4 and returns

the entire output using the following fixed function:

Σ wi y i
O5 ,i = ………………………………………………………………………… (24)
Σ

Similar to the calculation of energy consumption using the ANN algorithm, the ANFIS

algorithm performs forward propagation using the best weights and biases from the network

training stages. The resulting energy consumption is compared with the actual energy

consumption to obtain the accuracy of the training and testing results.

PAGE \* MERGEFORMAT 179


2.4 Summary

From the references that have been discussed in chapter 2, it can be concluded that the

research steps.

The data used as a reference in this study were obtained from research conducted for

approximately three months, namely January– April 2021. Data In this study, we used 3350

with 2450 training data, 833 test data, and 67 predictive data. Data are collected on the basis

of recorded data using a microcontroller with a sensor input with 9 input variables. Pressure,

temperature I, current, voltage, power, light, temperature II, altitude, and humidity.

Parameter Input

PCA feature
selection

BGA feature
selection

Stage 1
Stage 2
Prediction of energy
Prediction of energy
consumption using
consumption using
the ANFIS
the ANN algorithm
algorithm

Output PAGE \* MERGEFORMAT 179


• The application of the looping process to the
selection of features combining PCA and BGA
predicts energy consumption.
Figure 13. PCA+BGA Research Contribution

Before processing the data obtained during the research in the electrical engineering

vocational education laboratory (PVTE). The main purpose of feature selection here is that

the reduction function reduces the number of variables that were initially vast to fewer to

facilitate analysis at a later stage. The transformation function is used to change the variables

that were initially correlated to become uncorrelated.

Stage 1

After obtaining the looping process, feature selection is processed using artificial intelligence

using the ANFIS, PCA+ANFIS, PCA+BGA+ANFIS models.

Stage 2

This stage is almost the same as stage 1, the difference being that the model used is an ANN.

• Application of the looping process to the selection of features that combine PCA and BGA

to predict energy consumption.

PAGE \* MERGEFORMAT 179


• Evaluation of model performance by comparing the actual energy data with the predicted

data from the ANN, PCA+ANN, PCA+BGA+ANN and ANFIS, PCA+ANFIS,

PCA+BGA+ANFIS models.

A number of selected features are then used as inputs in the energy consumption prediction

stage. The prediction of energy consumption is performed using two different algorithms,

namely the Artificial Neural Network (ANN) algorithm and the Adaptive Neuro Fuzzy

Inference System (ANFIS).

Figure 14 Research framework

From Figure 14, it can be seen how the steps that will be taken when the research focuses

on grouping data on combining PCA and BGA features selection. The purpose of grouping

data on feature selection on PCA is to speed up data computing in predicting energy in a

room using machine learning (ML). The grouping of data on the PCA feature selections a

PAGE \* MERGEFORMAT 179


grouping by combining the value of PC with a value of 1 into PC1 and PC2 with a value of 0,

but the looping process in BGA with data grouping will determine the value of 1 and if BGA

has a logic 0 then the data will stop when logic 1 is found. After the data have stopped when

the looping process is optimal, the data are tested with the ANN and ANFIS models.

The BGA algorithm is a binary genetics algorithm. The input and output are binary

numbers used to find the most optimal solution to an optimization problem. In the

implementation of how to predict energy consumption in a laboratory by applying the PCA

and BGA features selection by grouping the data before the data is trained in the training and

testing process. The initial stage of this research is to select the input variable by grouping the

data based on the analysis component formed from the input variable as many as nine

variables. The data of the nine variables is reduced to PC1,PC2……..PC9, then performs the

transformation by selecting the PC value that value of one which will be chosen to influence

the prediction results using the ANN and ANFIS models according to the block diagram in

Figure 15.

PCA feature
selection
1. Pressure
2. Temperature
(DHT22) Dimension reduction
3. Electric current, TransformationP
PC1,PC2,PC3,PC3,PC
4. Voltage C1,PC2,PC3,PC
4,PC5,PC6,PC7,PC8,P
5. Power 4
C9
6. Luminous intensity
7. Temperature (BMP
180)
8. Altitude BGA feature
9. Humidity ANN Model and selection
ANFIS Model 1011

Figure 15 Block diagram of PCA aplication to the ANFIS model [6][5]

PAGE \* MERGEFORMAT 179


Implement the prediction of energy consumption in the room with a new one with the

model that I will propose by looking at the characteristics of the PCA model which has a lack

of data similarity level by choosing a number that has a value of 1 and a number below 0 is

not continued to the next process with experiments that have been carried out, it can be

concluded that only PC1, PC2, PC3, and PC4 are selected to enter the second feature

selection stage using BGA. `Only 2 components will be formed, namely PC1 and PC2, where

PC1 is the result of data grouping which is worth one while PC2 contains a value of 0.

However, there are shortcomings in the PCA method.

1. Each principal component (PC) is a linear combination of all variables. In other words,

each PC is a linear combination of all variables with a load assigned to each variable. The

resulting load value is usually not zero. These results in the PC results obtained will be

difficult to interpret.

2. PCA is not good at separating classes; therefore, accuracy decreases.

3. Requires more data for training and several data samples.

4. The dimension of the feature vector space is higher.

PAGE \* MERGEFORMAT 179


Genetic algorithms are a branch of evolutionary algorithms that are optimization

techniques based on natural genetics. In the genetic algorithm used to produce an optimal

solution, the search process is performed among several alternative optimal points based

on a probabilistic function [117].

Genetic algorithms with binary coding are commonly used in function optimization

problems. The weakness of binary coding is that the solution range is in a continuous

region. [118]. Early convergence often occurs because the solution population is trapped

in the local optimum. Distributed genetic algorithms (DGA) use several subpopulations,

and each subpopulation uses operators. The advantages of binary genetic algorithm

(BGA) compared with other search algorithm areas are as follows:

1. This algorithm performs only a few mathematical calculations related to the problem to

be solved. Because of the changing nature of natural evolution, this algorithm seeks a

solution without regard to the processes associated with the problem being solved

directly. This algorithm can also control the objective function and defined constraints in

either a discrete or analog search space.

2. Evolution operators make this algorithm very effective in global search.

3. This algorithm has high flexibility and can be combined with other search methods to

be more effective.

The weakness of genetic algorithms when compared with other search algorithms is the

uncertainty in producing a global optimum solution, because most of these algorithms

deal with probabilistic random numbers. The results of this data grouping will be

processed using the BGA algorithm with only two variables. The data to be processed are

PAGE \* MERGEFORMAT 179


two PCs that have the largest eigenvalues after which the data are processed using the

ANN or ANFIS model. By looking at the characteristics of PCA where the data grouping

is PC1 to PC9, then the input parameters are determined based on the eigenvalue and

eigenvector values where PC1 is the highest value of the formation PC2 to PC9 is the

lowest value, while BGA with its characteristics results obtained by setting 50% of the

data then based on the characteristics of BGA. [6][5]

PCA feature
selection
Dimension reduction
1. Pressure
PC1 and PC2 values Transformation
2. Temperature <=1 PC1 and PC3
(DHT22) If PC2 value = 0, then values >=1
3. Electric current, PC3<=1
4. Voltage
5. Power
6. Luminous
intensity BGA feature
7. Temperature ANN Model and selection
(BMP 180)
8. Altitude ANFIS Model PC1 and PC3

Figure 16 Block Diagram of PCA application to ANN and ANFIS models

From the model I propose to reduce the training time for the desired target is:

Selection of 4 variable features


using PCA

.
Dimension reduction
PC1 > PC2 PAGE \* MERGEFORMAT 179
If PC2 value = 0, and PC3 value = 1
Transformation
PC1 and PC3 value =1

No
BGA termination
PC1 and PC3

Yes

ANN Models and Models


ANFIS
Figure 17 Flow chart combining PCA and BGA as a feature selection

From Figure 19, it can be seen that the feature selection process uses PCA. Then,

the selection process from the BGA which consists of selection, crossover, mutation, and

regeneration, is replaced by using the PCA stages of dimension reduction and transformation

while testing the validity of the data using the process from BGA if the selection data is

appropriate. with the desired target; if not, the selection conditions will loop until the target

conditions match the energy prediction data to be achieved. The selection condition will stop

when it matches the desired target, and then the next process will continue testing with the

ANFIS and ANN models.

I will focus on the looping process resulting from the merging of PCA and BGA.

This flow looping process improves the computations in the training process without

considering the results of the PCA and BGA merging models. However, previous studies

PAGE \* MERGEFORMAT 179


have also carried out parallel merging of genetic algorithms[119]. The application of feature

selection for classification has also been performed with CMIM, and BGA has also been

performed to speed up computing [15]; however, this research does not explain how much

computational speed is generated with the model being tested.

PAGE \* MERGEFORMAT 179


CHAPTER 3

METHODOLOGY

3.1 Introduction

The theoretical review in chapter 2 discussed how the features selection that have existed

in previous studies have been discussed. The application of feature selection in energy

management in a building has not been widely applied, especially in the prediction of

energy use in a room. The application of feature selection in buildings has also been

widely applied in an area or city. In this case, there has been no application of feature

selection that has contributed to this dissertation research. The application of the previous

feature selection is also discussed, and how the weaknesses and shortcomings of previous

research have become the basis for this dissertation. Finally, at the end of chapter 2,

examples of several methods used in this dissertation are provided.

The data used in this study comprised training and testing data. This dataset has several

proportional attributes and a missing value. This dataset is divided into 80% as training

data, 20% as prediction data, and 10% as prediction data. In this study, we used 3283 with

2450 training data, 833 test data, and 67 predictive data. The data were collected based on

real retrieval in the PVTE UNTIRTA laboratory for 3 months. The first retrieval is from

January 06, 2021 to April 10, 2021. Data collection is based on data read by sensors

connected to a microcontroller. by using a data recording tool consisting of several

supporting tools for data recording. In addition, a PC/laptop is required where the data will

PAGE \* MERGEFORMAT 179


be recorded and displayed on the Excel PLX-DAQ software. In this trial, it was performed

with a delay of 10 min.

Recorded variables such as pressure, current, voltage, power, light, humidity, temperature,

and altitude. Recording is only performed on weekdays and operating hours because the

laboratory room is only active on weekdays and operating hours. Starting Monday– Friday,

09.00 to 17.00 pm. Prediction of electrical energy consumption in the laboratory where this

dissertation only focuses on the features selection used when managing energy in a

building or more, especially in this research, was carried out in the UNTIRTA PVTE

laboratory, with the help of Matlab R2018a. The working procedure of this research can be

seen as a whole in Figure 17.

Output: Testing
Input : PCA feature BGA feature
with ANN and
Dataset selection selection
ANFIS Models

Figure 18 Stages of Research Image

3.2 Research Design

In this study, several processes were conducted to solve the research problems. The

procedure for solving these problems is shown in Figure 18. In this study, there are 2

stages that will be performed before testing with the tested model, namely ANN and

ANFIS. In stage 1, the dataset is completed using feature selection using PCA Figure 3.2.

in stage 2, completion using feature selection using BGA. In stage 1, nine variables were

PAGE \* MERGEFORMAT 179


used: pressure, temperature, current, voltage, power, light, temperature (BMP), altitude,

and humidity.

Real Measurement Data processing Traning : Prediction :


Data from 9 sensors Training : 2450 2450 67
( mxn ) Testing Data : 833
m(Number of input Prediction 67
variables )
n(Number of data) Calculate
Data Testing : Correlation
833

Determine the
Eigenvalues
(Eigenvalues)

Calculating
Attribute Weights
using
Eigenvectors

Choose the attribute


with the highest
eigenvector value
weight >=1

Data After attribute


selection from 9
measurement sensors

Figure 19 PCA feature selection procedure

PAGE \* MERGEFORMAT 179


3.2.1 PCA feature selection procedure

3.2.1.1 Data Input

In this study, using Metlab 2018a, the attributes of the preprocessing results are repre-

sented in the form of labels (x1, x2, x3…x9), which represent the sequence of data that

corresponds to the dataset being tested.

The following is the result of attribute representation in the form of labels:

X1 = Pressure (atm)

X2 = Temp@ ( DHT22 )

X3 = Current (A)

X4 = Voltage ( V )

X5 = Power ( W )

X6 = Light ( Lux )

X7 = Temp© (BMP180)

X8 = Altitude (m)

X9 = Humidity(%)

3.2.1.2 Calculating the correlation between observational data

Covariance is used to measure the magnitude of the relationship between the two

attributes. The types of relationships that can occur between two attributes based on their

covariance values are: a. Positive: if the covariance value is positive or > 0 b. Negative: the

covariance value is negative or < 0 c. Zero : if the covariance value is zero or = 0 .

Therefore, to obtain the covariance value, equation 3. Next is the covariance matrix

PAGE \* MERGEFORMAT 179


obtained by counting the pairs of 2 attributes from a total of n attributes. For example:

attributes x, y, and z, with the number of attributes = 3,

Then the covariance matrix is 3x3 = nxn, namely:

Where cov(x,x), cov(y,y) and cov(z,z) the value is the same as the calculation of the

variance value of the x, y and z attributes. so the covariance matrix is:

If the value of the diagonal component (var(x), var(y), and var(z)) is 1, then the covariance

matrix is the same as the correlation matrix because the correlation result from the non-

standardized data is the same as the covariance result from the standardized data. This

means that the covariance matrix is a symmetrical. When decomposed with Eigenvalue,

then the value on the diagonal eigenvalue 0. Thus, the value:

cov(y,x) = cov(x,y)

PAGE \* MERGEFORMAT 179


cov(z,x) = cov(x,z)

cov(z,y) = cov(y,z)

3.2.1.3 Determine the eigenvalue of the covariance matrix

The eigenvalue (λ ) is a scalar number that defines the covariance matrix, which is a square

matrix measuring m m. Then, the eigenvalue (λ) corresponding to the covariance matrix is

obtained through equation 4, so that each scalar (λ1, 2……..m) satisfies equation 5 to form

an eigenvector matrix.

3.2.1.4 Selecting Principal Components (PCs)

Principal component analysis (PCA) is an analytical technique that transforms the original

attributes that are still correlated with one another into a new set of uncorrelated attributes.

In this study, the selected principal component has the maximum contribution amount with

a value of 99% (variance threshold) based on the proportion of variance of each selected

principal component because this value is sufficient to explain the total covariance

variance of the original attribute. the proportion of variance of each principal component is

obtained through equation 9, where the covariance variance is obtained from the sum of

the total values of the diagonal covariance matrix.

3.2.1.5 Calculating attribute weights using the eigenvector equation

To determine which attributes are included in the principal component (attributes that

greatly affect the covariance variance of the observation data) with a maximum

PAGE \* MERGEFORMAT 179


contribution of 99%, the eigenvector value weight calculation is performed using equation

9, where each x vector value formed corresponds to one eigenvalue for each attribute.

3.2.1.6 Choose the original attribute based on the weight of the highest eigenvector value.

In this study, the interpretation of the results of feature selection using PCA is obtained by

selecting the attribute that has the highest weighted eigenvector value from several

principal components that have a fairly large correlation with the formation of the original

attribute with a total proportion of the variance covariance of 99%. Thus, many attributes

are obtained from the attribute selection from the dataset of 9 input variables . Other

attributes that have low value weights do not have enough influence on the formation of

the original attributes of each data taken in real time at the PVTE Laboratory. It can be

concluded that this selected attribute is modeled into the BGA feature selection

Processing results with PCA


input feature selection BGA

Population initialization

Evaluation with RSME,


MSE and MAPE

Roulette Wheel Selection

Crossover

Mutation

PAGE \* MERGEFORMAT 179


New generation

Search
Termination?

Optimal fitness
results

Finish

Figure 20 BGA feature selection procedure.

3.2.2. The data used by BGA feature selection

The data used for feature selection use PCA BGA results, which are formed on the basis of

the eigenvalues that have the highest values. The data with the highest value are grouped

into PC1, PC2, PC3, and PC4 on the basis of the latent value obtained from Metlab 2018a.

Steps in the BGA feature selection with the help of the supporting software, namely

Matlab R2018a. At this stage, several steps are performed

a. The initial step of this GA method is to determine the initial population with fitness

values from numerous populations and then represent the genes and chromosomes

of the population in the form of variable data.

b. Determine the fitness probability value of some fitness measures over several

generations and determine the fitness function. Then, the value is evaluated.

c. The next step is to select the population. A good or optimal value is obtained.

PAGE \* MERGEFORMAT 179


d. The crossover step involves two chromosomes or the most dominant value to form

a new value. Not all values in a population will undergo this process.

e. The next step is mutation. This mutation technique plays a role in returning the

optimal solution that can be lost due to the previous crossover process.

Figure 21 Matlab R2018a Support Software

3.3 Proposed Methodology

This research study on how to manage energy in a building, especially in a building

using machine learning (ML), focuses on how to apply feature selection to speed up data

computing during the training and testing process. The feature selection observed based on

the references used were PCA and BGA. The data used as a reference in this study

wereobtained for approximately three months, namely January– April 2021, from the

PVTE laboratory. 3350 with 2450 training data, 833 test data, and 67 predictive data will

be used. Data sharing is based on the journal Adedeji et al. [33] [120] training data using

70% and 30% testing [121]. Other studies also discuss about 60% and 40% [122]. Other

studies 80% and 10% and predictions 10% [123]. Besides training and testing, there are

also those who use testing with a larger sample than training [124] Data are collected on
PAGE \* MERGEFORMAT 179
the basis of recorded data using a microcontroller with a sensor input with 9 input

variables. Pressure, temperature I, current, voltage, power, light, temperature II, altitude,

and humidity. Before processing the data obtained during the research in the electrical

engineering vocational education laboratory (PVTE). The main purpose of feature

selection here is that the reduction function reduces the number of variables that were

initially vast to fewer to facilitate analysis at a later stage. The transformation function is

used to change the variables that were initially correlated to become uncorrelated. There

are two stages in data collection. The first stage is to compare the data when the PCA and

BGA features are selected. The second stage involves data that do not use the selection

feature. Model used to test the ANN and ANFIS.

The BGA algorithm is a binary genetics algorithm. The input and output are binary

numbers used to find the most optimal solution to an optimization problem. In the

implementation of how to predict energy consumption in a laboratory by applying the PCA

and BGA features selection by grouping the data before the data is trained in the training

and testing process. The initial stage of this research is to select the input variable by

grouping the data based on the analysis component formed from the input variable as many

as nine variables. The data of the nine variables is reduced to PC1,PC2……..PC9, then

performs the transformation by selecting the PC value that value of one which will be

chosen to influence the prediction results using the ANN and ANFIS models according to

the block diagram in Figure 16.

PAGE \* MERGEFORMAT 179


PCA selection
feature
10. Pressure
11. Temperature
(DHT22)
Dimension reduction
12. Electric current,
13. Voltage
PC1,PC2,PC3,PC3,PC Transformation
14. Power 4,PC5,PC6,PC7,PC8,P PC1,PC2,PC3,PC4
15. Luminous intensity C9
16. Temperature (BMP
180)
17. Altitude
ANN Model and BGA selection
18. Humidity
ANFIS Model feature
1011

Figure 22 Block diagram of PCA aplication to the ANFIS model [6][5]

Implement energy management to predict energy consumption in a room using an

updated model based on the combination of PCA and BGA feature selection. By looking at

the characteristics of the PCA model, which has a similar level of lack of data by choosing

a number that has a value of 1 and a number below 0 does not proceed to the next process

with the experiments that have been carried out, it can be concluded that only PC1, PC2,

PC3, and PC4 were selected to enter the next stage. second feature selection using BGA.

Only 2 components will be formed, namely PC1 and PC2, where PC1 is the result of

grouping data with a value of one, while PC2 contains a value of 0. However, there are

shortcomings in the PCA method, namely

1. Each principal component (PC) is a linear combination of all variables. In other

words, each PC is a linear combination of all variables with a load assigned to each
PAGE \* MERGEFORMAT 179
variable. The resulting load value is usually not zero. These results in the PC results

obtained will be difficult to interpret.

2. PCA is not good at separating classes; therefore, accuracy decreases.

3. Requires more data for training and several data samples.

4. The dimension of the feature vector space is higher.

Genetic algorithms are a branch of evolutionary algorithms that are optimization

techniques based on natural genetics. In the genetic algorithm used to produce an optimal

solution, the search process is performed among several alternative optimal points based

on a probabilistic function [117].

Genetic algorithms with binary coding are commonly used in function optimization

problems. The weakness of binary coding is that the solution range is in a continuous

region. [118]. A weakness of the genetic algorithm is that early convergence often occurs

because the solution population is trapped in the local optimum. Advantages of the Binary

Genetic Algorithm:

1. This algorithm performs only a few mathematical calculations related to the

problem to be solved. Because of the changing nature of natural evolution, this algorithm

seeks a solution without regard to the processes associated with the problem being solved

directly. This algorithm can also control the objective function and defined constraints in

either a discrete or analog search space.

PAGE \* MERGEFORMAT 179


2. Evolution operators make this algorithm very effective in global search.

3. This algorithm has high flexibility and can be combined with other search

methods to be more effective.

The weakness of genetic algorithms when compared with other search algorithms

is the uncertainty in producing a global optimum solution, because most of these

algorithms deal with probabilistic random numbers. The results of this data grouping will

be processed using the BGA algorithm with only two variables. The data to be processed

are two PCs that have the largest eigenvalues after which the data are processed using the

ANN or ANFIS model. By looking at the characteristics of PCA where the data grouping

is PC1 to PC9, then the input parameters are determined based on the eigenvalue and

eigenvector values where PC1 is the highest value of the formation PC2 to PC9 is the

lowest value, while BGA with its characteristics results obtained by setting 50% of the data

then based on the characteristics of BGA. [6][5].

PCA
selection
Dimension reduction Transformatio
feature
1. Pressure PC1 and PC2 values n
2. Temperatur <=1 PC1 and PC3
e (DHT22)
3. Electric
If PC2 value = 0, then values >=1
current, PC3<=1
4. Voltage
5. Power
6. Luminous
intensity
BGA feature
7. Temperatur ANN Model and
e (BMP selection
ANFIS Model
180) PC1 and PC3
8. Altitude
9. Humidity

PAGE \* MERGEFORMAT 179


Figure 23 Block Diagram of PCA application to ANN and ANFIS models

From the model 1 propose to reduce the training time in the desired target is:

Selection of 4 variable features


using PCA

Dimension reduction
PC1 > PC2
If PC2 value = 0, and PC3 value = 1

Transformation
PC1 and PC3 value =1
.

BGA termination
PC1 and PC3
No

Yes
ANN Models and Models
ANFIS

Figure 24 Flow chart combining PCA and BGA as a selection feature

From Figure 24, it can be seen that the feature selection process uses PCA. Then,

the selection process from the BGA which consists of selection, crossover, mutation, and

regeneration, is replaced by using the PCA stages of dimension reduction and

transformation while testing the validity of the data using the process from BGA if the

selection data is appropriate. with the desired target; if not, the selection conditions will
PAGE \* MERGEFORMAT 179
loop until the target conditions match the energy prediction data to be achieved. The

selection condition will stop when it matches the desired target, and then the next process

will continue testing with the ANFIS and ANN models.

3.4. Limitation of Proposed Methodology

PAGE \* MERGEFORMAT 179


No Task 2018 2019 2020 2021

2-4 5-7 8-10 11-12 1-3 4-6 7-9 10-12 1-3 4-6 7-9 10-12 1-2

1 Study of
literature
Sensor Design
2 AC Voltage,
WCS1700 for
measuring AC
Current Sensor,
DC Voltage
module, ACS712
Sensor for
measuring
current sensor
DC, dan modul
PAGE \* MERGEFORMAT 179
BH1750

3 Research data
collection
4 Research data
processing
Publications,
3 patents and
dissertations

PAGE \* MERGEFORMAT 179


3.5 Summary

In this research, I will focus on the looping process between combining PCA

and BGA feature selection to reduce input parameters and speed up time when con-

ducting training and testing. PCA feature selection is the grouping of data based on

eigenvalues and eigenfactors that affect data grouping. The function of feature se-

lection in PCA is to reduce the input variables and data computation time during the

training and testing process. As discussed above, the input variable consists of 9

input variables that will be reduced and transformed so that only PC1, PC2, PC3,

and PC4 will be processed according to the data grouping value of 1, while the data

grouping value below 0 will not be processed. After the data are processed with

PCA feature selection , the data are processed using BGA feature selection . The

looping process that will occur in the training and testing process makes the process

only binary numbers 1 and 0, which will be processed to the next stage. based on

the journals discussed above [6][5].

PAGE \* MERGEFORMAT 179


CHAPTER 4

RESULTS AND DISCUSSION

4.1 Introduction

In this chapter, we explain PCA – BGA feature selection. Testing using 3350 data

points with 2450 training data, 833 testing data, and 67 prediction data. The distri-

bution of data based on the journal Adedeji et al. [33] [120] training data using 70%

and 30% testing [121][33] other studies also discuss about 60% and 40%[122].

Other studies also discuss 60% and 40%[123] in this study used 70% training, 20%

testing, and 10% prediction data.

The main contribution of this dissertation is presented in the previous chapter. Con-

sidering that the combination of PCA and BGA feature selection aims to increase

the speed of data computing when the training and testing process is carried out, the

advantages and limitations that characterize the selection of PCA and BGA features.

previous feature selection ability[6][5] BGA and PCA feature selection.

4.2. BGA and PCA feature selection.

This feature extraction process considers the characteristics of the training and test-

ing data. The difference between taking features from the training data and testing

data is the feature taking in the training data and the distribution of data according

to data collection for approximately 3 months. The choice of PCA features relates to

the explanation of the structure of the variance and covariance of variable clusters

PAGE \* MERGEFORMAT 179


in linear combinations of these variables. The main goal is to reduce data interpreta-

tion without losing the original data. If we have some p The variable is then con-

verted into k variable, where is the expected component k smaller than component p

( k<p ) variable . In selecting PCA features, there are several steps that need to be

performed:

4.2.1. Data standardization

Data normalization or standardization is required to eliminate the unit effect during

data collection. Standardization is required because it greatly affects the covariance

matrix [48][125] standardization results for training data 2450. Suppose attributes

A1, A2, A3, A4, A3, A4, A5, A6, A7, A8, A9.. Amount of data n = 2450.

The variance of each attribute can be determined using the equation:

n
2
∑ ( X 1−X )
var ( A 1 )= i=1
( n−1 )

Attribute searched until var ( A 9 ) according to table 5 standardization is used to get

the variance of each attribute

Table 3 Data Standardization (calculate correlation)

No A1 A2 … A9

1 0.50521656 -0.17815 … 1.495913

PAGE \* MERGEFORMAT 179


1

0.50521656

2 1 -0.17815 … 1.485819

0.56244173

3 8 -0.17815 … 1.506006

0.56244173

4 8 -0.17815 … 1.506006

0.50521656

5 1 -0.17815 … 1.495913

0.56244173

6 8 -0.17815 … 1.5161

0.56244173

7 8 -0.17815 … 1.485819

0.50521656

8 1 -0.17815 … 1.485819

0.50521656

9 1 -0.17815 … 1.495913

PAGE \* MERGEFORMAT 179


0.56244173

10 8 -0.17815 … 1.485819

2450 -0.12426038 -0.27912 … -1.45153

… … … … …

2.06102E-

total 09 7.49E-11 … -4.5E-12

Averat 8.41231E-

e 13 3.06E-14 … -1.8E-15

4.2.2. Calculating the Covariance Matrix

To get the covariance of two attributes

∑ ( X 1− X ) ( Y 1−Y )
var ( A 1 , A2 ) = i=1
( n−1 )

To calculate the correlation value between attributes in the laboratory

data collection dataset, equation 2.7 is used. Covariance matrix based on the

coefficient matrix of the input variable [112]. The following is the calculation of

the covariance value between attributes A1 and A2:

PAGE \* MERGEFORMAT 179


Cov(A1,A2) =

1
( 0.505216561−8.41231E-13 )(−0.178149221−3.05657E-14 )+ …=−0.160532037
2450−1

To get a more complete covariance value can be seen from table

4.4.Calculation of the Covariance value on the diagonal attribute A1

1
Cov(A1,A1) = ((0.505216561)2 +( 0.505216561)2+….+( -
2450−1

0.124260384)2=1

The calculation of the Covariance value on the diagonal attribute

A2,A3.A4.A5.A6,A7,A8,A9 can be seen from table 6

Next, we enter the results of the calculation of the covariance value in each

attribute pair into the form of a covariance matrix (Covariance Matrix)

measuring 9x9 where the value of Cov(A1,A2) is the same as Cov(A2,A1), the

value of Cov(A1,A3) is the same as that of Cov (A3,A1), and so on in the same

way also applies to each attribute pair. The following is the diagonal form of

the covariance matrix for each attribute pair:


Cov(A1,A1 Cov(A1,A3 Cov(A1,A5 Cov(A1,A7
) Cov(A1,A2) ) Cov(A1,A4) ) Cov(A1,A6) ) Cov(A1,A8) Cov(A1,A9)
Cov(A2,A1 Cov(A2,A3 Cov(A2,A5 Cov(A2,A7
) Cov(A2,A2) ) Cov(A2,A4) ) Cov(A2,A6) ) Cov(A2,A8) Cov(A2,A9)
Cov(A3,A1 Cov(A3,A3 Cov(A3,A5 Cov(A3,A7
) Cov(A3,A2) ) Cov(A3,A4) ) Cov(A3,A6) ) Cov(A3,A8) Cov(A3,A9)
Cov(A4,A1 Cov(A4,A3 Cov(A4,A5 Cov(A4,A7
) Cov(A4,A2) ) Cov(A4,A4) ) Cov(A4,A6) ) Cov(A4,A8) Cov(A4,A9)
Cov(A5,A1 Cov(A5,A3 Cov(A5,A5 Cov(A5,A7
) Cov(A5,A2) ) Cov(A5,A4) ) Cov(A5,A6) ) Cov(A5,A8) Cov(A5,A9)
Cov(A6,A1 Cov(A6,A3 Cov(A6,A5 Cov(A6,A7
) Cov(A6,A2) ) Cov(A6,A4) ) Cov(A6,A6) ) Cov(A6,A8) Cov(A6,A9)
Cov(A7,A1 Cov(A7,A3 Cov(A7,A5 Cov(A7,A7
) Cov(A7,A2) ) Cov(A7,A4) ) Cov(A7,A6) ) Cov(A7,A8) Cov(A7,A9)
Cov(A8,A1 Cov(A8,A3 Cov(A8,A5 Cov(A8,A7
) Cov(A8,A2) ) Cov(A8,A4) ) Cov(A8,A6) ) Cov(A8,A8) Cov(A8,A9)
Cov(A9,A1 Cov(A9,A3 Cov(A9,A5 Cov(A9,A7
) Cov(A9,A2) ) Cov(A9,A4) ) Cov(A9,A6) ) Cov(A9,A8) Cov(A9,A9)

Table 4 Covariance Matrix Results

Covariance C1 C2 C3 C4 C5 C6 C7 C8 C9

C1 1 -0.16053 -0.117 -0.06617 -0.03183 0.109248 -0.26982 -0.73521 -0.06013


0.17478
C2 -0.16053 1 0.03519 0.059189 0.137312 0.802889 0.23801 -0.09937
2

PAGE \* MERGEFORMAT 179


C3 -0.117 0.174782 1 -0.0282 0.58612 0.03748 0.151572 0.166233 -0.14694

C4 -0.06617 0.03519 -0.0282 1 0.110819 -0.04246 0.03957 0.069132 0.076691

C5 -0.03183 0.059189 0.58612 0.110819 1 -0.04506 0.043089 0.048115 -0.09342

C6 0.109248 0.137312 0.03748 -0.04246 -0.04506 1 0.13015 -0.0973 -0.28134


0.15157
C7 -0.26982 0.802889 0.03957 0.043089 0.13015 1 0.321778 -0.03871
2
0.16623
C8 -0.73521 0.23801 0.069132 0.048115 -0.0973 0.321778 1 0.000146
3
C9 -0.06013 -0.09937 -0.14694 0.076691 -0.09342 -0.28134 -0.03871 0.000146 1

To see the results of the covariance matrix from the complete electrical energy

consumption dataset, see Appendix 2.

4.2.3. Eigenvalue Decomposition of Covariance Matrix

The covariance matrix formed from the training data has a square matrix

measuring m, i.e., 9 9. Then, the eigenvalue (λ) corresponding to the covariance

matrix is obtained through equation 2.7, so that each scalar (λ1, λ 2……..λm)

satisfies equation 2.8 to calculate attribute weights using eigenvectors. To obtain the

eigenvalues (λ1, λ 2……λm), the lambda values are obtained according to table 7:

Table 5 Eigenvalue

λ Eigenvalue

λ1 2.397030227
λ2 1.586244249
λ3 1.46894975
λ4 1.079500193
λ5 0.95746891
λ6 0.684931487
λ7 0.380870837
λ8 0.258642745
λ9 0.186361601

PAGE \* MERGEFORMAT 179


So first a singular matrix is formed from the covariance matrix of the

training data as follows:

Cov(A1,A1)- λ1 Cov(A1,A2) Cov(A1,A3) Cov(A1,A4) …. Cov(A1,A9)


Cov(A2,A1) Cov(A2,A2)- λ2 Cov(A2,A3) Cov(A2,A4) …. Cov(A2,A9)
Cov(A3,A1) Cov(A3,A2) Cov(A3,A3)- λ3 Cov(A3,A4) …. Cov(A3,A9)
Cov(A4,A1) Cov(A4,A2) Cov(A4,A3) Cov(A4,A4)- λ4 …. Cov(A4,A9)
Cov(A5,A1) Cov(A5,A2) Cov(A5,A3) Cov(A5,A4) …. Cov(A5,A9)
Cov(A6,A1) Cov(A6,A2) Cov(A6,A3) Cov(A6,A4) …. Cov(A6,A9)
Cov(A7,A1) Cov(A7,A2) Cov(A7,A3) Cov(A7,A4) …. Cov(A7,A9)
Cov(A8,A1) Cov(A8,A2) Cov(A8,A3) Cov(A8,A4) …. Cov(A8,A9)
Cov(A9,A1) Cov(A9,A2) Cov(A9,A3) Cov(A9,A4) …. Cov(A9,A9)- λ9

The eigenvalues are formed from the diagonal of the covariance matrix

(Cov(A1,A1), Cov(A2,A2), Cov(A3,A3), Cov(A4,A4),….,Cov(A32,A32)). The

diagonal value is the same as the result of the calculation of the covariance variance

value of each training data attribute as follows:

Covariance variance = Cov(A1,A1)+Cov(A2,A2)+Cov(A3,A3)+….+Cov(A9,A9)

= 1+1+1+1+1+1+1+1+1

=9

By calculating the percentage value of the covariance variance proportion of

9 attributes for each principal component as follows:

Eigenvalue PC 1
1. Value Proportion PC 1 ( % )= x 100 %
Variance Covariance

2.39703022735865
= x 100 %=26.63366919 %
9

Eigenvaluen PC 2
2. Value Proportion PC 2 ( % )= x 100 %
Variance Covariance
PAGE \* MERGEFORMAT 179
1.58624424912761
= x 100 %=17.6249361 %
9

Eigenvalue PC 3
3. Value Proportion PC 3 ( % )= x 100 %
Variance Covariance

1.46894975
= x 100 %=16.32166389 %
9

Eigenvalue PC 4
4. Value Proportion PC 4 ( % )= x 100 %
Variance Covariance

1.079500193
= x 100 %=11.99444659 %
9

Eigenvalue PC 5
5. Value Proportion PC 5 ( % )= x 100 %
Variance Covariance

0.957468910371012
= x 100 %=10.63854345 %
9

Eigenvalue PC 6
6. Value Proportion PC 6 ( % )= x 100 %
Variance Covariance

0.684931486861588
= x 100 %=7.610349854 %
9

Eigenvalue PC 7
7. Value Proportion PC 7 ( % )= x 100 %
Variance Covariance

0.380870837201489
= x 100 %=4.231898191 %
9

Eigenvalue PC 8
8. Value Proportion PC 8 ( % )= x 100 %
Variance Covariance

0.258642745253906
= x 100 %=2.873808281 %
9

Eigenvalue PC 9
9. Value Proportion PC 9 ( % )= x 100 %
Variance Covariance

0.186361601262716
= x 100 %=2.070684458 %
9

PAGE \* MERGEFORMAT 179


4.2.4. Calculating PC

Based on the 2018a metlab calculation, PC1 to PC9 is obtained which can

be seen in table 8

Table 6 PC Data

No. PC1 PC2 .. PC9


-
1 0.918725956 .. -0.31702
1.206156448
-
2 0.907285002 .. -0.31687
1.202639303
-
3 0.850211139 .. -0.30846
1.215012028
-
4 0.8243564 .. -0.31006
1.224005147
-
5 1.313609383 .. -0.33026
1.451273306
-
6 1.289161884 .. -0.32476
1.491026019
-
7 0.750883957 .. -0.3138
1.196913487
-
8 1.310084695 .. -0.33003
1.443649662
-
9 1.108077184 .. -0.29218
1.295219786
-
10 1.111633393 .. -0.29268
1.413583272
.. .. .. .. ..
-
2450 0.600069098 .. 0.162893
1.301710535

4.2.5. Dimensional Reduction

From the results of PC calculations obtained PC1, PC2, PC3 and PC4

which have eigenvalues greater than one.

Table 7 Data Reduction

PAGE \* MERGEFORMAT 179


PC1 (eigen = PC2 (eigen = PC3 (eigen = PC4 (eigen =
No.
2.397) 1.586) 1.468) 1.079)
1 -1.206156448 0.918725956 0.24578191 1.147101731
2 -1.202639303 0.907285002 0.252701965 1.141858956
3 -1.215012028 0.850211139 0.302832342 1.18677732
4 -1.224005147 0.8243564 0.318263702 1.206819643
-
5 -1.451273306 1.313609383 1.080955924
0.239744563
-
6 -1.491026019 1.289161884 1.122461273
0.269041623
7 -1.196913487 0.750883957 0.249049253 1.139528298
-
8 -1.443649662 1.310084695 1.059502197
0.275423334
-
9 -1.295219786 1.108077184 1.10373548
0.039837188
-
10 -1.413583272 1.111633393 1.112555506
0.209027957
.. .. .. .. ..
- -
2450 0.600069098 1.016548816
1.301710535 0.837112294

4.3 BGA Feature Selection

To obtain the covariance of the two attributes [15] . Data are taken on the

basis of input variables from 9 sensors that have been selected using PCA feature

selection. The results of feature selection from PCA are presented in Table 5. In this

study, the dataset was used to find the fitness value according to Table 5 using a

linear equation. Feature selection is bypassed by the steps of BGA feature selection.

Results of processing after PCA using BGA and processing with BGA after

PCA. The initial step in processing using BGA first determines the function of the

training data as much as 2450 Feature selection is done after PCA using BGA

(Binary Genetic Algorithm) using tournament selection. Results of selection using

PCA.

PAGE \* MERGEFORMAT 179


The size or number of digits of the chromosomes depends on the optimal

problem to be solved. The maximum limit of the variable to be searched for is 2450.

Based on the available data, it can be seen that the value of x will be in the range of

0 < x < 2450. Thus, the most optimal data or the best is between 0 and 2450.

Chromosome size 2450 = 100110010010. To determine the size of the population

is done randomly or randomly. To generate random numbers, you can use exell ( (

RANDBETWEEN(0,1)) between 0 and 1

Table 8 Data processed after going through the BGA process

No. PC1 (eigen = PC2 (eigen = PC3 (eigen = PC4 (eigen =


2.397) 1.586) 1.468) 1.079)
1 -1.206156448 0.918725956 0.24578191 1.147101731
2 -1.202639303 0.907285002 0.252701965 1.141858956
3 -1.215012028 0.850211139 0.302832342 1.18677732
4 -1.224005147 0.8243564 0.318263702 1.206819643
5 -1.451273306 1.313609383 -0.239744563 1.080955924
6 -1.491026019 1.289161884 -0.269041623 1.122461273
7 -1.196913487 0.750883957 0.249049253 1.139528298
8 -1.443649662 1.310084695 -0.275423334 1.059502197
9 -1.295219786 1.108077184 -0.039837188 1.10373548
10 -1.413583272 1.111633393 -0.209027957 1.112555506
11 -1.211798447 0.89156391 0.138654915 1.122189258
12 -1.17290858 0.871081124 0.2327218 1.130308197
13 -1.230450823 0.807118855 0.170943721 1.149895983
14 -1.161475831 0.753197371 0.111681226 1.049283446
15 -1.129514733 0.602219352 0.175580782 1.122177138
16 -1.076917727 0.554813572 0.109367372 0.978785827
17 -1.464606009 1.102330014 -0.449657473 1.024605236
18 -1.294695044 0.910627893 -0.090320189 1.124518617
19 -1.334096204 1.026175637 -0.105880765 1.142976588
20 -1.391078612 1.300405338 -0.294091826 1.085146637
… … … … …
2450 0.600069098 -1.301710535 1.016548816 -0.837112294

PAGE \* MERGEFORMAT 179


The initial population was conducted based on data from PCA table 10

processing table with 2450 training data. Then, a gene mutation in the second

iteration on the k3 gene using one point crossover in k3. Max =3 iterations using the

roulette wheel .

Steps to solve iteration 1

4.3.2 Evaluation of the fitness of each chromosome is based on %Pi, the

selected chromosomes are crossover

In the third step of BGA processing, the chromosomes of each

data from PC1, PC2, PC3, and PC4 are converted into binary In the third

step of BGA processing, the chromosomes of each data from PC1, PC2,

PC3, and PC4 are converted into binary.

The next process of BGA is the process of selecting individuals from

the population, where the selection is based on the fitness value. Individuals will be

selected on the basis of the smallest fitness value because in predicting energy use

in the room they get the smallest value, to get an estimate of the minimum value

obtained in achieving energy efficiency in the PVTE laboratory.

In evaluating using the roulette wheel based on the fitness value f(x) %. The

roulette machine is rotated as much as the amount of data that is 2450.

Table 9 Chromosome % Pi

PAGE \* MERGEFORMAT 179


f(x) Fitness
No PC1 PC2 PC3 PC4 Fitness Function f(x) Actual
Function % count
1 -1.20616 0.918726 0.245782 1.147102 192.67 0.02776971 947
2 -1.20264 0.907285 0.252702 1.141859 195.35 0.028155938 963
3 -1.21501 0.850211 0.302832 1.186777 213.35 0.03075103 1,087
4 -1.22401 0.824356 0.318264 1.20682 219.96 0.031703716 1,168
5 -1.45127 1.313609 -0.23974 1.080956 34.89 0.005029182 218
6 -1.49103 1.289162 -0.26904 1.122461 32.40 0.004670535 199
7 -1.19691 0.750884 0.249049 1.139528 215.14 0.031007955 1,110
8 -1.44365 1.310085 -0.27542 1.059502 28.25 0.004072448 187
9 -1.29522 1.108077 -0.03984 1.103735 108.64 0.015658917 410
10 -1.41358 1.111633 -0.20903 1.112556 70.57 0.010170942 277
1035 1.82474 1.326941 -2.35692 -2.95315 (353.54) -0.050957067 1
… … … … … … … …
1520 3.652535 -7.02332 8.614358 0.347582 2,997.92 0.432095878 931
… … … … … … … …
2450 0.600069 -1.30171 1.016549 -0.83711 629.63 0.090750253 1

Total 693,808.15

Average 283.19

MIN (353.54)

MAX 2,997.92

Fitness Function f(x) %


100.00

The selection of individuals from the population was based on fitness. Individuals

with the lowest value were selected because the minimum value was used to predict

the minimum energy consumption in the PVTE laboratory.

Calculate the cumulative probability q i for each chromosome

i
q 1= ∑ P j ......................................................................................... (2)
j=i

PAGE \* MERGEFORMAT 179


Carry out the selection process by generating a random number r (float) in the

range{ 0,1 } if r < qi, then the first chromosome will be selected. q i−1< r <qi the th

chromosome −i detected. From equation 3 it can be seen that the fitness function is

divided by the total number of the fitness functions used.

From equation 9 it is found that the largest % Pi is obtained in percent 2,

because what is sought is minimal, then what is taken is % Pi which has smaller

than percent 2

Table 10 Chromosome PC1, PC2, PC3 and PC4

No. PC1 PC2 PC3 PC4

1 1.001001011
1.0011010011 0.111010110011 0.0011111011
0
2 1.001001000
1.0011001111 0.111010000100 0.0100000010
1
3 1.001011111
1.0011011100 0.000111000110 0.0100110110
1
4 1.001101001
1.0011100101 0.110100110000 0.0101000101
1
5 1.000101001
1.0111001110 1.010100000100 -0.0011110101
0
6 1.000111110
1.0111110110 1.010010100000 -0.0100010011
1
7 1.001000111
1.0011001001 0.110000000011 0.0011111111
0
8 1.000011110
1.0111000110 1.010011110110 -0.0100011010
0
9 1.000101001
1.0100101110 1.000110111010 -0.0000101000
0
10 1.000111001
1.0110100111 1.000111001001 -0.0011010110
1
11 1.000111110
1.0011011000 0.111001000011 0.0010001101
1
12 1.001000010
1.0010110001 0.110111101111 0.0011101110
1
13 1.0011101011 0.110011101001 0.0010101111 1.001001100
PAGE \* MERGEFORMAT 179
1
14 1.000011001
1.0010100101 0.110000001101 0.0011111011
0
15 1.000111110
1.0010000100 0.100110100010 0.0011111011
1
16 0.111110101
1.0001001110 0.100011100000 0.0001101111
0
17 1.000001100
1.0111011011 1.000110100011 -0.0111001100
1
18 1.000111111
1.0100101101 0.111010010001 -0.0001011100
1
19 1.001001001
1.0101010110 0.111010110011 -0.0001101100
0
… … … … …
245
0.162887706 1.162887706 2.162887706 3.162887706
0

From table 10, it can be seen that binary conversion is used for the

next step by performing cross-breeding for the selected percentages.

In this dissertation, because for minimal cases, the lowest

percentage P is taken, 2454 parents are in accordance with table 10

4.3.3. Perform a crossover for the selected parent (crossover)

Crossover is performed on two parent chromosomes to produce

offspring chromosomes. The formation of offspring chromosomes will

inherit some characteristics of their parents. The crossover used in this study

was a one-point crossover. The procedure for choosing which parent will

undergo a crossover process is by determining the crossover probability (Pc)

2. Generate random numbers 0 to 1 as many as i (number of chromosomes

in one population). The next step is to compare the number of random

PAGE \* MERGEFORMAT 179


chromosomes with the crossover probability (Pc). The next step is to select

the parent chromosome if the ith chromosome is less than or equal to the

crossover probability (Pc). The next process occurs when the number of

selected parent chromosomes is only one; this process is repeated until the

number of parents is more than one. In this study, it is necessary to

determine the crossover probability value, which represents the ratio of

offspring produced in the crossover process to population size, so that Pc

population_number will be produced. This study uses a one-point crossover.

1. Determine Pc = 0.5 2. Generate random numbers as large as the total

population. Like table 15.

The point of intersection between Perent 1 and Perent 2 was determined

randomly using Metlab or Excel with RANDBETWEN (1,9). Crossing is performed

using probability (Pc), and no pair may experience a cross. If this occurs, the

offspring will have the same chromosomes as the parent (the chromosomes of the

offspring are the same as those of the parent).

1 2 3 4 .... k-4 k-3 k-2 k-1

k-0

1 0 0 1 ...... 0 0 1 1 0

0 0 1 1 ...... 1 1 1 0 0

PAGE \* MERGEFORMAT 179


1 2 3 4 .... j-3 j-2 j-1 j

The crossing point is determined randomly. You can use a random

number generator, for example, Excel RANDBETWEEN (1,k-1)

Mating Parent 1 with Parent 2 and Parent 2 with Parent 3.

Performing a crossover operator with the following steps: generate a random

number r (float) in the range [0, 1].- if r < pc then the chromosome is

selected for the crossover process- determine crossover point and perform a

crossover.

Crossover between parents in this study is parent 1 with parent 2,

parent 2 with parent 3, and so on until the 2449 data. It can be seen in table

15 that the crossover that then occurs with the child resulting from the

crossover is child 1 to child 2449

Steps of the crossing process in a population resulting from reproduction,

offspring, or children from marriage. Before performing the cross, the probability of

crossover ( Pc ) is determined first. For each parent pair that is formed, we first

determine a random number between 0 and 1. For the crossover process using Excel

RAND(). For each pair, first note the Pc value. If Pc > is greater than a random

number, perform a crossover; if not, do not. The crossover point is determined

randomly, for example, using the help of exel RANBETWEEN ( 1,9 ). Suppose that

the cross is made with Pc = 0.6.


PAGE \* MERGEFORMAT 179
Table 11 Cross-marriage of parents

No. Perent 1 and Perent 2


1 -1.00
0.0 0111
1101 0.000 1110 1.00 0100
1101 1101
0011 1011 0011 1011 0101
0111
0001
2 -1.00
0.0 0100
0001 0.111 0100 1.00 0010
0000 1011
1001 0010 0001 0100 0101
0001
1111
Chilld 1 dan Child 2
1 -1.00
0.0 0111
1101 0.000 1110 1.00 0100
1101 1011
0011 1011 0001 0100 0101
0111
1111
2 -1.00
0.0 0100
0001 0.111 0100 1.00 0010
0000 1101
1001 0010 0011 0000 0101
0001
0011
No. Perent 2 and Perent 3
2 -1.00
0.0 0100
0001 0.111 0100 1.00 0010
0000 1011
1001 0010 0001 0100 0101
0001
1111
3 -1.00
0.0 0010
0110 0.001 1100 1.00 0010
0110 1100
1110 0110 1011 1111 1101
0011
0001
Chilld 2 dan Child 3
2 -1.00
0.0 0100
0001 0.111 0100 1.00 0010
0000 1001
1001 0010 1011 11000101
0001
0001
3 -1.00
0.0 0010
0110 0.001 1100 1.00 0010
0110
1110 0110 0001 11111101
10110011
1111
No. Perent 3 and Perent 4
3 -1.00
0.0 0010
0110 0.001 1100 1.00 0010
0110 1100
1110 0110 1011 1111 1101
0011
0001
4 -1.00
0.0 0101
0111 0.001 1010 1.00 0011
0001 0111
0010 0110 0001 0100 1111
1001
1011
PAGE \* MERGEFORMAT 179
Chilld 3 dan Child 4
3 -1.00
0.0 0010
0110 0.001 1100 1.00 0010
0110 0100
1110 0110 1001 1110 1101
0011
0011
4 -1.00
0.0 0101
0111 0.001 1010 1.00 0011
0001 1111
0010 0110 0011 0101 1111
1001
0001
No. Perent 4 and Perent 5
4 -1.00
0.0 0101
0111 0.001 1010 1.00 0011
0001 0111
0010 0110 0001 0100 1111
1001
1011
5 -1.01
-0.0 0111
1100 1.000 1010 1.00 0010
1010 1011
1110 0000 1001 1001 0111
1111
0001
Chilld 4 dan Child 5
4 -1.00
0.0 0101
0111 0.001 1010 1.00 0011
0001 0111
0010 0110 0001 0100 1111
1001
1011
5 -1.01
-0.0 0111
1100 1.000 1010 1.00 0010
1010 1011
1110 0000 1001 1001 0111
1111
0001
No. Perent 5 and Perent 6
5 -1.01
-0.0 0111
1100 1.000 1010 1.00 0010
1010 1011
1110 0000 1001 1001 0111
1111
0001
6 -1.00
-0.0 0100
0111 1.001 0010 1.00 0011
0100 1101
1101 1000 0001 1110 1011
1111
1011
Chilld 5 dan Child 6
5 -1.01
-0.0 0111
1100 1.000 1010 1.00 0010
1010 1001
1110 0000 0001 1000 0111
1111
0011
6 -1.00
-0.0 0100
0111 1.001 0010 1.00 0011
0100 1111
1101 1000 1001 1111 1011
1111
1001
No. Perent 6 and Perent 7

PAGE \* MERGEFORMAT 179


6 -1.00
-0.0 0100
0111 1.001 0010 1.00 0011
0100 1101
1101 1000 0001 1110 1011
1111
1011
7 -1.00
0.0 0111
0110 0.001 1000 1.0 0100
1111 1000
0100 0000 0111 0111 0111
0011
1101
Chilld 6 dan Child 7
6 -1.00
-0.0 0100
0111 1.001 0010 1.00 0011
0100 1100
1101 1000 0011 1111 1011
1111
1111
7 -1.00
0.0 0111
0110 0.001 1000 1.0 0100
1111 1001
0100 0000 0101 0110 0111
0011
1001
No. Perent 7 and Perent 8
7 -1.00
0.0 0111
0110 0.001 1000 1.0 0100
1111 1000
0100 0000 0111 0111 0111
0011
1101
8 -1.00
0111 1.000 0010 -0.0 010 0011 1.0 0001
0001 0111 1011 0100 0001 1110 0111
1001
Chilld 7 dan Child 8
7 -1.00
0.0 0111
0110 0.001 1000 1.0 0100
1111 1100
0100 0000 1111 0110 0111
0011
1001
8 -1.00
0111 1.000 0010 -0.0 010 0011 1.0 0001
0001 0111 0011 0000 0001 1111 0111
1101
No. Perent 8 and Perent 9
8 -1.00
0111 1.000 0010 -0.0 010 0011 1.0 0001
0100 0111 1011 0100 0001 1110 0111
1001
9 -1.00
-0.00 0010
0100 1.000 0011 1.00 0010
1000 1100
0001 0111 0101 1001 0111
1011
1001
Chilld 8 dan Child 9
8 -1.00 1.000 0010 -0.0 010 0011 1.0 0001
0111 0111 0011 0000 0001 1111 0111
PAGE \* MERGEFORMAT 179
0000
1001
9 -1.00
-0.00 0010
0100 1.000 0011 1.00 0010
1100 1100
0101 0111 1101 1000 0111
1011
1001
No. Perent 9 and Perent 10
9 -1.00
-0.00 0010
0100 1.000 0011 1.00 0010
1000 1100
0001 0111 0101 1001 0111
1011
1001
10 -1.00
-0.0 0110
0011 1.000 0111 1.00 0001
1011 0000
0100 0010 0101 1100 1101
0101
1111
Chilld 9 dan Child 10
9 -1.00
-0.00 0010
0100 1.000 0011 1.00 0010
1001 1100
0001 0110 0101 1000 0111
1011
1001
10 -1.00
-0.0 0110
0011 1.000 0111 1.00 0001
1010 0000
0100 0011 0101 1101 1101
0101
1111
No. Perent 10 and Perent 11
10 -1.00
-0.0 0110
0011 1.000 0111 1.00 0001
1011 0000
0100 0010 0101 1100 1101
0101
1111
11 -1.00
0.0 0100
0110 0.011 1001 1.00 0111
0110 1111
1100 0000 1111 1101 0001
1101
0111
Chilld 10 dan Child 11
10 -1.00
-0.0 0110
0011 1.000 0111 1.00 0001
1111 0000
1100 0000 0101 1101 1101
0101
1111
11 -1.00
0.0 0100
0110 0.011 1001 1.00 0111
0010 1111
0100 0010 1111 1100 0001
1101
0111
No. Perent 11 and Perent 12
11 -1.00
0.0 0100
0110 0.011 1001 1.00 0111
0110 1111
1100 0000 1111 1101 0001
1101
0111
PAGE \* MERGEFORMAT 179
12 -1.00
0.0 0111
0000 0.110 1111 1.0 0100
0111 0010
1011 0111 1111 0010 1011
0111
0001
Chilld 11 dan Child 12
11 -1.00
0.0 0100
0110 0.011 1001 1.00 0111
0111 1111
1101 0001 1111 1100 0001
1101
0111
12 -1.00
0.0 0111
0000 0.110 1111 1.0 0100
0110 0010
1010 0110 1111 0011 1011
0111
0001
No. Perent 12 and Perent 13
12 -1.00
0.0 0111
0000 0.110 1111 1.0 0100
0111 0010
1011 0111 1111 0010 1011
0111
0011
13 -1.00
0.0 0101
1110 0.110 0111 1.00 1001
0111 1000
1011 0100 1111 1001 0111
0101
0001
Chilld 12 dan Child 13
12 -1.00
0.0 0111
0000 0.110 1111 1.0 0100
0111 0000
1011 0101 1111 0000 1011
0111
0001
13 -1.00
0.0 0101
1110 0.110 0111 1.00 1001
0111 1010
1011 0110 1111 1011 0111
0101
0011
No. Perent 13 and Perent 15
12 -1.00
0.0 0111
0000 0.110 1111 1.0 0100
0111 0000
1011 0101 1111 0000 1011
0111
0001
15 -1.00
0.0 0111
1000 0.100 1101 1.00 1001
0111 1000
0100 0001 0101 1011 0111
0101
1001
Chilld 13 dan Child 15
12 -1.00
0.0 0111
0000 0.110 1111 1.0 0100
0111 1000
1011 0001 1111 1000 1011
0111
1001
15 -1.00 0.100 1101 0.0 0111 1.00 1001
1000 0101 0101 0111 0000 0011 0111
PAGE \* MERGEFORMAT 179
0100
0101
0001
No. Perent 15 and Perent 16
15 -1.00
0.0 0111
1000 0.100 1101 1.00 1001
0111 0000
0100 0101 0101 0011 0111
0101
0001
16 -1.00
0.0 0011
0001 0.001 0001 0.00 1111
0111 1111
0011 1100 0001 1010 1001
1111
1011
Chilld 15 dan Child 16
15 -1.00
0.0 0111
1000 0.100 1101 1.00 1001
0111 1000
0100 0100 0101 1011 0111
0101
0011
16 -1.00
0.0 0011
0001 0.001 0001 0.00 1111
0111 0111
0011 1101 0001 0010 1001
1111
1001
No. Perent 16 and Perent 17
16 -1.00
0.0 0011
0001 0.001 0001 0.00 1111
0111 1111
0011 1100 0001 1010 1001
1111
1011
17 -1.00
-0.0 0001
0111 1.000 0001 1.00 0001
1100 1100
0110 1010 0011 1001 0011
0111
1111
Chilld 16 dan Child 17
16 -1.00
0.0 0011
0001 0.001 0001 0.00 1111
0111 1110
0011 1000 0001 1011 1001
1111
1111
17 -1.00
-0.0 0001
0111 1.000 0001 1.00 0001
1100 1101
0110 1110 0011 1000 0011
0111
1011
No. Perent 17 and Perent 18
17 -1.00
-0.0 0001
0111 1.000 0001 1.00 0001
1100 1101
0110 1110 0011 1000 0011
0111
1011
18 -1.00
-0.0 0101
0100 0.111 0100 1.00 0000
1100 0111
1011 1000 1111 1111 1111
1101
0111
PAGE \* MERGEFORMAT 179
Chilld 17 dan Child 18
17 -1.00
-0.0 0001
0111 1.000 0001 1.00 0001
1100 0101
0110 1010 0011 1100 0011
0111
0011
18 -1.00
-0.0 0101
0100 0.111 0100 1.00 0000
1100 1111
1011 1100 1111 1011 1111
1101
1111
No. Perent 18 and Perent 19
18 -1.00
-0.0 0101
0100 0.111 0100 1.00 0000
1100 1111
1011 1100 1111 1011 1111
1101
1111
19 -1.01
-0.0 0001
0101 0.000 1110 1.00 0100
1011 0001
0110 1011 0011 1001 0011
1011
0001
Chilld 18 dan Child 19
18 -1.00
-0.0 0101
0100 0.111 0100 1.00 0000
1100 1101
1011 1000 1111 1001 1111
1101
1101
19 -1.01
-0.0 0001
0101 0.000 1110 1.00 0100
1011 0011
0110 1111 0011 1011 0011
1011
0011
No. Perent 19 and Perent 20
19 -1.01
-0.0 0001
0101 0.000 1110 1.00 0100
1011 0111
0110 1011 0011 0011 0011
1011
0111
20 -1.01
-0.0 0100
1001 1.001 0011 1.00 0101
1011 0000
0000 0111 1001 1111 0011
1001
0011
Chilld 19 dan Child 20
19 -1.01
-0.0 0001
0101 0.000 1110 1.00 0100
1011 0110
0110 0011 0011 0111 0011
1011
0011
20 -1.01
-0.0 0100
1001 1.001 0011 1.00 0101
1011 0001
0000 1111 1001 1011 0011
1001
0111
… … … … …

PAGE \* MERGEFORMAT 179


245 0.162887 1.16288770 3.16288770
2.162887706
0 7 6 6

In iteration 1 we have a new population that is

Table 12 New population

Chilld 1 dan Child 2


1 -1.00
0.000 1110 0.0 0111 1101 1.00 0100
1101 0011
1011 0001 1111 0111 1111 0101
0011
2 -1.00
0.111 0100 0.0 0100 0000 1.00 0010
0001 1001
0010 0011 1001 0001 0000 0101
1101
Chilld 2 dan Child 3
2 -1.00
0.111 0100 0.0 0100 0000 1.00 0010
0001 1001
0010 0011 1001 0001 0101 0101
1101
3 -1.00
0.001 1100 0.0 0010 0110 1.00 0010
0110 1110
0110 1001 1110 0011 1110 1101
0011
Chilld 3 dan Child 4
3 -1.00
0.001 1100 0.0 0010 0110 1.00 0010
0110 1110
0110 1001 0100 0011 1110 1101
0011
4 -1.00
0.001 1010 0.0 0101 0001 1.00 0011
0111 0010
0110 0011 1111 1001 0101 1111
1001
Chilld 4 dan Child 5
4 -1.00
0.001 1010 0.0 0101 0001 1.00 0011
0111 0010
0110 0001 0111 1001 0100 1111
1011
5 -1.01 -0.0 0111
1.000 1010 1.00 0010
1100 1110 1010 1011
0000 1001 1001 0111
0001 1111
Chilld 6 dan Child 7
6 -1.00 -0.0 0100
1.001 0010 1.00 0011
0111 1101 0100 1100
1000 0011 1111 1011
1111 1111
7 -1.00
0.001 1000 0.0 0111 1111 1.0 0100
0110 0100
0000 0101 1001 0011 0110 0111
1001
Chilld 7 dan Child 8
PAGE \* MERGEFORMAT 179
7 -1.00
0.001 1000 0.0 0111 1111 1.0 0100
0110 0100
0000 1111 1100 0011 0110 0111
1001
8 -1.00
1.000 0010 -0.0 010 0011 1.0 0001
0111 0001
0111 0011 0000 0001 1111 0111
1101
Chilld 8 dan Child 9
8 -1.00
1.000 0010 -0.0 010 0011 1.0 0001
0111 0000
0111 0011 0000 0001 1111 0111
1001
9 -1.00 -0.00 0010
1.000 0011 1.00 0010
0100 0101 1100 1100
0111 1101 1000 0111
1001 1011
Chilld 9 dan Child 10
9 -1.00 -0.00 0010
1.000 0011 1.00 0010
0100 0001 1001 1100
0110 0101 1000 0111
1001 1011
10 -1.00 -0.0 0110
1.000 0111 1.00 0001
0011 0100 1010 0000
0011 0101 1101 1101
1111 0101
Chilld 10 dan Child 11
10 -1.00 -0.0 0110
1.000 0111 1.00 0001
0011 1100 1111 0000
0000 0101 1101 1101
1111 0101
11 -1.00
0.011 1001 0.0 0100 0010 1.00 0111
0110 0100
0010 1111 1111 1101 1100 0001
0111
Chilld 11 dan Child 12
11 -1.00
0.011 1001 0.0 0100 0111 1.00 0111
0110 1101
0001 1111 1111 1101 1100 0001
0111
12 -1.00
0.110 1111 0.0 0111 0110 1.0 0100
0000 1010
0110 1111 0010 0111 0011 1011
0001
Chilld 12 dan Child 13
12 -1.00
0.110 1111 0.0 0111 0111 1.0 0100
0000 1011
0101 1111 0000 0111 0000 1011
0001
13 -1.00
0.110 0111 0.0 0101 0111 1.00 1001
1110 1011
0110 1111 1010 0101 1011 0111
0011
Chilld 13 dan Child 15
12 -1.00
0.110 1111 0.0 0111 0111 1.0 0100
0000 1011
0001 1111 1000 0111 1000 1011
1001
15 -1.00 0.100 1101 0.0 0111 0111 1.00 1001
PAGE \* MERGEFORMAT 179
1000 0100
0101 0101 0000 0101 0011 0111
0001
Chilld 15 dan Child 16
15 -1.00
0.100 1101 0.0 0111 0111 1.00 1001
1000 0100
0100 0101 1000 0101 1011 0111
0011
16 -1.00
0.001 0001 0.0 0011 0111 0.00 1111
0001 0011
1101 0001 0111 1111 0010 1001
1001
Chilld 16 dan Child 17
16 -1.00
0.001 0001 0.0 0011 0111 0.00 1111
0001 0011
1000 0001 1110 1111 1011 1001
1111
17 -1.00 -0.0 0001
1.000 0001 1.00 0001
0111 0110 1100 1101
1110 0011 1000 0011
1011 0111
Chilld 17 dan Child 18
17 -1.00 -0.0 0001
1.000 0001 1.00 0001
0111 0110 1100 0101
1010 0011 1100 0011
0011 0111
18 -1.00 -0.0 0101
0.111 0100 1.00 0000
0100 1011 1100 1111
1100 1111 1011 1111
1111 1101
Chilld 18 dan Child 19
18 -1.00 -0.0 0101
0.111 0100 1.00 0000
0100 1011 1100 1101
1000 1111 1001 1111
1101 1101
19 -1.01 -0.0 0001
0.000 1110 1.00 0100
0101 0110 1011 0011
1111 0011 1011 0011
0011 1011
Chilld 19 dan Child 20
19 -1.01 -0.0 0001
0.000 1110 1.00 0100
0101 0110 1011 0110
0011 0011 0111 0011
0011 1011
20 -1.01 -0.0 0100
1.001 0011 1.00 0101
1001 0000 1011 0001
1111 1001 1011 0011
0111 1001
… … … … …
2449 0.162887
1.162887706 2.162887706 3.162887706
7

This iteration process is up to 2450 data points after obtaining a new population,

then continued with iteration 2 until the fitness function f(x) is found according to

PAGE \* MERGEFORMAT 179


the most optimal value according to the steps of the Binery Genetic Algorithm, after

obtaining the minimum value for the next process.

4.3.4. Mutation

The mutation process is a genetic operator that produces random changes in

one chromosome. The easiest way to perform a mutation is to change one or more

parts of the chromosome, which depends on the mutation rate. The mutation rate

determines the probability of the number of bits in a population that are expected to

undergo mutations. If the value of the mutation rate is too small, many useful bits

may not appear in the population; however, if it is too high, the resulting offspring

will lose traits that may be superior to their parents. The mutation operator is used

to change some elements in selected individuals with the mutation probability (Pm)

(mutation rate or mutation probability), leading to genetic diversity to help avoid

the local optimal search process. The mutation that will be performed in this study

is a swap mutation. The number of chromosomes that undergo mutations in a

population is determined by the mutation rate parameter. The mutation process

involves replacing a randomly selected gene with a new value obtained at random.

The process of mutation is as follows: mutation), which leads to genetic diversity

and helps avoid the local optimal search process. The mutation that will be

performed in this study is a swap mutation. The number of mutated chromosomes in

a population is determined by mutation.

Steps for solution iteration 2. The repetition of the process will continue

until the most optimal value is found using Metlab 2018a. The most optimal value

is 1011 so that binary data are processed from the BGA.


PAGE \* MERGEFORMAT 179
Because this is a minimal case, we take the lowest percentage P if we obtain

chromosomes 3 and 4 . Thus, we have two parents, namely.

During the mutation process, a gene in a chromosome is chosen at random

and its number is changed to produce a new chromosome. Mutations were

performed using probability (Pm). Pm is usually small and ranges from 0.001 to

0.1. This is so that mutations are not carried out too often. The position of the

mutation is determined randomly. By using Excel: RANBETWEEN(1,k).

Table 13 First mutation process

Parent 3
0001 .011 001. 0101 000. 0100 01.0 0010
1001 1100 0000 0100 1101 1000 1001 0111
0011 0101 1000 1011 0110 0111 0011
Parent 4
001. 0011 000. 1101 000. 0100 0001 .001
0111 0000 1001 1010 1101 1000 0111 0110
1011 1111 0111 0110 0111 1000 0101

Cross-breed for selected percents between 2 and 3 %

Table 14 Second mutation process

Parent 3 Parent 4
001. 000. 000. 000.
0001 .01 01.0 001. 01.0
0101 0100 1101 0100
1 1001 0010 0011 0010
0000 1101 1001 1101
1100 1001 0111 1001
0100 1000 0 1010 1000
0011 0111 0000 0111
1000 110 1111 0110
0101 0011 1011 0011
1011 0111 0111 0111

Child 3 Child 4
0001 .01 001. 000. 01.0 001. 000. 000. 01.0
PAGE \* MERGEFORMAT 179
0101 0100 1101 0100
1 1001 0000 1101 0010 0011 1001 1101 0010
1100 0100 1000 1001 0111 1010 1000 1001
0011 1000 0110 0111 0000 1111 0110 0111
0101 1011 0111 0011 1011 0111 0111 0011

Do the k2 and k3 gene mutations by replacing 0 to 1 and 1 to 0, so that a

new population of mutations is obtained as follows:

Table 15 The third mutation process

Change bit 1 to 0; bit Change bit 1 to 0; bit 0


0 becomes 1 becomes 1
Child 3 Child 3
000. 000.
0001 .01 01.0 001. 01.0
0100 0100
1 1001 001. 0101 0010 0011 000. 1101 0010
1101 1101
1100 0000 0100 1001 0111 1001 1010 1001
1000 1000
0011 1000 1011 0111 0000 1111 0111 0111
0110 0110
0101 0011 1011 0011
0111 0111

Child Child
M1 M2
111. 111.
0001 .01 01.0 001. 01.0
1011 1011
1 1001 110. 1010 0010 0011 111. 0010 0010
0010 0010
1100 1111 1011 1001 0111 0110 0101 1001
0111 0111
0011 0111 0100 0111 0000 0000 1000 0111
1001 1001
0101 0011 1011 0011
1000 1000

Based on the results of the Child M1 mutation and the results of the Child mutation

of the M2 Child mutation, the new participants had better scores than the previous

participants. In MATLAB 2018a, the training process obtained a value of 1011;

therefore, this data was processed using the ANN and ANFIS models according to

the desired target when calculating energy efficiency in the Electrical Engineering

Vocational Education Laboratory (PVTE) room.

PAGE \* MERGEFORMAT 179


4.4. Comparison 1 of PCA and BGA with the ANN and ANFIS models

Based on the results of the experiments conducted, how do the comparisons when

the PCA and BGA feature selection are implemented? The models being tested are

ANN and ANFIS based on table 20.

Table 16 Comparison of Energy Predictions 1

Correla-
tion coeffi- PCA+AN- PCA+BGA+ PCA+BGA
ANN ANFIS PCA+ANN
cient train- FIS ANN +ANFIS
ing
Time com- 396.34 1397.40
6.621 5.506 4.640 0.403
putation 9 1
Accuration 99% 100% 98% 98% 93% 86%
MSE 817 0 1,572.40 2,090.20 8,451.30 16,084.90
RSME 28.6 0 39.70 45.7 91.9 126.80
MAPE Inf Inf Inf Inf Inf Inf

Correla-
tion coeffi- PCA+AN- PCA+BGA+ PCA+BGA
ANN ANFIS PCA+ANN
cient test- FIS ANN +ANFIS
ing
Time com-
0.047 0.097 0.237 0.003 0.027 0.002
putation
Accuration 99% 100% 95% 93% 94% 93%
3,227.
MSE
70 537.09 39,563.65 42,852.96 21,772.39 33,933.49
RSME 56.81 23.18 198.91 207.01 147.55 184.21
MAPE 7.4 0.39 80.41 79.51 48.46 42.17

The test results in table 20 use 50% for training and 50% for testing use 10% of the

testing data.

PAGE \* MERGEFORMAT 179


Table 17 Comparison of Energy Predictions 2

Correlation
PCA+ PCA+ PCA+BGA PCA+BGA+
coefficient ANN ANFIS
ANN ANFIS +ANN ANFIS
training
Time com- 460.20
69394.828 1308.493 61.377 431.35 8.448
putation 8
Accuration 98% 97% 96% 97% 96% 95%
MSE 0.016 0.03 0.035 0.034 0.039 0.044
RSME 0.128 0.174 0.188 0.183 0.198 0.21
MAPE Inf Inf Inf Inf Inf Inf

Correlation
PCA+ PCA+ PCA+BGA PCA+BGA+
coefficient ANN ANFIS
ANN ANFIS +ANN ANFIS
testing
Time com-
0.455 4.064 0.439 0.599 0.433 0.29
putation
Accuration 98% 98% 96% 97% 96% 95%
MSE 0.033 0.016 0.035 0.034 0.039 0.044
RSME 0.18 0.128 0.19 0.183 0.198 0.21
MAPE Inf Inf Inf Inf Inf Inf

4.4.1 ANN model

In the ANN model, the input parameters are used as the input for all. Table 17

shows the comparison of accuracy and speed of computation time when the data are

trained and tested with input variables Pressure, Temperature (DHT22), Electric

current, Voltage, Power, Luminous intensity, Temperature ( BMP 180), Altitude,

and Humidity when the model is tested with the ANN model. In this experiment,

3332 data points were tested, which were taken every 10 min from Monday. Janu-

ary 6, 2020 to Friday, April 10, 2020. The data are then divided 70% for training,

20% for testing, and 10% for testing.

PAGE \* MERGEFORMAT 179


4.4.1.1 ANN model with training data

Training 70% of the data used is 1 to 2332.

Figure 25 ANN model network during training in predicting the use of

electrical energy in the PVTE room

PAGE \* MERGEFORMAT 179


Gradient = 3.2241, at epoch 1000
5
10

gradient

100

Mu = 1000, at epoch 1000

100
mu

10-5

Validation Checks = 0, at epoch 1000


1
val fail

-1
0 100 200 300 400 500 600 700 800 900 1000
1000 Epochs

Figure 26 ANN training state model in training in predicting electricity usage

PAGE \* MERGEFORMAT 179


Best Training Performance is 817.1559 at epoch 1000
106
Train
Best

105
Mean Squared Error (mse)

104

103

102

101
0 100 200 300 400 500 600 700 800 900 1000
1000 Epochs

Figure 27 The MSE model of ANN in training predicts the use of electrical

energy

Elapsed time is 396.349490 seconds. R value = 0.99342 MSE value = 816.9668

RMSE value = 28.5826 MAPE value = Inf

PAGE \* MERGEFORMAT 179


Pelatihan: R=0.99342

2500 Data
Fit
Y=T

2000
Output ~= 0.99*Target + 3.6

1500

1000

500

0
0 500 1000 1500 2000 2500
Target

Figure 28 The correlation coefficient of the ANN model in training predicts the

use of electrical energy.

PAGE \* MERGEFORMAT 179


Pelatihan (Training)
3000
Target
Output
2500

2000
Output

1500

1000

500

0
0 500 1000 1500 2000 2500
Dataset Index

Figure 29 The correlation coefficient of 2 ANN models in training predicts the

use of electrical energy

4.4.1.2 ANN model with testing data

Testing based on table 16 shows that the data used is 20% of the total data obtained

in the measurements in the PVTE 2333 to 2998 laboratory. The elapsed time was

0.047319 s, R = 0.99072, MSE = 3227.7018, RMSE = 56.8129, and MAPE =

7.3961.

PAGE \* MERGEFORMAT 179


Figure 30 The ANN model network at the time of testing in predicting the use

of electrical energy in the PVTE room

PAGE \* MERGEFORMAT 179


Best Training Performance is 817.1559 at epoch 1000
106
Train
Best

105
Mean Squared Error (mse)

104

103

102

101
0 100 200 300 400 500 600 700 800 900 1000
1000 Epochs

Figure 31 The MSE model of ANN in testing predicts the use of electrical

energy.

PAGE \* MERGEFORMAT 179


Pengujian: R=0.99072
2500 Data
Fit
Y=T

2000
Output ~= 0.99*Target + 5.7

1500

1000

500

500 1000 1500 2000 2500


Target

Figure 32 The correlation coefficient of the ANN model in testing predicts the

use of electrical energy.

PAGE \* MERGEFORMAT 179


Pengujian (Testing)
3000
Target
Output
2500

2000
Output

1500

1000

500

0
0 100 200 300 400 500 600 700
Dataset Index

Figure 33 The correlation coefficient of 2 ANN models in testing predicts the

use of electrical energy.

PAGE \* MERGEFORMAT 179


Gradient = 3.2241, at epoch 1000
5
10

gradient

100

Mu = 1000, at epoch 1000

100
mu

10-5

Validation Checks = 0, at epoch 1000


1
val fail

-1
0 100 200 300 400 500 600 700 800 900 1000
1000 Epochs

Figure 34 Training state model ANN in testing in predicting electricity usage

PAGE \* MERGEFORMAT 179


4.4.1.3 ANN model with prediction data

The prediction data used in data processing is 10% of the data used in experiments

2998 to 3332.

Prediction energiy actual Vs prediction Model ANN


450
400
350
300
250
200
150
100
50
0


10:20

10:30

10:40

10:50

11:00

11:10

11:20

11:30

11:40

11:50

17:00
KAMIS, KAMIS, KAMIS, KAMIS, KAMIS, KAMIS, KAMIS, KAMIS, KAMIS, KAMIS, … JUMAT,
2 APRIL 2 APRIL 2 APRIL 2 APRIL 2 APRIL 2 APRIL 2 APRIL 2 APRIL 2 APRIL 2 APRIL 10
2020 2020 2020 2020 2020 2020 2020 2020 2020 2020 APRIL
2020
1 2 3 4 5 6 7 8 9 10 … 335

Data Actual Data Prediksi ANN

Figure 35 Prediction energiy actual Vs prediction Model ANN

Table 18 ANN Actual Vs Predicted Data

ANN
O'cloc Actua
No Day/date/year Predictio
k l Data
n Data
THURSDAY, 2 400.1
1 10:20 427.7216
APRIL 2020 6
THURSDAY, 2 398.6
2 10:30 424.4568
APRIL 2020 4
THURSDAY, 2 401.2
3 10:40 423.8666
APRIL 2020 8
THURSDAY, 2
4 10:50 179.6 156.1327
APRIL 2020

PAGE \* MERGEFORMAT 179


THURSDAY, 2 184.1
5 11:00 159.3772
APRIL 2020 6
THURSDAY, 2 196.1
6 11:10 158.0072
APRIL 2020 6
THURSDAY, 2 169.4
7 11:20 162.4046
APRIL 2020 4
THURSDAY, 2
8 11:30 184.4 164.1906
APRIL 2020
THURSDAY, 2 169.2
9 11:40 154.2004
APRIL 2020 8
THURSDAY, 2 255.7
10 11:50 290.0342
APRIL 2020 6
… … … … …
33 THURSDAY, 2 315.1
17:00 316.5912
5 APRIL 2020 2

4.5.1 ANFIS model

In the ANFIS model, only 6 input parameters are used: pressure, temperature

(DHT22), electric current, voltage, and power. As with the ANN model, the data are

divided into 70% for training, 20% for testing, and 10% for prediction.

4.5.1.1 ANFIS model training data

In processing the training data used 70%, namely data 1 to 2332 with minimal train-

ing RMSE = 0.0475313 Elapsed time is 1397.401058 seconds, R value = 1, MSE

value = 0.0022592, RMSE value = 0.047531, MAPE value = Inf

PAGE \* MERGEFORMAT 179


Pelatihan: R=1

2500 Data
Fit
Y=T
Output ~= 1*Target + 3.7e-06

2000

1500

1000

500

0
0 500 1000 1500 2000 2500
Target

Figure 36 The correlation coefficient of the ANFIS model in training predicts

the use of electrical energy.

PAGE \* MERGEFORMAT 179


Pelatihan (Training)
3000
Target
Output
2500

2000
Output

1500

1000

500

0
0 500 1000 1500 2000 2500
Dataset Index

Figure 37 The correlation coefficient of the 2 ANFIS models in testing predicts

the use of electrical energy

4.5.1.2 ANFIS model testing data

For the ANFIS model the data used is 20% 2333 to 2998

PAGE \* MERGEFORMAT 179


Pengujian: R=0.99853

Data
2500 Fit
Y=T
Output ~= 0.99*Target + 4

2000

1500

1000

500

500 1000 1500 2000 2500


Target

Figure 38 The correlation coefficient of the ANFIS model in testing predicts

electricity consumption

PAGE \* MERGEFORMAT 179


Pengujian (Testing)
3000
Target
Output
2500

2000
Output

1500

1000

500

0
0 100 200 300 400 500 600 700
Dataset Index

Figure 39 The correlation coefficient of the 2 ANFIS models in testing predicts

the use of electrical energy

4.5.1.2 ANFIS data prediction model

PAGE \* MERGEFORMAT 179


Prediction energiy actual Vs prediction
Model ANFIS
450
400
350
300
250
200
150
100
50
0


10:20

10:30

10:40

10:50

11:00

11:10

11:20

11:30

11:40

11:50

17:00
KAMIS, KAMIS, KAMIS, KAMIS, KAMIS, KAMIS, KAMIS, KAMIS, KAMIS, KAMIS, … JUMAT,
2 APRIL 2 APRIL 2 APRIL 2 APRIL 2 APRIL 2 APRIL 2 APRIL 2 APRIL 2 APRIL 2 APRIL 10
2020 2020 2020 2020 2020 2020 2020 2020 2020 2020 APRIL
2020
1 2 3 4 5 6 7 8 9 10 … 335

Data Actual Data Prediksi ANFIS


Figure 40 Prediction energiy actual Vs prediction Model ANFIS

Table 19 ANFIS Actual Vs Predicted Data

ANFIS
O'cloc Actua
No Day/date/year Predictio
k l Data
n Data
THURSDAY, 2 400.1
1 10:20 399.1512
APRIL 2020 6
THURSDAY, 2 398.6
2 10:30 402.7014
APRIL 2020 4
THURSDAY, 2 401.2
3 10:40 180.1402
APRIL 2020 8
THURSDAY, 2
4 10:50 179.6 184.5721
APRIL 2020
THURSDAY, 2 184.1
5 11:00 196.293
APRIL 2020 6
THURSDAY, 2 196.1
6 11:10 169.9352
APRIL 2020 6
THURSDAY, 2 169.4
7 11:20 184.6343
APRIL 2020 4
THURSDAY, 2
8 11:30 184.4 169.4262
APRIL 2020
THURSDAY, 2 169.2
9 11:40 256.1654
APRIL 2020 8

PAGE \* MERGEFORMAT 179


THURSDAY, 2 255.7
10 11:50 201.2323
APRIL 2020 6
… … … … …
THURSDAY, 2 315.1
335 17:00 316.5912
APRIL 2020 2

4.6.1. PCA dan model ANN

The combination of PCA feature selection results in an elapsed time of 6.620856 s.


R = 0.98733, MSE = 1572.4298, RMSE = 39.6539, MAPE value = Inf.

4.6.1.1 PCA and model ANN training data

It can be seen that the training in Figures 40 and 41 has an R that is almost the
same, close to 1, around 98.7%.

PAGE \* MERGEFORMAT 179


Figure 41 ANN model network with PCA during training in predicting the use

of electrical energy in the PVTE room

PAGE \* MERGEFORMAT 179


Gradient = 48.8164, at epoch 1000
104

gradient

102

Validation Checks = 0, at epoch 1000


1
val fail

-1

Learning Rate = 0.0022747, at epoch 1000


0.06

0.04
lr

0.02

0
0 100 200 300 400 500 600 700 800 900 1000
1000 Epochs

Figure 42 ANN training state model with PCA in training in predicting

electricity consumption

PAGE \* MERGEFORMAT 179


Pelatihan: R=0.98733

2500 Data
Fit
Y=T

2000
Output ~= 0.97*Target + 8.5

1500

1000

500

0
0 500 1000 1500 2000 2500
Target

Figure 43 The correlation coefficient of the PCA model and the ANN model in

testing predicts electricity consumption

PAGE \* MERGEFORMAT 179


Pelatihan (Training)
3000
Target
Output
2500

2000
Output

1500

1000

500

0
0 500 1000 1500 2000 2500
Dataset Index

Figure 44 The correlation coefficient of the PCA model and the ANN model in

testing predicts electricity consumption

4.6.1.2 PCA dan model ANN testing data

When the data are processed by PCA with the model being tested, the ANN model

has an elapsed time of 0.236994 s, R = 0.95205, MSE = 39563.6518, RMSE =

198.9061, and MAPE value = 80.4095

PAGE \* MERGEFORMAT 179


Pengujian: R=0.95205
2500 Data
Fit
Y=T
Output ~= 0.82*Target + 2.1e+02

2000

1500

1000

500

500 1000 1500 2000 2500


Target

Figure 45 The PCA correlation coefficient with the ANN model in testing

predicts the use of electrical energy

PAGE \* MERGEFORMAT 179


Pengujian (Testing)
3000
Target
Output
2500

2000
Output

1500

1000

500

0
0 100 200 300 400 500 600 700
Dataset Index

Figure 46 The correlation coefficient of 2 ANN models in testing predicts the

use of electrical energy

4.6.1.3 PCA and model ANN with prediction data

PAGE \* MERGEFORMAT 179


Prediction energiy actual Vs prediction PCA and Model
ANN
600
500
400
300
200
100
0


10:20

10:30

10:40

10:50

11:00

11:10

11:20

11:30

11:40

11:50

17:00
KAMIS, KAMIS, KAMIS, KAMIS, KAMIS, KAMIS, KAMIS, KAMIS, KAMIS, KAMIS, … JUMAT,
2 APRIL 2 APRIL 2 APRIL 2 APRIL 2 APRIL 2 APRIL 2 APRIL 2 APRIL 2 APRIL 2 APRIL 10
2020 2020 2020 2020 2020 2020 2020 2020 2020 2020 APRIL
2020
1 2 3 4 5 6 7 8 9 10 … 335

Data Actual Data Prediksi ANN

Figure 47 Prediction energiy actual Vs prediction PCA and Model ANN


Table 20 Data Actual Vs Data Prediksi PCA and Model ANN

Data ANN
O'cloc
No Day/date/year Actua Predictio
k
l n Data
THURSDAY, 2 400.1
1 10:20 499.115
APRIL 2020 6
THURSDAY, 2 398.6
2 10:30 495.3655
APRIL 2020 4
THURSDAY, 2 401.2
3 10:40 521.9342
APRIL 2020 8
THURSDAY, 2
4 10:50 179.6 428.5002
APRIL 2020
THURSDAY, 2 184.1
5 11:00 405.5754
APRIL 2020 6
THURSDAY, 2 196.1
6 11:10 369.7468
APRIL 2020 6
THURSDAY, 2 169.4
7 11:20 324.2928
APRIL 2020 4
THURSDAY, 2
8 11:30 184.4 381.2793
APRIL 2020
THURSDAY, 2 169.2
9 11:40 402.9195
APRIL 2020 8

PAGE \* MERGEFORMAT 179


THURSDAY, 2 255.7
10 11:50 392.3155
APRIL 2020 6
… … … … …
33 THURSDAY, 2 315.1
17:00 473.3933
5 APRIL 2020 2

4.7.1. PCA and ANFIS models

4.7.1.1 PCA with ANFIS training data model

Based on the results of the comparison of feature selection, the computational speed

of elapsed time is 2.840663 s, R = 0.98309, MSE value = 2090.1693, RMSE value

= 45.7184, MAPE value = Inf

Pelatihan: R=0.98309

2500 Data
Fit
Y=T

2000
Output ~= 0.97*Target + 9.4

1500

1000

500

0
0 500 1000 1500 2000 2500
Target

PAGE \* MERGEFORMAT 179


Figure 48 The correlation coefficient of the ANFIS model in training predicts

the use of electrical energy

Pelatihan (Training)
3000
Target
Output
2500

2000
Output

1500

1000

500

0
0 500 1000 1500 2000 2500
Dataset Index

Figure 49 The correlation coefficient of the 2 ANFIS models in testing predicts

the use of electrical energy

4.7.1.2 PCA with ANFIS testing data model

The application of PCA feature selection contributes to reducing input variables so

that data computation speeds are faster. Elapsed time: 0.002887 s, R = 0.92482,

PAGE \* MERGEFORMAT 179


MSE = 42852.9601, RMSE = 207.0096, MAPE = 79.5074

Pengujian: R=0.92482
2500 Data
Fit
Y=T
Output ~= 0.72*Target + 2.1e+02

2000

1500

1000

500

500 1000 1500 2000 2500


Target

Figure 50 The PCA correlation coefficient with the ANFIS model in testing

predicts electricity consumption

PAGE \* MERGEFORMAT 179


Pengujian (Testing)
3000
Target
Output
2500

2000
Output

1500

1000

500

0
0 100 200 300 400 500 600 700
Dataset Index

Figure 51 The PCA correlation coefficient with the ANFIS model in testing

predicts electricity consumption.

4.7.1.3 PCA with ANFIS data prediction model

Figure 52 shows the difference between the prediction of the actual data by apply-

ing the PCA feature selection with the ANFIS model. You can see that the signifi -

cant difference between the actual data and the prediction data using PCA with the

ANFIS model is the average difference of 175.28, and the data closest to the actual

data is on Thursday at 10.40 with a difference that is not too far from the actual data

401.28, while the predicted data using PCA feature selection with the ANFIS model

392.071271 differs only 9.208729.

PAGE \* MERGEFORMAT 179


Prediction energiy actual Vs prediction PCA and Model
ANFIS
700
600
500
400
300
200
100
0


10:20

10:30

10:40

10:50

11:00

11:10

11:20

11:30

11:40

11:50

16:50
KAMIS, KAMIS, KAMIS, KAMIS, KAMIS, KAMIS, KAMIS, KAMIS, KAMIS, KAMIS, … JUMAT,
2 APRIL 2 APRIL 2 APRIL 2 APRIL 2 APRIL 2 APRIL 2 APRIL 2 APRIL 2 APRIL 2 APRIL 10
2020 2020 2020 2020 2020 2020 2020 2020 2020 2020 APRIL
2020
1 2 3 4 5 6 7 8 9 10 … 334

Data Actual Data Prediksi PCA+ANFIS

Figure 52 Prediction energiy actual Vs prediction PCA and Model ANFIS

Table 21 Actual Data vs Predicted Data PCA and ANFIS Models

PCA+ANFI
Day/date/ O'cloc Data S
No
year k Actual Prediction
Data
THURSDAY, 549.166462
1 10:20 400.16
2 APRIL 2020 3
THURSDAY, 612.242093
2 10:30 398.64
2 APRIL 2020 3
THURSDAY,
3 10:40 401.28 392.071271
2 APRIL 2020
KAMIS, 2 382.095496
4 10:50 179.6
APRIL 2020 2
THURSDAY, 393.145309
5 11:00 184.16
2 APRIL 2020 2
THURSDAY,
6 11:10 196.16 376.041908
2 APRIL 2020
THURSDAY, 410.265384
7 11:20 169.44
2 APRIL 2020 3

PAGE \* MERGEFORMAT 179


THURSDAY, 394.480481
8 11:30 184.4
2 APRIL 2020 3
THURSDAY, 412.141212
9 11:40 169.28
2 APRIL 2020 2
THURSDAY, 370.013468
10 11:50 255.76
2 APRIL 2020 4
… … … … …
THURSDAY, 445.458557
334 16:50 328.32
2 APRIL 2020 9

4.8.1 PCA and BGA with ANN models

The selection of PCA and BGA features with parameter 9 yields good results,

where it is possible that the results are no better than those of the ANN model

without added feature selection.

4.8.1.1 PCA and BGA model ANN training data

Combining PCA and BGA feature selection, the computational speed of the elapsed

time is 4.640324 s, R = 0.92987, MSE = 8451.3352, RMSE value = 91.9311,

MAPE value = Inf.

PAGE \* MERGEFORMAT 179


Pelatihan: R=0.92987

2500 Data
Fit
Y=T

2000
Output ~= 0.85*Target + 39

1500

1000

500

0
0 500 1000 1500 2000 2500
Target

Figure 53 Correlation coefficient of PCA and BGA model of ANN training

data predicts electricity consumption

PAGE \* MERGEFORMAT 179


Pelatihan (Training)
3000
Target
Output
2500

2000
Output

1500

1000

500

0
0 500 1000 1500 2000 2500
Dataset Index

Figure 54 Correlation coefficient of PCA and BGA model of ANN training

data predicts electricity consumption

4.8.1.2 PCA and BGA model ANN testing data

Elapsed time is 0.027179 seconds, R value = 0.93683, MSE value = 21772.386,

RMSE value = 147.5547, MAPE value = 48.4587

PAGE \* MERGEFORMAT 179


Pengujian: R=0.93683
2500 Data
Fit
Y=T

2000
Output ~= 0.84*Target + 50

1500

1000

500

500 1000 1500 2000 2500


Target

Figure 55 The correlation coefficient of PCA and BGA models of ANN testing

data predicts electricity consumption

PAGE \* MERGEFORMAT 179


Pengujian (Testing)
3000
Target
Output
2500

2000
Output

1500

1000

500

0
0 100 200 300 400 500 600 700
Dataset Index

Figure 56 The correlation coefficient of PCA and BGA models of ANN testing

data predicts electricity consumption

4.8.1.3 PCA and BGA ANN models with prediction data

Prediction energiy actual Vs prediction PCA and BGA


with Model ANN
400

300

200

100

0

10:20

10:30

10:40

10:50

11:00

11:10

11:20

11:30

11:40

11:50

16:50

KAMIS, KAMIS, KAMIS, KAMIS, KAMIS, KAMIS, KAMIS, KAMIS, KAMIS, KAMIS, … JUMAT,
2 APRIL 2 APRIL 2 APRIL 2 APRIL 2 APRIL 2 APRIL 2 APRIL 2 APRIL 2 APRIL 2 APRIL 10
2020 2020 2020 2020 2020 2020 2020 2020 2020 2020 APRIL
2020
1 2 3 4 5 6 7 8 9 10 … 334

Data Actual Data Prediksi PCA+BGA+ANN

PAGE \* MERGEFORMAT 179


Table 22 Data Actual Vs prediction PCA and BGA with Model ANN

PCA+BGA+A
Actu
N Day/date/ O'clo NN
al
o year ck Prediction
Data
Data
THURSDAY, 2 400.1
1 10:20 240.4725
APRIL 2020 6
THURSDAY, 2 398.6
2 10:30 269.1279
APRIL 2020 4
THURSDAY, 2 401.2
3 10:40 269.3941
APRIL 2020 8
THURSDAY, 2
4 10:50 179.6 269.4417
APRIL 2020
THURSDAY, 2 184.1
5 11:00 251.2617
APRIL 2020 6
THURSDAY, 2 196.1
6 11:10 226.6766
APRIL 2020 6
THURSDAY, 2 169.4
7 11:20 165.5239
APRIL 2020 4
THURSDAY, 2
8 11:30 184.4 149.6222
APRIL 2020
THURSDAY, 2 169.2
9 11:40 159.3401
APRIL 2020 8
THURSDAY, 2 255.7
10 11:50 163.8037
APRIL 2020 6
… … … … …
33 THURSDAY, 2 328.3
16:50 389.8919
4 APRIL 2020 2

PAGE \* MERGEFORMAT 179


4.9.1. PCA and BGA with ANFIS models

4.9.1.1 PCA and BGA with ANFIS training data models

Selection of PCA and BGA features results in calculating the elapsed time is

0.402992 seconds, R value = 0.86134, MSE value = 16084.854, RMSE value =

126.8261, MAPE value = Inf

Pelatihan: R=0.86134

2500 Data
Fit
Y=T

2000
Output ~= 0.74*Target + 67

1500

1000

500

0
0 500 1000 1500 2000 2500
Target

Figure 57 The correlation coefficient of combining PCA and BGA feature

selection in the ANFIS model in training predicts electricity consumption

PAGE \* MERGEFORMAT 179


Pelatihan (Training)
3000
Target
Output
2500

2000
Output

1500

1000

500

0
0 500 1000 1500 2000 2500
Dataset Index

Figure 58 The correlation coefficient of 2 PCA and BGA ANFIS models in

training predicts the use of electrical energy

4.9.1.2 PCA and BGA with ANFIS testing data models

Merging PCA and BGA feature selection. Elapsed time is 0.002002 s. R = 0.92923,

MSE = 33933.4865, RMSE = 184.2104, MAPE value = 42.1658.

PAGE \* MERGEFORMAT 179


Pengujian: R=0.92923
2500 Data
Fit
Y=T
Output ~= 0.64*Target + 1.2e+02

2000

1500

1000

500

500 1000 1500 2000 2500


Target

Figure 59 Correlation coefficient combining PCA and BGA feature selection of

the ANFIS model in testing predicts electricity consumption

PAGE \* MERGEFORMAT 179


Pengujian (Testing)
3000
Target
Output
2500

2000
Output

1500

1000

500

0
0 100 200 300 400 500 600 700
Dataset Index

Figure 60 The correlation coefficient of 2 PCA and BGA models of ANFIS in

testing predicts electricity consumption

4.8.1.2 PCA and BGA models of ANFIS prediction data

PAGE \* MERGEFORMAT 179


Prediction energiy actual Vs prediction PCA and BGA
with Model ANFIS
450

400

350

300

250

200

150

100

50


10:20

10:30

10:40

10:50

11:00

11:10

11:20

11:30

11:40

11:50

16:50
KAMIS, KAMIS, KAMIS, KAMIS, KAMIS, KAMIS, KAMIS, KAMIS, KAMIS, KAMIS, … JUMAT,
2 APRIL 2 APRIL 2 APRIL 2 APRIL 2 APRIL 2 APRIL 2 APRIL 2 APRIL 2 APRIL 2 APRIL 10
2020 2020 2020 2020 2020 2020 2020 2020 2020 2020 APRIL
2020
1 2 3 4 5 6 7 8 9 10 … 334

Data Actual Data Prediksi PCA+BGA+ANFIS

Figure 61 Prediction energiy actual Vs prediction PCA and BGA with Model

ANFIS

Table 23 Data Actual Vs prediction PCA and BGA with Model ANFIS

PCA+BGA+A
Data
Day/date/ O'clo NFIS
No Actu
year ck Prediction
al
Data

THURSDA
400.1
1 Y, 2 APRIL 10:20 240.4725
6
2020
THURSDA
398.6
2 Y, 2 APRIL 10:30 269.1279
4
2020

PAGE \* MERGEFORMAT 179


THURSDA
401.2
3 Y, 2 APRIL 10:40 269.3941
8
2020
THURSDA
4 Y, 2 APRIL 10:50 179.6 269.4417
2020
THURSDA
184.1
5 Y, 2 APRIL 11:00 251.2617
6
2020
THURSDA
196.1
6 Y, 2 APRIL 11:10 226.6766
6
2020
THURSDA
169.4
7 Y, 2 APRIL 11:20 165.5239
4
2020
THURSDA
8 Y, 2 APRIL 11:30 184.4 149.6222
2020
THURSDA
169.2
9 Y, 2 APRIL 11:40 159.3401
8
2020
THURSDA
255.7
10 Y, 2 APRIL 11:50 163.8037
6
2020
… … … … …
THURSDA
328.3
334 Y, 2 APRIL 16:50 389.8919
2
2020

4.7. Comparison 3 of PCA and BGA with ANN and ANFIS models

To show the reliability of this feature selection model, the proposed feature selec-

tion is implemented on how to manage energy in a housing according to a previous

study [5][6]. Where in this test consisted of 42 houses analyzed, but in this study it

was only tested using 1 house with 462240, with the input parameters being AC in

Drawing Room (AC_DR_kW), UPS (UPS_kW), Lounge Room (LR_kW), Kitchen

(Kitchen_kW), AC_Dr_kW (AC in Dining Room), AC in Bed Room

(AC_BR_kW), while the total energy consumption used is in one minute

Usage_kW. Data collection starts on 7/15/2018 0:00 until 5/31/2019 23:59 collec-

PAGE \* MERGEFORMAT 179


tion for 8 months. In this study, the data is divided 70% for training, 20% for test-

ing, and 10% for prediction. Therefore, the results are presented in table 28.

Table 24 Comparison of Energy Predictions

Correla-
tion coeffi- PCA PCA+ PCA+BGA PCA+BGA
ANN ANFIS
cient train- +ANN ANFIS +ANN +ANFIS
ing
Time com- 548.30
14461.238 1135.118 44.067 523.485 7.242
putation 8
Accuration 98% 97% 96% 97% 96% 95%
MSE 0.016 0.03 0.035 0.034 0.043 0.046
RSME 0.128 0.174 0.188 0.183 0.208 0.214
MAPE Inf Inf Inf Inf Inf Inf

Correla-
tion coeffi- PCA PCA+ PCA+BGA PCA+BGA
ANN ANFIS
cient test- +ANN ANFIS +ANN +ANFIS
ing
Time com-
0.455 4.064 0.439 0.599 0.433 0.29
putation
Accuration 90% 89% 86% 86% 86% 85%
MSE 0.023 0.027 0.034 0.033 0.037 0.039
RSME 0.15 0.164 0.185 0.182 0.192 0.197
MAPE Inf Inf Inf Inf Inf Inf

In table 24, it can be seen that the test was carried out by testing data with a total of

70% of the total data usage so that testing data, testing data, and predictions were

obtained 323568, as shown in Table 28. When the ANN model was tested, it

obtained an accuracy value equal to testing and training, which is 0.983 but has not

too much difference between MSE and RSME in training and testing. However, it

still has the most campuses in the ANN and ANFIS models.

Table 25 Comparison of Energy Predictions 3

Correla-
tion coeffi- PCA+ PCA+ PCA+BGA+ PCA+BGA+
ANN ANFIS
cient train- ANN ANFIS ANN ANFIS
ing
PAGE \* MERGEFORMAT 179
Time com- 13479. 1049.7 486.78
37.998 460.268 6.986
putation 543 84 5
Accuration 98% 97% 96% 97% 96% 95%
MSE 0.016 0.03 0.035 0.034 0.043 0.214
RSME 0.128 0.174 0.188 0.183 0.208 0.046
MAPE Inf Inf Inf Inf Inf Inf

Correla-
tion coeffi- PCA+ PCA+A PCA+BGA+ PCA+BGA+
ANN ANFIS
cient test- ANN NFIS ANN ANFIS
ing
Time com-
1.503 1.872 0.501 0.728 0.454 0.296
putation
Accuration 0.983 0.968 0.963 0.965 0.955 0.952
MSE 0.016 0.03 0.035 0.034 0.043 0.046
RSME 0.128 0.174 0.188 0.183 0.208 0.215
MAPE Inf Inf Inf Inf Inf Inf

The graph of training time and testing computational time can be seen in Figure 55,

where the largest computation time is found in the ANN model when presenting

presented. The computation time is quite long, i.e., 13479,543 s for training and

1,503 s for testing. When the ANN model is added and the PCA feature is selected,

the speed of data computation becomes shorter, i.e., 486.785 s for training and test-

ing becomes 0.501 s. When ANN is added, select PCA and BGA features so that

the data speed is faster by utilizing the looping process from BGA to increase the

data speed for training 460.268 seconds for the testing process 0.454 seconds.

PAGE \* MERGEFORMAT 179


Computasi time 3 Training Vs Testing
15000

13000

11000

9000

7000

5000

3000

1000
ANN ANFIS PCA+ANN PCA+ANFIS PCA+BGA+A PCA+BGA+A
NN NFIS
Training Time ( NaN NaN 13479.5427 1049.78350 486.784799 37.998437 460.267827 6.986356
seconds) 8 3
Testing Time NaN NaN 1.503112 1.872113 0.501056 0.728083 0.454095 0.295787
( seconds)

Training Time ( seconds) Testing Time ( seconds)

Figure 62 Computasi time 3 Training Vs Testing

The computation speed in table 29 can be seen from the graph in Figure 55, which

shows that the largest computation occurs in the ANN. It can be seen how the sig-

nificant difference from the ANN model then ANFIS shows that the acceleration for

quite a long time is found in the ANN model and while the fastest speed is found in

the model that has added PCA and BGA feature selection with the model being

tested is ANFIS.

PAGE \* MERGEFORMAT 179


Comparison of Energy Predictions 4 Vs
Usage_KW
2.5

1.5

0.5

0
1 2 3 4 5 6 7 8 9 10 11

Usage_kW ANN ANFIS


PCA+ANN PCA+ANFIS PCA+BGA+ANN
PCA+BGA+ANFIS

Figure 63 Comparison of Energy Predictions 4 Vs Usage_KW

As seen in Figure 62, a comparison of the differences between feature selection and

the model being tested shows that the prediction of energy use in house 1 is feature

selection with the addition of PCA and BGA, with the models tested being ANN

and ANFIS.

PAGE \* MERGEFORMAT 179


Table 26 Prediction Data Actual and prediction of housing energy

consumption (House 1)

PCA+ PCA+ PCA+BGA PCA+BGA


sage_kW ANN ANFIS
ANN ANFIS +ANN +ANFIS
1.88411 1.850
1.731 3 4 1.831 1.807 1.753058 1.766117
1.932 1.871
1.7415 1.97178 6 1.893 4 1.770998 1.828634
1.96536 1.926 1.867
1.7391 1 9 1.89 2 1.77015 1.824078
1.97565 1.932 1.871
1.7371 1 2 1.892 2 1.771098 1.827619
1.95157 1.914 1.857
1.7268 3 8 1.878 8 1.76715 1.814591
1.94790
1.7246 1 1.916 1.878 1.858 1.766769 1.816363
1.94564 1.856
1.7215 5 1.914 1.875 1 1.765985 1.814842
2.04649 2.020 1.944
1.755 9 2 1.962 3 1.796393 1.903371
2.06304 2.043 1.964
1.7706 7 7 1.98 4 1.804305 1.923246
.. .. .. .. .. .. ..
2.003 1.929
1.757 2.03486 3 1.945 8 1.790107 1.888261

As seen in table 26, it can be seen how close the actual data is to the prediction data

when added to the selection of PCA and BGA features with the models tested,

namely ANN and ANFIS. It can be seen from the first data that the target energy

use in house 1 is 1,731 kW before adding the ANN feature selection to 1.884113

kW. Meanwhile, when the PCA feature selection is added, the data becomes 1,807

kW, and when the PCA feature selection is combined with BGA, the data is closest

to the actual data, namely 1.753058 kW. Likewise, the model tested with the

ANFIS model is shown in Table 26. It can be seen that the ANFIS model when

added with the selection of PCA and BGA data is closer to the actual data, where
PAGE \* MERGEFORMAT 179
the actual data has a goal of 1,731 kW, whereas when the selection of PCA and

BGA features becomes 1.766117 kW.

4.5. Comparison 1 of PCA and BGA with ANN and ANFIS models

On the basis of the comparison of the time when added to the feature selection by

not using the feature selection in accordance with Figure 85 and Figure 24, the

efficiency of the given time is shown. In Figure 25, it can be seen that the accuracy

results from table 16.

Perbandingan korekasi R1 dan R2


1.05 1.000
0.993
0.991 0.999 0.987 0.983
1 0.952
0.925 0.939
0.930 0.929
0.95
0.9 0.861
0.85
0.8
0.75
N S N S N
AN FI AN FI AN FI
S
AN A+ AN A+ AN
PC CA+ BG A+
P
A+ BG
PC
CA+
P

Prediction Energy Correlation coefficient R1 Training


Correlation coefficient R2 Testing

Figure 64 Comparison correlation coefficient R1 and R2

ANN,ANFIS,PCA+ANN,PCA+ANFIS

From Figure 85, the comparison correlation coefficient R1 and R2 ANN,

ANFIS, PCA + ANN, and PCA + ANFIS training shows that the difference is not

too far from the linearity of the comparison of each model tested, as shown in table

16. The contribution of this research is how data speed when using PCA and BGA

PAGE \* MERGEFORMAT 179


feature selection significantly contributes to data speed when conducting training

and testing processes.

Perbandingan korekasi R1 dan R2 dengan waktu


komputasi training dan testing
1397.401058
1300
900
396.34949
500
0.04731 6.620856 0.002887
0.097438 0.236994 5.505738 0.027179
4.640324 0.002002
0.402992
100
R1 Train- ANN ANFIS 1 PCA+ANN PCA+ANFI PCA+BGA PCA+BGA
ing R2 0.993 0.999 0.987 S 0.983 +ANN +ANFIS
Testing 0.991 0.952 0.925 0.93 0.939 0.861
0.929
Train- NaN NaN 396.34949 1397.4010 6.620856 5.505738 4.640324 0.402992
ing 58
Time
( sec-
onds)
Test- NaN NaN 0.04731 0.097438 0.236994 0.002887 0.027179 0.002002
ing
Time
(sec-
onds)

Figure 65 Comparison of corrections R1 and R2 with computational training

and testing time

In Figure 64, a comparison of the correlation with training and testing computation

time shows that the longest computation time is found in the ANFIS model with

1397.401058 s for a training time of 0.097438 s and a computation time of 0.04731

s. Computational training and testing time will be faster when PCA feature selection

is added when the model is tested by ANFIS, with a computation time of 5.505738

s for training and 0.002887 s for testing. When the selection of BGA and PCA

features is added to the ANFIS model, the computation time becomes faster at

0.402992 s for training and 0.002002 for testingLikewise, if the ANN model is

tested without using feature selection, the computation time is relatively longer

PAGE \* MERGEFORMAT 179


compared to the model added with feature selection, either added with PCA or

BGA.

4.6. Comparison 2 of PCA and BGA with ANN and ANFIS models

Table 27 shows a comparison of the six parameters used as input for the ANN.AN-

FIS, PCA+ANN.PCA+BGA+ANFIS, and PCA+BGA+ANN models. The best ac-

curacy value was found in the ANFIS model with a correlation R = 1 training, and

the same test had the best accuracy value. Training for the ANFIS model with MSE

0.004 and RSME 0.061, but the computational speed is still relatively long, namely

7.447 s, while for testing, the computation time is 0.009 s. As for the testing compu-

tational speed, MSE was 1.217, RSME was 1.103, and MAPE was 0.038.

Table 27 Comparison of Energy Predictions 2

Correlation
PCA+ PCA+ PCA+BGA+ PCA+BGA
coefficient ANN ANFIS
ANN ANFIS ANN +ANFIS
training
Time com- 149.059
putation 586 7.447 5.387 0.51 4.841 0.258
Accuration 98% 100% 99% 99% 97% 95%
1,233.0 1,474.0
MSE
11.256 0.004 0 0 3,202.10 5,972.90
RSME 3.355 0.061 35.115 38.393 56.587 77.284
MAPE 25.829 Inf Inf Inf Inf Inf

Correlation
PCA+ PCA+ PCA+BGA+ PCA+BGA
coefficient ANN ANFIS
ANN ANFIS ANN +ANFIS
testing
Time com- 149.059
7.447 5.387 0.51 4.841 0.258
putation 586
Accuration 84% 100% 97% 96% 96% 95%
19187. 22071.
MSE 0.058 1.217 29350.787 57349.905
743 79
148.56
RSME 0.241 1.103 138.52 171.321 239.478
6
MAPE 35.244 0.038 55.894 50.265 75.175 100.431
PAGE \* MERGEFORMAT 179
The ANN model has a correlation value of R1 testing 0.980 with a testing value of

R2 0.836 but has a poor MSE of 11.256, RSME 3.355, and MAPE 25.829.

Comparison of Energy Predictions 2 Vs data


actual
Data Aktual Prediksi Model ANFIS Prediksi PCA+ANFIS
PCA+BGA+ANFIS Prediction model ANN Prediksi PCA+ANN
PCA+BGA+ANN
700
600
500
400
300
200
100
0
1 2 3 4 5 6 7 8 9 10 11

Figure 66 Comparison of Energy Predictions 2 Vs data actual

Figure 65 shows a comparison between the actual data and the tested model. It can

be seen that the models closest to the actual data are the ANN and AN-FIS models

but have the longest data rates from the comparison of the application of the PCA

and BGA selection features based on table 26. Comparison of the energy predic-

tions 2

Predicted data can be seen from table 27. There is a huge difference between the

actual and predicted data using PCA and BGA feature selection with the ANN and

ANFIS models.

PAGE \* MERGEFORMAT 179


Figure 66 shows a comparison of the computational time in training and testing

where the longest computation time is in the ANN and ANFIS models.

Figure 67 Computasi time 3 Training Vs Testing

The computational time for the ANN model is 149.059586 s for training and 0.403 s

for testing. A comparison of the predicted and actual data is shown in Table 28.

Table 28 Prediction Data Actual and prediction model ANN

Prediksi Prediksi Prediction Prediksi


Data PCA+BG PCA+BGA
Model PCA+ANF model PCA+AN
Aktual A+ANFIS +ANN
ANFIS IS ANN N
405.92 402.217 435.969 552.9151 0.9222831 659.7036 380.3045
400.16 405.933 453.5528 629.0109 1.006332 543.2813 381.59681
398.64 400.164 436.6114 391.0779 1.0633949 612.4135 380.35673
401.28 398.651 434.7164 377.9204 1.0427047 559.7441 380.36728
179.6 401.289 325.3847 387.6834 1.0027885 559.6891 379.56789
184.16 179.557 320.0591 372.3232 1.1854849 401.6422 378.31417
196.16 184.124 362.1311 418.0254 1.0510269 374.6153 380.13855
169.44 196.087 300.7418 399.0994 1.054565 453.0023 379.34326
184.4 169.426 331.9784 416.2774 1.0285209 358.6416 380.23803
.. .. .. .. .. .. ..
368.32 328.335 387.0654 602.7284 0.4436742 458.8344 380.82032

PAGE \* MERGEFORMAT 179


4.8. Comparison 4 of PCA and BGA with ANN and ANFIS models

Comparison of the implementation of PCA and BGA feature selection with the

models tested by ANN and ANFIS. The input parameters used were six parameters

with training, testing, and prediction data equal to 70% of the data used in predict-

ing energy use in the electrical engineering vocational education laboratory

(PVTE). Table 29 shows a comparison of the energy use with feature selection. It

can be seen that the ANFIS model shows an accuracy value of R=1 with an MSE

value of 0.00, while the RSME is 3.355. However, the computation time is rela-

tively long, training is 7.925, and testing is 0.015. Selection of PCA and BGA fea -

tures contributes to computation time.

Table 29 Comparison of Energy Predictions 3

Correlation
PCA+ PCA+ PCA+BGA PCA+BGA
coefficient ANN ANFIS
ANN ANFIS +ANN +ANFIS
training
Time com-
151.187 7.925 5.746 2.408 4.694 0.601
putation
Accuration 98% 100% 99% 99% 97% 95%
1,233.0 1,473.9
MSE 11.26 0 3,202.08 5,972.86
4 8
RSME 0.128 3.355 35.115 38.393 56.587 77.284
MAPE Inf Inf Inf Inf Inf Inf

Correlation
PCA+ PCA+ PCA+BGA PCA+BGA
coefficient ANN ANFIS
ANN ANFIS +ANN +ANFIS
testing
Time com-
0.082 0.015 0.011 0.007 0.01 0.004
putation
Accuration 98% 100% 99% 99% 97% 95%
1,233.0 1,473.9
MSE 11.26 0 3,202.08 5,972.96
4 8
RSME 0.128 0.061 35.115 38.393 56.587 77.285
MAPE Inf Inf Inf Inf Inf Inf
PAGE \* MERGEFORMAT 179
Comparison of the prediction results based on table 29 of the actual data with the

predicted results of the ANN model. The best accuracy value can be seen in the

ANFIS model with an accuracy value of R = 1 or equal to 100% with less computa-

tion time than the ANN model, which is 7,925 s for training, while the computing

time for testing is smaller, namely 0,015 s. However, if the ANFIS model is added

by selecting the PCA feature, the accuracy value becomes R = 0.988 with a smaller

computation time than the model that is not added by selecting the training feature

(2.408 s) and testing 0.007 seconds. If the data are added to the selection of PCA

and BGA features, the data speed when performing computations becomes faster

compared with models that are not added with feature selection.

PAGE \* MERGEFORMAT 179


Comparison of Energy Predicti ons 4 Vs
data actual
Data actual Linear (Data actual) ANN
ANFIS PCA+ANFIS PCA+ANN
PCA+BGA+ANN PCA+BGA+ANFIS

225

175

125

75

25

1 2 3 4 5 6 7 8 9
Dat 191.6 196.88 207.6 219.04 62.64 55.68 219.92 0 51.6
a
ac-
tual
AN 176.33431 176.34236 176.34271 176.34281 103.57535 103.56794 176.34282 0 103.56595
N 3267363 1610564 1664145 8312647 0166998 4417556 7130621 8858151

AN 191.59625 196.87641 207.59604 219.03668 62.633608 55.678863 219.91503 0 51.596520


FIS 5275566 0432179 4833362 9665632 0643273 6962813 2696164 03793

PCA 202.50957 204.29249 215.38603 224.26485 60.561870 61.327724 223.47703 0 55.992856


+A 5937606 9919689 2247164 4445316 4134037 7261681 1657644 3102102
NFI
S
PCA 193.78232 194.96226 204.41658 210.94210 63.712110 61.911469 199.63069 0 57.394177
+A 4494463 0620146 9817626 3446784 29696 008234 456836 6343954
NN
PCA 198.47563 200.53158 209.55137 212.24379 48.678738 46.646161 212.88439 0 45.042331
+B 3171623 2006098 6098532 4978928 3555632 3994621 7459378 8007661
GA
+A
NN
PCA 175.06616 177.46563 190.92551 196.51163 69.686448 68.600715 199.16839 0 67.743352
+B 7268275 6713749 8337078 0347292 9185772 6067367 6398044 2123207
GA
+A
NFI
S

Figure 68 Comparison of Energy Predictions 4 Vs data actual


PAGE \* MERGEFORMAT 179
Figure 67 shows how the actual data compare to the model being tested, where the

ANFIS model is the closest model to the actual data according to Table 29, where

the correlation coefficient value is R=1.

Table 30 Comparison of Energy Predictions 4

Data PCA+ PCA+ PCA+BGA PCA+BGA


ANN ANFIS
actual ANFIS ANN +ANN +ANFIS

176.334 191.596 193.7


191.6 3133 26 202.51 82324 198.475633 175.06617
176.342 196.876 194.9
196.88 3616 41 204.29 62261 200.531582 177.46564
176.342 207.596 204.4
207.6 7117 04 215.39 1659 209.551376 190.92552
176.342 219.036 210.9
219.04 8183 69 224.26 42103 212.243795 196.51163
103.575 62.6336 63.71
62.64 3502 08 60.562 21103 48.6787384 69.686449
103.567 55.6788 61.91
55.68 9444 64 61.328 1469 46.6461614 68.600716
176.342 219.915 199.6
219.92 8271 03 223.48 30695 212.884397 199.1684
… … … … … … …
103.565 51.5965 57.39
51.6 9589 2 55.993 41776 45.0423318 67.743352

PAGE \* MERGEFORMAT 179


Computasi time 4 Training Vs Testing
151.187192
150

130

110

90

70

50

30
7.925059 5.746121 4.693943
10 0.082346 0.014511 0.011144 2.4077
0.006839 0.010058 0.600842
0.003721
ANN ANFIS PCA+ANN PCA+ANFI PCA+BGA+ PCA+BGA+
S ANN ANFIS
Training NaN NaN 151.18719 7.925059 5.746121 2.4077 4.693943 0.600842
Time ( sec- 2
onds)
Testing Time NaN NaN 0.082346 0.014511 0.011144 0.006839 0.010058 0.003721
( seconds)

Training Time ( seconds) Testing Time ( seconds)

Figure 69 Comparison of Energy Predictions 4 Vs data actual

A comparison of training time with testing time is shown in Figure 68, which shows

that the longest computation occurs in the ANN and ANFIS models. However, if

PCA feature selection is added, both the ANN and ANFIS models have relatively

faster computational speeds compared with models that are not selected selection.

This feature is observed when the ANN model has a computing speed of 151,187 s

for training and 0,082 s for testing. When added with the selection of the PCA

feature, the speed of the ANN model becomes 5,746 s for training and 0,011 s for

testing. When data are added, the selection of PCA and BGA features is relatively

faster, with 4,694 s for training and 0,010 s for testing. Similar to the ANFIS model,

PAGE \* MERGEFORMAT 179


where the computational speed is 7,925 s for training and 0,015 s for testing when

added to the PCA feature selection, the speed of the ANFIS model will be relatively

fast, i.e., 2,408 s for training and 0,007 s for testing. When the ANFIS model is

added, the PCA and BGA feature selection is 0.601 s for training and 0.004 s for

speed testing, which is faster than the model that has not been added and has been

added with feature selection.

4.9. Comparison 4 of PCA and BGA with ANN and ANFIS models and data
normalisasi

Table 31 shows the comparison between the models when normalized. It can be

seen that there is an increase and improvement in the data when normalized. It in-

creases when the ANN model without training and testing normalization shows R1

and R2 99%, but when it is normalized, the accuracy value becomes 100% with

relatively smaller errors. RSME when training and testing shows 0.00, so with MSE

it can be seen that the error is below 0%, namely 4.66E-10 for training and 8.40E-

07 for testing. You can see how the difference is when the data before being nor-

malized the error is too large even though R1 and R2 show a fairly good correlation

because it is almost close to 100%.

Table 31 Comparison of Energy Predictions 5

Correlation
PCA+ PCA+ PCA+BGA+ PCA+BGA
coefficient ANN ANFIS
ANN ANFIS ANN +ANFIS
training
Time com- 389.93 1329.2
6.38 0.791 5.183 0.636
putation 6 11
Accuration 99% 100% 99% 98% 92% 86%
1,572. 2,090.1
MSE 816.97 0 9,404.61 16,084.90
43 7
RSME 28.583 0.048 39.654 45.718 39.654 126.826

PAGE \* MERGEFORMAT 179


MAPE Inf Inf Inf Inf Inf Inf

Correlation
PCA+ PCA+ PCA+BGA+ PCA+BGA
coefficient ANN ANFIS
ANN ANFIS ANN +ANFIS
testing
Time com-
0.01 0.244 0.034 0.003 0.014 0.003
putation
Accuration 99% 92% 95% 92% 94% 93%
3,227. 39,563 42,852.
MSE
70 537.09 .65 96 20,721.77 16,084.85
198.90
RSME
56.813 23.175 6 207.01 143.951 126.826
MAPE 7.396 0.39 80.41 79.507 48.985 Inf

Correla-
tion coeffi-
cient train- PCA+ PCA+ PCA+BGA+ PCA+BGA
ANN
ing (nor- ANN ANFIS ANN +ANFIS
maliza-
tion)
Time com-
178.323 6.047 0.698 14.063 13.875
putation
Accuration 100% 92% 98% 96% 96%
MSE 4.66E-10 0 2,090.17 0 0
RSME 0 0.037 45.718 0.036 0.036
MAPE Inf Inf Inf Inf Inf

Correla-
tion coeffi-
cient test- PCA+ PCA+ PCA+BGA+ PCA+BGA
ANN
ing (nor- ANN ANFIS ANN +ANFIS
maliza-
tion)
Time com-
0.093 0.017 0.005 0.005 0.004
putation
Accuration 100% 92% 92% 96% 96%
42,852.9
MSE 8.40E-07 0.01 0 0
6
RSME 0.001 0.1 207.01 0.046 0.046
MAPE 0.081 96.483 79.507 44.578 44.578

PAGE \* MERGEFORMAT 179


4.1.0 Summary

The application of PCA and BGA features selection in energy management in a

building has contributed significantly. It can be seen that there is a decrease in the

computational speed of almost 99% of the tested models, namely the ANN and

ANFIS models. In previous studies, we discussed how to implement feature

selection using BGA and PCA [6][5], However, in this study, how the application

of PCA and BGA feature selection has contributed to improving computational

training and testing with the models tested are ANN and ANFIS. However, for

development in this research, it is expected to be able to use the accuracy model

with other algorithms such as SVM, Fuzzy and each of the many more models that

can be tested.

PAGE \* MERGEFORMAT 179


CHAPTER 5

CONCLUSION AND RECOMMENDATIONS

5.1 Conclusion

In the reference that is read as a reference in writing this dissertation, there are

many ways that can be done in managing energy in a building, one of which is to

increase accuracy in controlling in a room, especially air conditioning, lights used,

all connected devices that use a voltage source, especially in a laboratory that uses a

lot of electric voltage, in research on how to efficiently and predict energy use of

energy in a building or room with 9 variables. From the research here, the

application of feature selection in energy management using a learning machine

(ML) has already been implemented using the BGA and PCA feature selection, but

in this dissertation the process and calculations are also the same, but in this

dissertation I apply the PCA feature selection first and only after it is a new BGA

feature selection .

In this dissertation, the application of feature selection in energy management in a

PVTE UNTIRTA laboratory to predict energy in the laboratory focuses on data

speed in predicting energy, does not focus on the accuracy results obtained, and is

expected to consider speed and accuracy in predicting energy in the laboratory or in

the laboratory. a building. In the reference that is read as a reference in writing this

dissertation, there are many ways that can be done in managing energy in a

building, one of which is to increase accuracy in controlling in a room, especially

air conditioning, lights used, all connected devices that use a voltage source,

PAGE \* MERGEFORMAT 179


especially in a laboratory that uses a lot of electric voltage, in research on how to

efficiently and predict energy use of energy in a building or room with 9 variables.

From the research here, the application of feature selection in energy management

using a learning machine (ML) has already been implemented using the BGA and

PCA feature selection, but in this dissertation the process and calculations are also

the same, but in this dissertation I apply the PCA feature selection first and only

after it is a new BGA feature selection .

In this dissertation, the application of feature selection in energy management in a

PVTE UNTIRTA laboratory to predict energy in the laboratory focuses on data

speed in predicting energy, does not focus on the accuracy results obtained, and is

expected to consider speed and accuracy in predicting energy in the laboratory or in

the laboratory. a building.

Combining PCA and BGA feature selection and utilizing the looping process in

this dissertation, the stages of each stage are the same as mathematical calculations

for each of the characteristics of mature selection. However, it is hoped that in the

future, a combination of several methods can be developed to obtain better accuracy

values while still considering the speed and computation of the data when the data

is tested with the training and testing process..

5.2 Recommendations

For future improvements, it is expected that the accuracy value will also be

considered from the incorporation of the proposed feature selection.

PAGE \* MERGEFORMAT 179


i) Each model in the intelligent system has different characteristics and accuracy

values. It is hoped that in testing for subsequent developments, models other than

ANN and ANFIS can be used.

ii) The feature selection used here is connected in series or each feature selection

works according to the stages possessed by each feature selection used. It is hoped

that the incorporation of feature selection for energy management in a room or

building can be further simplified by merging in parallel so that you can find the

latest models in the feature selection that will come, but still consider the speed and

accuracy in controlling and predicting energy consumption in a room.

iii) The application of the ANFIS model in energy management can only be

simulated with 6 input variables; therefore, all nine variables cannot be used. In the

future, the application of ANFIS can be used using other models or two or more

ANFIS models.

iv) Feature selection using PCA and BGA are two very different characteristics of

feature selection when selecting input variables.

v) Consider combining the features selection used.

PAGE \* MERGEFORMAT 179


REFERENCES

[1] T. Parhizkar, E. Rafieipour, and A. Parhizkar, “Evaluation and improvement


of energy consumption prediction models using principal component analysis
based feature reduction,” J. Clean. Prod., vol. 279, 2021, doi:
10.1016/j.jclepro.2020.123866.
[2] I. E. Guyon., “An Introduction to Variable and Feature Selection,” Anal.
Chim. Acta, vol. 703, no. 2, pp. 152–162, 2011, doi:
10.1016/j.aca.2011.07.027.
[3] R. Kaur and E. Himanshi, “Face recognition using Principal Component
Analysis,” Souvenir 2015 IEEE Int. Adv. Comput. Conf. IACC 2015, pp.
585–589, 2015, doi: 10.1109/IADCC.2015.7154774.
[4] P. Wagh, R. Thakare, J. Chaudhari, and S. Patil, “Attendance system based
on face recognition using eigen face and PCA algorithms,” Proc. 2015 Int.
Conf. Green Comput. Internet Things, ICGCIoT 2015, pp. 303–308, 2016,
doi: 10.1109/ICGCIoT.2015.7380478.
[5] A. Yousaf, R. M. Asif, M. Shakir, A. U. Rehman, and M. S. Adrees, “An
improved residential electricity load forecasting using a machine-learning-
based feature selection approach and a proposed integration strategy,”
Sustain., vol. 13, no. 11, 2021, doi: 10.3390/su13116199.
[6] A. Yousaf et al., “A novel machine learning-based price forecasting for
energy management systems,” Sustain., vol. 13, no. 22, pp. 1–26, 2021, doi:
10.3390/su132212693.
[7] J. C. Lam, K. K. W. Wan, K. L. Cheung, and L. Yang, “Principal component
analysis of electricity use in office buildings,” Energy Build., vol. 40, no. 5,
pp. 828–836, 2008, doi: 10.1016/j.enbuild.2007.06.001.
[8] R. Gottumukkal and V. K. Asari, “An improved face recognition technique
based on modular PCA approach,” Pattern Recognit. Lett., vol. 25, no. 4, pp.
429–436, 2004, doi: 10.1016/j.patrec.2003.11.005.
[9] R. Platon, V. R. Dehkordi, and J. Martel, “Hourly prediction of a building’s
electricity consumption using case-based reasoning, artificial neural networks
and principal component analysis,” Energy Build., vol. 92, pp. 10–18, 2015,

PAGE \* MERGEFORMAT 179


doi: 10.1016/j.enbuild.2015.01.047.
[10] K. F. Man, K. S. Tang, and S. Kwong, “Genetic algorithms: Concepts and
applications,” IEEE Trans. Ind. Electron., vol. 43, no. 5, pp. 519–534, 1996,
doi: 10.1109/41.538609.
[11] L. Qu and W. Chen, “ENERGY CONSUMPTION PREDICTION OF
UNIVERSITY BUILDINGS IN CHINA,” pp. 1–6, 2020.
[12] M. K. M. Shapi, N. A. Ramli, and L. J. Awalin, “Energy consumption
prediction by using machine learning for smart building: Case study in
Malaysia,” Dev. Built Environ., vol. 5, p. 100037, 2021, doi:
10.1016/j.dibe.2020.100037.
[13] K. Li, C. Hu, G. Liu, and W. Xue, “Building’s electricity consumption
prediction using optimized artificial neural networks and principal
component analysis,” Energy Build., vol. 108, pp. 106–113, 2015, doi:
10.1016/j.enbuild.2015.09.002.
[14] C. Hu, K. Li, G. Liu, and L. Pan, “Forecasting building energy consumption
based on hybrid PSO-ANN prediction model,” Chinese Control Conf. CCC,
vol. 2015-Septe, pp. 8243–8247, 2015, doi: 10.1109/ChiCC.2015.7260948.
[15] A. K. Shukla, P. Singh, and M. Vardhan, “A New Hybrid Feature Subset
Selection Framework Based on Binary Genetic Algorithm and Information
Theory,” Int. J. Comput. Intell. Appl., vol. 18, no. 3, pp. 1–22, 2019, doi:
10.1142/S1469026819500202.
[16] GPI, “Energy Sector Jobs to 2030 : A Global analysis, Grrenpeace
International (GPI), Amasterdem,” 2009.
[17] J. Iwaro and A. Mwasha, “A review of building energy regulation and policy
for energy conservation in developing countries,” Energy Policy, vol. 38, no.
12, pp. 7744–7755, 2010, doi: 10.1016/j.enpol.2010.08.027.
[18] X. C. Yuan, X. Sun, W. Zhao, Z. Mi, B. Wang, and Y. M. Wei, “Forecasting
China’s regional energy demand by 2030: A Bayesian approach,” Resour.
Conserv. Recycl., vol. 127, no. May, pp. 85–95, 2017, doi:
10.1016/j.resconrec.2017.08.016.
[19] J. Eom, L. Clarke, S. H. Kim, P. Kyle, and P. Patel, “China’s building energy
demand: Long-term implications from a detailed assessment,” Energy, vol.
46, no. 1, pp. 405–419, 2012, doi: 10.1016/j.energy.2012.08.009.
PAGE \* MERGEFORMAT 179
[20] M. V. Moreno, B. Úbeda, A. F. Skarmeta, and M. A. Zamora, “How can we
tackle energy efficiency in iot based smart buildings?,” Sensors
(Switzerland), vol. 14, no. 6, pp. 9582–9614, 2014, doi:
10.3390/s140609582.
[21] A. D. Pham, N. T. Ngo, T. T. Ha Truong, N. T. Huynh, and N. S. Truong,
“Predicting energy consumption in multiple buildings using machine learning
for improving energy efficiency and sustainability,” J. Clean. Prod., vol.
260, p. 121082, 2020, doi: 10.1016/j.jclepro.2020.121082.
[22] M. Shao, X. Wang, Z. Bu, X. Chen, and Y. Wang, “Prediction of energy
consumption in hotel buildings via support vector machines,” Sustain. Cities
Soc., vol. 57, no. December 2019, p. 102128, 2020, doi:
10.1016/j.scs.2020.102128.
[23] L. L. Li, Y. W. Liu, M. L. Tseng, G. Q. Lin, and M. H. Ali, “Reducing
environmental pollution and fuel consumption using optimization algorithm
to develop combined cooling heating and power system operation strategies,”
J. Clean. Prod., vol. 247, 2020, doi: 10.1016/j.jclepro.2019.119082.
[24] Y. Chen, Z. Tong, Y. Zheng, H. Samuelson, and L. Norford, “Transfer
learning with deep neural networks for model predictive control of HVAC
and natural ventilation in smart buildings,” J. Clean. Prod., vol. 254, p.
119866, 2020, doi: 10.1016/j.jclepro.2019.119866.
[25] B. Azhagusundari and A. S. Thanamani, “Feature Selection based on
Information Gain,” Int. J. Innov. Technol. Explor. Eng., vol. 2, no. 2, pp. 18–
21, 2013.
[26] W. Q. Li and X. F. Wang, “Improved method of feature selection based on
information gain,” 2012 Spring World Congr. Eng. Technol. SCET 2012 -
Proc., pp. 6–9, 2012, doi: 10.1109/SCET.2012.6342005.
[27] G. P. McCabe, “Principle variables.” pp. 1–32, 1982.
[28] P. P. Lid and S. Planning, PRINCIPAL COMPONENTS ANALYSIS ( PCA )*
Xln 1, vol. 19, no. 3. 1993.
[29] J. Yang and V. Honavar, “Feature subset selection using genetic algorithm,”
IEEE Intell. Syst. Their Appl., vol. 13, no. 2, pp. 44–48, 1998, doi:
10.1109/5254.671091.
[30] W. F. Punch, E. D. Goodman, and R. J. Enbody, “Further Research on
PAGE \* MERGEFORMAT 179
Feature Selection and Classi cation Using Genetic Algorithms,” no. January,
1993.
[31] S. Mukherjee et al., “Version Terms of Use Feature Selection for SVMs,”
2000.
[32] L. K. Kong X, Liu X, Shi R, “Wind speed prediction using reduced support
vector machines with feature selection. Neurocomputing,” doi:
10.1016/S0164-1212(15)00073-4.
[33] P. A. Adedeji, S. Akinlabi, N. Madushele, and O. Obafemi, “Hybrid adaptive
neuro-fuzzy inference system ( ANFIS ) for a multi-campus university
energy consumption forecast,” Int. J. Ambient Energy, vol. 0, no. 0, pp. 1–10,
2020, doi: 10.1080/01430750.2020.1719885.
[34] X. Zhou, J. Yan, J. Zhu, and P. Cai, “Survey of energy consumption and
energy conservation measures for colleges and universities in Guangdong
province,” Energy Build., vol. 66, pp. 112–118, 2013, doi:
10.1016/j.enbuild.2013.07.055.
[35] K. P. Amber et al., “Energy Consumption Forecasting for University Sector
Buildings,” pp. 1–18, 2017, doi: 10.3390/en10101579.
[36] Y. Ding, Q. Wang, Z. Wang, S. Han, and N. Zhu, “Energy & Buildings An
occupancy-based model for building electricity consumption prediction : A
case study of three campus buildings in Tianjin,” Energy Build., vol. 202, p.
109412, 2019, doi: 10.1016/j.enbuild.2019.109412.
[37] H. Im and R. Srinivasan, “Building Energy Use Prediction Owing to Climate
Change : A Case Study of a University Campus,” pp. 43–50, 2019.
[38] T. Zhao, J. Xu, C. Zhang, and P. Wang, “A monitoring data based bottom-up
modeling method and its application for energy consumption prediction of
campus building,” J. Build. Eng., no. October, p. 101962, 2020, doi:
10.1016/j.jobe.2020.101962.
[39] M. V. Moreno et al., “Big data: the key to energy efficiency in smart
buildings,” Soft Comput., vol. 20, no. 5, pp. 1749–1762, 2016, doi:
10.1007/s00500-015-1679-4.
[40] M. Park, “Adaptive forecasting,” pp. 1757–1767.
[41] M. Peng, “I : nc Normalized Frequency,” vol. 8, no. 3, pp. 1195–1203, 1993.
[42] A. Azadeh, S. F. Ghaderi, and S. Sohrabkhani, “Annual electricity
PAGE \* MERGEFORMAT 179
consumption forecasting by neural network in high energy consuming
industrial sectors,” vol. 49, pp. 2272–2278, 2008, doi:
10.1016/j.enconman.2008.01.035.
[43] K. Y. Lee and S. Member, “t " I,” no. 2, pp. 450–457, 1991.
[44] H. Chen and C. Lee, “Electricity consumption prediction for buildings using
multiple adaptive network-based fuzzy inference system models and gray
relational analysis,” Energy Reports, vol. 5, pp. 1509–1524, 2019, doi:
10.1016/j.egyr.2019.10.009.
[45] G. Shi, D. Liu, and Q. Wei, “Neurocomputing Energy consumption
prediction of of fi ce buildings based on echo state networks,”
Neurocomputing, pp. 1–11, 2016, doi: 10.1016/j.neucom.2016.08.004.
[46] M . A . Girshick, “Principal Components Author ( s ): M . A . Girshick
Source : Journal of the American Statistical Association , Vol . 31 , No . 195
( Sep ., 1936 ), pp . 519- PRINCIPAL COMPONENTS * a set of statistical
PROBLEM of analyzing variables into a set of,” vol. 31, no. 195, pp. 519–
528, 1936.
[47] P. Magdalinos, C. Doulkeridis, and M. Vazirgiannis, “Enhancing Clustering
Quality through Landmark-Based,” vol. 5, no. 2, pp. 1–44, 2011, doi:
10.1145/1921632.1921637.
[48] J. Yan et al., “Effective and Efficient Dimensionality Reduction for Large-
Scale and Streaming Data,” vol. 18, no. 3, pp. 320–333, 2006.
[49] C. J. V. A. N. R. D. J. Harper and M. F. Porter, “THE SELECTION OF
GOOD SEARCH TERMS ?,” vol. 16, no. 6, pp. 271–285.
[50] C. Lee and G. G. Lee, “Information gain and divergence-based feature
selection for machine learning-based text categorization q,” vol. 42, pp. 155–
165, 2006, doi: 10.1016/j.ipm.2004.08.006.
[51] A. Hyvärinen and E. Oja, “Independent Component Analysis : Algorithms
and Applications,” vol. 1, no. 1, 2000.
[52] “Book reviews A Practical Approach to Microarray Data Analysis Edited by
Daniel P . Berrar , Werner Dubitzky and Martin Granzow Kluwer Academic
Publishers , Book reviews,” vol. 2, no. 1, pp. 82–84, 2003.
[53] M. Rezghi, “Expert Systems with Applications Noise-free principal
component analysis : An efficient dimension reduction technique for high
PAGE \* MERGEFORMAT 179
dimensional molecular data,” Expert Syst. Appl., no. June, 2014, doi:
10.1016/j.eswa.2014.06.024.
[54] Q. Zhang and C. Wang, “Using Genetic Algorithm to Optimize Artificial
Neural Network : A Case Study on Earthquake Prediction,” pp. 128–131,
2008, doi: 10.1109/WGEC.2008.96.
[55] W. Huang and H. N. Lam, “3UILDII ’, IG Using genetic algorithms to
optimize controller parameters for HVAC systems,” vol. 26, pp. 277–282,
1997.
[56] F. Boithias, M. El Mankibi, and P. Michel, “Genetic algorithms based
optimization of artificial neural network architecture for buildings ’ indoor
discomfort and energy consumption prediction,” pp. 95–106, 2012, doi:
10.1007/s12273-012-0059-6.
[57] B. B. Ekici and U. T. Aksoy, “Expert Systems with Applications Prediction
of building energy needs in early stage of design by using ANFIS,” Expert
Syst. Appl., vol. 38, no. 5, pp. 5352–5358, 2011, doi:
10.1016/j.eswa.2010.10.021.
[58] H. Alasha, B. Moghtaderi, A. Page, and H. Sugo, “A neuro – fuzzy model for
prediction of the indoor temperature in typical Australian residential
buildings,” vol. 41, pp. 703–710, 2009, doi: 10.1016/j.enbuild.2009.02.002.
[59] K. Kandananond, “Forecasting Electricity Demand in Thailand with an
Artificial Neural Network Approach,” pp. 1246–1257, 2011, doi:
10.3390/en4081246.
[60] A. Abraham and B. Nath, “A neuro-fuzzy approach for modelling electricity
demand in Victoria,” vol. 1, pp. 127–138, 2001.
[61] M. Zamarren and P. A. Gonza, “Prediction of hourly energy consumption in
buildings based on a feedback artificial neural network,” vol. 37, pp. 595–
601, 2005, doi: 10.1016/j.enbuild.2004.09.006.
[62] A. Moghaddamnia, M. G. Gousheh, J. Piri, S. Amin, and D. Han, “Advances
in Water Resources Evaporation estimation using artificial neural networks
and adaptive neuro-fuzzy inference system techniques,” Adv. Water Resour.,
vol. 32, no. 1, pp. 88–97, 2009, doi: 10.1016/j.advwatres.2008.10.005.
[63] S. Jassar, Z. Liao, and L. Zhao, “Adaptive neuro-fuzzy based inferential
sensor model for estimating the average air temperature in space heating
PAGE \* MERGEFORMAT 179
systems,” Build. Environ., vol. 44, no. 8, pp. 1609–1616, 2009, doi:
10.1016/j.buildenv.2008.10.002.
[64] K. Li, H. Su, and J. Chu, “Forecasting building energy consumption using
neural networks and hybrid neuro-fuzzy system: A comparative study,”
Energy Build., vol. 43, no. 10, pp. 2893–2899, 2011, doi:
10.1016/j.enbuild.2011.07.010.
[65] K. P. Amber, R. Ahmad, M. W. Aslam, A. Kousar, M. Usman, and M. S.
Khan, “Intelligent Techniques for Forecasting Electricity Consumption of
Buildings,” Energy, 2018, doi: 10.1016/j.energy.2018.05.155.
[66] D. Hawkins, S. M. Hong, R. Raslan, D. Mumovic, and S. Hanna,
“Determinants of energy use in UK higher education buildings using
statistical and artificial neural network methods,” Int. J. Sustain. Built
Environ., vol. 1, no. 1, pp. 50–63, 2012, doi: 10.1016/j.ijsbe.2012.05.002.
[67] E. Momeni, R. Nazir, D. J. Armaghani, and H. Maizir, “Prediction of pile
bearing capacity using a hybrid genetic algorithm-based ANN,”
MEASUREMENT, vol. 57, pp. 122–131, 2014, doi:
10.1016/j.measurement.2014.08.007.
[68] H. C. W. Lau, “A fuzzy logic approach to forecast energy consumption
change in a manufacturing system,” vol. 34, pp. 1813–1824, 2008, doi:
10.1016/j.eswa.2007.02.015.
[69] M. Collan, “Fuzzy logic and intelligent agents : towards the next step of
capital budgeting decision support,” pp. 410–422, 2000, doi:
10.1108/02635570310479981.
[70] G. T. S. Ho, H. C. W. Lau, C. K. M. Lee, and A. W. H. Ip, “An intelligent
forward quality enhancement system to achieve product customization,”
2004, doi: 10.1108/02635570510590174.
[71] J. O. Jaber and R. Mamlook, “Evaluation of energy conservation programs in
residential sector using fuzzy logic methodology,” vol. 33, pp. 1329–1338,
2005, doi: 10.1016/j.enpol.2003.12.009.
[72] S. S. Pappas, L. Ekonomou, D. C. Karamousantas, G. E. Chatzarakis, S. K.
Katsikas, and P. Liatsis, “Electricity demand loads modeling using
AutoRegressive Moving Average ( ARMA ) models,” vol. 33, pp. 1353–
1360, 2008, doi: 10.1016/j.energy.2008.05.008.
PAGE \* MERGEFORMAT 179
[73] H. P. Ã, “Comparing linear and nonlinear forecasts for Taiwan ’ s electricity
consumption,” vol. 31, pp. 2129–2141, 2006, doi:
10.1016/j.energy.2005.08.010.
[74] K. Kavaklioglu, H. Ceylan, H. Kemal, and O. Ersel, “Modeling and
prediction of Turkey ’ s electricity consumption using Artificial Neural
Networks,” Energy Convers. Manag., vol. 50, no. 11, pp. 2719–2727, 2009,
doi: 10.1016/j.enconman.2009.06.016.
[75] S. Kucukali and K. Baris, “Turkey ’ s short-term gross annual electricity
demand forecast by fuzzy logic approach,” Energy Policy, vol. 38, no. 5, pp.
2438–2445, 2010, doi: 10.1016/j.enpol.2009.12.037.
[76] H. Kemal, H. Ceylan, O. Ersel, and A. Hepbasli, “Electricity estimation
using genetic algorithm approach : a case study of Turkey,” vol. 30, pp.
1003–1012, 2005, doi: 10.1016/j.energy.2004.08.008.
[77] S. Önüt and C. Kahraman, “Expert Systems with Applications A decision
support system for demand forecasting with artificial neural networks and
neuro-fuzzy models : A comparative analysis,” vol. 36, pp. 6697–6707, 2009,
doi: 10.1016/j.eswa.2008.08.058.
[78] A. Al-ghandoor, M. Samhouri, I. Al-hinti, J. Jaber, and M. Al-rawashdeh,
“Projection of future transport energy demand of Jordan using adaptive
neuro-fuzzy technique,” Energy, vol. 38, no. 1, pp. 128–135, 2012, doi:
10.1016/j.energy.2011.12.023.
[79] A. Khosravi, S. Syri, J. J. G. Pabon, M. H. Barrientos, O. R. Sandoval, and
B. C. Caetano, “Energy modeling of a solar dish / Stirling by arti fi cial
intelligence approach,” Energy Convers. Manag., vol. 199, no. March, p.
112021, 2019, doi: 10.1016/j.enconman.2019.112021.
[80] M. Kim, S. Jung, and J. Kang, “Artificial Neural Network-Based Residential
Energy Consumption Prediction Models Considering Residential Building
Information and User Features in South Korea,” 2020.
[81] D. Bui, T. N. Nguyen, and T. D. Ngo, “An artificial neural network (ANN)
expert system enhanced with the electromagnetism-based firefly algorithm
(EFA) for predicting the energy consumption in buildings,” Energy, 2019,
doi: 10.1016/j.energy.2019.116370.
[82] N. Modeling, I. Rivals, and L. Personnaz, “Neural-Network Construction and
PAGE \* MERGEFORMAT 179
Selection in,” vol. 14, no. 4, pp. 804–819, 2003.
[83] S. Karatasou, M. Santamouris, and V. Geros, “Modeling and predicting
building’s energy use with artificial neural networks: Methods and results,”
Energy Build., vol. 38, no. 8, pp. 949–958, 2006, doi:
10.1016/j.enbuild.2005.11.005.
[84] R. Chauhan and H. Kaur, Predictive analytics and data mining: A framework
for optimizing decisions with R tool. 2013.
[85] R. A.Johnson, “Applierd Multivariate Statistical Analisis,” sixth Ed.
( Pearson) Prentice Hall), p. 174, 2005.
[86] I. T. Jolliffe, “Principal Component Analysis, Second Edition,” Encycl. Stat.
Behav. Sci., vol. 30, no. 3, p. 487, 2002, doi: 10.2307/1270093.
[87] R.A Johnson, Aplplied Multivariate Statistical Analysis, 6th ed. person
pratice hall, 2007.
[88] J. Cai, J. Luo, S. Wang, and S. Yang, “Feature selection in machine learning:
A new perspective,” Neurocomputing, vol. 300, pp. 70–79, 2018, doi:
10.1016/j.neucom.2017.11.077.
[89] S. Lei, “A feature selection method based on information gain and genetic
algorithm,” Proc. - 2012 Int. Conf. Comput. Sci. Electron. Eng. ICCSEE
2012, vol. 2, pp. 355–358, 2012, doi: 10.1109/ICCSEE.2012.97.
[90] “Feature selection in machine learning prediction systems for renewable
energy applications,” Renew. Sustain. Energy Rev., vol. 90, no. December
2016, pp. 728–741, 2018, doi: 10.1016/j.rser.2018.04.008.
[91] J. H. Holland, C. Langton, and S. W. Wilson, Complex Adaptive Systems
Bradford Books edition. 1992.
[92] David E. Goldberg, “Genetic Algorithms in Search, Optimization, and
Machine Learning,” American Society of Mechanical Engineers, Dynamic
Systems and Control Division (Publication) DSC, vol. 22. pp. 39–45, 1989.
[93] M. Melanie and A, An Introduction to Genetic Algorithms. A Bradford Book
The MIT Press Cambridge, Massachusetts • London, England Fifth, 1999.
[94] S. S. Rao, Engineering Optimization - Theory and Practice-Wiley (2009).
2009.
[95] C. Yang and Q. Wu, “Improved Genetic Algorithm based on Binary
Coding,” vol. 7, no. 8, pp. 201–207, 2020.
PAGE \* MERGEFORMAT 179
[96] R. C. Mitsuo Gen, “Genetic Algorithms and Engineering Design (Book
Review),” Int. J. Hum. Comput. Interact., vol. 9, no. 4, pp. 457–458, 1997,
doi: 10.1207/s15327590ijhc0904_9.
[97] J. J. Grefenstette, “Optimization of Control Parameters for Genetic
Algorithms,” IEEE Trans. Syst. Man Cybern., vol. 16, no. 1, pp. 122–128,
1986, doi: 10.1109/TSMC.1986.289288.
[98] A. Konar, Computational Intelligence Principles, Techniques and
Applications. 2005.
[99] S. S. Shreem, H. Turabieh, S. Al Azwari, and F. Baothman, “Enhanced
binary genetic algorithm as a feature selection to predict student
performance,” Soft Comput., vol. 26, no. 4, pp. 1811–1823, 2022, doi:
10.1007/s00500-021-06424-7.
[100] A. Kumar, “Encoding Schemes in Genetic Algorithm,” Int. J. Adv. Res. IT
Eng., vol. 2, no. 3, pp. 1–7, 2013, [Online]. Available: www.garph.co.uk.
[101] R. A. Gupta, R. Kumar, and A. K. Bansal, “Selection of input variables for
the prediction of wind speed in wind farms based on genetic algorithm,”
Wind Eng., vol. 35, no. 6, pp. 649–660, 2011, doi: 10.1260/0309-
524X.35.6.649.
[102] H. Mühlenbein and D. Schlierkamp-Voosen, “Predictive Models for the
Breeder Genetic Algorithm I. Continuous Parameter Optimization,” Evol.
Comput., vol. 1, no. 1, pp. 25–49, 1993, doi: 10.1162/evco.1993.1.1.25.
[103] N. M. Razali and J. Geraghty, “Genetic algorithm performance with different
selection strategiesin solving TSP,” Proc. World Congr. Eng. 2011, WCE
2011, vol. 2, pp. 1134–1139, 2011.
[104] K. Deb and R. B. Agrawal, “Simulated Binary Crossover for Continuous
Search Space The crossover operator is believed to be the main search
operator in the working of a genetic,” Complex Syst., vol. 9, no. 2, pp. 115–
148, 1994.
[105] Y. Xue, H. Zhu, J. Liang, and A. Słowik, “Knowledge-Based Systems
Adaptive crossover operator based multi-objective binary genetic algorithm
for feature selection in classification,” Knowledge-Based Syst., vol. 227, p.
107218, 2021, doi: 10.1016/j.knosys.2021.107218.
[106] A. J. Umbarkar and P. D. Sheth, “CROSSOVER OPERATORS IN
PAGE \* MERGEFORMAT 179
GENETIC ALGORITHMS : A REVIEW,” vol. 6956, no. October, pp.
1083–1092, 2015, doi: 10.21917/ijsc.2015.0150.
[107] W. M. Spears and K. A. De Jong, An Analysis of Multi-Point Crossover, vol.
1. Morgan Kaufmann Publishers, Inc.
[108] R. Poli and W. B. Langdon, “On the Ability to Search the Space of Programs
of Standard , One-point and Uniform Crossover in Genetic Programming,”
no. February 1998, 2015.
[109] F. J. Burkowski, “Shuffle Crossover and Mutual Information,” no. DeJong
1992, pp. 1–7, 1997.
[110] V. A. Cicirello and R. Cernera, “Profiling the distance characteristics of
mutation operators for permutation-based genetic algorithms,” FLAIRS 2013
- Proc. 26th Int. Florida Artif. Intell. Res. Soc. Conf., pp. 46–51, 2013.
[111] S. Barak and S. S. Sadegh, “Forecasting energy consumption using ensemble
ARIMA-ANFIS hybrid algorithm,” Int. J. Electr. Power Energy Syst., vol.
82, pp. 92–104, 2016, doi: 10.1016/j.ijepes.2016.03.012.
[112] D. He and R. Y. Liu, “Ultra-short-term wind power prediction using ANN
ensemble based on the principal components analysis,” Dianli Xitong Baohu
yu Kongzhi/Power Syst. Prot. Control, vol. 41, no. 4, pp. 50–54, 2013.
[113] F. Salo, A. B. Nassif, and A. Essex, “Dimensionality reduction with IG-PCA
and ensemble classifier for network intrusion detection,” Comput. Networks,
vol. 148, pp. 164–175, 2019, doi: 10.1016/j.comnet.2018.11.010.
[114] J. Yun and K. H. Won, “Building environment analysis based on temperature
and humidity for smart energy systems,” Sensors (Switzerland), vol. 12, no.
10, pp. 13458–13470, 2012, doi: 10.3390/s121013458.
[115] Y. Zhou, G. Cheng, S. Jiang, and M. Dai, “Building an efficient intrusion
detection system based on feature selection and ensemble classifier,”
Comput. Networks, vol. 174, no. April, 2020, doi:
10.1016/j.comnet.2020.107247.
[116] Z. Hu, Y. Bao, R. Chiong, and T. Xiong, “Mid-term interval load forecasting
using multi-output support vector regression with a memetic algorithm for
feature selection,” Energy, vol. 84, pp. 419–431, 2015, doi:
10.1016/j.energy.2015.03.054.
[117] Z. Michalewicz, Genetic Algorithms+Data Stuctures= Evolution Programs,
PAGE \* MERGEFORMAT 179
no. July. 2016.
[118] F. Herrera, M. Lozano, and J. L. Verdegay, “Tackling Real-Coded Genetic
Algorithms: Operators and Tools for Behavioural Analysis,” Artif. Intell.
Rev., vol. 12, no. 4, pp. 265–319, 1998, doi: 10.1023/A:1006504901164.
[119] C. To and J. Vohradsky, “Binary classification using parallel genetic
algorithm,” 2007 IEEE Congr. Evol. Comput. CEC 2007, no. October 2007,
pp. 1281–1287, 2007, doi: 10.1109/CEC.2007.4424618.
[120] X. J. Luo, L. O. Oyedele, A. O. Ajayi, O. O. Akinade, H. A. Owolabi, and A.
Ahmed, “Feature extraction and genetic algorithm enhanced adaptive deep
neural network for energy consumption prediction in buildings,” Renew.
Sustain. Energy Rev., vol. 131, no. June, p. 109980, 2020, doi:
10.1016/j.rser.2020.109980.
[121] F. Wahid, R. Ghazali, A. S. Shah, and M. Fayaz, “Prediction of Energy
Consumption in the Buildings Using Multi-Layer Perceptron and Random
Forest,” Int. J. Adv. Sci. Technol., vol. 101, pp. 13–22, 2017, doi:
10.14257/ijast.2017.101.02.
[122] “Forecasting Energy Consumption of Institutional Buildings in(2015).” .
[123] S. Hong et al., “Improved benchmarking comparability for energy
consumption in schools Improved benchmarking comparability for energy
consumption in schools,” vol. 3218, 2014, doi:
10.1080/09613218.2013.814746.
[124] B. Bektas Ekici and U. T. Aksoy, “Prediction of building energy needs in
early stage of design by using ANFIS,” Expert Syst. Appl., vol. 38, no. 5, pp.
5352–5358, 2011, doi: 10.1016/j.eswa.2010.10.021.
[125] Y. Liu, H. Chen, L. Zhang, and Z. Feng, “Enhancing building energy
efficiency using a random forest model: A hybrid prediction approach,”
Energy Reports, vol. 7, pp. 5003–5012, 2021, doi:
10.1016/j.egyr.2021.07.135.

PAGE \* MERGEFORMAT 179


APPENDIX A Prediction energy model ANN

Quantity
Date Tem Predic-
Tem Cur- Pow p© Alti- Data
No and Light tion
Pres- p© rent Volt- er (BM tude Hu- Actual
time (lux) Result
sure (DH (A) age (W) P180 (m) midity
(atm) T22) (V) ) (%)
Mon-
day, 0.992 26. 225.2 16.8 26.0 135.11 270.512
1 0.07 38.06 41.15 75.60
6/01/2 8 28 3 9 3 184 162
020
Tues-
day, 0.992 25. 225.9 12.4 26.7 99.660 268.907
2 0.06 42.54 47.32 74.06
7/01/2 1 74 1 6 1 408 36
020
Wedn
esday, 0.991 26. 225.6 14.9 26.1 119.27 269.163
3 0.07 37.75 47.88 76.25
8/01/2 7 34 6 1 9 51 045
020
Thus-
day, 0.992 26. 221.2 39.0 27.4 312.56 275.908
4 0.18 44.63 43.94 63.34
9/01/2 3 76 3 7 1 163 288
020
Friday,
0.991 26. 225.8 16.5 25.9 132.15 269.108
5 10/01/ 0.07 37.31 49.71 76.48
6 04 7 2 1 347 408
2020
Mon-
day, 0.991 26. 226.2 29.7 26.9 237.80 271.659
6 0.13 67.50 48.39 64.03
13/01/ 7 32 6 3 6 898 958
2020
Tues-
day,14 0.992 25. 226.0 17.1 25.5 136.77 269.544
7 0.08 37.76 46.45 71.38
/01/20 2 70 9 0 4 061 788
20
Wedn
esday, 0.991 27. 223.1 32.6 100.4 28.2 261.10 273.060
8 0.15 51.40 56.96
15/01/ 6 36 1 4 0 4 204 935
2020
9 Thus- 0.990 27. 0.11 223.2 24.4 67.16 27.8 49.99 62.17 195.77 269.807
PAGE \* MERGEFORMAT 179
day,
16/01/ 5 03 0 7 4 796 936
2020
Friday,
0.991 26. 170.4 29.1 27.0 233.27 273.882
10 17/01/ 0.16 32.04 47.04 55.62
9 63 4 6 3 184 953
2020
Mon-
day, 0.992 25. 225.2 34.0 25.6 272.54 273.079
11 0.15 66.43 45.04 63.24
20/01/ 3 78 4 7 4 857 341
2020
Tues-
day, 0.991 25. 224.7 38.8 26.9 310.41 273.112
12 0.17 50.74 51.35 64.33
21/01/ 6 99 4 0 3 469 956
2020
Wedn
esday, 0.992 25. 225.7 26.9 25.7 215.62 272.010
13 0.12 37.01 42.51 66.68
22/01/ 6 94 1 5 6 612 646
2020
Thus-
day, 0.992 25. 226.0 25.8 25.3 206.41 271.311
14 0.11 52.76 42.67 66.77
23/01/ 6 29 3 0 0 633 731
2020
Friday,
0.992 21. 227.0 24.9 21.9 199.90 268.753
15 24/01/ 0.11 58.83 44.05 60.93
4 66 8 9 2 204 55
2020
Mon-
day, 0.991 28. 220.0 27.3 100.0 29.1 219.07 272.450
16 0.13 53.00 55.09
27/01/ 5 07 6 8 6 3 102 673
2020
Tues-
day, 0.992 27. 220.5 26.2 104.6 27.7 209.62 273.043
17 0.12 42.35 54.92
28/01/ 6 04 7 0 6 9 612 253
2020
Wedn
esday, 0.992 26. 223.5 27.6 27.1 221.32 272.351
18 0.12 66.53 45.29 56.63
29/01/ 4 52 6 7 0 898 57
2020
Thus-
day, 0.992 26. 226.7 38.0 27.6 304.44 275.374
19 0.17 54.76 43.56 55.34
30/01/ 5 66 3 6 2 571 407
2020
Friday,
0.992 28. 226.3 37.5 28.9 300.68 276.766
20 31/01/ 0.17 48.40 46.10 55.45
2 60 3 9 4 571 017
2020
PAGE \* MERGEFORMAT 179
Mon-
day, 0.990 24. 223.8 25.2 26.3 201.82 269.441
21 0.12 44.14 55.76 66.98
3/02/2 9 90 8 3 4 204 702
020
Tues-
day, 0.991 27. 212.7 29.4 28.5 235.41 271.853
22 0.13 52.15 51.84 60.93
4/02/2 3 15 4 3 2 388 2
020
Wedn
esday, 0.991 25. 222.2 28.5 28.1 228.03 270.602
23 0.13 52.47 53.91 62.14
5/02/2 1 99 2 0 7 592 723
020
Thus-
day, 0.992 26. 226.1 39.5 26.6 316.50 275.735
24 0.17 31.11 41.07 61.32
6/02/2 8 03 7 6 4 449 264
020
Tues-
day, 0.992 26. 221.2 40.2 27.3 322.31 276.047
25 0.18 43.61 44.04 63.11
7/02/2 5 66 1 9 3 51 946
020
Mon-
day, 0.992 26. 225.2 36.7 26.4 294.15 273.845
26 0.16 49.49 44.98 67.68
10/02/ 2 02 3 7 1 51 679
2020
Tues-
day, 0.992 26. 224.3 36.1 104.7 27.0 289.53 274.442
27 0.16 43.41 61.40
11/02/ 5 11 6 9 2 3 959 906
2020
Wedn
esday, 0.992 26. 226.2 40.0 26.9 320.56 274.779
28 0.17 38.13 49.21 63.51
12/02/ 1 50 3 7 7 816 996
2020
Thus-
day, 0.991 26. 224.9 31.8 27.2 254.53 272.673
29 0.14 44.57 49.52 58.78
13/02/ 8 66 9 2 3 224 058
2020
Friday,
0.992 26. 227.5 52.6 26.8 421.01 278.159
30 14/02/ 0.23 16.75 47.32 61.14
1 19 8 3 5 878 088
2020
Mon-
day 0.993 28. 222.8 38.0 28.4 304.42 278.052
31 0.17 68.54 39.45 55.43
17/02/ 0 09 5 5 8 286 701
2020
32 Tues- 0.991 25. 0.54 221.8 116. 53.80 26.5 50.80 61.27 928.88 288.617
PAGE \* MERGEFORMAT 179
day,
18/02/ 7 63 6 11 1 327 293
2020
Wedn
esday, 0.992 25. 211.6 38.1 26.4 305.42 273.646
33 0.18 55.11 47.87 58.40
19/02/ 1 06 6 8 7 367 135
2020
Thus-
day, 0.991 26. 221.8 63.0 26.5 504.52 272.283
34 0.15 25.22 51.98 55.98
20/02/ 5 16 8 7 6 408 657
2020
Friday,
0.993 27. 219.2 43.7 27.7 349.75 275.491
35 21/02/ 0.13 48.61 34.28 59.57
6 46 6 2 8 837 78
2020
Mon-
day, 0.992 25. 155.8 20.0 26.5 160.75 273.082
36 0.15 58.85 44.42 56.83
24/02/ 3 79 2 9 9 265 496
2020
Tues-
day, 0.992 25. 223.4 28.4 25.8 227.52 274.648
37 0.18 57.91 43.73 57.34
25/02/ 5 19 9 4 4 49 942
2020
Wedn
esday, 0.990 27. 222.4 14.9 28.2 119.75 275.473
38 0.21 44.58 60.78 61.09
26/02/ 5 61 5 7 5 51 611
2020
Thus-
day, 0.992 25. 105.3 13.3 26.3 106.72 273.936
39 0.16 63.98 44.62 54.94
27/02/ 4 64 3 4 3 163 294
2020
Friday,
0.991 26. 197.4 26.9 27.4 215.59 272.203
40 28/02/ 0.15 36.91 54.36 58.78
0 89 1 5 6 347 852
2020
Mon-
day, 0.991 27. 223.8 24.0 28.4 192.01 271.321
41 0.10 47.30 47.82 57.97
2/03/2 9 62 1 0 3 796 223
020
Tues-
day, 0.992 26. 225.3 31.1 27.6 249.04 277.419
42 0.22 41.31 45.76 55.36
3/03/2 2 24 6 3 1 163 829
020
43 Wedn 0.992 26. 0.10 225.8 25.6 102.6 28.2 40.98 53.55 205.08 271.959
esday, 7 74 4 4 7 0 735 468
4/03/2
PAGE \* MERGEFORMAT 179
020
Thus-
day, 0.992 27. 225.7 30.8 27.6 246.80 277.011
44 0.19 55.70 44.72 43.30
5/03/2 4 33 4 5 3 327 372
020
Friday,
0.992 26. 225.3 31.1 27.6 249.04 277.419
45 6/03/2 0.22 41.31 45.76 55.36
2 24 6 3 1 163 829
020
Thurs-
day, 0.990 29. 222.5 35.6 30.5 285.02 287.349
46 0.41 67.76 56.35 71.18
12/03/ 9 50 6 3 4 531 502
2020
Friday,
0.989 30. 222.0 63.7 31.3 510.12 286.190
47 13/03/ 0.40 66.46 70.64 72.58
3 26 3 7 3 245 786
2020
Mon-
day, 0.992 26. 221.1 151. 27.2 1210.1 286.351
48 0.38 32.98 46.21 57.00
16/03/ 2 46 4 27 7 976 938
2020
Tues-
day, 0.990 28. 218.7 43.5 29.1 348.72 285.427
49 0.38 46.22 62.31 55.61
17/03/ 3 36 1 9 9 163 97
2020
Wedn
esday, 0.992 26. 223.3 29.4 27.0 235.40 287.491
50 0.42 68.11 44.96 48.49
18/03/ 3 48 8 3 9 082 411
2020
Thus-
day, 0.993 24. 219.2 43.6 25.7 349.33 287.597
51 0.43 53.38 37.56 46.33
19/03/ 2 95 3 7 2 551 534
2020
Friday,
0.991 26. 224.0 26.8 27.2 215.00 287.309
52 20/03/ 0.43 46.53 49.40 50.49
8 61 8 8 3 082 03
2020
Mon-
day, 0.991 28. 223.4 28.2 28.9 225.66 287.294
53 0.41 78.28 52.20 57.66
23/03/ 5 21 3 1 9 857 074
2020
Tues-
day, 0.993 27. 222.7 23.7 28.3 190.08 288.061
54 0.40 79.08 33.92 62.23
24/03/ 6 41 9 6 0 98 576
2020
Wedn 0.990 28. 225.9 40.7 29.2 326.23 285.183
55 0.38 86.43 63.70 56.99
esday, 1 56 2 8 1 673 355
PAGE \* MERGEFORMAT 179
25/03/
2020
Thus-
day, 0.991 27. 225.6 38.6 28.0 309.29 285.416
56 0.37 76.53 56.24 60.79
26/03/ 2 45 7 6 5 469 668
2020
Friday,
0.991 29. 222.3 27.9 30.5 223.44 287.641
57 27/03/ 0.41 65.71 54.08 77.59
3 71 8 3 3 163 872
2020
Mon-
day, 0.991 26. 221.2 27.9 27.1 223.81 282.856
58 0.34 65.39 56.90 45.41
30/03/ 0 11 1 8 8 388 667
2020
Tues-
day, 0.991 27. 223.4 29.6 29.3 237.32 286.801
59 0.40 73.30 52.58 51.91
31/03/ 4 89 9 7 0 735 456
2020
Wedn
esday, 0.990 28. 226.1 28.7 29.4 230.29 283.739
60 0.33 60.48 57.85 47.23
1/04/2 8 14 8 9 9 224 466
020
Thus-
day, 0.991 27. 226.3 39.6 29.0 317.43 286.400
61 0.40 64.27 54.54 45.26
2/04/2 2 16 8 8 6 837 581
020
Friday,
0.990 30. 225.1 45.9 32.4 367.34 286.698
62 3/04/2 0.40 66.34 64.21 74.14
1 19 3 2 8 694 352
020
Mon-
day, 0.991 25. 225.4 36.1 27.9 289.20 285.682
63 0.40 64.63 52.94 44.42
6/04/2 4 84 7 5 4 49 957
020
Tues-
day, 0.992 27. 225.5 48.5 28.5 387.99 287.158
64 0.40 74.63 46.34 53.21
7/04/2 2 31 9 0 1 837 152
020
Wedn
esday, 0.992 26. 225.6 45.4 27.1 363.24 286.953
65 0.42 61.85 47.96 53.69
8/04/2 0 10 3 1 5 898 195
020
Thus-
day, 0.993 28. 227.1 54.0 29.6 432.04 287.680
66 0.38 66.22 39.93 68.84
9/04/2 1 44 5 1 1 082 642
020
PAGE \* MERGEFORMAT 179
Friday,
0.993 27. 224.7 48.9 29.0 391.76 287.549
67 10/04/ 0.38 67.52 37.75 61.26
2 73 5 7 0 653 815
2020
44.64 118 9801. 1446 2424. 1215. 2107. 2747. 11572. 12309.9
Total 25 5.63 6.90 73 .60 22 28 59 41 82 6
0.992 26. 217.8 32.1 27.0
Average 1 35 0.15 2 5 53.87 1 46.84 61.05 257.17 273.55

APPENDIX B Prediction energy model ANFIS

Quantity
Date Tem Predic-
Lig Data
No and Tem Cur- p© Alti- tion
Power ht Actual
Pres- p© rent Volt- (BM tude Hu- Result
time (W) (lux
sure (DHT (A) age P180 (m) midity
)
(atm) 22) (V) ) (%)
Mon-
day, 26.2 0.0 225.2 38. 26.0 41.1 135.11
1 0.9928 16.89 75.60 155.826
6/01/2 8 7 3 06 3 5 184
020
Tues-
day, 25.7 0.0 225.9 42. 26.7 47.3 99.660 151.950
2 0.9921 12.46 74.06
7/01/2 4 6 1 54 1 2 408 8
020
Wedne
sday, 26.3 0.0 225.6 37. 26.1 47.8 119.27 154.656
3 0.9917 14.91 76.25
8/01/2 4 7 6 75 9 8 51 9
020
Thus-
day, 26.7 0.1 221.2 44. 27.4 43.9 312.56 315.844
4 0.9923 39.07 63.34
9/01/2 6 8 3 63 1 4 163 3
020
Friday,
26.0 0.0 225.8 37. 25.9 49.7 132.15 170.597
5 10/01/ 0.9916 16.52 76.48
4 7 7 31 1 1 347 9
2020
Mon-
day, 26.3 0.1 226.2 67. 26.9 48.3 237.80 240.238
6 0.9917 29.73 64.03
13/01/ 2 3 6 50 6 9 898 4
2020
Tues-
day,14/ 25.7 0.0 226.0 37. 25.5 46.4 136.77 174.910
7 0.9922 17.10 71.38
01/202 0 8 9 76 4 5 061 5
0
8 Wedne 0.9916 27.3 0.1 223.1 32.64 100 28.2 51.4 56.96 261.10 256.254
PAGE \* MERGEFORMAT 179
sday,
15/01/ 6 5 1 .40 4 0 204 4
2020
Thus-
day, 27.0 0.1 223.2 67. 27.8 49.9 195.77 205.570
9 0.9905 24.47 62.17
16/01/ 3 1 0 16 4 9 796 6
2020
Friday,
26.6 0.1 170.4 32. 27.0 47.0 233.27 282.835
10 17/01/ 0.9919 29.16 55.62
3 6 4 04 3 4 184 4
2020
Mon-
day, 25.7 0.1 225.2 66. 25.6 45.0 272.54 261.788
11 0.9923 34.07 63.24
20/01/ 8 5 4 43 4 4 857 7
2020
Tues-
day, 25.9 0.1 224.7 50. 26.9 51.3 310.41 283.976
12 0.9916 38.80 64.33
21/01/ 9 7 4 74 3 5 469 2
2020
Wedne
sday, 25.9 0.1 225.7 37. 25.7 42.5 215.62 221.888
13 0.9926 26.95 66.68
22/01/ 4 2 1 01 6 1 612 9
2020
Thus-
day, 25.2 0.1 226.0 52. 25.3 42.6 206.41 213.978
14 0.9926 25.80 66.77
23/01/ 9 1 3 76 0 7 633 4
2020
Friday,
21.6 0.1 227.0 58. 21.9 44.0 199.90 200.038
15 24/01/ 0.9924 24.99 60.93
6 1 8 83 2 5 204 4
2020
Mon-
day, 28.0 0.1 220.0 100 29.1 53.0 219.07 226.518
16 0.9915 27.38 55.09
27/01/ 7 3 6 .06 3 0 102 5
2020
Tues-
day, 27.0 0.1 220.5 104 27.7 42.3 209.62 225.877
17 0.9926 26.20 54.92
28/01/ 4 2 7 .66 9 5 612 8
2020
Wedne
sday, 26.5 0.1 223.5 66. 27.1 45.2 221.32
18 0.9924 27.67 56.63 229.466
29/01/ 2 2 6 53 0 9 898
2020
19 Thus- 0.9925 26.6 0.1 226.7 38.06 54. 27.6 43.5 55.34 304.44 299.536
day, 6 7 3 76 2 6 571 5
30/01/
PAGE \* MERGEFORMAT 179
2020
Friday,
28.6 0.1 226.3 48. 28.9 46.1 300.68
20 31/01/ 0.9922 37.59 55.45 293.357
0 7 3 40 4 0 571
2020
Mon-
day, 24.9 0.1 223.8 44. 26.3 55.7 201.82 240.154
21 0.9909 25.23 66.98
3/02/2 0 2 8 14 4 6 204 7
020
Tues-
day, 27.1 0.1 212.7 52. 28.5 51.8 235.41 238.161
22 0.9913 29.43 60.93
4/02/2 5 3 4 15 2 4 388 5
020
Wedne
sday, 25.9 0.1 222.2 52. 28.1 53.9 228.03
23 0.9911 28.50 62.14 238.168
5/02/2 9 3 2 47 7 1 592
020
Thus-
day, 26.0 0.1 226.1 31. 26.6 41.0 316.50 301.261
24 0.9928 39.56 61.32
6/02/2 3 7 7 11 4 7 449 2
020
Tues-
day, 26.6 0.1 221.2 43. 27.3 44.0 322.31
25 0.9925 40.29 63.11 315.725
7/02/2 6 8 1 61 3 4 51
020
Mon-
day, 26.0 0.1 225.2 49. 26.4 44.9 294.15 280.125
26 0.9922 36.77 67.68
10/02/ 2 6 3 49 1 8 51 3
2020
Tues-
day, 26.1 0.1 224.3 104 27.0 43.4 289.53 282.877
27 0.9925 36.19 61.40
11/02/ 1 6 6 .72 3 1 959 7
2020
Wedne
sday, 26.5 0.1 226.2 38. 26.9 49.2 320.56
28 0.9921 40.07 63.51 299.138
12/02/ 0 7 3 13 7 1 816
2020
Thus-
day, 26.6 0.1 224.9 44. 27.2 49.5 254.53 256.377
29 0.9918 31.82 58.78
13/02/ 6 4 9 57 3 2 224 4
2020
Friday,
26.1 0.2 227.5 16. 26.8 47.3 421.01 372.520
30 14/02/ 0.9921 52.63 61.14
9 3 8 75 5 2 878 8
2020
31 Mon- 0.9930 28.0 0.1 222.8 38.05 68. 28.4 39.4 55.43 304.42 317.438
PAGE \* MERGEFORMAT 179
day
17/02/ 9 7 5 54 8 5 286
2020
Tues-
day, 25.6 0.5 221.8 53. 26.5 50.8 928.88 736.931
32 0.9917 116.11 61.27
18/02/ 3 4 6 80 1 0 327 6
2020
Wedne
sday, 25.0 0.1 211.6 55. 26.4 47.8 305.42 277.162
33 0.9921 38.18 58.40
19/02/ 6 8 6 11 7 7 367 3
2020
Thus-
day, 26.1 0.1 221.8 25. 26.5 51.9 504.52 263.690
34 0.9915 63.07 55.98
20/02/ 6 5 8 22 6 8 408 9
2020
Friday,
27.4 0.1 219.2 48. 27.7 34.2 349.75 240.449
35 21/02/ 0.9936 43.72 59.57
6 3 6 61 8 8 837 2
2020
Mon-
day, 25.7 0.1 155.8 58. 26.5 44.4 160.75 261.416
36 0.9923 20.09 56.83
24/02/ 9 5 2 85 9 2 265 7
2020
Tues-
day, 25.1 0.1 223.4 57. 25.8 43.7 227.52 282.923
37 0.9925 28.44 57.34
25/02/ 9 8 9 91 4 3 49 5
2020
Wedne
sday, 27.6 0.2 222.4 44. 28.2 60.7 119.75
38 0.9905 14.97 61.09 301.184
26/02/ 1 1 5 58 5 8 51
2020
Thus-
day, 25.6 0.1 105.3 63. 26.3 44.6 106.72 276.206
39 0.9924 13.34 54.94
27/02/ 4 6 3 98 3 2 163 7
2020
Friday,
26.8 0.1 197.4 36. 27.4 54.3 215.59 259.893
40 28/02/ 0.9910 26.95 58.78
9 5 1 91 6 6 347 1
2020
Mon-
day, 27.6 0.1 223.8 47. 28.4 47.8 192.01 186.151
41 0.9919 24.00 57.97
2/03/2 2 0 1 30 3 2 796 2
020
42 Tues- 0.9922 26.2 0.2 225.3 31.13 41. 27.6 45.7 55.36 249.04 355.350
day, 4 2 6 31 1 6 163 6
3/03/2
PAGE \* MERGEFORMAT 179
020
Wedne
sday, 26.7 0.1 225.8 102 28.2 40.9 205.08 192.979
43 0.9927 25.64 53.55
4/03/2 4 0 4 .67 0 8 735 7
020
Thus-
day, 27.3 0.1 225.7 55. 27.6 44.7 246.80 328.752
44 0.9924 30.85 43.30
5/03/2 3 9 4 70 3 2 327 1
020
Friday,
26.2 0.2 225.3 41. 27.6 45.7 249.04 355.350
45 6/03/2 0.9922 31.13 55.36
4 2 6 31 1 6 163 6
020
Thurs-
day, 29.5 0.4 222.5 67. 30.5 56.3 285.02 280.445
46 0.9909 35.63 71.18
12/03/ 0 1 6 76 4 5 531 2
2020
Friday,
30.2 0.4 222.0 66. 31.3 70.6 510.12
47 13/03/ 0.9893 63.77 72.58 394.035
6 0 3 46 3 4 245
2020
Mon-
day, 26.4 0.3 221.1 32. 27.2 46.2 1210.1 581.557
48 0.9922 151.27 57.00
16/03/ 6 8 4 98 7 1 976 6
2020
Tues-
day, 28.3 0.3 218.7 46. 29.1 62.3 348.72 363.252
49 0.9903 43.59 55.61
17/03/ 6 8 1 22 9 1 163 6
2020
Wedne
sday, 26.4 0.4 223.3 68. 27.0 44.9 235.40 647.672
50 0.9923 29.43 48.49
18/03/ 8 2 8 11 9 6 082 7
2020
Thus-
day, 24.9 0.4 219.2 53. 25.7 37.5 349.33 778.715
51 0.9932 43.67 46.33
19/03/ 5 3 3 38 2 6 551 9
2020
Friday,
26.6 0.4 224.0 46. 27.2 49.4 215.00 586.281
52 20/03/ 0.9918 26.88 50.49
1 3 8 53 3 0 082 7
2020
Mon-
day, 28.2 0.4 223.4 78. 28.9 52.2 225.66 409.236
53 0.9915 28.21 57.66
23/03/ 1 1 3 28 9 0 857 4
2020
Tues- 27.4 0.4 222.7 79. 28.3 33.9 190.08
54 0.9936 23.76 62.23 630.428
day, 1 0 9 08 0 2 98
PAGE \* MERGEFORMAT 179
24/03/
2020
Wedne
sday, 28.5 0.3 225.9 86. 29.2 63.7 326.23 352.720
55 0.9901 40.78 56.99
25/03/ 6 8 2 43 1 0 673 9
2020
Thus-
day, 27.4 0.3 225.6 76. 28.0 56.2 309.29 446.408
56 0.9912 38.66 60.79
26/03/ 5 7 7 53 5 4 469 5
2020
Friday,
29.7 0.4 222.3 65. 30.5 54.0 223.44 233.673
57 27/03/ 0.9913 27.93 77.59
1 1 8 71 3 8 163 3
2020
Mon-
day, 26.1 0.3 221.2 65. 27.1 56.9 223.81 428.572
58 0.9910 27.98 45.41
30/03/ 1 4 1 39 8 0 388 1
2020
Tues-
day, 27.8 0.4 223.4 73. 29.3 52.5 237.32 441.053
59 0.9914 29.67 51.91
31/03/ 9 0 9 30 0 8 735 4
2020
Wedne
sday, 28.1 0.3 226.1 60. 29.4 57.8 230.29 380.463
60 0.9908 28.79 47.23
1/04/2 4 3 8 48 9 5 224 3
020
Thus-
day, 27.1 0.4 226.3 64. 29.0 54.5 317.43 478.461
61 0.9912 39.68 45.26
2/04/2 6 0 8 27 6 4 837 6
020
Friday,
30.1 0.4 225.1 66. 32.4 64.2 367.34 311.875
62 3/04/2 0.9901 45.92 74.14
9 0 3 34 8 1 694 3
020
Mon-
day, 25.8 0.4 225.4 64. 27.9 52.9 289.20 512.220
63 0.9914 36.15 44.42
6/04/2 4 0 7 63 4 4 49 8
020
Tues-
day, 27.3 0.4 225.5 74. 28.5 46.3 387.99 541.833
64 0.9922 48.50 53.21
7/04/2 1 0 9 63 1 4 837 6
020
Wedne
sday, 26.1 0.4 225.6 61. 27.1 47.9 363.24 608.622
65 0.9920 45.41 53.69
8/04/2 0 2 3 85 5 6 898 5
020
PAGE \* MERGEFORMAT 179
Thus-
day, 28.4 0.3 227.1 66. 29.6 39.9 432.04 427.570
66 0.9931 54.01 68.84
9/04/2 4 8 5 22 1 3 082 3
020
Friday,
27.7 0.3 224.7 67. 29.0 37.7 391.76
67 10/04/ 0.9932 48.97 61.26 544.482
3 8 5 52 0 5 653
2020
44.642 1185 9801. 1446.6 242 1215 2107 2747. 11572. 12025.5
Total 5 .63 6.90 73 0 4.22 .28 .59 41 82 0
26.3 0.1 217.8 53. 27.0 46.8
Average 0.9921 5 5 2 32.15 87 1 4 61.05 257.17 267.23

PAGE \* MERGEFORMAT 179

You might also like