Professional Documents
Culture Documents
• Activation function: After each convolution operation, where Wij are the weights between the visible vi and
an activation function (such as ReLU - Rectified Linear hidden hj units, bj and cj are the biases of the visible
Unit) is applied element by element to the resulting and hidden units respectively.
activation map. This introduces non-linearity into the • Top-down Propagation (Fine-Tuning): Once the BRMs
model, enabling it to learn complex patterns. have been pre-trained layer by layer, the network is
usually fine-tuned using a technique such as gradient
ReLU (x) = max(0, x) (7) back-propagation to adjust the weights and biases of the
• Pooling layer: After certain convolution layers, pooling entire network, thus optimizing model performance for
operations (such as max-pooling) are performed to reduce the specific task.
the spatial dimension of the activation map and the DBN is a deep neural network composed of visible and hidden
number of parameters in the model. Pooling retains the layers, where each pair of adjacent layers is modeled as an
essential features. For a given region of the activation BRM. These BRMs are pre-trained in an unsupervised way,
map, max-pooling retains the maximum value: then the overall network is fine-tuned for specific tasks using
supervised learning techniques.
M ax − P ooling(X) = max(region) (8)
Stacked Autoencoder
• Fully Connected (Dense) layers: One or more fully Sparse Autoencoders (SAEs) are a variant of the classic
connected layers can follow the convolution and pooling autoencoders used for unsupervised learning and feature
layers. These layers perform linear operations on the extraction in data. The special feature of SAEs is that they
extracted features to perform final classification or other impose a parsimony constraint on the internal representation
tasks. Let X be the output of the previous layer, W of the data, meaning that only a few units/neurons are activated
the weight matrix and b the associated bias. The fully at any one time. This can lead to a better representation
connected operation is given by: of the data, highlighting important features and reducing the
dimensionality of the data. An SAE typically consists of:
Y = XW + b (9)
• Encoder: The encoder transforms the input data into
• Output layer: The last layer of the CNN produces the an internal representation (encoding) using a layer of
output results, usually using an appropriate activation neurons called the encoding layer. This layer is generally
function (such as softmax for classification) to obtain narrower than the input layer, resulting in information
probabilities or class scores. compression.
• Decoder: The decoder reconstructs the input data from will be processed to produce the formatted Word, PDF, and
the internal representation obtained from the encoder. The HTML5 output formats, which will be provided to you for
aim of the decoder is to minimize the reconstruction error review, revision/resubmission (if applicable), and approval.
between the input data and the reconstructed data.
III. C ONCLUSION
2) Hyper parameter optimization: As presented previously,
LSTM, CNN, DBN and SAE networks are made up of several R EFERENCES
layers whose configurations can influence their learning [1] N. Funabiki, K.S. Lwin, Y. Aoyagi, M. Kuribayashi, and W.-C. Kao,
“A user-pc computing system as ultralow-cost computation platform for
performance. Thus, for a given forecasting task, the optimal small groups,” Appl. Theory. Comput. Tech., vol.2, no.3, pp.10-24, 2017.
parameterization to achieve good forecasting performance [2] H. Htet, N. Funabiki, A. Kamoyedji and M. Kuribayashi, “Design
with reduced computation time must be sought. In general, and implementation of improved user-pc computing system,” IEICE
Technical Report, vol.120, no.69, pp.37-42, 2020.
building an efficient machine learning model is a complex [3] N. Funabiki, Y. Asai, Y. Aoyagi and W. -C. Kao, “A Worker
and tedious process, involving determining the appropriate Implementation on Windows OS for User-PC Computing System,”
algorithm and obtaining an optimal model architecture by 2015 Third International Symposium on Computing and Networking
(CANDAR), Sapporo, Japan, pp. 279-282 2015.
tuning its hyperparameters (HP). To build an optimal model, [4] A. Kamoyedji, N. Funabiki, H. Htet and M. Kuribayashi, “A proposal
we need to explore a range of possibilities. In Table 1 below, of static job scheduling algorithm considering cpu core utilization
models hyperparameters and variation ranges are presented. for user-pc computing system,” 2021 9th International Conference on
Information and Education Technology (ICIET), Okayama, Japan, pp.
Table 1.Values of different integrals φi,j,k for Cosh pulses 374-379, 2021.
from ∀ n ∈ N [5] N. Funabiki, Y. Aoyagi, M. Kuribayashi, W.-C. Kao, “Worker pc
performance measurements using benchmarks for user-pc computing
system,” Consumer Electronics-Taiwan (ICCE-TW) 2016 IEEE
International Conference on Consumer Electronics-Taiwan (ICCE-TW),
• We vary input data size between 2h and 24h; Taiwan, pp.1-2, 2016.
• We vary the forecast horizon size between 1h and 23h in [6] T. Xie, A. Sung, and X. Qin, “Dynamic task scheduling with security
awareness in real-time systems,” in Proc. IEEE International Parallel and
steps of 1 hour; Distributed Processing Symposium, pp. 8-15, 2005.
• The number of units in the LSTM, CNN and SAE models
is varied between 10 and 1000 in steps of 10;
• The number of filters in the CNN model ranges from 10
to 300 in steps of 10;
• The number of hidden layers in the DBN model ranges
from 1 to 6 in steps of 1;
• The size of the hidden layers in the DBN model is varied
between 32 and 260 in steps of 32;
−2
• Three learning rates are tested: 10 , 10−3 , 10−4
Hyper-parameter optimization is performed using the Keras-
tuner library in Python3 [32].
E. Models training
Once the optimal parameters of the models have been
determined, they are then trained on the dataset. The models
are developed and trained using Tensorflow under python3.
TensorFlow is a leading open source platform for machine
learning. It features a comprehensive and flexible ecosystem
of tools, libraries and community resources that enable
researchers to advance machine learning and developers to
easily create and deploy machine learning-based applications
[33].
F. Model performance evaluation
More about the submission template
This submission version of your paper should not have headers
or footers, these will be added when your manuscript is
processed after acceptance. It should remain in a one-column
format—please do not alter any of the styles or margins.
If a paper is accepted for publication, authors will be
instructed on the next steps. Authors must then follow the
submission instructions found on their respective publication’s
web page. Once your submission is received, your paper