Professional Documents
Culture Documents
366
William Willibrordus Damopolii et al., International Journal of Advanced Trends in Computer Science and Engineering, 10(1), January – February 2021, 366 - 371
2. RELATED WORKS
In 2019 Saa et al [10]. conducted a study to predict student
Yadav et al [4]. conducted a study to compare multiple academic performance using educational data mining. The
decision tree algorithm such as ID3, C4.5, and CART using author used student demographics, course instructor
90 engineering student’s data obtained from VBS Purvanchal information, student general information, and student
University (Uttar Pradesh) on the sampling method for previous performance information from a private university
Institute Technology for session 2010 show that C4.5 in United Emirates Arab. Random Forest algorithm
algorithm can best classify student performance with 78.6% outperformed the other classifiers with 75.52% accuracy
accuracy. followed by Logistic Regression algorithm.
Chaudhari et al [5]. conducted a study to predict student Yao et al [11]. conduct a study to predict secondary school
performance acquired from SSBT College of Engineering and students' final score using their personal data. The dataset
Technology, Jalgaon. Using various student behaviour consists of several variables such as parent information,
variables such as number of library visits, hours spent on student health condition, financial condition, attendance etc.
study, ability to time management etc. Naive bayes algorithm With feature selection, the J48 algorithm showed the best
shows the best performance with 96% accuracy, compared to result with 84.39% accuracy, while without feature selection
C-means algorithm at 95% and K-means algorithm 94% OneR algorithm showed the best performance with 84.19%
accuracy.
Amrieh et al [6]. conducted a study to predict their academic
achievement based on Kalboard 360 learning management Rifat et al [12]. conducted a study to predict students'
system data, implementing a decision tree, artificial neural performance using student transcript data from a renowned
network, and naive bayes algorithm and using ensemble university in Bangladesh. To predict students' final score, the
methods improve the performance of the classifier up to authors used six state-of-the-art classification algorithms.
22.1% and by utilizing student behaviour features increase The results showed that the Random Forest algorithm gave
the classifier accuracy up to 25.8% resulting in 82% accuracy the best performance with 94.1% accuracy, followed by the
in decision tree, 80% accuracy in artificial neural network, Tree Ensemble algorithm.
and 80% accuracy in naive bayes.
3. METHODOLOGY
Hussain et al [7]. conducted a study to identify academically
weak students using data from Digboi College, Duliajan This study is conducted using naive bayes, decision tree, and
College, and Doomdooma College. The study implemented k-nearest neighbor algorithms with RapidMiner software due
deep learning using the sequential neural model with the to its extensive set of classification and optimization
adam optimization method. The study then also compared algorithm [13], the work focuses on comparing different
other classification methods such as the artificial immune algorithm’s performance combined with feature selection and
recognition system v2.0 and adaboost. The highest accuracy parameter optimization to find the best combination based on
achieved was 95.34% produced by deep learning technology. accuracy performance.
Ahmad et al [8]. conducted a study to predict the student’s 3.1 Data Collection
academic performance of a first year bachelor student in
computer science course. The data is collected from The first step in this research is data collection. We need to
eight-year period intakes from July 2006/2007 until July find and gather the right data for the algorithm, data may be
2013/2014 that contain student’s demographic, previous scattered in different spreadsheets, databases, or websites. We
academic record and family background information. Best used SMPN 124 Jakarta class of 2020 and 2019 with a total of
prediction result was achieved by implementing the rule 432 number of initial student data with 30 variables, which
are entrance grades, gender, religion, type of living,
based algorithm with 71.3% accuracy followed by the
transportation method, parent's education, parent's
decision tree with 68.8% and naive bayes with 67%.
occupation, parent's income, and students' grades in 1st and
2nd semester. The grades in each semester include religion
Almarabeh [9] conducted a study to analyze student education, civic, bahasa indonesia, english, mathematics,
performance using classification techniques. The author used natural science, social science, art and culture, and sports
various data such as midterm score, student attendance, subject. After we gather all data, we join them into a single
laboratory experiment, workshop and other factors to predict dataset.
student final score. With Train-Test 80:20, Rule Based
algorithms showed the best performance with 71.3%
accuracy.
367
William Willibrordus Damopolii et al., International Journal of Advanced Trends in Computer Science and Engineering, 10(1), January – February 2021, 366 - 371
368
William Willibrordus Damopolii et al., International Journal of Advanced Trends in Computer Science and Engineering, 10(1), January – February 2021, 366 - 371
Classification aims to identify characteristics that indicate the classification accuracy, and learning runtime required. There
group to which each case belongs. This pattern can be used to are many feature selection algorithms, one of them is forward
understand existing data and predict behavior of new data. selection [14].
Data mining creates classification models by examining
already classified data (cases) and finding predictive pattern Forward selection is a method of adding variables to the
inductively. These existing cases may come from historical model one at a time [15]. It starts with an empty selection of
databases [17]. variables, and then adds every unused variable of the given
data in each round. For each added variable, then calculate
The data classified using following classification algorithms: the performance. Only the variable with the highest
performance improvement are added to the selection. Then
A. Naïve Bayes
start a new round with the modified selection [16].
Naive Bayes is one of the simplest and the most commonly
used classifiers [18]. It assumes the conditional independence
3.6 Parameter Optimization
of a class, that is, given the class label of a tuple, it assumes
that the values of the variables are conditionally independent Parameter Optimization is one technique to increase the
of the others [1]. accuracy of the data mining algorithm by tuning its
parameters based on testing dataset performance. It runs the
It is based on the application of Bayes' theorem to handle
algorithm with all the predefined parameter configuration
simple probabilistic classification. It assumes that the
possibilities, finding the best possible parameter
existence of a particular characteristic of a class is unrelated
configuration based on the algorithm’s model accuracy.
to the existence of any other characteristic, Even if these
characteristics depend on the existence of another
The parameters that will be tuned in the decision tree are the
characteristic, the naive bayes classifier will treat all these
criterion on which attribute will be selected for splitting,
variables as independently contributing to the possibility of
tree’s maximal depth, pruning and minimal gain value.
classifying them into a specific class [19].
K-NN parameters that will be tuned are k value, measure type
and weighted vote. The parameters that will be tuned in naive
B. Decision Tree bayes are the laplace correction techniques.
4. RESULT AND ANALYSIS with feature selection and parameter optimization. Among
Table 3 summarizes the algorithm performance for naive these algorithms, the best accuracy is achieved by K-Nearest
bayes, decision tree, and k-nearest neighbor (K-NN), with 4 Neighbor with feature selection and parameter optimization
results of accuracy for each algorithm, without any that show the same accuracy result with Decision Tree with
optimization, using only features selection (FS), using only feature selection at 77.36% accuracy. There are limitations on
parameter optimization (PO), and with both optimization this study such as lack of data varieties, and the number of
(FS+PO). data processed due the original dataset containing a lot of
Table 3: Accuracy Result missing values.
Without FS +
Algorithm FS PO REFERENCES
Optimization PO
1. J. Han, J. Pei and M. Kamber, DATA MINING:
Naive 75.47 73.47
69.81% 76.92% Concepts and Techniques, Morgan Kaufmann, 2011.
Bayes % %
2. C. Romero and S. Ventura, Educational Data
Decision 77.36 64.15
62.26% 73.58% Mining: A Review of the State of the Art, IEEE
Tree % %
Transactions on Systems, Man, and Cybernetics, Part C
69.81 69.81
K-NN 67.92% 77.36% (Applications and Reviews), 40(6), pp. 601-618, 2010.
% %
3. M. Goyal and R. Vohra, Applications of Data
Mining in Higher Education, International Journal of
Naive bayes’ best performance at 76.92% is achieved by
Computer Science Issues (IJCSI) 9, no. 2, pp. 113, 2012.
applying parameter optimization by setting the laplace 4. S. K. Yadav and S. Pal, Data Mining: A Prediction
correction parameter to false and implementing feature for Performance Improvement of Engineering
selection which results in only using Bahasa Indonesia 2nd Students using Classification, World of Computer
semester grade, natural science 2nd semester grade, english Science and Information Technology Journal, vol. 2, no.
2nd semester grade, gender and type of living variables. 2. pp. 51-56, 2012.
5. K. P. Chaudhari, R. A. Sharma, S. S. Jha and R. J.
Decision tree best performance at 77.36% is achieved by Bari, Student Performance Prediction System using
implementing parameter optimization, tuning the criterion Data Mining Approach, Int J Adv Res Comput
based on gain ratio, set the maximal depth to 10, applying Commun Eng 6, no. 3, pp. 833-839, 2017.
pruning and with 0.01 minimal gain. 6. E. A. Amrieh, T. Hamtini and I. Aljarah, Mining
Educational Data to Predict Student’s Academic
Performance using Ensemble Methods, International
K-NN best performance at 77.36% is achieved by using Journal of Database Theory and Application 9, no. 8, pp.
parameter optimization by tuning the k value to 100, set the 119-136, 2016.
KNN to nominal measure and weighted vote to false and 7. S. Hussain, Z. F. Muhsion, Y. K. Salal, P. Theodoru,
applying feature selection that results in only using Bahasa F. Kurtoğlu and G. C. Hazarika, Prediction Model on
Indonesia 1st semester grade, english 2nd semester grade, Student Performance based on Internal Assessment
gender and mother’s education variable. using Deep Learning, International Journal of
Emerging Technologies in Learning (iJET) 14, no. 08,
In general, the experiment shows that feature selection and pp. 4-22, 2019.
parameter optimization improve the accuracy of the classifier 8. F. Ahmad, N. H. Ismail and A. A. Aziz, The
Prediction of Students’ Academic Performance using
algorithm. However, when combined with optimization, it
Classification Data Mining Techniques, Applied
does not always result in better accuracy, such as that in the
Mathematical Sciences 9, no. 129, pp. 6415-6426, 2015.
Decision Tree experimental result. Feature selection works
9. H. Almarabeh, Analysis of Students' Performance
better compared to parameter optimization in improving the by using Different Data Mining Classifiers,
accuracy due to the large amount of data variables. It also International Journal of Modern Education and
shows that various algorithms show different accuracy Computer Science 9, no. 8, pp. 9, 2017.
results, K-Nearest Neighbor with feature selection and 10. A. A. Saa, M. Al-Emran and K. Shaalan, Mining
parameter optimization shows the same result as decision tree Student Information System Records to Predict
with features selection display best accuracy value of 77.36%. Students’ Academic Performance, In International
conference on advanced machine learning technologies
5. CONCLUSION and applications, pp. 229-239, 2019.
11. Y. Yao, Z. Chen, S. Byun and Y. Liu, Using Data
This study is designed to compare various classification
Mining Classifiers to Predict Academic Performance
algorithms and optimization to predict student’s grade, we of High School Students, Scientific Cyber Security
applied Naive Bayes, Decision Tree, and K-Nearest Neighbor Association (SCSA), pp. 18-35, 2019.
370
William Willibrordus Damopolii et al., International Journal of Advanced Trends in Computer Science and Engineering, 10(1), January – February 2021, 366 - 371
371