Professional Documents
Culture Documents
vi) Sensitive: Pattern based Sensitivity analysed using Regular Expression - 0 : No, 1 :
Yes
Data Preprocessor
STRATIFIED TRAIN-TEST SPLIT:
• It splits the dataset into train and test sets in a way that preserves the same
proportions of examples in each class as observed in the original dataset. This is
called a stratified train-test split.
• The dataset is split into training and testing set of ratio 80:20.
LABEL ENCODER:
• It converts the labels into numeric form so as to convert it into the machine-
readable form. Machine learning algorithms can then decide in a better way on
how those labels must be operated.
Sensitivity Classifier
• RANDOM FOREST
• Classifier that contains a number of decision trees on
various subsets of the given dataset and takes the
average to improve the predictive accuracy.
Hyperparameter Tuning
CROSS VALIDATION:
• CV is used to evaluate machine
learning models on a limited data
sample.
ADVANTAGES:
• AES-GCM is not only efficient and
secure, but hardware implementations
can achieve high speeds with low cost
and low latency, because the mode can
be pipelined.
Password Authenticated Key Exchange (PAKE)
3 phases are performed in this protocol:
1)Initialization Phase:
A large prime number p is chosen and a multiplicative group Zp is found and
two generators g1 and g2 are found out. These are shared between the user and
the cloud server.
2)Registration Phase:
pake Authenticated Key Exchange (PAKE)
Password
3)Authentication Phase:
Result and Analysis
Algorithm Cross Training Testing CLASSIFIER COMPARISON
Validation Accuracy Accuracy 105
Accuracy
100
SVM 89 90.25 85
95
Percentage Accuracy
MNB 85.99 95.76 91.67
90
DT 94.11 100 96.67
RF 91.58 97.67 95.34 85
Classifiers
Accuracy score of various Classifiers Cross Validation Accuracy Training Accuracy Testing Accuracy
Result and Analysis
Algorithm Precision Recall F1 Score COMPARISON OF CLASSIFIERS
1.2
700
3DES 0.24 16.12 ✖ ✖ ✖
600
3 200
+ scrypt 0
10 100 1000 10000