You are on page 1of 81

Bidarahalli, Virgonagar Post, Bangalore-560 049

Department of Computer Science and Engineering &


Information Science and Engineering

ACADEMIC YEAR: 2018 - 19

VII Semester

MACHINE LEARNING Laboratory Manual

(Subject Code: 15CSL76)


Faculty Incharge: Mrs. Manimegalai
Mr. Prabhuraj Metipatil
Mrs. Ashwini V
Dr. Anitha N
Mrs. Shefali

Name of the Student :_________________________________

University Seat No : _________________________________

Batch : _________________________________
MACHINE LEARNING LABORATORY- 15CSL76

MACHINE LEARNING LABORATORY


[As per Choice Based Credit System (CBCS) scheme]

SEMESTER – VII Subject Code 15CSL76

Course objectives:

This course will enable students to

1. Make use of Data sets in implementing the machine learning algorithms


2. Implement the machine learning concepts and algorithms in Python Programming language.

Course outcomes:

The students should be able to

1. Understand the implementation procedures for the machine learning algorithms


2. Design Python programs for various Learning algorithms
3. Apply appropriate data sets to the Machine Learning algorithms
4. Identify and apply Machine Learning algorithms to solve real world problems

Description

1. The programs can be implemented in Python


2. Data sets can be taken from standard repositories (https://archive.ics.uci.edu/ml/datasets.html) or
can be constructed by the students

Conduction of Practical Examination

• All laboratory experiments are to be included for practical examination.


• Students are allowed to pick one experiment from the lot.
• Strictly follow the instructions as printed on the cover page of answer script
• Marks distribution: Procedure + Conduction + Viva:20 + 50 +10 (80)

Dept of CSE & ISE, EPCET B’LURU Page 1


MACHINE LEARNING LABORATORY- 15CSL76

Machine learning
Machine learning is a subset of artificial intelligence in the field of computer science that often uses
statistical techniques to give computers the ability to "learn" (i.e., progressively improve performance on
a specific task) with data, without being explicitly programmed. In the past decade, machine learning has
given us self-driving cars, practical speech recognition, effective web search, and a vastly improved
understanding of the human genome.

Machine learning tasks


Machine learning tasks are typically classified into two broad categories, depending on whether there is a
learning "signal" or "feedback" available to a learning system:

Supervised learning: The computer is presented with example inputs and their desired outputs, given by
a "teacher", and the goal is to learn a general rule that maps inputs to outputs. As special cases, the input
signal can be only partially available, or restricted to special feedback:

Semi-supervised learning: The computer is given only an incomplete training signal: a training set with
some (often many) of the target outputs missing.

Active learning: The computer can only obtain training labels for a limited set of instances (based on a
budget), and also has to optimize its choice of objects to acquire labels for. When used interactively, these
can be presented to the user for labeling.

Reinforcement learning: training data (in form of rewards and punishments) is given only as feedback to
the program's actions in a dynamic environment, such as driving a vehicle or playing a game against an
opponent.

Unsupervised learning: No labels are given to the learning algorithm, leaving it on its own to find
structure in its input. Unsupervised learning can be a goal in itself (discovering hidden patterns in data) or
a means towards an end (feature learning).

Supervised Learning Unsupervised Learning Instance Based Learning


Find-s algorithm EM algorithm
Candidate elimination algorithm
Decision tree algorithm
Back propagation Algorithm Locally Weighted Regression
Naïve Bayes Algorithm K Means Algorithm algorithm
K nearest neighbour
algorithm(lazy learning
algorithm)

Machine learning applications

General Applications

In classification, inputs are divided into two or more classes, and the learner must produce a model that
assigns unseen inputs to one or more (multi-label classification) of these classes. This is typically tackled

Dept of CSE & ISE, EPCET B’LURU Page 2


MACHINE LEARNING LABORATORY- 15CSL76

in a supervised manner. Spam filtering is an example of classification, where the inputs are email (or
other) messages and the classes are "spam" and "not spam".

In regression, also a supervised problem, the outputs are continuous rather than discrete.

In clustering, a set of inputs is to be divided into groups. Unlike in classification the groups are not
known beforehand, making this typically an unsupervised task.

Density estimation finds the distribution of inputs in some space.

Dimensionality reduction simplifies inputs by mapping them into a lower dimensional space. Topic
modeling is a related problem, where a program is given a list of human language documents and is
tasked with finding out which documents cover similar topics.

Some Specific Applications

• Healthcare
▪ Drug Discovery/ Manufacturing
▪ Personalized Treatment or Medication
• Finance
▪ Fraud Detection
▪ Focused Account Holder Targeting
• Retail
▪ Product recommendations using Recommender Systems
▪ Improved customer service
• Travel
▪ Dynamic Pricing
▪ Travel Sentiment analysis
• Media
▪ Sentiment Analysis
▪ Serving content to targeted audience
• Automobile Industry
• Robotics

Machine learning Approaches

Decision tree learning


Decision tree learning uses a decision tree as a predictive model, which maps observations about an item
to conclusions about the item's target value.

Association rule learning


Association rule learning is a method for discovering interesting relations between variables in large
databases.

Artificial neural networks


An artificial neural network (ANN) learning algorithm, usually called "neural network" (NN), is a
learning algorithm that is vaguely inspired by biological neural networks. Computations are structured in
terms of an interconnected group of artificial neurons, processing information using a connectionist

Dept of CSE & ISE, EPCET B’LURU Page 3


MACHINE LEARNING LABORATORY- 15CSL76

approach to computation. Modern neural networks are non-linear statistical data modeling tools. They are
usually used to model complex relationships between inputs and outputs, to find patterns in data, or to
capture the statistical structure in an unknown joint probability distribution between observed variables.

Deep learning
Falling hardware prices and the development of GPUs for personal use in the last few years have
contributed to the development of the concept of deep learning which consists of multiple hidden layers in
an artificial neural network. This approach tries to model the way the human brain processes light and
sound into vision and hearing. Some successful applications of deep learning are computer vision and
speech recognition.

Inductive logic programming


Inductive logic programming (ILP) is an approach to rule learning using logic programming as a uniform
representation for input examples, background knowledge, and hypotheses. Given an encoding of the
known background knowledge and a set of examples represented as a logical database of facts, an ILP
system will derive a hypothesized logic program that entails all positive and no negative examples.
Inductive programming is a related field that considers any kind of programming languages for
representing hypotheses (and not only logic programming), such as functional programs.

Support vector machines


Support vector machines (SVMs) are a set of related supervised learning methods used for classification
and regression. Given a set of training examples, each marked as belonging to one of two categories, an
SVM training algorithm builds a model that predicts whether a new example falls into one category or the
other.

Clustering
Cluster analysis is the assignment of a set of observations into subsets (called clusters) so that
observations within the same cluster are similar according to some pre designated criterion or criteria,
while observations drawn from different clusters are dissimilar. Different clustering techniques make
different assumptions on the structure of the data, often defined by some similarity metric and evaluated
for example by internal compactness (similarity between members of the same cluster) and separate on
between different clusters. Other methods are based on estimated density and graph connectivity.
Clustering is a method of unsupervised learning, and a common technique for statistical data analysis.

Bayesian networks
A Bayesian network, belief network or directed acyclic graphical model is a probabilistic graphical model
that represents a set of random variables and their conditional independencies via a directed acyclic graph
(DAG). For example, a Bayesian network could represent the probabilistic relationships between diseases
and symptoms. Given symptoms, the network can be used to compute the probabilities of the presence of
various diseases. Efficient algorithms exist that perform inference and learning.

Reinforcement learning
Reinforcement learning is concerned with how an agent ought to take actions in an environment so as to
maximize some notion of long-term reward. Reinforcement learning algorithms attempt to find a policy
that maps states of the world to the actions the agent ought to take in those states. Reinforcement learning
differs from the supervised learning problem in that correct input/output pairs are never presented, nor
sub-optimal actions explicitly corrected.

Dept of CSE & ISE, EPCET B’LURU Page 4


MACHINE LEARNING LABORATORY- 15CSL76

Similarity and metric learning


In this problem, the learning machine is given pairs of examples that are considered similar and pairs of
less similar objects. It then needs to learn a similarity function (or a distance metric function) that can
predict if new objects are similar. It is sometimes used in Recommendation systems.

Genetic algorithms
A genetic algorithm (GA) is a search heuristic that mimics the process of natural selection, and uses
methods such as mutation and crossover to generate new genotype in the hope of finding good solutions
to a given problem. In machine learning, genetic algorithms found some uses in the 1980s and 1990s.
Conversely, machine learning techniques have been used to improve the performance of genetic
and evolutionary algorithms.

Rule-based machine learning


Rule-based machine learning is a general term for any machine learning method that identifies, learns, or
evolves "rules" to store, manipulate or apply, knowledge. The defining characteristic of a rule-based
machine learner is the identification and utilization of a set of relational rules that collectively represent
the knowledge captured by the system. This is in contrast to other machine learners that commonly
identify a singular model that can be universally applied to any instance in order to make a prediction
Rule-based machine learning approaches include learning classifier systems, association rule learning,
and artificial immune systems.

Feature selection approach


Feature selection is the process of selecting an optimal subset of relevant features for use in model
construction. It is assumed the data contains some features that are either redundant or irrelevant, and can
thus be removed to reduce calculation cost without incurring much loss of information. Common
optimality criteria include accuracy, similarity and information measures.

Dept of CSE & ISE, EPCET B’LURU Page 5


MACHINE LEARNING LABORATORY- 15CSL76

LAB EXPERIMENTS

1. Implement and demonstrate the FIND-S algorithm for finding the most specific hypothesis based
on a given set of training data samples. Read the training data from a .CSV file

2. For a given set of training data examples stored in a .CSV file, implement and demonstrate the
Candidate-Elimination algorithm to output a description of the set of all hypotheses consistent
with the training examples

3. Write a program to demonstrate the working of the decision tree based ID3 algorithm. Use an
appropriate data set for building the decision tree and apply this knowledge to classify a new
sample

4. Build an Artificial Neural Network by implementing the Back propagation algorithm and test the
same using appropriate data sets

5. Write a program to implement the naïve Bayesian classifier for a sample training data set stored as
a .CSV file. Compute the accuracy of the classifier, considering few test data sets.

6. Assuming a set of documents that need to be classified, use the naïve Bayesian Classifier model to
perform this task. Built-in Java classes/API can be used to write the program. Calculate the
accuracy, precision, and recall for your data set

7. Write a program to construct a Bayesian network considering medical data. Use this model to
demonstrate the diagnosis of heart patients using standard Heart Disease Data Set. You can use
Java/Python ML library classes/API

8. Apply EM algorithm to cluster a set of data stored in a .CSV file. Use the same data set for
clustering using k-Means algorithm. Compare the results of these two algorithms and comment on
the quality of clustering. You can add Java/Python ML library classes/API in the program

9. Write a program to implement k-Nearest Neighbour algorithm to classify the iris data set. Print
both correct and wrong predictions. Java/Python ML library classes can be used for this problem.

10. Implement the non-parametric Locally Weighted Regression algorithm in order to fit data points.
Select appropriate data set for your experiment and draw graphs

Dept of CSE & ISE, EPCET B’LURU Page 6


MACHINE LEARNING LABORATORY- 15CSL76

Program 1

Implement and demonstrate the FIND-S algorithm for finding the most specific hypothesis
based on a given set of training data samples. Read the training data from a .CSV file

Task: Find-S algorithm is used to find a maximally specific hypothesis.

Positive and negative training examples for the target concept EnjoySport

Example Sky AirTemp Humidity Wind Water Forecast EnjoySport


1 Sunny Warm Normal Strong Warm Same Yes
2 Sunny Warm High Strong Warm Same Yes
3 Rainy Cold High Strong Warm Change No
4 Sunny Warm High Strong Cool Change Yes

Given:

✓ Instances X: Possible days, each described by the attributes


• Sky (with possible values Sunny, Cloudy, and Rainy),
• AirTemp (with values Warm and Cold),
• Humidity (with values Normal and High),
• Wind (with values Strong and Weak),
• Water (with values Warm and Cool), and
• Forecast (with values Same and Change).

✓ Hypotheses H: Each hypothesis is described by a conjunction of constraints on the attributes


Sky, AirTemp, Humidity, Wind, Water, and Forecast. The constraints may be "?" (any value is
acceptable), "0 (no value is acceptable), or a specific value

✓ Target concept c: EnjoySport : X →(0,l)

✓ Training examples D: Positive and negative examples of the target function

The steps of FINDS are given below:

Step1: The first step of FIND-S is to initialize h to the most specific hypothesis in H

h0  { ɸ, ɸ, ɸ, ɸ, ɸ, ɸ }

Step 2: Positive training example x1= <Sunny, Warm, Normal, Strong, Warm, Same>
h1  <Sunny, Warm, Normal, Strong, Warm, Same>

Dept of CSE & ISE, EPCET B’LURU Page 7


MACHINE LEARNING LABORATORY- 15CSL76

Step 3: Positive training instance x2= <Sunny, Warm, High, Strong, Warm, Same>
h2  <Sunny, Warm,?, Strong, Warm, Same>

Step 4: Negative training example x3= <Rainy, Cold, High, Strong, Warm, Change>
h3  <Sunny, Warm,? , Strong, Warm, Same>

Step 5: Positive training example x4= <Sunny,Warm, High, Strong, Cool, change>
h4  <Sunny, Warm,? , Strong, ?, ?>

Dept of CSE & ISE, EPCET B’LURU Page 8


MACHINE LEARNING LABORATORY- 15CSL76

Find-s Algorithm:

Step 1: Initialize h to the most specific hypothesis in H.


Step 2: For each positive training instance x.
For each attribute constraint ai in h
If the constraint ai is satisfied by x
Then do nothing
Else replace ai in h by the next more general constraint that is satisfied
by x.
Step 3: Output hypothesis h.

Find-s in Python:

import random
import csv
attributes = [['Sunny','Rainy'],
['Warm','Cold'],
['Normal','High'],
['Strong','Weak'],
['Warm','Cool'],
['Same','Change']]
num_attributes = len(attributes)
print (" \n The most general hypothesis : ['?','?','?','?','?','?']\n")
print ("\n The most specific hypothesis : ['0','0','0','0','0','0']\n")
a = []
print("\n The Given Training Data Set \n")

with open('D:/prabhu/ws.csv','r') as csvFile:


reader = csv.reader(csvFile)
for row in reader:
a.append (row)
print(row)
print("\n The initial value of hypothesis: ")
hypothesis = ['0'] * num_attributes
print(hypothesis)

# Comparing with First Training Example


for j in range(0,num_attributes):
hypothesis[j] = a[0][j];

# Comparing with Remaining Training Examples of Given Data Set


print("\n Find S: Finding a Maximally Specific Hypothesis\n")
for i in range(0,len(a)):
if a[i][num_attributes]=='Yes':
for j in range(0,num_attributes):
if a[i][j]!=hypothesis[j]:
hypothesis[j]='?'
else :
hypothesis[j]= a[i][j]

Dept of CSE & ISE, EPCET B’LURU Page 9


MACHINE LEARNING LABORATORY- 15CSL76

print(" For Training Example No :{0} the hypothesis is


".format(i),hypothesis)

print("\n The Maximally Specific Hypothesis for a given Training Examples


:\n")
print(hypothesis)

OUTPUT

Dept of CSE & ISE, EPCET B’LURU Page 10


MACHINE LEARNING LABORATORY- 15CSL76

Dept of CSE & ISE, EPCET B’LURU Page 11


MACHINE LEARNING LABORATORY- 15CSL76

Program 2

For a given set of training data examples stored in a .CSV file, implement and demonstrate the
Candidate-Elimination algorithm to output a description of the set of all hypotheses consistent
with the training examples.

Task: The CANDIDATE-ELIMINATION algorithm computes the version space containing all
hypotheses from H that are consistent with an observed sequence of training examples.
Dataset: EnjoySport training examples

Dept of CSE & ISE, EPCET B’LURU Page 12


MACHINE LEARNING LABORATORY- 15CSL76

Dept of CSE & ISE, EPCET B’LURU Page 13


MACHINE LEARNING LABORATORY- 15CSL76

Dept of CSE & ISE, EPCET B’LURU Page 14


MACHINE LEARNING LABORATORY- 15CSL76

Candidate-Elimination Algorithm:

Dept of CSE & ISE, EPCET B’LURU Page 15


MACHINE LEARNING LABORATORY- 15CSL76

class Holder:
factors={} #Initialize an empty dictionary
attributes = () #declaration of dictionaries parameters with an
arbitrary length
'''
Constructor of class Holder holding two parameters,
self refers to the instance of the class
'''
def __init__(self,attr): #
self.attributes = attr
for i in attr:
self.factors[i]=[]

def add_values(self,factor,values):
self.factors[factor]=values

class CandidateElimination:
Positive={} #Initialize positive empty dictionary
Negative={} #Initialize negative empty dictionary

def __init__(self,data,fact):
self.num_factors = len(data[0][0])
self.factors = fact.factors
self.attr = fact.attributes
self.dataset = data

#print self.attr

def run_algorithm(self):
# print self.dataset
'''
Initialize the specific and general boundaries, and loop the
dataset against the algorithm
'''
G = self.initializeG()
S = self.initializeS()

'''
Programmatically populate list in the iterating variable
trial_set
'''
count=0
for trial_set in self.dataset:
if self.is_positive(trial_set): #if trial set/example
consists of positive examples

Dept of CSE & ISE, EPCET B’LURU Page 16


MACHINE LEARNING LABORATORY- 15CSL76

G = self.remove_inconsistent_G(G,trial_set[0]) #remove
inconsitent data from the general boundary
S_new = S[:] #initialize the dictionary with no key-
value pair
print(S_new)
for s in S:
if not self.consistent(s,trial_set[0]):
S_new.remove(s)
generalization =
self.generalize_inconsistent_S(s,trial_set[0])
generalization =
self.get_general(generalization,G)
if generalization:
S_new.append(generalization)
S = S_new[:]
S = self.remove_more_general(S)

print (S)
else:#if it is negative
S = self.remove_inconsistent_S(S,trial_set[0]) #remove
inconsitent data from the specific boundary
G_new = G[:] #initialize the dictionary with no key-
value pair (dataset can take any value)
print (G_new)
for g in G:
if self.consistent(g,trial_set[0]):
G_new.remove(g)
specializations =
self.specialize_inconsistent_G(g,trial_set[0])
specializationss =
self.get_specific(specializations,S)
if specializations != []:
G_new += specializations
G = G_new[:]
print(G)
G = self.remove_more_specific(G)

print (S)
print (G)

def initializeS(self):
''' Initialize the specific boundary '''
S = tuple(['-' for factor in range(self.num_factors)]) #6
constraints in the vector
return [S]

Dept of CSE & ISE, EPCET B’LURU Page 17


MACHINE LEARNING LABORATORY- 15CSL76

def initializeG(self):
''' Initialize the general boundary '''
G = tuple(['?' for factor in range(self.num_factors)]) # 6
constraints in the vector
return [G]

def is_positive(self,trial_set):
''' Check if a given training trial_set is positive '''
if trial_set[1] == 'Y':
return True
elif trial_set[1] == 'N':
return False
else:
raise TypeError("invalid target value")

def is_negative(self,trial_set):
''' Check if a given training trial_set is negative '''
if trial_set[1] == 'N':
return False
elif trial_set[1] == 'Y':
return True
else:
raise TypeError("invalid target value")

def match_factor(self,value1,value2):
''' Check for the factors values match,
necessary while checking the consistency of
training trial_set with the hypothesis '''
if value1 == '?' or value2 == '?':
return True
elif value1 == value2 :
return True
return False

def consistent(self,hypothesis,instance):
''' Check whether the instance is part of the hypothesis '''
for i,factor in enumerate(hypothesis):
if not self.match_factor(factor,instance[i]):
return False
return True

def remove_inconsistent_G(self,hypotheses,instance):
''' For a positive trial_set, the hypotheses in G
inconsistent with it should be removed '''

Dept of CSE & ISE, EPCET B’LURU Page 18


MACHINE LEARNING LABORATORY- 15CSL76

G_new = hypotheses[:]
for g in hypotheses:
if not self.consistent(g,instance):
G_new.remove(g)
return G_new

def remove_inconsistent_S(self,hypotheses,instance):
''' For a negative trial_set, the hypotheses in S
inconsistent with it should be removed '''
S_new = hypotheses[:]
for s in hypotheses:
if self.consistent(s,instance):
S_new.remove(s)
return S_new

def remove_more_general(self,hypotheses):
''' After generalizing S for a positive trial_set, the
hypothesis in S
general than others in S should be removed '''
S_new = hypotheses[:]
for old in hypotheses:
for new in S_new:
if old!=new and self.more_general(new,old):
S_new.remove[new]
return S_new

def remove_more_specific(self,hypotheses):
''' After specializing G for a negative trial_set, the
hypothesis in G
specific than others in G should be removed '''
G_new = hypotheses[:]
for old in hypotheses:
for new in G_new:
if old!=new and self.more_specific(new,old):
G_new.remove[new]
return G_new

def generalize_inconsistent_S(self,hypothesis,instance):
''' When a inconsistent hypothesis for positive trial_set is
seen in the specific boundary S,
it should be generalized to be consistent with the
trial_set ... we will get one hypothesis'''
hypo = list(hypothesis) # convert tuple to list for mutability
for i,factor in enumerate(hypo):
if factor == '-':

Dept of CSE & ISE, EPCET B’LURU Page 19


MACHINE LEARNING LABORATORY- 15CSL76

hypo[i] = instance[i]
elif not self.match_factor(factor,instance[i]):
hypo[i] = '?'
generalization = tuple(hypo) # convert list back to tuple for
immutability
return generalization

def specialize_inconsistent_G(self,hypothesis,instance):
''' When a inconsistent hypothesis for negative trial_set is
seen in the general boundary G
should be specialized to be consistent with the
trial_set.. we will get a set of hypotheses '''
specializations = []
hypo = list(hypothesis) # convert tuple to list for mutability
for i,factor in enumerate(hypo):
if factor == '?':
values = self.factors[self.attr[i]]
for j in values:
if instance[i] != j:
hyp=hypo[:]
hyp[i]=j
hyp=tuple(hyp) # convert list back to tuple
for immutability
specializations.append(hyp)
return specializations

def get_general(self,generalization,G):
''' Checks if there is more general hypothesis in G
for a generalization of inconsistent hypothesis in S
in case of positive trial_set and returns valid
generalization '''

for g in G:
if self.more_general(g,generalization):
return generalization
return None

def get_specific(self,specializations,S):
''' Checks if there is more specific hypothesis in S
for each of hypothesis in specializations of an
inconsistent hypothesis in G in case of negative trial_set
and return the valid specializations'''
valid_specializations = []
for hypo in specializations:

Dept of CSE & ISE, EPCET B’LURU Page 20


MACHINE LEARNING LABORATORY- 15CSL76

for s in S:
if self.more_specific(s,hypo) or
s==self.initializeS()[0]:
valid_specializations.append(hypo)
return valid_specializations

def exists_general(self,hypothesis,G):
'''Used to check if there exists a more general hypothesis in
general boundary for version space'''

for g in G:
if self.more_general(g,hypothesis):
return True
return False

def exists_specific(self,hypothesis,S):
'''Used to check if there exists a more specific hypothesis in
general boundary for version space'''

for s in S:
if self.more_specific(s,hypothesis):
return True
return False

def get_version_space(self,specific,general):
''' Given the specific and the general boundary of the
version space, evaluate the version space in between '''
while get_order(VS):
for hypothesis in VS[:]:
hypo = list(hypothesis) # convert tuple to list for
mutability
for i,factor in enumerate(hypo):
if factor != '?':
hyp=hypo[:]
hyp[i]='?'
if self.exists_general(hyp,general)and
self.exists_specific(hyp,specific):
VS.append(tuple(hyp))

return VS
def get_order(self,hypothesis):
pass

def more_general(self,hyp1,hyp2):
''' Check whether hyp1 is more general than hyp2 '''

Dept of CSE & ISE, EPCET B’LURU Page 21


MACHINE LEARNING LABORATORY- 15CSL76

hyp = zip(hyp1,hyp2)
for i,j in hyp:
if i == '?':
continue
elif j == '?':
if i != '?':
return False
elif i != j:
return False
else:
continue
return True

def more_specific(self,hyp1,hyp2):
''' hyp1 more specific than hyp2 is
equivalent to hyp2 being more general than hyp1 '''
return self.more_general(hyp2,hyp1)

dataset=[(('sunny','warm','normal','strong','warm','same'),'Y'),(('sun
ny','warm','high','strong','warm','same'),'Y'),(('rainy','cold','high'
,'strong','warm','change'),'N'),(('sunny','warm','high','strong','cool
','change'),'Y')]
attributes =('Sky','Temp','Humidity','Wind','Water','Forecast')
f = Holder(attributes)
f.add_values('Sky',('sunny','rainy','cloudy')) #sky can be sunny rainy
or cloudy
f.add_values('Temp',('cold','warm')) #Temp can be sunny cold or warm
f.add_values('Humidity',('normal','high')) #Humidity can be normal or
high
f.add_values('Wind',('weak','strong')) #wind can be weak or strong
f.add_values('Water',('warm','cold')) #water can be warm or cold
f.add_values('Forecast',('same','change')) #Forecast can be same or
change
a = CandidateElimination(dataset,f) #pass the dataset to the algorithm
class and call the run algoritm method
a.run_algorithm()

Dept of CSE & ISE, EPCET B’LURU Page 22


MACHINE LEARNING LABORATORY- 15CSL76

OUTPUT

Dept of CSE & ISE, EPCET B’LURU Page 23


MACHINE LEARNING LABORATORY- 15CSL76

Dept of CSE & ISE, EPCET B’LURU Page 24


MACHINE LEARNING LABORATORY- 15CSL76

Program 3

Write a program to demonstrate the working of the decision tree based ID3 algorithm. Use an
appropriate data set for building the decision tree and apply this knowledge to classify a new
sample.

Task: ID3 determines the information gain for each candidate attribute (i.e., Outlook,
Temperature, Humidity, and Wind), then selects the one with highest information gain as the
root node of the tree. The information gain values for all four attributes are calculated using the
following formula:

Entropy(S)=Σ- P(I).log2P(I)

Gain(S, A)=Entropy(S)-Σ [ P(S/A).Entropy(S/A) ]

Dataset:

Calculation:
Decision/play column consists of 14 instances and includes two labels: yes and no.
There are 9 decisions labeled yes and 5 decisions labeled no.

Entropy[Decision] = -P(yes).log2P(yes) – P(no).log2P(no)


= - (9/14).log2(9/14) – (5/14).log2(5/14)
= 0.940
Now, we need to find out the most dominant factor for decision.

Dept of CSE & ISE, EPCET B’LURU Page 25


MACHINE LEARNING LABORATORY- 15CSL76

import ast
import csv
import sys
import math
import os

def load_csv_to_header_data(filename):
path = os.path.normpath(os.getcwd() + filename)
fs = csv.reader(open('D:/prabhu/data/credit_history.csv'))
all_row = []
for r in fs:
all_row.append(r)

headers = all_row[0]
idx_to_name, name_to_idx = get_header_name_to_idx_maps(headers)

data = {
'header': headers,
'rows': all_row[1:],
'name_to_idx': name_to_idx,
'idx_to_name': idx_to_name
}
return data

def get_header_name_to_idx_maps(headers):
name_to_idx = {}
idx_to_name = {}
for i in range(0, len(headers)):
name_to_idx[headers[i]] = i
idx_to_name[i] = headers[i]
return idx_to_name, name_to_idx

def project_columns(data, columns_to_project):


data_h = list(data['header'])
data_r = list(data['rows'])
all_cols = list(range(0, len(data_h)))
columns_to_project_ix = [data['name_to_idx'][name] for name in
columns_to_project]
columns_to_remove = [cidx for cidx in all_cols if cidx not in
columns_to_project_ix]

for delc in sorted(columns_to_remove, reverse=True):


del data_h[delc]
for r in data_r:
del r[delc]

Dept of CSE & ISE, EPCET B’LURU Page 26


MACHINE LEARNING LABORATORY- 15CSL76

idx_to_name, name_to_idx = get_header_name_to_idx_maps(data_h)

return {'header': data_h, 'rows': data_r,


'name_to_idx': name_to_idx,
'idx_to_name': idx_to_name}

def get_uniq_values(data):
idx_to_name = data['idx_to_name']
idxs = idx_to_name.keys()

val_map = {}
for idx in iter(idxs):
val_map[idx_to_name[idx]] = set()

for data_row in data['rows']:


for idx in idx_to_name.keys():
att_name = idx_to_name[idx]
val = data_row[idx]
if val not in val_map.keys():
val_map[att_name].add(val)
return val_map

def get_class_labels(data, target_attribute):


rows = data['rows']
col_idx = data['name_to_idx'][target_attribute]
labels = {}
for r in rows:
val = r[col_idx]
if val in labels:
labels[val] = labels[val] + 1
else:
labels[val] = 1
return labels

def entropy(n, labels):


ent = 0
for label in labels.keys():
p_x = labels[label] / n
ent += - p_x * math.log(p_x, 2)
return ent
def partition_data(data, group_att):
partitions = {}
data_rows = data['rows']
partition_att_idx = data['name_to_idx'][group_att]

Dept of CSE & ISE, EPCET B’LURU Page 27


MACHINE LEARNING LABORATORY- 15CSL76

for row in data_rows:


row_val = row[partition_att_idx]
if row_val not in partitions.keys():
partitions[row_val] = {
'name_to_idx': data['name_to_idx'],
'idx_to_name': data['idx_to_name'],
'rows': list()
}
partitions[row_val]['rows'].append(row)
return partitions

def avg_entropy_w_partitions(data, splitting_att, target_attribute):


# find uniq values of splitting att
data_rows = data['rows']
n = len(data_rows)
partitions = partition_data(data, splitting_att)
avg_ent = 0

for partition_key in partitions.keys():


partitioned_data = partitions[partition_key]
partition_n = len(partitioned_data['rows'])
partition_labels = get_class_labels(partitioned_data,
target_attribute)
partition_entropy = entropy(partition_n, partition_labels)
avg_ent += partition_n / n * partition_entropy
return avg_ent, partitions

def most_common_label(labels):
mcl = max(labels, key=lambda k: labels[k])
return mcl

def id3(data, uniqs, remaining_atts, target_attribute):


labels = get_class_labels(data, target_attribute)
node = {}
if len(labels.keys()) == 1:
node['label'] = next(iter(labels.keys()))
return node
if len(remaining_atts) == 0:
node['label'] = most_common_label(labels)
return node
n = len(data['rows'])
ent = entropy(n, labels)
max_info_gain = None
max_info_gain_att = None
max_info_gain_partitions = None

Dept of CSE & ISE, EPCET B’LURU Page 28


MACHINE LEARNING LABORATORY- 15CSL76

for remaining_att in remaining_atts:


avg_ent, partitions = avg_entropy_w_partitions(data,
remaining_att, target_attribute)
info_gain = ent - avg_ent
if max_info_gain is None or info_gain > max_info_gain:
max_info_gain = info_gain
max_info_gain_att = remaining_att
max_info_gain_partitions = partitions

if max_info_gain is None:
node['label'] = most_common_label(labels)
return node

node['attribute'] = max_info_gain_att
node['nodes'] = {}

remaining_atts_for_subtrees = set(remaining_atts)
remaining_atts_for_subtrees.discard(max_info_gain_att)

uniq_att_values = uniqs[max_info_gain_att]

for att_value in uniq_att_values:


if att_value not in max_info_gain_partitions.keys():
node['nodes'][att_value] = {'label':
most_common_label(labels)}
continue
partition = max_info_gain_partitions[att_value]
node['nodes'][att_value] = id3(partition, uniqs,
remaining_atts_for_subtrees, target_attribute)
return node

def load_config(config_file):
with open('D:/prabhu/data/credit_history.cfg', 'r') as myfile:
data = myfile.read().replace('\n', '')
return ast.literal_eval(data)

def pretty_print_tree(root):
stack = []
rules = set()

def traverse(node, stack, rules):


if 'label' in node:
stack.append(' THEN ' + node['label'])
rules.add(''.join(stack))

Dept of CSE & ISE, EPCET B’LURU Page 29


MACHINE LEARNING LABORATORY- 15CSL76

stack.pop()
elif 'attribute' in node:
ifnd = 'IF ' if not stack else ' AND '
stack.append(ifnd + node['attribute'] + ' EQUALS ')
for subnode_key in node['nodes']:
stack.append(subnode_key)
traverse(node['nodes'][subnode_key], stack, rules)
stack.pop()
stack.pop()
traverse(root, stack, rules)
print(os.linesep.join(rules))

def main():
config = load_config('D:/prabhu/data/credit_history.cfg')
data = load_csv_to_header_data(config['data_file'])
data = project_columns(data, config['data_project_columns'])
target_attribute = config['target_attribute']
remaining_attributes = set(data['header'])
remaining_attributes.remove(target_attribute)
uniqs = get_uniq_values(data)
root = id3(data, uniqs, remaining_attributes, target_attribute)
pretty_print_tree(root)
main()

OUTPUT

Dept of CSE & ISE, EPCET B’LURU Page 30


MACHINE LEARNING LABORATORY- 15CSL76

Dept of CSE & ISE, EPCET B’LURU Page 31


MACHINE LEARNING LABORATORY- 15CSL76

Program 4:

Build an Artificial Neural Network by implementing the Back propagation algorithm and
test the same using appropriate data set.

Back propagation Algorithm

The Backpropagation algorithm is a supervised learning method for multilayer feed-forward networks
from the field of Artificial Neural Networks, the backpropagation algorithm is a method for training the
weights in a multilayer feed-forward neural network. As such, it requires a network structure to be refined
of one or more layers where one layer is fully connected to the next layer.A standard network structure is
one input layer, one hidden layer, and one output layer. Backpropagation can be used for both
classification and regression problems.

Working of algorithm:

1. Initialize Network.
2. Forward Propagate.
3. Back Propagate Error.
4. Train the network

1. Initialize Network

Initialize the network weights to small random numbers in the range of 0 to 1. Below is a function named
initialize_network() that creates a new neural network ready for training. It accepts three parameters,
i. the number of inputs
ii. the number of neurons to have in the hidden layer and
iii. the number of outputs.

The hidden layer we create n_hidden neurons and each neuron in the hidden layer has n_inputs + 1
weights, one for each input column in a dataset and an additional one for the bias. The output layer that
connects to the hidden layer has n_outputs neurons, each with n_hidden + 1 weights. This means that
each neuron in the output layer connects to (has a weight for) each neuron in the hidden layer. Let’s test
out this function. Below is a complete example that creates a small network.

2. Forward Propagate

We can calculate an output from a neural network by propagating an input signal through each layer
until the output layer outputs its values. We call this forward-propagation.
We can break forward propagation down into three parts:
• Neuron Activation.
• Neuron Transfer.
• Forward Propagation.
Neuron Activation
The first step is to calculate the activation of one neuron given an input. The input could be a row from
our training dataset, as in the case of the hidden layer. It may also be the outputs from each neuron in the
hidden layer, in the case of the output layer. Neuron activation is calculated as the weighted sum of the
inputs. Much like linear regression.
activation = sum(weight_i * input_i) + bias

Dept of CSE & ISE, EPCET B’LURU Page 32


MACHINE LEARNING LABORATORY- 15CSL76

Where weight is a network weight, input is an input, i is the index of a weight or an input and bias is a
special weight that has no input to multiply with (or you can think of the input as always being 1.0).
Below is an implementation of this in a function named activate(). You can see that the function
assumes that the bias is the last weight in the list of weights. This helps here and later to make the code
easier to read.

Neuron Transfer

Once a neuron is activated, we need to transfer the activation to see what the neuron output actually is.
Different transfer functions can be used. It is traditional to use the sigmoid activation function. It is also
a function of which we can easily calculate the derivative (slope) that we will need later when
backpropagating error. We can transfer an activation function using the sigmoid function as follows:

output = 1 / (1 + e^(-activation))

Where e is the base of the natural logarithms (Euler’s number).

Forward Propagation

Forward propagating an input is straightforward. All of the outputs from one layer become inputs to the
neurons on the next layer. Below is a function named forward_propagate() that implements the
forward propagation for a row of data from our dataset with our neural network.

3. Back Propagate Error

The backpropagation algorithm is named for the way in which weights are trained. Error is calculated
between the expected outputs and the outputs forward propagated from the network. These errors are
then propagated backward through the network from the output layer to the hidden layer, assigning
blame for the error and updating weights as they go. The math for backpropagating error is rooted in
calculus, but we will remain high level in this section and focus on what is calculated and how rather
than why the calculations take this particular form. This part is broken down into two sections.

• Transfer Derivative.
• Error Backpropagation.

Transfer Derivative

Given an output value from a neuron, we need to calculate it’s slope. We are using the sigmoid transfer
function, the derivative of which can be calculated as follows:
derivative = output * (1.0 - output)
Error Backpropagation

The first step is to calculate the error for each output neuron; this will give us our error signal (input) to
propagate backwards through the network. The error for a given neuron can be calculated as follows:

error = (expected - output) * transfer_derivative(output)

Dept of CSE & ISE, EPCET B’LURU Page 33


MACHINE LEARNING LABORATORY- 15CSL76

Where expected is the expected output value for the neuron, output is the output value for the neuron
and transfer_derivative() calculates the slope of the neuron’s output value, as shown above. The error
signal for a neuron in the hidden layer is calculated as the weighted error of each neuron in the output
layer. Think of the error traveling back along the weights of the output layer to the neurons in the hidden
layer.

The back-propagated error signal is accumulated and then used to determine the error for the neuron in
the hidden layer, as follows:

error = (weight_k * error_j) * transfer_derivative(output)

Where error_j is the error signal from the jth neuron in the output layer, weight_k is the weight that
connects the kth neuron to the current neuron and output is the output for the current neuron.

4. Train Network

The network is trained using stochastic gradient descent. This involves multiple iterations of exposing a
training dataset to the network and for each row of data forward propagating the inputs, backpropagating
the error and updating the network weights. This part is broken down into two sections:

• Update Weights.
• Train Network.

Update Weights

Once errors are calculated for each neuron in the network via the back propagation method above, they
can be used to update weights. Network weights are updated as follows:

weight = weight + learning_rate * error * input

Where weight is a given weight, learning_rate is a parameter that you must specify, error is the error
calculated by the backpropagation procedure for the neuron and input is the input value that caused the
error.

Train Network

As mentioned, the network is updated using stochastic gradient descent. This involves first looping for
a fixed number of epochs and within each epoch updating the network for each row in the training
dataset. Below is a function that implements the training of an already initialized neural network with a
given training dataset, learning rate, fixed number of epochs and an expected number of output values.

Dept of CSE & ISE, EPCET B’LURU Page 34


MACHINE LEARNING LABORATORY- 15CSL76

from math import exp


from random import seed
from random import random
# Initialize a network
def initialize_network(n_inputs, n_hidden, n_outputs):
network = list()
hidden_layer = [{'weights':[random() for i in range(n_inputs +
1)]} for i in range(n_hidden)]
network.append(hidden_layer)
output_layer = [{'weights':[random() for i in range(n_hidden +
1)]} for i in range(n_outputs)]
network.append(output_layer)
return network

# Calculate neuron activation for an input


def activate(weights, inputs):
activation = weights[-1]
for i in range(len(weights)-1):
activation += weights[i] * inputs[i]
return activation

# Transfer neuron activation


def transfer(activation):
return 1.0 / (1.0 + exp(-activation))

# Forward propagate input to a network output


def forward_propagate(network, row):
inputs = row
for layer in network:
new_inputs = []
for neuron in layer:
activation = activate(neuron['weights'], inputs)
neuron['output'] = transfer(activation)
new_inputs.append(neuron['output'])
inputs = new_inputs
return inputs

# Calculate the derivative of an neuron output


def transfer_derivative(output):
return output * (1.0 - output)

# Backpropagate error and store in neurons


def backward_propagate_error(network, expected):
for i in reversed(range(len(network))):
layer = network[i]

Dept of CSE & ISE, EPCET B’LURU Page 35


MACHINE LEARNING LABORATORY- 15CSL76

errors = list()
if i != len(network)-1:
for j in range(len(layer)):
error = 0.0
for neuron in network[i + 1]:
error += (neuron['weights'][j] * neuron['delta'])
errors.append(error)
else:
for j in range(len(layer)):
neuron = layer[j]
errors.append(expected[j] - neuron['output'])
for j in range(len(layer)):
neuron = layer[j]
neuron['delta'] = errors[j] *
transfer_derivative(neuron['output'])

# Update network weights with error


def update_weights(network, row, l_rate):
for i in range(len(network)):
inputs = row[:-1]
if i != 0:
inputs = [neuron['output'] for neuron in network[i - 1]]
for neuron in network[i]:
for j in range(len(inputs)):
neuron['weights'][j] += l_rate * neuron['delta'] *
inputs[j]
neuron['weights'][-1] += l_rate * neuron['delta']

# Train a network for a fixed number of epochs


def train_network(network, train, l_rate, n_epoch, n_outputs):
for epoch in range(n_epoch):
sum_error = 0
for row in train:
outputs = forward_propagate(network, row)
expected = [0 for i in range(n_outputs)]
expected[row[-1]] = 1
sum_error += sum([(expected[i]-outputs[i])**2 for i in
range(len(expected))])
backward_propagate_error(network, expected)
update_weights(network, row, l_rate)
print('>epoch=%d, lrate=%.3f, error=%.3f' % (epoch, l_rate,
sum_error))

# Test training backprop algorithm


seed(1)

Dept of CSE & ISE, EPCET B’LURU Page 36


MACHINE LEARNING LABORATORY- 15CSL76

dataset = [[2.7810836,2.550537003,0],
[1.465489372,2.362125076,0],
[3.396561688,4.400293529,0],
[1.38807019,1.850220317,0],
[3.06407232,3.005305973,0],
[7.627531214,2.759262235,1],
[5.332441248,2.088626775,1],
[6.922596716,1.77106367,1],
[8.675418651,-0.242068655,1],
[7.673756466,3.508563011,1]]
n_inputs = len(dataset[0]) - 1
n_outputs = len(set([row[-1] for row in dataset]))

network = initialize_network(n_inputs, 2, n_outputs)


train_network(network, dataset, 0.5, 20, n_outputs)
for layer in network:
print(layer)

OUTPUT

Dept of CSE & ISE, EPCET B’LURU Page 37


MACHINE LEARNING LABORATORY- 15CSL76

Dept of CSE & ISE, EPCET B’LURU Page 38


MACHINE LEARNING LABORATORY- 15CSL76

Dept of CSE & ISE, EPCET B’LURU Page 39


MACHINE LEARNING LABORATORY- 15CSL76

Program 5

Write a program to implement the naïve Bayesian classifier for a sample training data set stored
as a .CSV file. Compute the accuracy of the classifier, considering few test data sets.

Task: It is a classification technique based on Bayes’ Theorem with an assumption of independence


among predictors. In simple terms, a Naive Bayes classifier assumes that the presence of a particular
feature in a class is unrelated to the presence of any other feature. For example, a fruit may be considered
to be an apple if it is red, round, and about 3 inches in diameter. Even if these features depend on each
other or upon the existence of the other features, all of these properties independently contribute to the
probability that this fruit is an apple and that is why it is known as ‘Naive’.

Bayes theorem provides a way of calculating posterior probability

P(c|x) from P(c), P(x) and P(x|c).

Above,

• P(c|x) is the posterior probability of class (c, target) given predictor (x, attributes).
• P(c) is the prior probability of class.
• P(x|c) is the likelihood which is the probability of predictor given class.
• P(x) is the prior probability of predictor

Dept of CSE & ISE, EPCET B’LURU Page 40


MACHINE LEARNING LABORATORY- 15CSL76

import csv
import math
import random

#Handle data
def loadCsv(filename):
lines = csv.reader(open(filename, "r"))
dataset = list(lines)
for i in range(len(dataset)):
dataset[i] = [float(x) for x in dataset[i]]
return dataset

#Split dataset with ratio


def splitDataset(dataset, splitRatio):
trainSize = int(len(dataset) * splitRatio)
trainSet = []
copy = list(dataset)
while len(trainSet) < trainSize:
index = random.randrange(len(copy))
trainSet.append(copy.pop(index))
return [trainSet, copy]

#Separate by Class
def separateByClass(dataset):
separated = {}
for i in range(len(dataset)):
vector = dataset[i]
if (vector[-1] not in separated):
separated[vector[-1]] = []
separated[vector[-1]].append(vector)
return separated

#Calculate Mean
def mean(numbers):
return sum(numbers)/float(len(numbers))

def stdev(numbers):
avg = mean(numbers)
variance = sum([pow(x-avg,2) for x in
numbers])/float(len(numbers)-1)
return math.sqrt(variance)

#Summarize Dataset
def summarize(dataset):

Dept of CSE & ISE, EPCET B’LURU Page 41


MACHINE LEARNING LABORATORY- 15CSL76

summaries = [(mean(attribute), stdev(attribute)) for attribute in


zip(*dataset)]
del summaries[-1]
return summaries

#Summarize attributes by class


def summarizeByClass(dataset):
separated = separateByClass(dataset)
summaries = {}
for classValue, instances in separated.items():
summaries[classValue] = summarize(instances)
return summaries

#Calculate Gaussian Probability Density Function


def calculateProbability(x, mean, stdev):
exponent = math.exp(-(math.pow(x-mean,2)/(2*math.pow(stdev,2))))
return (1/(math.sqrt(2*math.pi)*stdev))*exponent

#Calculate Class Probabilities


def calculateClassProbabilities(summaries, inputVector):
probabilities = {}
for classValue, classSummaries in summaries.items():
probabilities[classValue] = 1
for i in range(len(classSummaries)):
mean, stdev = classSummaries[i]
x = inputVector[i]
probabilities[classValue] *= calculateProbability(x,
mean, stdev)
return probabilities

#Make a prediction
def predict(summaries, inputVector):
probabilities = calculateClassProbabilities(summaries,
inputVector)
bestLabel, bestProb = None, -1
for classValue, probability in probabilities.items():
if bestLabel is None or probability > bestProb:
bestProb = probability
bestLabel = classValue
return bestLabel

#Get predictions

def getPredictions(summaries, testSet):


predictions = []

Dept of CSE & ISE, EPCET B’LURU Page 42


MACHINE LEARNING LABORATORY- 15CSL76

for i in range(len(testSet)):
result = predict(summaries, testSet[i])
predictions.append(result)
return predictions

#Get Accuracy
def getAccuracy(testSet, predictions):
correct = 0
for x in range(len(testSet)):
if testSet[x][-1] == predictions[x]:
correct += 1
return (correct/float(len(testSet)))*100.0

def main():
filename = 'D:\prabhu\pima.csv'
splitRatio = 0.67
dataset = loadCsv(filename)
trainingSet, testSet = splitDataset(dataset, splitRatio)
print('Split {0} rows into train = {1} and test = {2}
rows'.format(len(dataset),len(trainingSet),len(testSet)))
#prepare model
summaries = summarizeByClass(trainingSet)
#test model
predictions = getPredictions(summaries, testSet)
accuracy = getAccuracy(testSet, predictions)
print('Accuracy: {0}%'.format(accuracy))

main()

OUTPUT

Dept of CSE & ISE, EPCET B’LURU Page 43


MACHINE LEARNING LABORATORY- 15CSL76

Dept of CSE & ISE, EPCET B’LURU Page 44


MACHINE LEARNING LABORATORY- 15CSL76

Program 6:

Assuming a set of documents that need to be classified, use the naïve Bayesian Classifier model to
perform this task. Built-in Java classes/API can be used to write the program. Calculate the
accuracy, precision, and recall for your data set.

Multinomial Naïve Bayes Classifier:

Dept of CSE & ISE, EPCET B’LURU Page 45


MACHINE LEARNING LABORATORY- 15CSL76

import pandas as pd
msg=pd.read_csv('bayes.csv',names=['message','label'])
print('The dimensions of the dataset',msg.shape)
msg['labelnum']=msg.label.map({'pos':1,'neg':0})
X=msg.message
y=msg.labelnum
print(X)
print(y)
#splitting the dataset into train and test data
from sklearn.model_selection import train_test_split
xtrain,xtest,ytrain,ytest=train_test_split(X,y)
print(xtest.shape)
print(xtrain.shape)
print(ytest.shape)
print(ytrain.shape)
#output of count vectoriser is a sparse matrix
from sklearn.feature_extraction.text import CountVectorizer
count_vect = CountVectorizer()
xtrain_dtm = count_vect.fit_transform(xtrain)
xtest_dtm=count_vect.transform(xtest)
print(count_vect.get_feature_names())
df=pd.DataFrame(xtrain_dtm.toarray(),columns=count_vect.get_feature_na
mes())
print(df)#tabular representation
print(xtrain_dtm) #sparse matrix representation
# Training Naive Bayes (NB) classifier on training data.
from sklearn.naive_bayes import MultinomialNB
clf = MultinomialNB().fit(xtrain_dtm,ytrain)
predicted = clf.predict(xtest_dtm)
#printing accuracy metrics
from sklearn import metrics
print('Accuracy metrics')
print('Accuracy of the classifer
is',metrics.accuracy_score(ytest,predicted))
print('Confusion matrix')
print(metrics.confusion_matrix(ytest,predicted))
print('Recall and Precison ')
print(metrics.recall_score(ytest,predicted))
print(metrics.precision_score(ytest,predicted))

Dept of CSE & ISE, EPCET B’LURU Page 46


MACHINE LEARNING LABORATORY- 15CSL76

OUTPUT

Dept of CSE & ISE, EPCET B’LURU Page 47


MACHINE LEARNING LABORATORY- 15CSL76

Dept of CSE & ISE, EPCET B’LURU Page 48


MACHINE LEARNING LABORATORY- 15CSL76

Dept of CSE & ISE, EPCET B’LURU Page 49


MACHINE LEARNING LABORATORY- 15CSL76

Program 7

Write a program to construct a Bayesian network considering medical data. Use this model to
demonstrate the diagnosis of heart patients using standard Heart Disease Data Set. You can use
Java/Python ML library classes/API.

A Bayesian belief network describes the probability distribution over a set of variables.

Probability

P(A) is used to denote the probability of A. For example if A is discrete with states {True, False} then
P(A) might equal [0.2, 0.8]. I.e. 20% chance of being True, 80% chance of being False.

Joint probability

A joint probability refers to the probability of more than one variable occurring together, such as the
probability of A and B, denoted P(A,B).

Conditional probability

Conditional probability is the probability of a variable (or set of variables) given another variable (or set
of variables), denoted P(A|B).For example, the probability of Windy being True, given that Raining is
True might equal 50%.This would be denoted P(Windy = True | Raining = True) = 50%.

Once the structure has been defined (i.e. nodes and links), a Bayesian network requires a probability
distribution to be assigned to each node. Each node X in a Bayesian network requires a probability
distribution P(X | pa(X)).Note that if a node X has no parents pa(X) is empty, and the required
distribution is just P(X) sometimes referred to as the prior. This is the probability of itself given its parent
nodes.

If U = {A1,...,An} is the universe of variables (all the variables) in a Bayesian network, and pa(Ai) are
the parents of Ai then the joint probability distribution P(U) is the simply the product of all the
probability distributions (prior and conditional) in the network, as shown in the equation below.This
equation is known as the chain rule.

Dept of CSE & ISE, EPCET B’LURU Page 50


MACHINE LEARNING LABORATORY- 15CSL76

from pgmpy.models import BayesianModel

cancer_model = BayesianModel([('Pollution', 'Cancer'),


('Smoker', 'Cancer'),
('Cancer', 'Xray'),
('Cancer', 'Dyspnoea')])
print(cancer_model)
cancer_model.nodes()
cancer_model.edges()
cancer_model.get_cpds()
from pgmpy.factors.discrete import TabularCPD

cpd_poll = TabularCPD(variable='Pollution', variable_card=2,


values=[[0.9], [0.1]])
cpd_smoke = TabularCPD(variable='Smoker', variable_card=2,
values=[[0.3], [0.7]])
cpd_cancer = TabularCPD(variable='Cancer', variable_card=2,
values=[[0.03, 0.05, 0.001, 0.02],
[0.97, 0.95, 0.999, 0.98]],
evidence=['Smoker', 'Pollution'],
evidence_card=[2, 2])
cpd_xray = TabularCPD(variable='Xray', variable_card=2,
values=[[0.9, 0.2], [0.1, 0.8]],
evidence=['Cancer'], evidence_card=[2])
cpd_dysp = TabularCPD(variable='Dyspnoea', variable_card=2,
values=[[0.65, 0.3], [0.35, 0.7]],
evidence=['Cancer'], evidence_card=[2])
# Associating the parameters with the model structure.
cancer_model.add_cpds(cpd_poll, cpd_smoke, cpd_cancer, cpd_xray,
cpd_dysp)

# Checking if the cpds are valid for the model.


cancer_model.check_model()
# Doing some simple queries on the network
cancer_model.is_active_trail('Pollution', 'Smoker')
cancer_model.is_active_trail('Pollution', 'Smoker',
observed=['Cancer'])
cancer_model.get_cpds()
print(cancer_model.get_cpds('Pollution'))
print(cancer_model.get_cpds('Smoker'))
print(cancer_model.get_cpds('Xray'))
print(cancer_model.get_cpds('Dyspnoea'))
print(cancer_model.get_cpds('Cancer'))
cancer_model.local_independencies('Xray')
cancer_model.local_independencies('Pollution')

Dept of CSE & ISE, EPCET B’LURU Page 51


MACHINE LEARNING LABORATORY- 15CSL76

cancer_model.local_independencies('Smoker')
cancer_model.local_independencies('Dyspnoea')
cancer_model.local_independencies('Cancer')
cancer_model.get_independencies()

from pgmpy.inference import VariableElimination


cancer_infer = VariableElimination(cancer_model)

# Computing the probability of bronc given smoke.


q = cancer_infer.query(variables=['Cancer'], evidence={'Smoker': 1})
print(q['Cancer'])

# Computing the probability of bronc given smoke.


q = cancer_infer.query(variables=['Cancer'], evidence={'Smoker': 1})
print(q['Cancer'])

# Computing the probability of bronc given smoke.


q = cancer_infer.query(variables=['Cancer'], evidence={'Smoker':
1,'Pollution': 1})
print(q['Cancer'])

OUTPUT

Dept of CSE & ISE, EPCET B’LURU Page 52


MACHINE LEARNING LABORATORY- 15CSL76

Dept of CSE & ISE, EPCET B’LURU Page 53


MACHINE LEARNING LABORATORY- 15CSL76

import numpy as np
from urllib.request import urlopen
import urllib
import matplotlib.pyplot as plt # Visuals
import seaborn as sns
import sklearn as skl
import pandas as pd

Cleveland_data_URL = 'http://archive.ics.uci.edu/ml/machine-learning-
databases/heart-disease/processed.hungarian.data'
np.set_printoptions(threshold=np.nan)
names = ['age', 'sex', 'cp', 'trestbps', 'chol', 'fbs', 'restecg',
'thalach', 'exang', 'oldpeak', 'slope', 'ca', 'thal', 'heartdisease']
heartDisease = pd.read_csv(urlopen(Cleveland_data_URL), names = names)
heartDisease.head()
del heartDisease['ca']
del heartDisease['slope']
del heartDisease['thal']
del heartDisease['oldpeak']

heartDisease = heartDisease.replace('?', np.nan)


heartDisease.dtypes
heartDisease.columns

from pgmpy.models import BayesianModel


from pgmpy.estimators import MaximumLikelihoodEstimator,
BayesianEstimator

model = BayesianModel([('age', 'trestbps'), ('age', 'fbs'), ('sex',


'trestbps'), ('sex', 'trestbps'),
('exang',
'trestbps'),('trestbps','heartdisease'),('fbs','heartdisease'),

('heartdisease','restecg'),('heartdisease','thalach'),('heartdisease',
'chol')])
model.fit(heartDisease, estimator=MaximumLikelihoodEstimator)
#print(model.get_cpds('age'))

print(model.get_cpds('chol'))

print(model.get_cpds('sex'))

model.get_independencies()

from pgmpy.inference import VariableElimination

Dept of CSE & ISE, EPCET B’LURU Page 54


MACHINE LEARNING LABORATORY- 15CSL76

HeartDisease_infer = VariableElimination(model)

# Computing the probability of bronc given smoke.


q = HeartDisease_infer.query(variables=['heartdisease'],
evidence={'age': 28})
print(q['heartdisease'])

q = HeartDisease_infer.query(variables=['heartdisease'],
evidence={'chol': 100})
print(q['heartdisease'])

OUTPUT

Dept of CSE & ISE, EPCET B’LURU Page 55


MACHINE LEARNING LABORATORY- 15CSL76

Dept of CSE & ISE, EPCET B’LURU Page 56


MACHINE LEARNING LABORATORY- 15CSL76

Program 8

Apply EM algorithm to cluster a set of data stored in a .CSV file. Use the same data set for
clustering using k-Means algorithm. Compare the results of these two algorithms and comment on
the quality of clustering. You can add Java/Python ML library classes/API in the program.

Introduction to Expectation-Maximization (EM)

The EM algorithm tends to get stuck less than K-means algorithm. The idea is to assign data
points partially to different clusters instead of assigning to only one cluster. To do this partial
assignment, we model each cluster using a probabilistic distribution So a data point associates
with a cluster with certain probability and it belongs to the cluster with the highest probability in
the final assignment.

Expectation-Maximization (EM) algorithm

Dept of CSE & ISE, EPCET B’LURU Page 57


MACHINE LEARNING LABORATORY- 15CSL76

import numpy as np
import pandas as pd
from matplotlib import pyplot as plt
fromsklearn.mixture import GMM
fromsklearn.cluster import KMeans

# Importing the dataset


data = pd.read_csv('xclara.csv')
print("Input Data and Shape")
print(data.shape)
data.head()

# Getting the values and plotting it


f1 = data['V1'].values
f2 = data['V2'].values
X = np.array(list(zip(f1, f2)))
plt.scatter(f1, f2, c='black', s=7)
plt.show()
##########################################
kmeans = KMeans(3, random_state=0)
labels = kmeans.fit(X).predict(X)
centroids = kmeans.cluster_centers_

plt.scatter(X[:, 0], X[:, 1], c=labels, s=40, cmap='viridis');


plt.scatter(centroids[:, 0], centroids[:, 1], marker='*', s=200, c='#050505')
plt.show()

#gmm demo

gmm = GMM(n_components=3).fit(X)
labels = gmm.predict(X)
# forploting
probs = gmm.predict_proba(X)
size = 10 * probs.max(1) **15
print(probs[:300].round(3))
plt.scatter(X[:, 0], X[:, 1], c=labels, s=size, cmap='viridis');
plt.show()

Dept of CSE & ISE, EPCET B’LURU Page 58


MACHINE LEARNING LABORATORY- 15CSL76

[[1. 0. 0. ]
[1. 0. 0. ]
[1. 0. 0. ]
[1. 0. 0. ]
[1. 0. 0. ]
[0.848 0.152 0. ]
[1. 0. 0. ]
[1. 0. 0. ]
[1. 0. 0. ]
[1. 0. 0. ]
[1. 0. 0. ]
[1. 0. 0. ]
[1. 0. 0. ]
[1. 0. 0. ]
[1. 0. 0. ]
[1. 0. 0. ]
[1. 0. 0. ]
[1. 0. 0. ]
[1. 0. 0. ]
[1. 0. 0. ]
[0.999 0.001 0. ]
[1. 0. 0. ]
[1. 0. 0. ]
[1. 0. 0. ]
[1. 0. 0. ]
[1. 0. 0. ]
[1. 0. 0. ]
[1. 0. 0. ]
[1. 0. 0. ]
[0.067 0.933 0. ]
[1. 0. 0. ]
[1. 0. 0. ]
[1. 0. 0. ]
[0.983 0. 0.017]
[1. 0. 0. ]
[1. 0. 0. ]
[1. 0. 0. ]
[1. 0. 0. ]
[1. 0. 0. ]

Dept of CSE & ISE, EPCET B’LURU Page 59


MACHINE LEARNING LABORATORY- 15CSL76

[1. 0. 0. ]
[1. 0. 0. ]
[1. 0. 0. ]
[1. 0. 0. ]
[1. 0. 0. ]
[1. 0. 0. ]
[1. 0. 0. ]
[1. 0. 0. ]
[1. 0. 0. ]
[1. 0. 0. ]
[1. 0. 0. ]
[1. 0. 0. ]
[1. 0. 0. ]
[1. 0. 0. ]
[1. 0. 0. ]
[1. 0. 0. ]
[1. 0. 0. ]
[1. 0. 0. ]
[1. 0. 0. ]
[1. 0. 0. ]
[1. 0. 0. ]
[0.997 0. 0.003]
[1. 0. 0. ]
[1. 0. 0. ]
[1. 0. 0. ]
[1. 0. 0. ]
[1. 0. 0. ]
[1. 0. 0. ]
[1. 0. 0. ]
[1. 0. 0. ]
[1. 0. 0. ]
[0.904 0.096 0. ]
[1. 0. 0. ]
[1. 0. 0. ]
[1. 0. 0. ]
[1. 0. 0. ]
[1. 0. 0. ]
[1. 0. 0. ]
[1. 0. 0. ]
[1. 0. 0. ]
[1. 0. 0. ]
[1. 0. 0. ]

Dept of CSE & ISE, EPCET B’LURU Page 60


MACHINE LEARNING LABORATORY- 15CSL76

[1. 0. 0. ]
[1. 0. 0. ]
[1. 0. 0. ]
[0.995 0.005 0. ]
[1. 0. 0. ]
[0.992 0.008 0. ]
[1. 0. 0. ]
[1. 0. 0. ]
[1. 0. 0. ]
[0.999 0. 0.001]
[1. 0. 0. ]
[1. 0. 0. ]
[1. 0. 0. ]
[1. 0. 0. ]
[0.996 0.004 0. ]
[1. 0. 0. ]
[1. 0. 0. ]
[1. 0. 0. ]
[1. 0. 0. ]
[1. 0. 0. ]
[1. 0. 0. ]
[1. 0. 0. ]
[1. 0. 0. ]
[0.997 0.001 0.003]
[1. 0. 0. ]
[1. 0. 0. ]
[1. 0. 0. ]
[1. 0. 0. ]
[0.993 0.007 0. ]
[1. 0. 0. ]
[1. 0. 0. ]
[1. 0. 0. ]
[1. 0. 0. ]
[0.998 0. 0.002]
[1. 0. 0. ]
[1. 0. 0. ]
[1. 0. 0. ]
[1. 0. 0. ]
[1. 0. 0. ]
[1. 0. 0. ]
[1. 0. 0. ]
[1. 0. 0. ]

Dept of CSE & ISE, EPCET B’LURU Page 61


MACHINE LEARNING LABORATORY- 15CSL76

[1. 0. 0. ]
[1. 0. 0. ]
[1. 0. 0. ]
[1. 0. 0. ]
[1. 0. 0. ]
[1. 0. 0. ]
[1. 0. 0. ]
[1. 0. 0. ]
[1. 0. 0. ]
[1. 0. 0. ]
[1. 0. 0. ]
[1. 0. 0. ]
[1. 0. 0. ]
[0.964 0. 0.036]
[1. 0. 0. ]
[1. 0. 0. ]
[1. 0. 0. ]
[1. 0. 0. ]
[1. 0. 0. ]
[1. 0. 0. ]
[1. 0. 0. ]
[1. 0. 0. ]
[1. 0. 0. ]
[1. 0. 0. ]
[1. 0. 0. ]
[1. 0. 0. ]
[1. 0. 0. ]
[1. 0. 0. ]
[1. 0. 0. ]
[1. 0. 0. ]
[1. 0. 0. ]
[1. 0. 0. ]
[1. 0. 0. ]
[1. 0. 0. ]
[1. 0. 0. ]
[1. 0. 0. ]
[1. 0. 0. ]
[1. 0. 0. ]
[1. 0. 0. ]
[1. 0. 0. ]
[1. 0. 0. ]
[1. 0. 0. ]

Dept of CSE & ISE, EPCET B’LURU Page 62


MACHINE LEARNING LABORATORY- 15CSL76

[1. 0. 0. ]
[1. 0. 0. ]
[1. 0. 0. ]
[1. 0. 0. ]
[1. 0. 0. ]
[1. 0. 0. ]
[1. 0. 0. ]
[1. 0. 0. ]
[1. 0. 0. ]
[1. 0. 0. ]
[1. 0. 0. ]
[1. 0. 0. ]
[1. 0. 0. ]
[1. 0. 0. ]
[1. 0. 0. ]
[0.997 0.003 0. ]
[0.921 0.079 0. ]
[1. 0. 0. ]
[1. 0. 0. ]
[1. 0. 0. ]
[1. 0. 0. ]
[1. 0. 0. ]
[1. 0. 0. ]
[0.989 0.011 0. ]
[0.999 0.001 0. ]
[1. 0. 0. ]
[1. 0. 0. ]
[1. 0. 0. ]
[1. 0. 0. ]
[1. 0. 0. ]
[1. 0. 0. ]
[1. 0. 0. ]
[1. 0. 0. ]
[1. 0. 0. ]
[1. 0. 0. ]
[1. 0. 0. ]
[1. 0. 0. ]
[0.922 0.078 0. ]
[1. 0. 0. ]
[1. 0. 0. ]
[1. 0. 0. ]
[1. 0. 0. ]

Dept of CSE & ISE, EPCET B’LURU Page 63


MACHINE LEARNING LABORATORY- 15CSL76

[1. 0. 0. ]
[1. 0. 0. ]
[1. 0. 0. ]
[1. 0. 0. ]
[1. 0. 0. ]
[1. 0. 0. ]
[1. 0. 0. ]
[1. 0. 0. ]
[1. 0. 0. ]
[1. 0. 0. ]
[1. 0. 0. ]
[1. 0. 0. ]
[0.996 0.004 0. ]
[1. 0. 0. ]
[1. 0. 0. ]
[1. 0. 0. ]
[0.998 0.002 0. ]
[1. 0. 0. ]
[1. 0. 0. ]
[1. 0. 0. ]
[1. 0. 0. ]
[1. 0. 0. ]
[1. 0. 0. ]
[1. 0. 0. ]
[1. 0. 0. ]
[1. 0. 0. ]
[1. 0. 0. ]
[1. 0. 0. ]
[1. 0. 0. ]
[1. 0. 0. ]
[1. 0. 0. ]
[1. 0. 0. ]
[1. 0. 0. ]
[1. 0. 0. ]
[1. 0. 0. ]
[0.888 0. 0.112]
[1. 0. 0. ]
[0.719 0. 0.281]
[1. 0. 0. ]
[1. 0. 0. ]
[1. 0. 0. ]
[1. 0. 0. ]

Dept of CSE & ISE, EPCET B’LURU Page 64


MACHINE LEARNING LABORATORY- 15CSL76

[1. 0. 0. ]
[1. 0. 0. ]
[1. 0. 0. ]
[1. 0. 0. ]
[1. 0. 0. ]
[0.999 0. 0.001]
[1. 0. 0. ]
[1. 0. 0. ]
[1. 0. 0. ]
[1. 0. 0. ]
[1. 0. 0. ]
[1. 0. 0. ]
[1. 0. 0. ]
[1. 0. 0. ]
[1. 0. 0. ]
[1. 0. 0. ]
[1. 0. 0. ]
[0.995 0. 0.005]
[1. 0. 0. ]
[1. 0. 0. ]
[1. 0. 0. ]
[1. 0. 0. ]
[1. 0. 0. ]
[1. 0. 0. ]
[1. 0. 0. ]
[0.999 0.001 0. ]
[1. 0. 0. ]
[1. 0. 0. ]
[1. 0. 0. ]
[1. 0. 0. ]
[0.993 0.007 0. ]
[1. 0. 0. ]
[1. 0. 0. ]
[1. 0. 0. ]
[1. 0. 0. ]
[1. 0. 0. ]
[1. 0. 0. ]
[1. 0. 0. ]
[1. 0. 0. ]
[1. 0. 0. ]
[1. 0. 0. ]
[1. 0. 0. ]

Dept of CSE & ISE, EPCET B’LURU Page 65


MACHINE LEARNING LABORATORY- 15CSL76

[1. 0. 0. ]
[1. 0. 0. ]
[1. 0. 0. ]
[1. 0. 0. ]
[1. 0. 0. ]
[1. 0. 0. ]
[1. 0. 0. ]
[1. 0.0. ][1.0. 0. ]]

OUTPUT

Dept of CSE & ISE, EPCET B’LURU Page 66


MACHINE LEARNING LABORATORY- 15CSL76

Dept of CSE & ISE, EPCET B’LURU Page 67


MACHINE LEARNING LABORATORY- 15CSL76

Program 9

Write a program to implement k-Nearest Neighbor algorithm to classify the iris data set. Print both
correct and wrong predictions. Java/Python ML library classes can be used for this problem.

TASK: The task of this program is to classify the IRIS data set examples by using the k-Nearest neighbor
algorithm. The new instance has to be classified based on its k nearest neighbors.

KNN falls in the supervised learning family of algorithms. Informally, this means that we are given a
labeled dataset consisting of training observations (x,y) and would like to capture the relationship
between x and y. More formally, our goal is to learn a function h:X→Y so that given an unseen
observation x, h(x) can confidently predict the corresponding output y.

The KNN classifier is also a non parametric and instance-based learning algorithm.

• Non-parametric means it makes no explicit assumptions about the functional form of h,


avoiding the dangers of mismodeling the underlying distribution of the data. For example,
suppose our data is highly non-Gaussian but the learning model we choose assumes a
Gaussian form. In that case, our algorithm would make extremely poor predictions.
• Instance-based learning means that our algorithm doesn’t explicitly learn a model. Instead, it
chooses to memorize the training instances which are subsequently used as “knowledge” for
the prediction phase. Concretely, this means that only when a query to our database is made
(i.e. when we ask it to predict a label given an input), will the algorithm use the training
instances to spit out an answer.

In the classification setting, the K-nearest neighbor algorithm essentially boils down to forming a majority
vote between the K most similar instances to a given “unseen” observation. Similarity is defined
according to a distance metric between two data points. A popular choice is the Euclidean distance given
by

but other measures can be more suitable for a given setting and include the Manhattan, Chebyshev and
Hamming distance

More formally, given a positive integer K, an unseen observation x and a similarity metric d, KNN
classifier performs the following two steps:

1. It runs through the whole dataset computing d between x and each training observation. The K
points in the training data that are closest to x are called the set A. Note that K is usually odd
to prevent tie situations.

2. It then estimates the conditional probability for each class, that is, the fraction of points
in A with that given class label. (Note I(x) is the indicator function which evaluates to 1 when
the argument x is true and 0 otherwise)

Dept of CSE & ISE, EPCET B’LURU Page 68


MACHINE LEARNING LABORATORY- 15CSL76

Finally, the input x gets assigned to the class with the largest probability

EXAMPLE:

Consider the following data concerning credit default. Age and Loan are two numerical variables
(predictors) and Default is the target.

We can now use the training set to classify an unknown case (Age=48 and Loan=$142,000) using
Euclidean distance. If K=1 then the nearest neighbor is the last case in the training set with Default=Y.

D = Sqrt [(48-33)^2 + (142000-150000)^2] = 8000.01 >> Default=Y

Dept of CSE & ISE, EPCET B’LURU Page 69


MACHINE LEARNING LABORATORY- 15CSL76

With K=3, there are two Default=Y and one Default=N out of three closest neighbors. The prediction for
the unknown case is again Default=Y.

ALGORITHM:

Let m be the number of training data samples. Let p be an unknown point.


1. Store the training samples in an array of data points arr[]. This means each element of this
array represents a tuple (x, y).
2. for i=0 to m:
Calculate Euclidean distance d(arr[i], p).
3. Make set S of K smallest distances obtained. Each of these distances correspond to an already
classified data point.
4. Return the majority label among S.

Dept of CSE & ISE, EPCET B’LURU Page 70


MACHINE LEARNING LABORATORY- 15CSL76

import csv
import random
import math
import operator

# Split the data into training and test data


def loadDataset(filename, split, trainingSet=[] , testSet=[]):
with open(filename, 'r') as csvfile:
lines = csv.reader(csvfile)
dataset = list(lines)
for x in range(len(dataset)):
for y in range(4):
dataset[x][y] = float(dataset[x][y])
if random.random() < split:
trainingSet.append(dataset[x])
else:
testSet.append(dataset[x])

def euclideanDistance(instance1, instance2, length):


distance = 0
for x in range(length):
distance += pow((instance1[x] - instance2[x]), 2)
return math.sqrt(distance)

def getNeighbors(trainingSet, testInstance, k):


distances = []
length = len(testInstance)-1
for x in range(len(trainingSet)):
dist = euclideanDistance(testInstance, trainingSet[x], length)
distances.append((trainingSet[x], dist))
distances.sort(key = operator.itemgetter(1))
neighbors = []
for x in range(k):
neighbors.append(distances[x][0])
return neighbors

def getResponse(neighbors):
# Creating a list with all the possible neighbors
classVotes = {}
for x in range(len(neighbors)):
response = neighbors[x][-1]
if response in classVotes:
classVotes[response] += 1
else:
classVotes[response] = 1

Dept of CSE & ISE, EPCET B’LURU Page 71


MACHINE LEARNING LABORATORY- 15CSL76

sortedVotes = sorted(classVotes.items(),
key=operator.itemgetter(1), reverse=True)
return sortedVotes[0][0]

def getAccuracy(testSet, predictions):


correct = 0
for x in range(len(testSet)):
if testSet[x][-1] == predictions[x]:
correct += 1
return (correct/float(len(testSet))) * 100.0

def main():
trainingSet=[]
testSet=[]
split = 0.67
loadDataset('iris.csv', split, trainingSet, testSet)
print('Train set: ' + repr(len(trainingSet)))
print('Test set: ' + repr(len(testSet)))
predictions=[]
k = 3
for x in range(len(testSet)):
neighbors = getNeighbors(trainingSet, testSet[x], k)
result = getResponse(neighbors)
predictions.append(result)
#print('> predicted=' + repr(result) + ', actual=' +
repr(testSet[x][-1]))
accuracy = getAccuracy(testSet, predictions)
print('Accuracy: ', accuracy)

main()

OUTPUT

Dept of CSE & ISE, EPCET B’LURU Page 72


MACHINE LEARNING LABORATORY- 15CSL76

Dept of CSE & ISE, EPCET B’LURU Page 73


MACHINE LEARNING LABORATORY- 15CSL76

Program 10

Implement the non-parametric Locally Weighted Regression algorithm in order to fit data points.
Select appropriate data set for your experiment and draw graphs.

Nonparametric regression: is a category of regression analysis in which the predictor does not take a
predetermined form but is constructed according to information derived from the data. Nonparametric
regression requires larger sample sizes than regression based on parametric models because the data must
supply the model structure as well as the model estimates.

Nonparametric regression is used for prediction and is reliable even if hypotheses of linear regression are
not verified.

Locally weighted Learning also known as memory-based learning, instance-based learning, lazy
learning, and closely related to kernel density estimation, similarity searching and case-based reasoning.

LOWESS (Locally Weighted Scatterplot Smoothing), sometimes called LOESS (locally weighted
smoothing), is a popular tool used in regression analysis that creates a smooth line through a timeplot or
scatter plot to help you to see relationship between variables and foresee trends.

Locally weighted regression is a very powerful non-parametric model used in statistical learning.

Introduction :

Scatter-diagram smoothing (e.g. using the lowess() or loess() functions) involves drawing a smooth curve
on a scatter diagram to summarize a relationship, in a fashion that makes few assumptions initially about
the form or strength of the relationship. It is related to (and is a special case of) nonparametric regression,
in which the objective is to represent the relationship between a response variable and one or more
predictor variables, again in way that makes few assumptions about the form of the relationship. In other
words, in contrast to “standard” linear regression analysis, no assumption is made that the relationship is
represented by a straight line (although one could certainly think of a straight line as a special case of
nonparametric regression).

Locally weighted regression

Local means using nearby points (i.e. a nearest neighbors approach)

Weighted means we value points based upon how far away they are.

Regression means approximating a function

This is an instance-based learning method

The idea: whenever you want to classify a sample:

• Build a local model of the function (using a linear function, quadratic, neural network, etc.)
• Use the model to predict the output value
• Throw the model away.

Dept of CSE & ISE, EPCET B’LURU Page 74


MACHINE LEARNING LABORATORY- 15CSL76

Locally Weighted Regression

Our final method combines advantages of parametric methods with non-parametric. The idea is to fit a
regression model locally, weighting examples by the kernel K.

Note that we can do the same for classification, fitting a locally weighted logistic regression:

LOWESS is typically used for:

• Fitting a line to a scatter plot or time plot where noisy data values, sparse data points or weak
interrelationships interfere with your ability to see a line of best fit.
• Linear regression where least squares fitting don’t create a line of good fit or is too labor
intensive to use.
• Data exploration and analysis in the social sciences, particularly in elections and voting behavior.

Parametric and Non-Parametric Fitting

LOWESS, and least squares fitting in general, is non-parametric strategies for fitting a smooth curve to
data points. “Parametric” means that the researcher or analyst assumes in advance that the data fits some
type of distribution (i.e. the normal distribution). Because some type of distribution is assumed in
advance, parametric fitting can lead to fitting a smooth curve that misrepresents the data. In those cases,
on-parametric smoothers may be a better choice. Non-parametric smoothers like LOESS try to find a
curve of best fit without assuming the data must fit some distribution shape. In general, both types of
smoothers are used for the same set of data to offset the advantages and disadvantages of each type of
smoother.

Benefits of Non-Parametric Smoothing


• Provides a flexible approach to representing data.
• Ease of use.
• Computations are relatively easy.

Disadvantages of Non-Parametric Smoothing

• Can’t be used to obtain a simple equation for a set of data


• Less well understood than parametric smoothers.
• Requires the analyst to use a little guesswork to obtain a result

Dept of CSE & ISE, EPCET B’LURU Page 75


MACHINE LEARNING LABORATORY- 15CSL76

import pandas as pd
from sklearn.linear_model import LinearRegression
import matplotlib.pyplot as plt

# Assign the dataframe to this variable.


# Load the data
bmi_life_data = pd.read_csv('bmi_and_life_expectancy.csv')
x_values = bmi_life_data[['BMI']]
y_values = bmi_life_data[['Life expectancy']]
# Make and fit the linear regression model
# Fit the model and Assign it to bmi_life_model
bmi_life_model = LinearRegression()
bmi_life_model.fit(x_values, y_values)

# Mak a prediction using the model


# Predict life expectancy for a BMI value of 21.07931
value_for_prediction = 21.07931
laos_life_exp = bmi_life_model.predict(value_for_prediction)

print(laos_life_exp)

# Plot outputs
plt.scatter(x_values, y_values, color='blue')
plt.plot(value_for_prediction, laos_life_exp, color='black',
linewidth=3)

plt.show()

Dept of CSE & ISE, EPCET B’LURU Page 76


MACHINE LEARNING LABORATORY- 15CSL76

OUTPUT

Dept of CSE & ISE, EPCET B’LURU Page 77


MACHINE LEARNING LABORATORY- 15CSL76

VIVA QUESTIONS

1. What is machine learning?


2. Define supervised learning
3. Define unsupervised learning
4. Define semi supervised learning
5. Define reinforcement learning
6. What do you mean by hypotheses
7. What is classification
8. What is clustering
9. Define precision, accuracy and recall
10. Define entropy
11. Define regression
12. How Knn is different from k-means clustering
13. What is concept learning
14. Define specific boundary and general boundary
15. Define target function
16. Define decision tree
17. What is ANN
18. Explain gradient descent approximation
19. State Bayes theorem
20. Define Bayesian belief networks
21. Differentiate hard and soft clustering
22. Define variance
23. What is inductive machine learning
24. Why K nearest neighbour algorithm is lazy learning algorithm
25. Why naïve Bayes is naïve
26. Mention classification algorithms
27. Define pruning
28. Differentiate Clustering and classification
29. Mention clustering algorithms
30. Define Bias

Dept of CSE & ISE, EPCET B’LURU Page 78


MACHINE LEARNING LABORATORY- 15CSL76

NOTES

Dept of CSE & ISE, EPCET B’LURU Page 79


MACHINE LEARNING LABORATORY- 15CSL76

Dept of CSE & ISE, EPCET B’LURU Page 80

You might also like