Professional Documents
Culture Documents
❖ It is one of the most widely used and practical methods for supervised learning.
❖Decision Trees are a non-parametric supervised learning method used for both
❖ classification and regression tasks.
❖ The goal is to create a model that predicts the value of a target variable by
❖ learning simple decision rules inferred from the data features.
❖The decision rules are generally in form of if-then-else statements.
❖ID3 is one of the most common decision tree algorithm. Firstly, It was introduced in 1986
and it is acronym of Iterative Dichotomiser.
❖The ID3 algorithm begins with the original set as the root node.
❖On each iteration of the algorithm, it iterates through every unused attribute of the set
and calculates the entropy or the information gain of that attribute.
❖It then selects the attribute which has the largest information gain value.
❖The set is then split or partitioned by the selected attribute to produce subsets of the
data.
❖Information gain is a statistical property that measures how well a given attribute
separates the training examples according to their target classification.
TCE Online Course - APPLIED DATA SCIENCE WITH PYTHONDecision tree for the concept Play Cricket 7
TCE Online Course - APPLIED DATA SCIENCE WITH PYTHON 8
TCE Online Course - APPLIED DATA SCIENCE WITH PYTHON 9
Entropy of Class Label
Since, the basic version of the ID3 algorithm deal with the case where classification are either positive or negative,
we can define entropy as :
H(Outlook=overcast) = -(4/4)*log2(4/4)-0 = 0