The document outlines 12 assignment questions related to machine learning classification algorithms including decision trees, candidate generation and pruning in decision trees, expressing attribute conditions, Hunt's algorithm, constructing a decision tree on a sample data set, the Apriori algorithm and frequent item set generation, alternative frequent item set generation methods, constructing a hash tree and identifying leaf nodes for a transaction, constructing an FP-tree in the FP-Growth algorithm from a transaction data set, explaining classification problems and approaches, measures for selecting the best split in decision trees, overfitting in models and how to address it, and characteristics of decision tree induction.
The document outlines 12 assignment questions related to machine learning classification algorithms including decision trees, candidate generation and pruning in decision trees, expressing attribute conditions, Hunt's algorithm, constructing a decision tree on a sample data set, the Apriori algorithm and frequent item set generation, alternative frequent item set generation methods, constructing a hash tree and identifying leaf nodes for a transaction, constructing an FP-tree in the FP-Growth algorithm from a transaction data set, explaining classification problems and approaches, measures for selecting the best split in decision trees, overfitting in models and how to address it, and characteristics of decision tree induction.
The document outlines 12 assignment questions related to machine learning classification algorithms including decision trees, candidate generation and pruning in decision trees, expressing attribute conditions, Hunt's algorithm, constructing a decision tree on a sample data set, the Apriori algorithm and frequent item set generation, alternative frequent item set generation methods, constructing a hash tree and identifying leaf nodes for a transaction, constructing an FP-tree in the FP-Growth algorithm from a transaction data set, explaining classification problems and approaches, measures for selecting the best split in decision trees, overfitting in models and how to address it, and characteristics of decision tree induction.
1. Describe the different methods of candidate generation and pruning.
2. Explain various methods for Expressing Attribute Test Conditions. 3. How decision trees are used for classification. Write Hunts algorithm and illustrate it’s working
4. Construct decision tree for the following data set
Age competition type profit
old yes s/w Down old no s/w Down old no h/w Down mid yes s/w Down mid yes h/w Down mid no h/w Up mid no s/w Up new yes s/w Up new no h/w Up new no s/w Up
5. Describe frequent item set generation in Aprior algorithm with example.
6. Describe alternative methods for generating frequent item set. 7. Consider the following set of candidate three-itemsets {1,2,3} {1,2,6} {1,2,7} {1,3,4}{2,3,4}{2,4,6}{2,7,8}{1,7,8}{3,4,8}{3,7,9}{4,7,6}{4,5,9}{6,8,9}{3,6,7} {3,7,6}{456} assume hash function hp=p mod 3 i)construct a hash tree ii)given a transaction that contains items{1,2,5,6,8} which of the hash tree leaf nodes will be visited when finding the candidates of the transaction. 8. Consider the following transaction data set. Describe the construction of FP-Tree in FP-Growth algorithm. TID items 1 {a,b} 2 {b,c,d} 3 {a,c,d,e} 4 {a,d,e} 5 {a,b,c} 6 {a,b,c,d} 7 {a} 8 {a,b,c} 9 {a,b,d} 10 {b,c,e} 9. What is classification. Explain the general approach for solving a classification problem with an example. 10. Explain various measures for selecting the best split with an example 11. Explain Model Over fitting. What are the reasons for overfitting? How to address overfitting problems 12. List characteristics of decision tree induction.