Unformatted text preview: Determine a decision tree based upon choosing the largest information gain using the entropy measure and splitting attribute C as in (c). Justify your choices. 2. Suppose we have the training set of data given in the notes to classify vertebrates. Use a general to speci±c approach to obtain a rule to classify ±sh, i.e., start with the rule ( ) ⇒ ±sh and determine a rule by considering the coverage of an attribute and its accuracy. Justify why you made your choice. You do not have to compute coverage or accuracy if no outcomes are “±sh”. Make a table at each step of the test, how many records it covers, how many are classi±ed as ±sh, the accuracy and coverage. 1...
View Full Document
- Spring '11
- Data Mining, #, largest information gain