Beruflich Dokumente
Kultur Dokumente
Outline
Decision tree representation Decision tree learning (ID3) Information gain Overtting Extensions
Example problem
Problem: decide whether to wait for a table at a restaurant, based on the following attributes: 1. Alternate: is there an alternative restaurant nearby? 2. Bar: is there a comfortable bar area to wait in? 3. Fri/Sat: is today Friday or Saturday? 4. Hungry: are we hungry? 5. Patrons: number of people in the restaurant (None, Some, Full) 6. Price: price range ($, $$, $$$) 7. Raining: is it raining outside? 8. Reservation: have we made a reservation? 9. Type: kind of restaurant (French, Italian, Thai, Burger) 10. WaitEstimate: estimated waiting time (0-10, 10-30, 30-60, >60)
3
Example problem
Examples described by attribute values (Boolean, discrete, continuous, etc.) E.g., situations where I will/wont wait for a table:
Example X1 X2 X3 X4 X5 X6 X7 X8 X9 X10 X11 X12 Attributes Target Alt Bar F ri Hun P at P rice Rain Res T ype Est WillWait T F F T Some $$$ F T French 010 T T F F T Full $ F F Thai 3060 F F T F F Some $ F F Burger 010 T T F T T Full $ F F Thai 1030 T T F T F Full $$$ F T French >60 F F T F T Some $$ T T Italian 010 T F T F F None $ T F Burger 010 F F F F T Some $$ T T Thai 010 T F T T F Full $ T F Burger >60 F T T T T Full $$$ F T Italian 1030 F F F F F None $ F F Thai 010 F T T T T Full $ F F Burger 3060 T
Decision trees
One possible representation for hypotheses
Patrons?
None
F
Some
T
Full
WaitEstimate?
>60
F
3060
Alternate?
1030
Hungry?
010
T
No
Reservation?
Yes
Fri/Sat?
No
T
Yes
Alternate?
No
Bar?
Yes
T
No
F
Yes
T
No
T
Yes
Raining?
No
F
Yes
T
No
F
Yes
T
Decision trees
Decision tree representation each internal node tests on an attribute each branch corresponds to an attribute value each leaf node corresponds to a class label When to consider decision trees Produce comprehensible results Decision trees are especially well suited for representing simple rules for classifying instances that are described by discrete attribute values Decision tree learning algorithms are relatively ecient linear in the size of the decision tree and the size of the data set Are often among the rst to be tried on a new data set
Decision trees
We consider discrete valued function (classication) First consider discrete valued attributes (ID3 Ross Quinlan) Then extensions (C4.5 Ross Quinlan) Ross Quinlan, C4.5: Programs for machine learning, 1993. CART, Breiman etal, Classication and Regression Trees, 1984.
Expressiveness
Decision trees can express any function of the input attributes. E.g., for Boolean functions, truth table row path to leaf:
A F F T T B F T F T A xor B F T T F
B A
T
B
F
F
T
T
F
T
T
F
Trivially, there is a consistent decision tree for any training set w/ one path to leaf for each example (unless f nondeterministic in x) but it probably wont generalize to new examples Prefer to nd more compact decision trees Ockhams razor: maximize a combination of consistency and simplicity
Hypothesis spaces
How many distinct decision trees with n Boolean attributes = number of Boolean functions n = number of distinct truth tables with 2n rows = 22 E.g., with 6 Boolean attributes, there are 18,446,744,073,709,551,616 trees
10
Patrons?
None Some Full French Italian
Type?
Thai Burger
11
Base cases: -uniform example classication -empty examples: majority classication at the nodes parent -empty attributes: use a majority vote?
12
Suppose we have p positive and n negative examples in the training set E at the root H ( ) = H (E ) = H ( p/(p + n), n/(p + n) )
A chosen attribute A divides the training set E into subsets E1, . . . , Ev according to their values for A, where A has v distinct values. Let Ei have pi positive and ni negative examples. The conditional entropy H ( |A = ai) = H (Ei) = H ( pi/(pi + ni), ni/(pi + ni) )
13
Information Gain (mutual information) or reduction in entropy from the attribute test: Gain(A) = I (, A) = H (E ) Remainder (A)
Choose the attribute with the largest Information Gain = choose the attribute that minimizes the remaining information needed
14
Information Gain
E.g., for 12 restaurant examples, p = n = 6 so we need H (E ) = H (6/12, 6/12) = 1 bit
Patrons?
None Some Full French Italian
Type?
Thai Burger
Remaider (P atrons) = 2/12H (0, 1)+4/12H (1, 0)+6/12H (2/6, 4/6) = 0.459 bits Remaider (T ype) = 2/12H (1/2, 1/2) + 2/12H (1/2, 1/2) + 4/12H (2/4, 2/4) + 4/12H (2/4, 2/4) = 1 bit Patrons has the highest IG of all attributes and so is chosen by the DTL algorithm as the root
15
Example contd.
Decision tree learned from the 12 examples:
Patrons?
None
F
Some
T
Full
Hungry?
Yes
Type?
No
F
French
T
Italian
F
Thai
Fri/Sat?
Burger
T
No
F
Yes
T
16
17
18
Avoiding Overtting
How can we avoid overtting? stop growing earlier - Stop when further split fails to yield statistically signicant information gain grow full tree, then prune - more successful in practice
19
Reduced-Error Pruning
Split data into training and validation set Do until further pruning is harmful: 1. Evaluate impact on validation set of pruning each possible node 2. Greedily remove the one that most improves validation set accuracy Pruning a decision node consists of - removing the sub tree rooted at that node, - making it a leaf node, and - assigning it the most common label at that node The validation set has to be large enough not desirable when data set is small
20
Rule Post-Pruning
Outlook
Sunny Humidity
Overcast
Rain Wind
Yes
High No
Normal Yes
Strong No
Weak Yes
Convert tree to equivalent set of rules IF THEN IF THEN ... (Outlook = Sunny ) (Humidity = High) P layT ennis = N o (Outlook = Sunny ) (Humidity = N ormal) P layT ennis = Y es
21
Rule Post-Pruning
1. Convert tree to equivalent set of rules 2. Prune each rule independently of others by removing any preconditions that result in improving its estimated accuracy 3. Sort nal rules in order of lowest to highest error for classifying new instances Perhaps most frequently used method (e.g., C4.5)
22
Temperature: 40 48 60 72 80 90 PlayTennis: No No Yes Yes Yes No Preprocess: Discretize data Or dynamically dening new discrete-valued attributes that partition the continuous attribute value into a discrete set of intervals Find the best split point A > c that gives the highest information gain Sort instances by value of numeric attribute under consideration; Identifying adjacent examples that dier in their target classication to generate candidate splits T > (48 + 60)/2?, T > (80 + 90)/2? Split into multiple intervals?
23
SplitInf ormation(E, A)
i=1
where Ei is subset of examples E for which A has value ai A number of alternative selection measures - small impact on nal accuracy after pruning
24
25
26
27
28
29