KEMBAR78
Decision Trees | PDF | Data Analysis | Applied Mathematics
0% found this document useful (0 votes)
67 views26 pages

Decision Trees

Uploaded by

khanakggupta.123
Copyright
© © All Rights Reserved
We take content rights seriously. If you suspect this is your content, claim it here.
Available Formats
Download as PDF, TXT or read online on Scribd
0% found this document useful (0 votes)
67 views26 pages

Decision Trees

Uploaded by

khanakggupta.123
Copyright
© © All Rights Reserved
We take content rights seriously. If you suspect this is your content, claim it here.
Available Formats
Download as PDF, TXT or read online on Scribd
You are on page 1/ 26

Decision Trees

A decision tree is a non-parametric supervised


learning algorithm, which is utilized for both
classification and regression tasks. It has a
hierarchical, tree structure, which consists of a
root node, branches, internal nodes and leaf nodes.

Decision Trees 1
Outline
• Decision Tree Representations
– ID3 and C4.5 learning algorithms (Quinlan
1986)
– CART learning algorithm (Breiman et al. 1985)
• Entropy, Information Gain
• Overfitting

Decision Trees 2
Training Data Example: Goal is to Predict
When This Player Will Play Tennis?

Intro AI Decision Trees 3


Intro AI Decision Trees 4
Intro AI Decision Trees 5
Intro AI Decision Trees 6
Intro AI Decision Trees 7
Learning Algorithm for Decision
Trees
x = (x1 ,..., xd )
S = {(x1 , y1 ),..., (x N , yN )}
x j , y Î {0,1}

What happens if features are not binary? What about regression?


Intro AI Decision Trees 8
Choosing the Best Attribute
A1 and A2 are “attributes” (i.e. features or inputs).

Number +
and – examples
before and after
a split.

- Many different frameworks for choosing BEST have been


proposed!
- We will look at Entropy Gain.
Intro AI Decision Trees 9
Entropy

Intro AI Decision Trees 10


Entropy is like a measure of impurity…

Intro AI Decision Trees 11


Entropy

Intro AI Decision Trees 12


Intro AI Decision Trees 13
Information Gain

Intro AI Decision Trees 14


Intro AI Decision Trees 15
Intro AI Decision Trees 16
Intro AI Decision Trees 17
Training Example

Intro AI Decision Trees 18


Selecting the Next Attribute

Intro AI Decision Trees 19


Intro AI Decision Trees 20
Non-Boolean Features
• Features with multiple discrete values
– Multi-way splits
– Test for one value versus the rest
– Group values into disjoint sets
• Real-valued features
– Use thresholds
• Regression
– Splits based on mean squared error metric
Intro AI Decision Trees 21
Hypothesis Space Search

You do not get the globally


optimal tree!
- Search space is exponential.

Intro AI Decision Trees 22


Overfitting

Intro AI Decision Trees 23


Overfitting in Decision Trees

Intro AI Decision Trees 24


Validation Data is Used to Control
Overfitting
• Prune tree to reduce error on validation set

Intro AI Decision Trees 25


Homework
• Which feature will be at the root node of the decision tree
trained for the following data? In other words which
attribute makes a person most attractive?
Height Hair Eyes Attractive?
small blonde brown No
tall dark brown No
tall blonde blue Yes
tall dark Blue No
small dark Blue No
tall red Blue Yes
tall blonde brown No
small blonde blue Yes

Intro AI Decision Trees 26

You might also like