Evaluation of Explosion-Resistant Seals, Stoppings, and Overcast
Chapter 6 Decision Trees. 2 An Example Outlook overcast humiditywindy highnormal false true sunny...
-
Upload
bridget-patricia-richard -
Category
Documents
-
view
216 -
download
2
Transcript of Chapter 6 Decision Trees. 2 An Example Outlook overcast humiditywindy highnormal false true sunny...
![Page 1: Chapter 6 Decision Trees. 2 An Example Outlook overcast humiditywindy highnormal false true sunny rain NNPP P overcast.](https://reader037.fdocuments.us/reader037/viewer/2022110323/56649d8e5503460f94a76e2a/html5/thumbnails/1.jpg)
Chapter 6Decision Trees
![Page 2: Chapter 6 Decision Trees. 2 An Example Outlook overcast humiditywindy highnormal false true sunny rain NNPP P overcast.](https://reader037.fdocuments.us/reader037/viewer/2022110323/56649d8e5503460f94a76e2a/html5/thumbnails/2.jpg)
2
An ExampleOutlook Tempreature Humidity Windy Classsunny hot high false Nsunny hot high true Novercast hot high false Prain mild high false Prain cool normal false Prain cool normal true Novercast cool normal true Psunny mild high false Nsunny cool normal false Prain mild normal false Psunny mild normal true Povercast mild high true Povercast hot normal false Prain mild high true N
Outlook
overcast
humidity windy
high normal falsetrue
sunny rain
N NP P
P
overcast
![Page 3: Chapter 6 Decision Trees. 2 An Example Outlook overcast humiditywindy highnormal false true sunny rain NNPP P overcast.](https://reader037.fdocuments.us/reader037/viewer/2022110323/56649d8e5503460f94a76e2a/html5/thumbnails/3.jpg)
3
Another Example - Grades
Percent >= 90%?
Yes Grade = A
No 89% >= Percent >= 80%?
Yes Grade = B
No 79% >= Percent >= 70%?
Yes Grade = C
No Etc...
![Page 4: Chapter 6 Decision Trees. 2 An Example Outlook overcast humiditywindy highnormal false true sunny rain NNPP P overcast.](https://reader037.fdocuments.us/reader037/viewer/2022110323/56649d8e5503460f94a76e2a/html5/thumbnails/4.jpg)
4
Yet Another Example1 of 2
![Page 5: Chapter 6 Decision Trees. 2 An Example Outlook overcast humiditywindy highnormal false true sunny rain NNPP P overcast.](https://reader037.fdocuments.us/reader037/viewer/2022110323/56649d8e5503460f94a76e2a/html5/thumbnails/5.jpg)
5
Yet Another Example
• English Rules (for example):
2 of 2
If tear production rate = reduced then recommendation = none.If age = young and astigmatic = no and tear production rate = normal then recommendation = softIf age = pre-presbyopic and astigmatic = no and tear production rate = normal then recommendation = softIf age = presbyopic and spectacle prescription = myope and astigmatic = no then recommendation = noneIf spectacle prescription = hypermetrope and astigmatic = no and tear production rate = normal then recommendation = softIf spectacle prescription = myope and astigmatic = yes and tear production rate = normal then recommendation = hardIf age = young and astigmatic = yes and tear production rate =
normal then recommendation = hardIf age = pre-presbyopic and spectacle prescription = hypermetrope and astigmatic = yes then recommendation = noneIf age = presbyopic and spectacle prescription = hypermetrope and astigmatic = yes then recommendation = none
![Page 6: Chapter 6 Decision Trees. 2 An Example Outlook overcast humiditywindy highnormal false true sunny rain NNPP P overcast.](https://reader037.fdocuments.us/reader037/viewer/2022110323/56649d8e5503460f94a76e2a/html5/thumbnails/6.jpg)
6
Decision Tree Template
• Drawn top-to-bottom or left-
to-right
• Top (or left-most) node =
Root Node
• Descendent node(s) = Child
Node(s)
• Bottom (or right-most)
node(s) = Leaf Node(s)
• Unique path from root to
each leaf = Rule
Root
Child Child Leaf
LeafChild
Leaf
![Page 7: Chapter 6 Decision Trees. 2 An Example Outlook overcast humiditywindy highnormal false true sunny rain NNPP P overcast.](https://reader037.fdocuments.us/reader037/viewer/2022110323/56649d8e5503460f94a76e2a/html5/thumbnails/7.jpg)
7
Introduction
• Decision Trees– Powerful/popular for classification & prediction– Represent rules
• Rules can be expressed in English– IF Age <=43 & Sex = Male
& Credit Card Insurance = NoTHEN Life Insurance Promotion = No
• Rules can be expressed using SQL for query– Useful to explore data to gain insight into relationships
of a large number of candidate input variables to a target (output) variable
• You use mental decision trees often!• Game: “I’m thinking of…” “Is it …?”
![Page 8: Chapter 6 Decision Trees. 2 An Example Outlook overcast humiditywindy highnormal false true sunny rain NNPP P overcast.](https://reader037.fdocuments.us/reader037/viewer/2022110323/56649d8e5503460f94a76e2a/html5/thumbnails/8.jpg)
8
Decision Tree – What is it?
• A structure that can be used to divide up a large
collection of records into successively smaller
sets of records by applying a sequence of simple
decision rules
• A decision tree model consists of a set of rules
for dividing a large heterogeneous population
into smaller, more homogeneous groups with
respect to a particular target variable
![Page 9: Chapter 6 Decision Trees. 2 An Example Outlook overcast humiditywindy highnormal false true sunny rain NNPP P overcast.](https://reader037.fdocuments.us/reader037/viewer/2022110323/56649d8e5503460f94a76e2a/html5/thumbnails/9.jpg)
9
Decision Tree Types
• Binary trees – only two choices in each split. Can be non-uniform (uneven) in depth
• N-way trees or ternary trees – three or more choices in at least one of its splits (3-way, 4-way, etc.)
![Page 10: Chapter 6 Decision Trees. 2 An Example Outlook overcast humiditywindy highnormal false true sunny rain NNPP P overcast.](https://reader037.fdocuments.us/reader037/viewer/2022110323/56649d8e5503460f94a76e2a/html5/thumbnails/10.jpg)
10
Scoring
• Often it is useful to show the proportion of the data in each of the desired classes
• Clarify Fig 6.2
![Page 11: Chapter 6 Decision Trees. 2 An Example Outlook overcast humiditywindy highnormal false true sunny rain NNPP P overcast.](https://reader037.fdocuments.us/reader037/viewer/2022110323/56649d8e5503460f94a76e2a/html5/thumbnails/11.jpg)
11
Decision Tree Splits (Growth)
• The best split at root or child nodes is defined as
one that does the best job of separating the data
into groups where a single class predominates in
each group
– Example: US Population data input categorical
variables/attributes include:• Zip code
• Gender
• Age
– Split the above according to the above “best split” rule
![Page 12: Chapter 6 Decision Trees. 2 An Example Outlook overcast humiditywindy highnormal false true sunny rain NNPP P overcast.](https://reader037.fdocuments.us/reader037/viewer/2022110323/56649d8e5503460f94a76e2a/html5/thumbnails/12.jpg)
12
Example: Good & Poor Splits
Good Split
![Page 13: Chapter 6 Decision Trees. 2 An Example Outlook overcast humiditywindy highnormal false true sunny rain NNPP P overcast.](https://reader037.fdocuments.us/reader037/viewer/2022110323/56649d8e5503460f94a76e2a/html5/thumbnails/13.jpg)
13
Split Criteria
• The best split is defined as one that does the best job of separating the data into groups where a single class predominates in each group
• Measure used to evaluate a potential split is purity– The best split is one that increases purity of
the sub-sets by the greatest amount– A good split also creates nodes of similar size
or at least does not create very small nodes
![Page 14: Chapter 6 Decision Trees. 2 An Example Outlook overcast humiditywindy highnormal false true sunny rain NNPP P overcast.](https://reader037.fdocuments.us/reader037/viewer/2022110323/56649d8e5503460f94a76e2a/html5/thumbnails/14.jpg)
14
Tests for Choosing Best Split
• Purity (Diversity) Measures:
– Gini (population diversity)
– Entropy (information gain)
– Information Gain Ratio
– Chi-square Test
We will only explore Gini in class
![Page 15: Chapter 6 Decision Trees. 2 An Example Outlook overcast humiditywindy highnormal false true sunny rain NNPP P overcast.](https://reader037.fdocuments.us/reader037/viewer/2022110323/56649d8e5503460f94a76e2a/html5/thumbnails/15.jpg)
15
Gini (Population Diversity)
• The Gini measure of a node is the sum of the squares of the proportions of the classes.
Root Node: 0.5^2 + 0.5^2 = 0.5 (even balance)
Leaf Nodes: 0.1^2 + 0.9^2 = 0.82 (close to pure)
![Page 16: Chapter 6 Decision Trees. 2 An Example Outlook overcast humiditywindy highnormal false true sunny rain NNPP P overcast.](https://reader037.fdocuments.us/reader037/viewer/2022110323/56649d8e5503460f94a76e2a/html5/thumbnails/16.jpg)
16
Pruning
• Decision Trees can often be simplified or pruned:– CART– C5– Stability-based
We will not cover these in detail
![Page 17: Chapter 6 Decision Trees. 2 An Example Outlook overcast humiditywindy highnormal false true sunny rain NNPP P overcast.](https://reader037.fdocuments.us/reader037/viewer/2022110323/56649d8e5503460f94a76e2a/html5/thumbnails/17.jpg)
17
Decision Tree Advantages
1. Easy to understand
2. Map nicely to a set of business rules
3. Applied to real problems
4. Make no prior assumptions about the data
5. Able to process both numerical and
categorical data
![Page 18: Chapter 6 Decision Trees. 2 An Example Outlook overcast humiditywindy highnormal false true sunny rain NNPP P overcast.](https://reader037.fdocuments.us/reader037/viewer/2022110323/56649d8e5503460f94a76e2a/html5/thumbnails/18.jpg)
18
Decision Tree Disadvantages
1. Output attribute must be categorical
2. Limited to one output attribute
3. Decision tree algorithms are unstable
4. Trees created from numeric datasets can
be complex
![Page 19: Chapter 6 Decision Trees. 2 An Example Outlook overcast humiditywindy highnormal false true sunny rain NNPP P overcast.](https://reader037.fdocuments.us/reader037/viewer/2022110323/56649d8e5503460f94a76e2a/html5/thumbnails/19.jpg)
19
Alternative Representations
• Box Diagram
• Tree Ring Diagram
• Decision Table
• Supplementary Material
![Page 20: Chapter 6 Decision Trees. 2 An Example Outlook overcast humiditywindy highnormal false true sunny rain NNPP P overcast.](https://reader037.fdocuments.us/reader037/viewer/2022110323/56649d8e5503460f94a76e2a/html5/thumbnails/20.jpg)
20
End of Chapter 6