K Nearest Neighbor Classification Methods
-
Upload
davis-hebert -
Category
Documents
-
view
49 -
download
0
description
Transcript of K Nearest Neighbor Classification Methods
![Page 1: K Nearest Neighbor Classification Methods](https://reader036.fdocuments.us/reader036/viewer/2022082404/56812d5b550346895d92652c/html5/thumbnails/1.jpg)
K Nearest Neighbor Classification Methods
Qiang Yang
![Page 2: K Nearest Neighbor Classification Methods](https://reader036.fdocuments.us/reader036/viewer/2022082404/56812d5b550346895d92652c/html5/thumbnails/2.jpg)
The K-Nearest Neighbor Method
• Used for prediction/classification
• Given input x, (e.g., <sunny, normal, ..?>
• #neighbors = K (e.g., k=3)– Often a parameter to be determined
• The form of the distance function
– K neighbors in training data to the input data x:
• Break ties arbitrarily
• All k neighbors will vote: majority wins
![Page 3: K Nearest Neighbor Classification Methods](https://reader036.fdocuments.us/reader036/viewer/2022082404/56812d5b550346895d92652c/html5/thumbnails/3.jpg)
How to decide the distance?Try 3-NN on this data: assume distance function =‘ # of
different attributes.’
Outlook Temperature Humidity Windy Play
sunny hot high FALSE no
sunny hot high TRUE no
overcast hot high FALSE yes
rainy mild high FALSE yes
rainy cool normal FALSE yes
rainy cool normal TRUE no
overcast cool normal TRUE yes
sunny mild high FALSE no
sunny cool normal FALSE yes
rainy mild normal FALSE yes
sunny mild normal TRUE yes
overcast mild high TRUE yes
overcast hot normal FALSE yes
rainy mild high TRUE ?testing
![Page 4: K Nearest Neighbor Classification Methods](https://reader036.fdocuments.us/reader036/viewer/2022082404/56812d5b550346895d92652c/html5/thumbnails/4.jpg)
Why important?• Often a baseline
– Must beat this one to claim innovation• Forms of KNN
– Weighted KNN– “K” is a variable:
• Often we experiment with different values of K=1, 3, 5, to find out the optimal one
– Document similarity• Cosine
– Case based reasoning• Edited data base• Sometimes, the accuracy (CBR)/accuracy (KNN) can be better than
100%: why?– Image understanding
• Manifold learning• Distance metric
![Page 5: K Nearest Neighbor Classification Methods](https://reader036.fdocuments.us/reader036/viewer/2022082404/56812d5b550346895d92652c/html5/thumbnails/5.jpg)
K-NN can be misleading
• Consider applying K-NN on the training data– What is the accuracy? 100%– Why? Distance to self is zero– What should we do in testing?
• Use new data for testing, rather than training data.