BSAN 400 Introduction to Machine...

22

Transcript of BSAN 400 Introduction to Machine...

Page 1: BSAN 400 Introduction to Machine Learninghomepages.uc.edu/~lis6/Teaching/ML19Spring/Lecture/3_LogisticReg.pdfBSAN 400 Introduction to Machine Learning Lecture 4. Logistic Regression

BSAN 400 Introduction to Machine Learning

Lecture 4. Logistic Regression and Classi�cations1

Shaobo Li

University of Kansas

1Partially based on Hastie, et al. (2009) ESL, and James, et al. (2013) ISLR

Intro to ML Lecture 4. Logistic Regression 1 / 21

Page 2: BSAN 400 Introduction to Machine Learninghomepages.uc.edu/~lis6/Teaching/ML19Spring/Lecture/3_LogisticReg.pdfBSAN 400 Introduction to Machine Learning Lecture 4. Logistic Regression

From Continuous to Categorical Outcome

The response variable, Y , is categorical

Examples:- Banking: default vs. nondefault- Medical: disease positive vs. negative- Computer vision: sign recognition by self-driving cars- Many others...

Intro to ML Lecture 4. Logistic Regression 2 / 21

Page 3: BSAN 400 Introduction to Machine Learninghomepages.uc.edu/~lis6/Teaching/ML19Spring/Lecture/3_LogisticReg.pdfBSAN 400 Introduction to Machine Learning Lecture 4. Logistic Regression

Classi�cation

Denote C(X ) as a classi�er

Most DM algorithms estimate the probability that X belongs toeach class

Based on speci�c decision rules, classi�cation can be produced

Example: The model prediction tells that the probability ofdefault is 0.2, then

Threshold <0.1 >0.1

Class Nondefault Default

Intro to ML Lecture 4. Logistic Regression 3 / 21

Page 4: BSAN 400 Introduction to Machine Learninghomepages.uc.edu/~lis6/Teaching/ML19Spring/Lecture/3_LogisticReg.pdfBSAN 400 Introduction to Machine Learning Lecture 4. Logistic Regression

Classi�cation Methods

K-Nearest Neighbor

Logistic regression

Classi�cation tree

Discriminant analysis

Support vector machine

Neural networks

Deep learning

...

Is clustering a classi�cation model?

Intro to ML Lecture 4. Logistic Regression 4 / 21

Page 5: BSAN 400 Introduction to Machine Learninghomepages.uc.edu/~lis6/Teaching/ML19Spring/Lecture/3_LogisticReg.pdfBSAN 400 Introduction to Machine Learning Lecture 4. Logistic Regression

Why Not Linear Regression

Example: default prediction- Default (Y = 1) vs. Nondefault (Y = 0)- X1: credit card balance level, X2: income level

Suppose the estimated linear regression is

Y = −1.5+ 2X1 − X2

What is the predicted value if a person's balance level is 1 andincome level is 3?

How to interpret this value?

Intro to ML Lecture 4. Logistic Regression 5 / 21

Page 6: BSAN 400 Introduction to Machine Learninghomepages.uc.edu/~lis6/Teaching/ML19Spring/Lecture/3_LogisticReg.pdfBSAN 400 Introduction to Machine Learning Lecture 4. Logistic Regression

An Illustration

Intro to ML Lecture 4. Logistic Regression 6 / 21

Page 7: BSAN 400 Introduction to Machine Learninghomepages.uc.edu/~lis6/Teaching/ML19Spring/Lecture/3_LogisticReg.pdfBSAN 400 Introduction to Machine Learning Lecture 4. Logistic Regression

An Illustration

Intro to ML Lecture 4. Logistic Regression 6 / 21

Page 8: BSAN 400 Introduction to Machine Learninghomepages.uc.edu/~lis6/Teaching/ML19Spring/Lecture/3_LogisticReg.pdfBSAN 400 Introduction to Machine Learning Lecture 4. Logistic Regression

Logistic Regression

Generalized linear model

Logistic model for binary response

P(yi = 1|xi ) =1

1+ exp(−β0 −∑p

j=1 βjxij)

Sigmoid function

Function of X

Outcome is predicted probability of event

More than two classes: multinomial logistic model

Intro to ML Lecture 4. Logistic Regression 7 / 21

Page 9: BSAN 400 Introduction to Machine Learninghomepages.uc.edu/~lis6/Teaching/ML19Spring/Lecture/3_LogisticReg.pdfBSAN 400 Introduction to Machine Learning Lecture 4. Logistic Regression

Loss Function (in machine learning)

Recall LS

LOLS(β) =n∑

i=1

yi − β0 −p∑

j=1

βjxij

2

For logistic regression, we have a di�erent loss

Llogit(β) =n∑

i=1

−2 [yi log pi + (1− yi ) log(1− pi )]

where pi =1

1+exp(−β0−∑p

j=1 βjxij )

Intro to ML Lecture 4. Logistic Regression 8 / 21

Page 10: BSAN 400 Introduction to Machine Learninghomepages.uc.edu/~lis6/Teaching/ML19Spring/Lecture/3_LogisticReg.pdfBSAN 400 Introduction to Machine Learning Lecture 4. Logistic Regression

Maximum Likelihood Estimation (in statistics)

Maximum likelihood estimation

yi |xi ∼ Ber(pi (xi ))

Likelihood function of yi |xi :

L(yi ; xi ,β) = pyii × (1− pi )1−yi

=

(exp(xTi β)

1+ exp(xTi β)

)yi

×(

1

1+ exp(xTi β)

)1−yi

By simple algebra, total log-likelihood is (show in exercise)

`(β) =n∑

i=1

{yix

Ti β − log

(1+ exp(xTi β)

)}

Intro to ML Lecture 4. Logistic Regression 9 / 21

Page 11: BSAN 400 Introduction to Machine Learninghomepages.uc.edu/~lis6/Teaching/ML19Spring/Lecture/3_LogisticReg.pdfBSAN 400 Introduction to Machine Learning Lecture 4. Logistic Regression

How to solve for β?

No analytical solution like OLS

Solve numerically (univariate example)

Gradient descent

β(n+1) = β(n) − α ∗ L′(β(n))

where α is called learning rate.Newton's method (a very good tutorial)

β(n+1) = β(n) − L′(β(n))

L′′(β(n))

- We call derivative for univariate: f ′(x), f ′′(x)- We call gradient for multivariate: ∇f (x), ∇2f (x)

Intro to ML Lecture 4. Logistic Regression 10 / 21

Page 12: BSAN 400 Introduction to Machine Learninghomepages.uc.edu/~lis6/Teaching/ML19Spring/Lecture/3_LogisticReg.pdfBSAN 400 Introduction to Machine Learning Lecture 4. Logistic Regression

Odds and Interpretation of β

Logistic model is also called log odds model

Odds: ratio of probabilities:

Odds(X ) = P(Y = 1|X )/P(Y = 0|X )

Logit link (logit transformation)

logit(P) = log

(P

1− P

)= β0 + β1X1 + . . .+ βpXp

By simple algebra, given all X 's are �xed except Xj

βj = log

(Odds(Xj + 1)

Odds(Xj)

)which is log of odds ratio.

Intro to ML Lecture 4. Logistic Regression 11 / 21

Page 13: BSAN 400 Introduction to Machine Learninghomepages.uc.edu/~lis6/Teaching/ML19Spring/Lecture/3_LogisticReg.pdfBSAN 400 Introduction to Machine Learning Lecture 4. Logistic Regression

Multinomial Logit Model

Response Y = 1, 2, ...,K , K classes

Given predictors xi

log

(P(yi = 2)

P(Yi = 1)

)= βT

2 xi

log

(P(yi = 3)

P(Yi = 1)

)= βT

3 xi

...

log

(P(yi = K )

P(Yi = 1)

)= βT

Kxi

The �rst class �1� is the reference

There are (K − 1)× (p + 1) coe�cients need to be estimated.

Intro to ML Lecture 4. Logistic Regression 12 / 21

Page 14: BSAN 400 Introduction to Machine Learninghomepages.uc.edu/~lis6/Teaching/ML19Spring/Lecture/3_LogisticReg.pdfBSAN 400 Introduction to Machine Learning Lecture 4. Logistic Regression

Prediction � From Probability to Class

Direct outcome of model: probability

Next step: classi�cation

Need decision rule (cut-o� probability � p-cut)

Not unique

Intro to ML Lecture 4. Logistic Regression 13 / 21

Page 15: BSAN 400 Introduction to Machine Learninghomepages.uc.edu/~lis6/Teaching/ML19Spring/Lecture/3_LogisticReg.pdfBSAN 400 Introduction to Machine Learning Lecture 4. Logistic Regression

Confusion Matrix

Classi�cation table based on a speci�c cut-o� probability

Used for model assessment

Pred=1 Pred=0

True=1 True Positive (TP) False Negative (FN)

True=0 False Positive (FP) True Negative (TN)

FP: type I error; FN: type II error

Di�erent p-cut results in di�erent confusion matrix

Try to understand this table instead of memorizing!

Intro to ML Lecture 4. Logistic Regression 14 / 21

Page 16: BSAN 400 Introduction to Machine Learninghomepages.uc.edu/~lis6/Teaching/ML19Spring/Lecture/3_LogisticReg.pdfBSAN 400 Introduction to Machine Learning Lecture 4. Logistic Regression

Some Useful Measures

Misclassi�cation rate (MR) = FP+FNTotal

True positive rate (TPR) = TPTP+FN

: Sensitivity or Recall

True negative rate (TNR) = TNFP+TN

: Speci�city

False positive rate (FPR) = FPFP+TN

: 1 � Speci�city

True negative rate (FNR) = FNTP+FN

: 1 � Sensitivity

Positive predictive rate (PPR) = TPTP+FP

: Precision

False discovery rate (FDR) = 1 � Precision

Intro to ML Lecture 4. Logistic Regression 15 / 21

Page 17: BSAN 400 Introduction to Machine Learninghomepages.uc.edu/~lis6/Teaching/ML19Spring/Lecture/3_LogisticReg.pdfBSAN 400 Introduction to Machine Learning Lecture 4. Logistic Regression

ROC Curve

Receiver Operating Characteristic

Plot of FPR (X) against TPR (Y) at various p-cut values

Overall model assessment (not for a particular decision rule)

Unique for a given model

Area under the curve (AUC): a measure of goodness-of-�t

Intro to ML Lecture 4. Logistic Regression 16 / 21

Page 18: BSAN 400 Introduction to Machine Learninghomepages.uc.edu/~lis6/Teaching/ML19Spring/Lecture/3_LogisticReg.pdfBSAN 400 Introduction to Machine Learning Lecture 4. Logistic Regression

ROC Curve

Intro to ML Lecture 4. Logistic Regression 17 / 21

Page 19: BSAN 400 Introduction to Machine Learninghomepages.uc.edu/~lis6/Teaching/ML19Spring/Lecture/3_LogisticReg.pdfBSAN 400 Introduction to Machine Learning Lecture 4. Logistic Regression

Precision and Recall

More accurate measure for imbalanced data

Widely used in document retrieval

Precision = TPTP+FP

:- fraction of retrieved instances that are relevant

Recall = TPTP+FN

:- fraction of relevant instances that are retrieved

Neither incorporates TN (Y = 1 is of more interest)

F -score: F = 2× precision×recallprecision+recall

More details: see this highly cited paper

Intro to ML Lecture 4. Logistic Regression 18 / 21

Page 20: BSAN 400 Introduction to Machine Learninghomepages.uc.edu/~lis6/Teaching/ML19Spring/Lecture/3_LogisticReg.pdfBSAN 400 Introduction to Machine Learning Lecture 4. Logistic Regression

Precision-Recall Curve

Intro to ML Lecture 4. Logistic Regression 19 / 21

Page 21: BSAN 400 Introduction to Machine Learninghomepages.uc.edu/~lis6/Teaching/ML19Spring/Lecture/3_LogisticReg.pdfBSAN 400 Introduction to Machine Learning Lecture 4. Logistic Regression

Asymmetric Cost

Example: compare following two confusion matrices based ontwo p-cut values

Pred=1 Pred=0

True=1 10 40

True=0 10 440

Pred=1 Pred=0

True=1 40 10

True=0 130 320

Which one is better? In terms of what?

What if this is about loan application- Y = 1: default customer- Default will cost much more than reject a loan application

Intro to ML Lecture 4. Logistic Regression 20 / 21

Page 22: BSAN 400 Introduction to Machine Learninghomepages.uc.edu/~lis6/Teaching/ML19Spring/Lecture/3_LogisticReg.pdfBSAN 400 Introduction to Machine Learning Lecture 4. Logistic Regression

Choice of Decision Threshold (p-cut)

Do NOT simply use 0.5!

In general, we use grid search method to optimize a measure ofclassi�cation accuracy/loss- Cost function (symmetric or asymmetric)- F-score based on precision and recall

Grid search with cross-validation

Intro to ML Lecture 4. Logistic Regression 21 / 21