Linear Regression and least square error solution.

Post on 31-Mar-2015

237 views 2 download

Tags:

Transcript of Linear Regression and least square error solution.

Linear Regression and least square error solution

What is “Linear”?

• Remember this:• Y=mX+B?

B

m

Simple linear regression

The linear regression model:

Love of Math = 5 + .01*math SAT score

intercept

slope

P=.22; not significant

Prediction

If you know something about X, this knowledge helps you predict something about Y. (Sound familiar?…sound like conditional probabilities?)

EXAMPLE

• The distribution of baby weights at Stanford ~ N(3400, 360000)

Your “Best guess” at a random baby’s weight, given no information about the baby, is what?

3400 grams

But, what if you have relevant information? Can you make a better guess?

Predictor variable

• X=gestation time

• Assume that babies that gestate for longer are born heavier, all other things being equal.

• Pretend (at least for the purposes of this example) that this relationship is linear.

• Example: suppose a one-week increase in gestation, on average, leads to a 100-gram increase in birth-weight

Y depends on X

Y=birth- weight

(g)

X=gestation time (weeks)

Best fit line is chosen such that the sum of the squared (why squared?) distances of the points (Yi’s) from the line is minimized:

Or mathematically… (remember max and mins from calculus)…

Derivative[(Yi-(mx+b))2]=0

But…Note that not every Y-value (Yi) sits on the line. There’s

variability.

Y=baby weights

(g)

X=gestation times (weeks)

20 30 40

Y/X=40 weeks ~ N(4000, 2)

Y/X=30 weeks ~ N(3000, 2)

Y/X=20 weeks ~ N(2000, 2)

Mean values fall on the line

• E(Y/X=40 weeks)=4000• E(Y/X=30 weeks)=3000• E(Y/X=20 weeks)=2000

E(Y/X)= Y/X = 100 grams/week*X weeks

Linear Regression Model

Y’s are modeled…

Yi= 100*X + random errori

Follows a normal distribution

Fixed – exactly on the line

Assumptions (or the fine print)

• Linear regression assumes that… – 1. The relationship between X and Y is linear– 2. Y is distributed normally at each value of X– 3. The variance of Y at every value of X is the same

(homogeneity of variances)

• Why? The math requires it—the mathematical process is called “least squares” because it fits the regression line by minimizing the squared errors from the line (mathematically easy, but not general—relies on above assumptions).

Non-homogenous variance

Y=birth-weight

(100g)

X=gestation time (weeks)

Least squares estimation

** Least Squares EstimationA little calculus….What are we trying to estimate? β, the slope, from

What’s the constraint? We are trying to minimize the squared distance (hence the “least squares”) between the observations themselves and the predicted values , or (also called the “residuals”, or left-over unexplained variability)

Differencei = yi – (βx + α) Differencei2 = (yi – (βx + α)) 2

Find the β that gives the minimum sum of the squared differences. How do you maximize a function? Take the derivative; set it equal to zero; and solve. Typical max/min problem from calculus….

From here takes a little math trickery to solve for β…

...0))((2

)))(((2))((

1

2

11

2

n

iiiii

n

iiii

n

iii

xxxy

xxyxyd

d

Residual

Residual = observed value – predicted value

At 33.5 weeks gestation, predicted baby weight is 3350 grams

33.5 weeks

This baby was actually 3380 grams.

His residual is +30 grams:

3350 grams

Y=baby weights

(g)

X=gestation times (weeks)

20 30 40

The standard error of Y given X is the average variability around the regression line at any given value of X. It is assumed to be equal at all values of X.

Sy/x

Sy/x

Sy/x

Sy/x

Sy/x

Sy/x

Residual Analysis: check assumptions

• The residual for observation i, ei, is the difference between its observed and predicted value

• Check the assumptions of regression by examining the residuals– Examine for linearity assumption– Examine for constant variance for all levels of X (homoscedasticity) – Evaluate normal distribution assumption– Evaluate independence assumption

• Graphical Analysis of Residuals– Can plot residuals vs. X

iii YYe ˆ

Residual Analysis for Linearity

Not Linear Linear

x

resi

dua

ls

x

Y

x

Y

x

resi

dua

ls

Slide from: Statistics for Managers Using Microsoft® Excel 4th Edition, 2004 Prentice-Hall

Residual Analysis for Homoscedasticity

Non-constant variance Constant variance

x x

Y

x x

Y

resi

dua

ls

resi

dua

ls

Slide from: Statistics for Managers Using Microsoft® Excel 4th Edition, 2004 Prentice-Hall

Residual Analysis for Independence

Not IndependentIndependent

X

Xresi

dua

ls

resi

dua

ls

X

resi

dua

ls

Slide from: Statistics for Managers Using Microsoft® Excel 4th Edition, 2004 Prentice-Hall

Other types of multivariate regression

Multiple linear regression is for normally distributed outcomes

Logistic regression is for binary outcomes

Cox proportional hazards regression is used when time-to-event is the outcome

Principal Component Analysis (PCA)

• Given: n d-dimensional points x1, . . . , xn• Goal: find the “right” features from the data

Zero-D Representation

• Task: find x0 to “represent” x1, . . . , xn• Criterion: find x0 such that the sum of the

squared distances between x0 and the various xk is as small as possible

• the “best” zero-dimensional representation of the data set is the sample mean

One-D Representation

• Consider: represent the set of points by a line through m

• x = m+ ae, e: unit vector along the line

Cont’d

Finding eigenvector problem

Geometrical Interpretation

Finding least square error solution

• Finding the direction such that the least square errors is minimized

• Solution: Eigenvector with smallest eigenvalue

Minimize

Solving big matrix systems

• Ax=b• You can use Matlab’s \

– But not very scalable

• There is also sparse matrix library in C\C++, e.g. TAUCS, that provides routine for solving this sparse linear system

• Good News !You can use existing library to avoid the ``trouble’’ implementation of linear equation solver

• But, you need to understand what is happening within the linear solver

32

Conjugate gradient

• “The Conjugate Gradient Method is the most prominent iterative method for solving sparse systems of linear equations. Unfortunately, many textbook treatments of the topic are written with neither illustrations nor intuition, and their victims can be found to this day babbling senselessly in the corners of dusty libraries. For this reason, a deep, geometric understanding of the method has been reserved for the elite brilliant few who have painstakingly decoded the mumblings of their forebears. Nevertheless, the Conjugate Gradient Method is a composite of simple, elegant ideas that almost anyone can understand. Of course, a reader as intelligent as yourself will learn them almost effortlessly.”

33

Ax=b

• A is square, symmetric and positive-definite• When the A is dense, you’re stuck, use backsubstitution• When A is sparse, iterative techniques (such as

Conjugate Gradient) are faster and more memory efficient

• Simple example:

(Yeah yeah, it’s not sparse)

34

Turn Ax=b into a minimization problem• Minimization is more logical to analyze iteration (gradient ascent/descent)• Quadratic form

– c can be ignored because we want to minimize• Intuition:

– the solution of a linear system is always the intersection of n hyperplanes– Take the square distance to them– A needs to be positive-definite so that we have a nice parabola

35

Gradient of the quadratic form

36

since

And since A is symmetric

Not surprising: we turned Ax=b into the quadratic minimization

(if A is not symmetric, conjuagte gradient finds solution for

– Not our image gradient!– Multidimensional gradient (as

many dim as rows in matrix)

Steepest descent/ascent• Pick

gradient direction

• Find optimum in this direction

37

Gradient direction

Gradient dire

ction

Energy along the gradient

Residual

• At iteration i, we are at a point x(i)• Residual r(i)=b-Ax(i)• Cool property of quadratic form:

residual = - gradient

38

Behavior of gradient descent• Zigzag or goes straight depending if we’re lucky

– Ends up doing multiple steps in the same direction

39

Conjugate gradient

• Smarter choice of direction– Ideally, step directions should be orthogonal to

one another (no redundancy)– But tough to achieve– Next best thing: make them A-orthogonal

(conjugate)That is, orthogonal when transformed by A:

40

Conjugate gradient

• For each step: – Take the residual (gradient)– Make it A-orthogonal to the previous ones– Find minimum along this direction

• Plus life is good:– In practice, you only

need the previous one– You can show that the new

residual r(i+1) is already A-orthogonal to all previous directions p but p(i)

41