Chapter 8imakemathmoreradicalvaldez.weebly.com/uploads/3/0/... · appstats8.notebook 1 October 19,...

28
appstats8.notebook 1 October 19, 2012 Aug 43:41 PM Chapter 8 Linear Regression Aug 43:41 PM Objective: Students will construct and analyze a linear model for a given set of data.

Transcript of Chapter 8imakemathmoreradicalvaldez.weebly.com/uploads/3/0/... · appstats8.notebook 1 October 19,...

Page 1: Chapter 8imakemathmoreradicalvaldez.weebly.com/uploads/3/0/... · appstats8.notebook 1 October 19, 2012 Aug 43:41 PM Chapter 8 Linear Regression Aug 43:41 PM Objective: Students will

appstats8.notebook

1

October 19, 2012

Aug 4­3:41 PM

Chapter 8

Linear Regression

Aug 4­3:41 PM

Objective: Students will construct and analyze a linear model for a given set of data.

Page 2: Chapter 8imakemathmoreradicalvaldez.weebly.com/uploads/3/0/... · appstats8.notebook 1 October 19, 2012 Aug 43:41 PM Chapter 8 Linear Regression Aug 43:41 PM Objective: Students will

appstats8.notebook

2

October 19, 2012

Aug 4­3:41 PM

Fat Versus Protein: An Example

• The following is a scatterplot of total fat versus protein for 30 items on the Burger King menu:

pg 168

• Correlation says “There seems to be a linear association between these two variables,” but it doesn’t tell what that association is.

Aug 4­3:41 PM

The Linear Model

• We can say more about the linear relationship between two quantitative variables with a model.

• A model simplifies reality to help us understand underlying patterns and relationships.

pg 169

Page 3: Chapter 8imakemathmoreradicalvaldez.weebly.com/uploads/3/0/... · appstats8.notebook 1 October 19, 2012 Aug 43:41 PM Chapter 8 Linear Regression Aug 43:41 PM Objective: Students will

appstats8.notebook

3

October 19, 2012

Jul 28­11:12 AM

Linear RegressionNow that we know how to observe thedirection and strength of a linear relationship, we can model the relationship. Just as we used a Normal Model to approximate distributions of data that were unimodal and symmetric, We can use a linear model to make approximate predictions about a data set showing a linear relationship .

Aug 4­3:41 PM

The Linear Model (cont.)

• The linear model is just an equation of a straight line through the data.

• The points in the scatterplot don’t all line up, but a straight line can summarize the general pattern.

• The linear model can help us understand how the values are associated.

pg 169

Page 4: Chapter 8imakemathmoreradicalvaldez.weebly.com/uploads/3/0/... · appstats8.notebook 1 October 19, 2012 Aug 43:41 PM Chapter 8 Linear Regression Aug 43:41 PM Objective: Students will

appstats8.notebook

4

October 19, 2012

Oct 4­7:42 PM

Aug 4­3:41 PM

Correlation and the Line

• Moving one standard deviation away from the mean in x moves us r standard deviations away from the mean in y.

• This relationship is shown in a scatterplot of z­scores for fat and protein:

pg 170

Page 5: Chapter 8imakemathmoreradicalvaldez.weebly.com/uploads/3/0/... · appstats8.notebook 1 October 19, 2012 Aug 43:41 PM Chapter 8 Linear Regression Aug 43:41 PM Objective: Students will

appstats8.notebook

5

October 19, 2012

Aug 4­3:41 PM

Correlation and the Line (cont.)• Put generally, moving any number of standard deviations away from the mean in x moves us r times that number of standard deviations away from the mean in y.

pg 170

Aug 4­3:41 PM

How Big Can Predicted Values Get?

• r cannot be bigger than 1 (in absolute value), so each predicted y tends to be closer to its mean (in standard deviations) than its corresponding x was.

• This property of the linear model is called regression to the mean; the line is called the regression line.

pg 172

Page 6: Chapter 8imakemathmoreradicalvaldez.weebly.com/uploads/3/0/... · appstats8.notebook 1 October 19, 2012 Aug 43:41 PM Chapter 8 Linear Regression Aug 43:41 PM Objective: Students will

appstats8.notebook

6

October 19, 2012

Aug 4­3:41 PM

The Least Squares Line

• We write our model as

• This model says that our predictions from our model follow a straight line.• If the model is a good one, the data values will scatter closely around it.

pg 174

Aug 4­3:41 PM

The Least Squares Line (cont.)

• In our model, we also have an intercept (b0).• The intercept is built from the means and the slope:

• Our intercept is always in units of y.

pg 174

Page 7: Chapter 8imakemathmoreradicalvaldez.weebly.com/uploads/3/0/... · appstats8.notebook 1 October 19, 2012 Aug 43:41 PM Chapter 8 Linear Regression Aug 43:41 PM Objective: Students will

appstats8.notebook

7

October 19, 2012

Aug 4­3:41 PM

The Least Squares Line (cont.)

• In our model, we have a slope (b1):• The slope is built from the correlation and the standard deviations:

• Our slope is always in units of y per unit of x.

pg 174

Aug 4­3:41 PM

The Least Squares Line (cont.)

• Since regression and correlation are closely related, we need to check the same conditions for regressions as we did for correlations:

• Quantitative Variables Condition

• Straight Enough Condition

• Outlier Condition

pg 176see step­by­step

Page 8: Chapter 8imakemathmoreradicalvaldez.weebly.com/uploads/3/0/... · appstats8.notebook 1 October 19, 2012 Aug 43:41 PM Chapter 8 Linear Regression Aug 43:41 PM Objective: Students will

appstats8.notebook

8

October 19, 2012

Aug 4­3:41 PM

Fat Versus Protein: An Example

• The regression line for the Burger King data fits the data well:• The equation is

• The predicted fat content for a BK Broiler chicken sandwich is 6.8 + 0.97(30) = 35.9 grams of fat.

pg 175

Aug 4­3:41 PM

Assumptions and Conditions• Quantitative Variables Condition:

• Regression can only be done on two quantitative variables, so make sure to check this condition.

• Straight Enough Condition:

• The linear model assumes that the relationship between the variables is linear.

• A scatterplot will let you check that the assumption is reasonable.

pg 176

Page 9: Chapter 8imakemathmoreradicalvaldez.weebly.com/uploads/3/0/... · appstats8.notebook 1 October 19, 2012 Aug 43:41 PM Chapter 8 Linear Regression Aug 43:41 PM Objective: Students will

appstats8.notebook

9

October 19, 2012

Aug 4­3:41 PM

Assumptions and Conditions (cont.)

• Outlier Condition: • Watch out for outliers.

• Outlying points can dramatically change a regression model.

• Outliers can even change the sign of the slope, misleading us about the underlying relationship between the variables.

Aug 4­3:41 PM

Assumptions and Conditions (cont.)

1. It’s a good idea to check linearity again after computing the regression when we can examine the residuals.

2. You should also check for outliers, which could change the regression.

3. If the data seem to clump or cluster in the scatterplot, that could be a sign of trouble worth looking into further.

pg 176

Page 10: Chapter 8imakemathmoreradicalvaldez.weebly.com/uploads/3/0/... · appstats8.notebook 1 October 19, 2012 Aug 43:41 PM Chapter 8 Linear Regression Aug 43:41 PM Objective: Students will

appstats8.notebook

10

October 19, 2012

Aug 4­3:41 PM

Assumptions and Conditions (cont.)

• If the scatterplot is not straight enough, stop here.

• You can’t use a linear model for any two variables, even if they are related.

• They must have a linear association or the model won’t mean a thing.

• Some nonlinear relationships can be saved by re­expressing the data to make the scatterplot more linear.

pg 176see step by step go over

Oct 4­7:47 PM

Example 1

Page 11: Chapter 8imakemathmoreradicalvaldez.weebly.com/uploads/3/0/... · appstats8.notebook 1 October 19, 2012 Aug 43:41 PM Chapter 8 Linear Regression Aug 43:41 PM Objective: Students will

appstats8.notebook

11

October 19, 2012

Oct 15­10:22 PM

Objectives:

Students will be able to interpret the slope and y intercept of a linear model in context of the data.

Students will be able to informally assess the fit of a function by plotting and analyzing residuals.

Aug 4­3:41 PM

Residuals• The model won’t be perfect, regardless of the line we draw.

• Some points will be above the line and some will be below.

• The estimate made from a model is the predicted value (denoted as ).

pg 169

Page 12: Chapter 8imakemathmoreradicalvaldez.weebly.com/uploads/3/0/... · appstats8.notebook 1 October 19, 2012 Aug 43:41 PM Chapter 8 Linear Regression Aug 43:41 PM Objective: Students will

appstats8.notebook

12

October 19, 2012

Aug 4­3:41 PM

Residuals Revisited

• The linear model assumes that the relationship between the two variables is a perfect straight line. The residuals are the part of the data that hasn’t been modeled.

Data = Model + Residual or (equivalently)

Residual = Data – ModelOr, in symbols,

pg 178

Aug 4­3:41 PM

Residuals (cont.)• The difference between the observed value and its associated predicted value is called the residual.

• To find the residuals, we always subtract the predicted value from the observed one:

pg 169

Page 13: Chapter 8imakemathmoreradicalvaldez.weebly.com/uploads/3/0/... · appstats8.notebook 1 October 19, 2012 Aug 43:41 PM Chapter 8 Linear Regression Aug 43:41 PM Objective: Students will

appstats8.notebook

13

October 19, 2012

Oct 4­7:57 PM

Residual Plots

A residual is the difference between the observed y‐value and the _______ y ­value for a given x‐value.

residual = y ­ yˆ

The ________________________________________ (SSR) is used to determine the Least‐Squares Regression Line for a given set of data.

predicted

Sum of the least Square Residuals

Aug 4­3:41 PM

“Best Fit” Means Least Squares

• Some residuals are positive, others are negative, and, on average, they cancel each other out.

• So, we can’t assess how well the line fits by adding up all the residuals.

• Similar to what we did with deviations, we square the residuals and add the squares.

• The smaller the sum, the better the fit.

• The line of best fit is the line for which the sum of the squared residuals is smallest.

pg 170

Page 14: Chapter 8imakemathmoreradicalvaldez.weebly.com/uploads/3/0/... · appstats8.notebook 1 October 19, 2012 Aug 43:41 PM Chapter 8 Linear Regression Aug 43:41 PM Objective: Students will

appstats8.notebook

14

October 19, 2012

Aug 4­3:41 PM

Residuals (cont.)• A negative residual means the predicted value’s too big (an overestimate).

• A positive residual means the predicted value’s too small (an underestimate).

pg 169

Aug 4­3:41 PM

Residuals Revisited (cont.)

• Residuals help us to see whether the model makes sense.

• When a regression model is appropriate, nothing interesting should be left behind.

• After we fit a regression model, we usually plot the residuals in the hope of finding…nothing.

pg 178

Page 15: Chapter 8imakemathmoreradicalvaldez.weebly.com/uploads/3/0/... · appstats8.notebook 1 October 19, 2012 Aug 43:41 PM Chapter 8 Linear Regression Aug 43:41 PM Objective: Students will

appstats8.notebook

15

October 19, 2012

Aug 4­3:41 PM

Residuals Revisited (cont.)

• The residuals for the BK menu regression look appropriately boring:

pg 178

Oct 4­8:13 PM

A ____________________ is a scaerplot which graphs the residuals on the _______________ axis and the values of the explanatory variable on the _______________ axis for each data point,( )

The residual plot gives a visual representaon of the amount of error in the model. The closer the residuals are to __________, the smaller the error and the more accurate the model.

The LSRL is a good model if the residual plot shows random _______________ relavely close to the horizontal axis (zero). The horizontal axis represents the _______________.

Points in the residual plot that lie directly on the horizontal axis lie directly on the ___________.

Points in the residual plot that lie above the horizontal axis lie __________ the LSRL. Therefore, the model gives an underesmate at that point. Therefore _______________ residuals represent underesmates.

Points in the residual plot that lie below the horizontal axis, lie __________ the LSRL. Therefore the model gives an overesmate at that point. Therefore _______________ residuals represent overesmates.

The LSRL is not a good model if the residual plot shows ___________________ ____________.

closure

Page 16: Chapter 8imakemathmoreradicalvaldez.weebly.com/uploads/3/0/... · appstats8.notebook 1 October 19, 2012 Aug 43:41 PM Chapter 8 Linear Regression Aug 43:41 PM Objective: Students will

appstats8.notebook

16

October 19, 2012

Oct 15­10:06 PM

Oct 15­10:22 PM

Objectives:

Students will be able calculate the coefficient of determination and interpret its value for evaluating the regression line.

Page 17: Chapter 8imakemathmoreradicalvaldez.weebly.com/uploads/3/0/... · appstats8.notebook 1 October 19, 2012 Aug 43:41 PM Chapter 8 Linear Regression Aug 43:41 PM Objective: Students will

appstats8.notebook

17

October 19, 2012

Aug 4­3:41 PM

R2—The Variation Accounted For

• The variation in the residuals is the key to assessing how well the model fits.

• In the BK menu items example, total fat has a standard deviation of 16.4 grams. The standard deviation of the residuals is 9.2 grams.

pg 179

Aug 4­3:41 PM

R2—The Variation Accounted For (cont.)

• If the correlation were 1.0 and the model predicted the fat values perfectly, the residuals would all be zero and have no variation.

• As it is, the correlation is 0.83—not perfection.

• However, we did see that the model residuals had less variation than total fat alone.

• We can determine how much of the variation is accounted for by the model and how much is left in the residuals.

pg 179

Page 18: Chapter 8imakemathmoreradicalvaldez.weebly.com/uploads/3/0/... · appstats8.notebook 1 October 19, 2012 Aug 43:41 PM Chapter 8 Linear Regression Aug 43:41 PM Objective: Students will

appstats8.notebook

18

October 19, 2012

Aug 4­3:41 PM

R2—The Variation Accounted For (cont.)

• The squared correlation, r2, gives the fraction of the data’s variance accounted for by the model.

• Thus, 1– r2 is the fraction of the original variance left in the residuals.

• For the BK model, r2 = 0.832 = 0.69, so 31% of the variability in total fat has been left in the residuals.

pg 179

Aug 4­3:41 PM

R2—The Variation Accounted For (cont.)

• All regression analyses include this statistic, although by tradition, it is written R2 (pronounced “R­squared”).

• An R2 of 0 means that none of the variance in the data is in the model; all of it is still in the residuals.

• When interpreting a regression model you need to Tell what R2 means.

• In the BK example, 69% of the variation in total fat is accounted for by the model.

pg 179

Page 19: Chapter 8imakemathmoreradicalvaldez.weebly.com/uploads/3/0/... · appstats8.notebook 1 October 19, 2012 Aug 43:41 PM Chapter 8 Linear Regression Aug 43:41 PM Objective: Students will

appstats8.notebook

19

October 19, 2012

Aug 4­3:41 PM

How Big Should R2 Be?

• R2 is always between 0% and 100%. What makes a “good” R2 value depends on the kind of data you are analyzing and on what you want to do with it.

• The standard deviation of the residuals can give us more information about the usefulness of the regression by telling us how much scatter there is around the line.

pg 180

Aug 4­3:41 PM

Reporting R2

• Along with the slope and intercept for a regression, you should always report R2 so that readers can judge for themselves how successful the regression is at fitting the data.

• Statistics is about variation, and R2 measures the success of the regression model in terms of the fraction of the variation of y accounted for by the regression.

pg 181­183see step by step

see TI Tips

Page 20: Chapter 8imakemathmoreradicalvaldez.weebly.com/uploads/3/0/... · appstats8.notebook 1 October 19, 2012 Aug 43:41 PM Chapter 8 Linear Regression Aug 43:41 PM Objective: Students will

appstats8.notebook

20

October 19, 2012

Oct 4­8:35 PM

Coefficient of Determination

The coefficient of determinaon, also called R2, is the square of the _______________. The R2 valuetells how much of the variaon in the response variable is accounted for by the linear regression model.

For example, if R2 = 1, then _____% of the variability in the response variable is accounted for by thelinear model. In other words, the relaonship between the two variables is perfectly linear.

If R2 = 0.95,we can conclude that _____ % of the variability in the response variable is accounted for by the linear relaonship with the explanatory variable.

correlation

100

95

2. A study of class aendance and grades earned among first‐year students at a state universityshowed that in general students who aended a higher percent of their classes earned highergrades. Class aendance explained 16% of the variaon in grades among the students. What isthe numerical correlaon between percent of classes aended and grades earned? __________

Example

Oct 4­8:43 PM

Closure

Page 21: Chapter 8imakemathmoreradicalvaldez.weebly.com/uploads/3/0/... · appstats8.notebook 1 October 19, 2012 Aug 43:41 PM Chapter 8 Linear Regression Aug 43:41 PM Objective: Students will

appstats8.notebook

21

October 19, 2012

Oct 15­10:12 PM

Aug 4­3:41 PM

Reality Check: Is the Regression Reasonable?

• Statistics don’t come out of nowhere. They are based on data.

• The results of a statistical analysis should reinforce your common sense, not fly in its face.

• If the results are surprising, then either you’ve learned something new about the world or your analysis is wrong.

• When you perform a regression, think about the coefficients and ask yourself whether they make sense.

pg 184

Page 22: Chapter 8imakemathmoreradicalvaldez.weebly.com/uploads/3/0/... · appstats8.notebook 1 October 19, 2012 Aug 43:41 PM Chapter 8 Linear Regression Aug 43:41 PM Objective: Students will

appstats8.notebook

22

October 19, 2012

Aug 4­3:41 PM

What Can Go Wrong?

• Don’t fit a straight line to a nonlinear relationship.

• Beware extraordinary points (y­values that stand off from the linear pattern or extreme x­values).

• Don’t extrapolate beyond the data—the linear model may no longer hold outside of the range of the data.

• Don’t infer that x causes y just because there is a good linear model for their relationship—association is not causation.

• Don’t choose a model based on R2 alone.

pg 185

Aug 4­3:41 PM

What have we learned?

• When the relationship between two quantitative variables is fairly straight, a linear model can help summarize that relationship.

• The regression line doesn’t pass through all the points, but it is the best compromise in the sense that it has the smallest sum of squared residuals.

pg 186

Page 23: Chapter 8imakemathmoreradicalvaldez.weebly.com/uploads/3/0/... · appstats8.notebook 1 October 19, 2012 Aug 43:41 PM Chapter 8 Linear Regression Aug 43:41 PM Objective: Students will

appstats8.notebook

23

October 19, 2012

Aug 4­3:41 PM

What have we learned? (cont.)• The correlation tells us several things about the regression:

• 1)The slope of the line is based on the correlation, adjusted for the units of x and y.

• 2)For each SD in x that we are away from the x mean, we expect to be r SDs in y away from the y mean.

• 3)Since r is always between ­1 and +1, each predicted y is fewer SDs away from its mean than the corresponding x was (regression to the mean).

• 4)R2 gives us the fraction of the response accounted for by the regression model.

pg 186

Aug 4­3:41 PM

What have we learned?• The residuals also reveal how well the model works.

• If a plot of the residuals against predicted values shows a pattern, we should re­examine the data to see why.

• The standard deviation of the residuals quantifies the amount of scatter around the line.

pg 186

Page 24: Chapter 8imakemathmoreradicalvaldez.weebly.com/uploads/3/0/... · appstats8.notebook 1 October 19, 2012 Aug 43:41 PM Chapter 8 Linear Regression Aug 43:41 PM Objective: Students will

appstats8.notebook

24

October 19, 2012

Oct 4­8:55 PM

Oct 4­8:57 PM

Page 25: Chapter 8imakemathmoreradicalvaldez.weebly.com/uploads/3/0/... · appstats8.notebook 1 October 19, 2012 Aug 43:41 PM Chapter 8 Linear Regression Aug 43:41 PM Objective: Students will

appstats8.notebook

25

October 19, 2012

Oct 4­8:56 PM

Oct 4­8:59 PM

Page 26: Chapter 8imakemathmoreradicalvaldez.weebly.com/uploads/3/0/... · appstats8.notebook 1 October 19, 2012 Aug 43:41 PM Chapter 8 Linear Regression Aug 43:41 PM Objective: Students will

appstats8.notebook

26

October 19, 2012

Oct 4­8:43 PM

1. Below are heights and weights data for male and female AP Stat students. Create TI lists MHT,MWT, FHT, FWT, and enter the data. You will need to keep these data for several days.(Remember, to create a list: (1) put the cursor atop the names (L1,L2,...), then (2) space to theright, and (3) type the new name in the blank space. You can then access these names via theLIST NAMES menu.)

2. Is it reasonable to assume these data are drawn from populations that are normally distributed?Check summary statistics and histograms for each variable.

3. a) Make a scatterplot for each gender (STAT PLOT, first plot type). Which is the explanatoryvariable?b) Describe the relationship (form, strength, direction, outliers, etc.) for each gender.

4. a) Calculate and interpret r for each gender.b) What would you predict about the weight of. . . someone of average height?. . . a male 2 standard deviations above average in height.. . . a female 1 standard deviation below average in height?c) Explain how this illustrates the concept of regression toward the mean.d) Calculate the coefficients of determination, r2, and interpret each in the context of thisrelationship.

5. a) Write an equation of the least squares line of best fit for each gender.b) Check the residuals plots. Do you think that a line is a good model? Explain.c) Explain what the slope of each line means in the context of this relationship.d) Predict weights for : a 60” male; a 60” female; a 70” male; a 70” femalee) Predict the weight of a 7’2” male; of a 20” newborn baby girl. Comment on these results.

Oct 4­8:52 PM

Page 27: Chapter 8imakemathmoreradicalvaldez.weebly.com/uploads/3/0/... · appstats8.notebook 1 October 19, 2012 Aug 43:41 PM Chapter 8 Linear Regression Aug 43:41 PM Objective: Students will

appstats8.notebook

27

October 19, 2012

Oct 4­8:53 PM

Oct 4­8:54 PM

Page 28: Chapter 8imakemathmoreradicalvaldez.weebly.com/uploads/3/0/... · appstats8.notebook 1 October 19, 2012 Aug 43:41 PM Chapter 8 Linear Regression Aug 43:41 PM Objective: Students will

appstats8.notebook

28

October 19, 2012

Oct 4­8:54 PM

Oct 4­9:04 PM