Leadership Academy “Data Analysis” June 2010. Some thoughts about data analysis… We have moved...

Post on 19-Jan-2016

212 views 0 download

Tags:

Transcript of Leadership Academy “Data Analysis” June 2010. Some thoughts about data analysis… We have moved...

Leadership Academy“Data Analysis”

June 2010

Some thoughts about data analysis…

We have moved from a system where hunches and “cardiac data” were followed, to a system that is data driven.

We have moved from assuming good things are happening to insisting good things are happening and proving this is so.

Teaching is an art……and becoming a science.

Learning Outcomes

• Participants will identify:– Key ideas about effective data-based decision

making– Four types of data– Challenges within our district ITBS/ITED results– Findings & results from last year’s Leadership

Academy– Effective strategies to help improve student

achievement– Ideas to include within our revised Strategic Plan

8 Key Points

1.Data provides insight

2.Data are nothing without analysis.

3.Interpret graphs accurately

4.Improvement means more success, less failure, and less unwanted variation!

5.Real-time, frequent data are necessary to monitor improvement.

6.Data must be accessible to users.

7.To improve results, work on both product and processes.

8.Take action!

8 Key Points (cont.)

Leaders Ask Good Questions

What does the graph look like?

What patterns or trends do the data show?

Special or common cause?

Over-reacting or under-reacting?

“Is this the complete picture?”

Are there other assessments we need to look at?

What other data might we need to confirm, support and or clarify what we are seeing now?

How do I know if what I’m seeing is real? Try to use multiple data sources. Triangulation occurs if three different data sources all indicate the same thing.

ITBS scores

Unit test scores

District test score

Why do some people dislike (fear?) data analysis?

Data

Share your thoughts with someone at your

table?

Why do some people dislike (fear?) data analysis?

Data analysis… unlocks the door to the

traditionally private domain of the classroom,

inevitably results in a mandate for change, and

increases pressure for accountability, especially if results are poor.

Data

What did we learn from last year’s Leadership Academy?

Global Positioning System (GPS)

Mapping the Way

Few credits earned

Proficient math scores

Reading scores were low.

Did not attend

Estimate time of arrival…

Commencement Day!

12th11th10th9th8th7th6th5th4th3rd2nd1stKPre K

Write data in each arrow indicating your student's successes and/or challenges at each grade level.

Student GPS Timeline

Student dropped out

Start of poor attendance

Good attendance

Data DrillsDropouts

Struggling Students

Walkthoughs

Perception Surveys

Four Types of Data

Classroom Data ______________

_____

______________

______________

_____

______________

_____

Findings & Results of Last Year’s Leadership Academy

1. Struggling students & dropouts are a K-12 concern.-“Intervention Record” piloted in four elementary schools.

-Pinnacle Analytics set up to red flag struggling students & potential dropouts

2. Our data is inconsistent, poorly organized, and not effectively used to guide our instruction.

-Pinnacle Analytics has been designed to:

*Manage, organize & store our data,

*Visually help us correlate & analyze our data, and

*Help us work smarter, not harder.

Informing Our Strategic Plan

• Last year’s Leadership Academy findings were incorporated into our Strategic Plan.

• This year’s Leadership Academy findings will be as well.

Strategic Plan GoalsAchievement

Perception

Demographic

Classroom

NEEDS, GOAL(S), FOR STRATEGIC PLAN

Throughout the Leadership Academy, record relevant data along with corresponding needs and goals of our district that you believe

should be included in our next revision of the Strategic Plan.

Student Achievement Data

1. What is our current reality?

2. Is it “good enough”?

3. What are we doing to cause this?

4. What will we do to maintain or

improve?

Initial 2009-2010 ITBS/ITED Results

Reading Comprehension Grades 3-5 FAY Students

0

20

40

60

80

100

% Pr

ofici

ent

FAY 66.7 67.5 68.9 64.9 68.3 70.3

Trajecory 69.4 69.4 69.4 75.5 75.5 75.5 81.6 87.8 93.9 100

2004 2005 2006 2007 2008 2009 2010 2011 2012 2013

Trajectory

District FAY Score Trend line

Reading Comprehension Grades 6-8 FAY Students

0

20

40

60

80

100

% Pr

ofici

ent

Fay 58 60.2 62 61 62 63.7

Trajectory 64.4 64.4 64.4 71.5 71.5 71.5 78.7 85.8 92.9 100

2004 2005 2006 2007 2008 2009 2010 2011 2012 2013

Reading Comprehension Grade 11 FAY Students

0

20

40

60

80

100

% Pr

ofici

ent

FAY 61.8 61.7 68.0 67.5 68.4 68.5 74.6

Trajectory 69.0 74.2 74.2 74.2 79.3 79.3 79.3 84.5 89.7 94.8 100

2003 2004 2005 2006 2007 2008 2009 2010 2011 2012 2013

Math Total Grades 3-5 FAY Students

0

20

40

60

80

100

% Pr

ofici

ent

FAY 66.3 66.9 69.1 66.4 68.3 70

Trajectory 68.8 68.8 68.8 75.1 75.1 75.1 81.3 87.5 93.8 100

2004 2005 2006 2007 2008 2009 2010 2011 2012 2013

Math Total Grades 6-8 FAY Students

0

20

40

60

80

100

% Pr

ofici

ent

FAY 61 62.4 64.3 63 63.7 65

Trajectory 65.3 65.3 65.3 72.3 72.3 72.3 79.2 86.1 93.1 100

2004 2005 2006 2007 2008 2009 2010 2011 2012 2013

Math Total Grade 11 FAY Students

0

20

40

60

80

100

% Pr

ofici

ent

FAY 64.0 67.7 63.8 71.2 66.6 68.6 71.5

Trajectory 69.0 74.2 74.2 74.2 79.3 79.3 79.3 84.5 89.7 94.8 100

2003 2004 2005 2006 2007 2008 2009 2010 2011 2012 2013

Science Grades 3-5 FAY Students

0

20

40

60

80

100

% Pr

ofici

ent

FAY 69.7 73.2 74 72.9 73.1 75.9

2004 2005 2006 2007 2008 2009

Science Grades 6-8 FAY Students

0

20

40

60

80

100

% Pr

ofici

ent

FAY 68.9 69.3 74.5 72.9 73 76.6

2004 2005 2006 2007 2008 2009

Science Grade 11 FAY Students

0

20

40

60

80

100

% Pr

ofici

ent

FAY 61.8 66.8 66.4 73.8 73.6 77 77.8

2003 2004 2005 2006 2007 2008 2009

Reading Comprehension Grades 3-5 FAY Students

0

20

40

60

80

100

% Pr

ofici

ent

FAY 66.7 67.5 68.9 64.9 68.3 70.3

Trajecory 69.4 69.4 69.4 75.5 75.5 75.5 81.6 87.8 93.9 100

2004 2005 2006 2007 2008 2009 2010 2011 2012 2013

Trajectory

District FAY Score Trend line

1. What is our current reality?

2. Is it “good enough”?

3. What are we doing to cause this?

4. What will we do to maintain or

improve?

MathTotal Grades 6-8 FAY Students

0

10

20

30

40

50

60

70

80

90

100

% Pr

ofici

ent

2006 72.1 51.1 73.8 46.2 47.3

2007 71.1 50.4 64 40.9 48.3

2008 71.6 51.8 64.2 42.6 51.5

2009 73.3 58.2 65.9 47.6 53.1

White Native Am. Asian African Am Hispanic

% Proficient Trajectory

Reading Comprehension Grades 3-5 FAY Students

0

10

20

30

40

50

60

70

80

90

100

% P

rofic

ient

2006 68.9 58.9 33.8 50.1

2007 64.9 54.4 28.1 46.9

2008 68.3 60.6 33.8 51.7

2009 70.3 62.3 34.9 52.6

FAY Students Low SES IEP ELL

% Proficient Trajectory

1. What is our current reality?

2. Is it “good enough”?

3. What are we doing to cause this?

4. What will we do to maintain or

improve?

Full Academic Year (FAY) StudentsCohort Group from 3rd Grade to 11th Grade

40

60

80

100

% P

rofic

ient

Reading 64 69.8 69.3 55.3 63.7 63.3 56 63.2 74.6

Math 54.7 66.2 62.4 56.8 63.1 64.8 66 59.4 71.5

Science 64.9 76.6 72.6 65.8 70.1 77.4 75.9 69.8 77.8

3rd 4th 5th 6th 7th 8th 9th 10th 11th

FINAL District AYP Results of 2008-09 Iowa Tests

 Elementary

MathElementary

ReadingMiddle Math

Middle Reading

High MathHigh

Reading

All MissedMet - Safe

HarborMissed Missed Missed Missed

Low SES Missed Missed Missed Missed Missed Missed

IEP MissedMet - Safe

HarborMissed Missed

Met - Safe Harbor

Missed

ELL MissedMet - Safe

HarborMet - Safe

HarborMissed

Met - Safe Harbor

Missed

African American Met- GHS Missed Missed Met - GSHMet - Safe

HarborMet - Safe

Harbor

Asian MissedMet - Safe

HarborMet

Triennium Met - Trien Met

Met Triennium

Hispanic MissedMet - Safe

HarborMissed Missed Missed Missed

Native American Missed Missed Missed MissedMet - Safe

HarborMissed

White Met Met Met - Trien Met - Trien Met - Trien Missed

“Final” District AYP Results of 2008-09 Iowa Tests

Reading Math Partic Reading Math Partic Reading Math Partic

District delay 2 delay 2 DINA 3 DINA 3 DINA 4 DINA 4

East High watch SINA 1 SINA 1 SINA 2 SINA 2 SINA 3

North High SINA 2 SINA 3 watch SINA 3 SINA 4 SINA 1 SINA 4 SINA 5 delay 1

West High SINA 3 SINA 1 SINA 4 SINA 2 watch SINA 5 SINA 3

East Middle delay 2 watch SINA 3 SINA 1 SINA 4

North Middle watch watch SINA 1 SINA 1

West Middle SINA 3 delay 2 SINA 4 SINA 3 SINA 5 SINA 4

Central Campus delay 2 delay 2 watch SINA 3 SINA 3 SINA 1 SINA 4 SINA 4 SINA 2

Bryant watch watch SINA 1

ClarkCrescent Park watch

Emerson watch watch

Everett watch watch SINA 1 SINA 1

HuntIrving watch watch SINA 1

JoyLeeds watch watch

LincolnLongfellow watch SINA 1

Riverside watch watch SINA 1

Roosevelt watch watch SINA 1

Smith watch watch SINA 1 SINA 1

SunnysideUnity watch

WashingtonWhittier watch watch SINA 1

# of watch 1 3 2 12 10 2

# of SINA 3 5 6 1 11 14 1

# of delay 1 3 1

Disrict's & Schools' AYP Status Based Upon Previous Year's ITBS/ITED Results

2007-08 2008-09 2009-10

1. What is our current reality?

2. Is it “good enough”?

3. What are we doing to cause this?

4. What will we do to maintain or

improve?

We are moving to spring ITBS/ITED testing next year.

How might this change help our current reality?

Since we will continue to be evaluated by the

ITBS/ITED annual “snapshot”, how can we use our transition

to spring testing to help us better prepare

for the tests?

How can we provide a better picture of learning over a particular school year using the spring ITBS/ITED?

How can we better address our dip in transition year scores?

How can we better address the concerns of summer learning loss?

Can we avoid having to estimate the effects of one grade level from another since we no long would take the tests only three months into the school year?

1. How might this change help our current reality?

2. How can we use our transition to spring testing to better prepare for the tests?

3. How can we provide a better picture of learning over a particular school year?

4. How can we better address our dip in transition year scores?

5. How can we better address the concerns of summer learning loss?

6. Can we avoid having to estimate the effects of one grade level from another?

Moving to Spring ITBS/ITED

Will simply changing to spring testing with four additional months to prepare, automatically improve our scores?

What about norms?

Norms compare a student’s raw score, the number of problems the student got correct, with the scores from other students who took the same test at approximately the same time of the school year.

How much a student “knows” is inferred from

their standing or rank within that comparison or

norm group.

ITBS & ITEDNumber of Correct Items Required for Proficiency

Form

Code

ContentProficiency

Fall Midyear

Spring

# of items

3rd Grade-- Level 9

A RC Reading Comprehension

15 17 19 37

M1 Math Concepts & Estimation

16 17 19 31

M2 Math P. S*. & Data Interpret.

10 12 13 22

SC Science 13 14 15 30

B RC Reading Comprehension

17 19 21 37

M1 Math Concepts & Estimation

15 16 18 31

M2 Math P. S. & Data Interpret.

11 12 14 22

SC Science 13 14 16 30

Simply changing to spring testing with four additional months to prepare, will not automatically improve our scores.

We may even experience a dip in scores?

Spring ITBS / ITED Testing

• We are confident

• Improved curriculum, instruction & assessments

• Increased ownership and motivation

• Improved student achievement and proficiency on the ITBS/ITED.

Spring ITBS / ITED Testing?

• Beginning with the 2010-11 school year, we will be moving to spring testing.

• Consensus from all stakeholder groups

• Board acknowledgement and support

Preparing for Spring Testing

• March 14 – April 1• Results will arrive mid-May • District Assessment Calendar Adjustments

(ELDA in Feb, DRA-2 to 4th qrt.)• Improve curriculum, instruction & assessment• Increase formative assessment & differentiation• Use Pinnacle to analyze and triangulate data

throughout the year• Increase ownership and motivation• Learn from each other

What other achievement data should we be looking at?