Do You: Discover the Missing Ingredient in Today’s Modern ...
Diagnosis the missing ingredient in RTI assessment
description
Transcript of Diagnosis the missing ingredient in RTI assessment
![Page 1: Diagnosis the missing ingredient in RTI assessment](https://reader036.fdocuments.us/reader036/viewer/2022081821/549b3cffb4795924098b4723/html5/thumbnails/1.jpg)
204
The Reading Teacher Vol. 65 Issue 3 pp. 204–208 DOI:10.1002/TRTR.01031 © 2011 International Reading AssociationR T
P E R S P E C T I V E S O N R T I
DIAGNOSIS
Descriptions and advice about Response
to Intervention (RTI) implementations
often focus on assessment. In an earlier
column, Wixson and Valencia (2011)
described the multiple purposes for assessment in an
RTI system:
■ Screening
■ Diagnostics
■ Formative progress monitoring
■ Benchmark progress monitoring
■ Summative outcome assessment
Because RTI is an approach to identifying students
as learning disabled, it is important to attend to the
quality of the measurements used to make high-
stakes decisions about student placements. However,
RTI is also intended to reduce the number of
students who become identified as having a learning
disability by preventing reading difficulties. This is an
instructional problem and, as Scanlon and Anderson
(2010) noted, too little attention is given to “the nature
and qualities of the instruction that is offered” (p. 21).
All students deserve high-quality first instruction (we
will address this topic in an upcoming column).
However, good teaching requires thoughtful
assessment. Furthermore, some students continue to
struggle, even when they receive very good classroom
instruction. These students need more tailored
instruction that is responsive to their specific strengths
and areas of need. Screening measures designed to
identify struggling students are usually quite general or
sample only a snippet of skilled reading performance.
They rarely provide the specific information needed
to determine the most appropriate intervention or
instruction.
By itself, this would not be problematic. However,
many schools and districts have created RTI systems
that move straight from screening to instruction
without looking more closely at the individual
student, an approach called the direct route (Johnson,
Jenkins, & Petscher, 2010). Of course, when the
direct route decision is based on only one screening
instrument, there is quite a grave possibility of error
in classification (false negatives or false positives). As
well, these single measures do not provide enough
information to make good instructional decisions
about individuals or small groups of students
(Valencia et al., 2010).
This problem is exacerbated when the school adopts
a one-size-fits-all intervention, a practice sometimes
called standard protocol (Fuchs et al., 2003). Although
Fuchs and his colleagues were clear that a standard
protocol was to be employed with groups of students
with similar reading difficulties, a single screening
measure often invites only one possible response, and,
Marjorie Y. Lipson and Pam Chomsky-Higgins provide research and professional development support to schools through the Vermont Reads Institute at the University of Vermont, Montpelier, USA; e-mail [email protected] and [email protected].
Jane Kanfer is a reading specialist at Milton Elementary School in Vermont; e-mail [email protected].
The department editors welcome reader comments. Karen K. Wixson is dean of the school of education at the University of North Carolina at Greensboro, USA; e-mail [email protected]. Marjorie Y. Lipson is Professor Emerita at the University of Vermont, Burlington, USA; e-mail [email protected].
The Missing Ingredient in RTI Assessment
Marjorie Y. Lipsonwith Pam Chomsky-Higgins ■ Jane Kanfer
TRTR_1031.indd 204TRTR_1031.indd 204 11/7/2011 4:05:47 PM11/7/2011 4:05:47 PM
![Page 2: Diagnosis the missing ingredient in RTI assessment](https://reader036.fdocuments.us/reader036/viewer/2022081821/549b3cffb4795924098b4723/html5/thumbnails/2.jpg)
DI AGNOSIS: T H E M ISSI NG I NGR EDI E N T I N RT I A SSESSM E N T
www.reading.org R T
205
for many schools, the direct route leads to
just one instructional option.
A very large body of research is
emerging to confirm what good teachers
and specialists have always known: the
underlying roots of students’ reading
difficulties are diverse (Aaron et al.,
2008; Valencia & Buly, 2004). In addition,
it is becoming quite clear that instruction
focused on the wrong thing not only
does not help students, but it may
actually be harmful (Connor, Morrison,
& Katch, 2004). It should be a very high
priority for teachers and specialists
to gather specific information about
individual students to make appropriate
instructional decisions. So, we turn
our attention to an essential, but often
neglected, type of assessment: diagnosis.
Using Assessment for Diagnostic PurposesBecause the word diagnosis is so often
associated with identifying disease or
its symptoms, we want to be clear that
we do not assume a medical model of
reading difficulties. Instead, we want
to promote the original meaning of
the word diagnosis, from the Greek to discern the nature and cause of anything.
According to Merriam-Webster, a
diagnosis is “an investigation or analysis
of the cause or nature of a condition,
situation or problem” (www.aolsvc
.merriam-webster.aol.com/dictionary/
diagnosis). Without very good diagnostic
information and/or a flexible formative
assessment system, our instructional
programs and student performance will
not improve, and RTI will simply be
an alternate route to special education
placement or to permanent membership
in Title I classrooms.
We have been collaborating with
many Vermont schools to improve
literacy success for all students and have
recently turned our attention to helping
schools adopt a systemic approach to
RTI. Pam and Marge started to work
at Jane’s school 2 years ago. This large
Vermont school was entering year 3 of
corrective action. Despite several years of
attention to literacy, too many students
were identified for special education,
and too few were benefitting from the
literacy instruction they received.
The reading teachers in the building
were knowledgeable and committed.
However, they were relying on a
conventional approach to services. Using
universal assessment data, they identified
students near the beginning of the year,
created a schedule for the students on
“their caseload,” and then pulled those
students from the classroom for the
remainder of the year. The nature of the
instruction provided varied from teacher
to teacher and was not well documented.
More troubling were the longitudinal
data suggesting that students who
participated in reading services (provided
by reading teachers or special educators)
were not gaining ground against their
peers. Although there were exceptions,
students typically continued to qualify for
reading services from year to year, and
their overall performance was below the
state standards.
During year 1 of our collaboration,
Jane and her colleagues began to
examine their available assessment data
much more closely. They moved from
global reading levels to looking at the
component areas of reading (e.g., word
recognition, fluency, comprehension),
and they constructed “typical profiles”
of student difficulties (see Figure for the
framework). The first profiles focused not
Figure Template for Analyzing Student Data and Creating Profiles
Intervention Plan for Literacy—TemplateStudent Profile: [Describe or name the type of profile—for example, Struggling Grade 1 Student or Intermediate Comprehension Difficulty]Intervention Focus: [Describe or name primary focus for intervention]
Area(s) of Primary Concern Evidence and Data Source
Phonological awarenessLetter IDSight vocabularyFluencyDecodingVocabularyComprehensionMotivationStaminaWriting about readingText levelOther
Goal(s) of the Intervention:Describe the Intervention: [Include Information about:
• Duration• Frequency• Materials to be used• Strategies, techniques, approaches, etc.• Instructional framework• Person responsible• Relationship to classroom instruction]
Progress Monitoring: [How will instructional impact and effectiveness be assessed?]
TRTR_1031.indd 205TRTR_1031.indd 205 11/7/2011 4:05:48 PM11/7/2011 4:05:48 PM
![Page 3: Diagnosis the missing ingredient in RTI assessment](https://reader036.fdocuments.us/reader036/viewer/2022081821/549b3cffb4795924098b4723/html5/thumbnails/3.jpg)
DI AGNOSIS: T H E M ISSI NG I NGR EDI E N T I N RT I A SSESSM E N T
206
The Reading Teacher Vol. 65 Issue 3 November 2011R T
on the students with the most severe (and
least frequent) types of difficulties, but on
the most common types of difficulties—
arguing that these would have the
broadest impact and might also provide
insights into classroom practices that
could be strengthened. These students
are not necessarily the so-called “bubble
kids,” those whose test results suggest
that they are very nearly proficient (or “on
the bubble”). Cynical practices that direct
teachers to focus on those students so
that schools can realize “improvement”
have been criticized roundly elsewhere
(Koretz, 2008). In some schools, the most
common profile might be students who
are close to proficiency; in others, it could
be other students.
The most common profile when
we started involved struggling grade
1 students who showed very limited
proficiency. We called this profile
“beginning struggling word learner,”
and the intervention focus was strategic
decoding. These were students who had,
with good instruction in kindergarten,
come to grade 1 knowing all or most of
the letter names and many/most letter
sounds. They typically knew a small
collection of high-frequency sight words.
However, they had not met the leveled
text standard for end of kindergarten
(Level B, guided reading level) and were
making very slow progress in letter-
sound blending and in problem solving.
These students were an early focus for
us because we knew that helping them
was pivotal in preventing reading failure
later on, especially because another
common profile in the school involved
students in grades 3 through 6 who had
acquired good decoding skills but whose
comprehension was weak. These students
were especially worrisome because their
performance actually declined as they
moved through the grades.
As we created diagnostic profiles,
Jane and her colleagues detailed the
research-based
intervention
plans
appropriate for
each student.
The result
was similar
to what Dorn
and Henderson
(2010) called
a “portfolio of
interventions.”
Although
Jane’s school
has not fully
implemented
the portfolio of interventions, the results
of the early interventions have been
excellent. By the middle of the second
year of collaboration, every student in
grade 1 who had been identified as most
struggling met or exceeded the standard
on the periodic benchmark assessment
used by the school, and most were
released from intervention. Importantly,
these students did not receive more
intervention than they had in the past;
they simply received more tailored and
focused instruction based on careful
and comprehensive assessment and
the research-based interventions,
confirming what others have observed:
More intervention is not necessarily
better (Wanzek & Vaughn, 2008).
Diagnostic Assessment PracticesClassroom teachers and reading
specialists may already have excellent
assessment information available to
them collected for screening, periodic
benchmarking, or outcomes evaluation.
In these cases, the challenge is to
adopt systems that permit effective use
of that information for instructional
decision making. Designing data-team
formats and/or profiles can facilitate that
considerably. Other schools and districts
either have few data or have restricted
teachers’ access to the information.
There are two actions that may be
needed: (1) disaggregating existing
data to mine it for insights and/
or (2) planning to gather additional
information. In either event, it is
important to start with a matrix of the
most important component areas of
reading (and writing). These are visible
on the profile planner (see Figure). Of
course, not all information is equally
relevant for all students or all ages/
grades. One of the major difficulties
with a standard protocol approach is
that it may not recognize that there
“The challenge is to adopt systems that
permit effective use of that information
for instructional decision making.”
TRTR_1031.indd 206TRTR_1031.indd 206 11/7/2011 4:05:49 PM11/7/2011 4:05:49 PM
![Page 4: Diagnosis the missing ingredient in RTI assessment](https://reader036.fdocuments.us/reader036/viewer/2022081821/549b3cffb4795924098b4723/html5/thumbnails/4.jpg)
DI AGNOSIS: T H E M ISSI NG I NGR EDI E N T I N RT I A SSESSM E N T
www.reading.org R T
207
are developmental changes in the
relationships between and among
various factors and reading performance
(Valencia et al., 2010).
Whenever possible, it is efficient
to gather observations about multiple
types of reading performance with one
tool. At the very least, this can help
narrow the need for further assessment.
For example, Jane’s school, like many
we work with, uses one of the many
structured inventories or benchmark
assessment systems available (i.e., DRA,
Benchmark Assessment System, QRI-4),
which yields a text level for each student.
Over the past 2 years, the reading
professionals in that building have begun
to examine the results more closely for
evidence of word recognition accuracy,
fluency, and comprehension. Whereas
they had previously used only the text
level to make decisions about student
placements and programs, they now
look more closely at their data. They now
examine (and keep track of) differences
between students’ word recognition
accuracy as distinct from fluency. They
also consider whether students’ accuracy
is affected by phonics decoding, sight
word recognition, or both. Finally, of
course (and especially for older students),
they examine differences between
and among results
in comprehension,
fluency, and word
identification. They have
differentiated instruction
for those students who
were struggling with
comprehension even
though they were quite
fluent and vice versa.
It is important to
examine assessment
results in more than one
way because different
measures can provide
different insights. As
Valencia and her colleagues (2010)
demonstrated, we might get quite a
different picture of a student’s strength
and areas of need with an array of
information than if we relied on only
one score. The student profile form was
intended to encourage teachers to look
at multiple measures to identify areas of
concern that were not at all evident with
only word-level assessment data. This
comprehensive range of information is
important for all struggling readers, but
it is especially important to make good
instructional decisions for older students
and English learners (Klingner, Soltero-
Gonzales, & Lesaux, 2010; Vaughn &
Fletcher, 2010).
Every bit of this diagnostic
information can be gathered during
comprehensive benchmarking events.
Once we have looked at the data, we
can make additional decisions about
both assessment and instruction.
Good diagnosis can sometimes lead
to more assessment. For example, if a
student has made many miscues during
reading, further examination using
a modified miscue analysis can be
helpful (Lipson & Wixson, 2009). If the
student is exceptionally accurate, there
is no need to explore that component.
Similarly, if the student’s comprehension
is strong, we probably do not need
to look to vocabulary. This might
change, of course, if we have evidence
that the student performs less well on
expository text than narrative. Well-
planned diagnostic teaching episodes
can be helpful as well (Fuchs et al., 2007;
Lipson & Wixson, 1986, 2009).
Using Assessment to DifferentiateWe want to emphasize that this type
of closer look, engaging in diagnostic
assessment, is not necessary for
everyone. One of the advantages of an
effective overall assessment system is
that it can identify the students who are
benefitting from classroom instruction
and who are doing well. If we test those
children just a bit less (or use good,
formative assessment effectively), then it
frees up time to assess the students who
really have us worried.
Of course, diagnostic assessment
will not make any difference if it does
not lead to action. Matching students’
needs with instruction by using “if-then”
thinking (Strickland, 2005) is at the heart
of effective and efficient assessment
instruction and is essential if RTI is going
to make a difference. Students who are
struggling for success in the classroom
deserve differentiated instruction in the
classroom and tailored interventions in
supplemental settings. Fortunately, there
is an increasingly strong body of research
that points to the types of instruction
“Of course, diagnostic
assessment will not
make any difference
if it does not lead
to action.”
TRTR_1031.indd 207TRTR_1031.indd 207 11/7/2011 4:05:50 PM11/7/2011 4:05:50 PM
![Page 5: Diagnosis the missing ingredient in RTI assessment](https://reader036.fdocuments.us/reader036/viewer/2022081821/549b3cffb4795924098b4723/html5/thumbnails/5.jpg)
DI AGNOSIS: T H E M ISSI NG I NGR EDI E N T I N RT I A SSESSM E N T
208
The Reading Teacher Vol. 65 Issue 3 November 2011R T
and intervention that can help to prevent
reading disabilities and/or improve
students’ abilities during later years.
RE F ERENC ES
Aaron, P.G., Joshi, R.M., Gooden, R., & Bentum, K.W. (2008). Diagnosis and treatment of reading disabilities based on the component model of reading: An alternative to the discrepancy model of LD. Journal of Learning Disabilities, 41(1), 67–84. doi:10.1177/0022219407310838
Connor, C.M., Morrison, F.J., & Katch, L.E. (2004). Beyond the reading wars: Exploring the effect of child-instruction interactions on growth in early reading. Scientific Studies of Reading, 8(4), 305–336. doi:10.1207/s1532799xssr0804_1
Dorn, L. J., & Henderson, S. C. (2010). The comprehensive intervention model: A systems approach to RTI. In M.Y. Lipson & K.W. Wixson (Eds.), Successful approaches to RTI (pp. 88-120). Newark, DE: International Reading Association
Fuchs, D., Fuchs, L.S., Compton, D.L., Bouton, B., Caffrey, E., & Hill, L. (2007). Dynamic assessment as responsiveness to intervention. Teaching Exceptional Children, 39(5), 58–63.
Fuchs, D., Mock, D., Morgan, P.L., & Young, C. (2003). Responsiveness to intervention: Definitions, evidence, and implications for the learning disabilities construct. Learning Disabilities Research & Practice, 18(3), 157–171. doi:10.1111/1540-5826.00072
Johnson, E.S., Jenkins, J.R., & Petscher, Y. (2010). Improving the accuracy of a direct route screening process. Assessment for Effective Intervention, 35(3), 131–140. doi:10.1177/1534508409348375
Klingner, J.K., Soltero-Gonzales, L., & Lesaux, N. (2010). RTI for English-language learners. In M.Y. Lipson & K.K. Wixson (Eds.), Successful approaches to RTI: Collaborative practices for improving K-12 literacy (pp. 134–162). Newark, DE: International Reading Association.
IRA Books ■ RTI and the Adolescent Reader by William G.
Brozo (copublished with Teachers College
Press) ■ RTI in Literacy—Responsive and Comprehensive
edited by Peter H. Johnston ■ Successful Approaches to RTI: Collaborative
Practices for Improving K –12 Literacy
edited by Marjorie Y. Lipson and
Karen K. Wixson
IRA Journal Articles ■ “Response to Intervention (RTI):
What Teachers of Reading Need to Know”
by Eric M. Mesmer and Heidi Anne
E. Mesmer, The Reading Teacher,
December 2008 ■ “Responsiveness to Intervention: Multilevel
Assessment and Instruction as Early
Intervention and Disability Identification” by
Douglas Fuchs and Lynn S. Fuchs,
The Reading Teacher,
November 2009
MOR E TO E X PLOR E
Koretz, D. (2008). Measuring up: What educational testing really tells us. Boston: Harvard University Press.
Lipson, M.Y., & Wixson, K.K. (1986). Reading disability research: An interactionist pers pec tive. Review of Educational Research, 56(1), 111–136.
Lipson, M.Y., & Wixson, K.K. (2009). Assessment and instruction of reading and writing difficulties: An interactive approach (4th ed.). New York: Pearson Education.
Scanlon, D.M., & Anderson, K.L. (2010). Using the interactive strategies approach to prevent reading difficulties in an RTI context. In M.Y. Lipson & K.W. Wixson (Eds.), Successful approaches to RTI (pp. 20–65). Newark, DE: International Reading Association.
Strickland, K. (2005). What’s after assessment? Follow-up instruction for phonics, fluency, and comprehension. Portsmouth, NH: Heinemann.
Valencia, S.W., & Buly, M.R. (2004). Behind test scores: What struggling readers really need. The Reading Teacher, 57(6), 520–531.
Valencia, S.W., Smith, A.T., Reece, A.M., Li, M., Wixson, K.K., & Newman, H. (2010). Oral reading fluency assessment: Issues of construct, criterion, and consequential validity. Reading Research Quarterly, 45(3), 270–291. doi:10.1598/RRQ.45.3.1
Vaughn, S., & Fletcher, J. (2010). Thoughts on rethinking response to intervention with secondary students. School Psychology Review, 39(2), 296–299.
Wanzek, J., & Vaughn, S. (2008). Response to varying amounts of time in reading intervention for students with low response to intervention. Journal of Learning Disabilities, 41(2), 126–142. doi:10.1177/0022219407313426
Wixson, K.K., & Valencia, S.W. (2011). Assessment in RTI: What teachers and specialists need to know. The Reading Teacher, 64(6), 466–469.
TRTR_1031.indd 208TRTR_1031.indd 208 11/7/2011 4:05:52 PM11/7/2011 4:05:52 PM
![Page 6: Diagnosis the missing ingredient in RTI assessment](https://reader036.fdocuments.us/reader036/viewer/2022081821/549b3cffb4795924098b4723/html5/thumbnails/6.jpg)
Copyright of Reading Teacher is the property of International Reading Association and its content may not be
copied or emailed to multiple sites or posted to a listserv without the copyright holder's express written
permission. However, users may print, download, or email articles for individual use.