Modeling “The Cause”: Assessing Implementation Fidelity and Achieved Relative Strength in RCTs...
-
Upload
dortha-parks -
Category
Documents
-
view
218 -
download
2
Transcript of Modeling “The Cause”: Assessing Implementation Fidelity and Achieved Relative Strength in RCTs...
Modeling “The Cause”: Assessing Implementation
Fidelity and Achieved Relative Strength in RCTs
David S. CordrayVanderbilt University
IES/NCER Summer Research Training Institute, 2010
Overview• Define implementation fidelity and achieved relative strength• A 4-step approach to assessment and analysis of implementation
fidelity (IF) and achieved relative strength (ARS): – Model(s)-based– Quality Measures of Core Causal Components– Creating Indices– Integrating implementation assessments with models of effects (next Integrating implementation assessments with models of effects (next
week)week)• Strategy:
– Describe each step – Illustrate with an example– Planned Q&A segments
Caveat and Precondition• Caveat
– The black box (ITT model) is still #1 priority• Implementation fidelity and achieve relative
strength are supplemental to ITT-based results.
• Precondition– But, we consider implementation fidelity in
RCTs that are conducted on mature (enough) interventions
– That is, the intervention is stable enough to describe an underlying—
• Model/theory of change, and • Operational (logic and context) models.
Dimensions Intervention Fidelity
• Operative definitions:– True Fidelity = Adherence or compliance:
• Program components are delivered/used/received, as prescribed
• With a stated criteria for success or full adherence• The specification of these criteria is relatively rare
– Intervention Exposure:• Amount of program content, processes, activities
delivered/received by all participants (aka: receipt, responsiveness)
• This notion is most prevalent– Intervention Differentiation:
• The unique features of the intervention are distinguishable from other programs, including the control condition
• A unique application within RCTs
Linking Intervention Fidelity Assessment to Contemporary Models of Causality
• Rubin’s Causal Model:– True causal effect of X is (Yi
Tx – YiC)
– In RCTs, the difference between outcomes, on average, is the causal effect
• Fidelity assessment within RCTs also entails examining the difference between causal components in the intervention and control conditions.
• Differencing causal conditions can be characterized as achieved relative strength of the contrast. – Achieved Relative Strength (ARS) = tTx – tC
– ARS is a default index of fidelity
Achieved Relative Strength =.15
Infidelity
“Infidelity”
0.50d
85 700.50
30d
t c
pooled
Y Yd
sd
(85)-(70) = 15
tC
t tx
cY
tY
TTx
TC
.45
.40
.35
.30
.25
.20
.15
.10
.05
.00
Treatment Strength
with fidelity
with fidelity
90 650.83
30
T C
pooled
Y Yd
sd
d
Expected Relative Strength = (0.40-0.15) = 0.25
100
90
85
80
75
70
65
60
55
50
Outcome
TY
CY
Why is this Important?
• Statistical conclusion validity• Construct Validity:
– Which is the cause? (TTx - TC) or (tTx – tC)• Poor implementation: essential elements of the treatment are
incompletely implemented. • Contamination: The essential elements of the treatment group are
found in the control condition (to varying degrees).• Pre-existing similarities between T and C on intervention
components.
• External validity – generalization is about (tTx - tC)– This difference needs to be known for proper
generalization and future specification of the intervention components
So what is the cause? …The achieved relative difference in conditions across
components
0123456789
10
PD DifInst
Dif in Theory
Dif as Obs
0123456789
10
PD Asmt Dif Inst
C Planned
C Observ
Augmentation of Control
0123456789
10
PD Asmt Dif Inst
T Planned
T Obser
Infidelity
PD= Professional Development
Asmt=Formative Assessment
Diff Inst= Differentiated Instruction
In Practice….• Step 1: Identify core components in the
intervention group– e.g., via a Model of Change– Establish bench marks (if possible) for TTX and TC
• Step 2: Measure core components to derive tTx and tC
– e.g., via a “Logic model” based on Model of Change
• Step 3: Deriving indicators• Step 4: Indicators of IF and ARSI
Incorporated into the analysis of effects
Focused assessment is needed
What are the options?(1) Essential or core components (activities, processes);(2) Necessary, but not unique, activities, processes and structures (supporting the essential components of T); and(3) Ordinary features of the setting (shared with the control group)
• Focus on 1 and 2.
Step 1: Specifying Intervention Models
• Simple version of the question: What was intended?
• Interventions are generally multi-component, sequences of actions
• Mature-enough interventions are specifiable as:– Conceptual model of change – Intervention-specific model – Context-specific model
• Start with a specific example MAP RCT
Example –The Measuring Academic Progress (MAP) RCT
• The Northwest Evaluation Association (NWEA) developed the Measures of Academic Progress (MAP) program to enhance student achievement
• Used in 2000+ school districts, 17,500 schools
• No evidence of efficacy or effectiveness
Measures of Academic Progress (MAP): Model of Change
MAP Intervention:
4 days of training
On-demand consultation
Formative Testing
Student Reports
On-line resources
Formative Assessment
Differentiated Instruction
Achievement
Delivery – NWEA trainers
Receipt – Teachers and School Leaders
Enactment -- Teachers Outcomes -- Students
Implementation Issues:
Logic Model for MAP
Resources
Testing System
Multiple Assessment Reports
NWEA Trainers
NWEA Consultants
On-line teaching resources
Activities
4 training sessions
Follow-up Consultation
Access resources
Outputs
Use of Formative Assessment
Differentiated Instruction
Outcomes
State tests
MAP tests
Impacts
Improved Student Achievement
Program-specific implementation fidelity assessments: MAP only
Comparative implementation assessments: MAP and Non-MAP classes
Focus of implementation fidelity and achieved relative strength
Context-Specific Model: MAP
Acad
emic S
ched
ule
Fall Semester Spring Semester
Aug Sept Oct Nov Dec Jan Feb Mar Apr May
Major MAP Program Components and Activities
PD1 Con PD2 Con PD3 Con PD4 Con
Data
Sys
Use
Data
Diff Instr
Ch
ange
State Testing
Implementation
Two POINTS:
1. This tells us when assessments should be undertaken; and
2. Provides as basis for determining the length of the intervention study and the ultimate RCT design.
Step 2: Quality Measures of Core Components
• Measures of resources, activities, outputs
• Range from simple counts to sophisticated scaling of constructs
• Generally involves multiple methods
• Multiple indicators for each major component/activity
• Reliable scales (3-4 items per sub-scale)
Measuring Program-Specific Components
MAP Resources
Testing System
Multiple Assessment Reports
NWEA Trainers
NWEA Consultants
On-line teaching resources
MAP Activities
4 training sessions
Follow-up Consultation
Access resources
Criterion:
Present or Absent
Source or Method:
MAP Records
Criterion: Attendance
Source or method:
MAP records
Criterion:
Use
Source or Method:
Web-records
MAP Outputs
Measuring Outputs: Both MAP and Non-MAP Conditions
MAP Outputs
Use of Formative Assessment Data
Differentiated Instruction
Method:
End of year Teacher Survey
Methods:
End or year teacher survey
Observations (3)
Teacher Logs (10)
Criterion:
Achieved Relative Strength
Indices:
Difference in differentiated instruction (high v. low readiness students)
Proportion of observations segments with any differentiated instruction
Step 3: Indexing Fidelity and Achieved Relative Strength
True Fidelity – relative to a benchmark;
Intervention Exposure – amount of sessions, time, frequency
Achieved Relative Strength (ARS) Index
Standardized difference in fidelity index across Tx and C• Based on Hedges’ g (Hedges, 2007)• Corrected for clustering in the classroom
Tx C ARS Index
S
t t
Calculating ARSI When There Are Multiple Components
0123456789
10
PD DifInst
Dif in Theory
Dif as Obs
0123456789
10
PD Asmt Dif Inst
C Planned
C Observ
Augmentation of Control
0123456789
10
PD Asmt Dif Inst
T Planned
T Obser
Infidelity
PD= Professional Development
Asmt=Formative Assessment
Diff Inst= Differentiated Instruction
3 2.50.25
26 3.5
0.833
7 40.86
3.5
.25(.25) .33(.83) .42(.86) 0.69
E Ct t
PD
Assess
DI
Weighted j j
X XARSI
Sd
ARSI
ARSI
ARSI w ARSI
3 76%U
Weighted Achieved Relative Strength
Preliminary Conclusions for the MAP Implementation Assessment
• The developer (NWEA) – Complete implementation of resources, training, and
consultation• Teachers: Program-specific implementation
outcomes– Variable attendance at training and use of training
sessions– Moderate use of data, differentiation activities
services– Training extended through May, 2009
• Teachers: Achieved Relative Strength– No between-group differences in enactment of
differentiated instruction
Step 4: Indexing Cause-Effect Linkage
• Analysis Type 1:– Congruity of Cause-Effect in ITT analyses
• Effect = Average difference on outcomes ES• Cause = Average difference in causal components ARS
(Achieved Relative Strength) • Descriptive reporting of each, separately
• Analysis Type 2:– Variation in implementation fidelity linked to variation
in outcomes– Hierarchy of approaches (ITT LATE/CACE
Regression Descriptive)• TO BE CONTINUED ……