Session 853 Developing a Framework that Integrates Program Design, Management, Accountability &...
-
Upload
aileen-bridges -
Category
Documents
-
view
216 -
download
1
Transcript of Session 853 Developing a Framework that Integrates Program Design, Management, Accountability &...
Session 853Session 853
Extending Organizational Capacity and Capability to Evaluate Federal Environmental Research Programs
Developing a Framework that Integrates Developing a Framework that Integrates Program Design, Management, Accountability & Program Design, Management, Accountability &
EvaluationEvaluation
Dale Pahl* and Emma NorlandU.S. EPA Office of Research and Development
October 29, 2005
2005 Joint Conference: Crossing Borders, Crossing Boundaries Canada Evaluation Society & American Evaluation Association
*Corresponding author contact: [email protected]
2005 Joint Conference: Crossing Borders, Crossing Boundaries Canada Evaluation Society & American Evaluation Association
Pahl & Norland Oct’05
Presentation FocusPresentation Focus
Question: How can we develop a Question: How can we develop a framework thatframework that
Integrates program design, management, accountability, and evaluation?
Responds to OMB guidance about the Research & Development Investment Criteria?
Communicates clearly—to evaluators, clients, and external stakeholders—about the program’s environmental research and outcomes?
2005 Joint Conference: Crossing Borders, Crossing Boundaries Canada Evaluation Society & American Evaluation Association
Pahl & Norland Oct’05
Presentation FocusPresentation Focus
Response: Articulating the Response: Articulating the program program theorytheory for EPA’s environmental research for EPA’s environmental research creates a logical framework thatcreates a logical framework that . . . . . .
Integrates program design, management, accountability, & evaluation;
Engages research managers, clients, scientists, and stakeholders across the program’s scope and lifetime; and
Enables independent expert panels to evaluate evidence about program relevance, quality, performance, and leadership—with client input
2005 Joint Conference: Crossing Borders, Crossing Boundaries Canada Evaluation Society & American Evaluation Association
Pahl & Norland Oct’05
Communicating Program Theory for Environmental Communicating Program Theory for Environmental ResearchResearch Helps Integrate Design, Management, Helps Integrate Design, Management,
Accountability, & EvaluationAccountability, & Evaluation
Intermediate
Outcomes
Long-Term
Outcomes
MissionStrategic Goals & Objectives
Outcomes and Environmental Results
ClientsShort-Term
Outcomes
Effective Transfer
Clients
Client Decisions & Actions
Specific Organizations & Individuals
Research Topics & Activities
Research
Outputs
Resources
Research Program
Program Managers Program Managers Have Direct ControlHave Direct Control
Program Program Managers Have Managers Have
Direct Direct InfluenceInfluence
Agencies Have Agencies Have Indirect ImpactIndirect Impact
Clients use research (short-term outcomes) …
… e.g., to make environmental decisions
2005 Joint Conference: Crossing Borders, Crossing Boundaries Canada Evaluation Society & American Evaluation Association
Pahl & Norland Oct’05
Communicating Program Theory for Environmental Communicating Program Theory for Environmental ResearchResearch Helps Integrate Design, Management, Helps Integrate Design, Management,
Accountability, & EvaluationAccountability, & Evaluation
)
Specific Clients
Research Topics & Activities
Key Research
OutputsResources
Intended Changes in
Decisions or Actions by
Specific Clients
Intermediat
e Outcomes
(e.g., Improved
environmental quality, reduced human
exposure)
Long-Term Outcomes
(e.g., Improved human &
ecosystem health)
Short-term Outcomes
Programs are designed from RIGHT to LEFT
synthesis analysis
2005 Joint Conference: Crossing Borders, Crossing Boundaries Canada Evaluation Society & American Evaluation Association
Pahl & Norland Oct’05
Environmental Outcomes, Risk Assessment, & Environmental Outcomes, Risk Assessment, & AccountabilityAccountability Adapted from Presentations to EPA’s Board of Scientific Counselors by Hugh Tilson, Larry Cupitt, and John VandenbergAdapted from Presentations to EPA’s Board of Scientific Counselors by Hugh Tilson, Larry Cupitt, and John Vandenberg
Risk assessment helps identify & prioritize scientific questions & Risk assessment helps identify & prioritize scientific questions & knowledge gaps across a program’s environmental knowledge gaps across a program’s environmental outcomesoutcomes
Risk assessment is essential to help:Risk assessment is essential to help: Decide whether or not to take regulatory actionDecide whether or not to take regulatory action . . .
Is there an environmental hazard? Decide what actions are most effectiveDecide what actions are most effective . . .
What actions do we take to protect human health? Understand how to evaluate the effectiveness of our decisionsUnderstand how to evaluate the effectiveness of our decisions . . .
Were we effective?
AmbientConditions
Fate andTransport
SourceEmissions
ExposureAnd Dose
Early Signsof Effects
HealthImpacts
2005 Joint Conference: Crossing Borders, Crossing Boundaries Canada Evaluation Society & American Evaluation Association
Pahl & Norland Oct’05
Environmental Outcomes, Risk Assessment, & Environmental Outcomes, Risk Assessment, &
AccountabilityAccountability Adapted from Risk Assessment in the Federal Government: Managing the Process (NRC, 1983); 1997 Update to ORD’s Strategic Plan
(EPA, 1997); and OIG-ORD Presentation to EPA’s Deputy Administrator (Pahl & Norland, March 2002)
EXPOSURE RISK ASSESSMENT
RISKMANAGEMENT
3. Accountability Developing and measuring appropriate environmental
indicators demonstrates whether environmental decisions result in improved human and environmental health.
PollutionSources
Emissions
Transport &Transformation
Environmental Concentrations
Exposures Exposure Assessment
RiskCharacterization
Legal Considerations
Ecosystem and Human HealthConsiderations
Social, Economic,& PoliticalFactors
Exposure-Dose Relationships
Dose-ResponseAssessment
RiskManagementOptions
2. Implementation Decisions Managers
make decisions about how to implement, comply with & enforce regulations or how to remedy environmental problems.
Internal Dose
Biological Effect
Adverse Health Effect
Health Assessment
Hazard Identification
Dale Pahl Nov ‘02
1.Environmental Decisions and Regulations
2005 Joint Conference: Crossing Borders, Crossing Boundaries Canada Evaluation Society & American Evaluation Association
Pahl & Norland Oct’05
Short-Term Outcomes:Short-Term Outcomes: A Critical Link between Research & ImpactA Critical Link between Research & Impact
Intended Changes in Decisions
or Actions by
Specific Clients
Limitations or Gaps in Knowledge about the
Environmental Problem
Limitatio
ns or Gaps in
Attitudes
Limitations or Gaps in Skills and
Abilities Needed to
Respond to the
Environmental Problem
Transfer
(e.g., dissemination
thru publications)
Transfer
(e.g., guideline manuals & training)
Specific Clients
Short-Term Outcomes
(Research Contributions to Environmental Decisions)
Intermediat
e Outcomes
(e.g., improved
environmental quality, reduced human
exposure)
Long-Term Outcomes
(e.g., improved human &
ecosystem health)
Research Topics & Activities
Research Outputs
2005 Joint Conference: Crossing Borders, Crossing Boundaries Canada Evaluation Society & American Evaluation Association
Pahl & Norland Oct’05
Short-Term Outcomes: Short-Term Outcomes: A Critical Link between Research & ImpactA Critical Link between Research & Impact
Short-term outcomes are achieved when key research contributions are transferred to, and used by, intended clients
Research creates improved knowledge and applications for risk assessment, environmental decisions, regulatory decisions, and accountability:
• Setting or revising standards—formal EPA rulemaking
• Implementing standards—e.g., Regions or states develop plans to comply with standards, restore ecosystems, or manage environmental exposure & risk
• Applying environmental indicators to “measure” progress to achieve regulations and environmental outcomes—e.g., assessing whether legislation & regulations have the intended environmental impact
________1 See, for example: Strategic Research Plan for Particulate Matter, Air Quality Subcommittee of the Committee on the Environment and Natural resources (CENR), December, 2002; Science to Support Rulemaking, EPA Office of Inspector General (OIG) Report 2003-P-00003, November 15, 2002; and Air Quality Management in the United States, National Research Council, 2004.
2005 Joint Conference: Crossing Borders, Crossing Boundaries Canada Evaluation Society & American Evaluation Association
Pahl & Norland Oct’05
Communicating Program Theory for Environmental Communicating Program Theory for Environmental ResearchResearch Helps Integrate Design, Management, Helps Integrate Design, Management,
Accountability, & EvaluationAccountability, & Evaluation
)
Specific Clients
Research Topics & Activities
Key Research
OutputsResources
Intended Changes in
Decisions or Actions by
Specific Clients
Intermediat
e Outcomes
(e.g., improved
environmental quality, reduced human
exposure)
Long-Term Outcomes
(e.g., improved human &
ecosystem health)
Short-term Outcomes
Programs are designed from RIGHT to LEFT
synthesis analysis
2005 Joint Conference: Crossing Borders, Crossing Boundaries Canada Evaluation Society & American Evaluation Association
Pahl & Norland Oct’05
Organizing Research, Topics, & Organizing Research, Topics, & ActivitiesActivities
Both applied1 and use-inspired basic research2 contribute to regulatory decisions and improved understanding about environmental outcomes:
• Applied1 environmental research is targeted at understanding and solving particular environmental problems.
• Basic2 environmental research elucidates problems that involve complex environmental processes and nonlinear systems with multiple causes and effects.
• EPA’s basic research is “use-inspired” basic research2—targeted to answering questions about complex problems that cut across EPA’s programs to ensure that decisions are based on a foundation of sound science.
Typically, the distinction between these types of research is not clear cut.
__________ __________ 1 EPA's applied and basic research programs implement recommendations from the National Research Council in Building a Foundation for Sound Environmental Decisions, (NRC, 1997).
2 Stokes, D.E. “Renewing the Compact between Science and Government,” in 1995 Forum Proceedings, Vannevar Bush II—Science for the 21st Century. Pages 15-32. Sigma Xi, 1995.
2005 Joint Conference: Crossing Borders, Crossing Boundaries Canada Evaluation Society & American Evaluation Association
Pahl & Norland Oct’05
Organizing Research, Topics, & ActivitiesOrganizing Research, Topics, & ActivitiesAdapted from National Research Council, Building a Foundation for Sound Environmental Decisions (NRC, 1997).
Use risk assessment to rank issues and pinpoint largest uncertainties
Identify existing and emerging issues for a specific problem
Narrow EPA focus based on client needs and recognition of what
others are doing
Identify research topics that improve understanding, reduce uncertainties,
and develop client applications
Problem-DrivenResearch
Core Research
Select projects based on broad applicability, relevance to EPA,
and scientific merit.
Elucidation of Complex
Environmental Processes
Development of Tools
Collection of Data
FEEDBACK
R
e-ev
alu
ate
pri
ori
ties
re
gu
larl
y
2005 Joint Conference: Crossing Borders, Crossing Boundaries Canada Evaluation Society & American Evaluation Association
Pahl & Norland Oct’05
Organizing Research, Topics, & Organizing Research, Topics, & ActivitiesActivities
Question:
• How do we organize and “measure” research when there are no objective methods to:
Measure new knowledge as it develops? Manage the pace at which research progresses? Measure research quality and impact?
2005 Joint Conference: Crossing Borders, Crossing Boundaries Canada Evaluation Society & American Evaluation Association
Pahl & Norland Oct’05
Organizing Research, Topics, & Organizing Research, Topics, & ActivitiesActivities
Response:
• Organize the research with priority research topics• Assess research progress and priorities with
periodic meetings • Convene independent expert panels to evaluate
the improved knowledge and its applications with indicators that span a program’s scope & lifetime
__________1 For example, see Averch, H.A., “The Systematic Use of Expert Judgment,” pages 294-295 in Handbook of Practical Program Evaluation.
J.S. Wholey, H.P. Hatry, and K.E. Newcomer (eds). San Francisco, Jossey-Bass, 1994. 2 The Government Performance and Results Act: 1997 Governmentwide Implementation Will Be Uneven. U.S. GAO/GGD (1997)
2005 Joint Conference: Crossing Borders, Crossing Boundaries Canada Evaluation Society & American Evaluation Association
Pahl & Norland Oct’05
Developing Indicators Across a Program’s Scope & Developing Indicators Across a Program’s Scope & LifetimeLifetime
Potential Indicators: Potential Indicators: Evaluation CriteriaEvaluation Criteria Quality Relevance Performance Scientific
Leadership
Peer Review (multi-year research plan)
Peer Review (topic / project / activity)
Bibliometric Analysis (publications)
Client Feedback (client use of research)
Independent Expert Review (including client feedback)
Very useful Useful Limited usefulness
2005 Joint Conference: Crossing Borders, Crossing Boundaries Canada Evaluation Society & American Evaluation Association
Pahl & Norland Oct’05
The Importance of Independent Expert The Importance of Independent Expert EvaluationEvaluation
The systematic use of independent expert judgment is important1,2 when:
It is difficult to measure program progress or outputs (e.g., advances in research knowledge) needed to achieve outcomes
Multidisciplinary expertise is needed to evaluate scientific progress that responds to research topics and scientific questions
It is difficult to determine when outcomes can be attributed to the program
Agency programs involve research, regulation, or external partners such as state agencies
These criteria illustrate why independent expert These criteria illustrate why independent expert review is important for evaluating EPA’s research review is important for evaluating EPA’s research programsprograms
__________1 For example, see Averch, H.A., “The Systematic Use of Expert Judgment,” pages 294-295 in Handbook of Practical Program Evaluation. J.S. Wholey, H.P. Hatry, and K.E.
Newcomer (eds). San Francisco, Jossey-Bass, 1994. 2 The Government Performance and Results Act: 1997 Governmentwide Implementation Will Be Uneven. U.S. GAO/GGD (1997)
2005 Joint Conference: Crossing Borders, Crossing Boundaries Canada Evaluation Society & American Evaluation Association
Pahl & Norland Oct’05
Summary:Summary: EPA follows a systematic approachEPA follows a systematic approach1,2 1,2 to organize, to organize,
integrate, synthesize, and evaluate research integrate, synthesize, and evaluate research that informs environmental decisions by specific that informs environmental decisions by specific clients:clients:
• Develop an integrated risk assessment framework to synthesize available information about a specific environmental problem
• Assess client needs, knowledge gaps, research questions, and uncertainties• Create research topics needed to develop knowledge & evaluate evidence related to the
gaps, questions, and uncertainties• Monitor progress in implementing the research topics and in applying improved
knowledge base• Evaluate the improved knowledge base and client applications
The next slide illustrates how program theory The next slide illustrates how program theory creates a framework that integrates these creates a framework that integrates these objectives . . .objectives . . .
____________________1 For an example of EPA’s systematic approach, see Strategic Research Plan for Particulate Matter, Air Quality Subcommittee of the Committee on the Environment and
Natural resources (CENR), December, 2002.2For additional background information ,see Risk Assessment in the Federal Government: Managing the Process. National Research Council, 1983; and Environmental
Research and Development: Strengthening the Federal Infrastructure. Carnegie Commission on Science, Technology and Government, 1992.
2005 Joint Conference: Crossing Borders, Crossing Boundaries Canada Evaluation Society & American Evaluation Association
Pahl & Norland Oct’05
A risk-based assessment of knowledge gaps & needs related to EPA’s strategic goals—including strategic guidance from independent advisors such as the NAS
The number, sequence, and distribution of research projects / activities that respond to the priority research topics
Peer-reviewed publications that respond to the priority research topics, as measured by bibliometric analysis
The program integrates new knowledge and key research contributions for application at key decision-points (e.g., regulations) by specific clients
Client use of research knowledge and applications at key decision-points
The information-value or decision-value of the research knowledge or applications for understanding complex environmental processes
The information-value or decision-value of the research knowledge for developing indicators of public health or ecosystem health
Research Coordination Teams assess and prioritize client needs for research to inform policy, decision-making, and accountability
The number of research projects / activities completed during the past 5 years that respond to the priority topics
The information value of the publications (for example, in developing new knowledge or in reducing uncertainty) as measured by bibliometric analysis
The program collaborates with clients to transfer & demonstrate key research contributions
The information-value or decision-value of the research knowledge and applications used by clients
The information-value or decision-value of the research knowledge or applications for measuring environmental progress, developing risk mitigation approaches, or evaluating the effectiveness of regulations
The relevance of the program’s outcomes to EPA’s mission and legislative mandates
Research Coordination Teams select priority topics or questions to organize the research that leads to outcomes
The program ensures high quality research with merit-based competitive awards (e.g., through grant mechanisms) that respond to the priority research topics
Research Coordination Teams assess client feedback on key research contributions
Progress to achieve outcomes is measured with a small number of long-term goals and measures
The information-value or decision-value of the research knowledge or applications for helping EPA to achieve its strategic goals.
Partnerships provide FTE or $ to support priority research topics linked to EPA outcomes & strategic goals
The program sponsors periodic meetings to assess ongoing research progress and priorities
Scientific leadership combines several elements:
(1) The program has developed priority research topics that help EPA achieve its strategic goals;
(2) The program has developed partnerships that support & coordinate research which focuses on one or more of these topics; (3) Bibliometric analysis indicates that the program has made significant contributions that respond to the priority research topics; and (4) The program’s principal investigators are recognized leaders in their research disciplines
Evidence about program relevance, quality, performance, and leadership—evaluated by independent expert panels
Long-Term Outcomes
Environmental Outcomes and Strategic Goals
Resources &
Inputs
Research Topics & Activities
Research Outputs
Specific Clients
Short-Term Outcomes
WHAT?The right science
HOW? Programs are designed from RIGHT TO LEFT
WHY? The right questions