Background and History
EDN 568 Program Design and Evaluation
The Birth of Evaluation
In the beginning,
God created Heaven and Earth… And God saw everything he had made.
“Behold,” God said, “it is very good.”
And the evening and the morning were the sixth day.
And on the seventh day, God rested from all his work.
And on the seventh day, God’s ArchAngel came unto Him asking,
“God, how do you know what you created is ‘good?’
What are your criteria?
On what data do you base your judgment?
Just exactly what results were you expecting to attain?
And God, aren’t you a little too close to the situation to make a fair and unbiased evaluation?”
God thought very carefully about these questions the ArchAngel had asked him all that day…and His rest was greatly disturbed.
So on the eighth day God rose up, faced the ArchAngel, and said…
Well, Lucifer…
“YOU CAN JUST GO TO HELL!”
And thus, evaluation was born in a blaze of glory
• Evaluation is a tool which can be used to help judge whether a school program or instructional approach is being implemented as planned
• It allows us to assess the extent to which our goals and objectives are being achieved.
So, now that we know how evaluation was conceived, What do we need to know about it?
Helps Answer These Questions
• Are we doing the things for our students and teachers that we said we would?
• Are our students learning what we set out to teach them?
• How can we make improvements to the curriculum, teaching methods, and school programs?
Evaluation Myths
• Many people believed evaluation was a useless activity that generates lots of boring data with useless conclusions.
• This was a problem with evaluations in the past when program evaluation methods were chosen largely on the basis of achieving complete scientific accuracy, reliability and validity.
• This approach often generated extensive data from which very carefully chosen conclusions were drawn.
• Many people used to believe that evaluation was about proving the success or failure of a program.
• This myth assumed that success is implementing the perfect program and never having to deal with it again- the program will now run itself perfectly.
• This doesn't happen in real life. • Success is remaining open to continuing
feedback and adjusting the program accordingly.
• Evaluation gives you this continuing feedback.
• Many people believed that evaluation was a highly unique and complex process that occurred at a certain time in a certain way, and almost always had to include the use of outside experts.
• Many people also believed they must always demonstrate validity and reliability.
Evaluation Myths
Evaluation Myths
• For many years, people did not make generalizations & recommendations following program evaluations.
• As a result, evaluation reports usually stated the obvious and left program administrators disappointed and confused.
• Thanks to Michael Patton's development of utilization assessment, program evaluation has focused on utility, relevance and practicality.
Qualitative Case Study Analysis
Patton, M. Q. (2002). Qualitative research and evaluation methods (3rd ed.). Thousand Oaks, CA: Sage.
The Guru of Qualitative Program Evaluation & Analysis
NCLB & Qualitative Studies Case study research analysis allows us to
measure the extent to which the educational performance and the future potential for students has improved.
Case study analysis is context specific and therefore a “good fit” for educational program evaluation (when coupled with quantitative data).
Quantitative versus Qualitative Quantitative Evaluation is based the principles of
the scientific method. The goal to “prove”
something based on valid and reliable outcome
measures.
Qualitative Evaluation is context-based and
analyzes the comprehensive program impact
effects on participants.
Program Evaluation Defined
• Public school programs are organized methods to provide certain related services to students and other educational stakeholders
• Public School programs must be evaluated to decide if the programs are indeed useful for our stakeholders.
Program Evaluation Defined
• Program evaluation is a process by which society learns about itself.
• Program evaluations contribute to enlightened discussion of alternative plans for social action.
• Program evaluation must also be as much about political interaction as it is about determining facts.
The Politics of Program Evaluation
It very important to remember in debates over controversial educational programsthat sometimes liars figure & figures lie.
The educational program evaluator has the responsibility to protect students and schools from both types of deception.
The Politics of Educational Program Evaluation
The evaluator is an educator; his or her success is to
be judged by what others learn.
Those who shape policy should reach decisions with
their eyes wide open…not half shut.
It is the evaluator's task to illuminate the situation,
not to dictate the decision.
How Did Program Evaluation Developed Into What is Today?
History of Program Evaluation
During WW II, Program Evaluation developed at light speed to monitor soldier morale, evaluate personnel policies, and develop propaganda techniques.
The 1950s
Following the evaluation boom after WWII, program evaluation became a common tool for measuring all kinds of things in organizations.
Investigators assessed medical treatments, housing programs, educational activities, etc.
The 1960s
Dramatic increase in articles and books on program evaluation in the 60s.
President Johnson’s War on Poverty and Great Society Initiative provided a need to evaluate new social programs throughout the country.
The 1970s
By the 70s, evaluation research became a specialty field in all the social sciences.
The first journal in evaluation, Evaluation Review, began in 1976.
After that we were well on our way into the 1980s,1990s, and the 21st Century to….
EDN 568
Program Design & Evaluation Surfing Data &Evaluating Programs
Understanding Program Design and Evaluation
In the words of Stephen Covey, first things first….
1. What is a program?
2. What is program evaluation?
What Exactly Is a Program?
• Organizations usually try to identify several goals which must be reached to accomplish their mission.
• In nonprofit organizations, each of these goals often becomes a program.
Organizations & Program Evaluation
If programs represent the goals of an organization,
then there are two (2) basic ingredients required
for program evaluation:
1. An organization &
2. A program (goal)
Program Evaluation
• An assessment of an organization’s goals
• An assessment of the worth, value, or merit of an intervention, innovation, or service
• An assessment that identifies change – past and future
• An assessment that tries to solve a problem
• An assessment that carefully collects information about a program or some aspect of a program in order to make necessary decisions about the program.
Purpose & Uses of Program Evaluation
Evaluations of educational programs have expanded considerably over the past 40 years.
Title I of the Elementary and Secondary Education Act (ESEA) of 1965 represented the first major piece of federal social legislation that included a mandate for evaluation
Title I Program Evaluation Legislation
The 1965 legislation was passed with the evaluation requirement stated in very general language. This created quite a bit of controversy
State and local school systems were allowed considerable flexibility for interpretation and discretion.
Title I Evaluation Act
This evaluation requirement had two purposes:
(1) to ensure that the funds were being used to address the needs of disadvantaged children; and
(2) to provide information that would empower parents and communities to push for better education.
Title I Program Evaluation: A Means to Upgrade Schools
Many people saw the use of evaluation information on Title I programs and their effectiveness as a means of encouraging schools to improve performance.
Federal staff in the U.S. HEW welcomed the opportunity to have information about programs, populations served, and educational strategies used.
The Secretary of HEW in 1965 promoted the evaluation requirement as a means of finding out "what works" as a first step to promoting the dissemination of effective practices.
1965 Viewpoints on Title I Program Evaluation
• Expectation of reform and the view that evaluation was central to the development of change.
• A common assumption that evaluation activities would generate objective, reliable, and useful reports, and that findings would be used as the basis of decision-making and improvement.
What Happened?
• Unfortunately, the Title I Program Evaluations did not produce any of the intended results due to local schools’ failing to comply with the mandate
• Widespread support for evaluation did not take place at the local level.
• There was a concern that federal requirements for reporting would eventually lead to more federal control over schooling
The 1970s & Title I Program Evaluation
• It became clear in the 70s that the evaluation requirements in federal education legislation were not generating their desired results.
• The reauthorization of Title I of the federal Elementary and Secondary Education Act (ESEA) in 1974 strengthened the requirement for collecting information and reporting data by local grantees.
• It required the U.S. Office of Education to develop evaluation standards and models for state and local agencies.
• It also required the Office to provide technical assistance nationwide so exemplary programs could be identified and evaluation results could be disseminated.
TIERSThe Title I Evaluation and Reporting System (TIERS) was part of a continuing development effort to improve accountability and program services.
In 1980, the U.S. Department of Education created general administration regulations which established criteria for judging evaluation components of grant applications. These various changes in legislation and regulation reflected a continuing federal interest in evaluation data.
Use of Evaluation ResultsA system at the federal, state, or local levels was not in place to collect, analyze, and use the evaluation results to effect program or project improvements.
In 1988, amendments to ESEA reauthorized the Title I program, and strengthened the emphasis on evaluation and local program improvement.
1988 AmendmentsThe legislation required that state agencies identify programs:
1. That did not show aggregate student achievement gains2. That did not make substantial progress toward the goals
set by the local school district.
Those programs that were identified as needing improvementwere required to write program improvement plans. If, after one year, improvement was not sufficient, then the stateagency was required to work with the local program to developa program improvement process to raise student achievement
The 1990sThere was a further call for school reform, improvement, and accountability in the 1990s. The National Education Goals were formalized through the Educate America Act of 1994.
The new law issued a call for "world class" standards, assessment and accountability to challenge the nation's educators, parents, and students
2001 No Child Left Behind Act
The Elementary and Secondary Education Act (ESEA), renamed "No Child Left Behind" (NCLB) in 2001, established goals of high standards, accountability for all, and the belief that all children can learn, regardless of their background or ability.
NCLB
A Law of Accountability for Public Schools
On January 8, 2002, President Bush signed into law the No Child Left Behind (NCLB) Act, the most sweeping education reform legislation in decades
No Child Left Behind A New Level of Educational Program Evaluation
All educational programs evaluated on basis of student achievement and school performance measures
An almost totally “quantitative” evaluation of public schools.
Purpose of NCLB
To ensure that allall children have fair and equal opportunities to reach reach proficiency proficiency on state academic achievement standards.
Note emphasis on the word “opportunities.”
NCLB and Program Evaluation
The goal, worth, merit, and value of public schools shall be determined by student achievement and school performance measures.
A series of performance targets that states, school districts, and schools must achieve each year to meet the proficiency requirements.
To “meet” AYP, schools must be making adequate yearly progress towards the 2013/2014 NCLB goal.
NC AYP Proficiency Targets
Year Reading
3 - 8
Math
3 - 8
Reading/LA
Grade 10
Math
Grade 10
2005-2006 76.7% 81.0 % 35.4 % 70.8 %
2006-2007 76.7 % 81.0 % 35.4 % 70.8 %
2007-2008 84.4 % 87.3 % 56.9 % 80.5 %
2008-2009 84.4 % 87.3 % 56.9 % 80.5 %
2009-2010 84.4 % 87.3 % 56.9 % 80.5 %
2010-2011 92.2 % 93.7 % 78.4 % 90.2 %
AYP and Proficiency
• The ultimate goal of NCLB is to bring all students to PROFICIENCY (Level III or Above) as defined by North Carolina, no later than 2013-14.
• For the purpose of school, district and state accountability, the interim benchmark for progressing toward the goal is Adequate Yearly Progress (AYP) in raising student achievement.
Target Groups and Subgroups
• AYP focuses on all students and sub-groups of students in schools, school districts, and states, with a goal of closing achievement gaps and increasing proficiency to 100 percent.
• Ten student groups defined in NC public schools.
Ten Student Subgroups in NC1. School as a whole (all students)2. American Indian3. Asian4. Black5. Hispanic6. Multi-racial7. White8. Economically Disadvantaged (Free & Reduced Lunch)9. Limited English Proficient (LEP)10. Students with Disabilities (SWD)
Program Evaluation & NCLB
• All program evaluation (until further notice) must include proficiency achievement, school safety, and teacher quality goals and objectives.
• Regardless of the type of program evaluation conducted, NCLB requires program evaluators to include these standards and expectations.
Definitions & Concepts
Program evaluation has a specific language relative to terms, vocabulary, concepts, standards, and strategies. It is very important to have an adequate content knowledge before creating, implementing, and reporting program evaluations.
Important Program Evaluation Content Knowledge
Programs – an organization’s goalsProgram Evaluation’s Basic Ingredients – two (2)
1. An organization
2. A program Program Theory – Rationale, logic, and assumptions about the relationship of the program to expected outcomes.Inputs – Resources to run a program evaluationProcess – How a program evaluation is implementedOutputs – Units of program services
Outcomes – Benefits, changes, and effects on program participants Formative/Process Evaluation – An evaluation conducted to improve processes and
verify the relationship between activities and outcomes.Summative/Outcome-Based Evaluation - An evaluation to determine if goals and objectives were met, and the benefits, changes, and effects on program participants
Research – The scientific method of determining causality and validity.Program Evaluation – A method of determining worth, value, & decision making.
Important Program Evaluation Content Knowledge (cont.)
Needs Assessment - The process of diagnosing, identifying, describing, and
framing the nature, magnitude, & extent of a program’s purpose, goals, & activities.
Stakeholders - Individuals, groups, or organizations with an interest in how well a
program works.
Evaluation Sponsor - Person, group, or funding agency requesting a program
evaluation
Program Evaluation Standards – Utility, Feasibility, Propriety, & Accuracy
Utility Standard - serve the information needs of the intended users
Feasibility Standard – be realistic, prudent, diplomatic, and frugal
Propriety Standard – have legal, ethical, and due regard for program participants
Accuracy Standard - reveal and convey information to determine the worth, value, and merit of the program evaluated
Metaevaluation – The process of delineating, obtaining, and applying descriptive & judgmental information about the utility, feasibility, propriety, & accuracy of an evaluation.
Important Program Evaluation Content Knowledge (cont.)Program Evaluation Principles
1. Systematic Inquiry
2. Competence,
3. Integrity/Honesty
4. Respect for People
5. Responsibilities for General & Public Welfare
Educational Program Evaluation’s Six Steps
1. Define the Purpose & Engage Stakeholders
2. Describe & Specify Evaluation Questions, Focus Design
3. Develop Evaluation Methodology, Data Collection Plan
4. Collect the Data, Gather Evidence
5. Analyze the Data & Justify Conclusions
6. Ensure Use of Evaluation Report & Lessons for School Improvement
Program EvaluationSix Steps
1. Defining Purpose & Engaging Stakeholders2. Describe Program, Specify Evaluation Questions, Focus Design3. Develop Evaluation Methodology, Data Collection Plan4. Collect the Data, Gather Evidence5. Analyze the Data & Justify Conclusions6. Ensure Use of Evaluation Report & Lessons for School Improvement
1. Defining Purpose & Engaging Stakeholders
• Stakeholders must be engaged in the inquiry to ensure that their perspectives are understood.
• When stakeholders are not engaged, evaluation findings might be ignored, criticized, or resisted because they do not address the stakeholders' questions or values.
• Stakeholders must have a voice in determining purpose.
• After becoming involved, stakeholders help to execute the other steps
2. Describe Program, Specify Evaluation Questions, & Focus Design
• Needs Statement - describes the problem or opportunity that the program addresses and implies how the program will respond.
• Program descriptions convey the mission and objectives of the program being evaluated
• Descriptions should be sufficiently detailed to ensure understanding of program goals and strategies.
• Purpose, inputs, processes, output, & outcomes• Having a clear design that is focused on use helps persons who
will conduct the evaluation to know precisely who will do what with the findings and who will benefit from being a part of the program evaluation.
3. Develop Evaluation Methodology & Data Collection Plan
• Principals should strive to develop evaluation methods that will convey a well-rounded picture of the program and be seen as credible by the school’s stakeholders.
• Typically, qualitative assessments, along with academic quantitative measurements, are the preferred evaluation methodology.
• Information & evidence should be perceived by stakeholders as believable and relevant to increasing student achievement and overall school improvement.
• The data collection plan should align with the SMART Goals identified in the program’s purpose and questions to be answered.
4. Collect Data and Gather Evidence
1. Sources: Sources of evidence in a qualitative program evaluation are the persons, documents, observations, and test scores.
2. Quality: Quality refers to the appropriateness and integrity of information used in an evaluation.
3. Quantity: Quantity refers to the amount of evidence gathered in an evaluation. The amount of information required should be estimated in advance, or where evolving processes are used, criteria should be set for
deciding when to stop collecting data 4. Logistics: encompass the methods, timing, and physical infrastructure for
gathering and handling evidence
5. Analyze the Data & Justify Conclusions
Analyzing & justifying conclusions on the basis of evidence includes thefollowing five elements:
1. Standards - reflects the values held by stakeholders and provide the basis for forming judgments concerning program performance.
2. Analysis and synthesis - methods for examining and summarizing an evaluation’s findings. They detect patterns in evidence, either by isolating important findings (analysis) or by combining sources of information to reach a larger understanding (synthesis)
3. Interpretation - the effort of figuring out what the findings mean and is part of the overall effort to make sense of the evidence gathered in an evaluation.
4. Judgments - statements concerning the merit, worth, or significance of the program. They are formed by comparing the findings and interpretations regarding the program against one or more selected standards.
5. Recommendations - are actions for consideration resulting from the evaluation
6. Ensure Use and Share Lessons Learned
• Feedback is the communication that occurs among all parties to the evaluation. Giving and receiving feedback creates an atmosphere of trust among stakeholders; it keeps an evaluation on track by letting those involved stay informed regarding how the evaluation is proceeding.
• Follow-up refers to the technical and emotional support that users need during the evaluation and after they receive evaluation findings
• Dissemination is the process of communicating either the procedures or the lessons learned from an evaluation to relevant audiences in a timely, unbiased and consistent fashion.
Program EvaluationSix Steps
1. Defining Purpose & Engaging Stakeholders
2. Describe Program, Specify Evaluation Questions, Focus Design
3. Develop Evaluation Methodology, Data Collection Plan
4. Collect the Data, Gather Evidence
5. Analyze the Data & Justify Conclusions
6. Ensure Use of Report & Lessons for School Improvement
Types of Program Evaluation
1. Formative Evaluation - provides information to improve a program or process.
2. Summative Evaluation - provides short-term effectiveness or long-term impact information to decide whether or not to adopt a product or process.
Formative evaluation implies that the results will be used in
the formation and revision process of an educational effort.
More narrative, qualitative methods are
especially useful in formative assessments.
Formative assessments are typically used in
the improvement of educational programs.
Formative Program EvaluationExamples
• Formative evaluation of new instructional materials would ideally be conducted with experts and selected target audience members prior to full-scale implementation.
• Expert review of content by instructional designers or subject-matter experts could provide useful information for modifying or revising selected strategies.
• Learner review is a process of determining if students can use the new materials, if they lack prerequisites, if they are motivated, and if they learn.
Formative/Process EvaluationFormative evaluations are conducted during program implementation in order to provide information that will strengthen or improve the program being studied.
Example – An out-of-school tutoring program or initiative.
Formative evaluation findings typically point to aspects of program implementation that can be improved for better results, like how services are provided, how staff are trained, or how leadership and staff decisions are made.
Summative/Outcome-Based EvaluationOn the opposite pole, summative evaluation is for the purpose of documentation and dissemination of outcomes.
It is used for reporting to stakeholders and granting agencies, producing self-reports for accreditation, and certifying the achievement of students
Because of its documentary nature and the fact that its purpose is to judge value, summative assessment is frequently more quantitative than formative assessment. (However, this is not a hard and fast rule.)
Summative/Outcome-Based Program Evaluation Examples
• It is important to specify what decisions will be made as a result of the evaluation, then develop a list of questions to be answered by the evaluation.
• You may want to know if students met your objectives, if the innovation was cost-effective, if the innovation was efficient in terms of time to completion, or if the innovation had any unexpected outcomes.
Formative/Process vs. Summative/Outcome-Based Program Evaluation
Formative/Process Program Evaluation examines and assesses the implementation and effectiveness of specific instructional activities in order to make adjustments or changes in those activities. Process program evaluations are formative assessments, although summative assessments can also included in process evaluations.
Summative/Outcome-Based Program Evaluation determines what benefits (if any) there were for participants and any changes that occurred with the participants after program implementation. Summative/Outcome-based program evaluation is a summary assessment of the worth, value, merit, and benefits of a program to its participants.
Formative/Process Evaluation in Schools
The focus of formative/process evaluation could be a description and assessment of the curriculum, teaching methods used, staff experience and performance, in-service training, and adequacy of equipment and facilities.
The changes made as a result of process evaluation may involve immediate small adjustments (e.g., a change in how one particular curriculum unit is presented), minor changes in design (e.g., a change in how aides are assigned to classrooms), or major design changes (e.g., dropping the use of ability grouping in classrooms).
Formative/Process Evaluation Theory
• Formative/Process evaluation occurs on a continuous basis. • At an informal level, whenever a teacher talks to another
teacher or an administrator, they may be discussing adjustments to the curriculum or teaching methods.
• More formally, Formative/Process evaluation refers to a set of activities in which administrator and/or evaluators observe classroom activities and interact with teaching staff and/or students in order to define and communicate more effective ways of addressing curriculum goals.
• Formative/Process evaluation can be distinguished from outcome-based evaluation on the basis of the primary evaluation emphasis.
• Process evaluation is focused on a continuing series of decisions concerning program improvements, while outcome evaluation is focused on the effects of a program on its intended target audience (i.e., the students).
Formative/Process Program Evaluations
• Formative/Process Program Evaluations focus on understanding how the program really works, and its strengths and weaknesses
• Formative/Process Program Evaluations attempt to fully understand how a program works -- how does it produce that results that it does.
• These evaluations are useful if programs are long-standing, have changed over the years, students and/or teachers report a large number of complaints about the program, and there appear to be large inefficiencies in delivering program services.
Formative/Process Evaluation Questions What is required of students and teachers in this program?
What is the general process that students and teachers go through with the program or service?
What do students and teachers consider to be strengths of the instructional program or service?
What complaints are heard from students & teachers about the program or services?
What do students & teachers recommend to improve the program or services?
Summative/Outcome-Based Evaluation Also called “Impact Assessment”
A Summative/Outcomes-Based evaluation
seeks to determine if your organization is really
doing the right program activities to bring about
the outcomes you believe (or better yet, you've
verified) to be needed by students, rather
than just engaging in busy activities which
seem reasonable to do at the time.
Summative Outcome-Based Program Evaluation
Outcomes are benefits to students from participationin the program or support services. Outcomes areusually in terms of enhanced learning (knowledge, perceptions/attitudes or skills) or conditions, e.g., increased literacy, etc.
Outcomes are often confused with program outputsor units of services, e.g., the number of studentsWho went through a program.
Summative/Outcome-Based Evaluation ModelDesign Components
• Inputs include the resources dedicated to or consumed by the program.
• Process/Activities are what the program does with the inputs to fulfill its mission.
• Outputs are the direct products of program activities
• Outcomes – benefits, changes, and decisions produced by program or services
Inputs – Resources
• Examples are money, staff and staff time, volunteers and volunteer time, facilities, equipment, and supplies. For instance, inputs for a parent education class include the hours of staff time spent designing and delivering the program. Inputs also include constraints on the program, such as laws, regulations, and requirements for receipt of funding.
Process/Activities
How the program is run. How the strategies, techniques, and types of treatment that comprise the program's service methodology are implemented. Examples would be training students in study skills and counseling them to prepare for and find jobs.
Outputs – Units of Program Service
Measured in terms of the volume of work
accomplished - for example, the numbers of
classes taught, counseling sessions conducted,
educational materials distributed, and
participants served.
Outputs have little inherent value in themselves.
They are important because they are intended to
lead to a desired benefit for participants or target
populations.
Outcomes – Impact on Students
Outcomes are benefits or changes for individuals or populations during or after participating in program activities.
Outcomes may relate to behavior, skills, knowledge, attitudes, values, condition, or other attributes.
Outcomes are what students know, think, or can do; or how they behave; or what their condition is or how they are different following the program.
Summative/Outcome-Based Model of Evaluation
(Process)
Summative/Outcome-Based Program Evaluation Model
Summative/Outcome-Based Program Evaluation
Nature’s Example: The Mother Eagle’s intended outcome is
not how many worms she feeds her chick, but if and how well the young bird can fly
The extent to which program participants experience the intended benefits or changes.
Definition of Program EvaluationProgram evaluation is carefully collecting information about a program or some aspect of a program in order to make necessary decisions about the program.
Program evaluation can include any or a variety of at least 35 different types of evaluation, such as for needs assessments, accreditation, cost/benefit analysis, effectiveness, efficiency, formative, summative, goal-based, process, outcomes, etc.
We will be concerned primarily with Formative/Process & Summative/Outcome-Based Program Evaluation.
What is Program Design?Inputs, Process, Outputs, and Outcomes
Inputs are the various resources needed to run the program, e.g.,money, facilities, program staff, etc.
The Processes is how the program is carried out, e.g., studentsare counseled, supported, cared for, and taught (activities) .
The Outputs are the units of service, e.g., number of children tobe served, the number of teaches to receive staff development.
Outcomes are the impacts on the people receiving services, e.g., Increased mental health, safe and secure development
Developing a Program Design• Conducting a Needs Assessment
• Determining who is to be served
• Determining the purpose and the goals
• Defining inputs for implementation
• Creating implementation processes
• Identifying program outputs
• Determining anticipated & desired outcomes
Major Types of Program Evaluation Models
• Formative/Process-Based
• Summative/Outcomes-Based
Einstein
“Not everything that can be counted counts…and
not everything that counts can be counted.”
When conducting program evaluations, we must always remember the words of wisdom coined by a famous researcher and evaluator
Top Related