Incoming and Outgoing Shipments in 1 STEP Using Odoo 17
Â
PDE Week 5: Developing an Evaluation Plan
1. Putting It All Together:
Producing and Implementing
an Evaluation Plan
2. Overview*
I. Introduction and Key Considerations
II. What Should Be Included in an Evaluation Plan
III. Developing Tangible Plans for Evaluating Your Programs
IV. Implementing the Plan
V. Reporting the Results
VI. Pitfalls to Avoid and Things to Remember
VII. Using the Evaluation Report to Improve Your Programs
*The following content is based on W.K. Kellogg Foundation Evaluation Handbook (2004), Basic Guide to
Program Evaluation by Carter McNamara and The Program Managerâs Guide to Evaluation, Second Edition,
the U.S. Department of Health and Human Services (2001).
4. Why Do We Need an Evaluation Plan?
Understanding the Functions of an Evaluation Plan
An evaluation plan will:
⢠Identify the purpose, users, resources and timelines of the evaluation
⢠Select the key evaluation questions and indicators and the best design to
measure intended results
⢠Sequence evaluation activities (such as completing baseline
documentation, conducting participant follow-ups, materials pre-tests,
project reviews and special studies)
⢠Prepare data collection and data analysis plans, including cost as well as
results or program data
⢠Plan for communication, dissemination and use of evaluation results
⢠Identify the technical competencies needed on the evaluation team(s).
Excerpted from http://www.prime2.org/sst/step9-1.html
5. You need to take a few more steps in order to pull everything
together and develop a tangible plan for evaluating your
programs
⢠Have you completed the logic model?
⢠Have you determined what questions need
to be answered ?
⢠Have you decided on the types of
information you need to collect and data
collection methods?
Are You Ready?
6. Key Considerations
When designing a program evaluation it is
important to consider the following questions:
⢠For what purpose is the evaluation being done, i.e., what do you
want to be able to decide as a result of the evaluation?
⢠What kinds of information are needed to make the decision you
need to make?
⢠From what sources should the information be collected?
⢠How can that information be collected in a reasonable fashion?
⢠When is the information needed (so, by when must it be
collected)?
⢠What resources are available to collect the information?
7. What is an Evaluation Plan?
An evaluation plan is a written document that
specifies the evaluation design and details the
procedures for conducting the evaluation.
9. Typically an evaluation plan includes:
⢠Executive Summary
⢠Program objectives and outcomes
⢠Evaluation questions
⢠Types of data/information that need to be collected
⢠Data collection methods
⢠Timeframe for the evaluation (when collection of evaluation
information will begin and end)
⢠Staff requirements (who will be involved in the evaluation process)
⢠Budgeting requirements
⢠Attachments (surveys, interviews and other relevant information)
11. Steps in Developing an Evaluation Plan
1. Identifying Stakeholders and
Establishing an Evaluation Team
Why?
To ensure that you have gathered multiple
perspectives about the issues that require evaluation.
Who are the stakeholders?
Stakeholders include funders, project staff
and administrators, project participants
community leaders, and others with a direct,
or indirect, interest in program effectiveness.
12. Steps in Developing an Evaluation Plan
2. Budgeting for an Evaluation
Why?
To ensure that an organization has resources,
including time and money,
to invest in an evaluation
How much to spend on an evaluation ?
An evaluation typically costs 5% to 10% of a projectâs cost.
What expenses to include in the budget?
Evaluation staff salary and benefits; Consultant fees; Travel expenses for
staff and/or evaluators; Communications; Printing; Costs of acquiring data
collection instruments and library materials; Supplies and equipment
13. Steps in Developing an Evaluation Plan
3. Establishing Procedures for Managing
and Monitoring the Evaluation Process
Why?
To ensure consistency, confidentiality and accuracy in the process.
What kind of procedures?
- Appointing (hiring) and training staff that are to be responsible for an
evaluation;
- Deciding where the data that needs to be collected will be stored;
- Developing a data collection manual;
- Developing an evaluation timeline with key deadlines and responsible parties.
14. Create a Flexible and Responsive
Evaluation Design
Evaluations must be carefully designed
if they are to strengthen project
activities.
The evaluation design should avoid procedures that
require inhibiting controls.
Rather, the design should permit redirection and
revision as appropriate.
15. Create a Flexible and Responsive
Evaluation Design, contd.
Your evaluation design is flexible if it:
⢠âFitsâ the needs of the target populations and other stakeholders
⢠Produces data relevant to specific questions and project needs
⢠Allows revising evaluation questions and plans as project conditions
change
⢠Is being sensitive to cultural issues in the community
⢠Fits into the constraints of available resources and allows requesting
additional resources if necessary.
18. Conducting Data Collection and Analysis:
What is Involved?
⢠Gathering and organizing data in a systematic way to reduce
sources of bias and increase validity
⢠Discerning patterns, trends and comparisons from qualitative
and quantitative data
⢠Involving client and stakeholders in interpreting data
⢠Employing standards to arrive at conclusions
Source: http://www.prime2.org/sst/step9-2.html
19. Data Collection and Analysis
⢠Collect only the data you will use and that are relevant to
your
evaluation questions and purposes.
⢠Involve all staff involved in the data-collection phase in up-
front question formation.
⢠Revise data-collection strategies based on initial analysis.
What is working? What is not working? What pieces of data
are still missing?
⢠Base changes to existing tracking/data-collection strategies
on what is learned from evaluation.
21. The Evaluation Report
The report should present findings so an audience can
clearly see:
⢠Changes in performance
⢠How these changes can be attributed to the interventions
⢠Cost of the interventions
⢠If the evaluation design warrants, the report should also
present the effects (if any) of alternative interventions or the
absence of interventions in control areas and discuss
differences between those areas and the âcaseâ area.
Source: http://www.prime2.org/sst/step9-3.html
22. Communicating Your Evaluation Results
The Evaluation Report Includes:
⢠Purpose
⢠Methodology
⢠Findings
⢠Conclusions
⢠Recommendations
24. Pitfalls to Avoid
1. Assuming that the program is the only cause of positive changes
documented. Several factors may be responsible for changes in
participants or in a community
2. Forgetting that the same evaluation method may give different
results when used by different people.
3. Choose groups to compare that are different in too many ways. For
example, gender, age, race, economic status, and many other
factors can all have an impact on project outcomes.
4. Claiming that the results of a small-scale evaluation also apply to a
wide group or geographic area.
25. Pitfalls to Avoid
Carter McNamaraâs Basic Guide to Program Evaluation
5. Balking at evaluation because it seems far too "scientific.â Usually
the first 20% of effort will generate the first 80% of the plan.
6. Failing to include some interviews in your evaluation methods.
Questionnaires don't capture the story and the story is usually the most
powerful depiction of the benefits of your services.
7. Denying you can learn a great deal about the program by
understanding its failures, dropouts, etc.
26. Pitfalls to Avoid
8. Throwing away evaluation results once a report
has been generated. Results can provide precious
information later when trying to understand
changes in the program.
27. Things to Remember
âRemember to âcollect only the information you
are going to use, and use all the information you
collect.ââ
W.K. Kellogg Foundation Evaluation Handbook (2004), p. 99
28. Things to Remember
âAn evaluation report that sits on someoneâs shelf will
not lead us to improved program design and
management. Effective program evaluation supports
action.â
W.K. Kellogg Foundation Evaluation Handbook (2004), p. 99
29. Things to Remember
âThere is no "perfect" evaluation design. Don't worry
about the plan being perfect. It's far more important
to do something, than to wait until every last detail
has been tested.â
Carter McNamaraâs Basic Guide to
Program Evaluation
30. VII. Using the evaluation report to
improve your programs
31. Using The Evaluation Results
The Evaluation Report should be used to:
⢠Identify strengths and weaknesses of your program or
provide strategies for continuous improvement
⢠Discover new knowledge about effective practice
⢠Improve communication and shared understanding
between different stakeholders involved in the program
⢠Strengthen the organization's position in the community
⢠Enhance the overall capacity of the organization.