The following checklist provides a summary of the issues you should consider in evaluating a problem-solving effort. It should be interpreted as a general guide, and not as a set of rigid rules. This checklist is most helpful if used throughout the problem-solving process, beginning in the scanning stage.
I. Early Considerations
You should consider the following questions during the scanning, analysis and response stages.
A. What will the evaluation help you decide?
B. Do you know the problem? (You need to answer these questions with some precision to develop and evaluate a costeffective response. If you cannot answer them with some precision, then you should do more to analyze the problem.)
C. Do you know how the response works? (You need to answer these questions to determine if the response is likely to be effective, and to ensure accountability during implementation. If you cannot answer them, then your response plans are inadequate, and you need to focus more on the response stage.)
II. Process Evaluation
The process evaluation begins toward the end of the response stage, and continues well into the assessment stage.
A. Did you implement the response? (The closer the actual implementation is to the planned response, the greater confidence you have that the response caused the problem change documented in the impact evaluation. The more variation between what you intended and what occurred, the greater the likelihood that factors other than the response caused changes in the problem.)
B. Did you implement enough of the response? (You may have implemented the response, but without the resources, duration or intensity needed to make it effective.)
III. Impact Evaluation
Many of the decisions you need to make to conduct an impact evaluation should be considered in the analysis and response stages. This is particularly true of measurement decisions.
A. Do you need a control group? (Answering these questions helps you decide on the complexity of the evaluation design.)
B. How often can you measure the problem? (Answering these questions helps you to decide whether a time series design is possible.)
C. What type of evaluation design should you use? (Your answers to the questions in sections A and B, immediately above, provide some basic guidance for answering this question, as shown in Table E.1. Obviously, precise answers depend on the particular circumstances of each problemsolving effort.)
B. Question Checked | A. Question Checked | |
---|---|---|
1 | 2 | |
1 | Interrupted time series design | Multiple time series design |
2 | Pre-post design | Pre-post design with a control group |
3 | Combination of designs above | Combination of designs above |
Table E.1 Which Evaluation Design Makes the Most Sense? |
D. What type of control group do you need? (This question applies only if you chose one of the options from column 2 under "A. Question Checked" above. If you chose an option from column 1, then skip this section and go to part IV.)
If you checked one of the first four questions above, then systematically compare the response group's characteristics with the control group's characteristics, and list the major differences. In part V, you will consider whether other factors might have caused the change in the problem. Your list of differences is a list of potential "other factors."
IV. Evaluation Conclusions
The following questions fall within the assessment stage and are applicable once you have documented your evaluation results. These questions are designed to help you draw conclusions consistent with your process and impact evaluation results and your evaluation design. You will have to ask more questions than listed here to fully interpret your particular evaluation results.
A. What are your findings from the process evaluation?
B. What are your findings from the impact evaluation? (Select the design you used–pre-post, pre-post with a control group, time series, or multiple time series. If you used a combination of designs, then interpret your evaluation for each design separately, using tables E.2 and E.3.)
Pre-post design: Use Table E.2 to interpret your evaluation.
Pre-post design with a control group: Use Table E.3 to interpret your evaluation.
Time series design: Use Table E.3 to interpret your evaluation.
Multiple time series design: Use Table E.3 to interpret your evaluation.
V. Overall Impact Evaluation Conclusions
The answers to the following questions are judgment calls and reflect your degree of confidence in the findings, rather than a totally objective assessment of what occurred. Other people, examining the same evidence, could come to different conclusions. For this reason, you should answer these questions (and the question that follows) after several people with different perspectives have examined the assessment information.
Based on your answers to the preceding questions, are you reasonably confident that the response caused the decline (if any) in the problem?
Process Evaluation Results Answers to Question IV.A |
|||
---|---|---|---|
4 checked: You implemented the response almost as planned. | 1, 2 or 3 checked: You did not implement the response; implemented it in a radically different manner than planned; or implemented it with insufficient resources, for too short a time, or without the required intensity. | ||
Impact Evaluation Results Answers to Question IV.B (pre-post design) |
3 checked: The problem declined. | A. The response may or may not have caused the decline in the problem. Nevertheless, thedecline occurred. | C. This suggests that other factors may have caused the decline in the problem, or the response was accidentally effective. Nevertheless, the decline occurred. |
1 or 2 checked:The problem got worse or did not change. | B. The response does not seem to have worked, though it is possible the problem would have increased (or increased even more) without it. | D. You have learned little from this evaluation. It is unclear whether you should implement the planned response, or reanalyze the problem and try adifferent response. | |
Regardless of the interpretation (A, B, C, or D), you have insufficient evidence to link the response to the problem level. The impact evaluation results neither support nor rule out using the response for similar problems. |
|||
Table E.2 Interpreting Results of Process and Impact Evaluations (Pre-Post Designs) |
Process Evaluation Results Answers to Question IV.A |
|||
---|---|---|---|
4 checked: You implemented the response almost as planned. | 1, 2 or 3 checked: You did not implement the response; implemented it in a radically different manner than planned; or implemented it with insufficient resources, for too short a time, or without the required intensity. | ||
Impact Evaluation Results Answers to Question IV.B (pre-post design with a control group, time series design or multiple time series design) |
3 checked: The problem declined. | A. This is evidence that the response caused the decline in the problem. The response is a potentially useful option for similar problems. | C. This suggests that other factors may have caused the decline in the problem, or the response was accidentally effective. You should not recommend this response to address similar problems, since you do not know if it would have an impact. |
1 or 2 checked: The problem got worse or did not change. | B. This is evidence that the response was ineffective. The response probably should not be used for similar problems. You should reanalyze the problem and try a different response. | D. You have learned little from this evaluation. Perhaps if you had implemented the response as planned, you would have had better results, but this is speculative. No recommendations-either for or against the response-are valid. | |
Table E.3 Interpreting Results of Process and Impact Evaluations (Other Designs) |
You may order free bound copies in any of three ways:
Online: Department of Justice COPS Response Center
Email: askCopsRC@usdoj.gov
Phone: 800-421-6770 or 202-307-1480
Allow several days for delivery.
Send an e-mail with a link to this guide.
* required
Error sending email. Please review your enteries below.