OES develops resources to help agencies learn from leading evaluation practices.

Evaluation Working Group Case Study

GSA formed an Evaluation Working Group (EWG) in 2020 to provide feedback on the agency’s FY22 Annual Evaluation Plan and other Evidence Act deliverables. Meeting three times between April and June, the group helped develop criteria for significant evaluations and drafts of the Annual Evaluation Plan and GSA Evaluation Policy. The case study summary describes the EWG’s purpose and how they went about their work, as well as provides materials and agendas for forming similar groups at other agencies.
Evaluation Working Group Case Study
Appendix: Planning Materials for an Evaluation Working Group

Preregistration as a Tool for Strengthening Federal Evaluation

In order to ensure that evaluation findings are reliable and that statistical results are well founded, it is essential that evaluators commit to specific design choices and analytic methods in advance. By making these details publicly available – a practice known as preregistration – we promote transparency and reduce the risk of inadvertently tailoring methods to obtain certain results or selectively reporting positive results. This guidance paper describes the importance and benefits of preregistration and addresses concerns that Federal evaluators might have.
Preregistration handout

How to Use Unexpected and Null Results

Recent research shows that null results in federal evaluations are more common than we think, and occur for a variety of reasons. When agencies share both expected and unexpected results, we can learn about what programs work, what effect sizes are realistic, and improve Federal evaluations. This post dispels misconceptions about null results and highlights different uses and lessons from null results.
Unexpected results handout