Our library of practical resources is intended for researchers and research staff undertaking randomized evaluations, as well as those teaching the technique to others, and anyone interested in how randomized evaluations are conducted.
Incorporating lessons learned through our own experience and through guidance from researchers and research organizations, we provide practical advice for designing, implementing, and communicating about evaluations. These resources are a collaborative effort. We credit the authors of all the resources we post here, and link to their original work wherever possible.
Please reach out to us at [email protected] or fill out this form with questions or feedback.
Introduction to Randomized Evaluations
Resources
A non-technical overview and step-by-step introduction for those who are new to randomized evaluations, as well as case studies and other teaching resources.
Before Starting a Project
Resources
Tips on successful field management and implementation partnerships for researchers who are new to fieldwork.
Project Planning
Resources
Highlights include: annotated checklist for designing an informed consent process, detailed advice on grant proposals and budgeting, and suggestions for proactive measures to help ensure ethical principles are followed in research design and implementation.
Research Design
Resources
Data Collection and Access
Resources
This section contains guidance specific to working with surveyors or survey companies, information about administrative data collection, and information applicable to all modes of data collection, such as on data security, data quality, and grant management.
Processing and Analysis
Resources
All the steps in a research project after the data was collected or assembled, from data cleaning to communicating results.
Design and iterate implementation strategy
Implementing partners and researchers should work closely together during the study design phase of a randomized evaluation to create a feasible implementation strategy. This resource is intended to...
Élaborer et tester la stratégie de mise en œuvre
Pendant la phase de conception d’une évaluation aléatoire, les partenaires de mise en œuvre et les chercheurs doivent travailler en étroite collaboration afin d’élaborer une stratégie de mise en œuvre...
Evaluating technology-based interventions
This resource provides guidance for evaluations that use technology as a key part of the intervention being tested. Examples of such interventions might include automated alerts embedded into an...
Évaluer les interventions basées sur la technologie
Cette ressource donne des conseils pour l’évaluation des interventions dans le cadre desquelles la technologie joue un rôle clé. Ces interventions peuvent par exemple prendre la forme d’alertes...
Coding resources for randomized evaluations
This page compiles links to resources on software, user-written commands for randomized evaluations, coding in teams, and writing reproducible code. User-written commands listed below include common...
Ressources sur la programmation des évaluations aléatoires
Cette page rassemble des ressources sur les logiciels utiles pour la programmation, les commandes écrites par les utilisateurs pour les évaluations aléatoires, la programmation en équipe et la...
Pre-publication planning and proofing
This resource is intended for research teams who have drafted a paper with results from a randomized evaluation and are preparing to submit to a journal or publish a working paper.
Randomized...
Planification et relecture préalables à la publication
Cette ressource s’adresse aux équipes de recherche qui ont rédigé un article sur les résultats d’une évaluation aléatoire et qui s’apprêtent à le soumettre à une revue ou à publier un document de...