Indicative content of evaluation final reports Draft, Jérémie Toubkiss, Evaluation Office, NYHQ (updated 4 February 2016)

Similar documents
GEROS Evaluation Quality Assurance Tool Version

UNICEF-Adapted UNEG Quality Checklist for Evaluation Terms of Reference

Terms of reference Evaluator for mid-term review of 4.5-year EuropeAid Grant Agreement

Section & Overall Rating Highly Satisfactory Satisfactory Fair Unsatisfactory Missing Not Rated Weighting (Score) Implication

Technical Note Integrating Gender in WFP Evaluations

Guidance: Quality Criteria for Evaluation Reports

An independent review of ILO Global Supply Chains interventions undertaken between with a focus on lessons learned, what works and why

Terms of Reference (ToR) End-of-the Project Evaluation UNDP Support to the Strategic Capacity Building Initiative

Evaluation policy PURPOSE

Terms of Reference (ToR) End-of-the Programme Evaluation UNDP Support to Inclusive Participation in Governance May 2013

Terms of Reference. Final Evaluation ALERT Project

Evaluation Inception Report Draft Template

Evaluation of Environmental and Social Impact Management at ADA Terms of Reference

Mid-term Project Evaluation Guidance Note For EIF TIER 1 Funded Projects: Support to National Implementation Arrangements

UNICEF Global Evaluation Report Oversight System (GEROS) Review Template

Oxfam PNG - WASH Program Final Evaluation

TERMS OF REFERENCE. Independent Evaluation of the ILO Action Plan for Gender Equality

Haiti Earthquake 2010 Response and Recovery Programme Final Evaluation

Terms of Reference for Crisis Management and Recovery Programme Final Evaluation

TERMS OF REFERENCE FOR THE FINAL EVALUATION 1. BACKGROUND: ONE UN S PARTNERSHIP WITH THE GIHUNDWE HOSPITAL

WHO reform. WHO evaluation policy

UNICEF Global Evaluation Report Oversight System (GEROS) Review Template

Terms of Reference. Projects Outputs Evaluation

How to write a logframe

United Nations Development Programme South Sudan. Terms of Reference Review of Rapid Capacity Placement Initiative (RCPI) May 2012

GUIDELINES FOR INCEPTION REPORTS

SUBPROGRAMME EVALUATION: SUSTAINABLE DEVELOPMENT POLICIES DIVISION

Monitoring and Evaluation Policy

REQUEST FOR PROPOSALS

TERMS OF REFERENCE. The project intervention logicis is described as bellow: 1 P a g e

UNICEF Evaluation Office Terms of Reference: External Assessment of UNICEF s Global Evaluation Reports Oversight System (GEROS)

Guidelines for UNODC Evaluation Terms of Reference

THE DEPARTMENT OF NUTRITION FOR HEALTH AND DEVELOPMENT, WORLD HEALTH ORGANISATION APRIL 2015

PROCEDURE AND TOOLS FOR EVALUABILITY REVIEW OF ILO PROJECTS OVER US$5 MILLION

Evaluation purpose and scope The overall purpose of the evaluation, as laid out in the ToR, is to:

UNICEF/EAPRO, Bangkok

Terms of Reference (ToR)

UNITED NATIONS DEVELOPMENT PROGRAMME TERMS OF REFERENCE

Provision of Support Services: Office space: Yes x No Equipment (laptop etc): Yes x No Secretarial Services Yes x No. Signature of the Budget Owner:.

Mid-term Evaluation for the Gender and Generational Empowerment project in Tanzania.

Terms of Reference Independent Final Evaluation of Zambia Green Jobs Programme March 2018

TIPS PREPARING AN EVALUATION STATEMENT OF WORK ABOUT TIPS

End-Phase Project Evaluation Guidance Note. For EIF Tier I and Tier II-funded projects

TERMS OF REFERENCE FOR ENDLINE EVALUATION OF FOOD AND NUTRITION SECURITY PROJECT IN BORNO STATE, NIGERIA

Terms of Reference (ToR)

Terms of Reference: Project Final Evaluation

SEARCH FOR COMMON GROUND MOROCCO is seeking Midterm Evaluator [Consultant] Region Tangier - Tetouan, Morocco

Evaluation Policy for GEF Funded Projects

TERMS OF REFERENCE FOR THE FINAL EVALUATION OF HAKI MKONONI RIGHTS IN OUR HANDS

INDIVIDUAL CONSULTANT PROCUREMENT NOTICE / TERMS OF REFERENCE Description of the assignment: Final Evaluation National Consultant

TERMS OF REFERENCE FOR REVIEW OF CPAP OUTCOME KEN 44

EVALUATION STATEMENT OF WORK CHECKLIST AND REVIEW TEMPLATE

UNICEF Evaluation Office Evaluation of UNICEF s Policy Advocacy Work: Mapping and Scoping Exercise

American Red Cross Haiti Delegation 22, rue Metellus Petion Ville Port au Prince, Haiti

Evaluation Terms of Reference Template

Terms of Reference EXTERNAL EVALUATION ( ) OF FOKUS PROGRAMME WOMEN S ECONOMIC PARTICIPATION AND RIGHTS IN UGANDA

Terms of Reference. Senior Researcher / Study Consultant for the Educational Demand and Needs Assessment Study

Executing Organisations: SOCADIDO - Soroti Catholic Diocese Integrated Development Organisation Hoffnungszeichen Sign of Hope e.v.

Terms of Reference for Final Evaluation of Oxfam s Syria Crisis Response programme in Syria

TERMS OF REFERENCE. Location

GLOBAL EVALUATION REPORT OVERSIGHT SYSTEM

INDIVIDUAL CONSULTANT PROCUREMENT NOTICE National Consultant To support UNDAF Evaluation for Nepal

See Annex 2 for the terms of reference and Annex 3 regarding progress in the development of high-level indicators

UNITED NATIONS DEVELOPMENT PROGRAMME JOB DESCRIPTION

Training on Project Cycle Management for Researchers Ethiopian Academy of Sciences May 2016, Ghion Hotel, Addis Ababa

Terms of Reference (TOR)

Terms of Reference Final Evaluation

Terms of Reference (TOR)

PROJECT PREPARATION GRANT (PPG) PROJECT TYPE: Full-sized Project TYPE OF TRUST FUND: GEF Trust Fund

Terms of Reference Final Evaluation Open for business : Promoting equitable land rights protection in Sierra Leone, Liberia and Guinea

C-18: Checklist for Assessing USAID Evaluation Reports

(FY ) TERMS OF REFERENCE DRAFT [DATE]

TERMS OF REFERENCE. 1. Background on the project

MONITORING AND EVALUATIONS MANUAL PIDF SECRETARIAT

Terms of Reference for a Gender Analysis

CI-GEF PROJECT AGENCY MONITORING AND EVALUATION POLICY FOR GEF-FUNDED PROJECTS

UNICEF Lao PDR TERMS OF REFERENCE FOR SERVICES CONTRACT

Executive Board of the United Nations Development Programme and of the United Nations Population Fund

UNICEF Global Evaluation Report Oversight System (GEROS) Review Template

IAS EVALUATION POLICY

UNDAF MID-TERM REVIEW. TERMS OF REFERENCE FOR INTERNATIONAL CONSULTANT (Team Leader)

UNITED NATIONS CHILDREN'S FUND (UNICEF) CALL FOR INSTITUTIONAL EXPRESSIONS OF INTEREST

Consultancy Vacancy UNHCR Evaluation Service

Call for Expression of Interest

Country Portfolio Evaluations I. GUIDANCE FOR PROCESS & CONTENT

Monitoring and Evaluation Training Modules Table of Contents

CARE International Evaluation Policy 1

Terms of Reference (TOR) Evaluation of UN Volunteers intervention ( ) in the Tonle Sap Conservation Project

CONSULTANCY SERVICES FOR EVALUATION OF THE ECOWAS PROGRAMME ON GENDER MAINSTREAMING IN ENERGY ACCESS (ECOW-GEN) IN ECOWAS COUNTRIES

TERMS OF REFERENCE SOUTH AFRICA FINANCIAL SECTOR DEVELOPMENT AND REFORM PROGRAM INDEPENDENT PROGRAM EVALUATION. I. Background

TERMS OF REFERENCE (TOR) Title: Voices from the Underground: End-of-Project Evaluation Mozambique and South Africa

Frequently Asked Questions: UN-SWAP Evaluation Performance Indicator Reporting

Terms of Reference Mid-term Review (MTR) of Girls Advocacy Alliance (GAA) Project in Nepal

TERMS OF REFERENCE (TOR) CONSULTANCY SERVICE CALL FOR BASELINE SURVEY

UNITED NATIONS CHILDREN'S FUND (UNICEF) CALL FOR INSTITUTIONAL EXPRESSIONS OF INTEREST (EOI)

World Vision s Approach to Local Ownership in Evaluation. Korça, Albania Development Program a case study

UNLIREC TERMS OF REFERENCE

Terms of Reference (TOR)

Making the choice: Decentralized Evaluation or Review? (Orientation note) Draft for comments

Terms of Reference Final Evaluation I love my country : Strategic Communications for Peace building in South Sudan

Transcription:

Indicative content of evaluation final reports Draft, Jérémie Toubkiss, Evaluation Office, NYHQ (updated 4 February 2016) 0. Opening pages - The first pages provides the key information on the programme evaluation 0.1 Name and timeframe of the evaluated programme 0.2 Name of the country (and regions) 0.3 Type of evaluation (mid-term, final, etc.) 0.4 Name of the commissioning agency 0.5 Authors of the report (institution, names of individuals) 0.6 Date of the evaluation report 0.7 Status of the report (draft, final ) 0.8 Table of content 0.9 List of acronyms 0.10 List of tables and figures 0.11 Executive summary presenting only the following information that are intended to decision makers and senior management (max 5 pages, can stand-alone): 0.12 Brief description of the evaluated programme 0.13 Evaluation purpose, objectives, criteria, scope 0.14 Methodology (list of main data source, collection methods, triangulation approach, equity lens, ethical aspects, and main limitations) 0.15 2 to 3 key findings and the overall conclusion on the programme performance for each evaluation criteria 0.16 2 to 3 key lessons learnt 0.17 2 to 4 priority recommendations (urgent, strategic or easy-to-implement) by stakeholder 1. Context and objectives of the evaluation - The report includes sufficient and relevant contextual information, and explains relevance and usefulness of the evaluation. 1.1. Context of the evaluation and description of the evaluated programme 1.1.1 The ToR adequately describes the particular policy, social, economic, governance and programmatic environment of the country/region/area and of the programme/project/intervention to be evaluated, in a way that illuminates/contextualize the findings 1.1.2 Purpose and specific objectives of the programme/project/intervention to be evaluated, and its theory of change or logical framework, which should be well articulated (can be revised/updated by the evaluator) 1.1.3 Main assumptions/hypotheses and risks associated with the theory of change or logframe (based on the programme documents or on the evaluator's analysis) 1.1.4 Timeframe of the programme/project/intervention: starting and end dates (or implementation status); milestones in the project history; timing of the evaluation within the project management cycle 1.1.5 Geographical boudanies: number and name of regions and districts; rural/small towns/peri-urban/urban areas 1.1.6 Key stakeholders and their respective roles (UNICEF, government counterparts, implementing partners, service providers, beneficiaries) 1.1.7 Financial aspects (amount and origin of funds, future financing prospects...) 1.1.8 Existing knowledge and evidence on the programme/project/intervention (previous, on-going and planned evaluations & studies, incl. findings from previous evaluation or evaluations of similar interventions), knowledge gaps and added-value of this evaluation 1.1.9 Aspects of complexity or characteristics that can potentially make the evaluation work difficult, e.g. absence of logframe, of baseline, of M&E system, changes in program objectives, strategy or implementation over time, multiplicity of actors, multiplicity of program components, complexity of the policy, institutional, cultural or security context etc. 1.2. Purpose and objectives of the evaluation 1.2.1 Purpose of the evaluation (transparency & accountability; learning for decision making; advocacy etc.) 1.2.2 Specific objectives of the evaluation specified and coherent with purpose 1.2.3 Who has commissioned / initiatied the evaluation (UNICEF, donor, etc.) 1.2.4 Primary and secondary audiences for the evaluation 1.2.5 Why evaluation is being done at this time and how the evaluation will be useful (intended use, incl. what decisions might likely be influenced by the findings of this evaluation) 1

2. Methodology - The report adequately defines the criteria that have been utilized to guide the evaluation, the evaluation questions, the scope of the evaluation, the methodology used for data collection and analysis, and the overall organisation of the evaluation work. 2.1. Evaluation criteria Evaluation criteria used such as: relevance, efficiency, effectiveness (incl. cost-effectiveness or value for money), impact, and/or sustainability - and additional evaluation criteria if any (equity, 2.1.1 gender, disability, vulnerable / deprived segment of the population, human/children rights-based approach, results based management, robustness of the monitoring-evaluation-accountabilitylearning system, innovation, humanitarian principles, quality of process, exportability / potential for scaling up, added-value, coherence Why some standard evaluation criteria may not have been used 2.1.2 Definition of each evaluation criteria used 2.1.3 Description and justification of any change made to what was foreseen in the ToR 2.2. Evaluation questions 2.2.1 Final evaluation questions and by who and how they have been chosen and formulated (incl. any change made to the ones proposed in the initial ToR, for example knowledge gaps identified during the initial document review and recommendations drawn from previous evaluations & studies) 2.2.2 Sub-questions or hypothesis to be tested, if any 2.2.3 Description and justification of any change made to what was foreseen in the ToR 2.3. Scope of the evaluation 2.3.1 Components of the programme/project/intervention that have been evaluated (and what was excluded or impossible to evaluate) 2.3.2 Geographical coverage of the evaluation (with map comparing intervention areas with areas visited) 2.3.3 Time boundaries of the evaluation (or phase in the project) 2.3.4 Activities funded by specific donors (if relevant) 2.3.5 Stakeholders and beneficiaries included / excluded 2.3.6 Description and justification of any change made to what was foreseen in the ToR 2.4. Evaluation design, and methodology for data collection and analysis 2.4.1 Conceptual framework and evaluation design (non-experimental, quasi-experimental, experimental) 2.4.2 Overall methodology: quantitative, qualitative, participatory, self-evaluation where the evaluation team will play a facilitation role only, utilization-focused, theory-based, etc. 2.4.3 Data to be collected for each evaluation question Data source: M&E data generated by the evaluated programme; documents reviewed (number and list in annex) and their credibility, stakeholders consulted (program staff; implementing 2.4.4 partners such as gov, NGOs, and consultants; contractors such as construction companies and engineering firms; direct and indirect beneficiaries (specify any focus on sex, age, religion, ethnic group, professional occupation, poverty level, vulnerability status...); external partners or resource persons (other NGOs, donors, universities and researchers...) etc. 2.4.5 Triangulation plan: how key information have been triangulated 2.4.6 Methods used to collect each data from each data source: semi-structured interviews, focus-group discussions, problem tree/analysis, community mapping, document review, field observation, surveys, measures/tests, case studies / monography, etc. 2.4.7 Description and justification of any change made to what was foreseen in the ToR 2.4.8 Description of data collection instruments used (actual tools in annex) 2.4.9 Sample size and sampling method (for key informants, implementing partners, beneficiaries, intervention sites) 2.4.10 Training of enumerators' and pilot testing of survey tools (if applicable) 2.4.11 Supervision of & support provided to enumerators' 2.4.12 Other methods used for quality assurance of collected data 2.4.13 Data analysis methods: how data have be coded, displayed, processed, aggregated, synthetized, compared use of quantitative and qualitative data analysis methods; basis for jugment (subcriteria, performance indicators, benchmarks, standards, comparisons etc.); how Unicef's contribution to results has been assessed e.g. causal contribution analysis, rival hypothesis etc. 2.4.14 To what extend the evaluation has been participatory (involving stakeholders including beneficiary at every stage of the evaluation, from design to formulation of recommendations) 2.4.15 How aspects of complexity have been addressed 2.5. Equity aspects 2.5.1 How equity and gender have been taken into account 2

2.6. Ethical aspects Identification of ethical issues throughout the evaluation project as well as the measures and methods adopted to mitigate against these issues: methods or practices to ensure the avoidance or minimization of harm and stress to participants; security matters and protection protocols utilized - both for enumerators and people interviewed; protection of privacy and safety of all 2.6.1 participants; obtention of informed consent / verbal assent from participants; confidentiality & anonymity of data collected; absence of benefit or compensation offered to interviewees; training of enumerators in these issues and on enumeration/communication skills; declaration of any conflict of interests; if needed: ethical review and registration on clinicaltrial.gov. Reference to UNEG/UNICEF Ethical guidelines for evaluations, Evaluators' code of conduct, and Procedure for ethical standards. URL link to these documents are included. 2.7. Organisation of the evaluation work 2.7.1 List of main phases (inception phase, data collection, data analysis and report drafting, stakeholders workshop on preliminary findings and formulation of recommendations, etc.) and timeline 2.7.2 Description of each phase: objective and activities undertaken, duration, profile and role of people involved, logistics, support received from Unicef and other stakeholders, challenges faced and measures taken 2.7.3 Quality assurance process used for each deliverable (internally within the evaluation team, externally with the evaluation manager, steering committee, technical reference group or during the final stakeholders workshop) and reference to the UNICEF standards for evaluation reports 2.8. Methodological and organizational limitations 2.8.1 Methodological limitations: Issues of sampling and sample size; over-reliance on self-report; lack of use of relevant documents or M&E data; identification of potential bias (e.g. from respondents); lack of triangulation of key findings etc. 2.8.2 Organizational limitations: lack of access to relevant documents, M&E data, key informants; logistical issues, impossibility to make relevant observations in the field; etc. 2.8.3 How these limitations have been managed and to what extend they weaken the credibility of findings 3. Findings for evaluation criterion #1 - It is suggested here to dedicate one section per evaluation criterion 3.1. Methodological approach 3.1.1 Detailed definition of the evaluation criterion and of potential sub-criteria 3.1.2 Description of the evaluations questions and sub-questions related to the evaluated criterion, and hypotheses related to this criterion; definition of key concepts 3.1.3 3.1.4 3.2. Results and analysis 3.2.1 3.2.2 3.2.3 Factors explaining the findings (causal analysis) or formulation and assessment of hypotheses 3.2.4 Good or bad practice that can be documented, useful for and shared with the various categories of readers (examples, case studies, quotes from the interviews etc.) 3.3. Conclusion on (criterion #1) 3.3.1 For each question and programme component: clear, short, objective, evidence-supported and nuanced response 3.3.2 Overall assessment of the performance of the programme with reference to the evaluated criterion 4. Findings for evaluation criterion #2 4.1. Methodological approach 4.1.1 Detailed definition of the evaluation criterion and of potential sub-criteria 4.1.2 Description of the evaluations questions and sub-questions related to the evaluated criterion, and hypotheses related to this criterion; definition of key concepts 4.1.3 4.1.4 3

4.2. Results and analysis 4.2.1 4.2.2 4.2.3 Factors explaining the findings (causal analysis) or formulation and assessment of hypotheses 4.2.4 Good or bad practice that can be documented, useful for and shared with the various categories of readers (examples, case studies, quotes from the interviews etc.) 4.3. Conclusion on (criterion #2) 4.3.1 For each question and programme component: clear, short, objective, evidence-supported and nuanced response 4.3.2 Overall assessment of the performance of the programme with reference to the evaluated criterion 5. Findings for evaluation criterion #3 5.1. Methodological approach 5.1.1 Detailed definition of the evaluation criterion and of potential sub-criteria 5.1.2 Description of the evaluations questions and sub-questions related to the evaluated criterion, and hypotheses related to this criterion; definition of key concepts 5.1.3 5.1.4 5.2. Results and analysis 5.2.1 5.2.2 5.2.3 Factors explaining the findings (causal analysis) or formulation and assessment of hypotheses 5.2.4 Good or bad practice that can be documented, useful for and shared with the various categories of readers (examples, case studies, quotes from the interviews etc.) 5.3. Conclusion on (criterion #3) 5.3.1 For each question and programme component: clear, short, objective, evidence-supported and nuanced response 5.3.2 Overall assessment of the performance of the programme with reference to the evaluated criterion 6. Findings for evaluation criterion #4 6.1. Methodological approach 6.1.1 Detailed definition of the evaluation criterion and of potential sub-criteria 6.1.2 Description of the evaluations questions and sub-questions related to the evaluated criterion, and hypotheses related to this criterion; definition of key concepts 6.1.3 6.1.4 6.2. Results and analysis 6.2.1 6.2.2 6.2.3 Factors explaining the findings (causal analysis) or formulation and assessment of hypotheses 6.2.4 Good or bad practice that can be documented, useful for and shared with the various categories of readers (examples, case studies, quotes from the interviews etc.) 4

6.3. Conclusion on (criterion #4) 6.3.1 For each question and programme component: clear, short, objective, evidence-supported and nuanced response 6.3.2 Overall assessment of the performance of the programme with reference to the evaluated criterion 7. Findings for evaluation criterion #5 7.1. Methodological approach 7.1.1 Detailed definition of the evaluation criterion and of potential sub-criteria 7.1.2 Description of the evaluations questions and sub-questions related to the evaluated criterion, and hypotheses related to this criterion; definition of key concepts 7.1.3 7.1.4 7.2. Results and analysis 7.2.1 7.2.2 7.2.3 Factors explaining the findings (causal analysis) or formulation and assessment of hypotheses 7.2.4 Good or bad practice that can be documented, useful for and shared with the various categories of readers (examples, case studies, quotes from the interviews etc.) 7.3. Conclusion on (criterion #5) 7.3.1 For each question and programme component: clear, short, objective, evidence-supported and nuanced response 7.3.2 Overall assessment of the performance of the programme with reference to the evaluated criterion 8. Findings for evaluation criterion #6 8.1. Methodological approach 8.1.1 Detailed definition of the evaluation criterion and of potential sub-criteria 8.1.2 Description of the evaluations questions and sub-questions related to the evaluated criterion, and hypotheses related to this criterion; definition of key concepts 8.1.3 8.1.4 For each question and programme component: Presentation of the type of data collected, data source and data collection method used to answer each question (bias mitigation, triangulation) 8.2. Results and analysis 8.2.1 8.2.2 8.2.3 Factors explaining the findings (causal analysis) or formulation and assessment of hypotheses 8.2.4 Good or bad practice that can be documented, useful for and shared with the various categories of readers (examples, case studies, quotes from the interviews etc.) 8.3. Conclusion on (criterion #6) 8.3.1 For each question and programme component: clear, short, objective, evidence-supported and nuanced response 8.3.2 Overall assessment of the performance of the programme with reference to the evaluated criterion 9. Conclusions - The report objectively summarizes the key findings by evaluation criterion, programme component, and relevant unit of disaggregation 9.1 Summary of conclusions by evaluation criterion (hightlighing both strengths and weaknesses), and implications for the programme 9.2 Summary of conclusions by programme component (hightlighing both strengths and weaknesses, and specifying which activities or stakeholders contributed the most/least to the results?) and implications for the programme 9.3 Summary of conclusions by sub-category of population, geographical area, etc. and implications for the programme 5

10. Recommendations - The report formulates relevant, feasible, users-oriented and prioritized actions logically following the findings and conclusions 10.1 Priority recommendations: Urgent or quick-wins actions (listed by stakeholders/target group, and by programme component) 10.2 More longer term recommendations, likely to be slower to plan and implement (listed by stakeholders/target group, and by programme component) 11. Lessons learnt - Contribution of the evaluation to general knowledge outside of the particular scope of this evaluated programme; good practices worth replicating in other areas, countries or sectors; bad practices to avoid and also to communicate to a wider audience as part of the sector learning 11.1 For the sector stakeholders (programmatic aspects) 11.2 For evaluators of similar programmes (methodological aspects) 11.3 Others 12. Annexes - Provides all data and information needed to better all aspects of the evaluation but that are too long to be included in the body of the report 12.1 Full evaluation matrix linking eval criteria to questions, to hypothesis, to data to be collected, to data source, to data collection methods and sampling (and possibly to criteria or standards that will be used by the evaluation team to make a judgement on whether or not UNICEF as performed well in each particular evaluation question) 12.2 Detailed agenda of the evaluation work 12.3 List of documents read and used 12.4 List of key informants interviewed (if they accept to be identified in the report) 12.5 List of sites visited 12.6 Data collection tools 12.7 Enumerators interview/survey guide (if applicable) 12.8 Full table/set of cleaned data 12.9 Synthesis of comments received on the draft report and how they have been taken into account in the final report (or why not) 12.10 ToR of the evaluation 12.11 Short biography / professional background of the evaluators 12.12 Others 6