IMPORTANT LEGAL NOTICE - The information on this site is subject todisclaimercopyright notice
Evaluation
  EUROPA > European Commission > EuropeAid > Evaluation > Methodology > Guidelines > Check lists
Last updated: 12/01/2006
EuropeAid

Check-lists
Quality Assurance

EuropeAid
 

Quality of first phase report (desk)

 


Methodology
• Evaluation Guidelines
• Methodological bases
• Evaluation tools
• Examples
• Glossary
• Sitemap

Check lists
• content of
documents

• quality
assurance

• meetings

Quality assurance
• Inception
• First phase (desk)
• Final report

 
 

The criteria include and complete those from the previous stage (inception report).


1. Meeting needs

This report expresses clearly and thoroughly the way in which the evaluation team understands the following points:

  • The requirements which appear in the regulatory framework of the evaluation, especially as regards accountability.
  • The expectations expressed in the terms of reference.
  • The requirements expressed by the reference group members.

The report places the evaluation into its context, in connection to the bases of the development, cooperation or foreign policies and to any other EC policy or partner country's policy.

The proposed evaluation questions and their relevant judgment criteria reflect the identified requirements and the evaluation's intended use. They do not contain any ambiguity.

The report provides a first part of the answers to the evaluation questions and these contributions reflect the identified requirements.


2. Relevant scope

The report delineates the evaluation's central scope as regards its temporal, geographic and regulatory dimensions. It justifies the choices made for delineating the scope. It includes an analysis of major overlaps with related policies and justifies the choice of overlaps examined.

The themes, evaluation questions and judgment criteria reflect:

  • The results and intended impacts identified through the reconstruction of the intervention logic.
  • The various sectors, themes and instruments.
  • DAC's evaluation criteria, without leaving aside efficiency and sustainability, or coherence/complementarity and community added value.
3. Defensible design

The report describes the data collection and analysis method actually applied in the desk phase. It accounts for problems encountered and possible limitations.

The report describes the data collection and analysis method to be applied in the field phase. It shows how this method will allow all the evaluation questions to be addressed appropriately, and an overall assessment to be produced. The choices are discussed and defended against other options.

The method is feasible within the evaluation context. Both the risks and limitations are clearly specified. The report states the risks that would be taken if other methodological options were adopted.


4. Reliable data

Sources of qualitative and quantitative data are stated. The evaluation team provides a self-assessment of the reliability of data. Limitations to validity are clearly stated.


5. Sound analysis

The report includes a first analysis of available data with a view to answering evaluation questions, and deduces assumptions to be tested in the field. The reasoning is explicit and well-founded.


Top