Specify the Key Evaluation Questions

Key Evaluation Questions (KEQs) are the high-level questions that an evaluation is designed to answer - not specific questions that are asked in an interview or a questionnaire. Having an agreed set of Key Evaluation Questions (KEQs) makes it easier to decide what data to collect, how to analyze it, and how to report it.

KEQs usually need to be developed and agreed on at the beginning of evaluation planning - however sometimes KEQs are already prescribed by an evaluation system or a previously developed evaluation framework. 

Try not to have too many Key Evaluation Questions - a maximum of 5-7 main questions will be sufficient. It might also be useful to have some more specific questions under the KEQs.

Key Evaluation Questions should be developed by considering  the type of evaluation being done, its intended users, its intended uses (purposes), and the evaluative criteria being used.  In particular, it can be helpful to imagine scenarios where the answers to the KEQs being used - to check the KEQs are likely to be relevant and useful and that they cover the range of issues that the evaluation is intended to address. ​(This process can also help to review the types of data that might be feasible and credible to use to answer the KEQs).

The following information has been taken from the New South Wales Government, Department of Premier and Cabinet Evaluation Toolkit, which BetterEvaluation helped to develop.

Here are some typical key evaluation questions for the 3 main types of evaluation:

Key evaluation questions for the main types of evaluation 

Type Typical key evaluation questions
Process evaluation How is the program being implemented?
How appropriate are the processes compared with quality standards?
Is the program being implemented correctly?
Are participants being reached as intended?
How satisfied are program clients? For which clients?
What has been done in an innovative way?
Outcome evaluation (or impact evaluation) 

How well did the program work?
Did the program produce or contribute to the intended outcomes in the short, medium and long term?
For whom, in what ways and in what circumstances? What unintended outcomes (positive and negative) were produced?
To what extent can changes be attributed to the program? 
What were the particular features of the program and context that made a difference?
What was the influence of other factors?

Economic evaluation (costeffectiveness analysis and cost-benefit analysis)

What has been the ratio of costs to benefits?
What is the most cost-effective option?
Has the intervention been cost-effective (compared to alternatives)?
Is the program the best use of resources?

 

Appropriateness, effectiveness and efficiency

Three broad categories of key evaluation questions to assess whether the program is appropriate, effective and efficient are often used.

Organising key evaluation questions under these categories, allows an assessment of the degree to which a particular program in particular circumstances is appropriate, effective and efficient. Suitable questions under these categories will vary with the different types of evaluation (process, outcome or economic). 

  Typical key evaluation questions
Appropriateness  To what extent does the program address an identified need?How well does the program align with government and agency priorities?

Does the program represent a legitimate role for government?

 Effectiveness To what extent is the program achieving the intended outcomes, in the short, medium and long term?
To what extent is the program producing worthwhile results (outputs, outcomes) and/or meeting each of its objectives?
Efficiency Do the outcomes of the program represent value for money?
To what extent is the relationship between inputs and outputs timely, cost-effective and to expected standards?

Example

The Evaluation of the Stronger Families and Communities Strategy used clear Key Evaluation Questions to ensure a coherent evaluation despite the scale and diversity of what was being evaluated – an evaluation over 3 years, covering more than 600 different projects funded through 5 different funding initiatives, and producing 7 issues papers and 11 case study reports (including studies of particular funding initiatives) as well as ongoing progress reports and a final report.  

The Key Evaluation Questions were developed through an extensive consultative process to develop the evaluation framework, which was done before advertising the contract to conduct the actual evaluation.

1. How is the Strategy contributing to family and community strength in the short-term, medium-term, and longer-term?

2. To what extent has the Strategy produced unintended outcomes (positive and negative)?

3. What were the costs and benefits of the Strategy relative to similar national and international interventions? (Given data limitations, this was revised to ask the question in ‘broad, qualitative terms’

4. What were the particular features of the Strategy that made a difference?

5. What is helping or hindering the initiatives to achieve their objectives? What explains why some initiatives work? In particular, does the interaction between different initiatives contribute to achieving better outcomes?

 6. How does the Strategy contribute to the achievement of outcomes in conjunction with other initiatives, programs or services in the area?

 7. What else is helping or hindering the Strategy to achieve its objectives and outcomes? What works best for whom, why and when?

8. How can the Strategy achieve better outcomes?

CIRCLE (2008) Stronger Families and Communities Strategy 2000-2004: Final Report. Melbourne: RMIT University. 

The KEQs were used to structure progress reports and the final report, providing a clear framework for bringing together diverse evidence and an emerging narrative about the findings.

The Managers' Guide

Coming at this from a manager or commissioner's perspective? Step 2: Scope the evaluation in our Managers' Guide has some specific information geared towards making decisions about what the evaluation needs to do, including how to develop agreed key evaluation questions
 
Inline image 1

Resources

Guides

Tools

Examples

 

 

Cite this page

'Develop the Evaluation brief' (2016) New South Wales Government Department of Premier & Cabinet Evaluation Toolkit. © State of New South Wales Department of Premier and Cabinet 2015. Retrieved from: http://www.dpc.nsw.gov.au/programs_and_services/policy_makers_toolkit/st...

Comments

Anonymous's picture
Robin Kuwahara

This is great. I would like to share another related resource. Colleagues of mine at CDC’s National Asthma Control Program created a useful checklist for assessing potential evaluation questions. The list is grounded in the evaluation literature and has benefitted from the practice wisdom of evaluators who serve in a range of capacities and agencies.  It can be found here:

http://www.cdc.gov/asthma/program_eval/other_resources.htm

Add new comment

Login Login and comment as BetterEvaluation member or simply fill out the fields below.