We are currently moving our web services and information to Canada.ca.

The Treasury Board of Canada Secretariat website will remain available until this move is complete.

Evaluation Guidebook for Small Agencies

Archived information

Archived information is provided for reference, research or recordkeeping purposes. It is not subject à to the Government of Canada Web Standards and has not been altered or updated since it was archived. Please contact us to request a format other than those available.



Evaluation Guidebook
for Small Agencies

Acknowledgements

Section One: Introduction

Section Two: Defining Evaluation

Section Three: Developing an Evaluation Function

Section Four: Evaluation Cycle

Section Five: Managing Evaluations

Section Six: Using Evaluations

Appendix ATypes of Small Agencies
Appendix B
Horizontal Initiatives
Appendix C
Seeking External Advice and Support 
Appendix D
Expenditure Review Committee's 7 Tests
Appendix E
"How to" Information for Planning and Conducting Evaluations
Appendix F
Terms of Reference Template
Appendix G
Glossary 
Appendix H
Evaluation Web Sites

 

 

 



Section One: Introduction

Section One: Introduction

This guidebook is a reference tool designed to assist small agencies to develop their capacity to plan, implement and manage evaluations. The guidebook will help small agencies meet the requirements of the Treasury Board of Canada (TB) Evaluation Policy .

This "how-to" guidebook presents effective practices for planning and managing evaluation projects, provides strategies for fostering the use of information from evaluations in strategic decision making within small agencies, and provides examples of key documents used in the evaluation process.

This guidebook builds on the work that was previously conducted in 2003–04 with respect to developing appropriate models of evaluation and performance measurement in small agencies. [1] The guidebook represents the next step in assisting small agencies to build their capacity to plan, implement, and manage evaluations.

1.1  Purpose of the Guidebook

This guidebook should help you with the following:

1.2  Who Should Use This Guidebook?

This guidebook is designed for small agencies that require assistance in implementing the Treasury Board of Canada (TB) Evaluation Policy requirements. The primary target audience is the personnel who are responsible for evaluation within small agencies. For those readers who have limited experience in the field of evaluation, the guidebook is designed to provide detailed information on most aspects of evaluation and includes helpful material such as context, definitions, checklists, management tips and references for more detailed information on certain subjects. For those readers who have more experience with evaluations, they may find various parts or sections of the guidebook more useful than others, such as the checklists. It is also recognized that small agencies' needs are varied with respect to evaluation and information needs. The guidebook attempts to address these diverse needs.

1.3  Small Agency Context

Most small agencies were created by the federal government to independently carry out specific mandates. For example, some agencies render impartial decisions (e.g., tribunals). Other agencies perform a facilitation or intermediary role. Others have regulatory and inspection mandates. See Appendix A for an overview of different types of small agencies.

With respect to evaluation, previous studies [2] have noted that many small agencies face challenges in developing evaluation and performance measurement functions for their agencies. These challenges include having adequate resources and capacity to develop these functions, having to adapt models from larger organizations, and integrating the function with the day-to-day business of the agencies.

Small agencies as a group also have unique challenges and characteristics when compared with medium and large federal departments. One difference is that small agencies often have one or two main business lines in comparison with medium and large departments that often have multiple business lines containing many programs, initiatives, and policies. Another difference is that, in comparison with medium and large departments, small agencies often have more limited flexibility with respect to financial resources. Given these and other differences, many aspects of the models in meeting accountability and performance requirements in medium and large departments are not applicable to small agencies.

When considering what type of evaluation and performance measurement models are required in the small agency community, it is important to realize that, although the community shares commonalities when compared with medium and large departments, there is also a great amount of diversity within the small agency community itself. The small agency community in the federal government is diverse on various dimensions such as organizational structure, relationship with larger departments, nature of work, and organization size. These dimensions, in addition to others, contribute to the type and nature of information that agencies need for decision making and ensuring accountability within their organizations.

A 2003 Treasury Board of Canada Secretariat (TBS) project entitled Models for Evaluation and Performance Measurement for Small Agencies [3] presented three models of evaluation and performance measurement functions within federal small agencies. Two key considerations in developing the models were the types of management decision making and the types of information needed to make decisions within agencies.

These models are as follows:

Model A – Straightforward Information Needs

Model B – Blend of Straightforward and Complex Information Needs

Model C – Complex Information Needs

For a more detailed description of the models and how to classify your own agency, please go to the TBS Web site at http://www.tbs-sct.gc.ca/cee/tools-outils/model-eng.asp .

1.4  Structure of the Guidebook

In addition, there are a number of appendices that contain more detailed information that can be used as references.

1.5  How to Use this Guidebook

The guidebook is designed to be used by readers in a number of different ways. For those readers who are relatively new to the field of evaluation, you may want to work your way systematically through the guide. For more experienced readers, it may be more useful to go directly to the specific section that you need. Regardless of how you choose to use the guide, you should be aware of the symbols used throughout the text as indicated below.

Small Agency highlights

Small Agency highlights are contained in shaded boxes and are preceded by a circular symbol.

arrow symbol

Checklists are contained within a template and preceded by an arrow symbol.

This icon highlights excerpts from the TB Evaluation Policy

This icon highlights excerpts from the TB Evaluation Policy .

This file folder provides reminders with respect to key learning points

This file folder provides reminders with respect to key learning points.

This icon provides additional emphasis on learning points.

This icon provides additional emphasis on learning points.

If you require a more detailed guide with respect to planning or implementing an evaluation, please refer to Appendix E.

Terminology

Throughout this guidebook we will use the word "program" to refer to programs, policies, or initiatives.

Key References

Treasury Board of Canada Secretariat. Models for Evaluation and Performance Measurement for Small Agencies: Summary Report , 2003.

Treasury Board of Canada Secretariat. Independence vs. Partnering – Finding the Right Balance: A Dialogue on Values and Ethical Decision-Making in Small Agencies , 2003.

Treasury Board of Canada Secretariat. Evaluation Policy and Standards , 2001.

 



Section Two: Defining Evaluation

Section Two: Defining Evaluation

 

This section

  • defines evaluation;
  • describes different types of evaluation;
  • compares evaluation to other accountability activities such as performance measurement, internal audit, and management or operational reviews; and
  • places evaluation in the context of other government initiatives such as the Management Accountability Framework (MAF) and Management Resources and Results Structure (MRRS).

2.1  What is Evaluation?

Evaluation can be defined as the systematic collection and analysis of information on the performance of a policy, program, or initiative to make judgments about relevance, progress or success and cost-effectiveness, and/or to inform future programming decisions about design and implementation.

RBM E-Learning Tool ; TBS Web site

Note that evaluation

Evaluation provides a periodic opportunity to take an in-depth look at how a program, policy or initiative is doing. The primary focus is usually on being able to bring about improvements to facilitate the achievement of results or to determine the degree to which the program, policy or initiative led to the achievement of desired results (i.e., attribution).

RBM E-Learning Tool ; TBS Web site

2.2  Why Evaluation?

Following the release of the document Results for Canadians: A Management Framework for the Government of Canada in April 2001, TB released an Evaluation Policy "to ensure that the government has timely, strategically focussed, objective and evidence-based information on the performance of its policies, programs, and initiatives to produce better results for Canadians."

 Evaluation has the following two main purposes:

1.  Improvement: to help managers design or improve policies, programs, and initiatives; and

2.  Accountability: to provide, where appropriate, periodic assessments of policy, program, or initiative effectiveness, of impacts both intended and unintended, and of alternative ways of achieving expected results.

Within the small agency context, both improvement and accountability objectives are relevant. Evaluations can assist small agencies

2.3  Evaluation and the Program Development Cycle

Ideally, planning and evaluation are interdependent processes. Evaluation can and should be built into the planning process for agencies. Combined with the feedback from performance measurement activities, information from evaluations can help to guide planning within the agency.

Evaluation and the Program Development Cycle

2.4  Overview of Tasks in Conducting Evaluations

The illustration below provides an overview of the tasks required to conduct an evaluation. This guide examines these tasks in more detail in Sections 4, 5, and 6.

Overview of the tasks required to conduct an evaluation

 

2.5  Types of Evaluation

There are two main types of evaluations: formative and summative.

Formative Evaluation

This type of evaluation examines the effectiveness of implementation for facilitating improvement. A formative evaluation may be conducted in mid-cycle of the program, policy or initiative (i.e., within two years). Its intent is to provide information to improve the program, policy or initiative.

Formative evaluations should be used judiciously. They may not be required for ongoing programs. If compliance audits include operational questions, then a formative evaluation may not be needed. Formative evaluations may also target particular aspects of a program – for example, the performance measurement system.

Formative evaluations focus on the following:

  • management issues of how the policy, project or initiative is being implemented and delivered;
  • how risk is being managed;
  • if the performance measurement system is generating valid and reliable performance data;
  • verifying if adjustments are necessary; and
  • to what extent progress toward the achievement of the desired results is occurring.

Where "full formative evaluations" are undertaken, there is an expectation that outputs, early results, validation of program logic, and the likelihood of long-term results achievement are assessed. [4]

All evaluations, whether formative or summative, should address the Expenditure Review Committee's questions. (See Appendix D.)

 

Summative Evaluation

This type of evaluation examines impacts in order to make a decision about overall effectiveness. They have primarily an accountability function and are generally conducted towards the end of the cycle.

Summative evaluations focus on the following:

  • Relevance: Does the program continue to be consistent with departmental and government-wide priorities and does it realistically address an actual need?
  • Success: The degree to which desired results have been achieved and the extent to which the policy, program or initiative has contributed to the achievement of results.
  • Cost-effectiveness: Are the most appropriate and efficient means being used to achieve objectives, relative to alternative design and delivery approaches?

All evaluations, whether formative or summative, should address the Expenditure Review Committee's questions. (See Appendix D.)

Another type of evaluation often referred to within the federal context is a Horizontal Evaluation . These are evaluations of initiatives that involve the co-ordinated activities of several federal departments or agencies. See Appendix B for more details.

2.6  Other Accountability Activities

As previously mentioned, evaluation is one type of activity or tool that can be used by managers to demonstrate accountability. This section contains brief descriptions of other tools (e.g., performance measurement, audit), a discussion of how to determine when each is most appropriate, and how they can be used in conjunction with one another.

2.6.1  Performance Measurement

Performance measurement is the regular collection of information for monitoring how a policy, program, or initiative is doing at any point in time. It generally focuses on providing operational performance information to program managers. For many small agencies, one of the first tasks in developing an evaluation function is to develop a performance measurement strategy that encompasses periodic evaluation.

Evaluation and performance measurement are two complementary activities. Evaluation provides accountability by ensuring performance measurement is on track and indicators are appropriate. In turn, performance measurement data are often important data sources for periodic evaluations.

Small Agency highlights Small agencies vary in terms of the relative emphasis they place on performance measurement and evaluation. Some may rely more heavily on ongoing performance measurement, while others may place more emphasis on periodic evaluations. The most important thing is to determine what are an agency's information requirements with respect to decision making, and then ensure that the performance measurement and evaluation strategy is meeting these requirements.

2.6.2  Internal Audit

Internal audit is a function that provides assurances on a department or agency's risk management strategy, management control framework and information, both financial and non-financial, used for decision making and reporting.

The internal audit function does the following:

Small Agency highlights In many small agencies , the same person or small groups of people may be responsible for planning, implementing, and managing performance measurement, evaluations, audits, and reviews.

 

"When used in combination, performance measurement, audit and evaluation serve as an effective means to monitor the performance of an initiative throughout its life-cycle."

— RBM E-Learning Tool; TBS Web site

Audit and evaluation

2.6.3  Review

Reviews are often conducted in response to a pressing or immediate need of management. As such, the emphasis is usually on quick generation of sufficient information to inform decision making or reassure senior management of the dimensions of a problem or situation. The methodology used to gather information is usually secondary to developing an adequate answer in a timely fashion (i.e., evaluation or audit protocols and approaches are not adhered to). Although they are useful to address targeted issues, reviews or special studies do not conform to external reporting requirements, project control processes, or standards which delineate a discipline such as audit or evaluation.

2.7  Evaluation in Context

When would you conduct an evaluation?

You would conduct an evaluation when you need information about how and why results were achieved, the extent to which something has been implemented, whether your initiative is relevant, the extent to which the intended impacts have been achieved, what unintended impacts have resulted, and when you want to consider future options or strategies for improvement.

When would you conduct an audit?

You would conduct an internal audit when you need information about the adequacy of internal controls, risk management strategies, extent of compliance with rules, regulations or procedures, management practices and controls, and the extent to which financial, operational, or management information is valid.

When would you conduct a review?

You would conduct a review when there is a pressing need for management information that may be more limited in scope than an evaluation (or outside the scope of an evaluation) and when time or resources do not permit the rigour expected in an evaluation.

Evaluation has a long history in the federal government. More recently, evaluation is perceived as a key tool for public service managers as they address the requirements of recent approaches to management in the public sector. Some of these are briefly described below to situate evaluation in its current context. More detailed information on each of these approaches or initiatives is available directly from TBS. [5] All of the initiatives described below will continue to have an impact on managers in small agencies as they develop capacity in areas such as evaluation, performance measurement, and internal audit.

2.7.1  Results-based Management

Results-based management is a comprehensive approach to management aimed at improving performance through achieving better results.

Shortly after the federal government introduced Results for Canadians in 2000, the Results-based Management and Accountability Framework (RMAF) was introduced. The RMAF is a tool used to plan, monitor, evaluate, and report on the results of a program. The RMAF integrates the evaluation function within the context of results-based management. It is also a link to the Management Resources and Results Structure (see below). RMAFs will be discussed in more detail in Section 4 of this guidebook.

2.7.2  Management Accountability Framework (MAF)

The MAF defines and clarifies management expectations. It is a set of ten statements that summarizes TBS's expectations for modern public service management.

The MAF aims to

The MAF will be used in the following ways:

A graphical representation of the Management Accountability Framework follows.

Management Accountability Framework: Ten Essential Elements of Modern Managemen
Management Accountability Framework: Ten Essential Elements of Modern Management

2.7.3  Management Resources and Results Structure (MRRS)

The MRRS replaces the Planning, Reporting and Accountability Structure (PRAS) policy. In accordance with MAF expectations, MRRS supports governance and strategic direction, accountabilities and results and performance. The new policy is directed to the organizational level and encourages the alignment of programs, resources, and management practices with expected results.

The Program Activity Architecture (PAA) is an element of the MRRS and comprises

Over time, an integrated MRRS should also include

The Expenditure Management Information System (EMIS) will provide a common framework that aligns information on priorities, plans, actual expenditures, and results.

2.7.4  Where Does Evaluation Fit In?

With its focus on the "how" and "why" of results, evaluation is a key tool in managing for results. As a part of management practices, evaluation can help to design and improve programs. Evaluation should be used in conjunction with other management tools to improve accountability and decision making.

Evaluations are a critical tool for demonstrating results and performance and as such help to support the other elements contained in the MAF. Moreover, through the inclusion of results and performance as a key management expectation, the MAF makes an explicit commitment to evaluation.

Evaluation can provide timely information on impacts and relevance with respect to strategic objectives of an agency. Hence, evaluation will address some of the information needs as required by MRRS. Annual evaluation plans (organizational level) and RMAFs (program level) provide links to the MRRS.

The graphic below places evaluation within the context of other government initiatives.

Evaluation in Context

Evaluation in Context

Key References

Canadian Evaluation Society. Project In Support of Advocacy and Professional Development: Evaluation Benefits, Outputs, and Knowledge Elements , September 16, 2002.

Goss Gilroy Inc. Review of Approaches in Other Jurisdictions. Lessons Learned from International Experience with Performance Measurement/Evaluation in Small Agencies , 2003.

National Science Foundation. User-Friendly Handbook for Project Evaluation.

Treasury Board Canada Secretariat Web site RBM E-Learning Tool .

Treasury Board Canada Secretariat Web site. Management Accountability Framework

Treasury Board of Canada Secretariat. Evaluation Policy and Standards , 2001.

Treasury Board of Canada Secretariat. Preparing and Using Results-based Management and Accountability Frameworks , April 2004.

Treasury Board of Canada Secretariat. Results for Canadians , 2000.

 



Section Three: Developing an Evaluation Function

Section Three: Developing an Evaluation Function

This section

  • provides an overview of small agency evaluation capacity;
  • reviews the requirements of the Evaluation Policy ;
  • reviews best practices for establishing an Agency Evaluation Plan; and
  • outlines strategies for building evaluation capacity.

3.1  Challenges in Developing an Evaluation Function

As mentioned previously, many small agencies are currently facing or have faced considerable challenges in developing evaluation and performance measurement capacity in their organizations. Despite these challenges, some small agencies have made considerable progress with respect to implementing evaluation and performance activities in their organizations in a meaningful way. In these agencies, information and findings from performance measurement and evaluations is valued and is actively used to support decision making and planning within the organization.

Some factors that have been found to be associated with enhanced evaluation capacity in smaller organizations include the following:

Small Agency highlights Did you know?

In New Zealand, all government agencies and departments are required to develop Statements of Intent , reflecting the agencies' outcome targets. It was reported that in the smaller organizations, their targets were less the result of compromise and more reflective of the "big picture," with better links between the overall mandate of the organization and the performance targets.

Small agencies face many challenges with respect to enhancing evaluation capacity within their organization. Examples include the following:

As illustrated, the small agency environment creates unique circumstances and challenges for building capacity in evaluation and performance measurement. Suggestions for building capacity are outlined in later subsections.

3.2  Checklists for Implementing the Evaluation Policy

excerpts from TB Evaluation PolicyREMEMBER...

Organizational positioning of evaluation should also reflect the unique needs of the department or agency.

— TB Evaluation Policy, 2001

In this section, there are a number of checklists available that will assist the reader in becoming familiar with various sections of the TB Evaluation Policy . For readers already familiar with the policy, it may be a good exercise to go through the various lists to identify potential gaps or areas covered by your agency in implementing the policy.

  ChecklistChecklist for Required Elements of the Evaluation Policy

Required Element

Evaluation Capacity (Deputy Heads):

1.  Appoint a senior head of evaluation

2.  Establish an evaluation committee* and designate a senior departmental executive to chair it

3.  Ensure that TBS is given access to annual evaluation plans and early warning of evaluation findings that are of concern

Leadership and Direction (Departmental Heads of Evaluation):

4.  Ensure agency evaluation plans are strategic and adequately cover policies, programs, initiatives

5.  Ensure departmental heads of evaluation work with managers to enhance design, delivery and performance measurement of policies, programs, and initiatives

6.  Conduct evaluation studies in accordance with the evaluation plan

7.  Inform senior management and other appropriate players promptly of any findings of major concern

8.  Make completed evaluation reports available to the TB and to the public in both official languages

9.  Apply evaluation standards outlined in the policy (Evaluation Planning and Issues, Competency, Objectivity and Integrity, Consultation and Advice, Measurement and Analysis, Reporting)

Managing for Results (departmental managers):

10.  Ensure that there is reliable, timely, objective, and accessible information for decision making and performance improvement

11.  Use evaluation findings and measures for improvement in priority setting, planning, reporting, and decision making

*  For small agencies a departmental evaluation committee or a combined audit and evaluation committee may serve the same role. In the smaller agencies (say less than 50 FTEs) the Evaluation Committee, Senior Departmental or Management Committee is often comprised of the same people.

3.2.1  Establishing an Agency Evaluation Plan

The Evaluation Plan is strategically focussed and designed to balance the evaluation workload between meeting internal agency information needs and serving external reporting against federal government priorities. It should be developed or reviewed annually, although it may span several years. Many small agencies may conduct one or possibly two evaluations every five years. As a result, the annual evaluation plan is more likely to be reviewed rather than developed each year. During the review, the plan should be updated and modified to address any changes or shifts that have occurred within the previous twelve months.

The evaluation plan incorporates measures against the MRRS strategic outcomes, covers the program inventory, utilizes the ongoing performance measures, and feeds the products into the governance structure to support the MRRS decision-making needs.

Risk Management

In keeping with demands for a more integrated approach to management (i.e., MAF, Evaluation Policy ), the annual evaluation plan should also take into account the risk management profile developed for the agency. The framework can be based on systematically measuring risks (e.g., risk self-assessments). The risk framework can include risks with respect to strategic outcomes, finances, health and safety, corporate priorities, and government commitments.

  • The risk framework can be used as a foundation for establishing evaluation priorities.

For more information on risk management please refer to the TBS document entitled Meeting the Expectations of the TBS Policy on Internal Audit: A Handbook for Small Departments and Agencies , 2003.

You can also refer to the Integrated Risk Management Framework located on the TBS Web site at http://www.tbs-sct.gc.ca/pubs_pol/dcgpubs/RiskManagement/rmf-cgr_e.asp.

 

excerpts from TB Evaluation PolicyREMEMBER...

Departments should undertake an appropriate balance of evaluation work. To achieve this balance, evaluators should develop a strategically focussed plan that is based on

  • assessments of risk; and
  • departmental priorities and priorities of the government as a whole.

— TB Evaluation Policy, 2001

See the TBS Web site at http://www.tbs-sct.gc.ca/cee/pubs/guide/dep-epe-eng.asp for a detailed guide to good practices for developing and tabling annual evaluation plans. The following checklist outlines an abbreviated version of these emerging or best practices. Please note that not all of the points are required elements. (See previous checklist for required elements of the Evaluation Policy .)

  ChecklistChecklist for Your Agency Evaluation Plan

Considerations

Does the plan address the following?

Needs Assessment

1.  Identify management, client and stakeholder information needs?

Context

2.  Outline how evaluation will be used in the agency?

3.  Link evaluation to strategic concerns? (The evaluation plan should reflect MRRS strategic outcomes, program inventory, and performance measures)

4.  Refer to the TB Evaluation Policy and Standards?

Rationale and Priority Setting

5.  Take into account the priority setting and risk management approach?

6.  Identify methodology used for determining projects?

7.  Link to agency service, business lines and strategic priorities?

Scope and Coverage

8.  Indicate multi-year priorities for agency?

9.  Provide indication of scope of study (for those projects included in plan)? Outline the rationale for including study in the Plan?

10.  Give an appreciation of the proportion of the agency's evaluation universe  that the current year's projects represent?

11.  Consider cross-jurisdictional evaluations?

Management Expectations and Resources

12.  Identify which projects were completed within the fiscal year?

13.  Estimate costs for completing each project and/or planned expenditure in current fiscal year?

14.  Include a summary sheet of projects, costs, total expenditure on evaluation, funding received in addition to A-base for evaluation?

  Checklist for Your Agency Evaluation Plan (cont'd)

Considerations

Does the plan address the following?

Credibility and Timeframe (Does the plan address the following?)

15.  Identify project teams and schedules?

16.  Consider TBS standards during development of Plan?

17.  Identify key assumptions in order to achieve deliverables as per Plan?

Reporting (Have you…?)

18.  Tabled plan with Evaluation Committee for approval?

19.  Reviewed plan after six months or provided status report? (Six-month review or status report on Plan by Evaluation Committee)

20.  Posted evaluation reports on Agency Web site?*

21.  Forwarded evaluation reports to TBS for inclusion in database?

22.  Forwarded approved Evaluation Plan to CEE for review?

Evaluation Policy requires evaluation reports to be made public.

3.3  Building Internal Evaluation Capacity

The concept of capacity building is similar to the concepts of organizational development, organizational effectiveness and/or organizational performance.

Capacity building involves a variety of activities such as the following:

 

3.3.1  Steps for Building Evaluation Capacity


Building capacity within an organization usually follows three basic steps.

Step One – Identify Gaps

Identify gaps and issues with respect to evaluation capacity.

a.  Conduct inventory of existing data sources and information (i.e., operational, financial, administrative, strategic management, and accountability data).

b.  Undertake assessment with managers to identify current information needs.

c.  Identify information gaps and needs.

Step Two – Develop Strategies

Identify the changes or strategies needed to address these gaps.

d.  Consider strategies related to organizational change, human resources, infrastructure, networking. (See examples and suggested strategies below.)

Step Three – Action Plan for Change

e.  Develop an action plan for change and assign roles and responsibilities.

3.3.2  Strategies for Building Evaluation Capacity

Some examples of how small agencies have developed performance measurement and evaluation capacity within their agency are presented below.

 

Example One:
Patent Medicines Prices Review Board
(PMPRB)

The PMPRB is developing internal capacity in performance measurement and evaluation by ensuring that measurement is an integrated component of activities at various levels. Information from performance measurement systems is supplied to the senior levels of the Agency to assist in strategic planning and review activities at various periods throughout the year. The staff is provided with strategic planning documents so there is a general understanding of the needs and rationale for the different types of performance information that is collected throughout the Agency. Many staff members are then involved directly in the collection and/or processing of performance information as part of their ongoing activities throughout the year.

 

Example Two:
Office of the Commissioner of Official Languages (OCOL)

The OCOL is working to develop capacity in their organization by demonstrating the need for good performance information to various levels of staff, and then having direct involvement of various staff in the development of the necessary frameworks. This approach is augmented with the identification of a few individuals within the organization who could be described as "evaluation champions." As well, senior management's support and direction for performance measurement and evaluation activities is leading the organization in making a gradual shift in OCOL's culture towards one that is more consistent with a results-based management environment.

 

Some approaches to enhancing capacity that were suggested during the exercise to develop models for performance measurement and evaluation in small agencies include the following:

How do you get senior management support?

  • A critical element of success for enhancing evaluation capacity is senior management support. Achieving this type of change can be difficult. Below are some suggestions.
  • Develop or enhance leadership training within the agency so there is a clear understanding of the role of evaluation within a results-based management environment.
  • Build an understanding of the agencies' information needs.
  • Develop an understanding of the barriers and incentives for change and implement appropriate strategies.
  • Offer management training for results.
  • Consult with TBS and other external resources.

 

Small Agency highlights Lessons Learned: Feedback from small agencies during the Modern Comptrollership (MC) capacity assessment process

  • Results have to be communicated on a regular basis to sustain managers' interest.
  • The active commitment of senior management and buy-in from employees is a critical success factor.
  • Off-site workshops allow managers to explore MC concepts and discuss their application within the specific operational context of the agency.
  • Selection of managers to participate in pilot projects is a means of spreading workload, developing synergies and fostering commitment. Involving managers who already have an interest in the topic also helps.
  • Information overload is avoided by using shorter, more frequent sessions rather than full-blown courses.
  • Training must be adapted to the audience and make use of concrete, practical examples.
  • Collaborating with other similar agencies is a good idea when training resources are tight.
  • Resources need to be dedicated to focussed training on MC for middle and senior managers.

 

  ChecklistChecklist for Development and Maintenance of Internal Evaluation Capacity

Considerations

1.  Is there a leader or champion for evaluation? Have people been identified as responsible for developing evaluation capacity?

2.  Is there an understanding and awareness (at the executive level) of information needs and solutions?

3.  Do management and staff understand reasons and need for a stronger evaluation function?

4.  What changes are required to build evaluation? Is the change and the direction well defined? Is the change translated into goals, objectives and behaviours that team members can understand? Have the concepts been turned into a set of organization actions that can be implemented within the Agency?

5.  Are there gaps in knowledge and skills with respect to evaluation?

6.  Are there necessary resources to make these changes? Have possible resources been identified/leveraged?

7.  Have other stakeholders been identified that need to be involved and committed to the change?

8.  Have the key decisions been identified that will enable building of evaluation capacity?

9.  Is it clear who is accountable for these decisions and when?

10.  How have others developed their evaluation capacities?

11.  Have sources of external support and assistance been identified and consulted? (e.g., CEE, CES, SAAN)

12.  Have the barriers or sources of resistance to developing internal evaluation capacity been identified?

13.  Are there appropriate incentives and reinforcements in place?

14.  Has the commitment to evaluation been communicated?

15.  Have the process and steps been communicated to staff?

16.  Is evaluation linked to planning activities?

17.  Is evaluation linked to training and staff development activities?

18.  Is evaluation linked to other accountability activities such as performance measurement?

19.  Is evaluation monitored and followed up?

20.  Is results reporting linked to individuals that manage accountability reports?

General project management skills are required for the effective planning, implementation, and management of evaluations. In this manner, evaluation projects do not differ substantially from many other projects that public service managers will plan and implement in a results-based environment. The checklist below is a generic guide to be used for those who wish to develop and monitor capacity in the area of project management.

Checklist Generic Checklist for Internal Project Management

Considerations

1.  Are the goals, objectives and rationale clear?

2.  Is the scope of the project clear and consistent with the project objectives?

3.  Do stated deliverables show that an objective has been achieved or is progressing towards achievement?

4.  Is there a project start and end date?

5.  Have problems or barriers to completing the project been defined?

6.  Have project development costs been laid out?

7.  Have key roles and responsibilities been assigned and identified?

8.  Is the Agency's governance model reflected in the project's description?

9.  Does the governance model provide for the following: Scope/change management decisions? Fiscal/cost decisions? HR decisions? Issues management? Risk Management? Quality control? Transition to operational decisions?

10.  Are the reporting relationships clear?

11.  Has due consideration been given to risk management, management control framework, scheduling and task plans, communications plan, implementation plan, reporting plan, training plan, and values and ethics?

12.  Are performance planning and budgeting integrated?

13.  Are there adequate management tools, support, and training?

Key References

Community of Small Organizations Web site: .

London School of Economics and Political Science. Juggling On a Unicycle: A Short Guide to Organizing A Small Agency , 1999.

Performance and Planning Exchange. Making Change Personal: Why we don't know what we know , November 2003.

Treasury Board of Canada Secretariat. Models for Evaluation and Performance Measurement for Small Agencies. Summary Report , 2003.

Treasury Board of Canada Secretariat. Changing Management Culture: Models and Strategies to Make It Happen , 2003.

Treasury Board of Canada Secretariat. Evaluation Policy and Standards , 2001.

Treasury Board of Canada Secretariat. Modern Comptrollership Practices: Toward Management Excellence , 2003.

Treasury Board of Canada Secretariat. Review of Approaches in Other Jurisdictions. Lessons Learned from International Experience with Performance Measurement/Evaluation in Small Agencies , 2003.

 



Section Four: Evaluation Cycle

This section

  • provides an overview of the evaluation life cycle;
  • outlines considerations for planning an evaluation at the program level;
  • describes Results-based Management and Accountability Framework (RMAF) and its components;
  • outlines evaluation methods;
  • describes steps for carrying out an evaluation, including analyzing data; and
  • provides an overview of evaluation report writing.
Overview of Evaluation Life Cycle


This section provides an overview of evaluation from preparing an RMAF to collecting the information and writing the report. The graphic below illustrates the evaluation cycle.

The focus of the section will be on the first three steps: (1) planning; (2) implementing; and (3) reporting. More specifically, this section is presented as follows:

Drafting the plan at the program level (or RMAF)

Carrying out the evaluation

Writing the report

The fourth step in the overall evaluation life cycle, using the evaluation results, will be discussed in Section 6.

4.1  Planning the Evaluation

In order to ensure that the evaluation will be a useful product, the details need to be worked out early in the evaluation life cycle. In the beginning it is important to establish a basic understanding of why the evaluation is to be carried out. Below is a checklist for planning your evaluation at the program level.

Checklist  Checklist for Planning an Evaluation (Program Level)

Considerations

1.  Establish understanding of why evaluation is being carried out.

2.  Identify who will use the evaluation to make decisions (e.g., individual administrators, program staff, clients or consumers, legislators, senior management, other stakeholders); where the evaluation findings will be reported (DPR, annual report, departmental Web site); and what types of decisions might be made.

3.  Determine whether management responses/action plan will be required.

4.  Develop an evaluation strategy that includes the following:

  • a description of the program
  • scope and objectives of evaluation
  • evaluation issues and questions
  • data collection methods and sources of information

 

4.2  RMAF: Tool for Planning

Since the Results-based Management and Accountability Framework (RMAF) is a common tool in the federal government for evaluation planning, this subsection will describe the RMAF and its components. This section will focus on the profile, logic model and the evaluation strategy but will only provide a high level description of the ongoing performance measurement strategy.

While the previous section dealt with the overall agency evaluation plan, this subsection focusses on planning an evaluation for a specific program . The evaluation framework component of the RMAF should, however, link to the Agency Evaluation Plan.

Small Agency highlightsIn small agencies with only one or two business lines, the evaluation may be organization wide (e.g., evaluation of an agency's core business). In many instances, RMAFs are being developed for entire small agencies, rather than individual business lines.

 

RMAFs are a management tool, somewhat like a guidepost or compass for an organization, program, policy or initiative.

 

4.2.1  Frequently Asked Questions on RMAFs

 

What is an RMAF?

An RMAF is a plan that describes how a program will be measured, evaluated and reported.

 

Why Use an RMAF?

It serves as a useful guide in helping managers to measure, evaluate and report on their programs. It is a good idea to develop an RMAF (or framework) when a program is being designed to establish reasonable links between the proposed activities and the results and to set out the data collection requirements. An RMAF is also a link to an agency's Management Results and Resources Structure.

 

What is the difference between an RMAF and an Evaluation Framework?

RMAFs have generally replaced evaluation frameworks in the federal government. RMAFs include plans for both performance measurement and evaluation activities in one document, all of which will build on the program theory/logic model as the cornerstone of the RMAF. Stand-alone evaluation frameworks do not necessarily include a performance measurement and reporting strategy.

 

When is an RMAF required?

An RMAF is mandatory for certain categories of programs with transfer payments. These include grant programs (class grants), individual contributions, and contribution programs.

RMAFs are considered to be a good management practice and their use is generally encouraged in the Evaluation Policy and the TBS RMAF Guidance document.

 

Who is Involved in Developing an RMAF?

There are two key parties involved in the development and implementation of an RMAF: program managers and evaluation managers. In the case of those involving Treasury Board submissions, analysts of the Treasury Board of Canada Secretariat may also be involved.

An RMAF is manager led, with evaluators acting as facilitators. Managers hold the primary responsibility for the development and implementation of the RMAF. Managers are responsible for

  • ensuring that the content of the framework is accurate; and
  • implementing the RMAF.

The evaluation function is responsible for the "Evaluation Plan" section.

Key stakeholders should also be consulted in preparing elements of the RMAF. Their early buy-in helps to support the implementation process.

 

What are the Components of an RMAF?

Profile – contains a description of the program including context and need, stakeholders and beneficiaries, organizational and governance structures, and resource allocations.

Planned Results and Program Theory – includes a description (planned results and delivery strategy) and a graphical illustration (logic model) that shows how the activities of a program are expected to lead to the achievement of the planned results.

Monitoring and Evaluation Plan – a plan for ongoing performance measurement and evaluation activities. This component also includes a matrix of monitoring and evaluation reporting commitments.

 

additional emphasis on learning points.

An RMAF should be concise and focussed. This will help to support its implementation.

 

Checklist  Checklist for Developing an RMAF

Required Element

1.  Establish RMAF working group

2.  Assess internal capacity

Prepare Description

3.  Profile

4.  Planned Results and Program theory (planned results, delivery strategy and logic model)

Prepare Monitoring Plan (or Ongoing Performance Measurement Plan)

5.  Determine indicators (using logic model as guide)

6.  Determine data sources, data collection methods and timing

7.  Identify responsibility for data collection

8.  Estimate costs for monitoring activities

Prepare an Evaluation Plan

9.  Establish understanding of why evaluation is being carried out

10.  Determine issues and evaluation questions

11.  Cover issues of relevance, success and cost effectiveness

12.  Consider Expenditure Review Committee questions

13.  Determine appropriate evaluation design, data collection methods, data source, and frequency

14.  Identify responsibility for data collection

15.  Estimate costs for evaluation activities

Prepare a Reporting Strategy

16.  Identify all monitoring and evaluation reports (include DPR, RPP, annual performance report, compliance audit and a summative evaluation)

17.  Indicate timeframe for reporting performance information

18.  Indicate responsibility for reporting the performance information and evaluation results

19.  Indicate who will use the report

4.2.2  Profiling the Program

A clear understanding of the organization and the program is needed to guide monitoring and evaluation activities. The profile typically includes a summary of the context, objectives, key stakeholders and beneficiaries, organization and governance structures, and resources. The profile should provide a clear understanding of what the program aims to achieve and how.

Typical Profile Components*

Context

Clearly outlines the need and rationale for the program.

Objectives

Clearly states the objectives of the program. Describes how the objectives link to the department's strategic outcomes as identified in its Program Activity Architecture.

Key Stakeholders and Beneficiaries

This section of the profile should provide the reader with a precise understanding of who is involved in the program. Programs may involve many stakeholders with different roles, perspectives, and management information needs. If information is available, identify targets in terms of reach to project beneficiaries.

Organization and Governance Structures

Describes the organization and governance structures. Identifies decision-making authority and main roles and responsibilities of all project stakeholders (including delivery partners).

Resources

Identify annual resources allocated to the agency and each delivery partner (where applicable). Specify costs for monitoring and evaluation activities.

*  In the past, profiles have typically included planned results and the delivery strategy. These components may now be included in the Planned Results and Program Theory (Logic Model) section.

 

Example:
The Canadian forces Grievance Board (CFGB)

The following is an excerpt from the CFGB's Profile regarding governance structure:

The Board is presently made up of a Chairperson, a full-time Vice-Chairperson, a part-time Vice-Chairperson and three part-time Members. All are appointed by the Governor-in-Council, for terms that initially do not exceed four years.

Grievance Officers, working in the Grievance Analysis and Operations unit are responsible for analyzing grievances, conducting research, including the research of relevant jurisprudence, and drafting the initial findings and recommendations on grievances, in order to assist the Board Members in their work. Lawyers in the Legal Services unit are responsible for conducting a legal review of the findings and recommendations before submission to the Board Members, and the Board Members are accountable for the findings and recommendations that are submitted to the Chief of Defence Staff. The Chairperson of the Board is ultimately accountable for the work of the Board.

The Executive Director, who oversees the delivery of corporate support services, is accountable for the overall sound management of the Board, including its financial management. However, the Chairperson is ultimately accountable for all facets of Board management.


 

Checklist  Checklist for Developing a Profile for an RMAF

Considerations

1.  Have you consulted appropriate strategic and descriptive documents?

2.  Have you consulted with appropriate stakeholders to obtain missing or additional information?

Does the profile…?

3.  Include the main components (context, objectives, key stakeholders and beneficiaries, organization and governance structures, and resources)?

4.  Provide a clear understanding of what the program intends to achieve as well as an appreciation for how it intends to do so?

5.  Clearly describe the context for the program?

6.  Explain need and relevance?

7.  Fully, but concisely, describe the program? (5-7 pages as a general rule)

8.  Use neutral language? (avoid cheerleading)

9.  Identify the scope and magnitude of the program?

10.  Describe how the objectives link to the agency strategic objectives as identified in its Program Activity Architecture?

11.  Provide a clear statement of the roles and responsibilities of the main stakeholders (including delivery partners)?

12.  Outline governance structure from the perspective of accountability?

4.2.3  Developing the Logic Model

 

What is a Logic Model?

A logic model is a diagram or picture that shows the causal links from the activities to the results. Logic models illustrate the cause-effect relationship between activities and outputs through to the final results. It is a visual way of expressing the rationale, thought process or theory behind an organization, program or initiative. It is a representation of how the organization or initiative is expected to lead to the results.


A logic model can be applied to an organization, policy, program or initiative. It can be used for the purposes of planning, project management, evaluation and communication.

additional emphasis on learning points.

Logic models can help clarify objectives and focus the evaluation on results

 

Components of a Logic Model

While logic models can vary considerably in terms of how they look, they typically have three main components – activities, outputs, and results.

Components

Key Attribute

Description

Activities

What we do

The main actions of the project.

The description may begin with an action verb (e.g., market, provide, facilitate, deliver).

Outputs

What we produce

Outputs are the tangible products or services produced as a result of the activities. They are usually expressed as nouns. They typically do not have modifiers. They are tangible and can be counted.

Results

Why we do it

Results are the changes or the differences that result from the project outputs. Note that there can be up to three levels of results (immediate, intermediate, and ultimate or final). Results are usually modified (e.g., increased, decreased, enhanced, improved, maintained).

Immediate Results

Those changes that result from the outputs. These results are most closely associated with or attributed to the project.

Intermediate Results

Those changes that result from immediate results and will lead to the ultimate outcomes.

Ultimate Results

Those changes that result from the intermediate results. Generally considered a change in overall "state." Can be similar to strategic objectives. Link final results to the agency's strategic results as specified in the MRRS.

 

Some logic models also include other features, such as:


 

Example: National Parole Board's (NPB) Logic Model
for the Aboriginal Corrections Component of the
Effective Corrections Initiative

Examples of short-term results

  • Communities are better informed about the NPB and conditional release.
  • Hearing processes for offenders in the Nunavut Territory are culturally appropriate.

Examples of long-term results

  • The conditional release decision-making process is responsive to the diversity within the Aboriginal offender population.
  • The NPB has better information for decision making, including information on the effects of their history, when conducting hearings.

 

Are there Different Types of Logic Models?

Logic models vary considerably in terms of how they look. They can flow horizontally or vertically. The logic model type you choose should be appropriate to your agency and to your stakeholders. Whatever type is chosen, the model should provide sufficient direction and clarity for your planning and evaluation purposes. Flow charts or tables are the most common formats used to illustrate logic models.

Note that the logic model, irrespective of the types described below, will help to focus the evaluation on the results of your program.

 

TYPE 1: Flow Chart or Classic Logic Model

The flow chart or classic logic model illustrates the sequence of results that flow (or result) from activities and outputs. It is a very flexible logic model as long as the three core components of the logic model are presented: activities, outputs, and results. You can have any number of result levels to ensure that your logic model accurately depicts the sequence of outcome results.

The cause-effect linkages can be explained by using "if-then" statements. For example, if the activity is implemented, then these outputs will be produced. If the immediate result is achieved, then this leads to the intermediate result, and so on.


 

 

additional emphasis on learning points.

The flow chart logic model makes you think carefully about the linkages between specific activities, outputs and outcomes. What outputs result from each activity? What outcome resulted from the output?

 

TYPE 2: Results Chain Model

This type of model is also referred to as a performance chain. While it is similar to the flow chart model, it does not isolate the specific activities, outputs or results. The results chain, therefore, does not show the same detail with respect to the causal sequence of outputs and results.

Both types of logic models, however, are used as a structure for describing the expectations of a program and as a basis for reporting on performance. Like the flow chart model it is based on the rationale or theory of the program.

 


SourceSix Easy Steps to Managing For Results: A Guide for Managers , April 2003, Evaluation Division, Department of Foreign Affairs and International Trade.

Other Considerations

 



How Do I Build a Logic Model?


The following graphic presents an overview of the three steps for logic model development.

See Appendix E for more detailed information on building logic models.

 

Is my agency ready to build a logic model?

  • Is there sufficient time and commitment to develop the logic model internally?
  • Is there familiarity with respect to logic model development?
  • Are there sufficient planning and communication skills (key to building consensus and obtaining commitment)?
  • Is there sufficient objectivity or neutrality?
  • Does the program involve only my Agency in the federal government?

If you answered "yes" to these questions, you are probably ready to build a logic model. For details on how to build a logic model, please refer to Appendix E.

If you answered "no" to any of the first four questions, you may wish to contract out the development of the logic model.

If you answered "no" to the last question, then the initiative may be considered a "horizontal initiative." There are typically more challenges to developing a logic model for a horizontal initiative since you have to involve more stakeholders with different perspectives and opinions.

For further information on RMAFs, see Preparing and Using Results-based Management Accountability Frameworks, April 2004.

 

4.2.4  Developing the Performance Measurement Monitoring Plan for an RMAF

While this is an evaluation guidebook, RMAFs also include a monitoring plan. This strategy should generate a timely flow of information to support decision making on an ongoing basis. It is important to note that some data required for evaluation purposes can be collected on an ongoing basis as part of the performance measurement system.

The indicators for the performance measurement strategy are developed from the logic model's outputs and results. For each indicator, the data source, collection method, timing and frequency of the data collection, and responsibility for measurement must be identified.


4.2.5  Developing the Evaluation Strategy

 

This subsection presents

  • an overview and description of an evaluation strategy at the program level;
  • development of evaluation issues and questions;
  • development of indicators;
  • an overview of evaluation designs; and
  • an overview of data collection methods.

 

Overview of an Evaluation Strategy

The evaluation strategy may be presented in matrix format similar to the example below.

Evaluation Issue

Evaluation Question

Indicator

Data Source

Data Collection Method

Timing

Success

To what extent has the initiative improved staff evaluation capacity?

Quality of evaluation reports

Evaluation experts

Expert review

Year 2

 

Articulating the Strategy Efficiently and Effectively to Meet the Agency's Needs

An evaluation strategy should balance the need for timely and credible information with the need for practicality. Good linkages between evaluation and performance measurement will help to ensure that performance information is used as a source of information for evaluations. Linkages between planning and evaluation will help to ensure that the evaluation strategy is appropriately focussed and directed towards information needs.

Some things to think about when developing an evaluation strategy:

 

Sample Evaluation Matrix:
The Canadian Forces Grievance Board (CFGB)

Evaluation Issues/Questions

Indicators

Data Source/Methodology

ISSUE 2.5 – Cost-effectiveness

1. Can the quality of CFGBs F&R be maintained at a lower cost and in less time?

Reduction of average cost per grievance since CFGBs establishment.

Level of satisfaction among key stakeholders with the quality of the F&R

Interviews with CFGBs senior operational managers

Interviews with key stakeholders (CDS, DG-CFGA, ADM HR-Mil)

Case Management and Tracking Systems

 

 

Developing Evaluation Issues and Questions

Evaluation issues are the broad areas which need to be explored within an evaluation while evaluation questions are the more specific research questions that need to be answered in order to be able to address each evaluation issue.

The identification of the evaluation issues and questions provides a guide for the development of the strategy that ensures all essential issues will be addressed during later evaluation. The issues are used to elaborate a set of indicators and data collection strategies, which, on implementation, helps to ensure that information necessary for evaluation is available when it is needed. As such, the evaluation strategy needs to be linked to the ongoing performance measurement strategy, as some evaluation data will be collected through ongoing performance measurement activities.

For information on how to develop a specific list of issues and questions, see Appendix E. For the main evaluation issue areas please refer to the following diagram.

Expenditure Review Questions

All evaluations should address the Expenditure Review Committee's Seven Areas to Question. In addition to the three traditional evaluation issue areas listed above, program spending will also be assessed against specific questions in relation to the following:

  • Public Interest
  • Role of Government
  • Federalism
  • Partnership
  • Value for Money
  • Efficiency
  • Affordability

See Appendix D for the specific tests.


Determining Appropriate Indicators

What are Performance Indicators?

Performance indicators are a direct or indirect measure of an event or condition. An indicator is a measuring device showing change over time. Indicators are often quantitative (i.e., based on numbers or objective information) but can also be qualitative (i.e., narrative or subjective information). The indicator is a means to compare planned results with actual results. There are many ways to think about indicators.

  • Proxy indicators. Proxy indicators are sometimes used to provide information on results where direct information is not available. For example, the percentage of cases that are upheld on appeal could be a proxy indicator for the quality of decisions.
  • Quantitative indicators. Quantitative indicators are statistical measures such as number, frequency, percentile, ratios, and variance. For example, percentage of Web site users who find and obtain what they are looking for.
  • Qualitative indicators. Qualitative indicators are judgment and perception measures of congruence with established standards, the presence or absence of specific conditions, the extent and quality of participation, or the level of beneficiary satisfaction, etc. An example would be opinions on the timeliness of services.
  • Output and result indicators. There are also output and result indicators. Output indicators are those indicators that measure the outputs (products and services). Result indicators measure the results or changes of a program.

 

An example of various indicators is illustrated in the table below. See Appendix E for a review of how to develop indicators for your agency.

Measure

Types

Indicator Examples

Outputs

Quantity Produced/Delivered/Served

Number of clients served per month

(If your agency produces policy papers or research studies, the output indicator might be number or quality of policy papers/research studies.)

Quality of service

Achievement of standards for service delivery

Client Satisfaction

Per cent of clients satisfied with product and service delivery

Efficiency

Average cost per unit delivered

Results

Immediate

Number of person-weeks of training and career placement projects completed

(If your output is quality of policy papers, then an immediate result might be "increased awareness of the policy" or "better incorporation of policy principles with other relevant programs/policies.")

Intermediate

Number of successful job placements resulting from training and career projects

Ultimate

Individuals' self-rated health status in terms of well-being and functional abilities

Source :  Adapted from First Nation Self-Evaluation of Community Projects: A Guidebook on Performance Measurement .

 

Overview of Evaluation Designs

Evaluation design is the process of thinking about what you want to do and how you want to go about doing it. 

additional emphasis on learning points. A good research design will:

  1. improve the reliability and consistency of your results;
  2. eliminate (or minimize) bias; and
  3. answer what you need to know.

 

The most practical approach to determining evaluation design is to consider your information needs (i.e., evaluation questions) and use this to guide your design. Key considerations for selecting an appropriate design are feasibility and practicality .

Evaluation designs are typically placed into the following three categories:

Implicit designs are the most frequently used evaluation design. In the public service context, it is often the only design that can be used, when no pre-program measures exist and there is no obvious control group available, or it is not reasonable to assign interventions on a random basis. This design type is flexible and practical to implement.

It should be noted, however, that there are considerably more challenges in attributing impacts to specific interventions as we move away from experimental designs (the strongest for attributing impacts) to implicit designs.

Checklist  Checklist for Choosing an Evaluation Design

Considerations

Have you considered the following…?

1.  Information and decision-making needs

2.  Type of evaluation

3.  Practicality and costs

4.  Appropriate balance between information needs and costs

5.  Research concerns (i.e., related to the quality of evidence to be gathered)

6.  Other internal and external factors that may influence the program. How can the evaluation design minimize these factors?

7.  Targeted evaluation questions (i.e., those that take into account the most pressing evaluation concerns)

8.  Consider existing data, secondary data, and performance measurement information as potential sources of information for the evaluation

9.  Use multiple lines of evidence to ensure reliability of findings and conclusions

 

Overview of Data Collection Methods

 

excerpts from TB Evaluation PolicyREMEMBER...

Measurement and Analysis: Evaluation work must produce timely, pertinent and credible findings and conclusions that managers and other stakeholders can use with confidence, based on practical, cost-effective and objective data collection and analysis.

— TB Evaluation Policy, 2001

The table below provides an overview of various data collection methods available to evaluators. Note that these data collection methods involve either primary or secondary data. The investigator collects primary data directly. Secondary data have been collected and recorded by another person or organization, sometimes for different purposes.

In choosing appropriate data collection methods you can consider the following:

More information on choosing appropriate data collection methods is located in Appendix E.

 

Data Collection Method

When to Use

External Administrative Systems and Records: use of data collected by other institutions or agencies

  • Need information about context
  • Need historical information
  • When comparing program data to comparable data

Internal Administrative Data: data collected for management purposes

  • Need information on management practices, service delivery, clients' characteristics

Literature Review: review of past research and evaluation on a particular topic

  • To identify additional evaluation questions or issues and methodologies
  • Need information on conceptual and empirical background information
  • Need information on a specific issue
  • Need information about comparable programs, best practices

Interviews: a discussion covering a list of topics or specific questions, undertaken to gather information or views from an expert, stakeholder, and/or client; can be conducted face to face or by phone

  • Complex subject matter
  • Busy high-status respondents
  • Sensitive subject matter (in-person interviews)
  • Flexible, in-depth approach
  • Smaller populations

Focus groups: a group of people brought together to discuss a certain issue guided by a facilitator who notes the interaction and results of the discussion

  • Depth of understanding required
  • Weighted opinions
  • Testing ideas, products or services
  • Where there are a limited volume of issues to cover
  • Where interaction of participants may stimulate richer responses (people consider their own views in the context of others)

Case studies: a way of collecting and organizing information on people, institutions, events, and beliefs pertaining to an individual situation

  • When detailed information about a program is required
  • To explore the consequences of a program
  • To add sensitivity to the context in which the program actions are taken
  • To identify relevant intervening variables

Questionnaire or Survey: a paper or electronic list of questions designed to collect information from respondents on their knowledge and perceptions of a program (See Appendix E.)

  • Useful for large target audiences
  • Can provide both qualitative and quantitative information

Expert panels: the considered opinion of a panel of knowledgeable outsiders

  • Experts can share lessons learned and best practices
  • Where outside validation is required
  • Where diversity of opinion is sought on complex issues
  • Where there is a need to draw on specialized knowledge and expertise

Comparative studies: a range of studies which collect comparative data (e.g., cohort studies, case-control studies, experimental studies)

  • For summative evaluations

 

Depth vs. Breadth

Some data collection methods provide more depth of information, while others provide more breadth .

  • Depth – understanding of impact of program on an individual person or case
  • Breadth – understanding of impact of program on large group of people, but in less detail

For example, case studies provide depth of information while surveys provide more breadth. Each type of information is important for an evaluation depending on the specific questions being asked, and the integration of the various methods. Many evaluators attempt to combine methods that will provide both depth and breadth to the findings.

 

4.3  Collecting the Information

4.3.1  Gathering Data

Data collection should follow the plans developed in the previous step. The individuals assigned to the various data collection tasks need to be thoroughly trained in the data collection requirements and procedure.

Appropriate quality control procedures should be implemented and maintained during the evaluation study. If you are managing an evaluation you need to be aware of the progress of the data collection and any other issues of concern.


To facilitate analysis, information collected during an evaluation should

4  use appropriate methods to organize and record the information collected (e.g., frequency distributions, categories, tables); and

4  implement effective quality control procedures to ensure recorded information is accurate and original information is labelled and secure.

4.3.2  Analyzing Data

Once data are collected, they need to be analyzed and interpreted. Data analysis may take many forms from basic description to complex statistical analysis depending on the type of data and the complexity of the issues. For more detail as to how to analyze data, please refer to Appendix E.

Cause and Effect Inferences

The choice of analysis techniques is influenced by the evaluation questions and the evaluation design (i.e., experimental or implicit). For example, drawing inferences about causality is dependent upon the evaluation design rather than the analysis technique.

Generalizing the Findings

The only valid way of generalizing findings to an entire or target population--where you cannot survey or study everyone--is to use findings from a statistically representative random sample of the population you wish to study. Caution must therefore be exercised when analyzing data from non-randomized samples.

Qualitative and Quantitative Analysis

Analyzing qualitative data requires effective synthesis and interpretative skills. Qualitative information can be used, for example, to provide contextual information, explain how a program works, and to identify barriers to implementation. Qualitative data can be analyzed for patterns and themes that may be relevant to the evaluation questions. Qualitative material can be organized using categories and/or tables making it easier to find patterns, discrepancies, and themes.

Quantitative data analysis assigns numerical values to information. It can range from simple descriptive statistics (e.g., frequency, range, percentile, mean or average) to more complicated statistical analysis (e.g., t-test, analysis of variance). Computer software packages such as Statistical Package for Social Sciences (SPSS), Minitab, and Mystat can be used for more complicated analysis. Quantitative data analysis also requires interpretation skills. Quantitative findings should be considered within the context of the program.


Checklist  Checklist for Analyzing Data

Considerations

1.  Understand the problem before you analyze data (i.e., know what is being measured and why)

2.  Understand the program and how contextual factors link together

3.  Find out how the data were collected and how reliable they are

4.  Use your common sense; ask yourself if the analysis and interpretation seem appropriate

5.  Try to identify patterns, associations, and causal relationships

6.  Utilize statistical analyses when appropriate

7.  Are there any deviations in these patterns? Are there any factors that might explain these deviations?

8.  Compare findings to expected results (i.e., industry standards)

9.  Consider strengthening your analyses by combining evaluation data with risk data collected from periodic environmental scans

10.  The logic of each method of analysis should be made explicit (e.g., specify what constitutes reasonable evidence, identify underlying assumptions)

11.  Where possible, use several methods of analysis

12.  Use appropriate tests of significance whenever findings are generalized to the population from which samples were drawn*

13.  Use caution when generalizing evaluation results to other settings

*  In advance of gathering data, the evaluator needs to calculate the probability that the findings are not "accidental." With tests of significance an evaluator can decide how strong the results must be in order to be reasonably confident that the results are not due to chance.


4.4  Writing the Evaluation Report

excerpts from TB Evaluation PolicyREMEMBER...

Evaluation reports must present the findings, conclusions and recommendations in a clear and objective manner.

— TB Evaluation Policy, 2001

A good evaluation report responds effectively to the evaluation questions. Recommendations and lessons should be conclusive, concise, and practical. The executive summary should be a summary of the overall report. Often the executive summary is the most widely read section of the report so it should be detailed enough to give the reader a good sense of the highlights of the evaluation.

4.4.1  Table of Contents

An evaluation report typically contains the following sections:

4.4.2  Linking Findings, Conclusions, and Recommendations

There needs to be a clear link between findings, analysis, conclusions and recommendations. Practically speaking, the findings may not answer specific evaluation questions conclusively. Conclusions are formulated by combining the best evidence. Gathering different types of evidence relating to the same evaluation question can enhance credibility. Recommendations should link to the analysis and the conclusions.

Checklist Checklist for Evaluation Report Writing

Considerations

1.  Have the audience(s) and required information needs been identified?

2.  Is the report clear and concise?

3.  Are the reasons for carrying out the evaluation logical and clear?

4.  Does the report identify evaluation issues in accordance with evaluation policy (i.e., relevance, success and cost-effectiveness)?

5.  Does the report start with the most important information? (Each chapter, subsection, or paragraph should begin with the key point.)

6.  Is the context adequately explained?

7.  Is there a description of the general approach used, main data sources, data collection methods?

8.  Does the report clearly articulate the limits of the evaluation in terms of scope, methods, and conclusions?

9.  Are the findings substantiated by the evidence, as described in the evaluation report?

10.  Do the findings provide a good understanding of what was learned from this evaluation?

11.  Is it clear how the subject program or project is really performing?

12.  Does the presentation of results facilitate informed decision making?

13.  Is only the information that is needed for a proper understanding of the findings, conclusions, and recommendations included?

14.  Do the conclusions address the evaluation questions and are they supported by the findings?

15.  Are recommendations realistic and doable? Are the number of recommendations limited based on significance and value?

16.  Does the report present the conclusions and recommendations so that they flow logically from evaluation findings?

17.  Is the report in accordance with external reporting requirements?

18.  Does the report provide an accurate assessment of the results that have been achieved?

19.  Does the report provide relevant analysis and explanation of the exposure to risks for any significant problems identified and in respect of key recommendations?

 

Key References

CIDA, Performance Review Branch. CIDA Evaluation Guide , 2004.

Canadian Evaluation Society. Evaluation Methods Sourcebook , 1991.

National Science Foundation. A User-friendly Guide to Mixed Method Evaluations , 1997.

Forest Research Extension Project. Conducting Project and Project Evaluations: A Primer for Natural Resource Project Managers in British Columbia , 2003.

Gray & Guppy. Successful Surveys: Research Methods and Practice , 2003.

McLaughlin and Jordan. Logic Models: A Tool for Telling Your Program's Performance Story , 1999.

Stufflebeam, D.L. A Checklist Organizer. Guidelines for Choosing and Applying Evaluation Checklists:

Treasury Board of Canada Secretariat. Evaluation Policy and Standards , 2001.

Treasury Board of Canada Secretariat. The Art and Architecture of Writing Evaluation Reports , 2004.

Treasury Board of Canada Secretariat. Guide for the Review of Evaluation Reports , January, 2004.

Treasury Board of Canada Secretariat. Principles for the Evaluation of Programs by Federal Departments and Agencies , 1984.

Treasury Board of Canada Secretariat. Preparing and Using Results-based Management and Accountability Frameworks , April 2004.

Treasury Board of Canada Secretariat. Program Evaluation Methods: Measurement and Attribution of Program Results.

Treasury Board of Canada Secretariat. RBM E-Learning Tool.

 



 

Section Five: Managing Evaluations

 

 

This section provides advice on

  • when to utilize external resources;
  • how to prepare for evaluations (Roles and Responsibilities and Preparing Terms of Reference);
  • how to decide on contracting options;
  • choosing consultants; and
  • best practices for managing consultants.

5.1  When to Utilize External Resources


This subsection outlines guidelines for determining when to use external resources. Following are the two basic questions to be answered when deciding whether to employ external resources in an evaluation project:

  1. Are there staff members with the requisite background and knowledge available within the agency to conduct the evaluation?
  2. Is there a need to employ an external consultant to maintain objectivity or the appearance of such?

The first factor relates to the availability of the internal evaluation resources. When an agency lacks specific experience or knowledge, it may wish to contract for that aspect of the evaluation. In such circumstances, it may be useful to set up the contract so that there is some knowledge transfer to the agency. An agency may also lack the resources to do things like telephone surveys. Finally, an agency simply may not have the resources available internally to conduct the evaluation project and will have to resort to employing an external resource.

For the second factor, the agency needs to consider the use and target audience of the report and the risk associated with using an internal versus an external resource. Staff involvement in some evaluation tasks may bias the results (e.g., focus groups, administering a survey on client satisfaction) and for these specific tasks, an agency may wish to employ an external resource. For controversial programs where the report will be used as a public accountability tool, employing external resources may mitigate risks.

Checklist  Checklist to Assist in Deciding
When to Utilize External Resources

Considerations

P

1.  Are there sufficient funds designated for evaluation purposes?

 

2.  Has similar work been undertaken in-house? (e.g., previous evaluations of similar programs)

 

3.  Is there sufficient time and commitment to conduct the work?

 

4.  Is the information available from other sources?

 

5.  Are there existing measures or indicators of performance currently in place?

 

6.  Are existing program practices and methods of information collection useful for evaluation purposes?

 

7.  Is there sufficient objectivity to conduct evaluation work internally?

 

8.  Are there management or staff members who have training and experience in evaluation-related tasks?

 

Source :  Adapted from Who Should Conduct Your Evaluation? http://www.bja.evaluationwebsite.org/guide/documents/chapter_3_housing.htm

 

If you answered "yes" to Question 1, but "no" to all other questions, then you will likely need external assistance in conducting your evaluation. If you answered "no" to Question 1 but "yes" to most of the other resource questions, then an in-house evaluation may be a good choice.

Have you considered other options?

One option is to contract out portions of the evaluation or study to consultants while some evaluation tasks are performed in-house. This would also help an agency develop their internal capacity. Management and staff could be involved in different evaluation activities. Following are some sample activities.

  • Managers, staff, key stakeholders, and/or partner organizations may participate in conducting a literature review and/or secondary research.
  • Management, staff (or even graduate students) could conduct a systematic examination of available data.
  • Consider integrating additional data collection into current service delivery mechanisms or products.

— Adapted from Program Evaluation Kit: First 5 LA .

5.2  Preparing for Evaluation

In this subsection, two important steps to preparing for an evaluation will be described. They are as follows:

5.2.1  Roles and Responsibilities

Project Coordinator

You will need to decide who the project coordinator will be before proceeding with the evaluation. If the coordinator is to be managing an external consultant, he/she should be familiar with the project, understand the basics of evaluation, and have good project management skills.

Advisory Committee

You may also need to set up an advisory committee comprised of agency and other stakeholder representatives that have an interest in the project or organization being evaluated. It should be noted, however, that setting up an Advisory Committee on an as needed basis may be very resource intensive for a small agency. The need for an Advisory Committee must therefore be weighed very carefully. Where appropriate, a small agency may want to consider existing work teams to act in an advisory capacity.

Tips for setting up an Advisory Committee

  • The roles and responsibilities of the advisory committee members should be clearly laid out.
  • There should be one designated primary contact person on the committee.
  • Outline methods and frequency of communication and allow for ad-hoc meetings.
  • Address how formal committee documents will be prepared, distributed, and approved.
  • Provide for review and amendment of the Terms of Reference.

 

Steering Committee

A Steering Committee is usually comprised of senior managers and might also include central agency representatives and, where appropriate, regional representatives. When you want the advisory committee to have decision-making powers (i.e., final word on evaluation deliverables), then you may consider setting up a Steering Committee.

When would you use an Advisory Committee?

  • When technical advice is needed and evaluation findings need to be situated within the overall policy context and agency environment.

When would you use a Steering Committee?

  • When you want to provide the opportunity for senior management to indicate its support for the project. When you want the committee to have decision-making powers and when the evaluation is high profile in nature, you may want to consider a Steering Committee.

Many smaller agencies may not require advisory or steering committees to support the evaluation.

Advisory or steering committees may not be necessary in the following situations:

  • Senior management has provided clear direction as to the evaluation's TOR; and
  • Managers have a clear knowledge of organization policy, context and appropriate sources of information.

 

5.2.2  Setting the Evaluation's Terms of Reference (TORs)

TORs provide an overview of the evaluation and they make explicit management's initial requirements and expectations for the evaluation. The TORs are a useful tool for senior management in that they can guide the process until the evaluation work plan becomes the primary document.

TORs are used for many purposes.

To prepare the terms of reference, you may consider the following:


 

Main Elements of the Terms of Reference

Program Background

Program context and rationale

Identification of key stakeholders, clients, and partners

Program description

Reasons for the Evaluation

Statement of purpose of the study

Expected value-added

Intended use of results

Scope and Focus

Broad issues to be addressed/specific evaluation questions

Type of analysis to be used/level of detail

Specify the audience(s) for the reports and findings

Statement of Work

How purposes of study are to be achieved

Describe approaches

Describe data collection methods

The tasks required to undertake study

What groups will be consulted

Expectations with respect to communications and ongoing progress reports

Evaluation Team

Required professional qualifications/expertise/experience

Role and responsibilities of evaluation team, role of agency (program and/or evaluation managers)

Timetable

Approximate timetable to guide the preparation of the work plan

Budget

A specification of the estimated resources to be committed to the study and its different parts

Deliverables

Identification of key deliverables (e.g., work plan or methodology report, draft evaluation report, final evaluation report)


 


Guidelines for Developing TORs

  • Be clear about the scope of the evaluation and your expectations.
  • Accurately describe the rationale for the program being evaluated.
  • Include specific evaluation questions.
  • Identify the separate manageable project tasks and the results that are expected.
  • Do not assume that data sources will be available or accessible. Check to make sure the methodology is feasible.
  • Clearly identify the specific abilities, qualifications and skills required to carry out the study.
  • Establish expectations for deliverables, work scheduling and costs.
  • Plan accordingly as writing TORs can be time-consuming (i.e., 2 to 5 days).
  • Review other TORs for ideas.

 

The terms of reference provide the basis for the next step in the process – selecting the evaluator or evaluation team. Evaluators are selected through the established contracting process.

Contracting externally for an evaluation team involves the following four steps:

1.  determine the sourcing options;

2.  identify the best value from potential candidates;

3.  notify the successful candidate (posting on MERX for competitive process); and

4.  negotiate and sign the contract.

Bidders often have questions about the Terms of Reference. You should designate one contact person to co-ordinate the responses to these questions. You must ensure that you provide the same information to all proponents and that you do not identify the source of the question. You may want to consider faxing or e-mailing responses to all potential bidders.

Guidelines for Budgeting

Budgeting for evaluation should be part of the upfront planning. It is important to identify a budget amount or range in the TORs. It also allows you to judge which proposals offer the best value per dollar.

Typically, budgeting involves breaking down the evaluation into components or tasks and providing estimates for each component. You may consider costs with respect to evaluation staff, consultants, travel, communications, printing and photocopying, supplies and equipment, and translation and editing.


Guidelines for Budgeting for Evaluations

  • The evaluation budget is typically between 1 and 3 per cent of the program's overall budget.
  • Consider costs of previous or similar evaluations.
  • Consider trade-off between evaluation quality and budget.
  • Data collection typically takes up about half the budget; the other half of the evaluation budget goes to evaluation design and reporting.

 

Leveraging Resources for Evaluation

A small agency can supplement its limited resources by leveraging its needs and resources. While not all strategies might be applicable to your agency, here are some suggestions:

Preparing the Statement of Work

The statement of work is the main document between the manager and the consultant. The consultant responds to this document by developing a work plan. It outlines the study purpose and objectives, approach, data collection methods, and tasks. If an RMAF or evaluation framework has been prepared, this may be used as a guide when preparing the Statement of Work. As you may recall from the previous section, an RMAF outlines the sources of information and data collection methods for each evaluation question.

One key decision is the level of detail you may want to include in this section. You may choose a less detailed statement of work relying on the proponent to suggest specific approaches and methodologies. However, a detailed statement of work is useful as it gives the proponent a better idea of your expectations. It is also important to keep in mind that the proponent can suggest changes to approach and methods. It is helpful to request that the proponents include a discussion of possible methodological challenges and solutions. This may help you assess the proponent's expertise and may also serve to strengthen the proposed work plan.

5.3  Contracting Options

If you do decide to use external resources you then have to consider the contracting options. You can review the Contracting Policy on the TBS Web site at
http://www.tbs-sct.gc.ca/pubs_pol/dcgpubs/contracting/contractingpol_2_e.asp .

These options include the following:

TBS Standing Offer for Small Agency Evaluation Services. (This option is under development at time of printing.)

Proposal calls on MERX – Canada's Official Electronic Tendering Service.

Selective Tendering – The requirement for bids may be set aside when the total costs of the contract (including GST) do not exceed $25,000. In such cases where a fully open competitive process would not really be cost effective, a limited tendering approach may be used. This is done by inviting a minimum of three firms from a source list to bid. This approach is useful for obtaining the best value for money and providing a fair and competitive access to government procurement. Where can you locate sources? Consider the branch's corporate memory and contractor inventory along with the contracting authority's corporate memory, inventory and short lists obtained from other departments and agencies. In this situation, a formal Request For Proposals (RFP) is not required.

Sole Source – used in cases where only one person or firm is capable of performing the job.

Standing Offer – can be used for services of a repetitive nature.

It should be noted that the Government Contracts Regulations contain only four exceptions that permit the contracting authority to set aside the requirement to solicit bids.

Please Note: You may opt to secure the services of Consulting and Audit Canada (CAC). With respect to federal government agencies, there is a Memorandum of Understanding rather than a contract. Federal government contracting rules do not apply when this option is used.

A Request for Proposals (RFP) includes the following:

  • Terms of Reference
  • Statement of Work
  • Provisions of the contract
  • Standard contract clauses
  • Proposal format requirements
  • How proposals will be rated and selected
  • Submissions procedures and deadlines
  • Contact information for questions or clarification

5.4 Choosing Consultants

 

When reviewing proposals, think about the following questions

Does the evaluator possess knowledge, skills, and experience in

  • the application of sound research design to answer the chosen questions?
  • the collection and analysis of reliable quantitative and qualitative data?
  • the development of valid, credible, and unbiased conclusions and recommendations?

 

The terms of reference provides the basis for choosing the consultant. It is important to develop the criteria to be used for selecting a consultant during the planning stages (i.e., before sending out the RFP).

Mandatory and Rated Requirements

The criteria you use to select the proposal should reflect the requirements of the specific project. The RFP typically contains a set of mandatory and rated criteria. To be considered, a proposal must meet the mandatory requirements set out in the RFP. Mandatory requirements set the minimum requirements for the bid to be considered. They are assessed on a simple met/not met basis. These requirements are expressed by using clear and strong terms such as "shall," "must" and "will." When mandatory requirements are used, the RFP shall clearly indicate that failure to meet any of the mandatory criteria will render the bid non-compliant and that it will be given no further consideration.

Typical Mandatory Requirements

  • minimum experience level in area of work
  • documentation requirements (e.g., résumés, references, samples of work)
  • language requirements
  • security requirements
  • budget (maximum amount)
  • certification (e.g., state that the information provided is accurate)
  • availability (will be available to do work at time contract is awarded)

 

Proposals that meet the mandatory requirements may go on to be rated against other criteria and may be required to achieve a minimum level for the point-rated requirements to be given further consideration (e.g., score at least 70 points out of 100 to move to the next step in the evaluation process). Each criterion must have associated scoring criteria (the basis for scoring). Contractors can be asked to provide evidence to demonstrate their capacity to respond to a criterion (e.g., résumé).

When you are deciding what rated criteria to include in the RFP, you may want to consider what the most important factors are that will affect the quality of the work to be done (e.g., subject matter expertise, methodological experience, academic qualifications).

Quantitative rating scales are a useful decision-making device. However, these scales have to be constructed and weighted very carefully to ensure you are selecting the right consultant. The rate assigned to each criterion should reflect its importance.

Typical Rated Requirements

  • Firm/resources (e.g., qualifications, knowledge, experience, abilities, skills, references). The rated requirements can be very specific (e.g., experience conducting surveys).
  • Criteria related to the proposal (e.g., understanding of work, approach, management).
  • Cost (e.g., price of bid)

 

Hexagon: Did you know? Bidders may also be invited to make verbal presentations subsequent to the proposal submission.


5.4.1  Methods of Selection

The following methods are commonly used to assess proposals. The method to be used should be articulated in the RFP.

1.  The lowest price proposal method. This may be used when cost is the most important factor. This method can be used when mandatory requirements are the main tool for evaluating the proposals and suppliers offer uniform services.

2.  The best technical proposal method. The proposal that receives the highest score within budget is the winner. This method may be used when technical merit is the most important factor. A maximum budget is identified in the RFP. The point-rated requirements are typically used to assess proposals.

3.  The best overall value for money method or cost-per-point method. The proposal that has the best ratio score/price is the winning proposal. It is used when technical merit and price are both important factors .


5.4.2  Criteria to Consider When Reviewing Proposals

When reviewing the proposal, you need to ask yourself the following question: Does the evaluator have a proven ability to deliver results and meet expectations? Other criteria are presented in the checklist below.

Checklist  Checklist for Reviewing Proposals

Considerations

P

1.  Qualifications of the evaluation team

 

2.  Professional background and experience (e.g., federal government evaluation experience)

 

3.  Personal qualities: ability to communicate, teamwork capabilities, leadership skills. Note that leadership skills are particularly important when dealing with numerous stakeholders and partners.

 

4.  Evaluation skills (e.g., knowledge and practical application of evaluation methodologies)

 

5.  Subject matter expertise

 

6.  Demonstrated performance levels (check references)

 

7.  Avoid boiler-plate solutions. The proposed approach should be relevant to the needs of your agency. The bidder should demonstrate an understanding of the agency's needs and challenges.

 

8.  Proposal demonstrates that they can do the project within the designated time frame.

 

9.  Soundness of methodology and work plan

 

10.  Demonstrates understanding of challenges and methodological limitations

 

11.  Balance of junior and senior evaluators

 

12.  Existence of qualified back-up personnel (particularly important when using smaller firms)

 

 

5.5  Best Practices for Managing Consultants

Establishing clear expectations at the beginning of a project is crucial to its success. This subsection considers best practices in terms of scope of work, managing the project, and follow-up.

The following checklist provides an overview of the management process from its inception to the reporting of findings.

Checklist  Checklist for Working with Consultants

Considerations

P

1.  Set up initial meeting with the consultant and steering or advisory committee members (if appropriate).

 

2.  Review evaluation scope, objectives, work plan and timelines.

 

3.  Review, as necessary, the Terms of Reference and/or Statement of Work with all parties.

 

4.  Set up a communications plan with the consultant and other key stakeholders for the life of the project (i.e., this refers to a feedback process for sharing information on reports or data submitted by the consultants).

 

5.  Inform key stakeholders about the nature and purpose of the project.

 

6.  Alert consultant to confidential or sensitive issue requirements at the outset.

 

7.  Sign a formal contract with the revised Terms of Reference appended.

 

8.  Ensure that program data and any other information necessary are available to the consultants.

 

9.  Ensure contractors understand that they are working for and reporting to the evaluation function, not program management.

 

10.  Plan for interim reports to monitor progress.

 

11.  Adjust budget as necessary.

 

12.  Review final products to ensure that they are consistent with requirements and agreed upon expectations.

 

13.  Ensure all contributors are recognized and thanked.

 

14.  Debrief consultant and stakeholders and assess the evaluation.

 


5.5.1  Defining the Scope of Work

The first planning meeting or initial start-up meeting is critical for clarifying expectations and the evaluation work plan. The project manager should also ensure that the consultant has all the information needed to carry out the project – relevant documents and contact information.

For the start-up meeting you should

ensure that the evaluator(s) have full access to files, reports, publications and any other relevant information;

ensure there is adequate administrative and logistical support during the evaluation;

establish project management and reporting expectations (how often and in what format (written, oral); and

identify the major issues and priorities.

 

5.5.2  Preparation of the Work Plan

The evaluation work plan or methodology report will likely be the first deliverable produced by the consultant. The work plan should provide a clear description of what the evaluation team is expected to do, as well as where, when, how, and why. The work plan typically builds on the proposal and statement of work. Some changes to the original proposed work plan may be suggested given new information or other considerations.

Here are some key elements of the work plan:

5.5.3  Overseeing the Day-to-Day Operations with Consultants

The relationship between the consultant and evaluation manager is a partnership in the sense that you bring subject matter expertise and he/she brings evaluation expertise. However, the main role of the evaluation manager is to ensure the consultant follows the agreed upon work plan and provides a satisfactory level of quality. The evaluation manager must also address issues that the consultant might raise when preparing and implementing the work plan. While there is no single good strategy for managing consultants, the checklist below may offer some helpful considerations.

Checklist  Checklist for Best Practices in
Managing Evaluation Consultants

Considerations

P

1.  Set realistic time frames for deliverables

 

2.  Maintain an active role in the project

 

3.  Anticipate what might go wrong and develop strategies to deal with it

 

4.  Keep formal and informal lines of communication open. You should have a good sense of the status of the project at all times (i.e., status of data collection, status of preliminary findings)

 

5.  Establish a positive working relationship

 

6.  In addition to regular meetings and other contacts, another way to keep the project on track is through interim reports

 

7.  Regularly check on the progress of the work

 

8.  Realize that it is not unusual to have problems or misunderstandings during a project

 

9.  Address problems quickly as they arise

 

10.  Discuss any deviations from the TORs

 

11.  Provide timely and considered reviews of all reports (e.g., methodology reports, interim, and final); ask questions

 

12.  Keep key stakeholders informed about the progress of the project

 

13.  Ensure that there is sufficient time to review draft and final reports

 


5.5.Managing Follow-ups

In the 3 to 6 months following the completion of the evaluation report, you may develop a better sense of the usefulness of the evaluation. You may want to consider providing follow-up feedback to the evaluator in terms of the usefulness of the evaluation. This information serves to build the evaluator's capacity and benefits the agency by improving the available resource pool.

Key References

Canadian Evaluation Society. Evaluation Methods Sourcebook , 1991.

Financial Management Board Secretariat. NWT. Working Well With Consultants : http://www.gov.nt.ca/FMBS/documents/dox/Consultant%20Guide.pdf .

HRDC. Evaluation Tool Kit. User Guide on Contracting HRDC Evaluation Studies , 1999.

Treasury Board of Canada Secretariat. TBS Contracting Policies , 2003.

UNFPA, Office of Oversight and Evaluation. Planning and Managing an Evaluation , http://www.unfpa.org/monitoring/toolkit/tool5_4.doc .

 



 

Section Six: Using Evaluations

 

 

This section covers

  • developing a communications strategy;
  • guidelines for effective communication of evaluation results;
  • mechanisms for communicating findings;
  • communicating to managers and stakeholders; and
  • using evaluation findings – managing an action plan.

 

A successful evaluation is one that is used. Communications is an essential tool that can directly influence the application of results generated from evaluations. The adaptation of lessons learned and recommendations from evaluation findings increase with an efficient and systematic communication strategy.

In this section, strategies for communicating and using evaluation findings will be highlighted. Strategies for enhancing utilization of findings will also be discussed.

6.1  Communicating Evaluation Findings

 
Developing a communications strategy is a good first step in communicating evaluation findings. You may want to consider the following steps when developing a communications strategy:

 


Guidelines for Effective Communication of Evaluation Results

  1. Communicate results to key stakeholders; first, to those who are key decision-makers and then to other stakeholders.
  2. Present findings to stakeholders in person.
  3. Make report available in both official languages.
  4. Frame information and results according to information needs and to facilitate decision-making.
  5. Involve stakeholders throughout evaluation – this helps to avoid unnecessary surprises at the reporting stage.
  6. Communicate sensitive information with care.

 

How do you Communicate the Findings?

Consider some or all of the following:

  • detailed written report;
  • executive summary;
  • brochure on lessons and recommendations;
  • annual report;
  • other strategic documents (e.g., Departmental Performance Reports);
  • organizational newsletter;
  • public meeting;
  • lessons learned seminar, workshops; and
  • e-mail or Internet.

 

6.1.1  Communication to Managers

Communication of results to managers should consider their unique information and decision-making needs. Communications may focus on

Managers need a clear idea of the perspectives and opinions of different stakeholder groups. It is therefore important to report findings according to these various groups.

It may also be advisable to provide preliminary findings to program managers via a presentation. A presentation can offer the evaluator with additional information to more effectively communicate the findings and can also help prepare the management response to the evaluation.


6.1.2  Communication to Stakeholders

Stakeholder groups can include the public, special interest groups, partners, third party deliverers, other federal government departments and agencies, and provincial or municipal governments.

Stakeholders may want information with respect to the program or project's broader impact (e.g., on society) and value for money. They may also be interested in understanding the relationship of the project to overall government goals and activities (e.g., are there duplications?)

Appropriate strategies for widely disseminating results to stakeholders can include general meetings, seminars, conferences, Web sites, and annual reports.

6.1.3  Communication to Senior Managers [8]

Generally, senior managers require more "high-level" strategic information. Senior managers require an understanding of how the findings relate to strategic objectives. Any concerns with respect to legislation, regulations, and policy should be promptly communicated to this group. Findings relating to values and ethics of the organization may also be given emphasis when communicating to this group.

Remember that evaluations are generally submitted to the evaluation committee for approval.

6.2  Using Evaluation Findings

Remember the evaluation and program management cycle presented in Section Two? An effective feedback system needs to be incorporated within the program or project management cycle. Ideally, evaluation findings should be used to make informed decisions. In practice this may not happen. Why?

Sometimes lack of use relates to poor communication of the evaluation findings. At other times, the evaluation report may lack relevance to the manager's information and decision-making needs.

Effective strategies for use of evaluation reports begin in the planning stage. A good evaluation design needs to correspond with information needs. It will be easier to communicate and use results within the Agency if the evaluation questions are of relevance to decision makers.

Management Action Plan

An effective management response to the evaluation report is also a critical step in ensuring that the evaluation is used. The action plan should adequately address the findings and recommendations of the report. The plan should identify the required action, timelines and who is responsible for carrying out the action. The head of evaluation may play a role in monitoring the implementation of an action plan.

Checklists  Checklist for Strategies to Ensure that Evaluations are Used

Considerations

P

1.  Recommendations should be clear, explicit, and feasible.

 

2.  Consider timing of report. The evaluation report should be completed prior to important decisions or planning activities.

 

3.  Link findings to planning activities. Relevant evaluations can be reviewed before new projects or activities are planned.

 

4.  The management response or action plan adequately addresses findings and recommendations.

 

5.  The action plan describes what will be done, when, and who will do it.

 

6.  Consider tabling action plans with or soon after evaluation reports.

 

7.  Consider posting evaluation reports on agency Web site.

 

8.  Include findings in strategic reports (e.g., annual reports can be a useful way of sharing results since they present important information in relation to specific themes).

 

9.  Findings and recommendations may be summarized according to useful categories or themes.

 

10.  Arrange meetings or seminars with the various stakeholders either during the evaluation work or after to share evaluation results.

 

11.  Prepare thematic reports on the basis of evaluations and reviews.

 

12.  Reports of this type typically contain high-level information and consequently may appeal to decision makers.

 

13.  Prepare and publish summaries of major evaluations and reviews for wider distribution.

 

14.  Use of databases can also facilitate the exchange of information on various topics.

 

15.  Consider linkages to training. Develop training manuals, tools, checklists of lessons learned or best practices and circulate accordingly.

 

16.  Identify, disseminate, and apply lessons learned from best practices to ongoing training, leadership, and policy development activities.

 

17.  Best practices can be consolidated on a Web site or published in a document.

 

 

Key References

Annie E. Casey Foundation, Baltimore, MD. Getting Smart, Getting Real: Using Research and Evaluation Information to Improve Projects and Policies , September 1995. http://www.aecf.org/publications/data/getsmartgetreal.pdf

National Institute of Justice, Washington, D.C. Using Evaluation Findings for Decision Making , 1989. http://www.bja.evaluationwebsite.org/guide/documents/chapter_5_nij_guide.htm

National Institute of Law Enforcement and Criminal Justice, Washington, D.C. Achieving Utilization of Evaluation Findings , 1975. http://www.bja.evaluationwebsite.org

Treasury Board of Canada Secretariat. Evaluation Policy and Standards , 2001.

UN-HABITAT. Monitoring and Evaluation Guide , 2003.

 



 

Appendix A―Types of Small Agencies

Regulatory: Agencies that grant approvals or licences based on criteria set out in legislation or regulation.

Judicial: Courts presided over by federally appointed judges. (Please note that the Federal Court and Tax Court recently amalgamated to form the Courts Administrative Service, which leaves only the Supreme Court.)

Quasi-judicial Tribunal: Agencies that hear evidence under oath and render decisions based on that evidence alone in conjunction with the applicable statutes and precedents but independent of government policy.

Investigative Agency: Agencies that investigate a complaint or inquiry and report or make recommendations on their findings.

Parliamentary Agency (Agents of Parliament): Agencies that report directly to Parliament (i.e., Information Commissioner and Privacy Commissioner, Commissioner of Official Languages, Auditor General and Chief Electoral Officer).

Policy Development and Advisory: Agencies that develop policy and make recommendations to the government on issues such as health, the economy, or environment.

Other: Small agencies that do not fall under any of the above categories.


Appendix B―Horizontal Initiatives

Horizontal initiatives are efforts involving the co-ordinated activities of several federal departments and/or agencies focussed on specific objectives of national interest. They must have an RMAF and associated evaluation plans and strategies to which all partners are expected to contribute information and possibly resources.

These types of initiatives may provide opportunities for small agencies to share and/or develop internal evaluation capacity (e.g., through case studies or peer review steps in an evaluation design, through existing internal information, and benefiting from evaluation expertise or consulting expertise provided or paid for by other partners). Horizontal initiatives also provide opportunities for an appropriate grouping of agencies to develop common indicators and share experience in performance measurement and evaluation.

For some general guidance on developing Results-based Management and Accountability Frameworks, please refer to the TBS Web site at http://www.tbs-sct.gc.ca/cee/pubs/guide/sarmaf-ascgrr-fra.as .

Types of horizontal initiatives include:

Some challenges for evaluating horizontal initiatives include:

 

Appendix C―Seeking External Advice and Support

It is important to have adequate advice and support when attempting to build evaluation capacity. Consider the following possible sources of support.

TBS/CEE Support for Small Agencies

According to the TB Evaluation Policy , the Treasury Board of Canada Secretariat must provide central direction for evaluation; use evaluation results where appropriate in decision-making at the centre; and set standards and monitor capacity in the government.

At TBS, the Centre for Excellence in Evaluation (CEE) was established to

Other particularly relevant TBS areas include those involved in the change functions associated with Modern Comptrollership, Internal Audit, Results-based Management, Horizontal Reporting, and Expenditure Reporting. For more information, see the following Web site http://www.tbs-sct.gc.ca/cee/about-apropos-eng.asp .

Canadian Evaluation Society (CES)

In the field of evaluation, the CES promotes leadership, knowledge, advocacy, and professional development. The CES provides access to a community of evaluators, annual conferences, the Essential Skills Series of courses in evaluation, and reserved resources on the CES Web site ( http://www.evaluationcanada.ca/ ). The CES has various provincial chapters as well as a National Capital Chapter.

Small Agency Administrator's Network (SAAN)

SAAN's mission is to provide opportunities for small agencies to share information and practices as well as to discuss issues of common concern and to provide a common voice to Central and Common Service Agencies with respect to small agency issues. For more information, see the following Web site: http://www.cso-cpo.gc.ca/saan-rapo/charter_e.html .


Appendix D ―Expenditure Review Committee's 
7 Tests

Program spending will be assessed against the following specific tests:

1.  Public Interest Test – Does the program area or activity continue to serve the public interest?

2.  Role of Government Test – Is there a legitimate and necessary role for government in this program area or activity?

3.  Federalism Test – Is the current role of the federal government appropriate, or is the program a candidate for realignment with the provinces?

4.   Partnership Test – What activities or programs should or could be transferred in whole or in part to the private or voluntary sector?

5.  Value for Money Test – Are Canadians getting value for their tax dollars?

6.  Efficiency Test – If the program or activity continues, how could its efficiency be improved?

7.  Affordability Test – Is the resultant package of programs and activities affordable? If not, which programs or activities could be abandoned?


Appendix E―"How to" Information for Planning and Conducting Evaluations

1.0  How do you build a logic model?

The following graphic presents an overview of the three steps for logic model development.

 


 

Step 1a:  Preparing for Logic Model Development
Determining Internal Capacity

 

Is your agency ready to embark on building a logic model?

Ask yourself the following questions:

  • Is there sufficient time and commitment to develop the logic model internally?
  • Is there familiarity with respect to logic model development?
  • Are there sufficient planning and communication skills, which are key to building consensus and obtaining commitment?
  • Is there sufficient objectivity, neutrality?
  • Does the program involve only my agency?

If you answered "yes" to these questions, you are probably ready to build a logic model.

If you answered "no" to any of the first four questions, you may wish to contract out the development of the logic model.

If you answered "no" to the last question, then the initiative is considered a "horizontal initiative." There are typically more challenges to developing a logic model for a horizontal initiative since you have to involve many stakeholders with different perspectives and opinions.

For further information on RMAFs, see Preparing and Using Results-based Management Accountability Frameworks , April 2004.

 

Step 1b:  Preparing for Logic Model Development
Collecting Relevant Information

 

What are the key sources of information for developing a logic model?

Review the following documents:

  • relevant legislation, regulations, and policy
  • performance reports, business plans and other strategic documents
  • monitoring, audit, and evaluation reports
  • narrative descriptions or overview documents
  • documents or information from similar projects

Consult the following people:

  • senior management
  • board members
  • program or policy staff
  • stakeholders

Key Questions to Ask

  • What is the rationale for the program?
  • What key results do you expect from this program?  
  • How should this program be undertaken in order to achieve these results?
  • Who are the clients? Who are the other stakeholders?
  • What activities need to be in place to achieve those results? (A relevant question when developing a logic model for a planned initiative.)

 

Note that the extent to which you consult with all of the above groups depends on your information needs and resource constraints. However, you should note that perspectives from a variety of stakeholders provide you with a better understanding of the program.

Step 2:  Building the Logic Model

There are different strategies for building a logic model. Two options include

The key advantage of the first option is that sometimes the session proceeds more efficiently if the core elements of the logic model are developed beforehand. If there is limited time to conduct a working session, you may want to consider drafting the logic model prior to the working session.

There are also advantages of the second option in that stakeholders develop the logic model. It helps to build internal capacity with respect to logic model development and group processes. It can also lead to an enhanced understanding of the initiative. Finally, this approach may also strengthen commitment of stakeholders to the process.

When making the decision, consider stakeholder and facilitator preferences and timeframe.

Where do you start? Results or Activities?

While there is no right way to build a logic model, some experts suggest that you start with the selection of key activities if it is an existing program or policy. Start identifying the key results if it is a planned initiative. Remember that you can start wherever you prefer.

With an existing program you might start by asking: What is it we do? Then you can ask why? For example, why are we aiming for enhanced skills? The next result statement should provide the answer to this question (i.e., so that staff will work more efficiently).

When planning a program, you can start developing the logic model from the results. Once you have identified an appropriate result you can ask "How do we achieve this ultimate result?" The previous result statement (i.e., intermediate result) should provide the answer to this question.



Guidelines for constructing logic models

  • No logic model is ever perfect! It should be a reasonably accurate picture of the program.
  • Keep the logic model focussed.
  • Get feedback from a variety of key stakeholders, including program and/or policy staff.
  • The logic model components and linkages have to make sense. Can you spot any leaps in logic?
  • Link final results to the agency's strategic outcomes as specified in its Program Activity Architecture.
  • Ensure that the logic model demonstrates the "if...then", cause-effect relationship, from activities to outputs through to results.
  • Begin activity statements with an action verb.
  • Keep the number of activities to a minimum. Some activities may be merged with another activity.
  • Do not include administrative activities that are not directly involved in delivering your mandate (e.g., HR, IT, Finance, Corporate Services).
  • "If you control it, then it's an activity or an output, if you can only influence it, then it's an outcome."
  • Question activities with no outputs or results.
  • Results are modified (e.g., increase, decrease, improve, maintain).
  • Some programs may have more than one result track.
  • Does it build on, or is it situated in relation to, the business plan or strategic objectives of the department or agency?
  • Results have a "who," a "what," and a "when" (e.g., What change? In whom? By when?)
  • Results demonstrate that you are making a difference.
  • Immediate, intermediate, and ultimate results are presented as a sequence of results, but are not necessarily tied to particular timeframes.
  • You can add the connections after the component boxes are completed. (This applies to flow chart model only.)
  • You can use sticky paper to note activities, outputs, and results. This gives you flexibility to move the components around.
  • Remember that, as you move from immediate to final results, there are decreased levels of control with shared accountability and increased difficulty in evaluating attribution (i.e., the degree to which the program produced the results).
 

Step 3:  Validation of Logic Model

Consult with working groups and stakeholders. It is often helpful to solicit the feedback of individuals who are familiar with the program but who were not part of the working session to verify that all necessary elements are represented in the model.

Build awareness of the logic model. The working group or individual can create awareness on an informal basis by referring to the logic model in conversations with staff and stakeholders. Management and program teams can use the model as a consistent reference for all aspects of the management cycle: planning, monitoring, and reporting. Increased awareness will lead to feedback and insights on how to improve the model.

The model will never be prefect. Use feedback from consultations and from using it as a management reference to update it. Remember that as the context of the program changes over time, so will the underlying logic.

Key References

Treasury Board of Canada Secretariat. RBM E-Learning Tool.



2.0  How do you develop evaluation questions?

 



Guidelines for Developing Evaluation Questions

Start with the broad issues

  • Consider standard evaluation questions associated with each of the issues in accordance with the TB Evaluation Policy :
    • Relevance – Does the program continue to be consistent with agency and government-wide priorities and does it realistically address an actual need?
    • Success – Is the program effective in meeting its objectives, within budget, and without unwanted results?
    • Cost-effectiveness – Are the most appropriate and efficient means being used to achieve objectives relative to alternative design and delivery approaches?
  • Consider Expenditure Review Questions – See Appendix D.

Tailor the questions to your program

  • Use your logic model as a guide. Review your outputs as an aid to developing questions relating to efficiency and service delivery. Review the results as an aid to developing questions relevant to effectiveness.
  • Consult with key stakeholders to clarify key evaluation interests.
  • Consider the audience for the report and what action might be taken based on the report.

Prioritize

  • Consider accountability and information requirements.
  • Consider previous evaluation, audit, and monitoring reports.
  • Consider risks.
  • Consider costs and benefits associated with addressing each issue.
  • Separate "nice to know" from "need to know."


3.0  How do you identify the right performance indicators?

Step 1:  Review Logic Model

Go through each row of the logic model (except activities) and determine what specific piece of information or particular data would be required to assess whether each output has been produced or result achieved. A working session is an effective method for brainstorming indicators.

Step 2:  Prioritize

Identify the "need to have" versus the "nice to have" for each component.

 

Once a comprehensive set of performance indicators and associated measurement strategies have been identified, a smaller set of the best indicators needs to be identified. Check the top ranked indicators against the selection criteria described.

 

Step 3:  Check Against Criteria

What are criteria for selecting indicators?

  • Relevant : Is the indicator meaningful? Is it directly linked to the output or result in question?
  • Reliable : Is it a consistent measure over time?
  • Valid : Does it measure the result?
  • Practical : Will it be easy to collect and analyze? Is it affordable?
  • Comparable : Is it similar to what other organizations or areas in your organization already measure?
  • Useful : Is it useful? Will it be useful for decision making?

Useful tips

  • Begin by developing a few indicators. (Over time, additional indicators can be added, if necessary.)
  • Keep the number of indicators to a minimum.
  • Few indicators are good – but be aware of their limitations.
  • Try to keep a core set of indicators which can be maintained over time to allow for comparison between past and present performance.
  • Consider proxy indicators. Proxy indicators are sometimes used to provide information on results where direct information is not available. For example, the percentage of cases that are upheld at appeal could be a proxy indicator for the quality of decisions.

 

4.0  How do you choose an appropriate evaluation design?

Evaluation design is the process of thinking about what you want to do and how you want to go about doing it.


Guidelines for Choosing Appropriate Evaluation Designs

  • Consider the following:
  • What are the information and decision-making needs of the agency with respect to the evaluation?
  • What type of evaluation would be most appropriate given the life cycle of the program?
  • What considerations should be made with respect to practicality and costs?
  • What would be an appropriate balance between information needs and costs?
  • What level of concern exists with respect to the program to be evaluated (i.e., related to the quality of evidence to be gathered)?
  • What are other internal and external factors that may influence the program? How can the evaluation design minimize these factors?
  • How can the evaluation be designed to target evaluation questions to the most pressing concerns?
  • What sources of information exist for the evaluation? Consider existing data, secondary data, and performance measurement information as potential sources of information for the evaluation.
  • Are there multiple lines of evidence? (More than one line of evidence improves reliability of findings.)
  • To what extent will a rigorous design be required to accept findings and conclusions and implement recommendations?

 

Considerations of threats to validity in choosing evaluation design

When developing an evaluation design, you have to consider whether other factors are affecting the results of the program. It is particularly important when you are trying to determine the impacts or effectiveness that these factors or threats to validity are considered. These factors can be due to real changes in the environment or changes in participants involved in the program.


Overview of Evaluation Designs

 

Type 1: Implicit or Non-experimental Designs

In this type of design, "changes" to the program participants are measured. There is no comparison group of non-participants in the design. Using this design type, it is difficult to determine the extent to which the results can be attributed to the program. However, this design is useful for obtaining information relating to service delivery, extent of reach of the intervention, and progress towards objectives.

The post-test-only design and the pre-test/post-test design are two common types of non-experimental design.

Single group post-test-only design

In this design, beneficiaries or clients of an intervention are measured after the intervention. Participants, for example, can be simply asked about the impact of the intervention.

Single group pre-test/post-test design

This design uses before-and-after measures on a single group. For example, when measuring the impact of a training program, a knowledge test may be administered before and after the training program to help assess the impact of the training.

This design can be used

  • to answer certain types of information requests (e.g., questions about management issues relating to how the program is being implemented or whether risk is being managed, strategies for improvement);
  • when no pre-program measures exist;
  • where there is no obvious control or comparison group available; and
  • where practicality and costs are important considerations.

This type of design can be enhanced by

  • using varied quantitative and qualitative data collection methods and sources of information; and
  • ensuring the collection of "high-quality" data.

 

Type 2: Quasi-experimental Designs

The key distinction that separates experimental designs from non- or quasi-experimental designs is the random assignment of subjects into the intervention (treatment) groups and non-intervention (control) groups. Quasi-experimental designs involve comparison groups that are not randomly selected nor randomly assigned to the intervention. Efforts are usually made to match the comparison and the "treatment" groups as closely as possible according to a predetermined set of characteristics.

Quasi-experiments require analysis techniques that are much more complicated than those for true experiments. High-level statistics (e.g., econometric models) are required to deal with the differences between groups and isolate the effect of the program.

 

Type 3: Experimental Designs

Random assignment of subjects to the intervention (i.e., treatment) and control groups helps ensure that subjects in the groups will be equal before the intervention is introduced. Although experimental designs are considered ideal for measuring impact, they are rarely practical.

 

Both quasi-experimental and experimental designs involve some type of pre-test followed by a post-test. Both design types are appropriate for conducting summative evaluations. However, practicality and costs must also be considered.

Key References

Treasury Board of Canada Secretariat. Program Evaluation Methods: Measurement and Attribution of Program Results , 1998.
http://www.tbs-sct.gc.ca/cee/pubs/meth/pem-mep00-eng.asp


5.0  How do you choose appropriate data collection methods?

To choose an appropriate data collection method, you may consider the following:

For more detailed information, see the two tables below that compare the quantitative and qualitative methods, as well as describe the specific data collection methods available for evaluations.


A Comparison of Quantitative and Qualitative Methods

 

Quantitative Methods

Qualitative Methods

Use

to numerically measure "who, what, when, where, how much, how many, how often"

when you need to generalize findings

to qualitatively analyze "how and why"

to clarify issues and discover new issues

when you need a better understanding of context

Data Collection Methods

standardized interviews; surveys using closed-ended questions; observation using coded guides

administrative data

open and semi-structured interviews; surveys using open-ended questions; observation; interpretation of documents, case studies, and focus groups

Strengths

provides quantitative, accurate, and precise "hard data" to prove that certain problems exist

can test statistical relationships between a problem and apparent causes

can provide a broad view of a whole population

enables comparisons

establishes baseline information which can be used for evaluating impact

useful when planning an initiative concerned with social change

particularly in formative evaluations, investigators may need to know participant attitudes about a program, their ideas about how it could be improved, or their explanations about why they performed in a particular way

provides a thorough understanding of context to aid in interpretation of quantitative data

provides insights into attitudes and behaviours of a small sample population

establishes baseline information which can be used for evaluating qualitative outcomes

useful for getting feedback from stakeholders

Weaknesses

may be precise but may not measure what is intended

cannot explain the underlying causes of situations (i.e., it may tell you that the program had no effect, but will not be able to tell you why)

information may not be representative

more susceptible to biases of interviewers, observers, and informants

time-consuming to collect and analyze data

Source:  Adapted from the Program Manager's Monitoring and Evaluation Toolkit Number 5, Part III: Planning and Managing the Evaluation – the Data Collection Process . May 2001 ( www.unfpa.org United Nations Population Fund, Office of Oversight and Evaluation).

 

Overview of Data Collection Methods

Data Collection Method

When to Use

Strengths

Challenges

External Administrative Systems and Records: use of data collected by other institutions or agencies (e.g., Statistics Canada)

need information about context

need historical information

to compare program/ initiative data to comparable data

It is efficient and avoids duplication.

Is the information accurate, applicable, and available?

Are we comparing apples to apples?

Internal Administrative Data: program data collected internally for management purposes

need information about management, service delivery

It is efficient and can provide information about management activities and outputs.

It can be designed to collect performance information related to the program.

Is the information accurate and complete?

Literature Review: review of past research and evaluation on a particular topic

to identify additional evaluation questions/ issues, and methodologies

need information on conceptual and empirical background information

need information on a specific issue

need information about comparable programs, best practices

make the best use of previous related work

best practices

may suggest evaluation issues or methodologies for current study

can be secondary source of data helping to avoid duplication

Data and information gathered from a literature search may not be relevant to evaluation issues.

It can be difficult to determine the accuracy of secondary data in the early stages of a study.

Interview: a discussion covering a list of topics or specific questions, undertaken to gather information or views from an expert, stakeholder, and/or client; can be conducted face to face or by phone

complex subject matter

busy high-status respondents

sensitive subject matter (in-person interviews)

flexible, in-depth approach

smaller populations

inexpensive method for collecting contextual and systematic information about a program or service

flexible method (can occur either in person or remotely and can be either open-ended or structured)

danger of interviewer bias

The response rate to requests for phone and/or electronic interviews are often much lower than for in-person interviews.

Travel costs for in-person interviews can be high.

Focus groups: a group of people brought together to discuss a certain issue guided by a facilitator who notes the interaction and results of the discussion

depth of understanding required

weighted opinions

testing ideas, products, or services

where there are a limited number of issues to cover

where interaction of participants may stimulate richer responses (people consider their own views in the context of others')

Group processes can be helpful in revealing interactions and relationships within an organization.

The discussion may uncover insights on the rationale behind common perceptions and reactions, as well as demonstrate how differences in opinion are resolved.

Focus groups are short-lived, artificial situations.

Group situations may not put participants at ease to discuss personal beliefs and attitudes especially if the people have to relate to each other after leaving the focus group.

The data generated in a focus group tends to be quick response instead of considered answers.

Case studies: a way of collecting and organizing information on people, institutions, events and beliefs pertaining to an individual situation

when detailed information about a program is required

to explore the consequences of a program

to add sensitivity to the context in which the program actions are taken

to identify relevant intervening variables

permits a more holistic analysis and consideration of the inter-relationships among the elements of a particular situation

permits an in-depth analysis of a situation

provides depth of information

complex method of data organization

difficult to make conclusions that can be applied to other situations

Questionnaire/Survey: (paper, on-line, or telephone) a list of questions designed to collect information from respondents on their knowledge and perceptions of a program or service

useful for large target audiences

can provide both qualitative and quantitative information

tend to be less time and money intensive than interviewing large numbers of people

questions can cover a range of topics (on-line or mail-out)

respondents can take time to consider their answer and look up information

provides a breadth of information

may allow you to make statistically valid inferences about the entire population

low response rates

possibility that those who returned their questionnaire are not typical of the general population being surveyed

requires considerable expertise in their design, conduct, and interpretation

Expert panels: the considered opinion of a panel of knowledgeable outsiders

experts can share lessons learned and best practices

where outside validation is required

where diversity of opinion is sought on complex issues

where there is a need to draw on specialized knowledge and expertise

An expert panel can draw on the knowledge and experience of the panel members to provide opinions and recommendations on a program or approach.

efficient especially if done electronically or by phone

Unless the experts know a great deal about the program and context within which it operates, their opinion may offer very little useful insight.

Experts tend to hold a particular worldview or opinion that may affect their perception of a program or approach.

Comparative studies: a range of studies that collects comparative data (e.g., cohort studies, case-control studies, experimental studies)

summative evaluations

a powerful way of collecting data for comparative purposes

finding reasonable comparative groups

structuring valid studies

analyzing data is time and money intensive

Source :  Adapted from TBS, RBM E-Learning Tool

 Key References

Treasury Board of Canada Secretariat. RBM E-Learning Tool .

 

6.0  How do you design a survey questionnaire?

 

When you may need information about a large group or population, a survey is typically conducted. A sampling strategy should be such that the information obtained from a sample is information that is representative of the entire population. The more representative a sample is, the more confidence there is that you can attach to your findings. Representativeness is generally related to the sample size and lack of bias.

Step 1:  Sampling Procedures

Clearly define purpose of survey

In order to develop sampling procedures, you need to clearly define the purpose of the survey. The sampling strategy must be designed to answer the evaluation questions.

Sample Size

Determining the sample size will help determine which type of survey to use.

Note: A sample is part of entire population that possesses the characteristics you wish to study.

 

The following are considerations for determining sample size:

If you wish to draw conclusions about the entire population, you will require a certain sample size that is based on statistical parameters. For example, when polling firms say such a sample is accurate with a 5 per cent margin of error, 90 per cent of the time, this is based on a certain size of sample. The required sample size is related to the size of your population, the confidence level required, and the allowable margin of error. For example, if you have a population of 1,000 people from which you need a representative sample and you require a 95 per cent confidence interval with a 5 per cent margin of error, then you will require a maximum sample of 278. In social science research, a 95 per cent confidence level and a 5 per cent allowable margin of error is typically specified. You may need to consult with a statistical expert regarding appropriate sample size.

Sampling Techniques

There are a number of different methods that can be used to select a sample. Ideally you want the sample to represent the whole group so that you can generalize the findings to the program. Types of sampling include simple random sampling, stratified random sampling, systematic sampling, and cluster sampling.

With simple random sampling a list of all people (i.e., the survey population) is made and then individuals are selected randomly for inclusion in the sample. Random sampling means that everyone in the target group has an equal chance of being included in the study. One challenge with this approach is obtaining a complete list of the group.

The steps for sampling procedures are as follows:

1.  obtain a sampling frame;

2.  check for bias;

3.  assess the potential sampling source in advance; and

4.  apply sampling procedure.

Sampling Frames

Sampling frames are listings of people that represent or approximate the population of interest. Sampling frames should be comprehensive and representative. The list should be unbiased.

Check for Biases

Some common biases include:

Assess the potential sampling sources in advance

What if contact information doesn't exist? Options include

What if information about potential respondents is limited?

Examples of Random Sampling Technique

Step 2:  Determining the Survey Format

Step 3:  Develop Survey Questionnaire

A key consideration in developing your questionnaire is determining what types of questions to use. While open-ended questions can provide detailed information, they are time consuming to record and analyze. If you have a large group to survey, the questionnaire should be largely comprised of closed-ended questions.

You should also prepare a script and instructions for the interviewers (if by telephone or in-person). The script should include how to greet the respondent, how to invite them to participate, how to respond to their answers, how to keep the respondent on the line, how to thank the respondent, how to code each survey (completed, no answer).

Types of Survey Questions

Open-ended questions provide no structured answers. These types of questions are time-consuming to record and analyze. They should be kept to a minimum in survey research where there are a large number of respondents. Skilled interviewers are required to adequately probe and record these questions. They allow you to probe more deeply into issues of interest being raised. Open-ended questions are useful for exploring issues and providing more detailed information as to why and how.

Closed-ended – Scaled-response. These list alternative responses that increase or decrease in intensity along a continuum (e.g., very dissatisfied/ dissatisfied/neither satisfied nor dissatisfied/satisfied/very satisfied; strongly agree/agree). A 5-point scale is common and allows you to keep a neutral position (neither agree nor disagree). Try to include all possible answers among answer categories (e.g., don't know and not applicable).

Close-ended – Fixed Response questions. These involve choosing one or more options from a list. A category of "other" should be included so that the respondent is not forced to select an inappropriate answer. Ensure that categories are mutually exclusive (e.g., 0-9, 10-19). Avoid long lists of categories.

 

Checklist  Checklist for Developing a Survey Questionnaire

Considerations

P

1.  Relevance to research questions

 

2.  Consider your population (e.g., age, gender)

 

3.  Sample size will help to determine which type of survey to use

 

4.  Budget (Mail and Web surveys are cost-effective.)

 

5.  Subject matter (e.g., sensitive subject matter, requires thought)

 

6.  Appropriate length (Length influences response rate.) As a general rule, phone interviews should last between 10-20 minutes; 10 minutes is the ideal.

 

7.  Consider type of questions (open-ended, closed-ended – scale, fixed response)

 

8.  Consider what kind of scale. Choose from 3-point, 5-point, 7-point, 10-point and 100-point scales. Level of satisfaction?

 

9.  Keep questions short, simple, and clear.

 

10.  Keep questions as specific as possible.

 

11.  Avoid the use of double negatives.

 

12.  Avoid double-barrelled questions. These are single questions that ask for responses about two or more different things. For example: To what extent are you satisfied with the telephone and in-person service?

 

13.  Establish a relevant time frame for questions. When asking about past events it is important to establish an appropriate time frame. Respondents often can only recall general information. Example: Over the last seven days, how often have you exercised? In the past six months, how often have you gone to your doctor?

 

14.  Consider whether respondents have the knowledge, opinions, or experience necessary to answer the question.

 

15.  Make every effort to be consistent (e.g., one scale, one wording choice)

 

16.  Use social conversation as a guide to organizing the questionnaire (i.e., introduction, building up to main topic, main topic, closing).

 

17.  Develop script and instructions.

 

 

Step 4:  Pre-testing the Questionnaire

Surveys can be pre-tested and then adjusted. The purpose of pilot testing is to identify and resolve problems and deficiencies in the information collection methods or in the form and usefulness of the information gathered. Based on the results of the pre-test you can modify the survey accordingly. Conducting a small number of about 10 interviews can provide higher quality survey results.

You should determine what questions your pre-test will answer. For example, your pre-test might want to determine the following:

Step 5:  Implement the Survey

During implementation

If you expect your results to be comparable with existing results for benchmarking purposes, the research design, questionnaire, and administration need to be very similar, if not identical. It is okay to make some improvements over the previous iteration, but the bigger the change, the less comparable the results.

You should ensure there are procedures in place for following up non-response and compensating for lower response rates. Changes in the original composition of a sample are usually inevitable during the course of an evaluation study. Individuals may drop out from the sample and others may provide incomplete information. These changes may bias the study if they are not addressed.

Key References

Goss Gilroy Inc., Designing Effective Client Satisfaction Surveys , Strategic Management Conference, Montreal, 2003.

SPSS BI Survey Tips, A Handy Guide to Help You Save Time and Money as You Plan, Develop and Execute Your Surveys . http://www.spss.com/uk/SurveyTips booklet.pdf


7.0  How do you analyze data?

Step 1: Start with the evaluation objectives

When analyzing any type of data, review the purpose of the evaluation. This will help you organize your data and focus your analysis.

Step 2: Review for accuracy, completeness and consistency

Step 3: Summarize and organize data

Generalizing the Findings

The only valid way of generalizing findings to an entire or target population (where you cannot survey or study everyone) is to use findings from a random sample of the population you wish to study. Caution must therefore be exercised when analyzing data from non-randomized samples.

Qualitative and Quantitative Analysis

Analyzing qualitative data requires effective synthesis and interpretative skills. Qualitative information can be used to provide contextual information, explain how a program works, or to identify barriers to implementation. Qualitative data can be analyzed for patterns and themes that may be relevant to the evaluation questions. Qualitative material can be organized using categories and/or tables making it easier to find patterns, discrepancies and themes.

Quantitative data analysis provides numerical values to information. It can range from simple descriptive statistics (e.g., frequency, range, percentile, mean or average) to more complicated statistical analysis (e.g., t-test, analysis of variance). Computer software packages such as Statistical Package for Social Sciences (SPSS), Minitab, and Mystat can be used for more complicated analysis. Quantitative data analysis also requires interpretation skills. Quantitative findings should be considered within the context of the program.

About quantitative data

Frequencies, range, percentile, and standard deviation are used for descriptive statistics. Measures of central tendencies – mean, median, or mode – are also calculated.

Examples of inferential statistics:

Key References

Treasury Board of Canada Secretariat. RBM E-Learning Tool.


Appendix F―Terms of Reference Template

 

Main Elements of the Terms of Reference

Project Background

Project context and rationale

Identification of key stakeholders, clients and partners

Project description

Reasons for the Evaluation

Statement of purpose of the study

Expected value-added

Intended use of results

Scope and Focus

Broad issues to be addressed/specific evaluation questions

Type of analysis to be used/level of detail

Specify who the audience(s) will be for the reports and findings

Statement of Work

How purposes of study are to be achieved

Describe approaches

Describe data collection methods

Outline the tasks required to undertake study

State what groups will be consulted

List expectations with respect to communications and ongoing progress reports

Evaluation Team

Required professional qualifications/expertise/experience

Role responsibilities of evaluation team, role of agency (program and/or evaluation managers)

Timetable

Approximate timetable to guide the preparation of the work plan

Budget

A specification of the estimated resources to be committed to the study and its different parts

Deliverables

Identification of key deliverables (e.g., work plan or methodology report, draft evaluation report, final evaluation report)

 

Toolkit: Templates
Agency Audit and Evaluation Plan Template
[9]

Introduction and Context

Methodology/Approach

Rationale

Evaluation and Audit Plan Summary


Detailed Evaluation and Audit Plan

Appendices

 

Appendix G―Glossary

Accountability ( Responsabilisation) – The obligation to demonstrate and take responsibility for performance in light of agreed expectations. There is a difference between responsibility and accountability: responsibility includes the obligation to act whereas accountability includes the obligation to answer for an action

Activity ( Activit̩ ) РAn operation or work process internal to an organization, intended to produce specific outputs (e.g., products or services). Activities are the primary link in the chain through which outcomes are achieved.

Attribution ( Attribution ) – The assertion that certain events or conditions were, to some extent, caused or influenced by other events or conditions. This means a reasonable connection can be made between a specific outcome and the actions and outputs of a government policy, program, or initiative.

Departmental Performance Reports (DPR) ( Rapport minist̩riel sur le rendement (RMR) ) РDepartmental Performance Reports, tabled in the fall of each year by the President of the Treasury Board on behalf of all federal departments and agencies named in Schedule I, I.1 and II of the Financial Administration Act , are part of the Estimates and Supply process. The reports explain what the government has accomplished with the resources and authorities provided by Parliament. The performance information in the reports is intended to help members of Parliament advise the government on resource allocation in advance of the annual budget and Supply process in the spring.

Effectiveness ( Efficacit̩ ) РThe extent to which an organization, policy, program, or initiative is meeting its planned results. (A related term is Cost Effectiveness РThe extent to which an organization, policy, program, or initiative is producing its planned outcomes in relation to expenditure of resources.)

Efficiency ( Efficience ) – The extent to which an organization, policy, program, or initiative is producing its planned outputs in relation to expenditure of resources.

Evaluation ( Évaluation ) – The systematic collection and analysis of information on the performance of a policy, program, or initiative to make judgements about relevance, progress, or success and cost-effectiveness and/or to inform future programming decisions about design and implementation.

Expenditure Management Information System (EMIS) ( Syst̬me d'information sur la gestion des d̩penses (SIGD) ) РThis is a common information framework that supports the Expenditure Review Committee and departmental assessments related to the Management Accountability Framework. Program Activity Architecture and EMIS are to become the basis for the following:

Final Outcome ( R̩sultat final ) РThese are generally outcomes that take a longer period to be realized, are subject to influences beyond the policy, program, or initiative, and can also be at a more strategic level.

Goal ( But ) – A general statement of desired outcome to be achieved over a specified period of time. The term goal is roughly equivalent to Strategic Outcome. For technical precision, the Treasury Board of Canada Secretariat recommends that Strategic Outcome be used instead of goal. See also Objective.

Horizontal Result ( R̩sultat horizontal ) РAn outcome that is produced through the contributions of two or more departments or agencies, jurisdictions, or non-governmental organizations.

Impact ( Impact ) – Impact is a synonym for outcome, although an impact is somewhat more direct than an effect. Both terms are commonly used, but neither is a technical term. The Treasury Board of Canada Secretariat recommends that result be used instead of impact.

Indicator ( Indicateur ) – A statistic or parameter that provides information on trends in the condition of a phenomenon and has significance extending beyond that associated with the properties of the statistic itself.

Input ( Intrant ) – Resources (e.g., human, material, financial) used to carry out activities, produce outputs, and/or accomplish results.

Logic Model ( Mod̬le logique ) Р(also referred to as Results-based Logic Model) An illustration of the results chain or how the activities of a policy, program, or initiative are expected to lead to the achievement of the final results. Usually displayed as a flow chart. See also Results Chain.

Management Accountability Framework (MAF) ( Cadre de responsabilisation de la gestion (CRG) ) – It is a set of expectations for modern public service management. Its purpose is to provide a clear list of management expectations within an overall framework for high organizational performance.

Management of Resources and Results Structure (MRRS) ( Structure des ressources et des r̩sultats de gestion (SRRG) ) РThe MRRS replaces the Planning, Reporting, and Accountability Structure (PRAS) policy as the new reporting regime. The MRRS

Mission Statement ( Énoncé de mission ) – A formal, public statement of an organization's purpose. It is used by departmental management to set direction and values.

Objective ( Objectif ) – The high-level, enduring benefit towards which effort is directed.

Outcome ( R̩sultat ) РAn external consequence attributed to an organization, policy, program, or initiative that is considered significant in relation to its commitments. Outcomes may be described as immediate or intermediate; final, direct or indirect; intended or unintended. See also Result.

Output ( Extrant ) – Direct products or services stemming from the activities of a policy, program, or initiative, and delivered to a target group or population.

Performance (Rendement) – How well an organization, policy, program, or initiative is achieving its planned results measured against targets, standards, or criteria. In results-based management, performance is measured, assessed, reported, and used as a basis for management decision-making.

Performance Measurement Strategy ( Strat̩gie de mesure du rendement ) РSelection, development, and ongoing use of performance measures to guide corporate decision-making. The range of information in a performance measurement strategy could include reach; outputs and results; performance indicators; data sources; methodology; and costs.

Performance Measures ( Mesures du rendement ) – An indicator that provides information (either qualitative or quantitative) on the extent to which a policy, program, or initiative is achieving its results.

Performance Monitoring ( Suivi du rendement ) – The ongoing process of collecting information in order to assess progress in meeting Strategic Outcomes, and, if necessary, provide warning if progress is not meeting expectations.

Performance Reporting ( Rapport sur le rendement ) – The process of communicating evidence-based performance information. Performance reporting supports decision-making, serves to meet accountability requirements, and provides a basis for citizen engagement and a performance dialogue with parliamentarians.

Planned Results (Targets) ( R̩sultats pr̩vus (Cibles) ) РClear and concrete statement of results (including outputs and results) to be achieved within the time frame of parliamentary and departmental planning and reporting (1 to 3 years), against which actual results can be compared.

Reach ( Port̩e ) РThe individuals and organizations targeted and directly affected by a policy, program, or initiative.

Reliability ( Fiabilit̩ ) РRefers to the consistency or dependability of the data. The idea is simple: if the same test, questionnaire, or evaluation procedure is used a second time, or by a different research team, would it obtain the same results? If so, the test is reliable. In any evaluation or research design, the data collected are useful only if the measures used are reliable.

Reports on Plans and Priorities (RPP) ( Rapport sur les plans et les priorit̩s (RPP) ) РAs part of the Main Estimates, the RPPs provide information on departmental plans and expected performance over a three-year period. These reports are tabled in Parliament each spring, after resource allocation deliberations. They generally include information such as mission or mandate, strategies, as well as Strategic Outcomes and performance targets.

Result ( R̩sultat ) РThe consequence attributed to the activities of an organization, policy, program, or initiative. Results is a general term that often includes both outputs produced and outcomes achieved by a given organization, policy, program, or initiative. In the government's agenda for results-based management and in the document Results for Canadians: A Management Framework for the Government of Canada , the term result refers exclusively to outcomes.


Results Chain (also results-based logic model, results sequence) ( Enchąnement des r̩sultats (mod̬le logique ax̩ sur les r̩sultats, s̩quence de r̩sultats) ) РThe causal or logical relationship between activities and outputs and the outcomes of a given policy, program, or initiative, that they are intended to produce. Usually displayed as a flow chart.

Results for Canadians: A Management Framework for the Government of Canada ( Des résultats pour les Canadiens et les Canadiennes : un cadre de gestion pour le gouvernement du Canada ) – A document published in early 2000 that describes the management framework for the Government of Canada. This key document outlines the four management commitments for the federal government: citizen focus, values, results, and responsible spending.

Results-based Management ( Gestion ax̩e sur les r̩sultats ) РA comprehensive, life-cycle approach to management that integrates business strategy, people, processes, and measurements to improve decision-making and drive change. The approach focuses on getting the right design early in a process, implementing performance measurement, learning and changing, and reporting performance.

Results-based Management and Accountability Framework (RMAF) (Cadre de gestion et de responsabilisation ax̩ sur les r̩sultats (CGRR)) РA document that sets out the performance monitoring, evaluation, and reporting strategies for a policy, program, or initiative.

Service Commitment ( Engagement en mati̬re de service ) РService commitments or standards generally set performance objectives for the delivery of government products or services to the public, specifying the quality or level of service to which a department or agency commits, or can be expected to deliver to clients.

Strategic Outcome ( R̩sultat strat̩gique ) РA Strategic Outcome is a long-term and enduring benefit to Canadians that stems from a department's mandate, vision, and efforts. This Outcome represents the difference a department wants to make for Canadians and should be measurable. The achievement of or progress towards a strategic outcome will require, and Canadians will expect, the sustained leadership of a federal department or agency, especially in developing partnerships and alliances with other stakeholders and organizations.

Canadians also expect that departments will strive for excellence by establishing challenging outcomes that are within their sphere of control or influence. These outcomes will form the standards by which a department's performance is assessed through departmentally derived measures.

Target Group (Target Population) ( Groupe cible (Population cible) ) – The set of individuals that an activity is intended to influence.

Validity ( Validit̩ ) РThe extent to which the questions or procedures actually measure what they claim to measure. In other words, valid data are not only reliable, but are also true and accurate. Measures used to collect data about a variable in an evaluation study must be both reliable and valid if the overall evaluation is to produce useful data.

Source: TBS Guide to RMAFs .
http://www.tbs-sct.gc.ca/cee/tools-outils/rmaf-cgrr/guide00-eng.asp


Appendix H―Evaluation Web Sites

1.  http://www.tbs-sct.gc.ca/pol/doc-eng.aspx?id=15024

This is the Treasury Board of Canada Secretariat (TBS) Evaluation Policy .

2.  http://www.oecd.org/dataoecd/29/21/2754804.pdf

This link is a comprehensive glossary of key terms in evaluation and results-based management.

3.  http://www.wkkf.org/Pubs/Tools/Evaluation/Pub770.pdf

This is an evaluation handbook by Kellogg.

4.  http://www.evaluationcanada.ca

This is the Canadian Evaluation Society homepage. It contains information on courses and special events, various resources on evaluation, and unpublished documents for evaluators.

5.  http://www.phac-aspc.gc.ca/ncfv-cnivf/familyviolence/html/fvprojevaluation_e.html

This document is entitled Guide to Project Evaluation: A Participatory Approach . It was developed by the Population Health Directorate at Health Canada in 1996. The Guide provides an easy-to-use, comprehensive framework for project evaluation. This framework can be used to strengthen evaluation skills and knowledge to assist in the development and implementation of effective project evaluations.

6.  http://www.mapnp.org/library/evaluatn/fnl_eval.htm#anchor1585345

This link contains a Basic Guide to Program Evaluation. This document provides guidance toward planning and implementing an evaluation process for non-profit or for-profit organizations.

7.  http://www11.hrdc-drhc.gc.ca/pls/edd/toolkit.list

This link contains HRDC's Evaluation Tool Kit series developed by Evaluation and Data Development (EDD). This is a series of publications that provides pertinent information about designing, planning, and conducting an evaluation. Publications include the following:

Logic Models

8.  http://www.ed.gov/teachtech/logicmodels.doc

Logic Models: A Tool for Telling Your Program's Performance Story , describes the Logic Model process in detail and how logic models can be used to develop and tell the performance story for a program.

9.  http://national.unitedway.org/outcomes/resources/mpo/

The manual Measuring Program Outcomes: A Practical Approach , is a good source for information on logic models and performance indicator development.

10.  http://www.wkkf.org/Pubs/Tools/Evaluation/Pub3669.pdf

See the W.K. Kellogg Foundation Logic Model Development Guide for a good overview on logic model development. It also provides information on variations and types of logic models.

11.  http://www.impactalliance.org/file_download.php/prevent+1.pdf?URL_ID=2744& filename=10196046740prevent_1.pdf&filetype=application%2Fpdf&filesize=646378&name=prevent+1.pdf&location=user-S/

Prevention Works! A Practitioner's Guide to Achieving Outcomes

12.  http://www.insites.org/documents/logmod.htm

Everything you wanted to know about logic models but were afraid to ask.

13.  http://www.calib.com/home/work_samples/files/logicmdl.pdf

This paper provides a description of logic models and discusses their uses in treatment services planning and evaluation.

14.  http://www.uwex.edu/ces/lmcourse/

Enhancing Program Performance with Logic Models , University of Wisconsin, an online self-study course and excellent resource.

15.  http://www.gse.harvard.edu/hfrp/content/pubs/onlinepubs/rrb/learning.pdf

Learning from logic models. An example of a family/school partnership

RMAF Links

16.  http://www.tbs-sct.gc.ca/cee/tools-outils/rmaf-cgrr/guide00-eng.asp

This is the August 2001 Guide for the Development of Results-based Management and Accountability Frameworks . It contains guidelines for developing the Profile, Logic Model, Performance Measurement Strategy, Evaluation Strategy, and Reporting Strategy.

17.  http://www.tbs-sct.gc.ca/cee/tools-outils/comp-acc00-eng.asp

This is a companion guide for the development of RMAFs for horizontal initiatives. Horizontal initiatives often need to integrate vertical and horizontal accountabilities, various resource pools, as well as a variety of departmental mandates, performance measurement strategies, and reporting structures. This guide is designed to complement the Guide for the Development of Results-based Management and Accountability Frameworks by addressing the unique challenges encountered when diverse organizations work together to achieve common objectives. While it does not provide answers to every question, it does provide guidance based on the most important lessons learned to date.

18.  http://www.tbs-sct.gc.ca/cee/pubs/guide/sarmaf-ascgrr-eng.asp

This document Guidance for Strategic Approach to RMAFs complements the August 2001 Guide for the Development of Results-based Management and Accountability Frameworks . The purpose of this document is to help managers tailor the development of the RMAF to specific circumstances, taking into account such factors as overall risk, program complexity, and reporting requirements so as to ensure that RMAFs remain responsive to evolving needs.

19.  http://www.tbs-sct.gc.ca/res_can/rc_e.html

This document is Results for Canadians: A Management Framework for the Government of Canada . It outlines what public service managers are expected to do to improve the efficiency and effectiveness of their programs. The RMAF is an important management tool in meeting the four main objectives of Results for Canadians : a citizen focus in all government activities; emphasis on values; achievement of results; and responsible use of public funds.

20.  http://www.tbs-sct.gc.ca/pol/doc-eng.aspx?id=12257

This is the Treasury Board of Canada (TB) Policy on Transfer Payments , which formalizes the requirement of the RMAF as part of a TB submission involving transfer payments.


[1] .  See the TBS Web site at http://www.tbs-sct.gc.ca/cee/tools-outils/model-eng.asp .

[2] .  Treasury Board of Canada Secretariat. Models for Evaluation and Performance Measurement for Small Agencies , 2003.

Treasury Board of Canada Secretariat. Interim Evaluation of the Evaluation Policy , 2002.

[3] .  Treasury Board of Canada Secretariat. Models for Evaluation and Performance Measurement for Small Agencies , 2003.

[4] .  Treasury Board of Canada Secretariat. Preparing and Using Results-based Management and Accountability Frameworks. April 2004.

[5] .  Refer to the following Web site: http://www.tbs-sct.gc.ca/cee/ .

[6] .  Treasury Board of Canada Secretariat, Models for Evaluation and Performance Measurement for Small Agencies , 2003.

[7] .  For more information on the organizations mentioned, please refer to Appendix C.

[8] .  Smaller agencies may have only one level of management.

[9] .  Many small agencies submit an Annual Audit and Evaluation Plan. The template incorporates both audit and evaluation. Internal audit plans are required where internal audit priorities have been identified.

 



 

Acknowledgements

An Inter-Agency Steering Committee provided direction and context for the development of the Evaluation Guidebook for Small Agencies. We would like to thank the following participants:

Sylvie Charbonneau , Office of the Commissioner of Official Languages

Pierre Couturier , National Parole Board

Annette Ducharme , Canadian Forces Grievance Board

Yolaine Gauthier , Canadian Forces Grievance Board

Michael Glynn , Canadian Human Rights Tribunal

Janine Sherman , Canada School of Public Service

Robert Sauvé , Patented Medicine Prices Review Board

Greg Smith , Canadian Human Rights Tribunal

 Sharon Watts, Vice-President, Corporate Services & Adjudication, Hazardous Materials Information Review Commission

They reviewed drafts, and provided guidance and feedback to assist the consulting team to understand the context of small agencies.

We are most grateful to Glenn Crone, Kim Cronkwright and Zeljka Spasojevic of the Centre of Excellence for Evaluation, Treasury Board of Canada Secretariat, for their ongoing support.

Members of the Inter-Agency Steering Committee worked in collaboration with Zelda Yule, Sandy Moir, Celine Pinsent and Simon Roy of Goss Gilroy Inc.

 


Date modified: