Archived information

Information identified as archived is provided for reference, research or recordkeeping purposes. It is not subject to the Government of Canada Web Standards and has not been altered or updated since it was archived. Please contact us to request a format other than those available.

Development Effectiveness Review of the United Nations Children's Fund (UNICEF)

Table of Contents

Annexes

List of Tables

List of Figures

Acknowledgments

The Department of Foreign Affairs, Trade and Development's (DFATD) Development Evaluation Division wishes to thank all who have contributed to this review exercise.  DFATD is grateful for the valued input and support received from all those involved in this collaborative effort.

Our thanks go first to those who collectively facilitated the donor-neutral assessment of the development effectiveness of UNICEF.   The assessment was led by the Netherlands Ministry of Foreign Affairs under the leadership of Mr. Ted Kliest.  A consulting team from Goss Gilroy Inc. conducted the analysis and writing of the donor-neutral assessment.  The team included Mr. Ted Freeman, Ms. Sheila Dohoo Faure, Ms. Louise Mailloux, Ms. Tasha Truant and Mr. Robert Barnes.  The donor-neutral assessment is published on the website of the Organization for Economic Cooperation and Development (OECD).

DFATD produced the chapter related to Canada's relationship with UNICEF. We wish to thank the United Nations, Commonwealth and Francophonie Program Team in the Global Issues and Development Branch at DFATD for their engagement in the review.

From DFATD's Development Evaluation Division, we wish to thank Ms. Lamia Naji, Junior Evaluation Officer, and Dr. Tricia Vanderkooy, Evaluation Manager, for conducting the assessment of Canada's engagement with UNICEF.  The chapter on Canada's engagement with UNICEF was written by Ms. Naji, with editorial assistance from Dr. Vanderkooy.  We also thank Mr. Andres Velez-Guerra, Team Leader and Mr. James Melanson, Evaluation Director, for overseeing the review.

Caroline Leclerc
Head of Development Evaluation

List of Acronyms and Abbreviations

CCC
Core Commitments to Children in Humanitarian Emergencies
CIDA
Canadian International Development Agency
DAC
Development Assistance Committee
DAC-EVALNET
DFATD
DAC Network on Development Evaluation
Foreign Affairs, Trade and Development Canada
DFID
Department for International Development (United Kingdom)
GID
Global Issues and Development Branch (DFATD)
MFM
Global Issues and Development Branch (DFATD)Footnote 1
MOPAN
Multilateral Organizations Performance and Assessment Network
MoRES
Monitoring Results for Equity System
QCPR
Quadrennial Comprehensive Policy Review
UN
United Nations
UNAIDS
Joint United Nations Programme on HIV/AIDS
UNDAF
United Nations Development Assistance Framework
UNEG
United Nations Evaluation Group
UNICEF
United Nations Children's Fund
UNGEI
United Nations Girls' Education Initiative
VISION
Virtual Integrated System of Information
WEOG
Western Europe and Others Group

Executive Summary

Background

This report presents the results of a review of the effectiveness of development and humanitarian programming supported by the United Nations Children's Fund (UNICEF). It was commissioned by the Netherlands Ministry of Foreign Affairs and carried out by a team from Goss Gilroy Inc. of Ottawa, CanadaFootnote 2.  In addition, a review of Canada's engagement with UNICEF has been undertaken by Foreign Affairs, Trade and Development Canada (DFATD),  and is presented in chapter 5.

The common approach and methodology for this review were developed under the guidance of the Organisation for Economic Co-operation and Development's Development Assistance Committee (DAC) Network on Development Evaluation (DAC-EVALNET) and relies on the content of published evaluation reports produced by UNICEF, supplemented with a review of UNICEF corporate documents and consultation with staff at UNICEF headquarters in New York.

Five UNICEF Program Focus Areas

  1. Young child survival and development
  2. Basic education and gender equality
  3. HIV/AIDS and children
  4. Child protection from violence, exploitation and abuse
  5. Policy advocacy and partnerships for children’s rights

Medium-Term Strategic Plan, 2006 – 2013.

Purpose

The purpose of the review is to provide an independent, evidence-based assessment of the humanitarian and development effectiveness of UNICEF operations (hereafter referred to as "programs") for use by all stakeholders. The approach is intended to work in a coordinated way with initiatives such as the DAC-EVALNET/United Nations Evaluation Group (UNEG) Peer Reviews of United Nations organization evaluation functions and assessments carried out by the Multilateral Organizations Performance and Assessment Network (MOPAN). It also recognizes that multilateral organizations continue to make improvements to strengthen their reporting on development effectiveness and should eventually be providing regular, evidence-based, field-tested reporting on effectiveness themselves.

Approach and Methodology

The review, carried out from October 2012 to May 2013, began with a preliminary review of UNICEF documents and the identification of the population of UNICEF evaluation reports found in UNICEF's Global Database and reports from its Global Evaluation Report Oversight System. Interviews were then conducted with evaluation and program staff at UNICEF headquarters in New York. In consultation with evaluation staff, the review team drew a sample of seventy UNICEF evaluation reports, published between 2009 and 2011, that was illustrative of UNICEF programming. Evaluations were selected to include development and humanitarian programming, programming from the various regions and types of countries in which UNICEF works and programming from each of the five focus areas in UNICEF's Medium-Term Strategic Plan. However, it should be noted that the population of evaluations identified at UNICEF under-represented humanitarian action and programming in some of the countries that received the largest amounts of development or humanitarian funding.

The quality of each evaluation was assessed against criteria derived from the UNEG Norms and Standards for evaluation. Since the sample did not include any evaluations found to be "not confident to act" by UNICEF's own Global Evaluation Report Oversight System, only four reports were excluded due to quality concerns. Four others were excluded, either because they were too narrowly focused, did not provide coverage of a minimum number of effectiveness criteria or duplicated evaluation results reported in other evaluations in the sample. As a result, 62 evaluation reports were retained for systematic rating in the review.

Each evaluation report was rated, using a four-point scale that ranged from Highly Unsatisfactory to Highly Satisfactory, on six key Development Effectiveness Criteria and nineteen sub-criteria. The review team also identified the factors contributing to both positive and negative findings for each of the six key criteria used to assess the development effectiveness as reported by the evaluations. The results of the meta-synthesis of evaluation findings were then summarized and presented to UNICEF staff prior to the development of this report.

Assessment Criteria

  1. Relevance of Interventions
  2. The Achievement of Humanitarian / Development Objectives and Expected Results
  3. Cross Cutting Themes (Environmental Sustainability and Gender Equality)
  4. Sustainability of Results/Benefits
  5. Efficiency
  6. Using Evaluation and Monitoring to Improve Humanitarian and Development Effectiveness

Coverage of the Assessment Criteria in Reviewed Evaluations

The review established ranges for assessing how well the sub-criteria were covered in the sixty-two evaluations, based on the number of evaluations that addressed each of the sub-criteria. Coverage could be strong, moderate or weak. The results for sub-criteria with weak coverage were not reported. The coverage was rated as weak for only two sub criteria: the extent to which program-supported changes are environmentally sustainable and the effectiveness of systems for results-based-management. The findings for these sub-criteria are not reported.

Limitation: Retrospective Nature of the Review

The methodology section of the report describes in more detail some limitations of the review related to sampling issues and coverage. It is worth noting here that a review of development effectiveness that relies mainly on published evaluation documents is inherently retrospective, rather than forward-looking. Some important issues identified in the evaluations reviewed have been or are already being addressed by UNICEF. As one prominent example, the results of recent UNICEF global evaluations have been used by Programme Division in the preparation of inputs to next strategic plan. Where this has been pointed out to the team by UNICEF staff, or where there are examples in the documents of UNICEF taking action, the actions are noted in this report.

Coverage Ratings

Strong coverage: Sub-criterion was addressed in 45 – 62 evaluations
Moderate coverage: 30 – 44 evaluations
Weak coverage: less than 30 evaluations

Findings of the Development Effectiveness Review

Relevance of UNICEF-Supported Programs

The relevance of UNICEF's programming is very well covered in the evaluations and the findings reflect that programming is highly relevant to the needs of the target groups. The vast majority of evaluations reported high suitability to target group needs, alignment with national development goals and effective partnerships. However, programming in middle-income countries was somewhat more likely to have a higher rating for effective partnerships than that in least-developed countries.

Alignment of UNICEF's programming with government priorities was the major factor contributing to the relevance of programming, particularly to the extent that programming was designed to support the implementation of public policies, support programming and capacity building in national governments or institutions and was based on partnerships with key stakeholders. When these conditions were not met, notably when there were gaps in UNICEF's partnerships with national government (often because government is not taking adequate ownership) and other UN agencies, this detracts from the relevance of programming.

Objectives Achievement

Objectives achievement is generally very well covered in the evaluations, with the exception of the sub-criterion related to addressing changes in national development policies and programs. The findings are also largely positive. Three-quarters of the evaluations reported positive findings on the extent to which UNICEF-supported programs achieve their stated objectives. Positive results were even stronger for the programs' ability to provide positive benefits for the target population, with nearly nine out of ten evaluations reporting satisfactory or better findings. Results were somewhat less positive on the scale of program benefits measured in terms of the number of program beneficiaries. Finally, three-quarters of the programs were rated positively for their ability to support positive changes in national policies and programs.

It was the high quality of program design (in terms of the focus, alignment and engagement of national partners) that contributed positively to objectives achievement. UNICEF's role in influencing national policy and the appropriateness of the scope and resources for programs also contributed positively to objectives achievement. However, the opposite was also found. Weaknesses in program design, inadequate budgets and insufficient outputs detracted from the achievement of objectives.

Gender Equality

The evaluations presented a challenging picture on the effectiveness of gender equality. Firstly, the coverage of the effectiveness of gender equality in the evaluations was not high. Secondly, fewer than half of the evaluations that did address gender equality contained findings of satisfactory or better for this sub-criterion. Considering that twenty evaluations did not address gender equality, another way of expressing the result would be to note that only one-third of evaluations reflected that the programs effectively addressed gender equality. This finding is surprising given that addressing gender equality represents a "foundation strategy" for UNICEF programming. In addition, it seems surprising that findings of effective gender equality were somewhat more likely to be reported in evaluations of programs in least-developed countries than in middle-income countries.

UNICEF has recently been engaged in efforts to improve effectiveness in the area of gender equality. In response to the 2008 Gender Evaluation, the agency launched the Strategic Priority Action Plan for Gender Equality Action Plan 2010-2012 with eight areas of organizational action identified. UNICEF also reports its intent to take visible action on gender equality in the development of its new strategic plan.

As with other sub-criteria, the presence of specific factors contributed to the effectiveness of gender equality; whereas the absence of these same factors detracted from effectiveness. The factor identified most often as contributing to positive results in gender equality was the inclusion of gender-specific objectives and targets. The most common factors detracting from gender equality effectiveness were the lack of any specific gender equality objectives and/or the lack of a gender perspective in program design and implementation.

Sustainability

Overall the results with respect to sustainability are mixed. Generally the coverage of the sustainability sub-criteria in the evaluations was good. About half the evaluations reflected positive ratings for the likelihood that benefits would continue and the support for sustainability through developing institutional and/or community capacity. Evaluations of programming in middle income countries appeared to be somewhat more likely to get positive ratings for institutional and/or community capacity building than those from least-developed countries. Three-quarters of the evaluation reports were more positive about the strengthening of the enabling environment for development. This may be explained by the fact that efforts to improve the enabling environment do not, in themselves, have to be ongoing activities and, as a result, it may be easier to achieve positive findings for this sub-criterion than for other sub-criteria, which require evidence of the long-term impacts.

Strong national government and community engagement contributed positively to sustainability. On the other hand, weaknesses in program design and implementation (notably the failure to plan for sustainability), the lack of adequate capacity at the national level and an ongoing dependence on donor support detracted from sustainability.

Efficiency

The coverage of efficiency in the evaluations was relatively good, but findings are mixed and require careful interpretation. The fact that two-thirds of the evaluations reported positive findings on the cost efficiency of programs represents a reasonably good result.

Nonetheless, it is a concern that, of the evaluations that addressed these sub-criteria, half reported negative findings on timeliness and two-thirds reported negative findings on the efficiency of administrative systems. Also worrying is the fact that about one in five reported a finding of highly unsatisfactory for both sub-criteria. However, since there may be a tendency to under-report findings with respect to timeliness and the efficiency of administrative systems when no problems are encountered at field level, the results may be biased toward negative findings.

Effective monitoring systems, the use of low-cost approaches to service delivery and strong financial planning and cash-flow management contributed to cost efficiency. On the other hand, the lack of regular and timely reporting of appropriate cost data detracted from cost efficiency. Some evaluations also noted that delays in program start-up and the supply of inputs was a factor that reduced cost efficiency.

Similar themes are reflected in the factors that contributed to the timeliness of, and the effectiveness of systems and procedures for, program implementation and follow-up. The strong management and programming capacity of UNICEF country and regional offices, timely coordination among partner organizations, effective program monitoring and regular supervision and effective cash flow management all contributed positively to a rapid respond to program demands. The factors that detract from timeliness are related to UNICEF's administrative systems and procedures, notably lengthy delays in procurement and rigid and cumbersome financial systems and procedures for funds disbursement delayed program implementation.

Using Evaluation and Monitoring to Improve Humanitarian and Development Effectiveness

Coverage for the sub-criteria regarding the use of monitoring and evaluation was strong and the results are somewhat positive. Between half and two-thirds of the evaluations received positive ratings for the effectiveness of systems and processes for evaluation, and for the use of evaluation to improve effectiveness. However, just over one-third of the evaluations got a positive rating for the effectiveness of systems and processes for results monitoring and reporting.

The evaluations identified few factors that contributed positively to the use of monitoring and evaluation results to improve development effectiveness, beyond the fact that being part of a broader evaluation process (either UNICEF's country program evaluations or a donor-led global evaluation) contributed to the use of monitoring and evaluation. The use of evaluation is supported by the increasing tendency to prepare evaluation management responses that include action plans for the implementation of evaluation recommendations. However, two factors detracted from the use of monitoring and evaluation: the lack of a clear results framework and inadequate or inappropriate indicators, and inadequate baseline information.

Since 2011, UNICEF has engaged in a very significant effort to strengthen program results definition, monitoring and reporting, as it developed and implemented the Monitoring Results for Equity System (MoRES). This system is to be subject to a formative evaluation to be carried out by UNICEF's Evaluation Office in 2013. In addition, UNICEF recently used existing global evaluation reports to inform the development of its upcoming strategic plan. These initiatives appear to contribute to addressing gaps in monitoring and evaluation that were identified in the current review of evaluations.

Conclusions

  1. UNICEF-supported programs are highly relevant to the needs of target group members and are supportive of the development plans and priorities of program countries. UNICEF has also had success in developing effective partnerships with government and non-governmental organizations, especially in responding to humanitarian situations. The relevance of UNICEF programming has also been supported by efforts to ensure alignment with key national development plans and policy documents.
  2. UNICEF has largely been effective in achieving the objectives of the development and humanitarian programs, in securing positive benefits for target group members and in supporting positive changes in national policies and programs. Where UNICEF programs have achieved success in objectives achievement, it has often been supported by high quality program design and by UNICEF's role in influencing national sector policies. On the other hand, when UNICEF-supported programs failed to meet their objectives, it was most often due to weaknesses in project design, often linked to unclear causal relationships and lack of a results orientation.
  3. These weaknesses in project design were contributing factors to mixed results with respect to sustainability. UNICEF achieves fairly positive ratings for its contributions to strengthening the enabling environment for development. However, the results for the likely continuation of results or the development of institutional and community capacity are not as good. A key factor to explain these results is the failure to plan for sustainability and integrate sustainability into program designs.
  4. UNICEF's performance with respect to gender equality is a serious concern. Coverage of gender equality in the evaluations was not strong and, for those that did address it, the results were weak. A major factor contributing to poor results was the lack of specific objectives regarding gender equality or the lack of a gender perspective in program design and implementation. Given the identification of gender equality as a "foundation strategy" at UNICEF, it is surprising that gender is adequately addressed in only two-thirds of the evaluations.
  5. There was insufficient coverage of environmental sustainability to warrant presentation of the results. It appears that environmental sustainability or the impact of UNICEF-supported programs on their environment is not addressed in most evaluations, although some evaluations of humanitarian action and programs in water, sanitation and hygiene did address the issue. Given increasing emphasis on programs to mitigate the effects of climate change in some UNICEF country programs, coverage of environmental sustainability may be expected to improve in the future.
  6. The results for efficiency of programming are mixed and the interpretation of the results is difficult, as efficiency is not covered systematically in all evaluations. It is likely that factors related specifically to timeliness and implementation systems are only addressed in evaluations if they are problematic and can help to explain weakness in objectives achievement. However, the sub-criterion related to cost efficiency was covered more systematically in the evaluations and showed somewhat positive results. However, this reflects ways in which programs have tried to reduce unit costs and increase the efficiency of resource use, rather than an analysis of overall program costs, as these costs were not identified in nearly half the evaluations. The factors that enhance cost efficiency include the establishment of effective monitoring systems to track costs and improve efficiency through lower delivery costs or unit prices and efforts to identify low-cost approaches to programming. However, where the results are not so positive, the detracting factor is most often the lack of appropriate cost data.
  7. The evaluations reflected somewhat positive findings with respect to the use of evaluation at UNICEF, but less so for monitoring and reporting. The use of evaluations is supported by an increasing tendency to prepare evaluation management responses that include action plans. There were insufficient ratings for the use of results-based management systems to report on the results. The lack of clear results frameworks and appropriate indicators and baseline information were factors that detracted from UNICEF's effective use of monitoring and evaluation systems. The current development and implementation of the Monitoring Results for Equity Systems (MoRES) represents a significant effort to address this issue.
  8. While the sample of evaluations reviewed reflects reasonably the population of UNICEF evaluations between 2009 and 2011, a review of the profile, by country and by type of programming, of the evaluations conducted suggests that UNICEF does not have adequate coverage of the programming in the countries that receive the largest amounts of both development and humanitarian funding. There is noticeably limited coverage of UNICEF's humanitarian action.
  9. UNICEF is investing considerable time and effort into the development of systems for results monitoring, which is being implemented at all levels of the organization and in using the results of global evaluations for strategic planning. These initiatives hold out the promise of strengthened results reporting which could negate the need for development effectiveness reviews of this type in the future. That could depend on how well the evaluation function is able to be incorporated into the system as a means of verifying UNICEF's contribution and testing the validity of theories of change.

UNICEF has undertaken a number of initiatives in the period following publication of the evaluations reviewed in order to address some of the issues reported here. Examples of the main initiatives include:

While these have not been assessed, documents and interviews at UNICEF suggest that they represent a significant effort to respond to the issues identified. Their effectiveness will doubtless be the subject of future evaluations.

Canada's Relationship with UNICEF

The review of Canada's engagement with UNICEF identified the following strengths:

The Canada-specific review identified the following areas of opportunity:

The Canada-specific review concludes with recommendations for DFATD to consider in its engagement with UNICEF: 

  1. DFATD should continue to emphasize the integration of gender equality in UNICEF programs and should highlight the importance of performance measurement and reporting in this area.   
  2. DFATD should develop a performance measurement framework to assess progress on Canada's strategic objectives for its relationship with UNICEF.
  3. As Canada and other members of the General Assembly are seeking mechanisms to harmonize the UN development system, DFATD should continue to work with UNICEF to identify realistic and tangible ways to collaborate with other UN organizations.
  4. DFATD should encourage UNICEF to improve its sustainability planning. This could involve efforts to strengthen UNICEF's program design and implementation plans.
  5. DFATD should encourage UNICEF to increase its outcome-level reporting and evaluation of humanitarian assistance.

1.0  Introduction

1.1 Background

This report presents the results of a review of the effectiveness of development and humanitarian programming supported by the United Nations Children's Fund (UNICEF). It was commissioned by the Netherlands Ministry of Foreign Affairs and carried out by a team from Goss Gilroy Inc. of Ottawa, CanadaFootnote 3.  In addition, a review of Canada's engagement with UNICEF has been undertaken by Foreign Affairs, Trade and Development Canada (DFATD),  and is presented in chapter 5.

The common approach and methodology for reviews of this type were developed under the guidance of the Organisation for Economic Co-operation and Development's Development Assistance Committee (DAC) Network on Development Evaluation (DAC-EVALNET). The review relies on the content of published evaluation reports produced by UNICEF, supplemented with a review of UNICEF corporate documents and consultation with staff at UNICEF headquarters in New York.

The method uses a common set of assessment criteria derived from the DAC's evaluation criteria (Annex 1). It was pilot tested during 2010 using evaluation material from the Asian Development Bank and the World Health Organization. The overall approach and methodology were endorsed by the members of the DAC-EVALNET as an acceptable approach for assessing the development effectiveness of multilateral organizations in June 2011. The first full reviews using the approved methodology were conducted in 2011/2012. The Canadian International Development Agency (CIDA)Footnote 4 led the review of the United Nations Development Fund and the World Food Programme, with support and participation from the Netherlands Ministry of Foreign Affairs.

In 2012/13, another two development effectiveness reviews were carried out. Canada took the lead role in an assessment of the African Development Bank while the Netherlands Ministry of Foreign Affairs was responsible for the review of UNICEF. The review of UNICEF took place from October 2012 to May 2013.  It included three rounds of consultations with UNICEF, including two days of meetings at project startup, a presentation and review of preliminary findings held in March 2013 and the presentation and discussion of the final report on May 6, 2013.   The final report was also presented and discussed at a meeting of interested UNICEF donor agencies, hosted by the Netherlands Permanent Mission to the United Nations in New York City on May 7, 2013.

From its beginnings, the process of developing and implementing the reviews of development effectiveness has been coordinated with the work of the Multilateral Organization Performance Assessment Network (MOPAN). By focusing on development effectiveness and carefully selecting assessment criteria, the reviews seek to avoid duplication or overlap with the MOPAN process.

The intent behind the initiative is that EVALNET members engage in development reviews of multilateral organizations as distinct assessments in parallel to the MOPAN reviews. The planning of the reviews should take place in consultation with the evaluation department of the respective multilateral organization. The results of the development reviews will then be of use for the multilateral organizations and their stakeholders.

1.2 Purpose

The purpose of the review is to provide an independent, evidence-based assessment of the humanitarian and development effectiveness of UNICEF operations (hereafter referred to as "programs") for use by all stakeholders.

The current approach to assessing the development effectiveness of multilateral organizations was developed in order to address a gap in the information available to bilateral development agencies. While MOPAN provides regular reports on the organizational effectiveness of multilateral organizations, only in 2012 has it begun to examine development effectiveness and has not yet fully addressed the information gap this review is meant to fill. Other options such as large-scale, joint donor-funded evaluations of a given multilateral organization are much more time-consuming, more costly and result in a significant management burden to the organization being evaluated before, during and after such evaluations.

The current approach is intended to work in a coordinated way with initiatives such as the DAC-EVALNET/United Nations Evaluation Group Peer Reviews of United Nations organization evaluation functions. It also recognizes that multilateral organizations continue to make improvements and strengthen their reporting on development effectiveness. The ultimate aim of the approach is to be replaced by regular, evidence-based, field-tested reporting on development effectiveness provided by multilateral organizations themselves.

1.3 Structure of the Report

The report is structured as follows:

Box 1: Five UNICEF Program Focus Areas

  1. Young child survival and development
  2. Basic education and gender equality
  3. HIV/AIDS and children
  4. Child protection from violence, exploitation and abuse
  5. Policy advocacy and partnerships for children's rights

Medium-Term Strategic Plan, 2006 – 2013.

1.4   UNICEF: A Global Organization Focused on Equity for Children

1.4.1   UNICEF's Strategic Direction and Focus on Equity

UNICEF's mission statement reiterates its mandate from the United Nations to "advocate for the protection of children's rights, to help meet their basic needs and to expand their opportunities to reach their full potential."Footnote 5 Since 2006, it has been guided in carrying out this mission by a Medium-Term Strategic Plan,Footnote 6 which originally covered the period from 2006 to 2009 but has been extended twice and is now referred to as the Medium-Term Strategic Plan for 2006-2013.

The Medium-Term Strategic Plan sets out fifteen guiding principles, five program focus areas and two foundation strategies. The guiding principles link UNICEF's programming to normative documents such as the Convention on the Rights of the Child, the Convention on the Elimination of All Forms of Discrimination Against Women and the Millennium Declaration and Millennium Development Goals (MDG). They also commit the organization, inter alia, to a strategy of capacity development, working firmly within the United Nations system, and adapting program strategies to the needs of program countries.

The five program focus areas are a central organizing principle of the Medium-Term Strategic Plan (Table 1). They are intended to "have a decisive and sustained impact on realizing children's rights and achieving the commitments of the Millennium Declaration and Goals."Footnote 7

The Medium-Term Strategic Plan provides a description of the types of activities UNICEF will undertake and the programs it will support in each of its five focus areas (Table 1).

Table 1: UNICEF Activities and Programs in Five Focus Areas of the Medium-Term Strategic PlanFootnote 8
UNICEF Focus AreasUNICEF Program Contents
1.  Young child survival and developmentSupport in regular, emergency and transitional situations for essential health, nutrition, water and sanitation programs, and for young child and maternal care at the family, community, service-provider and policy levels.
2.  Basic education and gender equalityFocus on improved developmental readiness for school; access, retention and completion, especially for girls; improved education quality; education in emergency situations and continued leadership of the United Nations Girls' Education Initiative (UNGEI).
3.  HIV/AIDS and childrenEmphasis on increased care and services for children orphaned and made vulnerable by HIV/AIDS, on promoting expanded access to treatment for children and women and on preventing infections among children and adolescents; continued strong participation in the Joint United Nations Programme on HIV/AIDS (UNAIDS).
4.  Child protection from violence, exploitation and abuseStrengthening of country environments, capacities and responses to prevent and protect children from violence, exploitation, abuse, neglect and the effects of conflict.
5.  Policy advocacy and partnerships for children's rightsPutting children at the centre of policy, legislative and budgetary provisions by: generating high-quality, gender-disaggregated data and analysis; using these for advocacy in the best interests of children; supporting national emergency preparedness capacities; leveraging resources through partnerships for investing in children; and fostering children's and young people's participation as partners in development.

In addition to the five focus areas, the Medium-Term Strategic Plan reiterates UNICEF's intention to build its capacities to respond to emergencies in a timely and effective manner. It notes that responding to humanitarian situations remains an essential element of the work of UNICEF.

Finally, the Medium-Term Strategic Plan includes a major commitment to the pursuit of equity by focusing on gender equality and the rights of the most vulnerable. It identifies a human-rights based approach to programming and promoting gender equality as "foundation strategies" for UNICEF.

Applying a human rights-based approach and promoting gender equality, as "foundation strategies" for UNICEF work will improve and help to sustain the results of development efforts to reduce poverty and reach the Millennium Development Goals by directing attention, long-term commitment, resources and assistance from all sources to the poorest, most vulnerable, excluded, discriminated and marginalized groups.Footnote 9

The principles, focus areas and foundation strategies outlined in UNICEF's Medium-Term Strategic Plan have been ratified by two Mid-Term Review Reports (2008 and 2010).Footnote 10

The most recent annual report of the Executive Director of UNICEF placed special emphasis on the organization's "refocus on equity":

The refocus on equity holds abundant promise for children, especially through faster and more economical achievement of the Millennium Development Goals. UNICEF emphasized the implementation of the refocus during 2011 at the local and national levels in partnership with governments, civil society organizations and United Nations partners. Measuring results for the most disadvantaged populations, particularly at the local level, has proven critical to accelerate and sustain progress in reducing disparities.Footnote 11

Thus, UNICEF's strategy for 2006-2013 has placed considerable emphasis on: addressing inequities faced by vulnerable children through programming in its five program focus areas; responding effectively to humanitarian situations; and pursuing foundation strategies of a human rights-based approach to programming and promoting gender equality.

1.4.2   UNICEF Operations and Program Expenditures

UNICEF is active in over 190 countries and territories through its programs and the work of national committees. In 2011, it had programs of cooperation in 151 countries, areas and territories: forty-five in sub-Saharan Africa; thirty-five in Latin America and the Caribbean; thirty-five in Asia; sixteen in the Middle East and North Africa; and twenty in Central and Eastern Europe and the Commonwealth of Independent States.Footnote 12

UNICEF's spending on program assistance in 2011 totalled $3,472 million. Over half of that amount went to the Young Child Survival and Development focus area. Similarly, 57% of total program assistance in 2011 was directed to sub-Saharan Africa, which has the majority of the world's Least Developed Countries.Footnote 13

Figure 1 shows the portion of UNICEF direct program expenditures in each focus area and in humanitarian action from 2009 to 2011.

Figure 1: UNICEF Expenditures, by Focus Area and Humanitarian Action, 2009-2011 Footnote 14

Figure 1 Text Alternative

1. Young child survival and development 33%
2. Basic education and gender equality 16%
3. HIV/AIDs and children 5%
4. Child Protection 7%
5. Policy advocacy and partnership 10%
6. Other 1%
7. Humanitarian Emergency 27%

Expenditures in the five focus areas are development expenditures funded with a) Regular Resources and b) Regular Resources – Other, using UNICEF's system for accounting for resources. Expenditures on humanitarian action are those financed by c) Other Resources – Emergency. The distribution of expenditures by type of funding from 2009 to 2011 is shown in Figure 2.

Figure 2: UNICEF Expenditures, by Type, 2009 - 2011 Footnote 15

Figure 2 Text Alternative

Regular Resources 24%
Other Resources - Regular 49%
Other Resources - Emergency 27%

1.5   Program Evaluation at UNICEF

One of the conditions for undertaking a development effectiveness review is the availability of enough evaluation reports of reasonable quality to provide an illustrative sample covering an agency's activities and programs (Annex 2). In order to satisfy that condition, the review team examined program evaluation policies and practices and the quality of evaluation reports produced and published by UNICEF. This short survey is by no means exhaustive and should not be read as an overall assessment of the function. It is only intended to establish the feasibility of conducting a development effectiveness review of UNICEF based on its own published evaluations.

1.5.1 Evaluation Policies and Practices at UNICEF

UNICEF's current evaluation policyFootnote 16 was prepared, at least in part, in response to the report of a peer review panel of international evaluation experts produced under the auspices of the DAC-EVALNET and the United Nations Evaluation Group in 2006. That report was just the second in the ongoing series of peer reviews of evaluation functions in the multilateral system. In its judgment statement, the report recognized the strength of UNICEF's Evaluation Office and the ongoing challenge of ensuring evaluation quality and adequate coverage in a decentralized system of program evaluation:

Evaluation at UNICEF is highly useful for learning and decision-making purposes and, to a lesser extent, for accountability in achieving results.

UNICEF's central Evaluation Office is considered to be strong, independent and credible. Its leadership by respected professional evaluators is a major strength. The EO has played an important leadership role in UN harmonization through the UN Evaluation Group.

The Peer Review Panel considers that a decentralized system of evaluation is well-suited to the operational nature of UNICEF. However, there are critical gaps in quality and resources at the regional and country levels that weaken the usefulness of the evaluation function as a management tool.Footnote 17

The 2008 evaluation policy for UNICEF retained the decentralized system wherein responsibility for different types of evaluations rested at different levels in the organization. You will below an overview of evaluation roles and responsibilities by organizational location at UNICEF as described in the policy.

Organizational Responsibilities for Program Evaluation at UNICEF Footnote 18

Organizational Location

Country Offices/Country Representatives

Evaluation Roles and Responsibilities

Types of Evaluations Commissioned

Organizational Location

Offices/Regional Directors

Evaluation Roles and Responsibilities

Focuses on oversight and strengthening the capacity of country office evaluation functions by:

Types of Evaluations Commissioned

Organizational Location

Directors at Headquarters

Evaluation Roles and Responsibilities

Types of Evaluations Commissioned

Organizational Location

Evaluation Office

Evaluation Roles and Responsibilities

Types of Evaluations Commissioned

Organizational Location

Evaluation Committee

Evaluation Roles and Responsibilities

UNICEF reports annually to the Executive Board on the performance of the evaluation function. This report encompasses both the number and types of evaluations produced and progress in monitoring and improving the quality of the evaluations produced.

For the past three years, the quality of UNICEF evaluations produced by country and regional offices has been assessed independently by a contracted external agency applying a consistent methodology. The results of the Global Evaluation Report Oversight System (GEROS) are published annually. The results of the Global Evaluation Report Oversight System (GEROS) annual review are incorporated into the annual report on the evaluation function. Table 2 summarizes the evaluation results indicators reported in the 2012 report.

Table 2: Performance Indicators for the Evaluation Function at UNICEF 2011 Footnote 19
Evaluation Performance IndicatorResults for 2011

1. Number of evaluations managed and submitted to the Global Database

  • 140 evaluations
  • 700 studies
  • Avg. = 1 evaluation and 7 studies per office

2. Topical distribution

  • 25% Child survival and development
  • 15% Education and gender
  • 10% HIV/AIDS
  • 16% Child protection
  • 1% Policy, advocacy and partnership
  • 33% Cross cutting themes and organizational performance

3. Type of evaluations conducted

  • 33% Impact
  • 26% Outcomes
  • 33% Outputs

4. Quality of UNICEF evaluations

Global evaluation report oversight system (GEROS)  

Rating for evaluations assessed in 2011, conducted in 2010

  • 40% satisfactory
  • 30% nearly satisfactory
  • 30% poor quality

5. Use of evaluation, including management responses

  • 65% of evaluations upload management response to tracking system

6. Corporate-level evaluations

  • 13 corporate level evaluations (from a target of 15) completed or in progress by April 2012

The annual GEROS report represents an important effort by the Evaluation Office to monitor and influence the quality of evaluations produced by regional and country offices. This has been further strengthened by the development of a dashboard for summarizing information on the decentralized evaluation function (regional and country office level) which is part of UNICEF's Virtual Integrated System of Information (VISION) launched in January 2012.

The decentralized evaluation function dashboard is uploaded quarterly and presents (for each UNICEF regional office) information on: evaluation report submission rates; the quality of evaluation reports as assessed by Global Evaluation Report Oversight System (GEROS)  in the previous year; management responses uploaded to the Global Tracking System; and the implementation rates for management responses.

What is interesting is the absence of a metric used to measure coverage of country office programs in the Virtual Integrated System of Information (VISION) dashboard for decentralized evaluation. The review team found (Section 2.0) considerable variation in the number of evaluations carried out over three years (2009 to 2011) by the country offices. Nonetheless, it is clear that UNICEF's Evaluation Office is engaged in ongoing efforts to upgrade the quality of decentralized evaluations.
Finally, it is worth pointing out that recent global evaluations managed by the Evaluation Office have been used by the Programme Division at UNICEF to inform the planning process for the 2014 to 2017 Medium-Term Strategic Plan.Footnote 20

1.5.2   Quality of Evaluation Reporting at UNICEF

Data on evaluation quality at UNICEF is available for the review from two sources: UNICEF's independent Global Evaluation Report Oversight System (GEROS) reports and the results of the review team's quality assessment of evaluations in the sample.

The latest report of the Global Evaluation Report Oversight System (GEROS) was produced in 2012 and covers 87 evaluation reports produced by UNICEF in 2011. It reports a three-year improving trend with 42% of reports meeting UNICEF evaluation standards in 2011 versus 40% in 2010 and 36% in 2009.Footnote 21

The review took the Global Evaluation Report Oversight System (GEROS) ratings into account when developing the evaluation sample as noted in Annex 3. There are four evaluation report quality ratings used in the Global Evaluation Report Oversight System (GEROS) system: Not confident to act; Almost confident to act; Confident to act; and, Very confident to act. In developing a purposive sample of the evaluation reports produced from 2009 to 2011, the review team included only those evaluations rated almost confident to act or better. The review team assessed a total of 70 evaluations for quality and rejected only four for quality reasons. Four others were not included in the review for other reasons (Section 2.1 and Annex 3).

1.6   Monitoring and Reporting Results

1.6.1   Elements of UNICEF's Results Monitoring and Reporting Systems

In recent years, UNICEF has invested considerable effort in modifying and strengthening its systems and processes for monitoring and reporting on the results of programs, and for allocating resources to specific results. Currently, some of the most notable elements of the system include the Monitoring Results for Equity Systems (MoRES), the Virtual Integrated System of Information (VISION) and the Executive Director's annual reports on progress and achievement against the Medium-Term Strategic Plan.

Monitoring Results for Equity Systems

The Monitoring Results for Equity Systems (MoRES) system is one of the core elements in UNICEF's "refocus on equity" which began in 2010 as an initiative of the Executive Director. Monitoring Results for Equity Systems (MoRES) is UNICEF's conceptual framework for planning, programming, implementation and management of results.Footnote 22 It is intended to address the need for intermediate process and outcome measures between the routine monitoring of program and project inputs and outputs on one hand and the monitoring of higher level outcomes; which is done using internationally agreed indicators every three to five years.

The conceptual model for Monitoring Results for Equity Systems (MoRES) includes four levels of country office programmatic actions, each linked to different planning, monitoring, and evaluation activities. Level three monitoring, which is the primary focus of MoRES, is focused on the questions: "Are we going in the right direction? Are bottlenecks and barriers to equity changing?"Footnote 23

In order to undertake level three monitoring, country offices must first identify country specific indicators relating to ten determinants of bottlenecks hindering equity for children. These determinants are gathered under four major headings: the quality, demand for, and supply of services; and the enabling environment.

In 2012, the system of indicators for results in humanitarian action at UNICEF, the Humanitarian Performance Monitoring System, was incorporated into Monitoring Results for Equity Systems (MoRES) to allow for a single, integrated system of performance monitoring.

It is worth noting that the conceptual model of Monitoring Results for Equity Systems (MoRES) includes the evaluation function as the means of verifying final outcomes and impacts of programming at country level.

The global two-year management plan for Monitoring Results for Equity Systems (MoRES) was adopted by UNICEF in January 2012, and the level three monitoring module was implemented by 27 country offices in the March-June period of the same yearFootnote 24. Mainstreaming of Monitoring Results for Equity Systems (MoRES) into the annual review process took place in the last half of 2012. As an important element of the refocus on equity, Monitoring Results for Equity Systems (MoRES) is to be the subject of a formative evaluation managed by the Evaluation Office in 2013.

Virtual Integrated System of Information

The Virtual Integrated System of Information (VISION) system, fully launched in January 2012, allows UNICEF offices at all levels to develop and input data on different types and levels of results. Perhaps most importantly it links together budgeting data and key results so that country offices can now input their budget information under key UNICEF results areas rather than program expenditure activities.

A key component of VISION is its Results Assessment Module which can be used to extract information on program planned results and compare those to reported achievements. During 2011, Monitoring Results for Equity Systems (MoRES) level three monitoring elements were integrated into Virtual Integrated System of Information.

Annual Report of the Executive Director of UNICEF: progress and achievement against the medium term plan

The Executive Director of UNICEF reports annually on the progress made and achievements secured against the key results identified in the Medium-Term Strategic Plan. This report combines information on the global trends of indicators relating to child-specific Millennium Development Goals (and other global, regional and country-specific indicators of child welfare and disparity) with examples of UNICEF's contribution to results in each of the five focus areas.

The annual report also addresses developments in the implementation of programme principles and strategies and on internal measures of programme performance. In 2012, the report highlighted developments in evaluation, country office efficiency, shared services with other UN agencies, and new initiatives in recruiting and human resource management.

The annual report is supported by a large data companion report which tracks a significant number of key performance indicators across the five focus areas and humanitarian action.

1.6.2 Strengths and Weaknesses of the Results Monitoring and Reporting System

The main strength of UNICEF's results monitoring and reporting system as it currently stands is its multi-layered nature and its ability to link standardized categories of results to resource allocations and budgets through the Virtual Integrated System of Information (VISION) system. With Monitoring Results for Equity Systems (MoRES) level three monitoring integrated into the Virtual Integrated System of Information (VISION) system there should be strong linkages from development and humanitarian program inputs to the intermediate level of results.

What is lacking currently from the annual report of the Executive Director and its data companion is a clear explanation of how UNICEF's activities and its support to programs contribute to the higher level outcomes and impacts being reported. The contribution is asserted and examples are given, but it is not clear how the link between UNICEF's contribution and the outcomes is established. At the same time, it is important to recognize that Monitoring Results for Equity Systems (MoRES), by focusing on determinants of inequality for children and addressing indicators of intermediate results for UNICEF-supported programs, goes some way in addressing the issue of UNICEF's contribution.

The need for strengthened monitoring is recognized in Programme Division's recent work on the challenges of the upcoming Medium-Term Strategic Plan.Footnote 25

Monitoring and demonstrating tangible results is also central for UNICEFs results and performance-based management and reporting to donors, particularly in the light of increasing competition for humanitarian and development funding. The expansion of the assessment, monitoring, analysis and evaluation functions across country offices and Government partners will require additional investments in staff, training and technical assistance. The next Medium-Term Strategic Plan should emphasize monitoring as a key program function and highlight the investments needed to improve monitoring functions on the ground.

One development which may potentially strengthen the link between higher-level outcomes and UNICEF's contribution is the explicit emphasis in Monitoring Results for Equity Systems (MoRES) of the role which evaluation can play in verifying outcomes and impacts of country level programming. This will depend, of course, on the development of credible theories of change for country programs and their elements. To some extent, the Monitoring Results for Equity Systems (MoRES) model addresses this question through its identification of determinants of equity. However, this is not yet made explicit in reporting on the progress against the Medium-Term Strategic Plan.

In summary, UNICEF is investing considerable time and effort into the development of quite new systems for results monitoring which are being implemented at all levels of the organization. They hold out the promise of strengthened results reporting which could negate the need for development effectiveness reviews of this type in the future. That could depend on how well the evaluation function is able to be incorporated into the system as a means of verifying UNICEF's contribution and testing the validity of theories of change.

2.0 Methodology

This section describes the methodology used for the review, including the identification of the population and the sampling process, the review criteria, the review and analysis processes and other data collection used. It concludes with a discussion of the limitations of the review.

2.1   Evaluation Population and Sample

A population of UNICEF evaluations was identified from two sources:

These sources identified a population of 341 evaluations. It was decided, in consultation with UNICEF's Office of Evaluation, that the review should focus on evaluations conducted since 2008. UNICEF had adopted a new evaluation policy in that year and it was expected that evaluations after 2008 would reflect this new policy. It was also determined that the sample would not include those evaluations deemed, in the GEROS quality rating, to be of poor quality and "not confident to act." Evaluations that focused on global programming were not included in the population of evaluations for the quantitative review, but included in a qualitative review by senior team members. In addition, some evaluations in the population were duplicates, were not deemed to be evaluations or did not focus on UNICEF programming and were dropped from the population. This resulted in a population of 197 from which to draw the sample.

A sample of 70 evaluations was drawn from the population, with the intention of having 60 evaluations that would pass the quality review and be available for rating. Initially the sample was random, stratified by Medium-Term Strategic Plan theme area. Then the sample was adjusted manually to ensure adequate coverage on two other dimensions: region and type of country (least-developed or middle-income countries). The resulting sample was no longer random, but rather a purposive sample that is illustrative of UNICEF's programming across a number of dimensions. Eight evaluations were eliminated from the initial sample – four because they did not pass the quality review, two because they did not cover UNICEF programming or address program effectiveness and two because they covered activities that were already covered in other evaluations. This left a sample of 62 evaluations that were rated by the team (Annex 2).

The key characteristics of the population and the sample include:

See Annex 3 for details of the comparison of the evaluation population and sample.
It was not possible to identify the value of UNICEF programming that was covered by the evaluation sample. While the review process required the raters to identify the overall costs of the programs, this information was found in only 34 evaluations and the nature of the information presented was inconsistent across evaluations.Footnote 28

Apart from the fact that the evaluations do not cover the countries receiving the largest amounts of funding, or the full range of humanitarian action, the evaluations covered in the sample are, in all other aspects, illustrative of UNICEF's global programming.

2.2   Criteria

The methodology does not rely on a particular definition of development and humanitarian effectiveness. As agreed with the Management Group and the Task Team that were created by the DAC-EVALNET to develop the methodology of these reviews, the methodology focuses on some essential characteristics of effective multilateral organization programming, derived from the DAC evaluation criteria:

Box 2: Assessment Criteria

  1. Relevance of Interventions
  2. The Achievement of Humanitarian and Development Objectives and Expected Results
  3. Cross Cutting Themes (Environmental Sustainability and Gender Equality)
  4. Sustainability of Results/Benefits
  5. Efficiency
  6. Using Evaluation and Monitoring to Improve Humanitarian and Development Effectiveness

The review methodology, therefore, involves a systematic and structured review of the findings of UNICEF evaluations, as they relate to six main criteria (Box 2) and 19 sub-criteria that are considered to be essential elements of effective development and humanitarian programming (Annex 1).

2.3   Review Process and Data Analysis

Each evaluation was reviewed by one member of a small review team that included three reviewers and two senior members (including the team leader).Footnote 29 Each team member reviewed a set of evaluations. The first task of the reviewer was to assess the quality of the evaluation report to ensure that it was of sufficiently high quality to provide reliable information on development and humanitarian effectiveness. This was done using a quality scoring grid (Annex 3). If the evaluation met the minimum score required, the reviewers continued to provide a rating on each sub-criterion, based on information in the evaluations and standard review grid (Annex 4). They also provided evidence from the evaluations to substantiate the ratings.

All efforts were made to ensure consistency in the ratings by team members. The reviewers were trained by the senior team members; two workshops were held at which all team members reviewed and compared the ratings for the same three evaluations, a mid-term workshop was held to address any issues faced by the reviewers, and, following completion of the reviews and the documentation of the qualitative evidence to support the ratings, the team leader reviewed all ratings to ensure there was sufficient evidence provided and it was consistent with the rating. Senior team members then reviewed the qualitative evidence for each sub-criterion to identify factors contributing to, or detracting from, the achievement of the sub-criteria.

2.4   Other Data Collection

After the rating of the evaluations in the sample, further information about UNICEF's effectiveness was also gleaned from a qualitative review of the global evaluations conducted in the 2009 - 2011 period. These were not included in the sample for rating in order to avoid double-counting results. Some global evaluations included evidence drawn from other evaluations and, as such, there was a risk that evaluative information on the same program would be counted twice. The separate review of the global evaluations also gave them a higher profile in the analysis of the findings for the six criteria.

The review of evaluation reports was also supplemented by a review of UNICEF corporate documents and interviews with UNICEF staff. These were done to contextualize the results of the review. (A list of the global evaluations and documents consulted is provided in Annex 5.)

2.5   Limitations

As with any meta-synthesis, there are methodological challenges that limit the findings. For this review, the limitations include sampling bias, the challenge of assessing overall programming effectiveness when important variations in programming exist (for example, evaluations covering multiple programming components or only a specific theme or project as part of a program area) and the retrospective nature of a meta-synthesis.

Sampling Bias

The sample selected for this review is not a random sample. Even as a purposive sample, it is still illustrative of UNICEF-supported programming. However, caution must be exercised in generalizing from the findings of this review to all UNICEF programming. In addition, although the UNICEF's evaluations cover a range of countries and each focus area, Medium-Term Strategic Plan there are fewer evaluations of programming in countries receiving the largest amounts of UNICEF funding and its humanitarian action than would be expected, given its proportion of overall funding at UNICEF. Howeve r, the review did attempt to compensate for this by over-sampling from countries with the largest amounts of funding and those with humanitarian crises.  As already noted, neither the population nor the sample included evaluations of small (Level One) humanitarian actions at UNICEF.

There was generally adequate coverage of the criteria. Seventeen of the nineteen sub-criteria used to assess development and humanitarian effectiveness are sufficiently well covered in the evaluations included in this report. Two received a weak coverage rating and their results are not reflected in this report.

Variations in Programming

The review was not able to report systematically on the effectiveness of UNICEF's programming by focus area or by type of country. There were not sufficient evaluations in each focus area or type of country included in the meta-synthesis to allow for them to be analyzed separately. In addition, some evaluations cover multiple types of programming in the same evaluation. This means that it is not possible to draw conclusions by type of programming or type of country. Although not analyzed separately, where qualitative observations can be made about the focus area and/or the country type, they are reflected in the report. However, these observations should be treated with caution, as they can only be illustrative. The quantitative findings, where appropriate, for these analyses are reported in Annex 6.

Retrospective Nature of Meta-synthesis

Evaluations are, by definition, retrospective and a meta-synthesis is even more retrospective, as it is based on a body of evaluations that address policies and programming implemented over a much earlier period of time. UNICEF's evaluations published in 2009 - 2011, covered years beginning in 2000 and running to 2011.Footnote 30UNICEF's policies, strategies and approaches to programming have changed over these years, but the changes will not be reflected in all the evaluations. For this reason, the findings may be somewhat dated. To the extent possible, the review addressed this through observations gleaned from recent interviews with UNICEF staff and a review of UNICEF documents.

3.0  Findings on UNICEF's Development Effectiveness

This section presents the results of the review as they relate to the six main criteria and their associated sub-criteria. For each criterion, the report presents firstly the extent to which the review sub-criterion was addressed in the evaluation reports (coverage). Then the results are presented (findings) – including both the quantitative findings for each sub-criterion and the results of the qualitative analysis of the factors contributing to, or detracting from, the achievement of the sub-criteria. This section also includes evidence from global evaluations that were not included in the quantitative review.

In reporting on the factors, the report makes use of the terms "most", "many", "some" and "few" to describe the frequency with which an observation was noted, as a percentage of the number of evaluations for which the sub-criterion was covered (Box 3). In addition, for the most part, the order in which the factors are presented reflects the frequency with which they were mentioned.

Box 3: Frequency of Observations

Most = over three-quarters of the evaluations for which the sub-criterion was covered

Many = between half and three-quarters

Some = between one-fifth and half

Few = less than one-fifth

Table 3 summarizes the findings with respect to coverage (represented by the letter n), the team's assessment of whether this is strong, moderate or weak, and the ratings assigned by the review team of "satisfactory" or "unsatisfactory" for each of the six major criteria and their associated sub-criteria.

Table 3:  Coverage and Summary Results for Each Sub-criterionFootnote 31
Criteria and Sub-Criterian*Coverage Level**Satisfactory Ratings (%) ***Unsatisfactory Ratingss
(%)***
Relevance
1.1 Programs and projects are suited to the needs and/or priorities of the target group.60Strong90%10%
1.2 Projects and programs align with national humanitarian and development goals.61Strong98%2%
1.3 Effective partnerships with governments, bilateral and multilateral development and humanitarian organizations and Non-governmental organizations for planning, coordination and implementation of support to development and/or emergency preparedness, humanitarian relief and rehabilitation efforts.56Strong88%13%
Objectives Achievement
2.1 Programs and projects achieve their stated humanitarian and development objectives and attain expected results.61Strong77%23%
2.2 Programs and projects have resulted in positive benefits for target group members.62Strong85%15%
2.3 Programs and projects made differences for a substantial number of beneficiaries and where appropriate contributed to national humanitarian and development goals.61Strong72%28%
2.4 Programs contributed to significant changes in national humanitarian and development policies and programs (including for disaster preparedness, emergency response and rehabilitation) (policy impacts) and/or to needed system reforms.43Moderate77%23%
Cross-Cutting Themes – Inclusive Humanitarian and Development which is Sustainable
3.1   Extent to which multilateral organization supported activities effectively address the crosscutting issue of gender equality.42Moderate48%52%
3.2   Extent to which changes are environmentally sustainable.15Weak
Sustainability of Results/Benefits
4.1   Benefits continuing or likely to continue after project or program completion or there are effective measures to link the humanitarian relief operations, to rehabilitation, reconstructions and, eventually, to longer term humanitarian and development results.59Strong51%49%
4.2   Projects and programs are reported as sustainable in terms of institutional and/or community capacity.59Strong47%53%
4.3   Programming contributes to strengthening the enabling environment for humanitarian and development.40Strong70%30%
Efficiency
5.1 Program activities are evaluated as cost/resource efficient.48Strong63%38%
5.2 Implementation and objectives achieved on time (given the context, in the case of humanitarian programming).39Moderate51%49%
5.3 Systems and procedures for project/program implementation and follow up are efficient (including systems for engaging staff, procuring project inputs, disbursing payment, logistical arrangements etc.).35Moderate37%63%
Using Evaluation and Monitoring to Improve Humanitarian and Development Effectiveness
6.1 Systems and process for evaluation are effective.42Moderate64%36%
6.2 Systems and processes for monitoring and reporting on program results are effective.59Strong39%61%
6.3 Results based management systems are effective.24Weak
6.4 Evaluation is used to improve humanitarian and development effectiveness.61Strong57%43%

* n=number of evaluations addressing the given sub-criterion
** Strong: n=45 – 62; Moderate: n=30 – 44; Weak: n=less than 30
*** Satisfactory ratings includes "satisfactory" and "highly satisfactory"; unsatisfactory ratings includes "unsatisfactory" and "highly unsatisfactory"

3.1   Relevance

3.1.1   Coverage — Relevance

There is strong coverage of all sub-criteria with respect to relevance (Figure 3). Almost all evaluations covered the Sub-criterion 1.1 "Programs suited to the needs of the target group" and Sub-criterion 1.2 "Programs aligned with national development goals." Only slightly fewer evaluations address Sub-criterion 1.3 of "Effective partnerships."

Figure 3: Number of Evaluations Addressing Sub-criteria for Relevance

Figure 3 Text Alternative
Number of Evaluations Addressing Sub-criteria for Relevance
Sub-criterionNot addressedAddressed
1.1 Programs suited to the needs of target group260
1.2 Programs align with national development goals 161
1.3 Effective partnerships656

3.1.2 Key Findings — Relevance

The findings reflect that UNICEF's programming is highly relevant to the needs of the target groups (Figure 4). Ninety percent (90%) or more of the evaluations reported suitability to target group needs and alignment with national development goals as satisfactory or better. Eighty-eight percent (88%) of the evaluations reported partnerships as being satisfactory or better. Evaluations of programming in middle-income countries (MICs) are somewhat more likely to have satisfactory or better ratings for effective partnerships than those from least-developed countries (LDCs) (Annex 6). Although UNICEF's programming reflects a high level of relevance overall, the ratings for Sub-criterion 1.1 "Programs suited to the needs of the target group" and Sub-criterion 1.2 "Programs aligned with national development goals" for programming in the area of Child Protection tend to reflect somewhat lower ratings. This likely illustrates the extent to which child protection is not as well recognized, or as well reflected in national policies and priorities, as other sectors of UNICEF programming.

Figure 4: Findings for Relevance

Figure 4 Text Alternative
Findings for Relevance
Sub-criterion(4) Highly Satisfactory(3) Satisfactory(2) Unsatisfactory(1) Highly Unsatisfactory
1.3 Effective partnerships (n=56)30%57%7%5%
1.2 Alignment with national development goals (n=61)43%56%2%0%
1.1 Programs suited to needs and priorities of target group (n=60)32%58%5%5%

Humanitarian evaluations reflected a somewhat lower rating with respect to being suited to the needs of target groups, perhaps because of the challenges of conducting adequate needs assessments. One evaluation of a major humanitarian response in 2010 noted that there was "virtually no systematic needs assessment that might have informed subsequent cluster coordination and programme design, helped management monitor UNICEF‘s progress over time or kept key stakeholders abreast of developments with accurate, reliable information about the situation of women and children on the ground."Footnote 32 However, the same evaluation later notes that UNICEF's Core Commitments for Children (CCCs) in Humanitarian Action, adopted later in 2010, reflect a commitment to ensuring that assessment-based data is available in the future.Footnote 33

While less likely to be suited to the needs of the target groups, humanitarian evaluations were more likely to have a higher rating with respect to partnership, likely because of UNICEF's delivery of humanitarian action through non-governmental organizations (non-governmental organizations). Humanitarian evaluations were also as likely to be aligned with national goals as programming in other Medium-Term Strategic Plan focus areas. The global evaluation of UNICEF's education programming in emergencies highlighted the relevance of this programming, emphasizing the overlap between the countries in which UNICEF is programming and the lists of fragile and conflict-affected countries developed by other organizations. It notes that the "strong focus on capacity-building of government personnel and the heavy investment of funds and resources in delivery of learning materials supports the case for relevance at country and local levels."Footnote 34

Partnership was a sub-criterion for assessing the relevance of UNICEF-supported programming and the evaluations identified a wide range of partners involved in programming, including:

Box 4: Burden-sharing by UNICEF Partners

"[The campaign] was effective because of the establishment of social mobilization committees that included health and other governmental sectors [twelve ministries and institutions, including universities, training institutions etc.] ... non-governmental organizations and partners [nineteen national and international partners]. The contribution of the representatives of these sectors increased the strength of the social mobilization campaign. Different actors propose activities to be implemented during the campaign or make a material contribution to facilitate the campaign." [Translation by authors]

Evaluation of the Social Mobilization Activities and Communication for the Maternal and Child Health Week (2006 – 2011), 2012 p. 24

Specific examples of relevance from the evaluation reports include:

3.1.3   Contributing Factors — Relevance

A number of factors were identified as enhancing the relevance of UNICEF-supported programming.Footnote 35 Most evaluations reflect that UNICEF-supported programming was aligned with national priorities, as expressed in national plans, legal or development frameworks or national programs, policies and action plans. In addition to alignment with government priorities, UNICEF's programming was designed to support the implementation of public policies, programming and capacity building in national governments or institutions (Box 5). Most evaluations also noted the extent to which partnerships contribute to relevance.

Box 5: Alignment of UNICEF-supported programming

"The objectives of UNICEF's Education Programme are highly congruent with international priorities. UNICEF's education programming in Bangladesh reflects the principles of the Convention on the Rights of the Child  … [is] aligned with MDGs [Millennium Development Goals]  to achieve universal primary education (MDG [Millennium Development Goal]   2) and promote gender equality (MDG [Millennium Development Goal]   3) and article 10 of the Convention on the Elimination of Discrimination against Women. The programme reflects the aims of the Education for All movement and the Dakar Framework for Action…UNICEF's Education Programme is well aligned with national priorities as expressed in GoB [Government of Bangladesh]   policies, acts, and strategies. …UNICEF is very responsive to country needs, thanks to its close relationship with the government and its participatory approaches and grassroots consultations. The Education Programme is strongly aligned with UNICEF's corporate priorities for education. Its objectives … are fully aligned with the Medium-Term Strategic Plan 2006-2013 Keys Result Areas under Focus Area 2 (Basic Education and Gender Equality), and with specific aspects of the Key Result Areas under Focus Areas 1 (Young Child Survival and Development) and 3 (HIV-AIDS and Children)."

Evaluation of UNICEF Bangladesh Education and Child Protection Programmes: Final Report, December 2011, p. 33-34

Some evaluations noted that:

Although UNICEF-supported programming is considered highly relevant, there are a number of factors that detract from the relevance of the programming. These relate primarily to weaknesses in the same areas as the enhancing factors. Some evaluations noted gaps in UNICEF's partnerships, notably with national government (often because government is not taking adequate ownership) and other UN agencies (Box 6).

Box 6: Positive and Negative Aspects of Partnership and Coordination

"Working closely with the RC/HC [Resident Coordinator/Humanitarian Coordinator], UNICEF coordinated its response including needs assessments and external communication with the UNCT [UN Country Team]. In Georgia, coordination was influenced by additional factors, such as the arrival of many external non-governmental organizations and donor-led partners who had no previous experience of Georgia or previous relationships with existing humanitarian partners, and a government and UNCT that had had no previous experience of the new humanitarian coordinating mechanism namely the cluster approach. In addition there were several kinds of conflicting pressures by both donors and Government agencies to adopt certain approaches to relief and other interventions that were not all agreed on by the current actors.

In the WASH [water, sanitation and hygiene] sector ... the coordination was focused primarily on the operational aspects of provision of WASH [water, sanitation and hygiene] facilities in collective centres and tents. The cluster found it difficult to engage with government, donors and the shelter cluster to ensure that minimum standards in were followed in building settlements, with the result that many of the settlements which the government has now come up with do not have even the minimum habitable infrastructure on WASH [water, sanitation and hygiene] ."

UNICEF's Response to the Georgia Crisis: Real Time Evaluation, March 2009, p. 30-1

A few evaluations noted:

3.2   Objectives Achievement

3.2.1   Coverage – Objectives Achievement

There is generally strong coverage of sub-criteria relating to objectives achievement in the evaluations (Figure 5). This holds true for objectives achievement (2.1), benefits for target group members (2.2), and for program results reaching substantial numbers of beneficiaries (2.3). Only for the sub-criterion of positive changes in national policies and programs (2.4) does coverage decline to the moderate level, with forty-three evaluations including relevant findings.

The evaluations dealing with programming in Basic Education and Gender Equality (ten of eleven evaluations); Child Protection (five of six evaluations); and evaluations of Humanitarian programs (ten of thirteen evaluations) do tend to address UNICEF's success in influencing national policies and programs quite often. Only two of seven evaluations in the sample from the focus area Policy Advocacy and Partnership included findings on sub-criterion 2.4.

One explanation for only moderate levels of coverage of the extent of programs' influence on changes in national policies may be the close relationship between UNICEF and national policy makers in some countries: extending over a number of policy making cycles. At the time of the evaluations, there may have been little difference between UNICEF priorities and national policies, leaving little need for UNICEF to influence changes or to evaluate those changes.

3.2.2   Key Findings – Objectives Achievement

The review findings are largely positive (Figure 6). A full 77% of the evaluations reviewed reported satisfactory or better findings regarding the extent that UNICEF supported programs achieve their stated objectives (2.1). This does not mean that UNICEF programs were found to achieve 77% of their objectives across the board. Rather, it means that UNICEF programs achieved better than half their objectives (including the most important ones) in 77% of the evaluations reviewed.
Results were even stronger for the programs' ability to provide positive benefits for the target population (2.2), with 86% of evaluations reporting satisfactory or better findings. They were somewhat less positive for the scale of program benefits measured in terms of the number of program beneficiaries (2.3), with 72% of evaluations reporting satisfactory or better results.

Finally, programs were rated satisfactory or better for their ability to support positive changes in national policies and programs (2.4) in 77% of the evaluations reporting on this sub-criterion.

Figure 5: Number of Evaluations Addressing Sub-criteria for Objectives Achievement

Figure 5 Text Alternative
Number of Evaluations Addressing Sub-criteria for Objectives Achievement
Sub-criterionNot addressedAddressed
2.1 Programs and projects achieve stated objectives161
2.2 Positive benefits for target group members 062
2.3 Differences for a substantial numbers of beneficiaries/contributed to national development goals161
2.4 Significant changes in national development policies and programs1943

Figure 6: Findings for Objectives Achievement

Figure 6 Text Alternative
Findings for Objectives Achievement
Sub-criterion(4) Highly Satisfactory(3) Satisfactory(2) Unsatisfactory(1) Highly Unsatisfactory
2.4 Changes to national policies/programs (n=43)37%40%12%12%
2.3 Substantial numbers of beneficiaries (n=61)31%41%23%5%
2.2 Positive benefits for target group members (n=62)21%65%11%3%
2.1 Programs achieve stated objectives (n=61)16%61%16%7%

Of the four sub-criteria relating to objectives achievement, the last three showed no notable difference in the pattern of findings when evaluations of programs in Least Developed Countries were compared to those of programming in Middle Income Countries.

On the other hand, for the key sub-criterion of whether programs achieve their stated objectives (2.1), evaluations of programs in Middle Income Countries reported satisfactory or better findings somewhat more often than for programs in Least Developed Countries (Annex 6).

This does not seem counterintuitive since the program implementation capacity of UNICEF's partner agencies in Middle Income Countries might in general be higher than in Least Developed Countries. The pattern of more positive performance for programs in Middle Income Countries does not hold, however, for all the criteria, as can be seen from the results reported for gender equality (Section 3.3).

Types of Objectives Achieved by UNICEF-Supported Programs

It is interesting to consider what types of objectives are being achieved by UNICEF-supported programs according to the evaluations. It is clear from the pattern of programming represented by UNICEF's Medium-Term Strategic Plan focus areas and the sub-sectors which they encompass, that a wide variety of program outputs and outcomes are included in the objectives of UNICEF programming.

Providing an exhaustive portrait of the positive achievements of UNICEF-supported programs and how they contribute to the goals of the Medium-Term Strategic Plan is beyond the scope of this review. Instead, this section presents a small sampling of the specific achievements noted in the evaluations. These objectives were achieved in partnership with governments, non-governmental organizations and local community members. The evaluations do not attempt to attribute their achievement solely to UNICEF support. The achievements reported in the evaluations include:

Box 7: Improvements in Police Services in Indonesia

"By 2007, the creation of children's desks in every district in Indonesia was agreed to by the national police. A component on child-friendly procedures was also incorporated into the police training curriculum. By 2008, 22 children's desks had been re-established in all districts of NAD; one child courtroom established in Banda Aceh district; instructions on diversion [of youth to the restorative justice system] adopted by the police; standard procedures and guidelines for restorative justice for police officers developed; and a case management database was under development."

 2009: Children and the 2004 Indian Ocean Tsunami: UNICEF's response in Indonesia (2005-2008: Child Protection, p. 21-22

Box 8: Contributing to Improved Sanitation and Hygiene in Ethiopia

"The survey results showed that 93% of the households interviewed had access to an improved water source which is above the stated level of achievement while access to a latrine was 74% which is slightly less…The improvement in service levels achieved in the target woredas [administrative districts in Ethiopia managed by local governments] can be regarded as one of the measures of impact of the project. Additionally, woreda staff report a genuine change in attitude and commitment of the community as a demonstrable impact of the project. The role of water, sanitation and hygiene [WASH] committees in ensuring sustained operation of water schemes and monitoring of the sanitation condition of their villages was also considered important. In general there was considerable anecdotal support behind the project's contribution to minimize community susceptibility to diseases emanating from poor hygiene and sanitation conditions."

 Ethiopia: Mid-Term Evaluation of EU/UNICEF Supported WASH Programme, 2010, p. 21 and 32

3.2.3   Contributing Factors – Objectives Achievement

Factors which made a positive contribution to the achievement of program objectives are presented in this sub-section.

Many evaluations pointed to the following factors:

Some evaluations pointed to the following positive factors:

Box 9: Strengthened Programming through Participatory Processes: United Nations Girls Education Initiative in Uganda

"Indeed, UNGEI Uganda provides a model for multi-stakeholder ownership and management of girls' education, whereby the school community has been firmly placed "in the driver's seat". The school communities are directly consulted on matters relating to their priority needs and perceptions about the quality of education service delivery. The findings from such consultative efforts are factored into the work plans and are used to inform resource allocation decision-making. As a result, there is now increased prioritization of school community-specific interventions and raised grassroots' consciousness."

Formative Evaluation of the UNGEI Country Report: Uganda, 2011. p. 66

Global evaluations echoed some of the same factors for success in securing positive program results and achieving objectives. The synthesis report on Protecting Children from Violence, for example, highlighted advocacy and the incorporation of child protection into national and decentralized planning processes as an essential tool for strengthening child protection systems.Footnote 36

The evaluations also provided a diagnosis of some factors which negatively affect the ability of UNICEF-supported programs to achieve their stated objectives.
Many evaluation reports referred to weakness in program design as the crucial factor inhibiting program success. These included lack of a results orientation, absence of linkages or a total absence of a causal chain, over-ambitious program goals, poorly targeted programming, failure to address the rights of target group members, an overly fragmented approach to program delivery, and poor technical design of program inputs. Poor program design was more often noted in evaluations of programming in Least Developed Countries than in Middle Income Countries.

Some evaluations pointed out that program budgets were inadequate and that outputs were insufficient to address the identified problem and to reach the coverage targets set for the program.

A few evaluations highlighted the following negative factors influencing objectives achievement.

3.3   Cross-Cutting Themes

3.3.1   Coverage – Cross-cutting Themes

Compared to Relevance and Objectives Achievement, the sub-criteria relating to Crosscutting Themes present a very different picture in terms of coverage in the UNICEF evaluations (Figure 7). The sub-criterion on the extent to which UNICEF supported activities effectively address gender equality received a moderate level of coverage with forty-two of the sixty-two evaluations containing relevant findings. Coverage of sub-criterion on the environmental sustainability of changes resulting from UNICEF supported programs was weak. It was only addressed in fifteen evaluation reports.

Figure 7: Number of Evaluations Addressing Sub-criteria for Crosscutting Themes

Figure 7 Text Alternative
Number of Evaluations Addressing Sub-criteria for Crosscutting Themes
Sub-criterionNot addressedAddressed
3.1 Activities effectively address gender equality2042
3.2 Changes are environmentally sustainable4715

Gender Equality

It is remarkable that almost one-third of the evaluations in the sample did not address the question of gender equality. It is important to note that in order for the review to rate this sub-criterion, evaluations had to identify the extent to which the program or project in question had incorporated gender equality objectives and whether they had been achieved. Gender equality could also be addressed in an evaluation report if it assessed the program's success in mainstreaming gender equality. On the other hand, if an evaluation simply reported the portion of girls or women receiving benefits compared to boys or men, not addressed was the rating provided for this criteria because there was no reference to the program's success in addressing gender equality.

It is also worthwhile considering which evaluation reports in the sample tended to address gender equality more often than others. Evaluations of UNICEF support to programs in the Medium-Term Strategic Plan focus area of Basic Education and Gender Equality tended to address gender equality with ten of eleven reports covering the issue. Similarly, eleven of thirteen evaluations of humanitarian action adequately covered gender equality.

Lower levels of coverage were found in evaluations of programming in the Young Child Survival and Development focus area with only thirteen of twenty-one evaluations addressing the subject adequately. This may reflect the operational nature of much programming in this area, which includes support to water, sanitation and hygiene programs. On the other hand, programs in child health could conceivably often face issues of differential access for boys and girls.

Finally, some evaluation reports pointed out that UNICEF and partner staff felt that gender equality was implied in programs which implemented a results-based approach or which focused on child protection and/or child rights.

The recent global evaluation of the application of the Human Rights Based-Approach to programming notes that UNICEF responded to an earlier global evaluation of its gender equality policy by re-emphasizing its organizational commitment:

The evaluation found that the strategies advocated by the original policy [mainstreaming and gender-specific programming activities] remain sound, and these have been retained. However, it indicated that the policy required updating to respond to new program priorities. Such priorities included the commitment to work more explicitly with men and boys as both agents and beneficiaries of gender equality, and to improve the priority and resourcing given to gender-equality programming by the organization and the grounding of its actions in Convention on the Elimination of All Forms of Discrimination Against Women together with the Convention on the Rights of the Child.Footnote 37

Given the strength of UNICEF's strategic commitment to gender equality and its re-emphasis in the 2010 policy,Footnote 38 it is surprising that one-third of the evaluations reviewed fail to address gender equality.

Environmental Sustainability

Given the very weak level of coverage of environmental sustainability, with just fifteen evaluations addressing the sub-criteria, the review is not able to report on the effectiveness of UNICEF-supported programs in this area.

It is worth noting that evaluations of Young Child Survival and Development programming do sometimes address environmental sustainability (eight of twenty-one evaluations). This is largely due to the presence of water, sanitation and hygiene programs in this focus area with consequent attention paid to the environmental effects of investments in water and sanitation infrastructure.

Similarly, evaluations of humanitarian action also sometimes addressed environmental sustainability (six of thirteen evaluations). This is not surprising since emergency programs often include a component focused on the rehabilitation of water, sanitation, education and health infrastructure.

The Programme Division of UNICEF reports that programmatic attention to climate change adaptation grew in 2012, especially in relation to water, sanitation and hygiene programs. Given this trend and the activation (in 2012) of an inter-divisional working group on climate change adaptation, it would be reasonable to expect increased attention to issues of environmental sustainability in evaluations carried out by UNICEF in the future.

3.3.2   Key Findings in Gender Equality

The evaluations present a challenging picture of the effectiveness of UNICEF-supported programs in addressing gender equality. Less than half (47%) of the forty-two evaluations that addressed gender equality contained findings of satisfactory or better for this sub-criterion (Figure 8).

Considering that twenty-two evaluations did not address gender equality, another way of expressing the result would be to note that only one-third (32%) of the sixty-two evaluations reviewed were able to demonstrate that UNICEF-supported programs effectively address gender equality.

Figure 8: Findings for Effectiveness in Supporting Gender Equality

Figure 8 Text Alternative
Findings for Effectiveness in Supporting Gender Equality
Sub-criterion(4) Highly Satisfactory(3) Satisfactory(2) Unsatisfactory(1) Highly Unsatisfactory
3.1 Effectively Address gender equality (n=42)7%40%38%14%

Surprisingly perhaps, results for effectively supporting gender equality were somewhat more positive when reported in evaluations of programs in Least Developed Countries than those in Middle Income Countries (Annex 6).

At the same time, results for Sub-criterion 3.1 were notably more negative for humanitarian action than for the samples of evaluations taken as a whole (Annex 6). Therefore, while evaluations of humanitarian action were more likely to address gender equality, they reported more negative findings than other evaluations.

However, the evaluations point to some important successes for UNICEF in supporting programs that effectively address gender equality. Where programs are effective, they have been reported to achieve various types of successful results in gender equality. A few examples include:

Box 10: Gender Sensitisation and People Friendly Police (GSPP) Initiative in Karnataka

"Definite changes were seen in the way women and children are treated at all the police stations visited. This was also seen in instances where we could see women interacting with the police at the station. This was also echoed through our discussions with non-governmental organizations and community members….Over and above all these, the GSPP training was seen as an important factor for bringing about this change. This was expressed during discussions and in the questionnaire survey."

Evaluation of the Gender Sensitization and People Friendly Police Initiative: Karnataka, 2011 p. 45

The global evaluations reviewed qualitatively also point to a variety of positive results in gender equality for UNICEF supported programmes. The global evaluation of Child Friendly Schools Programming, for example, found that:

Students mostly feel that their schools provide female and male students with equal access to opportunities: about three quarters or more in each country responded ‘mostly true' or ‘very true' to the statement ‘Both boys and girls have equal opportunities to succeed at this school'.Footnote 39

It is worth noting that UNICEF has recognized issues relating to gender equality and has been engaged in efforts to improve effectiveness in the area. In response to the 2008 Gender Evaluation, the agency launched the Strategic Priority Action Plan for Gender Equality Action Plan 2010-2012 with eight areas of organizational action identified. UNICEF also reports its intent to take visible action on gender equality in the new strategic plan.

3.3.3   Contributing Factors – Gender

Given the generally negative findings for gender equality, it is not surprising that reviewers noted more negative than positive contributing factors. Nonetheless, the evaluations discussed and presented a number of contributing factors.

The factor identified most often as contributing to positive results in gender equality was the inclusion of gender specific objectives and targets, including a focus on women's rights and (for a few evaluations) effective mainstreaming of gender equality beyond a single thematic area (such as Child Protection).

Some evaluations pointed to the active promotion of women as leaders and as members of water and other community management committees. They also pointed to the promotion of women to key national policy making positions.

A few evaluations noted each of the following as factors contributing to positive results in gender equality;

The reviewed evaluations identified a wide range of factors inhibiting achievements in gender equality.

Most evaluations highlighted the fact that programs lacked any specific objectives regarding gender equality and/or they simply lacked any gender perspective in their design and implementation. For a few evaluations, this took the form of treating gender as a completely "neutral" issue in program design, with no distinction made between boys and girls or men and women.

Some evaluations pointed to:

A few evaluations identified negative factors of:

Box 11: Gender-Based Violence and the Response to the Earthquake in Haiti

"...in the early months of the response, UNICEF had only one gender-based violence focal point. (It took a month for the organization to deploy a gender-based violence specialist), and to this day it remains unclear where the issue sits within child protection. This was despite the emergence of rape in the camps as an issue before the end of January 2010….It was reported by UNICEF surge staff that it had taken up to four months to install lighting for toilet areas in some camps using UNFPA materials – a key security precaution that might have helped combat sexual violence.

Independent Review of UNICEF's Operational Response to the January 2010 Earthquake in Haiti, 2011, p. 15

Some global evaluations identified similar issues for UNICEF in effectively addressing gender equality issues. The synthesis report on Protecting Children from Violence, for example, noted a number of challenges, constraints and gaps faced by UNICEF. The first two mentioned are consistent with the factors reported above:

Gaps: In some programmes, evaluators noted that gender issues are not addressed adequately in planning and implementation (Bosnia Herzegovina, Thailand). In some evaluations concerns were raised that programmes were not using their programmes to raise awareness around gender and/or stimulate discussion in this regard (India).

Knowledge Management: Failure to adequately conduct an in-depth analysis of gender issues at the outset, hindered programmes from effectively enduring gender equity (Nepal, Maldives).Footnote 40

The global evaluation of Human-Rights Based Approach to UNICEF programming examines the human-rights based approach under five key principles. While all five are relevant to gender-equality, perhaps the most central is the principle of non-discrimination. The global evaluation also identifies data disaggregation and weaknesses in reporting as factors limiting the application of the principle of non-discrimination in programming by UNICEF country offices:

The operationalization of this principle in UNICEF's work rests to a large extent on the degree to which data disaggregation is sufficiently carried out across all levels of vulnerability so as to help inform the development of targeted or universal programs.

The scoring on non-discrimination across the Country Office Assessment was ‘satisfactory' to ‘weak', suggesting that it is applied with mixed results. More specifically, the Country Office Assessment rated more than one third of COs (15) as ‘weak' on this principle. The SitAn [Situation Analysis] often reveals who is being excluded in society, who the marginalized groups are, etc. However, SitAns [Situation Analyses] often use generic descriptions for identifying the most vulnerable, and programming documents and annual reports do not provide sufficiently detailed information about the specific efforts Country Offices make to ensure they reach the most vulnerable/marginalized. Moreover, in many cases, documents did not provide sufficient detail to determine whether the most vulnerable have been reached.Footnote 41

3.4 Sustainability

3.4.1 Coverage - Sustainability

The coverage for two of the three sub-criteria with respect to sustainability was strong (Figure 9). Both sub-criterion 4.1 "Benefits continuing after program completion" and sub-criterion 4.2 "Institutional/community capacity for sustainability" were covered in fifty-nine evaluations. The coverage was moderate for Sub-criterion 4.3 "Strengthened enabling environment for development", as this sub-criterion was addressed in forty evaluations.

Figure 9: Number of Evaluations Addressing Sub-criteria for Sustainability

Figure 9 Text Alternative
Number of Evaluations Addressing Sub-criteria for Sustainability
Sub-criterionNot addressedAddressed
4.1 Benefits continuing or likely to continue359
4.2 Programs support sustainability through Institutional and/or community capacity359
4.3 Programs strengthen enabling environment for development 2240

3.4.2 Key Findings - Sustainability

Overall, the results with respect to sustainability are mixed (Figure 10). About half the evaluations reflected ratings of satisfactory or better on two sub-criteria – sub-criterion 4.1 "Benefits continuing or likely to continue" (51%) and sub-criterion 4.2 "Programs support sustainability through institutional and/or community capacity" (48%). Evaluations of programming in Middle Income Countries appear to be somewhat more likely to get satisfactory or better ratings for sustainability through institutional and/or community capacity than those from Least Developed Countries (Annex 6). The evaluation reports were more positive for sub-criterion 4.3 "Programs strengthen enabling environment for development", with 71% reflecting findings of satisfactory or better.

Figure 10: Findings for Sustainability

Figure 10 Text Alternative
Findings for Sustainability
Sub-criterion(4) Highly Satisfactory(3) Satisfactory(2) Unsatisfactory(1) Highly Unsatisfactory
4.3 Enabling environment for development (n=40)8%63%20%10%
4.2 Institutional and community capacity (n=59)12%36%36%17%
4.1 Benefits continue after program completion (n=59)7%44%37%12%

Activities often associated with strengthening the environment for sustainability included:

Box 12: Advocacy and Policy Dialogue with Government

Having established a strong advocacy role with the GoSL [Government of Sri Lanka], UNICEF now has the potential to influence policy changes in water quality surveillance and water supply subsidies [in the water, sanitation and hygiene sector]. ... [In child protection], UNICEF has successfully lobbied for changes in practice and attitudes away from institutions and towards safe placement of children with legal guardians. ... Attributing change to any one agency would be dishonest, but UNICEF sizeable interventions over four years, and the encouraging manner in which it has assimilated lessons from these interventions, have given it a unique opportunity to guide and influence national and sub-national government.

Children and the 2004 Indian Ocean Tsunami: Evaluation of UNICEF's Response in Sri Lanka (2005-2008), p. 57-8

Specific examples from the evaluations include:

It was noted that the results for sub-criterion 4.3 are more likely to be positive than those for other sub-criteria. It is suggested that this is because efforts to improve the enabling environment do not, in themselves, have to be on-going activities. For example, support for the development of national legislation or policies is a time-limited activity, which would be expected to have a long-term impact on the sector. As a result, it may be easier to achieve satisfactory findings for this sub-criterion than for other sub-criteria, which require evidence of the long-term impacts. This is reflected in some of the achievements with respect to strengthening the enabling environment:

3.4.3 Contributing Factors — Sustainability

A number of factors identified in the evaluations were linked to enhancing the sustainability of UNICEF programming. Some evaluations reflected:

A few evaluations noted that:

The major factors detracting from sustainability were, for the most part, the failure to achieve the factors that enhance sustainability. Some evaluations noted:

Box 13: Lack of strategy for sustainability

"There was insufficient explicit attention to sustainability of results in the CPC 5 and CPC 6 [UNICEF Country Program for Children 5 and 6] programs. While capacity building was central to both programs, the absence of a common understanding, clear strategy, and systematic approach to capacity building within the UNICEF Philippines Country Office (PCO) contributed to its mixed performance in supporting sustainable capacity building results."

 UNICEF Philippines Country Program Evaluation: Final Report, 2012 p. iii

In addition, a few evaluations noted that, in spite of efforts at capacity building, a challenge to sustainability lies in the loss of previously developed capacity – for example, the faltering of stakeholder engagement, the disbanding of community committees and the dropout of trained resources.

A few evaluations of humanitarian action reflected that weaknesses in program design detracted from sustainability because of the limited linking of relief and development (Box 14). Linking relief and development is reportedly being addressed by UNICEF in its upcoming, new strategic plan, which will highlight UNICEF's commitments to building resilience and addressing vulnerability and fragility through its development programming.Footnote 42

Box 14: Lack of linking of relief and development

"In the health sector, a successful emergency phase was not matched by strategic multi-year planning in the recovery phase. More attention should have been paid to the development of district- and provincial-level capacity for planning and administration. Structural weaknesses in government, apparent before the tsunami, persist; UNICEF's preoccupation with bricks and mortar has prevented appropriate attention to this. Nevertheless, UNICEF has helped to build institutional capacity through training of health care workers, social workers and volunteers…"

 Children and the 2004 Indian Ocean Tsunami: Evaluation of UNICEF's Response in Indonesia (2005-2008) Country Synthesis Report, 2009, p. 63

3.5   Efficiency

3.5.1   Coverage — Efficiency

Coverage levels for the three sub-criteria of efficiency range from strong for sub-criterion 5.1 on the cost and resource efficiency of programming, to moderate for sub-criteria 5.2 and 5.3 (Figure 11). The latter two deal with the timeliness of program implementation and the effect of UNICEF's administrative systems and procedures on the efficiency of program delivery.

It is important to point out that the fairly strong level of coverage of cost efficiency (5.1) does not mean that UNICEF programs are regularly subjected to cost effectiveness measurement in the evaluations. In fact, twenty-eight evaluations in the sample did not include any information on the overall cost of the program.Footnote 43 What is meant by coverage of cost and resource efficiency is that the evaluations made some judgment on program efforts to control the resource requirements and/or the unit cost of program inputs or outputs. Nonetheless, the fact is that forty-eight of sixty-two evaluations in the sample (77%) made some effort to assess the cost efficiency of the programs under review.

Figure 11: Number of Evaluations Addressing Sub-criteria for Efficiency

Figure 11 Text Alternative
Number of Evaluations Addressing Sub-criteria for Efficiency
Sub-criterionNot addressedAddressed
5.1 Program activities are efficient1448
5.2 Implementation and objectives achieved on time2349
5.3 Systems and procedures are efficient 2735

The moderate level of coverage of sub-criteria on timeliness (5.2) and administrative systems and procedures (5.3) is important. The review team suspects that these two criteria tend not to be addressed routinely or be included systematically in evaluation terms of reference. They may be more likely to be included in evaluation reports when problems in timely delivery of program inputs or problems with administrative systems and procedures are encountered during field level evaluations. The information may then be presented to explain limitations in the achievement of program objectives.

For sub-criterion 5.1 on cost and resource efficiency there is no particular pattern in terms of which Medium-Term Strategic Plan focus areas tend to address the issue. Evaluations in all focus areas (and humanitarian program evaluations) examine cost efficiency in about three-quarters of evaluations in the sample.

The pattern is similar for timeliness of program delivery (5.2) with the exception that only one of six evaluations in Child Protection addressed this sub-criterion. For sub-Criterion 5.3 on the efficiency of implementation systems and procedures, coverage levels are low for all focus areas except for Young Child Survival and Development (addressed in 14 of 21 evaluations) and for humanitarian action (addressed in nine of 13 evaluations).

For both these two focus areas, on-time delivery and installation of elements of infrastructure are often crucial to achieving program objectives, with the result that evaluations often report findings relating to the effectiveness of program implementation systems, particularly systems of procurement and supply.

3.5.2   Key Findings — Efficiency

The findings for efficiency are mixed and require careful interpretation. The fact that 62% of evaluations reported satisfactory or better findings on the cost efficiency of programs represents a reasonably positive result for sub-criterion 5.1 (Figure 12). Also notable is the fact that no evaluations reported highly unsatisfactory results on cost efficiency.

As already noted, there may be a tendency to under-report findings with respect to timeliness and the efficiency of administrative systems when no problems are encountered at field level. Nonetheless, it is a concern that 49% of the evaluations that addressed timeliness reported findings of unsatisfactory or worse. Also worrying is the fact that 18% reported a finding of highly unsatisfactory.
Similarly, the findings for sub-criterion 5.3 on the efficiency of administrative systems and procedures are largely negative, with 63% unsatisfactory or worse, and 17% highly unsatisfactory.

Figure 12: Findings on Efficiency

Figure 12 Text Alternative
Findings on Efficiency
Sub-criterion(4) Highly Satisfactory(3) Satisfactory(2) Unsatisfactory(1) Highly Unsatisfactory
5.3 Systems for program implementation and follow up efficient (n=35)3%34%46%17%
5.2 Implementation and objectives achieved on time (n=39)5%46%31%18%
5.1 Program activities cost/resource efficient (n=48)8%54%38%0%

Interestingly, this is an area where results for evaluations of programs in Least Developed Countries are somewhat more positive than for Middle Income Countries. In particular, results on the timeliness of program implementation are notably more positive for programs in Least Developed Countries (Annex 6).

The evaluation reports contain some interesting examples of the ways UNICEF-supported programs have tried to reduce unit program costs and increase the efficiency of resource use by:

3.5.3   Contributing Factors — Efficiency

Because very different factors contribute to results regarding cost efficiency, timeliness and the efficiency of administrative systems are discussed separately in this section.

Cost/Resource Efficiency

Positive factors promoting the cost efficiency of programs included the following:

A few evaluations pointed to:

Box 15: Open bidding for well drilling in Ethiopia

"Among donor financed WASH [water, sanitation and hygiene] programs however, it is the UNICEF program alone that allows public enterprises, non-governmental organizations and private contractors to participate equally in the bidding process. Particularly in the SNNPR [Southern Nations and Nationalities' People's Region] region, church-based organizations and other non-governmental organizations are awarded contracts to drill UNICEF financed wells and the regional water resource bureau also possesses [a] rig donated by UNICEF which is mainly used to drill UNICEF financed wells. This has contributed to keeping the cost of drilling down."

Mid-Term Evaluation of EU/UNICEF Supported WASH Program, Ethiopia, 2010. p. 14

Negative factors contributing to reduced cost and resource efficiency for UNICEF-supported programs were also identified in the review of evaluations:

A very diverse set of other negative factors for cost efficiency were identified and reported in a few evaluations. They include:

Box 16: Complex and Expensive Sanitation Systems for the Maldives

"The potential cost of running these systems for the islands is high, the human and institutional capacity to manage the complex systems is low, and key policy frameworks are still in development. The sanitation systems built as a response to the tsunami are complex engineering feats that are over-designed and expensive. This new method necessitates cooperative management and financial recovery, which is not yet available on the islands."

Children and the 2004 Indian Ocean Tsunami: Evaluation of UNICEF's Response in Maldives: Country Study, 2009, p. 20

Timeliness of Program Implementation and
Effective Systems and Procedures for Program Implementation and Follow-Up

There was a very close relationship between the timeliness of UNICEF-supported programs and the efficiency of UNICEF systems and procedures for program implementation as reported in the evaluations. For that reason, the factors that supported or impeded positive results for these two sub-criteria tended to be identical. They are presented together in this section.

Positive factors for sub-criteria 5.2 and 5.3 reported in a few evaluations include:

Evaluations also identified negative factors that tended to limit the timeliness of UNICEF-supported programs, often related to administrative systems and procedures. Some evaluations pointed to:

Box 17: After a Rapid Start, Delays in Emergency Response in Yemen

"Despite its general lack of preparedness for emergencies and its dominant development orientation, the UNICEF Country Office in Yemen acted relatively quickly in 2009 to mobilise existing staff and supplies for the emergency response. A loan of two million US$ from UNICEF's internal emergency financing mechanism, the Emergency Programme Fund (EPF) in September 2009, enabled this rapid response. Subsequently, however, it experienced often significant delays in procuring or shipping additional supplies or implementing activities that are mainly due to the fact that the Country Office is relying on its development structures and processes which grant little autonomy to local staff and have slow procurement and supply procedures (cf. section 3.6). These delays undermine UNICEF's credibility in the eyes of its partners and beneficiaries. As a result, beneficiaries appear less open to accepting important "soft" interventions by UNICEF and its implementing partners, such as hygiene education or awareness raising activities on child protection issues."

Real-Time Evaluation of UNICEF's Response to the Sa'ada Conflict in Northern Yemen, 2010, p.23

A few evaluations identified:

The global evaluations of humanitarian action also identified similar strengths and challenges with UNICEF's internal procedures (administration, finance and procurement). The evaluation of UNICEF's education in emergencies programming noted that the extent to which the evaluation could assess efficiency was limited because the program had failed to produce significant outputs and/or because of a lack of data. However, it did find that an efficient use of resources was reflected in the ability to leverage additional resources for the humanitarian response. On the other hand, there were delays in disbursement of funds in the first two years, which led to the reallocation of unspent funds from slow-performing country programs to those that were performing better.Footnote 44

Similarly, the recent real-time assessment of interventions in the Sahel did not include information on cost efficiency but did note that there were delays in both the deployment of staff and funding. The delays in funding are attributed in the evaluation to the lack of a clear strategy for integrating other sectors (for example, water, sanitation and hygiene, education, child protection) into the nutrition response in the Sahel.Footnote 45 The evaluation of the DFID-UNICEF program to strengthen UNICEF's humanitarian capacity reflected positives findings with respect to the efficiency of the joint program. However, it also noted that "Attempts have been made from Headquarters to streamline and clarify finance and administrative systems for emergencies. However, at the operational level, these are yet to make any significant difference and delays in funds release, Programme Cooperation Agreements and supplies remain common, except in a few countries."Footnote 46 UNICEF staff report that these issues are currently being addressed through revisions to standard operating procedures for the different levels of emergencies. A number of other recent changes were outlined in the 2012 UNICEF Humanitarian Action for Children.Footnote 47

3.6   Monitoring and Evaluation

3.6.1 Coverage — Monitoring and Evaluation

Coverage for two of the four sub-criteria for monitoring and evaluation is strong – sub-criterion 6.2 "Systems and processes for results monitoring and reporting are effective" (fifty-nine evaluations) and sub-criterion 6.4 "Evaluation used to improve effectiveness" (sixty-one evaluations) (Figure 13). Coverage for sub-criterion 6.1 "Systems and processes for evaluation are effective" is moderate. However, coverage for sub-criterion 6.3 "Results-based management   systems are effective" is weak and, as such, the findings will not be included in this report. In assessing this sub-criterion, the reviewers were looking for evidence of a focus on results and the integration of results monitoring into management decision-making. This emphasis on a quite literal interpretation of the term results-based management was developed so that the analysts could readily distinguish between findings on results monitoring and those relating to a more fully developed system for managing by results. This distinction may have resulted in a finding of "not addressed" for some evaluations, even if some elements of an results-based management system were addressed in the evaluation report.

This was not found in sufficient evaluations, in spite of the emphasis in UNICEF on results-based programming.

Figure 13: Number of Evaluations Addressing Sub-criteria for Using Evaluation and Monitoring

Figure 13 Text Alternative
Number of Evaluations Addressing Sub-criteria for Using Evaluation and Monitoring
Sub-criterionNot addressedAddressed
6.1 Systems and process for evaluation are effective2042
6.2 Systems and processes for results monitoring and reporting are effective359
6.3 RBM systems are effective 3824
6.4 Evaluation used to improve effectiveness161

3.6.2 Key Findings — Monitoring and Evaluation

The results for two sub-criteria are somewhat positive (Figure 14). The finding for sub-criteria 6.1 "Systems and processes for evaluation are effective" and sub-criterion 6.4 "Evaluation used to improve effectiveness" indicate that more than half the evaluations were rated as satisfactory or better, 64% and 57%, respectively. However, the findings for sub-criterion 6.2 "Systems and processes for results monitoring and reporting are effective" are considerably less positive. Only 39% of the evaluations reflected a rating of satisfactory or better for this sub-criterion. As will be seen below, this is one of the factors detracting from the use of monitoring and evaluation.

As noted in Section 1.6, since 2011, UNICEF has made a significant effort to strengthen results definition, monitoring and reporting through the development and implementation of Monitoring Results for Equity Systems (MoRES). The effect of this system on the strength of UNICEF monitoring will be clearer on completion of the planned formative evaluation of Monitoring Results for Equity Systems (MoRES) in 2013.

Figure 14: Findings for Using Evaluation and Monitoring

Figure 14 Text Alternative
Findings for Using Evaluation and Monitoring
Sub-criterion(4) Highly Satisfactory(3) Satisfactory(2) Unsatisfactory(1) Highly Unsatisfactory
6.4 Use of evaluation to improve effectiveness (n= 61)36%21%5%38%
6.3 RBM systems are effective (n=24)8%25%50%17%
6.2 Monitoring and reporting on program results effective (n=59)12%27%49%12%
6.1 Systems and process for evaluation are effective (n=42)19%45%31%5%

Effective Systems and Processes for Evaluation and Effective Systems for Monitoring and Reporting

The findings for Sub-criterion 6.2 "Systems and processes for results monitoring and reporting are effective" tended to be somewhat more positive for programming in the Medium-Term Strategic Plan focus area of Young Child Survival and Development. This is perhaps explained by the fact that many of these activities tend to be more operational and are long-standing UNICEF activities, which are easier to monitor than other areas. On the other hand, the findings for Sub-criterion 6.1 "Systems and processes for evaluation are effective" for Child Protection tend to be less positive than for other focus areas. This is possibly a reflection of the challenges of monitoring and evaluation in an area that is less operational.

The evaluations identify a range of UNICEF's monitoring and evaluation tools including baseline studies, mid-term or formative evaluations, regular evaluations and other evaluative-type studies, such as costing studies and participation in lessons learned exercises and joint evaluations. While some evaluations noted the positive achievements with respect to monitoring and evaluation (for example, tools for monitoring and evaluation were well established and regularly implemented and good quality reports are being produced) (Box 18), others reported weaknesses, including:

Box 18: Evaluation and Implementation of Lessons Learned

At the end of 2005, UNICEF conducted a major evaluation of the emergency response and initial phase (first six months after the tsunami) for Sri Lanka, Indonesia and Maldives.31. UNICEF also participated actively in the Tsunami Evaluation Coalition (TEC), which produced a series of evaluations and reports covering thematic topics, including: coordination; needs assessment; the impact of the international response on local and national capacities; links between relief, rehabilitation and development; and the funding response. In addition to this, there were regional consultations and ‘lessons learned' exercises that captured some of the key findings.32. The recommendations and lessons from these evaluations have influenced adjustments in programme design and management, as well as the formulation of UNICEF‘s emergency/early recovery response policies and capacities. More recently (end-2008), a follow-up Linking Relief Rehabilitation and Development (LRRD) study was undertaken by the Tsunami Evaluation Coalition.

Children and the 2004 Indian Ocean Tsunami: Evaluation of UNICEF's Response in Sri Lanka (2005-2008), November 2009, p. 5

Similar findings in: Children and the 2004 Indian Ocean Tsunami: Evaluation of UNICEF's Response in Indonesia (2005-2008): Country Synthesis Report, November 2009, p. 7

Specific examples from the evaluations include:

A few evaluations noted specifically that these weaknesses meant that UNICEF could not measure the impact of its programming and that there was a tendency to focus on short-term accomplishments. This was noted in the global formative evaluation of the United Nations Girls' Education Initiative: "Assessing the partnership's impact in its outcome areas is challenging, in part because of the relatively short time frame for many of the initiatives, but also because of the absence of systems to track the impact, or the absence of clear targets for what the various initiatives are meant to accomplish."Footnote 48

Use of Evaluations to Improve Effectiveness

The results for Sub-criterion 6.4 reflect the extent to which UNICEF is preparing management responses for its evaluations and whether these include adequate responses to the recommendations, including an action plan and clear responsibility for its implementation. In 2008, the Executive Director was required by the Executive Board to ensure the preparation and availability of management responses for all evaluation reportsFootnote 49 and this requirement has been implemented gradually over the past few years. As a result, evaluations reflecting a satisfactory or better rating for this sub-criterion are much more likely to be found for evaluations published in 2010 and 2011, than in 2009. The data also suggest that adequate management responses are more likely to be found for evaluations of programming in Middle Income Countries.

3.6.3 Contributing Factors — Monitoring and Evaluation

The reviews identified very few factors that contributed positively to monitoring and evaluation. There were one or two references to the fact that the evaluations were part of a broader process of evaluation – either as part of UNICEF's evaluation of country programs, or as part of a donor-led global evaluation. In one case it was noted that self-evaluations were conducted in anticipation of an external evaluation. A few evaluations noted that the need for data for program management contributed to the development of a monitoring system.

However, a range of factors were identified that detract from UNICEF's ability to monitor and evaluate its programming adequately. Some evaluation reflected that:

Box 19: Lack of focus on results

"Management for results remains a major challenge. There is … a growing commitment to monitoring girls' education processes and outputs but with less attention being paid to outcomes and impact that can be directly attributed to UNGEI. Results continue to be advanced in terms of (i) commitment of stakeholders; (ii) processes such as community participation at the investment stage; and (iii) direct project outputs such as girls' and boys' access and retention rates. Less interest has been directed to outcomes such as (a) community participation at the operational/implementation stage; and (b) the effects of training on actual teaching practices and how girls experience school differently as a result of teachers' training. Strong beliefs in UNGEI‟s effectiveness are thus not always backed by strong objective evidence."

Formative Evaluation of the United Nations Girls' Education Initiative: Country Report – Uganda: Final Report, August 2011, p.70

Box 20: Challenges with baseline data and indicators

"Measuring the results of the interventions is a challenge because of the absence of baseline data and that the statement of goals and objectives of all the four projects could have been more results-oriented and formulated in "SMART" way."

 Evaluation of Adolescent and Youth Participation in UNICEF Cambodia, 2011, [no page number]

A few evaluations reflected that:

4.0  Conclusions

  1. UNICEF-supported programs are highly relevant to the needs of target group members and are supportive of the development plans and priorities of program countries (satisfactory or better in almost all evaluations). UNICEF has also had success in developing effective partnerships with government and non-governmental organizations, especially in responding to humanitarian situations – which has, in itself, contributed to program relevance. The relevance of UNICEF programming has also been supported by efforts to ensure alignment with key national development plans and policy documents. It has also benefited from careful use of research into country conditions and the specific needs of target group members.
  2. UNICEF has largely been effective in achieving the objectives of the development and humanitarian programs it supports, with three-quarters of evaluations reporting findings of satisfactory or better. UNICEF-supported programs have also been effective in securing positive benefits for target group members and in supporting positive changes in national policies and programs. Where UNICEF programs have achieved success in objectives achievement, it has often been supported by high quality program design, with clear objectives that were integrated into national programs. It was also supported by UNICEF's active role in influencing national sector policies. On the other hand, when UNICEF-supported programs failed to meet their objectives, it was most often due to weaknesses in project design, often linked to unclear causal relationships and lack of a results orientation.
  3. These weaknesses in project design were contributing factors to mixed results with respect to sustainability. UNICEF achieves fairly positive ratings (more than two-thirds satisfactory or better) for its contributions to strengthening the enabling environment for development. However, the results for the likely continuation of results or the development of institutional and community capacity are not as good. Less than half the evaluations were rated as satisfactory or better on these two sub-criteria. A key factor to explain these results is the failure to plan for sustainability and integrate sustainability into program designs.
  4. UNICEF's performance with respect to gender equality is a serious concern. About one-third of evaluations did not address gender equality and, for those that did address it, the results were weak. As a result, only one-third of the evaluations were able to demonstrate that UNICEF-supported programs effectively address gender equality. A major factor contributing to these poor results was the lack of specific objectives regarding gender equality or the lack of a gender perspective in program design and implementation. Given the identification of gender equality as a foundation strategy, it is surprising that gender is adequately addressed in only two-thirds of the reviewed evaluations.
  5. There was insufficient coverage of environmental sustainability to warrant the presentation of related results. It appears that environmental sustainability or the impact of UNICEF-supported programs on their environment is not addressed in most evaluations, although some evaluations of humanitarian action and programs in water, sanitation and hygiene did address the issue. Given increasing emphasis on programs to mitigate the effects of climate change in some UNICEF country programs, coverage of environmental sustainability may be expected to improve in future evaluations.
  6. The results for efficiency of programming are mixed and the interpretation of the results is difficult because the sub-criteria associated with efficiency are not covered systematically in all evaluations. It is likely that factors related specifically to timeliness and implementation systems are only addressed in evaluations if they are problematic and can help to explain weakness in objectives achievement. However, the sub-criterion related to cost efficiency was more likely to be covered in the evaluations and showed somewhat positive results – almost two-thirds received a rating of satisfactory or better. However, this reflects ways in which programs have tried to reduce unit costs and increase the efficiency of resource use, rather than an analysis of overall program costs, as these costs were not identified in nearly half the evaluations. The factors that enhance cost efficiency include the establishment of effective monitoring systems to track costs and improve efficiency through lower delivery costs or unit prices and efforts to identify low-cost approaches to programming. However, where the results are not so positive, the detracting factor is most often the lack of appropriate cost data.
  7. The evaluations reflected somewhat positive findings with respect to the use of evaluation at UNICEF, but less so for monitoring and reporting. Nearly two-thirds of the evaluations were rated as satisfactory or better in relation to UNICEF's systems and processes for evaluation; but only two-fifths were given the same rating for UNICEF's monitoring and reporting on effectiveness. The use of evaluations is supported by an increasing tendency to prepare management responses for evaluations that include action plans for the implementation of evaluation recommendations. There were insufficient ratings for the use of results-based management systems to report on the results. The lack of clear results frameworks and appropriate indicators and baseline information were factors that detracted from UNICEF's effective use of monitoring and evaluation systems. The current development and implementation of the Monitoring Results for Equity Systems (MoRES) represents a significant effort to address this issue.
  8. While the sample of evaluations reviewed reasonably reflects the population of UNICEF evaluations between 2009 and 2011, a review of the profile, by country and by type of programming, suggests that UNICEF does not have adequate coverage of programming in countries that received the largest amounts of both development and humanitarian funding. There is noticeably limited coverage of UNICEF's humanitarian action.
  9. UNICEF is investing considerable time and effort into the development of quite new systems for results monitoring that are being implemented at all levels of the organization and in using the results of global evaluations for strategic planning. These initiatives hold out the promise of strengthened results reporting which could negate the need for development effectiveness reviews of this type in the future. That could depend on how well the evaluation function is able to be incorporated into the system as a means of verifying UNICEF's contribution and testing the validity of theories of change.

As noted in Section 3.0, UNICEF has undertaken a number of initiatives in the period following publication of the evaluations reviewed, in order to address some of the issues reported here. Examples of the main initiatives include:

While the review did not carry out an assessment of the expected results of these initiatives, it was able to verify, from the documents reviewed and the interviews carried out, that they represent a significant effort to respond to the issues identified. Their effectiveness will doubtless be the subject of future evaluations.

5.0 Canada's Relationship with UNICEF

This chapter provides an overview of Canada's relationship with UNICEF, with a focus on Canada's international development priorities and strategic objectives for engagement with the institution.  The purpose of this chapter is to fulfill evaluation requirements mandated by the Government of Canada's Policy on Evaluation to present evidence of the relevance, efficiency and effectiveness of its engagement with UNICEF; and to provide DFATD with evidence-based guidance on its future engagement with the institution.

5.1   Introduction

In May 2009, the Minister for International Cooperation announced that CIDA's development assistance would be focused on three thematic priorities: increasing food security; stimulating sustainable economic growth; and securing the future of children and youth. This chapter considers the extent to which UNICEF contributes to these priorities.  It begins with a description of Canada's financial support for, and engagement with UNICEF.Footnote 50  It then assesses the relevance, effectiveness, and efficiency of that engagement.  Conclusions and recommendations for DFATD's future engagement with UNICEF are provided at the end of the chapter.

Canada's Thematic Priorities for Development Assistance

DFATD, Priority Themes (2011)Footnote 51

5.1.1   Methodology

This assessment relies on data from multiple sources, including document review and interviews with DFATD staff, UNICEF, and other donors.  Among the documents reviewed were descriptions of Canada's overall foreign and development objectives, Canada's Due Diligence Assessment, and materials used at UNICEF Executive Board meetings. The review also included input provided by DFATD to UNICEF on strategic documents, including its draft Strategic Plan for 2014 – 2017 and the Annual Report of the Executive Director.Footnote 52 

Interviews were conducted with DFATD staff (7) from the Global Issues and Development Branch (GID), including project officers responsible for the institutional relationship with UNICEF and specific thematic projects. Staff from UNICEF and two member countries from UNICEF's Executive Board were also interviewed.Footnote 53

5.1.2   Canada's Support for UNICEF

Canada has been supporting UNICEF since its inception in 1946.Footnote 54 Table 4 summarizes Canada's funding to UNICEF between 2008-2009 and 2012-2013, when it totalled $772.25 million.

In 2011-2012, UNICEF received the second highest allotment of CIDA's international assistance delivered through multilateral organizationsFootnote 55, when it was the fifth largest donor to UNICEF's total budget.Footnote 56

Table 4: Canada's Development Assistance to UNICEF: 2008-2009 to 2012-2013 ($ millions)
Source and Type of Funding2008-20092009-20102010-20112011-20122012-2013Total
Multilateral Programming71.1892.8170.70101.6691.80428.15
Long-Term institutional support-18.0018.0036.0017.1089.10
International Humanitarian Assistance19.7025.656.7542.4536.10130.65
Initiative-specific51.4849.1645.9523.2138.60208.40
Bilateral Programming53.9865.2579.6873.8271.37344.10
Total125.16158.06150.38175.49163.17772.25

Chief Financial Officer Branch, DFATD (2013).

5.1.3   Canada's Relationship with UNICEF

Canada's engagement with UNICEF is facilitated through the Global Issues and Development Branch (GID), which is responsible for managing the institutional relationship. Other branches, such as geographic program branches and other thematic directorates in DFATD, also engage with UNICEF as an implementing partner for specific initiatives.

Overall, Canada's engagement with UNICEF is guided by five strategic objectives:Footnote 57

These objectives are pursued through formal and informal channels with the organisation, and at the project level.

Formal channels:

DFATD is directly involved in UNICEF's Executive Board meetings, which take place three times per year. Based on its regular annual contributions, Canada sits on the Board eight years out of fifteen, in accordance with an agreed rotational scheme. Canada was a member in 2013, and its next turns on the Board will be in 2015 and 2017.Footnote 58   As part of its preparation for Board meetings, GID solicits and incorporates feedback from geographic project officers working with UNICEF as an implementing partner. 

Informal channels:

Canada engages with UNICEF during informal meetings that are held leading up to UNICEF Executive Board meetings on specific agenda items. GID designates one headquarters officer as the UNICEF institutional lead.Footnote 59 Senior level engagement is also undertaken with UNICEF by MFM management. In addition, through Canada's Permanent Mission to the UN in New York, ongoing engagement with other donors and UNICEF staff is maintained outside of Executive Board meetings. These informal meetings facilitate collaboration among donors, particularly within the Western Europe and Others Group (WEOG), intended to influence decisions at the Executive Board and contribute to adoption of appropriate policies and institutional change.

Project Level:

Country programs and thematic directorates at DFATD engage with UNICEF for ongoing project management of Canada-supported projects.  Engagement with UNICEF in these instances can take place between DFATD's field offices and UNICEF's regional or country offices as well as between the headquarters of each organization in North America.

5.2   UNICEF and Canada's International Development Priorities

5.2.1   Relevance of UNICEF to Canadian Priorities

UNICEF's work aligns closely with two of Canada's three international development priorities, namely increasing food security and securing the future of children and youth. Initiatives in these areas can contribute to sustainable economic growth.

UNICEF programs contribute to food security

Canada's Food Security Strategy recognizes food assistance and nutrition programming as key interventions to address food insecurity and malnourishment.Footnote 60 UNICEF conducts programming in health and nutrition through its priority focus on Young Child Survival and Development. One-third (33%) of UNICEF's expenditures between 2009 and 2011 were distributed to this focus area.Footnote 61 UNICEF is one of the main actors in the UN Comprehensive Framework for Action on the Global Food Crisis, which seeks to develop a comprehensive approach to food and nutrition security.Footnote 62

UNICEF's initiatives in the area of health and nutrition include collaborative efforts to improve nutritional outcomes and access to water and sanitation. UNICEF also strengthens community-based integrated primary health care systems for women and children. Such efforts have contributed to the global reduction of under-five child mortality.Footnote 63 Progress has been further accelerated through the expansion of basic child health interventions and therapeutic feeding for the treatment of severe acute malnutrition. In addition, UNICEF is the Global Cluster lead for the Nutrition Cluster, which works to improve the nutritional status of emergency affected populations by ensuring a coordinated, appropriate response that is predictable, timely, effective and at scale.Footnote 64

Through such initiatives, UNICEF directly contributes to Canada's commitments to improving maternal and newborn child health.Footnote 65 UNICEF is presently implementing a number of Canadian-funded projects that support results achievement in this area.  For example, a project in Malawi, Mozambique and Zimbabwe aims to reduce the vulnerability of children to under nutrition as well as HIV/AIDS.Footnote 66 Canadian development assistance is also supporting a UNICEF project in Nigeria that aims to reduce maternal, newborn and child mortality.

UNICEF's mandate and focus center on securing the future of children and youth

UNICEF is mandated by the United Nations General Assembly to advocate for the protection of children's rights, to help meet children's basic needs, and to expand their opportunities to reach their full potential.Footnote 67  As such, UNICEF's programming directly relates to Canada's international development priority of child protection and education. Between 2009 and 2011, UNICEF channeled 7% of its expenditures to child protection initiatives.Footnote 68 This included improving justice systems, and advocating against child labour and the recruitment and use of children in armed conflict. In 2011 for example, UNICEF supported the community reintegration of more than 11,600 children formerly associated with armed forces or groups.Footnote 69

In addition to protection, UNICEF aims to foster basic quality education for boys and girls, including in humanitarian situations. Between 2009 and 2011, 16% of UNICEF's expenditures were for education and gender equality.Footnote 70 UNICEF particularly promotes girls' education to improve gender parity in primary and secondary education.Footnote 71 It also works to ensure that national policy frameworks integrate appropriate policy, legislation and budget allocations aimed at universal school readiness.Footnote 72

UNICEF builds foundations for sustainable economic growth

UNICEF helps to build the foundations necessary for long-term economic development through its interventions in food security, health and education.Footnote 73 However, UNICEF's direct contribution to sustainable economic growth is minimal. UNICEF's partnerships with governments, private sector organizations and civil society do contribute to long-term economic sustainability by improving governance and strengthening development planning and results.

UNICEF's contributions to gender equality, environmental sustainability and improved governance

UNICEF has committed to mainstreaming gender equality and environmental sustainability at all levels.  Nonetheless, UNICEF demonstrates weaknesses in integrating gender equality and environmental sustainability into its planning and programming. This is an area of serious concern, as one-third of UNICEF evaluations included in the sample (32%) for this review failed to consider gender equality.Footnote 74 There remains a need to identify gender-sensitive objectives throughout all of UNICEF's programming, and to collect sex-disaggregated data and report on gender-sensitive results.Footnote 75

Despite the fact that UNICEF has an environmental policy and environmental assessment practices, this review was not able to report on the effectiveness of UNICEF programs in this regard.  Therefore, due to a lack of available evidence, UNICEF's performance in environmental sustainability could not be assessed.  The review suggests that environmental sustainability is gaining increased attention in UNICEF, as demonstrated by the creation of a working group on climate change and programming related to water, sanitation and hygiene.Footnote 76  These developments suggest that future UNICEF evaluations may pay increased attention to issues of environmental sustainability.

The methodology for this review did not include a criterion to assess governance.Footnote 77  Several of UNICEF's program interventions align with good governance, such as capacity building in policy, legislative and budgetary matters.  However, findings on governance are beyond the scope of this exercise.

5.2.2   Effectiveness: Achievement of Canada's Strategic Objectives

This section of the report reviews Canada's progress in achieving its five strategic objectives for its engagement with UNICEF. While GID has not to date developed a formal results framework, this section uses available evidence to review progress towards their achievement.

Broadening UNICEF's reach to the most vulnerable and under-served

In an effort to reach the most vulnerable target groups, including young girls, Canada encourages UNICEF to scale up successful initiatives and to use evidence-based programming, and emphasizes the collection of sex-disaggregated data. UNICEF demonstrates weaknesses in planning and reporting on gender-sensitive results. For example, the 2013 Annual Report of the Executive Director lacked gender-specific analysis despite the availability of sex-disaggregated data.Footnote 78 The development of a Strategic Priority Action Plan for Gender (SPAP) (2010-2013) demonstrates some progress towards gender-sensitive planning and is a key tool to improving the integration of gender equality in UNICEF programs. Further effort is needed to ensure that gender equality is integrated into UNICEF's overall programming.Footnote 79

UNICEF demonstrates a strong focus on equity as exemplified by, for example, inclusive initiatives that support children with disabilities as well as out-of-school children.Footnote 80 UNICEF has taken steps to identify equity gaps through its global Social Protection Strategic Framework, which allows for sustainability audits and social budgeting.

Supporting UNICEF's efforts to work more effectively with partners and other UN agencies

Canada has encouraged UNICEF to enhance its engagement with partners to facilitate a higher level of effectiveness. UNICEF evaluations demonstrate that partnerships are a key element to the relevance of UNICEF's programming, with stronger partnerships promoting a higher degree of both effectiveness and sustainability.Footnote 81 Gaps in partnerships (which occur more commonly in Least Developed Countries than Middle Income Countries) may arise because of weak coordination and reliance on poor data, the instability of partnerships during the course of programming, and frequent staff changes.

UNICEF is increasingly working with non-governmental organizations, national governments and the private sector to develop and implement programming that advances the long-term sustainability of results.Footnote 82 UNICEF aligns its programming with national priorities and works with governments to build capacity and implement public policy. UNICEF also increasingly engages in partnerships with other UN agencies to maximize development impacts and reduce duplication.Footnote 83 Partnerships such as the GAVI Alliance, Scaling Up Nutrition and Global Education initiatives promote stronger national health, education and protection systems.Footnote 84 However, UNICEF has occasionally demonstrated weaknesses in sharing information and harmonizing its aid efforts with other UN agencies.Footnote 85

Strengthening UNICEF's capacity in humanitarian crises

UNICEF is a central actor in providing humanitarian assistance, and focuses on child protection, nutrition, and water and sanitation needs. Its mandate is to meet the needs of the most vulnerable, to save lives and protect rights as defined in the Core Commitments to Children in Humanitarian Action.Footnote 86 UNICEF plays a vital role in the Cluster System, which designates lead agencies to coordinate specific sectors of an emergency response with UN agencies and other humanitarian organizations. UNICEF currently leads on the Nutrition Cluster, the Water, Sanitation and Hygiene Cluster and is the co-lead for the Education Cluster with Save the Children. In addition, UNICEF has responsibilities within the Protection Cluster for child protection and gender-based violence.

UNICEF is increasingly demonstrating progress in the area of humanitarian assistance. For instance, Canada and other donors have stressed the importance of including humanitarian assistance in UNICEF's strategic plans, given that it represents a quarter of its budget. As a result, UNICEF's 2014-2017 Strategic Plan will integrate humanitarian assistance throughout. Canada and other donors continue to encourage strong cluster leadership in the field and coordination across clusters to ensure an effective humanitarian response. Initiatives demonstrating action in this area include UNICEF's response strategy to complex emergencies, such as Haiti and Pakistan,Footnote 87 and its first global evaluation of child protection in emergencies in 2012.Footnote 88

To further improve its capacity for humanitarian response, Canada has encouraged UNICEF to strengthen its performance monitoring and reporting in emergencies, better staff humanitarian coordinator positions for UNICEF-led clusters, and strengthen planning and reporting on risk assessment and mitigation.Footnote 89

Working with other donors to support a more active role by UNICEF in UN Reform

Canada and other members of the UN General Assembly are taking steps to enhance the effectiveness and relevance of the UN development system through the Quadrennial Comprehensive Policy Review (QCPR), which seeks to strengthen coordination across the UN. The QCPR requests that member states support system-wide cost sharing for the Resident CoordinatorFootnote 90 and ensure coherent messaging across Executive Boards and governing bodies of funds, programs and specialized agencies.Footnote 91 UNICEF recognizes Canada's active engagement with the Utstein Group, an initiative that seeks to make the UN system more effective.Footnote 92 UNICEF also considers Canada a leader in good donor practices, such as reducing the burden on international institutions.Footnote 93

In the context of these efforts, Canada encourages UNICEF to better collaborate with the broader UN development system. However, this has been a challenge as UNICEF has a strong independent identity and significant funding sources independent of the UN allocation process.Footnote 94  UNICEF has occasionally demonstrated reluctance to engage in the wider agenda of UN reform.Footnote 95 Canada has encouraged UNICEF to improve its collaboration with other UN agencies, but this remains an area for improvement.

Promoting the objectives of the Joint-Institutional Approach, with emphasis on gender equality and results-based management

In 2006, Canada, Sweden and the United Kingdom established the Joint-Institutional Approach (JIA) in an effort to assist UNICEF with the implementation of its Medium Strategic Plan for 2006 – 2013.Footnote 96 The JIA acted as a partnership framework, which "intended to guide the three donors in working more coherently and effectively with UNICEF in the spirit of good donorship".Footnote 97 The JIA's seven areas of work included: human rights based approaches; gender equality; humanitarian capacity; results-based management; evaluation; UN reform; and human resources.

While the JIA is no longer active, Canada continues to work alongside donors through the Western European and Others Group (WEOG) to advocate for objectives that are similar to those of the JIA, particularly with respect to the integration and application of results-based management and gender equality. Interviews with other donors indicate that Canada plays a key leadership role in promoting both of these areas. In particular, other donors highlighted Canada's ongoing efforts to engage additional member countries during Executive Board meetings, issue-specific committees and the peer review created to provide input to UNICEF's Strategic Plan (2014-2017). Other donors also attest to Canada's collaborative and transparent approach when providing feedback to UNICEF.

Canada and other donors are working with UNICEF to identify and refine appropriate indicators and measurable results at the institutional level. The aim of these efforts is for UNICEF to provide more demonstration of both outcomes and the "performance story" of the institution.Footnote 98 In this regard, Canada encourages UNICEF to collect sex-disaggregated data. Canada also seeks to ensure that sex-disaggregated data is incorporated into UNICEF's reporting.Footnote 99 Canada has encouraged UNICEF to integrate gender sensitive information, and specifically, gender-sensitive indicators, into all seven focus areas of UNICEF's Strategic Plan (2014-2017).Footnote 100

Canada continues to focus on gender equality and results-based management in its efforts to influence UNICEF's institutional development. Through its leadership role at formal and informal meetings, Canada recognizes improvements within UNICEF, particularly progress made in the area of gender equality, while also continuously providing constructive feedback. UNICEF describes Canada's current engagement as "very good, fruitful and fluid".Footnote 101

Opportunities for future consideration

It has been several years since Canada's institutional objectives for engagement with UNICEF were developed.  An update of these objectives may be in order, especially given areas of potential future emphasis that were identified in this review.

One area identified by the review that was not included as a priority in Canada's current objectives for its relationship with UNICEF is the sustainability of programming. Despite its progress in gender equality and equity, some UNICEF programs demonstrate weaknesses in continuity and sustainability. These challenges stem from a lack of planning and an absence of exit strategies. Greater consideration to sustainability during program design and implementation will be crucial if UNICEF programs are to make the greatest possible contribution to the most vulnerable and under-served.Footnote 102 Section 3.4 of the review provides detail on this finding.

A second area for improvement flagged by the review is UNICEF's evaluation-based reporting on humanitarian action. While humanitarian programming accounted for over one-quarter of UNICEF expenditures between 2009 and 2011, few humanitarian programs were evaluated for this period.Footnote 103 The evaluations that were available on humanitarian programming covered a narrow range of programming, as, for instance, five of the humanitarian evaluations included in the sample for this review reported on the same crisis – the 2004 Indian Ocean Tsunami.Footnote 104 Humanitarian-specific evaluations are mostly limited to the headquarters level.Footnote 105

Canada's advocacy related to humanitarian assistance actively contributes to strengthened capacity within UNICEF. However, Canada's current strategic objectives do not specifically mention reporting on outcomes in humanitarian crises.

5.2.3   Management Practices: Assessing the efficiency of Canada's engagement with UNICEF

The management of Canada's institutional relationship with UNICEF relies on a single institutional lead at DFATD headquarters and the First Secretary for Development in New York. All non-DFATD interviewees remarked on the efficiency of this staff allocation. The presence of an experienced staff member in both New York and Ottawa now contributes to a sustained engagement with UNICEF, a critical element according to UNICEF.Footnote 106 The institutional lead's outreach to project officers in DFATD proves efficient during preparations for Executive Board meetings, which requires input from many staff and good coordination across DFATD. However, feedback to UNICEF can be delayed due to DFATD's internal processes. For example, UNICEF encounters delays in communication from DFATD regarding the status of funding. Timely communication from DFATD to UNICEF regarding funding status is an area for improvement.

Engagement between DFATD and UNICEF can be complex, given differing management structures.  DFATD has separate development and humanitarian funding channels while UNICEF has multi-faceted programming, including areas where humanitarian and development programming are merged, as in the case of Syria.Footnote 107 Thus, UNICEF staff may interact with both development and humanitarian funding channels at DFATD to assemble funds for crosscutting initiatives. There was observation of a need in some circumstances to coordinate messaging between development and humanitarian funding channels within DFATD, in order to ensure coherent outward communication with UNICEF.

5.3   Conclusions

Based on the above, this chapter concludes that:

  1. UNICEF's efforts are highly relevant to two of Canada's international development priorities: increasing food security and securing the future of children and youth. UNICEF programs in health and education also indirectly contribute to Canada's third development priority of long-term sustainable economic growth. 
  2. UNICEF demonstrates progress in integrating gender equality into programming, but continues to lack sufficient performance measurement and reporting in this area. UNICEF's contribution to environmental sustainability and governance could not be assessed in this review. 
  3. Progress has been made towards the achievement of Canada's strategic objectives for its engagement with UNICEF. However, DFATD has not developed performance measurement tools to track and assess its performance in achieving these objectives.
  4. UNICEF increasingly engages in partnerships with civil society, national governments, and with other UN agencies, all of which contributes to the effectiveness of its programs. Despite these partnerships, UNICEF has occasionally demonstrated reluctance to engage in the wider agenda of UN reform, an effort that Canada and other donors actively champion.
  5. In its engagement with UNICEF, Canada has not emphasized the sustainability of programming, an area of weakness highlighted by UNICEF's own evaluations.
  6. Canada is influential in advocating for strengthened humanitarian capacity within UNICEF. Performance monitoring and reporting on outcomes of humanitarian action remain a challenge.
  7. Canada uses its human and financial resources efficiently in its engagement with UNICEF, a conclusion echoed by both other donors and UNICEF.

5.4   Recommendations for Canada

This section contains recommendations to DFATD based on the conclusions derived from the assessment of Canada's engagement with UNICEF. As one of several donors working with UNICEF, Canada is limited in the extent to which it alone can influence improvements in the development effectiveness of the organization. Accordingly, DFATD should continue to engage with other donors to advocate for these improvements. 

  1. DFATD should continue to emphasize the integration of gender equality in UNICEF programs and should highlight the importance of performance measurement and reporting in this area.   
  2. DFATD should develop a performance measurement framework to assess progress on Canada's strategic objectives for its relationship with UNICEF
  3. As Canada and other members of the General Assembly are seeking mechanisms to harmonize the UN development system, DFATD should continue to work with UNICEF to identify realistic and tangible ways to collaborate with other UN organizations.
  4. DFATD should encourage UNICEF to improve its sustainability planning. This could involve efforts to strengthen UNICEF's program design and implementation plans
  5. DFATD should encourage UNICEF to increase its outcome-level reporting, analysis and evaluation of humanitarian assistance programming.

Annex 1: Effectiveness Criteria

Relevance

Achievement of Humanitarian and Development Objectives and Expected Results

Cross Cutting Themes: Inclusive Humanitarian and Development Which can be Sustained (Gender Equality and Environmental Sustainability)

Sustainability

Efficiency

Using Evaluation and Monitoring to Improve Humanitarian and Development Effectiveness

Annex 2: Evaluation Sample

  1. 2009 Maldives: Children and the 2004 Indian Ocean Tsunami: Evaluation of UNICEF’s Response in Maldives (2005-2008) Country Synthesis Report
  2. 2009 Sri-Lanka: Children and the 2004 Indian Ocean Tsunami: Evaluation of UNICEF’s Response in Sri Lanka (2005-2008) Country Synthesis Report 2009 Thailand: Children and the 2004 Indian Ocean Tsunami - Evaluation of UNICEF's Response in Thailand (2005-2008)
  3. 2009 Thailand: Children and the 2004 Indian Ocean Tsunami - Evaluation of UNICEF's Response in Thailand (2005-2008)
  4. 2009 Indonesia: Children and the 2004 Indian Ocean Tsunami: Evaluation of UNICEF’s Response in Indonesia (2005-2008) Country Synthesis Report
  5. 2011 Global: Independent Review of UNICEF's Operational Response to the January 2010 Earthquake in Haiti
  6. 2009 Indonesia: Children and the 2004 Indian Ocean Tsunami: UNICEF's Response in Indonesia (2005-2008) Child Protection
  7. 2009 Colombia: Educación en el Riesgo de Minas (ERM) y Asistencia a Víctimas en los departamentos de Cauca, Chocó, Nariño y la región de La Mojana /Sur de Bolívar
  8. Evaluación Del Programa "Escuela Busca al Niño"
  9. Rapport De L’évaluation Externe De La Phase Pilote Du Projet De Prise En Charge Des Orphelins Et Autres Enfants Vulnérables En République De Djibouti
  10. L’Evaluation des résultats des 12 Associations de Services Financiers (ASF) pour le développement Communautaire dans la zone de Kissidougou
  11. Evaluation of the actions to attend to the water sanitation and hygiene sector among populations affected by Hurricane Felix
  12. Evaluation de la couverture de la campagne nationale de distribution des moustiquaires imprégnées en 2008 et de l'impact des interventions de lutte contre le paludisme au Togo/Evaluation of the impact of anti malaria interventions, including 2008 national campaign to distribute ITNs
  13. Evaluation of the UNICEF-DIPECHO programme “Supporting Disaster Risk Reduction Amongst Vulnerable Communities and Institutions in Central Asia and South Caucasus
  14. Evaluation of the impact of educational innovations (IECD centers, satellite schools, non formal basic education centers) on Burkina Faso’s educational development system
  15. Monitoring & Evaluation of Child Health Days in Madagascar
  16. Mid-term Evaluation of project of Basic Education in Eastern DRC (BEED)
  17. Recovery Action and Rehabilitation Project (RARP) Evaluation
  18. Evaluation of the National Plan for the Eradication of Child Labour 2000-2010
  19. Evaluation des campagnes de supplémentassions en Vitamine et de la déparasitage (2002-2006) et des Semaines de la Sante de la Mère et de l'Enfant (2006-2011) a Madagascar
  20. 2010 Ethiopia: Mid-Term Evaluation of EU/UNICEF Supported WASH Programme
  21. Assistance aux personnes expulsées de la Tanzanie et appui à la réintégration des personnes rapatriées, expulsées et déplacées (2006-2009) Assistance to persons expelled from Tanzania and support to the reintegration of repatriated, expelled and displaced person (2006-2009)
  22. Evaluation, consolidation and strengthening of the “I'm a Person, Too” programme for optimum psychosocial development of children under six in the Department of Huila, Colombia - Evaluación del PROGRAMA “TAMBIÉN SOY PERSONA”
  23. Evaluation of the YCSD programme
  24. Evaluation of community-led total sanitation (CLTS)
  25. Evaluation of the community- based rehabilitation project for disabled children in 4 municipalities of Oaxaca
  26. Impact du Changement de Normes Sociales sur les Comportements en Milieu Rural au Sénégal
  27. 2011 Ghana: Evaluation of UNICEF's Early Childhood Development Programme with Focus on Government of Netherlands Funding (2008-2010)
  28. 2009 Georgia: UNICEF’s Response to Georgia Crisis: Real Time Evaluation
  29. 2010 Kenya: Cash Transfer Programme for Orphans and Vulnerable Children (CT-OVC), Kenya: Operational and Impact Evaluation, 2007-2009
  30. 2009 Cambodia: Evaluation of Community-Led Total Sanitation
  31. Community Radio Listening Groups Project
  32. Evaluation of the EthioInfo Utilization in Ethiopia
  33. Evaluation of PHAST tool for the promotion hygiene Sanitation in the GOK/UNICEF Programme of cooperation
  34. 2010 Ghana: Review of Second Performance Monitoring of the IWASH Project
  35. Evaluation of Integrated Management of Childhood Illnesses Initiative in the Republic of Moldova Years 2000-2010
  36. Evaluation of Adolescent and Youth Participation in UNICEF Cambodia
  37. Evaluation of Child and Youth Participation Initiatives in UNICEF Mozambique
  38. Evaluation of the UNICEF/UNFPA Joint Programme “Support to Ghana’s National HIV AND AIDS Response: Scaling up Best Practices on Prevention, Care and Support Interventions for Young People”
  39. Emergency Water Supply to un-served/underserved/ Vulnerable Areas in Baghdad and the IDPs
  40. Evaluation of UNICEF Nepal Mine Action Activities: Victim-Activated Explosion Injury Surveillance and Mine Risk Education
  41. 2009 Uganda: Final Review of UNICEF-supported Programmes for Children Affected by Conflict in Kitgum, Northern Uganda
  42. Supporting Sustainable water management and governance for the poor in drought and flood prone areas in Kenya
  43. 2010 WCARO Regional: Roll Out Evaluation of Community Led Total Sanitation in West & Central
  44. Evaluation of Gender Sensitisation and People Friendly Police Initiative in Karnataka
  45. 2010 Iraq: Evaluation of Water Quality Control and Surveillance in Iraq
  46. ZIMWASH Project End- Term Evaluation Report: ACP EU Water Facility Project - 2006-2011 - Addressing water and sanitation needs of the rural poor in the context of HIV and AIDS in Zimbabwe
  47. 2010 Sudan: Go To School Evaluation
  48. Evaluation of UNICEF Bangladesh Education and Child Protection Programmes
  49. Evaluation of Social Work Coaching
  50. Evaluation of Sustained Outreach Services (SOS) for immunization/ Vitamin A (Indonesia)
  51. 2010 Zimbabwe: Evaluation of Programme of Support for National Action Plan for Orphans and Vulnerable Children Impact/ Outcome Assessment
  52. 2010 Myanmar: Evaluation of UNICEF Education Programme - Improving Access to Quality Basic Education in Myanmar (2006-2010)
  53. 2010 Madagascar: Evaluation on girl-to-girl mentorship strategy
  54. Summary Report and Evaluation: Child Injury Prevention Project (2005-2010)
  55. Evaluation of Netherlands-UNICEF Water Initiative (NUWI)
  56. Real-Time Evaluation of UNICEF’s Response to the Sa’ada Conflict in Northern Yemen
  57. 2011 Tanzania: Evaluation of UNICEF's Early Childhood Development Programme with Focus on Government of Netherlands Funding (2008-2010)
  58. Formative Evaluation of the United Nations Girls Education Initiative (UNGEI)
  59. 2011 Mozambique: Impact evaluation of drinking water supply and sanitation interventions in rural Mozambique: More than Water
  60. Assessing of WASH package interventions in 5 counties of Liberia
  61. 2010 Niger: Evaluation of Cash Transfer for Protection of Blanket Feeding: UNICEF Emergency Project Niger
  62. 2010 Philippines: UNICEF Philippines Country Program Evaluation

Annex 3: Methodology

This annex provides a more detailed explanation of the population identification and sampling methodology used for the review of UNICEF's development and humanitarian effectiveness and includes a comparison of the sample to the population of evaluations and, on the issue of UNICEF funding, to global funding for the 2009 – 2011 period.

UNICEF Evaluation Population

UNICEF evaluations were identified from two sources:

This included evaluations conducted or commissioned by the central Office of Evaluation, as well as evaluations conducted at the decentralized level by UNICEF Country or Regional Offices.

These sources identified a population of 341 evaluations. It was decided, in consultation with the UNICEF EO that the review should focus on evaluations conducted since 2008. UNICEF had implemented a new evaluation policy in 2008 and evaluations conducted after that year would be likely to reflect the new policy. As a result, the scope of the review would include only evaluations conducted in 2009, 2010 and 2011.Footnote 108 It was also agreed that the review should not include those evaluations deemed, in the Global Evaluation Report Oversight System (GEROS) quality rating, to be of poor quality and "not confident to act." Finally, evaluations that focused on global programming were dropped from the population of evaluations from which to draw the sample for the qualitative review. These global evaluations were not included in the sample, but were rather subject to a qualitative review by senior team members. In addition, some evaluations in the population were duplicates, were not deemed to be evaluations or did not focus on UNICEF programming and were dropped from the population. This resulted in a population of 197 from which to draw the sample (Table 5).

Table 5: Development of Populations of UNICEF Evaluations

Evaluation Sample

A sample of 70 evaluations was drawn from the population, with the intent that, after the quality review, there would be 60 evaluations available for review. Initially the sample was random, stratified by Medium-Term Strategic Plan theme area. Then the sample was adjusted manually to ensure adequate coverage on two other dimensions: region and type of country (low- or middle-income countries). The resulting sample was no longer random, but rather a purposive sample that was illustrative of UNICEF's programming across a number of dimensions – year, commissioning office, region, type of country, country and Medium-Term Strategic Plan focus area.

Quality Review

The first task of each reviewer was to review the quality of the evaluation, using a standard review grid (Table 6). The grid reflects the criteria being rated and how the maximum number of points is allocated for each criterion.

The purpose of this review was two-fold:

Table 6: Quality Review Grid
NoPoints for criteria scoredMaximum PointsScore
APurpose of the evaluation
  • why the evaluation was done (1)
  • what triggered the evaluation (including timing in the project/program cycle) (1)
  • how evaluation is to be used (1)
3 
BEvaluation objectives
  • evaluation objectives are clearly stated (1)
  • objectives logically flow from purpose (1)
2 
COrganization of the evaluation
  • logical structure to the organization of the evaluation (1)
  • evaluation is well written (1)
  • clear distinction between evidence, findings, conclusions, and recommendations (1)
3 
DSubject evaluated is clearly described
Evaluation describes:
  • the activity/program being evaluated (1)
  • the program's expected achievements (for humanitarian: identification of needs of target population and agency specific role in addressing needs) (1)
  • how the program addresses the development problem (1)
  • the implementation modalities used (1)
4 
EScope of the evaluation
Evaluation defines the boundaries of the evaluation in terms of:
  • time period covered (1)
  • implementation phase under review (1)
  • geographic area (1)
  • dimensions of stakeholder involvement being examined (1)
4 
FEvaluation criteria
Evaluation criteria include:
  • the achievement of development objectives and expected results (including impacts) (1)
  • cross-cutting issues: inclusive development which is gender sensitive and environmentally sustainable (1)
  • the sustainability of benefits and positive results achieved (1)
  • the relevance of MO activities and supported projects and programs (1)
  • the efficiency of MO operations in support of projects and programs (1)
5 
GMultiple lines of evidence
  • one point (1) for each line of evidence used (case studies, surveys, site visits, interviews, etc.), up to a maximum of four points (4)
4 
HEvaluation design
Elements of a good evaluation design include:
  • an explicit theory of how objectives and results were to be achieved (1)
  • specification of the level of results achieved (output, outcome, impact) (1)
  • baseline data (quantitative or qualitative) on conditions prior to program implementation (1)
  • comparison of conditions after program delivery to those before (1)
  • a qualitative or quantitative comparison of conditions among program participants and a control group (not applicable to humanitarian evaluation, assign 1 point) (1)
5 
IEvaluation findings and conclusions are relevant and evidence based
Evaluation report includes:
  • evaluation findings relevant to the assessment criteria (1)
  • findings that are supported by the chosen methodologies (1)
  • a clear logical link between the evidence and the finding (1)
  • conclusions which are linked to the evaluation findings as reported (1)
4 
JEvaluation limitations
  • statement of the limitations of the methodology (1)
  • impact of limitations on evaluation (1)
  • remedies of limitations (1)
3 
KEvaluation Recommendations
  • evaluation contains recommendations that flow from findings and conclusions (1)
  • recommendations are directed to one or more organization (1)
  • recommendations are aimed at improving development effectiveness (1)
3 
 Total (required to have a minimum of 25 points) 40 
 Total for Criteria G, H and I (required to have minimum of 9 points) 13 

Any evaluation that failed to meet either one of these conditions was rejected from the sample. The initial sample of evaluations included 70 evaluations. Of these four were rejected because:

Of the remaining 66 evaluations, all achieved sufficient points in the overall quality score, but four did not receive sufficient points for the specific criteria related to measuring effectiveness and were rejected from the sample (Table 7).

Table 7: Results of Quality Review Scoring
Overall evaluation quality scoresDevelopment effectiveness key criteria
Max points = 40 (Min required = 25)No. of evaluations% of evaluationsMaximum de points = 13
(minimum requis = 9)
No. of evaluations% of evaluations
36 - 401522.7130 
31 - 353654.51269.1
26 - 301522.7112436.4
21 - 25102537.9
16 - 209710..6
11 - 15823.0
6 - 10711.5
0 - 56 0.0
511.5
Total66100.0Total66100.0

Comparison of Evaluation Population and Sample

This section provides a profile of the evaluation population and sample and comments on the extent to which the sample is illustrative of the population of UNICEF evaluations and UNICEF programming for the period 2009 – 2011.

Evaluation and Programming Years

The sample was spread across the three years 2009 – 2011 (Table 8). The spilt by year parallels that of the population of evaluations.

Table 8: Profile of Population and Sample of Evaluations, by Year
YearPopulationSample
 NumberPercentNumberPercent
20097538%2235%
20106131%2337%
20116131%1727%
Total197100%62100%

However, the programming covered by these evaluations spanned the period 2000 to 2010. Just over one-third of the evaluations covered three-year programs and another 40% four- to five-year programs (Table 9).Footnote 110 However, a few (16%) covered programs that lasted between six and eleven years.

Table 9: Profile of Sample of Evaluations, by Range of Years Covered
Span of Years Covered by EvaluationsNumberPercent
1 - 3 years2337.1%
4 - 5 years2540.3%
6 - 8 years69.7%
9 - 11 years58.1%
Not available34.8%
 Total62100.0%

This represents a limitation of evaluation meta-syntheses. Evaluations are, by definition, retrospective and a meta-synthesis is even more retrospective, as it is based on a body of evaluations conducted over a much earlier period of time. UNICEF's policies, strategies and approaches to programming have changed over these years, but the changes will not be reflected in all the evaluations. For this reason, the findings may be somewhat dated. To the extent possible, the review addresses this through observations gleaned from recent interviews with UNICEF staff and a review of UNICEF documents.

Commissioning Office

One-fifth of the evaluations included in the sample were commissioned by the Central Evaluation Office (Table 10). The reviewers were unable to identify the commissioning office in one-third of the evaluations (not available). Nearly half the evaluations (46%) were commissioned by UNICEF country or regional offices.

Table 10: Profile of Sample of Evaluations, by Commissioning Office
Evaluation Commission OfficeNumberPercent
Central Evaluation Office1219.4%
Country Office2743.5%
Regional Office23.2%
Not available2133.9%
Total62100.0%

Regions and Type of Country

The sample was spread across all regions and types of countries in which UNICEF is programming (Tables 11 et 12).

Table 11: Profile of Population and Sample of Evaluations, by Region
RegionPopulationSample
NumberPercentNumberPercent
Middle East and North Africa189711
West and Central Africa24121118
Eastern and Southern Africa44221626
Latin America and Caribbean2211711
CEE/CIS281423
South Asia189610
East Asia and Pacific2915915
Multi-region14746
Total 19710062100

* Regional classification was based on information from The State of The World's Children 2012: Children in an Urban World, UNICEF, 2012, p. 124  

Table 12: Profile of Population and Sample of Evaluations, by Country Classification Footnote 111
Country ClassificationPopulationSample
 NumberPercentNumberPercent
Least-developed countries76393455
Middle-income countries90462642
Regional/multi-country12623
Not identified1910
Total19710062100

The sample included slightly more evaluations from West and Central Africa and slightly fewer evaluations from Central and Eastern Europe/Commonwealth of Independent States (CES/CIS) than would have been suggested by the population. As would be consistent with this, the sample included slightly more evaluations from least-developed countries that would have been suggested by the evaluation population.

Countries

This oversampling of least-developed countries in West and Central Africa occurred because the review had to oversample from the countries receiving the largest amount of UNICEF funding. Table 13 reflects the countries that were in the top funded countries for development programming in 2007 – 2011. Nineteen evaluations (7% of all evaluations in the population) and nine evaluations in the sample (18%) come from the top funded countries of the five-year period 2007 - 2011. This suggests that UNICEF is not achieving strong evaluation coverage in the countries that receive the largest amounts of development funding. The team oversampled from the top funded countries to increase the coverage of countries that received the most UNICEF development funding.

Table 13: Profile of Population and Sample of Development Evaluations, by Funding and Country, by Funding for 2007 – 2011 Footnote 112
CountryNo. of years country included in top funded countries with development funding*
(Maximum 5 year)
Evaluation identified
(2009 – 2011)
Population
(182 evaluations)
Sample(49 evaluations)
India531
Nigeria510
Ethiopia533
Afghanistan511
Pakistan500
Bangladesh521
Mozambique531
Malawi500
Southern Sudan532
Somalia530
Total  199
As % of all evaluations in population or sample6.6%18.4%

* Regular resources and Other resources - Regular

There is a similar pattern with respect to humanitarian action. Table 14 reflects the countries that were in the top funded countries for humanitarian action in 2007 – 2011. Seven evaluations in the population (47%) and three evaluations in the sample (23%) came from the countries that received the most humanitarian funding in the years 2007 – 2011. The coverage of the top funded countries is better for humanitarian action than development programming because the population of evaluations included Inter-Agency Steering Committee evaluations, which do not address UNICEF programming specifically. However, this suggests, as with development evaluations, that UNICEF is not achieving strong evaluation coverage of its programming in the countries that receive the largest amounts of humanitarian funding. The team oversampled from the top countries to increase the coverage of countries that received the most UNICEF humanitarian funding.

Table 14: Profile of Population and Sample of Humanitarian Evaluations, by Funding and Country, by Funding for 2007 - 2011 Footnote 113
CountryNo. of years country included in top funded countries with humanitarian funding*
(Maximum 5 years)
Evaluation identified
(2009 – 2011)
Population
(15 evaluations)
Sample(13 evaluations)
Somalia000
Pakistan**220
Northern Sudan000
Ethiopia000
Zimbabwe000
Southern Sudan100
Sri Lanka221
Indonesia532
Total73
As % of all evaluations in population or sample46.7%23.1%

* Funding from Other resources – emergency
** Evaluations from Pakistan were not included in the sample because they did not cover UNICEF programming

Medium-Term Strategic Plan Focus Areas

The evaluations in the sample covered all priority areas of UNICEF programming, including all themes in the Medium-Term Strategic Plan and humanitarian action (see Table 15). However, not all types of UNICEF programming are covered equally by the evaluations. Only fifteen evaluations in the population (8%) and thirteen (19%) in the sample covered humanitarian action; whereas this programming accounted for just over one-quarter of UNICEF expenditures in 2008 – 2011. There is proportionally less coverage of humanitarian action by UNICEF's evaluations that development programming. It should also be noted that five of the humanitarian evaluations included in the sample, although not for programming in the same country, covered programming for the same humanitarian crisis – the 2004 Indian Ocean tsunami. Only one of the hardest hit countries receiving funding in response to the tsunami is included in the list of top humanitarian funded countries for 2009 – 2011. This means that, although the evaluations were published in years included in the sample, the funding occurred several years earlier.

Table 15: Profile of Population and Sample of Evaluations, by Medium-Term Strategic Plan Focus Area and Humanitarian Action Footnote 114
Medium-Term Strategic Plan Focus Area/Humanitarian Action% of Expenditures (2009 – 2011)*PopulationSample
 PercentNumberPercentNumberPercent
Development (Regular and Other Resources - Regular)      
1. Young Child Survival and Development33.40%4322%2235%
2. Basic Education and Gender Equality16.40%6533%1118%
3. HIV/AIDS and Children5.20%116%35%
4. Child Protection from Violence, Exploitation and Abuse7.40%3518%610%
5. Policy Advocacy and Partnership9.80%179%711%
Sub-total73.40%    
Humanitarian (Other Resources - Emergency) 26.60%158%1219%
Total 100.00% 94%  
Other (Country Program, Other, Blank) 116%12%
Total Evaluations 197100%62100%

Apart from the fact that the evaluations do not cover the countries receiving the largest amounts of funding, or the full range of humanitarian action, the evaluations covered in the sample are, in all other aspects, illustrative of UNICEF's global programming.

Annex 4: Guide for Review Team to Classify Evaluation Findings

Common Development and Humanitarian Evaluation Assessment Criteria
Criteria(1) Highly Unsatisfactory(2) Unsatisfactory(3) Satisfactory(4) Highly Satisfactory
1. Relevance
1.1 Multilateral organization supported programs and projects are suited to the needs and/or priorities of the target groupEvaluation finds that substantial elements of program or project activities and outputs were unsuited to the needs and priorities of the target group.Evaluation finds that no systematic analysis of target group needs and priorities took place during the design phase of developmental or relief and rehabilitation programming or the evaluation report indicates some evident mismatch between program and project activities and outputs and the needs and priorities of the target group.Evaluation finds that the multilateral organization supported activity, program or project is designed taking into account the needs of the target group as identified through a process of situation or problem analysis (including needs assessment for relief operations) and that the resulting activities are designed to meet the needs of the target group.Evaluation finds methods used in program and project humanitarian and development (including needs assessment for relief operations) to identify target group needs and priorities (including consultations with target group members) and finds that the program and project takes those needs into account and is designed to meet those needs and priorities (whether or not it does so successfully).
1.2 Multilateral organization supported humanitarian and development projects and programs align with national humanitarian and development goals:The evaluation reports that significant elements of multilateral organization supported humanitarian and development program and project activity run counter to national humanitarian and development priorities with a resulting loss of effectiveness.The evaluation reports a significant portion (1/4 or more) of the multilateral organization supported humanitarian and development programs and projects subject to the evaluation are not aligned with national plans and priorities but there is no evidence that they run counter to those priorities.Most multilateral organization supported humanitarian and development programs and projects are reported in the evaluation to be fully aligned with national plans and priorities as expressed in national poverty eradication and sector plans and priorities. Wherever multilateral organization supported programs and projects are reported in the evaluation as not directly supportive of national plans and priorities they do not run counter to those priorities.All multilateral organization supported humanitarian and development projects and programs subject to the evaluation are reported in the evaluation to be fully aligned to national humanitarian and development goals as described in national and sector plans and priorities, especially including the national poverty eradication strategy and sector strategic priorities.
1.3 Multilateral organization has developed an effective partnership with governments, bilateral and multilateral development and humanitarian organizations and non-governmental organizations for planning, coordination and implementation of support to development and/or emergency preparedness, humanitarian relief and rehabilitation efforts.Evaluation finds that the multilateral organization experiences significant divergence in priorities from those of its (government, non-governmental organization or donor) partners and lacks a strategy or plan which will credibly address the divergence and which should result in strengthened partnership over time.Evaluation finds that multilateral organization has experienced significant difficulties in developing an effective relationship with partners and that there has been significant divergence in the priorities of the multilateral organization and its partners.Evaluation finds that multilateral organization has improved the effectiveness of its partnership relationship with partners over time during the evaluation period and that this partnership was effective at the time of the evaluation or was demonstrably improved.Evaluation finds that multilateral organization has consistently achieved a high level of partnership during the evaluation period.
2. Achievement of Humanitarian and Development Objectives and Expected Results
2.1 MO supported programs and projects achieve their stated development and/or humanitarian objectives and attain expected results.Less than half of stated output and outcome objectives have been achieved including one or more very important output and/or outcome level objectives.Half or less than half of stated output and outcome level objectives are achieved.MO supported programs and projects either achieve at least a majority of stated output and outcome objectives (more than 50% if stated) or that the most important of stated output and outcome objectives are achieved.MO supported programs and projects achieve all or almost all significant development and/or humanitarian objectives at the output and outcome level.
2.2 MO supported programs and projects have resulted in positive benefits for target group members.Problems in the design or delivery of MO supported activities mean that expected positive benefits for target group members have not occurred or are unlikely to occur.MO supported projects and programs result in no or very few positive changes experienced by target group members. These benefits may include the avoidance or reduction of negative effects of a sudden onset or protracted emergency.MO supported projects and programs have resulted in positive changes experienced by target group members (at the individual, household or community level). These benefits may include the avoidance or reduction of negative effects of a sudden onset or protracted emergency.MO supported projects and programs have resulted in widespread and significant positive changes experienced by target group members as measured using either quantitative or qualitative methods (possibly including comparison of impacts with non-program participants). These benefits may include the avoidance or reduction of negative effects of a sudden onset or protracted emergency.
2.3 MO programs and projects made differences for a substantial number of beneficiaries and where appropriate contributed to national development goals.MO supported projects and programs have not contributed to positive changes in the lives of beneficiaries as measured quantitatively or qualitatively.MO supported projects and programs have contributed to positive changes in the lives of only a small number of beneficiaries (when compared to project or program targets and local or national goals if established).MO supported projects and programs have contributed to positive changes in the lives of substantial numbers of beneficiaries as measured quantitatively or qualitatively. These may result from development, relief, or protracted relief and rehabilitation operations and may include the avoidance of negative effects of emergencies.MO supported projects and programs have contributed to positive changes in the lives of substantial numbers of beneficiaries. Further, they have contributed to the achievement of specific national development goals or have contributed to meeting humanitarian relief objectives agreed to with the national government and/or national and international development and relief organizations.
2.4 MO activities contributed to significant changes in national development policies and programs (including for disaster preparedness, emergency response and rehabilitation) (policy impacts) and/or to needed system reforms.National policies and programs in a given sector or area of development (including disaster preparedness, emergency response and rehabilitation) were deficient and required strengthening but MO activities have not addressed these deficiencies.MO activities have not made a significant contribution to the development of national policies and programs in a given sector or area of development, disaster preparedness, emergency response or rehabilitation. (Policy changes in humanitarian situations may include allowing access to the effected populations).MO activities have made a significant contribution to either re-orienting or sustaining effective national policies and programs in a given sector or area of development disaster preparedness, emergency response or rehabilitation.MO activities have substantial contribution to either re-orienting or sustaining effective national policies and programs in a given sector or area of development disaster preparedness, emergency response or rehabilitation. Further, the supported policies and program implementation modalities have resulted in improved positive impacts for target group members.
3. Cross Cutting Themes: Inclusive Humanitarian Assistance and Development Which can be Sustained
3.1 Extent MO supported activities effectively address the cross-cutting issue of gender equality.MO supported activities are unlikely to contribute to gender equality or may in fact lead to increases in gender inequalities.MO supported activities either lack gender equality objectives or achieve less than half of their stated gender equality objectives. (Note: where a program or activity is clearly gender-focused (maternal health programming for example) achievement of more than half its stated objectives warrants a satisfactory rating).MO supported programs and projects achieve a majority (more than 50%) of their stated gender equality objectives.MO supported programs and projects achieve all or nearly all of their stated gender equality objectives.
3.2 Extent changes are environmentally sustainable.MO supported programs and projects do not include planned activities or project design criteria intended to promote environmental sustainability. In addition changes resulting from MO supported programs and projects are not environmentally sustainable.MO supported programs and projects do not include planned activities or project design criteria intended to promote environmental sustainability. There is, however, no direct indication that project or program results are not environmentally sustainable. OR MO supported programs and projects include planned activities or project design criteria intended to promote sustainability but these have not been successful.MO supported programs and projects include some planned activities and project design criteria to ensure environmental sustainability. These activities are implemented successfully and the results are environmentally sustainable.MO supported programs and projects are specifically designed to be environmentally sustainable and include substantial planned activities and project design criteria to ensure environmental sustainability. These plans are implemented successfully and the results are environmentally sustainable.
4. Sustainability
4.1 Benefits continuing or likely to continue after project or program completion or there are effective measures to link the humanitarian relief operations, to rehabilitation, reconstructions and, eventually, to longer-term developmental results.There is a very low probability that the program/project will result in continued intended benefits for the target group after project completion. For humanitarian relief operations, the evaluation finds no strategic or operational measures to link relief, to rehabilitation, reconstruction and, eventually, to development.There is a low probability that the program/project will result in continued benefits for the target group after completion. For humanitarian relief operations, efforts to link the relief phase to rehabilitation, reconstruction and, eventually, to development are inadequate. (Note, in some circumstances such linkage may not be possible due to the context of the emergency. If this is stated in the evaluation, a rating of satisfactory can be given)Likely that the program or project will result in continued benefits for the target group after completion. For humanitarian relief operations, the strategic and operational measures to link relief to rehabilitation, reconstruction and, eventually, development are credible.Highly likely that the program or project will result in continued benefits for the target group after completion. For humanitarian relief operations, the strategic and operational measures to link relief to rehabilitation, reconstruction and, eventually, development are credible. Further, they are likely to succeed in securing continuing benefits for target group members.
4.2 Extent MO supported projects and programs are reported as sustainable in terms of institutional and/or community capacity.The design of MO supported programs and projects failed to address the need to strengthen institutional and/or community capacity as required. In the case of humanitarian operations, the design of programs and projects failed to take account of identified needs to strengthen local capacities for delivery of relief operations and/or for managing the transition to rehabilitation and/or development.MO programs and projects may have failed to contribute to strengthening institutional and/or community capacity or, where appropriate, to strengthen local capacities for delivery of relief operations and/or for managing the transition to rehabilitation and/or development.MO programs and projects have contributed to strengthening institutional and/or community capacity but with limited success.Either MO programs and projects have contributed to significantly strengthen institutional and/or community capacity as required or institutional partners and communities already had the required capacity to sustain program results.
4.3 Extent MO development programming contributes to strengthening the enabling environment for development.For development programs, there were important weaknesses in the enabling environment for development (the overall framework and process for national development planning; systems and processes for public consultation and for participation by civil society in development planning; governance structures and the rule of law; national and local mechanisms for accountability for public expenditures, service delivery and quality; and necessary improvements to supporting structures such as capital and labour markets). Further, the MO activities and support provided to programs and projects failed to address the identified weakness successfully, further limiting program results.MO development activities and/or MO supported projects and programs have not made a notable contribution to changes in the enabling environment for development.MO development activities and/or MO supported projects and programs have made a notable contribution to changes in the enabling environment for development including one or more of: the overall framework and process for national development planning; systems and processes for public consultation and for participation by civil society in development planning; governance structures and the rule of law; national and local mechanisms for accountability for public expenditures, service delivery and quality; and necessary improvements to supporting structures such as capital and labour markets.MO development activities and/or MO supported projects and programs have made a significant contribution to changes in the enabling environment for development including one or more of: the overall framework and process for national development planning; systems and processes for public consultation and for participation by civil society in development planning; governance structures and the rule of law; national and local mechanisms for accountability for public expenditures, service delivery and quality; and necessary improvements to supporting structures such as capital and labour markets. Further, these improvements in the enabling environment are leading to improved development outcomes.
5. Efficiency
5.1 Program activities are evaluated as cost/resource efficient:Credible information indicating that MO supported programs and projects (development, emergency preparedness, relief and rehabilitation) are not cost/resource efficient.MO supported programs and projects under evaluation (development, emergency preparedness, relief and rehabilitation) do not have credible, reliable information on the costs of activities and inputs and therefore the evaluation is not able to report on cost/resource efficiency. OR MO supported programs and projects under evaluation present mixed findings on the cost/resource efficiency of the inputs.Level of program outputs achieved (development, emergency preparedness, relief and rehabilitation) when compared to the cost of program activities and inputs are appropriate even when the program design process did not directly consider alternative program delivery methods and their associated costs.MO supported (development, emergency preparedness, relief and rehabilitation) programs and projects are designed to include activities and inputs that produce outputs in the most cost/resource efficient manner available at the time.
5.2 Evaluation indicates implementation and objectives achieved on time (given the context, in the case of humanitarian programming)Less than half of stated output and outcome level objectives of MO supported programs and projects are achieved on time, there is no credible plan or legitimate explanation found by the evaluation which would suggest significant improvement in on-time objectives achievement in the future.Less than half of stated output and outcome level objectives of MO supported programs and projects are achieved on time but the program or project design has been adjusted to take account of difficulties encountered and can be expected to improve the pace of objectives achievement in the future. In the case of humanitarian programming, there was a legitimate explanation for the delays.More than half of stated output and outcome level objectives of MO supported programs and projects are achieved on time and that this level is appropriate to the context faced by the program during implementation, particularly for humanitarian programming.Nearly all stated output and outcome level objectives of MO supported programs and projects are achieved on time or, in the case of humanitarian programming, a legitimate explanation for delays in the achievement of some outputs/outcomes is provided.
5.3 Evaluation indicates that MO systems and procedures for project/program implementation and follow up are efficient (including systems for engaging staff, procuring project inputs, disbursing payment, logistical arrangements etc.)Serious deficiencies in agency systems and procedures for project/program implementation that result in significant delays in project start-up, implementation or completion and/or significant cost increases.Some deficiencies in agency systems and procedures for project/program implementation but does not indicate that these have contributed to delays in achieving project/program objectives.Agency systems and procedures for project implementation are reasonably efficient and have not resulted in significant delays or increased costs.Efficiency of agency systems and procedures for project implementation represent an important organizational strength in the implementation of the program under evaluation.  
6. Using Evaluation and Monitoring to Improve humanitarian and development Effectiveness
6.1 Systems and process for evaluation are effective.Evaluation practices in use for programs and projects of this type (development, emergency preparedness, relief and rehabilitation) are seriously deficient.No indication that programs and projects of this type (development, emergency preparedness, relief and rehabilitation) are subject to systematic and regular evaluations.Program being evaluated is subject to systematic and regular evaluations or describes significant elements of such practice. No mention of policy and practice regarding similar programs and projects. This may include specialized evaluation methods and approaches to emergency preparedness, relief and rehabilitation programming.Program being evaluated (along with similar programs and projects) is subject to systematic regular evaluations or describes significant elements of such practice.
6.2 Systems and processes for monitoring and reporting on program results are effectiveAbsence of monitoring and reporting systems for the development and humanitarian programming. This would include the absence of adequate monitoring of outputs during the implementation of humanitarian programming.While monitoring and reporting systems for the development and humanitarian programming exist, they either do not report on a regular basis or they are inadequate in frequency, coverage or reliability.Monitoring and reporting systems for development and humanitarian programming as appropriate are well-established and report regularly.Monitoring and reporting systems for the program are well-established and report regularly. The quality of regular reports is rated highly by the evaluation and results are reportedly used in the management of the program.
6.3 Results Based Management (results-based management  ) systems are effectiveNo evidence of the existence of a results-based management   system for the program and no system is being developed.While a results-based management   system is in place, or being developed, it is unreliable and does not produce regular reports on program performance.Results-based management   system is in place and produces regular reports on program performance.Results-based management   system is in place for the program and there is evidence noted in the evaluation that the system is used to make changes in the program to improve effectiveness.
6.4 MO makes use of evaluation to improve development/humanitarian effectivenessReport does not include a management response and does not have one appended to it or associated with it. There is no indication of how the evaluation results will be used. There is no indication that similar evaluations have been used to improve effectiveness in the past.Report includes a management response (or has one attached or associated with it) but it does not indicate which recommendations have been accepted.
OR
There is some, non-specific indication that similar evaluations have been used to improve program effectiveness in the past.
Report includes a management response (or has one attached or associated with it) that indicates which recommendations have been accepted.
OR
There is a clear indication that similar evaluations in the past have been used to make clearly identified improvements in program effectiveness.
Report includes a management response (or has one attached or associated with it) describes a response to each major recommendation which is appropriate and likely to result in the organizational and programmatic changes needed to achieve their intent.

Annex 5: Global Evaluations and UNICEF Documents

UNICEF Global Evaluations

Other Documents

UNICEF Strategy and Policy Documents

UNICEF Reports

UNICEF Monitoring and Evaluation Documents

Other UNICEF Documents

Global Evaluation Report Oversight System (GEROS) Documents

Documents from Other Organizations

Annex 6: Selected Results for Comparisons of Least-Developed and Middle-Income Countries and by Focus Area and Humanitarian Action

Comparison of Least-developed and Middle-income Countries

1.3 Effective partnerships
Country Type(4)
Highly Satisfactory
(3)
Satisfactory
(2) Unsatisfactory(1)
Highly Unsatisfactory
TotalValid CasesMean ScoreTotal CasesSatisfactory or Better
Least-developed country28.1%53.1%9.4%9.4%100.0%323.003481%
Middle-income country27.3%68.2%4.5%.0%100.0%223.232695%
2.1 Programs achieve their state objectives
Country Type(4)
Highly Satisfactory
(3)
Satisfactory
(2) Unsatisfactory(1)
Highly Unsatisfactory
TotalValid CasesMean ScoreTotal CasesSatisfactory or Better
Least-developed country15.2%51.5%27.3%6.1%100.0%332.7634.0067%
Middle-income country19.2%69.2%3.8%7.7%100.0%26326.0088%
4.2 Institutional and community capacity
Country Type(4)
Highly Satisfactory
(3)
Satisfactory
(2) Unsatisfactory(1)
Highly Unsatisfactory
TotalValid CasesMean ScoreTotal CasesSatisfactory or Better
Least-developed country9.1%39.4%33.3%18.2%100.0%332.283448%
Middle-income country16.0%40.0%40.0%4.0%100.0%252.682656%
5.2 Programs implemented/objectives achieved on time
Country Type(4)
Highly Satisfactory
(3)
Satisfactory
(2) Unsatisfactory(1)
Highly Unsatisfactory
TotalValid CasesMean ScoreTotal CasesSatisfactory or Better
Least-developed country10.0%50.0%20.0%20.0%100.0%202.503460%
Middle-income country.0%41.2%41.2%17.6%100.0%172.242641%
Criteria 1.1 Suited to the needs of target group members
Focus AreaHighly Un-satisfactoryUn-satisfactorySatisfactoryHighly SatisfactoryNot AddressedTotalAddressed
1. Young Child Survival and Development1011812120
2. Basic Education and Gender Equality008211110
3. HIV/AIDS and Children0012033
4. Child Protection from Violence, Exploitation and Abuse0042066
5. Policy Advocacy and Partnership0043077
6. Humanitarian125501313
7. Country Program0100011
Total23332226260
% of All Valid Cases Excluding Not Addressed3%5%55%37% 100% 
Criteria 1.3 Effective partnerships
Focus AreaHighly Un-satisfactoryUn-satisfactorySatisfactoryHighly SatisfactoryNot AddressedTotalAddressed
1. Young Child Survival and Development109832118
2. Basic Education and Gender Equality126111110
3. HIV/AIDS and Children0021033
4. Child Protection from Violence, Exploitation and Abuse0213066
5. Policy Advocacy and Partnership0041275
6. Humanitarian109301313
7. Country Program0 10011
Total34321766256
% of All Valid Cases Excluding Not Addressed5%7%57%30% 100% 
Criteria 6.1 Systems for evaluation effective
Focus AreaHighly Un-satisfactoryUn-satisfactorySatisfactoryHighly SatisfactoryNot AddressedTotalAddressed
1. Young Child Survival and Development155462115
2. Basic Education and Gender Equality00515116
3. HIV/AIDS and Children0020132
4. Child Protection from Violence, Exploitation and Abuse1210264
5. Policy Advocacy and Partnership0240176
6. Humanitarian03235138
7. Country Program01 0011
Total213198206242
% of All Valid Cases Excluding Not Addressed5%31%45%19% 100% 
Criteria 6.2 Systems for monitoring and reporting results effective
Focus AreaHighly Un-satisfactoryUn-satisfactorySatisfactoryHighly SatisfactoryNot AddressedTotalAddressed
1. Young Child Survival and Development348512120
2. Basic Education and Gender Equality262101111
3. HIV/AIDS and Children0210033
4. Child Protection from Violence, Exploitation and Abuse1410066
5. Policy Advocacy and Partnership1411077
6. Humanitarian083021311
7. Country Program01 0011
Total72916736259
% of All Valid Cases Excluding Not Addressed12%49%27%12% 100% 

Annex 7: UNICEF Staff Interviewed
(Development Effectiveness Review)

Evaluation Office

Division of Financial and Administrative Management

Division of Governance UN and Multilateral Affairs

Division of Policy and Strategy

Office of Emergency Programs

Programme Division

Public Sector Alliances and Resource Mobilization

Annex 8: Data sources for Section 5.0
(Canada's Relationship with UNICEF)

Documents Consulted

DFATD Interviewees

UNICEF Interviewees

Other Interviewees

Annex 9: Management Response from DFATD's Global Issues and Development Branch

The Canadian Chapter of the Development Effectiveness Review of the United Nations Children's Fund (UNICEF), prepared by Foreign Affairs, Trade and Development (DFATD) Canada's Evaluation Directorate, provides an overview of Canada's relationship with UNICEF.  It focuses on Canada's international development priorities and strategic objectives for engagement with the institution. It presents evidence of the relevance, efficiency and effectiveness of this relationship, along with evidence-based guidance on Canada's future engagement with the institution.

UNICEF's work is carried out in more than 190 countries through country programmes and national committees, including all of DFATD's countries of focus.   The institution contributes to the achievement of numerous Canadian development priorities including Increasing Food Security: Canada's Food Security Strategy, Securing the Future of Children and Youth and Canada's Muskoka Initiative for Maternal, Newborn and Child Health.  As well, UNICEF is a key partner in Canada's peace and security and humanitarian goals and advancing Canadian human rights values

The DFATD Global Issues and Development Branch agrees with the majority of the recommendations made by the Evaluation Directorate informed from its assessment of Canada's engagement with UNICEF. Namely, continuing to emphasize the integration of gender equality within UNICEF at all levels and the importance of performance measurement and reporting in this area; continued work with UNICEF to foster collaboration with other UN organizations; encouraging UNICEF to improve its sustainability programming; and, encouraging UNICEF to increase  its evaluation of its humanitarian programming.  However, with regard to the recommendation related to the development of a performance measurement framework, it is felt that this recommendation should reflect a multi-donor approach, consistent with the principles of aid effectiveness.  Consistent with this view, DFATD welcomes the acknowledgement by the Directorate that Canada is one of several donors working with UNICEF at the institutional level and is limited in the extent to which it alone can influence improvements in the development effectiveness of the institution.  As such, DFATD will continue to work with other donors to advocate for improvements within UNICEF. 

The Global Issues and Development Branch will take concrete action as outlined in the table below, to address the Chapter's recommendations.  It will use the opportunities made available through its work as a member of UNICEF's Executive Board and its ongoing dialogue with the institution.  It will also reach out to other parts of DFATD in an effort to ensure a more coherent approach in the Department's interaction with UNICEF on these issues, both at Headquarters, in the field, and in Canada's Permanent Mission in New York.

Management Response
RecommendationsCommitments/measuresResponsibleCompletion dateStatus

1. DFATD should continue to emphasize the integration of gender equality in UNICEF programs and should highlight the importance of performance measurement and reporting in this area.   

Agree.  DFATD has prioritized the integration of gender equality in its engagement with UNICEF.  Most recently, Canada advocated for the inclusion of gender equality results in all focus areas and at all levels in UNICEF's 2014-2017 Strategic Plan (SP).  However, despite efforts by Canada and other donors, gender equality was not adequately integrated.

As a result, Canada and other donors on the UNICEF Executive Board have requested that the upcoming 2014-2017 Gender Equality Action Plan (GEAP) address the shortcomings of the Strategic Plan. 

1.1 Canada will continue to work with other donors and engage with UNICEF throughout the development of their 2014-2017 Gender Equality Action Plan to recommend that it is costed with dedicated resources, contains a comprehensive gender analysis, is aligned with system-level gender mainstreaming efforts, and includes a performance measurement framework.

1.1 MFM/Global Institutions Bureau

1.1 June 2014

1.2 Canada will monitor future reporting on UNICEF's 2014-2017 Strategic Plan, the 2014-2017 Integrated Budget and its 2014-2017 Gender Equality Action Plan, with particular attention paid to the first progress report for each document.  Canada will make concrete suggestions for improvement to each document on reporting formats through National and joint statements, bilateral meetings and additional interventions at the Executive Board.

1.2 MFM/Global Institutions Bureau

1.2 June 2015

 2. DFATD should develop a performance measurement framework to assess progress on Canada's strategic objectives for its relationship with UNICEF.

Partially agree. In the context of aid effectiveness principles, it would be more effective to develop a common performance measurement framework for all donors, rather than focusing on Canada's bilateral engagement.

2.1. The program will engage with other donors to review joint strategic objectives for engagement with UNICEF and work to define expected outcomes going forward.  

2.1. MFM/Global Institutions Bureau

2.1  December 2015

3. As Canada and other members of the General Assembly are seeking mechanisms to harmonize the UN development system, DFATD should continue to work with UNICEF to identify realistic and tangible ways to collaborate with other UN organizations.

 Agree.  This recommendation is consistent with the work already being undertaken by UNICEF, with input from DFATD and other donors, to improve harmonization and collaboration between UN Funds and Programs. 

3.1 Canada will continue its efforts to encourage  UNICEF to identify realistic and tangible ways to collaborate with other UN organizations.  This will include working with other donors and a specific focus on the following:

  • greater efficiency within the UN system as a whole
  • monitoring the implementation of strategic guidance from the QCPR within the  UN development system through our participation within the ECOSOC and GA processes
  • monitoring the implementation of the Standard Operating Procedures for the Delivering as One approach in UN operations, including through harmonized Country Program Documents (CPD) at country level.

3.1 MFM/Global Institutions Bureau

3.1. No end date as these are the responsibilities required in the ongoing management of Canada's institutional relationship with UNICEF.

4. DFATD should encourage UNICEF to improve its sustainability planning. This could involve efforts to strengthen UNICEF's program design and implementation plans.

Agree. DFATD recognizes the complex factors involved in fostering sustainable development and the need for involvement of a variety of stakeholders.

4.1. The key entry point for Canada to encourage UNICEF to continue to improve the sustainability of its programming at the institutional level is through our participation on UNICEF's Executive Board. DFATD will:

  • review the Result Framework of UNICEF Strategic Plan from a sustainability perspective (June 2014)
  • closely monitor the implementation of UNICEF's commitments through its annual reporting to the Executive Board; and,
  • engage in the oversight of the institution's risk management practices and review program-specific sources of risk. 

4.1. MFM/Global Institutions Bureau

4.1. No end date as these are the responsibilities required in the ongoing management of Canada's institutional relationship with UNICEF.

5. DFATD should encourage UNICEF to increase its outcome-level reporting, analysis and evaluation of humanitarian assistance programming.

Agree.  Canada recognizes that UNICEF's 2014-2017 Strategic Plan reflects the institution's attempt at better integrating its humanitarian assistance programming within its institutional strategic planning process.   This recommendation reflects a natural next step in the process.
DFATD will:

5.1. Engage with UNICEF's Office of Emergency Programmes to better understand and encourage an increased focus on outcome-level reporting and analysis.

5.1 MFM/International Humanitarian Assistance Bureau 

5.1. June 2015

5.2. Reinforce the importance of UNICEF's role in humanitarian assistance by advocating for greater coverage at the Executive Board meetings, and increased evaluations of humanitarian assistance programming.

5.2. MFM/Global Institutions Bureau  

5.2. No end date as these are the responsibilities required in the ongoing management of Canada's institutional relationship with UNICEF.

Date Modified: