86
i | Page THE REPUBLIC OF UGANDA MINISTRY OF HEALTH Monitoring and Evaluation Plan for Implementation of the Health Sector Development Plan 2015/16 - 2019/20 May 2016

Monitoring and Evaluation Plan for Implementation of the

  • Upload
    others

  • View
    1

  • Download
    0

Embed Size (px)

Citation preview

i | P a g e

THE REPUBLIC OF UGANDA

MINISTRY OF HEALTH

Monitoring and Evaluation Plan for

Implementation of the

Health Sector Development Plan

2015/16 - 2019/20

May 2016

Any part of this document may be freely reviewed, quoted, reproduced or translated in full or in part, provided the source is acknowledged. It may not be sold or used in conjunction with commercial purposes or for profit. Government of Uganda, Ministry of Health: Monitoring & Evaluation Plan for the Health Sector Development Plan, 2015/16 – 2019/20. Published by: Ministry of Health PO Box 7272 Kampala, Uganda Email: [email protected] Website: www.health.go.ug

iii | P a g e

Table of Contents

TABLE OF CONTENTS .......................................................................................................................................III

ACRONYMS ..................................................................................................................................................... V

DEFINITIONS ................................................................................................................................................. VII

FOREWORD ..................................................................................................................................................... 8

ACKNOWLEDGEMENTS .................................................................................................................................... 9

EXECUTIVE SUMMARY ................................................................................................................................... 10

1 INTRODUCTION ..................................................................................................................................... 12

1.1 GOALS AND OBJECTIVES OF THE HSDP 2015/16 - 2019/20 ........................................................................... 12

1.2 GOAL AND OBJECTIVES OF THE M&E PLAN FOR THE HSDP.............................................................................. 13

1.3 SITUATIONAL ANALYSIS OF THE M&E PRACTICES AND MECHANISM .................................................................... 13

1.4 PROCESS FOR DEVELOPMENT OF THE M&E PLAN ........................................................................................... 16

1.5 DISEASE AND PROGRAM SPECIFIC M&E ALIGNMENT ....................................................................................... 17

2 INSTITUTIONAL CAPACITY ..................................................................................................................... 18

2.1 M&E COORDINATION MECHANISMS ........................................................................................................... 19

2.2 ROLES AND RESPONSIBILITIES ..................................................................................................................... 20

2.3 CAPACITY BUILDING STRATEGY ................................................................................................................... 26

3 MONITORING AND EVALUATION FRAMEWORK .................................................................................... 28

3.1 THE M&E FRAMEWORK ............................................................................................................................ 28

3.2 M&E EVENTS ......................................................................................................................................... 29

3.3 INDICATORS ............................................................................................................................................ 29

3.4 DATA SOURCES ........................................................................................................................................ 36

3.5 DATA MANAGEMENT ................................................................................................................................ 37

3.6 DATA DISSEMINATION AND USE ................................................................................................................... 44

3.7 EVALUATION ........................................................................................................................................... 45

3.8 SURVEYS AND SURVEILLANCE ...................................................................................................................... 46

4 PROGRESS AND PERFORMANCE REVIEWS ............................................................................................. 48

4.1 PROGRESS REVIEW ................................................................................................................................... 50

4.2 ANNUAL REVIEW ...................................................................................................................................... 52

4.3 MID TERM REVIEW .................................................................................................................................. 53

4.4 END TERM REVIEW .................................................................................................................................. 53

4.5 LINKAGES BETWEEN HEALTH SECTOR REVIEWS, PROGRAM / PROJECT SPECIFIC REVIEWS, AND GLOBAL REPORTING ....... 54

4.6 DECISION-MAKING PROCESSES AND TOOLS FOR REMEDIAL ACTION ..................................................................... 55

5 IMPLEMENTATION PLAN ....................................................................................................................... 56

6 MONITORING THE IMPLEMENTATION OF THE M&E PLAN FOR THE HSDP ............................................. 58

6.1 M&E PLAN IMPLEMENTATION MONITORING INDICATORS ............................................................................... 59

7 BUDGET ................................................................................................................................................. 60

8 ANNEX ...................................................................................................................................................... I

8.1 QUARTERLY OUTPUT REPORT FORMAT FOR THE WORKPLANS & BUDGETS ............................................................. I

8.2 DETAILED 5 YEAR BUDGET FOR IMPLEMENTATION OF THE PLAN ........................................................................... I

8.3 DISTRICT LEAGUE TABLE INDICATORS ............................................................................................................ IV

8.4 INDICATOR DEFINITIONS FOR THE HSDP KEY PERFORMANCE INDICATORS ............................................................. V

8.5 INDICATOR DEFINITIONS FOR THE ADDITIONAL INDICATORS IN THE DLT ............................................................ XXII

8.6 CYP CONVERSION FACTORS .................................................................................................................... XXIV

v | P a g e

Acronyms AHSPR Annual Health Sector Performance Report

BFP Budget Framework Paper

CAO Chief Administrative Officer

CCM Country Coordinating Mechanism

CHeSS Country Health Systems Surveillance

CHMIS Community Health Management Information System

CHEW Community Health Extension Worker

CPD Continuous Professional Development

CSO Civil Society Organisation

DHIS District Health Information System

DHMT District Health Management Team

DHO District Health Office

DHT District Health Team

DQA Data Quality Assessment

DQAA Data Quality Assessment and Adjustment

GAVI Global Alliance for Vaccines and Immunisation

GoU Government of Uganda

HC Health Center

HDP Health Development Partners

HID Health Information Division

HMIS Health Management Information System

HRIS Human Resource Information System

HSD Health Sub-District

HSDP Health Sector Development Plan

HSSIP Health Sector Strategic and Investment Plan

HPAC Health Policy Advisory Committee

HUMC Health Unit Management Committee

ICD International Classification of Diseases

ICT Information and Communication Technology

IFMS Integrated Financial Management System

IHP+ International Health Partnerships and related Initiatives

IDSR Integrated Disease Surveillance and Response

JAF Joint Assessment Framework

JRM Joint Review Mission

KM Knowledge Management

LC Local Council

LG Local Government

MDAs Ministries, Departments and Agencies

MDGs Millennium Development Goals

MPS Ministerial Policy Statement

M&E Monitoring and Evaluation

MoH Ministry of Health

MoFPED Ministry of Finance, Planning and Economic Development

MoLG Ministry of Local Government

MTR Mid Term Review

NCD Non-Communicable Diseases

NDP National Development Plan

NHA National Health Accounts

NHP National Health Policy

NIRA National Identification & Registration Authority

NSDS National Service Delivery Survey

NMS National Medical Stores

NPA National Planning Authority

NRH National Referral Hospital

OBT Output Budgeting Tool

OPM Office of the Prime Minister

PHP Private Health Practitioners

PNFP Private-Not-For-Profit Organizations

QAD Quality Assurance Department

RPMT Regional Performance Monitoring Team

RRH Regional Referral Hospital

SCMS Supply Chain Management System

SDG Sustainable Development Goal

SMC Senior Management Committee

SME&R Supervision, Monitoring, Evaluation and Research

SOP Standard Operating Procedure

SWAp Sector Wide Approach

TWG Technical Working Group

UBOS Uganda Bureau of Statistics

UDHS Uganda Demographic Health Survey

UNHRO Uganda National Health Research Organisation

UNPS Uganda National Panel Survey

VHT Village Health Team

WHO World Health Organisation

vii | P a g e

Definitions

Civil Society Organization Is the aggregate of non-governmental organizations and institutions

that manifest interests and will of citizens.

Data Management Comprises all processes related to data collection, analysis,

synthesis and dissemination.

Data Quality Assurance Is the process of profiling the data to discover inconsistencies and

other anomalies in the data, as well as performing data cleansing

activities (e.g. removing outliers, missing data interpolation) to

improve the data quality.

Evaluation Rigorous, science-based analysis of information about program

activities, characteristics, outcomes and impact that determines the

merit or worth of a specific program or intervention.

IHP+ Is a group of partners who share a common interest in improving

health services and health outcomes by putting the principles of the

Paris Declaration on Aid Effectiveness into practice.

Impact Fundamental intended or unintended changes in the conditions of

the target group, population, system or organization.

Knowledge Management Is a set of principles, tools and practices that enable people to create

knowledge, and to share, translate and apply what they know to

create value and improve effectiveness.

Monitoring The routine tracking, analysis and reporting of priority information

about a program and its intended outputs and outcomes.

Outcome Actual or intended changes in use, satisfaction levels or behaviour

that a planned intervention seeks to support.

Performance The extent to which relevance, effectiveness, efficiency, economy,

sustainability and impact (expected and unexpected) are achieved

by an initiative, program or policy.

Review Organizational processes for assessing health systems progress and

performance.

Foreword The Ministry of Health launched the Health Sector Development Plan 2015/16 – 2019/20

which defines the medium term health agenda and operationalize Uganda’s aspirations as

outlined in the second National Development Plan. The goal of this Monitoring & Evaluation

(M&E) Plan is to contribute to the attainment of the HSDP 2015/16 - 2019/20 goal,

objectives and performance targets by providing a health sector-wide framework for regular

and systematic tracking and documenting progress in the health sector performance.

The development of M&E Plan for HSDP 2015/16 – 2019/20 has been largely informed by

lessons from the Mid Term Review of the Health Sector Strategic and Investment Plan

(HSSIP) 2010/11 - 2014/15 in addition to a rapid assessment conducted to provide insight

into the current M&E practices and mechanisms. The process of development of this plan was

highly consultative, participatory and transparent. The Supervision Monitoring Evaluation

& Research Technical Working Group comprising of MoH officials, Health Development

Partners, Civil Society, UBOS, Medical Bureaus, private sector and academia was responsible

for overseeing the development of the M&E plan.

It is envisaged that this comprehensive M&E plan to which all health partners subscribe

under the IHP+ arrangements, shall be the basis for improving the quality of routine

information systems and be used to institutionalize mechanisms and tools for measuring

quality of both facility and community based services. It should also strengthen dissemination

and use of information at national, sub national and global levels.

I wish to express my appreciation to all of you who worked tirelessly to develop the M&E

Plan for HSDP on behalf of the people of Uganda. I look forward to the acceleration of the

implementation of this Plan and hold each other accountable for the resources provided for

implementation of the HSDP.

For God and My Country

Dr. Jane Ruth Aceng

MINISTER OF HEALTH

9 | P a g e

Acknowledgements

The Ministry of Health would like to express its appreciation to all key stakeholders who

supported the process of developing this Monitoring & Evaluation Plan for Implementation

of the Health Sector Development Plan (HSDP) 2015/16 - 2019/20 through technical and

financial support. The development process was consultative covering a wide range of

stakeholders at national and sub-national levels including private and public sectors.

The funding for the development of this plan was generously supported by Government of

Uganda, the WHO Country Office and the USAID Monitoring and Evaluation Technical

Support (METS) Project under the School of Public Health.

Special thanks go to the Ministry of Health Quality Assurance Department for taking the

lead in reviewing and developing this plan and the Supervision, Monitoring, Evaluation and

Research Technical Working Group for the valuable input.

All stakeholders in the health sector are thus encouraged to make use of this valuable

document to keep track of the implementation of the HSDP and make overall contribution to

the attainment of Universal Health Coverage goals.

Prof. Anthony Mbonye

AG. DIRECTOR GENERAL HEALTH SERVICES

Executive Summary The M&E Plan for the HSDP 2015/2016 – 2019/20 has been developed to operationalize the strategic

orientation provided for comprehensive M&E in HSDP 2015/2016 – 2019/20. It focuses on the main

M&E activities and aligns them to the existing national and international structures and frameworks.

This Plan describes the M&E framework, indicators, processes, sources of data, methods and tools

that the sector will use to collect, compile, report and use data, and provide feed-back as part of the

national Health Sector M&E mechanisms. It is intended to document what needs to be monitored,

with whom, by whom, when, how, and how the M&E data will be used. It also outlines how and when

the different types of studies and evaluations will be conducted by the sector. It translates these

processes into annualized and costed activities, and assigns responsibilities for implementation.

The specific objectives of the M&E Plan for the HSDP are;

1. To define mechanisms for assessment of the health sector performance in accordance with the

agreed objectives, performance indicators and reporting timelines for the HSDP.

2. To support management for results through evidence based decision-making, policy development

and advocacy; sector learning and improvement at all levels during implementation of the HSDP.

3. To enhance compliance with government policies (accountability), and constructive engagement

with stakeholders (policy dialogue) in the health sector.

4. To build capacity of the health sector to generate quality data and track progress of

implementation of the HSDP.

5. To facilitate continuous learning (document and share the challenges and lessons learnt) at all

levels during implementation of the HSDP.

A review of a wide range of national and health sector documents was done to provide in-depth

analysis and understanding of the sector M&E system; such as NDP II M&E Strategy, NHP II, M&E Plan

for the HSSIP 2010/11 – 2014/15, HSDP 2015/16 - 2019/20, Program specific M&E plans and the

Health Management Information System (HMIS) Manual, among others.

The sector-wide M&E system for effective tracking, evaluation and feedback on HSDP implementation

and results will be followed as defined in the NDP II M&E Strategy. This implies that Central

Government, MoH, LGs, HDPs, CSOs and other stakeholders will be involved directly or indirectly in

the M&E activities. The relevant institutions shall ensure availability of adequate skilled human

resources in all aspects of M&E at all levels. To ensure this, the sector shall develop/improve M&E

human resource capacity through training, providing guidelines and tools, supportive supervision and

mentoring.

The health sector shall apply the IHP+ common M&E Framework which provides a logical and results-

chain representation of the key components of the HSDP M&E. It comprises four major indicator

categories: system input and processes, outputs, outcomes, and impact (Figure 2). 41 key

performance indicators (see Table 3) were identified based on the priority diseases and the health

systems components addressed in the HSDP.

Specific programs and projects have their own strategic and operational plans that are more detailed

than the HSDP. They are also expected to develop specific program / project M&E Plans which will

include more indicators to measure process, outputs and immediate outcomes. These should

11 | P a g e

however, be aligned with the goal, objectives and strategic interventions of the HSDP to demonstrate

the effects of health system strengthening to health outcomes.

Special focus will be on the data management practices namely; collection, validation, feedback,

analysis / synthesis, dissemination and use. Specific efforts will be made to undertake Data Quality

Audits and routine (quarterly) data on a sample of districts among others. Regular data verification for

facility based data for accuracy and completeness will be carried out at least monthly by the health

facility in-charges. Data analysis and synthesis will be done at all levels (national, sub-national to

community) to enhance evidence-based decision-making. All relevant data will be synthesized based

on determined parameters (disaggregated) where applicable and analyzed for use at various levels of

the sector. Regular reporting using the HMIS and other sources of data will be used to generate

relevant information at all levels. The analytical reports will provide information for decision-making,

program management, financial disbursements and global reporting.

The health sector will undertake evaluation of the overall HSDP implementation as well as evaluation

of specific programs, projects, policies and interventions. The type of evaluation to be planned for and

conducted should reflect the nature and scope of the public investment.

During this HSDP period the following major surveys are planned; the UDHS-16, AIDS Behavioral

Survey, National Health Accounts and efficiency analyses. Each program will identify and include all

surveys to be conducted in the respective M&E Plans. Health Institutions are encouraged to conduct

operational research. Operational surveys shall be carried as planned on approval by the relevant

institutions and findings disseminated through the existing structures.

The HSDP monitoring and review process will be interlinked across the different planning entities. All

reviews will be presented and endorsed at the respective management and partnership structures for

the level. The outcomes of the review process will be used to guide decision making during planning,

resource mobilisation and allocation, accountability, policy dialogue, review and development.

Global reporting requirements shall be based on ongoing country processes. The common M&E

platform (See Figure 6) serves as the guide for all M&E related processes such as the health sector

component of the Sector Wide Approaches (SWAp) and IHP+, for program activities supported by

Global Funding Agencies e.g. GAVI, the Global Fund for TB, AIDs and Malaria, and other DPs, and for

disease and program / project specific needs.

Over the 5-year period, an estimated total of UGX 89,653,447,000/- will be required for

implementation of this plan. Of this 35.4% is required for compilation and submission of performance

reports, 31.5% for performance reviews and 23% for surveys (Table 8).

The implementation plan is annualized for the key M&E activities, showing the responsible persons or

departments and frequency of the activities (Table 7).

1 Introduction

The purpose of the sector Monitoring and Evaluation (M&E) system is to coordinate and support the

Ministry of Health (MoH), related Ministries, Health Development Partners (HDPs), Local

Governments (LGs) and stakeholders to regularly and systematically track progress of implementation

of priority interventions of the sector plans and assess performance of the sector and LGs in

accordance with the agreed objectives and performance indicators over the Medium Term (output,

outcome and impact monitoring). In order to do this there is need for an M&E system that provides

timely and accurate information to government and partners in order to inform performance reviews,

policy discussions and periodic revisions to the national strategic and operational plans.

The health sector carries out three types of monitoring which address different stages in the results

chain, namely;

(i) Financial implementation monitoring addresses whether or not budgets have been released

and spent in line with allocations;

(ii) Physical implementation monitoring addresses whether activities have taken place in line

with targets; and

(iii) Outputs, outcome and impact monitoring trace whether or not results are occurring

amongst the target population.

The M&E Plan for the Health Sector Development Plan (HSDP) 2015/2016 – 2019/20 has been

developed to operationalize the strategic orientation provided for comprehensive M&E in

HSDP2015/2016 – 2019/20. It focuses on the main M&E activities and aligns them to the existing

national and international structures and frameworks.

This Plan describes the M&E framework, indicators, processes, sources of data, methods and tools

that the sector will use to collect, compile, report and use data, and provide feed-back as part of the

national Health Sector M&E mechanisms. It is intended to document what needs to be monitored,

with whom, by whom, when, how, and how the M&E data will be used. It also outlines how and when

the different types of studies and evaluations will be conducted by the sector. It translates these

processes into annualized and costed activities, and assigns responsibilities for implementation.

1.1 Goals and objectives of the HSDP 2015/16 - 2019/20

1.1.1 Goal

To accelerate movement towards Universal Health Coverage (UHC) with essential health and related

services needed for promotion of a healthy and productive life.

This goal will be achieved through a focus on the following specific objectives.

1.1.2 Objectives

1) To contribute to the production of a healthy human capital for wealth creation through provision

of equitable, safe and sustainable health services.

2) To increase financial risk protection of households against impoverishment due to health

expenditures.

13 | P a g e

3) To address the key determinants of health through strengthening inter-sectoral collaboration and

partnerships.

4) To enhance health sector competitiveness in the region and globally.

1.2 Goal and Objectives of the M&E Plan for the HSDP

1.2.1 Goal

The goal of this M&E Plan is to contribute to the attainment of the HSDP 2015/16 - 2019/20 objectives

and performance targets by providing a health sector-wide framework for regular and systematic

tracking and documenting progress in the health sector performance.

1.2.2 Specific Objectives

The specific objectives are;

1. To define mechanisms for assessment of the health sector performance in accordance with the

agreed objectives, performance indicators and reporting timelines for the HSDP.

2. To support management for results through evidence based decision-making, policy development

and advocacy; sector learning and improvement at all levels during implementation of the HSDP.

3. To enhance compliance with government policies (accountability), and constructive engagement

with stakeholders (policy dialogue) in the health sector.

4. To build capacity of the health sector to generate quality data and track progress of

implementation of the HSDP.

5. To facilitate continuous learning (document and share the challenges and lessons learnt) at all

levels during implementation of the HSDP.

1.2.3 Outcomes

This M&E Plan should result in:

i) Timely reporting on progress of implementation of the HSDP

ii) Objective evidence based decision making for performance improvement, planning, resource

allocation, policy development and advocacy.

iii) Timely accountability to government and stakeholders.

iv) Competent and well equipped human resource for M&E.

v) Documentation and sharing of results and lessons learnt.

1.3 Situational Analysis of the M&E practices and mechanism The M&E Plan for HSSIP 2010/11 – 2014/15 was developed and disseminated aiming at establishing

an M&E system that is robust, comprehensive, fully integrated, harmonized and well coordinated

to guide monitoring of the implementation of the HSSIP and evaluate impact. This resulted in a

Country-led M&E Platform especially in the use of harmonized data collection tools and reporting

system. With health information, research and evidence generation, the country was able to

transition to the District Health Information System (DHIS)-2, which is an electronic web based

reporting mechanism and also revised the HMIS reporting tools to ensure information

disaggregation. A Data Quality Assessment (DQA) manual and harmonized DQA tools were

developed. Surveys and facility assessments specific for some high burden conditions (HIV, TB

prevalence, malaria, NCDs, etc) were successfully initiated / conducted (HSDP MoH 2015). During the

HSSIP period, timeliness of HMIS monthly reporting by the districts increased from 77% in 2010/11

to 88% in 2014/15; and completeness of facility reporting increased from 85% in 2010/11 to 96%

(AHSPR 2014/15, MoH).

E-health has become a stronger area of focus, with the national e-health technology framework

completed and draft e-Health policy developed. The Government has maintained strong stewardship

over this developing area, to ensure the emerging e-health architecture is aligned to the pillars of the

e-health house of value.

With support from the Global Fund for TB, HIV/AIDs and Malaria (GFTAM), 12 Regional Performance

Monitoring Teams were established to strengthen the decentralized capacity for active performance

monitoring and surveillance of program outputs and support the performance management of

implementing agencies at all levels. A Health Facility and community feedback mechanism i.e. mTRAC

exists and has been rolled in all the 112 districts. However, UNICEF has also been able to roll out the

U- report-which is also a community feedback mechanism. On annual basis Annual Health Sector

Performance Reports were compiled and disseminated. Annual Sector Performance Reviews were

conducted in addition to biannual reviews at national level. Quarterly performance reviews were

conducted at sub-national levels. Program review conducted for Malaria Program in 2011, TB

program in 2013 and HIV/AIDS program review to be conducted in 2015 (HSDP, MoH 2015). The MoH

also undertook a Mid-Term review of the HSSIP and the findings were used to inform the planning

process in the subsequent years as well as inform the development of the HSDP.

The M&E Structure for the MoH is not defined to streamline the M&E functions at national level. Weak

linkages between MoH and UBOS leading to poor timing in provision of results for guiding sector

planning. There is also lack of annual vital statistics on births and deaths. The sector has made progress

towards aligning the previously fragmented information systems within the health sector to ensure

contribution to the one M&E system.

The inadequate supply of the revised HMIS tools in all facilities is still hampering data collection and

reporting which compromises the data quality. There is still a challenge in the generation, compilation

and submission of programmatic and departmental reports that are expected to feed into the overall

sector performance reports. There are still irregular program and sub-national performance reviews

due to inadequate funds. Evaluations for most programs not conducted due to inadequate funding

and as a result impact of interventions is not well documented. The low reporting rates from the

private sector limits the ability to appropriately monitor overall sector outcome performance. Finally,

the utilization of data for decision-making is still minimal and technical guidance and coordination

needs to be given and enforced from the MoH to the lower levels and Partners (HSDP, MoH 2015).

The reliable collection of Civil Registration information and its accurate and timely processing into

Vital Statistics is critical to Government planning for the provision of social services. By December

2014, the Civil Registration and Vital Statistics (CVRS) was operational in 135 Government, Missionary

and Industrial hospitals country wide to record births in hospitals, 62 sub counties and 5 divisions

within Kampala Capital City Authority (KCCA) to record births in communities. Operators of this system

have been trained and equipped with registration equipment and materials. (Uganda National

Assessment on CRVS, UBOS 2014)

15 | P a g e

The national assessment revealed that whereas people are knowledgeable about birth and death

registration, they seldom register birth and death, except when a birth or death certificate is needed

as a requirement to access a certain service. It was further established that the majority of the duty

bearers who are charged with registration of births and deaths do not consider making birth and death

registration returns as an important issue. The Country lacks a policy on BDR which makes it difficult

to enforce the BDR Act.

Further review of the M&E Plan for implementation of the HSSIP 2010/11 - 2014/15 was undertaken

to inform the development of this plan using the WHO Accountability Assessment Framework. The

objective of the assessment was to ensure that the HSDP M&E Plan practices and mechanisms are

appropriate to the country needs and that the plan will have wide ownership and commitment. This

was through identification of the strengths and weaknesses in the country M&E system, including

identification of the major actions required to address the gaps and needs.

Overall, there was compliance of the M&E plan for the HSSIP 2010/11-2015/15 basing on the WHO

Accountability Assessment Framework.

The key gaps identified using the WHO Accountability Assessment Framework are outlined below:

1) The baseline information for most of the impact indicators and other population based data

indicators for the HSSIP is not available to provide the current baseline or progress during the

HSSIP period. This information is obtained from the Uganda Demographic and Household

Survey (UDHS), which takes place every five years. It was noted that baseline information for

these impact indicators for the current strategic planning period would only be available

between December 2016 and March 2017.

2) The level of disaggregation (gender, age category, literacy, poverty, region) is not routinely

built into the analysis of the core set of indicators. Although some of the latter (age, sex,

region) are well captured in the HMIS and population-based surveys e.g. UDHS, it was noted

that the data users (programs) do not regularly demand for disaggregated data from the

Resource Center.

3) The M&E Plan specified the related costs for the Plan but many important areas remained

unfunded. There was some off-budget support had been obtained for some of these unfunded

areas but that this information had not been subsequently captured.

4) The main Human Resource (HR) challenge is the absence of an HR structure for M&E at the

MoH even though a submission to redress this was made to the Ministry of Public Service

(MoPS). So far there are only scattered uncoordinated efforts that have nevertheless enabled

recruitment of a few staff. There is still need for skills development for M&E.

5) There is an unclear process for setting and reviewing annual targets. For example, during the

drafting of the M&E plan for the HSSIP, Program Managers made their submissions to

Supervision Monitoring Evaluation & Research Technical Working Group (SMER TWG) and

there was no balance between the expected resources and targets being set leading to

underperformance against a number of indicators. However, there is a process to review some

of the key sector performance indicators that are reported to the Office of the Prime Minister

(OPM) annually.

6) There is a small task force under the Country Information and Accountability Framework

(COIA) Taskforce working on vital statistics with representation from the SMER TWG. However

there is a weak linkage between the MoH and the National Identification & Registration

Authority (NIRA), which undermines opportunities for strengthening the institutional linkage

between the MoH and the Ministry of Justice and Constitutional Affairs.

7) Disease burden among adults is summarized as broad categories – non-

communicable/communicable largely because there is weak diagnostic capacity at lower

levels e.g. the Hospital Census shows that only 13 out of 335 facilities are using the

International Classification of Diseases (ICD) 10 classification.

1.3.1 Recommendations

Prioritisation of provision of HMIS data collection tools at all levels including the private sector.

Include the levels / categories of disaggregation for all core indicators in the indicator definition

manual.

The impact indicator baselines should be updated as soon as information becomes available, if

possible during the Mid-Term Review (MTR) of the HSDP. The related targets and strategies should

be revised at the time of updating the baseline information.

SMER TWG to develop and disseminate the process for adjusting indicator targets and streamline

the commitment of targets with actual resources at least once during the MTR.

Ensure data synthesis and analysis based on disaggregated data and support skills strengthening

for this process at all levels. Sub-national analyses of the already available HMIS data as well as

the population-based survey results could identify challenged regions or sub-populations

deserving more targeted approach for achieving higher impact.

Linkage of the Human Resource Information System (HRIS), and Supply Chain management system

and other Health Information Systems to the DHIS 2 for better management.

Strengthen coordination of existing M&E human resource and activities at all levels for provision

of timely and accurate information, effective use of statistical data and health research at all levels

is a critical issue that can impact successful implementation of the HSDP.

Strengthen partner coordination for support and reporting through regular partner meetings and

joint planning.

Strengthen linkages and coordination between MoH and key partners e.g. UBOS, NIRA to address

the MoH data needs in a timely manner.

Institutionalization of the ICD 10 at all levels.

1.4 Process for development of the M&E Plan The SME&R TWG comprising of MoH officials, HDPs, Civil Society, UBOS, Medical Bureaus, private

sector and academia was responsible for overseeing the development of the HSDP M&E plan. A Task

Force was formed which worked in close collaboration with all relevant stakeholders in the

development process. This was in conjunction with the development of the HSDP and took into

consideration the NDP II M&E Strategy and HSDP requirements, as well as international treaties and

17 | P a g e

conventions to which Uganda is a signatory more especially the Global Health Development Agenda

including the Sustainable Development Goals (SDGs) and the International Health Partnerships and

related Initiatives (IHP+) which seek to achieve better health results and provide a framework for

increased aid effectiveness; among others.

A review of a wide range of national and health sector documents was done to provide in-depth

analysis and understanding of the sector M&E system; such as NDP II M&E Strategy, NHP II, M&E Plan

for the HSSIP 2010/11 – 2014/15, HSDP 2015/16 - 2019/20, Program specific M&E plans and the

Health Management Information System (HMIS) Manual, among others. The aim of the document

review was to harmonize this Plan with the other government and inter-sectoral M&E Plans.

A rapid assessment of the HSSIP M&E Plan was conducted to provide insight into the current M&E

practices and mechanisms as well as insight into the unfinished agenda from the previous strategic

period and the reasons for this. The findings were used to guide the development of this plan. The

findings were presented and discussed in the SMER TWG and were used to enrich the Plan. A national

stakeholders meeting was conducted to obtain consensus on the plan. The involvement of the

different stakeholders was important in order to ensure transparency and ownership of the Plan. The

final draft was presented to the MoH Senior Management Committee for input and finally to the

Health Policy Advisory Committee (HPAC) for approval.

1.5 Disease and program specific M&E alignment Development of program and project specific M&E plans is a requirement for institutions, programs

and projects. Under the country-led platform operationalized during the last five years it is a

requirement for programs and projects to develop M&E plans aligned to the national strategy M&E

plan. All disease and program specific M&E Plans will use the same technical framework and M&E

platform defined in the HSDP M&E Plan.

The sector will also use a common platform for household surveys, facility assessments and

evaluations. This implies that data collection, aggregation, transfer and analysis are all coordinated

under the national M&E system.

Performance monitoring and review by institutions, programs and projects shall be based on the

institutional, program or project specific indicators in the respective M&E plans.

2 Institutional Capacity Chapter two outlines the institutional environment defining the coordination mechanisms, roles and

responsibilities for the different stakeholders and capacity building strategy. The health sector should

ensure availability of sufficient funding and financial management capacity as well as human resources

with adequate technical capacity to manage the various components of the M&E system (Table 1) in

support of progress and performance reviews.

Table 1: The 12 Components of a functional M&E system

Component Performance goal for this component

1. Organizational structures with M&E functions

Establish and maintain a network of organizations responsible for M&E at the national, sub-national and service delivery levels.

2. Human capacity for M&E Ensure adequate skilled human resources at all levels of the M&E system to ensure completion of all tasks defined in the annual costed M&E workplan. This requires sufficient capacity to use the data and produce relevant reports.

3. Partnerships to plan, coordinate and manage the M&E system

Establish and maintain partnerships among in-country and international stakeholders involved in planning and managing the national M&E system.

4. National multisectoral M&E plan

Develop and regularly update the national M&E plan, including identified data needs, national standardised indicators, data collection procedures and tools as well as roles and responsibilities for implementation.

5. Annual costed M&E workplan

Develop an annual costed M&E workplan including specified and costed M&E activities of all relevant stakeholders and identified sources of funding; use this plan for coordination and for assessing the progress of M&E implementation throughout the year.

6. Advocacy, communication and culture for M&E

Ensure knowledge of and commitment to M&E and the M&E system among policymakers, program managers, program staff and other stakeholders.

7. Routine program monitoring

Produce timely and high quality (valid, reliable, comprehensive and timely) routine program monitoring data.

8. Surveys and surveillance Produce timely, valid and reliable data from surveys and surveillance systems.

9. National and sub-national databases

Develop and maintain national and sub-national databases that enable stakeholders to access relevant data for formulating policy and for managing and improving programs.

10. Supportive supervision and data auditing

Monitor data quality periodically and address obstacles to producing high quality (valid, reliable, comprehensive and timely) data.

11. Evaluation and research Identify evaluation and research questions, coordinate studies to meet identified needs and enhance the use to evaluation and research findings.

12. Data dissemination and use

Disseminate and use data from the M&E system to guide the formulation of policy and the planning and improvement of programs.

19 | P a g e

2.1 M&E Coordination Mechanisms The sector-wide M&E system for effective tracking, evaluation and feedback on HSDP implementation

and results will be followed as defined in the NDP II M&E Strategy. This implies that Central

Government, MoH, LGs, HDPs, CSOs and other stakeholders will be involved directly or indirectly in

the M&E activities. Consequently, a participatory approach that entails the involvement of all key

actors and primary stakeholders will be adopted. This will enable all key actors to fully internalize and

own the system as well as use the results to inform their actions.

Better coordination of service delivery is a key element required to maximize the outputs of the health

sector. Coordination of M&E activities in the health sector will be through the existing organizational

structures of the health sector (See Figure 1).

Figure 1: Health Sector M&E National Coordination Structure

Quality Assurance

Department

Department M&E

Focal Persons /

Health Facility Manager

Health Information

Division

Planning Department

Senior Management Committee

HPAC

Top Management

SMER TWG

Regional M&E

Teams

District Health

Team

Regional M&E

Teams

District Health

Team

Figure 2: District Level M&E National Coordination Structure

2.2 Roles and Responsibilities In order to avoid over-laps, role conflicts, and uncertainty in the M&E function during the

implementation of the HSDP, roles and responsibilities of key actors are specified below.

2.2.1 Roles and Responsibilities of Senior Top Management

The Senior Top Management is comprised of the 3 Ministers. They shall work closely with the

Cabinet through the Parliamentary Health Committee for;

i) Overall political, and policy oversight.

ii) Articulating the policy direction for the sector, taking broader Government objectives into

consideration.

iii) Review of sector progress against the policy imperatives set out in contribution towards the NHP

II and NDP II.

This function will be supported by reports of the Auditor General, NPA and other statutory agencies.

District Technical

Planning Committee

District Social Services

Committee

District Health

Management Team

Community / Households / Individuals

District Executive Committee

District Health Team

Health Sub District

Management Team

Health Facility

Manager

Health Facility

Manager

CHEW / VHT CHEW / VHT

Health Sub District

Management Team

Health Facility

Manager

Health Facility

Manager

CHEW / VHT CHEW / VHT

21 | P a g e

2.2.2 Roles and Responsibilities of Top Management

The Top Management is comprised of the Permanent Secretary, Director General of Health Services,

Directors and Commissioners.

The M&E roles and responsibilities for the top management are;

i) Providing governance and partnership oversight to the sector.

ii) Reviewing of sector progress against the policy imperatives set out in the NHP II and NDP II.

iii) Monitoring adherence to the policy direction (One M&E platform) of the sector.

iv) Mobilizing resources for operationalizing the plan.

v) Monitoring health sector performance and six-monthly reporting to OPM on sector performance.

This performance reporting will be based on the quarterly submissions to OPM on progress against

key actions, and outputs towards outcomes.

2.2.3 Roles and Responsibilities of Health Policy Advisory Committee (HPAC)

HPAC is a forum for the Government, HDPs, private sector, PNFPs, CSOs, academia and related line

ministries.

The role of HPAC in implementation of this plan is to;

i) Monitor and advise on health policy issues.

ii) Monitoring adherence to the policy direction (One M&E platform) of the sector.

iii) Monitor implementation of the partnership arrangements e.g. the Compact and aide memoire

recommendations.

iv) Provide guidance for the sector performance reviews.

2.2.4 Roles and Responsibilities of SME&R TWG

The SME&R TWG is comprised of MoH representatives from the different programs, CSOs, HDPs,

Medical Bureaus, Private Sector and academia and reports to the Senior Management Committee.

The TWG will be responsible for;

i) Developing and reviewing the results framework for the HSDP, and for ensuring that relevant

departments (and relevant non-state actors) develop results indicators that are consistent with

the HSDP.

ii) Reviewing, consolidating and validating the various sector reports before dissemination to the

relevant stakeholders.

iii) Facilitating utilization of M&E and Research Information (knowledge translation and

dissemination) in liaison with the relevant sector departments, programs and the Policy Analysis

Unit.

iv) Periodically reviewing available research/survey information to update the current evidence on

best practices in order to guide Planning, decision-making or Policy Formulation.

2.2.5 Roles and Responsibilities of Quality Assurance Department (QAD)

The QAD is responsible for overall coordination of the implementation of the HSDP M&E Plan.

However, MoH has proposed to the MoPS the establishment of a fully fledged M&E Unit to strengthen

the M&E function. The MoH M&E sub-units will work under the guidance of the QAD in

implementation of this plan.

QAD is responsible for;

i) Overall coordination and oversight (monitoring and supervision) of M&E activities in the sector.

ii) Development and dissemination of the sector M&E plan for the HSDP.

iii) Ensuring the harmonization of the institutional, program and project M&E plans with the HSDP

M&E Plan

iv) Identifying capacity building needs and training for health workers and managers in M&E.

v) Organizing sector performance reviews.

vi) Supporting LGs to organize regular performance review meetings.

vii) Producing the quarterly sector progress reports.

viii) Providing on a quarterly basis, data and explanatory information on progress against performance

indicators to MoFPED and OPM through the Output-based Budgeting Tool (OBT).

ix) Maintaining a Recommendations Implementation Tracking Plan which will keep track of review

and evaluation recommendations, agreed follow-up actions, and status of these actions.

x) M&E of the implementation of the HSDP M&E Plan.

2.2.6 Roles and Responsibilities of Health Information Division

The HID will have the primary responsibility of;

i) Coordinating, harmonizing and operationalizing the HMIS and e-HMIS at all levels.

ii) Strengthening capacity for collection, validation, analysis, dissemination and utilization of health

data at all levels.

iii) Conducting regular data validation in the districts, health facilities and other health institutions to

ensure quality data.

iv) Ensure that HMIS and e-HMIS data is made easily available to all stakeholders in a timely manner,

while ensuring that the sharing of reports respects the Access to Information Act, 2005.

v) Conduct data review meetings include data use conferences to enhance data utilization at all

levels.

vi) Updating the master health facility inventory of all reporting health facilities in the country.

vii) Ensure the ICD 10 coding of all diagnoses as outlines in the HMIS, and consequently updating these

in the eHMIS.

viii) Generating the health statistical report annually.

2.2.7 Roles and Responsibilities of the Departmental Heads and M&E Focal Persons /

Specialists

Central level departments and institutions will be centers for performance monitoring as well as

reporting on progress of their plans against the set targets and actions. They will also be the direct

consumers of the outputs and outcomes of this M&E Plan.

The roles and responsibilities of the departmental heads or M&E focal persons include;

i) Provide oversight for monitoring implementation of work plans and preparation of quarterly and

annual performance reports.

ii) Participating in data quality assurance.

iii) Providing quality data on relevant performance indicators to MoH and relevant stakeholders.

iv) Participate in M&E capacity building activities.

v) M&E support supervision and mentoring.

23 | P a g e

vi) Participating in development of the program M&E plans, Mid- and end-term evaluation of national

strategic plan and preparation of the periodic sector performance reports.

vii) Maintaining a Recommendations Implementation Tracking Plan which will keep track of review

and evaluation recommendations, agreed follow-up actions, and status of these actions.

viii) Utilizing M&E findings to inform program, policy, and resource allocation decisions.

2.2.8 Roles and Responsibilities of Regional Monitoring Teams

M&E functions at the regional level will be coordinated through designated regional teams.

The regional teams will be responsible for;

i) Liaising between national level and the districts on M&E at regional level.

ii) Capacity building for data collection, validation, analysis, dissemination and utilization of health

data at regional level.

iii) Conducting data validation in the region.

iv) Supporting the development and implementation of the M&E plans of the districts and RRH in the

region.

v) Monitoring and reviewing the implementation of the M&E plans in the region by compiling and

analyzing quarterly and annual reports.

vi) Maintaining a Recommendations Implementation Tracking Plan which will keep track of review

and evaluation recommendations, agreed follow-up actions, and progress of these actions.

vii) Supporting operational research and survey activities.

2.2.9 Roles and Responsibilities of the District Executive Committee (DEC)

At district level the DEC will be responsible for;

i) Providing governance and leadership oversight in the district.

ii) Monitoring implementation of the annual workplan and District Development Plan.

iii) Monitoring adherence of all stakeholders to the policy direction (One M&E platform).

2.2.10 Roles and Responsibilities of the District Social Services Committee (DSSC)

The DSSC is a committee of the District Council and will be responsible for;

i) Monitoring implementation of the annual workplan and District Development Plan.

ii) Mobilizing resources for operationalizing the district M&E plan.

iii) Participating in the district performance review meetings.

2.2.11 Roles and Responsibilities of the DTPC

The DTPC is chaired by the CAO is comprised of all Heads of Departments in the district. The DPTC

will be responsible for;

i) Monitoring implementation of the annual workplan and District Development Plan.

ii) Mobilizing resources for operationalizing the district M&E plan.

iii) Ensure timely reporting to the respective entities (MoFPED, MoLG and MoH).

iv) Coordinating all IPs and CSOs to ensure alignment with institutional arrangements and district

priorities.

2.2.12 Roles and Responsibilities of the DHMT

The DHMT is comprised of the DHT and representatives of the key stakeholders in the district including

the CAO, Secretary for Health, CSOs and IPs in the district. The DHMT will be responsible for;

i) Monitoring implementation of the annual workplan and District Development Plan (Performance

Reviews).

ii) Participating in the district performance review meetings.

2.2.13 Roles and Responsibilities of the DHT

At district level the District Health Team will be responsible for;

i) Development of a district M&E plan.

ii) Identifying key performance indicators and targets.

iii) Coordination of M&E activities in the district to ensure alignment with institutional arrangements

and district priorities.

iv) Supervision and mentoring of HSDs in M&E.

v) Conducting data quality audits.

vi) Training of health workers in M&E.

vii) Maintaining a functional district HMIS with up to date district database.

viii) Compile and submit periodic district reports to the district, MoFPED and MoH

ix) Conducting district performance Reviews.

x) Maintaining a Recommendations Implementation Tracking Plan which will keep track of review

and evaluation recommendations, agreed follow-up actions, and progress of these actions.

xi) Utilization of M&E results.

2.2.14 Roles and Responsibilities of the HSD Team

At HSD level the HSD Team will be responsible for;

i) Supervision and mentoring of lower level health facilities in M&E.

ii) Aggregation, validation, analysis, dissemination and utilization of district data.

iii) Maintaining an up to date HSD database.

iv) Compile and submit periodic HSD reports to the district.

v) Conducting HSD performance Reviews.

vi) Maintaining a Recommendations Implementation Tracking Plan which will keep track of review

and evaluation recommendations, agreed follow-up actions, and progress of these actions.

vii) Utilization of M&E results.

2.2.15 Roles and Responsibilities of Health Facility Managers

All facility managers in the public and private health facilities will be responsible for;

i) Ensuring development of facility M&E plans.

ii) Determining performance targets for the key output indicators.

iii) Resource mobilization and allocation for M&E activities.

iv) Maintaining an up to date health facility database.

v) Compile and submit periodic reports to the relevant bodies.

vi) Conduct data verification before submission of reports.

vii) Conducting health facility performance reviews.

viii) Dissemination and utilization of data.

ix) Maintaining a Recommendations Implementation Tracking Plan which will keep track of review

and evaluation recommendations, agreed follow-up actions, and progress of these actions.

2.2.16 Roles and Responsibilities of the Village Health Team (VHT)

The roles of the VHTs include;

25 | P a g e

i) Collection, compilation, analysis and reporting on community health data including births and

deaths through the community registers.

ii) Use data to discuss performance within the community, and agree on priorities to focus on.

2.2.17 Roles and Responsibilities of Community Level Actors

These comprise of community based CSOs, administrative units at grassroots and health consumers. Their role is; i) To provide information on; i) delivery of various services, ii) transparency and accountability of

resources accorded; and iii) challenges and gaps experienced in delivery of various services.

ii) They will also participate in validation of the outcomes of implementation of the HSDP in their respective areas.

iii) Communities will also be engaged in the review process using participatory appraisal mechanisms like focus group discussions and community meetings or dialogues like the Constituency (HSD) Health Assemblies, barazas, open days, etc.

2.2.18 Roles and responsibilities of other executing agencies and partners

The MoH will work in close collaboration with other agencies and partners in the implementation of this plan. These include UBOS, NIRA, Development Partners and CSOs.

2.2.18.1 Roles and Responsibilities of Uganda Bureau of Statistics (UBOS)

Will be responsible for;

i) Coordinating, supporting, validating and designating as official any statistics produced by UBOS,

MDAs and LGs.

ii) Coordinating and clearing all censuses and nationally representative household economic surveys.

iii) Ensuring production, harmonization and dissemination of statistical information.

iv) Strengthening statistical capacity of planning units in MoH and LGs for data production and use.

v) Ensuring best practice and adherence to standards, classifications, and procedures for statistical

collection, analysis and dissemination in MoH and LGs.

vi) Ensuring that complete and approved health statistical data are made easily available to the public

in a timely manner, while ensuring that the sharing of reports respects the Access to Information

Act, 2005.

Research activities by academic institutions that are related to the M&E discipline will also contribute

in this regard.

2.2.18.2 Roles and Responsibilities of NIRA

NIRA is responsible for;

i) Operating and managing the Civil Events Registry.

ii) Timely processing and dissemination of Vital Statistics.

iii) Capacity building of health providers and community in registration of births and deaths.

iv) Provision of registration equipment and materials.

2.2.19 Roles and responsibilities of CSOs

The CSO coalitions e.g. Uganda Coalition for Access to Essential Medicines (UCAEM), Maternal Health

CSOs, Civil Society Budget advocacy group, Voices for health right, among others will play a coordination

role in monitoring all CSOs to ensure alignment with national priorities.

The role of the CSOs and IPs in the implementation of the M&E Plan will be;

i) Contribution in the development of M&E standards and plans.

ii) Participating in sector monitoring processes at LG and national level.

iii) Providing performance reports and quality data to the relevant program managers at national and

district level. These will be compiled as part of departmental reports to be reviewed by relevant

working groups for onward transmission to SMC or DTPC.

iv) Participating in the M&E related committee meetings at all levels.

v) Conduct independent M&E audits and share findings for performance improvement.

vi) Community sensitization and advocacy for accountability mechanisms.

2.2.20 Roles and responsibilities of HDPs and IPs

The HDPs and IPs in the sector will work closely with the national and sub-national levels to ensure

alignment with national priorities and use of one M&E platform.

The role of the HDPs and IPs in the implementation of the M&E Plan will be;

i) Contribution in the development of M&E standards and plans.

ii) Participating in sector monitoring processes at LG and national level.

iii) Participating in the M&E related committee meetings at all levels.

iv) Utilizing M&E findings for policy dialogue, resource mobilization and planning.

v) Providing feedback to domestic and international constituencies on health sector performance and results.

vi) Supporting the health sector through financial, technical and other forms of assistance to strengthen M&E performance.

2.3 Capacity Building Strategy The sector institutions shall ensure availability of adequate skilled human resources in all aspects of

M&E at all levels of the M&E system. M&E personnel require sufficient analytical capacity to use the

data and produce relevant reports. To ensure this, the sector shall develop/improve M&E human

resource capacity through training, providing guidelines and tools, supportive supervision and

mentoring.

2.3.1 Training

Training needs assessments in M&E will be conducted to identify actual needs and tailored program

developed to address capacity gaps. Training approaches to include on-job training, coaching /

mentoring, Continuous Medical Education sessions, workshops or approved M&E courses (distance

and otherwise).

All District Health Teams and M&E staff will be oriented on the M&E Plan and current M&E tools.

Particular attention will be paid to strengthen capacity for data quality assessment, data analysis and

synthesis and use of data for progress and performance reviews within LGs, MoH Departments, semi-

autonomous Institutions, PNFP & PFP facilities and CSOs. Other areas of focus will include training on

the concept of performance based monitoring and management and target setting. The MoH will

leverage the expertise and capacity of in-country institutions such as academic, public health and

27 | P a g e

research institutions to conduct in-service training in analytical skills that will facilitate knowledge

management.

2.3.2 Providing Guidelines and Standard Operating Procedures

One of the ways to enable health workers generate quality data and routinely monitor service delivery

is to ensure that current standards, guidelines and Standard Operating Procedures (SOPs) are

reviewed and disseminated. The M&E Plan is one such document that will be disseminated to all

districts and health facilities. In addition other documents like the HMIS manual, Data Quality Manual,

program M&E plans will be reviewed and disseminated.

2.3.3 Supportive Supervision and mentoring

Supportive supervision is mainly conducted to assess level of performance against standards. This

process involves “directing and supporting staff so that they may effectively perform their duties

through observation, discussion, guidance/support and on-the-job training using mentoring and

coaching methods.

During implementation of this M&E Plan the health sector shall undertake the following;

Review and disseminate M&E support supervision tools.

Conduct regular M&E support supervision and mentorship at all levels.

The supervisors are to;

Work with staff and the community to identify appropriate interventions that will lead to

improved M&E system.

Facilitate teamwork at organisational level and between the sub-national, national level and the

facilities.

Ensure that facilities have relevant M&E standards, guidelines and tools.

Facilitate districts and facilities to develop M&E plans.

Review completeness and accuracy of reports (data quality) as well as timeliness of reporting.

Work with staff to use data to evaluate results (keep track of action plans, other tools).

Work with staff to assess training needs, plan and conduct training.

Mobilize resources from many different sources to implement M&E workplans.

3 Monitoring and Evaluation Framework Chapter three defines the M&E Framework for the HSDP, main M&E events, key indicators and targets

for monitoring performance in the sector, data sources, data management, dissemination, use,

evaluation, surveys and surveillance.

3.1 The M&E Framework The monitoring framework for tracking progress is informed by the need to comprehensively monitor

and review sector progress. The framework for the analysis is based on the common steps of the

M&E logical framework, which shows the way in which inputs may lead to desirable health impact.

The health sector shall apply the IHP+ common M&E Framework which provides a logical and results-

chain representation of the key components of the HSDP M&E (see Figure 3). It comprises four major

indicator categories: system input and processes, outputs, outcomes, and impact.

The framework shows how inputs to the system (e.g. financing, infrastructure) and processes (e.g.

supply chain) are reflected in outputs (such as availability of services and interventions) and eventual

outcomes (e.g. intervention coverage) and impact (e.g. improved health outcomes). Figure 3: IHP+ Common M&E Framework

Source: Monitoring, Evaluation and Review of National health Strategies, WHO 2011

Inputs and processes Outputs Outcomes Impact

Go

vern

ance

Fin

anci

ng

Infrastructure, Information and communication

technologies

Health workforce

Supply chain

Information

Coverage of interventions

Prevalence risk behaviours and

factors

Improved health

outcomes and equity

Social and

financial risk protection

Responsiveness

Efficiency

Intervention access and

services readiness

Intervention

quality, safety

Administrative sources Financial tracking system; National Health Accounts; Databases and records, HR, infrastructure, medicines, etc Policy data

Facility assessments Service readiness, quality, coverage, health status

Population based surveys, coverage, health status, equity, risk protection, responsiveness

Clinical reporting systems

Civil registration

Indicator domains

Analysis

Communication

and Use

Data quality assessment; Review of progress and performance; Evaluation

Country review process; Global reporting

Data

Collection

29 | P a g e

3.2 M&E Events The main M&E events, purpose / description, main actors and timeframe are summarized in Table 2.

Table 2: The M&E events and processes

Main M&E Events Purpose and Description Main actors Frequency

Baseline Surveys Establish basis for measuring progress of the HSDP Implementation overtime

UBOS, NIRA, MOH, Academia

Every 5 years, annually

Surveys/studies Basis for a 'before and after' assessment of the HSDP progress

UBOS, NIRA, MOH, Academia, Partners

Ongoing

Establishing Management Information Systems

Management of M&E data UBOS, MOH, LGs, Facilities, Partners

Ongoing

Evaluations To provide feedback on effectiveness, efficiency, and adequacy, of the HSDP implementation and to identify probable areas for improvement.

MoH, HDPs, Private Sector, CSOs, Researchers, Academia and partner organizations

Mid term and end term

Performance Reviews

To review progress in implementation of the Ministerial Policy Statement and workplans against the targets sets.

MOH, HDPs, Private Sector, CSOs, Researchers and Partner Organizations

Quarterly, Semi-Annually and Annually

Technical Reviews To review thematic areas and discuss key successes and challenges, identify areas for modification of operations / implementation and make adjustments.

MOH, HDPs, Private Sector, CSOs, Researchers, Academia and Partner Organizations

Once a year - March - April.

M&E capacity building

Increase knowledge regarding the development monitoring and evaluation systems among key stakeholders involved in the HSDP implementation

MOH, LGs & Partner Organizations

Continuous

Information sharing

Enhancing smooth implementation of the HSDP

MOH, HDPs, Private Sector, CSOs, Researchers, Academia and Partner Organizations

Continuous

3.3 Indicators The ability to monitor and evaluate in the health sector is essential if we are to correctly target

interventions and assess whether they are having the desired impact. Different indicators are available

to guide performance in health. The sector shall use both program based, and sector level monitoring,

and use of indicators for each.

Program based monitoring is able to inform on progress being made at the program level, at a level

of detail sufficient to target implementation of available interventions. As such, many indicators are

usually employed, each providing specific information, on its own, about a different aspect of the

program. On the other hand, sector monitoring is higher level monitoring of whether the multitude of

activities being carried out are having the desired impact on the sector goals. A single indicator on its

own is insufficient to inform on sector progress – rather a number of different indicators are

employed. These shall be called key indicators.

3.3.1 Key Indicators

Key indicators are defined, and structured to inform on, and compare trends across the different

indicator domains. The key indicators guide analysis of sector progress, at all the indicator domains –

input / process, output, outcome, and impact.

Selection of the key indicators was through a participatory process, bringing together key decision-

makers across priority diseases addressed in the HSDP and health system components. The 41 key

indicators are based on the following critical variables:

They reflect all domains in the M&E conceptual framework,

Are aligned to existing sector and global monitoring commitments.

Considerations of scientific soundness, relevance, usefulness for decision-making,

responsiveness to change, and data availability.

Responsiveness to the information needs for monitoring progress and performance towards

the main objectives of the HSDP.

Table 3 presents all the key indicators showing the domain, data collection method, frequency of data

collection, agency responsible for data collection and reporting, baseline values with dates and source

of data and targets according to frequency of measurement.

The annual targets were set based on computations that include the likely availability of funding, on

how this can be translated into intervention access and coverage, and ultimately on health impact in

addition to taking the international targets of the post 2015 MDG agenda into consideration. The

assumption was that all factors contributing to the attainment of the targets will be addressed. There

will be a review of the indicator targets during the midterm review of the HSDP and adjustment made

where necessary based on the level of achievement and other factors like funding and responsiveness

to the information needs.

The indicator matrix in Table 4 shows the HSDP objectives, key result areas, indicators, data collection

methods, frequency of data collection, agency responsible, baseline and annualized indicator targets.

31 | P a g e

Table 3: HSDP 2014/15 – 2015/16 Key Indicators by Domain

INPUT & PROCESS (7) OUTPUT (13) OUTCOME (17) IMPACT (5)

Health financing Service access and availability Coverage of interventions Health status

1 General

Government budget

allocated on health

as % of total

government budget

8 New OPD utilization rate 21 ANC 4 Coverage 36 Maternal Mortality

Ratio

2 Out of pocket health

expenditure as a %

of Total Health

Expenditure

9 Hospital (inpatient)

admissions per 100

population

22 DPT3HibHeb3 coverage 37 Neonatal mortality

rate

Workforce 10 Bed occupancy rate

(Hospitals & HC IVs)

23 Measles coverage under

1 year

38 Infant Mortality

Rate

3 Annual reduction in

absenteeism rate

11 HC IVs offering CEmOC

Services

24 Malaria cases per 1,000

persons per year

39 Under 5 mortality

rate

4 Approved posts in

public facilities filled

with qualified

personnel

12 Availability of a basket of

commodities in the previous

quarter (% of facilities that

had over 95%)

25 Under-five Vitamin A

second dose coverage

40 Total Fertility Rate

5 Number of health

workers (doctors,

midwives, nurses)

per 1,000

population

Service quality and safety 26 Intermittent Presumptive

Treatment 3 or more

doses coverage for

pregnant women

41 Adolescent

Pregnancy Rate

Infrastructure 13 Client satisfaction index 27 TB Case Detection Rate

6 Population living

within 5km of a

health facility

14 Average length of stay

(Hospitals & HC IVs)

28 ART Coverage

Health Information 15 Facility based fresh still

births (per 1,000 deliveries)

29 HIV+ pregnant women not

receiving HAART

receiving ARVs for

eMTCT during pregnancy,

labour, delivery &

postpartum

7 Timeliness of

reporting (HMIS

105)

16 Maternal deaths among

100,000 health facility

deliveries

30 Contraceptive Prevalence

Rate

17 Maternal death reviews

conducted

31 Couple years of

protection

18 Under five deaths among

1,000 under 5 admissions

Risk factors and behaviours

(4)

19 ART Retention rate 32 Latrine coverage

20 TB Treatment Success

Rate

33 Villages/ wards with a

functional VHT

34 Children below 5 years

who are stunted

35 Children below 5 years

who are under weight

Table 4: HSDP 2015/16 – 2019/20 Indicator Matrix

Objective Key Result Area Indicator Data

collection

method

Frequency

of data

collection

Agency

Responsible

Baseline Annual Targets

2015/16 2016/17 2017/18 2018/19 2019/20

To accelerate

movement

towards UHC

with essential

health and

related services

needed for

promotion of a

healthy and

productive life.

HEALTH IMPACT

Health impact

trends

Maternal Mortality Ratio (per 100,000)

Survey Every 5

years

UBOS 438 (2011)

425 320

Neonatal Mortality Rate (per 1,000)

Survey Every 5

years

UBOS 26 (2011)

26 16

Infant Mortality rate (per 1,000)

Survey Every 5 years

UBOS 54 (2011)

50 44

Under five mortality rate (per 1,000)

Survey Every 5

years

UBOS 90 (2011)

64 51

Total Fertility Rate Survey Every 5

years

UBOS 6.2 (2011)

5.9 5.1

Adolescent Pregnancy Rate

Survey Every 5

years

UBOS 24% (2011)

23% 14%

To contribute to

the production

of a healthy

human capital

for wealth

creation through

provision of

equitable, safe

and sustainable

health services.

HEALTH & RELATED SERVICES OUTCOME

TARGETS

Communicable

disease prevention

& control

ART Coverage HMIS Annually MoH 42% (2013/14)

57% 65% 72% 80% 80%

HIV+ pregnant women not

on HAART receiving

ARVs for eMTCT during

pregnancy, labour,

delivery and postpartum

HMIS Annually MoH 72% (2013/14)

85% 87% 90% 95% 95%

TB Case Detection Rate (all forms)

HMIS Annually MoH 80% (2014/15)

73% 75% 78% 81% 84%

IPT3 or more doses coverage for pregnant women

HMIS Annually MoH 50% (2013/14)

64% 71% 79% 86% 93%

In patient malaria deaths per 100,000 persons per year

HMIS Annually MoH 30 (2013/14)

13 5 5 5 5

33 | P a g e

Objective Key Result Area Indicator Data

collection

method

Frequency

of data

collection

Agency

Responsible

Baseline Annual Targets

2015/16 2016/17 2017/18 2018/19 2019/20

Malaria cases per 1,000 persons per year

HMIS Annually MoH 460

(2013/14)

373 329 285 242 198

Under-five Vitamin A second dose coverage

HMIS Annually MoH 26.6% (2013/14)

58% 60% 62% 64% 66%

DPT3HibHeb3 coverage HMIS Annually MoH 93% (2013/14)

95% 95% 95% 97% 97%

Measles coverage under 1 year

HMIS Annually MoH 87% (2013/14)

90% 92% 95% 95% 95%

Essential clinical

and rehabilitative

care

Bed occupancy rate (Hospitals & HC IVs)

HMIS Annually MoH 59% (2013/14)

62% 70% 80% 85% 90%

MoH 50% (2013/14)

55% 60% 65% 70% 75%

Average length of stay (Hospitals & HC IVs)

HMIS Annually MoH Hospital 4 (2013/14)

4 3 3 3 3

MoH HC IV 3 (2013/14)

3 3 3 3 3

Contraceptive Prevalence Rate

Survey Every 5 years

MoH 30% (2011)

39% 50%

Couple years of protection HMIS Annually MoH 4,074,673 (2013/14)

4.3 M 4.4 M 4.5 M 4.6 M 4.7 M

ANC 4 coverage HMIS Annually MoH 32.4% (2013/14)

37.5% 40% 42.5% 45% 47.5%

Health Facility deliveries HMIS Annually MoH 44.4% (2013/14)

54% 56% 58% 60% 64%

HC IVs offering CEmOC Services

HMIS Annually MoH 37% (2013/14)

55% 55% 57% 59% 60%

HEALTH INVESTMENT OUTPUT TARGETS

Health

Infrastructure

New OPD utilization rate HMIS Annually MoH 1.0

(2013/14)

1.1 1.2 1.3 1.4 1.5

Hospital (inpatient) admissions per 100 population

HMIS Annually MoH 6

(2013/14)

7 8 9 10 10

Objective Key Result Area Indicator Data

collection

method

Frequency

of data

collection

Agency

Responsible

Baseline Annual Targets

2015/16 2016/17 2017/18 2018/19 2019/20

Population living within

5km of a health facility

Household

Survey

Annually MoH 75% 77% 80% 82% 84% 85%

Medicines and

health supplies

Availability of a basket of

commodities in the

previous quarter (% of

facilities that had over

95%)

HMIS Annually MoH To be

determined

86% 100% 100% 100% 100%

Improving quality of

care

Facility based fresh still births (per 1,000 deliveries)

HMIS Annually MoH 16 (2013/14)

15 14 13 12 11

Maternal deaths among 100,000 health facility deliveries

HMIS Annually MoH 132

(2013/14)

114 110 106 102 98

Maternal death reviews HMIS Annually MoH 33.3% (2012/13)

38% 45% 52% 59% 65%

Under five deaths among 1,000 under 5 admissions

HMIS Annually MoH 17

(2013/14)

17.6 17.3 16.9 16.5 16.1

ART Retention rate HMIS Annually MoH 79.7%

(2013/14)

83% 84% 84% 84% 84%

TB Treatment Success

Rate

Program Reports

Annually MoH 80%

(2013

Cohort)

82% 84% 86% 88% 90%

Responsiveness Client satisfaction index Surveys Annually MoH 69%

(2014)

71% 73% 75% 77% 79%

Health Information Timeliness of reporting

(HMIS 105)

HMIS Annually MoH 85% (2013/14)

88% 90% 93% 95% 97%

To increase

financial risk

protection of

households

Health Financing Out of pocket health expenditure as a % of Total Health Expenditure

NHA Annually MoH 37% (2011/12)

37% 35% 33% 31% 30%

General Government allocation for health as %

MTEF Annually MoH 8.7% (2013/14)

10% 12% 14% 15% 15%

35 | P a g e

Objective Key Result Area Indicator Data

collection

method

Frequency

of data

collection

Agency

Responsible

Baseline Annual Targets

2015/16 2016/17 2017/18 2018/19 2019/20

against

impoverishment

due to health

expenditures.

of total government budget

Human Resources Approved posts in public

facilities filled with

qualified personnel

HR

Biannual

Report

Annually MoH 69%

(April

2015)

69% 70% 72% 76% 80%

Number of health workers

(doctors, midwives,

nurses) per 1,000

population

HRIS

Annually MoH Doctors

1:24,725

1:24,725 1:24,500 1:24,300 1:24:000 1:23,500

Midwives

1:11,000

1:11,500 1:11,000 1:10,500 1:10,000 1:9,500

Nurses

1:18,000

(HSSIP

MTR 2013)

1:18,500 1:18,300 1:18,000 1:17,500 1:17,000

To address the

key

determinants of

health through

strengthening

intersectoral

collaboration

and

partnerships.

Social and

economic

determinants of

health

Children below 5 years who are stunted

Survey Every 5 years

UBOS 33% (2011)

31% 29%

Children below 5 years who are under weight

Survey Every 5 years

UBOS 14% (2011)

14% 10%

Health promotion &

environmental

health

Latrine coverage Program Annually MoH 73% (2013/14)

74.6% 76% 78% 80% 82%

Villages/ wards with a functional VHT

Program Annually MoH 72% (2013/14)

75% 80% 85% 85% 85%

3.3.2 Program and Project specific indicators

Specific programs and projects have their own strategic and operational plans that are more detailed

than the HSDP. They are also expected to develop specific program / project M&E Plans which will

include more indicators to measure process, outputs and immediate outcomes. These should

however, be aligned with the goal, objectives and strategic interventions of the HSDP to demonstrate

the effects of health system strengthening to health outcomes.

3.4 Data Sources The data needs of the HSDP are based on agreed performance indicators (core and program specific)

to facilitate M&E, reporting and decision-making. The main data sources will include;

Community generated data will be collected by the approved community based health structures

like the VHTs and Community Health Extension Workers (CHEWs) through the Community Health

Management Information System (CHMIS) linked to the e-HMIS.

Facility generated data will be collected by all public and private health service delivery facilities.

In addition different programs and projects managed at the MoH/national level shall provide

reports to the RC and other stakeholders on program-specific activities. Health projects managed

by implementing partners (DPs and CSOs) at district or community level shall provide reports

through the district health system. This data will be collected routinely using the established HMIS.

Administrative data sources will provide information on health inventories, supervision,

management meetings, logistics management, human resource, financial resource flows and

expenditures at national and sub-national levels.

Population based health surveys mainly carried out by Uganda Bureau of Statistics (UBOS) and

other institutions that generate data relative to populations (population studies) as a whole.

Research Institutions and academia that carry out health systems research, clinical trials and

longitudinal community studies will also provide data for interpretation and possible use by the

sector.

Civil registration and vital statistics system is operated and managed by the NIRA. It is responsible

for registering all vital events including births and deaths.

Population and Household Census is carried out every ten years and will be the primary source

of data on size of the population, its geographic distribution, and the social, demographic and

economic characteristics. Annual projections at national and sub-national level will be provided

by UBOS.

Other surveys carried out depending on the data needs e.g. facility assessments, financial studies.

Overall, the sources of M&E information will be guided by different information needs, particularly

the Government, Parliament, DPs, private sector and the community.

3.4.1 Critical data gaps and weaknesses and how to address these

As noted above in section 1.3 on the situation analysis of the M&E practices and mechanisms, the

sector has registered improvements in most of the M&E system components however, there are still

a number of data gaps and weaknesses in data management as indicated below;

1) Low reporting rates from the private sector.

2) Data quality gaps

3) Delayed and incomplete financial data reporting.

37 | P a g e

4) Underutilization of the existing IT to generate data on health infrastructure, equipment and

commodities, service delivery, and health workforce.

5) Inadequate / irregular funding for surveys.

6) Analysis, synthesis, effective dissemination and use of information to guide policy dialogue and

implementation of health programs remain a challenge.

7) Inadequate HMIS equipment (computers, modems) at hospital and lower levels.

8) Inadequate reporting on the vital statistics (births and deaths)

Recommendations for improving data gaps are;

1) Appropriate and strategic advocacy should be carried out for the private sector to report through

the national system.

2) Increased supervision, monitoring and feedback to districts to address data quality gaps.

3) Improving administrative data sources including systems of tracking financial resource flows and

expenditures to sub-national levels.

4) IT skills development for health workers.

5) Advocacy for appropriate funding (level, mechanisms) for surveys.

6) Use of data should be enhanced through meaningful analysis, appropriate presentation and

effective dissemination.

7) Skills development for data disaggregation. Rational distribution of IT equipment for data

management.

8) Provision and maintenance of HMIS equipment.

9) Scale up birth and death reporting to all districts and health facilities and community through the

Mobile Vital Registration System.

3.5 Data management

3.5.1 Responsibilities for data management

There are various players involved in data management. Data management involves

data collection,

data validation,

feedback,

analysis / synthesis,

dissemination and use.

Table 5 below shows the key data management activities and the responsible persons.

3.5.2 Data collection

The methods of data collection will be a combination of quantitative and qualitative methods using

standardized data collection tools and techniques. Most data in respect of some indicators will be

collected annually, and any survey-based indicators will be collected as predetermined. Efforts will be

made to ensure that data is provided timely as indicated in the indicator matrix.

Table 5: Responsibilities for data management

Activity Responsibility

1. Compile and process administrative data into

departmental / Institutional records (minutes,

inventory) and reports (supervision, activity)

Managers at all level

2. Conduct surveys and evaluations and compile

reports

UBOS, Research Institutions, Program

Managers

3. Compile and submit routine facility HMIS

reports

Medical Records Officers / Health

Information Assistants

4. Receive and enter health statistical data into

the DHIS-2

Biostasticians / Information Officers at

District, HSD and Hospitals

5. Data validation MoH RC, Program Managers, DHTs,

Health Facility Managers

6. Provide feedback on reports submitted MoH RC, DHOs, HSD In-charges, Health

Facility Managers

7. Data analysis and synthesis MoH RC, M&E Officers, DHTs, HSD Team,

Health Facility Managers, Health

Information Officers / Assistants, UBOS,

NIRA

8. Data dissemination DGHS, Program Managers, DHO, HSD I/C,

Health Facility Managers, UBOS, NIRA

3.5.3 Data analysis and synthesis

The framework for reviewing health progress and performance covers the M&E process from routine

performance monitoring, quarterly reviews, annual review and evaluation of all the HSDP indicator

domains (Figure 3 above). Specific questions will have to be answered during the different review

processes, especially the annual reviews, but also the performance monitoring.

Health progress and performance assessment will bring together the different dimensions of

quantitative and qualitative analyses and will include analyses on:

(i) Progress towards the HSDP goals;

(ii) Equity

(iii) Efficiency

(iv) Qualitative analyses of contextual changes; and

(v) Benchmarking.

Data analysis and synthesis will be done at all levels (national, sub-national to community) to enhance

evidence-based decision-making. All relevant data will be synthesized based on determined

parameters (disaggregated) where applicable and analyzed for use at various levels of the sector. The

results obtained will be summarized into a consistent assessment of the health situation and trends,

using key sector performance indicators and targets.

Basic indicator information shall be the national average achievement. This is obtained from collating

all the available information from all reporting units into the national figure. Sub analyses of the

indicator information shall be carried out, to provide information on the impact of multi-dimensional

39 | P a g e

poverty on actual coverage, and health status and financial risk protection achievements. This shall

enable better targeting of strategies to address the multi dimensional poverty issues impacting on the

results being sought.

The required levels of disaggregation may not be possible on an annual basis. As a proxy, therefore,

the sector will use district rankings for the different poverty and equity dimensions to separate

districts with high and low attainment of the respective index. The indicator achievements for the top

quintile of districts will be compared with the achievements of the bottom quintile of districts to

illustrate any differences. The focus of analysis will be on comparing planned results with actual ones,

understanding the reasons for divergences and comparing the performance at different levels.

Information on indicators will therefore be analysed in the following lines;

Overall national achievement

Disaggregation of achievement, by

o Age

o Sex

o Region, and district (new and old)

o Poverty index

o Literacy level

Progress towards HSDP goals: The monitoring and review process will measure the extent to which

the objectives and goals of the HSDP (core indicators and their targets) have been attained. This will

be complemented by a stepwise analysis to assess which policies and programs were successful; from

inputs such as finances and policies to service access and quality, utilization, coverage of interventions,

and health outcomes, financial risk protection and responsiveness.

Qualitative assessment and analysis of contextual change: This takes into account non-health system

changes, such as socio-economic development that affect both implementation and the outcomes

and impact observed. Qualitative information on the leadership, policy environment and context is

crucial to understand how well and by whom government policies are translated into practice. A

systematic and participatory approach will be used to gather, analyze and communicate qualitative

information. This will be combined with quantitative data as basis for a policy dialogue. This will then

become a solid basis to inform planning cycles, regular reviews and M&E. Such analysis shall include:

- Health sector reforms and policies stating the importance of assuring service coverage for all

communities;

- A critical assessment of the health sector or disease program response and adhering to policy

direction to date;

- How cultural and political factors impact on health or multi-sectoral programs and outcomes;

- Assessment of organizational context, leadership and accountability mechanisms;

- Assessment of the regulatory environment and how it enables or hinders improvements in health

systems and program delivery.

Figure 4: Framework for reviewing health progress and performance

Quality of care assessment will be undertaken for all health facilities annually using the Health Facility

Quality of Care Assessment tool. Each facility will have a scorecard to enable district level assessment

of facilities. At national level, the district average score will be used to determine the quality of services

provided.

Benchmarking: Benchmarking refers to comparisons between and within entities to assess

performance. There are different types of benchmarking which may vary according to level of

comparison (national or sub-national comparisons), level of assessment (individual service provider –

facility – care organization – district – region – national), measurement focus (process, outcomes,

quality, performance), and use of data (public reporting, accountability, internal reporting only, self-

learning and improvement).

The following tools will be used for benchmarking:

Program specific scorecards

District League Table to be compiled and disseminated quarterly and annually.

HC IV Performance Assessment compiled annually as part of the AHSPR.

General Hospital Performance Assessment compiled annually.

Regional referral and large PNFP hospital performance assessment (SUO) compiled annually.

Health facility quality of care assessment scorecard.

3.5.4 Reporting

Regular reporting using the HMIS and other sources of data will be used to generate relevant

information at all levels. The analytical outputs that will be produced as the basis for decision-making,

program management, financial disbursements and global reporting are;

41 | P a g e

1) Annual Health Sector Performance Report

2) Annual Health Statistical Abstract

3) Annual Program / Project reports

4) Annual Report to National Planning Authority

5) Quarterly Progress Reports

6) Quarterly fiscal performance reports (OBT reports)

1. The Annual Health Sector Performance Report (AHSPR)

The AHSPR is a mandatory report by the constitution for all MDAs. The AHSPR brings together all data

from different sources, including the facility reporting system, household surveys, administrative data

(minutes, supervision reports, financial reports, SCM reports, HRIS reports, etc) and research studies,

to answer the key questions on progress and performance using the HSDP key output indicators and

health goals. The AHSPR will present a detailed account of annual performance against the key output

and programmatic indicators of the sector strategic plan, comparing current results with results of

previous years, and formulate challenges and recommendations by cluster and program. The AHSPR

should be produced by August of the following Financial Year, and will be a key input to the joint sector

review that should also take place in August/September. Annual sub-national performance reports

shall be presented and discussed at the relevant annual stakeholders’ fora.

2. Annual Health Statistical Abstract

The Annual Health Statistical abstract is compiled from the statistical reports submitted through the

HMIS. This feeds into the UBOS statistical abstract for MDAs ahead of the government planning cycle.

3. Annual Program / Project Reports

Annual program / project reports are a requirement from all programs and projects based on their

specific M&E plans. Progress review reports shall be submitted to the MoH M&E unit before the

annual sector reviews.

4. Annual Sector Report to National Planning Authority

The annual sector report to NPA is required to report progress on implementation of the NDP

intervention matrix by August of the following Financial Year. This feeds into the National

Development Report which is presented to the National Planning Forum chaired by H.E. the President.

5. Quarterly Progress Reports

Regular performance monitoring is key in keeping track of progress in implementation of workplans. All service delivery levels will compile quarterly progress reports to be discussed at the quarterly review meetings. At National Level (MoH), quarterly progress reports will be presented and discussed during the sector quarterly review meetings. Quarterly district performance (Quarterly District League Table) will also be disseminated and discussed at this forum. At sub-national level (Regional and district), quarterly performance reports will be presented and discussed at the quarterly review meetings attended by the key implementers.

6. Quarterly Fiscal Performance Reports

MDAs are required to report quarterly to the MoFPED and OPM on a limited number of indicators that

reflect performance in terms of service delivery outcomes (health service coverage), contributory

outputs and main actions. The quarterly fiscal performance reports are generated from the OBT and

aim at linking inputs to activities, outputs and outcomes. The OBT is also used for financial

implementation monitoring addressing whether or not budgets have been released and spent in line

with the planned activities in the Ministerial Policy Statements and BFPs. Reports are submitted to

MoFPED and OPM and are the basis for subsequent financial releases and quarterly reviews for LGs.

The M&E reports will be validated by key stakeholders to:-

i) Obtain stakeholder insight on the information generated;

ii) Mitigate bias through discussion of the information generated with key M&E strategy actors

and beneficiaries;

iii) Generate consensus on the data findings and gaps; and

iv) Strengthen ownership and commitment to M&E activities.

In order to ensure coordinated reporting and routine feedback on performance to sub-national and

private providers, it is crucial that all service delivery and administrative structures adhere to the

following data flow mechanism.

43 | P a g e

Figure 5: Data Flow Diagram

3.5.5 Data Quality

The objective of Data Quality Audits (Assessments) and Adjustment (DQAA) is to ensure that the data

used by the stakeholders to make decisions is sound and accurate.

3.5.5.1 Data Quality Audits (DQA)

DQAs will be conducted in line at all levels of health care service delivery with the MoH DQA guidelines

and using the approved tools.

Specific efforts will be made to undertake DQAs including: application of the computed

validation/data accuracy index into district annual reports; specific support for outliers; routine

(quarterly) data checks on a sample of districts among others. Regular data verification for facility

based data for accuracy and completeness will be carried out at least monthly by the health facility in-

charges.

The various processes in the DQA will take care of: data collection methods, aggregation and analysis

on a quarterly basis. The key DQA activities are;

1) Training of staff in conducting DQAs

Technical Support

Supervision, DQA,

feedback, M&E

Technical Support

Supervision, DQA,

feedback, M&E

Feedback,

M&E

District Database (DHIS, HRIS, SCMS)

Resource Centre (DHIS, HRIS, SCMS)

UBOS

Implementers Public / Private Health Facilities CSOs/Projects

Reporting,

sharing &

dissemination

Reporting, sharing & dissemination

Data collection,

cleaning,

aggregation,

reporting

Data collection,

cleaning,

aggregation,

reporting

National Reporting to OPM and NPA

DQA,

feedback

2) Conducting regular DQAs

3) Data adjustments

4) Compiling and disseminating DQA reports.

This shall be undertaken by;

Health Facility In-charges at facility and community level.

HSD HMIS Focal Person within the HSD.

District Biostatistician within the districts at the various HSDs and health units.

Technical staff of the MoH/HID and Regional Teams at the district level.

Independent DQAAs will be conducted at the mid-term and end evaluation of the HSDP.

3.6 Data dissemination and use Data will be packaged and disseminated in formats that are determined by management at the various

levels. Service delivery data shall be packaged and displayed at the various levels using the HMIS

formats already provided. The timing of information dissemination should fit in the planning cycles

and needs of the users through reports, score cards, bulletins, case studies, among others.

3.6.1 Decision-support tools and approaches

Data should be presented in usable formats to allow policy makers to make well informed decisions

within the limited timeframe. The MoH will use a variety of decision-support tools and approaches

such as data dashboards, health summary bulletins, health status report cards, and colour-coded data

presentation techniques depending on the audience.

Health summary bulletins -These will be used to disseminate data in the AHSPR and program

reports which provide an overall picture of the health status in the country.

Health status report cards - Will be used to report on HSDP core indicators or program areas to

compare current progress to a target or to past report card trends.

Policy briefs - Will be used for conveying specific evidence-based policy recommendations arising from

the M&E system e.g. research findings.

Data dashboards visually present critical data in summary form so that decisions can be made quickly.

Dashboards assist in the management of the large amounts of data that are being collected by health

programs by tracking key program metrics and displaying trends. This allows users to identify

problems and target specific follow-up activities to improve services.

3.6.2 Data Communication

The MoH will use various communication channels including Information Communication Technology

(ICT) in order to ensure public access to data and reports. Quantitative and qualitative data will be

made publicly accessible through the MoH repository under the RC. The public will also be able to

access health information on the MoH website, www.health.go.ug.

45 | P a g e

The Health Education and Promotion Unit in the MoH to translate data and information according to

the target audience and utilize various communication channels e.g. radio, T.V, video conferencing,

teleconferencing, newsletters, booklets, etc.

3.7 Evaluation Evaluation is defined as a systematic and objective assessment of an ongoing or completed program

or policy, its design, implementation and results. The aim is to determine the relevance and fulfilment

of objectives, development efficiency, effectiveness, impact and sustainability. An evaluation should

answer specific evaluation questions and should provide information that is credible and useful,

enabling the incorporation of lessons learned into the decision-making process of both recipients and

donors. Evaluation also refers to the process of determining the worth or significance of an activity,

policy or program.

Evaluations are separately scheduled activities performed at specific intervals (for example baseline,

midterm or at the end of a program / project). The health sector will address different evaluation

questions through:

Evaluations to assess the performance, efficiency and quality of the priority programs;

Thematic evaluations to assess the success and impact of key initiatives.

3.7.1 Evaluations

The health sector will undertake evaluation of the overall HSDP implementation as well as evaluation

of specific programs, projects, policies and interventions. The evaluations will be commissioned by the

MoH and conducted jointly by the MoH M&E Unit, country partners and international agencies. Where

possible, the evaluations will build on previous progress reviews. The type of evaluation to be planned

for and conducted should reflect the nature and scope of the public investment. For example, pilot

projects that are being conducted amongst a random group of participants shall be selected for impact

evaluation to determine whether or not the investment should be scaled up.

As a minimum requirement, investment projects will be required to conduct the following:

(i) A Baseline study during the preparatory design phase of the project

(ii) A Mid-term review at the mid-point in the project to assess progress against objectives and provide recommendations for corrective measures

(iii) A final evaluation or value-for-money audit will be carried out for key front-line service delivery projects where value for money is identified as a primary criterion.

(iv) All other projects will be subjected to standard rigorous final evaluation.

The MoH through the specific program / project managers will be responsible for the design,

management and follow-up of the program and project evaluations (including baseline and mid-term

reviews). All projects are required to budget for periodic project evaluations. All project evaluations

will be conducted by external evaluators to ensure independence. Value for money will be undertaken

by the Office of the Auditor General.

3.7.2 Thematic Reviews

Thematic reviews examine and aggregate information on a specific thematic area of the HSDP. They

look across the portfolio of investments to examine wider systems effects, including health systems

strengthening and institutional development, gender and human rights. The MoH will undertake

annual of thematic reviews in relation to specific strategic interventions focusing on identifying and

discussing key successes and challenges, identify areas for modification of operations /

implementation and make adjustments to facilitate achievement of the planned targets. These will be

conducted at least once annually.

3.8 Surveys and surveillance

3.8.1 Surveys

During the last 5 years a number of major surveys related to the health sector were undertaken.

These include;

i) UDHS 2011

ii) HIV/AIDS Indicator Survey (AIS) - 2011

iii) Uganda Health Systems Assessment - 2011

iv) National Health Accounts (NHA) 2010/11 and 2012

v) Service Availability and Readiness Assessment (SARA) 2011 and 2012

vi) Non-Communicable Diseases Survey - 2014

vii) Hospital and HC IV Census - 2014

viii) Client Satisfaction Survey - 2014

ix) Malaria Indicator Survey (MIS) 2014

x) TB Prevalence Survey (Ongoing 2015)

During this HSDP period the following major surveys are planned; the UDHS-16, HIV/AIDS Impact

Assessment Survey, NHA and efficiency analyses. Each program will identify and include all surveys to

be conducted in the respective M&E Plans.

3.8.2 Surveillance

The MoH undertakes disease surveillance under the; Integrated Disease Surveillance and Response

(IDSR) and active surveillance of malaria infections as outlined below.

1) Weekly Surveillance

Weekly surveillance reporting is undertaken through the routine HMIS using the weekly surveillance

report code named as HMIS 033b.Every health facility is expected to submit a weekly surveillance

report to the DHO every Monday. The areas of reporting are; Diseases of epidemic potential where

number of cases and deaths for each of the diseases is reported, Malaria case management and the

ACT stock balances and the number of Maternal and Perinatal Deaths that have occurred in the health

facility.

2) Malaria Sentinel Surveillance

Uganda Malaria Surveillance Project developed the sentinel site malaria surveillance project to

provide high quality and timely malaria surveillance data beyond the scope of the HMIS surveillance

system. The aim of this approach is to obtain high quality individual patient data from a limited number

of sites to provide accurate malaria surveillance data for quick response systems, to support ongoing

47 | P a g e

malaria control activities and strengthen HMIS at the sentinel sites. Data is aggregated, analysed and

disseminated monthly.

3.8.3 Research and Special Studies

Health Institutions are encouraged to conduct operational research. The results of these studies are

intended to inform decision making hence contribute to improving delivery of and access to health

care and nutrition services, by developing practical solutions to common, critical problems in

implementing these interventions. Operational surveys shall be carried as planned on approval by the

relevant institutions and findings disseminated through the existing structures. Operational research

shall encompass a wide range of problem-solving techniques and methods applied in the pursuit of

improved decision-making and efficiency.

Operation research (using qualitative and quantitative methodologies) should be focused on

understanding of the reasons why key targets (in particular related to RMNCAH, Malaria, TB and

HIV/AIDS) are not being met, and results be critically discussed and synthesized during mid-tern or

final review of the HSDP.

The Uganda National Health Research Organization (UNHRO) is responsible for coordinating all the

health related research in Uganda. UNHRO will be responsible for mobilizing resources, setting,

developing and disseminating health services research agenda for the 5 year period, commissioning

and organizing health research in collaboration with other research and academic institutions and

NGOs.

4 Progress and Performance Reviews Progress and performance reviews are critical for updating all stakeholders on program progress,

discussing problems and challenges, and developing a consensus on corrective measures or actions.

Progress and performance reviews are part of the governance mechanisms that help ensure

transparency and allow for debate between partners.

Figure 5 illustrates how progress and performance are assessed by annual reviews, which result in

adjustments to annual operational plans by a mid-term review, and by a final evaluation.

Figure 6: Progress and Performance Review of the National Health Strategy

Source: Monitoring and Evaluation of National Health Strategies, WHO 2011

The HSDP monitoring and review process will be interlinked across the different planning entities. This

process will be based on a national platform approach that: uses the district as the unit of design and

analysis; based on continuous monitoring of different levels of indicators; gathers additional data

before, during, and after the period to be assessed by multiple methods; uses several analytical

techniques to deal with various data gaps and biases; and includes interim and summative evaluation

analyses. This implies that information at each level will be provided from the planning entities below

it. Management support, on the other hand, as well as governance/partnership information will be

analyzed at the same level it is to be provided.

Respective reviews will be guided by information developed by the Government management

structures at each level. These will compile the review information with inputs from the other

stakeholders. All reviews will be presented and endorsed at the respective management and

partnership structures for the level. These are illustrated in the table below.

Year 1 Year 2 Year 3 Year 4 Year 5

Operational Plans

Annu

al

Annu

al

Annu

al

Progress and performance reviews and policy dialogue

Annu

al Final and situation

Data analysis and synthesis Systematic qualitative assessment

Surveys Context Facility

data Civil

registration

Administrative

data

Research Leadership Policies

49 | P a g e

Table 6: HSDP Monitoring and Review process

Process Frequency Focus Level

Progress review Quarterly Done by Joint (public + private) Performance Assessment Teams and peers per level or institution. A review of progress against planned activities and key output targets for the workplan

National Regional

District HSD Health Facility

Joint Annual review Annually Done Jointly with key stakeholders,

and planning entities as from HSD level onwards Review progress against annual output targets for the workplan and HSDP targets

National

Regional

Mid Term Review 2nd year of implementation

Done jointly with key stakeholders Review progress against targets for outcome and impact indicators, as well as contextual changes

National

End Review Last year of HSDP - June - Dec 2019

Comprehensive analysis of progress and performance. Includes results of specific research and prospective evaluations from the beginning of the HSDP

National

The performance review process will be one of the learning mechanisms in the sector. For proper

follow up and learning:

i) All performance reviews and evaluations will contain specific, targeted and actionable recommendations.

ii) All target institutions will provide a response to the recommendation(s) within a stipulated timeframe, and outlining a) agreement or disagreement with said recommendation(s), b) proposed action(s) to address said recommendation(s), c) timeframe for implementation of said recommendation(s).

iii) All institutions will be required to maintain a Recommendation Implementation Tracking Plan which will keep track of review and evaluation recommendations, agreed follow-up actions, and status of these actions.

iv) Institutions which have an oversight responsibility on the implementation of public policy will monitor the implementation of agreed actions utilizing the Recommendation Implementation Tracking Plan. These institutions are as follows: OPM overseeing MDAs, Ministries overseeing the LGs.

The budget of collating performance data and conducting sector performance reviews will be provided

for under the sector monitoring budget.

4.1 Progress review

4.1.1 Progress Review at Central Level

The Quarterly Sector Performance Review will assess progress on the quarterly work plans of MoH

departments, semi-autonomous institutions, Health Professional Councils, National Referral Hospitals

(NRHs, Mulago and Butabika Hospitals) and LGs. These quarterly reviews are focused on the indicators

and targets in the operational plans and these are mainly input, process and output indicators. The

aim of these reviews is to:

Assess progress made on action points/recommendations of previous quarterly reviews.

Assess implementation of planned activities against set targets.

Highlight budget performance during the quarter (utilization against allocation).

Propose strategies to address challenges in subsequent quarters.

This shall be carried out using a standardized reporting format (Ref. Annex 8.1) during the quarterly

performance review workshops, where reports produced by each department/institution and NRHs

will be presented and discussed in plenary sessions. The report should focus on the results achieved

against the set targets in the workplans. The meetings will be attended by representatives of all

reporting units.

The departmental and institutional reports are to be compiled into quarterly sector performance

reports which will feed into the AHSPR. The QAD will be responsible for organizing the quarterly review

meetings, compiling and disseminating the quarterly reports.

4.1.2 Progress Review at Regional Level

Regional Performance Monitoring Teams are responsible for organising quarterly performance review

workshops, where reports produced by each district in the catchment area will be presented and

discussed in plenary sessions. The meetings will be attended by representatives of all reporting

districts and key stakeholders.

These stakeholders will include;

- Members of Parliament

- LC Chairpersons

- Secretary for Health / Social Services Committee

- DHT Representatives

- Representatives of Public and private health facility managers

- Representatives of health related sectors

- Representatives of HDPs/ IPs

- Representatives of CSOs

- Community, and / or cultural leaders

MoH will provide standardized planning and reporting formats, tools and process to all the regional

coordinators, for guidance.

51 | P a g e

4.1.3 Progress Review at District Level

The DHMT shall review the district performance reports focusing on progress of implementation of

the workplans, timeliness, completeness and accuracy of the reports. The district quarterly and annual

performance reports will be presented and discussed during the quarterly district review meetings.

4.1.4 Performance Monitoring and Review at HSD Level

Like the district, each HSD Team shall review their HSD reports performance focusing on

implementation of the HSD workplans, timeliness, completeness and accuracy of the reports. The HSD

quarterly and annual performance reports will be presented and discussed during the quarterly HSD

review meetings and stakeholders’ forum (HSD / Constituency Health Assembly).

HSD stakeholders’ fora

These stakeholders include;

- Members of Parliament

- RDC

- Political, administrative and technical leadership

- Public and private health providers operating

- IPs supporting, or facilitating activities

- Critical civil society groups operating

- Representatives of health related sectors

4.1.5 Progress Review at Health Facility Level

Performance monitoring and review in health facilities (including private health facilities and

community) will be carried out using standard planning and reporting formats (Ref. Annex 8.1). During

the monthly health facility meetings performance review should focus on progress in implementation

of the workplan, timeliness, completeness and accuracy of the reports. The health facility quarterly

assessment reports will be used for performance review during the quarterly HUMC or Hospital Board

meetings. Annual report to be discussed at the annual health facility stakeholders' fora.

Health facility stakeholders’ fora

Each health facility in the country has a defined catchment area, for which it is responsible for

coordinating delivery of services. All stakeholders in the catchment area of the facility shall come

together to discuss health and health related issues affecting them at least once a year. These

stakeholders include;

- In-charges of the departments where applicable

- Community and / or political leaders

- Representatives of health related sectors

- IPs supporting, or facilitating activities

- CBOs operating

- Community Health Workers

The fora shall use health unit data to discuss performance of health within the catchment area, and

agree on priorities to guide the facility and other service providers in their respective planning and

implementation processes. In addition, the HUMCs and Hospital Boards are expected to use the M&E

results to sensitize the community and accountability through the community dialogue meetings.

4.1.6 Progress Review at Community Level

Stakeholders in the Sub counties shall come together to discuss health and health related issues

affecting their SC quarterly.

These stakeholders include;

- Political, administrative and technical leadership

- Public and private health providers

- HDPs / IPs

- Critical civil society groups

- Representatives of health related sectors

The fora shall use SC data to discuss performance within the SC, and agree on priorities to guide

facilities and other service providers in their respective planning and implementation processes. In

addition, the SCs are expected to use the M&E results to sensitize the community and accountability

through the community dialogue meetings.

4.1.7 Performance Monitoring for CSOs and Private Sector

CSOs and the private sector contribute significantly to health service delivery in the country. Most

times their input is not captured in the sector performance reports. In order to measure their

contribution to the overall sector performance they will be required to report to the relevant sector

entities (Districts / MoH programs) using the existing monitoring tools. They will as well be required

to participate in progress reviews following the above review process. The CSO and Private Providers’

umbrella bodies will play a coordination role in monitoring all CSOs and private sector providers to

ensure alignment with national priorities. MoH will support this process through providing the

necessary M&E tools.

4.2 Annual review The annual performance review will be conducted through the Joint Review Mission (JRM) which is

the national forum that brings together all relevant stakeholders in the health sector. The review is

focused on input, process, output and coverage indicators and targets specified in the annual plans.

Performance of the previous fiscal year is assessed, actions and plans for the year ahead (current

year+1) are also determined. The actions and plans should be clearly spelt out in amendments to the

HSDP operations, and the Budget Framework Papers and Ministerial Policy Statements of the Ministry.

Annual Sector Reviews should be completed by the 30th September each year, to ensure that the

findings feed into the planning and budget process of the coming year, and annual reporting to Cabinet

and Parliament. (National Policy on Public Sector Monitoring and Evaluation, October 2010). The main

purpose of the annual review is to;

(a) Receive and discuss the previous financial year sector annual performance report against

priorities.

(b) Raise and respond to issues arising from the AHSPR.

(c) Identify and agree on sector priorities for following financial year.

53 | P a g e

The annual review shall be organized by the MoH in collaboration with HDPs. The participants include;

key Policy makers (Parliamentary Social Services Committee), Health Service Commission, key MoH

staff and departments, Professional Councils, representatives of LGs, other Health Related Ministries

(e.g. MoFPED, MoLG, MoPS, Ministry of Education and Sports, Ministry of Lands, Water and

Environment, Ministry of Agriculture, Animal Industries and Fisheries, Ministry of Gender, Labour and

Social Development), UN agencies, DPs, non-state implementing partners (PNFPs, CSOs, private

providers), representatives of Health Consumers Organizations and CSO advocacy organizations.

The proceedings of the JRM will be documented in the Aide Memoire to be signed by the MoH, DPs,

Private Sector and CSOs. The budget for conducting annual sector reviews will be provided for under

the sector monitoring budget.

4.3 Mid Term Review A Mid-Term review of the HSDP 2015/16 – 2019/20 will be conducted after the second year (2016/17)

of implementation and report presented at the JRM. The purpose of the MTR is to review the progress

of implementation; identify and propose adjustments to the HSDP and other government policies as

required. The analysis will focus on progress of the entire sector against planned impact, but will also

include an assessment of inputs, processes, outputs and outcomes, using the HSDP indicators (key and

others). The main result will be a list of recommendations for the remaining HSDP years. The MTR

performance shall be used to review the HSDP key indicators for the remaining HSDP implementation

period. This will be an internal, joint exercise involving all stakeholders. External expertise may be

sought to support some aspects of the review process.

4.4 End Term Review The HSDP final evaluation will be conducted in the final year of implementation (2019/20). The

findings and recommendations will be used for the formulation of the next strategic plan. The analysis

will focus on progress of the entire sector against planned impact and will also include an assessment

of inputs, processes, outputs and outcomes, using the HSDP indicators (key and others). It will focus

on expected and achieved accomplishments, examining the results chain, processes, contextual

factors and causality, in order to understand achievements or the lack thereof. The evaluation will

have to answer questions of attribution (What made the difference?) and counterfactual (what would

have happened if we had not done A or B?) and take into account contextual changes (economic

growth, social changes, environmental factors etc.), as well as policies and resource flows. The reviews

will be conducted by a team of independent in-country institutions in close collaboration with

international consultants. The End Term Evaluation findings shall guide the development of the

subsequent HSDP.

4.5 Linkages between health sector reviews, program / project specific

reviews, and global reporting

4.5.1 Program / project specific reviews

Detailed program / project specific reviews shall be linked to the overall health sector review and

contribute to it. Program / project specific reviews shall be conducted prior to the overall health sector

review, and help inform the content of the health sector review in relation to that specific program

area. To ensure credibility and an objective view of progress, the specific program reviews will involve

all stakeholders.

4.5.2 Global reporting

Global reporting requirements shall be based on ongoing country processes of data generation,

compilation, analysis and synthesis, communication and use for decision making. This has been clearly

spelt out in the Country Compact for implementation of the HSDP. The common M&E platform (See

Figure 7) serves as the guide for all M&E related processes such as the health sector component of

the Sector Wide Approaches (SWAp) and IHP+, for program activities supported by Global Funding

Agencies e.g. GAVI, the Global Fund for TB, AIDs and Malaria, and other DPs, and for disease and

program / project specific needs.

Figure 7: Country-led platform for monitoring & review of the national health strategy

Source: WHO 2011

Country data Information generation

Analysis and synthesis

Data quality assessment

Ind

epe

nd

ent

revi

ew

s

Statistical

repo

rts

Programme reports

Monitoring reports

Review

s

Re

view

s

SDG / UN

Reporting

Program

Reporting

e.g TB,

RMNCAH,

HIV, etc

Evaluation

GAVI

Reporting

Global Fund

Reporting

PEPFAR

Reporting

External validation and estimates Common standards and tools

Minimization of

reporting requirements Harmonization of

reporting requirements

55 | P a g e

4.6 Decision-making processes and tools for remedial action 4.6.1 Decision-making processes

The outcomes of the review process will be used to guide decision making during planning, resource

mobilisation and allocation, accountability, policy dialogue, review and development.

1. Planning

Periodically, the achievements, best practices and challenges identified during the review process

will be used to determine priorities for the subsequent planning period and set new performance

targets.

2. Resource Mobilization and Allocation

The MoH will move towards evidence based priority setting to guide resource mobilisation and more

effective and efficient resource allocation levels, rather than proportionate allocation of resources.

Other approaches like Results Based Financing which relies on quality data, will be institutionalized.

3. Accountability

All public and private institutions will be held accountable for the achievement of targets set and

agreed upon annually as documented in policy statements and work plans. These achievements will,

where appropriate, include targets linked to the charters and service delivery standards. For instance,

at the HSD level, a new concept of Constituency /HSD Task Forces will be rolled out in the country; all

HSDs will be required to hold biannual assemblies where they will present and discuss their

performance with key stakeholders.

4. Policy Dialogue, Review and Development

At national level policy issues / problems are raised through reports presented by SMC to HPAC which

meets every two months. At district and sub-county level technical committees will be responsible for

generating policy issues / problems for discussion during the Technical Planning Committee meetings

and forwarding to the respective Local Council (LC) Committees. The LC will be responsible for

analyzing and making policy recommendations to the District / Sub-county Council. The policy analysis

process shall take into consideration the "Wants", "Needs" and "Feasibility" of strategic interventions.

The findings shall be packaged to enable policy dialogue, review and development.

5 Implementation Plan M&E is not only a semi-annual or quarterly event, but a set of activities all of which provide data, which add to the periodic data collections, reviews and data

use. The implementation plan is annualized for the key M&E activities, showing the responsible persons or departments and frequency of the activities.

Table 7: M&E Implementation Plan

Activity Responsible

Person/Department

Frequency

2015/16 2016/17 2017/18 2018/19 2019/20

Compilation and

submission of

reports

Provide the data collection tools ACHS HID Ongoing Ongoing Ongoing Ongoing Ongoing

Weekly Surveillance data Surveillance Focal Person Weekly Weekly Weekly Weekly Weekly HMIS 105 (OPD) & HMIS 108 (Inpatient)

HMIS 123 (District OPD) & HMIS 124 (District

Inpatient) Births and deaths report

Health Unit I/C

HSD I/C, DHO

Monthly Monthly Monthly Monthly Monthly

HMIS 106a Health Facility Manager Quarterly Quarterly Quarterly Quarterly Quarterly Annual Health Unit performance report (HMIS 107) Health Facility Manager July July July July July

Annual District performance report (HMIS 127) DHO July July July July July

OBT (Financial performance) Report All Accounting Officers Quarterly Quarterly Quarterly Quarterly Quarterly Quarterly sector performance report CHS QAD Quarterly Quarterly Quarterly Quarterly Quarterly

Annual program / project reports Program Managers July July July July July Annual Sector Statistical Abstract ACHS HID Sept Sept Sept Sept Sept

Annual health sector performance report DGHS Sept Sept Sept Sept Sept

Annual Government Performance Report Director Planning Sept Sept Sept Sept Sept Annual sector performance report for the NDP Director Planning Sept Sept Sept Sept Sept

Data Quality

Assurance

Health Facility Data verification Facility Manager/ HSD

Manager / DHO

Monthly Monthly Monthly Monthly Monthly

District Data Quality Audit ACHS HID Quarterly Quarterly Quarterly Quarterly Quarterly Data Quality Assessment and Adjustment University May May May May May

Feedback Feedback to all reporting entities on reports submitted

to higher levels

All Facility / M&E Officers

ACHS HID

Monthly Monthly Monthly Monthly Monthly

Performance

Reviews

Performance Review meetings HSD Manager/ DHO / RPMTs /

CHS QAD

Quarterly Quarterly Quarterly Quarterly Quarterly

Program review meetings Program Managers Annual Annual Annual Annual Annual

57 | P a g e

Activity Responsible

Person/Department

Frequency

2015/16 2016/17 2017/18 2018/19 2019/20

Technical Review Meeting CHS P April April April April April

Joint Review Mission DGHS Sep Sep Sep Sep Sep

Conduct Surveys Service Availability Readiness Assessment CHS QA May - June May - June

Efficiency analysis ACHS B&F Apr - Jun Apr - Jun Apr - Jun Apr - Jun Apr - Jun

National Health Accounts CHS P Jul - Dec Jul - Dec Jul - Dec Jul - Dec Jul - Dec HIV/AIDS Epidemiological Surveillance PM ACP Annual Annual Annual Annual Annual HIV/AIDS Indicator Survey ACHS CD Feb to Jun

National Client Satisfaction Survey QAD Jan Jan

Uganda Demographic Health Survey-6 UBOS Jan - Jun

Panel Survey (Annual) UBOS Jul - Dec Jul - Dec Jul - Dec Jul - Dec Jul - Dec

National Household Survey (Every 3 years) UBOS Jul - Dec Jul - Dec

National Service Delivery Survey (Every 5 years) UBOS Jul to Dec

Evaluations Specific program / project evaluations (Baseline, Mid &

End Term)

Program/ Project Managers Ongoing Ongoing Ongoing Ongoing Ongoing

HSDP Mid Term Review DGHS Sept to April

HSDP End Term Evaluation DGHS Jan to Jun

Capacity Building Printing of the HSDP M&E Plan ACHS QA April

Dissemination of the HSDP M&E Plan ACHS QA May - Dec

Conduct Data Use Workshop ACHS HID Annual Annual Annual Annual Annual

Conduct M&E in-service training and mentorship at all

levels

ACHS QA / HID / RPMTs /

Managers / DHO

Ongoing Ongoing Ongoing Ongoing Ongoing

Recruitment and deployment of required human

resource for effective M&E at all levels.

HSC / DSC Ongoing Ongoing Ongoing Ongoing Ongoing

M&E Plan

Monitoring &

Evaluation

SME&R TWG meetings CHS QA Monthly Monthly Monthly Monthly Monthly

M&E technical supervision to Institutions and LGs ACHS QA / HID / RPMTs /

M&E Officers / DHO

Quarterly Quarterly Quarterly Quarterly Quarterly

M&E Stakeholders meetings CHS QA / HID / RPMTs / DHOs Jul Jul Jul Jul Jul

Assessment of the M&E Plan implementation CHS QA April - Jun April - Jun

6 Monitoring the Implementation of the M&E Plan for the HSDP Existing sector M&E structures (Ref. to chapter 2.2 - roles and responsibilities) will be used for

monitoring implementation of the M&E Plan for the HSDP. Support supervision, meetings and periodic

reports will be used for monitoring at the various levels as outlined below.

At national Level:

1. Monthly SME&R TWG meetings will be carried out to track progress on implementation of the

M&E Plan.

2. Quarterly M&E technical supervision to institutions and LGs.

3. Quarterly M&E progress reports to be presented at the sector review meetings.

4. Annual M&E stakeholders meetings.

5. Mid-term assessment and end evaluation of the implementation of HSDP M&E plan.

At district Level:

1. Monthly District Health Team Meetings to track progress on implementation of the district M&E

plan.

2. Quarterly DHT supervision to HSDs

3. Quarterly DHMT performance review meetings

At Health Sub-District level:

1. Monthly HSD Team Meetings to track progress on implementation of the HSD M&E plan.

2. Quarterly HSD supervision to health facilities.

3. Quarterly HSD Team performance review meetings.

At facility level:

1. Monthly health facility meetings to track progress on implementation of the facility workplan.

2. Quarterly health facility meetings.

3. Quarterly supervision to VHTs and CHEWs.

At community level:

1. Quarterly CHEW and VHT review meetings.

59 | P a g e

6.1 M&E Plan Implementation Monitoring Indicators The indicators and levels of reporting for monitoring implementation of this M&E Plan are shown in

the table below;

No. Indicator Level of reporting

1. % of period reports (weekly, quarterly, annual compiled and

submitted timely

All levels

2. % of monthly and quarterly HMIS reports verified by the

Health Facility In-charges before submission to the

DHO/MoH.

Facility

3. % of quarterly data validations conducted. National and district

4. % of districts providing quarterly feedback to health facilities. District / HSD

5. % of quarterly feedback reports to districts. National

6. % of planned performance reviews conducted. All levels

7. Number /% of planned evaluations conducted and

disseminated.

National

8. % of districts oriented on the M&E Plan for the HSDP

(dissemination).

National

9. Number of health workers receiving in-service training

(workshops, seminar or CMEs) in M&E.

All levels

7 Budget Cost estimates for implementation of the M&E plan are based on the HSDP costs estimates for M&E.

Costs for managing the health information systems infrastructure are captured under the HMIS

Strategic Plan by the MoH/HID and respective health information systems (HRIS, SCMS) budgets. Over

the 5-year period, an estimated total of UGX 89,653,447,000/- will be required for implementation of

this plan. Of this 35.4% is required for compilation and submission of performance reports, 31.5% for

performance reviews and 23% for surveys (Table 8). The detailed budget is in Annex 8.2.

The funds for the Plan implementation will be managed under the responsible departments, programs, districts and stakeholders.

Table 8: Budget Summary for the M&E plan activities

Activity Amount UShs. '000 Total %

2015/16 2016/17 2017/18 2018/19 2019/20 UGX USD

Compilation and submission of

performance reports

6,043,205 6,194,285 6,349,142 6,507,871 6,670,568 31,765,071 8,823.63 35.4%

Data Quality Assurance &

feedback

123,760 126,854 190,025 133,276 199,608 773,524 214.87 0.9%

Performance Reviews 5,469,000 5,349,475 5,745,712 5,620,292 6,036,424 28,220,904 7,839.14 31.5%

Surveys 8,012,600 2,984,915 3,159,538 3,136,026 3,324,427 20,617,506 5,727.09 23.0%

Evaluations 200,000 200,000 500,000 200,000 550,000 1,650,000 458.33 1.8%

Dissemination 648,600 664,815 701,435 698,471 735,933 3,449,255 958.13 3.8%

Capacity Building 469,000 316,000 181,200 187,430 195,691 1,293,321 359.26 1.4%

Monitoring & Evaluation 358,400 367,360 376,544 385,958 395,607 1,883,868 523.30 2.1%

Total 21,324,565 16,203,704 17,203,597 16,869,324 18,108,257 89,653,447 24,903.74 100%

Exchange USD: UShs. 1:3,400

I | P a g e

8 Annex

8.1 Quarterly Output Report Format for the Workplans & Budgets

Output Code Output

Description

Planned Annual

Outputs

Planned Output

Target (Qty and

Location)

Planned

Expenditure

(Ush)

Actual Output (Qty

and Location)

Actual

Expenditure (Ush)

Reasons for

variation in

performance

I | P a g e

8.2 Detailed 5 Year Budget for Implementation of the Plan

Activity Output Indicator Target Unit Cost

UGX '000

No. of

Units

Amount UGX. '000 Total

2015/16 2016/17 2017/18 2018/19 2019/20

Compilation and submission of performance reports

HMIS tools (registers & report

forms)

No. of tools printed 1 6,000,000 1 6,000,000 6,150,000 6,303,750 6,461,344 6,622,877 31,537,971

Compile Annual project reports No. of annual reports 20 15.00 10 3,000 3,075 3,152 3,231 3,311 15,769

Compile & submit Annual Central

Level performance report (OBT)

No. of annual reports 20 10 1 200 205 210 215 221 1,051

Compile & submit Annual health

sector performance report

Annual Report 1 40,000 1 40,000 41,000 42,025 43,076 44,153 210,253

Compile & submit statistical abstract No. of Reports 1 5 1 5 5 5 5 6 26

Subtotal 6,043,205 6,194,285 6,349,142 6,507,871 6,670,568 31,765,071

Data Quality Assurance

Health Facility Data Quality Audit No. of reports 4 150 112 67,200 68,880 70,602 72,367 74,176 353,225

District Data Validation in selected

districts

No. of reports 4 700 20 56,000 57,400 58,835 60,306 61,814 294,354

Data Quality Assessment and

Adjustment

No. of reports 1 60,000 1 0 60,000 63,000 123,000

Subtotal 123,200 126,280 189,437 132,673 198,990 770,580

Feedback

Routine feedback to sub-national

and key stakeholders on reports

submitted to Resource Center -

Quarterly

No. of feedback

reports to all

stakeholders

4 1 140 560 574 588 603 618 2,944

Subtotal 560 574 588 603 618 2,944

Performance Reviews

Quarterly Health Facility Review

meetings

No. of meetings 4 100 3,500 1,400,000 1,435,000 1,470,875 1,507,647 1,545,338 7,358,860

Quarterly Sub County Review

meetings

No. of meetings 4 250 1,117 1,117,000 1,144,925 1,173,548 1,202,887 1,232,959 5,871,319

Quarterly HSD Review meetings No. of meetings 4 1,000 238 952,000 975,800 1,000,195 1,025,200 1,050,830 5,004,025

Quarterly District Review meetings No. of meetings 4 2,000 120 960,000 984,000 1,008,600 1,033,815 1,059,660 5,046,075

Activity Output Indicator Target Unit Cost

UGX '000

No. of

Units

Amount UGX. '000 Total

2015/16 2016/17 2017/18 2018/19 2019/20

Quarterly Regional Review

meetings

No. of meetings 4 5,000 12 240,000 246,000 252,150 258,454 264,915 1,261,519

Quarterly Sector Performance

review meetings

No. of meetings 4 25,000 1 100,000 102,500 105,063 107,689 110,381 525,633

Technical Review Meeting No. of meetings 1 50,000 1 50,000 51,250 52,531 53,845 55,191 262,816

Joint Review Mission No. of JRM 1 400,000 1 400,000 410,000 420,250 430,756 441,525 2,102,531

National Health Assembly No. of NHA 1 250,000 1 250,000 0 262,500 0 275,625 788,125

Subtotal 5,469,000 5,349,475 5,745,712 5,620,292 6,036,424 28,220,904

Surveys

ANC Sentinel Surveillance (HIV) No. of reports 1 50,000 1 50,000 51,250 52,531 53,845 55,191 262,816

HIV/AIDS Indicator Survey Survey report 1 5,000,000 1 5,000,000 0 0 0 0 5,000,000

SARA Survey report 1 120,000 1 120,000 123,000 126,075 129,227 132,458 630,759

Health Facility Quality of Care

Assessments

Report 1 23,000 112 2,576,000 2,640,400 2,706,410 2,774,070 2,843,422 13,540,302

Health Facility Client Satisfaction

Surveys (Hospitals & HC IVs)

CSS Report 1 200 333 66,600 68,265 69,972 71,721 73,514 350,071

National Client Satisfaction Survey NCSS Report 1 100,000 1 100,000 0 100,000 0 110,000 310,000

Other operational surveys Reports Assorted 10,000 10 100,000 102,000 104,550 107,164 109,843 523,557

Subtotal 8,012,600 2,984,915 3,159,538 3,136,026 3,324,427 20,617,506

Evaluations

Specific program / project

evaluations

No. of evaluations Assorted 100,000 10 200,000 200,000 200,000 200,000 200,000 1,000,000

HSDP Mid Term Review MTR 1 300,000 1 0 300,000 0 0 0 300,000

HSDP End Term Evaluation ETR Report 1 350,000 1 0 0 0 0 350,000 350,000

Subtotal 200,000 200,000 500,000 200,000 550,000 1,650,000

Dissemination

Weekly surveillance bulletins No. of bulletins

disseminated

52 1,000 12 624,000 639,600 655,590 671,980 688,779 3,279,949

Quarterly performance reports No. of reports printed

& disseminated

200 20 1 4,000 4,100 4,203 4,308 4,415 21,025

Annual Health Sector Performance

Report

AHSPR printed &

disseminated

1 20 1,000 20,000 20,500 21,013 21,538 22,076 105,127

III | P a g e

Activity Output Indicator Target Unit Cost

UGX '000

No. of

Units

Amount UGX. '000 Total

2015/16 2016/17 2017/18 2018/19 2019/20

Mid term review Report MTR printed &

disseminated

1 20 1,000 0 0 20,000 0 0 20,000

End Term Evaluation Report printed

and disseminated

ETE Report

disseminated

1 20 1,000 0 0 0 0 20,000 20,000

Survey Reports No. of survey reports

disseminated

Assorted 20 30 600 615 630 646 662 3,154

Subtotal 648,600 664,815 701,435 698,471 735,933 3,449,255

Capacity Building

Printing of the HSDP M&E Plan No. of plans printed 1 25 6,000 150,000 0 0 0 0 150,000

Develop and print an indicator

booklet

Booklet developed 1 10 1,000 10,000 0 0 0 0 10,000

Orientation and Dissemination of the

M&E Plan at district level

No. of dissemination

meetings

1 2,500 112 140,000 140,000 0 0 0 280,000

In-service training in M&E No. of health workers

trained

112 500 1 56,000 58,000 60,000 62,000 64,000 244,000

National Data use workshop Workshop conducted 1 65,000 1 65,000 70,000 72,000 75,000 80,000 362,000

Vehicle Maintenance and Operation No. of vehicles

maintained

12 1,000 4 48,000 48,000 49,200 50,430 51,691 247,321

Birth and Death registration (MVRS) No. of health workers

trained

500 125 100 12,500 12,500 12,500 12,500 12,500 62,500

Subtotal 481,500 328,500 193,700 199,930 208,191 1,355,821

Monitoring & Evaluation

M&E technical supervision and

mentoring to Institutions and LGs

No. of reports 4 800 112 358,400 367,360 376,544 385,958 395,607 1,883,868

Subtotal 358,400 367,360 376,544 385,958 395,607 1,883,868

Total 21,337,065 16,216,204 17,216,097 16,881,824 18,120,757 89,715,947

8.3 District League Table Indicators Indicator Category

Indicator National Baseline

Target 2019/20

Human Resources

1. Approved posts filled with qualified personnel in public health facilities (%)

71% (2014/15)

80%

Service quality 2. Facility based fresh still births (per 1,000 deliveries)

16 / 1,000 (2013/14)

11/1,000

3. Maternal deaths reported which are audited / reviewed (%)

33% (2012/13)

65%

4. TB Treatment Success Rate (%) 79% (2013 cohort)

90%

5. Patients diagnosed with malaria that are laboratory confirmed (%)

NA 100%

Coverage of interventions

6. PCV3 Coverage (%) 79% (2014/15)

94%

7. ANC 4 (%) 36.6% (2014/15)

47.5%

8. IPT3+ (%) NA

93%

9. HIV+ women receiving ARVs for eMTCT during pregnancy & delivery (%)

72% (2013/14)

95%

10. Deliveries in facilities (%) 52.7% (2014/15)

64%

11. Latrine Coverage (%) 77% (2014/15)

82%

Health Information

12. Timeliness of HMIS (105) reporting (%) 88% (2014/15)

97%

13. Completeness of HMIS report 105 (%) 83% (2014/15)

97%

Management 14. Timeliness of quarterly district financial reporting (%)

NA 60%

15. Average district quality of care score (%) NA 75%

V | P a g e

8.4 Indicator Definitions for the HSDP Key Performance Indicators 1. Maternal Mortality Ratio

Definition: Annual number of female deaths from any cause related to or aggravated by

pregnancy or its management (excluding accidental or incidental causes) during

pregnancy and childbirth or within 42 days of termination of pregnancy,

irrespective of the duration and site of the pregnancy, per 100 000 live births,

for a specified time period.

Numerator: Number of maternal deaths

Denominator: Number of live births

Calculation: Numerator

__________________ x 100, 000

Denominator

Data Sources: Household surveys – e.g. UDHS

Frequency of

Analysis,

Interpretation

and Response

Every 5 years

Disaggregation By age and parity, location (major regions), and socio-economic characteristics

(e.g. education level, wealth quintile).

2. Neonatal Mortality Rate

Definition: Number of deaths during the first 28 completed days of life per 1000 live births

in a given year or other period.

Numerator: The number of neonatal deaths in a given period

Denominator: The total number of live births surveyed

Calculation: Numerator

__________________ x 1,000

Denominator

Data Sources: Household surveys – e.g. UDHS.

Frequency of

Analysis,

Interpretation

and Response

Every 5 years

Disaggregation By sex, location (major regions), and socio-economic characteristics (e.g.

mother’s education level, wealth quintile).

3. Infant Mortality Rate

Definition: Probability of a child born in a specific year or period dying before reaching the

age of one, if subject to age-specific mortality rates of that period.

Numerator: The number of infant deaths

Denominator: The total number of live births that year

Calculation: Numerator

__________________ x 1,000

Denominator Data Sources: Household surveys – e.g. UDHS.

Frequency of

Analysis,

Interpretation

and Response

Every 5 years

Disaggregation By sex, location (major regions), and socio-economic characteristics (e.g. mother’s

education level, wealth quintile).

4. Under Five Mortality Rate

Definition: Probability of a child born in a specific year or period dying before reaching the

age of five, if subject to age-specific mortality rates of that period. Numerator: The number of under five deaths in a given period

Denominator: The total number of live births surveyed.

Calculation: Numerator

__________________ x 1,000

Denominator

Data Sources: Household surveys – e.g. UDHS.

Frequency of

Analysis,

Interpretation

and Response

Every 5 years

Disaggregation By sex, location (major regions), and socio-economic characteristics (e.g. mother’s

education level, wealth quintile).

5. Total Fertility Rate

Definition: The number of live births per 1000 women in a specific age group for a specified geographic area and for a specific point in time

Numerator: Number of live births to women in specified age group

Denominator: Number of women in same age group

Calculation: Numerator

__________________ x 1,000

Denominator

Data Sources: Household surveys – e.g. UDHS.

Frequency of

Analysis,

Interpretation

and Response

Every 5 years

Disaggregation By location (major regions), and socio-economic characteristics (e.g. education

level, wealth quintile).

6. Adolescent Pregnancy (Fertility) Rate (per 1,000 women)

Definition: Number of live births to adolescents (women 15-19 years) per 1,000 in that

age group.

Numerator: Number of live births to adolescents (women 15-19 years) in a specified period

Denominator: Number of adolescents (women 15-19 years) in a specified period

Calculation: Numerator

__________________x 1,000

Denominator

VII | P a g e

Data Sources: Household surveys – e.g. UDHS.

Frequency of

Analysis,

Interpretation

and Response

Every 5 years

Disaggregation By location (major regions), and socio-economic characteristics (e.g. education

level, wealth quintile).

Financial risk protection

1. Out of pocket health expenditure (payment) for health as a % of Total Health Expenditure

Definition: Share of total current expenditure on health paid by households out-of-pocket

(expressed as a % of total current expenditure on health, this is the households’

out-of-pocket expenditure)

Numerator: Total current expenditure on health paid by households out-of-pocket

Denominator: Total current expenditure on health (all sources)

Calculation: Numerator

__________________ x 100

Denominator Data Sources: Household surveys – e.g. UDHS.

Frequency of

Analysis,

Interpretation

and Response

Every 2 years

Disaggregation Financing source institutional unit, Disease, Input, Sub-National Level , Socio-

economic status

OUTCOME

1. ANC 4 Coverage

Definition: Percentage of women aged 15-49 with a live birth in a given time period that

received antenatal care four times.

Numerator: Number of pregnant women that received antenatal care at least four times in a

given period.

Denominator: Number of expected pregnancies in the catchment area in a given period.

Calculation: Numerator

__________________ x 100

Denominator

Data Sources: Health service statistics: - HMIS

Household surveys – that can generate this indicator includes UDHS, Multiple

Indicator Cluster Surveys (MICS), Fertility and Family Surveys, Reproductive Health

Surveys (RHS) and other surveys based on similar methodologies. The UDHS uses

a different indicator definition targeting women aged 15 – 49 years with full term

birth in a given period bay taking birth history. However, in Uganda a number of

young girls are giving birth below the age of 15 years and there are also women

older than 49 years giving birth.

Frequency of

Analysis,

Interpretation

and Response:

The frequency of data collection should be daily, with monthly summaries and

monthly reporting within the schedule of the HMIS. (Health Unit Monthly Report,

Health Unit Quarterly Assessment Report, Health Unit Annual Report).

Surveys - Every 3 – 5 years

Disaggregation: By location (urban/rural, major regions), level of care, parity and socio-economic

characteristics (e.g. mother’s education level, wealth quintile).

2. Health Facility (Institutional) Deliveries

Definition: The proportion of births / deliveries in public and private health facilities.

Numerator: Number of births / deliveries in public and PNFP health facilities.

Denominator: Total number of expected births / deliveries in the same period.

Calculation: Numerator

__________________ x 100

Denominator

Data Sources: Health service statistics: As the point of contact with women, this is the main and

most obvious routine source of information for the numerator however it

excludes deliveries conducted by Private Health Providers not reporting through

the HMIS. The results from the health service statistics should be validated against

the UDHS results.

Population Surveys e.g. UDHS.

Census: Health service information cannot provide accurate size of the

denominator population. Projections or in some cases vital registration data are

used to provide the denominator (number of live births). Frequency of

Analysis,

Interpretation

and Response

The frequency of data collection should be daily, with monthly summaries and

monthly reporting within the schedule of the HMIS.

Analysis, interpretation and response should be done monthly, quarterly and

annually.

Disaggregation By facility level, facility ownership, type of skilled health personnel, location

(urban/rural, districts, regions), parity and socio-economic characteristics (e.g.

mother's education, wealth quintile).

3. DPT3Hib3Heb3 coverage

Definition: The percentage of under one-year olds who have received three valid doses of the

combined DPTHibHeb vaccine in a given year.

Numerator: Number of under one year old who have received the 3rd valid dose of DPTHibHeb

vaccine.

Denominator: Number of children under one year in the target population in one year period.

Calculation: Numerator

__________________x 100

Denominator

Data Sources: Health service statistics - HMIS. The results from the health service statistics

should be validated against the UDHS results.

Household surveys: The principle types of surveys are the Expanded Program on

Immunization (EPI) 30-cluster survey, the UNICEF MICS, and the UDHS.

IX | P a g e

Census: Projections or in some cases vital registration data are used to provide the

denominator (number of children under one year in the target population).

Frequency of

Analysis,

Interpretation

and Response:

The frequency of data collection should be daily, with monthly summaries and

monthly reporting within the schedule of the HMIS.

Analysis, interpretation and response should be done monthly, quarterly and

annually.

Disaggregation: By sex, location (urban/rural, districts, regions) and socio-economic characteristics

(e.g. mother's education, wealth quintile).

4. Measles coverage under 1 year

Definition: The percentage of children under one year of age who have received at least one

valid dose of measles containing vaccine in a given year.

Numerator: Number of children under one year who have received at least one valid dose of

measles vaccine.

Denominator: Number of children under one year in the target population in one year period.

Calculation: Numerator

__________________x 100

Denominator

Data Sources: Health service statistics - HMIS. The results from the health service statistics

should be validated against the UDHS results.

Household surveys: The principle types of surveys are the Expanded Program on

Immunization (EPI) 30-cluster survey, the UNICEF MICS, and the UDHS.

Census: Projections or in some cases vital registration data are used to provide

the denominator (number of children under one year in the target population).

Frequency of

Analysis,

Interpretation

and Response:

The frequency of data collection should be daily, with monthly summaries and

monthly reporting within the schedule of the HMIS.

Analysis, interpretation and response should be done monthly, quarterly and

annually.

Disaggregation: By sex, location (urban/rural, districts, regions) and socio-economic characteristics

(e.g. mother's education, wealth quintile).

5. Malaria cases per 1,000 persons per year

Definition: Number of confirmed reported malaria (microscopy or RDT) cases per 1000

persons per year.

Numerator: Number of confirmed reported malaria (microscopy or RDT) in a year

Denominator: Population in specified geographical location

Calculation: Numerator

__________________x 1000

Denominator

Data Sources: Health service statistics - HMIS.

Frequency of

Analysis,

Interpretation

and Response:

The frequency of data collection should be daily, with monthly summaries and

monthly reporting within the schedule of the HMIS.

Analysis, interpretation and response should be done monthly, quarterly and

annually.

Disaggregation: Age, sex, place of residence, season (year and month)

6. Under five Vitamin A second dose coverage

Definition: Percentage of children 6 – 59 months who received 2 age-appropriate doses of

vitamin A in the past 12 months.

Numerator: Number of children 6 – 59 months who received 2 age-appropriate doses of vitamin A in the past 12 months.

Denominator: Number of children 6 - 59 months in the target population in one year period

Calculation: Numerator

__________________ x 100

Denominator

Data Sources: Health service statistics - HMIS. The results from the health service statistics

should be validated against the UDHS results.

Household surveys: e.g. UDHS.

Frequency of

Analysis,

Interpretation

and Response:

The frequency of data collection should be daily, with monthly summaries and

monthly reporting within the schedule of the HMIS.

Analysis, interpretation and response should be done monthly, quarterly and

annually.

Disaggregation: By sex, location (urban/rural, districts, regions) and socio-economic

characteristics (e.g. mother's education, wealth quintile).

7. Intermittent Presumptive Treatment 3 or more doses coverage for pregnant women

Definition: The proportion of pregnant women attending antenatal care who have received

the 3 or more doses of Sulphadoxine / Pyrimethamine for malaria prophylaxis

during the last pregnancy.

Numerator: Number of pregnant women who received 3 or more doses of Sulphadoxine /

Pyrimethamine for malaria prophylaxis during the last pregnancy.

Denominator: Number of first antenatal clinic visits during a specified time

Calculation: Numerator

__________________x 100

Denominator

Data Sources: Health service statistics - HMIS. The results from the health service statistics

should be validated against the UDHS results.

Household surveys: e.g. UDHS.

Frequency of

Analysis,

Interpretation

and Response:

The frequency of data collection should be daily, with monthly summaries and

monthly reporting within the schedule of the HMIS.

Analysis, interpretation and response should be done monthly, quarterly and

annually.

Disaggregation: By sex, location (urban/rural, districts, regions), parity and socio-economic

characteristics (e.g. mother's education, wealth quintile).

8. TB Case Detection Rate

Definition: Proportion of estimated new and relapse TB cases detected in a given year.

Numerator: Number of new TB cases detected / notified in a given year / period.

Denominator: Total catchment population

Calculation: Numerator

__________________x 100,000

Denominator

Data Sources: Health service statistics: - HMIS

XI | P a g e

Frequency of

Analysis,

Interpretation

and Response:

The frequency of data collection should be daily, with monthly summaries and

monthly reporting within the schedule of the HMIS.

Analysis, interpretation and response should be done monthly, quarterly and

annually.

Disaggregation: By sex, age, location (urban/rural, subcounties, districts, regions) and socio-

economic characteristics (e.g. education, wealth quintile).

9. Anti-Retroviral Therapy Coverage

Definition: The percentage of adults and children with advanced HIV infection currently

receiving antiretroviral combination therapy in accordance with the nationally

approved treatment protocols (or WHO/UNAIDS standards) among the estimated

number of adults and children with advanced HIV infection.

Numerator: Number of adults and children eligible for ART who are currently receiving ARV in

accordance with the nationally approved treatment protocol at the end of the

reporting period.

Denominator: Estimated number of adults and children eligible for ART

(The denominator of the coverage estimate is obtained from models that also

generate the HIV prevalence, incidence and mortality estimates)

Calculation: Numerator

__________________ x 100

Denominator

Data Sources: Health service statistics - HMIS

Estimates for the denominator based on modeling

Frequency of

Analysis,

Interpretation

and Response:

The frequency of data collection should be daily, with monthly summaries and

monthly reporting within the schedule of the HMIS.

Analysis, interpretation and response should be done monthly, quarterly and

annually.

Disaggregation: By sex, Age, Key populations, pregnant women, Persons newly initiated on ART

during the last year, provider type (public/private).

10. HIV + pregnant women not on HAART receiving ARVs for eMTCT during pregnancy, labour,

delivery and postpartum

Definition: Percentage of HIV-infected pregnant women not on HAART provided with ARV

drugs to reduce the risk of mother-to-child transmission during pregnancy, labour,

delivery and postpartum in a specified period.

Numerator: Number of HIV-infected pregnant women not on HAART provided with ARV

drugs to reduce the risk of mother-to-child transmission during pregnancy,

labour, delivery and postpartum in a specified period.

Denominator: Number of HIV-infected pregnant women not on HAART in a specified period.

Calculation: Numerator

__________________x 100

Denominator

Data Sources: Health service statistics – HMIS

Frequency of

Analysis,

The frequency of data collection should be daily, with monthly summaries and

monthly reporting within the schedule of the HMIS.

Interpretation

and Response:

Analysis, interpretation and response should be done monthly, quarterly and

annually.

Disaggregation: By age, location (urban/rural, districts, regions), socio-economic characteristics

(e.g. education, wealth quintile) and provider type (public/private).

11. Contraceptive Prevalence Rate

Definition: The proportion of women aged 15 – 49 years, married or in union, who are using

or whose partner is using at least one method of contraception, regardless of the

method used.

Numerator: Number of women in reproductive age (15 – 49) who are married or in union and

are currently using any method of contraception in a given period.

Denominator: Total number of women in reproductive age (15 – 49) who are married or in union

in a given period.

Calculation: Numerator

__________________x 100

Denominator

Data Sources: Household surveys (such as UDHS, MICS).

Frequency of

Analysis,

Interpretation

and Response:

Every three to five years.

Disaggregation: By age, method of contraception, location (urban/rural, major regions), and socio-

economic characteristics (e.g. education level, quintile).

12. Couple Years of Protection

Definition: The estimated protection provided by family planning (FP) services during a one-

year period, based upon the volume of all contraceptives sold or distributed free

of charge to clients during that period.

Calculation: The CYP is calculated by multiplying the quantity of each method distributed to

clients by a conversion factor (Ref. Annex 8.4), to yield an estimate of the

duration of contraceptive protection provided per unit of that method. The CYPs

for each method are then summed over all methods to obtain a total CYP figure.

Data Sources: Health service statistics: HMIS or Supply Chain Management Information System

Frequency of

Analysis,

Interpretation

and Response:

The frequency of data collection should be daily, with monthly summaries and

monthly reporting within the schedule of the HMIS.

Analysis, interpretation and response should be done monthly, quarterly and

annually.

Disaggregation: By method, location (urban/rural, districts, regions) level and type of facility.

RISK FACTORS

1. Latrine Coverage

Definition: The percentage of the households using proper excreta disposal facilities.

Numerator: Number of households using proper1 excreta disposal facilities.

1 Proper excreta disposal facilities include: Flush or pour-flush to piped sewer system, septic tank or Pit latrine

(Ventilated improved, with san-plat, with slab, ecological sanitation toilets, composting toilet).

XIII | P a g e

Denominator: Number of households in catchment.

Calculation: Numerator

__________________x 100

Denominator

Data Sources: Administrative reporting - using the Environmental Health Assessment Book.

Verified with the survey results.

Household surveys – These include National Household Survey.

Population census – Provides the estimates on number of households which is the

denominator.

Frequency of

Analysis,

Interpretation

and Response:

Annually

Disaggregation: By location (urban and rural, district) and socio-economic characteristics of head

of household (e.g. education level, wealth quintile).

2. Villages/ wards with a functional VHT

Definition: The percentage of villages / wards with functional2 VHTs.

Numerator: Number of villages / wards with functional VHTs in the catchment under

consideration.

Denominator: Number of villages / wards in the catchment under consideration.

Calculation: Numerator

__________________x 100

Denominator

Data Sources: Health Service Statistics - Community HMIS

Frequency of

Analysis,

Interpretation

and Response:

Analysis, interpretation and response should be done quarterly and annually.

Disaggregation: By location (subcounty, district, urban/rural)

3. Children below 5 years who are stunted

Definition: Percentage of stunting (height-for-age less than -2 standard deviations of the

WHO Child Growth Standards median) among children below five years (aged 0-4

years)

Numerator: Number of children below 5 years (aged 0-4 years) that fall below minus two

standard deviations from the median height-for-age during a specified period.

Denominator: Total number of children below 5 years (aged 0-4 years) that were measured

during a specified period.

Calculation: Numerator

__________________x 100

Denominator

Data Sources: Household surveys – UDHS

Specific population surveys – national nutrition surveys

2 VHT Functionality: VHTs submitting reports through the Community HMIS.

Frequency of

Analysis,

Interpretation

and Response:

Every 5 years

Disaggregation: By sex, age, location (urban/rural, major regions, districts) and socio-economic

characteristics (e.g. education level, wealth quintile).

4. Children below 5 years who are under weight

Definition: Percentage of underweight (weight-for-age less than 2 standard deviations of the

WHO Child Growth Standards median among children below 5 years (0 – 4 years). Numerator: Number of children below 5 years (0 – 4 years) that fall below minus two standard

deviations from the median weight-for-age during a specified period.

Denominator: Total number of children below 5 years (0 – 4 years) that were weighed during a

specified period.

Calculation: Numerator

__________________x 100

Denominator

Data Sources: Household surveys – UDHS

Specific population surveys – national nutrition surveys

Frequency of

Analysis,

Interpretation

and Response:

Every 5 years

Disaggregation: By sex, age, location (urban/rural, major regions, districts), and socio-economic

characteristics (e.g. education level, wealth quintile).

Service access and availability

1. New OPD utilization rate

Definition: The number of new outpatient visits3 during a specified period relative to the total

population of the same geographical area.

Numerator: Number of new outpatient visits to health facilities for ambulant care in a specified

period.

Denominator: Total health facility catchment population in a specified period.

Calculation: Numerator

__________________

Denominator

Data Sources: Health service statistics: HMIS Frequency of

Analysis,

Interpretation

and Response:

The frequency of data collection should be daily, with monthly summaries and

monthly reporting within the schedule of the HMIS.

Analysis, interpretation and response should be done monthly, quarterly and

annually.

Disaggregation: By sex, age group, location (urban/rural, districts, regions) and socio-economic

characteristics (e.g. education, wealth quintile), level and type of facility.

3 New OPD attendance is the occurrence of a patient attending a clinic for a new episode of illness.

XV | P a g e

2. Hospital (inpatient) admissions per 100 population

Definition: The number of hospital (inpatient) admissions per 100 population.

Numerator: Number of hospital (inpatient) admissions in a specified period.

Denominator: Total catchment population in a specified period.

Calculation: Numerator

__________________x 100

Denominator

Data Sources: Health service statistics: HMIS Frequency of

Analysis,

Interpretation

and Response:

The frequency of data collection should be daily, with monthly summaries and

monthly reporting within the schedule of the HMIS.

Analysis, interpretation and response should be done monthly, quarterly and

annually.

Disaggregation: By sex, age group, location (urban/rural, districts, regions) and socio-economic

characteristics (e.g. education, wealth quintile), level and type of facility.

3. Bed occupancy rate (Hospitals & HC IVs)

Definition: The number of hospital beds occupied by patients expressed as a percentage of the total beds available in the ward, specialty, hospital, area, or region in a specified period.

Numerator: Number of inpatient days in a specified period

Denominator: Number of available hospital beds in a specified period

Calculation: Numerator

__________________x 100

Denominator

Data Sources: Health service statistics: HMIS Frequency of

Analysis,

Interpretation

and Response:

The frequency of data collection should be daily, with monthly summaries and

monthly reporting within the schedule of the HMIS.

Analysis, interpretation and response should be done monthly, quarterly and

annually.

Disaggregation: By location (urban/rural, districts, regions), level and type of facility.

4. Health Centre IVs offering CEmOC Services (Functionality)

Definition: The percentage of HC IVs offering CEmOC services.

Numerator: Number of HC IVs offering CEmOC services

Denominator: Total number of HC IVs

Calculation: Numerator

__________________x 100

Denominator

Data Sources: Health service statistics: HMIS Frequency of

Analysis,

Interpretation

and Response:

The frequency of data collection should be daily, with monthly summaries and

monthly reporting within the schedule of the HMIS.

Analysis, interpretation and response should be done monthly, quarterly and

annually.

Disaggregation: By location (urban/rural, districts, regions) and ownership.

5. Availability of a basket of commodities in the previous quarter (% of facilities that had over

95%)

Definition: The percentage of health facilities with over 95% of the basket of commodities in

the previous quarter.

Numerator: Number of health facilities with over 95% of the basket of commodities in the

previous quarter.

Denominator: Number of health facilities reporting in a given period

Calculation: Numerator

__________________ x 100

Denominator

Data Sources: Health service statistics: HMIS

Special facility surveys – Medicines availability study carried out where data on

availability of a specific list of medicines are collected in at least four geographic

or administrative areas in a sample of medicines dispensing points. Frequency of

Analysis,

Interpretation

and Response:

The frequency of data collection should be daily, with monthly summaries and

monthly reporting within the schedule of the HMIS.

Analysis, interpretation and response should be done monthly, quarterly and

annually.

Disaggregation: By provider type (public/private), health facility level, district, product type. Service quality and safety

6. Client satisfaction index

Definition: The proportion of clients expressing satisfaction with health services.

Numerator: Number of people surveyed expressing satisfaction with health services.

Denominator: Number of clients surveyed

Calculation: Numerator

__________________x 100

Denominator

Data Sources: Health Facility Surveys – for facility specific client satisfaction

Population based surveys - for National level determination of client satisfaction

Frequency of

Analysis,

Interpretation

and Response:

Health facility specific data should be collected and analyzed annually.

Data from population based surveys is collected and analyzed every 2 - 3 years.

Disaggregation: By sex, location (rural / urban), type and ownership of facility

7. Average length of stay (Hospitals & HC IVs)

Definition: Average number of days patients spend in hospitals / HC IVs.

Numerator: Total inpatient days of care in a specified period

Denominator: Total admissions in a specified period

Calculation: Numerator

__________________x 100

Denominator

Data Sources: Health service statistics: HMIS Frequency of

Analysis,

Interpretation

and Response:

The frequency of data collection should be daily, with monthly summaries and

monthly reporting within the schedule of the HMIS.

Analysis, interpretation and response should be done monthly, quarterly and

annually.

XVII | P a g e

Disaggregation: By location (urban/rural, districts, regions), facility level, service provided

(Department) and ownership.

8. Facility based fresh still births (per 1,000 deliveries)

Definition: Number of fresh still births among 1,000 deliveries in health facilities.

Numerator: Number of fresh still births.

Denominator: Number of deliveries in health facilities.

Calculation: Numerator

__________________x 1000

Denominator

Data Sources: Health service statistics: HMIS Frequency of

Analysis,

Interpretation

and Response:

The frequency of data collection should be daily, with monthly summaries and

monthly reporting within the schedule of the HMIS.

Analysis, interpretation and response should be done monthly, quarterly and

annually.

Disaggregation: By location (urban/rural, districts, regions), facility level and ownership, parity,

and socio-economic characteristics (e.g. education, wealth quintile).

9. Institutional Maternal Mortality

Definition: Number of maternal deaths among 100,000 health facility deliveries.

Numerator: Number of maternal deaths in a specified period.

Denominator: Number of health facility deliveries in a specified period.

Calculation: Numerator

__________________x 100,000

Denominator

Data Sources: Health service statistics: HMIS Frequency of

Analysis,

Interpretation

and Response:

The frequency of data collection should be daily, with monthly summaries and

monthly reporting within the schedule of the HMIS.

Analysis, interpretation and response should be done monthly, quarterly and

annually.

Disaggregation: By location (urban/rural, districts, regions), facility level and ownership, parity

and socio-economic characteristics (e.g. education, wealth quintile).

10. Maternal death reviews conducted

Definition: Percentage of maternal deaths occurring in the facility that were audited.

Numerator: Number of maternal deaths reviewed in a specified period.

Denominator: Number of maternal deaths in a specified period.

Calculation: Numerator

__________________x 100,000

Denominator

Data Sources: Health service statistics: HMIS Frequency of

Analysis,

Interpretation

and Response:

The frequency of data collection should be daily, with monthly summaries and

monthly reporting within the schedule of the HMIS.

Analysis, interpretation and response should be done monthly, quarterly and

annually.

Disaggregation: By location (urban/rural, districts, regions), facility level and ownership.

11. Institutional under five mortality

Definition: Under five deaths among 1,000 under 5 admissions.

Numerator: Number of under five deaths in health facilities in a specified period.

Denominator: Number of under five admissions in a specified period.

Calculation: Numerator

__________________x 100,000

Denominator

Data Sources: Health service statistics: HMIS Frequency of

Analysis,

Interpretation

and Response:

The frequency of data collection should be daily, with monthly summaries and

monthly reporting within the schedule of the HMIS.

Analysis, interpretation and response should be done monthly, quarterly and

annually.

Disaggregation: By age, sex, location (urban/rural, districts, regions), facility level and ownership.

12. ART Retention rate

Definition: Percentage of adults and children with HIV alive and on antiretroviral therapy

(ART) 12, 24, 36 months (etc.) after initiating treatment among patients initiating

antiretroviral therapy during a specified time period.

Numerator: Number of patients (adults and children) with HIV alive and on antiretroviral

therapy (ART) 12, 24, 36 months (etc.) after initiating treatment.

Denominator: Number of patients (adult and children) alive initiated on antiretroviral therapy during a specified time period.

Calculation: Numerator

__________________x 100,000

Denominator

Data Sources: Health service statistics: HMIS Frequency of

Analysis,

Interpretation

and Response:

The frequency of data collection should be daily, with monthly summaries and

monthly reporting within the schedule of the HMIS.

Analysis, interpretation and response should be done monthly, quarterly and

annually.

Disaggregation: By age, sex, location (urban/rural, districts, regions), facility level and ownership,

and socio-economic characteristics (e.g. education, wealth quintile).

13. TB Treatment Success Rate

Definition: Percentage of TB cases successfully treated (cured plus treatment completed)

among all new TB cases notified to the national health authorities during a

specified period.

Numerator: Number of TB cases successfully treated (cured plus treatment completed).

Denominator: Number of new TB cases notified to the national health authorities during a specified period.

Calculation: Numerator

__________________x 100

Denominator

Data Sources: Health service statistics: HMIS Frequency of

Analysis,

The frequency of data collection should be daily, with monthly summaries and

monthly reporting within the schedule of the HMIS.

XIX | P a g e

Interpretation

and Response:

Analysis, interpretation and response should be done monthly, quarterly and

annually.

Disaggregation: Age, sex, treatment history: new and relapse/previously treated (excl. relapse),

HIV, Bacteriology, facility level and ownership.

Health Financing

1. Current expenditure on health by general government General Government allocation for

health as % of total government budget

Definition: General Government allocation for health as a % of the total government

budget.

Numerator: Government of Uganda allocation for health.

Denominator: Total government budget.

Calculation: Numerator

__________________x 100

Denominator

Data Sources: Administrative reporting system - General Government Accounts, Public

Expenditure Reviews, Institutional reports of public entities involved in health care

provision or financing, notably MoFPED reports (social security and other health

insurance compulsory agencies when established). GGA reported by the Central

Bank and the MoFPED.

National Health Accounts – resources are tracked for all public entities acting

sources, financing agents (managing health funds and purchasing or paying for

health goods and services) and providers.

Frequency of

Analysis,

Interpretation

and Response:

Analysis, interpretation and response should be done quarterly and annually.

Disaggregation: By organizational, district and health facility level.

Workforce

1. Absenteeism rate

Definition: Percentage of health workers absent from duty at a given time.

Numerator: Number of health workers absent from duty at a given time.

Denominator: The total number of health workers employed at the given time.

Calculation: Numerator

__________________ x 100

Denominator

Data Sources: Administrative records – attendance registers, duty rosters

Health service surveys - population based and facility sample surveys e.g.

National Panel Survey, Service Provision Assessment (SPA) or Service Availability

Readiness Assessment (SARA)

Frequency of

Analysis,

Interpretation

and Response:

At facility level administrative records should be analyzed and reported quarterly

and annually.

Periodically validated and adjusted against data from health service surveys.

Disaggregation: By reason (e.g. off duty, study leave, workshops, family circumstances, dual

employment, etc), cadre, sex, type of facility, location (urban/rural) district,

regions).

2. Approved posts in public facilities filled with qualified personnel

Definition: Percentage of approved posts4 filled by trained health workers.

Numerator: Number of approved posts filled by trained health workers.

Denominator: Total number of approved posts for trained health workers in the Public and

Private Not-for-Profit Health Facilities.

Calculation: Numerator

__________________x 100

Denominator

Data Sources: Human Resource Information System

Frequency of

Analysis,

Interpretation

and Response:

At district level analysis should be carried out at least twice a year.

At national level reported at least annually, and periodically validated and

adjusted against data from a population census or other nationally representative

source.

Disaggregation: By cadre, sex, age, facility type and ownership, location (urban / rural), district,

region.

3. Health worker density

Definition: Number of health workers per 1,000 population.

Numerator: Number of health workers (per cadre) in a specified period.

Denominator: Catchment population in a specified period.

Calculation: Numerator

__________________x 1,000

Denominator

Data Sources: Human Resource Information System

Frequency of

Analysis,

Interpretation

and Response:

At district level analysis should be carried out at least twice a year.

At national level reported at least annually, and periodically validated and

adjusted against data from a population census or other nationally representative

source.

Disaggregation: By Cadre: core professionals (physicians, midwives, nurses); specific cadres such

as specialists (surgeons, psychiatrists, etc); other cadres (dentists, pharmacists)

Distribution: Place of employment (urban/rural); Sub-national (district, region).

Infrastructure

1. Population living within 5km of a health facility

Definition: Percentage of population living within 5km of a health facility (total number of

health facilities per 10,000 population).

Numerator: Population living within 5km of a health facility.

Denominator: Total population in a specified period.

Calculation: Numerator

4 Approved posts are the numbers (staffing levels) and types (cadres) of employees established for

the organization.

XXI | P a g e

__________________x 1,000

Denominator

Data Sources: Human Facility Inventory

Frequency of

Analysis,

Interpretation

and Response:

Annually

Disaggregation: By location - (urban/rural); district, region.

Health Information

1. Timeliness of reporting (HMIS 105)

Definition: Percentage of health facilities that submit reports within the required deadline.

Numerator: Number of health facilities that submit reports within the required deadline.

Denominator: Total number of health facilities registered.

Calculation: Numerator

__________________x 100

Denominator

Data Sources: Health Service Statistics – HMIS

Frequency of

Analysis,

Interpretation

and Response:

The frequency of data collection should be monthly within the schedule of the

HMIS.

Analysis, interpretation and response should be done monthly, quarterly and

annually.

Disaggregation: By facility type and ownership, location (urban / rural), district, region.

8.5 Indicator Definitions for the Additional Indicators in the DLT

1. Patients diagnosed with malaria that are laboratory confirmed

Definition: Percentage of patients diagnosed with malaria that are laboratory confirmed

(rapid diagnostic test or microscopy) in a specified period.

Numerator: Number of patients diagnosed with malaria that are laboratory confirmed in a

specified period.

Denominator: Number of patients diagnosed with malaria in a specified period.

Calculation: Numerator

____________x 100

Denominator

Data Sources: Health service statistics: HMIS Frequency of

Analysis,

Interpretation

and Response

The frequency of data collection should be daily, with monthly summaries and

monthly reporting within the schedule of the HMIS.

Analysis, interpretation and response should be done monthly, quarterly and

annually.

Disaggregation Age, sex, facility level and ownership.

2. PCV3 Coverage

Definition: The percentage of under one-year olds who have received three valid

doses of PCV vaccine in a specified period.

Numerator: Number of under one year old who have received the 3rdvalid dose of PCV vaccine

in a specified period.

Denominator: Number of children under one year in the target population in a specified period.

Calculation: Numerator

__________________x 100

Denominator

Data Sources: Health service statistics - HMIS. The results from the health service statistics

should be validated against the UDHS results.

Frequency of

Analysis,

Interpretation

and Response

The frequency of data collection should be daily, with monthly summaries and

monthly reporting within the schedule of the HMIS.

Analysis, interpretation and response should be done monthly, quarterly and

annually.

Disaggregation By sex, location (urban/rural, districts, regions) and socio-economic characteristics

(e.g. mother's education, wealth quintile).

3. Timeliness of quarterly district financial reporting (OBT)

XXIII | P a g e

Definition: Percentage of districts that submit quarterly district financial reports within the

required deadline.

Numerator: Number of districts that submit quarterly district financial reports within the

required deadline.

Denominator: Total number of districts

Calculation: Numerator

__________________x 100

Denominator

Data Sources: Financial Reports

Frequency of

Analysis,

Interpretation

and Response

The frequency of data collection should be monthly within the schedule of the

HMIS.

Analysis, interpretation and response should be done monthly, quarterly and

annually.

Disaggregation By district and region.

4. Average district quality of care score

Definition: Percentage average district quality of care score

Numerator: Total district quality of care scores

Denominator: Number of health facilities assessed in the district

Calculation: Numerator

____________x 100

Denominator

Data Sources: Health Service Statistics – HFQAP database

Frequency of

Analysis,

Interpretation

and Response

The frequency of data collection should be annually within the schedule of the

program.

Analysis, interpretation and response should be done annually.

Disaggregation By status – new or old district, number of health facilities per district, functionality

of the DHT.

8.6 CYP Conversion Factors

Method CYP Per Unit

Oral Contraceptives 15 cycles per CYP

Condoms (male and female) 120 units per CYP

Monthly Vaginal Ring/Patch 15 units per CYP

Vaginal Foaming Tablets 120 units per CYP

DepoProvera Injectable 4 doses (ml) per CYP

Noristerat Injectable 6 doses per CYP

Cyclofem Monthly Injectable 13 doses per CYP

Copper-T 380-A IUD 4.6 CYP per IUD inserted (3.3 for 5 year IUD,

e.g. LNG-IUS)

3 Year Implant (e.g. Implanon) 2.5 CYP per implant

4 Year Implant (e.g. Sino-Implant) 3.2 CYP per implant

5 Year Implant (e.g. Jadelle) 3.8 CYP per implant

Emergency Contraceptive Pills 20 doses per CYP

Standard Days Method 1.5 CYP per trained adopter

Lactational Amenorrhea Method (LAM) 4 active users per CYP (or .25 CYP per user)

Sterilization (male and female) Africa 9.3 CYP