Skip to main content
Advertisement
Browse Subject Areas
?

Click through the PLOS taxonomy to find articles in your field.

For more information about PLOS Subject Areas, click here.

  • Loading metrics

“It might be a statistic to me, but every death matters.: An assessment of facility-level maternal and perinatal death surveillance and response systems in four sub-Saharan African countries

  • Mary V. Kinney ,

    Roles Data curation, Formal analysis, Investigation, Supervision, Validation, Visualization, Writing – original draft, Writing – review & editing

    mkinney@uwc.ac.za

    Affiliations Save the Children US, Washington, DC, United States of America, University of the Western Cape, Cape Town, South Africa

  • Gbaike Ajayi,

    Roles Data curation, Formal analysis, Investigation, Methodology, Project administration, Visualization, Writing – original draft, Writing – review & editing

    Affiliations US Agency for International Development (USAID)’s Maternal and Child Survival Program (MCSP), Washington, DC, United States of America, Jhpiego, Baltimore, Maryland, United States of America

  • Joseph de Graft-Johnson,

    Roles Formal analysis, Investigation, Methodology, Supervision, Writing – review & editing

    Affiliations Save the Children US, Washington, DC, United States of America, US Agency for International Development (USAID)’s Maternal and Child Survival Program (MCSP), Washington, DC, United States of America

  • Kathleen Hill,

    Roles Conceptualization, Funding acquisition, Investigation, Methodology, Supervision, Validation, Writing – review & editing

    Affiliations US Agency for International Development (USAID)’s Maternal and Child Survival Program (MCSP), Washington, DC, United States of America, Jhpiego, Baltimore, Maryland, United States of America

  • Neena Khadka,

    Roles Funding acquisition, Investigation, Methodology, Supervision, Validation, Writing – review & editing

    Affiliations Save the Children US, Washington, DC, United States of America, US Agency for International Development (USAID)’s Maternal and Child Survival Program (MCSP), Washington, DC, United States of America

  • Alyssa Om’Iniabohs,

    Roles Data curation, Investigation, Project administration, Visualization, Writing – review & editing

    Affiliations Save the Children US, Washington, DC, United States of America, US Agency for International Development (USAID)’s Maternal and Child Survival Program (MCSP), Washington, DC, United States of America

  • Fadzai Mukora-Mutseyekwa,

    Roles Investigation, Supervision, Writing – review & editing

    Affiliation USAID’s Maternal and Child Health Integrated Program/John Snow Inc., Harare, Zimbabwe

  • Edwin Tayebwa,

    Roles Investigation, Supervision, Writing – review & editing

    Affiliation USAID’s MCSP/Jhpiego, Kigali, Rwanda

  • Oladapo Shittu,

    Roles Investigation, Supervision, Writing – review & editing

    Affiliation Ahmadu Bello University, Zaria, Kaduna State, Nigeria

  • Chrisostom Lipingu,

    Roles Investigation, Supervision, Writing – review & editing

    Affiliation Bukoba Regional Referral Hospital, Kagera, Tanzania

  • Kate Kerber,

    Roles Conceptualization, Data curation, Investigation, Methodology, Supervision, Validation, Visualization, Writing – review & editing

    Affiliation Save the Children US, Washington, DC, United States of America

  • Juma Daimon Nyakina,

    Roles Investigation, Writing – review & editing

    Affiliation Bukoba Regional Referral Hospital, Kagera, Tanzania

  • Perpetus Chudi Ibekwe,

    Roles Investigation, Writing – review & editing

    Affiliation Maternal and perinatal death surveillance and response, Abakaliki, Ebonyi State, Nigeria

  • Felix Sayinzoga,

    Roles Investigation, Supervision, Writing – review & editing

    Affiliation Maternal, Child, and Community Health Division, Rwanda Biomedical Center, Kigali, Rwanda

  • Bernard Madzima,

    Roles Investigation, Writing – review & editing

    Affiliation Family Health Directorate, Ministry of Health and Child Care, Harare, Zimbabwe

  • Asha S. George,

    Roles Supervision, Validation, Visualization, Writing – review & editing

    Affiliation University of the Western Cape, Cape Town, South Africa

  •  [ ... ],
  • Kusum Thapa

    Roles Data curation, Formal analysis, Investigation, Methodology, Supervision, Validation, Visualization, Writing – original draft, Writing – review & editing

    Affiliations US Agency for International Development (USAID)’s Maternal and Child Survival Program (MCSP), Washington, DC, United States of America, Jhpiego, Baltimore, Maryland, United States of America

  • [ view all ]
  • [ view less ]

Abstract

Background

Maternal and perinatal death surveillance and response (MPDSR) systems aim to understand and address key contributors to maternal and perinatal deaths to prevent future deaths. From 2016–2017, the US Agency for International Development’s Maternal and Child Survival Program conducted an assessment of MPDSR implementation in Nigeria, Rwanda, Tanzania, and Zimbabwe.

Methods

A cross-sectional, mixed-methods research design was used to assess MPDSR implementation. The study included a desk review, policy mapping, semistructured interviews with 41 subnational stakeholders, observations, and interviews with key informants at 55 purposefully selected facilities. Using a standardised tool with progress markers defined for six stages of implementation, each facility was assigned a score from 0–30. Quantitative and qualitative data were analysed from the 47 facilities with a score above 10 (‘evidence of MPDSR practice’).

Results

The mean calculated MPDSR implementation progress score across 47 facilities was 18.98 out of 30 (range: 11.75–27.38). The team observed variation across the national MPDSR guidelines and tools, and inconsistent implementation of MPDSR at subnational and facility levels. Nearly all facilities had a designated MPDSR coordinator, but varied in their availability and use of standardised forms and the frequency of mortality audit meetings. Few facilities (9%) had mechanisms in place to promote a no-blame environment. Some facilities (44%) could demonstrate evidence that a change occurred due to MPDSR. Factors enabling implementation included clear support from leadership, commitment from staff, and regular occurrence of meetings. Barriers included lack of health worker capacity, limited staff time, and limited staff motivation.

Conclusion

This study was the first to apply a standardised scoring methodology to assess subnational- and facility-level MPDSR implementation progress. Structures and processes for implementing MPDSR existed in all four countries. Many implementation gaps were identified that can inform priorities and future research for strengthening MPDSR in low-capacity settings.

Introduction

Despite gradual progress, women and their babies continue to die of complications of gravidity and childbirth or complications in the first month after birth; an estimated 303,000 global maternal deaths, 2.6 million stillbirths, and 2.5 million newborn deaths occur per year [1,2]. Over 40% of these deaths occur in sub-Saharan Africa, and one-half occur in the perinatal period [35]. Many of these deaths are preventable through timely access to high-quality, safe care that delivers evidence-based interventions and avoids harmful practices for women and newborns during gravidity, childbirth, and the postnatal period [5]. To achieve the Sustainable Development Goal targets to end preventable maternal and newborn deaths by 2030, there has been a renewed focus on improving quality of care [6,7], as reflected in multiple global and country efforts [814]. Concurrently, there has been momentum to strengthen maternal and perinatal death surveillance and response (MPDSR) as one mechanism to help address quality of care deficits and other important contributors to preventable maternal and newborn deaths [1518].

MPDSR is a systematic process used to understand the medical causes and the modifiable factors that contribute to maternal and perinatal deaths to identify actions to prevent future deaths [18]. MPDSR operates at all levels of the health system. Its aims are to ensure accurate documentation and reporting of deaths, identify modifiable systemic and social factors at various levels (e.g., delays in care seeking, lack of access to care, quality of care gaps), and link recommendations and accountability for follow-up actions [1921].

The World Health Organization (WHO) has distinct guidelines for maternal death surveillance and response and for perinatal death audit [19,20]. WHO promotes an integrated approach when appropriate, and many countries have adopted integrated national MPDSR guidelines and policies in recent years [17,21]. A number of studies and reviews have explored facilitators and inhibitors of implementation or sustainability of maternal and perinatal mortality audit systems [17,2124]. Challenges to effective implementation of MPDSR have been identified, including not having a national MPDSR policy, weak information and surveillance systems (e.g., lack of vital registration systems and lack of primary data on cause of death), lack of diagnostic capacity for accurate classification of cause of death, and gaps in identifying and documenting maternal and perinatal deaths. Even when data do exist and deaths are reviewed, identified modifiable factors may not be addressed, undermining the “response” component of MPDSR [21,25].

Despite some knowledge of the high-level factors enabling or preventing implementation, there is limited understanding of subnational and facility-based MPDSR activities in sub-Saharan African countries. Better understanding of MPDSR implementation status at subnational and facility levels, including enablers and barriers, can help countries to strengthen MPDSR systems as an important element of their efforts to reduce preventable deaths.

Methodology

Aim and design

The aim of this study was to systematically assess the level of implementation of MPDSR in four sub-Saharan African countries, applying a standardised scoring methodology, and to describe common facilitators and barriers to sustainable MPDSR practice. A cross-sectional, mixed-methods research design was used to assess MPDSR implementation at subnational and facility levels. Qualitative and quantitative data collection methods were employed, including observations (e.g., onsite review of facility documents) and semistructured key informant interviews with subnational and facility managers and staff. The US Agency for International Development (USAID)’s Maternal and Child Survival Program (MCSP) led the study with support from ministries of health. Country visits took place between October 2016 and May 2017. Country study protocols and tools were approved by in-country ethics committees, including the Rwanda National Ethics Committee, Tanzania’s National Institute for Medical Research, the Medical Research Council of Zimbabwe, and Nigeria’s National Health Research Ethics Committee. The study received a nonhuman subjects research determination by the Johns Hopkins Bloomberg School of Public Health Institutional Review Board. The data collected in this assessment did not include any personal identifiers from respondents. Before review of facility documents and before every key informant interview, the interviewer read an oral consent script and asked the participant to respond “yes” or “no”. Oral consent was obtained in Nigeria, Rwanda, and Zimbabwe and written consent obtained in Tanzania, in accordance with ethics committee approvals in each local setting.

Sampling

Four countries—Nigeria, Rwanda, Tanzania, and Zimbabwe—were purposively selected as countries from which a more detailed picture of district- and facility-based MPDSR activities could be gathered. Factors that influenced the selection of the four countries included: (1) having existing national guidelines for MPDSR (or any form of maternal and/or perinatal death audit policy), (2) country government interest and approval, (3) in-country presence of MCSP (or affiliated organization) to support the assessment, and (4) presence of other in-country partners supporting maternal and/or perinatal death review and response. Table 1 presents selected statistics for the four countries, demonstrating the range of maternal and perinatal death rates and ratios, and institutional birth coverage across the four countries.

thumbnail
Table 1. Selection of maternal and newborn health information for the four countries.

https://doi.org/10.1371/journal.pone.0243722.t001

National and subnational stakeholders were identified for interview by MCSP in-country staff and/or the ministry of health. A total of 41 stakeholders were interviewed, including four national stakeholders in Zimbabwe and Tanzania, and 37 regional and district government health officials supporting MPDSR in Zimbabwe, Tanzania, and Nigeria. No stakeholder interviews were conducted in Rwanda due to the unavailability of identified interviewees, who were all engaged in a national meeting at the time of the assessment. Selection of facilities was purposeful and done in collaboration with the ministries of health and included the following criteria: provision of childbirth services and current or previous experience conducting maternal and/or perinatal death audits. Facilities were based on a convenience sample rather than a true probability sample and differed between countries with respect to geographic spread and levels of care. For example, two regions (states) were targeted in Nigeria and Tanzania due to MCSP presence in these areas at the time of the assessment, whereas facilities in all major geographic areas were targeted in Rwanda and Zimbabwe. In total, 55 health facilities (41 hospitals and 14 health centres) received onsite visits. Table 2 summarises the geographic distribution and types of facilities and subnational stakeholders selected in each country.

Data collection

Data collectors included MCSP technical staff and in-country staff from MCSP partner organisations (Save the Children and Jhpiego), national and subnational ministry of health representatives, professional association members (in Nigeria only), and local consultants as needed. The size of the assessment teams for each facility varied from two to five people. Each country’s data collection team received standardised training on completion of the data collection tools and assessment methodology. Data collection tools included a semistructured questionnaire for subnational managers and stakeholders (S1 Table) to explore district and regional MPDSR activities, and subnational support of facility-level MPDSR implementation. The facility assessments included two types of data collection: 1) administration of a standardised, semistructured questionnaire to facility health workers supporting MPDSR-related activities who were present on the day of the visit, and 2) observations by assessors of MPDSR-related documents and activities in the facility (e.g., review of MPDSR meeting notes). Generally, facility-level interviews were conducted with health workers as a team, with individual staff selected by the facility manager.

An implementation tool was developed specifically for this study, adapted from the work by Bergh and colleagues for understanding facility-based kangaroo mother care implementation status [30,31]. The tool designed for this study was developed by grounding the constructs in the literature on the topic, engaging experts in the development of the criteria and consulting global guidelines (Table 3). It was also informed by a set of potential questions and progress markers proposed for measuring the status of perinatal death audit implementation [24].

Data analysis

To understand the context and history of implementation, a desk review of related national MPDSR guidelines and literature on implementation of MPDSR in these countries was conducted. A linked policy mapping set out to determine the content of each national guideline in relation to instructions that have been provided to subnational and facility levels regarding implementation.

To derive a cumulative implementation progress score for each facility, the quantitative data were analysed using the adapted implementation progress monitoring model. An implementation progress score was calculated for each facility across six stages of implementation, with each stage having a weighted score based on specific points (Fig 1). For each stage, the assessors considered all relevant collected data to assign stage-specific points, contributing to a possible total score of 30 (see Table 3). Any discrepancies between the data collectors’ score assignment and progress marker results were resolved through discussion and consensus, with the final score determined by the lead investigators (KK for Zimbabwe, KK and OS for Nigeria, KT and GA for Rwanda, and KT and MK for Tanzania). The lead investigators also met with in-country ministry of health and partner stakeholders before and after assessments to present the study design and discuss interpretation of the findings before scores were finalised. Facilities that scored greater than or equal to 10 met at least the fourth stage of ‘evidence of practice’. Eight facilities were excluded from the qualitative and quantitative analyses because they did not meet the facility inclusion criteria of ‘evidence of practice’ (seven in Nigeria and one in Tanzania).

Data from the facility and subnational key informant questionnaires were extracted into a database to tabulate descriptive means and frequencies of explanatory variables and progress markers (S1 Data). Qualitative data were analysed using thematic content analysis. Team members (KT, MK, and JJ) independently coded qualitative responses, consulted, and reached consensus on data interpretation. The team mapped national guidelines and tools using a content analysis and verified data with national stakeholders (S2 Table). To determine the leading facility-reported barriers and enablers to MPDSR implementation, the team analysed the frequency of qualitative responses from facility interviews based on the thematic content analysis and considered the frequency of relevant progress markers (S3 Table).

Results

National and subnational enabling structures

The history of introducing and implementing maternal and perinatal death audits or reviews varied among the four countries (S1 File). National MPDSR guidelines, tools, and forms varied in content across the four countries, including guidance on methods to classify deaths and timeline for death notification (S2 Table). Paper-based systems were used in all four countries. In addition, Rwanda used electronic tools for documenting and reporting maternal deaths, and one province in Zimbabwe was piloting an electronic data system for both maternal and neonatal data. Subnational managers interviewed in Tanzania, Nigeria, and Zimbabwe expressed concerns about the quality of data in facility MPDSR reports in their district or region. All countries had active national MPDSR committees, but subnational support structures varied among countries.

Facility-based implementation of MPDSR

Across those facilities with evidence of practice, the stage of facility-based MPDSR implementation varied within and across countries (Fig 2). The mean implementation progress score across the 47 facilities was 18.98 (evidence of practice, Stage 4), with a range from 11.75–27.38. One-third of facilities (34%) had reached the evidence of practice stage (Stage 4); over half of facilities (55%) were assessed to be at the stage of routine and integrated practice (Stage 5); and 11% demonstrated implementation at the level of sustainable practice (Stage 6). Overall, hospitals scored higher on average (19.68) than health centres (16.01).

thumbnail
Fig 2. Implementation progress score and distribution of facilities by country.

https://doi.org/10.1371/journal.pone.0243722.g002

Results by stage of facility-based MPDSR implementation

Results are reported for both specific progress markers and questionnaire items across stages that represent a linked implementation progression. Table 4 presents the results for all progress markers by individual country and cumulatively across the four countries. S4 Table provides the ranking of the progress markers by frequency overall. Progress markers for earlier stages (Stages 1–3) were mostly achieved by all facilities, which was consistent with facility selection criteria. Fewer facilities met the progress markers for higher stages of implementation (Stages 5 and 6), and wide variation was observed for some progress markers in the higher stages across countries (e.g., plans to ensure training). This section summarises results for each of the six stages of facility-based MPDSR implementation.

thumbnail
Table 4. Proportion of facilities meeting the progress markers for each stage of implementation (n = 47).

https://doi.org/10.1371/journal.pone.0243722.t004

Stage 1—Creating awareness.

The two progress markers for this stage were mostly achieved (by at least 68% of facilities). In most facilities (89%), leaders were fully involved in championing death audits, and nearly all facilities (98%) had a focal person responsible for conducting death audits. The individual assigned as the MPDSR coordinator varied by facility level. The facility in-charge was cited most commonly as the MPDSR coordinator in health centres and in small hospitals; the regional/district health officer for provincial, regional, and district hospitals; and the head of the obstetrics and gynaecology, paediatric, or neonatology department for tertiary and private hospitals. Introduction of MPDSR to facility staff varied by country and facility except in Rwanda, where respondents all reported a similar orientation process.

Stage 2—Adopting the concept.

The two progress markers for this stage were mostly achieved. A ‘formal decision to implement MPDSR’ was recalled by facility staff in Nigeria, Rwanda, and Tanzania. However, some facility respondents in Zimbabwe could not recall the decision to begin implementing MPDSR. All facilities in Rwanda and Tanzania had established MPDSR steering committees, whereas only two of three facilities in Nigeria and 13 of 16 facilities in Zimbabwe had established committees.

Stage 3—Taking ownership.

Among the seven progress markers in this stage, four were mostly achieved, one was moderately achieved (34–67% of facilities), and two were rarely achieved (< 33% of facilities), though findings varied among and within countries. Nearly all facilities (94%) could describe or show documentation of MPDSR processes. Standard MPDSR data collection forms were available in 84% of health facilities. Most facilities reported having a policy, guideline, or protocol available at the facility, which was shown to assessors, and for the most part, it was the national guideline. Nigeria was the exception, as facilities reported no written MPDSR policy, guidelines, or tools available in the facility. MPDSR tools included cause of death and modifiable factors in facilities in Rwanda, Tanzania, and Zimbabwe. Most facility tools across the four countries lacked a designated place to document follow-up on actions taken (i.e., response), except for in Rwanda, where the standard MPDSR form includes a place to document follow-up of actions. There was strong awareness of national MPDSR guidelines among facility interviewees in Rwanda and Zimbabwe. Few of the facilities in Nigeria were aware of the national guidelines. In Tanzania, all facilities were aware of the national guideline, but five hospitals demonstrated gaps in adhering with the national guideline, notably around information flow to other levels and community follow-up. Respondents at both the facility and subnational levels described how they valued the process of reviewing cases:

Providing information about preventable factors that contribute to maternal death and using information to guide actions is key for preventing similar death in the future’.

Facility interview, Rwanda

We may think it’s too much to review every death, but each one death is crucial to someone. It might be a statistic to me, but every death matters’.

Stakeholder interview, Zimbabwe

Few facilities had agreements or procedures in place regarding the conduct of MPDSR meetings (9%). Nearly one-quarter of facilities (23%) reported a connection between professional disciplinary actions and MPDSR activities, including one facility in Rwanda, three in Tanzania, two in Nigeria, and six in Zimbabwe. In Nigeria, only one of three facilities reported a nonpunitive, no-blame environment. Respondents described different approaches to assigning blame within MPDSR activities:

Review meetings are where people learn to “stick to the rules”. … Some staff are reprimanded verbally and [receive] other punishments’.

Facility interview, Nigeria

The health worker involved is requested to provide a statement of how the incident happened and may be given a verbal warning or a written one… and in one incident, the responsible person did not work for 1 month’.

Facility interview, Tanzania

One-third of all facilities reported financial or in-kind support from the hospital budget or partner allocations to establish or support MPDSR activities. Hospital or district budget support to establish MPDSR processes varied starkly across facilities, ranging from 15% of facilities in Rwanda, to 33% of facilities in Nigeria and Tanzania, to 69% of facilities in Zimbabwe.

Stage 4—Evidence of practice.

Four of the six progress markers were mostly achieved in this stage. Minutes of MPDSR meetings were observed in 74% of facilities; meeting minutes included action items and respected the confidentiality of staff and patients in two-thirds (68%) of facilities. One-third of facilities (30%) presented meeting minutes with documented follow-up of prioritised actions from previous meetings. Qualitative interviews emphasised the importance of meeting minutes and written recommendations:

We need to document the meetings better with minutes and give the designated actions to the responsible persons in writing’.

Facility interview, Tanzania

One of the most challenging parts of the review process is the formulation of appropriate recommendations, but this step is critical to successful MPDSR’.

Facility interview, Rwanda

Overall, 71% of facilities provided some sort of orientation on MPDSR to facility staff members, ranging from 53% of facilities in Zimbabwe to 100% in Nigeria. The assessment did not explore who attended orientations, how an orientation was conducted, or why one was not conducted.

Only one-third of facilities demonstrated the display or sharing of data trends (e.g., run charts with key statistics posted on a wall). The most commonly mentioned sources of data on death were the labour and delivery registers, followed by the postnatal register. At facilities responsible for capturing information on maternal and perinatal deaths in the community (four of six health centres in Tanzania, nine of 16 facilities in Zimbabwe, and three of 13 facilities in Rwanda), assessors observed gaps in the information provided in the case files. Data sources for compiling case reports in advance of death audit meetings included patient clinical records, registers, transfer/referral forms, and ambulance records. Guidance on methods to classify deaths varied from an optional checklist approach, to open-ended questions on apparent causes of death, to ICD-10 classification (The 10th revision of the International Statistical Classification of Diseases and Related Health Problems [ICD-10]). Less than one-half of the facility respondents (47%) reported that the medical records and registers captured the information necessary to determine cause of death and identify contributing factors (ranging from 27% of facilities in Tanzania to 75% of facilities in Zimbabwe). Cause of death classification systems varied among and within countries. Two-thirds of facility respondents reported using some form of standard coding system aligned with the national guideline on the mortality audit forms (66%). For modifiable factors, almost all facilities reported classifying deaths as avoidable, possibly avoidable, or not avoidable, and/or used the three delays model or a root cause analysis [35]. Facility respondents expressed varying perceptions of the accuracy of data:

One cannot vouch for the accuracy of data being collected because staff are not motivated. They do not know what it will be used for’.

Facility interview, Nigeria

I strongly believe the forms provide adequate information, but the big challenge here resides in providers who do not fill in the necessary information. In general, information is not filled in the forms’.

Stakeholder interview, Zimbabwe

We always need to reconcile the cause of death data from the MPDSR form and register to avoid discrepancies of deaths in facilities’.

Facility interview, Tanzania

Stage 5—Evidence of routine and integrated practice.

Only one of the four progress markers in this stage (multidisciplinary engagement) was mostly achieved in at least two-thirds of facilities, while the other three progress markers were only moderately achieved. Most facilities reported that they assigned specific follow-up actions to individuals with timelines (79%). Less than one-half of the facilities (44%) could actually demonstrate or show any evidence of change(s) made based on recommendations from death reviews (Fig 3). Examples of changes described by facility respondents included improved clinical practices, referrals, documentation, and procurement of essential commodities (e.g., blood). The quote below by a facility respondent provides an example of a successful local response:

Now that the perinatal death is audited, they have started resuscitation of babies who are not crying or breathing. Also, proper use of partographs is now in place’.

Facility interview, Tanzania

thumbnail
Fig 3. Proportion of facilities reporting follow-up of recommended actions from death reviews (N = 47 facilities).

https://doi.org/10.1371/journal.pone.0243722.g003

Though national guidelines included schematics on the reporting structure, including how responses should be tracked, less than one-third (28%) of facilities reported a formal written documentation system for tracking follow-up of recommended actions. Only one facility each in Zimbabwe and Tanzania and three in Rwanda demonstrated a formal process for follow-up of recommendations, apart from reviewing minutes at the next mortality audit meeting. None of the facilities in Nigeria had a systematic process for following up on recommendations.

One-half of facilities held meetings on a predetermined schedule (53%), ranging from 47% in Zimbabwe to 73% in Rwanda. Other facilities held meetings only after a death occurred or on an ad hoc basis. The reporting of regular MPDSR meetings by facility respondents was generally greater than observable evidence of regular meetings (e.g., through review of meeting minutes).

Most facilities demonstrated evidence of multidisciplinary participation in death audit meetings (86%) with representation of a range of health workers from different units, especially in larger facilities. Respondents explained the value of the multidisciplinary nature of the meetings and some of the challenges posed around attendance given staff shortages.

Everyone attends our maternal and perinatal meetings, all the way to the driver, because when we have a case to transfer, he knows why we need to move now’.

Facility interview, Zimbabwe

It’s helping [the MPDSR process]. One person wouldn’t have noted these gaps alone. But together, we are improving the quality of services’.

Facility interview, Zimbabwe

There are not enough staff to attend meeting as well as tend to patients’.

Facility interview, Tanzania

We have difficulty finding an opportunity to gather everyone due to busy schedules’.

Facility interview, Nigeria

Three-quarters of health facilities reported regularly linking MPDSR to other quality improvement activities at their facilities (74%). However, none of the national guidelines included clear guidance on linking MPDSR to quality improvement activities, and the team did not systematically assess the linkages.

One-third of the facilities reported sharing death audit findings, recommendations, and progress with the community (34%), including four facilities in Rwanda, seven in Zimbabwe, and two in Tanzania (none in Nigeria). The reported channels of communication varied among and within countries. Audit recommendations were typically shared with community health workers to disseminate to the community in Rwanda, whereas in Zimbabwe, some facilities reported that a facility staff member was designated as a community liaison and was responsible for sharing recommendations with the community. One facility respondent in Tanzania reflected the desire to provide feedback but did not have a mechanism to do so, a sentiment echoed by other facilities:

We wish that there was a specific mechanism to ensure that MPDSR feedback is shared with the community’.

Facility interview, Tanzania

Stage 6—Evidence of sustained practice.

The three progress markers in this stage ranged from rarely achieved to mostly achieved. Most facilities assessed (83%) achieved the progress marker for demonstrating occurrence of death audit meetings for at least 1 year (irrespective of regularity). Evidence of staff development to sustain MPDSR practice was partially achieved, with only 45% of facilities reporting that staff had received MPDSR training in the past year. A plan in place to ensure all staff receive MPDSR training was rarely achieved by the assessed facilities (24%), with no future plans observed at the facilities in Rwanda and Tanzania. The qualitative responses supported these findings:

By policy, the ward in-charge is supposed to be trained in MPDSR, but she has not had any training, even though she is preparing the case summary’.

Facility interview, Tanzania

Enablers and barriers to MPDSR

Table 5 summarises the top three barriers and enablers of MPDSR implementation as observed by the assessors and as reported by facility informants. The top three enablers observed by the assessors included leadership, regular meeting conducted with participation from a multidisciplinary team, and availability and use of the MPDSR-related guidelines and tools. The top three barriers observed by the assessors included lack of health worker capacity to capture and use data analytically to inform the review process, limited plans for training health workers on the MPDSR process, and limited accountability for the follow-up actions identified during the review process. S3 Table provides detailed results of the identified MPDSR implementation enablers and barriers analyses by country.

thumbnail
Table 5. Top enablers and barriers to MPDSR implementation.

https://doi.org/10.1371/journal.pone.0243722.t005

The most commonly described enabling factors by informants across countries included teamwork, communication between staff, staff commitment, and multidisciplinary participation during meetings. Other reported enablers across the countries included national and subnational support through MPDSR training support and evidence of MPDSR process leading to change or having improved health services. Additional cited enablers included availability of MPDSR guidelines and tools, facility leadership for MPDSR, observed positive effect of MPDSR process on reducing deaths, and staff motivation to support MPDSR due to concern about high number of deaths. The most commonly cited barriers to implementing MPDSR processes described by facility staff included limited staff time, heavy workloads preventing participation in meetings, general staff shortages, and high staff turnover. Other reported barriers included lack of motivation due to absence of incentives for participation in meetings (e.g., travel support) or perceived lack of effect of death audit meetings (e.g., audit recommendations not implemented, health services unchanged.) The most commonly cited changes to improve the utility of MPDSR included actions to motivate staff, such as providing incentives for participation in MPDSR processes, increasing facility staff numbers, increasing MPDSR capacity and skills through additional training and mentorship, more funding and specific resources to facilitate meeting and data collection processes, stronger facility leadership of MPDSR, more regular death review meetings, multidisciplinary participation, and reducing the blame environment.

Discussion

This assessment of MPDSR implementation aimed to characterise the stages of MPDSR implementation progress across several countries using a standardised scoring methodology. The assessment results reinforce previous findings [17,2123] and highlight important implementation gaps and priority areas to strengthen MPDSR systems in low-capacity settings.

Implementation factors

A supportive policy and political environment for MPDSR facilitates implementation but does not guarantee translation into practice [22,23,32,33]. Components in national guidelines that are more straightforward to implement, such as establishment of a steering committee or assigning an MDSR or perinatal death surveillance and response coordinator, generally had greater uptake in facilities. Components of the national guidelines with fewer details (e.g., cause of death classification, or follow-up on action plans or community linkage) demonstrated more variable practice across facilities. Ensuring onsite availability of practical guidance and tools is a critical component at the pre-implementation phase [21]. The history of MPDSR introduction and implementation also matters for sustaining and institutionalising MPDSR practice [24,36], as demonstrated by Zimbabwe, which had the highest overall score (27.38) and has a long history of practising MPDSR in central-level hospitals. While the national guidelines could be strengthened in some areas, such as not having clear instructions on how to follow up on the recommendations, they were mostly aligned with the WHO global guidelines and all had useful tools for implementation, which would enable a supportive policy and political environment to initiate and support implementation [33]. The primary challenge of implementation appears to be at the organizational and individual levels, which are the coalface of implementation [33].

This study confirmed previously reported common facilitators of MPDSR, including the importance of strong leadership and effective teamwork [2124,3741]. Engagement of subnational managers promotes accountability and supports MPDSR practice at facility level through cross-facility/-district learning, capacity-building, and mentorship [24,33,40]. Multifaceted efforts to improve quality of care, including MPDSR, emphasise leadership and teamwork, understanding of the root causes of local quality of care gaps, and the systematic implementation of changes to close gaps [23,32,33,42]. There are many opportunities to strengthen alignment of broader quality improvement and MPDSR processes. For example, MPDSR generates essential information about the local causes of maternal and perinatal deaths and the key contributors to these deaths, which is important for designing robust quality improvement efforts that are responsive to local needs. Quality improvement efforts typically include a systematic change management and monitoring strategy. They can help bolster the systematic follow-up and measurement of the effect of death audit recommendations, an area of weakness identified in this assessment.

Linked to teamwork, the organisational culture around the death audit process can either facilitate or inhibit implementation of MPDSR. Previous studies have found that a lack of trust between health professionals and service administrators, a culture of blame and fear of potential legal ramifications, and the lack of ownership of a process prevent successful implementation [22,32,43]. Failure to comply with principles of confidentiality and anonymity can inhibit implementation practice [22,23,32,41,4346]. A culture of safety in which staff feel protected from disciplinary action and in which death audit data are de-identified and/or kept confidential is a WHO-recommended practice [19,20]. If staff fear repercussions, they are unlikely to support MPDSR or engage fully and productively in an audit process. Elements of individual-level fault-finding and/or disciplinary processes were reported in one-quarter of the facilities in this study, though comments made by respondents during the interview process suggested blame and disciplinary action occurred more than was reported. A study in Nigeria found that the interactional processes among those involved in audit meetings affect the meaningfulness of the death review and may inhibit their impact [34]. Deeper investigation is needed to better characterise and understand the impact that a ‘blame culture’ has on the effectiveness of the MPDSR process. Strategies, such as official audit charters or codes of conduct that are mentioned in the national guidelines, may minimize acrimony and prevent (or reduce) blame and recriminations [47,48]. Few facilities in this assessment had formal agreements or procedures in place regarding the conduct of MPDSR meetings despite facility staff undergoing some type of training or having access to guidelines, which made this recommendation.

Poor staff motivation, limited time and capacity, poorly functioning health systems, and general human resource challenges have also been shown to undermine MPDSR efforts [25,36,37,44,49,50]. Success of MPDSR relies on an individual’s and team’s willingness to ‘self-correct’; commit to honest, open discussions with peers about a traumatic event; and implement recommended actions [33]. When problems identified during review meetings are not followed up on and addressed, staff lose motivation to participate in MPDSR activities [22,34,51,52]. At the facility level, this assessment demonstrated a lack of consistent follow-up of recommended actions and infrequent sharing of success stories arising from the audit process. Further investigation is needed to determine how this affects the motivation of facility staff.

Prior studies demonstrate that the confidence and capability of health workers to complete the review process and analyse death audit data strongly influence implementation of effective MPDSR processes [21,23,24,32,36,41,49,52,53]. Low confidence of managers and health workers to assess causes of deaths and modifiable factors documented in this assessment confirm the findings of prior studies and illustrate the importance of strengthening health worker confidence, skills, and information systems to support MPDSR. Several studies have shown that stronger health information systems, including improved data capture, use, and reliability, can facilitate MPDSR processes [23,32,3638,40,45,47,52]. The common lack of mortality and patient care data in routine health information systems in low-resource settings (e.g., patient records/case notes, facility registers) hinders robust MPDSR implementation, including accurate assignment of cause of death and identification of critical gaps in quality of care [42]. In this assessment, subnational managers expressed concern about the quality of data in facility MPDSR reports, and less than one-half of facility respondents reported that the health information available in their facility was sufficient to classify cause of death and analyse contributing factors. None of the national guidelines in the four assessment countries explicitly aligned with the WHO ICD-10 maternal mortality guidelines [54], published before the most recently updated guidelines in each country, nor the WHO ICD-10 perinatal mortality guidelines, published at the time of the assessment [55]. There is a need to strengthen health information systems and assignment of cause of death guidance in both policy and practice.

Reliance on external funds and/or goodwill of professional organisations to support administration, training, and implementation of MPDSR processes have previously been identified as a barrier to sustainable practice [23,47,56,57]. It is unclear whether designated funding (e.g., a budget line item) is important for effective MPDSR implementation. This assessment did not demonstrate a close relationship between reported budgetary or in-kind support and facility conduct of death audits. Presence of donor support in some areas may have boosted findings of sustainable practice but this would need to be investigated further.

Community engagement may strengthen collective ownership, responsibility (e.g., for referral), and quality of maternal and perinatal care, and may contribute to more robust implementation of MPDSR processes [21,22,32,44,57,58]. The small proportion of facilities reporting sharing death audit recommendations with the community in all four countries deserves greater exploration. Learning from studies of facilities undertaking intentional efforts to engage communities should be further explored to determine how such community engagement might influence the accountability mechanism of death audits and how this may influence community behaviours [22,59].

Measuring implementation

This assessment was the first to our knownledge to apply a standardised implementation progress scoring model to assess MPDSR implementation. The related tool developed for the assessment sought to classify progress markers of MPDSR processes derived from the literature. Its sensitivity in being able to correctly identify a facility’s ability to demonstrate specific implementation markers could not be formally assessed in comparison to alternative tools for MPDSR since it was the first of its kind. The progress markers measure the current status of implementation, especially in terms of tangible and immediate indicators of organizational commitment to implement MPDSR processes including committees formed, training, focal point identified, and availability of tools. It is important to note, however, that the tool was not designed to assess the quality of specific MPDSR processes (e.g. correct assignment of causes of death; robust identification of modifiable contributors to deaths audited; development and follow up of actionable responses to address identified contributors, ability to correct mismanagement etc…). Future applications of this standardised implementation progress scoring model methodology for MPDSR should review the stage-specific progress markers, data collection tools, and process of assigning a standardised implementation score based on learnings from this assessment. Additional progress markers of implementation coverage, such as proportion of deaths reviewed based on national recommendations, should also be considered. Clear operational definitions for each marker will strengthen inter-rater reliability and systematic measurement across sites.

Limitations

The assessment was conducted in a relatively small number of nonrandomly selected facilities in only four countries; therefore, it is not possible to generalise the assessment findings at the country subnational or national level or for the continent of Africa. Given the purposeful, nonrepresentative sample of facilities, the team was not able to analyse potential patterns or differences in MPDSR implementation by facility type (e.g., rural versus urban, primary versus secondary). The nature of the study is a source of possible biases [60]. First, the choice of facilities was made on the basis of a specific program favouring MPDSR. Second, interviews were led by people who may have had an interest in presenting the program in a favourable light. Third, the assessors had a background in clinical care for maternal and newborn health and/or worked for non-governmental organizations, professional associations, or Ministry of Health bringing their own professional background, experiences and prior assumptions. Power dynamics between assessors and those interviewed may have impacted on participants’ willingness to talk openly about experiences. Despite efforts to standardise data collection across countries, the variation in individual assessors and the modest adaptation of data collection tools in each country may have also contributed to some variation in the scoring approach in individual facilities and countries. Data were collected from health workers present at the facility on the specific day of the facility visit; thus, the views and MPDSR activities reported by facility respondents may not capture all facility-specific MPDSR activities or reflect the views of all health care staff, including junior staff, who may be subject to more blame or scrutiny during mortality audit meetings and who may have been absent on the day of the assessment or more hesitant to share their views during group interviews. The non-availablity of subnational stakeholders in Rwanda at the time of the assessment is another limitation to note.

For the most part, this assessment did not differentiate between maternal and perinatal death audit processes. Further research is needed to distinguish differences in death audits and responses for maternal and perinatal deaths. The study included both health centres and hospitals but was not designed to investigate differences in implementation between the two different levels. Further research is needed to explore characteristics of implementing MPDSR in a health center versus a hospital setting.

The assessment set out to measure implementation status and did not evaluate the quality of MPDSR processes (e.g., surveillance completeness, accuracy of cause of death assignment, analysis of modifiable factors, development and follow-up of actions).

Conclusion

This assessment is the first attempt, to the authors’ knowledge, to assess facility-level MPDSR implementation progress using a standardised scoring methodology in multiple countries. Structures and processes for implementing MPDSR existed in all four countries, with over two-thirds of the assessed facilities reaching at least stage 5 –evidence of routine and integrated practice. Many implementation gaps were identified that can inform priorities for strengthening MPDSR implementation. These gaps include ensuring availability of onsite MPDSR guidelines and forms, developing more explicit guidance on cause of death assignment and follow-up of audit recommendations across system levels as part of national guidelines, instituting regular mechanisms to build manager and health worker confidence and skills to implement MPDSR (e.g., training, supervision), strengthening health information systems to permit accurate classification of cause of death and support robust death reviews, strengthening alignment of MPDSR and broader quality improvement efforts, and increasing linkages across system-level MPDSR activities, from community, to facilities, to regional and district health managers. Further implementation research is needed to assess the quality of MPDSR implementation processes and to identify and test mechanisms to overcome common MPDSR implementation gaps in low-capacity settings.

Supporting information

S1 Table. Data sources and collection methods.

https://doi.org/10.1371/journal.pone.0243722.s001

(DOCX)

S2 Table. Mapping content of national MPDSR policy by country.

https://doi.org/10.1371/journal.pone.0243722.s002

(DOCX)

S3 Table. Summary of MPDSR implementation enablers and barriers most commonly cited by facility staff in four countries.

https://doi.org/10.1371/journal.pone.0243722.s003

(DOCX)

S4 Table. Ranking of progress markers by frequency across 47 facilities.

https://doi.org/10.1371/journal.pone.0243722.s004

(DOCX)

S1 File. Brief historical summary of MPDSR processes by country.

https://doi.org/10.1371/journal.pone.0243722.s005

(DOCX)

S1 Data. Database for MCSP multicountry assessment of MPDSR implementation.

https://doi.org/10.1371/journal.pone.0243722.s006

(DOCX)

Acknowledgments

We would like to thank the ministry of health officials, USAID Missions, professional association members, and health workers from Nigeria, Rwanda, Tanzania, and Zimbabwe for their input and support. A very special thanks to many in-country MCSP staff and technical experts who supported the research implementation process. In addition, we would like to thank, Lara Vaz (Save the Children US), Stella Abwao (MCSP), John Varallo (MCSP), Lisa Noguchi (MCSP), Ayne Worku (MCSP), and Brianne Kallam (MCSP) for their technical input and review of drafts of the individual country and multicountry reports. We appreciate the support from Edward Kenyi (MCSP) on data extraction. We thank Sylvia Alford with USAID for her strong support during all stages of planning and implementing this assessment and analyzing and disseminating assessment results.

References

  1. 1. WHO. World Health Statistics 2019. Geneva: World Health Organization, 2019 [cited 19 Sept 2019]. https://apps.who.int/iris/bitstream/handle/10665/324835/9789241565707-eng.pdf?ua=1.
  2. 2. Blencowe H, Cousens S, Jassir FB, Say L, Chou D, Mathers C, et al. National, regional, and worldwide estimates of stillbirth rates in 2015, with trends from 2000: a systematic analysis. Lancet Glob Health. 2016;4(2):e98–e108. pmid:26795602
  3. 3. Oza S, Lawn JE, Hogan DR, Mathers C, Cousens SN. Neonatal cause-of-death estimates for the early and late neonatal periods for 194 countries: 2000–2013. Bull World Health Organ. 2015;93(1):19–28. pmid:25558104
  4. 4. Lawn JE, Blencowe H, Waiswa P, Amouzou A, Mathers C, Hogan D, et al. Stillbirths: rates, risk factors, and acceleration towards 2030. Lancet. 2016;387(10018):587–603. pmid:26794078
  5. 5. Chou D, Daelmans B, Jolivet RR, Kinney M, Say L, Every Newborn Action P, et al. Ending preventable maternal and newborn mortality and stillbirths. BMJ. 2015;351:h4255. pmid:26371222
  6. 6. Tuncalp , Were WM, MacLennan C, Oladapo OT, Gulmezoglu AM, Bahl R, et al. Quality of care for pregnant women and newborns-the WHO vision. BJOG. 2015;122(8):1045–9. pmid:25929823
  7. 7. Kruk ME, Gage AD, Arsenault C, Jordan K, Leslie HH, Roder-DeWan S, et al. High-quality health systems in the Sustainable Development Goals era: time for a revolution. Lancet Global Health. 2018;6(11):e1196–e252. pmid:30196093
  8. 8. Every Woman Every Child. The Global Strategy for Women’s, Children’s and Adolescents’ Health. New York, NY: Every Woman Every Child, 2015 [cited 15 Sept 2019]. http://globalstrategy.everywomaneverychild.org/.
  9. 9. WHO. Every Newborn: An action plan to end preventable newborn deaths. Geneva: World Health Organization, 2014 [cited 11 Nov 2019]. https://apps.who.int/iris/bitstream/handle/10665/127938/9789241507448_eng.pdf;jsessionid=9C37B71BDAF7F50A5096E3FD39D51995?sequence=1.
  10. 10. WHO. Strategies toward ending preventable maternal mortality (EPMM). Geneva: World Health Organization, 2015 [cited 15 Nov 2019]. http://who.int/reproductivehealth/topics/maternal_perinatal/epmm/en/.
  11. 11. Commission on Information and Accountability. Keeping promises, measuring results: Commission on Information and Accountability for Women’s and Children’s Health. Geneva: World Health Organization, 2011 [cited 15 Nov 2019]. https://www.who.int/topics/millennium_development_goals/accountability_commission/Commission_Report_advance_copy.pdf.
  12. 12. Independent Accountability Panel. Old challenges, new hopes: Accountability for the Global Strategy for Women’s, Children’s and Adolescents’ Health. New York, NY: Every Women Every Child, 2016 [cited 13 Sept 2019]. https://www.everywomaneverychild.org/wp-content/uploads/2017/03/IAP_Report_September2016.pdf.
  13. 13. WHO. Standards for improving quality of maternal and newborn care in health facilities. Geneva: World Health Organization, 2016 [cited 5 Sept 2017]. http://www.who.int/maternal_child_adolescent/documents/improving-maternal-newborn-care-quality/en/.
  14. 14. WHO. Quality, Equity, Dignity: A network for improving quality of care for maternal, newborn and child health 2017 [cited 15 Feb 2019]. http://qualityofcarenetwork.org/.
  15. 15. Evidence 4 Action. MDSR Action Network: Evidence 4 Action; 2017 [cited 5 Sept 2019]. http://mdsr-action.net/.
  16. 16. WHO. Global Monitoring of Implementation of Maternal Death Surveillance and Response (MDSR) 2017 [cited 15 Nov 2019]. http://www.who.int/maternal_child_adolescent/epidemiology/maternal-death-surveillance/global-monitoring/en/.
  17. 17. WHO. Time to respond: a report on the global implementation of maternal death surveillance and response (MDSR). Geneva: World Health Organization, 2016 [cited 15 Nov 2019]. http://apps.who.int/iris/bitstream/10665/249524/1/9789241511230-eng.pdf?ua=1.
  18. 18. WHO. Beyond the numbers: Reviewing maternal deaths and complications to make pregnancy safer. Geneva: World Health Organization; 2004.
  19. 19. WHO. Making every baby count: audit and review of stillbirths and neonatal deaths. Geneva: World Health Organization, 2016 [cited 15 Nov 2019]. https://www.who.int/maternal_child_adolescent/documents/stillbirth-neonatal-death-review/en/.
  20. 20. WHO. Maternal death surveillance and response: technical guidance. Geneva: World Health Organization, 2013 [cited 6 May 2015]. http://www.who.int/maternal_child_adolescent/documents/maternal_death_surveillance/en/.
  21. 21. Kerber KJ, Mathai M, Lewis G, Flenady V, Erwich JJ, Segun T, et al. Counting every stillbirth and neonatal death through mortality audit to improve quality of care for every pregnant woman and her baby. BMC Pregnancy Childbirth. 2015;15 Suppl 2:S9. pmid:26391558
  22. 22. Martin Hilber A, Blake C, Bohle LF, Bandali S, Agbon E, Hulton L. Strengthening accountability for improved maternal and newborn health: A mapping of studies in Sub-Saharan Africa. Int J Gynaecol Obstet. 2016;135(3):345–57. pmid:27802869
  23. 23. Smith H, Ameh C, Roos N, Mathai M, Broek NVD. Implementing maternal death surveillance and response: a review of lessons from country case studies. BMC Pregnancy Childbirth. 2017;17(1):233. pmid:28716124
  24. 24. Belizan M, Bergh AM, Cilliers C, Pattinson RC, Voce A, Synergy G. Stages of change: A qualitative study on the implementation of a perinatal audit programme in South Africa. BMC Health Serv Res. 2011;11:243. pmid:21958353
  25. 25. Bandali S, Thomas C, Hukin E, Matthews Z, Mathai M, Ramachandran Dilip T, et al. Maternal Death Surveillance and Response Systems in driving accountability and influencing change. Int J Gynaecol Obstet. 2016;135(3):365–71. pmid:27836470
  26. 26. Healthy Newborn Network. Newborn Numbers Database [Internet]. Save the Children. 2018 [cited. https://www.healthynewbornnetwork.org/hnn-content/uploads/Newborn-Numbers-excel-30-Jan-2018-FINAL.xlsx.
  27. 27. World Bank. World Bank Open Data [Internet]. The World Bank Group. 2019 [cited. https://data.worldbank.org.
  28. 28. Tanzania National Bureau of Statistics. Sub-Divisional Population Projection for Year 2016 and 2017 Based on 2012 Population and Housing Census. Dodoma: Tanzania National Bureau of Statistics, 2019 [cited 15 Sept 2019]. https://www.nbs.go.tz/nbs/takwimu/census2012/Tanzania_Total_Population_by_District-Regions-2016_2017r.pdf.
  29. 29. Nigeria National Bureau of Statistics. National Population Commission and National Bureau of Statistics Estimates: Population 2006–2016. Abuja, Nigeria: Nigerial National Bureau of Statistics 2019 [cited 15 Sept 2019]. https://nigerianstat.gov.ng/download/474.
  30. 30. Bergh A, Kerber K, Abwao S, Johnson Jd-G, Aliganyira P, Davy K, et al. Implementing facility-based kangaroo mother care services: lessons from a multi-country study in Africa. BMC Health Serv Res. 2014;14(1):293. pmid:25001366
  31. 31. Bergh A, Arsalo I, Malan AF, Patrick M, Pattinson RC, Phillips N. Measuring implementation progress in kangaroo mother care. Acta Paediatr. 2005;94(8):1102–8. pmid:16188856
  32. 32. Abebe B, Busza J, Hadush A, Usmael A, Zeleke AB, Sita S, et al. 'We identify, discuss, act and promise to prevent similar deaths': a qualitative study of Ethiopia’s Maternal Death Surveillance and Response system. BMJ Global Health. 2017;2(2):e000199. pmid:28589016
  33. 33. Lewis G. The cultural environment behind successful maternal death and morbidity reviews. BJOG. 2014;121 Suppl 4:24–31. pmid:25236630
  34. 34. de Kok B, Imamura M, Kanguru L, Owolabi O, Okonofua F, Hussein J. Achieving accountability through maternal death reviews in Nigeria: a process analysis. Health Policy Plan. 2017. pmid:28666342
  35. 35. Thaddeus S, Maine D. Too far to walk: maternal mortality in context. Soc Sci Med. 1994;38 (8):1091–110. pmid:8042057
  36. 36. Richard F, Ouedraogo C, Zongo V, Ouattara F, Zongo S, Gruenais ME, et al. The difficulty of questioning clinical practice: experience of facility-based case reviews in Ouagadougou, Burkina Faso. BJOG. 2009;116(1):38–44. pmid:18503575
  37. 37. Filippi V, Brugha R, Browne E, Gohou V, Bacci A, De Brouwere V, et al. Obstetric audit in resource-poor settings: lessons from a multi-country project auditing ‘near miss’ obstetrical emergencies. Health Policy Plan. 2004;19(1):57–66. pmid:14679286
  38. 38. Kongnyuy EJ, Leigh B, van den Broek N. Effect of audit and feedback on the availability, utilisation and quality of emergency obstetric care in three districts in Malawi. Women Birth. 2008;21(4):149–55. pmid:18842471
  39. 39. Purandare C, Bhardwaj A, Malhotra M, Bhushan H, Shah PK. Every death counts: electronic tracking systems for maternal death review in India. Int J Gynaecol Obstet. 2014;127 Suppl 1:S35–9. pmid:25262442
  40. 40. Rhoda NR, Greenfield D, Muller M, Prinsloo R, Pattinson RC, Kauchali S, et al. Experiences with perinatal death reviews in South Africa—the Perinatal Problem Identification Programme: scaling up from programme to province to country. BJOG. 2014;121 Suppl 4:160–6.
  41. 41. van Hamersveld KT, den Bakker E, Nyamtema AS, van den Akker T, Mfinanga EH, van Elteren M, et al. Barriers to conducting effective obstetric audit in Ifakara: a qualitative assessment in an under-resourced setting in Tanzania. Trop Med Int Health. 2012;17(5):652–7. pmid:22469464
  42. 42. Akachi Y, Kruk ME. Quality of care: measuring a neglected driver of improved health. Bull World Health Organ. 2017;95(6):465–72. pmid:28603313
  43. 43. Agaro C, Beyeza-Kashesya J, Waiswa P, Sekandi JN, Tusiime S, Anguzu R, et al. The conduct of maternal and perinatal death reviews in Oyam District, Uganda: a descriptive cross-sectional study. BMC Womens Health. 2016;16:38. pmid:27418127
  44. 44. Kongnyuy EJ, van den Broek N. The difficulties of conducting maternal death reviews in Malawi. BMC Pregnancy Childbirth. 2008;8:42. pmid:18786267
  45. 45. Mutsigiri-Murewanhema F, Mafaune PT, Juru T, Gombe NT, Bangure D, Mungati M, et al. Evaluation of the maternal mortality surveillance system in Mutare district, Zimbabwe, 2014–2015: a cross sectional study. Pan Afr Med J. 2017;27:204. pmid:28904729
  46. 46. Nyamtema AS, Urassa DP, Pembe AB, Kisanga F, van Roosmalen J. Factors for change in maternal and perinatal audit systems in Dar es Salaam hospitals, Tanzania. BMC Pregnancy Childbirth. 2010;10:29. pmid:20525282
  47. 47. Congo B, Sanon D, Millogo T, Ouedraogo CM, Yameogo WME, Meda ZC, et al. Inadequate programming, insufficient communication and non-compliance with the basic principles of maternal death audits in health districts in Burkina Faso: a qualitative study. Reprod Health. 2017;14(1):121. pmid:28969656
  48. 48. Dartey AF. The role of midwives in the implementation of maternal death review (MDR) in health facilities in Ashanti region, Ghana. Cape Town: University of the Western Cape; 2012.
  49. 49. Muffler N, Trabelssi Mel H, De Brouwere V. Scaling up clinical audits of obstetric cases in Morocco. Trop Med Int Health. 2007;12(10):1248–57. pmid:17956508
  50. 50. Bakker W, van den Akker T, Mwagomba B, Khukulu R, van Elteren M, van Roosmalen J. Health workers’ perceptions of obstetric critical incident audit in Thyolo District, Malawi. Trop Med Int Health. 2011;16(10):1243–50. pmid:21767335
  51. 51. Smith H, Ameh C, Godia P, Maua J, Bartilol K, Amoth P, et al. Implementing Maternal Death Surveillance and Response in Kenya: Incremental Progress and Lessons Learned. Glob Health Sci Pract. 2017;5(3):345–54. pmid:28963171
  52. 52. Armstrong CE, Lange IL, Magoma M, Ferla C, Filippi V, Ronsmans C. Strengths and weaknesses in the implementation of maternal and perinatal death reviews in Tanzania: perceptions, processes and practice. Trop Med Int Health. 2014;19(9):1087–95. pmid:25039579
  53. 53. Tapesana S, Chirundu D, Juru T, Shambira G, Gombe NT, Tshimanga M. Evaluation of the Maternal Death Surveillance and Response System, Sanyati, Zimbabwe 2017. TIJPH. 2017;5(2).
  54. 54. WHO. The WHO application of ICD-10 to deaths during pregnancy, childbirth and puerperium: ICD MM. Geneva: World Health Organization, 2012 [cited 15 Nov 2019]. http://www.who.int/reproductivehealth/publications/monitoring/9789241548458/en/.
  55. 55. WHO. The WHO Application of ICD-10 to perinatal deaths: ICD-PM. Geneva, Switzerland: World Health Organization, 2016 [cited 15 Nov 2019]. https://www.who.int/reproductivehealth/publications/monitoring/icd-10-perinatal-deaths/en/.
  56. 56. De Brouwere V, Delvaux T, Leke RJ. Achievements and lessons learnt from facility-based maternal death reviews in Cameroon. BJOG. 2014;121 Suppl 4:71–4. pmid:25236637
  57. 57. Hofman JJ, Mohammed H. Experiences with facility-based maternal death reviews in northern Nigeria. Int J Gynaecol Obstet. 2014;126(2):111–4. pmid:24834852
  58. 58. Magoma M, Massinde A, Majinge C, Rumanyika R, Kihunrwa A, Gomodoka B. Maternal death reviews at Bugando hospital north-western Tanzania: a 2008–2012 retrospective analysis. BMC pregnancy and childbirth. 2015;15:333. pmid:26670664
  59. 59. Bayley O, Chapota H, Kainja E, Phiri T, Gondwe C, King C, et al. Community-linked maternal death review (CLMDR) to measure and prevent maternal mortality: a pilot study in rural Malawi. BMJ Open. 2015;5(4):e007753. pmid:25897028
  60. 60. Malterud K. Qualitative research: standards, challenges, and guidelines. Lancet. 2001;358(9280):483–8. pmid:11513933