Tools for Assessment of Country Preparedness for Public Health Emergencies: A Critical Review

Recent international communicable disease crises have highlighted the need for countries to assure their preparedness to respond effectively to public health emergencies. The objective of this study was to critically review existing tools to support a country’s assessment of its health emergency preparedness. We developed a framework to analyze the expected effectiveness and utility of these tools. Through mixed search strategies, we identified 12 tools with relevance to public health emergencies. There was considerable consensus concerning the critical preparedness system elements to be assessed, although their relative emphasis and means of assessment and measurement varied considerably. Several tools identified appeared to have reporting requirements as their primary aim, rather than primary utility for system self-assessment of the countries and states using the tool. Few tools attempted to give an account of their underlying evidence base. Only some tools were available in a user-friendly electronic modality or included quantitative measures to support the monitoring of system preparedness over time. We conclude there is still a need for improvement in tools available for assessment of country preparedness for public health emergencies, and for applied research to increase identification of system measures that are valid indicators of system response capability.

I nternational communicable disease crises of recent years have highlighted the need for countries to assure their preparedness to respond effectively to public health emergencies. [1][2][3] Further, implementation of the revised International Health Regulations (IHR), a 2013 decision of the European Parliament and Commission on cross border health threats, and concerns over the effectiveness of the initial response to the Ebola crisis, have led to calls for countries to assess and report on the effectiveness of their public health emergency preparedness and response organization and plans. [4][5][6] Despite this, there is only limited knowledge of the methods by which countries assess the state of their preparedness for public health emergencies. There is also little consensus regarding the system elements that should be included in a preparedness system assessment and their validity as predictors of country responsiveness during an actual emergency. 7,8 Available assessment tools were reviewed by Asch et al. 9 and Nelson et al. 10 in 2005 and 2007, respectively, expressing concerns over the validity of methodologies then in use. Since then, a number of assessment methodologies and tools have been developed, with varying relevance to public health emergency preparedness.
The objective of this study was to review the characteristics and utility of available tools and methodologies for assessment of countries' public health emergency preparedness. The underlying purpose of this review was to guide development by the European Centre for Disease Prevention and Control (ECDC) of methodological concepts and tools to support public health emergency preparedness in European Union (EU)/European Economic Area (EEA) member states.

METHODS
Existing tools were located through 3 lines of inquiry: (1) MEDLINE search of peer-reviewed literature, (2) online search of gray literature in public health and civil defense websites, and (3) e-mail contact with ECDC designated emergency preparedness respondents for all EU/EEA member states, to identify nationally developed tools that had not been published elsewhere. Methods 1 and 2 were independently performed by 2 reviewers. The review was conducted between October and December 2014, with updates up to August 2018.
For the MEDLINE search, following a pilot exploration of the data sources and literature, all key words/MeSH terms that could be related with public health emergency preparedness assessment tools were included, and search filters were chosen to be non-restrictive in order to increase search sensitivity. These are given in Table 1. Inclusion criteria were (1) Period: 2000-2017; (2) Languages: reports published in English, Spanish, French, German, or Italian; (3) Category: humans; (4) Scope: subnational, national, or international; (5) Type of hazard: generic (ie, all-hazards approach), or pandemic influenza; and (6) Presence of a checklist, indicators, or measures to assess national public health emergency preparedness status. Civil protection emergency assessment tools were excluded if they did not include public health or health care aspects of the emergency response. The gray literature search included international and national public health and civil defense websites.
A framework to review and compare the identified tools was developed by the investigators, drawing substantively on criteria developed by Nelson et al. 10 and Asch et al. 9 Complementary indicators were extracted from these publications and combined in a single analytical framework, together with 2 further indicators developed by the authors ("completeness," "main advantages") ( Table 2).

RESULTS
A total of 1459 peer-reviewed articles were retrieved and screened for eligibility. Only 2 assessment tools meeting the quality criteria were identified through a peer-review literature search: United States National Health Security Preparedness Index (NHSPI™) 11 and Harvard School of Public Health Emergency Preparedness Exercise Evaluation Tool a (H-EPREP). 12 Another 9 tools were identified from the search of public health and civil defense websites: European Commission technical guidance on generic preparedness planning for public health emergencies 13 and template for providing information on preparedness and response planning in relation to serious cross-border threats to health 14 ; World Health Organization (WHO) questionnaire for monitoring progress in the implementation of IHR Core Capacities In States Parties 15 and Joint External Evaluation Tool (JEE) 16 ; Toolkit for assessing health system capacity for crisis management of the WHO Regional Office for Europe 17 ; Joint European Pandemic Preparedness Self-Assessment Indicators 18 ; National Health Services England Core Standards for Emergency Preparedness, Resilience, and Response (EPRR) 19 ; New Zealand Civil Defence and Emergency Management (CDEM) Capability Assessment Tool 20 ; and United States Centers for Disease Control and Prevention (CDC) Public Health Preparedness Capabilities. 21 One further tool was identified by an ECDC national respondent: the EpiSouth National Generic Emergency Preparedness Plans a (E-EPREP). 22 The 12 tools identified are summarized in Table 3. All were published between 2009 and 2016. Seven of the 12 tools were developed by international authorities or organizations. [13][14][15][16][17][18]22 The other 5 were country-specific: from England, 19 New Zealand, 20 and the United States. 11,12,21 Some of the tools developed by international organizations had a primary focus on voluntary country level implementation. 13,17,18,22 Others had an apparent primary rationale of required or recommended reporting under international or European regulations. [14][15][16]  Appraisal of the tools according to the evaluation framework is summarized in Table 4. All tools identified had governmental or institutional authorities as the principal target audience and specified an accountable entity, but with varying degrees of detail. Although all tools had public health emergencies as a primary focus, they varied in their relative emphasis on various aspects of emergency preparedness, including health system resilience and the wider civil emergency protection function. All except 1 18 took an all-hazards approach, although they mainly focused on communicable (infectious) disease emergencies with some additional sections for other types of public health hazard such as chemical or radiological events.
The key assessment areas included in each of the tools are outlined in Table 5. Some areas were common to nearly all tools,

Evidence based
Extent to which the observational or experimental evidence was clearly provided for the indicated actions or capacities.

Feasibility
Extent to which the measures were user friendly and did not impose excessive burdens. This was assessed on author's judgment when they have tried to use and/or complete the different tools.

Utility
Extent to which the tool supported decisions related to improvement, accountability, and other functions. Improvement-oriented assessments are usually aimed at internal audiences. Accountability-oriented assessments are usually aimed at external stakeholders. 10 In this case, utility was linked to the actual use of the tool output (eg, list of tasks to perform vs external evaluation measures) and the methodology used (eg, qualitative vs quantitative).

Specification of an accountable entity
Extent to which the tool identified an individual or group of the evaluated institution as responsible for completion of each component indicator.

Main advantages
Characteristics of the tool that could be useful for a European tool development.
Modified from Asch et al. 9 and Nelson et al. 10     with varying degrees of detail and methodological approaches, for example, interoperability and inter-sectoral collaboration, crisis management and operations, planning, communication and information systems, and human resources and capability development. Other assessment areas were addressed less frequently, for example, recovery, community preparedness, cross-border issues, or ethical aspects.
Most tools identified provided little or no information on the criteria or decision processes used to identify the measures included in them, or the evidential approach taken for their development. Exceptions included the JEE, 16 CDC, 21 H-EPREP, 12 and NHSPI 11 tools. In most cases, the development of preparedness standards appeared to be based primarily on consultations with groups of experts. Only a minority of tools attempted to describe a conceptual and strategic framework underlying their design. 11,12,21 The Civil Defence and Emergency Management Tool 20 from New Zealand had the most comprehensive, logical, and updated framework, consistent with current concepts of health emergency preparedness. 23,28 Most of the selected tools had clear measurement parameters, with different methodological formats and complexity. These varied from a detailed list of tasks [13][14][15]21,22 to simple qualitative scales, 12,[16][17][18][19] through to more complex scoring systems. 11,20 Four of the tools included a quantitative element: JEE, 16 CDEM, 20 H-EPREP, 12 and NHSPI. 11 The CDEM tool had a scoring system with weighted indicators that can be customized according to national or regional priorities. 20 Seven of the tools were paper-based only, with no electronic informatics to facilitate use. [13][14][15][16][17][18]21 Two were presented for use as an Excel file (NHS 19 and CDEM 20 ), and 3 had online modalities (H-EPREP, 12 NHSPI, 11 and E-EPREP 22 ). Two allowed a degree of customization by the user: CDEM 20 and H-EPREP. 12 H-EPREP 12 was an exception, providing for the generation of

DISCUSSION
The use of systematic methods and tools for system assessment should have substantive benefits for the preparedness of countries for public health emergencies. The tool infrastructure should in itself have symbolic value to help communicate a coherent view of the emergency preparedness system to all participants. This should cover all of the elements critical to ensuring an effective response, including effective collaboration across sectors and between countries in responding to cross-border events. The systematic assessment of these elements should enable gaps and weaknesses to be proactively identified and addressed. To achieve this, tools should include assessment items, which are valid indicators of actual performance in an emergency. They should be available in userfriendly electronic modalities and include quantitative elements to support the monitoring of system preparedness over time and voluntary benchmarking with others, to promote learning and system improvement.
We have identified 12 presently available tools to support assessment of country preparedness for public health emergencies. Most tools were found through national and international websites, and it is possible that more may have been identified through gray literature in languages other than those included in this search, at subnational level, and sources such as postgraduate theses.
Few of the identified tools meet all of the above requirements. We acknowledge a potential limitation of our appraisal in that the tools were evaluated as a desktop exercise based on a priori criteria; however, the evidence base from user experience of the presently available tools is almost non-existent.
Our review suggests some possible contributing perspectives on this present situation. Available tools appear to have been developed with somewhat different primary aims and methodological approaches. Most tools developed by international agencies and 1 in the United States appeared to focus primarily on standard reporting requirements to which countries and states are subject. Exceptions included the self-evaluation checklists developed by the European Commission 13 and the WHO Regional Office for Europe, 17 which appear to be designed explicitly for country use. Tools developed by national authorities provide a primary focus on the evaluation needs of the country but may not extrapolate well for use by others, given country-specific characteristics of health and public health emergency response systems. Further, countrylevel tools may have less utility for subnational (regional, local) jurisdictions, and vice versa.
Country preparedness evaluations need to assess not only plans and capacities, but also system capabilities for effective response to actual emergencies. Several tools relied heavily on input data relating to system capacities and resources; while information concerning these is often readily available, it may be only indirectly predictive of the capability to respond to an emergency. Nelson et al. observed in 2007 that the few tools then available to assess preparedness status tended to focus on capacities, and little evidence existed that linked specific structures with the ability to execute effective response processes, noting that "structural measures may not be valid indicators of preparedness." 10 In reporting on a review of national influenza pandemic preparedness plans in the EU in 2012, Nicoll noted that some national authorities had ceased further preparedness development after producing written plans and had neither developed operational aspects nor tried to assess whether they would work in practice. 24 The present study suggests only modest advance in this respect; among the identified tools, only the CDEM, 20 H-EPREP, 12 and JEE 16 tools included significant consideration of system capabilities, as well as capacities.
The evidence base linking preparedness capacities and capabilities to health outcomes remain weak. 7,10 Asch et al. noted in 2005 that most instruments for assessing public health emergency preparedness relied excessively on subjective or structural measures and lacked a scientific evidence base. 9 Previous literature reviews have found that the majority of journal articles were commentaries and anecdotal case studies, based on qualitative analyses, 8,25-27 a situation unchanged in our present literature search in support of this critical tools analysis (to be reported separately). One systematic review concluded that most studies lacked a rigorous design, raising questions about the validity of the results. 7 It appears that more and better quality research into public health emergency management is needed for the development of useful assessment tools, and the validity of presently assessed system elements as predictive of actual response capability remains largely unverified. This is also the conclusion of the developers of other tools, which attempted to provide some evidence-based approach, who ended up relying mainly on lessons learned documents 11,12,16,21 (see Table 4). A focus of future research should include the comparison of preparedness system a priori assessment scores and the actual system performance outcomes in real-life incidents and emergencies.
As the tools reviewed did not have a documented strong evidence base, there was only partial consensus on the system elements critical for public health emergency preparedness, and how they may be assessed or measured. Although some system areas were common to most tools, there was significant diversity in the system elements included and their emphasis across the tools reviewed, and in the indicators or standards used to measure their effective presence. "The problem lies not in the absence of standards per se, but in the multiplicity of overlapping (and sometimes conflicting) standards." 10,28 One issue underlying indicator development appears as differing preferences for standardizing all system measures, or leaving countries' flexibility to modify, add, or delete them. Some authors have recommended standardization of all assessment measures 9,29 in order to facilitate comparisons, either to a "gold standard" or between countries. However, some emergency response leaders consider that this is less useful than a flexible country-specific tool, given different country administrative structures and health care systems. Respondents to the EU pandemic influenza preparedness review in 2009 considered that "instead of [standardized] indicators, it would be more useful to develop a tool describing the main areas for consideration in pandemic influenza preparedness planning. Each country may then add its own criteria, indicators or outcomes for determining whether something is in place." 23 This choice, in turn, appears to also reflect divergent views on the perceived value of sharing country information and benchmarking with others. In the same review, "a number of member states made it clear to the ECDC that the country specific results should only be known to the country […] and that specifically there would be no 'league tables.'" 23 Few tools were available in user-friendly, electronic modalities that could facilitate data gathering, analysis, and dissemination and discussion of results by participants and stakeholders. H-EPREP 12 was an exception, as it also allowed the generation of customized exercise evaluation forms, storage, transmission to multiple evaluators by e-mail, and generation of basic reports. Developers should therefore be encouraged to produce assessment tools in more user-friendly modalities. Inclusion of quantitative scoring systems usefully support the monitoring of progress in the development of a country's public health emergency preparedness over time. Such quantitative scoring systems can also facilitate voluntary benchmarking with other countries. However, few tools included this feature. Only 2 tools had been published in a manner accessible to a conventional literature search 28,30 ; most were available only through the websites of the organizations that developed them.

CONCLUSIONS
Methods and processes for assessment of country systems are an integral part of a holistic approach to assuring country emergency preparedness, including simulation exercises, after action reports and peer reviews. 31 We conclude, however, that few of the existing tools satisfy all or most of the requirements for utility and effectiveness discussed previously. There is a continuing need for further improvement in tools available for countries' assessment of their preparedness for public health emergencies. Existing tools could be revised with critical review of the validity of their assessment elements and indicators, and availability in more userfriendly electronic format with analytical and reporting modalities. New tools could be developed de novo at country and supranational level based on both a country's needs and best available evidence relating to the validity of its assessment elements and indicators.
The paucity of applied emergency response systems research remains a significant impediment to achieving these improvements. In particular, the elements of the preparedness system that are valid indicators of actual response capability remain poorly understood. Reporting and critical review of user experience of all of the different means of evaluating country preparedness should contribute to this goal. 31