CDC Program Evaluation Framework, 2024
Recommendations and Reports / September 26, 2024 / 73(6);1–37
Daniel P. Kidder, PhD1; Leslie A. Fierro, PhD2,3; Elena Luna, DrPH1; Heather Salvaggio, MPH1; Amanda McWhorter, DrPH4; Shelly-Ann Bowen, PhD5; Rebecca Murphy-Hoefer, PhD6; Sally Thigpen, MPA7; Dayna Alexander, DrPH5; Theresa L. Armstead, PhD8; Euna August, PhD9; Destiny Bruce, MPH1; Seseni Nu Clarke, MPH, MA10; Cassandra Davis, MPH1; Amia Downes, DrPH11; Sarah Gill, MA12; L. Duane House, PhD13; Michael Kerzner, MPH1; Karen Kun, MPH, MA14; Karen Mumford, PhD15; Leah Robin, PhD16; Dara Schlueter, MPH17; Michael Schooley, MPH18; Eduardo Valverde, DrPH19; Linda Vo, MPH1; Donjanea Williams, EdD20; Kai Young, MPH9; CDC Evaluation Framework Work Group (View author affiliations)
View suggested citationAltmetric:
Summary
Program evaluation is a critical tool for understanding and improving organizational activities and systems. This report updates the 1999 CDC Framework for Program Evaluation in Public Health (CDC. Framework for program evaluation in public health. MMWR Recomm Rep 1999;48[No. RR-11];1–40) by integrating major advancements in the fields of evaluation and public health, lessons learned from practical applications of the original framework, and current Federal agency policies and practices. A practical, nonprescriptive tool, the updated 2024 framework is designed to summarize and organize essential elements of program evaluation, and can be applied at any level from individual programs to broader systems by novices and experts for planning and implementing an evaluation. Although many of the key aspects from the 1999 framework remain, certain key differences exist. For example, this updated framework also includes six steps that describe the general process of evaluation planning and implementation, but some content and step names have changed (e.g., the first step has been renamed Assess context). The standards for high-quality evaluation remain central to the framework, although they have been updated to the five Federal evaluation standards. The most substantial change from the 1999 framework is the addition of three cross-cutting actions that are core tenets to incorporate within each evaluation step: engage collaboratively, advance equity, and learn from and use insights. The 2024 framework provides a guide for designing and conducting evaluation across many topics within and outside of public health that anyone involved in program evaluation efforts can use alone or in conjunction with other evaluation approaches, tools, or methods to build evidence, understand programs, and refine evidence-based decision-making to improve all program outcomes.
Introduction
Program evaluation is a critical function that communities and organizations should undertake to improve and strengthen their activities and systems (1). As a scientific activity, program evaluation uses systematic data collection and analysis of programs, policies, and organizations to assess their effectiveness and efficiency (2–4). Evaluation can provide insights to many questions, including the strengths of current programs and areas for improvement throughout a program life cycle, such as the adequacy of program resources, accuracy of program assumptions, quality or fidelity of program operations, and the intended and unintended effects of a program (5).
Although program evaluation often uses methods that are also used in research, the purpose is different. Whereas research primarily aims to contribute to generalizable knowledge, evaluation aims to continuously improve programs and organizations and produce findings and recommendations for decision-making (4,6). Program evaluation can help clarify how to improve existing programs and build upon their strengths, why a program is or is not being implemented as planned or producing intended results, and why certain trends or patterns are observed in existing data sources.
The understanding of evaluation and its importance for improving programs has increased over the past several decades. At the Federal level, the Foundations for Evidence-based Policymaking Act of 2018 (Evidence Act) prioritizes and “elevates program evaluation as a critical agency function,” and recognizes that “pressing challenges face our nation today, with urgent needs for evidence about the approaches that work best to … address current and future challenges” (4). As such, Federal agencies are investing in building evaluation capacity and focusing more efforts on evaluating their programs and activities and using the findings about program processes and outcomes for decision-making and continuous program improvement. Program evaluation is most effective when it is appropriately resourced and fully integrated into the entire lifecycle of a program (i.e., program design to program conclusion) (7).
CDC’s program evaluation portfolio has been guided by the 1999 Framework for Program Evaluation in Public Health (8) (the 1999 framework). The 1999 framework is a “practical, nonprescriptive tool, designed to summarize and organize essential elements of program evaluation” (8). It placed evaluation within the broader context of public health practice and described a clear six-step process for conducting evaluation as well as four central standards to guide the production of high-quality evaluations (9). The original framework also represented a major milestone for evaluation practice at CDC by providing a central organizing process for planning and implementing evaluations that could be used by agency programs both to inform their own evaluative efforts and to provide guidance to funding recipients about the agency’s expectations for how evaluations would be performed (6). Since its publication, the 1999 framework has been cited in approximately 300 peer-reviewed articles and has grown from an internal tool used to guide effective program evaluation within CDC to a foundational element of projects worldwide, reaching approximately 50 countries on six continents. In addition, the application of the framework has extended beyond public health into areas such as clinical research, education, and the military (6,10,11).
CDC recognized the need to update the original CDC framework with advancements in the fields of evaluation and public health over the past 25 years; integrate lessons learned from practical applications of the framework; and align with current Federal agency policies, practices, and priorities (3,4,12–16). To reflect the current state of program evaluation and extend the framework’s value and use, CDC updated the framework through a multiyear process. Two key principles guided the questions, discussions, and decisions that informed this update: 1) maintain the practicality and simplicity of the framework and 2) refresh the framework rather than engage in a wholesale change. The widespread use and application of the original framework across many different contexts suggested that many aspects of the framework were useful. Thus, the updated framework retains the aspects that guided evaluations over the past several decades and includes modifications and new materials that allow evaluators continued flexibility in applying the framework across a wide array of contexts, including those that are emerging and evolving.
Methods for Updating the Program Evaluation Framework
The process for updating the framework included gathering information and insights from various evaluation framework users, including input from Federal and non-Federal evaluators, and a literature review of evaluation and public health publications. The methods and procedures are described briefly here, and additional information is available (https://stacks.cdc.gov/view/cdc/160381). To guide the process, an 80-person Evaluation Framework Work Group was convened, comprising CDC staff with expertise in evaluation representing a range of CDC programs. A subgroup of volunteers from the larger work group formed a steering committee with 21 members that guided development and implementation, reviewed results, discussed and determined revisions, and contributed to revising the framework. Feedback on framework drafts was gathered from reviewers within and outside CDC and incorporated into the final document.
Groups of framework users were identified. These groups and persons provided information to understand how the framework was adapted and used in different settings, what aspects of the framework were useful, challenges experienced when using the framework, and gaps identified in the framework’s content. A mixed methods approach for gathering information included in-depth interviews, surveys, listening sessions, and a request for public comment (17). Data were collected during March 2022–February 2023 from approximately 850 groups or persons who had experience using the evaluation framework. Information was collected through a survey of Federal employees and staff (n = 123) within and outside of CDC, virtual listening sessions with approximately 450 CDC-funded recipients and evaluators working with members of American Indian or Alaska Native communities (n = 172), telephone interviews with Federal evaluation leaders (n = 11) from seven Federal agencies, relevant responses to a public request for information (n = 22), a virtual session with work group members (n = 53), and attendees at a CDC Evaluation Day session (n = 38).
Data were cleaned and analyzed in Microsoft Word and Excel, and descriptive statistics for quantitative data and thematic analysis for qualitative data were generated. Information was synthesized across all activities with major themes related to conceptual design recommendations, terminology additions or revisions, and areas for additional detail or clarity.
A literature review was conducted to identify key conceptual evaluation advancements to include in the framework. The literature review included English language scholarship and practice published during 2013–2023 in a purposive sample of evaluation (n = 15) and public health journals (n = 2) selected by the steering committee that were indexed in Scopus and the Education Resources Information Center library. Search criteria included 41 search terms in article title, abstract, or keywords related to the cross-framework actions and steps. Results of the literature search yielded 3,436 publications from Scopus, 290 of which were analyzed for full-text review. Data were extracted using Covidence literature management software and analyzed in Excel to identify key themes. This activity was reviewed by CDC, deemed not research, and was conducted consistent with applicable Federal law and CDC policy.*
Defining Key Concepts
Before describing the updated framework in detail, multiple key definitions are provided to facilitate a common understanding of the program evaluation framework content:
- Evaluation: An assessment using systematic data collection and analysis of one or more programs, policies, and organizations intended to assess their effectiveness and efficiency (2).
- Program: Any set of related activities undertaken to achieve an intended outcome (18). In this context, program is used to describe the object of evaluation, which could be any organized public health action. This definition is deliberately broad because the framework can be applied to almost any organized public health activity including interventions, surveillance systems, policy development and implementation, outbreak investigations, emergency response efforts, laboratory diagnostics, mass media and communication initiatives, infrastructure projects, training and educational activities, community efforts, research initiatives, and systems (8).
- Interest holder: Any person or organization having an investment in the evaluation, such as those served or affected by the program, those planning or implementing the program, those who might use the evaluation findings, and those who are skeptical about the program. Previously referred to as “stakeholder” (8), a term that can indicate a power differential between groups and that is recognized as having a violent connotation for certain American Indian or Alaska Native tribes and tribal members (19–22). Advancing equity requires many actions, one of which is using inclusive and respectful language in communications. Replacing the term stakeholders aligns with an equity-centered approach to communications because it recognizes the cultural, linguistic, environmental, and historical experiences of the many audiences of this evaluation framework and persons who might be affected by use of the framework (19,23,24). Stakeholder was replaced with interest holder to emphasize that anyone with an interest in the evaluation or program that is the subject of the evaluation are to be engaged in this collaborative process.
- Performance measurement: Ongoing monitoring and reporting of program accomplishments, particularly progress toward pre-established goals (4).
- Standards: Factors that guide evaluation decisions and are used to determine what constitutes high-quality program evaluation (16).
- Health equity: The state in which everyone has a fair and just opportunity to attain their highest level of health. Achieving this requires ongoing societal efforts to address historical and contemporary injustices; overcoming economic, social, and other obstacles to health and health care; and eliminating preventable health disparities (14).
Types of Evaluation
Various evaluation terms are used and many types of evaluation can be applied in different contexts, which are continuously evolving. Summarizing all types of evaluation is beyond the scope of this report. Certain key types of evaluation are as follows:
- Formative evaluation: Assesses whether a program, policy, or organizational approach, or some aspect of these, is feasible, appropriate, and acceptable before it is fully implemented. It can include process and outcome measures (4).
- Process or implementation evaluation: Assesses how the program, intervention, operation, or regulation is implemented relative to its intended theory of change. It often includes information on processes, content, quantity, quality, and structure of what is being assessed (4).
- Outcome evaluation: Measures the extent to which a program, policy, or organization has achieved its intended outcome(s). It cannot attribute causality (4).
- Impact evaluation: Estimates and compares outcomes with and without the program, policy, or organization, usually seeking to determine whether a causal relation can be established between the activity and the observed outcomes (4).
- Economic evaluation: Examines program effects relative to the costs of the program. Common approaches include cost analysis, cost-benefit analysis, cost-effectiveness analysis, and cost-utility analysis. It might overlap with other evaluation types depending on the evaluation question(s) and type of economic evaluation used (25).
Program Evaluation Framework
The program evaluation framework guides the design and implementation of high-quality evaluations by providing a structure that summarizes and organizes the essential elements of program evaluation. The framework can be applied at any level: a single intervention, multicomponent interventions, programs comprising multiple projects, and even broader systems. In addition, the framework can be used in conjunction with other frameworks from within or outside of public health.
Although much of the updated framework is similar to the previous version, certain key differences exist (Table 1). Like the original, the updated version includes six steps that describe the general process of evaluation planning and implementation and a set of evaluation standards that describe what constitutes high-quality evaluation. Some step names and content have changed and been updated, and the first step is new. The standards remain central to the framework, although they have been changed to the Federal evaluation standards (16). The most substantial change from the original framework is the addition of cross-cutting actions, which represent three core tenets to be addressed in each framework step: 1) engage collaboratively, 2) advance equity, and 3) learn from and use insights. The remainder of this report describes in detail the Federal evaluation standards, the cross-cutting actions, and each step (Figure 1) (Box 1).
Program Evaluation Standards
Evaluation standards such as relevance and utility, rigor, independence and objectivity, transparency, and ethics are designed to improve the quality of evaluations by guiding decisions made in each step of the planning and implementation process. These standards are not intended to be applied as rigid rules and are intentionally broad to give evaluators and interest holders the flexibility to consider their unique circumstances, weigh the various options, and determine the best course of action. Depending on circumstances, the importance of one evaluation standard might need to be balanced with the relative importance of another standard, and decisions made based on this balance might need to be revisited if the relative importance of the standards changes during an evaluation.
Federal evaluation practice adheres to five broad evaluation standards that are part of the Evidence Act implementation and described in detail in the Office of Management and Budget M-20–12 (16). These are similar to evaluation standards used by other organizations (26). The standards are intentionally broad to provide flexibility to adapt to the unique circumstances, weigh the various options, and determine the best course of action. Flexibility should not be interpreted as ignoring or considering options that are contrary to the evaluation standards. Because evaluation is a scientific activity, those conducting evaluations should “uphold scientific integrity as they undertake evaluation activities” (4).
The evaluation standards summarized in this report reflect current thinking on evaluation standards and practices (16,27). They might be updated to address changes in the evaluation landscape and to incorporate new practices developed over time (16). These standards are complementary to, and can be used in conjunction with, evaluator competencies (15), evaluator guiding principles (13), cultural competence statements (12), and other evaluation standards (26) to ensure robust evaluation practice.
Relevance and Utility
Evaluations should address questions of importance and serve the information needs of interest holders to be useful. Findings should be actionable; available in time for use; and presented in ways that are understandable, culturally responsive, and informative for taking action (e.g., budgeting, program improvement, accountability, management, regulatory action, policy development, and strategic planning).
Rigor
Evaluations should produce findings that interest holders can confidently rely on while providing clear explanations of limitations. The rigor of an evaluation depends on thoughtful planning and implementation of the underlying design and methods as well as how findings are interpreted and reported. Credible evaluations should be planned, implemented, and interpreted by qualified evaluators in collaboration with interest holders. These evaluators should have the relevant education, skills, and experience for the methods undertaken. An evaluation should use the most appropriate design and methods to answer the evaluation questions while balancing the evaluation goals, scale, timeline, feasibility, and available resources.
Independence and Objectivity
Evaluations should strive to be as independent and objective as possible for interest holders, experts, and the public to accept their findings. The implementation of evaluation activities should be appropriately insulated from political and other undue influences that might affect their objectivity, impartiality, and professional judgment. Evaluators should strive for objectivity in the planning and conducting of evaluations and in the interpretation and dissemination of findings, avoiding conflicts of interest, bias, and any other partiality. To enhance objectivity, evaluators can regularly assess their potential biases, which might affect whom they choose to engage in an evaluation, to what they pay the most attention, and what they might be overlooking as a result.
Transparency
Evaluations should be transparent in the planning, implementation, and reporting phases to enable accountability and help ensure that aspects of an evaluation are not tailored to generate specific findings. Decisions about the evaluation’s purpose and objectives, the range of interest holders who will have access to details of the work and findings, the design and methods, and the timeline and strategy for releasing findings should be clearly documented before conducting the evaluation. Once evaluations are complete, comprehensive reporting of the findings should be released in a timely manner and provide sufficient detail so that others can review, interpret, replicate, or reproduce the work.
Ethics
Evaluations should be conducted to the highest ethical standards to maintain trust in the process and products. Evaluations should be planned and implemented to safeguard the dignity, rights, safety, and privacy of participants and other interest holders or affected entities. Evaluators should abide by current professional standards pertaining to treatment of participants. Evaluations should be equitable, fair, and just and should account for cultural and contextual factors that could influence the findings or their use.
Cross-Cutting Actions
How evaluators and interest holders approach the act of evaluation is as important as what steps are used to carry out the evaluation process. The following three cross-cutting actions are core tenets of evaluation practice that need to be addressed within each of the steps: 1) engage collaboratively, 2) advance equity, and 3) learn from and use insights. Intentionally integrating each of these actions throughout an evaluation is more likely to produce rigorous evidence that is informed by multiple perspectives derived from both lived and professional experiences; generate insights that improve understanding about how to design and implement public health activities that further health equity; and provide evidence that is meaningful, informative, timely, and actionable. In addition, engaging in evaluations that embrace collaboration, welcome all voices, and create safe environments that support learning from both successes and failures can help to instill ongoing reflection and continuous improvement that are rooted in equitable processes.
The cross-cutting actions described in this section are depicted as wrapping around the framework steps (Figure 1). This is intentional because these actions affect how all six steps are performed. As such, though these actions are described, their use also is integrated into the discussion of each step later in the document.
Engage Collaboratively
Central to the framework is facilitating co-ownership of the program evaluation with interest holders to improve collaboration and input on decision-making. This collaboration starts at the beginning of the evaluation planning process and continues into the implementation and interpretation phases. The process actively engages persons with a broad range of perspectives derived from both lived and professional experiences. Within each step of the evaluation process, evaluators are responsible for intentionally creating an environment that is open to and respects all views by building trusting relationships and encouraging active sharing and listening to ideas among all parties involved (15).
Collaborative approaches to evaluation can have multiple benefits. For example, engaging a diverse set of interest holders can improve the program evaluation process and outcomes by providing a more complete understanding of the context, including its complexities (28), and help decrease overemphasis on values held by specific persons or groups (29). In addition, this approach can increase the validity of findings (30) and improve the likelihood that results are used by interest holders (30–32). Persons reviewing the evaluation findings might view the results as less credible if interest holders are not engaged in the evaluation (33) in a manner that respects and integrates their perspectives (34). Engaging interest holders has the additional potential benefit of improving the evaluation capacity of the persons who take part (as well as the organizations with which they are affiliated) because of learning through hands-on experiences (30,35,36).
The extent to which interest holders are involved directly in planning and implementing the program evaluation will vary and depend on factors such as interest holder availability (30,37). Collaboratively engaging interest holders in evaluation might require a shift in thinking among evaluators and those involved with evaluations toward the allocation of sufficient time for engagement and inclusion throughout an evaluation (38). Evaluations that fully engage interest holders in true collaboration often require more time (30), so evaluators need to consider how to best balance providing timely evidence for decision making while ensuring adequate time for collaborative engagement. In addition, collaborative evaluations require that evaluators effectively facilitate power differentials and dynamics among types of interest holders and between the evaluator and the interest holders (30).
Respecting and understanding the availability of collaborators throughout the evaluation process is important. Persons who can collaborate at the beginning of an evaluation might not be available throughout the entire evaluation, as contexts and circumstances might naturally change. Evaluators are responsible for communicating with collaborators throughout the evaluation to determine whether the level and type of collaboration continue to be effective for those involved while maintaining a range of perspectives. Engaging in evaluation planning and implementation processes requires commitment, and it is important that a person’s time and expertise are valued appropriately within the context.
Advance Equity
Evaluation can contribute to advancing equity and eliminating health inequities in multiple ways (14). First, by using collaborative and equitable evaluation approaches, evaluators can create environments where everyone is respected and heard (39). Such environments can advance equity by creating forums where interest holders who might otherwise not have been as involved are able to share their perspectives (14,29,30). Evaluators should be attentive to the range of interest holder perspectives and the diverse (and sometimes divergent) input that might arise, understand how power differentials between interest holders can influence equity in the evaluation process (28), and consider how to balance these perspectives and power differentials. Including members of typically underrepresented groups (e.g., affected communities) and ensuring that they be represented, involved, and heard within the evaluation development, implementation, and interpretation phases is important, with lived experiences welcomed and valued alongside technical or professional experiences. When engaging with different interest holders, employing equitable communication principles throughout the evaluation can reduce bias in language and enable positive and constructive interactions (40).
Second, in each step of the framework and when applying the standards, evaluators and interest holders can think about how to advance equity and potential effects of decisions. Doing so can create an intentional process in which evaluation discussions can provide insights about how the program can address drivers, which are “factors that create, perpetuate, or exacerbate a health inequity” (14,39). For example, when describing the program (Step 2), evaluators can ask questions about how the program’s underlying theory of change addresses the drivers of health inequities. If the theory of change does not address these topics, interest holders can discuss what opportunities exist. In addition, care should be taken during evaluation planning to ask key evaluation questions, data collection methodologies, and data analyses (Steps 3–5) that explore and examine drivers of health inequities (14,41). These actions can produce evaluation findings that provide valuable insights about how the public health program is already contributing to health equity or how it can be adjusted to facilitate change.
Third, evaluators can conduct evaluation in a culturally responsive way that recognizes “the shared experiences of people, including their languages, values, customs, beliefs, and mores … worldviews, ways of knowing, and ways of communicating” (12,28). Culturally responsive evaluation integrates the uniqueness of each context into the design and implementation of an evaluation, including the history, systems, and structures that can contribute to health inequities. In understanding and respecting different cultures within evaluation, it is important to acknowledge that the data provided for an evaluation depict the experiences of communities. As such, evaluators can consult with communities on the best approaches to share their story with others (e.g., release of findings), incorporate ways to benefit those providing insights throughout the evaluation and reciprocate their contributions throughout the evaluation, collaboratively interpret evaluation findings, and review communications products with interest holders so they appropriately reflect the context and avoid unintended bias (19). In addition, it is important to be aware of, and adhere to, the specific rights, rules, and procedures that exist within certain contexts relating to data collection; data ownership; and how data, knowledge, and insights are shared (e.g., tribal data sovereignty) (42).
Finally, evaluators should continuously consider how their professional and lived experiences might affect what they see and hear, their decisions, and how they engage with others in each context (12,43–45). Doing so can help evaluators “better understand how their own backgrounds and other life experiences serve as assets or limitations in the conduct of an evaluation” (12) and bring awareness to who or what they might inadvertently pay the most attention to in various contexts, how they might hear or interpret information shared by others, and generally engage more effectively with interest holders (46).
Learn From and Use Insights
Evaluations are conducted to provide results that inform decision making (2,16,36). Although the focus is often on the final evaluation findings and recommendations to inform action (Step 6), opportunities exist throughout the evaluation to learn about the program and evaluation itself and to use these insights for improvement and decision making (47). Evaluators have an important role in facilitating continuous learning, use of insights, and improvement throughout the evaluation (48,49). By approaching each evaluation with this role in mind, evaluators can enable learning and use from the beginning of evaluation planning. Successful evaluators build relationships, cultivate trust, and model the way for interest holders to see value and utility in evaluation insights. This takes dedication, skill, and perseverance on the part of the evaluator. The many aspects of facilitating learning and use with interest holders include planning for use of findings; sharing, discussing, and interpreting insights during the evaluation process (rather than waiting until the end); and discussing how to implement recommendations. These types of activities can be integrated into the evaluation plan at the beginning of an evaluation process to ensure these opportunities for learning occur (47).
Interest holders can experience changes in thinking and behavior from engaging in the process of planning and implementing an evaluation, including learning from and using insights throughout the process (36). For example, when newcomers to evaluation begin to think evaluatively, fundamental shifts in perspective can occur (50). Evaluation prompts staff to clarify their understanding of program goals, and this clarity might allow staff to function more cohesively as a team with a shared vision of the common endpoint(s). Immersion in the logic, reasoning, and values of evaluation can lead to lasting effects (e.g., basing decisions on data and insights instead of assumptions). As interest holders become more familiar with the evaluation process, they might develop new knowledge about evaluation, acquire additional evaluation skills, and place increased importance on the value of evaluation (30,36).
Changes in attitudes, knowledge, and skill among persons also might translate into changes within the organization, such as developing an infrastructure (e.g., systems, structures, policies, and procedures) that is more supportive of evaluation or enhancing an organization’s evaluation culture as more persons engage in evaluative thinking by regularly questioning program assumptions and asking questions such as what is working, for whom, and under what conditions (36,51,52). Such changes at the individual and organizational levels represent enhancements in evaluation capacity and can be an expressed intention of engaging in a collaborative evaluation process (30,35,51,53,54). The benefits that arise from these and other uses provide further rationale for initiating evaluation activities at the beginning of a program and ensuring evaluators and interest holders learn from and use insights throughout the evaluation.
Evaluators and interest holders also can take the following actions to increase the likelihood that evaluation findings will be useful and used:
- Ensure that interest holders are engaged in the evaluation and their perspectives and needs are understood.
- Consider evaluation use from the beginning of the process.
- Clarify who is likely to use the findings and for what purposes.
- Conduct the evaluation flexibly (i.e., modify methods and approach to align with changing or emerging needs).
- Share findings in time to inform decision making and in a manner that is responsive to the decision-making context (32,36).
Program Evaluation Steps
The framework is composed of six steps that are important to use in any evaluation to improve how evaluations are conceived and conducted. Each step is considered when planning an evaluation and revisited during evaluation implementation. Although the steps are arranged in a linear sequence, all steps are highly interdependent, and might be encountered in a nonlinear sequence. Program evaluation, like these steps suggest, is an iterative process. For example, while formulating evaluation questions (Step 3), evaluators can consider how audiences might act on findings (Step 6) so the answers provided through the evaluation will be more useful. Earlier steps provide the foundation for subsequent progress; however, it is often the case that contextual constraints or nuances are revealed only in later steps (e.g., Steps 4 and 5) and might require revisiting and revising decisions made in earlier steps (e.g., Step 3). Furthermore, the evaluation standards and cross-cutting actions are important to consider in each step; examples of ways to apply both are provided at the end of each step.
Step 1: Assess Context
A first step in the evaluation process is understanding the context in which the program operates (43,55). Contextual details can include various features or components of settings in which evaluation occurs, such as location and environment; persons and their values or beliefs; political, economic, cultural, and historical circumstances; how power and privilege manifest in context; and other underlying factors that contribute to the condition(s) the program was designed to address (12,15,56).
Understanding the context sets the stage for a meaningful, actionable, and culturally responsive evaluation as it provides essential insights for understanding what is most important to know about the program, what constitutes program success, when evaluative insights are needed, and who needs to learn from the insights. In addition, a deeper understanding of the context can increase the validity of evaluations (57,58) and provide essential insights for interpreting the evaluation findings correctly and formulating a feasible action plan. This section describes four factors to consider when assessing context: 1) readiness for evaluation, 2) interest holders, 3) place, and 4) evaluation capacity.
Readiness for evaluation. Before starting an evaluation, consider whether the appropriate conditions exist to conduct an evaluation that is likely to produce relevant, useful, and rigorous insights (59,60). Evaluability assessments are a tool for examining evaluation readiness and might be particularly helpful for public health programs (61). Various factors can be examined through an evaluability assessment, including the extent to which the goals of the program are clearly articulated and potentially attainable through the proposed activities, the clarity of the program theory (e.g., how the proposed activities are envisioned to lead to the desired programmatic results), the availability of appropriate resources to support the proposed evaluation (e.g., funding, staff member availability, and data), and the existing level of interest in the proposed evaluation (43). Multiple factors examined in an evaluability assessment could be reviewed at the start of a program’s lifecycle and addressed to ensure or increase the likelihood of readiness when a program or organization needs to begin an evaluation.
Evaluability assessment findings might indicate that it would be helpful for the program to reach a common understanding of how the program is envisioned to work before engaging in an evaluation. Ensuring consensus on the underlying theory and how the program will proceed from activities to outcomes might create better conditions for describing the program (Step 2), developing clearer evaluation questions (Step 3), and defining what to measure through data collection efforts (Step 4). The assessment might suggest that although persons who are most likely to make use of the evaluation are specifically interested in learning whether the program has produced the intended outcomes, that it is too early in the program’s lifecycle for the intended outcomes to occur. Such a finding would suggest that it is appropriate to refocus the purpose of the evaluation on program implementation rather than outcomes (43). Alternatively, the assessment could show that conditions are optimal and the evaluation can proceed.
Interest holders. Ensuring an evaluation is relevant and useful requires understanding who is invested in, and might be affected by, the evaluation (16). To increase the likelihood that the perspectives of persons with a broad range of professional and lived experiences are included in the evaluation, it can be helpful to assign persons or groups to the following categories. Persons might be assigned to more than one category.
- Persons served or affected by the program. These interest holders include persons or organizations affected by the program, either directly (e.g., receive services) or indirectly (e.g., benefit from enhanced community assets). They have a vested interest in the evaluation because changes made to the program that are based on the evaluation findings might alter how they experience the program and what benefits they derive.
- Persons who plan or implement the program. These persons have a vested interest in the evaluation because they might need to modify the program based on the evaluation findings. Although the efforts of various groups and persons (e.g., program funders, staff members who engage in day-to-day program operations, and partners who support programming) contribute collectively to program delivery, they are not necessarily a single interest group. Subgroups of persons and organizations involved in program delivery can have different perspectives and might have alternative agendas.
- Persons who might use the evaluation findings. These persons learn from, and act on, the evaluation findings. This is a broad category that includes persons or groups who have the authority to make decisions regarding the program (e.g., can adjust funding allocations or modify program processes) or have a general interest in the results because they design, implement, evaluate, or advocate on behalf of the program being evaluated or similar programs. The primary intended users of the evaluation, those who have a specific interest in the evaluation and clear idea for how to use the findings, should also be considered and highlighted (36). Consider identifying primary intended users early in the development of the evaluation and maintain frequent interactions with them so the evaluation addresses their unique information needs (36) while also integrating the perspectives and values of other categories of persons identified in this section.
- Persons who are skeptical about the program. These persons or groups are skeptical or antagonistic toward the program. Opposition to a program might stem from differing values regarding what change is needed or how to achieve it. When these types of opposition exist, it might be helpful to engage the help of program opponents in the inquiry to strengthen the evaluation’s credibility. These contrasting views also might reveal additional needs and uses for the evaluation.
When assessing the context, it is important to understand the different perspectives or values that the identified persons hold about the program and what might be examined through the evaluation. Persons might view program activities and outcomes differently (Step 2), propose different questions to answer in the evaluation resulting from their unique perspectives (Step 3), value different types of designs and methods (Steps 3 and 4), and differ in their views regarding what constitutes programmatic success and how to interpret (Step 5) and act upon the evaluation findings (Step 6).
Place. Well-designed evaluations recognize, acknowledge, and integrate the uniqueness of the place-based context in which the program and evaluation are conducted. The place dimensions can include program and community history, power dynamics, and the systems and structures that exist, and how these factors intersect with the current-day realities of marginalized communities (28). Aspects of the context that are important to understand, include but are not limited to (43) the following:
- Program features. These features can include why the program was developed, where it operates, how it came into existence, the specific needs the program intends to address, who it intends to serve, how it operates (including the funding mechanisms used), who is directly involved in delivering the program, who has the authority for decision-making, and who and what might influence their decision-making process (36).
- Program environment. Environmental aspects include the current and historical features of the environment in which the program operates (physical and virtual). Specific topics of importance might include the “historical, economic, health, and social dimensions of the communities” (43) and can highlight the strengths and assets (including the talent and expertise) of persons who interact with the program (46). As part of this assessment, it is important to understand how power is distributed among persons who interact with or influence the program, or who might be engaged in the evaluation (e.g., evaluation funders, planners, implementers, and users of findings). For example, understanding whose perspectives have and have not been previously heard or included provides valuable insights for considering how to engage persons in the evaluation. To advance health equity through evaluation, it is also important to understand what health inequities exist within the program environment as well as the drivers of these inequities (14). Such an understanding might inform the role that the program plays in the pathway between drivers and health inequities (Step 2), what elements an evaluation examines (Step 3) and how (Steps 3 and 4), and how the findings from an evaluation are interpreted and acted on (Steps 5 and 6).
Evaluation capacity. Understanding the program’s existing capacity to “do and use” evaluation can be helpful in engaging with persons in a way that takes into account their current understanding and valuing of evaluation (53). In addition, assessing existing capacity can help in identifying the strengths persons and organizations involved in the evaluation might bring to support the planning and implementation of the evaluation, as well as sharing, learning from, and using the findings (36,53,62). Evaluation capacity can be examined at the organizational and individual levels (54,63).
- Organizational. The organization(s) involved in the program (and therefore potentially the evaluation) have an important role in facilitating a high-quality evaluation. The process can benefit from evaluators understanding early in the process an organization’s capacity to support the evaluation so they can leverage existing strengths and, if needed, identifying gaps that might need to be filled to support evaluation activities. Example questions evaluators might ask about existing organizational evaluation capacity to support the conduct and use of evaluations include the following (54,64) (Table 2):
- What is the organizational culture with respect to evaluation and using evaluation findings? Is there support for evaluation generally? To what extent?
- What resources (e.g., funds, staff members, volunteers, time, technology, and data) are available to support planning and implementation of the evaluation? Are there partnering organizations (e.g., community-based organizations and health departments) that are available, willing, and able to support the evaluation?
- Are there internal evaluation champions who can support and effectively encourage engagement in evaluation as well as communicate about the evaluation process and findings?
- What mechanisms already exist to share products from the evaluation with others in the organization who could benefit from the evaluation (e.g., learning forums, online repositories accessible by different organizational units, and workplace collaboration tools)?
- Are there opportunities within the organization to reflect insights that arise throughout the course of the evaluation (e.g., monthly staff meetings)?
- Individual. Interest holders who help plan and implement the evaluation as well as use the findings, might hold assumptions about what evaluation is (or is not) and beliefs about the usefulness of evaluation, which might come from previous experiences with evaluation (36). The process might benefit from inquiring about these past experiences and learning how these have shaped interest holders’ attitudes about evaluation and the extent to which they value the information evaluations can provide.
Understanding these assumptions and beliefs can help evaluators engage in more meaningful conversations about what to anticipate from the evaluation process. Discussing how interest holders understand evaluation can lead to clarity about what questions program evaluation can address and how it is similar to, and different from, other evidence-building functions (e.g., research, performance measurement, and policy analysis). For example, interest holders might have extensive experience with performance measurement or monitoring and incorrectly interpret this activity as synonymous with program evaluation. Clarifying and addressing early in the evaluation planning process how performance measurement and evaluation are related (e.g., the former potentially serving as a data source for the latter) and how the evaluation process might differ (e.g., ability to address questions about program effects with rigor) can reduce misunderstandings and areas of confusion (65).
Certain persons who are invested in the evaluation might not have extensive evaluation experience; however, they might still bring valuable knowledge and skills to support the evaluation activities. Evaluators should consider the interest holders’ existing evaluation capacity (e.g., knowledge and skills) when bringing them into the evaluation steps and consider opportunity areas throughout the evaluation process for further enhancing individual evaluation capacities.
Implementation considerations. When assessing context, and throughout the other evaluation steps, it is important that evaluators recognize and are responsive to different perspectives, cultures, and approaches (46). Culture is multidimensional and includes “the shared experiences of people, including their languages, values, customs, beliefs, and mores. It also includes worldviews, ways of knowing, and ways of communicating” (12,28). Implementing evaluations that appreciate and account for culture advances equity by answering evaluation questions that matter, by analyzing data collected through ways that are meaningful and understood by those responding, and by interpreting the data within the specific context (28). Evaluators bring their own cultural norms to the evaluation and should engage in self-reflective practices to understand how their own culture might affect what they ask, how they ask it, what they perceive, and whether they might be inadvertently favoring certain voices over others in this evaluation context (12,28,46,66).
Evaluators and interest holders need to consider that the context can change during an evaluation and be prepared to adjust accordingly (36). For example, at the start of an evaluation, the interest holders who are engaged might be very supportive of the evaluation and see value and utility in the potential findings. Later in the evaluation process, interest holders might change, and new interest holders might not be as supportive (or not supportive) of the evaluation. Evaluators should stay appraised of the context so they can readily identify such changes and consider in advance how their approach to collaboration includes engaging new interest holders in the process.
Step 2: Describe the Program
Program descriptions identify the outcomes the program intends to achieve and the key activities that are expected to lead to those outcomes. The program description is the foundation for all subsequent steps in the evaluation, and without this description, it would be challenging to design and implement an effective evaluation. The aim is to produce a program description that is clear and concise with enough detail to facilitate an understanding of the program roadmap. The roadmap often takes the form of a one-page graphic (e.g., logic model, theory of change, or rich picture) that is accompanied by a narrative explanation providing more detailed information.
Collaborating with interest holders can aid in developing a description that is comprehensive and inclusive of different perspectives while bringing clarity and program benefits beyond evaluation planning and implementation. Collaboration also can provide an opportunity for reaching agreement about what the program is doing and aims to achieve, and how the program intends to advance health equity. Evaluations conducted without agreement on key activities and outcomes might be of limited use.
Narrative description. Aspects to consider in a narrative description include:
- Need. A statement of need describes the issue, challenge, or opportunity the program is intending to solve or is contributing to solving. Potential components include the nature and magnitude of the problem or opportunity, which persons or groups are affected and how they are affected (e.g., health disparities between groups), and how the need is changing and in what manner(s) (e.g., disease trends). The statement of need also can reflect existing evidence that informs the understanding of the need, any previous efforts the program has taken to address the issue, and any potentially related factors (e.g., drivers of health inequities).
- Inputs. Inputs are the resources needed for conducting program activities, such as personnel, partners, materials, funding, equipment, data (e.g., surveillance), and the existing evidence base. Descriptions of program inputs need to convey the amount and intensity of program services and highlight areas in which there is a potential mismatch between the activities and resources available to implement those activities. Assumptions regarding base needs for the program (e.g., culturally responsive training curriculum) can be specified as inputs if they are expected to be in place or available at the beginning of the program.
- Activities. The activities identify what the program is implementing to effect change and achieve the intended outcomes. The description can include higher level strategies as well as the activities associated with each. For example, a public health emergency readiness and response strategy might include activities such as developing and implementing readiness and response plans, developing rapid forecasting capabilities, conducting emerging disease surveillance, and coordinating outbreak and emergency response activities.
Well-designed program descriptions clearly identify whether the activities occur sequentially or simultaneously and how activities relate to each other. The description of activities clarifies the program’s hypothesized mechanism or theory of change that explains how the activities are presumed to lead to the intended outcomes (67). Any existing evidence supporting the mechanisms of change can be identified and cited, and evidence gaps are also identified. In addition, it is helpful to describe the intentions regarding what needs to happen with respect to program implementation to achieve the intended outcomes (e.g., essential training and skills of program implementers, program dose and duration, key characteristics of settings where the program is implemented, and characteristics of program participants) (68). In public health, various groups often implement activities with the intention of contributing to shared outcomes, so program descriptions also can clarify which activities are the direct responsibility of the program or related programs or partners. - Outcomes. The outcomes identify who or what is expected to change as a result of the program’s efforts (i.e., program effects). For most programs, outcomes can be temporally sequenced, with shorter term (often more specific) outcomes leading to intermediate then longer-term (broader) outcomes. Many potential sources exist for identifying intended program outcomes, such as a program’s mission and objectives (often shorter or intermediate outcomes), vision and goals (often longer-term outcomes), input from interest holders (e.g., program designers, participants, partners, and funders), social science or other theories (e.g., theory of planned behavior and diffusion of innovation), research, and findings from evaluations of similar programs (69). It is also useful to anticipate and include unintended program consequences or outcomes to the extent that is possible.
- Contextual factors. Public health programs often operate in settings where factors outside the program exist that might affect achieving the desired outcomes. Understanding and accounting for contextual factors that might affect the program’s success is required to design a context-sensitive evaluation to account for and interpret findings accurately. Documenting these observations might be helpful for potential users of the evaluation who wish to transfer learnings from the evaluation to another context. Much of this information will have been identified in Step 1 (Assess context) but can be documented here as part of the program description.
- Stage of development. Programs mature and change; therefore, it is important to consider a program’s stage of development in the evaluation. For example, programs that are in planning stages will differ from those that have operated continuously for a decade. The evaluation purpose and questions posed about a program (Step 3) need to align with the program’s stage of development.
In the planning stage when program activities are untested, and during the implementation stage when program activities are being field-tested and modified, a process evaluation can provide valuable insights about “how the program or service is delivered relative to its intended theory of change, and often includes information on content, quantity, quality, and structure of services provided” (4). Programs in a more mature stage are often well-positioned to examine the extent to which the intended program outcomes have occurred through an outcome evaluation. Programs that assess the causal impact of their activities on outcomes relative to those of a counterfactual (i.e., a condition in which the program does not take place) will conduct an impact evaluation (16).
Logic model or program roadmap. A logic model is a graphic depiction of the relation between a program’s activities and its intended outcomes or effects. It shows the sequence of events for bringing about intended change by synthesizing the main program elements into a roadmap of how the program is supposed to work (rather than how it does work, which could be a question answered through an evaluation). It reflects the underlying theory of change, showing the “if-then” relation between earlier and later activities and the connection between activities and outcomes and earlier to later outcomes (70).
A simple, high-level, one-page logic model can concisely synthesize most of the program elements described above. Logic models can take many forms, though the most common elements include activities and short-term, intermediate, and long-term outcomes (and sometimes inputs) with arrows showing the connections between or among each of these (depicting the underlying theory of change). For example, a logic model was adapted from the CDC Tips from Former Smokers campaign evaluation (Figure 2).
When designed in a collaborative manner, the logic modeling process can improve the level of clarity and agreement among partners about the main strategies and activities and intended program outcomes. During these discussions, persons can work together to clarify the program’s chain of events and identify key assumptions or gaps in the logic of the program’s effects. Ensuring a clear program logic exists without apparent gaps is important to ensuring the program is ready for evaluation (60).
Families of logic models or nested logic models can be created to display a program at different levels of detail, from different perspectives, or for different audiences. For example, a high-level logic model could show an entire organization. A second logic model could focus on a specific program or component within the broader logic model, showing the specific activities for that program and aligning these activities with relevant programmatic outcomes. Finally, a third nested logic model could show a specific intervention within the program. Viewed together, the group of logic models can comprehensively show all aspects of a program, which can be useful for program planning and the next steps in the evaluation process.
Although logic models are a useful tool, they are not the only method to visually depict a program (43). The linearity of a logic model might not resonate in all contexts, so it is important that interest holders are engaged early to determine a method that will resonate within their context.
Implementation considerations. Logic models and the narrative that accompanies them are “living documents” and it is important to update them as program changes occur or are anticipated. A few examples of scenarios which might require logic model revisions include changes in context (e.g., pandemic or policy changes that affect the program) or inputs (e.g., unanticipated resource needs during a pandemic), advancements in knowledge or practice from research (e.g., new research suggesting proposed connections between activities and outcomes would not operate as depicted), or evaluations of the program (e.g., modifications to program activities are made in response to an evaluation’s findings).
Sometimes inconsistencies arise among collaborators in how program activities or outcomes are described and the envisioned connections among them. In addition, apparent gaps or inconsistencies in the program logic might arise such as when a particular dose or level of an activity described is unlikely to effect change in an outcome or an outcome in the presumed causal pathway is skipped (e.g., knowledge or skill development is not depicted before a behavior change that would rely on this). Often these inconsistencies can be overcome through further discussion among the interest holders or through further examination of existing evidence. Occasionally, these issues point to an underlying lack of clarity in the program logic suggesting that the program might need more design work before being ready for an evaluation (60,61); in this case, the issues need to be resolved before moving forward with the evaluation.
Collaboratively engaging with interest holders on understanding and describing the program can take various forms and could include a series of meetings to discuss and reach consensus on program descriptions, methods for receiving feedback, and working to facilitate a clear understanding and consensus. To advance equity through evaluation, these discussions also can include conversations about whether (or how) the activities advance equity, whether the program has (or could have) any long-term intended health equity outcomes, and the pathway to achieve those outcomes.
Step 3: Focus the Evaluation Questions and Design
After the program has been described, the next step is to determine on which parts of the program to focus the evaluation efforts. Because most parts of a program logic model or roadmap can be evaluated, this step prioritizes the information needs for learning and use. The main products of this step are the following:
- Purpose statement explaining why the evaluation is being performed, how the findings are likely to be used, and who is likely to learn from or use the findings,
- Statement about the type of evaluation that will be conducted (e.g., process and outcome),
- List of intended users and use of the evaluation findings,
- List of evaluation questions (Table 3), and
- Description of the overarching evaluation design that will be used to answer the evaluation questions.
The objective in this step is to develop collaboratively an optimal, culturally responsive evaluation design that accommodates the program context and available resources, anticipates intended uses, and incorporates all relevant evaluation standards. A well-developed and articulated purpose statement and a clear set of evaluation questions can be referred to throughout the evaluation to help decision-making regarding how the evaluation will be conducted, analyzed, and interpreted.
Flexibility is important when focusing the evaluation because what is learned in Steps 4 and 5 might affect the decisions made in Step 3. For example, perhaps interest holders would like to answer a specific evaluation question, but in Step 4 it is discovered that the data interest holders would find most credible in answering this question cannot feasibly be collected given available resources.
Purpose. Articulating an evaluation’s purpose (i.e., intent and aims) can help to prevent premature decision-making regarding how the evaluation will be conducted and maintain the intended scope of the evaluation efforts. Characteristics of the program, particularly its stage of development and context, will influence the evaluation’s purpose. There are many potential purposes for conducting an evaluation, although they are all aimed at learning about and understanding a program and using the findings for program improvement. Evaluations might have more than one purpose, although it is important to gain clarity with interest holders about the highest priority purpose(s) so the scope of the evaluation does not become too broad.
The evaluation’s purpose helps to focus the evaluation by identifying the most appropriate uses of the evaluation’s findings, the types of evaluation questions likely to be within the evaluation’s scope, and the strength of evidence needed from the evaluation. Some example evaluation purposes include identifying opportunities for improving a program (36,71), providing insights in support of innovation (36,72), examining the program’s effectiveness (36), accountability (i.e., program oversight) (16,36), advancing human rights and social justice (71), and building evaluation capacity (30,35,36) (Table 4).
Evaluation types. A clear evaluation purpose statement provides valuable insights into what type of evaluation is most appropriate to meet the specific information needs given the program’s stage of development. Several types of evaluations are possible and have been defined earlier in this report. They include, but are not limited to, formative evaluation, process or implementation evaluation, outcome evaluation, impact evaluation, and economic evaluation.
Intended users and uses. Users are the specific persons or groups who learn from and act on the evaluation findings. They typically are a subset of interest holders, as not all interest holders will use evaluation findings. Identifying intended users early (during Step 1) and engaging them collaboratively in conversations about what they would like to know about the program, how they intend to use this knowledge, and when they are most likely to use the findings are important considerations in focusing the evaluation.
Evaluation findings can be used in several ways (Table 4). When identifying the intended uses, items to consider include when the findings are needed, the program’s stage of development (e.g., pilot phase, early implementation, or mature program), and the broader context that might affect the ability to use evaluation findings, as assessed in Step 1. All uses are linked to one or more specific users and align directly with the purpose of the evaluation. Stating uses in vague terms that appeal to many persons decreases the chances that the evaluation will fully address anyone’s needs.
Evaluation questions. Evaluation questions tend to be broad in scope, are open-ended, establish the boundaries for the evaluation by stating what aspects of the program will be addressed, and can be answered with the data gathered from the evaluation. Evaluation questions are not the same as data collection questions (e.g., survey, interview, and focus group questions) and typically do not ask about what or how something should be done in the future.
When developing evaluation questions, engage collaboratively with interest holders to identify and prioritize the questions that they would like the evaluation to answer (Table 5). Factors that can be helpful when developing and prioritizing evaluation questions with interest holders include
- evaluation purpose,
- resources available for planning and implementing the evaluation (e.g., funding and staffing),
- timeline for conducting the evaluation,
- when interest holders need information to make decisions,
- how long the program has been in place (e.g., has it been long enough for outcomes to be achieved?),
- availability of similar insights from prior evaluations or other evidence activities, and
- whether answering the question will provide insights for advancing health equity.
The number of evaluation questions should be limited to those that can be answered in a timely manner, with the resources available, and by using appropriately rigorous methods. Formulating and prioritizing evaluation questions should establish the specific aspects of a program to be evaluated. For example, certain persons might want to understand how programs operate together as a system of interventions to effect change within a community, others might have questions concerning the performance of a single program or a local project within a program, and others might want to concentrate on specific subcomponents or processes of a project.
Clear decisions regarding what aspects of a program are within the evaluation’s scope and related to the evaluation purpose will be important in subsequent steps of the evaluation to guide method selection. The final evaluation questions should align clearly with the evaluation purpose and the intended uses of the findings while addressing the information needs of interest holders to the extent possible. Ensuring the evaluation provides insights for funders and implementers, as well as community members who might be affected by the program, is important to make sure that all perspectives are represented in the evaluation aims.
Evaluation design. The evaluation design provides the overarching structure for an evaluation, determining important methodological decisions (e.g., whether comparisons will be made, and if so, what types; and whether sampling will be needed, and if so, what type) (4). No design is better than another under all circumstances. Selection of the appropriate evaluation design is guided by multiple factors, such as the purpose of the evaluation, the evaluation questions, and the evaluation context (e.g., budget, timeline, setting, and responsiveness to interest holder needs). The selected evaluation design should be implemented in a manner that upholds the evaluation standards to the greatest extent possible. Each type of design has strengths and limitations. The choice of design has implications for what is considered evidence, how that evidence will be gathered, and what kinds of claims can be made (including the internal and external validity of conclusions).
Because evaluation is a scientific activity, evaluation design options often have drawn from other scientific disciplines such as the social and behavioral sciences (4). These designs are typically classified as experimental, quasi-experimental, or observational.
- Experimental designs use random assignment to compare the effect of a program on one group with an otherwise equivalent group that did not receive the program.
- Quasi-experimental designs do not use random assignment. Instead, they compare between nonequivalent groups (e.g., program participants versus those on a waiting list) or between a group at different points in time (e.g., time series).
- Observational designs (e.g., case studies and post-test only) are typically considered most useful for evaluations that seek answers to various noncausal questions, such as the fidelity to an original design, quality of implementation, efficiency of the program, or other general operating practices (e.g., process and implementation evaluation) (16).
Experimental and quasi-experimental designs are typically viewed as more rigorous design options (compared with observational) for answering questions about whether a program has contributed to or resulted in the outcomes of interest (e.g., outcome and impact evaluation) because they can often rule out other reasons why an observed outcome occurred (73,74).
As these designs have been applied in various contexts, there has been a need to develop new or modify existing approaches to accommodate nuances that often arise in evaluation contexts (e.g., the complexity of settings in which programs are implemented, the complexity of programs themselves, and contextual constraints and conditions that render certain traditional designs infeasible or inappropriate). As a result, evaluation methodologists have created and implemented various of approaches that include principles for evaluation designs that might better accommodate certain contexts (75). Examples include outcome harvesting (76), process tracing (75), contribution analysis (77), and the success case method (78).
Additional factors to consider when selecting an evaluation design extend beyond ensuring that the selected design aligns directly with the evaluation questions and purpose. One factor is the likelihood of implementing the design as intended. Evaluators should collaborate with interest holders who are familiar with the context to examine whether it will be possible to implement the proposed design in a manner that upholds the underlying scientific principles within the contextual constraints that exist. Examples of factors in the environment that might affect successful implementation of the design include the ability to access and engage participants successfully, the timeline for conducting the evaluation, and the resources available (e.g., funds and staff members) (16). Designs can only be considered rigorous when they are implemented with high quality (i.e., in alignment with underlying scientific principles).
The selected design should be culturally responsive (28). Interest holders might have viewpoints on the credibility of information from specific evaluation designs, which can affect whether they participate in, and act on, the evaluation findings. Evaluators need to consider how their own experiences and background might influence their design preferences. Engaging in reflective practice throughout the evaluation process can help evaluators better understand the viewpoints of others when weighing different design options and ultimately arrive at a design that produces relevant, useful, and rigorous insights in an ethical manner.
Implementation considerations. When implementing an evaluation, the context or data needs might change. For example, the intended use for the evaluation might shift from improving a program’s current activities to determining whether to expand program services to a new population group (36). Interest holders and other persons involved in the process might find that something was overlooked within the context during the planning phase that makes it difficult or impossible to continue with the design as planned and still achieve the evaluation aims (e.g., envisioned participants are no longer available). These types of shifts in the environment require that evaluators are adaptive in their approach (36) and might result in a need to revisit and modify the evaluation design, evaluation questions, and even the purpose.
Certain decisions about the evaluation design can make it more challenging to adjust once the evaluation is in the implementation stage (e.g., the design includes multiple intervention and comparison sites that already have been approved through ethics reviews). These challenges do not indicate that certain designs are to be avoided, rather they emphasize the importance of ensuring that thorough discussions are held during the planning phase with careful consideration to the feasibility of implementing the selected design (Table 6).
Step 4: Gather Credible Evidence
This step builds on the high-level evaluation design (evaluation purpose, evaluation questions, and design) developed in Step 3 to determine the evidence needed to answer the evaluation questions, including what data will be collected, how, when, and from who or what. The product of this step includes a data collection strategy that defines expectations for credible evidence, methods that will be used to ensure data quality, indicators and associated measures of interest, and data sources. Evaluators collaboratively engage with interest holders to consider the credibility of the approaches and the rigor of resulting data and decide what specific data to collect, and how, in response to these information needs (Table 7) (Table 8).
Establish expectations. Establishing expectations involves evaluators and interest holders engaging collaboratively to determine what evidence will be used to answer the evaluation questions; what expectations they have about the type, quality, and quantity of data needed; and what changes, trends, or patterns suggest the program is on track or doing well. These discussions might include the types of evidence that are most valued by different groups (e.g., quantitative and qualitative) and the perceived credibility of data sources. In an outcome evaluation, this also might include discussing which outcomes will be examined and identifying the accountable outcome (i.e., the most distal outcome interest holders expect the program to show progress toward achieving). Establishing these expectations is critical before determining the methods and measures to use in answering the evaluation questions.
Another type of expectation relates to understanding what type and level of results will be used to answer the evaluation questions. For example, an evaluation inquiring about a program’s efficiency will need to demonstrate an understanding of what level of efficiency is considered excellent, good, adequate, or poor. Explicitly identifying these expectations improves transparency and provides a point of reference with which to compare results and see if expectations were met. Various methods can be used to identify these expectations. For example, evaluators and interest holders could examine previous patterns in program data, consult the literature for research and evaluation studies conducted on the same or similar topics, review industry standards, or consult with interest holders who might offer their own perspectives regarding these expectations. Discussing, understanding, and coming to consensus on these expectations can facilitate the use of evaluation findings and create transparency around how the evaluation findings will be interpreted (43).
Cultural norms might vary regarding what constitutes credible evidence (79,80). Understanding these norms, respecting different vantage points, and collaboratively engaging with interest holders to identify how to collect data in a way that is meaningful and useful in that context are important for conducting a high-quality evaluation (13). In addition, interest holders might have different perspectives regarding what constitutes rigorous and credible data. When discussing how to answer the evaluation questions, certain interest holders might value quantitative data and associated statistics, whereas other interest holders might value narratives from qualitative data. An evaluation should strive to collect data that will convey a well-rounded picture of the program in a manner that is rigorous and credible for interest holders. When interest holders are involved in discussing and defining data that will be credible in their contexts, they will be more likely to trust the results and be more invested in the evaluation’s conclusions and recommendations, which will enhance the likelihood that findings will be acted on (Step 6).
Methods. The overall evaluation design was identified in Step 3. During this step, evaluators and interest holders will make decisions about how to gather evaluation data. Each data collection method has strengths and limitations, and no data collection method provides a complete picture. Evaluators and interest holders will need to weigh the pros and cons of methods to arrive at the best approach for the evaluation in the specific context taking into account any associated constraints. Consulting statistics experts might be necessary, such as in situations where there is a need for methods or types of data to make inferences about the program’s success.
Describing the many quantitative (i.e., numeric) or qualitative (i.e., narrative) data collection methods available (e.g., surveys, interviews, focus groups, observations, document and record reviews, and journals or diaries) is beyond the scope of this report. Although quantitative and qualitative methods are often implemented separately, a mixed methods approach in which the two are purposefully integrated can improve the accuracy of the results by compensating for limitations in one method with another that is strong in that area or acquiring deeper insights on a topic to improve understanding (e.g., explaining the “why” behind quantitative survey results through follow-up interviews) (43,81).
Ensuring that the evaluation data collection methods are culturally responsive is fundamental to ensuring trustworthy and accurate data (82). Various considerations exist when adapting or developing instruments, such as appropriately matching the instrument type to the context (e.g., not using a written survey in an oral-based culture), ensuring the appropriate reading level is used for written instruments, taking steps to confirm that translations of instruments into different languages are accurate, and allowing respondents to share information in their preferred language (28,83,84).
Although new data might need to be collected to answer the evaluation questions, before committing to gathering new data, evaluators and interest holders might explore whether there are data already available that might be able to answer some or all evaluation questions. For example, public health surveillance, education, census, or other large data sets might be available and analyzed at a much lower cost than primary data collection. These data sources might not perfectly align with the ideal data or participant group, but if they align well with the evaluation question of interest and are trustworthy, the tradeoff might be worth using data that do not involve additional expenditures.
Indicators. Indicators are measurable statements and serve as a bridge between general program constructs or concepts and specific metrics or measures that can be interpreted (70,74,85,86). For example, a construct of social connectedness might have an indicator of the quality of relationships that could be measured by the percent of persons in a community who report having a close bond with at least one person. Checking whether there are metrics with sufficient specificity and detail for accurate data collection is important. The program roadmap or logic model, when used in conjunction with the evaluation questions, is useful for guiding development and decisions related to indicators, which can relate to any part of the program. Common indicator categories include the following:
- Inputs. The resources needed for conducting program activities (e.g., personnel, materials, funding, equipment, surveillance data, partnerships, and existing evidence base).
- Activities. Characteristics or qualities of the program implementation efforts (e.g., screening for disease, delivering an immunization campaign, or conducting training). This category includes outputs, which are the products of program activities (e.g., children screened who meet risk profile, vaccinations administered, or community members who completed training using culturally appropriate curriculum).
- Outcomes. The expected program effects or changes in the short, intermediate, and long term (e.g., increased antitobacco attitudes, increased intentions to exercise, increased immunization rates, and decrease in morbidity and mortality due to breast cancer).
Outcomes can be more challenging to measure than inputs or outputs. In addition, long-term outcomes are often difficult to attribute to a single program, because usually, multiple factors contribute to changes in outcomes. As a result, it is sometimes tempting to focus evaluation questions and the indicators that align with these on inputs, activities, and outputs. However, programs are encouraged to measure outcomes rather than just outputs (if appropriate for the stage of the program and evaluation questions).
To advance equity, it is important to consider whether the proposed indicators will provide valuable information about the drivers of health inequities. Indicators, and the measures associated with them, not only provide insights about what or how much of something is happening, but also for whom and under what conditions. Logic models and discussions with interest holders can provide helpful ideas for indicators to use in responding to the evaluation questions.
Measures associated with indicators might be quantitative or qualitative, depending on the evaluation question being answered. For example, understanding why a training did not achieve increases in knowledge or changes in attitudes might be best captured through qualitative data, whereas changes in knowledge might be captured using quantitative measures.
Multiple indicators are often needed for responding to the evaluation questions; however, too many indicators can detract from the evaluation goals, take valuable resources to collect and analyze, and be burdensome for persons or organizations providing the data. Thus, evaluators and interest holders might want to consider that certain indicators will be more time-consuming and costly than others to measure and carefully consider the level of effort associated with each before making a final decision on which to include.
Data sources. Data might be provided from various sources and might include new data collected specifically for the evaluation (primary data) or existing data (secondary data that might be available within the program being evaluated or in external organizations). If possible, using multiple sources provides an opportunity to include different modes and perspectives, potentially enhancing the evaluation’s rigor and credibility. For example, a perspective from inside a program could come from internal documents and interviews with staff or program managers, whereas clients and persons with lived experience, neutral observers, or those who do not support the program might provide a different but equally relevant perspective. Considering these and other perspectives provides a more comprehensive view of the program.
A key decision point related to data sources is whether information needed to answer evaluation questions will be collected from all units of a specific source (e.g., all recipient reports submitted) or a subset (e.g., random sample of all recipient reports). If sampling is needed, the criteria used and rationale for the sampling strategy should be stated clearly to provide information that interest holders can use to interpret the evidence accurately and assess potential biases (16).
Data quantity and quality. When collecting data, consider the quantity needed. Collecting the appropriate amount and types of data to answer the evaluation questions sufficiently (i.e., need to know) is important, as is avoiding the desire to collect data that might be tangential to answering evaluation questions (i.e., nice to know). Balancing the amount of data with the burden (in terms of time commitment and effort) data collection can place on the respondents and others who might be involved in data collection and processing can be challenging. Collaboratively engaging with communities can help ensure the right balance is struck for the specific context.
Data quality refers to the appropriateness and integrity of the data used in an evaluation (58). High-quality data are reliable, valid, authentic, and informative for their intended use. Well-defined indicators enable easier collection of quality data because they clarify what specific data are viewed as credible and necessary to answer the evaluation questions. Other factors affecting quality include instrument design, data collection procedures, training of data collectors, source selection, coding, data management, data cleaning, and error checking. Obtaining quality data will entail tradeoffs (e.g., breadth versus depth), and discussing the options with interest holders when planning the evaluation can highlight how certain tradeoffs might affect perceived data credibility.
Data collection and context considerations. The timing and infrastructure for collecting, handling, and storing data, and the cultural context need to be considered when making decisions regarding gathering data. Persons providing data should be knowledgeable about their rights; any associated risks; and how the data will be handled, stored, and used, including how privacy and confidentiality will be protected in the process (16).
Persons and organizations might have cultural norms regarding appropriate permissions to engage in data collection, identifying who will collect the data, data governance processes, and acceptable ways of asking questions and collecting data (42). For example, certain participants might be willing to discuss their health behaviors with a stranger, whereas others might be more at ease with someone they know. Working with interest holders on the evaluation data collection procedures can help ensure they align with the project setting, that privacy and confidentiality are protected, and ethical practices are upheld (16).
Persons who provide data for the evaluation contribute time, energy, and knowledge. Without their involvement, evaluations would not be possible. In developing a data collection plan, discuss how persons will be acknowledged for their important contributions (38). Collecting data that are not extractive is important, and evaluators need to work with the project team to share data with communities and respect data sovereignty.
Implementation considerations. Despite the best planning, data collection challenges are common once an evaluation has commenced. For example, even though evaluators planned for and addressed data concerns in advance, once the evaluation starts, persons or organizations might have unease and concern about providing access to data sources or might answer questions in a more favorable light than truly represents a situation for fear that the evaluation will show inadequacies in the program and consequences will result (e.g., less funding and program cancellation). Such reactions are demonstrations of evaluation anxiety (87). Engaging collaboratively with those invested in or affected by the evaluation early in the evaluation planning process, and often throughout the evaluation implementation, can help to establish trusting relationships and reduce concerns that might stem from the unknown.
Data collection plans might change once implementation begins. Access to certain data might not be available as originally anticipated, response rates might be lower than necessary, and information needs might change (e.g., evaluation findings are needed sooner than originally expected because of a change in circumstance) requiring a modification to the data collection approach. Such changes need to be documented throughout an evaluation so that persons making use of the findings can make a well-informed decision about the quality and trustworthiness of the work performed. Documenting modifications provides a level of transparency required of high-quality evaluations (13,16,26).
Step 5: Generate and Support Conclusions
Activities in Step 5 focus on generating answers to the evaluation questions (Table 9). These answers are presented as evaluation conclusions, align with the evaluation questions (Step 3), and demonstrate how the conclusions are supported by the data collected (Step 4). This step involves reviewing the evidence expectations identified previously, conducting robust data analysis, interpreting findings, and developing recommendations. Engaging interest holders in the process of interpreting analyses, drawing evaluative conclusions, and testing the feasibility of potential recommendations will help ensure the conclusions and suggested actions are tied to the underlying data and responsive to the context.
Analysis. The plan for analyzing evaluation data should be established in advance of implementing the evaluation and use the most robust methods possible to answer the evaluation questions. Developing an analysis plan before data collection will increase the likelihood that data collection instruments include questions necessary to acquire the data needed to produce measures aligned with the indicators established in Step 4.
Whether conducting an analysis of quantitative, qualitative, or both types of data, each type of analysis has established procedures for upholding rigor and objectivity and considerations for protecting privacy and confidentiality that should be followed. Identifying and describing the multitude of analytic methods available is beyond the scope of this framework. Regardless, decisions about which analytic approach(es) to use need to be guided by the evaluation questions and characteristics of the data collected. As noted in Step 4, involving statistics experts might be necessary for analyses and interpretation, particularly for complex analyses, as incorrect or inappropriate analysis or interpretation can lead to false claims and potentially result in decreased trust among interest holders.
The decisions in Step 4 regarding measures and sources will inform the analysis plan and, if there are multiple measures or data sources to answer evaluation questions, describe how to synthesize across them. Having multiple sources and methods can help in interpretation, drawing conclusions, and making recommendations because they provide more information to learn from than singular data sources.
Engaging collaboratively with interest holders regarding the types of analysis that will be needed is important. Certain evaluation questions might be addressed through descriptive analyses, whereas others might require more advanced analyses. Discussions with interest holders also should include how to incorporate analyses that might contribute valuable insights for advancing health equity.
Interpretation. Simply reporting analytic results is insufficient to draw evaluative conclusions. Results of data analyses are compared with the expectations identified earlier (Step 4) and interpreted within context (Step 1) to determine the practical application and implications of what has been learned. In Step 5, evaluators and interest holders work together to translate what the findings mean, identifying existing strengths, successes, and areas for improvement including opportunities to advance health equity (Box 2). Engaging collaboratively to interpret the findings has multiple benefits including producing a more robust understanding of the findings and their implications and enhancing interest holders’ receptivity and commitment to learning from and using the evaluation findings.
Where an existing evidence base exists, evaluation conclusions can be further strengthened by interpreting the analytic findings within the context of this evidence base. Furthermore, scientific theories or models (e.g., theory of planned behavior or diffusion of innovation) identified in earlier steps or in existing literature also might be used to explain findings.
Recommendations. Recommendations are actions for consideration resulting from the evaluation and can suggest how improvements could be made and how existing successes and strengths can be leveraged (88). Similar to the evaluative conclusions formed when interpreting findings, recommendations also are rooted in the evaluation findings and need to be supported by the evidence.
When formulating recommendations, the broader framework steps should be considered, particularly Step 6 (Act on findings). Using insights to create actionable recommendations is critical to creating meaningful program changes. Recommendations are more actionable when they are clearly worded, provide multiple potential options for action, are in alignment with potential users’ roles and responsibilities, and are feasible to implement in context (i.e., can be implemented within fiscal, time, staff, and other constraints) (88).
Engaging collaboratively with persons who are aware of the context and potential constraints and opportunities within the environment(s) in which recommendations will be implemented can be done while formulating potential recommendations so evaluators can learn about the potential feasibility of acting on them and adjust accordingly. To further increase the likelihood of facilitating action, evaluators might consider limiting the number of recommendations and prioritizing them (88) and providing suggestions for who could be responsible for taking action, on what timeline, and how the implementation of actions might be monitored (if known and appropriate for the context).
Implementation considerations. All evaluations have strengths and limitations. Ensuring that both are articulated alongside the analysis approaches, interpretations, and recommendations is important for transparency. Much of the time spent on an evaluation is often allocated to earlier steps in the framework. Implementation of an evaluation plan in large part includes interpreting and understanding the findings resulting from data collection and analysis and using that information for recommendations and acting on findings (Figure 3). It is important that sufficient time is allocated for synthesis and working with interest holders on interpretation and recommendations (43).
When working with interest holders who might not be as familiar with analysis and interpretation methods, it might be helpful to engage in ways that clearly and plainly summarize the procedures and findings, including strengths and limitations, to ensure the connection between findings and data are transparent and clear. This is also an opportunity for a collaborative approach to understanding and interpreting the meanings of the findings and to hear from interest holders who might have a different perspective or interpretation. Understanding and incorporating these perspectives into the products will improve the likelihood that the results and recommendations will accurately represent the context and be accepted and used by interest holders.
Step 6: Act on Findings
Evaluation findings, recommendations, and lessons learned are crucial for improving programs; however, they do not automatically translate into action for informed decision-making. Using evaluation data and insights remains an elusive goal for many organizations. To ensure evaluation insights are used requires early planning, collaboration, and commitment from the evaluator and all interest holders to act on the findings and recommendations. This step is an essential element in the evaluation cycle and is important not to overlook (Table 10). Key elements for acting on the findings of an evaluation include planning, preparing findings for use, and facilitating insights to action.
Planning. Much of the planning for acting on the findings and recommendations has been discussed in previous steps of the framework. For example, in Steps 3 and 4, collaborative engagement with interest holders about evaluation questions, methods, and indicators has included how they plan to use the findings and recommendations.
Evaluation planning needs to begin with the end in mind (89), asking questions about who will use the evaluation insights, what their needs are, how and when they intend to use the evaluation insights, what potential uses exist (beyond those already anticipated), and how to best facilitate and promote the use of findings and recommendations (36).
A well-crafted strategy that lays out a plan for who will use the findings, when, and how to facilitate learning and use is beneficial for facilitating action (48). Furthermore, the process of creating this strategy will highlight how interest holders can enhance the relevance, credibility, and overall utility of the evaluation. When designing this strategy, consider who could learn from the findings, how the findings can be provided quickly and often, what ways of presenting the insights will resonate most given the context, and what opportunities exist to facilitate decision-making conversations using evaluation insights.
Preparing findings for use. Preparing findings for use refers to the steps needed to ready the insights and recommendations for dissemination and action. Various methodologies and frameworks offer evaluators a structured approach to move from the evaluation findings to learning and use (e.g., knowledge to action and data to action) (90–92), including guiding users through how potential findings (including negative findings) might affect decision-making. This can prepare interest holders for eventually using the evidence and make space for identifying options for program improvement. Evaluators can consider ways to prime users for uptake and learning by considering issues such as how users will receive and understand the findings; how users can apply the insights in their work; and how the evaluator can help persons to use and understand the findings.
Preparation also can include various ways to disseminate evaluation findings to all interest holders in a timely, unbiased, and consistent fashion (Box 3). In addition, well-designed plans include opportunities for dialogue about how to use the insights and implement recommendations. Interest holder communication and feedback are an integral part of evaluation, particularly for learning from and using evaluation findings.
Dissemination is not the final act of the evaluation; it is a cycle that evaluators conduct regularly. Dissemination can take multiple forms at each stage of an evaluation. During the evaluation, these include in-process data sharing, user check-ins, and feedback sessions with interest holders. Sharing findings early and often, even when data analysis is still in-process, and seeking input from users create an atmosphere of trust. This communication also can keep an evaluation on course by keeping those involved informed regarding how the evaluation is proceeding and how to make ongoing adjustments to the program. Evaluators can hold periodic discussions during each step of the evaluation process and routinely share interim findings, provisional interpretations, draft highlights, lessons learned, and promising practices (89).
Although evaluation documentation is needed, a formal evaluation report is often not the most critical product (93,94). As with other evaluation elements, the reporting strategy can be discussed in advance with intended users and other interest holders. Such consultation increases the likelihood that the information needs of relevant audiences will be met. Planning effective communications requires that evaluators consider the timing, style, tone, message source, vehicle, and format of information products. Items to consider specifically when developing evaluation products include tailoring the content for the audience; explaining the program roadmap and focus of the evaluation; and sharing data methods and findings, recommendations, and the evaluation strengths and limitations. Ensuring the use of simple, culturally responsive, and effective data visualization techniques is also important (95–98). Materials need to be clear and communicated in plain language that can be understood by the intended audience(s). Furthermore, the dissemination format, content, and language used should be informed by and responsive to the context and audience and adhere to principles of equitable communication (19,20,99).
Facilitating insights into action. Evaluators take on the role of facilitators when they commit to seeing the evaluation insights used (49). Evaluators help to make sense of and interpret the findings, uncover and apply insights, encourage learning, and lead groups to see ways they can be used to improve the program. Users require support from evaluators when they receive evaluation findings, with each discussion of evaluation findings offering an opportunity for users to engage with the insights. Active facilitation is necessary to guide groups to understand and use the insights as well as new uses that might emerge. Follow-up also might be required to prevent lessons learned from being lost or ignored in the process of making complex or sensitive decisions. To guard against such oversight, the evaluator serves as an advocate for the evaluation findings during the decision-making phase, facilitating understanding of what was discovered and what actions are consistent with the findings.
Facilitating the use of evaluation findings also includes preventing misuse (90). Evaluation results are always bound by the context in which the evaluation was conducted. However, certain interest holders might be tempted to take results out of context or to use them for purposes other than those agreed on (100,101). An example of misinterpretation of results is intentionally selecting certain results that do not reflect the overall analyses and interpretation of the evaluation. Those seeking to undermine a program might misuse results by overemphasizing negative findings without considering the program’s positive results, which is a violation of scientific integrity (81). Evaluators can work to prevent misinterpretations and misuse by collaboratively engaging interest holders who intend to use the findings throughout the evaluation process (101), ensuring that evidence is well-understood and that it is not applied to questions other than those in the evaluation, and that the findings are shared holistically rather than picked to support a particular point of view.
Implementation considerations. Evaluators might have additional opportunities to share information about the evaluation throughout the implementation process as opportunities arise. For situational awareness, evaluators can actively seek out and ask questions of those with whom they are collaborating about innovative ways to engage interest holders (36).
Although discussions regarding how interest holders will make use of the evaluation findings will have occurred in earlier steps, it is important to revisit the planned actions after the evaluation has been implemented because evaluations do not always occur as planned. For example, modifications to data collection procedures (e.g., types of data available, response rates, and sampling) might have changed during implementation and affect how the findings might be best used.
Applying the Framework: Addressing Evaluation Misconceptions
Three common misconceptions regarding program evaluation are clarified by using this framework. First, the perceived cost of and time required for evaluation can deter their use. The cost of an evaluation depends on the questions asked and the level of precision desired for the answers (36,47,48). A simple, low-cost evaluation can deliver valuable results. However, the expense of an evaluation is relative, and it is important to align the investment in evaluation with program needs. Rather than discounting evaluations as time-consuming and tangential to program operations (e.g., left to the end of a program’s project period), the framework encourages conducting evaluations from the beginning that are timed strategically to provide the necessary feedback to guide action. This makes integrating evaluation with program practice possible.
A second misconception centers on the perceived technical demands of designing and conducting an evaluation. Although circumstances exist where controlled environments and elaborate analytic techniques are needed, most public health program evaluations do not require such methods. Instead, the practical approach endorsed by this framework focuses on questions that will improve the program by using context-sensitive methods and analytic techniques that accurately summarize the meaning of quantitative and qualitative information.
Finally, certain program staff might have concerns about evaluation due to perceptions that it is punitive, exclusionary, or adversarial. The framework encourages an evaluation approach that is designed to be helpful and engages all interest holders in a process that welcomes their participation. Penalties to be applied, if any, should not result from discovering negative findings but from failing to use the learning to change for greater effectiveness.
Conclusion
Program evaluation is an essential activity for any organization interested in understanding and improving their programs and services. The systematic development and implementation of a well-conceived and culturally responsive evaluation can provide insights and recommendations that can only be the result of an evaluation inquiry process. The process of learning and using insights as a collaborative endeavor with interest holders can advance health equity and result in benefits beyond the individual evaluation, such as increased evaluative thinking across organizations, where it becomes part of the culture to ask why something is happening as it is and how to continue to learn and improve.
This framework provides a practical approach to actions, steps, and standards to consider when designing and implementing an evaluation. Because the framework is purposefully general, it provides a guide for designing and conducting specific evaluation projects across many different areas. In addition, using this framework does not preclude using other evaluation approaches, tools, or methods, which can be overlayed and used in conjunction with this framework. Thus, this framework is one of multiple tools that organizations can use to improve their programs and activities.
CDC Evaluation Framework Work Group Members
Anita Alston Jones, National Institute for Occupational Safety and Health CDC; Bayo Arthur, Immunization Services Division National Center for Immunization; Respiratory Diseases, CDC; Omoshalewa Bamkole, Division of Workforce Development National Center for State, Tribal Local Territorial Public Health Infrastructure and Workforce CDC; Diana Bartlett, Office of Science CDC; Mohamed Bouras, Office of the Director National Center for Chronic Disease Prevention; Health Promotion, CDC; Christopher J. Cappelli, Division of Global HIV & TB, Global Health Center CDC; Denise C. Carty, Office of Women’s Health CDC; Jessica Crowell, Division of Violence Prevention, National Center for Injury Prevention and Control CDC; Shivani Dama, Division of STD Prevention National Center for HIV; Viral Hepatitis, STD; Tuberculosis Prevention, CDC; Jina Dcruz, Office of the Director National Center for Immunization; Respiratory Diseases, CDC; Dora Ducak, Division of Violence Prevention, National Center for Injury Prevention and Control CDC; Tambra Dunams, Division of Emergency Operations, Office of Readiness and Response CDC; Arlene Edwards, Division of HIV Prevention National Center for HIV; Viral Hepatitis, STD; Tuberculosis Prevention, CDC; Oluwayomi Fabayo, Division of Diabetes Translation National Center for Chronic Disease Prevention; Health Promotion, CDC; Leah S. Fischer, Division of Infectious Disease Readiness and Innovation National Center for Emerging; Zoonotic Infectious Diseases, CDC; Holly H. Fisher, Division of Viral Hepatitis National Center for HIV; Viral Hepatitis, STD; Tuberculosis Prevention, CDC; Cassandra Martin Frazier, Division of Partnership Support National Center for State, Tribal Local Territorial Public Health Infrastructure and Workforce CDC; Sherese Bleechington Garrett, Division of HIV Prevention National Center for HIV; Viral Hepatitis, STD; Tuberculosis Prevention, CDC; Danique Gigger, Division of Global Health Protection, Global Health Center CDC; Danielle Gilliard Pella, Office of the Director National Center for Emerging; Zoonotic Infectious Diseases, CDC; Jeffrey M. Gordon, Division of Workforce Development Public Health Infrastructure Center CDC; Bradford Greening, Division of Infectious Disease Readiness and Innovation National Center for Emerging; Zoonotic Infectious Diseases, CDC; Jordan D. Helms, Division of Laboratory Systems, Center for Laboratory Systems and Responses CDC; Sara Jacenko, Global Immunization Division, Global Health Center CDC; Jill Kuhn, Office of Integration and Coordination, Global Health Center CDC; Nicole Kuiper, Division of Population Health National Center for Chronic Disease Prevention; Health Promotion, CDC; S. Rene Lavinghouze, Division of Nutrition Physical Activity Obesity National Center for Chronic Disease Prevention; Health Promotion, CDC; Neranga Liyanaarachchige, Division of Global HIV and TB, Global Health Center CDC; Elizabeth Lummus, Division of Overdose Prevention, National Center for Injury Prevention and Control CDC; Brandy L. Peterson; Maddox, Office on Smoking and Health, National Center for Chronic Disease Prevention and Promotion CDC; Angèle Marandet, Division of Global HIV and TB, Global Health Center CDC; Mariette Marano-Lee, Division of HIV Prevention National Center for HIV; Viral Hepatitis, STD; Tuberculosis Prevention, CDC; Brittany Marshall, Division of HIV Prevention National Center for HIV; Viral Hepatitis, STD; Tuberculosis Prevention, CDC; Elizabeth Martin, Office of Equal Employment Opportunity; Workplace Equity, CDC; Caitlin McColloch, Division of Workforce Development National Center for State, Tribal Local Territorial Public Health Infrastructure and Workforce CDC; Susi McGhee, Division of Injury Prevention, National Center for Injury Prevention and Control CDC; Carol Nixon, Spokane Mining Research Division, National Institute for Occupational Safety and Health CDC; Lijing Ouyang, Division of Reproductive Health National Center for Chronic Disease Prevention; Health Promotion, CDC; Jasmine R. Owens, Division of Overdose Prevention, National Center for Injury Prevention and Control CDC; Margaret Paek, Division of Healthcare Quality Promotion National Center for Emerging; Zoonotic Infectious Diseases, CDC; Suchita Patel, Immunization Services Division National Center for Immunization; Respiratory Diseases, CDC; Miriam Phields, Division of HIV Prevention National Center for HIV; Viral Hepatitis, STD; Tuberculosis Prevention, CDC; Shubha Rao, Division of HIV Prevention Division of Viral Hepatitis National Center for HIV; Viral Hepatitis, STD; Tuberculosis Prevention, CDC; Minda D. Reed, Division of Overdose Prevention, National Center for Injury Prevention and Control CDC; Michelle Roth, Division of Foodborne Waterborne; Environmental Diseases, National Center for Emerging; Zoonotic Infectious Diseases, CDC; Maria Sanchez, Division of Human Development and Disability National Center on Birth Defects; Developmental Disabilities, CDC; Delight Satter, Office of the Director National Center for State, Tribal Local Territorial Public Health Infrastructure and Workforce CDC; Arielle Shiver, Division of Nutrition Physical Activity Obesity National Center for Chronic Disease Prevention; Health Promotion, CDC; Kat Sisler, Division of Environmental Health Science and Practice National Center for Environmental Health CDC; Amrita Tailor, Division of HIV Prevention National Center for HIV; Viral Hepatitis, STD; Tuberculosis Prevention, CDC; Affaud Tanon, Global Immunization Division, Global Health Center CDC; Andrea Torres, Division of Cancer Prevention and Control National Center for Chronic Disease Prevention; Health Promotion, CDC; Silvia M. Trigoso, Division of Emergency Operations, Office of Readiness and Response CDC; Ann Ussery-Hall, Division of Environmental Health Science and Practice National Center for Environmental Health CDC; Ijeoma Uzoezie, Division of Infectious Disease Readiness and Innovation National Center for Emerging; Zoonotic Infectious Diseases, CDC; Nga Vuong, Division of Healthcare Quality Promotion National Center for Emerging; Zoonotic Infectious Diseases, CDC; Maureen Wilce, Division of Environmental Health Science and Practice National Center for Environmental Health CDC; Shaunta S. Wright, Division of Jurisdictional Support National Center for State, Tribal Local Territorial Public Health Infrastructure and Workforce CDC; Monique Young, Office of Public Health Practice National Center for Chronic Disease Prevention; Health Promotion, CDC.
Acknowledgments
U.S. Office of Management and Budget; Office of Policy, Performance, and Evaluation, CDC; Office of Tribal Affairs and Strategic Alliances, CDC; Melissa Abelev, U.S. Department of Agriculture Food and Nutrition Service; Jami Anderson, Immunization Services Division, National Center for Immunization and Respiratory Diseases, CDC; Danielle Berman, Office of Management and Budget, Executive Office of the President; Kari Cruz, Division of Overdose Prevention, National Center for Injury Prevention and Control, CDC; Debra Dekker, National Association of County and City Health Officials; Amanda Doreson, Substance Abuse and Mental Health Services Administration; Diana Epstein, Office of Management and Budget, Executive Office of the President; Nick Hart, Data Foundation; Kristen Hudgins, Office of the Assistant Secretary for Planning and Evaluation, U.S. Department of Health and Human Services; Susan Jenkins, U.S. Department of Health and Human Services; Laura Kollar, Division of Overdose Prevention, National Center for Injury Prevention and Control, CDC; Melvin M. Mark, Pennsylvania State University; Kathryn E. Newcomer, The Trachtenberg School of Public Policy and Public Administration; Esther C. Nolton, American University; Joanne Odenkirchen, Office of Tribal Affairs and Strategic Alliances, CDC; Allison Pompey, Centers for Medicare and Medicaid Services; Elizabeth Rudd, U.S. Department of Housing and Urban Development; Erin Smith, Office of Science Quality and Library Services, Office of Science, CDC; Andrea Williams, Public Health Infrastructure Center, CDC; Stacey Willocks, Division of Injury Prevention, National Center for Injury Prevention and Control, CDC; Daniel J. Woytowich, Division of Global Health Protection, Global Health Center, CDC; Seh Welch, Division of Global HIV and TB, Global Health Center, CDC; Erica Zielewski, Office of Management and Budget, Executive Office of the President.
Corresponding author: Daniel P. Kidder, Office of Policy, Performance, and Evaluation, CDC. Telephone: 404-639-6270; Email: dkidder@cdc.gov.
1Office of Policy, Performance, and Evaluation, CDC, Atlanta, Georgia; 2Fierro Consulting, Inc., Thousand Oaks, California; 3Max Bell School of Public Policy, McGill University, Montreal, Quebec, Canada; 4Office of Readiness and Response, Division of State and Local Readiness, CDC, Atlanta, Georgia; 5Division of Diabetes Translation, National Center for Chronic Disease Prevention and Health Promotion, CDC, Atlanta, Georgia; 6Office on Smoking and Health, Center for Chronic Disease Prevention and Health Promotion, CDC, Atlanta, Georgia; 7Office of the Director, National Center for Injury Prevention and Control, CDC, Atlanta, Georgia; 8Office of Data Reporting and Evaluation, National Center for State, Tribal, Local and Territorial Infrastructure and Workforce, CDC, Atlanta, Georgia; 9Office of Health Equity, CDC, Atlanta, Georgia; 10Division of Global HIV and TB, Global Health Center, CDC, Atlanta, Georgia; 11Office of the Director, National Institute for Occupational Safety and Health, CDC, Atlanta, Georgia; 12Division of Environmental Health Science and Practice, National Center for Environmental Health, CDC, Atlanta, Georgia; 13Division of Population Health, National Center for Chronic Disease Prevention and Health Promotion, CDC, Atlanta, Georgia; 14Division of HIV Prevention, National Center for HIV, Viral Hepatitis, STD, and Tuberculosis Prevention, CDC, Atlanta, Georgia; 15Division of Jurisdictional Support, Public Health Infrastructure Center, CDC, Atlanta, Georgia; 16Division of Adolescent and School Health, National Center for Chronic Disease Prevention and Health Promotion, CDC, Atlanta, Georgia; 17Division of Cancer Prevention and Control, National Center for Chronic Disease Prevention and Health Promotion, CDC, Atlanta, Georgia; 18Division for Heart Disease and Stroke Prevention, National Center for Chronic Disease Prevention and Health Promotion, CDC, Atlanta, Georgia; 19Office of the Director, National Center for HIV, Viral Hepatitis, STD, and Tuberculosis Prevention, CDC, Atlanta, Georgia; 20Division of Overdose Prevention, National Center for Injury Prevention and Control, CDC, Atlanta, Georgia
Conflicts of Interest
All authors have completed and submitted the ICMJE form for disclosure of potential conflicts of interest. Leslie A. Fierro reported payment for work on this project as an independent consultant under a subcontract arrangement with the prime contractor GINTECH, LLC. No other potential conflicts of interest were disclosed.
* 45 C.F.R. part 46, 21 C.F.R. part 56; 42 U.S.C. Sect. 241(d); 5 U.S.C. Sect. 552a; 44 U.S.C. Sect. 3501 et seq.
References
- de Beaumont Foundation, Public Health National Center for Innovations. Defining public health practice: 25 years of the 10 essential public health services. Alexandria, VA: Public Health Accreditation Board; 2020. https://phaboard.org/wp-content/uploads/Defining-Public-Health-Practice-25-Years-of-the-10-Essential-Public-Health-Services-1.pdf
- Foundations for Evidence-Based Policymaking Act of 2018. 115th Congress. Pub. L. No. 115–435. Washington, DC: US Government Publishing Office; 2019. https://www.congress.gov/bill/115th-congress/house-bill/4174/text
- Office of Management and Budget. M-19–23. Memorandum on phase 1 implementation of the Foundations for Evidence-Based Policymaking Act of 2018: learning agendas, personnel, and planning guidance. Washington, DC: Office of Management and Budget; 2019. https://www.whitehouse.gov/wp-content/uploads/2019/07/M-19-23.pdf
- Office of Management and Budget. M-21–27. Memorandum on evidence-based policymaking: learning agendas and annual evaluation plans. Washington, DC: Office of Management and Budget; 2021. https://www.whitehouse.gov/wp-content/uploads/2021/06/M-21-27.pdf
- Mark MM, Henry GT, Julnes G. Evaluation: an integrated framework for understanding, guiding, and improving policies and programs. San Francisco, CA: Jossey-Bass; 2000.
- Kidder DP, Chapel TJ. CDC’s program evaluation journey: 1999 to present. Public Health Rep 2018;133:356–9. https://doi.org/10.1177/0033354918778034 PMID:29928844
- Frechtling JA. Logic modeling methods in program evaluation. San Francisco, CA: Jossey-Bass; 2007.
- CDC. Framework for program evaluation in public health. MMWR Recomm Rep 1999;48(No. RR-11):1–40. PMID:10499397
- The Joint Committee on Standards for Educational Evaluation. The program evaluation standards: how to assess evaluations of educational programs. 2nd ed. Thousand Oaks, CA: Sage; 1994.
- Christie CA, Fierro LA. Program evaluation [Volume 1]. In: Baker E, McGaw B, Peterson P, eds. International encyclopedia of education. 3rd ed. Amsterdam, The Netherlands: Elsevier; 2010:706–12.
- CDC. CDC’s Framework for Program Evaluation at 20 years: origins and influence. Atlanta, GA: US Department of Health and Human Services, CDC; 2019. https://intranet.cdc.gov/oppe/evaluation/20thanniversary/CDCs-Framework-Program-Evaluation.pdf
- American Evaluation Association. American Evaluation Association public statement on cultural competence in evaluation. Washington, DC: American Evaluation Association; 2011. https://www.eval.org/About/Competencies-Standards/undefined
- American Evaluation Association. Guiding principles for evaluators. Washington, DC: American Evaluation Association; 2018. https://www.eval.org/About/Guiding-Principles
- Burton DC, Kelly A, Cardo D, et al. Principles of health equity science for public health action. Public Health Rep 2024;139:277–83. https://doi.org/10.1177/00333549231213162 PMID:38044623
- King JA, Stevahn L. Presenting the 2018 AEA evaluator competencies. New Dir Eval 2020;2020:49–61. https://doi.org/10.1002/ev.20435
- Office of Management and Budget. M-20–12. Memorandum on phase 4 implementation of the Foundations for Evidence-Based Policymaking Act of 2018: program evaluation standards and practice. Washington, DC: Office of Management and Budget; 2020. https://www.whitehouse.gov/wp-content/uploads/2020/03/M-20-12.pdf
- CDC. Proposed update to the CDC Framework for Program Evaluation in Public Health. Request for information. Fed Regis 2022 Nov 29;87:73311. https://www.federalregister.gov/documents/2022/11/29/2022-25997/proposed-update-to-the-cdc-framework-for-program-evaluation-in-public-health
- US Department of Health and Human Services. Improving the use of program evaluation for maximum health impact: guidelines and recommendations. Washington, DC: US Department of Health and Human Services; 2012. https://www.cdc.gov/evaluation/materials/FinalCDCEvaluationRecommendations_Formatted_120412.pdf
- Calanan RM, Bonds ME, Bedrosian SR, Laird SK, Satter D, Penman-Aguilar A. CDC’s guiding principles to promote an equity-centered approach to public health communication. Prev Chronic Dis 2023;20:230061. https://doi.org/10.5888/pcd20.230061 PMID:37410939
- CDC. Health equity guiding principles for inclusive communication. Atlanta, GA: US Department of Health and Human Services, CDC; 2023. https://www.cdc.gov/healthcommunication/Health_Equity.html
- Joseph B. 9 Terms to avoid in communications with indigenous peoples. Port Coquitlam, Canada: Indigenous Corporate Training Inc; 2018. https://www.ictinc.ca/blog/9-terms-to-avoid-in-communications-with-indigenous-peoples
- Sharfstein JM. Banishing “stakeholders.” Milbank Q 2016;94:476–9. https://doi.org/10.1111/1468-0009.12208 PMID:27620681
- American Medical Association. Advancing health equity: guide on language, narrative and concepts. Chicago, IL: American Medical Association; 2021. https://www.ama-assn.org/about/ama-center-health-equity/advancing-health-equity-guide-language-narrative-and-concepts-0
- Reed MS, Merkle BG, Cook EJ, et al. Reimagining the language of engagement in a post-stakeholder world. Sustain Sci 2024;19:1481–90. https://doi.org/10.1007/s11625-024-01496-4
- Levin HM, McEwan PJ. Cost effectiveness analysis: methods and applications. 2nd ed. Thousand Oaks, CA: Sage; 2001.
- Yarbrough DB, Shulha LM, Hopson RK, Caruthers FA. The program evaluation standards: a guide for evaluators and evaluation users. 3rd ed. Thousand Oaks, CA: Corwin Press; 2010.
- Office of the Assistant Secretary for Planning and Evaluation. Evaluation policy for the Department of Health and Human Services. Washington, DC: US Department of Health and Human Services, Office of the Assistant Secretary for Planning and Evaluation; 2021. https://aspe.hhs.gov/sites/default/files/migrated_legacy_files//200386/hhs-evaluation-policy.pdf
- Hood S, Hopson R, Kirkhart K. Culturally responsive evaluation [Chapter 12]. In: Newcomer KE, Hatry HP, Wholey JS, eds. Handbook of practical program evaluation. 4th ed. San Francisco, CA: Jossey-Bass; 2015:281–317.
- Mertens DM. Transformative research and evaluation. New York, NY: Guilford Press; 2008.
- Cousins JB, Chouinard JA. Participatory evaluation up close: an integration of research-based knowledge. Greenwich, CT: Information Age Publishing; 2012.
- Brandon PR, Fukunaga LL. The state of the empirical research literature on stakeholder involvement in program evaluation. Am J Eval 2014;35:26–44. https://doi.org/10.1177/1098214013503699
- Searle M, Cooper A, Worthington P, Hughes J, Gokiert R, Poth C. Mapping evaluation use: a scoping review of extant literature (2005–2022). Am J Eval 2024;10982140241234841. https://doi.org/10.1177/10982140241234841
- Mark MM, Allen JB, Goodwin JL. Stakeholder involvement in evaluation: does it affect observers’ perceptions of an evaluation? And which stakeholder group(s) do people think should to participate? Eval Rev 2021;45:166–90. https://doi.org/10.1177/0193841X211055937 PMID:34693763
- Froncek B, Rohmann A. “You get the great feeling that you’re being heard but in the end you realize that things will be done differently and in others’ favor”: an experimental investigation of negative effects of participation in evaluation. Am J Eval 2019;40:19–34. https://doi.org/10.1177/1098214018813447
- Bourgeois I, Lemire ST, Fierro LA, Castleman AM, Cho M. Laying a solid foundation for the next generation of evaluation capacity building: findings from an integrative review. Am J Eval 2023;44:29–49. https://doi.org/10.1177/10982140221106991 PMID:37069836
- Patton MQ, Campbell-Patton CE. Utilization-focused evaluation. 5th ed. Thousand Oaks, CA: Sage; 2021.
- King JA, Stevahn LA. Interactive evaluation practice: mastering the interpersonal dynamics of program evaluation. Thousand Oaks, CA: Sage; 2012.
- Nolton EC, Forsythe LP. Engagement in evaluation [Chapter 12]. In: Newcomer KE, Mumford SW, eds. Research handbook on program evaluation. Northampton, MA: Edward Elgar Publishing Inc; 2024:214–42.
- Equitable Evaluation Initiative. The equitable evaluation framework: May 2023 expansion. Seattle, WA: Equitable Evaluation Initiative; 2023. https://www.equitableeval.org/_files/ugd/21786c_aab47695b0d2476d8de5d32f19bd6df9.pdf
- Office of Science and Data Policy. Equitable evaluation series: principles of equitable communication. Washington, DC: US Department of Health and Human Services, Assistant Secretary for Planning and Evaluation, Office on Science and Data Policy; 2022. https://aspe.hhs.gov/sites/default/files/documents/d4bfa84a6fc3ac13904525c86e1078ee/ees-principles-of-equitable-communication.pdf
- Hacker K, Auerbach J, Ikeda R, Philip C, Houry D. Social determinants of health—an approach taken at CDC. J Public Health Manag Pract 2022;28:589–94. https://doi.org/10.1097/PHH.0000000000001626 PMID:36194813
- Eakins D, Gaffney A, Marum C, Wangmo T, Parker M, Magarati M. Indigenous evaluation toolkit for tribal public health programs: an actionable guide for organizations serving American Indian/Alaska Native communities through opioid prevention programming. Seattle, WA: Seven Directions; 2023. https://assets-global.website-files.com/5d4b3177c03a6439be501a14/63f550f6aca5a76fe89c290a_FINAL_7D_EvalToolKit_FullDoc_022123_WEB_compressed.pdf
- Wilce M, Fierro L, Gill S, et al. Planting the seeds for high-quality program evaluation in public health. Atlanta, GA: US Department of Health and Human Services, CDC, National Center for Environmental Health; 2021. https://www.cdc.gov/asthma/program_eval/PlantingSeeds_eTextbook-508.pdf
- Mezirow J. Fostering critical reflection in adulthood: a guide to transformative and emancipatory learning. Hoboken, NJ: John Wiley & Sons; 1991.
- Mezirow J. An overview on transformative learning [Chapter 3]. In: Crowther J, Sutherland P, eds. Lifelong learning. 1st ed. Milton Park, UK: Taylor & Francis Group; 2008:40–54.
- CDC. Practical strategies for culturally competent evaluation. Atlanta, GA: US Department of Health and Human Services, CDC; 2014. https://www.cdc.gov/national-asthma-control-program/media/pdfs/2024/05/cultural_competence_guide.pdf
- Patton MQ. Utilization-focused evaluation checklist. Kalamazoo, MI: The Evaluation Center, Western Michigan University; 2013. https://wmich.edu/sites/default/files/attachments/u350/2018/ufe-patton.pdf
- Russ-Eft D, Preskill H. Evaluation in organizations: a systematic approach to enhancing learning, performance, and change. 2nd ed. New York, NY: Basic Books; 2009.
- Patton MQ. Facilitating evaluation: principles in practice. Thousand Oaks, CA: Sage; 2017.
- Fierro LA, Codd H, Gill S, Pham PK, Grandjean Targos PT, Wilce M. Evaluative thinking in practice: the National Asthma Control Program. New Dir Eval 2018;2018:49–72. https://doi.org/10.1002/ev.20322 PMID:29950803
- Preskill H, Boyle S. A multidisciplinary model of evaluation capacity building. Am J Eval 2008;29:443–59. https://doi.org/10.1177/1098214008324182
- Vo AT, Archibald T. New directions for evaluative thinking. New Dir Eval 2018;2018:139–47. https://doi.org/10.1002/ev.20317
- Bourgeois I, Cousins JB. Understanding dimensions of organizational evaluation capacity. Am J Eval 2013;34:299–319. https://doi.org/10.1177/1098214013477235
- Fierro LA, Christie CA. Evaluator and program manager perceptions of evaluation capacity and evaluation practice. Am J Eval 2017;38:376–92. https://doi.org/10.1177/1098214016667581
- Greene JC, Boyce AS, Ahn J. Value-engaged, educative evaluation guidebook. Urbana-Champaign, IL: University of Illinois at Urbana-Champaign; 2011. https://comm.eval.org/viewdocument/eval11-session-316
- American Evaluation Association. The 2018 evaluator competencies. Washington, DC: American Evaluation Association; 2018; https://www.eval.org/Portals/0/Docs/AEA%20Evaluator%20Competencies.pdf
- Kirkhart KE. Advancing considerations of culture and validity: honoring the key evaluation checklist [Chapter 9]. In: Donaldson S, ed. The future of evaluation in society: a tribute to Michael Scriven. Greenwich, CT: Information Age Publishing; 2013:129–59.
- Newcomer K, Hart N. Evidence-building and evaluation in government. Thousand Oaks, CA: Sage; 2021.
- Davies R. Evaluability assessment. Melbourne, Australia: BetterEvaluation; 2021. https://www.betterevaluation.org/methods-approaches/themes/evaluability-assessment
- Wholey JS. Exploratory evaluation [Chapter 4]. In: Newcomer K, Hatry HP, Wholey JS, eds. Handbook of practical program evaluation. 4th ed. San Francisco, CA: Jossey-Bass; 2015:88–107.
- Leviton LC, Khan LK, Rog D, Dawkins N, Cotton D. Evaluability assessment to improve public health policies, programs, and practices. Annu Rev Public Health 2010;31:213–33. https://doi.org/10.1146/annurev.publhealth.012809.103625 PMID:20235852
- Cousins JB, Goh SC, Elliott CJ, Bourgeois I. Framing the capacity to do and use evaluation. New Dir Eval 2014;2014:7–23. https://doi.org/10.1002/ev.20076
- Nielsen SB, Lemire S, Bourgeois I, Fierro LA. Capturing evaluation capacity: findings from a mapping of evaluation capacity instruments. Can J Program Eval 2024;38:433–60. https://doi.org/10.3138/cjpe-2024-0001
- Preskill H, Torres RT. Evaluative inquiry for learning in organizations. Thousand Oaks, CA: Sage; 1999.
- Nielsen BS, Hunter DEK. Special issue: performance management and evaluation. New Dir Eval 2013;2013:1–123.
- Symonette H. Culturally responsive evaluation as a resource for helpful-help [Chapter 6]. In: Hood S, Hopson R, Frierson H, eds. Continuing the journey to reposition culture and cultural context in evaluation theory and practice. Greenwich, CT: Information Age Publishing; 2015:109–30.
- Weiss CH. Evaluation. 2nd ed. London, UK: Pearson; 1997.
- Chen HT. Practical program evaluation: theory-driven evaluation and the integrated evaluation perspective. Thousand Oaks, CA: Sage; 2014.
- Donaldson SI. Program theory–driven evaluation science: strategies and applications. London, UK: Routledge; 2007.
- McDavid JC, Huse I, Hawthorn LR. Program evaluation and performance measurement: an introduction to practice. 3rd ed. Thousand Oaks, CA: Sage; 2018.
- Mertens DM, Wilson AT. Program evaluation theory and practice. New York, NY: Guilford Press; 2018.
- Patton MQ. Developmental evaluation: applying complexity concepts to enhance innovation and use. New York, NY: Guilford Press; 2010.
- Fink A. Evaluation fundamentals: insights into program effectiveness, quality, and value. 3rd ed. Thousand Oaks, CA: Sage; 2014.
- Trochim WMK. Research methods knowledge base. Washington, DC: World Bank; 2021. https://conjointly.com/kb/navigating-the-kb/
- Vaessen J, Lemire S, Befani B. Evaluation of international development interventions: an overview of approaches and methods. Washington, DC: World Bank; 2020.
- Wilson-Grau R. Outcome harvesting: principles, steps, and evaluation applications. Greenwich, CT: Information Age Publishing; 2018.
- Mayne J. Contribution analysis: addressing cause and effect [Chapter 3]. In: Forss K, Marra M, Schwartz R, eds. Evaluating the complex: attribution, contribution and beyond. London, UK: Routledge; 2011:53–96.
- Brinkerhoff R. The success case method: find out quickly what’s working and what’s not. Oakland, CA: Berrett Koehler Publishers; 2003.
- Donaldson S, Christie CA, Mark MM. What counts as credible evidence in applied research and evaluation practice? Thousand Oaks, CA: Sage; 2008.
- Executive Office of the President. Guidance for Federal departments and agencies on indigenous knowledge. Washington, DC: Executive Office of the President, Office of Science and Technology Policy; 2022. https://www.whitehouse.gov/wp-content/uploads/2022/12/OSTP-CEQ-IK-Guidance.pdf
- Greene JC. Mixed methods in social inquiry. Hoboken, NJ: John Wiley & Sons; 2007.
- Garibay C, Teasdale RM. Equity and evaluation in informal STEM education. New Dir Eval 2019;2019:87–106. https://doi.org/10.1002/ev.20352
- Ghanbarpour S, Noguez Mercado AP, Palotai A. A language justice framework for culturally responsive and equitable evaluation. New Dir Eval 2020;2020:37–47. https://doi.org/10.1002/ev.20412
- Kirkhart K. Through a cultural lens: reflections on validity and theory in evaluation [Chapter 3]. In: Hood S, Hopson R, Frierson H, eds. The role of culture and cultural context in evaluation: a mandate for inclusion, the discovery of truth and understanding in evaluative theory and practice. Greenwich, CT: Information Age Publishing; 2005:21–38.
- Groves RM, Fowler FJ Jr, Couper MP, Lepkowski JM, Singer E, Tourangeau R. Survey methodology. 2nd ed. Hoboken, NJ: John Wiley & Sons; 2009.
- Davidson EJ. Evaluation methodology basics: the nuts and bolts of sound evaluation. Thousand Oaks, CA: Sage; 2005.
- Donaldson SI, Gooler LE, Scriven M. Strategies for managing evaluation anxiety: toward a psychology of program evaluation. Am J Eval 2002;23:261–73. https://doi.org/10.1177/109821400202300303
- Grob GF. Providing recommendations, suggestions, and options for improvement [Chapter 27]. In: Newcomer KE, Hatry HP, Wholey JS, eds. Handbook of practical program evaluation. 4th ed. San Francisco, CA: Jossey-Bass; 2015:725–38.
- Alkin MC, Vo AT. Evaluation essentials: from A to Z. New York, NY: Guilford Press; 2018.
- Wilson KM, Brady TJ, Lesesne C; NCCDPHP Work Group on Translation. An organizing framework for translation in public health: the Knowledge to Action Framework. Prev Chronic Dis 2011;8:A46. PMID:21324260
- CDC. Applying the knowledge to action (K2A) framework: questions to guide planning. Atlanta, GA: US Department of Health and Human Services, CDC, National Centers for Chronic Disease Prevention and Health Promotion; 2014. https://stacks.cdc.gov/view/cdc/103484
- Zakocs R, Hill JA, Brown P, Wheaton J, Freire KE. The Data-to-Action Framework: a rapid program improvement process. Health Educ Behav 2015;42:471–9. https://doi.org/10.1177/1090198115595010 PMID:26245935
- Johnson J, Hall J, Greene JC, Ahn J. Exploring alternative approaches for presenting evaluation results. Am J Eval 2013;34:486–503. https://doi.org/10.1177/1098214013492995
- Torres RT, Preskill H, Piontek ME. Evaluation strategies for communicating and reporting: enhancing learning in organizations. Thousand Oaks, CA: Sage; 2005.
- Azzam T, Evergreen S. Special issue: data visualization, part 1. New Dir Eval 2013;2013:1–84.
- Azzam T, Evergreen S. Special issue: data visualization, part 2. New Dir Eval 2013;2013:1–84.
- Evergreen SD. Effective data visualization: the right chart for the right data. Thousand Oaks, CA: Sage; 2019.
- O’Connor J, Parman M, Bowman N, Evergreen S. Decolonizing data visualization: a history and future of indigenous data visualization. J Multidiscip Eval 2023;19:62–79. https://doi.org/10.56645/jmde.v19i44.783
- Plain Language Action and Information Network. Federal plain language guidelines. Washington, DC: Plain Language Action and Information Network; 2011. https://www.plainlanguage.gov/media/FederalPLGuidelines.pdf
- Alkin MC, King JA. Definitions of evaluation use and misuse, evaluation influence, and factors affecting use. Am J Eval 2017;38:434–50. https://doi.org/10.1177/1098214017717015
- Patton M. Misuse of evaluations. In: Mathison M, ed. Encyclopedia of evaluation. Thousand Oaks, CA; Sage; 2005; 255–6.
FIGURE 1. CDC Program Evaluation Framework including three cross-cutting actions, six evaluation planning and implementation steps, and five evaluation standards
BOX 1. 2024 CDC Program Evaluation Framework cross-cutting actions, steps, and standards
Cross-cutting actions: incorporate core tenets into each framework step
Steps: describe the general process of evaluation planning and implementation
Standards: guide what constitutes high-quality evaluation
|
FIGURE 2. Example logic model* for Step 2 (Describe the program) of CDC Program Evaluation Framework†,§
* Based on CDC’s Tips From Former Smokers Campaign. R Murphy-Hoefer, CDC, personal communication, August 2024.
† Contextual factors: social determinants of health, U.S. national media infrastructure, cessation support infrastructure, and other media campaigns.
§ Theoretical foundations: theory of reasoned action/planned behavior, health belief model, social learning theory, transtheoretical model of change (stages of change), or diffusion of innovation theory.
BOX 2. 2024 CDC Program Evaluation Framework (Step 5 — Generate and support conclusions): considerations for involving interest holders when interpreting analytic results
Planning for interpretation of results with interest holders
During the collaborative interpretation process
|
Source: Krause, H., and Richburg-Hayes, L. The Data Equity Framework: a concrete and systematic equity-oriented approach to quantitative data projects; 2023. https://doi.org/10.31235/osf.io/sqt4u
FIGURE 3. Sequence of the CDC Program Evaluation Framework informing development* and implementation† of an evaluation plan to generate and support conclusions
* Example evaluation plan outline; might differ depending on program evaluation and context.
† Continued application of cross-cutting actions and evaluation standards. Ongoing consideration of culture and awareness of and adaptation to changing context.
BOX 3. CDC Program Evaluation Framework (Step 6 — Act on findings): potential methods for sharing interim and final results
|
Suggested citation for this article: Kidder DP, Fierro LA, Luna E, et al. CDC Program Evaluation Framework, 2024. MMWR Recomm Rep 2024;73(No. RR-6):1–37. DOI: http://dx.doi.org/10.15585/mmwr.rr7306a1.
MMWR and Morbidity and Mortality Weekly Report are service marks of the U.S. Department of Health and Human Services.
Use of trade names and commercial sources is for identification only and does not imply endorsement by the U.S. Department of
Health and Human Services.
References to non-CDC sites on the Internet are
provided as a service to MMWR readers and do not constitute or imply
endorsement of these organizations or their programs by CDC or the U.S.
Department of Health and Human Services. CDC is not responsible for the content
of pages found at these sites. URL addresses listed in MMWR were current as of
the date of publication.
All HTML versions of MMWR articles are generated from final proofs through an automated process. This conversion might result in character translation or format errors in the HTML version. Users are referred to the electronic PDF version (https://www.cdc.gov/mmwr) and/or the original MMWR paper copy for printable versions of official text, figures, and tables.
Questions or messages regarding errors in formatting should be addressed to mmwrq@cdc.gov.