Original Research ARTICLE
Methods for streamlining intervention fidelity checklists: an example from the Chronic Disease Self-Management Program
- 1Division of Health Systems Management and Policy, School of Public Health, The University of Memphis, Memphis, TN, USA
- 2Department of Health Promotion and Community Health Sciences, Texas A&M Health Science Center, School of Public Health, College Station, TX, USA
- 3Department of Health Promotion and Behavior, College of Public Health, The University of Georgia, Athens, GA, USA
- 4Center for Health Promotion and Disease Prevention, The University of North Carolina at Chapel Hill, Chapel Hill, NC, USA
- 5School of Nursing, Biobehavioral Nursing and Health Systems, The University of Washington, Seattle, WA, USA
Maintaining intervention fidelity should be part of any programmatic quality assurance (QA) plan and is often a licensure requirement. However, fidelity checklists designed by original program developers are often lengthy, which makes compliance difficult once programs become widely disseminated in the field. As a case example, we used Stanford’s original Chronic Disease Self-Management Program (CDSMP) fidelity checklist of 157 items to demonstrate heuristic procedures for generating shorter fidelity checklists. Using an expert consensus approach, we sought feedback from active master trainers registered with the Stanford University Patient Education Research Center about which items were most essential to, and also feasible for, assessing fidelity. We conducted three sequential surveys and one expert group-teleconference call. Three versions of the fidelity checklist were created using different statistical and methodological criteria. In a final group-teleconference call with seven national experts, there was unanimous agreement that all three final versions (e.g., a 34-item version, a 20-item version, and a 12-item version) should be made available because the purpose and resources for administering a checklist might vary from one setting to another. This study highlights the methodology used to generate shorter versions of a fidelity checklist, which has potential to inform future QA efforts for this and other evidence-based programs (EBP) for older adults delivered in community settings. With CDSMP and other EBP, it is important to differentiate between program fidelity as mandated by program developers for licensure, and intervention fidelity tools for providing an “at-a-glance” snapshot of the level of compliance to selected program indicators.
Chronic conditions have received nationwide attention because of their adverse impact on individuals’ daily functioning, social interaction, and self-reported quality of life (1) as well as their association with rising healthcare costs (2). Self-management has been viewed as a key factor enabling patients to deal with the everyday challenges of chronic conditions through medical, emotional, and/or role management (3, 4). Despite national calls for more attention to public health strategies that empower Americans to be more involved in their own health (5), many Americans do not inherently possess the skills for actively engaging in self-management behaviors that can help ameliorate the effects of living with chronic diseases.
Evidence-based programs such as the Chronic Disease Self-Management Program (CDSMP) have gained national and international recognition for helping people with chronic conditions learn self-management skills (6, 7). While positive outcomes of CDSMP have been well documented (8), less is known about the actual implementation processes at the state or national level other than gross indicators of program completion or adherence to recommendations regarding class size (9, 10). Thus, the primary purpose of this project was to address intervention fidelity and describe a methodological approach to streamline a fidelity checklist. As a heuristic example, we used the fidelity checklist contained within the CDSMP Fidelity Toolkit (11). A secondary purpose of this project was to use this methodological process to evoke expert opinions about how leaders in the aging services field view the fidelity and quality assurance (QA) processes. As a note, we will consistently use “intervention” fidelity throughout the current study because we focus on a fidelity checklist assessing processes beyond the program itself (e.g., training, before the program, after the program, and evaluation).
Public Health Initiatives and Evaluations
The U.S. Administration on Aging (AoA) Evidence-Based Disease Prevention Grant Programs, initiated in 2003, have stimulated the development and implementation of evidence-based programs (EBP) for seniors, which dispel earlier myths that health promotion efforts were futile in older populations (12). With this greater national appreciation for the potential of evidence-based health promotion programing for improving health and functioning among older adults, the research questions have shifted from “do we know what works?” to “can we do what is known to work?” This change in focus is now seen with service providers having ready access to a growing list of EBP, which have been widely tested in community and clinical settings and within the aging services network (13, 14).
Yet, the translation of scientifically tested research findings to community-based programs is often slow, fragmented, and subject to speculation by the practitioner community (15). As more EBP are offered by diverse host agencies in more diverse communities, evidence is mounting that their successful dissemination occurs sporadically (15–18). Translational research is coming of age, and models such as RE-AIM are being formulated to serve as guiding frameworks for planning implementation efforts and evaluating the public health impacts of EBP (19–21). More specifically, the RE-AIM framework seeks to identify and overcome the challenges facing program planners and practitioners when moving an EBP from the research setting in which it was developed to the less-than-perfect, resource-limited, and real-world practice environment (19–22).
The RE-AIM framework contains the following five key elements: reach, effectiveness, adoption, implementation, and maintenance (21, 23). Some studies examine all five elements, while others examine outcomes using one or two key elements (16). Our current study focuses on the “I” in RE-AIM, program implementation processes, specifically fidelity monitoring, which can be neglected because of funding and logistic issues in large-scale community-based disease prevention efforts for older adults.
As EBP become widely disseminated, there has been growing attention to program fidelity in implementation science (16). In terms of translational research, there has been a strong programmatic emphasis on fidelity, which can be defined as the adherence of actual treatment delivery to the protocol originally developed (24). A breach in intervention fidelity, defined as the adherent and competent delivery of an intervention by the interventionists (e.g., trainers, course leaders, and program coordinators) as set forth in the intervention manual (25), threatens licensure and makes it difficult to interpret study results. For example, if the program is not delivered as intended, it is difficult to know if the resulting health outcomes can be attributed to receipt of the intervention or to some other variation in the intervention’s delivery (26).
While maintaining fidelity during program implementation is essential, ensuring the feasibility of monitoring fidelity is also important, especially for organizations with relatively limited capacity to administer the intervention (27). Fidelity to treatment or intervention delivery is one subset of overall treatment fidelity (28) and has often been monitored through observation, interviews, self-assessed fidelity checklists, and pairing of trained facilitators (26). Recent articles have highlighted the importance of having high resource commitment to better monitor fidelity in evidence-based health promotion programs such as CDSMP (29) or EnhanceFitness (30).
Strategies for Ensuring CDSMP Program Fidelity
Programmatic adherence to implementation aspects of CDSMP is supported by a centralized training and certification system that provides for scripted small-group and participatory workshops (2.5 h a week for 6 consecutive weeks) focused on self-management strategies that provide medical, emotional, and role management skills (4). In regard to the training and certification system, it is noted that there are three hierarchical levels of trainers (31). First, a person can be a certified (lay) leader when she or he completes 4-day Leader Training and facilitates one 6-week workshop within 12 months from the training date. Second, master trainer certification can be obtained when a person completes 4.5-day Master Training and facilitates two 6-week workshops within 12 months of completion of training. Third, a person can be a T-Trainer when she or he completes 4.5-day apprenticeship under supervision of a Stanford approved certifying T-Trainer and conducts at least one Master Training within 12 months completion. In addition, the program coordinator is another important workforce member who plays a key role in implementing CDSMP. The program coordinator, who may be a master trainer or lay leader, typically engages in a variety of tasks such as: identifying community partners, recruiting and supervising workshop leaders and participants, arranging for workshop sites, monitoring intervention fidelity, and evaluating program processes and outcomes (32).
A standardized resource material (e.g., the program guide “Living a Healthy Life with Chronic Conditions” revised in 2012) (4) helps provide general guidance behind the theory and activities. An implementation manual provides more detailed guidance to trainers (33), and a fidelity manual outlines mandatory program requirements. The 2010 CDSMP Fidelity Toolkit (11) contains a fidelity checklist with key aspects listed in the following link: http://patienteducation.stanford.edu/licensing/Fidelity_ToolKit2010.pdf. This fidelity checklist as part of the Toolkit is called “Must Do’s Fidelity Checklist” and provides guidance for personnel with regard to the implementation of CDSMP (e.g., program coordinators, leaders, master trainers, or T-trainers). Personnel are advised to go through the list and check “Yes” for all the items they are currently doing, and are encouraged to incorporate these items with their fidelity plan for the future if they are not able to implement the entire fidelity task right away. These 157 items are chronologically categorized under 7 headings (16 subheadings or blocks): (1) personnel; (2) delivery before training; (3) fidelity during training; (4) fidelity after training; (5) fidelity during workshops; (6) fidelity for leaders and master trainer retention; and (7) fidelity after workshops. Each heading was further divided into a couple of subheadings.
Implementing EBP can require detailed monitoring and tracking information, placing substantial administrative burdens on program coordinators (15). As such, shorter fidelity tools were developed by some states implementing CDSMP including Missouri1 and New Jersey2. However, none of these fidelity tools was systematically tested. In our role as technical advisors to the AoA Evidence-Based Disease Prevention Grant Programs, we were asked to explore methods for streamlining a fidelity checklist (e.g., CDSMP) and use these methodological processes to seek expert opinions about how leaders in the aging services field view the fidelity and QA processes.
Materials and Methods
The expert consensus method refers to a multi-phase approach for statistically analyzing pooled opinions that minimizes biases inherent in other systems of summarizing expert viewpoints (34, 35). We used this method to gather and analyze expert opinion on streamlining the 157-item CDSMP fidelity checklist in 2010–2011. We employed three rounds of data collection using Qualtrics software (36) to streamline the original 157-item CDSMP fidelity checklist without losing essential fidelity items but improving feasibility of administration. As a final effort, we held a telephone conference to solicit expert opinion for making final recommendations regarding the use of fidelity checklists. Table 1 displays these four rounds of checklist streamlining and the number of items remaining after each round. Human Subjects approval was obtained from the Texas A&M University Institutional Review Board.
As the credentialing unit, the Stanford University Patient Education Research Center compiles a list of all CDSMP master trainers and manages a listserv for exchange of information. Using this distribution list, we invited all active master trainers in 2010 to participate in our study. In Round One, we collected 119 responses (114 master trainers, 5 others), where 5 responses were eliminated from the final analysis since they were not master trainers. Twenty-six out of 114 were both master trainers and program coordinators. Out of these 114 master trainers, 47 agreed to participate in the second survey. In Round Two, 24 out of 47 master trainers responded to the survey (51% response rate). From this group, nine master trainers were willing to respond to the third survey. In Round Three, seven out of nine master trainers responded to the third survey (78% response rate). In Round Four, eight out of nine master trainers from Round Three agreed to participate in a telephone conference to obtain feedback, advice, and concerns from their experiences delivering and overseeing programs. In this last round, seven of out of eight experts were able to join the group-teleconference call with the study team. Each of the four rounds to streamline the CDSMP fidelity checklist contained multiple steps as shown below.
Round One: Streamlining the Original CSDMP by “Ensuring” the Overall Fidelity of CDSMP
The survey in Round One was conducted in November and December 2010 with master trainers (n = 114) who identified through the Stanford CDSMP master trainer listserv. These experts helped us identify the items in the original 157-item CDSMP fidelity checklist, which they believed to be most important to ensuring the overall fidelity of the program. Participants were asked to rank statements on a scale of one (least important) to five (most important). During this initial step, we eliminated 8 items related to master trainer training and another item referring to language literacy, which resulted in an initial portfolio of 148 unique items to consider. To obtain the most relevant information from participants, the survey included skip patterns that presented participants with a list of items most appropriate to them based on their roles (i.e., master trainers or master trainers/program coordinators). Because of their universal relevance, some items were presented to every group.
To streamline the CDSMP fidelity checklist, we used both statistical and practical approaches based upon participants’ ranking of each checklist item. First, the statistical approach involved eliminating items based on their distance from the mean score. Because of testing the statistical significance of multiple comparisons, we used the Bonferroni technique by adjusting the significance level (0.05) to avoid the risk of Type I error (37). Second, the practical version involved selecting only those responses that were rated as four or five in terms of importance. We eliminated any items that “failed” to meet criteria for either the statistical or practical cut-offs. For instance, there were 26 items in a question block (or subheading) asking the importance of fidelity before lay leader training. Relying on the statistical approach, the mean of these items was 4.15, and the adjusted significance level for this question block was 0.0019 (i.e., 0.05÷26). This process helped us identify six items that showed significantly lower importance ratings relative to the mean after adjusting multiple comparisons. In the same question block, the practical approach helped us identify another six items that were not rated as four or five. Of the 12 items identified for elimination, 2 items were mutually exclusive and 5 items were common to both the statistical and practical approaches. Thus, we were able to eliminate seven items that failed to meet the statistical or practical cutoff. In other question blocks (e.g., fidelity during workshops), we continued to use both approaches to eliminate items that failed to meet the statistical criteria or the practical criteria and we were able to delete 32 (22%) out of 148 checklist items. At the conclusion of Round One, we asked for volunteers to complete a second survey used to continue this expert consensus process.
Round Two: Streamlining the Original CDSMP by Ascertaining “Feasibility” of Administration and “Endorsement” of Predetermined Number of Items
The second survey was conducted in June 2011. In this round, we solicited the opinions of the 47 master trainers who agreed to participate after Round One. Round Two involved asking participants to provide feedback via a two-step process that assessed both “feasibility” and “endorsement.” The first step asked master trainers to rank how feasible they think it is to monitor each fidelity checklist item. Feasibility response categories included “not at all,” “somewhat,” and “very high.” Based on the items identified as having “very high” feasibility in the first step, the master trainers were then asked to select a predetermined number of items (total 34) from each question block that they would endorse for inclusion in the final, shorter fidelity checklist versions. We had 24 master trainers respond to our second survey, resulting in a 51% response rate.
Employing the practical methodology based on the feasibility and endorsement responses, we were able to generate a shorter Fidelity-12 checklist (12 items), a medium-length Fidelity-20 checklist (20 items), and a longer version Fidelity-34 checklist (34 items). First, we calculated the percent of participants who endorsed each item relative to the total number of endorsements received for each question block (using a scale from 0 to 100%). We also calculated the percent of participants who endorsed each item received relative to the total number of feasibility ratings of “very high” (using a scale from 0 to 100%). The aggregate of this process was assessed on a combined scale of 0–200%. Fidelity-34 included 34 items that received the highest aggregate numbers of each question block (1, 2, or 3 items based on number of items in each question block). Due to our desire to streamline the checklist further, we selected only items with scores of 90 and 100 as cut-points on the combined 0–200 scale. Fidelity-20 (90 as a cut-off) included 20 items while Fidelity-12 (100 as a cut-off) included 12 items. At the conclusion of Round Two, 9 out of 24 master trainers agreed to participate in the survey in Round Three.
Round Three: Finalizing the Shortened CDSMP Fidelity Checklists
We surveyed nine experts to review the three shortened versions of the fidelity checklist and to help us address some remaining questions related to their perceptions about the most critical items for assessing intervention fidelity using Fidelity-34 as the referent. We had seven experts respond to the short survey, and the response rate was 78%. A report of findings was prepared indicating that Fidelity-34 was the one that experts felt best balanced the inclusion of key fidelity items with the feasibility of administration.
Round Four: Confirming the Necessity of a Shorter CDSMP Fidelity Checklist
We held a telephone conference with seven experts identified in Round Three to report on our findings and solicit expert opinion regarding the use of fidelity checklists. This conference call was critically important for obtaining feedback and advice from those with experience in delivering and overseeing programs. The experts confirmed a preference for Fidelity-34, but felt that the other two shorter versions should be available since the purpose and resources for administering the fidelity checklist might vary from one setting to another.
Expert Consensus Results: Summary Recommendations
Table 2 displays the three versions of the CDSMP fidelity checklist (i.e., Fideity-12, Fidelity-20, and Fidelity-34). To sum up these streamlining processes, we included 14 question blocks out of the original 16 blocks (subheadings) that related to personnel to administer and the chronological tasks of CDSMP. In total, 36 items were included in three versions of fidelity checklist. Among these 36 items, 2 items were only included in the Fidelity-20 (“Have all Leaders facilitate at least once a year,” and “Have trainees participate in two practice teaching activities during training”). Just like Fidelity-34, Fidelity-20 included at least one item of each question block. However, Fidelity-12, as the shortest version of CDSMP fidelity checklist, did not include any items from four question blocks: “Fidelity after lay leader training;” “Fidelity before master training;” “Fidelity when counseling out leaders/master trainers during training;” and “Fidelity during workshops: Physical environment and material resources.” Although Fidelity-34 was highly recommended in Rounds Three and Four, CDSMP experts from the telephone conference also recommended providing organizational partners with all three versions and allowing them to select the best checklist based on their resources. For instance, organizations with time restraints and limited staffing may prefer shorter checklist versions to the longer version, whereas organizations with better staffing and time resources may want to utilize the more thorough version of the checklist.
Using the original CDSMP fidelity checklist as a case example, this research provides a methodology for streamlining fidelity checklists that have many unique items, making field implementation resource-intensive and challenging. We see the methodology described in this paper as our key contribution, which can be applied to different EBPs. It should be noted that these shortened checklists are for research only and require further field-testing before specific endorsements can be made. The CDSMP fidelity checklist has been updated since the time of this study, and a new fidelity manual was developed that stresses the importance of setting intervention fidelity within an overall fidelity plan. These updated materials also distinguish “must do” fidelity strategies from those that are “nice to do” to strengthen program fidelity (33). For the most current requirements, please refer to the Stanford University Patient Education Center website3.
Applying the expert consensus technique, we consolidated the 157-item CDSMP fidelity checklist into three abbreviated versions without sacrificing fidelity items deemed essential by master trainers. Due to its overall length, the original Stanford checklist was often used more as a self-assessment reminder rather than an actual fidelity checklist. Given the importance of program implementation as a core component of program evaluation (21, 23), we believe shorter fidelity checklists will prove beneficial to current and future program leaders and coordinators who are trying to implement EBP with limited financial or time resources.
Though the abbreviated checklists are likely less time-consuming than the original version, further improvements may be needed. First, any checklist should be seen as just one element in an overall QA plan (38). Additionally, one might explore different delivery modalities. For example, enabling users to access fidelity checklists online could further enhance their usefulness. An online monitoring system would allow for real-time review and feedback so that program coordinators and trainers can evaluate their progress and fidelity as implementation is occurring, making changes when necessary to adhere to program guidelines.
In addition to generating three abbreviated versions of CDSMP fidelity checklist, this study also demonstrated an effective use of expert consensus method for generating consensus from a broad field of CDSMP experts with varying experience and perspectives. When dealing with a diversity of opinions, problems may arise due to conflicting viewpoints, self-censorship due to lack of anonymity, incomplete feedback loops or poor communication, or lack of defined statistical methods for attributing quantitative values to subjective factors. Many of these potential pitfalls were avoided by using three iterative rounds of online surveys. Because respondents did not communicate directly, they were free to express their opinions. Statistical and practical methods were used to give each respondent’s perspective equal weight in reaching consensus. In the Round Four telephone conference, expert participants did directly communicate, but because much of the work related to consolidating the list was already concluded, the participants were able to reach consensus easily.
At the end of the process, we were able to identify the most relevant and applicable items and garner experts’ endorsement of the abbreviated checklists as useful. The methods employed in this study could be used as a model for administrators of other EBP aiming to reduce the length of a fidelity checklist for program monitoring. However, it is important to note that CDSMP sites must comply with licensure requirements and be familiar with the official CDSMP Implementation and Fidelity Manuals, which will need to be updated on a regular basis, if programmatic or licensure requirements change.
Several limitations should be considered. First, our relying on small-group processes should be understood for its lack of true representation of the general CDSMP master trainer population. Nevertheless, we believe the current study contributes to the literature related to streamlining a fidelity checklist because of the relative large sample size (n = 114) in Round One survey, and multiple rounds of survey, and a final telephone conference with content experts. Second, our study is potentially limited by focusing on one major stakeholder group, i.e., master trainers. The inclusion of various opinions from CDSMP completers, lay leaders, T-trainers, program developers, and researchers in academia in the Round Three survey could have ensured better representation by various CDSMP implementers and stakeholders. Third, the CDSMP program was revamped after the initiation of this study. While some aspects of the program have been improved upon, we believe the basic elements have remained consistent to ensuring fundamental intervention fidelity based on what is needed before, during, and after workshop delivery. Hence, the need for shortened fidelity checklists is still relevant. Fourth, the study was conducted within a specific time period when federal funding was available for CDSMP implementation. Such changes in the availability of resources and program demands may influence, which fidelity checklist items are most feasible to monitor. However, the ever-changing context and limited funding resources might make three short versions of the CDSMP fidelity checklist helpful and valuable. Fifth, there were noticeable dropouts between streamlining rounds. Since data were not collected from the purposive list of participants, it was not possible to make real comparisons between responders and non-responders. Sixth, we note that generating one intervention fidelity tool may not be the most effective or efficient way to capture all aspects of fidelity, which could be implementation fidelity for leaders or program fidelity for program coordinators and agency administrators. Last, we do not know the level of expertise and knowledge among participated master trainers since we did not collect this information. Nevertheless, we strongly believe master trainers are appropriate experts given the required level of training and experience with delivering CDSMP. To retain their certificates, master trainers are required to conduct 4-day Leader training within 18 months of original training and conduct either a 4-day Leader Training, a 6-week series of community workshops, or a Leader cross training (31).
Fidelity is critical to the successful dissemination of EBP. The challenge for the field is balancing resources for program delivery with those for assuring intervention fidelity. In this study we explore one avenue for reducing the administrative resources needed for maintaining fidelity. The study demonstrates the importance of finding ways to streamline intervention fidelity checklists for EBP, suggesting several key points. First, an expert consensus method is a viable approach to assessing the usability of fidelity checklists. Second, online software (e.g., Qualtrics) can be used for efficient data collection, analyses, and tracking of participant response with built-in reminder systems. Third, it is important to have input from stakeholders with various roles to have a comprehensive picture of intervention fidelity. We encourage researchers to apply this expert consensus model to other EBP, and to conduct further study of the reliability, validity, and practicability of fidelity checklists.
Conflict of Interest Statement
The authors declare that the research was conducted in the absence of any commercial or financial relationships that could be construed as a potential conflict of interest.
This paper is included in the Research Topic, “Evidence-Based Programming for Older Adults.” This Research Topic received partial funding from multiple government and private organizations/agencies; however, the views, findings, and conclusions in these articles are those of the authors and do not necessarily represent the official position of these organizations/agencies. All papers published in the Research Topic received peer review from members of the Frontiers in Public Health (Public Health Education and Promotion section) panel of Review Editors. Because this Research Topic represents work closely associated with a nationwide evidence-based movement in the US, many of the authors and/or Review Editors may have worked together previously in some fashion. Review Editors were purposively selected based on their expertise with evaluation and/or evidence-based programming for older adults. Review Editors were independent of named authors on any given article published in this volume.
This work was supported by the Administration on Aging [90OP0001/03]; and the National Institute of Child Health and Human Development [R01HD047143]. The opinions expressed are those of the authors and do not reflect those of the funding agency. The authors acknowledge the university members of the CDC’s Prevention Research Centers Healthy Aging Research Network (PRC-HAN). The authors also thank Ms. Patti Smith, who helped to review this manuscript.
1. Rothrock NE, Hays RD, Spritzer K, Yount SE, Riley W, Cella D. Relative to the general US population, chronic diseases are associated with poorer health-related quality of life as measured by the patient-reported outcomes measurement information system (PROMIS). J Clin Epidemiol (2010) 63:1195–204. doi:10.1016/j.jclinepi.2010.04.012
4. Lorig K, Holman H, Sobel D, Laurent D. Living a Healthy Life with Chronic Conditions: Self Management of Heart Disease, Arthritis, Diabetes, Asthma, Bronchitis, Emphysema and Others. Boulder, CO: Bull Publishing Company (2006).
7. Ory MG, Smith ML, Patton K, Lorig K, Zenker W, Whitelaw N. Self-management at the tipping point: reaching 100,000 Americans with evidence-based programs. J Am Geriatr Soc (2013) 61:821–3. doi:10.1111/jgs.12239
8. Ory MG, Ahn S, Jiang L, Smith ML, Ritter PL, Whitelaw N, et al. Successes of a national study of the Chronic Disease Self-Management Program: meeting the triple aim of health care reform. Med Care (2013) 51:992–8. doi:10.1097/MLR.0b013e3182a95dd1
9. Smith ML, Ory MG, Belza B, Altpeter M. Personal and delivery site characteristics associated with intervention dosage in an evidence-based fall risk reduction program for older adults. Transl Behav Med (2012) 2:188–98. doi:10.1007/s13142-012-0133-8
10. Smith ML, Hochhalter AK, Cheng Y, Wang S, Ory MG. Programmatic influences on outcomes of an evidence-based fall prevention program for older adults: a translational assessment. Transl Behav Med (2011) 1:384–93. doi:10.1007/s13142-011-0058-7
12. Ory MG, Hoffman M, Hawkins M, Sanner B, Mockenhaupt R. Challenging aging stereotypes: designing and evaluating physical activity programs. Am J Prev Med (2003) 25:164–71. doi:10.1016/S0749-3797(03)00181-8
15. Glasgow RE, Emmons KM. How can we increase translation of research into practice? Types of evidence needed. Annu Rev Public Health (2007) 28:413–33. doi:10.1146/annurev.publhealth.28.021406.144145
17. Jernigan VBB. Community-based participatory research with Native American communities: the Chronic Disease Self-Management Program. Health Promot Pract (2009) 11(6):888–99. doi:10.1177/1524839909333374
18. Green LW, Glasgow RE. Evaluating the relevance, generalization, and applicability of research: issues in external validation and translation methodology. Eval Health Prof (2006) 29:126. doi:10.1177/0163278705284445
22. Ory MG, Altpeter M, Belza B, Helduser J, Zhang C, Smith ML. Perceived utility of the RE-AIM framework for health promotion/disease prevention initiatives for older adults: a case study from the U.S. evidence-based disease prevention initiative. Front Public Health (2015) 2:143. doi:10.3389/fpubh.2014.0014
23. Glasgow RE, Klesges LM, Dzewaltowski DA, Estabrooks PA, Vogt TM. Evaluating the impact of health promotion programs: using the RE-AIM framework to form summary measures for decision making involving complex issues. Health Educ Res (2006) 21:688–94. doi:10.1093/her/cyl081
28. Bellg AJ, Borrelli B, Resnick B, Hecht J, Minicucci DS, Ory M, et al. Enhancing treatment fidelity in health behavior change studies: best practices and recommendations from the NIH behavior change consortium. Health Psychol (2004) 23:443–51. doi:10.1037/0278-6126.96.36.1993
29. Tomioka M, Braun KL, Compton M, Tanoue L. Adapting Stanford’s Chronic Disease Self-Management Program to Hawaii’s multicultural population. Gerontologist (2012) 52:121–32. doi:10.1093/geront/gnr054
33. Stanford University Patient Education Research Center. Program Fidelity Manual: Stanford Self-Management Programs 2012 Update. San Jose, CA: Stanford University Patient Education Research Center (2012).
Keywords: intervention fidelity, quality assurance, Chronic Disease Self-Management Program, aging, evidence-based programs, expert consensus
Citation: Ahn S, Smith ML, Altpeter M, Belza B, Post L and Ory MG (2015) Methods for streamlining intervention fidelity checklists: an example from the Chronic Disease Self-Management Program. Front. Public Health 2:294. doi: 10.3389/fpubh.2014.00294
Received: 19 September 2014; Accepted: 29 December 2014;
Published online: 27 April 2015.
Edited by:Katherine Henrietta Leith, University of South Carolina, USA
Reviewed by:Jay E. Maddock, University of Hawaii at Manoa, USA
Katherine Henrietta Leith, University of South Carolina, USA
Copyright: © 2015 Ahn, Smith, Altpeter, Belza, Post and Ory. This is an open-access article distributed under the terms of the Creative Commons Attribution License (CC BY). The use, distribution or reproduction in other forums is permitted, provided the original author(s) or licensor are credited and that the original publication in this journal is cited, in accordance with accepted academic practice. No use, distribution or reproduction is permitted which does not comply with these terms.
*Correspondence: SangNam Ahn, Division of Health Systems Management and Policy, School of Public Health, The University of Memphis, 133 Robison Hall, Memphis, TN 38152-3530, USA e-mail: email@example.com