Skip to main content

PERSPECTIVE article

Front. Public Health, 17 December 2019
Sec. Public Health Education and Promotion
This article is part of the Research Topic Use of the RE-AIM Framework: Translating Research to Practice with Novel Applications and Emerging Directions View all 21 articles

Evaluating “Take the Stairs, Wyoming!” Through the RE-AIM Framework: Challenges and Opportunities

  • 1Division of Agriculture Cooperative Extension Service, University of Arkansas System, Little Rock, AR, United States
  • 2Center for Quality Aging, Vanderbilt University Medical Center, Nashville, TN, United States

Introduction: Health promotion delivery systems are increasingly being asked to implement policy, systems, and environmental interventions (PSEs). However, evaluating PSEs is challenging, especially in low-resource community settings. This paper describes the use of RE-AIM to evaluate a physical activity PSE delivered through University of Wyoming Extension and highlights challenges and opportunities in pragmatic, real-world program evaluation.

Methods: Extension health educators adapted a point-of-decision prompt intervention encouraging stairway use through posters, called Take the Stairs, Wyoming! Reach was assessed through estimates of daily traffic, effectiveness was assessed through opportunistic interviews, adoption was calculated as the number and proportion of sites that agreed to hang posters, implementation was calculated as the proportion of sites with a poster in place at a 2-weeks follow-up visit, and maintenance was assessed through 6-months opportunistic interviews (individual level) and proportion of sites with a poster in place (organizational level).

Results: Overall, the posters were widely adopted and most posters were implemented as intended. However, capturing reach, effectiveness, and maintenance was challenging, as health educators found the evaluation burdensome. Therefore, it was difficult to determine if the posters were effective at increasing physical activity levels.

Discussion: Suggestions are provided for capturing reach, effectiveness, and maintenance data in community settings. Future efforts are needed to create evaluation tools to pragmatically measure effectiveness of PSEs on changing behaviors, as well as to prioritize program evaluation in Extension.

Introduction

Health promotion delivery systems are increasingly being asked to implement policy, systems, and environmental interventions (PSEs). PSEs, such as creating or improving places for physical activity (1) and providing healthier food and beverages in schools (2), focus on changing the environment to support healthy behaviors. One system tasked with implementing PSEs is the nationwide Land-Grant University Cooperative Extension System (Extension). In Extension, campus-based specialists support county-based educators who deliver programs in agriculture, natural resources, 4-H/youth development, community development, and family and consumer science (3). Within family and consumer science, Extension delivers health promotion programming addressing physical activity (since 2014) (4) and nutrition.

With its roots in home economics and agricultural education, Extension has a long history of implementing individual-level educational programs; however, implementing PSEs is a relatively new focus area. One driver of this change was the 2014 release of Cooperative Extension's National Framework for Health and Wellness, which outlined health promotion efforts based on the social-ecological model that included both “healthy and safe choices” and “healthy and safe environments” and identified PSEs as Extension priorities (5). Another factor is the Healthy, Hunger-Free Kids Act, which was released in 2010 and required the Supplemental Nutrition Assistance Program Education (SNAP-Ed, administered by Extension in some states) to implement comprehensive, multi-level interventions in addition to direct education (6). Lastly, funding opportunities available to Extension (e.g., Centers for Disease Control and Prevention grants) have shifted focus to increasing access to healthier foods and places for physical activity in an effort to create long-lasting health impacts (7).

Implementing PSEs in community settings has the potential for broad impacts on population health (1). However, evaluating PSEs can be challenging, as it is difficult to determine who is influenced by PSEs and track changes in their behavior. Evaluation of health promotion interventions (both PSEs and individual-level interventions) can be especially challenging in low-resource community settings (i.e., those that may not have funding or personnel dedicated to program evaluation) (8, 9). One challenge is that PSEs that were not designed and tested in community settings may include evaluations that are difficult to replicate (e.g., using many hours of observation pre- and post-intervention) (10). Adding to this challenge, interventions that are designed and evaluated in community settings as part of funded, researcher-initiated studies may also be difficult to replicate. Without funded research trials and dedicated evaluation staff, programs may not have the institutional support to be widely adopted and effectively evaluated, and consequently may not achieve the desired results (9).

Another challenge is that existing PSEs evaluation measures often only capture adoption and implementation at the organization level rather than measuring behavior change. For example, the PSEs listed in the SNAP-Ed Toolkit (a repository of practice-tested interventions used in SNAP-Ed) are primarily evaluated through indicators such as organizational-level adoption of nutrition or physical activity supports (11). The SNAP-Ed Evaluation Framework does include individual-level behavior change indicators; however, they are primarily designed to evaluate direct education (11).

While evaluating the impact of PSEs is difficult, it is necessary for stakeholder and funder accountability (12), as well as demonstrating the public value of federally funded programs, like Extension and SNAP-Ed (13). The reach, effectiveness, adoption, implementation, maintenance framework (RE-AIM) has been suggested for robustly evaluating PSEs (14), as well as for planning and evaluating Extension programs (15, 16). RE-AIM has been used for pragmatic program evaluation in community settings (8) and may help practitioners overcome the challenges to evaluating PSEs by providing a comprehensive evaluation framework. The purpose of this paper is to describe the use of RE-AIM as a planning and evaluation framework for a physical activity PSE delivered through University of Wyoming Extension (UWE).

Methods

Setting and Intervention

In Wyoming, five county-based Extension health educators deliver programs in three initiative areas: healthy eating, active living, and food safety; each educator covers multiple counties. Additionally, Cent$ible Nutrition Program (CNP) educators are located in most counties and are federally funded to serve limited-resource audiences through SNAP-Ed and the Expanded Food and Nutrition Education Program (EFNEP). The Extension health educators identified a need for an intervention to increase physical activity levels that was feasible to implement with a small number of Extension health educators covering the state. Collections of evidence-based interventions were searched, and point-of-decision prompts, recommended by the Community Guide (the Community Preventive Services Task Force's list of evidence-based strategies and interventions) (10, 17) were selected. The prompts encourage stairway use through posters to increase physical activity levels (1820). The posters were adapted to give them a more modern look (see Figure 1). UWE program funds were used to print posters for statewide dissemination.

FIGURE 1
www.frontiersin.org

Figure 1. Take the Stairs, Wyoming! posters.

Implementation Process and Research Design

The Extension health educators partnered with CNP educators to implement the intervention, titled Take the Stairs, Wyoming! As there was no database available listing all buildings with elevators in the state, each Extension health educator worked together with the CNP educator(s) in their area to identify businesses and organizations with elevators. Wyoming is a large, rural state with primarily small cities (95% of Wyoming cities have a population under 10,000); this made it possible for educators to identify buildings in their counties with elevators based on local knowledge.

Data were collected through an observational design. Both Extension health educators and CNP educators were asked to approach the identified businesses or organizations in their counties to hang the stairway posters and collect initial data (reach and adoption). After the initial visit, Extension health educators were responsible for completing data collection (effectiveness, implementation, and maintenance) through 2-weeks and 6-months follow-up visits. Educators were asked to implement the intervention between February and August 2018. The University of Wyoming Institutional Review Board approved this study.

RE-AIM Measures

Detailed aims and outcome measures for each RE-AIM dimension are described below and summarized in Table 1. Means and standard deviations of continuous variables and frequencies and proportions of nominal variables were calculated in SPSS (IBM, Version 25).

TABLE 1
www.frontiersin.org

Table 1. RE-AIM dimensions and measures.

Reach

Each business or organization was asked to provide an estimate of daily traffic. For example, this could include average daily patrons at a library.

Effectiveness

Extension health educators conducted opportunistic interviews (i.e., using a convenience sample of all individuals who walked past the poster) (21) for 1 h at each poster site at a 2-weeks follow up visit. The opportunistic interviews consisted of three questions: (1) “Did you see the poster?,” if yes, (2) “Did you feel that your behavior changed in response to the poster?,” and, if yes, (3) “How did the posters change your behavior?” This evaluation measure was selected after reviewing evaluation methods of all the literature that was included in the Community Guide recommendation and was selected as the most feasible. The other studies included in the Community Guide used up to 9 h of pre- and post-implementation observations, which was determined not feasible due to competing demands on Extension health educators' time.

Adoption

Adoption was calculated as the number and proportion of businesses and organizations that agreed to hang the posters.

Implementation

Implementation of the intervention (fidelity of posters implemented) was calculated as the proportion of sites that had a poster in place at the 2-weeks follow up visit to each poster site.

Maintenance

Maintenance was assessed at a 6-months follow up visit to each poster site through opportunistic interviews (individual level) and the proportion of sites that had a poster in place (organizational level).

Results

Eight Extension personnel approached businesses and organizations to place stairway posters across the state: three Extension health educators, four CNP educators, and one campus-based Specialist within the Department of Agriculture and Applied Economics who volunteered to assist. Posters were placed in eight of the state's 23 counties.

Reach

At 38 of the 47 poster sites (81%), the estimated daily traffic was left blank or recorded as unknown, varies, not sure, or not available. Of the nine sites (19%) that did provide estimated daily traffic, an average of 99 (SD ± 127) individuals per site were reached.

Effectiveness

Opportunistic interviews were conducted at 10 poster sites (21%). Across these sites, 42 interviews were conducted. Twenty-four interviewees (57%) responded “yes” to question one indicating that they had seen the poster. Of these twenty-four, eight (33%) responded yes to question two indicating that they felt their behavior had changed in response to the poster. Of those eight, five (63%) responded that they had taken the stairs more often (e.g., “I came in with bags and would have taken the elevator, but saw the sign and took the stairs.”). Three (38%) indicated a change in their thoughts rather than their behavior (e.g., “It made me think twice about taking the elevator.”). Of the 16 who indicated they had seen the posters but their behavior had not changed, 11 provided unsolicited feedback indicating that they already take the stairs (e.g., “I always take the stairs.”) and three indicated that they had thought about changing their behavior (e.g., “I thought about it more seeing the poster.”). See Figure 2 for details.

FIGURE 2
www.frontiersin.org

Figure 2. Results of opportunistic interviews (N = 42).

These data were used to iteratively improve the intervention during the implementation phase (22). For example, when interviewees indicated that they had not seen the posters, this information provided an opportunity to place additional posters in locations that may have been more visible. This can be seen as a real-time adaptation to the intervention.

Adoption

A total of 32 businesses and organizations were asked for approval to hang a poster. Of these, all but two (94%) provided approval, and a total of 44 posters were placed. During 2-weeks follow-up visits to these buildings, three additional posters were placed for a total of 47 posters.

Implementation

Two-weeks follow-up visits were conducted at 42 of the original 44 poster sites; no additional follow-up visits were completed for the three additional posters placed during 2-weeks follow-up visits. Of the 42 follow-up visits completed, 29 posters (69%) were still in place. At two sites where the posters were no longer in place, staff reported the reason (e.g., the elevator was no longer working).

Maintenance

As no follow-up visits were conducted, maintenance was not able to be reported.

Discussion

Overall, the posters were widely adopted by the businesses and organizations that were approached and most posters were implemented as intended (i.e., still in place after 2 weeks). However, capturing reach, effectiveness, and maintenance was challenging. Taken together, these results suggest that the posters may not have been effective at increasing levels of physical activity through increased stair use, but it is difficult to determine due to the limited data. The challenges experienced with data collection as well as suggestions for improvement are presented.

Data Collection Challenges

Determining reach was difficult, as most staff at the participating businesses and organizations were unsure of daily traffic. Additionally, proportion and representativeness (e.g., age, gender) were not captured, so it is unknown if the intervention reached those most in need. While a limitation for impact, this type of barrier is not uncommon in pragmatic settings, as this study actively worked in organizations focused on their own daily activities rather than those specifically recruited for research (23).

As for effectiveness, there were multiple issues with data collection. Three Extension health educators and the Specialist completed the 2-weeks follow-up visits, but only two of them completed the opportunistic interviews. Via email, one who did not complete interviews reported that it was too time consuming and not a good use of her time. Of the two staff members who did complete opportunistic interviews, at four sites they were not completed as they were not able to obtain permission from staff at the business or organization. Additionally, at six of the sites where interviews were conducted, the interview period lasted for less than the prescribed 1 h; one Extension health educator reported that this was due to time constraints when traveling to distant sites. Of the 42 interviews that were completed, only five interviewees reported an actual change in behavior as a result of seeing the posters. Additionally, as the majority of interviewees who had seen the posters indicated that they already take the stairs, the targeted population may not have been reached through the poster intervention. In the future, a follow-up question for those who saw the poster but did not change their behavior may be useful to provide insight into improving effectiveness.

The issues faced in collecting effectiveness data also made collecting maintenance data challenging; if staff experience difficulty with data collection methods at the start of the intervention, it is likely that they will continue to struggle with completing evaluations 6 months or more post-program. No staff completed the prescribed 6-months follow-up visits, so no maintenance data were able to be reported.

Challenges in Evaluation

The barriers experienced in evaluating this PSE—especially effectiveness and maintenance data—are common among community organizations, as they often do not have the means to monitor impacts of PSEs on behavior change (14). Overall, more work is needed to evaluate PSEs in low-resource, community settings. Organizational changes, along with more feasible measures, could improve PSE evaluation in the future.

Need for Organizational Changes

One of the main barriers in this study was the lack of adherence to data collection by Extension health educators. Indeed, Extension struggles with program evaluation; collecting empirical data on behavior change as a program outcome is still relatively novel to the system (8, 9, 24, 25). In the case of PSEs, which are also fairly new to Extension and more difficult to evaluate than direct education interventions, matching evaluation methods to staff resources and expectations is key (Balis et al., under review). While this intervention was selected and planned by a fellow Extension health educator through a participatory approach (2628), the evaluation was still considered a burden. This perception of evaluation as onerous highlights the need to change Extension culture to prioritize time spent evaluating programs rather than only time spent delivering programs. However, part of this burden must still remain on intervention developers to continuously consider the feasibility of the intervention's outcome measures.

Need for Feasible Measures

To improve data collection adherence, feasible measures that are less of a burden on staff need to be available. Intervention developers should consider including pragmatic, low-cost evaluation measures with their interventions for community organization staff to select. For example, with additional funding, infrared people counters or open/close sensors on doors throughout adoption organizations are relatively low-cost solutions that could be used to collect pre- and post-intervention data. These types of measures reduce staff time while providing an estimate of people using stairs and also estimate (if placed at multiple levels) how many flights of stairs individuals will use. Additionally, they would provide an objective measure of physical activity rather than the subjective measure used in this study. These feasible, objective measures need to be tested and, if successful, included in program repositories (e.g., the SNAP-Ed toolkit and evaluation framework) (11) to be used by professionals in community-based organizations. Finally, engaging in partnerships may also reduce evaluation burden. For example, students could complete observations or interviews for research experience; however, this can present an obstacle for Extension interventions that are located throughout the state rather than clustered near campus. Partnering with the organizations and businesses that adopt stairway posters and training their staff to collect effectiveness and maintenance data (e.g., through systematic observations) could also result in better data completion (14). The intervention may have been improved by engaging these stakeholders during the planning process.

There were some limitations to this study, including small sample sizes and incomplete data collection. However, we believe that it is important to include these data in an effort to highlight the reality of real-world program implementation and evaluation. There have been calls from organizations such as the National Institute of Aging (29), funding announcements from the National Institute of Health (30), and commentary pieces from the New England Journal of Medicine (31) that all discuss the various important reasons for conducting pragmatic research. To summarize these points, the real world does not conform to the unrealistic expectations of a randomized-control trial, and while these trials are incredibly important during efficacy testing, it is equally important that intervention are adaptable to real-world uncontrolled settings. The barriers within this study highlight these pragmatic needs.

Overall, RE-AIM was a useful tool for both planning and evaluating this intervention; as recommended, it can also be used after delivery to iteratively refine the intervention (22). For example, for the next iteration, the needs of Extension health educators who did not adhere to data collection procedures can be considered to tailor the evaluation plan to better meet their needs or provide training and technical assistance. Additionally, future iterations could be adapted through RE-AIM to reach businesses and organizations with populations that do not already take the stairs (e.g., through engaging the organizations to complete pre-intervention observations).

Implications for intervention developers include providing PSE evaluation tools that go beyond assessing adoption and implementation and are feasible to use in low-resource community settings. Using pragmatic measures (32) could allow community organizations to confirm effectiveness of PSEs while also collecting data on the other RE-AIM dimensions to ensure these interventions work in the “real world.”

Data Availability Statement

The dataset analyzed during the current study is available from the corresponding author on reasonable request.

Ethics Statement

This study involving human participants was reviewed and approved by University of Wyoming IRB. Written informed consent for participation was not required for this study in accordance with the national legislation and the institutional requirements.

Author Contributions

LB conceived of the study, participated in its design and coordination, and led the manuscript preparation. TS contributed to data analysis and manuscript preparation. All authors read, contributed to, and approved the final manuscript.

Conflict of Interest

The authors declare that the research was conducted in the absence of any commercial or financial relationships that could be construed as a potential conflict of interest.

Acknowledgments

The authors would like to acknowledge Samantha Harden for her assistance with conceptualization.

References

1. Kahn E, Ramsey L, Brownson RC, Heath GW, Howze EH, Powell KE, et al. The effectiveness of interventions to increase physical activity: a systematic review. Am J Prev Med. (2002) 22:73–107. doi: 10.1016/S0749-3797(02)00434-8

PubMed Abstract | CrossRef Full Text | Google Scholar

2. Hanks AS, Just DR, Wansink B. Smarter lunchrooms can address new school lunchroom guidelines and childhood obesity. J Pediatr. (2013) 162:867–9. doi: 10.1016/j.jpeds.2012.12.031

PubMed Abstract | CrossRef Full Text | Google Scholar

3. Franz NK, Townson L. The nature of complex organizations: the case of Cooperative Extension. New Direct Eval. (2008) 2008:5–14. doi: 10.1002/ev.272

CrossRef Full Text | Google Scholar

4. United States Department of Agriculture. The Farm Bill 2015. Washington, DC: United States Department of Agriculture (2015). Available online at: http://www.usda.gov/wps/portal/usda/usdahome?navid=farmbill (accessed November 13, 2016).

Google Scholar

5. Braun B, Bruns K, Cronk L, Kirk Fox L, Koukel S, Le Menestrel S, et al. Cooperative Extension's National Framework for Health and Wellness. Washington, DC: United States Department of Agriculture (2014).

Google Scholar

6. United States Department of Agriculture. Healthy, Hunger-Free Kids Act of 2010. Washington, DC: United States Department of Agriculture (2010) Available online at: https://www.fns.usda.gov/school-meals/healthy-hunger-free-kids-act (accessed September 4, 2019).

Google Scholar

7. Centers for Disease Control and Prevention. High Obesity Program. Washington, DC: Centers for Disease Control and Prevention (2019) Available online at: https://www.cdc.gov/nccdphp/dnpao/state-local-programs/hop-1809/high-obesity-program-1809.html (accessed September 13, 2019).

Google Scholar

8. Balis LE, Strayer TE III, Ramalingam N, Harden SM. Beginning with the end in mind: contextual considerations for scaling-out a community-based intervention. Front Pub Health. (2018) 6:357. doi: 10.3389/fpubh.2018.00357

PubMed Abstract | CrossRef Full Text | Google Scholar

9. Balis L, Harden S. Scaling out a 4-H healthy meeting initiative: challenges in implementation and comprehensive evaluation. J Nutrit Educ Behav. (2019) 51:1020–4. doi: 10.1016/j.jneb.2019.05.288

PubMed Abstract | CrossRef Full Text | Google Scholar

10. Community Preventive Services Task Force. The Community Guide. Washington, DC: Community Preventive Services Task Force (2019). Available online at: https://www.thecommunityguide.org/ (accessed September 4, 2019).

Google Scholar

11. United States Department of Agriculture. SNAP-Ed Toolkit: Obesity Prevention Interventions and Evaluation Framework. Washington, DC (2019). Available online at: https://snapedtoolkit.org/ (accessed September 4, 2019).

Google Scholar

12. Fetsch RJ, MacPhee D, Boyer LK. Evidence-based programming: what is a process an Extension Agent can use to evaluate a program's effectiveness? J Extens. (2012) 50:5FEA2.

Google Scholar

13. Franz N. The role of evaluation in determining the public value of Extension. J Extens. (2014) 52:4COM3.

Google Scholar

14. King D, Glasgow R, Leeman-Castillo B. Reaiming RE-AIM: using the model to plan, implement, and evaluate the effects of environmental change approaches to enhancing population health. Am J Public Health. (2010) 100:2076–84. doi: 10.2105/AJPH.2009.190959

PubMed Abstract | CrossRef Full Text | Google Scholar

15. Balis L, John D, Harden S. Beyond evaluation: using the RE-AIM framework for program planning in Extension. J Extens. (2019) 57:2TOT1.

Google Scholar

16. Downey LH, Peterson DJ, Donaldson J, Hardman A. An extension application of the RE-AIM evaluation framework. J Extens. (2017) 55:3IAW1.

Google Scholar

17. Truman BI, Smith-Akin CK, Hinman AR, Gebbie KM, Brownson R, Novick LF, et al. Developing the guide to community preventive services—overview and rationale. Am J Prev Med. (2000) 18:18–26. doi: 10.1016/S0749-3797(99)00172-5

PubMed Abstract | CrossRef Full Text | Google Scholar

18. Task Force on Community Preventive Services. Recommendations to increase physical activity in communities. Am J Prev Med. (2002) 22:67–72. doi: 10.1016/S0749-3797(02)00433-6

CrossRef Full Text | Google Scholar

19. Task Force on Community Preventive Services. Recommendation for use of point-of-decision prompts to increase stair use in communities. Am J Prev Med. (2010) 38:S290–1. doi: 10.1016/j.amepre.2009.10.031

CrossRef Full Text | Google Scholar

20. Soler RE, Leeks KD, Buchanan LR, Brownson RC, Heath GW, Hopkins DH, et al. Point-of-decision prompts to increase stair use: a systematic review update. Am J Prev Med. (2010) 38:S292–300. doi: 10.1016/j.amepre.2009.10.028

PubMed Abstract | CrossRef Full Text | Google Scholar

21. Adams J, White M. A systematic approach to the development and evaluation of an intervention promoting stair use. Health Educ J. (2002) 61:272–86. doi: 10.1177/001789690206100308

CrossRef Full Text | Google Scholar

22. Harden S, Smith M, Smith-Ray R, Ory M, Estabrooks P, Glasgow R. RE-AIM in clinical, community, and corporate settings: perspectives, strategies, and recommendations to enhance public health impact. Front Public Health. (2018) 6:71. doi: 10.3389/fpubh.2018.00071

PubMed Abstract | CrossRef Full Text | Google Scholar

23. Israel B, Eng E, Schulz A, Parker E. Methods in Community-Based Participatory Research for Health. San Francisco: Jossey-Bass (2005).

Google Scholar

24. Harden SM, Ramalingam N, Breig S, Estabrooks P. Walk this way: our perspectives on challenges and opportunities for Extension statewide walking promotion programs. J Nutrit Educ Behav. (2019) 51:636–43. doi: 10.1016/j.jneb.2018.12.010

CrossRef Full Text | Google Scholar

25. Balis L, Strayer T, Ramalingam N, Wilson M, Harden S. Open-access physical activity programs for older adults: a pragmatic and systematic review. Gerontologist. (2018) 59:e268–78. doi: 10.1093/geront/gnx195

PubMed Abstract | CrossRef Full Text | Google Scholar

26. Harden S, Johnson SB, Almeida F, Estabrooks P. Improving physical activity program adoption using integrated research-practice partnerships: an effectiveness-implementation trial. Trans Behav Med. (2017) 7:28–38. doi: 10.1007/s13142-015-0380-6

PubMed Abstract | CrossRef Full Text | Google Scholar

27. Johnson SB, Harden S, Estabrooks P. Uptake of evidence-based physical activity programs: comparing perceptions of adopters and nonadopters. Trans Behav Med. (2016) 6:629–37. doi: 10.1007/s13142-015-0371-7

PubMed Abstract | CrossRef Full Text | Google Scholar

28. Wilson ML, Strayer T III, Davis R, Harden SM. Use of an integrated research-practice partnership to improve outcomes of a community-based strength-training intervention for older adults: reach and effect of lifelong improvements through fitness together (LIFT). Int J Environ Res Pub Health. (2018) 15:237. doi: 10.3390/ijerph15020237

CrossRef Full Text | Google Scholar

29. Salive M. Pragmatic clinical trials: testing treatments in the real world. In: Inside NIA: A Blog for Researchers. Washington, DC: National Institute on Aging (2017). Available online at: https://www.nia.nih.gov/research/blog/2017/06/pragmatic-clinical-trials-testing-treatments-real-world (accessed September 12, 2019).

Google Scholar

30. National Institutes of Health. Pragmatic Research in Healthcare Settings to Improve Diabetes and Obesity Prevention and Care (R18 Clinical Trial Required). Washington, DC: National Institutes of Health (2019). Available online at: https://grants.nih.gov/grants/guide/pa-files/par-18-925.html (accessed September 12, 2019).

Google Scholar

31. Ford I, Norrie J. Pragmatic trials. N Eng J Med. (2016) 375:454–63. doi: 10.1056/NEJMra1510059

PubMed Abstract | CrossRef Full Text | Google Scholar

32. Glasgow R. What does it mean to be pragmatic? Pragmatic methods, measures, and models to facilitate research translation. Health Educ Behav. (2013) 40:257–65. doi: 10.1177/1090198113486805

PubMed Abstract | CrossRef Full Text | Google Scholar

Keywords: PSEs, RE-AIM, extension, evaluation, point-of-decision prompt, physical activity

Citation: Balis LE and Strayer T III (2019) Evaluating “Take the Stairs, Wyoming!” Through the RE-AIM Framework: Challenges and Opportunities. Front. Public Health 7:368. doi: 10.3389/fpubh.2019.00368

Received: 17 September 2019; Accepted: 19 November 2019;
Published: 17 December 2019.

Edited by:

Matthew Lee Smith, Texas A&M University, United States

Reviewed by:

Jo Ann Shoup, Kaiser Permanente, United States
Christopher Mierow Maylahn, New York State Department of Health, United States

Copyright © 2019 Balis and Strayer. This is an open-access article distributed under the terms of the Creative Commons Attribution License (CC BY). The use, distribution or reproduction in other forums is permitted, provided the original author(s) and the copyright owner(s) are credited and that the original publication in this journal is cited, in accordance with accepted academic practice. No use, distribution or reproduction is permitted which does not comply with these terms.

*Correspondence: Laura E. Balis, lbalis@uaex.edu

Disclaimer: All claims expressed in this article are solely those of the authors and do not necessarily represent those of their affiliated organizations, or those of the publisher, the editors and the reviewers. Any product that may be evaluated in this article or claim that may be made by its manufacturer is not guaranteed or endorsed by the publisher.