Skip to main content

METHODS article

Front. Genet., 14 October 2022
Sec. Human and Medical Genomics
This article is part of the Research Topic DNA-based Population Screening for Precision Public Health View all 28 articles

A mixed-methods protocol to develop and validate a stewardship maturity matrix for human genomic data in the cloud

  • 1Baylor College of Medicine, One Baylor Plaza, Houston, TX, (United States)
  • 2Earth System Science Center/NASA MSFC IMPACT, The University of Alabama in Huntsville, Huntsville, AL, (United States)
  • 3Stanford Center for Biomedical Ethics, Stanford University, Stanford, CA, (United States)

This article describes a mixed-methods protocol to develop and test the implementation of a stewardship maturity matrix (SMM) for repositories which govern access to human genomic data in the cloud. It is anticipated that the cloud will host most human genomic and related health datasets generated as part of publicly funded research in the coming years. However, repository managers lack practical tools for identifying what stewardship outcomes matter most to key stakeholders as well as how to track progress on their stewardship goals over time. In this article we describe a protocol that combines Delphi survey methods with SMM modeling first introduced in the earth and planetary sciences to develop a stewardship impact assessment tool for repositories that manage access to human genomic data. We discuss the strengths and limitations of this mixed-methods design and offer points to consider for wrangling both quantitative and qualitative data to enhance rigor and representativeness. We conclude with how the empirical methods bridged in this protocol have potential to improve evaluation of data stewardship systems and better align them with diverse stakeholder values in genomic data science.

1 Introduction

Genomics is a data-intensive science requiring extensive research collaboration across institutions and international borders. Research institutions face mounting pressure co-locate secure access, use and exchange of data to drive innovation in genomics (Langmead and Nellore, 2018). In addition to decentralized and federated access models, national research agencies are heavily invested in cloud technologies to enable controlled data access (Stein et al., 2015). This migration to the cloud represents an important shift not only in how data repositories stand up their privacy and security infrastructures, but also in how repository managers steward the data resources generated by research supported through public funds (Grzesik et al., 2021). Genomic data are uniquely identifying not only for the individual about whom data specifically relate, but also for their biological relatives and communities (Song et al., 2022) in which they live and work. Sharing genomic data also comes with increased risk of re-identification. Recent studies have shown, for example, that individuals can be re-identified from aggregate datasets with few record linkages (Dwork et al., 2017). These properties affect how genomic and related data are collected, regulated, and shared.

We refer to data repositories in this article as entities which store, organize, validate, archive, preserve and distribute genomic and related health data submitted by the community related to particular system(s) in compliance with the FAIR (findable, accessible, reusable and interoperable) Data Principles (NIH, 2022a). At a minimum, data stewardship can refer to the institutional practices and policies meant to calibrate appropriate data protection with compliant data access and use. Data stewardship is thus integral to well-functioning data governance systems (Boeckhout et al., 2018) that requires practical frameworks for compliance as well as stakeholder-engaged research on values and priorities.

Yet while commitments to responsible stewardship are outlined in repository data sharing policies, and methods for evaluating stewardship impact have been proposed (Wilkinson et al., 2016), these are largely underdeveloped for cloud-native environments with few exceptions [see for example access policies for the research analysis platform of the United Kingdom Biobank (UK Biobank, 2022) and NIH Cloud Guidebook (NIH, 2022b)].

We lack empirical data, for example, on what stewardship outcomes matter most to key stakeholders and how we should measure them over time. Examples of stewardship outcomes could include concordance between consent permissions and data use restrictions, ethics review of proposed data uses, processing times for data access requests, and the number of successful data access requests among researchers working in low-and middle-income countries. According to its access procedures, for example, United Kingdom Biobank’s cloud services charges fees for tiered access as well as data storage and analysis of data. While reduced access options are available, it is unclear whether pay-for-access policies affect who can afford to conduct the research in the first place.

In this article we describe a mixed-methods study design to identify stewardship outcomes and develop assessment criteria for assessing them in cloud-native environments. We first discuss the unique properties of genomic data and the ethical, legal and social issues of migrating such data to the cloud. We then explain how current genomic data management and access challenges the ways that repositories practice responsible stewardship in these new computing environments. In response to these practical challenges, we describe how a modified Delphi together with stewardship maturity modeling can be used to develop, validate and test the implementation of a stewardship impact assessment tool for global repositories which host data in the cloud. Next, we discuss analytical approaches for wrangling both quantitative and qualitative data generated in the proposed study, raising points to consider for ensuring rigor and representativeness. We conclude with how adapting SMMs for tracking progress on data stewardship can advance a new research agenda for evidence-based stewardship in human genomics as computing capabilities evolve.

1.1 Cloud infrastructures and the need to store, analyze and share human genomic data at enterprise scale

New digital infrastructures powered by cloud technologies transform how researchers interact with, analyze, and share data at scale including in clinical areas such as cancer (Langmead and Nellore, 2018) (Lau et al., 2017) and rare disease (Zurek et al., 2021). Using cloud services as infrastructure to host the largescale genomic data collections—one of four distinct types of cloud service separate from software as service (SaaS), platform as service (PaaS) and serverless (O’Driscoll et al., 2013)—offers powerful advantages (Stein, 2010). These include simplifying management (Schatz et al., 2022), overcoming security risks associated with traditional copy and download, and making data available in organized, searchable formats which reduce time and resource burdens (Kudtarkar et al., 2010).

However unique features of these computing environments compel new ethical, legal and social questions about how to responsibly access and steward genomic data in the cloud (Carter, 2019) (Filippi and Vieira, 2014). For example, data protection laws are jurisdiction-specific while actual data users may be based all over the world. This complicates which data protections regulations should principally apply: those in the jurisdiction where the repository is based, where the user resides, or both? Many repositories purchase cloud services from commercial providers (e.g., Google, Amazon Web Services), raising some concerns about the dependence on third parties and potential for interference (Molnár-Gábor et al., 2017). As Philipps and colleagues argue, “service outages caused by technical problems, changes to the company”s terms of service or even sudden closure of the company could block researchers’ access to data at any time. Also, it is often unclear to what extent researchers using cloud services can ensure that their data are not disclosed to third parties, such as those conducting abusive state-level “surveillance” (Phillips et al., 2020).

While there is broad consensus on data stewardship principles outlined in frameworks such as FAIR, TRUST, and CARE (Table 1), their assessment has been computationally difficult to perform in practice (Anjaria, 2020). It has been shown how modeling a stewardship maturity matrix (SMM) can be effective at capturing the FAIRness of datasets and TRUSTworthiness of repositories in the earth and planetary sciences (Downs et al., 2015) (22). SMMs are often presented by a two dimensional array mapping n stewardship outcomes of interest onto various levels of organizational development (Peng et al., 2015): ad hoc, minimal, intermediate, advanced and optimal. A sample SMM is presented in Table 2. Across the rows of the matrix reflect “various facets of core stewardship functionality, (e.g., data management), while the columns describe typical behaviours representing increasing maturity in practices and capability against each aspect, ranging from a poorly-managed or no-capability state to an advanced, well-managed state” (23). Once developed, the SMM “can be used not only as a guide to users about the rigour of data stewardship practices, but also as a tool for monitoring and improving aspects of organizational performance in producing, managing, or servicing climate data” (Dunn et al., 2021).

TABLE 1
www.frontiersin.org

TABLE 1. Data stewardship frameworks.

TABLE 2
www.frontiersin.org

TABLE 2. Template stewardship maturity matrix that charts n stewardship outcomes of interest onto five descriptive layers of organizational development.

Several reasons justify exploring how SMMs can be adapted to study human genomic data stewardship outcomes. First, advances in human genomics, like earth and planetary sciences, depend on sharing high quality and well managed data resources. Second, large, publicly funded repositories are among the primary sources where researchers access the data they need to conduct rigorous genomics research. Therefore data access and release activities catalyzed by repositories makes them strategic focal points for assessing stewardship outcomes (Dunn et al., 2021).

2 Methods

In the sections that follow, we provide methods and instructions for how to first develop (phase 1) validate (phase 2) and then test the implementation (phase 3) of a SMM for human genomic and related health data managed in the cloud. An overview of the protocol, as well as the specific materials and equipment used are provided in Figure 1 and Table 3, respectively. First, a scoping review of data sharing, management and access policies inform an initial core outcomes set for responsible data stewardship bespoke to cloud-native repositories. These core outcomes are then evaluated and further refined by actual repository managers, privacy officers and other institutional data stewards in a Delphi study. Institutional stakeholders engaged in the Delphi will also work to develop assessment criteria specific to each core outcome in a process that will result in a draft SMM. The SMM will be field tested with topic experts and piloted within repositories that currently host genomic data in the cloud.

FIGURE 1
www.frontiersin.org

FIGURE 1. Step-by-step outline of a protocol to develop, validate and pilot the implementation of a stewardship maturity matrix tool to track progress on human genomic data stewarded in the cloud.

TABLE 3
www.frontiersin.org

TABLE 3. Materials and equipment used in the protocol organized by study phase.

2.1 Phase 1: Identifying core outcomes of genomic data stewardship

The objective of Phase 1 is to inform a core outcomes set (COS) for genomic data stewarded in the cloud following a scoping literature review of data sharing, management and access policies (see for example Ethics and Governance Framework for the United Kingdom Biobank); published data stewardship frameworks, empirical studies, guidelines, and best practices. A detailed search strategy will be developed with guidance from a reference librarian, and which will include relevant search terms such as “genomic data,” “stewardship,” “cloud,” “infrastructure,” “data sharing,” “outcomes” among others to best capture existing stewardship measurements and approaches. An example search strategy is provided in the Supplementary Material S1.

2.2 Phase 2. Developing the stewardship maturity matrix

Findings from the literature review will inform an initial COS that will be refined in a three-round Delphi survey involving institutional data stewards, repository managers and other data access and privacy officers working at genomic data repositories globally.

Delphi methods are particularly well suited to refining COS and have been used in previous bioethics work to guide genomics policy (Stevens Smith et al., 2020). Delphi studies engage informed stakeholders through iterative rounds of structured communication and feedback (Banno et al., 2019). A Delphi facilitator collects panel responses, usually anonymously, and statistically aggregates and analyzes them (Rowe et al., 2001). The facilitator then provides summaries back to panelists who are invited to re-evaluate their position after considering responses from fellow panelists. This process is iterated across several rounds until reaching a pre-specified threshold indicating a consensus pattern.

The Delphi survey will enable panelists to evaluate each outcome for its relative importance and feasibility, suggest new outcomes and vote to eliminate those that are either infeasible to implement or unable to be measured in practice. In the final round of the Delphi, panelists will convene to develop assessment criteria specific to each core outcome and map these onto a two-dimensional array shown in Table 2.

2.2.1 Phase 2 participant selection

Prospective panelists should represent institutional stakeholders with expertise in data management and data access review (e.g., data access committee members, privacy officers, managers) across repositories which currently use cloud services or plan to in the future. Panel membership is critical to the external validity of the resulting SMM. We will therefore carefully consider personal attributes such as relevant expertise, experience, availability, and representativeness to guide recruitment decisions using Table 4 as a guide. Published studies also reported that offering incentives improved panel retention and enhanced the quality of participation (Belton et al., 2019) without unduly pressuring participation. As is customary, we plan to compensate Delphi panelists using rates typical of professional consultation in their respective fields.

TABLE 4
www.frontiersin.org

TABLE 4. Practical guidance for planning an expert Delphi panel.

2.2.2 Phase 2 data collection

In Round 1 of the Delphi, we will capture panelists’ perspectives on the relative importance and feasibility of each core outcome (Sinha et al., 2011) and allow panelists the opportunity to contribute additional outcomes. We intend to pilot each round of surveys among a group of topic-naïve experts to ensure overall comprehension. To discourage ambivalent responses, we will adopt a three point Likert scale for rating exercises (Lange et al., 2020). Embedding free text responses in the survey will allow us to triangulate quantitative survey data with qualitative analysis of the rationales panelists provide for each core outcome. In Round 2 of the Delphi, panelists will re-rate outcomes that failed to reach consensus in Round 1 after reviewing the results and panel summaries. A summary report of survey results and qualitative rationales from Round 2 will be given to panelists prior to a 60 min virtual consensus workshop in Round 3. During the workshop, panelists will provide input on draft assessment criteria specific to core outcomes deemed to be essential after Rounds 1 and 2. We will use a progressive maturity scale—the capability maturity model integration™ (Carnegie Mellon University, 2001)—to match core outcomes with assessment criteria.

2.2.3 Phase 2 data analysis

Practical guidance is limited on developing core outcome sets for organizations rather than individuals such as clinicians or policy makers (Sinha et al., 2011). We will therefore look to consensus building frameworks and psychometrically-validated tools used in the clinical (Kirkham et al., 2017) and other data science research contexts for guidance (Board, 2019). Descriptive statistics–including median, mean, interquartile range and standard deviation—will benchmark consensus on the core outcomes set (von der Gracht, 2012) when there is >70% agreement on one rating, or 80% agreement across two contiguous ratings (Needham and de Loe, 1990). We will generate a core-outcomes set from those outcomes which are considered essential via panel consensus and which demonstrate low to no polarity based on IQRs less than 1 (Raskin, 1994; Rayens and Hahn, 2000).

2.3 Phase 3 validation of the stewardship maturity matrix tool

Borrowing from approaches used in the environmental impact assessment literature (Bockstaller and Girardin, 2003), two validation exercises will serve to test the tool’s “output” and “usability” among prospective end users.

2.3.1 Phase 3 data collection

We will first develop hypothetical vignettes of stewardship practices that correspond to each of the five stewardship maturity levels outlined in the SMM and assign reference scores to them. Next, we will conduct cognitive interviews with prospective end users to validate how well user scores align with the reference (output validation). Cognitive interviewing is a specific approach to structured interviewing during which we will capture real-time feedback on user experience (Willis et al., 2004; Willis, 2005; Boeije and Willis, 2013). Interviewees ‘think aloud’ as they apply the SMM to assign an overall stewardship maturity score to each vignette until assessments reach a recommended interrater reliability score of 0.8 (Burla et al., 2008). Following the interviews participants will complete a System Usability Survey (Bangor et al., 2008; Lewis, 2018) to complement output validation data about the tool’s overall ease of use (user validation).

2.3.2 Phase 3 participation selection

Interviewees will be purposively recruited from expert communities who have experience developing data management and release policies, standards and executable data access workflows in cloud environments.

2.3.3 Phase 3 data analysis

We expect the validation exercises to generate quantitative as well as qualitative data. Both datasets will require their own analytical approaches. Pearson’s chi square test will enable us to compare reference scores with scores assigned by end users. User experience themes will also be synthesized from qualitative data emerging from the cognitive interviews using a content analysis approach. To enhance rigor, independent coders will develop an initial codebook from analyzing a sample of interview transcripts. Coders will then meet to resolve any discrepancies and revise the codebook as appropriate.

2.4 Pilot testing and implementation

Should we fail to reach interrater consensus during the cognitive interviews, or the usability tests reveal issues with internal validity, we will re-engage Delphi participants to further refine the SMM based on feedback from the validation studies. Upon successfully demonstrating the tool’s output validity and usability, we will pursue a pilot program with repository managers affiliated with cloud-native repositories. Pilot testing will inform the organizational factors to consider for implementation.

3 Limitations

The mixed-methods study design described in this protocol should be considered in light of several limitations and considerations. Delphi studies can be both time and resource intensive. It is possible that panelists are lost to attrition, which may skew the rating distributions. Second, engaging primarily institutional stakeholders to help develop the tool, may not adequately capture the perspectives and experiences of data contributors. Researchers could consider adapting the protocol in the future to solicit input directly from individuals who have previously shared their data, or plan to contribute their genomic data to cloud-native repositories in the future. Third, cloud computing and software engineering professionals skew largely white, European and male. Therefore, oversampling participants from groups commonly underrepresented in these technical fields, particularly during the Phase 2 validation phase, is critically important for promoting equity and representation as well as to ensuring external validity. Fourth, usability testing may not capture all relevant errors end users could make. Participants’ unfamiliarity with the concepts measured in Phase 2—for example ethics, stewardship and governance, time spent working in one’s role—as well as biases that can carry over from institutional environments are among the most common reasons why usability testing fails.

4 Conclusion and future directions

The development, validation, and implementation of an impact assessment tool is an important practical solution to a growing infrastructure problem for institutions that endeavor to track progress on genomic data stewardship in the cloud. This article outlines a mixed-methods protocol to rigorously develop and validate an assessment tool to monitor human genomic data stewardship in novel cloud environments. Research and development of a SMM for genomic data stewardship is especially timely as government investment in cloud-based data infrastructures expands (e.g., NIH STRIDES Initiative, https://cloud.nih.gov/about-strides/). Both institutional and public stakeholders benefit from transparent reporting of stewardship outcomes at the repository level. A reliable and usable SMM tool allows data managers, data access committee members, privacy officers, and other institutional officials to self-assess stewardship practices early and often. Scores generated from periodic assessment using the SMM tool could enable data stewards to identify ‘“quick wins” where higher ratings for some aspects require little effort to obtain” (Dunn et al., 2021). With the stewardship assessment criteria in mind, genomic researchers could proactively practice good stewardship when sharing or curating data they generate in their work. Researchers could also use stewardship scores to help guide their choices about which datasets to use for their projects. Finally, periodic assessment and routine reporting of stewardship outcomes using a standard SMM tool can improve repository practices in the long term while helping to sustain public trust in publicly funded genomic research in the future.

Future work will be needed to determine repository preparedness for implementing stewardship assessments as part of their annual reporting. Rigorous studies investigating the effects of transparent reporting of stewardship outcomes on more diverse data stakeholders (e.g., individual and community data contributors) are also needed. Cloud-native repositories could in the future seek certification for their commitment to responsible stewardship practice through programs sponsored under the CoreTrustSeal (https://www.coretrustseal.org/) and strike an advisory committee to review and assess new data infrastructure proposals. “If cloud technology is the future of biomedical science then, for genomics, the future is already here” (44). It is incumbent on data producers, users and regulators alike to prepare for this future in ways that are concordant with diverse value systems and as computer science and genomic data discovery evolve.

Data availability statement

The original contributions presented in the study are included in the article/Supplementary Material, further inquiries can be directed to the corresponding author.

Author contributions

VR conceptualized the study and developed the protocol with supervisory input from authors GP and MC. VR prepared initial drafts of this manuscript. All authors reviewed and approved the submitted version of the manuscript.

Funding

VR is supported through the National Human Genome Research Institute Ethical, Legal and Social Issues in Genetics/Genomics training grant T32 HG008953. GP is supported by the National Aeronautics and Space Administration (NASA) through the Marshall Space Flight Center (MSFC) Interagency Implementation and Advanced Concepts Team (IMPACT) project under a cooperative agreement with the University of Alabama in Huntsville (NASA Grant NNM11AA01A).

Acknowledgments

We thank the peer reviewers for their thoughtful comments on earlier versions of this paper.

Conflict of interest

The authors declare that the research was conducted in the absence of any commercial or financial relationships that could be construed as a potential conflict of interest.

Publisher’s note

All claims expressed in this article are solely those of the authors and do not necessarily represent those of their affiliated organizations, or those of the publisher, the editors and the reviewers. Any product that may be evaluated in this article, or claim that may be made by its manufacturer, is not guaranteed or endorsed by the publisher.

Supplementary material

The Supplementary Material for this article can be found online at: https://www.frontiersin.org/articles/10.3389/fgene.2022.876869/full#supplementary-material

Abbreviations

ELSI—Ethical, legal, and social issues. FAIR—Findable, Accessible, Interoperable, Reusable. IQR—interquartile range. PaaS—Platform as a service. SMM—stewardship maturity matrix. SaaS—Software as a service.

References

Anjaria, K. A. (2020). Computational implementation and formalism of FAIR data stewardship principles. DTA 54 (2), 193–214. doi:10.1108/dta-09-2019-0164

CrossRef Full Text | Google Scholar

Bangor, A., Kortum, P. T., and Miller, J. T. (2008). An empirical evaluation of the system usability scale. Int. J. Human–Computer Interact. 24 (6), 574–594. doi:10.1080/10447310802205776

CrossRef Full Text | Google Scholar

Banno, M., Tsujimoto, Y., and Kataoka, Y. (2019). Reporting quality of the Delphi technique in reporting guidelines: A protocol for a systematic analysis of the EQUATOR network library. BMJ Open 9 (4), e024942. doi:10.1136/bmjopen-2018-024942

PubMed Abstract | CrossRef Full Text | Google Scholar

Belton, I., MacDonald, A., Wright, G., and Hamlin, I. (2019). Improving the practical application of the Delphi method in group-based judgment: A six-step prescription for a well-founded and defensible process. Technol. Forecast. Soc. Change 147, 72–82. doi:10.1016/j.techfore.2019.07.002

CrossRef Full Text | Google Scholar

Board, C. S. (2019). CoreTrustSeal trustworthy data repositories requirements: Extended guidance 2020–2022. Available from: https://zenodo.org/record/3632533 (accessed on May 17, 2022).

Google Scholar

Bockstaller, C., and Girardin, P. (2003). How to validate environmental indicators. Agric. Syst. 76 (2), 639–653. doi:10.1016/s0308-521x(02)00053-7

CrossRef Full Text | Google Scholar

Boeckhout, M., Zielhuis, G. A., and Bredenoord, A. L. (2018). The FAIR guiding principles for data stewardship: Fair enough? Eur. J. Hum. Genet. 26 (7), 931–936. doi:10.1038/s41431-018-0160-0

PubMed Abstract | CrossRef Full Text | Google Scholar

Boeije, H., and Willis, G. (2013). The cognitive interviewing reporting framework (CIRF): Towards the harmonization of cognitive testing reports. Methodology 9 (3), 87–95. doi:10.1027/1614-2241/a000075

CrossRef Full Text | Google Scholar

Burla, L., Knierim, B., Barth, J., Liewald, K., Duetz, M., and Abel, T. (2008). From text to codings: Intercoder reliability assessment in qualitative content analysis. Nurs. Res. 57 (2), 113–117. doi:10.1097/01.NNR.0000313482.33917.7d

PubMed Abstract | CrossRef Full Text | Google Scholar

Carnegie Mellon University (2001). Capability maturity model integration (CMMISM), version 1.1. Continuous representation. Pittsburgh, United States: Carnegie Mellon University, 645.

Google Scholar

Carroll, S. R., Herczog, E., Hudson, M., Russell, K., and Stall, S. (2021). Operationalizing the CARE and FAIR principles for indigenous data futures. Sci. Data 8 (1), 108. doi:10.1038/s41597-021-00892-0

PubMed Abstract | CrossRef Full Text | Google Scholar

Carter, A. B. (2019). Considerations for genomic data privacy and security when working in the cloud. J. Mol. Diagn. 21 (4), 542–552. doi:10.1016/j.jmoldx.2018.07.009

PubMed Abstract | CrossRef Full Text | Google Scholar

Downs, R. R., Duerr, R., Hills, D. J., and Ramapriyan, H. K. (2015). Data stewardship in the earth sciences 21 (7/8). D-Lib Magazine.

CrossRef Full Text | Google Scholar

Dunn, R., Lief, C., Peng, G., Wright, W., Baddour, O., Donat, M., et al. (2021). Stewardship maturity assessment tools for modernization of climate data management. Data Sci. J. 20, 7. doi:10.5334/dsj-2021-007

CrossRef Full Text | Google Scholar

Dwork, C., Smith, A., Steinke, T., and Ullman, J. (2017). Exposed! A survey of attacks on private data. Annu. Rev. Stat. Appl. 4 (1), 61–84. doi:10.1146/annurev-statistics-060116-054123

CrossRef Full Text | Google Scholar

Filippi, P. D., and Vieira, M. S. (2014). THe commodification of information commons: The case of cloud computing. Sci. Tech. L. Rev 102, 42.

Google Scholar

Grzesik, P., Augustyn, D. R., Wyciślik, Ł., and Mrozek, D. (2021). Serverless computing in omics data analysis and integration. Briefings Bioinforma. 23, bbab349. doi:10.1093/bib/bbab349

CrossRef Full Text | Google Scholar

Kirkham, J. J., Davis, K., Altman, D. G., Blazeby, J. M., Clarke, M., Tunis, S., et al. (2017). Core outcome set-STAndards for development: The COS-STAD recommendations. PLoS Med. 14 (11), e1002447. doi:10.1371/journal.pmed.1002447

PubMed Abstract | CrossRef Full Text | Google Scholar

Kudtarkar, P., DeLuca, T. F., Fusaro, V. A., Tonellato, P. J., and Wall, D. P. (2010). Cost-effective cloud computing: A case study using the comparative genomics tool, roundup. Evol. Bioinform Online 6, 197–203. doi:10.4137/EBO.S6259

PubMed Abstract | CrossRef Full Text | Google Scholar

Lange, T., Kopkow, C., Lützner, J., Günther, K. P., Gravius, S., Scharf, H. P., et al. (2020). Comparison of different rating scales for the use in Delphi studies: Different scales lead to different consensus and show different test-retest reliability. BMC Med. Res. Methodol. 20 (1), 28. doi:10.1186/s12874-020-0912-8

PubMed Abstract | CrossRef Full Text | Google Scholar

Langmead, B., and Nellore, A. (2018). Cloud computing for genomic data analysis and collaboration. Nat. Rev. Genet. 19 (4), 208–219. doi:10.1038/nrg.2017.113

PubMed Abstract | CrossRef Full Text | Google Scholar

Lau, J. W., Lehnert, E., Sethi, A., Malhotra, R., Kaushik, G., Onder, Z., et al. (2017). The cancer genomics cloud: Collaborative, reproducible, and democratized—a new paradigm in large-scale computational research. Cancer Res. 77 (21), e3–e6. doi:10.1158/0008-5472.CAN-17-0387

PubMed Abstract | CrossRef Full Text | Google Scholar

Lewis, J. R. (2018). The system usability scale: Past, present, and future. Int. J. Human–Computer. Interact. 34 (7), 577–590. doi:10.1080/10447318.2018.1455307

CrossRef Full Text | Google Scholar

Lin, D., Crabtree, J., Dillo, I., Downs, R. R., Edmunds, R., Giaretta, D., et al. (2020). The TRUST Principles for digital repositories. Sci. Data 7 (1), 144. doi:10.1038/s41597-020-0486-7

PubMed Abstract | CrossRef Full Text | Google Scholar

Molnár-Gábor, F., Lueck, R., Yakneen, S., and Korbel, J. O. (2017). Computing patient data in the cloud: Practical and legal considerations for genetics and genomics research in europe and internationally. Genome Med. 9 (1), 58. doi:10.1186/s13073-017-0449-6

PubMed Abstract | CrossRef Full Text | Google Scholar

Needham, R. D., and de Loe, R. C. (1990). The policy Delphi: Purpose, structure and application. Can. Geogr. 34 (2), 133–142. doi:10.1111/j.1541-0064.1990.tb01258.x

CrossRef Full Text | Google Scholar

NIH, (2022). Biomedical data repositories and knowledgebases. Available from: https://datascience.nih.gov/data-ecosystem/biomedical-data-repositories-and-knowledgebases#repositories.

Google Scholar

NIH, (2022). Cloud Guidebook. Available from: http://nih-data-commons.us/cloud-guidebook/.

Google Scholar

O’Driscoll, A., Daugelaite, J., and Sleator, R. D. (2013). ‘Big data’, Hadoop and cloud computing in genomics. J. Biomed. Inf. 46 (5), 774–781. doi:10.1016/j.jbi.2013.07.001

CrossRef Full Text | Google Scholar

Peng, G., Privette, J. L., Kearns, E. J., Ritchey, N. A., and Ansari, S. (2015). A unified framework for measuring stewardship practices applied to digital environmental datasets. Data Sci. J. 13 (0), 231–252. doi:10.2481/dsj.14-049

CrossRef Full Text | Google Scholar

Phillips, M., Molnár-gábor, F., Korbel, J. O., Thorogood, A., Joly, Y., Chalmers, D., et al. (2020). Genomics: data sharing needs an international code of conduct. Nature. 578, 31–33.

PubMed Abstract | CrossRef Full Text | Google Scholar

Raskin, M. S. (1994). The Delphi study in field instruction revisited: Expert consensus on issues and research priorities. J. Soc. Work Educ. 30 (1), 75–89. doi:10.1080/10437797.1994.10672215

CrossRef Full Text | Google Scholar

Rayens, M. K., and Hahn, E. J. (2000). Building consensus using the policy Delphi method. Policy, Polit. Nurs. Pract. 1 (4), 308–315. doi:10.1177/152715440000100409

CrossRef Full Text | Google Scholar

Rowe, G., and Wright, G. (2001). “Expert opinions in forecasting: The role of the Delphi technique,” in International series in operations research & management science; vol. 30. Editor F. S. Hillier (Boston, MA: Springer US), 125–144.

CrossRef Full Text | Google Scholar

Schatz, M. C., Philippakis, A. A., Afgan, E., Banks, E., Carey, V. J., Carroll, R. J., et al. (2022). Inverting the model of genomics data sharing with the NHGRI Genomic Data Science Analysis, Visualization, and Informatics Lab-space. Cell Genomics. 2, 100085. doi:10.1016/j.xgen.2021.100085

PubMed Abstract | CrossRef Full Text | Google Scholar

Sinha, I. P., Smyth, R. L., and Williamson, P. R. (2011). Using the Delphi technique to determine which outcomes to measure in clinical trials: Recommendations for the future based on a systematic review of existing studies. PLoS Med. 8 (1), e1000393. doi:10.1371/journal.pmed.1000393

PubMed Abstract | CrossRef Full Text | Google Scholar

Song, L., Liu, H., Brinkman, F. S. L., Gill, E., Griffiths, E. J., Hsiao, W. W. L., et al. (2022). Addressing privacy concerns in sharing viral sequences and minimum contextual data in a public repository during the COVID-19 pandemic. Front. Genet. 12, 716541. doi:10.3389/fgene.2021.716541

PubMed Abstract | CrossRef Full Text | Google Scholar

Stein, L. D., Knoppers, B. M., Campbell, P., and Korbel, J. O. (2015). Data analysis: Create a cloud commons. Nature 523, 149–151. doi:10.1038/523149a

PubMed Abstract | CrossRef Full Text | Google Scholar

Stein, L. D. (2010). The case for cloud computing in genome informatics. Genome Biol. 11 (5), 207. doi:10.1186/gb-2010-11-5-207

PubMed Abstract | CrossRef Full Text | Google Scholar

Stevens Smith, H., Russell, H. V., Lee, B. H., and Morain, S. R. (2020). Using the Delphi method to identify clinicians’ perceived importance of pediatric exome sequencing results. Genet. Med. 22 (1), 69–76. doi:10.1038/s41436-019-0601-3

PubMed Abstract | CrossRef Full Text | Google Scholar

UK Biobank, (2022). UK Biobank access procedures: Application and review procedures for access to the UK Biobank resource. Available from: https://www.ukbiobank.ac.uk/media/hfnf2w0f/20220722-access-procedures-v2-1-final.pdf.

PubMed Abstract | Google Scholar

Van der Auwera, G., and O’Connor, B. (2020). Genomics in the cloud: Using docker, GATK, and WDL in terra. 1st ed. California, United States: O’Reilly Media.

PubMed Abstract | Google Scholar

von der Gracht, H. A. (2012). Consensus measurement in Delphi studies. Technol. Forecast. Soc. Change 79 (8), 1525–1536. doi:10.1016/j.techfore.2012.04.013

PubMed Abstract | CrossRef Full Text | Google Scholar

Wilkinson, M. D., Dumontier, M., Aalbersberg, IjJ., Appleton, G., Axton, M., Baak, A., et al. (2016). Comment: The FAIR Guiding Principles for scientific data management and stewardship. Sci. Data 3, 160018–160019. doi:10.1038/sdata.2016.18

PubMed Abstract | CrossRef Full Text | Google Scholar

Willis, G. (2005). Cognitive interviewing. California, United States: SAGE Publications, Inc.

PubMed Abstract | Google Scholar

Willis, G. B. (2004). “Cognitive interviewing revisited: A useful technique, in theory?,” in Wiley series in survey methodology. S. Presser, J. M. Rothgeb, M. P. Couper, J. T. Lessler, E. Martin, J. Martinet al. (Hoboken, NJ, USA: John Wiley & Sons), 23–43.

Google Scholar

Zurek, B., Ellwanger, K., Vissers, L. E. L. M., Schüle, R., Synofzik, M., Töpf, A., et al. (2021). Solve-RD: Systematic pan-European data sharing and collaborative analysis to solve rare diseases. Eur. J. Hum. Genet. 29 (9), 1325–1331. doi:10.1038/s41431-021-00859-0

PubMed Abstract | CrossRef Full Text | Google Scholar

Keywords: stewardship, human genomics, ELSI (ethical, legal, and social implications), data governance, cloud, Delphi

Citation: Rahimzadeh V, Peng G and Cho M (2022) A mixed-methods protocol to develop and validate a stewardship maturity matrix for human genomic data in the cloud. Front. Genet. 13:876869. doi: 10.3389/fgene.2022.876869

Received: 15 February 2022; Accepted: 28 September 2022;
Published: 14 October 2022.

Edited by:

Laura V. Milko, University of North Carolina at Chapel Hill, United States

Reviewed by:

Michael Adams, University of North Carolina Hospitals, United States
Jared C Roach, Institute for Systems Biology (ISB), United States

Copyright © 2022 Rahimzadeh, Peng and Cho. This is an open-access article distributed under the terms of the Creative Commons Attribution License (CC BY). The use, distribution or reproduction in other forums is permitted, provided the original author(s) and the copyright owner(s) are credited and that the original publication in this journal is cited, in accordance with accepted academic practice. No use, distribution or reproduction is permitted which does not comply with these terms.

*Correspondence: Vasiliki Rahimzadeh, vaso.rahimzadeh@bcm.edu

Disclaimer: All claims expressed in this article are solely those of the authors and do not necessarily represent those of their affiliated organizations, or those of the publisher, the editors and the reviewers. Any product that may be evaluated in this article or claim that may be made by its manufacturer is not guaranteed or endorsed by the publisher.