<?xml version="1.0" encoding="UTF-8"?>
<!DOCTYPE article PUBLIC "-//NLM//DTD Journal Archiving and Interchange DTD v2.3 20070202//EN" "archivearticle.dtd">
<article article-type="methods-article" dtd-version="2.3" xml:lang="EN" xmlns:mml="http://www.w3.org/1998/Math/MathML" xmlns:xlink="http://www.w3.org/1999/xlink">
<front>
<journal-meta>
<journal-id journal-id-type="publisher-id">Front. Genet.</journal-id>
<journal-title>Frontiers in Genetics</journal-title>
<abbrev-journal-title abbrev-type="pubmed">Front. Genet.</abbrev-journal-title>
<issn pub-type="epub">1664-8021</issn>
<publisher>
<publisher-name>Frontiers Media S.A.</publisher-name>
</publisher>
</journal-meta>
<article-meta>
<article-id pub-id-type="publisher-id">764020</article-id>
<article-id pub-id-type="doi">10.3389/fgene.2021.764020</article-id>
<article-categories>
<subj-group subj-group-type="heading">
<subject>Genetics</subject>
<subj-group>
<subject>Methods</subject>
</subj-group>
</subj-group>
</article-categories>
<title-group>
<article-title>Biological Network Inference With GRASP: A Bayesian Network Structure Learning Method Using Adaptive Sequential Monte Carlo</article-title>
<alt-title alt-title-type="left-running-head">Yu et&#x20;al.</alt-title>
<alt-title alt-title-type="right-running-head">Biological Network Inference with GRASP</alt-title>
</title-group>
<contrib-group>
<contrib contrib-type="author" corresp="yes">
<name>
<surname>Yu</surname>
<given-names>Kaixian</given-names>
</name>
<xref ref-type="aff" rid="aff1">
<sup>1</sup>
</xref>
<xref ref-type="corresp" rid="c001">&#x2a;</xref>
<uri xlink:href="https://loop.frontiersin.org/people/515853/overview"/>
</contrib>
<contrib contrib-type="author">
<name>
<surname>Cui</surname>
<given-names>Zihan</given-names>
</name>
<xref ref-type="aff" rid="aff1">
<sup>1</sup>
</xref>
</contrib>
<contrib contrib-type="author">
<name>
<surname>Sui</surname>
<given-names>Xin</given-names>
</name>
<xref ref-type="aff" rid="aff1">
<sup>1</sup>
</xref>
</contrib>
<contrib contrib-type="author">
<name>
<surname>Qiu</surname>
<given-names>Xing</given-names>
</name>
<xref ref-type="aff" rid="aff2">
<sup>2</sup>
</xref>
<uri xlink:href="https://loop.frontiersin.org/people/1157377/overview"/>
</contrib>
<contrib contrib-type="author" corresp="yes">
<name>
<surname>Zhang</surname>
<given-names>Jinfeng</given-names>
</name>
<xref ref-type="aff" rid="aff1">
<sup>1</sup>
</xref>
<xref ref-type="corresp" rid="c001">&#x2a;</xref>
<uri xlink:href="https://loop.frontiersin.org/people/684424/overview"/>
</contrib>
</contrib-group>
<aff id="aff1">
<label>
<sup>1</sup>
</label>Department of Statistics, Florida State University, <addr-line>Tallahassee</addr-line>, <addr-line>FL</addr-line>, <country>United&#x20;States</country>
</aff>
<aff id="aff2">
<label>
<sup>2</sup>
</label>Department of Biostatistics and Computational Biology, University of Rochester, <addr-line>Rochester</addr-line>, <addr-line>NY</addr-line>, <country>United&#x20;States</country>
</aff>
<author-notes>
<fn fn-type="edited-by">
<p>
<bold>Edited by:</bold> <ext-link ext-link-type="uri" xlink:href="https://loop.frontiersin.org/people/690813/overview">Yong He</ext-link>, Shandong University, China</p>
</fn>
<fn fn-type="edited-by">
<p>
<bold>Reviewed by:</bold> <ext-link ext-link-type="uri" xlink:href="https://loop.frontiersin.org/people/1465720/overview">Hao Chen</ext-link>, Peking University, China</p>
<p>
<ext-link ext-link-type="uri" xlink:href="https://loop.frontiersin.org/people/1466921/overview">Xin Liu</ext-link>, Shanghai University of Finance and Economics, China</p>
</fn>
<corresp id="c001">&#x2a;Correspondence: Jinfeng Zhang, <email>jinfeng@stat.fsu.edu</email>; Kaixian Yu, <email>kaixiany@gmail.com</email>
</corresp>
<fn fn-type="other">
<p>This article was submitted to Statistical Genetics and Methodology, a section of the journal Frontiers in Genetics</p>
</fn>
</author-notes>
<pub-date pub-type="epub">
<day>29</day>
<month>11</month>
<year>2021</year>
</pub-date>
<pub-date pub-type="collection">
<year>2021</year>
</pub-date>
<volume>12</volume>
<elocation-id>764020</elocation-id>
<history>
<date date-type="received">
<day>24</day>
<month>08</month>
<year>2021</year>
</date>
<date date-type="accepted">
<day>25</day>
<month>10</month>
<year>2021</year>
</date>
</history>
<permissions>
<copyright-statement>Copyright &#xa9; 2021 Yu, Cui, Sui, Qiu and Zhang.</copyright-statement>
<copyright-year>2021</copyright-year>
<copyright-holder>Yu, Cui, Sui, Qiu and Zhang</copyright-holder>
<license xlink:href="http://creativecommons.org/licenses/by/4.0/">
<p>This is an open-access article distributed under the terms of the Creative Commons Attribution License (CC BY). The use, distribution or reproduction in other forums is permitted, provided the original author(s) and the copyright owner(s) are credited and that the original publication in this journal is cited, in accordance with accepted academic practice. No use, distribution or reproduction is permitted which does not comply with these&#x20;terms.</p>
</license>
</permissions>
<abstract>
<p>Bayesian networks (BNs) provide a probabilistic, graphical framework for modeling high-dimensional joint distributions with complex correlation structures. BNs have wide applications in many disciplines, including biology, social science, finance and biomedical science. Despite extensive studies in the past, network structure learning from data is still a challenging open question in BN research. In this study, we present a sequential Monte Carlo (SMC)-based three-stage approach, GRowth-based Approach with Staged Pruning (GRASP). A double filtering strategy was first used for discovering the overall skeleton of the target BN. To search for the optimal network structures we designed an adaptive SMC (adSMC) algorithm to increase the quality and diversity of sampled networks which were further improved by a third stage to reclaim edges missed in the skeleton discovery step. GRASP gave very satisfactory results when tested on benchmark networks. Finally, BN structure learning using multiple types of genomics data illustrates GRASP&#x2019;s potential in discovering novel biological relationships in integrative genomic studies.</p>
</abstract>
<kwd-group>
<kwd>Bayesian network</kwd>
<kwd>Bayesian network structure learning</kwd>
<kwd>sequential Monte Carlo</kwd>
<kwd>adaptive sequential Monte Carlo</kwd>
<kwd>GRASP for BN structure learning</kwd>
<kwd>biological network inference</kwd>
</kwd-group>
<contract-num rid="cn001">R01GM126558</contract-num>
<contract-sponsor id="cn001">National Institutes of Health<named-content content-type="fundref-id">10.13039/100000002</named-content>
</contract-sponsor>
</article-meta>
</front>
<body>
<sec id="s1">
<title>Introduction</title>
<p>A Bayesian network (BN) is a graphical representation of the joint probability distribution of a set of variables (called nodes in the graph). BNs have been widely used in various fields, such as computational biology (<xref ref-type="bibr" rid="B11">Friedman, Linial, Nachman and Pe&#x2019;er 2000</xref>; <xref ref-type="bibr" rid="B31">Raval, Ghahramani and Wild 2002</xref>; <xref ref-type="bibr" rid="B40">Vignes, et&#x20;al., 2011</xref>), document classification (<xref ref-type="bibr" rid="B8">Denoyer and Gallinari 2004</xref>), and decision support system (<xref ref-type="bibr" rid="B20">Kristensen and Rasmussen 2002</xref>). A BN encodes conditional dependencies and independencies (CDIs) among variables into a directed acyclic graph (DAG). And this DAG is called the structure of a BN. When the structure of a BN is given, the parameters that quantify the CDIs can be estimated from observed data. If neither the structure nor parameters are given, they can be inferred from observed data. In this study, we will be focusing on the structure estimation of a BN and its application in learning biological networks using heterogeneous genomics&#x20;data.</p>
<p>The technical difficulties of structure learning are mainly due to the super-exponential cardinality of the DAG spaces, which are also quite rugged for most commonly used score functions. Estimating the structure exactly is an NP-hard problem (<xref ref-type="bibr" rid="B6">Cooper 1990</xref>; <xref ref-type="bibr" rid="B19">Koller and Friedman 2009</xref>). There have been many inexact and heuristic methods proposed in the past 2&#xa0;decades (<xref ref-type="bibr" rid="B48">Zhang, Li, Zhou and Wei 2013</xref>; <xref ref-type="bibr" rid="B1">Adabor, Acquaah-Mensah and Oduro 2015</xref>; <xref ref-type="bibr" rid="B21">Larjo and Lahdesmaki 2015</xref>; <xref ref-type="bibr" rid="B3">Amirkhani, Rahmati, Lucas and Hommersom 2017</xref>; <xref ref-type="bibr" rid="B10">Franzin, Sambo and Di Camillo 2017</xref>; <xref ref-type="bibr" rid="B16">Han, Zhang, Homayouni and Karmaus 2017</xref>; <xref ref-type="bibr" rid="B9">Ferreira-Santos, Monteiro-Soares and Rodrigues 2018</xref>; <xref ref-type="bibr" rid="B18">Jabbari, Visweswaran and Cooper 2018</xref>; <xref ref-type="bibr" rid="B23">Li and Guo 2018</xref>; <xref ref-type="bibr" rid="B36">Tang, Wang, Nguyen and Altintas 2019</xref>; <xref ref-type="bibr" rid="B49">Zhang, Wang, Duan and Sun 2019</xref>; <xref ref-type="bibr" rid="B47">Zhang, Rodrigues, Narain and Akmaev 2020</xref>; <xref ref-type="bibr" rid="B28">Liu, Gao, Wang and Ru 2021</xref>). The strategy of these methods can be classified mainly into three categories: constraint-based, score-based, and hybrid, which combines both constraint-based and score-based approaches.</p>
<p>A constraint-based method utilizes the conditional dependency test to identify the conditional dependencies and independencies among all the nodes (<xref ref-type="bibr" rid="B4">Campos 1998</xref>; <xref ref-type="bibr" rid="B7">de Campos and Huete 2000</xref>; <xref ref-type="bibr" rid="B29">Margaritis 2003</xref>; <xref ref-type="bibr" rid="B38">Tsamardinos, Aliferis and Statnikov 2003</xref>; <xref ref-type="bibr" rid="B42">Yaramakala and Margaritis 2005</xref>; <xref ref-type="bibr" rid="B2">Aliferis, Statnikov, Tsamardinos, Mani and Koutsoukos 2010</xref>; <xref ref-type="bibr" rid="B28">Liu, et&#x20;al., 2021</xref>). A major disadvantage of such a method is that a large number of tests have to be conducted; therefore, an appropriate method to adjust the <italic>p</italic>-values obtained from all the tests have to be applied. The fact that not all the tests are mutually independent further complicates the <italic>p</italic>-value adjustment. Another issue is that the goodness-of-fit of the obtained network is usually not considered in such an approach; therefore, the estimated BN may not fit the observed data&#x20;well.</p>
<p>A score-based method uses a score function to evaluate the structures of BNs on observed data (<xref ref-type="bibr" rid="B22">Larra&#xf1;aga, Poza, Yurramendi, Murga and Kuijpers 1996</xref>; <xref ref-type="bibr" rid="B12">Friedman, Nachman and Pe&#xe9;r 1999</xref>; <xref ref-type="bibr" rid="B14">G&#xe1;mez, Mateo and Puerta 2011</xref>). A searching algorithm is employed to search the best BN (with the highest score) with respect to certain score function. Various Bayesian and non-Bayesian score functions have been proposed in the past. As exact search is not feasible, over the past 2&#xa0;decades, various heuristic searching methods, such as hill climbing, tabu search, and simulated annealing were proposed to search for the optimal BN structures. The problem with score-based method is that the searching space is often very large and complicated; therefore, the searching algorithm either will take too much time to find the optimum or be trapped in local optima. Many efforts have been made to overcome this challenging issue, such as searching using an ordered DAG space to reduce the searching space (<xref ref-type="bibr" rid="B37">Teyssier and Koller 2012</xref>). In the ordered DAG space, the nodes are given an order such that edges will only be searched from higher orders to lower orders. The practical issue is that determining the orders and finding the optimal structure is equally difficult. More recently, various penalty-based methods were proposed to estimate the structures for Gaussian BN (GBN) (<xref ref-type="bibr" rid="B13">Fu and Zhou 2013</xref>; <xref ref-type="bibr" rid="B17">Huang, et&#x20;al., 2013</xref>; <xref ref-type="bibr" rid="B41">Xiang and Kim 2013</xref>). These methods have been shown to be quite efficient for GBN structure learning and are able to handle structure learning and parameter estimation simultaneously; however, these methods are quite restrictive: the joint distributions must approximately follow a multivariate Gaussian distribution and dependencies among nodes are assumed to be linear.</p>
<p>Hybrid methods which combine a constraint-based method and a score-based method were proposed to combine the advantages of both methods (<xref ref-type="bibr" rid="B39">Tsamardinos, Brown and Aliferis 2006</xref>). Such methods often contain two stages: first pruning the searching space by a constraint-based methods, then searching using a score function over the much smaller pruned space. In the pruning stage, the goal is to identify the so-called skeleton of the network, which is the undirected graph of the target DAG. Later in the second stage, the direction of each edge will be determined by optimizing the score function. In a hybrid method, it is important that the first stage identifies as many true undirected edges as possible, since only the identified undirected edges will be considered in the second&#x20;stage.</p>
<p>In this study, we developed a novel BN structure learning method named GRASP (GRowth-based Approach with Staged Pruning). It is a three-stage method: in stage one, we used a double filtering method to discover a cover of the true skeleton. Unlike the traditional constraint-based methods, which try to obtain the true skeleton exactly, our method only estimates a super set of the undirected edges and it only conditions on at most one node other than the pair of nodes being tested, which dramatically reduces the number of observations needed to make the test results robust. In stage two, we designed an adaptive sequential Monte Carlo (adSMC) (<xref ref-type="bibr" rid="B27">Liu and Chen 1998</xref>; <xref ref-type="bibr" rid="B26">Liu 2008</xref>) approach to search for a BN structure with optimal score based on constructed skeleton. SMC has been successfully adopted to solve optimization problems in the past (<xref ref-type="bibr" rid="B15">Grassberger 1997</xref>; <xref ref-type="bibr" rid="B51">Liang et&#x20;al., 2002</xref>; <xref ref-type="bibr" rid="B46">Zhang and Liu 2002</xref>; <xref ref-type="bibr" rid="B52">Zhang et&#x20;al., 2003</xref>; <xref ref-type="bibr" rid="B53">Zhang et&#x20;al., 2004</xref>; <xref ref-type="bibr" rid="B50">Zhang and Liu 2006</xref>; <xref ref-type="bibr" rid="B44">Zhang et&#x20;al., 2007</xref>; <xref ref-type="bibr" rid="B45">Zhang et&#x20;al., 2009</xref>). Compared to most greedy searching methods, SMC is less likely to be trapped in local optima. Another advantage of SMC is that it can be run in parallel for each SMC sample, making it suitable for distributed or GPU-based implementations. To further increase the efficiency of the sampling, an adaptive SMC strategy was used to generate higher scored networks. After these two stages, we enhanced the traditional two-stage approach with a third stage which adds possible missed edges back into the network using Random Order Hill Climbing (ROHC).</p>
</sec>
<sec id="s2">
<title>Methods and Data</title>
<sec id="s2-1">
<title>GRASP: GRowth-Based Approach With Staged Pruning</title>
<p>GRASP is a three-stage algorithm for learning the structure of a BN. In the first (pruning) stage, we designed a Double Filtering (DF) method to find the cover of the skeleton of the BN, where the skeleton of a BN is defined as the BN structure after removing the direction of all the edges, and the cover is defined as a superset of undirected edges containing all the edges of the skeleton. In the second (structure searching) stage, we developed an adaptive sequential Monte Carlo (adSMC) method to search the BN structure on the undirected network found in the first stage based on Bayesian information criterion (BIC) score. To reclaim the potentially missed edges, we designed a Random Order Hill Climbing (ROHC) method as the third&#x20;stage.</p>
</sec>
<sec id="s2-2">
<title>First Stage: Double Filtering (DF) Method to Infer the Skeleton</title>
<p>The first stage, namely Double Filtering (DF) method, contains two filtering processes. The first filtering was done by unconditioned tests, filtering out the nodes that are not ancestors or descendants of a given node <inline-formula id="inf1">
<mml:math id="m1">
<mml:mrow>
<mml:msub>
<mml:mi>X</mml:mi>
<mml:mi>i</mml:mi>
</mml:msub>
</mml:mrow>
</mml:math>
</inline-formula>. The second filtering was built on conditioned tests, further filtering out the nodes that are not direct neighbors (parents or children) of&#x20;<inline-formula id="inf2">
<mml:math id="m2">
<mml:mrow>
<mml:msub>
<mml:mi>X</mml:mi>
<mml:mi>i</mml:mi>
</mml:msub>
</mml:mrow>
</mml:math>
</inline-formula>.</p>
<p>Suppose we have <inline-formula id="inf3">
<mml:math id="m3">
<mml:mi>p</mml:mi>
</mml:math>
</inline-formula> nodes. For a given node <inline-formula id="inf4">
<mml:math id="m4">
<mml:mrow>
<mml:msub>
<mml:mi>X</mml:mi>
<mml:mi>i</mml:mi>
</mml:msub>
<mml:mo>,</mml:mo>
</mml:mrow>
</mml:math>
</inline-formula> let <inline-formula id="inf5">
<mml:math id="m5">
<mml:mrow>
<mml:mtext>nbr</mml:mtext>
<mml:mrow>
<mml:mo>(</mml:mo>
<mml:mrow>
<mml:msub>
<mml:mi>X</mml:mi>
<mml:mi>i</mml:mi>
</mml:msub>
</mml:mrow>
<mml:mo>)</mml:mo>
</mml:mrow>
</mml:mrow>
</mml:math>
</inline-formula> be the set of nodes that have an undirected edge with <inline-formula id="inf6">
<mml:math id="m6">
<mml:mrow>
<mml:msub>
<mml:mi>X</mml:mi>
<mml:mi>i</mml:mi>
</mml:msub>
</mml:mrow>
</mml:math>
</inline-formula>. Initialize <inline-formula id="inf7">
<mml:math id="m7">
<mml:mrow>
<mml:mtext>nbr</mml:mtext>
<mml:mrow>
<mml:mo>(</mml:mo>
<mml:mrow>
<mml:msub>
<mml:mi>X</mml:mi>
<mml:mi>i</mml:mi>
</mml:msub>
</mml:mrow>
<mml:mo>)</mml:mo>
</mml:mrow>
<mml:mo>&#x3d;</mml:mo>
<mml:mo>&#x2205;</mml:mo>
</mml:mrow>
</mml:math>
</inline-formula> the empty set for all <inline-formula id="inf8">
<mml:math id="m8">
<mml:mi>i</mml:mi>
</mml:math>
</inline-formula>. The procedure of the DF method is as follows:<list list-type="simple">
<list-item>
<p>1. <italic>First filtering</italic>. For each pair of nodes (<inline-formula id="inf9">
<mml:math id="m9">
<mml:mrow>
<mml:msub>
<mml:mi>X</mml:mi>
<mml:mi>i</mml:mi>
</mml:msub>
</mml:mrow>
</mml:math>
</inline-formula> and <inline-formula id="inf10">
<mml:math id="m10">
<mml:mrow>
<mml:msub>
<mml:mi>X</mml:mi>
<mml:mi>j</mml:mi>
</mml:msub>
</mml:mrow>
</mml:math>
</inline-formula>, <inline-formula id="inf11">
<mml:math id="m11">
<mml:mrow>
<mml:mi>i</mml:mi>
<mml:mo>&#x2260;</mml:mo>
<mml:mi>j</mml:mi>
</mml:mrow>
</mml:math>
</inline-formula>), record the <italic>p</italic>-value of their Mutual Information (MI) test as <inline-formula id="inf12">
<mml:math id="m12">
<mml:mrow>
<mml:msub>
<mml:mi>p</mml:mi>
<mml:mrow>
<mml:mtext>ij</mml:mtext>
</mml:mrow>
</mml:msub>
</mml:mrow>
</mml:math>
</inline-formula>. If <inline-formula id="inf13">
<mml:math id="m13">
<mml:mrow>
<mml:msub>
<mml:mi>p</mml:mi>
<mml:mrow>
<mml:mtext>ij</mml:mtext>
</mml:mrow>
</mml:msub>
<mml:mo>&#x3c;</mml:mo>
<mml:mi>&#x3b1;</mml:mi>
</mml:mrow>
</mml:math>
</inline-formula>, <inline-formula id="inf14">
<mml:math id="m14">
<mml:mi>&#x3b1;</mml:mi>
</mml:math>
</inline-formula> being a predetermined significance level for the test, add <inline-formula id="inf15">
<mml:math id="m15">
<mml:mrow>
<mml:msub>
<mml:mi>X</mml:mi>
<mml:mi>j</mml:mi>
</mml:msub>
</mml:mrow>
</mml:math>
</inline-formula> into <inline-formula id="inf16">
<mml:math id="m16">
<mml:mrow>
<mml:mi>n</mml:mi>
<mml:mi>b</mml:mi>
<mml:mi>r</mml:mi>
<mml:mrow>
<mml:mo>(</mml:mo>
<mml:mrow>
<mml:msub>
<mml:mi>X</mml:mi>
<mml:mi>i</mml:mi>
</mml:msub>
</mml:mrow>
<mml:mo>)</mml:mo>
</mml:mrow>
<mml:mo>&#xa0;</mml:mo>
</mml:mrow>
</mml:math>
</inline-formula>and <inline-formula id="inf17">
<mml:math id="m17">
<mml:mrow>
<mml:msub>
<mml:mi>X</mml:mi>
<mml:mi>i</mml:mi>
</mml:msub>
</mml:mrow>
</mml:math>
</inline-formula> into <inline-formula id="inf18">
<mml:math id="m18">
<mml:mrow>
<mml:mi>n</mml:mi>
<mml:mi>b</mml:mi>
<mml:mi>r</mml:mi>
<mml:mrow>
<mml:mo>(</mml:mo>
<mml:mrow>
<mml:msub>
<mml:mi>X</mml:mi>
<mml:mi>j</mml:mi>
</mml:msub>
</mml:mrow>
<mml:mo>)</mml:mo>
</mml:mrow>
</mml:mrow>
</mml:math>
</inline-formula>. The MI test is also used by other BN structure learning methods (<xref ref-type="bibr" rid="B4">Campos 1998</xref>). Let <inline-formula id="inf19">
<mml:math id="m19">
<mml:mrow>
<mml:msub>
<mml:mi>N</mml:mi>
<mml:mi>i</mml:mi>
</mml:msub>
</mml:mrow>
</mml:math>
</inline-formula> be the number of elements in <inline-formula id="inf20">
<mml:math id="m20">
<mml:mrow>
<mml:mtext>nbr</mml:mtext>
<mml:mrow>
<mml:mo>(</mml:mo>
<mml:mrow>
<mml:msub>
<mml:mi>X</mml:mi>
<mml:mi>i</mml:mi>
</mml:msub>
</mml:mrow>
<mml:mo>)</mml:mo>
</mml:mrow>
</mml:mrow>
</mml:math>
</inline-formula> after first filtering. Sort <inline-formula id="inf21">
<mml:math id="m21">
<mml:mrow>
<mml:mtext>nbr</mml:mtext>
<mml:mrow>
<mml:mo>(</mml:mo>
<mml:mrow>
<mml:msub>
<mml:mi>X</mml:mi>
<mml:mi>i</mml:mi>
</mml:msub>
</mml:mrow>
<mml:mo>)</mml:mo>
</mml:mrow>
</mml:mrow>
</mml:math>
</inline-formula> using their <inline-formula id="inf22">
<mml:math id="m22">
<mml:mi>p</mml:mi>
</mml:math>
</inline-formula>-values in ascending order and denote these nodes as <inline-formula id="inf23">
<mml:math id="m23">
<mml:mrow>
<mml:msubsup>
<mml:mi>X</mml:mi>
<mml:mn>1</mml:mn>
<mml:mrow>
<mml:mrow>
<mml:mo>(</mml:mo>
<mml:mi>i</mml:mi>
<mml:mo>)</mml:mo>
</mml:mrow>
</mml:mrow>
</mml:msubsup>
<mml:mo>,</mml:mo>
<mml:mo>&#xa0;</mml:mo>
<mml:msubsup>
<mml:mi>X</mml:mi>
<mml:mn>2</mml:mn>
<mml:mrow>
<mml:mrow>
<mml:mo>(</mml:mo>
<mml:mi>i</mml:mi>
<mml:mo>)</mml:mo>
</mml:mrow>
</mml:mrow>
</mml:msubsup>
<mml:mo>,</mml:mo>
<mml:mo>&#xa0;</mml:mo>
<mml:mo>&#x2026;</mml:mo>
<mml:mo>,</mml:mo>
<mml:msubsup>
<mml:mi>X</mml:mi>
<mml:mrow>
<mml:msub>
<mml:mi>N</mml:mi>
<mml:mi>i</mml:mi>
</mml:msub>
</mml:mrow>
<mml:mrow>
<mml:mrow>
<mml:mo>(</mml:mo>
<mml:mi>i</mml:mi>
<mml:mo>)</mml:mo>
</mml:mrow>
</mml:mrow>
</mml:msubsup>
</mml:mrow>
</mml:math>
</inline-formula>.</p>
</list-item>
<list-item>
<p>2. <italic>Second filtering.</italic> For every node <inline-formula id="inf24">
<mml:math id="m24">
<mml:mrow>
<mml:msub>
<mml:mi>X</mml:mi>
<mml:mi>i</mml:mi>
</mml:msub>
</mml:mrow>
</mml:math>
</inline-formula>, initialize the set of its final neighbors <inline-formula id="inf25">
<mml:math id="m25">
<mml:mrow>
<mml:msup>
<mml:mrow>
<mml:mtext>nbr</mml:mtext>
</mml:mrow>
<mml:mi>C</mml:mi>
</mml:msup>
<mml:mrow>
<mml:mo>(</mml:mo>
<mml:mrow>
<mml:msub>
<mml:mi>X</mml:mi>
<mml:mi>i</mml:mi>
</mml:msub>
</mml:mrow>
<mml:mo>)</mml:mo>
</mml:mrow>
<mml:mo>&#x3d;</mml:mo>
<mml:mtext>nbr</mml:mtext>
<mml:mrow>
<mml:mo>(</mml:mo>
<mml:mrow>
<mml:msub>
<mml:mi>X</mml:mi>
<mml:mi>i</mml:mi>
</mml:msub>
</mml:mrow>
<mml:mo>)</mml:mo>
</mml:mrow>
</mml:mrow>
</mml:math>
</inline-formula>. Loop over its elements in the order of <inline-formula id="inf26">
<mml:math id="m26">
<mml:mrow>
<mml:msubsup>
<mml:mi>X</mml:mi>
<mml:mn>1</mml:mn>
<mml:mrow>
<mml:mrow>
<mml:mo>(</mml:mo>
<mml:mi>i</mml:mi>
<mml:mo>)</mml:mo>
</mml:mrow>
</mml:mrow>
</mml:msubsup>
<mml:mo>,</mml:mo>
<mml:mo>&#xa0;</mml:mo>
<mml:msubsup>
<mml:mi>X</mml:mi>
<mml:mn>2</mml:mn>
<mml:mrow>
<mml:mrow>
<mml:mo>(</mml:mo>
<mml:mi>i</mml:mi>
<mml:mo>)</mml:mo>
</mml:mrow>
</mml:mrow>
</mml:msubsup>
<mml:mo>,</mml:mo>
<mml:mo>&#xa0;</mml:mo>
<mml:mo>&#x2026;</mml:mo>
<mml:mo>,</mml:mo>
<mml:msubsup>
<mml:mi>X</mml:mi>
<mml:mrow>
<mml:msub>
<mml:mi>N</mml:mi>
<mml:mi>i</mml:mi>
</mml:msub>
</mml:mrow>
<mml:mrow>
<mml:mrow>
<mml:mo>(</mml:mo>
<mml:mi>i</mml:mi>
<mml:mo>)</mml:mo>
</mml:mrow>
</mml:mrow>
</mml:msubsup>
</mml:mrow>
</mml:math>
</inline-formula>.</p>
<list list-type="simple">
<list-item>
<p>(a). For every element <inline-formula id="inf27">
<mml:math id="m27">
<mml:mrow>
<mml:msubsup>
<mml:mi>X</mml:mi>
<mml:mi>j</mml:mi>
<mml:mrow>
<mml:mrow>
<mml:mo>(</mml:mo>
<mml:mi>i</mml:mi>
<mml:mo>)</mml:mo>
</mml:mrow>
</mml:mrow>
</mml:msubsup>
</mml:mrow>
</mml:math>
</inline-formula>, find <inline-formula id="inf28">
<mml:math id="m28">
<mml:mrow>
<mml:mtext>nbr</mml:mtext>
<mml:mrow>
<mml:mo>(</mml:mo>
<mml:mrow>
<mml:msub>
<mml:mi>X</mml:mi>
<mml:mi>i</mml:mi>
</mml:msub>
</mml:mrow>
<mml:mo>)</mml:mo>
</mml:mrow>
<mml:mo>&#x2229;</mml:mo>
<mml:mtext>nbr</mml:mtext>
<mml:mrow>
<mml:mo>(</mml:mo>
<mml:mrow>
<mml:msubsup>
<mml:mi>X</mml:mi>
<mml:mi>j</mml:mi>
<mml:mrow>
<mml:mrow>
<mml:mo>(</mml:mo>
<mml:mi>i</mml:mi>
<mml:mo>)</mml:mo>
</mml:mrow>
</mml:mrow>
</mml:msubsup>
</mml:mrow>
<mml:mo>)</mml:mo>
</mml:mrow>
</mml:mrow>
</mml:math>
</inline-formula>, the intersection of the neighbors of <inline-formula id="inf29">
<mml:math id="m29">
<mml:mrow>
<mml:msub>
<mml:mi>X</mml:mi>
<mml:mi>i</mml:mi>
</mml:msub>
</mml:mrow>
</mml:math>
</inline-formula> and <inline-formula id="inf30">
<mml:math id="m30">
<mml:mrow>
<mml:msubsup>
<mml:mi>X</mml:mi>
<mml:mi>j</mml:mi>
<mml:mrow>
<mml:mrow>
<mml:mo>(</mml:mo>
<mml:mi>i</mml:mi>
<mml:mo>)</mml:mo>
</mml:mrow>
</mml:mrow>
</mml:msubsup>
</mml:mrow>
</mml:math>
</inline-formula>.</p>
</list-item>
<list-item>
<p>(b). For every element <inline-formula id="inf31">
<mml:math id="m31">
<mml:mrow>
<mml:msubsup>
<mml:mi>X</mml:mi>
<mml:mi>k</mml:mi>
<mml:mrow>
<mml:mrow>
<mml:mo>(</mml:mo>
<mml:mrow>
<mml:mi>i</mml:mi>
<mml:mo>,</mml:mo>
<mml:mi>j</mml:mi>
</mml:mrow>
<mml:mo>)</mml:mo>
</mml:mrow>
</mml:mrow>
</mml:msubsup>
</mml:mrow>
</mml:math>
</inline-formula> in the set of intersection, perform a conditional dependency test for <inline-formula id="inf32">
<mml:math id="m32">
<mml:mrow>
<mml:msub>
<mml:mi>X</mml:mi>
<mml:mi>i</mml:mi>
</mml:msub>
</mml:mrow>
</mml:math>
</inline-formula> and <inline-formula id="inf33">
<mml:math id="m33">
<mml:mrow>
<mml:msubsup>
<mml:mi>X</mml:mi>
<mml:mi>k</mml:mi>
<mml:mrow>
<mml:mrow>
<mml:mo>(</mml:mo>
<mml:mrow>
<mml:mi>i</mml:mi>
<mml:mo>,</mml:mo>
<mml:mi>j</mml:mi>
</mml:mrow>
<mml:mo>)</mml:mo>
</mml:mrow>
</mml:mrow>
</mml:msubsup>
</mml:mrow>
</mml:math>
</inline-formula>, given <inline-formula id="inf34">
<mml:math id="m34">
<mml:mrow>
<mml:msubsup>
<mml:mi>X</mml:mi>
<mml:mi>j</mml:mi>
<mml:mrow>
<mml:mrow>
<mml:mo>(</mml:mo>
<mml:mi>i</mml:mi>
<mml:mo>)</mml:mo>
</mml:mrow>
</mml:mrow>
</mml:msubsup>
<mml:mo>.</mml:mo>
</mml:mrow>
</mml:math>
</inline-formula> If the <inline-formula id="inf35">
<mml:math id="m35">
<mml:mi>p</mml:mi>
</mml:math>
</inline-formula>-value <inline-formula id="inf36">
<mml:math id="m36">
<mml:mrow>
<mml:mo>&#x3e;</mml:mo>
<mml:mi>&#x3b1;</mml:mi>
</mml:mrow>
</mml:math>
</inline-formula>, remove <inline-formula id="inf37">
<mml:math id="m37">
<mml:mrow>
<mml:msubsup>
<mml:mi>X</mml:mi>
<mml:mi>j</mml:mi>
<mml:mrow>
<mml:mrow>
<mml:mo>(</mml:mo>
<mml:mi>i</mml:mi>
<mml:mo>)</mml:mo>
</mml:mrow>
</mml:mrow>
</mml:msubsup>
</mml:mrow>
</mml:math>
</inline-formula> from <inline-formula id="inf38">
<mml:math id="m38">
<mml:mrow>
<mml:msup>
<mml:mrow>
<mml:mtext>nbr</mml:mtext>
</mml:mrow>
<mml:mi>S</mml:mi>
</mml:msup>
<mml:mrow>
<mml:mo>(</mml:mo>
<mml:mrow>
<mml:msub>
<mml:mi>X</mml:mi>
<mml:mi>i</mml:mi>
</mml:msub>
</mml:mrow>
<mml:mo>)</mml:mo>
</mml:mrow>
</mml:mrow>
</mml:math>
</inline-formula>.</p>
</list-item>
</list>
</list-item>
</list>
</p>
<p>After applying the DF method on all <inline-formula id="inf39">
<mml:math id="m39">
<mml:mi>p</mml:mi>
</mml:math>
</inline-formula> nodes, the collection of <inline-formula id="inf40">
<mml:math id="m40">
<mml:mrow>
<mml:msup>
<mml:mrow>
<mml:mtext>nbr</mml:mtext>
</mml:mrow>
<mml:mi>C</mml:mi>
</mml:msup>
<mml:mrow>
<mml:mo>(</mml:mo>
<mml:mrow>
<mml:msub>
<mml:mi>X</mml:mi>
<mml:mi>i</mml:mi>
</mml:msub>
</mml:mrow>
<mml:mo>)</mml:mo>
</mml:mrow>
<mml:mo>,</mml:mo>
<mml:mo>&#xa0;</mml:mo>
<mml:mi>i</mml:mi>
<mml:mo>&#x3d;</mml:mo>
<mml:mn>1,2</mml:mn>
<mml:mo>,</mml:mo>
<mml:mn>...</mml:mn>
<mml:mo>,</mml:mo>
<mml:mi>p</mml:mi>
</mml:mrow>
</mml:math>
</inline-formula> gives us the skeleton of&#x20;BN.</p>
</sec>
<sec id="s2-3">
<title>Second Stage: Structure Searching</title>
<p>In the pruned space, we designed an adaptive sequential Monte Carlo (adSMC) method to search the structure of the Bayesian network. In a traditional sequential Monte Carlo, the random variable of all <inline-formula id="inf41">
<mml:math id="m41">
<mml:mi>p</mml:mi>
</mml:math>
</inline-formula> nodes (or features) <inline-formula id="inf42">
<mml:math id="m42">
<mml:mrow>
<mml:mi mathvariant="bold-italic">X</mml:mi>
<mml:mo>&#x2208;</mml:mo>
<mml:msup>
<mml:mi mathvariant="normal">&#x211d;</mml:mi>
<mml:mi>p</mml:mi>
</mml:msup>
</mml:mrow>
</mml:math>
</inline-formula> is decomposed into <inline-formula id="inf43">
<mml:math id="m43">
<mml:mi>M</mml:mi>
</mml:math>
</inline-formula> blocks <inline-formula id="inf44">
<mml:math id="m44">
<mml:mrow>
<mml:mrow>
<mml:mo>(</mml:mo>
<mml:mrow>
<mml:msub>
<mml:mi mathvariant="bold-italic">x</mml:mi>
<mml:mn>1</mml:mn>
</mml:msub>
<mml:mo>,</mml:mo>
<mml:mi mathvariant="bold-italic">&#xa0;</mml:mi>
<mml:msub>
<mml:mi mathvariant="bold-italic">x</mml:mi>
<mml:mn>2</mml:mn>
</mml:msub>
<mml:mo>,</mml:mo>
<mml:mi mathvariant="bold-italic">&#xa0;</mml:mi>
<mml:mo>&#x2026;</mml:mo>
<mml:mo>,</mml:mo>
<mml:mi mathvariant="bold-italic">&#xa0;</mml:mi>
<mml:msub>
<mml:mi mathvariant="bold-italic">x</mml:mi>
<mml:mi mathvariant="bold-italic">M</mml:mi>
</mml:msub>
</mml:mrow>
<mml:mo>)</mml:mo>
</mml:mrow>
</mml:mrow>
</mml:math>
</inline-formula> each with <inline-formula id="inf45">
<mml:math id="m45">
<mml:mrow>
<mml:msub>
<mml:mi>p</mml:mi>
<mml:mi>i</mml:mi>
</mml:msub>
</mml:mrow>
</mml:math>
</inline-formula> features <inline-formula id="inf46">
<mml:math id="m46">
<mml:mrow>
<mml:msub>
<mml:mi mathvariant="bold-italic">x</mml:mi>
<mml:mi mathvariant="bold-italic">i</mml:mi>
</mml:msub>
<mml:mo>&#x2208;</mml:mo>
<mml:msup>
<mml:mi mathvariant="normal">&#x211d;</mml:mi>
<mml:mrow>
<mml:msub>
<mml:mi>p</mml:mi>
<mml:mi>i</mml:mi>
</mml:msub>
</mml:mrow>
</mml:msup>
</mml:mrow>
</mml:math>
</inline-formula> and <inline-formula id="inf47">
<mml:math id="m47">
<mml:mrow>
<mml:mstyle displaystyle="true">
<mml:munderover>
<mml:mo>&#x2211;</mml:mo>
<mml:mrow>
<mml:mi>i</mml:mi>
<mml:mo>&#x3d;</mml:mo>
<mml:mn>1</mml:mn>
</mml:mrow>
<mml:mi>M</mml:mi>
</mml:munderover>
<mml:mrow>
<mml:msub>
<mml:mi>p</mml:mi>
<mml:mi>i</mml:mi>
</mml:msub>
<mml:mo>&#x3d;</mml:mo>
<mml:mi>p</mml:mi>
</mml:mrow>
</mml:mstyle>
</mml:mrow>
</mml:math>
</inline-formula>, and the decomposition is predefined and fixed throughout the whole sampling procedure. One usually samples<inline-formula id="inf48">
<mml:math id="m48">
<mml:mrow>
<mml:mi mathvariant="bold-italic">&#xa0;</mml:mi>
<mml:msub>
<mml:mi mathvariant="bold-italic">x</mml:mi>
<mml:mn>1</mml:mn>
</mml:msub>
</mml:mrow>
</mml:math>
</inline-formula> at first, then <inline-formula id="inf49">
<mml:math id="m49">
<mml:mrow>
<mml:msub>
<mml:mi mathvariant="bold-italic">x</mml:mi>
<mml:mn>2</mml:mn>
</mml:msub>
</mml:mrow>
</mml:math>
</inline-formula>, and so on. However, the sequence each variable is sampled (namely sampling sequence in this study) based on any prior decomposition may not be the most efficient one. The optimal sampling sequence may need to be decided dynamically. For example, when <inline-formula id="inf50">
<mml:math id="m50">
<mml:mrow>
<mml:msub>
<mml:mi mathvariant="bold-italic">x</mml:mi>
<mml:mn>1</mml:mn>
</mml:msub>
<mml:mo>,</mml:mo>
<mml:mi mathvariant="bold-italic">&#xa0;</mml:mi>
<mml:msub>
<mml:mi mathvariant="bold-italic">x</mml:mi>
<mml:mn>2</mml:mn>
</mml:msub>
<mml:mo>,</mml:mo>
<mml:mo>&#x2026;</mml:mo>
<mml:mo>,</mml:mo>
<mml:mi mathvariant="bold-italic">&#xa0;</mml:mi>
<mml:msub>
<mml:mi mathvariant="bold-italic">x</mml:mi>
<mml:mrow>
<mml:mi mathvariant="bold-italic">m</mml:mi>
<mml:mo>&#x2212;</mml:mo>
<mml:mn>1</mml:mn>
</mml:mrow>
</mml:msub>
</mml:mrow>
</mml:math>
</inline-formula> have been sampled for some <inline-formula id="inf51">
<mml:math id="m51">
<mml:mrow>
<mml:mn>0</mml:mn>
<mml:mo>&#x3c;</mml:mo>
<mml:mi>m</mml:mi>
<mml:mo>&#x2264;</mml:mo>
<mml:mi>M</mml:mi>
</mml:mrow>
</mml:math>
</inline-formula>, the conditional distribution <inline-formula id="inf52">
<mml:math id="m52">
<mml:mrow>
<mml:mi>f</mml:mi>
<mml:mrow>
<mml:mo>(</mml:mo>
<mml:mrow>
<mml:mo>&#xa0;</mml:mo>
<mml:mrow>
<mml:mrow>
<mml:msub>
<mml:mi mathvariant="bold-italic">x</mml:mi>
<mml:mi mathvariant="bold-italic">m</mml:mi>
</mml:msub>
</mml:mrow>
<mml:mo>&#x7c;</mml:mo>
</mml:mrow>
<mml:msub>
<mml:mi mathvariant="bold-italic">x</mml:mi>
<mml:mn>1</mml:mn>
</mml:msub>
<mml:mo>,</mml:mo>
<mml:mi mathvariant="bold-italic">&#xa0;</mml:mi>
<mml:msub>
<mml:mi mathvariant="bold-italic">x</mml:mi>
<mml:mn>2</mml:mn>
</mml:msub>
<mml:mo>,</mml:mo>
<mml:mo>&#x2026;</mml:mo>
<mml:mo>,</mml:mo>
<mml:mi mathvariant="bold-italic">&#xa0;</mml:mi>
<mml:msub>
<mml:mi mathvariant="bold-italic">x</mml:mi>
<mml:mrow>
<mml:mi mathvariant="bold-italic">m</mml:mi>
<mml:mo>&#x2212;</mml:mo>
<mml:mn>1</mml:mn>
</mml:mrow>
</mml:msub>
</mml:mrow>
<mml:mo>)</mml:mo>
</mml:mrow>
</mml:mrow>
</mml:math>
</inline-formula> may have a small set of candidate decompositions (to satisfy the acyclic condition) which limits the diversity of the SMC samples. Therefore, we designed our sampling block <inline-formula id="inf53">
<mml:math id="m53">
<mml:mrow>
<mml:msub>
<mml:mi mathvariant="bold-italic">x</mml:mi>
<mml:mi mathvariant="bold-italic">m</mml:mi>
</mml:msub>
</mml:mrow>
</mml:math>
</inline-formula> conditioning on the current sampled structure <inline-formula id="inf54">
<mml:math id="m54">
<mml:mrow>
<mml:msub>
<mml:mi mathvariant="bold-italic">x</mml:mi>
<mml:mn>1</mml:mn>
</mml:msub>
<mml:mo>,</mml:mo>
<mml:mi mathvariant="bold-italic">&#xa0;</mml:mi>
<mml:msub>
<mml:mi mathvariant="bold-italic">x</mml:mi>
<mml:mn>2</mml:mn>
</mml:msub>
<mml:mo>,</mml:mo>
<mml:mo>&#x2026;</mml:mo>
<mml:mo>,</mml:mo>
<mml:mi mathvariant="bold-italic">&#xa0;</mml:mi>
<mml:msub>
<mml:mi mathvariant="bold-italic">x</mml:mi>
<mml:mrow>
<mml:mi mathvariant="bold-italic">m</mml:mi>
<mml:mo>&#x2212;</mml:mo>
<mml:mn>1</mml:mn>
</mml:mrow>
</mml:msub>
</mml:mrow>
</mml:math>
</inline-formula> to increase the diversity and quality of obtained samples (see <xref ref-type="sec" rid="s10">Supplementary Figure S1</xref> in Supplementary Materials for an example).</p>
<p>For each SMC sample, we start with all possible fully connected triplets (three nodes connected by three undirected edges) discovered in the first stage. We sample one such triplet having the least outside connection, e.g., the one having the least undirected edges connected to its nodes (<xref ref-type="fig" rid="F1">Figure&#x20;1A</xref>). These triplets are likely to be restricted to certain configuration by the sampled structure; therefore, to sample them earlier allows more variety in their configurations. When all fully connected triplets are sampled, partially connected triplets (two undirected edges among three nodes) are considered (<xref ref-type="fig" rid="F1">Figure&#x20;1B</xref>). Lastly, we consider pairs (the remaining undirected edges, <xref ref-type="fig" rid="F1">Figure&#x20;1C</xref>). For partially connected triplets and pairs, the configurations with the least outside connections are sampled&#x20;first.</p>
<fig id="F1" position="float">
<label>FIGURE 1</label>
<caption>
<p>Structure discovering procedure.</p>
</caption>
<graphic xlink:href="fgene-12-764020-g001.tif"/>
</fig>
<p>The probabilities of possible configurations of triplets and pairs are proportional to their BIC (Bayesian Information Criterion) score defined as <inline-formula id="inf55">
<mml:math id="m55">
<mml:mrow>
<mml:mtext>BIC</mml:mtext>
<mml:mo>&#x3d;</mml:mo>
<mml:mtext>ln</mml:mtext>
<mml:mrow>
<mml:mo>(</mml:mo>
<mml:mi>n</mml:mi>
<mml:mo>)</mml:mo>
</mml:mrow>
<mml:mo>&#x22c5;</mml:mo>
<mml:mi>k</mml:mi>
<mml:mo>&#x2212;</mml:mo>
<mml:mn>2</mml:mn>
<mml:mo>&#x22c5;</mml:mo>
<mml:mtext>ln</mml:mtext>
<mml:mrow>
<mml:mo>(</mml:mo>
<mml:mrow>
<mml:mover accent="true">
<mml:mi>L</mml:mi>
<mml:mo>&#x5e;</mml:mo>
</mml:mover>
</mml:mrow>
<mml:mo>)</mml:mo>
</mml:mrow>
</mml:mrow>
</mml:math>
</inline-formula>, where<inline-formula id="inf56">
<mml:math id="m56">
<mml:mrow>
<mml:mtext>&#xa0;</mml:mtext>
<mml:mi>n</mml:mi>
</mml:mrow>
</mml:math>
</inline-formula> represents sample size, <inline-formula id="inf57">
<mml:math id="m57">
<mml:mi>k</mml:mi>
</mml:math>
</inline-formula> represents number of parameters of the partial BN model, and <inline-formula id="inf58">
<mml:math id="m58">
<mml:mrow>
<mml:mrow>
<mml:mover accent="true">
<mml:mi>L</mml:mi>
<mml:mo>&#x5e;</mml:mo>
</mml:mover>
</mml:mrow>
<mml:mo>&#x3d;</mml:mo>
<mml:mi>p</mml:mi>
<mml:mrow>
<mml:mo>(</mml:mo>
<mml:mrow>
<mml:mrow>
<mml:mi>x</mml:mi>
<mml:mo>&#x7c;</mml:mo>
</mml:mrow>
<mml:mrow>
<mml:mover accent="true">
<mml:mi>&#x3b8;</mml:mi>
<mml:mo>&#x5e;</mml:mo>
</mml:mover>
</mml:mrow>
<mml:mo>,</mml:mo>
<mml:mtext>Model</mml:mtext>
</mml:mrow>
<mml:mo>)</mml:mo>
</mml:mrow>
</mml:mrow>
</mml:math>
</inline-formula> is the maximized value of the likelihood function of the model, with estimated parameters <inline-formula id="inf59">
<mml:math id="m59">
<mml:mrow>
<mml:mover accent="true">
<mml:mi>&#x3b8;</mml:mi>
<mml:mo>&#x5e;</mml:mo>
</mml:mover>
</mml:mrow>
</mml:math>
</inline-formula> from observed data<inline-formula id="inf60">
<mml:math id="m60">
<mml:mrow>
<mml:mo>&#xa0;</mml:mo>
<mml:mi>x</mml:mi>
</mml:mrow>
</mml:math>
</inline-formula>. The probability is set to be 0 if certain configuration fails to satisfy the acyclicity condition. In summary, for triplets (pairs), we calculate the BIC for all possible configurations of this triplet (pair) and sample one configuration with probability<disp-formula id="e1">
<mml:math id="m61">
<mml:mrow>
<mml:mi>P</mml:mi>
<mml:mrow>
<mml:mo>(</mml:mo>
<mml:mrow>
<mml:mtext>configuration</mml:mtext>
<mml:mtext>&#x2009;</mml:mtext>
<mml:mi>i</mml:mi>
</mml:mrow>
<mml:mo>)</mml:mo>
</mml:mrow>
<mml:mo>&#x221d;</mml:mo>
<mml:mrow>
<mml:mo>{</mml:mo>
<mml:mrow>
<mml:mtable>
<mml:mtr>
<mml:mtd>
<mml:mrow>
<mml:mi>exp</mml:mi>
<mml:mrow>
<mml:mo>(</mml:mo>
<mml:mrow>
<mml:mfrac>
<mml:mrow>
<mml:msub>
<mml:mrow>
<mml:mtext>BIC</mml:mtext>
</mml:mrow>
<mml:mi>i</mml:mi>
</mml:msub>
</mml:mrow>
<mml:mi>T</mml:mi>
</mml:mfrac>
</mml:mrow>
<mml:mo>)</mml:mo>
</mml:mrow>
<mml:mo>,</mml:mo>
</mml:mrow>
</mml:mtd>
<mml:mtd>
<mml:mrow>
<mml:mtext>&#xa0;if&#xa0;</mml:mtext>
<mml:mtext>&#x2009;</mml:mtext>
<mml:mtext>configuration</mml:mtext>
<mml:mtext>&#x2009;</mml:mtext>
<mml:mi>i</mml:mi>
<mml:mtext>&#x2009;</mml:mtext>
<mml:mtext>does</mml:mtext>
<mml:mtext>&#x2009;</mml:mtext>
<mml:mtext>not</mml:mtext>
<mml:mtext>&#x2009;</mml:mtext>
<mml:mtext>result</mml:mtext>
<mml:mtext>&#x2009;</mml:mtext>
<mml:mtext>in</mml:mtext>
<mml:mtext>&#x2009;</mml:mtext>
<mml:mtext>a</mml:mtext>
<mml:mtext>&#x2009;</mml:mtext>
<mml:mtext>loop</mml:mtext>
</mml:mrow>
</mml:mtd>
</mml:mtr>
<mml:mtr>
<mml:mtd>
<mml:mrow>
<mml:mn>0</mml:mn>
<mml:mo>,</mml:mo>
</mml:mrow>
</mml:mtd>
<mml:mtd>
<mml:mrow>
<mml:mtext>if&#xa0;</mml:mtext>
<mml:mtext>&#x2009;</mml:mtext>
<mml:mtext>configuration</mml:mtext>
<mml:mtext>&#x2009;</mml:mtext>
<mml:mi>i</mml:mi>
<mml:mtext>&#x2009;</mml:mtext>
<mml:mtext>result</mml:mtext>
<mml:mtext>&#x2009;</mml:mtext>
<mml:mtext>in</mml:mtext>
<mml:mtext>&#x2009;</mml:mtext>
<mml:mtext>a</mml:mtext>
<mml:mtext>&#x2009;</mml:mtext>
<mml:mtext>loop</mml:mtext>
</mml:mrow>
</mml:mtd>
</mml:mtr>
</mml:mtable>
</mml:mrow>
</mml:mrow>
<mml:mo>,</mml:mo>
</mml:mrow>
</mml:math>
<label>(1)</label>
</disp-formula>Where <inline-formula id="inf61">
<mml:math id="m62">
<mml:mrow>
<mml:mtext>exp</mml:mtext>
<mml:mrow>
<mml:mo>(</mml:mo>
<mml:mo>&#x22c5;</mml:mo>
<mml:mo>)</mml:mo>
</mml:mrow>
</mml:mrow>
</mml:math>
</inline-formula> is the exponential function, <inline-formula id="inf62">
<mml:math id="m63">
<mml:mi>T</mml:mi>
</mml:math>
</inline-formula> is temperature controlling how greedy we want the searching to be, and <inline-formula id="inf63">
<mml:math id="m64">
<mml:mo>&#x221d;</mml:mo>
</mml:math>
</inline-formula> means proportional&#x20;to.</p>
<p>The main algorithm used in the second step is as follows.</p>
<p>
<statement content-type="step" id="Step_1">
<label>Step 1</label>
<p>Sample one fully connected triplet <inline-formula id="inf64">
<mml:math id="m65">
<mml:mrow>
<mml:mrow>
<mml:mo>{</mml:mo>
<mml:mrow>
<mml:msub>
<mml:mi>X</mml:mi>
<mml:mi>I</mml:mi>
</mml:msub>
<mml:mo>,</mml:mo>
<mml:msub>
<mml:mi>X</mml:mi>
<mml:mi>J</mml:mi>
</mml:msub>
<mml:mo>,</mml:mo>
<mml:msub>
<mml:mi>X</mml:mi>
<mml:mi>K</mml:mi>
</mml:msub>
</mml:mrow>
<mml:mo>}</mml:mo>
</mml:mrow>
</mml:mrow>
</mml:math>
</inline-formula> with the least outside connection; Choose a configuration between these three nodes with probability described in <xref ref-type="disp-formula" rid="e1">Eq. 1</xref>; Then remove the connection between <inline-formula id="inf65">
<mml:math id="m66">
<mml:mrow>
<mml:msub>
<mml:mi>X</mml:mi>
<mml:mi>I</mml:mi>
</mml:msub>
</mml:mrow>
</mml:math>
</inline-formula>, <inline-formula id="inf66">
<mml:math id="m67">
<mml:mrow>
<mml:msub>
<mml:mi>X</mml:mi>
<mml:mi>J</mml:mi>
</mml:msub>
</mml:mrow>
</mml:math>
</inline-formula> and <inline-formula id="inf67">
<mml:math id="m68">
<mml:mrow>
<mml:msub>
<mml:mi>X</mml:mi>
<mml:mi>K</mml:mi>
</mml:msub>
</mml:mrow>
</mml:math>
</inline-formula> from the skeleton.</p>
</statement>
</p>
<p>
<statement content-type="step" id="Step_2">
<label>Step 2</label>
<p>Repeat step 1 until all fully connected triplets are sampled.</p>
</statement>
</p>
<p>
<statement content-type="step" id="Step_3">
<label>Step 3</label>
<p>Sample one partially connected triplet <inline-formula id="inf68">
<mml:math id="m69">
<mml:mrow>
<mml:mrow>
<mml:mo>{</mml:mo>
<mml:mrow>
<mml:msub>
<mml:mi>X</mml:mi>
<mml:mi>I</mml:mi>
</mml:msub>
<mml:mo>,</mml:mo>
<mml:msub>
<mml:mi>X</mml:mi>
<mml:mi>J</mml:mi>
</mml:msub>
<mml:mo>,</mml:mo>
<mml:msub>
<mml:mi>X</mml:mi>
<mml:mi>K</mml:mi>
</mml:msub>
</mml:mrow>
<mml:mo>}</mml:mo>
</mml:mrow>
</mml:mrow>
</mml:math>
</inline-formula> with the least outside connection. Choose a configuration between these three nodes with probability described in <xref ref-type="disp-formula" rid="e1">Eq. 1</xref>. Then remove the connection between <inline-formula id="inf69">
<mml:math id="m70">
<mml:mrow>
<mml:msub>
<mml:mi>X</mml:mi>
<mml:mi>I</mml:mi>
</mml:msub>
</mml:mrow>
</mml:math>
</inline-formula>, <inline-formula id="inf70">
<mml:math id="m71">
<mml:mrow>
<mml:msub>
<mml:mi>X</mml:mi>
<mml:mi>J</mml:mi>
</mml:msub>
</mml:mrow>
</mml:math>
</inline-formula> and <inline-formula id="inf71">
<mml:math id="m72">
<mml:mrow>
<mml:msub>
<mml:mi>X</mml:mi>
<mml:mi>K</mml:mi>
</mml:msub>
</mml:mrow>
</mml:math>
</inline-formula> (if applicable) from the skeleton.</p>
</statement>
</p>
<p>
<statement content-type="step" id="Step_4">
<label>Step 4</label>
<p>Repeat step 3 until all partially connected triplets are sampled.</p>
</statement>
</p>
<p>
<statement content-type="step" id="Step_5">
<label>Step 5</label>
<p>Sample one pair <inline-formula id="inf72">
<mml:math id="m73">
<mml:mrow>
<mml:mrow>
<mml:mo>{</mml:mo>
<mml:mrow>
<mml:msub>
<mml:mi>X</mml:mi>
<mml:mi>I</mml:mi>
</mml:msub>
<mml:mo>,</mml:mo>
<mml:msub>
<mml:mi>X</mml:mi>
<mml:mi>J</mml:mi>
</mml:msub>
</mml:mrow>
<mml:mo>}</mml:mo>
</mml:mrow>
</mml:mrow>
</mml:math>
</inline-formula> with the least outside connection. Choose a configuration between them (either<inline-formula id="inf73">
<mml:math id="m74">
<mml:mrow>
<mml:mtext>&#xa0;</mml:mtext>
<mml:msub>
<mml:mi>X</mml:mi>
<mml:mi>I</mml:mi>
</mml:msub>
<mml:mo>&#x2192;</mml:mo>
<mml:msub>
<mml:mi>X</mml:mi>
<mml:mi>J</mml:mi>
</mml:msub>
<mml:mo>&#xa0;</mml:mo>
</mml:mrow>
</mml:math>
</inline-formula>or<inline-formula id="inf74">
<mml:math id="m75">
<mml:mrow>
<mml:mtext>&#xa0;</mml:mtext>
<mml:msub>
<mml:mi>X</mml:mi>
<mml:mi>I</mml:mi>
</mml:msub>
<mml:mo>&#x2190;</mml:mo>
<mml:msub>
<mml:mi>X</mml:mi>
<mml:mi>J</mml:mi>
</mml:msub>
</mml:mrow>
</mml:math>
</inline-formula>) with probability described in <xref ref-type="disp-formula" rid="e1">Eq. 1</xref>. Then remove the connection between <inline-formula id="inf75">
<mml:math id="m76">
<mml:mrow>
<mml:msub>
<mml:mi>X</mml:mi>
<mml:mi>I</mml:mi>
</mml:msub>
</mml:mrow>
</mml:math>
</inline-formula> and <inline-formula id="inf76">
<mml:math id="m77">
<mml:mrow>
<mml:msub>
<mml:mi>X</mml:mi>
<mml:mi>J</mml:mi>
</mml:msub>
</mml:mrow>
</mml:math>
</inline-formula> from the skeleton.</p>
</statement>
</p>
<p>
<statement content-type="step" id="Step_6">
<label>Step 6</label>
<p>Repeat step 5 until all pairs are sampled and no more unsampled edges in the skeleton.</p>
<p>Since each SMC sample is generated independently, we can run our algorithm in parallel on multiple CPUs/GPU cores to speed up the sampling process.</p>
</statement>
</p>
</sec>
<sec id="s2-4">
<title>Third Stage: Reclaiming Missed Edges</title>
<p>We mentioned earlier that one disadvantage of the traditional two-stage method was that the edges missed in the first stage will never be recovered. Therefore, in the third stage we designed a Random Order Hill Climbing (ROHC) method to identify the possible missed edges and refine the network. The general idea is described as follows:<list list-type="simple">
<list-item>
<p>1. Generate a permutation of <inline-formula id="inf77">
<mml:math id="m78">
<mml:mrow>
<mml:mn>1,2</mml:mn>
<mml:mo>,</mml:mo>
<mml:mo>&#x2026;</mml:mo>
<mml:mo>,</mml:mo>
<mml:mo>&#xa0;</mml:mo>
<mml:mi>p</mml:mi>
</mml:mrow>
</mml:math>
</inline-formula> for each network sampled by adSMC in stage 2, suppose <inline-formula id="inf78">
<mml:math id="m79">
<mml:mrow>
<mml:msub>
<mml:mi>m</mml:mi>
<mml:mn>1</mml:mn>
</mml:msub>
<mml:mo>,</mml:mo>
<mml:mo>&#x2026;</mml:mo>
<mml:mo>,</mml:mo>
<mml:mo>&#xa0;</mml:mo>
<mml:msub>
<mml:mi>m</mml:mi>
<mml:mi>p</mml:mi>
</mml:msub>
</mml:mrow>
</mml:math>
</inline-formula> is such a permutation.</p>
</list-item>
<list-item>
<p>2. For every node <inline-formula id="inf79">
<mml:math id="m80">
<mml:mrow>
<mml:msub>
<mml:mi>X</mml:mi>
<mml:mi>i</mml:mi>
</mml:msub>
</mml:mrow>
</mml:math>
</inline-formula>, iterate <italic>j</italic> from <inline-formula id="inf80">
<mml:math id="m81">
<mml:mrow>
<mml:msub>
<mml:mi>m</mml:mi>
<mml:mn>1</mml:mn>
</mml:msub>
</mml:mrow>
</mml:math>
</inline-formula> through <inline-formula id="inf81">
<mml:math id="m82">
<mml:mrow>
<mml:msub>
<mml:mi>m</mml:mi>
<mml:mi>p</mml:mi>
</mml:msub>
</mml:mrow>
</mml:math>
</inline-formula>. If <inline-formula id="inf82">
<mml:math id="m83">
<mml:mrow>
<mml:msub>
<mml:mi>X</mml:mi>
<mml:mi>i</mml:mi>
</mml:msub>
<mml:mo>&#x2190;</mml:mo>
<mml:msub>
<mml:mi>X</mml:mi>
<mml:mi>j</mml:mi>
</mml:msub>
</mml:mrow>
</mml:math>
</inline-formula> does not create loop and result in an increasing in BIC, we add edge <inline-formula id="inf83">
<mml:math id="m84">
<mml:mrow>
<mml:msub>
<mml:mi>X</mml:mi>
<mml:mi>i</mml:mi>
</mml:msub>
<mml:mo>&#x2190;</mml:mo>
<mml:msub>
<mml:mi>X</mml:mi>
<mml:mi>j</mml:mi>
</mml:msub>
</mml:mrow>
</mml:math>
</inline-formula>.</p>
</list-item>
<list-item>
<p>3. Repeat (2) until there is no possible edge to add or the searching limit is reached.</p>
</list-item>
</list>
</p>
<p>One could also view this stage as a further ascent to the local optima to ensure we achieve the best possible BIC&#x20;score.</p>
<p>In `general, generating more SMC samples gives a higher chance to reach the optimum. However, more samples also require more computation time; therefore, a balance between running time and sample sizes must be made. In most of our simulation study and practical problems, we found that around 20,000 samples were often good enough for finding a network with a satisfactory BIC&#x20;score.</p>
</sec>
<sec id="s2-5">
<title>Performance Evaluation</title>
<p>To measure the effectiveness of edge screening methods, we employed the precision, recall and f-score measurements. Precision is defined as TP/(TP &#x2b; FP), recall is defined as TP/(TP &#x2b; FN), and f-score is the harmonic mean of precision and recall, 2 (precision &#xd7; <inline-formula id="inf84">
<mml:math id="m85">
<mml:mo>&#xd7;</mml:mo>
</mml:math>
</inline-formula>recall)/(precision &#x2b; recall), where TP means true positive (number of true undirected edges identified), FP false positive (number of non-edges identified as undirected edges), and FN false negative (number of undirected edges not identified).</p>
<p>In our study, recall measures the percentage of true edges (irrespective of their directions) identified; therefore, it is the most important measurement in edge screening stage, since as we discussed earlier, any missed edges in stage one may never be reclaimed in a traditional two stage approach. Besides the recall, f-score is also important since it measures a balanced performance in terms of both precision and recall. It is obvious that if we propose all possible edges, we will always identify all true edges, but that will not do any pruning to the searching space. Thus, a high f-score is desired for a decent edge screening strategy.</p>
<p>We used Bayesian Information Criterion (BIC) as the score function in both second stage and third stage. BIC has the score-equivalent property (Appendix definition 10), which can reduce the searching space, since if we could find one network in the equivalent class, we found the true network. And the consistency property of BIC score guarantees that the true network has the highest score asymptotically.</p>
</sec>
<sec id="s2-6">
<title>Benchmark Networks</title>
<p>The networks used to generate simulated data (<xref ref-type="table" rid="T1">Table&#x20;1</xref>) are from actual decision making processes of a wide range of real applications including risk management, tech support, and disease diagnosis. All networks are obtained from Bayesian Network Repository maintained by M. Scutari <ext-link ext-link-type="uri" xlink:href="http://www.bnlearn.com/bnrepository/">http://www.bnlearn.com/bnrepository/</ext-link>.</p>
<table-wrap id="T1" position="float">
<label>TABLE 1</label>
<caption>
<p>Bayesian networks used in the simulation&#x20;study.</p>
</caption>
<table>
<thead valign="top">
<tr>
<th align="left">Name</th>
<th align="center">&#x23; of nodes</th>
<th align="center">&#x23; of edges</th>
<th align="center">&#x23; of parameters</th>
<th align="center">Max in-degree</th>
</tr>
</thead>
<tbody valign="top">
<tr>
<td align="left">Alarm</td>
<td align="char" char=".">37</td>
<td align="char" char=".">46</td>
<td align="char" char=".">509</td>
<td align="char" char=".">4</td>
</tr>
<tr>
<td align="left">Andes</td>
<td align="char" char=".">223</td>
<td align="char" char=".">338</td>
<td align="char" char=".">1,157</td>
<td align="char" char=".">6</td>
</tr>
<tr>
<td align="left">Child</td>
<td align="char" char=".">20</td>
<td align="char" char=".">25</td>
<td align="char" char=".">230</td>
<td align="char" char=".">2</td>
</tr>
<tr>
<td align="left">Hailfinder</td>
<td align="char" char=".">56</td>
<td align="char" char=".">66</td>
<td align="char" char=".">2,656</td>
<td align="char" char=".">4</td>
</tr>
<tr>
<td align="left">Hepar2</td>
<td align="char" char=".">70</td>
<td align="char" char=".">1,236</td>
<td align="char" char=".">1,453</td>
<td align="char" char=".">6</td>
</tr>
<tr>
<td align="left">Insurance</td>
<td align="char" char=".">27</td>
<td align="char" char=".">52</td>
<td align="char" char=".">984</td>
<td align="char" char=".">3</td>
</tr>
<tr>
<td align="left">Win95pts</td>
<td align="char" char=".">76</td>
<td align="char" char=".">112</td>
<td align="char" char=".">574</td>
<td align="char" char=".">7</td>
</tr>
</tbody>
</table>
</table-wrap>
<p>We randomly generated data with 1,000, 2,000, and 5,000 observations, and we generated 10 datasets for each size of observations. All results reported in this section are based on averages of 10 datasets. Observation size in this article refers to the number of data points, and shall not be confused with number of sequential Monte Carlo samples. The datasets were generated using R package <italic>bnlearn</italic> (<xref ref-type="bibr" rid="B33">Scutari 2009</xref>; <xref ref-type="bibr" rid="B30">Nagarajan, Scutari and L&#xe8;bre 2013</xref>).</p>
</sec>
<sec id="s2-7">
<title>Real Data</title>
<sec id="s2-7-1">
<title>Flow Cytometry Dataset</title>
<p>In the flow cytometry dataset (<xref ref-type="bibr" rid="B32">Sachs, Perez, Pe&#x2019;er, Lauffenburger and Nolan 2005</xref>), there are 11 proteins and phospholipid components of the signaling network. The original data was collected from 7,466 cells, containing continuous variables. Sachs et&#x20;al. suggested to get rid of the potential outliers by removing data that are three standard deviations away from any attribute. Thus the data we are analyzing contains 6,814 observations. We discretized each variable into three categories, practically stands for high/medium/low, with each category containing 33% of the&#x20;data.</p>
</sec>
<sec id="s2-7-2">
<title>Genomics and Epigenomics Data From the Cancer Genome Atlas (TCGA)</title>
<p>We used several different types of data obtained from TCGA: RNA-seq, protein expression, DNA methylation, and microRNA-seq, which have been used in our previous studies (<xref ref-type="bibr" rid="B35">Stewart, Luks, Roycik, Sang and Zhang 2013</xref>; <xref ref-type="bibr" rid="B25">Li, et&#x20;al., 2017</xref>; <xref ref-type="bibr" rid="B34">Shi, et&#x20;al., 2017</xref>; <xref ref-type="bibr" rid="B24">Li, et&#x20;al., 2020</xref>). These data can be freely downloaded from TCGA data portal (<ext-link ext-link-type="uri" xlink:href="https://portal.gdc.cancer.gov/">https://portal.gdc.cancer.gov/</ext-link>), which has detailed description on each of the data&#x20;types.</p>
</sec>
</sec>
</sec>
<sec sec-type="results" id="s3">
<title>Results</title>
<sec id="s3-1">
<title>Edge Screening</title>
<p>The principal of the edge screening stage is pruning the searching space as much as possible while the remaining edges in the pruned space still possess as many true edges as possible. We compare our method to five other methods including max-min parent-child (mmpc) (<xref ref-type="bibr" rid="B39">Tsamardinos, et&#x20;al., 2006</xref>), grow-shrink (gs) (<xref ref-type="bibr" rid="B29">Margaritis 2003</xref>), incremental association (iamb) (<xref ref-type="bibr" rid="B38">Tsamardinos, et&#x20;al., 2003</xref>), fast iamb, and inter iamb (<xref ref-type="bibr" rid="B42">Yaramakala, et&#x20;al., 2005</xref>). For all methods, we fixed the significance level (&#x3b1;) to&#x20;0.01.</p>
<p>The simulation study results (<xref ref-type="fig" rid="F2">Figure&#x20;2</xref> and Figure S2) showed that our double filtering (DF) method was able to identify the most edges (highest recall) for each of the observation size we tested. In some cases we observed that with even 1,000 observations, our method achieved a higher recall than the other methods using 5,000 observations and the f-scores are still comparable (e.g., Alarm, Hepar2 and etc.). For some networks (Child, Insurance), not only the recalls were higher but also the f-scores were higher for DF. The results confirmed that DF identifies true edges more accurately than other methods and it often requires fewer observations. Higher recall is desired in the first stage (the edge screening stage) since any missed edges will not be sampled in the second&#x20;stage.</p>
<fig id="F2" position="float">
<label>FIGURE 2</label>
<caption>
<p>Recall and f1 score of different methods with observation size 1,000. One can see that DF generally has higher recalls with higher or comparable F1-scores for the same network.</p>
</caption>
<graphic xlink:href="fgene-12-764020-g002.tif"/>
</fig>
<sec id="s3-1-1">
<title>Effect of Temperature</title>
<p>The temperature parameter in SMC has the same effect as that in MCMC (Markov Chain Monte Carlo) simulations. A lower temperature will cause searching to become greedier, and higher temperatures make it less greedy. When <inline-formula id="inf85">
<mml:math id="m86">
<mml:mrow>
<mml:mi>T</mml:mi>
<mml:mo>&#x2192;</mml:mo>
<mml:mn>0</mml:mn>
</mml:mrow>
</mml:math>
</inline-formula> the searching procedure becomes a local greedy search. On the other hand, when<inline-formula id="inf86">
<mml:math id="m87">
<mml:mrow>
<mml:mtext>T</mml:mtext>
<mml:mo>&#x2192;</mml:mo>
<mml:mi>&#x221e;</mml:mi>
</mml:mrow>
</mml:math>
</inline-formula>, the configuration is sampled uniformly. The optimal temperature is usually a value in between. In this simulation study, we fixed SMC sample size to 20,000, and rounds of ROHC to 5. The temperature was set to between <inline-formula id="inf87">
<mml:math id="m88">
<mml:mrow>
<mml:msup>
<mml:mrow>
<mml:mn>10</mml:mn>
</mml:mrow>
<mml:mrow>
<mml:mo>&#x2212;</mml:mo>
<mml:mn>7</mml:mn>
</mml:mrow>
</mml:msup>
</mml:mrow>
</mml:math>
</inline-formula> and<inline-formula id="inf88">
<mml:math id="m89">
<mml:mrow>
<mml:mo>&#xa0;</mml:mo>
<mml:msup>
<mml:mrow>
<mml:mn>10</mml:mn>
</mml:mrow>
<mml:mrow>
<mml:mo>&#x2212;</mml:mo>
<mml:mn>1</mml:mn>
</mml:mrow>
</mml:msup>
</mml:mrow>
</mml:math>
</inline-formula>, increased by 10-time each time (Figure S3). The performance is shown in the relative scale (BIC of true network/BIC of the learned network), where higher ratio means higher BIC score; thus, better network structure. Lower temperature in most cases gave a lower score, as well as the higher temperature, consistent with what we would expect. Most of the optimal scores happened around T &#x3d; 0.001 or 0.01. We can also see that the optimal temperature does not depend on the observation sizes, since the optimal temperatures are the same across the three different observation sizes. Another observation we had was that the optimal temperatures do not change much when the number of variables (nodes) changes. From figure S3 we can see that for Andes (with 223 nodes) and child (20 nodes), the optimal temperature is both around 0.01 and&#x20;0.001.</p>
</sec>
<sec id="s3-1-2">
<title>Effect of Adaptive SMC</title>
<p>To show the improvement of using adaptive SMC, we compared the BICs of 20,000 SMC samples between the adSMC and traditional SMC (Figure S4). In the traditional SMC, we designed the sampling block in the order of fully connected triplets, partially connected triplets and pairs, and started from least outside connected ones. Clearly, the adSMC generates higher scored networks in general.</p>
</sec>
<sec id="s3-1-3">
<title>Effect of the Edge Reclaiming Step</title>
<p>We discussed earlier that there could be some true edges missed in the first stage due to the test power and limited data. Here we will show that Random Order Hill Climbing (ROHC) indeed improves the learned BN structure in stage 2. We used <italic>alarm</italic> and <italic>win95pts</italic> networks to illustrate the improvement made by ROHC (Figure S5). They both had significance level cut-off of 0.01, temperature 0.001, and 20,000 SMC samples. As we can see, the improvements were substantial, demonstrating that it is necessary to have the third stage to further refine the learned network. However, one should notice that the complexity level of ROHC is approximately <inline-formula id="inf89">
<mml:math id="m90">
<mml:mrow>
<mml:mi>O</mml:mi>
<mml:mrow>
<mml:mo>(</mml:mo>
<mml:mrow>
<mml:msup>
<mml:mi>N</mml:mi>
<mml:mn>2</mml:mn>
</mml:msup>
</mml:mrow>
<mml:mo>)</mml:mo>
</mml:mrow>
</mml:mrow>
</mml:math>
</inline-formula>; therefore, in a typical network with hundreds of nodes only 1 or 2 rounds of ROHC are affordable.</p>
</sec>
<sec id="s3-1-4">
<title>Performance on Benchmark Networks</title>
<p>We evaluated the overall performance of our method and the general two stage methods (five edge screening methods, gs, mmpc, iamb, fast.iamb, and inter.iamb combined with two optimization methods, Hill climbing and tabu search) on seven benchmark networks. The results are shown in <xref ref-type="fig" rid="F3">Figure&#x20;3</xref> and <xref ref-type="sec" rid="s10">Supplementary Figure S6</xref> (Supplementary Material). For three different observation sizes, our method outperformed all the general two-stage methods on almost all benchmark networks except on the hepar2 network where all methods achieved similar scores, which are very close to the BIC of the true network.</p>
<fig id="F3" position="float">
<label>FIGURE 3</label>
<caption>
<p>BIC scores of all methods on seven benchmark networks with observation size 1,000. GRASP has higher BIC scores for all the benchmark networks.</p>
</caption>
<graphic xlink:href="fgene-12-764020-g003.tif"/>
</fig>
</sec>
<sec id="s3-1-5">
<title>Performance on the Flow Cytometry Data</title>
<p>We first compared our method to the general 2-stage methods and the CD method (<xref ref-type="bibr" rid="B13">Fu, et&#x20;al., 2013</xref>) on the flow cytometry data. GRASP achieved the highest BIC score (<xref ref-type="fig" rid="F4">Figure&#x20;4</xref>), which is consistent with the simulation&#x20;study.</p>
<fig id="F4" position="float">
<label>FIGURE 4</label>
<caption>
<p>BIC scores for the flow cytometry data, comparing 12 methods, and GRASP has the highest BIC score. The y-axis value is the ratio of the BIC score of the sampled network and the true network. It is possible that a sampled network has even higher BIC score than the true network, hence the value can be higher than 1.</p>
</caption>
<graphic xlink:href="fgene-12-764020-g004.tif"/>
</fig>
</sec>
<sec id="s3-1-6">
<title>An Integrative Genomic Study Using TCGA Data</title>
<p>An advantage of BN models is that they can handle heterogeneous data well. In this section, we will test our method using a heterogeneous genomics dataset from TCGA through learning network structures that may shed light on real biological problems. In a previous study of ours (<xref ref-type="bibr" rid="B35">Stewart, et&#x20;al., 2013</xref>), we have identified a long non-coding RNA, LOC90784, which is strongly associated with breast cancer health disparity between African American and Caucasian American breast cancer patients. However, literature search resulted in no information about it since it had not been studied by any researchers in the past. Using several different types of genomics data, we applied GRASP to perform an integrative study to build a Bayesian network with different genomics features to shed some light on the function of this transcript. All the data were first discretized into a small number of categories, usually 2&#x2013;4. We first used RNA-seq data to identify transcripts highly correlated with LOC90784. This gave us eight transcripts with absolute value of correlation coefficient greater than 0.27. We then found other genomic features, including microRNAs, DNA methylations and protein expressions that are highly correlated with these transcripts, which gave us 13 microRNAs, 5 DNA methylation regions (aggregated around genes) and five proteins. Using the samples with all the above measurements, we inferred the BN structure for these genomics features as shown in <xref ref-type="fig" rid="F5">Figure&#x20;5</xref>. As a comparison, bnlearn, a R package for BN structure learning, gave a network without LOC90784 (<xref ref-type="sec" rid="s10">Supplementary Figure S7</xref>). <xref ref-type="fig" rid="F5">Figure&#x20;5</xref> showed rather complex relationships among all these genomic features. A thorough investigation of this network is beyond the scope of this work. However, some literature search on the nodes around LOC90784 provided interesting hypotheses, which could be followed up with experiments. Specifically, TET3, an upstream gene, was found to inhibit TGF-&#x3b2;1-induced epithelial-mesenchymal transition in ovarian cancer cells (<xref ref-type="bibr" rid="B43">Ye, et&#x20;al., 2016</xref>). High frequency of PIK3R2 mutations in endometrial cancer was found to be related to the regulation of protein stability of PTEN (<xref ref-type="bibr" rid="B5">Cheung, et&#x20;al., 2011</xref>), which is a well-known cancer related gene. There are not a lot of published studies on IRGQ. From the Human Protein Atlas database (<ext-link ext-link-type="uri" xlink:href="https://www.proteinatlas.org/ENSG00000167378-IRGQ/pathology">https://www.proteinatlas.org/ENSG00000167378-IRGQ/pathology</ext-link>) we found that this gene is a prognostic biomarker and significant for survival for several cancer types including pancreatic cancer, renal cancer, cervical cancer and liver cancer. It would be interesting to see how perturbations of TET3, PIK3R2, such as knockdown/knockout experiments, affect LOC90784 and how perturbation of LOC90784 affects IRGQ. These hypotheses demonstrated the potential of GRASP for discovering new biology through integrative genomic studies.</p>
<fig id="F5" position="float">
<label>FIGURE 5</label>
<caption>
<p>The BN structure learned by GRASP using multiple different genomic features which are highly correlated with the expression of LOC90784. Orange nodes: mRNA transcripts; Red nodes: microRNAs; Blue nodes: protein expressions; Green nodes: DNA methylations.</p>
</caption>
<graphic xlink:href="fgene-12-764020-g005.tif"/>
</fig>
</sec>
</sec>
</sec>
<sec sec-type="discussion" id="s4">
<title>Discussion</title>
<p>In this study, we developed a three-stage Bayesian network structure learning method, GRASP. The first stage is an edge screening method, Double Filtering, which recovers a super set of true edges and proposes as few edges as possible. The second stage is an adaptive SMC (adSMC) approach to optimize a score function (BIC in this study) that measures the fitness of a BN structure to the given data. To reclaim the possible missed edges from the first two stages, we developed a random order hill climbing method (ROHC) to recover the missed edges as the last stage. The principal of double filtering is quite different from the well-known mmpc method or other similar constraint-based methods, where the algorithm is trying to identify the skeleton of the BN (undirected true edges). Double filtering method focuses on identifying a set of undirected edges that contains all the true edges, at the same time tries to propose as few edges as possible. The advantage of mmpc is that given enough observations it identifies the true network skeleton; however, it may not be feasible when the observations are limited since mmpc conducts conditional dependency test conditioning on all previously identified dependent (connected) nodes, and it requires more observations when the number of conditioned nodes increases. On the other hand, double filtering only conditions on one node at a time, so the required observation size can be much smaller.</p>
<p>The adSMC approach in structure sampling stage can find better BN structure than greedy searching algorithms or traditional SMC. The algorithm takes into account the currently sampled partial BN structures to make more informed decisions on the sampling of new edges. In addition, adSMC sampling is completely parallelizable, and multiple CPUs/GPU implementations will likely further improve the computational efficiency substantially.</p>
<p>Although in this study we focused on categorical variables (nodes) with multinomial distribution, one may extend our approach to other types of variables including Gaussian ones, as long as all nodes have the same distribution and the local conditional distribution can be estimated. Imposing distributions that are easier to be estimated on the nodes will in general make the searching more efficient. Practically, it is not an easy task to find appropriate distribution for all nodes. For BNs with mixed node types, where nodes do not necessarily have the same distribution, our method could handle them indirectly by discretizing the observations making each node distributed as multinomial distribution.</p>
<p>The application of GRASP on heterogeneous genomics data showed its potential to infer complex biological networks, which may shed light on the functions of unknown genes or epigenetic features. The learned structures of BN also provide guidance on formulating specific hypotheses that can be tested experimentally.</p>
</sec>
</body>
<back>
<sec id="s5">
<title>Data Availability Statement</title>
<p>Publicly available datasets were analyzed in this study. Benchmark Bayesian networks can be found at: <ext-link ext-link-type="uri" xlink:href="https://www.bnlearn.com/bnrepository/">https://www.bnlearn.com/bnrepository/</ext-link>. Genomics data analyzed can be found at: <ext-link ext-link-type="uri" xlink:href="https://portal.gdc.cancer.gov/">https://portal.gdc.cancer.gov/</ext-link>.</p>
</sec>
<sec id="s6">
<title>Author Contributions</title>
<p>KY designed the study, performed the research, analyzed the results, and wrote the paper; ZC analyzed the results, XS analyzed the results, XQ revised the paper, JZ designed the study and wrote the paper.</p>
</sec>
<sec id="s7">
<title>Funding</title>
<p>This work was supported by the National Institute of General Medical Sciences of the National Institutes of Health under award number R01GM126558. The funders had no role in study design, data collection and analysis, decision to publish, or preparation of the manuscript.</p>
</sec>
<sec sec-type="COI-statement" id="s8">
<title>Conflict of Interest</title>
<p>The authors declare that the research was conducted in the absence of any commercial or financial relationships that could be construed as a potential conflict of interest.</p>
</sec>
<sec sec-type="disclaimer" id="s9">
<title>Publisher&#x2019;s Note</title>
<p>All claims expressed in this article are solely those of the authors and do not necessarily represent those of their affiliated organizations, or those of the publisher, the editors and the reviewers. Any product that may be evaluated in this article, or claim that may be made by its manufacturer, is not guaranteed or endorsed by the publisher.</p>
</sec>
<sec id="s10">
<title>Supplementary material</title>
<p>The supplementary material for this article can be found online at: <ext-link ext-link-type="uri" xlink:href="https://www.frontiersin.org/articles/10.3389/fgene.2021.764020/full#supplementary-material">https://www.frontiersin.org/articles/10.3389/fgene.2021.764020/full&#x23;supplementary-material</ext-link>
</p>
<supplementary-material xlink:href="DataSheet1.docx" id="SM1" mimetype="application/docx" xmlns:xlink="http://www.w3.org/1999/xlink"/>
</sec>
<ref-list>
<title>References</title>
<ref id="B1">
<citation citation-type="journal">
<person-group person-group-type="author">
<name>
<surname>Adabor</surname>
<given-names>E. S.</given-names>
</name>
<name>
<surname>Acquaah-Mensah</surname>
<given-names>G. K.</given-names>
</name>
<name>
<surname>Oduro</surname>
<given-names>F. T.</given-names>
</name>
</person-group> (<year>2015</year>). <article-title>Saga: A Hybrid Search Algorithm for Bayesian Network Structure Learning of Transcriptional Regulatory Networks</article-title>. <source>J.&#x20;Biomed. Inform.</source> <volume>53</volume>, <fpage>27</fpage>&#x2013;<lpage>35</lpage>. <pub-id pub-id-type="doi">10.1016/j.jbi.2014.08.010</pub-id> </citation>
</ref>
<ref id="B2">
<citation citation-type="journal">
<person-group person-group-type="author">
<name>
<surname>Aliferis</surname>
<given-names>C. F.</given-names>
</name>
<name>
<surname>Statnikov</surname>
<given-names>A.</given-names>
</name>
<name>
<surname>Tsamardinos</surname>
<given-names>I.</given-names>
</name>
<name>
<surname>Mani</surname>
<given-names>S.</given-names>
</name>
<name>
<surname>Koutsoukos</surname>
<given-names>X. D.</given-names>
</name>
</person-group> (<year>2010</year>). <article-title>Local Causal and Markov Blanket Induction for Causal Discovery and Feature Selection for Classification Part I: Algorithms and Empirical Evaluation</article-title>. <source>J.&#x20;Mach. Learn. Res.</source> <volume>11</volume>, <fpage>171</fpage>&#x2013;<lpage>234</lpage>. <pub-id pub-id-type="doi">10.1145&#x002F;175600.61756013</pub-id> </citation>
</ref>
<ref id="B3">
<citation citation-type="journal">
<person-group person-group-type="author">
<name>
<surname>Amirkhani</surname>
<given-names>H.</given-names>
</name>
<name>
<surname>Rahmati</surname>
<given-names>M.</given-names>
</name>
<name>
<surname>Lucas</surname>
<given-names>P. J.&#x20;F.</given-names>
</name>
<name>
<surname>Hommersom</surname>
<given-names>A.</given-names>
</name>
</person-group> (<year>2017</year>). <article-title>Exploiting Experts&#x27; Knowledge for Structure Learning of Bayesian Networks</article-title>. <source>IEEE Trans. Pattern Anal. Mach. Intell.</source> <volume>39</volume>, <fpage>2154</fpage>&#x2013;<lpage>2170</lpage>. <pub-id pub-id-type="doi">10.1109/tpami.2016.2636828</pub-id> </citation>
</ref>
<ref id="B4">
<citation citation-type="journal">
<person-group person-group-type="author">
<name>
<surname>Campos</surname>
<given-names>L. M. d.</given-names>
</name>
<name>
<surname>Campos</surname>
<given-names>L. M. d.</given-names>
</name>
</person-group> (<year>1998</year>). <article-title>Independency Relationships and Learning Algorithms for Singly Connected NetworksA Scoring Function for Learning Bayesian Networks Based on Mutual Information and Conditional Independence Tests</article-title>. <source>J.&#x20;Machine Learn. Res.</source> <volume>7</volume>, <fpage>2149</fpage>&#x2013;<lpage>2187</lpage>. <pub-id pub-id-type="doi">10.1080/095281398146743</pub-id> </citation>
</ref>
<ref id="B5">
<citation citation-type="journal">
<person-group person-group-type="author">
<name>
<surname>Cheung</surname>
<given-names>L. W. T.</given-names>
</name>
<name>
<surname>Hennessy</surname>
<given-names>B. T.</given-names>
</name>
<name>
<surname>Li</surname>
<given-names>J.</given-names>
</name>
<name>
<surname>Yu</surname>
<given-names>S.</given-names>
</name>
<name>
<surname>Myers</surname>
<given-names>A. P.</given-names>
</name>
<name>
<surname>Djordjevic</surname>
<given-names>B.</given-names>
</name>
<etal/>
</person-group> (<year>2011</year>). <article-title>High Frequency of Pik3r1 and Pik3r2 Mutations in Endometrial Cancer Elucidates a Novel Mechanism for Regulation of Pten Protein Stability</article-title>. <source>Cancer Discov.</source> <volume>1</volume>, <fpage>170</fpage>&#x2013;<lpage>185</lpage>. <pub-id pub-id-type="doi">10.1158/2159-8290.cd-11-0039</pub-id> </citation>
</ref>
<ref id="B6">
<citation citation-type="journal">
<person-group person-group-type="author">
<name>
<surname>Cooper</surname>
<given-names>G. F.</given-names>
</name>
</person-group> (<year>1990</year>). <article-title>The Computational Complexity of Probabilistic Inference Using Bayesian Belief Networks</article-title>. <source>Artif. Intelligence</source> <volume>42</volume>, <fpage>393</fpage>&#x2013;<lpage>405</lpage>. <pub-id pub-id-type="doi">10.1016/0004-3702(90)90060-d</pub-id> </citation>
</ref>
<ref id="B7">
<citation citation-type="journal">
<person-group person-group-type="author">
<name>
<surname>de Campos</surname>
<given-names>L. M.</given-names>
</name>
<name>
<surname>Huete</surname>
<given-names>J.&#x20;F.</given-names>
</name>
</person-group> (<year>2000</year>). <article-title>A New Approach for Learning Belief Networks Using Independence Criteria</article-title>. <source>Int. J.&#x20;Approximate Reasoning</source> <volume>24</volume>, <fpage>11</fpage>&#x2013;<lpage>37</lpage>. <pub-id pub-id-type="doi">10.1016/s0888-613x(99)00042-0</pub-id> </citation>
</ref>
<ref id="B8">
<citation citation-type="journal">
<person-group person-group-type="author">
<name>
<surname>Denoyer</surname>
<given-names>L.</given-names>
</name>
<name>
<surname>Gallinari</surname>
<given-names>P.</given-names>
</name>
</person-group> (<year>2004</year>). <article-title>Bayesian Network Model for Semi-structured Document Classification</article-title>. <source>Inf. Process. Manage.</source> <volume>40</volume>, <fpage>807</fpage>&#x2013;<lpage>827</lpage>. <pub-id pub-id-type="doi">10.1016/j.ipm.2004.04.009</pub-id> </citation>
</ref>
<ref id="B9">
<citation citation-type="journal">
<person-group person-group-type="author">
<name>
<surname>Ferreira-Santos</surname>
<given-names>D.</given-names>
</name>
<name>
<surname>Monteiro-Soares</surname>
<given-names>M.</given-names>
</name>
<name>
<surname>Rodrigues</surname>
<given-names>P. P.</given-names>
</name>
</person-group> (<year>2018</year>). <article-title>Impact of Imputing Missing Data in Bayesian Network Structure Learning for Obstructive Sleep Apnea Diagnosis</article-title>. <source>Stud. Health Technol. Inform.</source> <volume>247</volume>, <fpage>126</fpage>&#x2013;<lpage>130</lpage>. </citation>
</ref>
<ref id="B10">
<citation citation-type="journal">
<person-group person-group-type="author">
<name>
<surname>Franzin</surname>
<given-names>A.</given-names>
</name>
<name>
<surname>Sambo</surname>
<given-names>F.</given-names>
</name>
<name>
<surname>Di Camillo</surname>
<given-names>B.</given-names>
</name>
</person-group> (<year>2017</year>). <article-title>Bnstruct: An R Package for Bayesian Network Structure Learning in the Presence of Missing Data</article-title>. <source>Bioinformatics</source> <volume>33</volume>, <fpage>1250</fpage>&#x2013;<lpage>1252</lpage>. <pub-id pub-id-type="doi">10.1093/bioinformatics/btw807</pub-id> </citation>
</ref>
<ref id="B11">
<citation citation-type="journal">
<person-group person-group-type="author">
<name>
<surname>Friedman</surname>
<given-names>N.</given-names>
</name>
<name>
<surname>Linial</surname>
<given-names>M.</given-names>
</name>
<name>
<surname>Nachman</surname>
<given-names>I.</given-names>
</name>
<name>
<surname>Pe&#x27;er</surname>
<given-names>D.</given-names>
</name>
</person-group> (<year>2000</year>). <article-title>Using Bayesian Networks to Analyze Expression Data</article-title>. <source>J.&#x20;Comput. Biol.</source> <volume>7</volume>, <fpage>601</fpage>&#x2013;<lpage>620</lpage>. <pub-id pub-id-type="doi">10.1089/106652700750050961</pub-id> </citation>
</ref>
<ref id="B12">
<citation citation-type="book">
<person-group person-group-type="author">
<name>
<surname>Friedman</surname>
<given-names>N.</given-names>
</name>
<name>
<surname>Nachman</surname>
<given-names>I.</given-names>
</name>
<name>
<surname>Pe&#xe9;r</surname>
<given-names>D.</given-names>
</name>
</person-group> (<year>1999</year>). &#x201c;<article-title>Learning Bayesian Network Structure from Massive Datasets: The &#xab;Sparse Candidate &#xab;Algorithm</article-title>&#x201d;, In. <conf-name>Proceedings of the Fifteenth conference on Uncertainty in artificial intelligence</conf-name>, <conf-date>30 July 1999</conf-date>, <conf-loc>Stockholm, Sweden</conf-loc>, <publisher-name>Morgan Kaufmann Publishers Inc.</publisher-name>, <fpage>206</fpage>&#x2013;<lpage>215</lpage>. </citation>
</ref>
<ref id="B13">
<citation citation-type="journal">
<person-group person-group-type="author">
<name>
<surname>Fu</surname>
<given-names>F.</given-names>
</name>
<name>
<surname>Zhou</surname>
<given-names>Q.</given-names>
</name>
</person-group> (<year>2013</year>). <article-title>Learning Sparse Causal Gaussian Networks with Experimental Intervention: Regularization and Coordinate Descent</article-title>. <source>J.&#x20;Am. Stat. Assoc.</source> <volume>108</volume>, <fpage>288</fpage>&#x2013;<lpage>300</lpage>. <pub-id pub-id-type="doi">10.1080/01621459.2012.754359</pub-id> </citation>
</ref>
<ref id="B14">
<citation citation-type="journal">
<person-group person-group-type="author">
<name>
<surname>G&#xe1;mez</surname>
<given-names>J.&#x20;A.</given-names>
</name>
<name>
<surname>Mateo</surname>
<given-names>J.&#x20;L.</given-names>
</name>
<name>
<surname>Puerta</surname>
<given-names>J.&#x20;M.</given-names>
</name>
</person-group> (<year>2011</year>). <article-title>Learning Bayesian Networks by Hill Climbing: Efficient Methods Based on Progressive Restriction of the Neighborhood</article-title>. <source>Data Min Knowl Disc</source> <volume>22</volume>, <fpage>106</fpage>&#x2013;<lpage>148</lpage>. <pub-id pub-id-type="doi">10.1007/s10618-010-0178-6</pub-id> </citation>
</ref>
<ref id="B15">
<citation citation-type="journal">
<person-group person-group-type="author">
<name>
<surname>Grassberger</surname>
<given-names>P.</given-names>
</name>
</person-group> (<year>1997</year>). <article-title>Pruned-enriched Rosenbluth Method: Simulations Of&#x3b8;polymers of Chain Length up to 1&#x20;000 000</article-title>. <source>Phys. Rev. E</source> <volume>56</volume>, <fpage>3682</fpage>&#x2013;<lpage>3693</lpage>. <pub-id pub-id-type="doi">10.1103/physreve.56.3682</pub-id> </citation>
</ref>
<ref id="B16">
<citation citation-type="journal">
<person-group person-group-type="author">
<name>
<surname>Han</surname>
<given-names>S.</given-names>
</name>
<name>
<surname>Zhang</surname>
<given-names>H.</given-names>
</name>
<name>
<surname>Homayouni</surname>
<given-names>R.</given-names>
</name>
<name>
<surname>Karmaus</surname>
<given-names>W.</given-names>
</name>
</person-group> (<year>2017</year>). <article-title>An Efficient Bayesian Approach for Gaussian Bayesian Network Structure Learning</article-title>. <source>Commun. Stat. - Simulation Comput.</source> <volume>46</volume>, <fpage>5070</fpage>&#x2013;<lpage>5084</lpage>. <pub-id pub-id-type="doi">10.1080/03610918.2016.1143103</pub-id> </citation>
</ref>
<ref id="B17">
<citation citation-type="journal">
<person-group person-group-type="author">
<name>
<surname>Huang</surname>
<given-names>S.</given-names>
</name>
<name>
<surname>Li</surname>
<given-names>J.</given-names>
</name>
<name>
<surname>Ye</surname>
<given-names>J.</given-names>
</name>
<name>
<surname>Fleisher</surname>
<given-names>A.</given-names>
</name>
<name>
<surname>Chen</surname>
<given-names>K.</given-names>
</name>
<name>
<surname>Wu</surname>
<given-names>T.</given-names>
</name>
<etal/>
</person-group> (<year>2013</year>). <article-title>A Sparse Structure Learning Algorithm for Gaussian Bayesian Network Identification from High-Dimensional Data</article-title>. <source>IEEE Trans. Pattern Anal. Mach. Intell.</source> <volume>35</volume>, <fpage>1328</fpage>&#x2013;<lpage>1342</lpage>. <pub-id pub-id-type="doi">10.1109/tpami.2012.129</pub-id> </citation>
</ref>
<ref id="B18">
<citation citation-type="journal">
<person-group person-group-type="author">
<name>
<surname>Jabbari</surname>
<given-names>B.</given-names>
</name>
<name>
<surname>Visweswaran</surname>
<given-names>S.</given-names>
</name>
<name>
<surname>Cooper</surname>
<given-names>G. F.</given-names>
</name>
</person-group> (<year>2018</year>). <article-title>Botulinum Toxin Treatment in Children</article-title>. <source>Proc. Mach Learn. Res.</source> <volume>72</volume>, <fpage>169</fpage>&#x2013;<lpage>180</lpage>. <pub-id pub-id-type="doi">10.1007/978-3-319-99945-6_14</pub-id> </citation>
</ref>
<ref id="B19">
<citation citation-type="book">
<person-group person-group-type="author">
<name>
<surname>Koller</surname>
<given-names>D.</given-names>
</name>
<name>
<surname>Friedman</surname>
<given-names>N.</given-names>
</name>
</person-group> (<year>2009</year>). <source>Probabilistic Graphical Models: Principles and Techniques</source>. <publisher-loc>Cambridge, MA</publisher-loc>: <publisher-name>MIT Press.</publisher-name> </citation>
</ref>
<ref id="B20">
<citation citation-type="journal">
<person-group person-group-type="author">
<name>
<surname>Kristensen</surname>
<given-names>K.</given-names>
</name>
<name>
<surname>Rasmussen</surname>
<given-names>I. A.</given-names>
</name>
</person-group> (<year>2002</year>). <article-title>The Use of a Bayesian Network in the Design of a Decision Support System for Growing Malting Barley without Use of Pesticides</article-title>. <source>Comput. Electro. Agric.</source> <volume>33</volume>, <fpage>197</fpage>&#x2013;<lpage>217</lpage>. <pub-id pub-id-type="doi">10.1016/s0168-1699(02)00007-8</pub-id> </citation>
</ref>
<ref id="B21">
<citation citation-type="journal">
<person-group person-group-type="author">
<name>
<surname>Larjo</surname>
<given-names>A.</given-names>
</name>
<name>
<surname>L&#xe4;hdesm&#xe4;ki</surname>
<given-names>H.</given-names>
</name>
</person-group> (<year>2015</year>). <article-title>Using Multi-step Proposal Distribution for Improved Mcmc Convergence in Bayesian Network Structure Learning</article-title>. <source>EURASIP J.&#x20;Bioinform Syst. Biol.</source> <volume>2015</volume>, <fpage>6</fpage>. <pub-id pub-id-type="doi">10.1186/s13637-015-0024-7</pub-id> </citation>
</ref>
<ref id="B22">
<citation citation-type="journal">
<person-group person-group-type="author">
<name>
<surname>Larra&#xf1;aga</surname>
<given-names>P.</given-names>
</name>
<name>
<surname>Poza</surname>
<given-names>M.</given-names>
</name>
<name>
<surname>Yurramendi</surname>
<given-names>Y.</given-names>
</name>
<name>
<surname>Murga</surname>
<given-names>R. H.</given-names>
</name>
<name>
<surname>Kuijpers</surname>
<given-names>C. M. H.</given-names>
</name>
</person-group> (<year>1996</year>). <article-title>Structure Learning of Bayesian Networks by Genetic Algorithms: A Performance Analysis of Control Parameters</article-title>. <source>IEEE Trans. Pattern Anal. Machine Intell.</source> <volume>18</volume>, <fpage>912</fpage>&#x2013;<lpage>926</lpage>. <pub-id pub-id-type="doi">10.1109/34.537345</pub-id> </citation>
</ref>
<ref id="B51">
<citation citation-type="journal">
<person-group person-group-type="author">
<name>
<surname>Liang</surname>
<given-names>J.</given-names>
</name>
<name>
<surname>Zhang</surname>
<given-names>J.</given-names>
</name>
<name>
<surname>Chen</surname>
<given-names>R.</given-names>
</name>
</person-group> (<year>2002</year>). <article-title>Statistical Geometry of Packing Defects of Lattice Chain Polymer from Enumeration and Sequential Monte Carlo Method</article-title>. <source>J. Chem. Phys.</source> <volume>117</volume>, <fpage>3511</fpage>&#x2013;<lpage>3521</lpage>. </citation>
</ref>
<ref id="B23">
<citation citation-type="book">
<person-group person-group-type="author">
<name>
<surname>Li</surname>
<given-names>H.</given-names>
</name>
<name>
<surname>Guo</surname>
<given-names>H.</given-names>
</name>
</person-group> (<year>2018</year>). <source>A Hybrid Structure Learning Algorithm for Bayesian Network Using Experts&#x27; Knowledge</source>. <publisher-loc>Basel)</publisher-loc>: <publisher-name>Entropy</publisher-name>, <fpage>20</fpage>. </citation>
</ref>
<ref id="B25">
<citation citation-type="journal">
<person-group person-group-type="author">
<name>
<surname>Li</surname>
<given-names>Y.</given-names>
</name>
<name>
<surname>Steppi</surname>
<given-names>A.</given-names>
</name>
<name>
<surname>Zhou</surname>
<given-names>Y.</given-names>
</name>
<name>
<surname>Mao</surname>
<given-names>F.</given-names>
</name>
<name>
<surname>Miller</surname>
<given-names>P. C.</given-names>
</name>
<name>
<surname>He</surname>
<given-names>M. M.</given-names>
</name>
<etal/>
</person-group> (<year>2017</year>). <article-title>Tumoral Expression of Drug and Xenobiotic Metabolizing Enzymes in Breast Cancer Patients of Different Ethnicities with Implications to Personalized Medicine</article-title>. <source>Sci. Rep.</source> <volume>7</volume>, <fpage>4747</fpage>. <pub-id pub-id-type="doi">10.1038/s41598-017-04250-2</pub-id> </citation>
</ref>
<ref id="B26">
<citation citation-type="book">
<person-group person-group-type="author">
<name>
<surname>Liu</surname>
<given-names>J.</given-names>
</name>
</person-group> (<year>2008</year>). <source>Monte Carlo Strategies in Scientific Computing</source>. <publisher-name>Springer</publisher-name> </citation>
</ref>
<ref id="B27">
<citation citation-type="journal">
<person-group person-group-type="author">
<name>
<surname>Liu</surname>
<given-names>J.&#x20;S.</given-names>
</name>
<name>
<surname>Chen</surname>
<given-names>R.</given-names>
</name>
</person-group> (<year>1998</year>). <article-title>Sequential Monte Carlo Methods for Dynamic Systems</article-title>. <source>J.&#x20;Am. Stat. Assoc.</source> <volume>93</volume>, <fpage>1032</fpage>&#x2013;<lpage>1044</lpage>. <pub-id pub-id-type="doi">10.1080/01621459.1998.10473765</pub-id> </citation>
</ref>
<ref id="B28">
<citation citation-type="book">
<person-group person-group-type="author">
<name>
<surname>Liu</surname>
<given-names>X.</given-names>
</name>
<name>
<surname>Gao</surname>
<given-names>X.</given-names>
</name>
<name>
<surname>Wang</surname>
<given-names>Z.</given-names>
</name>
<name>
<surname>Ru</surname>
<given-names>X.</given-names>
</name>
</person-group> (<year>2021</year>). <source>Improved Local Search with Momentum for Bayesian Networks Structure Learning</source>. <publisher-loc>Basel)</publisher-loc>: <publisher-name>Entropy</publisher-name>, <fpage>23</fpage> </citation>
</ref>
<ref id="B29">
<citation citation-type="book">
<person-group person-group-type="author">
<name>
<surname>Margaritis</surname>
<given-names>D.</given-names>
</name>
</person-group> (<year>2003</year>). <source>Learning Bayesian Network Model Structure from Data</source>. <publisher-name>US Army.</publisher-name> </citation>
</ref>
<ref id="B30">
<citation citation-type="book">
<person-group person-group-type="author">
<name>
<surname>Nagarajan</surname>
<given-names>R.</given-names>
</name>
<name>
<surname>Scutari</surname>
<given-names>M.</given-names>
</name>
<name>
<surname>L&#xe8;bre</surname>
<given-names>S.</given-names>
</name>
</person-group> (<year>2013</year>). <source>Bayesian Networks in R</source>. <publisher-loc>New York</publisher-loc>: <publisher-name>Springer</publisher-name> </citation>
</ref>
<ref id="B31">
<citation citation-type="journal">
<person-group person-group-type="author">
<name>
<surname>Raval</surname>
<given-names>A.</given-names>
</name>
<name>
<surname>Ghahramani</surname>
<given-names>Z.</given-names>
</name>
<name>
<surname>Wild</surname>
<given-names>D. L.</given-names>
</name>
</person-group> (<year>2002</year>). <article-title>A Bayesian Network Model for Protein Fold and Remote Homologue Recognition</article-title>. <source>Bioinformatics</source> <volume>18</volume>, <fpage>788</fpage>&#x2013;<lpage>801</lpage>. <pub-id pub-id-type="doi">10.1093/bioinformatics/18.6.788</pub-id> </citation>
</ref>
<ref id="B32">
<citation citation-type="journal">
<person-group person-group-type="author">
<name>
<surname>Sachs</surname>
<given-names>K.</given-names>
</name>
<name>
<surname>Perez</surname>
<given-names>O.</given-names>
</name>
<name>
<surname>Pe&#x27;er</surname>
<given-names>D.</given-names>
</name>
<name>
<surname>Lauffenburger</surname>
<given-names>D. A.</given-names>
</name>
<name>
<surname>Nolan</surname>
<given-names>G. P.</given-names>
</name>
</person-group> (<year>2005</year>). <article-title>Causal Protein-Signaling Networks Derived from Multiparameter Single-Cell Data</article-title>. <source>Science</source> <volume>308</volume>, <fpage>523</fpage>&#x2013;<lpage>529</lpage>. <pub-id pub-id-type="doi">10.1126/science.1105809</pub-id> </citation>
</ref>
<ref id="B33">
<citation citation-type="journal">
<person-group person-group-type="author">
<name>
<surname>Scutari</surname>
<given-names>M.</given-names>
</name>
</person-group> (<year>2009</year>). <article-title>Learning Bayesian Networks with the Bnlearn R Package</article-title>. <comment>arXiv preprint arXiv:0908.3817</comment> </citation>
</ref>
<ref id="B34">
<citation citation-type="journal">
<person-group person-group-type="author">
<name>
<surname>Shi</surname>
<given-names>Y.</given-names>
</name>
<name>
<surname>Steppi</surname>
<given-names>A.</given-names>
</name>
<name>
<surname>Cao</surname>
<given-names>Y.</given-names>
</name>
<name>
<surname>Wang</surname>
<given-names>J.</given-names>
</name>
<name>
<surname>He</surname>
<given-names>M. M.</given-names>
</name>
<name>
<surname>Li</surname>
<given-names>L.</given-names>
</name>
<etal/>
</person-group> (<year>2017</year>). <article-title>Integrative Comparison of Mrna Expression Patterns in Breast Cancers from Caucasian and Asian Americans with Implications for Precision Medicine</article-title>. <source>Cancer Res.</source> <volume>77</volume>, <fpage>423</fpage>&#x2013;<lpage>433</lpage>. <pub-id pub-id-type="doi">10.1158/0008-5472.can-16-1959</pub-id> </citation>
</ref>
<ref id="B35">
<citation citation-type="journal">
<person-group person-group-type="author">
<name>
<surname>Stewart</surname>
<given-names>P. A.</given-names>
</name>
<name>
<surname>Luks</surname>
<given-names>J.</given-names>
</name>
<name>
<surname>Roycik</surname>
<given-names>M. D.</given-names>
</name>
<name>
<surname>Sang</surname>
<given-names>Q. X.</given-names>
</name>
<name>
<surname>Zhang</surname>
<given-names>J.</given-names>
</name>
</person-group> (<year>2013</year>), <article-title>Differentially Expressed Transcripts and Dysregulated Signaling Pathways and Networks in African American Breast Cancer</article-title>, <source>PLoS One</source> <volume>8</volume>. <fpage>e82460</fpage>. <pub-id pub-id-type="doi">10.1371/journal.pone.0082460</pub-id> </citation>
</ref>
<ref id="B36">
<citation citation-type="journal">
<person-group person-group-type="author">
<name>
<surname>Tang</surname>
<given-names>Y.</given-names>
</name>
<name>
<surname>Wang</surname>
<given-names>J.</given-names>
</name>
<name>
<surname>Nguyen</surname>
<given-names>M.</given-names>
</name>
<name>
<surname>Altintas</surname>
<given-names>I.</given-names>
</name>
</person-group> (<year>2019</year>), <article-title>Penbayes: A Multi-Layered Ensemble Approach for Learning Bayesian Network Structure from Big Data</article-title>, <volume>19</volume> <source>Sensors (Basel)</source>, <issue>20</issue>. <fpage>4400</fpage>. <pub-id pub-id-type="doi">10.3390/s19204400</pub-id> </citation>
</ref>
<ref id="B37">
<citation citation-type="journal">
<person-group person-group-type="author">
<name>
<surname>Teyssier</surname>
<given-names>M.</given-names>
</name>
<name>
<surname>Koller</surname>
<given-names>D.</given-names>
</name>
</person-group> (<year>2012</year>). <article-title>Ordering-Based Search: A Simple and Effective Algorithm for Learning Bayesian Networks</article-title>. <comment>arXiv preprint arXiv:1207.1429</comment> </citation>
</ref>
<ref id="B38">
<citation citation-type="book">
<person-group person-group-type="author">
<name>
<surname>Tsamardinos</surname>
<given-names>I.</given-names>
</name>
<name>
<surname>Aliferis</surname>
<given-names>C. F.</given-names>
</name>
<name>
<surname>Statnikov</surname>
<given-names>A.</given-names>
</name>
</person-group> (<year>2003</year>). &#x201c;<article-title>Time and Sample Efficient Discovery of Markov Blankets and Direct Causal Relations</article-title>&#x201d;, In. <conf-name>Proceedings of the Ninth ACM SIGKDD International Conference on Knowledge Discovery and Data Mining</conf-name>, <conf-date>24 August 2003</conf-date>, <conf-loc>Washington, D.C.US</conf-loc>, <publisher-name>Association for Computing Machinery</publisher-name>, <fpage>673</fpage>&#x2013;<lpage>678</lpage>. <pub-id pub-id-type="doi">10.1145/956750.956838</pub-id> </citation>
</ref>
<ref id="B39">
<citation citation-type="journal">
<person-group person-group-type="author">
<name>
<surname>Tsamardinos</surname>
<given-names>I.</given-names>
</name>
<name>
<surname>Brown</surname>
<given-names>L. E.</given-names>
</name>
<name>
<surname>Aliferis</surname>
<given-names>C. F.</given-names>
</name>
</person-group> (<year>2006</year>). <article-title>The Max-Min Hill-Climbing Bayesian Network Structure Learning Algorithm</article-title>. <source>Mach Learn.</source> <volume>65</volume>, <fpage>31</fpage>&#x2013;<lpage>78</lpage>. <pub-id pub-id-type="doi">10.1007/s10994-006-6889-7</pub-id> </citation>
</ref>
<ref id="B40">
<citation citation-type="journal">
<person-group person-group-type="author">
<name>
<surname>Vignes</surname>
<given-names>M.</given-names>
</name>
<name>
<surname>Vandel</surname>
<given-names>J.</given-names>
</name>
<name>
<surname>Allouche</surname>
<given-names>D.</given-names>
</name>
<name>
<surname>Ramadan-Alban</surname>
<given-names>N.</given-names>
</name>
<name>
<surname>Cierco-Ayrolles</surname>
<given-names>C.</given-names>
</name>
<name>
<surname>Schiex</surname>
<given-names>T.</given-names>
</name>
<etal/>
</person-group> (<year>2011</year>). <article-title>Gene Regulatory Network Reconstruction Using Bayesian Networks, the Dantzig Selector, the Lasso and Their Meta-Analysis</article-title>. <source>PLoS ONE</source> <volume>6</volume>, <fpage>e29165</fpage>. <pub-id pub-id-type="doi">10.1371/journal.pone.0029165</pub-id> </citation>
</ref>
<ref id="B41">
<citation citation-type="book">
<person-group person-group-type="author">
<name>
<surname>Xiang</surname>
<given-names>J.</given-names>
</name>
<name>
<surname>Kim</surname>
<given-names>S.</given-names>
</name>
</person-group> (<year>2013</year>). <source>A\\Ast Lasso for Learning a Sparse Bayesian Network Structure for Continuous Variables</source>, <fpage>2418</fpage>&#x2013;<lpage>2426</lpage>. </citation>
</ref>
<ref id="B42">
<citation citation-type="book">
<person-group person-group-type="author">
<name>
<surname>Yaramakala</surname>
<given-names>S.</given-names>
</name>
<name>
<surname>Margaritis</surname>
<given-names>D.</given-names>
</name>
</person-group> (<year>2005</year>).&#x201d;<article-title>Speculative Markov Blanket Discovery for Optimal Feature Selection</article-title>&#x201d;, In. <conf-name>Proceedings of the Data mining, fifth IEEE international conference on</conf-name>, <conf-date>27 November 2005</conf-date>, <conf-loc>NW Washington, DC, United&#x20;States</conf-loc>. <publisher-name>IEEE</publisher-name>, <fpage>4</fpage> . </citation>
</ref>
<ref id="B24">
<citation citation-type="journal">
<person-group person-group-type="author">
<name>
<surname>Yan</surname>
<given-names>L.</given-names>
</name>
<name>
<surname>Xiaodong</surname>
<given-names>P.</given-names>
</name>
<name>
<surname>Zihan</surname>
<given-names>C.</given-names>
</name>
<name>
<surname>Yidong</surname>
<given-names>Z.</given-names>
</name>
<name>
<surname>Feng</surname>
<given-names>M.</given-names>
</name>
<name>
<surname>Yan</surname>
<given-names>L.</given-names>
</name>
<etal/>
</person-group> (<year>2020</year>). <article-title>Genetic Factors Associated with Cancer Racial Disparity - an Integrative Study across Twenty-One Cancer Types</article-title>. <source>Mol. Oncol.</source> <volume>14</volume>. </citation>
</ref>
<ref id="B43">
<citation citation-type="journal">
<person-group person-group-type="author">
<name>
<surname>Ye</surname>
<given-names>Z.</given-names>
</name>
<name>
<surname>Li</surname>
<given-names>J.</given-names>
</name>
<name>
<surname>Han</surname>
<given-names>X.</given-names>
</name>
<name>
<surname>Hou</surname>
<given-names>H.</given-names>
</name>
<name>
<surname>Chen</surname>
<given-names>H.</given-names>
</name>
<name>
<surname>Zheng</surname>
<given-names>X.</given-names>
</name>
<etal/>
</person-group> (<year>2016</year>). <article-title>TET3 Inhibits TGF-&#x392;1-Induced Epithelial-Mesenchymal Transition by Demethylating miR-30d Precursor Gene in Ovarian Cancer Cells</article-title>. <source>J.&#x20;Exp. Clin. Cancer Res.</source> <volume>35</volume>, <fpage>72</fpage>. <pub-id pub-id-type="doi">10.1186/s13046-016-0350-y</pub-id> </citation>
</ref>
<ref id="B52">
<citation citation-type="journal">
<person-group person-group-type="author">
<name>
<surname>Zhang</surname>
<given-names>J.</given-names>
</name>
<name>
<surname>Chen</surname>
<given-names>R.</given-names>
</name>
<name>
<surname>Tang</surname>
<given-names>C.</given-names>
</name>
<name>
<surname>Liang</surname>
<given-names>J.</given-names>
</name>
</person-group> (<year>2003</year>). <article-title>Origin of Scaling Behavior of Protein Packing Density: A Sequential Monte Carlo Study of Compact Long Chain Polymers</article-title>. <source>J. Chem. Phys.</source> <volume>118</volume> (<issue>13</issue>), <fpage>6102</fpage>&#x2013;<lpage>6109</lpage>. </citation>
</ref>
<ref id="B53">
<citation citation-type="journal">
<person-group person-group-type="author">
<name>
<surname>Zhang</surname>
<given-names>J.</given-names>
</name>
<name>
<surname>Chen</surname>
<given-names>Y.</given-names>
</name>
<name>
<surname>Chen</surname>
<given-names>R.</given-names>
</name>
<name>
<surname>Liang</surname>
<given-names>J.</given-names>
</name>
</person-group> (<year>2004</year>). <article-title>Importance of Chirality and Reduced Flexibility of Protein Side Chains: A Study with Square and Tetrahedral Lattice Models</article-title>. <source>J. Chem. Phys.</source> <volume>121</volume>, <fpage>592</fpage>&#x2013;<lpage>603</lpage>. </citation>
</ref>
<ref id="B44">
<citation citation-type="journal">
<person-group person-group-type="author">
<name>
<surname>Zhang</surname>
<given-names>J.</given-names>
</name>
<name>
<surname>Lin</surname>
<given-names>M.</given-names>
</name>
<name>
<surname>Chen</surname>
<given-names>R.</given-names>
</name>
<name>
<surname>Liang</surname>
<given-names>J.</given-names>
</name>
<name>
<surname>Liu</surname>
<given-names>J.&#x20;S.</given-names>
</name>
</person-group> (<year>2007</year>). <article-title>Monte Carlo Sampling of Near-Native Structures of Proteins with Applications</article-title>. <source>Proteins</source> <volume>66</volume>, <fpage>61</fpage>&#x2013;<lpage>68</lpage>. <pub-id pub-id-type="doi">10.1002/prot.21203</pub-id> </citation>
</ref>
<ref id="B50">
<citation citation-type="journal">
<person-group person-group-type="author">
<name>
<surname>Zhang</surname>
<given-names>J.</given-names>
</name>
<name>
<surname>Liu</surname>
<given-names>J. S.</given-names>
</name>
</person-group> (<year>2006</year>). <article-title>On Side-chain Conformational Entropy of Proteins</article-title>. <source>PLoS Comput. Biol.</source> <volume>2</volume> (<issue>12</issue>), <fpage>e168</fpage>. <pub-id pub-id-type="doi">10.1371/journal.pcbi.0020168</pub-id> </citation>
</ref>
<ref id="B45">
<citation citation-type="journal">
<person-group person-group-type="author">
<name>
<surname>Zhang</surname>
<given-names>J.</given-names>
</name>
<name>
<surname>Dundas</surname>
<given-names>J.</given-names>
</name>
<name>
<surname>Lin</surname>
<given-names>M.</given-names>
</name>
<name>
<surname>Chen</surname>
<given-names>R.</given-names>
</name>
<name>
<surname>Wang</surname>
<given-names>W.</given-names>
</name>
<name>
<surname>Liang</surname>
<given-names>J.</given-names>
</name>
</person-group> (<year>2009</year>). <article-title>Prediction of Geometrically Feasible Three-Dimensional Structures of Pseudoknotted Rna through Free Energy Estimation</article-title>. <source>Rna</source> <volume>15</volume>, <fpage>2248</fpage>&#x2013;<lpage>2263</lpage>. <pub-id pub-id-type="doi">10.1261/rna.1723609</pub-id> </citation>
</ref>
<ref id="B46">
<citation citation-type="journal">
<person-group person-group-type="author">
<name>
<surname>Zhang</surname>
<given-names>J.&#x20;L.</given-names>
</name>
<name>
<surname>Liu</surname>
<given-names>J.&#x20;S.</given-names>
</name>
</person-group> (<year>2002</year>). <article-title>A New Sequential Importance Sampling Method and its Application to the Two-Dimensional Hydrophobic-Hydrophilic Model</article-title>. <source>J.&#x20;Chem. Phys.</source> <volume>117</volume>, <fpage>3492</fpage>&#x2013;<lpage>3498</lpage>. <pub-id pub-id-type="doi">10.1063/1.1494415</pub-id> </citation>
</ref>
<ref id="B47">
<citation citation-type="journal">
<person-group person-group-type="author">
<name>
<surname>Zhang</surname>
<given-names>L.</given-names>
</name>
<name>
<surname>Rodrigues</surname>
<given-names>L. O.</given-names>
</name>
<name>
<surname>Narain</surname>
<given-names>N. R.</given-names>
</name>
<name>
<surname>Akmaev</surname>
<given-names>V. R.</given-names>
</name>
</person-group> (<year>2020</year>). <article-title>Baicis: A Novel Bayesian Network Structural Learning Algorithm and its Comprehensive Performance Evaluation against Open-Source Software</article-title>. <source>J.&#x20;Comput. Biol.</source> <volume>27</volume>, <fpage>698</fpage>&#x2013;<lpage>708</lpage>. <pub-id pub-id-type="doi">10.1089/cmb.2019.0210</pub-id> </citation>
</ref>
<ref id="B48">
<citation citation-type="journal">
<person-group person-group-type="author">
<name>
<surname>Zhang</surname>
<given-names>Q.</given-names>
</name>
<name>
<surname>Li</surname>
<given-names>Z.</given-names>
</name>
<name>
<surname>Zhou</surname>
<given-names>C. J.</given-names>
</name>
<name>
<surname>Wei</surname>
<given-names>X. P.</given-names>
</name>
</person-group> (<year>2013</year>). <article-title>Bayesian Network Structure Learning Based on the Chaotic Particle Swarm Optimization Algorithm</article-title>. <source>Genet. Mol. Res.</source> <volume>12</volume>, <fpage>4468</fpage>&#x2013;<lpage>4479</lpage>. <pub-id pub-id-type="doi">10.4238/2013.october.10.12</pub-id> </citation>
</ref>
<ref id="B49">
<citation citation-type="book">
<person-group person-group-type="author">
<name>
<surname>Zhang</surname>
<given-names>Y.</given-names>
</name>
<name>
<surname>Wang</surname>
<given-names>L.</given-names>
</name>
<name>
<surname>Duan</surname>
<given-names>Z.</given-names>
</name>
<name>
<surname>Sun</surname>
<given-names>M.</given-names>
</name>
</person-group> (<year>2019</year>). <source>Structure Learning of Bayesian Network Based on Adaptive Thresholding</source>. <publisher-loc>Basel)</publisher-loc>: <publisher-name>Entropy</publisher-name>,&#x20;<fpage>21</fpage>. </citation>
</ref>
</ref-list>
</back>
</article>
