<?xml version="1.0" encoding="UTF-8" standalone="no"?>
<!DOCTYPE article PUBLIC "-//NLM//DTD Journal Publishing DTD v2.3 20070202//EN" "journalpublishing.dtd">
<article xmlns:mml="http://www.w3.org/1998/Math/MathML" xmlns:xlink="http://www.w3.org/1999/xlink" article-type="review-article">
<front>
<journal-meta>
<journal-id journal-id-type="publisher-id">Front. Chem.</journal-id>
<journal-title>Frontiers in Chemistry</journal-title>
<abbrev-journal-title abbrev-type="pubmed">Front. Chem.</abbrev-journal-title>
<issn pub-type="epub">2296-2646</issn>
<publisher>
<publisher-name>Frontiers Media S.A.</publisher-name>
</publisher>
</journal-meta>
<article-meta>
<article-id pub-id-type="doi">10.3389/fchem.2019.00809</article-id>
<article-categories>
<subj-group subj-group-type="heading">
<subject>Chemistry</subject>
<subj-group>
<subject>Review</subject>
</subj-group>
</subj-group>
</article-categories>
<title-group>
<article-title>Deep Learning for Deep Chemistry: Optimizing the Prediction of Chemical Patterns</article-title>
</title-group>
<contrib-group>
<contrib contrib-type="author" corresp="yes">
<name><surname>Cova</surname> <given-names>T&#x000E2;nia F. G. G.</given-names></name>
<xref ref-type="corresp" rid="c001"><sup>&#x0002A;</sup></xref>
<uri xlink:href="http://loop.frontiersin.org/people/565756/overview"/>
</contrib>
<contrib contrib-type="author" corresp="yes">
<name><surname>Pais</surname> <given-names>Alberto A. C. C.</given-names></name>
<xref ref-type="corresp" rid="c002"><sup>&#x0002A;</sup></xref>
<uri xlink:href="http://loop.frontiersin.org/people/579794/overview"/>
</contrib>
</contrib-group>
<aff><institution>Coimbra Chemistry Centre, CQC, Department of Chemistry, Faculty of Sciences and Technology, University of Coimbra</institution>, <addr-line>Coimbra</addr-line>, <country>Portugal</country></aff>
<author-notes>
<fn fn-type="edited-by"><p>Edited by: Emilio Martinez-Nunez, University of Santiago de Compostela, Spain</p></fn>
<fn fn-type="edited-by"><p>Reviewed by: Feng Wang, Swinburne University of Technology, Australia; Anakuthil Anoop, Indian Institute of Technology Kharagpur, India; Ruby Srivastava, Centre for Cellular and Molecular Biology (CCMB), India</p></fn>
<corresp id="c001">&#x0002A;Correspondence: T&#x000E2;nia F. G. G. Cova <email>tfirmino&#x00040;qui.uc.pt</email></corresp>
<corresp id="c002">Alberto A. C. C. Pais <email>pais&#x00040;qui.uc.pt</email></corresp>
<fn fn-type="other" id="fn001"><p>This article was submitted to Theoretical and Computational Chemistry, a section of the journal Frontiers in Chemistry</p></fn></author-notes>
<pub-date pub-type="epub">
<day>26</day>
<month>11</month>
<year>2019</year>
</pub-date>
<pub-date pub-type="collection">
<year>2019</year>
</pub-date>
<volume>7</volume>
<elocation-id>809</elocation-id>
<history>
<date date-type="received">
<day>16</day>
<month>07</month>
<year>2019</year>
</date>
<date date-type="accepted">
<day>11</day>
<month>11</month>
<year>2019</year>
</date>
</history>
<permissions>
<copyright-statement>Copyright &#x000A9; 2019 Cova and Pais.</copyright-statement>
<copyright-year>2019</copyright-year>
<copyright-holder>Cova and Pais</copyright-holder>
<license xlink:href="http://creativecommons.org/licenses/by/4.0/"><p>This is an open-access article distributed under the terms of the Creative Commons Attribution License (CC BY). The use, distribution or reproduction in other forums is permitted, provided the original author(s) and the copyright owner(s) are credited and that the original publication in this journal is cited, in accordance with accepted academic practice. No use, distribution or reproduction is permitted which does not comply with these terms.</p></license>
</permissions>
<abstract><p>Computational Chemistry is currently a synergistic assembly between <italic>ab initio</italic> calculations, simulation, machine learning (ML) and optimization strategies for describing, solving and predicting chemical data and related phenomena. These include accelerated literature searches, analysis and prediction of physical and quantum chemical properties, transition states, chemical structures, chemical reactions, and also new catalysts and drug candidates. The generalization of scalability to larger chemical problems, rather than specialization, is now the main principle for transforming chemical tasks in multiple fronts, for which systematic and cost-effective solutions have benefited from ML approaches, including those based on deep learning (e.g. quantum chemistry, molecular screening, synthetic route design, catalysis, drug discovery). The latter class of ML algorithms is capable of combining raw input into layers of intermediate features, enabling bench-to-bytes designs with the potential to transform several chemical domains. In this review, the most exciting developments concerning the use of ML in a range of different chemical scenarios are described. A range of different chemical problems and respective rationalization, that have hitherto been inaccessible due to the lack of suitable analysis tools, is thus detailed, evidencing the breadth of potential applications of these emerging multidimensional approaches. Focus is given to the models, algorithms and methods proposed to facilitate research on compound design and synthesis, materials design, prediction of binding, molecular activity, and soft matter behavior. The information produced by pairing Chemistry and ML, through data-driven analyses, neural network predictions and monitoring of chemical systems, allows (i) prompting the ability to understand the complexity of chemical data, (ii) streamlining and designing experiments, (ii) discovering new molecular targets and materials, and also (iv) planning or rethinking forthcoming chemical challenges. In fact, optimization engulfs all these tasks directly.</p></abstract>
<kwd-group>
<kwd>machine-learning</kwd>
<kwd>deep-learning</kwd>
<kwd>optimization</kwd>
<kwd>models</kwd>
<kwd>molecular simulation</kwd>
<kwd>chemistry</kwd>
</kwd-group>
<contract-num rid="cn001">UID/QUI/00313/2019</contract-num>
<contract-sponsor id="cn001">Funda&#x000E7;&#x000E3;o para a Ci&#x000EA;ncia e a Tecnologia<named-content content-type="fundref-id">10.13039/501100001871</named-content></contract-sponsor>
<counts>
<fig-count count="7"/>
<table-count count="0"/>
<equation-count count="0"/>
<ref-count count="171"/>
<page-count count="22"/>
<word-count count="17039"/>
</counts>
</article-meta>
</front>
<body>
<sec sec-type="intro" id="s1">
<title>Introduction</title>
<p>Patterns are ubiquitous in Chemistry. From the crystalline structures of solid forms to the branched chains of lipids, or the complex combinations of functional groups, chemical patterns determine the underlying properties of molecules and materials, essential to address important issues of societal concern. Artificial Intelligence (AI), and machine learning (ML) in particular, are committed to recognizing and learn from these patterns (Mitchell, <xref ref-type="bibr" rid="B114">2014</xref>; Rupp, <xref ref-type="bibr" rid="B130">2015</xref>; Goh et al., <xref ref-type="bibr" rid="B59">2017</xref>; Li et al., <xref ref-type="bibr" rid="B99">2017</xref>; Butler et al., <xref ref-type="bibr" rid="B21">2018</xref>; Fleming, <xref ref-type="bibr" rid="B52">2018</xref>; Gao et al., <xref ref-type="bibr" rid="B56">2018</xref>; Kishimoto et al., <xref ref-type="bibr" rid="B92">2018</xref>; Popova et al., <xref ref-type="bibr" rid="B123">2018</xref>; Aspuru-Guzik et al., <xref ref-type="bibr" rid="B7">2019</xref>; Elton et al., <xref ref-type="bibr" rid="B48">2019</xref>; Gromski et al., <xref ref-type="bibr" rid="B65">2019</xref>; Mater and Coote, <xref ref-type="bibr" rid="B106">2019</xref>; Schleder et al., <xref ref-type="bibr" rid="B139">2019</xref>; Venkatasubramanian, <xref ref-type="bibr" rid="B157">2019</xref>).</p>
<p>Recent evidence on the most interesting and challenging prospects for accelerating discoveries in various chemistry fields, reported under &#x0201C;Charting a course for chemistry&#x0201D; (Aspuru-Guzik et al., <xref ref-type="bibr" rid="B7">2019</xref>), indicate that the terms often used by the scientific community for describing the future trends in their field of research include &#x0201C;big data,&#x0201D; &#x0201C;machine learning,&#x0201D; and &#x0201C;artificial intelligence.&#x0201D;</p>
<p>It is recognized that ML is already boosting computational chemistry, at different levels. Different aspects have been affected, and it is not easy to summarize developments in a consistent way. In what follows, the main areas in which ML has been employed are enumerated. These are extracted from recent contributions, that can be regarded as complementary and providing an overall perspective of the applications. These include different approaches for (i) understanding and controlling chemical systems and related behavior (Chakravarti, <xref ref-type="bibr" rid="B23">2018</xref>; Fuchs et al., <xref ref-type="bibr" rid="B54">2018</xref>; Janet et al., <xref ref-type="bibr" rid="B81">2018</xref>; Elton et al., <xref ref-type="bibr" rid="B48">2019</xref>; Mezei and Von Lilienfeld, <xref ref-type="bibr" rid="B110">2019</xref>; Sanchez-Lengeling et al., <xref ref-type="bibr" rid="B137">2019</xref>; Venkatasubramanian, <xref ref-type="bibr" rid="B157">2019</xref>; Xu et al., <xref ref-type="bibr" rid="B166">2019</xref>; Zhang et al., <xref ref-type="bibr" rid="B168">2019</xref>), (ii) calculating, optimizing, or predicting structure-property relationships (Varnek and Baskin, <xref ref-type="bibr" rid="B156">2012</xref>; Ramakrishnan et al., <xref ref-type="bibr" rid="B125">2014</xref>; Goh et al., <xref ref-type="bibr" rid="B59">2017</xref>; Sim&#x000F5;es et al., <xref ref-type="bibr" rid="B145">2018</xref>; Chandrasekaran et al., <xref ref-type="bibr" rid="B25">2019</xref>), density functional theory (DFT) functionals, and interatomic potentials (Snyder et al., <xref ref-type="bibr" rid="B150">2012</xref>; Ramakrishnan et al., <xref ref-type="bibr" rid="B126">2015</xref>; Faber et al., <xref ref-type="bibr" rid="B50">2017</xref>; Hegde and Bowen, <xref ref-type="bibr" rid="B74">2017</xref>; Smith et al., <xref ref-type="bibr" rid="B147">2017</xref>; Pronobis et al., <xref ref-type="bibr" rid="B124">2018</xref>; Mezei and Von Lilienfeld, <xref ref-type="bibr" rid="B110">2019</xref>; Schleder et al., <xref ref-type="bibr" rid="B139">2019</xref>), (iii) driving generative models for inverse design (i.e., produce stable molecules from a set of desired properties) (White and Wilson, <xref ref-type="bibr" rid="B162">2010</xref>; Benjamin et al., <xref ref-type="bibr" rid="B14">2017</xref>; Kadurin et al., <xref ref-type="bibr" rid="B87">2017</xref>; Harel and Radinsky, <xref ref-type="bibr" rid="B70">2018</xref>; J&#x000F8;rgensen et al., <xref ref-type="bibr" rid="B86">2018b</xref>; Kang and Cho, <xref ref-type="bibr" rid="B89">2018</xref>; Li et al., <xref ref-type="bibr" rid="B101">2018b</xref>; Sanchez-Lengeling and Aspuru-Guzik, <xref ref-type="bibr" rid="B136">2018</xref>; Schneider, <xref ref-type="bibr" rid="B140">2018</xref>; Ar&#x000FA;s-Pous et al., <xref ref-type="bibr" rid="B6">2019</xref>; Freeze et al., <xref ref-type="bibr" rid="B53">2019</xref>; Jensen, <xref ref-type="bibr" rid="B83">2019</xref>), (iv) screening, synthesizing, and characterizing new compounds and materials (Ahneman et al., <xref ref-type="bibr" rid="B3">2018</xref>; Coley et al., <xref ref-type="bibr" rid="B35">2018a</xref>; Granda et al., <xref ref-type="bibr" rid="B61">2018</xref>; Segler et al., <xref ref-type="bibr" rid="B142">2018</xref>; Li and Eastgate, <xref ref-type="bibr" rid="B100">2019</xref>), (v) improving catalytic technologies and analytical tools (Li et al., <xref ref-type="bibr" rid="B99">2017</xref>; Gao et al., <xref ref-type="bibr" rid="B56">2018</xref>; Huang et al., <xref ref-type="bibr" rid="B78">2018</xref>; Durand and Fey, <xref ref-type="bibr" rid="B43">2019</xref>; Freeze et al., <xref ref-type="bibr" rid="B53">2019</xref>; Schleder et al., <xref ref-type="bibr" rid="B139">2019</xref>), (vi) developing quantum algorithms for molecular simulations, and (vii) progressing quantum sensing (Ramakrishnan et al., <xref ref-type="bibr" rid="B125">2014</xref>; Ramakrishnan and Von Lilienfeld, <xref ref-type="bibr" rid="B127">2017</xref>; Xia and Kais, <xref ref-type="bibr" rid="B165">2018</xref>; Ahn et al., <xref ref-type="bibr" rid="B2">2019</xref>; Christensen et al., <xref ref-type="bibr" rid="B31">2019</xref>; Mezei and Von Lilienfeld, <xref ref-type="bibr" rid="B110">2019</xref>; Zaspel et al., <xref ref-type="bibr" rid="B167">2019</xref>; Zhang et al., <xref ref-type="bibr" rid="B168">2019</xref>), just to name a few examples. In fact, Chemistry is a data-rich area, encompassing complex information which is often unstructured and poorly understood.</p>
<p>Deep learning (DL) approaches can also be particularly useful to solving a variety of chemical problems, including compound identification and classification, and description of soft matter behavior (Huang et al., <xref ref-type="bibr" rid="B78">2018</xref>; Jha et al., <xref ref-type="bibr" rid="B84">2018</xref>; J&#x000F8;rgensen et al., <xref ref-type="bibr" rid="B86">2018b</xref>; Popova et al., <xref ref-type="bibr" rid="B123">2018</xref>; Segler et al., <xref ref-type="bibr" rid="B142">2018</xref>; Zhou et al., <xref ref-type="bibr" rid="B169">2018</xref>; Chandrasekaran et al., <xref ref-type="bibr" rid="B25">2019</xref>; Degiacomi, <xref ref-type="bibr" rid="B41">2019</xref>; Elton et al., <xref ref-type="bibr" rid="B48">2019</xref>; Ghosh et al., <xref ref-type="bibr" rid="B58">2019</xref>; Mater and Coote, <xref ref-type="bibr" rid="B106">2019</xref>; Matsuzaka and Uesawa, <xref ref-type="bibr" rid="B107">2019</xref>; Xu et al., <xref ref-type="bibr" rid="B166">2019</xref>).</p>
<p>The design of generalized cause/effect models, and the scaling-up of the contributions that are being made, containing high-dimensional data, and following the open-science basis (i.e., completely accessible, with precise metadata and practical formats) are critical challenges, that may, however, facilitate the routine implementation of data mining in chemistry and expedite new discoveries.</p>
<p>The amount and quality of chemical data generated by experiments and simulations have been the mainstay of the new data-driven paradigm, that establishes the bridge between theory, experiment, computation, and simulation.</p>
<p>This review describes, in a critical and comprehensive way, relevant contributions carried out recently and involving the development of chemistry ML approaches. An exhaustive account of the theoretical foundations and applications published in the early years of AI and ML in Chemistry falls beyond the scope of this review. The reader is referred to Lecun et al. (<xref ref-type="bibr" rid="B96">2015</xref>), Coveney Peter et al. (<xref ref-type="bibr" rid="B38">2016</xref>), Goh et al. (<xref ref-type="bibr" rid="B59">2017</xref>), Elton et al. (<xref ref-type="bibr" rid="B48">2019</xref>), Gromski et al. (<xref ref-type="bibr" rid="B65">2019</xref>), and Mater and Coote (<xref ref-type="bibr" rid="B106">2019</xref>) for a full description of these efforts.</p>
<p>Until 10 years ago, only a few 100 studies on the use of ML in Chemistry were published, resulting from the contributions made over four decades. In 2018, ca. 8,000 articles in the Web of Science database included these keywords, corresponding to an increase in ca. 35% for just one decade. In this review, there is room to mention only a small fraction of these applications.</p>
<p>Despite the increasing number of works on the topic, the models proposed and practices carried out by chemists are entailing serious concerns (Chuang and Keiser, <xref ref-type="bibr" rid="B32">2018a</xref>). Several technical challenges, pitfalls, and potentials of ML, and also the reliability of the results, have been discussed by some authors (Ahneman et al., <xref ref-type="bibr" rid="B3">2018</xref>; Chuang and Keiser, <xref ref-type="bibr" rid="B32">2018a</xref>,<xref ref-type="bibr" rid="B33">b</xref>; Estrada et al., <xref ref-type="bibr" rid="B49">2018</xref>) corroborating some critical remarks on the fragility of purely data-based approaches (Microsoft, <xref ref-type="bibr" rid="B111">2018</xref>). &#x0201C;If data can speak for themselves, they can also lie for themselves.&#x0201D; This reflects the need for an in-depth understanding of chemical patterns, data-driven and theory-driven models, and algorithms, before their application.</p>
<p>Although significant progress has been made connecting specific neural network predictions to chemical input features, understanding how scientists should analyze and interpret these models to produce valid and conclusive assumptions about the system under study, still remains to be fully defined.</p>
<sec>
<title>Co-occurring Machine-Learning Contributions in Chemical Sciences</title>
<p>Scientific production covering ML-based approaches for dealing with chemical patterns has increased exponentially in recent years. However, the establishment and understanding of holistic, or macro insights on the major research trends in Chemistry sub-fields, are critical tasks. The challenge relies on how the analysis of these sub-fields, with thousands published works, reveals the most prominent applications supported by ML approaches (Butler et al., <xref ref-type="bibr" rid="B21">2018</xref>; Chmiela et al., <xref ref-type="bibr" rid="B28">2018</xref>; Chuang and Keiser, <xref ref-type="bibr" rid="B32">2018a</xref>; Coley et al., <xref ref-type="bibr" rid="B35">2018a</xref>; Gao et al., <xref ref-type="bibr" rid="B56">2018</xref>; Lo et al., <xref ref-type="bibr" rid="B103">2018</xref>; Panteleev et al., <xref ref-type="bibr" rid="B122">2018</xref>; Xia and Kais, <xref ref-type="bibr" rid="B165">2018</xref>; Ceriotti, <xref ref-type="bibr" rid="B22">2019</xref>; Chan et al., <xref ref-type="bibr" rid="B24">2019</xref>; Christensen et al., <xref ref-type="bibr" rid="B31">2019</xref>; Gallidabino et al., <xref ref-type="bibr" rid="B55">2019</xref>; H&#x000E4;se et al., <xref ref-type="bibr" rid="B71">2019</xref>; Iype and Urolagin, <xref ref-type="bibr" rid="B80">2019</xref>; Mezei and Von Lilienfeld, <xref ref-type="bibr" rid="B110">2019</xref>; Schleder et al., <xref ref-type="bibr" rid="B139">2019</xref>; Stein et al., <xref ref-type="bibr" rid="B151">2019a</xref>; Wang et al., <xref ref-type="bibr" rid="B158">2019</xref>).</p>
<p>In <xref ref-type="fig" rid="F1">Figure 1</xref> an overview of the information generated during the last decade and ranked in the research domain of &#x0201C;Science Technology&#x0201D; of the Web of Science database, is presented.</p>
<fig id="F1" position="float">
<label>Figure 1</label>
<caption><p>A holistic view of ML-based contributions in Chemistry. The clustering heatmap displays the relative counts of ML outcomes, within each area of Chemistry (organic, inorganic, analytical, physical, and biochemistry), in the 2008&#x02013;2019 (30 June) period. Data are expressed as fractions of the highest number of publications, including articles, reviews and books, containing specific co-occurring keywords, and following a standard normalization procedure. Hierarchical clustering with Euclidean distances and Ward linkage was performed on both Chemistry sub-fields and type of application. Co-occurrences are colored using a yellow-to-red color scheme. Highest and lowest relative contributions correspond to 1 (red) and 0 (yellow) values, respectively.</p></caption>
<graphic xlink:href="fchem-07-00809-g0001.tif"/>
</fig>
<p>The purpose of assessing the different facets of ML in Chemistry across the respective sub-fields is 3-fold: (i) to be able to quickly identify areas that have benefited most from the development and implementation of ML approaches, and those that still lack of such an optimization, as evidenced by the type of outcome, (ii) to identify the most relevant ML outcomes in each sub-field, and (iii) to assess the dynamics of ML outcomes over the 2008&#x02013;2019 period and how these are related, giving rise to relevant research trends.</p>
<p>An extensive literature search on ML contributions in 30 Chemistry sub-fields is carried out, using a global set of 270 co-occurring keywords, each composed of three main terms, <italic>machine learning, type of outcome</italic> and the <italic>sub-field</italic> in which they co-occur (e.g., first co-occurrence: <italic>Machine learning</italic> AND <italic>Quantum chemistry</italic> AND <italic>Quantum models</italic>, second co-occurrence: <italic>Machine learning</italic> AND <italic>Medicinal Chemistry</italic> AND <italic>Molecular screening</italic>). A total of 5,279 contributions (including books, articles, reviews, editorials and letters) on ML in Chemistry, with 81,248 citations, and published between 2008 and June 30, 2019, are found in the worldwide Web of Science database, corresponding to a 4-fold increase over the previous four decades. Considering the compiled data and the selected Chemistry fields (organic, inorganic, physical, analytical, and biochemical), nine different ML outcomes embracing the most frequent chemical challenges are defined, including (i) text mining and system description, (ii) quantitative structure-activity/property relationships, (iii) DFT functionals and interatomic potentials, (iv) generative models and inverse molecular design, (v) molecular screening, (vi) synthesis/characterization of new compounds and materials, (vii) catalytic technologies, (viii) analytical techniques, and (ix) quantum models, algorithms, and quantum sensing. Note how these have a strong relation with the seven overall applications presented above (i&#x02013;vii).</p>
<p>The heatmap represented in <xref ref-type="fig" rid="F1">Figure 1</xref> reflects the impact of each type of ML outcomes on Chemistry sub-fields. The analysis of co-occurring keywords is thus performed in order to find the number of publications that appeared simultaneously in the selected sub-field. This relation is established with greater or lesser impact depending on the frequency of each set of keywords in the selected time-span.</p>
<p>The natural clusters generated from the most important co-occurring relationships are also identified. Considering the dendrogram for the Chemistry sub-fields, it can be observed that these are organized in two main groups, which discriminates, in general, classical Chemistry sub-fields (organic, inorganic, and physical) from analytical and biochemical sub-fields. This structure suggests a significant similarity in the type of ML outcomes within each group. Group 1 have benefitted from a significant production on catalytic technologies, DFT functionals and interatomic potentials, quantum models and quantum sensing. The most representative ML outcomes in group 2 are associated to text mining, analytical techniques, generative models and inverse design, molecular screening, structure activity relationships, and synthesis of new compounds and materials. Examination of the similarity between the type of ML outcomes reveals that there are three main groups, corresponding to (i) text mining, analytical techniques, generative modes and inverse design, and molecular screening (group 1), (ii) structure-activity relationships and synthesis of new compounds and materials (group 2), and (iii) catalytic technologies, DFT functionals and interatomic potentials, and quantum models and quantum sensing (group 3).</p>
<p>Historically, researchers have introduced numerical approximations to Schr&#x000F6;dinger&#x00027;s equation, and the popular DFT calculations in <italic>ab initio</italic> approaches. However, the computational cost inherent to these classical approximations have limited the size, flexibility, and extensibility of the studies. Larger searches on relevant chemical patterns, have been successfully conducted since several research groups have developed ML models and algorithms to predict chemical properties using training data generated by DFT, which have also contributed to the increase of public collections of molecules coupled with vibrational, thermodynamic and DFT computed electronic properties (e.g., Behler and Parrinello, <xref ref-type="bibr" rid="B13">2007</xref>; Rupp et al., <xref ref-type="bibr" rid="B132">2012</xref>; Behler, <xref ref-type="bibr" rid="B12">2016</xref>; Hegde and Bowen, <xref ref-type="bibr" rid="B74">2017</xref>; Pronobis et al., <xref ref-type="bibr" rid="B124">2018</xref>; Chandrasekaran et al., <xref ref-type="bibr" rid="B25">2019</xref>; Iype and Urolagin, <xref ref-type="bibr" rid="B80">2019</xref>; Marques et al., <xref ref-type="bibr" rid="B105">2019</xref>; Schleder et al., <xref ref-type="bibr" rid="B139">2019</xref>).</p>
<p>Based on the heatmap it can be determined that groups of Chemistry sub-fields have similar, but distinct ML-based contributions.</p>
<p>The increase in chemical data and scientific documents has boosted data mining and text mining processes to manage the huge amount of chemical information and to extract useful and non-trivial knowledge in different scenarios (Krallinger et al., <xref ref-type="bibr" rid="B95">2017</xref>).</p>
<p>It is interesting to inspect if certain ML outcomes are produced in combination with each other.</p>
<p>In this context, the strongest correlation (0.97), shown in <xref ref-type="fig" rid="F2">Figure 2</xref>, is observed between text mining and molecular screening, which is to be expected as a large number of molecules has been collected and screened systematically, by combining different text mining processes and chemoinformatics techniques (e.g., pharmacophore-based similarity and docking). These integrated approaches have allowed (i) extracting and collecting, in a systematic and high-throughput way, the available chemical and biological information from different sources (e.g., scientific documents) (Krallinger et al., <xref ref-type="bibr" rid="B95">2017</xref>; Grzybowski et al., <xref ref-type="bibr" rid="B66">2018</xref>), (ii) predicting activity based on chemical structure (Granda et al., <xref ref-type="bibr" rid="B61">2018</xref>; Sim&#x000F5;es et al., <xref ref-type="bibr" rid="B145">2018</xref>; Ar&#x000FA;s-Pous et al., <xref ref-type="bibr" rid="B6">2019</xref>; Gromski et al., <xref ref-type="bibr" rid="B65">2019</xref>; Lee et al., <xref ref-type="bibr" rid="B97">2019</xref>; Li and Eastgate, <xref ref-type="bibr" rid="B100">2019</xref>), and (iii) selecting promising molecular targets and candidates for further experimental validation (e.g., <italic>in vitro</italic> tests) (Ramakrishnan et al., <xref ref-type="bibr" rid="B125">2014</xref>; Gupta et al., <xref ref-type="bibr" rid="B67">2018</xref>; Segler et al., <xref ref-type="bibr" rid="B142">2018</xref>; Brown et al., <xref ref-type="bibr" rid="B20">2019</xref>; Elton et al., <xref ref-type="bibr" rid="B48">2019</xref>; Li and Eastgate, <xref ref-type="bibr" rid="B100">2019</xref>; Schleder et al., <xref ref-type="bibr" rid="B139">2019</xref>; Xu et al., <xref ref-type="bibr" rid="B166">2019</xref>).</p>
<fig id="F2" position="float">
<label>Figure 2</label>
<caption><p>Pairwise Pearson correlations between the different types of ML outcomes in Chemistry, produced in the 2008&#x02013;2019 (30 June) period (darker colors reflect stronger correlations).</p></caption>
<graphic xlink:href="fchem-07-00809-g0002.tif"/>
</fig>
<p>Other strong correlations are found between generative models &#x00026; inverse design and the two abovementioned ML applications, molecular screening (0.95) and text mining (0.93). This can be explained by the fact that many researchers have proposed machine learning frameworks based on a variety of generative models for modeling molecules, which differ in the respective model structure and in the selected input features (Kadurin et al., <xref ref-type="bibr" rid="B87">2017</xref>; Gupta et al., <xref ref-type="bibr" rid="B67">2018</xref>; J&#x000F8;rgensen et al., <xref ref-type="bibr" rid="B86">2018b</xref>; Ar&#x000FA;s-Pous et al., <xref ref-type="bibr" rid="B6">2019</xref>; Brown et al., <xref ref-type="bibr" rid="B20">2019</xref>; Jensen, <xref ref-type="bibr" rid="B83">2019</xref>; Xu et al., <xref ref-type="bibr" rid="B166">2019</xref>).</p>
<p>Also relevant are the correlations between generative models and inverse design and synthesis of new compounds and materials (0.90), and between generative models and inverse design and analytical techniques (0.85). The former relation evidences the significant effort that has been made on applying ML models, in particular those based on accurate DL architectures, to find and select lead molecules (e.g., drugs), displaying desired properties (Varnek and Baskin, <xref ref-type="bibr" rid="B156">2012</xref>; Mitchell, <xref ref-type="bibr" rid="B114">2014</xref>; Rupp, <xref ref-type="bibr" rid="B130">2015</xref>; Lo et al., <xref ref-type="bibr" rid="B103">2018</xref>). These properties are to be translated into a more simplified information on the molecular structures, and encoded into the respective chemical fingerprint (i.e., a set of binary characteristics of molecules). The process continues with the screening of the available databases for finding molecules that possess similar fingerprints to the generated ones. Generative models and deep neural networks (DLNs) have thus allowed generating molecules and promising candidates for useful drugs, basically from scratch, making it possible to &#x0201C;design perfect needles instead of searching for a needle in a haystack&#x0201D; (White and Wilson, <xref ref-type="bibr" rid="B162">2010</xref>; Benjamin et al., <xref ref-type="bibr" rid="B14">2017</xref>; G&#x000F3;mez-Bombarelli et al., <xref ref-type="bibr" rid="B60">2018</xref>; Harel and Radinsky, <xref ref-type="bibr" rid="B70">2018</xref>; Kang and Cho, <xref ref-type="bibr" rid="B89">2018</xref>; Li et al., <xref ref-type="bibr" rid="B101">2018b</xref>; Merk et al., <xref ref-type="bibr" rid="B109">2018</xref>; Nouira et al., <xref ref-type="bibr" rid="B121">2018</xref>; Popova et al., <xref ref-type="bibr" rid="B123">2018</xref>; Sanchez-Lengeling and Aspuru-Guzik, <xref ref-type="bibr" rid="B136">2018</xref>; Schneider, <xref ref-type="bibr" rid="B140">2018</xref>).</p>
<p>It is also observed that there are other ML contributions that are interrelated: structure activity relationships with (i) molecular screening and (0.84), (ii) synthesis/characterization of new compounds and materials (0.78), and (iii) generative models and inverse design (0.75), DFT functionals and interatomic potentials with quantum models and quantum sensing (0.83), and synthesis/characterization of new compounds and materials with analytical techniques (0.79).</p>
<p>Both generative models and analytical techniques have been extensively used in the qualitative/quantitative search of patterns underlying chemical systems (Elton et al., <xref ref-type="bibr" rid="B48">2019</xref>; Ghosh et al., <xref ref-type="bibr" rid="B58">2019</xref>; Stein et al., <xref ref-type="bibr" rid="B151">2019a</xref>,<xref ref-type="bibr" rid="B152">b</xref>). It should be noted the use data from large repositories (e.g., Protein Data Bank and Cambridge Structural Database) and ML methods are not new (Hiller et al., <xref ref-type="bibr" rid="B75">1973</xref>; Gasteiger and Zupan, <xref ref-type="bibr" rid="B57">1993</xref>; Behler, <xref ref-type="bibr" rid="B12">2016</xref>). The latter have been employed as classification tools in pioneering works, encompassing, for e.g., the analysis of spectra (Thomsen and Meyer, <xref ref-type="bibr" rid="B155">1989</xref>), quantification of structure-activity relationships (QSARs) (Agrafiotis et al., <xref ref-type="bibr" rid="B1">2002</xref>), and prediction of binding sites of biomolecules (Keil et al., <xref ref-type="bibr" rid="B91">2004</xref>).</p>
<p>The range of ML applications is now quite extended as a result of a deep integration of ML in analytical, theoretical and computational chemistry. Despite of some initial skepticism in understanding the foundations and structure of ML methods, their use has been accelerated and maturated in recent years essentially due to their suitability to new applications and industry needs, including chemical and pharmaceutical sectors.</p>
</sec>
</sec>
<sec id="s2">
<title>Machine Learning For Optimization: Challenges and Opportunities</title>
<p>Designing models from chemical observations to study, control, and improve chemical processes and properties is the basis of optimization approaches. The understanding of chemical systems, and the respective underlying behavior, mechanisms and dynamics, is currently facilitated by the development of descriptive, interpretative, and predictive models, i.e., approximations that represent the target system or process. Applications of such models have included the (i) optimization of reaction parameters and process conditions, e.g., changing the type of reagents, catalysts, and solvents, and also varying systematically, concentration, addition rate, time, temperature, or solvent polarity, (ii) suggestion of new reactions based on critical functional groups, (iii) prediction of reaction/catalyst design, and optimization of heterogeneous/homogeneous catalytic reactions, (iv) acceleration and discovery of new process strategies for batch reactions, (v) establishment of trade-offs in the reaction rate and yield of organic compounds, (vi) description and maximization of the production rate and conversion efficiency of chemical reactions, (vii) prediction of the potential toxicity of different compounds, and also the (viii) rational design of target molecules and guided exploration of chemical space (Kowalik et al., <xref ref-type="bibr" rid="B94">2012</xref>; Houben and Lapkin, <xref ref-type="bibr" rid="B76">2015</xref>; Houben et al., <xref ref-type="bibr" rid="B77">2015</xref>; Zielinski et al., <xref ref-type="bibr" rid="B171">2017</xref>; H&#x000E4;se et al., <xref ref-type="bibr" rid="B72">2018</xref>; Min et al., <xref ref-type="bibr" rid="B113">2018</xref>; Zhou et al., <xref ref-type="bibr" rid="B169">2018</xref>; Ahn et al., <xref ref-type="bibr" rid="B2">2019</xref>; Choi et al., <xref ref-type="bibr" rid="B30">2019</xref>; Gromski et al., <xref ref-type="bibr" rid="B65">2019</xref>; Matsuzaka and Uesawa, <xref ref-type="bibr" rid="B107">2019</xref>).</p>
<p>ML provides the tools to scrutinize and extract useful information to be employed in modeling and system-solving solutions (Artrith and Urban, <xref ref-type="bibr" rid="B5">2016</xref>; Ward and Wolverton, <xref ref-type="bibr" rid="B159">2017</xref>). In Chemistry domains, researchers have had access to multidimensional data of unprecedented scale and accuracy, that characterize the systems/processes to be modeled. A collection of different examples of optimization based on ML approaches can be found in Kowalik et al. (<xref ref-type="bibr" rid="B94">2012</xref>); Houben and Lapkin (<xref ref-type="bibr" rid="B76">2015</xref>); Houben et al. (<xref ref-type="bibr" rid="B77">2015</xref>); Cort&#x000E9;s-Borda et al. (<xref ref-type="bibr" rid="B37">2016</xref>); Wei et al. (<xref ref-type="bibr" rid="B160">2016</xref>); Benjamin et al. (<xref ref-type="bibr" rid="B14">2017</xref>); Ahneman et al. (<xref ref-type="bibr" rid="B3">2018</xref>); Gao et al. (<xref ref-type="bibr" rid="B56">2018</xref>); Granda et al. (<xref ref-type="bibr" rid="B61">2018</xref>); Min et al. (<xref ref-type="bibr" rid="B113">2018</xref>); Ahn et al. (<xref ref-type="bibr" rid="B2">2019</xref>); Elton et al. (<xref ref-type="bibr" rid="B48">2019</xref>); Matsuzaka and Uesawa (<xref ref-type="bibr" rid="B107">2019</xref>).</p>
<p>Specifically, ML contributions have involved a variety of systems including drugs (Griffen et al., <xref ref-type="bibr" rid="B63">2018</xref>), polymers (Li et al., <xref ref-type="bibr" rid="B98">2018a</xref>), polypeptides (Grisoni et al., <xref ref-type="bibr" rid="B64">2018</xref>; M&#x000FC;ller et al., <xref ref-type="bibr" rid="B120">2018</xref>), energetic materials (Elton et al., <xref ref-type="bibr" rid="B47">2018</xref>), metal organic frameworks (He et al., <xref ref-type="bibr" rid="B73">2018</xref>; J&#x000F8;rgensen et al., <xref ref-type="bibr" rid="B85">2018a</xref>; Shen et al., <xref ref-type="bibr" rid="B144">2018</xref>), and organic solar cells (J&#x000F8;rgensen et al., <xref ref-type="bibr" rid="B85">2018a</xref>).</p>
<p>Advances in analytical methods, laboratory equipment and automation have rapidly improved the performance of experimental procedures (e.g., miniaturizing experiments for reactions, and connecting analytical instruments to advanced software based on decision-making algorithms and optimization tools) (Stevens et al., <xref ref-type="bibr" rid="B153">2010</xref>; Smith et al., <xref ref-type="bibr" rid="B146">2011</xref>; Richmond et al., <xref ref-type="bibr" rid="B128">2012</xref>; Houben and Lapkin, <xref ref-type="bibr" rid="B76">2015</xref>). The implementation of ML-based approaches have allowed developing innovative capabilities, such as cost-effective experiments, advanced algorithms for automation, and designing of experiments, chemoinformatics tools for dealing with high-dimensional analytical data, and accelerated <italic>in situ</italic>/in line analysis of chemical transformations (e.g., polymerization reactions, heterogeneous catalytic processes, aggregation of nanoparticles) (Houben and Lapkin, <xref ref-type="bibr" rid="B76">2015</xref>; H&#x000E4;se et al., <xref ref-type="bibr" rid="B72">2018</xref>).</p>
<p>However, there are critical challenges that ML in Chemistry must face, including the control of experiments, the detailed description of chemical space, the flexibility and generalization of models, robustness, and accuracy of predictions, and the establishment of effective cross-disciplinary collaborations (Montavon et al., <xref ref-type="bibr" rid="B117">2013</xref>; Hansen et al., <xref ref-type="bibr" rid="B68">2015</xref>; Kishimoto et al., <xref ref-type="bibr" rid="B92">2018</xref>; Smith et al., <xref ref-type="bibr" rid="B148">2018a</xref>).</p>
<p>A clear definition of ML, as well as the distinction from other purely mathematical regression methods is not straightforward, and can be associated to some degree of arbitrariness (Behler, <xref ref-type="bibr" rid="B12">2016</xref>). Standard ML methods include, artificial neural networks, support vector machines, and Gaussian processes, which have contributed to the rational design of compounds and materials, and to the improvement of computational frameworks (Goh et al., <xref ref-type="bibr" rid="B59">2017</xref>; Mater and Coote, <xref ref-type="bibr" rid="B106">2019</xref>). The latter have been applied for e.g., in QSAR models and drug design (Kadurin et al., <xref ref-type="bibr" rid="B87">2017</xref>; Chen et al., <xref ref-type="bibr" rid="B26">2018</xref>; Fleming, <xref ref-type="bibr" rid="B52">2018</xref>; Green et al., <xref ref-type="bibr" rid="B62">2018</xref>; Gupta et al., <xref ref-type="bibr" rid="B67">2018</xref>; Li et al., <xref ref-type="bibr" rid="B101">2018b</xref>; Lo et al., <xref ref-type="bibr" rid="B103">2018</xref>; Popova et al., <xref ref-type="bibr" rid="B123">2018</xref>; Sim&#x000F5;es et al., <xref ref-type="bibr" rid="B145">2018</xref>) aiming at identifying systems, molecules and materials with optimal properties (e.g., conductivity, aqueous solubility, bioavailability, bioactivity, or toxicity) (Kadurin et al., <xref ref-type="bibr" rid="B87">2017</xref>; Freeze et al., <xref ref-type="bibr" rid="B53">2019</xref>). This can be made via extensive searches, in large databases, of latent relationships between the atomic structures. The structures, can thus be encoded using multiple descriptors, and target properties.</p>
<p>The possibilities of applying ML for optimization in Chemistry are endless. There are studies focused on ML approaches for inferring on the optimized geometry of a system (Zielinski et al., <xref ref-type="bibr" rid="B171">2017</xref>; Venkatasubramanian, <xref ref-type="bibr" rid="B157">2019</xref>), and finding minima on complex potential energy surfaces (Chen et al., <xref ref-type="bibr" rid="B27">2015</xref>; Chmiela et al., <xref ref-type="bibr" rid="B28">2018</xref>; Kanamori et al., <xref ref-type="bibr" rid="B88">2018</xref>; Xia and Kais, <xref ref-type="bibr" rid="B165">2018</xref>; Hughes et al., <xref ref-type="bibr" rid="B79">2019</xref>), such as those of large water clusters (Bose et al., <xref ref-type="bibr" rid="B15">2018</xref>; Chan et al., <xref ref-type="bibr" rid="B24">2019</xref>).</p>
<p>The most innovative aspects of ML in Chemistry are related to the availability of large volumes of theoretical data (e.g., electrostatic energy contributions in force fields, atomic charges, structural properties, and representations of the potential energies), obtained from automatic and accurate electronic structure calculations (Behler, <xref ref-type="bibr" rid="B12">2016</xref>).</p>
<p>However, the intricate nature of the configuration space and its exponential dependence on system size and composition, have hampered the screening of the entire set of candidate structures directly by electronic structure calculations (Behler, <xref ref-type="bibr" rid="B12">2016</xref>; Welborn et al., <xref ref-type="bibr" rid="B161">2018</xref>).</p>
<sec>
<title>Signs of Controversy</title>
<p>Despite the usefulness of ML approaches being indisputable, with the promise to modernize molecular simulations, synthesis, materials science, and drug discovery, the respective endorsement and practical aspects in some chemical sub-fields is far from consensual (Ahneman et al., <xref ref-type="bibr" rid="B3">2018</xref>; Chuang and Keiser, <xref ref-type="bibr" rid="B32">2018a</xref>,<xref ref-type="bibr" rid="B33">b</xref>).</p>
<p>Ten years ago, there were only a few publications on applications of ML in Chemistry, but currently there are thousands of published works. The controversy has highlighted the potential (instructive) pitfalls of some practices using ML. It has been argued that ML algorithms may lead to overestimated performances and deficient model generalizations, due to their sensitivity to the presence of maze-like variables and experimental artifacts (Chuang and Keiser, <xref ref-type="bibr" rid="B32">2018a</xref>). For instance, Ahneman et al. (<xref ref-type="bibr" rid="B3">2018</xref>) have recently designed a ML model to predict yields of cross coupling reactions with high accuracy, containing isoxazoles, as reaction inhibitors, which were incorporated for assessing the robustness of the reaction. Input data for the proposed algorithm included yields and reagent parameters of 3,000 reactions, such as NMR shifts, dipole moments, and orbital energies. The most significant features of the proposed algorithm were found to be the descriptors of additives. However, the experimental design of this original work has been contested by Chuang and Keiser (<xref ref-type="bibr" rid="B33">2018b</xref>), who warned for potential artifacts associated to the original work. These authors demonstrated that the model also identified reaction additives as the descriptors displaying the greatest impact on the reactions, suggesting that high additive feature contributions cannot be discriminated from the hidden structure within the dataset, i.e., the procedure of the original paper was not sufficient for establishing isoxazole additives as the most important descriptors (Chuang and Keiser, <xref ref-type="bibr" rid="B33">2018b</xref>). A meticulous preprocessing of input data and validation of the model hypothesis was then suggested. The Y-randomization test in the original work was taken into account just the information rooted in the structure of the data set, irrespective to the intended outcome. The classical approach based on multiple hypotheses to assess alternative descriptions of the performance of the ML model was implemented (Chuang and Keiser, <xref ref-type="bibr" rid="B33">2018b</xref>). The effect of different reaction parameters (e.g., additives, catalyst, and aryl halide) in an extensive combinatorial layout generated over several independent reactions was duly explored, providing the underlying structure of the data (Chuang and Keiser, <xref ref-type="bibr" rid="B33">2018b</xref>).</p>
<p>An alternative assumption considering that ML algorithms deal with patterns within the experimental design, instead of learning from the most relevant chemical features was therefore investigated. It was concluded that ML is prone to explore data irrespective to their size and structure. This aspect was illustrated by extracting and replacing the chemical features (e.g., electrostatics, NMR shifts, dipole moments) from each molecule with random (Gaussian distributed) numeric strings. It was shown that the predictions were similar to the original ones. Chuang and Keiser (<xref ref-type="bibr" rid="B32">2018a</xref>) have also introduced technical and conceptual standpoints, including the use of adversarial controls to evaluate the predictive performance of ML models, focusing on the design of rigorous and deliberated experiments, ensuring accurate predictions from suitable and significant models (Chuang and Keiser, <xref ref-type="bibr" rid="B32">2018a</xref>). By revising the original information, a number of variations of the test sets was introduced by Estrada et al. (<xref ref-type="bibr" rid="B49">2018</xref>) for assessing the performance of predictions, considering alternatives to the random-forest model. It was therefore demonstrated that ML models are in fact quite sensitive to such imposed features, and the reagent-label models are relevant representations of the data set and useful for comparing performances in generalization assessments.</p>
<p>The original assumptions regarding the significance and validity of the random-forest (chemical-feature) model to describe important and general chemical features were also confirmed (Estrada et al., <xref ref-type="bibr" rid="B49">2018</xref>).</p>
<p>A lesson that chemists may draw from such constructive discussions is that as the size of the data set increases, the performance of ML models also increases, but with the possibility of obtaining unexpected results and irrelevant patterns, as the rules for ML algorithms to detect and deal with potential technical and conceptual gaps are not well-established. Specifically, the description of chemical reactivity underlying a data set is required in order to ensure the reaction prediction, by using data and reagent-label models to evaluate the scope and restraints of chemical characterization.</p>
<p>ML provides new opportunities to increase the quality and quantity of chemical data, which are essential to promote optimization, implementation of rational design and synthetic approaches, prioritization of candidate molecules, decision-making, and also for guiding of innovative ideas.</p>
</sec>
<sec>
<title>Deep Learning, Deep Chemistry</title>
<p>In this section, an introductory overview into the core concepts of DL, and DLNs is provided. Focus is given to the unique properties of DL, that distinguish these algorithms from traditional machine learning approaches, with emphasis on chemical applications rather than providing theoretical and mathematical details.</p>
<p>ML is a branch of computer science dedicated to the development of algorithms capable of learning and making decisions on complex data (Samuel, <xref ref-type="bibr" rid="B134">1959</xref>; Mitchell, <xref ref-type="bibr" rid="B115">1997</xref>). This learning process involves specific tasks that are commonly classified in (i) supervised learning, for establishing the relationship between input and output data (e.g., linear regressions and classification techniques), (ii) unsupervised learning, for finding hidden patterns or features in data, without any previous information on such characteristics and interrelations (e.g., clustering and dimension reduction techniques), and (iii) reinforcement learning, for performing a particular task through repeated dynamic interactions e.g., optimization of molecules (Zhou et al., <xref ref-type="bibr" rid="B169">2018</xref>) and chemical reactions (Zhou et al., <xref ref-type="bibr" rid="B170">2017</xref>).</p>
<p>Deep learning is a fast-moving sub-area of ML, focused on sophisticated learning and extrapolation tasks, fostered by the wide range of chemistry literature, open-source code, and datasets (Goh et al., <xref ref-type="bibr" rid="B59">2017</xref>).</p>
<p>The ability of DL to establish the relevant phenomena, expedite chemical reactions, and predict relevant properties, optimal synthesis routes, solve critical analytical uncertainties, and reduce costs and resources, is invaluable in Chemistry. Its success in modeling compound properties and reactions, depends, among other aspects, on the access to comprehensive, historical repositories of published chemical data (Venkatasubramanian, <xref ref-type="bibr" rid="B157">2019</xref>).</p>
<p>There are barriers to be surpassed, including cleaning data, production of meaningful and accurate chemical information (free of bias), lack of standardization of chemical data, expertise and familiarity with ML and DL in chemistry sectors, and also lack of collaboration opportunities) (Mater and Coote, <xref ref-type="bibr" rid="B106">2019</xref>).</p>
<p>The majority of DL algorithms currently developed are based on artificial neural networks (Lecun et al., <xref ref-type="bibr" rid="B96">2015</xref>).</p>
<p>DLNs are now a proving-ground for research in chemical sciences (Goh et al., <xref ref-type="bibr" rid="B59">2017</xref>; Jha et al., <xref ref-type="bibr" rid="B84">2018</xref>; Popova et al., <xref ref-type="bibr" rid="B123">2018</xref>; Segler et al., <xref ref-type="bibr" rid="B142">2018</xref>; Elton et al., <xref ref-type="bibr" rid="B48">2019</xref>; Mater and Coote, <xref ref-type="bibr" rid="B106">2019</xref>; Xu et al., <xref ref-type="bibr" rid="B166">2019</xref>). Similarly to artificial neural networks, DLNs are produced to resemble the brain, in which the information passes through a series of interconnected nodes comparable to neurons (Lecun et al., <xref ref-type="bibr" rid="B96">2015</xref>). Each node analyzes segments of information and transfer that information to adjacent nodes (see <xref ref-type="fig" rid="F3">Figure 3</xref>).</p>
<fig id="F3" position="float">
<label>Figure 3</label>
<caption><p>Schematic representation of an artificial neuron (top), and a simple neural network displaying three basic elements: input, hidden and output layers (bottom-left), and a deep neural network showing at least two hidden layers, or nodes (bottom-right). The calculation is performed through the connections, which contain the input data, the pre-assigned weights, and the paths defined by the activation function. If the result is far from expected, the weights of the connections are recalibrated, and the analysis continues, until the outcome is as accurate as possible.</p></caption>
<graphic xlink:href="fchem-07-00809-g0003.tif"/>
</fig>
<p>The computational model consists of multiple hidden layers (in higher number comparing to more conventional approaches) which confer the ability of DLNs to learn from highly complex data and perform correlation and reduction. This means that the algorithm discovers correlated data, while discarding irrelevant information. Each layer combines information collected from the previous layer, and subsequently infers on the respective significance and send the relevant information to the next layer. The hidden term is used to represent layers that are not direct neighbors of the input or output layers.</p>
<p>The process allows constructing increasingly complex and abstract features, by adding layers and/or increasing the number of neurons per layer. However, the use of more than a single hidden layer requires determining error attributions and corrections to the respective weights. This is carried out via a backpropagation, i.e., a backward process starting from the predicted output, and back through the neural network (Goh et al., <xref ref-type="bibr" rid="B59">2017</xref>). In this process a gradient descent algorithm is employed to determine the minimum in the error surface created by each respective neuron, when generating the output. Note that, this gradient descent approach is conceptually similar to the steepest descent algorithm implemented in classical MD simulations (Goh et al., <xref ref-type="bibr" rid="B59">2017</xref>). The major difference lies on the use of an iterative process, in which an error function of the target output of the neural network is minimized, and the weights of the neurons are updated, instead of iteratively minimizing an energy function and updating atomic coordinates for each step.</p>
<p>A complete description of the main core concepts and architecture of DL applied to chemistry is given in Goh et al. (<xref ref-type="bibr" rid="B59">2017</xref>) and Mater and Coote (<xref ref-type="bibr" rid="B106">2019</xref>).</p>
<p>Other interesting reviews covering theoretical aspects (Goh et al., <xref ref-type="bibr" rid="B59">2017</xref>), available descriptors and datasets, and also comparing model performances (Wu et al., <xref ref-type="bibr" rid="B164">2017</xref>) have been published. Moreover, a wide range of ML applications, including drug design (Ekins, <xref ref-type="bibr" rid="B45">2016</xref>; Chen et al., <xref ref-type="bibr" rid="B26">2018</xref>; Fleming, <xref ref-type="bibr" rid="B52">2018</xref>), synthesis planning (Coley et al., <xref ref-type="bibr" rid="B35">2018a</xref>), medicinal chemistry (Panteleev et al., <xref ref-type="bibr" rid="B122">2018</xref>), cheminformatics (Lo et al., <xref ref-type="bibr" rid="B103">2018</xref>), quantum mechanical calculations (Rupp, <xref ref-type="bibr" rid="B130">2015</xref>), and materials science (Butler et al., <xref ref-type="bibr" rid="B21">2018</xref>) have been collected.</p>
<p>A summary of the main contributions of DL for solving relevant chemical challenges, as well as an illustration of the general components of a DL framework are presented in <xref ref-type="fig" rid="F4">Figure 4</xref>.</p>
<fig id="F4" position="float">
<label>Figure 4</label>
<caption><p>Overview of (top) the contribution of DL algorithms for solving different chemical challenges and the respective tasks, and (bottom) the general components of a DL framework, including the input data, the learning model able to interpret the data and the prediction space, from which the model performance can be inspected. The model represents an optimization cycle containing interconnected components: prediction, evaluation, and optimization. Reprinted with permission from Mater and Coote (<xref ref-type="bibr" rid="B106">2019</xref>). Copyright (2019) American Chemical Society.</p></caption>
<graphic xlink:href="fchem-07-00809-g0004.tif"/>
</fig>
<p>DL algorithms are particularly attractive for accelerating discoveries in pharmaceutical, medicinal and environmental chemistry (El-Atta and Hassanien, <xref ref-type="bibr" rid="B46">2017</xref>; Goh et al., <xref ref-type="bibr" rid="B59">2017</xref>; Klucznik et al., <xref ref-type="bibr" rid="B93">2018</xref>; Miller et al., <xref ref-type="bibr" rid="B112">2018</xref>; Panteleev et al., <xref ref-type="bibr" rid="B122">2018</xref>; Smith et al., <xref ref-type="bibr" rid="B149">2018b</xref>; Wu and Wang, <xref ref-type="bibr" rid="B163">2018</xref>; Molga et al., <xref ref-type="bibr" rid="B116">2019</xref>), since they have made possible, for e.g., to simulate millions of toxic compounds and identify those compounds displaying target properties, safely, economically, and sustainably. These types of applications have been thoroughly revised in various publications and will not be further addressed in what follows [see for e.g., (Kadurin et al., <xref ref-type="bibr" rid="B87">2017</xref>; Chen et al., <xref ref-type="bibr" rid="B26">2018</xref>; Fleming, <xref ref-type="bibr" rid="B52">2018</xref>; Green et al., <xref ref-type="bibr" rid="B62">2018</xref>; Gupta et al., <xref ref-type="bibr" rid="B67">2018</xref>; Li et al., <xref ref-type="bibr" rid="B101">2018b</xref>; Lo et al., <xref ref-type="bibr" rid="B103">2018</xref>; Panteleev et al., <xref ref-type="bibr" rid="B122">2018</xref>; Popova et al., <xref ref-type="bibr" rid="B123">2018</xref>; Smith et al., <xref ref-type="bibr" rid="B149">2018b</xref>)].</p>
<p>DL is not only a cost-cutting effort, but also an innovative source of new perspectives.</p>
</sec>
</sec>
<sec id="s3">
<title>Cutting-Edge Applications</title>
<p>In recent years, ML has been evoked in chemistry-related tasks. The use of ML and, in particular, DL-based approaches across prediction of binding, activity and other relevant molecular properties, compound/material design and synthesis, as well as applications of genetic algorithms are highlighted in what follows.</p>
<p>Researchers in chemical sciences have started exploring the capabilities of ML using data collected from computations and experimental measurements. Data mining is traditionally adopted to explore high-dimensional data sets, in order to identify and establish relevant connections for the chemical features of compounds and materials.</p>
<p>Other more ambitious approaches, including quantum mechanics, which integrates physics-based computations (e.g., DFT) and ML methods in the search for novel molecular components, have also been implemented (Curtarolo et al., <xref ref-type="bibr" rid="B39">2013</xref>).</p>
<p>Amongst the major achievements of DL in Chemistry, are the outstanding performances in predicting activity and toxicity, in the context of the Merck activity prediction challenge in 2012, and the Tox21 toxicity prediction challenge launched by NIH in 2014, respectively. In the former, DL was very successful in the competition outperforming Merck&#x00027;s internal baseline model. In the second challenge, DL models also achieved top positions (Goh et al., <xref ref-type="bibr" rid="B59">2017</xref>).</p>
<p>Similarly to what happens to the majority of the modern computational chemists who no longer build their own code to perform MD simulations or quantum chemical calculations, due to the existence and availability of well-established software packages, DL researchers have also use several software packages for training neural networks including Torch, Caffe, Theano, and Tensorflow (Goh et al., <xref ref-type="bibr" rid="B59">2017</xref>).</p>
<p>Apart from the influence of software improvements, the continuous growth of chemical data in public databases, such as PubChem and Protein Data Bank has also facilitated the raise of ML and DL applications in Chemistry, including quantum chemistry, property prediction and materials design, drug discovery, QSAR, virtual screening, and protein structure prediction (Goh et al., <xref ref-type="bibr" rid="B59">2017</xref>; Christensen et al., <xref ref-type="bibr" rid="B31">2019</xref>).</p>
<sec>
<title>Improving Computational and Quantum Chemistry</title>
<p>Computational chemistry is naturally a sub-field that has been increasingly boosted by the advances and unique capabilities of ML (Rupp et al., <xref ref-type="bibr" rid="B132">2012</xref>; Ramakrishnan et al., <xref ref-type="bibr" rid="B125">2014</xref>, <xref ref-type="bibr" rid="B126">2015</xref>; Dral et al., <xref ref-type="bibr" rid="B42">2015</xref>; S&#x000E1;nchez-Lengeling and Aspuru-Guzik, <xref ref-type="bibr" rid="B135">2017</xref>; Christensen et al., <xref ref-type="bibr" rid="B31">2019</xref>; Iype and Urolagin, <xref ref-type="bibr" rid="B80">2019</xref>; Mezei and Von Lilienfeld, <xref ref-type="bibr" rid="B110">2019</xref>; Zaspel et al., <xref ref-type="bibr" rid="B167">2019</xref>).</p>
<p>Also, recent progresses have enabled the acceleration of MD simulations (atomistic and coarse-grained), contributing to increase knowledge on the interactions within quantum many-body systems and efficiency of DFT-based quantum mechanical modeling methods (Bart&#x000F3;k et al., <xref ref-type="bibr" rid="B9">2010</xref>, <xref ref-type="bibr" rid="B8">2013</xref>; Behler, <xref ref-type="bibr" rid="B10">2011a</xref>,<xref ref-type="bibr" rid="B11">b</xref>, <xref ref-type="bibr" rid="B12">2016</xref>; Rupp et al., <xref ref-type="bibr" rid="B132">2012</xref>, <xref ref-type="bibr" rid="B131">2015</xref>; Snyder et al., <xref ref-type="bibr" rid="B150">2012</xref>; Hansen et al., <xref ref-type="bibr" rid="B69">2013</xref>, <xref ref-type="bibr" rid="B68">2015</xref>; Montavon et al., <xref ref-type="bibr" rid="B117">2013</xref>; Sch&#x000FC;tt et al., <xref ref-type="bibr" rid="B141">2014</xref>; Alipanahi et al., <xref ref-type="bibr" rid="B4">2015</xref>; Botu and Ramprasad, <xref ref-type="bibr" rid="B18">2015b</xref>; De et al., <xref ref-type="bibr" rid="B40">2016</xref>; Faber et al., <xref ref-type="bibr" rid="B51">2016</xref>; Sadowski et al., <xref ref-type="bibr" rid="B133">2016</xref>; Wei et al., <xref ref-type="bibr" rid="B160">2016</xref>; Brockherde et al., <xref ref-type="bibr" rid="B19">2017</xref>; Chmiela et al., <xref ref-type="bibr" rid="B29">2017</xref>, <xref ref-type="bibr" rid="B28">2018</xref>; Smith et al., <xref ref-type="bibr" rid="B147">2017</xref>; Wu et al., <xref ref-type="bibr" rid="B164">2017</xref>; G&#x000F3;mez-Bombarelli et al., <xref ref-type="bibr" rid="B60">2018</xref>). This field is still in its infancy and have offered invaluable opportunities for dealing with a wide range of challenges and unsolved questions, including but not limited to model accuracy, interpretability, and causality.</p>
<p>For instance, the prediction of the refractive index of ionic liquids based on quantum chemistry calculations and an extreme learning machine (ELM) algorithm has been conducted (Kang et al., <xref ref-type="bibr" rid="B90">2018</xref>). Specifically, 1,194 experimental data points for 115 ionic liquids at different temperatures were collected from more than 100 literature reports. Quantum chemistry calculations were performed for obtaining the structures and descriptors of the ionic liquids. The model was designed using a stepwise regression algorithm and the <italic>R</italic><sup>2</sup> and AARD% values were 0.841 and 0.855%, respectively. It was found that prediction of the refractive index was significantly affected by ionic liquid anions, comparing to the cations. Better performances were achieve using the ELM algorithm, with the <italic>R</italic><sup>2</sup> and AARD% values of 0.957 and 0.295%, respectively (Kang et al., <xref ref-type="bibr" rid="B90">2018</xref>).</p>
<p>ML has also contributed for modeling the water behavior, shedding light on important phenomena related to water molecules interactions and the resulting density. Morawietz et al. (<xref ref-type="bibr" rid="B118">2016</xref>) have calculated ice&#x00027;s melting point from fundamental quantum mechanics, demonstrating the predictive power of ab initio MD simulations and highlighting the critical role of van der Waals forces (Morawietz et al., <xref ref-type="bibr" rid="B118">2016</xref>). It was evidenced that ice occupies a larger volume than liquid water as hydrogen bonds display water molecules in a rigid 3D network. These hydrogen bonds weaken when ice melts, and water molecules approximate, becoming dense with an extreme value at 4&#x000B0;C (Morawietz et al., <xref ref-type="bibr" rid="B118">2016</xref>). Note that these processes can also be rationalized resorting to <italic>ab initio</italic> MD approaches based on DFT; however, such calculations are associated to highly demanding computations. In addition to this, DFT approaches are not able to accurately reproduce minute but relevant van der Waals forces. The same authors have trained a neural network to reproduce DFT results with less computer power, and employed a previously-existing van der Waals correction. Water density changes, hydrogen bond network flexibility, and competition effects in terms of the nearest shell&#x00027;s contraction, after cooling, were explained based on the simulations (Morawietz et al., <xref ref-type="bibr" rid="B118">2016</xref>).</p>
<p>One of the current challenges is to answer the question of whether chemical-physical properties, that often require quantum mechanics (e.g., dipole moments, binding and potential energies, and thermodynamics), can be represented and predicted by ML methods (Hansen et al., <xref ref-type="bibr" rid="B69">2013</xref>, <xref ref-type="bibr" rid="B68">2015</xref>; Montavon et al., <xref ref-type="bibr" rid="B117">2013</xref>; Faber et al., <xref ref-type="bibr" rid="B51">2016</xref>; Iype and Urolagin, <xref ref-type="bibr" rid="B80">2019</xref>; Jaquis et al., <xref ref-type="bibr" rid="B82">2019</xref>). Several attempts have been made on the topic with some successful examples (Rupp et al., <xref ref-type="bibr" rid="B132">2012</xref>; Faber et al., <xref ref-type="bibr" rid="B50">2017</xref>).</p>
<p>Rupp et al. (<xref ref-type="bibr" rid="B132">2012</xref>) have developed a model based on nuclear charges and atomic positions for predicting molecular atomization energies of various organic compounds. A matrix composed of molecular elements and configuration was built, describing the potential energy of each individual atom and the Coulomb repulsion between nuclear charges. A non-linear regression scheme was employed for solving and mapping the molecular Schr&#x000F6;dinger equation.</p>
<p>The regression models were trained and compared to atomization energies calculated with hybrid DFT, transforming a 1-h run (on average) of hybrid DFT per each atomization energy into milliseconds using ML. Cross-validation over more than seven thousand organic molecules yielded a mean absolute error below 10 kcal/mol. The authors have trained the ML algorithm on a set of compounds in a database, comparing the respective matrices to determine differences between molecules, so as to develop a landscape of such differences. Based on the atomic composition and configuration, the unknown molecule can be positioned in the landscape and the respective atomization energy can be estimated by the contributions (weights) obtained from the differences between the unknown and all known molecules (Rupp et al., <xref ref-type="bibr" rid="B132">2012</xref>).</p>
<p>More recently, the impact of selecting regressors and molecular representations on the construction of fast ML models of several electronic ground-state properties of organic molecules has also been investigated (Faber et al., <xref ref-type="bibr" rid="B50">2017</xref>). The performance of each combination between regressor, representation, and property was evaluated with learning curves, which allowed reporting out-of-sample errors, as a function of the size if the training set (ca. 118 k molecules). The QM9 database (Ramakrishnan et al., <xref ref-type="bibr" rid="B125">2014</xref>) was used for extracting the molecular structures and properties at the hybrid DFT level of theory, and included data on dipole moment, polarizability, enthalpies and free-energies of atomization, HOMO/LUMO energies and gap, heat capacity, zero point vibrational energy, and the highest fundamental vibrational frequency.</p>
<p>Several regression methods including linear models (Bayesian ridge regression and elastic net regularization), random-forest, kernel ridge regression, and neural networks (graph convolutions and gated graph networks) were tested. It was concluded that out-of-sample errors were strongly affected by the molecular properties, and by the type of representation and regression method. Molecular graphs and graph convolutions displayed better performances for electronic properties, while kernel ridge regression and histograms of dihedrals were suitable for describing energy-related properties [see Faber et al. (<xref ref-type="bibr" rid="B50">2017</xref>) for details on other relevant combinations]. Predictions based on the ML model for all properties have shown lower deviations from DFT (B3LYP) than the latter deviated from experiment. ML models displayed thus an improved prediction accuracy than hybrid DFT, since experimental or explicitly electron correlated quantum data were available.</p>
<p>In terms of drug development Brockherde et al. (<xref ref-type="bibr" rid="B19">2017</xref>) have developed a ML algorithm for predicting the behavior of molecules with potential to be used as pharmaceuticals and in the design of new molecules, able to enhance the performance of emerging energetic materials, including solar cells, battery technologies, and digital displays. The main goal was to identify the underlying patterns in the molecular behavior, by employing the ML algorithm for understanding atomic interactions within a molecule and using such information to predict new molecular phenomena.</p>
<p>Specifically, the algorithm was created and trained on the basis of a small sample set of the molecule under study, and applied to simulate the intricate chemical behavior within selected molecules, including malonaldehyde. A directed learning of the density-potential and energy-density maps was conducted, as illustrated in <xref ref-type="fig" rid="F5">Figure 5</xref>, and the first MD simulation of with a ML density functional on malonaldehyde was performed, allowing to describe the intramolecular proton transfer process (Brockherde et al., <xref ref-type="bibr" rid="B19">2017</xref>).</p>
<fig id="F5" position="float">
<label>Figure 5</label>
<caption><p><bold>(A)</bold> Illustrative summary of the mappings proposed by Brockherde et al. (<xref ref-type="bibr" rid="B19">2017</xref>). <italic>E</italic>[<italic>v</italic>] is a conventional electronic structure calculation, i.e., Kohn&#x02013;Sham density functional theory (KS-DFT) and is represented by the bottom vector. The ground-state energy is determined by solving KS equations given the external potential, <italic>v</italic>. <italic>E</italic>[<italic>n</italic>] corresponds to the total energy density functional. The Hohenberg&#x02013;Kohn map n[v] (red vector) from external potential to its ground state density is also presented. <bold>(B)</bold> Top: graphical representation of the dependency of the energy error on the number of training points (M), for ML-OF and ML-HK, considering different basis sets for the one-dimensional problem. Bottom: errors in the Perdew-Burke-Ernzerhof (PBE) energies and the ML maps as a function of interatomic spacing, R, for H<sub>2</sub> with M &#x0003D; 7. <bold>(C)</bold> Schematic illustration of the strategy for obtaining predictions based on the proposed machine learning Hohenberg&#x02013;Kohn (ML-HK) map. Molecular geometry is represented by Gaussians, several independent Kernel ridge regression models allows predicting each basis coefficient of the density. The performance of data-driven (ML) and common physical basis representations for the electron density is assessed.</p></caption>
<graphic xlink:href="fchem-07-00809-g0005.tif"/>
</fig>
<p>In more detail, one of the key tasks in atomistic modeling is the prompt and automated analysis of the simulation results, in order to provide a comprehensive understanding of the behavior of individual atoms and target collective properties. The main supervised and unsupervised machine-learning methods directed at classifying and coarse-graining of molecular simulations were recently summarized and discussed in Ceriotti (<xref ref-type="bibr" rid="B22">2019</xref>). A schematic overview of these methods, and also of a workflow reflecting the application of a ML scheme to an atomic-scale system is presented in <xref ref-type="fig" rid="F6">Figure 6</xref>.</p>
<fig id="F6" position="float">
<label>Figure 6</label>
<caption><p><bold>(I)</bold> Schematic representation of the main components of atomistic ML. <bold>(a)</bold> the inputs of the model are structures <italic>A</italic> or local environments <italic>X</italic>, <bold>(b)</bold> the mathematical representation of the inputs, based on vectors of features |<italic>X</italic> &#x0232A;, a measure of similarity d, or a kernel k, <bold>(c)</bold> the ML model, controlled by a series of parameters &#x003B8;, and trained based on a set of inputs. <bold>(II)</bold> An overview of the clustering methods, including <bold>(a)</bold> a set of data points clustered according to their hidden common features, <bold>(b)</bold> a density-based clustering for identifying maxima in the probability distribution of inputs, <bold>(c)</bold> distribution-based clustering for finding a model of the data distribution based on the combination of clustering probabilities, and <bold>(d)</bold> hierarchical clustering for identifying natural clusters of the inputs. <bold>(III)</bold> Summary of dimensional reduction techniques, including principal component analysis (PCA) for establishing the most relevant subspace retaining the largest fraction of the input data variance, <bold>(b)</bold> a kernel-based method, <bold>(c)</bold> multidimensional scaling for reproducing in low dimension the similarity between high-dimensional data points. Reprinted with permission from Ceriotti (<xref ref-type="bibr" rid="B22">2019</xref>).</p></caption>
<graphic xlink:href="fchem-07-00809-g0006.tif"/>
</fig>
<p>Also relevant is the development of improved molecular force fields, commonly used in MD simulations, using ML. On the other hand, the intrinsic operational aspects of MD simulations, in which the dynamic evolution of the chemical system is detailed in a fixed period of time, and for which interparticle forces and potential energies are often estimated using interatomic potentials, or molecular mechanics force fields, are perfectly suited for ML. In fact, some of the timesteps can be used as a training phase for estimating consecutive ones, assuming that each of the timesteps of MD simulation is strongly correlated with the preceding timestep and is adequate for sampling the phase space rapidly and accurately, allowing to estimate any meaningful property (Behler, <xref ref-type="bibr" rid="B12">2016</xref>). MD simulations often sample abnormal, but probably relevant configurations, requiring the implementation of a decision tool for dealing with the unusual configuration, and from which ML may turn off and start learning (Botu and Ramprasad, <xref ref-type="bibr" rid="B17">2015a</xref>; Smith et al., <xref ref-type="bibr" rid="B148">2018a</xref>). These conditions have also been previously discussed and applied to <italic>ab initio</italic> MD (Botu and Ramprasad, <xref ref-type="bibr" rid="B17">2015a</xref>).</p>
<p>In MD, the energies and forces for a vast number of atomic configurations are required, which can be obtained by performing the electronic structure calculations along the trajectory, or by evaluating the direct functional relation between the atomic configuration and the energy (Mansbach and Ferguson, <xref ref-type="bibr" rid="B104">2015</xref>). This analytic expression, defined before running the simulation, is often recognized as a force field, an interatomic potential, or a potential-energy surface. Calculations of electronic structures are very demanding, even for DFT. DFT-based <italic>ab initio</italic> MD simulations are restricted to a few 100 atoms and shorter simulation times (Ahn et al., <xref ref-type="bibr" rid="B2">2019</xref>).</p>
<p>The requirements for calculating ML potentials are very similar to conventional empirical potentials, and are duly discussed in Behler (<xref ref-type="bibr" rid="B12">2016</xref>). More recent conventional force fields are developed and validated for very specific systems, being limited by the functional form upon which they were constructed. On the other hand, despite requiring a training set, ML-based force fields are adaptive and more robust upon configurations not previously sampled (Botu and Ramprasad, <xref ref-type="bibr" rid="B17">2015a</xref>). Furthermore, these force fields can be extended rapidly to different types of atoms and molecules, as they can learn and apply the physical laws, rather than starting from strarch (Botu et al., <xref ref-type="bibr" rid="B16">2017</xref>).</p>
<p>Several improved force fields, and accurate predictions of thermodynamics and kinetics signatures, as well as their influence in molecular structures have been provided by performing ML-based atomistic and <italic>ab initio</italic> MD simulations. For instance, Chmiela et al. (<xref ref-type="bibr" rid="B28">2018</xref>) have incorporated spatial and temporal physical symmetries into a gradient-domain machine learning (sGDML) model for constructing flexible molecular force fields from high-level <italic>ab initio</italic> calculations, with a great potential to be used to improve spectroscopic accuracy in molecular simulations. The sGDML model was able to reproduce global force fields at quantum-chemical CCSD(T) level of accuracy and produced converged MD simulations with fully quantized electrons and nuclei (Chmiela et al., <xref ref-type="bibr" rid="B28">2018</xref>).</p>
<p>The parameterization of force fields and semiempirical quantum mechanics have also been performed integrating ML and evolutionary algorithms (Wang et al., <xref ref-type="bibr" rid="B158">2019</xref>), which were successfully applied in MD (Wang et al., <xref ref-type="bibr" rid="B158">2019</xref>). Constructing coarse-grained molecular models has been a common approach to extend the time/length-scales accessible to large or complex systems (Wang et al., <xref ref-type="bibr" rid="B158">2019</xref>). These models have allowed establishing suitable interaction potentials for properties of high-resolution models or experimental data. Wang et al. (<xref ref-type="bibr" rid="B158">2019</xref>) have reformulated coarse-graining as a supervised machine learning problem, by using statistical learning theory for decoupling the coarse-graining error, and cross-validation for choosing and comparing the performance of distinct models. For that purpose, the authors developed a DL model, that learned coarse-grained free-energy functions and was trained by a force-matching strategy (see <xref ref-type="fig" rid="F7">Figure 7</xref>).</p>
<fig id="F7" position="float">
<label>Figure 7</label>
<caption><p><bold>(I)</bold> Machine-learned coarse-graining of dynamics in <bold>(a)</bold> a two-dimensional potential, showing the <bold>(b)</bold> exact free-energy along x, comparison of <bold>(c)</bold> the instantaneous forces and the learned mean forces using feature regression and coarse-grained neural network models with the exact forces, and <bold>(d)</bold> the potential-of-mean-force along x, predicted by feature regression, and coarse-grained neural network models with the exact free energy. <bold>(II)</bold> Free-energy profiles and representative structures of alanine dipeptide simulated using all-atom and machine-learned coarse-grained models: <bold>(a)</bold> free-energy reference as a function of the dihedral angles, obtained from the histograms of all-atom simulations, <bold>(b)</bold> standard coarse-grained model using a sum of splines of individual internal coordinates, <bold>(c)</bold> regularized coarse-grained neural network models, <bold>(d)</bold> unregularized networks, <bold>(e)</bold> representative structures extracted from the free-energy minima, from atomistic simulation (ball-and-stick representation) and regularized coarse-grained neural network simulation (licorice representation). <bold>(III)</bold> Free-energy landscape of Chignolin for the different models, obtained from the <bold>(a)</bold> all-atom simulation, as a function of the first two TICA coordinates, <bold>(b)</bold> spline model, as a function of the same two coordinates used in the all-atom model, <bold>(c)</bold> coarse-grained neural network model, as a function of the same two coordinates. <bold>(d)</bold> Comparison of the one-dimensional free-energy profile as a function of the first TICA coordinate, reflecting the folding/unfolding transition, for the all-atom (blue), spline (green), and coarse-grained neural network models (red). <bold>(e)</bold> Representative Chignolin conformations in the three minima from (a&#x02013;c) all-atom simulation and (a&#x02032;-c&#x02032;) coarse-grained neural network model. Reprinted with permission from Wang et al. (<xref ref-type="bibr" rid="B158">2019</xref>).</p></caption>
<graphic xlink:href="fchem-07-00809-g0007.tif"/>
</fig>
<p>The proposed framework automatically learned multiple terms necessary for accurate coarse-grained force fields, i.e., was able to keep relevant invariances and incorporate physics knowledge, avoiding the sampling of unphysical structures.</p>
<p>The class of coarse-grained directed neural networks can thus be trained with the force-matching principle and can encode all physically relevant invariances and constraints, including invariance of (i) the free-energy and mean force with respect to translation of the molecule, (ii) the free-energy and variance of the mean force associated to molecular rotation, and considering (iii) the mean force being a conservative force field generated by the free-energy, and (iv) a prior energy for preventing deviations of the simulations with coarse-grained neural networks into unphysical state space regions, i.e., states displaying overstretched bonds or clashing atoms, which are captured out of the training data.</p>
<p>The proposed strategy also outperformed classical coarse-graining approaches, which generally failed to capture relevant features of the free-energy surface, providing the all-atom explicit-solvent free-energy surfaces estimated with models including just a few coarse-grained beads, in the absence of solvent (Wang et al., <xref ref-type="bibr" rid="B158">2019</xref>).</p>
<p>The integration of ML in MD simulations have also been useful for understanding the rate and yield of chemical reactions and providing key mechanistic details (Christensen et al., <xref ref-type="bibr" rid="B31">2019</xref>; H&#x000E4;se et al., <xref ref-type="bibr" rid="B71">2019</xref>). For instance, an unsupervised ML analysis tool based on Bayesian neural networks (BNNs) was proposed by H&#x000E4;se et al. (<xref ref-type="bibr" rid="B71">2019</xref>) to extract relevant information from <italic>ab initio</italic> MD simulation of chemical reactions (H&#x000E4;se et al., <xref ref-type="bibr" rid="B71">2019</xref>). BNNs have been optimized to predict a specific outcome of an <italic>ab initio</italic> MD simulation corresponding to the dissociation time of the unmethylated and tetramethylated 1,2-dioxetane molecules, from the initial nuclear geometry and velocities. Predictions based on BNNs showed that an earlier dissociation was related to the planarization of the two formaldehyde moieties and also to the symmetric shortening of the C&#x02013;O bonds, respecting the octet rule, i.e., the relation between bond order and bond length and orbital hybridization (H&#x000E4;se et al., <xref ref-type="bibr" rid="B71">2019</xref>).</p>
<p>Rupp et al. (<xref ref-type="bibr" rid="B132">2012</xref>) have developed a ML algorithm based on non-linear statistical regression to predict the atomization energies of organic molecules. The proposed model employed a subset of seven thousand elements of the database, and a library of more than 100 stable and synthetically-tractable organic compounds. The target data used to train the model included atomization energies of the compounds calculated using the PBE0 hybrid functional. Cartesian coordinated and nuclear charge were used as descriptors in a &#x0201C;Coulomb&#x0201D; matrix representation. A mean-absolute error accuracy of 14.9 kcal/mol was achieved using a small fraction of the compounds for the training set. Similar accuracy, ca. 15.3 kcal/mol, was obtained considering an external validation set of 6,000 compounds showing the potential transferability of the model within in-class compounds. It was notable to outline QM-calculated energies, with a mean-absolute error of ca. 15 kcal/mol, without using the Schrodinger Equation in the ML algorithm. It was also suggested that the DLNs-based model should outperform the traditional ML-approach (Goh et al., <xref ref-type="bibr" rid="B59">2017</xref>).</p>
<p>More recently, an alternative approach based ML algorithms for supplementing existing QM algorithms was proposed (Ramakrishnan et al., <xref ref-type="bibr" rid="B126">2015</xref>). A &#x00394;-learning approach, involving a ML correction term was developed. Such correction was used in DFT calculated properties for predicting the corresponding quantity at the G4MP2 level of theory. This combined QM/ML approach gathers approximate but fast legacy QM approximations and big-data based QM estimates, trained on results across chemical space, despite being applied using only traditional ML algorithms (Ramakrishnan et al., <xref ref-type="bibr" rid="B126">2015</xref>).</p>
<p>G&#x000F3;mez-Bombarelli et al. (<xref ref-type="bibr" rid="B60">2018</xref>) have applied DL for generating and optimizing functional compounds, such as drug-like molecules. The proposed model allowed converting discrete representations of molecules from and into a multidimensional continuous representation, and generating new molecules for exploration and optimization.</p>
<p>A DLN was trained on a a large set of existing chemical structures to build an encoder, which converts the discrete representation of a molecule into a continuous vector, a decoder, that transforms the continuous vector into discrete molecular representations (e.g., SMILES string), and a predictor, which estimates chemical properties from the latent continuous vector representation of the molecule. These representations allowed generating new chemical structures automatically by employing simple operations in the latent space (e.g., decoding random vectors, perturbing defined chemical structures, and interpolating between molecules), and applying gradient-based optimization for a oriented-search of functional molecules (G&#x000F3;mez-Bombarelli et al., <xref ref-type="bibr" rid="B60">2018</xref>).</p>
<p>DLNs have also been applied for exploring the molecular conformational space of proteins. Some authors (Degiacomi, <xref ref-type="bibr" rid="B41">2019</xref>) have demonstrated that generative neural networks trained on protein structures, extracted from molecular simulation, can be employed to create new conformations complementing pre-existing ones. The model was trained and tested in a protein-protein docking scenario to account for specific motions occurring upon binding.</p>
<p>The fewer examples of DLNs applications in quantum chemistry suggest that it is still in an earlier stage of development compared to other approaches including computational structural biology and computer-aided drug design.</p>
</sec>
<sec>
<title>Planning and Predicting Reactions and Routes</title>
<p>Some practical questions in organic chemistry have been addressed by ML approaches, including the identification of the most suitable synthesis method for a specific compound and the optimal conditions (reactants, solvent, catalyst, temperature, and among others) for ensuring region/chemo/stereo selectivity and obtaining the highest yields, estimating the precise rate, yield and time for the reaction, predicting major/minor product, and also evaluating similarity between reactions (Wei et al., <xref ref-type="bibr" rid="B160">2016</xref>; Ahneman et al., <xref ref-type="bibr" rid="B3">2018</xref>).</p>
<p>Making predictions in reactive chemical systems can also resort to DL. Segler and Waller (<xref ref-type="bibr" rid="B143">2017</xref>) and Segler et al. (<xref ref-type="bibr" rid="B142">2018</xref>) have predicted reaction rules considering fundamental substructures of reactants and products, allowing to return a product, given a reactant as input, and vice versa. In simple terms, a reaction rule is a pattern guiding the interaction process for a set of reactants and suggesting potential chemical products. As the knowledge available in often inaccurate, such rules are often ambiguous or even incomplete (Kishimoto et al., <xref ref-type="bibr" rid="B92">2018</xref>). However, there are some successful examples, such as the recent outcomes of Chematica. Grzybowski et al. (<xref ref-type="bibr" rid="B66">2018</xref>) have assembled the relevant transformations that connect chemical species into a large network. The latter have codified and organized the known pathways through chemical space and displays nodes of molecules, elements and chemical reactions, collected by linking reactants to products on the basis of core reactions.</p>
<p>The Chematica platform comprises network theory, high-performance computing, artificial intelligence, and expert chemical knowledge to accelerate the design of synthetic pathways leading to new targets. However, the experimental verification of the respective predictions was carried out recently (Grzybowski et al., <xref ref-type="bibr" rid="B66">2018</xref>). The authors have described the results of a systematic approach in which synthetic pathways leading to eight targets with distinct structures and of medicinal relevance were designed without human supervision and experimentally validated. There are other prominent products such as ChemPlanner, and Synthia created from databases of rules for chemical transformations. Both platforms incorporate ML algorithms and allows navigating through chemical space using those rules and suggesting to the user possible ways to synthesize a target molecule. Synthia also employs MD, quantum mechanics, and electronic properties to infer on the viability of a transformation and on the stability of an intermediate along a synthesis route (Klucznik et al., <xref ref-type="bibr" rid="B93">2018</xref>).</p>
<p>Reaction prediction and retrosynthesis are the mainstays of organic chemistry. Retrosynthesis has been used for planning synthesis of small organic molecules, in which target molecules are recursively converted into progressively simpler precursors (Segler and Waller, <xref ref-type="bibr" rid="B143">2017</xref>). However, the results obtained from the <italic>in silico</italic> version of this process are not, in general, adequate. Rule-based procedures have been extensively employed for solving, computationally, both reaction prediction and retrosynthesis. However, reactivity conflicts are often generated, since reaction rules tend to ignore the molecular context. It is often difficult to predict how a compound would behave in practice, unless an experiment is carried out (Granda et al., <xref ref-type="bibr" rid="B61">2018</xref>). Evaluating a candidate sequence of reaction steps means that the synthesis of a given compound is also difficult. In chemical synthesis planning, Szymku&#x00107; et al. (<xref ref-type="bibr" rid="B154">2016</xref>) have discussed these issues. Segler and Waller have reported (Segler et al., <xref ref-type="bibr" rid="B142">2018</xref>) that the prioritization of the most suitable conversion rules, as well as the approach to conflicting or complexity raising issues can be achieved by learning with DLNs. The authors have trained their model on ca. three million reactions, exhibiting accuracies of 97 and 95% for reaction prediction and retrosynthesis, respectively, on a validation set of ca. one million reactions. Following this procedure, the same authors have applied Monte Carlo tree search and symbolic artificial intelligence to find retrosynthetic routes. DLNs were trained on the whole set of published organic reactions (Segler et al., <xref ref-type="bibr" rid="B142">2018</xref>).</p>
<p>Coley et al. (<xref ref-type="bibr" rid="B34">2017</xref>, <xref ref-type="bibr" rid="B36">2018b</xref>) have performed DL with features based on the alterations of reactants and have determined scores for putative products. The product was modeled as a true target molecule (product) if it was generated by a reaction covered by the patent literature, and as a false product otherwise. More recently Coley et al. (<xref ref-type="bibr" rid="B36">2018b</xref>) have put forward a new definition addressing the synthetic complexity in order to compare with the expected number of reaction steps required for producing target molecules, with known compounds as reasonable starting materials. Specifically, a neural network model was trained on 12 million reactions from the Reaxys database, imposing a pairwise inequality constraint and showing that the products of published chemical reaction are, on average, more synthetically complex than their corresponding reactants.</p>
<p>A graph-link-prediction-based procedure was formulated by Savage et al. (<xref ref-type="bibr" rid="B138">2017</xref>) to predict candidate molecules (reactants), given a target molecule (product) as input and to discover adequate synthesis routes for producing the targets. This was employed over the Network of Organic Chemistry constructed from eight million chemical reactions described in the US patent literature in the 1976&#x02013;2013 period (Savage et al., <xref ref-type="bibr" rid="B138">2017</xref>). The proposed evaluation demonstrated that Factorization Machines, trained with chemistry-specific information, outperforms similarity-based methods of chemical structures. In these approaches, a fingerprint is built from a graphical representation of the molecule, containing the respective structural information and chemical features. The latter can be selected using different approaches (Morgan, <xref ref-type="bibr" rid="B119">1965</xref>; Rogers and Hahn, <xref ref-type="bibr" rid="B129">2010</xref>). Some neural graph fingerprints have displayed significant predictive performance (Duvenaud et al., <xref ref-type="bibr" rid="B44">2015</xref>). The detection of molecular active substructures (e.g., a moiety impacting on a disease and a moiety that confers structural stability) can also be performed with ML (Duvenaud et al., <xref ref-type="bibr" rid="B44">2015</xref>).</p>
<p>Researchers have also designed a chemical-handling robot for screening and predicting chemical reactivity using ML. The authors have found four novel reactions, demonstrating the respective potential in discovering reactions. Chemical reactions related to many different pathways can lead to a desired molecule. To find the best pathways, discovering new chemical reactivity is crucial to make the processes that produce chemicals, pharmaceuticals and materials more sustainable, environmentally-friendly and efficient. However, discovering new reactions is usually an unpredictable and time-consuming process that&#x00027;s constrained by a top-down approach involving expert knowledge to target a particular molecule.</p>
<p>Other researchers (Granda et al., <xref ref-type="bibr" rid="B61">2018</xref>) have created an organic synthesis robotic ML system able to explore the reactivity several reagents from the bottom-up with no specific target. By performing ca. 10% of 969 possible reactions from a set of 18 reagents, the proposed system allowed predicting the reactivity of the remaining 90% of reactions with an accuracy of 86%. The database was continuously updated by performing multiple experiments based on the reactivity data collected. This allowed discovering new reactions that were inspected to isolate and characterize the new compounds (Granda et al., <xref ref-type="bibr" rid="B61">2018</xref>).</p>
</sec>
<sec>
<title>Supporting Analytical Chemistry and Catalysis</title>
<p>Analytical chemistry is possibly the area corresponding to the longest history, but also one that mostly displays embryonic applications of ML. A large number of statistical analyses and ML expert systems have been implemented in analytical chemistry for a long time (e.g., comparing and classifying mass spectra, NMR, or IR through assessments on available compounds) (Lipkowitz and Boyd, <xref ref-type="bibr" rid="B102">1995</xref>; Mayer and Baeumner, <xref ref-type="bibr" rid="B108">2019</xref>). Until recently, ML approaches were mainly employed to explain chemical reactions and to provide valuable predictive insights. Currently, it is possible to predict unexpected reactive outcomes, or relevant mechanistic insights for catalytic processes. A survey of some of these contributions can be found in Durand and Fey (<xref ref-type="bibr" rid="B43">2019</xref>).</p>
<p>Other groups (Ghosh et al., <xref ref-type="bibr" rid="B58">2019</xref>) have proposed DL methods for predicting molecular excitation spectra. Considering the electronic density of the states of 132 k organic compounds, the authors have built three different neural network architectures: a multilayer perceptron (MLP), a convolutional neural network (CNN), and a DLNs. The coordinates and charge of the atoms in each molecule were used as inputs for the neural networks. The DLNs reached the best performance with a root-mean-square error (RMSE) of 0.19 eV, while MLP and CNN were able to learn spectra with a RMSE of 0.3 and 0.23 eV, respectively. Both CNN and DLNs allowed identifying subtle variations in the spectral shape. The structures of 10 k organic molecules previously unseen were scanned and the instant predictions on spectra were obtained to identify molecules for further applications (Ghosh et al., <xref ref-type="bibr" rid="B58">2019</xref>).</p>
<p>A new computational approach, denoted as quantitative profile-profile relationship (QPPR) modeling, and based on ML techniques, has been proposed for predicting the pre-discharge chemical profiles of ammunition components from the components of the respective post-discharge gunshot residue (Gallidabino et al., <xref ref-type="bibr" rid="B55">2019</xref>). The predicted profiles can be compared with other measured profiles to perform evidential associations in forensic investigations. Specifically, the approach was optimized and assessed for the prediction of GC-MS profiles of smokeless powders (SLPs) obtained from organic gunshot residues, considering nine ammunition types. A high degree of similarity between predicted and experimentally measured profiles was found, after applying 14 ML techniques, with a median correlation of 0.982 (Gallidabino et al., <xref ref-type="bibr" rid="B55">2019</xref>). Receiver operating characteristic (ROC) analysis was employed to assess association performances, and allowed comparing predicted&#x02013;predicted and predicted&#x02013;measured profiles, producing areas under the curve (AUCs) of 0.976 and 0.824, respectively, in extrapolation mode. On the other hand, AUCs of 0.962 and 0.894 were obtained in the interpolation mode. These results were approximated to the values obtained from the comparison of the measured SLP profiles (AUC &#x0003D; 0.998), demonstrating excellent potential to correctly associate evidence in a number of different forensic situations (Gallidabino et al., <xref ref-type="bibr" rid="B55">2019</xref>). The advantages of this approach are numerous and may be extended to other fields in analytical sciences that routinely experience mutable chemical signatures, including the analysis of explosive devices, toxicological samples and environmental pollutants (Gallidabino et al., <xref ref-type="bibr" rid="B55">2019</xref>).</p>
<p>The integration of ML-based algorithms in a chemosensor has also pointed out the future of ML and the artificial internet of things applicability, i.e., optimized sensors, linked to a central data analysis unit via wireless (Mayer and Baeumner, <xref ref-type="bibr" rid="B108">2019</xref>).</p>
<p>Additionally, researchers have used ML to develop tools for predicting catalytic components and dynamics. For instance, the identification and prediction of ligands for metal-catalyzed coupling reaction have been conducted for designing a synthetic economic and ecological route, with the potential to be expanded into a system of pharmaceutical interest (Durand and Fey, <xref ref-type="bibr" rid="B43">2019</xref>). Durand and Fey have recently summarized calculations of several ligand descriptors, focusing on homogeneous organometallic catalysis. Different approaches for calculating steric and electronic parameters were also reviewed and assessed, and a set of descriptors for a wide range of ligands (e.g., 30 monodentate phosphorus (III) donor ligands, 23 bidentate P,P-donor ligands, and 30 carbenes) were collected.</p>
<p>Different case studies covering the application of these descriptors, including maps and models and DFT calculations, have been discussed, demonstrating the usefulness of descriptor-oriented studies of catalysis for guiding experiments and successfully evaluate and compare the proposed models (Durand and Fey, <xref ref-type="bibr" rid="B43">2019</xref>).</p>
<p>Li and Eastgate (<xref ref-type="bibr" rid="B100">2019</xref>) have designed a ML-based tool for acting on transition metal-catalyzed carbon&#x02013;nitrogen coupling reactions encompassing phosphine ligands, which are often involved in pharmaceutical syntheses. The data set of the system was composed of literature documents reporting coupling reactions with phosphine ligands. The input variables were the molecular features of ligand electrophiles and nucleophiles, and the phosphine ligands were de output obtained in successful reactions. The tools used substrate fingerprints, to build a multiclass predictive model and identify the ligands prone to function in a Pd-catalyzed C&#x02013;N coupling reaction. The resulting probabilities were associated to the corresponding ligand (cPMIs) to determine a probability-weighted predicted holistic PMI for the transformation, considering the synthesis of the ligand. This novel ML approach were developed for estimating the probability of success for ligands, given specified electrophile and nucleophile combinations, illustrated in the a Pd-catalyzed C&#x02013;N coupling context. The neural network allowed thus improving the predictive performance of the top-N accuracy over other ML approaches. Further application of this tool will foster the development of frameworks based on criteria-decision analytics, optimizing the design of manufacturing processes.</p>
<p>Designing catalysts using computational approaches is also a major challenge in chemistry. Conventional approaches have been restricted to calculate properties for a complex and large number of potential catalysts. More recently, innovative approaches for inverse design have emerged, for finding the desired property and optimizing the respective chemical structure. The chemical space has been explored by combining gradient-based optimization, alchemical transformations, and ML. These efforts have been duly reviewed in the context of inverse design and relevance to developing catalytic technologies (Freeze et al., <xref ref-type="bibr" rid="B53">2019</xref>). These approaches have offered new opportunities for identifying catalysts using efficient methods that circumvent the need for high-throughput screening and reduce the array of compounds and materials displaying the target properties and can be experimentally validated. For instance, inverse design can be employed for modulating catalytic activity via alterations in the first and second coordination spheres of the catalyst binding site (e.g., functionality of catalytic cofactors in enzymes).</p>
<p>One possible approach to inverse design is to use the synthetic accessibility score, commonly used for drug molecules, in the scoring functions of inverse design for ensuring synthetic feasibility. For that purpose, empirical parameters can be used to describe molecules without the cost of using 3D coordinates for an entire structure and without using a model to describe the complex interactions from geometries.</p>
<p>The major progress on inverse design relies on optimization algorithms, which govern the process for exploring a specific space, improving identification rates of parameters that allows optimizing the value of the scoring function. For example, the Classical Optimal Control Optimization algorithm, used for global energy minimization, is based on the diffeomorphic modulation under the observable-response-preserving homotopy algorithm, and lead the classical dynamics of a probe particle, driven by an external field for reaching the optimal value of a multidimensional function, by adjusting iteratively field control parameters over the gradient of the scoring function related to the controls. However, the respective use for scoring functions in inverse design applications still remain a challenge (Freeze et al., <xref ref-type="bibr" rid="B53">2019</xref>). Scoring functions allow correlating molecular descriptors to catalytic properties for finding catalysts via gradient-based optimization. In a simple example, similar molecules often display distinct catalytic activity due to subtle effects that must be detected by scoring functions. Such effects may be determined by combining experimentation to build adequate training sets of systems with different values of selected properties for determining feature sets able to detect such properties. ML can also be used to evaluate performance scores for GA-based methods.</p>
<p>The application of autoencoders have allowed transforming SMILES representations of compounds into a continuous latent space in order to optimize chemical properties, including synthetic accessibility score and Quantitative Estimation of Drug Likeness. Additionally, by resorting to gradient-based methods the latent space can be intersected to predict new candidate structures for being synthesized and tested.</p>
<p>The integration of inverse design, gradient-based optimization and ML is a very promising strategy to shorten the long path toward catalyst discovery (Freeze et al., <xref ref-type="bibr" rid="B53">2019</xref>). Also, other related methods that have been implemented to scrutinize the chemical space for drug development can be applied for catalyst discovery, as described in Freeze et al. (<xref ref-type="bibr" rid="B53">2019</xref>).</p>
</sec>
</sec>
<sec id="s4">
<title>Concluding Remarks</title>
<p>This review has sought to provide a sample of ML approaches that support the major research trends in Chemistry, especially in computational chemistry, focusing on DLNs. Such an approaches have offered the possibility of solving chemical problems that cannot be described and explained via conventional methods. In the last few years, the application of ML to the optimization and prediction of molecular properties has become very popular, since more researchers are trained and acquired technical skills to develop and use such methods. ML applications are area-dependent and follow, in fact, a more or less obvious pattern. For instance, medicinal chemistry excels in structure-activity relationships. In other words, each sub-field is progressing essentially in activities that belong to its core subjects. It seems that these fields are evolving naturally, and we cannot identify significant disruptive trends.</p>
<p>Despite the historical route of ML methods involving the implementation of clustering or dimensionality reduction approaches, to provide a simple, low dimensional, or coarse-grained representations of structural and dynamical patterns of complex chemical systems, the interplay between innovative ML-driven predictions and molecular simulations can be combined to make time-consuming electronic calculations feasible, obtain accurate interatomic potentials on complex systems, and provide a rational design of molecules and materials. However, the convergence between different ML algorithms is a major challenge to achieve a definite progress in the chemistry fields.</p>
<p>Unsupervised learning may also contribute to elucidate the operating aspects of supervised algorithms, while supervised approaches may contribute to the formulation of objective metrics to evaluate the performance of unsupervised approaches.</p>
<p>In Chemistry DL is still at an incipient stage, particularly in computational chemistry, although the pace of contributions has been increasing very recently. One of the major challenges is the disparity, quality and interpretability of the generated outcomes. Paired with the sophistication and ability of GPU-accelerated computing for training DLNs and the massive growth of chemical information used for training DLNs, it is anticipated that DL algorithms will be an invaluable engine for computational chemistry. DL uses a hierarchical cascade of non-linear functions allowing to learn representations and capture the required features from raw chemical data, necessary for predicting target physicochemical properties.</p>
<p>Considering the recent effort on the topic, DL models have been implemented in various Chemistry sub-fields, including quantum-chemistry, compound and materials design, with superior performances to conventional ML algorithms. There is still tremendous room for improved model accuracy and interpretability. While industrial sectors will continue driving many advances, academia will continue playing a critical role in supplying innovative technical and practical contributions, as well as in fostering cross-disciplinary cooperation.</p>
</sec>
<sec id="s5">
<title>Author Contributions</title>
<p>TC performed the bibliometrics analysis, collected the relevant studies in the context of the review, structured, and wrote the paper. AP directed the work, contributed to the interpretation of the data, and to the structure of the review. Both authors reviewed the manuscript.</p>
<sec>
<title>Conflict of Interest</title>
<p>The authors declare that the research was conducted in the absence of any commercial or financial relationships that could be construed as a potential conflict of interest.</p>
</sec>
</sec>
</body>
<back>
<ref-list>
<title>References</title>
<ref id="B1">
<citation citation-type="journal"><person-group person-group-type="author"><name><surname>Agrafiotis</surname> <given-names>D. K.</given-names></name> <name><surname>Cede&#x000F1;o</surname> <given-names>W.</given-names></name> <name><surname>Lobanov</surname> <given-names>V. S.</given-names></name></person-group> (<year>2002</year>). <article-title>On the use of neural network ensembles in QSAR and QSPR</article-title>. <source>J. Chem. Inf. Comput. Sci.</source> <volume>42</volume>, <fpage>903</fpage>&#x02013;<lpage>911</lpage>. <pub-id pub-id-type="doi">10.1021/ci0203702</pub-id><pub-id pub-id-type="pmid">12132892</pub-id></citation></ref>
<ref id="B2">
<citation citation-type="journal"><person-group person-group-type="author"><name><surname>Ahn</surname> <given-names>S.</given-names></name> <name><surname>Hong</surname> <given-names>M.</given-names></name> <name><surname>Sundararajan</surname> <given-names>M.</given-names></name> <name><surname>Ess</surname> <given-names>D. H.</given-names></name> <name><surname>Baik</surname> <given-names>M.-H.</given-names></name></person-group> (<year>2019</year>). <article-title>Design and optimization of catalysts based on mechanistic insights derived from quantum chemical reaction modeling</article-title>. <source>Chem. Rev</source>. <volume>119</volume>, <fpage>6509</fpage>&#x02013;<lpage>6560</lpage>. <pub-id pub-id-type="doi">10.1021/acs.chemrev.9b00073</pub-id><pub-id pub-id-type="pmid">31066549</pub-id></citation></ref>
<ref id="B3">
<citation citation-type="journal"><person-group person-group-type="author"><name><surname>Ahneman</surname> <given-names>D. T.</given-names></name> <name><surname>Estrada</surname> <given-names>J. G.</given-names></name> <name><surname>Lin</surname> <given-names>S.</given-names></name> <name><surname>Dreher</surname> <given-names>S. D.</given-names></name> <name><surname>Doyle</surname> <given-names>A. G.</given-names></name></person-group> (<year>2018</year>). <article-title>Predicting reaction performance in C&#x02013;N cross-coupling using machine learning</article-title>. <source>Science</source> <volume>360</volume>, <fpage>186</fpage>&#x02013;<lpage>190</lpage>. <pub-id pub-id-type="doi">10.1126/science.aar5169</pub-id><pub-id pub-id-type="pmid">29449509</pub-id></citation></ref>
<ref id="B4">
<citation citation-type="journal"><person-group person-group-type="author"><name><surname>Alipanahi</surname> <given-names>B.</given-names></name> <name><surname>Delong</surname> <given-names>A.</given-names></name> <name><surname>Weirauch</surname> <given-names>M. T.</given-names></name> <name><surname>Frey</surname> <given-names>B. J.</given-names></name></person-group> (<year>2015</year>). <article-title>Predicting the sequence specificities of DNA- and RNA-binding proteins by deep learning</article-title>. <source>Nat. Biotechnol.</source> <volume>33</volume>, <fpage>831</fpage>&#x02013;<lpage>838</lpage>. <pub-id pub-id-type="doi">10.1038/nbt.3300</pub-id><pub-id pub-id-type="pmid">26213851</pub-id></citation></ref>
<ref id="B5">
<citation citation-type="journal"><person-group person-group-type="author"><name><surname>Artrith</surname> <given-names>N.</given-names></name> <name><surname>Urban</surname> <given-names>A.</given-names></name></person-group> (<year>2016</year>). <article-title>An implementation of artificial neural-network potentials for atomistic materials simulations: performance for TiO<sub>2</sub></article-title>. <source>Comput. Mater. Sci.</source> <volume>114</volume>, <fpage>135</fpage>&#x02013;<lpage>150</lpage>. <pub-id pub-id-type="doi">10.1016/j.commatsci.2015.11.047</pub-id></citation></ref>
<ref id="B6">
<citation citation-type="journal"><person-group person-group-type="author"><name><surname>Ar&#x000FA;s-Pous</surname> <given-names>J.</given-names></name> <name><surname>Blaschke</surname> <given-names>T.</given-names></name> <name><surname>Ulander</surname> <given-names>S.</given-names></name> <name><surname>Reymond</surname> <given-names>J.-L.</given-names></name> <name><surname>Chen</surname> <given-names>H.</given-names></name> <name><surname>Engkvist</surname> <given-names>O.</given-names></name></person-group> (<year>2019</year>). <article-title>Exploring the GDB-13 chemical space using deep generative models</article-title>. <source>J. Cheminform.</source> <volume>11</volume>:<fpage>20</fpage>. <pub-id pub-id-type="doi">10.1186/s13321-019-0341-z</pub-id><pub-id pub-id-type="pmid">30868314</pub-id></citation></ref>
<ref id="B7">
<citation citation-type="journal"><person-group person-group-type="author"><name><surname>Aspuru-Guzik</surname> <given-names>A.</given-names></name> <name><surname>Baik</surname> <given-names>M.-H.</given-names></name> <name><surname>Balasubramanian</surname> <given-names>S.</given-names></name> <name><surname>Banerjee</surname> <given-names>R.</given-names></name> <name><surname>Bart</surname> <given-names>S.</given-names></name> <name><surname>Borduas-Dedekind</surname> <given-names>N.</given-names></name> <etal/></person-group>. (<year>2019</year>). <article-title>Charting a course for chemistry</article-title>. <source>Nat. Chem.</source> <volume>11</volume>, <fpage>286</fpage>&#x02013;<lpage>294</lpage>. <pub-id pub-id-type="doi">10.1038/s41557-019-0236-7</pub-id><pub-id pub-id-type="pmid">30903035</pub-id></citation></ref>
<ref id="B8">
<citation citation-type="journal"><person-group person-group-type="author"><name><surname>Bart&#x000F3;k</surname> <given-names>A. P.</given-names></name> <name><surname>Kondor</surname> <given-names>R.</given-names></name> <name><surname>Cs&#x000E1;nyi</surname> <given-names>G.</given-names></name></person-group> (<year>2013</year>). <article-title>On representing chemical environments</article-title>. <source>Phys. Rev. B</source> <volume>87</volume>:<fpage>184115</fpage>. <pub-id pub-id-type="doi">10.1103/PhysRevB.87.184115</pub-id></citation></ref>
<ref id="B9">
<citation citation-type="journal"><person-group person-group-type="author"><name><surname>Bart&#x000F3;k</surname> <given-names>A. P.</given-names></name> <name><surname>Payne</surname> <given-names>M. C.</given-names></name> <name><surname>Kondor</surname> <given-names>R.</given-names></name> <name><surname>Cs&#x000E1;nyi</surname> <given-names>G.</given-names></name></person-group> (<year>2010</year>). <article-title>Gaussian approximation potentials: the accuracy of quantum mechanics, without the electrons</article-title>. <source>Phys. Rev. Lett.</source> <volume>104</volume>:<fpage>136403</fpage>. <pub-id pub-id-type="doi">10.1103/PhysRevLett.104.136403</pub-id><pub-id pub-id-type="pmid">20481899</pub-id></citation></ref>
<ref id="B10">
<citation citation-type="journal"><person-group person-group-type="author"><name><surname>Behler</surname> <given-names>J.</given-names></name></person-group> (<year>2011a</year>). <article-title>Atom-centered symmetry functions for constructing high-dimensional neural network potentials</article-title>. <source>J. Chem. Phys.</source> <volume>134</volume>:<fpage>074106</fpage>. <pub-id pub-id-type="doi">10.1063/1.3553717</pub-id><pub-id pub-id-type="pmid">21341827</pub-id></citation></ref>
<ref id="B11">
<citation citation-type="journal"><person-group person-group-type="author"><name><surname>Behler</surname> <given-names>J.</given-names></name></person-group> (<year>2011b</year>). <article-title>Neural network potential-energy surfaces in chemistry: a tool for large-scale simulations</article-title>. <source>Phys. Chem. Chem. Phys.</source> <volume>13</volume>, <fpage>17930</fpage>&#x02013;<lpage>17955</lpage>. <pub-id pub-id-type="doi">10.1039/c1cp21668f</pub-id><pub-id pub-id-type="pmid">21915403</pub-id></citation></ref>
<ref id="B12">
<citation citation-type="journal"><person-group person-group-type="author"><name><surname>Behler</surname> <given-names>J.</given-names></name></person-group> (<year>2016</year>). <article-title>Perspective: machine learning potentials for atomistic simulations</article-title>. <source>J. Chem. Phys.</source> <volume>145</volume>:<fpage>170901</fpage>. <pub-id pub-id-type="doi">10.1063/1.4966192</pub-id></citation></ref>
<ref id="B13">
<citation citation-type="journal"><person-group person-group-type="author"><name><surname>Behler</surname> <given-names>J.</given-names></name> <name><surname>Parrinello</surname> <given-names>M.</given-names></name></person-group> (<year>2007</year>). <article-title>Generalized neural-network representation of high-dimensional potential-energy surfaces</article-title>. <source>Phys. Rev. Lett.</source> <volume>98</volume>:<fpage>146401</fpage>. <pub-id pub-id-type="doi">10.1103/PhysRevLett.98.146401</pub-id><pub-id pub-id-type="pmid">17501293</pub-id></citation></ref>
<ref id="B14">
<citation citation-type="journal"><person-group person-group-type="author"><name><surname>Benjamin</surname> <given-names>S.-L.</given-names></name> <name><surname>Carlos</surname> <given-names>O.</given-names></name> <name><surname>Gabriel</surname> <given-names>L. G.</given-names></name> <name><surname>Alan</surname> <given-names>A.-G.</given-names></name></person-group> (<year>2017</year>). <article-title>Optimizing Distributions Over Molecular Space. An Objective-Reinforced Generative Adversarial Network for Inverse-design Chemistry (ORGANIC)</article-title>. <source>ChemRxiv [Preprint].</source> <pub-id pub-id-type="doi">10.26434/chemrxiv.5309668.v3</pub-id></citation></ref>
<ref id="B15">
<citation citation-type="journal"><person-group person-group-type="author"><name><surname>Bose</surname> <given-names>S.</given-names></name> <name><surname>Dhawan</surname> <given-names>D.</given-names></name> <name><surname>Nandi</surname> <given-names>S.</given-names></name> <name><surname>Sarkar</surname> <given-names>R. R.</given-names></name> <name><surname>Ghosh</surname> <given-names>D.</given-names></name></person-group> (<year>2018</year>). <article-title>Machine learning prediction of interaction energies in rigid water clusters</article-title>. <source>Phys. Chem. Chem. Phys.</source> <volume>20</volume>, <fpage>22987</fpage>&#x02013;<lpage>22996</lpage>. <pub-id pub-id-type="doi">10.1039/C8CP03138J</pub-id><pub-id pub-id-type="pmid">30156235</pub-id></citation></ref>
<ref id="B16">
<citation citation-type="journal"><person-group person-group-type="author"><name><surname>Botu</surname> <given-names>V.</given-names></name> <name><surname>Batra</surname> <given-names>R.</given-names></name> <name><surname>Chapman</surname> <given-names>J.</given-names></name> <name><surname>Ramprasad</surname> <given-names>R.</given-names></name></person-group> (<year>2017</year>). <article-title>Machine learning force fields: construction, validation, and outlook</article-title>. <source>J. Phys. Chem. C</source> <volume>121</volume>, <fpage>511</fpage>&#x02013;<lpage>522</lpage>. <pub-id pub-id-type="doi">10.1021/acs.jpcc.6b10908</pub-id></citation></ref>
<ref id="B17">
<citation citation-type="journal"><person-group person-group-type="author"><name><surname>Botu</surname> <given-names>V.</given-names></name> <name><surname>Ramprasad</surname> <given-names>R.</given-names></name></person-group> (<year>2015a</year>). <article-title>Adaptive machine learning framework to accelerate ab initio molecular dynamics</article-title>. <source>Int. J. Quantum Chem.</source> <volume>115</volume>, <fpage>1074</fpage>&#x02013;<lpage>1083</lpage>. <pub-id pub-id-type="doi">10.1002/qua.24836</pub-id></citation></ref>
<ref id="B18">
<citation citation-type="journal"><person-group person-group-type="author"><name><surname>Botu</surname> <given-names>V.</given-names></name> <name><surname>Ramprasad</surname> <given-names>R.</given-names></name></person-group> (<year>2015b</year>). <article-title>Learning scheme to predict atomic forces and accelerate materials simulations</article-title>. <source>Phys. Rev. B</source> <volume>92</volume>:<fpage>094306</fpage>. <pub-id pub-id-type="doi">10.1103/PhysRevB.92.094306</pub-id></citation></ref>
<ref id="B19">
<citation citation-type="journal"><person-group person-group-type="author"><name><surname>Brockherde</surname> <given-names>F.</given-names></name> <name><surname>Vogt</surname> <given-names>L.</given-names></name> <name><surname>Li</surname> <given-names>L.</given-names></name> <name><surname>Tuckerman</surname> <given-names>M. E.</given-names></name> <name><surname>Burke</surname> <given-names>K.</given-names></name> <name><surname>M&#x000FC;ller</surname> <given-names>K.-R.</given-names></name></person-group> (<year>2017</year>). <article-title>Bypassing the Kohn-Sham equations with machine learning</article-title>. <source>Nat. Commun.</source> <volume>8</volume>:<fpage>872</fpage>. <pub-id pub-id-type="doi">10.1038/s41467-017-00839-3</pub-id><pub-id pub-id-type="pmid">29021555</pub-id></citation></ref>
<ref id="B20">
<citation citation-type="journal"><person-group person-group-type="author"><name><surname>Brown</surname> <given-names>N.</given-names></name> <name><surname>Fiscato</surname> <given-names>M.</given-names></name> <name><surname>Segler</surname> <given-names>M. H. S.</given-names></name> <name><surname>Vaucher</surname> <given-names>A. C.</given-names></name></person-group> (<year>2019</year>). <article-title>GuacaMol: benchmarking models for <italic>de novo</italic> molecular design</article-title>. <source>J. Chem. Inf. Model.</source> <volume>59</volume>, <fpage>1096</fpage>&#x02013;<lpage>1108</lpage>. <pub-id pub-id-type="doi">10.1021/acs.jcim.8b00839</pub-id><pub-id pub-id-type="pmid">30887799</pub-id></citation></ref>
<ref id="B21">
<citation citation-type="journal"><person-group person-group-type="author"><name><surname>Butler</surname> <given-names>K. T.</given-names></name> <name><surname>Davies</surname> <given-names>D. W.</given-names></name> <name><surname>Cartwright</surname> <given-names>H.</given-names></name> <name><surname>Isayev</surname> <given-names>O.</given-names></name> <name><surname>Walsh</surname> <given-names>A.</given-names></name></person-group> (<year>2018</year>). <article-title>Machine learning for molecular and materials science</article-title>. <source>Nature</source> <volume>559</volume>, <fpage>547</fpage>&#x02013;<lpage>555</lpage>. <pub-id pub-id-type="doi">10.1038/s41586-018-0337-2</pub-id><pub-id pub-id-type="pmid">30046072</pub-id></citation></ref>
<ref id="B22">
<citation citation-type="journal"><person-group person-group-type="author"><name><surname>Ceriotti</surname> <given-names>M.</given-names></name></person-group> (<year>2019</year>). <article-title>Unsupervised machine learning in atomistic simulations, between predictions and understanding</article-title>. <source>J. Chem. Phys.</source> <volume>150</volume>:<fpage>150901</fpage>. <pub-id pub-id-type="doi">10.1063/1.5091842</pub-id><pub-id pub-id-type="pmid">31005087</pub-id></citation></ref>
<ref id="B23">
<citation citation-type="journal"><person-group person-group-type="author"><name><surname>Chakravarti</surname> <given-names>S. K.</given-names></name></person-group> (<year>2018</year>). <article-title>Distributed representation of chemical fragments</article-title>. <source>ACS Omega</source> <volume>3</volume>, <fpage>2825</fpage>&#x02013;<lpage>2836</lpage>. <pub-id pub-id-type="doi">10.1021/acsomega.7b02045</pub-id><pub-id pub-id-type="pmid">30023852</pub-id></citation></ref>
<ref id="B24">
<citation citation-type="journal"><person-group person-group-type="author"><name><surname>Chan</surname> <given-names>H.</given-names></name> <name><surname>Cherukara</surname> <given-names>M. J.</given-names></name> <name><surname>Narayanan</surname> <given-names>B.</given-names></name> <name><surname>Loeffler</surname> <given-names>T. D.</given-names></name> <name><surname>Benmore</surname> <given-names>C.</given-names></name> <name><surname>Gray</surname> <given-names>S. K.</given-names></name> <etal/></person-group>. (<year>2019</year>). <article-title>Machine learning coarse grained models for water</article-title>. <source>Nat. Commun.</source> <volume>10</volume>:<fpage>379</fpage>. <pub-id pub-id-type="doi">10.1038/s41467-018-08222-6</pub-id><pub-id pub-id-type="pmid">30670699</pub-id></citation></ref>
<ref id="B25">
<citation citation-type="journal"><person-group person-group-type="author"><name><surname>Chandrasekaran</surname> <given-names>A.</given-names></name> <name><surname>Kamal</surname> <given-names>D.</given-names></name> <name><surname>Batra</surname> <given-names>R.</given-names></name> <name><surname>Kim</surname> <given-names>C.</given-names></name> <name><surname>Chen</surname> <given-names>L.</given-names></name> <name><surname>Ramprasad</surname> <given-names>R.</given-names></name></person-group> (<year>2019</year>). <article-title>Solving the electronic structure problem with machine learning</article-title>. <source>NPJ Comput. Mater.</source> <volume>5</volume>:<fpage>22</fpage>. <pub-id pub-id-type="doi">10.1038/s41524-019-0162-7</pub-id></citation></ref>
<ref id="B26">
<citation citation-type="journal"><person-group person-group-type="author"><name><surname>Chen</surname> <given-names>H.</given-names></name> <name><surname>Engkvist</surname> <given-names>O.</given-names></name> <name><surname>Wang</surname> <given-names>Y.</given-names></name> <name><surname>Olivecrona</surname> <given-names>M.</given-names></name> <name><surname>Blaschke</surname> <given-names>T.</given-names></name></person-group> (<year>2018</year>). <article-title>The rise of deep learning in drug discovery</article-title>. <source>Drug Discov. Today</source> <volume>23</volume>, <fpage>1241</fpage>&#x02013;<lpage>1250</lpage>. <pub-id pub-id-type="doi">10.1016/j.drudis.2018.01.039</pub-id><pub-id pub-id-type="pmid">29366762</pub-id></citation></ref>
<ref id="B27">
<citation citation-type="journal"><person-group person-group-type="author"><name><surname>Chen</surname> <given-names>M.</given-names></name> <name><surname>Yu</surname> <given-names>T.-Q.</given-names></name> <name><surname>Tuckerman</surname> <given-names>M. E.</given-names></name></person-group> (<year>2015</year>). <article-title>Locating landmarks on high-dimensional free energy surfaces</article-title>. <source>Proc. Natl. Acad. Sci. U.S.A.</source> <volume>112</volume>:<fpage>3235</fpage>. <pub-id pub-id-type="doi">10.1073/pnas.1418241112</pub-id><pub-id pub-id-type="pmid">25737545</pub-id></citation></ref>
<ref id="B28">
<citation citation-type="journal"><person-group person-group-type="author"><name><surname>Chmiela</surname> <given-names>S.</given-names></name> <name><surname>Sauceda</surname> <given-names>H. E.</given-names></name> <name><surname>M&#x000FC;ller</surname> <given-names>K.-R.</given-names></name> <name><surname>Tkatchenko</surname> <given-names>A.</given-names></name></person-group> (<year>2018</year>). <article-title>Towards exact molecular dynamics simulations with machine-learned force fields</article-title>. <source>Nat. Commun.</source> <volume>9</volume>, <fpage>3887</fpage>&#x02013;<lpage>3887</lpage>. <pub-id pub-id-type="doi">10.1038/s41467-018-06169-2</pub-id><pub-id pub-id-type="pmid">30250077</pub-id></citation></ref>
<ref id="B29">
<citation citation-type="journal"><person-group person-group-type="author"><name><surname>Chmiela</surname> <given-names>S.</given-names></name> <name><surname>Tkatchenko</surname> <given-names>A.</given-names></name> <name><surname>Sauceda</surname> <given-names>H. E.</given-names></name> <name><surname>Poltavsky</surname> <given-names>I.</given-names></name> <name><surname>Sch&#x000FC;tt</surname> <given-names>K. T.</given-names></name> <name><surname>M&#x000FC;ller</surname> <given-names>K.-R.</given-names></name></person-group> (<year>2017</year>). <article-title>Machine learning of accurate energy-conserving molecular force fields</article-title>. <source>Sci. Adv.</source> <volume>3</volume>:<fpage>e1603015</fpage>. <pub-id pub-id-type="doi">10.1126/sciadv.1603015</pub-id><pub-id pub-id-type="pmid">28508076</pub-id></citation></ref>
<ref id="B30">
<citation citation-type="journal"><person-group person-group-type="author"><name><surname>Choi</surname> <given-names>H.</given-names></name> <name><surname>Kang</surname> <given-names>H.</given-names></name> <name><surname>Chung</surname> <given-names>K.-C.</given-names></name> <name><surname>Park</surname> <given-names>H.</given-names></name></person-group> (<year>2019</year>). <article-title>Development and application of a comprehensive machine learning program for predicting molecular biochemical and pharmacological properties</article-title>. <source>Phys. Chem. Chem. Phys.</source> <volume>21</volume>, <fpage>5189</fpage>&#x02013;<lpage>5199</lpage>. <pub-id pub-id-type="doi">10.1039/C8CP07002D</pub-id><pub-id pub-id-type="pmid">30775759</pub-id></citation></ref>
<ref id="B31">
<citation citation-type="journal"><person-group person-group-type="author"><name><surname>Christensen</surname> <given-names>A. S.</given-names></name> <name><surname>Faber</surname> <given-names>F. A.</given-names></name> <name><surname>von Lilienfeld</surname> <given-names>O. A.</given-names></name></person-group> (<year>2019</year>). <article-title>Operators in quantum machine learning: Response properties in chemical space</article-title>. <source>J. Chem. Phys.</source> <volume>150</volume>:<fpage>064105</fpage>. <pub-id pub-id-type="doi">10.1063/1.5053562</pub-id><pub-id pub-id-type="pmid">30769998</pub-id></citation></ref>
<ref id="B32">
<citation citation-type="journal"><person-group person-group-type="author"><name><surname>Chuang</surname> <given-names>K. V.</given-names></name> <name><surname>Keiser</surname> <given-names>M. J.</given-names></name></person-group> (<year>2018a</year>). <article-title>Adversarial controls for scientific machine learning</article-title>. <source>ACS Chem. Biol.</source> <volume>13</volume>, <fpage>2819</fpage>&#x02013;<lpage>2821</lpage>. <pub-id pub-id-type="doi">10.1021/acschembio.8b00881</pub-id><pub-id pub-id-type="pmid">30336670</pub-id></citation></ref>
<ref id="B33">
<citation citation-type="journal"><person-group person-group-type="author"><name><surname>Chuang</surname> <given-names>K. V.</given-names></name> <name><surname>Keiser</surname> <given-names>M. J.</given-names></name></person-group> (<year>2018b</year>). <article-title>Comment on &#x0201C;predicting reaction performance in C&#x02013;N cross-coupling using machine learning&#x0201D;</article-title>. <source>Science</source> <volume>362</volume>:<fpage>eaat8603</fpage>. <pub-id pub-id-type="doi">10.1126/science.aat8603</pub-id></citation></ref>
<ref id="B34">
<citation citation-type="journal"><person-group person-group-type="author"><name><surname>Coley</surname> <given-names>C. W.</given-names></name> <name><surname>Barzilay</surname> <given-names>R.</given-names></name> <name><surname>Jaakkola</surname> <given-names>T. S.</given-names></name> <name><surname>Green</surname> <given-names>W. H.</given-names></name> <name><surname>Jensen</surname> <given-names>K. F.</given-names></name></person-group> (<year>2017</year>). <article-title>Prediction of organic reaction outcomes using machine learning</article-title>. <source>ACS Central Sci.</source> <volume>3</volume>, <fpage>434</fpage>&#x02013;<lpage>443</lpage>. <pub-id pub-id-type="doi">10.1021/acscentsci.7b00064</pub-id><pub-id pub-id-type="pmid">28573205</pub-id></citation></ref>
<ref id="B35">
<citation citation-type="journal"><person-group person-group-type="author"><name><surname>Coley</surname> <given-names>C. W.</given-names></name> <name><surname>Green</surname> <given-names>W. H.</given-names></name> <name><surname>Jensen</surname> <given-names>K. F.</given-names></name></person-group> (<year>2018a</year>). <article-title>Machine learning in computer-aided synthesis planning</article-title>. <source>Acc. Chem. Res.</source> <volume>51</volume>, <fpage>1281</fpage>&#x02013;<lpage>1289</lpage>. <pub-id pub-id-type="doi">10.1021/acs.accounts.8b00087</pub-id><pub-id pub-id-type="pmid">29715002</pub-id></citation></ref>
<ref id="B36">
<citation citation-type="journal"><person-group person-group-type="author"><name><surname>Coley</surname> <given-names>C. W.</given-names></name> <name><surname>Rogers</surname> <given-names>L.</given-names></name> <name><surname>Green</surname> <given-names>W. H.</given-names></name> <name><surname>Jensen</surname> <given-names>K. F.</given-names></name></person-group> (<year>2018b</year>). <article-title>SCScore: synthetic complexity learned from a reaction corpus</article-title>. <source>J. Chem. Inf. Model.</source> <volume>58</volume>, <fpage>252</fpage>&#x02013;<lpage>261</lpage>. <pub-id pub-id-type="doi">10.1021/acs.jcim.7b00622</pub-id><pub-id pub-id-type="pmid">29309147</pub-id></citation></ref>
<ref id="B37">
<citation citation-type="journal"><person-group person-group-type="author"><name><surname>Cort&#x000E9;s-Borda</surname> <given-names>D.</given-names></name> <name><surname>Kutonova</surname> <given-names>K. V.</given-names></name> <name><surname>Jamet</surname> <given-names>C.</given-names></name> <name><surname>Trusova</surname> <given-names>M. E.</given-names></name> <name><surname>Zammattio</surname> <given-names>F.</given-names></name> <name><surname>Truchet</surname> <given-names>C.</given-names></name> <etal/></person-group>. (<year>2016</year>). <article-title>Optimizing the Heck&#x02013;Matsuda reaction in flow with a constraint-adapted direct search algorithm</article-title>. <source>Organ. Process Res. Dev.</source> <volume>20</volume>, <fpage>1979</fpage>&#x02013;<lpage>1987</lpage>. <pub-id pub-id-type="doi">10.1021/acs.oprd.6b00310</pub-id></citation></ref>
<ref id="B38">
<citation citation-type="journal"><person-group person-group-type="author"><name><surname>Coveney Peter</surname> <given-names>V.</given-names></name> <name><surname>Dougherty Edward</surname> <given-names>R.</given-names></name> <name><surname>Highfield Roger</surname> <given-names>R.</given-names></name></person-group> (<year>2016</year>). <article-title>Big data need big theory too</article-title>. <source>Philos. Trans. R. Soc. Math. Phys. Eng. Sci.</source> <volume>374</volume>:<fpage>20160153</fpage>. <pub-id pub-id-type="doi">10.1098/rsta.2016.0153</pub-id><pub-id pub-id-type="pmid">27698035</pub-id></citation></ref>
<ref id="B39">
<citation citation-type="journal"><person-group person-group-type="author"><name><surname>Curtarolo</surname> <given-names>S.</given-names></name> <name><surname>Hart</surname> <given-names>G. L. W.</given-names></name> <name><surname>Nardelli</surname> <given-names>M. B.</given-names></name> <name><surname>Mingo</surname> <given-names>N.</given-names></name> <name><surname>Sanvito</surname> <given-names>S.</given-names></name> <name><surname>Levy</surname> <given-names>O.</given-names></name></person-group> (<year>2013</year>). <article-title>The high-throughput highway to computational materials design</article-title>. <source>Nat. Mater.</source> <volume>12</volume>, <fpage>191</fpage>&#x02013;<lpage>201</lpage>. <pub-id pub-id-type="doi">10.1038/nmat3568</pub-id><pub-id pub-id-type="pmid">23422720</pub-id></citation></ref>
<ref id="B40">
<citation citation-type="journal"><person-group person-group-type="author"><name><surname>De</surname> <given-names>S.</given-names></name> <name><surname>Bart&#x000F3;k</surname> <given-names>A. P.</given-names></name> <name><surname>Cs&#x000E1;nyi</surname> <given-names>G.</given-names></name> <name><surname>Ceriotti</surname> <given-names>M.</given-names></name></person-group> (<year>2016</year>). <article-title>Comparing molecules and solids across structural and alchemical space</article-title>. <source>Phys. Chem. Chem. Phys.</source> <volume>18</volume>, <fpage>13754</fpage>&#x02013;<lpage>13769</lpage>. <pub-id pub-id-type="doi">10.1039/C6CP00415F</pub-id><pub-id pub-id-type="pmid">27101873</pub-id></citation></ref>
<ref id="B41">
<citation citation-type="journal"><person-group person-group-type="author"><name><surname>Degiacomi</surname> <given-names>M. T.</given-names></name></person-group> (<year>2019</year>). <article-title>Coupling molecular dynamics and deep learning to mine protein conformational space</article-title>. <source>Structure</source> <volume>27</volume>, <fpage>1034</fpage>&#x02013;<lpage>1040</lpage>.e1033. <pub-id pub-id-type="doi">10.1016/j.str.2019.03.018</pub-id><pub-id pub-id-type="pmid">31031199</pub-id></citation></ref>
<ref id="B42">
<citation citation-type="journal"><person-group person-group-type="author"><name><surname>Dral</surname> <given-names>P. O.</given-names></name> <name><surname>Von Lilienfeld</surname> <given-names>O. A.</given-names></name> <name><surname>Thiel</surname> <given-names>W.</given-names></name></person-group> (<year>2015</year>). <article-title>Machine learning of parameters for accurate semiempirical quantum chemical calculations</article-title>. <source>J. Chem. Theory Comput.</source> <volume>11</volume>, <fpage>2120</fpage>&#x02013;<lpage>2125</lpage>. <pub-id pub-id-type="doi">10.1021/acs.jctc.5b00141</pub-id><pub-id pub-id-type="pmid">26146493</pub-id></citation></ref>
<ref id="B43">
<citation citation-type="journal"><person-group person-group-type="author"><name><surname>Durand</surname> <given-names>D. J.</given-names></name> <name><surname>Fey</surname> <given-names>N.</given-names></name></person-group> (<year>2019</year>). <article-title>Computational ligand descriptors for catalyst design</article-title>. <source>Chem. Rev</source>. <volume>119</volume>, <fpage>6561</fpage>&#x02013;<lpage>6594</lpage>. <pub-id pub-id-type="doi">10.1021/acs.chemrev.8b00588</pub-id><pub-id pub-id-type="pmid">30802036</pub-id></citation></ref>
<ref id="B44">
<citation citation-type="book"><person-group person-group-type="author"><name><surname>Duvenaud</surname> <given-names>D. K.</given-names></name> <name><surname>Maclaurin</surname> <given-names>D.</given-names></name> <name><surname>Iparraguirre</surname> <given-names>J.</given-names></name> <name><surname>Bombarell</surname> <given-names>R.</given-names></name> <name><surname>Hirzel</surname> <given-names>T.</given-names></name> <name><surname>Aspuru-Guzik</surname> <given-names>A.</given-names></name> <etal/></person-group>. (<year>2015</year>). <article-title>Convolutional networks on graphs for learning molecular fingerprints</article-title>, in <source>Advances in Neural Information Processing Systems</source>, eds <person-group person-group-type="editor"><name><surname>Cortes</surname> <given-names>C.</given-names></name> <name><surname>Lawrence</surname> <given-names>N. D.</given-names></name> <name><surname>Lee</surname> <given-names>D. D</given-names></name> <name><surname>Sugiyama</surname> <given-names>M.</given-names></name> <name><surname>Garnett</surname> <given-names>R.</given-names></name></person-group> (<publisher-loc>Montreal, QC</publisher-loc>), <fpage>2224</fpage>&#x02013;<lpage>2232</lpage>.</citation></ref>
<ref id="B45">
<citation citation-type="journal"><person-group person-group-type="author"><name><surname>Ekins</surname> <given-names>S.</given-names></name></person-group> (<year>2016</year>). <article-title>The next Era: deep learning in pharmaceutical research</article-title>. <source>Pharm. Res.</source> <volume>33</volume>, <fpage>2594</fpage>&#x02013;<lpage>2603</lpage>. <pub-id pub-id-type="doi">10.1007/s11095-016-2029-7</pub-id><pub-id pub-id-type="pmid">27599991</pub-id></citation></ref>
<ref id="B46">
<citation citation-type="journal"><person-group person-group-type="author"><name><surname>El-Atta</surname> <given-names>A. H. A.</given-names></name> <name><surname>Hassanien</surname> <given-names>A. E.</given-names></name></person-group> (<year>2017</year>). <article-title>Two-class support vector machine with new kernel function based on paths of features for predicting chemical activity</article-title>. <source>Inf. Sci.</source> <fpage>403</fpage>&#x02013;<lpage>404</lpage>, 42&#x02013;54. <pub-id pub-id-type="doi">10.1016/j.ins.2017.04.003</pub-id></citation></ref>
<ref id="B47">
<citation citation-type="journal"><person-group person-group-type="author"><name><surname>Elton</surname> <given-names>D. C.</given-names></name> <name><surname>Boukouvalas</surname> <given-names>Z.</given-names></name> <name><surname>Butrico</surname> <given-names>M. S.</given-names></name> <name><surname>Fuge</surname> <given-names>M. D.</given-names></name> <name><surname>Chung</surname> <given-names>P. W.</given-names></name></person-group> (<year>2018</year>). <article-title>Applying machine learning techniques to predict the properties of energetic materials</article-title>. <source>Sci. Rep.</source> <volume>8</volume>:<fpage>9059</fpage>. <pub-id pub-id-type="doi">10.1038/s41598-018-27344-x</pub-id><pub-id pub-id-type="pmid">29899464</pub-id></citation></ref>
<ref id="B48">
<citation citation-type="journal"><person-group person-group-type="author"><name><surname>Elton</surname> <given-names>D. C.</given-names></name> <name><surname>Boukouvalas</surname> <given-names>Z.</given-names></name> <name><surname>Fuge</surname> <given-names>M. D.</given-names></name> <name><surname>Chung</surname> <given-names>P. W.</given-names></name></person-group> (<year>2019</year>). <article-title>Deep learning for molecular design&#x02014;a review of the state of the art</article-title>. <source>Mol. Syst. Design Eng.</source> <volume>4</volume>, <fpage>828</fpage>&#x02013;<lpage>849</lpage>. <pub-id pub-id-type="doi">10.1039/C9ME00039A</pub-id></citation></ref>
<ref id="B49">
<citation citation-type="journal"><person-group person-group-type="author"><name><surname>Estrada</surname> <given-names>J. G.</given-names></name> <name><surname>Ahneman</surname> <given-names>D. T.</given-names></name> <name><surname>Sheridan</surname> <given-names>R. P.</given-names></name> <name><surname>Dreher</surname> <given-names>S. D.</given-names></name> <name><surname>Doyle</surname> <given-names>A. G.</given-names></name></person-group> (<year>2018</year>). <article-title>Response to comment on &#x0201C;predicting reaction performance in C&#x02013;N cross-coupling using machine learning&#x0201D;</article-title>. <source>Science</source> <volume>362</volume>:<fpage>eaat8763</fpage>. <pub-id pub-id-type="doi">10.1126/science.aat8763</pub-id></citation></ref>
<ref id="B50">
<citation citation-type="journal"><person-group person-group-type="author"><name><surname>Faber</surname> <given-names>F. A.</given-names></name> <name><surname>Hutchison</surname> <given-names>L.</given-names></name> <name><surname>Huang</surname> <given-names>B.</given-names></name> <name><surname>Gilmer</surname> <given-names>J.</given-names></name> <name><surname>Schoenholz</surname> <given-names>S. S.</given-names></name> <name><surname>Dahl</surname> <given-names>G. E.</given-names></name> <etal/></person-group>. (<year>2017</year>). <article-title>Prediction errors of molecular machine learning models lower than hybrid DFT error</article-title>. <source>J. Chem. Theory Comput.</source> <volume>13</volume>, <fpage>5255</fpage>&#x02013;<lpage>5264</lpage>. <pub-id pub-id-type="doi">10.1021/acs.jctc.7b00577</pub-id><pub-id pub-id-type="pmid">28926232</pub-id></citation></ref>
<ref id="B51">
<citation citation-type="journal"><person-group person-group-type="author"><name><surname>Faber</surname> <given-names>F. A.</given-names></name> <name><surname>Lindmaa</surname> <given-names>A.</given-names></name> <name><surname>Von Lilienfeld</surname> <given-names>O. A.</given-names></name> <name><surname>Armiento</surname> <given-names>R.</given-names></name></person-group> (<year>2016</year>). <article-title>Machine learning energies of 2 million elpasolite (ABC$_2$D$_6$) crystals</article-title>. <source>Phys. Rev. Lett.</source> <volume>117</volume>:<fpage>135502</fpage>. <pub-id pub-id-type="doi">10.1103/PhysRevLett.117.135502</pub-id><pub-id pub-id-type="pmid">27715098</pub-id></citation></ref>
<ref id="B52">
<citation citation-type="journal"><person-group person-group-type="author"><name><surname>Fleming</surname> <given-names>N.</given-names></name></person-group> (<year>2018</year>). <article-title>How artificial intelligence is changing drug discovery</article-title>. <source>Nature</source> <volume>557</volume>, <fpage>S55</fpage>&#x02013;<lpage>S55</lpage>. <pub-id pub-id-type="doi">10.1038/d41586-018-05267-x</pub-id><pub-id pub-id-type="pmid">29849160</pub-id></citation></ref>
<ref id="B53">
<citation citation-type="journal"><person-group person-group-type="author"><name><surname>Freeze</surname> <given-names>J. G.</given-names></name> <name><surname>Kelly</surname> <given-names>H. R.</given-names></name> <name><surname>Batista</surname> <given-names>V. S.</given-names></name></person-group> (<year>2019</year>). <article-title>Search for catalysts by inverse design: artificial intelligence, mountain climbers, and alchemists</article-title>. <source>Chem. Rev</source>. <volume>119</volume>, <fpage>6595</fpage>&#x02013;<lpage>6612</lpage>. <pub-id pub-id-type="doi">10.1021/acs.chemrev.8b00759</pub-id><pub-id pub-id-type="pmid">31059236</pub-id></citation></ref>
<ref id="B54">
<citation citation-type="journal"><person-group person-group-type="author"><name><surname>Fuchs</surname> <given-names>J.-A.</given-names></name> <name><surname>Grisoni</surname> <given-names>F.</given-names></name> <name><surname>Kossenjans</surname> <given-names>M.</given-names></name> <name><surname>Hiss</surname> <given-names>J. A.</given-names></name> <name><surname>Schneider</surname> <given-names>G.</given-names></name></person-group> (<year>2018</year>). <article-title>Lipophilicity prediction of peptides and peptide derivatives by consensus machine learning</article-title>. <source>Medchemcomm</source> <volume>9</volume>, <fpage>1538</fpage>&#x02013;<lpage>1546</lpage>. <pub-id pub-id-type="doi">10.1039/C8MD00370J</pub-id><pub-id pub-id-type="pmid">30288227</pub-id></citation></ref>
<ref id="B55">
<citation citation-type="journal"><person-group person-group-type="author"><name><surname>Gallidabino</surname> <given-names>M. D.</given-names></name> <name><surname>Barron</surname> <given-names>L. P.</given-names></name> <name><surname>Weyermann</surname> <given-names>C.</given-names></name> <name><surname>Romolo</surname> <given-names>F. S.</given-names></name></person-group> (<year>2019</year>). <article-title>Quantitative profile&#x02013;profile relationship (QPPR) modelling: a novel machine learning approach to predict and associate chemical characteristics of unspent ammunition from gunshot residue (GSR)</article-title>. <source>Analyst</source> <volume>144</volume>, <fpage>1128</fpage>&#x02013;<lpage>1139</lpage>. <pub-id pub-id-type="doi">10.1039/C8AN01841C</pub-id><pub-id pub-id-type="pmid">30474092</pub-id></citation></ref>
<ref id="B56">
<citation citation-type="journal"><person-group person-group-type="author"><name><surname>Gao</surname> <given-names>H.</given-names></name> <name><surname>Struble</surname> <given-names>T. J.</given-names></name> <name><surname>Coley</surname> <given-names>C. W.</given-names></name> <name><surname>Wang</surname> <given-names>Y.</given-names></name> <name><surname>Green</surname> <given-names>W. H.</given-names></name> <name><surname>Jensen</surname> <given-names>K. F.</given-names></name></person-group> (<year>2018</year>). <article-title>Using machine learning to predict suitable conditions for organic reactions</article-title>. <source>ACS Central Sci.</source> <volume>4</volume>, <fpage>1465</fpage>&#x02013;<lpage>1476</lpage>. <pub-id pub-id-type="doi">10.1021/acscentsci.8b00357</pub-id><pub-id pub-id-type="pmid">30555898</pub-id></citation></ref>
<ref id="B57">
<citation citation-type="journal"><person-group person-group-type="author"><name><surname>Gasteiger</surname> <given-names>J.</given-names></name> <name><surname>Zupan</surname> <given-names>J.</given-names></name></person-group> (<year>1993</year>). <article-title>Neural networks in chemistry</article-title>. <source>Angew. Chem. Int. Ed. Eng.</source> <volume>32</volume>, <fpage>503</fpage>&#x02013;<lpage>527</lpage>. <pub-id pub-id-type="doi">10.1002/anie.199305031</pub-id></citation></ref>
<ref id="B58">
<citation citation-type="journal"><person-group person-group-type="author"><name><surname>Ghosh</surname> <given-names>K.</given-names></name> <name><surname>Stuke</surname> <given-names>A.</given-names></name> <name><surname>Todorovi&#x00107;</surname> <given-names>M.</given-names></name> <name><surname>J&#x000F8;rgensen</surname> <given-names>P. B.</given-names></name> <name><surname>Schmidt</surname> <given-names>M. N.</given-names></name> <name><surname>Vehtari</surname> <given-names>A.</given-names></name> <etal/></person-group>. (<year>2019</year>). <article-title>Deep learning spectroscopy: neural networks for molecular excitation spectra</article-title>. <source>Adv. Sci.</source> <volume>6</volume>:<fpage>1801367</fpage>. <pub-id pub-id-type="doi">10.1002/advs.201801367</pub-id><pub-id pub-id-type="pmid">31065514</pub-id></citation></ref>
<ref id="B59">
<citation citation-type="journal"><person-group person-group-type="author"><name><surname>Goh</surname> <given-names>G. B.</given-names></name> <name><surname>Hodas</surname> <given-names>N. O.</given-names></name> <name><surname>Vishnu</surname> <given-names>A.</given-names></name></person-group> (<year>2017</year>). <article-title>Deep learning for computational chemistry</article-title>. <source>J. Comput. Chem.</source> <volume>38</volume>, <fpage>1291</fpage>&#x02013;<lpage>1307</lpage>. <pub-id pub-id-type="doi">10.1002/jcc.24764</pub-id><pub-id pub-id-type="pmid">28272810</pub-id></citation></ref>
<ref id="B60">
<citation citation-type="journal"><person-group person-group-type="author"><name><surname>G&#x000F3;mez-Bombarelli</surname> <given-names>R.</given-names></name> <name><surname>Wei</surname> <given-names>J. N.</given-names></name> <name><surname>Duvenaud</surname> <given-names>D.</given-names></name> <name><surname>Hern&#x000E1;ndez-Lobato</surname> <given-names>J. M.</given-names></name> <name><surname>S&#x000E1;nchez-Lengeling</surname> <given-names>B.</given-names></name> <name><surname>Sheberla</surname> <given-names>D.</given-names></name> <etal/></person-group>. (<year>2018</year>). <article-title>Automatic chemical design using a data-driven continuous representation of molecules</article-title>. <source>ACS Central Sci.</source> <volume>4</volume>, <fpage>268</fpage>&#x02013;<lpage>276</lpage>. <pub-id pub-id-type="doi">10.1021/acscentsci.7b00572</pub-id><pub-id pub-id-type="pmid">29532027</pub-id></citation></ref>
<ref id="B61">
<citation citation-type="journal"><person-group person-group-type="author"><name><surname>Granda</surname> <given-names>J. M.</given-names></name> <name><surname>Donina</surname> <given-names>L.</given-names></name> <name><surname>Dragone</surname> <given-names>V.</given-names></name> <name><surname>Long</surname> <given-names>D.-L.</given-names></name> <name><surname>Cronin</surname> <given-names>L.</given-names></name></person-group> (<year>2018</year>). <article-title>Controlling an organic synthesis robot with machine learning to search for new reactivity</article-title>. <source>Nature</source> <volume>559</volume>, <fpage>377</fpage>&#x02013;<lpage>381</lpage>. <pub-id pub-id-type="doi">10.1038/s41586-018-0307-8</pub-id><pub-id pub-id-type="pmid">30022133</pub-id></citation></ref>
<ref id="B62">
<citation citation-type="journal"><person-group person-group-type="author"><name><surname>Green</surname> <given-names>C. P.</given-names></name> <name><surname>Engkvist</surname> <given-names>O.</given-names></name> <name><surname>Pairaudeau</surname> <given-names>G.</given-names></name></person-group> (<year>2018</year>). <article-title>The convergence of artificial intelligence and chemistry for improved drug discovery</article-title>. <source>Future Med. Chem.</source> <volume>10</volume>, <fpage>2573</fpage>&#x02013;<lpage>2576</lpage>. <pub-id pub-id-type="doi">10.4155/fmc-2018-0161</pub-id><pub-id pub-id-type="pmid">30499699</pub-id></citation></ref>
<ref id="B63">
<citation citation-type="journal"><person-group person-group-type="author"><name><surname>Griffen</surname> <given-names>E. J.</given-names></name> <name><surname>Dossetter</surname> <given-names>A. G.</given-names></name> <name><surname>Leach</surname> <given-names>A. G.</given-names></name> <name><surname>Montague</surname> <given-names>S.</given-names></name></person-group> (<year>2018</year>). <article-title>Can we accelerate medicinal chemistry by augmenting the chemist with Big Data and artificial intelligence?</article-title> <source>Drug Discov. Today</source> <volume>23</volume>, <fpage>1373</fpage>&#x02013;<lpage>1384</lpage>. <pub-id pub-id-type="doi">10.1016/j.drudis.2018.03.011</pub-id><pub-id pub-id-type="pmid">29577971</pub-id></citation></ref>
<ref id="B64">
<citation citation-type="journal"><person-group person-group-type="author"><name><surname>Grisoni</surname> <given-names>F.</given-names></name> <name><surname>Neuhaus</surname> <given-names>C. S.</given-names></name> <name><surname>Gabernet</surname> <given-names>G.</given-names></name> <name><surname>M&#x000FC;ller</surname> <given-names>A. T.</given-names></name> <name><surname>Hiss</surname> <given-names>J. A.</given-names></name> <name><surname>Schneider</surname> <given-names>G.</given-names></name></person-group> (<year>2018</year>). <article-title>Designing anticancer peptides by constructive machine learning</article-title>. <source>ChemMedChem</source> <volume>13</volume>, <fpage>1300</fpage>&#x02013;<lpage>1302</lpage>. <pub-id pub-id-type="doi">10.1002/cmdc.201800204</pub-id><pub-id pub-id-type="pmid">29679519</pub-id></citation></ref>
<ref id="B65">
<citation citation-type="journal"><person-group person-group-type="author"><name><surname>Gromski</surname> <given-names>P. S.</given-names></name> <name><surname>Henson</surname> <given-names>A. B.</given-names></name> <name><surname>Granda</surname> <given-names>J. M.</given-names></name> <name><surname>Cronin</surname> <given-names>L.</given-names></name></person-group> (<year>2019</year>). <article-title>How to explore chemical space using algorithms and automation</article-title>. <source>Nat. Rev. Chem.</source> <volume>3</volume>, <fpage>119</fpage>&#x02013;<lpage>128</lpage>. <pub-id pub-id-type="doi">10.1038/s41570-018-0066-y</pub-id></citation></ref>
<ref id="B66">
<citation citation-type="journal"><person-group person-group-type="author"><name><surname>Grzybowski</surname> <given-names>B. A.</given-names></name> <name><surname>Szymku&#x00107;</surname> <given-names>S.</given-names></name> <name><surname>Gajewska</surname> <given-names>E. P.</given-names></name> <name><surname>Molga</surname> <given-names>K.</given-names></name> <name><surname>Dittwald</surname> <given-names>P.</given-names></name> <name><surname>Wo&#x00142;os</surname> <given-names>A.</given-names></name> <etal/></person-group>. (<year>2018</year>). <article-title>Chematica: a story of computer code that started to think like a chemist</article-title>. <source>Chem</source> <volume>4</volume>, <fpage>390</fpage>&#x02013;<lpage>398</lpage>. <pub-id pub-id-type="doi">10.1016/j.chempr.2018.02.024</pub-id></citation></ref>
<ref id="B67">
<citation citation-type="journal"><person-group person-group-type="author"><name><surname>Gupta</surname> <given-names>A.</given-names></name> <name><surname>M&#x000FC;ller</surname> <given-names>A. T.</given-names></name> <name><surname>Huisman</surname> <given-names>B. J. H.</given-names></name> <name><surname>Fuchs</surname> <given-names>J. A.</given-names></name> <name><surname>Schneider</surname> <given-names>P.</given-names></name> <name><surname>Schneider</surname> <given-names>G.</given-names></name></person-group> (<year>2018</year>). <article-title>Generative recurrent networks for <italic>de novo</italic> drug design</article-title>. <source>Mol. Inform.</source> <volume>37</volume>:<fpage>1700111</fpage>. <pub-id pub-id-type="doi">10.1002/minf.201700111</pub-id></citation></ref>
<ref id="B68">
<citation citation-type="journal"><person-group person-group-type="author"><name><surname>Hansen</surname> <given-names>K.</given-names></name> <name><surname>Biegler</surname> <given-names>F.</given-names></name> <name><surname>Ramakrishnan</surname> <given-names>R.</given-names></name> <name><surname>Pronobis</surname> <given-names>W.</given-names></name> <name><surname>Von Lilienfeld</surname> <given-names>O. A.</given-names></name> <name><surname>M&#x000FC;ller</surname> <given-names>K.-R.</given-names></name> <etal/></person-group>. (<year>2015</year>). <article-title>Machine learning predictions of molecular properties: accurate many-body potentials and nonlocality in chemical space</article-title>. <source>J. Phys. Chem. Lett.</source> <volume>6</volume>, <fpage>2326</fpage>&#x02013;<lpage>2331</lpage>. <pub-id pub-id-type="doi">10.1021/acs.jpclett.5b00831</pub-id><pub-id pub-id-type="pmid">26113956</pub-id></citation></ref>
<ref id="B69">
<citation citation-type="journal"><person-group person-group-type="author"><name><surname>Hansen</surname> <given-names>K.</given-names></name> <name><surname>Montavon</surname> <given-names>G.</given-names></name> <name><surname>Biegler</surname> <given-names>F.</given-names></name> <name><surname>Fazli</surname> <given-names>S.</given-names></name> <name><surname>Rupp</surname> <given-names>M.</given-names></name> <name><surname>Scheffler</surname> <given-names>M.</given-names></name> <etal/></person-group>. (<year>2013</year>). <article-title>Assessment and validation of machine learning methods for predicting molecular atomization energies</article-title>. <source>J. Chem. Theory Comput.</source> <volume>9</volume>, <fpage>3404</fpage>&#x02013;<lpage>3419</lpage>. <pub-id pub-id-type="doi">10.1021/ct400195d</pub-id><pub-id pub-id-type="pmid">26584096</pub-id></citation></ref>
<ref id="B70">
<citation citation-type="journal"><person-group person-group-type="author"><name><surname>Harel</surname> <given-names>S.</given-names></name> <name><surname>Radinsky</surname> <given-names>K.</given-names></name></person-group> (<year>2018</year>). <article-title>Prototype-based compound discovery using deep generative models</article-title>. <source>Mol. Pharm.</source> <volume>15</volume>, <fpage>4406</fpage>&#x02013;<lpage>4416</lpage>. <pub-id pub-id-type="doi">10.1021/acs.molpharmaceut.8b00474</pub-id><pub-id pub-id-type="pmid">30063142</pub-id></citation></ref>
<ref id="B71">
<citation citation-type="journal"><person-group person-group-type="author"><name><surname>H&#x000E4;se</surname> <given-names>F.</given-names></name> <name><surname>Fdez. Galv&#x000E1;n</surname> <given-names>I.</given-names></name> <name><surname>Aspuru-Guzik</surname> <given-names>A.</given-names></name> <name><surname>Lindh</surname> <given-names>R.</given-names></name> <name><surname>Vacher</surname> <given-names>M.</given-names></name></person-group> (<year>2019</year>). <article-title>How machine learning can assist the interpretation of ab initio molecular dynamics simulations and conceptual understanding of chemistry</article-title>. <source>Chem. Sci.</source> <volume>10</volume>, <fpage>2298</fpage>&#x02013;<lpage>2307</lpage>. <pub-id pub-id-type="doi">10.1039/C8SC04516J</pub-id><pub-id pub-id-type="pmid">30881655</pub-id></citation></ref>
<ref id="B72">
<citation citation-type="journal"><person-group person-group-type="author"><name><surname>H&#x000E4;se</surname> <given-names>F.</given-names></name> <name><surname>Roch</surname> <given-names>L. M.</given-names></name> <name><surname>Aspuru-Guzik</surname> <given-names>A.</given-names></name></person-group> (<year>2018</year>). <article-title>Chimera: enabling hierarchy based multi-objective optimization for self-driving laboratories</article-title>. <source>Chem. Sci.</source> <volume>9</volume>, <fpage>7642</fpage>&#x02013;<lpage>7655</lpage>. <pub-id pub-id-type="doi">10.1039/C8SC02239A</pub-id><pub-id pub-id-type="pmid">30393525</pub-id></citation></ref>
<ref id="B73">
<citation citation-type="journal"><person-group person-group-type="author"><name><surname>He</surname> <given-names>Y.</given-names></name> <name><surname>Cubuk</surname> <given-names>E. D.</given-names></name> <name><surname>Allendorf</surname> <given-names>M. D.</given-names></name> <name><surname>Reed</surname> <given-names>E. J.</given-names></name></person-group> (<year>2018</year>). <article-title>Metallic metal&#x02013;organic frameworks predicted by the combination of machine learning methods and Ab initio calculations</article-title>. <source>J. Phys. Chem. Lett.</source> <volume>9</volume>, <fpage>4562</fpage>&#x02013;<lpage>4569</lpage>. <pub-id pub-id-type="doi">10.1021/acs.jpclett.8b01707</pub-id><pub-id pub-id-type="pmid">30052453</pub-id></citation></ref>
<ref id="B74">
<citation citation-type="journal"><person-group person-group-type="author"><name><surname>Hegde</surname> <given-names>G.</given-names></name> <name><surname>Bowen</surname> <given-names>R. C.</given-names></name></person-group> (<year>2017</year>). <article-title>Machine-learned approximations to density functional theory hamiltonians</article-title>. <source>Sci. Rep.</source> <volume>7</volume>:<fpage>42669</fpage>. <pub-id pub-id-type="doi">10.1038/srep42669</pub-id><pub-id pub-id-type="pmid">28198471</pub-id></citation></ref>
<ref id="B75">
<citation citation-type="journal"><person-group person-group-type="author"><name><surname>Hiller</surname> <given-names>S. A.</given-names></name> <name><surname>Golender</surname> <given-names>V. E.</given-names></name> <name><surname>Rosenblit</surname> <given-names>A. B.</given-names></name> <name><surname>Rastrigin</surname> <given-names>L. A.</given-names></name> <name><surname>Glaz</surname> <given-names>A. B.</given-names></name></person-group> (<year>1973</year>). <article-title>Cybernetic methods of drug design. I. Statement of the problem&#x02014;the perceptron approach</article-title>. <source>Comput. Biomed. Res.</source> <volume>6</volume>, <fpage>411</fpage>&#x02013;<lpage>421</lpage>. <pub-id pub-id-type="doi">10.1016/0010-4809(73)90074-8</pub-id><pub-id pub-id-type="pmid">4747104</pub-id></citation></ref>
<ref id="B76">
<citation citation-type="journal"><person-group person-group-type="author"><name><surname>Houben</surname> <given-names>C.</given-names></name> <name><surname>Lapkin</surname> <given-names>A. A.</given-names></name></person-group> (<year>2015</year>). <article-title>Automatic discovery and optimization of chemical processes</article-title>. <source>Curr. Opin. Chem. Eng.</source> <volume>9</volume>, <fpage>1</fpage>&#x02013;<lpage>7</lpage>. <pub-id pub-id-type="doi">10.1016/j.coche.2015.07.001</pub-id></citation></ref>
<ref id="B77">
<citation citation-type="journal"><person-group person-group-type="author"><name><surname>Houben</surname> <given-names>C.</given-names></name> <name><surname>Peremezhney</surname> <given-names>N.</given-names></name> <name><surname>Zubov</surname> <given-names>A.</given-names></name> <name><surname>Kosek</surname> <given-names>J.</given-names></name> <name><surname>Lapkin</surname> <given-names>A. A.</given-names></name></person-group> (<year>2015</year>). <article-title>Closed-loop multitarget optimization for discovery of new emulsion polymerization recipes</article-title>. <source>Organ. Process Res. Dev.</source> <volume>19</volume>, <fpage>1049</fpage>&#x02013;<lpage>1053</lpage>. <pub-id pub-id-type="doi">10.1021/acs.oprd.5b00210</pub-id><pub-id pub-id-type="pmid">26435638</pub-id></citation></ref>
<ref id="B78">
<citation citation-type="journal"><person-group person-group-type="author"><name><surname>Huang</surname> <given-names>S.-D.</given-names></name> <name><surname>Shang</surname> <given-names>C.</given-names></name> <name><surname>Kang</surname> <given-names>P.-L.</given-names></name> <name><surname>Liu</surname> <given-names>Z.-P.</given-names></name></person-group> (<year>2018</year>). <article-title>Atomic structure of boron resolved using machine learning and global sampling</article-title>. <source>Chem. Sci.</source> <volume>9</volume>, <fpage>8644</fpage>&#x02013;<lpage>8655</lpage>. <pub-id pub-id-type="doi">10.1039/C8SC03427C</pub-id><pub-id pub-id-type="pmid">30627388</pub-id></citation></ref>
<ref id="B79">
<citation citation-type="journal"><person-group person-group-type="author"><name><surname>Hughes</surname> <given-names>Z. E.</given-names></name> <name><surname>Thacker</surname> <given-names>J. C. R.</given-names></name> <name><surname>Wilson</surname> <given-names>A. L.</given-names></name> <name><surname>Popelier</surname> <given-names>P. L. A.</given-names></name></person-group> (<year>2019</year>). <article-title>Description of potential energy surfaces of molecules using FFLUX machine learning models</article-title>. <source>J. Chem. Theory Comput.</source> <volume>15</volume>, <fpage>116</fpage>&#x02013;<lpage>126</lpage>. <pub-id pub-id-type="doi">10.1021/acs.jctc.8b00806</pub-id><pub-id pub-id-type="pmid">30507180</pub-id></citation></ref>
<ref id="B80">
<citation citation-type="journal"><person-group person-group-type="author"><name><surname>Iype</surname> <given-names>E.</given-names></name> <name><surname>Urolagin</surname> <given-names>S.</given-names></name></person-group> (<year>2019</year>). <article-title>Machine learning model for non-equilibrium structures and energies of simple molecules</article-title>. <source>J. Chem. Phys.</source> <volume>150</volume>:<fpage>024307</fpage>. <pub-id pub-id-type="doi">10.1063/1.5054968</pub-id><pub-id pub-id-type="pmid">30646726</pub-id></citation></ref>
<ref id="B81">
<citation citation-type="journal"><person-group person-group-type="author"><name><surname>Janet</surname> <given-names>J. P.</given-names></name> <name><surname>Chan</surname> <given-names>L.</given-names></name> <name><surname>Kulik</surname> <given-names>H. J.</given-names></name></person-group> (<year>2018</year>). <article-title>Accelerating chemical discovery with machine learning: simulated evolution of spin crossover complexes with an artificial neural network</article-title>. <source>J. Phys. Chem. Lett.</source> <volume>9</volume>, <fpage>1064</fpage>&#x02013;<lpage>1071</lpage>. <pub-id pub-id-type="doi">10.1021/acs.jpclett.8b00170</pub-id><pub-id pub-id-type="pmid">29425453</pub-id></citation></ref>
<ref id="B82">
<citation citation-type="journal"><person-group person-group-type="author"><name><surname>Jaquis</surname> <given-names>B. J.</given-names></name> <name><surname>Li</surname> <given-names>A.</given-names></name> <name><surname>Monnier</surname> <given-names>N. D.</given-names></name> <name><surname>Sisk</surname> <given-names>R. G.</given-names></name> <name><surname>Acree</surname> <given-names>W. E.</given-names></name> <name><surname>Lang</surname> <given-names>A. S.</given-names></name></person-group> (<year>2019</year>). <article-title>Using machine learning to predict enthalpy of solvation</article-title>. <source>J. Solution Chem.</source> <volume>48</volume>, <fpage>564</fpage>&#x02013;<lpage>573</lpage>. <pub-id pub-id-type="doi">10.1007/s10953-019-00867-1</pub-id></citation></ref>
<ref id="B83">
<citation citation-type="journal"><person-group person-group-type="author"><name><surname>Jensen</surname> <given-names>J. H.</given-names></name></person-group> (<year>2019</year>). <article-title>A graph-based genetic algorithm and generative model/Monte Carlo tree search for the exploration of chemical space</article-title>. <source>Chem. Sci.</source> <volume>10</volume>, <fpage>3567</fpage>&#x02013;<lpage>3572</lpage>. <pub-id pub-id-type="doi">10.1039/C8SC05372C</pub-id><pub-id pub-id-type="pmid">30996948</pub-id></citation></ref>
<ref id="B84">
<citation citation-type="journal"><person-group person-group-type="author"><name><surname>Jha</surname> <given-names>D.</given-names></name> <name><surname>Ward</surname> <given-names>L.</given-names></name> <name><surname>Paul</surname> <given-names>A.</given-names></name> <name><surname>Liao</surname> <given-names>W.-K.</given-names></name> <name><surname>Choudhary</surname> <given-names>A.</given-names></name> <name><surname>Wolverton</surname> <given-names>C.</given-names></name> <etal/></person-group>. (<year>2018</year>). <article-title>ElemNet: deep learning the chemistry of materials from only elemental composition</article-title>. <source>Sci. Rep.</source> <volume>8</volume>:<fpage>17593</fpage>. <pub-id pub-id-type="doi">10.1038/s41598-018-35934-y</pub-id><pub-id pub-id-type="pmid">30514926</pub-id></citation></ref>
<ref id="B85">
<citation citation-type="journal"><person-group person-group-type="author"><name><surname>J&#x000F8;rgensen</surname> <given-names>P. B.</given-names></name> <name><surname>Mesta</surname> <given-names>M.</given-names></name> <name><surname>Shil</surname> <given-names>S.</given-names></name> <name><surname>Lastra</surname> <given-names>J. M. G.</given-names></name> <name><surname>Jacobsen</surname> <given-names>K. W.</given-names></name> <name><surname>Thygesen</surname> <given-names>K. S.</given-names></name> <etal/></person-group>. (<year>2018a</year>). <article-title>Machine learning-based screening of complex molecules for polymer solar cells</article-title>. <source>J. Chem. Phys.</source> <volume>148</volume>:<fpage>241735</fpage>. <pub-id pub-id-type="doi">10.1063/1.5023563</pub-id><pub-id pub-id-type="pmid">29960358</pub-id></citation></ref>
<ref id="B86">
<citation citation-type="journal"><person-group person-group-type="author"><name><surname>J&#x000F8;rgensen</surname> <given-names>P. B.</given-names></name> <name><surname>Schmidt</surname> <given-names>M. N.</given-names></name> <name><surname>Winther</surname> <given-names>O.</given-names></name></person-group> (<year>2018b</year>). <article-title>Deep generative models for molecular science</article-title>. <source>Mol. Inform.</source> <volume>37</volume>:<fpage>1700133</fpage>. <pub-id pub-id-type="doi">10.1002/minf.201700133</pub-id><pub-id pub-id-type="pmid">29405647</pub-id></citation></ref>
<ref id="B87">
<citation citation-type="journal"><person-group person-group-type="author"><name><surname>Kadurin</surname> <given-names>A.</given-names></name> <name><surname>Nikolenko</surname> <given-names>S.</given-names></name> <name><surname>Khrabrov</surname> <given-names>K.</given-names></name> <name><surname>Aliper</surname> <given-names>A.</given-names></name> <name><surname>Zhavoronkov</surname> <given-names>A.</given-names></name></person-group> (<year>2017</year>). <article-title>druGAN: an advanced generative adversarial autoencoder model for <italic>de novo</italic> generation of new molecules with desired molecular properties <italic>in silico</italic></article-title>. <source>Mol. Pharm.</source> <volume>14</volume>, <fpage>3098</fpage>&#x02013;<lpage>3104</lpage>. <pub-id pub-id-type="doi">10.1021/acs.molpharmaceut.7b00346</pub-id><pub-id pub-id-type="pmid">28703000</pub-id></citation></ref>
<ref id="B88">
<citation citation-type="journal"><person-group person-group-type="author"><name><surname>Kanamori</surname> <given-names>K.</given-names></name> <name><surname>Toyoura</surname> <given-names>K.</given-names></name> <name><surname>Honda</surname> <given-names>J.</given-names></name> <name><surname>Hattori</surname> <given-names>K.</given-names></name> <name><surname>Seko</surname> <given-names>A.</given-names></name> <name><surname>Karasuyama</surname> <given-names>M.</given-names></name> <etal/></person-group>. (<year>2018</year>). <article-title>Exploring a potential energy surface by machine learning for characterizing atomic transport</article-title>. <source>Phys. Rev. B</source> <volume>97</volume>:<fpage>125124</fpage>. <pub-id pub-id-type="doi">10.1103/PhysRevB.97.125124</pub-id></citation></ref>
<ref id="B89">
<citation citation-type="journal"><person-group person-group-type="author"><name><surname>Kang</surname> <given-names>S.</given-names></name> <name><surname>Cho</surname> <given-names>K.</given-names></name></person-group> (<year>2018</year>). <article-title>Conditional molecular design with deep generative models</article-title>. <source>J. Chem. Inf. Model.</source> <volume>59</volume>, <fpage>43</fpage>&#x02013;<lpage>52</lpage>. <pub-id pub-id-type="doi">10.1021/acs.jcim.8b00263</pub-id><pub-id pub-id-type="pmid">30016587</pub-id></citation></ref>
<ref id="B90">
<citation citation-type="journal"><person-group person-group-type="author"><name><surname>Kang</surname> <given-names>X.</given-names></name> <name><surname>Zhao</surname> <given-names>Y.</given-names></name> <name><surname>Li</surname> <given-names>J.</given-names></name></person-group> (<year>2018</year>). <article-title>Predicting refractive index of ionic liquids based on the extreme learning machine (ELM) intelligence algorithm</article-title>. <source>J. Mol. Liq.</source> <volume>250</volume>, <fpage>44</fpage>&#x02013;<lpage>49</lpage>. <pub-id pub-id-type="doi">10.1016/j.molliq.2017.11.166</pub-id></citation></ref>
<ref id="B91">
<citation citation-type="journal"><person-group person-group-type="author"><name><surname>Keil</surname> <given-names>M.</given-names></name> <name><surname>Exner</surname> <given-names>T. E.</given-names></name> <name><surname>Brickmann</surname> <given-names>J.</given-names></name></person-group> (<year>2004</year>). <article-title>Pattern recognition strategies for molecular surfaces: III. Binding site prediction with a neural network</article-title>. <source>J. Comput. Chem.</source> <volume>25</volume>, <fpage>779</fpage>&#x02013;<lpage>789</lpage>. <pub-id pub-id-type="doi">10.1002/jcc.10361</pub-id><pub-id pub-id-type="pmid">15011250</pub-id></citation></ref>
<ref id="B92">
<citation citation-type="book"><person-group person-group-type="author"><name><surname>Kishimoto</surname> <given-names>A.</given-names></name> <name><surname>Buesser</surname> <given-names>B.</given-names></name> <name><surname>Botea</surname> <given-names>A.</given-names></name></person-group> (<year>2018</year>). <article-title>AI meets chemistry</article-title>, in <source>Thirty-Second AAAI Conference on Artificial Intelligence</source>. <publisher-loc>Ireland</publisher-loc>: <publisher-name>IBM Research</publisher-name>.</citation></ref>
<ref id="B93">
<citation citation-type="journal"><person-group person-group-type="author"><name><surname>Klucznik</surname> <given-names>T.</given-names></name> <name><surname>Mikulak-Klucznik</surname> <given-names>B.</given-names></name> <name><surname>Mccormack</surname> <given-names>M. P.</given-names></name> <name><surname>Lima</surname> <given-names>H.</given-names></name> <name><surname>Szymku&#x00107;</surname> <given-names>S.</given-names></name> <name><surname>Bhowmick</surname> <given-names>M.</given-names></name> <etal/></person-group>. (<year>2018</year>). <article-title>Efficient syntheses of diverse, medicinally relevant targets planned by computer and executed in the laboratory</article-title>. <source>Chem</source> <volume>4</volume>, <fpage>522</fpage>&#x02013;<lpage>532</lpage>. <pub-id pub-id-type="doi">10.1016/j.chempr.2018.02.002</pub-id></citation></ref>
<ref id="B94">
<citation citation-type="journal"><person-group person-group-type="author"><name><surname>Kowalik</surname> <given-names>M.</given-names></name> <name><surname>Gothard</surname> <given-names>C. M.</given-names></name> <name><surname>Drews</surname> <given-names>A. M.</given-names></name> <name><surname>Gothard</surname> <given-names>N. A.</given-names></name> <name><surname>Weckiewicz</surname> <given-names>A.</given-names></name> <name><surname>Fuller</surname> <given-names>P. E.</given-names></name> <etal/></person-group>. (<year>2012</year>). <article-title>Parallel optimization of synthetic pathways within the network of organic chemistry</article-title>. <source>Angew. Chem. Int. Ed.</source> <volume>51</volume>, <fpage>7928</fpage>&#x02013;<lpage>7932</lpage>. <pub-id pub-id-type="doi">10.1002/anie.201202209</pub-id><pub-id pub-id-type="pmid">22807100</pub-id></citation></ref>
<ref id="B95">
<citation citation-type="journal"><person-group person-group-type="author"><name><surname>Krallinger</surname> <given-names>M.</given-names></name> <name><surname>Rabal</surname> <given-names>O.</given-names></name> <name><surname>Louren&#x000E7;o</surname> <given-names>A.</given-names></name> <name><surname>Oyarzabal</surname> <given-names>J.</given-names></name> <name><surname>Valencia</surname> <given-names>A.</given-names></name></person-group> (<year>2017</year>). <article-title>Information retrieval and text mining technologies for chemistry</article-title>. <source>Chem. Rev.</source> <volume>117</volume>, <fpage>7673</fpage>&#x02013;<lpage>7761</lpage>. <pub-id pub-id-type="doi">10.1021/acs.chemrev.6b00851</pub-id><pub-id pub-id-type="pmid">28475312</pub-id></citation></ref>
<ref id="B96">
<citation citation-type="journal"><person-group person-group-type="author"><name><surname>Lecun</surname> <given-names>Y.</given-names></name> <name><surname>Bengio</surname> <given-names>Y.</given-names></name> <name><surname>Hinton</surname> <given-names>G.</given-names></name></person-group> (<year>2015</year>). <article-title>Deep learning</article-title>. <source>Nature</source> <volume>521</volume>:<fpage>436</fpage>. <pub-id pub-id-type="doi">10.1038/nature14539</pub-id><pub-id pub-id-type="pmid">26017442</pub-id></citation></ref>
<ref id="B97">
<citation citation-type="journal"><person-group person-group-type="author"><name><surname>Lee</surname> <given-names>A. A.</given-names></name> <name><surname>Yang</surname> <given-names>Q.</given-names></name> <name><surname>Bassyouni</surname> <given-names>A.</given-names></name> <name><surname>Butler</surname> <given-names>C. R.</given-names></name> <name><surname>Hou</surname> <given-names>X.</given-names></name> <name><surname>Jenkinson</surname> <given-names>S.</given-names></name> <etal/></person-group>. (<year>2019</year>). <article-title>Ligand biological activity predicted by cleaning positive and negative chemical correlations</article-title>. <source>Proc. Natl. Acad. Sci. U.S.A.</source> <volume>116</volume>:<fpage>3373</fpage>. <pub-id pub-id-type="doi">10.1073/pnas.1810847116</pub-id><pub-id pub-id-type="pmid">30808733</pub-id></citation></ref>
<ref id="B98">
<citation citation-type="journal"><person-group person-group-type="author"><name><surname>Li</surname> <given-names>H.</given-names></name> <name><surname>Collins</surname> <given-names>C. R.</given-names></name> <name><surname>Ribelli</surname> <given-names>T. G.</given-names></name> <name><surname>Matyjaszewski</surname> <given-names>K.</given-names></name> <name><surname>Gordon</surname> <given-names>G. J.</given-names></name> <name><surname>Kowalewski</surname> <given-names>T.</given-names></name> <etal/></person-group>. (<year>2018a</year>). <article-title>Tuning the molecular weight distribution from atom transfer radical polymerization using deep reinforcement learning</article-title>. <source>Mol. Syst. Design Eng.</source> <volume>3</volume>, <fpage>496</fpage>&#x02013;<lpage>508</lpage>. <pub-id pub-id-type="doi">10.1039/C7ME00131B</pub-id></citation></ref>
<ref id="B99">
<citation citation-type="journal"><person-group person-group-type="author"><name><surname>Li</surname> <given-names>H.</given-names></name> <name><surname>Zhang</surname> <given-names>Z.</given-names></name> <name><surname>Liu</surname> <given-names>Z.</given-names></name></person-group> (<year>2017</year>). <article-title>Application of artificial neural networks for catalysis: a review</article-title>. <source>Catalysts</source> <volume>7</volume>:<fpage>306</fpage>. <pub-id pub-id-type="doi">10.3390/catal7100306</pub-id></citation></ref>
<ref id="B100">
<citation citation-type="journal"><person-group person-group-type="author"><name><surname>Li</surname> <given-names>J.</given-names></name> <name><surname>Eastgate</surname> <given-names>M. D.</given-names></name></person-group> (<year>2019</year>). <article-title>Making better decisions during synthetic route design: leveraging prediction to achieve greenness-by-design</article-title>. <source>React. Chem. Eng</source>. <volume>4</volume>, <fpage>1595</fpage>&#x02013;<lpage>1607</lpage>. <pub-id pub-id-type="doi">10.1039/C9RE00019D</pub-id></citation></ref>
<ref id="B101">
<citation citation-type="journal"><person-group person-group-type="author"><name><surname>Li</surname> <given-names>Y.</given-names></name> <name><surname>Zhang</surname> <given-names>L.</given-names></name> <name><surname>Liu</surname> <given-names>Z.</given-names></name></person-group> (<year>2018b</year>). <article-title>Multi-objective <italic>de novo</italic> drug design with conditional graph generative model</article-title>. <source>J. Cheminform.</source> <volume>10</volume>:<fpage>33</fpage>. <pub-id pub-id-type="doi">10.1186/s13321-018-0287-6</pub-id><pub-id pub-id-type="pmid">30043127</pub-id></citation></ref>
<ref id="B102">
<citation citation-type="book"><person-group person-group-type="author"><name><surname>Lipkowitz</surname> <given-names>K. B.</given-names></name> <name><surname>Boyd</surname> <given-names>D. B.</given-names></name></person-group> (<year>1995</year>). <source>Reviews in Computational Chemistry 6.</source> <publisher-loc>New York, NY</publisher-loc>: <publisher-name>Wiley Online Library</publisher-name>. <pub-id pub-id-type="doi">10.1002/9780470125830</pub-id></citation></ref>
<ref id="B103">
<citation citation-type="journal"><person-group person-group-type="author"><name><surname>Lo</surname> <given-names>Y.-C.</given-names></name> <name><surname>Rensi</surname> <given-names>S. E.</given-names></name> <name><surname>Torng</surname> <given-names>W.</given-names></name> <name><surname>Altman</surname> <given-names>R. B.</given-names></name></person-group> (<year>2018</year>). <article-title>Machine learning in chemoinformatics and drug discovery</article-title>. <source>Drug Discov. Today</source> <volume>23</volume>, <fpage>1538</fpage>&#x02013;<lpage>1546</lpage>. <pub-id pub-id-type="doi">10.1016/j.drudis.2018.05.010</pub-id><pub-id pub-id-type="pmid">29750902</pub-id></citation></ref>
<ref id="B104">
<citation citation-type="journal"><person-group person-group-type="author"><name><surname>Mansbach</surname> <given-names>R. A.</given-names></name> <name><surname>Ferguson</surname> <given-names>A. L.</given-names></name></person-group> (<year>2015</year>). <article-title>Machine learning of single molecule free energy surfaces and the impact of chemistry and environment upon structure and dynamics</article-title>. <source>J. Chem. Phys.</source> <volume>142</volume>:<fpage>105101</fpage>. <pub-id pub-id-type="doi">10.1063/1.4914144</pub-id><pub-id pub-id-type="pmid">25770561</pub-id></citation></ref>
<ref id="B105">
<citation citation-type="journal"><person-group person-group-type="author"><name><surname>Marques</surname> <given-names>M. R. G.</given-names></name> <name><surname>Wolff</surname> <given-names>J.</given-names></name> <name><surname>Steigemann</surname> <given-names>C.</given-names></name> <name><surname>Marques</surname> <given-names>M. A. L.</given-names></name></person-group> (<year>2019</year>). <article-title>Neural network force fields for simple metals and semiconductors: construction and application to the calculation of phonons and melting temperatures</article-title>. <source>Phys.istry Chem. Phys.</source> <volume>21</volume>, <fpage>6506</fpage>&#x02013;<lpage>6516</lpage>. <pub-id pub-id-type="doi">10.1039/C8CP05771K</pub-id><pub-id pub-id-type="pmid">30843548</pub-id></citation></ref>
<ref id="B106">
<citation citation-type="journal"><person-group person-group-type="author"><name><surname>Mater</surname> <given-names>A. C.</given-names></name> <name><surname>Coote</surname> <given-names>M. L.</given-names></name></person-group> (<year>2019</year>). <article-title>Deep learning in chemistry</article-title>. <source>J. Chem. Inf. Model.</source> <volume>59</volume>, <fpage>2545</fpage>&#x02013;<lpage>2559</lpage>. <pub-id pub-id-type="doi">10.1021/acs.jcim.9b00266</pub-id><pub-id pub-id-type="pmid">31194543</pub-id></citation></ref>
<ref id="B107">
<citation citation-type="journal"><person-group person-group-type="author"><name><surname>Matsuzaka</surname> <given-names>Y.</given-names></name> <name><surname>Uesawa</surname> <given-names>Y.</given-names></name></person-group> (<year>2019</year>). <article-title>Optimization of a deep-learning method based on the classification of images generated by parameterized deep snap a novel molecular-image-input technique for quantitative structure-activity relationship (QSAR) analysis</article-title>. <source>Front. Bioeng. Biotechnol.</source> <volume>7</volume>, <fpage>65</fpage>&#x02013;<lpage>65</lpage>. <pub-id pub-id-type="doi">10.3389/fbioe.2019.00065</pub-id><pub-id pub-id-type="pmid">30984753</pub-id></citation></ref>
<ref id="B108">
<citation citation-type="journal"><person-group person-group-type="author"><name><surname>Mayer</surname> <given-names>M.</given-names></name> <name><surname>Baeumner</surname> <given-names>A. J.</given-names></name></person-group> (<year>2019</year>). <article-title>A megatrend challenging analytical chemistry: biosensor and chemosensor concepts ready for the internet of things</article-title>. <source>Chem. Rev</source>. <volume>119</volume>, <fpage>7996</fpage>&#x02013;<lpage>8027</lpage>. <pub-id pub-id-type="doi">10.1021/acs.chemrev.8b00719</pub-id><pub-id pub-id-type="pmid">31070892</pub-id></citation></ref>
<ref id="B109">
<citation citation-type="journal"><person-group person-group-type="author"><name><surname>Merk</surname> <given-names>D.</given-names></name> <name><surname>Grisoni</surname> <given-names>F.</given-names></name> <name><surname>Friedrich</surname> <given-names>L.</given-names></name> <name><surname>Schneider</surname> <given-names>G.</given-names></name></person-group> (<year>2018</year>). <article-title>Tuning artificial intelligence on the <italic>de novo</italic> design of natural-product-inspired retinoid X receptor modulators</article-title>. <source>Commun. Chem.</source> <volume>1</volume>:<fpage>68</fpage>. <pub-id pub-id-type="doi">10.1038/s42004-018-0068-1</pub-id></citation></ref>
<ref id="B110">
<citation citation-type="journal"><person-group person-group-type="author"><name><surname>Mezei</surname> <given-names>P. D.</given-names></name> <name><surname>Von Lilienfeld</surname> <given-names>O. A.</given-names></name></person-group> (<year>2019</year>). <article-title>Non-covalent quantum machine learning corrections to density functionals</article-title>. <source>arXiv [preprint]. arXiv</source>:1903.09010.</citation></ref>
<ref id="B111">
<citation citation-type="book"><person-group person-group-type="author"><collab>Microsoft</collab></person-group> (<year>2018</year>). <source>Machine Learning, Data Mining and Rethinking Knowledge at KDD 2018</source>. <publisher-loc>London, UK</publisher-loc>: <publisher-name>Microsoft</publisher-name>.</citation></ref>
<ref id="B112">
<citation citation-type="journal"><person-group person-group-type="author"><name><surname>Miller</surname> <given-names>T. H.</given-names></name> <name><surname>Gallidabino</surname> <given-names>M. D.</given-names></name> <name><surname>Macrae</surname> <given-names>J. I.</given-names></name> <name><surname>Hogstrand</surname> <given-names>C.</given-names></name> <name><surname>Bury</surname> <given-names>N. R.</given-names></name> <name><surname>Barron</surname> <given-names>L. P.</given-names></name> <etal/></person-group>. (<year>2018</year>). <article-title>Machine learning for environmental toxicology: a call for integration and innovation</article-title>. <source>Environ. Sci. Technol.</source> <volume>52</volume>, <fpage>12953</fpage>&#x02013;<lpage>12955</lpage>. <pub-id pub-id-type="doi">10.1021/acs.est.8b05382</pub-id><pub-id pub-id-type="pmid">30338686</pub-id></citation></ref>
<ref id="B113">
<citation citation-type="journal"><person-group person-group-type="author"><name><surname>Min</surname> <given-names>K.</given-names></name> <name><surname>Choi</surname> <given-names>B.</given-names></name> <name><surname>Park</surname> <given-names>K.</given-names></name> <name><surname>Cho</surname> <given-names>E.</given-names></name></person-group> (<year>2018</year>). <article-title>Machine learning assisted optimization of electrochemical properties for Ni-rich cathode materials</article-title>. <source>Sci. Rep.</source> <volume>8</volume>:<fpage>15778</fpage>. <pub-id pub-id-type="doi">10.1038/s41598-018-34201-4</pub-id><pub-id pub-id-type="pmid">30361533</pub-id></citation></ref>
<ref id="B114">
<citation citation-type="journal"><person-group person-group-type="author"><name><surname>Mitchell</surname> <given-names>J. B. O.</given-names></name></person-group> (<year>2014</year>). <article-title>Machine learning methods in chemoinformatics</article-title>. <source>Wiley interdisciplinary reviews. Comput. Mol. Sci.</source> <volume>4</volume>, <fpage>468</fpage>&#x02013;<lpage>481</lpage>. <pub-id pub-id-type="doi">10.1002/wcms.1183</pub-id><pub-id pub-id-type="pmid">25285160</pub-id></citation></ref>
<ref id="B115">
<citation citation-type="book"><person-group person-group-type="author"><name><surname>Mitchell</surname> <given-names>T. M.</given-names></name></person-group> (<year>1997</year>). <source>Machine Learning</source>. <publisher-loc>Burr Ridge, IL</publisher-loc>: <publisher-name>McGraw Hill</publisher-name>.</citation></ref>
<ref id="B116">
<citation citation-type="journal"><person-group person-group-type="author"><name><surname>Molga</surname> <given-names>K.</given-names></name> <name><surname>Dittwald</surname> <given-names>P.</given-names></name> <name><surname>Grzybowski</surname> <given-names>B. A.</given-names></name></person-group> (<year>2019</year>). <article-title>Navigating around patented routes by preserving specific motifs along computer-planned retrosynthetic pathways</article-title>. <source>Chem</source> <volume>5</volume>, <fpage>460</fpage>&#x02013;<lpage>473</lpage>. <pub-id pub-id-type="doi">10.1016/j.chempr.2018.12.004</pub-id></citation></ref>
<ref id="B117">
<citation citation-type="journal"><person-group person-group-type="author"><name><surname>Montavon</surname> <given-names>G.</given-names></name> <name><surname>Rupp</surname> <given-names>M.</given-names></name> <name><surname>Gobre</surname> <given-names>V.</given-names></name> <name><surname>Vazquez-Mayagoitia</surname> <given-names>A.</given-names></name> <name><surname>Hansen</surname> <given-names>K.</given-names></name> <name><surname>Tkatchenko</surname> <given-names>A.</given-names></name> <etal/></person-group>. (<year>2013</year>). <article-title>Machine learning of molecular electronic properties in chemical compound space</article-title>. <source>New J. Phys.</source> <volume>15</volume>:<fpage>095003</fpage>. <pub-id pub-id-type="doi">10.1088/1367-2630/15/9/095003</pub-id></citation></ref>
<ref id="B118">
<citation citation-type="journal"><person-group person-group-type="author"><name><surname>Morawietz</surname> <given-names>T.</given-names></name> <name><surname>Singraber</surname> <given-names>A.</given-names></name> <name><surname>Dellago</surname> <given-names>C.</given-names></name> <name><surname>Behler</surname> <given-names>J.</given-names></name></person-group> (<year>2016</year>). <article-title>How van der Waals interactions determine the unique properties of water</article-title>. <source>Proc. Natl. Acad. Sci. U.S.A.</source> <volume>113</volume>, <fpage>8368</fpage>&#x02013;<lpage>8373</lpage>. <pub-id pub-id-type="doi">10.1073/pnas.1602375113</pub-id><pub-id pub-id-type="pmid">27402761</pub-id></citation></ref>
<ref id="B119">
<citation citation-type="journal"><person-group person-group-type="author"><name><surname>Morgan</surname> <given-names>H. L.</given-names></name></person-group> (<year>1965</year>). <article-title>The generation of a unique machine description for chemical structures-a technique developed at chemical abstracts service</article-title>. <source>J. Chem. Doc.</source> <volume>5</volume>, <fpage>107</fpage>&#x02013;<lpage>113</lpage>. <pub-id pub-id-type="doi">10.1021/c160017a018</pub-id></citation></ref>
<ref id="B120">
<citation citation-type="journal"><person-group person-group-type="author"><name><surname>M&#x000FC;ller</surname> <given-names>A. T.</given-names></name> <name><surname>Hiss</surname> <given-names>J. A.</given-names></name> <name><surname>Schneider</surname> <given-names>G.</given-names></name></person-group> (<year>2018</year>). <article-title>Recurrent neural network model for constructive peptide design</article-title>. <source>J. Chem. Inf. Model.</source> <volume>58</volume>, <fpage>472</fpage>&#x02013;<lpage>479</lpage>. <pub-id pub-id-type="doi">10.1021/acs.jcim.7b00414</pub-id><pub-id pub-id-type="pmid">29355319</pub-id></citation></ref>
<ref id="B121">
<citation citation-type="journal"><person-group person-group-type="author"><name><surname>Nouira</surname> <given-names>A.</given-names></name> <name><surname>Crivello</surname> <given-names>J.-C.</given-names></name> <name><surname>Sokolovska</surname> <given-names>N.</given-names></name></person-group> (<year>2018</year>). <article-title>CrystalGAN: learning to discover crystallographic structures with generative adversarial networks</article-title>. <source>arXiv [preprint]. arXiv</source>:1810.11203.</citation></ref>
<ref id="B122">
<citation citation-type="journal"><person-group person-group-type="author"><name><surname>Panteleev</surname> <given-names>J.</given-names></name> <name><surname>Gao</surname> <given-names>H.</given-names></name> <name><surname>Jia</surname> <given-names>L.</given-names></name></person-group> (<year>2018</year>). <article-title>Recent applications of machine learning in medicinal chemistry</article-title>. <source>Bioorgan. Med. Chem. Lett.</source> <volume>28</volume>, <fpage>2807</fpage>&#x02013;<lpage>2815</lpage>. <pub-id pub-id-type="doi">10.1016/j.bmcl.2018.06.046</pub-id><pub-id pub-id-type="pmid">30122222</pub-id></citation></ref>
<ref id="B123">
<citation citation-type="journal"><person-group person-group-type="author"><name><surname>Popova</surname> <given-names>M.</given-names></name> <name><surname>Isayev</surname> <given-names>O.</given-names></name> <name><surname>Tropsha</surname> <given-names>A.</given-names></name></person-group> (<year>2018</year>). <article-title>Deep reinforcement learning for <italic>de novo</italic> drug design</article-title>. <source>Sci. Adv.</source> <volume>4</volume>:<fpage>eaap7885</fpage>. <pub-id pub-id-type="doi">10.1126/sciadv.aap7885</pub-id><pub-id pub-id-type="pmid">30050984</pub-id></citation></ref>
<ref id="B124">
<citation citation-type="journal"><person-group person-group-type="author"><name><surname>Pronobis</surname> <given-names>W.</given-names></name> <name><surname>Sch&#x000FC;tt</surname> <given-names>K. T.</given-names></name> <name><surname>Tkatchenko</surname> <given-names>A.</given-names></name> <name><surname>M&#x000FC;ller</surname> <given-names>K.-R.</given-names></name></person-group> (<year>2018</year>). <article-title>Capturing intensive and extensive DFT/TDDFT molecular properties with machine learning</article-title>. <source>Eur. Phys. J. B</source> <volume>91</volume>:<fpage>178</fpage>. <pub-id pub-id-type="doi">10.1140/epjb/e2018-90148-y</pub-id></citation></ref>
<ref id="B125">
<citation citation-type="journal"><person-group person-group-type="author"><name><surname>Ramakrishnan</surname> <given-names>R.</given-names></name> <name><surname>Dral</surname> <given-names>P. O.</given-names></name> <name><surname>Rupp</surname> <given-names>M.</given-names></name> <name><surname>Von Lilienfeld</surname> <given-names>O. A.</given-names></name></person-group> (<year>2014</year>). <article-title>Quantum chemistry structures and properties of 134 kilo molecules</article-title>. <source>Sci. Data</source> <volume>1</volume>:<fpage>140022</fpage>. <pub-id pub-id-type="doi">10.1038/sdata.2014.22</pub-id><pub-id pub-id-type="pmid">25977779</pub-id></citation></ref>
<ref id="B126">
<citation citation-type="journal"><person-group person-group-type="author"><name><surname>Ramakrishnan</surname> <given-names>R.</given-names></name> <name><surname>Dral</surname> <given-names>P. O.</given-names></name> <name><surname>Rupp</surname> <given-names>M.</given-names></name> <name><surname>Von Lilienfeld</surname> <given-names>O. A.</given-names></name></person-group> (<year>2015</year>). <article-title>Big data meets quantum chemistry approximations: the &#x00394;-machine learning approach</article-title>. <source>J. Chem. Theory Comput.</source> <volume>11</volume>, <fpage>2087</fpage>&#x02013;<lpage>2096</lpage>. <pub-id pub-id-type="doi">10.1021/acs.jctc.5b00099</pub-id><pub-id pub-id-type="pmid">26574412</pub-id></citation></ref>
<ref id="B127">
<citation citation-type="journal"><person-group person-group-type="author"><name><surname>Ramakrishnan</surname> <given-names>R.</given-names></name> <name><surname>Von Lilienfeld</surname> <given-names>O. A.</given-names></name></person-group> (<year>2017</year>). <article-title>Machine learning, quantum chemistry, and chemical space</article-title>, in <source>Reviews in Computational Chemistry</source>, Vol. <volume>30</volume>, eds <person-group person-group-type="editor"><name><surname>Parrill</surname> <given-names>A. L.</given-names></name> <name><surname>Lipkowitz</surname> <given-names>K. B.</given-names></name></person-group> (<publisher-loc>Wiley</publisher-loc>), <fpage>225</fpage>&#x02013;<lpage>256</lpage>. <pub-id pub-id-type="doi">10.1002/9781119356059.ch5</pub-id></citation></ref>
<ref id="B128">
<citation citation-type="journal"><person-group person-group-type="author"><name><surname>Richmond</surname> <given-names>C. J.</given-names></name> <name><surname>Miras</surname> <given-names>H. N.</given-names></name> <name><surname>De La Oliva</surname> <given-names>A. R.</given-names></name> <name><surname>Zang</surname> <given-names>H.</given-names></name> <name><surname>Sans</surname> <given-names>V.</given-names></name> <name><surname>Paramonov</surname> <given-names>L.</given-names></name> <etal/></person-group>. (<year>2012</year>). <article-title>A flow-system array for the discovery and scale up of inorganic clusters</article-title>. <source>Nat. Chem.</source> <volume>4</volume>, <fpage>1037</fpage>&#x02013;<lpage>1043</lpage>. <pub-id pub-id-type="doi">10.1038/nchem.1489</pub-id><pub-id pub-id-type="pmid">23174985</pub-id></citation></ref>
<ref id="B129">
<citation citation-type="journal"><person-group person-group-type="author"><name><surname>Rogers</surname> <given-names>D.</given-names></name> <name><surname>Hahn</surname> <given-names>M.</given-names></name></person-group> (<year>2010</year>). <article-title>Extended-connectivity fingerprints</article-title>. <source>J. Chem. Inf. Model.</source> <volume>50</volume>, <fpage>742</fpage>&#x02013;<lpage>754</lpage>. <pub-id pub-id-type="doi">10.1021/ci100050t</pub-id><pub-id pub-id-type="pmid">20426451</pub-id></citation></ref>
<ref id="B130">
<citation citation-type="journal"><person-group person-group-type="author"><name><surname>Rupp</surname> <given-names>M.</given-names></name></person-group> (<year>2015</year>). <article-title>Machine learning for quantum mechanics in a nutshell</article-title>. <source>Int. J. Quantum Chem.</source> <volume>115</volume>, <fpage>1058</fpage>&#x02013;<lpage>1073</lpage>. <pub-id pub-id-type="doi">10.1002/qua.24954</pub-id></citation></ref>
<ref id="B131">
<citation citation-type="journal"><person-group person-group-type="author"><name><surname>Rupp</surname> <given-names>M.</given-names></name> <name><surname>Ramakrishnan</surname> <given-names>R.</given-names></name> <name><surname>Von Lilienfeld</surname> <given-names>O. A.</given-names></name></person-group> (<year>2015</year>). <article-title>Machine learning for quantum mechanical properties of atoms in molecules</article-title>. <source>J. Phys. Chem. Lett.</source> <volume>6</volume>, <fpage>3309</fpage>&#x02013;<lpage>3313</lpage>. <pub-id pub-id-type="doi">10.1021/acs.jpclett.5b01456</pub-id></citation></ref>
<ref id="B132">
<citation citation-type="journal"><person-group person-group-type="author"><name><surname>Rupp</surname> <given-names>M.</given-names></name> <name><surname>Tkatchenko</surname> <given-names>A.</given-names></name> <name><surname>M&#x000FC;ller</surname> <given-names>K.-R.</given-names></name> <name><surname>Von Lilienfeld</surname> <given-names>O. A.</given-names></name></person-group> (<year>2012</year>). <article-title>Fast and accurate modeling of molecular atomization energies with machine learning</article-title>. <source>Phys. Rev. Lett.</source> <volume>108</volume>:<fpage>058301</fpage>. <pub-id pub-id-type="doi">10.1103/PhysRevLett.108.058301</pub-id><pub-id pub-id-type="pmid">22400967</pub-id></citation></ref>
<ref id="B133">
<citation citation-type="journal"><person-group person-group-type="author"><name><surname>Sadowski</surname> <given-names>P.</given-names></name> <name><surname>Fooshee</surname> <given-names>D.</given-names></name> <name><surname>Subrahmanya</surname> <given-names>N.</given-names></name> <name><surname>Baldi</surname> <given-names>P.</given-names></name></person-group> (<year>2016</year>). <article-title>Synergies between quantum mechanics and machine learning in reaction prediction</article-title>. <source>J. Chem. Inf. Model.</source> <volume>56</volume>, <fpage>2125</fpage>&#x02013;<lpage>2128</lpage>. <pub-id pub-id-type="doi">10.1021/acs.jcim.6b00351</pub-id><pub-id pub-id-type="pmid">27749058</pub-id></citation></ref>
<ref id="B134">
<citation citation-type="journal"><person-group person-group-type="author"><name><surname>Samuel</surname> <given-names>A. L.</given-names></name></person-group> (<year>1959</year>). <article-title>Some studies in machine learning using the game of checkers</article-title>. <source>IBM J. Res. Dev.</source> <volume>3</volume>, <fpage>210</fpage>&#x02013;<lpage>229</lpage>. <pub-id pub-id-type="doi">10.1147/rd.33.0210</pub-id></citation></ref>
<ref id="B135">
<citation citation-type="journal"><person-group person-group-type="author"><name><surname>S&#x000E1;nchez-Lengeling</surname> <given-names>B.</given-names></name> <name><surname>Aspuru-Guzik</surname> <given-names>A.</given-names></name></person-group> (<year>2017</year>). <article-title>Learning more, with less</article-title>. <source>ACS Central Sci.</source> <volume>3</volume>, <fpage>275</fpage>&#x02013;<lpage>277</lpage>. <pub-id pub-id-type="doi">10.1021/acscentsci.7b00153</pub-id><pub-id pub-id-type="pmid">28470043</pub-id></citation></ref>
<ref id="B136">
<citation citation-type="journal"><person-group person-group-type="author"><name><surname>Sanchez-Lengeling</surname> <given-names>B.</given-names></name> <name><surname>Aspuru-Guzik</surname> <given-names>A.</given-names></name></person-group> (<year>2018</year>). <article-title>Inverse molecular design using machine learning: generative models for matter engineering</article-title>. <source>Science</source> <volume>361</volume>, <fpage>360</fpage>&#x02013;<lpage>365</lpage>. <pub-id pub-id-type="doi">10.1126/science.aat2663</pub-id><pub-id pub-id-type="pmid">30049875</pub-id></citation></ref>
<ref id="B137">
<citation citation-type="journal"><person-group person-group-type="author"><name><surname>Sanchez-Lengeling</surname> <given-names>B.</given-names></name> <name><surname>Roch</surname> <given-names>L. M.</given-names></name> <name><surname>Perea</surname> <given-names>J. D.</given-names></name> <name><surname>Langner</surname> <given-names>S.</given-names></name> <name><surname>Brabec</surname> <given-names>C. J.</given-names></name> <name><surname>Aspuru-Guzik</surname> <given-names>A.</given-names></name></person-group> (<year>2019</year>). <article-title>A Bayesian approach to predict solubility parameters</article-title>. <source>Adv. Theory Simul.</source> <volume>2</volume>:<fpage>1800069</fpage>. <pub-id pub-id-type="doi">10.1002/adts.201800069</pub-id></citation></ref>
<ref id="B138">
<citation citation-type="book"><person-group person-group-type="author"><name><surname>Savage</surname> <given-names>J.</given-names></name> <name><surname>Kishimoto</surname> <given-names>A.</given-names></name> <name><surname>Buesser</surname> <given-names>B.</given-names></name> <name><surname>Diaz-Aviles</surname> <given-names>E.</given-names></name> <name><surname>Alzate</surname> <given-names>C.</given-names></name></person-group> (<year>2017</year>). <article-title>Chemical reactant recommendation using a network of organic chemistry</article-title>, in <source>Proceedings of the Eleventh ACM Conference on Recommender Systems</source> (<publisher-loc>New York, NY</publisher-loc>: <publisher-name>ACM</publisher-name>), <fpage>210</fpage>&#x02013;<lpage>214</lpage>.</citation></ref>
<ref id="B139">
<citation citation-type="journal"><person-group person-group-type="author"><name><surname>Schleder</surname> <given-names>G. R.</given-names></name> <name><surname>Padilha</surname> <given-names>A. C. M.</given-names></name> <name><surname>Acosta</surname> <given-names>C. M.</given-names></name> <name><surname>Costa</surname> <given-names>M.</given-names></name> <name><surname>Fazzio</surname> <given-names>A.</given-names></name></person-group> (<year>2019</year>). <article-title>From DFT to machine learning: recent approaches to materials science&#x02013;a review</article-title>. <source>J. Phys. Mater.</source> <volume>2</volume>:<fpage>032001</fpage>. <pub-id pub-id-type="doi">10.1088/2515-7639/ab084b</pub-id></citation></ref>
<ref id="B140">
<citation citation-type="journal"><person-group person-group-type="author"><name><surname>Schneider</surname> <given-names>G.</given-names></name></person-group> (<year>2018</year>). <article-title>Generative models for artificially-intelligent molecular design</article-title>. <source>Mol. Inform.</source> <volume>37</volume>:<fpage>1880131</fpage>. <pub-id pub-id-type="doi">10.1002/minf.201880131</pub-id><pub-id pub-id-type="pmid">29442446</pub-id></citation></ref>
<ref id="B141">
<citation citation-type="journal"><person-group person-group-type="author"><name><surname>Sch&#x000FC;tt</surname> <given-names>K.</given-names></name> <name><surname>Glawe</surname> <given-names>H.</given-names></name> <name><surname>Brockherde</surname> <given-names>F.</given-names></name> <name><surname>Sanna</surname> <given-names>A.</given-names></name> <name><surname>M&#x000FC;ller</surname> <given-names>K.</given-names></name> <name><surname>Gross</surname> <given-names>E.</given-names></name></person-group> (<year>2014</year>). <article-title>How to represent crystal structures for machine learning: towards fast prediction of electronic properties</article-title>. <source>Phys. Rev. B</source> <volume>89</volume>:<fpage>205118</fpage>. <pub-id pub-id-type="doi">10.1103/PhysRevB.89.205118</pub-id></citation></ref>
<ref id="B142">
<citation citation-type="journal"><person-group person-group-type="author"><name><surname>Segler</surname> <given-names>M. H. S.</given-names></name> <name><surname>Preuss</surname> <given-names>M.</given-names></name> <name><surname>Waller</surname> <given-names>M. P.</given-names></name></person-group> (<year>2018</year>). <article-title>Planning chemical syntheses with deep neural networks and symbolic AI</article-title>. <source>Nature</source> <volume>555</volume>, <fpage>604</fpage>&#x02013;<lpage>610</lpage>. <pub-id pub-id-type="doi">10.1038/nature25978</pub-id><pub-id pub-id-type="pmid">29595767</pub-id></citation></ref>
<ref id="B143">
<citation citation-type="journal"><person-group person-group-type="author"><name><surname>Segler</surname> <given-names>M. H. S.</given-names></name> <name><surname>Waller</surname> <given-names>M. P.</given-names></name></person-group> (<year>2017</year>). <article-title>Neural-symbolic machine learning for retrosynthesis and reaction prediction</article-title>. <source>Chem. A Eur. J.</source> <volume>23</volume>, <fpage>5966</fpage>&#x02013;<lpage>5971</lpage>. <pub-id pub-id-type="doi">10.1002/chem.201605499</pub-id><pub-id pub-id-type="pmid">28134452</pub-id></citation></ref>
<ref id="B144">
<citation citation-type="journal"><person-group person-group-type="author"><name><surname>Shen</surname> <given-names>X.</given-names></name> <name><surname>Zhang</surname> <given-names>T.</given-names></name> <name><surname>Broderick</surname> <given-names>S.</given-names></name> <name><surname>Rajan</surname> <given-names>K.</given-names></name></person-group> (<year>2018</year>). <article-title>Correlative analysis of metal organic framework structures through manifold learning of Hirshfeld surfaces</article-title>. <source>Mol. Syst. Design Eng.</source> <volume>3</volume>, <fpage>826</fpage>&#x02013;<lpage>838</lpage>. <pub-id pub-id-type="doi">10.1039/C8ME00014J</pub-id></citation></ref>
<ref id="B145">
<citation citation-type="journal"><person-group person-group-type="author"><name><surname>Sim&#x000F5;es</surname> <given-names>R. S.</given-names></name> <name><surname>Maltarollo</surname> <given-names>V. G.</given-names></name> <name><surname>Oliveira</surname> <given-names>P. R.</given-names></name> <name><surname>Honorio</surname> <given-names>K. M.</given-names></name></person-group> (<year>2018</year>). <article-title>Transfer and multi-task learning in QSAR modeling: advances and challenges</article-title>. <source>Front. Pharmacol.</source> <volume>9</volume>:<fpage>74</fpage>. <pub-id pub-id-type="doi">10.3389/fphar.2018.00074</pub-id><pub-id pub-id-type="pmid">29467659</pub-id></citation></ref>
<ref id="B146">
<citation citation-type="journal"><person-group person-group-type="author"><name><surname>Smith</surname> <given-names>C. J.</given-names></name> <name><surname>Nikbin</surname> <given-names>N.</given-names></name> <name><surname>Ley</surname> <given-names>S. V.</given-names></name> <name><surname>Lange</surname> <given-names>H.</given-names></name> <name><surname>Baxendale</surname> <given-names>I. R.</given-names></name></person-group> (<year>2011</year>). <article-title>A fully automated, multistep flow synthesis of 5-amino-4-cyano-1,2,3-triazoles</article-title>. <source>Organ. Biomol. Chem.</source> <volume>9</volume>, <fpage>1938</fpage>&#x02013;<lpage>1947</lpage>. <pub-id pub-id-type="doi">10.1039/c0ob00815j</pub-id><pub-id pub-id-type="pmid">21283874</pub-id></citation></ref>
<ref id="B147">
<citation citation-type="journal"><person-group person-group-type="author"><name><surname>Smith</surname> <given-names>J. S.</given-names></name> <name><surname>Isayev</surname> <given-names>O.</given-names></name> <name><surname>Roitberg</surname> <given-names>A. E.</given-names></name></person-group> (<year>2017</year>). <article-title>ANI-1: an extensible neural network potential with DFT accuracy at force field computational cost</article-title>. <source>Chem. Sci.</source> <volume>8</volume>, <fpage>3192</fpage>&#x02013;<lpage>3203</lpage>. <pub-id pub-id-type="doi">10.1039/C6SC05720A</pub-id><pub-id pub-id-type="pmid">28507695</pub-id></citation></ref>
<ref id="B148">
<citation citation-type="journal"><person-group person-group-type="author"><name><surname>Smith</surname> <given-names>J. S.</given-names></name> <name><surname>Nebgen</surname> <given-names>B.</given-names></name> <name><surname>Lubbers</surname> <given-names>N.</given-names></name> <name><surname>Isayev</surname> <given-names>O.</given-names></name> <name><surname>Roitberg</surname> <given-names>A. E.</given-names></name></person-group> (<year>2018a</year>). <article-title>Less is more: Sampling chemical space with active learning</article-title>. <source>J. Chem. Phys.</source> <volume>148</volume>:<fpage>241733</fpage>. <pub-id pub-id-type="doi">10.1063/1.5023802</pub-id><pub-id pub-id-type="pmid">29960353</pub-id></citation></ref>
<ref id="B149">
<citation citation-type="journal"><person-group person-group-type="author"><name><surname>Smith</surname> <given-names>J. S.</given-names></name> <name><surname>Roitberg</surname> <given-names>A. E.</given-names></name> <name><surname>Isayev</surname> <given-names>O.</given-names></name></person-group> (<year>2018b</year>). <article-title>Transforming computational drug discovery with machine learning and AI</article-title>. <source>ACS Med. Chem. Lett.</source> <volume>9</volume>, <fpage>1065</fpage>&#x02013;<lpage>1069</lpage>. <pub-id pub-id-type="doi">10.1021/acsmedchemlett.8b00437</pub-id><pub-id pub-id-type="pmid">30429945</pub-id></citation></ref>
<ref id="B150">
<citation citation-type="journal"><person-group person-group-type="author"><name><surname>Snyder</surname> <given-names>J. C.</given-names></name> <name><surname>Rupp</surname> <given-names>M.</given-names></name> <name><surname>Hansen</surname> <given-names>K.</given-names></name> <name><surname>M&#x000FC;ller</surname> <given-names>K.-R.</given-names></name> <name><surname>Burke</surname> <given-names>K.</given-names></name></person-group> (<year>2012</year>). <article-title>Finding density functionals with machine learning</article-title>. <source>Phys. Rev. Lett.</source> <volume>108</volume>:<fpage>253002</fpage>. <pub-id pub-id-type="doi">10.1103/PhysRevLett.108.253002</pub-id><pub-id pub-id-type="pmid">23004593</pub-id></citation></ref>
<ref id="B151">
<citation citation-type="journal"><person-group person-group-type="author"><name><surname>Stein</surname> <given-names>H. S.</given-names></name> <name><surname>Guevarra</surname> <given-names>D.</given-names></name> <name><surname>Newhouse</surname> <given-names>P. F.</given-names></name> <name><surname>Soedarmadji</surname> <given-names>E.</given-names></name> <name><surname>Gregoire</surname> <given-names>J. M.</given-names></name></person-group> (<year>2019a</year>). <article-title>Machine learning of optical properties of materials &#x02013; predicting spectra from images and images from spectra</article-title>. <source>Chem. Sci.</source> <volume>10</volume>, <fpage>47</fpage>&#x02013;<lpage>55</lpage>. <pub-id pub-id-type="doi">10.1039/C8SC03077D</pub-id><pub-id pub-id-type="pmid">30746072</pub-id></citation></ref>
<ref id="B152">
<citation citation-type="journal"><person-group person-group-type="author"><name><surname>Stein</surname> <given-names>H. S.</given-names></name> <name><surname>Soedarmadji</surname> <given-names>E.</given-names></name> <name><surname>Newhouse</surname> <given-names>P. F.</given-names></name> <name><surname>Dan</surname> <given-names>G.</given-names></name> <name><surname>Gregoire</surname> <given-names>J. M.</given-names></name></person-group> (<year>2019b</year>). <article-title>Synthesis, optical imaging, and absorption spectroscopy data for 179072 metal oxides</article-title>. <source>Sci. Data</source> <volume>6</volume>:<fpage>9</fpage>. <pub-id pub-id-type="doi">10.1038/s41597-019-0019-4</pub-id><pub-id pub-id-type="pmid">30918263</pub-id></citation></ref>
<ref id="B153">
<citation citation-type="journal"><person-group person-group-type="author"><name><surname>Stevens</surname> <given-names>J. G.</given-names></name> <name><surname>Bourne</surname> <given-names>R. A.</given-names></name> <name><surname>Twigg</surname> <given-names>M. V.</given-names></name> <name><surname>Poliakoff</surname> <given-names>M.</given-names></name></person-group> (<year>2010</year>). <article-title>Real-time product switching using a twin catalyst system for the hydrogenation of furfural in supercritical CO<sub>2</sub></article-title>. <source>Angew. Chem. Int. Ed.</source> <volume>49</volume>, <fpage>8856</fpage>&#x02013;<lpage>8859</lpage>. <pub-id pub-id-type="doi">10.1002/anie.201005092</pub-id><pub-id pub-id-type="pmid">20928878</pub-id></citation></ref>
<ref id="B154">
<citation citation-type="journal"><person-group person-group-type="author"><name><surname>Szymku&#x00107;</surname> <given-names>S.</given-names></name> <name><surname>Gajewska</surname> <given-names>E. P.</given-names></name> <name><surname>Klucznik</surname> <given-names>T.</given-names></name> <name><surname>Molga</surname> <given-names>K.</given-names></name> <name><surname>Dittwald</surname> <given-names>P.</given-names></name> <name><surname>Startek</surname> <given-names>M.</given-names></name> <etal/></person-group>. (<year>2016</year>). <article-title>Computer-assisted synthetic planning: the end of the beginning</article-title>. <source>Angew. Chem. Int. Ed.</source> <volume>55</volume>:<fpage>5904</fpage>. <pub-id pub-id-type="doi">10.1002/anie.201506101</pub-id><pub-id pub-id-type="pmid">27062365</pub-id></citation></ref>
<ref id="B155">
<citation citation-type="journal"><person-group person-group-type="author"><name><surname>Thomsen</surname> <given-names>J. U.</given-names></name> <name><surname>Meyer</surname> <given-names>B.</given-names></name></person-group> (<year>1989</year>). <article-title>Pattern recognition of the 1H NMR spectra of sugar alditols using a neural network</article-title>. <source>J. Magnetic Reson.</source> <volume>84</volume>, <fpage>212</fpage>&#x02013;<lpage>217</lpage>. <pub-id pub-id-type="doi">10.1016/0022-2364(89)90021-8</pub-id></citation></ref>
<ref id="B156">
<citation citation-type="journal"><person-group person-group-type="author"><name><surname>Varnek</surname> <given-names>A.</given-names></name> <name><surname>Baskin</surname> <given-names>I.</given-names></name></person-group> (<year>2012</year>). <article-title>Machine learning methods for property prediction in chemoinformatics: quo vadis?</article-title> <source>J. Chem. Inf. Model.</source> <volume>52</volume>, <fpage>1413</fpage>&#x02013;<lpage>1437</lpage>. <pub-id pub-id-type="doi">10.1021/ci200409x</pub-id><pub-id pub-id-type="pmid">22582859</pub-id></citation></ref>
<ref id="B157">
<citation citation-type="journal"><person-group person-group-type="author"><name><surname>Venkatasubramanian</surname> <given-names>V.</given-names></name></person-group> (<year>2019</year>). <article-title>The promise of artificial intelligence in chemical engineering: is it here, finally?</article-title> <source>AIChE J.</source> <volume>65</volume>, <fpage>466</fpage>&#x02013;<lpage>478</lpage>. <pub-id pub-id-type="doi">10.1002/aic.16489</pub-id></citation></ref>
<ref id="B158">
<citation citation-type="journal"><person-group person-group-type="author"><name><surname>Wang</surname> <given-names>J.</given-names></name> <name><surname>Olsson</surname> <given-names>S.</given-names></name> <name><surname>Wehmeyer</surname> <given-names>C.</given-names></name> <name><surname>P&#x000E9;rez</surname> <given-names>A.</given-names></name> <name><surname>Charron</surname> <given-names>N. E.</given-names></name> <name><surname>De Fabritiis</surname> <given-names>G.</given-names></name> <etal/></person-group>. (<year>2019</year>). <article-title>Machine learning of coarse-grained molecular dynamics force fields</article-title>. <source>ACS Central Sci.</source> <volume>5</volume>, <fpage>755</fpage>&#x02013;<lpage>767</lpage>. <pub-id pub-id-type="doi">10.1021/acscentsci.8b00913</pub-id><pub-id pub-id-type="pmid">31139712</pub-id></citation></ref>
<ref id="B159">
<citation citation-type="journal"><person-group person-group-type="author"><name><surname>Ward</surname> <given-names>L.</given-names></name> <name><surname>Wolverton</surname> <given-names>C.</given-names></name></person-group> (<year>2017</year>). <article-title>Atomistic calculations and materials informatics: a review</article-title>. <source>Curr. Opin. Solid State Mater. Sci.</source> <volume>21</volume>, <fpage>167</fpage>&#x02013;<lpage>176</lpage>. <pub-id pub-id-type="doi">10.1016/j.cossms.2016.07.002</pub-id></citation></ref>
<ref id="B160">
<citation citation-type="journal"><person-group person-group-type="author"><name><surname>Wei</surname> <given-names>J. N.</given-names></name> <name><surname>Duvenaud</surname> <given-names>D.</given-names></name> <name><surname>Aspuru-Guzik</surname> <given-names>A.</given-names></name></person-group> (<year>2016</year>). <article-title>Neural networks for the prediction of organic chemistry reactions</article-title>. <source>ACS Central Sci.</source> <volume>2</volume>, <fpage>725</fpage>&#x02013;<lpage>732</lpage>. <pub-id pub-id-type="doi">10.1021/acscentsci.6b00219</pub-id><pub-id pub-id-type="pmid">27800555</pub-id></citation></ref>
<ref id="B161">
<citation citation-type="journal"><person-group person-group-type="author"><name><surname>Welborn</surname> <given-names>M.</given-names></name> <name><surname>Cheng</surname> <given-names>L.</given-names></name> <name><surname>Miller</surname> <given-names>T. F.</given-names></name></person-group> (<year>2018</year>). <article-title>Transferability in machine learning for electronic structure via the molecular orbital basis</article-title>. <source>J. Chem. Theory Comput.</source> <volume>14</volume>, <fpage>4772</fpage>&#x02013;<lpage>4779</lpage>. <pub-id pub-id-type="doi">10.1021/acs.jctc.8b00636</pub-id><pub-id pub-id-type="pmid">30040892</pub-id></citation></ref>
<ref id="B162">
<citation citation-type="journal"><person-group person-group-type="author"><name><surname>White</surname> <given-names>D.</given-names></name> <name><surname>Wilson</surname> <given-names>R. C.</given-names></name></person-group> (<year>2010</year>). <article-title>Generative models for chemical structures</article-title>. <source>J. Chem. Inf. Model.</source> <volume>50</volume>, <fpage>1257</fpage>&#x02013;<lpage>1274</lpage>. <pub-id pub-id-type="doi">10.1021/ci9004089</pub-id><pub-id pub-id-type="pmid">20666408</pub-id></citation></ref>
<ref id="B163">
<citation citation-type="journal"><person-group person-group-type="author"><name><surname>Wu</surname> <given-names>Y.</given-names></name> <name><surname>Wang</surname> <given-names>G.</given-names></name></person-group> (<year>2018</year>). <article-title>Machine learning based toxicity prediction: from chemical structural description to transcriptome analysis</article-title>. <source>Int. J. Mol. Sci.</source> <volume>19</volume>:<fpage>2358</fpage>. <pub-id pub-id-type="doi">10.3390/ijms19082358</pub-id><pub-id pub-id-type="pmid">30103448</pub-id></citation></ref>
<ref id="B164">
<citation citation-type="web"><person-group person-group-type="author"><name><surname>Wu</surname> <given-names>Z.</given-names></name> <name><surname>Ramsundar</surname> <given-names>B.</given-names></name> <name><surname>Feinberg</surname> <given-names>E. N.</given-names></name> <name><surname>Gomes</surname> <given-names>J.</given-names></name> <name><surname>Geniesse</surname> <given-names>C.</given-names></name> <name><surname>Pappu</surname> <given-names>A. S.</given-names></name> <etal/></person-group>. (<year>2017</year>). <source>MoleculeNet: a benchmark for molecular machine learning. arXiv e-prints</source>. Available online at: <ext-link ext-link-type="uri" xlink:href="https://ui.adsabs.harvard.edu/abs/2017arXiv170300564W">https://ui.adsabs.harvard.edu/abs/2017arXiv170300564W</ext-link> (accessed March 01, 2017).</citation></ref>
<ref id="B165">
<citation citation-type="journal"><person-group person-group-type="author"><name><surname>Xia</surname> <given-names>R.</given-names></name> <name><surname>Kais</surname> <given-names>S.</given-names></name></person-group> (<year>2018</year>). <article-title>Quantum machine learning for electronic structure calculations</article-title>. <source>Nat. Commun.</source> <volume>9</volume>:<fpage>4195</fpage>. <pub-id pub-id-type="doi">10.1038/s41467-018-06598-z</pub-id><pub-id pub-id-type="pmid">30305624</pub-id></citation></ref>
<ref id="B166">
<citation citation-type="journal"><person-group person-group-type="author"><name><surname>Xu</surname> <given-names>Y.</given-names></name> <name><surname>Lin</surname> <given-names>K.</given-names></name> <name><surname>Wang</surname> <given-names>S.</given-names></name> <name><surname>Wang</surname> <given-names>L.</given-names></name> <name><surname>Cai</surname> <given-names>C.</given-names></name> <name><surname>Song</surname> <given-names>C.</given-names></name> <etal/></person-group>. (<year>2019</year>). <article-title>Deep learning for molecular generation</article-title>. <source>Future Med. Chem.</source> <volume>11</volume>, <fpage>567</fpage>&#x02013;<lpage>597</lpage>. <pub-id pub-id-type="doi">10.4155/fmc-2018-0358</pub-id><pub-id pub-id-type="pmid">30698019</pub-id></citation></ref>
<ref id="B167">
<citation citation-type="journal"><person-group person-group-type="author"><name><surname>Zaspel</surname> <given-names>P.</given-names></name> <name><surname>Huang</surname> <given-names>B.</given-names></name> <name><surname>Harbrecht</surname> <given-names>H.</given-names></name> <name><surname>Von Lilienfeld</surname> <given-names>O. A.</given-names></name></person-group> (<year>2019</year>). <article-title>Boosting quantum machine learning models with a multilevel combination technique: pople diagrams revisited</article-title>. <source>J. Chem. Theory Comput.</source> <volume>15</volume>, <fpage>1546</fpage>&#x02013;<lpage>1559</lpage>. <pub-id pub-id-type="doi">10.1021/acs.jctc.8b00832</pub-id><pub-id pub-id-type="pmid">30516999</pub-id></citation></ref>
<ref id="B168">
<citation citation-type="journal"><person-group person-group-type="author"><name><surname>Zhang</surname> <given-names>P.</given-names></name> <name><surname>Shen</surname> <given-names>L.</given-names></name> <name><surname>Yang</surname> <given-names>W.</given-names></name></person-group> (<year>2019</year>). <article-title>Solvation free energy calculations with quantum mechanics/molecular mechanics and machine learning models</article-title>. <source>J. Phys. Chem. B</source> <volume>123</volume>, <fpage>901</fpage>&#x02013;<lpage>908</lpage>. <pub-id pub-id-type="doi">10.1021/acs.jpcb.8b11905</pub-id><pub-id pub-id-type="pmid">30557020</pub-id></citation></ref>
<ref id="B169">
<citation citation-type="journal"><person-group person-group-type="author"><name><surname>Zhou</surname> <given-names>Z.</given-names></name> <name><surname>Kearnes</surname> <given-names>S.</given-names></name> <name><surname>Li</surname> <given-names>L.</given-names></name> <name><surname>Zare</surname> <given-names>R. N.</given-names></name> <name><surname>Riley</surname> <given-names>P.</given-names></name></person-group> (<year>2018</year>). <article-title>Optimization of molecules via deep reinforcement learning</article-title>. <source>arXiv preprint arXiv</source>:1810.08678. <pub-id pub-id-type="doi">10.1038/s41598-019-47148-x</pub-id><pub-id pub-id-type="pmid">31341196</pub-id></citation></ref>
<ref id="B170">
<citation citation-type="journal"><person-group person-group-type="author"><name><surname>Zhou</surname> <given-names>Z.</given-names></name> <name><surname>Li</surname> <given-names>X.</given-names></name> <name><surname>Zare</surname> <given-names>R. N.</given-names></name></person-group> (<year>2017</year>). <article-title>Optimizing chemical reactions with deep reinforcement learning</article-title>. <source>ACS Central Sci.</source> <volume>3</volume>, <fpage>1337</fpage>&#x02013;<lpage>1344</lpage>. <pub-id pub-id-type="doi">10.1021/acscentsci.7b00492</pub-id><pub-id pub-id-type="pmid">29296675</pub-id></citation></ref>
<ref id="B171">
<citation citation-type="journal"><person-group person-group-type="author"><name><surname>Zielinski</surname> <given-names>F.</given-names></name> <name><surname>Maxwell</surname> <given-names>P. I.</given-names></name> <name><surname>Fletcher</surname> <given-names>T. L.</given-names></name> <name><surname>Davie</surname> <given-names>S. J.</given-names></name> <name><surname>Di Pasquale</surname> <given-names>N.</given-names></name> <name><surname>Cardamone</surname> <given-names>S.</given-names></name> <etal/></person-group>. (<year>2017</year>). <article-title>Geometry optimization with machine trained topological atoms</article-title>. <source>Sci. Rep.</source> <volume>7</volume>:<fpage>12817</fpage>. <pub-id pub-id-type="doi">10.1038/s41598-017-12600-3</pub-id><pub-id pub-id-type="pmid">28993674</pub-id></citation></ref>
</ref-list>
<fn-group>
<fn fn-type="financial-disclosure"><p><bold>Funding.</bold> The Coimbra Chemistry Centre (CQC) was supported by the Portuguese Agency for Scientific Research, Funda&#x000E7;&#x000E3;o para a Ci&#x000EA;ncia e a Tecnologia (FCT), through Project UID/QUI/00313/2019.</p>
</fn>
</fn-group>
</back>
</article>