<?xml version="1.0" encoding="UTF-8"?>
<!DOCTYPE article PUBLIC "-//NLM//DTD Journal Publishing DTD v2.3 20070202//EN" "journalpublishing.dtd">
<article article-type="research-article" dtd-version="2.3" xml:lang="EN" xmlns:mml="http://www.w3.org/1998/Math/MathML" xmlns:xlink="http://www.w3.org/1999/xlink">
<front>
<journal-meta>
<journal-id journal-id-type="publisher-id">Front. Pharmacol.</journal-id>
<journal-title>Frontiers in Pharmacology</journal-title>
<abbrev-journal-title abbrev-type="pubmed">Front. Pharmacol.</abbrev-journal-title>
<issn pub-type="epub">1663-9812</issn>
<publisher>
<publisher-name>Frontiers Media S.A.</publisher-name>
</publisher>
</journal-meta>
<article-meta>
<article-id pub-id-type="publisher-id">606668</article-id>
<article-id pub-id-type="doi">10.3389/fphar.2020.606668</article-id>
<article-categories>
<subj-group subj-group-type="heading">
<subject>Pharmacology</subject>
<subj-group>
<subject>Original Research</subject>
</subj-group>
</subj-group>
</article-categories>
<title-group>
<article-title>Improvement of Prediction Performance With Conjoint Molecular Fingerprint in Deep Learning</article-title>
<alt-title alt-title-type="left-running-head">Xie et&#x0020;al.</alt-title>
<alt-title alt-title-type="right-running-head">Conjoint Fingerprint Boosts Prediction Performance</alt-title>
</title-group>
<contrib-group>
<contrib contrib-type="author">
<name>
<surname>Xie</surname>
<given-names>Liangxu</given-names>
</name>
<xref ref-type="aff" rid="aff1">
<sup>1</sup>
</xref>
<xref ref-type="aff" rid="aff2">
<sup>2</sup>
</xref>
<uri xlink:href="http://loop.frontiersin.org/people/1088088/overview"/>
</contrib>
<contrib contrib-type="author" corresp="yes">
<name>
<surname>Xu</surname>
<given-names>Lei</given-names>
</name>
<xref ref-type="aff" rid="aff1">
<sup>1</sup>
</xref>
<uri xlink:href="http://loop.frontiersin.org/people/794274/overview"/>
</contrib>
<contrib contrib-type="author">
<name>
<surname>Kong</surname>
<given-names>Ren</given-names>
</name>
<xref ref-type="aff" rid="aff1">
<sup>1</sup>
</xref>
<uri xlink:href="http://loop.frontiersin.org/people/830481/overview"/>
</contrib>
<contrib contrib-type="author" corresp="yes">
<name>
<surname>Chang</surname>
<given-names>Shan</given-names>
</name>
<xref ref-type="aff" rid="aff1">
<sup>1</sup>
</xref>
<xref ref-type="corresp" rid="c001">&#x002a;</xref>
<uri xlink:href="http://loop.frontiersin.org/people/854058/overview"/>
</contrib>
<contrib contrib-type="author" corresp="yes">
<name>
<surname>Xu</surname>
<given-names>Xiaojun</given-names>
</name>
<xref ref-type="aff" rid="aff1">
<sup>1</sup>
</xref>
<xref ref-type="corresp" rid="c001">&#x002a;</xref>
<uri xlink:href="https://loop.frontiersin.org/people/1166035/overview"/>
</contrib>
</contrib-group>
<aff id="aff1">
<label>
<sup>1</sup>
</label>Institute of Bioinformatics and Medical Engineering, School of Electrical and Information Engineering, Jiangsu University of Technology, <addr-line>Changzhou</addr-line>, <country>China</country>
</aff>
<aff id="aff2">
<label>
<sup>2</sup>
</label>Jiangsu Sino-Israel Industrial Technology Research Institute, <addr-line>Changzhou</addr-line>, <country>China</country>
</aff>
<author-notes>
<fn fn-type="edited-by">
<p>
<bold>Edited by:</bold> <ext-link ext-link-type="uri" xlink:href="https://loop.frontiersin.org/people/18269/overview">Salvatore Salomone</ext-link>, University of Catania, Italy</p>
</fn>
<fn fn-type="edited-by">
<p>
<bold>Reviewed by:</bold> <ext-link ext-link-type="uri" xlink:href="https://loop.frontiersin.org/people/303971/overview">Vishnu Sresht</ext-link>, Massachusetts Institute of Technology, United&#x0020;States</p>
<p>
<ext-link ext-link-type="uri" xlink:href="https://loop.frontiersin.org/people/483996/overview">Emilio Xavier Esposito</ext-link>, exeResearch LLC, United&#x0020;States</p>
</fn>
<corresp id="c001">&#x2a;Correspondence: Shan Chang, <email>schang@jsut.edu.cn</email>; Xiaojun Xu, <email>xuxiaojun@jsut.edu.cn</email>
</corresp>
<fn fn-type="other">
<p>This article was submitted to Experimental Pharmacology and Drug&#x0020;Discovery, a section of the journal Frontiers in Pharmacology</p>
</fn>
</author-notes>
<pub-date pub-type="epub">
<day>18</day>
<month>12</month>
<year>2020</year>
</pub-date>
<pub-date pub-type="collection">
<year>2020</year>
</pub-date>
<volume>11</volume>
<elocation-id>606668</elocation-id>
<history>
<date date-type="received">
<day>21</day>
<month>09</month>
<year>2020</year>
</date>
<date date-type="accepted">
<day>23</day>
<month>11</month>
<year>2020</year>
</date>
</history>
<permissions>
<copyright-statement>Copyright &#x00A9; 2020 Xie, Xu, Kong, Chang and Xu.</copyright-statement>
<copyright-year>2020</copyright-year>
<copyright-holder>Xie, Xu, Kong, Chang and Xu</copyright-holder>
<license xlink:href="http://creativecommons.org/licenses/by/4.0/">
<p>This is an open-access article distributed under the terms of the Creative Commons Attribution License (CC BY). The use, distribution or reproduction in other forums is permitted, provided the original author(s) and the copyright owner(s) are credited and that the original publication in this journal is cited, in accordance with accepted academic practice. No use, distribution or reproduction is permitted which does not comply with these&#x0020;terms.</p>
</license>
</permissions>
<abstract>
<p>The accurate predicting of physical properties and bioactivity of drug molecules in deep learning depends on how molecules are represented. Many types of molecular descriptors have been developed for quantitative structure-activity/property relationships quantitative structure-activity relationships (QSPR). However, each molecular descriptor is optimized for a specific application with encoding preference. Considering that standalone featurization methods may only cover parts of information of the chemical molecules, we proposed to build the conjoint fingerprint by combining two supplementary fingerprints. The impact of conjoint fingerprint and each standalone fingerprint on predicting performance was systematically evaluated in predicting the logarithm of the partition coefficient (logP) and binding affinity of protein-ligand by using machine learning/deep learning (ML/DL) methods, including random forest (RF), support vector regression (SVR), extreme gradient boosting (XGBoost), long short-term memory network (LSTM), and deep neural network (DNN). The results demonstrated that the conjoint fingerprint yielded improved predictive performance, even outperforming the consensus model using two standalone fingerprints among four out of five examined methods. Given that the conjoint fingerprint scheme shows easy extensibility and high applicability, we expect that the proposed conjoint scheme would create new opportunities for continuously improving predictive performance of deep learning by harnessing the complementarity of various types of fingerprints.</p>
</abstract>
<kwd-group>
<kwd>artificial intelligence</kwd>
<kwd>deep learning</kwd>
<kwd>fingerprints</kwd>
<kwd>quantitative structure-activity relationship</kwd>
<kwd>molecular descriptors</kwd>
</kwd-group>
<counts>
<page-count count="0"/>
</counts>
</article-meta>
</front>
<body>
<sec id="s1">
<title>Introduction</title>
<p>Predicting molecular properties plays important roles in guiding drug discovery. In the last decade, applying machine learning to predict physical or chemical properties of molecular drugs gains great interest, especially since the emergence of deep learning (<xref ref-type="bibr" rid="B26">LeCun et&#x0020;al., 2015</xref>; <xref ref-type="bibr" rid="B34">Min et&#x0020;al., 2016</xref>; <xref ref-type="bibr" rid="B52">Shen et&#x0020;al., 2020</xref>). By converting molecules into computer readable formats, such as molecular descriptors, machine learning will map features through hierarchical non-linear functions to required outputs. Deep learning with matched input molecular descriptors has achieved breakthrough improvements in biology and chemistry fields, such as predicting quantitative structure-activity relationships (QSAR) (<xref ref-type="bibr" rid="B4">Butler et&#x0020;al., 2018</xref>), modeling absorption, distribution, metabolism, excretion and toxicity (ADMET) (<xref ref-type="bibr" rid="B27">Lei et&#x0020;al., 2017</xref>; <xref ref-type="bibr" rid="B70">Wu et&#x0020;al., 2019</xref>), virtual screening (<xref ref-type="bibr" rid="B6">Cereto-Massagu&#x00e9; et&#x0020;al., 2015</xref>; <xref ref-type="bibr" rid="B63">Wang et&#x0020;al., 2020</xref>), drug design (<xref ref-type="bibr" rid="B35">Morrone et&#x0020;al., 2020</xref>), materials design (<xref ref-type="bibr" rid="B73">Xie and Grossman, 2018</xref>), chemical reactions (<xref ref-type="bibr" rid="B18">Grambow et&#x0020;al., 2020</xref>), and protein structure prediction (<xref ref-type="bibr" rid="B51">Senior et&#x0020;al., 2020</xref>),&#x0020;etc.</p>
<p>The accumulated experience in bioinformatics studies shows that the accurate predictions of machine learning heavily depend on the effective molecular representations (<xref ref-type="bibr" rid="B49">Schneider, 2010</xref>). Researchers from chemistry and biology field adopt many ways to design proper molecular descriptors, which requires strong experience and professional knowledge (<xref ref-type="bibr" rid="B33">Mater and Coote, 2019</xref>). Many types of molecular descriptors have been designed based on professional knowledge and specific demands. In the early days, the primary aim is to store and retrieve molecules, so that the molecular representations are compact and simple. The famous example is simplified input line entry system (SMILES) (<xref ref-type="bibr" rid="B65">Weininger, 1988</xref>; <xref ref-type="bibr" rid="B66">Weininger et&#x0020;al., 1989</xref>; <xref ref-type="bibr" rid="B65">Weininger, 1990</xref>). Later, the aim to search substructures drives to develop key-based fingerprints, such as molecular access system (MACCS) keys (<xref ref-type="bibr" rid="B12">Durant et&#x0020;al., 2002a</xref>). To meet growing need to model structure-activity and bioactivity, more effective fingerprints are designed, such as pharmacophore fingerprint and topological fingerprints. Recently, researchers are trying to incorporating 3D information in fingerprints for accurately predicting bioactivity of molecular drugs. More expert-designed fingerprints are continuously to be developed, such as 4D-fingerprints (<xref ref-type="bibr" rid="B50">Senese et&#x0020;al., 2004</xref>), molecular graphs (<xref ref-type="bibr" rid="B24">Kearnes et&#x0020;al., 2016</xref>), coulomb matrices and atomic coordinates (<xref ref-type="bibr" rid="B47">Sanchez-Lengeling and Aspuru-Guzik, 2018</xref>) or properties extracted from molecular dynamics simulations (<xref ref-type="bibr" rid="B44">Riniker, 2017</xref>).</p>
<p>Though many types of molecular descriptors have been proposed, there is not &#x201c;one size fits all&#x201d; molecular representation. The domain expert-engineered molecular features sometimes becomes one main obstacle sitting on the road to deep learning (<xref ref-type="bibr" rid="B8">Chuang et&#x0020;al., 2020</xref>). Molecular descriptors represent molecular structures from holistic representations, such as molecular size, weight, molecular shape. In the contrast, molecular fingerprints describe the local aspect of chemical structures and exploding whether the presence of substructure patterns. Molecular fingerprints have been optimized based on the particular tasks. Existing molecular fingerprints encodes different information with preference to reproduce the best results for the designated tasks. The available fingerprints can be classified into five types: topological, geometrical, thermodynamic, electronic and constitutional fingerprints (<xref ref-type="bibr" rid="B9">Danishuddin and Khan, 2016</xref>). Several studies have been reported to check the performance of different fingerprint schemes (<xref ref-type="bibr" rid="B11">Duan et&#x0020;al., 2010</xref>; <xref ref-type="bibr" rid="B43">Riniker and Landrum, 2013</xref>; <xref ref-type="bibr" rid="B71">Wu et&#x0020;al., 2018</xref>). Each type of molecular descriptors combined with machine learning methods fits into the matching scope of applicability. Two most used molecular descriptors are MACCS keys and extended connectivity fingerprints (ECFP) (<xref ref-type="bibr" rid="B45">Rogers and Hahn, 2010a</xref>). MACCS key is the substructure key-based fingerprints, which includes predefined atom symbols, bond types, atom environment properties, atom properties (<xref ref-type="fig" rid="F1">Figure&#x0020;1</xref>) (<xref ref-type="bibr" rid="B13">Durant et&#x0020;al., 2002b</xref>; <xref ref-type="bibr" rid="B6">Cereto-Massagu&#x00e9; et&#x0020;al., 2015</xref>). ECFP encodes local neighborhoods around each atom and bonding connectivity in molecules (<xref ref-type="bibr" rid="B46">Rogers and Hahn, 2010b</xref>). Both MACCS keys and ECFP have gained wide applications in similarity searching (<xref ref-type="bibr" rid="B60">Vilar et&#x0020;al., 2014</xref>; <xref ref-type="bibr" rid="B6">Cereto-Massagu&#x00e9; et&#x0020;al., 2015</xref>), modeling QSAR (<xref ref-type="bibr" rid="B16">Glen et&#x0020;al., 2006</xref>; <xref ref-type="bibr" rid="B36">Myint et&#x0020;al., 2012</xref>), and predicting chemical reactivity (<xref ref-type="bibr" rid="B48">Sandfort et&#x0020;al., 2020</xref>). Wei group has adopted MACCS keys to encode protein and ligand pharmacological space and realized high predictive accuracy and improved high-throughput performance in drug discovery (<xref ref-type="bibr" rid="B28">Li et&#x0020;al., 2019</xref>). Recently, deep learning combined with ECFP fingerprint has been shown as a robust method for high throughput logP predictions, which obtained the root mean square error of 0.61 logP units and ranks as top quarter out of the 92 submissions in the sixth round of Statistical Assessment of the modeling of Proteins and ligands (SAMPL6) competition (<xref ref-type="bibr" rid="B40">Prasad and Brooks, 2020</xref>).</p>
<fig id="F1" position="float">
<label>FIGURE 1</label>
<caption>
<p>The schematic computing procedures of MACCS keys, ECFP fingerprints and the conjoint fingerprints for LSTM and DNN. One molecule is transformed to its feature (vector) space representation using MACCS keys and ECFP fingerprints. Two types of fingerprint are combined as conjoint fingerprint. Output of deep neural network is the predicted properties. <bold>(A)</bold> schematic computing procedures of MACCS keys; <bold>(B)</bold> schematic computing procedures of ECFP fingerprints; <bold>(C)</bold> LSTM trained with conjoint fingerprint; <bold>(D)</bold> DNN trained with conjoint fingerprint.</p>
</caption>
<graphic xlink:href="fphar-11-606668-g001.tif"/>
</fig>
<p>The importance of effective representations of molecules has been recognized when seeking higher accuracy of predicting results (<xref ref-type="bibr" rid="B14">Feinberg et&#x0020;al., 2020</xref>; <xref ref-type="bibr" rid="B30">Lui et&#x0020;al., 2020</xref>). Professional knowledge-based molecular descriptors will be one straightforward way but will bring great challenges for general users who are not familiar with computer techniques. Considering that the standalone molecular descriptors inherently cover parts of information of chemical molecules, we can develop new computing schemes to utilize the existing molecular descriptors. With realization of preference encoding in each molecular descriptor, combining two fingerprints together shows great potential in improving prediction performance in addition to design novel types of fingerprints (<xref ref-type="bibr" rid="B59">Tseng et&#x0020;al., 2012</xref>).</p>
<p>Though the combining of different classes of the molecular fingerprint was proposed by Tseng et&#x0020;al., only important molecular features were selected from the trial descriptor pool that is constructed from molecular representations (<xref ref-type="bibr" rid="B59">Tseng et&#x0020;al., 2012</xref>). The hybrid fingerprints with feature engineering have been actively employed in the field (<xref ref-type="bibr" rid="B37">Nisius and Bajorath, 2009</xref>; <xref ref-type="bibr" rid="B61">Wang et&#x0020;al., 2016</xref>). Features selection can be completed by genetic algorithms, least absolute shrinkage and selection operator (LASSO), or partial least square (PLS), etc. For example, Tseng applied genetic function approximation and multi-dimensional linear regression to select important descriptors from the entire descriptor pool (<xref ref-type="bibr" rid="B50">Senese et&#x0020;al., 2004</xref>). They also employed PLS to highlight import features from multiple descriptor pool in the predictive toxicology modeling (<xref ref-type="bibr" rid="B56">Su et&#x0020;al., 2012</xref>). P&#x00e9;rez-Castillo reported an automatic genetic algorithm to select features for binary classification when they built QSAR modeling (<xref ref-type="bibr" rid="B39">P&#x00e9;rez-Castillo et&#x0020;al., 2012</xref>). Algamal employed the adaptive LASSO method to study high-dimensional QSAR prediction of the anticancer potency. Feature selection methods are active not in QSAR modeling but also in machine learning fields (<xref ref-type="bibr" rid="B1">Algamal et&#x0020;al., 2015</xref>). Bajorath and coworkers extracted main features from MACCS keys, typed-graph distances (TGD) (<xref ref-type="bibr" rid="B54">Sheridan et&#x0020;al., 1996</xref>), typed-graph triangles (TGT) (<xref ref-type="bibr" rid="B58">Tovar et&#x0020;al., 2007</xref>) to form hybrid fingerprints for similarity searching (<xref ref-type="bibr" rid="B37">Nisius and Bajorath, 2009</xref>). In the latest research, Hou et&#x0020;al. also found that proper molecular descriptors selection was able to yield satisfied performance of machine learning (<xref ref-type="bibr" rid="B15">Fu et&#x0020;al., 2020</xref>; <xref ref-type="bibr" rid="B22">Jiang et&#x0020;al., 2020</xref>). These works proved that building proper hybrid fingerprints was one of important techniques for traditional machine learning methods (<xref ref-type="bibr" rid="B5">Cai et&#x0020;al., 2018</xref>). However, feature engineering is required to identify significant molecular features among molecular descriptor pools. Feature engineering process is a tedious and error-prone process and also requires professional knowledge (<xref ref-type="bibr" rid="B62">Wang and Bajorath, 2008</xref>; <xref ref-type="bibr" rid="B21">Hu et&#x0020;al., 2009</xref>).</p>
<p>Nowadays, deep learning shows capability of feature engineering and can automatically train algorithms to learn which fingerprints are important, leading to unique advantages in dealing with complex patterns of big data (<xref ref-type="bibr" rid="B57">Taherkhani et&#x0020;al., 2018</xref>). As been reported, automatically feature extraction endows deep learning with incomparable advantages in predicting physical and chemical properties of molecules in bioinformatics, chemistry, material science and drug discovery fields (<xref ref-type="bibr" rid="B17">Goh et&#x0020;al., 2017</xref>; <xref ref-type="bibr" rid="B47">Sanchez-Lengeling and Aspuru-Guzik, 2018</xref>; <xref ref-type="bibr" rid="B74">Yang et&#x0020;al., 2019</xref>; <xref ref-type="bibr" rid="B22">Jiang et&#x0020;al., 2020</xref>). Can we avoid feature engineering in the days of deep learning? Hop and coworkers proved that machine learned features outperformed than the domain expert engineered features (<xref ref-type="bibr" rid="B20">Hop et&#x0020;al., 2018</xref>). Tseng and coworkers reported that using raw data as molecular representations for deep learning can efficiently learn the most informative features (<xref ref-type="bibr" rid="B7">Chen and Tseng, 2020</xref>). As novel architectures have been developed, molecular descriptors even can be learned from low-level to high-level encodings of molecules during the training process (<xref ref-type="bibr" rid="B24">Kearnes et&#x0020;al., 2016</xref>; <xref ref-type="bibr" rid="B68">Winter et&#x0020;al., 2019</xref>). From the previsou success, we can find that more features lead to better prediction results by using deep learning.</p>
<p>Considering the limitations of standalone featurization and the automatic feature engineering ability of deep learning, we hypothesized that combining two complementary fingerprints rather than relying on expert engineering fingerprints may have room to improve performance of deep learning. As prospected by Tseng et&#x0020;al., the future of QSAR lies in developing new strategies that combine and use 1D through <italic>n</italic>D molecular fingerprints. Endeavors are worthy to be paid for optimizing the use of available descriptors or fingerprints (<xref ref-type="bibr" rid="B59">Tseng et&#x0020;al., 2012</xref>). By harnessing the automatic learning ability of deep learning, combining two types of fingerprints as input data for deep learning rather than manually feature engineering shows great potential in improving predicting performance. Combining multi-dimensional fingerprints preserves as much information as possible. The selecting fingerprints will be automatically completed during the training of machine learning or deep learning (ML/DL) methods by leveraging the learning ability of ML/DL methods. Under this context, we tried to circumvent feature engineering by selecting a proper pair of conjoint fingerprints and ML/DL methods. To facilitate practical applications of deep learning, we adopt two well established molecular fingerprints, MACCS keys and ECFP fingerprints, to construct conjoint fingerprints for deep learning. MACCS keys contain the constituent elements and predefined substructural keys of molecules while lacking substructure connectivity. ECFP contains bonding connectivity and topological features. These two types of molecular fingerprints can provide supplementary information in predicting physicochemical properties. However, the evaluation of combining of MACCS keys and ECFP fingerprints has not been reported.</p>
<p>In this study, we validate the performance of conjoint fingerprints by using three classic machine learning methods (RF, SVR, and XGBoost) and two deep learning methods (LSTM and DNN) in the predicting the logarithm of partition coefficient of a molecule between water and the lipid phase (abbreviated as logP), and the binding affinities of protein-ligand. To demonstrate the generalizability of the proposed conjoint fingerprint, we conducted the regression task on three tasks: predicting logP of DrugBank database (<xref ref-type="bibr" rid="B69">Wishart et&#x0020;al., 2018</xref>), predicting logP of the Lipophilicity dataset that is collected in the MoleculeNet (<xref ref-type="bibr" rid="B70">Wu et&#x0020;al., 2018</xref>), and predicting binding affinities for protein-ligand complex in the PDBbind database (<xref ref-type="bibr" rid="B29">Liu et&#x0020;al., 2014</xref>). It is expected that deep learning can automatically learn the proper representations from conjoint fingerprints, which will overcome the limitations of feature engineering in machine learning methods.</p>
</sec>
<sec sec-type="materials|methods" id="s2">
<title>Materials and Methods</title>
<sec id="s2-1">
<title>Data Preparation</title>
<p>DrugBank database. The molecular structures and the corresponding logP values were obtained from DrugBank database (<xref ref-type="bibr" rid="B69">Wishart et&#x0020;al., 2018</xref>). DrugBank database collects the detailed drug data and the comprehensive drug target information. The logP value is one of most concerned properties of drug molecules, which measures the solubility, absorption, and membrane penetration of drug molecules in the tissues. The DrugBank database contains two subsets: the FDA approved drug molecules, which will be named as &#x201c;Approved&#x201d; subset, and all molecules including potential drugs under study, which will be referred as &#x201c;All&#x201d; subset in this study. Until 2020, there are 13,566 drug entries. Among them, 2011 and 8,656 drug molecules contain logP value entries in the &#x201c;Approved&#x201d; and &#x201c;All&#x201d; subset. In current study, 20&#x0025; of data were randomly selected as test set and the remaining data were further separated as training and validation dataset with the ratio 4:1 in the hyperparameter optimization using Grid Search with cross-validation (GridSearchCV) method (<xref ref-type="bibr" rid="B19">GridSearchCV, 2020</xref>). The dataset was split using the same random seed to keep reproducibility for different validated models.</p>
<p>Lipophilicity dataset. We also selected the Lipophilicity dataset that is collected in the MoleculeNet to present the general applicability of conjoint fingerprints in a different dataset. The Lipophilicity dataset consists of the experimental value of the octanol/water distribution coefficient, which is curated from ChEMBL database. Based on this high-quality dataset, we further validated the performance of standalone and conjoint fingerprints when using 5&#x00a0;ML/DL methods. For comparison, we also computed SlogP by using one traditional logP prediction approach proposed by Wildman-Crippen logP prediction approach (<xref ref-type="bibr" rid="B67">Wildman and Crippen, 1999</xref>), which is implemented in RDKit (<xref ref-type="bibr" rid="B42">RDKit, 2017</xref>). Moreover, we checked the &#x201c;random&#x201d; splitting and &#x201c;scaffold&#x201d; splitting effect on the performance. &#x201c;Scaffold&#x201d; splitter in the DeepChem was used to split the Lipophilicity dataset into training and test subsets (<xref ref-type="bibr" rid="B10">DeepChem, 2018</xref>).</p>
<p>PDBbind dataset. The refined subset of PDBbind was selected because the refined subset contains high quality experimental dissociation constant or inhibition constant (referred as pKi) data for the reasonable number of protein-ligand structures. We used the MACCS keys and ECFP fingerprint to predict pKi of the refined subset of PDBbind. The bound ligands and the binding pockets of protein within 4.5&#x00a0;&#x00c5; of ligand were converted into MACCS keys and ECFP fingerprint, respectively. Water molecule and metal ions within the pocket were deleted due to the technical limitations of RDKit. 4,752 structures were successfully converted to fingerprints. The conjoint fingerprints were built by concatenating MACCS keys and ECFP fingerprint strings. The major focus is on the performance comparison between separated and conjoint fingerprint, therefore the same set of hyperparameters that optimized for DrugBank was adopted.</p>
<p>Fingerprint conversion. The molecular structures and logP were extracted from the SDF files of DrugBank database. The molecules were converted from Cartesian coordinates into vector space representation. Specifically, MACCS keys use a dictionary to check whether the atom types and substructure exist. MACCS keys only cover information of atom and bond types for one molecule and provide limited connecting information in chemical molecules. While ECFP includes the information of how atoms bonded with each other but does not include the chemical properties of each atoms. The combining of MACCS keys and ECFP fingerprints can provide supplementary information in the predicting physicochemical properties. For MACCS keys, the type with 166 keys is the most commonly used in virtual screening. Therefore, each drug molecule was converted into a 166-bit structural MACCS key by checking whether the substructures exist. MACCS keys were computed by using RDKit (<xref ref-type="bibr" rid="B42">RDKit, 2017</xref>). ECFP fingerprints analyze the bonded structural information within the circular radius of atoms. The local structural information around each atom was converted into integer identifiers and then hashed to a bit vector. A radius of two bond lengths was usually used in ECFP. A fixed number of vectors of 2048-bit circular fingerprint were adopted in this study. The ECFP fingerprints were converted by DeepChem open-source package which was developed in Pande group (<xref ref-type="bibr" rid="B41">Ramsundar et&#x0020;al., 2017</xref>).</p>
<p>Machine learning and deep learning algorithms were trained with conjoint fingerprints along with MACCS keys and ECFP fingerprint separately. The conjoint scheme was built by concatenating two strings into one input string (<xref ref-type="fig" rid="F1">Figure&#x0020;1</xref>). The conjoint fingerprint is expected to be more informative by covering both substructural and topological information. The impact of the conjoint fingerprint on performance will be checked in comparison with MACCS keys and ECFP using five learning algorithms.</p>
<p>Random forests. Random forests (RF) was normally selected as a baseline to compare with deep learning methods. RF attracts much interest in QSAR/QSPR studies because it is not sensitive to the hyperparamters. RF outstands from other machine learning methods with advantages of high accuracy (<xref ref-type="bibr" rid="B3">Breiman, 2001</xref>). RF is an ensemble prediction method, which consists of many individual decision trees and the final results are averaged over each individual tree. RF can complete random feature selections in the trees. The difference between RF and DNN is that RF split the whole feature into fragment for each individual tree while DNN can simultaneously process whole features. The number of estimators, tree depth and the number of leafs were selected based on GridSearchCV method.</p>
<p>Support vector regression. The support vector machine (SVM) is designed to classification problems. To do regression, SVR tries to find a hyperplane with the minimized sum of distance from data to the hyperplane. The hyperplane is the combination of functions that parameterized by support vectors. SVR is one popular machine learning methods in QSAR/QSPR with advantages in modeling nonlinear problems. In the &#x201c;RBF&#x201d; kernel, &#x201c;C&#x201d; is the regularization parameter, which is inversely proportional to the strength of the regularization. A higher &#x201c;C&#x201d; value leads to lower tolerance toward to misclassification of training data. &#x201c;Gamma&#x201d; is the coefficient of &#x201c;RBF&#x201d; kernel, which is inversely proportional to the variance of Gaussian distribution. It controls how far the influence of a selected support vector reaches. The value of &#x201c;C&#x201d; and &#x201c;gamma&#x201d; was chosen from a GridSearchCV method using ECFP fingerprint. The &#x201c;RBF&#x201d; kernel function with &#x201c;C&#x201d; equals to five and &#x201c;gamma&#x201d; equals to 0.015 was adopted in this&#x0020;study.</p>
<p>Extreme gradient boosting. Extreme gradient boosting (XGBoost) model is recognized as a new generation of ensemble learning model. It is developed under the Gradient Boosting framework and is developed sequentially in a stagewise additive model. It can solve many data science problems with improved speed and accuracy. It has dominated in machine learning and Kaggle competition with higher performance and robust speed. XGBoost has been reported to achieve comparative performance than deep neural network (<xref ref-type="bibr" rid="B53">Sheridan et&#x0020;al., 2016</xref>).</p>
<p>Architecture of long short-term memory network. Long short-term memory network (LSTM) is improved based on the recurrent neural network (RNN). The advantage of LSTM is its ability to process sequence information with long-term dependency information. LSTM may be benefited from conjoint fingerprints, where two types of fingerprints are kept. The general architecture of LSTM unit is composed of an input gate, a forget gate, an output gate and a memory block. The forget gate is used to decide what information will be forgot from previous cell states. The input gate controls how much information will be kept for new cell states. The output gate determines the output information for new state. LSTM passes information selectively through gating mechanism by incorporating the memory cell that learns when to forget previous hidden states and when to update new hidden states. In this study, we adopted two LSTM layers, which were connected sequentially with one dense layer and one output layer. The time step was set to 1. The output dimension of the first LSTM layer was set to the same dimension of input data. To the best of our knowledge, it is the first time to implement LSTM in predicting logP value for drug molecules.</p>
<p>Architecture of deep neural network. Deep neural network (DNN) is a prototypical deep learning architecture. The important advantage of using DNN is that it can extract useful features from the raw input data. The typical DNN contains three parts: input, hidden and output layer. Each layer contains a set of neurons. We trained different DNN which varied in the size and number of neurons of hidden layers in their architecture. The number of neurons, batch size, epoch number, dropout rate and activation were searched over the hyperparameter space using K-fold cross validation over the training set using GridSearchCV method. In the study, five-fold was employed for the training dataset during hyperparameter optimization. Specifically, the number of neurons of hidden layer used is 10, 20, 40, 50, 60, 100, 300, and 500. Activation function of &#x201c;softsign&#x201d;, &#x201c;rectified linear unit (relu)&#x201d;, &#x201c;linear&#x201d; and &#x201c;tanh&#x201d; were tested each by each. The optimizer of &#x201c;adaptive moment estimation (Adam)&#x201d; with the default learning rate of 0.001 was employed in this study because Adam optimizer uses the adaptive learning momentum and also performs efficiently (<xref ref-type="bibr" rid="B25">Kingma and Ba, 2015</xref>). To reduce the overfitting, the dropout rate from 0 to 0.6 with interval of 0.1 was optimized using GridSearchCV method. During training and validation, the batch size and number of epochs were searched. The processes were repeated 20&#x0020;times to calculate ensemble averages.</p>
<p>Consensus model. We also build the hybrid network models using the consensus model idea of DeepDTA (<xref ref-type="bibr" rid="B38">&#x00d6;zt&#x00fc;rk et&#x0020;al., 2018</xref>). The consensus model has been reported to provide superior performance than single model in some recent researches (<xref ref-type="bibr" rid="B38">&#x00d6;zt&#x00fc;rk et&#x0020;al., 2018</xref>; <xref ref-type="bibr" rid="B15">Fu et&#x0020;al., 2020</xref>). Different from Hou&#x2019;s work (<xref ref-type="bibr" rid="B15">Fu et&#x0020;al., 2020</xref>), we trained machine learning methods with two standalone fingerprints rather than using different types of methods. Consensus model was constructed from different inputs for the same machine learning methods and can reduce statistical bias brought from single learning algorithm.</p>
<p>Performance evaluation. The mean squared errors (MSE) were calculated based on the following equation as the loss function during hyperparameter tuning.<disp-formula id="equ1">
<mml:math id="m1">
<mml:mrow>
<mml:mi>M</mml:mi>
<mml:mi>S</mml:mi>
<mml:mi>E</mml:mi>
<mml:mo>&#x003d;</mml:mo>
<mml:mfrac>
<mml:mrow>
<mml:msup>
<mml:mstyle displaystyle="true">
<mml:mo>&#x2211;</mml:mo>
</mml:mstyle>
<mml:mtext>&#x200b;</mml:mtext>
</mml:msup>
<mml:msup>
<mml:mrow>
<mml:mrow>
<mml:mo>(</mml:mo>
<mml:mrow>
<mml:mi>log</mml:mi>
<mml:mo>&#x2061;</mml:mo>
<mml:msup>
<mml:mi>P</mml:mi>
<mml:mrow>
<mml:mi>e</mml:mi>
<mml:mi>x</mml:mi>
<mml:mi>p</mml:mi>
</mml:mrow>
</mml:msup>
<mml:mo>&#x2212;</mml:mo>
<mml:mi>log</mml:mi>
<mml:mo>&#x2061;</mml:mo>
<mml:msup>
<mml:mi>P</mml:mi>
<mml:mrow>
<mml:mi>c</mml:mi>
<mml:mi>a</mml:mi>
<mml:mi>l</mml:mi>
</mml:mrow>
</mml:msup>
</mml:mrow>
<mml:mo>)</mml:mo>
</mml:mrow>
</mml:mrow>
<mml:mn>2</mml:mn>
</mml:msup>
</mml:mrow>
<mml:mi>n</mml:mi>
</mml:mfrac>
</mml:mrow>
</mml:math>
</disp-formula>The root mean squared errors (RMSE) were computed to present the accuracy of examined learning algorithms. The predicting performance was checked based on a linear correlation between predicted and true logP value in DrugBank database on the given set of drug molecules. The overall agreement between experimental and predicated value was assessed by computing Pearson correlation coefficient according to the following equation<disp-formula id="equ2">
<mml:math id="m2">
<mml:mrow>
<mml:msup>
<mml:mi>R</mml:mi>
<mml:mn>2</mml:mn>
</mml:msup>
<mml:mo>&#x003d;</mml:mo>
<mml:mfrac>
<mml:mrow>
<mml:msup>
<mml:mstyle displaystyle="true">
<mml:mo>&#x2211;</mml:mo>
</mml:mstyle>
<mml:mtext>&#x200b;</mml:mtext>
</mml:msup>
<mml:mrow>
<mml:mo>(</mml:mo>
<mml:mrow>
<mml:mi>log</mml:mi>
<mml:mo>&#x2061;</mml:mo>
<mml:msup>
<mml:mi>P</mml:mi>
<mml:mrow>
<mml:mi>e</mml:mi>
<mml:mi>x</mml:mi>
<mml:mi>p</mml:mi>
</mml:mrow>
</mml:msup>
<mml:mo>&#x2212;</mml:mo>
<mml:mrow>
<mml:mover accent="true">
<mml:mrow>
<mml:mi>log</mml:mi>
<mml:mo>&#x2061;</mml:mo>
<mml:msup>
<mml:mi>P</mml:mi>
<mml:mrow>
<mml:mi>e</mml:mi>
<mml:mi>x</mml:mi>
<mml:mi>p</mml:mi>
</mml:mrow>
</mml:msup>
</mml:mrow>
<mml:mo stretchy="true">&#x00af;</mml:mo>
</mml:mover>
</mml:mrow>
</mml:mrow>
<mml:mo>)</mml:mo>
</mml:mrow>
<mml:mrow>
<mml:mo>(</mml:mo>
<mml:mrow>
<mml:mi>log</mml:mi>
<mml:mo>&#x2061;</mml:mo>
<mml:msup>
<mml:mi>P</mml:mi>
<mml:mrow>
<mml:mi>c</mml:mi>
<mml:mi>a</mml:mi>
<mml:mi>l</mml:mi>
</mml:mrow>
</mml:msup>
<mml:mo>&#x2212;</mml:mo>
<mml:mrow>
<mml:mover accent="true">
<mml:mrow>
<mml:mi>log</mml:mi>
<mml:mo>&#x2061;</mml:mo>
<mml:msup>
<mml:mi>P</mml:mi>
<mml:mrow>
<mml:mi>c</mml:mi>
<mml:mi>a</mml:mi>
<mml:mi>l</mml:mi>
</mml:mrow>
</mml:msup>
</mml:mrow>
<mml:mo stretchy="true">&#x00af;</mml:mo>
</mml:mover>
</mml:mrow>
</mml:mrow>
<mml:mo>)</mml:mo>
</mml:mrow>
</mml:mrow>
<mml:mrow>
<mml:msqrt>
<mml:mrow>
<mml:msup>
<mml:mstyle displaystyle="true">
<mml:mo>&#x2211;</mml:mo>
</mml:mstyle>
<mml:mtext>&#x200b;</mml:mtext>
</mml:msup>
<mml:msup>
<mml:mrow>
<mml:mrow>
<mml:mo>(</mml:mo>
<mml:mrow>
<mml:mi>log</mml:mi>
<mml:mo>&#x2061;</mml:mo>
<mml:msup>
<mml:mi>P</mml:mi>
<mml:mrow>
<mml:mi>e</mml:mi>
<mml:mi>x</mml:mi>
<mml:mi>p</mml:mi>
</mml:mrow>
</mml:msup>
<mml:mo>&#x2212;</mml:mo>
<mml:mrow>
<mml:mover accent="true">
<mml:mrow>
<mml:mi>log</mml:mi>
<mml:mo>&#x2061;</mml:mo>
<mml:msup>
<mml:mi>P</mml:mi>
<mml:mrow>
<mml:mi>e</mml:mi>
<mml:mi>x</mml:mi>
<mml:mi>p</mml:mi>
</mml:mrow>
</mml:msup>
</mml:mrow>
<mml:mo stretchy="true">&#x00af;</mml:mo>
</mml:mover>
</mml:mrow>
</mml:mrow>
<mml:mo>)</mml:mo>
</mml:mrow>
</mml:mrow>
<mml:mn>2</mml:mn>
</mml:msup>
<mml:msup>
<mml:mstyle displaystyle="true">
<mml:mo>&#x2211;</mml:mo>
</mml:mstyle>
<mml:mtext>&#x200b;</mml:mtext>
</mml:msup>
<mml:msup>
<mml:mrow>
<mml:mrow>
<mml:mo>(</mml:mo>
<mml:mrow>
<mml:mi>log</mml:mi>
<mml:mo>&#x2061;</mml:mo>
<mml:msup>
<mml:mi>P</mml:mi>
<mml:mrow>
<mml:mi>c</mml:mi>
<mml:mi>a</mml:mi>
<mml:mi>l</mml:mi>
</mml:mrow>
</mml:msup>
<mml:mo>&#x2212;</mml:mo>
<mml:mrow>
<mml:mover accent="true">
<mml:mrow>
<mml:mi>log</mml:mi>
<mml:mo>&#x2061;</mml:mo>
<mml:msup>
<mml:mi>P</mml:mi>
<mml:mrow>
<mml:mi>c</mml:mi>
<mml:mi>a</mml:mi>
<mml:mi>l</mml:mi>
</mml:mrow>
</mml:msup>
</mml:mrow>
<mml:mo stretchy="true">&#x00af;</mml:mo>
</mml:mover>
</mml:mrow>
</mml:mrow>
<mml:mo>)</mml:mo>
</mml:mrow>
</mml:mrow>
<mml:mn>2</mml:mn>
</mml:msup>
</mml:mrow>
</mml:msqrt>
</mml:mrow>
</mml:mfrac>
</mml:mrow>
</mml:math>
</disp-formula>The Keras-2.2.2 was used to build the models and to optimize hyperparameters. Tensorflow version 1.14 and scikit-learn version 0.20 were used for training and evaluating for five learning algorithms.</p>
</sec>
</sec>
<sec sec-type="results" id="s3">
<title>Results</title>
<sec id="s3-1">
<title>The Distribution of logP in DrugBank Database</title>
<p>The logP is the partition coefficient of a chemical molecule between water and lipid phase, which measures the ability of molecular absorption and excretion. The computational methods of logP prediction can be classified into two major categories: substructure-based and property-based methods. Mannhold et&#x0020;al.&#x2019;s review summarizes the available logP prediction approaches and provides benchmarked results for 30 methods (<xref ref-type="bibr" rid="B32">Mannhold et&#x0020;al., 2009b</xref>). The logP has been used to estimate transport ability of molecules through membranes and metabolisms in tissues, which has been included in the Lipinski&#x2019;s rule of five. Considering the importance of predicting logP, we selected the chemical molecules and the corresponding logP values in DrugBank. The empirical logP values ranged from -4.21 to 9.72. the proportion of 93% of drug molecules shows logP smaller than 5 (<xref ref-type="fig" rid="F2">Figure&#x0020;2</xref>). The distribution is consistent with Lipinski&#x2019;s rule of five, which states that the logP should ideally be not greater than five for orally bioavailable druglike small molecules.</p>
<fig id="F2" position="float">
<label>FIGURE 2</label>
<caption>
<p>The logP value distributions in &#x201c;Approved&#x201d; and &#x201c;All&#x201d; subset of DrugBank. The red line represents the upper value of logP in Lipinski&#x2019;s rule of five invented by Pfizer.</p>
</caption>
<graphic xlink:href="fphar-11-606668-g002.tif"/>
</fig>
<p>The encoded chemical space of MACCS keys, ECFP and conjoint fingerprints can be projected on the principle components to aid visualization. From principle components analysis (PCA), the conjoint fingerprint shows the more degree of dispersion in comparison with MACCS keys and ECFP. As shown in <xref ref-type="fig" rid="F3">Figure&#x0020;3</xref>, MACCS keys and ECFP distributed around a local region and the represented chemical space was not as wide as the conjoint fingerprint, implying more chemical space was kept in conjoint fingerprints. The training set and test set share the same distribution and may guarantee reasonable prediction performance.</p>
<fig id="F3" position="float">
<label>FIGURE 3</label>
<caption>
<p>The principal component analysis plot of the first two principal components based on MACCS keys, ECFP and the conjoint fingerprint.</p>
</caption>
<graphic xlink:href="fphar-11-606668-g003.tif"/>
</fig>
</sec>
<sec id="s3-2">
<title>Hyperparameter Optimization</title>
<p>Tuning hyperparamters is critical for the predicting performance. We conducted Grid Search with cross-validation (GridSearchCV) method to tune hyperparamters with 5-fold cross validation scheme by using &#x201c;Approved&#x201d; and &#x201c;All&#x201d; data subset of DrugBank. Each data set was further separated as training, validation and test sets. We examined the predicting performance by using MACCS keys, ECFP and the conjoint fingerprints. The negative of mean squared error acted as mean score to evaluate the results as shown in <xref ref-type="sec" rid="s10">Supplementary Figures S1&#x2013;S4</xref>. Clearly, the optimal hyperparameters should be tuned in a statistical way as the mean score fluctuated for each examined parameter. The parameters were chosen based on the 20 round cross validations rather than a chance encounter. The selected parameters were summarized in <xref ref-type="sec" rid="s10">Supplementary Tables S1 andS2</xref>.</p>
</sec>
<sec id="s3-3">
<title>Conjoint Fingerprint Improved Predictive Accuracy</title>
<p>The predictive performance for unrecognized molecules was validated in the test subsets using in total five machine learning and deep learning algorithms. The scatter plots of predicted logP against stored logP value in DrugBank were shown in <xref ref-type="fig" rid="F4">Figures 4</xref>, <xref ref-type="fig" rid="F5">5</xref>. Clearly, conjoint fingerprint can provide the better distribution and higher predictive accuracy for the test set than that of MACCS keys and ECFP by using SVR, XGBoost, LSTM and DNN. RMSE was calculated to evaluate the overall error for the test set and was shown in <xref ref-type="table" rid="T1">Table&#x0020;1</xref>. Overall, &#x201c;All&#x201d; subset displayed smaller root mean squared error (RMSE) than &#x201c;Approved&#x201d; subset. From our results, the obvious improvement is observed when dataset changes from &#x201c;Approved&#x201d; to &#x201c;All&#x201d; subset. If there are more high-quality data, the predictive performance of deep learning can be further improved.</p>
<fig id="F4" position="float">
<label>FIGURE 4</label>
<caption>
<p>Scatter plot of the predicted logP vs. stored data in &#x201c;Approved&#x201d; dataset of DrugBank for five models with MACCS, ECFP, consensus model and the conjoint fingerints.</p>
</caption>
<graphic xlink:href="fphar-11-606668-g004.tif"/>
</fig>
<fig id="F5" position="float">
<label>FIGURE 5</label>
<caption>
<p>Scatter plot of the predicted logP vs. stored data in &#x201c;All&#x201d; dataset of DrugBank for five models with MACCS, ECFP, consensus model and the conjoint fingerprints.</p>
</caption>
<graphic xlink:href="fphar-11-606668-g005.tif"/>
</fig>
<table-wrap id="T1" position="float">
<label>TABLE 1</label>
<caption>
<p>Root mean square error and distribution for each validated fingerprints.</p>
</caption>
<table frame="hsides" rules="groups">
<thead>
<tr>
<th rowspan="2"/>
<th rowspan="2" align="center"/>
<th colspan="5" align="center">Approved</th>
<th colspan="5" align="center">All</th>
</tr>
<tr>
<th align="center"/>
<th align="center"/>
<th colspan="3" align="center">% Of molecules within error range</th>
<th align="center"/>
<th align="center"/>
<th colspan="3" align="center">% Of molecules within error range</th>
</tr>
<tr>
<th>Methods</th>
<th align="left">Fingerprint</th>
<th align="center">
<italic>R</italic>
<sup>2</sup>
</th>
<th align="center">RMSE</th>
<th align="center">&#x003c;0.5</th>
<th align="center">0.5&#x2013;1</th>
<th align="center">&#x003e;1</th>
<th align="center">
<italic>R</italic>
<sup>2</sup>
</th>
<th align="center">RMSE</th>
<th align="center">&#x003c;0.5</th>
<th align="center">0.5&#x2013;1</th>
<th align="center">&#x003e;1</th>
</tr>
</thead>
<tbody>
<tr>
<td>RF</td>
<td align="left">MACCS</td>
<td align="center">0.85</td>
<td align="center">1.04</td>
<td align="center">32.3</td>
<td align="center">29.8</td>
<td align="center">37.9</td>
<td align="center">0.89</td>
<td align="center">0.87</td>
<td align="center">37.0</td>
<td align="center">28.5</td>
<td align="center">34.5</td>
</tr>
<tr>
<td>RF</td>
<td align="left">ECFP</td>
<td align="center">0.80</td>
<td align="center">1.24</td>
<td align="center">28.8</td>
<td align="center">21.6</td>
<td align="center">49.6</td>
<td align="center">0.86</td>
<td align="center">1.02</td>
<td align="center">31.7</td>
<td align="center">26.4</td>
<td align="center">41.9</td>
</tr>
<tr>
<td>RF-Cons</td>
<td align="left">MACCS &#x002b; ECFP</td>
<td align="center">0.83</td>
<td align="center">1.01</td>
<td align="center">30.8</td>
<td align="center">25.6</td>
<td align="center">43.67</td>
<td align="center">0.87</td>
<td align="center">0.91</td>
<td align="center">35.4</td>
<td align="center">27.7</td>
<td align="center">36.9</td>
</tr>
<tr>
<td>RF</td>
<td align="left">Conjoint</td>
<td align="center">0.79</td>
<td align="center">1.35</td>
<td align="center">24.1</td>
<td align="center">22.8</td>
<td align="center">53.1</td>
<td align="center">0.85</td>
<td align="center">1.07</td>
<td align="center">30.4</td>
<td align="center">24.8</td>
<td align="center">44.8</td>
</tr>
<tr>
<td>SVR</td>
<td align="left">MACCS</td>
<td align="center">0.89</td>
<td align="center">0.82</td>
<td align="center">43.9</td>
<td align="center">28.3</td>
<td align="center">27.8</td>
<td align="center">0.92</td>
<td align="center">0.63</td>
<td align="center">52.3</td>
<td align="center">27.0</td>
<td align="center">20.6</td>
</tr>
<tr>
<td>SVR</td>
<td align="left">ECFP</td>
<td align="center">0.89</td>
<td align="center">0.79</td>
<td align="center">47.9</td>
<td align="center">25.1</td>
<td align="center">27.0</td>
<td align="center">0.94</td>
<td align="center">0.56</td>
<td align="center">57.8</td>
<td align="center">26.0</td>
<td align="center">16.2</td>
</tr>
<tr>
<td>SVR-Cons</td>
<td align="center">MACCS &#x002b; ECFP</td>
<td align="center">0.89</td>
<td align="center">0.72</td>
<td align="center">49.1</td>
<td align="center">29.0</td>
<td align="center">21.9</td>
<td align="center">0.93</td>
<td align="center">0.52</td>
<td align="center">60.2</td>
<td align="center">26.8</td>
<td align="center">13.0</td>
</tr>
<tr>
<td>SVR</td>
<td align="left">Conjoint</td>
<td align="center">0.92</td>
<td align="center">0.69</td>
<td align="center">52.4</td>
<td align="center">27.3</td>
<td align="center">20.4</td>
<td align="center">0.96</td>
<td align="center">0.48</td>
<td align="center">63.1</td>
<td align="center">25.8</td>
<td align="center">11.1</td>
</tr>
<tr>
<td>XGB</td>
<td align="left">MACCS</td>
<td align="center">0.90</td>
<td align="center">0.75</td>
<td align="center">47.9</td>
<td align="center">26.8</td>
<td align="center">25.3</td>
<td align="center">0.93</td>
<td align="center">0.61</td>
<td align="center">53.7</td>
<td align="center">27.1</td>
<td align="center">19.2</td>
</tr>
<tr>
<td>XGB</td>
<td align="left">ECFP</td>
<td align="center">0.88</td>
<td align="center">0.81</td>
<td align="center">42.9</td>
<td align="center">27.1</td>
<td align="center">30.0</td>
<td align="center">0.93</td>
<td align="center">0.64</td>
<td align="center">50.0</td>
<td align="center">29.5</td>
<td align="center">20.5</td>
</tr>
<tr>
<td>XGB-Cons</td>
<td align="left">MACCS &#x002b; ECFP</td>
<td align="center">0.89</td>
<td align="center">0.71</td>
<td align="center">50.1</td>
<td align="center">26.6</td>
<td align="center">23.3</td>
<td align="center">0.93</td>
<td align="center">0.55</td>
<td align="center">56.6</td>
<td align="center">28.0</td>
<td align="center">15.4</td>
</tr>
<tr>
<td>XGB</td>
<td align="left">Conjoint</td>
<td align="center">0.91</td>
<td align="center">0.69</td>
<td align="center">52.1</td>
<td align="center">27.3</td>
<td align="center">20.6</td>
<td align="center">0.95</td>
<td align="center">0.52</td>
<td align="center">60.0</td>
<td align="center">26.2</td>
<td align="center">13.8</td>
</tr>
<tr>
<td>LSTM</td>
<td align="left">MACCS</td>
<td align="center">0.87</td>
<td align="center">0.82</td>
<td align="center">44.9</td>
<td align="center">27.8</td>
<td align="center">27.3</td>
<td align="center">0.92</td>
<td align="center">0.69</td>
<td align="center">48.3</td>
<td align="center">28.3</td>
<td align="center">23.3</td>
</tr>
<tr>
<td>LSTM</td>
<td align="left">ECFP</td>
<td align="center">0.88</td>
<td align="center">0.89</td>
<td align="center">39.5</td>
<td align="center">29.3</td>
<td align="center">31.3</td>
<td align="center">0.91</td>
<td align="center">0.66</td>
<td align="center">51.0</td>
<td align="center">26.8</td>
<td align="center">22.2</td>
</tr>
<tr>
<td>LSTM-Cons</td>
<td align="left">MACCS &#x002b; ECFP</td>
<td align="center">0.87</td>
<td align="center">0.74</td>
<td align="center">47.9</td>
<td align="center">26.3</td>
<td align="center">25.8</td>
<td align="center">0.92</td>
<td align="center">0.57</td>
<td align="center">56.4</td>
<td align="center">27.2</td>
<td align="center">16.4</td>
</tr>
<tr>
<td>LSTM</td>
<td align="left">Conjoint</td>
<td align="center">0.91</td>
<td align="center">0.75</td>
<td align="center">46.2</td>
<td align="center">28.5</td>
<td align="center">25.3</td>
<td align="center">0.95</td>
<td align="center">0.54</td>
<td align="center">59.6</td>
<td align="center">24.9</td>
<td align="center">15.5</td>
</tr>
<tr>
<td>DNN</td>
<td align="left">MACCS</td>
<td align="center">0.88</td>
<td align="center">0.81</td>
<td align="center">45.7</td>
<td align="center">25.3</td>
<td align="center">29.0</td>
<td align="center">0.91</td>
<td align="center">0.70</td>
<td align="center">47.2</td>
<td align="center">30.0</td>
<td align="center">22.3</td>
</tr>
<tr>
<td>DNN</td>
<td align="left">ECFP</td>
<td align="center">0.89</td>
<td align="center">0.81</td>
<td align="center">43.4</td>
<td align="center">28.8</td>
<td align="center">27.8</td>
<td align="center">0.94</td>
<td align="center">0.60</td>
<td align="center">53.2</td>
<td align="center">29.9</td>
<td align="center">16.9</td>
</tr>
<tr>
<td>DNN-Cons</td>
<td align="left">MACCS &#x002b; ECFP</td>
<td align="center">0.89</td>
<td align="center">0.72</td>
<td align="center">49.9</td>
<td align="center">25.8</td>
<td align="center">24.3</td>
<td align="center">0.92</td>
<td align="center">0.56</td>
<td align="center">55.6</td>
<td align="center">29.1</td>
<td align="center">15.3</td>
</tr>
<tr>
<td>DNN</td>
<td align="left">Conjoint</td>
<td align="center">0.92</td>
<td align="center">0.69</td>
<td align="center">47.6</td>
<td align="center">30.0</td>
<td align="center">22.4</td>
<td align="center">0.96</td>
<td align="center">0.53</td>
<td align="center">57.4</td>
<td align="center">29.3</td>
<td align="center">13.3</td>
</tr>
</tbody>
</table>
</table-wrap>
<p>To facilitate the comparison between different schemes, the same set of hyperparameters that selected based on MACCS keys was used for conjoint fingerprint. From <xref ref-type="table" rid="T1">Table&#x0020;1</xref>, we can notice that the smallest RMSE for &#x201c;Approved&#x201d; and &#x201c;All&#x201d; dataset were 0.686 and 0.475, which obtained from XGBoost and SVR with conjoint fingerprint, respectively. Conjoint fingerprint increased prediction accuracy for SVR, XGBoost, LSTM, and DNN when predicting logP values.</p>
<p>Furthermore, the predictive accuracy was quantified using the deviation counting statistics. We classified the prediction accuracy using the same criteria used by Tetko (<xref ref-type="bibr" rid="B31">Mannhold et&#x0020;al., 2009a</xref>), the deviation between predicted and true logP in the range of 0.0&#x2013;0.5 as considered as &#x201c;acceptable&#x201d;, 0.5&#x2013;1.0 as &#x201c;disputable&#x201d;, and larger than 1.0 as &#x201c;unacceptable&#x201d;. Therefore, counting statistics for RMSE was classified into three regions. For &#x201c;All&#x201d; dataset, the percentages within &#x201c;acceptable&#x201d; range took up to 63.1% when using conjoint fingerprints in SVR, which was higher than that of each standalone fingerprint (52.3 and 57.8% for MACCS and ECFP). Except RF, other methods also achieved similar conclusion. The results demonstrated that the conjoint fingerprint could improve predictive performance and also showed satisfactory generalization ability in predicting logP values of drug molecules. Overall, conjoint fingerprint reproduced the least RMSE than each standalone fingerprint even without optimal hyperparameters.</p>
</sec>
<sec id="s3-4">
<title>Conjoint Fingerprint Boosted Overall Performance</title>
<p>We compared the overall predicting results among RF, SVR, XGBoost, LSTM, and DNN. The Pearson coefficients of the same test set were calculated for all examined methods. The generalization ability is another important indicator to examine the predictive performance of deep learning. We run 20 individual training by randomly separate dataset into training and testing set. The average Pearson coefficients and error bars were computed to present generalization ability. From <xref ref-type="fig" rid="F6">Figure&#x0020;6</xref>, the conjoint fingerprint improved predictive performance over MACCS keys or ECFP, suggesting that the conjoint fingerprints achieved complementarity of two types of fingerprints. DNN generally outperformed over other methods when predicting logP values in &#x201c;Approved&#x201d; subset. The Pearson coefficient of DNN with conjoint fingerprint reached to 0.910. When data becomes more, the kernel-based method, SVR, showed remarkable predictive performance by reproducing the highest Pearson coefficient of 0.959 in &#x201c;All&#x201d; subset. With enough data, SVR displays increasingly performance in treating nonlinear problems and presents better generalization performance. In general, the improvements benefited from the conjoint fingerprint have been realized in SVR, XGBoost, LSTM, and DNN. In this study, we adopted the same set of hyperparamters tuned based on MACCS keys. The performance can be improved with fine-tuned hyperparameters (see <xref ref-type="sec" rid="s10">Supplementary Table S3</xref> for more information). Conjoint fingerprints increase prediction accuracy, implying that the logP of molecules is relevant with both substructures and its neighboring atomic environment. Therefore, the standalone fingerprints cannot surpass the conjoint fingerprints.</p>
<fig id="F6" position="float">
<label>FIGURE 6</label>
<caption>
<p>Pearson coefficients for &#x201c;Approved&#x201d; and &#x201c;All&#x201d; dataset calculated by five learning algorithms using MACCS keys, ECFP, consensus model and the conjoint fingerprints. The similar set of color represents one learning algorithm using different molecular fingerprints. Error bars were computed from 20&#x0020;times individual training and testing processes.</p>
</caption>
<graphic xlink:href="fphar-11-606668-g006.tif"/>
</fig>
</sec>
<sec id="s3-5">
<title>Comparison Between Ensemble Learning and Deep Learning</title>
<p>RF and XGBoost are the ensemble learning methods. The remarkable performance of XGBoost has been demonstrated in previous studies (<xref ref-type="bibr" rid="B27">Lei et&#x0020;al., 2017</xref>). We also obtained consistent results as shown in <xref ref-type="fig" rid="F6">Figure&#x0020;6</xref>. For RF, the Pearson coefficient even decreased for conjoint fingerprint. This is consistent with previous studies that the feature engineering is required for traditional machine learning method (<xref ref-type="bibr" rid="B55">Solorio-Fern&#x00e1;ndez et&#x0020;al., 2020</xref>). As has been pointed by Hou et&#x0020;al., the machine learning methods displayed different prediction capabilities and some machine learning methods showed comparative performance as deep learning (<xref ref-type="bibr" rid="B15">Fu et&#x0020;al., 2020</xref>). Therefore, prediction models should be adopted on a case-by-case&#x0020;basis.</p>
<p>RF employs different approaches to process input fingerprints. RF consists of many decision trees and it splits the fingerprints for each individual tree. Each tree of RF samples parts of input fingerprints and cannot harness the complementarity information from conjoint fingerprint. The presence of irrelevant or redundant fingerprints even reduces the predictive accuracy for machine learning methods (<xref ref-type="bibr" rid="B5">Cai et&#x0020;al., 2018</xref>). While DNN or LSTM can process all input fingerprints at the same time, from which it automatically learns and identifies useful features. The results demonstrate that the proposed conjoint fingerprints can be combined with deep learning to improve predicting accuracy by taking full advantages of automatic feature engineering in DNN and&#x0020;LSTM.</p>
</sec>
<sec id="s3-6">
<title>Comparison Between Conjoint Fingerprint and Consensus Model</title>
<p>Consensus model showed superior performance than each standalone method but did not surpass the performance of conjoint fingerprint. The loss can be tracked during the training process of LSTM and DNN. As revealed in <xref ref-type="fig" rid="F7">Figure&#x0020;7</xref>, MACCS keys and ECFP showed the larger deviation between training and validation subsets in consensus model than that of the conjoint fingerprint. Conjoint fingerprints reproduced the least deviation for both &#x201c;Approved&#x201d; and &#x201c;All&#x201d; subset. The deviation decreased from 0.827 to 0.583 for LSTM when dataset changed from &#x201c;Approved&#x201d; to &#x201c;All&#x201d; (see <xref ref-type="sec" rid="s10">Supplementary Table S4</xref> for more information). The loss value of consensus model in &#x201c;All&#x201d; subset was 1.772 while it was 0.656 for the conjoint fingerprint for DNN. From <xref ref-type="fig" rid="F7">Figure&#x0020;7</xref>, the loss for LSTM and DNN with conjoint fingerprint leveled off within 20 epochs. Consensus models required more training cycle. The loss did not level off until after &#x223c;80 epochs for consensus models. As revealed from the training and validation loss, conjoint fingerprints required less training cycle and increased robustness than consensus&#x0020;model.</p>
<fig id="F7" position="float">
<label>FIGURE 7</label>
<caption>
<p>The loss in the training and validation dataset for MACCS keys, ECFP, the conjoint fingerprints.</p>
</caption>
<graphic xlink:href="fphar-11-606668-g007.tif"/>
</fig>
<p>Conjoint fingerprints scheme outperformed over consensus model that uses standalone fingerprint. The reason may be that the input was trained separately in the consensus model and some information may be lost along with dimension reduction during the training through trees or neural network layers. The results are controlled by &#x201c;buckets effect&#x201d;, which limits further improved predictive accuracy. In contrast, the conjoint fingerprint conserves all information, which can be leveraged by deep learning to reproduce more accurate results.</p>
</sec>
<sec id="s3-7">
<title>The Generalizability of Conjoint Fingerprints for Other Regression Tasks</title>
<p>Conjoint fingerprint is applicable to the Lipophilicity dataset from MoleculeNet. For all of five examined ML/DL methods, the predicted performance was improved by using conjoint fingerprints as shown in <xref ref-type="fig" rid="F8">Figure&#x0020;8</xref>. The predicted Pearson coefficient exceeded 0.8 by using SVR and XGBoost. The results were compared with one available computational method, SlogP computed by Wildman-Crippen logP prediction approach. On the same test set, the ML/DL methods outperformed over Wildman-Crippen logP computation method when using the current dataset. Wildman-Crippen logP reproduced different Pearson coefficient on the different split subset, implying that Wildman-Crippen logP computational method may also depend on the training dataset.</p>
<fig id="F8" position="float">
<label>FIGURE 8</label>
<caption>
<p>Pearson coefficients for Lipophilicity dataset calculated by five learning algorithms using MACCS keys, ECFP, consensus model and the conjoint fingerprints. The similar set of color represents one learning algorithm using different molecular fingerprints. Wildman-Crippen logP labeled as W-C logP.</p>
</caption>
<graphic xlink:href="fphar-11-606668-g008.tif"/>
</fig>
<p>We also noticed that random splitting led to better performance of ML/DL methods than scaffold splitting. This is consistent with previous conclusions that substructure-based fingerprints likely result in better performance during random splitting than scaffold splitting. Scaffold splitting attempts to separate different chemical scaffold molecules into different subsets. Therefore, scaffold splitting can reveal the true learning abilities of ML/DL methods. In <xref ref-type="fig" rid="F8">Figure&#x0020;8</xref>, the Pearson coefficient difference between conjoint fingerprints and consensus model become more obvious, suggesting that the superiority of conjoint fingerprint over the consensus model. The result reminds us that we can quickly evaluate prediction quality by checking the substructure similarity between the training dataset and the test samples during practical applications.</p>
<p>To demonstrate the generalizability of the proposed conjoint fingerprint, we conducted the regression task for PDBbind dataset. The Pearson coefficient between predicted and experimental pKi was computed for each ML/DL methods using MACCS keys, ECFP and conjoint fingerprints. Among evaluated methods, RF, SVR, and XGBboost produced the similar Pearson coefficient for ECFP and conjoint fingerprint as shown in <xref ref-type="fig" rid="F9">Figure&#x0020;9</xref>. LSTM and DNN lead to a higher Pearson coefficient for conjoint fingerprint than MACCS keys or ECFP. The Pearson coefficients obtained from conjoint fingerprints were higher than that obtained from the consensus model, implying that the combination of fingerprints can at least act as an alternative approach to the consensus model. The best predicting performance was achieved by the pairing of SVR and conjoint fingerprint, reaching the highest Pearson coefficient of 0.74, which is comparable to the predicted result with the grid featurization (<xref ref-type="bibr" rid="B72">Xie et&#x0020;al., 2020</xref>). Therefore, the conjoint fingerprint also contributed to the improved predicting performance in the regression task for PDBbind. The combination of two fingerprints will embody the information from each fingerprint. Without feature engineering, that taking all the combined fingerprints as the input for the ML/DL methods will provide more information while it also brings challenges for ML/DL at the meantime. Therefore we should select the matched ML/DL methods for the conjoint fingerprint via trial and error process. We believed that more improvement can be realized after optimizing hyperparameters for each ML/DL methods.</p>
<fig id="F9" position="float">
<label>FIGURE 9</label>
<caption>
<p>Pearson coefficients for PDBbind dataset calculated by five learning algorithms using MACCS keys, ECFP, consensus model and the conjoint fingerprints. The similar set of color represents one learning algorithm using different molecular fingerprints.</p>
</caption>
<graphic xlink:href="fphar-11-606668-g009.tif"/>
</fig>
<p>From our evaluation, we can notice that combining two types of fingerprints can obtain improved predicting performance than consensus model. Our manuscript acted as the preliminary demonstration on how to select multi-dimensional molecular fingerprints with matched ML/DL methods to circumvent feature selection. The combining scheme can be generally extended to other types of molecular descriptors and fingerprints. A rigorous evaluation of the conjoint fingerprints to check whether the conjoint fingerprint&#x0027;s superiority is statistically significant will be conducted in the future&#x0020;work.</p>
</sec>
</sec>
<sec sec-type="discussion" id="s4">
<title>Discussion</title>
<p>When developing molecular representations, molecular descriptors have been optimized for the specific applications. Recombination of different types of molecular descriptors would be a convenient forward way to improved performance, especially for general users with no knowledge of molecular descriptor design. From our evaluation, we can see that conjoint fingerprint can improve predictive accuracy and reduce training cycles by leveraging automatic feature learning ability of deep learning. Using conjoint fingerprints, SVR and XGBoost achieved the comparative performance as that of LSTM and DNN. In applications, the choice of machine learning or deep learning depended on the task. The standalone featurization may inherently cover parts of information in the chemical molecules and thus the combination of accessible fingerprints would improve the predictive power of deep learning. The following points of view can be considered to improve predictive performance of deep learning when using conjoint fingerprints.<list list-type="order">
<list-item>
<p>(1) Both MACCS keys and ECFP have been well documented in open-source software and thus other researchers can adopt them in their researches, which should facilitate applications of deep learning. Besides, we have witnessed great development in novel types of molecular descriptors in the last decade. Besides MACCS keys and ECFP, three dimensional types of ECFP (<xref ref-type="bibr" rid="B2">Axen et&#x0020;al., 2017</xref>), molecular graph convolutions (<xref ref-type="bibr" rid="B24">Kearnes et&#x0020;al., 2016</xref>) and atomic convolutional networks (<xref ref-type="bibr" rid="B23">Gomes et&#x0020;al., 2017</xref>) have been developed. The conjoint fingerprints can be built from other types of molecular descriptors besides the substructure based fingerprints. For example, conjoint fingerprints can be extended to include atomic or fragment-based molecular descriptors in the future work. Each new types of molecular descriptors show different merits. If they can provide open-source tools, it is worthy of conducting systematic search to find out the optimal combination of different types of molecular descriptors.</p>
</list-item>
<list-item>
<p>(2) For architectures of neural networks, convolutional neural networks and recurrent neural networks present as another exciting starting point to improve predictive performance of deep learning. Deep learning uses the hierarchical learning of representations (<xref ref-type="bibr" rid="B75">Zeiler and Fergus, 2014</xref>). The lowest layers of neural networks learn simple features that will be used to build higher order information along with their propagation through the networks. The informative features can be captured during hidden layers by automatically constructing one intermediate feature space. Deep learning will be expected reduce tedious works on intricate feature engineering. Experts from computer or related fields can provide more valuable insights if they have access to structural, topological and graphical fingerprints and other powerful deep learning architectures by following current protocol.</p>
</list-item>
</list>
</p>
</sec>
<sec sec-type="conclusion" id="s5">
<title>Conclusion</title>
<p>We validated the impact of the conjoint fingerprints on three well established machine learning methods and two emerging deep learning methods, including RF, SVR, XGBoost, LSTM, and DNN. Combining MACCS keys with ECFP achieved complementarity in substructural and topological fingerprints, which can be processed by machine learning and deep learning algorithms to find the inherent rules between the demanded activity/property and their structures of drug molecules. Our results demonstrated that the conjoint fingerprints achieved the least loss and the highest Pearson coefficients than that of each standalone fingerprint for SVR, XGBoost, LSTM, and DNN, even surpassing the consensus model. By complementarily combining two types of fingerprints, boosted performance can be achieved than that of using single molecular descriptor. The proposed conjoint fingerprint scheme can be generally extended to other types of molecular descriptor. We anticipate that our proposed conjoint scheme would invoke following studies by integrating structural, topological or spatial fingerprints in deep learning&#x0020;area.</p>
</sec>
<sec id="s6">
<title>Data Availability Statement</title>
<p>The datasets presented in this study can be found in online repositories. The names of the repository/repositories and accession numbers can be found below: <ext-link ext-link-type="uri" xlink:href="https://github.com/xlxgit/AlogP-DL.git,%20github">https://github.com/xlxgit/AlogP-DL.git, github</ext-link>.</p>
</sec>
<sec id="s7">
<title>Author Contributions</title>
<p>LiX and LeX contributed equally. The manuscript was written through contributions of all authors. All authors have given approval to the final version of the manuscript.</p>
</sec>
<sec id="s8">
<title>Funding</title>
<p>This work was supported by the fund of the Natural Science Foundation of Jiangsu Province (BK20191032 and BE2019650), the National Natural Science Foundation of China (22003020, 12074151, and 81803430), Changzhou Sci. and Tech. Program (CJ20200045), and open funding from Jiangsu Sino-Israel Industrial Technology Research Institute (JSIITRI202009).</p>
</sec>
<sec sec-type="COI-statement" id="s9">
<title>Conflict of Interest</title>
<p>The authors declare that the research was conducted in the absence of any commercial or financial relationships that could be construed as a potential conflict of interest.</p>
</sec>
</body>
<back>
<sec id="s10">
<title>Supplementary Material</title>
<p>The Supplementary Material for this article can be found online at: <ext-link ext-link-type="uri" xlink:href="https://www.frontiersin.org/articles/10.3389/fphar.2020.606668/full#supplementary-material">https://www.frontiersin.org/articles/10.3389/fphar.2020.606668/full&#x0023;supplementary-material</ext-link>.</p>
<supplementary-material xlink:href="datasheet1.docx" id="SM1" mimetype="application/docx" xmlns:xlink="http://www.w3.org/1999/xlink"/>
</sec>
<ref-list>
<title>References</title>
<ref id="B1">
<citation citation-type="journal">
<person-group person-group-type="author">
<name>
<surname>Algamal</surname>
<given-names>Z. Y.</given-names>
</name>
<name>
<surname>Lee</surname>
<given-names>M. H.</given-names>
</name>
<name>
<surname>Al-Fakih</surname>
<given-names>A. M.</given-names>
</name>
<name>
<surname>Aziz</surname>
<given-names>M.</given-names>
</name>
</person-group> (<year>2015</year>). <article-title>High-dimensional QSAR prediction of anticancer potency of imidazo[4,5-b]pyridine derivatives using adjusted adaptive LASSO</article-title>. <source>J.&#x0020;Chemometr.</source> <volume>29</volume>, <fpage>547</fpage>&#x2013;<lpage>556</lpage>. <pub-id pub-id-type="doi">10.1002/cem.2741</pub-id> </citation>
</ref>
<ref id="B2">
<citation citation-type="journal">
<person-group person-group-type="author">
<name>
<surname>Axen</surname>
<given-names>S. D.</given-names>
</name>
<name>
<surname>Huang</surname>
<given-names>X. P.</given-names>
</name>
<name>
<surname>C&#x00e1;ceres</surname>
<given-names>E. L.</given-names>
</name>
<name>
<surname>Gendelev</surname>
<given-names>L.</given-names>
</name>
<name>
<surname>Roth</surname>
<given-names>B. L.</given-names>
</name>
<name>
<surname>Keiser</surname>
<given-names>M. J.</given-names>
</name>
</person-group> (<year>2017</year>). <article-title>A simple representation of three-dimensional molecular structure</article-title>. <source>J.&#x0020;Med. Chem.</source> <volume>60</volume>, <fpage>7393</fpage>&#x2013;<lpage>7409</lpage>. <pub-id pub-id-type="doi">10.1021/acs.jmedchem.7b00696</pub-id> </citation>
</ref>
<ref id="B3">
<citation citation-type="journal">
<person-group person-group-type="author">
<name>
<surname>Breiman</surname>
<given-names>L.</given-names>
</name>
</person-group> (<year>2001</year>). <article-title>Random forests</article-title>. <source>Mach. Learn.</source> <volume>45</volume>, <fpage>5</fpage>&#x2013;<lpage>32</lpage>. <pub-id pub-id-type="doi">10.1023/A:1010933404324</pub-id> </citation>
</ref>
<ref id="B4">
<citation citation-type="journal">
<person-group person-group-type="author">
<name>
<surname>Butler</surname>
<given-names>K. T.</given-names>
</name>
<name>
<surname>Davies</surname>
<given-names>D. W.</given-names>
</name>
<name>
<surname>Cartwright</surname>
<given-names>H.</given-names>
</name>
<name>
<surname>Isayev</surname>
<given-names>O.</given-names>
</name>
<name>
<surname>Walsh</surname>
<given-names>A.</given-names>
</name>
</person-group> (<year>2018</year>). <article-title>Machine learning for molecular and materials science</article-title>. <source>Nature</source> <volume>559</volume>, <fpage>547</fpage>&#x2013;<lpage>555</lpage>. <pub-id pub-id-type="doi">10.1038/s41586-018-0337-2</pub-id> </citation>
</ref>
<ref id="B5">
<citation citation-type="journal">
<person-group person-group-type="author">
<name>
<surname>Cai</surname>
<given-names>J.</given-names>
</name>
<name>
<surname>Luo</surname>
<given-names>J.</given-names>
</name>
<name>
<surname>Wang</surname>
<given-names>S.</given-names>
</name>
<name>
<surname>Yang</surname>
<given-names>S.</given-names>
</name>
</person-group> (<year>2018</year>). <article-title>Feature selection in machine learning: a new perspective</article-title>. <source>Neurocomputing</source> <volume>300</volume>, <fpage>70</fpage>&#x2013;<lpage>79</lpage>. <pub-id pub-id-type="doi">10.1016/j.neucom.2017.11.077</pub-id> </citation>
</ref>
<ref id="B6">
<citation citation-type="journal">
<person-group person-group-type="author">
<name>
<surname>Cereto-Massagu&#x00e9;</surname>
<given-names>A.</given-names>
</name>
<name>
<surname>Ojeda</surname>
<given-names>M. J.</given-names>
</name>
<name>
<surname>Valls</surname>
<given-names>C.</given-names>
</name>
<name>
<surname>Mulero</surname>
<given-names>M.</given-names>
</name>
<name>
<surname>Garcia-Vallv&#x00e9;</surname>
<given-names>S.</given-names>
</name>
<name>
<surname>Pujadas</surname>
<given-names>G.</given-names>
</name>
</person-group> (<year>2015</year>). <article-title>Molecular fingerprint similarity search in virtual screening</article-title>. <source>Methods</source> <volume>71</volume>, <fpage>58</fpage>&#x2013;<lpage>63</lpage>. <pub-id pub-id-type="doi">10.1016/j.ymeth.2014.08.005</pub-id> </citation>
</ref>
<ref id="B7">
<citation citation-type="journal">
<person-group person-group-type="author">
<name>
<surname>Chen</surname>
<given-names>J.-H.</given-names>
</name>
<name>
<surname>Tseng</surname>
<given-names>Y. J.</given-names>
</name>
</person-group> (<year>2020</year>). <article-title>Different molecular enumeration influences in deep learning: an example using aqueous solubility</article-title>. <source>Briefings Bioinf.</source> [<comment>Epub ahead of print</comment>]. <pub-id pub-id-type="doi">10.1093/bib/bbaa092</pub-id> </citation>
</ref>
<ref id="B8">
<citation citation-type="journal">
<person-group person-group-type="author">
<name>
<surname>Chuang</surname>
<given-names>K. V.</given-names>
</name>
<name>
<surname>Gunsalus</surname>
<given-names>L. M.</given-names>
</name>
<name>
<surname>Keiser</surname>
<given-names>M. J.</given-names>
</name>
</person-group> (<year>2020</year>). <article-title>Learning molecular representations for medicinal chemistry</article-title>. <source>J.&#x0020;Med. Chem.</source> <volume>63</volume> (<issue>16</issue>), <fpage>8705</fpage>&#x2013;<lpage>8722</lpage>. <pub-id pub-id-type="doi">10.1021/acs.jmedchem.0c00385</pub-id> </citation>
</ref>
<ref id="B9">
<citation citation-type="journal">
<person-group person-group-type="author">
<name>
<surname>Danishuddin</surname>
</name>
<name>
<surname>Khan</surname>
<given-names>A. U.</given-names>
</name>
</person-group> (<year>2016</year>). <article-title>Descriptors and their selection methods in QSAR analysis: paradigm for drug design</article-title>. <source>Drug Discov. Today</source> <volume>21</volume>, <fpage>1291</fpage>&#x2013;<lpage>1302</lpage>. <pub-id pub-id-type="doi">10.1016/j.drudis.2016.06.013</pub-id> </citation>
</ref>
<ref id="B10">
<citation citation-type="web">
<collab>DeepChem</collab> (<year>2018</year>). <article-title>Deep-learning models for drug discovery and quantum chemistry</article-title>. <comment>Available at: <ext-link ext-link-type="uri" xlink:href="https://github.com/deepchem/deepchem">https://github.com/deepchem/deepchem</ext-link> (Accessed January 3, 2018)</comment>. </citation>
</ref>
<ref id="B11">
<citation citation-type="journal">
<person-group person-group-type="author">
<name>
<surname>Duan</surname>
<given-names>J.</given-names>
</name>
<name>
<surname>Dixon</surname>
<given-names>S. L.</given-names>
</name>
<name>
<surname>Lowrie</surname>
<given-names>J.&#x0020;F.</given-names>
</name>
<name>
<surname>Sherman</surname>
<given-names>W.</given-names>
</name>
</person-group> (<year>2010</year>). <article-title>Analysis and comparison of 2D fingerprints: insights into database screening performance using eight fingerprint methods</article-title>. <source>J.&#x0020;Mol. Graph. Model.</source> <volume>29</volume>, <fpage>157</fpage>&#x2013;<lpage>170</lpage>. <pub-id pub-id-type="doi">10.1016/j.jmgm.2010.05.008</pub-id> </citation>
</ref>
<ref id="B12">
<citation citation-type="journal">
<person-group person-group-type="author">
<name>
<surname>Durant</surname>
<given-names>J.&#x0020;L.</given-names>
</name>
<name>
<surname>Leland</surname>
<given-names>B. A.</given-names>
</name>
<name>
<surname>Henry</surname>
<given-names>D. R.</given-names>
</name>
<name>
<surname>Nourse</surname>
<given-names>J.&#x0020;G.</given-names>
</name>
</person-group> (<year>2002a</year>). <article-title>Reoptimization of MDL keys for use in drug discovery</article-title>. <source>J.&#x0020;Chem. Inf. Comput. Sci.</source> <volume>42</volume>, <fpage>1273</fpage>. <pub-id pub-id-type="doi">10.1021/ci010132r</pub-id> </citation>
</ref>
<ref id="B13">
<citation citation-type="journal">
<person-group person-group-type="author">
<name>
<surname>Durant</surname>
<given-names>J.&#x0020;L.</given-names>
</name>
<name>
<surname>Leland</surname>
<given-names>B. A.</given-names>
</name>
<name>
<surname>Henry</surname>
<given-names>D. R.</given-names>
</name>
<name>
<surname>Nourse</surname>
<given-names>J.&#x0020;G.</given-names>
</name>
</person-group> (<year>2002b</year>). <article-title>Reoptimization of MDL keys for use in drug discovery</article-title>. <source>J.&#x0020;Chem. Inf. Comput. Sci.</source> <volume>42</volume>, <fpage>1273</fpage>&#x2013;<lpage>1280</lpage>. <pub-id pub-id-type="doi">10.1021/ci010132r</pub-id> </citation>
</ref>
<ref id="B14">
<citation citation-type="journal">
<person-group person-group-type="author">
<name>
<surname>Feinberg</surname>
<given-names>E. N.</given-names>
</name>
<name>
<surname>Joshi</surname>
<given-names>E.</given-names>
</name>
<name>
<surname>Pande</surname>
<given-names>V. S.</given-names>
</name>
<name>
<surname>Cheng</surname>
<given-names>A. C.</given-names>
</name>
</person-group> (<year>2020</year>). <article-title>Improvement in ADMET prediction with multitask deep featurization</article-title>. <source>J.&#x0020;Med. Chem.</source> <volume>63</volume> (<issue>16</issue>), <fpage>8835</fpage>&#x2013;<lpage>8848</lpage>. <pub-id pub-id-type="doi">10.1021/acs.jmedchem.9b02187</pub-id> </citation>
</ref>
<ref id="B15">
<citation citation-type="journal">
<person-group person-group-type="author">
<name>
<surname>Fu</surname>
<given-names>L.</given-names>
</name>
<name>
<surname>Liu</surname>
<given-names>L.</given-names>
</name>
<name>
<surname>Yang</surname>
<given-names>Z. J.</given-names>
</name>
<name>
<surname>Li</surname>
<given-names>P.</given-names>
</name>
<name>
<surname>Ding</surname>
<given-names>J.&#x0020;J.</given-names>
</name>
<name>
<surname>Yun</surname>
<given-names>Y. H.</given-names>
</name>
<etal/>
</person-group> (<year>2020</year>). <article-title>Systematic modeling of log&#x202f;D7.4 based on ensemble machine learning, group contribution, and matched molecular pair analysis</article-title>. <source>J.&#x0020;Chem. Inf. Model.</source> <volume>60</volume>, <fpage>63</fpage>&#x2013;<lpage>76</lpage>. <pub-id pub-id-type="doi">10.1021/acs.jcim.9b00718</pub-id> </citation>
</ref>
<ref id="B16">
<citation citation-type="journal">
<person-group person-group-type="author">
<name>
<surname>Glen</surname>
<given-names>R.</given-names>
</name>
<name>
<surname>Bender</surname>
<given-names>A.</given-names>
</name>
<name>
<surname>Arnby</surname>
<given-names>C.</given-names>
</name>
<name>
<surname>Carlsson</surname>
<given-names>L.</given-names>
</name>
<name>
<surname>Boyer</surname>
<given-names>S.</given-names>
</name>
<name>
<surname>Smith</surname>
<given-names>J.</given-names>
</name>
</person-group> (<year>2006</year>). <article-title>Circular fingerprints: flexible molecular descriptors with applications from physical chemistry to ADME</article-title>. <source>IDrugs</source> <volume>9</volume>, <fpage>199</fpage>&#x2013;<lpage>204</lpage>. </citation>
</ref>
<ref id="B17">
<citation citation-type="journal">
<person-group person-group-type="author">
<name>
<surname>Goh</surname>
<given-names>G. B.</given-names>
</name>
<name>
<surname>Hodas</surname>
<given-names>N. O.</given-names>
</name>
<name>
<surname>Vishnu</surname>
<given-names>A.</given-names>
</name>
</person-group> (<year>2017</year>). <article-title>Deep learning for computational chemistry</article-title>. <source>J.&#x0020;Comput. Chem.</source> <volume>38</volume>, <fpage>1291</fpage>&#x2013;<lpage>1307</lpage>. <pub-id pub-id-type="doi">10.1002/jcc.24764</pub-id> </citation>
</ref>
<ref id="B18">
<citation citation-type="journal">
<person-group person-group-type="author">
<name>
<surname>Grambow</surname>
<given-names>C. A.</given-names>
</name>
<name>
<surname>Pattanaik</surname>
<given-names>L.</given-names>
</name>
<name>
<surname>Green</surname>
<given-names>W. H.</given-names>
</name>
</person-group> (<year>2020</year>). <article-title>Deep learning of activation energies</article-title>. <source>J.&#x0020;Phys. Chem. Lett.</source> <volume>11</volume>, <fpage>2992</fpage>&#x2013;<lpage>2997</lpage>. <pub-id pub-id-type="doi">10.1021/acs.jpclett.0c00500</pub-id> </citation>
</ref>
<ref id="B19">
<citation citation-type="web">
<collab>GridSearchCV</collab> (<year>2020</year>). <article-title>klearn.model_selection.GridSearchCV [Online]</article-title>. <comment>Available: <ext-link ext-link-type="uri" xlink:href="https://scikit-learn.org/stable/modules/generated/sklearn.model_selection.GridSearchCV.html">https://scikit-learn.org/stable/modules/generated/sklearn.model_selection.GridSearchCV.html</ext-link> (Accessed February 14, 2020)</comment>. </citation>
</ref>
<ref id="B20">
<citation citation-type="journal">
<person-group person-group-type="author">
<name>
<surname>Hop</surname>
<given-names>P.</given-names>
</name>
<name>
<surname>Allgood</surname>
<given-names>B.</given-names>
</name>
<name>
<surname>Yu</surname>
<given-names>J.</given-names>
</name>
</person-group> (<year>2018</year>). <article-title>Geometric deep learning autonomously learns chemical features that outperform those engineered by domain experts</article-title>. <source>Mol. Pharm.</source> <volume>15</volume>, <fpage>4371</fpage>&#x2013;<lpage>4377</lpage>. <pub-id pub-id-type="doi">10.1021/acs.molpharmaceut.7b01144</pub-id> </citation>
</ref>
<ref id="B21">
<citation citation-type="journal">
<person-group person-group-type="author">
<name>
<surname>Hu</surname>
<given-names>Y.</given-names>
</name>
<name>
<surname>Lounkine</surname>
<given-names>E.</given-names>
</name>
<name>
<surname>Bajorath</surname>
<given-names>J.</given-names>
</name>
</person-group> (<year>2009</year>). <article-title>Improving the search performance of extended connectivity fingerprints through activity-oriented feature filtering and application of a bit-density-dependent similarity function</article-title>. <source>ChemMedChem</source> <volume>4</volume>, <fpage>540</fpage>&#x2013;<lpage>548</lpage>. <pub-id pub-id-type="doi">10.1002/cmdc.200800408</pub-id> </citation>
</ref>
<ref id="B22">
<citation citation-type="journal">
<person-group person-group-type="author">
<name>
<surname>Jiang</surname>
<given-names>D.</given-names>
</name>
<name>
<surname>Lei</surname>
<given-names>T.</given-names>
</name>
<name>
<surname>Wang</surname>
<given-names>Z.</given-names>
</name>
<name>
<surname>Shen</surname>
<given-names>C.</given-names>
</name>
<name>
<surname>Cao</surname>
<given-names>D.</given-names>
</name>
<name>
<surname>Hou</surname>
<given-names>T.</given-names>
</name>
</person-group> (<year>2020</year>). <article-title>ADMET evaluation in drug discovery. 20. Prediction of breast cancer resistance protein inhibition through machine learning</article-title>. <source>J.&#x0020;Cheminf.</source> <volume>12</volume>, <fpage>16</fpage>. <pub-id pub-id-type="doi">10.1186/s13321-020-00421-y</pub-id> </citation>
</ref>
<ref id="B23">
<citation citation-type="journal">
<person-group person-group-type="author">
<name>
<surname>Gomes</surname>
<given-names>J.</given-names>
</name>
<name>
<surname>Ramsundar</surname>
<given-names>B.</given-names>
</name>
<name>
<surname>Feinberg</surname>
<given-names>E. N.</given-names>
</name>
<name>
<surname>Pande</surname>
<given-names>V. S.</given-names>
</name>
</person-group> (<year>2017</year>). <article-title>Atomic convolutional networks for predicting protein-ligand binding affinity</article-title>. <source>arXiv:1703.10603</source> </citation>
</ref>
<ref id="B24">
<citation citation-type="journal">
<person-group person-group-type="author">
<name>
<surname>Kearnes</surname>
<given-names>S.</given-names>
</name>
<name>
<surname>Mccloskey</surname>
<given-names>K.</given-names>
</name>
<name>
<surname>Berndl</surname>
<given-names>M.</given-names>
</name>
<name>
<surname>Pande</surname>
<given-names>V.</given-names>
</name>
<name>
<surname>Riley</surname>
<given-names>P.</given-names>
</name>
</person-group> (<year>2016</year>). <article-title>Molecular graph convolutions: moving beyond fingerprints</article-title>. <source>J.&#x0020;Comput. Aided Mol. Des.</source> <volume>30</volume>, <fpage>595</fpage>&#x2013;<lpage>608</lpage>. <pub-id pub-id-type="doi">10.1007/s10822-016-9938-8</pub-id> </citation>
</ref>
<ref id="B25">
<citation citation-type="journal">
<person-group person-group-type="author">
<name>
<surname>Kingma</surname>
<given-names>D. P.</given-names>
</name>
<name>
<surname>Ba</surname>
<given-names>J.&#x0020;L.</given-names>
</name>
</person-group> (<year>2015</year>). <article-title>Adam: a method for stochastic optimization</article-title>. <source>arXiv:1412.6980</source> </citation>
</ref>
<ref id="B26">
<citation citation-type="journal">
<person-group person-group-type="author">
<name>
<surname>Lecun</surname>
<given-names>Y.</given-names>
</name>
<name>
<surname>Bengio</surname>
<given-names>Y.</given-names>
</name>
<name>
<surname>Hinton</surname>
<given-names>G.</given-names>
</name>
</person-group> (<year>2015</year>). <article-title>Deep learning</article-title>. <source>Nature</source> <volume>521</volume>, <fpage>436</fpage>. <pub-id pub-id-type="doi">10.1038/nature14539</pub-id> </citation>
</ref>
<ref id="B27">
<citation citation-type="journal">
<person-group person-group-type="author">
<name>
<surname>Lei</surname>
<given-names>T.</given-names>
</name>
<name>
<surname>Sun</surname>
<given-names>H.</given-names>
</name>
<name>
<surname>Kang</surname>
<given-names>Y.</given-names>
</name>
<name>
<surname>Zhu</surname>
<given-names>F.</given-names>
</name>
<name>
<surname>Liu</surname>
<given-names>H.</given-names>
</name>
<name>
<surname>Zhou</surname>
<given-names>W.</given-names>
</name>
<etal/>
</person-group> (<year>2017</year>). <article-title>ADMET evaluation in drug discovery. 18. Reliable prediction of chemical-induced urinary tract toxicity by boosting machine learning approaches</article-title>. <source>Mol. Pharm.</source> <volume>14</volume>, <fpage>3935</fpage>&#x2013;<lpage>3953</lpage>. <pub-id pub-id-type="doi">10.1021/acs.molpharmaceut.7b00631</pub-id> </citation>
</ref>
<ref id="B28">
<citation citation-type="journal">
<person-group person-group-type="author">
<name>
<surname>Li</surname>
<given-names>L.</given-names>
</name>
<name>
<surname>Koh</surname>
<given-names>C. C.</given-names>
</name>
<name>
<surname>Reker</surname>
<given-names>D.</given-names>
</name>
<name>
<surname>Brown</surname>
<given-names>J.&#x0020;B.</given-names>
</name>
<name>
<surname>Wang</surname>
<given-names>H.</given-names>
</name>
<name>
<surname>Lee</surname>
<given-names>N. K.</given-names>
</name>
<etal/>
</person-group> (<year>2019</year>). <article-title>Predicting protein-ligand interactions based on bow-pharmacological space and Bayesian additive regression trees</article-title>. <source>Sci. Rep.</source> <volume>9</volume>, <fpage>7703</fpage>. <pub-id pub-id-type="doi">10.1038/s41598-019-43125-6</pub-id> </citation>
</ref>
<ref id="B29">
<citation citation-type="journal">
<person-group person-group-type="author">
<name>
<surname>Liu</surname>
<given-names>Z.</given-names>
</name>
<name>
<surname>Li</surname>
<given-names>Y.</given-names>
</name>
<name>
<surname>Han</surname>
<given-names>L.</given-names>
</name>
<name>
<surname>Li</surname>
<given-names>J.</given-names>
</name>
<name>
<surname>Liu</surname>
<given-names>J.</given-names>
</name>
<name>
<surname>Zhao</surname>
<given-names>Z.</given-names>
</name>
<etal/>
</person-group> (<year>2014</year>). <article-title>PDB-wide collection of binding data: current status of the PDBbind database</article-title>. <source>Bioinformatics</source> <volume>31</volume>, <fpage>405</fpage>&#x2013;<lpage>412</lpage>. <pub-id pub-id-type="doi">10.1093/bioinformatics/btu626</pub-id> </citation>
</ref>
<ref id="B30">
<citation citation-type="journal">
<person-group person-group-type="author">
<name>
<surname>Lui</surname>
<given-names>R.</given-names>
</name>
<name>
<surname>Guan</surname>
<given-names>D.</given-names>
</name>
<name>
<surname>Matthews</surname>
<given-names>S.</given-names>
</name>
</person-group> (<year>2020</year>). <article-title>A comparison of molecular representations for lipophilicity quantitative structure-property relationships with results from the SAMPL6 logP Prediction Challenge</article-title>. <source>J.&#x0020;Comput. Aided Mol. Des.</source> <volume>34</volume>, <fpage>523</fpage>&#x2013;<lpage>534</lpage>. <pub-id pub-id-type="doi">10.1007/s10822-020-00279-0</pub-id> </citation>
</ref>
<ref id="B31">
<citation citation-type="journal">
<person-group person-group-type="author">
<name>
<surname>Mannhold</surname>
<given-names>R.</given-names>
</name>
<name>
<surname>Poda</surname>
<given-names>G. I.</given-names>
</name>
<name>
<surname>Ostermann</surname>
<given-names>C.</given-names>
</name>
<name>
<surname>Tetko</surname>
<given-names>I. V.</given-names>
</name>
</person-group> (<year>2009a</year>). <article-title>Calculation of molecular lipophilicity: state-of-the-art and comparison of log P methods on more than 96,000 compounds</article-title>. <source>J.&#x0020;Pharm. Sci.</source> <volume>98</volume>, <fpage>861</fpage>&#x2013;<lpage>893</lpage>. <pub-id pub-id-type="doi">10.1002/jps.21494</pub-id> </citation>
</ref>
<ref id="B32">
<citation citation-type="journal">
<person-group person-group-type="author">
<name>
<surname>Mannhold</surname>
<given-names>R.</given-names>
</name>
<name>
<surname>Poda</surname>
<given-names>G. I.</given-names>
</name>
<name>
<surname>Ostermann</surname>
<given-names>C.</given-names>
</name>
<name>
<surname>Tetko</surname>
<given-names>I. V.</given-names>
</name>
</person-group> (<year>2009b</year>). <article-title>Calculation of molecular lipophilicity: state-of-the-art and comparison of log P methods on more than 96,000 compounds</article-title>. <source>J.&#x0020;Pharm. Sci.</source> <volume>98</volume>, <fpage>861</fpage>&#x2013;<lpage>893</lpage>. <pub-id pub-id-type="doi">10.1002/jps.21494</pub-id> </citation>
</ref>
<ref id="B33">
<citation citation-type="journal">
<person-group person-group-type="author">
<name>
<surname>Mater</surname>
<given-names>A. C.</given-names>
</name>
<name>
<surname>Coote</surname>
<given-names>M. L.</given-names>
</name>
</person-group> (<year>2019</year>). <article-title>Deep learning in chemistry</article-title>. <source>J.&#x0020;Chem. Inf. Model.</source> <volume>59</volume>, <fpage>2545</fpage>&#x2013;<lpage>2559</lpage>. <pub-id pub-id-type="doi">10.1021/acs.jcim.9b00266</pub-id> </citation>
</ref>
<ref id="B34">
<citation citation-type="journal">
<person-group person-group-type="author">
<name>
<surname>Min</surname>
<given-names>S.</given-names>
</name>
<name>
<surname>Lee</surname>
<given-names>B.</given-names>
</name>
<name>
<surname>Yoon</surname>
<given-names>S.</given-names>
</name>
</person-group> (<year>2016</year>). <article-title>Deep learning in bioinformatics</article-title>. <source>Brief Bioinform.</source> <volume>18</volume>, <fpage>851</fpage>&#x2013;<lpage>869</lpage>. <pub-id pub-id-type="doi">10.1093/bib/bbw068</pub-id> </citation>
</ref>
<ref id="B35">
<citation citation-type="journal">
<person-group person-group-type="author">
<name>
<surname>Morrone</surname>
<given-names>J.&#x0020;A.</given-names>
</name>
<name>
<surname>Weber</surname>
<given-names>J.&#x0020;K.</given-names>
</name>
<name>
<surname>Huynh</surname>
<given-names>T.</given-names>
</name>
<name>
<surname>Luo</surname>
<given-names>H.</given-names>
</name>
<name>
<surname>Cornell</surname>
<given-names>W. D.</given-names>
</name>
</person-group> (<year>2020</year>). <article-title>Combining docking pose rank and structure with deep learning improves protein&#x2013;ligand binding mode prediction over a baseline docking approach</article-title>. <source>J.&#x0020;Chem. Inf. Model.</source> <volume>60</volume> (<issue>9</issue>), <fpage>4170</fpage>&#x2013;<lpage>4179</lpage>. <pub-id pub-id-type="doi">10.1021/acs.jcim.9b00927</pub-id> </citation>
</ref>
<ref id="B36">
<citation citation-type="journal">
<person-group person-group-type="author">
<name>
<surname>Myint</surname>
<given-names>K. Z.</given-names>
</name>
<name>
<surname>Wang</surname>
<given-names>L.</given-names>
</name>
<name>
<surname>Tong</surname>
<given-names>Q.</given-names>
</name>
<name>
<surname>Xie</surname>
<given-names>X. Q.</given-names>
</name>
</person-group> (<year>2012</year>). <article-title>Molecular fingerprint-based artificial neural networks QSAR for ligand biological activity predictions</article-title>. <source>Mol. Pharm.</source> <volume>9</volume>, <fpage>2912</fpage>&#x2013;<lpage>2923</lpage>. <pub-id pub-id-type="doi">10.1021/mp300237z</pub-id> </citation>
</ref>
<ref id="B37">
<citation citation-type="journal">
<person-group person-group-type="author">
<name>
<surname>Nisius</surname>
<given-names>B.</given-names>
</name>
<name>
<surname>Bajorath</surname>
<given-names>J.</given-names>
</name>
</person-group> (<year>2009</year>). <article-title>Molecular fingerprint recombination: generating hybrid fingerprints for similarity searching from different fingerprint types</article-title>. <source>ChemMedChem</source> <volume>4</volume>, <fpage>1859</fpage>&#x2013;<lpage>1863</lpage>. <pub-id pub-id-type="doi">10.1002/cmdc.200900243</pub-id> </citation>
</ref>
<ref id="B38">
<citation citation-type="journal">
<person-group person-group-type="author">
<name>
<surname>&#x00d6;zt&#x00fc;rk</surname>
<given-names>H.</given-names>
</name>
<name>
<surname>&#x00d6;zg&#x00fc;r</surname>
<given-names>A.</given-names>
</name>
<name>
<surname>Ozkirimli</surname>
<given-names>E.</given-names>
</name>
</person-group> (<year>2018</year>). <article-title>DeepDTA: deep drug&#x2013;target binding affinity prediction</article-title>. <source>Bioinformatics</source> <volume>34</volume>, <fpage>i821</fpage>&#x2013;<lpage>i829</lpage>. <pub-id pub-id-type="doi">10.1093/bioinformatics/bty593</pub-id> </citation>
</ref>
<ref id="B39">
<citation citation-type="journal">
<person-group person-group-type="author">
<name>
<surname>P&#x00e9;rez-Castillo</surname>
<given-names>Y.</given-names>
</name>
<name>
<surname>Lazar</surname>
<given-names>C.</given-names>
</name>
<name>
<surname>Taminau</surname>
<given-names>J.</given-names>
</name>
<name>
<surname>Froeyen</surname>
<given-names>M.</given-names>
</name>
<name>
<surname>Cabrera-P&#x00e9;rez</surname>
<given-names>M. &#x00c1;.</given-names>
</name>
<name>
<surname>Now&#x00e9;</surname>
<given-names>A.</given-names>
</name>
</person-group> (<year>2012</year>). <article-title>GA(M)E-QSAR: a novel, fully automatic genetic-algorithm-(meta)-ensembles approach for binary classification in ligand-based drug design</article-title>. <source>J.&#x0020;Chem. Inf. Model.</source> <volume>52</volume>, <fpage>2366</fpage>&#x2013;<lpage>2386</lpage>. <pub-id pub-id-type="doi">10.1021/ci300146h</pub-id> </citation>
</ref>
<ref id="B40">
<citation citation-type="journal">
<person-group person-group-type="author">
<name>
<surname>Prasad</surname>
<given-names>S.</given-names>
</name>
<name>
<surname>Brooks</surname>
<given-names>B. R.</given-names>
</name>
</person-group> (<year>2020</year>). <article-title>A deep learning approach for the blind logP prediction in SAMPL6 challenge</article-title>. <source>J.&#x0020;Comput. Aided Mol. Des.</source> <volume>34</volume>, <fpage>535</fpage>&#x2013;<lpage>542</lpage>. <pub-id pub-id-type="doi">10.1007/s10822-020-00292-3</pub-id> </citation>
</ref>
<ref id="B41">
<citation citation-type="journal">
<person-group person-group-type="author">
<name>
<surname>Ramsundar</surname>
<given-names>B.</given-names>
</name>
<name>
<surname>Liu</surname>
<given-names>B.</given-names>
</name>
<name>
<surname>Wu</surname>
<given-names>Z.</given-names>
</name>
<name>
<surname>Verras</surname>
<given-names>A.</given-names>
</name>
<name>
<surname>Tudor</surname>
<given-names>M.</given-names>
</name>
<name>
<surname>Sheridan</surname>
<given-names>R. P.</given-names>
</name>
<etal/>
</person-group> (<year>2017</year>). <article-title>Is multitask deep learning practical for pharma?</article-title>. <source>J.&#x0020;Chem. Inf. Model.</source> <volume>57</volume>, <fpage>2068</fpage>&#x2013;<lpage>2076</lpage>. <pub-id pub-id-type="doi">10.1021/acs.jcim.7b00146</pub-id> </citation>
</ref>
<ref id="B42">
<citation citation-type="web">
<collab>RDKit (</collab>
<year>2017</year>). <article-title>Open-source cheminformatics</article-title>. <comment>Available: <ext-link ext-link-type="uri" xlink:href="http://www.rdkit.org">http://www.rdkit.org</ext-link>
</comment> (<comment>Accessed</comment> <month>September</month> <day>09</day>, <year>2017</year>). </citation>
</ref>
<ref id="B43">
<citation citation-type="journal">
<person-group person-group-type="author">
<name>
<surname>Riniker</surname>
<given-names>S.</given-names>
</name>
<name>
<surname>Landrum</surname>
<given-names>G. A.</given-names>
</name>
</person-group> (<year>2013</year>). <article-title>Open-source platform to benchmark fingerprints for ligand-based virtual screening</article-title>. <source>J.&#x0020;Cheminf.</source> <volume>5</volume>, <fpage>26</fpage>. <pub-id pub-id-type="doi">10.1186/1758-2946-5-26</pub-id> </citation>
</ref>
<ref id="B44">
<citation citation-type="journal">
<person-group person-group-type="author">
<name>
<surname>Riniker</surname>
<given-names>S.</given-names>
</name>
</person-group> (<year>2017</year>). <article-title>Molecular dynamics fingerprints (MDFP): machine learning from MD data to predict free-energy differences</article-title>. <source>J.&#x0020;Chem. Inf. Model.</source> <volume>57</volume>, <fpage>726</fpage>&#x2013;<lpage>741</lpage>. <pub-id pub-id-type="doi">10.1021/acs.jcim.6b00778</pub-id> </citation>
</ref>
<ref id="B45">
<citation citation-type="journal">
<person-group person-group-type="author">
<name>
<surname>Rogers</surname>
<given-names>D.</given-names>
</name>
<name>
<surname>Hahn</surname>
<given-names>M.</given-names>
</name>
</person-group> (<year>2010a</year>). <article-title>Extended-connectivity fingerprints</article-title>. <source>J.&#x0020;Chem. Inf. Model.</source> <volume>50</volume>, <fpage>742</fpage>. <pub-id pub-id-type="doi">10.1021/ci100050t</pub-id> </citation>
</ref>
<ref id="B46">
<citation citation-type="journal">
<person-group person-group-type="author">
<name>
<surname>Rogers</surname>
<given-names>D.</given-names>
</name>
<name>
<surname>Hahn</surname>
<given-names>M.</given-names>
</name>
</person-group> (<year>2010b</year>). <article-title>Extended-connectivity fingerprints</article-title>. <source>J.&#x0020;Chem. Inf. Model.</source> <volume>50</volume>, <fpage>742</fpage>&#x2013;<lpage>754</lpage>. <pub-id pub-id-type="doi">10.1021/ci100050t</pub-id> </citation>
</ref>
<ref id="B47">
<citation citation-type="journal">
<person-group person-group-type="author">
<name>
<surname>Sanchez-Lengeling</surname>
<given-names>B.</given-names>
</name>
<name>
<surname>Aspuru-Guzik</surname>
<given-names>A.</given-names>
</name>
</person-group> (<year>2018</year>). <article-title>Inverse molecular design using machine learning: generative models for matter engineering</article-title>. <source>Science</source> <volume>361</volume>, <fpage>360</fpage>. <pub-id pub-id-type="doi">10.1126/science.aat2663</pub-id> </citation>
</ref>
<ref id="B48">
<citation citation-type="journal">
<person-group person-group-type="author">
<name>
<surname>Sandfort</surname>
<given-names>F.</given-names>
</name>
<name>
<surname>Strieth-Kalthoff</surname>
<given-names>F.</given-names>
</name>
<name>
<surname>K&#x00fc;hnemund</surname>
<given-names>M.</given-names>
</name>
<name>
<surname>Beecks</surname>
<given-names>C.</given-names>
</name>
<name>
<surname>Glorius</surname>
<given-names>F.</given-names>
</name>
</person-group> (<year>2020</year>). <article-title>A structure-based platform for predicting chemical reactivity</article-title>. <source>Inside Chem.</source> <volume>6</volume>, <fpage>1379</fpage>&#x2013;<lpage>1390</lpage>. <pub-id pub-id-type="doi">10.1016/j.chempr.2020.02.017</pub-id> </citation>
</ref>
<ref id="B49">
<citation citation-type="journal">
<person-group person-group-type="author">
<name>
<surname>Schneider</surname>
<given-names>G.</given-names>
</name>
</person-group> (<year>2010</year>). <article-title>Virtual screening: an endless staircase?</article-title>. <source>Nat. Rev. Drug Discov.</source> <volume>9</volume>, <fpage>273</fpage>&#x2013;<lpage>276</lpage>. <pub-id pub-id-type="doi">10.1038/nrd3139</pub-id> </citation>
</ref>
<ref id="B50">
<citation citation-type="journal">
<person-group person-group-type="author">
<name>
<surname>Senese</surname>
<given-names>C. L.</given-names>
</name>
<name>
<surname>Duca</surname>
<given-names>J.</given-names>
</name>
<name>
<surname>Pan</surname>
<given-names>D.</given-names>
</name>
<name>
<surname>Hopfinger</surname>
<given-names>A. J.</given-names>
</name>
<name>
<surname>Tseng</surname>
<given-names>Y. J.</given-names>
</name>
</person-group> (<year>2004</year>). <article-title>4D-Fingerprints, universal QSAR and QSPR descriptors</article-title>. <source>J.&#x0020;Chem. Inf. Comput. Sci.</source> <volume>44</volume>, <fpage>1526</fpage>&#x2013;<lpage>1539</lpage>. <pub-id pub-id-type="doi">10.1021/ci049898s</pub-id> </citation>
</ref>
<ref id="B51">
<citation citation-type="journal">
<person-group person-group-type="author">
<name>
<surname>Senior</surname>
<given-names>A. W.</given-names>
</name>
<name>
<surname>Evans</surname>
<given-names>R.</given-names>
</name>
<name>
<surname>Jumper</surname>
<given-names>J.</given-names>
</name>
<name>
<surname>Kirkpatrick</surname>
<given-names>J.</given-names>
</name>
<name>
<surname>Sifre</surname>
<given-names>L.</given-names>
</name>
<name>
<surname>Green</surname>
<given-names>T.</given-names>
</name>
<etal/>
</person-group> (<year>2020</year>). <article-title>Improved protein structure prediction using potentials from deep learning</article-title>. <source>Nature</source> <volume>577</volume>, <fpage>706</fpage>&#x2013;<lpage>710</lpage>. <pub-id pub-id-type="doi">10.1038/s41586-019-1923-7</pub-id> </citation>
</ref>
<ref id="B52">
<citation citation-type="journal">
<person-group person-group-type="author">
<name>
<surname>Shen</surname>
<given-names>C.</given-names>
</name>
<name>
<surname>Ding</surname>
<given-names>J.</given-names>
</name>
<name>
<surname>Wang</surname>
<given-names>Z.</given-names>
</name>
<name>
<surname>Cao</surname>
<given-names>D.</given-names>
</name>
<name>
<surname>Ding</surname>
<given-names>X.</given-names>
</name>
<name>
<surname>Hou</surname>
<given-names>T.</given-names>
</name>
</person-group> (<year>2020</year>). <article-title>From machine learning to deep learning: advances in scoring functions for protein&#x2013;ligand docking</article-title>. <source>WIREs Comput. Mol. Sci.</source> <volume>10</volume>, <fpage>e1429</fpage>. <pub-id pub-id-type="doi">10.1002/wcms.1429</pub-id> </citation>
</ref>
<ref id="B53">
<citation citation-type="journal">
<person-group person-group-type="author">
<name>
<surname>Sheridan</surname>
<given-names>R. P.</given-names>
</name>
<name>
<surname>Wang</surname>
<given-names>W. M.</given-names>
</name>
<name>
<surname>Liaw</surname>
<given-names>A.</given-names>
</name>
<name>
<surname>Ma</surname>
<given-names>J.</given-names>
</name>
<name>
<surname>Gifford</surname>
<given-names>E. M.</given-names>
</name>
</person-group> (<year>2016</year>). <article-title>Extreme gradient boosting as a method for quantitative structure-activity relationships</article-title>. <source>J.&#x0020;Chem. Inf. Model.</source> <volume>56</volume>, <fpage>2353</fpage>&#x2013;<lpage>2360</lpage>. <pub-id pub-id-type="doi">10.1021/acs.jcim.6b00591</pub-id> </citation>
</ref>
<ref id="B54">
<citation citation-type="journal">
<person-group person-group-type="author">
<name>
<surname>Sheridan</surname>
<given-names>R. P.</given-names>
</name>
<name>
<surname>Miller</surname>
<given-names>M. D.</given-names>
</name>
<name>
<surname>Underwood</surname>
<given-names>D. J.</given-names>
</name>
<name>
<surname>Kearsley</surname>
<given-names>S. K.</given-names>
</name>
</person-group> (<year>1996</year>). <article-title>Chemical similarity using geometric atom pair descriptors</article-title>. <source>J.&#x0020;Chem. Inf. Comput. Sci.</source> <volume>36</volume>, <fpage>128</fpage>&#x2013;<lpage>136</lpage>. <pub-id pub-id-type="doi">10.1021/ci950275b</pub-id> </citation>
</ref>
<ref id="B55">
<citation citation-type="journal">
<person-group person-group-type="author">
<name>
<surname>Solorio-Fern&#x00e1;ndez</surname>
<given-names>S.</given-names>
</name>
<name>
<surname>Carrasco-Ochoa</surname>
<given-names>J.&#x0020;A.</given-names>
</name>
<name>
<surname>Mart&#x00ed;nez-Trinidad</surname>
<given-names>J.&#x0020;F.</given-names>
</name>
</person-group> (<year>2020</year>). <article-title>A review of unsupervised feature selection methods</article-title>. <source>Artif. Intell. Rev.</source> <volume>53</volume>, <fpage>907</fpage>&#x2013;<lpage>948</lpage>. <pub-id pub-id-type="doi">10.1007/s10462-019-09682-y</pub-id> </citation>
</ref>
<ref id="B56">
<citation citation-type="journal">
<person-group person-group-type="author">
<name>
<surname>Su</surname>
<given-names>B. H.</given-names>
</name>
<name>
<surname>Tu</surname>
<given-names>Y. S.</given-names>
</name>
<name>
<surname>Esposito</surname>
<given-names>E. X.</given-names>
</name>
<name>
<surname>Tseng</surname>
<given-names>Y. J.</given-names>
</name>
</person-group> (<year>2012</year>). <article-title>Predictive toxicology modeling: protocols for exploring hERG classification and Tetrahymena pyriformis end point predictions</article-title>. <source>J.&#x0020;Chem. Inf. Model.</source> <volume>52</volume>, <fpage>1660</fpage>&#x2013;<lpage>1673</lpage>. <pub-id pub-id-type="doi">10.1021/ci300060b</pub-id> </citation>
</ref>
<ref id="B57">
<citation citation-type="journal">
<person-group person-group-type="author">
<name>
<surname>Taherkhani</surname>
<given-names>A.</given-names>
</name>
<name>
<surname>Cosma</surname>
<given-names>G.</given-names>
</name>
<name>
<surname>Mcginnity</surname>
<given-names>T. M.</given-names>
</name>
</person-group> (<year>2018</year>). <article-title>Deep-FS: a feature selection algorithm for deep boltzmann machines</article-title>. <source>Neurocomputing</source> <volume>322</volume>, <fpage>22</fpage>&#x2013;<lpage>37</lpage>. <pub-id pub-id-type="doi">10.1016/j.neucom.2018.09.040</pub-id> </citation>
</ref>
<ref id="B58">
<citation citation-type="journal">
<person-group person-group-type="author">
<name>
<surname>Tovar</surname>
<given-names>A.</given-names>
</name>
<name>
<surname>Eckert</surname>
<given-names>H.</given-names>
</name>
<name>
<surname>Bajorath</surname>
<given-names>J.</given-names>
</name>
</person-group> (<year>2007</year>). <article-title>Comparison of 2D fingerprint methods for multiple-template similarity searching on compound activity classes of increasing structural diversity</article-title>. <source>ChemMedChem</source> <volume>2</volume>, <fpage>208</fpage>&#x2013;<lpage>217</lpage>. <pub-id pub-id-type="doi">10.1002/cmdc.200600225</pub-id> </citation>
</ref>
<ref id="B59">
<citation citation-type="journal">
<person-group person-group-type="author">
<name>
<surname>Tseng</surname>
<given-names>Y. J.</given-names>
</name>
<name>
<surname>Hopfinger</surname>
<given-names>A. J.</given-names>
</name>
<name>
<surname>Esposito</surname>
<given-names>E. X.</given-names>
</name>
</person-group> (<year>2012</year>). <article-title>The great descriptor melting pot: mixing descriptors for the common good of QSAR models</article-title>. <source>J.&#x0020;Comput. Aided Mol. Des.</source> <volume>26</volume>, <fpage>39</fpage>&#x2013;<lpage>43</lpage>. <pub-id pub-id-type="doi">10.1007/s10822-011-9511-4</pub-id> </citation>
</ref>
<ref id="B60">
<citation citation-type="journal">
<person-group person-group-type="author">
<name>
<surname>Vilar</surname>
<given-names>S.</given-names>
</name>
<name>
<surname>Uriarte</surname>
<given-names>E.</given-names>
</name>
<name>
<surname>Santana</surname>
<given-names>L.</given-names>
</name>
<name>
<surname>Lorberbaum</surname>
<given-names>T.</given-names>
</name>
<name>
<surname>Hripcsak</surname>
<given-names>G.</given-names>
</name>
<name>
<surname>Friedman</surname>
<given-names>C.</given-names>
</name>
<etal/>
</person-group> (<year>2014</year>). <article-title>Similarity-based modeling in large-scale prediction of drug-drug interactions</article-title>. <source>Nat. Protoc.</source> <volume>9</volume>, <fpage>2147</fpage>&#x2013;<lpage>2163</lpage>. <pub-id pub-id-type="doi">10.1038/nprot.2014.151</pub-id> </citation>
</ref>
<ref id="B61">
<citation citation-type="journal">
<person-group person-group-type="author">
<name>
<surname>Wang</surname>
<given-names>L.</given-names>
</name>
<name>
<surname>Wang</surname>
<given-names>Y.</given-names>
</name>
<name>
<surname>Chang</surname>
<given-names>Q.</given-names>
</name>
</person-group>. (<year>2016</year>). <article-title>Feature selection methods for big data bioinformatics: a survey from the search perspective</article-title>. <source>Methods</source> <volume>111</volume>, <fpage>21</fpage>&#x2013;<lpage>31</lpage>. <pub-id pub-id-type="doi">10.1016/j.ymeth.2016.08.014</pub-id> </citation>
</ref>
<ref id="B62">
<citation citation-type="journal">
<person-group person-group-type="author">
<name>
<surname>Wang</surname>
<given-names>Y.</given-names>
</name>
<name>
<surname>Bajorath</surname>
<given-names>J.</given-names>
</name>
</person-group> (<year>2008</year>). <article-title>Bit silencing in fingerprints enables the derivation of compound class-directed similarity metrics</article-title>. <source>J.&#x0020;Chem. Inf. Model.</source> <volume>48</volume>, <fpage>1754</fpage>&#x2013;<lpage>1759</lpage>. <pub-id pub-id-type="doi">10.1021/ci8002045</pub-id> </citation>
</ref>
<ref id="B63">
<citation citation-type="journal">
<person-group person-group-type="author">
<name>
<surname>Wang</surname>
<given-names>Z.</given-names>
</name>
<name>
<surname>Sun</surname>
<given-names>H.</given-names>
</name>
<name>
<surname>Shen</surname>
<given-names>C.</given-names>
</name>
<name>
<surname>Hu</surname>
<given-names>X.</given-names>
</name>
<name>
<surname>Gao</surname>
<given-names>J.</given-names>
</name>
<name>
<surname>Li</surname>
<given-names>D.</given-names>
</name>
<etal/>
</person-group> (<year>2020</year>). <article-title>Combined strategies in structure-based virtual screening</article-title>. <source>Phys. Chem. Chem. Phys.</source> <volume>22</volume>, <fpage>3149</fpage>&#x2013;<lpage>3159</lpage>. <pub-id pub-id-type="doi">10.1039/c9cp06303j</pub-id> </citation>
</ref>
<ref id="B64">
<citation citation-type="journal">
<person-group person-group-type="author">
<name>
<surname>Weininger</surname>
<given-names>D.</given-names>
</name>
</person-group> (<year>1988</year>). <article-title>SMILES, a chemical language and information system. 1. Introduction to methodology and encoding rules</article-title>. <source>J.&#x0020;Chem. Inf. Comput. Sci.</source> <volume>28</volume>, <fpage>31</fpage>&#x2013;<lpage>36</lpage>. <pub-id pub-id-type="doi">10.1021/ci00057a005</pub-id> </citation>
</ref>
<ref id="B65">
<citation citation-type="journal">
<person-group person-group-type="author">
<name>
<surname>Weininger</surname>
<given-names>D.</given-names>
</name>
</person-group> (<year>1990</year>). <article-title>SMILES. 3. DEPICT. Graphical depiction of chemical structures</article-title>. <source>J.&#x0020;Chem. Inf. Comput. Sci.</source> <volume>30</volume>, <fpage>237</fpage>&#x2013;<lpage>243</lpage>. <pub-id pub-id-type="doi">10.1021/ci00067a005</pub-id> </citation>
</ref>
<ref id="B66">
<citation citation-type="journal">
<person-group person-group-type="author">
<name>
<surname>Weininger</surname>
<given-names>D.</given-names>
</name>
<name>
<surname>Weininger</surname>
<given-names>A.</given-names>
</name>
<name>
<surname>Weininger</surname>
<given-names>J.&#x0020;L.</given-names>
</name>
</person-group> (<year>1989</year>). <article-title>SMILES. 2. Algorithm for generation of unique SMILES notation</article-title>. <source>J.&#x0020;Chem. Inf. Comput. Sci.</source> <volume>29</volume>, <fpage>97</fpage>&#x2013;<lpage>101</lpage>. <pub-id pub-id-type="doi">10.1021/ci00062a008</pub-id> </citation>
</ref>
<ref id="B67">
<citation citation-type="journal">
<person-group person-group-type="author">
<name>
<surname>Wildman</surname>
<given-names>S. A.</given-names>
</name>
<name>
<surname>Crippen</surname>
<given-names>G. M.</given-names>
</name>
</person-group> (<year>1999</year>). <article-title>Prediction of physicochemical parameters by atomic contributions</article-title>. <source>J.&#x0020;Chem. Inf. Comput. Sci.</source> <volume>39</volume>, <fpage>868</fpage>&#x2013;<lpage>873</lpage>. <pub-id pub-id-type="doi">10.1021/ci990307l</pub-id> </citation>
</ref>
<ref id="B68">
<citation citation-type="journal">
<person-group person-group-type="author">
<name>
<surname>Winter</surname>
<given-names>R.</given-names>
</name>
<name>
<surname>Montanari</surname>
<given-names>F.</given-names>
</name>
<name>
<surname>No&#x00e9;</surname>
<given-names>F.</given-names>
</name>
<name>
<surname>Clevert</surname>
<given-names>D. A.</given-names>
</name>
</person-group> (<year>2019</year>). <article-title>Learning continuous and data-driven molecular descriptors by translating equivalent chemical representations</article-title>. <source>Chem. Sci.</source> <volume>10</volume>, <fpage>1692</fpage>&#x2013;<lpage>1701</lpage>. <pub-id pub-id-type="doi">10.1039/c8sc04175j</pub-id> </citation>
</ref>
<ref id="B69">
<citation citation-type="journal">
<person-group person-group-type="author">
<name>
<surname>Wishart</surname>
<given-names>D. S.</given-names>
</name>
<name>
<surname>Feunang</surname>
<given-names>Y. D.</given-names>
</name>
<name>
<surname>Guo</surname>
<given-names>A. C.</given-names>
</name>
<name>
<surname>Lo</surname>
<given-names>E. J.</given-names>
</name>
<name>
<surname>Marcu</surname>
<given-names>A.</given-names>
</name>
<name>
<surname>Grant</surname>
<given-names>J.&#x0020;R.</given-names>
</name>
<etal/>
</person-group> (<year>2018</year>). <article-title>DrugBank 5.0: a major update to the DrugBank database for 2018</article-title>. <source>Nucleic Acids Res.</source> <volume>46</volume>, <fpage>D1074</fpage>&#x2013;<lpage>D1082</lpage>. <pub-id pub-id-type="doi">10.1093/nar/gkx1037</pub-id> </citation>
</ref>
<ref id="B70">
<citation citation-type="journal">
<person-group person-group-type="author">
<name>
<surname>Wu</surname>
<given-names>Z.</given-names>
</name>
<name>
<surname>Lei</surname>
<given-names>T.</given-names>
</name>
<name>
<surname>Shen</surname>
<given-names>C.</given-names>
</name>
<name>
<surname>Wang</surname>
<given-names>Z.</given-names>
</name>
<name>
<surname>Cao</surname>
<given-names>D.</given-names>
</name>
<name>
<surname>Hou</surname>
<given-names>T.</given-names>
</name>
</person-group> (<year>2019</year>). <article-title>ADMET evaluation in drug discovery. 19. Reliable prediction of human cytochrome P450 inhibition using artificial intelligence approaches</article-title>. <source>J.&#x0020;Chem. Inf. Model.</source> <volume>59</volume>, <fpage>4587</fpage>&#x2013;<lpage>4601</lpage>. <pub-id pub-id-type="doi">10.1021/acs.jcim.9b00801</pub-id> </citation>
</ref>
<ref id="B71">
<citation citation-type="journal">
<person-group person-group-type="author">
<name>
<surname>Wu</surname>
<given-names>Z.</given-names>
</name>
<name>
<surname>Ramsundar</surname>
<given-names>B.</given-names>
</name>
<name>
<surname>Feinberg</surname>
<given-names>E. N.</given-names>
</name>
<name>
<surname>Gomes</surname>
<given-names>J.</given-names>
</name>
<name>
<surname>Geniesse</surname>
<given-names>C.</given-names>
</name>
<name>
<surname>Pappu</surname>
<given-names>A. S.</given-names>
</name>
<etal/>
</person-group> (<year>2018</year>). <article-title>MoleculeNet: a benchmark for molecular machine learning</article-title>. <source>Chem. Sci.</source> <volume>9</volume>, <fpage>513</fpage>&#x2013;<lpage>530</lpage>. <pub-id pub-id-type="doi">10.1039/c7sc02664a</pub-id> </citation>
</ref>
<ref id="B72">
<citation citation-type="journal">
<person-group person-group-type="author">
<name>
<surname>Xie</surname>
<given-names>L.</given-names>
</name>
<name>
<surname>Xu</surname>
<given-names>L.</given-names>
</name>
<name>
<surname>Chang</surname>
<given-names>S.</given-names>
</name>
<name>
<surname>Xu</surname>
<given-names>X.</given-names>
</name>
<name>
<surname>Meng</surname>
<given-names>L.</given-names>
</name>
</person-group> (<year>2020</year>). <article-title>Multitask deep networks with grid featurization achieve improved scoring performance for protein-ligand binding</article-title>. <source>Chem. Biol. Drug Des.</source> <volume>96</volume>, <fpage>973</fpage>&#x2013;<lpage>983</lpage>. <pub-id pub-id-type="doi">10.1111/cbdd.13648</pub-id> </citation>
</ref>
<ref id="B73">
<citation citation-type="journal">
<person-group person-group-type="author">
<name>
<surname>Xie</surname>
<given-names>T.</given-names>
</name>
<name>
<surname>Grossman</surname>
<given-names>J.&#x0020;C.</given-names>
</name>
</person-group> (<year>2018</year>). <article-title>Crystal graph convolutional neural networks for an accurate and interpretable prediction of material properties</article-title>. <source>Phys. Rev. Lett.</source> <volume>120</volume>, <fpage>145301</fpage>. <pub-id pub-id-type="doi">10.1103/PhysRevLett.120.145301</pub-id> </citation>
</ref>
<ref id="B74">
<citation citation-type="journal">
<person-group person-group-type="author">
<name>
<surname>Yang</surname>
<given-names>X.</given-names>
</name>
<name>
<surname>Wang</surname>
<given-names>Y.</given-names>
</name>
<name>
<surname>Byrne</surname>
<given-names>R.</given-names>
</name>
<name>
<surname>Schneider</surname>
<given-names>G.</given-names>
</name>
<name>
<surname>Yang</surname>
<given-names>S.</given-names>
</name>
</person-group> (<year>2019</year>). <article-title>Concepts of artificial intelligence for computer-assisted drug discovery</article-title>. <source>Chem. Rev.</source> <volume>119</volume>, <fpage>10520</fpage>&#x2013;<lpage>10594</lpage>. <pub-id pub-id-type="doi">10.1021/acs.chemrev.8b00728</pub-id> </citation>
</ref>
<ref id="B75">
<citation citation-type="book">
<person-group person-group-type="author">
<name>
<surname>Zeiler</surname>
<given-names>M. D.</given-names>
</name>
<name>
<surname>Fergus</surname>
<given-names>R.</given-names>
</name>
</person-group> (<year>2014</year>). &#x201c;<article-title>Visualizing and understanding convolutional networks</article-title>,&#x201d; in <source>Computer vision &#x2013; ECCV 2014</source>. Editors <person-group person-group-type="editor">
<name>
<surname>Fleet</surname>
<given-names>D.</given-names>
</name>
<name>
<surname>Pajdla</surname>
<given-names>T.</given-names>
</name>
<name>
<surname>Schiele</surname>
<given-names>B.</given-names>
</name>
<name>
<surname>Tuytelaars</surname>
<given-names>T.</given-names>
</name>
</person-group> (<publisher-name>Springer International Publishing</publisher-name>), <fpage>818</fpage>&#x2013;<lpage>833</lpage>. </citation>
</ref>
</ref-list>
</back>
</article>
