<?xml version="1.0" encoding="UTF-8"?>
<!DOCTYPE article PUBLIC "-//NLM//DTD Journal Archiving and Interchange DTD v2.3 20070202//EN" "archivearticle.dtd">
<article article-type="methods-article" dtd-version="2.3" xml:lang="EN" xmlns:mml="http://www.w3.org/1998/Math/MathML" xmlns:xlink="http://www.w3.org/1999/xlink">
<front>
<journal-meta>
<journal-id journal-id-type="publisher-id">Front. Genet.</journal-id>
<journal-title>Frontiers in Genetics</journal-title>
<abbrev-journal-title abbrev-type="pubmed">Front. Genet.</abbrev-journal-title>
<issn pub-type="epub">1664-8021</issn>
<publisher>
<publisher-name>Frontiers Media S.A.</publisher-name>
</publisher>
</journal-meta>
<article-meta>
<article-id pub-id-type="publisher-id">834488</article-id>
<article-id pub-id-type="doi">10.3389/fgene.2021.834488</article-id>
<article-categories>
<subj-group subj-group-type="heading">
<subject>Genetics</subject>
<subj-group>
<subject>Methods</subject>
</subj-group>
</subj-group>
</article-categories>
<title-group>
<article-title>Identifying Membrane Protein Types Based on Lifelong Learning With Dynamically Scalable Networks</article-title>
<alt-title alt-title-type="left-running-head">Lu et&#x20;al.</alt-title>
<alt-title alt-title-type="right-running-head">Identifying Membrane Protein Types</alt-title>
</title-group>
<contrib-group>
<contrib contrib-type="author">
<name>
<surname>Lu</surname>
<given-names>Weizhong</given-names>
</name>
<xref ref-type="aff" rid="aff1">
<sup>1</sup>
</xref>
<xref ref-type="aff" rid="aff2">
<sup>2</sup>
</xref>
<xref ref-type="aff" rid="aff3">
<sup>3</sup>
</xref>
</contrib>
<contrib contrib-type="author">
<name>
<surname>Shen</surname>
<given-names>Jiawei</given-names>
</name>
<xref ref-type="aff" rid="aff1">
<sup>1</sup>
</xref>
<uri xlink:href="https://loop.frontiersin.org/people/1591463/overview"/>
</contrib>
<contrib contrib-type="author">
<name>
<surname>Zhang</surname>
<given-names>Yu</given-names>
</name>
<xref ref-type="aff" rid="aff4">
<sup>4</sup>
</xref>
</contrib>
<contrib contrib-type="author" corresp="yes">
<name>
<surname>Wu</surname>
<given-names>Hongjie</given-names>
</name>
<xref ref-type="aff" rid="aff1">
<sup>1</sup>
</xref>
<xref ref-type="aff" rid="aff2">
<sup>2</sup>
</xref>
<xref ref-type="corresp" rid="c001">&#x2a;</xref>
</contrib>
<contrib contrib-type="author">
<name>
<surname>Qian</surname>
<given-names>Yuqing</given-names>
</name>
<xref ref-type="aff" rid="aff1">
<sup>1</sup>
</xref>
</contrib>
<contrib contrib-type="author">
<name>
<surname>Chen</surname>
<given-names>Xiaoyi</given-names>
</name>
<xref ref-type="aff" rid="aff1">
<sup>1</sup>
</xref>
</contrib>
<contrib contrib-type="author">
<name>
<surname>Fu</surname>
<given-names>Qiming</given-names>
</name>
<xref ref-type="aff" rid="aff1">
<sup>1</sup>
</xref>
</contrib>
</contrib-group>
<aff id="aff1">
<sup>1</sup>
<institution>School of Electronic and Information Engineering</institution>, <institution>Suzhou University of Science and Technology</institution>, <addr-line>Suzhou</addr-line>, <country>China</country>
</aff>
<aff id="aff2">
<sup>2</sup>
<institution>Suzhou Key Laboratory of Virtual Reality Intelligent Interaction and Application Technology</institution>, <institution>Suzhou University of Science and Technology</institution>, <addr-line>Suzhou</addr-line>, <country>China</country>
</aff>
<aff id="aff3">
<sup>3</sup>
<institution>Provincial Key Laboratory for Computer Information Processing Technology</institution>, <institution>Soochow University</institution>, <addr-line>Suzhou</addr-line>, <country>China</country>
</aff>
<aff id="aff4">
<sup>4</sup>
<institution>Suzhou Industrial Park Institute of Services Outsourcing</institution>, <addr-line>Suzhou</addr-line>, <country>China</country>
</aff>
<author-notes>
<fn fn-type="edited-by">
<p>
<bold>Edited by:</bold> <ext-link ext-link-type="uri" xlink:href="https://loop.frontiersin.org/people/589586/overview">Leyi Wei</ext-link>, Shandong University, China</p>
</fn>
<fn fn-type="edited-by">
<p>
<bold>Reviewed by:</bold> <ext-link ext-link-type="uri" xlink:href="https://loop.frontiersin.org/people/1599121/overview">Chunhua Li</ext-link>, Beijing University of Technology, China</p>
<p>
<ext-link ext-link-type="uri" xlink:href="https://loop.frontiersin.org/people/854549/overview">Jieming Ma</ext-link>, Xi&#x2019;an Jiaotong-Liverpool University, China</p>
</fn>
<corresp id="c001">&#x2a;Correspondence: Hongjie Wu, <email>hongjiewu@usts.edu.cn</email>
</corresp>
<fn fn-type="other">
<p>This article was submitted to Computational Genomics, a section of the journal Frontiers in Genetics</p>
</fn>
</author-notes>
<pub-date pub-type="epub">
<day>14</day>
<month>03</month>
<year>2022</year>
</pub-date>
<pub-date pub-type="collection">
<year>2021</year>
</pub-date>
<volume>12</volume>
<elocation-id>834488</elocation-id>
<history>
<date date-type="received">
<day>13</day>
<month>12</month>
<year>2021</year>
</date>
<date date-type="accepted">
<day>21</day>
<month>12</month>
<year>2021</year>
</date>
</history>
<permissions>
<copyright-statement>Copyright &#xa9; 2022 Lu, Shen, Zhang, Wu, Qian, Chen and Fu.</copyright-statement>
<copyright-year>2022</copyright-year>
<copyright-holder>Lu, Shen, Zhang, Wu, Qian, Chen and Fu</copyright-holder>
<license xlink:href="http://creativecommons.org/licenses/by/4.0/">
<p>This is an open-access article distributed under the terms of the Creative Commons Attribution License (CC BY). The use, distribution or reproduction in other forums is permitted, provided the original author(s) and the copyright owner(s) are credited and that the original publication in this journal is cited, in accordance with accepted academic practice. No use, distribution or reproduction is permitted which does not comply with these&#x20;terms.</p>
</license>
</permissions>
<abstract>
<p>Membrane proteins are an essential part of the body&#x2019;s ability to maintain normal life activities. Further research into membrane proteins, which are present in all aspects of life science research, will help to advance the development of cells and drugs. The current methods for predicting proteins are usually based on machine learning, but further improvements in prediction effectiveness and accuracy are needed. In this paper, we propose a dynamic deep network architecture based on lifelong learning in order to use computers to classify membrane proteins more effectively. The model extends the application area of lifelong learning and provides new ideas for multiple classification problems in bioinformatics. To demonstrate the performance of our model, we conducted experiments on top of two datasets and compared them with other classification methods. The results show that our model achieves high accuracy (95.3 and 93.5%) on benchmark datasets and is more effective compared to other methods.</p>
</abstract>
<kwd-group>
<kwd>lifelong learning</kwd>
<kwd>membrane proteins</kwd>
<kwd>dynamically scalable networks</kwd>
<kwd>position specific scoring matrix</kwd>
<kwd>evolutionary features</kwd>
</kwd-group>
</article-meta>
</front>
<body>
<sec id="s1">
<title>1 Introduction</title>
<p>The biological cell&#x2019;s daily activities are associated with membranes, without which it would not be possible to form a living structure. The essential proteins that make up membranes are the lipids and proteins that are the main components of membranes. In the present biological research, there are eight types of membrane proteins: 1) single-span 1; 2) single-span 2; 3) single-span 3; 4) single-span 4; 5) multi-span; 6) lipid-anchor; 7) GPI-anchor and (8) peripheral (<xref ref-type="bibr" rid="B5">Cedano et&#x20;al., 1997</xref>).</p>
<p>Bioinformatics is present in all aspects of the biological sciences, and how to use computers to classify proteins efficiently and accurately has been a hot research problem in the direction of bioinformatics and computer science. Although traditional physicochemical as well as biological experiments are desirable in terms of predictive accuracy, these methods are too cumbersome and require a great deal of human and material resources. To save time and financial costs, and to better understand the structure and function of membrane proteins, a number of calculations have been developed to efficiently discriminate between protein types (<xref ref-type="bibr" rid="B22">Feng and Zhang, 2000</xref>; <xref ref-type="bibr" rid="B4">Cai et&#x20;al., 2004</xref>; <xref ref-type="bibr" rid="B49">Zou et&#x20;al., 2013</xref>; <xref ref-type="bibr" rid="B44">Wei et&#x20;al., 2017</xref>; <xref ref-type="bibr" rid="B48">Zhou et&#x20;al., 2021</xref>; <xref ref-type="bibr" rid="B51">Zou et&#x20;al., 2020</xref>; <xref ref-type="bibr" rid="B30">Qian et&#x20;al., 2021</xref>; <xref ref-type="bibr" rid="B18">Ding et&#x20;al., 2021</xref>; <xref ref-type="bibr" rid="B21">Ding et&#x20;al., 2021</xref>; <xref ref-type="bibr" rid="B50">Zou et&#x20;al., 2021</xref>). The extant methods are in large part improvements on Chou&#x2019;s algorithm (<xref ref-type="bibr" rid="B11">Chou and Elrod, 1999</xref>). Song et&#x20;al. (<xref ref-type="bibr" rid="B26">Lu et&#x20;al., 2020</xref>) used Chou&#x2019;s 5-step method to extract evolutionary information to input to a support vector machine for protein prediction. Cao and Lu (<xref ref-type="bibr" rid="B27">Lu et&#x20;al., 2021</xref>) avoided loss of information due to truncation by introducing a fag vector and used a variable length dynamic two-way gated cyclic unit model to predict protein. Yang (<xref ref-type="bibr" rid="B45">Wu et&#x20;al., 2019</xref>) designed a reward function to model the protein input under full-state reinforcement learning. Wu and Huang (<xref ref-type="bibr" rid="B46">Wu et&#x20;al., 2019</xref>) et&#x20;al. used random forests to build their own model and used binary reordering to make their predictions more efficient. To avoid the limitations of overfitting, Lu and Tang (<xref ref-type="bibr" rid="B28">Lu et&#x20;al., 2019</xref>) et&#x20;al. used an energy filter to make the sequence length follow the model adaptively.</p>
<p>In most methods of machine learning, predictions are made using fixed models for different kinds of proteins, which generally suffer from two problems. Firstly, they do not allow for incremental learning, and secondly, they do not consider task-to-task connections at the task level. Lifelong learning approaches aim to bridge these two issues. Lifelong machine learning methods were first proposed by Thrun and Mitchell (<xref ref-type="bibr" rid="B41">Thrun and Mitchell, 1995</xref>), who viewed each task as a binary problem to classify all tasks. A number of memory-based and neural network-based approaches to lifelong learning were then proposed and refined by Silver (<xref ref-type="bibr" rid="B40">Silver, 1996</xref>; <xref ref-type="bibr" rid="B39">Silver and Mercer, 2002</xref>; <xref ref-type="bibr" rid="B38">Silver et&#x20;al., 2015</xref>) et&#x20;al. Ruvolo and Eaton (<xref ref-type="bibr" rid="B32">Ruvolo and Eaton, 2013</xref>) proposed the Efficient Lifelong Learning Algorithm (ELLA), which greatly enhanced the algorithm proposed by Kumar (<xref ref-type="bibr" rid="B24">Kumar and Daume, 2012</xref>) et&#x20;al. for multi-task learning (MTL). Ruvolo and Eaton (<xref ref-type="bibr" rid="B31">Ruvolo and Eaton, 2013</xref>) viewed lifelong learning as a real-time task selection process, Chen (<xref ref-type="bibr" rid="B10">Chen et&#x20;al., 2015</xref>) proposed a lifelong learning algorithm based on plain Bayesian classification, and Shu (<xref ref-type="bibr" rid="B37">Shu et&#x20;al., 2017</xref>) et&#x20;al. investigated the direction of lifelong learning by improving the conditional random field model. Mazumder (<xref ref-type="bibr" rid="B9">Chen and Liu, 2014</xref>) investigated human-machine conversational machines and enabled chatbots to learn new knowledge in the process of chatting with humans. Chen, Liu and Wang (<xref ref-type="bibr" rid="B42">Wang et&#x20;al., 2016</xref>) proposed a number of lifelong topic modeling methods to mine topics from historical tasks and apply them to new topic discovery. Shu et&#x20;al. proposed a relaxed labeling approach to solve lifelong unsupervised learning tasks. Chen and Liu (<xref ref-type="bibr" rid="B8">Chen and Liu, 2019</xref>) provide more detailed information on the direction of lifelong machine learning in this&#x20;book.</p>
<p>After a lot of research and careful selection, we ended up using a DSN model based on sequence information and lifetime learning of membrane proteins themselves. First, we processed the membrane protein sequence dataset based on BLAST (<xref ref-type="bibr" rid="B2">Altschul et&#x20;al., 1997</xref>) to obtain the scoring matrix (PSSM) (<xref ref-type="bibr" rid="B17">Dehzangi et&#x20;al., 2017</xref>; <xref ref-type="bibr" rid="B33">Sharma et&#x20;al., 2018</xref>; <xref ref-type="bibr" rid="B47">Yosvany et&#x20;al., 2018</xref>; <xref ref-type="bibr" rid="B6">Chandra et&#x20;al., 2019</xref>). Then, we extracted valid features from the PSSM by the averaging block method (Avblock) (<xref ref-type="bibr" rid="B36">Shen et&#x20;al., 2019</xref>), the discrete wavelet transforms method (DWT) (<xref ref-type="bibr" rid="B35">Shen et&#x20;al., 2017</xref>; <xref ref-type="bibr" rid="B43">Wang et&#x20;al., 2017</xref>), the discrete cosine transforms method (DCT) (<xref ref-type="bibr" rid="B1">Ahmed et&#x20;al., 1974</xref>), the histogram of oriented gradients method (HOG) (<xref ref-type="bibr" rid="B30">Qian et&#x20;al., 2021</xref>)and the Pse-PSSM method. The features extracted by these five methods are then stitched end-to-end and fed into our model for prediction. Finally, the performance is evaluated by random validation tests and independent tests. Through the results we can see that our model achieves good prediction results in the case of predicting membrane protein types. <xref ref-type="fig" rid="F1">Figure&#x20;1</xref> shows a sketch of the main research in this&#x20;paper.</p>
<fig id="F1" position="float">
<label>FIGURE 1</label>
<caption>
<p>The summary of the main research.</p>
</caption>
<graphic xlink:href="fgene-12-834488-g001.tif"/>
</fig>
</sec>
<sec id="s2">
<title>2 Materials and Methods</title>
<p>In this experiment, the sequence of discriminating membrane protein types can be broadly divided into three steps: creating the required model, performing training and testing of the model, and making predictions and conducting analysis of the results. Firstly, the features are extracted from the processed dataset. Next, the features are integrated into a lifelong learning model for prediction. Finally, the sequence information of the membrane proteins is transformed into algorithmic information, which is then analyzed and predicted using the model. <xref ref-type="fig" rid="F2">Figure&#x20;2</xref> shows the research infrastructure for this approach.</p>
<fig id="F2" position="float">
<label>FIGURE 2</label>
<caption>
<p>The research infrastructure for this approach.</p>
</caption>
<graphic xlink:href="fgene-12-834488-g002.tif"/>
</fig>
<sec id="s2-1">
<title>2.1 Analysis of Data Sets</title>
<p>In order to test the performance of our lifelong learning model, we experimentally selected two membrane protein datasets for testing, Data 1 and Data 2. The specifics of the two datasets are shown in <xref ref-type="table" rid="T1">Table&#x20;1</xref>. Data 1 and Data 2 include eight membrane protein types. Data 1 is from the work of Chou (<xref ref-type="bibr" rid="B12">Chou and Shen, 2007</xref>) et&#x20;al. The training and test sets were randomly obtained from Swiss-Prot (<xref ref-type="bibr" rid="B3">Boeckmann et&#x20;al., 2003</xref>) by percentage assignment, which ensured that the quantities of these two sequences are consistent. Data 2 is from the work of Chen (<xref ref-type="bibr" rid="B7">Chen and Li, 2013</xref>) et&#x20;al. where they used the CD-hit (<xref ref-type="bibr" rid="B25">Li and Godzik, 2006</xref>) method to remove redundant sequences from dataset 1 so that no two-by-two sequences would have less than 40% identity.</p>
<table-wrap id="T1" position="float">
<label>TABLE 1</label>
<caption>
<p>The sample sizes for the two different data sets used in this experiment.</p>
</caption>
<table>
<thead valign="top">
<tr>
<th rowspan="2" align="left">The specific type of classification</th>
<th colspan="2" align="center">Dataset 1</th>
<th colspan="2" align="center">Dataset 2</th>
</tr>
<tr>
<th align="center">Train</th>
<th align="center">Test</th>
<th align="center">Train</th>
<th align="center">Test</th>
</tr>
</thead>
<tbody valign="top">
<tr>
<td align="left">Single-span type 1</td>
<td align="center">610</td>
<td align="center">444</td>
<td align="center">388</td>
<td align="center">223</td>
</tr>
<tr>
<td align="left">Single-span type 2</td>
<td align="center">312</td>
<td align="center">78</td>
<td align="center">218</td>
<td align="center">39</td>
</tr>
<tr>
<td align="left">Single-span type 3</td>
<td align="center">24</td>
<td align="center">6</td>
<td align="center">19</td>
<td align="center">6</td>
</tr>
<tr>
<td align="left">Single-span type 4</td>
<td align="center">44</td>
<td align="center">12</td>
<td align="center">35</td>
<td align="center">10</td>
</tr>
<tr>
<td align="left">Multi-span type 5</td>
<td align="center">1,316</td>
<td align="center">3,265</td>
<td align="center">936</td>
<td align="center">1,673</td>
</tr>
<tr>
<td align="left">Lipid-anchor type 6</td>
<td align="center">151</td>
<td align="center">38</td>
<td align="center">98</td>
<td align="center">26</td>
</tr>
<tr>
<td align="left">GPI-anchor type 7</td>
<td align="center">182</td>
<td align="center">46</td>
<td align="center">122</td>
<td align="center">24</td>
</tr>
<tr>
<td align="left">Peripheral type 8</td>
<td align="center">610</td>
<td align="center">444</td>
<td align="center">472</td>
<td align="center">305</td>
</tr>
<tr>
<td align="left">Totality</td>
<td align="center">3,249</td>
<td align="center">4,333</td>
<td align="center">2,288</td>
<td align="center">2,306</td>
</tr>
</tbody>
</table>
</table-wrap>
</sec>
<sec id="s2-2">
<title>2.2 Extracting the Message of Evolutionary Conservatism</title>
<p>The PSSM used in this experiment is the &#x201c;Position-Specific Scoring Matrix.&#x201d; This scoring matrix stores the sequence information of membrane proteins. We use the PSSM matrix (<xref ref-type="bibr" rid="B34">Shen et&#x20;al., 2017</xref>; <xref ref-type="bibr" rid="B19">Ding et&#x20;al., 2017</xref>; <xref ref-type="bibr" rid="B36">Shen et&#x20;al., 2019</xref>) for membrane protein prediction because it reflects the evolutionary information of membrane proteins very well. For any membrane protein sequence, such as Q, the PSSM can be derived by PSI-BLAST (<xref ref-type="bibr" rid="B2">Altschul et&#x20;al., 1997</xref>), after several iterations. First it forms the PSSM based on the first search result, then it performs the next step which is the second search based on the first search result, then it continues with the second search result for another time and repeats this process until the target is searched for the best result. As the performance of the experimental results is best after three iterations, we generally adopt three iterations as the setting. The value of its E is 0.001. Assume that the sequence Q &#x3d; <inline-formula id="inf1">
<mml:math id="m1">
<mml:mrow>
<mml:msub>
<mml:mi>q</mml:mi>
<mml:mn>1</mml:mn>
</mml:msub>
<mml:msub>
<mml:mi>q</mml:mi>
<mml:mn>2</mml:mn>
</mml:msub>
<mml:msub>
<mml:mi>q</mml:mi>
<mml:mn>3</mml:mn>
</mml:msub>
<mml:mn>...</mml:mn>
<mml:mtext>&#xa0;</mml:mtext>
<mml:mo>.</mml:mo>
<mml:mtext>&#xa0;</mml:mtext>
<mml:msub>
<mml:mi>q</mml:mi>
<mml:mi>L</mml:mi>
</mml:msub>
</mml:mrow>
</mml:math>
</inline-formula>, whose length is L. This is followed by storing the PSSM containing the membrane protein evolution information inside a matrix with a size-area of L&#x20;<bold>&#xd7;</bold> 20. The matrix is represented as follows:<disp-formula id="e1">
<mml:math id="m2">
<mml:mrow>
<mml:mi mathvariant="bold-italic">PSS</mml:mi>
<mml:msub>
<mml:mi mathvariant="bold-italic">M</mml:mi>
<mml:mrow>
<mml:mi mathvariant="bold-italic">original</mml:mi>
</mml:mrow>
</mml:msub>
<mml:mi mathvariant="bold-italic">&#x3d;</mml:mi>
<mml:msub>
<mml:mrow>
<mml:mrow>
<mml:mo>[</mml:mo>
<mml:mrow>
<mml:mtable>
<mml:mtr>
<mml:mtd>
<mml:mrow>
<mml:mtable>
<mml:mtr>
<mml:mtd>
<mml:mrow>
<mml:msub>
<mml:mi mathvariant="bold">p</mml:mi>
<mml:mrow>
<mml:mi mathvariant="bold">1,1</mml:mi>
</mml:mrow>
</mml:msub>
</mml:mrow>
</mml:mtd>
<mml:mtd>
<mml:mrow>
<mml:msub>
<mml:mi mathvariant="bold">p</mml:mi>
<mml:mrow>
<mml:mi mathvariant="bold">1,2</mml:mi>
</mml:mrow>
</mml:msub>
</mml:mrow>
</mml:mtd>
</mml:mtr>
<mml:mtr>
<mml:mtd>
<mml:mrow>
<mml:msub>
<mml:mi mathvariant="bold">p</mml:mi>
<mml:mrow>
<mml:mi mathvariant="bold">2,1</mml:mi>
</mml:mrow>
</mml:msub>
</mml:mrow>
</mml:mtd>
<mml:mtd>
<mml:mrow>
<mml:msub>
<mml:mi mathvariant="bold">p</mml:mi>
<mml:mrow>
<mml:mi mathvariant="bold">2,2</mml:mi>
</mml:mrow>
</mml:msub>
</mml:mrow>
</mml:mtd>
</mml:mtr>
<mml:mtr>
<mml:mtd>
<mml:mi mathvariant="bold-italic">&#x22EE;</mml:mi>
</mml:mtd>
<mml:mtd>
<mml:mi mathvariant="bold-italic">&#x22EE;</mml:mi>
</mml:mtd>
</mml:mtr>
</mml:mtable>
</mml:mrow>
</mml:mtd>
<mml:mtd>
<mml:mrow>
<mml:mtable>
<mml:mtr>
<mml:mtd>
<mml:mi mathvariant="bold-italic">&#x2026;</mml:mi>
</mml:mtd>
<mml:mtd>
<mml:mrow>
<mml:msub>
<mml:mi mathvariant="bold">p</mml:mi>
<mml:mrow>
<mml:mi mathvariant="bold">1,20</mml:mi>
</mml:mrow>
</mml:msub>
</mml:mrow>
</mml:mtd>
</mml:mtr>
<mml:mtr>
<mml:mtd>
<mml:mi mathvariant="bold-italic">&#x2026;</mml:mi>
</mml:mtd>
<mml:mtd>
<mml:mrow>
<mml:msub>
<mml:mi mathvariant="bold">p</mml:mi>
<mml:mrow>
<mml:mi mathvariant="bold">2,20</mml:mi>
</mml:mrow>
</mml:msub>
</mml:mrow>
</mml:mtd>
</mml:mtr>
<mml:mtr>
<mml:mtd>
<mml:mi mathvariant="bold-italic">&#x22F1;</mml:mi>
</mml:mtd>
<mml:mtd>
<mml:mi mathvariant="bold-italic">&#x22EE;</mml:mi>
</mml:mtd>
</mml:mtr>
</mml:mtable>
</mml:mrow>
</mml:mtd>
</mml:mtr>
<mml:mtr>
<mml:mtd>
<mml:mrow>
<mml:mtable>
<mml:mtr>
<mml:mtd>
<mml:mrow>
<mml:msub>
<mml:mi mathvariant="bold">p</mml:mi>
<mml:mrow>
<mml:mi mathvariant="bold-italic">i</mml:mi>
<mml:mo>,</mml:mo>
<mml:mi mathvariant="bold">1</mml:mi>
</mml:mrow>
</mml:msub>
</mml:mrow>
</mml:mtd>
<mml:mtd>
<mml:mrow>
<mml:msub>
<mml:mi mathvariant="bold">p</mml:mi>
<mml:mrow>
<mml:mi mathvariant="bold-italic">i</mml:mi>
<mml:mo>,</mml:mo>
<mml:mi mathvariant="bold">2</mml:mi>
</mml:mrow>
</mml:msub>
</mml:mrow>
</mml:mtd>
</mml:mtr>
<mml:mtr>
<mml:mtd>
<mml:mi mathvariant="bold-italic">&#x22EE;</mml:mi>
</mml:mtd>
<mml:mtd>
<mml:mi mathvariant="bold-italic">&#x22EE;</mml:mi>
</mml:mtd>
</mml:mtr>
<mml:mtr>
<mml:mtd>
<mml:mrow>
<mml:msub>
<mml:mi mathvariant="bold">p</mml:mi>
<mml:mrow>
<mml:mi mathvariant="bold-italic">L</mml:mi>
<mml:mo>,</mml:mo>
<mml:mi mathvariant="bold">1</mml:mi>
</mml:mrow>
</mml:msub>
</mml:mrow>
</mml:mtd>
<mml:mtd>
<mml:mrow>
<mml:msub>
<mml:mi mathvariant="bold">p</mml:mi>
<mml:mrow>
<mml:mi mathvariant="bold-italic">L</mml:mi>
<mml:mo>,</mml:mo>
<mml:mi mathvariant="bold">2</mml:mi>
</mml:mrow>
</mml:msub>
</mml:mrow>
</mml:mtd>
</mml:mtr>
</mml:mtable>
</mml:mrow>
</mml:mtd>
<mml:mtd>
<mml:mrow>
<mml:mtable>
<mml:mtr>
<mml:mtd>
<mml:mi mathvariant="bold-italic">&#x2026;</mml:mi>
</mml:mtd>
<mml:mtd>
<mml:mrow>
<mml:msub>
<mml:mi mathvariant="bold">p</mml:mi>
<mml:mrow>
<mml:mi mathvariant="bold-italic">i</mml:mi>
<mml:mo>,</mml:mo>
<mml:mi mathvariant="bold">20</mml:mi>
</mml:mrow>
</mml:msub>
</mml:mrow>
</mml:mtd>
</mml:mtr>
<mml:mtr>
<mml:mtd>
<mml:mi mathvariant="bold-italic">&#x22F1;</mml:mi>
</mml:mtd>
<mml:mtd>
<mml:mi mathvariant="bold-italic">&#x22EE;</mml:mi>
</mml:mtd>
</mml:mtr>
<mml:mtr>
<mml:mtd>
<mml:mi mathvariant="bold-italic">&#x2026;</mml:mi>
</mml:mtd>
<mml:mtd>
<mml:mrow>
<mml:msub>
<mml:mi mathvariant="bold">p</mml:mi>
<mml:mrow>
<mml:mi mathvariant="bold-italic">L</mml:mi>
<mml:mo>,</mml:mo>
<mml:mi mathvariant="bold">20</mml:mi>
</mml:mrow>
</mml:msub>
</mml:mrow>
</mml:mtd>
</mml:mtr>
</mml:mtable>
</mml:mrow>
</mml:mtd>
</mml:mtr>
</mml:mtable>
</mml:mrow>
<mml:mo>]</mml:mo>
</mml:mrow>
</mml:mrow>
<mml:mrow>
<mml:mi mathvariant="bold-italic">L&#xd7;</mml:mi>
<mml:mi mathvariant="bold">20</mml:mi>
</mml:mrow>
</mml:msub>
</mml:mrow>
</mml:math>
<label>(1)</label>
</disp-formula>
</p>
<p>In addition, the expression below shows the representation of <inline-formula id="inf2">
<mml:math id="m3">
<mml:mrow>
<mml:mtext>&#xa0;</mml:mtext>
<mml:mi>P</mml:mi>
<mml:mi>S</mml:mi>
<mml:mi>S</mml:mi>
<mml:msub>
<mml:mi>M</mml:mi>
<mml:mrow>
<mml:mi>o</mml:mi>
<mml:mi>r</mml:mi>
<mml:mi>i</mml:mi>
<mml:mi>g</mml:mi>
<mml:mi>i</mml:mi>
<mml:mi>n</mml:mi>
<mml:mi>a</mml:mi>
<mml:mi>l</mml:mi>
</mml:mrow>
</mml:msub>
<mml:mrow>
<mml:mo>(</mml:mo>
<mml:mrow>
<mml:mi>i</mml:mi>
<mml:mo>,</mml:mo>
<mml:mi>j</mml:mi>
</mml:mrow>
<mml:mo>)</mml:mo>
</mml:mrow>
</mml:mrow>
</mml:math>
</inline-formula>:<disp-formula id="e2">
<mml:math id="m4">
<mml:mrow>
<mml:mi mathvariant="bold-italic">PSS</mml:mi>
<mml:msub>
<mml:mi mathvariant="bold-italic">M</mml:mi>
<mml:mrow>
<mml:mi mathvariant="bold-italic">original</mml:mi>
</mml:mrow>
</mml:msub>
<mml:mrow>
<mml:mo>(</mml:mo>
<mml:mrow>
<mml:mi mathvariant="bold-italic">i,j</mml:mi>
</mml:mrow>
<mml:mo>)</mml:mo>
</mml:mrow>
<mml:mi mathvariant="bold-italic">&#x3d;</mml:mi>
<mml:munderover>
<mml:mstyle displaystyle="true">
<mml:mi mathvariant="bold-italic">&#x2211;</mml:mi>
</mml:mstyle>
<mml:mrow>
<mml:mi mathvariant="bold-italic">k&#x3d;1</mml:mi>
</mml:mrow>
<mml:mrow>
<mml:mi mathvariant="bold">20</mml:mi>
</mml:mrow>
</mml:munderover>
<mml:mi mathvariant="bold-italic">&#x3c9;</mml:mi>
<mml:mrow>
<mml:mo>(</mml:mo>
<mml:mrow>
<mml:mi mathvariant="bold-italic">i,k</mml:mi>
</mml:mrow>
<mml:mo>)</mml:mo>
</mml:mrow>
<mml:mi mathvariant="bold-italic">&#xd7;D</mml:mi>
<mml:mrow>
<mml:mo>(</mml:mo>
<mml:mrow>
<mml:mi mathvariant="bold-italic">k,j</mml:mi>
</mml:mrow>
<mml:mo>)</mml:mo>
</mml:mrow>
<mml:mi mathvariant="bold">,&#xa0;i&#x3d;1,&#xa0;&#x2026;,L</mml:mi>
<mml:mi mathvariant="bold">.&#xa0;&#xa0;j&#x3d;1,&#x2026;,20</mml:mi>
</mml:mrow>
</mml:math>
<label>(2)</label>
</disp-formula>where <inline-formula id="inf3">
<mml:math id="m5">
<mml:mrow>
<mml:mi>&#x3c9;</mml:mi>
<mml:mrow>
<mml:mo>(</mml:mo>
<mml:mi>i</mml:mi>
<mml:mo>,</mml:mo>
<mml:mi>k</mml:mi>
<mml:mo>)</mml:mo>
</mml:mrow>
</mml:mrow>
</mml:math>
</inline-formula> is the frequencies of the k-th type of amino acid at the i-th position and <inline-formula id="inf4">
<mml:math id="m6">
<mml:mrow>
<mml:mi>D</mml:mi>
<mml:mrow>
<mml:mo>(</mml:mo>
<mml:mi>k</mml:mi>
<mml:mo>,</mml:mo>
<mml:mi>j</mml:mi>
<mml:mo>)</mml:mo>
</mml:mrow>
</mml:mrow>
</mml:math>
</inline-formula> is the mutation percentage of the substitution matrix from the k-th molecular substances into the sequence of the protein. The larger the value, the more conserved the position is. If this is not present, the contrary will be achieved.</p>
<sec id="s2-2-1">
<title>2.2.1&#x20;Pse-Pssm</title>
<p>Pse-PSSM (pseudo-PSSM) is a feature extraction method often used in membrane protein prediction (<xref ref-type="bibr" rid="B12">Chou and Shen, 2007</xref>). PSSM matrices are often used in the characterization of membrane proteins. This feature extraction method, which aims to preserve PSSM biological information through pseudo amino acids, is expressed as follows:<disp-formula id="e3">
<mml:math id="m7">
<mml:mrow>
<mml:msub>
<mml:mi mathvariant="italic">f</mml:mi>
<mml:mrow>
<mml:mi mathvariant="bold-italic">i,j</mml:mi>
</mml:mrow>
</mml:msub>
<mml:mi mathvariant="bold-italic">&#x3d;</mml:mi>
<mml:mfrac>
<mml:mrow>
<mml:msub>
<mml:mi mathvariant="italic">p</mml:mi>
<mml:mrow>
<mml:mi mathvariant="bold-italic">i,j</mml:mi>
</mml:mrow>
</mml:msub>
<mml:mi mathvariant="bold-italic">-</mml:mi>
<mml:mfrac>
<mml:mi mathvariant="bold">1</mml:mi>
<mml:mrow>
<mml:mi mathvariant="bold">20</mml:mi>
</mml:mrow>
</mml:mfrac>
<mml:msubsup>
<mml:mstyle displaystyle="true">
<mml:mi mathvariant="bold-italic">&#x2211;</mml:mi>
</mml:mstyle>
<mml:mrow>
<mml:mi mathvariant="bold-italic">k&#x3d;1</mml:mi>
</mml:mrow>
<mml:mrow>
<mml:mi mathvariant="bold">20</mml:mi>
</mml:mrow>
</mml:msubsup>
<mml:msub>
<mml:mi mathvariant="italic">p</mml:mi>
<mml:mrow>
<mml:mi mathvariant="bold-italic">i,k</mml:mi>
</mml:mrow>
</mml:msub>
</mml:mrow>
<mml:mrow>
<mml:msqrt>
<mml:mrow>
<mml:mfrac>
<mml:mi mathvariant="bold">1</mml:mi>
<mml:mrow>
<mml:mi mathvariant="bold">20</mml:mi>
</mml:mrow>
</mml:mfrac>
<mml:msubsup>
<mml:mstyle displaystyle="true">
<mml:mi mathvariant="bold-italic">&#x2211;</mml:mi>
</mml:mstyle>
<mml:mrow>
<mml:mi mathvariant="bold-italic">l&#x3d;1</mml:mi>
</mml:mrow>
<mml:mrow>
<mml:mi mathvariant="bold">20</mml:mi>
</mml:mrow>
</mml:msubsup>
<mml:msup>
<mml:mrow>
<mml:mrow>
<mml:mo>(</mml:mo>
<mml:mrow>
<mml:msub>
<mml:mi mathvariant="italic">p</mml:mi>
<mml:mrow>
<mml:mi mathvariant="bold-italic">i,l</mml:mi>
</mml:mrow>
</mml:msub>
<mml:mi mathvariant="bold-italic">-</mml:mi>
<mml:mfrac>
<mml:mi mathvariant="bold">1</mml:mi>
<mml:mrow>
<mml:mi mathvariant="bold">20</mml:mi>
</mml:mrow>
</mml:mfrac>
<mml:msubsup>
<mml:mstyle displaystyle="true">
<mml:mi mathvariant="bold-italic">&#x2211;</mml:mi>
</mml:mstyle>
<mml:mrow>
<mml:mi mathvariant="bold-italic">k&#x3d;1</mml:mi>
</mml:mrow>
<mml:mrow>
<mml:mi mathvariant="bold">20</mml:mi>
</mml:mrow>
</mml:msubsup>
<mml:msub>
<mml:mi mathvariant="italic">p</mml:mi>
<mml:mrow>
<mml:mi mathvariant="bold-italic">i,k</mml:mi>
</mml:mrow>
</mml:msub>
</mml:mrow>
<mml:mo>)</mml:mo>
</mml:mrow>
</mml:mrow>
<mml:mi mathvariant="bold">2</mml:mi>
</mml:msup>
</mml:mrow>
</mml:msqrt>
</mml:mrow>
</mml:mfrac>
<mml:mi mathvariant="italic">,i&#x3d;1,&#x2026;,L;j&#x3d;1,&#x2026;,20</mml:mi>
</mml:mrow>
</mml:math>
<label>(3)</label>
</disp-formula>
</p>
<p>The <inline-formula id="inf5">
<mml:math id="m8">
<mml:mrow>
<mml:msub>
<mml:mi>P</mml:mi>
<mml:mrow>
<mml:mi>n</mml:mi>
<mml:mi>o</mml:mi>
<mml:mi>r</mml:mi>
<mml:mi>m</mml:mi>
<mml:mi>a</mml:mi>
<mml:mi>l</mml:mi>
<mml:mi>i</mml:mi>
<mml:mi>z</mml:mi>
<mml:mi>e</mml:mi>
<mml:mi>d</mml:mi>
</mml:mrow>
</mml:msub>
</mml:mrow>
</mml:math>
</inline-formula> is as follows:<disp-formula id="e4">
<mml:math id="m9">
<mml:mrow>
<mml:msub>
<mml:mi mathvariant="italic">P</mml:mi>
<mml:mrow>
<mml:mi mathvariant="italic">normalized</mml:mi>
</mml:mrow>
</mml:msub>
<mml:mi mathvariant="bold-italic">&#x3d;</mml:mi>
<mml:msub>
<mml:mrow>
<mml:mrow>
<mml:mo>[</mml:mo>
<mml:mrow>
<mml:mtable>
<mml:mtr>
<mml:mtd>
<mml:mrow>
<mml:msub>
<mml:mi mathvariant="italic">f</mml:mi>
<mml:mrow>
<mml:mi mathvariant="bold">1,1</mml:mi>
</mml:mrow>
</mml:msub>
</mml:mrow>
</mml:mtd>
<mml:mtd>
<mml:mi mathvariant="bold-italic">&#x2026;</mml:mi>
</mml:mtd>
<mml:mtd>
<mml:mrow>
<mml:msub>
<mml:mi mathvariant="italic">f</mml:mi>
<mml:mrow>
<mml:mi mathvariant="bold">1,20</mml:mi>
</mml:mrow>
</mml:msub>
</mml:mrow>
</mml:mtd>
</mml:mtr>
<mml:mtr>
<mml:mtd>
<mml:mrow>
<mml:mtable>
<mml:mtr>
<mml:mtd>
<mml:mi mathvariant="bold-italic">&#x22EE;</mml:mi>
</mml:mtd>
</mml:mtr>
<mml:mtr>
<mml:mtd>
<mml:mrow>
<mml:mtable>
<mml:mtr>
<mml:mtd>
<mml:mrow>
<mml:msub>
<mml:mi mathvariant="italic">f</mml:mi>
<mml:mrow>
<mml:mi mathvariant="bold-italic">i</mml:mi>
<mml:mo>,</mml:mo>
<mml:mi mathvariant="bold">1</mml:mi>
</mml:mrow>
</mml:msub>
</mml:mrow>
</mml:mtd>
</mml:mtr>
<mml:mtr>
<mml:mtd>
<mml:mi mathvariant="bold-italic">&#x22EE;</mml:mi>
</mml:mtd>
</mml:mtr>
</mml:mtable>
</mml:mrow>
</mml:mtd>
</mml:mtr>
</mml:mtable>
</mml:mrow>
</mml:mtd>
<mml:mtd>
<mml:mrow>
<mml:mtable>
<mml:mtr>
<mml:mtd>
<mml:mi mathvariant="bold-italic">&#x22F1;</mml:mi>
</mml:mtd>
</mml:mtr>
<mml:mtr>
<mml:mtd>
<mml:mrow>
<mml:mtable>
<mml:mtr>
<mml:mtd>
<mml:mi mathvariant="bold-italic">&#x2026;</mml:mi>
</mml:mtd>
</mml:mtr>
<mml:mtr>
<mml:mtd>
<mml:mi mathvariant="bold-italic">&#x22F1;</mml:mi>
</mml:mtd>
</mml:mtr>
</mml:mtable>
</mml:mrow>
</mml:mtd>
</mml:mtr>
</mml:mtable>
</mml:mrow>
</mml:mtd>
<mml:mtd>
<mml:mrow>
<mml:mtable>
<mml:mtr>
<mml:mtd>
<mml:mi mathvariant="bold-italic">&#x22EE;</mml:mi>
</mml:mtd>
</mml:mtr>
<mml:mtr>
<mml:mtd>
<mml:mrow>
<mml:mtable>
<mml:mtr>
<mml:mtd>
<mml:mrow>
<mml:msub>
<mml:mi mathvariant="italic">f</mml:mi>
<mml:mrow>
<mml:mi mathvariant="bold-italic">i</mml:mi>
<mml:mo>,</mml:mo>
<mml:mi mathvariant="bold">20</mml:mi>
</mml:mrow>
</mml:msub>
</mml:mrow>
</mml:mtd>
</mml:mtr>
<mml:mtr>
<mml:mtd>
<mml:mi mathvariant="bold-italic">&#x22EE;</mml:mi>
</mml:mtd>
</mml:mtr>
</mml:mtable>
</mml:mrow>
</mml:mtd>
</mml:mtr>
</mml:mtable>
</mml:mrow>
</mml:mtd>
</mml:mtr>
<mml:mtr>
<mml:mtd>
<mml:mrow>
<mml:msub>
<mml:mi mathvariant="italic">f</mml:mi>
<mml:mrow>
<mml:mi mathvariant="bold-italic">L</mml:mi>
<mml:mo>,</mml:mo>
<mml:mi mathvariant="bold">1</mml:mi>
</mml:mrow>
</mml:msub>
</mml:mrow>
</mml:mtd>
<mml:mtd>
<mml:mi mathvariant="bold-italic">&#x2026;</mml:mi>
</mml:mtd>
<mml:mtd>
<mml:mrow>
<mml:msub>
<mml:mi mathvariant="italic">f</mml:mi>
<mml:mrow>
<mml:mi mathvariant="bold-italic">L</mml:mi>
<mml:mo>,</mml:mo>
<mml:mi mathvariant="bold">20</mml:mi>
</mml:mrow>
</mml:msub>
</mml:mrow>
</mml:mtd>
</mml:mtr>
</mml:mtable>
</mml:mrow>
<mml:mo>]</mml:mo>
</mml:mrow>
</mml:mrow>
<mml:mrow>
<mml:mi mathvariant="bold-italic">L</mml:mi>
<mml:mo>&#xd7;</mml:mo>
<mml:mi mathvariant="bold">20</mml:mi>
</mml:mrow>
</mml:msub>
</mml:mrow>
</mml:math>
<label>(4)</label>
</disp-formula>where <inline-formula id="inf6">
<mml:math id="m10">
<mml:mrow>
<mml:msub>
<mml:mi>f</mml:mi>
<mml:mrow>
<mml:mi>i</mml:mi>
<mml:mo>,</mml:mo>
<mml:mi>j</mml:mi>
</mml:mrow>
</mml:msub>
<mml:mtext>&#xa0;</mml:mtext>
</mml:mrow>
</mml:math>
</inline-formula> is the normalised PSSM score with a mean of 0 for the 20 amino acids. And the <inline-formula id="inf7">
<mml:math id="m11">
<mml:mrow>
<mml:msub>
<mml:mi>p</mml:mi>
<mml:mrow>
<mml:mi>i</mml:mi>
<mml:mo>,</mml:mo>
<mml:mi>j</mml:mi>
</mml:mrow>
</mml:msub>
</mml:mrow>
</mml:math>
</inline-formula> is the raw score. While a positive score refers to the occurrence of the corresponding homozygous mutation, which is more frequent in multiple reciprocals, over and above chance mutations, a negative score is the opposite of a positive&#x20;score.</p>
</sec>
<sec id="s2-2-2">
<title>2.2.2 Average Blocks</title>
<p>The AB method was first proposed by Huang et&#x20;al. Its full name is the averaging block methodology (AvBlock) (<xref ref-type="bibr" rid="B23">Jong Cheol Jeong et&#x20;al., 2011</xref>). When feature extraction is performed for PSSM, the extracted feature values are diverse because the size of individual features is different and the abundance of amino acids also varies in the individual membrane proteins. To solve this type of problem, we can average the features for the local features of the PSSMs. Inside each module after averaging, 5% of the membrane protein sequences are covered, and this method is the AB feature extraction method. When performing the AB method feature extraction on the PSSM, it is not necessary to consider the sequence length of the membrane proteins. When we split the PSSM matrix by rows, it becomes a block of size L/20 each, with 20 blocks. After this operation, every 20 features form a block. the AB formulation is as follows:<disp-formula id="e5">
<mml:math id="m12">
<mml:mrow>
<mml:mi mathvariant="bold-italic">AB</mml:mi>
<mml:mrow>
<mml:mo>(</mml:mo>
<mml:mi mathvariant="bold-italic">k</mml:mi>
<mml:mo>)</mml:mo>
</mml:mrow>
<mml:mi mathvariant="bold-italic">&#x3d;</mml:mi>
<mml:mfrac>
<mml:mrow>
<mml:mi mathvariant="bold">20</mml:mi>
</mml:mrow>
<mml:mi mathvariant="bold-italic">N</mml:mi>
</mml:mfrac>
<mml:munderover>
<mml:mstyle displaystyle="true">
<mml:mi mathvariant="bold-italic">&#x2211;</mml:mi>
</mml:mstyle>
<mml:mrow>
<mml:mi mathvariant="bold">p&#x3d;1</mml:mi>
</mml:mrow>
<mml:mrow>
<mml:mfrac>
<mml:mi mathvariant="bold-italic">N</mml:mi>
<mml:mrow>
<mml:mi mathvariant="bold">20</mml:mi>
</mml:mrow>
</mml:mfrac>
</mml:mrow>
</mml:munderover>
<mml:mi mathvariant="bold-italic">Mt</mml:mi>
<mml:mrow>
<mml:mo>(</mml:mo>
<mml:mrow>
<mml:mi mathvariant="bold">p&#x2b;</mml:mi>
<mml:mrow>
<mml:mo>(</mml:mo>
<mml:mrow>
<mml:mi mathvariant="bold-italic">i-</mml:mi>
<mml:mi mathvariant="bold">1</mml:mi>
</mml:mrow>
<mml:mo>)</mml:mo>
</mml:mrow>
<mml:mi mathvariant="bold-italic">&#xd7;</mml:mi>
<mml:mfrac>
<mml:mrow>
<mml:mi mathvariant="bold">20</mml:mi>
</mml:mrow>
<mml:mi mathvariant="bold-italic">N</mml:mi>
</mml:mfrac>
<mml:mi mathvariant="bold-italic">,j</mml:mi>
</mml:mrow>
<mml:mo>)</mml:mo>
</mml:mrow>
<mml:mo>,</mml:mo>
<mml:mi mathvariant="bold-italic">&#xa0;i&#x3d;</mml:mi>
<mml:mi mathvariant="bold">1,</mml:mi>
<mml:mi mathvariant="bold">&#x2026;,20;</mml:mi>
<mml:mi mathvariant="bold-italic">j&#x3d;</mml:mi>
<mml:mi mathvariant="bold">1&#x2026;,20;</mml:mi>
<mml:mi mathvariant="bold-italic">j</mml:mi>
<mml:mi mathvariant="bold">&#x3d;1,&#x2026;,20;</mml:mi>
<mml:mi mathvariant="bold-italic">k&#x3d;j&#x2b;</mml:mi>
<mml:mi mathvariant="bold">20&#xd7;</mml:mi>
<mml:mi mathvariant="bold-italic">(i-</mml:mi>
<mml:mi mathvariant="bold">1)</mml:mi>
</mml:mrow>
</mml:math>
<label>(5)</label>
</disp-formula>Where N/20 is the size of j blocks and <inline-formula id="inf8">
<mml:math id="m13">
<mml:mrow>
<mml:mi>M</mml:mi>
<mml:mi>t</mml:mi>
<mml:mrow>
<mml:mo>(</mml:mo>
<mml:mrow>
<mml:mi>p</mml:mi>
<mml:mo>&#x2b;</mml:mo>
<mml:mrow>
<mml:mo>(</mml:mo>
<mml:mrow>
<mml:mi>i</mml:mi>
<mml:mo>&#x2212;</mml:mo>
<mml:mn>1</mml:mn>
</mml:mrow>
<mml:mo>)</mml:mo>
</mml:mrow>
<mml:mo>&#xd7;</mml:mo>
<mml:mfrac>
<mml:mrow>
<mml:mn>20</mml:mn>
</mml:mrow>
<mml:mi>N</mml:mi>
</mml:mfrac>
<mml:mo>,</mml:mo>
<mml:mi>j</mml:mi>
</mml:mrow>
<mml:mo>)</mml:mo>
</mml:mrow>
</mml:mrow>
</mml:math>
</inline-formula> is a vector of size 1&#x20;&#xd7; 20 from position <inline-formula id="inf9">
<mml:math id="m14">
<mml:mrow>
<mml:msub>
<mml:mi>i</mml:mi>
<mml:mrow>
<mml:mi>t</mml:mi>
<mml:mi>h</mml:mi>
</mml:mrow>
</mml:msub>
</mml:mrow>
</mml:math>
</inline-formula> of the <inline-formula id="inf10">
<mml:math id="m15">
<mml:mrow>
<mml:msub>
<mml:mi>j</mml:mi>
<mml:mrow>
<mml:mi>t</mml:mi>
<mml:mi>h</mml:mi>
</mml:mrow>
</mml:msub>
</mml:mrow>
</mml:math>
</inline-formula> block that is taken in the&#x20;pssms.</p>
</sec>
<sec id="s2-2-3">
<title>2.2.3 Discrete Wavelet Transform</title>
<p>We refer to a discrete wavelet transform feature extraction method as DWT, which uses the concepts of frequency and position (<xref ref-type="bibr" rid="B29">Nanni et&#x20;al., 2012</xref>). It is because we can consider the membrane protein sequence as a picture, then matrix the sequence and extract the coefficient information from the matrix by DWT. This method was first suggested by Nanni et&#x20;al.</p>
<p>In addition to this, we refer to the projection of the signal <inline-formula id="inf11">
<mml:math id="m16">
<mml:mrow>
<mml:mi>f</mml:mi>
<mml:mrow>
<mml:mo>(</mml:mo>
<mml:mi>t</mml:mi>
<mml:mo>)</mml:mo>
</mml:mrow>
</mml:mrow>
</mml:math>
</inline-formula> onto the wavelet function as the wavelet transform (WT). This is shown below:<disp-formula id="e6">
<mml:math id="m17">
<mml:mrow>
<mml:mi mathvariant="bold-italic">T</mml:mi>
<mml:mrow>
<mml:mo>(</mml:mo>
<mml:mrow>
<mml:mi mathvariant="bold-italic">a,&#xa0;b</mml:mi>
</mml:mrow>
<mml:mo>)</mml:mo>
</mml:mrow>
<mml:mi mathvariant="bold-italic">&#x3d;</mml:mi>
<mml:mfrac>
<mml:mi mathvariant="bold">1</mml:mi>
<mml:mrow>
<mml:msqrt>
<mml:mi mathvariant="bold-italic">a</mml:mi>
</mml:msqrt>
</mml:mrow>
</mml:mfrac>
<mml:munderover>
<mml:mstyle displaystyle="true">
<mml:mi mathvariant="bold-italic">&#x222B;</mml:mi>
</mml:mstyle>
<mml:mi mathvariant="bold-italic">0</mml:mi>
<mml:mi mathvariant="bold-italic">t</mml:mi>
</mml:munderover>
<mml:mi mathvariant="italic">f(t)&#x3c8;</mml:mi>
<mml:mrow>
<mml:mo>(</mml:mo>
<mml:mrow>
<mml:mfrac>
<mml:mrow>
<mml:mi mathvariant="bold-italic">t-b</mml:mi>
</mml:mrow>
<mml:mi mathvariant="bold-italic">a</mml:mi>
</mml:mfrac>
</mml:mrow>
<mml:mo>)</mml:mo>
</mml:mrow>
<mml:msub>
<mml:mi mathvariant="bold-italic">d</mml:mi>
<mml:mi mathvariant="bold-italic">t</mml:mi>
</mml:msub>
</mml:mrow>
</mml:math>
<label>(6)</label>
</disp-formula>where in the above equation, the scale variable is denoted by a, the translational variable by b, and <inline-formula id="inf12">
<mml:math id="m18">
<mml:mrow>
<mml:mi>&#x3c8;</mml:mi>
<mml:mrow>
<mml:mo>(</mml:mo>
<mml:mrow>
<mml:mfrac>
<mml:mrow>
<mml:mi>t</mml:mi>
<mml:mo>&#x2212;</mml:mo>
<mml:mi>b</mml:mi>
</mml:mrow>
<mml:mi>a</mml:mi>
</mml:mfrac>
</mml:mrow>
<mml:mo>)</mml:mo>
</mml:mrow>
</mml:mrow>
</mml:math>
</inline-formula> refers to the wavelet parsing and analysis function. T <inline-formula id="inf13">
<mml:math id="m19">
<mml:mrow>
<mml:mrow>
<mml:mo>(</mml:mo>
<mml:mrow>
<mml:mi>a</mml:mi>
<mml:mo>,</mml:mo>
<mml:mtext>&#xa0;</mml:mtext>
<mml:mi>b</mml:mi>
</mml:mrow>
<mml:mo>)</mml:mo>
</mml:mrow>
</mml:mrow>
</mml:math>
</inline-formula> refers to the transform&#x20;coefficients used in conjunction with a particular location when performing a specific wavelet period signal transform. Further, an efficient DWT algorithm was submitted by Nanni et&#x20;al.; they denote the discrete signal <inline-formula id="inf14">
<mml:math id="m20">
<mml:mrow>
<mml:mi>f</mml:mi>
<mml:mrow>
<mml:mo>(</mml:mo>
<mml:mi>t</mml:mi>
<mml:mo>)</mml:mo>
</mml:mrow>
</mml:mrow>
</mml:math>
</inline-formula> by x [n] and perform a DWT on it. It is expressed in terms of the coefficients as follows:<disp-formula id="e7">
<mml:math id="m21">
<mml:mrow>
<mml:msub>
<mml:mi mathvariant="bold-italic">y</mml:mi>
<mml:mrow>
<mml:mi mathvariant="bold-italic">j,low</mml:mi>
</mml:mrow>
</mml:msub>
<mml:mrow>
<mml:mo>[</mml:mo>
<mml:mi mathvariant="bold-italic">n</mml:mi>
<mml:mo>]</mml:mo>
</mml:mrow>
<mml:mi mathvariant="bold-italic">&#x3d;</mml:mi>
<mml:munderover>
<mml:mstyle displaystyle="true">
<mml:mi mathvariant="bold-italic">&#x2211;</mml:mi>
</mml:mstyle>
<mml:mrow>
<mml:mi mathvariant="bold-italic">k&#x3d;1</mml:mi>
</mml:mrow>
<mml:mi mathvariant="bold-italic">N</mml:mi>
</mml:munderover>
<mml:mi mathvariant="bold-italic">x</mml:mi>
<mml:mrow>
<mml:mo>[</mml:mo>
<mml:mi mathvariant="bold-italic">k</mml:mi>
<mml:mo>]</mml:mo>
</mml:mrow>
<mml:mi mathvariant="bold-italic">g</mml:mi>
<mml:mrow>
<mml:mo>[</mml:mo>
<mml:mrow>
<mml:mi mathvariant="bold">2</mml:mi>
<mml:mi mathvariant="bold-italic">n-k</mml:mi>
</mml:mrow>
<mml:mo>]</mml:mo>
</mml:mrow>
</mml:mrow>
</mml:math>
<label>(7)</label>
</disp-formula>
<disp-formula id="e8">
<mml:math id="m22">
<mml:mrow>
<mml:msub>
<mml:mi mathvariant="bold-italic">y</mml:mi>
<mml:mrow>
<mml:mi mathvariant="bold-italic">j,high</mml:mi>
</mml:mrow>
</mml:msub>
<mml:mrow>
<mml:mo>[</mml:mo>
<mml:mi mathvariant="bold-italic">n</mml:mi>
<mml:mo>]</mml:mo>
</mml:mrow>
<mml:mi mathvariant="bold-italic">&#x3d;</mml:mi>
<mml:munderover>
<mml:mstyle displaystyle="true">
<mml:mi mathvariant="bold-italic">&#x2211;</mml:mi>
</mml:mstyle>
<mml:mrow>
<mml:mi mathvariant="bold-italic">k&#x3d;1</mml:mi>
</mml:mrow>
<mml:mi mathvariant="bold-italic">N</mml:mi>
</mml:munderover>
<mml:mi mathvariant="bold-italic">x</mml:mi>
<mml:mrow>
<mml:mo>[</mml:mo>
<mml:mi mathvariant="bold-italic">k</mml:mi>
<mml:mo>]</mml:mo>
</mml:mrow>
<mml:mi mathvariant="bold-italic">h</mml:mi>
<mml:mrow>
<mml:mo>[</mml:mo>
<mml:mrow>
<mml:mi mathvariant="bold">2</mml:mi>
<mml:mi mathvariant="bold-italic">n-k</mml:mi>
</mml:mrow>
<mml:mo>]</mml:mo>
</mml:mrow>
</mml:mrow>
</mml:math>
<label>(8)</label>
</disp-formula>where N is the length of the discrete signal and the low-pass and high-pass filters are g and h, respectively. <inline-formula id="inf15">
<mml:math id="m23">
<mml:mrow>
<mml:msub>
<mml:mi>y</mml:mi>
<mml:mrow>
<mml:mi>j</mml:mi>
<mml:mo>,</mml:mo>
<mml:mi>l</mml:mi>
<mml:mi>o</mml:mi>
<mml:mi>w</mml:mi>
</mml:mrow>
</mml:msub>
<mml:mrow>
<mml:mo>[</mml:mo>
<mml:mi>n</mml:mi>
<mml:mo>]</mml:mo>
</mml:mrow>
</mml:mrow>
</mml:math>
</inline-formula> is the approximate coefficient when the signal is in the low-frequency part, while <inline-formula id="inf16">
<mml:math id="m24">
<mml:mrow>
<mml:msub>
<mml:mi>y</mml:mi>
<mml:mrow>
<mml:mi>j</mml:mi>
<mml:mo>,</mml:mo>
<mml:mi>h</mml:mi>
<mml:mi>i</mml:mi>
<mml:mi>g</mml:mi>
<mml:mi>h</mml:mi>
</mml:mrow>
</mml:msub>
<mml:mrow>
<mml:mo>[</mml:mo>
<mml:mi>n</mml:mi>
<mml:mo>]</mml:mo>
</mml:mrow>
</mml:mrow>
</mml:math>
</inline-formula> is the detailed coefficient when the signal is in the high-frequency band. In our study, their mean, standard deviation, maximum and minimum values are computed through the DWT quadruple layer. In addition, the PSSM discrete signal after transforming four times, consists of 20 discrete signals. The 4-stage DWT structure can be seen in <xref ref-type="fig" rid="F3">Figure&#x20;3</xref>.</p>
<fig id="F3" position="float">
<label>FIGURE 3</label>
<caption>
<p>The 4-stage DWT structure.</p>
</caption>
<graphic xlink:href="fgene-12-834488-g003.tif"/>
</fig>
</sec>
<sec id="s2-2-4">
<title>2.2.4 Discrete Cosine Transform</title>
<p>The DCT (<xref ref-type="bibr" rid="B1">Ahmed et&#x20;al., 1974</xref>), known as the Discrete Cosine Transform, converts a signal to its fundamental frequency by means of a linearly separable transform. This method has been widely used in the field of image compression. In this experiment,&#x20;we compressed the PSSM matrix of membrane proteins using a 2-dimensional DCT (2D-DCT). 2D-DCT is defined as follows:<disp-formula id="e9">
<mml:math id="m25">
<mml:mrow>
<mml:msub>
<mml:mi mathvariant="bold-italic">F</mml:mi>
<mml:mrow>
<mml:mi mathvariant="bold-italic">PSSM-DCT</mml:mi>
</mml:mrow>
</mml:msub>
<mml:mi mathvariant="bold-italic">&#x3d;</mml:mi>
<mml:msub>
<mml:mi mathvariant="bold-italic">&#x3b1;</mml:mi>
<mml:mi mathvariant="bold-italic">i</mml:mi>
</mml:msub>
<mml:msub>
<mml:mi mathvariant="bold-italic">&#x3b1;</mml:mi>
<mml:mi mathvariant="bold-italic">j</mml:mi>
</mml:msub>
<mml:munderover>
<mml:mstyle displaystyle="true">
<mml:mi mathvariant="bold-italic">&#x2211;</mml:mi>
</mml:mstyle>
<mml:mrow>
<mml:mi mathvariant="bold-italic">m&#x3d;</mml:mi>
<mml:mi mathvariant="bold">0</mml:mi>
</mml:mrow>
<mml:mrow>
<mml:mi mathvariant="bold-italic">M-</mml:mi>
<mml:mi mathvariant="bold">1</mml:mi>
</mml:mrow>
</mml:munderover>
<mml:munderover>
<mml:mstyle displaystyle="true">
<mml:mi mathvariant="bold-italic">&#x2211;</mml:mi>
</mml:mstyle>
<mml:mrow>
<mml:mi mathvariant="bold-italic">n&#x3d;</mml:mi>
<mml:mi mathvariant="bold">0</mml:mi>
</mml:mrow>
<mml:mrow>
<mml:mi mathvariant="bold-italic">N-</mml:mi>
<mml:mi mathvariant="bold">1</mml:mi>
</mml:mrow>
</mml:munderover>
<mml:mi mathvariant="bold-italic">PSSM(m,n)cos</mml:mi>
<mml:mfrac>
<mml:mrow>
<mml:mi mathvariant="bold-italic">&#x3c0;</mml:mi>
<mml:mrow>
<mml:mo>(</mml:mo>
<mml:mrow>
<mml:mi mathvariant="bold-italic">2m&#x2b;</mml:mi>
<mml:mi mathvariant="bold">1</mml:mi>
</mml:mrow>
<mml:mo>)</mml:mo>
</mml:mrow>
<mml:mi mathvariant="bold-italic">i</mml:mi>
</mml:mrow>
<mml:mrow>
<mml:mi mathvariant="bold-italic">2M</mml:mi>
</mml:mrow>
</mml:mfrac>
<mml:mi mathvariant="bold-italic">cos</mml:mi>
<mml:mfrac>
<mml:mrow>
<mml:mi mathvariant="bold-italic">&#x3c0;</mml:mi>
<mml:mrow>
<mml:mo>(</mml:mo>
<mml:mrow>
<mml:mi mathvariant="bold-italic">2n&#x2b;</mml:mi>
<mml:mi mathvariant="bold">1</mml:mi>
</mml:mrow>
<mml:mo>)</mml:mo>
</mml:mrow>
<mml:mi mathvariant="bold-italic">j</mml:mi>
</mml:mrow>
<mml:mrow>
<mml:mi mathvariant="bold-italic">2N</mml:mi>
</mml:mrow>
</mml:mfrac>
</mml:mrow>
</mml:math>
<label>(9)</label>
</disp-formula>
<disp-formula id="e10">
<mml:math id="m26">
<mml:mrow>
<mml:msub>
<mml:mi mathvariant="bold-italic">&#x3b1;</mml:mi>
<mml:mi mathvariant="bold-italic">i</mml:mi>
</mml:msub>
<mml:mrow>
<mml:mo>{</mml:mo>
<mml:mrow>
<mml:mtable>
<mml:mtr>
<mml:mtd>
<mml:mrow>
<mml:msqrt>
<mml:mrow>
<mml:mfrac>
<mml:mi mathvariant="bold">1</mml:mi>
<mml:mi mathvariant="bold-italic">M</mml:mi>
</mml:mfrac>
</mml:mrow>
</mml:msqrt>
<mml:mi mathvariant="bold-italic">,i&#x3d;</mml:mi>
<mml:mi mathvariant="bold">0</mml:mi>
</mml:mrow>
</mml:mtd>
</mml:mtr>
<mml:mtr>
<mml:mtd>
<mml:mrow>
<mml:msqrt>
<mml:mrow>
<mml:mfrac>
<mml:mi mathvariant="bold">2</mml:mi>
<mml:mi mathvariant="bold-italic">M</mml:mi>
</mml:mfrac>
</mml:mrow>
</mml:msqrt>
<mml:mo>,</mml:mo>
<mml:mi mathvariant="bold">1</mml:mi>
<mml:mi mathvariant="bold-italic">&#x2264;i&#x2264;M-</mml:mi>
<mml:mi mathvariant="bold">1</mml:mi>
</mml:mrow>
</mml:mtd>
</mml:mtr>
</mml:mtable>
</mml:mrow>
</mml:mrow>
<mml:msub>
<mml:mi mathvariant="bold-italic">&#x3b1;</mml:mi>
<mml:mi mathvariant="bold-italic">j</mml:mi>
</mml:msub>
<mml:mrow>
<mml:mo>{</mml:mo>
<mml:mrow>
<mml:mtable>
<mml:mtr>
<mml:mtd>
<mml:mrow>
<mml:msqrt>
<mml:mrow>
<mml:mfrac>
<mml:mi mathvariant="bold">1</mml:mi>
<mml:mi mathvariant="bold-italic">N</mml:mi>
</mml:mfrac>
</mml:mrow>
</mml:msqrt>
<mml:mi mathvariant="bold-italic">,j&#x3d;</mml:mi>
<mml:mi mathvariant="bold">0</mml:mi>
</mml:mrow>
</mml:mtd>
</mml:mtr>
<mml:mtr>
<mml:mtd>
<mml:mrow>
<mml:msqrt>
<mml:mrow>
<mml:mfrac>
<mml:mi mathvariant="bold">2</mml:mi>
<mml:mi mathvariant="bold-italic">N</mml:mi>
</mml:mfrac>
</mml:mrow>
</mml:msqrt>
<mml:mo>,</mml:mo>
<mml:mi mathvariant="bold">1</mml:mi>
<mml:mi mathvariant="bold-italic">&#x2264;j&#x2264;N-</mml:mi>
<mml:mi mathvariant="bold">1</mml:mi>
</mml:mrow>
</mml:mtd>
</mml:mtr>
</mml:mtable>
</mml:mrow>
</mml:mrow>
</mml:mrow>
</mml:math>
<label>(10)</label>
</disp-formula>
</p>
<p>The mission of the DCT is to convert a uniformly distributed information density into an uneven distribution. Once its length and signal have been converted, the most important part of the information is collected in the low frequency section of the PSSM, that is in the middle and top-left corner.</p>
</sec>
<sec id="s2-2-5">
<title>2.2.5 Histogram of Oriented Gradient</title>
<p>Histogram of Oriented Gradients (HOG), which is a method for describing features, is mainly used in computer vision. In this experiment, to handle a PSSM matrix using the HOG method, it is first necessary to look at it as a particular image matrix. In the first step, the horizontal gradient values and vertical gradient values of the PSSM are used to derive the direction and size of the gradient matrix. In the second step, the gradient matrix is divided into 25&#x20;sub-matrices by direction and size. In the third step, conversion of the results generated in the second step is carried out according to the requirement to generate 10 histogram channels per sub-matrix.</p>
</sec>
</sec>
<sec id="s2-3">
<title>2.3 Lifelong Learning</title>
<p>Lifelong learning (<xref ref-type="bibr" rid="B41">Thrun and Mitchell, 1995</xref>), like machine learning, can be divided into the directions of lifelong supervised learning, lifelong unsupervised learning and lifelong reinforcement learning. The lifelong machine learning part of the study focuses on whether the model can be extended when new categories of categories are added to the model. When the current model has been classified into n categories, if a new class of data is added, the model can somehow be adaptively expanded to classify n &#x2b; 1 category. Multi-task learning models and lifelong machine learning are easily translatable to each other if we have all the original data. Whenever a new category is added to the original category and needs to be classified, only one new category needs to be added and then all the training data can be trained again to expand the new category. One obvious disadvantage of this strategy is that it wastes a lot of computational time to compute each new class, and if too many new classes are added, it may lead to changes in the model architecture for multi-task learning. This model therefore uses a dynamically scalable network to better perform incremental learning of the added tasks. Assume that the current model has successfully classified Class 1, Class 2, ... , Class <italic>n</italic>. When the new data class Class-new is added, the model does not need to train all the data from scratch, but only needs to expand the overall model by adding n new binary classification models. The simple flow of the lifelong learning model is shown in <xref ref-type="fig" rid="F4">Figure&#x20;4</xref>.</p>
<fig id="F4" position="float">
<label>FIGURE 4</label>
<caption>
<p>The simple flow of the lifelong learning&#x20;model.</p>
</caption>
<graphic xlink:href="fgene-12-834488-g004.tif"/>
</fig>
</sec>
<sec id="s2-4">
<title>2.4 Dynamically Scalable Networks</title>
<p>Dynamically scalable networks are incremental training of deep neural networks for lifelong learning, for which there will be an unknown amount and unknown distribution of data to be trained fed into our model in turn. To expand on this, there are now T a sequence of task learning models, t &#x3d; 1, .... , t, .... , T is unbounded T in which the tasks at time point t carry training data <inline-formula id="inf17">
<mml:math id="m27">
<mml:mrow>
<mml:msub>
<mml:mi>D</mml:mi>
<mml:mi>t</mml:mi>
</mml:msub>
<mml:mo>&#x3d;</mml:mo>
<mml:msubsup>
<mml:mrow>
<mml:mrow>
<mml:mo>{</mml:mo>
<mml:mrow>
<mml:msub>
<mml:mi>x</mml:mi>
<mml:mi>i</mml:mi>
</mml:msub>
<mml:mo>,</mml:mo>
<mml:msub>
<mml:mi>y</mml:mi>
<mml:mi>i</mml:mi>
</mml:msub>
</mml:mrow>
<mml:mo>}</mml:mo>
</mml:mrow>
</mml:mrow>
<mml:mrow>
<mml:mi>i</mml:mi>
<mml:mo>&#x3d;</mml:mo>
<mml:mn>1</mml:mn>
</mml:mrow>
<mml:mrow>
<mml:msub>
<mml:mi>N</mml:mi>
<mml:mi>t</mml:mi>
</mml:msub>
</mml:mrow>
</mml:msubsup>
</mml:mrow>
</mml:math>
</inline-formula>. It is important to note that each subtask can be a single or a group of tasks. For simplicity, even though our approach is general for any kind of task, we only consider the two-classification problem. That is, input features <inline-formula id="inf18">
<mml:math id="m28">
<mml:mrow>
<mml:mi>x</mml:mi>
<mml:mtext>&#xa0;</mml:mtext>
<mml:mo>&#x2208;</mml:mo>
<mml:mtext>&#xa0;</mml:mtext>
<mml:msup>
<mml:mi>R</mml:mi>
<mml:mi>d</mml:mi>
</mml:msup>
</mml:mrow>
</mml:math>
</inline-formula> of <inline-formula id="inf19">
<mml:math id="m29">
<mml:mrow>
<mml:mi>y</mml:mi>
<mml:mtext>&#xa0;</mml:mtext>
<mml:mo>&#x2208;</mml:mo>
<mml:mtext>&#xa0;</mml:mtext>
<mml:mrow>
<mml:mo>{</mml:mo>
<mml:mn>0</mml:mn>
<mml:mo>,</mml:mo>
<mml:mtext>&#xa0;</mml:mtext>
<mml:mn>1</mml:mn>
<mml:mo>}</mml:mo>
</mml:mrow>
</mml:mrow>
</mml:math>
</inline-formula>. One challenge with lifelong learning is that at the current time t, all previous training datasets are unavailable (if any, only from previous model parameters). The lifelong learning agent learns the model parameters <inline-formula id="inf20">
<mml:math id="m30">
<mml:mrow>
<mml:msup>
<mml:mi>W</mml:mi>
<mml:mi>t</mml:mi>
</mml:msup>
</mml:mrow>
</mml:math>
</inline-formula> by solving the following problem in a reasonable amount of time t:<disp-formula id="e11">
<mml:math id="m31">
<mml:mrow>
<mml:mi mathvariant="bold-italic">minimiz</mml:mi>
<mml:msub>
<mml:mi mathvariant="bold-italic">e</mml:mi>
<mml:mrow>
<mml:msup>
<mml:mi mathvariant="bold-italic">W</mml:mi>
<mml:mi mathvariant="bold-italic">t</mml:mi>
</mml:msup>
</mml:mrow>
</mml:msub>
<mml:mi mathvariant="bold-italic">L</mml:mi>
<mml:mrow>
<mml:mo>(</mml:mo>
<mml:mrow>
<mml:msup>
<mml:mi mathvariant="bold-italic">W</mml:mi>
<mml:mi mathvariant="bold-italic">t</mml:mi>
</mml:msup>
<mml:mi mathvariant="bold-italic">&#xa0;;</mml:mi>
<mml:msup>
<mml:mi mathvariant="bold-italic">W</mml:mi>
<mml:mrow>
<mml:mi mathvariant="bold-italic">t-</mml:mi>
<mml:mi mathvariant="bold">1</mml:mi>
</mml:mrow>
</mml:msup>
<mml:mo>,</mml:mo>
<mml:msub>
<mml:mi mathvariant="bold-italic">D</mml:mi>
<mml:mi mathvariant="bold-italic">t</mml:mi>
</mml:msub>
</mml:mrow>
<mml:mo>)</mml:mo>
</mml:mrow>
<mml:mi mathvariant="bold-italic">&#x2b;&#xa0;&#x3bb;</mml:mi>
<mml:mi mathvariant="normal">&#x3a9;</mml:mi>
<mml:mrow>
<mml:mo>(</mml:mo>
<mml:mrow>
<mml:msup>
<mml:mi mathvariant="bold-italic">W</mml:mi>
<mml:mi mathvariant="bold-italic">t</mml:mi>
</mml:msup>
<mml:mi mathvariant="bold-italic">&#xa0;</mml:mi>
</mml:mrow>
<mml:mo>)</mml:mo>
</mml:mrow>
<mml:mi mathvariant="bold-italic">,&#xa0;t&#xa0;&#x3d;&#xa0;1</mml:mi>
<mml:mi mathvariant="bold">1</mml:mi>
<mml:mi mathvariant="bold">,&#xa0;&#x2026;</mml:mi>
</mml:mrow>
</mml:math>
<label>(11)</label>
</disp-formula>where L is task specific loss function, <inline-formula id="inf21">
<mml:math id="m32">
<mml:mrow>
<mml:msup>
<mml:mi>W</mml:mi>
<mml:mi>t</mml:mi>
</mml:msup>
</mml:mrow>
</mml:math>
</inline-formula> is the parameter for task t, and <inline-formula id="inf22">
<mml:math id="m33">
<mml:mtext>&#x3a9;</mml:mtext>
</mml:math>
</inline-formula> (<inline-formula id="inf23">
<mml:math id="m34">
<mml:mrow>
<mml:msup>
<mml:mi>W</mml:mi>
<mml:mi>t</mml:mi>
</mml:msup>
</mml:mrow>
</mml:math>
</inline-formula>) is the regularization (e.g. element-wise <inline-formula id="inf24">
<mml:math id="m35">
<mml:mrow>
<mml:msub>
<mml:mi>&#x2113;</mml:mi>
<mml:mn>2</mml:mn>
</mml:msub>
</mml:mrow>
</mml:math>
</inline-formula> norm) to enforce our model <inline-formula id="inf25">
<mml:math id="m36">
<mml:mrow>
<mml:msup>
<mml:mi>W</mml:mi>
<mml:mi>t</mml:mi>
</mml:msup>
</mml:mrow>
</mml:math>
</inline-formula> appropriately. In case of a neural network which is our primary interest, <inline-formula id="inf26">
<mml:math id="m37">
<mml:mrow>
<mml:msup>
<mml:mi>W</mml:mi>
<mml:mi>t</mml:mi>
</mml:msup>
<mml:mo>&#x3d;</mml:mo>
<mml:msubsup>
<mml:mrow>
<mml:mrow>
<mml:mo>{</mml:mo>
<mml:mrow>
<mml:msub>
<mml:mi>W</mml:mi>
<mml:mi>l</mml:mi>
</mml:msub>
</mml:mrow>
<mml:mo>}</mml:mo>
</mml:mrow>
</mml:mrow>
<mml:mrow>
<mml:mi>l</mml:mi>
<mml:mo>&#x3d;</mml:mo>
<mml:mn>1</mml:mn>
</mml:mrow>
<mml:mi>L</mml:mi>
</mml:msubsup>
</mml:mrow>
</mml:math>
</inline-formula> is the weight tensor.</p>
<p>To counter these problems that arise in the course of lifelong learning, we allow the knowledge generated in previous tasks to be used to the maximum extent possible. At the same time, it is allowed&#x20;to dynamically extend its capabilities when mechanically accumulated knowledge does not explain well for emerging tasks. <xref ref-type="fig" rid="F5">Figure&#x20;5</xref> and <xref ref-type="table" rid="T3">Algorithm 1</xref> illustrate our progressive learning process.</p>
<fig id="F5" position="float">
<label>FIGURE 5</label>
<caption>
<p>Incremental learning for dynamically scalable networks.</p>
</caption>
<graphic xlink:href="fgene-12-834488-g005.tif"/>
</fig>
</sec>
</sec>
<sec id="s3">
<title>3 Experiment Results</title>
<p>In this subsection we will have an analysis of the capabilities of the&#x20;respective modelling and methodological approaches. Furthermore, the modelling we used in that context was compared with other available methods on separate datasets.</p>
<sec id="s3-1">
<title>3.1 Assessment Measurements</title>
<p>To evaluate our lifelong learning model better, we chose several parameters: sample all-prediction accuracy (ACC), single sample specificity (SP), single sample sensitivity (SN), and Mathews correlation coefficient. These metrics are widely used in the analysis of biological sequence information:<disp-formula id="e12">
<mml:math id="m38">
<mml:mrow>
<mml:mrow>
<mml:mo>{</mml:mo>
<mml:mrow>
<mml:mtable>
<mml:mtr>
<mml:mtd>
<mml:mrow>
<mml:mi mathvariant="bold-italic">SN&#x3d;</mml:mi>
<mml:mfrac>
<mml:mrow>
<mml:mi mathvariant="bold-italic">TP</mml:mi>
</mml:mrow>
<mml:mrow>
<mml:mi mathvariant="bold-italic">TP&#x2b;FN</mml:mi>
</mml:mrow>
</mml:mfrac>
</mml:mrow>
</mml:mtd>
</mml:mtr>
<mml:mtr>
<mml:mtd>
<mml:mrow>
<mml:mi mathvariant="bold-italic">SP&#x3d;</mml:mi>
<mml:mfrac>
<mml:mrow>
<mml:mi mathvariant="bold-italic">TN</mml:mi>
</mml:mrow>
<mml:mrow>
<mml:mi mathvariant="bold-italic">TN&#x2b;FP</mml:mi>
</mml:mrow>
</mml:mfrac>
</mml:mrow>
</mml:mtd>
</mml:mtr>
<mml:mtr>
<mml:mtd>
<mml:mrow>
<mml:mi mathvariant="bold-italic">ACC&#x3d;</mml:mi>
<mml:mfrac>
<mml:mrow>
<mml:mi mathvariant="bold-italic">TP&#x2b;TN</mml:mi>
</mml:mrow>
<mml:mrow>
<mml:mi mathvariant="bold-italic">TP&#x2b;FP&#x2b;TN&#x2b;FN</mml:mi>
</mml:mrow>
</mml:mfrac>
</mml:mrow>
</mml:mtd>
</mml:mtr>
<mml:mtr>
<mml:mtd>
<mml:mrow>
<mml:mi mathvariant="bold-italic">MCC&#x3d;</mml:mi>
<mml:mfrac>
<mml:mrow>
<mml:mi mathvariant="bold-italic">TP&#xd7;TN-FP&#xd7;FN</mml:mi>
</mml:mrow>
<mml:mrow>
<mml:msqrt>
<mml:mrow>
<mml:mrow>
<mml:mo>(</mml:mo>
<mml:mrow>
<mml:mi mathvariant="bold-italic">TP&#x2b;FN</mml:mi>
</mml:mrow>
<mml:mo>)</mml:mo>
</mml:mrow>
<mml:mi mathvariant="bold-italic">&#xd7;</mml:mi>
<mml:mrow>
<mml:mo>(</mml:mo>
<mml:mrow>
<mml:mi mathvariant="bold-italic">TN&#x2b;FP</mml:mi>
</mml:mrow>
<mml:mo>)</mml:mo>
</mml:mrow>
<mml:mi mathvariant="bold-italic">&#xd7;</mml:mi>
<mml:mrow>
<mml:mo>(</mml:mo>
<mml:mrow>
<mml:mi mathvariant="bold-italic">TP&#x2b;FP</mml:mi>
</mml:mrow>
<mml:mo>)</mml:mo>
</mml:mrow>
<mml:mi mathvariant="bold-italic">&#xd7;</mml:mi>
<mml:mrow>
<mml:mo>(</mml:mo>
<mml:mrow>
<mml:mi mathvariant="bold-italic">TN&#x2b;FN</mml:mi>
</mml:mrow>
<mml:mo>)</mml:mo>
</mml:mrow>
</mml:mrow>
</mml:msqrt>
</mml:mrow>
</mml:mfrac>
</mml:mrow>
</mml:mtd>
</mml:mtr>
</mml:mtable>
</mml:mrow>
</mml:mrow>
<mml:mi mathvariant="bold-italic">&#xa0;&#xa0;&#xa0;</mml:mi>
</mml:mrow>
</mml:math>
<label>(12)</label>
</disp-formula>
</p>
<p>For the above equation, true trueness (TP) refers to the number of true samples correctly predicted; false positivity (FP) refers to the number of true samples incorrectly predicted; true negativity (TN) refers to the number of negative samples correctly predicted; and false negativity (FN) refers to the number of negative samples erroneously predicted (<xref ref-type="bibr" rid="B14">Chou et&#x20;al., 2011a</xref>; <xref ref-type="bibr" rid="B16">Chou et&#x20;al., 2011b</xref>; <xref ref-type="bibr" rid="B13">Chou, 2013</xref>).</p>
</sec>
<sec id="s3-2">
<title>3.2 Situational Analysis of Two Data Sets</title>
<p>The lengths of the datasets used in our experiments are shown in <xref ref-type="fig" rid="F6">Figure&#x20;6</xref>. Most of the membrane proteins in dataset 1 and dataset 2 have a similar length distribution because of their specific type. To better demonstrate the superiority of lifelong learning for membrane protein classification, we calculated the amino acid frequencies for all protein types in the experiment, as shown in <xref ref-type="fig" rid="F7">Figure&#x20;7</xref>.</p>
<fig id="F6" position="float">
<label>FIGURE 6</label>
<caption>
<p>Distribution of the lengths of the training and test sets in the two datasets.</p>
</caption>
<graphic xlink:href="fgene-12-834488-g006.tif"/>
</fig>
<fig id="F7" position="float">
<label>FIGURE 7</label>
<caption>
<p>Component composition of the training and test sets of the two datasets.</p>
</caption>
<graphic xlink:href="fgene-12-834488-g007.tif"/>
</fig>
</sec>
<sec id="s3-3">
<title>3.3 The Forecasting Results for Dataset 1</title>
<p>The PSSM sequence matrix contains important genetic information required for protein prediction. Many elements of biological evolution, such as the stability of the three-dimensional structure and the aggregation of proteins, can have an impact on the storage and alteration of sequences. These elements demonstrate that PSSM captures important information about ligand binding. Thus, proving the validity of the PSSM characterization method.</p>
<p>We will compare the model methods we used in this use&#x20;with other existing methods in terms of prediction accuracy on dataset 1. The methods involved in the comparison are MemType-2L (<xref ref-type="bibr" rid="B12">Chou and Shen, 2007</xref>) and&#x20;predMPT (<xref ref-type="bibr" rid="B7">Chen and Li, 2013</xref>). Details can be found in&#x20;<xref ref-type="table" rid="T2">Table&#x20;2</xref>, where it is clear that our model approach has&#x20;an overall ACC of 95.3%. 3.7% higher than MemType-2L&#x2019;s 91.6%, 2.7% higher than predMPT&#x2019;s 92.6%, and 2.4%&#x20;higher than Average weights&#x2019; 92.7%. In the independent test set, our method was superior for membrane protein type 2 (88.4%), type 4 (83.3%), type 5 (96.3%) and type 7 (100%).</p>
<table-wrap id="T2" position="float">
<label>TABLE 2</label>
<caption>
<p>The performance exhibited by different models on dataset 1.</p>
</caption>
<table>
<thead valign="top">
<tr>
<th align="left">The specific type of classification</th>
<th align="center">
<italic>Ave-WT</italic>
<xref ref-type="table-fn" rid="Tfn1">
<sup>a</sup>
</xref> (%)</th>
<th align="center">
<italic>Mem Type-2L</italic>
<xref ref-type="table-fn" rid="Tfn2">
<sup>b</sup>
</xref> (%)</th>
<th align="center">
<italic>preMPT</italic>
<xref ref-type="table-fn" rid="Tfn3">
<sup>c</sup>
</xref> (%)</th>
<th align="center">
<italic>Our method</italic>
</th>
</tr>
</thead>
<tbody valign="top">
<tr>
<td align="left">Single-span type 1</td>
<td align="char" char="(">93.9 (417/444)</td>
<td align="char" char="(">86.9 (386/444)</td>
<td align="char" char="(">94.6 (420/444)</td>
<td align="char" char="(">94.8 (421/444)</td>
</tr>
<tr>
<td align="left">Single-span type 2</td>
<td align="char" char="(">87.2 (68/78)</td>
<td align="char" char="(">70.5 (55/78)</td>
<td align="char" char="(">79.5 (62/78)</td>
<td align="char" char="(">88.4 (69/78)</td>
</tr>
<tr>
<td align="left">Single-span type 3</td>
<td align="char" char="(">0 (0/6)</td>
<td align="char" char="(">33.3 (2/6)</td>
<td align="char" char="(">33.3 (2/6)</td>
<td align="char" char="(">33.3 (2/6)</td>
</tr>
<tr>
<td align="left">Single-span type 4</td>
<td align="char" char="(">66.7 (8/12)</td>
<td align="char" char="(">66.7 (8/12)</td>
<td align="char" char="(">41.7 (5/12)</td>
<td align="char" char="(">83.3 (10/12)</td>
</tr>
<tr>
<td align="left">Multi-span type 5</td>
<td align="char" char="(">93.9 (3,065/3,265)</td>
<td align="char" char="(">95.0 (3,103/3,265)</td>
<td align="char" char="(">94.9 (3,097/3,265)</td>
<td align="char" char="(">96.3 (3,147/3,265)</td>
</tr>
<tr>
<td align="left">Lipid-anchor type 6</td>
<td align="char" char="(">29.0 (11/38)</td>
<td align="char" char="(">42.1 (16/38)</td>
<td align="char" char="(">65.8 (25/38)</td>
<td align="char" char="(">52.6 (20/38)</td>
</tr>
<tr>
<td align="left">GPI-anchor type 7</td>
<td align="char" char="(">84.8 (39/46)</td>
<td align="char" char="(">76.1 (35/46)</td>
<td align="char" char="(">93.5 (43/46)</td>
<td align="char" char="(">100.0 (46/46)</td>
</tr>
<tr>
<td align="left">Peripheral type 8</td>
<td align="char" char="(">91.9 (408/444)</td>
<td align="char" char="(">82.2 (365/444)</td>
<td align="char" char="(">81.1 (360/444)</td>
<td align="char" char="(">93.2 (414/444)</td>
</tr>
<tr>
<td align="left">Overall</td>
<td align="char" char="(">92.7 (4,016/4,333)</td>
<td align="char" char="(">91.6 (3,970/4,333)</td>
<td align="char" char="(">92.6 (4,014/4,333)</td>
<td align="char" char="(">95.3 (4,129/4,333)</td>
</tr>
</tbody>
</table>
<table-wrap-foot>
<fn id="Tfn1">
<label>a</label>
<p>Mean-weighted MKSVM,&#x20;based.</p>
</fn>
<fn id="Tfn2">
<label>b</label>
<p>The results are taken from (<xref ref-type="bibr" rid="B12">Chou and Shen, 2007</xref>).</p>
</fn>
<fn id="Tfn3">
<label>c</label>
<p>The results are taken from (<xref ref-type="bibr" rid="B7">Chen and Li, 2013</xref>).</p>
</fn>
</table-wrap-foot>
</table-wrap>
</sec>
<sec id="s3-4">
<title>3.4 The Forecasting Results for Dataset 2</title>
<p>As a solution to the possible problem of untimely updates in Data 1, Chen and Li (<xref ref-type="bibr" rid="B7">Chen and Li, 2013</xref>) used the Swissprot annotation method to update Data Set 1, resulting in a new dataset of membrane proteins (Data Set 2). The results of the comparison using Dataset 2 are presented in <xref ref-type="table" rid="T3">Table&#x20;3</xref>. The overall average accuracy of our models was 3.2% higher than the predMPT method (90.3%). Even though they added features such as 3D structure to the predMPT prediction session, our model performance was clearly higher than it. We outperformed it by 2.6% in terms of prediction accuracy for type 1 (94.1 vs. 91.5%). In contrast, analog 5 outperformed it by 1.3% (94.1 vs. 92.8%).</p>
<table-wrap id="T3" position="float">
<label>ALGORITHM 1</label>
<caption>
<p>Incremental Learning for Dynamically Scalable Networks.</p>
</caption>
<table>
<tbody>
<tr>
<td>
<inline-graphic xlink:href="fgene-12-834488-fx1.tif"/>
</td>
</tr>
</tbody>
</table>
</table-wrap>
<table-wrap id="T4" position="float">
<label>TABLE 3</label>
<caption>
<p>The performance exhibited by different models on dataset 2.</p>
</caption>
<table>
<thead valign="top">
<tr>
<th align="left">The specific type of classification</th>
<th align="center">
<italic>Ave-WT</italic>
<xref ref-type="table-fn" rid="Tfn4">
<sup>a</sup>
</xref> (%)</th>
<th align="center">
<italic>Mem Type-2L</italic>
<xref ref-type="table-fn" rid="Tfn5">
<sup>b</sup>
</xref> (%)</th>
<th align="center">
<italic>preMPT</italic>
<xref ref-type="table-fn" rid="Tfn6">
<sup>c</sup>
</xref> (%)</th>
<th align="center">
<italic>Our method</italic>
</th>
</tr>
</thead>
<tbody valign="top">
<tr>
<td align="left">Single-span type 1</td>
<td align="char" char="(">89.2 (199/223)</td>
<td align="char" char="(">76.7 (171/223)</td>
<td align="char" char="(">91.5 (204/223)</td>
<td align="char" char="(">94.1 (210/223)</td>
</tr>
<tr>
<td align="left">Single-span type 2</td>
<td align="char" char="(">79.5 (31/39)</td>
<td align="char" char="(">66.7 (26/39)</td>
<td align="char" char="(">74.4 (29/39)</td>
<td align="char" char="(">87.2 (34/39)</td>
</tr>
<tr>
<td align="left">Single-span type 3</td>
<td align="char" char="(">33.3 (2/6)</td>
<td align="char" char="(">33.3 (2/6)</td>
<td align="char" char="(">16.7 (1/6)</td>
<td align="char" char="(">50.0 (3/6)</td>
</tr>
<tr>
<td align="left">Single-span type 4</td>
<td align="char" char="(">90.0 (9/10)</td>
<td align="char" char="(">70.0 (7/10)</td>
<td align="char" char="(">80.0 (8/10)</td>
<td align="char" char="(">90.0 (9/10)</td>
</tr>
<tr>
<td align="left">Multi-span type 5</td>
<td align="char" char="(">91.1 (1,524/1,673)</td>
<td align="char" char="(">91.4 (1,529/1,673)</td>
<td align="char" char="(">92.8 (1,552/1,673)</td>
<td align="char" char="(">94.1 (1,575/1,673)</td>
</tr>
<tr>
<td align="left">Lipid-anchor type 6</td>
<td align="char" char="(">30.8 (8/26)</td>
<td align="char" char="(">23.1 (6/26)</td>
<td align="char" char="(">53.8 (14/26)</td>
<td align="char" char="(">57.6 (15/26)</td>
</tr>
<tr>
<td align="left">GPI-anchor type 7</td>
<td align="char" char="(">91.7 (22/26)</td>
<td align="char" char="(">70.8 (17/24)</td>
<td align="char" char="(">95.8 (23/24)</td>
<td align="char" char="(">100.0 (24/24)</td>
</tr>
<tr>
<td align="left">Peripheral type 8</td>
<td align="char" char="(">88.9 (271/305)</td>
<td align="char" char="(">68.2 (208/305)</td>
<td align="char" char="(">82.6 (252/305)</td>
<td align="char" char="(">93.7 (286/305)</td>
</tr>
<tr>
<td align="left">Overall</td>
<td align="char" char="(">89.6 (2066/2,306)</td>
<td align="char" char="(">85.3 (1966/2,306)</td>
<td align="char" char="(">90.3 (2083/2,306)</td>
<td align="char" char="(">93.5 (2,156/2,306)</td>
</tr>
</tbody>
</table>
<table-wrap-foot>
<fn id="Tfn4">
<label>a</label>
<p>Mean-weighted MKSVM,&#x20;based.</p>
</fn>
<fn id="Tfn5">
<label>b</label>
<p>The results are taken from (<xref ref-type="bibr" rid="B12">Chou and Shen, 2007</xref>).</p>
</fn>
<fn id="Tfn6">
<label>c</label>
<p>The results are taken from (<xref ref-type="bibr" rid="B7">Chen and Li, 2013</xref>).</p>
</fn>
</table-wrap-foot>
</table-wrap>
</sec>
</sec>
<sec id="s4">
<title>4 Conclusion and Discussion</title>
<p>In previous work, investigators have often used the PseAAC (<xref ref-type="bibr" rid="B15">Chou, 2001</xref>) approach to identify membrane protein types, and this approach has indeed performed well in the field of protein classification. Using Chou&#x2019;s operation (<xref ref-type="bibr" rid="B12">Chou and Shen, 2007</xref>) on feature extraction from PSSM, we were inspired to use the five methods of Pse-pssm, DCT, AvBlock, HOG and DWT to extract features. In order to avoid the low accuracy of a single feature extraction method, we integrated the above five methods together and fed the integrated features into our DSN model method.</p>
<p>Our constructed lifelong learning dynamic network model proved to achieve superior results on different datasets (95.3 and 93.5%). However, the prediction of some small sample affiliations by the methodology has not been as accurate as we had anticipated. In order to improve the performance of this model, we will consider improving our own features and combining some other feature extraction methods, and adjusting the parameters of our model in our future research.</p>
</sec>
</body>
<back>
<sec id="s5">
<title>Data Availability Statement</title>
<p>Publicly available datasets were analyzed in this study. This data can be found here: <ext-link ext-link-type="uri" xlink:href="https://github.com/sjw-cmd/DATA.git">https://github.com/sjw-cmd/DATA.git</ext-link>.</p>
</sec>
<sec id="s6">
<title>Author Contributions</title>
<p>WL proposed the original idea. JS and HW designed the framework and the experiments. WL, JS and YZ performed the experiments and the primary data analysis. WL and JS wrote the manuscript. YQ, QF and XC modified the codes and the manuscript. All authors contributed to the manuscript. All authors read and approved the final manuscript.</p>
</sec>
<sec id="s7">
<title>Funding</title>
<p>This paper is supported by the National Natural Science Foundation of China (No.61902272, 62073231, 61772357, 62176175, 61876217, 61902271), National Research Project (2020YFC2006602), Provincial Key Laboratory for Computer Information Processing Technology, Soochow University (KJS2166).</p>
</sec>
<sec sec-type="COI-statement" id="s8">
<title>Conflict of Interest</title>
<p>The authors declare that the research was conducted in the absence of any commercial or financial relationships that could be construed as a potential conflict of interest.</p>
</sec>
<sec sec-type="disclaimer" id="s9">
<title>Publisher&#x2019;s Note</title>
<p>All claims expressed in this article are solely those of the authors and do not necessarily represent those of their affiliated organizations, or those of the publisher, the editors and the reviewers. Any product that may be evaluated in this article, or claim that may be made by its manufacturer, is not guaranteed or endorsed by the publisher.</p>
</sec>
<ack>
<p>The authors acknowledge and thank the reviewers for their suggestions that allowed the improvement of our manuscript.</p>
</ack>
<ref-list>
<title>References</title>
<ref id="B1">
<citation citation-type="journal">
<person-group person-group-type="author">
<name>
<surname>Ahmed</surname>
<given-names>N.</given-names>
</name>
<name>
<surname>Natarajan</surname>
<given-names>T.</given-names>
</name>
<name>
<surname>Rao</surname>
<given-names>K. R.</given-names>
</name>
</person-group> (<year>1974</year>). <article-title>Discrete Cosine Transform</article-title>. <source>IEEE Trans. Comput.</source> <volume>C-23</volume> (<issue>1</issue>), <fpage>90</fpage>&#x2013;<lpage>93</lpage>. <pub-id pub-id-type="doi">10.1109/t-c.1974.223784</pub-id> </citation>
</ref>
<ref id="B2">
<citation citation-type="journal">
<person-group person-group-type="author">
<name>
<surname>Altschul</surname>
<given-names>S.</given-names>
</name>
<name>
<surname>Madden</surname>
<given-names>T.</given-names>
</name>
<name>
<surname>Schffer</surname>
<given-names>A.</given-names>
</name>
<name>
<surname>Zhang</surname>
<given-names>J.</given-names>
</name>
<name>
<surname>Zhang</surname>
<given-names>Z.</given-names>
</name>
<name>
<surname>Miller</surname>
<given-names>W.</given-names>
</name>
<etal/>
</person-group> (<year>1997</year>). <article-title>Gapped Blast and Psi-Blast: a New Generation of Protein Database Search Programs</article-title>. <source>Nucleic Acids Res.</source> <volume>25</volume> (<issue>17</issue>), <fpage>3389</fpage>&#x2013;<lpage>3402</lpage>. <pub-id pub-id-type="doi">10.1093/nar/25.17.3389</pub-id> </citation>
</ref>
<ref id="B3">
<citation citation-type="journal">
<person-group person-group-type="author">
<name>
<surname>Boeckmann</surname>
<given-names>B.</given-names>
</name>
<name>
<surname>Bairoch</surname>
<given-names>A.</given-names>
</name>
<name>
<surname>Apweiler</surname>
<given-names>R.</given-names>
</name>
<etal/>
</person-group> (<year>2003</year>). <article-title>The Swiss-Prot Protein Knowledgebase and its Supplement Trembl in 2003</article-title>. <source>Nucleic Acids Res.</source> <volume>31</volume> (<issue>1</issue>), <fpage>365</fpage>&#x2013;<lpage>370</lpage>. <pub-id pub-id-type="doi">10.1093/nar/gkg095</pub-id> </citation>
</ref>
<ref id="B4">
<citation citation-type="journal">
<person-group person-group-type="author">
<name>
<surname>Cai</surname>
<given-names>Y.-D.</given-names>
</name>
<name>
<surname>Ricardo</surname>
<given-names>P.-W.</given-names>
</name>
<name>
<surname>Jen</surname>
<given-names>C.-H.</given-names>
</name>
<name>
<surname>Chou</surname>
<given-names>K.-C.</given-names>
</name>
</person-group> (<year>2004</year>). <article-title>Application of Svm to Predict Membrane Protein Types</article-title>. <source>J.&#x20;Theor. Biol.</source> <volume>226</volume> (<issue>4</issue>), <fpage>373</fpage>&#x2013;<lpage>376</lpage>. <pub-id pub-id-type="doi">10.1016/j.jtbi.2003.08.015</pub-id> </citation>
</ref>
<ref id="B5">
<citation citation-type="journal">
<person-group person-group-type="author">
<name>
<surname>Cedano</surname>
<given-names>J.</given-names>
</name>
<name>
<surname>Aloy</surname>
<given-names>P.</given-names>
</name>
<name>
<surname>P&#xe9;rez-Pons</surname>
<given-names>J.&#x20;A.</given-names>
</name>
<name>
<surname>Querol</surname>
<given-names>E.</given-names>
</name>
</person-group> (<year>1997</year>). <article-title>Relation between Amino Acid Composition and Cellular Location of Proteins 1&#x20;1Edited by F. E. Cohen</article-title>. <source>J.&#x20;Mol. Biol.</source> <volume>266</volume> (<issue>3</issue>), <fpage>594</fpage>&#x2013;<lpage>600</lpage>. <pub-id pub-id-type="doi">10.1006/jmbi.1996.0804</pub-id> </citation>
</ref>
<ref id="B6">
<citation citation-type="journal">
<person-group person-group-type="author">
<name>
<surname>Chandra</surname>
<given-names>A. A.</given-names>
</name>
<name>
<surname>Sharma</surname>
<given-names>A.</given-names>
</name>
<name>
<surname>Dehzangi</surname>
<given-names>A.</given-names>
</name>
<name>
<surname>Tsunoda</surname>
<given-names>T.</given-names>
</name>
</person-group> (<year>2019</year>). <article-title>Evolstruct-phogly: Incorporating Structural Properties and Evolutionary Information from Profile Bigrams for the Phosphoglycerylation Prediction</article-title>. <source>BMC Genomics</source> <volume>19</volume> (<issue>9</issue>), <fpage>984</fpage>. <pub-id pub-id-type="doi">10.1186/s12864-018-5383-5</pub-id> </citation>
</ref>
<ref id="B7">
<citation citation-type="journal">
<person-group person-group-type="author">
<name>
<surname>Chen</surname>
<given-names>Y.-K.</given-names>
</name>
<name>
<surname>Li</surname>
<given-names>K.-B.</given-names>
</name>
</person-group> (<year>2013</year>). <article-title>Predicting Membrane Protein Types by Incorporating Protein Topology, Domains, Signal Peptides, and Physicochemical Properties into the General Form of Chou&#x27;s Pseudo Amino Acid Composition</article-title>. <source>J.&#x20;Theor. Biol.</source> <volume>318</volume> (<issue>1</issue>), <fpage>1</fpage>&#x2013;<lpage>12</lpage>. <pub-id pub-id-type="doi">10.1016/j.jtbi.2012.10.033</pub-id> </citation>
</ref>
<ref id="B8">
<citation citation-type="book">
<person-group person-group-type="author">
<name>
<surname>Chen</surname>
<given-names>Z.</given-names>
</name>
<name>
<surname>Liu</surname>
<given-names>B.</given-names>
</name>
</person-group> (<year>2019</year>). <source>Lifelong Machine learning[M]</source>. <publisher-loc>San Rafael, CA</publisher-loc>: <publisher-name>Morgan &#x26; Claypool Publishers</publisher-name>, <fpage>1207</fpage>. </citation>
</ref>
<ref id="B9">
<citation citation-type="confproc">
<person-group person-group-type="author">
<name>
<surname>Chen</surname>
<given-names>Z.</given-names>
</name>
<name>
<surname>Liu</surname>
<given-names>B.</given-names>
</name>
</person-group> (<year>2014</year>). &#x201c;<article-title>Topic Modeling Using Topics from many Domains, Lifelong Learning and Big Data[C]</article-title>,&#x201d; in <conf-name>Proceedings of the 31th International Conference on Machine Learning</conf-name> (<publisher-loc>Beijing</publisher-loc>), <fpage>703711</fpage>. </citation>
</ref>
<ref id="B10">
<citation citation-type="confproc">
<person-group person-group-type="author">
<name>
<surname>Chen</surname>
<given-names>Z.</given-names>
</name>
<name>
<surname>Ma</surname>
<given-names>N.</given-names>
</name>
<name>
<surname>Liu</surname>
<given-names>B.</given-names>
</name>
</person-group> (<year>2015</year>). &#x201c;<article-title>Lifelong Learning for Sentiment Classification[C]</article-title>,&#x201d; in <conf-name>Proceedings of the 53rd Annual Meeting of the Association for Computational Linguistics</conf-name> (<publisher-loc>Beijing</publisher-loc>), <fpage>750756</fpage>. </citation>
</ref>
<ref id="B11">
<citation citation-type="journal">
<person-group person-group-type="author">
<name>
<surname>Chou</surname>
<given-names>K.-C.</given-names>
</name>
<name>
<surname>Elrod</surname>
<given-names>D. W.</given-names>
</name>
</person-group> (<year>1999</year>). <article-title>Prediction of Membrane Protein Types and Subcellular Locations</article-title>. <source>Proteins</source> <volume>34</volume> (<issue>1</issue>), <fpage>137</fpage>&#x2013;<lpage>153</lpage>. <pub-id pub-id-type="doi">10.1002/(sici)1097-0134(19990101)34:1&#x3c;137:aid-prot11&#x3e;3.0.co;2-o</pub-id> </citation>
</ref>
<ref id="B12">
<citation citation-type="journal">
<person-group person-group-type="author">
<name>
<surname>Chou</surname>
<given-names>K.-C.</given-names>
</name>
<name>
<surname>Shen</surname>
<given-names>H.-B.</given-names>
</name>
</person-group> (<year>2007</year>). <article-title>Memtype-2L: a Web Server for Predicting Membrane Proteins and Their Types by Incorporating Evolution Information through Pse-Pssm</article-title>. <source>Biochem. Biophysical Res. Commun.</source> <volume>360</volume> (<issue>2</issue>), <fpage>339</fpage>&#x2013;<lpage>345</lpage>. <pub-id pub-id-type="doi">10.1016/j.bbrc.2007.06.027</pub-id> </citation>
</ref>
<ref id="B13">
<citation citation-type="journal">
<person-group person-group-type="author">
<name>
<surname>Chou</surname>
<given-names>K.-C.</given-names>
</name>
</person-group> (<year>2013</year>). <article-title>Some Remarks on Predicting Multi-Label Attributes in Molecular Biosystems</article-title>. <source>Mol. Biosyst.</source> <volume>9</volume> (<issue>6</issue>), <fpage>1092</fpage>&#x2013;<lpage>1100</lpage>. <pub-id pub-id-type="doi">10.1039/c3mb25555g</pub-id> </citation>
</ref>
<ref id="B14">
<citation citation-type="journal">
<person-group person-group-type="author">
<name>
<surname>Chou</surname>
<given-names>K.-C.</given-names>
</name>
<name>
<surname>Wu</surname>
<given-names>Z.-C.</given-names>
</name>
<name>
<surname>Xiao</surname>
<given-names>X.</given-names>
</name>
</person-group> (<year>2011</year>). <article-title>iLoc-Euk: A Multi-Label Classifier for Predicting the Subcellular Localization of Singleplex and Multiplex&#x20;Eukaryotic Proteins</article-title>. <source>Plos One</source> <volume>6</volume> (<issue>3</issue>), <fpage>e18258</fpage>. <pub-id pub-id-type="doi">10.1371/journal.pone.0018258</pub-id> </citation>
</ref>
<ref id="B15">
<citation citation-type="journal">
<person-group person-group-type="author">
<name>
<surname>Chou</surname>
<given-names>K. C.</given-names>
</name>
</person-group> (<year>2001</year>). <article-title>Prediction of Protein Cellular Attributes Using Pseudo-amino Acid Composition</article-title>. <source>Proteins</source> <volume>43</volume> (<issue>6</issue>), <fpage>246</fpage>&#x2013;<lpage>255</lpage>. <pub-id pub-id-type="doi">10.1002/prot.1035</pub-id> </citation>
</ref>
<ref id="B16">
<citation citation-type="journal">
<person-group person-group-type="author">
<name>
<surname>Chou</surname>
<given-names>K. C.</given-names>
</name>
<name>
<surname>Wu</surname>
<given-names>Z. C.</given-names>
</name>
<name>
<surname>Xiao</surname>
<given-names>X.</given-names>
</name>
</person-group> (<year>2011</year>). <article-title>Iloc-hum: Using the Accumulation-Label Scale to Predict Subcellular Locations of Human Proteins with Both Single and Multiple Sites</article-title>. <source>Mol. Biosyst.</source> <volume>8</volume> (<issue>2</issue>), <fpage>629</fpage>&#x2013;<lpage>641</lpage>. <pub-id pub-id-type="doi">10.1039/c1mb05420a</pub-id> </citation>
</ref>
<ref id="B17">
<citation citation-type="journal">
<person-group person-group-type="author">
<name>
<surname>Dehzangi</surname>
<given-names>A.</given-names>
</name>
<name>
<surname>L&#xf3;pez</surname>
<given-names>Y.</given-names>
</name>
<name>
<surname>Lal</surname>
<given-names>S. P.</given-names>
</name>
<name>
<surname>Taherzadeh</surname>
<given-names>G.</given-names>
</name>
<name>
<surname>Michaelson</surname>
<given-names>J.</given-names>
</name>
<name>
<surname>Sattar</surname>
<given-names>A.</given-names>
</name>
<etal/>
</person-group> (<year>2017</year>). <article-title>Pssm-suc: Accurately Predicting Succinylation Using Position Specific Scoring Matrix into Bigram for Feature Extraction</article-title>. <source>J.&#x20;Theor. Biol.</source> <volume>425</volume>, <fpage>97</fpage>&#x2013;<lpage>102</lpage>. <pub-id pub-id-type="doi">10.1016/j.jtbi.2017.05.005</pub-id> </citation>
</ref>
<ref id="B18">
<citation citation-type="journal">
<person-group person-group-type="author">
<name>
<surname>Ding</surname>
<given-names>Y.</given-names>
</name>
<name>
<surname>Tang</surname>
<given-names>J.</given-names>
</name>
<name>
<surname>Guo</surname>
<given-names>F.</given-names>
</name>
</person-group> (<year>2021</year>). <article-title>Identification of Drug-Target Interactions via Multi-View Graph Regularized Link Propagation Model</article-title>. <source>Neurocomputing</source> <volume>461</volume>, <fpage>618</fpage>&#x2013;<lpage>631</lpage>. <pub-id pub-id-type="doi">10.1016/j.neucom.2021.05.100</pub-id> </citation>
</ref>
<ref id="B19">
<citation citation-type="journal">
<person-group person-group-type="author">
<name>
<surname>Ding</surname>
<given-names>Y.</given-names>
</name>
<name>
<surname>Tang</surname>
<given-names>J.</given-names>
</name>
<name>
<surname>Guo</surname>
<given-names>F.</given-names>
</name>
</person-group> (<year>2017</year>). <article-title>Identification of Protein-Ligand Binding Sites by Sequence Information and Ensemble Classifier</article-title>. <source>J.&#x20;Chem. Inf. Model.</source> <volume>57</volume> (<issue>12</issue>), <fpage>3149</fpage>&#x2013;<lpage>3161</lpage>. <pub-id pub-id-type="doi">10.1021/acs.jcim.7b00307</pub-id> </citation>
</ref>
<ref id="B21">
<citation citation-type="journal">
<person-group person-group-type="author">
<name>
<surname>Ding</surname>
<given-names>Y.</given-names>
</name>
<name>
<surname>Yang</surname>
<given-names>C.</given-names>
</name>
<name>
<surname>Tang</surname>
<given-names>J.</given-names>
</name>
<name>
<surname>Guo</surname>
<given-names>F.</given-names>
</name>
</person-group> (<year>2021</year>). <article-title>Identification of Protein-Nucleotide Binding Residues via Graph Regularized K-Local Hyperplane Distance Nearest Neighbor Model</article-title>. <source>Appl. Intell.</source> (<issue>10</issue>), <fpage>1</fpage>&#x2013;<lpage>15</lpage>. <pub-id pub-id-type="doi">10.1007/s10489-021-02737-0</pub-id> </citation>
</ref>
<ref id="B22">
<citation citation-type="journal">
<person-group person-group-type="author">
<name>
<surname>Feng</surname>
<given-names>Z.-P.</given-names>
</name>
<name>
<surname>Zhang</surname>
<given-names>C.-T.</given-names>
</name>
</person-group> (<year>2000</year>). <article-title>Prediction of Membrane Protein Types Based on the Hydrophobic index of Amino Acids</article-title>. <source>J.&#x20;Protein Chem.</source> <volume>19</volume> (<issue>4</issue>), <fpage>269</fpage>&#x2013;<lpage>275</lpage>. <pub-id pub-id-type="doi">10.1023/a:1007091128394</pub-id> </citation>
</ref>
<ref id="B23">
<citation citation-type="journal">
<person-group person-group-type="author">
<name>
<surname>Jong Cheol Jeong</surname>
<given-names>J.</given-names>
</name>
<name>
<surname>Xiaotong Lin</surname>
<given-names>X.</given-names>
</name>
<name>
<surname>Xue-Wen Chen</surname>
<given-names>X.</given-names>
</name>
</person-group> (<year>2011</year>). <article-title>On Position-specific Scoring Matrix for Protein Function Prediction</article-title>. <source>Ieee/acm Trans. Comput. Biol. Bioinf.</source> <volume>8</volume> (<issue>2</issue>), <fpage>308</fpage>&#x2013;<lpage>315</lpage>. <pub-id pub-id-type="doi">10.1109/tcbb.2010.93</pub-id> </citation>
</ref>
<ref id="B24">
<citation citation-type="confproc">
<person-group person-group-type="author">
<name>
<surname>Kumar</surname>
<given-names>A.</given-names>
</name>
<name>
<surname>DaumeIII</surname>
<given-names>H.</given-names>
</name>
</person-group> (<year>2012</year>). &#x201c;<article-title>Learning Task Grouping and Overlap in Multitask Learning[C]</article-title>,&#x201d; in <conf-name>Proceedings of the 29th International Conference on Machine Learning</conf-name> (<publisher-loc>Edinburgh</publisher-loc>), <fpage>112</fpage>. </citation>
</ref>
<ref id="B25">
<citation citation-type="journal">
<person-group person-group-type="author">
<name>
<surname>Li</surname>
<given-names>W.</given-names>
</name>
<name>
<surname>Godzik</surname>
<given-names>A</given-names>
</name>
</person-group>.(<year>2006</year>). <article-title>Cd-hit: a Fast Program for Clustering and Comparing Large Sets of Protein or Nucleotide Sequences</article-title>. <source>Bioinformatics</source> <volume>22</volume> (<issue>13</issue>), <fpage>1658</fpage>&#x2013;<lpage>1659</lpage>. <pub-id pub-id-type="doi">10.1093/bioinformatics/btl158</pub-id> </citation>
</ref>
<ref id="B26">
<citation citation-type="journal">
<person-group person-group-type="author">
<name>
<surname>Lu</surname>
<given-names>W.</given-names>
</name>
<name>
<surname>Song</surname>
<given-names>Z.</given-names>
</name>
<name>
<surname>Ding</surname>
<given-names>Y.</given-names>
</name>
<name>
<surname>Wu</surname>
<given-names>H.</given-names>
</name>
<name>
<surname>Cao</surname>
<given-names>Y.</given-names>
</name>
<name>
<surname>Zhang</surname>
<given-names>Y.</given-names>
</name>
<etal/>
</person-group> (<year>2020</year>). <article-title>Use Chou&#x27;s 5-Step Rule to Predict DNA-Binding Proteins with Evolutionary Information</article-title>. <source>Biomed. Res. Int.</source> <volume>2020</volume>, <fpage>6984045</fpage>. <pub-id pub-id-type="doi">10.1155/2020/6984045</pub-id> </citation>
</ref>
<ref id="B27">
<citation citation-type="journal">
<person-group person-group-type="author">
<name>
<surname>Lu</surname>
<given-names>W.</given-names>
</name>
<name>
<surname>Cao</surname>
<given-names>Y.</given-names>
</name>
<name>
<surname>Wu</surname>
<given-names>H.</given-names>
</name>
<name>
<surname>Ding</surname>
<given-names>Y.</given-names>
</name>
<name>
<surname>Song</surname>
<given-names>Z.</given-names>
</name>
<name>
<surname>Zhang</surname>
<given-names>Y.</given-names>
</name>
<etal/>
</person-group> (<year>2021</year>). <article-title>Research on RNA Secondary Structure Predicting via Bidirectional Recurrent Neural Network</article-title>. <source>BMC Bioinformatics</source> <volume>22</volume>, <fpage>431</fpage>. <pub-id pub-id-type="doi">10.1186/s12859-021-04332-z</pub-id> </citation>
</ref>
<ref id="B28">
<citation citation-type="journal">
<person-group person-group-type="author">
<name>
<surname>Lu</surname>
<given-names>W.</given-names>
</name>
<name>
<surname>Tang</surname>
<given-names>Y.</given-names>
</name>
<name>
<surname>Wu</surname>
<given-names>H.</given-names>
</name>
<name>
<surname>Huang</surname>
<given-names>H.</given-names>
</name>
<name>
<surname>Fu</surname>
<given-names>Q.</given-names>
</name>
<name>
<surname>Qiu</surname>
<given-names>J.</given-names>
</name>
<etal/>
</person-group> (<year>2019</year>). <article-title>Predicting RNA Secondary Structure via Adaptive Deep Recurrent Neural Networks with Energy-Based Filter</article-title>. <source>BMC Bioinformatics</source> <volume>20</volume>, <fpage>684</fpage>. <pub-id pub-id-type="doi">10.1186/s12859-019-3258-7</pub-id> </citation>
</ref>
<ref id="B29">
<citation citation-type="journal">
<person-group person-group-type="author">
<name>
<surname>Nanni</surname>
<given-names>L.</given-names>
</name>
<name>
<surname>Brahnam</surname>
<given-names>S.</given-names>
</name>
<name>
<surname>Lumini</surname>
<given-names>A.</given-names>
</name>
</person-group> (<year>2012</year>). <article-title>Wavelet Images and Chou&#x27;s Pseudo Amino Acid Composition for Protein Classification</article-title>. <source>Amino Acids</source> <volume>43</volume> (<issue>2</issue>), <fpage>657</fpage>&#x2013;<lpage>665</lpage>. <pub-id pub-id-type="doi">10.1007/s00726-011-1114-9</pub-id> </citation>
</ref>
<ref id="B30">
<citation citation-type="journal">
<person-group person-group-type="author">
<name>
<surname>Qian</surname>
<given-names>Y.</given-names>
</name>
<name>
<surname>Meng</surname>
<given-names>H.</given-names>
</name>
<name>
<surname>Lu</surname>
<given-names>W.</given-names>
</name>
<name>
<surname>Liao</surname>
<given-names>Z.</given-names>
</name>
<name>
<surname>Ding</surname>
<given-names>Y.</given-names>
</name>
<name>
<surname>Wu</surname>
<given-names>H.</given-names>
</name>
</person-group> (<year>2021</year>). <article-title>Identification of DNA-Binding Proteins via Hypergraph Based Laplacian Support Vector Machine</article-title>. <source>Cbio</source> <volume>16</volume>. <pub-id pub-id-type="doi">10.2174/1574893616666210806091922</pub-id> </citation>
</ref>
<ref id="B31">
<citation citation-type="confproc">
<person-group person-group-type="author">
<name>
<surname>Ruvolo</surname>
<given-names>P.</given-names>
</name>
<name>
<surname>Eaton</surname>
<given-names>E.</given-names>
</name>
</person-group> (<year>2013</year>). &#x201c;<article-title>Active Task Selection for Lifelong Machine Learning[C]</article-title>,&#x201d; in <conf-name>Proceedings of the Twenty-seventh AAAI Conference on Artificial Intelligence</conf-name> (<publisher-loc>Washington</publisher-loc>), <fpage>862868</fpage>. </citation>
</ref>
<ref id="B32">
<citation citation-type="confproc">
<person-group person-group-type="author">
<name>
<surname>Ruvolo</surname>
<given-names>P.</given-names>
</name>
<name>
<surname>Eaton</surname>
<given-names>E.</given-names>
</name>
</person-group> (<year>2013</year>). &#x201c;<article-title>Ella: An Efficient Lifelong Learning Algorithm[C]</article-title>,&#x201d; in <conf-name>Proceedings of the 30th International Conference on Machine Learning</conf-name> (<publisher-loc>Atlanta</publisher-loc>), <fpage>507515</fpage>. </citation>
</ref>
<ref id="B33">
<citation citation-type="journal">
<person-group person-group-type="author">
<name>
<surname>Sharma</surname>
<given-names>R.</given-names>
</name>
<name>
<surname>Raicar</surname>
<given-names>G.</given-names>
</name>
<name>
<surname>Tsunoda</surname>
<given-names>T.</given-names>
</name>
<name>
<surname>Patil</surname>
<given-names>A.</given-names>
</name>
<name>
<surname>Sharma</surname>
<given-names>A.</given-names>
</name>
</person-group> (<year>2018</year>). <article-title>Opal: Prediction of Morf Regions in Intrinsically Disordered Protein Sequences</article-title>. <source>Bioinformatics</source> <volume>34</volume> (<issue>11</issue>), <fpage>1850</fpage>&#x2013;<lpage>1858</lpage>. <pub-id pub-id-type="doi">10.1093/bioinformatics/bty032</pub-id> </citation>
</ref>
<ref id="B34">
<citation citation-type="journal">
<person-group person-group-type="author">
<name>
<surname>Shen</surname>
<given-names>C.</given-names>
</name>
<name>
<surname>Ding</surname>
<given-names>Y.</given-names>
</name>
<name>
<surname>Tang</surname>
<given-names>J.</given-names>
</name>
<name>
<surname>Song</surname>
<given-names>J.</given-names>
</name>
<name>
<surname>Guo</surname>
<given-names>F.</given-names>
</name>
</person-group> (<year>2017</year>). <article-title>Identification of DNA-Protein Binding Sites through Multi-Scale Local Average Blocks on Sequence Information</article-title>. <source>Molecules</source> <volume>22</volume> (<issue>12</issue>), <fpage>2079</fpage>. <pub-id pub-id-type="doi">10.3390/molecules22122079</pub-id> </citation>
</ref>
<ref id="B35">
<citation citation-type="journal">
<person-group person-group-type="author">
<name>
<surname>Shen</surname>
<given-names>C.</given-names>
</name>
<name>
<surname>Ding</surname>
<given-names>Y.</given-names>
</name>
<name>
<surname>Tang</surname>
<given-names>J.</given-names>
</name>
<name>
<surname>Xu</surname>
<given-names>X.</given-names>
</name>
<name>
<surname>Guo</surname>
<given-names>F.</given-names>
</name>
</person-group> (<year>2017</year>). <article-title>An Ameliorated Prediction of Drug-Target Interactions Based on Multi-Scale Discrete Wavelet Transform and Network Features</article-title>. <source>Ijms</source> <volume>18</volume> (<issue>8</issue>), <fpage>1781</fpage>. <pub-id pub-id-type="doi">10.3390/ijms18081781</pub-id> </citation>
</ref>
<ref id="B36">
<citation citation-type="journal">
<person-group person-group-type="author">
<name>
<surname>Shen</surname>
<given-names>Y.</given-names>
</name>
<name>
<surname>Tang</surname>
<given-names>J.</given-names>
</name>
<name>
<surname>Guo</surname>
<given-names>F.</given-names>
</name>
</person-group> (<year>2019</year>). <article-title>Identification of Protein Subcellular Localization via Integrating Evolutionary and Physicochemical Information into Chou&#x27;s General PseAAC</article-title>. <source>J.&#x20;Theor. Biol.</source> <volume>462</volume>, <fpage>230</fpage>&#x2013;<lpage>239</lpage>. <pub-id pub-id-type="doi">10.1016/j.jtbi.2018.11.012</pub-id> </citation>
</ref>
<ref id="B37">
<citation citation-type="confproc">
<person-group person-group-type="author">
<name>
<surname>Shu</surname>
<given-names>L.</given-names>
</name>
<name>
<surname>Xu</surname>
<given-names>H.</given-names>
</name>
<name>
<surname>Liu</surname>
<given-names>B.</given-names>
</name>
</person-group> (<year>2017</year>). &#x201c;<article-title>Lifelong Learning Crf for Supervised Aspect Extraction[C]</article-title>,&#x201d; in <conf-name>Proceedings of the 55th Annual Meeting of the Association for Computational Linguistics</conf-name> (<publisher-loc>Vancouver</publisher-loc>), <fpage>148154</fpage>. </citation>
</ref>
<ref id="B38">
<citation citation-type="confproc">
<person-group person-group-type="author">
<name>
<surname>Silver</surname>
<given-names>D. L.</given-names>
</name>
<name>
<surname>Mason</surname>
<given-names>G.</given-names>
</name>
<name>
<surname>Eljabu</surname>
<given-names>L.</given-names>
</name>
</person-group> (<year>2015</year>). &#x201c;<article-title>Consolidation Using Sweep Task Rehearsal: Overcoming the Stability Plasticity Problem[C]</article-title>,&#x201d; in <conf-name>Canadian Conference on Artificial Intelligence</conf-name> (<publisher-loc>Halifax</publisher-loc>), <fpage>307322</fpage>. </citation>
</ref>
<ref id="B39">
<citation citation-type="confproc">
<person-group person-group-type="author">
<name>
<surname>Silver</surname>
<given-names>D. L.</given-names>
</name>
<name>
<surname>Mercer</surname>
<given-names>R. E.</given-names>
</name>
</person-group> (<year>2002</year>). &#x201c;<article-title>The Task Rehearsal Method of Lifelong Learning: Overcoming Impoverished Data[C]</article-title>,&#x201d; in <conf-name>Conference of the Canadian Society for Computational Studies of Intelligence</conf-name> (<publisher-loc>Calgary</publisher-loc>), <fpage>90101</fpage>. </citation>
</ref>
<ref id="B40">
<citation citation-type="journal">
<person-group person-group-type="author">
<name>
<surname>Silver</surname>
<given-names>D. L.</given-names>
</name>
</person-group> (<year>1996</year>). <article-title>The Parallel Transfer of Task Knowledge Using Dynamic Learning Rates Based on a Measure of Relatedness[J]</article-title>. <source>Connect. Sci.</source> <volume>8</volume> (<issue>2</issue>), <fpage>277294</fpage>. <pub-id pub-id-type="doi">10.1080/095400996116929</pub-id> </citation>
</ref>
<ref id="B41">
<citation citation-type="journal">
<person-group person-group-type="author">
<name>
<surname>Thrun</surname>
<given-names>S.</given-names>
</name>
<name>
<surname>Mitchell</surname>
<given-names>T. M.</given-names>
</name>
</person-group> (<year>1995</year>). <article-title>Lifelong Robot Learning[J]</article-title>. <source>Robotics Autonomous Syst.</source> <volume>15</volume> (<issue>12</issue>), <fpage>2546</fpage>. <pub-id pub-id-type="doi">10.1016/0921-8890(95)00004-Y</pub-id> </citation>
</ref>
<ref id="B42">
<citation citation-type="confproc">
<person-group person-group-type="author">
<name>
<surname>Wang</surname>
<given-names>S.</given-names>
</name>
<name>
<surname>Chen</surname>
<given-names>Z.</given-names>
</name>
<name>
<surname>Liu</surname>
<given-names>B.</given-names>
</name>
</person-group> (<year>2016</year>). &#x201c;<article-title>Mining Aspect-specific Opinion Using a Holistic Lifelong Topic Model[C]</article-title>,&#x201d; in <conf-name>Proceedings of the 25th International Conference on World Wide Web</conf-name> (<publisher-loc>Montreal</publisher-loc>), <fpage>167176</fpage>. </citation>
</ref>
<ref id="B43">
<citation citation-type="journal">
<person-group person-group-type="author">
<name>
<surname>Wang</surname>
<given-names>Y.</given-names>
</name>
<name>
<surname>Ding</surname>
<given-names>Y.</given-names>
</name>
<name>
<surname>Guo</surname>
<given-names>F.</given-names>
</name>
<name>
<surname>Wei</surname>
<given-names>L.</given-names>
</name>
<name>
<surname>Tang</surname>
<given-names>J.</given-names>
</name>
</person-group> (<year>2017</year>). <article-title>Improved Detection of Dna-Binding Proteins via Compression Technology on Pssm Information</article-title>. <source>Plos One</source> <volume>12</volume> (<issue>9</issue>), <fpage>e0185587</fpage>. <pub-id pub-id-type="doi">10.1371/journal.pone.0185587</pub-id> </citation>
</ref>
<ref id="B44">
<citation citation-type="journal">
<person-group person-group-type="author">
<name>
<surname>Wei</surname>
<given-names>L.</given-names>
</name>
<name>
<surname>Tang</surname>
<given-names>J.</given-names>
</name>
<name>
<surname>Zou</surname>
<given-names>Q.</given-names>
</name>
</person-group> (<year>2017</year>). <article-title>Local-dpp: an Improved Dna-Binding Protein Prediction Method by Exploring Local Evolutionary Information</article-title>. <source>Inf. Sci.</source> <volume>384</volume>, <fpage>135</fpage>&#x2013;<lpage>144</lpage>. <pub-id pub-id-type="doi">10.1016/j.ins.2016.06.026</pub-id> </citation>
</ref>
<ref id="B45">
<citation citation-type="journal">
<person-group person-group-type="author">
<name>
<surname>Wu</surname>
<given-names>H.</given-names>
</name>
<name>
<surname>Yang</surname>
<given-names>R.</given-names>
</name>
<name>
<surname>Fu</surname>
<given-names>Q.</given-names>
</name>
<name>
<surname>Chen</surname>
<given-names>J.</given-names>
</name>
<name>
<surname>Lu</surname>
<given-names>W.</given-names>
</name>
<name>
<surname>Li</surname>
<given-names>H.</given-names>
</name>
</person-group> (<year>2019a</year>). <article-title>Research on Predicting 2D-HP Protein Folding Using Reinforcement Learning with Full State Space</article-title>. <source>BMC Bioinformatics</source> <volume>20</volume>, <fpage>685</fpage>. <pub-id pub-id-type="doi">10.1186/s12859-019-3259-6</pub-id> </citation>
</ref>
<ref id="B46">
<citation citation-type="journal">
<person-group person-group-type="author">
<name>
<surname>Wu</surname>
<given-names>H.</given-names>
</name>
<name>
<surname>Huang</surname>
<given-names>H.</given-names>
</name>
<name>
<surname>Lu</surname>
<given-names>W.</given-names>
</name>
<name>
<surname>Fu</surname>
<given-names>Q.</given-names>
</name>
<name>
<surname>Ding</surname>
<given-names>Y.</given-names>
</name>
<name>
<surname>Qiu</surname>
<given-names>J.</given-names>
</name>
<etal/>
</person-group> (<year>2019b</year>). <article-title>Ranking Near-Native Candidate Protein Structures via Random forest Classification</article-title>. <source>BMC Bioinformatics</source> <volume>20</volume>, <fpage>683</fpage>. <pub-id pub-id-type="doi">10.1186/s12859-019-3257-8</pub-id> </citation>
</ref>
<ref id="B47">
<citation citation-type="journal">
<person-group person-group-type="author">
<name>
<surname>Yosvany</surname>
<given-names>L.</given-names>
</name>
<name>
<surname>Alok</surname>
<given-names>S.</given-names>
</name>
<name>
<surname>Abdollah</surname>
<given-names>D.</given-names>
</name>
<name>
<surname>Pranit</surname>
<given-names>L. S.</given-names>
</name>
<name>
<surname>Ghazaleh</surname>
<given-names>T.</given-names>
</name>
<name>
<surname>Abdul</surname>
<given-names>S.</given-names>
</name>
<etal/>
</person-group> (<year>2018</year>). <article-title>Success: Evolutionary and Structural Properties of Amino Acids Prove Effective for Succinylation Site Prediction</article-title>. <source>BMC Genom.</source> <volume>19</volume> (<issue>Suppl. 1</issue>), <fpage>923</fpage>. <pub-id pub-id-type="doi">10.1186/s12864-017-4336-8</pub-id> </citation>
</ref>
<ref id="B48">
<citation citation-type="journal">
<person-group person-group-type="author">
<name>
<surname>Zhou</surname>
<given-names>H.</given-names>
</name>
<name>
<surname>Wang</surname>
<given-names>H.</given-names>
</name>
<name>
<surname>Ding</surname>
<given-names>Y.</given-names>
</name>
<name>
<surname>Tang</surname>
<given-names>J.</given-names>
</name>
</person-group> (<year>2021</year>). <article-title>Multivariate Information Fusion for Identifying Antifungal Peptides with Hilbert-Schmidt Independence Criterion</article-title>. <source>Cbio</source> <volume>16</volume>. <pub-id pub-id-type="doi">10.2174/1574893616666210727161003</pub-id> </citation>
</ref>
<ref id="B49">
<citation citation-type="journal">
<person-group person-group-type="author">
<name>
<surname>Zou</surname>
<given-names>Q.</given-names>
</name>
<name>
<surname>Li</surname>
<given-names>X.</given-names>
</name>
<name>
<surname>Jiang</surname>
<given-names>Y.</given-names>
</name>
<name>
<surname>Zhao</surname>
<given-names>Y.</given-names>
</name>
<name>
<surname>Wang</surname>
<given-names>G.</given-names>
</name>
</person-group> (<year>2013</year>). <article-title>Binmempredict: a Web Server and Software for Predicting Membrane Protein Types</article-title>. <source>Cp</source> <volume>10</volume> (<issue>1</issue>), <fpage>2</fpage>&#x2013;<lpage>9</lpage>. <pub-id pub-id-type="doi">10.2174/1570164611310010002</pub-id> </citation>
</ref>
<ref id="B50">
<citation citation-type="journal">
<person-group person-group-type="author">
<name>
<surname>Zou</surname>
<given-names>Y.</given-names>
</name>
<name>
<surname>Ding</surname>
<given-names>Y.</given-names>
</name>
<name>
<surname>Peng</surname>
<given-names>L.</given-names>
</name>
<name>
<surname>Zou</surname>
<given-names>Q.</given-names>
</name>
</person-group> (<year>2021</year>). <article-title>FTWSVM-SR: DNA-Binding Proteins Identification via Fuzzy Twin Support Vector Machines on Self-Representation</article-title>. <source>Interdiscip. Sci. Comput. Life Sci.</source>. <pub-id pub-id-type="doi">10.1007/s12539-021-00489-6</pub-id> </citation>
</ref>
<ref id="B51">
<citation citation-type="journal">
<person-group person-group-type="author">
<name>
<surname>Zou</surname>
<given-names>Y.</given-names>
</name>
<name>
<surname>Wu</surname>
<given-names>H.</given-names>
</name>
<name>
<surname>Guo</surname>
<given-names>X.</given-names>
</name>
<name>
<surname>Peng</surname>
<given-names>L.</given-names>
</name>
<name>
<surname>Ding</surname>
<given-names>Y.</given-names>
</name>
<name>
<surname>Tang</surname>
<given-names>J.</given-names>
</name>
<etal/>
</person-group> (<year>2020</year>). <article-title>A Multiple Kernel-Based Fuzzy SVM Model for Predicting DNA-Binding Proteins via Support Vector Data Description</article-title>. <source>Curr. Bioinformatics</source> <volume>16</volume> (<issue>2</issue>), <fpage>274</fpage>&#x2013;<lpage>283</lpage>. </citation>
</ref>
</ref-list>
</back>
</article>