<?xml version="1.0" encoding="UTF-8"?>
<!DOCTYPE article PUBLIC "-//NLM//DTD Journal Publishing DTD v2.3 20070202//EN" "journalpublishing.dtd">
<article article-type="research-article" dtd-version="2.3" xml:lang="EN" xmlns:mml="http://www.w3.org/1998/Math/MathML" xmlns:xlink="http://www.w3.org/1999/xlink">
<front>
<journal-meta>
<journal-id journal-id-type="publisher-id">Front. Appl. Math. Stat.</journal-id>
<journal-title>Frontiers in Applied Mathematics and Statistics</journal-title>
<abbrev-journal-title abbrev-type="pubmed">Front. Appl. Math. Stat.</abbrev-journal-title>
<issn pub-type="epub">2297-4687</issn>
<publisher>
<publisher-name>Frontiers Media S.A.</publisher-name>
</publisher>
</journal-meta>
<article-meta>
<article-id pub-id-type="publisher-id">693126</article-id>
<article-id pub-id-type="doi">10.3389/fams.2021.693126</article-id>
<article-categories>
<subj-group subj-group-type="heading">
<subject>Applied Mathematics and Statistics</subject>
<subj-group>
<subject>Original Research</subject>
</subj-group>
</subj-group>
</article-categories>
<title-group>
<article-title>New Developments in Sparse PLS Regression</article-title>
<alt-title alt-title-type="left-running-head">Magnanensi et&#x20;al.</alt-title>
<alt-title alt-title-type="right-running-head">New Developments in sPLS Regression</alt-title>
</title-group>
<contrib-group>
<contrib contrib-type="author">
<name>
<surname>Magnanensi</surname>
<given-names>J&#xe9;r&#xe9;my</given-names>
</name>
<xref ref-type="aff" rid="aff1">
<sup>1</sup>
</xref>
<xref ref-type="aff" rid="aff2">
<sup>2</sup>
</xref>
</contrib>
<contrib contrib-type="author">
<name>
<surname>Maumy-Bertrand</surname>
<given-names>Myriam</given-names>
</name>
<xref ref-type="aff" rid="aff1">
<sup>1</sup>
</xref>
<xref ref-type="aff" rid="aff3">
<sup>3</sup>
</xref>
</contrib>
<contrib contrib-type="author">
<name>
<surname>Meyer</surname>
<given-names>Nicolas</given-names>
</name>
<xref ref-type="aff" rid="aff2">
<sup>2</sup>
</xref>
</contrib>
<contrib contrib-type="author" corresp="yes">
<name>
<surname>Bertrand</surname>
<given-names>Fr&#xe9;d&#xe9;ric</given-names>
</name>
<xref ref-type="aff" rid="aff1">
<sup>1</sup>
</xref>
<xref ref-type="aff" rid="aff3">
<sup>3</sup>
</xref>
<xref ref-type="corresp" rid="c001">&#x2a;</xref>
<uri xlink:href="https://loop.frontiersin.org/people/1270416/overview"/>
</contrib>
</contrib-group>
<aff id="aff1">
<label>
<sup>1</sup>
</label>IRMA, CNRS UMR 7501, Labex IRMIA, Universit&#xe9; de Strasbourg and CNRS, <addr-line>Strasbourg</addr-line>, <country>France</country>
</aff>
<aff id="aff2">
<label>
<sup>2</sup>
</label>Laboratoire de Biostatistique et Informatique M&#xe9;dicale, Facult&#xe9; de M&#xe9;decine, EA3430, Universit&#xe9; de Strasbourg and CNRS, <addr-line>Strasbourg</addr-line>, <country>France</country>
</aff>
<aff id="aff3">
<label>
<sup>3</sup>
</label>LIST3N, Universit&#xe9; de Technologie de Troyes, <addr-line>Troyes</addr-line>, <country>France</country>
</aff>
<author-notes>
<fn fn-type="edited-by">
<p>
<bold>Edited by:</bold> <ext-link ext-link-type="uri" xlink:href="https://loop.frontiersin.org/people/1126913/overview">Hong Zhu</ext-link>, Jiangsu University, China</p>
</fn>
<fn fn-type="edited-by">
<p>
<bold>Reviewed by:</bold> <ext-link ext-link-type="uri" xlink:href="https://loop.frontiersin.org/people/506351/overview">Yunlong Feng</ext-link>, University at Albany, United&#x20;States</p>
<p>
<ext-link ext-link-type="uri" xlink:href="https://loop.frontiersin.org/people/1058596/overview">Deepak Gupta</ext-link>, National Institute of Technology, Arunachal Pradesh, India</p>
</fn>
<corresp id="c001">&#x2a;Correspondence: Fr&#xe9;d&#xe9;ric Bertrand, <email>frederic.bertrand@utt.fr</email>
</corresp>
<fn fn-type="other">
<p>This article was submitted to Mathematics of Computation and Data Science, a section of the journal Frontiers in Applied Mathematics and Statistics</p>
</fn>
</author-notes>
<pub-date pub-type="epub">
<day>16</day>
<month>07</month>
<year>2021</year>
</pub-date>
<pub-date pub-type="collection">
<year>2021</year>
</pub-date>
<volume>7</volume>
<elocation-id>693126</elocation-id>
<history>
<date date-type="received">
<day>09</day>
<month>04</month>
<year>2021</year>
</date>
<date date-type="accepted">
<day>21</day>
<month>06</month>
<year>2021</year>
</date>
</history>
<permissions>
<copyright-statement>Copyright &#xa9; 2021 Magnanensi, Maumy-Bertrand, Meyer and Bertrand.</copyright-statement>
<copyright-year>2021</copyright-year>
<copyright-holder>Magnanensi, Maumy-Bertrand, Meyer and Bertrand</copyright-holder>
<license xlink:href="http://creativecommons.org/licenses/by/4.0/">
<p>This is an open-access article distributed under the terms of the Creative Commons Attribution License (CC BY). The use, distribution or reproduction in other forums is permitted, provided the original author(s) and the copyright owner(s) are credited and that the original publication in this journal is cited, in accordance with accepted academic practice. No use, distribution or reproduction is permitted which does not comply with these&#x20;terms.</p>
</license>
</permissions>
<abstract>
<p>Methods based on partial least squares (PLS) regression, which has recently gained much attention in the analysis of high-dimensional genomic datasets, have been developed since the early 2000s for performing variable selection. Most of these techniques rely on tuning parameters that are often determined by cross-validation (CV) based methods, which raises essential stability issues. To overcome this, we have developed a new dynamic bootstrap-based method for significant predictor selection, suitable for both PLS regression and its incorporation into generalized linear models (GPLS). It relies on establishing bootstrap confidence intervals, which allows testing of the significance of predictors at preset type I risk <italic>&#x3b1;</italic>, and avoids CV. We have also developed adapted versions of sparse PLS (SPLS) and sparse GPLS regression (SGPLS), using a recently introduced non-parametric bootstrap-based technique to determine the numbers of components. We compare their variable selection reliability and stability concerning tuning parameters determination and their predictive ability, using simulated data for PLS and real microarray gene expression data for PLS-logistic classification. We observe that our new dynamic bootstrap-based method has the property of best separating random noise in <bold>y</bold> from the relevant information with respect to other methods, leading to better accuracy and predictive abilities, especially for non-negligible noise levels.</p>
</abstract>
<kwd-group>
<kwd>variable selection</kwd>
<kwd>partial least squares</kwd>
<kwd>sparse partial least squares</kwd>
<kwd>generalized partial least squares</kwd>
<kwd>bootstrap</kwd>
<kwd>stability</kwd>
</kwd-group>
<contract-num rid="cn001">ANR-11-LABX-0055 IRMIA</contract-num>
<contract-sponsor id="cn001">Agence Nationale de la Recherche10.13039/501100001665</contract-sponsor>
<contract-sponsor id="cn002">Centre National de la Recherche Scientifique10.13039/501100004794</contract-sponsor>
</article-meta>
</front>
<body>
<sec id="s1">
<title>1. Introduction</title>
<p>Partial least squares (PLS) regression, introduced by [<xref ref-type="bibr" rid="B1">1</xref>], is a well-known dimension-reduction method, notably in chemometrics and spectrometric modeling [<xref ref-type="bibr" rid="B2">2</xref>]. In this paper, we focus on the PLS univariate response framework, better known as PLS1. Let <italic>n</italic> be the number of observations and <italic>p</italic> the number of covariates. Then, <inline-formula id="inf1">
<mml:math id="m1">
<mml:mrow>
<mml:mi mathvariant="bold">y</mml:mi>
<mml:mo>&#x3d;</mml:mo>
<mml:msup>
<mml:mrow>
<mml:mrow>
<mml:mo>(</mml:mo>
<mml:mrow>
<mml:msub>
<mml:mi>y</mml:mi>
<mml:mn>1</mml:mn>
</mml:msub>
<mml:mo>,</mml:mo>
<mml:mo>&#x2026;</mml:mo>
<mml:mo>,</mml:mo>
<mml:msub>
<mml:mi>y</mml:mi>
<mml:mi>n</mml:mi>
</mml:msub>
</mml:mrow>
<mml:mo>)</mml:mo>
</mml:mrow>
</mml:mrow>
<mml:mi>T</mml:mi>
</mml:msup>
<mml:mo>&#x2208;</mml:mo>
<mml:msup>
<mml:mi mathvariant="normal">&#x211d;</mml:mi>
<mml:mi>n</mml:mi>
</mml:msup>
</mml:mrow>
</mml:math>
</inline-formula> represents the response vector, with <inline-formula id="inf2">
<mml:math id="m2">
<mml:mrow>
<mml:msup>
<mml:mrow>
<mml:mrow>
<mml:mo>(</mml:mo>
<mml:mo>.</mml:mo>
<mml:mo>)</mml:mo>
</mml:mrow>
</mml:mrow>
<mml:mi>T</mml:mi>
</mml:msup>
</mml:mrow>
</mml:math>
</inline-formula> denoting the transpose. The original underlying algorithm, developed to deal with continuous responses, consists of building latent variables <inline-formula id="inf3">
<mml:math id="m3">
<mml:mrow>
<mml:msub>
<mml:mi mathvariant="bold">t</mml:mi>
<mml:mi>k</mml:mi>
</mml:msub>
<mml:mo>,</mml:mo>
<mml:mtext>&#x2002;</mml:mtext>
<mml:mn>1</mml:mn>
<mml:mi mathvariant="normal">&#x2a7d;</mml:mi>
<mml:mi>k</mml:mi>
<mml:mi mathvariant="normal">&#x2a7d;</mml:mi>
<mml:mi>K</mml:mi>
</mml:mrow>
</mml:math>
</inline-formula>, also called components, as linear combinations of the original predictors <inline-formula id="inf4">
<mml:math id="m4">
<mml:mrow>
<mml:mi mathvariant="bold">X</mml:mi>
<mml:mo>&#x3d;</mml:mo>
<mml:mrow>
<mml:mo>(</mml:mo>
<mml:mrow>
<mml:msub>
<mml:mi mathvariant="bold">x</mml:mi>
<mml:mn>1</mml:mn>
</mml:msub>
<mml:mo>,</mml:mo>
<mml:mo>&#x2026;</mml:mo>
<mml:mo>,</mml:mo>
<mml:msub>
<mml:mi mathvariant="bold">x</mml:mi>
<mml:mi>p</mml:mi>
</mml:msub>
</mml:mrow>
<mml:mo>)</mml:mo>
</mml:mrow>
<mml:mo>&#x2208;</mml:mo>
<mml:msub>
<mml:mi mathvariant="normal">&#x2133;</mml:mi>
<mml:mrow>
<mml:mi>n</mml:mi>
<mml:mo>,</mml:mo>
<mml:mi>p</mml:mi>
</mml:mrow>
</mml:msub>
<mml:mrow>
<mml:mo>(</mml:mo>
<mml:mi mathvariant="normal">&#x211d;</mml:mi>
<mml:mo>)</mml:mo>
</mml:mrow>
</mml:mrow>
</mml:math>
</inline-formula>, where <inline-formula id="inf5">
<mml:math id="m5">
<mml:mrow>
<mml:msub>
<mml:mi mathvariant="normal">&#x2133;</mml:mi>
<mml:mrow>
<mml:mi>n</mml:mi>
<mml:mo>,</mml:mo>
<mml:mi>p</mml:mi>
</mml:mrow>
</mml:msub>
<mml:mrow>
<mml:mo>(</mml:mo>
<mml:mi mathvariant="normal">&#x211d;</mml:mi>
<mml:mo>)</mml:mo>
</mml:mrow>
</mml:mrow>
</mml:math>
</inline-formula> represents the set of matrices of <italic>n</italic> rows and <italic>p</italic> columns. Thus,<disp-formula id="e1">
<mml:math id="m6">
<mml:mrow>
<mml:msub>
<mml:mi mathvariant="bold">t</mml:mi>
<mml:mi>k</mml:mi>
</mml:msub>
<mml:mo>&#x3d;</mml:mo>
<mml:msub>
<mml:mi mathvariant="bold">X</mml:mi>
<mml:mrow>
<mml:mi>k</mml:mi>
<mml:mo>&#x2212;</mml:mo>
<mml:mn>1</mml:mn>
</mml:mrow>
</mml:msub>
<mml:msub>
<mml:mi mathvariant="bold">w</mml:mi>
<mml:mi>k</mml:mi>
</mml:msub>
<mml:mo>,</mml:mo>
<mml:mtext>&#x2002;</mml:mtext>
<mml:mn>1</mml:mn>
<mml:mi mathvariant="normal">&#x2a7d;</mml:mi>
<mml:mi>k</mml:mi>
<mml:mi mathvariant="normal">&#x2a7d;</mml:mi>
<mml:mi>K</mml:mi>
<mml:mo>,</mml:mo>
</mml:mrow>
</mml:math>
<label>(1)</label>
</disp-formula>where <inline-formula id="inf6">
<mml:math id="m7">
<mml:mrow>
<mml:msub>
<mml:mi mathvariant="bold">X</mml:mi>
<mml:mn>0</mml:mn>
</mml:msub>
<mml:mo>&#x3d;</mml:mo>
<mml:mi mathvariant="bold">X</mml:mi>
</mml:mrow>
</mml:math>
</inline-formula>, and <inline-formula id="inf7">
<mml:math id="m8">
<mml:mrow>
<mml:msub>
<mml:mi mathvariant="bold">X</mml:mi>
<mml:mrow>
<mml:mi>k</mml:mi>
<mml:mo>&#x2212;</mml:mo>
<mml:mn>1</mml:mn>
</mml:mrow>
</mml:msub>
<mml:mo>,</mml:mo>
<mml:mtext>&#x2002;</mml:mtext>
<mml:mi>k</mml:mi>
<mml:mi mathvariant="normal">&#x2a7e;</mml:mi>
<mml:mn>2</mml:mn>
</mml:mrow>
</mml:math>
</inline-formula> represents the residual covariate matrix obtained through the ordinary least squares regression (OLSR) of <inline-formula id="inf8">
<mml:math id="m9">
<mml:mrow>
<mml:msub>
<mml:mi mathvariant="bold">X</mml:mi>
<mml:mrow>
<mml:mi>k</mml:mi>
<mml:mo>&#x2212;</mml:mo>
<mml:mn>2</mml:mn>
</mml:mrow>
</mml:msub>
</mml:mrow>
</mml:math>
</inline-formula> on <inline-formula id="inf9">
<mml:math id="m10">
<mml:mrow>
<mml:msub>
<mml:mi mathvariant="bold">t</mml:mi>
<mml:mrow>
<mml:mi>k</mml:mi>
<mml:mo>&#x2212;</mml:mo>
<mml:mn>1</mml:mn>
</mml:mrow>
</mml:msub>
</mml:mrow>
</mml:math>
</inline-formula>. Here, <inline-formula id="inf10">
<mml:math id="m11">
<mml:mrow>
<mml:msub>
<mml:mi mathvariant="bold">w</mml:mi>
<mml:mi>k</mml:mi>
</mml:msub>
<mml:mo>&#x3d;</mml:mo>
<mml:msup>
<mml:mrow>
<mml:mrow>
<mml:mo>(</mml:mo>
<mml:mrow>
<mml:msub>
<mml:mi>w</mml:mi>
<mml:mrow>
<mml:mn>1</mml:mn>
<mml:mi>k</mml:mi>
</mml:mrow>
</mml:msub>
<mml:mo>,</mml:mo>
<mml:mo>&#x2026;</mml:mo>
<mml:mo>,</mml:mo>
<mml:msub>
<mml:mi>w</mml:mi>
<mml:mrow>
<mml:mi>p</mml:mi>
<mml:mi>k</mml:mi>
</mml:mrow>
</mml:msub>
</mml:mrow>
<mml:mo>)</mml:mo>
</mml:mrow>
</mml:mrow>
<mml:mi>T</mml:mi>
</mml:msup>
<mml:mo>&#x2208;</mml:mo>
<mml:msup>
<mml:mi mathvariant="normal">&#x211d;</mml:mi>
<mml:mi>p</mml:mi>
</mml:msup>
</mml:mrow>
</mml:math>
</inline-formula> is obtained as the solution of the following maximization problem [<xref ref-type="bibr" rid="B3">3</xref>]:<disp-formula id="e2">
<mml:math id="m12">
<mml:mrow>
<mml:msub>
<mml:mi mathvariant="bold">w</mml:mi>
<mml:mi>k</mml:mi>
</mml:msub>
<mml:mo>&#x3d;</mml:mo>
<mml:munder>
<mml:mrow>
<mml:mtext>argmax</mml:mtext>
</mml:mrow>
<mml:mrow>
<mml:mi mathvariant="bold">w</mml:mi>
<mml:mo>&#x2208;</mml:mo>
<mml:msup>
<mml:mi mathvariant="normal">&#x211d;</mml:mi>
<mml:mi>p</mml:mi>
</mml:msup>
</mml:mrow>
</mml:munder>
<mml:mrow>
<mml:mo>{</mml:mo>
<mml:mrow>
<mml:msup>
<mml:mrow>
<mml:mtext>Cov</mml:mtext>
</mml:mrow>
<mml:mn>2</mml:mn>
</mml:msup>
<mml:mrow>
<mml:mo>(</mml:mo>
<mml:mrow>
<mml:msub>
<mml:mi mathvariant="bold">y</mml:mi>
<mml:mrow>
<mml:mi>k</mml:mi>
<mml:mo>&#x2212;</mml:mo>
<mml:mn>1</mml:mn>
</mml:mrow>
</mml:msub>
<mml:mo>,</mml:mo>
<mml:msub>
<mml:mi mathvariant="bold">t</mml:mi>
<mml:mi>k</mml:mi>
</mml:msub>
</mml:mrow>
<mml:mo>)</mml:mo>
</mml:mrow>
</mml:mrow>
<mml:mo>}</mml:mo>
</mml:mrow>
</mml:mrow>
</mml:math>
<label>(2)</label>
</disp-formula>
<disp-formula id="e3">
<mml:math id="m13">
<mml:mrow>
<mml:mo>&#x3d;</mml:mo>
<mml:munder>
<mml:mrow>
<mml:mtext>argmax</mml:mtext>
</mml:mrow>
<mml:mrow>
<mml:mi mathvariant="bold">w</mml:mi>
<mml:mo>&#x2208;</mml:mo>
<mml:msup>
<mml:mi mathvariant="normal">&#x211d;</mml:mi>
<mml:mi>p</mml:mi>
</mml:msup>
</mml:mrow>
</mml:munder>
<mml:mrow>
<mml:mo>{</mml:mo>
<mml:mrow>
<mml:msup>
<mml:mi mathvariant="bold">w</mml:mi>
<mml:mi>T</mml:mi>
</mml:msup>
<mml:msubsup>
<mml:mi mathvariant="bold">X</mml:mi>
<mml:mrow>
<mml:mi>k</mml:mi>
<mml:mo>&#x2212;</mml:mo>
<mml:mn>1</mml:mn>
</mml:mrow>
<mml:mi>T</mml:mi>
</mml:msubsup>
<mml:msub>
<mml:mi mathvariant="bold">y</mml:mi>
<mml:mrow>
<mml:mi>k</mml:mi>
<mml:mo>&#x2212;</mml:mo>
<mml:mn>1</mml:mn>
</mml:mrow>
</mml:msub>
<mml:msubsup>
<mml:mi mathvariant="bold">y</mml:mi>
<mml:mrow>
<mml:mi>k</mml:mi>
<mml:mo>&#x2212;</mml:mo>
<mml:mn>1</mml:mn>
</mml:mrow>
<mml:mi>T</mml:mi>
</mml:msubsup>
<mml:msub>
<mml:mi mathvariant="bold">X</mml:mi>
<mml:mrow>
<mml:mi>k</mml:mi>
<mml:mo>&#x2212;</mml:mo>
<mml:mn>1</mml:mn>
</mml:mrow>
</mml:msub>
<mml:mi mathvariant="bold">w</mml:mi>
</mml:mrow>
<mml:mo>}</mml:mo>
</mml:mrow>
<mml:mo>,</mml:mo>
</mml:mrow>
</mml:math>
<label>(3)</label>
</disp-formula>with the constraint <inline-formula id="inf11">
<mml:math id="m14">
<mml:mrow>
<mml:msubsup>
<mml:mrow>
<mml:mrow>
<mml:mo>&#x2016;</mml:mo>
<mml:mrow>
<mml:msub>
<mml:mi mathvariant="bold">w</mml:mi>
<mml:mi>k</mml:mi>
</mml:msub>
</mml:mrow>
<mml:mo>&#x2016;</mml:mo>
</mml:mrow>
</mml:mrow>
<mml:mn>2</mml:mn>
<mml:mn>2</mml:mn>
</mml:msubsup>
<mml:mo>&#x3d;</mml:mo>
<mml:mn>1</mml:mn>
</mml:mrow>
</mml:math>
</inline-formula>, and where <inline-formula id="inf12">
<mml:math id="m15">
<mml:mrow>
<mml:msub>
<mml:mi mathvariant="bold">y</mml:mi>
<mml:mn>0</mml:mn>
</mml:msub>
<mml:mo>&#x3d;</mml:mo>
<mml:mi mathvariant="bold">y</mml:mi>
</mml:mrow>
</mml:math>
</inline-formula>, and <inline-formula id="inf13">
<mml:math id="m16">
<mml:mrow>
<mml:msub>
<mml:mi mathvariant="bold">y</mml:mi>
<mml:mrow>
<mml:mi>k</mml:mi>
<mml:mo>&#x2212;</mml:mo>
<mml:mn>1</mml:mn>
</mml:mrow>
</mml:msub>
<mml:mo>,</mml:mo>
<mml:mtext>&#x2002;</mml:mtext>
<mml:mi>k</mml:mi>
<mml:mi mathvariant="normal">&#x2a7e;</mml:mi>
<mml:mn>2</mml:mn>
</mml:mrow>
</mml:math>
</inline-formula> represents the residual vector obtained from the OLSR of <inline-formula id="inf14">
<mml:math id="m17">
<mml:mrow>
<mml:msub>
<mml:mi mathvariant="bold">y</mml:mi>
<mml:mrow>
<mml:mi>k</mml:mi>
<mml:mo>&#x2212;</mml:mo>
<mml:mn>2</mml:mn>
</mml:mrow>
</mml:msub>
</mml:mrow>
</mml:math>
</inline-formula> on <inline-formula id="inf15">
<mml:math id="m18">
<mml:mrow>
<mml:msub>
<mml:mi mathvariant="bold">t</mml:mi>
<mml:mrow>
<mml:mi>k</mml:mi>
<mml:mo>&#x2212;</mml:mo>
<mml:mn>1</mml:mn>
</mml:mrow>
</mml:msub>
</mml:mrow>
</mml:math>
</inline-formula>.</p>
<p>The final regression model is thus:<disp-formula id="e4">
<mml:math id="m19">
<mml:mrow>
<mml:mi mathvariant="bold">y</mml:mi>
<mml:mo>&#x3d;</mml:mo>
<mml:mstyle displaystyle="true">
<mml:munderover>
<mml:mo>&#x2211;</mml:mo>
<mml:mrow>
<mml:mi>k</mml:mi>
<mml:mo>&#x3d;</mml:mo>
<mml:mn>1</mml:mn>
</mml:mrow>
<mml:mi>K</mml:mi>
</mml:munderover>
<mml:mrow>
<mml:msub>
<mml:mi>c</mml:mi>
<mml:mi>k</mml:mi>
</mml:msub>
<mml:msub>
<mml:mi mathvariant="bold">t</mml:mi>
<mml:mi>k</mml:mi>
</mml:msub>
<mml:mo>&#x2b;</mml:mo>
<mml:mi mathvariant="normal">&#x3f5;</mml:mi>
</mml:mrow>
</mml:mstyle>
</mml:mrow>
</mml:math>
<label>(4)</label>
</disp-formula>
<disp-formula id="e5">
<mml:math id="m20">
<mml:mrow>
<mml:mo>&#x3d;</mml:mo>
<mml:mstyle displaystyle="true">
<mml:munderover>
<mml:mo>&#x2211;</mml:mo>
<mml:mrow>
<mml:mi>j</mml:mi>
<mml:mo>&#x3d;</mml:mo>
<mml:mn>1</mml:mn>
</mml:mrow>
<mml:mi>p</mml:mi>
</mml:munderover>
<mml:mrow>
<mml:msubsup>
<mml:mi>&#x3b2;</mml:mi>
<mml:mi>j</mml:mi>
<mml:mrow>
<mml:mtext>PLS</mml:mtext>
</mml:mrow>
</mml:msubsup>
<mml:msub>
<mml:mi mathvariant="bold">x</mml:mi>
<mml:mi>j</mml:mi>
</mml:msub>
<mml:mo>&#x2b;</mml:mo>
<mml:mi mathvariant="normal">&#x3f5;</mml:mi>
<mml:mo>,</mml:mo>
</mml:mrow>
</mml:mstyle>
</mml:mrow>
</mml:math>
<label>(5)</label>
</disp-formula>with <inline-formula id="inf16">
<mml:math id="m21">
<mml:mrow>
<mml:mi>&#x3f5;</mml:mi>
<mml:mo>&#x3d;</mml:mo>
<mml:msup>
<mml:mrow>
<mml:mrow>
<mml:mo>(</mml:mo>
<mml:mrow>
<mml:msub>
<mml:mi mathvariant="normal">&#x3f5;</mml:mi>
<mml:mn>1</mml:mn>
</mml:msub>
<mml:mo>,</mml:mo>
<mml:mo>&#x2026;</mml:mo>
<mml:mo>,</mml:mo>
<mml:msub>
<mml:mi mathvariant="normal">&#x3f5;</mml:mi>
<mml:mi>n</mml:mi>
</mml:msub>
</mml:mrow>
<mml:mo>)</mml:mo>
</mml:mrow>
</mml:mrow>
<mml:mi>T</mml:mi>
</mml:msup>
</mml:mrow>
</mml:math>
</inline-formula> the <inline-formula id="inf17">
<mml:math id="m22">
<mml:mrow>
<mml:mi>n</mml:mi>
<mml:mo>&#xd7;</mml:mo>
<mml:mn>1</mml:mn>
</mml:mrow>
</mml:math>
</inline-formula> error vector and <inline-formula id="inf18">
<mml:math id="m23">
<mml:mrow>
<mml:mrow>
<mml:mo>(</mml:mo>
<mml:mrow>
<mml:msub>
<mml:mi>c</mml:mi>
<mml:mn>1</mml:mn>
</mml:msub>
<mml:mo>,</mml:mo>
<mml:mo>&#x2026;</mml:mo>
<mml:mo>,</mml:mo>
<mml:msub>
<mml:mi>c</mml:mi>
<mml:mi>K</mml:mi>
</mml:msub>
</mml:mrow>
<mml:mo>)</mml:mo>
</mml:mrow>
</mml:mrow>
</mml:math>
</inline-formula> the regression coefficients related to the OLSR of <inline-formula id="inf19">
<mml:math id="m24">
<mml:mrow>
<mml:msub>
<mml:mi mathvariant="bold">y</mml:mi>
<mml:mrow>
<mml:mi>k</mml:mi>
<mml:mo>&#x2212;</mml:mo>
<mml:mn>1</mml:mn>
</mml:mrow>
</mml:msub>
</mml:mrow>
</mml:math>
</inline-formula> on <inline-formula id="inf20">
<mml:math id="m25">
<mml:mrow>
<mml:msub>
<mml:mi mathvariant="bold">t</mml:mi>
<mml:mi>k</mml:mi>
</mml:msub>
</mml:mrow>
</mml:math>
</inline-formula>, <inline-formula id="inf21">
<mml:math id="m26">
<mml:mrow>
<mml:mo>&#x2200;</mml:mo>
<mml:mi>k</mml:mi>
<mml:mo>&#x2208;</mml:mo>
<mml:mrow>
<mml:mo>&#x301a;</mml:mo>
<mml:mrow>
<mml:mn>1</mml:mn>
<mml:mo>,</mml:mo>
<mml:mi>K</mml:mi>
</mml:mrow>
<mml:mo>&#x301b;</mml:mo>
</mml:mrow>
</mml:mrow>
</mml:math>
</inline-formula>, also known as <bold>y</bold>-loadings. More details are available, notably in H&#xf6;skuldsson [<xref ref-type="bibr" rid="B4">4</xref>] and Tenenhaus&#x20;[<xref ref-type="bibr" rid="B5">5</xref>].</p>
<p>This particular regression technique, based on reductions in the original dimension, avoids matrix inversion and diagonalization, using only deflation. It allows us to deal with high-dimensional datasets efficiently, and notably solves the collinearity problem&#x20;[<xref ref-type="bibr" rid="B6">6</xref>].</p>
<p>With great technological advances in recent decades, PLS regression has been gaining attention, and has been successfully applied to many domains, notably genomics. Indeed, the development of both microarray and allelotyping techniques result in high-dimensional datasets from which information has to be efficiently extracted. To this end, PLS regression has become a benchmark as an efficient statistical method for prediction, regression and dimension reduction [<xref ref-type="bibr" rid="B3">3</xref>]. Practically speaking, the observed response related to such studies does commonly not follow a continuous distribution. Frequent goals with gene expression datasets involve classification problems, such as cancer stage prediction, disease relapse, and tumor classification. For such reasons, PLS regression has had to be adapted to take into account discrete responses. This has been an intensive research subject in recent years, leading globally to two types of adapted PLS regression for classification. The first, studied and developed notably by Nguyen and Rocke [<xref ref-type="bibr" rid="B7">7</xref>], Nguyen and Rocke [<xref ref-type="bibr" rid="B8">8</xref>] and Boulesteix [<xref ref-type="bibr" rid="B9">9</xref>], is a two-step method. The first step consists of building standard PLS components by treating the response as continuous. In the second step, classification methods are run, e.g., logistic discrimination (LD) or quadratic discriminant analysis (QDA). The second type of adapted PLS regression consists of building PLS components using either an adapted version of or the original iteratively reweighted least squares (IRLS) algorithm, followed by generalized linear regression on these components. This type of method was first introduced by Marx [<xref ref-type="bibr" rid="B10">10</xref>]. Different modifications and improvements, using notably ridge regression [<xref ref-type="bibr" rid="B11">11</xref>] or Firth&#x2019;s procedure [<xref ref-type="bibr" rid="B12">12</xref>] to avoid non-convergence and infinite parameter-value estimations, have been developed, notably by Nguyen and Rocke [<xref ref-type="bibr" rid="B13">13</xref>], Ding and Gentleman [<xref ref-type="bibr" rid="B14">14</xref>], Fort and Lambert-Lacroix [<xref ref-type="bibr" rid="B15">15</xref>] and Bastien et&#x20;al. [<xref ref-type="bibr" rid="B16">16</xref>]. In this work, we focus on the second type of adapted PLS regression, referred to from now on as&#x20;GPLS.</p>
<p>As previously mentioned, a feature of datasets of interest is their high-dimensional setting, i.e.,&#x20;<inline-formula id="inf22">
<mml:math id="m27">
<mml:mrow>
<mml:mi>n</mml:mi>
<mml:mo>&#x226a;</mml:mo>
<mml:mi>p</mml:mi>
</mml:mrow>
</mml:math>
</inline-formula>. Chun and Kele&#x15f; [<xref ref-type="bibr" rid="B17">17</xref>] have shown that the asymptotic consistency of PLS estimators does not hold in this situation, so filtering or predictor selection become necessary in order to obtain consistent parameter estimation. However, all methods described above proceed to classification using the entire set of predictors. For datasets that frequently contain thousands of predictors, such as microarray or RNAseq ones, a variable filtering pre-processing thus needs to be applied. A commonly used pre-processing method when performing classification uses the BSS/WSS-statistic:<disp-formula id="e6">
<mml:math id="m28">
<mml:mrow>
<mml:mrow>
<mml:mrow>
<mml:msub>
<mml:mrow>
<mml:mtext>BSS</mml:mtext>
</mml:mrow>
<mml:mi>j</mml:mi>
</mml:msub>
</mml:mrow>
<mml:mo>/</mml:mo>
<mml:mrow>
<mml:msub>
<mml:mrow>
<mml:mtext>WSS</mml:mtext>
</mml:mrow>
<mml:mi>j</mml:mi>
</mml:msub>
</mml:mrow>
</mml:mrow>
<mml:mo>&#x3d;</mml:mo>
<mml:mfrac>
<mml:mrow>
<mml:mstyle displaystyle="true">
<mml:munderover>
<mml:mo>&#x2211;</mml:mo>
<mml:mrow>
<mml:mi>q</mml:mi>
<mml:mo>&#x3d;</mml:mo>
<mml:mn>1</mml:mn>
</mml:mrow>
<mml:mi>Q</mml:mi>
</mml:munderover>
<mml:mrow>
<mml:mstyle displaystyle="true">
<mml:munder>
<mml:mo>&#x2211;</mml:mo>
<mml:mrow>
<mml:mi>i</mml:mi>
<mml:mo>:</mml:mo>
<mml:msub>
<mml:mi>y</mml:mi>
<mml:mi>i</mml:mi>
</mml:msub>
<mml:mo>&#x2208;</mml:mo>
<mml:msub>
<mml:mi>G</mml:mi>
<mml:mi>q</mml:mi>
</mml:msub>
</mml:mrow>
</mml:munder>
<mml:mrow>
<mml:msup>
<mml:mrow>
<mml:mrow>
<mml:mo>(</mml:mo>
<mml:mrow>
<mml:msub>
<mml:mrow>
<mml:mover accent="true">
<mml:mi>&#x3bc;</mml:mi>
<mml:mo>&#x5e;</mml:mo>
</mml:mover>
</mml:mrow>
<mml:mrow>
<mml:mi>j</mml:mi>
<mml:mi>q</mml:mi>
</mml:mrow>
</mml:msub>
<mml:mo>&#x2212;</mml:mo>
<mml:msub>
<mml:mrow>
<mml:mover accent="true">
<mml:mi>&#x3bc;</mml:mi>
<mml:mo>&#x5e;</mml:mo>
</mml:mover>
</mml:mrow>
<mml:mi>j</mml:mi>
</mml:msub>
</mml:mrow>
<mml:mo>)</mml:mo>
</mml:mrow>
</mml:mrow>
<mml:mn>2</mml:mn>
</mml:msup>
</mml:mrow>
</mml:mstyle>
</mml:mrow>
</mml:mstyle>
</mml:mrow>
<mml:mrow>
<mml:mstyle displaystyle="true">
<mml:munderover>
<mml:mo>&#x2211;</mml:mo>
<mml:mrow>
<mml:mi>q</mml:mi>
<mml:mo>&#x3d;</mml:mo>
<mml:mn>1</mml:mn>
</mml:mrow>
<mml:mi>Q</mml:mi>
</mml:munderover>
<mml:mrow>
<mml:mstyle displaystyle="true">
<mml:munder>
<mml:mo>&#x2211;</mml:mo>
<mml:mrow>
<mml:mi>i</mml:mi>
<mml:mo>:</mml:mo>
<mml:msub>
<mml:mi>y</mml:mi>
<mml:mi>i</mml:mi>
</mml:msub>
<mml:mo>&#x2208;</mml:mo>
<mml:msub>
<mml:mi>G</mml:mi>
<mml:mi>q</mml:mi>
</mml:msub>
</mml:mrow>
</mml:munder>
<mml:mrow>
<mml:msup>
<mml:mrow>
<mml:mrow>
<mml:mo>(</mml:mo>
<mml:mrow>
<mml:msub>
<mml:mi>x</mml:mi>
<mml:mrow>
<mml:mi>i</mml:mi>
<mml:mi>j</mml:mi>
</mml:mrow>
</mml:msub>
<mml:mo>&#x2212;</mml:mo>
<mml:msub>
<mml:mrow>
<mml:mover accent="true">
<mml:mi>&#x3bc;</mml:mi>
<mml:mo>&#x5e;</mml:mo>
</mml:mover>
</mml:mrow>
<mml:mrow>
<mml:mi>j</mml:mi>
<mml:mi>q</mml:mi>
</mml:mrow>
</mml:msub>
</mml:mrow>
<mml:mo>)</mml:mo>
</mml:mrow>
</mml:mrow>
<mml:mn>2</mml:mn>
</mml:msup>
</mml:mrow>
</mml:mstyle>
</mml:mrow>
</mml:mstyle>
</mml:mrow>
</mml:mfrac>
<mml:mo>,</mml:mo>
</mml:mrow>
</mml:math>
<label>(6)</label>
</disp-formula>with <inline-formula id="inf23">
<mml:math id="m29">
<mml:mrow>
<mml:msub>
<mml:mrow>
<mml:mover accent="true">
<mml:mi>&#x3bc;</mml:mi>
<mml:mo>&#x5e;</mml:mo>
</mml:mover>
</mml:mrow>
<mml:mi>j</mml:mi>
</mml:msub>
</mml:mrow>
</mml:math>
</inline-formula> the sample mean of <inline-formula id="inf24">
<mml:math id="m30">
<mml:mrow>
<mml:msub>
<mml:mi mathvariant="bold">x</mml:mi>
<mml:mi>j</mml:mi>
</mml:msub>
</mml:mrow>
</mml:math>
</inline-formula>, and <inline-formula id="inf25">
<mml:math id="m31">
<mml:mrow>
<mml:msub>
<mml:mrow>
<mml:mover accent="true">
<mml:mi>&#x3bc;</mml:mi>
<mml:mo>&#x5e;</mml:mo>
</mml:mover>
</mml:mrow>
<mml:mrow>
<mml:mi>j</mml:mi>
<mml:mi>q</mml:mi>
</mml:mrow>
</mml:msub>
</mml:mrow>
</mml:math>
</inline-formula> the sample mean of <inline-formula id="inf26">
<mml:math id="m32">
<mml:mrow>
<mml:msub>
<mml:mi mathvariant="bold">x</mml:mi>
<mml:mi>j</mml:mi>
</mml:msub>
</mml:mrow>
</mml:math>
</inline-formula> in class <inline-formula id="inf27">
<mml:math id="m33">
<mml:mrow>
<mml:msub>
<mml:mi>G</mml:mi>
<mml:mi>q</mml:mi>
</mml:msub>
</mml:mrow>
</mml:math>
</inline-formula> for <inline-formula id="inf28">
<mml:math id="m34">
<mml:mrow>
<mml:mi>q</mml:mi>
<mml:mo>&#x2208;</mml:mo>
<mml:mrow>
<mml:mo>{</mml:mo>
<mml:mrow>
<mml:mn>1</mml:mn>
<mml:mo>,</mml:mo>
<mml:mo>&#x2026;</mml:mo>
<mml:mo>,</mml:mo>
<mml:mi>Q</mml:mi>
</mml:mrow>
<mml:mo>}</mml:mo>
</mml:mrow>
</mml:mrow>
</mml:math>
</inline-formula>. Then, predictors associated with the highest values of this are retained, but no specific rule exists to choose the number of predictors to retain. A Bayesian-based technique, available in the R-package <italic>limma</italic>, has become a common way to deal with this, computing a Bayesian-based <italic>p</italic>-value for each predictor, therefore allowing users to choose the number of relevant predictors based on a threshold <italic>p</italic>-value&#x20;[<xref ref-type="bibr" rid="B18">18</xref>].</p>
<p>This method cannot be considered parsimonious, but rather as a pre-processing stage for exclusion of uninformative covariates. Reliably selecting relevant predictors in PLS regression is of interest for several reasons. Practically speaking, it would allow users to identify the original covariates which are significantly linked to the response, as is done in OLS regression with Student-type tests. Statistically speaking, it would avoid the establishment of over-complex models and ensure consistency of PLS estimators. Several methods for variable selection have already been developed [<xref ref-type="bibr" rid="B19">19</xref>]. Lazraq et&#x20;al. [<xref ref-type="bibr" rid="B20">20</xref>] group these techniques into two main categories. The first, model-wise selection, consists of first establishing the PLS model before performing a variable selection. The second, dimension-wise selection, consists of selecting variables on one PLS component at a&#x20;time.</p>
<p>A dimension-wise method, introduced by Chun and Kele&#x15f; [<xref ref-type="bibr" rid="B17">17</xref>] and called Sparse PLS (SPLS), has become the benchmark for selecting relevant predictors using PLS methodology. The technique is for continuous responses and consists of simultaneous dimension reduction and variable selection, computing sparse linear combinations of covariates as PLS components. This is achieved by introducing an <inline-formula id="inf29">
<mml:math id="m35">
<mml:mrow>
<mml:msub>
<mml:mi>L</mml:mi>
<mml:mn>1</mml:mn>
</mml:msub>
</mml:mrow>
</mml:math>
</inline-formula> constraint on the weight vectors <inline-formula id="inf30">
<mml:math id="m36">
<mml:mrow>
<mml:msub>
<mml:mtext>w</mml:mtext>
<mml:mi>k</mml:mi>
</mml:msub>
</mml:mrow>
</mml:math>
</inline-formula>, leading to the following formulation of the objective function:<disp-formula id="equ1">
<mml:math id="m37">
<mml:mrow>
<mml:msub>
<mml:mi mathvariant="bold">w</mml:mi>
<mml:mi>k</mml:mi>
</mml:msub>
<mml:mo>&#x3d;</mml:mo>
<mml:munder>
<mml:mrow>
<mml:mtext>argmax</mml:mtext>
</mml:mrow>
<mml:mrow>
<mml:mi mathvariant="bold">w</mml:mi>
<mml:mo>&#x2208;</mml:mo>
<mml:msup>
<mml:mi mathvariant="normal">&#x211d;</mml:mi>
<mml:mi>p</mml:mi>
</mml:msup>
</mml:mrow>
</mml:munder>
<mml:mrow>
<mml:mo>{</mml:mo>
<mml:mrow>
<mml:msup>
<mml:mi mathvariant="bold">w</mml:mi>
<mml:mi>T</mml:mi>
</mml:msup>
<mml:msubsup>
<mml:mi mathvariant="bold">X</mml:mi>
<mml:mrow>
<mml:mi>k</mml:mi>
<mml:mo>&#x2212;</mml:mo>
<mml:mn>1</mml:mn>
</mml:mrow>
<mml:mi>T</mml:mi>
</mml:msubsup>
<mml:msub>
<mml:mi mathvariant="bold">y</mml:mi>
<mml:mrow>
<mml:mi>k</mml:mi>
<mml:mo>&#x2212;</mml:mo>
<mml:mn>1</mml:mn>
</mml:mrow>
</mml:msub>
<mml:msubsup>
<mml:mi mathvariant="bold">y</mml:mi>
<mml:mrow>
<mml:mi>k</mml:mi>
<mml:mo>&#x2212;</mml:mo>
<mml:mn>1</mml:mn>
</mml:mrow>
<mml:mi>T</mml:mi>
</mml:msubsup>
<mml:msub>
<mml:mi mathvariant="bold">X</mml:mi>
<mml:mrow>
<mml:mi>k</mml:mi>
<mml:mo>&#x2212;</mml:mo>
<mml:mn>1</mml:mn>
</mml:mrow>
</mml:msub>
<mml:mi mathvariant="bold">w</mml:mi>
</mml:mrow>
<mml:mo>}</mml:mo>
</mml:mrow>
<mml:mo>,</mml:mo>
</mml:mrow>
</mml:math>
</disp-formula>
<disp-formula id="e7">
<mml:math id="m38">
<mml:mrow>
<mml:mtext>s</mml:mtext>
<mml:mo>.</mml:mo>
<mml:mtext>c</mml:mtext>
<mml:mo>.</mml:mo>
<mml:mtext>&#x2002;</mml:mtext>
<mml:mtext>&#xa0;</mml:mtext>
<mml:msubsup>
<mml:mrow>
<mml:mrow>
<mml:mo>&#x2016;</mml:mo>
<mml:mi mathvariant="bold">w</mml:mi>
<mml:mo>&#x2016;</mml:mo>
</mml:mrow>
</mml:mrow>
<mml:mn>2</mml:mn>
<mml:mn>2</mml:mn>
</mml:msubsup>
<mml:mo>&#x3d;</mml:mo>
<mml:mn>1</mml:mn>
<mml:mo>,</mml:mo>
<mml:mtext>&#x2002;</mml:mtext>
<mml:msub>
<mml:mrow>
<mml:mrow>
<mml:mo>&#x2016;</mml:mo>
<mml:mi mathvariant="bold">w</mml:mi>
<mml:mo>&#x2016;</mml:mo>
</mml:mrow>
</mml:mrow>
<mml:mn>1</mml:mn>
</mml:msub>
<mml:mo>&#x3c;</mml:mo>
<mml:mi>&#x3bb;</mml:mi>
<mml:mo>,</mml:mo>
</mml:mrow>
</mml:math>
<label>(7)</label>
</disp-formula>where <italic>&#x3bb;</italic> determines the amount of sparsity. More details are available in Chun and Kele&#x15f; [<xref ref-type="bibr" rid="B17">17</xref>]. Two tuning parameters are thus involved: <inline-formula id="inf31">
<mml:math id="m39">
<mml:mrow>
<mml:mi>&#x3b7;</mml:mi>
<mml:mo>&#x2208;</mml:mo>
<mml:mrow>
<mml:mo>[</mml:mo>
<mml:mrow>
<mml:mn>0,1</mml:mn>
</mml:mrow>
<mml:mo>]</mml:mo>
</mml:mrow>
</mml:mrow>
</mml:math>
</inline-formula> as a rescaled parameter equivalent to <italic>&#x3bb;</italic>, and the number of PLS components <italic>K</italic>, which are determined through CV-based mean squared error (CV MSE). We refer to this technique as SPLS CV in the following.</p>
<p>Chung and Keles [<xref ref-type="bibr" rid="B21">21</xref>] have developed an extension of this technique by integrating it into the generalized linear model (GLM) framework, leading it to be able to solve classification problems. They also integrate Firth&#x2019;s procedure, in order to deal with non-convergence issues. Both tuning parameters are selected using CV MSE. We refer to this method as SGPLS CV in the following.</p>
<p>A well-known model-wise selection method is the one introduced by Lazraq et&#x20;al. [<xref ref-type="bibr" rid="B20">20</xref>]. It consists of bootstrapping pairs <inline-formula id="inf32">
<mml:math id="m40">
<mml:mrow>
<mml:mrow>
<mml:mo>(</mml:mo>
<mml:mrow>
<mml:msub>
<mml:mi mathvariant="bold">y</mml:mi>
<mml:mi>i</mml:mi>
</mml:msub>
<mml:mo>,</mml:mo>
<mml:msub>
<mml:mi mathvariant="bold">x</mml:mi>
<mml:mrow>
<mml:mi>i</mml:mi>
<mml:mo>&#x2022;</mml:mo>
</mml:mrow>
</mml:msub>
</mml:mrow>
<mml:mo>)</mml:mo>
</mml:mrow>
<mml:mo>,</mml:mo>
<mml:mtext>&#x2002;</mml:mtext>
<mml:mn>1</mml:mn>
<mml:mi mathvariant="normal">&#x2a7d;</mml:mi>
<mml:mi>i</mml:mi>
<mml:mi mathvariant="normal">&#x2a7d;</mml:mi>
<mml:mi>n</mml:mi>
</mml:mrow>
</mml:math>
</inline-formula>, where <inline-formula id="inf33">
<mml:math id="m41">
<mml:mrow>
<mml:msub>
<mml:mi mathvariant="bold">x</mml:mi>
<mml:mrow>
<mml:mi>i</mml:mi>
<mml:mo>&#x2022;</mml:mo>
</mml:mrow>
</mml:msub>
</mml:mrow>
</mml:math>
</inline-formula> represents the <italic>i</italic>th row of <inline-formula id="inf34">
<mml:math id="m42">
<mml:mi mathvariant="bold">X</mml:mi>
</mml:math>
</inline-formula>, before applying PLS regression with a preset number of components <italic>K</italic> on each bootstrap sample. By performing this method, approximations of distributions related to predictors&#x2019; coefficients can be achieved. This leads to bootstrap-based confidence intervals (CI) for each predictor, and so opens up the possibility of directly testing their significance with a fixed type I risk <italic>&#x3b1;</italic>. The advantages of this method are twofold. First, by focusing on PLS regression coefficients, it is relevant for both the PLS and GPLS frameworks. Second, it only depends on one tuning parameter <italic>K</italic>, which must be determined earlier.</p>
<p>Two important related issue should be mentioned. Firstly, the presence of variables unrelated to the phenomenon under study strongly increases the level of inherent noise in the response and in the predictors. This phenomenon occurs frequently in various contexts and has been studied by several authors both in the context of PLS regression and its extensions [<xref ref-type="bibr" rid="B22">22</xref>, <xref ref-type="bibr" rid="B23">23</xref>] and in the more general context of machine learning where the use of techniques related to least squares have proven to be relevant and effective [<xref ref-type="bibr" rid="B24">24</xref>&#x2013;<xref ref-type="bibr" rid="B28">28</xref>].</p>
<p>Secondly, while performing this technique, approximations of distributions are achieved conditionally on the fixed dimension of the extracted subspace. In other words, it approximates the uncertainty of these coefficients, conditionally on the fact that estimations are performed in a <italic>K</italic>-dimensional subspace for each bootstrap sample. The determination of an optimal number of components is crucial for achieving reliable estimations of the regression coefficients [<xref ref-type="bibr" rid="B29">29</xref>]. Thus, since this determination is specific to the dataset in question, it must be performed for each bootstrap sample, in order to obtain reliable bootstrap-based CI. We have established some theoretical results which confirm this (<xref ref-type="sec" rid="s9">Supplementary Section&#x20;1.3</xref>).</p>
<p>Determining tuning parameters by using <italic>q</italic>-fold cross-validation (<italic>q</italic>-CV) based criteria may induce important issues concerning the stability of extracted results (Hastie et&#x20;al. [<xref ref-type="bibr" rid="B30">30</xref>], p.249; Boulesteix [<xref ref-type="bibr" rid="B31">31</xref>]; Magnanensi et&#x20;al. [<xref ref-type="bibr" rid="B32">32</xref>]). Thus, using such criteria for successive choosing of the number of components should be avoided. As mentioned, amongst others, by Efron and Tibshirani [<xref ref-type="bibr" rid="B33">33</xref>], p.255 and Kohavi [<xref ref-type="bibr" rid="B34">34</xref>], bootstrap-based criteria are known to be more stable than CV-based ones. In this context, Magnanensi et&#x20;al. [<xref ref-type="bibr" rid="B32">32</xref>] developed a robust bootstrap-based criterion for the determination of the number of PLS components, characterized by a high level of stability and suitable for both the PLS and GPLS regression frameworks. Thus, this criterion opens up the possibility of reliable successive choosing for each bootstrap sample.</p>
<p>In this article, we introduce a new dynamic bootstrap-based technique for covariate selection suitable for both the PLS and GPLS frameworks. It consists of bootstrapping pairs <inline-formula id="inf35">
<mml:math id="m43">
<mml:mrow>
<mml:mrow>
<mml:mo>(</mml:mo>
<mml:mrow>
<mml:msub>
<mml:mi mathvariant="bold">y</mml:mi>
<mml:mi>i</mml:mi>
</mml:msub>
<mml:mo>,</mml:mo>
<mml:msub>
<mml:mi mathvariant="bold">x</mml:mi>
<mml:mrow>
<mml:mi>i</mml:mi>
<mml:mo>&#x2022;</mml:mo>
</mml:mrow>
</mml:msub>
</mml:mrow>
<mml:mo>)</mml:mo>
</mml:mrow>
</mml:mrow>
</mml:math>
</inline-formula>, and successive extraction of the optimal number of components for each bootstrap sample, using the previously mentioned bootstrap-based criterion. Here, our goal is to better approximate the uncertainty related to regression coefficients by removing the condition of extracting a fixed <italic>K</italic>-dimensional subspace for each bootstrap sample, leading to more reliable CI. This new method both avoids the use of CV, and features the same advantages as those previously mentioned related to the technique introduced by Lazraq et&#x20;al. [<xref ref-type="bibr" rid="B20">20</xref>]. This method also has the advantage of behaving consistently with the variation of the noise level in the dataset. We refer to this new dynamic method as BootYTdyn in the following.</p>
<p>We also succeed in adapting the bootstrap-based criterion introduced by Magnanensi et&#x20;al. [<xref ref-type="bibr" rid="B32">32</xref>] to the determination of a unique optimal number of components related to each preset value of <italic>&#x3b7;</italic> in both the SPLS and SGPLS frameworks. Thus, these adapted versions, by reducing the use of CV, improve the reliability of the hyper-parameter tuning. We will refer to these adapted techniques as SPLS BootYT and SGPLS BootYT, respectively.</p>
<p>There is some constructive attributes of our proposed approach as follows:<list list-type="simple">
<list-item>
<p>1. theoretical results are developed and presented to give indications on the number of components related to a bootstrap sample and showed the necessity for a dynamic algorithm for Sparse&#x20;PLS;</p>
</list-item>
<list-item>
<p>2. adapt our stable stopping criterion in the construction of PLS components to the Sparse&#x20;PLS;</p>
</list-item>
<list-item>
<p>3. for each value of the parsimony parameter, a determination of the optimal number of components is performed using our new algorithm;</p>
</list-item>
<list-item>
<p>4. the number of models to be compared is considerably reduced;</p>
</list-item>
<list-item>
<p>5. possible thanks to the adaptation of our stable stopping criterion for PLS components and by its properties, especially its stability.</p>
</list-item>
</list>
</p>
<p>The article is organized as follows. In <xref ref-type="sec" rid="s2">Section 2</xref>, we introduce our new dynamic bootstrap-based technique, followed by the description of our adaptation of the BootYT stopping criterion to the SPLS and SGPLS frameworks. In <xref ref-type="sec" rid="s3">Section 3</xref>, we present simulations related to the PLS framework, and summarize the results, depending notably on different noise levels in <bold>y</bold>. In <xref ref-type="sec" rid="s4">Section 4</xref>, we treat a real microarray gene expression dataset with a binary response, here the original localization of colon tumors, by benchmarking our new dynamic bootstrap-based approach for GPLS regression. Lastly, we discuss results and conclusions in <xref ref-type="sec" rid="s5">Section 5</xref> as well as possible future&#x20;work.</p>
</sec>
<sec id="s2">
<title>2&#x20;Bootstrap-Based Approaches for Predictor Selection</title>
<sec id="s2-1">
<title>2.1 A New Dynamic Bootstrap-Based Technique</title>
<p>As mentioned in <xref ref-type="sec" rid="s1">Section 1</xref>, the selected number of extracted components is crucial for reliable estimation of <inline-formula id="inf36">
<mml:math id="m44">
<mml:mrow>
<mml:msubsup>
<mml:mi>&#x3b2;</mml:mi>
<mml:mi>j</mml:mi>
<mml:mrow>
<mml:mtext>PLS</mml:mtext>
</mml:mrow>
</mml:msubsup>
<mml:mo>,</mml:mo>
<mml:mn>1</mml:mn>
<mml:mi mathvariant="normal">&#x2a7d;</mml:mi>
<mml:mi>j</mml:mi>
<mml:mi mathvariant="normal">&#x2a7d;</mml:mi>
<mml:mi>p</mml:mi>
</mml:mrow>
</mml:math>
</inline-formula> [<xref ref-type="bibr" rid="B29">29</xref>]. We have shown (<xref ref-type="sec" rid="s9">Supplementary Section 1.3</xref>) that selecting an optimal number of components on the original dataset and using it to perform PLS regression on the constructed bootstrap samples, as done by Lazraq et&#x20;al. [<xref ref-type="bibr" rid="B20">20</xref>], is not correct for the obtention of reliable&#x20;CI.</p>
<p>In order to take into account these theoretical results, we have developed a new dynamic bootstrap-based approach for variable selection relevant for both the PLS and GPLS frameworks. The approach consists of estimating the optimal number of components for each bootstrap sample created during the algorithm. To obtain consistent results, a robust and resample-stable stopping criterion in component construction has to be used. Let us use <inline-formula id="inf37">
<mml:math id="m45">
<mml:mrow>
<mml:msub>
<mml:mi>&#x3b2;</mml:mi>
<mml:mi>j</mml:mi>
</mml:msub>
</mml:mrow>
</mml:math>
</inline-formula> to mean <inline-formula id="inf38">
<mml:math id="m46">
<mml:mrow>
<mml:msubsup>
<mml:mi>&#x3b2;</mml:mi>
<mml:mi>j</mml:mi>
<mml:mrow>
<mml:mtext>PLS</mml:mtext>
</mml:mrow>
</mml:msubsup>
</mml:mrow>
</mml:math>
</inline-formula> in the following, in order to lighten notation. The algorithm for this new method is as follows:<list list-type="simple">
<list-item>
<p>1. Let <inline-formula id="inf39">
<mml:math id="m47">
<mml:mrow>
<mml:msup>
<mml:mi>D</mml:mi>
<mml:mrow>
<mml:mi>o</mml:mi>
<mml:mi>r</mml:mi>
<mml:mi>i</mml:mi>
</mml:mrow>
</mml:msup>
</mml:mrow>
</mml:math>
</inline-formula> be the original dataset and <italic>R</italic> the total number of bootstrap samples <inline-formula id="inf40">
<mml:math id="m48">
<mml:mrow>
<mml:msubsup>
<mml:mi>D</mml:mi>
<mml:mi>r</mml:mi>
<mml:mi>b</mml:mi>
</mml:msubsup>
<mml:mo>,</mml:mo>
<mml:mtext>&#x2009;</mml:mtext>
<mml:mi>r</mml:mi>
<mml:mo>&#x2208;</mml:mo>
<mml:mrow>
<mml:mo>&#x301a;</mml:mo>
<mml:mrow>
<mml:mn>1</mml:mn>
<mml:mo>,</mml:mo>
<mml:mi>R</mml:mi>
</mml:mrow>
<mml:mo>&#x301b;</mml:mo>
</mml:mrow>
</mml:mrow>
</mml:math>
</inline-formula>.</p>
</list-item>
<list-item>
<p>2. <inline-formula id="inf41">
<mml:math id="m49">
<mml:mrow>
<mml:mo>&#x2200;</mml:mo>
<mml:mi>r</mml:mi>
<mml:mo>&#x2208;</mml:mo>
<mml:mrow>
<mml:mo>&#x301a;</mml:mo>
<mml:mrow>
<mml:mn>1</mml:mn>
<mml:mo>,</mml:mo>
<mml:mi>R</mml:mi>
</mml:mrow>
<mml:mo>&#x301b;</mml:mo>
</mml:mrow>
</mml:mrow>
</mml:math>
</inline-formula>:</p>
<list list-type="simple">
<list-item>
<p>&#x2022; Extract the number of PLS components that is needed for <inline-formula id="inf42">
<mml:math id="m50">
<mml:mrow>
<mml:msubsup>
<mml:mi>D</mml:mi>
<mml:mi>r</mml:mi>
<mml:mi>b</mml:mi>
</mml:msubsup>
</mml:mrow>
</mml:math>
</inline-formula> using a preset stopping criterion.</p>
</list-item>
<list-item>
<p>&#x2022; Compute the estimations <inline-formula id="inf43">
<mml:math id="m51">
<mml:mrow>
<mml:msubsup>
<mml:mrow>
<mml:mover accent="true">
<mml:mi>&#x3b2;</mml:mi>
<mml:mo>&#x5e;</mml:mo>
</mml:mover>
</mml:mrow>
<mml:mi>j</mml:mi>
<mml:mi>r</mml:mi>
</mml:msubsup>
<mml:mo>,</mml:mo>
<mml:mtext>&#x2009;</mml:mtext>
<mml:mo>&#x2200;</mml:mo>
<mml:mi>j</mml:mi>
<mml:mo>&#x2208;</mml:mo>
<mml:mrow>
<mml:mo>&#x301a;</mml:mo>
<mml:mrow>
<mml:mn>1</mml:mn>
<mml:mo>,</mml:mo>
<mml:mi>p</mml:mi>
</mml:mrow>
<mml:mo>&#x301b;</mml:mo>
</mml:mrow>
</mml:mrow>
</mml:math>
</inline-formula>, by fitting the relevant PLS or GPLS&#x20;model.</p>
</list-item>
</list>
</list-item>
<list-item>
<p>3. <inline-formula id="inf44">
<mml:math id="m52">
<mml:mrow>
<mml:mo>&#x2200;</mml:mo>
<mml:mi>j</mml:mi>
<mml:mo>&#x2208;</mml:mo>
<mml:mrow>
<mml:mo>&#x301a;</mml:mo>
<mml:mrow>
<mml:mn>1</mml:mn>
<mml:mo>,</mml:mo>
<mml:mi>p</mml:mi>
</mml:mrow>
<mml:mo>&#x301b;</mml:mo>
</mml:mrow>
</mml:mrow>
</mml:math>
</inline-formula>, construct a <inline-formula id="inf45">
<mml:math id="m53">
<mml:mrow>
<mml:mrow>
<mml:mo>(</mml:mo>
<mml:mrow>
<mml:mn>100</mml:mn>
<mml:mo>&#xd7;</mml:mo>
<mml:mrow>
<mml:mo>(</mml:mo>
<mml:mrow>
<mml:mn>1</mml:mn>
<mml:mo>&#x2212;</mml:mo>
<mml:mi>&#x3b1;</mml:mi>
</mml:mrow>
<mml:mo>)</mml:mo>
</mml:mrow>
</mml:mrow>
<mml:mo>)</mml:mo>
</mml:mrow>
</mml:mrow>
</mml:math>
</inline-formula> % bilateral <inline-formula id="inf46">
<mml:math id="m54">
<mml:mrow>
<mml:mi>B</mml:mi>
<mml:msub>
<mml:mi>C</mml:mi>
<mml:mi>a</mml:mi>
</mml:msub>
</mml:mrow>
</mml:math>
</inline-formula> CI for <inline-formula id="inf47">
<mml:math id="m55">
<mml:mrow>
<mml:msub>
<mml:mi>&#x3b2;</mml:mi>
<mml:mi>j</mml:mi>
</mml:msub>
</mml:mrow>
</mml:math>
</inline-formula>, noted:</p>
</list-item>
</list>
<disp-formula id="equ2">
<mml:math id="m56">
<mml:mrow>
<mml:msub>
<mml:mrow>
<mml:mtext>CI</mml:mtext>
</mml:mrow>
<mml:mi>j</mml:mi>
</mml:msub>
<mml:mo>&#x3d;</mml:mo>
<mml:mrow>
<mml:mo>[</mml:mo>
<mml:mrow>
<mml:msub>
<mml:mrow>
<mml:mtext>CI</mml:mtext>
</mml:mrow>
<mml:mrow>
<mml:mi>j</mml:mi>
<mml:mo>,</mml:mo>
<mml:mn>1</mml:mn>
</mml:mrow>
</mml:msub>
<mml:mo>,</mml:mo>
<mml:msub>
<mml:mrow>
<mml:mtext>CI</mml:mtext>
</mml:mrow>
<mml:mrow>
<mml:mi>j</mml:mi>
<mml:mo>,</mml:mo>
<mml:mn>2</mml:mn>
</mml:mrow>
</mml:msub>
</mml:mrow>
<mml:mo>]</mml:mo>
</mml:mrow>
<mml:mo>.</mml:mo>
</mml:mrow>
</mml:math>
</disp-formula>
<list list-type="simple">
<list-item>
<p>4. <inline-formula id="inf48">
<mml:math id="m57">
<mml:mrow>
<mml:mo>&#x2200;</mml:mo>
<mml:mi>j</mml:mi>
<mml:mo>&#x2208;</mml:mo>
<mml:mrow>
<mml:mo>&#x301a;</mml:mo>
<mml:mrow>
<mml:mn>1</mml:mn>
<mml:mo>,</mml:mo>
<mml:mi>p</mml:mi>
</mml:mrow>
<mml:mo>&#x301b;</mml:mo>
</mml:mrow>
</mml:mrow>
</mml:math>
</inline-formula>, If <inline-formula id="inf49">
<mml:math id="m58">
<mml:mrow>
<mml:mn>0</mml:mn>
<mml:mo>&#x2209;</mml:mo>
<mml:msub>
<mml:mrow>
<mml:mtext>CI</mml:mtext>
</mml:mrow>
<mml:mi>j</mml:mi>
</mml:msub>
</mml:mrow>
</mml:math>
</inline-formula> then retain <inline-formula id="inf50">
<mml:math id="m59">
<mml:mrow>
<mml:msub>
<mml:mtext>x</mml:mtext>
<mml:mtext>j</mml:mtext>
</mml:msub>
</mml:mrow>
</mml:math>
</inline-formula> else delete&#x20;<inline-formula id="inf51">
<mml:math id="m60">
<mml:mrow>
<mml:msub>
<mml:mtext>x</mml:mtext>
<mml:mtext>j</mml:mtext>
</mml:msub>
</mml:mrow>
</mml:math>
</inline-formula>.</p>
</list-item>
<list-item>
<p>5. Obtain the reduced model <inline-formula id="inf52">
<mml:math id="m61">
<mml:mrow>
<mml:msub>
<mml:mi mathvariant="normal">&#x2133;</mml:mi>
<mml:mrow>
<mml:mtext>sel</mml:mtext>
</mml:mrow>
</mml:msub>
</mml:mrow>
</mml:math>
</inline-formula> by only integrating the significant predictors, and extracting the number of components <inline-formula id="inf53">
<mml:math id="m62">
<mml:mrow>
<mml:msub>
<mml:mi>K</mml:mi>
<mml:mrow>
<mml:mi>s</mml:mi>
<mml:mi>e</mml:mi>
<mml:mi>l</mml:mi>
</mml:mrow>
</mml:msub>
</mml:mrow>
</mml:math>
</inline-formula> determined by the preset stopping criterion.</p>
</list-item>
</list>
</p>
<p>Note that here, we set <inline-formula id="inf54">
<mml:math id="m63">
<mml:mrow>
<mml:mi>&#x3b1;</mml:mi>
<mml:mo>&#x3d;</mml:mo>
<mml:mn>0.05</mml:mn>
</mml:mrow>
</mml:math>
</inline-formula>.</p>
</sec>
<sec id="s2-2">
<title>2.2 An Adapted Bootstrap-Based Sparse PLS Implementation</title>
<p>As mentioned by Boulesteix [<xref ref-type="bibr" rid="B31">31</xref>], using <italic>q</italic>-CV based methods for tuning parameters potentially induces problems, notably concerning variability of results due to dependency on the way folds are randomly chosen. However, as detailed in Section New developments in Sparse PLS regression, the selection of both tuning parameters involved in the SPLS regression developed by Chun and Kele&#x15f; [<xref ref-type="bibr" rid="B17">17</xref>] is performed using <italic>q</italic>-CV MSE. Therefore, to improve the reliability of this selection, we adapt the bootstrap-based stopping criterion to this method, which gives the following algorithm:<list list-type="simple">
<list-item>
<p>1. Let <inline-formula id="inf55">
<mml:math id="m64">
<mml:mrow>
<mml:mrow>
<mml:mo>{</mml:mo>
<mml:mrow>
<mml:msub>
<mml:mi>&#x3b7;</mml:mi>
<mml:mn>1</mml:mn>
</mml:msub>
<mml:mo>,</mml:mo>
<mml:mo>&#x2026;</mml:mo>
<mml:mo>,</mml:mo>
<mml:msub>
<mml:mi>&#x3b7;</mml:mi>
<mml:mi>s</mml:mi>
</mml:msub>
</mml:mrow>
<mml:mo>}</mml:mo>
</mml:mrow>
</mml:mrow>
</mml:math>
</inline-formula> be a set of pre-chosen values for the sparsity parameter and <inline-formula id="inf56">
<mml:math id="m65">
<mml:mrow>
<mml:mrow>
<mml:mo>{</mml:mo>
<mml:mrow>
<mml:msub>
<mml:mi>k</mml:mi>
<mml:mn>1</mml:mn>
</mml:msub>
<mml:mo>,</mml:mo>
<mml:mo>&#x2026;</mml:mo>
<mml:mo>,</mml:mo>
<mml:msub>
<mml:mi>k</mml:mi>
<mml:mi>s</mml:mi>
</mml:msub>
</mml:mrow>
<mml:mo>}</mml:mo>
</mml:mrow>
<mml:mo>&#x3d;</mml:mo>
<mml:mrow>
<mml:mo>{</mml:mo>
<mml:mrow>
<mml:mn>1</mml:mn>
<mml:mo>,</mml:mo>
<mml:mo>&#x2026;</mml:mo>
<mml:mo>,</mml:mo>
<mml:mn>1</mml:mn>
</mml:mrow>
<mml:mo>}</mml:mo>
</mml:mrow>
</mml:mrow>
</mml:math>
</inline-formula> the set of initial numbers of components for each <inline-formula id="inf57">
<mml:math id="m66">
<mml:mrow>
<mml:msub>
<mml:mi>&#x3b7;</mml:mi>
<mml:mi>i</mml:mi>
</mml:msub>
</mml:mrow>
</mml:math>
</inline-formula>. Let <inline-formula id="inf58">
<mml:math id="m67">
<mml:mrow>
<mml:mi>i</mml:mi>
<mml:mo>&#x3d;</mml:mo>
<mml:mn>1</mml:mn>
</mml:mrow>
</mml:math>
</inline-formula>.</p>
</list-item>
<list-item>
<p>2. Let <inline-formula id="inf59">
<mml:math id="m68">
<mml:mrow>
<mml:msubsup>
<mml:mi>c</mml:mi>
<mml:mi>j</mml:mi>
<mml:mrow>
<mml:msub>
<mml:mi>&#x3b7;</mml:mi>
<mml:mi>i</mml:mi>
</mml:msub>
</mml:mrow>
</mml:msubsup>
<mml:mo>,</mml:mo>
<mml:mtext>&#x2002;</mml:mtext>
<mml:mi>j</mml:mi>
<mml:mo>&#x2208;</mml:mo>
<mml:mrow>
<mml:mo>&#x301a;</mml:mo>
<mml:mrow>
<mml:mn>1</mml:mn>
<mml:mo>,</mml:mo>
<mml:msub>
<mml:mi>k</mml:mi>
<mml:mi>i</mml:mi>
</mml:msub>
</mml:mrow>
<mml:mo>&#x301b;</mml:mo>
</mml:mrow>
</mml:mrow>
</mml:math>
</inline-formula>, be the regression coefficients of <inline-formula id="inf60">
<mml:math id="m69">
<mml:mi mathvariant="bold">y</mml:mi>
</mml:math>
</inline-formula> on <inline-formula id="inf61">
<mml:math id="m70">
<mml:mrow>
<mml:msub>
<mml:mi mathvariant="bold">T</mml:mi>
<mml:mrow>
<mml:msub>
<mml:mi>k</mml:mi>
<mml:mi>i</mml:mi>
</mml:msub>
</mml:mrow>
</mml:msub>
<mml:mo>&#x3d;</mml:mo>
<mml:mrow>
<mml:mo>(</mml:mo>
<mml:mrow>
<mml:msub>
<mml:mi mathvariant="bold">t</mml:mi>
<mml:mn>1</mml:mn>
</mml:msub>
<mml:mo>,</mml:mo>
<mml:mo>&#x2026;</mml:mo>
<mml:mo>,</mml:mo>
<mml:msub>
<mml:mi mathvariant="bold">t</mml:mi>
<mml:mrow>
<mml:msub>
<mml:mi>k</mml:mi>
<mml:mi>i</mml:mi>
</mml:msub>
</mml:mrow>
</mml:msub>
</mml:mrow>
<mml:mo>)</mml:mo>
</mml:mrow>
<mml:mo>&#x2208;</mml:mo>
<mml:msub>
<mml:mi mathvariant="normal">&#x2133;</mml:mi>
<mml:mrow>
<mml:mi>n</mml:mi>
<mml:mo>,</mml:mo>
<mml:msub>
<mml:mi>k</mml:mi>
<mml:mi>i</mml:mi>
</mml:msub>
</mml:mrow>
</mml:msub>
<mml:mrow>
<mml:mo>(</mml:mo>
<mml:mi mathvariant="normal">&#x211d;</mml:mi>
<mml:mo>)</mml:mo>
</mml:mrow>
</mml:mrow>
</mml:math>
</inline-formula>. Obtain <inline-formula id="inf62">
<mml:math id="m71">
<mml:mrow>
<mml:msub>
<mml:mi>k</mml:mi>
<mml:mi>i</mml:mi>
</mml:msub>
</mml:mrow>
</mml:math>
</inline-formula> <inline-formula id="inf63">
<mml:math id="m72">
<mml:mrow>
<mml:mi>B</mml:mi>
<mml:msub>
<mml:mi>C</mml:mi>
<mml:mi>a</mml:mi>
</mml:msub>
</mml:mrow>
</mml:math>
</inline-formula> CI for <inline-formula id="inf64">
<mml:math id="m73">
<mml:mrow>
<mml:msubsup>
<mml:mi>c</mml:mi>
<mml:mi>j</mml:mi>
<mml:mrow>
<mml:msub>
<mml:mi>&#x3b7;</mml:mi>
<mml:mi>i</mml:mi>
</mml:msub>
</mml:mrow>
</mml:msubsup>
<mml:mo>,</mml:mo>
<mml:mtext>&#x2002;</mml:mtext>
<mml:mi>j</mml:mi>
<mml:mo>&#x2208;</mml:mo>
<mml:mrow>
<mml:mo>&#x301a;</mml:mo>
<mml:mrow>
<mml:mn>1</mml:mn>
<mml:mo>,</mml:mo>
<mml:msub>
<mml:mi>k</mml:mi>
<mml:mi>i</mml:mi>
</mml:msub>
</mml:mrow>
<mml:mo>&#x301b;</mml:mo>
</mml:mrow>
</mml:mrow>
</mml:math>
</inline-formula>, using the bootstrap-based stopping criterion, noted:</p>
</list-item>
</list>
<disp-formula id="equ3">
<mml:math id="m74">
<mml:mrow>
<mml:msubsup>
<mml:mrow>
<mml:mtext>CI</mml:mtext>
</mml:mrow>
<mml:mi>j</mml:mi>
<mml:mrow>
<mml:msub>
<mml:mi>k</mml:mi>
<mml:mi>i</mml:mi>
</mml:msub>
</mml:mrow>
</mml:msubsup>
<mml:mo>&#x3d;</mml:mo>
<mml:mrow>
<mml:mo>[</mml:mo>
<mml:mrow>
<mml:msubsup>
<mml:mrow>
<mml:mtext>CI</mml:mtext>
</mml:mrow>
<mml:mrow>
<mml:mi>j</mml:mi>
<mml:mo>,</mml:mo>
<mml:mn>1</mml:mn>
</mml:mrow>
<mml:mrow>
<mml:msub>
<mml:mi>k</mml:mi>
<mml:mi>i</mml:mi>
</mml:msub>
</mml:mrow>
</mml:msubsup>
<mml:mo>,</mml:mo>
<mml:msubsup>
<mml:mrow>
<mml:mtext>CI</mml:mtext>
</mml:mrow>
<mml:mrow>
<mml:mi>j</mml:mi>
<mml:mo>,</mml:mo>
<mml:mn>2</mml:mn>
</mml:mrow>
<mml:mrow>
<mml:msub>
<mml:mi>k</mml:mi>
<mml:mi>i</mml:mi>
</mml:msub>
</mml:mrow>
</mml:msubsup>
</mml:mrow>
<mml:mo>]</mml:mo>
</mml:mrow>
<mml:mo>.</mml:mo>
</mml:mrow>
</mml:math>
</disp-formula>
<list list-type="simple">
<list-item>
<p>3. If <inline-formula id="inf65">
<mml:math id="m75">
<mml:mrow>
<mml:mo>&#x2203;</mml:mo>
<mml:mi>j</mml:mi>
<mml:mo>&#x2208;</mml:mo>
<mml:mrow>
<mml:mo>&#x301a;</mml:mo>
<mml:mrow>
<mml:mn>1</mml:mn>
<mml:mo>,</mml:mo>
<mml:msub>
<mml:mi>k</mml:mi>
<mml:mi>i</mml:mi>
</mml:msub>
</mml:mrow>
<mml:mo>&#x301b;</mml:mo>
</mml:mrow>
<mml:mtext>&#xa0;such&#xa0;that&#xa0;</mml:mtext>
<mml:mtext>&#x2009;</mml:mtext>
<mml:mn>0</mml:mn>
<mml:mo>&#x2208;</mml:mo>
<mml:msubsup>
<mml:mrow>
<mml:mtext>CI</mml:mtext>
</mml:mrow>
<mml:mi>j</mml:mi>
<mml:mrow>
<mml:msub>
<mml:mi>k</mml:mi>
<mml:mi>i</mml:mi>
</mml:msub>
</mml:mrow>
</mml:msubsup>
</mml:mrow>
</mml:math>
</inline-formula> then <inline-formula id="inf66">
<mml:math id="m76">
<mml:mrow>
<mml:msubsup>
<mml:mi>K</mml:mi>
<mml:mrow>
<mml:mi>o</mml:mi>
<mml:mi>p</mml:mi>
<mml:mi>t</mml:mi>
</mml:mrow>
<mml:mrow>
<mml:msub>
<mml:mi>&#x3b7;</mml:mi>
<mml:mi>i</mml:mi>
</mml:msub>
</mml:mrow>
</mml:msubsup>
<mml:mo>&#x3d;</mml:mo>
<mml:msub>
<mml:mi>k</mml:mi>
<mml:mi>i</mml:mi>
</mml:msub>
<mml:mo>&#x2212;</mml:mo>
<mml:mn>1</mml:mn>
</mml:mrow>
</mml:math>
</inline-formula> else <inline-formula id="inf67">
<mml:math id="m77">
<mml:mrow>
<mml:mrow>
<mml:mo>{</mml:mo>
<mml:mrow>
<mml:msub>
<mml:mi>k</mml:mi>
<mml:mi>i</mml:mi>
</mml:msub>
<mml:mo>&#x3d;</mml:mo>
<mml:msub>
<mml:mi>k</mml:mi>
<mml:mi>i</mml:mi>
</mml:msub>
<mml:mo>&#x2b;</mml:mo>
<mml:mn>1</mml:mn>
<mml:mtext>&#xa0;and&#xa0;return&#xa0;to&#xa0;step&#xa0;</mml:mtext>
<mml:mn>2</mml:mn>
</mml:mrow>
<mml:mo>}</mml:mo>
</mml:mrow>
</mml:mrow>
</mml:math>
</inline-formula>.</p>
</list-item>
<list-item>
<p>4. While <inline-formula id="inf68">
<mml:math id="m78">
<mml:mrow>
<mml:mi>i</mml:mi>
<mml:mo>&#x2260;</mml:mo>
<mml:mi>s</mml:mi>
</mml:mrow>
</mml:math>
</inline-formula> then <inline-formula id="inf69">
<mml:math id="m79">
<mml:mrow>
<mml:mrow>
<mml:mo>{</mml:mo>
<mml:mrow>
<mml:mi>i</mml:mi>
<mml:mo>&#x3d;</mml:mo>
<mml:mi>i</mml:mi>
<mml:mo>&#x2b;</mml:mo>
<mml:mn>1</mml:mn>
<mml:mtext>&#xa0;and&#xa0;return&#xa0;to&#xa0;step&#xa0;</mml:mtext>
<mml:mn>2</mml:mn>
</mml:mrow>
<mml:mo>}</mml:mo>
</mml:mrow>
</mml:mrow>
</mml:math>
</inline-formula>
<bold>.</bold>
</p>
</list-item>
<list-item>
<p>5. Return the set of the extracted numbers of components <inline-formula id="inf70">
<mml:math id="m80">
<mml:mrow>
<mml:mrow>
<mml:mo>{</mml:mo>
<mml:mrow>
<mml:msubsup>
<mml:mi>K</mml:mi>
<mml:mrow>
<mml:mi>o</mml:mi>
<mml:mi>p</mml:mi>
<mml:mi>t</mml:mi>
</mml:mrow>
<mml:mrow>
<mml:msub>
<mml:mi>&#x3b7;</mml:mi>
<mml:mn>1</mml:mn>
</mml:msub>
</mml:mrow>
</mml:msubsup>
<mml:mo>,</mml:mo>
<mml:mo>&#x2026;</mml:mo>
<mml:mo>,</mml:mo>
<mml:msubsup>
<mml:mi>K</mml:mi>
<mml:mrow>
<mml:mi>o</mml:mi>
<mml:mi>p</mml:mi>
<mml:mi>t</mml:mi>
</mml:mrow>
<mml:mrow>
<mml:msub>
<mml:mi>&#x3b7;</mml:mi>
<mml:mi>s</mml:mi>
</mml:msub>
</mml:mrow>
</mml:msubsup>
</mml:mrow>
<mml:mo>}</mml:mo>
</mml:mrow>
</mml:mrow>
</mml:math>
</inline-formula> related to <inline-formula id="inf71">
<mml:math id="m81">
<mml:mrow>
<mml:mrow>
<mml:mo>{</mml:mo>
<mml:mrow>
<mml:msub>
<mml:mi>&#x3b7;</mml:mi>
<mml:mn>1</mml:mn>
</mml:msub>
<mml:mo>,</mml:mo>
<mml:mo>&#x2026;</mml:mo>
<mml:mo>,</mml:mo>
<mml:msub>
<mml:mi>&#x3b7;</mml:mi>
<mml:mi>s</mml:mi>
</mml:msub>
</mml:mrow>
<mml:mo>}</mml:mo>
</mml:mrow>
</mml:mrow>
</mml:math>
</inline-formula>.</p>
</list-item>
<list-item>
<p>6. Return the pair <inline-formula id="inf72">
<mml:math id="m82">
<mml:mrow>
<mml:mrow>
<mml:mo>(</mml:mo>
<mml:mrow>
<mml:msub>
<mml:mi>&#x3b7;</mml:mi>
<mml:mrow>
<mml:mi>o</mml:mi>
<mml:mi>p</mml:mi>
<mml:mi>t</mml:mi>
</mml:mrow>
</mml:msub>
<mml:mo>,</mml:mo>
<mml:msubsup>
<mml:mi>K</mml:mi>
<mml:mrow>
<mml:mi>o</mml:mi>
<mml:mi>p</mml:mi>
<mml:mi>t</mml:mi>
</mml:mrow>
<mml:mrow>
<mml:msub>
<mml:mi>&#x3b7;</mml:mi>
<mml:mrow>
<mml:mi>o</mml:mi>
<mml:mi>p</mml:mi>
<mml:mi>t</mml:mi>
</mml:mrow>
</mml:msub>
</mml:mrow>
</mml:msubsup>
</mml:mrow>
<mml:mo>)</mml:mo>
</mml:mrow>
</mml:mrow>
</mml:math>
</inline-formula> having the lowest CV-based&#x20;MSE.</p>
</list-item>
</list>
</p>
<p>Retesting all components obtained after each increase in <inline-formula id="inf73">
<mml:math id="m83">
<mml:mrow>
<mml:msub>
<mml:mi>k</mml:mi>
<mml:mi>i</mml:mi>
</mml:msub>
</mml:mrow>
</mml:math>
</inline-formula> is essential since the original predictors involved in the components construction change when <inline-formula id="inf74">
<mml:math id="m84">
<mml:mrow>
<mml:msub>
<mml:mi>k</mml:mi>
<mml:mi>i</mml:mi>
</mml:msub>
</mml:mrow>
</mml:math>
</inline-formula> increases [<xref ref-type="bibr" rid="B17">17</xref>]. This fact, combined with the aim of retaining orthogonality between components, leads the components themselves to change, so that the significance of each component has to be retested at each&#x20;step.</p>
<p>While the original implementation compares <inline-formula id="inf75">
<mml:math id="m85">
<mml:mrow>
<mml:msup>
<mml:mi>K</mml:mi>
<mml:mrow>
<mml:mtext>max</mml:mtext>
</mml:mrow>
</mml:msup>
<mml:mo>&#xd7;</mml:mo>
<mml:mi>s</mml:mi>
</mml:mrow>
</mml:math>
</inline-formula> models through CV MSE, with <inline-formula id="inf76">
<mml:math id="m86">
<mml:mrow>
<mml:msup>
<mml:mi>K</mml:mi>
<mml:mrow>
<mml:mtext>max</mml:mtext>
</mml:mrow>
</mml:msup>
</mml:mrow>
</mml:math>
</inline-formula> the maximal number of components (set by the user), this new bootstrap-based version only focuses on <italic>s</italic> models, since only one number of components is determined for each preset value of &#x3b7;. An illustration of the stability improvement obtained by using this new implementation, based on the simulated dataset introduced in <xref ref-type="sec" rid="s3-3-1">Section 3.3.1</xref> with <inline-formula id="inf77">
<mml:math id="m87">
<mml:mrow>
<mml:mi>s</mml:mi>
<mml:mi>d</mml:mi>
<mml:mrow>
<mml:mo>(</mml:mo>
<mml:mi mathvariant="normal">&#x3f5;</mml:mi>
<mml:mo>)</mml:mo>
</mml:mrow>
<mml:mo>&#x3d;</mml:mo>
<mml:mn>5</mml:mn>
</mml:mrow>
</mml:math>
</inline-formula>, is shown in <xref ref-type="fig" rid="F1">Figure&#x20;1</xref>.</p>
<fig id="F1" position="float">
<label>FIGURE 1</label>
<caption>
<p>Repartition of 100 selections of <inline-formula id="inf78">
<mml:math id="m88">
<mml:mrow>
<mml:mrow>
<mml:mo>(</mml:mo>
<mml:mrow>
<mml:msub>
<mml:mi>&#x3b7;</mml:mi>
<mml:mrow>
<mml:mi>o</mml:mi>
<mml:mi>p</mml:mi>
<mml:mi>t</mml:mi>
</mml:mrow>
</mml:msub>
<mml:mo>,</mml:mo>
<mml:msub>
<mml:mi>K</mml:mi>
<mml:mrow>
<mml:mi>o</mml:mi>
<mml:mi>p</mml:mi>
<mml:mi>t</mml:mi>
</mml:mrow>
</mml:msub>
</mml:mrow>
<mml:mo>)</mml:mo>
</mml:mrow>
</mml:mrow>
</mml:math>
</inline-formula> using the original SPLS approach <bold>(A)</bold> and the new bootstrap-based implementation <bold>(B)</bold>.</p>
</caption>
<graphic xlink:href="fams-07-693126-g001.tif"/>
</fig>
</sec>
</sec>
<sec id="s3">
<title>3 Simulations Studies</title>
<sec id="s3-1">
<title>3.1 Simulations for Accuracy Comparisons</title>
<p>These simulations are based on a simulation scheme proposed by Chun and Kele&#x15f; [<xref ref-type="bibr" rid="B17">17</xref>, p.14] and modified in order to study high-dimensional settings. We consider the case where there are less observations than predictors, i.e.,&#x20;<inline-formula id="inf79">
<mml:math id="m89">
<mml:mrow>
<mml:mi>n</mml:mi>
<mml:mo>&#x3c;</mml:mo>
<mml:mi>p</mml:mi>
</mml:mrow>
</mml:math>
</inline-formula>, and set <inline-formula id="inf80">
<mml:math id="m90">
<mml:mrow>
<mml:mi>n</mml:mi>
<mml:mo>&#x3d;</mml:mo>
<mml:mn>100</mml:mn>
</mml:mrow>
</mml:math>
</inline-formula>, and <inline-formula id="inf81">
<mml:math id="m91">
<mml:mrow>
<mml:mi>p</mml:mi>
<mml:mo>&#x3d;</mml:mo>
<mml:mn>200</mml:mn>
</mml:mrow>
</mml:math>
</inline-formula> or 1,000. Let <italic>q</italic> be the number of spurious predictors. While Chun and Kele&#x15f; [<xref ref-type="bibr" rid="B17">17</xref>, p.14] only consider a ratio <inline-formula id="inf82">
<mml:math id="m92">
<mml:mrow>
<mml:mi>q</mml:mi>
<mml:mo>/</mml:mo>
<mml:mi>p</mml:mi>
</mml:mrow>
</mml:math>
</inline-formula> equal to 0.25 and 0.5, both the 0.05 and 0.95 ratio values have been added here. Four independent and identically distributed hidden variables <inline-formula id="inf83">
<mml:math id="m93">
<mml:mrow>
<mml:msub>
<mml:mi mathvariant="bold">h</mml:mi>
<mml:mn>1</mml:mn>
</mml:msub>
<mml:mo>,</mml:mo>
<mml:mo>&#x2026;</mml:mo>
<mml:mo>,</mml:mo>
<mml:msub>
<mml:mi mathvariant="bold">h</mml:mi>
<mml:mn>4</mml:mn>
</mml:msub>
</mml:mrow>
</mml:math>
</inline-formula>, following a <inline-formula id="inf84">
<mml:math id="m94">
<mml:mrow>
<mml:mi mathvariant="script">N</mml:mi>
<mml:mrow>
<mml:mo>(</mml:mo>
<mml:mrow>
<mml:mn>0,25</mml:mn>
<mml:msub>
<mml:mi>I</mml:mi>
<mml:mi>n</mml:mi>
</mml:msub>
</mml:mrow>
<mml:mo>)</mml:mo>
</mml:mrow>
</mml:mrow>
</mml:math>
</inline-formula> distribution, were computed. Then, columns of the covariate matrix <inline-formula id="inf85">
<mml:math id="m95">
<mml:mtext>X</mml:mtext>
</mml:math>
</inline-formula> were generated by <inline-formula id="inf86">
<mml:math id="m96">
<mml:mrow>
<mml:msub>
<mml:mtext>x</mml:mtext>
<mml:mi>j</mml:mi>
</mml:msub>
<mml:mo>&#x3d;</mml:mo>
<mml:msub>
<mml:mi mathvariant="bold">h</mml:mi>
<mml:mi>l</mml:mi>
</mml:msub>
<mml:mo>&#x2b;</mml:mo>
<mml:msub>
<mml:mi mathvariant="normal">&#x3f5;</mml:mi>
<mml:mi>j</mml:mi>
</mml:msub>
</mml:mrow>
</mml:math>
</inline-formula> for <inline-formula id="inf87">
<mml:math id="m97">
<mml:mrow>
<mml:msub>
<mml:mi>p</mml:mi>
<mml:mrow>
<mml:mi>l</mml:mi>
<mml:mo>&#x2212;</mml:mo>
<mml:mn>1</mml:mn>
</mml:mrow>
</mml:msub>
<mml:mo>&#x2b;</mml:mo>
<mml:mn>1</mml:mn>
<mml:mtext>&#x2009;</mml:mtext>
<mml:mi mathvariant="normal">&#x2a7d;</mml:mi>
<mml:mtext>&#x2009;</mml:mtext>
<mml:mi>j</mml:mi>
<mml:mtext>&#x2009;</mml:mtext>
<mml:mi mathvariant="normal">&#x2a7d;</mml:mi>
<mml:mtext>&#x2009;</mml:mtext>
<mml:msub>
<mml:mi>p</mml:mi>
<mml:mi>l</mml:mi>
</mml:msub>
</mml:mrow>
</mml:math>
</inline-formula>, where <inline-formula id="inf88">
<mml:math id="m98">
<mml:mrow>
<mml:mi>l</mml:mi>
<mml:mo>&#x3d;</mml:mo>
<mml:mn>1</mml:mn>
<mml:mo>,</mml:mo>
<mml:mo>&#x2026;</mml:mo>
<mml:mo>,</mml:mo>
<mml:mn>4</mml:mn>
</mml:mrow>
</mml:math>
</inline-formula>, <inline-formula id="inf89">
<mml:math id="m99">
<mml:mrow>
<mml:mrow>
<mml:mo>(</mml:mo>
<mml:mrow>
<mml:msub>
<mml:mi>p</mml:mi>
<mml:mn>0</mml:mn>
</mml:msub>
<mml:mo>,</mml:mo>
<mml:mo>&#x2026;</mml:mo>
<mml:mo>,</mml:mo>
<mml:msub>
<mml:mi>p</mml:mi>
<mml:mn>4</mml:mn>
</mml:msub>
</mml:mrow>
<mml:mo>)</mml:mo>
</mml:mrow>
<mml:mo>&#x3d;</mml:mo>
<mml:mrow>
<mml:mo>(</mml:mo>
<mml:mrow>
<mml:mn>0</mml:mn>
<mml:mo>,</mml:mo>
<mml:mrow>
<mml:mrow>
<mml:mrow>
<mml:mo>(</mml:mo>
<mml:mrow>
<mml:mi>p</mml:mi>
<mml:mo>&#x2212;</mml:mo>
<mml:mi>q</mml:mi>
</mml:mrow>
<mml:mo>)</mml:mo>
</mml:mrow>
</mml:mrow>
<mml:mo>/</mml:mo>
<mml:mn>2</mml:mn>
</mml:mrow>
<mml:mo>,</mml:mo>
<mml:mi>p</mml:mi>
<mml:mo>&#x2212;</mml:mo>
<mml:mi>q</mml:mi>
<mml:mo>,</mml:mo>
<mml:mi>p</mml:mi>
<mml:mo>&#x2212;</mml:mo>
<mml:mi>r</mml:mi>
<mml:mo>,</mml:mo>
<mml:mi>p</mml:mi>
</mml:mrow>
<mml:mo>)</mml:mo>
</mml:mrow>
</mml:mrow>
</mml:math>
</inline-formula>, <inline-formula id="inf90">
<mml:math id="m100">
<mml:mrow>
<mml:mi>r</mml:mi>
<mml:mo>&#x3d;</mml:mo>
<mml:mn>5</mml:mn>
</mml:mrow>
</mml:math>
</inline-formula> when <inline-formula id="inf91">
<mml:math id="m101">
<mml:mrow>
<mml:mi>p</mml:mi>
<mml:mo>&#x3d;</mml:mo>
<mml:mn>200</mml:mn>
</mml:mrow>
</mml:math>
</inline-formula> and <inline-formula id="inf92">
<mml:math id="m102">
<mml:mrow>
<mml:mi>r</mml:mi>
<mml:mo>&#x3d;</mml:mo>
<mml:mn>10</mml:mn>
</mml:mrow>
</mml:math>
</inline-formula> when <inline-formula id="inf93">
<mml:math id="m103">
<mml:mrow>
<mml:mi>p</mml:mi>
<mml:mo>&#x3d;</mml:mo>
<mml:mn>1000</mml:mn>
</mml:mrow>
</mml:math>
</inline-formula>, and <inline-formula id="inf94">
<mml:math id="m104">
<mml:mrow>
<mml:msub>
<mml:mi mathvariant="normal">&#x3f5;</mml:mi>
<mml:mn>1</mml:mn>
</mml:msub>
<mml:mo>,</mml:mo>
<mml:mo>&#x2026;</mml:mo>
<mml:mo>,</mml:mo>
<mml:msub>
<mml:mi mathvariant="normal">&#x3f5;</mml:mi>
<mml:mi>p</mml:mi>
</mml:msub>
</mml:mrow>
</mml:math>
</inline-formula> are drawn independently from a <inline-formula id="inf95">
<mml:math id="m105">
<mml:mrow>
<mml:mi mathvariant="script">N</mml:mi>
<mml:mrow>
<mml:mo>(</mml:mo>
<mml:mrow>
<mml:mn>0,0.1</mml:mn>
<mml:msub>
<mml:mi>I</mml:mi>
<mml:mi>n</mml:mi>
</mml:msub>
</mml:mrow>
<mml:mo>)</mml:mo>
</mml:mrow>
</mml:mrow>
</mml:math>
</inline-formula>. Also, <bold>y</bold> is generated by <inline-formula id="inf96">
<mml:math id="m106">
<mml:mrow>
<mml:mn>3</mml:mn>
<mml:msub>
<mml:mi mathvariant="bold">h</mml:mi>
<mml:mn>1</mml:mn>
</mml:msub>
<mml:mo>&#x2212;</mml:mo>
<mml:mn>4</mml:mn>
<mml:msub>
<mml:mi mathvariant="bold">h</mml:mi>
<mml:mn>2</mml:mn>
</mml:msub>
<mml:mo>&#x2b;</mml:mo>
<mml:mi>f</mml:mi>
</mml:mrow>
</mml:math>
</inline-formula>, where <italic>f</italic> is normally distributed with mean 0 and variance such that the signal-to-noise ratio (SNR) equals&#x20;10.</p>
<p>Using this simulation scheme, accuracy of the SPLS technique using 10&#x20;fold-CV for selecting tuning parameters (SPLS CV), and our new dynamic bootstrap method combined with the bootstrap-based stopping criterion (BootYTdyn), is compared. In order to do so, for each parameter setting, 50 selections of the sparse support related to both methods were established. Lastly, mean accuracy values over the 50 trials were calculated. Results are summarized in <xref ref-type="table" rid="T1">Table&#x20;1</xref>.</p>
<table-wrap id="T1" position="float">
<label>TABLE 1</label>
<caption>
<p>Mean accuracy values (SNR).</p>
</caption>
<table>
<thead valign="top">
<tr>
<th align="left">
<italic>p</italic>
</th>
<th colspan="4" align="center">200</th>
<th colspan="4" align="center">1,000</th>
</tr>
<tr>
<th align="left">
<inline-formula id="inf97">
<mml:math id="m107">
<mml:mrow>
<mml:mi>q</mml:mi>
<mml:mo>/</mml:mo>
<mml:mi>p</mml:mi>
</mml:mrow>
</mml:math>
</inline-formula>
</th>
<th align="center">0.05</th>
<th align="center">0.25</th>
<th align="center">0.5</th>
<th align="center">0.95</th>
<th align="center">0.05</th>
<th align="center">0.25</th>
<th align="center">0.5</th>
<th align="center">0.95</th>
</tr>
</thead>
<tbody valign="top">
<tr>
<td align="left">SPLS CV</td>
<td align="char" char=".">0.986</td>
<td align="char" char=".">0.961</td>
<td align="char" char=".">0.849</td>
<td align="char" char=".">0.591</td>
<td align="char" char=".">0.998</td>
<td align="char" char=".">0.997</td>
<td align="char" char=".">0.989</td>
<td align="char" char=".">0.963</td>
</tr>
<tr>
<td align="left">BootYTdyn</td>
<td align="char" char=".">1.000</td>
<td align="char" char=".">0.867</td>
<td align="char" char=".">0.805</td>
<td align="char" char=".">0.982</td>
<td align="char" char=".">0.967</td>
<td align="char" char=".">0.827</td>
<td align="char" char=".">0.893</td>
<td align="char" char=".">0.985</td>
</tr>
</tbody>
</table>
</table-wrap>
<p>Based on these results, SPLS CV gives better accuracy than BootYTdyn for ratio values <inline-formula id="inf98">
<mml:math id="m108">
<mml:mrow>
<mml:mi>q</mml:mi>
<mml:mo>/</mml:mo>
<mml:mi>p</mml:mi>
</mml:mrow>
</mml:math>
</inline-formula> that are not close to 0 or 1. While both give good performance when the ratio is close to 0, i.e.,&#x20;when a major part of predictors are significant, BootYTdyn outperforms SPLS CV when only a small proportion of predictors are significant.</p>
<p>Nevertheless, in this simulation set-up, covariates are collected into four groups. While within-group correlations between covariates are close to one, between-group ones are close to zero. This unrealistic situation makes irrelevant the determination of an optimal support, and seems more appropriate to selecting the number of components. As an illustration, 50 additional samples in the <inline-formula id="inf99">
<mml:math id="m109">
<mml:mrow>
<mml:mi>p</mml:mi>
<mml:mo>&#x3d;</mml:mo>
<mml:mn>1000</mml:mn>
</mml:mrow>
</mml:math>
</inline-formula> and <inline-formula id="inf100">
<mml:math id="m110">
<mml:mrow>
<mml:mi>q</mml:mi>
<mml:mo>/</mml:mo>
<mml:mi>p</mml:mi>
<mml:mo>&#x3d;</mml:mo>
<mml:mn>0.5</mml:mn>
</mml:mrow>
</mml:math>
</inline-formula> case were simulated. We then calculated the predictive MSE (PMSE) based on four different supports <inline-formula id="inf101">
<mml:math id="m111">
<mml:mrow>
<mml:msub>
<mml:mi>S</mml:mi>
<mml:mn>1</mml:mn>
</mml:msub>
</mml:mrow>
</mml:math>
</inline-formula>, <inline-formula id="inf102">
<mml:math id="m112">
<mml:mrow>
<mml:msub>
<mml:mi>S</mml:mi>
<mml:mn>2</mml:mn>
</mml:msub>
</mml:mrow>
</mml:math>
</inline-formula>, <inline-formula id="inf103">
<mml:math id="m113">
<mml:mrow>
<mml:msub>
<mml:mi>S</mml:mi>
<mml:mn>3</mml:mn>
</mml:msub>
</mml:mrow>
</mml:math>
</inline-formula> and <inline-formula id="inf104">
<mml:math id="m114">
<mml:mrow>
<mml:msub>
<mml:mi>S</mml:mi>
<mml:mn>4</mml:mn>
</mml:msub>
</mml:mrow>
</mml:math>
</inline-formula>, where <inline-formula id="inf105">
<mml:math id="m115">
<mml:mrow>
<mml:msub>
<mml:mi>S</mml:mi>
<mml:mn>1</mml:mn>
</mml:msub>
<mml:mo>&#x3d;</mml:mo>
<mml:mrow>
<mml:mo>{</mml:mo>
<mml:mrow>
<mml:msub>
<mml:mtext>x</mml:mtext>
<mml:mi>j</mml:mi>
</mml:msub>
<mml:mo>,</mml:mo>
<mml:mn>1</mml:mn>
<mml:mi mathvariant="normal">&#x2a7d;</mml:mi>
<mml:mi>j</mml:mi>
<mml:mi mathvariant="normal">&#x2a7d;</mml:mi>
<mml:mi>p</mml:mi>
</mml:mrow>
<mml:mo>}</mml:mo>
</mml:mrow>
</mml:mrow>
</mml:math>
</inline-formula>, <inline-formula id="inf106">
<mml:math id="m116">
<mml:mrow>
<mml:msub>
<mml:mi>S</mml:mi>
<mml:mn>2</mml:mn>
</mml:msub>
<mml:mo>&#x3d;</mml:mo>
<mml:mrow>
<mml:mo>{</mml:mo>
<mml:mrow>
<mml:msub>
<mml:mtext>x</mml:mtext>
<mml:mi>j</mml:mi>
</mml:msub>
<mml:mo>,</mml:mo>
<mml:mn>1</mml:mn>
<mml:mi mathvariant="normal">&#x2a7d;</mml:mi>
<mml:mi>j</mml:mi>
<mml:mi mathvariant="normal">&#x2a7d;</mml:mi>
<mml:mrow>
<mml:mo>(</mml:mo>
<mml:mrow>
<mml:mi>p</mml:mi>
<mml:mo>&#x2212;</mml:mo>
<mml:mi>q</mml:mi>
</mml:mrow>
<mml:mo>)</mml:mo>
</mml:mrow>
</mml:mrow>
<mml:mo>}</mml:mo>
</mml:mrow>
</mml:mrow>
</mml:math>
</inline-formula>, <inline-formula id="inf107">
<mml:math id="m117">
<mml:mrow>
<mml:msub>
<mml:mi>S</mml:mi>
<mml:mn>3</mml:mn>
</mml:msub>
<mml:mo>&#x3d;</mml:mo>
<mml:mrow>
<mml:mo>{</mml:mo>
<mml:mrow>
<mml:msub>
<mml:mi mathvariant="bold">x</mml:mi>
<mml:mn>1</mml:mn>
</mml:msub>
<mml:mtext>,</mml:mtext>
<mml:msub>
<mml:mi mathvariant="bold">x</mml:mi>
<mml:mrow>
<mml:mn>251</mml:mn>
</mml:mrow>
</mml:msub>
</mml:mrow>
<mml:mo>}</mml:mo>
</mml:mrow>
</mml:mrow>
</mml:math>
</inline-formula> and <inline-formula id="inf108">
<mml:math id="m118">
<mml:mrow>
<mml:msub>
<mml:mi>S</mml:mi>
<mml:mn>4</mml:mn>
</mml:msub>
<mml:mo>&#x3d;</mml:mo>
<mml:mrow>
<mml:mo>{</mml:mo>
<mml:mrow>
<mml:msub>
<mml:mi mathvariant="bold">x</mml:mi>
<mml:mn>1</mml:mn>
</mml:msub>
<mml:mo>,</mml:mo>
<mml:msub>
<mml:mi mathvariant="bold">x</mml:mi>
<mml:mrow>
<mml:mn>251</mml:mn>
</mml:mrow>
</mml:msub>
</mml:mrow>
<mml:mo>}</mml:mo>
</mml:mrow>
<mml:mstyle displaystyle="true">
<mml:mo>&#x222a;</mml:mo>
<mml:mrow>
<mml:mrow>
<mml:mo>{</mml:mo>
<mml:mrow>
<mml:msub>
<mml:mi mathvariant="bold">x</mml:mi>
<mml:mi>j</mml:mi>
</mml:msub>
<mml:mo>,</mml:mo>
<mml:mrow>
<mml:mo>(</mml:mo>
<mml:mrow>
<mml:mi>p</mml:mi>
<mml:mo>&#x2212;</mml:mo>
<mml:mi>q</mml:mi>
</mml:mrow>
<mml:mo>)</mml:mo>
</mml:mrow>
<mml:mo>&#x2b;</mml:mo>
<mml:mn>1</mml:mn>
<mml:mi mathvariant="normal">&#x2a7d;</mml:mi>
<mml:mi>j</mml:mi>
<mml:mi mathvariant="normal">&#x2a7d;</mml:mi>
<mml:mi>p</mml:mi>
</mml:mrow>
<mml:mo>}</mml:mo>
</mml:mrow>
</mml:mrow>
</mml:mstyle>
</mml:mrow>
</mml:math>
</inline-formula>. Results are summarized in <xref ref-type="table" rid="T2">Table&#x20;2</xref>.</p>
<table-wrap id="T2" position="float">
<label>TABLE 2</label>
<caption>
<p>PMSE values for different supports.</p>
</caption>
<table>
<thead valign="top">
<tr>
<th align="left"/>
<th align="center">
<inline-formula id="inf109">
<mml:math id="m119">
<mml:mrow>
<mml:msub>
<mml:mi>S</mml:mi>
<mml:mn>1</mml:mn>
</mml:msub>
</mml:mrow>
</mml:math>
</inline-formula>
</th>
<th align="center">
<inline-formula id="inf110">
<mml:math id="m120">
<mml:mrow>
<mml:msub>
<mml:mi>S</mml:mi>
<mml:mn>2</mml:mn>
</mml:msub>
</mml:mrow>
</mml:math>
</inline-formula>
</th>
<th align="center">
<inline-formula id="inf111">
<mml:math id="m121">
<mml:mrow>
<mml:msub>
<mml:mi>S</mml:mi>
<mml:mn>3</mml:mn>
</mml:msub>
</mml:mrow>
</mml:math>
</inline-formula>
</th>
<th align="center">
<inline-formula id="inf112">
<mml:math id="m122">
<mml:mrow>
<mml:msub>
<mml:mi>S</mml:mi>
<mml:mn>4</mml:mn>
</mml:msub>
</mml:mrow>
</mml:math>
</inline-formula>
</th>
</tr>
</thead>
<tbody valign="top">
<tr>
<td align="left">K &#x3d; 1</td>
<td align="char" char=".">65.383</td>
<td align="char" char=".">62.702</td>
<td align="center">63.988</td>
<td align="char" char=".">856.779</td>
</tr>
<tr>
<td align="left">K &#x3d; 2</td>
<td align="char" char=".">63.957</td>
<td align="char" char=".">64.443</td>
<td align="center">65.695</td>
<td align="char" char=".">209.873</td>
</tr>
<tr>
<td align="left">K &#x3d; 3</td>
<td align="char" char=".">65.745</td>
<td align="char" char=".">76.197</td>
<td align="center">NA</td>
<td align="char" char=".">61.801</td>
</tr>
</tbody>
</table>
</table-wrap>
<p>In the light of these observations, the aim of this simulation scheme is instead the extraction if an optimal number of components rather than an optimal support. We thus decided to use a real dataset as covariate matrix for a more general and relevant comparison.</p>
</sec>
<sec id="s3-2">
<title>3.2 Simulations for Global Comparison</title>
<sec id="s3-3-1">
<title>3.3.1 Dataset Simulations</title>
<p>In this study, we used a real microarray gene expression dataset, which was created using fresh-frozen primary tumors samples, from a multi-center cohort, with stage I to IV colon cancer. 566 samples fulfilled RNA quality requirement, and constituted our database. These samples were split into a 443 sample discovery set and a 123 sample test set, well balanced for the main anatomo-clinical characteristics. This database has already been studied by Marisa et&#x20;al. [<xref ref-type="bibr" rid="B35">35</xref>] and more details on it are available in their article.</p>
<p>In order to reduce computational time, a preliminary selection of 100 predictors was performed. Based on the original localization of the tumors as response vector, and the full 566 samples, the 100 most differentially expressed probe sets were extracted. As mentioned in Section New developments in Sparse PLS regression, this pre-processing is based on a Bayesian technique and gives us our benchmark predictors matrix.</p>
<p>Then, based on correlation values, four positively-correlated predictors were selected to form the set of significant covariates (<xref ref-type="sec" rid="s9">Supplementary Section 1.1</xref>). To this end, let <inline-formula id="inf113">
<mml:math id="m123">
<mml:mrow>
<mml:msub>
<mml:mtext>X</mml:mtext>
<mml:mrow>
<mml:mi>s</mml:mi>
<mml:mi>e</mml:mi>
<mml:mi>l</mml:mi>
</mml:mrow>
</mml:msub>
<mml:mo>&#x3d;</mml:mo>
<mml:mrow>
<mml:mo>(</mml:mo>
<mml:mrow>
<mml:msub>
<mml:mtext>x</mml:mtext>
<mml:mn>1</mml:mn>
</mml:msub>
<mml:mo>,</mml:mo>
<mml:msub>
<mml:mtext>x</mml:mtext>
<mml:mrow>
<mml:mn>12</mml:mn>
</mml:mrow>
</mml:msub>
<mml:mo>,</mml:mo>
<mml:msub>
<mml:mtext>x</mml:mtext>
<mml:mrow>
<mml:mn>15</mml:mn>
</mml:mrow>
</mml:msub>
<mml:mo>,</mml:mo>
<mml:msub>
<mml:mtext>x</mml:mtext>
<mml:mrow>
<mml:mn>59</mml:mn>
</mml:mrow>
</mml:msub>
</mml:mrow>
<mml:mo>)</mml:mo>
</mml:mrow>
<mml:mo>&#x2208;</mml:mo>
<mml:msub>
<mml:mi mathvariant="normal">&#x2133;</mml:mi>
<mml:mrow>
<mml:mi>n</mml:mi>
<mml:mo>,</mml:mo>
<mml:mn>4</mml:mn>
</mml:mrow>
</mml:msub>
<mml:mrow>
<mml:mo>(</mml:mo>
<mml:mi mathvariant="normal">&#x211d;</mml:mi>
<mml:mo>)</mml:mo>
</mml:mrow>
</mml:mrow>
</mml:math>
</inline-formula> be the matrix composed of these predictors, so that <bold>y</bold> is simulated as follows:<disp-formula id="e8">
<mml:math id="m124">
<mml:mrow>
<mml:mtext>y</mml:mtext>
<mml:mo>&#x3d;</mml:mo>
<mml:msub>
<mml:mtext>X</mml:mtext>
<mml:mrow>
<mml:mi>s</mml:mi>
<mml:mi>e</mml:mi>
<mml:mi>l</mml:mi>
</mml:mrow>
</mml:msub>
<mml:mi mathvariant="italic">&#x3b2;</mml:mi>
<mml:mo>&#x2b;</mml:mo>
<mml:mi mathvariant="normal">&#x3f5;</mml:mi>
<mml:mo>,</mml:mo>
</mml:mrow>
</mml:math>
<label>(8)</label>
</disp-formula>with <inline-formula id="inf115">
<mml:math id="m125">
<mml:mrow>
<mml:mi>&#x3b2;</mml:mi>
<mml:mo>&#x3d;</mml:mo>
<mml:msup>
<mml:mrow>
<mml:mrow>
<mml:mo>(</mml:mo>
<mml:mrow>
<mml:mn>3.559</mml:mn>
<mml:mo>,</mml:mo>
<mml:mn>2.071</mml:mn>
<mml:mo>,</mml:mo>
<mml:mn>1.440</mml:mn>
<mml:mo>,</mml:mo>
<mml:mn>1.770</mml:mn>
</mml:mrow>
<mml:mo>)</mml:mo>
</mml:mrow>
</mml:mrow>
<mml:mi>T</mml:mi>
</mml:msup>
</mml:mrow>
</mml:math>
</inline-formula>, <inline-formula id="inf116">
<mml:math id="m126">
<mml:mrow>
<mml:mi mathvariant="double-struck">E</mml:mi>
<mml:mrow>
<mml:mo>(</mml:mo>
<mml:mi mathvariant="normal">&#x3f5;</mml:mi>
<mml:mo>)</mml:mo>
</mml:mrow>
<mml:mo>&#x3d;</mml:mo>
<mml:mn>0</mml:mn>
</mml:mrow>
</mml:math>
</inline-formula> and Var <inline-formula id="inf117">
<mml:math id="m127">
<mml:mrow>
<mml:mrow>
<mml:mo>(</mml:mo>
<mml:mi mathvariant="normal">&#x3f5;</mml:mi>
<mml:mo>)</mml:mo>
</mml:mrow>
<mml:mo>&#x3d;</mml:mo>
<mml:msup>
<mml:mi>&#x3c3;</mml:mi>
<mml:mn>2</mml:mn>
</mml:msup>
<mml:msub>
<mml:mi>I</mml:mi>
<mml:mi>n</mml:mi>
</mml:msub>
</mml:mrow>
</mml:math>
</inline-formula>.</p>
<p>We performed simulations for six distinct levels of random noise standard deviation in order to investigate the performance of the different methods. Both these standard deviations and their related SNR are shown in <xref ref-type="table" rid="T3">Table&#x20;3</xref>.</p>
<table-wrap id="T3" position="float">
<label>TABLE 3</label>
<caption>
<p>Noise standard deviation (SNR).</p>
</caption>
<table>
<thead valign="top">
<tr>
<th align="left">Dataset 1</th>
<th align="center">Dataset 2</th>
<th align="center">Dataset 3</th>
<th align="center">Dataset 4</th>
<th align="center">Dataset 5</th>
<th align="center">Dataset 6</th>
</tr>
</thead>
<tbody valign="top">
<tr>
<td align="left">0.5 (810.603)</td>
<td align="center">1 (202.651)</td>
<td align="center">3 (22.517)</td>
<td align="center">4 (12.666)</td>
<td align="center">5 (8.106)</td>
<td align="center">6.366 (5.000)</td>
</tr>
</tbody>
</table>
</table-wrap>
</sec>
<sec id="s3-3-2">
<title>3.3.2 Benchmarked Methods</title>
<p>Using these simulated datasets, eight methods were analyzed and compared.<list list-type="simple">
<list-item>
<p>1. <inline-formula id="inf118">
<mml:math id="m128">
<mml:mrow>
<mml:msup>
<mml:mi>Q</mml:mi>
<mml:mn>2</mml:mn>
</mml:msup>
</mml:mrow>
</mml:math>
</inline-formula>. The original bootstrap-based method, introduced by Lazraq et&#x20;al. [<xref ref-type="bibr" rid="B20">20</xref>], combined with the 10-fold CV-based <inline-formula id="inf119">
<mml:math id="m129">
<mml:mrow>
<mml:msup>
<mml:mi>Q</mml:mi>
<mml:mn>2</mml:mn>
</mml:msup>
</mml:mrow>
</mml:math>
</inline-formula> criterion [<xref ref-type="bibr" rid="B5">5</xref>, p.83] for pre-selecting the number of components.</p>
</list-item>
<list-item>
<p>2. BIC. The bootstrap-based method, introduced by Lazraq et&#x20;al. [<xref ref-type="bibr" rid="B20">20</xref>], combined with the corrected BIC using estimated degrees of freedom (DoF) [<xref ref-type="bibr" rid="B36">36</xref>] for pre-selecting the number of components.</p>
</list-item>
<list-item>
<p>3. BootYT. The bootstrap-based method, introduced by Lazraq et&#x20;al. [<xref ref-type="bibr" rid="B20">20</xref>], combined with the bootstrap-based criterion [<xref ref-type="bibr" rid="B32">32</xref>] for previously selecting the number of components.</p>
</list-item>
<list-item>
<p>4. BICdyn. Our new dynamic bootstrap-based method combined with the corrected BIC criterion for successive selections of number of components.</p>
</list-item>
<list-item>
<p>5. BootYTdyn. Our new dynamic bootstrap-based method combined with the bootstrap-based criterion for successive selections of number of components.</p>
</list-item>
<list-item>
<p>6. SPLS CV. The original SPLS method using 10-fold CV for tuning parameter determination&#x20;[<xref ref-type="bibr" rid="B17">17</xref>].</p>
</list-item>
<list-item>
<p>7. SPLS BootYT. The new adapted SPLS version using the bootstrap-based criterion for component selection.</p>
</list-item>
<list-item>
<p>8. Lasso. Lasso regression, included as a benchmark&#x20;[<xref ref-type="bibr" rid="B37">37</xref>].</p>
</list-item>
</list>
</p>
</sec>
<sec id="s3-3-3">
<title>3.3.3 Simulation Scheme and Notation</title>
<p>In order to perform reliable comparisons between the eight methods, each type of trial was performed one hundred times. Numbers of components, sparse supports and sparse tuning parameters are the main examples of these. Results linked to the highest occurrence rates are then chosen for method comparison. All bootstrap-based techniques were performed with <inline-formula id="inf120">
<mml:math id="m130">
<mml:mrow>
<mml:mi>R</mml:mi>
<mml:mo>&#x3d;</mml:mo>
<mml:mn>1000</mml:mn>
</mml:mrow>
</mml:math>
</inline-formula> bootstrap samples, and each related CI was constructed with type I risk <inline-formula id="inf121">
<mml:math id="m131">
<mml:mrow>
<mml:mi>&#x3b1;</mml:mi>
<mml:mo>&#x3d;</mml:mo>
<mml:mn>0.05</mml:mn>
</mml:mrow>
</mml:math>
</inline-formula>.</p>
<p>The global comparison has two main parts. First, in order to compare accuracy and stability related to each technique, we focus both on different supports, and models extracted by the different variable selection methods. Indeed, in the PLS framework, a specific model results from both a set of predictors and a specific number of components. Due to the sparsity parameter in SPLS approaches, the same support can be extracted, but with a different number of components leading to different models. Lasso regression can also extract the same support for several different sparsity parameters, leading to different estimations of model coefficients. Therefore, for clarity, the following notation, related to each specific variable selection technique, is introduced.<list list-type="simple">
<list-item>
<p>&#x2022; <inline-formula id="inf122">
<mml:math id="m132">
<mml:mrow>
<mml:mrow>
<mml:mo>{</mml:mo>
<mml:mrow>
<mml:msub>
<mml:mi mathvariant="script">S</mml:mi>
<mml:mn>1</mml:mn>
</mml:msub>
<mml:mo>,</mml:mo>
<mml:mo>&#x2026;</mml:mo>
<mml:mo>,</mml:mo>
<mml:msub>
<mml:mi mathvariant="script">S</mml:mi>
<mml:mrow>
<mml:msub>
<mml:mtext>&#x393;</mml:mtext>
<mml:mn>1</mml:mn>
</mml:msub>
</mml:mrow>
</mml:msub>
</mml:mrow>
<mml:mo>}</mml:mo>
</mml:mrow>
</mml:mrow>
</mml:math>
</inline-formula>, the set of extracted supports.</p>
</list-item>
<list-item>
<p>&#x2022; <inline-formula id="inf123">
<mml:math id="m133">
<mml:mrow>
<mml:mrow>
<mml:mo>{</mml:mo>
<mml:mrow>
<mml:msub>
<mml:mi mathvariant="normal">&#x2133;</mml:mi>
<mml:mn>1</mml:mn>
</mml:msub>
<mml:mo>,</mml:mo>
<mml:mo>&#x2026;</mml:mo>
<mml:mo>,</mml:mo>
<mml:msub>
<mml:mi mathvariant="normal">&#x2133;</mml:mi>
<mml:mrow>
<mml:msub>
<mml:mtext>&#x393;</mml:mtext>
<mml:mn>2</mml:mn>
</mml:msub>
</mml:mrow>
</mml:msub>
</mml:mrow>
<mml:mo>}</mml:mo>
</mml:mrow>
</mml:mrow>
</mml:math>
</inline-formula>, the set of fitted models.</p>
</list-item>
<list-item>
<p>&#x2022; <inline-formula id="inf124">
<mml:math id="m134">
<mml:mrow>
<mml:msub>
<mml:mi mathvariant="script">S</mml:mi>
<mml:mrow>
<mml:mi>s</mml:mi>
<mml:mi>e</mml:mi>
<mml:mi>l</mml:mi>
</mml:mrow>
</mml:msub>
</mml:mrow>
</mml:math>
</inline-formula>, the selected support, i.e.,&#x20;the one that appears the most&#x20;often.</p>
</list-item>
<list-item>
<p>&#x2022; <inline-formula id="inf125">
<mml:math id="m135">
<mml:mrow>
<mml:msub>
<mml:mi mathvariant="normal">&#x2133;</mml:mi>
<mml:mrow>
<mml:mi>s</mml:mi>
<mml:mi>e</mml:mi>
<mml:mi>l</mml:mi>
</mml:mrow>
</mml:msub>
</mml:mrow>
</mml:math>
</inline-formula>, the selected model, i.e.,&#x20;the one that appears the most&#x20;often.</p>
</list-item>
<list-item>
<p>&#x2022; <inline-formula id="inf126">
<mml:math id="m136">
<mml:mrow>
<mml:mtext>%</mml:mtext>
<mml:msub>
<mml:mi mathvariant="script">S</mml:mi>
<mml:mrow>
<mml:mi>s</mml:mi>
<mml:mi>e</mml:mi>
<mml:mi>l</mml:mi>
</mml:mrow>
</mml:msub>
</mml:mrow>
</mml:math>
</inline-formula>, rate of occurrence of the selected support.</p>
</list-item>
<list-item>
<p>&#x2022; <inline-formula id="inf127">
<mml:math id="m137">
<mml:mrow>
<mml:mtext>%</mml:mtext>
<mml:msub>
<mml:mi mathvariant="normal">&#x2133;</mml:mi>
<mml:mrow>
<mml:mi>s</mml:mi>
<mml:mi>e</mml:mi>
<mml:mi>l</mml:mi>
</mml:mrow>
</mml:msub>
</mml:mrow>
</mml:math>
</inline-formula>, rate of occurrence of the selected&#x20;model.</p>
</list-item>
<list-item>
<p>&#x2022; <inline-formula id="inf128">
<mml:math id="m138">
<mml:mrow>
<mml:msub>
<mml:mi>K</mml:mi>
<mml:mrow>
<mml:mi>s</mml:mi>
<mml:mi>e</mml:mi>
<mml:mi>l</mml:mi>
</mml:mrow>
</mml:msub>
</mml:mrow>
</mml:math>
</inline-formula>, the number of components related to the selected&#x20;model.</p>
</list-item>
</list>
</p>
<p>Second, in order to compare the predictive ability of models, 10-fold CV MSE, related to each selected sparse model through PLS regression, were computed one hundred times. The test set was also used in order to confirm results obtained by&#x20;CV.</p>
<p>Note that, concerning the dynamic BIC-based method for <inline-formula id="inf129">
<mml:math id="m139">
<mml:mrow>
<mml:mtext>sd</mml:mtext>
<mml:mrow>
<mml:mo>(</mml:mo>
<mml:mi mathvariant="normal">&#x3f5;</mml:mi>
<mml:mo>)</mml:mo>
</mml:mrow>
<mml:mo>&#x3d;</mml:mo>
<mml:mn>0.5</mml:mn>
</mml:mrow>
</mml:math>
</inline-formula>, only 97 trials performed well. Lastly, due to equality of occurrence rates between the two most-represented pairs of tuning parameters, results for SPLS CV and <inline-formula id="inf130">
<mml:math id="m140">
<mml:mrow>
<mml:mtext>sd</mml:mtext>
<mml:mrow>
<mml:mo>(</mml:mo>
<mml:mi mathvariant="normal">&#x3f5;</mml:mi>
<mml:mo>)</mml:mo>
</mml:mrow>
<mml:mo>&#x3d;</mml:mo>
<mml:mn>5</mml:mn>
</mml:mrow>
</mml:math>
</inline-formula> come instead from 150 trials.</p>
</sec>
<sec id="s3-3-4">
<title>3.3.4 Stability and Accuracy Results</title>
<p>Both the mean accuracy values over trials in each case, and stability results based on extracted supports, are given in <xref ref-type="table" rid="T4">Table&#x20;4</xref>. The numbers of components used for the original bootstrap-based approach [<xref ref-type="bibr" rid="B20">20</xref>] are summarized in <xref ref-type="sec" rid="s9">Supplementary Section&#x20;1.2</xref>.</p>
<table-wrap id="T4" position="float">
<label>TABLE 4</label>
<caption>
<p>Mean accuracy values over trials in each case and stability results based on extracted supports and number of predictors.</p>
</caption>
<table>
<thead valign="top">
<tr>
<th align="left"/>
<th align="center">
<inline-formula id="inf131">
<mml:math id="m141">
<mml:mrow>
<mml:msup>
<mml:mi>Q</mml:mi>
<mml:mn>2</mml:mn>
</mml:msup>
</mml:mrow>
</mml:math>
</inline-formula>
</th>
<th align="center">BIC</th>
<th align="center">BICdyn</th>
<th align="center">BootYT</th>
<th align="center">BootYTdyn</th>
<th align="center">SPLS CV</th>
<th align="center">SPLS BootYT</th>
<th align="center">Lasso</th>
</tr>
</thead>
<tbody valign="top">
<tr>
<td colspan="9" align="left">
<bold>Mean accuracy values</bold>
</td>
</tr>
<tr>
<td align="left">&#x2003;<inline-formula id="inf132">
<mml:math id="m142">
<mml:mrow>
<mml:mtext>sd</mml:mtext>
<mml:mrow>
<mml:mo>(</mml:mo>
<mml:mi mathvariant="normal">&#x3f5;</mml:mi>
<mml:mo>)</mml:mo>
</mml:mrow>
<mml:mo>&#x3d;</mml:mo>
<mml:mn>0.5</mml:mn>
</mml:mrow>
</mml:math>
</inline-formula>
</td>
<td align="center">0.9331</td>
<td align="center">0.9710</td>
<td align="center">0.9882</td>
<td align="center">0.9587</td>
<td align="center">0.9718</td>
<td align="center">0.9914</td>
<td align="center">0.9960</td>
<td align="center">0.9572</td>
</tr>
<tr>
<td align="left">&#x2003;<inline-formula id="inf133">
<mml:math id="m143">
<mml:mrow>
<mml:mtext>sd</mml:mtext>
<mml:mrow>
<mml:mo>(</mml:mo>
<mml:mi mathvariant="normal">&#x3f5;</mml:mi>
<mml:mo>)</mml:mo>
</mml:mrow>
<mml:mo>&#x3d;</mml:mo>
<mml:mn>1</mml:mn>
</mml:mrow>
</mml:math>
</inline-formula>
</td>
<td align="center">0.9370</td>
<td align="center">0.9503</td>
<td align="center">0.9575</td>
<td align="center">0.9557</td>
<td align="center">0.9781</td>
<td align="center">0.9915</td>
<td align="center">1.0000</td>
<td align="center">0.9689</td>
</tr>
<tr>
<td align="left">&#x2003;<inline-formula id="inf134">
<mml:math id="m144">
<mml:mrow>
<mml:mtext>sd</mml:mtext>
<mml:mrow>
<mml:mo>(</mml:mo>
<mml:mi mathvariant="normal">&#x3f5;</mml:mi>
<mml:mo>)</mml:mo>
</mml:mrow>
<mml:mo>&#x3d;</mml:mo>
<mml:mn>3</mml:mn>
</mml:mrow>
</mml:math>
</inline-formula>
</td>
<td align="center">0.8730</td>
<td align="center">0.9353</td>
<td align="center">0.9576</td>
<td align="center">0.9614</td>
<td align="center">0.9837</td>
<td align="center">0.9821</td>
<td align="center">0.9799</td>
<td align="center">0.9741</td>
</tr>
<tr>
<td align="left">&#x2003;<inline-formula id="inf135">
<mml:math id="m145">
<mml:mrow>
<mml:mtext>sd</mml:mtext>
<mml:mrow>
<mml:mo>(</mml:mo>
<mml:mi mathvariant="normal">&#x3f5;</mml:mi>
<mml:mo>)</mml:mo>
</mml:mrow>
<mml:mo>&#x3d;</mml:mo>
<mml:mn>4</mml:mn>
</mml:mrow>
</mml:math>
</inline-formula>
</td>
<td align="center">0.8004</td>
<td align="center">0.9289</td>
<td align="center">0.9397</td>
<td align="center">0.9692</td>
<td align="center">0.9928</td>
<td align="center">0.9771</td>
<td align="center">0.9799</td>
<td align="center">0.9686</td>
</tr>
<tr>
<td align="left">&#x2003;<inline-formula id="inf136">
<mml:math id="m146">
<mml:mrow>
<mml:mtext>sd</mml:mtext>
<mml:mrow>
<mml:mo>(</mml:mo>
<mml:mi mathvariant="normal">&#x3f5;</mml:mi>
<mml:mo>)</mml:mo>
</mml:mrow>
<mml:mo>&#x3d;</mml:mo>
<mml:mn>5</mml:mn>
</mml:mrow>
</mml:math>
</inline-formula>
</td>
<td align="center">0.8327</td>
<td align="center">0.9176</td>
<td align="center">0.9444</td>
<td align="center">0.9557</td>
<td align="center">0.9876</td>
<td align="center">0.9790</td>
<td align="center">0.9841</td>
<td align="center">0.9676</td>
</tr>
<tr>
<td align="left">&#x2003;<inline-formula id="inf137">
<mml:math id="m147">
<mml:mrow>
<mml:mtext>sd</mml:mtext>
<mml:mrow>
<mml:mo>(</mml:mo>
<mml:mi mathvariant="normal">&#x3f5;</mml:mi>
<mml:mo>)</mml:mo>
</mml:mrow>
<mml:mo>&#x3d;</mml:mo>
<mml:mn>6.366</mml:mn>
</mml:mrow>
</mml:math>
</inline-formula>
</td>
<td align="center">0.8970</td>
<td align="center">0.8755</td>
<td align="center">0.9347</td>
<td align="center">0.9625</td>
<td align="center">0.9820</td>
<td align="center">0.9745</td>
<td align="center">0.9714</td>
<td align="center">0.9731</td>
</tr>
<tr>
<td colspan="9" align="left">
<bold>Number</bold> <inline-formula id="inf138">
<mml:math id="m148">
<mml:mrow>
<mml:msub>
<mml:mtext>&#x393;</mml:mtext>
<mml:mn>1</mml:mn>
</mml:msub>
</mml:mrow>
</mml:math>
</inline-formula> <bold>of different extracted supports</bold> <inline-formula id="inf139">
<mml:math id="m149">
<mml:mrow>
<mml:mrow>
<mml:mo>(</mml:mo>
<mml:mrow>
<mml:mtext>%</mml:mtext>
<mml:msub>
<mml:mi mathvariant="script">S</mml:mi>
<mml:mrow>
<mml:mi>s</mml:mi>
<mml:mi>e</mml:mi>
<mml:mi>l</mml:mi>
</mml:mrow>
</mml:msub>
</mml:mrow>
<mml:mo>)</mml:mo>
</mml:mrow>
</mml:mrow>
</mml:math>
</inline-formula>
</td>
</tr>
<tr>
<td align="left">&#x2003;<inline-formula id="inf140">
<mml:math id="m150">
<mml:mrow>
<mml:mtext>sd</mml:mtext>
<mml:mrow>
<mml:mo>(</mml:mo>
<mml:mi mathvariant="normal">&#x3f5;</mml:mi>
<mml:mo>)</mml:mo>
</mml:mrow>
<mml:mo>&#x3d;</mml:mo>
<mml:mn>0.5</mml:mn>
</mml:mrow>
</mml:math>
</inline-formula>
</td>
<td align="center">20 (17)</td>
<td align="center">23 (10)</td>
<td align="center">6 (73.2)</td>
<td align="center">11 (30)</td>
<td align="center">16 (35)</td>
<td align="center">5 (56)</td>
<td align="center">2 (90)</td>
<td align="center">4 (48)</td>
</tr>
<tr>
<td align="left">&#x2003;<inline-formula id="inf141">
<mml:math id="m151">
<mml:mrow>
<mml:mtext>sd</mml:mtext>
<mml:mrow>
<mml:mo>(</mml:mo>
<mml:mi mathvariant="normal">&#x3f5;</mml:mi>
<mml:mo>)</mml:mo>
</mml:mrow>
<mml:mo>&#x3d;</mml:mo>
<mml:mn>1</mml:mn>
</mml:mrow>
</mml:math>
</inline-formula>
</td>
<td align="center">18 (30)</td>
<td align="center">8 (57)</td>
<td align="center">7 (38)</td>
<td align="center">11 (26)</td>
<td align="center">17 (23)</td>
<td align="center">6 (53)</td>
<td align="center">1 (100)</td>
<td align="center">5 (49)</td>
</tr>
<tr>
<td align="left">&#x2003;<inline-formula id="inf142">
<mml:math id="m152">
<mml:mrow>
<mml:mtext>sd</mml:mtext>
<mml:mrow>
<mml:mo>(</mml:mo>
<mml:mi mathvariant="normal">&#x3f5;</mml:mi>
<mml:mo>)</mml:mo>
</mml:mrow>
<mml:mo>&#x3d;</mml:mo>
<mml:mn>3</mml:mn>
</mml:mrow>
</mml:math>
</inline-formula>
</td>
<td align="center">41 (19)</td>
<td align="center">16 (47)</td>
<td align="center">14 (39)</td>
<td align="center">26 (16)</td>
<td align="center">6 (44)</td>
<td align="center">5 (34)</td>
<td align="center">4 (90)</td>
<td align="center">4 (58)</td>
</tr>
<tr>
<td align="left">&#x2003;<inline-formula id="inf143">
<mml:math id="m153">
<mml:mrow>
<mml:mtext>sd</mml:mtext>
<mml:mrow>
<mml:mo>(</mml:mo>
<mml:mi mathvariant="normal">&#x3f5;</mml:mi>
<mml:mo>)</mml:mo>
</mml:mrow>
<mml:mo>&#x3d;</mml:mo>
<mml:mn>4</mml:mn>
</mml:mrow>
</mml:math>
</inline-formula>
</td>
<td align="center">96 (2)</td>
<td align="center">6 (58)</td>
<td align="center">11 (26)</td>
<td align="center">18 (30)</td>
<td align="center">6 (48)</td>
<td align="center">12 (48)</td>
<td align="center">4 (67)</td>
<td align="center">4 (69)</td>
</tr>
<tr>
<td align="left">&#x2003;<inline-formula id="inf144">
<mml:math id="m154">
<mml:mrow>
<mml:mtext>sd</mml:mtext>
<mml:mrow>
<mml:mo>(</mml:mo>
<mml:mi mathvariant="normal">&#x3f5;</mml:mi>
<mml:mo>)</mml:mo>
</mml:mrow>
<mml:mo>&#x3d;</mml:mo>
<mml:mn>5</mml:mn>
</mml:mrow>
</mml:math>
</inline-formula>
</td>
<td align="center">88 (3)</td>
<td align="center">11 (48)</td>
<td align="center">17 (33)</td>
<td align="center">12 (22)</td>
<td align="center">6 (54)</td>
<td align="center">11 (25.33)</td>
<td align="center">4 (45)</td>
<td align="center">3 (64)</td>
</tr>
<tr>
<td align="left">&#x2003;<inline-formula id="inf145">
<mml:math id="m155">
<mml:mrow>
<mml:mtext>sd</mml:mtext>
<mml:mrow>
<mml:mo>(</mml:mo>
<mml:mi mathvariant="normal">&#x3f5;</mml:mi>
<mml:mo>)</mml:mo>
</mml:mrow>
<mml:mo>&#x3d;</mml:mo>
<mml:mn>6.366</mml:mn>
</mml:mrow>
</mml:math>
</inline-formula>
</td>
<td align="center">47 (10)</td>
<td align="center">25 (24)</td>
<td align="center">9 (57)</td>
<td align="center">10 (39)</td>
<td align="center">5 (38)</td>
<td align="center">10 (18)</td>
<td align="center">4 (46)</td>
<td align="center">3 (64)</td>
</tr>
<tr>
<td colspan="9" align="left">
<bold>Number of predictors in</bold> <inline-formula id="inf146">
<mml:math id="m156">
<mml:mrow>
<mml:msub>
<mml:mi mathvariant="script">S</mml:mi>
<mml:mrow>
<mml:mi>s</mml:mi>
<mml:mi>e</mml:mi>
<mml:mi>l</mml:mi>
</mml:mrow>
</mml:msub>
</mml:mrow>
</mml:math>
</inline-formula> <inline-formula id="inf147">
<mml:math id="m157">
<mml:mrow>
<mml:mrow>
<mml:mo>(</mml:mo>
<mml:mrow>
<mml:msub>
<mml:mi>K</mml:mi>
<mml:mrow>
<mml:mi>s</mml:mi>
<mml:mi>e</mml:mi>
<mml:mi>l</mml:mi>
</mml:mrow>
</mml:msub>
</mml:mrow>
<mml:mo>)</mml:mo>
</mml:mrow>
</mml:mrow>
</mml:math>
</inline-formula>
</td>
</tr>
<tr>
<td align="left">&#x2003;<inline-formula id="inf148">
<mml:math id="m158">
<mml:mrow>
<mml:mtext>sd</mml:mtext>
<mml:mrow>
<mml:mo>(</mml:mo>
<mml:mi mathvariant="normal">&#x3f5;</mml:mi>
<mml:mo>)</mml:mo>
</mml:mrow>
<mml:mo>&#x3d;</mml:mo>
<mml:mn>0.5</mml:mn>
</mml:mrow>
</mml:math>
</inline-formula>
</td>
<td align="center">11 (5)</td>
<td align="center">6 (4)</td>
<td align="center">5 (5)</td>
<td align="center">8 (5)</td>
<td align="center">7 (6)</td>
<td align="center">4 (4)</td>
<td align="center">4 (4)</td>
<td align="center">4 (4)</td>
</tr>
<tr>
<td align="left">&#x2003;<inline-formula id="inf149">
<mml:math id="m159">
<mml:mrow>
<mml:mtext>sd</mml:mtext>
<mml:mrow>
<mml:mo>(</mml:mo>
<mml:mi mathvariant="normal">&#x3f5;</mml:mi>
<mml:mo>)</mml:mo>
</mml:mrow>
<mml:mo>&#x3d;</mml:mo>
<mml:mn>1</mml:mn>
</mml:mrow>
</mml:math>
</inline-formula>
</td>
<td align="center">10 (4)</td>
<td align="center">9 (5)</td>
<td align="center">9 (5)</td>
<td align="center">9 (5)</td>
<td align="center">6 (4)</td>
<td align="center">4 (4)</td>
<td align="center">4 (4)</td>
<td align="center">5 (3)</td>
</tr>
<tr>
<td align="left">&#x2003;<inline-formula id="inf150">
<mml:math id="m160">
<mml:mrow>
<mml:mtext>sd</mml:mtext>
<mml:mrow>
<mml:mo>(</mml:mo>
<mml:mi mathvariant="normal">&#x3f5;</mml:mi>
<mml:mo>)</mml:mo>
</mml:mrow>
<mml:mo>&#x3d;</mml:mo>
<mml:mn>3</mml:mn>
</mml:mrow>
</mml:math>
</inline-formula>
</td>
<td align="center">16 (4)</td>
<td align="center">11 (6)</td>
<td align="center">8 (6)</td>
<td align="center">7 (4)</td>
<td align="center">6 (4)</td>
<td align="center">8 (4)</td>
<td align="center">6 (4)</td>
<td align="center">7 (4)</td>
</tr>
<tr>
<td align="left">&#x2003;<inline-formula id="inf151">
<mml:math id="m161">
<mml:mrow>
<mml:mtext>sd</mml:mtext>
<mml:mrow>
<mml:mo>(</mml:mo>
<mml:mi mathvariant="normal">&#x3f5;</mml:mi>
<mml:mo>)</mml:mo>
</mml:mrow>
<mml:mo>&#x3d;</mml:mo>
<mml:mn>4</mml:mn>
</mml:mrow>
</mml:math>
</inline-formula>
</td>
<td align="center">24 (3)</td>
<td align="center">11 (5)</td>
<td align="center">10 (5)</td>
<td align="center">6 (4)</td>
<td align="center">5 (3)</td>
<td align="center">6 (4)</td>
<td align="center">6 (4)</td>
<td align="center">7 (4)</td>
</tr>
<tr>
<td align="left">&#x2003;<inline-formula id="inf152">
<mml:math id="m162">
<mml:mrow>
<mml:mtext>sd</mml:mtext>
<mml:mrow>
<mml:mo>(</mml:mo>
<mml:mi mathvariant="normal">&#x3f5;</mml:mi>
<mml:mo>)</mml:mo>
</mml:mrow>
<mml:mo>&#x3d;</mml:mo>
<mml:mn>5</mml:mn>
</mml:mrow>
</mml:math>
</inline-formula>
</td>
<td align="center">21 (3)</td>
<td align="center">12 (5)</td>
<td align="center">10 (5)</td>
<td align="center">9 (4)</td>
<td align="center">3 (3)</td>
<td align="center">5 (3)</td>
<td align="center">3 (3)</td>
<td align="center">7 (4)</td>
</tr>
<tr>
<td align="left">&#x2003;<inline-formula id="inf153">
<mml:math id="m163">
<mml:mrow>
<mml:mtext>sd</mml:mtext>
<mml:mrow>
<mml:mo>(</mml:mo>
<mml:mi mathvariant="normal">&#x3f5;</mml:mi>
<mml:mo>)</mml:mo>
</mml:mrow>
<mml:mo>&#x3d;</mml:mo>
<mml:mn>6.366</mml:mn>
</mml:mrow>
</mml:math>
</inline-formula>
</td>
<td align="center">15 (3)</td>
<td align="center">16 (4)</td>
<td align="center">11 (4)</td>
<td align="center">7 (4)</td>
<td align="center">3 (3)</td>
<td align="center">4 (3)</td>
<td align="center">5 (3)</td>
<td align="center">7 (4)</td>
</tr>
</tbody>
</table>
</table-wrap>
<p>Concerning the number of different models, results related to lasso regression are the same as those concerning the number of different supports. Only the result for <inline-formula id="inf154">
<mml:math id="m164">
<mml:mrow>
<mml:mtext>sd</mml:mtext>
<mml:mrow>
<mml:mo>(</mml:mo>
<mml:mi mathvariant="normal">&#x3f5;</mml:mi>
<mml:mo>)</mml:mo>
</mml:mrow>
<mml:mo>&#x3d;</mml:mo>
<mml:mn>0.5</mml:mn>
</mml:mrow>
</mml:math>
</inline-formula> differs, since one trial finished with a fifth value of the sparsity parameter and the same support as the model that was selected. Therefore, only results concerning models established using SPLS methods are summarized in <xref ref-type="table" rid="T5">Table&#x20;5</xref>. In the case of bootstrap-based techniques, supports and models are similar, since no sparsity parameter is needed.</p>
<table-wrap id="T5" position="float">
<label>TABLE 5</label>
<caption>
<p>Results for SPLS model stability.</p>
</caption>
<table>
<thead valign="top">
<tr>
<th align="left"/>
<th colspan="2" align="center">&#x23; <inline-formula id="inf155">
<mml:math id="m165">
<mml:mrow>
<mml:mrow>
<mml:mo>(</mml:mo>
<mml:mrow>
<mml:msub>
<mml:mi>&#x3b7;</mml:mi>
<mml:mrow>
<mml:mi mathvariant="bold-italic">o</mml:mi>
<mml:mi mathvariant="bold-italic">p</mml:mi>
<mml:mi mathvariant="bold-italic">t</mml:mi>
</mml:mrow>
</mml:msub>
<mml:mo>,</mml:mo>
<mml:msub>
<mml:mi mathvariant="bold-italic">K</mml:mi>
<mml:mrow>
<mml:mi mathvariant="bold-italic">o</mml:mi>
<mml:mi mathvariant="bold-italic">p</mml:mi>
<mml:mi mathvariant="bold-italic">t</mml:mi>
</mml:mrow>
</mml:msub>
</mml:mrow>
<mml:mo>)</mml:mo>
</mml:mrow>
</mml:mrow>
</mml:math>
</inline-formula> <inline-formula id="inf156">
<mml:math id="m166">
<mml:mrow>
<mml:mrow>
<mml:mo>(</mml:mo>
<mml:mrow>
<mml:mtext>%</mml:mtext>
<mml:msub>
<mml:mrow>
<mml:mrow>
<mml:mo>(</mml:mo>
<mml:mrow>
<mml:mi>&#x3b7;</mml:mi>
<mml:mo>,</mml:mo>
<mml:mi mathvariant="bold-italic">K</mml:mi>
</mml:mrow>
<mml:mo>)</mml:mo>
</mml:mrow>
</mml:mrow>
<mml:mrow>
<mml:mi mathvariant="bold-italic">s</mml:mi>
<mml:mi mathvariant="bold-italic">e</mml:mi>
<mml:mi mathvariant="bold-italic">l</mml:mi>
</mml:mrow>
</mml:msub>
</mml:mrow>
<mml:mo>)</mml:mo>
</mml:mrow>
</mml:mrow>
</mml:math>
</inline-formula>
</th>
<th colspan="2" align="center">
<inline-formula id="inf157">
<mml:math id="m167">
<mml:mrow>
<mml:msub>
<mml:mtext>&#x393;</mml:mtext>
<mml:mn>2</mml:mn>
</mml:msub>
<mml:mrow>
<mml:mo>(</mml:mo>
<mml:mrow>
<mml:mtext>%</mml:mtext>
<mml:msub>
<mml:mi mathvariant="normal">&#x2133;</mml:mi>
<mml:mrow>
<mml:mi mathvariant="bold-italic">s</mml:mi>
<mml:mi mathvariant="bold-italic">e</mml:mi>
<mml:mi mathvariant="bold-italic">l</mml:mi>
</mml:mrow>
</mml:msub>
</mml:mrow>
<mml:mo>)</mml:mo>
</mml:mrow>
</mml:mrow>
</mml:math>
</inline-formula>
</th>
</tr>
<tr>
<th align="left"/>
<th align="center">SPLS CV</th>
<th align="center">SPLS BootYT</th>
<th align="center">SPLS CV</th>
<th align="center">SPLS BootYT</th>
</tr>
</thead>
<tbody valign="top">
<tr>
<td align="left">
<inline-formula id="inf158">
<mml:math id="m168">
<mml:mrow>
<mml:mtext>sd</mml:mtext>
<mml:mrow>
<mml:mo>(</mml:mo>
<mml:mi mathvariant="normal">&#x3f5;</mml:mi>
<mml:mo>)</mml:mo>
</mml:mrow>
<mml:mo>&#x3d;</mml:mo>
<mml:mn>0.5</mml:mn>
</mml:mrow>
</mml:math>
</inline-formula>
</td>
<td align="char" char="(">9 (46)</td>
<td align="char" char="(">3 (81)</td>
<td align="char" char="(">8 (55)</td>
<td align="char" char="(">2 (90)</td>
</tr>
<tr>
<td align="left">
<inline-formula id="inf159">
<mml:math id="m169">
<mml:mrow>
<mml:mtext>sd</mml:mtext>
<mml:mrow>
<mml:mo>(</mml:mo>
<mml:mi mathvariant="normal">&#x3f5;</mml:mi>
<mml:mo>)</mml:mo>
</mml:mrow>
<mml:mo>&#x3d;</mml:mo>
<mml:mn>1</mml:mn>
</mml:mrow>
</mml:math>
</inline-formula>
</td>
<td align="char" char="(">13 (36)</td>
<td align="char" char="(">2 (73)</td>
<td align="char" char="(">9 (45)</td>
<td align="char" char="(">1 (100)</td>
</tr>
<tr>
<td align="left">
<inline-formula id="inf160">
<mml:math id="m170">
<mml:mrow>
<mml:mtext>sd</mml:mtext>
<mml:mrow>
<mml:mo>(</mml:mo>
<mml:mi mathvariant="normal">&#x3f5;</mml:mi>
<mml:mo>)</mml:mo>
</mml:mrow>
<mml:mo>&#x3d;</mml:mo>
<mml:mn>3</mml:mn>
</mml:mrow>
</mml:math>
</inline-formula>
</td>
<td align="char" char="(">13 (17)</td>
<td align="char" char="(">5 (59)</td>
<td align="char" char="(">5 (34)</td>
<td align="char" char="(">4 (90)</td>
</tr>
<tr>
<td align="left">
<inline-formula id="inf161">
<mml:math id="m171">
<mml:mrow>
<mml:mtext>sd</mml:mtext>
<mml:mrow>
<mml:mo>(</mml:mo>
<mml:mi mathvariant="normal">&#x3f5;</mml:mi>
<mml:mo>)</mml:mo>
</mml:mrow>
<mml:mo>&#x3d;</mml:mo>
<mml:mn>4</mml:mn>
</mml:mrow>
</mml:math>
</inline-formula>
</td>
<td align="char" char="(">15 (27)</td>
<td align="char" char="(">6 (37)</td>
<td align="char" char="(">12 (48)</td>
<td align="char" char="(">4 (67)</td>
</tr>
<tr>
<td align="left">
<inline-formula id="inf162">
<mml:math id="m172">
<mml:mrow>
<mml:mtext>sd</mml:mtext>
<mml:mrow>
<mml:mo>(</mml:mo>
<mml:mi mathvariant="normal">&#x3f5;</mml:mi>
<mml:mo>)</mml:mo>
</mml:mrow>
<mml:mo>&#x3d;</mml:mo>
<mml:mn>5</mml:mn>
</mml:mrow>
</mml:math>
</inline-formula>
</td>
<td align="char" char="(">20 (19.33)</td>
<td align="char" char="(">6 (45)</td>
<td align="char" char="(">12 (25.33)</td>
<td align="char" char="(">4 (45)</td>
</tr>
<tr>
<td align="left">
<inline-formula id="inf163">
<mml:math id="m173">
<mml:mrow>
<mml:mtext>sd</mml:mtext>
<mml:mrow>
<mml:mo>(</mml:mo>
<mml:mi mathvariant="normal">&#x3f5;</mml:mi>
<mml:mo>)</mml:mo>
</mml:mrow>
<mml:mo>&#x3d;</mml:mo>
<mml:mn>6.366</mml:mn>
</mml:mrow>
</mml:math>
</inline-formula>
</td>
<td align="char" char="(">16 (18)</td>
<td align="char" char="(">4 (46)</td>
<td align="char" char="(">11 (18)</td>
<td align="char" char="(">4 (46)</td>
</tr>
</tbody>
</table>
</table-wrap>
<p>Concerning the three bootstrap-based techniques and in the light of accuracy results (<xref ref-type="table" rid="T4">Table&#x20;4</xref>), BootYT outperforms both the others, except in the case where <inline-formula id="inf164">
<mml:math id="m174">
<mml:mrow>
<mml:mtext>sd</mml:mtext>
<mml:mrow>
<mml:mo>(</mml:mo>
<mml:mi mathvariant="normal">&#x3f5;</mml:mi>
<mml:mo>)</mml:mo>
</mml:mrow>
<mml:mo>&#x3d;</mml:mo>
<mml:mn>0.5</mml:mn>
</mml:mrow>
</mml:math>
</inline-formula>, for which BIC should be used. This exception is confirmed through the comparison of the dynamic bootstrap-based methods, where the <inline-formula id="inf165">
<mml:math id="m175">
<mml:mrow>
<mml:mtext>sd</mml:mtext>
<mml:mrow>
<mml:mo>(</mml:mo>
<mml:mi mathvariant="normal">&#x3f5;</mml:mi>
<mml:mo>)</mml:mo>
</mml:mrow>
<mml:mo>&#x3d;</mml:mo>
<mml:mn>0.5</mml:mn>
</mml:mrow>
</mml:math>
</inline-formula> case is the only one where using the BIC criterion also represents the most relevant choice. This phenomenon matches with conclusions obtained by Magnanensi et&#x20;al. [<xref ref-type="bibr" rid="B32">32</xref>], in that the BIC criterion is well-designed for small values of noise variance, while BootYT outperforms it for non-negligible levels of random noise. The use of the <inline-formula id="inf166">
<mml:math id="m176">
<mml:mrow>
<mml:msup>
<mml:mi>Q</mml:mi>
<mml:mn>2</mml:mn>
</mml:msup>
</mml:mrow>
</mml:math>
</inline-formula> criterion for selecting the number of components would appear never be a reliable option, so combining this criterion with our new dynamical approach was not done. The accuracy values highlight the fact that our new method is always an improvement over the original one. Concerning the two versions of SPLS regression, both gave similar accuracy. These stand to perform better than others for the smallest levels of random noise variance, while BootYTdyn outperforms all others for the highest values of noise variability.</p>
<p>Based on results introduced in <xref ref-type="table" rid="T4">Tables 4</xref>,<xref ref-type="table" rid="T5">5</xref>, the <inline-formula id="inf167">
<mml:math id="m177">
<mml:mrow>
<mml:msup>
<mml:mi>Q</mml:mi>
<mml:mn>2</mml:mn>
</mml:msup>
</mml:mrow>
</mml:math>
</inline-formula>-based approach for predictor selection has non-negligable stability issues, providing between 18 and 96 different models in the 100 simulations. Depending on both the noise variability and the criterion combined with our dynamic approach, the latter improves over the original one by stabilizing the choice of sparse support. Cases where this is observed match with previous conclusions established when analyzing accuracy results, namely both the BIC-based dynamic approach for small values of noise variability, and BootYTdyn used for datasets with non-negligible noise variances. This strengthens the conclusion that BIC is well-designed for small values of noise variance, while BootYT performs best for non-negligible noise. As for SPLS methods, our new bootstrap-based version gains in stability in that it retains fewer different optimal pairs <inline-formula id="inf168">
<mml:math id="m178">
<mml:mrow>
<mml:mrow>
<mml:mo>(</mml:mo>
<mml:mrow>
<mml:msub>
<mml:mi>&#x3b7;</mml:mi>
<mml:mrow>
<mml:mi>o</mml:mi>
<mml:mi>p</mml:mi>
<mml:mi>t</mml:mi>
</mml:mrow>
</mml:msub>
<mml:mo>,</mml:mo>
<mml:msub>
<mml:mi>K</mml:mi>
<mml:mrow>
<mml:mi>o</mml:mi>
<mml:mi>p</mml:mi>
<mml:mi>t</mml:mi>
</mml:mrow>
</mml:msub>
</mml:mrow>
<mml:mo>)</mml:mo>
</mml:mrow>
</mml:mrow>
</mml:math>
</inline-formula>, and also sparse models, than the original does. Moreover, since <inline-formula id="inf169">
<mml:math id="m179">
<mml:mrow>
<mml:msub>
<mml:mtext>&#x393;</mml:mtext>
<mml:mn>1</mml:mn>
</mml:msub>
<mml:mo>&#x3d;</mml:mo>
<mml:msub>
<mml:mtext>&#x393;</mml:mtext>
<mml:mn>2</mml:mn>
</mml:msub>
</mml:mrow>
</mml:math>
</inline-formula> for all studied datasets, this directly implies that it retains a unique optimal number of components for each sparse support. It thus permits a choice of optimal model in a more reliable way than using the CV-based technique. Lastly, lasso regression has good stability performances, leading it, BootYTdyn and SPLS BootYT to be recommended when stability is important.</p>
<p>The last descriptive statistic concerns the number of significant predictors retained (<xref ref-type="table" rid="T4">Table&#x20;4</xref>). The <inline-formula id="inf170">
<mml:math id="m180">
<mml:mrow>
<mml:msup>
<mml:mi>Q</mml:mi>
<mml:mn>2</mml:mn>
</mml:msup>
</mml:mrow>
</mml:math>
</inline-formula>-based approach is the least sparse, selecting the highest number of covariates. Both BICdyn BootYTdyn improve respectively BIC and BootYT, showing better accuracy related to their selected support. Indeed, the four expected covariates are always included in these selected supports. Once more, BootYT can be recommended for datasets with non-negligible random noise variability compared with corrected BIC, which has to be applied for small values of random variance. Indeed, both the BIC and BICdyn approaches, by retaining globally increasing numbers of predictors while the random noise standard deviation increases, lead us to suppose that they use some predictors to model this random noise, leading to over-fitting. On the contrary, while BootYT selects a stable number of significant predictors, BootYTdyn selects a decreasing number of significant predictors as the random noise standard deviation increases, which is expected. As a confirmation of over-fitting issues with BIC, we present the 10-fold CV-based MSE in <xref ref-type="fig" rid="F2">Figure&#x20;2</xref>.</p>
<fig id="F2" position="float">
<label>FIGURE 2</label>
<caption>
<p>Boxplots of 10-CV MSE based on <bold>y</bold> with&#x20;noise.</p>
</caption>
<graphic xlink:href="fams-07-693126-g002.tif"/>
</fig>
<p>These results tend to confirm our suspicion of over-fitting since, except for results related to BootYTdyn, the others have MSE that do not match with the theoretical random noise variances, suggesting that a part of the noise is being modeled. As for the two SPLS methods, there is no pertinent difference to mention, both of them concluding on similar numbers of selected predictors. Lastly, like in BIC-based approaches, the lasso extracts an increasing number of significant predictors.</p>
<p>As a first conclusion, we can thus reasonably conclude that, based on these first simulation results, BootYTdyn and SPLS BootYT should be used in practice.</p>
</sec>
<sec id="s3-3-5">
<title>3.3.5 Complexity and Predictive Ability Results</title>
<p>To confirm and strengthen the conclusions of <xref ref-type="sec" rid="s3-3-2">Section 3.3.2</xref>, we will now focus on the predictive abilities of the models selected by the various approaches. We calculate one hundred times the 10-fold CV MSE based on the original simulated response values (without noise) of the various selected models. These results thus reflect the accuracy in predicting the original information by leaving out random noise. We also compute the Predictive MSE (PMSE) based on the test set, by using its simulated response <inline-formula id="inf171">
<mml:math id="m181">
<mml:mrow>
<mml:msub>
<mml:mtext>y</mml:mtext>
<mml:mrow>
<mml:mtext>test</mml:mtext>
</mml:mrow>
</mml:msub>
</mml:mrow>
</mml:math>
</inline-formula> without including noise. Lastly, we extract the DoF of each selected sparse model (<inline-formula id="inf172">
<mml:math id="m182">
<mml:mrow>
<mml:msub>
<mml:mrow>
<mml:mtext>DoF</mml:mtext>
</mml:mrow>
<mml:mrow>
<mml:mi>s</mml:mi>
<mml:mi>e</mml:mi>
<mml:mi>l</mml:mi>
</mml:mrow>
</mml:msub>
</mml:mrow>
</mml:math>
</inline-formula>) to compare the respective complexities.</p>
<p>Graphical results for these statistics, related to <inline-formula id="inf173">
<mml:math id="m183">
<mml:mrow>
<mml:msup>
<mml:mi>Q</mml:mi>
<mml:mn>2</mml:mn>
</mml:msup>
</mml:mrow>
</mml:math>
</inline-formula>, BIC and BootYT, are shown in the first row of <xref ref-type="fig" rid="F3">Figure&#x20;3</xref>.</p>
<fig id="F3" position="float">
<label>FIGURE 3</label>
<caption>
<p>From left to right: DoF of the extracted sparse models, PMSE based on <inline-formula id="inf174">
<mml:math id="m184">
<mml:mrow>
<mml:msub>
<mml:mi mathvariant="bold">y</mml:mi>
<mml:mrow>
<mml:mtext>test</mml:mtext>
</mml:mrow>
</mml:msub>
</mml:mrow>
</mml:math>
</inline-formula> without noise, and boxplots of 10-CV MSE based on <bold>y</bold> without&#x20;noise.</p>
</caption>
<graphic xlink:href="fams-07-693126-g003.tif"/>
</fig>
<p>The evolution of estimated DoF both highlights and confirms BIC and <inline-formula id="inf175">
<mml:math id="m185">
<mml:mrow>
<mml:msup>
<mml:mi>Q</mml:mi>
<mml:mn>2</mml:mn>
</mml:msup>
</mml:mrow>
</mml:math>
</inline-formula> over-fitting issues. Indeed, as the random noise standard deviation increases, these methods globally build sparse models with increasing complexity. Thus, they model increasing parts of the inserted random noise, implying poor predictive abilities of their selected models compared to those obtained by applying BootYT. This is confirmed through higher values of PMSE and CV MSE, especially for datasets with non-negligible random noise variability. These results confirm the conclusions from the previous section in that using the <inline-formula id="inf176">
<mml:math id="m186">
<mml:mrow>
<mml:msup>
<mml:mi>Q</mml:mi>
<mml:mn>2</mml:mn>
</mml:msup>
</mml:mrow>
</mml:math>
</inline-formula> criterion for selecting the number of PLS components should be avoided, and that BootYT outperforms corrected BIC, except for responses with negligible random noise levels. Therefore, only BIC and BootYT are retained for further comparison.</p>
<p>The results shown in the second row of <xref ref-type="fig" rid="F3">Figure&#x20;3</xref> highlight that BootYTdyn is the only method that models with decreasing DoF, ensuring a complexity reduction suitable to the avoidance of prediction issues. Indeed, BootYTdyn selects models with the lowest PMSE and 10-CV&#x20;MSE.</p>
<p>In light of these results, only BootYTyn is retained for further comparison. Comparing the two SPLS implementations with respect to their predictive abilities lead us to recommend SPLS BootYT, since models selected by this bootstrap-adapted SPLS technique feature comparable if not lower PMSE and 10-CV MSE (third row of <xref ref-type="fig" rid="F3">Figure&#x20;3</xref>). Let us clarify that, in order to ensure relevant comparisons, we used ordinary PLS regressions with both the support and the number of components selected by the SPLS methods, and not SPLS methods with selected tuning parameters, for computing the 10-CV&#x20;MSE.</p>
<p>For datasets characterized by a low level of random noise variability in <inline-formula id="inf177">
<mml:math id="m187">
<mml:mtext>y</mml:mtext>
</mml:math>
</inline-formula>, SPLS BootYT builds models inducing the smallest CV MSE and PMSE. By focusing on non-negligible random variability, BootYTdyn gives the smallest predictive errors, thus showing the high level of robustness of this approach against random noise. This robustness is, as previously mentioned, due to its decreasing number of significant predictors, and also components, leading to decreasing DoF, i.e.,&#x20;a loss of complexity.</p>
<p>Lastly, we compare the two retained methods BootYTdyn and SPLS BootYT, and the lasso is run. As for SPLS methods, in order to perform relevant comparisons, the supports extracted by the lasso are used as sets of covariates for a PLS regression. The number of PLS components is then established by performing one hundred times their selection using the bootstrap-based stopping criterion; the number of components related to the highest occurrence rate was selected. In this way, results shown in the fourth row of <xref ref-type="fig" rid="F3">Figure&#x20;3</xref> concerning 10-CV MSE give the predictive ability of the extracted supports for PLS regression. In order to provide a clear picture of the impact of this choice, the PMSE obtained through the lasso is also displayed on the fourth row of <xref ref-type="fig" rid="F3">Figure&#x20;3</xref>. These approaches are referred to as Lasso and Lasso.supp.</p>
<p>Except for negligible values of random noise variability in <bold>y</bold>, the support extracted by the lasso regression and applied as explanatory variables for PLS regression are linked to both the highest 10-CV MSE and highest PMSE. This is a direct consequence of the lasso&#x2019;s accuracy issues mentioned in <xref ref-type="sec" rid="s3-3-4">Section 3.3.4</xref>, notably the increasing number of extracted covariate while the random noise variability rises. However, performing predictions using the model obtained by the original lasso regression lead to the lowest PMSE values. This is due to the <inline-formula id="inf178">
<mml:math id="m188">
<mml:mrow>
<mml:msup>
<mml:mi>L</mml:mi>
<mml:mn>1</mml:mn>
</mml:msup>
</mml:mrow>
</mml:math>
</inline-formula> penalization which is applied on the vector of estimated parameters, thus permitting correction of the relative lack of accuracy of this technique.</p>
<p>To conclude, we summarize our conclusions in the following <xref ref-type="table" rid="T6">Table&#x20;6</xref>, and recommend certain approaches, depending on whether the initial aim was to select significant predictors or to obtain a sparse model with attractive predictive ability.</p>
<table-wrap id="T6" position="float">
<label>TABLE 6</label>
<caption>
<p>Recommended approaches.</p>
</caption>
<table>
<thead valign="top">
<tr>
<th align="left"/>
<th align="center">Accuracy</th>
<th align="center">Predictive ability</th>
</tr>
</thead>
<tbody valign="top">
<tr>
<td align="left">Low noise variability</td>
<td align="left">SPLS BootYT</td>
<td align="left">SPLS BootYT</td>
</tr>
<tr>
<td align="left">High noise variability</td>
<td align="left">BootYTdyn</td>
<td align="left">Lasso/BootYTdyn</td>
</tr>
</tbody>
</table>
</table-wrap>
</sec>
</sec>
</sec>
<sec id="s4">
<title>4 Real Dataset Application</title>
<p>In this section, we deal with the predictors matrix introduced in <xref ref-type="sec" rid="s3-3-1">Section 3.3.1</xref> and the original binary response vector. Five approaches for variable selection, adapted for the GPLS framework, are considered for comparison.<list list-type="simple">
<list-item>
<p>1. BootYT. The bootstrap-based method, introduced by Lazraq et&#x20;al. [<xref ref-type="bibr" rid="B20">20</xref>], combined with the bootstrap-based criterion [<xref ref-type="bibr" rid="B32">32</xref>] for pre-selecting the number of components.</p>
</list-item>
<list-item>
<p>2. BootYTdyn. The new dynamic bootstrap-based method combined with the bootstrap-based criterion for successive determinations of the number of components.</p>
</list-item>
<list-item>
<p>3. SGPLS CV. The original SGPLS method using 10-fold CV for tuning parameter selection&#x20;[<xref ref-type="bibr" rid="B21">21</xref>].</p>
</list-item>
<list-item>
<p>4. SGPLS BootYT. The new adapted SGPLS version using the bootstrap-based criterion for component selection.</p>
</list-item>
<list-item>
<p>5. RSGPLS. An approach developed by Durif et&#x20;al. [<xref ref-type="bibr" rid="B38">38</xref>]. It consists of adapting the SGPLS method by introducing a ridge penalty to ensure convergence of parameter estimations, and stability in hyper-parameter tuning. They also propose an adjustment of the <inline-formula id="inf179">
<mml:math id="m189">
<mml:mrow>
<mml:msub>
<mml:mi>L</mml:mi>
<mml:mn>1</mml:mn>
</mml:msub>
</mml:mrow>
</mml:math>
</inline-formula> constraint in order to further penalize less significant predictors. Hyper-parameters are tuned using CV&#x20;MSE.</p>
</list-item>
<list-item>
<p>6. Lasso. The adapted lasso regression for logistic framework, available in the R package <italic>glmnet</italic>, as a benchmark.</p>
</list-item>
</list>
</p>
<p>Concerning bootstrap-based approaches, the incorporation of PLS methodology into GLM, developed by Bastien et&#x20;al. [<xref ref-type="bibr" rid="B16">16</xref>], was used. Due to non-convergence issues for parameter estimations, some samples were excluded using a threshold for parameter estimations. Indeed, a model built using a bootstrap sample with at least one parameter estimate that is higher in absolute value than <inline-formula id="inf180">
<mml:math id="m190">
<mml:mrow>
<mml:msup>
<mml:mrow>
<mml:mn>10</mml:mn>
</mml:mrow>
<mml:mn>4</mml:mn>
</mml:msup>
</mml:mrow>
</mml:math>
</inline-formula> times the one (in absolute value as well) estimated on the original dataset, leads to the exclusion of this bootstrap sample. Thus, to ensure a sufficient number of relevant bootstrap samples, the preset number of computed samples was increased to <inline-formula id="inf181">
<mml:math id="m191">
<mml:mrow>
<mml:mi>R</mml:mi>
<mml:mo>&#x3d;</mml:mo>
<mml:mn>4000</mml:mn>
</mml:mrow>
</mml:math>
</inline-formula>. As for the lasso, three different loss functions for the establishment of the sparsity parameter using 10-fold CV were used: the number of misclassified values, the MSE, and the deviance. We refer to them as Lasso.Cl, Lasso.MSE and Lasso.Dev, respectively. The set of values defined for the sparsity parameter is preset by the &#x201c;glmnet&#x201d; package. For the SGPLS and RSGPLS approaches, the number of components <italic>K</italic> varies from 1 to 10, the sparsity parameter &#x3b7; varies from 0.04 to 0.99 (by steps of 0.05), and the ridge parameter involved in the RSGPLS technique is selected from 31 points <inline-formula id="inf182">
<mml:math id="m192">
<mml:mrow>
<mml:msub>
<mml:mrow>
<mml:mtext>log</mml:mtext>
</mml:mrow>
<mml:mrow>
<mml:mn>10</mml:mn>
</mml:mrow>
</mml:msub>
</mml:mrow>
</mml:math>
</inline-formula>-linearly spaced in the range <inline-formula id="inf183">
<mml:math id="m193">
<mml:mrow>
<mml:mrow>
<mml:mo>[</mml:mo>
<mml:mrow>
<mml:msup>
<mml:mrow>
<mml:mn>10</mml:mn>
</mml:mrow>
<mml:mrow>
<mml:mo>&#x2212;</mml:mo>
<mml:mn>2</mml:mn>
</mml:mrow>
</mml:msup>
<mml:mo>,</mml:mo>
<mml:msup>
<mml:mrow>
<mml:mn>10</mml:mn>
</mml:mrow>
<mml:mn>3</mml:mn>
</mml:msup>
</mml:mrow>
<mml:mo>]</mml:mo>
</mml:mrow>
</mml:mrow>
</mml:math>
</inline-formula>, as in Durif et&#x20;al.&#x20;[<xref ref-type="bibr" rid="B38">38</xref>].</p>
<p>Each method was performed one hundred times in order to obtain relevant results. However, due to the high observed stability of results extracted with the BootYTdyn approach, and in order to save computational time, this was only performed twenty times instead of a hundred.</p>
<p>Stability results for tuning parameter selection, except for the bootstrap-based methods, are shown in <xref ref-type="fig" rid="F4">Figure&#x20;4</xref> and <xref ref-type="table" rid="T7">Table&#x20;7</xref>.</p>
<fig id="F4" position="float">
<label>FIGURE 4</label>
<caption>
<p>Repartition of selected sets of tuning parameters over the 100 trials, for SGPLS CV <bold>(A)</bold> and SGPLS BootYT <bold>(B)</bold>.</p>
</caption>
<graphic xlink:href="fams-07-693126-g004.tif"/>
</fig>
<table-wrap id="T7" position="float">
<label>TABLE 7</label>
<caption>
<p>Number of different selected sets of tuning parameters over the 100 trials (rate of occurrence of the retained set of tuning parameters).</p>
</caption>
<table>
<thead valign="top">
<tr>
<th align="left">SGPLS CV</th>
<th align="center">SGPLS BootYT</th>
<th align="center">RSGPLS</th>
<th align="center">Lasso.Cl</th>
<th align="center">Lasso.MSE</th>
<th align="center">Lasso.Dev</th>
</tr>
</thead>
<tbody valign="top">
<tr>
<td align="left">44 (9)</td>
<td align="center">26 (16)</td>
<td align="center">64 (5)</td>
<td align="center">33 (8)</td>
<td align="center">6 (44)</td>
<td align="center">5 (48)</td>
</tr>
</tbody>
</table>
</table-wrap>
<p>Based on results summarized in <xref ref-type="table" rid="T7">Table&#x20;7</xref>, the lasso methodology, using CV-based MSE or deviance values for the selection of its hyper-parameter, is the most stable. This can be explained by the fact that only one hyper-parameter is involved, while the other techniques are based on two or three tuning parameters. Using the number of misclassified values for CV has to be avoided for stable selection of the sparsity parameter. Our SGPLS adaptation with the BootYT criterion improves reliability in selecting the set of tuning parameters, as previously observed in the PLS framework (<xref ref-type="sec" rid="s2-2">Sections 2.2</xref>, <xref ref-type="sec" rid="s3-3-4">3.3.4</xref>). Note that, concerning RSGPLS, three different sets of optimal parameters were extracted with maximal occurrence rate of five, all of them selecting the same set of predictors. Thus, the set of parameters which leads to the smallest number of misclassified values on the training dataset was retained. As already mentioned in <xref ref-type="sec" rid="s3-3-3">Section 3.3.3</xref>, extracting the same support does not necessarily lead to the same model when sparsity or ridge parameters are involved. Therefore, the numbers of sparse supports and models retained are summarized in <xref ref-type="table" rid="T8">Table&#x20;8</xref>.</p>
<table-wrap id="T8" position="float">
<label>TABLE 8</label>
<caption>
<p>Number <inline-formula id="inf184">
<mml:math id="m194">
<mml:mrow>
<mml:msub>
<mml:mtext>&#x393;</mml:mtext>
<mml:mn>1</mml:mn>
</mml:msub>
</mml:mrow>
</mml:math>
</inline-formula> of different extracted supports <inline-formula id="inf185">
<mml:math id="m195">
<mml:mrow>
<mml:mrow>
<mml:mo>(</mml:mo>
<mml:mrow>
<mml:mtext>%</mml:mtext>
<mml:msub>
<mml:mi mathvariant="script">S</mml:mi>
<mml:mrow>
<mml:mi>s</mml:mi>
<mml:mi>e</mml:mi>
<mml:mi>l</mml:mi>
</mml:mrow>
</mml:msub>
</mml:mrow>
<mml:mo>)</mml:mo>
</mml:mrow>
</mml:mrow>
</mml:math>
</inline-formula> (first row) and number <inline-formula id="inf186">
<mml:math id="m196">
<mml:mrow>
<mml:msub>
<mml:mtext>&#x393;</mml:mtext>
<mml:mn>2</mml:mn>
</mml:msub>
</mml:mrow>
</mml:math>
</inline-formula> of different extracted sparse models <inline-formula id="inf187">
<mml:math id="m197">
<mml:mrow>
<mml:mrow>
<mml:mo>(</mml:mo>
<mml:mrow>
<mml:mtext>%</mml:mtext>
<mml:msub>
<mml:mi mathvariant="normal">&#x2133;</mml:mi>
<mml:mrow>
<mml:mi>s</mml:mi>
<mml:mi>e</mml:mi>
<mml:mi>l</mml:mi>
</mml:mrow>
</mml:msub>
</mml:mrow>
<mml:mo>)</mml:mo>
</mml:mrow>
</mml:mrow>
</mml:math>
</inline-formula> (second row).</p>
</caption>
<table>
<thead valign="top">
<tr>
<th align="left">SGPLS CV</th>
<th align="center">SGPLS BootYT</th>
<th align="center">RSGPLS</th>
<th align="center">Lasso.Cl</th>
<th align="center">Lasso.MSE</th>
<th align="center">Lasso.Dev</th>
<th align="center">BootYT</th>
<th align="center">BootYTdyn</th>
</tr>
</thead>
<tbody valign="top">
<tr>
<td align="left">44 (9)</td>
<td align="char" char="(">14 (16)</td>
<td align="char" char="(">26 (40)</td>
<td align="char" char="(">26 (13)</td>
<td align="char" char="(">5 (83)</td>
<td align="char" char="(">4 (83)</td>
<td align="char" char="(">4 (35)</td>
<td align="char" char="(">2 (80)</td>
</tr>
<tr>
<td align="left">44 (9)</td>
<td align="char" char="(">16 (16)</td>
<td align="char" char="(">60 (5)</td>
<td align="char" char="(">33 (8)</td>
<td align="char" char="(">6 (44)</td>
<td align="char" char="(">5 (48)</td>
<td align="char" char="(">4 (35)</td>
<td align="char" char="(">2 (80)</td>
</tr>
</tbody>
</table>
</table-wrap>
<p>In light of these results, BootYTdyn and both the MSE- and deviance-based lasso techniques are the most stable in extracting supports and models. This could be due in part to the fact than all of them depend on only one tuning parameter. Even if this hyper-parameter for the BootYTdyn approach, i.e.,&#x20;the number of components, has to be chosen <italic>R</italic> times, the high stability of the bootstrap-based stopping criterion introduced by Magnanensi et&#x20;al. [<xref ref-type="bibr" rid="B32">32</xref>] endows this approach with good stability in selecting the sparse support. As for the PLS framework, our new bootstrap-based SGPLS implementation improves the stability here. The lack of stability of the lasso based on misclassified values is directly induced by the discrete form of this statistic. This issue was also observed and mentioned in Magnanensi et&#x20;al.&#x20;[<xref ref-type="bibr" rid="B32">32</xref>].</p>
<p>The various selected supports are displayed in <xref ref-type="fig" rid="F5">Figure&#x20;5</xref>. Note that both the MSE- and deviance-based lasso regressions select the same support and the same model, noted Lasso.MSE.Dev in the following.</p>
<fig id="F5" position="float">
<label>FIGURE 5</label>
<caption>
<p>Summary of selected supports.</p>
</caption>
<graphic xlink:href="fams-07-693126-g005.tif"/>
</fig>
<p>In the following, two additional independent public datasets, stated by Marisa et&#x20;al. [<xref ref-type="bibr" rid="B35">35</xref>] as being comparable with our original dataset, were included for comparison of predictive abilities. These datasets are named GSE18088 (<italic>n</italic>&#x20;&#x3d; 53) [<xref ref-type="bibr" rid="B39">39</xref>] and GSE14333 (<italic>n</italic>&#x20;&#x3d; 247) [<xref ref-type="bibr" rid="B40">40</xref>]. Both the MSE and number of misclassified (MC) value of the selected models were computed on both the training and test parts of the original dataset, as well as on the two additional datasets. These results are summarized in <xref ref-type="table" rid="T9">Table&#x20;9</xref>. Since independent datasets were available, we decided to follow the suggestion of Van Wieringen et&#x20;al. [<xref ref-type="bibr" rid="B41">41</xref>, p.1596]: &#x201c;the true evaluation of a predictor&#x2019;s performance is to be done on independent data&#x201d;.</p>
<table-wrap id="T9" position="float">
<label>TABLE 9</label>
<caption>
<p>Summary of model fitting and predictive abilities.</p>
</caption>
<table>
<thead valign="top">
<tr>
<th rowspan="2" align="left">GSE</th>
<th colspan="4" align="center">MC</th>
<th colspan="4" align="center">MSE</th>
</tr>
<tr>
<th align="center">39582trai</th>
<th align="center">39582test</th>
<th align="center">18088</th>
<th align="center">14333</th>
<th align="center">39582trai</th>
<th align="center">39582test</th>
<th align="center">18088</th>
<th align="center">14333</th>
</tr>
</thead>
<tbody valign="top">
<tr>
<td align="left">SGPLS CV</td>
<td align="char" char=".">48</td>
<td align="char" char=".">19</td>
<td align="char" char=".">10</td>
<td align="char" char=".">45</td>
<td align="char" char=".">0.1372</td>
<td align="char" char=".">0.1374</td>
<td align="char" char=".">0.1743</td>
<td align="char" char=".">0.1520</td>
</tr>
<tr>
<td align="left">SGPLS BootYT</td>
<td align="char" char=".">40</td>
<td align="char" char=".">15</td>
<td align="char" char=".">12</td>
<td align="char" char=".">46</td>
<td align="char" char=".">0.1295</td>
<td align="char" char=".">0.1328</td>
<td align="char" char=".">0.1848</td>
<td align="char" char=".">0.1601</td>
</tr>
<tr>
<td align="left">RSGPLS</td>
<td align="char" char=".">53</td>
<td align="char" char=".">13</td>
<td align="char" char=".">17</td>
<td align="char" char=".">45</td>
<td align="char" char=".">0.1005</td>
<td align="char" char=".">0.0912</td>
<td align="char" char=".">0.1909</td>
<td align="char" char=".">0.1276</td>
</tr>
<tr>
<td align="left">Lasso.Cl</td>
<td align="char" char=".">50</td>
<td align="char" char=".">11</td>
<td align="char" char=".">12</td>
<td align="char" char=".">44</td>
<td align="char" char=".">0.0928</td>
<td align="char" char=".">0.0827</td>
<td align="char" char=".">0.1283</td>
<td align="char" char=".">0.1425</td>
</tr>
<tr>
<td align="left">Lasso.MSE.Dev</td>
<td align="char" char=".">46</td>
<td align="char" char=".">15</td>
<td align="char" char=".">12</td>
<td align="char" char=".">47</td>
<td align="char" char=".">0.0875</td>
<td align="char" char=".">0.0828</td>
<td align="char" char=".">0.1242</td>
<td align="char" char=".">0.1406</td>
</tr>
<tr>
<td align="left">BootYT</td>
<td align="char" char=".">73</td>
<td align="char" char=".">23</td>
<td align="char" char=".">13</td>
<td align="char" char=".">111</td>
<td align="char" char=".">0.1272</td>
<td align="char" char=".">0.1256</td>
<td align="char" char=".">0.1817</td>
<td align="char" char=".">0.2924</td>
</tr>
<tr>
<td align="left">BootYTdyn</td>
<td align="char" char=".">92</td>
<td align="char" char=".">25</td>
<td align="char" char=".">13</td>
<td align="char" char=".">35</td>
<td align="char" char=".">0.1507</td>
<td align="char" char=".">0.1446</td>
<td align="char" char=".">0.1695</td>
<td align="char" char=".">0.1325</td>
</tr>
</tbody>
</table>
</table-wrap>
<p>In this real data study, BootYTdyn retains only one predictor, which is also retained by all other methods. Thus, as expected, this most sparse support induced the highest values of both the MSE and number of misclassified observations, based on the training subset of the original dataset. It also provided the highest values based on the test subset of the original dataset, which is to be expected since, as explained in <xref ref-type="sec" rid="s3-3-1">Section 3.3.1</xref>, these two parts are well-balanced in terms of anatomo-clinical characteristics. Thus, this causes a bias in evaluation of model&#x2019;s predictive abilities, by making comparable MSE and misclassified values based on both the training and testing subsets. Results in <xref ref-type="table" rid="T9">Table&#x20;9</xref> confirm this property, and also highlight the usefulness of additional independent datasets for reliably comparing predictive abilities. While a well-designed model for predictive purposes will provide similar PMSE values on comparable independent additional datasets compared to MSE obtained on the training dataset, an over-fitted one would be related to higher PMSE values, due to its dependance on training data. Thus, the differences, noted <inline-formula id="inf188">
<mml:math id="m198">
<mml:mtext>&#x394;</mml:mtext>
</mml:math>
</inline-formula> MSE, between the MSE obtained on the training subset of the original dataset, and those obtained on the three test datasets, are shown in <xref ref-type="fig" rid="F6">Figure&#x20;6</xref>.</p>
<fig id="F6" position="float">
<label>FIGURE 6</label>
<caption>
<p>Differences between the MSE computed on the training subset of the original dataset and the PMSE obtained on test&#x20;sets.</p>
</caption>
<graphic xlink:href="fams-07-693126-g006.tif"/>
</fig>
<p>Our new dynamic bootstrap-based approach is the only one that exhibits this MSE stability property. In contrast, all other approaches provided higher PMSE values on the two other datasets than on the original. This led BootYTdyn to have only 48 (16%) misclassified values on these two additional datasets, representing the best result among all studied approaches. Thus, we can reasonably assume that our new method has helped to remove non-informative predictors, in the sense of not being relevant for improving predictive ability.</p>
<p>Lastly, the unique extracted probe set, named 230784_at, is already known to be related to the original location in the distal colon. The sign of the regression coefficient obtained with BootYTdyn is coherent with this state-of-the-art result, and thus strengthens our conclusion.</p>
</sec>
<sec sec-type="discussion" id="s5">
<title>5 Discussion</title>
<p>In this article, we developed a new dynamic bootstrap-based technique for variable selection, suitable for both the PLS and GPLS frameworks, and proposed a bootstrap-based version of the SPLS and SGPLS methods for selecting the number of components. While the first of these lets us completely avoid CV, the second lets us select the set of tuning parameters in a more reliable&#x20;way.</p>
<p>In state-of-the-art approaches, the use of CV-based techniques for selection of hyper-parameters is common, and can lead to important stability issues, observed notably by Boulesteix [<xref ref-type="bibr" rid="B31">31</xref>] and Magnanensi et&#x20;al. [<xref ref-type="bibr" rid="B32">32</xref>], and confirmed in our studies. Our new dynamic bootstrap-based technique represents a useful method for avoiding CV-related issues, since the unique hyper-parameter is successively selected using a bootstrap-based criterion. This new technique improves on the original bootstrap-based methodology introduced by Lazraq et&#x20;al. [<xref ref-type="bibr" rid="B20">20</xref>], in that it permits us to approximate the distribution of covariates&#x2019; regression coefficients by removing the condition of working in a subspace of fixed dimension <italic>K</italic>. Theoretical results have been established that strengthen the usefulness of building subspaces spanned by a dynamic number of components for performing PLS regressions on bootstrap samples.</p>
<p>In the PLS framework, conclusions based on our simulations are twofold. First, for datasets with negligible random noise in <inline-formula id="inf189">
<mml:math id="m199">
<mml:mtext>y</mml:mtext>
</mml:math>
</inline-formula>, SPLS BootYT is recommended. Indeed, both in terms of accuracy and predictive abilities, it outperforms all other techniques compared here. Second, for datasets with non-negligible random noise in <bold>y</bold>, which represents the more realistic case, our new dynamic bootstrap-based method is recommended. As for SPLS BootYT for the negligible random noise variability framework, BootYTdyn outperform all other methods for each property studied. Furthermore, it is the only method which concludes as to a decreasing number of significant predictors, while the random noise variability increases, which was expected&#x20;here.</p>
<p>Results obtained from our classification study using real datasets match with previous conclusions. Indeed, BootYTdyn is the only one which leads to the expected PMSE values on two additional independent datasets, which allows us to suppose that over-fitting issues were avoided, and that these PMSE are induced by noise or information which cannot be modeled using only gene expression. Furthermore, the extracted probe set is already known to be linked to the relevant location in the distal colon, which strengthens our confidence in this new dynamic approach.</p>
<p>Lastly, our new bootstrap-based SPLS implementation improves the stability of this method. Indeed, in all cases studied, both the SPLS CV and SGPLS CV select a higher number of unique sets of hyper-parameters than our bootstrap-based versions do, leading to higher numbers of unique supports and models&#x20;too.</p>
<p>In the future, our work may be extended in the following three directions. Firstly, Sun et&#x20;al. [<xref ref-type="bibr" rid="B42">42</xref>] also worked on this subject, and proposed a methodology for selecting tuning parameters of penalized regressions in order to stabilize variable selection. Even if their method is not applicable to the selection of the number of components in PLS regression, it would be interesting to adapt it to both SPLS CV and our new SPLS BootYT implementation, in order to look for potential stability&#x20;gains.</p>
<p>Secondly combining partial least squares based regular or sparse techniques with the existing techniques that were developed in [<xref ref-type="bibr" rid="B24">24</xref>&#x2013;<xref ref-type="bibr" rid="B28">28</xref>] may speed up, specialize toward a specific target or improve the accuracy of those existing algorithms by taking into account some latent structure of the&#x20;data.</p>
<p>Thirdly, simulations may be performed to strengthen the results obtained on real data sets for the logistic framework (<xref ref-type="sec" rid="s4">Section 4</xref>). In addition, testing the performance of these new approaches for responses following other distributions -such as Poisson, Gamma-should also be done. However, based on all the results obtained in this paper, our new dynamic method seems to be the most efficient compared to the state of the art approaches for datasets with non-negligible noise variability, a common situation in daily practice.</p>
</sec>
</body>
<back>
<sec id="s6">
<title>Data Availability Statement</title>
<p>The datasets analyzed for this study can be found in the Datasets_benchmark Github repository, <ext-link ext-link-type="uri" xlink:href="https://github.com/fbertran/Datasets_benchmark">https://github.com/fbertran/Datasets_benchmark</ext-link>. The packages used for the simulations in this article are available on the CRAN, <ext-link ext-link-type="uri" xlink:href="https://cran.r-project.org">https://cran.r-project.org</ext-link> and especially the plsRglm package that implements partial least squares generalized regression, <ext-link ext-link-type="uri" xlink:href="https://CRAN.R-project.org/package=plsRglm">https://CRAN.R-project.org/package&#x3d;plsRglm</ext-link> and the bootPLS package that implements number of components selection using bootstrap for PLS regression and generalized regression models, <ext-link ext-link-type="uri" xlink:href="https://github.com/fbertran/bootPLS">https://github.com/fbertran/bootPLS</ext-link>.</p>
</sec>
<sec id="s7">
<title>Author Contributions</title>
<p>JM, FB, MM-B and NM designed the study. JM did the simulations. JM, FB and MM-B wrote the manuscript.</p>
</sec>
<sec id="s8">
<title>Funding</title>
<p>This work was supported by grants from the Agence Nationale de la Recherche (ANR) (ANR-11-LABX-0055_IRMIA); the CNRS (UMR 7501) LabEx IRMIA to FB, NM and MM-B; Mast&#x00E9;re Sp&#x00E9;cialis&#x00E9;&#x00AE; Expert Big Data Engineer of the University of Technology of Troyes to FB and MM.</p>
</sec>
<sec sec-type="COI-statement" id="s9">
<title>Conflict of Interest</title>
<p>The authors declare that the research was conducted in the absence of any commercial or financial relationships that could be construed as a potential conflict of interest.</p>
</sec>
<sec id="s10">
<title>Supplementary Material</title>
<p>The Supplementary Material for this article can be found online at: <ext-link ext-link-type="uri" xlink:href="https://www.frontiersin.org/articles/10.3389/fams.2021.693126/full#supplementary-material">https://www.frontiersin.org/articles/10.3389/fams.2021.693126/full&#x23;supplementary-material</ext-link>
</p>
<supplementary-material xlink:href="DataSheet1.PDF" id="SM1" mimetype="application/PDF" xmlns:xlink="http://www.w3.org/1999/xlink"/>
</sec>
<ref-list>
<title>References</title>
<ref id="B1">
<label>1.</label>
<citation citation-type="book">
<person-group person-group-type="author">
<name>
<surname>Wold</surname>
<given-names>S</given-names>
</name>
<name>
<surname>Martens</surname>
<given-names>H</given-names>
</name>
<name>
<surname>Wold</surname>
<given-names>H</given-names>
</name>
</person-group>. <article-title>The Multivariate Calibration Problem in Chemistry Solved by the PLS Method</article-title>. In: <source>Matrix Pencils</source>. <publisher-loc>Heidelberg, Deutschland</publisher-loc>: <publisher-name>Springer-Verlag</publisher-name> (<year>1983</year>) p. <fpage>286</fpage>&#x2013;<lpage>93</lpage>. <pub-id pub-id-type="doi">10.1007/bfb0062108</pub-id> </citation>
</ref>
<ref id="B2">
<label>2.</label>
<citation citation-type="journal">
<person-group person-group-type="author">
<name>
<surname>Wold</surname>
<given-names>S</given-names>
</name>
<name>
<surname>Sj&#xf6;str&#xf6;m</surname>
<given-names>M</given-names>
</name>
<name>
<surname>Eriksson</surname>
<given-names>L</given-names>
</name>
</person-group>. <article-title>PLS-regression: a Basic Tool of Chemometrics</article-title>. <source>Chemometrics Intell Lab Syst</source> (<year>2001</year>) <volume>58</volume>:<fpage>109</fpage>&#x2013;<lpage>30</lpage>. <pub-id pub-id-type="doi">10.1016/s0169-7439(01)00155-1</pub-id> </citation>
</ref>
<ref id="B3">
<label>3.</label>
<citation citation-type="journal">
<person-group person-group-type="author">
<name>
<surname>Boulesteix</surname>
<given-names>AL</given-names>
</name>
<name>
<surname>Strimmer</surname>
<given-names>K</given-names>
</name>
</person-group>. <article-title>Partial Least Squares: a Versatile Tool for the Analysis of High-Dimensional Genomic Data</article-title>. <source>Brief Bioinform</source> (<year>2007</year>) <volume>8</volume>:<fpage>32</fpage>&#x2013;<lpage>44</lpage>. <pub-id pub-id-type="doi">10.1093/bib/bbl016</pub-id> </citation>
</ref>
<ref id="B4">
<label>4.</label>
<citation citation-type="journal">
<person-group person-group-type="author">
<name>
<surname>H&#xf6;skuldsson</surname>
<given-names>A</given-names>
</name>
</person-group>. <article-title>PLS Regression Methods</article-title>. <source>J&#x20;Chemometrics</source> (<year>1988</year>) <volume>2</volume>:<fpage>211</fpage>&#x2013;<lpage>28</lpage>. <pub-id pub-id-type="doi">10.1002/cem.1180020306</pub-id> </citation>
</ref>
<ref id="B5">
<label>5.</label>
<citation citation-type="book">
<person-group person-group-type="author">
<name>
<surname>Tenenhaus</surname>
<given-names>M</given-names>
</name>
</person-group>. <source>La r&#xe9;gression PLS, Th&#xe9;orie et pratique</source>. <publisher-loc>Paris, France</publisher-loc>: <publisher-name>Editions Technip</publisher-name> (<year>1998</year>).</citation>
</ref>
<ref id="B6">
<label>6.</label>
<citation citation-type="journal">
<person-group person-group-type="author">
<name>
<surname>Wold</surname>
<given-names>S</given-names>
</name>
<name>
<surname>Ruhe</surname>
<given-names>A</given-names>
</name>
<name>
<surname>Wold</surname>
<given-names>H</given-names>
</name>
<name>
<surname>Dunn, III</surname>
<given-names>WJ</given-names>
<suffix>III</suffix>
</name>
</person-group>. <article-title>The Collinearity Problem in Linear Regression. The Partial Least Squares (PLS) Approach to Generalized Inverses</article-title>. <source>SIAM J&#x20;Sci Stat Comput</source> (<year>1984</year>) <volume>5</volume>:<fpage>735</fpage>&#x2013;<lpage>43</lpage>. <pub-id pub-id-type="doi">10.1137/0905052</pub-id> </citation>
</ref>
<ref id="B7">
<label>7.</label>
<citation citation-type="journal">
<person-group person-group-type="author">
<name>
<surname>Nguyen</surname>
<given-names>DV</given-names>
</name>
<name>
<surname>Rocke</surname>
<given-names>DM</given-names>
</name>
</person-group>. <article-title>Multi-class Cancer Classification via Partial Least Squares with Gene Expression Profiles</article-title>. <source>Bioinformatics</source> (<year>2002</year>) <volume>18</volume>:<fpage>1216</fpage>&#x2013;<lpage>26</lpage>. <pub-id pub-id-type="doi">10.1093/bioinformatics/18.9.1216</pub-id> </citation>
</ref>
<ref id="B8">
<label>8.</label>
<citation citation-type="journal">
<person-group person-group-type="author">
<name>
<surname>Nguyen</surname>
<given-names>DV</given-names>
</name>
<name>
<surname>Rocke</surname>
<given-names>DM</given-names>
</name>
</person-group>. <article-title>Tumor Classification by Partial Least Squares Using Microarray Gene Expression Data</article-title>. <source>Bioinformatics</source> (<year>2002</year>) <volume>18</volume>:<fpage>39</fpage>&#x2013;<lpage>50</lpage>. <pub-id pub-id-type="doi">10.1093/bioinformatics/18.1.39</pub-id> </citation>
</ref>
<ref id="B9">
<label>9.</label>
<citation citation-type="journal">
<person-group person-group-type="author">
<name>
<surname>Boulesteix</surname>
<given-names>A-L</given-names>
</name>
</person-group>. <article-title>PLS Dimension Reduction for Classification with Microarray Data</article-title>. <source>Stat Appl Genet Mol Biol</source> (<year>2004</year>) <volume>3</volume>:<fpage>1</fpage>&#x2013;<lpage>30</lpage>. <pub-id pub-id-type="doi">10.2202/1544-6115.1075</pub-id> </citation>
</ref>
<ref id="B10">
<label>10.</label>
<citation citation-type="journal">
<person-group person-group-type="author">
<name>
<surname>Marx</surname>
<given-names>BD</given-names>
</name>
</person-group>. <article-title>Iteratively Reweighted Partial Least Squares Estimation for Generalized Linear Regression</article-title>. <source>Technometrics</source> (<year>1996</year>) <volume>38</volume>:<fpage>374</fpage>&#x2013;<lpage>81</lpage>. <pub-id pub-id-type="doi">10.1080/00401706.1996.10484549</pub-id> </citation>
</ref>
<ref id="B11">
<label>11.</label>
<citation citation-type="journal">
<person-group person-group-type="author">
<name>
<surname>Cessie</surname>
<given-names>SL</given-names>
</name>
<name>
<surname>Houwelingen</surname>
<given-names>JCV</given-names>
</name>
</person-group>. <article-title>Ridge Estimators in Logistic Regression</article-title>. <source>Appl Stat</source> (<year>1992</year>) <volume>41</volume>:<fpage>191</fpage>&#x2013;<lpage>201</lpage>. <pub-id pub-id-type="doi">10.2307/2347628</pub-id> </citation>
</ref>
<ref id="B12">
<label>12.</label>
<citation citation-type="journal">
<person-group person-group-type="author">
<name>
<surname>Firth</surname>
<given-names>D</given-names>
</name>
</person-group>. <article-title>Bias Reduction of Maximum Likelihood Estimates</article-title>. <source>Biometrika</source> (<year>1993</year>) <volume>80</volume>:<fpage>27</fpage>&#x2013;<lpage>38</lpage>. <pub-id pub-id-type="doi">10.1093/biomet/80.1.27</pub-id> </citation>
</ref>
<ref id="B13">
<label>13.</label>
<citation citation-type="journal">
<person-group person-group-type="author">
<name>
<surname>Nguyen</surname>
<given-names>DV</given-names>
</name>
<name>
<surname>Rocke</surname>
<given-names>DM</given-names>
</name>
</person-group>. <article-title>On Partial Least Squares Dimension Reduction for Microarray-Based Classification: a Simulation Study</article-title>. <source>Comput Stat Data Anal</source> (<year>2004</year>) <volume>46</volume>:<fpage>407</fpage>&#x2013;<lpage>25</lpage>. <pub-id pub-id-type="doi">10.1016/j.csda.2003.08.001</pub-id> </citation>
</ref>
<ref id="B14">
<label>14.</label>
<citation citation-type="journal">
<person-group person-group-type="author">
<name>
<surname>Ding</surname>
<given-names>B</given-names>
</name>
<name>
<surname>Gentleman</surname>
<given-names>R</given-names>
</name>
</person-group>. <article-title>Classification Using Generalized Partial Least Squares</article-title>. <source>J&#x20;Comput Graphical Stat</source> (<year>2005</year>) <volume>14</volume>:<fpage>280</fpage>&#x2013;<lpage>98</lpage>. <pub-id pub-id-type="doi">10.1198/106186005x47697</pub-id> </citation>
</ref>
<ref id="B15">
<label>15.</label>
<citation citation-type="journal">
<person-group person-group-type="author">
<name>
<surname>Fort</surname>
<given-names>G</given-names>
</name>
<name>
<surname>Lambert-Lacroix</surname>
<given-names>S</given-names>
</name>
</person-group>. <article-title>Classification Using Partial Least Squares with Penalized Logistic Regression</article-title>. <source>Bioinformatics</source> (<year>2005</year>) <volume>21</volume>:<fpage>1104</fpage>&#x2013;<lpage>11</lpage>. <pub-id pub-id-type="doi">10.1093/bioinformatics/bti114</pub-id> </citation>
</ref>
<ref id="B16">
<label>16.</label>
<citation citation-type="journal">
<person-group person-group-type="author">
<name>
<surname>Bastien</surname>
<given-names>P</given-names>
</name>
<name>
<surname>Vinzi</surname>
<given-names>VE</given-names>
</name>
<name>
<surname>Tenenhaus</surname>
<given-names>M</given-names>
</name>
</person-group>. <article-title>PLS Generalised Linear Regression</article-title>. <source>Comput Stat Data Anal</source> (<year>2005</year>) <volume>48</volume>:<fpage>17</fpage>&#x2013;<lpage>46</lpage>. <pub-id pub-id-type="doi">10.1016/j.csda.2004.02.005</pub-id> </citation>
</ref>
<ref id="B17">
<label>17.</label>
<citation citation-type="journal">
<person-group person-group-type="author">
<name>
<surname>Chun</surname>
<given-names>H</given-names>
</name>
<name>
<surname>Kele&#xc5;&#x178;</surname>
<given-names>S</given-names>
</name>
</person-group>. <article-title>Sparse Partial Least Squares Regression for Simultaneous Dimension Reduction and Variable Selection</article-title>. <source>J&#x20;R Stat Soc Ser B (Statistical Methodology)</source> (<year>2010</year>) <volume>72</volume>:<fpage>3</fpage>&#x2013;<lpage>25</lpage>. <pub-id pub-id-type="doi">10.1111/j.1467-9868.2009.00723.x</pub-id> </citation>
</ref>
<ref id="B18">
<label>18.</label>
<citation citation-type="journal">
<person-group person-group-type="author">
<name>
<surname>Smyth</surname>
<given-names>GK</given-names>
</name>
</person-group>. <article-title>Linear Models and Empirical Bayes Methods for Assessing Differential Expression in Microarray Experiments</article-title>. <source>Stat Appl Genet Mol Biol</source> (<year>2004</year>) <volume>3</volume>:<fpage>Article3</fpage>. <pub-id pub-id-type="doi">10.2202/1544-6115.1027</pub-id> </citation>
</ref>
<ref id="B19">
<label>19.</label>
<citation citation-type="journal">
<person-group person-group-type="author">
<name>
<surname>Mehmood</surname>
<given-names>T</given-names>
</name>
<name>
<surname>Liland</surname>
<given-names>KH</given-names>
</name>
<name>
<surname>Snipen</surname>
<given-names>L</given-names>
</name>
<name>
<surname>S&#xe6;b&#xf8;</surname>
<given-names>S</given-names>
</name>
</person-group>. <article-title>A Review of Variable Selection Methods in Partial Least Squares Regression</article-title>. <source>Chemometrics Intell Lab Syst</source> (<year>2012</year>) <volume>118</volume>:<fpage>62</fpage>&#x2013;<lpage>9</lpage>. <pub-id pub-id-type="doi">10.1016/j.chemolab.2012.07.010</pub-id> </citation>
</ref>
<ref id="B20">
<label>20.</label>
<citation citation-type="journal">
<person-group person-group-type="author">
<name>
<surname>Lazraq</surname>
<given-names>A</given-names>
</name>
<name>
<surname>Cl&#xe9;roux</surname>
<given-names>R</given-names>
</name>
<name>
<surname>Gauchi</surname>
<given-names>J-P</given-names>
</name>
</person-group>. <article-title>Selecting Both Latent and Explanatory Variables in the PLS1 Regression Model</article-title>. <source>Chemometrics Intell Lab Syst</source> (<year>2003</year>) <volume>66</volume>:<fpage>117</fpage>&#x2013;<lpage>26</lpage>. <pub-id pub-id-type="doi">10.1016/s0169-7439(03)00027-3</pub-id> </citation>
</ref>
<ref id="B21">
<label>21.</label>
<citation citation-type="journal">
<person-group person-group-type="author">
<name>
<surname>Chung</surname>
<given-names>D</given-names>
</name>
<name>
<surname>Keles</surname>
<given-names>S</given-names>
</name>
</person-group>. <article-title>Sparse Partial Least Squares Classification for High Dimensional Data</article-title>. <source>Stat Appl Genet Mol Biol</source> (<year>2010</year>) <volume>9</volume>. <fpage>Article 17</fpage>. <pub-id pub-id-type="doi">10.2202/1544-6115.1492</pub-id> </citation>
</ref>
<ref id="B22">
<label>22.</label>
<citation citation-type="journal">
<person-group person-group-type="author">
<name>
<surname>Bastien</surname>
<given-names>P</given-names>
</name>
<name>
<surname>Bertrand</surname>
<given-names>F</given-names>
</name>
<name>
<surname>Meyer</surname>
<given-names>N</given-names>
</name>
<name>
<surname>Maumy-Bertrand</surname>
<given-names>M</given-names>
</name>
</person-group>. <article-title>Deviance Residuals-Based Sparse PLS and Sparse Kernel PLS Regression for Censored Data</article-title>. <source>Bioinformatics</source> (<year>2015</year>) <volume>31</volume>:<fpage>397</fpage>&#x2013;<lpage>404</lpage>. <pub-id pub-id-type="doi">10.1093/bioinformatics/btu660</pub-id> </citation>
</ref>
<ref id="B23">
<label>23.</label>
<citation citation-type="journal">
<person-group person-group-type="author">
<name>
<surname>Meyer</surname>
<given-names>N</given-names>
</name>
<name>
<surname>Maumy-Bertrand</surname>
<given-names>M</given-names>
</name>
<name>
<surname>Bertrand</surname>
<given-names>F</given-names>
</name>
</person-group>. <article-title>Comparaison de variantes de r&#xe9;gressions logistiques PLS et de r&#xe9;gression PLS sur variables qualitatives: application aux donn&#xe9;es d&#x2019;all&#xe9;lotypage</article-title>. <source>J&#x20;de la Soci&#xe9;t&#xe9; Fran&#xe7;aise de Statistique</source> (<year>2010</year>) <volume>151</volume>:<fpage>1</fpage>&#x2013;<lpage>18</lpage>. </citation>
</ref>
<ref id="B24">
<label>24.</label>
<citation citation-type="journal">
<person-group person-group-type="author">
<name>
<surname>Gupta</surname>
<given-names>D</given-names>
</name>
<name>
<surname>Richhariya</surname>
<given-names>B</given-names>
</name>
</person-group>. <article-title>Entropy Based Fuzzy Least Squares Twin Support Vector Machine for Class Imbalance Learning</article-title>. <source>Appl Intell</source> (<year>2018</year>) <volume>48</volume>:<fpage>4212</fpage>&#x2013;<lpage>31</lpage>. <pub-id pub-id-type="doi">10.1007/s9489-018-1204-4</pub-id> </citation>
</ref>
<ref id="B25">
<label>25.</label>
<citation citation-type="confproc">
<person-group person-group-type="author">
<name>
<surname>Gupta</surname>
<given-names>U</given-names>
</name>
<name>
<surname>Gupta</surname>
<given-names>D</given-names>
</name>
<name>
<surname>Prasad</surname>
<given-names>M</given-names>
</name>
</person-group>. <article-title>Kernel Target Alignment Based Fuzzy Least Square Twin Bounded Support Vector Machine</article-title>. In: <conf-name>Proceedings of the 2018 IEEE Symposium Series on Computational Intelligence</conf-name>; <conf-date>18-21 Nov. 2018</conf-date>; <conf-loc>Bangalore, India</conf-loc>
<publisher-name>SSCI</publisher-name> (<year>2018</year>) p. <fpage>228</fpage>&#x2013;<lpage>35</lpage>. <pub-id pub-id-type="doi">10.1109/SSCI.2018.8628903</pub-id> </citation>
</ref>
<ref id="B26">
<label>26.</label>
<citation citation-type="book">
<person-group person-group-type="author">
<name>
<surname>Borah</surname>
<given-names>P</given-names>
</name>
<name>
<surname>Gupta</surname>
<given-names>D</given-names>
</name>
</person-group>. <article-title>A Two-Norm Squared Fuzzy-Based Least Squares Twin Parametric-Margin Support Vector Machine</article-title>. In: <person-group person-group-type="editor">
<name>
<surname>Tanveer</surname>
<given-names>M</given-names>
</name>
<name>
<surname>Pachori</surname>
<given-names>RB</given-names>
</name>
</person-group>, editors. <source>
<italic>Machine Intelligence and Signal Analysis</italic> Advances in Intelligent Systems and Computing</source>, <volume>748</volume>. <publisher-loc>Singapore</publisher-loc>: <publisher-name>Springer Singapore</publisher-name> (<year>2019</year>) p. <fpage>119</fpage>&#x2013;<lpage>34</lpage>. <pub-id pub-id-type="doi">10.1007/978-981-13-0923-610.1007/978-981-13-0923-6_11</pub-id> </citation>
</ref>
<ref id="B27">
<label>27.</label>
<citation citation-type="confproc">
<person-group person-group-type="author">
<name>
<surname>Borah</surname>
<given-names>P</given-names>
</name>
<name>
<surname>Gupta</surname>
<given-names>D</given-names>
</name>
<name>
<surname>Prasad</surname>
<given-names>M</given-names>
</name>
</person-group>. <article-title>Improved 2-norm Based Fuzzy Least Squares Twin Support Vector Machine</article-title>. In: <conf-name>2018 IEEE Symposium Series on Computational Intelligence (SSCI)</conf-name>; <conf-date>18-21 Nov. 2018</conf-date>; <conf-loc>Bangalore, India</conf-loc>
<publisher-name>IEEE</publisher-name> (<year>2018</year>) p. <fpage>412</fpage>&#x2013;<lpage>9</lpage>. <pub-id pub-id-type="doi">10.1109/SSCI.2018.8628818</pub-id> </citation>
</ref>
<ref id="B28">
<label>28.</label>
<citation citation-type="journal">
<person-group person-group-type="author">
<name>
<surname>Gupta</surname>
<given-names>U</given-names>
</name>
<name>
<surname>Gupta</surname>
<given-names>D</given-names>
</name>
</person-group>. <article-title>Least Squares Large Margin Distribution Machine for Regression</article-title>. <source>Appl Intell</source> (<year>2021</year>). <pub-id pub-id-type="doi">10.1007/s10489-020-02166-5</pub-id> </citation>
</ref>
<ref id="B29">
<label>29.</label>
<citation citation-type="journal">
<person-group person-group-type="author">
<name>
<surname>Wiklund</surname>
<given-names>S</given-names>
</name>
<name>
<surname>Nilsson</surname>
<given-names>D</given-names>
</name>
<name>
<surname>Eriksson</surname>
<given-names>L</given-names>
</name>
<name>
<surname>Sj&#xf6;str&#xf6;m</surname>
<given-names>M</given-names>
</name>
<name>
<surname>Wold</surname>
<given-names>S</given-names>
</name>
<name>
<surname>Faber</surname>
<given-names>K</given-names>
</name>
</person-group>. <article-title>A Randomization Test for PLS Component Selection</article-title>. <source>J&#x20;Chemometrics</source> (<year>2007</year>) <volume>21</volume>:<fpage>427</fpage>&#x2013;<lpage>39</lpage>. <pub-id pub-id-type="doi">10.1002/cem.1086</pub-id> </citation>
</ref>
<ref id="B30">
<label>30.</label>
<citation citation-type="book">
<person-group person-group-type="author">
<name>
<surname>Hastie</surname>
<given-names>T</given-names>
</name>
<name>
<surname>Tibshirani</surname>
<given-names>R</given-names>
</name>
<name>
<surname>Friedman</surname>
<given-names>JJH</given-names>
</name>
</person-group>. <source>The Elements of Statistical Learning</source>. <edition>2nd ed.</edition>, <volume>Vol. 1</volume>. <publisher-loc>New York, USANew York</publisher-loc>: <publisher-name>Springer</publisher-name> (<year>2009</year>).</citation>
</ref>
<ref id="B31">
<label>31.</label>
<citation citation-type="book">
<person-group person-group-type="author">
<name>
<surname>Boulesteix</surname>
<given-names>AL</given-names>
</name>
</person-group>. <source>Accuracy Estimation for PLS and Related Methods via Resampling-Based Procedures</source>. <publisher-loc>Paris</publisher-loc>: <publisher-name>PLS&#x2019;14 Book of Abstracts</publisher-name> (<year>2014</year>). p. <fpage>13</fpage>&#x2013;<lpage>4</lpage>.</citation>
</ref>
<ref id="B32">
<label>32.</label>
<citation citation-type="journal">
<person-group person-group-type="author">
<name>
<surname>Magnanensi</surname>
<given-names>J</given-names>
</name>
<name>
<surname>Bertrand</surname>
<given-names>F</given-names>
</name>
<name>
<surname>Maumy-Bertrand</surname>
<given-names>M</given-names>
</name>
<name>
<surname>Meyer</surname>
<given-names>N</given-names>
</name>
</person-group>. <article-title>A New Universal Resample-Stable Bootstrap-Based Stopping Criterion for PLS Component Construction</article-title>. <source>Stat Comput</source> (<year>2017</year>) <volume>27</volume>:<fpage>1</fpage>&#x2013;<lpage>18</lpage>. <pub-id pub-id-type="doi">10.1007/s11222-016-9651-4</pub-id> </citation>
</ref>
<ref id="B33">
<label>33.</label>
<citation citation-type="book">
<person-group person-group-type="author">
<name>
<surname>Efron</surname>
<given-names>B</given-names>
</name>
<name>
<surname>Tibshirani</surname>
<given-names>RJ</given-names>
</name>
</person-group>. <source>An Introduction to the Bootstrap</source>, <volume>Vol. 57</volume>. <publisher-loc>Boca Raton, USA</publisher-loc>: <publisher-name>Chapman &#x26; Hall/CRC</publisher-name> (<year>1993</year>).</citation>
</ref>
<ref id="B34">
<label>34.</label>
<citation citation-type="confproc">
<person-group person-group-type="author">
<name>
<surname>Kohavi</surname>
<given-names>R</given-names>
</name>
</person-group>. <article-title>A Study of Cross-Validation and Bootstrap for Accuracy Estimation and Model Selection</article-title>. In: <conf-name>Proceedings of the 14th international joint conference on Artificial intelligence-Volume 2</conf-name>, <conf-loc>Montreal, Quebec, Canada</conf-loc>, <conf-date>August 20&#x2013;25, 1995</conf-date>. <publisher-loc>USA</publisher-loc>: <publisher-name>Morgan Kaufmann Publishers Inc.)</publisher-name> (<year>1995</year>). p. <fpage>1137</fpage>&#x2013;<lpage>43</lpage>. </citation>
</ref>
<ref id="B35">
<label>35.</label>
<citation citation-type="journal">
<person-group person-group-type="author">
<name>
<surname>Marisa</surname>
<given-names>L</given-names>
</name>
<name>
<surname>de Reyni&#xe8;s</surname>
<given-names>A</given-names>
</name>
<name>
<surname>Duval</surname>
<given-names>A</given-names>
</name>
<name>
<surname>Selves</surname>
<given-names>J</given-names>
</name>
<name>
<surname>Gaub</surname>
<given-names>MP</given-names>
</name>
<name>
<surname>Vescovo</surname>
<given-names>L</given-names>
</name>
<etal/>
</person-group> <article-title>Gene Expression Classification of colon Cancer into Molecular Subtypes: Characterization, Validation, and Prognostic Value</article-title>. <source>Plos Med</source> (<year>2013</year>) <volume>10</volume>:<fpage>e1001453</fpage>. <pub-id pub-id-type="doi">10.1371/journal.pmed.1001453</pub-id> </citation>
</ref>
<ref id="B36">
<label>36.</label>
<citation citation-type="journal">
<person-group person-group-type="author">
<name>
<surname>Kr&#xe4;mer</surname>
<given-names>N</given-names>
</name>
<name>
<surname>Sugiyama</surname>
<given-names>M</given-names>
</name>
</person-group>. <article-title>The Degrees of freedom of Partial Least Squares Regression</article-title>. <source>J&#x20;Am Stat Assoc</source> (<year>2011</year>) <volume>106</volume>:<fpage>697</fpage>&#x2013;<lpage>705</lpage>. <pub-id pub-id-type="doi">10.1198/jasa.2011.tm10107</pub-id> </citation>
</ref>
<ref id="B37">
<label>37.</label>
<citation citation-type="journal">
<person-group person-group-type="author">
<name>
<surname>Efron</surname>
<given-names>B</given-names>
</name>
<name>
<surname>Hastie</surname>
<given-names>T</given-names>
</name>
<name>
<surname>Johnstone</surname>
<given-names>I</given-names>
</name>
<name>
<surname>Tibshirani</surname>
<given-names>R</given-names>
</name>
<etal/>
</person-group> <article-title>Least Angle Regression</article-title>. <source>Ann Stat</source> (<year>2004</year>) <volume>32</volume>:<fpage>407</fpage>&#x2013;<lpage>99</lpage>. <pub-id pub-id-type="doi">10.1214/009053604000000067</pub-id> </citation>
</ref>
<ref id="B38">
<label>38.</label>
<citation citation-type="journal">
<person-group person-group-type="author">
<name>
<surname>Durif</surname>
<given-names>G</given-names>
</name>
<name>
<surname>Modolo</surname>
<given-names>L</given-names>
</name>
<name>
<surname>Michaelsson</surname>
<given-names>J</given-names>
</name>
<name>
<surname>Mold</surname>
<given-names>JE</given-names>
</name>
<name>
<surname>Lambert-Lacroix</surname>
<given-names>S</given-names>
</name>
<name>
<surname>Picard</surname>
<given-names>F</given-names>
</name>
</person-group>. <article-title>High Dimensional Classification with Combined Adaptive Sparse Pls and Logistic Regression</article-title>. <source>Bioinformatics</source> (<year>2017</year>) <volume>34</volume>:<fpage>485</fpage>&#x2013;<lpage>93</lpage>. <pub-id pub-id-type="doi">10.1093/bioinformatics/btx571</pub-id> </citation>
</ref>
<ref id="B39">
<label>39.</label>
<citation citation-type="journal">
<person-group person-group-type="author">
<name>
<surname>Gr&#xf6;ne</surname>
<given-names>J</given-names>
</name>
<name>
<surname>Lenze</surname>
<given-names>D</given-names>
</name>
<name>
<surname>Jurinovic</surname>
<given-names>V</given-names>
</name>
<name>
<surname>Hummel</surname>
<given-names>M</given-names>
</name>
<name>
<surname>Seidel</surname>
<given-names>H</given-names>
</name>
<name>
<surname>Leder</surname>
<given-names>G</given-names>
</name>
<etal/>
</person-group> <article-title>Molecular Profiles and Clinical Outcome of Stage UICC II colon Cancer Patients</article-title>. <source>Int J&#x20;Colorectal Dis</source> (<year>2011</year>) <volume>26</volume>:<fpage>847</fpage>&#x2013;<lpage>58</lpage>. <pub-id pub-id-type="doi">10.1007/s00384-011-1176-x</pub-id> </citation>
</ref>
<ref id="B40">
<label>40.</label>
<citation citation-type="journal">
<person-group person-group-type="author">
<name>
<surname>Jorissen</surname>
<given-names>RN</given-names>
</name>
<name>
<surname>Gibbs</surname>
<given-names>P</given-names>
</name>
<name>
<surname>Christie</surname>
<given-names>M</given-names>
</name>
<name>
<surname>Prakash</surname>
<given-names>S</given-names>
</name>
<name>
<surname>Lipton</surname>
<given-names>L</given-names>
</name>
<name>
<surname>Desai</surname>
<given-names>J</given-names>
</name>
<etal/>
</person-group> <article-title>Metastasis-associated Gene Expression Changes Predict Poor Outcomes in Patients with Dukes Stage B and C Colorectal Cancer</article-title>. <source>Clin Cancer Res</source> (<year>2009</year>) <volume>15</volume>:<fpage>7642</fpage>&#x2013;<lpage>51</lpage>. <pub-id pub-id-type="doi">10.1158/1078-0432.ccr-09-1431</pub-id> </citation>
</ref>
<ref id="B41">
<label>41.</label>
<citation citation-type="journal">
<person-group person-group-type="author">
<name>
<surname>Van Wieringen</surname>
<given-names>WN</given-names>
</name>
<name>
<surname>Kun</surname>
<given-names>D</given-names>
</name>
<name>
<surname>Hampel</surname>
<given-names>R</given-names>
</name>
<name>
<surname>Boulesteix</surname>
<given-names>A-L</given-names>
</name>
</person-group>. <article-title>Survival Prediction Using Gene Expression Data: a Review and Comparison</article-title>. <source>Comput Stat Data Anal</source> (<year>2009</year>) <volume>53</volume>:<fpage>1590</fpage>&#x2013;<lpage>603</lpage>. <pub-id pub-id-type="doi">10.1016/j.csda.2008.05.021</pub-id> </citation>
</ref>
<ref id="B42">
<label>42.</label>
<citation citation-type="journal">
<person-group person-group-type="author">
<name>
<surname>Sun</surname>
<given-names>W</given-names>
</name>
<name>
<surname>Wang</surname>
<given-names>J</given-names>
</name>
<name>
<surname>Fang</surname>
<given-names>Y</given-names>
</name>
</person-group>. <article-title>Consistent Selection of Tuning Parameters via Variable Selection Stability</article-title>. <source>J&#x20;Machine Learn Res</source> (<year>2013</year>) <volume>14</volume>:<fpage>3419</fpage>&#x2013;<lpage>40</lpage>. </citation>
</ref>
</ref-list>
</back>
</article>