<?xml version="1.0" encoding="UTF-8" standalone="no"?>
<!DOCTYPE article PUBLIC "-//NLM//DTD Journal Publishing DTD v2.3 20070202//EN" "journalpublishing.dtd">
<article xmlns:mml="http://www.w3.org/1998/Math/MathML" xmlns:xlink="http://www.w3.org/1999/xlink" article-type="research-article">
<front>
<journal-meta>
<journal-id journal-id-type="publisher-id">Front. Phys.</journal-id>
<journal-title>Frontiers in Physics</journal-title>
<abbrev-journal-title abbrev-type="pubmed">Front. Phys.</abbrev-journal-title>
<issn pub-type="epub">2296-424X</issn>
<publisher>
<publisher-name>Frontiers Media S.A.</publisher-name>
</publisher>
</journal-meta>
<article-meta>
<article-id pub-id-type="doi">10.3389/fphy.2019.00210</article-id>
<article-categories>
<subj-group subj-group-type="heading">
<subject>Physics</subject>
<subj-group>
<subject>Original Research</subject>
</subj-group>
</subj-group>
</article-categories>
<title-group>
<article-title>Tackling the Trade-Off Between Information Processing Capacity and Rate in Delay-Based Reservoir Computers</article-title>
</title-group>
<contrib-group>
<contrib contrib-type="author" corresp="yes">
<name><surname>Ort&#x000ED;n</surname> <given-names>Silvia</given-names></name>
<xref ref-type="aff" rid="aff1"><sup>1</sup></xref>
<xref ref-type="corresp" rid="c001"><sup>&#x0002A;</sup></xref>
<uri xlink:href="http://loop.frontiersin.org/people/793475/overview"/>
</contrib>
<contrib contrib-type="author">
<name><surname>Pesquera</surname> <given-names>Luis</given-names></name>
<xref ref-type="aff" rid="aff2"><sup>2</sup></xref>
<uri xlink:href="http://loop.frontiersin.org/people/671041/overview"/>
</contrib>
</contrib-group>
<aff id="aff1"><sup>1</sup><institution>Unidad Cientifica de Innovaci&#x000F3;n empresarial, Instituto de Neurociencias, CSIC-UMH</institution>, <addr-line>Sant Joan d&#x00027;Alacant</addr-line>, <country>Spain</country></aff>
<aff id="aff2"><sup>2</sup><institution>Departamento de Estructura de la Materia, Instituto de F&#x000ED;sica de Cantabria, CSIC-UC</institution>, <addr-line>Santander</addr-line>, <country>Spain</country></aff>
<author-notes>
<fn fn-type="edited-by"><p>Edited by: V&#x000ED;ctor M. Egu&#x000ED;luz, Institute of Interdisciplinary Physics and Complex Systems (IFISC), Spain</p></fn>
<fn fn-type="edited-by"><p>Reviewed by: Alexander Vladimirovich Bogdanov, Saint Petersburg State University, Russia; Ignazio Licata, Institute for Scientific Methodology (ISEM), Italy; Guy Verschaffelt, Vrije University Brussel, Belgium; Apostolos Argyris, Institute of Interdisciplinary Physics and Complex Systems (IFISC), Spain</p></fn>
<corresp id="c001">&#x0002A;Correspondence: Silvia Ort&#x000ED;n <email>silortin&#x00040;gmail.com</email></corresp>
<fn fn-type="other" id="fn001"><p>This article was submitted to Interdisciplinary Physics, a section of the journal Frontiers in Physics</p></fn></author-notes>
<pub-date pub-type="epub">
<day>12</day>
<month>12</month>
<year>2019</year>
</pub-date>
<pub-date pub-type="collection">
<year>2019</year>
</pub-date>
<volume>7</volume>
<elocation-id>210</elocation-id>
<history>
<date date-type="received">
<day>20</day>
<month>08</month>
<year>2019</year>
</date>
<date date-type="accepted">
<day>21</day>
<month>11</month>
<year>2019</year>
</date>
</history>
<permissions>
<copyright-statement>Copyright &#x000A9; 2019 Ort&#x000ED;n and Pesquera.</copyright-statement>
<copyright-year>2019</copyright-year>
<copyright-holder>Ort&#x000ED;n and Pesquera</copyright-holder>
<license xlink:href="http://creativecommons.org/licenses/by/4.0/"><p>This is an open-access article distributed under the terms of the Creative Commons Attribution License (CC BY). The use, distribution or reproduction in other forums is permitted, provided the original author(s) and the copyright owner(s) are credited and that the original publication in this journal is cited, in accordance with accepted academic practice. No use, distribution or reproduction is permitted which does not comply with these terms.</p></license>
</permissions>
<abstract><p>We study the role of the system response time in the computational capacity of delay-based reservoir computers. Photonic hardware implementation of these systems offers high processing speed. However, delay-based reservoir computers have a trade-off between computational capacity and processing speed due to the non-zero response time of the non-linear node. The reservoir state is obtained from the sampled output of the non-linear node. We show that the computational capacity is degraded when the sampling output rate is higher than the inverse of the system response time. We find that the computational capacity depends not only on the sampling output rate but also on the misalignment between the delay time of the non-linear node and the data injection time. We show that the capacity degradation due to the high sampling output rate can be reduced when the delay time is greater than the data injection time. We find that this mismatch gives an improvement of the performance of delay-based reservoir computers for several benchmarking tasks. Our results show that the processing speed of delay-based reservoir computers can be increased while keeping a good computational capacity by using a mismatch between delay and data injection times. It is also shown that computational capacity for high sampling output rates can be further increased by using an extra feedback line and delay times greater than the data injection time.</p></abstract>
<kwd-group>
<kwd>reservoir computing</kwd>
<kwd>delayed-feedback systems</kwd>
<kwd>memory capacity</kwd>
<kwd>system response time</kwd>
<kwd>information processing rate</kwd>
</kwd-group>
<counts>
<fig-count count="8"/>
<table-count count="0"/>
<equation-count count="11"/>
<ref-count count="25"/>
<page-count count="12"/>
<word-count count="8654"/>
</counts>
</article-meta> 
</front>
<body>
<sec sec-type="intro" id="s1">
<title>1. Introduction</title>
<p>Reservoir computing (RC) is a successful brain-inspired concept to process information with temporal dependencies [<xref ref-type="bibr" rid="B1">1</xref>, <xref ref-type="bibr" rid="B2">2</xref>]. RC conceptually belongs to the field of recurrent neural networks (RNN) [<xref ref-type="bibr" rid="B3">3</xref>]. In these systems, the input signal is non-linearly projected onto a high-dimensional state space where the task can be solved much more easily than in the original input space. The high-dimensional space is typically a network of interconnected non-linear nodes (called neurons). The ensemble of neurons is called the reservoir. RC implementations are generally composed of three layers: input, reservoir, and output (see <xref ref-type="fig" rid="F1">Figure 1</xref>). The input layer feeds the input signal to the reservoir via fixed weighted connections. The input weights are often chosen randomly. These weights determine how strongly each of the inputs couples to each of the neurons. In traditional RNN the connections among the neurons are optimized to solve the task. Nevertheless, in RC, the coupling weights in the reservoir are not trained and can be chosen at random. The reservoir state is given by the combined states of all the individual nodes. Under the influence of input signals, the nodes of the reservoir remain in a transient state such that each input is injected in the presence of the response to the previous input. As a result the reservoir can retain input data for a finite amount of time (short-term memory [<xref ref-type="bibr" rid="B4">4</xref>]), and it can compute linear and non-linear functions of the retained information. The reservoir output is constructed through a linear combination of neural responses, with readout weights that are trained for the specific task. These weights are typically obtained by a simple linear regression. The strength of the reservoir computing scheme lies in the simplicity of its training method, where only the connections with the output are optimized.</p>
<fig id="F1" position="float">
<label>Figure 1</label>
<caption><p>Schematic illustration of delay-based RC. NL stands for Non-linear Node. The NL can have one (&#x003B2;<sub>2</sub> &#x0003D; 0) or two delay lines. The points <italic>r</italic><sub><italic>i</italic></sub>(<italic>n</italic>) represent the virtual nodes separated by time intervals &#x003B8;. The masked input <italic>u</italic>(<italic>n</italic>&#x0002B;1) &#x02297; Mask is injected directly following <italic>u</italic>(<italic>n</italic>) &#x02297;Mask.</p></caption>
<graphic xlink:href="fphy-07-00210-g0001.tif"/>
</fig>
<p>Hardware implementations of RC are sought because they offer high processing speed [<xref ref-type="bibr" rid="B5">5</xref>], parallelism, and low power consumption [<xref ref-type="bibr" rid="B6">6</xref>] compared to digital implementations. However, traditional RC involves a large number of interconnected non-linear neurons, so the hardware implementation is very challenging. Recently, it has been shown that RC can be efficiently implemented using a single non-linear dynamical system (neuron) subject to delayed feedback (delay-based RC) [<xref ref-type="bibr" rid="B7">7</xref>]. This architecture emulates the dynamic complexity traditionally achieved by a network of neurons. In delay-based RC, the spatial multiplexing of the input in standard RC systems with N neurons is replaced by time-multiplexing (see <xref ref-type="fig" rid="F1">Figure 1</xref>). The reservoir is composed of N sampled outputs of the non-linear node distributed along the delay line, called virtual nodes. Connections between these N virtual nodes are established through the delayed feedback when a mismatch between the delay and data injection times is introduced [<xref ref-type="bibr" rid="B8">8</xref>]. Delay-based RC has facilitated hardware implementation in photonic systems that have the potential to develop high-speed information processing. An overview of recent advances is given in Van der Sande et al. [<xref ref-type="bibr" rid="B9">9</xref>]. However, the information processing rate is limited by the non-zero response time of the system. The reservoir state is obtained from the sampled output of the non-linear node. The information processing (or data injection) time is given by <italic>T</italic><sub><italic>p</italic></sub> &#x0003D; <italic>N&#x003B8;</italic>, where &#x003B8; is the inverse of the output sampling rate, i.e., the time interval between two virtual nodes (see <xref ref-type="fig" rid="F1">Figure 1</xref>). The information processing rate <inline-formula><mml:math id="M1"><mml:msup><mml:mrow><mml:msub><mml:mrow><mml:mi>T</mml:mi></mml:mrow><mml:mrow><mml:mi>p</mml:mi></mml:mrow></mml:msub></mml:mrow><mml:mrow><mml:mo>-</mml:mo><mml:mn>1</mml:mn></mml:mrow></mml:msup></mml:math></inline-formula> can be increased by decreasing the node distance (higher sampling output rate). However, when &#x003B8; is less than the response time of the system <italic>T</italic>, virtual nodes are connected through the non-linear node dynamics. Network connections due to inertia lead to virtual node-states with similar dependence on inputs. Then the number of independent virtual nodes decreases and the diversity of the reservoir states is reduced. As a consequence computational capacity is degraded. Then there is a trade-off between information processing capacity and rate in delay-based reservoir computers.</p>
<p>In this work we show, using numerical simulations, that the computational capacity is degraded when the sampling output rate is higher than the inverse of the system response time. We obtain the memory capacities for different values of &#x003B8;/<italic>T</italic> and the mismatch between the delay and data injection times. Until now only two different delay-based reservoir architectures have been considered: &#x003B8; &#x0003C; <italic>T</italic> without mismatch [<xref ref-type="bibr" rid="B7">7</xref>] and &#x003B8; &#x0226B; <italic>T</italic> with mismatch time &#x003B8; [<xref ref-type="bibr" rid="B8">8</xref>]. We find that the computational capacity depends not only on the sampling output rate but also on the misalignment between the delay time of the non-linear node and the data injection time. We show that the capacity degradation due to high sampling output rate can be reduced when the delay time is greater than the data injection time. We also find that this mismatch gives an improvement of the performance of delay-based reservoir computers for several benchmarking tasks. Then, delay-based reservoir computers can achieve a high processing speed and good computational capacity using a mismatch between delay and data injection times.</p>
<p>We first consider a simple architecture of a single non-linear node with one feedback delay line. The linear and non-linear information processing capacities are obtained for different values of &#x003B8;/<italic>T</italic>. It is found that information processing capacity is boosted for small values of &#x003B8;/<italic>T</italic> if the delay of the non-linear node &#x003C4; is greater than <italic>T</italic><sub><italic>p</italic></sub>. A similar performance is obtained for small and large values of &#x003B8;/<italic>T</italic> for channel equalization and also for NARMA-10 task if values of the delay time greater than <italic>T</italic><sub><italic>p</italic></sub> are used. Then the information processing rate is increased without causing system performance degradation. This is due to the increase in reservoir diversity. Another strategy to increase reservoir diversity is to use an extra feedback line. We show that memory capacity can be further increased with this architecture for small values of &#x003B8;/<italic>T</italic> when the delay time is greater than the information processing time.</p></sec>
<sec sec-type="materials and methods" id="s2">
<title>2. Materials and Methods</title>
<sec>
<title>2.1. Delay-Based Reservoir Computers</title>
<p>Traditional RC implementations consist of a large number <italic>N</italic> of randomly interconnected non-linear nodes [<xref ref-type="bibr" rid="B3">3</xref>]. The state of the reservoir at time step <italic>n</italic>, <bold>r</bold>(<italic>n</italic>), is determined by:</p>
<disp-formula id="E1"><label>(1)</label><mml:math id="M2"><mml:mtable class="eqnarray" columnalign="right center left"><mml:mtr><mml:mtd><mml:mstyle mathvariant="bold"><mml:mtext>r</mml:mtext></mml:mstyle><mml:mrow><mml:mo stretchy="false">(</mml:mo><mml:mrow><mml:mi>n</mml:mi></mml:mrow><mml:mo stretchy="false">)</mml:mo></mml:mrow><mml:mo>=</mml:mo><mml:mi>f</mml:mi><mml:mrow><mml:mo stretchy="false">(</mml:mo><mml:mrow><mml:mi>&#x003B3;</mml:mi><mml:msup><mml:mrow><mml:mstyle mathvariant="bold"><mml:mtext>W</mml:mtext></mml:mstyle></mml:mrow><mml:mrow><mml:mi>i</mml:mi><mml:mi>n</mml:mi></mml:mrow></mml:msup><mml:mstyle mathvariant="bold"><mml:mtext>u</mml:mtext></mml:mstyle><mml:mrow><mml:mo stretchy="false">(</mml:mo><mml:mrow><mml:mi>n</mml:mi></mml:mrow><mml:mo stretchy="false">)</mml:mo></mml:mrow><mml:mo>&#x0002B;</mml:mo><mml:mi>&#x003B2;</mml:mi><mml:mstyle mathvariant="bold"><mml:mtext>W</mml:mtext></mml:mstyle><mml:mstyle mathvariant="bold"><mml:mtext>r</mml:mtext></mml:mstyle><mml:mrow><mml:mo stretchy="false">(</mml:mo><mml:mrow><mml:mi>n</mml:mi><mml:mo>-</mml:mo><mml:mn>1</mml:mn></mml:mrow><mml:mo stretchy="false">)</mml:mo></mml:mrow></mml:mrow><mml:mo stretchy="false">)</mml:mo></mml:mrow><mml:mo>,</mml:mo></mml:mtd></mml:mtr></mml:mtable></mml:math></disp-formula>
<p>where <bold>u</bold>(<italic>n</italic>) is sequentially injected input data and <italic>f</italic> is the reservoir activation function. The matrices <bold>W</bold> and <bold>W</bold><sup><italic>in</italic></sup> contain the (generally random) reservoir and input connection weights, respectively. The matrix <bold>W</bold> (<bold>W</bold><sup><italic>in</italic></sup>) is rescaled with a connection (input) scaling factor &#x003B2; (&#x003B3;). The exact internal connectivity is not crucial. In fact, it has been shown that simple non-random connection topologies (e.g., a simple chain or ring) gives a good performance [<xref ref-type="bibr" rid="B10">10</xref>].</p>
<p>Delay-based RC is a minimal approach to information processing based on the emulation of a recurrent network via a single non-linear dynamical node subject to delayed feedback. The reservoir nodes (called virtual nodes) are the sampled outputs of the non-linear node distributed along the delay line (see <xref ref-type="fig" rid="F1">Figure 1</xref>). In the time delay-based approach there is only one real non-linear node. Thus, the spatial multiplexing of the input in standard RC is replaced here by time multiplexing. The advantage of delay-based RC lies in the minimal hardware requirements. There is a price to pay for this hardware simplification: compared to an <italic>N</italic>-node standard spatially-distributed reservoir, the dynamical behaviour in the system has to run at an <italic>N</italic>-times higher speed in order to have equal input-throughput.</p>
<p>The dynamics of a delay-based reservoir has been described as [<xref ref-type="bibr" rid="B7">7</xref>, <xref ref-type="bibr" rid="B11">11</xref>&#x02013;<xref ref-type="bibr" rid="B16">16</xref>]:</p>
<disp-formula id="E2"><label>(2)</label><mml:math id="M3"><mml:mtable class="eqnarray" columnalign="right center left"><mml:mtr><mml:mtd><mml:mrow><mml:mi>T</mml:mi><mml:mover accent='true'><mml:mi>x</mml:mi><mml:mo>&#x002D9;</mml:mo></mml:mover><mml:mo stretchy="false">(</mml:mo><mml:mrow><mml:mi>t</mml:mi></mml:mrow><mml:mo stretchy="false">)</mml:mo></mml:mrow><mml:mo>=</mml:mo><mml:mo>-</mml:mo><mml:mi>x</mml:mi><mml:mrow><mml:mo stretchy="false">(</mml:mo><mml:mrow><mml:mi>t</mml:mi></mml:mrow><mml:mo stretchy="false">)</mml:mo></mml:mrow><mml:mo>&#x0002B;</mml:mo><mml:mi>f</mml:mi><mml:mrow><mml:mo stretchy="true">(</mml:mo><mml:mrow><mml:mi>&#x003B2;</mml:mi><mml:mi>x</mml:mi><mml:mrow><mml:mo stretchy="false">(</mml:mo><mml:mrow><mml:mi>t</mml:mi><mml:mo>-</mml:mo><mml:mi>&#x003C4;</mml:mi></mml:mrow><mml:mo stretchy="false">)</mml:mo></mml:mrow><mml:mo>&#x0002B;</mml:mo><mml:mi>&#x003B3;</mml:mi><mml:mi>J</mml:mi><mml:mrow><mml:mo stretchy="false">(</mml:mo><mml:mrow><mml:mi>t</mml:mi></mml:mrow><mml:mo stretchy="false">)</mml:mo></mml:mrow></mml:mrow><mml:mo stretchy="true">)</mml:mo></mml:mrow><mml:mo>,</mml:mo></mml:mtd></mml:mtr></mml:mtable></mml:math></disp-formula>
<p>where <italic>T</italic> is the response time of the system, &#x003C4; the delay time, &#x003B2; &#x0003E; 0 the feedback strength and &#x003B3; the input scaling. The masked input <italic>J</italic>(<italic>t</italic>) is the continuous version of the discrete random mapping of the original input <bold>W</bold><sup><italic>in</italic></sup><bold>u</bold>(<italic>n</italic>). In our approach, every time interval of the data injection/processing time <italic>T</italic><sub><italic>p</italic></sub> represents another discrete time step. This time is given by <italic>T</italic><sub><italic>p</italic></sub> &#x0003D; <italic>N&#x003B8;</italic>, where &#x003B8; is the temporal separation between virtual nodes. Individual virtual nodes are addressed by time-multiplexing the input signal. An input mask is used to emulate the input weights of traditional RC. This mask function is a piecewise constant function, constant over an interval of &#x003B8;, and periodic with period <italic>T</italic><sub><italic>p</italic></sub>. The N mask values <italic>m</italic><sub><italic>i</italic></sub> are drawn from a random uniform distribution in the interval [&#x02013;1,1] The procedure to construct the continuous data <italic>J</italic>(<italic>t</italic>) is the following. First, the input stream <italic>u</italic>(<italic>n</italic>) undergoes a sample and hold operation to define a stream which is constant during one <italic>T</italic><sub><italic>p</italic></sub>, before it is updated. Every segment of length <italic>T</italic><sub><italic>p</italic></sub> is multiplied by the mask (see <xref ref-type="fig" rid="F1">Figure 1</xref>). The masked input <italic>u</italic>(<italic>n</italic>&#x0002B;1)&#x02297; Mask is injected directly following <italic>u</italic>(<italic>n</italic>)&#x02297;Mask. After a time <italic>T</italic><sub><italic>p</italic></sub>, each virtual node is updated.</p>
<p>The reservoir state that corresponds to the input <italic>u</italic>(<italic>n</italic>), <bold>r</bold>(<italic>n</italic>) &#x0003D; [<italic>r</italic><sub>1</sub>(<italic>n</italic>)&#x02026;<italic>r</italic><sub><italic>N</italic></sub>(<italic>n</italic>)], is the collection of <italic>N</italic> outputs of the dynamical system, <italic>r</italic><sub><italic>i</italic></sub>(<italic>n</italic>) &#x0003D; <italic>x</italic>(<italic>nT</italic><sub><italic>p</italic></sub> &#x02212; (<italic>N</italic> &#x02212; <italic>i</italic>)&#x003B8;), where <italic>i</italic> &#x0003D; 1, &#x02026;, <italic>N</italic> (see <xref ref-type="fig" rid="F1">Figure 1</xref>). These <italic>N</italic> points are called virtual nodes because they correspond to taps in the delay line and play the same role as the neurons in standard RC. The node responses <italic>r</italic><sub><italic>i</italic></sub>(<italic>n</italic>) are used to train the reservoir to perform a specific task. As in the standard RC [<xref ref-type="bibr" rid="B1">1</xref>, <xref ref-type="bibr" rid="B17">17</xref>], only the output weights <bold>W</bold><sup><italic>out</italic></sup> are computed to obtain the output <italic>&#x00177;</italic> &#x0003D; <bold>W</bold><sup><italic>out</italic></sup> <bold>r</bold>. A linear regression method is used to minimize the error between the output &#x00177; and the desired target <italic>y</italic> in the training phase. The testing is then performed using previously unseen input data of the same kind as those used for training.</p>
<sec>
<title>2.1.1. Interconnection Structure of Delay-Based Reservoir Computers</title>
<p>In delay-based reservoir computers virtual nodes are connected through the feedback loop with nodes affected by previous inputs. Virtual node states also depend on close (in time) nodes through the inherent dynamics of the non-linear node. We can identify four time scales in the delayed feedback system with external input described by Equation (2): the response time <italic>T</italic> of the non-linear node, the delay time &#x003C4;, the separation of the virtual nodes &#x003B8;, and the data injection/processing time <italic>T</italic><sub><italic>p</italic></sub>. Setting the values of the different time scales creates a fixed interconnection structure. The virtual nodes can set up a network structure via the feedback loop by introducing a mismatch between <italic>T</italic><sub><italic>p</italic></sub> and &#x003C4;. Interconnection between virtual nodes due to the inherent dynamics of the non-linear node is obtained if the node separation &#x003B8; is smaller than the response time of the system <italic>T</italic>. Due to inertia the response of the system is not instantaneous. Therefore, the state of a virtual node depends on the states of nodes that correspond to previous taps in the delay line. However, if &#x003B8; is too short, the non-linear node will not be able to follow the changes in the input signal and the response signal will be too small to measure. Typically, a number of &#x003B8; &#x0003D; 0.2<italic>T</italic> is quoted [<xref ref-type="bibr" rid="B7">7</xref>, <xref ref-type="bibr" rid="B11">11</xref>&#x02013;<xref ref-type="bibr" rid="B16">16</xref>, <xref ref-type="bibr" rid="B18">18</xref>].</p>
<p>When &#x003B8; &#x0226B; <italic>T</italic> the state of a given virtual node is independent of the states of the neighboring virtual nodes. Then virtual nodes are not coupled through the non-linear node dynamics. The reservoir state is only determined by the instantaneous value of the input <italic>J</italic>(<italic>t</italic>) and the delayed reservoir state. The system given by Equation (2) can then be described with a map:</p>
<disp-formula id="E3"><label>(3)</label><mml:math id="M4"><mml:mtable class="eqnarray" columnalign="right center left"><mml:mtr><mml:mtd><mml:mi>x</mml:mi><mml:mrow><mml:mo stretchy="false">(</mml:mo><mml:mrow><mml:mi>t</mml:mi></mml:mrow><mml:mo stretchy="false">)</mml:mo></mml:mrow><mml:mo>=</mml:mo><mml:mi>f</mml:mi><mml:mrow><mml:mo stretchy="true">(</mml:mo><mml:mrow><mml:mi>&#x003B2;</mml:mi><mml:mi>x</mml:mi><mml:mrow><mml:mo stretchy="false">(</mml:mo><mml:mrow><mml:mi>t</mml:mi><mml:mo>-</mml:mo><mml:mi>&#x003C4;</mml:mi></mml:mrow><mml:mo stretchy="false">)</mml:mo></mml:mrow><mml:mo>&#x0002B;</mml:mo><mml:mi>&#x003B3;</mml:mi><mml:mi>J</mml:mi><mml:mrow><mml:mo stretchy="false">(</mml:mo><mml:mrow><mml:mi>t</mml:mi></mml:mrow><mml:mo stretchy="false">)</mml:mo></mml:mrow></mml:mrow><mml:mo stretchy="true">)</mml:mo></mml:mrow><mml:mo>.</mml:mo></mml:mtd></mml:mtr></mml:mtable></mml:math></disp-formula>
<p>A network structure can be obtained via the feedback loop by introducing a mismatch between <italic>T</italic><sub><italic>p</italic></sub> and &#x003C4;. This mismatch can be quantified in terms of the number of virtual nodes by &#x003B1; &#x0003D; (&#x003C4; &#x02212; <italic>N&#x003B8;</italic>)/&#x003B8;. In the case of 0 &#x02264; &#x003B1; &#x0003C; <italic>N</italic> and &#x003B8; &#x0226B; <italic>T</italic>, the virtual node states are given by:</p>
<disp-formula id="E4"><mml:math id="M5"><mml:mrow><mml:msub><mml:mi>r</mml:mi><mml:mi>i</mml:mi></mml:msub><mml:mo stretchy='false'>(</mml:mo><mml:mi>n</mml:mi><mml:mo stretchy='false'>)</mml:mo><mml:mo>=</mml:mo><mml:mrow><mml:mo>{</mml:mo><mml:mrow><mml:mtable columnalign='left'><mml:mtr columnalign='left'><mml:mtd columnalign='left'><mml:mrow><mml:mi>f</mml:mi><mml:mo stretchy='false'>(</mml:mo><mml:mi>&#x003B2;</mml:mi><mml:msub><mml:mi>r</mml:mi><mml:mrow><mml:mi>i</mml:mi><mml:mo>&#x02212;</mml:mo><mml:mi>&#x003B1;</mml:mi></mml:mrow></mml:msub><mml:mo stretchy='false'>(</mml:mo><mml:mi>n</mml:mi><mml:mo>&#x02212;</mml:mo><mml:mn>1</mml:mn><mml:mo stretchy='false'>)</mml:mo><mml:mo>+</mml:mo><mml:mi>&#x003B3;</mml:mi><mml:msub><mml:mi>m</mml:mi><mml:mi>i</mml:mi></mml:msub><mml:mi>u</mml:mi><mml:mo stretchy='false'>(</mml:mo><mml:mi>n</mml:mi><mml:mo stretchy='false'>)</mml:mo><mml:mo stretchy='false'>)</mml:mo></mml:mrow></mml:mtd><mml:mtd columnalign='left'><mml:mrow><mml:mtext>&#x000A0;if&#x000A0;</mml:mtext><mml:mi>&#x003B1;</mml:mi><mml:mo>&#x0003C;</mml:mo><mml:mi>i</mml:mi><mml:mo>&#x02264;</mml:mo><mml:mi>N</mml:mi></mml:mrow></mml:mtd></mml:mtr><mml:mtr columnalign='left'><mml:mtd columnalign='left'><mml:mrow><mml:mi>f</mml:mi><mml:mo stretchy='false'>(</mml:mo><mml:mi>&#x003B2;</mml:mi><mml:msub><mml:mi>r</mml:mi><mml:mrow><mml:mi>N</mml:mi><mml:mo>+</mml:mo><mml:mi>i</mml:mi><mml:mo>&#x02212;</mml:mo><mml:mi>&#x003B1;</mml:mi></mml:mrow></mml:msub><mml:mo stretchy='false'>(</mml:mo><mml:mi>n</mml:mi><mml:mo>&#x02212;</mml:mo><mml:mn>2</mml:mn><mml:mo stretchy='false'>)</mml:mo><mml:mo>+</mml:mo><mml:mi>&#x003B3;</mml:mi><mml:msub><mml:mi>m</mml:mi><mml:mi>i</mml:mi></mml:msub><mml:mi>u</mml:mi><mml:mo stretchy='false'>(</mml:mo><mml:mi>n</mml:mi><mml:mo stretchy='false'>)</mml:mo><mml:mo stretchy='false'>)</mml:mo></mml:mrow></mml:mtd><mml:mtd columnalign='left'><mml:mrow><mml:mtext>&#x000A0;if&#x000A0;</mml:mtext><mml:mi>i</mml:mi><mml:mo>&#x02264;</mml:mo><mml:mi>&#x003B1;</mml:mi></mml:mrow></mml:mtd></mml:mtr></mml:mtable></mml:mrow></mml:mrow></mml:mrow></mml:math></disp-formula>
<p>The network topology depends on the value of &#x003B1;. When &#x003B1; &#x0003D; 1 (i.e., &#x003C4; &#x0003D; <italic>T</italic><sub><italic>p</italic></sub>&#x0002B;&#x003B8;) the topology is equivalent to the ring topology in standard RC systems [<xref ref-type="bibr" rid="B10">10</xref>]. When &#x003B1; &#x0003C; 0, a number |&#x003B1;| of virtual nodes are not connected through the feedback line with nodes at a previous time. When &#x003B1; and <italic>N</italic> have no common divisors, all virtual nodes are connected through feedback in a single ring. However, when <italic>N</italic> and &#x003B1; are not coprimes, subnetworks are formed with a similar dependence on inputs and the reservoir diversity is reduced.</p>
<p>Although the two types of virtual node connections are not exclusive, only two cases have been considered until now: delay-based reservoirs connected through system dynamics (&#x003B1; &#x0003D; 0 and &#x003B8; &#x0003C; <italic>T</italic>) [<xref ref-type="bibr" rid="B7">7</xref>, <xref ref-type="bibr" rid="B12">12</xref>&#x02013;<xref ref-type="bibr" rid="B18">18</xref>], or by the feedback line (&#x003B8; &#x0226B; <italic>T</italic>) [<xref ref-type="bibr" rid="B8">8</xref>, <xref ref-type="bibr" rid="B15">15</xref>, <xref ref-type="bibr" rid="B19">19</xref>].</p>
<p>It is clear that the information processing rate of delay-based reservoir computers <inline-formula><mml:math id="M6"><mml:msup><mml:mrow><mml:msub><mml:mrow><mml:mi>T</mml:mi></mml:mrow><mml:mrow><mml:mi>p</mml:mi></mml:mrow></mml:msub></mml:mrow><mml:mrow><mml:mo>-</mml:mo><mml:mn>1</mml:mn></mml:mrow></mml:msup></mml:math></inline-formula> depends on the node separation. Then reservoir computers with nodes connected only through the feedback line (&#x003B8; &#x0226B; <italic>T</italic>) are slower than a counterpart exploiting the virtual connections through the system dynamics (&#x003B8; &#x0003C; <italic>T</italic>). However, as we will show in 3.1, information processing capacity is degraded when &#x003B8; &#x0003C; <italic>T</italic>. In this case, the computational capacity increases with the mismatch between the delay and data injection times (see section 3.1).</p>
</sec></sec>
<sec>
<title>2.2. Computational Capacity</title>
<p>Delay-based reservoir computers can reconstruct functions of <italic>h</italic> previous inputs <bold><italic>y</italic><sub><italic>k</italic></sub></bold>(<italic>n</italic>) &#x0003D; <italic>y</italic>(<italic>u</italic>(<italic>n</italic> &#x02212; <italic>k</italic><sub>1</sub>), &#x02026;, <italic>u</italic>(<italic>n</italic> &#x02212; <italic>k</italic><sub><italic>h</italic></sub>)) from the state of a dynamical system using a linear estimator <inline-formula><mml:math id="M7"><mml:msub><mml:mrow><mml:mover accent="true"><mml:mrow><mml:mstyle mathvariant="bold"><mml:mtext>y</mml:mtext></mml:mstyle></mml:mrow><mml:mo>^</mml:mo></mml:mover></mml:mrow><mml:mrow><mml:mstyle mathvariant="bold"><mml:mtext>k</mml:mtext></mml:mstyle></mml:mrow></mml:msub></mml:math></inline-formula>. Here <bold>k</bold> denotes the vector (<italic>k</italic><sub>1</sub>, &#x02026;, <italic>k</italic><sub><italic>h</italic></sub>). The estimator <inline-formula><mml:math id="M8"><mml:msub><mml:mrow><mml:mover accent="true"><mml:mrow><mml:mstyle mathvariant="bold"><mml:mtext>y</mml:mtext></mml:mstyle></mml:mrow><mml:mo>^</mml:mo></mml:mover></mml:mrow><mml:mrow><mml:mstyle mathvariant="bold"><mml:mtext>k</mml:mtext></mml:mstyle></mml:mrow></mml:msub></mml:math></inline-formula> is obtained from <italic>N</italic> internal variables (node states) of the system. The suitability of a reservoir to reconstruct <bold>y<sub>k</sub></bold> can be quantified by using the capacity [<xref ref-type="bibr" rid="B20">20</xref>]:</p>
<disp-formula id="E5"><label>(4)</label><mml:math id="M9"><mml:mtable class="eqnarray" columnalign="right center left"><mml:mtr><mml:mtd><mml:mrow><mml:mi>C</mml:mi><mml:mrow><mml:mo>[</mml:mo><mml:mrow><mml:msub><mml:mstyle mathvariant='bold' mathsize='normal'><mml:mi>y</mml:mi></mml:mstyle><mml:mstyle mathvariant='bold' mathsize='normal'><mml:mi>k</mml:mi></mml:mstyle></mml:msub></mml:mrow><mml:mo>]</mml:mo></mml:mrow><mml:mo>=</mml:mo><mml:mo stretchy='false'>(</mml:mo><mml:mn>1</mml:mn><mml:mo>&#x02212;</mml:mo><mml:mfrac><mml:mrow><mml:mstyle displaystyle='true'><mml:msub><mml:mo>&#x02211;</mml:mo><mml:mi>n</mml:mi></mml:msub><mml:mrow><mml:mo stretchy='false'>(</mml:mo><mml:msub><mml:mrow><mml:mover accent='true'><mml:mstyle mathvariant='bold' mathsize='normal'><mml:mi>y</mml:mi></mml:mstyle><mml:mo stretchy='true'>&#x0005E;</mml:mo></mml:mover></mml:mrow><mml:mstyle mathvariant='bold' mathsize='normal'><mml:mi>k</mml:mi></mml:mstyle></mml:msub><mml:mo stretchy='false'>(</mml:mo></mml:mrow></mml:mstyle><mml:mi>n</mml:mi><mml:mo stretchy='false'>)</mml:mo><mml:mo>&#x02212;</mml:mo><mml:msub><mml:mstyle mathvariant='bold' mathsize='normal'><mml:mi>y</mml:mi></mml:mstyle><mml:mstyle mathvariant='bold' mathsize='normal'><mml:mi>k</mml:mi></mml:mstyle></mml:msub><mml:mo stretchy='false'>(</mml:mo><mml:mi>n</mml:mi><mml:mo stretchy='false'>)</mml:mo><mml:msup><mml:mo stretchy='false'>)</mml:mo><mml:mn>2</mml:mn></mml:msup></mml:mrow><mml:mrow><mml:mstyle displaystyle='true'><mml:msub><mml:mo>&#x02211;</mml:mo><mml:mi>n</mml:mi></mml:msub><mml:mrow><mml:mo stretchy='false'>(</mml:mo><mml:msub><mml:mstyle mathvariant='bold' mathsize='normal'><mml:mi>y</mml:mi></mml:mstyle><mml:mstyle mathvariant='bold' mathsize='normal'><mml:mi>k</mml:mi></mml:mstyle></mml:msub><mml:mo stretchy='false'>(</mml:mo></mml:mrow></mml:mstyle><mml:mi>n</mml:mi><mml:mo stretchy='false'>)</mml:mo><mml:msup><mml:mo stretchy='false'>)</mml:mo><mml:mn>2</mml:mn></mml:msup></mml:mrow></mml:mfrac><mml:mo stretchy='false'>)</mml:mo><mml:mo>.</mml:mo></mml:mrow></mml:mtd></mml:mtr></mml:mtable></mml:math></disp-formula>
<p>The capacity is <italic>C</italic>[<bold>y<sub>k</sub></bold>] &#x0003D; 1 when the reconstruction error for <bold>y<sub>k</sub></bold> is zero. The capacity for reconstructing a function of the inputs <bold>y</bold>, <italic>C</italic>[<bold>y</bold>], is given by the sum of <italic>C</italic>[<bold>y<sub>k</sub></bold>] over all sequences of past inputs [<xref ref-type="bibr" rid="B20">20</xref>]:</p>
<disp-formula id="E6"><label>(5)</label><mml:math id="M10"><mml:mtable class="eqnarray" columnalign="right center left"><mml:mtr><mml:mtd><mml:mrow><mml:mi>C</mml:mi><mml:mo stretchy='false'>[</mml:mo><mml:mstyle mathvariant='bold'><mml:mtext>y</mml:mtext></mml:mstyle><mml:mo stretchy='false'>]</mml:mo><mml:mo>=</mml:mo><mml:mstyle displaystyle='true'><mml:munder><mml:mo>&#x02211;</mml:mo><mml:mstyle mathvariant='bold'><mml:mi>k</mml:mi></mml:mstyle></mml:munder><mml:mrow><mml:mi>C</mml:mi><mml:mo stretchy='true'>[</mml:mo><mml:msub><mml:mstyle mathvariant='bold'><mml:mtext>y</mml:mtext></mml:mstyle><mml:mstyle mathvariant='bold'><mml:mtext>k</mml:mtext></mml:mstyle></mml:msub><mml:mo stretchy='true'>]</mml:mo><mml:mo>.</mml:mo></mml:mrow></mml:mstyle></mml:mrow></mml:mtd></mml:mtr></mml:mtable></mml:math></disp-formula>
<p>The total computational capacity <italic>C</italic><sub><italic>T</italic></sub> is the sum of <italic>C</italic>[<bold>y<sub>k</sub></bold>] over all sequences of past inputs and a complete orthonormal set of functions. When <bold>y<sub>k</sub></bold> is a linear function of one of the past inputs, <bold>y<sub>k</sub></bold>(<italic>n</italic>) &#x0003D; <italic>u</italic>(<italic>n</italic> &#x02212; <italic>k</italic>), the capacity <italic>C</italic>[<bold>y</bold>] corresponds to the linear memory capacity introduced in Jaeger [<xref ref-type="bibr" rid="B4">4</xref>]. The capacity of the system to compute non-linear functions of the retained information is given by the non-linear memory capacity [<xref ref-type="bibr" rid="B20">20</xref>]. The computational capacity is given by the sum of the linear and non-linear memory capacities. The total capacity is limited by the dimension of the reservoir. As a consequence, there is a trade-off between linear and non-linear memory capacities [<xref ref-type="bibr" rid="B20">20</xref>].</p>
<p>The total computational capacity of delay-based reservoirs is given by the number of linearly independent virtual nodes. The computational power of delay-based reservoir computers is therefore hidden in the diversity of the reservoir states. In the presence of inertia (&#x003B8; &#x0003C; <italic>T</italic>) non-linear node dynamics couples close (in time) virtual nodes. This coupling reduces reservoir diversity, and then computational capacity is degraded. The computational capacity of delay-based reservoir depends not only on the separation between the virtual nodes but also on the misalignment between <italic>T</italic><sub><italic>p</italic></sub> and &#x003C4;, given by &#x003B1;. When &#x003B1; &#x0003C; 0, the state of a virtual node of index <italic>i</italic> &#x0003E; (<italic>N</italic> &#x02212; |&#x003B1;|), <italic>r</italic><sub><italic>i</italic></sub>(<italic>n</italic>), is a function of the virtual node state <italic>r</italic><sub><italic>i</italic>&#x02212;<italic>N</italic>&#x0002B;|&#x003B1;|</sub>(<italic>n</italic>) at the same time. Then the reservoir diversity and computational capacity are reduced. Computational capacity is also reduced if |&#x003B1;| and <italic>N</italic> are not coprimes. In this case, the <italic>N</italic> virtual nodes form <italic>gcd</italic>(|&#x003B1;|, <italic>N</italic>) ring subnetworks, where <italic>gcd</italic> is the greatest common divisor. Each subnetwork has <italic>p</italic> &#x0003D; <italic>N</italic>/<italic>gcd</italic>(|&#x003B1;|, <italic>N</italic>) virtual nodes. Virtual node-states belonging to different subnetworks have a similar dependence on inputs and reservoir diversity is reduced.</p></sec>
<sec>
<title>2.3. Reservoir Computers With Two Delay Lines</title>
<p>An architecture with several delay lines has been proposed [<xref ref-type="bibr" rid="B21">21</xref>, <xref ref-type="bibr" rid="B22">22</xref>] to increase the memory capacity of delay-based reservoir computers with virtual nodes connected only through non-linear system dynamics (&#x003B8; &#x0003C; <italic>T</italic> and &#x003B1; &#x0003D; 0). Several delay lines are added to preserve older information. The longer the delay, the older the response that is being fed back. Even without explicitly reading the older states from the delay line, the information is re-injected into the system and its memory can be extended. We apply this approach to delay-based reservoir computers with virtual nodes that are connected through non-linear node dynamics and by the feedback line.</p>
<p>The dynamics of reservoir computers with two delay lines is described by:</p>
<disp-formula id="E7"><label>(6)</label><mml:math id="M11"><mml:mtable class="eqnarray" columnalign="right center left"><mml:mtr><mml:mtd><mml:mrow><mml:mi>T</mml:mi><mml:mover accent='true'><mml:mi>x</mml:mi><mml:mo>&#x002D9;</mml:mo></mml:mover><mml:mo stretchy="false">(</mml:mo><mml:mrow><mml:mi>t</mml:mi></mml:mrow><mml:mo stretchy="false">)</mml:mo></mml:mrow><mml:mo>=</mml:mo><mml:mo>-</mml:mo><mml:mi>x</mml:mi><mml:mrow><mml:mo stretchy="false">(</mml:mo><mml:mrow><mml:mi>t</mml:mi></mml:mrow><mml:mo stretchy="false">)</mml:mo></mml:mrow><mml:mo>&#x0002B;</mml:mo><mml:mi>f</mml:mi><mml:mrow><mml:mo stretchy="true">(</mml:mo><mml:mrow><mml:msub><mml:mrow><mml:mi>&#x003B2;</mml:mi></mml:mrow><mml:mrow><mml:mn>1</mml:mn></mml:mrow></mml:msub><mml:mi>x</mml:mi><mml:mrow><mml:mo stretchy="false">(</mml:mo><mml:mrow><mml:mi>t</mml:mi><mml:mo>-</mml:mo><mml:msub><mml:mrow><mml:mi>&#x003C4;</mml:mi></mml:mrow><mml:mrow><mml:mn>1</mml:mn></mml:mrow></mml:msub></mml:mrow><mml:mo stretchy="false">)</mml:mo></mml:mrow><mml:mo>&#x0002B;</mml:mo><mml:msub><mml:mrow><mml:mi>&#x003B2;</mml:mi></mml:mrow><mml:mrow><mml:mn>2</mml:mn></mml:mrow></mml:msub><mml:mi>x</mml:mi><mml:mrow><mml:mo stretchy="false">(</mml:mo><mml:mrow><mml:mi>t</mml:mi><mml:mo>-</mml:mo><mml:msub><mml:mrow><mml:mi>&#x003C4;</mml:mi></mml:mrow><mml:mrow><mml:mn>2</mml:mn></mml:mrow></mml:msub></mml:mrow><mml:mo stretchy="false">)</mml:mo></mml:mrow><mml:mo>&#x0002B;</mml:mo><mml:mi>&#x003B3;</mml:mi><mml:mi>J</mml:mi><mml:mrow><mml:mo stretchy="false">(</mml:mo><mml:mrow><mml:mi>t</mml:mi></mml:mrow><mml:mo stretchy="false">)</mml:mo></mml:mrow></mml:mrow><mml:mo stretchy="true">)</mml:mo></mml:mrow><mml:mo>,</mml:mo></mml:mtd></mml:mtr></mml:mtable></mml:math></disp-formula>
<p>where &#x003B2;<sub><italic>i</italic></sub> &#x02265; 0 is the feedback strength of the delay line <italic>i</italic>. The total feedback strength is &#x003B2; &#x0003D; &#x003B2;<sub>1</sub> &#x0002B; &#x003B2;<sub>2</sub>. The corresponding delays are given by &#x003C4;<sub>1</sub> &#x0003D; <italic>N&#x003B8;</italic> &#x0002B; &#x003B1;<sub>1</sub> and &#x003C4;<sub>2</sub> &#x0003D; 2<italic>N&#x003B8;</italic> &#x0002B; &#x003B1;<sub>2</sub>, where 0 &#x02264; &#x003B1;<sub><italic>i</italic></sub> &#x0003C; <italic>N&#x003B8;</italic>. The reservoir state is the same as in one delay-based RC, i.e., the virtual nodes correspond to taps only in the shorter (&#x003C4;<sub>1</sub>) delay line. In the case of &#x003B1;<sub>1</sub> &#x0003D; 0, it has been shown [<xref ref-type="bibr" rid="B23">23</xref>] that the best performance for NARMA-10 task is obtained when &#x003C4;<sub>1</sub> and &#x003C4;<sub>2</sub> are coprimes. In this case, the number of virtual nodes that are mixed together within the history of each virtual node is maximized.</p>
<p>If the mismatches &#x003B1;<sub><italic>i</italic></sub> (<italic>i</italic> &#x0003D; 1, 2) are zero, the virtual node states at time <italic>n</italic> depend on the reservoir state at time (<italic>n</italic> &#x02212; 1) and (<italic>n</italic> &#x02212; 2) via the delay line 1 and 2, respectively. In one-delay reservoirs (&#x003B2;<sub>2</sub> &#x0003D; 0), the number of virtual nodes whose state at time <italic>n</italic> depends on the reservoir state at time (<italic>n</italic> &#x02212; 2) increases with the mismatch (see Equation 2.1.1 for the case without inertia). When a second delay is added with a mismatch &#x003B1;<sub>2</sub> &#x0003E; 0, some virtual nodes at time <italic>n</italic> are connected with nodes at time (<italic>n</italic> &#x02212; 3). The number of virtual nodes with states at time <italic>n</italic> that depend on the reservoir state at time (<italic>n</italic> &#x02212; 3) increases with &#x003B1;<sub>2</sub>. These connections with older states can extend the memory of the two-delay reservoir computer.</p>
</sec></sec>
<sec sec-type="results" id="s3">
<title>3. Results</title>
<p>In this section, we show the numerical results obtained for the memory capacities and performance of a non-linear delay-based RC system. We study a delay-based reservoir computer with a single non-linear node for the one and two delay lines architectures. The one-delay system is governed by Equation (2) and the two-delay reservoir by Equation (6). In both cases the reservoir activation function <italic>f</italic> is given by:</p>
<disp-formula id="E8"><label>(7)</label><mml:math id="M12"><mml:mtable class="eqnarray" columnalign="right center left"><mml:mtr><mml:mtd><mml:mi>f</mml:mi><mml:mrow><mml:mo stretchy="false">(</mml:mo><mml:mrow><mml:mi>z</mml:mi></mml:mrow><mml:mo stretchy="false">)</mml:mo></mml:mrow><mml:mo>=</mml:mo><mml:msub><mml:mrow><mml:mi>f</mml:mi></mml:mrow><mml:mrow><mml:mi>s</mml:mi></mml:mrow></mml:msub><mml:mfrac><mml:mrow><mml:mn>1</mml:mn><mml:mo>-</mml:mo><mml:mo class="qopname">exp</mml:mo><mml:mrow><mml:mo stretchy="false">(</mml:mo><mml:mrow><mml:mo>-</mml:mo><mml:mo>&#x003BB;</mml:mo><mml:mi>z</mml:mi></mml:mrow><mml:mo stretchy="false">)</mml:mo></mml:mrow></mml:mrow><mml:mrow><mml:mi>a</mml:mi><mml:mo>&#x0002B;</mml:mo><mml:mo class="qopname">exp</mml:mo><mml:mrow><mml:mo stretchy="false">(</mml:mo><mml:mrow><mml:mo>-</mml:mo><mml:mo>&#x003BB;</mml:mo><mml:mi>z</mml:mi></mml:mrow><mml:mo stretchy="false">)</mml:mo></mml:mrow></mml:mrow></mml:mfrac><mml:mo>,</mml:mo></mml:mtd></mml:mtr></mml:mtable></mml:math></disp-formula>
<p>where <italic>a</italic> &#x0003D; 2 and &#x003BB; &#x0003D; 1. The value of <italic>f</italic><sub><italic>s</italic></sub> &#x0003D; 2.5 is chosen to have, when &#x003B2; &#x0003C; 1, a stable fixed point for the system defined by Equation (2) in absence of input (&#x003B3; &#x0003D; 0). This non-linear function is asymmetric to allow that the reservoir computer reconstructs even functions of the input. Similar results are obtained for different reservoir activation functions, in particular for a <italic>sin</italic><sup>2</sup> function, that corresponds to an optoelectronic implementation [<xref ref-type="bibr" rid="B8">8</xref>, <xref ref-type="bibr" rid="B11">11</xref>, <xref ref-type="bibr" rid="B13">13</xref>&#x02013;<xref ref-type="bibr" rid="B15">15</xref>].</p>
<p>The number of virtual nodes used in the numerical simulations is a prime number, <italic>N</italic> &#x0003D; 97, to avoid the capacity degradation due to the formation of subnetworks. The rest of fixed parameters are: <italic>T</italic> &#x0003D; 1 and &#x003B2; &#x0003D; &#x003B2;<sub>1</sub> &#x0003D; 0.8 for the one-delay reservoir computer and &#x003B2;<sub>1</sub> &#x0002B; &#x003B2;<sub>2</sub> &#x0003D; &#x003B2; &#x0003D; 0.8 for the two-delay reservoir computer. The effective non-linearity of the delay-based reservoir computer can be changed with the scaling input parameter &#x003B3;. In this work, we consider &#x003B3; &#x0003D; 0.1 and &#x003B3; &#x0003D; 1 that correspond to low-to-moderate and strong non-linearity, respectively. The total capacity of a linear reservoir computer with <italic>f</italic>(<italic>z</italic>) &#x0003D; <italic>z</italic> will also be analyzed.</p>
<p>All the results presented in this paper are the average over 5 simulation runs with different training/test sets and different masks. A total of 8,000 inputs (6,000 for training and 2,000 for testing) are used for computational capacities and the NARMA-10 task. The dataset for the channel equalization task has 10,000 points for training and 6,000 for testing.</p>
<sec>
<title>3.1. Computational Capacity</title>
<p>To analyze the computational capacity of the non-linear delay-based reservoir computer, we calculate by using (Equations 4 and 5) four capacities as in Duport et al. [<xref ref-type="bibr" rid="B19">19</xref>], namely linear (LMC), quadratic (QMC), cubic (CMC) and cross (XMC) memory capacities, which correspond to functions <bold>y</bold> given by the first, second and third order Legendre polynomials, respectively. In order to obtain these capacities a series of i.i.d. input samples drawn uniformly from the interval [&#x02013;1, 1] is injected into the reservoir. The LMC is obtained by summing over <italic>k</italic> the capacity <italic>C</italic>[<bold>y<sub>k</sub></bold>] for reconstructing <bold>y</bold><sub><italic>k</italic></sub>(<italic>n</italic>) &#x0003D; <italic>u</italic>(<italic>n</italic> &#x02212; <italic>k</italic>). It corresponds to the linear memory capacity introduced in Jaeger [<xref ref-type="bibr" rid="B4">4</xref>]. The QMC and CMC are obtained by summing over <italic>k</italic> the capacity for <inline-formula><mml:math id="M13"><mml:msub><mml:mrow><mml:mstyle mathvariant="bold"><mml:mtext>y</mml:mtext></mml:mstyle></mml:mrow><mml:mrow><mml:mi>k</mml:mi></mml:mrow></mml:msub><mml:mrow><mml:mo stretchy="false">(</mml:mo><mml:mrow><mml:mi>n</mml:mi></mml:mrow><mml:mo stretchy="false">)</mml:mo></mml:mrow><mml:mo>=</mml:mo><mml:mrow><mml:mo stretchy="false">(</mml:mo><mml:mrow><mml:mn>3</mml:mn><mml:msup><mml:mrow><mml:mi>u</mml:mi></mml:mrow><mml:mrow><mml:mn>2</mml:mn></mml:mrow></mml:msup><mml:mrow><mml:mo stretchy="false">(</mml:mo><mml:mrow><mml:mi>n</mml:mi><mml:mo>-</mml:mo><mml:mi>k</mml:mi></mml:mrow><mml:mo stretchy="false">)</mml:mo></mml:mrow><mml:mo>-</mml:mo><mml:mn>1</mml:mn></mml:mrow><mml:mo stretchy="false">)</mml:mo></mml:mrow><mml:mo>/</mml:mo><mml:mn>2</mml:mn></mml:math></inline-formula> and <inline-formula><mml:math id="M14"><mml:msub><mml:mrow><mml:mstyle mathvariant="bold"><mml:mtext>y</mml:mtext></mml:mstyle></mml:mrow><mml:mrow><mml:mi>k</mml:mi></mml:mrow></mml:msub><mml:mrow><mml:mo stretchy="false">(</mml:mo><mml:mrow><mml:mi>n</mml:mi></mml:mrow><mml:mo stretchy="false">)</mml:mo></mml:mrow><mml:mo>=</mml:mo><mml:mrow><mml:mo stretchy="false">(</mml:mo><mml:mrow><mml:mn>5</mml:mn><mml:msup><mml:mrow><mml:mi>u</mml:mi></mml:mrow><mml:mrow><mml:mn>3</mml:mn></mml:mrow></mml:msup><mml:mrow><mml:mo stretchy="false">(</mml:mo><mml:mrow><mml:mi>n</mml:mi><mml:mo>-</mml:mo><mml:mi>k</mml:mi></mml:mrow><mml:mo stretchy="false">)</mml:mo></mml:mrow><mml:mo>-</mml:mo><mml:mn>3</mml:mn><mml:mi>u</mml:mi><mml:mrow><mml:mo stretchy="false">(</mml:mo><mml:mrow><mml:mi>n</mml:mi><mml:mo>-</mml:mo><mml:mi>k</mml:mi></mml:mrow><mml:mo stretchy="false">)</mml:mo></mml:mrow></mml:mrow><mml:mo stretchy="false">)</mml:mo></mml:mrow><mml:mo>/</mml:mo><mml:mn>2</mml:mn></mml:math></inline-formula>, respectively. The XMC is obtained by summing over <italic>k</italic>, <italic>k</italic>&#x02032; for <italic>k</italic> &#x0003C; <italic>k</italic>&#x02032; the capacities for the product of two inputs, <inline-formula><mml:math id="M15"><mml:msub><mml:mrow><mml:mstyle mathvariant="bold"><mml:mtext>y</mml:mtext></mml:mstyle></mml:mrow><mml:mrow><mml:mi>k</mml:mi><mml:mo>,</mml:mo><mml:msup><mml:mrow><mml:mi>k</mml:mi></mml:mrow><mml:mrow><mml:mi>&#x02032;</mml:mi></mml:mrow></mml:msup></mml:mrow></mml:msub><mml:mo>=</mml:mo><mml:mi>u</mml:mi><mml:mrow><mml:mo stretchy="false">(</mml:mo><mml:mrow><mml:mi>n</mml:mi><mml:mo>-</mml:mo><mml:mi>k</mml:mi></mml:mrow><mml:mo stretchy="false">)</mml:mo></mml:mrow><mml:mo>&#x000B7;</mml:mo><mml:mi>u</mml:mi><mml:mrow><mml:mo stretchy="false">(</mml:mo><mml:mrow><mml:mi>n</mml:mi><mml:mo>-</mml:mo><mml:msup><mml:mrow><mml:mi>k</mml:mi></mml:mrow><mml:mrow><mml:mi>&#x02032;</mml:mi></mml:mrow></mml:msup></mml:mrow><mml:mo stretchy="false">)</mml:mo></mml:mrow></mml:math></inline-formula>. In non-linear systems, the sum <italic>C</italic><sub><italic>s</italic></sub> &#x0003D; <italic>LMC</italic>&#x0002B;<italic>QMC</italic>&#x0002B;<italic>CMC</italic>&#x0002B;<italic>XMC</italic> does not include all possible contributions to <italic>C</italic><sub><italic>T</italic></sub>, so <italic>C</italic><sub><italic>s</italic></sub> &#x02264; <italic>C</italic><sub><italic>T</italic></sub>, whereas for linear systems <italic>C</italic><sub><italic>s</italic></sub> &#x0003D; <italic>LMC</italic> &#x0003D; <italic>C</italic><sub><italic>T</italic></sub>. Finally, note that in some cases the main contribution to the LMC is due to the sum of <italic>C</italic>[<bold>y</bold><sub><italic>k</italic></sub>] over a large range of values of <italic>k</italic> greater than a certain value <italic>k</italic><sub><italic>c</italic></sub> with large normalized-root-mean-square reconstruction errors NRMSRE(<italic>k</italic>) &#x0003D; <inline-formula><mml:math id="M16"><mml:msqrt><mml:mrow><mml:mn>1</mml:mn><mml:mo>-</mml:mo><mml:mi>C</mml:mi><mml:mrow><mml:mo>[</mml:mo><mml:mrow><mml:msub><mml:mrow><mml:mstyle mathvariant="bold"><mml:mtext>y</mml:mtext></mml:mstyle></mml:mrow><mml:mrow><mml:mi>k</mml:mi></mml:mrow></mml:msub></mml:mrow><mml:mo>]</mml:mo></mml:mrow></mml:mrow></mml:msqrt></mml:math></inline-formula>. This corresponds to a memory function <italic>m</italic>(<italic>k</italic>) &#x0003D; <italic>C</italic>[<bold>y</bold><sub><italic>k</italic></sub>] with a long tail. In these cases a high LMC can be obtained but the reconstruction error for <bold>y</bold><sub><italic>k</italic></sub> when <italic>k</italic> &#x0003E; <italic>k</italic><sub><italic>c</italic></sub> is large. This low quality memory capacity leads to poor performance for tasks requiring long memory, such as NARMA-10 task [<xref ref-type="bibr" rid="B10">10</xref>]. A memory capacity with good quality (quality memory capacity) can be calculated by summing only the capacities for <bold>y</bold><sub><italic>k</italic></sub> over <italic>k</italic> until they drop below a certain value <italic>q</italic>. If we consider that the error is small when NRMSRE(<italic>k</italic>) &#x0003C; 0.3, this corresponds to <italic>C</italic>[<bold>y</bold><sub><italic>k</italic></sub>] &#x0003E; 0.91. Then we consider a value <italic>q</italic> &#x0003D; 0.9 to obtain the quality memory capacity <italic>C</italic>[<bold>y</bold> ]<sup><italic>q</italic> &#x0003D; 0.9</sup>.</p>
<sec>
<title>3.1.1. Memory Capacities of One-Delay Reservoir Computers</title>
<p>First, we simulate a delay-based reservoir computer with a single delay line. We focus on the influence of the system response time on the computational capacity for different values of the mismatch &#x003B1; between the data injection and delay times. Until now two values of the mismatch have been used: &#x003B1; &#x0003D; 0 with &#x003B8; &#x0003D; 0.2<italic>T</italic> [<xref ref-type="bibr" rid="B7">7</xref>, <xref ref-type="bibr" rid="B12">12</xref>&#x02013;<xref ref-type="bibr" rid="B18">18</xref>], and &#x003B1; &#x0003D; 1 with &#x003B8; &#x0226B; <italic>T</italic> [<xref ref-type="bibr" rid="B8">8</xref>, <xref ref-type="bibr" rid="B15">15</xref>, <xref ref-type="bibr" rid="B19">19</xref>].</p>
<p>We first consider a linear system with <italic>f</italic>(<italic>z</italic>) &#x0003D; <italic>z</italic> in Equation (2). As stated before, the total computational capacity of this system can be obtained from the linear memory capacity, e.g., <inline-formula><mml:math id="M18"><mml:msubsup><mml:mrow><mml:mi>C</mml:mi></mml:mrow><mml:mrow><mml:mi>T</mml:mi></mml:mrow><mml:mrow><mml:mi>l</mml:mi></mml:mrow></mml:msubsup><mml:mo>=</mml:mo><mml:mi>L</mml:mi><mml:mi>M</mml:mi><mml:mi>C</mml:mi></mml:math></inline-formula>. <xref ref-type="fig" rid="F2">Figure 2</xref> shows the total computational capacity of the linear reservoir computer as a function of the node separation for two different values of the detuning between <italic>T</italic><sub><italic>p</italic></sub> and &#x003C4;: &#x003B1; &#x0003D; 0 and &#x003B1; &#x0003D; 1. For &#x003B1; &#x0003D; 1 (<xref ref-type="fig" rid="F2">Figure 2B</xref>), <inline-formula><mml:math id="M19"><mml:msubsup><mml:mrow><mml:mi>C</mml:mi></mml:mrow><mml:mrow><mml:mi>T</mml:mi></mml:mrow><mml:mrow><mml:mi>l</mml:mi></mml:mrow></mml:msubsup></mml:math></inline-formula> increases with &#x003B8;/<italic>T</italic> and the upper bound <italic>C</italic><sub><italic>T</italic></sub> &#x0003D; <italic>N</italic> &#x0003D; 97 is almost reached for &#x003B8;/<italic>T</italic> &#x0003D; 10. Similar behaviour is obtained for detuning values 1 &#x0003C; &#x003B1; &#x0003C; <italic>N</italic>. Then almost all the nodes are linearly independent for &#x003B8;/<italic>T</italic> &#x0003D; 10 and non-zero &#x003B1;. The quality memory <inline-formula><mml:math id="M20"><mml:msubsup><mml:mrow><mml:mi>C</mml:mi></mml:mrow><mml:mrow><mml:mi>T</mml:mi></mml:mrow><mml:mrow><mml:mi>l</mml:mi><mml:mrow><mml:mo stretchy="false">(</mml:mo><mml:mrow><mml:mi>q</mml:mi><mml:mo>=</mml:mo><mml:mn>0</mml:mn><mml:mo>.</mml:mo><mml:mn>9</mml:mn></mml:mrow><mml:mo stretchy="false">)</mml:mo></mml:mrow></mml:mrow></mml:msubsup><mml:mo>=</mml:mo><mml:mi>L</mml:mi><mml:mi>M</mml:mi><mml:msup><mml:mrow><mml:mi>C</mml:mi></mml:mrow><mml:mrow><mml:mi>q</mml:mi><mml:mo>=</mml:mo><mml:mn>0</mml:mn><mml:mo>.</mml:mo><mml:mn>9</mml:mn></mml:mrow></mml:msup></mml:math></inline-formula> of the linear delay-based reservoir computer also increases with &#x003B8;/<italic>T</italic> following the same behavior than <inline-formula><mml:math id="M21"><mml:msubsup><mml:mrow><mml:mi>C</mml:mi></mml:mrow><mml:mrow><mml:mi>T</mml:mi></mml:mrow><mml:mrow><mml:mi>l</mml:mi></mml:mrow></mml:msubsup></mml:math></inline-formula> for &#x003B1; &#x0003D; 1. However, when &#x003B8; &#x0003C; <italic>T</italic> a total capacity <inline-formula><mml:math id="M22"><mml:msubsup><mml:mrow><mml:mi>C</mml:mi></mml:mrow><mml:mrow><mml:mi>T</mml:mi></mml:mrow><mml:mrow><mml:mi>l</mml:mi></mml:mrow></mml:msubsup><mml:mo>&#x0003C;</mml:mo><mml:mn>50</mml:mn></mml:math></inline-formula> is obtained. Then a clear degradation of the capacity is observed with respect to its upper bound, given by <italic>N</italic> &#x0003D; 97, when the node separation is smaller than the response time of the non-linear node dynamics. In this case virtual nodes with an index difference smaller than <italic>T</italic> have similar states. Then reservoir diversity is reduced and the information processing capacity is degraded. When &#x003B8;/<italic>T</italic> increases the coupling between close (in time) virtual nodes decreases, and the capacity increases.</p>
<fig id="F2" position="float">
<label>Figure 2</label>
<caption><p>Computational capacity of the linear delay-based RC with one delay line as a function of &#x003B8;/<italic>T</italic> for <bold>(A)</bold> &#x003B1; &#x0003D; 0 and <bold>(B)</bold> &#x003B1; &#x0003D; 1. The solid line with blue circles is the total computational capacity (<inline-formula><mml:math id="M17"><mml:msubsup><mml:mrow><mml:mi>C</mml:mi></mml:mrow><mml:mrow><mml:mi>T</mml:mi></mml:mrow><mml:mrow><mml:mi>l</mml:mi></mml:mrow></mml:msubsup></mml:math></inline-formula>) and the dashed line with black points is the total quality computational capacity calculated for <italic>q</italic> &#x0003D; 0.9.</p></caption>
<graphic xlink:href="fphy-07-00210-g0002.tif"/>
</fig>
<p>In the special case of zero detuning (&#x003B1; &#x0003D; 0), the only coupling between the virtual nodes is through the system dynamics with non-zero response time. For &#x003B1; &#x0003D; 0, the total capacity of the linear delay-based reservoir computer has a maximum value <inline-formula><mml:math id="M23"><mml:msubsup><mml:mrow><mml:mi>C</mml:mi></mml:mrow><mml:mrow><mml:mi>T</mml:mi></mml:mrow><mml:mrow><mml:mi>l</mml:mi></mml:mrow></mml:msubsup><mml:mo>=</mml:mo><mml:mn>38</mml:mn></mml:math></inline-formula> at &#x003B8;/<italic>T</italic> &#x0007E; 1.2 (see <xref ref-type="fig" rid="F2">Figure 2A</xref>). In this case a clear degradation of the capacity is observed for any value of &#x003B8;/<italic>T</italic>. The maximum is due to the trade-off between the fading of the coupling through the system dynamics for low sampling output rates and the very similar responses to different inputs for small &#x003B8;. Furthermore, for &#x003B1; &#x0003D; 0, the quality memory capacity decreases with &#x003B8;/<italic>T</italic> and the maximum <inline-formula><mml:math id="M24"><mml:msubsup><mml:mrow><mml:mi>C</mml:mi></mml:mrow><mml:mrow><mml:mi>T</mml:mi></mml:mrow><mml:mrow><mml:mi>l</mml:mi><mml:mrow><mml:mo stretchy="false">(</mml:mo><mml:mrow><mml:mi>q</mml:mi><mml:mo>=</mml:mo><mml:mn>0</mml:mn><mml:mo>.</mml:mo><mml:mn>9</mml:mn></mml:mrow><mml:mo stretchy="false">)</mml:mo></mml:mrow></mml:mrow></mml:msubsup></mml:math></inline-formula> is obtained at &#x003B8;/<italic>T</italic> &#x0003D; 0.2. For low inertia, &#x003B8;/<italic>T</italic> &#x0003D; 4, we obtain a normalized-root-mean-square reconstruction error NRMSRE(<italic>k</italic>) &#x0003E; 0.6 when <italic>k</italic> &#x0003E; 2. For &#x003B8;/<italic>T</italic> = 1 a NRMSRE(<italic>k</italic>) &#x0003E; 0.3 when <italic>k</italic> &#x0003E; 12 is obtained.</p>
<p>We consider now a non-linear delay-based reservoir computer with an activation function given by Equation (7) and a low-to-moderate non-linearity (&#x003B3; &#x0003D; 0.1). In this case, the capacity <italic>C</italic><sub><italic>s</italic></sub> has a behaviour as a function of &#x003B8; similar to that of the total capacity of the linear case <inline-formula><mml:math id="M25"><mml:msubsup><mml:mrow><mml:mi>C</mml:mi></mml:mrow><mml:mrow><mml:mi>T</mml:mi></mml:mrow><mml:mrow><mml:mi>l</mml:mi></mml:mrow></mml:msubsup></mml:math></inline-formula> (see <xref ref-type="fig" rid="F3">Figure 3</xref>). For &#x003B1; &#x0003D; 1, <italic>C</italic><sub><italic>s</italic></sub> increases with &#x003B8;/<italic>T</italic>, and a value of <italic>C</italic><sub><italic>s</italic></sub> &#x0003D; 93 is obtained at &#x003B8;/<italic>T</italic> &#x0003D; 4. If all the capacities would be considered for &#x003B1; &#x0003D; 1, <italic>C</italic><sub><italic>T</italic></sub> &#x0007E; <italic>N</italic>. The increase in <italic>C</italic><sub><italic>s</italic></sub> with &#x003B8;/<italic>T</italic> is mainly due to the <italic>XMC</italic> and to the <italic>LMC</italic>. When &#x003B8;/<italic>T</italic> &#x0003C; 1 a capacity <italic>C</italic><sub><italic>s</italic></sub> &#x0003C; 75 is obtained. However, this degradation in <italic>C</italic><sub><italic>s</italic></sub> is smaller than in the linear case. It is worth mentioning that for &#x003B1; &#x0003D; 1, <italic>C</italic><sub><italic>s</italic></sub> is greater than the total capacity of the linear case <inline-formula><mml:math id="M26"><mml:msubsup><mml:mrow><mml:mi>C</mml:mi></mml:mrow><mml:mrow><mml:mi>T</mml:mi></mml:mrow><mml:mrow><mml:mi>l</mml:mi></mml:mrow></mml:msubsup></mml:math></inline-formula>. Then we have <inline-formula><mml:math id="M27"><mml:msubsup><mml:mrow><mml:mi>C</mml:mi></mml:mrow><mml:mrow><mml:mi>T</mml:mi></mml:mrow><mml:mrow><mml:mi>l</mml:mi></mml:mrow></mml:msubsup><mml:mo>&#x0003C;</mml:mo><mml:msub><mml:mrow><mml:mi>C</mml:mi></mml:mrow><mml:mrow><mml:mi>s</mml:mi></mml:mrow></mml:msub><mml:mo>&#x02264;</mml:mo><mml:msubsup><mml:mrow><mml:mi>C</mml:mi></mml:mrow><mml:mrow><mml:mi>T</mml:mi></mml:mrow><mml:mrow><mml:mi>n</mml:mi><mml:mi>l</mml:mi></mml:mrow></mml:msubsup></mml:math></inline-formula>, where <inline-formula><mml:math id="M28"><mml:msubsup><mml:mrow><mml:mi>C</mml:mi></mml:mrow><mml:mrow><mml:mi>T</mml:mi></mml:mrow><mml:mrow><mml:mi>n</mml:mi><mml:mi>l</mml:mi></mml:mrow></mml:msubsup></mml:math></inline-formula> is the total capacity of the non-linear system. This is due to the fact that non-linearity increases the number of linearly independent virtual node states, since correlations between virtual nodes are smaller for non-linear delay-based reservoir computer. In the case without mismatch (&#x003B1; &#x0003D; 0) the capacity <italic>C</italic><sub><italic>s</italic></sub> of the non-linear reservoir computer (see <xref ref-type="fig" rid="F3">Figure 3A</xref>) has a maximum as in the linear case at &#x003B8;/<italic>T</italic> &#x0007E; 1.2. The degradation of <italic>C</italic><sub><italic>s</italic></sub> is smaller than that of <inline-formula><mml:math id="M29"><mml:msubsup><mml:mrow><mml:mi>C</mml:mi></mml:mrow><mml:mrow><mml:mi>T</mml:mi></mml:mrow><mml:mrow><mml:mi>l</mml:mi></mml:mrow></mml:msubsup></mml:math></inline-formula> in the linear case.</p>
<fig id="F3" position="float">
<label>Figure 3</label>
<caption><p>Memory capacities of the non-linear delay-based RC with one delay line as a function of &#x003B8;/<italic>T</italic> for <bold>(A)</bold> &#x003B1; &#x0003D; 0 and <bold>(B)</bold> &#x003B1; &#x0003D; 1 when &#x003B3; &#x0003D; 0.1. The blue stars, red circles, green crosses, pink diamonds correspond to the LMC, QMC, CMC, and XMC. The black solid line is the <italic>C</italic><sub><italic>s</italic></sub>.</p></caption>
<graphic xlink:href="fphy-07-00210-g0003.tif"/>
</fig>
<p>We have shown that the computational capacity is degraded when the sampling output rate is higher than the inverse of the system response time. However, the information processing capacity of delay-based reservoir computers depends not only on output sampling rate (i.e., the separation between the virtual nodes) but also on the detuning between <italic>T</italic><sub><italic>p</italic></sub> and &#x003C4;, i.e., &#x003B1;. To study this dependency, we calculate the memory capacities as a function of &#x003B1; for a non-linear delay-based reservoir computer with two different response times: an instantaneous response to the input <italic>T</italic> &#x0003D; 0 (<xref ref-type="fig" rid="F4">Figures 4C,D</xref>) and <italic>T</italic> &#x0003D; &#x003B8;/0.2 (<xref ref-type="fig" rid="F4">Figures 4A,B</xref>). This node separation &#x003B8; &#x0003D; 0.2<italic>T</italic> is the one used in most of the reservoirs with connections through system dynamics [<xref ref-type="bibr" rid="B7">7</xref>, <xref ref-type="bibr" rid="B12">12</xref>&#x02013;<xref ref-type="bibr" rid="B18">18</xref>]. The capacities for <italic>T</italic> &#x0003D; 0 correspond to a node separation much larger than <italic>T</italic>. When &#x003B8;/<italic>T</italic> &#x0226B; 1 the nodes response to an input reach the steady state after a time &#x003B8;. Then the reservoir state is given by Equation (2) for <italic>T</italic> &#x0003D; 0. As a consequence, when &#x003B8;/<italic>T</italic> &#x0226B; 1 the computational capacity tends to the value obtained for <italic>T</italic> &#x0003D; 0. For a mismatch &#x003B1; &#x0003D; 1 this limit is reached for &#x003B8;/<italic>T</italic> &#x0003E; 4 (see <xref ref-type="fig" rid="F3">Figure 3B</xref>). Two values of &#x003B3; &#x0003D; 0.1 and &#x003B3; &#x0003D; 1 that correspond to low-to-moderate and strong non-linearity, respectively are considered. We also calculate the total capacity as a function of &#x003B1; for a linear reservoir computer with &#x003B8; &#x0003D; 0.2<italic>T</italic> (<xref ref-type="fig" rid="F4">Figure 4B</xref>).</p>
<fig id="F4" position="float">
<label>Figure 4</label>
<caption><p>Memory capacities of the one delay-based RC as a function of &#x003B1;. Left panels <bold>(A,C)</bold>: &#x003B3; &#x0003D; 1. Right panels <bold>(B,D)</bold>: &#x003B3; &#x0003D; 0.1. Top panels <bold>(A,B)</bold>: <italic>T</italic> &#x0003D; &#x003B8;/0.2. Bottom panels <bold>(C,D)</bold> <italic>T</italic> &#x0003D; 0. The blue stars, red circles, green crosses, pink diamonds correspond to the LMC, QMC, CMC, XMC, respectively. The solid black line is the <italic>C</italic><sub><italic>s</italic></sub>. The dotted black line in <bold>(B)</bold> is the <inline-formula><mml:math id="M30"><mml:msubsup><mml:mrow><mml:mi>C</mml:mi></mml:mrow><mml:mrow><mml:mi>T</mml:mi></mml:mrow><mml:mrow><mml:mi>l</mml:mi></mml:mrow></mml:msubsup></mml:math></inline-formula>.</p></caption>
<graphic xlink:href="fphy-07-00210-g0004.tif"/>
</fig>
<p>The virtual states of delay-based systems with an instantaneous response to the input are given by the map of Equation (3). When <italic>N</italic> and &#x003B1; are coprimes, we have for 0 &#x0003C; &#x003B1; &#x0003C; <italic>N</italic> a total capacity <italic>C</italic><sub><italic>T</italic></sub> &#x02248; <italic>N</italic>. Thus, increasing &#x003B1; in the case of <italic>T</italic> &#x0003D; 0 does not increase the total capacity; it only changes the relative contribution of the different capacities to <inline-formula><mml:math id="M31"><mml:msubsup><mml:mrow><mml:mi>C</mml:mi></mml:mrow><mml:mrow><mml:mi>T</mml:mi></mml:mrow><mml:mrow><mml:mi>n</mml:mi><mml:mi>l</mml:mi></mml:mrow></mml:msubsup></mml:math></inline-formula>. This is clearly shown in <xref ref-type="fig" rid="F4">Figure 4D</xref> where a low-to-moderate non-linearity (&#x003B3; &#x0003D; 0.1) is considered. Here, the non-linear memory capacities of degree greater than two are zero (i.e., CMC), and <italic>C</italic><sub><italic>s</italic></sub> &#x0007E; 95 for 0 &#x0003C; &#x003B1; &#x0003C; 90. This value is very close to the upper bound for the capacity <italic>C</italic><sub><italic>T</italic></sub> &#x0003D; <italic>N</italic> &#x0003D; 97. Since <italic>C</italic><sub><italic>T</italic></sub> is limited by <italic>N</italic>, there is a trade-off between the linear and non-linear capacities. Then the increase in the <italic>LMC</italic> with &#x003B1; is compensated by a decrease of the <italic>XMC</italic> in <xref ref-type="fig" rid="F4">Figure 4D</xref>. In the case of strong non-linearity (&#x003B3; &#x0003D; 1), <xref ref-type="fig" rid="F4">Figure 4C</xref> shows that <italic>C</italic><sub><italic>s</italic></sub> is not close to the upper bound for the capacity <italic>C</italic><sub><italic>T</italic></sub> &#x0003D; <italic>N</italic> &#x0003D; 97. Then there is a significant contribution to <inline-formula><mml:math id="M32"><mml:msubsup><mml:mrow><mml:mi>C</mml:mi></mml:mrow><mml:mrow><mml:mi>T</mml:mi></mml:mrow><mml:mrow><mml:mi>n</mml:mi><mml:mi>l</mml:mi></mml:mrow></mml:msubsup></mml:math></inline-formula> of capacities with a non-linear degree greater than the ones considered in <italic>C</italic><sub><italic>s</italic></sub>. An increase in <italic>C</italic><sub><italic>s</italic></sub> with &#x003B1; is obtained. This increase is mainly due to <italic>LMC</italic> and <italic>XMC</italic>. It only indicates that the contribution to <inline-formula><mml:math id="M33"><mml:msubsup><mml:mrow><mml:mi>C</mml:mi></mml:mrow><mml:mrow><mml:mi>T</mml:mi></mml:mrow><mml:mrow><mml:mi>n</mml:mi><mml:mi>l</mml:mi></mml:mrow></mml:msubsup></mml:math></inline-formula> of the capacities with a lower non-linear degree considered in <italic>C</italic><sub><italic>s</italic></sub> increases.</p>
<p>Now we analyze the capacity dependence on &#x003B1; when &#x003B8;/<italic>T</italic> &#x0003D; 0.2. We consider integer values of &#x003B1;. Similar results are obtained when &#x003B1; is not an exact integer. We first consider the linear system. In this case the total capacity <inline-formula><mml:math id="M34"><mml:msubsup><mml:mrow><mml:mi>C</mml:mi></mml:mrow><mml:mrow><mml:mi>T</mml:mi></mml:mrow><mml:mrow><mml:mi>l</mml:mi></mml:mrow></mml:msubsup></mml:math></inline-formula> is given by the <italic>LMC</italic>. As seen in <xref ref-type="fig" rid="F2">Figure 2A</xref> the capacity is degraded when &#x003B8; &#x0003C; <italic>T</italic> due to the similar evolution in time of close (in time) virtual nodes connected through non-linear node dynamics. <xref ref-type="fig" rid="F4">Figure 4B</xref> shows that <inline-formula><mml:math id="M35"><mml:msubsup><mml:mrow><mml:mi>C</mml:mi></mml:mrow><mml:mrow><mml:mi>T</mml:mi></mml:mrow><mml:mrow><mml:mi>l</mml:mi></mml:mrow></mml:msubsup></mml:math></inline-formula> increases with &#x003B1;. A significant increase of nearly 50% is obtained for the capacity when the mismatch is large. This is due to an increase in reservoir diversity. When the mismatch &#x003B1; is increased, virtual nodes are connected through feedback to nodes that are not connected through system dynamics. This improves reservoir diversity, and a larger capacity can be achieved.</p>
<p>In the non-linear case with &#x003B8;/<italic>T</italic> &#x0003D; 0.2, <xref ref-type="fig" rid="F4">Figures 4A,B</xref> show that regardless of the non-linearity, <italic>C</italic><sub><italic>s</italic></sub> increases with &#x003B1;. This increase can not be attributed only to a change in the contribution of linear and non-linear capacities to the total capacity <inline-formula><mml:math id="M36"><mml:msubsup><mml:mrow><mml:mi>C</mml:mi></mml:mrow><mml:mrow><mml:mi>T</mml:mi></mml:mrow><mml:mrow><mml:mi>n</mml:mi><mml:mi>l</mml:mi></mml:mrow></mml:msubsup></mml:math></inline-formula>. As seen for the linear case, when &#x003B8;/<italic>T</italic> &#x0003D; 0.2 the total capacity <inline-formula><mml:math id="M37"><mml:msubsup><mml:mrow><mml:mi>C</mml:mi></mml:mrow><mml:mrow><mml:mi>T</mml:mi></mml:mrow><mml:mrow><mml:mi>l</mml:mi></mml:mrow></mml:msubsup></mml:math></inline-formula> increases with &#x003B1; due to an increase in reservoir diversity. This should also lead in the non-linear case to an increase in the total capacity <inline-formula><mml:math id="M38"><mml:msubsup><mml:mrow><mml:mi>C</mml:mi></mml:mrow><mml:mrow><mml:mi>T</mml:mi></mml:mrow><mml:mrow><mml:mi>n</mml:mi><mml:mi>l</mml:mi></mml:mrow></mml:msubsup></mml:math></inline-formula> with &#x003B1;. It is worth mentioning that in the case of <italic>T</italic> &#x0003D; &#x003B8;/0.2 we obtain a similar <italic>C</italic><sub><italic>s</italic></sub> for low-to-moderate (see <xref ref-type="fig" rid="F4">Figure 4B</xref>) and strong (<xref ref-type="fig" rid="F4">Figure 4A</xref>) non-linearity. However, the relative contribution of the linear memory capacity is higher for low non-linearity. Finally, note that regardless of the non-linearity and <italic>T</italic>, higher order capacities such as QMC and CMC remain almost constant with &#x003B1; and the change of <italic>C</italic><sub><italic>s</italic></sub> is due to LMC and XMC.</p></sec>
<sec>
<title>3.1.2. Memory Capacities of Two-Delay Reservoir Computers</title>
<p>We have shown that the computational capacity is boosted for small values of &#x003B8;/<italic>T</italic> when the delay time of the non-linear node is greater than the data injection time. This mismatch between &#x003C4; and <italic>T</italic><sub><italic>p</italic></sub> allows higher processing speeds of delay-based reservoir computers without performance degradation. This is due to the increase in reservoir diversity. To further increase reservoir diversity in the case of <italic>T</italic> &#x0003D; &#x003B8;/0.2, we explore the effect of adding a extra feedback line to the non-linear node. <xref ref-type="fig" rid="F5">Figure 5</xref> shows the <italic>C</italic><sub><italic>s</italic></sub> of the two-delay reservoir computer vs. the misalignment of the second delay when &#x003B3; &#x0003D; 0.1. The mismatch of the first delay is fixed at &#x003B1;<sub>1</sub> &#x0003D; 73 (<xref ref-type="fig" rid="F5">Figure 5</xref>, left) and &#x003B1; &#x0003D; 1 (<xref ref-type="fig" rid="F5">Figure 5</xref>, right). In both cases the maximum of <italic>C</italic><sub><italic>s</italic></sub> reached for the two-delay system is <italic>C</italic><sub><italic>s</italic></sub> &#x0007E; 61. This value is obtained in the two cases, &#x003B1;<sub>1</sub> &#x0003D; 1 and &#x003B1;<sub>1</sub> &#x0003D; 73, for &#x003B1;<sub>2</sub> &#x0007E; 70 when &#x003B2;<sub>2</sub> &#x0003D; 0.75 and just in the case of &#x003B1;<sub>1</sub> &#x0003D; 73 also for &#x003B1;<sub>2</sub> &#x0007E; 82 and &#x003B2;<sub>2</sub> &#x0003D; &#x003B2;<sub>1</sub> &#x0003D; 0.4. The maximum <italic>C</italic><sub><italic>s</italic></sub> obtained for the two-delay system is slightly higher than the one reached for its one-delay counterpart. In the one-delay system the maximum capacity is <italic>C</italic><sub><italic>s</italic></sub> &#x0007E; 57 that is obtained for &#x003B1; &#x0007E; 80 (see <xref ref-type="fig" rid="F4">Figure 4B</xref>). Therefore, the calculated information processing capacity for high sampling output rates can be further increased by using an extra feedback line and delay times greater than the information processing time. However, the second delay does not significantly improve the computational capacity of the one-delay system. Moreover, when the first delay mismatch is fixed near its optimal value for the one-delay system (&#x003B1; &#x0007E; 80), the effect of the second delay feedback strength or misalignment is small [see <xref ref-type="fig" rid="F5">Figure 5</xref> (right)]. However, when the first delay mismatch is not close to its optimal value for the one-delay system, the maximum <italic>C</italic><sub><italic>s</italic></sub> reached for the one-delay system is outperformed by adding a second delay with a high strength (&#x003B2;<sub>2</sub> &#x0003D; 0.75) and a mismatch 10 &#x0003C; &#x003B1;<sub>2</sub> &#x0003C; 90 [see <xref ref-type="fig" rid="F5">Figure 5</xref> (left)].</p>
<fig id="F5" position="float">
<label>Figure 5</label>
<caption><p><italic>C</italic><sub><italic>s</italic></sub> of the two-delay-based RC as function of &#x003B1;<sub>2</sub>. <bold>Left</bold>: &#x003B1;<sub>1</sub> &#x0003D; 1. <bold>Right</bold>: &#x003B1;<sub>1</sub> &#x0003D; 73. The solid black line is the value of <italic>C</italic><sub><italic>s</italic></sub> for the one-delay case with &#x003B1; &#x0003D; &#x003B1;<sub>1</sub>. Red circles, green diamonds and blue starts correspond to the <italic>C</italic><sub><italic>s</italic></sub> with two delays and a &#x003B2;<sub>2</sub> of 0.05, 0.4, and 0.75, respectively. These results are obtained for <italic>T</italic> &#x0003D; &#x003B8;/0.2 and &#x003B3; &#x0003D; 0.1.</p></caption>
<graphic xlink:href="fphy-07-00210-g0005.tif"/>
</fig>
<p>The contributions of the individual memory capacities to <italic>C</italic><sub><italic>s</italic></sub> for the two-delay system are depicted in <xref ref-type="fig" rid="F6">Figures 6</xref>, <xref ref-type="fig" rid="F7">7</xref> for &#x003B1;<sub>1</sub> &#x0003D; 1 and &#x003B1;<sub>1</sub> &#x0003D; 73, respectively. <xref ref-type="fig" rid="F6">Figure 6</xref> shows that the increase in <italic>C</italic><sub><italic>s</italic></sub> obtained for &#x003B1;<sub>1</sub> &#x0003D; 1 is mainly due to the increase in LMC and QMC. It is interesting that in the case of &#x003B1;<sub>2</sub> &#x0003D; 73, the same <italic>C</italic><sub><italic>s</italic></sub> &#x0007E; 61 can be obtained with different relative contributions of the memory capacities to <italic>C</italic><sub><italic>s</italic></sub>. The case of &#x003B1;<sub>2</sub> &#x0007E; 70 and &#x003B2;<sub>2</sub> &#x0003D; 0.75 yields to a higher LMC and a lower XMC than in the one-delay system. The case of &#x003B1;<sub>2</sub> &#x0007E; 82 and &#x003B2;<sub>2</sub> &#x0003D; 0.4 gives the <italic>C</italic><sub><italic>s</italic></sub> &#x0007E; 61 thanks mainly to the increase in the XMC.</p>
<fig id="F6" position="float">
<label>Figure 6</label>
<caption><p>Memory capacities for the two-delay RC as function of &#x003B1;<sub>2</sub> for a fixed &#x003B1;<sub>1</sub> &#x0003D; 1, <italic>T</italic> &#x0003D; &#x003B8;/0.2 and &#x003B3; &#x0003D; 0.1. The red circles, green diamonds and blue stars correspond to &#x003B2;<sub>2</sub> equal to 0.05, 0.4, and 0.75, respectively. The solid black line is for &#x003B2;<sub>2</sub> &#x0003D; 0 and corresponds to the one-delay system with &#x003B1; &#x0003D; 1 and &#x003B2; &#x0003D; 0.8.</p></caption>
<graphic xlink:href="fphy-07-00210-g0006.tif"/>
</fig>
<fig id="F7" position="float">
<label>Figure 7</label>
<caption><p>Memory capacities for the two-delay-based RC as function of &#x003B1;<sub>2</sub> for a fixed &#x003B1;<sub>1</sub> &#x0003D; 73, <italic>T</italic> &#x0003D; &#x003B8;/0.2 and &#x003B3; &#x0003D; 0.1. The red circles, green diamonds and blue stars correspond to &#x003B2;<sub>2</sub> equal to 0.05, 0.4, and 0.75, respectively. The solid black line is for &#x003B2;<sub>2</sub> &#x0003D; 0 and corresponds to the one-delay case with &#x003B1; &#x0003D; 1 and &#x003B2; &#x0003D; 0.8.</p></caption>
<graphic xlink:href="fphy-07-00210-g0007.tif"/>
</fig>
</sec></sec>
<sec>
<title>3.2. Delay-Based Reservoir Computer Performance</title>
<p>Finally we study the effect of increasing the mismatch &#x003B1; on the performance of a delay-based reservoir computer for two different response times of the non-linear node dynamics: <italic>T</italic> &#x0003D; 0 and <italic>T</italic> &#x0003D; 0.2&#x003B8;. Two tasks are considered: the NARMA-10 task and the equalization of a wireless communication channel. These two tasks are benchmarking tasks used to assess the performance of RC [<xref ref-type="bibr" rid="B1">1</xref>, <xref ref-type="bibr" rid="B10">10</xref>].</p>
<p>The NARMA-10 task consists in predicting the output of an auto-regressive moving average from the input <italic>u</italic>(<italic>t</italic>). The output <italic>y</italic>(<italic>t</italic>&#x0002B;1) is given by:</p>
<disp-formula id="E9"><label>(8)</label><mml:math id="M39"><mml:mtable class="eqnarray" columnalign="right center left"><mml:mtr><mml:mtd><mml:mrow><mml:mi>y</mml:mi><mml:mo stretchy='false'>(</mml:mo><mml:mi>t</mml:mi><mml:mo>+</mml:mo><mml:mn>1</mml:mn><mml:mo stretchy='false'>)</mml:mo><mml:mo>=</mml:mo><mml:mn>0.3</mml:mn><mml:mi>y</mml:mi><mml:mo stretchy='false'>(</mml:mo><mml:mi>t</mml:mi><mml:mo stretchy='false'>)</mml:mo><mml:mo>+</mml:mo><mml:mn>0.05</mml:mn><mml:mi>y</mml:mi><mml:mo stretchy='false'>(</mml:mo><mml:mi>t</mml:mi><mml:mo stretchy='false'>)</mml:mo><mml:mrow><mml:mstyle displaystyle='true'><mml:munderover><mml:mo>&#x2211;</mml:mo><mml:mrow><mml:mi>i</mml:mi><mml:mo>=</mml:mo><mml:mn>0</mml:mn></mml:mrow><mml:mn>9</mml:mn> </mml:munderover> <mml:mrow><mml:mi>y</mml:mi><mml:mo stretchy='false'>(</mml:mo><mml:mi>k</mml:mi><mml:mo>&#x2212;</mml:mo><mml:mi>i</mml:mi><mml:mo stretchy='false'>)</mml:mo> </mml:mrow></mml:mstyle></mml:mrow><mml:mo>+</mml:mo><mml:mn>1.5</mml:mn><mml:mi>u</mml:mi><mml:mo stretchy='false'>(</mml:mo><mml:mi>t</mml:mi><mml:mo>&#x02212;</mml:mo><mml:mn>9</mml:mn><mml:mo stretchy='false'>)</mml:mo><mml:mi>u</mml:mi><mml:mo stretchy='false'>(</mml:mo><mml:mi>t</mml:mi><mml:mo stretchy='false'>)</mml:mo><mml:mo stretchy='false'>)</mml:mo><mml:mo>+</mml:mo><mml:mn>0.1</mml:mn></mml:mrow></mml:mtd></mml:mtr></mml:mtable></mml:math></disp-formula>
<p>The input <italic>u</italic>(<italic>t</italic>) is independently and identically drawn from the uniform distribution in [0, 0.5]. Solving the NARMA-10 task requires both memory and non-linearity. <xref ref-type="fig" rid="F8">Figure 8</xref> (left) shows the normalized-root-mean-square error (NRMSE) of the NARMA-10 task as a function of &#x003B1; for &#x003B3; &#x0003D; 0.1. We consider a small value of &#x003B3; &#x0003D; 0.1 because a long memory is required to obtain a good performance for NARMA-10 task. Regardless the response time (<italic>T</italic> &#x0003D; 0 or <italic>T</italic> &#x0003D; &#x003B8;/0.2), the NRMSE decreases when the processing and delay times are mismatched (&#x003B1; &#x0003E; 0). However, for <italic>T</italic> &#x0003D; 0 the NRMSE is almost the same for a wide variety of values of &#x003B1;, and a mismatch &#x003B1; &#x0003D; 1 is enough to obtain a NRMSE &#x0003D; 0.31 close to the absolute minimum (NRMSE &#x0003D; 0.28 for &#x003B1; &#x0003D; 78). When the response time of the non-linear node is larger than node separation (<italic>T</italic> &#x0003D; &#x003B8;/0.2), the NRMSE decreases from a NRMSE &#x02248; 0.46 at &#x003B1; &#x0003D; (0, 1) to a NRMSE &#x0003D; 0.34 at &#x003B1; &#x0007E; 72. This is due to the long memory required to obtain a good performance for NARMA-10 task. In the case of <italic>T</italic> &#x0003D; &#x003B8;/0.2, the required LMC is not reached until &#x003B1; &#x0007E; 72 (see <xref ref-type="fig" rid="F4">Figure 4B</xref>). Our results show that a similar performance can be obtained for small and large values of <italic>T</italic>/&#x003B8; thanks to the mismatch &#x003B1;. Therefore, increasing &#x003B1; allows a faster processing information (higher sampling output rate) without causing system performance degradation.</p>
<fig id="F8" position="float">
<label>Figure 8</label>
<caption><p>Performance of the non-linear one delay-based RC for two tasks as function of &#x003B1;. <bold>Left</bold>: NARMA-10 for &#x003B3; &#x0003D; 0.1. <bold>Right</bold>: Equalization with <italic>SNR</italic> &#x0003D; 20 dB and &#x003B3; &#x0003D; 1. The blue stars correspond to the case of <italic>T</italic> &#x0003D; 0 and the red circles to the case of <italic>T</italic> &#x0003D; &#x003B8;/0.2.</p></caption>
<graphic xlink:href="fphy-07-00210-g0008.tif"/>
</fig>
<p>The equalization of a wireless communication channel consists in reconstructing the input signal <italic>s</italic>(<italic>i</italic>) from the output sequence of the channel <italic>u</italic>(<italic>i</italic>) [<xref ref-type="bibr" rid="B1">1</xref>]. The input to the channel is a random sequence of values <italic>s</italic>(<italic>i</italic>) taken in {&#x02212;3, &#x02212;1, 1, 3}. The input <italic>s</italic>(<italic>i</italic>) first goes through a linear channel yielding:</p>
<disp-formula id="E10"><mml:math id="M40"><mml:mtable class="multline-star"><mml:mtr><mml:mtd><mml:mi>q</mml:mi><mml:mrow><mml:mo stretchy="false">(</mml:mo><mml:mrow><mml:mi>i</mml:mi></mml:mrow><mml:mo stretchy="false">)</mml:mo></mml:mrow><mml:mo>=</mml:mo><mml:mn>0</mml:mn><mml:mo>.</mml:mo><mml:mn>08</mml:mn><mml:mi>s</mml:mi><mml:mrow><mml:mo stretchy="false">(</mml:mo><mml:mrow><mml:mi>i</mml:mi><mml:mo>&#x0002B;</mml:mo><mml:mn>2</mml:mn></mml:mrow><mml:mo stretchy="false">)</mml:mo></mml:mrow><mml:mo>-</mml:mo><mml:mn>0</mml:mn><mml:mo>.</mml:mo><mml:mn>12</mml:mn><mml:mi>s</mml:mi><mml:mrow><mml:mo stretchy="false">(</mml:mo><mml:mrow><mml:mi>i</mml:mi><mml:mo>&#x0002B;</mml:mo><mml:mn>1</mml:mn></mml:mrow><mml:mo stretchy="false">)</mml:mo></mml:mrow><mml:mo>&#x0002B;</mml:mo><mml:mi>s</mml:mi><mml:mrow><mml:mo stretchy="false">(</mml:mo><mml:mrow><mml:mi>i</mml:mi></mml:mrow><mml:mo stretchy="false">)</mml:mo></mml:mrow><mml:mo>&#x0002B;</mml:mo><mml:mn>0</mml:mn><mml:mo>.</mml:mo><mml:mn>18</mml:mn><mml:mi>s</mml:mi><mml:mrow><mml:mo stretchy="false">(</mml:mo><mml:mrow><mml:mi>i</mml:mi><mml:mo>-</mml:mo><mml:mn>1</mml:mn></mml:mrow><mml:mo stretchy="false">)</mml:mo></mml:mrow><mml:mo>-</mml:mo><mml:mn>0</mml:mn><mml:mo>.</mml:mo><mml:mn>1</mml:mn><mml:mi>s</mml:mi><mml:mrow><mml:mo stretchy="false">(</mml:mo><mml:mrow><mml:mi>i</mml:mi><mml:mo>-</mml:mo><mml:mn>2</mml:mn></mml:mrow><mml:mo stretchy="false">)</mml:mo></mml:mrow></mml:mtd></mml:mtr><mml:mtr><mml:mtd><mml:mo>&#x0002B;</mml:mo><mml:mn>0</mml:mn><mml:mo>.</mml:mo><mml:mn>091</mml:mn><mml:mi>s</mml:mi><mml:mrow><mml:mo stretchy="false">(</mml:mo><mml:mrow><mml:mi>i</mml:mi><mml:mo>-</mml:mo><mml:mn>3</mml:mn></mml:mrow><mml:mo stretchy="false">)</mml:mo></mml:mrow><mml:mo>-</mml:mo><mml:mn>0</mml:mn><mml:mo>.</mml:mo><mml:mn>05</mml:mn><mml:mi>s</mml:mi><mml:mrow><mml:mo stretchy="false">(</mml:mo><mml:mrow><mml:mi>i</mml:mi><mml:mo>-</mml:mo><mml:mn>4</mml:mn></mml:mrow><mml:mo stretchy="false">)</mml:mo></mml:mrow><mml:mo>&#x0002B;</mml:mo><mml:mn>0</mml:mn><mml:mo>.</mml:mo><mml:mn>04</mml:mn><mml:mi>s</mml:mi><mml:mrow><mml:mo stretchy="false">(</mml:mo><mml:mrow><mml:mi>i</mml:mi><mml:mo>-</mml:mo><mml:mn>5</mml:mn></mml:mrow><mml:mo stretchy="false">)</mml:mo></mml:mrow><mml:mo>&#x0002B;</mml:mo><mml:mn>0</mml:mn><mml:mo>.</mml:mo><mml:mn>03</mml:mn><mml:mi>s</mml:mi><mml:mrow><mml:mo stretchy="false">(</mml:mo><mml:mrow><mml:mi>i</mml:mi><mml:mo>-</mml:mo><mml:mn>6</mml:mn></mml:mrow><mml:mo stretchy="false">)</mml:mo></mml:mrow><mml:mo>&#x0002B;</mml:mo><mml:mn>0</mml:mn><mml:mo>.</mml:mo><mml:mn>01</mml:mn><mml:mi>s</mml:mi><mml:mrow><mml:mo stretchy="false">(</mml:mo><mml:mrow><mml:mi>i</mml:mi><mml:mo>-</mml:mo><mml:mn>7</mml:mn></mml:mrow><mml:mo stretchy="false">)</mml:mo></mml:mrow></mml:mtd></mml:mtr></mml:mtable></mml:math></disp-formula>
<p>It then goes through a noisy non-linear channel:</p>
<disp-formula id="E11"><label>(9)</label><mml:math id="M41"><mml:mtable class="eqnarray" columnalign="right center left"><mml:mtr><mml:mtd><mml:mi>u</mml:mi><mml:mrow><mml:mo stretchy="false">(</mml:mo><mml:mrow><mml:mi>i</mml:mi></mml:mrow><mml:mo stretchy="false">)</mml:mo></mml:mrow><mml:mo>=</mml:mo><mml:mi>q</mml:mi><mml:mrow><mml:mo stretchy="false">(</mml:mo><mml:mrow><mml:mi>i</mml:mi></mml:mrow><mml:mo stretchy="false">)</mml:mo></mml:mrow><mml:mo>&#x0002B;</mml:mo><mml:mn>0</mml:mn><mml:mo>.</mml:mo><mml:mn>036</mml:mn><mml:mi>q</mml:mi><mml:msup><mml:mrow><mml:mrow><mml:mo stretchy="false">(</mml:mo><mml:mrow><mml:mi>i</mml:mi></mml:mrow><mml:mo stretchy="false">)</mml:mo></mml:mrow></mml:mrow><mml:mrow><mml:mn>2</mml:mn></mml:mrow></mml:msup><mml:mo>-</mml:mo><mml:mn>0</mml:mn><mml:mo>.</mml:mo><mml:mn>011</mml:mn><mml:mi>q</mml:mi><mml:msup><mml:mrow><mml:mrow><mml:mo stretchy="false">(</mml:mo><mml:mrow><mml:mi>i</mml:mi></mml:mrow><mml:mo stretchy="false">)</mml:mo></mml:mrow></mml:mrow><mml:mrow><mml:mn>3</mml:mn></mml:mrow></mml:msup><mml:mo>&#x0002B;</mml:mo><mml:mi>v</mml:mi><mml:mrow><mml:mo stretchy="false">(</mml:mo><mml:mrow><mml:mi>i</mml:mi></mml:mrow><mml:mo stretchy="false">)</mml:mo></mml:mrow><mml:mo>,</mml:mo></mml:mtd></mml:mtr></mml:mtable></mml:math></disp-formula>
<p>where <italic>v</italic>(<italic>i</italic>) is a Gaussian noise with zero mean adjusted in power to give a signal-to-noise ratio (SNR) of 20 dB. The performance is measured using the Symbol Error Rate (SER), that is the fraction of inputs <italic>s</italic> that are misclassified. The SER for the equalization with a SNR of 20dB is depicted as a function of &#x003B1; for &#x003B3; &#x0003D; 1 in <xref ref-type="fig" rid="F8">Figure 8</xref> (right). In the case of <italic>T</italic> &#x0003D; 0, there is a clear improvement of the performance from &#x003B1; &#x0003D; 0 to &#x003B1; &#x0003D; 1 but the errors are almost constant when &#x003B1; is further increased. When <italic>T</italic> &#x0003D; &#x003B8;/0.2 performance improves with &#x003B1; until a minimum SER &#x0003D; 0.012 is reached when &#x003B1; &#x0007E; 4. This SER is similar to that obtained when <italic>T</italic> &#x0003D; 0. Then, regardless the value of <italic>T</italic>/&#x003B8;, a similar performance is obtained by using the mismatch &#x003B1;. A SER of 0.01 for the channel equalization task has been obtained using an optoelectronic reservoir computer [<xref ref-type="bibr" rid="B15">15</xref>].</p>
<p>It is not straightforward how the processing capacity will translate into the performance for specific tasks. Different tasks require to compute functions with different degrees of non-linearity and memory. Information processing capacity should be complemented with those requirements to identify optimized operating conditions for the reservoir. For the channel equalization task, when <italic>T</italic> &#x0003D; 0 the capacities LMC and XMC increase with &#x003B1; showing a very large increase from &#x003B1; &#x0003D; 0 to &#x003B1; &#x0003D; 1 (see <xref ref-type="fig" rid="F4">Figure 4C</xref>). The SER shows also a clear decrease from &#x003B1; &#x0003D; 0 to &#x003B1; &#x0003D; 1 but it is almost constant when &#x003B1; &#x0003E; 1 [see <xref ref-type="fig" rid="F8">Figure 8</xref> (right)]. The capacities LMC and XMC achieved for &#x003B1; &#x0003D; 1 when <italic>T</italic> &#x0003D; 0 are enough to solve the channel equalization task. However, the quadratic capacity QMC is almost constant when &#x003B1; &#x0003E; 1. As a consequence the SER is almost constant for &#x003B1; &#x0003E; 1. When taking a small node separation (&#x003B8; &#x0003D; 0.2<italic>T</italic>) the capacities LMC and XMC increase with &#x003B1; (see <xref ref-type="fig" rid="F4">Figure 4A</xref>). This increase in processing capacity leads to a better performance with &#x003B1; and the SER decreases from 0.017 for &#x003B1; &#x0003D; 0 to a minimum error of 0.012 for &#x003B1; &#x0003D; 4. This is an improvement in performance of around 30%. However, the increase in the total capacity for &#x003B1; &#x0003E; 4 (mainly due to the LMC) does not translate into the performance. The reason is the same as for the case of <italic>T</italic> &#x0003D; 0. The capacities LMC and XMC achieved for &#x003B1; &#x0003D; 4 are enough to solve the channel equalization task while the capacities QMC and CMC do not increase with &#x003B1;.</p>
<p>The addition of the second delay line to the non-linear node does not improve the performance for the equalization task. In the case of <italic>T</italic> &#x0003D; 0, the extra delay line slightly improves the performance for the NARMA-10 task. The minimum error is NRMSE &#x0007E; 0.25 when &#x003B1;<sub>1</sub> &#x0003D; 77, &#x003B1;<sub>2</sub> &#x0003D; 20 and &#x003B2;<sub>1</sub> &#x0003D; &#x003B2;<sub>2</sub> &#x0003D; 0.4. When <italic>T</italic> &#x0003D; &#x003B8;/0.2 a NRMSE &#x0003D; 0.27 is obtained for &#x003B1;<sub>1</sub> &#x0003D; 77, &#x003B1;<sub>2</sub> &#x0003D; 86, &#x003B2;<sub>1</sub> &#x0003D; 0.05, and &#x003B2;<sub>2</sub> &#x0003D; 0.75, while a minimum NRMSE=0.34 was obtained with one delay line for &#x003B1; &#x0007E; 72. This performance improvement for the NARMA-10 task when <italic>T</italic> &#x0003D; &#x003B8;/0.2 is at the cost of adding second delay line and optimizing more parameters to minimize the error. A NRMSE of 0.22 for the NARMA-10 task has been obtained using a photonic reservoir computer based on a coherently driven passive cavity with a greater number of virtual nodes <italic>N</italic> &#x0003D; 300 [<xref ref-type="bibr" rid="B24">24</xref>] than the one we used, <italic>N</italic> &#x0003D; 97.</p>
</sec></sec>
<sec sec-type="discussion" id="s4">
<title>4. Discussion</title>
<p>We have investigated the role of the system response time in the computational capacity of delay-based reservoir computers with a single non-linear neuron. These reservoir computers can be easily implemented in hardware, potentially allowing for high-speed information processing. The information processing rate, given by <inline-formula><mml:math id="M42"><mml:mn>1</mml:mn><mml:mo>/</mml:mo><mml:msub><mml:mrow><mml:mi>T</mml:mi></mml:mrow><mml:mrow><mml:mi>p</mml:mi></mml:mrow></mml:msub><mml:mo>=</mml:mo><mml:msup><mml:mrow><mml:mrow><mml:mo stretchy="false">(</mml:mo><mml:mrow><mml:mi>N</mml:mi><mml:mi>&#x003B8;</mml:mi></mml:mrow><mml:mo stretchy="false">)</mml:mo></mml:mrow></mml:mrow><mml:mrow><mml:mo>-</mml:mo><mml:mn>1</mml:mn></mml:mrow></mml:msup></mml:math></inline-formula>, can be increased by using a high sampling output rate (small node separation &#x003B8;). However, we have shown that the computational capacity is reduced when node separation is smaller than system response time. We can thus conclude that there is a trade-off between information capacity and rate in delay-based reservoir computers. In this context, parallel-based architectures with <italic>k</italic> non-linear nodes reduce the information processing time by a factor of <italic>k</italic> for the same total number of virtual nodes. It has been shown [<xref ref-type="bibr" rid="B16">16</xref>, <xref ref-type="bibr" rid="B25">25</xref>] that for (&#x003B8;/<italic>T</italic>) &#x0003C; 1 and without mismatch between <italic>T</italic><sub><italic>p</italic></sub> and &#x003C4;, performance is improved when different activation functions are used for the non-linear nodes. However, the hardware implementation becomes more involved than the one of a delay-based reservoir computer with a single non-linear node.</p>
<p>We have considered a different strategy still based on the simple architecture of a single non-linear node to tackle the trade-off between information capacity and rate. In this strategy, the mismatch between delay and data injection times &#x003B1; is used to increase reservoir diversity when &#x003B8; &#x0003C; <italic>T</italic>. For small values of (&#x003B8;/<italic>T</italic>) and &#x003B1;, the states of virtual nodes that are separated by less than <italic>T</italic> (i.e., with an index difference smaller than <italic>T</italic>/&#x003B8;) are similar. When the mismatch is increased, virtual nodes are connected through feedback to nodes that are not connected through non-linear node dynamics. Reservoir diversity is then increased. Our results show that the linear memory capacity increases the mismatch &#x003B1;. In this way the capacity degradation due to high sampling output rate is reduced by increasing &#x003B1;.</p>
<p>Another strategy to increase reservoir diversity when &#x003B8; &#x0003C; <italic>T</italic> is to use an extra feedback line. We show that the linear memory capacity can be further increased with this architecture by using long delay times (large mismatch &#x003B1;). However, only a slight increase in the calculated capacity is obtained.</p>
<p>We have also obtained the performance of delay-based reservoir computers for two benchmarking tasks: channel equalization and NARMA-10. Our results show that for fast reservoirs with &#x003B8; &#x0003C; <italic>T</italic> performance improves when the mismatch &#x003B1; increases. A similar performance is obtained for small and large values of (&#x003B8;/<italic>T</italic>) for channel equalization and also for NARMA-10 tasks if delay and injection times are mismatched.</p>
<p>We can thus conclude that the processing speed of delay-based reservoir computers can be increased while keeping a good computational capacity by using a mismatch between delay and data injection times.</p></sec>
<sec sec-type="data-availability-statement" id="s5">
<title>Data Availability Statement</title>
<p>The datasets generated for this study are available on request to the corresponding author.</p></sec>
<sec id="s6">
<title>Author Contributions</title>
<p>SO implemented the program and performed the numerical calculations. All authors contributed to the conception, design of the study, contributed to the discussion of the results, and to the writing of the manuscript.</p>
<sec>
<title>Conflict of Interest</title>
<p>The authors declare that the research was conducted in the absence of any commercial or financial relationships that could be construed as a potential conflict of interest.</p>
</sec></sec>
</body>
<back>
<ref-list>
<title>References</title>
<ref id="B1">
<label>1.</label>
<citation citation-type="journal"><person-group person-group-type="author"><name><surname>Jaeger</surname> <given-names>H</given-names></name> <name><surname>Haas</surname> <given-names>H</given-names></name></person-group>. <article-title>Harnessing nonlinearity: predicting chaotic systems and saving energy in wireless communication</article-title>. <source>Science</source>. (<year>2004</year>) <volume>304</volume>:<fpage>78</fpage>&#x02013;<lpage>80</lpage>. <pub-id pub-id-type="doi">10.1126/science.1091277</pub-id><pub-id pub-id-type="pmid">15064413</pub-id></citation></ref>
<ref id="B2">
<label>2.</label>
<citation citation-type="journal"><person-group person-group-type="author"><name><surname>Maass</surname> <given-names>W</given-names></name> <name><surname>Natschl&#x000E4;ger</surname> <given-names>T</given-names></name> <name><surname>Markram</surname> <given-names>H</given-names></name></person-group>. <article-title>Real-time computing without stable states: a new framework for neural computation based on perturbations</article-title>. <source>Neural Comput.</source> (<year>2002</year>) <volume>14</volume>:<fpage>2531</fpage>&#x02013;<lpage>60</lpage>. <pub-id pub-id-type="doi">10.1162/089976602760407955</pub-id><pub-id pub-id-type="pmid">12433288</pub-id></citation></ref>
<ref id="B3">
<label>3.</label>
<citation citation-type="journal"><person-group person-group-type="author"><name><surname>Luko&#x00161;evicius</surname> <given-names>M</given-names></name> <name><surname>Jaeger</surname> <given-names>H</given-names></name></person-group>. <article-title>Reservoir computing approaches to recurrent neural network training</article-title>. <source>Comput Sci Rev</source>. (<year>2009</year>) <volume>3</volume>:<fpage>127</fpage>&#x02013;<lpage>49</lpage>. <pub-id pub-id-type="doi">10.1016/j.cosrev.2009.03.005</pub-id></citation></ref>
<ref id="B4">
<label>4.</label>
<citation citation-type="book"><person-group person-group-type="author"><name><surname>Jaeger</surname> <given-names>H</given-names></name></person-group>. <source>Short Term Memory in Echo State Networks</source>. GMD Forschungszentrum Informationstechnik GmbH. GMD Report 152, Sankt Augustin (<year>2002</year>).</citation></ref>
<ref id="B5">
<label>5.</label>
<citation citation-type="journal"><person-group person-group-type="author"><name><surname>Larger</surname> <given-names>L</given-names></name> <name><surname>Bayl&#x000F3;n-Fuentes</surname> <given-names>A</given-names></name> <name><surname>Martinenghi</surname> <given-names>R</given-names></name> <name><surname>Udaltsov</surname> <given-names>VS</given-names></name> <name><surname>Chembo</surname> <given-names>YK</given-names></name> <name><surname>Jacquot</surname> <given-names>M</given-names></name></person-group>. <article-title>High-speed photonic reservoir computing using a time-delay-based architecture: million words per second classification</article-title>. <source>Phys Rev X.</source> (<year>2017</year>) <volume>7</volume>:<fpage>11015</fpage>. <pub-id pub-id-type="doi">10.1103/PhysRevX.7.011015</pub-id></citation></ref>
<ref id="B6">
<label>6.</label>
<citation citation-type="journal"><person-group person-group-type="author"><name><surname>Moon</surname> <given-names>J</given-names></name> <name><surname>Ma</surname> <given-names>W</given-names></name> <name><surname>Shin</surname> <given-names>JH</given-names></name> <name><surname>Cai</surname> <given-names>F</given-names></name> <name><surname>Du</surname> <given-names>C</given-names></name> <name><surname>Lee</surname> <given-names>SH</given-names></name> <etal/></person-group>. <article-title>Temporal data classification and forecasting using a memristor-based reservoir computing system</article-title>. <source>Nat Electr.</source> (<year>2019</year>) <volume>2</volume>:<fpage>480</fpage>&#x02013;<lpage>7</lpage>. <pub-id pub-id-type="doi">10.1038/s41928-019-0313-3</pub-id></citation></ref>
<ref id="B7">
<label>7.</label>
<citation citation-type="journal"><person-group person-group-type="author"><name><surname>Appeltant</surname> <given-names>L</given-names></name> <name><surname>Soriano</surname> <given-names>MC</given-names></name> <name><surname>Van Der Sande</surname> <given-names>G</given-names></name> <name><surname>Danckaert</surname> <given-names>J</given-names></name> <name><surname>Massar</surname> <given-names>S</given-names></name> <name><surname>Dambre</surname> <given-names>J</given-names></name> <etal/></person-group>. <article-title>Information processing using a single dynamical node as complex system</article-title>. <source>Nat Communicat.</source> (<year>2011</year>) <volume>2</volume>:<fpage>468</fpage>. <pub-id pub-id-type="doi">10.1038/ncomms1476</pub-id><pub-id pub-id-type="pmid">21915110</pub-id></citation></ref>
<ref id="B8">
<label>8.</label>
<citation citation-type="journal"><person-group person-group-type="author"><name><surname>Paquot</surname> <given-names>Y</given-names></name> <name><surname>Duport</surname> <given-names>F</given-names></name> <name><surname>Smerieri</surname> <given-names>A</given-names></name> <name><surname>Dambre</surname> <given-names>J</given-names></name> <name><surname>Schrauwen</surname> <given-names>B</given-names></name> <name><surname>Haelterman</surname> <given-names>M</given-names></name> <etal/></person-group>. <article-title>Optoelectronic reservoir computing</article-title>. <source>Sci Rep</source>. (<year>2012</year>) <volume>2</volume>:<fpage>287</fpage>. <pub-id pub-id-type="doi">10.1038/srep00287</pub-id><pub-id pub-id-type="pmid">22371825</pub-id></citation></ref>
<ref id="B9">
<label>9.</label>
<citation citation-type="journal"><person-group person-group-type="author"><name><surname>Van der Sande</surname> <given-names>G</given-names></name> <name><surname>Brunner</surname> <given-names>D</given-names></name> <name><surname>Soriano</surname> <given-names>MC</given-names></name></person-group>. <article-title>Advances in photonic reservoir computing</article-title>. <source>Nanophotonics</source>. (<year>2017</year>) <volume>6</volume>:<fpage>561</fpage>&#x02013;<lpage>76</lpage>. <pub-id pub-id-type="doi">10.1515/nanoph-2016-0132</pub-id></citation></ref>
<ref id="B10">
<label>10.</label>
<citation citation-type="journal"><person-group person-group-type="author"><name><surname>Rodan</surname> <given-names>A</given-names></name> <name><surname>Tino</surname> <given-names>P</given-names></name></person-group>. <article-title>Minimum complexity echo state network</article-title>. <source>IEEE T Neural Netw</source>. (<year>2011</year>) <volume>22</volume>:<fpage>131</fpage>&#x02013;<lpage>44</lpage>. <pub-id pub-id-type="doi">10.1109/TNN.2010.2089641</pub-id><pub-id pub-id-type="pmid">21075721</pub-id></citation></ref>
<ref id="B11">
<label>11.</label>
<citation citation-type="journal"><person-group person-group-type="author"><name><surname>Martinenghi</surname> <given-names>R</given-names></name> <name><surname>Rybalko</surname> <given-names>S</given-names></name> <name><surname>Jacquot</surname> <given-names>M</given-names></name> <name><surname>Chembo</surname> <given-names>YK</given-names></name> <name><surname>Larger</surname> <given-names>L</given-names></name></person-group>. <article-title>Photonic nonlinear transient computing with multiple-delay wavelength dynamics</article-title>. <source>Phys Rev Let.</source> (<year>2012</year>) <volume>108</volume>:<fpage>244101</fpage>. <pub-id pub-id-type="doi">10.1103/PhysRevLett.108.244101</pub-id><pub-id pub-id-type="pmid">23004274</pub-id></citation></ref>
<ref id="B12">
<label>12.</label>
<citation citation-type="journal"><person-group person-group-type="author"><name><surname>Soriano</surname> <given-names>MC</given-names></name> <name><surname>Ort&#x000ED;n</surname> <given-names>S</given-names></name> <name><surname>Keuninckx</surname> <given-names>L</given-names></name> <name><surname>Appeltant</surname> <given-names>L</given-names></name> <name><surname>Danckaert</surname> <given-names>J</given-names></name> <name><surname>Pesquera</surname> <given-names>L</given-names></name> <etal/></person-group>. <article-title>Delay-based reservoir computing: noise effects in a combined analog and digital implementation</article-title>. <source>IEEE Trans Neural Netw Learn Syst</source>. (<year>2015</year>) <volume>26</volume>:<fpage>388</fpage>&#x02013;<lpage>93</lpage>. <pub-id pub-id-type="doi">10.1109/TNNLS.2014.2311855</pub-id><pub-id pub-id-type="pmid">25608295</pub-id></citation></ref>
<ref id="B13">
<label>13.</label>
<citation citation-type="journal"><person-group person-group-type="author"><name><surname>Larger</surname> <given-names>L</given-names></name> <name><surname>Soriano</surname> <given-names>MC</given-names></name> <name><surname>Brunner</surname> <given-names>D</given-names></name> <name><surname>Appeltant</surname> <given-names>L</given-names></name> <name><surname>Gutierrez</surname> <given-names>JM</given-names></name> <name><surname>Pesquera</surname> <given-names>L</given-names></name> <etal/></person-group>. <article-title>Photonic information processing beyond Turing: an optoelectronic implementation of reservoir computing</article-title>. <source>Opt Express</source>. (<year>2012</year>) <volume>20</volume>:<fpage>3241</fpage>&#x02013;<lpage>49</lpage>. <pub-id pub-id-type="doi">10.1364/OE.20.003241</pub-id><pub-id pub-id-type="pmid">22330562</pub-id></citation></ref>
<ref id="B14">
<label>14.</label>
<citation citation-type="journal"><person-group person-group-type="author"><name><surname>Soriano</surname> <given-names>MC</given-names></name> <name><surname>Ort&#x000ED;n</surname> <given-names>S</given-names></name> <name><surname>Brunner</surname> <given-names>D</given-names></name> <name><surname>Larger</surname> <given-names>L</given-names></name> <name><surname>Mirasso</surname> <given-names>CR</given-names></name> <name><surname>Fischer</surname> <given-names>I</given-names></name> <etal/></person-group>. <article-title>Optoelectronic reservoir computing: tackling noise-induced performance degradation</article-title>. <source>Opt Express</source>. (<year>2013</year>) <volume>21</volume>:<fpage>12</fpage>&#x02013;<lpage>20</lpage>. <pub-id pub-id-type="doi">10.1364/OE.21.000012</pub-id><pub-id pub-id-type="pmid">23388891</pub-id></citation></ref>
<ref id="B15">
<label>15.</label>
<citation citation-type="journal"><person-group person-group-type="author"><name><surname>Ort&#x000ED;n</surname> <given-names>S</given-names></name> <name><surname>Soriano</surname> <given-names>MC</given-names></name> <name><surname>Pesquera</surname> <given-names>L</given-names></name> <name><surname>Brunner</surname> <given-names>D</given-names></name> <name><surname>San-Mart&#x000ED;n</surname> <given-names>D</given-names></name> <name><surname>Fischer</surname> <given-names>I</given-names></name> <etal/></person-group>. <article-title>A unified framework for reservoir computing and extreme learning machines based on a single time-delayed neuron</article-title>. <source>Sci Rep</source>. (<year>2015</year>) <volume>5</volume>:<fpage>14945</fpage>. <pub-id pub-id-type="doi">10.1038/srep14945</pub-id><pub-id pub-id-type="pmid">26446303</pub-id></citation></ref>
<ref id="B16">
<label>16.</label>
<citation citation-type="journal"><person-group person-group-type="author"><name><surname>Ort&#x000ED;n</surname> <given-names>S</given-names></name> <name><surname>Pesquera</surname> <given-names>L</given-names></name></person-group>. <article-title>Reservoir computing with an ensemble of time-delay reservoirs</article-title>. <source>Cogn Comput.</source> (<year>2017</year>) <volume>9</volume>:<fpage>327</fpage>&#x02013;<lpage>36</lpage>. <pub-id pub-id-type="doi">10.1007/s12559-017-9463-7</pub-id></citation></ref>
<ref id="B17">
<label>17.</label>
<citation citation-type="book"><person-group person-group-type="author"><name><surname>Jaeger</surname> <given-names>H</given-names></name></person-group>. <source>Tutorial on training recurrent neural networks, covering BPTT, RTRL, EKF and the &#x02018;echo state network&#x00027; approach</source>. Technical Report GMD Report 159, German National Research Center for Information Technology, Sankt Augustin (<year>2002</year>).</citation></ref>
<ref id="B18">
<label>18.</label>
<citation citation-type="journal"><person-group person-group-type="author"><name><surname>Brunner</surname> <given-names>D</given-names></name> <name><surname>Soriano</surname> <given-names>MC</given-names></name> <name><surname>Mirasso</surname> <given-names>CR</given-names></name> <name><surname>Fischer</surname> <given-names>I</given-names></name></person-group>. <article-title>Parallel photonic information processing at gigabyte per second data rates using transient states</article-title>. <source>Nat Commun</source>. (<year>2013</year>) <volume>4</volume>:<fpage>1364</fpage>. <pub-id pub-id-type="doi">10.1038/ncomms2368</pub-id><pub-id pub-id-type="pmid">23322052</pub-id></citation></ref>
<ref id="B19">
<label>19.</label>
<citation citation-type="journal"><person-group person-group-type="author"><name><surname>Duport</surname> <given-names>F</given-names></name> <name><surname>Schneider</surname> <given-names>B</given-names></name> <name><surname>Smerieri</surname> <given-names>A</given-names></name> <name><surname>Haelterman</surname> <given-names>M</given-names></name> <name><surname>Massar</surname> <given-names>S</given-names></name></person-group>. <article-title>All-optical reservoir computing</article-title>. <source>Opt Express</source>. (<year>2012</year>) <volume>20</volume>:<fpage>22783</fpage>&#x02013;<lpage>95</lpage>. <pub-id pub-id-type="doi">10.1364/OE.20.022783</pub-id><pub-id pub-id-type="pmid">23037429</pub-id></citation></ref>
<ref id="B20">
<label>20.</label>
<citation citation-type="journal"><person-group person-group-type="author"><name><surname>Dambre</surname> <given-names>J</given-names></name> <name><surname>Verstraeten</surname> <given-names>D</given-names></name> <name><surname>Schrauwen</surname> <given-names>B</given-names></name> <name><surname>Massar</surname> <given-names>S</given-names></name></person-group>. <article-title>Information processing capacity of dynamical systems</article-title>. <source>Sci Rep.</source> (<year>2012</year>) <volume>2</volume>:<fpage>514</fpage>. <pub-id pub-id-type="doi">10.1038/srep00514</pub-id><pub-id pub-id-type="pmid">22816038</pub-id></citation></ref>
<ref id="B21">
<label>21.</label>
<citation citation-type="book"><person-group person-group-type="author"><name><surname>Appeltant</surname> <given-names>L</given-names></name></person-group>. <source>Reservoir Computing Based on Delay-Dynamical Systems</source>. Vrije Universiteit Brussel/Universitat de les Illes Balears, Brussels (<year>2012</year>).</citation></ref>
<ref id="B22">
<label>22.</label>
<citation citation-type="book"><person-group person-group-type="author"><name><surname>Ort&#x000ED;n</surname> <given-names>S</given-names></name> <name><surname>Appeltant</surname> <given-names>L</given-names></name> <name><surname>Pesquera</surname> <given-names>L</given-names></name> <name><surname>der Sande</surname> <given-names>G</given-names></name> <name><surname>Danckaert</surname> <given-names>J</given-names></name> <name><surname>Gutierrez</surname> <given-names>JM</given-names></name></person-group>. <article-title>Information processing using an electro-optic oscillator subject to multiple delay lines</article-title>. In: <source>International Quantum Electronics Conference</source>. <publisher-loc>Piscataway, NJ</publisher-loc>: <publisher-name>Optical Society of America</publisher-name> (<year>2013</year>).</citation></ref>
<ref id="B23">
<label>23.</label>
<citation citation-type="journal"><person-group person-group-type="author"><name><surname>Nieters</surname> <given-names>P</given-names></name> <name><surname>Leugering</surname> <given-names>J</given-names></name> <name><surname>Pipa</surname> <given-names>G</given-names></name></person-group>. <article-title>Neuromorphic computation in multi-delay coupled models</article-title>. <source>IBM J Res Dev</source>. (<year>2017</year>) <volume>61</volume>:<fpage>8:1</fpage>&#x02013;<lpage>8:9</lpage>. <pub-id pub-id-type="doi">10.1147/JRD.2017.2664698</pub-id></citation></ref>
<ref id="B24">
<label>24.</label>
<citation citation-type="journal"><person-group person-group-type="author"><name><surname>Vinckier</surname> <given-names>Q</given-names></name> <name><surname>Duport</surname> <given-names>F</given-names></name> <name><surname>Smerieri</surname> <given-names>A</given-names></name> <name><surname>Vandoorne</surname> <given-names>K</given-names></name> <name><surname>Bienstman</surname> <given-names>P</given-names></name> <name><surname>Haelterman</surname> <given-names>M</given-names></name> <etal/></person-group>. <article-title>High-performance photonic reservoir computer based on a coherently driven passive cavity</article-title>. <source>Optica</source>. (<year>2015</year>) <volume>2</volume>:<fpage>438</fpage>&#x02013;<lpage>46</lpage>. <pub-id pub-id-type="doi">10.1364/OPTICA.2.000438</pub-id></citation></ref>
<ref id="B25">
<label>25.</label>
<citation citation-type="journal"><person-group person-group-type="author"><name><surname>Ort&#x000ED;n</surname> <given-names>S</given-names></name> <name><surname>Pesquera</surname> <given-names>L</given-names></name> <name><surname>Guti&#x000E9;rrez</surname> <given-names>JM</given-names></name></person-group>. <article-title>Memory and nonlinear mapping in reservoir computing with two uncoupled nonlinear delay nodes</article-title>. In: <source>Proceedings of the European Conference on Complex Systems 2012</source>. Berlin: Springer (<year>2013</year>). p. <fpage>895</fpage>&#x02013;<lpage>9</lpage>. <pub-id pub-id-type="doi">10.1007/978-3-319-00395-5_107</pub-id></citation></ref>
</ref-list>
<fn-group>
<fn fn-type="financial-disclosure"><p><bold>Funding.</bold> This work has been funded by the Spanish Ministerio de Ciencia, Innovaci&#x000F3;n y Universidades and Fondo Europeo de Desarrollo Regional (FEDER) through project RTI2018-094118-B-C22.</p>
</fn>
</fn-group>
</back>
</article>