<?xml version="1.0" encoding="UTF-8" standalone="no"?>
<!DOCTYPE article PUBLIC "-//NLM//DTD Journal Publishing DTD v2.3 20070202//EN" "journalpublishing.dtd">
<article xmlns:mml="http://www.w3.org/1998/Math/MathML" xmlns:xlink="http://www.w3.org/1999/xlink" article-type="research-article">
<front>
<journal-meta>
<journal-id journal-id-type="publisher-id">Front. Robot. AI</journal-id>
<journal-title>Frontiers in Robotics and AI</journal-title>
<abbrev-journal-title abbrev-type="pubmed">Front. Robot. AI</abbrev-journal-title>
<issn pub-type="epub">2296-9144</issn>
<publisher>
<publisher-name>Frontiers Media S.A.</publisher-name>
</publisher>
</journal-meta>
<article-meta>
<article-id pub-id-type="doi">10.3389/frobt.2019.00010</article-id>
<article-categories>
<subj-group subj-group-type="heading">
<subject>Robotics and AI</subject>
<subj-group>
<subject>Original Research</subject>
</subj-group>
</subj-group>
</article-categories>
<title-group>
<article-title>Importance of Parameter Settings on the Benefits of Robot-to-Robot Learning in Evolutionary Robotics</article-title>
</title-group>
<contrib-group>
<contrib contrib-type="author" corresp="yes">
<name><surname>Heinerman</surname> <given-names>Jacqueline</given-names></name>
<xref ref-type="corresp" rid="c001"><sup>&#x0002A;</sup></xref>
<uri xlink:href="http://loop.frontiersin.org/people/554836/overview"/>
</contrib>
<contrib contrib-type="author">
<name><surname>Haasdijk</surname> <given-names>Evert</given-names></name>
</contrib>
<contrib contrib-type="author">
<name><surname>Eiben</surname> <given-names>A. E.</given-names></name><uri xlink:href="http://loop.frontiersin.org/people/133074/overview"/>
</contrib>
</contrib-group>
<aff><institution>Department of Computer Science, Vrije Universiteit Amsterdam</institution>, <addr-line>Amsterdam</addr-line>, <country>Netherlands</country></aff>
<author-notes>
<fn fn-type="edited-by"><p>Edited by: Mel Slater, University of Barcelona, Spain</p></fn>
<fn fn-type="edited-by"><p>Reviewed by: Eiji Uchibe, Advanced Telecommunications Research Institute International (ATR), Japan; Andrej Gams, Jo&#x0017E;ef Stefan Institute (IJS), Slovenia</p></fn>
<corresp id="c001">&#x0002A;Correspondence: Jacqueline Heinerman <email>j.v.heinerman&#x00040;vu.nl</email></corresp>
<fn fn-type="other" id="fn001"><p>This article was submitted to Evolutionary Robotics, a section of the journal Frontiers in Robotics and AI</p></fn></author-notes>
<pub-date pub-type="epub">
<day>04</day>
<month>03</month>
<year>2019</year>
</pub-date>
<pub-date pub-type="collection">
<year>2019</year>
</pub-date>
<volume>6</volume>
<elocation-id>10</elocation-id>
<history>
<date date-type="received">
<day>02</day>
<month>08</month>
<year>2018</year>
</date>
<date date-type="accepted">
<day>04</day>
<month>02</month>
<year>2019</year>
</date>
</history>
<permissions>
<copyright-statement>Copyright &#x000A9; 2019 Heinerman, Haasdijk and Eiben.</copyright-statement>
<copyright-year>2019</copyright-year>
<copyright-holder>Heinerman, Haasdijk and Eiben</copyright-holder>
<license xlink:href="http://creativecommons.org/licenses/by/4.0/"><p>This is an open-access article distributed under the terms of the Creative Commons Attribution License (CC BY). The use, distribution or reproduction in other forums is permitted, provided the original author(s) and the copyright owner(s) are credited and that the original publication in this journal is cited, in accordance with accepted academic practice. No use, distribution or reproduction is permitted which does not comply with these terms.</p></license>
</permissions>
<abstract><p>Robot-to-robot learning, a specific case of social learning in robotics, enables multiple robots to share learned skills while completing a task. The literature offers various statements of its benefits. Robots using this type of social learning can reach a higher performance, an increased learning speed, or both, compared to robots using individual learning only. No general explanation has been advanced for the difference in observations, which make the results highly dependent on the particular system and parameter setting. In this paper, we perform a detailed analysis into the effects of robot-to-robot learning. As a result, we show that this type of social learning can reduce the sensitivity of the learning process to the choice of parameters in two ways. First, robot-to-robot learning can reduce the number of bad performing individuals in the population. Second, robot-to-robot learning can increase the chance of having a successful run, where success is defined as the presence of a high performing individual. Additionally, we show that robot-to-robot learning results in an increased learning speed for almost all parameter settings. Our results indicate that robot-to-robot learning is a powerful mechanism which leads to benefits in both performance and learning speed.</p></abstract>
<kwd-group>
<kwd>social learning</kwd>
<kwd>robot-to-robot learning</kwd>
<kwd>evolutionary robotics</kwd>
<kwd>parameter tuning</kwd>
<kwd>neural networks</kwd>
<kwd>evolutionary algorithms</kwd>
</kwd-group>
<contract-sponsor id="cn001">Horizon 2020<named-content content-type="fundref-id">10.13039/501100007601</named-content></contract-sponsor>
<counts>
<fig-count count="5"/>
<table-count count="0"/>
<equation-count count="0"/>
<ref-count count="32"/>
<page-count count="11"/>
<word-count count="7596"/>
</counts>
</article-meta>
</front>
<body>

<sec sec-type="intro" id="s1">
<title>1. Introduction</title>
<p>The widely used definition of social learning reflects animal behavior: social learning is learning through observation of conspecifics. Considering humans the definition can be extended: social learning is learning through observation of conspecifics or transferring knowledge through language. That is, the ability to use language offers a new method, a second tool in the toolbox of social learning. Regarding robots, we can add a third tool to this toolbox based on the ability to transfer robot controllers directly from one robot to another. (In common parlance this would be the robotic equivalent of telepathy.) Thus, the definition of social learning can be broadened again. If robots are concerned then social learning is learning through observation of conspecifics or transferring knowledge through language or direct exchange of (parts of) controllers.</p>
<p>In the current paper we focus on the third option for robots, the direct exchange of controllers, that is a special case of social learning that is only available for robots. To emphasize this we usethe term robot-to-robot learning.</p>
<p>Consider a collective of autonomous robots in an environment that is not well understood or modeled at design time. It is not possible to develop and validate adequate robot controllers without a thorough understanding of the environment, so the robots need to be able to adapt their behavior to suit. It is preferable that the robots are capable of learning autonomously, without the need for centraloversight: such a centralized scheme implies a single point of failure in the combined system.</p>
<p>To illustrate our concept of adaptation, consider the controller of a robot as a process that maps inputs, read from the robot&#x00027;s sensors and internal states to outputs, typically actuator and state settings. Learning can then be defined as any change to the mapping between inputs and outputs, cf. (Haasdijk et al., <xref ref-type="bibr" rid="B10">2013</xref>). In such a setting, the robots can learn individually, e.g., by encapsulating a self-sufficient learning algorithm within each robot, and they can learn collectively by sharing knowledge, called robot-to-robot learning.</p>
<p>Robot-to-robot learning in a robotic collective has been studied for different machine learning implementations such as Reinforcement Learning (e.g., Sutton and Barto, <xref ref-type="bibr" rid="B27">1998</xref>; Zhang et al., <xref ref-type="bibr" rid="B32">2010</xref>; Noble and Franks, <xref ref-type="bibr" rid="B19">2012</xref>; Wiering and van Otterlo, <xref ref-type="bibr" rid="B31">2012</xref>) and Evolutionary Algorithms (EAs) (e.g., Pugh and Martinoli, <xref ref-type="bibr" rid="B23">2009</xref>; Eiben and Smith, <xref ref-type="bibr" rid="B5">2015a</xref>,<xref ref-type="bibr" rid="B6">b</xref>). In this paper, we consider evolutionary algorithms applied to robotics, i.e., Evolutionary Robotics (Nolfi and Floreano, <xref ref-type="bibr" rid="B20">2000</xref>).</p>
<p>There is ample evidence that set-ups, where robots can share knowledge, outperform otherwise equivalent set-ups where robots learn in isolation.When robots share knowledge, they achieve better performance and/or the learning curve is steeper (Usui and Arita, <xref ref-type="bibr" rid="B29">2003</xref>; Curran and ORiordan, <xref ref-type="bibr" rid="B3">2007</xref>; Perez et al., <xref ref-type="bibr" rid="B22">2008</xref>; Pugh and Martinoli, <xref ref-type="bibr" rid="B23">2009</xref>; Garca-Sanchez et al., <xref ref-type="bibr" rid="B8">2012</xref>; Miikkulainen et al., <xref ref-type="bibr" rid="B17">2012</xref>; Tansey et al., <xref ref-type="bibr" rid="B28">2012</xref>; Heinerman et al., <xref ref-type="bibr" rid="B12">2015a</xref>,<xref ref-type="bibr" rid="B13">b</xref>; Jolley et al., <xref ref-type="bibr" rid="B16">2016</xref>). A higher overall performance can be observed when there is a quality or diversity assessment before the knowledge is sent or incorporated (Huijsman et al., <xref ref-type="bibr" rid="B15">2011</xref>; Garca-Sanchez et al., <xref ref-type="bibr" rid="B8">2012</xref>; Heinerman et al., <xref ref-type="bibr" rid="B13">2015b</xref>). Evidence by Huijsman et al. (<xref ref-type="bibr" rid="B15">2011</xref>) and Silva et al. (<xref ref-type="bibr" rid="B24">2015</xref>) show that robot-to-robot learning can linearly decrease learning time, e.g., the fitness measure that four robots can reach in 2 h can be reached by eight robots in 1 h when they learn socially. Although there is evidence that robot-to-robot learning can increase performance and/or learning speed, no general explanation has been advanced for the difference in observations. As an example, Usui and Arita (<xref ref-type="bibr" rid="B29">2003</xref>) showed that the speed of adaptation (not the finalperformance level) improves in hybrid set-ups compared to purely distributed ones, but that this improvement depends on the size of the encapsulated population. The authors of Pugh and Martinoli (<xref ref-type="bibr" rid="B23">2009</xref>) also varied the population size but did not find a significant effect of the population size on the performance. Because of these contradicting results, it is difficult to generalize on the benefits of robot-to-robot learning. This makes the results highly dependent on the particular system and parameter settings.</p>
<p>In our research, we increase our understanding of robot-to-robot learning by studying the dependence of the parameter settings on the benefits of robot-to-robot learning within one system. First, we observe the performance and learning speed of anindividual learning robot, learning a foraging task, when using different parameter settings. Then we compare the performance and learning speed with a setup where we enable the exchange of knowledge. This analysis enables us to observe when robot-to-robot learning leads to particular benefits. As a result, it brings us closer to understanding how robot-to-robot learning can improve performance and/or learning speed.</p>
<p>In previous work (Heinerman et al., <xref ref-type="bibr" rid="B14">2017</xref>), we have shown that the observed advantages of robot-to-robot learning depend on the quality of the parameter settings of the individual robotic learning process. In particular, we showed that parameter settings resulting in a median performance experienced more benefits from robot-to-robot learning than parameter setting that already gave a high performance for one robot. As a consequence, research in robot-to-robot learning must consider the quality of the used parameter settings, as they can drastically impact the conclusion.</p>
<p>While this result explains the difference in observations in the current literature, we discovered an additional benefit of robot-to-robot learning that we investigate further in this paper. We show that robot-to-robot learning can reduce the sensitivity of the learning process to the choice of parameters for the individual learning process in two ways. First, robot-to-robot learning can reduce the number of bad performing individuals in the population within one run. Second, robot-to-robot learning can increase the chance of having a high performing individual in the population over multiple runs.</p>
<p>We study robot-to-robot learning in the context of on-line evolutionary robotics. In terms of the taxonomy defined by Haasdijk et al. (<xref ref-type="bibr" rid="B11">2012</xref>), they are defined as <italic>hybrid</italic> systems where robots can adapt their controller individually, but can also exchange information. The field of evolutionary robotics originated in the late 1980s and aims to create robotic controllers with Evolutionary Algorithms (EAs). These algorithms are inspired by Darwin&#x00027;s theory of survival of the fittest. In nature, animals survive and procreate when they are more fit. Similarly, a robotic controller is tested by observing the behavior of the robot and is given a corresponding fitness measure. The higher the fitness, the more chance this controller has to procreate. Over generations, the quality of the controllers will improve and lead to robots that are capable of executing a predefined task properly. The robotic controller that we consider is a neural network. A neural network is a direct policy that maps the sensor inputs of the robot to actions. This mapping, consisting of nodes and connections between the nodes, are evolved with EAs.</p>
<p>This paper is structured as follows: In section 2 we explain our method to select parameter settings and we describe measurements that we use to observe the benefits ofsocial learning. This approach is independent of the system described in section 3. We continue with the experimental setup in section 4 and present the results in section 5. An in depth discussion of the results is given in section 6. Our concluding remarks are summarized in section 7.</p>
</sec>

<sec id="s2">
<title>2. Parameter Settings and Measurements</title>
<p>Our implementation of the individual learning process has 21 parameters. These parameters are related to the individual learning mechanism of one robot. Depending on the chosen learning mechanism, the number of parameters can vary. The approach described in this section is independent of the specific individual learning mechanism of the robot. The parameters, of which 2 are boolean and 19 continuous, are presented in <xref ref-type="app" rid="A1">Appendix</xref>. Every parameter could be assigned a low, middle or high value. A combination of all parameters with their corresponding values is called a <italic>configuration</italic>.</p>
<p>Testing all possible configurations would require too much computational power. Therefore, Design of Experiments (DoE) was used to create the configurations. DoE is an approach that creates a minimal number of configurations to test while preserving the possibility to perform statistical analysis on the data (Montgomery, <xref ref-type="bibr" rid="B18">2009</xref>). The DoE<xref ref-type="fn" rid="fn0001"><sup>1</sup></xref> gave, for the case at hand with 19 continuous and 2 boolean parameters, an experimental design of 50 different configurations.</p>
<p>The configurations provided by the DoE are ranked based on their quality. The quality of a configuration is defined as the median performance in the final generation. The performance of a generation is defined as the median fitness of theindividuals in that generation. Rank number 1 is the configuration resulting in the highest quality and rank number 50 is the configuration resulting in the lowest quality.</p>
<p>Using the median performance measure is the most suitable measure for online evolutionary roboticsfor two reasons. First, the distribution of the fitness within one run and over different runs is very skewed. Second, for on-line evolution, it is more important that all controllers perform well as the robots both learn and perform their task at the same time. Therefore, if we mention performance, either a generation or a configuration, we always refer to the median performance unless mentioned otherwise.</p>
<p>For the robot-to-robot learning experiments, we use the best and median configurations. To avoid confusion, note that this median is not the median of multiple data points but based on the rank of the parameter setting. The best configurations are defined as the 10 settings that lead to the highest quality (rank 1&#x02013;10). The median configurations are defined as the settings resulting in a median quality (rank 21&#x02013;30). How robot-to-robot learning is exactly implemented is explained in section 3.</p>
<p>For a detailed analysis of the effects of robot-to-robot learning, we also need more detailed measurements for performance and learning speed. The measures presented here are not new but they have not (all) been used to study the benefits of robot-to-robot learning.</p>
<p>For the performance we use the following measures:
<list list-type="simple">
<list-item><p><bold>Success Rate (SR)</bold> The SR is measured by the percentage of the 20 replicate runs that have a controller in the final generation with good fitness. Agood fitness is a fitness equal to or higher than 75% of the maximum observed fitness over all runs. The maximum observed fitness in all runs is 12, resulting in a good fitness of 9 or higher.</p></list-item>
<list-item><p>The Success Rate Ratio (SRR) is calculated by dividing the SR of the robot-to-robot learning experiment by the individual learning experiment. An SRR higher than 1 means that robot-to-robot learning results in more successful runs.</p></list-item>
<list-item><p><bold>Population Failure (PF)</bold> The PF is measured by the median percentage of bad controllers in the final generation over the 20 replicate runs. A bad performing robot is a robot with a fitness equal to zero.</p></list-item>
<list-item><p>The Population Failure Ratio (PFR) is calculated by dividing the PF of the individual learning experiment by the robot-to-robot learning experiment. A PFR higher than 1 means that robot-to-robot learning results in fewer bad performing robots in the final generation.</p></list-item>
</list></p>
<p>Note that these two measures act on a different level. The SR is calculated over multiple experiments, while a PF can be calculated for everyexperiment individually. The SR results in one value while the PF results in a value per experiment, of which the median is taken.</p>
<p>For the learning speed we use a measure that isolates the learning speed from the performance level:
<list list-type="simple">
<list-item><p><bold>Learning Speed (LS)</bold> The LS is measured as the numeric integral of the median performance over time (in a number of generations) of the 20 replicate runs. The performance is normalized with respect to the maximum median observed performance over time. In other words, it is the surface underthe median performance curve divided by the surface of a rectangle enclosing the highest median performance over time.</p></list-item>
<list-item><p>The Learning Speed Increase (LSI) is calculated by subtracting the learning speed of the individual learning experiment from the learning speed of the robot-to-robot learning experiment. An LSI higher than 0 means that robot-to-robot learning results in faster learning.</p></list-item>
</list></p>
</sec>

<sec id="s3">
<title>3. System Description</title>
<p>Parts of the system described in this section were already presented in previous work Heinerman et al. (<xref ref-type="bibr" rid="B14">2017</xref>). However, we decided to recite parts to increase the readability of this paper.</p>

<sec>
<title>3.1. Task, Robot, and Environment</title>
<p>The experiment we chose requires the robots to learn a foraging task. A foraging task requires the robot to collect pucks and bring them to the nest located in the center of the arena. The environment is a square arena as shown in <xref ref-type="fig" rid="F1">Figure 1</xref>. Five pucks are randomly placed in the arena at the start of a run. Once a puck is brought to the nest, it is immediately moved to a random location in the environment. The fitness of each robot is equal to the number of pucks collected during a trial lasting 1,000 time steps.</p>
<fig id="F1" position="float">
<label>Figure 1</label>
<caption><p>The environment with one robot searching for the blue pucks. The target location is indicated by the gray circle. The dashed line shows an example trajectory of a robot that picks up a puck to release at the target location.</p></caption>
<graphic xlink:href="frobt-06-00010-g0001.tif"/>
</fig>
<p>The experiments are conducted in simulation using JBotEvolver (Duarte et al., <xref ref-type="bibr" rid="B4">2014</xref>)<xref ref-type="fn" rid="fn0002"><sup>2</sup></xref>. JBotEvolver is a Java-based open-source, cross-platform framework for research and education in Evolutionary Robotics featuring a 2D differential-drive kinematics engine. The robots in our experiments simulate an e-puck robot. This robot is a small (7 cm) differential drive wheeled mobile robot equipped with 8 infrared proximity sensors. The range of these sensors is approximately 10% of the arena width. Additionally, the robots are equipped with the following task-specific sensors:
<list list-type="simple">
<list-item><p><bold>Puck carrying sensor</bold> Indicates if the robot is carrying a puck. The robot can carry one puck at a time;</p></list-item>
<list-item><p><bold>Puck sensor</bold> Indicates the distance to the closest puck within the 45&#x000B0; perception cone of the sensor. The range of this sensor is approximately 60.</p></list-item>
<list-item><p><bold>Nest sensor</bold> Indicates the distance to the nest if within the 45&#x000B0; perception cone of the sensor. The range of this sensor is approximately 60.</p></list-item>
</list></p>
</sec>

<sec>
<title>3.2. Controller and Individual Learning Mechanism</title>
<p>The robot&#x00027;s controller is an artificial neural network. The neural network has 11 input and two output nodes. The input nodes consist of 8 proximity sensors, a nest sensor, a puck sensor, and a puck carrying sensor; the output nodes provide the right and left motor speed. A neural network is a direct policy that maps the sensor inputs of the robot to actions. This mapping, consisting of nodes and connections between the nodes, are learned with an evolutionary algorithm.</p>
<p>Individual learning is implemented by an encapsulating, self-sufficient learning mechanism. The learning mechanism used in this paper is NEAT (Stanley and Miikkulainen, <xref ref-type="bibr" rid="B26">2002</xref>). NEAT is an evolutionary algorithm that evolves both the topology and the connectivity of artificial neural networks. The initial population is composed of randomly generated feedforward neural networks without hidden layers. Every neural network is called an individual. Over time, nodes and connections can be added to the neural network, including the possibility of forming recurrent connections. All nodes have sigmoid activation functions.</p>
<p>Each robot possesses its own instance of NEAT. This means that each robot has a population of individuals, i.e., a set of controllers. These controllers are sequentially evaluated directly on the robot for 1,000 time steps. One time stamp is one sequence of the neural network where the sensor inputs are translated to an action. The fitness after these 1,000 time steps, the number of collected pucks, is stored. This fitness is used to select individuals for reproduction to create a new population.</p>
<p>The learning is conducted online, i.e., the robot is not relocated between the evaluations and each controller is tested starting from the location reached by the previous one. A consequence of online learning is that a controller can suffer from a bad initial position caused by a previous evaluation. Having to recover from a bad starting position can impact the fitness of the new controller in a negative way (Bredeche et al., <xref ref-type="bibr" rid="B2">2009</xref>). In our set-up, the most common example of a bad starting position is being placed against a wall. To mitigate the negative effect of a bad starting position, we reposition the robot to a random location at the beginning of a controller evaluation when it was driving against the wall. This is to make sure that the learning does not stagnate because it is driving into the wall for a sequent number of evaluations and loses the diversity in the population of controllers.</p>
</sec>

<sec>
<title>3.3. Robot-to-Robot Learning Mechanism</title>
<p>When robot-to-robot learning is applied, every robot has its own arena as shown in <xref ref-type="fig" rid="F1">Figure 1</xref>. Every robot is learning in an online fashion, while the robot is performing the task. Robot-to-robot learning is implemented as follows: first, the robots sequentially evaluate all controllers in their current population. Then, the robots exchange information. This means that the information is sent to another arena where another robot is located. This is different from having multiple smaller populations in one robot, because the position of the robot is a result of all evaluated controllers before. The robot compares the received controller?s fitness to that of its own worst controller. The new controller replaces the worst controller if it is better. The NEAT algorithm uses the updated list of controllers and fitness values to create the next generation.</p>
<p>As noted earlier, NEAT can modify the topology of the neural networks during evolution. Every structural modification in the network is identified by a unique innovation number to enable alignment of genomes for recombination purposes. When implementing NEAT with the possibility to exchange individuals as described for robot-to-robot learning, care must be taken to avoid conflicting innovation numbers. In our implementation, we keep track of a centralized global innovation database. If this would not be possible or desirable, one can use a distributed systems such as odNEAT (Silva et al., <xref ref-type="bibr" rid="B25">2012</xref>) or Gene Clocks (Fern&#x000E1;ndez P&#x000E9;rez et al., <xref ref-type="bibr" rid="B7">2015</xref>)</p>
<p>Algorithm 1 summarizes the individual and robot-to-robot learning mechanism in pseudocode.</p>
<table-wrap position="float">
<label>Algorithm 1</label>
<caption><p>Pseudocode of the algorithm that runs on every robot</p></caption>
<table frame="hsides" rules="groups">
<tbody>
<tr>
<td valign="top" align="left">&#x00A0;&#x00A0;l initialise population of first generation (<italic>P</italic><sub>1</sub>) with</td>
</tr>
<tr>
<td valign="top" align="left">&#x00A0;&#x00A0;individuals <italic>i</italic><sub>1</sub>, &#x02026;, <italic>i</italic><sub><italic>n</italic></sub></td>
</tr>
<tr>
<td valign="top" align="left">&#x00A0;&#x00A0;<bold>while</bold> current generation &#x02264; final generation</td>
</tr>
<tr>
<td valign="top" align="left">&#x00A0;&#x00A0;&#x00A0;&#x00A0;<bold>for</bold> every <italic>i</italic> in <italic>P</italic></td>
</tr>
<tr>
<td valign="top" align="left">&#x00A0;&#x00A0;&#x00A0;&#x00A0;&#x00A0;&#x00A0;evaluate <italic>i</italic></td>
</tr>
<tr>
<td valign="top" align="left">&#x00A0;&#x00A0;&#x00A0;&#x00A0;&#x00A0;&#x00A0;store <italic>fitness</italic> of <italic>i</italic></td>
</tr>
<tr>
<td valign="top" align="left">&#x00A0;&#x00A0;&#x00A0;&#x00A0;sort the individuals based on fitness (<italic>i</italic><sub>1</sub> is best)</td>
</tr>
<tr>
<td valign="top" align="left">&#x00A0;&#x00A0;&#x00A0;&#x00A0;<bold>if</bold> robot-to-robot learning?</td>
</tr>
<tr>
<td valign="top" align="left">&#x00A0;&#x00A0;&#x00A0;&#x00A0;&#x00A0;&#x00A0;pick random other robot <italic>R</italic></td>
</tr>
<tr>
<td valign="top" align="left">&#x00A0;&#x00A0;&#x00A0;&#x00A0;&#x00A0;&#x00A0;receive it&#x00027;s best individual <italic>r</italic><sub>1</sub></td>
</tr>
<tr>
<td valign="top" align="left">&#x00A0;&#x00A0;&#x00A0;&#x00A0;&#x00A0;&#x00A0;<bold>if</bold> <italic>fitness</italic>(<italic>r</italic><sub>1</sub>)&#x0003E;<italic>fitness</italic>(<italic>i</italic><sub><italic>n</italic></sub>)</td>
</tr>
<tr>
<td valign="top" align="left">&#x00A0;&#x00A0;&#x00A0;&#x00A0;&#x00A0;&#x00A0;&#x00A0;&#x00A0;&#x00A0;&#x00A0;<italic>i</italic><sub><italic>n</italic></sub> &#x02190; <italic>r</italic><sub>1</sub></td>
</tr>
<tr>
<td valign="top" align="left">&#x00A0;&#x00A0;&#x00A0;&#x00A0;create new population</td>
</tr>
</tbody>
</table>
</table-wrap>
</sec>
</sec>
<sec id="s4">
<title>4. Experimental Setup</title>
<p>We distinguish two different sets of experiments: the individual learning experiments, also called the <italic>baseline experiments</italic>, and the robot-to-robot learning experiments. Robot-to-robot learning experiments are performed with a group of 2 and 4 robots.</p>
<p>The baseline experiments are the 50 configurations given by the DoE. They are ranked based on the quality where after the best configurations (rank 1&#x02013;10) and the median configurations (rank 21&#x02013;30) are chosen to apply robot-to-robot learning too.</p>
<p>For all configurations, we use a fixed number of 20 k fitness evaluations. The total number of evaluations is the number of generations times the population size. Because the population size is a parameter in the configuration, we set the number of generations accordingly. As a result, 20 k evaluations correspond to 200 generations for a population size of 100, 334 generations for a population size of 60 and 1,000 generations for a population size of 20.</p>
<p>When robot-to-robot learning is applied, the robots have the same configuration as the one robot setup except for the population size. The population size for the robot-to-robot learning setup is the population size from the 1 robot setup divided by the number of robots to ensure the same number of evaluations per generation. Thus, if the original setup specifies a population size of 100, the robot-to-robot learning experiments use a population size of 50 and 25 for the 2 robots and 4 robot setup, respectively.</p>
<p>The robots operate in their own arena but they communicate across arenas. Consequently, the performance of the robot is only due to its own actions and not influenced by other robots in the same arena. Removing this inter-robot collision allows for a better comparison between the individual and the robot-to-robot learning experiments.</p>
<p>For all experiments, 20 replicate runs are performed with different random seeds.</p>
</sec>

<sec sec-type="results" id="s5">
<title>5. Results</title>

<sec>
<title>5.1. Baseline Experiments</title>
<p>The 50 configurations from the DoE are referred to as the baseline experiments. For the baseline experiments, there is only 1 robot, and it is learning individually. <xref ref-type="fig" rid="F2">Figure 2</xref> shows the quality and the interquartile range<xref ref-type="fn" rid="fn0003"><sup>3</sup></xref> for all baseline experiments. If the median of two ranks is equal, the one with the highest value for the third quartile gets the better ranking. The quality differences between the configurations are clearly present: the lowest ranked setting has a quality of 0, while the highest ranked setting reached a quality of 4.5. The data in <xref ref-type="fig" rid="F2">Figure 2</xref> confirms that the configuration, e.g., the values of the parameters, significantly influence the quality of the individual learning result.</p>
<fig id="F2" position="float">
<label>Figure 2</label>
<caption><p>Median performance with an interquartile range of the baseline experiments for all DoE parameter settings at the final generation. The <italic>y</italic>-axis shows the performance, measured as the median number of collected pucks in the population. The <italic>x</italic>-axis shows the rank of the configuration. The results are compiled over 20 replicate runs.</p></caption>
<graphic xlink:href="frobt-06-00010-g0002.tif"/>
</fig>
<p>The data in <xref ref-type="fig" rid="F2">Figure 2</xref> shows the performance at the final generation for all 50 parameter configurations. It is not clear whether the performance for all configurations converged atthis point. To show that the performance has approximately converged at this point for most parameter configurations, we observe the increase in performance for all configurations from 95 to 100% of the evaluation budget. The median of this increase over the 50 configurations is 0 with an interquartile range of [0,0.1875], i.e., most of the 50 parameter configurations converged when they reach 95% of the evaluation budget. These statistics confirm that we used a sufficient number of evaluations for our experiments.</p>
</sec>

<sec>
<title>5.2. Robot-to-Robot Learning: Performance</title>
<p><xref ref-type="fig" rid="F3">Figure 3</xref> shows the fitness of the individuals at the final generation for all 20 independent runs and all robot-to-robot learning experiments. To explain this graph, we start at the first column (parameter configuration 1) and the first row of the 20 runs for 1 robot. This row, or bar, goes from blue to red, and actually consists of many dots, where each dot represents one individual in the final generation and the color represents the fitness of the individual. The bar consists of 100 dots, representing the fitness value of the individuals from all robots.</p>
<fig id="F3" position="float">
<label>Figure 3</label>
<caption><p>Fitness of the individuals at the final generation for all 20 runs for 1 robot <bold>(top)</bold>, 2 robot <bold>(middle)</bold> and 4 robots <bold>(bottom)</bold>. The columns refer to the rank of the configuration (rank 1&#x02013;10 and 21&#x02013;30). The 20 runs are sorted on the sum of the fitnesses of the individuals. Within one run, the individuals are sorted on the fitness of which the color reflects the value. When using multiple robots, the individuals of the final generation for all robots are combined and sorted on fitness. Therefore, every bar from blue to red actually consists of 100 dots where every dot represents one individual.</p></caption>
<graphic xlink:href="frobt-06-00010-g0003.tif"/>
</fig>
<p>There are groups of 20 bars where each bar represents one run. The 20 runs are sorted on the sum of the fitnesses of the individuals. This block of 20 runs is shown for every combination of rank and the number of robots. The column refers to the rank of the configuration. Rank 1&#x02013;10 (best configurations) and 21 to 30 (median configurations) are present. The rows refer to the number of robots. When using multiple robots, the individuals of the final generation for all robots are combined and sorted by fitness.</p>
<p><xref ref-type="fig" rid="F3">Figure 3</xref> is an important graph to understand the effect of robot-to-robot learning. For the baseline experiments, we can observe the difference between configurations. Looking at the top row, the individual learning experiments, rank 1 to 10 shows more good performing individuals (red) and fewer bad performing individuals (blue) compared to rank 21&#x02013;30. This is not a surprise since the configuration rank is based on the median performance of the individuals. In general, the runs with a high performing individual, have more red and yellow colored individuals than a run without a good performing individual. This results in an increase of the median performance measure.</p>
<p>However, an increase of median performing individuals is not a necessity for an increase in a number of successful runs. Looking at rank 28, we can clearly see an increase of performance of the best individuals within a run but not a decrease in the number of bad performing individuals. Therefore, we can conclude that the benefits of robot-to-robot learning highly depend on the particular configuration.</p>
<p>Due to the replacement of the worst performing individual for the robot-to-robot learning experiments, one might expect that the number of bad performing individuals decreases. This effect does happen for some parameter settings, such as setting 24, but not for all. The number of generations, and thus the number of times you receive the best controller, exceeds the size of the population. Especially for a population size of 20 (e.g., rank 21&#x02013;25), there are 1,000 generations and thus 1,000 times to receive the best controller. However, in the final generation when the population size is 20, there are still bad performing robots. Therefore, we can say that simply replacing the worst controller with a betterone is not the reason that there are fewer badly performing controllers.</p>
<p><xref ref-type="fig" rid="F4">Figure 4</xref> summarizes the information from <xref ref-type="fig" rid="F3">Figure 3</xref> by using the measures introduced in section 2. <xref ref-type="fig" rid="F4">Figures 4A,B</xref> show the SRR and the PFR for the best and median configurations for the robot-to-robot learning experiment with 2 and 4 robots, respectively. A value higher than 1 means that a benefit due to robot-to-robot learning is observed for that particular measure.</p>
<fig id="F4" position="float">
<label>Figure 4</label>
<caption><p>Population failure ratio (PFR) (<italic>y</italic>-axis) and success rate ratio (SRR) (<italic>x</italic>-axis) for 2 <bold>(A)</bold> and 4 <bold>(B)</bold> robots. The color presents the rank of the configuration. A value higher than 1 means that a benefit is observed when using robot-to-robot learning for that particular measure.</p></caption>
<graphic xlink:href="frobt-06-00010-g0004.tif"/>
</fig>
<p>From <xref ref-type="fig" rid="F4">Figure 4</xref> we can conclude that the benefits of robot-to-robot learning are more present for 4 robots than for 2 robots, indicated by fewer observations below the thicker white lines and the higher values for both measurements. Furthermore, we can observe that every configuration can have completely different benefits. An SRR of 8 can be observed for the 4 robot setup for rank 25 and 28 while the PFR increases slightly for rank 28 and a lot for rank 25. This is consistent with the observation of <xref ref-type="fig" rid="F3">Figure 3</xref>.</p>
<p>Besides the influence of the configuration on the benefits of robot-to-robot learning, the group size also seems of influence. This can be further clarified when looking at rank 1 in depth. When applying robot-to-robot learning for 2 robots, the SR decreases from 50 to 40% while for 4 robots the SR increases from 50 to 75%.</p>
<p>Configuration 29 did not have any successful runs for the individual learning experiments. For both robot-to-robot learning experiments, with 2 and 4 robots, there were successful runs. Therefore, we can conclude that robot-to-robot learning can reach performance levels unreachable for the individual learning counterpart. This conclusion has also been presented in Jolley et al. (<xref ref-type="bibr" rid="B16">2016</xref>). Configuration 29 has been excluded from figure4 because the ratio cannot be calculated.</p>
</sec>

<sec>
<title>5.3. Robot-to-Robot Learning: Learning Speed</title>
<p><xref ref-type="fig" rid="F5">Figures 5A,B</xref> show the LSI when using robot-to-robot learning for 2 and 4 robots, respectively. A value higher than 0 means that applying robot-to-robot learning results in faster learning compared to individual learning.</p>
<fig id="F5" position="float">
<label>Figure 5</label>
<caption><p>Learning speed increase (LSI) when using robot-to-robot learning with 2 <bold>(A)</bold> and 4 <bold>(B)</bold> robots. The LSI is calculated by subtraction the LS of one robot from the LS when using more robots. The LS is a numerical integral over the median performance over time and therefore one value. A value higher than 0 means that a benefit in learning speed is observed when using robot-to-robot learning.</p></caption>
<graphic xlink:href="frobt-06-00010-g0005.tif"/>
</fig>
<p>This graph shows an increased learning speed for most parameter setting for both 2 and 4 robots, indicated by the LSI values above 0. Robot-to-robot learning with 4 robots has a higher increase than robot-to-robot learning with 2 robots. The LSI for 2 robots correlates with the LSI for 4 robots [Pearson&#x00027;s <italic>r</italic><sub>(20)</sub> = 0.47, <italic>p</italic> &#x0003D; 0.0435]. This indicates that if there is a learning speed when using 2 robots, this is still also the case for 4 robots.</p>
</sec>
</sec>

<sec sec-type="discussion" id="s6">
<title>6. Discussion</title>
<p>In this paper, we performed a detailed analysis into the effects of robot-to-robot learning. In particular, we investigated the effect of the parameter settings, or configurations, of the individual learning mechanism on the benefits of robot-to-robot learning. The benefits of robot-to-robot learning were measured in three ways: (1) the success rate, which is the percentage of runs that have a good controller in the final generation (2) the population failure, which is the median of the percentage of bad individuals in the final generation and (3) the learning speed, which is the numerical integral of the median performance over time.</p>
<p>In previous work Heinerman et al. (<xref ref-type="bibr" rid="B14">2017</xref>) we concluded that configurations leading to a median quality for the individual learning process benefit more from robot-to-robot learning. In that work, we used the mean as the performance measurement. Due to the more specific measurements in this paper, we can conclude that the increase of performance of the median quality configurations is due to the increase in success rate. This can be seen in <xref ref-type="fig" rid="F4">Figure 4</xref> where the median parameter settings show a higher increase in the success rate. As a result, we can conclude that if the parameter settings are not optimal, individual learning needs more luck to have a good controller in the final generation. When we apply robot-to-robot learning this effect is mitigated with more successful runs as a result.</p>
<p>The parameter configurations impact the conclusion of the benefits of robot-to-robot learning. This might explain the contradicting observations in literature. However, we also have shown that robot-to-robot learning reduces the sensitivity to the choice of parameters. As a result, this might mean that even though different parameter setting has been used in literature which leads to different results, we could still compare the final performance in literature.</p>
<p>Additionally, from <xref ref-type="fig" rid="F4">Figure 4</xref> we conclude that robot-to-robot learning can reduce the number of bad performing individuals in the population. This number is important for online evolutionary robotics because the robots learn while executing a task, i.e., it is desirable that all robots have a good fitness instead of only one best.</p>
<p>This paper led to new insights into the effect of robot-to-robot learning. Robot-to-robot learning can reduce the sensitivity of the learning process to the choice of parameters in two ways: increase in the number of successful runs and a decrease of bad performing individuals. These two effects are similar to the effects desirable when tuning parameters. When tuning parameters the goal is to increase the average, median or maximum performance over all experiments. Preferably, these measures have a low variability. Increasing the success rate increases the average and possibly the maximum and median performance measures. Decreasing the population failure increases the average and potentially the median performance measure. Therefore, robot-to-robot learning can potentially reduce parameter tuning efforts.</p>
<p>We have seen that robot-to-robot learning can result in different benefits. Some parameter settings experience a small loss in performance and/or learning speed but most parameter settings experience a large gain in performance. While having observed the benefits in more detail, the question remains how the exchange of information results in these benefits. Due to the replacement of the worst performing individual for the robot-to-robot learning experiments, one might expect that the number of bad performing individuals decreases. But, we argued that this is not the reason due to the much higher number of controller exchanges compared to the population size. We believe it is an interplay between receiving new knowledge from other robots (resulting from a different search process) and less aggressive variation operators within one robot. This leads to more diverse quality solutions without increasing the effect of variation operators. However, we need extra data to confirm this hypothesis and this is left for future work.</p>
<p>Our specific implementation choice of robot-to-robot learning, has some similarities with parallel EAs and island models. These are commonly known to increase diversity which results in an increase in performance (Gordon and Whitley, <xref ref-type="bibr" rid="B9">1993</xref>; Whitley et al., <xref ref-type="bibr" rid="B30">1999</xref>; Alba and Tomassini, <xref ref-type="bibr" rid="B1">2002</xref>; Park and Ryu, <xref ref-type="bibr" rid="B21">2010</xref>). However, no research explains where the increase in diversity comes from and how this impacts the performance. Research in these areas mostly focusses on the decrease in runtime due to smaller population sizes (Alba and Tomassini, <xref ref-type="bibr" rid="B1">2002</xref>). Measuring the runtime in evolutionary robotics is useless because the evaluation time of the robot is much larger than the computational effort. Additionally, the fitness function in evolutionary robotics is extremely stochastic. Although the results of both fields cannot be used to interpret each others work at the moment, we do believe that there are some common elements in parallel EAs, island models and robot-to-robot learning in evolutionary robotics. Especially, studying the effect of the number of parallels/robots on the diversity of the whole population is of interest to both fields. This will be investigated in future work.</p>
<p>Other interesting aspects of the benefits of robot-to-robot learning include the group size and task complexity. Our results indicate that using 4 robots results in more benefits than using 2 robots. This could potentially be due to the total number of controller exchanges between robots that varies depending on the number of generations. The influence of the frequency of controller exchanges on the speed of convergence in robot-to-robot learning can be taken into account when understanding the scalability of the group size.</p>
<p>Lastly, we should note that for the robot-to-robot learning experiments, we used the parameters of the 1 robot setup while these parameters might not be optimal for the multiple robot setups. A multiple robot setup itself can also benefit from parameter tuning, just as the parameters for the exchange of knowledge. In future research, we will study the effect of different parameter settings on the performance of the multi-robot setup. Additionally, one could change and or adapt the parameter settings per robot.</p>
</sec>

<sec sec-type="conclusions" id="s7">
<title>7. Conclusions</title>
<p>In this paper, we investigated the benefits of enabling robots to share knowledge with others. Existing literature in robot-to-robot learning typically compares individual learning with robot-to-robot learning for only one parameter setting. Our study extended this comparison by using 50 different parameter settings.</p>
<p>We showed that robot-to-robot learning can reduce the sensitivity of the learning process to the choice of parameters in two ways. First, robot-to-robot learning can reduce the number of bad performing individuals in the population. Second, robot-to-robot learning can increase the chance of having a successful run, where success is defined as the presence of a high performing individual. While some parameter settings experience a small decrease in performance, most parameter settings benefit greatly from robot-to-robot learning in terms of performance and/or learning speed.</p>
<p>Our results indicate that robot-to-robot learning is a powerful mechanism which leads to benefits in both performance and learning speed. Additionally, this paper showed the importance of an in-depth analysis to draw conclusions that are not possible with aggregated statistics. We hope to inspire others to use our proposed measurements for an in-depth analysis of components that have an impact on the benefits of robot-to-robot learning such as different tasks, the frequency of exchange, different group sizes and different environments.</p>
</sec>

<sec id="s8">
<title>Author Contributions</title>
<p>JH contributed to the programming and running of the experiments, the visualization of the results, and the writing of the manuscript. EH contributed to the supervision of the experiments and the writing of the manuscript. AE contributed to the supervising the experiments and writing of the manuscript.</p>
<sec>
<title>Conflict of Interest Statement</title>
<p>The authors declare that the research was conducted in the absence of any commercial or financial relationships that could be construed as a potential conflict of interest.</p>
</sec>
</sec>
</body>
<back>
<ack>
<p>This work was made possible by the European Union FET Proactive Initiative: Knowing, Doing, Being: Cognition Beyond Problem Solving funding the Deferred Restructuring of Experience in Autonomous Machines (DREAM) project under grant agreement 640891.</p>
</ack>
<ref-list>
<title>References</title>
<ref id="B1">
<citation citation-type="journal"><person-group person-group-type="author"><name><surname>Alba</surname> <given-names>E.</given-names></name> <name><surname>Tomassini</surname> <given-names>M.</given-names></name></person-group> (<year>2002</year>). <article-title>Parallelism and evolutionary algorithms</article-title>. <source>IEEE Trans. Evol. Comput.</source> <volume>6</volume>, <fpage>443</fpage>&#x02013;<lpage>462</lpage>. <pub-id pub-id-type="doi">10.1109/TEVC.2002.800880</pub-id></citation></ref>
<ref id="B2">
<citation citation-type="book"><person-group person-group-type="author"><name><surname>Bredeche</surname> <given-names>N.</given-names></name> <name><surname>Haasdijk</surname> <given-names>E.</given-names></name> <name><surname>Eiben</surname> <given-names>A. E.</given-names></name></person-group> (<year>2009</year>). <article-title>&#x0201C;On-line, on-board evolution of robot controllers,&#x0201D;</article-title> in <source>International Conference on Artificial Evolution (Evolution Artificielle)</source> (<publisher-loc>Berlin; Heidelberg</publisher-loc>: <publisher-name>Springer</publisher-name>), <fpage>110</fpage>&#x02013;<lpage>121</lpage>.</citation></ref>
<ref id="B3">
<citation citation-type="web"><person-group person-group-type="author"><name><surname>Curran</surname> <given-names>D.</given-names></name> <name><surname>ORiordan</surname> <given-names>C.</given-names></name></person-group> (<year>2007</year>). <article-title>Cultural learning in a dynamic environment: an analysis of both fitness and diversity in populations of neural network agents</article-title>. <source>J. Arti. Soc. Soc. Simul.</source> <volume>10</volume>, <fpage>1</fpage>&#x02013;<lpage>3</lpage>. Available online at: <ext-link ext-link-type="uri" xlink:href="http://jasss.soc.surrey.ac.uk/10/4/3.html">http://jasss.soc.surrey.ac.uk/10/4/3.html</ext-link></citation></ref>
<ref id="B4">
<citation citation-type="book"><person-group person-group-type="author"><name><surname>Duarte</surname> <given-names>M.</given-names></name> <name><surname>Silva</surname> <given-names>F.</given-names></name> <name><surname>Rodrigues</surname> <given-names>T.</given-names></name> <name><surname>Oliveira</surname> <given-names>S. M.</given-names></name> <name><surname>Christensen</surname> <given-names>A. L.</given-names></name></person-group> (<year>2014</year>). <article-title>&#x0201C;Jbotevolver: a versatile simulation platform for evolutionary robotics,&#x0201D;</article-title> in <source>Proceedings of the 14th International Conference on the Synthesis &#x00026; Simulation of Living Systems.</source> (<publisher-loc>Cambridge, MA</publisher-loc>: <publisher-name>MIT Press</publisher-name>), <fpage>210</fpage>&#x02013;<lpage>211</lpage>.</citation></ref>
<ref id="B5">
<citation citation-type="journal"><person-group person-group-type="author"><name><surname>Eiben</surname> <given-names>A. E.</given-names></name> <name><surname>Smith</surname> <given-names>J. E.</given-names></name></person-group> (<year>2015a</year>). <article-title>From evolutionary computation to the evolution of things</article-title>. <source>Nature</source> <volume>521</volume>, <fpage>476</fpage>&#x02013;<lpage>482</lpage>. <pub-id pub-id-type="doi">10.1038/nature14544</pub-id><pub-id pub-id-type="pmid">26017447</pub-id></citation></ref>
<ref id="B6">
<citation citation-type="journal"><person-group person-group-type="author"><name><surname>Eiben</surname> <given-names>A. E.</given-names></name> <name><surname>Smith</surname> <given-names>J. E.</given-names></name></person-group> (<year>2015b</year>). <source>Introduction to Evolutionary Computing</source>, 2nd Edn. <publisher-loc>Berlin; Heidelberg</publisher-loc>: <publisher-name>Springer</publisher-name>.</citation></ref>
<ref id="B7">
<citation citation-type="book"><person-group person-group-type="author"><name><surname>Fern&#x000E1;ndez P&#x000E9;rez</surname> <given-names>I.</given-names></name> <name><surname>Boumaza</surname> <given-names>A.</given-names></name> <name><surname>Charpillet</surname> <given-names>F.</given-names></name></person-group> (<year>2015</year>). <article-title>&#x0201C;Decentralized innovation marking for neural controllers in embodied evolution,&#x0201D;</article-title> in <source>Proceedings of the 2015 Annual Conference on Genetic and Evolutionary Computation</source> (<publisher-loc>Madrid: ACM</publisher-loc>), <fpage>161</fpage>&#x02013;<lpage>168</lpage>.</citation></ref>
<ref id="B8">
<citation citation-type="book"><person-group person-group-type="author"><name><surname>Garcia-Sanchez</surname> <given-names>P.</given-names></name> <name><surname>Eiben</surname> <given-names>A. E.</given-names></name> <name><surname>Haasdijk</surname> <given-names>E.</given-names></name> <name><surname>Weel</surname> <given-names>B.</given-names></name> <name><surname>Merelo-Guervos</surname> <given-names>J.-J.</given-names></name></person-group> (<year>2012</year>). <article-title>&#x0201C;Testing diversity-enhancing migration policies for hybrid on-line evolution of robot controllers,&#x0201D;</article-title> in <source>European Conference on the Applications of Evolutionary Computation</source> (<publisher-loc>Berlin; Heidelberg</publisher-loc>: <publisher-name>Springer</publisher-name>), <fpage>52</fpage>&#x02013;<lpage>62</lpage>.</citation></ref>
<ref id="B9">
<citation citation-type="book"><person-group person-group-type="author"><name><surname>Gordon</surname> <given-names>V. S.</given-names></name> <name><surname>Whitley</surname> <given-names>D.</given-names></name></person-group> (<year>1993</year>). <article-title>&#x0201C;Serial and parallel genetic algorithms as function optimizers,&#x0201D;</article-title> in <source>ICGA</source> (<publisher-loc>Urbana</publisher-loc>, <publisher-name>IL</publisher-name>), <fpage>177</fpage>&#x02013;<lpage>183</lpage>.</citation></ref>
<ref id="B10">
<citation citation-type="journal"><person-group person-group-type="author"><name><surname>Haasdijk</surname> <given-names>E.</given-names></name> <name><surname>Eiben</surname> <given-names>A. E.</given-names></name> <name><surname>Winfield</surname> <given-names>A. F.</given-names></name></person-group> (<year>2013</year>). <source>Chapter 12: Individual, Social and Evolutionary Adaptation in Collective Systems</source>. <publisher-loc>Singapore</publisher-loc>: <publisher-name>Pan</publisher-name> <volume>Stanford</volume>, <fpage>413</fpage>&#x02013;<lpage>471</lpage>.</citation></ref>
<ref id="B11">
<citation citation-type="journal"><person-group person-group-type="author"><name><surname>Haasdijk</surname> <given-names>E.</given-names></name> <name><surname>Smit</surname> <given-names>S. K.</given-names></name> <name><surname>Eiben</surname> <given-names>A. E.</given-names></name></person-group> (<year>2012</year>). <article-title>Exploratory analysis of an on-line evolutionary algorithm in simulated robots</article-title>. <source>Evol. Intel.</source> <volume>5</volume>, <fpage>213</fpage>&#x02013;<lpage>230</lpage>. <pub-id pub-id-type="doi">10.1007/s12065-012-0083-6</pub-id></citation></ref>
<ref id="B12">
<citation citation-type="journal"><person-group person-group-type="author"><name><surname>Heinerman</surname> <given-names>J.</given-names></name> <name><surname>Drupsteen</surname> <given-names>D.</given-names></name> <name><surname>Eiben</surname> <given-names>A. E.</given-names></name></person-group> (<year>2015a</year>). <article-title>&#x0201C;Three-fold adaptivity in groups of robots: The effect of social learning,&#x0201D;</article-title> in <source>Proceedings of the 17th Annual Conference on Genetic and Evolutionary Computation</source>, ed <person-group person-group-type="editor"><name><surname>Silva</surname> <given-names>S.</given-names></name></person-group> (<publisher-loc>New York, NY</publisher-loc>: <publisher-name>ACM</publisher-name>), GECCO <volume>15</volume>, <fpage>177</fpage>&#x02013;<lpage>183</lpage>.</citation></ref>
<ref id="B13">
<citation citation-type="book"><person-group person-group-type="author"><name><surname>Heinerman</surname> <given-names>J.</given-names></name> <name><surname>Rango</surname> <given-names>M.</given-names></name> <name><surname>Eiben</surname> <given-names>A. E.</given-names></name></person-group> (<year>2015b</year>). <article-title>&#x0201C;Evolution, individual learning, and social learning in a swarm of real robots,&#x0201D;</article-title> in <source>Proceedings-2015 IEEE Symposium Series on Computational Intelligence, SSCI 2015</source> (<publisher-loc>Capetown</publisher-loc>: <publisher-name>IEEE</publisher-name>), <fpage>1055</fpage>&#x02013;<lpage>1062</lpage>.</citation></ref>
<ref id="B14">
<citation citation-type="book"><person-group person-group-type="author"><name><surname>Heinerman</surname> <given-names>J.</given-names></name> <name><surname>Stork</surname> <given-names>J.</given-names></name> <name><surname>Coy</surname> <given-names>R.</given-names></name> <name><surname>Hubert</surname> <given-names>J.</given-names></name> <name><surname>Eiben</surname> <given-names>A. E.</given-names></name> <name><surname>Bartz-Beielstein</surname> <given-names>T.</given-names></name> <etal/></person-group>. (<year>2017</year>). <article-title>&#x0201C;Can social learning increase learning speed, performance or both?,&#x0201D;</article-title> in <source>Proceedings of the 14th European Conference on Artificial Life ECAL 2017</source> (<publisher-loc>Lyon</publisher-loc>).</citation></ref>
<ref id="B15">
<citation citation-type="book"><person-group person-group-type="author"><name><surname>Huijsman</surname> <given-names>R.-J.</given-names></name> <name><surname>Haasdijk</surname> <given-names>E.</given-names></name> <name><surname>Eiben</surname> <given-names>A. E.</given-names></name></person-group> (<year>2011</year>). <article-title>&#x0201C;An on-line on-board distributed algorithm for evolutionary robotics,&#x0201D;</article-title> in <source>Arti?cial Evolution, 10th International Conference Evolution Arti?cielle</source>, eds <person-group person-group-type="editor"><name><surname>Hao</surname> <given-names>J.-K.</given-names></name> <name><surname>Legrand</surname> <given-names>P.</given-names></name> <name><surname>Collet</surname> <given-names>P.</given-names></name> <name><surname>Monmarche</surname> <given-names>N.</given-names></name> <name><surname>Lutton</surname> <given-names>E.</given-names></name> <name><surname>Schoenauer</surname> <given-names>M.</given-names></name></person-group> (<publisher-loc>Angers</publisher-loc>: <publisher-name>Springer</publisher-name>), <fpage>73</fpage>&#x02013;<lpage>84</lpage>.</citation></ref>
<ref id="B16">
<citation citation-type="book"><person-group person-group-type="author"><name><surname>Jolley</surname> <given-names>B. P.</given-names></name> <name><surname>Borg</surname> <given-names>J. M.</given-names></name> <name><surname>Channon</surname> <given-names>A.</given-names></name></person-group> (<year>2016</year>). <article-title>&#x0201C;Analysis of social learning strategies when discovering and maintaining behaviours inaccessible to incremental genetic evolution,&#x0201D;</article-title> in <source>International Conference on Simulation of Adaptive Behavior</source> (<publisher-loc>Springer</publisher-loc>), <fpage>293</fpage>&#x02013;<lpage>304</lpage>.</citation></ref>
<ref id="B17">
<citation citation-type="book"><person-group person-group-type="author"><name><surname>Miikkulainen</surname> <given-names>R.</given-names></name> <name><surname>Feasley</surname> <given-names>E.</given-names></name> <name><surname>Johnson</surname> <given-names>L.</given-names></name> <name><surname>Karpov</surname> <given-names>I.</given-names></name> <name><surname>Rajagopalan</surname> <given-names>P.</given-names></name> <name><surname>Rawal</surname> <given-names>A.</given-names></name> <etal/></person-group>. (<year>2012</year>). <article-title>&#x0201C;Multiagent learning through neuroevolution,&#x0201D;</article-title> in <source>IEEE World Congress on Computational Intelligence</source> (<publisher-loc>Berlin; Heidelberg</publisher-loc>: <publisher-name>Springer</publisher-name>), <fpage>24</fpage>&#x02013;<lpage>46</lpage>.</citation></ref>
<ref id="B18">
<citation citation-type="journal"><person-group person-group-type="author"><name><surname>Montgomery</surname> <given-names>D. C.</given-names></name></person-group> (<year>2009</year>). <source>Design and Analysis of Experiments.</source> <publisher-name>John Wiley &#x00026; Sons, Inc</publisher-name>.</citation></ref>
<ref id="B19">
<citation citation-type="web"><person-group person-group-type="author"><name><surname>Noble</surname> <given-names>J.</given-names></name> <name><surname>Franks</surname> <given-names>D. W.</given-names></name></person-group> (<year>2012</year>). <article-title>Social learning in a multi-agent system</article-title>. <source>Comput. Inform.</source> <volume>22</volume>, <fpage>561</fpage>&#x02013;<lpage>574</lpage>. Available online at: <ext-link ext-link-type="uri" xlink:href="http://www.cai.sk/ojs/index.php/cai/article/view/469/376">http://www.cai.sk/ojs/index.php/cai/article/view/469/376</ext-link></citation></ref>
<ref id="B20">
<citation citation-type="journal"><person-group person-group-type="author"><name><surname>Nolfi</surname> <given-names>S.</given-names></name> <name><surname>Floreano</surname> <given-names>D.</given-names></name></person-group> (<year>2000</year>). <source>Evolutionary Robotics: The Biology, Intelligence, and Technology of Self-Organizing Machines</source>. <publisher-loc>London</publisher-loc>: <publisher-name>MIT Press</publisher-name>.</citation></ref>
<ref id="B21">
<citation citation-type="journal"><person-group person-group-type="author"><name><surname>Park</surname> <given-names>T.</given-names></name> <name><surname>Ryu</surname> <given-names>K. R.</given-names></name></person-group> (<year>2010</year>). <article-title>A dual-population genetic algorithm for adaptive diversity control</article-title>. <source>IEEE Trans. Evol. Comput.</source> <volume>14</volume>, <fpage>865</fpage>&#x02013;<lpage>884</lpage>. <pub-id pub-id-type="doi">10.1109/TEVC.2010.2043362</pub-id></citation></ref>
<ref id="B22">
<citation citation-type="book"><person-group person-group-type="author"><name><surname>Perez</surname> <given-names>A. L. F.</given-names></name> <name><surname>Bittencourt</surname> <given-names>G.</given-names></name> <name><surname>Roisenberg</surname> <given-names>M.</given-names></name></person-group> (<year>2008</year>). <article-title>&#x0201C;Embodied evolution with a new genetic programming variation algorithm,&#x0201D;</article-title> in <source>International Conference on Autonomic and Autonomous Systems</source> (<publisher-loc>Los Alamitos, CA, USA</publisher-loc>: <publisher-name>IEEE Press</publisher-name>), <fpage>118</fpage>&#x02013;<lpage>123</lpage>.</citation></ref>
<ref id="B23">
<citation citation-type="journal"><person-group person-group-type="author"><name><surname>Pugh</surname> <given-names>J.</given-names></name> <name><surname>Martinoli</surname> <given-names>A.</given-names></name></person-group> (<year>2009</year>). <article-title>Distributed scalable multi-robot learning using particle swarm optimization</article-title>. <source>Swarm Intell</source>. <volume>3</volume>, <fpage>203</fpage>&#x02013;<lpage>222</lpage>. <pub-id pub-id-type="doi">10.1007/s11721-009-0030-z</pub-id></citation></ref>
<ref id="B24">
<citation citation-type="book"><person-group person-group-type="author"><name><surname>Silva</surname> <given-names>F.</given-names></name> <name><surname>Correia</surname> <given-names>L.</given-names></name> <name><surname>Christensen</surname> <given-names>A. L.</given-names></name></person-group> (<year>2015</year>). <article-title>&#x0201C;A case study on the scalability of online evolution of robotic controllers,&#x0201D;</article-title> in <source>Portuguese Conference on Artificial Intelligence</source> (<publisher-loc>Cham</publisher-loc>: <publisher-name>Springer International Publishing</publisher-name>), <fpage>189</fpage>&#x02013;<lpage>200</lpage>.</citation></ref>
<ref id="B25">
<citation citation-type="journal"><person-group person-group-type="author"><name><surname>Silva</surname> <given-names>F.</given-names></name> <name><surname>Urbano</surname> <given-names>P.</given-names></name> <name><surname>Oliveira</surname> <given-names>S.</given-names></name> <name><surname>Christensen</surname> <given-names>A. L.</given-names></name></person-group> (<year>2012</year>). <article-title>odneat: An algorithm for distributed online, onboard evolution of robot behaviours</article-title>. <source>Artif. Life</source> <volume>13</volume>, <fpage>251</fpage>&#x02013;<lpage>258</lpage>. <pub-id pub-id-type="doi">10.7551/978-0-262-31050-5-ch034</pub-id></citation></ref>
<ref id="B26">
<citation citation-type="journal"><person-group person-group-type="author"><name><surname>Stanley</surname> <given-names>K. O.</given-names></name> <name><surname>Miikkulainen</surname> <given-names>R.</given-names></name></person-group> (<year>2002</year>). <article-title>Evolving neural networks through augmenting topologies</article-title>. <source>Evol. Comput.</source> <volume>10</volume>, <fpage>99</fpage>&#x02013;<lpage>127</lpage>. <pub-id pub-id-type="doi">10.1162/106365602320169811</pub-id><pub-id pub-id-type="pmid">12180173</pub-id></citation></ref>
<ref id="B27">
<citation citation-type="journal"><person-group person-group-type="author"><name><surname>Sutton</surname> <given-names>R. S.</given-names></name> <name><surname>Barto</surname> <given-names>A. G.</given-names></name></person-group> (<year>1998</year>). <source>Reinforcement Learning : An Introduction</source>. <publisher-loc>London</publisher-loc>: <publisher-name>MIT Press</publisher-name>.</citation></ref>
<ref id="B28">
<citation citation-type="journal"><person-group person-group-type="author"><name><surname>Tansey</surname> <given-names>W.</given-names></name> <name><surname>Feasley</surname> <given-names>E.</given-names></name> <name><surname>Miikkulainen</surname> <given-names>R.</given-names></name></person-group> (<year>2012</year>). <article-title>&#x0201C;Accelerating evolution via egalitarian social learning,&#x0201D;</article-title> in <source>Proceedings of the 14th Annual Conference on Genetic and Evolutionary Computation</source>, ed <person-group person-group-type="editor"><name><surname>Soule</surname> <given-names>T.</given-names></name></person-group> (<publisher-loc>New York, NY</publisher-loc>: <publisher-name>ACM</publisher-name>), <fpage>919</fpage>&#x02013;<lpage>926</lpage>.</citation></ref>
<ref id="B29">
<citation citation-type="book"><person-group person-group-type="author"><name><surname>Usui</surname> <given-names>Y.</given-names></name> <name><surname>Arita</surname> <given-names>T.</given-names></name></person-group> (<year>2003</year>). <article-title>&#x0201C;Situated and Embodied Evolution in Collective Evolutionary Robotics,&#x0201D;</article-title> in <source>Proceedings of the 8th International Symposium on Arti?cial Life and Robotics</source> (<publisher-loc>Beppu</publisher-loc>; <publisher-name>Oita</publisher-name>), <fpage>212</fpage>&#x02013;<lpage>215</lpage>.</citation></ref>
<ref id="B30">
<citation citation-type="journal"><person-group person-group-type="author"><name><surname>Whitley</surname> <given-names>D.</given-names></name> <name><surname>Rana</surname> <given-names>S.</given-names></name> <name><surname>Heckendorn</surname> <given-names>R. B.</given-names></name></person-group> (<year>1999</year>). <article-title>The island model genetic algorithm: On separability, population size and convergence</article-title>. <source>J. Comput. Inform. Technol.</source> <volume>7</volume>, <fpage>33</fpage>&#x02013;<lpage>47</lpage>.</citation></ref>
<ref id="B31">
<citation citation-type="journal"><person-group person-group-type="editor"><name><surname>Wiering</surname> <given-names>M.</given-names></name> <name><surname>van Otterlo</surname> <given-names>M.</given-names></name></person-group> (eds.). (<year>2012</year>). <source>Reinforcement Learning: State of the Art</source>. <publisher-loc>Berlin</publisher-loc>: <publisher-name>Springer</publisher-name>.</citation>
</ref>
<ref id="B32">
<citation citation-type="book"><person-group person-group-type="author"><name><surname>Zhang</surname> <given-names>P.</given-names></name> <name><surname>Ma</surname> <given-names>X.</given-names></name> <name><surname>Pan</surname> <given-names>Z.</given-names></name> <name><surname>Li</surname> <given-names>X.</given-names></name> <name><surname>Xie</surname> <given-names>K.</given-names></name></person-group> (<year>2010</year>). <article-title>&#x0201C;Multi-agent cooperative reinforcement learning in 3d virtual world,&#x0201D;</article-title> in <source>International Conference in Swarm Intelligence</source> (<publisher-loc>Berlin; Heidelberg</publisher-loc>: <publisher-name>Springer</publisher-name>), <fpage>731</fpage>&#x02013;<lpage>739</lpage>.</citation></ref>
</ref-list>
<app-group>
<app id="A1">
<title>Appendix</title>
<sec>
<title>Neat Parameters</title>
<p>The NEAT parameters with description and values (low, middle and high) for the DoE. Low, middle (mid) and high levels are used for numeric parameters are shown in the table below.</p>
<p><italic>Mutation and crossover parameters</italic> _______________________________________</p>
<table-wrap position="float">
<table frame="hsides" rules="groups">
<thead><tr>
<th/>
<th valign="top" align="left">Low</th>
<th valign="top" align="center">Mid</th>
<th valign="top" align="left">High</th>
</tr>
</thead>
<tbody>
<tr>
<td valign="top" align="left"><bold>p</bold><sub><bold>Xover</bold></sub></td>
<td valign="top" align="left"><monospace>0.05</monospace></td>
<td valign="top" align="center"><monospace>0.20</monospace></td>
<td valign="top" align="left"><monospace>0.35</monospace></td>
</tr>
<tr>
<td valign="top" align="left">chance to apply crossover</td>
<td/>
<td/>
<td/>
</tr>
<tr>
<td valign="top" align="left"><bold>p</bold><sub><bold>Mutation</bold></sub></td>
<td valign="top" align="left"><monospace>0.1</monospace></td>
<td valign="top" align="center"><monospace>0.25</monospace></td>
<td valign="top" align="left"><monospace>0.4</monospace></td>
</tr>
<tr>
<td valign="top" align="left">chance to apply mutation on each node/link</td>
<td/>
<td/>
<td/>
</tr>
<tr>
<td valign="top" align="left"><bold>p</bold><sub><bold>WeightReplaced</bold></sub></td>
<td valign="top" align="left"><monospace>0.0</monospace></td>
<td valign="top" align="center"><monospace>0.25</monospace></td>
<td valign="top" align="left"><monospace>0.5</monospace></td>
</tr>
<tr>
<td valign="top" align="left">chance to replace weight</td>
<td/>
<td/>
<td/>
</tr>
<tr>
<td valign="top" align="left"><bold>max</bold><sub><bold>Perturb</bold></sub></td>
<td valign="top" align="left"><monospace>0.25</monospace></td>
<td valign="top" align="center"><monospace>0.5</monospace></td>
<td valign="top" align="left"><monospace>0.75</monospace></td>
</tr>
<tr>
<td valign="top" align="left">maximum allowed change on weight</td>
<td/>
<td/>
<td/>
</tr>
<tr>
<td valign="top" align="left"><bold>p</bold><sub><bold>AddLink</bold></sub></td>
<td valign="top" align="left"><monospace>0.01</monospace></td>
<td valign="top" align="center"><monospace>0.05</monospace></td>
<td valign="top" align="left"><monospace>0.1</monospace></td>
</tr>
<tr>
<td valign="top" align="left">chance to add a link</td>
<td/>
<td/>
<td/>
</tr>
<tr>
<td valign="top" align="left"><bold>p</bold><sub><bold>AddNode</bold></sub></td>
<td valign="top" align="left"><monospace>0.01</monospace></td>
<td valign="top" align="center"><monospace>0.03</monospace></td>
<td valign="top" align="left"><monospace>0.05</monospace></td>
</tr>
<tr>
<td valign="top" align="left">chance to add a node</td>
<td/>
<td/>
<td/>
</tr>
<tr>
<td valign="top" align="left"><italic>Species parameters</italic> _________________</td>
<td/>
<td/>
<td/>
</tr>
<tr>
<td valign="top" align="left"><bold>species</bold><sub><bold>Count</bold></sub></td>
<td valign="top" align="left"><monospace>3</monospace></td>
<td valign="top" align="center"><monospace>6</monospace></td>
<td valign="top" align="left"><monospace>9</monospace></td>
</tr>
<tr>
<td valign="top" align="left">Maximum number of species.</td>
<td/>
<td/>
<td/>
</tr>
<tr>
<td valign="top" align="left"><bold>max</bold><sub><bold>SpeciesAge</bold></sub></td>
<td valign="top" align="left"><monospace>6</monospace></td>
<td valign="top" align="center"><monospace>18</monospace></td>
<td valign="top" align="left"><monospace>30</monospace></td>
</tr>
<tr>
<td valign="top" align="left">maximum age of species</td>
<td/>
<td/>
<td/>
</tr>
<tr>
<td valign="top" align="left"><bold>coeff</bold><sub><bold>Excess</bold></sub></td>
<td valign="top" align="left"><monospace>0.5</monospace></td>
<td valign="top" align="center"><monospace>1.0</monospace></td>
<td valign="top" align="left"><monospace>1.5</monospace></td>
</tr>
<tr>
<td valign="top" align="left">used for species compatibility score</td>
<td/>
<td/>
<td/>
</tr>
<tr>
<td valign="top" align="left"><bold>coeff</bold><sub><bold>Disjoint</bold></sub></td>
<td valign="top" align="left"><monospace>0.5</monospace></td>
<td valign="top" align="center"><monospace>1.0</monospace></td>
<td valign="top" align="left"><monospace>1.5</monospace></td>
</tr>
<tr>
<td valign="top" align="left">used for species compatibility score</td>
<td/>
<td/>
<td/>
</tr>
<tr>
<td valign="top" align="left"><bold>coeff</bold><sub><bold>Weight</bold></sub></td>
<td valign="top" align="left"><monospace>0.1</monospace></td>
<td valign="top" align="center"><monospace>0.4</monospace></td>
<td valign="top" align="left"><monospace>0.7</monospace></td>
</tr>
<tr>
<td valign="top" align="left">used for species compatibility score</td>
<td/>
<td/>
<td/>
</tr>
<tr>
<td valign="top" align="left"><bold>threshold</bold></td>
<td valign="top" align="left"><monospace>0.3</monospace></td>
<td valign="top" align="center"><monospace>0.5</monospace></td>
<td valign="top" align="left"><monospace>0.7</monospace></td>
</tr>
<tr>
<td valign="top" align="left">used for species compatibility score</td>
</tr>
<tr>
<td valign="top" align="left"><bold>threshold</bold><sub><bold>Change</bold></sub></td>
<td valign="top" align="left"><monospace>0.01</monospace></td>
<td valign="top" align="center"><monospace>0.1</monospace></td>
<td valign="top" align="left"><monospace>0.2</monospace></td>
</tr>
<tr>
<td valign="top" align="left">used to change threshold value</td>
<td/>
<td/>
<td/>
</tr>
<tr>
<td valign="top" align="left"><bold>species</bold><sub><bold>AgeThreshold</bold></sub></td>
<td valign="top" align="left"><monospace>0.7</monospace></td>
<td valign="top" align="center"><monospace>0.75</monospace></td>
<td valign="top" align="left"><monospace>0.8</monospace></td>
</tr>
<tr>
<td valign="top" align="left">percentage of age to count as old</td>
<td/>
<td/>
<td/>
</tr>
<tr>
<td valign="top" align="left"><bold>species</bold><sub><bold>YouthThreshold</bold></sub></td>
<td valign="top" align="left"><monospace>0.2</monospace></td>
<td valign="top" align="center"><monospace>0.25</monospace></td>
<td valign="top" align="left"><monospace>0.3</monospace></td>
</tr>
<tr>
<td valign="top" align="left">percentage of age to count as young</td>
<td/>
<td/>
<td/>
</tr>
<tr>
<td valign="top" align="left"><bold>age</bold><sub><bold>Penalty</bold></sub></td>
<td valign="top" align="left"><monospace>0.5</monospace></td>
<td valign="top" align="center"><monospace>0.7</monospace></td>
<td valign="top" align="left"><monospace>0.9</monospace></td>
</tr>
<tr>
<td valign="top" align="left">fitness multiplier for old individual</td>
<td/>
<td/>
<td/>
</tr>
<tr>
<td valign="top" align="left"><bold>age</bold><sub><bold>Boost</bold></sub></td>
<td valign="top" align="left"><monospace>1.1</monospace></td>
<td valign="top" align="center"><monospace>1.25</monospace></td>
<td valign="top" align="left"><monospace>1.4</monospace></td>
</tr>
<tr>
<td valign="top" align="left">fitness multiplier for young individual</td>
<td/>
<td/>
<td/>
</tr>
<tr>
<td valign="top" align="left"><italic>Other parameters</italic> _________________</td>
<td/>
<td/>
<td/>
</tr>
<tr>
<td valign="top" align="left"><bold>size</bold></td>
<td valign="top" align="left"><monospace>20</monospace></td>
<td valign="top" align="center"><monospace>60</monospace></td>
<td valign="top" align="left"><monospace>100</monospace></td>
</tr>
<tr>
<td valign="top" align="left">population size of one robot</td>
<td/>
<td/>
<td/>
</tr>
<tr>
<td valign="top" align="left"><bold>survival</bold><sub><bold>Threshold</bold></sub></td>
<td valign="top" align="left"><monospace>0.1</monospace></td>
<td valign="top" align="center"><monospace>0.45</monospace></td>
<td valign="top" align="left"><monospace>0.8</monospace></td>
</tr>
<tr>
<td valign="top" align="left">top % individuals that can be parents</td>
<td/>
<td/>
<td/>
</tr>
<tr>
<td valign="top" align="left"><bold>copy</bold><sub><bold>Best</bold></sub></td>
<td valign="top" align="left"><monospace>TRUE</monospace></td>
<td valign="top" align="center"><monospace>FALSE</monospace></td>
<td/>
</tr>
<tr>
<td valign="top" align="left">clone best individual previous generation</td>
<td/>
<td/>
<td/>
</tr>
<tr>
<td valign="top" align="left"><bold>copy</bold><sub><italic><bold>BestEver</bold></italic></sub></td>
<td valign="top" align="left"><monospace>TRUE</monospace></td>
<td valign="top" align="center"><monospace>FALSE</monospace></td>
<td/>
</tr>
<tr>
<td valign="top" align="left">clone best individual so far.</td>
<td/>
<td/>
<td/>
</tr>
</tbody>
</table>
</table-wrap>
</sec>
</app>
</app-group>
<fn-group>
<fn id="fn0001"><p><sup>1</sup>created with the help of JMP software SAS Institute Inc, JMP, Version 11.1.0.</p></fn>
<fn id="fn0002"><p><sup>2</sup>The code is available on <ext-link ext-link-type="uri" xlink:href="https://github.com/ci-group/ECAL_SocialLearning">https://github.com/ci-group/ECAL_SocialLearning</ext-link>.</p></fn>
<fn id="fn0003"><p><sup>3</sup>The interquartile range fluctuates between configurations. This is due to the difference in population size. When the population size is small, the interquartile range is relatively large.</p></fn>
</fn-group>
</back>
</article>