REVIEW article

Front. Neurorobot., 06 July 2018

Volume 12 - 2018 | https://doi.org/10.3389/fnbot.2018.00035

A Survey of Robotics Control Based on Learning-Inspired Spiking Neural Networks

  • 1. Chair of Robotics, Artificial Intelligence and Real-time Systems, Department of Informatics, Technical University of Munich, Munich, Germany

  • 2. Department of Data and Computer Science, Sun Yat-Sen University, Guangzhou, China

Article metrics

View details

172

Citations

29,1k

Views

21,2k

Downloads

Abstract

Biological intelligence processes information using impulses or spikes, which makes those living creatures able to perceive and act in the real world exceptionally well and outperform state-of-the-art robots in almost every aspect of life. To make up the deficit, emerging hardware technologies and software knowledge in the fields of neuroscience, electronics, and computer science have made it possible to design biologically realistic robots controlled by spiking neural networks (SNNs), inspired by the mechanism of brains. However, a comprehensive review on controlling robots based on SNNs is still missing. In this paper, we survey the developments of the past decade in the field of spiking neural networks for control tasks, with particular focus on the fast emerging robotics-related applications. We first highlight the primary impetuses of SNN-based robotics tasks in terms of speed, energy efficiency, and computation capabilities. We then classify those SNN-based robotic applications according to different learning rules and explicate those learning rules with their corresponding robotic applications. We also briefly present some existing platforms that offer an interaction between SNNs and robotics simulations for exploration and exploitation. Finally, we conclude our survey with a forecast of future challenges and some associated potential research topics in terms of controlling robots based on SNNs.

1. Introduction

The mysterious biological intelligence of living creatures has long attracted us to explore their capabilities from perceiving, memorizing, to thinking, and then resulting in languages and behaviors. Nowadays, owing to the increasing efforts of mimicking those structural and functional principles, scientists have investigated how the brain, robot actuators, and sensors could work together to operate robots autonomously performing complex tasks, e.g., in the form of self-driving vehicles (Schoettle and Sivak, 2014), biomimetic robots (Ijspeert et al., 2007; Gong et al., 2016), collaborative industrial robots (Shen and Norrie, 1999). However, to acquire more autonomy and operate within the real world, robots should be further investigated with the following capacities: (1) perceiving their environments via sensors that typically deliver high-dimensional data; (2) processing redundant or sparse information with low response latency and energy efficiency; (3) behaving under dynamic and changing conditions, which requires a self-learning ability.

Meanwhile, neither traditional control strategies nor conventional artificial neural networks (ANNs) can meet those aforementioned needs. To be specific, traditional model-based control methods via numerical techniques, kinematics and dynamics approaches often fail to adapt to unknown situations (Ijspeert, 2008; Yu et al., 2014; Bing et al., 2017). On the other hand, conventional ANNs have difficulties in processing the high computational demands for a step further, despite the hardware progress that made large neural networks applicable to real-world problems. The main disadvantages are as follow. First, training artificial neural networks is time consuming (Krogh and Vedelsby, 1995) and can easily take multiple days for state-of-the-art architectures (Lee C. S. et al., 2016). Training large-scale networks is computationally expensive [AlphaGo 1,202 CPUs and 176 GPUs (Silver et al., 2016)], and running them typically produces high response latencies (Dong et al., 2009). Second, performing computations with large networks on traditional hardware usually consumes a lot of energy as well. In self-driving cars for example, this results in computational hardware configurations that consume a few thousand Watts compared to the human brain, which only needs around 20 Watts of Power (Drubach, 2000). Especially in mobile applications, these are considerable disadvantages, in which real-time responses are important and energy supply is limited.

In nature, information is processed using relatively small populations of spikes and their precise relative timing, which is sufficient to drive learning and behavior (VanRullen et al., 2005; Houweling and Brecht, 2008; Huber et al., 2008; Wolfe et al., 2010). Therefore, a promising solution to robotics control challenges could be given by spiking neural networks that mimic the underlying mechanisms of the brain much more realistically. Due to their functional similarity to the brain, SNNs have the capabilities for processing information and learning in a much better fashion, both in terms of energy and data, e.g., building large-scale brain model (Eliasmith et al., 2012) or using neurally inspired hardware such as the SpiNNaker board (Furber et al., 2014) or Dynamic Vision Sensors (DVS) (Lichtsteiner et al., 2008). Moreover, SNNs have offered solutions to a broad range of specific implementations, such as fast signal-processing (Rossello et al., 2014), speech recognition (Loiselle et al., 2005), robot navigation (Nichols et al., 2010), and other problems solved by non-spiking neural networks, but in fact showed even more superiorities. However, a comprehensive review on controlling robots based on spiking neural networks is still missing.

Therefore, in this article, we aim to survey the state-of-the-art SNN modeling, design, and training methods for controlling a variety of robotics applications since the recent decade. The overall motivation of this article is to ease the barrier for roboticists to understand the complicated biological knowledge of SNNs, meanwhile enlighten readers with some general learning-based SNN approaches to different robot control tasks. The contribution of this survey is three-fold. First, we try to set forth SNNs' superiorities in terms of speed, energy efficiency, and computation capabilities. And we outline a general design framework for controlling SNN-based robotics tasks. Then, our survey aims to summarize the learning-based SNNs in robotics tasks, ranging from the modeling, learning rules, and robot implementations and platforms. We generally categorize the selected robotic implementations according to different learning rules, this shows up in the source of learning signals, which could be acquired from different ways, such as the labeled dataset, neutral stimulus, rewards from environment or other external controllers. Finally, we attempt to point out the open topics that need to be addressed for implementing SNNs in robotics tasks.

The rest of the article is organized as follows. In section 2, the theoretical background of SNNs will be briefly introduced, which will include their biological foundations and the changing course of the artificial neural networks on a basis of learning rules. Section 3 presents the primary motivation and research framework for SNN-based robot control. Then we will discuss SNN implementations from the simple neuron unit to the topologies of more advanced systems (section 4). Various methods training SNNs for control tasks will be classified and explained with their corresponding robotic applications (section 5), as well as the existing platforms for exploring neurorobotics (section 6). Finally we will summarize future challenges and potential research topics in section 7 and conclude in section 8.

2. Theoretical background

Before studying in deep of the robotics control based on SNNs, it is worth briefly summarizing the biological mechanisms taking place in human nervous system. Therefore, this section serves as a short summary of the theoretical foundations as well as the vocabulary that is used in the following sections. An in-depth introduction of SNNs can be found in Vreeken (2003), Ghosh-Dastidar and Adeli (2009), Ponulak and Kasinski (2011), and Grüning and Bohte (2014).

2.1. Biological background

Even today's understanding of the human brain remains rather incomplete and challenging, some insights into our neural structure have been made over the past couple of decades. Since the initial discovery of neurons as the basic structure of the nervous system by Santiago Ramón y Cajal at the beginning of the twentieth century, a rough concept of how neurons might work has been developed. At the very basis, neurons can be understood as simple building blocks processing incoming information in the form of short pulses of electrical energy into output signals. By connecting neurons to huge networks, complex dynamics emerge that can process information and make sense of our world. This basic concept can be found all over nature, ranging from simpler organisms like jellyfish with a couple of thousand neurons to humans with an estimated number of 86 billion neurons on average in our nervous system (Herculano-Houzel, 2012).

The structure of a typical neuron of the human brain embedded in a salty extra-cellular fluid is shown in Figure 1A. Incoming signals from multiple dendrites alter the voltage of the neuronal membrane. When a threshold is reached, the cell body or soma sends out an action potential—also called spike or pulse—itself. This process of generating a short (1ms) and sudden increase in voltage is usually referred to as spiking or firing of a neuron. After firing, it is followed by a short inactive period called the refractory period, in which the neuron cannot send out other spikes regardless of any incoming signals.

Figure 1

Once the membrane potential threshold has been reached and the neuron fires, the generated output spike is transmitted via the axon of a neuron. These can grow quite long and branch out to a multitude of other nervous cells at the end.

At the end of an axon, the axon terminal, incoming signals are transmitted to other nervous cells, such as other neurons or muscular cells. There is now proven evidence that synapses are in fact one of the most complicated part of a neuron. On top of transmitting information, they work as a signal pre-processor and play a crucial part in learning and adaption for many neuroscientific models. When a traversing spike reaches an axon terminal, it can cause a synaptic vesicle to migrate toward the presynaptic membrane, as shown in Figure 1B. At the presynaptic membrane, the triggered vesicle will fuse with the membrane and release its stored neurotransmitters into the synaptic cleft filled with the extra-cellular fluid. After diffusing into this gap, neurotransmitter molecules have to reach a matching receptor at the postsynaptic side of the gap and bind with them. Directly or indirectly, this causes postsynaptic ion-channels to open or close. The resulting ion flux initiates a cascade that traverses the dendritic tree down to the trigger zone of the soma, changing the membrane potential of the postsynaptic cell. Therefore, different neurotransmitters can have opposing effects on the excitability of postsynaptic neurons, thus mediating the information transfer. These effects that make postsynaptic cells either more or less likely to fire action potentials are called excitatory postsynaptic potential or inhibitory postsynaptic potential, respectively. The dependence of postsynaptic potentials on different amounts and types of neurotransmitters released and the resulting number of ion-channels activated is often referred to as synaptic efficacy in short. After a while, neurotransmitter molecules are released again from their receptors into the synaptic cleft and either reabsorbed into the presynaptic axon terminal or decomposed by enzymes in the extra-cellular fluid.

The properties, characteristics, and capacities of synapses as signal pre-processors, e.g., chances of vesicle deployment or regeneration and amount of receptors, are not fixed, but always changing depending on the short and long-term history of its own and outside influences. Neuro-hormones in the extra-cellular fluid can influence both the pre and postsynaptic terminals temporarily, i.e., by enhancing vesicle regeneration or blocking neurotransmitters from activating ion-gate receptors. All these effects that change the influence of incoming spikes on the postsynaptic membrane potential are usually referred to as synaptic plasticity and form the basis of most models of learning in neuro and computer-sciences.

2.2. From mcculloch-pitts to backpropagation

In 1943, neurophysiologist Warren McCulloch and mathematician Walter Pitts wrote a theoretical paper on how neurons might work describing a simple neural network model using electrical circuits (McCulloch and Pitts, 1943). Capable of performing mathematical operations with boolean outputs, these first generation neural networks fire binary signals once a threshold of summed incoming signals is reached in a neuron. They have been successfully applied in powerful artificial neural networks such as multi-layer perceptrons and Hopfield nets (Hopfield, 1982).

With the advent of more powerful computing, this concept was later extended by introducing continuous activation functions, e.g., sigmoid (Han and Moraga, 1995) or hyperbolic tangent functions, to handle analog inputs and outputs as well. In fact, it has been proven that sufficiently large neural networks with continuous activation functions can approximate any analog function arbitrarily well by altering the network information flow through their synaptic weights (Hornik et al., 1989). The most commonly used and powerful supervised learning algorithm that takes advantage of continuous activation functions by using gradient-descent on an error function is called backpropagation (Hecht-Nielsen, 1992).

However, second generation neurons do not model electrical pulses that have been described in their biological counterparts, their analog information signals can actually be interpreted as an abstracted rate coding. Over a certain period of time, an averaging window mechanism can be used to code pulse frequencies into analog signals giving these models a more biologically plausible meaning.

2.3. Spiking neural networks

Following its biological counterpart, a third generation of neural networks (Maass, 1997, 2001) has been introduced that directly communicates by individual sequences of spikes. Instead of using abstracted information signals, they use pulse-coding mechanisms that allow for the incorporation of spatial-temporal information that would otherwise be lost by averaging over pulse frequencies. It becomes clear that these neural network models, referred to as Spiking-Neural-Networks (SNNs), can be understood as an extension to second generation neural networks and can, in fact, be applied to all problems solvable by non-spiking neural networks (Fiasché and Taisch, 2015). In theory, it has been shown that these models are even more computationally powerful than perceptrons and sigmoidal gates (Maass, 1997).

Due to their functional similarity to biological neurons (DasGupta and Schnitger, 1992), SNNs have become a scientific tool for analyzing brain processes, e.g., helping to explain how the human brain can process visual information in an incredibly short amount of time (Chun and Potter, 1995). Moreover, SNNs promise solutions for problems in applied engineering as well as power efficient, low-latency alternatives to second generation neural networks, e.g., for applications in robotics (Lee J. H. et al., 2016).

3. Primary motivation and framework

In this section, we will briefly introduce the research impetuses of SNN-based robotics control from multiple aspects. Core points and a major framework for generally organizing an SNN for robotic implementation are introduced as well.

3.1. Primary impetuses

As the third generation of the neural network model, SNNs have attracted more and more attention and gradually become an interdisciplinary research field for neuroscience as well as robotics. For clarity and simplicity, the fascinating features of SNNs, which apply well to robotic controllers, can be summarized as follows.

3.1.1. Biological plausibility

From the perspective of neuroscience, SNNs once again raise the level of biological realism by directly using individual sequences of spikes in communication and computation, like real neurons do (Ferster and Spruston, 1995). Experimental evidence accumulated during the last few years has indicated that many biological neural systems use the timing of single-action potentials (or “spikes”) to encode information (Maass, 1997), rather than the traditional rate-based models. In Walter et al. (2015a), it is explained that how the exact modeling of time in spiking neural networks serves as an important basis for powerful computation based on neurobiological principles.

3.1.2. Speed and energy efficiency

Despite the hardware upgrades that make large neural networks applicable to real-world problems, it usually does not apply to robotics platforms with limited energy and computing resources. Since SNNs are able to transmit and receive large volumes of data encoded by the relative timing of only a few spikes, this leads to the possibility of very fast and efficient implementations. For example, experiments have demonstrated that visual pattern analysis and pattern classification can be carried out by humans in just 100 ms, in spite of the fact that it involves a minimum of 10 synaptic stages from the retina to the temporal lobe (Thorpe et al., 2001). On the other hand, in terms of energy efficiency, maintaining sufficient working of the nervous system to perform various tasks requires a continuous energy supply (Sengupta and Stemmler, 2014). Yet, the human brain only needs very low power consumption, which is around 20 W of Power (Drubach, 2000).

3.1.3. Computational capabilities

Recently, established experiments in vivo have indicated that SNNs are capable of processing the information sufficiently using a relatively small number of spikes to drive learning and behavior (VanRullen et al., 2005; Houweling and Brecht, 2008; Huber et al., 2008; Wolfe et al., 2010); meanwhile, they can also handle a large-scale network containing up to a trillion neurons like elephants (Herculano-Houzel et al., 2014). Furthermore, SNNs are superior to non-spiking ones for utilizing the temporal information, referring to the precise timing of events to acquire the exact information with incredible precision and accuracy. For instance, the auditory system of the barn owl is able to locate sources of sound in the horizontal plane with a precision of 1 to 2 degrees, which equates to a temporal difference of only a few microseconds (5μs) between the arrival of sound waves at the left and right ears (Gerstner et al., 1999).

3.1.4. Information processing

Instead of using abstracted information signals, SNNs use pulse coding mechanisms that allow incorporating spatial-temporal information that would otherwise be lost by only averaging over pulse frequencies. This ability to learn and act in a dynamic environment, rich with temporal information, is a necessary quality for biological systems and for artificial systems that seek to perform similar tasks. Neurobiologists used weekly electric fish as a model to study the processing from stimulus encoding to feature extraction (Gabbiani et al., 1996; Metzner et al., 1998). They found that although pyramidal cells do not accurately convey detailed information about the time course of the stimulus, they reliably encode up- and down-strokes of random modulations by bursts of spikes. In addition, a problem referred to as “dynamic binding,” has at best remained elusive to implement in neural networks, referring to different types of sensor information together in an assembly. SNNs are able to efficiently detect conjunctions of primitives (features) anywhere on a-large-input grid in an efficient, position-invariant manner. Examples such as data classification and image recognition tasks can be found in Hopfield (1995), Thorpe et al. (2001), Bohte et al. (2002), Guyonneau et al. (2004), and Shin et al. (2010).

In conclusion, these interesting features make SNNs suitable for pursuing autonomy for robotics implementations. However, there is an implicit knowledge gap since SNNs are just investigated at the theoretical level, rather than widely adopted to practical robotics applications. Even so, the growing knowledge of spiking neural networks and their increasingly popularity consistently draw more research attention and have led to more and more SNN-based implementations, as illustrated in Figure 2.

Figure 2

3.2. Research directions

The major research directions of SNNs have focused on three aspects: SNN modeling, training, and implementations, which will be detailedly discussed or briefly introduced with other reference readings in the following sections. A general design framework for learning-inspired robot control is shown in Figure 3. Most robot control tasks chasing autonomy could be described as a cycle including three steps, namely, perception, decision, execution (Gspandl et al., 2012). Robots usually use their sensors and actuators to sense and interact with the environment. However, the SNNs can be regarded as the brains to make decision, which build up a bridge between perception and execution, by taking encoded information from environment and outputting decoded motor commands for robots. To be specific, first, the architecture and mathematical model of an SNN should be determined including the neuron and synapse. Neurons are known to be a major signaling unit of the nervous system, and synapses can be seen as signal transmitters that communicate among neurons. For this reason, modeling of an SNN is of great importance to characterize its properties. Then, the SNN should be initialized and trained with specific parameters and learning rules, as conventional neural networks. Choosing an appropriate learning rule directly impact the performance of the networks. For an SNN, the most common learning rule is the Hebbian rule, which will be explained in the following section. Finally, after training the SNN successfully, it should be validated in other scenarios and be optimized if necessary.

Figure 3

4. Modeling of spiking neural networks

At the very beginning of the construction of an SNN for robot control, an appropriate SNN control model should be decided on. The basic task is to determine the general topological structure of the SNN, as well as the neuron models in each layer of the SNN.

4.1. Neuron models

Generally, neuron models can be expressed in the form of ordinary differential equations. In the literature, many different mathematical descriptions of spiking neural models have been proposed, processing excitatory and inhibitory inputs using internal state variables. The most influential models used for SNNs are the Hodgkin-Huxley model (Hodgkin and Huxley, 1952) as well as the Integrate-and-Fire model and its variants (Burkitt, 2006). To find an appropriate one among existing diverse neuron models, there is usually a trade-off to be balanced between the biological plausibility and complexity. A detailed comparison of the neuro-computational properties of spiking and bursting models can be found in Izhikevich (2004).

One of the most widely used models is the so-called Leaky-Integrate-and-Fire (LIF) model (Stein, 1965) that can be easily explained by the principles of electronics. These models are based on the assumption that the timing of spikes, rather than the specific shape, carries neural information (Andrew, 2003). The sequences of firing times are called spike trains and can be described as

where f = 1, 2, … is the label of a spike and δ(·) is a Dirac function defined as

Passing a simplified synapse model, the incoming spike train will trigger a synaptic electric current into the postsynaptic neuron. This input signal i(t) induced by a presynaptic spike train Sj(t) can, in a simple form, be described by the exponential function (Ponulak and Kasinski, 2011):

Here, τs denotes the synaptic time constant. This synaptic transmission can be modeled by low-pass filter dynamics.

The postsynaptic current then charges the LIF neuron model increasing the membrane potential u according to

where τm = RC is the time constant of the neuron membrane, modeling the voltage leakage, depending on the resistance R. urest is the potential value after each reset. i0(t) denotes an external current driving the neural state, ij(t) is the input current from the jth synaptic input and wj represents the strength of the jth synapse. Once the membrane potential u reaches a certain firing threshold ϑ, the neuron fires a single spike and its membrane potential is set back to urest. Usually, this spiking event is followed by a refractory period in which the neuron stays inactive and can't be charged again.

It is worth pointing out that biological studies highlight the presence of another operational unit cell assemblies (Braitenberg, 1978) in the brain, which are defined as a group of neurons with strong mutual excitatory connections and tend to be activated as a whole. A deeper review of spiking neuron models can be found in Andrew (2003).

4.2. Information Encoding and Decoding

The term neural encoding refers to representing information from the physical world (such as direction of a moving stimulus) in the activity of a neuron (such as its firing rate). On the other hand, information decoding is a reverse process to interpret from neuron activity to electrical signal for actuators (such as muscle or motor). How the brain encodes information is to think of two spaces: the physical space and neural space. The physical space can be the physical properties of objects, such as color, speed, and temperature. Neural space consists of properties of a neuron, such as firing rate in most cases.

A number of neural information encoding methods have been proposed, such as binary coding (Gütig and Sompolinsky, 2006), population coding (Probst et al., 2012), temporal coding, and the most commonly used rate coding (Urbanczik and Senn, 2009; Wade et al., 2010). For binary coding, neurons are only modeled to take two values on/off , but it ignores the timed nature and multiplicity of spikes altogether. Due to is simplicity, this coding mechanism was used in early-stage implementations. Besides, binary coding is also used to represent pixel value of an image (Meschede, 2017). For rate coding, it is inspired by the observation that neurons tend to fire more often for stronger (sensory or artificial) stimulus. Scientists usually use a concept in probability theory known as the Poisson process to simulate spike trains that have characteristics close to real neurons. As the most intuitive and simple coding strategy, rate-coding has been adopted by most robotic implementations. For temporal coding, it is motivated by the evidence founded in neuroscience that spike-timing can be remarkably precise and reproducible Gerstner et al. (1996). With this encoding strategy, information is represented with the timing when the spike occurs. However, the underlying mechanism is still not so clear. The aforementioned coding solutions is usually for one single neuron. However, sometime a population of neurons is used as a whole to encode information. This is strongly supported by the brain of living creature, where functions are controlled by one area of neuron populations.

The goal of neural decoding is to characterize how the electrical activity of neurons elicit activity and responses in the brain. The most common used scheme for decoding is rate-based, where stronger neuron activity usually means higher motor speed or force. In Kaiser et al. (2016), a steering wheel model based on an agonist-antagonist muscle system was proposed according to the spike numbers of output neuron.

4.3. Synaptic plasticity models

Once the neuron model is decided on, the synapse model should be carefully chosen to connect those neurons inside and among the layers of SNNs. By influencing the membrane potentials of each connected neuron, synaptic plasticity was first proposed as a mechanism for learning and memory on the basis of theoretical analysis (Hebb, 1949). Up to this day, the synaptic plasticity models used for practical implementations are typically very simple. Based on an input-output relationship between neuronal activity and synaptic plasticity, they are roughly classified into two types, which are rate-based and spike based, that differ in the type of their input variables.

4.3.1. Rate-based

The first and most commonly used definition of a firing rate refers to a spike-count average over time (Andrew, 2003). The rate-based model is a popular approach for converting conventional ANNs into a spiking neural network that can still be trained by backpropagation. It has been successfully used in many aspects, especially in experiments on the sensory or motor system (Adrian, 1926; Bishop, 1995; Kubat, 1999; Kandel et al., 2000).

4.3.2. Spike-based

Spike-based learning rules were developed in Gerstner et al. (1993), Ruf and Schmitt (1997), Senn et al. (1997), Kempter et al. (1999), and Roberts (1999). Experiments showed that the synaptic plasticity is influenced by the exact timing of individual spikes, in particular, by their order (Markram et al., 1997; Bi and Poo, 1998). If a presynaptic spike preceded a postsynaptic spike, a potentiation of the synaptic strength could be observed, while the reversed order caused a depression. This phenomenon has been termed as Spike-Timing-Dependent-Plasticity (STDP) or anti-STDP for the exact opposite impact and explains the activity-dependent development of nervous systems. In other words, neural inputs that are likely to have contributed to the neurons' excitation are strengthened, while inputs that are less likely to have contributed are weakened. As for neuro-engineering, STDP has demonstrated to be successfully implemented as the underlying neural learning mechanism in robots and other autonomous systems in both simulated and real environments.

In the past, different mathematical models of STDP have been proposed, e.g., by Gerstner and Kistler (2002). For this work, the weight update rule under STDP as a function of the time difference between pre and postsynaptic spikes was defined as

with A+ and A representing positive constants scaling the strength of potentiation and depression, respectively. τ+ and τ are positive time constants defining the width of the positive and negative learning window. For deeper insights into the influence of the STDP mechanism, readers could refer to Song et al. (2000), Rubin et al. (2001), and Câteau and Fukai (2003).

A comparison of rate-based and spike-based spiking neural networks used for MNIST classification is shown in Diehl and Cook (2015).

4.4. Network models

The SNN network model resembles the synapse model in that it simulates synaptic interactions among neurons. Typical examples of neural networks consisting of neurons of these types are classified into two general categories:

4.4.1. Feed-forward networks

As the first and simplest type of network topology, information in feed-forward networks always travels from the input nodes, through hidden nodes (if any), to the output nodes and never goes backwards. In the biological nervous system, abstracted feed-forward networks are mainly found to acquire and transmit external information. Therefore, similarly, networks of this type are usually adopted for low-level sensory acquisition in robotic systems, such as vision (Perrinet et al., 2004), tactile sensing (Rochel et al., 2002), and olfaction (Cassidy and Ekanayake, 2006). For example, inspired by the structures and principles of primate visual cortex, Qiao et al. (2014, 2015, 2016) enhanced the feed-forward models including Hierarchical Max Pooling (HAMX) model and Convolutional Deep Belief Network (CDBN) with memory, association, active adjustment, semantic and episodic feature learning ability etc., and achieved good results in visual recognition task.

Taking the work from Meschede (2017) as an example, a two-layer feed-forward SNN was trained for a lane keeping vehicle. The control scheme is shown in Figure 4. In this work, the dynamic vision sensors (DVS) was used to detect the land markers by generating a sequence of events. The input layer consisted of 8 × 4 Poisson neurons and connected to the two LIF output motor neurons with R-STDP synapses in an “all to all” fashion. The learning phase was conducted by repeatedly training and switching the robot from the start positions in the inner and outer lanes. In comparison with other three learning methods, namely, the deep Q-learning (DQN), DQN-SNN, and Braitenberg Vehicle, the R-STDP SNN exhibited the best accuracy and adaptability in different lane scenarios.

Figure 4

4.4.2. Recurrent networks

Different from the feed-forward networks, recurrent neural networks (RNNs) transmit their information with a directed cycle and exhibit dynamic temporal behaviors. It is worth pointing out that recurrent neural networks are recursive neural networks (Wikipedia, 2017d) with a certain structure such as a linear chain. Living organisms seem to use this mechanism to process arbitrary sequences of inputs with their internal memory stored inside RNNs. As for robotics implementations, RNNs are widely used for vision (Kubota and Nishida, 2006), planning (Soula et al., 2005; Rueckert et al., 2016), and dynamic control (Probst et al., 2012).

In Rueckert et al. (2016), a recurrent SNN is proposed for solving planning tasks, which consists of two populations of neurons, namely, the state neuron population and the content neuron population. (see Figure 5) The state neuron population consists of K state neurons, which control all the state of a freely moving target. In their finite horizon planning task, the agent spatial position is controlled by nine state neurons. These state neurons are wired to each other and the content neuron populations by R-STDP synapse. The context neurons produce spatiotemporal spike patterns that represent high-level goals and context information. In this case, its average firing rate represents the target spatial position at different time step. A final reward is only received if the agent passes through two obstacles, one at time T/2 and one at time T. They show that the optimal planning policy can be learned using the reward modulated update rule in a network where the state neurons follow winner-take-all (WTA) dynamics. Due to the probability, in each time step exactly one state neuron is active and encodes the current position of the agent. Their results demonstrated a successful planner trajectory planning task using a recurrent SNN.

Figure 5

5. Learning and robotics applications

Changes in the strength of synaptic connections between neurons are thought to be the physiological basis of learning (Vasilaki et al., 2009). These changes can either be gated by neuromodulators that encode the presence of reward or inner co-activation among neurons and synapses. In control tasks presented in this section, the network is supposed to learn a function that maps some state input to a control or action output. When successfully learned, the network is able to perform simple tasks such as wall following, obstacle avoidance, target reaching, lane following, taxi behavior, or food foraging. In most cases, the network input directly comes from the robot's sensors, which range from binary sensors, e.g., olfactory, to multi-dimensional continuous sensors, such as cameras. In other cases, the input can be pre-processed data, e.g., coming from electroencephalography (EEG) data. Similarly, the output can range from one-dimensional, binary behavior control to multi-dimensional continuous output values, e.g., for motor control, as well.

Initially, solving simulated control tasks was done by manually setting network weights, e.g., in Lewis et al. (2000) and Ambrosano et al. (2016). However, this approach is limited to solving simple behavioral tasks such as wall following (Wang et al., 2009) or lane following (Kaiser et al., 2016), it is usually only feasible for very small network architectures with few weights.

Therefore, a variety of training methods for SNNs in control tasks has been researched and published. Instead of focusing on criteria such as field of research, biological plausibility or the specific task, this section is meant to serve as a classification of published algorithms into the basic underlying training mechanisms from a robotics and machine learning perspective. In the first part of this section, some implementations of SNN control are introduced that use some form of Hebbian-based learning. In the second part, publications are shown that try to bridge the gap between classical reinforcement learning and spiking neural networks. Finally, some alternative methods on how to train and implement spiking neural networks are discussed.

5.1. Hebbian-based learning

One of the earliest theories in neuroscience explaining the adaption of synaptic efficacies in the brain during the learning process was introduced by Donald Hebb in his 1949 book The Organization of Behavior (Hebb, 1949). Often summarized by the phrase “Cells that fire together, wire together,” his idea is usually expressed in mathematical terms as

where wij refers to the change of synaptic weight between the presynaptic neuron i and the postsynaptic cell j; and v represents the activities of those neurons, respectively.

Hebbian-based learning rule that rely on the precise timing of pre and post-synaptic spikes play a crucial part in the emergence of highly non-linear functions in SNNs. Learning based on Hebbs rule has been successfully applied to problems such as input clustering, pattern recognition, source separation, dimensionality reduction, formation of associative memories, or formation of self-organizing maps (Hinton and Sejnowski, 1999). Furthermore, different biologically plausible learning rules have been used for using Spiking Neural Networks in robot control tasks. However, as the basic underlying mechanism stays the same, training these networks can be achieved in different ways as follows (see Table 1). In the table, the two-wheel vehicle means a vehicle with two active wheels.

Table 1

Learning ruleRobotSensorMethodologyReference
UnsupervisedTwo-wheel vehicle5 Proximity SensorsImplementing an SNN on a resistive memory device and apply it to navigation tasksSarim et al., 2016a,b
Mobile vehicle Casis-I16 Ultrasonic SensorsA behavior-based target-approaching navigation controller composed of three sub-controllers: the obstacle-avoidance, wall-following, and goal-approaching SNN controllers.Wang et al., 2014, 2008
TriBot RobotDistance Sensor, Contact Sensor,Using SNN to make robot navigate in an unknown environment and avoid obstaclesArena et al., 2010
SupervisedTwo-wheel insect4 Proximity SensorsImplementing an indirect training SNN in digital CMOS to navigate with obstaclesHu et al., 2014; Mazumder et al., 2016
Two-wheel insect2 Terrain, 2 TargetIndirectly train an SNN by RBFs to determine precise spike timings and minimize a desired objective functionZhang et al., 2013
AircraftIMUIndirectly training an SNN to approximate an optimal flight controllerFoderaro et al., 2010
4-DoF robotic arm4 Joint Encoder, 3 Spatial direction of end-effectorUsing supervised learning to train a single-layer network to control a robotic arm with 4 degrees of freedom in 3D spaceBouganis and Shanahan, 2010
2-DoF robotic armSensorimotorUsing supervised learning to train a spiking model of the cerebellum to control a robotic armCarrillo et al., 2008
STDP / Hebbian Learning
ConditioningSimulated flyOlfactory ReceptorImplementing an SNN inspired by Drosophila olfactory system to simulate flightFaghihi et al., 2017
Lego EV3 robotic platformCamera, Infrared sensor Colour/light sensorLearning and unlearning autonomously locomotion based on visual-input with reinforced/aversive reflex-responseJimenez-Romero, 2017
Two-wheel robot3 Proximity Sensors, 1 RGB SensorUsing Reward-dependent STDP learning rule to allow OC and CC learningDumesnil et al., 2016a,b
Foraging AntsOlfactory Sensors, NociceptorLearns to associate olfactory sensor input with different behaviors through a single-layer SNNJimenez-Romero et al., 2015, 2016
Lego NXT 2.0Color sensor, Touch SensorUsing SNN to sustain OC in multiple learning scenariosCyr et al., 2014; Cyr and Thériault, 2015
Two-wheel VehicleLight SensorsUsing light sensors in a target-reaching task to punish wrongful behaviorIwadate et al., 2014
Two-wheel Vehicle5 Proximity Sensors, 9 IR Sensors, Vibration SensorUsing infrared, ultrasound and visual neurons as CS and vibration neurons as USCyr and Boukadoum, 2012
Mobile vehicle Casis-I16 Ultrasonic SensorsA learning algorithm combining operant conditioning and a shunting neural dynamics model is applied to the path planningWang et al., 2012
TriBot RobotDistance Sensor, Camera, Contact Sensor,Using target distance as CS, while contact sensors work as US causing an unconditioned responseArena et al., 2009a,b
R-STDPFlapping InsectGPS and IMUIndirectly training an SNN-based controller for adaptive flight controlClawson et al., 2016
1-DoF robotics arm5 Proximity SensorsUsing an SNN trained by a global reward and punishment signal to reach arbitrary targetsSpüler et al., 2015
Musculoskeletal arm, WAM robotEncodersUsing a cortical spiking model composed of several hundred spiking model-neurons to control a two-joint armDura-Bernal et al., 2015
CARL-SJRTactile SensorsUsing SNN to provide feedback to users by displaying bright colors on its surface.Chou et al., 2015
Two-wheel vehicle2 Proximity SensorsImplement a version of DA-modulated STDP on a food foraging taskEvans, 2015
Foraging SimulatorVisual SensorsUsing reward-STDP based SNN to solve a grid-based foraging taskSkorheim et al., 2014
DfRobotShop RoverCamera, Light SensorUsing an SNN and external flash to reinforce the goal-directed and adaptive behaviorsHelgadottir et al., 2013
2-DoF robotics armSensorimotorUsing an SNN based on R-STDP to control a two-joint virtual arm to reach to a fixed targetNeymotin et al., 2013
1-DoF robotics armEncoderUsing an SNN to control a single-joint arm for target reachingChadderdon et al., 2012

Learning rules based on STDP/Hebbian learning.

5.1.1. Unsupervised learning

According to STDP, if a presynaptic spike preceded a postsynaptic spike, a potentiation of the synaptic strength could be observed [Long Term Potentiation (LTP)], while the reversed order caused a depression [Long Term Depression (LTD)]. Because of the absence of direct goals, correction functions or a knowledgeable supervisor, this kind of learning is usually categorized as unsupervised learning (Hinton and Sejnowski, 1999). Learning based on STDP rule has been successfully applied to many problems such as input clustering, pattern recognition, and spatial navigation and mental exploration of the environment.

Wang et al. (2008) used this approach to train a behavior controller based on SNN to achieve obstacle avoidance using ultrasonic sensory signals with a mobile robot by driving it from different start positions. Compared with other classical NNs, they demonstrated that SNN needs fewer neurons and is relatively simple. Afterwards, they (Wang et al., 2014) extended their navigation controllers with with wall-following and goal-approaching abilities. In a similar research, Arena et al. (2010) presented an SNN based on an unsupervised learning paradigm to allow the robot to autonomously learn how to navigate in an unknown environment. Their controller allowed the robot to learn high-level sensor features, based on a set of basic reflexes, depending on some low-level sensor inputs by continuously strengthening the association between the unconditioned stimuli (contact and target sensors) and conditioned stimuli (distance and vision sensors).

5.1.2. Supervised learning

In non-spiking neural networks, many successes in recent years can be summarized as finding ways to efficiently learn from labeled data. This type of learning, where a neural network mimics a known outcome from given data is called supervised learning (Hastie et al., 2001). A variety of different neuroscientific studies has shown that this type of learning can also be found in the human brain (Knudsen, 1994), e.g., in motor control and motor learning (Thach, 1996; Montgomery et al., 2002). But despite the extensive exploration of these topics, the exact mechanisms of supervised learning in biological neurons remain unknown.

Accordingly, a simple way of training SNNs for robot control tasks is by providing an external training signal that adjusts the synapses in a supervised learning setting. As shown in Figure 6, when an external signal is induced into the network as a post-synaptic spike-train, the synapses can adjust their weights, for example, using learning rules such as STDP. After an initial training phase, this will cause the network to mimic the training signal with satisfactory precision. Even though this approach provides a simple, straight-forward way for training networks, it is dependent on an external controller. Especially for control tasks involving high-dimensional network inputs, this may not be feasible.

Figure 6

Several models have been proposed on how this might work, either by using activity templates to be reproduced (Miall and Wolpert, 1996) or error signals to be minimized (Kawato and Gomi, 1992; Montgomery et al., 2002). In the nervous system, these teaching signals might be provided by sensory feedback or other supervisory neural structures (Carey et al., 2005). One of these models that is primarily suitable for single-layer networks is called supervised Hebbian learning (SHL). Based on the learing rule derived in 8, a teaching signal is used to train the postsynaptic neuron to fire at target times and to remain silent at other times. It can be expressed as

where wij again is the synaptic efficacy between a presynaptic neuron i and a postsynaptic neuron j, α is the learning rate, vi is the presynaptic neurons activity and tj represents the postsynaptic teaching signal.

Carrillo et al. (2008) used this basic approach to train a spiking model of the cerebellum to control a robotic arm with 2 degrees of freedom in a target-reaching task taking joint angles and speeds, as well as target position as inputs. The spiking cerebellum model is trained by simulating the robotics arm to seven different targets repeatedly. In contrast to other STDP learning rules, only long-term depression was externally induced by a training signal, which relied on the motor error, namely the difference between the desired and actual state. In a similar experiment, Bouganis and Shanahan (2010) trained a single-layer network to control a robotic arm with 4 degrees of freedom in 3D space. As inputs, joint angles and the spatial During each training iteration, all four joints are driven with random motor commands, in the range of [−5°, 5°].direction of the end-effector were used, while outputs consisted of four motor-command neurons. The training signal was computed using an inverse kinematics model of the arm, adjusting the synaptic weights with a symmetric STDP learning rule. More examples can be found in Table 1 with an order by descending year.

5.1.3. Classical conditioning

Classical conditioning (Wikipedia, 2017a) refers to a learning procedure in which a biologically potent stimulus (e.g., food) is paired with a previously neutral stimulus (e.g., a bell). It will result that the neutral stimulus comes to elicit a response (e.g., salivation), which is usually elicited by the potent stimulus. In the famous experiment on classical conditioning (Pavlov and Anrep, 2003), Pavlov's dog learns to associate an unconditioned stimulus (US), in this case food, and a conditioned stimulus (CS), a bell, with each other. While, it is not clear how the high-level stimuli given in his experiment are processed within the brain, the same learning principle can be used for training on a neural level as well. Figure 7 shows how a synapse based on the STDP learning rule can associate US and CS provoking a response even in the absence of US.

Figure 7

Following this principle, bio-inspired robots can learn to associate a CS, e.g., sensory information, with a US that functions as an external reinforcer. That way, robots can learn to follow the desired behavior based on sensory inputs. Arena et al. (2009a,b, 2010) showed how classical conditioning can be used in an obstacle avoidance and target reaching task. In an SNN with two output motor neurons, distance and vision sensors function as CS, while contact and target sensors work as US causing an unconditioned response. By navigating the robot in a pre-designed enclosed environment, the robot successfully learned to associate the CS and the US together and reach the target without hitting obstacles. In a similar experiment, Cyr and Boukadoum (2012) carried out different classical conditioning tasks in a controlled virtual environment using infrared, ultrasound and visual neurons as CS and vibration neurons as US. Wang et al. (2008, 2014) constructed a controller that stimulated two motor neurons as US. A single-layer SNN using proximity sensor data as CS input was then trained in tasks such as obstacle avoidance and target reaching. Iwadate et al. (2014) used light sensors in a target-reaching task to punish wrongful behavior. Jimenez-Romero et al. (2015, 2016) and Jimenez-Romero (2017), implemented a virtual ant that learns to associate olfactory sensor input with different behaviors through a single-layer SNN. The robot was able to learn to recognize rewarding and harmful stimuli as well as simple navigation in a simulated environment. Casellato et al. (2014) proposed a realistic cerebellar SNN with a real haptic robotic arm to achieve diverse sensorimotor tasks. In all tasks, the robot learned to adjust timing and gain of the motor response and successfully reproduced human biological systems acquire, extinguish, and express knowledge in a noisy world.

In order to successfully learn such behavioral tasks, some unconditioned stimulus has to be given for every relevant conditioned stimulus that the robot should learn. This also means that the robot will learn to associate stimuli that are delayed in time. Taken together, using classical conditioning for robot control basically means constructing an external controller that provides unconditioned stimuli for every relevant state input, which may not be feasible in many tasks.

5.1.4. Operant conditioning

While classical conditioning is concerned with passively associating conditioned and unconditioned stimuli with each other, operant conditioning (OC) consists of associating stimuli with responses and actively changing behaviors thereafter. Conceptually, operant conditioning involves changing voluntary behaviors and is closely related to reinforcement learning and its agent-environment interaction cycle. A behavior response is followed by either reinforcement or punishment. Reinforcement following a behavior will cause the behavior to increase, but if behavior is followed by punishment the behavior will decrease. Instead of developing a formal mathematical model, operant conditioning has been mainly researched in biological and psychological domains. Despite advances in the understanding of operant conditioning, it is still not clear how this type of learning is implemented on a neural level.

In this context, Cyr et al. (2014) and Cyr and Thériault (2015) developed a spiking OC model that consists of an input feeding cue neuron, an action neuron and a predictor neuron that receives rewards or punishments. With this simple basic architecture and learning rules such as habituation and STDP, they were able solve simple OC-related tasks in a simulated environment, such as pushing blocks. In another publication by Dumesnil et al. (2016a,b) a reward-dependent STDP learning rule was implemented on a robot to allow for OC learning and demonstrated in a maze task. The RGB camera was used to capture the color information which represented the cue or the reward in the maze environment. Eventually, the robot learned the association, if an action was frequently followed by a reward.

5.1.5. Reward-modulated training

In Figure 8 the learning rule for reward-based training is shown. Using one or more chemicals emitted by a given neuron to regulate diverse populations of neurons is know as neuromodulation (Hasselmo, 1999). As one of the neuromodulators, dopamine neurons forming the midbrain dopaminergic cell groups are crucial for executive functions, motor control, motivation, reinforcement, and rewards. Most types of neurological rewards increase the level of dopamine in the brain, thus stimulating the dopamine neurons (Schultz, 1998). Inspired by dopaminergic neurons in the brain, the effects of STDP events are collected in an eligibility trace and a global reward signal induces synaptic weight changes. In contrast to supervised training as discussed before, rewards can be attributed to stimuli, even if they are delayed in time. This can be a very useful property for robot control, because it might simplify the requirements of an external training signal leading to more complex tasks. A simple learning rule combing models of STDP and a global reward signal was proposed by Florian (2007) and Izhikevich (2007). In the R-STDP, the synaptic weight w changes with the reward signal R. The eligibility trace of a synapse can be defined as,

where c is an eligibility trace. spre/post means the time of a pre- or post-synaptic spikes. C1 is a constant coefficient. τc is a time constant of the eligibility trace. δ(·) is the Dirac delta function.

where R(t) is the reward signal.

Figure 8

In the literature, a variety of algorithms has been published using this basic learning architecture for training. Even though they are all based on the same mechanism, the rewards can be constructed in different ways.

  • Rewarding Specific Events: The most straight-forward implementation of reward-based learning resembling classical reinforcement learning tasks uses rewards associated to specific events. Evans (2012) trained a simple, single-layer SNN in several food foraging tasks consisting of 4 input sensor neurons and 4 output motor neurons. In a separate network, other reward-related sensor neurons stimulated a dopaminergic neuron that in turn modulated the synaptic weight change. With this simulation setup, the robot was able to learn food-attraction behavior and subsequently unlearn this behavior when the environment changed. This was achieved by a training stage during which the robot were randomly droved to explore the environment effectively. By shifting the dopamine response from the primary to a secondary stimulus, the robot was able to learn, even with large temporal distance, between correct behavior and reward. Faghihi et al. (2017) showed an SNN model of a fruit fly that is able to execute both first and second order conditioning. In a simple task, the simulated fly learned to avoid getting close to an olfactory target emitting electric shocks. Furthermore, the same behavior can be transferred to a secondary stimulus that is associated to the primary stimulus without emitting electric shocks itself.

  • Control Error Minimization: As opposed to rewarding specific events, dopamine-modulated learning can also be used in an optimization task to minimize an objective function. This is usually achieved by strengthening or weakening the connections that lead to changes in the objective function based on their eligibility traces. Clawson et al. (2016) used this basic architecture to train an SNN to follow a trajectory. The network consisted of lateral state variables as inputs, a hidden layer and an output layer population decoding the lateral control output. Learning is achieved offline by minimizing the error between decoded actual and desired output, which is provided by an external linear controller.

  • Indirect Control Error Minimization: For some potential applications of SNNs, e.g., neuroprosthetic devices implanted in the brain, direct manipulation of synaptic weights might not be possible. Therefore, an indirect approach to training SNNs was shown by Foderaro et al. (2010) that induces changes in the synaptic efficacy through input spikes generated by a separate critic SNN. This external network was provided with control input as well as feedback signals and trained using a reward-based STDP learning rule. By minimizing the error between control output and optimal control law offline, it was able to learn adaptive control of an aircraft. Similar ideas were presented by Zhang et al. (2012), Zhang et al. (2013), Hu et al. (2014), and Mazumder et al. (2016) who trained a simple, virtual insect in a target reaching and obstacle avoidance task.

  • Metric Minimization: The same principle can also be applied to minimize a global metric that might be easier to construct and calculate than an external controller. Chadderdon et al. (2012) proposed a spiking-neuron model of the motor cortex which controlled a single-joint arm in a target-reaching task. The model consisted of 144 excitatory and 64 inhibitory neurons with proprioceptive inputs cells and output cells controlling the flexor and extensor muscles. A global reward or punishment signal was given depending on the change of hand-target distance during the learning phase, during which the robot was set with five different targets repeatedly. Neymotin et al. (2013) and Spüler et al. (2015) extended this architecture with two-joint robotic arm later. Similarly, Dura-Bernal et al. (2015) used a biomimetic cortical spiking model composed of several hundred spiking model-neurons to control a two-joint arm. With proprioceptive sensory input (muscle lengths) and muscle excitation output, the network was trained by minimizing the hand-target distance. The distance error was reduced by repeatedly move to the target with the guidance of a global reward/punish signal. Kocaturk et al. (2015) extended the same basic architecture in order to develop a brain-machine interface. Extracellularly recorded motor cortical neurons provide the network inputs used for prosthetic control. By pressing a button, the user can reward desired movements and guide the prosthetic arm toward a target. Using a miniaturized microprocessor with resistive crossbar memories implemented on a two-wheeled differential drive robot, Sarim et al. (2016a,b) showed how an STDP-based learning rule could lead to target approaching and obstacle avoidance behavior. Although, in this case, learning was implemented using if-then rules that relied on distance changes from target and obstacles, it is conceptually identical to reward-modulated learning. This can easily be seen by exchanging the if-rules with a reward of +1 or −1.

  • Reinforcing Associations: Chou et al. (2015) introduced a tactile robot that uses a network architecture inspired by the insular cortex. As in classical conditioning, a dopamine-modulated synaptic plasticity rule was used to reinforce associations between conditioned and unconditioned stimuli.

5.2. Reinforcement learning

In the previous subsection, a variety of approaches was presented for training SNNs based on Hebbian learning rules. This was done either by providing a supervised training signal through an external controller or by using a reward-based learning rule with different ways of constructing the reward. The latter type of learning, however, was shown to successfully train SNNs in simple tasks solely based on delayed rewards. In general, all of these approaches have been trained in tasks that don't require looking very far ahead, as reinforcement learning theories usually do.

In classical reinforcement learning theory, on the other hand, learning to look at multiple steps in advance in a Markov Decision Process (MDP) is one of the main concerns. Therefore, several algorithms have been published combining SNNs with classical reinforcement learning algorithms.

5.2.1. Temporal difference

The learning rule in which one looks at one or more steps forward in time was introduced as temporal difference (TD) learning. Hereby, Potjans et al. (2009) and Frémaux et al. (2013) used place cells to represent the state space in an MDP and single-layer SNNs for state evaluations and policies. Both algorithms were able to learn to navigate in a simple grid-world after some training. With a similar approach, Nichols et al. (2013) presented a robot controller inspired by the control structures of biological systems. In a self-organizing, multi-layered network structure, sensory data coming from distance and orientation sensors was gradually fused into state neurons representing distinct combinations of sensory inputs. On top, each individual state neuron was connected to 3 output motor neurons. By fusing the sensory input into distinct state neurons and connecting them to action neurons, a simplified TD learning rule could be used to set each synaptic weight in the last layer individually, when the robot conducted a trial locomotion. Performance of this controller was demonstrated in a wall-following task.

While these state representations work very well for relatively small state spaces, they are usually bound to fail for larger, high-dimensional state spaces, since the TD method can only obtain the reward in several steps. Therefore, it is less stable and may converge to the wrong solution, especially for high-dimensional state spaces. In fact, these approaches can conceptually be seen as an SNN implementation of table-based Q-learning.

5.2.2. Model-based

Although for robot control tasks, such as those shown in this paper, model-free reinforcement learning methods seem favorable, two recent publications are at least worth mentioning that presented SNN implementations of model-based reinforcement learning algorithms. Rueckert et al. (2016) presented a recurrent spiking neural network for planning tasks that was demonstrated on a real robot in an obstacle avoidance task. Friedrich and Lengyel (2016) implemented a biologically realistic network of spiking neurons for decision making. The network uses local plasticity rules to solve one-step as well as sequential decision making tasks, which mimics the neural responses recorded in frontal cortices during the execution of such similar tasks. Their model reproduced behavioral and neuro-physiological data on tasks ranging from simple binary choice to multi-step sequential decision making. They took a two-step maze navigation task as an illustration. During each state, the rat was rewarded with different values according to its actions. The reward was modeled as an external stimuli. The SNN learned a stable policy within 10 ms.

5.3. Others

Except for the two aforementioned major methods, there are also other training methods for SNNs in robot control tasks as follows (see Table 2).

Table 2

Learning ruleRobotSensorMethodologyReference
Evolutionary algorithmsNeural Racing gameSpeedometer, Proximity SensorsUsing evolutionary algorithm to train SNN and compare results with multi-layer perceptronMarkowska and Koldowski, 2015
QuadrotorGPSUsing evolutionary algorithm to generate high utility topology/weight combinations in the SNNHoward and Elfes, 2014
Two-wheel Vehicle5 IR SensorsUsing SNN to mimic the behaviors captured under control of a heuristic rule programBatllori et al., 2011
Khepera Robot (Two-wheel Vehicle)Linear CameraUsing evolution to rapidly generate SNN capable of navigating in a textual environmentFloreano et al., 2006; Floreano and Mattiussi, 2001
Two-wheel Vehicle4 IR SensorsA use-dependent synaptic modification algorithm of SNN for obstacle-avoidance vehicle behaviorAlnajjar and Murase, 2006
Two-wheel Vehicle9 Ultrasonic Sensors, 4 Bump SensorsUsing an adaptive GA to evolve the SNN online through interaction with the real environmentHagras et al., 2004
Fuzzy logicalTwo-wheel Vehicle7 Ultrasonic Sensors (5 in front, 2 at back)Using SNN to mimic the knowledge of a fuzzy controllerKubota, 2004
Liquid state machineHexapod RobotVisual Sensor (Distance, Height)Mushroom bodies in drosophila are modeled as a recurrent SNN under LSM paradigmArena et al., 2017
2-Dof Ball Balance PlatformPosition and VelocityUsing a cortical network (LSM) to learn under a supervised learning rule for position controlProbst et al., 2012
Khepera Robot8 IR SensorsUsing Randomly generated recurrent SNN to operate real-time obstacle avoidanceBurgsteiner, 2005

Other learning rules.

5.3.1. Evolutionary algorithms

In nature, evolution has produced a multitude of organisms in all kinds of shapes with survival strategies optimally aligned to environmental conditions. Based on these ideas, a class of algorithms has been developed for finding problem solutions by mimicking elementary natural processes called evolutionary algorithms (Michalewicz, 1996). Generally, evolutionary processes can be understood as some form of gradient-descent optimization. Therefore, a typical problem using these algorithms is getting stuck in local minima. In applications in robot control, evolving SNNs have been shown to work well in mostly static environments. Due to the training principle of trial and error, there are usually difficulties in dynamically changing environments.

Floreano and Mattiussi (2001) showed a vision-based controller in an irregularly textured environment that navigated without hitting obstacles. The predefined SNN consisted of 18 sensory-input receptors connected to 10 fully-connected hidden neurons and 2 motor-output neurons. Using static synaptic weight values, the algorithm was used to search the space of connectivity by genetically evolving only signs of weights (excitatory and inhibitory), when the robot was continuously driving around in the experiment setup. With a population of 60 individuals, fitness was evaluated by summing up over motor speeds at every time step, and new generations were created using one-point crossover, bit mutation and elitism. Hagras et al. (2004) later extended this approach to evolving SNN weights as well using adaptive crossover and mutation probabilities. They were able to evolve good SNN controllers in a small number of generations in a wall-following scenario. Howard and Elfes (2014) presented a quadrotor neurocontroller that performed a hovering task in challenging wind conditions. With a feed-forward network taking the differences between current position and target position as input and pitch, roll and thrust as output, weights and topology were evolved to minimize the spatial error. In a target-reaching and obstacle-avoidance task using binocular light sensors and proximity sensors, Batllori et al. (2011) evolved an SNN by minimizing the control error in order to mimic an external controller signal. Markowska and Koldowski (2015) used a feed-forward network architecture of predefined size to control a toy car. Based on speed, localization and road boarder input signals, the network controlled speed regulation and turn direction, and evolved its weights using a genetic algorithm.

5.3.2. Self-organizing algorithms

Alnajjar and Murase (2006) formulated a synaptic learning rule that enforced connections between neurons depending on their activities. During the learning phase, the robot gradually organized the network and the obstacle avoidance behavior was formed. With this self-organization algorithm that resembles other Hebbian-based learning methods, they were able to learn obstacle avoidance and simple navigation behavior.

5.3.3. Liquid state machine

As a particular kind of SNN, an liquid state machine (LSM) usually consists of a large assemblage of neurons that receives time-varying input from external sources as well as from other neural units (Yamazaki and Tanaka, 2007). All mixed and disorderly neuron units are randomly generated and then arranged under the activations of recurrent spatio-temporal patterns of the connections obtained from the time-varying input. Hence, the LSM is regarded as a large variety of nonlinear functions which is able to compute the output as linear combinations of the input. LSMs seem to be a potential and promising theory to explain brain operation mainly because neuron activities are not hard coded and limited for specific tasks. Burgsteiner (2005), Probst et al. (2012), and Arena et al. (2017) showed how liquid state machines can be trained for robot control tasks.

6. Simulators and platforms

With the fast development of neuroscience and chip industry, large-scale neuromorphic hardware using spiking neural networks has been studied to achieve the same capabilities as animal brains in terms of speed, efficiency, and mechanism. For examples, SpiNNaker (Furber et al., 2013) is a million-core system for modeling large-scale SNNs in real time. TruthNorth (Merolla et al., 2014) contains 1 million programmable spiking neurons and only consumes less than one hundred milliwatts. Other neuromorphic computing platforms such as Neural Grid (Benjamin et al., 2014), NeuroFlow (Cheung et al., 2016) can be found and introduced in Schuller and Stevens (2015). Meanwhile, a growing number of dynamic simulators has been developed to assist robotic research (Ivaldi et al., 2014), such as Gazebo (Koenig and Howard, 2004), ODE (Wikipedia, 2017c), and V-Rep (Rohmer et al., 2013). Those simulators greatly facilitate the research process that involving mechanical design, virtual sensors simulation, and control architecture.

Although adequate tools exist to simulate either spiking neural networks (Brette et al., 2007; Bekolay et al., 2014), or robots and their environments (Staranowicz and Mariottini, 2011; Harris and Conrad, 2011), tools that offer researchers joint interaction, including a realistic brain model, robot, and sensory-rich environment, are in need. Some existing platforms are listed in Table 3.

Table 3

PlatformNameMethodologyReference
PlatformNeurorobotics PlatformDesign, import, and simulate different robot bodies and diverse brain models in rich environmentsFalotico et al., 2017
Musculoskeletal RobotsCombining Myorobotics with SpiNNaker the proof of principle of a system that can scale to dozens of neurally controlled, physically compliant joints.Richter et al., 2016
Retina simulationThe retina simulation platform is integrated in the NRP.Ambrosano et al., 2016
Neural self-driving vehicle simulation frameworkA visual encoder from camera images to spikes inspired by the silicon retina, and a steering-wheel decoder based on an agonist antagonist muscle model.Kaiser et al., 2016
iSpikeInterface between SNN simulators and the iCub humanoid robotGamez et al., 2012
AnimatLabProvide functions, such as robot modeling, two neural models, and plugins for importing other models.Cofer et al., 2010

Taxonomy of platforms for robotics control based on SNN.

iSpike (Gamez et al., 2012), as the first attempt to combine spiking neural networks and robots, is a C++ library that provides an interface between SNN simulators and the iCub humanoid robot. It uses a biologically inspired approach to convert the robot's sensory information into spikes that are passed to the neural network simulator, and it decodes output spikes from the network into motor signals that are sent to control the robot. CLONES (Voegtlin, 2011) communicates between the Brian neural simulator (Goodman and Brette, 2009) and SOFA (Allard et al., 2007) and is also an interface used for shared memory and semaphores. A more generic system which permits dealing with simulated robotic platforms is AnimatLab (Cofer et al., 2010), which provides functionalities such as robot modeling, two neural models, and plugins for importing other models.

Recently, the first release of the HBP Neurorobotics Platform (NRP) (American Association for the Advancement of Science, 2016; Falotico et al., 2017) was presented, which was developed within the EU Flagship Human Brain Project. For the first time, it provides scientists with an integrated toolchain to connect pre-defined and customized brain models to detailed simulations of robot bodies and environments in in-silico experiments. In particular, NRP consists of six key components, which are essential to construct neurorobotics experiments from scratch. It can be seen that the NRP provides a complete framework for the coupled simulation of robots and brain models. The Brain Simulator simulates the brain by bio-inspired learning algorithms such as a spiking neural network to control the robot in a silico neurorobotics experiment. The World Simulator simulates the robots and their interacting environment. The Brain Interface and Body Integrator (BIBI) builds a communication channel between brain models and robot models. The Closed Loop Engine (CLE) is responsible for the control logic of experiments as well as for the data communication between different components. The Backend receives requests from the frontend for the neurorobotics experiment and distributes them to the corresponding component, mainly via ROS. The Frontend is a web-based user interface for neurorobotics experiments. Users are able to design a new experiment or edit existing template experiments.

7. Open research topics

In the previous sections, the state-of-the-art of SNN-based control for various robots has been surveyed in terms of learning methods. Although an increasing amount of work has been done to explore the theoretical foundations and practical implementations of SNNs for robotics control, many related topics need to be investigated, especially in the following areas.

7.1. Biological mechanism

Despite the extensive exploration of the functions and structure of the brain, the exact mechanisms of learning in biological neurons remain unknown. Some of those related to robotics applications are listed as: (1) How is diverse information coded in many neural activities other than the rates and timing of spikes? (2) How are memories distinguished, stored, and retrieved in such an efficient and precise manner? (3) How do brains simulate the future, since it involves the concept of “previous steps,” thus requiring some form of memory? As long as we can constantly address these unsolved mysteries of the brain, the robots of the future definitely can achieve more advanced intelligence.

7.2. Designing and training SNNs

None of the currently suggested algorithms are general-purpose able, at least in principle, to learn an arbitrary task in the way that backpropagation (through time) and its variants (with all their limitations) do for rate neurons (Grüning and Bohte, 2014). Therefore, there is no general design framework that could offer the functionalities of modeling and training, as well as those substantial tools for the conventional ANNs do, for instance, Tensorflow (Allaire et al., 2016), Theano (Theano Development Team, 2016), and Torch (Collobert et al., 2011). The nature of this situation is that training these kind of networks is notoriously difficult, especially when it comes to deep-network architectures. Since error backpropagation mechanisms commonly used in ANNs cannot be directly transferred to SNNs due to non-differentiabilities at spike times, there has been a void of practical learning methods.

Moreover, training should strengthen the combination with the burgeoning technologies of reinforcement learning, for instance, extending SNN into deep architecture or generating continuous action space (Lillicrap et al., 2015). In the future, combining the R-STDP with a reward-prediction model could lead to an algorithm that is actually capable of solving sequential decision tasks such as MDPs as well.

7.3. High performance computing with neuromorphic devices

Another important general issue that needs extensive research and is not clearly defined is how to integrate SNN-based controllers into neuromorphic devices, since they have the potential to offer fundamental improvements in computational capabilities such as speed and lower power consumption (Hang et al., 2003; Schuller and Stevens, 2015). These are of vital importance for robot applications, especially in mobile applications where real-time responses are important and energy supply is limited. An overview of how to program SNNs based on neuromorphic chips can be found (Walter et al., 2015b).

SNNs computation can highly benefit from parallel computing, substantially more so than conventional ANNs. Unlike a traditional neuron in rate coding, a spiking neuron does not need to receive weight values from each presynaptic neuron at each compution step. Since at each time step only a few neurons are active in an SNN, the classic bottleneck of message passing is removed. Moreover, computing the updated state of membrane potential is more complex than computing a weighted sum. Therefore communication time and computation cost are much more well-balanced in SNN parallel implementation as compared to conventional ANNs.

7.4. Interdisciplinary research of neuroscience and robotics

Another barrier that needs to be removed comes from a dilemma for the researchers of neuroscience and robotics: Roboticists often use a simplified brain model in a virtual robot to make a real-time simulation, while neuro-scientists develop detailed brain models that are not possible to be embedded into the real world due to their high complexity. Learning complex sensorimotor mapping of the robot generated in the interaction with dynamic and rich sensory environment is also required (Hwu et al., 2017). An ongoing solution is the Neurorobotics Platform, which offers adequate tools to model virtual robots, high-fidelity environments, and complex neural network models for both neuroscientists and roboticists.

8. Conclusion

By mimicking the underlying mechanisms of the brain much more realistically, spiking neural networks have showed great potential for achieving advanced robotic intelligence in terms of speed, energy efficiency, and computation capabilities. Therefore in this article, we seek to offer readers a comprehensive review of the literature about solving robotic control tasks based on SNNs as well as the related modeling and training approaches, and meanwhile offer inspiration to researchers. Specifically, we retrospect the biological evidences of SNNs and their major impetuses for being adopted for the area of robotics at the beginning. Then, we present the mainstream modeling approaches for designing SNNs in terms of neuron, synapse, and network. The learning solutions of SNNs are generally classified into two types based on Hebbian rule and reinforcement learning, illustrated and expounded with exhaustive robotic-related examples and summary tables. Finally, some popular interfaces or platforms for simulating SNNs for robotics are preliminarily investigated.

As indicated in the open topics, the biggest challenge for control tasks based on SNNs is a lack of a universal training method, as back-propagation is to the conventional ANNs. Therefore, more knowledge and interactions from the fields of neuroscience and robotics are needed to explore this area in the future.

Statements

Author contributions

ZB, FR, and AK brought up the core concept and architecture of this manuscript. ZB, CM, FR, and KH wrote the paper.

Funding

The research leading to these results has received funding from the European Union Research and Innovation Programme Horizon 2020 (H2020/2014-2020) under grant agreement No. 720270 (The Human Brain Project, HBP) and the Chinese Scholarship Council. Meanwhile it was also supported by the German Research Foundation (DFG) and the Technical University of Munich (TUM) in the framework of the Open Access Publishing Program.

Acknowledgments

We thank Mr. Philipp Neumann for his valuable suggestions for improving this paper.

Conflict of interest

The authors declare that the research was conducted in the absence of any commercial or financial relationships that could be construed as a potential conflict of interest.

References

  • 1

    AdrianE. D. (1926). The impulses produced by sensory nerve endings. J. Physiol.61, 4972.

  • 2

    AllaireJ.EddelbuettelD.GoldingN.TangY. (2016). tensorflow: R Interface to TensorFlow. Available online at: https://github.com/rstudio/tensorflow

  • 3

    AllardJ.CotinS.FaureF.BensoussanP. J.PoyerF.DuriezC.et al. (2007). SOFA-an open source framework for medical simulation, in MMVR 15-Medicine Meets Virtual Reality, Vol. 125 (Palm Beach, FL: IOP Press), 1318.

  • 4

    AlnajjarF.MuraseK. (2006). Self-organization of spiking neural network generating autonomous behavior in a miniature mobile robot, in Proceedings of the 3rd International Symposium on Autonomous Minirobots for Research and Edutainment (AMiRE 2005) (Berlin; Heidelberg: Springer), 255260.

  • 5

    AmbrosanoA.VannucciL.AlbaneseU.KirtayM.FaloticoE.Martínez-CañadaP.et al. (2016). Retina color-opponency based pursuit implemented through spiking neural networks in the neurorobotics platform, in Biomimetic and Biohybrid Systems: 5th International Conference, Living Machines 2016 (Edinburgh: Springer International Publishing).

  • 6

    American Association for the Advancement of Science (2016). Booklet Brain-inspired intelligent robotics: the intersection of robotics and neuroscience sciences. Science354, 1445. 10.1126/science.354.6318.1445-b

  • 7

    AndrewA. M. (2003). Spiking neuron models: single neurons, populations, plasticity. Kybernetes32, 78. 10.1108/k.2003.06732gae.003

  • 8

    ArenaE.ArenaP.StraussR.PatanéL. (2017). Motor-skill learning in an insect inspired neuro-computational control system. Front. Neurorobot.11:12. 10.3389/fnbot.2017.00012

  • 9

    ArenaP.De FioreS.PatanéL.PollinoM.VenturaC. (2009a). STDP-based behavior learning on the tribot robot, in Society of Photo-Optical Instrumentation Engineers (SPIE) Conference Series (Dresden).

  • 10

    ArenaP.De FioreS.PatanéL.PollinoM.VenturaC. (2010). Insect inspired unsupervised learning for tactic and phobic behavior enhancement in a hybrid robot, in The 2010 International Joint Conference on Neural Networks (IJCNN) (Barcelona: IEEE), 18.

  • 11

    ArenaP.FortunaL.FrascaM.PatanéL. (2009b). Learning anticipation via spiking networks: application to navigation control. IEEE Trans. Neural Netw.20, 202216. 10.1109/TNN.2008.2005134

  • 12

    BatlloriR.LarameeC. B.LandW.SchafferJ. D. (2011). Evolving spiking neural networks for robot control. Proced. Comput. Sci.6, 329334. 10.1016/j.procs.2011.08.060

  • 13

    BekolayT.BergstraJ.HunsbergerE.DeWolfT.StewartT. C.RasmussenD. (2014). Nengo: a python tool for building large-scale functional brain models. Front. Neuroinformatics7:48. 10.3389/fninf.2013.00048

  • 14

    BenjaminB. V.GaoP.McQuinnE.ChoudharyS.ChandrasekaranA. R.BussatJ. M.et al. (2014). Neurogrid: a mixed-analog-digital multichip system for large-scale neural simulations. Proc. IEEE102, 699716. 10.1109/JPROC.2014.2313565

  • 15

    BiG. Q.PooM. M. (1998). Synaptic modifications in cultured hippocampal neurons: dependence on spike timing, synaptic strength, and postsynaptic cell type. J. Neurosci.18, 1046410472.

  • 16

    BingZ.ChengL.ChenG.RöhrbeinF.HuangK.KnollA. (2017). Towards autonomous locomotion: CPG-based control of smooth 3D slithering gait transition of a snake-like robot. Bioinspir. Biomimet.12:035001. 10.1088/1748-3190/aa644c

  • 17

    BishopC. M. (1995). Neural Networks for Pattern Recognition. Cambridge, UK: Oxford University Press.

  • 18

    BohteS. M.La PoutreH.KokJ. N. (2002). Unsupervised clustering with spiking neurons by sparse temporal coding and multilayer rbf networks. IEEE Trans. Neural Netw.13, 426435. 10.1109/72.991428

  • 19

    BouganisA.ShanahanM. (2010). Training a spiking neural network to control a 4-dof robotic arm based on spike timing-dependent plasticity, in The 2010 International Joint Conference on Neural Networks (IJCNN) (Barcelona: IEEE), 18.

  • 20

    BraitenbergV. (1978). Cell assemblies in the cerebral cortex. in Theoretical Approaches to Complex Systems (Tübingen: Springer), 171188.

  • 21

    BretteR.RudolphM.CarnevaleT.HinesM.BeemanD.BowerJ. M.et al. (2007). Simulation of networks of spiking neurons: a review of tools and strategies. J. Comput. Neurosci.23, 349398. 10.1007/s10827-007-0038-6

  • 22

    BurgsteinerH. (2005). Training networks of biological realistic spiking neurons for real-time robot control, in Proceedings of the 9th International Conference on Engineering Applications of Neural Networks (Lille), 129136.

  • 23

    BurkittA. N. (2006). A review of the integrate-and-fire neuron model: I. homogeneous synaptic input. Biol. Cybern.95, 119. 10.1007/s00422-006-0068-6

  • 24

    CareyM. R.MedinaJ. F.LisbergerS. G. (2005). Instructive signals for motor learning from visual cortical area mt. Nat. Neurosci.8, 813819. 10.1038/nn1470

  • 25

    CarrilloR. R.RosE.BouchenyC.OlivierJ. M. C. (2008). A real-time spiking cerebellum model for learning robot control. Biosystems94, 1827. 10.1016/j.biosystems.2008.05.008

  • 26

    CasellatoC.AntoniettiA.GarridoJ. A.CarrilloR. R.LuqueN. R.RosE.et al. (2014). Adaptive robotic control driven by a versatile spiking cerebellar network. PLoS ONE9:e112265. 10.1371/journal.pone.0112265

  • 27

    CassidyA.EkanayakeV. (2006). A biologically inspired tactile sensor array utilizing phase-based computation, in Biomedical Circuits and Systems Conference, 2006. BioCAS 2006. IEEE (London, UK: IEEE), 4548.

  • 28

    CâteauH.FukaiT. (2003). A stochastic method to predict the consequence of arbitrary forms of spike-timing-dependent plasticity. Neural Comput.15, 597620. 10.1162/089976603321192095

  • 29

    ChadderdonG. L.NeymotinS. A.KerrC. C.LyttonW. W. (2012). Reinforcement learning of targeted movement in a spiking neuronal model of motor cortex. PLoS ONE7:e47251. 10.1371/journal.pone.0047251

  • 30

    CheungK.SchultzS. R.LukW. (2016). Neuroflow: a general purpose spiking neural network simulation platform using customizable processors. Front. Neurosci.9:516. 10.3389/fnins.2015.00516

  • 31

    ChouT. S.BucciL. D.KrichmarJ. L. (2015). Learning touch preferences with a tactile robot using dopamine modulated stdp in a model of insular cortex. Front. Neurorobot.9:6. 10.3389/fnbot.2015.00006

  • 32

    ChunM. M.PotterM. C. (1995). A two-stage model for multiple target detection in rapid serial visual presentation. J. Exp. Psychol. Hum. Percept. Perform.21, 109.

  • 33

    ClawsonT. S.FerrariS.FullerS. B.WoodR. J. (2016). Spiking neural network (SNN) control of a flapping insect-scale robot, in 2016 IEEE 55th Conference on Decision and Control (CDC) (Las Vegas, NV: IEEE), 33813388.

  • 34

    CoferD.CymbalyukG.ReidJ.ZhuY.HeitlerW. J.EdwardsD. H. (2010). Animatlab: a 3d graphics environment for neuromechanical simulations. J. Neurosci. Methods187, 280288. 10.1016/j.jneumeth.2010.01.005

  • 35

    CollobertR.KavukcuogluK.FarabetC. (2011). Torch7: A matlab-like environment for machine learning, in BigLearn, NIPS Workshop (Lake Tahoe, NV).

  • 36

    CyrA.BoukadoumM. (2012). Classical conditioning in different temporal constraints: an stdp learning rule for robots controlled by spiking neural networks. Adapt. Behav.20, 257272. 10.1177/1059712312442231

  • 37

    CyrA.BoukadoumM.ThériaultF. (2014). Operant conditioning: a minimal components requirement in artificial spiking neurons designed for bio-inspired robot's controller. Front. Neurorobot.8:21. 10.3389/fnbot.2014.00021

  • 38

    CyrA.ThériaultF. (2015). Action selection and operant conditioning: a neurorobotic implementation. J. Robot.2015:6. 10.1155/2015/643869

  • 39

    DasGuptaB.SchnitgerG. (1992). The power of approximating: a comparison of activation functions, in NIPS (Denver, CO), 615622.

  • 40

    DiehlP. U.CookM. (2015). Unsupervised learning of digit recognition using spike-timing-dependent plasticity. Front. Comput. Neurosci.9:99. 10.3389/fncom.2015.00099

  • 41

    DongY.WangY.LinZ.WatanabeT. (2009). High performance and low latency mapping for neural network into network on chip architecture, in 2009 IEEE 8th International Conference on ASIC (Changsha), 891894.

  • 42

    DrubachD. (2000). The Brain Explained. Upper Saddle River, NJ: Prentice Hall Health.

  • 43

    DumesnilE.BeaulieuP. O.BoukadoumM. (2016a). Robotic implementation of classical and operant conditioning as a single stdp learning process, in 2016 International Joint Conference on Neural Networks (IJCNN) (Vancouver, BC: IEEE), 52415247.

  • 44

    DumesnilE.BeaulieuP. O.BoukadoumM. (2016b). Robotic implementation of classical and operant conditioning within a single snn architecture, in 2016 IEEE 15th International Conference on Cognitive Informatics & Cognitive Computing (ICCI* CC) (Palo Alto, CA: IEEE), 322330.

  • 45

    Dura-BernalS.ZhouX.NeymotinS. A.PrzekwasA.FrancisJ. T.LyttonW. W. (2015). Cortical spiking network interfaced with virtual musculoskeletal arm and robotic arm. Front. Neurorobot.9:13. 10.3389/fnbot.2015.00013

  • 46

    EliasmithC.StewartT. C.ChooX.BekolayT.DeWolfT.TangY.et al. (2012). A large-scale model of the functioning brain. Science338, 12021205. 10.1126/science.1225266

  • 47

    EvansR. (2012). Reinforcement learning in a neurally controlled robot using dopamine modulated stdp. arXiv preprint arXiv:1502.06096.

  • 48

    EvansR. (2015). Reinforcement learning in a neurally controlled robot using dopamine modulated stdp. arXiv preprint arXiv:1502.06096.

  • 49

    FaghihiF.MoustafaA. A.HeinrichR. Wörgötter. F. (2017). A computational model of conditioning inspired by drosophila olfactory system. Neural Netw.87, 96108. 10.1016/j.neunet.2016.11.002

  • 50

    FaloticoE.VannucciL.AmbrosanoA.AlbaneseU.UlbrichS.Vasquez TieckJ. C.et al. (2017). Connecting artificial brains to robots in a comprehensive simulation framework: the neurorobotics platform. Front. Neurorobot.11:2. 10.3389/fnbot.2017.00002

  • 51

    FersterD.SprustonN. (1995). Cracking the neuronal code. Science270:756.

  • 52

    FiaschéM.TaischM. (2015). On the use of quantum-inspired optimization techniques for training spiking neural networks: a new method proposed, in Advances in Neural Networks: Computational and Theoretical Issues. Smart Innovation, Systems and Technologies, Vol. 37, eds BassisS.EspositoA.MorabitoF. (Cham: Springer).

  • 53

    FloreanoD.EparsY.ZuffereyJ.MattiussiC. (2006). Evolution of spiking neural circuits in autonomous mobile robots. Int. J. Intell. Syst.21, 10051024. 10.1002/int.20173

  • 54

    FloreanoD.MattiussiC. (2001). Evolution of spiking neural controllers for autonomous vision-based robots, in Evolutionary Robotics. From Intelligent Robotics to Artificial Life. EvoRobots 2001. Lecture Notes in Computer Science, Vol. 2217, ed GomiT. (Berlin; Heidelberg: Springer).

  • 55

    FlorianR. V. (2007). Reinforcement learning through modulation of spike-timing-dependent synaptic plasticity. Neural Comput.19, 14681502. 10.1162/neco.2007.19.6.1468

  • 56

    FoderaroG.HenriquezC.FerrariS. (2010). Indirect training of a spiking neural network for flight control via spike-timing-dependent synaptic plasticity, in 2010 49th IEEE Conference on Decision and Control (CDC) (Atlanta, GA: IEEE), 911917.

  • 57

    FrémauxN.SprekelerH.GerstnerW. (2013). Reinforcement learning using a continuous time actor-critic framework with spiking neurons. PLoS Comput. Biol.9:e1003024. 10.1371/journal.pcbi.1003024

  • 58

    FriedrichJ.LengyelM. (2016). Goal-directed decision making with spiking neurons. J. Neurosci.36, 15291546. 10.1523/JNEUROSCI.2854-15.2016

  • 59

    FurberS. B.GalluppiF.TempleS.PlanaL. A. (2014). The spinnaker project. Proc. IEEE102, 652665. 10.1109/JPROC.2014.2304638

  • 60

    FurberS. B.LesterD. R.PlanaL. A.GarsideJ. D.PainkrasE.TempleS.et al. (2013). Overview of the spinnaker system architecture. IEEE Trans. Comput.62, 24542467. 10.1109/TC.2012.142

  • 61

    GabbianiF.MetznerW.WesselR.KochC. (1996). From stimulus encoding to feature extraction in weakly electric fish. Nature384, 564.

  • 62

    GamezD.FidjelandA. K.LazdinsE. (2012). ispike: a spiking neural interface for the icub robot. Bioinspir. Biomimetics7:025008. 10.1088/1748-3182/7/2/025008

  • 63

    GerstnerW.KempterR.van HemmenJ.WagnerH. (1999). Hebbian learning of pulse timing in the barn owl auditory system, in Pulsed Neural Networks, eds MaassW.BishopC. M. (Cambridge, MA: MIT Press), 353377.

  • 64

    GerstnerW.KistlerW. M. (2002) Mathematical formulations of hebbian learning. Biol. Cybern.87, 404415. 10.1007/s00422-002-0353-y

  • 65

    GerstnerW.RitzR.van HemmenJ. L. (1993). Why spikes? hebbian learning and retrieval of time-resolved excitation patterns. Biol. Cybern.69, 503515.

  • 66

    GerstnerW.KempterR.van HemmenJ. L.WagnerH. (1996). A neuronal learning rule for sub-millisecond temporal coding. Nature383, 76c.

  • 67

    Ghosh-DastidarS.AdeliH. (2009). Spiking neural networks. Int. J. Neural Syst.19, 295308. 10.1142/S0129065709002002

  • 68

    GongC.TraversM. J.AstleyH. C.LiL.MendelsonJ. R.GoldmanD. I.et al. (2016). Kinematic gait synthesis for snake robots. Int. J. Robot. Res.35, 100113. 10.1177/0278364915593793

  • 69

    GoodmanD. F.BretteR. (2009). The brian simulator. Front. Neurosci.3:192. 10.3389/neuro.01.026.2009

  • 70

    GrüningA.BohteS. M. (2014). Spiking neural networks: principles and challenges, in ESANN (Bruges).

  • 71

    GütigR.SompolinskyH. (2006). The tempotron: a neuron that learns spike timing–based decisions. Nat. Neurosci.9, 420428. 10.1038/nn1643

  • 72

    GuyonneauR.VanrullenR.ThorpeS. J. (2004). Temporal codes and sparse representations: a key to understanding rapid processing in the visual system. J. Physiol.98, 487497. 10.1016/j.jphysparis.2005.09.004

  • 73

    GspandlS.PodesserS.ReipM.SteinbauerG.WolframM. (2012). A dependable perception-decision-execution cycle for autonomous robots, in 2012 IEEE International Conference on Robotics and Automation (Saint Paul, MN), 29922998.

  • 74

    HagrasH.Pounds-CornishA.ColleyM.CallaghanV.ClarkeG. (2004). Evolving spiking neural network controllers for autonomous robots, in IEEE International Conference on Robotics and Automation (ICRA) (New Orleans, LA: IEEE), 46204626.

  • 75

    HanJ.MoragaC. (1995). The influence of the sigmoid function parameters on the speed of backpropagation learning, in From Natural to Artificial Neural Computation (Berlin; Heidelberg), 195201.

  • 76

    HangQ.PengJ.XuZ. B.ZhangB. (2003). A reference model approach to stability analysis of neural networks. IEEE Trans. Syst. Man Cybern Pt B33, 925936.

  • 77

    HarrisA.ConradJ. M. (2011). Survey of popular robotics simulators, frameworks, and toolkits, in Southeastcon, 2011 Proceedings of IEEE (Nashville, TN: IEEE), 243249.

  • 78

    HasselmoM. E. (1999). Neuromodulation: acetylcholine and memory consolidation. Trends Cogn. Sci.3, 351359.

  • 79

    HastieT.FriedmanJ.TibshiraniR. (2001). Overview of supervised learning, in The Elements of Statistical Learning. Springer Series in Statistics (New York, NY: Springer).

  • 80

    HebbD. O. (1949). The Organization of Behavior: A Neuropsychological Approach. Mahwah, NJ: John Wiley & Sons.

  • 81

    Hecht-NielsenR. (1992). Theory of the backpropagation neural network, in Neural Networks for Perception, ed WechslerH. (Academic Press), 6593.

  • 82

    HelgadottirL. I.HaenickeJ.LandgrafT.RojasR.NawrotM. P. (2013). Conditioned behavior in a robot controlled by a spiking neural network, in Neural Engineering (NER), 2013 6th International IEEE/EMBS Conference on (San Diego, CA: IEEE), 891894.

  • 83

    Herculano-HouzelS. (2012). The remarkable, yet not extraordinary, human brain as a scaled-up primate brain and its associated cost. Proc. Natl. Acad. Sci. U.S.A.109(Suppl. 1), 1066110668. 10.1073/pnas.1201895109

  • 84

    Herculano-HouzelS.Avelino-de SouzaK.NevesK.PorfíoJ.MessederD.Mattos FeijóL.et al. (2014). The elephant brain in numbers. Front. Neuroanat.8:46. 10.3389/fnana.2014.00046

  • 85

    HintonG. E.SejnowskiT. J. (1999). Unsupervised Learning: Foundations of Neural Computation. Cambridge, MA: MIT Press.

  • 86

    HodgkinA. L.HuxleyA. F. (1952). A quantitative description of membrane current and its application to conduction and excitation in nerve. J. Physiol.117, 500544.

  • 87

    HopfieldJ. J. (1982). Neural networks and physical systems with emergent collective computational abilities. Proc. Natl. Acad. Sci. U.S.A.79, 25542558.

  • 88

    HopfieldJ. J. (1995). Pattern recognition computation using action potential timing for stimulus representation. Nature376, 33.

  • 89

    HornikK.StinchcombeM.WhiteH. (1989). Multilayer feedforward networks are universal approximators. Neural Netw.2, 359366.

  • 90

    HouwelingA. R.BrechtM. (2008). Behavioural report of single neuron stimulation in somatosensory cortex. Nature451, 6568. 10.1038/nature06447

  • 91

    HowardD.AlbertoE. (2014). Evolving spiking networks for turbulence-tolerant quadrotor control, in International Conference on Artificial Life (ALIFE14) (New York, NY), 431438.

  • 92

    HuD.ZhangX.XuZ.FerrariS.MazumderP. (2014). Digital implementation of a spiking neural network (SNN) capable of spike-timing-dependent plasticity (STDP) learning, in 2014 IEEE 14th International Conference on Nanotechnology (IEEE-NANO) (Toronto, ON: IEEE), 873876.

  • 93

    HuberD.PetreanuL.GhitaniN.RanadeS.HromádkaT.MainenZ.et al. (2008). Sparse optical microstimulation in barrel cortex drives learned behaviour in freely moving mice. Nature451, 61. 10.1038/nature06445

  • 94

    HwuT.IsbellJ.OrosN.KrichmarJ. (2017). A self-driving robot using deep convolutional neural networks on neuromorphic hardware, in Neural Networks (IJCNN), 2017 International Joint Conference on (Anchorage, AK: IEEE), 635641.

  • 95

    IjspeertA. J. (2008). Central pattern generators for locomotion control in animals and robots: a review. Neural Netw.21, 642653. 10.1016/j.neunet.2008.03.014

  • 96

    IjspeertA. J.CrespiA.RyczkoD.CabelguenJ. M. (2007). From swimming to walking with a salamander robot driven by a spinal cord model. Science315, 14161420. 10.1126/science.1138353

  • 97

    IvaldiS.PetersJ.PadoisV.NoriF. (2014). Tools for simulating humanoid robot dynamics: a survey based on user feedback, in 2014 14th IEEE-RAS International Conference on Humanoid Robots (Humanoids) (Madrid: IEEE), 842849.

  • 98

    IwadateK.SuzukiI.WatanabeM.YamamotoM.FurukawaM. (2014). An artificial neural network based on the architecture of the cerebellum for behavior learning, in Soft Computing in Artificial Intelligence. Advances in Intelligent Systems and Computing, Vol. 270, eds ChoY.MatsonE. (Cham: Springer).

  • 99

    IzhikevichE. M. (2004). Which model to use for cortical spiking neurons?IEEE Trans. Neural Netw.15, 10631070. 10.1109/TNN.2004.832719

  • 100

    IzhikevichE. M. (2007). Solving the distal reward problem through linkage of stdp and dopamine signaling. Cereb. Cortex17, 24432452. 10.1093/cercor/bhl152

  • 101

    Jimenez-RomeroC. (2017). A Heterosynaptic Spiking Neural System for the Development of Autonomous Agents. Ph.D. Thesis, The Open University.

  • 102

    Jimenez-RomeroC.Sousa-RodriguesD.JohnsonJ. (2016). Designing behaviour in bio-inspired robots using associative topologies of spiking-neural-networks, in Proceedings of the 9th EAI International Conference on Bio-inspired Information and Communications Technologies (formerly BIONETICS). ICST (Brussels: Institute for Computer Sciences, Social-Informatics and Telecommunications Engineering), 197200.

  • 103

    Jimenez-RomeroC.Sousa-RodriguesD.JohnsonJ. H.RamosV. (2015). A model for foraging ants, controlled by spiking neural networks and double pheromones. arXiv preprint arXiv:1507.08467.

  • 104

    KaiserJ.TieckJ. C. V.HubschneiderC.WolfP.WeberM.HoffM.et al. (2016). Towards a framework for end-to-end control of a simulated vehicle with spiking neural networks, in IEEE International Conference on Simulation, Modeling, and Programming for Autonomous Robots (SIMPAR) (San Francisco, CA: IEEE), 127134.

  • 105

    KandelE. R.SchwartzJ. H.JessellT. M.SiegelbaumS. A.HudspethA. J.et al. (2000). Principles of Neural Science, Vol. 4. New York, NY: McGraw-Hill.

  • 106

    KawatoM.GomiH. (1992). A computational model of four regions of the cerebellum based on feedback-error learning. Biol. Cybern.68, 95103.

  • 107

    KempterR.GerstnerW.Van HemmenJ. L. (1999). Hebbian learning and spiking neurons. Phys. Rev. E59, 4498.

  • 108

    KnudsenE. I. (1994). Supervised learning in the brain. J. Neurosci.14, 39853997.

  • 109

    KocaturkM.GulcurH. O.CanbeyliR. (2015). Toward building hybrid biological/in silico neural networks for motor neuroprosthetic control. Front. Neurorobot.9:8. 10.3389/fnbot.2015.00008

  • 110

    KoenigN.HowardA. (2004). Design and use paradigms for gazebo, an open-source multi-robot simulator, in 2004 IEEE/RSJ International Conference on Intelligent Robots and Systems (IROS) (IEEE Cat. No. 04CH37566), Vol. 3, (Sendai), 21492154.

  • 111

    KroghA.VedelsbyJ. (1995). Neural network ensembles, cross validation, and active learning, in Advances in Neural Information Processing Systems (Denver, CO), 231238.

  • 112

    KubatM. (1999). Neural networks: a comprehensive foundation by simon haykin, macmillan, 1994, isbn 0-02-352781-7. Knowl. Eng. Rev.13, 409412.

  • 113

    KubotaN. (2004). A spiking neural network for behavior learning of a mobile robot in a dynamic environment, in IEEE International Conference on Systems, Man and Cybernetics, Vol 6. (The Hague: IEEE), 57835788.

  • 114

    KubotaN.NishidaK. (2006). The role of spiking neurons for visual perception of a partner robot, in IEEE International Conference on Fuzzy Systems (Vancouver, BC: IEEE), 122129.

  • 115

    LeeC. S.WangM. H.YenS. J.WeiT. H.WuI. C.ChouP. C.et al. (2016) Human vs. computer go: Review prospect [discussion forum]. IEEE Comput. Intell. Mag.11, 6772. 10.1109/MCI.2016.2572559

  • 116

    LeeJ. H.DelbruckT.PfeifferM. (2016). Training deep spiking neural networks using backpropagation. Front. Neurosci.10:508. 10.3389/fnins.2016.00508

  • 117

    LewisM. A.Etienne-CummingsR.CohenA. H.HartmannM. (2000). Toward biomorphic control using custom avlsi cpg chips, in IEEE International Conference on Robotics and Automation, 2000. Proceedings. ICRA'00., Vol. 1 (San Francisco, CA: IEEE), 494500.

  • 118

    LichtsteinerP.PoschC.DelbruckT. (2008). A 128 × 128 120 db 15 μs latency asynchronous temporal contrast vision sensor. IEEE J. Solid State Circuits43, 566576. 10.1109/JSSC.2007.914337

  • 119

    LillicrapT. P.HuntJ. J.PritzelA.HeessN.ErezT.TassaY.et al. (2015). Continuous control with deep reinforcement learning. arXiv preprint arXiv:1509.02971.

  • 120

    LoiselleS.RouatJ.PressnitzerD.ThorpeS. (2005). Exploration of rank order coding with spiking neural networks for speech recognition, in Neural Networks, 2005. IJCNN'05. Proceedings. 2005 IEEE International Joint Conference on, Vol 4. (Montreal, QC: IEEE), 20762080.

  • 121

    MaassW. (1997). Networks of spiking neurons: the third generation of neural network models. Neural Netw.10, 16591671. 10.1016/S0893-6080(97)00011-7

  • 122

    MaassW. (2001). On the relevance of time in neural computation and learning. Theor. Comput. Sci.261, 157178. 10.1016/S0304-3975(00)00137-7

  • 123

    MarkowskaU.KoldowskiM. (2015). Spiking neural network vs multilayer perceptron: who is the winner in the racing car computer game. Soft Comput.19, 34653478. 10.1007/s00500-014-1515-2

  • 124

    MarkramH.LübkeJ.FrotscherM.SakmannB. (1997). Regulation of synaptic efficacy by coincidence of postsynaptic aps and epsps. Science275, 213215.

  • 125

    MazumderP.HuD.EbongI.ZhangX.XuZ.FerrariS. (2016). Digital implementation of a virtual insect trained by spike-timing dependent plasticity. Integr. VLSI J.54, 109117. 10.1016/j.vlsi.2016.01.002

  • 126

    McCullochW. S.PittsW. (1943). A logical calculus of the ideas immanent in nervous activity. Bull. Math. Biophys.5, 115133.

  • 127

    MerollaP. A.ArthurJ. V.Alvarez-IcazaR.CassidyA. S.SawadaJ.AkopyanF.et al. (2014). A million spiking-neuron integrated circuit with a scalable communication network and interface. Science345, 668673. 10.1126/science.1254642

  • 128

    MeschedeC. (2017). Training Neural Networks for Event-Based End-to-End Robot Control. Master Thesis, Technical University of Munich.

  • 129

    MetznerW.KochC.WesselR.GabbianiF. (1998). Feature extraction by burst-like spike patterns in multiple sensory maps. J. Neurosci.18, 22832300.

  • 130

    MiallR. C.WolpertD. M. (1996). Forward models for physiological motor control. Neural Netw.9, 12651279.

  • 131

    MichalewiczZ. (1996). GAs: what are they?, in Genetic Algorithms + Data Structures = Evolution Programs (Berlin; Heidelberg: Springer).

  • 132

    MontgomeryJ.CartonG.BodznickD. (2002). Error-driven motor learning in fish. Biol. Bull.203, 238239. 10.2307/1543417

  • 133

    NeymotinS. A.ChadderdonG. L.KerrC. C.FrancisJ. T.LyttonW. W. (2013). Reinforcement learning of two-joint virtual arm reaching in a computer model of sensorimotor cortex. Neural Comput.25, 32633293. 10.1162/NECO_a_00521

  • 134

    NicholsE.McDaidL. J.SiddiqueN. H. (2010). Case study on a self-organizing spiking neural network for robot navigation. Int. J. Neural Syst.20, 501508. 10.1142/S0129065710002577

  • 135

    NicholsE.McDaidL. J.SiddiqueN. (2013). Biologically inspired snn for robot control. IEEE Trans. Cybern.43, 115128. 10.1109/TSMCB.2012.2200674

  • 136

    PavlovI. P.AnrepG. V. (2003). Conditioned Reflexes. Mineola, NY: Courier Corporation.

  • 137

    PerrinetL.SamuelidesM.ThorpeS. (2004). Sparse spike coding in an asynchronous feed-forward multi-layer neural network using matching pursuit. Neurocomputing57, 125134. 10.1016/j.neucom.2004.01.010

  • 138

    PonulakF.KasinskiA. (2011). Introduction to spiking neural networks: information processing, learning and applications. Acta Neurobiol. Exp.71, 409433.

  • 139

    PotjansW.MorrisonA.DiesmannM. (2009). A spiking neural network model of an actor-critic learning agent. Neural Comput.21, 301339. 10.1162/neco.2008.08-07-593

  • 140

    ProbstD.MaassW.MarkramH.GewaltigM. O. (2012). Liquid computing in a simplified model of cortical layer IV: learning to balance a ball, in Artificial Neural Networks and Machine Learning - ICANN 2012. ICANN 2012. Lecture Notes in Computer Science, Vol. 7552, eds VillaA. E. P.DuchW.ÉrdiP.MasulliF.PalmG. (Berlin; Heidelberg: Springer).

  • 141

    QiaoH.LiY.TangT.WangP. (2014). Introducing memory and association mechanism into a biologically inspired visual model. IEEE Trans. Cybern.44, 14851496. 10.1109/tcyb.2013.2287014

  • 142

    QiaoH.LiY.LiF.XiX.WuW. (2016). Biologically inspired model for visual cognition achieving unsupervised episodic and semantic feature learning. IEEE Trans. Cybern.46, 23352347. 10.1109/tcyb.2015.2476706

  • 143

    QiaoH.XiX.LiY.WuW.LiF. (2015). Biologically inspired visual model with preliminary cognition and active attention adjustment. IEEE Trans. Cybern.45, 26122624. 10.1109/tcyb.2014.2377196

  • 144

    RichterC.JentzschS.HostettlerR.GarridoJ. A.RosE.KnollA.RöhrbeinF.et al. (2016). Musculoskeletal robots: scalability in neural control. IEEE Robot. Autom. Mag.23, 128137.

  • 145

    RobertsP. D. (1999). Computational consequences of temporally asymmetric learning rules: I. differential hebbian learning. J. Comput. Neurosci.7, 235246.

  • 146

    RochelO.MartinezD.HuguesE.SarryF. (2002). Stereo-olfaction with a sniffing neuromorphic robot using spiking neurons, in 16th European Conference on Solid-State Transducers-EUROSENSORS (Prague), 4.

  • 147

    RohmerE.SinghS. P. N.FreeseM. (2013). V-rep: a versatile and scalable robot simulation framework, in 2013 IEEE/RSJ International Conference on Intelligent Robots and Systems (Tokyo), 13211326.

  • 148

    RosselloJ.CanalsV.OliverA.AlomarM.MorroA. (2014). Spiking neural networks signal processing, in Conference on Design of Circuits and Integrated Circuits (DCIS) (Madrid: IEEE), 16.

  • 149

    RubinJ.LeeD. D.SompolinskyH. (2001). Equilibrium properties of temporally asymmetric hebbian plasticity. Phys. Rev. Lett.86:364. 10.1103/PhysRevLett.86.364

  • 150

    RueckertE.KappelD.TannebergD.PecevskiD.PetersJ. (2016). Recurrent spiking networks solve planning tasks. Sci. Rep.6:21142. 10.1038/srep21142

  • 151

    RufB.SchmittM. (1997). Unsupervised learning in networks of spiking neurons using temporal coding, in Artificial Neural Networks - ICANN'97. ICANN 1997. Lecture Notes in Computer Science, Vol. 1327, eds GerstnerW.GermondA.HaslerM.NicoudJ. D. (Berlin; Heidelberg: Springer).

  • 152

    SarimM.SchultzT.JhaR.KumarM. (2016a). Ultra-low energy neuromorphic device based navigation approach for biomimetic robots, in Aerospace and Electronics Conference (NAECON) and Ohio Innovation Summit (OIS), 2016 IEEE National (Dayton, OH: IEEE), 241247.

  • 153

    SarimM.SchultzT.KumarM.JhaR. (2016b). An artificial brain mechanism to develop a learning paradigm for robot navigation, in ASME 2016 Dynamic Systems and Control Conference (Minneapolis, MN: American Society of Mechanical Engineers), V001T03A004.

  • 154

    SchoettleB.SivakM. (2014). A Survey of Public Opinion About Autonomous and Self-Driving Vehicles in the US, the UK, and Australia. Report No. UMTRI-2014-21, the Transportation Research Institute, University of Michigan.

  • 155

    SchullerI. K.StevensR. (2015). Neuromorphic computing : from materials to systems architecture. Technical report, U.S. Department of Energy, Office of Science.

  • 156

    SchultzW. (1998). Predictive reward signal of dopamine neurons. J. Neurophysiol.80, 127.

  • 157

    SenguptaB.StemmlerM. B. (2014). Power consumption during neuronal computation. Proc. IEEE102, 738750. 10.1109/JPROC.2014.2307755

  • 158

    SennW.TsodyksM.MarkramH. (1997). An algorithm for synaptic modification based on exact timing of pre- and post-synaptic action potentials, in Artificial Neural Networks - ICANN'97. ICANN 1997. Lecture Notes in Computer Science, Vol. 1327, eds GerstnerW.GermondA.HaslerM.NicoudJ. D. (Berlin; Heidelberg: Springer).

  • 159

    ShenW.NorrieD. H. (1999). Agent-based systems for intelligent manufacturing: a state-of-the-art survey. Knowl. Inform. Syst.1, 129156.

  • 160

    ShinJ.-H.SmithD.SwierczW.StaleyK.RickardJ. T.MonteroJ.et al. (2010). Recognition of partially occluded and rotated images with a network of spiking neurons. IEEE Trans. Neural Netw.21, 16971709. 10.1109/TNN.2010.2050600

  • 161

    SilverD.HuangA.MaddisonC. J.GuezA.SifreL.Van Den DriesscheG.et al. (2016). Mastering the game of go with deep neural networks and tree search. Nature529, 484489. 10.1038/nature16961

  • 162

    SkorheimS.LonjersP.BazhenovM. (2014). A spiking network model of decision making employing rewarded stdp. PLoS ONE9:e90821. 10.1371/journal.pone.0090821

  • 163

    SongS.MillerK. D.AbbottL. F. (2000). Competitive hebbian learning through spike-timing-dependent synaptic plasticity. Nat. Neurosci.3, 919926. 10.1038/78829

  • 164

    SoulaH.AlwanA.BeslonG. (2005). Learning at the edge of chaos: Temporal coupling of spiking neurons controller for autonomous robotic, in Proceedings of the AAAI Spring Symposia on Developmental Robotics. AAAI Palo Alto, CA.

  • 165

    SpülerM.NagelS.RosenstielW. (2015). A spiking neuronal model learning a motor control task by reinforcement learning and structural synaptic plasticity, in International Joint Conference on Neural Networks (IJCNN), 2015 (Killarney: IEEE), 18.

  • 166

    StaranowiczA.MariottiniG. L. (2011). A survey and comparison of commercial and open-source robotic simulator software, in Proceedings of the 4th International Conference on PErvasive Technologies Related to Assistive Environments (Heraklion: ACM), 56.

  • 167

    SteinR. B. (1965). A theoretical analysis of neuronal variability. Biophys. J.5, 173194.

  • 168

    ThachW. (1996). On the specific role of the cerebellum in motor learning and cognition: clues from PET activation and lesion studies in man. Behav. Brain Sci.19, 411433. 10.1017/S0140525X00081504

  • 169

    Theano Development Team (2016). Theano: A Python Framework for Fast Computation of Mathematical Expressions. arXiv e-prints abs/1605.02688. Available online at: http://arxiv.org/abs/1605.02688

  • 170

    ThorpeS.DelormeA.Van RullenR. (2001). Spike-based strategies for rapid processing. Neural Netw.14, 715725. 10.1016/S0893-6080(01)00083-1

  • 171

    UrbanczikR.SennW. (2009). Reinforcement learning in populations of spiking neurons. Nat. Neurosci.12, 250. 10.1038/nn.2264

  • 172

    VanRullenR.GuyonneauR.ThorpeS. J. (2005). Spike times make sense. Trends Neurosci.28, 14. 10.1016/j.tins.2004.10.010

  • 173

    VasilakiE.FrémauxN.UrbanczikR.SennW.GerstnerW. (2009). Spike-based reinforcement learning in continuous state and action space: when policy gradient methods fail. PLoS Comput. Biol.5:e1000586. 10.1371/journal.pcbi.1000586

  • 174

    VoegtlinT. (2011). Clones: a closed-loop simulation framework for body, muscles and neurons. BMC Neurosci.12:P363. 10.1186/1471-2202-12-S1-P363

  • 175

    VreekenJ. (2003). Spiking Neural Networks, An Introduction. Utrecht: Utrecht University: Information and Computing Sciences.

  • 176

    WadeJ. J.McDaidL. J.SantosJ. A.SayersH. M. (2010). Swat: a spiking neural network training algorithm for classification problems. IEEE Trans. Neural Netw.21, 18171830. 10.1109/TNN.2010.2074212

  • 177

    WalterF.RöhrbeinF.KnollA. (2015a). Computation by time. Neural Process. Lett.44, 103124. 10.1007/s11063-015-9478-6

  • 178

    WalterF.RöhrbeinF.KnollA. (2015b). Neuromorphic implementations of neurobiological learning algorithms for spiking neural networks. Neural Netw.72, 152167. 10.1016/j.neunet.2015.07.004

  • 179

    WangL.YangS. X.BiglarbegianM. (2012). Bio-inspired navigation of mobile robots, in Autonomous and Intelligent Systems, eds KamelM.KarrayF.HagrasH. (Berlin; Heidelberg Springer), 5968.

  • 180

    WangX.HouZ. G.LvF.TanM.WangY. (2014). Mobile robots modular navigation controller using spiking neural networks. Neurocomputing134, 230238. 10.1016/j.neucom.2013.07.055

  • 181

    WangX.HouZ. G.TanM.WangY.HuL. (2009). The wall-following controller for the mobile robot using spiking neurons, in International Conference on Artificial Intelligence and Computational Intelligence, 2009. AICI'09., Vol. 1 (Shanghai: IEEE), 194199.

  • 182

    WangX.HouZ. G.ZouA.TanM.ChengL. (2008). A behavior controller based on spiking neural networks for mobile robots. Neurocomputing71, 655666. 10.1016/j.neucom.2007.08.025

  • 183

    Wikipedia (2017a) Classical Conditioning — Wikipedia the Free Encyclopedia. (Accessed July 05, 2017).

  • 184

    Wikipedia (2017b) Neuron — Wikipedia, the Free Encyclopedia. Available online at: http://en.wikipedia.org/wiki/Neuron (Accessed January 30, 2017).

  • 185

    Wikipedia (2017c) Open Dynamics Engine (ode) — Wikipedia the Free Encyclopedia. (Accessed July 05, 2017).

  • 186

    Wikipedia (2017d) Recursive Neural Network — Wikipedia the Free Encyclopedia. (Accessed July 05, 2017).

  • 187

    Wikipedia (2017e) Synapse — Wikipedia, the Free Encyclopedia. Available online at: en.wikipedia.org/wiki/Synapse (Accessed January 30, 2017).

  • 188

    WolfeJ.HouwelingA. R.BrechtM. (2010). Sparse and powerful cortical spikes. Curr. Opin. Neurobiol.20, 306312. 10.1016/j.conb.2010.03.006

  • 189

    YamazakiT.TanakaS. (2007). The cerebellum as a liquid state machine. Neural Netw.20, 290297. 10.1016/j.neunet.2007.04.004

  • 190

    YuJ.TanM.ChenJ.ZhangJ. (2014). A survey on cpg-inspired control models and system implementation. IEEE Trans. Neural Netw. Learn. Syst.25, 441456. 10.1109/TNNLS.2013.2280596

  • 191

    ZhangX.FoderaroG.HenriquezC.VanDongenA. M.FerrariS. (2012). A radial basis function spike model for indirect learning via integrate-and-fire sampling and reconstruction techniques. Adv. Artif. Neural Syst.2012:713581. 10.1155/2012/713581

  • 192

    ZhangX.XuZ.HenriquezC.FerrariS. (2013). Spike-based indirect training of a spiking neural network-controlled virtual insect. in IEEE 52nd Annual Conference on Decision and Control (Florence: IEEE), 67986805.

Summary

Keywords

spiking neural network, brain-inspired robotics, neurorobotics, learning control, survey

Citation

Bing Z, Meschede C, Röhrbein F, Huang K and Knoll AC (2018) A Survey of Robotics Control Based on Learning-Inspired Spiking Neural Networks. Front. Neurorobot. 12:35. doi: 10.3389/fnbot.2018.00035

Received

19 March 2018

Accepted

14 June 2018

Published

06 July 2018

Volume

12 - 2018

Edited by

Hong Qiao, University of Chinese Academy of Sciences (UCAS), China

Reviewed by

Claudia Casellato, University of Pavia, Italy; Fernando Perez-Peña, University of Cádiz, Spain

Updates

Copyright

*Correspondence: Zhenshan Bing

Disclaimer

All claims expressed in this article are solely those of the authors and do not necessarily represent those of their affiliated organizations, or those of the publisher, the editors and the reviewers. Any product that may be evaluated in this article or claim that may be made by its manufacturer is not guaranteed or endorsed by the publisher.

Outline

Figures

Cite article

Copy to clipboard


Export citation file


Share article

Article metrics