<?xml version="1.0" encoding="UTF-8" standalone="no"?>
<!DOCTYPE article PUBLIC "-//NLM//DTD Journal Publishing DTD v2.3 20070202//EN" "journalpublishing.dtd">
<article xmlns:mml="http://www.w3.org/1998/Math/MathML" xmlns:xlink="http://www.w3.org/1999/xlink" article-type="research-article">
<front>
<journal-meta>
<journal-id journal-id-type="publisher-id">Front. Neurosci.</journal-id>
<journal-title>Frontiers in Neuroscience</journal-title>
<abbrev-journal-title abbrev-type="pubmed">Front. Neurosci.</abbrev-journal-title>
<issn pub-type="epub">1662-453X</issn>
<publisher>
<publisher-name>Frontiers Media S.A.</publisher-name>
</publisher>
</journal-meta>
<article-meta>
<article-id pub-id-type="doi">10.3389/fnins.2020.00634</article-id>
<article-categories>
<subj-group subj-group-type="heading">
<subject>Neuroscience</subject>
<subj-group>
<subject>Original Research</subject>
</subj-group>
</subj-group>
</article-categories>
<title-group>
<article-title>Supervised Learning in All FeFET-Based Spiking Neural Network: Opportunities and Challenges</article-title>
</title-group>
<contrib-group>
<contrib contrib-type="author" corresp="yes">
<name><surname>Dutta</surname> <given-names>Sourav</given-names></name>
<xref ref-type="aff" rid="aff1"><sup>1</sup></xref>
<xref ref-type="corresp" rid="c001"><sup>&#x002A;</sup></xref>
</contrib>
<contrib contrib-type="author">
<name><surname>Schafer</surname> <given-names>Clemens</given-names></name>
<xref ref-type="aff" rid="aff2"><sup>2</sup></xref>
<uri xlink:href="http://loop.frontiersin.org/people/889479/overview"/>
</contrib>
<contrib contrib-type="author">
<name><surname>Gomez</surname> <given-names>Jorge</given-names></name>
<xref ref-type="aff" rid="aff1"><sup>1</sup></xref>
<uri xlink:href="http://loop.frontiersin.org/people/700078/overview"/>
</contrib>
<contrib contrib-type="author">
<name><surname>Ni</surname> <given-names>Kai</given-names></name>
<xref ref-type="aff" rid="aff3"><sup>3</sup></xref>
</contrib>
<contrib contrib-type="author">
<name><surname>Joshi</surname> <given-names>Siddharth</given-names></name>
<xref ref-type="aff" rid="aff2"><sup>2</sup></xref>
<uri xlink:href="http://loop.frontiersin.org/people/14426/overview"/>
</contrib>
<contrib contrib-type="author">
<name><surname>Datta</surname> <given-names>Suman</given-names></name>
<xref ref-type="aff" rid="aff1"><sup>1</sup></xref>
<uri xlink:href="http://loop.frontiersin.org/people/888757/overview"/>
</contrib>
</contrib-group>
<aff id="aff1"><sup>1</sup><institution>Department of Electrical Engineering, College of Engineering, University of Notre Dame</institution>, <addr-line>Notre Dame, IN</addr-line>, <country>United States</country></aff>
<aff id="aff2"><sup>2</sup><institution>Department of Computer Science and Engineering, College of Engineering, University of Notre Dame</institution>, <addr-line>Notre Dame, IN</addr-line>, <country>United States</country></aff>
<aff id="aff3"><sup>3</sup><institution>Department of Microsystems Engineering, Rochester Institute of Technology</institution>, <addr-line>Rochester, NY</addr-line>, <country>United States</country></aff>
<author-notes>
<fn fn-type="edited-by"><p>Edited by: Kaushik Roy, Purdue University, United States</p></fn>
<fn fn-type="edited-by"><p>Reviewed by: Guoqi Li, Tsinghua University, China; Lyes Khacef, Universit&#x00E9; C&#x00F4;te d&#x2019;Azur, France</p></fn>
<corresp id="c001">&#x002A;Correspondence: Sourav Dutta, <email>sdutta4@nd.edu</email></corresp>
<fn fn-type="other" id="fn004"><p>This article was submitted to Neuromorphic Engineering, a section of the journal Frontiers in Neuroscience</p></fn>
</author-notes>
<pub-date pub-type="epub">
<day>24</day>
<month>06</month>
<year>2020</year>
</pub-date>
<pub-date pub-type="collection">
<year>2020</year>
</pub-date>
<volume>14</volume>
<elocation-id>634</elocation-id>
<history>
<date date-type="received">
<day>16</day>
<month>01</month>
<year>2020</year>
</date>
<date date-type="accepted">
<day>22</day>
<month>05</month>
<year>2020</year>
</date>
</history>
<permissions>
<copyright-statement>Copyright &#x00A9; 2020 Dutta, Schafer, Gomez, Ni, Joshi and Datta.</copyright-statement>
<copyright-year>2020</copyright-year>
<copyright-holder>Dutta, Schafer, Gomez, Ni, Joshi and Datta</copyright-holder>
<license xlink:href="http://creativecommons.org/licenses/by/4.0/"><p>This is an open-access article distributed under the terms of the Creative Commons Attribution License (CC BY). The use, distribution or reproduction in other forums is permitted, provided the original author(s) and the copyright owner(s) are credited and that the original publication in this journal is cited, in accordance with accepted academic practice. No use, distribution or reproduction is permitted which does not comply with these terms.</p></license>
</permissions>
<abstract>
<p>The two possible pathways toward artificial intelligence (AI)&#x2014;(i) neuroscience-oriented neuromorphic computing [like spiking neural network (SNN)] and (ii) computer science driven machine learning (like deep learning) differ widely in their fundamental formalism and coding schemes (<xref ref-type="bibr" rid="B58">Pei et al., 2019</xref>). Deviating from traditional deep learning approach of relying on neuronal models with static nonlinearities, SNNs attempt to capture brain-like features like computation using spikes. This holds the promise of improving the energy efficiency of the computing platforms. In order to achieve a much higher areal and energy efficiency compared to today&#x2019;s hardware implementation of SNN, we need to go beyond the traditional route of relying on CMOS-based digital or mixed-signal neuronal circuits and segregation of computation and memory under the von Neumann architecture. Recently, ferroelectric field-effect transistors (FeFETs) are being explored as a promising alternative for building neuromorphic hardware by utilizing their non-volatile nature and rich polarization switching dynamics. In this work, we propose an all FeFET-based SNN hardware that allows low-power spike-based information processing and co-localized memory and computing (a.k.a. in-memory computing). We experimentally demonstrate the essential neuronal and synaptic dynamics in a 28 nm high-K metal gate FeFET technology. Furthermore, drawing inspiration from the traditional machine learning approach of optimizing a cost function to adjust the synaptic weights, we implement a surrogate gradient (SG) learning algorithm on our SNN platform that allows us to perform supervised learning on MNIST dataset. As such, we provide a pathway toward building energy-efficient neuromorphic hardware that can support traditional machine learning algorithms. Finally, we undertake synergistic device-algorithm co-design by accounting for the impacts of device-level variation (stochasticity) and limited bit precision of on-chip synaptic weights (available analog states) on the classification accuracy.</p>
</abstract>
<kwd-group>
<kwd>neuromorphic computing</kwd>
<kwd>supervised learning</kwd>
<kwd>surrogate gradient learning</kwd>
<kwd>ferroelectric FET</kwd>
<kwd>spiking neural network</kwd>
<kwd>spiking neuron</kwd>
<kwd>analog synapse</kwd>
</kwd-group>
<counts>
<fig-count count="5"/>
<table-count count="2"/>
<equation-count count="10"/>
<ref-count count="76"/>
<page-count count="14"/>
<word-count count="0"/>
</counts>
</article-meta>
</front>
<body>
<sec id="S1">
<title>Introduction</title>
<p>Machine learning, especially deep learning has been a <italic>de facto</italic> choice for solving a wide range of real-world complex tasks and has contributed to the unprecedented success story of artificial intelligence (AI) in recent years. Fueled by large datasets and high-performance processors like GPU and TPU, deep learning has exhibited similar or even superior performance compared to human capabilities over a broad spectrum of workloads. However, for applications like smart devices, wearables for healthcare monitoring, or autonomous drones for spatial exploration that require constant real-time information processing, we want to embed implementation of neural networks on the edge. This imposes stringent constraints in terms of power, latency, and footprint area and requires us to rethink the approach toward building hardware for deep learning. Although the architecture of deep neural networks like convolutional neural networks (CNNs) is strongly inspired by the cortical hierarchies, the implementation deviates significantly from the biological counterpart. One obvious point of difference is that neurons are implemented using continuous non-linear functions like sigmoid or ReLu, whereas biological neurons compute using asynchronous spikes that indicate the occurrence of an event. Using such asynchronous event-based information processing may significantly bring down the hardware resources in terms of computational power and footprint area. A recent work established a gain of 54% in area and 45% in power for 65 nm CMOS ASIC implementation of SNN over multi-layer perceptron (MLP) at iso-accuracy and similar architecture (<xref ref-type="bibr" rid="B37">Khacef et al., 2018</xref>). Furthermore, with event-based sensors like visual sensors having reached a matured state (<xref ref-type="bibr" rid="B42">Lichtsteiner et al., 2008</xref>), SNNs provide a natural choice to be interfaced with them. In the last decade, there has been enormous efforts to build and scale up neuromorphic hardware using CMOS based mixed-signal (<xref ref-type="bibr" rid="B7">Benjamin et al., 2014</xref>; <xref ref-type="bibr" rid="B11">Chicca et al., 2014</xref>; <xref ref-type="bibr" rid="B56">Park et al., 2014</xref>; <xref ref-type="bibr" rid="B61">Qiao et al., 2015</xref>) and fully digital (<xref ref-type="bibr" rid="B45">Merolla et al., 2014</xref>; <xref ref-type="bibr" rid="B13">Davies et al., 2018</xref>) designs. However, there lies several considerations for hardware implementation of SNN that must be undertaken to minimize hardware resources (area and energy), some of which are discussed below.</p>
<p>One major consideration is the choice of the neuronal model and its hardware emulation either in analog or digital domain that will ultimately dictate the compactness and energy efficiency. Biological neurons consist of thin lipid layer membrane whose potential is altered by the arrival of excitatory or inhibitory post-synaptic potentials (PSPs) through the dendrites of the neuron. Upon sufficient stimulation, the neuron generates an action potential and the event is commonly referred to as <italic>firing</italic> or <italic>spiking</italic> of the neuron. To emulate these neuronal dynamics in a hardware, including the transient dynamics as well as the mechanism for neurotransmission, the first ingredient of the implementation is an appropriate choice of the neuron model. Although numerous models have been proposed by drawing inspiration from neuroscience like the biologically plausible complex Hodgkin&#x2013;Huxley model (<xref ref-type="bibr" rid="B26">Hodgkin and Huxley, 1952</xref>) and the Izhikevich model (<xref ref-type="bibr" rid="B31">Izhikevich, 2003</xref>), we choose the bio-inspired leaky-integrate-and-fire (LIF) neuron model that provides reduced complexity for hardware implementation while producing the required key dynamics for computation. Spiking LIF neuron can be implemented either in analog or digital domain. While fully digital spiking neurons have been implemented (<xref ref-type="bibr" rid="B45">Merolla et al., 2014</xref>; <xref ref-type="bibr" rid="B13">Davies et al., 2018</xref>), using analog circuits provides an alternative promising pathway. By using transistors biased in the sub-threshold regime, exponential behaviors can be easily mimicked allowing non-discretized continuous-time neural emulation (<xref ref-type="bibr" rid="B28">Indiveri, 2003</xref>; <xref ref-type="bibr" rid="B11">Chicca et al., 2014</xref>; <xref ref-type="bibr" rid="B56">Park et al., 2014</xref>; <xref ref-type="bibr" rid="B61">Qiao et al., 2015</xref>). Recently, <xref ref-type="bibr" rid="B36">Joubert et al. (2012)</xref> provided a quantitative comparison between a digital and analog implementation of LIF neuron at 65 nm CMOS technology node with the same level of performance and established an area and energy benefit of 5x and 20x, respectively, for analog over digital design. One pitfall for analog implementation is, however, the usage of large capacitors for emulating the membrane potential. Even with the most drastically scaled technology node, realizing dense on-chip capacitance comparable to biological neuronal membranes (&#x223C;10<italic>f</italic><italic>F</italic>/&#x03BC;<italic>m</italic><sup>2</sup>; <xref ref-type="bibr" rid="B23">Gentet et al., 2000</xref>) is challenging. For example, <xref ref-type="bibr" rid="B36">Joubert et al. (2012)</xref> implemented the temporal integration property of an analog spiking neuron using a 500 <italic>fF</italic> metal-insulator-metal (MIM) capacitor that requires 100 &#x03BC;<italic>m</italic><sup>2</sup> silicon area while <xref ref-type="bibr" rid="B29">Indiveri et al. (2006)</xref> reports using a 432 <italic>fF</italic> capacitance occupying 244 &#x03BC;<italic>m</italic><sup>2</sup> silicon area. Additionally, biological neurons have been shown to be stochastic and this stochasticity adds to the richness of biological computation. With the recent focus on exploiting the physics of functional materials such as ferroelectrics, magnetics, and phase-change materials to build nano-scale devices that can emulate the characteristics of a low-power, stochastic, and capacitor-less spiking neuron, several proposals have been put forward (<xref ref-type="bibr" rid="B63">Sengupta et al., 2016</xref>; <xref ref-type="bibr" rid="B67">Tuma et al., 2016</xref>; <xref ref-type="bibr" rid="B35">Jerry et al., 2017</xref>). In this work, we experimentally demonstrate the essential neuronal dynamics in a 28 nm ferroelectric field-effect transistor (FeFET) technology with ultra-scaled gate length. The membrane potential is represented using the intrinsic ferroelectric polarization and the rich polarization switching dynamics is utilized to perform temporal integration of post-synaptic spikes, thus mimicking an LIF neuron.</p>
<p>The second consideration is the design of synaptic weight storage. Conventional von-Neumann architecture suffers from time and energy spent in moving data between a centralized memory and the processing units. In contrast, a non-von-Neumann architecture allows computation to be done at the location of the stored synaptic weights, thus circumventing the problem of data-movement. Typical examples of such neuromorphic hardware implementing distributed computing include Intel&#x2019;s Loihi chip with 128 cores each having a local 2 MB static random access memory (SRAM) (<xref ref-type="bibr" rid="B13">Davies et al., 2018</xref>) and IBM&#x2019;s TrueNorth with 4096 neurosynaptic cores each containing 12.75 kB local SRAM (<xref ref-type="bibr" rid="B45">Merolla et al., 2014</xref>; <xref ref-type="bibr" rid="B2">Akopyan et al., 2015</xref>). Additionally, novel techniques such as time-multiplexing has been proposed to reduce hardware resources or facilitate memory usage efficiently (<xref ref-type="bibr" rid="B2">Akopyan et al., 2015</xref>; <xref ref-type="bibr" rid="B13">Davies et al., 2018</xref>; <xref ref-type="bibr" rid="B68">Wang et al., 2018</xref>; <xref ref-type="bibr" rid="B1">Abderrahmane et al., 2020</xref>). Further improvement in energy efficient on-chip training and inference can come from replacing digital SRAM arrays with high density analog synapses that can encode the synaptic weight directly using a physical property of the device such as conductance. Such analog synaptic weight cells can substantially reduce power for both training and inference (<xref ref-type="bibr" rid="B46">Morie and Amemiya, 1994</xref>; <xref ref-type="bibr" rid="B10">Burr et al., 2015</xref>; <xref ref-type="bibr" rid="B24">Gokmen and Vlasov, 2016</xref>). Desirable characteristics of such analog devices include fast and low-power programming of multiple analog states (bit resolution), good retention of the multiple states, and high endurance. Specifically for achieving on-chip training, gradual and symmetric conductance update characteristic is extremely crucial. Recent research efforts have explored numerous potential candidates for building such analog synaptic weight cells including resistive random access memory (RRAM) (<xref ref-type="bibr" rid="B75">Yu et al., 2015</xref>; <xref ref-type="bibr" rid="B22">Gao et al., 2015</xref>; <xref ref-type="bibr" rid="B60">Prezioso et al., 2015</xref>; <xref ref-type="bibr" rid="B71">Wu et al., 2017</xref>), phase-change memory (PCM) (<xref ref-type="bibr" rid="B39">Kuzum et al., 2012</xref>; <xref ref-type="bibr" rid="B10">Burr et al., 2015</xref>; <xref ref-type="bibr" rid="B3">Ambrogio et al., 2018</xref>) and FeFETs (<xref ref-type="bibr" rid="B32">Jerry et al., 2018a</xref>, <xref ref-type="bibr" rid="B34">2019</xref>; <xref ref-type="bibr" rid="B44">Luo et al., 2019</xref>; <xref ref-type="bibr" rid="B65">Sun et al., 2019</xref>). In this work, we provide new experimental results of a FeFET-based synaptic weight cell at scaled device dimensions using 28 nm FeFET technology.</p>
<p>Finally, while deep learning, involving non-spiking and often CNNs, has made remarkable progress in achieving human-like performance at solving complex tasks, similar efficient training algorithms have been challenging to design for SNNs. The difficulty in applying traditional deep learning algorithms stems from various factors. First, the notion of time is an important aspect of SNN. As such, a different cost function has to be used that incorporates the notion of time while learning spatiotemporal patterns rather than what&#x2019;s commonly used in deep learning. Second, spiking neurons are inherently non-differentiable during their time of spike. Over the recent years, several efforts on training SNNs have been undertaken. These include indirect supervised learning like DNN to SNN conversion (<xref ref-type="bibr" rid="B54">O&#x2019;Connor et al., 2013</xref>; <xref ref-type="bibr" rid="B59">P&#x00E9;rez-Carrasco et al., 2013</xref>; <xref ref-type="bibr" rid="B16">Diehl et al., 2015</xref>; <xref ref-type="bibr" rid="B64">Sengupta et al., 2019</xref>), direct supervised learning such as spatiotemporal backpropagation (<xref ref-type="bibr" rid="B73">Wu Y. et al., 2018</xref>; <xref ref-type="bibr" rid="B69">Wu J. et al., 2019</xref>), and unsupervised training of SNNs using bio-inspired local Hebbian learning rule like spike-time-dependent-plasticity (STDP) (<xref ref-type="bibr" rid="B15">Diehl and Cook, 2015</xref>; <xref ref-type="bibr" rid="B55">Panda and Roy, 2016</xref>; <xref ref-type="bibr" rid="B38">Kheradpisheh et al., 2018</xref>). In this work, we focus on the direct supervised learning scheme. Recently, <xref ref-type="bibr" rid="B76">Zenke and Ganguli (2018)</xref> proposed a novel supervised learning algorithm to train multilayer SNNs using a surrogate gradient (SG) based on the membrane potential, known as SuperSpike. In this work, we follow their approach closely by substituting the non-differential derivative of the step-function in the backward pass with a normalized negative part of a fast sigmoid of the membrane potential. Furthermore, we account for the limited bit precision offered by the FeFET-based synapses by considering weight quantization during the training process itself (<xref ref-type="bibr" rid="B70">Wu S. et al., 2018</xref>). We quantize the weights in the forward pass while working with high precision gradients. Previous works have shown that DNNs are very well capable of achieving state-of-the-art results with limited precision weights and activations (<xref ref-type="bibr" rid="B12">Choi et al., 2019</xref>) as well as quantized errors and gradients (<xref ref-type="bibr" rid="B70">Wu S. et al., 2018</xref>). Most modern quantization schemes require additional modification of the learning and inference process by scaling, clipping, or stochastic rounding of variables. <xref ref-type="bibr" rid="B12">Choi et al. (2019)</xref>, for example, train a separate parameter exclusively for activation clipping and compute a scaling factor for the weights that minimize quantization error. However, note that Choi et al. achieve good results with weights quantized using 2 bits in the forward and backward pass by storing and updating a full precision copy of the weights as well as quantizing them under the consideration of the first and second moment of the weight distribution (SAWB). Since spikes only require 1 bit, the energy consumption in our case is mainly driven by weights. Hence, we focus exclusively on the weight quantization and use the weight quantization method as described by <xref ref-type="bibr" rid="B70">Wu S. et al. (2018)</xref> since it imposes marginal quantization overhead. We perform supervised learning on MNIST dataset as an example. We further discuss the impact of FeFET device scaling on the achievable number of analog synaptic weight states (bit resolution) leading to a loss of classification accuracy and potential new avenues for research to circumvent the problem.</p>
</sec>
<sec id="S2" sec-type="materials|methods">
<title>Materials and Methods</title>
<sec id="S2.SS1">
<title>FeFET-Based Analog Adaptive Spiking Neuron</title>
<p>The general working principle of a SNN is as follows. When a synapse receives an action potential, also known as a spike, from its pre-synaptic neuron, it emits a PSP. The PSP in turn stimulates the membrane potential of the post-synaptic neuron. The neuronal membrane potential exhibits temporal evolution where it integrates the PSPs. When the membrane potential crosses a threshold, the post-synaptic neuron fires, i.e., it emits an output spike. <xref ref-type="fig" rid="F1">Figure 1A</xref> illustrates the operation of a simple LIF neuron. Considering a generic LIF neuron, the membrane potential <italic>u</italic> is governed by the following equation:</p>
<disp-formula id="S2.Ex1">
<mml:math id="M1">
<mml:mrow>
<mml:mrow>
<mml:mi mathvariant="normal">&#x03C4;</mml:mi>
<mml:mo>&#x2062;</mml:mo>
<mml:mfrac>
<mml:mrow>
<mml:mi>d</mml:mi>
<mml:mo>&#x2062;</mml:mo>
<mml:mi>u</mml:mi>
</mml:mrow>
<mml:mrow>
<mml:mi>d</mml:mi>
<mml:mo>&#x2062;</mml:mo>
<mml:mi>t</mml:mi>
</mml:mrow>
</mml:mfrac>
</mml:mrow>
<mml:mo>=</mml:mo>
<mml:mrow>
<mml:mrow>
<mml:mi>f</mml:mi>
<mml:mo>&#x2062;</mml:mo>
<mml:mrow>
<mml:mo>(</mml:mo>
<mml:mi>u</mml:mi>
<mml:mo>)</mml:mo>
</mml:mrow>
</mml:mrow>
<mml:mo>+</mml:mo>
<mml:mrow>
<mml:mo movablelimits="false">&#x2211;</mml:mo>
<mml:mrow>
<mml:msub>
<mml:mi>w</mml:mi>
<mml:mi>i</mml:mi>
</mml:msub>
<mml:mo>&#x2062;</mml:mo>
<mml:msub>
<mml:mi>I</mml:mi>
<mml:mi>i</mml:mi>
</mml:msub>
</mml:mrow>
</mml:mrow>
</mml:mrow>
</mml:mrow>
</mml:math>
</disp-formula>
<fig id="F1" position="float">
<label>FIGURE 1</label>
<caption><p><bold>(A)</bold> Schematic of a spiking neural network consisting of an array of plastic synapses and a spiking neuron. A key element of the spiking neuron is the neuronal membrane, which is represented by the intrinsic state variable, i.e., the ferroelectric polarization. <bold>(B)</bold> Schematic of a metal-ferroelectric-metal (MFM) capacitor consisting of 10 nm ferroelectric Hf<sub>x</sub>Zr<sub>1</sub><sub>&#x2013;</sub><sub>x</sub>O<sub>2</sub> thin film sandwiched between two Tungsten (W) metal electrodes that is used to investigating the voltage-dependent polarization switching dynamics. <bold>(C)</bold> Experimentally measured polarization vs electric field (P-E) loop exhibiting saturation as well as minor loops owing to the presence of multiple domains in such ferroelectric thin film. By applying short sub-coercive voltage pulses <bold>(D)</bold>, we can measure the transient polarization switching highlighting the temporal integration of applied voltage pulses and relaxation during the absence of input pulse <bold>(E)</bold>. This inherent polarization dynamics closely resembles the neuronal membrane dynamics of the LIF neuron.</p></caption>
<graphic xlink:href="fnins-14-00634-g001.tif"/>
</fig>
<p>where <italic>f</italic>(<italic>u</italic>) is the leak term accounting for the leakage of accumulated charge in the cell membrane, <italic>w</italic><sub><italic>i</italic></sub> is the synaptic weight, and <italic>I</italic><sub><italic>i</italic></sub> is the input current that depends on the excitatory or inhibitory PSPs. Upon arrival of excitatory input voltage pulses, the membrane potential continuously evolves in time and as it crosses a threshold, the neuronal circuit sends out an output voltage pulse thereby creating a &#x201C;<italic>firing event</italic>.&#x201D; The key idea behind a FeFET-based spiking neuron is to represent the membrane potential <italic>u</italic> by the intrinsic state variable, i.e., ferroelectric polarization instead of the charge stored by a capacitor. As will be discussed next, such dynamics can be achieved within the ferroelectric gate stack of a FeFET that allows realizing compact and low-power spiking neuron.</p>
<sec id="S2.SS1.SSS1">
<title>Polarization Switching Dynamics</title>
<p>We start by investigating the voltage-dependent polarization switching dynamics in a 10nm ferroelectric Hf<sub>x</sub>Zr<sub>1</sub><sub>&#x2013;</sub><sub>x</sub>O<sub>2</sub> thin film sandwiched between two Tungsten (W) metal electrodes. Such a metal-ferroelectric-metal (MFM) capacitor is illustrated in <xref ref-type="fig" rid="F1">Figure 1B</xref>. The fabricated capacitors have lateral dimensions of 80 &#x03BC;<italic>m</italic> &#x00D7; 80 &#x03BC;<italic>m</italic>. <xref ref-type="fig" rid="F1">Figure 1C</xref> shows the experimentally measured polarization vs electric field (P-E) loop exhibiting saturation as well as minor loops owing to the presence of multiple domains in such ferroelectric thin film. Starting from a negative polarization state, where all the dipoles are pointing down, we apply as short voltage pulse. Since the coercive field (V<sub>C</sub>) exhibits a Gaussian distribution in such multidomain thin film, the applied short voltage pulse becomes larger than V<sub>C</sub> in some of the domains leading to a partial polarization switching in the MFM capacitor.</p>
<p>In order to study the temporal evolution of polarization switching, next we apply short sub-coercive voltage pulses (<xref ref-type="fig" rid="F1">Figure 1D</xref>) and measure the net switching current <italic>I</italic><sub><italic>Total</italic></sub> as a function of time. The total measured current <italic>I</italic><sub><italic>Total</italic></sub> will have contribution from two factors&#x2014;the ferroelectric switching and the linear dielectric response. We subtract the contribution from the dielectric portion to reveal the switching dynamics associated with the polarization alone. <xref ref-type="fig" rid="F1">Figure 1E</xref> shows the transient polarization switching dynamics highlighting the temporal integration of applied voltage pulses and relaxation during the absence of input pulse. The neuronal dynamics is emulated by utilizing the ferroelectric polarization accumulation property (<xref ref-type="bibr" rid="B53">Ni et al., 2018</xref>; <xref ref-type="bibr" rid="B62">Saha et al., 2019</xref>) that allows temporal integration of PSP. Such ferroelectric polarization switching dynamics bear close resemblance to that of a LIF spiking neuron. It is intriguing to compare the dynamics of the FeFET-based neuron with a standard LIF neuron realized using dielectric capacitor. It is seen that the integration behavior is similar for both the neurons. However, the leak characteristics indicate a surprisingly opposite behavior. As seen in <xref ref-type="fig" rid="F1">Figure 1E</xref>, the transient relaxation in ferroelectric polarization when the voltage pulse is switched off decreases with the increasing in the number of applied pulses. This is contrary to that of a standard LIF neuron built using linear dielectric capacitor, where the discharge rate of the capacitor increases with the pulse number. Such a deviation in polarization relaxation dynamics can be understood by considering the interaction among the ferroelectric domains within the thin film and has been recently studied using phase-field modeling approach (<xref ref-type="bibr" rid="B19">Dutta et al., 2019b</xref>; <xref ref-type="bibr" rid="B62">Saha et al., 2019</xref>).</p>
</sec>
<sec id="S2.SS1.SSS2">
<title>FeFET Switching Dynamics</title>
<p>Next, we extend the investigation of polarization switching dynamics to FeFETs that consist of a doped-HfO<sub>2</sub> ferroelectric layer integrated into the gate stack of a conventional MOSFET. <xref ref-type="fig" rid="F2">Figure 2A</xref> shows the schematic and TEM of a high-K metal gate FeFET with a poly-Si/TiN/Si:HfO<sub>2</sub>/SiON/p-Si gate stack fabricated at 28 nm technology node (<xref ref-type="bibr" rid="B66">Trentzsch et al., 2017</xref>). All experiments reported here have been performed on FeFETs with channel length of 34 nm and width of 80 nm. On application of successive sub-coercive voltage pulses to the gate of FeFET, the ferroelectric polarization within the Si:HfO<sub>2</sub> layer switches due to an accumulative effect (<xref ref-type="bibr" rid="B47">Mulaosmanovic et al., 2018a</xref>; <xref ref-type="bibr" rid="B53">Ni et al., 2018</xref>; <xref ref-type="bibr" rid="B62">Saha et al., 2019</xref>), resulting in the modulation of the threshold voltage (V<sub>T</sub>) of the FeFET. As seen in <xref ref-type="fig" rid="F2">Figure 2B</xref>, the V<sub>T</sub> gets modulated abruptly from a high-V<sub>T</sub> to low-V<sub>T</sub> state. The abrupt V<sub>T</sub> shift arises due to the presence of very few grains (hence ferroelectric domains) within such a scaled device. This in turn causes an abrupt increase in the drain-to-source channel conductance (G<sub>DS</sub>), thereby exhibiting the temporal integration of PSPs in FeFET. <xref ref-type="fig" rid="F2">Figure 2C</xref> shows the measured conductance modulation as a function of the number of applied pulses over multiple cycles. The cycle-to-cycle variation arises from the nucleation dominated ferroelectric polarization switching in FeFET which at the domain level is known to be a stochastic process (<xref ref-type="bibr" rid="B48">Mulaosmanovic et al., 2018b</xref>; <xref ref-type="bibr" rid="B18">Dutta et al., 2019a</xref>; <xref ref-type="bibr" rid="B51">Ni et al., 2019a</xref>). Once G<sub>DS</sub> exceeds a threshold, the drain current (I<sub>D</sub>) increases and the FeFET is said to &#x201C;<italic>fire</italic>.&#x201D; Once in the low-V<sub>T</sub> state, a negative voltage needs to be applied across the gate and drain/source in order to reset the FeFET to high-V<sub>T</sub> state. Additionally, since the V<sub>T</sub> can be gradually increased as well as decreased by applying positive and negative voltage pulses, respectively, this allows the incorporation of both excitatory (<italic>I</italic> &#x003E; 0) and inhibitory (<italic>I</italic> &#x003C; 0) inputs without any additional circuitry. <xref ref-type="fig" rid="F2">Figure 2D</xref> shows the continuous conductance modulation due to the application of PSPs and how the integrate-and-fire (IF) dynamics repeats after each reset. Owning to the inherent stochasticity, over multiple IF cycles, a single neuron exhibits a distribution of inter-spike intervals for a range of applied input voltage pulse amplitude or width. <xref ref-type="fig" rid="F2">Figures 2E,F</xref> show the distribution of inter-spike interval for a range of voltage amplitudes and the corresponding stochastic firing rate of the neuron. Similar impact of varying the input pulse width on the inter-spike interval and firing rate is shown in <xref ref-type="fig" rid="F2">Figures 2G,H</xref>. Such stochasticity can be harnessed for emulating the probabilistic activity exhibited by biological neurons (<xref ref-type="bibr" rid="B20">Faisal et al., 2008</xref>) without implementing any additional complex circuitry for randomness generation.</p>
<fig id="F2" position="float">
<label>FIGURE 2</label>
<caption><p><bold>(A)</bold> Schematic and TEM of a high-K metal gate FeFET with a poly-Si/TiN/Si:HfO2/SiON/p-Si gate stack fabricated at 28 nm technology node. <bold>(B)</bold> On application of successive sub-coercive voltage pulses to the gate of FeFET, the threshold voltage V<sub>T</sub> gets modulated abruptly from a high-VT to low-VT state. <bold>(C)</bold> Corresponding conductance modulation as a function of number of applied pulses, measured over multiple cycles. <bold>(D)</bold> The integrate-and-fire dynamics of the FeFET neuron. After reaching a conductance threshold, the FeFET is reset to the initial polarization state using a negative gate voltage, which results in a sequence of firing events. <bold>(E,F)</bold> Distribution of inter-spike interval for a range of voltage amplitudes and the corresponding stochastic firing rate of the FeFET neuron. Similar impact of varying the input pulse width on the inter-spike interval and firing rate is seen in <bold>(G,H)</bold>.</p></caption>
<graphic xlink:href="fnins-14-00634-g002.tif"/>
</fig>
</sec>
<sec id="S2.SS1.SSS3">
<title>Implementation of Adaptive Spiking Neuron</title>
<p>We leverage this rich dynamics of the FeFET to implement a low-power spiking neuron circuit consisting of three transistors and one FeFET. Utilizing the temporal integration property of FeFET also allows us to avoid using capacitors for membrane potential, thus providing us an area advantage as well. <xref ref-type="fig" rid="F3">Figure 3A</xref> illustrates the proposed neuron circuit. The input PSPs are applied to the PMOS M1. Initially, both the node voltages V<sub>0</sub> and V<sub>1</sub> are at 0 V. As the PSPs are applied, the node voltage V<sub>0</sub> increases and sub-coercive voltage pulses are applied to the gate of FeFET. Upon application of successive pulses, the FeFET abruptly changes from high-V<sub>T</sub> to low-V<sub>T</sub> state and the drain current I<sub>D</sub> increases. This sends out an output voltage pulse (&#x201C;<italic>spike</italic>&#x201D;) as well as increases the node voltage V<sub>1</sub>. Once an output spike is generated, a reset signal is applied to transistor M3. This external reset, initiated by an arbiter, enables array-based operation often seen in large-scale, event-driven, asynchronous systems such as (<xref ref-type="bibr" rid="B30">Indiveri et al., 2011</xref>; <xref ref-type="bibr" rid="B7">Benjamin et al., 2014</xref>; <xref ref-type="bibr" rid="B56">Park et al., 2014</xref>). With M1 being cut-off during the inter-spike intervals, the node voltage V<sub>0</sub> is pulled down to 0 V. This results in a negative V<sub>GS</sub> across the FeFET, thus switching the polarization in opposite direction and resetting the FeFET to high-V<sub>T</sub> state. We also incorporate bio-inspired homeostatic mechanism that regulates the activity of a neuron and lowers the firing rate after every output spike (<xref ref-type="bibr" rid="B43">Liu and Wang, 2001</xref>; <xref ref-type="bibr" rid="B6">Benda and Herz, 2003</xref>). The homeostatic spike frequency adaptation mechanism is introduced through three additional transistors M4&#x2013;M6 as shown in <xref ref-type="fig" rid="F3">Figure 3A</xref>. The capacitance C<sub>P</sub> can be realized by considering the parasitic capacitance of that node. During every output spike event, as the node voltage V<sub>1</sub> goes high, transistor M4 gets turned on and that in turn increases the node voltage V<sub>2</sub>. The discharge rate of V<sub>2</sub> can be controlled by adding an additional transistor. As V<sub>2</sub> increases, M5 gets turned on gradually with every output spike which in turn increases the discharge rate of node voltage V<sub>0</sub>. Thus, the neuron has to integrate over more input PSPs in order to spike which brings down the neuron&#x2019;s firing rate with every output spiking event, thereby implementing spike frequency adaptation. <xref ref-type="fig" rid="F3">Figure 3B</xref> shows the SPICE circuit simulation of the FeFET-based adaptive spiking neuron. To mimic the stochastic switching dynamics of the FeFET, we introduce a distribution of the coercive field (V<sub>C</sub>) for the ferroelectric domains. We performed Monte Carlo simulation to generate the stochastic spike frequency adaptation as shown in <xref ref-type="fig" rid="F3">Figure 3C</xref>, where the instantaneous firing rate goes down with each output spike. The implication of such a stochastic neuron on the classification accuracy is discussed in section &#x201C;Results.&#x201D;</p>
<fig id="F3" position="float">
<label>FIGURE 3</label>
<caption><p><bold>(A)</bold> Circuit implementation of a FeFET-based spiking neuron. The LIF neuron is implemented using one FeFET and three transistors (M1&#x2013;M3). Biologically inspired homeostatic plasticity is implemented using additional transistors (M4&#x2013;M6). <bold>(B)</bold> SPICE circuit simulation of the FeFET-based LIF spiking neuron with spike frequency adaptation. <bold>(C)</bold> Decrease in the instantaneous firing rate of the neuron with each output spike, exhibiting spike frequency adaptation.</p></caption>
<graphic xlink:href="fnins-14-00634-g003.tif"/>
</fig>
</sec>
</sec>
<sec id="S2.SS2">
<title>FeFET-Based Analog Synapse</title>
<p>The idea of voltage-dependent partial polarization switching in ferroelectric Hf<sub>x</sub>Zr<sub>1</sub><sub>&#x2013;</sub><sub>x</sub>O<sub>2</sub> can be leveraged to implement a non-volatile FeFET-based analog synapse. As illustrated in <xref ref-type="fig" rid="F4">Figure 4A</xref>, the FeFET synapse can be integrated into a pseudo-crossbar array that is suitable for row-wise weight update and column-wise summation. Recently, FeFET-based analog synapse has been experimentally demonstrated on 3 &#x03BC;m long and 20 &#x03BC;m wide devices that exhibited 32 non-volatile states (equivalent to 5-bit precision) and a dynamic range of 45x with amplitude modulated programming pulses (<xref ref-type="bibr" rid="B32">Jerry et al., 2018a</xref>, <xref ref-type="bibr" rid="B33">b</xref>). Here, we provide experimentally measured conductance modulation in a scaled 500 nm &#x00D7; 500 nm high-K metal gate FeFET fabricated at 28 nm technology node (<xref ref-type="bibr" rid="B66">Trentzsch et al., 2017</xref>). As shown in <xref ref-type="fig" rid="F4">Figure 4B</xref>, we used the amplitude modulation scheme with pulse widths of 1 &#x03BC;s to modulate the conductance of the FeFET. Applying progressively increasing gate pulses V<sub>P</sub> causes the FeFET to transition from the initial high-V<sub>T</sub> state to lower V<sub>T</sub> states as shown by the I<sub>D</sub>&#x2013;V<sub>G</sub> characteristics in <xref ref-type="fig" rid="F4">Figure 4B</xref>. The resulting channel conductance G<sub>DS</sub> progressively increases as shown in <xref ref-type="fig" rid="F4">Figure 4C</xref>. However, due to the lateral scaling of the device, the number of ferroelectric domains decreases resulting in a reduced number of non-volatile states. Since the typical grain size in 10 nm HfO<sub>2</sub> is around 10&#x2013;15 nm, it can be estimated that there will be around 1000 domains for a 500 nm &#x00D7; 500 nm FeFET. This also results in cycle-to-cycle (as well as device-to-device) variation, since the stochastic domain switching contribution from individual domains becomes more pronounced (<xref ref-type="bibr" rid="B51">Ni et al., 2019a</xref>). The inherent stochasticity results in a variation of the conductance states measured over multiple cycles for each voltage applied as shown in <xref ref-type="fig" rid="F4">Figure 4C</xref>. We choose eight non-overlapping G<sub>DS</sub> states obtained over multiple cycles using both potentiation and depression pulses as shown in <xref ref-type="fig" rid="F4">Figure 4D</xref> that allowed the representation of a 3-bit equivalent analog weight cell. <xref ref-type="fig" rid="F4">Figure 4E</xref> shows the cumulative distribution of the G<sub>DS</sub> states corresponding to potentiation pulse scheme obtained over multiple cycles. This indicates that while FeFETs are a promising candidate for non-volatile analog synapse, the number of available non-volatile states drastically reduce at the scaled node (<xref ref-type="bibr" rid="B18">Dutta et al., 2019a</xref>; <xref ref-type="bibr" rid="B51">Ni et al., 2019a</xref>). The implications of such a reduced bit-precision on learning algorithms are discussed next. This challenge also opens up new avenues of research both at the material/device/circuit level, as well as at the algorithmic level. For example, a FeFET-based synapse has been recently proposed that utilizes hybrid precision training and inference to overcome the challenge of limited bit precision (<xref ref-type="bibr" rid="B65">Sun et al., 2019</xref>).</p>
<fig id="F4" position="float">
<label>FIGURE 4</label>
<caption><p><bold>(A)</bold> FeFET-based pseudo-crossbar array enabling analog vector&#x2013;matrix multiplication and row-wise parallel weight update of the synaptic weight with column-wise summation. <bold>(B)</bold> Measured Id&#x2013;Vg characteristics of a 500 nm &#x00D7; 500 nm FeFET upon applying amplitude modulated voltage pulses. Applying progressively increasing gate pulses V<sub>P</sub> causes the FeFET to transition from the initial high-V<sub>T</sub> state to lower V<sub>T</sub> states. <bold>(C)</bold> Corresponding channel conductance of the FeFET upon applying progressively increasing gate pulses V<sub>P</sub>, measured over multiple cycles. <bold>(D)</bold> Measured eight non-overlapping G<sub>DS</sub> states obtained over multiple cycles using both potentiation and depression pulses that allows the representation of a 3-bit analog weight cell. <bold>(E)</bold> Cumulative distribution of the G<sub>DS</sub> states corresponding to potentiation pulse scheme obtained over multiple cycles.</p></caption>
<graphic xlink:href="fnins-14-00634-g004.tif"/>
</fig>
</sec>
<sec id="S2.SS3">
<title>Model of FeFET-Based Analog Spiking Neuron and Synapse</title>
<p>The inherent ferroelectric polarization switching dynamics closely resembles the neuronal membrane dynamics of the LIF neuron and can be captured by a modified quasi-LIF neuron model. Our description of the FeFET-based neuron model builds upon the traditional LIF neuron presented by <xref ref-type="bibr" rid="B15">Diehl and Cook (2015)</xref> as given below:</p>
<disp-formula id="S2.Ex2">
<mml:math id="M2">
<mml:mrow>
<mml:mpadded width="+2.8pt">
<mml:mfrac>
<mml:mrow>
<mml:mi>d</mml:mi>
<mml:mo>&#x2062;</mml:mo>
<mml:mi>v</mml:mi>
</mml:mrow>
<mml:mrow>
<mml:mi>d</mml:mi>
<mml:mo>&#x2062;</mml:mo>
<mml:mi>t</mml:mi>
</mml:mrow>
</mml:mfrac>
</mml:mpadded>
<mml:mo rspace="5.3pt">=</mml:mo>
<mml:mrow>
<mml:mpadded width="+2.8pt">
<mml:mfrac>
<mml:mrow>
<mml:mrow>
<mml:mi mathvariant="normal">&#x03B1;</mml:mi>
<mml:mo>&#x2062;</mml:mo>
<mml:msub>
<mml:mi>v</mml:mi>
<mml:mrow>
<mml:mi>r</mml:mi>
<mml:mo>&#x2062;</mml:mo>
<mml:mi>e</mml:mi>
<mml:mo>&#x2062;</mml:mo>
<mml:mi>s</mml:mi>
<mml:mo>&#x2062;</mml:mo>
<mml:mi>t</mml:mi>
</mml:mrow>
</mml:msub>
</mml:mrow>
<mml:mo>-</mml:mo>
<mml:mi>v</mml:mi>
</mml:mrow>
<mml:msub>
<mml:mi mathvariant="normal">&#x03C4;</mml:mi>
<mml:mrow>
<mml:mi>l</mml:mi>
<mml:mo>&#x2062;</mml:mo>
<mml:mi>e</mml:mi>
<mml:mo>&#x2062;</mml:mo>
<mml:mi>a</mml:mi>
<mml:mo>&#x2062;</mml:mo>
<mml:mi>k</mml:mi>
</mml:mrow>
</mml:msub>
</mml:mfrac>
</mml:mpadded>
<mml:mo rspace="5.3pt">+</mml:mo>
<mml:mfrac>
<mml:mrow>
<mml:mo>&#x2211;</mml:mo>
<mml:mrow>
<mml:mo>(</mml:mo>
<mml:mrow>
<mml:mrow>
<mml:msub>
<mml:mi>g</mml:mi>
<mml:mi>e</mml:mi>
</mml:msub>
<mml:mo>&#x2062;</mml:mo>
<mml:mrow>
<mml:mo>(</mml:mo>
<mml:mrow>
<mml:msub>
<mml:mi>E</mml:mi>
<mml:mrow>
<mml:mi>e</mml:mi>
<mml:mo>&#x2062;</mml:mo>
<mml:mi>x</mml:mi>
<mml:mo>&#x2062;</mml:mo>
<mml:mi>c</mml:mi>
</mml:mrow>
</mml:msub>
<mml:mo>-</mml:mo>
<mml:mi>v</mml:mi>
</mml:mrow>
<mml:mo>)</mml:mo>
</mml:mrow>
</mml:mrow>
<mml:mo>+</mml:mo>
<mml:mrow>
<mml:msub>
<mml:mi>g</mml:mi>
<mml:mi>i</mml:mi>
</mml:msub>
<mml:mo>&#x2062;</mml:mo>
<mml:mrow>
<mml:mo>(</mml:mo>
<mml:mrow>
<mml:msub>
<mml:mi>E</mml:mi>
<mml:mrow>
<mml:mi>i</mml:mi>
<mml:mo>&#x2062;</mml:mo>
<mml:mi>n</mml:mi>
<mml:mo>&#x2062;</mml:mo>
<mml:mi>h</mml:mi>
</mml:mrow>
</mml:msub>
<mml:mo>-</mml:mo>
<mml:mi>v</mml:mi>
</mml:mrow>
<mml:mo>)</mml:mo>
</mml:mrow>
</mml:mrow>
</mml:mrow>
<mml:mo>)</mml:mo>
</mml:mrow>
</mml:mrow>
<mml:msub>
<mml:mi mathvariant="normal">&#x03C4;</mml:mi>
<mml:mrow>
<mml:mi>i</mml:mi>
<mml:mo>&#x2062;</mml:mo>
<mml:mi>n</mml:mi>
<mml:mo>&#x2062;</mml:mo>
<mml:mi>t</mml:mi>
<mml:mo>&#x2062;</mml:mo>
<mml:mi>e</mml:mi>
<mml:mo>&#x2062;</mml:mo>
<mml:mi>g</mml:mi>
<mml:mo>&#x2062;</mml:mo>
<mml:mi>r</mml:mi>
<mml:mo>&#x2062;</mml:mo>
<mml:mi>a</mml:mi>
<mml:mo>&#x2062;</mml:mo>
<mml:mi>t</mml:mi>
<mml:mo>&#x2062;</mml:mo>
<mml:mi>e</mml:mi>
</mml:mrow>
</mml:msub>
</mml:mfrac>
</mml:mrow>
</mml:mrow>
</mml:math>
</disp-formula>
<disp-formula id="S2.Ex3">
<mml:math id="M3">
<mml:mrow>
<mml:mrow>
<mml:msub>
<mml:mi mathvariant="normal">&#x03C4;</mml:mi>
<mml:mrow>
<mml:mi>g</mml:mi>
<mml:mo>&#x2062;</mml:mo>
<mml:mi>e</mml:mi>
</mml:mrow>
</mml:msub>
<mml:mo>&#x2062;</mml:mo>
<mml:mpadded width="+2.8pt">
<mml:mfrac>
<mml:mrow>
<mml:mi>d</mml:mi>
<mml:mo>&#x2062;</mml:mo>
<mml:msub>
<mml:mi>g</mml:mi>
<mml:mi>e</mml:mi>
</mml:msub>
</mml:mrow>
<mml:mrow>
<mml:mi>d</mml:mi>
<mml:mo>&#x2062;</mml:mo>
<mml:mi>t</mml:mi>
</mml:mrow>
</mml:mfrac>
</mml:mpadded>
</mml:mrow>
<mml:mo rspace="5.3pt">=</mml:mo>
<mml:mrow>
<mml:mo>-</mml:mo>
<mml:msub>
<mml:mi>g</mml:mi>
<mml:mi>e</mml:mi>
</mml:msub>
</mml:mrow>
</mml:mrow>
</mml:math>
</disp-formula>
<disp-formula id="S2.Ex4">
<mml:math id="M4">
<mml:mrow>
<mml:mrow>
<mml:msub>
<mml:mi mathvariant="normal">&#x03C4;</mml:mi>
<mml:mrow>
<mml:mi>g</mml:mi>
<mml:mo>&#x2062;</mml:mo>
<mml:mi>i</mml:mi>
</mml:mrow>
</mml:msub>
<mml:mo>&#x2062;</mml:mo>
<mml:mpadded width="+2.8pt">
<mml:mfrac>
<mml:mrow>
<mml:mi>d</mml:mi>
<mml:mo>&#x2062;</mml:mo>
<mml:msub>
<mml:mi>g</mml:mi>
<mml:mi>i</mml:mi>
</mml:msub>
</mml:mrow>
<mml:mrow>
<mml:mi>d</mml:mi>
<mml:mo>&#x2062;</mml:mo>
<mml:mi>t</mml:mi>
</mml:mrow>
</mml:mfrac>
</mml:mpadded>
</mml:mrow>
<mml:mo>=</mml:mo>
<mml:mrow>
<mml:mo>-</mml:mo>
<mml:mpadded width="+2.8pt">
<mml:msub>
<mml:mi>g</mml:mi>
<mml:mi>i</mml:mi>
</mml:msub>
</mml:mpadded>
</mml:mrow>
</mml:mrow>
</mml:math>
</disp-formula>
<disp-formula id="S2.Ex5">
<mml:math id="M5">
<mml:mrow>
<mml:mrow>
<mml:msub>
<mml:mi mathvariant="normal">&#x03C4;</mml:mi>
<mml:mi mathvariant="normal">&#x03B1;</mml:mi>
</mml:msub>
<mml:mo>&#x2062;</mml:mo>
<mml:mpadded width="+2.8pt">
<mml:mfrac>
<mml:mrow>
<mml:mi>d</mml:mi>
<mml:mo>&#x2062;</mml:mo>
<mml:mi mathvariant="normal">&#x03B1;</mml:mi>
</mml:mrow>
<mml:mrow>
<mml:mi>d</mml:mi>
<mml:mo>&#x2062;</mml:mo>
<mml:mi>t</mml:mi>
</mml:mrow>
</mml:mfrac>
</mml:mpadded>
</mml:mrow>
<mml:mo>=</mml:mo>
<mml:mrow>
<mml:mo>-</mml:mo>
<mml:mpadded width="+2.8pt">
<mml:msub>
<mml:mi>g</mml:mi>
<mml:mi>e</mml:mi>
</mml:msub>
</mml:mpadded>
</mml:mrow>
</mml:mrow>
</mml:math>
</disp-formula>
<p>where <italic>v</italic> is the membrane potential, <italic>v</italic><sub><italic>rest</italic></sub> denotes the resting potential of the neuron, and <italic>E</italic><sub><italic>e</italic><italic>x</italic><italic>c</italic></sub> and <italic>E</italic><sub><italic>i</italic><italic>n</italic><italic>h</italic></sub> are the equilibrium potentials of excitatory and inhibitory synapses. &#x03C4;<sub><italic>leak</italic></sub> and &#x03C4;<sub><italic>integrate</italic></sub> are the time constants associated with for the leakage and integration phase of the neuron, respectively. When the neuron&#x2019;s membrane potential crosses the membrane threshold <italic>v</italic><sub><italic>thres</italic></sub>, the neuron fires and the membrane potential is reset to <italic>v</italic><sub><italic>reset</italic></sub>. We incorporate the quasi-leaky behavior (decrease in the leak rate of the neuron during the inter-spike interval) into the neuron model by using a variable resetting voltage by multiplying <italic>v</italic><sub><italic>reset</italic></sub> with a parameter &#x03B1; that changes with each incoming spike. Additionally, this quasi-leak behavior can also be incorporated into the model by using a variable &#x03C4;<sub><italic>leak</italic></sub> that also depends on the membrane potential <italic>v</italic> (<xref ref-type="bibr" rid="B19">Dutta et al., 2019b</xref>). However, owing to very small relaxation dynamics, one can also ignore the leaky behavior of the FeFET-based neuron and treat it as a perfect IF neuron (<xref ref-type="bibr" rid="B47">Mulaosmanovic et al., 2018a</xref>). Furthermore, we use an adaptive threshold regime to regulate the neuron&#x2019;s activity. Once a neuron hits the threshold and issues a spike, this neuron&#x2019;s threshold increases by a fixed amount, thereby making it harder for this neuron to spike again and prioritizing activities of other neurons. However, the threshold increase only happens until the neuron threshold reaches a maximum level, after which the threshold is not changed by issued spikes anymore.</p>
<p>Synapse models have been incorporated following <xref ref-type="bibr" rid="B15">Diehl and Cook (2015)</xref> where the synaptic conductance changes instantaneously by weight <italic>w</italic> when a presynaptic spike arrives at the synapse, else the conductance decays exponentially. <italic>g</italic><sub><italic>e</italic></sub> and <italic>g</italic><sub><italic>i</italic></sub> are the conductances of the excitatory and inhibitory synapse, respectively. &#x03C4;<sub><italic>ge</italic></sub> and &#x03C4;<sub><italic>gi</italic></sub> are the time constants of the excitatory and inhibitory PSPs, respectively.</p>
<p>This model is then discretized with a standard Euler method so we can use discrete time steps in our simulation. The discrete time version of the models is expressed as:</p>
<disp-formula id="S2.Ex6">
<mml:math id="M6">
<mml:mrow>
<mml:mi>v</mml:mi>
<mml:mo>&#x2062;</mml:mo>
<mml:mrow>
<mml:mo>[</mml:mo>
<mml:mrow>
<mml:mi>n</mml:mi>
<mml:mo>+</mml:mo>
<mml:mn>1</mml:mn>
</mml:mrow>
<mml:mo>]</mml:mo>
</mml:mrow>
</mml:mrow>
<mml:mo>=</mml:mo>
<mml:mrow>
<mml:mrow>
<mml:mi mathvariant="normal">&#x03B1;</mml:mi>
<mml:mo>&#x2062;</mml:mo>
<mml:msub>
<mml:mi>E</mml:mi>
<mml:mrow>
<mml:mi>r</mml:mi>
<mml:mo>&#x2062;</mml:mo>
<mml:mi>e</mml:mi>
<mml:mo>&#x2062;</mml:mo>
<mml:mi>s</mml:mi>
<mml:mo>&#x2062;</mml:mo>
<mml:mi>t</mml:mi>
</mml:mrow>
</mml:msub>
</mml:mrow>
<mml:mo>+</mml:mo>
<mml:mrow>
<mml:mi mathvariant="normal">&#x03B2;</mml:mi>
<mml:mo>&#x2062;</mml:mo>
<mml:mi>v</mml:mi>
<mml:mo>&#x2062;</mml:mo>
<mml:mrow>
<mml:mo>[</mml:mo>
<mml:mi>n</mml:mi>
<mml:mo>]</mml:mo>
</mml:mrow>
</mml:mrow>
<mml:mo>+</mml:mo>
<mml:mrow>
<mml:msub>
<mml:mi>g</mml:mi>
<mml:mi>e</mml:mi>
</mml:msub>
<mml:mo>&#x2062;</mml:mo>
<mml:mrow>
<mml:mo>[</mml:mo>
<mml:mi>n</mml:mi>
<mml:mo>]</mml:mo>
</mml:mrow>
<mml:mo>&#x2062;</mml:mo>
<mml:msub>
<mml:mi>E</mml:mi>
<mml:mrow>
<mml:mi>e</mml:mi>
<mml:mo>&#x2062;</mml:mo>
<mml:mi>x</mml:mi>
<mml:mo>&#x2062;</mml:mo>
<mml:mi>c</mml:mi>
</mml:mrow>
</mml:msub>
</mml:mrow>
</mml:mrow>
<mml:mrow>
<mml:mrow>
<mml:mo>+</mml:mo>
<mml:mrow>
<mml:mi mathvariant="normal">&#x03B2;</mml:mi>
<mml:mo>&#x2062;</mml:mo>
<mml:msub>
<mml:mi>g</mml:mi>
<mml:mi>e</mml:mi>
</mml:msub>
<mml:mo>&#x2062;</mml:mo>
<mml:mrow>
<mml:mo>[</mml:mo>
<mml:mi>n</mml:mi>
<mml:mo>]</mml:mo>
</mml:mrow>
<mml:mo>&#x2062;</mml:mo>
<mml:mi>v</mml:mi>
<mml:mo>&#x2062;</mml:mo>
<mml:mrow>
<mml:mo>[</mml:mo>
<mml:mi>n</mml:mi>
<mml:mo>]</mml:mo>
</mml:mrow>
</mml:mrow>
</mml:mrow>
<mml:mo>+</mml:mo>
<mml:mrow>
<mml:msub>
<mml:mi>g</mml:mi>
<mml:mi>i</mml:mi>
</mml:msub>
<mml:mo>&#x2062;</mml:mo>
<mml:mrow>
<mml:mo>[</mml:mo>
<mml:mi>n</mml:mi>
<mml:mo>]</mml:mo>
</mml:mrow>
<mml:mo>&#x2062;</mml:mo>
<mml:msub>
<mml:mi>E</mml:mi>
<mml:mrow>
<mml:mi>i</mml:mi>
<mml:mo>&#x2062;</mml:mo>
<mml:mi>n</mml:mi>
<mml:mo>&#x2062;</mml:mo>
<mml:mi>h</mml:mi>
</mml:mrow>
</mml:msub>
</mml:mrow>
<mml:mo>+</mml:mo>
<mml:mrow>
<mml:mi mathvariant="normal">&#x03B2;</mml:mi>
<mml:mo>&#x2062;</mml:mo>
<mml:msub>
<mml:mi>g</mml:mi>
<mml:mi>i</mml:mi>
</mml:msub>
<mml:mo>&#x2062;</mml:mo>
<mml:mrow>
<mml:mo>[</mml:mo>
<mml:mi>n</mml:mi>
<mml:mo>]</mml:mo>
</mml:mrow>
<mml:mo>&#x2062;</mml:mo>
<mml:mi>v</mml:mi>
<mml:mo>&#x2062;</mml:mo>
<mml:mrow>
<mml:mo>[</mml:mo>
<mml:mi>n</mml:mi>
<mml:mo>]</mml:mo>
</mml:mrow>
</mml:mrow>
</mml:mrow>
</mml:math>
</disp-formula>
<disp-formula id="S2.Ex8">
<mml:math id="M7">
<mml:mrow>
<mml:mrow>
<mml:msub>
<mml:mi>g</mml:mi>
<mml:mi>e</mml:mi>
</mml:msub>
<mml:mo>&#x2062;</mml:mo>
<mml:mrow>
<mml:mo>[</mml:mo>
<mml:mrow>
<mml:mi>n</mml:mi>
<mml:mo>+</mml:mo>
<mml:mn>1</mml:mn>
</mml:mrow>
<mml:mo>]</mml:mo>
</mml:mrow>
</mml:mrow>
<mml:mo rspace="5.3pt">=</mml:mo>
<mml:mrow>
<mml:mpadded width="+2.8pt">
<mml:msup>
<mml:mi>e</mml:mi>
<mml:mrow>
<mml:mo>-</mml:mo>
<mml:mfrac>
<mml:msub>
<mml:mo>&#x0394;</mml:mo>
<mml:mi>t</mml:mi>
</mml:msub>
<mml:msub>
<mml:mi mathvariant="normal">&#x03C4;</mml:mi>
<mml:msub>
<mml:mi>g</mml:mi>
<mml:mi>e</mml:mi>
</mml:msub>
</mml:msub>
</mml:mfrac>
</mml:mrow>
</mml:msup>
</mml:mpadded>
<mml:mo>&#x2062;</mml:mo>
<mml:msub>
<mml:mi>g</mml:mi>
<mml:mi>e</mml:mi>
</mml:msub>
<mml:mo>&#x2062;</mml:mo>
<mml:mrow>
<mml:mo stretchy="false">[</mml:mo>
<mml:mi>n</mml:mi>
<mml:mo stretchy="false">]</mml:mo>
</mml:mrow>
</mml:mrow>
</mml:mrow>
</mml:math>
</disp-formula>
<disp-formula id="S2.Ex9">
<mml:math id="M8">
<mml:mrow>
<mml:mrow>
<mml:msub>
<mml:mi>g</mml:mi>
<mml:mi>i</mml:mi>
</mml:msub>
<mml:mo>&#x2062;</mml:mo>
<mml:mrow>
<mml:mo>[</mml:mo>
<mml:mrow>
<mml:mi>n</mml:mi>
<mml:mo>+</mml:mo>
<mml:mn>1</mml:mn>
</mml:mrow>
<mml:mo>]</mml:mo>
</mml:mrow>
</mml:mrow>
<mml:mo rspace="5.3pt">=</mml:mo>
<mml:mrow>
<mml:mpadded width="+2.8pt">
<mml:msup>
<mml:mi>e</mml:mi>
<mml:mrow>
<mml:mo>-</mml:mo>
<mml:mfrac>
<mml:msub>
<mml:mo>&#x0394;</mml:mo>
<mml:mi>t</mml:mi>
</mml:msub>
<mml:msub>
<mml:mi mathvariant="normal">&#x03C4;</mml:mi>
<mml:msub>
<mml:mi>g</mml:mi>
<mml:mi>i</mml:mi>
</mml:msub>
</mml:msub>
</mml:mfrac>
</mml:mrow>
</mml:msup>
</mml:mpadded>
<mml:mo>&#x2062;</mml:mo>
<mml:msub>
<mml:mi>g</mml:mi>
<mml:mi>i</mml:mi>
</mml:msub>
<mml:mo>&#x2062;</mml:mo>
<mml:mrow>
<mml:mo stretchy="false">[</mml:mo>
<mml:mi>n</mml:mi>
<mml:mo stretchy="false">]</mml:mo>
</mml:mrow>
</mml:mrow>
</mml:mrow>
</mml:math>
</disp-formula>
<disp-formula id="S2.Ex10">
<mml:math id="M9">
<mml:mrow>
<mml:mrow>
<mml:mi mathvariant="normal">&#x03B1;</mml:mi>
<mml:mo>&#x2062;</mml:mo>
<mml:mrow>
<mml:mo>[</mml:mo>
<mml:mrow>
<mml:mi>n</mml:mi>
<mml:mo>+</mml:mo>
<mml:mn>1</mml:mn>
</mml:mrow>
<mml:mo>]</mml:mo>
</mml:mrow>
</mml:mrow>
<mml:mo rspace="5.3pt">=</mml:mo>
<mml:mrow>
<mml:mpadded width="+2.8pt">
<mml:msup>
<mml:mi>e</mml:mi>
<mml:mrow>
<mml:mo>-</mml:mo>
<mml:mfrac>
<mml:msub>
<mml:mo>&#x0394;</mml:mo>
<mml:mi>t</mml:mi>
</mml:msub>
<mml:msub>
<mml:mi mathvariant="normal">&#x03C4;</mml:mi>
<mml:mi mathvariant="normal">&#x03B1;</mml:mi>
</mml:msub>
</mml:mfrac>
</mml:mrow>
</mml:msup>
</mml:mpadded>
<mml:mo>&#x2062;</mml:mo>
<mml:msub>
<mml:mi>g</mml:mi>
<mml:mi>e</mml:mi>
</mml:msub>
<mml:mo>&#x2062;</mml:mo>
<mml:mrow>
<mml:mo stretchy="false">[</mml:mo>
<mml:mi>n</mml:mi>
<mml:mo stretchy="false">]</mml:mo>
</mml:mrow>
</mml:mrow>
</mml:mrow>
</mml:math>
</disp-formula>
<p>where <italic>v</italic>[<italic>n</italic>] is the discretized membrane potential of the neuron at time step <italic>n</italic>. We use a single membrane time constant &#x03C4;<sub><italic>v</italic></sub>, accounting for both &#x03C4;<sub><italic>leak</italic></sub> and &#x03C4;<sub><italic>integrate</italic></sub>. <inline-formula><mml:math id="INEQ10"><mml:mrow><mml:mi mathvariant="normal">&#x03B2;</mml:mi><mml:mo>=</mml:mo><mml:msup><mml:mi>e</mml:mi><mml:mrow><mml:mo>-</mml:mo><mml:mfrac><mml:msub><mml:mo>&#x0394;</mml:mo><mml:mi>t</mml:mi></mml:msub><mml:msub><mml:mi mathvariant="normal">&#x03C4;</mml:mi><mml:mi>v</mml:mi></mml:msub></mml:mfrac></mml:mrow></mml:msup></mml:mrow></mml:math></inline-formula> captures the decay in the membrane potential during a &#x0394;<sub><italic>t</italic></sub> time step.</p>
</sec>
<sec id="S2.SS4">
<title>Supervised Learning for SNNs</title>
<p>The success of deep learning in recent years has largely been attributed to the power of supervised learning techniques and gradient based learning (<xref ref-type="bibr" rid="B40">Lecun et al., 2015</xref>). Given an objective function, backpropagation adjusts parameters and weights of a given network so that its objective function is minimized. In DNNs, weights are updated in multiple layers organized hierarchically enabling it to learn complex classification or regression functions. Two critical challenges must be overcome in order for SNNs to gain similar success: (a) The development of hierarchical &#x201C;deep&#x201D; networks like (<xref ref-type="bibr" rid="B55">Panda and Roy, 2016</xref>; <xref ref-type="bibr" rid="B38">Kheradpisheh et al., 2018</xref>) which can learn complex representations and (b) enabling the application of gradient based learning to deep spiking neural networks (SSNs). Several studies have proposed ways to train SSNs in order to address the above challenges, such as (<xref ref-type="bibr" rid="B25">G&#x00FC;tig, 2014</xref>; <xref ref-type="bibr" rid="B4">Anwani and Rajendran, 2015</xref>). However, these approaches did not sufficiently enhance the representative power of SNNs, nor did they enable the development of deeper more complex networks. <xref ref-type="bibr" rid="B50">Neftci et al. (2019)</xref> identified four streams of research attempting to train SNNs with hidden units: (i) biologically inspired local learning rules, (ii) translating conventionally trained &#x201C;rate-based&#x201D; neural networks to SNNs, (iii) smoothing the network model to be continuously differentiable, and iv) defining a SG as a continuous relaxation. Fortuitously, the SG-based methods simultaneously address the two challenges presented and form the basis for the remainder of this article. This method allows us to build upon the solid research base of backpropagation with only marginal modifications of the spiking network model.</p>
<sec id="S2.SS4.SSS1">
<title>Surrogate Gradient Learning</title>
<p>Historically, the spike function in SNNs prevented the application of gradient based learning rules due to the discontinuities induced by the non-differentiable spikes, consequently &#x201C;<italic>stopping the gradient from flowing</italic>.&#x201D; This in turn results in backpropagation failing to function correctly. The SG method substitutes the gradient in the backward pass of the backpropagation with a differentiable proxy or surrogate. This surrogate is generally based on the membrane potential of a neuron. As a result of this new gradient, the non-differentiability is circumvented, and the gradient can propagate. Using this gradient-based update rule, standard solutions to the credit assignment problem can be applied. Thus, given a global loss function, we can apply traditional gradient-based learning methods such as backpropagation through time (BPTT) (<xref ref-type="bibr" rid="B27">Huh and Sejnowski, 2018</xref>) or other learning rules such as three factor learning rules (<xref ref-type="bibr" rid="B76">Zenke and Ganguli, 2018</xref>) to SNNs. Most modern machine learning libraries (e.g., PyTorch or TensorFlow) provide autograd functionalities which facilitate the gradient computation. Our experiments used BPTT as gradient-based learning method in conjunction with a SG in the backward pass. The SGs were computed by applying the normalized negative part of a fast sigmoid on the membrane potential.</p>
</sec>
<sec id="S2.SS4.SSS2">
<title>Quantization</title>
<p>As highlighted earlier, the on-chip FeFET-based analog synapse provides limited bit precision ranging from 5 to 3 bits in scaled devices. Efficient implementation of (i) off-chip training followed by reduced bit precision for inference mode and (ii) on-chip learning and inference with reduced bit precision, both demand efficient training algorithm taking into consideration the quantization in synaptic weights. To accurately model the effect of quantizing the weights, we follow the procedure outlined in <xref ref-type="bibr" rid="B70">Wu S. et al. (2018)</xref>. Weights are quantized by restricting them to a feasible range [&#x2212;1 + &#x03C3;(<italic>b</italic><sub><italic>w</italic></sub>), +1&#x2212;&#x03C3;(<italic>b</italic><sub><italic>w</italic></sub>)], where &#x03C3;(<italic>b</italic>) = 2<sup>1&#x2212;<italic>b</italic></sup> and <italic>b</italic><sub><italic>w</italic></sub> is the number of bits encoding the weight. Weights in each layer are further scaled by &#x03B3;:</p>
<disp-formula id="S2.Ex11">
<mml:math id="M10">
<mml:mrow>
<mml:mi mathvariant="normal">&#x03B3;</mml:mi>
<mml:mo>=</mml:mo>
<mml:msup>
<mml:mn>2</mml:mn>
<mml:mrow>
<mml:mi>r</mml:mi>
<mml:mo>&#x2062;</mml:mo>
<mml:mi>o</mml:mi>
<mml:mo>&#x2062;</mml:mo>
<mml:mi>u</mml:mi>
<mml:mo>&#x2062;</mml:mo>
<mml:mi>n</mml:mi>
<mml:mo>&#x2062;</mml:mo>
<mml:mi>d</mml:mi>
<mml:mo>&#x2062;</mml:mo>
<mml:mrow>
<mml:mo>[</mml:mo>
<mml:mrow>
<mml:msub>
<mml:mi>log</mml:mi>
<mml:mn>2</mml:mn>
</mml:msub>
<mml:mo>&#x2061;</mml:mo>
<mml:mrow>
<mml:mo>(</mml:mo>
<mml:mfrac>
<mml:mrow>
<mml:mrow>
<mml:mo>(</mml:mo>
<mml:mrow>
<mml:mfrac>
<mml:mn>1</mml:mn>
<mml:mrow>
<mml:mi mathvariant="normal">&#x03C3;</mml:mi>
<mml:mo>&#x2062;</mml:mo>
<mml:mrow>
<mml:mo stretchy="false">(</mml:mo>
<mml:msub>
<mml:mi>b</mml:mi>
<mml:mi>w</mml:mi>
</mml:msub>
<mml:mo stretchy="false">)</mml:mo>
</mml:mrow>
</mml:mrow>
</mml:mfrac>
<mml:mo>-</mml:mo>
<mml:mn>0.5</mml:mn>
</mml:mrow>
<mml:mo>)</mml:mo>
</mml:mrow>
<mml:mo>.</mml:mo>
<mml:mrow>
<mml:mi mathvariant="normal">&#x03C3;</mml:mi>
<mml:mo>&#x2062;</mml:mo>
<mml:mrow>
<mml:mo stretchy="false">(</mml:mo>
<mml:msub>
<mml:mi>b</mml:mi>
<mml:mi>w</mml:mi>
</mml:msub>
<mml:mo stretchy="false">)</mml:mo>
</mml:mrow>
</mml:mrow>
</mml:mrow>
<mml:msqrt>
<mml:mfrac>
<mml:mn>3</mml:mn>
<mml:mrow>
<mml:mi>f</mml:mi>
<mml:mo>&#x2062;</mml:mo>
<mml:mi>a</mml:mi>
<mml:mo>&#x2062;</mml:mo>
<mml:mpadded width="+2.8pt">
<mml:mi>n</mml:mi>
</mml:mpadded>
<mml:mo>&#x2062;</mml:mo>
<mml:mi>i</mml:mi>
<mml:mo>&#x2062;</mml:mo>
<mml:mi>n</mml:mi>
</mml:mrow>
</mml:mfrac>
</mml:msqrt>
</mml:mfrac>
<mml:mo>)</mml:mo>
</mml:mrow>
</mml:mrow>
<mml:mo>]</mml:mo>
</mml:mrow>
</mml:mrow>
</mml:msup>
</mml:mrow>
</mml:math>
</disp-formula>
<p>where <italic>f</italic><italic>a</italic><italic>n</italic><italic>i</italic><italic>n</italic> represents the number of connections into a layer. Weights are also uniformly initialized in their feasible range and clipped to the range after each update during training.</p>
</sec>
</sec>
</sec>
<sec id="S3">
<title>Results</title>
<p>We performed supervised learning on MNIST dataset (using the standard train/test split of 60,000/10,000) using a three-layer SNN as shown in <xref ref-type="fig" rid="F5">Figure 5A</xref>. The input layer consists of 784 neurons while the output layer has 10 neurons to classify the digits. The number of hidden layer units is an architectural question which can have significant impact on the performance. We used a Bayesian hyperparameter optimization approach (<xref ref-type="bibr" rid="B8">Bergstra et al., 2013</xref>) to determine the number of hidden layer neurons, learning rate, input multiplier (scaling of the input spikes), scaling coefficient for the &#x03C4;<sub><italic>leak</italic></sub> in relation to &#x03C4;<sub><italic>integrate</italic></sub>, size of the regularizer, batch size, and saturation threshold. All these parameters have an impact on the performance and are sensitive to the dataset of the network. We use the saturation threshold method taken from <xref ref-type="bibr" rid="B74">Yousefzadeh et al. (2018)</xref> which prevents the firing threshold from being further increased once it surpasses the saturation threshold, e.g., once it issued a certain number of spikes. For the hyperparameter optimization, we gave the optimizer ranges for the mentioned parameters and programmed it to train a sampled configuration for 12 epochs. Overall, we constrained the optimizer to use 75 evaluations and come up with the best configuration. The number of discrete time steps remained fixed at 80. In our simulations, we used a negative loglikelihood loss function, which performed classification by integrating the last layer&#x2019;s neurons membrane potential over time and selecting the class of the neuron with the largest integration value.</p>
<fig id="F5" position="float">
<label>FIGURE 5</label>
<caption><p><bold>(A)</bold> We performed supervised learning on MNIST dataset using a three-layer SNN. The input layer consists of 784 neurons while the output layer has 10 neurons to classify the digits. We additionally use a Bayesian hyperparameter optimization approach (<xref ref-type="bibr" rid="B8">Bergstra et al., 2013</xref>) to determine the number of neurons in the hidden layer. <bold>(B)</bold> Training accuracy as a function of the number of training epochs for different synaptic weight bit precision (analog states). <bold>(C)</bold> Impact of stochastic noise (incorporated as uniform threshold noise) on the test accuracy for different weight bit precision. <bold>(D)</bold> Comparison of test accuracy for different weight bit precision with and without noise.</p></caption>
<graphic xlink:href="fnins-14-00634-g005.tif"/>
</fig>
<p>Since the number of analog states that can be represented by a single FeFET-based synaptic weight cell decreases as we scale the device, it is important to consider the impact of bit precision of the synaptic weights (number of analog states) on training as well as test accuracy. Hence, in our simulation, we varied the weight bit precision from a high value of 8 bits (that would require a single FeFET to represent 256 analog states) to 5 bits (32 analog states demonstrated in a 3 &#x03BC;m &#x00D7; 20 &#x03BC;m device (<xref ref-type="bibr" rid="B32">Jerry et al., 2018a</xref>, <xref ref-type="bibr" rid="B33">b</xref>) down to 3 bits (eight analog states demonstrated in this work). We also compared the results against the full 32-bit floating point precision available on a CPU. <xref ref-type="fig" rid="F5">Figure 5B</xref> shows the training accuracy as a function of the number of training epochs for various precision of the synaptic weight without introducing any stochastic noise in the simulation. It is seen that while up to 6 bits, we get a test accuracy of 95.4% which is comparable (or even better) to that of the full precision accuracy of 95.1%. The accuracy starts decreasing to 91% for 5 bits and drastically down to 43.5% for 4 bit precision. The increase of accuracy at 6 bits can be seen as the regularizing effect of more coarse weights. The sharp decrease of accuracy for 5 or fewer bits likely indicate the tolerance threshold of SNNs toward reduced weight precision and the lack of information in spikes coupled with weights after a certain weight quantization level. Note that previous works like Choi et al. achieve good results with 2-bit weight quantization in the forward and backward pass by storing and updating a full precision copy of the weights as well as quantizing them under the consideration of the first and second moment of the weight distribution (SAWB). In contrast, our results are obtained with only quantized weights in both the forward and backward pass as well as a linear quantization step reflecting the capabilities of our proposed device.</p>
<p>We further studied the impact of stochastic neurons on the overall performance of the SNN by introducing a uniform noise around the membrane threshold which can be mimicked by the stochastic neuronal dynamics (as shown in <xref ref-type="fig" rid="F2">Figure 2</xref>). <xref ref-type="fig" rid="F5">Figure 5C</xref> shows the impact of noise on the test accuracy. The accuracy for 5&#x2013;8 bits increased to 96%. Interesting the accuracy for 4-bit weights improved substantially with more noise around the threshold which is in accordance with previous works on ordinary quantized DNNs (<xref ref-type="bibr" rid="B70">Wu S. et al., 2018</xref>; <xref ref-type="bibr" rid="B12">Choi et al., 2019</xref>). Over a population of neurons and multiple times steps, the threshold with more noise becomes more like a soft function (e.g., sigmoid, softmax, or tanh) rather than a hard threshold and hence more similar to ordinary DNNs which allows for reduced weight precision. In the case of 3-bit weights, we were not able to compensate for the granularity of the weights with noise around the threshold. <xref ref-type="fig" rid="F5">Figure 5D</xref> shows the testing accuracy as a function of various weight precision with and without noise indicating that having a stochastic SNN helps improve the classification accuracy in the presence of reduced weight precision. As mentioned earlier, another way to further improve the accuracy will be to resort to a CMOS-augmented FeFET-based hybrid synapse design that can provide hybrid precision training and inference to overcome the challenge of limited bit precision (<xref ref-type="bibr" rid="B65">Sun et al., 2019</xref>).</p>
</sec>
<sec id="S4">
<title>Discussion</title>
<p>In this work, we exploit the rich dynamics of ferroelectric polarization switching in FeFET to realize compact and low-power analog spiking neuron and synapse. The membrane potential of the spiking neuron is represented by the intrinsic ferroelectric polarization of the FeFET. The neuronal dynamics is emulated by utilizing the polarization accumulation property (<xref ref-type="bibr" rid="B53">Ni et al., 2018</xref>; <xref ref-type="bibr" rid="B62">Saha et al., 2019</xref>) that allows temporal integration of PSP. This allows the realization of a capacitor-less analog spiking neuron which proves to be compact and low power. <xref ref-type="table" rid="T1">Table 1</xref> shows a comparative study between our FeFET-based analog spiking neuron and various other proposals. Compared to CMOS-based realization of LIF neuron that requires more than 20 transistors and an explicit capacitor, our proposal of FeFET-based spiking neuron requires seven transistors including one FeFET. To estimate the areal requirements, we performed a layout using a 45nm technology node. The estimated area is approximately 1.74 &#x00D7; 1.18 &#x03BC;<italic>m</italic><sup>2</sup> which would be much smaller than the capacitor-based CMOS circuits. For example, <xref ref-type="bibr" rid="B36">Joubert et al. (2012)</xref> realized an analog spiking neuron with a footprint area of 120 &#x03BC;<italic>m</italic><sup>2</sup> at 65 nm technology node, of which 100 &#x03BC;<italic>m</italic><sup>2</sup> was dedicated to realizing the 500 <italic>fF</italic> capacitor. Our estimated footprint area in terms of feature size F is at least 4x lower than this. Similarly, <xref ref-type="bibr" rid="B29">Indiveri et al. (2006)</xref> report using a 432 <italic>fF</italic> capacitance occupying 244 &#x03BC;<italic>m</italic><sup>2</sup> silicon area. The energy dissipated by our FeFET-based analog neuron is comparable to the analog neuron implementation by <xref ref-type="bibr" rid="B36">Joubert et al. (2012)</xref> while it is 4x lower than the digital implementation and 90x lower than the energy dissipated by <xref ref-type="bibr" rid="B29">Indiveri et al. (2006)</xref>. Compared to PCM-based neuron that requires additional digital circuitry like a latch and a NOR logic gate (<xref ref-type="bibr" rid="B67">Tuma et al., 2016</xref>), our FeFET-based neuron dissipates 40x lower power and occupies at least 2.5x lower area in terms of feature size F. Compared to insulator-to-metel phase-transition vanadium dioxide (VO<sub>2</sub>)-based neuron (<xref ref-type="bibr" rid="B35">Jerry et al., 2017</xref>), FeFET-based neuron dissipates 300x lower power.</p>
<table-wrap position="float" id="T1">
<label>TABLE 1</label>
<caption><p>Comparative study between various hardware implementations of spiking neuron.</p></caption>
<table cellspacing="5" cellpadding="5" frame="hsides" rules="groups">
<thead>
<tr>
<td valign="top" align="left"></td>
<td valign="top" align="center"><xref ref-type="bibr" rid="B29">Indiveri et al., 2006</xref></td>
<td valign="top" align="center" colspan="2"><xref ref-type="bibr" rid="B36">Joubert et al., 2012</xref></td>
<td valign="top" align="center"><xref ref-type="bibr" rid="B67">Tuma et al., 2016</xref></td>
<td valign="top" align="center"><xref ref-type="bibr" rid="B63">Sengupta et al., 2016</xref></td>
<td valign="top" align="center"><xref ref-type="bibr" rid="B35">Jerry et al., 2017</xref></td>
<td valign="top" align="center">This work</td>
</tr>
</thead>
<tbody>
<tr>
<td valign="top" align="left">Neuron type</td>
<td valign="top" align="center">LIF</td>
<td valign="top" align="center">Analog LIF</td>
<td valign="top" align="center">Digital LIF</td>
<td valign="top" align="center">LIF</td>
<td valign="top" align="center">LIF</td>
<td valign="top" align="center">Piecewise linear FHN</td>
<td valign="top" align="center">LIF</td>
</tr>
<tr>
<td valign="top" align="left">Material</td>
<td valign="top" align="center">CMOS</td>
<td valign="top" align="center">CMOS</td>
<td valign="top" align="center">CMOS</td>
<td valign="top" align="center">Phase change (PCM)</td>
<td valign="top" align="center">Magnetic tunnel junction (MTJ)</td>
<td valign="top" align="center">Vanadium dioxide (VO<sub>2</sub>)</td>
<td valign="top" align="center">Ferroelectric HZO</td>
</tr>
<tr>
<td valign="top" align="left">Technology</td>
<td valign="top" align="center">800 nm</td>
<td valign="top" align="center">65 nm</td>
<td valign="top" align="center">65 nm</td>
<td valign="top" align="center">14 nm</td>
<td valign="top" align="center">&#x2013;</td>
<td valign="top" align="center">&#x2013;</td>
<td valign="top" align="center">45 nm</td>
</tr>
<tr>
<td valign="top" align="left">Integration mechanism</td>
<td valign="top" align="center">Capacitor charging</td>
<td valign="top" align="center">Capacitor charging</td>
<td valign="top" align="center">&#x2013;</td>
<td valign="top" align="center">Joule heating</td>
<td valign="top" align="center">Magnetization dynamics</td>
<td valign="top" align="center">Capacitor charging</td>
<td valign="top" align="center">Polarization accumulative</td>
</tr>
<tr>
<td valign="top" align="left">Circuit elements</td>
<td valign="top" align="center">22 Transistor + one capacitor</td>
<td valign="top" align="center">33 Transistor + one capacitor</td>
<td valign="top" align="center">Pulse generator, counter, and comparator</td>
<td valign="top" align="center">One PCM + digital circuit</td>
<td valign="top" align="center">Two MTJs + four transistors</td>
<td valign="top" align="center">One VO<sub>2</sub> + one transistor + one capacitor</td>
<td valign="top" align="center">One FeFET + six transistors</td>
</tr>
<tr>
<td valign="top" align="left">Stochasticity</td>
<td valign="top" align="center">Yes</td>
<td valign="top" align="center">No</td>
<td valign="top" align="center">No</td>
<td valign="top" align="center">Yes</td>
<td valign="top" align="center">Yes</td>
<td valign="top" align="center">Yes</td>
<td valign="top" align="center">Yes</td>
</tr>
<tr>
<td valign="top" align="left">Power or energy/spike</td>
<td valign="top" align="center">900 pJ</td>
<td valign="top" align="center">2 pJ</td>
<td valign="top" align="center">41.3 pJ</td>
<td valign="top" align="center">120 &#x03BC;W</td>
<td valign="top" align="center">&#x2013;</td>
<td valign="top" align="center">11.9 &#x03BC;W</td>
<td valign="top" align="center">1&#x2013;10 pJ</td>
</tr>
<tr>
<td valign="top" align="left">Firing rate</td>
<td valign="top" align="center">200 Hz</td>
<td valign="top" align="center">2 MHz</td>
<td valign="top" align="center">2 MHz</td>
<td valign="top" align="center">35&#x2013;40 kHz</td>
<td valign="top" align="center">&#x2013;</td>
<td valign="top" align="center">30 kHz</td>
<td valign="top" align="center">50 kHz</td>
</tr>
<tr>
<td valign="top" align="left">Area</td>
<td valign="top" align="center">2573 &#x03BC;<italic>m</italic><sup>2</sup></td>
<td valign="top" align="center">120 &#x03BC;<italic>m</italic><sup>2</sup></td>
<td valign="top" align="center">538 &#x03BC;<italic>m</italic><sup>2</sup></td>
<td valign="top" align="center">0.5&#x2013;1 &#x03BC;<italic>m</italic><sup>2</sup></td>
<td valign="top" align="center">&#x2013;</td>
<td valign="top" align="center">&#x2013;</td>
<td valign="top" align="center">2.05 &#x03BC;<italic>m</italic><sup>2</sup></td>
</tr>
</tbody>
</table></table-wrap>
<p>The intrinsic ferroelectric polarization switching mechanism being a stochastic process (<xref ref-type="bibr" rid="B48">Mulaosmanovic et al., 2018b</xref>, <xref ref-type="bibr" rid="B49">Mulaosmanovic et al., 2018c</xref>; <xref ref-type="bibr" rid="B18">Dutta et al., 2019a</xref>; <xref ref-type="bibr" rid="B51">Ni et al., 2019a</xref>), the FeFET-based spiking neuron exhibits stochastic firing that maybe useful for building stochastic neural networks like neural sampling machine with novel properties like inherent weight normalization (<xref ref-type="bibr" rid="B14">Detorakis et al., 2019</xref>), for applications like modeling uncertainties in neural networks (<xref ref-type="bibr" rid="B21">Gal and Ghahramani, 2016</xref>) and for probabilistic inferencing (<xref ref-type="bibr" rid="B57">Pecevski et al., 2011</xref>). One key limitation of FeFET-based neuron compared to generalized neuron model utilized in neuroscience and CMOS-based circuits is that the membrane potential is represented by the intrinsic ferroelectric polarization state variable and the associated stochasticity arises directly from the ferroelectric domain nucleation process. Hence, the degree of tuning the neuronal parameters and the stochastic response is limited which might be disadvantageous for algorithms in which the parameters and the stochasticity have to be tightly controlled.</p>
<p>The FeFET-based analog synapse is realized using voltage-dependent partial polarization switching in multi-domain ferroelectric thin film (<xref ref-type="bibr" rid="B32">Jerry et al., 2018a</xref>, <xref ref-type="bibr" rid="B33">b</xref>). Recent experimental works have shown the ability to program FeFETs with voltage pulse widths as low as 50 ns (<xref ref-type="bibr" rid="B33">Jerry et al., 2018b</xref>) while the programming voltage can be brought down from 4 to 1.8 V by engineering the gate stack by adding an additional metal layer between the ferroelectric capacitor and MOS capacitor (<xref ref-type="bibr" rid="B52">Ni et al., 2019b</xref>). <xref ref-type="table" rid="T2">Table 2</xref> shows a comparative study between FeFET-based analog synapse and various other candidates like PCM (<xref ref-type="bibr" rid="B9">Burr et al., 2010</xref>; <xref ref-type="bibr" rid="B5">Athmanathan et al., 2016</xref>; <xref ref-type="bibr" rid="B3">Ambrogio et al., 2018</xref>) and RRAM (<xref ref-type="bibr" rid="B41">Lee et al., 2012</xref>; <xref ref-type="bibr" rid="B71">Wu et al., 2017</xref>; <xref ref-type="bibr" rid="B72">Wu W. et al., 2018</xref>; <xref ref-type="bibr" rid="B44">Luo et al., 2019</xref>). One major benefit of using FeFET for implementing analog synapse is the reduced variability to less that 0.5% (<xref ref-type="bibr" rid="B44">Luo et al., 2019</xref>) and an order of magnitude reduction in write energy (<xref ref-type="bibr" rid="B17">D&#x00FC;nkel et al., 2018</xref>; <xref ref-type="bibr" rid="B52">Ni et al., 2019b</xref>). The cell area is comparable to that of PCM and RRAM. One limitation of FeFET-based analog synapse is the achievable number of non-volatile conductance states as we scale down the device. While a recent experiment on 60 &#x03BC;m<sup>2</sup> size FeFET devices exhibited 32 conductance states (equivalent to 5-bit precision) (<xref ref-type="bibr" rid="B32">Jerry et al., 2018a</xref>, <xref ref-type="bibr" rid="B33">b</xref>), in this work, we achieved eight non-overlapping conductance states (equivalent to 3 bits) in a 0.25 &#x03BC;m<sup>2</sup> size device. The precision of synaptic weight overed can be further improved by resorting to hybrid mechanisms like the recently proposed two transistor-one FEFET (2T1F) hybrid weight cell that allow up to 64 states with improved non-linearity and asymmetry factors (<xref ref-type="bibr" rid="B44">Luo et al., 2019</xref>; <xref ref-type="bibr" rid="B65">Sun et al., 2019</xref>). Similar hybrid schemes have been applied to other novel devices like the three-transistor, one-capacitor, and two PCM (3T1C+2PCM) weight cell (<xref ref-type="bibr" rid="B3">Ambrogio et al., 2018</xref>).</p>
<table-wrap position="float" id="T2">
<label>TABLE 2</label>
<caption><p>Comparative study between various hardware implementations of analog synaptic weight cell.</p></caption>
<table cellspacing="5" cellpadding="5" frame="hsides" rules="groups">
<thead>
<tr>
<td valign="top" align="left"></td>
<td valign="top" align="center">PCM</td>
<td valign="top" align="center">RRAM</td>
<td valign="top" align="center">FeFET</td>
</tr>
</thead>
<tbody>
<tr>
<td valign="top" align="left">Material</td>
<td valign="top" align="center">GST</td>
<td valign="top" align="center">TaO<sub>x</sub>/HfO<sub>x</sub></td>
<td valign="top" align="center">Hf<sub>x</sub>Zr<sub>1&#x2013;x</sub>O<sub>2</sub></td>
</tr>
<tr>
<td valign="top" align="left">States</td>
<td valign="top" align="center">8</td>
<td valign="top" align="center">128</td>
<td valign="top" align="center">8</td>
</tr>
<tr>
<td valign="top" align="left">Variation</td>
<td valign="top" align="center">&#x223C;1.5%</td>
<td valign="top" align="center">&#x223C;3.7%</td>
<td valign="top" align="center">&#x003C;0.5%</td>
</tr>
<tr>
<td valign="top" align="left">Write voltage</td>
<td valign="top" align="center">2.5 V</td>
<td valign="top" align="center">1.6 V</td>
<td valign="top" align="center">4 V</td>
</tr>
<tr>
<td valign="top" align="left">Write energy</td>
<td valign="top" align="center">30 pJ</td>
<td valign="top" align="center">&#x223C;10 pJ</td>
<td valign="top" align="center">0.1 pJ</td>
</tr>
<tr>
<td valign="top" align="left">Cell area</td>
<td valign="top" align="center">25F<sup>2</sup></td>
<td valign="top" align="center">24F<sup>2</sup></td>
<td valign="top" align="center">24F<sup>2</sup></td>
</tr>
</tbody>
</table></table-wrap>
</sec>
<sec id="S5">
<title>Conclusion</title>
<p>In summary, we explore the rich polarization switching dynamics and non-volatile nature of FeFETs and propose an all FeFET-based SNN neuromorphic hardware that can enable low-power spike-based information processing and co-localized memory and computing (a.k.a. in-memory computing). We experimentally demonstrate the essential neuronal and synaptic dynamics in a 28 nm high-K metal gate FeFET technology. Furthermore, we implement a SG learning algorithm on our SNN platform, thus enabling us to perform supervised learning. As such, the work provides a pathway toward building energy-efficient neuromorphic hardware that can support traditional machine learning algorithms. We also undertake synergistic device-algorithm co-design by accounting for the impacts of device-level variation (stochasticity) and limited bit precision of on-chip synaptic weights (available analog states) on the classification accuracy and highlight possible avenues of future work to overcome the current challenges such as resorting to hybrid precision training and inference.</p>
</sec>
<sec id="S6">
<title>Data Availability Statement</title>
<p>The datasets generated for this study are available on request to the corresponding author.</p>
</sec>
<sec id="S7">
<title>Author Contributions</title>
<p>SDu, SJ, and SDa developed the main idea. SDu performed all the measurements. SDu, JG, and KN performed the circuit simulations. CS performed the machine learning simulations. All authors discussed the results, agreed to the conclusions of the manuscript, and contributed to the writing of the manuscript.</p>
</sec>
<sec id="conf1">
<title>Conflict of Interest</title>
<p>The authors declare that the research was conducted in the absence of any commercial or financial relationships that could be construed as a potential conflict of interest.</p>
</sec>
</body>
<back>
<fn-group>
<fn fn-type="financial-disclosure">
<p><bold>Funding.</bold> This work was supported in part by the NSF sponsored ASSIST Engineering Research Center, Semiconductor Research Corporation (SRC), and DARPA.</p>
</fn>
</fn-group>
<ack>
<p>We are grateful to M. Trentzsch, S. Dunkel, S. Beyer, and W. Taylor at Globalfoundries Dresden, Germany, for providing 28 nm HKMG FeFET test devices.</p>
</ack>
<ref-list>
<title>References</title>
<ref id="B1"><citation citation-type="journal"><person-group person-group-type="author"><name><surname>Abderrahmane</surname> <given-names>N.</given-names></name> <name><surname>Lemaire</surname> <given-names>E.</given-names></name> <name><surname>Miramond</surname> <given-names>B.</given-names></name></person-group> (<year>2020</year>). <article-title>Design space exploration of hardware spiking neurons for embedded artificial intelligence.</article-title> <source><italic>Neural Networks</italic></source> <volume>121</volume> <fpage>366</fpage>&#x2013;<lpage>386</lpage>. <pub-id pub-id-type="doi">10.1016/j.neunet.2019.09.024</pub-id> <pub-id pub-id-type="pmid">31593842</pub-id></citation></ref>
<ref id="B2"><citation citation-type="journal"><person-group person-group-type="author"><name><surname>Akopyan</surname> <given-names>F.</given-names></name> <name><surname>Sawada</surname> <given-names>J.</given-names></name> <name><surname>Cassidy</surname> <given-names>A.</given-names></name> <name><surname>Alvarez-Icaza</surname> <given-names>R.</given-names></name> <name><surname>Arthur</surname> <given-names>J.</given-names></name> <name><surname>Merolla</surname> <given-names>P.</given-names></name><etal/></person-group> (<year>2015</year>). <article-title>TrueNorth: design and tool flow of a 65 mW 1 million neuron programmable neurosynaptic chip.</article-title> <source><italic>IEEE Trans. Comput. Des. Integr. Circuits Syst.</italic></source> <volume>34</volume> <fpage>1537</fpage>&#x2013;<lpage>1557</lpage>. <pub-id pub-id-type="doi">10.1109/TCAD.2015.2474396</pub-id></citation></ref>
<ref id="B3"><citation citation-type="journal"><person-group person-group-type="author"><name><surname>Ambrogio</surname> <given-names>S.</given-names></name> <name><surname>Narayanan</surname> <given-names>P.</given-names></name> <name><surname>Tsai</surname> <given-names>H.</given-names></name> <name><surname>Shelby</surname> <given-names>R. M.</given-names></name> <name><surname>Boybat</surname> <given-names>I.</given-names></name> <name><surname>Di Nolfo</surname> <given-names>C.</given-names></name><etal/></person-group> (<year>2018</year>). <article-title>Equivalent-accuracy accelerated neural-network training using analogue memory.</article-title> <source><italic>Nature</italic></source> <volume>558</volume> <fpage>60</fpage>&#x2013;<lpage>67</lpage>. <pub-id pub-id-type="doi">10.1038/s41586-018-0180-5</pub-id> <pub-id pub-id-type="pmid">29875487</pub-id></citation></ref>
<ref id="B4"><citation citation-type="journal"><person-group person-group-type="author"><name><surname>Anwani</surname> <given-names>N.</given-names></name> <name><surname>Rajendran</surname> <given-names>B.</given-names></name></person-group> (<year>2015</year>). &#x201C;<article-title>NormAD - normalized approximate descent based supervised learning rule for spiking neurons</article-title>,&#x201D; in <source><italic>Proceedings of the International Joint Conference on Neural Networks</italic></source>, <publisher-loc>Killarney</publisher-loc>.</citation></ref>
<ref id="B5"><citation citation-type="journal"><person-group person-group-type="author"><name><surname>Athmanathan</surname> <given-names>A.</given-names></name> <name><surname>Stanisavljevic</surname> <given-names>M.</given-names></name> <name><surname>Papandreou</surname> <given-names>N.</given-names></name> <name><surname>Pozidis</surname> <given-names>H.</given-names></name> <name><surname>Eleftheriou</surname> <given-names>E.</given-names></name></person-group> (<year>2016</year>). <article-title>Multilevel-cell phase-change memory: a viable technology.</article-title> <source><italic>IEEE J. Emerg. Sel. Top. Circuits Syst.</italic></source> <volume>6</volume> <fpage>87</fpage>&#x2013;<lpage>100</lpage>. <pub-id pub-id-type="doi">10.1109/JETCAS.2016.2528598</pub-id></citation></ref>
<ref id="B6"><citation citation-type="journal"><person-group person-group-type="author"><name><surname>Benda</surname> <given-names>J.</given-names></name> <name><surname>Herz</surname> <given-names>A. V. M.</given-names></name></person-group> (<year>2003</year>). <article-title>A universal model for spike-frequency adaptation.</article-title> <source><italic>Neural Comput.</italic></source> <volume>15</volume> <fpage>2523</fpage>&#x2013;<lpage>2564</lpage>. <pub-id pub-id-type="doi">10.1162/089976603322385063</pub-id> <pub-id pub-id-type="pmid">14577853</pub-id></citation></ref>
<ref id="B7"><citation citation-type="journal"><person-group person-group-type="author"><name><surname>Benjamin</surname> <given-names>B. V.</given-names></name> <name><surname>Gao</surname> <given-names>P.</given-names></name> <name><surname>McQuinn</surname> <given-names>E.</given-names></name> <name><surname>Choudhary</surname> <given-names>S.</given-names></name> <name><surname>Chandrasekaran</surname> <given-names>A. R.</given-names></name> <name><surname>Bussat</surname> <given-names>J. M.</given-names></name><etal/></person-group> (<year>2014</year>). <article-title>Neurogrid: a mixed-analog-digital multichip system for large-scale neural simulations.</article-title> <source><italic>Proc. IEEE.</italic></source> <volume>102</volume> <fpage>699</fpage>&#x2013;<lpage>716</lpage>. <pub-id pub-id-type="doi">10.1109/JPROC.2014.2313565</pub-id></citation></ref>
<ref id="B8"><citation citation-type="journal"><person-group person-group-type="author"><name><surname>Bergstra</surname> <given-names>J.</given-names></name> <name><surname>Yamins</surname> <given-names>D.</given-names></name> <name><surname>Cox</surname> <given-names>D. D.</given-names></name></person-group> (<year>2013</year>). &#x201C;<article-title>Making a science of model search: Hyperparameter optimization in hundreds of dimensions for vision architectures</article-title>,&#x201D; in <source><italic>Proceedings of the 30th International Conference on Machine Learning, ICML 2013</italic></source>, <publisher-loc>Atlanta</publisher-loc>.</citation></ref>
<ref id="B9"><citation citation-type="journal"><person-group person-group-type="author"><name><surname>Burr</surname> <given-names>G. W.</given-names></name> <name><surname>Breitwisch</surname> <given-names>M. J.</given-names></name> <name><surname>Franceschini</surname> <given-names>M.</given-names></name> <name><surname>Garetto</surname> <given-names>D.</given-names></name> <name><surname>Gopalakrishnan</surname> <given-names>K.</given-names></name> <name><surname>Jackson</surname> <given-names>B.</given-names></name><etal/></person-group> (<year>2010</year>). <article-title>Phase change memory technology.</article-title> <source><italic>J. Vac. Sci. Technol. B, Nanotechnol. Microelectron. Mater. Process. Meas. Phenom.</italic></source> <volume>28</volume> <fpage>223</fpage>&#x2013;<lpage>262</lpage>. <pub-id pub-id-type="doi">10.1116/1.3301579</pub-id></citation></ref>
<ref id="B10"><citation citation-type="journal"><person-group person-group-type="author"><name><surname>Burr</surname> <given-names>G. W.</given-names></name> <name><surname>Shelby</surname> <given-names>R. M.</given-names></name> <name><surname>Sidler</surname> <given-names>S.</given-names></name> <name><surname>Di Nolfo</surname> <given-names>C.</given-names></name> <name><surname>Jang</surname> <given-names>J.</given-names></name> <name><surname>Boybat</surname> <given-names>I.</given-names></name><etal/></person-group> (<year>2015</year>). <article-title>Experimental demonstration and tolerancing of a large-scale neural network (165 000 synapses) using phase-change memory as the synaptic weight element.</article-title> <source><italic>IEEE Trans. Electron Devices</italic></source> <volume>62</volume> <fpage>3498</fpage>&#x2013;<lpage>3507</lpage>. <pub-id pub-id-type="doi">10.1109/TED.2015.2439635</pub-id></citation></ref>
<ref id="B11"><citation citation-type="journal"><person-group person-group-type="author"><name><surname>Chicca</surname> <given-names>E.</given-names></name> <name><surname>Stefanini</surname> <given-names>F.</given-names></name> <name><surname>Bartolozzi</surname> <given-names>C.</given-names></name> <name><surname>Indiveri</surname> <given-names>G.</given-names></name></person-group> (<year>2014</year>). <article-title>Neuromorphic electronic circuits for building autonomous cognitive systems.</article-title> <source><italic>Proc. IEEE</italic></source> <volume>102</volume> <fpage>1367</fpage>&#x2013;<lpage>1388</lpage>. <pub-id pub-id-type="doi">10.1109/JPROC.2014.2313954</pub-id></citation></ref>
<ref id="B12"><citation citation-type="journal"><person-group person-group-type="author"><name><surname>Choi</surname> <given-names>J.</given-names></name> <name><surname>Venkataramani</surname> <given-names>S.</given-names></name> <name><surname>Srinivasan</surname> <given-names>V.</given-names></name> <name><surname>Gopalakrishnan</surname> <given-names>K.</given-names></name> <name><surname>Wang</surname> <given-names>Z.</given-names></name> <name><surname>Chuang</surname> <given-names>P.</given-names></name></person-group> (<year>2019</year>). &#x201C;<article-title>Accurate and efficient 2-Bit quantized neural networks</article-title>,&#x201D; in <source><italic>SysML</italic></source>, (<publisher-loc>Stanford, CA</publisher-loc>).</citation></ref>
<ref id="B13"><citation citation-type="journal"><person-group person-group-type="author"><name><surname>Davies</surname> <given-names>M.</given-names></name> <name><surname>Srinivasa</surname> <given-names>N.</given-names></name> <name><surname>Lin</surname> <given-names>T. H.</given-names></name> <name><surname>Chinya</surname> <given-names>G.</given-names></name> <name><surname>Cao</surname> <given-names>Y.</given-names></name> <name><surname>Choday</surname> <given-names>S. H.</given-names></name><etal/></person-group> (<year>2018</year>). <article-title>Loihi: a neuromorphic manycore processor with on-chip learning.</article-title> <source><italic>IEEE Micro</italic></source> <volume>38</volume> <fpage>82</fpage>&#x2013;<lpage>99</lpage>. <pub-id pub-id-type="doi">10.1109/MM.2018.112130359</pub-id></citation></ref>
<ref id="B14"><citation citation-type="journal"><person-group person-group-type="author"><name><surname>Detorakis</surname> <given-names>G.</given-names></name> <name><surname>Dutta</surname> <given-names>S.</given-names></name> <name><surname>Khanna</surname> <given-names>A.</given-names></name> <name><surname>Jerry</surname> <given-names>M.</given-names></name> <name><surname>Datta</surname> <given-names>S.</given-names></name> <name><surname>Neftci</surname> <given-names>E.</given-names></name></person-group> (<year>2019</year>). <article-title>Inherent weight normalization in stochastic neural networks.</article-title> <source><italic>Adv. Neural Informat. Proc. Syst.</italic></source> <fpage>3286</fpage>&#x2013;<lpage>3297</lpage>.</citation></ref>
<ref id="B15"><citation citation-type="journal"><person-group person-group-type="author"><name><surname>Diehl</surname> <given-names>P. U.</given-names></name> <name><surname>Cook</surname> <given-names>M.</given-names></name></person-group> (<year>2015</year>). <article-title>Unsupervised learning of digit recognition using spike-timing-dependent plasticity.</article-title> <source><italic>Front. Comput. Neurosci.</italic></source> <volume>9</volume>:<issue>99</issue>. <pub-id pub-id-type="doi">10.3389/fncom.2015.00099</pub-id> <pub-id pub-id-type="pmid">26941637</pub-id></citation></ref>
<ref id="B16"><citation citation-type="journal"><person-group person-group-type="author"><name><surname>Diehl</surname> <given-names>P. U.</given-names></name> <name><surname>Neil</surname> <given-names>D.</given-names></name> <name><surname>Binas</surname> <given-names>J.</given-names></name> <name><surname>Cook</surname> <given-names>M.</given-names></name> <name><surname>Liu</surname> <given-names>S. C.</given-names></name> <name><surname>Pfeiffer</surname> <given-names>M.</given-names></name></person-group> (<year>2015</year>). &#x201C;<article-title>Fast-classifying, high-accuracy spiking deep networks through weight and threshold balancing</article-title>,&#x201D; in <source><italic>Proceedings of the International Joint Conference on Neural Networks</italic></source>, <publisher-loc>Killarney</publisher-loc>.</citation></ref>
<ref id="B17"><citation citation-type="journal"><person-group person-group-type="author"><name><surname>D&#x00FC;nkel</surname> <given-names>S.</given-names></name> <name><surname>Trentzsch</surname> <given-names>M.</given-names></name> <name><surname>Richter</surname> <given-names>R.</given-names></name> <name><surname>Moll</surname> <given-names>P.</given-names></name> <name><surname>Fuchs</surname> <given-names>C.</given-names></name> <name><surname>Gehring</surname> <given-names>O.</given-names></name><etal/></person-group> (<year>2018</year>). &#x201C;<article-title>A FeFET based super-low-power ultra-fast embedded NVM technology for 22nm FDSOI and beyond</article-title>,&#x201D; in <source><italic>Proceedings of the Technical Digest - International Electron Devices Meeting, IEDM</italic></source>, <publisher-loc>San Francisco, CA</publisher-loc>.</citation></ref>
<ref id="B18"><citation citation-type="journal"><person-group person-group-type="author"><name><surname>Dutta</surname> <given-names>S.</given-names></name> <name><surname>Chakraborty</surname> <given-names>W.</given-names></name> <name><surname>Gomez</surname> <given-names>J.</given-names></name> <name><surname>Ni</surname> <given-names>K.</given-names></name> <name><surname>Joshi</surname> <given-names>S.</given-names></name> <name><surname>Datta</surname> <given-names>S.</given-names></name></person-group> (<year>2019a</year>). &#x201C;<article-title>Energy-Efficient Edge Inference on Multi-Channel Streaming Data</article-title>,&#x201D; in <source><italic>Proceedings of the 28nm HKMG FeFET Technology. in 2019 Symposium on VLSI Technology</italic></source>, <publisher-loc>Kyoto</publisher-loc>: <publisher-name>IEEE</publisher-name>, <fpage>T38</fpage>&#x2013;<lpage>T39</lpage>.</citation></ref>
<ref id="B19"><citation citation-type="journal"><person-group person-group-type="author"><name><surname>Dutta</surname> <given-names>S.</given-names></name> <name><surname>Saha</surname> <given-names>A.</given-names></name> <name><surname>Panda</surname> <given-names>P.</given-names></name> <name><surname>Chakraborty</surname> <given-names>W.</given-names></name> <name><surname>Gomez</surname> <given-names>J.</given-names></name> <name><surname>Khanna</surname> <given-names>A.</given-names></name><etal/></person-group> (<year>2019b</year>). &#x201C;<article-title>Biologically plausible ferroelectric quasi-leaky integrate and fire neuron</article-title>,&#x201D; in <source><italic>Proceedings of the 2019 Symposium on VLSI Technology</italic></source>, <publisher-loc>Kyoto</publisher-loc>: <publisher-name>IEEE</publisher-name>, <fpage>T140</fpage>&#x2013;<lpage>T141</lpage>.</citation></ref>
<ref id="B20"><citation citation-type="journal"><person-group person-group-type="author"><name><surname>Faisal</surname> <given-names>A. A.</given-names></name> <name><surname>Selen</surname> <given-names>L. P. J.</given-names></name> <name><surname>Wolpert</surname> <given-names>D. M.</given-names></name></person-group> (<year>2008</year>). <article-title>Noise in the nervous system.</article-title> <source><italic>Nat. Rev. Neurosci.</italic></source> <volume>9</volume> <fpage>292</fpage>&#x2013;<lpage>303</lpage>. <pub-id pub-id-type="doi">10.1038/nrn2258</pub-id> <pub-id pub-id-type="pmid">18319728</pub-id></citation></ref>
<ref id="B21"><citation citation-type="journal"><person-group person-group-type="author"><name><surname>Gal</surname> <given-names>Y.</given-names></name> <name><surname>Ghahramani</surname> <given-names>Z.</given-names></name></person-group> (<year>2016</year>). &#x201C;<article-title>Dropout as a bayesian approximation: representing model uncertainty in deep learning</article-title>,&#x201D; in <source><italic>Proceedings of the 33rd International Conference on Machine Learning</italic></source>, <volume>Vol. 48</volume> <publisher-loc>New York, NY</publisher-loc>: <publisher-name>PMLR</publisher-name>, <fpage>1050</fpage>&#x2013;<lpage>1059</lpage>.</citation></ref>
<ref id="B22"><citation citation-type="journal"><person-group person-group-type="author"><name><surname>Gao</surname> <given-names>L.</given-names></name> <name><surname>Wang</surname> <given-names>I. T.</given-names></name> <name><surname>Chen</surname> <given-names>P. Y.</given-names></name> <name><surname>Vrudhula</surname> <given-names>S.</given-names></name> <name><surname>Seo</surname> <given-names>J. S.</given-names></name> <name><surname>Cao</surname> <given-names>Y.</given-names></name><etal/></person-group> (<year>2015</year>). <article-title>Fully parallel write/read in resistive synaptic array for accelerating on-chip learning.</article-title> <source><italic>Nanotechnology</italic></source> <volume>26</volume>:<issue>455204</issue>. <pub-id pub-id-type="doi">10.1088/0957-4484/26/45/455204</pub-id></citation></ref>
<ref id="B23"><citation citation-type="journal"><person-group person-group-type="author"><name><surname>Gentet</surname> <given-names>L. J.</given-names></name> <name><surname>Stuart</surname> <given-names>G. J.</given-names></name> <name><surname>Clements</surname> <given-names>J. D.</given-names></name></person-group> (<year>2000</year>). <article-title>Direct measurement of specific membrane capacitance in neurons.</article-title> <source><italic>Biophys. J.</italic></source> <volume>79</volume> <fpage>314</fpage>&#x2013;<lpage>320</lpage>. <pub-id pub-id-type="doi">10.1016/S0006-3495(00)76293-X</pub-id></citation></ref>
<ref id="B24"><citation citation-type="journal"><person-group person-group-type="author"><name><surname>Gokmen</surname> <given-names>T.</given-names></name> <name><surname>Vlasov</surname> <given-names>Y.</given-names></name></person-group> (<year>2016</year>). <article-title>Acceleration of deep neural network training with resistive cross-point devices: design considerations.</article-title> <source><italic>Front. Neurosci.</italic></source> <volume>10</volume>:<issue>333</issue>. <pub-id pub-id-type="doi">10.3389/fnins.2016.00333</pub-id> <pub-id pub-id-type="pmid">27493624</pub-id></citation></ref>
<ref id="B25"><citation citation-type="journal"><person-group person-group-type="author"><name><surname>G&#x00FC;tig</surname> <given-names>R.</given-names></name></person-group> (<year>2014</year>). <article-title>To spike, or when to spike?</article-title> <source><italic>Curr. Opin. Neurobiol.</italic></source> <volume>25</volume> <fpage>134</fpage>&#x2013;<lpage>139</lpage>. <pub-id pub-id-type="doi">10.1016/j.conb.2014.01.004</pub-id> <pub-id pub-id-type="pmid">24468508</pub-id></citation></ref>
<ref id="B26"><citation citation-type="journal"><person-group person-group-type="author"><name><surname>Hodgkin</surname> <given-names>A. L.</given-names></name> <name><surname>Huxley</surname> <given-names>A. F.</given-names></name></person-group> (<year>1952</year>). <article-title>A quantitative description of membrane current and its application to conduction and excitation in nerve.</article-title> <source><italic>J. Physiol.</italic></source> <volume>117</volume> <fpage>500</fpage>&#x2013;<lpage>544</lpage>. <pub-id pub-id-type="doi">10.1113/jphysiol.1952.sp004764</pub-id> <pub-id pub-id-type="pmid">12991237</pub-id></citation></ref>
<ref id="B27"><citation citation-type="journal"><person-group person-group-type="author"><name><surname>Huh</surname> <given-names>D.</given-names></name> <name><surname>Sejnowski</surname> <given-names>T. J.</given-names></name></person-group> (<year>2018</year>). &#x201C;<article-title>Gradient descent for spiking neural networks</article-title>,&#x201D; in <source><italic>Proceedings of the Advances in Neural Information Processing Systems</italic></source>, <publisher-loc>Montr&#x00E9;al, QC</publisher-loc>, <fpage>1440</fpage>&#x2013;<lpage>1450</lpage>.</citation></ref>
<ref id="B28"><citation citation-type="journal"><person-group person-group-type="author"><name><surname>Indiveri</surname> <given-names>G.</given-names></name></person-group> (<year>2003</year>). &#x201C;<article-title>A low-power adaptive integrate-and-fire neuron circuit</article-title>,&#x201D; in <source><italic>Proceedings of the - IEEE International Symposium on Circuits and Systems</italic></source>, <publisher-loc>Bangkok</publisher-loc>.</citation></ref>
<ref id="B29"><citation citation-type="journal"><person-group person-group-type="author"><name><surname>Indiveri</surname> <given-names>G.</given-names></name> <name><surname>Chicca</surname> <given-names>E.</given-names></name> <name><surname>Douglas</surname> <given-names>R.</given-names></name></person-group> (<year>2006</year>). <article-title>A VLSI array of low-power spiking neurons and bistable synapses with spike-timing dependent plasticity.</article-title> <source><italic>IEEE Trans. Neural Networks</italic></source> <volume>17</volume> <fpage>211</fpage>&#x2013;<lpage>221</lpage>. <pub-id pub-id-type="doi">10.1109/TNN.2005.860850</pub-id> <pub-id pub-id-type="pmid">16526488</pub-id></citation></ref>
<ref id="B30"><citation citation-type="journal"><person-group person-group-type="author"><name><surname>Indiveri</surname> <given-names>G.</given-names></name> <name><surname>Linares-Barranco</surname> <given-names>B.</given-names></name> <name><surname>Hamilton</surname> <given-names>T. J.</given-names></name> <name><surname>van Schaik</surname> <given-names>A.</given-names></name> <name><surname>Etienne-Cummings</surname> <given-names>R.</given-names></name> <name><surname>Delbruck</surname> <given-names>T.</given-names></name><etal/></person-group> (<year>2011</year>). <article-title>Neuromorphic silicon neuron circuits.</article-title> <source><italic>Front. Neurosci.</italic></source> <volume>5</volume>:<issue>73</issue>. <pub-id pub-id-type="doi">10.3389/fnins.2011.00073</pub-id> <pub-id pub-id-type="pmid">21747754</pub-id></citation></ref>
<ref id="B31"><citation citation-type="journal"><person-group person-group-type="author"><name><surname>Izhikevich</surname> <given-names>E. M.</given-names></name></person-group> (<year>2003</year>). <article-title>Simple model of spiking neurons.</article-title> <source><italic>IEEE Trans. Neural Networks</italic></source> <volume>14</volume> <fpage>1569</fpage>&#x2013;<lpage>1572</lpage>. <pub-id pub-id-type="doi">10.1109/TNN.2003.820440</pub-id> <pub-id pub-id-type="pmid">18244602</pub-id></citation></ref>
<ref id="B32"><citation citation-type="journal"><person-group person-group-type="author"><name><surname>Jerry</surname> <given-names>M.</given-names></name> <name><surname>Chen</surname> <given-names>P. Y.</given-names></name> <name><surname>Zhang</surname> <given-names>J.</given-names></name> <name><surname>Sharma</surname> <given-names>P.</given-names></name> <name><surname>Ni</surname> <given-names>K.</given-names></name> <name><surname>Yu</surname> <given-names>S.</given-names></name><etal/></person-group> (<year>2018a</year>). &#x201C;<article-title>Ferroelectric FET analog synapse for acceleration of deep neural network training</article-title>,&#x201D; in <source><italic>Proceedings of the Technical Digest - International Electron Devices Meeting, IEDM</italic></source>, <publisher-loc>San Francisco, CA</publisher-loc>.</citation></ref>
<ref id="B33"><citation citation-type="journal"><person-group person-group-type="author"><name><surname>Jerry</surname> <given-names>M.</given-names></name> <name><surname>Dutta</surname> <given-names>S.</given-names></name> <name><surname>Kazemi</surname> <given-names>A.</given-names></name> <name><surname>Ni</surname> <given-names>K.</given-names></name> <name><surname>Zhang</surname> <given-names>J.</given-names></name> <name><surname>Chen</surname> <given-names>P.-Y.</given-names></name><etal/></person-group> (<year>2018b</year>). <article-title>A Ferroelectric field effect transistor based synaptic weight cell.</article-title> <source><italic>J. Phys. D. Appl. Phys.</italic></source> <volume>51</volume>:<issue>434001</issue>. <pub-id pub-id-type="doi">10.1088/1361-6463/aad6f8</pub-id></citation></ref>
<ref id="B34"><citation citation-type="journal"><person-group person-group-type="author"><name><surname>Jerry</surname> <given-names>M.</given-names></name> <name><surname>Dutta</surname> <given-names>S.</given-names></name> <name><surname>Ni</surname> <given-names>K.</given-names></name> <name><surname>Zhang</surname> <given-names>J.</given-names></name> <name><surname>Sharma</surname> <given-names>P.</given-names></name> <name><surname>Datta</surname> <given-names>S.</given-names></name><etal/></person-group> (<year>2019</year>). <source><italic>Ferroelectric FET based Non-Volatile Analog Synaptic Weight Cell.</italic></source> <publisher-loc>Notre Dame</publisher-loc>: <publisher-name>University of Notre Dame</publisher-name>.</citation></ref>
<ref id="B35"><citation citation-type="journal"><person-group person-group-type="author"><name><surname>Jerry</surname> <given-names>M.</given-names></name> <name><surname>Parihar</surname> <given-names>A.</given-names></name> <name><surname>Grisafe</surname> <given-names>B.</given-names></name> <name><surname>Raychowdhury</surname> <given-names>A.</given-names></name> <name><surname>Datta</surname> <given-names>S.</given-names></name></person-group> (<year>2017</year>). &#x201C;<article-title>Ultra-low power probabilistic IMT neurons for stochastic sampling machines</article-title>,&#x201D; in <source><italic>Proceedings of the IEEE Symposium on VLSI Circuits, Digest of Technical Papers</italic></source>, <publisher-loc>Kyoto</publisher-loc>.</citation></ref>
<ref id="B36"><citation citation-type="journal"><person-group person-group-type="author"><name><surname>Joubert</surname> <given-names>A.</given-names></name> <name><surname>Belhadj</surname> <given-names>B.</given-names></name> <name><surname>Temam</surname> <given-names>O.</given-names></name> <name><surname>H&#x00E9;liot</surname> <given-names>R.</given-names></name></person-group> (<year>2012</year>). &#x201C;<article-title>Hardware spiking neurons design: Analog or digital?</article-title>,&#x201D; in <source><italic>Proceedings of the International Joint Conference on Neural Networks</italic></source>, <publisher-loc>Brisbane, QLD</publisher-loc>.</citation></ref>
<ref id="B37"><citation citation-type="journal"><person-group person-group-type="author"><name><surname>Khacef</surname> <given-names>L.</given-names></name> <name><surname>Abderrahmane</surname> <given-names>N.</given-names></name> <name><surname>Miramond</surname> <given-names>B.</given-names></name></person-group> (<year>2018</year>). &#x201C;<article-title>Confronting machine-learning with neuroscience for neuromorphic architectures design</article-title>,&#x201D; in <source><italic>Proceedings of the International Joint Conference on Neural Networks</italic></source>, <publisher-loc>Rio de Janeiro</publisher-loc>.</citation></ref>
<ref id="B38"><citation citation-type="journal"><person-group person-group-type="author"><name><surname>Kheradpisheh</surname> <given-names>S. R.</given-names></name> <name><surname>Ganjtabesh</surname> <given-names>M.</given-names></name> <name><surname>Thorpe</surname> <given-names>S. J.</given-names></name> <name><surname>Masquelier</surname> <given-names>T.</given-names></name></person-group> (<year>2018</year>). <article-title>STDP-based spiking deep convolutional neural networks for object recognition.</article-title> <source><italic>Neural Networks</italic></source> <volume>99</volume> <fpage>S56</fpage>&#x2013;<lpage>S67</lpage>. <pub-id pub-id-type="doi">10.1016/j.neunet.2017.12.005</pub-id> <pub-id pub-id-type="pmid">29328958</pub-id></citation></ref>
<ref id="B39"><citation citation-type="journal"><person-group person-group-type="author"><name><surname>Kuzum</surname> <given-names>D.</given-names></name> <name><surname>Jeyasingh</surname> <given-names>R. G. D.</given-names></name> <name><surname>Lee</surname> <given-names>B.</given-names></name> <name><surname>Wong</surname> <given-names>H. S. P.</given-names></name></person-group> (<year>2012</year>). <article-title>Nanoelectronic programmable synapses based on phase change materials for brain-inspired computing.</article-title> <source><italic>Nano Lett</italic>.</source> <volume>12</volume> <fpage>2179</fpage>&#x2013;<lpage>2186</lpage>. <pub-id pub-id-type="doi">10.1021/nl201040y</pub-id> <pub-id pub-id-type="pmid">21668029</pub-id></citation></ref>
<ref id="B40"><citation citation-type="journal"><person-group person-group-type="author"><name><surname>Lecun</surname> <given-names>Y.</given-names></name> <name><surname>Bengio</surname> <given-names>Y.</given-names></name> <name><surname>Hinton</surname> <given-names>G.</given-names></name></person-group> (<year>2015</year>). <article-title>Deep learning.</article-title> <source><italic>Nature</italic></source> <volume>521</volume> <fpage>436</fpage>&#x2013;<lpage>444</lpage>. <pub-id pub-id-type="doi">10.1038/nature14539</pub-id> <pub-id pub-id-type="pmid">26017442</pub-id></citation></ref>
<ref id="B41"><citation citation-type="journal"><person-group person-group-type="author"><name><surname>Lee</surname> <given-names>S. R.</given-names></name> <name><surname>Kim</surname> <given-names>Y. B.</given-names></name> <name><surname>Chang</surname> <given-names>M.</given-names></name> <name><surname>Kim</surname> <given-names>K. M.</given-names></name> <name><surname>Lee</surname> <given-names>C. B.</given-names></name> <name><surname>Hur</surname> <given-names>J. H.</given-names></name><etal/></person-group> (<year>2012</year>). &#x201C;<article-title>Multi-level switching of triple-layered TaOx RRAM with excellent reliability for storage class memory</article-title>,&#x201D; in <source><italic>Proceedings of the Digest of Technical Papers - Symposium on VLSI Technology</italic></source>, <publisher-loc>Honolulu, HI</publisher-loc>.</citation></ref>
<ref id="B42"><citation citation-type="journal"><person-group person-group-type="author"><name><surname>Lichtsteiner</surname> <given-names>P.</given-names></name> <name><surname>Posch</surname> <given-names>C.</given-names></name> <name><surname>Delbruck</surname> <given-names>T.</given-names></name></person-group> (<year>2008</year>). <article-title>A 128 &#x00D7; 128 120 dB 15 &#x03BC;s latency asynchronous temporal contrast vision sensor.</article-title> <source><italic>IEEE J. Solid State Circuits</italic></source> <volume>43</volume> <fpage>566</fpage>&#x2013;<lpage>576</lpage>. <pub-id pub-id-type="doi">10.1109/JSSC.2007.914337</pub-id></citation></ref>
<ref id="B43"><citation citation-type="journal"><person-group person-group-type="author"><name><surname>Liu</surname> <given-names>Y. H.</given-names></name> <name><surname>Wang</surname> <given-names>X. J.</given-names></name></person-group> (<year>2001</year>). <article-title>Spike-frequency adaptation of a generalized leaky integrate-and-fire model neuron.</article-title> <source><italic>J. Comput. Neurosci.</italic></source> <volume>10</volume> <fpage>25</fpage>&#x2013;<lpage>45</lpage>. <pub-id pub-id-type="doi">10.1023/A:1008916026143</pub-id></citation></ref>
<ref id="B44"><citation citation-type="journal"><person-group person-group-type="author"><name><surname>Luo</surname> <given-names>Y.</given-names></name> <name><surname>Wang</surname> <given-names>P.</given-names></name> <name><surname>Peng</surname> <given-names>X.</given-names></name> <name><surname>Sun</surname> <given-names>X.</given-names></name> <name><surname>Yu</surname> <given-names>S.</given-names></name></person-group> (<year>2019</year>). <article-title>Benchmark of ferroelectric transistor based hybrid precision synapse for neural network accelerator.</article-title> <source><italic>IEEE J. Explor. Solid-State Comput. Devices Circuits</italic></source> <volume>5</volume> <fpage>142</fpage>&#x2013;<lpage>150</lpage>. <pub-id pub-id-type="doi">10.1109/JXCDC.2019.2925061</pub-id></citation></ref>
<ref id="B45"><citation citation-type="journal"><person-group person-group-type="author"><name><surname>Merolla</surname> <given-names>P. A.</given-names></name> <name><surname>Arthur</surname> <given-names>J. V.</given-names></name> <name><surname>Alvarez-Icaza</surname> <given-names>R.</given-names></name> <name><surname>Cassidy</surname> <given-names>A. S.</given-names></name> <name><surname>Sawada</surname> <given-names>J.</given-names></name> <name><surname>Akopyan</surname> <given-names>F.</given-names></name><etal/></person-group> (<year>2014</year>). <article-title>A million spiking-neuron integrated circuit with a scalable communication network and interface.</article-title> <source><italic>Science</italic></source> <volume>354</volume> <fpage>668</fpage>&#x2013;<lpage>673</lpage>. <pub-id pub-id-type="doi">10.1126/science.1254642</pub-id> <pub-id pub-id-type="pmid">25104385</pub-id></citation></ref>
<ref id="B46"><citation citation-type="journal"><person-group person-group-type="author"><name><surname>Morie</surname> <given-names>T.</given-names></name> <name><surname>Amemiya</surname> <given-names>Y.</given-names></name></person-group> (<year>1994</year>). <article-title>An all-analog expandable neural network lsi with on-chip backpropagation learning.</article-title> <source><italic>IEEE J. Solid State Circuits</italic></source> <volume>29</volume> <fpage>1086</fpage>&#x2013;<lpage>1093</lpage>. <pub-id pub-id-type="doi">10.1109/4.309904</pub-id></citation></ref>
<ref id="B47"><citation citation-type="journal"><person-group person-group-type="author"><name><surname>Mulaosmanovic</surname> <given-names>H.</given-names></name> <name><surname>Chicca</surname> <given-names>E.</given-names></name> <name><surname>Bertele</surname> <given-names>M.</given-names></name> <name><surname>Mikolajick</surname> <given-names>T.</given-names></name> <name><surname>Slesazeck</surname> <given-names>S.</given-names></name></person-group> (<year>2018a</year>). <article-title>Mimicking biological neurons with a nanoscale ferroelectric transistor.</article-title> <source><italic>Nanoscale</italic></source> <volume>10</volume> <fpage>21755</fpage>&#x2013;<lpage>21763</lpage>. <pub-id pub-id-type="doi">10.1039/c8nr07135g</pub-id> <pub-id pub-id-type="pmid">30431045</pub-id></citation></ref>
<ref id="B48"><citation citation-type="journal"><person-group person-group-type="author"><name><surname>Mulaosmanovic</surname> <given-names>H.</given-names></name> <name><surname>Mikolajick</surname> <given-names>T.</given-names></name> <name><surname>Slesazeck</surname> <given-names>S.</given-names></name></person-group> (<year>2018b</year>). <article-title>Accumulative polarization reversal in nanoscale ferroelectric transistors.</article-title> <source><italic>ACS Appl. Mater. Interfaces</italic></source> <volume>10</volume> <fpage>23997</fpage>&#x2013;<lpage>24002</lpage>. <pub-id pub-id-type="doi">10.1021/acsami.8b08967</pub-id> <pub-id pub-id-type="pmid">29947210</pub-id></citation></ref>
<ref id="B49"><citation citation-type="journal"><person-group person-group-type="author"><name><surname>Mulaosmanovic</surname> <given-names>H.</given-names></name> <name><surname>Mikolajick</surname> <given-names>T.</given-names></name> <name><surname>Slesazeck</surname> <given-names>S.</given-names></name></person-group> (<year>2018c</year>). <article-title>Random number generation based on ferroelectric switching.</article-title> <source><italic>IEEE Electron Device Lett.</italic></source> <volume>39</volume> <fpage>135</fpage>&#x2013;<lpage>138</lpage>. <pub-id pub-id-type="doi">10.1109/LED.2017.2771818</pub-id></citation></ref>
<ref id="B50"><citation citation-type="journal"><person-group person-group-type="author"><name><surname>Neftci</surname> <given-names>E. O.</given-names></name> <name><surname>Mostafa</surname> <given-names>H.</given-names></name> <name><surname>Zenke</surname> <given-names>F.</given-names></name></person-group> (<year>2019</year>). <article-title>Surrogate gradient learning in spiking neural networks: bringing the power of gradient-based optimization to spiking neural networks.</article-title> <source><italic>IEEE Signal Process. Mag.</italic></source> <volume>36</volume> <fpage>51</fpage>&#x2013;<lpage>63</lpage>. <pub-id pub-id-type="doi">10.1109/MSP.2019.2931595</pub-id></citation></ref>
<ref id="B51"><citation citation-type="journal"><person-group person-group-type="author"><name><surname>Ni</surname> <given-names>K.</given-names></name> <name><surname>Chakraborty</surname> <given-names>W.</given-names></name> <name><surname>Smith</surname> <given-names>J.</given-names></name> <name><surname>Grisafe</surname> <given-names>B.</given-names></name> <name><surname>Datta</surname> <given-names>S.</given-names></name></person-group> (<year>2019a</year>). &#x201C;<article-title>Fundamental understanding and control of device-to-device variation</article-title>,&#x201D; in <source><italic>Proceedings of the Deeply Scaled Ferroelectric FETs. 2019 Symposium on VLSI Technology</italic></source>, <publisher-loc>Kyoto</publisher-loc>.</citation></ref>
<ref id="B52"><citation citation-type="journal"><person-group person-group-type="author"><name><surname>Ni</surname> <given-names>K.</given-names></name> <name><surname>Smith</surname> <given-names>J. A.</given-names></name> <name><surname>Grisafe</surname> <given-names>B.</given-names></name> <name><surname>Rakshit</surname> <given-names>T.</given-names></name> <name><surname>Obradovic</surname> <given-names>B.</given-names></name> <name><surname>Kittl</surname> <given-names>J. A.</given-names></name><etal/></person-group> (<year>2019b</year>). &#x201C;<article-title>SoC logic compatible multi-bit FeMFET weight cell for neuromorphic applications</article-title>,&#x201D; in <source><italic>Proceedings of the Technical Digest - International Electron Devices Meeting, IEDM</italic></source>, <publisher-loc>San Francisco, CA</publisher-loc>.</citation></ref>
<ref id="B53"><citation citation-type="journal"><person-group person-group-type="author"><name><surname>Ni</surname> <given-names>K.</given-names></name> <name><surname>Grisafe</surname> <given-names>B.</given-names></name> <name><surname>Chakraborty</surname> <given-names>W.</given-names></name> <name><surname>Saha</surname> <given-names>A. K.</given-names></name> <name><surname>Dutta</surname> <given-names>S.</given-names></name> <name><surname>Jerry</surname> <given-names>M.</given-names></name><etal/></person-group> (<year>2018</year>). &#x201C;<article-title>In-memory computing primitive for sensor data fusion in 28 nm HKMG FeFET technology</article-title>,&#x201D; in <source><italic>Proceedings of the 2018 IEEE International Electron Devices Meeting (IEDM)</italic></source>, <publisher-loc>San Francisco, CA</publisher-loc>: <publisher-name>IEEE</publisher-name>, <fpage>11</fpage>&#x2013;<lpage>16</lpage>.</citation></ref>
<ref id="B54"><citation citation-type="journal"><person-group person-group-type="author"><name><surname>O&#x2019;Connor</surname> <given-names>P.</given-names></name> <name><surname>Neil</surname> <given-names>D.</given-names></name> <name><surname>Liu</surname> <given-names>S. C.</given-names></name> <name><surname>Delbruck</surname> <given-names>T.</given-names></name> <name><surname>Pfeiffer</surname> <given-names>M.</given-names></name></person-group> (<year>2013</year>). <article-title>Real-time classification and sensor fusion with a spiking deep belief network.</article-title> <source><italic>Front. Neurosci.</italic></source> <volume>7</volume>:<issue>178</issue>. <pub-id pub-id-type="doi">10.3389/fnins.2013.00178</pub-id> <pub-id pub-id-type="pmid">24115919</pub-id></citation></ref>
<ref id="B55"><citation citation-type="journal"><person-group person-group-type="author"><name><surname>Panda</surname> <given-names>P.</given-names></name> <name><surname>Roy</surname> <given-names>K.</given-names></name></person-group> (<year>2016</year>). &#x201C;<article-title>Unsupervised regenerative learning of hierarchical features in spiking deep networks for object recognition</article-title>,&#x201D; in <source><italic>Proceedings of the International Joint Conference on Neural Networks</italic></source>, <publisher-loc>Vancouver, BC</publisher-loc>.</citation></ref>
<ref id="B56"><citation citation-type="journal"><person-group person-group-type="author"><name><surname>Park</surname> <given-names>J.</given-names></name> <name><surname>Ha</surname> <given-names>S.</given-names></name> <name><surname>Yu</surname> <given-names>T.</given-names></name> <name><surname>Neftci</surname> <given-names>E.</given-names></name> <name><surname>Cauwenberghs</surname> <given-names>G.</given-names></name></person-group> (<year>2014</year>). &#x201C;<article-title>A 65k-neuron 73-Mevents/s 22-pJ/event asynchronous micro-pipelined integrate-and-fire array transceiver</article-title>,&#x201D; in <source><italic>Proceedings of the IEEE 2014 Biomedical Circuits and Systems Conference, BioCAS 2014 - Proceedings</italic></source>, <publisher-loc>Lausanne</publisher-loc>.</citation></ref>
<ref id="B57"><citation citation-type="journal"><person-group person-group-type="author"><name><surname>Pecevski</surname> <given-names>D.</given-names></name> <name><surname>Buesing</surname> <given-names>L.</given-names></name> <name><surname>Maass</surname> <given-names>W.</given-names></name></person-group> (<year>2011</year>). <article-title>Probabilistic inference in general graphical models through sampling in stochastic networks of spiking neurons.</article-title> <source><italic>PLoS Comput. Biol.</italic></source> <volume>7</volume>:<issue>e1002294</issue>. <pub-id pub-id-type="doi">10.1371/journal.pcbi.1002294</pub-id> <pub-id pub-id-type="pmid">22219717</pub-id></citation></ref>
<ref id="B58"><citation citation-type="journal"><person-group person-group-type="author"><name><surname>Pei</surname> <given-names>J.</given-names></name> <name><surname>Deng</surname> <given-names>L.</given-names></name> <name><surname>Song</surname> <given-names>S.</given-names></name> <name><surname>Zhao</surname> <given-names>M.</given-names></name> <name><surname>Zhang</surname> <given-names>Y.</given-names></name> <name><surname>Wu</surname> <given-names>S.</given-names></name><etal/></person-group> (<year>2019</year>). <article-title>Towards artificial general intelligence with hybrid Tianjic chip architecture.</article-title> <source><italic>Nature</italic></source> <volume>572</volume> <fpage>106</fpage>&#x2013;<lpage>111</lpage>. <pub-id pub-id-type="doi">10.1038/s41586-019-1424-8</pub-id> <pub-id pub-id-type="pmid">31367028</pub-id></citation></ref>
<ref id="B59"><citation citation-type="journal"><person-group person-group-type="author"><name><surname>P&#x00E9;rez-Carrasco</surname> <given-names>J. A.</given-names></name> <name><surname>Zhao</surname> <given-names>B.</given-names></name> <name><surname>Serrano</surname> <given-names>C.</given-names></name> <name><surname>Acha</surname> <given-names>B.</given-names></name> <name><surname>Serrano-Gotarredona</surname> <given-names>T.</given-names></name> <name><surname>Chen</surname> <given-names>S.</given-names></name><etal/></person-group> (<year>2013</year>). <article-title>Mapping from frame-driven to frame-free event-driven vision systems by low-rate rate coding and coincidence processing - Application to feedforward convnets.</article-title> <source><italic>IEEE Trans. Pattern Anal. Mach. Intell.</italic></source> <volume>35</volume> <fpage>2706</fpage>&#x2013;<lpage>2719</lpage>. <pub-id pub-id-type="doi">10.1109/TPAMI.2013.71</pub-id> <pub-id pub-id-type="pmid">24051730</pub-id></citation></ref>
<ref id="B60"><citation citation-type="journal"><person-group person-group-type="author"><name><surname>Prezioso</surname> <given-names>M.</given-names></name> <name><surname>Merrikh-Bayat</surname> <given-names>F.</given-names></name> <name><surname>Hoskins</surname> <given-names>B. D.</given-names></name> <name><surname>Adam</surname> <given-names>G. C.</given-names></name> <name><surname>Likharev</surname> <given-names>K. K.</given-names></name> <name><surname>Strukov</surname> <given-names>D. B.</given-names></name></person-group> (<year>2015</year>). <article-title>Training and operation of an integrated neuromorphic network based on metal-oxide memristors.</article-title> <source><italic>Nature</italic></source> <volume>521</volume> <fpage>61</fpage>&#x2013;<lpage>64</lpage>. <pub-id pub-id-type="doi">10.1038/nature14441</pub-id> <pub-id pub-id-type="pmid">25951284</pub-id></citation></ref>
<ref id="B61"><citation citation-type="journal"><person-group person-group-type="author"><name><surname>Qiao</surname> <given-names>N.</given-names></name> <name><surname>Mostafa</surname> <given-names>H.</given-names></name> <name><surname>Corradi</surname> <given-names>F.</given-names></name> <name><surname>Osswald</surname> <given-names>M.</given-names></name> <name><surname>Stefanini</surname> <given-names>F.</given-names></name> <name><surname>Sumislawska</surname> <given-names>D.</given-names></name><etal/></person-group> (<year>2015</year>). <article-title>A reconfigurable on-line learning spiking neuromorphic processor comprising 256 neurons and 128K synapses.</article-title> <source><italic>Front. Neurosci.</italic></source> <volume>9</volume>:<issue>141</issue>. <pub-id pub-id-type="doi">10.3389/fnins.2015.00141</pub-id> <pub-id pub-id-type="pmid">25972778</pub-id></citation></ref>
<ref id="B62"><citation citation-type="journal"><person-group person-group-type="author"><name><surname>Saha</surname> <given-names>A. K.</given-names></name> <name><surname>Ni</surname> <given-names>K.</given-names></name> <name><surname>Dutta</surname> <given-names>S.</given-names></name> <name><surname>Datta</surname> <given-names>S.</given-names></name> <name><surname>Gupta</surname> <given-names>S.</given-names></name></person-group> (<year>2019</year>). <article-title>Phase field modeling of domain dynamics and polarization accumulation in ferroelectric HZO.</article-title> <source><italic>Appl. Phys. Lett.</italic></source> <volume>114</volume>:<issue>202903</issue>. <pub-id pub-id-type="doi">10.1063/1.5092707</pub-id></citation></ref>
<ref id="B63"><citation citation-type="journal"><person-group person-group-type="author"><name><surname>Sengupta</surname> <given-names>A.</given-names></name> <name><surname>Panda</surname> <given-names>P.</given-names></name> <name><surname>Wijesinghe</surname> <given-names>P.</given-names></name> <name><surname>Kim</surname> <given-names>Y.</given-names></name> <name><surname>Roy</surname> <given-names>K.</given-names></name></person-group> (<year>2016</year>). <article-title>Magnetic tunnel junction mimics stochastic cortical spiking neurons.</article-title> <source><italic>Sci. Rep.</italic></source> <volume>6</volume>:<issue>30039</issue>. <pub-id pub-id-type="doi">10.1038/srep30039</pub-id> <pub-id pub-id-type="pmid">27443913</pub-id></citation></ref>
<ref id="B64"><citation citation-type="journal"><person-group person-group-type="author"><name><surname>Sengupta</surname> <given-names>A.</given-names></name> <name><surname>Ye</surname> <given-names>Y.</given-names></name> <name><surname>Wang</surname> <given-names>R.</given-names></name> <name><surname>Liu</surname> <given-names>C.</given-names></name> <name><surname>Roy</surname> <given-names>K.</given-names></name></person-group> (<year>2019</year>). <article-title>Going deeper in spiking neural networks: VGG and residual architectures.</article-title> <source><italic>Front. Neurosci.</italic></source> <volume>13</volume>:<issue>95</issue>. <pub-id pub-id-type="doi">10.3389/fnins.2019.00095</pub-id> <pub-id pub-id-type="pmid">30899212</pub-id></citation></ref>
<ref id="B65"><citation citation-type="journal"><person-group person-group-type="author"><name><surname>Sun</surname> <given-names>X.</given-names></name> <name><surname>Wang</surname> <given-names>P.</given-names></name> <name><surname>Ni</surname> <given-names>K.</given-names></name> <name><surname>Datta</surname> <given-names>S.</given-names></name> <name><surname>Yu</surname> <given-names>S.</given-names></name></person-group> (<year>2019</year>). &#x201C;<article-title>Exploiting hybrid precision for training and inference: A 2T-1FeFET based analog synaptic weight cell</article-title>,&#x201D; in <source><italic>Proceedings of the Technical Digest - International Electron Devices Meeting, IEDM</italic></source>, <publisher-loc>San Francisco, CA</publisher-loc>.</citation></ref>
<ref id="B66"><citation citation-type="journal"><person-group person-group-type="author"><name><surname>Trentzsch</surname> <given-names>M.</given-names></name> <name><surname>Flachowsky</surname> <given-names>S.</given-names></name> <name><surname>Richter</surname> <given-names>R.</given-names></name> <name><surname>Paul</surname> <given-names>J.</given-names></name> <name><surname>Reimer</surname> <given-names>B.</given-names></name> <name><surname>Utess</surname> <given-names>D.</given-names></name><etal/></person-group> (<year>2017</year>). &#x201C;<article-title>A 28nm HKMG super low power embedded NVM technology based on ferroelectric FETs</article-title>,&#x201D; in <source><italic>Proceedings of the Technical Digest - International Electron Devices Meeting, IEDM</italic></source>, <publisher-loc>San Francisco, CA</publisher-loc>.</citation></ref>
<ref id="B67"><citation citation-type="journal"><person-group person-group-type="author"><name><surname>Tuma</surname> <given-names>T.</given-names></name> <name><surname>Pantazi</surname> <given-names>A.</given-names></name> <name><surname>Le Gallo</surname> <given-names>M.</given-names></name> <name><surname>Sebastian</surname> <given-names>A.</given-names></name> <name><surname>Eleftheriou</surname> <given-names>E.</given-names></name></person-group> (<year>2016</year>). <article-title>Stochastic phase-change neurons.</article-title> <source><italic>Nat. Nanotechnol.</italic></source> <volume>11</volume> <fpage>693</fpage>&#x2013;<lpage>699</lpage>. <pub-id pub-id-type="doi">10.1038/nnano.2016.70</pub-id> <pub-id pub-id-type="pmid">27183057</pub-id></citation></ref>
<ref id="B68"><citation citation-type="journal"><person-group person-group-type="author"><name><surname>Wang</surname> <given-names>R. M.</given-names></name> <name><surname>Thakur</surname> <given-names>C. S.</given-names></name> <name><surname>van Schaik</surname> <given-names>A.</given-names></name></person-group> (<year>2018</year>). <article-title>An FPGA-based massively parallel neuromorphic cortex simulator.</article-title> <source><italic>Front. Neurosci.</italic></source> <volume>12</volume>:<issue>213</issue>. <pub-id pub-id-type="doi">10.3389/fnins.2018.00213</pub-id> <pub-id pub-id-type="pmid">29692702</pub-id></citation></ref>
<ref id="B69"><citation citation-type="journal"><person-group person-group-type="author"><name><surname>Wu</surname> <given-names>J.</given-names></name> <name><surname>Chua</surname> <given-names>Y.</given-names></name> <name><surname>Zhang</surname> <given-names>M.</given-names></name> <name><surname>Yang</surname> <given-names>Q.</given-names></name> <name><surname>Li</surname> <given-names>G.</given-names></name> <name><surname>Li</surname> <given-names>H.</given-names></name></person-group> (<year>2019</year>). &#x201C;<article-title>Deep Spiking Neural Network with Spike Count based Learning Rule</article-title>,&#x201D; in <source><italic>Proceedings of the International Joint Conference on Neural Networks</italic></source>, <publisher-loc>Budapest</publisher-loc>.</citation></ref>
<ref id="B70"><citation citation-type="journal"><person-group person-group-type="author"><name><surname>Wu</surname> <given-names>S.</given-names></name> <name><surname>Li</surname> <given-names>G.</given-names></name> <name><surname>Chen</surname> <given-names>F.</given-names></name> <name><surname>Shi</surname> <given-names>L.</given-names></name></person-group> (<year>2018</year>). &#x201C;<article-title>Training and inference with integers in deep neural networks</article-title>,&#x201D; in <source><italic>Proceedings of the 6th International Conference on Learning Representations, ICLR 2018 - Conference Track Proceedings</italic></source>, <publisher-loc>Vancouver, BC</publisher-loc>.</citation></ref>
<ref id="B71"><citation citation-type="journal"><person-group person-group-type="author"><name><surname>Wu</surname> <given-names>W.</given-names></name> <name><surname>Wu</surname> <given-names>H.</given-names></name> <name><surname>Gao</surname> <given-names>B.</given-names></name> <name><surname>Deng</surname> <given-names>N.</given-names></name> <name><surname>Yu</surname> <given-names>S.</given-names></name> <name><surname>Qian</surname> <given-names>H.</given-names></name></person-group> (<year>2017</year>). <article-title>Improving analog switching in HfOx-based resistive memory with a thermal enhanced layer.</article-title> <source><italic>IEEE Electron Device Lett.</italic></source> <volume>38</volume> <fpage>1019</fpage>&#x2013;<lpage>1022</lpage>. <pub-id pub-id-type="doi">10.1109/LED.2017.2719161</pub-id></citation></ref>
<ref id="B72"><citation citation-type="journal"><person-group person-group-type="author"><name><surname>Wu</surname> <given-names>W.</given-names></name> <name><surname>Wu</surname> <given-names>H.</given-names></name> <name><surname>Gao</surname> <given-names>B.</given-names></name> <name><surname>Yao</surname> <given-names>P.</given-names></name> <name><surname>Zhang</surname> <given-names>X.</given-names></name> <name><surname>Peng</surname> <given-names>X.</given-names></name><etal/></person-group> (<year>2018</year>). &#x201C;<article-title>A methodology to improve linearity of analog RRAM for neuromorphic computing</article-title>,&#x201D; in <source><italic>Proceedings of the Digest of Technical Papers - Symposium on VLSI Technology</italic></source>, <publisher-loc>Honolulu, HI</publisher-loc>.</citation></ref>
<ref id="B73"><citation citation-type="journal"><person-group person-group-type="author"><name><surname>Wu</surname> <given-names>Y.</given-names></name> <name><surname>Deng</surname> <given-names>L.</given-names></name> <name><surname>Li</surname> <given-names>G.</given-names></name> <name><surname>Zhu</surname> <given-names>J.</given-names></name> <name><surname>Shi</surname> <given-names>L.</given-names></name></person-group> (<year>2018</year>). <article-title>Spatio-temporal backpropagation for training high-performance spiking neural networks.</article-title> <source><italic>Front. Neurosci.</italic></source> <volume>12</volume>:<issue>331</issue>. <pub-id pub-id-type="doi">10.3389/fnins.2018.00331</pub-id> <pub-id pub-id-type="pmid">29875621</pub-id></citation></ref>
<ref id="B74"><citation citation-type="journal"><person-group person-group-type="author"><name><surname>Yousefzadeh</surname> <given-names>A.</given-names></name> <name><surname>Stromatias</surname> <given-names>E.</given-names></name> <name><surname>Soto</surname> <given-names>M.</given-names></name> <name><surname>Serrano-Gotarredona</surname> <given-names>T.</given-names></name> <name><surname>Linares-Barranco</surname> <given-names>B.</given-names></name></person-group> (<year>2018</year>). <article-title>On practical issues for stochastic STDP hardware with 1-bit synaptic weights.</article-title> <source><italic>Front. Neurosci.</italic></source> <volume>12</volume>:<issue>665</issue>. <pub-id pub-id-type="doi">10.3389/fnins.2018.00665</pub-id> <pub-id pub-id-type="pmid">30374283</pub-id></citation></ref>
<ref id="B75"><citation citation-type="journal"><person-group person-group-type="author"><name><surname>Yu</surname> <given-names>S.</given-names></name> <name><surname>Chen</surname> <given-names>P. Y.</given-names></name> <name><surname>Cao</surname> <given-names>Y.</given-names></name> <name><surname>Xia</surname> <given-names>L.</given-names></name> <name><surname>Wang</surname> <given-names>Y.</given-names></name> <name><surname>Wu</surname> <given-names>H.</given-names></name></person-group> (<year>2015</year>). &#x201C;<article-title>Scaling-up resistive synaptic arrays for neuro-inspired architecture: challenges and prospect</article-title>,&#x201D; in <source><italic>Proceedings of the Technical Digest - International Electron Devices Meeting, IEDM</italic></source>, <publisher-loc>Washington, DC</publisher-loc>.</citation></ref>
<ref id="B76"><citation citation-type="journal"><person-group person-group-type="author"><name><surname>Zenke</surname> <given-names>F.</given-names></name> <name><surname>Ganguli</surname> <given-names>S.</given-names></name></person-group> (<year>2018</year>). <article-title>SuperSpike: supervised learning in multilayer spiking neural networks.</article-title> <source><italic>Neural Comput.</italic></source> <volume>30</volume> <fpage>1514</fpage>&#x2013;<lpage>1541</lpage>. <pub-id pub-id-type="doi">10.1162/neco_a_01086</pub-id></citation></ref>
</ref-list>
</back>
</article>