<?xml version="1.0" encoding="UTF-8" standalone="no"?>
<!DOCTYPE article PUBLIC "-//NLM//DTD Journal Publishing DTD v2.3 20070202//EN" "journalpublishing.dtd">
<article xml:lang="EN" xmlns:mml="http://www.w3.org/1998/Math/MathML" xmlns:xlink="http://www.w3.org/1999/xlink" article-type="brief-report">
<front>
<journal-meta>
<journal-id journal-id-type="publisher-id">Front. Syst. Neurosci.</journal-id>
<journal-title>Frontiers in Systems Neuroscience</journal-title>
<abbrev-journal-title abbrev-type="pubmed">Front. Syst. Neurosci.</abbrev-journal-title>
<issn pub-type="epub">1662-5137</issn>
<publisher>
<publisher-name>Frontiers Media S.A.</publisher-name>
</publisher>
</journal-meta>
<article-meta>
<article-id pub-id-type="doi">10.3389/fnsys.2022.805990</article-id>
<article-categories>
<subj-group subj-group-type="heading">
<subject>Neuroscience</subject>
<subj-group>
<subject>Brief Research Report</subject>
</subj-group>
</subj-group>
</article-categories>
<title-group>
<article-title>Recurrent Connections Might Be Important for Hierarchical Categorization</article-title>
</title-group>
<contrib-group>
<contrib contrib-type="author" corresp="yes">
<name><surname>Matsumoto</surname> <given-names>Narihisa</given-names></name>
<xref ref-type="aff" rid="aff1"><sup>1</sup></xref>
<xref ref-type="corresp" rid="c001"><sup>&#x002A;</sup></xref>
<uri xlink:href="http://loop.frontiersin.org/people/1356179/overview"/>
</contrib>
<contrib contrib-type="author">
<name><surname>Taguchi</surname> <given-names>Yusuke</given-names></name>
<xref ref-type="aff" rid="aff1"><sup>1</sup></xref>
<xref ref-type="aff" rid="aff2"><sup>2</sup></xref>
<xref ref-type="author-notes" rid="fn002"><sup>&#x2020;</sup></xref>
</contrib>
<contrib contrib-type="author">
<name><surname>Shimizu</surname> <given-names>Masaumi</given-names></name>
<xref ref-type="aff" rid="aff3"><sup>3</sup></xref>
</contrib>
<contrib contrib-type="author">
<name><surname>Katakami</surname> <given-names>Shun</given-names></name>
<xref ref-type="aff" rid="aff3"><sup>3</sup></xref>
</contrib>
<contrib contrib-type="author">
<name><surname>Okada</surname> <given-names>Masato</given-names></name>
<xref ref-type="aff" rid="aff3"><sup>3</sup></xref>
<uri xlink:href="http://loop.frontiersin.org/people/508400/overview"/>
</contrib>
<contrib contrib-type="author">
<name><surname>Sugase-Miyamoto</surname> <given-names>Yasuko</given-names></name>
<xref ref-type="aff" rid="aff1"><sup>1</sup></xref>
<uri xlink:href="http://loop.frontiersin.org/people/29099/overview"/>
</contrib>
</contrib-group>
<aff id="aff1"><sup>1</sup><institution>Human Informatics and Interaction Research Institute, National Institute of Advanced Industrial Science and Technology (AIST)</institution>, <addr-line>Tsukuba</addr-line>, <country>Japan</country></aff>
<aff id="aff2"><sup>2</sup><institution>Graduate School of Science and Technology, University of Tsukuba</institution>, <addr-line>Tsukuba</addr-line>, <country>Japan</country></aff>
<aff id="aff3"><sup>3</sup><institution>Graduate School of Frontier Sciences, The University of Tokyo</institution>, <addr-line>Kashiwa</addr-line>, <country>Japan</country></aff>
<author-notes>
<fn fn-type="edited-by"><p>Edited by: Natasha Sigala, Brighton and Sussex Medical School, United Kingdom</p></fn>
<fn fn-type="edited-by"><p>Reviewed by: Benjamin D. Evans, University of Bristol, United Kingdom; Horacio Rostro Gonzalez, University of Guanajuato, Mexico</p></fn>
<corresp id="c001">&#x002A;Correspondence: Narihisa Matsumoto, <email>xmatumo@ni.aist.go.jp</email></corresp>
<fn fn-type="present-address" id="fn002"><p><sup>&#x2020;</sup>Present address: Yusuke Taguchi, IBM Consulting BTS DTT, IBM Japan, Chuo-ku, Japan</p></fn>
</author-notes>
<pub-date pub-type="epub">
<day>24</day>
<month>02</month>
<year>2022</year>
</pub-date>
<pub-date pub-type="collection">
<year>2022</year>
</pub-date>
<volume>16</volume>
<elocation-id>805990</elocation-id>
<history>
<date date-type="received">
<day>31</day>
<month>10</month>
<year>2021</year>
</date>
<date date-type="accepted">
<day>12</day>
<month>01</month>
<year>2022</year>
</date>
</history>
<permissions>
<copyright-statement>Copyright &#x00A9; 2022 Matsumoto, Taguchi, Shimizu, Katakami, Okada and Sugase-Miyamoto.</copyright-statement>
<copyright-year>2022</copyright-year>
<copyright-holder>Matsumoto, Taguchi, Shimizu, Katakami, Okada and Sugase-Miyamoto</copyright-holder>
<license xlink:href="http://creativecommons.org/licenses/by/4.0/"><p>This is an open-access article distributed under the terms of the Creative Commons Attribution License (CC BY). The use, distribution or reproduction in other forums is permitted, provided the original author(s) and the copyright owner(s) are credited and that the original publication in this journal is cited, in accordance with accepted academic practice. No use, distribution or reproduction is permitted which does not comply with these terms.</p></license>
</permissions>
<abstract>
<p>Visual short-term memory is an important ability of primates and is thought to be stored in area TE. We previously reported that the initial transient responses of neurons in area TE represented information about a global category of faces, e.g., monkey faces vs. human faces vs. simple shapes, and the latter part of the responses represented information about fine categories, e.g., facial expression. The neuronal mechanisms of hierarchical categorization in area TE remain unknown. For this study, we constructed a combined model that consisted of a deep neural network (DNN) and a recurrent neural network and investigated whether this model can replicate the time course of hierarchical categorization. The visual images were stored in the recurrent connections of the model. When the visual images with noise were input to the model, the model outputted the time course of the hierarchical categorization. This result indicates that recurrent connections in the model are important not only for visual short-term memory but for hierarchical categorization, suggesting that recurrent connections in area TE are important for hierarchical categorization.</p>
</abstract>
<kwd-group>
<kwd>visual category</kwd>
<kwd>visual cortex</kwd>
<kwd>short-term memory</kwd>
<kwd>deep learning</kwd>
<kwd>modeling</kwd>
</kwd-group>
<contract-num rid="cn001">KAKENHI (26120535,16H01561,16H01684, 18H0520, 19K07804, 19K12149)</contract-num>
<contract-sponsor id="cn001">Japan Society for the Promotion of Science<named-content content-type="fundref-id">10.13039/501100001691</named-content></contract-sponsor><contract-sponsor id="cn002">New Energy and Industrial Technology Development Organization<named-content content-type="fundref-id">10.13039/501100003051</named-content></contract-sponsor>
<counts>
<fig-count count="4"/>
<table-count count="1"/>
<equation-count count="5"/>
<ref-count count="25"/>
<page-count count="6"/>
<word-count count="3640"/>
</counts>
</article-meta>
</front>
<body>
<sec id="S1" sec-type="intro">
<title>Introduction</title>
<p>Visual short-term memory is an important ability of primates. When primates see objects, the information about the objects is processed from the retina to the visual cortex in the brain. In the visual cortex, the object information is processed from V1 to area TE of the inferior temporal cortex (<xref ref-type="bibr" rid="B16">Mishkin et al., 1983</xref>). Visual short-term memory is thought to be stored in area TE (<xref ref-type="bibr" rid="B24">Sugase-Miyamoto et al., 2008</xref>) and the prefrontal cortex (<xref ref-type="bibr" rid="B5">Freedman et al., 2001</xref>). In area TE, some neurons respond to complex objects, faces, and so on and represent information about a global category, e.g., human vs. monkey vs. simple shapes, earlier than fine category information about faces, e.g., facial expression or identity (<xref ref-type="bibr" rid="B23">Sugase et al., 1999</xref>; <xref ref-type="bibr" rid="B14">Matsumoto et al., 2005a</xref>; <xref ref-type="bibr" rid="B25">Sugase-Miyamoto et al., 2014</xref>). In our previous study, we constructed a deep neural network (DNN) to compare information representation in each layer and information encoded by a neural population in area TE with a visual stimulus set that included human and monkey faces (<xref ref-type="bibr" rid="B13">Matsumoto et al., 2021</xref>). We found that the time course of hierarchical categorization could not be replicated with the DNN. Furthermore, global categorization occurred in the lower layers of the DNN. In this study, we hypothesize that visual short-term memory is retrieved from global to fine information of images <italic>via</italic> recurrent connections in area TE. To test this hypothesis, we constructed a combined model of a DNN, i.e., Xception net (<xref ref-type="bibr" rid="B4">Chollet, 2017</xref>), and a recurrent neural network, i.e., Hopfield model (<xref ref-type="bibr" rid="B7">Hopfield, 1982</xref>). The Hopfield model is known as an associative memory model (<xref ref-type="bibr" rid="B2">Anderson, 1972</xref>; <xref ref-type="bibr" rid="B10">Kohonen, 1972</xref>; <xref ref-type="bibr" rid="B17">Nakano, 1972</xref>). An associative memory model is considered a short-term memory model because it can store and retrieve original images from noise-degraded images. The combined model performed better for adversarial examples than using only the Xception net. The combined model also outputs the time course of hierarchical categorization. This indicates that recurrent connections in the Hopfield model are important for hierarchical categorization, suggesting that recurrent connections in area TE are important for such categorization.</p>
</sec>
<sec id="S2" sec-type="materials|methods">
<title>Materials and Methods</title>
<sec id="S2.SS1">
<title>Model</title>
<p>We constructed our combined model consisting of an Xception net and a Hopfield model to investigate whether it can replicate the time course of hierarchical categorization (<xref ref-type="fig" rid="F1">Figure 1A</xref>). Model parameters including weight values of the original Xception net were downloaded from <ext-link ext-link-type="uri" xlink:href="https://github.com/keras-team/keras">https://github.com/keras-team/keras</ext-link>. The downloaded weight values were determined from images in the ImageNet database (<xref ref-type="bibr" rid="B19">Russakovsky et al., 2015</xref>). The weight values of the Xception net were fixed in this study. The top layer of the original Xception net is a fully connected layer that outputs the probability of each category. The fully connected layer was removed from the original Xception net, and the Hopfield model was inserted instead as a model of area TE. This was done because our previous studies showed that the information representation in fully connected layers of a DNN was similar to the representation in area TE (<xref ref-type="bibr" rid="B13">Matsumoto et al., 2021</xref>) and that an associative memory model was able to reproduce the neural activities of area TE (<xref ref-type="bibr" rid="B15">Matsumoto et al., 2005b</xref>). We compared the performance of the combined model with another model, i.e., the Xception model without the Hopfield model (<xref ref-type="fig" rid="F1">Figure 1B</xref>). The inputs to the models were visual images (250 &#x00D7; 250 pixels, RGB color) and the outputs were the image category probabilities. In the learning phase, the weights of a binary dense layer (<xref ref-type="bibr" rid="B8">Hubara et al., 2016</xref>) and fully connected layers were learned using a backpropagation algorithm (<xref ref-type="bibr" rid="B18">Rumelhart et al., 1986</xref>) in both models, and weights of the Hopfield model were learned by the Storkey rule (<xref ref-type="bibr" rid="B22">Storkey, 1997</xref>) or the covariance rule for the combined model. In the test phase, adversarial examples generated from the learned images or learned images with Gaussian noise were given as input to the combined model. The code of the model was written using TensorFlow (<xref ref-type="bibr" rid="B1">Abadi et al., 2015</xref>) and Keras (<xref ref-type="bibr" rid="B3">Chollet, 2015</xref>).</p>
<fig id="F1" position="float">
<label>FIGURE 1</label>
<caption><p>The model structures for <bold>(A)</bold> combined model and <bold>(B)</bold> Xception model.</p></caption>
<graphic mimetype="image" mime-subtype="tiff" xlink:href="fnsys-16-805990-g001.tif"/>
</fig>
<p>The Hopfield model consists of <italic>N</italic> neurons. The internal potential of neuron <italic>i</italic> at time <italic>t</italic> is denoted as <italic>h</italic><sub><italic>i</italic></sub>(<italic>t</italic>) and updated as given by the following equation,</p>
<disp-formula id="S2.E1">
<label>(1)</label>
<mml:math id="M1">
<mml:mrow>
<mml:mrow>
<mml:mrow>
<mml:msub>
<mml:mi>h</mml:mi>
<mml:mi>i</mml:mi>
</mml:msub>
<mml:mo>&#x2062;</mml:mo>
<mml:mrow>
<mml:mo>(</mml:mo>
<mml:mi>t</mml:mi>
<mml:mo>)</mml:mo>
</mml:mrow>
</mml:mrow>
<mml:mo>=</mml:mo>
<mml:mrow>
<mml:munderover>
<mml:mo largeop="true" movablelimits="false" symmetric="true">&#x2211;</mml:mo>
<mml:mrow>
<mml:mi>j</mml:mi>
<mml:mo>&#x2260;</mml:mo>
<mml:mi>i</mml:mi>
</mml:mrow>
<mml:mi>N</mml:mi>
</mml:munderover>
<mml:mrow>
<mml:msub>
<mml:mi>J</mml:mi>
<mml:mrow>
<mml:mi>i</mml:mi>
<mml:mo>&#x2062;</mml:mo>
<mml:mi>j</mml:mi>
</mml:mrow>
</mml:msub>
<mml:mo>&#x2062;</mml:mo>
<mml:msub>
<mml:mi>s</mml:mi>
<mml:mi>j</mml:mi>
</mml:msub>
<mml:mo>&#x2062;</mml:mo>
<mml:mrow>
<mml:mo stretchy="false">(</mml:mo>
<mml:mi>t</mml:mi>
<mml:mo stretchy="false">)</mml:mo>
</mml:mrow>
</mml:mrow>
</mml:mrow>
</mml:mrow>
<mml:mo>,</mml:mo>
</mml:mrow>
</mml:math>
</disp-formula>
<p>where <italic>J</italic><sub><italic>ij</italic></sub> denotes a synaptic weight of recurrent connection from neuron <italic>j</italic> to neuron <italic>i</italic>, and <italic>s</italic><sub><italic>j</italic></sub>(<italic>t</italic>) denotes the state of neuron <italic>j</italic> at time <italic>t</italic> (<italic>s</italic><sub><italic>j</italic></sub>(<italic>t</italic>) = {1, &#x2212;1}):</p>
<disp-formula id="S2.E2">
<label>(2)</label>
<mml:math id="M2">
<mml:mrow>
<mml:mrow>
<mml:mrow>
<mml:msub>
<mml:mi>s</mml:mi>
<mml:mi>j</mml:mi>
</mml:msub>
<mml:mo>&#x2062;</mml:mo>
<mml:mrow>
<mml:mo>(</mml:mo>
<mml:mrow>
<mml:mi>t</mml:mi>
<mml:mo>+</mml:mo>
<mml:mn>1</mml:mn>
</mml:mrow>
<mml:mo>)</mml:mo>
</mml:mrow>
</mml:mrow>
<mml:mo>=</mml:mo>
<mml:mrow>
<mml:mpadded width="+1.7pt">
<mml:mtext>sign</mml:mtext>
</mml:mpadded>
<mml:mo>&#x2062;</mml:mo>
<mml:mrow>
<mml:mo stretchy="false">(</mml:mo>
<mml:mrow>
<mml:msub>
<mml:mi>h</mml:mi>
<mml:mi>j</mml:mi>
</mml:msub>
<mml:mo>&#x2062;</mml:mo>
<mml:mrow>
<mml:mo>(</mml:mo>
<mml:mi>t</mml:mi>
<mml:mo>)</mml:mo>
</mml:mrow>
</mml:mrow>
<mml:mo stretchy="false">)</mml:mo>
</mml:mrow>
</mml:mrow>
</mml:mrow>
<mml:mo>,</mml:mo>
</mml:mrow>
</mml:math>
</disp-formula>
<p>where sign[<italic>h</italic><sub><italic>j</italic></sub>(<italic>t</italic>)] is a sign function: if <italic>h</italic><sub><italic>j</italic></sub>(t) &#x2265; 0, sign[<italic>h</italic><sub><italic>j</italic></sub>(<italic>t</italic>)] = 1: otherwise, sign[<italic>h</italic><sub><italic>j</italic></sub>(<italic>t</italic>)] = &#x2212;1. A feature vector of the binary dense layer was used as the memory pattern &#x03BE;<sup>&#x03BC;</sup> for each image and set as an initial state, <italic>s</italic>(0), of the Hopfield model. The weight was determined by the Storkey rule (results are shown in <xref ref-type="fig" rid="F2">Figure 2</xref>),</p>
<disp-formula id="S2.E3">
<label>(3)</label>
<mml:math id="M3">
<mml:mrow>
<mml:mrow>
<mml:msubsup>
<mml:mi>J</mml:mi>
<mml:mrow>
<mml:mi>i</mml:mi>
<mml:mo>&#x2062;</mml:mo>
<mml:mi>j</mml:mi>
</mml:mrow>
<mml:mi>&#x03BD;</mml:mi>
</mml:msubsup>
<mml:mo>=</mml:mo>
<mml:mrow>
<mml:mrow>
<mml:msubsup>
<mml:mi>J</mml:mi>
<mml:mrow>
<mml:mi>i</mml:mi>
<mml:mo>&#x2062;</mml:mo>
<mml:mi>j</mml:mi>
</mml:mrow>
<mml:mrow>
<mml:mi>&#x03BD;</mml:mi>
<mml:mo>-</mml:mo>
<mml:mn>1</mml:mn>
</mml:mrow>
</mml:msubsup>
<mml:mo>+</mml:mo>
<mml:mrow>
<mml:mfrac>
<mml:mn>1</mml:mn>
<mml:mi>N</mml:mi>
</mml:mfrac>
<mml:mo>&#x2062;</mml:mo>
<mml:msubsup>
<mml:mi>&#x03BE;</mml:mi>
<mml:mi>i</mml:mi>
<mml:mi>&#x03BD;</mml:mi>
</mml:msubsup>
<mml:mo>&#x2062;</mml:mo>
<mml:msubsup>
<mml:mi>&#x03BE;</mml:mi>
<mml:mi>j</mml:mi>
<mml:mi>&#x03BD;</mml:mi>
</mml:msubsup>
</mml:mrow>
</mml:mrow>
<mml:mo>-</mml:mo>
<mml:mrow>
<mml:mfrac>
<mml:mn>1</mml:mn>
<mml:mi>N</mml:mi>
</mml:mfrac>
<mml:mo>&#x2062;</mml:mo>
<mml:msubsup>
<mml:mi>&#x03BE;</mml:mi>
<mml:mi>i</mml:mi>
<mml:mi>&#x03BD;</mml:mi>
</mml:msubsup>
<mml:mo>&#x2062;</mml:mo>
<mml:msubsup>
<mml:mi>f</mml:mi>
<mml:mrow>
<mml:mi>j</mml:mi>
<mml:mo>&#x2062;</mml:mo>
<mml:mi>i</mml:mi>
</mml:mrow>
<mml:mi>&#x03BD;</mml:mi>
</mml:msubsup>
</mml:mrow>
<mml:mo>-</mml:mo>
<mml:mrow>
<mml:mfrac>
<mml:mn>1</mml:mn>
<mml:mi>N</mml:mi>
</mml:mfrac>
<mml:mo>&#x2062;</mml:mo>
<mml:msubsup>
<mml:mi>&#x03BE;</mml:mi>
<mml:mi>j</mml:mi>
<mml:mi>&#x03BD;</mml:mi>
</mml:msubsup>
<mml:mo>&#x2062;</mml:mo>
<mml:msubsup>
<mml:mi>f</mml:mi>
<mml:mrow>
<mml:mi>i</mml:mi>
<mml:mo>&#x2062;</mml:mo>
<mml:mi>j</mml:mi>
</mml:mrow>
<mml:mi>&#x03BD;</mml:mi>
</mml:msubsup>
</mml:mrow>
</mml:mrow>
</mml:mrow>
<mml:mo>,</mml:mo>
</mml:mrow>
</mml:math>
</disp-formula>
<fig id="F2" position="float">
<label>FIGURE 2</label>
<caption><p><bold>(A)</bold> Adversarial examples for perturbation parameter. The original image is taken from the ImageNet database. <bold>(B)</bold> Accuracy in estimating correct category for perturbation parameters of the Xception model (red line) and combined model (blue line). <bold>(C)</bold> Time course for the probability of each category for Retriever image <bold>(A)</bold> at perturbation parameter 0.26.</p></caption>
<graphic mimetype="image" mime-subtype="tiff" xlink:href="fnsys-16-805990-g002.tif"/>
</fig>
<p>where &#x03BD;={1,&#x2026;,&#x03BC;}, <italic>J<sub><italic>ij</italic></sub> = J<sub><italic>ij</italic></sub><sup>&#x03BC;</sup></italic>, and <italic>f<sub><italic>ij</italic></sub><sup>&#x03BD;</sup></italic> obeys:</p>
<disp-formula id="S2.E4">
<label>(4)</label>
<mml:math id="M4">
<mml:mrow>
<mml:mrow>
<mml:msubsup>
<mml:mi>f</mml:mi>
<mml:mrow>
<mml:mi>i</mml:mi>
<mml:mo>&#x2062;</mml:mo>
<mml:mi>j</mml:mi>
</mml:mrow>
<mml:mi>&#x03BD;</mml:mi>
</mml:msubsup>
<mml:mo>=</mml:mo>
<mml:mrow>
<mml:munderover>
<mml:mo largeop="true" movablelimits="false" symmetric="true">&#x2211;</mml:mo>
<mml:mrow>
<mml:mi>k</mml:mi>
<mml:mo>&#x2260;</mml:mo>
<mml:mrow>
<mml:mi>i</mml:mi>
<mml:mo>,</mml:mo>
<mml:mi>j</mml:mi>
</mml:mrow>
</mml:mrow>
<mml:mi>N</mml:mi>
</mml:munderover>
<mml:mrow>
<mml:msubsup>
<mml:mi>J</mml:mi>
<mml:mrow>
<mml:mi>i</mml:mi>
<mml:mo>&#x2062;</mml:mo>
<mml:mi>k</mml:mi>
</mml:mrow>
<mml:mrow>
<mml:mi>&#x03BD;</mml:mi>
<mml:mo>-</mml:mo>
<mml:mn>1</mml:mn>
</mml:mrow>
</mml:msubsup>
<mml:mo>&#x2062;</mml:mo>
<mml:msubsup>
<mml:mi>&#x03BE;</mml:mi>
<mml:mi>k</mml:mi>
<mml:mi>&#x03BD;</mml:mi>
</mml:msubsup>
</mml:mrow>
</mml:mrow>
</mml:mrow>
<mml:mo>.</mml:mo>
</mml:mrow>
</mml:math>
</disp-formula>
<p>The weight <italic>J</italic><sub><italic>ij</italic></sub> was also determined by the covariance rule (results are shown in <xref ref-type="fig" rid="F3">Figure 3</xref>),</p>
<disp-formula id="S2.E5">
<label>(5)</label>
<mml:math id="M5">
<mml:mrow>
<mml:mrow>
<mml:msub>
<mml:mi>J</mml:mi>
<mml:mrow>
<mml:mi>i</mml:mi>
<mml:mo>&#x2062;</mml:mo>
<mml:mi>j</mml:mi>
</mml:mrow>
</mml:msub>
<mml:mo>=</mml:mo>
<mml:mrow>
<mml:mfrac>
<mml:mn>1</mml:mn>
<mml:mi>N</mml:mi>
</mml:mfrac>
<mml:mo>&#x2062;</mml:mo>
<mml:mrow>
<mml:munderover>
<mml:mo largeop="true" movablelimits="false" symmetric="true">&#x2211;</mml:mo>
<mml:mi mathvariant="normal">&#x03BC;</mml:mi>
<mml:mi>p</mml:mi>
</mml:munderover>
<mml:mrow>
<mml:mrow>
<mml:mo>(</mml:mo>
<mml:mrow>
<mml:msubsup>
<mml:mi>&#x03BE;</mml:mi>
<mml:mi>i</mml:mi>
<mml:mi mathvariant="normal">&#x03BC;</mml:mi>
</mml:msubsup>
<mml:mo>-</mml:mo>
<mml:mi>m</mml:mi>
</mml:mrow>
<mml:mo>)</mml:mo>
</mml:mrow>
<mml:mo>&#x2062;</mml:mo>
<mml:mrow>
<mml:mo>(</mml:mo>
<mml:mrow>
<mml:msubsup>
<mml:mi>&#x03BE;</mml:mi>
<mml:mi>j</mml:mi>
<mml:mi mathvariant="normal">&#x03BC;</mml:mi>
</mml:msubsup>
<mml:mo>-</mml:mo>
<mml:mi>m</mml:mi>
</mml:mrow>
<mml:mo>)</mml:mo>
</mml:mrow>
</mml:mrow>
</mml:mrow>
</mml:mrow>
</mml:mrow>
<mml:mo>,</mml:mo>
</mml:mrow>
</mml:math>
</disp-formula>
<fig id="F3" position="float">
<label>FIGURE 3</label>
<caption><p>Time course of the probability of each category for Woman and Poodle images with Gaussian noise. <bold>(A,C)</bold> Woman and Poodle images with Gaussian noise. <bold>(B,D)</bold> Time course of the probability of each category for Woman and Poodle images. Cyan: Human, magenta: Woman, black: Japanese, red: Dog, green: Dalmatian, blue: Poodle.</p></caption>
<graphic mimetype="image" mime-subtype="tiff" xlink:href="fnsys-16-805990-g003.tif"/>
</fig>
<p>where <italic>m</italic> is the average of <italic>&#x03BE;<sub><italic>i</italic></sub><sup>&#x03BC;</sup></italic>.</p>
</sec>
</sec>
<sec id="S3" sec-type="results">
<title>Results</title>
<sec id="S3.SS1">
<title>Adversarial Examples</title>
<p>We tested whether our combined model can retrieve the correct category of images from noise-degraded images, i.e., adversarial examples. Adversarial examples were generated using VGG16 (<xref ref-type="bibr" rid="B20">Simonyan and Zisserman, 2014</xref>) and the fast gradient sign method (FGSM) (<xref ref-type="bibr" rid="B6">Goodfellow et al., 2014</xref>). We changed a perturbation parameter to obtain different amounts of noise (<xref ref-type="fig" rid="F2">Figure 2A</xref>). In the learning phase, the weights of the binary dense layers and fully connected layers were learned from the 250 images by using the backpropagation algorithm. The weights of the Hopfield model (<italic>N</italic> = 5,000) were learned using the Storkey rule (<xref ref-type="bibr" rid="B22">Storkey, 1997</xref>) with 250 original images of 50 categories (<xref ref-type="supplementary-material" rid="TS1">Supplementary Table 1</xref>) taken randomly from the ImageNet database. In the test phase, the largest difference between the accuracies of the estimating categories of adversarial examples for the combined model and the Xception model was 9.2%, i.e., the accuracies were 72.4% (combined model) and 63.2% (Xception model), at the perturbation parameter 0.26 (<xref ref-type="fig" rid="F2">Figure 2B</xref>). At the perturbation parameter 0.26, the combined model outputted the Rifle category at <italic>t</italic> = 0 for the image in <xref ref-type="fig" rid="F2">Figure 2A</xref> and then outputted the Retriever category (<xref ref-type="fig" rid="F2">Figure 2C</xref>). In other words, the model has an error-correcting ability of an associative memory model. At the perturbation parameter 0.26, the number of adversarial examples for each model performance is shown in <xref ref-type="table" rid="T1">Table 1</xref>.</p>
<table-wrap position="float" id="T1">
<label>TABLE 1</label>
<caption><p>Number of adversarial examples classified by performance for the Xception model and the combined model at the perturbation parameter 0.26.</p></caption>
<table cellspacing="5" cellpadding="5" frame="hsides" rules="groups">
<thead>
<tr>
<td valign="top" align="left"></td>
<td valign="top" align="center">Xception: correct</td>
<td valign="top" align="center">Xception: incorrect</td>
</tr>
</thead>
<tbody>
<tr>
<td valign="top" align="left">Combined: correct</td>
<td valign="top" align="center">148</td>
<td valign="top" align="center">34</td>
</tr>
<tr>
<td valign="top" align="left">Combined: incorrect</td>
<td valign="top" align="center">59</td>
<td valign="top" align="center">9</td>
</tr>
</tbody>
</table></table-wrap>
</sec>
<sec id="S3.SS2">
<title>Images With Gaussian Noise</title>
<p>To examine whether the hierarchical categorizations were observed in the combined model, the combined model was tested using images with Gaussian noise. In the learning phase, the weights of the binary dense layers and fully connected layers were learned from the 30 original images of six categories (Human, Woman, Japanese, Dogs, Dalmatian, and Poodle) by using the backpropagation algorithm. The weights of the Hopfield model (<italic>N</italic> = 2,048) were learned using the covariance rule with 20 original images of four categories (Woman, Japanese, Dalmatian, and Poodle). Images of super-categories, i.e., Human and Dog,</p>
<p>were not learned in the Hopfield model. In the test phase, the learned images with Gaussian noise (mean: 0, variance: 0.1, size: 15 &#x00D7; 15 pixels) (<xref ref-type="fig" rid="F3">Figures 3A,C</xref>) were given as input to the combined model. The model outputted the probability of each category at each time step. When a Woman or Poodle image with Gaussian noise (<xref ref-type="fig" rid="F3">Figures 3A,C</xref>) was presented to the combined model, the model initially responded with the Human or Dog category, then responded with the correct category, i.e., Woman or Poodle (<xref ref-type="fig" rid="F3">Figures 3B,D</xref>). The Hopfield model did not process information at the initial time step, <italic>t = 0</italic>. Therefore, the combined model was the same as the Xception model at only <italic>t = 0</italic>. The sum of the probability of each category was 1. At the initial time step <italic>t</italic> = 0, multiple categories had small probabilities, so the difference between Dog and Dalmatian became small. At <italic>t</italic> = 10 only a few categories had values of probability, and therefore, the difference among the categories became large. In <xref ref-type="fig" rid="F3">Figure 3B</xref>, the output was Human (super-category) at <italic>t = 0</italic>, followed by Woman (sub-category). In <xref ref-type="fig" rid="F3">Figure 3D</xref>, the output was Dog (super-category) at <italic>t = 0</italic>, followed by Poodle (sub-category), then Dalmatian, and finally Poodle again. In other words, the combined model has an error-correcting ability of an associative memory model as shown in the previous paragraph. Two of the three images that were assigned the correct category had this trend of hierarchical categorizations.</p>
<p>To understand the temporal behavior of the Hopfield model, we projected the neuronal states into this model, i.e., 2,048-dimensional vectors, for 20 images into a two-dimensional space by principal component analysis (PCA) (<xref ref-type="bibr" rid="B14">Matsumoto et al., 2005a</xref>), as shown in <xref ref-type="fig" rid="F4">Figure 4</xref>. The horizontal and vertical axes indicate the first and second principal components (PC1, PC2). The red points indicate Woman or Japanese. The blue points indicate Dalmatian or Poodle. At <italic>t</italic> = 0, the distributions for state vectors of Dalmatian and Poodle, and Woman and Japanese overlapped (<xref ref-type="fig" rid="F4">Figure 4A</xref>). At <italic>t</italic> = 5, many state vectors for Dalmatian and Poodle were projected into the left side of <xref ref-type="fig" rid="F4">Figure 4B</xref>, and most state vectors for Japanese were projected into the right side of <xref ref-type="fig" rid="F4">Figure 4B</xref>. At <italic>t</italic> = 30, there were four clusters. A cluster contained the vectors of Woman and Japanese (<xref ref-type="fig" rid="F4">Figure 4C</xref>). The others contained all four categories, i.e., Woman, Japanese, Dalmatian, and Poodle. Therefore, different categories were encoded in a different time course with the Hopfield model.</p>
<fig id="F4" position="float">
<label>FIGURE 4</label>
<caption><p>Two-dimensional space of state vectors of Hopfield model obtained by principal component analysis (PCA) at <bold>(A)</bold> <italic>t</italic> = 0, <bold>(B)</bold> <italic>t</italic> = 5, and <bold>(C)</bold> <italic>t</italic> = 30. Red circles: Woman, red crosses: Japanese, blue squares: Dalmatian, blue diamonds: Poodle.</p></caption>
<graphic mimetype="image" mime-subtype="tiff" xlink:href="fnsys-16-805990-g004.tif"/>
</fig>
</sec>
</sec>
<sec id="S4" sec-type="discussion">
<title>Discussion</title>
<p>We constructed a model that combined an Xception net and a Hopfield model to investigate whether it can replicate the time course of a hierarchical categorization. The combined model for adversarial examples performed better than the Xception model. The combined model also outputted different categories during the time course. These results indicate that recurrent connections in the Hopfield model are important not only for short-term memory but also for hierarchical categorization, suggesting that recurrent connections in area TE are important for hierarchical categorization.</p>
<p>In our previous study, we showed that the behavior of an associative memory model was qualitatively similar to that of neurons in area TE (<xref ref-type="bibr" rid="B15">Matsumoto et al., 2005b</xref>). The model we constructed for that study used random bit patterns not visual images as input. In another study, we constructed a DNN, i.e., AlexNet (<xref ref-type="bibr" rid="B12">Krizhevsky et al., 2012</xref>), to compare the information represented in each layer and the information encoded by a neural population in area TE with a visual stimulus set that included human and monkey faces (<xref ref-type="bibr" rid="B13">Matsumoto et al., 2021</xref>). Thus, the representation in the fully connected layers of AlexNet most resembled the representation of TE neurons for human and monkey faces. Studies have suggested that recurrent processing is important for visual recognition (<xref ref-type="bibr" rid="B21">Spoerer et al., 2017</xref>; <xref ref-type="bibr" rid="B9">Kar et al., 2019</xref>). These models consist of recurrent connections in all layers, and each layer is not a Hopfield model. In a combined model which consisted of a DNN and a recurrent network, e.g., long short-term memory (LSTM) in <xref ref-type="bibr" rid="B11">Koo et al. (2019)</xref>, to output hierarchical categories, a feature vector from top to bottom layer was given as input to LSTM at each time step. The feature vector in the top layer was inputted to LSTM at <italic>t</italic> = 0, the vector in the second-top layer was inputted to LSTM at <italic>t</italic> = 1. Therefore, the feature vectors in all layers should be stored in the memory. In our combined model a feature vector from a single layer of the Xception net was given as input to the Hopfield model at initial time step <italic>t</italic> = 0. The vector was updated by recurrent connections of the Hopfield model. Therefore, the structures of our combined model and the combined model of <xref ref-type="bibr" rid="B11">Koo et al. (2019)</xref> are different, and the structure of our model requires less memory consumption than that of the model of <xref ref-type="bibr" rid="B11">Koo et al. (2019)</xref>. In our combined model, we added recurrent connections only to the Hopfield model layer to investigate whether recurrent processing in area TE is important for hierarchical categorization. We considered the Hopfield model as modeling for area TE in the higher visual cortex. The fully connected layers in our model were considered to be the prefrontal cortex or other higher brain areas that judge categories of visual images. Thus, our model can retrieve hierarchical categorical information from noise-degraded images and be considered as a model for short-term memory.</p>
</sec>
<sec id="S5" sec-type="data-availability">
<title>Data Availability Statement</title>
<p>The raw data supporting the conclusions of this article will be made available by the corresponding author, NM (<email>xmatumo@ni.aist.go.jp</email>), upon reasonable request.</p>
</sec>
<sec id="S6">
<title>Author Contributions</title>
<p>NM, MO, and YS-M designed the research and discussed the data. NM, YT, MS, and SK conducted the modeling. NM wrote the draft of the article. MO and YS-M revised the manuscript. All authors approved the final version of the manuscript.</p>
</sec>
<sec id="conf1" sec-type="COI-statement">
<title>Conflict of Interest</title>
<p>YT is currently employed by the company IBM Japan. The remaining authors declare that the research was conducted in the absence of any commercial or financial relationships that could be construed as a potential conflict of interest.</p>
</sec>
<sec id="pudiscl1" sec-type="disclaimer">
<title>Publisher&#x2019;s Note</title>
<p>All claims expressed in this article are solely those of the authors and do not necessarily represent those of their affiliated organizations, or those of the publisher, the editors and the reviewers. Any product that may be evaluated in this article, or claim that may be made by its manufacturer, is not guaranteed or endorsed by the publisher.</p>
</sec>
</body>
<back>
<sec id="S7" sec-type="funding-information">
<title>Funding</title>
<p>This work was supported by the KAKENHI (26120535, 16H01561, 16H01684, 18H0520, 19K07804, and 19K12149) and this article was based on results obtained from a project commissioned by the New Energy and Industrial Technology Development Organization (NEDO).</p>
</sec>
<sec id="S8" sec-type="supplementary-material">
<title>Supplementary Material</title>
<p>The Supplementary Material for this article can be found online at: <ext-link ext-link-type="uri" xlink:href="https://www.frontiersin.org/articles/10.3389/fnsys.2022.805990/full#supplementary-material">https://www.frontiersin.org/articles/10.3389/fnsys.2022.805990/full#supplementary-material</ext-link></p>
<supplementary-material xlink:href="Table_1.docx" id="TS1" mimetype="application/vnd.openxmlformats-officedocument.wordprocessingml.document" xmlns:xlink="http://www.w3.org/1999/xlink"/>
</sec>
<ref-list>
<title>References</title>
<ref id="B1"><citation citation-type="journal"><person-group person-group-type="author"><name><surname>Abadi</surname> <given-names>M.</given-names></name> <name><surname>Agarwal</surname> <given-names>A.</given-names></name> <name><surname>Barham</surname> <given-names>P.</given-names></name> <name><surname>Brevdo</surname> <given-names>E.</given-names></name> <name><surname>Chen</surname> <given-names>Z.</given-names></name> <name><surname>Citro</surname> <given-names>C.</given-names></name><etal/></person-group> (<year>2015</year>). <source><italic>TensorFlow: Large-Scale Machine Learning on Heterogeneous Systems.</italic></source> Available online at: <ext-link ext-link-type="uri" xlink:href="https://www.tensorflow.org/">https://www.tensorflow.org/</ext-link> <comment>(accessed June 6, 2018)</comment>.</citation></ref>
<ref id="B2"><citation citation-type="journal"><person-group person-group-type="author"><name><surname>Anderson</surname> <given-names>J. A.</given-names></name></person-group> (<year>1972</year>). <article-title>A simple neural network generating an interactive memory.</article-title> <source><italic>Math. Biosci.</italic></source> <volume>14</volume> <fpage>197</fpage>&#x2013;<lpage>220</lpage>.</citation></ref>
<ref id="B3"><citation citation-type="journal"><person-group person-group-type="author"><name><surname>Chollet</surname> <given-names>F.</given-names></name></person-group> (<year>2015</year>). <source><italic>Keras.</italic></source> Available online at: <ext-link ext-link-type="uri" xlink:href="https://keras.io/">https://keras.io/</ext-link> <comment>(accessed 6, 2018)</comment>.</citation></ref>
<ref id="B4"><citation citation-type="journal"><person-group person-group-type="author"><name><surname>Chollet</surname> <given-names>F.</given-names></name></person-group> (<year>2017</year>). &#x201C;<article-title>Xception: deep learning with depthwise separable convolutions,</article-title>&#x201D; in <source><italic>Proceedings of the 30th IEEE Conference on Computer Vision and Pattern Recognition (CVPR 2017)</italic></source>, <publisher-loc>Honolulu, HI</publisher-loc>, <fpage>1800</fpage>&#x2013;<lpage>1807</lpage>.</citation></ref>
<ref id="B5"><citation citation-type="journal"><person-group person-group-type="author"><name><surname>Freedman</surname> <given-names>D. J.</given-names></name> <name><surname>Riesenhuber</surname> <given-names>M.</given-names></name> <name><surname>Poggio</surname> <given-names>T.</given-names></name> <name><surname>Miller</surname> <given-names>E. K.</given-names></name></person-group> (<year>2001</year>). <article-title>Categorical representation of visual stimuli in the primate prefrontal cortex.</article-title> <source><italic>Science</italic></source> <volume>291</volume> <fpage>312</fpage>&#x2013;<lpage>316</lpage>. <pub-id pub-id-type="doi">10.1126/science.291.5502.312</pub-id> <pub-id pub-id-type="pmid">11209083</pub-id></citation></ref>
<ref id="B6"><citation citation-type="journal"><person-group person-group-type="author"><name><surname>Goodfellow</surname> <given-names>I.</given-names></name> <name><surname>Pouget-Abadie</surname> <given-names>J.</given-names></name> <name><surname>Mirza</surname> <given-names>M.</given-names></name> <name><surname>Xu</surname> <given-names>B.</given-names></name> <name><surname>Warde-Farley</surname> <given-names>D.</given-names></name> <name><surname>Ozair</surname> <given-names>S.</given-names></name><etal/></person-group> (<year>2014</year>). &#x201C;<article-title>Generative adversarial nets,</article-title>&#x201D; in <source><italic>Proceedings of the 27th International Conference on Neural Information Processing Systems NIPS&#x2019;14</italic></source>, <publisher-loc>Montreal, QC</publisher-loc>, <fpage>2672</fpage>&#x2013;<lpage>2680</lpage>.</citation></ref>
<ref id="B7"><citation citation-type="journal"><person-group person-group-type="author"><name><surname>Hopfield</surname> <given-names>J. J.</given-names></name></person-group> (<year>1982</year>). <article-title>Neural networks and physical systems with emergent collective computational abilities.</article-title> <source><italic>Proc. Natl. Acad. Sci. U.S.A.</italic></source> <volume>79</volume> <fpage>2554</fpage>&#x2013;<lpage>2558</lpage>. <pub-id pub-id-type="doi">10.1073/pnas.79.8.2554</pub-id> <pub-id pub-id-type="pmid">6953413</pub-id></citation></ref>
<ref id="B8"><citation citation-type="journal"><person-group person-group-type="author"><name><surname>Hubara</surname> <given-names>I.</given-names></name> <name><surname>Courbariaux</surname> <given-names>M.</given-names></name> <name><surname>Soudry</surname> <given-names>D.</given-names></name> <name><surname>El-Yaniv</surname> <given-names>R.</given-names></name> <name><surname>Bengio</surname> <given-names>Y.</given-names></name></person-group> (<year>2016</year>). &#x201C;<article-title>Binarized neural networks,</article-title>&#x201D; in <source><italic>Proceedings of the 29th Advances in Neural Information Processing Systems (NIPS 2016)</italic></source>, <publisher-loc>Barcelona</publisher-loc>.</citation></ref>
<ref id="B9"><citation citation-type="journal"><person-group person-group-type="author"><name><surname>Kar</surname> <given-names>K.</given-names></name> <name><surname>Kubilius</surname> <given-names>J.</given-names></name> <name><surname>Schmidt</surname> <given-names>K.</given-names></name> <name><surname>Issa</surname> <given-names>E. B.</given-names></name> <name><surname>DiCarlo</surname> <given-names>J. J.</given-names></name></person-group> (<year>2019</year>). <article-title>Evidence that recurrent circuits are critical to the ventral stream&#x2019;s execution of core object recognition behavior.</article-title> <source><italic>Nat. Neurosci.</italic></source> <volume>22</volume> <fpage>974</fpage>&#x2013;<lpage>983</lpage>. <pub-id pub-id-type="doi">10.1038/s41593-019-0392-5</pub-id> <pub-id pub-id-type="pmid">31036945</pub-id></citation></ref>
<ref id="B10"><citation citation-type="journal"><person-group person-group-type="author"><name><surname>Kohonen</surname> <given-names>T.</given-names></name></person-group> (<year>1972</year>). <article-title>Correlation matrix memories.</article-title> <source><italic>IEEE Trans. Comput.</italic></source> <volume>C-21</volume>, <fpage>353</fpage>&#x2013;<lpage>359</lpage>. <pub-id pub-id-type="doi">10.1109/tc.1972.5008975</pub-id></citation></ref>
<ref id="B11"><citation citation-type="journal"><person-group person-group-type="author"><name><surname>Koo</surname> <given-names>J.</given-names></name> <name><surname>Klabjan</surname> <given-names>D.</given-names></name> <name><surname>Utke</surname> <given-names>J.</given-names></name></person-group> (<year>2019</year>). <source><italic>Combined Convolutional and Recurrent Neural Networks for Hierarchical Classification of Images.</italic></source> Available online at: <ext-link ext-link-type="uri" xlink:href="http://arxiv.org/abs/1809.09574">http://arxiv.org/abs/1809.09574</ext-link> <comment>(accessed December 24, 2021)</comment>.</citation></ref>
<ref id="B12"><citation citation-type="journal"><person-group person-group-type="author"><name><surname>Krizhevsky</surname> <given-names>A.</given-names></name> <name><surname>Sutskever</surname> <given-names>I.</given-names></name> <name><surname>Hinton</surname> <given-names>G. E.</given-names></name></person-group> (<year>2012</year>). <article-title>ImageNet classification with deep convolutional neural networks</article-title>. <source><italic>Adv. Neural Inf. Process. Syst.</italic></source> <volume>25</volume>, <fpage>1106</fpage>&#x2013;<lpage>1114</lpage>.</citation></ref>
<ref id="B13"><citation citation-type="journal"><person-group person-group-type="author"><name><surname>Matsumoto</surname> <given-names>N.</given-names></name> <name><surname>Mototake</surname> <given-names>Y.</given-names></name> <name><surname>Kawano</surname> <given-names>K.</given-names></name> <name><surname>Okada</surname> <given-names>M.</given-names></name> <name><surname>Sugase-Miyamoto</surname> <given-names>Y.</given-names></name></person-group> (<year>2021</year>). <article-title>Comparison of neuronal responses in primate inferior-temporal cortex and feed-forward deep neural network model with regard to information processing of faces.</article-title> <source><italic>J. Comput. Neurosci.</italic></source> <volume>49</volume> <fpage>251</fpage>&#x2013;<lpage>257</lpage>.</citation></ref>
<ref id="B14"><citation citation-type="journal"><person-group person-group-type="author"><name><surname>Matsumoto</surname> <given-names>N.</given-names></name> <name><surname>Okada</surname> <given-names>M.</given-names></name> <name><surname>Sugase-Miyamoto</surname> <given-names>Y.</given-names></name> <name><surname>Yamane</surname> <given-names>S.</given-names></name> <name><surname>Kawano</surname> <given-names>K.</given-names></name></person-group> (<year>2005a</year>). <article-title>Population dynamics of face-responsive neurons in the inferior temporal cortex.</article-title> <source><italic>Cereb. Cortex</italic></source> <volume>15</volume> <fpage>1103</fpage>&#x2013;<lpage>1112</lpage>. <pub-id pub-id-type="doi">10.1093/cercor/bhh209</pub-id> <pub-id pub-id-type="pmid">15563724</pub-id></citation></ref>
<ref id="B15"><citation citation-type="journal"><person-group person-group-type="author"><name><surname>Matsumoto</surname> <given-names>N.</given-names></name> <name><surname>Okada</surname> <given-names>M.</given-names></name> <name><surname>Sugase-Miyamoto</surname> <given-names>Y.</given-names></name> <name><surname>Yamane</surname> <given-names>S.</given-names></name></person-group> (<year>2005b</year>). <article-title>Neuronal mechanisms encoding global-to-fine information in inferior-temporal cortex.</article-title> <source><italic>J. Comput. Neurosci.</italic></source> <volume>18</volume> <fpage>85</fpage>&#x2013;<lpage>103</lpage>.</citation></ref>
<ref id="B16"><citation citation-type="journal"><person-group person-group-type="author"><name><surname>Mishkin</surname> <given-names>M.</given-names></name> <name><surname>Ungerleider</surname> <given-names>L. G.</given-names></name> <name><surname>Macko</surname> <given-names>K. A.</given-names></name></person-group> (<year>1983</year>). <article-title>Object vision and spatial vision: two cortical pathways.</article-title> <source><italic>Trends Neurosci.</italic></source> <volume>6</volume> <fpage>414</fpage>&#x2013;<lpage>417</lpage>.</citation></ref>
<ref id="B17"><citation citation-type="journal"><person-group person-group-type="author"><name><surname>Nakano</surname> <given-names>K.</given-names></name></person-group> (<year>1972</year>). <article-title>Associatron &#x2013; A model of associative memory.</article-title> <source><italic>IEEE Trans. Syst. Man Cybern.</italic></source> <volume>SMC-2</volume>, <fpage>380</fpage>&#x2013;<lpage>388</lpage>. <pub-id pub-id-type="doi">10.1364/AO.28.000291</pub-id> <pub-id pub-id-type="pmid">20548471</pub-id></citation></ref>
<ref id="B18"><citation citation-type="journal"><person-group person-group-type="author"><name><surname>Rumelhart</surname> <given-names>D.</given-names></name> <name><surname>Hinton</surname> <given-names>G.</given-names></name> <name><surname>Williams</surname> <given-names>R.</given-names></name></person-group> (<year>1986</year>). <article-title>Learning representations by back-propagating errors.</article-title> <source><italic>Nature</italic></source> <volume>323</volume> <fpage>533</fpage>&#x2013;<lpage>536</lpage>.</citation></ref>
<ref id="B19"><citation citation-type="journal"><person-group person-group-type="author"><name><surname>Russakovsky</surname> <given-names>O.</given-names></name> <name><surname>Deng</surname> <given-names>J.</given-names></name> <name><surname>Su</surname> <given-names>H.</given-names></name> <name><surname>Krause</surname> <given-names>J.</given-names></name> <name><surname>Satheesh</surname> <given-names>S.</given-names></name> <name><surname>Ma</surname> <given-names>S.</given-names></name><etal/></person-group> (<year>2015</year>). <article-title>ImageNet large scale visual recognition challenge</article-title>. <source><italic>Int. J. Comput. Vis.</italic></source> <volume>115</volume>, <fpage>211</fpage>&#x2013;<lpage>252</lpage>.</citation></ref>
<ref id="B20"><citation citation-type="journal"><person-group person-group-type="author"><name><surname>Simonyan</surname> <given-names>K.</given-names></name> <name><surname>Zisserman</surname> <given-names>A.</given-names></name></person-group> (<year>2014</year>). <source><italic>Very Deep Convolutional Networks for Large-Scale Image Recognition.</italic></source> Available online at: <ext-link ext-link-type="uri" xlink:href="https://arxiv.org/abs/1409.1556">https://arxiv.org/abs/1409.1556</ext-link> <comment>(accessed June 6, 2018)</comment>.</citation></ref>
<ref id="B21"><citation citation-type="journal"><person-group person-group-type="author"><name><surname>Spoerer</surname> <given-names>C. J.</given-names></name> <name><surname>McClure</surname> <given-names>P.</given-names></name> <name><surname>Kriegeskorte</surname> <given-names>N.</given-names></name></person-group> (<year>2017</year>). <article-title>Recurrent convolutional neural networks: a better model of biological object recognition.</article-title> <source><italic>Front. Psychol.</italic></source> <volume>8</volume>:<issue>1551</issue>. <pub-id pub-id-type="doi">10.3389/fpsyg.2017.01551</pub-id> <pub-id pub-id-type="pmid">28955272</pub-id></citation></ref>
<ref id="B22"><citation citation-type="journal"><person-group person-group-type="author"><name><surname>Storkey</surname> <given-names>A.</given-names></name></person-group> (<year>1997</year>). &#x201C;<article-title>Increasing the capacity of a hopfield network without sacrificing functionality,</article-title>&#x201D; in <source><italic>Artificial Neural Networks &#x2014; ICANN&#x2019;97. ICANN 1997. Lecture Notes in Computer Science</italic></source>, <volume>Vol. 1327</volume>, <role>eds</role> <person-group person-group-type="editor"><name><surname>Gerstner</surname> <given-names>W.</given-names></name> <name><surname>Germond</surname> <given-names>A.</given-names></name> <name><surname>Hasler</surname> <given-names>M.</given-names></name> <name><surname>Nicoud</surname> <given-names>J. D.</given-names></name></person-group> (<publisher-loc>Berlin</publisher-loc>: <publisher-name>Springer</publisher-name>).</citation></ref>
<ref id="B23"><citation citation-type="journal"><person-group person-group-type="author"><name><surname>Sugase</surname> <given-names>Y.</given-names></name> <name><surname>Yamane</surname> <given-names>S.</given-names></name> <name><surname>Ueno</surname> <given-names>S.</given-names></name> <name><surname>Kawano</surname> <given-names>K.</given-names></name></person-group> (<year>1999</year>). <article-title>Global and fine information coded by single neurons in the temporal visual cortex.</article-title> <source><italic>Nature</italic></source> <volume>400</volume> <fpage>869</fpage>&#x2013;<lpage>873</lpage>. <pub-id pub-id-type="doi">10.1038/23703</pub-id> <pub-id pub-id-type="pmid">10476965</pub-id></citation></ref>
<ref id="B24"><citation citation-type="journal"><person-group person-group-type="author"><name><surname>Sugase-Miyamoto</surname> <given-names>Y.</given-names></name> <name><surname>Liu</surname> <given-names>Z.</given-names></name> <name><surname>Wiener</surname> <given-names>M. C.</given-names></name> <name><surname>Optican</surname> <given-names>L. M.</given-names></name> <name><surname>Richmond</surname> <given-names>B. J.</given-names></name></person-group> (<year>2008</year>). <article-title>Short-term memory trace in rapidly adapting synapses of inferior temporal cortex</article-title>. <source><italic>PLoS Comput. Biol.</italic></source> <volume>4</volume>:<issue>e1000073</issue>. <pub-id pub-id-type="doi">10.1371/journal.pcbi.1000073</pub-id> <pub-id pub-id-type="pmid">18464917</pub-id></citation></ref>
<ref id="B25"><citation citation-type="journal"><person-group person-group-type="author"><name><surname>Sugase-Miyamoto</surname> <given-names>Y.</given-names></name> <name><surname>Matsumoto</surname> <given-names>N.</given-names></name> <name><surname>Ohyama</surname> <given-names>K.</given-names></name> <name><surname>Kawano</surname> <given-names>K.</given-names></name></person-group> (<year>2014</year>). <article-title>Face inversion decreased information about facial identity and expression in face-responsive neurons in macaque area TE.</article-title> <source><italic>J. Neurosci.</italic></source> <volume>34</volume> <fpage>12457</fpage>&#x2013;<lpage>12469</lpage>. <pub-id pub-id-type="doi">10.1523/JNEUROSCI.0485-14.2014</pub-id> <pub-id pub-id-type="pmid">25209284</pub-id></citation></ref>
</ref-list>
</back>
</article>