<?xml version="1.0" encoding="UTF-8" standalone="no"?>
<!DOCTYPE article PUBLIC "-//NLM//DTD Journal Publishing DTD v2.3 20070202//EN" "journalpublishing.dtd">
<article xmlns:mml="http://www.w3.org/1998/Math/MathML" xmlns:xlink="http://www.w3.org/1999/xlink" article-type="editorial">
<front>
<journal-meta>
<journal-id journal-id-type="publisher-id">Front. Neurosci.</journal-id>
<journal-title>Frontiers in Neuroscience</journal-title>
<abbrev-journal-title abbrev-type="pubmed">Front. Neurosci.</abbrev-journal-title>
<issn pub-type="epub">1662-453X</issn>
<publisher>
<publisher-name>Frontiers Media S.A.</publisher-name>
</publisher>
</journal-meta>
<article-meta>
<article-id pub-id-type="doi">10.3389/fnins.2019.00978</article-id>
<article-categories>
<subj-group subj-group-type="heading">
<subject>Neuroscience</subject>
<subj-group>
<subject>Editorial</subject>
</subj-group>
</subj-group>
</article-categories>
<title-group>
<article-title>Editorial: Bio-inspired Audio Processing, Models and Systems</article-title>
</title-group>
<contrib-group>
<contrib contrib-type="author">
<name><surname>Liu</surname> <given-names>Shih-Chii</given-names></name>
<xref ref-type="aff" rid="aff1"><sup>1</sup></xref>
<uri xlink:href="http://loop.frontiersin.org/people/14463/overview"/>
</contrib>
<contrib contrib-type="author">
<name><surname>Harris</surname> <given-names>John G.</given-names></name>
<xref ref-type="aff" rid="aff2"><sup>2</sup></xref>
<uri xlink:href="http://loop.frontiersin.org/people/14707/overview"/>
</contrib>
<contrib contrib-type="author">
<name><surname>Elhilali</surname> <given-names>Mounya</given-names></name>
<xref ref-type="aff" rid="aff3"><sup>3</sup></xref>
<uri xlink:href="http://loop.frontiersin.org/people/52170/overview"/>
</contrib>
<contrib contrib-type="author" corresp="yes">
<name><surname>Slaney</surname> <given-names>Malcolm</given-names></name>
<xref ref-type="aff" rid="aff4"><sup>4</sup></xref>
<xref ref-type="corresp" rid="c001"><sup>&#x0002A;</sup></xref>
<uri xlink:href="http://loop.frontiersin.org/people/416440/overview"/>
</contrib>
</contrib-group>
<aff id="aff1"><sup>1</sup><institution>Institute of Neuroinformatics, University of Zurich and ETH Zurich</institution>, <addr-line>Zurich</addr-line>, <country>Switzerland</country></aff>
<aff id="aff2"><sup>2</sup><institution>Department of Electrical &#x00026; Computer Engineering, University of Florida</institution>, <addr-line>Gainesville, FL</addr-line>, <country>United States</country></aff>
<aff id="aff3"><sup>3</sup><institution>Department of Electrical and Computer Engineering, Whiting School of Engineering, Johns Hopkins University</institution>, <addr-line>Baltimore, MD</addr-line>, <country>United States</country></aff>
<aff id="aff4"><sup>4</sup><institution>Google</institution>, <addr-line>Mountain View, CA</addr-line>, <country>United States</country></aff>
<author-notes>
<fn fn-type="edited-by"><p>Edited by: Timothy K. Horiuchi, University of Maryland, College Park, United States</p></fn>
<fn fn-type="edited-by"><p>Reviewed by: Soumyajit Mandal, Case Western Reserve University, United States; Chetan Singh Thakur, Indian Institute of Science (IISc), India</p></fn>
<corresp id="c001">&#x0002A;Correspondence: Malcolm Slaney <email>malcolm&#x00040;ieee.org</email></corresp>
<fn fn-type="other" id="fn001"><p>This article was submitted to Neuromorphic Engineering, a section of the journal Frontiers in Neuroscience</p></fn></author-notes>
<pub-date pub-type="epub">
<day>13</day>
<month>09</month>
<year>2019</year>
</pub-date>
<pub-date pub-type="collection">
<year>2019</year>
</pub-date>
<volume>13</volume>
<elocation-id>978</elocation-id>
<history>
<date date-type="received">
<day>02</day>
<month>06</month>
<year>2019</year>
</date>
<date date-type="accepted">
<day>30</day>
<month>08</month>
<year>2019</year>
</date>
</history>
<permissions>
<copyright-statement>Copyright &#x000A9; 2019 Liu, Harris, Elhilali and Slaney.</copyright-statement>
<copyright-year>2019</copyright-year>
<copyright-holder>Liu, Harris, Elhilali and Slaney</copyright-holder>
<license xlink:href="http://creativecommons.org/licenses/by/4.0/"><p>This is an open-access article distributed under the terms of the Creative Commons Attribution License (CC BY). The use, distribution or reproduction in other forums is permitted, provided the original author(s) and the copyright owner(s) are credited and that the original publication in this journal is cited, in accordance with accepted academic practice. No use, distribution or reproduction is permitted which does not comply with these terms.</p></license>
</permissions>
<related-article id="RA1" related-article-type="commentary-article" xlink:href="https://www.frontiersin.org/research-topics/5930/bio-inspired-audio-processing-models-and-systems" ext-link-type="uri">Editorial on the Research Topic <article-title>Bio-inspired Audio Processing, Models and Systems</article-title></related-article>
<kwd-group>
<kwd>neuromorphic audio sensors</kwd>
<kwd>spatial hearing</kwd>
<kwd>auditory perception</kwd>
<kwd>inter-aural time delay</kwd>
<kwd>spike-based computation</kwd>
<kwd>spatial-temporal response fields</kwd>
<kwd>attention decoding</kwd>
<kwd>saliency</kwd>
</kwd-group>
<counts>
<fig-count count="0"/>
<table-count count="0"/>
<equation-count count="0"/>
<ref-count count="0"/>
<page-count count="2"/>
<word-count count="1151"/>
</counts>
</article-meta> 
</front>
<body>
<sec sec-type="intro" id="s1">
<title>Introduction</title>
<p>Bio-inspired systems look at biology to inspire engineering solutions that help explain, emulate and complement the intricate processes that take place in a biological system. As such, they operate at the intersection of biology and engineering and leverage advantages from both disciplines. When applied to brain sciences, bio-inspired systems often use non-conventional approaches to solve complex sensory and cognitive tasks.</p>
<p>Recent developments in sensor design, algorithmic configurations, and network-level processing show the promise and efficacy of brain-like systems in solving complex tasks. While vision systems are widely explored in neuromorphic engineering design, audio systems offer unique challenges. These include careful handling of the time and space dimensions, issues related to temporal sampling and signal representation in both time and frequency, leveraging the redundancy in audio signals for complex detection and recognition tasks, as well as robust processing against noise and other interferers and maskers.</p>
<p>Our auditory systems have evolved highly efficient solutions to audio scene analysis, spatial understanding, and sound recognition. We wish to better understand the biological solutions that allow the brain to process sounds in unknown and highly distorted conditions; in order to help advance state-of-art audio systems that often operate well under well-controlled environments but fail to generalize, adapt and efficiently process unknown conditions. Furthermore, we want to apply engineering methods to better understand biological processes, using non-invasive methods. By leveraging both our knowledge of the biology in building better systems, as well as new technological advantages to unravel secrets of the brain, we hope to enrich the conversation across both disciplines in order to advance our understanding of the brain function and help improve technologies that impact our lives in a wide range of domains.</p>
</sec>
<sec id="s2">
<title>Overview</title>
<p>This special topic issue describes the latest advances in research on sensors, models, networks, and hardware for audio processing, hearing systems, and speech technologies. Broadly speaking, the papers in this special issue fall into four broad classes:
<list list-type="order">
<list-item><p>Bio-inspired implementations</p></list-item>
<list-item><p>Models based on spikes</p></list-item>
<list-item><p>Sound recognition</p></list-item>
<list-item><p>Attention decoding.</p></list-item>
</list></p>
<p>Bio-inspired systems often start with hardware designed to mimic and/or capitalize on the advantages of biological systems. With regards to processing acoustic cues, a paper by <ext-link ext-link-type="uri" xlink:href="https://doi.org/10.3389/fnins.2018.00198">Xu et al.</ext-link> describes a digital hardware FPGA implementation of a well-known CAR-FAC cochlear model that mimics the auditory physiology seen in the biological cochlea. Similarly, our auditory system is exquisitely sensitive to the differences in signals received between the two ears. The paper by <ext-link ext-link-type="uri" xlink:href="https://doi.org/10.3389/fnins.2018.00177">Isbell and Horiuchi</ext-link> explores how the auditory system might change the timing of pulses in an echo-location system. Finally a paper by <ext-link ext-link-type="uri" xlink:href="https://doi.org/10.3389/fnins.2018.00140">Encke and Hemmert</ext-link> introduces a spiking neuron model based on recent physiological findings in mammals for the detection of interaural time differences for sound localization.</p>
<p>The most obvious difference between conventional solutions to auditory processing and biological systems is the way that our biology depends on discrete spikes to represent the sensory signal. Toward this end, papers by <ext-link ext-link-type="uri" xlink:href="https://doi.org/10.3389/fnins.2018.00023">Anumula et al.</ext-link> and <ext-link ext-link-type="uri" xlink:href="https://doi.org/10.3389/fnins.2018.00160">Acharya et al.</ext-link> investigate different ways to represent the spiking information in ways amenable to conventional machine-learning methods. The paper by <ext-link ext-link-type="uri" xlink:href="https://doi.org/10.3389/fnins.2018.00836">Wu et al.</ext-link> takes these approaches to feature discovery a step further by using a self-organizing network to design the best feature representation. Then, the paper by <ext-link ext-link-type="uri" xlink:href="https://doi.org/10.3389/fnins.2018.00194">Li and Pr&#x000ED;ncipe</ext-link> looks at ways to extend the temporal information using kernel methods that can choose the optimal representation.</p>
<p>An important task for the auditory system is to understand and identify the sounds around us. The paper by <ext-link ext-link-type="uri" xlink:href="https://doi.org/10.3389/fnins.2017.00485">McWalter and Dau</ext-link> considers high-level features that combine information across time and frequency for synthesizing and perceiving auditory textures. A paper by <ext-link ext-link-type="uri" xlink:href="https://doi.org/10.3389/fnins.2018.00349">Zuk et al.</ext-link> looks at how we perceive musical beats, comparing the information from bottom-up (sensory) processes vs. top-down (cognitive) expectations. Finally a paper by <ext-link ext-link-type="uri" xlink:href="https://doi.org/10.3389/fnins.2018.00532">Huang et al.</ext-link> looks at ways to build models of what makes a sound salient in its environment.</p>
<p>To conclude this special issue, much effort recently has gone toward finding methods that allow us to monitor the attention of a user. In the visual world, the eyes provide an important clue, but no such obvious signal exists for the auditory world. The paper by <ext-link ext-link-type="uri" xlink:href="https://doi.org/10.3389/fnins.2019.00153">Alickovic et al.</ext-link> summarizes several approaches based on regression and correlation analysis that allow us to match the audio signal and brain&#x00027;s response. <ext-link ext-link-type="uri" xlink:href="https://doi.org/10.3389/fnins.2018.00531">Wong et al.&#x00027;s</ext-link> paper adds further details on regularization methods for regression-based methods, which are needed to make the computations stable. To put it all together, a paper by <ext-link ext-link-type="uri" xlink:href="https://doi.org/10.3389/fnins.2018.00262">Miran et al.</ext-link> builds an end-to-end solution that considers the statistics of the input signal and the output decision to build an optimal decoder of a user&#x00027;s attentional state.</p>
<p>We hope you find these 13 papers illuminating. They represent the state of the art in bio-inspired audio-processing models and systems.</p>
</sec>
<sec id="s3">
<title>Author Contributions</title>
<p>This editorial was written and edited by MS, S-CL, ME, and JH.</p>
<sec>
<title>Conflict of Interest Statement</title>
<p>The authors declare that the research was conducted in the absence of any commercial or financial relationships that could be construed as a potential conflict of interest.</p>
</sec>
</sec>
</body>
</article>