<?xml version="1.0" encoding="UTF-8"?>
<!DOCTYPE article PUBLIC "-//NLM//DTD Journal Publishing DTD v2.3 20070202//EN" "journalpublishing.dtd">
<article article-type="research-article" dtd-version="2.3" xml:lang="EN" xmlns:mml="http://www.w3.org/1998/Math/MathML" xmlns:xlink="http://www.w3.org/1999/xlink">
<front>
<journal-meta>
<journal-id journal-id-type="publisher-id">Front. Comput. Sci.</journal-id>
<journal-title>Frontiers in Computer Science</journal-title>
<abbrev-journal-title abbrev-type="pubmed">Front. Comput. Sci.</abbrev-journal-title>
<issn pub-type="epub">2624-9898</issn>
<publisher>
<publisher-name>Frontiers Media S.A.</publisher-name>
</publisher>
</journal-meta>
<article-meta>
<article-id pub-id-type="publisher-id">613981</article-id>
<article-id pub-id-type="doi">10.3389/fcomp.2021.613981</article-id>
<article-categories>
<subj-group subj-group-type="heading">
<subject>Computer Science</subject>
<subj-group>
<subject>Original Research</subject>
</subj-group>
</subj-group>
</article-categories>
<title-group>
<article-title>FusionNet: A Deep Fully Residual Convolutional Neural Network for Image Segmentation in Connectomics</article-title>
<alt-title alt-title-type="left-running-head">Quan et&#x20;al.</alt-title>
<alt-title alt-title-type="right-running-head">FusionNet: Image Segmentation in Connectomics</alt-title>
</title-group>
<contrib-group>
<contrib contrib-type="author">
<name>
<surname>Quan</surname>
<given-names>Tran Minh</given-names>
</name>
<xref ref-type="aff" rid="aff1">
<sup>1</sup>
</xref>
<uri xlink:href="https://loop.frontiersin.org/people/1321140/overview"/>
</contrib>
<contrib contrib-type="author">
<name>
<surname>Hildebrand</surname>
<given-names>David Grant Colburn</given-names>
</name>
<xref ref-type="aff" rid="aff2">
<sup>2</sup>
</xref>
<uri xlink:href="https://loop.frontiersin.org/people/104099/overview"/>
</contrib>
<contrib contrib-type="author" corresp="yes">
<name>
<surname>Jeong</surname>
<given-names>Won-Ki</given-names>
</name>
<xref ref-type="aff" rid="aff3">
<sup>3</sup>
</xref>
<xref ref-type="corresp" rid="c001">&#x2a;</xref>
<uri xlink:href="https://loop.frontiersin.org/people/1104888/overview"/>
</contrib>
</contrib-group>
<aff id="aff1">
<label>
<sup>1</sup>
</label>College of Engineering and Computer Science, VinUniversity, <addr-line>Ha Noi</addr-line>, <country>Vietnam</country>
</aff>
<aff id="aff2">
<label>
<sup>2</sup>
</label>Laboratory of Neural Systems, The Rockefeller University, <addr-line>New York</addr-line>, <addr-line>NY</addr-line>, <country>United&#x20;States</country>
</aff>
<aff id="aff3">
<label>
<sup>3</sup>
</label>Department of Computer Science and Engineering, Korea University, <addr-line>Seoul</addr-line>, <country>South Korea</country>
</aff>
<author-notes>
<fn fn-type="edited-by">
<p>
<bold>Edited by:</bold> <ext-link ext-link-type="uri" xlink:href="https://loop.frontiersin.org/people/670193/overview">Francesco Rundo</ext-link>, STMicroelectronics, Italy</p>
</fn>
<fn fn-type="edited-by">
<p>
<bold>Reviewed by:</bold> <ext-link ext-link-type="uri" xlink:href="https://loop.frontiersin.org/people/1237814/overview">Francesco Guarnera</ext-link>, University of Catania, Italy</p>
<p>
<ext-link ext-link-type="uri" xlink:href="https://loop.frontiersin.org/people/920878/overview">Francesca Trenta</ext-link>, University of Catania, Italy</p>
</fn>
<corresp id="c001">&#x2a;Correspondence: Won-Ki Jeong, <email>wkjeong@korea.ac.kr</email>
</corresp>
<fn fn-type="other">
<p>This article was submitted to Computer Vision, a section of the journal Frontiers in Computer Science</p>
</fn>
</author-notes>
<pub-date pub-type="epub">
<day>13</day>
<month>05</month>
<year>2021</year>
</pub-date>
<pub-date pub-type="collection">
<year>2021</year>
</pub-date>
<volume>3</volume>
<elocation-id>613981</elocation-id>
<history>
<date date-type="received">
<day>04</day>
<month>10</month>
<year>2020</year>
</date>
<date date-type="accepted">
<day>12</day>
<month>04</month>
<year>2021</year>
</date>
</history>
<permissions>
<copyright-statement>Copyright &#xa9; 2021 Quan, Hildebrand and Jeong.</copyright-statement>
<copyright-year>2021</copyright-year>
<copyright-holder>Quan, Hildebrand and Jeong</copyright-holder>
<license xlink:href="http://creativecommons.org/licenses/by/4.0/">
<p>This is an open-access article distributed under the terms of the Creative Commons Attribution License (CC BY). The use, distribution or reproduction in other forums is permitted, provided the original author(s) and the copyright owner(s) are credited and that the original publication in this journal is cited, in accordance with accepted academic practice. No use, distribution or reproduction is permitted which does not comply with these&#x20;terms.</p>
</license>
</permissions>
<abstract>
<p>Cellular-resolution connectomics is an ambitious research direction with the goal of generating comprehensive brain connectivity maps using high-throughput, nano-scale electron microscopy. One of the main challenges in connectomics research is developing scalable image analysis algorithms that require minimal user intervention. Deep learning has provided exceptional performance in image classification tasks in computer vision, leading to a recent explosion in popularity. Similarly, its application to connectomic analyses holds great promise. Here, we introduce a deep neural network architecture, FusionNet, with a focus on its application to accomplish automatic segmentation of neuronal structures in connectomics data. FusionNet combines recent advances in machine learning, such as semantic segmentation and residual neural networks, with summation-based skip connections. This results in a much deeper network architecture and improves segmentation accuracy. We demonstrate the performance of the proposed method by comparing it with several other popular electron microscopy segmentation methods. We further illustrate its flexibility through segmentation results for two different tasks: cell membrane segmentation and cell nucleus segmentation.</p>
</abstract>
<kwd-group>
<kwd>connectomic analysis</kwd>
<kwd>image segementation</kwd>
<kwd>deep learning</kwd>
<kwd>refinement</kwd>
<kwd>skip connection</kwd>
</kwd-group>
<contract-num rid="cn001">NRF-2017M3C7A1047904 NRF-2019M3E5D2A01063819 HI18C0316 IITP-2021-0-01819</contract-num>
<contract-sponsor id="cn001">National Research Foundation of Korea<named-content content-type="fundref-id">10.13039/501100003725</named-content>
</contract-sponsor>
</article-meta>
</front>
<body>
<sec id="s1">
<title>1 Introduction</title>
<p>The brain is considered the most complex organ in the human body. Despite decades of intense research, our understanding of how its structure relates to its function remains limited (<xref ref-type="bibr" rid="B34">Lichtman and Denk, 2011</xref>). Connectomics research seeks to disentangle the complicated neuronal circuits embedded within the brain. This field has gained substantial attention recently thanks to the advent of new serial-section electron microscopy (EM) technologies (<xref ref-type="bibr" rid="B5">Briggman and Bock, 2012</xref>; <xref ref-type="bibr" rid="B22">Hayworth et&#x20;al., 2014</xref>; <xref ref-type="bibr" rid="B15">Eberle and Zeidler, 2018</xref>; <xref ref-type="bibr" rid="B59">Zheng et&#x20;al., 2018</xref>; <xref ref-type="bibr" rid="B21">Graham et&#x20;al., 2019</xref>). The resolution afforded by EM is sufficient for resolving tiny but important neuronal structures that are often densely packed together, such as dendritic spine necks and synaptic vesicles. These structures are often only tens of nanometers in diameter (<xref ref-type="bibr" rid="B24">Helmstaedter, 2013</xref>). <xref ref-type="fig" rid="F1">Figure&#x20;1</xref> shows an example of such an EM image and its cell membrane segmentation. Such high-resolution imaging results in enormous datasets, approaching one petabyte for only the relatively small tissue volume of one cubic millimeter. Therefore, handling and analyzing EM datasets is one of the most challenging problems in connectomics.</p>
<fig id="F1" position="float">
<label>FIGURE 1</label>
<caption>
<p>An example EM image <bold>(left)</bold> and its manually extracted cellular membrane segmentation result <bold>(right)</bold> from the ISBI 2012 EM segmentation challenge (<xref ref-type="bibr" rid="B1">Arganda-Carreras et&#x20;al., 2015</xref>). Scale bar (green): 500&#xa0;nm.</p>
</caption>
<graphic xlink:href="fcomp-03-613981-g001.tif"/>
</fig>
<p>Early connectomics research focused on the sparse reconstruction of neuronal circuits (<xref ref-type="bibr" rid="B4">Bock et&#x20;al., 2011</xref>; <xref ref-type="bibr" rid="B6">Briggman et&#x20;al., 2011</xref>), meaning they focused reconstruction efforts on a subset of neurons in the data using manual or semi-automatic tools (<xref ref-type="bibr" rid="B29">Jeong et&#x20;al., 2010</xref>; <xref ref-type="bibr" rid="B47">Sommer et&#x20;al., 2011</xref>; <xref ref-type="bibr" rid="B8">Cardona et&#x20;al., 2012</xref>). Unfortunately, this approach requires too much human interaction to scale well over the vast amount of EM data that can be collected with new technologies. Therefore, developing scalable and automatic image analysis algorithms is an important and active research direction in the field of connectomics.</p>
<p>Although some EM image processing pipelines use conventional, light-weight pixel classifiers [e.g., RhoANA (<xref ref-type="bibr" rid="B30">Kaynig et&#x20;al., 2015</xref>)], the majority of automatic image segmentation algorithms for connectomics rely on deep learning. Earlier automatic segmentation work using deep learning mainly focused on patch-based pixel-wise classification based on a convolutional neural network (CNN) for affinity map generation (<xref ref-type="bibr" rid="B50">Turaga et&#x20;al., 2010</xref>) and cell membrane probability estimation (<xref ref-type="bibr" rid="B13">Ciresan et&#x20;al., 2012</xref>). However, one limitation of applying a conventional CNN to EM image segmentation is that per-pixel network deployment scaling becomes prohibitively expensive considering the tera-scale to peta-scale EM data size. For this reason, more efficient and scalable deep neural networks are important for image segmentation of the large datasets that can now be produced. One approach is to extend a fully convolutional neural network (FCN) (<xref ref-type="bibr" rid="B36">Long et&#x20;al., 2015</xref>), which uses encoding and decoding phases similar to an autoencoder for the end-to-end semantic segmentation problem (<xref ref-type="bibr" rid="B41">Ronneberger et&#x20;al., 2015</xref>; <xref ref-type="bibr" rid="B9">Chen et&#x20;al., 2016a</xref>).</p>
<p>The motivation of the proposed work stems from our recent research effort to develop a deeper neural network for end-to-end cell segmentation with higher accuracy. We observed that, like conventional CNNs, a popular deep neural network for end-to-end segmentation known as U-net (<xref ref-type="bibr" rid="B41">Ronneberger et&#x20;al., 2015</xref>) is limited by gradient vanishing with increasing network depth. To address this problem, we propose two extensions of U-net: using residual layers in each level of the network and introducing summation-based skip connections to make the entire network much deeper. Our segmentation method produces an accurate result that is competitive with similar EM segmentation methods. The main contribution of this study can be summarized as follows:<list list-type="bullet">
<list-item>
<p>We introduce an end-to-end automatic EM image segmentation method using deep learning. The proposed method combines a variant of U-net and residual CNN with novel summation-based skip connections to make the proposed architecture, a <italic>fully residual deep CNN</italic>. This new architecture directly employs residual properties within and across levels, thus providing a <italic>deeper</italic> network with higher accuracy.</p>
</list-item>
<list-item>
<p>We demonstrate the performance of the proposed deep learning architecture by comparing it with several EM segmentation methods listed in the leader board of the ISBI 2012 EM segmentation challenge (<xref ref-type="bibr" rid="B1">Arganda-Carreras et&#x20;al., 2015</xref>). Our method outperformed many of the top-ranked methods in terms of segmentation accuracy.</p>
</list-item>
<list-item>
<p>We introduce a <italic>data enrichment</italic> method specifically built for EM data by collecting all the orientation variants of the input images (eight in the 2D case, including all combinations of flipping and rotation). We used the same augmentation process for deployment: the final output is a combination of eight different probability values, which increases the accuracy of the method.</p>
</list-item>
<list-item>
<p>We demonstrate the flexibility of the proposed method on two different EM segmentation tasks. The first involves cell membrane segmentation on a fruit fly (<italic>Drosophila</italic>) EM dataset (<xref ref-type="bibr" rid="B1">Arganda-Carreras et&#x20;al., 2015</xref>). The second involves cell nucleus feature segmentation on a whole-brain larval zebrafish EM dataset (<xref ref-type="bibr" rid="B26">Hildebrand et&#x20;al., 2017</xref>).</p>
</list-item>
</list>
</p>
</sec>
<sec id="s2">
<title>2 Related Work</title>
<p>Deep neural networks (<xref ref-type="bibr" rid="B32">LeCun et&#x20;al., 2015</xref>) have surpassed human performance in solving many complex visual recognition problems. Systems using this method can flexibly learn to recognize patterns such as handwritten digits (<xref ref-type="bibr" rid="B31">Krizhevsky et&#x20;al., 2012</xref>) in images with increasing layers hierarchically corresponding to increasing feature complexity (<xref ref-type="bibr" rid="B57">Zeiler and Fergus, 2014</xref>). A major drawback of using deep neural networks is that they often require a huge amount of training data. In order to overcome this issue, researchers have started to collect large databases containing millions of images that span hundreds of categories (<xref ref-type="bibr" rid="B42">Russakovsky et&#x20;al., 2015</xref>). Largely thanks to such training datasets, many advanced architectures have been introduced, including VGG (<xref ref-type="bibr" rid="B46">Simonyan and Zisserman, 2014</xref>) and GoogleNet (<xref ref-type="bibr" rid="B49">Szegedy et&#x20;al., 2015</xref>). With these architectures, computers are now able to perform even more complex tasks, such as transferring artistic styles from a source image to an unrelated target (<xref ref-type="bibr" rid="B20">Gatys et&#x20;al., 2016</xref>). To leverage these new capabilities, researchers are actively working to extend deep learning methods for analyzing biomedical image data (<xref ref-type="bibr" rid="B12">Cicek et&#x20;al., 2016</xref>). Developing such methods for automatic classification and segmentation of different biomedical image modalities, such as CT (<xref ref-type="bibr" rid="B58">Zheng et&#x20;al., 2015</xref>) and MRI (<xref ref-type="bibr" rid="B28">Isin et&#x20;al., 2016</xref>), is leading to faster and more accurate decision-making processes in laboratory and clinical settings.</p>
<p>Similarly, deep learning has been quickly adopted by connectomics researchers to enhance automatic EM image segmentation. One of the earliest applications to EM segmentation involved the straightforward application of a convolutional neural network (CNN) for pixel-wise membrane probability estimation (<xref ref-type="bibr" rid="B13">Ciresan et&#x20;al., 2012</xref>), an approach that won the ISBI 2012 EM segmentation challenge (<xref ref-type="bibr" rid="B1">Arganda-Carreras et&#x20;al., 2015</xref>). As more deep learning methods are introduced, automatic EM segmentation techniques evolve and new groups overtake the title of state-of-the-art performance in such challenges. One notable recent advancement was the introduction of a fully convolutional neural network (FCN) (<xref ref-type="bibr" rid="B36">Long et&#x20;al., 2015</xref>) for end-to-end semantic segmentation. Inspired by this work, several modified FCNs have been proposed for EM image segmentation. One variant combined multi-level upscaling layers to produce a final segmentation (<xref ref-type="bibr" rid="B9">Chen et&#x20;al., 2016a</xref>). Additional post-processing steps such as lifted multi-cut (<xref ref-type="bibr" rid="B2">Beier et&#x20;al., 2016</xref>; <xref ref-type="bibr" rid="B37">Pape et&#x20;al., 2019</xref>) further refined this segmentation result.</p>
<p>Another approach added skip connections for concatenating feature maps into a &#x201c;U-net&#x201d; architecture (<xref ref-type="bibr" rid="B41">Ronneberger et&#x20;al., 2015</xref>). While U-net and its variants can learn multi-contextual information from input data, they are limited in the depth of the network they can construct because of the vanishing gradient problem. On the other hand, the addition of shortcut connections and direction summations (<xref ref-type="bibr" rid="B23">He et&#x20;al., 2016</xref>) allows gradients to flow across multiple layers during the training phase. This creates a fully residual CNN where the architecture is a fusion of the U-net design and networks with summation-based skip connections, similar to Fully Convolutional Residual Networks (FC-ResNets) (<xref ref-type="bibr" rid="B14">Drozdzal et&#x20;al., 2016</xref>) and Residual Deconvolutional Networks (RDN) (<xref ref-type="bibr" rid="B18">Fakhry et&#x20;al., 2017</xref>). These related studies inspired us to propose a fully residual CNN for analyzing connectomics&#x20;data.</p>
<p>Work that leverages recurrent neural network (RNN) architectures can also accomplish this segmentation task (<xref ref-type="bibr" rid="B48">Stollenga et&#x20;al., 2015</xref>). Instead of simultaneously considering all surrounding pixels and computing responses for the feature maps, RNN-based networks treat the pixels as a list or sequence with various routing rules and recurrently update each feature pixel. In fact, RNN-based membrane segmentation approaches are crucial for connected component labeling steps that can resolve false splits and merges during the post-processing of probability maps (<xref ref-type="bibr" rid="B16">Ensafi et&#x20;al., 2014</xref>; <xref ref-type="bibr" rid="B38">Parag et&#x20;al., 2015</xref>).</p>
</sec>
<sec sec-type="methods" id="s3">
<title>3 Methods</title>
<sec id="s3-1">
<title>3.1 Network Architecture</title>
<p>Our proposed network, FusionNet, is based on the architecture of a convolutional autoencoder and is illustrated in <xref ref-type="fig" rid="F2">Figure&#x20;2</xref>. It consists of an encoding path (upper half, from 640&#x20;&#xd7; 640 to 40&#x20;&#xd7; 40) that retrieves features of interest and a symmetric decoding path (lower half, from 40&#x20;&#xd7; 40 to 640&#x20;&#xd7; 640) that accumulates the feature maps from different scales to form the segmentation. Both the encoding and decoding paths consist of multiple levels (i.e.,&#x20;resolutions). Four basic building blocks are used to construct the proposed network. Each <italic>green block</italic> is a regular convolutional layer followed by rectified linear unit activation and batch normalization (omitted from the figure for simplicity). Each <italic>violet block</italic> is a residual layer that consists of three convolutional blocks and a residual skip connection. Each <italic>blue block</italic> is a maxpooling layer located between levels only in the encoding path to perform downsampling for feature compression. Each <italic>red block</italic> is a deconvolutional layer located between levels only in the decoding path to upsample the input data using learnable interpolations. A detailed specification of FusionNet, including the number of feature maps and their sizes, is provided in <xref ref-type="table" rid="T1">Table&#x20;1</xref>.</p>
<fig id="F2" position="float">
<label>FIGURE 2</label>
<caption>
<p>The proposed FusionNet architecture. An illustration of the encoding path <bold>(top to middle)</bold> and the decoding path <bold>(middle to bottom)</bold>. Each intermediate residual block contains a residual skip connection within the same path, while the nested residual skip connections connect two different&#x20;paths.</p>
</caption>
<graphic xlink:href="fcomp-03-613981-g002.tif"/>
</fig>
<table-wrap id="T1" position="float">
<label>TABLE 1</label>
<caption>
<p>Architecture of the proposed network.</p>
</caption>
<table>
<thead valign="top">
<tr>
<th align="left">Block type</th>
<th align="center">Ingredients</th>
<th align="center">Size of feature maps</th>
</tr>
</thead>
<tbody valign="top">
<tr>
<td align="left">input</td>
<td align="left"/>
<td align="center">640 &#xd7; 640 &#xd7; 1</td>
</tr>
<tr>
<td rowspan="2" align="left">down 1</td>
<td align="left">conv &#x2b; res &#x2b; conv</td>
<td align="center">640 &#xd7; 640 &#xd7; 64</td>
</tr>
<tr>
<td align="left">&#x2b; maxpooling</td>
<td align="center">320 &#xd7; 320 &#xd7; 64</td>
</tr>
<tr>
<td rowspan="2" align="left">down 2</td>
<td align="left">conv &#x2b; res &#x2b; conv</td>
<td align="center">320 &#xd7; 320 &#xd7; 128</td>
</tr>
<tr>
<td align="left">&#x2b; maxpooling</td>
<td align="center">160 &#xd7; 160 &#xd7; 128</td>
</tr>
<tr>
<td rowspan="2" align="left">down 3</td>
<td align="left">conv &#x2b; res &#x2b; conv</td>
<td align="center">160 &#xd7; 160 &#xd7; 256</td>
</tr>
<tr>
<td align="left">&#x2b; maxpooling</td>
<td align="center">80 &#xd7; 80 &#xd7; 256</td>
</tr>
<tr>
<td rowspan="2" align="left">down 4</td>
<td align="left">conv &#x2b; res &#x2b; conv</td>
<td align="center">80 &#xd7; 80 &#xd7; 512</td>
</tr>
<tr>
<td align="left">&#x2b; maxpooling</td>
<td align="center">40 &#xd7; 40 &#xd7; 512</td>
</tr>
<tr>
<td align="left">bridge</td>
<td align="left">conv &#x2b; res &#x2b; conv</td>
<td align="center">40 &#xd7; 40 &#xd7; 1024</td>
</tr>
<tr>
<td rowspan="2" align="left">up 4</td>
<td align="left">deconv &#x2b; merge &#x2b;</td>
<td align="center">80 &#xd7; 80 &#xd7; 512</td>
</tr>
<tr>
<td align="left">conv &#x2b; res &#x2b; conv</td>
<td align="center">80 &#xd7; 80 &#xd7; 512</td>
</tr>
<tr>
<td rowspan="2" align="left">up 3</td>
<td align="left">deconv &#x2b; merge &#x2b;</td>
<td align="center">160 &#xd7; 160 &#xd7; 256</td>
</tr>
<tr>
<td align="left">conv &#x2b; res &#x2b; conv</td>
<td align="center">160 &#xd7; 160 &#xd7; 256</td>
</tr>
<tr>
<td rowspan="2" align="left">up 2</td>
<td align="left">deconv &#x2b; merge &#x2b;</td>
<td align="center">320 &#xd7; 320 &#xd7; 128</td>
</tr>
<tr>
<td align="left">conv &#x2b; res &#x2b; conv</td>
<td align="center">320 &#xd7; 320 &#xd7; 128</td>
</tr>
<tr>
<td rowspan="2" align="left">up 1</td>
<td align="left">deconv &#x2b; merge &#x2b;</td>
<td align="center">640 &#xd7; 640 &#xd7; 64</td>
</tr>
<tr>
<td align="left">conv &#x2b; res &#x2b; conv</td>
<td align="center">640 &#xd7; 640 &#xd7; 64</td>
</tr>
<tr>
<td align="left">output</td>
<td align="left">conv</td>
<td align="center">640 &#xd7; 640 &#xd7; 1</td>
</tr>
</tbody>
</table>
</table-wrap>
<p>One major difference between the FusionNet and U-net architectures is the way in which skip connections are used (<xref ref-type="fig" rid="F3">Figure&#x20;3</xref>). In FusionNet, each level in the decoding path begins with a deconvolutional block (red) that un-pools the feature map from a coarser level (i.e.,&#x20;resolution), then merges it by pixel-wise <italic>addition</italic> with the feature map from the corresponding level in the encoding path by a <italic>long</italic> skip connection. There is also a <italic>short</italic> skip connection contained in each residual block (violet) that serves as a direct connection from the previous layer within the same encoding or decoding path. In contrast, U-net concatenates feature maps using only long skip connections. Additionally, by replacing concatenation with addition, FusionNet becomes a <italic>fully</italic> residual network, which resolves some common issues in deep networks (i.e.,&#x20;gradient vanishing). Furthermore, the nested short and long skip connections in FusionNet permit information flow within and across levels.</p>
<fig id="F3" position="float">
<label>FIGURE 3</label>
<caption>
<p>Difference between the core connections of U-net (<xref ref-type="bibr" rid="B41">Ronneberger et&#x20;al., 2015</xref>) <bold>(left)</bold> and FusionNet <bold>(right)</bold>. Note that FusionNet is a fully residual network due to the summation-based skip connections and is a much deeper network.</p>
</caption>
<graphic xlink:href="fcomp-03-613981-g003.tif"/>
</fig>
<p>In the FusionNet encoding path, the number of feature maps doubles whenever downsampling is performed. After passing through the encoding path, the bridge level (i.e.,&#x20;40 &#xd7; 40 layer) residual block starts to expand feature maps into the following decoding path. In the decoding path, the number of feature maps is halved at every level, which maintains network symmetry. Note that there are convolutional layers both before and after each residual block. These convolutional layers serve as portal gateways that effectively adjust the amount of feature maps before and after residual blocks to the appropriate numbers. The placement of these convolutional layers on either side of the residual block leads the entire network to be perfectly symmetric (see <xref ref-type="fig" rid="F2">Figure&#x20;2</xref>).</p>
<p>FusionNet performs end-to-end segmentation from the input EM data to the output segmentation label prediction. We train the network with pairs of EM images and their corresponding manually segmented label images as input. The training process involves comparing the output prediction with the input target labels using a mean-absolute-error (MAE) loss function to back-propagate adjustments to the connection weights. We considered the network sufficiently trained when its loss function values plateaued over several hundred epochs.</p>
</sec>
<sec id="s3-2">
<title>3.2 Data Augmentation</title>
<p>Our system involves data augmentation in multiple stages during both the training and deployment phases.</p>
<p>For training:</p>
<p>
<list list-type="bullet">
<list-item>
<p>The order of the image and label pairs are shuffled and organized with three-fold cross-validation to improve the generalization of our method.</p>
</list-item>
<list-item>
<p>Offline, all training images and labels are reoriented to first produce an enriched dataset.</p>
</list-item>
<list-item>
<p>Online, elastic field deformation is applied to both images and corresponding labels, followed by noise addition to only the images.</p>
</list-item>
</list>
</p>
<p>For prediction:<list list-type="bullet">
<list-item>
<p>Offline, input images are reoriented as for training.</p>
</list-item>
<list-item>
<p>Inference is performed on all reoriented images separately, then each intermediate result is reverted to the original orientation, and all intermediate results are averaged to produce the final prediction.</p>
</list-item>
</list>
</p>
<p>Boundary extension is performed for all input images and labels. We describe each augmentation step in more detail in the following subsections.</p>
<p>
<bold>Reorienation enrichment</bold>: Different EM images typically share similar orientation-independent textures in structures such as mitochondria, axons, and synapses. We reasoned that it should therefore be possible to enrich our input data with seven additional image and label pairs by reorienting the EM images, and in the case of training, their corresponding labels. <xref ref-type="fig" rid="F4">Figure&#x20;4</xref> shows all eight orientations resulting from a single EM image after performing this data enrichment, with an overlaid letter &#x201c;g&#x201d; in each panel to provide a simpler view of the generated orientation. To generate these permutations, we rotated each EM image (and corresponding label) by 90&#xb0;, 180&#xb0;, and 270&#xb0;. We then vertically reflected the original and rotated images. For training, each orientation was added as a new image and label pair. For prediction, inference was performed on each of these data orientations separately, then each prediction result was reverted to the original orientation before averaging to produce the final accumulation. Our intuition here is that, based on the equivariance of isotropic data, each orientation will contribute equally toward the final prediction result. Note that because the image and label pairs are enriched eight times by this process, other on-the-fly linear data augmentation techniques such as random rotation, flipping, or transposition are unnecessary.</p>
<fig id="F4" position="float">
<label>FIGURE 4</label>
<caption>
<p>Eight reoriented versions of the same EM image. The original image is outlined in blue. By adding these reoriented images, the input data size is increased by eight&#x20;times.</p>
</caption>
<graphic xlink:href="fcomp-03-613981-g004.tif"/>
</fig>
<p>
<bold>Elastic field deformation</bold>: To avoid overfitting (i.e.,&#x20;network remembering the training data), elastic deformation was performed on the entire enriched image dataset for every training epoch. This strategy is common in machine learning, especially for deep networks, to overcome limitations associated with small training dataset sizes. This procedure is illustrated in <xref ref-type="fig" rid="F5">Figure&#x20;5</xref>. We first initialized a random sparse 12&#x20;&#xd7; 12 vector field whose amplitudes at the image border boundaries vanish to zero. This field was then interpolated to the input size and used to warp both EM images and corresponding labels. The flow map was randomly generated for each epoch. No elastic field deformation was performed during deployment.</p>
<fig id="F5" position="float">
<label>FIGURE 5</label>
<caption>
<p>Elastic field deformation example. A randomly sparse vector field <bold>(A)</bold> is generated for each training image and label pair. This sparse vector field is then used to warp both the original image data (<bold>B</bold>, left) and its corresponding label (<bold>C</bold>, left) to form an augmentation pair consisting of warped image data (<bold>B</bold>, middle) and warped label (<bold>C</bold>, middle). The difference between the original and warped images (<bold>B</bold>, right) and labels (<bold>C</bold>, right) show the effect of deformation.</p>
</caption>
<graphic xlink:href="fcomp-03-613981-g005.tif"/>
</fig>
<p>
<bold>Random noise addition</bold>: During only the training phase, we randomly added Gaussian noise (mean &#xb5; &#x3d; 0, variance <inline-formula id="inf1">
<mml:math id="m2">
<mml:mrow>
<mml:mi>&#x3c3;</mml:mi>
<mml:mo>&#x3d;</mml:mo>
<mml:mn>0.1</mml:mn>
</mml:mrow>
</mml:math>
</inline-formula>) to each EM input image but not its corresponding&#x20;label.</p>
<p>
<bold>Boundary extension</bold>: FusionNet accepts an input image size of 512&#x20;&#xd7; 512. Each input image, and in the case of training its corresponding label, was automatically padded with the mirror reflections of itself across the image border boundary (radius &#x3d; 64&#xa0;px) to maintain similar statistics for pixels that are near the edges. This padding is the reason why FusionNet starts with a 640&#x20;&#xd7; 640 image, which is 128&#xa0;px larger along each edge than the original input. However, we performed convolution with 3&#x20;&#xd7; 3 kernel size and &#x201c;SAME&#x201d; mode, which leads the final segmentation to have the same padded size. To account for this, the final output prediction was cropped to eliminate the padded regions.</p>
</sec>
<sec id="s3-3">
<title>3.3 Experimental Setup</title>
<p>FusionNet was implemented using the Keras open-source deep learning library (<xref ref-type="bibr" rid="B11">Chollet, 2015</xref>). This library provides an easy-to-use, high-level programming API written in Python, with Theano or TensorFlow as a back-end engine. The model was trained with the Adam optimizer with a decaying learning rate of 2e<sup>&#x2212;4</sup> for over 50,000 epochs to harness the benefits of heavy elastic deformation on the small annotated datasets. FusionNet has also been translated to PyTorch and pure TensorFlow for other applications, such as Image-to-Image translation (<xref ref-type="bibr" rid="B33">Lee et&#x20;al., 2018</xref>) and MRI reconstruction (<xref ref-type="bibr" rid="B40">Quan et&#x20;al., 2018</xref>). All training and deployment presented here was conducted on a system with an Intel i7 CPU, 32&#xa0;GB RAM, and a NVIDIA GTX GeForce 1080&#x20;GPU.</p>
</sec>
<sec id="s3-4">
<title>3.4 Network Chaining</title>
<p>FusionNet by itself performs end-to-end segmentation from the EM data input to the final prediction output. In typical real world applications of end-to-end segmentation approaches, however, manual proofreading by human experts is usually performed in an attempt to &#x201c;correct&#x201d; any mistakes in the output labels. We therefore reasoned that concatenating a chain of several FusionNet units could serve as a form of built-in refinement similar to proofreading that could resolve ambiguities in the initial predictions. <xref ref-type="fig" rid="F6">Figure&#x20;6</xref> shows an example case with four chained FusionNet units (FusionNetW<sup>4</sup>). To impose a target-driven approach across the chained network during training, we calculate the loss between the output of each separate unit and the training labels. As a result, chained FusionNet architectures have a single input and multiple outputs, where the end of each decoding path serves as a checkpoint between units attempting to produce better and better segmentation results.</p>
<fig id="F6" position="float">
<label>FIGURE 6</label>
<caption>
<p>FusionNetW<sup>4</sup>, a chain of four concatenated FusionNet&#x20;units.</p>
</caption>
<graphic xlink:href="fcomp-03-613981-g006.tif"/>
</fig>
<p>Since the architecture of each individual unit is the same, the chained FusionNet model can be thought of as similar to an unfolded Recurrent Neural Network (RNN) with each FusionNet unit akin to a single feedback cycle but with weights that are <italic>not</italic> shared across cycles. Each FusionNet can be considered as a V-cycle in the multigrid method (<xref ref-type="bibr" rid="B43">Shapira, 2008</xref>) commonly used in numerical analysis, where the contraction in the encoding path is similar to <italic>restriction</italic> from a fine to a coarse grid, the expansion in the decoding is similar to the <italic>prolongation</italic> toward the final segmentation, and the skip connections play a role similar to <italic>relaxation</italic>. The simplest chain of two V-cycle units forms a W shape, so we refer to FusionNet chains using a &#x201c;FusionNetW&#x201d; terminology. To differentiate various configurations, we use the superscript to indicate how many FusionNet units are chained and the subscript to show the initial number of feature maps in the original resolution. For example, <inline-formula id="inf2">
<mml:math id="m3">
<mml:mrow>
<mml:mi mathvariant="normal">FusionNet</mml:mi>
<mml:msubsup>
<mml:mi mathvariant="normal">W</mml:mi>
<mml:mrow>
<mml:mn>64</mml:mn>
</mml:mrow>
<mml:mn>4</mml:mn>
</mml:msubsup>
</mml:mrow>
</mml:math>
</inline-formula> would signify a network that chains four FusionNet units, each of them with the base number of convolution kernels (in Keras, nb_filters parameter) set to 64. We chose this specific 4-chain example case as the maximum chain length used here ad hoc to roughly match the memory available on our GPU. We also used 64 for the base number of convolution kernels in every case to match the backbone architecture of U-net. During training, the weights of each FusionNet unit <inline-formula id="inf3">
<mml:math id="m4">
<mml:mrow>
<mml:mrow>
<mml:mo>(</mml:mo>
<mml:mrow>
<mml:msup>
<mml:mi>&#x3b8;</mml:mi>
<mml:mi>k</mml:mi>
</mml:msup>
<mml:mrow>
<mml:mo>[</mml:mo>
<mml:mi>i</mml:mi>
<mml:mo>]</mml:mo>
</mml:mrow>
</mml:mrow>
<mml:mo>)</mml:mo>
</mml:mrow>
</mml:mrow>
</mml:math>
</inline-formula> are updated independently, as opposed to the RNN strategy of averaging the gradients from shared weights. For the example FusionNetW<sup>4</sup> case, we trained with the input images <italic>S</italic> and corresponding manual labels <italic>L</italic>. Each FusionNet unit in FusionNetW<sup>4</sup>, which can be indexed as FusionNetW<sup>4</sup>[<italic>i</italic>] where i &#x3d; 1, 2, 3 or 4, generates the prediction P[i] by minimizing the MAE loss between its prediction values and the target labels <italic>L</italic>. For each epoch, we incrementally train FusionNetW<sup>4</sup>[<italic>i</italic>] and fix its weights before training FusionNetW<sup>4</sup>[<italic>i</italic>&#x20;&#x2b; 1]. This procedure can be summarized as follows:<disp-formula id="e1">
<mml:math id="m5">
<mml:mtable columnalign="left">
<mml:mtr>
<mml:mtd>
<mml:munder>
<mml:mrow>
<mml:mi>min</mml:mi>
</mml:mrow>
<mml:mrow>
<mml:msup>
<mml:mi>&#x3b8;</mml:mi>
<mml:mn>4</mml:mn>
</mml:msup>
<mml:mrow>
<mml:mo>[</mml:mo>
<mml:mn>1</mml:mn>
<mml:mo>]</mml:mo>
</mml:mrow>
</mml:mrow>
</mml:munder>
<mml:mi mathvariant="normal">MAE</mml:mi>
<mml:mrow>
<mml:mo>(</mml:mo>
<mml:mrow>
<mml:mi>P</mml:mi>
<mml:mrow>
<mml:mo>[</mml:mo>
<mml:mn>1</mml:mn>
<mml:mo>]</mml:mo>
</mml:mrow>
<mml:mo>,</mml:mo>
<mml:mi>L</mml:mi>
</mml:mrow>
<mml:mo>)</mml:mo>
</mml:mrow>
<mml:mi>s</mml:mi>
<mml:mo>.</mml:mo>
<mml:mi>t</mml:mi>
<mml:mo>.</mml:mo>
<mml:mtext>&#x2003;</mml:mtext>
<mml:mi>P</mml:mi>
<mml:mrow>
<mml:mo>[</mml:mo>
<mml:mn>1</mml:mn>
<mml:mo>]</mml:mo>
</mml:mrow>
<mml:mo>&#x3d;</mml:mo>
<mml:mi>F</mml:mi>
<mml:mi>u</mml:mi>
<mml:mi>s</mml:mi>
<mml:mi>i</mml:mi>
<mml:mi>o</mml:mi>
<mml:mi>n</mml:mi>
<mml:mi>N</mml:mi>
<mml:mi>e</mml:mi>
<mml:msup>
<mml:mi>t</mml:mi>
<mml:mn>4</mml:mn>
</mml:msup>
<mml:mrow>
<mml:mo>[</mml:mo>
<mml:mn>1</mml:mn>
<mml:mo>]</mml:mo>
</mml:mrow>
<mml:mrow>
<mml:mo>(</mml:mo>
<mml:mi>S</mml:mi>
<mml:mo>)</mml:mo>
</mml:mrow>
</mml:mtd>
</mml:mtr>
<mml:mtr>
<mml:mtd>
<mml:munder>
<mml:mrow>
<mml:mi>min</mml:mi>
</mml:mrow>
<mml:mrow>
<mml:msup>
<mml:mi>&#x3b8;</mml:mi>
<mml:mn>4</mml:mn>
</mml:msup>
<mml:mrow>
<mml:mo>[</mml:mo>
<mml:mn>2</mml:mn>
<mml:mo>]</mml:mo>
</mml:mrow>
</mml:mrow>
</mml:munder>
<mml:mi mathvariant="normal">MAE</mml:mi>
<mml:mrow>
<mml:mo>(</mml:mo>
<mml:mrow>
<mml:mi>P</mml:mi>
<mml:mrow>
<mml:mo>[</mml:mo>
<mml:mn>2</mml:mn>
<mml:mo>]</mml:mo>
</mml:mrow>
<mml:mo>,</mml:mo>
<mml:mi>L</mml:mi>
</mml:mrow>
<mml:mo>)</mml:mo>
</mml:mrow>
<mml:mi>s</mml:mi>
<mml:mo>.</mml:mo>
<mml:mi>t</mml:mi>
<mml:mo>.</mml:mo>
<mml:mtext>&#x2003;</mml:mtext>
<mml:mi>P</mml:mi>
<mml:mrow>
<mml:mo>[</mml:mo>
<mml:mn>2</mml:mn>
<mml:mo>]</mml:mo>
</mml:mrow>
<mml:mo>&#x3d;</mml:mo>
<mml:mi>F</mml:mi>
<mml:mi>u</mml:mi>
<mml:mi>s</mml:mi>
<mml:mi>i</mml:mi>
<mml:mi>o</mml:mi>
<mml:mi>n</mml:mi>
<mml:mi>N</mml:mi>
<mml:mi>e</mml:mi>
<mml:msup>
<mml:mi>t</mml:mi>
<mml:mn>4</mml:mn>
</mml:msup>
<mml:mrow>
<mml:mo>[</mml:mo>
<mml:mn>2</mml:mn>
<mml:mo>]</mml:mo>
</mml:mrow>
<mml:mrow>
<mml:mo>(</mml:mo>
<mml:mrow>
<mml:mi>P</mml:mi>
<mml:mrow>
<mml:mo>[</mml:mo>
<mml:mn>1</mml:mn>
<mml:mo>]</mml:mo>
</mml:mrow>
</mml:mrow>
<mml:mo>)</mml:mo>
</mml:mrow>
</mml:mtd>
</mml:mtr>
<mml:mtr>
<mml:mtd>
<mml:munder>
<mml:mrow>
<mml:mi>min</mml:mi>
</mml:mrow>
<mml:mrow>
<mml:msup>
<mml:mi>&#x3b8;</mml:mi>
<mml:mn>4</mml:mn>
</mml:msup>
<mml:mrow>
<mml:mo>[</mml:mo>
<mml:mn>3</mml:mn>
<mml:mo>]</mml:mo>
</mml:mrow>
</mml:mrow>
</mml:munder>
<mml:mi mathvariant="normal">MAE</mml:mi>
<mml:mrow>
<mml:mo>(</mml:mo>
<mml:mrow>
<mml:mi>P</mml:mi>
<mml:mrow>
<mml:mo>[</mml:mo>
<mml:mn>3</mml:mn>
<mml:mo>]</mml:mo>
</mml:mrow>
<mml:mo>,</mml:mo>
<mml:mi>L</mml:mi>
</mml:mrow>
<mml:mo>)</mml:mo>
</mml:mrow>
<mml:mi>s</mml:mi>
<mml:mo>.</mml:mo>
<mml:mi>t</mml:mi>
<mml:mo>.</mml:mo>
<mml:mtext>&#x2003;</mml:mtext>
<mml:mi>P</mml:mi>
<mml:mrow>
<mml:mo>[</mml:mo>
<mml:mn>3</mml:mn>
<mml:mo>]</mml:mo>
</mml:mrow>
<mml:mo>&#x3d;</mml:mo>
<mml:mi>F</mml:mi>
<mml:mi>u</mml:mi>
<mml:mi>s</mml:mi>
<mml:mi>i</mml:mi>
<mml:mi>o</mml:mi>
<mml:mi>n</mml:mi>
<mml:mi>N</mml:mi>
<mml:mi>e</mml:mi>
<mml:msup>
<mml:mi>t</mml:mi>
<mml:mn>4</mml:mn>
</mml:msup>
<mml:mrow>
<mml:mo>[</mml:mo>
<mml:mn>3</mml:mn>
<mml:mo>]</mml:mo>
</mml:mrow>
<mml:mrow>
<mml:mo>(</mml:mo>
<mml:mrow>
<mml:mi>P</mml:mi>
<mml:mrow>
<mml:mo>[</mml:mo>
<mml:mn>2</mml:mn>
<mml:mo>]</mml:mo>
</mml:mrow>
</mml:mrow>
<mml:mo>)</mml:mo>
</mml:mrow>
</mml:mtd>
</mml:mtr>
<mml:mtr>
<mml:mtd>
<mml:munder>
<mml:mrow>
<mml:mi>min</mml:mi>
</mml:mrow>
<mml:mrow>
<mml:msup>
<mml:mi>&#x3b8;</mml:mi>
<mml:mn>4</mml:mn>
</mml:msup>
<mml:mrow>
<mml:mo>[</mml:mo>
<mml:mn>4</mml:mn>
<mml:mo>]</mml:mo>
</mml:mrow>
</mml:mrow>
</mml:munder>
<mml:mi mathvariant="normal">MAE</mml:mi>
<mml:mrow>
<mml:mo>(</mml:mo>
<mml:mrow>
<mml:mi>P</mml:mi>
<mml:mrow>
<mml:mo>[</mml:mo>
<mml:mn>4</mml:mn>
<mml:mo>]</mml:mo>
</mml:mrow>
<mml:mo>,</mml:mo>
<mml:mi>L</mml:mi>
</mml:mrow>
<mml:mo>)</mml:mo>
</mml:mrow>
<mml:mi>s</mml:mi>
<mml:mo>.</mml:mo>
<mml:mi>t</mml:mi>
<mml:mo>.</mml:mo>
<mml:mtext>&#x2003;</mml:mtext>
<mml:mi>P</mml:mi>
<mml:mrow>
<mml:mo>[</mml:mo>
<mml:mn>4</mml:mn>
<mml:mo>]</mml:mo>
</mml:mrow>
<mml:mo>&#x3d;</mml:mo>
<mml:mi>F</mml:mi>
<mml:mi>u</mml:mi>
<mml:mi>s</mml:mi>
<mml:mi>i</mml:mi>
<mml:mi>o</mml:mi>
<mml:mi>n</mml:mi>
<mml:mi>N</mml:mi>
<mml:mi>e</mml:mi>
<mml:msup>
<mml:mi>t</mml:mi>
<mml:mn>4</mml:mn>
</mml:msup>
<mml:mrow>
<mml:mo>[</mml:mo>
<mml:mn>4</mml:mn>
<mml:mo>]</mml:mo>
</mml:mrow>
<mml:mrow>
<mml:mo>(</mml:mo>
<mml:mrow>
<mml:mi>P</mml:mi>
<mml:mrow>
<mml:mo>[</mml:mo>
<mml:mn>3</mml:mn>
<mml:mo>]</mml:mo>
</mml:mrow>
</mml:mrow>
<mml:mo>)</mml:mo>
</mml:mrow>
</mml:mtd>
</mml:mtr>
</mml:mtable>
</mml:math>
<label>(1)</label>
</disp-formula>
</p>
<p>The loss training curves decrease as <italic>i</italic> increases, eventually converging as the number of training epochs increases.</p>
</sec>
</sec>
<sec sec-type="results" id="s4">
<title>4 Results</title>
<sec id="s4-1">
<title>4.1 Fruit Fly Data</title>
<p>The fruit fly (<italic>Drosophila</italic>) ventral nerve cord EM data used here was captured from a first instar larva (<xref ref-type="bibr" rid="B7">Cardona et&#x20;al., 2010</xref>). Training and test datasets were provided as part of the ISBI 2012 EM segmentation challenge<xref ref-type="fn" rid="fn1">
<sup>1</sup>
</xref> (<xref ref-type="bibr" rid="B1">Arganda-Carreras et&#x20;al., 2015</xref>). Each dataset consisted of a 512&#x20;&#xd7; 512&#x20;&#xd7; 30 volume acquired at anisotropic <inline-formula id="inf4">
<mml:math id="m6">
<mml:mrow>
<mml:mn>4</mml:mn>
<mml:mtext>&#xa0;</mml:mtext>
<mml:mo>&#xd7;</mml:mo>
<mml:mn>4</mml:mn>
<mml:mtext>&#xa0;</mml:mtext>
<mml:mo>&#xd7;</mml:mo>
<mml:mo>&#x223c;</mml:mo>
<mml:mn>50</mml:mn>
<mml:mtext>&#x2009;</mml:mtext>
<mml:mi mathvariant="normal">n</mml:mi>
<mml:msup>
<mml:mi mathvariant="normal">m</mml:mi>
<mml:mn>3</mml:mn>
</mml:msup>
<mml:mtext>&#x2009;</mml:mtext>
<mml:mi mathvariant="normal">v</mml:mi>
<mml:msup>
<mml:mi mathvariant="normal">x</mml:mi>
<mml:mrow>
<mml:mo>&#x2212;</mml:mo>
<mml:mn>1</mml:mn>
</mml:mrow>
</mml:msup>
<mml:mtext>&#xa0;</mml:mtext>
</mml:mrow>
</mml:math>
</inline-formula> resolution with transmission EM. These datasets were originally chosen in part because they contained noise and small image alignment errors that frequently occur in serial-section EM. For training, the provided dataset included EM image data and publicly available manual segmentation labels. The first 20 of 30 slices of the training volume were used for training and the last 10 slices were used for validation. For testing, the provided dataset included only EM image data, while segmentation labels were kept private for the assessment of segmentation accuracy (<xref ref-type="bibr" rid="B1">Arganda-Carreras et&#x20;al., 2015</xref>). Test segmentations were produced for all 30 slices of the test volume and were then uploaded for comparison to the hidden ISBI Challenge segmentation labels.</p>
<p>
<xref ref-type="fig" rid="F7">Figure&#x20;7</xref> illustrates the <inline-formula id="inf5">
<mml:math id="m7">
<mml:mrow>
<mml:mi mathvariant="normal">FusionNet</mml:mi>
<mml:msubsup>
<mml:mi mathvariant="normal">W</mml:mi>
<mml:mrow>
<mml:mn>64</mml:mn>
</mml:mrow>
<mml:mn>2</mml:mn>
</mml:msubsup>
</mml:mrow>
</mml:math>
</inline-formula> probability map extraction results from test data without any post-processing steps (middle) and with lifted multi-cut (LMC) algorithm post-processing (right) (<xref ref-type="bibr" rid="B3">Beier et&#x20;al., 2017</xref>), which resulted in thinning of the probability map. As this shows, our chained FusionNet method is able to remove extraneous structures belonging to mitochondria (appearing as dark shaded textures) and vesicles (appearing as small circles). Uncertain regions in the prediction results without post-processing appear as blurry gray smears (highlighted by pink boxes). In cases like this, <inline-formula id="inf6">
<mml:math id="m8">
<mml:mrow>
<mml:mi mathvariant="normal">FusionNet</mml:mi>
<mml:msubsup>
<mml:mi mathvariant="normal">W</mml:mi>
<mml:mrow>
<mml:mn>64</mml:mn>
</mml:mrow>
<mml:mn>2</mml:mn>
</mml:msubsup>
</mml:mrow>
</mml:math>
</inline-formula> must decide whether or not the highlighted pixels should be segmented as membrane, but the region is ambiguous because of membrane smearing, likely due to anisotropy in the&#x20;data.</p>
<fig id="F7" position="float">
<label>FIGURE 7</label>
<caption>
<p>Example results of cellular membrane segmentation on test data from the ISBI 2012 EM segmentation challenge (slice 22/30) illustrating an input EM image <bold>(left)</bold>, the probability prediction from <inline-formula id="inf7">
<mml:math id="m9">
<mml:mrow>
<mml:mi mathvariant="normal">FusionNet</mml:mi>
<mml:msubsup>
<mml:mi mathvariant="normal">W</mml:mi>
<mml:mrow>
<mml:mn>64</mml:mn>
</mml:mrow>
<mml:mn>2</mml:mn>
</mml:msubsup>
</mml:mrow>
</mml:math>
</inline-formula> <bold>(middle)</bold>, and the thinned probability prediction after applying LMC (<xref ref-type="bibr" rid="B3">Beier et&#x20;al., 2017</xref>) post-processing <bold>(right)</bold>. Pink boxes highlight uncertain regions that are ambiguous because of membrane smearing, likely due to anisotropy in the&#x20;data.</p>
</caption>
<graphic xlink:href="fcomp-03-613981-g007.tif"/>
</fig>
<p>FusionNet approaches outperformed several other methods in segmenting the ISBI 2012 EM challenge data by several standard metrics. These metrics include foreground-restricted Rand scoring after border thinning <inline-formula id="inf8">
<mml:math id="m10">
<mml:mrow>
<mml:mrow>
<mml:mo>(</mml:mo>
<mml:mrow>
<mml:msub>
<mml:mi>V</mml:mi>
<mml:mrow>
<mml:mi>r</mml:mi>
<mml:mi>a</mml:mi>
<mml:mi>n</mml:mi>
<mml:mi>d</mml:mi>
</mml:mrow>
</mml:msub>
</mml:mrow>
<mml:mo>)</mml:mo>
</mml:mrow>
</mml:mrow>
</mml:math>
</inline-formula> and foreground-restricted information-theoretic scoring after border thinning (<italic>V<sub>info</sub>
</italic>) (<xref ref-type="bibr" rid="B1">Arganda-Carreras et&#x20;al., 2015</xref>). Quantitative comparisons with other methods are summarized in <xref ref-type="table" rid="T2">Table&#x20;2</xref>. Even using a single FusionNet unit (<inline-formula id="inf9">
<mml:math id="m11">
<mml:mrow>
<mml:mi mathvariant="normal">FusionNet</mml:mi>
<mml:msubsup>
<mml:mi mathvariant="normal">W</mml:mi>
<mml:mrow>
<mml:mn>64</mml:mn>
</mml:mrow>
<mml:mn>1</mml:mn>
</mml:msubsup>
</mml:mrow>
</mml:math>
</inline-formula>), we achieved better results compared to many well-known methods, such as U-net (<xref ref-type="bibr" rid="B41">Ronneberger et&#x20;al., 2015</xref>), network-in-network (<xref ref-type="bibr" rid="B35">Lin et&#x20;al., 2014</xref>), fused-architecture (<xref ref-type="bibr" rid="B9">Chen et&#x20;al., 2016a</xref>), and long short-term memory (LSTM) (<xref ref-type="bibr" rid="B48">Stollenga et&#x20;al., 2015</xref>) approaches. Using a chained FusionNet with two modules (<inline-formula id="inf10">
<mml:math id="m12">
<mml:mrow>
<mml:mi mathvariant="normal">FusionNet</mml:mi>
<mml:msubsup>
<mml:mi mathvariant="normal">W</mml:mi>
<mml:mrow>
<mml:mn>64</mml:mn>
</mml:mrow>
<mml:mn>2</mml:mn>
</mml:msubsup>
</mml:mrow>
</mml:math>
</inline-formula>) performed even better, surpassing the performance of many previous state-of-the-art deep learning methods (<xref ref-type="bibr" rid="B10">Chen et&#x20;al., 2016b</xref>; <xref ref-type="bibr" rid="B14">Drozdzal et&#x20;al., 2016</xref>). These results confirm that chaining a deeper architecture with a residual bottleneck helps to increase the accuracy of the EM segmentation task. Both with and without LMC post-processing, <inline-formula id="inf11">
<mml:math id="m13">
<mml:mrow>
<mml:mi mathvariant="normal">FusionNet</mml:mi>
<mml:msubsup>
<mml:mi mathvariant="normal">W</mml:mi>
<mml:mrow>
<mml:mn>64</mml:mn>
</mml:mrow>
<mml:mn>2</mml:mn>
</mml:msubsup>
</mml:mrow>
</mml:math>
</inline-formula> ranks among the top 10 in the ISBI 2012 EM segmentation challenge leaderboard (as of June 2020).</p>
<table-wrap id="T2" position="float">
<label>TABLE 2</label>
<caption>
<p>Accuracy of various segmentation methods on the <italic>Drosophila</italic> EM dataset (ISBI 2012 EM segmentation challenge leaderboard, June 2020). Bold values correspond to the method presented here.</p>
</caption>
<table>
<thead valign="top">
<tr>
<th align="left">Methods</th>
<th align="center">
<italic>V<sub>rand</sub>
</italic>
</th>
<th align="center">
<italic>V<sub>info</sub>
</italic>
</th>
</tr>
</thead>
<tbody valign="top">
<tr>
<td align="left">&#x2a;&#x2a;Human values&#x2a;&#x2a;</td>
<td align="char" char=".">0.997847778</td>
<td align="char" char=".">0.998997659</td>
</tr>
<tr>
<td align="left">PatchPerPix <xref ref-type="bibr" rid="B27">Hirsch et&#x20;al. (2020)</xref>
</td>
<td align="char" char=".">0.988290649</td>
<td align="char" char=".">0.991641507</td>
</tr>
<tr>
<td align="left">IAL MutexWS <xref ref-type="bibr" rid="B54">Wolf et&#x20;al. (2019)</xref>
</td>
<td align="char" char=".">0.987922250</td>
<td align="char" char=".">0.991833594</td>
</tr>
<tr>
<td align="left">CASIA MIRA <xref ref-type="bibr" rid="B56">Xiao et&#x20;al. (2018)</xref>
</td>
<td align="char" char=".">0.987877739</td>
<td align="char" char=".">0.990920188</td>
</tr>
<tr>
<td align="left">IAL - SFCNN <xref ref-type="bibr" rid="B51">Weiler et&#x20;al. (2017)</xref>
</td>
<td align="char" char=".">0.986800916</td>
<td align="char" char=".">0.991438892</td>
</tr>
<tr>
<td align="left">ACE-net <xref ref-type="bibr" rid="B60">Zhu et&#x20;al. (2019)</xref>
</td>
<td align="char" char=".">0.985032746</td>
<td align="char" char=".">0.989490497</td>
</tr>
<tr>
<td align="left">M2FCN-MFA <xref ref-type="bibr" rid="B45">Shen et&#x20;al. (2017)</xref>
</td>
<td align="char" char=".">0.983651122</td>
<td align="char" char=".">0.991303595</td>
</tr>
<tr>
<td align="left">
<inline-formula id="inf12">
<mml:math id="m14">
<mml:mrow>
<mml:mi mathvariant="bold">FusionNet</mml:mi>
<mml:msubsup>
<mml:mi mathvariant="bold">W</mml:mi>
<mml:mrow>
<mml:mi mathvariant="bold">64</mml:mi>
</mml:mrow>
<mml:mi mathvariant="bold">2</mml:mi>
</mml:msubsup>
</mml:mrow>
</mml:math>
</inline-formula> <bold>LMC</bold>
</td>
<td align="char" char=".">
<bold>0.983651122</bold>
</td>
<td align="char" char=".">
<bold>0.991303595</bold>
</td>
</tr>
<tr>
<td align="left">IAL MC/LMC <xref ref-type="bibr" rid="B3">Beier et&#x20;al. (2017)</xref>
</td>
<td align="char" char=".">0.982616131</td>
<td align="char" char=".">0.989461939</td>
</tr>
<tr>
<td align="left">IAL LMC <xref ref-type="bibr" rid="B2">Beier et&#x20;al. (2016)</xref>
</td>
<td align="char" char=".">0.982240005</td>
<td align="char" char=".">0.988448278</td>
</tr>
<tr>
<td align="left">
<inline-formula id="inf13">
<mml:math id="m15">
<mml:mrow>
<mml:mi mathvariant="bold">FusionNet</mml:mi>
<mml:msubsup>
<mml:mi mathvariant="bold">W</mml:mi>
<mml:mrow>
<mml:mi mathvariant="bold">64</mml:mi>
</mml:mrow>
<mml:mi mathvariant="bold">2</mml:mi>
</mml:msubsup>
</mml:mrow>
</mml:math>
</inline-formula>
</td>
<td align="char" char=".">
<bold>0.981586186</bold>
</td>
<td align="char" char=".">
<bold>0.990099898</bold>
</td>
</tr>
<tr>
<td align="left">PolyMtl <xref ref-type="bibr" rid="B14">Drozdzal et&#x20;al. (2016)</xref>
</td>
<td align="char" char=".">0.980582825</td>
<td align="char" char=".">0.988163049</td>
</tr>
<tr>
<td align="left">KUnet <xref ref-type="bibr" rid="B10">Chen et&#x20;al. (2016b)</xref>
</td>
<td align="char" char=".">0.980222514</td>
<td align="char" char=".">0.988967601</td>
</tr>
<tr>
<td align="left">
<inline-formula id="inf14">
<mml:math id="m16">
<mml:mrow>
<mml:mi mathvariant="bold">FusionNet</mml:mi>
<mml:msubsup>
<mml:mi mathvariant="bold">W</mml:mi>
<mml:mrow>
<mml:mi mathvariant="bold">64</mml:mi>
</mml:mrow>
<mml:mi mathvariant="bold">1</mml:mi>
</mml:msubsup>
</mml:mrow>
</mml:math>
</inline-formula>
</td>
<td align="char" char=".">
<bold>0.978042575</bold>
</td>
<td align="char" char=".">
<bold>0.989945379</bold>
</td>
</tr>
<tr>
<td align="left">IAL IC <xref ref-type="bibr" rid="B35">Lin et&#x20;al. (2014)</xref>
</td>
<td align="char" char=".">0.977345721</td>
<td align="char" char=".">0.989240736</td>
</tr>
<tr>
<td align="left">Masters <xref ref-type="bibr" rid="B53">Wiehman and Villiers (2016)</xref>
</td>
<td align="char" char=".">0.977141154</td>
<td align="char" char=".">0.987534429</td>
</tr>
<tr>
<td align="left">CUMedVision <xref ref-type="bibr" rid="B9">Chen et&#x20;al. (2016a)</xref>
</td>
<td align="char" char=".">0.976824580</td>
<td align="char" char=".">0.988645822</td>
</tr>
<tr>
<td align="left">ICNN <xref ref-type="bibr" rid="B55">Wu (2015)</xref>
</td>
<td align="char" char=".">0.976546913</td>
<td align="char" char=".">0.988341665</td>
</tr>
<tr>
<td align="left">DIVE-SCI <xref ref-type="bibr" rid="B17">Fakhry et&#x20;al. (2016)</xref>
</td>
<td align="char" char=".">0.976229111</td>
<td align="char" char=".">0.987392123</td>
</tr>
<tr>
<td align="left">LSTM <xref ref-type="bibr" rid="B48">Stollenga et&#x20;al. (2015)</xref>
</td>
<td align="char" char=".">0.975366444</td>
<td align="char" char=".">0.987425430</td>
</tr>
<tr>
<td align="left">U-net <xref ref-type="bibr" rid="B41">Ronneberger et&#x20;al. (2015)</xref>
</td>
<td align="char" char=".">0.972760748</td>
<td align="char" char=".">0.986616590</td>
</tr>
</tbody>
</table>
</table-wrap>
</sec>
<sec id="s4-2">
<title>4.2 Zebrafish Data</title>
<p>The zebrafish EM data used here was taken from a publicly available database<xref ref-type="fn" rid="fn2">
<sup>2</sup>
</xref>. It was captured from a 5.5&#xa0;days post-fertilization larval specimen. This specimen was cut into &#x223c;18,000 serial sections and collected onto a tape substrate with an automated tape-collecting ultramicrotome (ATUM) (<xref ref-type="bibr" rid="B22">Hayworth et&#x20;al., 2014</xref>). A series of images spanning the anterior quarter of the larval zebrafish was acquired at <inline-formula id="inf15">
<mml:math id="m17">
<mml:mrow>
<mml:mn>56.4</mml:mn>
<mml:mo>&#xd7;</mml:mo>
<mml:mn>56.4</mml:mn>
<mml:mo>&#xd7;</mml:mo>
<mml:mo>&#x223c;</mml:mo>
<mml:mn>60</mml:mn>
<mml:mtext>&#x2009;</mml:mtext>
<mml:mi mathvariant="normal">n</mml:mi>
<mml:msup>
<mml:mi mathvariant="normal">m</mml:mi>
<mml:mn>3</mml:mn>
</mml:msup>
<mml:mtext>&#x2009;</mml:mtext>
<mml:mi mathvariant="normal">v</mml:mi>
<mml:msup>
<mml:mi mathvariant="normal">x</mml:mi>
<mml:mrow>
<mml:mo>&#x2212;</mml:mo>
<mml:mn>1</mml:mn>
</mml:mrow>
</mml:msup>
</mml:mrow>
</mml:math>
</inline-formula> resolution from 16,000 sections using scanning EM (<xref ref-type="bibr" rid="B25">Hildebrand, 2015</xref>; <xref ref-type="bibr" rid="B26">Hildebrand et&#x20;al., 2017</xref>). All 2D images were then co-registered into a 3D volume using an FFT signal whitening approach (<xref ref-type="bibr" rid="B52">Wetzel et&#x20;al., 2016</xref>). For training, two small sub-volume crops were extracted from a near-final iteration of the full volume alignment in order to avoid deploying later segmentation runs on training data. Two training volumes that contained different tissue features were chosen. One volume was 512&#x20;&#xd7; 512&#x20;&#xd7; 512 and the other was 512&#x20;&#xd7; 512&#x20;&#xd7; 256. The blob-like features of interest&#x2014;neuronal nuclei&#x2014;were manually segmented as area-lists in each training volume using the Fiji (<xref ref-type="bibr" rid="B44">Schindelin et&#x20;al., 2012</xref>) TrakEM2&#x20;plug-in (<xref ref-type="bibr" rid="B8">Cardona et&#x20;al., 2012</xref>). From each of these two training volumes, three quarters were used for training and one quarter was used for validation. These area-lists were exported as binary masks for use in the training procedure. For accuracy assessments, an additional non-overlapping 512&#x20;&#xd7; 512&#x20;&#xd7; 512 testing sub-volume and corresponding manual segmentation labels were&#x20;used.</p>
<p>To assess the performance of <inline-formula id="inf16">
<mml:math id="m18">
<mml:mrow>
<mml:mi mathvariant="normal">FusionNet</mml:mi>
<mml:msubsup>
<mml:mi mathvariant="normal">W</mml:mi>
<mml:mrow>
<mml:mn>16</mml:mn>
</mml:mrow>
<mml:mn>4</mml:mn>
</mml:msubsup>
</mml:mrow>
</mml:math>
</inline-formula> on this segmentation task, we first deployed it on 512&#x20;&#xd7; 512&#x20;&#xd7; 512 test volume alongside the U-net (<xref ref-type="bibr" rid="B41">Ronneberger et&#x20;al., 2015</xref>) and RDN (<xref ref-type="bibr" rid="B18">Fakhry et&#x20;al., 2017</xref>) methods. <xref ref-type="fig" rid="F8">Figure&#x20;8</xref> displays volume renderings of the zebrafish test set EM data, its manual cell nucleus segmentation, and segmentation results from U-net, RDN, and <inline-formula id="inf17">
<mml:math id="m19">
<mml:mrow>
<mml:mi mathvariant="normal">FusionNet</mml:mi>
<mml:msubsup>
<mml:mi mathvariant="normal">W</mml:mi>
<mml:mrow>
<mml:mn>16</mml:mn>
</mml:mrow>
<mml:mn>4</mml:mn>
</mml:msubsup>
</mml:mrow>
</mml:math>
</inline-formula>. As this shows, <inline-formula id="inf18">
<mml:math id="m20">
<mml:mrow>
<mml:mi mathvariant="normal">FusionNet</mml:mi>
<mml:msubsup>
<mml:mi mathvariant="normal">W</mml:mi>
<mml:mrow>
<mml:mn>16</mml:mn>
</mml:mrow>
<mml:mn>4</mml:mn>
</mml:msubsup>
</mml:mrow>
</mml:math>
</inline-formula> introduced less false predictions compared to U-net and RDN. <xref ref-type="table" rid="T3">Table&#x20;3</xref> compares U-net, RDN, and <inline-formula id="inf19">
<mml:math id="m21">
<mml:mrow>
<mml:mi mathvariant="normal">FusionNet</mml:mi>
<mml:msubsup>
<mml:mi mathvariant="normal">W</mml:mi>
<mml:mrow>
<mml:mn>16</mml:mn>
</mml:mrow>
<mml:mn>4</mml:mn>
</mml:msubsup>
</mml:mrow>
</mml:math>
</inline-formula> using three quality metrics: foreground-restricted Rand scoring after border thinning (<italic>V</italic>
<sub>
<italic>rand</italic>
</sub>), foreground-restricted information theoretic scoring after border thinning (<italic>V</italic>
<sub>
<italic>info</italic>
</sub>), and the Dice coefficient (<italic>V</italic>
<sub>
<italic>dice</italic>
</sub>). By all of these metrics, <inline-formula id="inf20">
<mml:math id="m22">
<mml:mrow>
<mml:mi mathvariant="normal">FusionNet</mml:mi>
<mml:msubsup>
<mml:mi mathvariant="normal">W</mml:mi>
<mml:mrow>
<mml:mn>16</mml:mn>
</mml:mrow>
<mml:mn>4</mml:mn>
</mml:msubsup>
</mml:mrow>
</mml:math>
</inline-formula> produced more accurate segmentation results.</p>
<fig id="F8" position="float">
<label>FIGURE 8</label>
<caption>
<p>Visual comparison of the larval zebrafish EM volume segmentation. <bold>(A)</bold> Input serial-section EM volume. <bold>(B)</bold> Manual segmentation (ground truth). <bold>(C)</bold> U-net (<xref ref-type="bibr" rid="B41">Ronneberger et&#x20;al., 2015</xref>) result. <bold>(D)</bold> RDN (<xref ref-type="bibr" rid="B18">Fakhry et&#x20;al., 2017</xref>) result. <bold>(E)</bold> <inline-formula id="inf21">
<mml:math id="m23">
<mml:mrow>
<mml:mi mathvariant="normal">FusionNet</mml:mi>
<mml:msubsup>
<mml:mi mathvariant="normal">W</mml:mi>
<mml:mrow>
<mml:mn>16</mml:mn>
</mml:mrow>
<mml:mn>4</mml:mn>
</mml:msubsup>
</mml:mrow>
</mml:math>
</inline-formula> result. Red arrows indicate errors.</p>
</caption>
<graphic xlink:href="fcomp-03-613981-g008.tif"/>
</fig>
<table-wrap id="T3" position="float">
<label>TABLE 3</label>
<caption>
<p>Segmentation accuracy on a test volume from the zebrafish EM dataset. Bold values correspond to the method presented here.</p>
</caption>
<table>
<thead valign="top">
<tr>
<th align="left">Methods</th>
<th align="center">
<inline-formula id="inf22">
<mml:math id="m24">
<mml:mrow>
<mml:mi mathvariant="bold">FusionNet</mml:mi>
<mml:msubsup>
<mml:mi mathvariant="bold">W</mml:mi>
<mml:mrow>
<mml:mi mathvariant="bold">64</mml:mi>
</mml:mrow>
<mml:mi mathvariant="bold">2</mml:mi>
</mml:msubsup>
</mml:mrow>
</mml:math>
</inline-formula>
</th>
<th align="center">RDN <xref ref-type="bibr" rid="B18">Fakhry et&#x20;al. (2017)</xref>
</th>
<th align="center">U-net <xref ref-type="bibr" rid="B41">Ronneberger et&#x20;al. (2015)</xref>
</th>
</tr>
</thead>
<tbody valign="top">
<tr>
<td align="left">V<sub>
<italic>rand</italic>
</sub>
</td>
<td align="char" char=".">
<bold>0.998648782</bold>
</td>
<td align="char" char=".">0.991844302</td>
<td align="char" char=".">0.987366177</td>
</tr>
<tr>
<td align="left">V<sub>
<italic>info</italic>
</sub>
</td>
<td align="char" char=".">
<bold>0.996929124</bold>
</td>
<td align="char" char=".">0.994208722</td>
<td align="char" char=".">0.992482059</td>
</tr>
<tr>
<td align="left">V<sub>
<italic>dice</italic>
</sub>
</td>
<td align="char" char=".">
<bold>0.963047248</bold>
</td>
<td align="char" char=".">0.946099985</td>
<td align="char" char=".">0.908491647</td>
</tr>
</tbody>
</table>
</table-wrap>
<p>We also deployed the trained network to the complete set of 16,000 sections of the larval zebrafish brain imaged at <inline-formula id="inf23">
<mml:math id="m25">
<mml:mrow>
<mml:mn>56.4</mml:mn>
<mml:mo>&#xd7;</mml:mo>
<mml:mn>56.4</mml:mn>
<mml:mo>&#xd7;</mml:mo>
<mml:mo>&#x223c;</mml:mo>
<mml:mn>60</mml:mn>
<mml:mtext>&#x2009;</mml:mtext>
<mml:mi mathvariant="normal">n</mml:mi>
<mml:msup>
<mml:mi mathvariant="normal">m</mml:mi>
<mml:mn>3</mml:mn>
</mml:msup>
<mml:mtext>&#x2009;</mml:mtext>
<mml:mi mathvariant="normal">v</mml:mi>
<mml:msup>
<mml:mi mathvariant="normal">x</mml:mi>
<mml:mrow>
<mml:mo>&#x2212;</mml:mo>
<mml:mn>1</mml:mn>
</mml:mrow>
</mml:msup>
<mml:mtext>&#xa0;</mml:mtext>
</mml:mrow>
</mml:math>
</inline-formula> resolution, which is about 1.2&#xa0;terabytes in data size. <xref ref-type="fig" rid="F9">Figure&#x20;9</xref> shows EM dataset cross-sections in the transverse (top, <italic>x</italic>-<italic>y</italic>) and horizontal (bottom, <italic>x</italic>-<italic>z</italic>) planes of the larval zebrafish overlaid with the cell nucleus segmentation results. The transverse view overlay also shows the sphericity of each segmented cell nucleus in a blue to red color map, which can help to visually identify the location of false positives.</p>
<fig id="F9" position="float">
<label>FIGURE 9</label>
<caption>
<p>Cell nucleus segmentation results overlaid onto zebrafish EM volume cross-sections through the transverse (<bold>top</bold>, blue to red color map varies with cell sphericity) and horizontal <bold>(bottom)</bold> planes.</p>
</caption>
<graphic xlink:href="fcomp-03-613981-g009.tif"/>
</fig>
</sec>
</sec>
<sec sec-type="conclusions" id="s5">
<title>5 Conclusions</title>
<p>In this paper, we introduced a deep neural network architecture for image segmentation with a focus on connectomics EM image analysis. The proposed architecture, FusionNet, extends the U-net and residual CNN architectures to develop a deeper network for a more accurate end-to-end segmentation. We demonstrated the flexibility and performance of FusionNet in membrane- and blob-type EM segmentation&#x20;tasks.</p>
<p>Several other approaches share similarities with FusionNet, particularly in concatenated chain forms. Chen et&#x20;al. proposed concatenating multiple FCNs to build a RNN that extracts inter-slice contexts (<xref ref-type="bibr" rid="B10">Chen et&#x20;al., 2016b</xref>). Unlike FusionNet, this approach takes as input multiple different resolutions of the raw image to produce a single segmentation output and uses a single loss function. Wu proposed iteratively applying a pixel-wise CNN (ICNN) to refine membrane detection probability maps (MDPM) (<xref ref-type="bibr" rid="B55">Wu, 2015</xref>). In this method, a regular CNN for generating MDPM from the raw input images and an iterative CNN for refining MDPM are trained independently. In contrast, FusionNet is trained as a single chained network. Additionally, FusionNet can refine errors in MDPM more completely using a chained network (i.e.,&#x20;by correcting errors in the error-corrected results) and scales better to larger image sizes due to the end-to-end nature of the network. More in-depth analyses into why chaining approaches are beneficial to improve the prediction accuracy of such deep networks will be an important goal for future&#x20;work.</p>
</sec>
</body>
<back>
<sec id="s6">
<title>Data Availability Statement</title>
<p>The original contributions presented in the study are included in the article/Supplementary Material, further inquiries can be directed to the corresponding author.</p>
</sec>
<sec id="s7">
<title>Author Contributions</title>
<p>TMQ developed the methods and performed the experiments with input from W-KJ and DGCH. Project supervision and funding were provided by W-KJ. All authors wrote the&#x20;paper.</p>
</sec>
<sec id="s8">
<title>Funding</title>
<p>This work was supported by a NAVER Fellowship to TMQ and by a Leon Levy Foundation Fellowship in Neuroscience to DGCH.</p>
</sec>
<sec sec-type="COI-statement" id="s9">
<title>Conflict of Interest</title>
<p>The authors declare that the research was conducted in the absence of any commercial or financial relationships that could be construed as a potential conflict of interest.</p>
</sec>
<ack>
<p>We thank Woohyuk Choi and Jungmin Moon for assistance in creating the larval zebrafish EM volume renderings. This manuscript has been released as a pre-print at <ext-link ext-link-type="uri" xmlns:xlink="http://www.w3.org/1999/xlink" xlink:href="https://arxiv.org/abs/1612.05360">https://arxiv.org/abs/1612.05360</ext-link> (<xref ref-type="bibr" rid="B39">Quan et&#x20;al., 2016</xref>).</p>
</ack>
<fn-group>
<fn id="fn1">
<label>1</label>
<p>
<ext-link ext-link-type="uri" xlink:href="%20http://brainiac2.mit.edu/isbi_challenge/">http://brainiac2.mit.edu/isbi_challenge/</ext-link>
</p>
</fn>
<fn id="fn2">
<label>2</label>
<p>
<ext-link ext-link-type="uri" xlink:href="%20http://zebrafish.link/hildebrand16/">http://zebrafish.link/hildebrand16/</ext-link>
</p>
</fn>
</fn-group>
<ref-list>
<title>References</title>
<ref id="B1">
<citation citation-type="journal">
<person-group person-group-type="author">
<name>
<surname>Arganda-Carreras</surname>
<given-names>I.</given-names>
</name>
<name>
<surname>Turaga</surname>
<given-names>S. C.</given-names>
</name>
<name>
<surname>Berger</surname>
<given-names>D. R.</given-names>
</name>
<name>
<surname>Cire&#x15f;an</surname>
<given-names>D.</given-names>
</name>
<name>
<surname>Giusti</surname>
<given-names>A.</given-names>
</name>
<name>
<surname>Gambardella</surname>
<given-names>L. M.</given-names>
</name>
<etal/>
</person-group> (<year>2015</year>). <article-title>Crowdsourcing the Creation of Image Segmentation Algorithms for Connectomics</article-title>. <source>Front. Neuroanat.</source> <volume>910</volume>, <fpage>142</fpage>. <pub-id pub-id-type="doi">10.3389/fnana.2015.00142</pub-id> </citation>
</ref>
<ref id="B2">
<citation citation-type="book">
<person-group person-group-type="author">
<name>
<surname>Beier</surname>
<given-names>T.</given-names>
</name>
<name>
<surname>Andres</surname>
<given-names>B.</given-names>
</name>
<name>
<surname>K&#xf6;the</surname>
<given-names>U.</given-names>
</name>
<name>
<surname>Hamprecht</surname>
<given-names>F. A.</given-names>
</name>
</person-group> (<year>2016</year>). &#x201c;<article-title>An Efficient Fusion Move Algorithm for the Minimum Cost Lifted Multicut Problem</article-title>,&#x201d; in <conf-name>Proceedings of ECCV 2016</conf-name>, <conf-loc>Amsterdam, Netherlands</conf-loc>, <conf-date>October 8&#x2013;16, 2016</conf-date> (<publisher-name>Springer, Cham</publisher-name>), <fpage>715</fpage>&#x2013;<lpage>730</lpage>. </citation>
</ref>
<ref id="B3">
<citation citation-type="journal">
<person-group person-group-type="author">
<name>
<surname>Beier</surname>
<given-names>T.</given-names>
</name>
<name>
<surname>Pape</surname>
<given-names>C.</given-names>
</name>
<name>
<surname>Rahaman</surname>
<given-names>N.</given-names>
</name>
<name>
<surname>Prange</surname>
<given-names>T.</given-names>
</name>
<name>
<surname>Berg</surname>
<given-names>S.</given-names>
</name>
<name>
<surname>Bock</surname>
<given-names>D. D.</given-names>
</name>
<etal/>
</person-group> (<year>2017</year>). <article-title>Multicut Brings Automated Neurite Segmentation Closer to Human Performance</article-title>. <source>Nat. Methods</source> <volume>14</volume>, <fpage>101</fpage>&#x2013;<lpage>102</lpage>. <pub-id pub-id-type="doi">10.1038/nmeth.4151</pub-id> </citation>
</ref>
<ref id="B4">
<citation citation-type="journal">
<person-group person-group-type="author">
<name>
<surname>Bock</surname>
<given-names>D. D.</given-names>
</name>
<name>
<surname>Lee</surname>
<given-names>W.-C. A.</given-names>
</name>
<name>
<surname>Kerlin</surname>
<given-names>A. M.</given-names>
</name>
<name>
<surname>Andermann</surname>
<given-names>M. L.</given-names>
</name>
<name>
<surname>Hood</surname>
<given-names>G.</given-names>
</name>
<name>
<surname>Wetzel</surname>
<given-names>A. W.</given-names>
</name>
<etal/>
</person-group> (<year>2011</year>). <article-title>Network Anatomy and <italic>In Vivo</italic> Physiology of Visual Cortical Neurons</article-title>. <source>Nature</source> <volume>471</volume>, <fpage>177</fpage>&#x2013;<lpage>182</lpage>. <pub-id pub-id-type="doi">10.1038/nature09802</pub-id> </citation>
</ref>
<ref id="B5">
<citation citation-type="journal">
<person-group person-group-type="author">
<name>
<surname>Briggman</surname>
<given-names>K. L.</given-names>
</name>
<name>
<surname>Bock</surname>
<given-names>D. D.</given-names>
</name>
</person-group> (<year>2012</year>). <article-title>Volume Electron Microscopy for Neuronal Circuit Reconstruction</article-title>. <source>Curr. Opin. Neurobiol.</source> <volume>22</volume>, <fpage>154</fpage>&#x2013;<lpage>161</lpage>. <pub-id pub-id-type="doi">10.1016/j.conb.2011.10.022</pub-id> </citation>
</ref>
<ref id="B6">
<citation citation-type="journal">
<person-group person-group-type="author">
<name>
<surname>Briggman</surname>
<given-names>K. L.</given-names>
</name>
<name>
<surname>Helmstaedter</surname>
<given-names>M.</given-names>
</name>
<name>
<surname>Denk</surname>
<given-names>W.</given-names>
</name>
</person-group> (<year>2011</year>). <article-title>Wiring Specificity in the Direction-Selectivity Circuit of the Retina</article-title>. <source>Nature</source> <volume>471</volume>, <fpage>183</fpage>&#x2013;<lpage>188</lpage>. <pub-id pub-id-type="doi">10.1038/nature09818</pub-id> </citation>
</ref>
<ref id="B7">
<citation citation-type="journal">
<person-group person-group-type="author">
<name>
<surname>Cardona</surname>
<given-names>A.</given-names>
</name>
<name>
<surname>Saalfeld</surname>
<given-names>S.</given-names>
</name>
<name>
<surname>Preibisch</surname>
<given-names>S.</given-names>
</name>
<name>
<surname>Schmid</surname>
<given-names>B.</given-names>
</name>
<name>
<surname>Cheng</surname>
<given-names>A.</given-names>
</name>
<name>
<surname>Pulokas</surname>
<given-names>J.</given-names>
</name>
<etal/>
</person-group> (<year>2010</year>). <article-title>An Integrated Micro- and Macroarchitectural Analysis of the Drosophila Brain by Computer-Assisted Serial Section Electron Microscopy</article-title>. <source>PLoS Biol.</source> <volume>8</volume> (<issue>10</issue>), <fpage>e1000502</fpage>. <pub-id pub-id-type="doi">10.1371/journal.pbio.1000502</pub-id> </citation>
</ref>
<ref id="B8">
<citation citation-type="journal">
<person-group person-group-type="author">
<name>
<surname>Cardona</surname>
<given-names>A.</given-names>
</name>
<name>
<surname>Saalfeld</surname>
<given-names>S.</given-names>
</name>
<name>
<surname>Schindelin</surname>
<given-names>J.</given-names>
</name>
<name>
<surname>Arganda-Carreras</surname>
<given-names>I.</given-names>
</name>
<name>
<surname>Preibisch</surname>
<given-names>S.</given-names>
</name>
<name>
<surname>Longair</surname>
<given-names>M.</given-names>
</name>
<etal/>
</person-group> (<year>2012</year>). <article-title>TrakEM2 Software for Neural Circuit Reconstruction</article-title>. <source>PLoS ONE</source> <volume>7</volume> (<issue>6</issue>), <fpage>e38011</fpage>. <pub-id pub-id-type="doi">10.1371/journal.pone.0038011</pub-id> </citation>
</ref>
<ref id="B9">
<citation citation-type="book">
<person-group person-group-type="author">
<name>
<surname>Chen</surname>
<given-names>H.</given-names>
</name>
<name>
<surname>Qi</surname>
<given-names>X.</given-names>
</name>
<name>
<surname>Cheng</surname>
<given-names>J.</given-names>
</name>
<name>
<surname>Heng</surname>
<given-names>P.</given-names>
</name>
</person-group> (<year>2016</year>). &#x201c;<article-title>Deep Contextual Networks for Neuronal Structure Segmentation</article-title>,&#x201d; in <conf-name>Proceedings of the Thirtieth AAAI Conference on Artificial Intelligence</conf-name>, <conf-loc>Phoenix, AZ</conf-loc>, <conf-date>February 12&#x2013;17, 2016</conf-date> (<publisher-name>AAAI Press</publisher-name>), <fpage>1167</fpage>&#x2013;<lpage>1173</lpage>. </citation>
</ref>
<ref id="B10">
<citation citation-type="journal">
<person-group person-group-type="author">
<name>
<surname>Chen</surname>
<given-names>J.</given-names>
</name>
<name>
<surname>Yang</surname>
<given-names>L.</given-names>
</name>
<name>
<surname>Zhang</surname>
<given-names>Y.</given-names>
</name>
<name>
<surname>Alber</surname>
<given-names>M.</given-names>
</name>
<name>
<surname>Chen</surname>
<given-names>D. Z.</given-names>
</name>
</person-group> (<year>2016</year>). &#x201c;<article-title>Combining Fully Convolutional and Recurrent Neural Networks for 3D Biomedical Image Segmentation</article-title>,&#x201d; in <conf-name>Proceedings of NIPS 2016</conf-name>, <conf-loc>Barcelona, Spain</conf-loc>, <conf-date>September 5, 2016</conf-date> <comment>(Curran Associates, Inc.)</comment>, <fpage>3036</fpage>&#x2013;<lpage>3044</lpage>. </citation>
</ref>
<ref id="B11">
<citation citation-type="web">
<person-group person-group-type="author">
<name>
<surname>Chollet</surname>
<given-names>F.</given-names>
</name>
</person-group> (<year>2015</year>). <article-title>Keras: Deep Learning Library for Theano and Tensorflow</article-title>. <comment>Available at: <ext-link ext-link-type="uri" xlink:href="http://keras.io/">http://keras.io/</ext-link>
</comment> (<comment>Accessed May 20, 2017</comment>). </citation>
</ref>
<ref id="B12">
<citation citation-type="book">
<person-group person-group-type="author">
<name>
<surname>Cicek</surname>
<given-names>O.</given-names>
</name>
<name>
<surname>Abdulkadir</surname>
<given-names>A.</given-names>
</name>
<name>
<surname>Lienkamp</surname>
<given-names>S. S.</given-names>
</name>
<name>
<surname>Brox</surname>
<given-names>T.</given-names>
</name>
<name>
<surname>Ronneberger</surname>
<given-names>O.</given-names>
</name>
</person-group> (<year>2016</year>). &#x201c;<article-title>3D U-Net: Learning Dense Volumetric Segmentation from Sparse Annotation</article-title>,&#x2ee; in <conf-name>Proceedings of MICCAI 2016</conf-name>, <conf-loc>Athens, Greece</conf-loc>, <conf-date>October 17&#x2013;21, 2016</conf-date> (<publisher-name>Springer, Cham</publisher-name>), <fpage>424</fpage>&#x2013;<lpage>432</lpage>. </citation>
</ref>
<ref id="B13">
<citation citation-type="book">
<person-group person-group-type="author">
<name>
<surname>Ciresan</surname>
<given-names>D.</given-names>
</name>
<name>
<surname>Giusti</surname>
<given-names>A.</given-names>
</name>
<name>
<surname>Gambardella</surname>
<given-names>L. M.</given-names>
</name>
<name>
<surname>Schmidhuber</surname>
<given-names>J.</given-names>
</name>
</person-group> (<year>2012</year>). &#x201c;<article-title>Deep Neural Networks Segment Neuronal Membranes in Electron Microscopy Images</article-title>,&#x2ee; in <conf-name>Proceedings of NIPS 2012</conf-name>, <conf-loc>Stateline, NV</conf-loc>, <conf-date>December 3&#x2013;8, 2012</conf-date> (<publisher-name>Curran Associates Inc.</publisher-name>), <fpage>2843</fpage>&#x2013;<lpage>2851</lpage>. </citation>
</ref>
<ref id="B14">
<citation citation-type="book">
<person-group person-group-type="author">
<name>
<surname>Drozdzal</surname>
<given-names>M.</given-names>
</name>
<name>
<surname>Vorontsov</surname>
<given-names>E.</given-names>
</name>
<name>
<surname>Chartrand</surname>
<given-names>G.</given-names>
</name>
<name>
<surname>Kadoury</surname>
<given-names>S.</given-names>
</name>
<name>
<surname>Pal</surname>
<given-names>C.</given-names>
</name>
</person-group> (<year>2016</year>). &#x201c;<article-title>The Importance of Skip Connections in Biomedical Image Segmentation</article-title>,&#x2ee; in <conf-name>Proceedings of DLMIA 2016</conf-name>, <conf-loc>Athens, Greece</conf-loc>, <conf-date>October 21, 2016</conf-date> (<publisher-name>Springer, Cham</publisher-name>), <fpage>179</fpage>&#x2013;<lpage>187</lpage>. </citation>
</ref>
<ref id="B15">
<citation citation-type="journal">
<person-group person-group-type="author">
<name>
<surname>Eberle</surname>
<given-names>A. L.</given-names>
</name>
<name>
<surname>Zeidler</surname>
<given-names>D.</given-names>
</name>
</person-group> (<year>2018</year>). <article-title>Multi-Beam Scanning Electron Microscopy for High-Throughput Imaging in Connectomics Research</article-title>. <source>Front. Neuroanat.</source> <volume>12</volume>, <fpage>112</fpage>. <pub-id pub-id-type="doi">10.3389/fnana.2018.00112</pub-id> </citation>
</ref>
<ref id="B16">
<citation citation-type="book">
<person-group person-group-type="author">
<name>
<surname>Ensafi</surname>
<given-names>S.</given-names>
</name>
<name>
<surname>Lu</surname>
<given-names>S.</given-names>
</name>
<name>
<surname>Kassim</surname>
<given-names>A. A.</given-names>
</name>
<name>
<surname>Tan</surname>
<given-names>C. L.</given-names>
</name>
</person-group> (<year>2014</year>). &#x201c;<article-title>3D Reconstruction of Neurons in Electron Microscopy Images</article-title>,&#x201d; in <conf-name>Proceedings of IEEE EMBS 2014</conf-name>, <conf-loc>Chicago, Illinois</conf-loc>, <conf-date>August 27&#x2013;31, 2014</conf-date> (<publisher-name>IEEE</publisher-name>), <fpage>6732</fpage>&#x2013;<lpage>6735</lpage>. </citation>
</ref>
<ref id="B17">
<citation citation-type="journal">
<person-group person-group-type="author">
<name>
<surname>Fakhry</surname>
<given-names>A.</given-names>
</name>
<name>
<surname>Peng</surname>
<given-names>H.</given-names>
</name>
<name>
<surname>Ji</surname>
<given-names>S.</given-names>
</name>
</person-group> (<year>2016</year>). <article-title>Deep Models for Brain EM Image Segmentation: Novel Insights and Improved Performance</article-title>. <source>Bioinformatics</source> <volume>32</volume>, <fpage>2352</fpage>&#x2013;<lpage>2358</lpage>. <pub-id pub-id-type="doi">10.1093/bioinformatics/btw165</pub-id> </citation>
</ref>
<ref id="B18">
<citation citation-type="journal">
<person-group person-group-type="author">
<name>
<surname>Fakhry</surname>
<given-names>A.</given-names>
</name>
<name>
<surname>Zeng</surname>
<given-names>T.</given-names>
</name>
<name>
<surname>Ji</surname>
<given-names>S.</given-names>
</name>
</person-group> (<year>2017</year>). <article-title>Residual Deconvolutional Networks for Brain Electron Microscopy Image Segmentation</article-title>. <source>IEEE Trans. Med. Imaging</source> <volume>36</volume>, <fpage>447</fpage>&#x2013;<lpage>456</lpage>. <pub-id pub-id-type="doi">10.1109/tmi.2016.2613019</pub-id> </citation>
</ref>
<ref id="B20">
<citation citation-type="book">
<person-group person-group-type="author">
<name>
<surname>Gatys</surname>
<given-names>L. A.</given-names>
</name>
<name>
<surname>Ecker</surname>
<given-names>A. S.</given-names>
</name>
<name>
<surname>Bethge</surname>
<given-names>M.</given-names>
</name>
</person-group> (<year>2016</year>). &#x201c;<article-title>Image Style Transfer Using Convolutional Neural Networks</article-title>,&#x2ee; in <conf-name>Proceedings of IEEE CVPR 2016</conf-name>, <conf-loc>Las Vegas, NV</conf-loc>, <conf-date>June 27&#x2013;30, 2016</conf-date> (<publisher-name>IEEE</publisher-name>), <fpage>2414</fpage>&#x2013;<lpage>2423</lpage>. </citation>
</ref>
<ref id="B21">
<citation citation-type="web">
<person-group person-group-type="author">
<name>
<surname>Graham</surname>
<given-names>B. J.</given-names>
</name>
<name>
<surname>Hildebrand</surname>
<given-names>D. G. C.</given-names>
</name>
<name>
<surname>Kuan</surname>
<given-names>A. T.</given-names>
</name>
<name>
<surname>Maniates-Selvin</surname>
<given-names>J.&#x20;T.</given-names>
</name>
<name>
<surname>Thomas</surname>
<given-names>L. A.</given-names>
</name>
<name>
<surname>Shanny</surname>
<given-names>B. L.</given-names>
</name>
<etal/>
</person-group> (<year>2019</year>). <article-title>High-throughput Transmission Electron Microscopy With Automated Serial Sectioning</article-title>. <pub-id pub-id-type="doi">10.1101/657346</pub-id> <comment>Preprint. Available at: <ext-link ext-link-type="uri" xlink:href="https://doi.org/10.1101/657346">https://doi.org/10.1101/657346</ext-link>
</comment> (<comment>Accessed June 2, 2019</comment>). </citation>
</ref>
<ref id="B22">
<citation citation-type="journal">
<person-group person-group-type="author">
<name>
<surname>Hayworth</surname>
<given-names>K. J.</given-names>
</name>
<name>
<surname>Morgan</surname>
<given-names>J.&#x20;L.</given-names>
</name>
<name>
<surname>Schalek</surname>
<given-names>R.</given-names>
</name>
<name>
<surname>Berger</surname>
<given-names>D. R.</given-names>
</name>
<name>
<surname>Hildebrand</surname>
<given-names>D. G. C.</given-names>
</name>
<name>
<surname>Lichtman</surname>
<given-names>J.&#x20;W.</given-names>
</name>
</person-group> (<year>2014</year>). <article-title>Imaging ATUM Ultrathin Section Libraries With WaferMapper: A Multi-Scale Approach to EM Reconstruction of Neural Circuits</article-title>. <source>Front. Neural Circuits</source> <volume>8</volume>, <fpage>68</fpage>. <pub-id pub-id-type="doi">10.3389/fncir.2014.00068</pub-id> </citation>
</ref>
<ref id="B23">
<citation citation-type="book">
<person-group person-group-type="author">
<name>
<surname>He</surname>
<given-names>K.</given-names>
</name>
<name>
<surname>Zhang</surname>
<given-names>X.</given-names>
</name>
<name>
<surname>Ren</surname>
<given-names>S.</given-names>
</name>
<name>
<surname>Sun</surname>
<given-names>J.</given-names>
</name>
</person-group> (<year>2016</year>). &#x201c;<article-title>Deep Residual Learning for Image Recognition</article-title>,&#x2ee; in <conf-name>Conference on Computer Vision and Pattern Recognition (CVPR)</conf-name>, <conf-loc>Las Vegas, NV</conf-loc>, <conf-date>June 27&#x2013;30, 2016</conf-date> (<publisher-name>IEEE</publisher-name>), <fpage>770</fpage>&#x2013;<lpage>778</lpage>. </citation>
</ref>
<ref id="B24">
<citation citation-type="journal">
<person-group person-group-type="author">
<name>
<surname>Helmstaedter</surname>
<given-names>M.</given-names>
</name>
</person-group> (<year>2013</year>). <article-title>Cellular-Resolution Connectomics: Challenges of Dense Neural Circuit Reconstruction</article-title>. <source>Nat. Methods</source> <volume>10</volume>, <fpage>501</fpage>&#x2013;<lpage>507</lpage>. <pub-id pub-id-type="doi">10.1038/nmeth.2476</pub-id> </citation>
</ref>
<ref id="B25">
<citation citation-type="book">
<person-group person-group-type="author">
<name>
<surname>Hildebrand</surname>
<given-names>D. G. C.</given-names>
</name>
</person-group> (<year>2015</year>). <article-title>Whole-Brain Functional and Structural Examination in Larval Zebrafish</article-title>. <comment>PhD thesis</comment>. <publisher-loc>Cambridge (MA)</publisher-loc>: <publisher-name>Harvard University, Graduate School of Arts and Sciences</publisher-name>. </citation>
</ref>
<ref id="B26">
<citation citation-type="journal">
<person-group person-group-type="author">
<name>
<surname>Hildebrand</surname>
<given-names>D. G. C.</given-names>
</name>
<name>
<surname>Cicconet</surname>
<given-names>M.</given-names>
</name>
<name>
<surname>Torres</surname>
<given-names>R. M.</given-names>
</name>
<name>
<surname>Choi</surname>
<given-names>W.</given-names>
</name>
<name>
<surname>Quan</surname>
<given-names>T. M.</given-names>
</name>
<name>
<surname>Moon</surname>
<given-names>J.</given-names>
</name>
<etal/>
</person-group> (<year>2017</year>). <article-title>Whole-Brain Serial-Section Electron Microscopy in Larval Zebrafish</article-title>. <source>Nature</source> <volume>545</volume>, <fpage>345</fpage>&#x2013;<lpage>349</lpage>. <pub-id pub-id-type="doi">10.1038/nature22356</pub-id> </citation>
</ref>
<ref id="B27">
<citation citation-type="book">
<person-group person-group-type="author">
<name>
<surname>Hirsch</surname>
<given-names>P.</given-names>
</name>
<name>
<surname>Mais</surname>
<given-names>L.</given-names>
</name>
<name>
<surname>Kainmueller</surname>
<given-names>D.</given-names>
</name>
</person-group> (<year>2020</year>). &#x201c;<article-title>Patchperpix for Instance Segmentation</article-title>,&#x201d; in <conf-name>Proceedings of ECCV 2020</conf-name>, <conf-loc>Glasgow, United&#x20;Kingdom</conf-loc>, <conf-date>August 23&#x2013;28, 2020</conf-date> (<publisher-name>Springer, Cham</publisher-name>), <fpage>288</fpage>&#x2013;<lpage>304</lpage>. </citation>
</ref>
<ref id="B28">
<citation citation-type="journal">
<person-group person-group-type="author">
<name>
<surname>Isin</surname>
<given-names>A.</given-names>
</name>
<name>
<surname>Direkoglu</surname>
<given-names>C.</given-names>
</name>
<name>
<surname>Sah</surname>
<given-names>M.</given-names>
</name>
</person-group> (<year>2016</year>). <article-title>Review of MRI-Based Brain Tumor Image Segmentation Using Deep Learning Methods</article-title>. <source>Procedia Comput. Sci.</source> <volume>102</volume>, <fpage>317</fpage>&#x2013;<lpage>324</lpage>. <pub-id pub-id-type="doi">10.1016/j.procs.2016.09.407</pub-id> </citation>
</ref>
<ref id="B29">
<citation citation-type="journal">
<person-group person-group-type="author">
<name>
<surname>Jeong</surname>
<given-names>W.-K.</given-names>
</name>
<name>
<surname>Beyer</surname>
<given-names>J.</given-names>
</name>
<name>
<surname>Hadwiger</surname>
<given-names>M.</given-names>
</name>
<name>
<surname>Blue</surname>
<given-names>R.</given-names>
</name>
<name>
<surname>Law</surname>
<given-names>C.</given-names>
</name>
<name>
<surname>V&#xe1;zquez-Reina</surname>
<given-names>A.</given-names>
</name>
<etal/>
</person-group> (<year>2010</year>). <article-title>Secrett and NeuroTrace: Interactive Visualization and Analysis Tools for Large-Scale Neuroscience Data Sets</article-title>. <source>IEEE Comput. Graphics Appl.</source> <volume>30</volume>, <fpage>58</fpage>&#x2013;<lpage>70</lpage>. <pub-id pub-id-type="doi">10.1109/MCG.2010.56</pub-id> </citation>
</ref>
<ref id="B30">
<citation citation-type="journal">
<person-group person-group-type="author">
<name>
<surname>Kaynig</surname>
<given-names>V.</given-names>
</name>
<name>
<surname>Vazquez-Reina</surname>
<given-names>A.</given-names>
</name>
<name>
<surname>Knowles-Barley</surname>
<given-names>S.</given-names>
</name>
<name>
<surname>Roberts</surname>
<given-names>M.</given-names>
</name>
<name>
<surname>Jones</surname>
<given-names>T. R.</given-names>
</name>
<name>
<surname>Kasthuri</surname>
<given-names>N.</given-names>
</name>
<etal/>
</person-group> (<year>2015</year>). <article-title>Large-scale Automatic Reconstruction of Neuronal Processes From Electron Microscopy Images</article-title>. <source>Med. Image Anal.</source> <volume>22</volume>, <fpage>77</fpage>&#x2013;<lpage>88</lpage>. <pub-id pub-id-type="doi">10.1016/j.media.2015.02.001</pub-id> </citation>
</ref>
<ref id="B31">
<citation citation-type="book">
<person-group person-group-type="author">
<name>
<surname>Krizhevsky</surname>
<given-names>A.</given-names>
</name>
<name>
<surname>Sutskever</surname>
<given-names>I.</given-names>
</name>
<name>
<surname>Hinton</surname>
<given-names>G. E.</given-names>
</name>
</person-group> (<year>2012</year>). &#x201c;<article-title>ImageNet Classification With Deep Convolutional Neural Networks</article-title>, in <conf-name>Proceedings of NIPS 2012</conf-name>, <conf-loc>Stateline, NV</conf-loc>, <conf-date>December 3&#x2013;8, 2012</conf-date> (<publisher-name>Curran Associates Inc.</publisher-name>), <fpage>1097</fpage>&#x2013;<lpage>1105</lpage>. </citation>
</ref>
<ref id="B32">
<citation citation-type="journal">
<person-group person-group-type="author">
<name>
<surname>LeCun</surname>
<given-names>Y.</given-names>
</name>
<name>
<surname>Bengio</surname>
<given-names>Y.</given-names>
</name>
<name>
<surname>Hinton</surname>
<given-names>G.</given-names>
</name>
</person-group> (<year>2015</year>). <article-title>Deep Learning</article-title>. <source>Nature</source> <volume>521</volume>, <fpage>436</fpage>&#x2013;<lpage>444</lpage>. <pub-id pub-id-type="doi">10.1038/nature14539</pub-id> </citation>
</ref>
<ref id="B33">
<citation citation-type="book">
<person-group person-group-type="author">
<name>
<surname>Lee</surname>
<given-names>G.</given-names>
</name>
<name>
<surname>Oh</surname>
<given-names>J.-W.</given-names>
</name>
<name>
<surname>Kang</surname>
<given-names>M.-S.</given-names>
</name>
<name>
<surname>Her</surname>
<given-names>N.-G.</given-names>
</name>
<name>
<surname>Kim</surname>
<given-names>M.-H.</given-names>
</name>
<name>
<surname>Jeong</surname>
<given-names>W.-K.</given-names>
</name>
</person-group> (<year>2018</year>). &#x201c;<article-title>DeepHCS: Bright-Field to Fluorescence Microscopy Image Conversion Using Deep Learning for Label-Free High-Content Screening</article-title>, &#x2ee; in <conf-name>Medical Image Computing and Computer Assisted Intervention&#x2013;MICCAI 2018</conf-name>, <conf-loc>Granada, Spain</conf-loc>, <conf-date>September 16&#x2013;20, 2018</conf-date> (<publisher-name>Springer International Publishing</publisher-name>), <fpage>335</fpage>&#x2013;<lpage>343</lpage>. </citation>
</ref>
<ref id="B34">
<citation citation-type="journal">
<person-group person-group-type="author">
<name>
<surname>Lichtman</surname>
<given-names>J.&#x20;W.</given-names>
</name>
<name>
<surname>Denk</surname>
<given-names>W.</given-names>
</name>
</person-group> (<year>2011</year>). <article-title>The Big and the Small: Challenges of Imaging the Brain&#x2019;s Circuits</article-title>. <source>Science</source> <volume>334</volume>, <fpage>618</fpage>&#x2013;<lpage>623</lpage>. <pub-id pub-id-type="doi">10.1126/science.1209168</pub-id> </citation>
</ref>
<ref id="B35">
<citation citation-type="journal">
<person-group person-group-type="author">
<name>
<surname>Lin</surname>
<given-names>M.</given-names>
</name>
<name>
<surname>Chen</surname>
<given-names>Q.</given-names>
</name>
<name>
<surname>Yan</surname>
<given-names>S.</given-names>
</name>
</person-group> (<year>2014</year>). &#x201c;<article-title>Network in Network</article-title>, &#x2ee; in <conf-name>Proceedings of ICLR 2014</conf-name>. <comment>arXiv:1312.4400v3</comment>. </citation>
</ref>
<ref id="B36">
<citation citation-type="book">
<person-group person-group-type="author">
<name>
<surname>Long</surname>
<given-names>J.</given-names>
</name>
<name>
<surname>Shelhamer</surname>
<given-names>E.</given-names>
</name>
<name>
<surname>Darrell</surname>
<given-names>T.</given-names>
</name>
</person-group> (<year>2015</year>). &#x201c;<article-title>Fully Convolutional Networks for Semantic Segmentation</article-title>,&#x2ee; in <conf-name>Proceedings of IEEE CVPR 2015</conf-name>, <conf-loc>Boston, MA</conf-loc>, <conf-date>June 7&#x2013;12, 2015</conf-date> (<publisher-name>IEEE</publisher-name>), <fpage>3431</fpage>&#x2013;<lpage>3440</lpage>. </citation>
</ref>
<ref id="B37">
<citation citation-type="journal">
<person-group person-group-type="author">
<name>
<surname>Pape</surname>
<given-names>C.</given-names>
</name>
<name>
<surname>Matskevych</surname>
<given-names>A.</given-names>
</name>
<name>
<surname>Wolny</surname>
<given-names>A.</given-names>
</name>
<name>
<surname>Hennies</surname>
<given-names>J.</given-names>
</name>
<name>
<surname>Mizzon</surname>
<given-names>G.</given-names>
</name>
<name>
<surname>Louveaux</surname>
<given-names>M.</given-names>
</name>
<etal/>
</person-group> (<year>2019</year>). <article-title>Leveraging Domain Knowledge to Improve Microscopy Image Segmentation with Lifted Multicuts</article-title>. <source>Front. Comput. Sci.</source> <volume>1</volume>, <fpage>657</fpage>. <pub-id pub-id-type="doi">10.3389/fcomp.2019.00006</pub-id> </citation>
</ref>
<ref id="B38">
<citation citation-type="book">
<person-group person-group-type="author">
<name>
<surname>Parag</surname>
<given-names>T.</given-names>
</name>
<name>
<surname>Ciresan</surname>
<given-names>D. C.</given-names>
</name>
<name>
<surname>Giusti</surname>
<given-names>A.</given-names>
</name>
</person-group> (<year>2015</year>). &#x201c;<article-title>Efficient Classifier Training to Minimize False Merges in Electron Microscopy Segmentation</article-title>,&#x2ee; in <conf-name>Proceedings of IEEE ICCV 2015</conf-name>, <conf-loc>Santiago, Chile</conf-loc>, <conf-date>December 7&#x2013;13, 2015</conf-date> (<publisher-name>IEEE</publisher-name>), <fpage>657</fpage>&#x2013;<lpage>665</lpage>. </citation>
</ref>
<ref id="B39">
<citation citation-type="journal">
<person-group person-group-type="author">
<name>
<surname>Quan</surname>
<given-names>T. M.</given-names>
</name>
<name>
<surname>Hildebrand</surname>
<given-names>D. G. C.</given-names>
</name>
<name>
<surname>Jeong</surname>
<given-names>W.-K.</given-names>
</name>
</person-group> (<year>2016</year>). <article-title>FusionNet: A Deep Fully Residual Convolutional Neural Network for Image Segmentation in Connectomics</article-title>. <comment>arXiv preprint arXiv:1612.05360</comment>. </citation>
</ref>
<ref id="B40">
<citation citation-type="journal">
<person-group person-group-type="author">
<name>
<surname>Quan</surname>
<given-names>T. M.</given-names>
</name>
<name>
<surname>Nguyen-Duc</surname>
<given-names>T.</given-names>
</name>
<name>
<surname>Jeong</surname>
<given-names>W.-K.</given-names>
</name>
</person-group> (<year>2018</year>). <article-title>Compressed Sensing MRI Reconstruction Using a Generative Adversarial Network With a Cyclic Loss</article-title>. <source>IEEE Trans. Med. Imaging</source> <volume>37</volume>, <fpage>1488</fpage>&#x2013;<lpage>1497</lpage>. <pub-id pub-id-type="doi">10.1109/tmi.2018.2820120</pub-id> </citation>
</ref>
<ref id="B41">
<citation citation-type="book">
<person-group person-group-type="author">
<name>
<surname>Ronneberger</surname>
<given-names>O.</given-names>
</name>
<name>
<surname>Fischer</surname>
<given-names>P.</given-names>
</name>
<name>
<surname>Brox</surname>
<given-names>T.</given-names>
</name>
</person-group> (<year>2015</year>). &#x201c;<article-title>U-Net: Convolutional Networks for Biomedical Image Segmentation</article-title>. &#x2ee;in <conf-name>Proceedings of MICCAI 2015</conf-name>, <conf-loc>Munich, Germany</conf-loc>, <conf-date>October 5&#x2013;9, 2015</conf-date> (<publisher-name>Springer, Cham</publisher-name>), <fpage>234</fpage>&#x2013;<lpage>241</lpage>. </citation>
</ref>
<ref id="B42">
<citation citation-type="journal">
<person-group person-group-type="author">
<name>
<surname>Russakovsky</surname>
<given-names>O.</given-names>
</name>
<name>
<surname>Deng</surname>
<given-names>J.</given-names>
</name>
<name>
<surname>Su</surname>
<given-names>H.</given-names>
</name>
<name>
<surname>Krause</surname>
<given-names>J.</given-names>
</name>
<name>
<surname>Satheesh</surname>
<given-names>S.</given-names>
</name>
<name>
<surname>Ma</surname>
<given-names>S.</given-names>
</name>
<etal/>
</person-group> (<year>2015</year>). <article-title>ImageNet Large Scale Visual Recognition Challenge</article-title>. <source>Int. J.&#x20;Comput. Vision</source> <volume>115</volume>, <fpage>211</fpage>&#x2013;<lpage>252</lpage>. <pub-id pub-id-type="doi">10.1007/s11263-015-0816-y</pub-id> </citation>
</ref>
<ref id="B43">
<citation citation-type="book">
<person-group person-group-type="editor">
<name>
<surname>Shapira</surname>
<given-names>Y.</given-names>
</name>
</person-group> (Editor) (<year>2008</year>). <source>Matrix-Based Multigrid</source>. <publisher-loc>New York, NY</publisher-loc>: <publisher-name>Springer US</publisher-name>.</citation>
</ref>
<ref id="B44">
<citation citation-type="journal">
<person-group person-group-type="author">
<name>
<surname>Schindelin</surname>
<given-names>J.</given-names>
</name>
<name>
<surname>Arganda-Carreras</surname>
<given-names>I.</given-names>
</name>
<name>
<surname>Frise</surname>
<given-names>E.</given-names>
</name>
<name>
<surname>Kaynig</surname>
<given-names>V.</given-names>
</name>
<name>
<surname>Longair</surname>
<given-names>M.</given-names>
</name>
<name>
<surname>Pietzsch</surname>
<given-names>T.</given-names>
</name>
<etal/>
</person-group> (<year>2012</year>). <article-title>Fiji: an Open-Source Platform for Biological-Image Analysis</article-title>. <source>Nat. Methods</source> <volume>9</volume>, <fpage>676</fpage>&#x2013;<lpage>682</lpage>. <pub-id pub-id-type="doi">10.1038/nmeth.2019</pub-id> </citation>
</ref>
<ref id="B45">
<citation citation-type="book">
<person-group person-group-type="author">
<name>
<surname>Shen</surname>
<given-names>W.</given-names>
</name>
<name>
<surname>Wang</surname>
<given-names>B.</given-names>
</name>
<name>
<surname>Jiang</surname>
<given-names>Y.</given-names>
</name>
<name>
<surname>Wang</surname>
<given-names>Y.</given-names>
</name>
<name>
<surname>Yuille</surname>
<given-names>A.</given-names>
</name>
</person-group> (<year>2017</year>). &#x201c;<article-title>Multi-stage Multi-Recursive-Input Fully Convolutional Networks for Neuronal Boundary Detection</article-title>,&#x2ee; in <conf-name>IEEE International Conference on Computer Vision (ICCV)</conf-name>, <conf-loc>Venice, Italy</conf-loc>, <conf-date>October 22&#x2013;29, 2017</conf-date> (<publisher-name>IEEE</publisher-name>), <fpage>2391</fpage>&#x2013;<lpage>2400</lpage>. </citation>
</ref>
<ref id="B46">
<citation citation-type="journal">
<person-group person-group-type="author">
<name>
<surname>Simonyan</surname>
<given-names>K.</given-names>
</name>
<name>
<surname>Zisserman</surname>
<given-names>A.</given-names>
</name>
</person-group> (<year>2014</year>). <article-title>Very Deep Convolutional Networks for Large-Scale Image Recognition</article-title>. <comment>arXiv preprint arXiv:1409.1556</comment> </citation>
</ref>
<ref id="B47">
<citation citation-type="book">
<person-group person-group-type="author">
<name>
<surname>Sommer</surname>
<given-names>C.</given-names>
</name>
<name>
<surname>Str&#xe4;hle</surname>
<given-names>C.</given-names>
</name>
<name>
<surname>K&#xf6;the</surname>
<given-names>U.</given-names>
</name>
<name>
<surname>Hamprecht</surname>
<given-names>F. A.</given-names>
</name>
</person-group> (<year>2011</year>). &#x201c;<article-title>Ilastik: Interactive Learning and Segmentation Toolkit</article-title>,&#x2ee; in <conf-name>Proceedings of IEEE ISBI 2011</conf-name>, <conf-loc>Chicago, IL</conf-loc>, <conf-date>March 30&#x2013;April 2, 2011</conf-date> (<publisher-name>IEEE</publisher-name>), <fpage>230</fpage>&#x2013;<lpage>233</lpage>. </citation>
</ref>
<ref id="B48">
<citation citation-type="journal">
<person-group person-group-type="author">
<name>
<surname>Stollenga</surname>
<given-names>M. F.</given-names>
</name>
<name>
<surname>Byeon</surname>
<given-names>W.</given-names>
</name>
<name>
<surname>Liwicki</surname>
<given-names>M.</given-names>
</name>
<name>
<surname>Schmidhuber</surname>
<given-names>J.</given-names>
</name>
</person-group> (<year>2015</year>). &#x201c;<article-title>Parallel Multi-Dimensional LSTM, with Application to Fast Biomedical Volumetric Image Segmentation</article-title>,&#x2ee; in <conf-name>Proceedings of NIPS 2015</conf-name>, <conf-date>June 24, 2015</conf-date>, <conf-loc>Montreal, QC</conf-loc> <comment>(Curran Associates, Inc.)</comment>, <fpage>2998</fpage>&#x2013;<lpage>3006</lpage>. </citation>
</ref>
<ref id="B49">
<citation citation-type="book">
<person-group person-group-type="author">
<name>
<surname>Szegedy</surname>
<given-names>C.</given-names>
</name>
<name>
<surname>Liu</surname>
<given-names>W.</given-names>
</name>
<name>
<surname>Jia</surname>
<given-names>Y.</given-names>
</name>
<name>
<surname>Sermanet</surname>
<given-names>P.</given-names>
</name>
<name>
<surname>Reed</surname>
<given-names>S.</given-names>
</name>
<name>
<surname>Anguelov</surname>
<given-names>D.</given-names>
</name>
<etal/>
</person-group> (<year>2015</year>). &#x201c;<article-title>Going Deeper with Convolutions</article-title>,&#x2ee; in <conf-name>Proceedings of IEEE CVPR</conf-name>, <conf-loc>Boston, MA</conf-loc>, <conf-date>June 7&#x2013;12, 2015</conf-date> (<publisher-name>IEEE</publisher-name>) <fpage>1</fpage>&#x2013;<lpage>9</lpage>. </citation>
</ref>
<ref id="B50">
<citation citation-type="journal">
<person-group person-group-type="author">
<name>
<surname>Turaga</surname>
<given-names>S. C.</given-names>
</name>
<name>
<surname>Murray</surname>
<given-names>J.&#x20;F.</given-names>
</name>
<name>
<surname>Jain</surname>
<given-names>V.</given-names>
</name>
<name>
<surname>Roth</surname>
<given-names>F.</given-names>
</name>
<name>
<surname>Helmstaedter</surname>
<given-names>M.</given-names>
</name>
<name>
<surname>Briggman</surname>
<given-names>K.</given-names>
</name>
<etal/>
</person-group> (<year>2010</year>). <article-title>Convolutional Networks can Learn to Generate Affinity Graphs for Image Segmentation</article-title>. <source>Neural Comput.</source> <volume>22</volume>, <fpage>511</fpage>&#x2013;<lpage>538</lpage>. <pub-id pub-id-type="doi">10.1162/neco.2009.10-08-881</pub-id> </citation>
</ref>
<ref id="B51">
<citation citation-type="book">
<person-group person-group-type="author">
<name>
<surname>Weiler</surname>
<given-names>M.</given-names>
</name>
<name>
<surname>Hamprecht</surname>
<given-names>F. A.</given-names>
</name>
<name>
<surname>Storath</surname>
<given-names>M.</given-names>
</name>
</person-group> (<year>2017</year>). &#x201c;<article-title>Learning Steerable Filters for Rotation Equivariant cnns</article-title>,&#x2ee; in <conf-name>Computer Vision and Pattern Recognition</conf-name>, <conf-loc>Honolulu, HI</conf-loc>, <conf-date>July 21&#x2013;26, 2017</conf-date> (<publisher-name>IEEE</publisher-name>), <fpage>849</fpage>&#x2013;<lpage>858</lpage>. </citation>
</ref>
<ref id="B52">
<citation citation-type="book">
<person-group person-group-type="author">
<name>
<surname>Wetzel</surname>
<given-names>A. W.</given-names>
</name>
<name>
<surname>Bakal</surname>
<given-names>J.</given-names>
</name>
<name>
<surname>Dittrich</surname>
<given-names>M.</given-names>
</name>
<name>
<surname>Hildebrand</surname>
<given-names>D. G. C.</given-names>
</name>
<name>
<surname>Morgan</surname>
<given-names>J.&#x20;L.</given-names>
</name>
<name>
<surname>Lichtman</surname>
<given-names>J.&#x20;W.</given-names>
</name>
</person-group> (<year>2016</year>). &#x201c;<article-title>Registering Large Volume Serial-Section Electron Microscopy Image Sets for Neural Circuit Reconstruction Using FFT Signal Whitening</article-title>,&#x2ee; in <conf-name>Proceedings of AIPR Workshop 2016</conf-name>, <conf-loc>Washington, D.C., United&#x20;States</conf-loc>, <conf-date>October 18&#x2013;20, 2016</conf-date> (<publisher-name>IEEE</publisher-name>), <fpage>1</fpage>&#x2013;<lpage>10</lpage>. </citation>
</ref>
<ref id="B53">
<citation citation-type="book">
<person-group person-group-type="author">
<name>
<surname>Wiehman</surname>
<given-names>S.</given-names>
</name>
<name>
<surname>Villiers</surname>
<given-names>H. D.</given-names>
</name>
</person-group> (<year>2016</year>). &#x201c;<article-title>Semantic Segmentation of Bioimages Using Convolutional Neural Networks</article-title>,&#x2ee; in <conf-name>Proceedings of IJCNN 2016</conf-name>, <conf-loc>Vancouver, BC</conf-loc>, <conf-date>July 24&#x2013;29, 2016</conf-date> (<publisher-name>IEEE</publisher-name>), <fpage>624</fpage>&#x2013;<lpage>631</lpage>. </citation>
</ref>
<ref id="B54">
<citation citation-type="journal">
<person-group person-group-type="author">
<name>
<surname>Wolf</surname>
<given-names>S.</given-names>
</name>
<name>
<surname>Bailoni</surname>
<given-names>A.</given-names>
</name>
<name>
<surname>Pape</surname>
<given-names>C.</given-names>
</name>
<name>
<surname>Rahaman</surname>
<given-names>N.</given-names>
</name>
<name>
<surname>Kreshuk</surname>
<given-names>A.</given-names>
</name>
<name>
<surname>K&#xf6;the</surname>
<given-names>U.</given-names>
</name>
<etal/>
</person-group> (<year>2019</year>). <article-title>The Mutex Watershed and its Objective: Efficient, Parameter-Free Image Partitioning</article-title>. <source>IEEE Trans. Pattern Anal. Mach. Intell.</source> <pub-id pub-id-type="doi">10.1109/TPAMI.2020.2980827</pub-id> </citation>
</ref>
<ref id="B55">
<citation citation-type="journal">
<person-group person-group-type="author">
<name>
<surname>Wu</surname>
<given-names>X.</given-names>
</name>
</person-group> (<year>2015</year>). <article-title>An Iterative Convolutional Neural Network Algorithm Improves Electron Microscopy Image Segmentation</article-title>. <comment>arXiv preprint arXiv:150605849</comment>. </citation>
</ref>
<ref id="B56">
<citation citation-type="book">
<person-group person-group-type="author">
<name>
<surname>Xiao</surname>
<given-names>C.</given-names>
</name>
<name>
<surname>Liu</surname>
<given-names>J.</given-names>
</name>
<name>
<surname>Chen</surname>
<given-names>X.</given-names>
</name>
<name>
<surname>Han</surname>
<given-names>H.</given-names>
</name>
<name>
<surname>Shu</surname>
<given-names>C.</given-names>
</name>
<name>
<surname>Xie</surname>
<given-names>Q.</given-names>
</name>
</person-group> (<year>2018</year>). &#x201c;<article-title>Deep Contextual Residual Network for Electron Microscopy Image Segmentation in Connectomics</article-title>,&#x2ee; in <conf-name>IEEE 15th International Symposium On Biomedical Imaging (ISBI 2018)</conf-name>, <conf-loc>Washington, D.C., United&#x20;States</conf-loc>, <conf-date>April 4&#x2013;7, 2018</conf-date> (<publisher-name>IEEE</publisher-name>), <fpage>378</fpage>&#x2013;<lpage>381</lpage>. </citation>
</ref>
<ref id="B57">
<citation citation-type="book">
<person-group person-group-type="author">
<name>
<surname>Zeiler</surname>
<given-names>M. D.</given-names>
</name>
<name>
<surname>Fergus</surname>
<given-names>R.</given-names>
</name>
</person-group> (<year>2014</year>). &#x201c;<article-title>Visualizing and Understanding Convolutional Networks</article-title>,&#x2ee; in <conf-name>Proceedings of ECCV 2014</conf-name>, <conf-loc>Zurich, Switzerland</conf-loc>, <conf-date>September 6&#x2013;12, 2014</conf-date> (<publisher-name>Springer, Cham</publisher-name>), <fpage>818</fpage>&#x2013;<lpage>833</lpage>. </citation>
</ref>
<ref id="B58">
<citation citation-type="book">
<person-group person-group-type="author">
<name>
<surname>Zheng</surname>
<given-names>Y.</given-names>
</name>
<name>
<surname>Liu</surname>
<given-names>D.</given-names>
</name>
<name>
<surname>Georgescu</surname>
<given-names>B.</given-names>
</name>
<name>
<surname>Nguyen</surname>
<given-names>H.</given-names>
</name>
<name>
<surname>Comaniciu</surname>
<given-names>D.</given-names>
</name>
</person-group> (<year>2015</year>). &#x201c;<article-title>3D Deep Learning for Efficient and Robust Landmark Detection in Volumetric Data</article-title>,&#x2ee; in <conf-name>Proceedings of MICCAI 2015</conf-name>, <conf-loc>Munich, Germany</conf-loc>, <conf-date>October 5&#x2013;9, 2015</conf-date> (<publisher-name>Springer, Cham</publisher-name>), <fpage>565</fpage>&#x2013;<lpage>572</lpage>. </citation>
</ref>
<ref id="B59">
<citation citation-type="journal">
<person-group person-group-type="author">
<name>
<surname>Zheng</surname>
<given-names>Z.</given-names>
</name>
<name>
<surname>Lauritzen</surname>
<given-names>J.&#x20;S.</given-names>
</name>
<name>
<surname>Perlman</surname>
<given-names>E.</given-names>
</name>
<name>
<surname>Robinson</surname>
<given-names>C. G.</given-names>
</name>
<name>
<surname>Nichols</surname>
<given-names>M.</given-names>
</name>
<name>
<surname>Milkie</surname>
<given-names>D.</given-names>
</name>
<etal/>
</person-group> (<year>2018</year>). <article-title>A Complete Electron Microscopy Volume of the Brain of Adult drosophila Melanogaster</article-title>. <source>Cell</source> <volume>174</volume>, <fpage>730</fpage>&#x2013;<lpage>743</lpage>. <pub-id pub-id-type="doi">10.1016/j.cell.2018.06.019</pub-id>
<comment>e22</comment> </citation>
</ref>
<ref id="B60">
<citation citation-type="book">
<person-group person-group-type="author">
<name>
<surname>Zhu</surname>
<given-names>Y.</given-names>
</name>
<name>
<surname>Torrens</surname>
<given-names>Y.</given-names>
</name>
<name>
<surname>Chen</surname>
<given-names>Z.</given-names>
</name>
<name>
<surname>Zhao</surname>
<given-names>S.</given-names>
</name>
<name>
<surname>Xie</surname>
<given-names>H.</given-names>
</name>
<name>
<surname>Guo</surname>
<given-names>W.</given-names>
</name>
<name>
<surname>Zhang</surname>
<given-names>Y.</given-names>
</name>
</person-group> (<year>2019</year>). &#x201c;<article-title>Ace-net: Biomedical Image Segmentation with Augmented Contracting and Expansive Paths</article-title>,&#x2ee;in <conf-name>International Conference on Medical Image Computing and Computer-Assisted Intervention</conf-name>, <conf-loc>Shenzhen, China</conf-loc>, <conf-date>October 13&#x2013;17, 2019</conf-date> (<publisher-name>Springer, Cham</publisher-name>).<fpage>712</fpage>&#x2013;<lpage>720</lpage>. </citation>
</ref>
</ref-list>
</back>
</article>