<?xml version="1.0" encoding="UTF-8" standalone="no"?>
<!DOCTYPE article PUBLIC "-//NLM//DTD Journal Publishing DTD v2.3 20070202//EN" "journalpublishing.dtd">
<article xmlns:mml="http://www.w3.org/1998/Math/MathML" xmlns:xlink="http://www.w3.org/1999/xlink" xmlns:xsi="http://www.w3.org/2001/XMLSchema-instance" article-type="research-article" dtd-version="2.3" xml:lang="EN">
<front>
<journal-meta>
<journal-id journal-id-type="publisher-id">Front. Mar. Sci.</journal-id>
<journal-title>Frontiers in Marine Science</journal-title>
<abbrev-journal-title abbrev-type="pubmed">Front. Mar. Sci.</abbrev-journal-title>
<issn pub-type="epub">2296-7745</issn>
<publisher>
<publisher-name>Frontiers Media S.A.</publisher-name>
</publisher>
</journal-meta>
<article-meta>
<article-id pub-id-type="doi">10.3389/fmars.2023.1205142</article-id>
<article-categories>
<subj-group subj-group-type="heading">
<subject>Marine Science</subject>
<subj-group>
<subject>Original Research</subject>
</subj-group>
</subj-group>
</article-categories>
<title-group>
<article-title>Toward the development of smart capabilities for understanding seafloor stretching morphology and biogeographic patterns via DenseNet from high-resolution multibeam bathymetric surveys for underwater vehicles</article-title>
</title-group>
<contrib-group>
<contrib contrib-type="author" corresp="yes">
<name>
<surname>Nian</surname>
<given-names>Rui</given-names>
</name>
<xref ref-type="aff" rid="aff1">
<sup>1</sup>
</xref>
<xref ref-type="author-notes" rid="fn001">
<sup>*</sup>
</xref>
<uri xlink:href="https://loop.frontiersin.org/people/1160345"/>
</contrib>
<contrib contrib-type="author">
<name>
<surname>Liu</surname>
<given-names>Shasha</given-names>
</name>
<xref ref-type="aff" rid="aff2">
<sup>2</sup>
</xref>
</contrib>
<contrib contrib-type="author">
<name>
<surname>Lu</surname>
<given-names>Zongcan</given-names>
</name>
<xref ref-type="aff" rid="aff1">
<sup>1</sup>
</xref>
</contrib>
<contrib contrib-type="author">
<name>
<surname>Li</surname>
<given-names>Xiaoyu</given-names>
</name>
<xref ref-type="aff" rid="aff1">
<sup>1</sup>
</xref>
</contrib>
<contrib contrib-type="author">
<name>
<surname>Ren</surname>
<given-names>Shidong</given-names>
</name>
<xref ref-type="aff" rid="aff1">
<sup>1</sup>
</xref>
</contrib>
<contrib contrib-type="author">
<name>
<surname>Qian</surname>
<given-names>Yuqi</given-names>
</name>
<xref ref-type="aff" rid="aff1">
<sup>1</sup>
</xref>
<uri xlink:href="https://loop.frontiersin.org/people/1554556"/>
</contrib>
<contrib contrib-type="author">
<name>
<surname>Li</surname>
<given-names>Qiuying</given-names>
</name>
<xref ref-type="aff" rid="aff1">
<sup>1</sup>
</xref>
</contrib>
<contrib contrib-type="author">
<name>
<surname>He</surname>
<given-names>Guotong</given-names>
</name>
<xref ref-type="aff" rid="aff1">
<sup>1</sup>
</xref>
<uri xlink:href="https://loop.frontiersin.org/people/2572957"/>
</contrib>
<contrib contrib-type="author">
<name>
<surname>Shi</surname>
<given-names>Kexin</given-names>
</name>
<xref ref-type="aff" rid="aff1">
<sup>1</sup>
</xref>
</contrib>
<contrib contrib-type="author">
<name>
<surname>Zhang</surname>
<given-names>Guoyao</given-names>
</name>
<xref ref-type="aff" rid="aff1">
<sup>1</sup>
</xref>
</contrib>
<contrib contrib-type="author">
<name>
<surname>Zang</surname>
<given-names>Lina</given-names>
</name>
<xref ref-type="aff" rid="aff1">
<sup>1</sup>
</xref>
</contrib>
<contrib contrib-type="author">
<name>
<surname>Li</surname>
<given-names>Luyao</given-names>
</name>
<xref ref-type="aff" rid="aff1">
<sup>1</sup>
</xref>
</contrib>
<contrib contrib-type="author">
<name>
<surname>He</surname>
<given-names>Bo</given-names>
</name>
<xref ref-type="aff" rid="aff1">
<sup>1</sup>
</xref>
</contrib>
<contrib contrib-type="author">
<name>
<surname>Yan</surname>
<given-names>Tianhong</given-names>
</name>
<xref ref-type="aff" rid="aff3">
<sup>3</sup>
</xref>
<uri xlink:href="https://loop.frontiersin.org/people/156583"/>
</contrib>
<contrib contrib-type="author">
<name>
<surname>Li</surname>
<given-names>Xishuang</given-names>
</name>
<xref ref-type="aff" rid="aff4">
<sup>4</sup>
</xref>
<uri xlink:href="https://loop.frontiersin.org/people/1462774"/>
</contrib>
</contrib-group>
<aff id="aff1">
<sup>1</sup>
<institution>School of Electronic Engineering, Ocean University of China</institution>, <addr-line>Qingdao</addr-line>, <country>China</country>
</aff>
<aff id="aff2">
<sup>2</sup>
<institution>Bureau of Administrative Service, Qingdao Municipal Government</institution>, <addr-line>Qingdao</addr-line>, <country>China</country>
</aff>
<aff id="aff3">
<sup>3</sup>
<institution>School of Mechatronic Engineering, China Jiliang University</institution>, <addr-line>Hangzhou</addr-line>, <country>China</country>
</aff>
<aff id="aff4">
<sup>4</sup>
<institution>Key Laboratory of Marine Geology and Metallogeny, Ministry of Nature Resources of People&#x2019;s Republic of China</institution>, <addr-line>Qingdao</addr-line>, <country>China</country>
</aff>
<author-notes>
<fn fn-type="edited-by">
<p>Edited by: Benjamin Misiuk, Dalhousie University, Canada</p>
</fn>
<fn fn-type="edited-by">
<p>Reviewed by: Xiaodong Shang, Naval University of Engineering, China; Brandon Hobley, University of East Anglia, United Kingdom</p>
</fn>
<fn fn-type="corresp" id="fn001">
<p>*Correspondence: Rui Nian, <email xlink:href="mailto:nianrui_80@163.com">nianrui_80@163.com</email>
</p>
</fn>
</author-notes>
<pub-date pub-type="epub">
<day>24</day>
<month>11</month>
<year>2023</year>
</pub-date>
<pub-date pub-type="collection">
<year>2023</year>
</pub-date>
<volume>10</volume>
<elocation-id>1205142</elocation-id>
<history>
<date date-type="received">
<day>03</day>
<month>05</month>
<year>2023</year>
</date>
<date date-type="accepted">
<day>16</day>
<month>10</month>
<year>2023</year>
</date>
</history>
<permissions>
<copyright-statement>Copyright &#xa9; 2023 Nian, Liu, Lu, Li, Ren, Qian, Li, He, Shi, Zhang, Zang, Li, He, Yan and Li</copyright-statement>
<copyright-year>2023</copyright-year>
<copyright-holder>Nian, Liu, Lu, Li, Ren, Qian, Li, He, Shi, Zhang, Zang, Li, He, Yan and Li</copyright-holder>
<license xlink:href="http://creativecommons.org/licenses/by/4.0/">
<p>This is an open-access article distributed under the terms of the Creative Commons Attribution License (CC BY). The use, distribution or reproduction in other forums is permitted, provided the original author(s) and the copyright owner(s) are credited and that the original publication in this journal is cited, in accordance with accepted academic practice. No use, distribution or reproduction is permitted which does not comply with these terms.</p>
</license>
</permissions>
<abstract>
<p>The increasing use of underwater vehicles facilitates deep-sea exploration at a wide range of depths and spatial scales. In this paper, we make an initial attempt to develop online computing strategies to identify seafloor categories and predict biogeographic patterns with a deep learning-based architecture, DenseNet, integrated with joint morphological cues, with the expectation of potentially developing its embedded smart capacities. We utilized high-resolution multibeam bathymetric measurements derived from MBES and denoted a collection of joint morphological cues to help with semantic mapping and localization. We systematically strengthened dominant feature propagation and promoted feature reuse via DenseNet by applying the channel attention module and spatial pyramid pooling. From our experiment results, the seafloor classification accuracy reached up to 89.87%, 82.01%, and 73.52% on average in terms of PA, MPA, and MIoU metrics, achieving comparable performances with the state-of-the-art deep learning frameworks. We made a preliminary study on potential biogeographic distribution statistics, which allowed us to delicately distinguish the functionality of probable submarine benthic habitats. This study demonstrates the premise of using underwater vehicles through unbiased means or pre-programmed path planning to quantify and estimate seafloor categories and the exhibited fine-scale biogeographic patterns.</p>
</abstract>
<kwd-group>
<kwd>multi-beam bathymetric mapping</kwd>
<kwd>seafloor stretching morphology</kwd>
<kwd>DenseNet</kwd>
<kwd>channel attention module</kwd>
<kwd>spatial pyramid pooling</kwd>
<kwd>biogeographic patterns</kwd>
</kwd-group>
<counts>
<fig-count count="8"/>
<table-count count="7"/>
<equation-count count="24"/>
<ref-count count="88"/>
<page-count count="19"/>
<word-count count="10251"/>
</counts>
<custom-meta-wrap>
<custom-meta>
<meta-name>section-in-acceptance</meta-name>
<meta-value>Ocean Observation</meta-value>
</custom-meta>
</custom-meta-wrap>
</article-meta>
</front>
<body>
<sec id="s1" sec-type="intro">
<label>1</label>
<title>Introduction</title>
<p>Deep sea is viewed as one of the least explored ocean ecosystems (<xref ref-type="bibr" rid="B68">Snelgrove, 1998</xref>; <xref ref-type="bibr" rid="B80">Williams et&#xa0;al., 2010</xref>; <xref ref-type="bibr" rid="B71">Teixeira et&#xa0;al., 2013</xref>; <xref ref-type="bibr" rid="B10">Copley, 2014</xref>). The variety of physicochemical gradients in deep-sea habitats is quite remarkable (<xref ref-type="bibr" rid="B72">Thornton et&#xa0;al., 2016</xref>), with diverse biogeographic patterns. Due to difficulties in its sensing accessibility, it is of great challenging to discover how the seafloor surface stretches and functions as submarine benthic habitats in the extremely deep sea (<xref ref-type="bibr" rid="B30">Jannasch and Mottl, 1985</xref>; <xref ref-type="bibr" rid="B18">Fisher et&#xa0;al., 1994</xref>; <xref ref-type="bibr" rid="B32">Johnson et&#xa0;al., 1994</xref>; <xref ref-type="bibr" rid="B49">McCollom and Shock, 1997</xref>; <xref ref-type="bibr" rid="B63">Sarrazin et&#xa0;al., 1999</xref>; <xref ref-type="bibr" rid="B15">Desbruy&#xe8;res et&#xa0;al., 2000</xref>; <xref ref-type="bibr" rid="B43">Luther et&#xa0;al., 2001</xref>; <xref ref-type="bibr" rid="B75">Van Dover et&#xa0;al., 2002</xref>; <xref ref-type="bibr" rid="B3">Bergquist et&#xa0;al., 2007</xref>; <xref ref-type="bibr" rid="B52">Nakamura and Takai, 2014</xref>; <xref ref-type="bibr" rid="B37">Lamarche et&#xa0;al., 2016</xref>).</p>
<p>Over the decades, a multitude of geomorphometric techniques have utilized bathymetric sensors to characterize seafloor stretching attributes (<xref ref-type="bibr" rid="B7">Chakraborty et&#xa0;al., 2013</xref>; <xref ref-type="bibr" rid="B48">Masetti et&#xa0;al., 2018</xref>; <xref ref-type="bibr" rid="B53">Neil et&#xa0;al., 2019</xref>; <xref ref-type="bibr" rid="B59">Pillay et&#xa0;al., 2020</xref>; <xref ref-type="bibr" rid="B79">Wang et&#xa0;al., 2021a</xref>; <xref ref-type="bibr" rid="B78">Wang et&#xa0;al., 2021b</xref>). The increasing use of underwater vehicles, such as Autonomous Underwater Vehicles (AUVs) or Remotely Operated Vehicles (ROVs), e.g., Bluefin (<xref ref-type="bibr" rid="B55">Panish and Taylor, 2011</xref>), Hugin (<xref ref-type="bibr" rid="B46">Marthiniussen et&#xa0;al., 2004</xref>), Remus (<xref ref-type="bibr" rid="B19">Freitag et&#xa0;al., 2005</xref>), Autosub (<xref ref-type="bibr" rid="B21">Furlong et&#xa0;al., 2012</xref>), JAMSTEC (<xref ref-type="bibr" rid="B70">Tamura et&#xa0;al., 2000</xref>), Urashima (<xref ref-type="bibr" rid="B64">Sawa et&#xa0;al., 2005</xref>), and so forth, provides more opportunities to facilitate deep-sea inspection with flexibility and adaptability (<xref ref-type="bibr" rid="B66">Singh et&#xa0;al., 2004</xref>; <xref ref-type="bibr" rid="B4">Bewley et&#xa0;al., 2012</xref>; <xref ref-type="bibr" rid="B67">Smale et&#xa0;al., 2012</xref>; <xref ref-type="bibr" rid="B29">Huvenne et&#xa0;al., 2018</xref>).</p>
<p>Advanced sensing techniques have provided the opportunity to investigate the detailed seafloor stretching morphology over a wide range of spatial scales. Underwater vehicles could be equipped with a series of sensors and proceed at shallower depths closer to the seafloor surface, providing higher-resolution morphological images and benthic habitat observations. A multibeam echosounder (MBES) is one of the most widely mounted sensors in bathymetric surveys, which can give insight into noteworthy benthic habitats at large scales, like hydrothermal vent sites, cold springs, mud volcanoes, and seamounts along subduction zones and trench areas. Meanwhile, the optical sensors would enable the provision of video transects at increasing depths in the deep sea to explore the potential biogeographic distribution, the submarine benthic habitat, and biological community structure in detail at smaller scales in the field.</p>
<p>Therefore, increasing attention has been paid to exploring the regional and even global seafloor stretching morphology, the extent of geographical ranges, and the submarine benthic habitats in the deep sea (<xref ref-type="bibr" rid="B41">Lonsdale, 1977</xref>; <xref ref-type="bibr" rid="B1">Bach and Edwards, 2003</xref>; <xref ref-type="bibr" rid="B73">Trenkel et&#xa0;al., 2004</xref>; <xref ref-type="bibr" rid="B81">Williams et&#xa0;al., 2012</xref>; <xref ref-type="bibr" rid="B60">Pizarro et&#xa0;al., 2013</xref>; <xref ref-type="bibr" rid="B36">Kuhnz et&#xa0;al., 2014</xref>; <xref ref-type="bibr" rid="B52">Nakamura and Takai, 2014</xref>; <xref ref-type="bibr" rid="B72">Thornton et&#xa0;al., 2016</xref>; <xref ref-type="bibr" rid="B72">Thornton et&#xa0;al., 2016</xref>; <xref ref-type="bibr" rid="B16">Dunlop et&#xa0;al., 2018</xref>; <xref ref-type="bibr" rid="B51">Misiuk and Brown, 2022</xref>). The increased diversity of seafloor surface morphology may account for spatial habitat heterogeneity. The formation and types of benthic habitats are typically associated with the physical and geological attributes of seafloor surface stretching. While the benthic habitat, indicator taxa, biodiversity, and community assemblages at one site may still be greatly different from other sites, they may show significant differentiation along similar geomorphometric characteristics. The classification of the seafloor would probably serve as the physical and geological elements to exert a significant impact on the benthic habitats and the biodiversity of organisms inhabiting the underwater environments. <xref ref-type="bibr" rid="B11">Corr&#xea;a et&#xa0;al. (2022)</xref> conducted an exploration of the plateau and rifts in the Rio Grande Rise (RGR) area with the HyBIS robot and analyzed the description of the structuring factors regarding seafloor topography and habitat types, which revealed highly heterogeneous and rapidly changing habitats with differences in geomorphology, slope, and substrate textures. <xref ref-type="bibr" rid="B74">Urra et&#xa0;al. (2021)</xref> characterized the geomorphologic diversity, habitats, and associated biodiversity in the Gazul MV mud volcanoes with underwater imaging and multibeam bathymetry techniques, identifying habitats harboring a characteristic faunal assemblage and highlighting the slope and water depth as the main factors explaining the distribution of the assemblages. <xref ref-type="bibr" rid="B56">Perez et&#xa0;al. (2022)</xref> explored seamounts located within the Discovery Rise and classified the types of benthic habitats based on substrate hardness, texture, slope, and physical and biological modifiers. <xref ref-type="bibr" rid="B58">Pierdomenico et&#xa0;al. (2015)</xref> conducted a comprehensive analysis of acoustic mapping and optical surveys with underwater vehicles to complete ultra-high-resolution bathymetric and backscatter imagery of geomorphological features of seafloor stretching and to characterize benthic habitat variation in the Hudson Canyon. <xref ref-type="bibr" rid="B69">Swanborn et&#xa0;al. (2023)</xref> examined how multiscale seafloor heterogeneity influences commercially important fish families on seamounts of the Southwest Indian Ridge by quantifying seascape heterogeneity from bathymetry and geomorphological habitat maps. <xref ref-type="bibr" rid="B14">De la Torriente et&#xa0;al. (2018)</xref> observed a highly diverse range of habitats from multibeam bathymetry and high-resolution seismic profiles with ROV at the Seco de los Olivos Seamount and identified depth and slope as the main significant factors structuring epibenthic assemblages.</p>
<p>Most of the seafloor bathymetric surveys tend to retrieve comprehensive topographic and morphological factors and other auxiliary variables and identify seafloor stretching attributes during post-processing. With the rapid development of hardware embedded in underwater vehicles, more and more expectations are cast to promote online smart computing capabilities, enabling the immediate understanding of seafloor surface stretching and submarine benthic habitat on site from on-board perception sensors rather than relying only on routine post-processing back from missions. Due to the great success of deep learning, all kinds of emerging and advanced algorithms have been developed and enhanced recently, from Deep Belief Networks (DBN) (<xref ref-type="bibr" rid="B26">Hinton et&#xa0;al., 2006</xref>), Deep Convolutional Neural Networks (CNN) (<xref ref-type="bibr" rid="B33">Karpathy et&#xa0;al., 2014</xref>), and AlexNet (<xref ref-type="bibr" rid="B35">Krizhevsky et&#xa0;al., 2012</xref>) to more recent Generative Adversarial Networks (GANs) (<xref ref-type="bibr" rid="B22">Goodfellow et&#xa0;al., 2014</xref>), Deep Residual Networks (ResNet) (<xref ref-type="bibr" rid="B25">He et&#xa0;al., 2016</xref>), Densely Connected Convolutional Networks (DenseNet) (<xref ref-type="bibr" rid="B28">Huang et&#xa0;al., 2017</xref>), and Transformer (<xref ref-type="bibr" rid="B76">Vaswani et&#xa0;al., 2017</xref>). Recently, <xref ref-type="bibr" rid="B9">Conti et&#xa0;al. (2019)</xref> employed Marine Object-Based Image Analysis (MOBIA) and machine learning classification to identify the distribution and zonation of individual organisms on a cold-water coral (CWC), the Piddington Mound within the Porcupine Seabight, Ireland Margin, from a high-resolution reef-scale video mosaic and ROV-mounted multibeam data. <xref ref-type="bibr" rid="B61">Qin et&#xa0;al. (2021)</xref> applied shallow-water, side-scan sonar imaging from the Pearl River Estuary, established the ResNet architecture for acoustic seafloor classification (ASC), and explored the use of GANs for augmentation. <xref ref-type="bibr" rid="B62">Rimavicius and Gelzinis (2017)</xref> developed an accurate Norwegian seafloor interpretation and classification system with state-of-the-art deep learning techniques. <xref ref-type="bibr" rid="B47">Martin-Abadal et&#xa0;al. (2019)</xref> presented a highprecision semantic segmentation performed automatically in Posidonia oceanica meadows and its habitat by a deep learning-based network, VGG16. <xref ref-type="bibr" rid="B17">Dyer et&#xa0;al. (2020)</xref> proposed the identification of seafloor landslides in images with the deep learning model ResNet101 in the Gulf of Mexico, from the advanced bathymetry raster and its derivatives rendered to high-resolution seafloor topography.</p>
<p>However, the above has not yet been extensively applied to specify seafloor stretching characteristics and intrinsic relations with the limited computation resources in underwater vehicles due to the difficulties of simultaneously bathymetry mapping, classifying, and validating. The hardware configuration for embedded supercomputing, such as a high-performance Graphics Processing Unit (GPU), may promote online computational capacities. Hence, we attempted to accelerate the identification and localization process of seafloor morphology and to deduce possible biogeographic patterns with a deep learning-based architecture, DenseNet. DenseNet connects each layer to every other layer in a feed-forward fashion, where the feature maps of all preceding layers are inputs and its own feature maps are inputs to all subsequent layers (<xref ref-type="bibr" rid="B28">Huang et&#xa0;al., 2017</xref>). DenseNet offers significant advantages over many state-of-the-art deep learning algorithms. It addresses the vanishing gradient problem, enhances feature propagation, and significantly reduces the number of parameters, thereby demanding fewer computational resources. Moreover, various new variants have been developed recently (<xref ref-type="bibr" rid="B31">J&#xe9;gou et&#xa0;al., 2017</xref>; <xref ref-type="bibr" rid="B88">Zhu and Newsam, 2017</xref>; <xref ref-type="bibr" rid="B77">Wang et&#xa0;al., 2018</xref>; <xref ref-type="bibr" rid="B38">Lee et&#xa0;al., 2019</xref>; <xref ref-type="bibr" rid="B42">Lu et&#xa0;al., 2021</xref>; <xref ref-type="bibr" rid="B85">Xiao et&#xa0;al., 2021</xref>).</p>
<p>We aimed to establish a generalized deep learning-based architecture so as to develop smart capabilities to identify seafloor categories for underwater vehicles. This could facilitate online computing strategies for unbiased path planning with real-time perception and autonomous decision-making, and adapt to dynamic, unknown, and complex underwater conditions. Such unbiased path planning tends not to rely so much on predetermined global seafloor mappings or predefined routes but focuses on current environmental states, efficiently updating the path selection with the understanding of seafloor categories or potential biogeographic distribution retrieved along the mission route points. The model parameters could be updated and fine-tuned by the local underwater geology, dynamically promoting solutions to viable path planning optimization in an online manner. This could also benefit pre-programmed path planning that utilizes deterministic or optimization algorithms to search offline for optimal or shortest routes from the known underwater environment mapping. Such pre-programmed path planning typically requires prior acquisition of knowledge of regional and even global seafloor stretching morphology. The quality and strength of such approaches partially lie in the ability to identify seafloor categories with high quality and high computational efficiency.</p>
<p>In this paper, starting from the inspiration to explore the seafloor surface stretching and biogeographic patterns along the traces of underwater vesicles, we directly retrieved the multibeam bathymetric mapping from the MBES scans and calculated a variety of morphological parameters from the digital elevation to help semantic segmentation and localization. We established online computation strategies via DenseNet and took the multibeam bathymetric measurements and the joint morphological cues as inputs, with the annotations manually labeled as expected outputs. The minimalistic transition-up blocks, the channel attention module, and the spatial pyramid pooling have been seamlessly integrated to systematically strengthen the dominant feature propagation and encourage feature reuse with the global contextual prior. The developed model has been comprehensively evaluated in terms of PA (Pixel Accuracy), MPA (Mean Pixel Accuracy), and MIoU (Mean Intersection over Union Ratio). We further proposed a preliminary study on the potential biogeographic distribution statistics to provide initial insights into the connective and predictive evidence between seafloor categories, benthic habitats, and even species assemblages.</p>
<p>The remainder of the paper is organized as follows: Sections 2, 3, and 4 describe the basic principles of multibeam bathymetric mapping, the basics of seafloor surface morphological calculation, and DenseNet, respectively. Section 5 introduces the seafloor surface classification via DenseNet, coupled with the channel attention module and spatial pyramid pooling. Section 6 shows the simulation experiment and the analysis of the results. Finally, the conclusions are given in Section 7.</p>
</sec>
<sec id="s2">
<label>2</label>
<title>High-resolution multibeam bathymetric mapping</title>
<sec id="s2_1">
<label>2.1</label>
<title>MBES principle</title>
<p>MBES refers to a type of highly integrated multibeam bathymetric sensor. It could help with full-coverage depth measurements at high resolution and determine the nature of seafloor surfaces in the deep sea. The basic principle of MBES is shown in <xref ref-type="supplementary-material" rid="SM1">
<bold>Supplementary Figure A1</bold></xref>. The MBES transducer is essentially a combination of an acoustic projector array and a perpendicular hydrophone array. The former emits acoustic pulses at a specific frequency, with a narrow opening along-track angle and a wide across-track angle, in a given swath (<xref ref-type="bibr" rid="B45">Mahmud and Yusof, 2006</xref>; <xref ref-type="bibr" rid="B12">Costa et&#xa0;al., 2009</xref>). The latter is built to listen to echo reflections with received beams. Thus, the seafloor strips, ensonified by the projectors, will intersect with those observed by the hydrophones, producing the beam footprints. When receiving across-track beams of certain time intervals one after another, the position and depth of the seafloor measurement could be calculated, given the angle of incidence and the two-way travel time of each beam (<xref ref-type="bibr" rid="B87">Zhao et al., 2020</xref>; <xref ref-type="bibr" rid="B84">Wu et&#xa0;al., 2021</xref>). In a complete transmission and reception period, the projector array runs only once to generate acoustic pulses, while the hydrophone array acquires multiple received beams with appropriate delays. As underwater vehicles proceed forward, a strip of water depth measurements reflecting bathymetric mapping at a specific width could be derived from the MBES, providing full coverage of the seafloor surface morphology, which would benefit the identification and location of seafloor categories with high precision, high density, and high efficiency.</p>
</sec>
<sec id="s2_2">
<label>2.2</label>
<title>Digital elevation model</title>
<p>Essentially, we utilized the Digital Elevation Model (DEM) to solve the numerical problem of visualizing the geospatial entities of the seafloor surfaces with a finite set of depth measurements from the MBES. The core interpolation calculation allows the production of a gridded multibeam bathymetric map. We applied a weighted average point-to-point interpolation to generate the digital elevation. Assuming that the elevation point is to be inserted in the center of each sliding sampled window, the elevation value <inline-formula>
<mml:math display="inline" id="im1">
<mml:mrow>
<mml:msub>
<mml:mi>E</mml:mi>
<mml:mi>G</mml:mi>
</mml:msub>
</mml:mrow>
</mml:math>
</inline-formula> is determined by approximating the weighted averaging of the surrounding elevation values within the window, which can be formulated as</p>
<disp-formula>
<label>(1)</label>
<mml:math display="block" id="M1">
<mml:mrow>
<mml:msub>
<mml:mi>E</mml:mi>
<mml:mi>G</mml:mi>
</mml:msub>
<mml:mo>=</mml:mo>
<mml:mfrac>
<mml:mrow>
<mml:mstyle displaystyle="true">
<mml:munderover>
<mml:mo>&#x2211;</mml:mo>
<mml:mrow>
<mml:mi>i</mml:mi>
<mml:mo>=</mml:mo>
<mml:mn>1</mml:mn>
</mml:mrow>
<mml:mi>n</mml:mi>
</mml:munderover>
<mml:mrow>
<mml:msub>
<mml:mi>G</mml:mi>
<mml:mi>i</mml:mi>
</mml:msub>
<mml:mo>&#xd7;</mml:mo>
<mml:msub>
<mml:mi>E</mml:mi>
<mml:mi>i</mml:mi>
</mml:msub>
</mml:mrow>
</mml:mstyle>
</mml:mrow>
<mml:mrow>
<mml:mstyle displaystyle="true">
<mml:munderover>
<mml:mo>&#x2211;</mml:mo>
<mml:mrow>
<mml:mi>i</mml:mi>
<mml:mo>=</mml:mo>
<mml:mn>1</mml:mn>
</mml:mrow>
<mml:mi>n</mml:mi>
</mml:munderover>
<mml:mrow>
<mml:msub>
<mml:mi>G</mml:mi>
<mml:mi>i</mml:mi>
</mml:msub>
</mml:mrow>
</mml:mstyle>
</mml:mrow>
</mml:mfrac>
</mml:mrow>
</mml:math>
</disp-formula>
<p>where the number of neighboring elevation points in the sliding window is denoted as <inline-formula>
<mml:math display="inline" id="im2">
<mml:mi>n</mml:mi>
</mml:math>
</inline-formula>, <inline-formula>
<mml:math display="inline" id="im3">
<mml:mrow>
<mml:msub>
<mml:mi>E</mml:mi>
<mml:mi>i</mml:mi>
</mml:msub>
</mml:mrow>
</mml:math>
</inline-formula> refers to the <inline-formula>
<mml:math display="inline" id="im4">
<mml:mi>i</mml:mi>
</mml:math>
</inline-formula> th elevation value, and <inline-formula>
<mml:math display="inline" id="im5">
<mml:mrow>
<mml:msub>
<mml:mi>G</mml:mi>
<mml:mi>i</mml:mi>
</mml:msub>
</mml:mrow>
</mml:math>
</inline-formula> represents the corresponding weight. For the output elevation values, the sum of the products between the surrounding elevation values and their corresponding weights within the window is divided by the sum of all the weights. Each weight <inline-formula>
<mml:math display="inline" id="im6">
<mml:mrow>
<mml:msub>
<mml:mi>G</mml:mi>
<mml:mi>i</mml:mi>
</mml:msub>
</mml:mrow>
</mml:math>
</inline-formula> is defined as the reciprocal of the spatial distance <inline-formula>
<mml:math display="inline" id="im7">
<mml:mrow>
<mml:msub>
<mml:mi>D</mml:mi>
<mml:mi>i</mml:mi>
</mml:msub>
</mml:mrow>
</mml:math>
</inline-formula> between the surrounding elevation points and the center to be inserted,</p>
<disp-formula>
<label>(2)</label>
<mml:math display="block" id="M2">
<mml:mrow>
<mml:msub>
<mml:mi>G</mml:mi>
<mml:mi>i</mml:mi>
</mml:msub>
<mml:mo>=</mml:mo>
<mml:mfrac>
<mml:mn>1</mml:mn>
<mml:mrow>
<mml:msub>
<mml:mi>D</mml:mi>
<mml:mi>i</mml:mi>
</mml:msub>
</mml:mrow>
</mml:mfrac>
</mml:mrow>
</mml:math>
</disp-formula>
<p>The greater the spatial distance <inline-formula>
<mml:math display="inline" id="im8">
<mml:mrow>
<mml:msub>
<mml:mi>D</mml:mi>
<mml:mi>i</mml:mi>
</mml:msub>
</mml:mrow>
</mml:math>
</inline-formula>, the smaller the corresponding weight <inline-formula>
<mml:math display="inline" id="im9">
<mml:mrow>
<mml:msub>
<mml:mi>G</mml:mi>
<mml:mi>i</mml:mi>
</mml:msub>
</mml:mrow>
</mml:math>
</inline-formula>, and vice versa. An example of a seafloor strip before and after the interpolation is shown in <xref ref-type="fig" rid="f1">
<bold>Figure&#xa0;1</bold>
</xref>, where the color bar denotes the water depth values. We can see from the experimental results that the high-resolution multibeam bathymetric mapping could reasonably depict the integrity of the seafloor surface stretching in the DEM, especially the stitching of the gaps in the edges.</p>
<fig id="f1" position="float">
<label>Figure&#xa0;1</label>
<caption>
<p>Example strip of seafloor surface mapping. <bold>(A)</bold> before interpolation, <bold>(B)</bold> after interpolation.</p>
</caption>
<graphic mimetype="image" mime-subtype="tiff" xlink:href="fmars-10-1205142-g001.tif"/>
</fig>
</sec>
</sec>
<sec id="s3">
<label>3</label>
<title>Seafloor surface morphological calculation</title>
<p>We further elaborated on the thematic maps of seafloor surface morphological cues from high-resolution multibeam bathymetric mapping. Let the depth value of the given elevation point on a certain seafloor surface be <inline-formula>
<mml:math display="inline" id="im10">
<mml:mrow>
<mml:mtext>z</mml:mtext>
<mml:mo>=</mml:mo>
<mml:mi>f</mml:mi>
<mml:mo stretchy="false">(</mml:mo>
<mml:mi>x</mml:mi>
<mml:mo>,</mml:mo>
<mml:mi>y</mml:mi>
<mml:mo stretchy="false">)</mml:mo>
</mml:mrow>
</mml:math>
</inline-formula>, with <inline-formula>
<mml:math display="inline" id="im11">
<mml:mi>x</mml:mi>
</mml:math>
</inline-formula> and <inline-formula>
<mml:math display="inline" id="im12">
<mml:mi>y</mml:mi>
</mml:math>
</inline-formula> representing the horizontal and vertical coordinates at the seafloor location, respectively. We essentially employed several topological attributes in aid of the first and second derivatives calculated from the neighborhood within a sliding window to consider every elevation point in turn.</p>
<sec id="s3_1">
<label>3.1</label>
<title>Slope</title>
<p>The slope refers to the measurement that determines the steepness or degree of inclination in seafloor bathymetric mapping relative to the horizontal plane, which constitutes the fundamental index of benthic habitat and colonization at a variety of scales (<xref ref-type="bibr" rid="B20">Friedman et&#xa0;al., 2013</xref>). Multibeam bathymetric mapping can be approximated by a bivariate quadratic equation, and we compute the slope with the first derivative of the elevation values. The slope with the origin at the central point in the local coordinate system within the sliding window is hereby calculated as</p>
<disp-formula>
<label>(3)</label>
<mml:math display="block" id="M3">
<mml:mrow>
<mml:mi>S</mml:mi>
<mml:mo>=</mml:mo>
<mml:mi>tan</mml:mi>
<mml:msqrt>
<mml:mrow>
<mml:msubsup>
<mml:mi>S</mml:mi>
<mml:mi>x</mml:mi>
<mml:mrow>
<mml:mtext>&#xa0;&#xa0;</mml:mtext>
<mml:mn>2</mml:mn>
</mml:mrow>
</mml:msubsup>
<mml:mo>+</mml:mo>
<mml:msubsup>
<mml:mi>S</mml:mi>
<mml:mi>y</mml:mi>
<mml:mrow>
<mml:mtext>&#xa0;&#xa0;</mml:mtext>
<mml:mn>2</mml:mn>
</mml:mrow>
</mml:msubsup>
</mml:mrow>
</mml:msqrt>
</mml:mrow>
</mml:math>
</disp-formula>
<p>The slope direction <inline-formula>
<mml:math display="inline" id="im13">
<mml:mi>A</mml:mi>
</mml:math>
</inline-formula> could be defined as,</p>
<disp-formula>
<label>(4)</label>
<mml:math display="block" id="M4">
<mml:mrow>
<mml:mi>A</mml:mi>
<mml:mo>=</mml:mo>
<mml:mfrac>
<mml:mrow>
<mml:msub>
<mml:mi>S</mml:mi>
<mml:mi>x</mml:mi>
</mml:msub>
</mml:mrow>
<mml:mrow>
<mml:msub>
<mml:mi>S</mml:mi>
<mml:mi>y</mml:mi>
</mml:msub>
</mml:mrow>
</mml:mfrac>
</mml:mrow>
</mml:math>
</disp-formula>
<p>where <inline-formula>
<mml:math display="inline" id="im14">
<mml:mrow>
<mml:msub>
<mml:mi>S</mml:mi>
<mml:mi>x</mml:mi>
</mml:msub>
</mml:mrow>
</mml:math>
</inline-formula> and <inline-formula>
<mml:math display="inline" id="im90">
<mml:mrow>
<mml:msub>
<mml:mi>S</mml:mi>
<mml:mi>y</mml:mi>
</mml:msub>
</mml:mrow>
</mml:math>
</inline-formula> represent the slope with respect to <inline-formula>
<mml:math display="inline" id="im15">
<mml:mi>x</mml:mi>
</mml:math>
</inline-formula> and <inline-formula>
<mml:math display="inline" id="im16">
<mml:mi>y</mml:mi>
</mml:math>
</inline-formula> directions, which can take a variety of forms. We determine the slope value of the central point from the finite differential of the surrounding neighbors within the sliding window, as is shown in <xref ref-type="supplementary-material" rid="SM1">
<bold>Supplementary Figure B.1</bold>
</xref>. The slope <inline-formula>
<mml:math display="inline" id="im17">
<mml:mrow>
<mml:msub>
<mml:mi>S</mml:mi>
<mml:mi>x</mml:mi>
</mml:msub>
</mml:mrow>
</mml:math>
</inline-formula> and <inline-formula>
<mml:math display="inline" id="im18">
<mml:mrow>
<mml:msub>
<mml:mi>S</mml:mi>
<mml:mi>y</mml:mi>
</mml:msub>
</mml:mrow>
</mml:math>
</inline-formula> of the horizontal and vertical directions could be denoted as,</p>
<disp-formula>
<label>(5)</label>
<mml:math display="block" id="M5">
<mml:mrow>
<mml:msub>
<mml:mi>S</mml:mi>
<mml:mi>x</mml:mi>
</mml:msub>
<mml:mo>=</mml:mo>
<mml:mfrac>
<mml:mrow>
<mml:msub>
<mml:mi>z</mml:mi>
<mml:mn>1</mml:mn>
</mml:msub>
<mml:mo>&#x2212;</mml:mo>
<mml:msub>
<mml:mi>z</mml:mi>
<mml:mn>3</mml:mn>
</mml:msub>
</mml:mrow>
<mml:mrow>
<mml:mn>2</mml:mn>
<mml:mo>&#xd7;</mml:mo>
<mml:mtext>&#x394;</mml:mtext>
<mml:mi>l</mml:mi>
</mml:mrow>
</mml:mfrac>
</mml:mrow>
</mml:math>
</disp-formula>
<disp-formula>
<label>(6)</label>
<mml:math display="block" id="M6">
<mml:mrow>
<mml:msub>
<mml:mi>S</mml:mi>
<mml:mi>y</mml:mi>
</mml:msub>
<mml:mo>=</mml:mo>
<mml:mfrac>
<mml:mrow>
<mml:msub>
<mml:mi>z</mml:mi>
<mml:mn>4</mml:mn>
</mml:msub>
<mml:mo>&#x2212;</mml:mo>
<mml:msub>
<mml:mi>z</mml:mi>
<mml:mn>2</mml:mn>
</mml:msub>
</mml:mrow>
<mml:mrow>
<mml:mn>2</mml:mn>
<mml:mo>&#xd7;</mml:mo>
<mml:mtext>&#x394;</mml:mtext>
<mml:mi>l</mml:mi>
</mml:mrow>
</mml:mfrac>
</mml:mrow>
</mml:math>
</disp-formula>
<p>where <inline-formula>
<mml:math display="inline" id="im19">
<mml:mrow>
<mml:msub>
<mml:mi>z</mml:mi>
<mml:mn>1</mml:mn>
</mml:msub>
<mml:mo>&#x2212;</mml:mo>
<mml:msub>
<mml:mi>z</mml:mi>
<mml:mn>4</mml:mn>
</mml:msub>
</mml:mrow>
</mml:math>
</inline-formula> are the elevation values in the sliding window, respectively, and <inline-formula>
<mml:math display="inline" id="im20">
<mml:mrow>
<mml:mtext>&#x394;</mml:mtext>
<mml:mi>l</mml:mi>
</mml:mrow>
</mml:math>
</inline-formula> is the grid length.</p>
</sec>
<sec id="s3_2">
<label>3.2</label>
<title>Curvature</title>
<p>The curvature behaves as a quantitative measurement of the degree of distortion on the surface of the seafloor geomorphic changes, providing a possible assessment of uplift or depression (<xref ref-type="bibr" rid="B65">Shary, 1995</xref>). The profile curvature values stand for the stretching morphology of the seafloor surface, with positive curvature attesting to an upwardly concave and a negative curvature, indicating upwardly convex, and a value of zero indicating flat seafloor surfaces. It helps to delimit distinct habitat regions by identifying boundaries in seafloor morphology, delineating between favorable and unfavorable habitats for communities. The curvature is a second spatial derivative of the seabed terrain, which can be expressed as</p>
<disp-formula>
<label>(7)</label>
<mml:math display="block" id="M7">
<mml:mrow>
<mml:msub>
<mml:mi>C</mml:mi>
<mml:mi>v</mml:mi>
</mml:msub>
<mml:mo>=</mml:mo>
<mml:mo>&#x2212;</mml:mo>
<mml:mfrac>
<mml:mrow>
<mml:msup>
<mml:mi>l</mml:mi>
<mml:mn>2</mml:mn>
</mml:msup>
<mml:mi>r</mml:mi>
<mml:mo>+</mml:mo>
<mml:mn>2</mml:mn>
<mml:mi>l</mml:mi>
<mml:mi>q</mml:mi>
<mml:mi>s</mml:mi>
<mml:mo>+</mml:mo>
<mml:msup>
<mml:mi>q</mml:mi>
<mml:mn>2</mml:mn>
</mml:msup>
<mml:mi>t</mml:mi>
</mml:mrow>
<mml:mrow>
<mml:mrow>
<mml:mo stretchy="false">(</mml:mo>
<mml:mrow>
<mml:msup>
<mml:mi>l</mml:mi>
<mml:mn>2</mml:mn>
</mml:msup>
<mml:mo>+</mml:mo>
<mml:msup>
<mml:mi>q</mml:mi>
<mml:mn>2</mml:mn>
</mml:msup>
</mml:mrow>
<mml:mo stretchy="false">)</mml:mo>
</mml:mrow>
<mml:msqrt>
<mml:mrow>
<mml:mn>1</mml:mn>
<mml:mo>+</mml:mo>
<mml:msup>
<mml:mi>l</mml:mi>
<mml:mn>2</mml:mn>
</mml:msup>
<mml:mo>+</mml:mo>
<mml:msup>
<mml:mi>q</mml:mi>
<mml:mn>2</mml:mn>
</mml:msup>
</mml:mrow>
</mml:msqrt>
</mml:mrow>
</mml:mfrac>
</mml:mrow>
</mml:math>
</disp-formula>
<p>where <inline-formula>
<mml:math display="inline" id="im21">
<mml:mi>l</mml:mi>
</mml:math>
</inline-formula> and <inline-formula>
<mml:math display="inline" id="im22">
<mml:mi>q</mml:mi>
</mml:math>
</inline-formula> are the first derivatives of the elevation values in the horizontal and vertical directions, respectively; <inline-formula>
<mml:math display="inline" id="im23">
<mml:mi>r</mml:mi>
</mml:math>
</inline-formula>, <inline-formula>
<mml:math display="inline" id="im24">
<mml:mi>s</mml:mi>
</mml:math>
</inline-formula>, <inline-formula>
<mml:math display="inline" id="im25">
<mml:mi>t</mml:mi>
</mml:math>
</inline-formula> correspond to the derivative of the horizontal slope with respect to <inline-formula>
<mml:math display="inline" id="im26">
<mml:mi>x</mml:mi>
</mml:math>
</inline-formula> direction, the derivative of the horizontal slope with respect to <inline-formula>
<mml:math display="inline" id="im27">
<mml:mi>y</mml:mi>
</mml:math>
</inline-formula> direction, and the derivative of the vertical slope with respect to <inline-formula>
<mml:math display="inline" id="im28">
<mml:mi>y</mml:mi>
</mml:math>
</inline-formula> direction, respectively.</p>
<disp-formula>
<label>(8)</label>
<mml:math display="block" id="M8">
<mml:mrow>
<mml:mi>l</mml:mi>
<mml:mo>=</mml:mo>
<mml:mfrac>
<mml:mrow>
<mml:mtext>d</mml:mtext>
<mml:mi>z</mml:mi>
</mml:mrow>
<mml:mrow>
<mml:mi>d</mml:mi>
<mml:mi>x</mml:mi>
</mml:mrow>
</mml:mfrac>
<mml:mo>=</mml:mo>
<mml:mfrac>
<mml:mrow>
<mml:msub>
<mml:mi>z</mml:mi>
<mml:mn>1</mml:mn>
</mml:msub>
<mml:mo>&#x2212;</mml:mo>
<mml:msub>
<mml:mi>z</mml:mi>
<mml:mn>3</mml:mn>
</mml:msub>
</mml:mrow>
<mml:mrow>
<mml:mn>2</mml:mn>
<mml:mo>&#xd7;</mml:mo>
<mml:mtext>&#x394;</mml:mtext>
<mml:mi>l</mml:mi>
</mml:mrow>
</mml:mfrac>
</mml:mrow>
</mml:math>
</disp-formula>
<disp-formula>
<label>(9)</label>
<mml:math display="block" id="M9">
<mml:mrow>
<mml:mi>q</mml:mi>
<mml:mo>=</mml:mo>
<mml:mfrac>
<mml:mrow>
<mml:mi>d</mml:mi>
<mml:mi>z</mml:mi>
</mml:mrow>
<mml:mrow>
<mml:mi>d</mml:mi>
<mml:mi>y</mml:mi>
</mml:mrow>
</mml:mfrac>
<mml:mo>=</mml:mo>
<mml:mfrac>
<mml:mrow>
<mml:msub>
<mml:mi>z</mml:mi>
<mml:mn>4</mml:mn>
</mml:msub>
<mml:mo>&#x2212;</mml:mo>
<mml:msub>
<mml:mi>z</mml:mi>
<mml:mn>2</mml:mn>
</mml:msub>
</mml:mrow>
<mml:mrow>
<mml:mn>2</mml:mn>
<mml:mo>&#xd7;</mml:mo>
<mml:mtext>&#x394;</mml:mtext>
<mml:mi>l</mml:mi>
</mml:mrow>
</mml:mfrac>
</mml:mrow>
</mml:math>
</disp-formula>
<disp-formula>
<label>(10)</label>
<mml:math display="block" id="M10">
<mml:mrow>
<mml:mi>r</mml:mi>
<mml:mo>=</mml:mo>
<mml:mfrac>
<mml:mrow>
<mml:msup>
<mml:mo>&#x2202;</mml:mo>
<mml:mn>2</mml:mn>
</mml:msup>
<mml:mi>Z</mml:mi>
</mml:mrow>
<mml:mrow>
<mml:mo>&#x2202;</mml:mo>
<mml:msup>
<mml:mi>x</mml:mi>
<mml:mn>2</mml:mn>
</mml:msup>
</mml:mrow>
</mml:mfrac>
<mml:mo>=</mml:mo>
<mml:mfrac>
<mml:mrow>
<mml:msub>
<mml:msup>
<mml:mi>z</mml:mi>
<mml:mo>&#x2032;</mml:mo>
</mml:msup>
<mml:mrow>
<mml:mn>1</mml:mn>
<mml:mi>x</mml:mi>
</mml:mrow>
</mml:msub>
<mml:mo>&#x2212;</mml:mo>
<mml:msub>
<mml:msup>
<mml:mi>z</mml:mi>
<mml:mo>&#x2032;</mml:mo>
</mml:msup>
<mml:mrow>
<mml:mn>3</mml:mn>
<mml:mi>x</mml:mi>
</mml:mrow>
</mml:msub>
</mml:mrow>
<mml:mrow>
<mml:mn>2</mml:mn>
<mml:mo>&#xd7;</mml:mo>
<mml:mtext>&#x394;</mml:mtext>
<mml:mi>l</mml:mi>
</mml:mrow>
</mml:mfrac>
</mml:mrow>
</mml:math>
</disp-formula>
<disp-formula>
<label>(11)</label>
<mml:math display="block" id="M11">
<mml:mrow>
<mml:mi>s</mml:mi>
<mml:mo>=</mml:mo>
<mml:mfrac>
<mml:mrow>
<mml:msup>
<mml:mo>&#x2202;</mml:mo>
<mml:mn>2</mml:mn>
</mml:msup>
<mml:mi>Z</mml:mi>
</mml:mrow>
<mml:mrow>
<mml:mo>&#x2202;</mml:mo>
<mml:mi>x</mml:mi>
<mml:mo>&#x2202;</mml:mo>
<mml:mi>y</mml:mi>
</mml:mrow>
</mml:mfrac>
<mml:mo>=</mml:mo>
<mml:mfrac>
<mml:mrow>
<mml:msub>
<mml:msup>
<mml:mi>z</mml:mi>
<mml:mo>&#x2032;</mml:mo>
</mml:msup>
<mml:mrow>
<mml:mn>4</mml:mn>
<mml:mi>x</mml:mi>
</mml:mrow>
</mml:msub>
<mml:mo>&#x2212;</mml:mo>
<mml:msub>
<mml:msup>
<mml:mi>z</mml:mi>
<mml:mo>&#x2032;</mml:mo>
</mml:msup>
<mml:mrow>
<mml:mn>2</mml:mn>
<mml:mi>x</mml:mi>
</mml:mrow>
</mml:msub>
</mml:mrow>
<mml:mrow>
<mml:mn>2</mml:mn>
<mml:mo>&#xd7;</mml:mo>
<mml:mtext>&#x394;</mml:mtext>
<mml:mi>l</mml:mi>
</mml:mrow>
</mml:mfrac>
</mml:mrow>
</mml:math>
</disp-formula>
<disp-formula>
<label>(12)</label>
<mml:math display="block" id="M12">
<mml:mrow>
<mml:mi>t</mml:mi>
<mml:mo>=</mml:mo>
<mml:mfrac>
<mml:mrow>
<mml:msup>
<mml:mo>&#x2202;</mml:mo>
<mml:mn>2</mml:mn>
</mml:msup>
<mml:mi>Z</mml:mi>
</mml:mrow>
<mml:mrow>
<mml:mo>&#x2202;</mml:mo>
<mml:msup>
<mml:mi>y</mml:mi>
<mml:mn>2</mml:mn>
</mml:msup>
</mml:mrow>
</mml:mfrac>
<mml:mo>=</mml:mo>
<mml:mfrac>
<mml:mrow>
<mml:msub>
<mml:msup>
<mml:mi>z</mml:mi>
<mml:mo>&#x2032;</mml:mo>
</mml:msup>
<mml:mrow>
<mml:mn>4</mml:mn>
<mml:mi>y</mml:mi>
</mml:mrow>
</mml:msub>
<mml:mo>&#x2212;</mml:mo>
<mml:msub>
<mml:msup>
<mml:mi>z</mml:mi>
<mml:mo>&#x2032;</mml:mo>
</mml:msup>
<mml:mrow>
<mml:mn>2</mml:mn>
<mml:mi>y</mml:mi>
</mml:mrow>
</mml:msub>
</mml:mrow>
<mml:mrow>
<mml:mn>2</mml:mn>
<mml:mo>&#xd7;</mml:mo>
<mml:mtext>&#x394;</mml:mtext>
<mml:mi>l</mml:mi>
</mml:mrow>
</mml:mfrac>
</mml:mrow>
</mml:math>
</disp-formula>
<p>where <inline-formula>
<mml:math display="inline" id="im29">
<mml:mrow>
<mml:msub>
<mml:msup>
<mml:mi>z</mml:mi>
<mml:mo>&#x2032;</mml:mo>
</mml:msup>
<mml:mrow>
<mml:mn>1</mml:mn>
<mml:mi>x</mml:mi>
</mml:mrow>
</mml:msub>
<mml:mo>&#x2212;</mml:mo>
<mml:msub>
<mml:msup>
<mml:mi>z</mml:mi>
<mml:mo>&#x2032;</mml:mo>
</mml:msup>
<mml:mrow>
<mml:mn>4</mml:mn>
<mml:mi>x</mml:mi>
</mml:mrow>
</mml:msub>
</mml:mrow>
</mml:math>
</inline-formula>, <inline-formula>
<mml:math display="inline" id="im30">
<mml:mrow>
<mml:msub>
<mml:msup>
<mml:mi>z</mml:mi>
<mml:mo>&#x2032;</mml:mo>
</mml:msup>
<mml:mrow>
<mml:mn>1</mml:mn>
<mml:mi>y</mml:mi>
</mml:mrow>
</mml:msub>
<mml:mo>&#x2212;</mml:mo>
<mml:msub>
<mml:msup>
<mml:mi>z</mml:mi>
<mml:mo>&#x2032;</mml:mo>
</mml:msup>
<mml:mrow>
<mml:mn>4</mml:mn>
<mml:mi>y</mml:mi>
</mml:mrow>
</mml:msub>
</mml:mrow>
</mml:math>
</inline-formula> are the first derivatives of the elevation values in the horizontal and vertical directions within the sliding window, as is calculated in <xref ref-type="supplementary-material" rid="SM1">
<bold>Supplementary Figure C1</bold>
</xref>.</p>
</sec>
<sec id="s3_3">
<label>3.3</label>
<title>Roughness</title>
<p>Surface roughness reflects the degree of the structural complexity of the seafloor surface stretching, which to some extent indicates its macrotopographic characteristics and undulation status, and can be defined as the ratio of the total seafloor surface of the sampled region to a projected plane to decouple measurements from the overall slope (<xref ref-type="bibr" rid="B20">Friedman et&#xa0;al., 2013</xref>). Each topographic seafloor surface stretching can be divided into non-overlapping virtual quadrats, and the surface roughness value is derived from each virtual square as</p>
<disp-formula>
<label>(13)</label>
<mml:math display="block" id="M13">
<mml:mrow>
<mml:msub>
<mml:mi>R</mml:mi>
<mml:mi>s</mml:mi>
</mml:msub>
<mml:mo>=</mml:mo>
<mml:mfrac>
<mml:mrow>
<mml:msub>
<mml:mi>S</mml:mi>
<mml:mi>s</mml:mi>
</mml:msub>
</mml:mrow>
<mml:mrow>
<mml:msub>
<mml:mi>S</mml:mi>
<mml:mi>p</mml:mi>
</mml:msub>
</mml:mrow>
</mml:mfrac>
</mml:mrow>
</mml:math>
</disp-formula>
<p>where <inline-formula>
<mml:math display="inline" id="im31">
<mml:mrow>
<mml:msub>
<mml:mi>S</mml:mi>
<mml:mi>s</mml:mi>
</mml:msub>
</mml:mrow>
</mml:math>
</inline-formula> and <inline-formula>
<mml:math display="inline" id="im32">
<mml:mrow>
<mml:msub>
<mml:mi>S</mml:mi>
<mml:mi>p</mml:mi>
</mml:msub>
</mml:mrow>
</mml:math>
</inline-formula> are the seafloor surface area and the horizontal projected area, respectively, in a given virtual quadrat. Let the slope at a given <inline-formula>
<mml:math display="inline" id="im33">
<mml:mi>i</mml:mi>
</mml:math>
</inline-formula> th elevation point in the sliding window be <inline-formula>
<mml:math display="inline" id="im34">
<mml:mrow>
<mml:msub>
<mml:mi>S</mml:mi>
<mml:mi>i</mml:mi>
</mml:msub>
</mml:mrow>
</mml:math>
</inline-formula>, the corresponding surface roughness <inline-formula>
<mml:math display="inline" id="im35">
<mml:mrow>
<mml:msub>
<mml:mi>R</mml:mi>
<mml:mi>s</mml:mi>
</mml:msub>
</mml:mrow>
</mml:math>
</inline-formula> could then benefit from the calculation of this available topographic factor as follows:</p>
<disp-formula>
<label>(14)</label>
<mml:math display="block" id="M14">
<mml:mrow>
<mml:msub>
<mml:mi>R</mml:mi>
<mml:mi>s</mml:mi>
</mml:msub>
<mml:mo>=</mml:mo>
<mml:mfrac>
<mml:mn>1</mml:mn>
<mml:mrow>
<mml:mi>cos</mml:mi>
<mml:mrow>
<mml:mo stretchy="false">(</mml:mo>
<mml:mrow>
<mml:msub>
<mml:mi>S</mml:mi>
<mml:mi>i</mml:mi>
</mml:msub>
</mml:mrow>
<mml:mo stretchy="false">)</mml:mo>
</mml:mrow>
</mml:mrow>
</mml:mfrac>
</mml:mrow>
</mml:math>
</disp-formula>
<p>The surface roughness <inline-formula>
<mml:math display="inline" id="im36">
<mml:mrow>
<mml:msub>
<mml:mi>R</mml:mi>
<mml:mi>s</mml:mi>
</mml:msub>
</mml:mrow>
</mml:math>
</inline-formula> of each virtual quadrat with <inline-formula>
<mml:math display="inline" id="im37">
<mml:mi>n</mml:mi>
</mml:math>
</inline-formula> elevation points can then be expressed as</p>
<disp-formula>
<label>(15)</label>
<mml:math display="block" id="M15">
<mml:mrow>
<mml:msub>
<mml:mi>R</mml:mi>
<mml:mi>s</mml:mi>
</mml:msub>
<mml:mo>=</mml:mo>
<mml:mfrac>
<mml:mn>1</mml:mn>
<mml:mi>n</mml:mi>
</mml:mfrac>
<mml:mstyle displaystyle="true">
<mml:msubsup>
<mml:mo>&#x2211;</mml:mo>
<mml:mrow>
<mml:mi>i</mml:mi>
<mml:mo>=</mml:mo>
<mml:mn>1</mml:mn>
</mml:mrow>
<mml:mi>n</mml:mi>
</mml:msubsup>
<mml:mrow>
<mml:mfrac>
<mml:mn>1</mml:mn>
<mml:mrow>
<mml:mi>cos</mml:mi>
<mml:mrow>
<mml:mo stretchy="false">(</mml:mo>
<mml:mrow>
<mml:msub>
<mml:mi>S</mml:mi>
<mml:mi>i</mml:mi>
</mml:msub>
</mml:mrow>
<mml:mo stretchy="false">)</mml:mo>
</mml:mrow>
</mml:mrow>
</mml:mfrac>
</mml:mrow>
</mml:mstyle>
</mml:mrow>
</mml:math>
</disp-formula>
</sec>
<sec id="s3_4">
<label>3.4</label>
<title>Joint morphological cues</title>
<p>It was believed that such thematic maps of topological parameters reflecting the seafloor elements and types are effective in classifying seafloor categories in terms of their formation processes and evolution (<xref ref-type="bibr" rid="B5">Burrough and McDonnell, 1998</xref>). The topological parameters of high similarity would most likely be shared with the identical seafloor categories. Since most attempts to characterize seafloor elements are limited to a relatively restricted range of morphological attributes, while seafloor types represent characteristic patterns that repeat regardless of scales (<xref ref-type="bibr" rid="B44">MacMillan et&#xa0;al., 2000</xref>), we endeavored to utilize high-resolution multibeam bathymetric mapping to extract micro geomorphologic factors such as slope and curvature, as well as macro geomorphologic factors like surface roughness, and to assess the effectiveness of individual or joint morphological cues in distinguishing seafloor surface types. It should be noted that the formation of seafloor surfaces can be viewed from a variety of spatial scales, and the effect of scales involves geomorphology in a complex, hierarchical context. Thus, seafloor classification is related to the issue of scales in different geomorphological settings and the role that morphological cues play in seafloor surface stretching (<xref ref-type="bibr" rid="B13">De Boer, 1992</xref>).</p>
<p>The thematic maps of topological parameters for a few example MBES images are shown in <xref ref-type="fig" rid="f2">
<bold>Figure&#xa0;2</bold>
</xref>, with the original images, the slope, the surface roughness, and the curvature, respectively, displayed from top to bottom. The slope of the seamount generally approached a large value with high-level relief amplitude; the slope of the trench bottom basin was relatively small with nearly flat surfaces; and the slope of the island slope deepwater terrace shifted frequently, representing the divergence of the degree of seafloor surface steepness. The surface roughness provides a macroscopic view of the complexity of seafloor surfaces and reflects the degree to which the seabed terrain is susceptible to erosion. Higher surface roughness values corresponded to more complex or eroded seafloor terrain, e.g., around the island slope deep water terrace. Conversely, flat seafloor surfaces experienced less erosion and exhibited lower roughness values. The curvature directly affected the net erosion, reflecting the degree of seafloor surface fragmentation. When the curvature value of the sea mount was relatively small, the degree of fragmentation was the lowest, and the curvature value of the island slope deep water terrace was relatively large, representing a high degree of fragmentation. The curvature directly affected the net erosion, reflecting the degree of seafloor surface fragmentation. When the curvature value of the sea mount was relatively small, the degree of fragmentation was the lowest, and the curvature value of the island slope deep water terrace was relatively large, representing a high degree of fragmentation.</p>
<fig id="f2" position="float">
<label>Figure&#xa0;2</label>
<caption>
<p>Extraction of topographic factors from MBES imagery. <bold>(A)</bold> Multibeam bathymetric topographic mapping, <bold>(B)</bold> slope, <bold>(C)</bold> surface roughness, <bold>(D)</bold> curvature.</p>
</caption>
<graphic mimetype="image" mime-subtype="tiff" xlink:href="fmars-10-1205142-g002.tif"/>
</fig>
<p>We initially utilized basic clustering techniques (K-means) to agglomeratively assign elevation points with highly similar topological parameters into the same group and to deviate from the significantly inconsistent outlier elevation points. We could therefore locate and identify individual notions of landforms and geological structures at certain scales with specific physical attributes and translate them to the complete coverage of bathymetric mapping to estimate the potentially appropriate scales as a whole for reference. The individual and joint morphological cues in combinations have served as the input to assess the clustering performances, in terms of PA, MPA, and MIoU, as is shown in <xref ref-type="supplementary-material" rid="SM1">
<bold>Supplementary Table E1</bold>
</xref>, where the first row is the clustering evaluation of only the bathymetric topographic mapping from MBES, and the second, third, and fourth rows are the evaluation results when introducing the additional morphological cues, respectively, including the slope, surface roughness and curvature. Among them, the clustering performance was superior when both slope and surface roughness were fed as inputs together with the original bathymetric mapping. The comparison of clustering performance with the individual and joint morphological cues for example MBES imaging is shown in <xref ref-type="fig" rid="f3">
<bold>Figure&#xa0;3</bold>
</xref>, with the original example images, the clustering results from bathymetric mapping+slope, +surface roughness, +curvature, and the ground truth listed from left to right respectively. It was shown that some regions of the trench seamount group were quite easily misclassified as island slopes, leading to many mistakenly divided holes. Owing to the complexity and variability of seafloor surfaces, there exist large divergences even within identical seafloor types and possible similarities across distinct seafloor types, all of which would influence the discrimination process. We have tried to integrate the joint morphological cues into the deep learning-based models to improve the accuracy of distinguishing seafloor categories.</p>
<fig id="f3" position="float">
<label>Figure&#xa0;3</label>
<caption>
<p>Clustering comparison with joint morphological cues. <bold>(A)</bold> Bathymetric topographic mapping, <bold>(B)</bold> +slope, <bold>(C)</bold> +slope+roughness, <bold>(D)</bold> +slope+roughness+curvature, <bold>(E)</bold> ground truth.</p>
</caption>
<graphic mimetype="image" mime-subtype="tiff" xlink:href="fmars-10-1205142-g003.tif"/>
</fig>
</sec>
</sec>
<sec id="s4">
<label>4</label>
<title>Network construction</title>
<sec id="s4_1">
<label>4.1</label>
<title>Basics of DenseNet</title>
<p>The basic Densely Connected Convolutional Networks (DenseNet) embraces the hypothesis that shorter connections exhibit high performance in a substantially deeper network manner (<xref ref-type="bibr" rid="B28">Huang et&#xa0;al., 2017</xref>; <xref ref-type="bibr" rid="B31">J&#xe9;gou et&#xa0;al., 2017</xref>). The feature maps of all previous layers are used as inputs for each layer, and its own feature maps are introduced as inputs to all subsequent layers. Therefore, a basic DenseNet comprising <inline-formula>
<mml:math display="inline" id="im38">
<mml:mi>L</mml:mi>
</mml:math>
</inline-formula> layers will result in <inline-formula>
<mml:math display="inline" id="im39">
<mml:mrow>
<mml:mi>L</mml:mi>
<mml:mo stretchy="false">(</mml:mo>
<mml:mi>L</mml:mi>
<mml:mo>+</mml:mo>
<mml:mn>1</mml:mn>
<mml:mo stretchy="false">)</mml:mo>
<mml:mo stretchy="false">/</mml:mo>
<mml:mn>2</mml:mn>
</mml:mrow>
</mml:math>
</inline-formula> direct connections in a feed-forward fashion. Let <inline-formula>
<mml:math display="inline" id="im40">
<mml:mrow>
<mml:msub>
<mml:mi>H</mml:mi>
<mml:mi>i</mml:mi>
</mml:msub>
<mml:mo stretchy="false">(</mml:mo>
<mml:mo>&#xb7;</mml:mo>
<mml:mo stretchy="false">)</mml:mo>
</mml:mrow>
</mml:math>
</inline-formula> be the non-linear transformation implemented in the <inline-formula>
<mml:math display="inline" id="im41">
<mml:mi>i</mml:mi>
</mml:math>
</inline-formula> th layer, with the output of the <inline-formula>
<mml:math display="inline" id="im42">
<mml:mi>i</mml:mi>
</mml:math>
</inline-formula> th layer denoted as <inline-formula>
<mml:math display="inline" id="im43">
<mml:mrow>
<mml:msub>
<mml:mi>x</mml:mi>
<mml:mi>i</mml:mi>
</mml:msub>
</mml:mrow>
</mml:math>
</inline-formula>. DenseNet proposes a dense connectivity pattern that introduces direct connections from each layer to all subsequent layers. Consequently, the <inline-formula>
<mml:math display="inline" id="im44">
<mml:mi>i</mml:mi>
</mml:math>
</inline-formula> th layer receives the feature maps of all previous layers as the input</p>
<disp-formula>
<label>(16)</label>
<mml:math display="block" id="M16">
<mml:mrow>
<mml:msub>
<mml:mi>x</mml:mi>
<mml:mi>i</mml:mi>
</mml:msub>
<mml:mo>=</mml:mo>
<mml:msub>
<mml:mi>H</mml:mi>
<mml:mi>i</mml:mi>
</mml:msub>
<mml:mrow>
<mml:mo stretchy="false">(</mml:mo>
<mml:mrow>
<mml:mrow>
<mml:mo stretchy="false">[</mml:mo>
<mml:mrow>
<mml:msub>
<mml:mi>x</mml:mi>
<mml:mn>0</mml:mn>
</mml:msub>
<mml:mo>,</mml:mo>
<mml:msub>
<mml:mi>x</mml:mi>
<mml:mn>1</mml:mn>
</mml:msub>
<mml:mo>,</mml:mo>
<mml:mo>&#x2026;</mml:mo>
<mml:mo>,</mml:mo>
<mml:msub>
<mml:mi>x</mml:mi>
<mml:mrow>
<mml:mi>i</mml:mi>
<mml:mo>&#x2212;</mml:mo>
<mml:mn>1</mml:mn>
</mml:mrow>
</mml:msub>
</mml:mrow>
<mml:mo stretchy="false">]</mml:mo>
</mml:mrow>
</mml:mrow>
<mml:mo stretchy="false">)</mml:mo>
</mml:mrow>
</mml:mrow>
</mml:math>
</disp-formula>
<p>where <inline-formula>
<mml:math display="inline" id="im45">
<mml:mrow>
<mml:mo stretchy="false">[</mml:mo>
<mml:msub>
<mml:mi>x</mml:mi>
<mml:mn>0</mml:mn>
</mml:msub>
<mml:mo>,</mml:mo>
<mml:msub>
<mml:mi>x</mml:mi>
<mml:mn>1</mml:mn>
</mml:msub>
<mml:mo>,</mml:mo>
<mml:mo>&#x2026;</mml:mo>
<mml:mo>,</mml:mo>
<mml:msub>
<mml:mi>x</mml:mi>
<mml:mrow>
<mml:mi>i</mml:mi>
<mml:mo>&#x2212;</mml:mo>
<mml:mn>1</mml:mn>
</mml:mrow>
</mml:msub>
<mml:mo stretchy="false">]</mml:mo>
</mml:mrow>
</mml:math>
</inline-formula> refers to the concatenation of the feature maps produced in the previous layers. For ease of implementation, the multiple inputs of <inline-formula>
<mml:math display="inline" id="im46">
<mml:mrow>
<mml:msub>
<mml:mi>H</mml:mi>
<mml:mi>i</mml:mi>
</mml:msub>
</mml:mrow>
</mml:math>
</inline-formula>(&#xb7;) could be concatenated into a single tensor. Since the concatenation operation may not be feasible if the size of the feature maps changes during down-sampling, DenseNet would be further divided into multiple dense blocks, with the transition layers between them for convolution and pooling.</p>
</sec>
<sec id="s4_2">
<label>4.2</label>
<title>Backbone network architecture</title>
<p>We used DenseNet121 as the backbone network of our proposed scheme for seafloor surface classification. The non-linear transformation <inline-formula>
<mml:math display="inline" id="im47">
<mml:mrow>
<mml:msub>
<mml:mi>H</mml:mi>
<mml:mi>i</mml:mi>
</mml:msub>
</mml:mrow>
</mml:math>
</inline-formula>(&#xb7;) was initially defined as a composite function of consecutive operations, i.e., Batch Normalization (BN), followed by a Rectified Linear Unit (ReLU) and a Convolution (Conv). The design of a 1&#xd7;1 convolution was introduced as a bottleneck layer before each 3&#xd7;3 convolution to improve computational efficiency. The DenseNet121 network configuration was made up of four dense blocks. Before entering the first dense block, the initial convolution layer comprised <inline-formula>
<mml:math display="inline" id="im48">
<mml:mrow>
<mml:mn>2</mml:mn>
<mml:mi>k</mml:mi>
</mml:mrow>
</mml:math>
</inline-formula> convolutions of size 7&#xd7;7 with step size 2, and the number of feature maps in all other layers followed from the setting <inline-formula>
<mml:math display="inline" id="im49">
<mml:mi>k</mml:mi>
</mml:math>
</inline-formula>. The transition layers took a 1&#xd7;1 convolution, followed by a <inline-formula>
<mml:math display="inline" id="im50">
<mml:mrow>
<mml:mn>2</mml:mn>
<mml:mo>&#xd7;</mml:mo>
<mml:mn>2</mml:mn>
</mml:mrow>
</mml:math>
</inline-formula> Average pooling between two contiguous dense blocks. At the end of the last dense block, global Average pooling was performed and then a softmax classifier was applied. The number of feature maps in the four dense blocks was 6, 12, 24, and 16, respectively, and the corresponding size of features was <inline-formula>
<mml:math display="inline" id="im51">
<mml:mrow>
<mml:mn>1</mml:mn>
<mml:mo stretchy="false">/</mml:mo>
<mml:mn>4</mml:mn>
</mml:mrow>
</mml:math>
</inline-formula>, <inline-formula>
<mml:math display="inline" id="im52">
<mml:mrow>
<mml:mn>1</mml:mn>
<mml:mo stretchy="false">/</mml:mo>
<mml:mn>8</mml:mn>
</mml:mrow>
</mml:math>
</inline-formula>, <inline-formula>
<mml:math display="inline" id="im53">
<mml:mrow>
<mml:mn>1</mml:mn>
<mml:mo stretchy="false">/</mml:mo>
<mml:mn>16</mml:mn>
</mml:mrow>
</mml:math>
</inline-formula>, <inline-formula>
<mml:math display="inline" id="im54">
<mml:mrow>
<mml:mn>1</mml:mn>
<mml:mo stretchy="false">/</mml:mo>
<mml:mn>32</mml:mn>
</mml:mrow>
</mml:math>
</inline-formula> of the original input.</p>
</sec>
</sec>
<sec id="s5">
<label>5</label>
<title>Seafloor surface classification via DenseNet</title>
<p>Our proposed scheme consists of several correlative steps, as follows: (1) Manual annotation: at the beginning, the seafloor stretching annotation of the bathymetric mapping collected from MBES scans is manually labeled as the standard reference. (2) Network construction: an upgraded Densenet121 backbone is established, in aid of the minimalistic transition-up blocks in the upsampling path, the channel attention module, and the spatial pyramid pooling, to explore the potential seafloor stretching categories. (3) Morphological feature fusion: the morphological cues are incorporated into the context of the DenseNet architecture from the extracted topological parameters. (4) The seafloor stretching classification would go through the statistics on potential biogeographic distribution and jointly improve the understanding of delineating the submarine benthic habitats. The flowchart of seafloor surface classification via DenseNet is shown in <xref ref-type="fig" rid="f4">
<bold>Figure&#xa0;4</bold>
</xref>. It should be noted that our proposed scheme could actually be extended to accommodate multiple scales of input seafloor stretching surfaces. It is well known that the parameters for describing the geomorphological formations of seafloor surfaces can be quite sensitive to the scales involved (<xref ref-type="bibr" rid="B82">Wong, 1973</xref>; <xref ref-type="bibr" rid="B57">Phillips, 1988</xref>; <xref ref-type="bibr" rid="B45">Mahmud and Yusof, 2006</xref>; <xref ref-type="bibr" rid="B50">Millar, 2013</xref>). We have taken into account the issues of scales and specifically designed the geomorphological modeling via deep learning. We have tried to integrate the site-specific scales and responded to recognize the appropriate spatial scales from the individual physical attributes of the seafloor categories during the reasoning process under the classification criterion.</p>
<fig id="f4" position="float">
<label>Figure&#xa0;4</label>
<caption>
<p>Flowchart of seafloor surface classification using DenseNet.</p>
</caption>
<graphic mimetype="image" mime-subtype="tiff" xlink:href="fmars-10-1205142-g004.tif"/>
</fig>
<sec id="s5_1">
<label>5.1</label>
<title>Upsampling path</title>
<p>DenseNet121 transforms the input into a feature tensor by gradually reducing the spatial resolution and increasing the number of feature maps along a downsampling path. As for the design and the upsampling path, the Tiramisu model has had great success in the naive extension of DenseNet to fully convolutional networks, while mitigating the linear growth of the feature map explosion in very deep neural networks with very few parameters, replacing the convolution operation with a sequence of dense blocks and the transposed convolution referred to as transition-up (TU) blocks, with an approximately 10-fold reduction with respect to the state-of-the-art models (<xref ref-type="bibr" rid="B31">J&#xe9;gou et&#xa0;al., 2017</xref>). In this paper, in order to explore the possibilities of developing smart capabilities in understanding the seafloor stretching morphology for underwater vehicles, we have updated the DenseNet architecture with an upsampling path of a more simplified transition-up process, i.e., the minimalistic transition-up blocks, which could transform the low-resolution features into high-resolution predictions by recovering details from early layers with blending semantics from deeper layers (<xref ref-type="bibr" rid="B34">Kreso et&#xa0;al., 2017</xref>). The design of minimalistic TU blocks is introduced to play the role of the upsampling path in DenseNet121. TU blocks blend the smaller and larger representations whose spatial resolutions differ by a factor of 2 from the upsampling and downsampling paths, respectively, via a skip connection. The blending procedure is repeated recursively by simple summation along the upsampling path, with skip connections arriving from the outputs of each dense block instead of the symmetric encoder-decoder network. The final TU block produces logits at the resolution of the DenseNet stem. The dense predictions at the input resolution are finally obtained by 4&#xd7; bilinear upsampling. The minimalistic design helps lightweight semantic execution with a low memory footprint and low-dimensional feature tensors during upsampling and discourages overfitting to low-level textures, which potentially presents significant online computation capacities in distinguishing seafloor categories for underwater vehicles.</p>
</sec>
<sec id="s5_2">
<label>5.2</label>
<title>Channel attention module</title>
<p>We adaptively refined the input feature maps along channels by seamlessly integrating the Convolutional Block Attention Module (CBAM) (<xref ref-type="bibr" rid="B83">Woo et&#xa0;al., 2018</xref>) into DenseNet121. The CBAM module sequentially infers channel-wise attention maps, which are multiplied by input feature maps. Unlike the Squeeze-and-Excitation (SE) module (<xref ref-type="bibr" rid="B27">Hu et&#xa0;al., 2018</xref>), we have tried to exploit the inter-channel relationships by employing both Average pooling and Max pooling in parallel. Given an intermediate feature map <inline-formula>
<mml:math display="inline" id="im55">
<mml:mi>X</mml:mi>
</mml:math>
</inline-formula> of size <inline-formula>
<mml:math display="inline" id="im56">
<mml:mrow>
<mml:mi>H</mml:mi>
<mml:mo>&#xd7;</mml:mo>
<mml:mi>W</mml:mi>
<mml:mo>&#xd7;</mml:mo>
<mml:mi>C</mml:mi>
</mml:mrow>
</mml:math>
</inline-formula>, with <inline-formula>
<mml:math display="inline" id="im57">
<mml:mi>H</mml:mi>
</mml:math>
</inline-formula>, <inline-formula>
<mml:math display="inline" id="im58">
<mml:mi>W</mml:mi>
</mml:math>
</inline-formula>, <inline-formula>
<mml:math display="inline" id="im59">
<mml:mi>C</mml:mi>
</mml:math>
</inline-formula> being the height, width, and channel number of the feature map, respectively, the spatial dimension of the feature map is squeezed as follows:</p>
<disp-formula>
<label>(17)</label>
<mml:math display="block" id="M17">
<mml:mrow>
<mml:msubsup>
<mml:mi>X</mml:mi>
<mml:mrow>
<mml:mi>a</mml:mi>
<mml:mi>v</mml:mi>
<mml:mi>g</mml:mi>
</mml:mrow>
<mml:mi>c</mml:mi>
</mml:msubsup>
<mml:mo>=</mml:mo>
<mml:mi>A</mml:mi>
<mml:mi>v</mml:mi>
<mml:mi>g</mml:mi>
<mml:mi>p</mml:mi>
<mml:mi>o</mml:mi>
<mml:mi>o</mml:mi>
<mml:mi>l</mml:mi>
<mml:mrow>
<mml:mo stretchy="false">(</mml:mo>
<mml:mi>X</mml:mi>
<mml:mo stretchy="false">)</mml:mo>
</mml:mrow>
</mml:mrow>
</mml:math>
</disp-formula>
<disp-formula>
<label>(18)</label>
<mml:math display="block" id="M18">
<mml:mrow>
<mml:msubsup>
<mml:mi>X</mml:mi>
<mml:mrow>
<mml:mi>max</mml:mi>
</mml:mrow>
<mml:mi>c</mml:mi>
</mml:msubsup>
<mml:mo>=</mml:mo>
<mml:mi>M</mml:mi>
<mml:mi>a</mml:mi>
<mml:mi>x</mml:mi>
<mml:mi>p</mml:mi>
<mml:mi>o</mml:mi>
<mml:mi>o</mml:mi>
<mml:mi>l</mml:mi>
<mml:mrow>
<mml:mo stretchy="false">(</mml:mo>
<mml:mi>X</mml:mi>
<mml:mo stretchy="false">)</mml:mo>
</mml:mrow>
</mml:mrow>
</mml:math>
</disp-formula>
<p>where <inline-formula>
<mml:math display="inline" id="im60">
<mml:mrow>
<mml:msubsup>
<mml:mi>X</mml:mi>
<mml:mrow>
<mml:mi>a</mml:mi>
<mml:mi>v</mml:mi>
<mml:mi>g</mml:mi>
</mml:mrow>
<mml:mi>c</mml:mi>
</mml:msubsup>
</mml:mrow>
</mml:math>
</inline-formula> and <inline-formula>
<mml:math display="inline" id="im61">
<mml:mrow>
<mml:msubsup>
<mml:mi>X</mml:mi>
<mml:mrow>
<mml:mi>max</mml:mi>
</mml:mrow>
<mml:mi>c</mml:mi>
</mml:msubsup>
</mml:mrow>
</mml:math>
</inline-formula> are the outputs of the Average pooling and the Max pooling, respectively, with a size of <inline-formula>
<mml:math display="inline" id="im62">
<mml:mrow>
<mml:mn>1</mml:mn>
<mml:mo>&#xd7;</mml:mo>
<mml:mn>1</mml:mn>
<mml:mo>&#xd7;</mml:mo>
<mml:mi>C</mml:mi>
</mml:mrow>
</mml:math>
</inline-formula>. The Average pooling aggregates the spatial dimension to suggest the extent of the seafloor surface stretching, and the Max pooling gathers clues of distinctive seafloor surface features to simultaneously infer finer channel-wise attention. Both descriptors allow the global receptive fields to be embedded.</p>
<p>An excitation operation, where the specific activations govern the excitation of the channels by the dependency, feeds the two descriptors into a shared multi-layer perceptron (MLP) with a hidden layer to produce the channel attention map. To reduce the parameter overhead, the hidden activation size is set to <inline-formula>
<mml:math display="inline" id="im63">
<mml:mrow>
<mml:mi>C</mml:mi>
<mml:mo stretchy="false">/</mml:mo>
<mml:mi>r</mml:mi>
</mml:mrow>
</mml:math>
</inline-formula>, where <inline-formula>
<mml:math display="inline" id="im64">
<mml:mi>r</mml:mi>
</mml:math>
</inline-formula> is the reduction ratio. The output in MLP is recovered to generate the feature vectors of size <inline-formula>
<mml:math display="inline" id="im65">
<mml:mrow>
<mml:mn>1</mml:mn>
<mml:mo>&#xd7;</mml:mo>
<mml:mn>1</mml:mn>
<mml:mo>&#xd7;</mml:mo>
<mml:mi>C</mml:mi>
</mml:mrow>
</mml:math>
</inline-formula>. After the shared MLP is applied, the feature vectors are merged by the element-wise summation. In short, channel attention is computed as</p>
<disp-formula>
<label>(19)</label>
<mml:math display="block" id="M19">
<mml:mrow>
<mml:msub>
<mml:mi>M</mml:mi>
<mml:mi>c</mml:mi>
</mml:msub>
<mml:mrow>
<mml:mo stretchy="false">(</mml:mo>
<mml:mi>X</mml:mi>
<mml:mo stretchy="false">)</mml:mo>
</mml:mrow>
<mml:mo>=</mml:mo>
<mml:mi>&#x3c3;</mml:mi>
<mml:mrow>
<mml:mo stretchy="true">(</mml:mo>
<mml:mrow>
<mml:msub>
<mml:mi>W</mml:mi>
<mml:mn>2</mml:mn>
</mml:msub>
<mml:mi>&#x3b4;</mml:mi>
<mml:mrow>
<mml:mo stretchy="true">(</mml:mo>
<mml:mrow>
<mml:msub>
<mml:mi>W</mml:mi>
<mml:mn>1</mml:mn>
</mml:msub>
<mml:mrow>
<mml:mo stretchy="true">(</mml:mo>
<mml:mrow>
<mml:msubsup>
<mml:mi>X</mml:mi>
<mml:mrow>
<mml:mi>a</mml:mi>
<mml:mi>v</mml:mi>
<mml:mi>g</mml:mi>
</mml:mrow>
<mml:mi>c</mml:mi>
</mml:msubsup>
</mml:mrow>
<mml:mo stretchy="true">)</mml:mo>
</mml:mrow>
</mml:mrow>
<mml:mo stretchy="true">)</mml:mo>
</mml:mrow>
<mml:mo>+</mml:mo>
<mml:msub>
<mml:mi>W</mml:mi>
<mml:mn>2</mml:mn>
</mml:msub>
<mml:mi>&#x3b4;</mml:mi>
<mml:mrow>
<mml:mo stretchy="true">(</mml:mo>
<mml:mrow>
<mml:msub>
<mml:mi>W</mml:mi>
<mml:mn>1</mml:mn>
</mml:msub>
<mml:mrow>
<mml:mo stretchy="true">(</mml:mo>
<mml:mrow>
<mml:msubsup>
<mml:mi>X</mml:mi>
<mml:mrow>
<mml:mi>max</mml:mi>
</mml:mrow>
<mml:mi>c</mml:mi>
</mml:msubsup>
</mml:mrow>
<mml:mo stretchy="true">)</mml:mo>
</mml:mrow>
</mml:mrow>
<mml:mo stretchy="true">)</mml:mo>
</mml:mrow>
</mml:mrow>
<mml:mo stretchy="true">)</mml:mo>
</mml:mrow>
</mml:mrow>
</mml:math>
</disp-formula>
<p>where <inline-formula>
<mml:math display="inline" id="im66">
<mml:mrow>
<mml:msub>
<mml:mi>W</mml:mi>
<mml:mn>1</mml:mn>
</mml:msub>
</mml:mrow>
</mml:math>
</inline-formula> and <inline-formula>
<mml:math display="inline" id="im67">
<mml:mrow>
<mml:msub>
<mml:mi>W</mml:mi>
<mml:mn>2</mml:mn>
</mml:msub>
</mml:mrow>
</mml:math>
</inline-formula> respectively refer to the weights of the two layers, <inline-formula>
<mml:math display="inline" id="im68">
<mml:mi>&#x3b4;</mml:mi>
</mml:math>
</inline-formula> stands for the ReLU activation function, and <inline-formula>
<mml:math display="inline" id="im69">
<mml:mi>&#x3c3;</mml:mi>
</mml:math>
</inline-formula> denotes the sigmoid function. Finally, the channel attention output <inline-formula>
<mml:math display="inline" id="im70">
<mml:mrow>
<mml:msub>
<mml:mi>M</mml:mi>
<mml:mi>c</mml:mi>
</mml:msub>
</mml:mrow>
</mml:math>
</inline-formula> is multiplied with the initial feature map <inline-formula>
<mml:math display="inline" id="im71">
<mml:mi>X</mml:mi>
</mml:math>
</inline-formula> to retrieve the newly refined features with calibration,</p>
<disp-formula>
<label>(20)</label>
<mml:math display="block" id="M20">
<mml:mrow>
<mml:msup>
<mml:mi>X</mml:mi>
<mml:mo>&#x2032;</mml:mo>
</mml:msup>
<mml:mo>=</mml:mo>
<mml:msub>
<mml:mi>M</mml:mi>
<mml:mi>c</mml:mi>
</mml:msub>
<mml:mrow>
<mml:mo stretchy="false">(</mml:mo>
<mml:mi>X</mml:mi>
<mml:mo stretchy="false">)</mml:mo>
</mml:mrow>
<mml:mo>&#x2297;</mml:mo>
<mml:mi>X</mml:mi>
</mml:mrow>
</mml:math>
</disp-formula>
<p>where <inline-formula>
<mml:math display="inline" id="im72">
<mml:mo>&#x2297;</mml:mo>
</mml:math>
</inline-formula> denotes the element-wise multiplication. The weight coefficient from the channel attention values is broadcast along the spatial dimension during the multiplication to adaptively screen the optimal feature map along the channels.</p>
</sec>
<sec id="s5_3">
<label>5.3</label>
<title>Pyramid pooling module</title>
<p>We have further embraced the idea of a kind of spatial pyramid pooling module (SPP) (<xref ref-type="bibr" rid="B24">He et&#xa0;al., 2015</xref>) into our DenseNet121 architecture since it may not sufficiently incorporate the momentous global contextual prior for the receptive fields of the seafloor surface stretching, especially on high-level layers. The basic module of the pyramid scene parsing network (PSPNet) is developed to help exploit and enhance the capability of global context-aware features through aggregation along with sub-regions from multiple receptive fields. We have proposed the introduction of a global context with a sub-region context that enriches to distinguish seafloor surface categories in a pyramidal manner, using both the Average pooling and the Max pooling, as is shown in <xref ref-type="supplementary-material" rid="SM1">
<bold>Supplementary Figure H1</bold>
</xref>.</p>
<p>Let the number of channels from the channel attention module be <inline-formula>
<mml:math display="inline" id="im73">
<mml:mrow>
<mml:msub>
<mml:mi>C</mml:mi>
<mml:mi>D</mml:mi>
</mml:msub>
</mml:mrow>
</mml:math>
</inline-formula>; the dimensionality reduction is first performed on the input feature maps by a 1&#xd7;1 convolution. The Average pooling and the Max pooling simultaneously conclude the feature maps in sub-regions of pyramid scales, with the latter appropriately compensating for the former in detail, and then connect together at pyramid levels along the channel dimension. To maintain the weight of the global seafloor features, a 1&#xd7;1 convolution layer is applied after each pyramid level. The low-dimensional feature maps are directly upsampled to obtain feature maps of the same size before pooling by bilinear interpolation. Multiple levels of pyramid pooling features are concatenated with the original feature maps before the pooling stage as the final globally enhanced seafloor features, and then output with <inline-formula>
<mml:math display="inline" id="im74">
<mml:mrow>
<mml:mn>1</mml:mn>
<mml:mo>&#xd7;</mml:mo>
<mml:mn>1</mml:mn>
<mml:mo>&#xd7;</mml:mo>
</mml:mrow>
</mml:math>
</inline-formula> <inline-formula>
<mml:math display="inline" id="im75">
<mml:mrow>
<mml:msub>
<mml:mi>C</mml:mi>
<mml:mi>D</mml:mi>
</mml:msub>
<mml:mo stretchy="false">/</mml:mo>
<mml:mn>4</mml:mn>
</mml:mrow>
</mml:math>
</inline-formula> convolution for the next upsampling.</p>
</sec>
<sec id="s5_4">
<label>5.4</label>
<title>Morphological feature fusion</title>
<p>We further proposed fusion strategies to merge with the morphological cues in the context of DenseNet so as to enhance the semantic understanding among seafloor surface types, as is shown in <xref ref-type="supplementary-material" rid="SM1">
<bold>Supplementary Figures I1</bold>
</xref> and <xref ref-type="supplementary-material" rid="SM1">
<bold>I2</bold>
</xref>. The first one is that we have attempted to superimpose the morphological features as the input of DenseNet together with the bathymetric seafloor mapping, calibrating the deep-level feature mapping with the help of the channel attention module, enhancing the global feature extraction from the spatial pyramid pooling, and restoring the high-resolution predictions in the up-sampling path for the pixel-level seafloor surface classification. The second strategy is to make an up-sampling of those morphological features through a <inline-formula>
<mml:math display="inline" id="im76">
<mml:mrow>
<mml:mn>1</mml:mn>
<mml:mo>&#xd7;</mml:mo>
<mml:mn>1</mml:mn>
</mml:mrow>
</mml:math>
</inline-formula> convolution as a branch to join with the deep-level feature mapping of the same dimensionality extracted from the DenseNet branch to jointly contribute as the input of the residual block for the subsequent seafloor type prediction. Due to the existence of the identity mapping in ResNet, the residual block could at least copy the previous layer to prevent degradation and simultaneously refine morphological details. In addition, we have evaluated the impact of multiple morphological cues on promoting the descriptiveness and distinguishability of seafloor surface classification.</p>
</sec>
</sec>
<sec id="s6">
<label>6</label>
<title>Experimental results and analysis</title>
<sec id="s6_1">
<label>6.1</label>
<title>Dataset for seafloor surface classification</title>
<p>In our simulation experiment, the developed scheme has been verified by the high-resolution multibeam bathymetric data from the NOAA Office of Ocean Exploration and Research (OER) for the expeditions EX1605L1, EX1605L2, and EX1605L3, with Kongsberg EM302 multibeam echosounders on board the research vessel Okeanos Explorer. The total time of the expedition is 1631.269&#xa0;h, lasting for 59 days, from the 20th of April to the 10th of July 2016, with a track length of 26703.6897&#xa0;km and an average speed of 16.33 km/h in the Mariana Trench Marine National Monument and the Commonwealth of the Northern Mariana Islands, as is shown in <xref ref-type="supplementary-material" rid="SM1">
<bold>Supplementary Table K1</bold>
</xref>.</p>
<p>Meanwhile, the submersible ROV Deep Discoverer (D2), equipped with high-definition cameras and a lighting system, was connected to the camera platform Seirios and the research vessel via an umbilical cable, which provided the possibility of visual cues about the benthic habitat and colonization that are difficult to obtain in the deep sea (<xref ref-type="bibr" rid="B6">Cantwell, 2016</xref>). The detailed summary of the ROV Deep Discoverer dive log of EX1605L3 is listed in <xref ref-type="supplementary-material" rid="SM1">
<bold>Supplementary Table L1</bold>
</xref>, and it includes the latitude and longitude, bottom time, and maximum depth.</p>
<p>First, we essentially utilized the manually labeled seafloor surface annotation as the standard reference so as to identify eight seafloor stretching categories via DenseNet. The normative standard of our manual annotation is listed in <xref ref-type="table" rid="T1"><bold>Table 1</bold></xref>, where the descriptive morphological formation features are commonly known to systematically evaluate the seafloor surface categories (<xref ref-type="bibr" rid="B54">Nishizawa et al., 2009</xref>; <xref ref-type="bibr" rid="B23">Harris et al., 2014</xref>). In the beginning, we divided the original MBES images into overlapping sub-blocks based on their relatively independent physical attributes of morphological structures at the given scales. We normalized the above MBES images at multiple scales, with their corresponding morphological cues and manual labeling into the basic uniform size 256&#xd7;256. Such transformed sub-blocks were varied with multiple processing steps, such as random flip, rotation, translation, etc., to promote the diversity of the samples. The selection of the basic uniform size satisfied a comprehensive view of most seafloor topography in our experiment, allowing for interpretation, classification, and validation under the given average swath width of MBES scans. Once a variety of scales with regard to geomorphological formations of seafloor surfaces have been used, normalization would be taken to adapt to the proposed model. In total, 11,720 sub-blocks were chosen, with 8200 samples for training and 3520 for testing, of which 697 samples were originally labeled to the island slope ridge category, 2765 samples to the island slope category, 1145 samples to the island slope deep water terrace category, 2682 samples to the trench seamount group category, 1690 samples to the trench edge slope category, 1240 samples to the trench bottom basin category, 840 samples to the island platform category, and 661 samples to the slope fault basin category. We could further accumulate and refine the seafloor surface annotation as the ground truth through the acquisition of more MBES images.</p>
<table-wrap id="T1" position="float">
<label>Table&#xa0;1</label>
<caption>
<p>Annotation standard for seafloor surface categories.</p>
</caption>
<table frame="hsides">
<thead>
<tr>
<th valign="middle" align="center">Category</th>
<th valign="middle" align="center">Category features</th>
</tr>
</thead>
<tbody>
<tr>
<td valign="middle" align="center">
<bold>Island slope</bold>
</td>
<td valign="middle" align="left">The deepening of the seafloor out from the shelf edge to the upper limit of the continental rise around an island, or the point where there is a general decrease in steepness.</td>
</tr>
<tr>
<td valign="middle" align="center">
<bold>Island slope ridge</bold>
</td>
<td valign="middle" align="left">An isolated or a group of elongated, narrow elevations of varying complexity with steep sides around an island (length/width ratio &gt;2).</td>
</tr>
<tr>
<td valign="middle" align="center">
<bold>Island slope deep water terrace</bold>
</td>
<td valign="middle" align="left">An isolated or a group of relatively flat horizontal or gently inclined surfaces formed by the sloping area around an island.</td>
</tr>
<tr>
<td valign="middle" align="center">
<bold>Trench seamount group</bold>
</td>
<td valign="middle" align="left">A discrete or a group of large elevations of greater relief above the seafloor, characterized by conical shape (length/width ratio&lt;2), located on the trench axis, usually where plate subduction occurs.</td>
</tr>
<tr>
<td valign="middle" align="center">
<bold>Trench edge slope</bold>
</td>
<td valign="middle" align="left">The sloping area on either side of a trench.</td>
</tr>
<tr>
<td valign="middle" align="center">
<bold>Trench bottom basin</bold>
</td>
<td valign="middle" align="left">A depression, at the deepest part of a trench in the seafloor, is more or less equidimensional in plane and of variable extent defined by a closed bathymetric contour.</td>
</tr>
<tr>
<td valign="middle" align="center">
<bold>Slope fault basin</bold>
</td>
<td valign="middle" align="left">A subsided area controlled by faults within the slope area.</td>
</tr>
<tr>
<td valign="middle" align="center">
<bold>Island platform</bold>
</td>
<td valign="middle" align="left">The area of shallow water around an island.</td>
</tr>
</tbody>
</table>
</table-wrap>
</sec>
<sec id="s6_2">
<label>6.2</label>
<title>Configuration details</title>
<p>The configuration of the supercomputing solutions during the model building, training, and testing process was as follows: NVIDIA TITAN Xp graphics card and GeForce GTX 1080Ti graphics cards, an Intel Core i5-2410M CPU with a main frequency of 2.3GHZ, 32GB of memory cards, an Ubuntu 16.04 operating system, a Tensorflow 1.3.0 deep learning framework, a Python3.5 interpreter, data science libraries including Numpy and Pandas, and netCDF data viewers. For optimization, the best Adam optimizer was adopted, among which the exponential decay rate of the first-order moment estimation <inline-formula>
<mml:math display="inline" id="im78">
<mml:mrow>
<mml:msub>
<mml:mi>&#x3b2;</mml:mi>
<mml:mn>1</mml:mn>
</mml:msub>
</mml:mrow>
</mml:math>
</inline-formula> and the second-order moment estimation <inline-formula>
<mml:math display="inline" id="im79">
<mml:mrow>
<mml:msub>
<mml:mi>&#x3b2;</mml:mi>
<mml:mn>2</mml:mn>
</mml:msub>
</mml:mrow>
</mml:math>
</inline-formula> were 0.9 and 0.99, respectively, by using the cross entropy as the loss function, the learning rate was initially set to 0.001, with the batch size of 16. It should be noted that we examined the hyper-parameters in our simulation experiment, especially the learning rate and the batch size, to ensure the impact on the convergence of our developed model. When the batch size varied from 8 to 32 and the learning rate varied from 0.0005 to 0.01, it was demonstrated from our experimental results that the selected parameters exhibited quite comparable convergence for our proposed scheme.</p>
</sec>
<sec id="s6_3">
<label>6.3</label>
<title>Evaluation metrics</title>
<p>We employed PA, MPA, and MIoU metrics to quantify semantic segmentation performance with the help of manual annotation. Assuming that there are <inline-formula>
<mml:math display="inline" id="im80">
<mml:mi>k</mml:mi>
</mml:math>
</inline-formula> categories of seafloor surfaces, let <inline-formula>
<mml:math display="inline" id="im81">
<mml:mrow>
<mml:msub>
<mml:mi>n</mml:mi>
<mml:mrow>
<mml:mi>i</mml:mi>
<mml:mi>j</mml:mi>
</mml:mrow>
</mml:msub>
</mml:mrow>
</mml:math>
</inline-formula> be the total number of image pixels that originally belonged to the <inline-formula>
<mml:math display="inline" id="im82">
<mml:mi>i</mml:mi>
</mml:math>
</inline-formula> th category but have been incorrectly classified into the <inline-formula>
<mml:math display="inline" id="im83">
<mml:mi>j</mml:mi>
</mml:math>
</inline-formula> th category, and <inline-formula>
<mml:math display="inline" id="im84">
<mml:mrow>
<mml:msub>
<mml:mi>n</mml:mi>
<mml:mrow>
<mml:mi>j</mml:mi>
<mml:mi>i</mml:mi>
</mml:mrow>
</mml:msub>
</mml:mrow>
</mml:math>
</inline-formula> be the total number of image pixels that originally belonged to the <inline-formula>
<mml:math display="inline" id="im85">
<mml:mi>j</mml:mi>
</mml:math>
</inline-formula> th category but have been incorrectly classified into the <inline-formula>
<mml:math display="inline" id="im86">
<mml:mi>i</mml:mi>
</mml:math>
</inline-formula> th category, with <inline-formula>
<mml:math display="inline" id="im87">
<mml:mrow>
<mml:msub>
<mml:mi>n</mml:mi>
<mml:mrow>
<mml:mi>i</mml:mi>
<mml:mi>i</mml:mi>
</mml:mrow>
</mml:msub>
</mml:mrow>
</mml:math>
</inline-formula> the total number of image pixels that belonged to the <inline-formula>
<mml:math display="inline" id="im88">
<mml:mi>i</mml:mi>
</mml:math>
</inline-formula> th category and have been correctly classified into the <inline-formula>
<mml:math display="inline" id="im89">
<mml:mi>i</mml:mi>
</mml:math>
</inline-formula> th category.</p>
<p>PA refers to the ratio between the amount of properly classified image pixels and the total number, which can be expressed as the following formula:</p>
<disp-formula>
<label>(21)</label>
<mml:math display="block" id="M21">
<mml:mrow>
<mml:mi>P</mml:mi>
<mml:mi>A</mml:mi>
<mml:mo>=</mml:mo>
<mml:mfrac>
<mml:mrow>
<mml:mstyle displaystyle="true">
<mml:msubsup>
<mml:mo>&#x2211;</mml:mo>
<mml:mrow>
<mml:mi>i</mml:mi>
<mml:mo>=</mml:mo>
<mml:mn>1</mml:mn>
</mml:mrow>
<mml:mi>k</mml:mi>
</mml:msubsup>
<mml:mrow>
<mml:msub>
<mml:mi>n</mml:mi>
<mml:mrow>
<mml:mi>i</mml:mi>
<mml:mi>i</mml:mi>
</mml:mrow>
</mml:msub>
</mml:mrow>
</mml:mstyle>
</mml:mrow>
<mml:mrow>
<mml:mstyle displaystyle="true">
<mml:msubsup>
<mml:mo>&#x2211;</mml:mo>
<mml:mrow>
<mml:mi>i</mml:mi>
<mml:mo>=</mml:mo>
<mml:mn>1</mml:mn>
</mml:mrow>
<mml:mi>k</mml:mi>
</mml:msubsup>
<mml:mrow>
<mml:mstyle displaystyle="true">
<mml:msubsup>
<mml:mo>&#x2211;</mml:mo>
<mml:mrow>
<mml:mi>j</mml:mi>
<mml:mo>=</mml:mo>
<mml:mn>1</mml:mn>
</mml:mrow>
<mml:mi>k</mml:mi>
</mml:msubsup>
<mml:mrow>
<mml:msub>
<mml:mi>n</mml:mi>
<mml:mrow>
<mml:mi>i</mml:mi>
<mml:mi>j</mml:mi>
</mml:mrow>
</mml:msub>
</mml:mrow>
</mml:mstyle>
</mml:mrow>
</mml:mstyle>
</mml:mrow>
</mml:mfrac>
</mml:mrow>
</mml:math>
</disp-formula>
<p>MPA refers to the ratio of the number of correctly classified image pixels on a per-category basis, which is then averaged over the total number of categories,</p>
<disp-formula>
<label>(22)</label>
<mml:math display="block" id="M22">
<mml:mrow>
<mml:mi>M</mml:mi>
<mml:mi>P</mml:mi>
<mml:mi>A</mml:mi>
<mml:mo>=</mml:mo>
<mml:mfrac>
<mml:mn>1</mml:mn>
<mml:mi>k</mml:mi>
</mml:mfrac>
<mml:mstyle displaystyle="true">
<mml:msubsup>
<mml:mo>&#x2211;</mml:mo>
<mml:mrow>
<mml:mi>i</mml:mi>
<mml:mo>=</mml:mo>
<mml:mn>1</mml:mn>
</mml:mrow>
<mml:mi>k</mml:mi>
</mml:msubsup>
<mml:mrow>
<mml:mfrac>
<mml:mrow>
<mml:msub>
<mml:mi>n</mml:mi>
<mml:mrow>
<mml:mi>i</mml:mi>
<mml:mi>i</mml:mi>
</mml:mrow>
</mml:msub>
</mml:mrow>
<mml:mrow>
<mml:mstyle displaystyle="true">
<mml:msubsup>
<mml:mo>&#x2211;</mml:mo>
<mml:mrow>
<mml:mi>j</mml:mi>
<mml:mo>=</mml:mo>
<mml:mn>1</mml:mn>
</mml:mrow>
<mml:mi>k</mml:mi>
</mml:msubsup>
<mml:mrow>
<mml:msub>
<mml:mi>n</mml:mi>
<mml:mrow>
<mml:mi>i</mml:mi>
<mml:mi>j</mml:mi>
</mml:mrow>
</mml:msub>
</mml:mrow>
</mml:mstyle>
</mml:mrow>
</mml:mfrac>
</mml:mrow>
</mml:mstyle>
</mml:mrow>
</mml:math>
</disp-formula>
<p>MIoU calculates the average IoU ratio across all categories, which describes the degree of overlap ratio between the intersection and union of categories,</p>
<disp-formula>
<label>(23)</label>
<mml:math display="block" id="M23">
<mml:mrow>
<mml:mi>I</mml:mi>
<mml:mi>o</mml:mi>
<mml:mi>U</mml:mi>
<mml:mo>=</mml:mo>
<mml:mfrac>
<mml:mrow>
<mml:msub>
<mml:mi>n</mml:mi>
<mml:mrow>
<mml:mi>i</mml:mi>
<mml:mi>i</mml:mi>
</mml:mrow>
</mml:msub>
</mml:mrow>
<mml:mrow>
<mml:mstyle displaystyle="true">
<mml:msubsup>
<mml:mo>&#x2211;</mml:mo>
<mml:mrow>
<mml:mi>j</mml:mi>
<mml:mo>=</mml:mo>
<mml:mn>1</mml:mn>
</mml:mrow>
<mml:mi>k</mml:mi>
</mml:msubsup>
<mml:mrow>
<mml:msub>
<mml:mi>n</mml:mi>
<mml:mrow>
<mml:mi>i</mml:mi>
<mml:mi>j</mml:mi>
</mml:mrow>
</mml:msub>
<mml:mo>+</mml:mo>
<mml:mstyle displaystyle="true">
<mml:msubsup>
<mml:mo>&#x2211;</mml:mo>
<mml:mrow>
<mml:mi>j</mml:mi>
<mml:mo>=</mml:mo>
<mml:mn>1</mml:mn>
</mml:mrow>
<mml:mi>k</mml:mi>
</mml:msubsup>
<mml:mrow>
<mml:msub>
<mml:mi>n</mml:mi>
<mml:mrow>
<mml:mi>j</mml:mi>
<mml:mi>i</mml:mi>
</mml:mrow>
</mml:msub>
<mml:mo>&#x2212;</mml:mo>
<mml:msub>
<mml:mi>n</mml:mi>
<mml:mrow>
<mml:mi>i</mml:mi>
<mml:mi>i</mml:mi>
</mml:mrow>
</mml:msub>
</mml:mrow>
</mml:mstyle>
</mml:mrow>
</mml:mstyle>
</mml:mrow>
</mml:mfrac>
</mml:mrow>
</mml:math>
</disp-formula>
<disp-formula>
<label>(24)</label>
<mml:math display="block" id="M24">
<mml:mrow>
<mml:mi>M</mml:mi>
<mml:mi>I</mml:mi>
<mml:mi>o</mml:mi>
<mml:mi>U</mml:mi>
<mml:mo>=</mml:mo>
<mml:mfrac>
<mml:mn>1</mml:mn>
<mml:mi>k</mml:mi>
</mml:mfrac>
<mml:mstyle displaystyle="true">
<mml:msubsup>
<mml:mo>&#x2211;</mml:mo>
<mml:mrow>
<mml:mi>i</mml:mi>
<mml:mo>=</mml:mo>
<mml:mn>1</mml:mn>
</mml:mrow>
<mml:mi>k</mml:mi>
</mml:msubsup>
<mml:mrow>
<mml:mfrac>
<mml:mrow>
<mml:msub>
<mml:mi>n</mml:mi>
<mml:mrow>
<mml:mi>i</mml:mi>
<mml:mi>i</mml:mi>
</mml:mrow>
</mml:msub>
</mml:mrow>
<mml:mrow>
<mml:mstyle displaystyle="true">
<mml:msubsup>
<mml:mo>&#x2211;</mml:mo>
<mml:mrow>
<mml:mi>j</mml:mi>
<mml:mo>=</mml:mo>
<mml:mn>1</mml:mn>
</mml:mrow>
<mml:mi>k</mml:mi>
</mml:msubsup>
<mml:mrow>
<mml:msub>
<mml:mi>n</mml:mi>
<mml:mrow>
<mml:mi>i</mml:mi>
<mml:mi>j</mml:mi>
</mml:mrow>
</mml:msub>
<mml:mo>+</mml:mo>
<mml:mstyle displaystyle="true">
<mml:msubsup>
<mml:mo>&#x2211;</mml:mo>
<mml:mrow>
<mml:mi>j</mml:mi>
<mml:mo>=</mml:mo>
<mml:mn>1</mml:mn>
</mml:mrow>
<mml:mi>k</mml:mi>
</mml:msubsup>
<mml:mrow>
<mml:msub>
<mml:mi>n</mml:mi>
<mml:mrow>
<mml:mi>j</mml:mi>
<mml:mi>i</mml:mi>
</mml:mrow>
</mml:msub>
<mml:mo>&#x2212;</mml:mo>
<mml:msub>
<mml:mi>n</mml:mi>
<mml:mrow>
<mml:mi>i</mml:mi>
<mml:mi>i</mml:mi>
</mml:mrow>
</mml:msub>
</mml:mrow>
</mml:mstyle>
</mml:mrow>
</mml:mstyle>
</mml:mrow>
</mml:mfrac>
</mml:mrow>
</mml:mstyle>
</mml:mrow>
</mml:math>
</disp-formula>
<p>We further started to evaluate the semantic segmentation performance of our proposed scheme. First, we verified the configuration of a variety of backbone networks, such as ResNet50, ResNet101, and DenseNet121, to determine whether it would be more effective to extract the possibly deeper level features for the seafloor surface stretching by means of the identical upsampling modules. As shown in <xref ref-type="supplementary-material" rid="SM1">
<bold>Supplementary Figure J1</bold>
</xref>, the selection of DenseNet121 initially achieved comparable performance for semantic segmentation of seafloor surface stretching in terms of PA, MPA, and MIoU metrics.</p>
</sec>
<sec id="s6_4">
<label>6.4</label>
<title>Ablation studies</title>
<p>We carried out a series of ablation studies to quantitatively investigate the extent to which the progress of semantic segmentation performance could benefit individually from the improvement of the channel attention module and spatial pyramid pooling in our proposed model. The performance verification for each step is listed in <xref ref-type="table" rid="T2">
<bold>Tables&#xa0;2</bold>
</xref> and <xref ref-type="table" rid="T3">
<bold>3</bold>
</xref> in our ablation studies, respectively, in terms of PA, MPA, and MIoU metrics. The channel attention module combined both global average pooling and global maximum pooling to optimize the generation of the deep-level feature descriptors. We made the comparative evaluation of the baseline Densenet121, with either the global average pooling or the global maximum pooling, as well as with both types of the pooling. In our experimental results, it has been demonstrated that the effectiveness of both types of pooling behaved better in parallel, where the maximum pooling supplied the possible losses derived from the Average pooling. Spatial pyramid pooling was added to the baseline Densenet121 with the channel attention module, using various pooling selections at multiple pyramid scales. It was shown that the Average pooling alone outperformed the maximum pooling alone, while the two complementary poolings in parallel improved the semantic segmentation accuracy more.</p>
<table-wrap id="T2" position="float">
<label>Table&#xa0;2</label>
<caption>
<p>Evaluation of the channel attention module in our ablation studies.</p>
</caption>
<table frame="hsides">
<thead>
<tr>
<th valign="middle" align="center">Method</th>
<th valign="middle" align="center">PA (%)</th>
<th valign="middle" align="center">MPA (%)</th>
<th valign="middle" align="center">MIoU (%)</th>
</tr>
</thead>
<tbody>
<tr>
<td valign="middle" align="center">
<bold>Baseline</bold>
</td>
<td valign="middle" align="center">87.63</td>
<td valign="middle" align="center">79.65</td>
<td valign="middle" align="center">70.92</td>
</tr>
<tr>
<td valign="middle" align="center">
<bold>+cha. average pooling</bold>
</td>
<td valign="middle" align="center">88.12</td>
<td valign="middle" align="center">80.02</td>
<td valign="middle" align="center">71.45</td>
</tr>
<tr>
<td valign="middle" align="center">
<bold>+cha. max pooling</bold>
</td>
<td valign="middle" align="center">87.91</td>
<td valign="middle" align="center">79.96</td>
<td valign="middle" align="center">71.13</td>
</tr>
<tr>
<td valign="middle" align="center">
<bold>+cha. average pooling &amp; max pooling</bold>
</td>
<td valign="middle" align="center">
<bold>88.55</bold>
</td>
<td valign="middle" align="center">
<bold>80.42</bold>
</td>
<td valign="middle" align="center">
<bold>71.84</bold>
</td>
</tr>
</tbody>
</table>
</table-wrap>
<table-wrap id="T3" position="float">
<label>Table&#xa0;3</label>
<caption>
<p>Evaluation of the channel attention module in our ablation studies.</p>
</caption>
<table frame="hsides">
<thead>
<tr>
<th valign="middle" align="center">Method</th>
<th valign="middle" align="center">PA (%)</th>
<th valign="middle" align="center">MPA (%)</th>
<th valign="middle" align="center">MIoU (%)</th>
</tr>
</thead>
<tbody>
<tr>
<td valign="middle" align="center">
<bold>Baseline+ Channel</bold>
</td>
<td valign="middle" align="center">88.55</td>
<td valign="middle" align="center">80.42</td>
<td valign="middle" align="center">71.84</td>
</tr>
<tr>
<td valign="middle" align="center">
<bold>+ spatial average pooling</bold>
</td>
<td valign="middle" align="center">89.04</td>
<td valign="middle" align="center">80.87</td>
<td valign="middle" align="center">72.21</td>
</tr>
<tr>
<td valign="middle" align="center">
<bold>+ spatial max pooling</bold>
</td>
<td valign="middle" align="center">88.89</td>
<td valign="middle" align="center">80.76</td>
<td valign="middle" align="center">71.94</td>
</tr>
<tr>
<td valign="middle" align="center">
<bold>+ spatial average pooling &amp; max pooling</bold>
</td>
<td valign="middle" align="center">
<bold>89.21</bold>
</td>
<td valign="middle" align="center">
<bold>81.25</bold>
</td>
<td valign="middle" align="center">
<bold>72.71</bold>
</td>
</tr>
</tbody>
</table>
</table-wrap>
<p>We also examined which types of morphological cues are more relevant to the semantic segmentation of seafloor surface categories, together with the features retrieved directly from MBES imagery via Densenet. <xref ref-type="table" rid="T4">
<bold>Table&#xa0;4</bold>
</xref> lists the evaluation of the semantic segmentation accuracy by merging multiple morphological cues into the DenseNet backbone network with the embedded channel attention module (C) and spatial pyramid pooling module (S) in the context of two types of feature fusion strategies, including slope, roughness, curvature, slope + roughness, slope + curvature, roughness + curvature, and slope + roughness + curvature. The first mode concatenated the individual or joint morphological cues with multibeam bathymetric seafloor mapping in advance to generate the multi-channel input for DenseNet, and the resulting fused feature maps would be adaptively optimized with the channel attention module, advanced into global feature representation with the spatial pyramid pooling, then restored to high-resolution predictions from up-sampling with the aid of transition-up blocks, outputting the subsequent semantic seafloor classification. In the second mode, the bathymetric seafloor mapping was individually input into DenseNet, with the channel attention module and spatial pyramid pooling employed. Concatenated feature maps of the same dimensionality were extracted from up-sampled morphological cues by convolution from another branch in parallel, and then commonly fed the feature fusion into the residual block to output the seafloor type prediction. From our experimental results, the second mode achieved the overall performance improvement compared to the baseline and the first mode, which to a certain extent plays a role in compensating for the loss in down-sampling, thereby improving the descriptiveness and distinguishability of seafloor surface categories. The morphological cues of slope + roughness exhibited better performance, while the curvature did not show a significant improvement in accuracy. It was inferred that the slope tends to indicate the degree of steepness in seafloor surface stretching, and the surface roughness might display the extent of erosion in the seafloor surface topography, all of which contribute to the semantic segmentation. Also, the curvature reflects the degree of fragmentation, which might not be seen as a very distinguishable index and might lead to misclassification to a large extent.</p>
<table-wrap id="T4" position="float">
<label>Table&#xa0;4</label>
<caption>
<p>Evaluation of feature fusion in our ablation studies.</p>
</caption>
<table frame="hsides">
<thead>
<tr>
<th valign="middle" colspan="2" align="center">Method</th>
<th valign="middle" align="center">PA (%)</th>
<th valign="middle" align="center">MPA (%)</th>
<th valign="middle" align="center">MIoU (%)</th>
</tr>
</thead>
<tbody>
<tr>
<td valign="middle" colspan="2" align="center">
<bold>Baseline+CS</bold>
</td>
<td valign="middle" align="center">89.21</td>
<td valign="middle" align="center">81.25</td>
<td valign="middle" align="center">72.71</td>
</tr>
<tr>
<td valign="middle" rowspan="7" align="center">
<bold>Plan one</bold>
</td>
<td valign="middle" align="center">+slope</td>
<td valign="middle" align="center">89.47</td>
<td valign="middle" align="center">81.38</td>
<td valign="middle" align="center">72.92</td>
</tr>
<tr>
<td valign="middle" align="center">+roughness</td>
<td valign="middle" align="center">89.35</td>
<td valign="middle" align="center">81.31</td>
<td valign="middle" align="center">72.76</td>
</tr>
<tr>
<td valign="middle" align="center">+curvature</td>
<td valign="middle" align="center">89.17</td>
<td valign="middle" align="center">81.18</td>
<td valign="middle" align="center">72.58</td>
</tr>
<tr>
<td valign="middle" align="center">+slope + roughness</td>
<td valign="middle" align="center">
<bold>89.85</bold>
</td>
<td valign="middle" align="center">81.76</td>
<td valign="middle" align="center">73.29</td>
</tr>
<tr>
<td valign="middle" align="center">+slope + curvature</td>
<td valign="middle" align="center">89.48</td>
<td valign="middle" align="center">81.41</td>
<td valign="middle" align="center">72.83</td>
</tr>
<tr>
<td valign="middle" align="center">+ roughness + curvature</td>
<td valign="middle" align="center">89.39</td>
<td valign="middle" align="center">81.33</td>
<td valign="middle" align="center">72.73</td>
</tr>
<tr>
<td valign="middle" align="center">+ slope + roughness + curvature</td>
<td valign="middle" align="center">89.59</td>
<td valign="middle" align="center">
<bold>81.47</bold>
</td>
<td valign="middle" align="center">
<bold>73.02</bold>
</td>
</tr>
<tr>
<td valign="middle" rowspan="7" align="center">
<bold>Plan two</bold>
</td>
<td valign="middle" align="center">+slope</td>
<td valign="middle" align="center">89.74</td>
<td valign="middle" align="center">81.72</td>
<td valign="middle" align="center">73.24</td>
</tr>
<tr>
<td valign="middle" align="center">+roughness</td>
<td valign="middle" align="center">89.56</td>
<td valign="middle" align="center">81.52</td>
<td valign="middle" align="center">73.03</td>
</tr>
<tr>
<td valign="middle" align="center">+curvature</td>
<td valign="middle" align="center">89.18</td>
<td valign="middle" align="center">81.20</td>
<td valign="middle" align="center">72.65</td>
</tr>
<tr>
<td valign="middle" align="center">+slope + roughness</td>
<td valign="middle" align="center">
<bold>89.87</bold>
</td>
<td valign="middle" align="center">
<bold>82.01</bold>
</td>
<td valign="middle" align="center">
<bold>73.52</bold>
</td>
</tr>
<tr>
<td valign="middle" align="center">+slope + curvature</td>
<td valign="middle" align="center">89.72</td>
<td valign="middle" align="center">81.64</td>
<td valign="middle" align="center">73.13</td>
</tr>
<tr>
<td valign="middle" align="center">+ roughness + curvature</td>
<td valign="middle" align="center">89.47</td>
<td valign="middle" align="center">81.41</td>
<td valign="middle" align="center">72.89</td>
</tr>
<tr>
<td valign="middle" align="center">+ slope + roughness + curvature</td>
<td valign="middle" align="center">89.85</td>
<td valign="middle" align="center">
<bold>82.01</bold>
</td>
<td valign="middle" align="center">73.51</td>
</tr>
</tbody>
</table>
<table-wrap-foot>
<fn id="fnT1_1">
<p>The bold values highlight the scores that fit the best performances.</p>
</fn>
</table-wrap-foot>
</table-wrap>
<p>Furthermore, the semantic segmentation accuracy of each individual seafloor surface category was systematically evaluated against the classic Fully Convolutional Network (FCN) (<xref ref-type="bibr" rid="B40">Long et&#xa0;al., 2015</xref>) in terms of the IoU measure, as is shown in <xref ref-type="table" rid="T5">
<bold>Table&#xa0;5</bold>
</xref>. Since IoU describes the degree of overlap between the actual outputs and the expected outputs of each category prediction, the higher the IoU index value, the better the segmentation performance. It could be seen from our experimental results that our proposed scheme has made general progress in distinguishing individual seafloor surface categories from the rest via DenseNet with the channel attention module and spatial pyramid pooling strategies, which tended to enhance the feature mapping capability of the entire network and thus improve the segmentation accuracy accordingly.</p>
<table-wrap id="T5" position="float">
<label>Table&#xa0;5</label>
<caption>
<p>Performance of semantic segmentation across categories.</p>
</caption>
<table frame="hsides">
<thead>
<tr>
<th valign="middle" align="center">Category</th>
<th valign="middle" align="center">FCN-8s IoU(%)</th>
<th valign="middle" align="center">Ours IoU(%)</th>
</tr>
</thead>
<tbody>
<tr>
<td valign="middle" align="center">
<bold>Island slope ridge</bold>
</td>
<td valign="middle" align="center">68.75</td>
<td valign="middle" align="center">
<bold>75.12</bold>
</td>
</tr>
<tr>
<td valign="middle" align="center">
<bold>Island slope</bold>
</td>
<td valign="middle" align="center">68.21</td>
<td valign="middle" align="center">
<bold>75.95</bold>
</td>
</tr>
<tr>
<td valign="middle" align="center">
<bold>Island slope deep water terrace</bold>
</td>
<td valign="middle" align="center">58.74</td>
<td valign="middle" align="center">
<bold>67.28</bold>
</td>
</tr>
<tr>
<td valign="middle" align="center">
<bold>Trench seamount group</bold>
</td>
<td valign="middle" align="center">56.28</td>
<td valign="middle" align="center">
<bold>75.21</bold>
</td>
</tr>
<tr>
<td valign="middle" align="center">
<bold>Trench edge slope</bold>
</td>
<td valign="middle" align="center">55.69</td>
<td valign="middle" align="center">
<bold>73.53</bold>
</td>
</tr>
<tr>
<td valign="middle" align="center">
<bold>Trench bottom basin</bold>
</td>
<td valign="middle" align="center">
<bold>69.12</bold>
</td>
<td valign="middle" align="center">66.13</td>
</tr>
<tr>
<td valign="middle" align="center">
<bold>Slope fault basin</bold>
</td>
<td valign="middle" align="center">55.69</td>
<td valign="middle" align="center">
<bold>63.07</bold>
</td>
</tr>
<tr>
<td valign="middle" align="center">
<bold>Island platform</bold>
</td>
<td valign="middle" align="center">56.28</td>
<td valign="middle" align="center">
<bold>64.19</bold>
</td>
</tr>
</tbody>
</table>
<table-wrap-foot>
<fn>
<p>The bold values highlight the scores that fit the best performances.</p>
</fn>
</table-wrap-foot>
</table-wrap>
<p>The overall performance evaluation of our developed scheme has been illustrated in <xref ref-type="table" rid="T6">
<bold>Table&#xa0;6</bold>
</xref>, where the semantic segmentation accuracy calculation, from the baseline Densenet121 with the minimalistic transition-up (TU) blocks (DT), embedded with the channel attention module (DTC) and spatial pyramid pooling module (DTCS), to the coupled feature fusion with the morphological cues (DTCSF), was quantitatively measured step by step in terms of PA, MPA, and MIoU metrics.</p>
<table-wrap id="T6" position="float">
<label>Table&#xa0;6</label>
<caption>
<p>Evaluation of the overall network performance.</p>
</caption>
<table frame="hsides">
<thead>
<tr>
<th valign="middle" align="center">Method</th>
<th valign="middle" align="center">PA (%)</th>
<th valign="middle" align="center">MPA (%)</th>
<th valign="middle" align="center">MIoU (%)</th>
</tr>
</thead>
<tbody>
<tr>
<td valign="middle" align="center">
<bold>DenseNet+TU (DT)</bold>
</td>
<td valign="middle" align="center">87.63</td>
<td valign="middle" align="center">79.65</td>
<td valign="middle" align="center">70.92</td>
</tr>
<tr>
<td valign="middle" align="center">
<bold>+ Channel Attention (DTC)</bold>
</td>
<td valign="middle" align="center">88.55</td>
<td valign="middle" align="center">80.42</td>
<td valign="middle" align="center">71.84</td>
</tr>
<tr>
<td valign="middle" align="center">
<bold>+ Spatial Pooling (DTCS)</bold>
</td>
<td valign="middle" align="center">89.21</td>
<td valign="middle" align="center">81.25</td>
<td valign="middle" align="center">72.71</td>
</tr>
<tr>
<td valign="middle" align="center">
<bold>+ Feature Fusion (DTCSF)</bold>
</td>
<td valign="middle" align="center">
<bold>89.87</bold>
</td>
<td valign="middle" align="center">
<bold>82.01</bold>
</td>
<td valign="middle" align="center">
<bold>73.52</bold>
</td>
</tr>
</tbody>
</table>
<table-wrap-foot>
<fn>
<p>The bold values highlight the scores that fit the best performances.</p>
</fn>
</table-wrap-foot>
</table-wrap>
</sec>
<sec id="s6_5">
<label>6.5</label>
<title>Comparison with state-of-the-art methods</title>
<p>We further performed the comparative evaluation with some state-of-the-art models for semantic segmentation of seafloor surface stretching, including FCN-8s, SegNet (<xref ref-type="bibr" rid="B2">Badrinarayanan et&#xa0;al., 2017</xref>), RefineNet (<xref ref-type="bibr" rid="B39">Lin et&#xa0;al., 2017</xref>), PSPNet (<xref ref-type="bibr" rid="B86">Zhao et&#xa0;al., 2017</xref>), DeepLab v3+ (<xref ref-type="bibr" rid="B8">Chen et&#xa0;al., 2018</xref>), and our developed scheme, as shown in <xref ref-type="table" rid="T7">
<bold>Table&#xa0;7</bold>
</xref>. The classic FCN-8s network integrates the multi-layer feature maps during down-sampling in FCN. The SegNet network calls the pooling index at the corresponding encoder in the decoder to upsample the feature map through the unpooling operation. RefineNet explicitly exploits all the information available along the down-sampling process to enable high-resolution prediction through long-range residual connections. PSPNet captures global context through different-region-based context aggregation by the pyramid pooling module to improve network performance. DeepLab v3+ makes use of an encoder-decoder to perform multi-scale information fusion while retaining the dilated convolutions and Atrous Spatial Pyramid Pooling (ASPP) layer of the original DeepLab series. It should be noted that the above segmentation results for seafloor stretching categories were initially generated from the average product of 5-fold cross-validation with our developed model by dividing into&#xa0;mutually exclusive subsets with nearly equal numbers of randomly selected samples. From the experimental results, it was demonstrated that our proposed scheme had achieved a significant improvement in semantic segmentation performance, with PA, MPA, and MIoU metrics reaching up to 89.87%, 82.01%, and 73.52%, respectively. The model also exhibited a high level of stability in terms of PA, MPA, and MIoU metrics with a series of cross-validation rounds.</p>
<table-wrap id="T7" position="float">
<label>Table&#xa0;7</label>
<caption>
<p>Comparison of classification performance.</p>
</caption>
<table frame="hsides">
<thead>
<tr>
<th valign="middle" align="center">Method</th>
<th valign="middle" align="center">PA (%)</th>
<th valign="middle" align="center">MPA (%)</th>
<th valign="middle" align="center">MIoU (%)</th>
<th valign="middle" align="center">Time(ms)</th>
</tr>
</thead>
<tbody>
<tr>
<td valign="middle" align="center">
<bold>Random forest</bold>
</td>
<td valign="middle" align="center">80.05</td>
<td valign="middle" align="center">75.43</td>
<td valign="middle" align="center">58.15</td>
<td valign="middle" align="center">33.7</td>
</tr>
<tr>
<td valign="middle" align="center">
<bold>FCN-8s</bold>
</td>
<td valign="middle" align="center">84.62</td>
<td valign="middle" align="center">77.05</td>
<td valign="middle" align="center">62.76</td>
<td valign="middle" align="center">
<bold>33.2</bold>
</td>
</tr>
<tr>
<td valign="middle" align="center">
<bold>SegNet</bold>
</td>
<td valign="middle" align="center">86.73</td>
<td valign="middle" align="center">78.71</td>
<td valign="middle" align="center">68.63</td>
<td valign="middle" align="center">67.1</td>
</tr>
<tr>
<td valign="middle" align="center">
<bold>RefineNet</bold>
</td>
<td valign="middle" align="center">88.90</td>
<td valign="middle" align="center">80.46</td>
<td valign="middle" align="center">69.74</td>
<td valign="middle" align="center">39.5</td>
</tr>
<tr>
<td valign="middle" align="center">
<bold>PSPNet</bold>
</td>
<td valign="middle" align="center">88.62</td>
<td valign="middle" align="center">80.13</td>
<td valign="middle" align="center">70.09</td>
<td valign="middle" align="center">34.1</td>
</tr>
<tr>
<td valign="middle" align="center">
<bold>DeepLab v3+</bold>
</td>
<td valign="middle" align="center">89.08</td>
<td valign="middle" align="center">81.27</td>
<td valign="middle" align="center">72.47</td>
<td valign="middle" align="center">45.1</td>
</tr>
<tr>
<td valign="middle" align="center">
<bold>Ours</bold>
</td>
<td valign="middle" align="center">
<bold>89.87</bold>
</td>
<td valign="middle" align="center">
<bold>82.01</bold>
</td>
<td valign="middle" align="center">
<bold>73.52</bold>
</td>
<td valign="middle" align="center">35.7</td>
</tr>
</tbody>
</table>
<table-wrap-foot>
<fn>
<p>The bold values highlight the scores that fit the best performances.</p>
</fn>
</table-wrap-foot>
</table-wrap>
<p>The semantic segmentation of multibeam bathymetric seafloor mapping has been further visualized, as is shown in <xref ref-type="fig" rid="f5">
<bold>Figure&#xa0;5</bold>
</xref>, where the example MBES image, the segmentation results of both FCN-8s and our proposed scheme, and the corresponding labels are listed from left to right, respectively, with the island slope ridge in red, the island slope in green, the island slope deepwater terrace in cyan, the trench seamount group in yellow, the trench edge slope in blue,&#xa0;the trench bottom basin in purple, the island platform in orange, and the slope fault basin in black. From our experimental results, it was demonstrated that our developed scheme visually outperformed the classic FCN-8s, and simultaneously enhanced the details in-between edges, with the ability to preserve the salient features and eliminate redundancy on a global scale, showing its superiority in the descriptiveness and distinguishability of the seafloor surface categories. Some semantic segmentation results of the example multibeam bathymetric seafloor mapping along waypoints of the expedition track around the Mariana Trench are shown in <xref ref-type="fig" rid="f6">
<bold>Figure&#xa0;6</bold>
</xref>, where the location of the waypoint, the original example MBES images, the segmentation results, and the ground truth are listed from left to right.</p>
<fig id="f5" position="float">
<label>Figure&#xa0;5</label>
<caption>
<p>Visualization of an example MBES image segmentation result. <bold>(A)</bold> Example MBES image, <bold>(B)</bold> FCN-8s, <bold>(C)</bold> the proposed scheme, <bold>(D)</bold> ground truth.</p>
</caption>
<graphic mimetype="image" mime-subtype="tiff" xlink:href="fmars-10-1205142-g005.tif"/>
</fig>
<fig id="f6" position="float">
<label>Figure&#xa0;6</label>
<caption>
<p>Multibeam bathymetric seafloor mapping along waypoints around the Mariana Trench. <bold>(A)</bold> Location of waypoints, <bold>(B)</bold> Example MBES image, <bold>(C)</bold> segmentation result, <bold>(D)</bold> ground truth.</p>
</caption>
<graphic mimetype="image" mime-subtype="tiff" xlink:href="fmars-10-1205142-g006.tif"/>
</fig>
</sec>
<sec id="s6_6">
<label>6.6</label>
<title>Biogeographic patterns</title>
<p>Furthermore, we made an attempt to focus on a more delicate observation and a preliminary study of how the seafloor surface stretching functions as a submarine benthic habitat and what type of biogeographic pattern distribution of the benthic organisms are present in the extremely deep sea, with the help of both the acoustic sensor on board the research vessel Okeanos Explorer and the optical sensor mounted in the ROV Deep Discoverer. A total of 10,000 underwater images of the dominant resident biological species and their corresponding habitats, at dive depths within a range of 250-5000m, included Rimicaris, Austinograea, Symphurus thermophiles, Bathymodiolus, Phenacolepadidae, Shinkailepas, Thoridae Lebbeus, Lamellibrachia, etc., and were considered as our alternative underwater vision dataset for this preliminary study. The primary benthic species retrieved from the video of each dive by the ROV Deep Discoverer during the EX1605L3 expedition route are recorded in <xref ref-type="supplementary-material" rid="SM1">
<bold>Supplementary Table M1</bold>
</xref>.</p>
<p>We established a global geographical link between the ROV dive path and the MBES bathymetric mapping route. <xref ref-type="fig" rid="f7">
<bold>Figure&#xa0;7</bold>
</xref> shows the connection between the latitude and longitude of the example ROV&#xa0;dive paths and the location of the MBES imaging survey, including the original example MBES bathymetric mapping, the corresponding seafloor surface categories, the ROV dive paths on Eifuku Seamount and Daikoku Seamount, and the possible typical&#xa0;benthic habitats retrieved from visual cues along the paths,&#xa0;thus linking the seafloor surface topography with the primary benthic biogeographic patterns. We tried to statistically match the corresponding primary benthic habitats and species with the seafloor surface stretching by roughly retrieving the microgeographic cues from each ROV dive, and subdividing the benthic habitats with the visual cues from optical sensing. <xref ref-type="fig" rid="f8">
<bold>Figure&#xa0;8</bold>
</xref> lists some examples of dominant benthic habitats and species that visually reflect the possible biogeographic patterns that respectively appeared and were distributed at distinct locations of the seamount above <xref ref-type="fig" rid="f8">
<bold>Figures 8A&#x2013;F</bold>
</xref> and the seamount below <xref ref-type="fig" rid="f8">
<bold>Figures 8G&#x2013;L</bold>
</xref>, which also makes it possible to provide an initial insight into the diversity and distribution of the benthic community.</p>
<fig id="f7" position="float">
<label>Figure&#xa0;7</label>
<caption>
<p>The connection between the latitude and longitude of Dive8 and Dive9 path and the geographical location of MBES imaging survey. <bold>(A)</bold> MBES bathymetric mapping, <bold>(B)</bold> geographical linkage between ROV diving path and the route of MBES bathymetric mapping, <bold>(C)</bold> segmentation result of MBES imaging, <bold>(D)</bold>&#xa0;geographical linkage between ROV diving path and segmentation result of MBES imaging, <bold>(E)</bold> Dive8 diving path maps, <bold>(F)</bold> Dive9 diving path maps.</p>
</caption>
<graphic mimetype="image" mime-subtype="tiff" xlink:href="fmars-10-1205142-g007.tif"/>
</fig>
<fig id="f8" position="float">
<label>Figure&#xa0;8</label>
<caption>
<p>Preliminary study of benthic habitats with both acoustic and optical sensors along the ROV dive path above and below. <bold>(A)</bold> old hydrothermal chimney, <bold>(B)</bold> coronaster seastar, <bold>(C)</bold> crater wall, <bold>(D)</bold> grouper, <bold>(E)</bold> volcaniclastics, <bold>(F)</bold> octocoral fauna, <bold>(G)</bold> sulfur, <bold>(H)</bold> Symphurus thermophilus, <bold>(I)</bold> volcanic rock, <bold>(J)</bold> Gandalfus yunohana, <bold>(K)</bold> crater, <bold>(L)</bold> volcanic smoke.</p>
</caption>
<graphic mimetype="image" mime-subtype="tiff" xlink:href="fmars-10-1205142-g008.tif"/>
</fig>
<p>It could be seen from the visual clues from the ROV dive on Eifuku Seamount that the benthic species, especially fish and octocoral fauna, were unexpectedly diverse, and the typical geomorphology discovered included the crater wall and the hydrothermal chimney structure near the summit, while the visual clues from the ROV dive on Daikoku Seamount demonstrated the high activity of the hydrothermal vents, the possible evidence of the recent eruption, the volcaniclastics, the sulfur pond and the thick volcanic smoke, the plume, and the flatfish communities, e.g., Symphurus thermophilus and Gandalfus yunohana. The extent of the seafloor surface stretching and the estimation of the primary benthic biogeographic patterns reflect the coupling variation of multivariate environmental variables in the deep sea. The associative study derived from the sparse observation statistics through both acoustic and optical sensors not only produces the possibility of capturing the potential relationships between the full coverage of seafloor mapping and the benthic habitats, even the benthic species assemblage maps, but also provides the opportunity to examine the predicted biogeographic patterns with better-described variations and uncertainties towards the distinct geographical characteristics of seafloor surfaces.</p>
</sec>
</sec>
<sec id="s7" sec-type="conclusion">
<label>7</label>
<title>Conclusion</title>
<p>The extent to which the seafloor surface stretches and the benthic community survives and thrives, strongly reflecting the coupling variation of multivariate environmental factors, is still poorly understood in the extreme deep sea. With access to advanced sensing techniques, especially those related to the deployment of ROVs and AUVs, more attention has been paid to accumulating benthic biogeographic knowledge from these less-mapped or even unknown deep-sea regions. In this paper, an effort has been made to explore how to build online computing strategies for underwater vehicles that enable them to reasonably classify seafloor surface categories and identify the potential biogeographic patterns when underwater vehicles proceed along waypoints, in favor of a deep learning-based framework, DenseNet. We directly utilized multibeam bathymetric mapping from high-resolution MBES scans and manually labeled the seafloor surfaces as the standard references, i.e., island slope ridge, island slope, island slope deepwater terrace, trench seamount group, trench edge slope, trench bottom basin, island platform, and slope fault basin. We initially applied the individual and joint morphological cues in combination, in terms of slope, surface roughness, and curvature, in relation to the formation and evolution processes, to help semantic mapping and localization, which agglomerated assigned elevation points with highly similar topological parameters for the identical seafloor categories and deviated from the significantly inconsistent outlier elevation points. We then established an upgraded Densenet121 backbone from the minimalistic transition-up blocks in the upsampling path, systematically strengthening the dominant feature propagation and encouraging multi-scale feature reuse by employing both the channel attention module and the spatial pyramid pooling, and derived the types of seafloor categories with specific morphological parameters. We integrated the channel attention module by sequentially inferring channel-wise attention maps from both the Average pooling and the Max pooling in parallel. Meanwhile, we applied spatial pyramid pooling to incorporate the momentous global contextual prior by aggregation along with sub-regions from multiple receptive fields in a pyramidal manner. From the experiment results, it was demonstrated that the seafloor stretching classification accuracy of our proposed scheme could reach up to 89.87%, 82.01%, and 73.52% on average in terms of PA, MPA, and MIoU metrics, achieving comparable performances with state-of-the-art deep learning frameworks such as FCN-8s, SegNet, RefineNet, PSPNet, and DeepLab v3+, which permits us to delicately and adaptively distinguish the specific seafloor categories and connect the probable submarine benthic habitats. We also made a preliminary study on the potential biogeographic distribution statistics, showing the premise of deploying underwater vehicles through unbiased means or pre-programmed path planning to quantify and estimate the specific types of seafloor categories and the exhibiting fine-scale biogeographic patterns. The proposed scheme commits to developing smart capabilities embedded in underwater vehicles that could integrate the geometrical, topological, morphological, and biogeographic evidence yielded from MBES and optical sensing and hopefully perform habitat delineation, sampling programs, and other deep-sea tasks with more flexibility and adaptability. In the future, we will tentatively extend our proposed seafloor classification scheme with more advanced deep learning frameworks, such as Transformers, and further implant them in underwater vehicles through lightweight solutions.</p>
</sec>
<sec id="s8" sec-type="data-availability">
<title>Data availability statement</title>
<p>The original contributions presented in the study are included in the article/<xref ref-type="supplementary-material" rid="s13">
<bold>Supplementary Material</bold>
</xref>. Further inquiries can be directed to the corresponding author. Ocean Exploration and Research (OER) digital atlas is available in the official website of the National Oceanic and Atmospheric Administration (NOAA) (<ext-link ext-link-type="uri" xlink:href="http://www.noaa.gov/">http://www.noaa.gov/</ext-link>).</p>
</sec>
<sec id="s9" sec-type="author-contributions">
<title>Author contributions</title>
<p>RN: Conceptualization, Methodology, Supervision, Programming, Formal analysis, Writing, Funding acquisition. SL: Methodology, Software, Formal analysis, Writing. ZL: Data curation, Methodology, Formal analysis, Writing. XYL: Data curation, Methodology, Programming. SR: Data curation, Methodology, Software, Formal analysis, Programming. YQ: Data curation, Writing. QL: Data curation, Writing. GH: Data curation, Writing. KS: Data curation, Writing. GZ: Programming, Software. LZ: Data curation, Software. LL: Software. BH: Supervision, Funding acquisition. TY: Supervision, Funding acquisition. XSL: Supervision. All authors contributed to the article and approved the submitted version.</p>
</sec>
</body>
<back>
<sec id="s10" sec-type="funding-information">
<title>Funding</title>
<p>The author(s) declare financial support was received for the research, authorship, and/or publication of this article. This work was supported by the National Key R&amp;D Program (grant number 2019YFC1408304); the National High-Tech R&amp;D 863 Program (grant number 2014AA093410); the National Key R&amp;D Program (grant number 2016YFC0301400); the Natural Science Foundation of P. R. China (grant number 42376191); the Natural Science Foundation of P. R. China (grant number 31202036); the Natural Science Foundation of P. R. China (grant number 41376140).</p>
</sec>
<ack>
<title>Acknowledgments</title>
<p>We would like to acknowledge team member Jie Wang for his guidance in data analysis.</p>
</ack>
<sec id="s11" sec-type="COI-statement">
<title>Conflict of interest</title>
<p>The authors declare that the research was conducted in the absence of any commercial or financial relationships that could be construed as a potential conflict of interest.</p>
</sec>
<sec id="s12" sec-type="disclaimer">
<title>Publisher&#x2019;s note</title>
<p>All claims expressed in this article are solely those of the authors and do not necessarily represent those of their affiliated organizations, or those of the publisher, the editors and the reviewers. Any product that may be evaluated in this article, or claim that may be made by its manufacturer, is not guaranteed or endorsed by the publisher.</p>
</sec>
<sec id="s13" sec-type="supplementary-material">
<title>Supplementary material</title>
<p>The Supplementary Material for this article can be found online at: <ext-link ext-link-type="uri" xlink:href="https://www.frontiersin.org/articles/10.3389/fmars.2023.1205142/full#supplementary-material">https://www.frontiersin.org/articles/10.3389/fmars.2023.1205142/full#supplementary-material</ext-link>
</p>
<supplementary-material xlink:href="DataSheet_1.docx" id="SM1" mimetype="application/vnd.openxmlformats-officedocument.wordprocessingml.document"/>
</sec>
<ref-list>
<title>References</title>
<ref id="B1">
<citation citation-type="journal">
<person-group person-group-type="author">
<name>
<surname>Bach</surname> <given-names>W.</given-names>
</name>
<name>
<surname>Edwards</surname> <given-names>K. J.</given-names>
</name>
</person-group> (<year>2003</year>). <article-title>Iron and sulfide oxidation within the basaltic ocean crust: implications for chemolithoautotrophic microbial biomass production</article-title>. <source>Geochimica Cosmochimica Acta</source> <volume>67</volume> (<issue>20</issue>), <fpage>3871</fpage>&#x2013;<lpage>3887</lpage>. doi: <pub-id pub-id-type="doi">10.1016/S0016-7037(03)00304-1</pub-id>
</citation>
</ref>
<ref id="B2">
<citation citation-type="journal">
<person-group person-group-type="author">
<name>
<surname>Badrinarayanan</surname> <given-names>V.</given-names>
</name>
<name>
<surname>Kendall</surname> <given-names>A.</given-names>
</name>
<name>
<surname>Cipolla</surname> <given-names>R.</given-names>
</name>
</person-group> (<year>2017</year>). <article-title>SegNet: A deep convolutional encoder-decoder architecture for image segmentation</article-title>. <source>IEEE Trans. Pattern Anal. Mach. Intell.</source> <volume>39</volume> (<issue>12</issue>), <fpage>2481</fpage>&#x2013;<lpage>2495</lpage>. doi:&#xa0;<pub-id pub-id-type="doi">10.1109/TPAMI.2016.2644615</pub-id>
</citation>
</ref>
<ref id="B3">
<citation citation-type="journal">
<person-group person-group-type="author">
<name>
<surname>Bergquist</surname> <given-names>D. C.</given-names>
</name>
<name>
<surname>Eckner</surname> <given-names>J. T.</given-names>
</name>
<name>
<surname>Urcuyo</surname> <given-names>I. A.</given-names>
</name>
<name>
<surname>Cordes</surname> <given-names>E. E.</given-names>
</name>
<name>
<surname>Hourdez</surname> <given-names>S.</given-names>
</name>
<name>
<surname>Macko</surname> <given-names>S. A.</given-names>
</name>
<etal/>
</person-group>. (<year>2007</year>). <article-title>Using stable isotopes and quantitative community characteristics to determine a local hydrothermal vent food web</article-title>. <source>Mar. Ecol. Prog. Ser.</source> <volume>330</volume>, <fpage>49</fpage>&#x2013;<lpage>65</lpage>. doi: <pub-id pub-id-type="doi">10.3354/meps330049</pub-id>
</citation>
</ref>
<ref id="B4">
<citation citation-type="confproc">
<person-group person-group-type="author">
<name>
<surname>Bewley</surname> <given-names>M.</given-names>
</name>
<name>
<surname>Douillard</surname> <given-names>B.</given-names>
</name>
<name>
<surname>Nourani-Vatani</surname> <given-names>N.</given-names>
</name>
<name>
<surname>Friedman</surname> <given-names>A.</given-names>
</name>
<name>
<surname>Pizarro</surname> <given-names>O.</given-names>
</name>
<name>
<surname>Williams</surname> <given-names>S.</given-names>
</name>
</person-group> (<year>2012</year>). &#x201c;<article-title>Automated species detection: An experimental approach to kelp detection from sea-floor AUV images</article-title>,&#x201d; in <conf-name>Proc Australas Conf Rob Autom</conf-name>. <conf-loc>Wellington, New Zealand</conf-loc>. Vol. <volume>2012</volume>.</citation>
</ref>
<ref id="B5">
<citation citation-type="journal">
<person-group person-group-type="author">
<name>
<surname>Burrough</surname> <given-names>P. A.</given-names>
</name>
<name>
<surname>McDonnell</surname> <given-names>R.</given-names>
</name>
</person-group> (<year>1998</year>). <source>Principles of geographical information systems</source>. (<publisher-loc>New York</publisher-loc>: <publisher-name>Oxford University Press</publisher-name>).</citation>
</ref>
<ref id="B6">
<citation citation-type="book">
<person-group person-group-type="author">
<name>
<surname>Cantwell</surname> <given-names>K.</given-names>
</name>
</person-group> (<year>2016</year>). &#x201c;<article-title>Final Project Instructions</article-title>,&#x201d; in <source>EX-16-05 Leg 3: CAPSTONE CNMI &amp; Mariana Trench MNM (rov &amp; mapping), June 17-July 10, 2016.m depth of each dive</source>.</citation>
</ref>
<ref id="B7">
<citation citation-type="journal">
<person-group person-group-type="author">
<name>
<surname>Chakraborty</surname> <given-names>B.</given-names>
</name>
<name>
<surname>Haris</surname> <given-names>K.</given-names>
</name>
<name>
<surname>Latha</surname> <given-names>G.</given-names>
</name>
<name>
<surname>Maslov</surname> <given-names>N.</given-names>
</name>
<name>
<surname>Menezes</surname> <given-names>A.</given-names>
</name>
</person-group> (<year>2013</year>). <article-title>Multifractal approach for seafloor characterization</article-title>. <source>IEEE Geosci. Remote Sens. Lett.</source> <volume>11</volume> (<issue>1</issue>), <fpage>54</fpage>&#x2013;<lpage>58</lpage>. doi:&#xa0;<pub-id pub-id-type="doi">10.1109/LGRS.2013.2245856</pub-id>
</citation>
</ref>
<ref id="B8">
<citation citation-type="journal">
<person-group person-group-type="author">
<name>
<surname>Chen</surname> <given-names>L. C.</given-names>
</name>
<name>
<surname>Papandreou</surname> <given-names>G.</given-names>
</name>
<name>
<surname>Kokkinos</surname> <given-names>I.</given-names>
</name>
<name>
<surname>Murphy</surname> <given-names>K.</given-names>
</name>
<name>
<surname>Yuille</surname> <given-names>A. L.</given-names>
</name>
</person-group> (<year>2018</year>). <article-title>DeepLab: semantic image segmentation with deep convolutional nets, atrous convolution, and fully connected CRFs</article-title>. <source>IEEE Trans. Pattern Anal. Mach. Intell.</source> <volume>40</volume> (<issue>4</issue>), <fpage>834</fpage>&#x2013;<lpage>848</lpage>. doi:&#xa0;<pub-id pub-id-type="doi">10.1109/TPAMI.2017.2699184</pub-id>
</citation>
</ref>
<ref id="B9">
<citation citation-type="journal">
<person-group person-group-type="author">
<name>
<surname>Conti</surname> <given-names>L. A.</given-names>
</name>
<name>
<surname>Lim</surname> <given-names>A.</given-names>
</name>
<name>
<surname>Wheeler</surname> <given-names>A. J.</given-names>
</name>
</person-group> (<year>2019</year>). <article-title>High resolution mapping of a cold water coral mound</article-title>. <source>Sci. Rep.</source> <volume>9</volume> (<issue>1</issue>), <fpage>1016</fpage>. doi: <pub-id pub-id-type="doi">10.1038/s41598-018-37725-x</pub-id>
</citation>
</ref>
<ref id="B10">
<citation citation-type="book">
<person-group person-group-type="author">
<name>
<surname>Copley</surname> <given-names>J.</given-names>
</name>
</person-group> (<year>2014</year>). &#x201c;<article-title>Just how little do we know about the ocean floor</article-title>?,&#x201d; in <source>The conversation</source>. Available at: <uri xlink:href="https://theconversation.com/just-how-little-do-weknow-about-the-ocean-floor-32751">https://theconversation.com/just-how-little-do-weknow-about-the-ocean-floor-32751</uri>.</citation>
</ref>
<ref id="B11">
<citation citation-type="journal">
<person-group person-group-type="author">
<name>
<surname>Corr&#xea;a</surname> <given-names>P. V. F.</given-names>
</name>
<name>
<surname>Jovane</surname> <given-names>L.</given-names>
</name>
<name>
<surname>Murton</surname> <given-names>B. J.</given-names>
</name>
<name>
<surname>Sumida</surname> <given-names>P. Y. G.</given-names>
</name>
</person-group> (<year>2022</year>). <article-title>Benthic megafauna habitats, community structure and environmental drivers at Rio Grande Rise (SW Atlantic)</article-title>. <source>Deep Sea Res. Part I: Oceanographic Res. Papers</source> <volume>186</volume>, <fpage>103811</fpage>. doi:&#xa0;<pub-id pub-id-type="doi">10.1016/j.dsr.2022.103811</pub-id>
</citation>
</ref>
<ref id="B12">
<citation citation-type="journal">
<person-group person-group-type="author">
<name>
<surname>Costa</surname> <given-names>B. M.</given-names>
</name>
<name>
<surname>Battista</surname> <given-names>T. A.</given-names>
</name>
<name>
<surname>Pittman</surname> <given-names>S. J.</given-names>
</name>
</person-group> (<year>2009</year>). <article-title>Comparative evaluation of airborne LiDAR and ship-based multibeam SoNAR bathymetry and intensity for mapping coral reef ecosystems</article-title>. <source>Remote Sens. Environ.</source> <volume>113</volume> (<issue>5</issue>), <fpage>1082</fpage>&#x2013;<lpage>1100</lpage>. doi:&#xa0;<pub-id pub-id-type="doi">10.1016/j.rse.2009.01.015</pub-id>
</citation>
</ref>
<ref id="B13">
<citation citation-type="journal">
<person-group person-group-type="author">
<name>
<surname>De Boer</surname> <given-names>D. H.</given-names>
</name>
</person-group> (<year>1992</year>). <article-title>Hierarchies and spatial scale in process geomorphology: a review</article-title>. <source>Geomorphology</source> <volume>4</volume> (<issue>5</issue>), <fpage>303</fpage>&#x2013;<lpage>318</lpage>. doi: <pub-id pub-id-type="doi">10.1016/0169-555X(92)90026-K</pub-id>
</citation>
</ref>
<ref id="B14">
<citation citation-type="journal">
<person-group person-group-type="author">
<name>
<surname>De la Torriente</surname> <given-names>A.</given-names>
</name>
<name>
<surname>Serrano</surname> <given-names>A.</given-names>
</name>
<name>
<surname>Fern&#xe1;ndez-Salas</surname> <given-names>L. M.</given-names>
</name>
<name>
<surname>Garc&#xed;a</surname> <given-names>M.</given-names>
</name>
<name>
<surname>Aguilar</surname> <given-names>R.</given-names>
</name>
</person-group> (<year>2018</year>). <article-title>Identifying epibenthic habitats on the Seco de los Olivos Seamount: species assemblages and environmental characteristics</article-title>. <source>Deep Sea Res. Part I: Oceanographic Res. Papers</source> <volume>135</volume>, <fpage>9</fpage>&#x2013;<lpage>22</lpage>. doi: <pub-id pub-id-type="doi">10.1016/j.dsr.2018.03.015</pub-id>
</citation>
</ref>
<ref id="B15">
<citation citation-type="confproc">
<person-group person-group-type="author">
<name>
<surname>Desbruy&#xe8;res</surname> <given-names>D.</given-names>
</name>
<name>
<surname>Almeida</surname> <given-names>A.</given-names>
</name>
<name>
<surname>Biscoito</surname> <given-names>M.</given-names>
</name>
<name>
<surname>Comtet</surname> <given-names>T.</given-names>
</name>
<name>
<surname>Khripounoff</surname> <given-names>A.</given-names>
</name>
<name>
<surname>Le Bris</surname> <given-names>N.</given-names>
</name>
<etal/>
</person-group> (<year>2000</year>). <article-title>A review of the distribution of hydrothermal vent communities along the northern Mid-Atlantic Ridge: dispersal vs. environmental controls</article-title>. In <conf-name>Island, Ocean and Deep-Sea Biology: Proceedings of the 34th European Marine Biology Symposium, Ponta Delgada (Azores)</conf-name>. <conf-loc>Portugal</conf-loc>, <conf-date>September 1999</conf-date>.</citation>
</ref>
<ref id="B16">
<citation citation-type="journal">
<person-group person-group-type="author">
<name>
<surname>Dunlop</surname> <given-names>K. M.</given-names>
</name>
<name>
<surname>Jarvis</surname> <given-names>T.</given-names>
</name>
<name>
<surname>Benoit-Bird</surname> <given-names>K. J.</given-names>
</name>
<name>
<surname>Waluk</surname> <given-names>C. M.</given-names>
</name>
<name>
<surname>Caress</surname> <given-names>D. W.</given-names>
</name>
<name>
<surname>Thomas</surname> <given-names>H.</given-names>
</name>
<etal/>
</person-group>. (<year>2018</year>). <article-title>Detection and characterisation of deep-sea benthopelagic animals from an autonomous underwater vehicle with a multibeam echosounder: A proof of concept and description of data-processing methods</article-title>. <source>Deep Sea Res. Part I: Oceanographic Res. Papers</source> <volume>134</volume>, <fpage>64</fpage>&#x2013;<lpage>79</lpage>. doi:&#xa0;<pub-id pub-id-type="doi">10.1016/j.dsr.2018.01.006</pub-id>
</citation>
</ref>
<ref id="B17">
<citation citation-type="book">
<person-group person-group-type="author">
<name>
<surname>Dyer</surname> <given-names>A. S.</given-names>
</name>
<name>
<surname>Zaengle</surname> <given-names>D.</given-names>
</name>
<name>
<surname>Suhag</surname> <given-names>A.</given-names>
</name>
<name>
<surname>Mark-Moser</surname> <given-names>M. K.</given-names>
</name>
<name>
<surname>Duran</surname> <given-names>R.</given-names>
</name>
<name>
<surname>Rose</surname> <given-names>K.</given-names>
</name>
<etal/>
</person-group>. (<year>2020</year>). <source>Deep Learning to Locate Seafloor Landslides in High Resolution Bathymetry (No. RSS624)</source> (<publisher-loc>Pittsburgh, PA, Morgantown, WV, and Albany, OR (United States</publisher-loc>: <publisher-name>National Energy Technology Laboratory (NETL</publisher-name>).</citation>
</ref>
<ref id="B18">
<citation citation-type="journal">
<person-group person-group-type="author">
<name>
<surname>Fisher</surname> <given-names>C. R.</given-names>
</name>
<name>
<surname>Childress</surname> <given-names>J. J.</given-names>
</name>
<name>
<surname>Macko</surname> <given-names>S. A.</given-names>
</name>
<name>
<surname>Brooks</surname> <given-names>J. M.</given-names>
</name>
</person-group> (<year>1994</year>). <article-title>Nutritional interactions in Galapagos Rift hydrothermal vent communities: inferences from stable carbon and nitrogen isotope analyses</article-title>. <source>Mar. Ecol. Prog. Ser.</source> <volume>103</volume>, <fpage>45</fpage>&#x2013;<lpage>55</lpage>. doi: <pub-id pub-id-type="doi">10.3354/meps103045</pub-id>
</citation>
</ref>
<ref id="B19">
<citation citation-type="journal">
<person-group person-group-type="author">
<name>
<surname>Freitag</surname> <given-names>L.</given-names>
</name>
<name>
<surname>Grund</surname> <given-names>M.</given-names>
</name>
<name>
<surname>Von Alt</surname> <given-names>C.</given-names>
</name>
<name>
<surname>Stokey</surname> <given-names>R.</given-names>
</name>
<name>
<surname>Austin</surname> <given-names>T.</given-names>
</name>
</person-group> (<year>2005</year>). <article-title>A shallow water acoustic network for mine countermeasures operations with autonomous underwater vehicles</article-title>. <source>Underwater Defense Technol. (UDT)</source>, <fpage>1</fpage>&#x2013;<lpage>6</lpage>.</citation>
</ref>
<ref id="B20">
<citation citation-type="journal">
<person-group person-group-type="author">
<name>
<surname>Friedman</surname> <given-names>A.</given-names>
</name>
<name>
<surname>Pizarro</surname> <given-names>O.</given-names>
</name>
<name>
<surname>Williams</surname> <given-names>S. B.</given-names>
</name>
<name>
<surname>Johnson-Roberson</surname> <given-names>M.</given-names>
</name>
</person-group> (<year>2013</year>). <article-title>Correction: multi-scale measures of rugosity, slope and aspect from benthic stereo image reconstructions</article-title>. <source>PloS One</source> <volume>7</volume> (<issue>12</issue>), <elocation-id>e50440</elocation-id>. doi:&#xa0;<pub-id pub-id-type="doi">10.1371/annotation/55ee98d1-6731-4bee-81d6-03ce0259c191</pub-id>
</citation>
</ref>
<ref id="B21">
<citation citation-type="confproc">
<person-group person-group-type="author">
<name>
<surname>Furlong</surname> <given-names>M. E.</given-names>
</name>
<name>
<surname>Paxton</surname> <given-names>D.</given-names>
</name>
<name>
<surname>Stevenson</surname> <given-names>P.</given-names>
</name>
<name>
<surname>Pebody</surname> <given-names>M.</given-names>
</name>
<name>
<surname>McPhail</surname> <given-names>S. D.</given-names>
</name>
<name>
<surname>Perrett</surname> <given-names>J.</given-names>
</name>
</person-group> (<year>2012</year>). &#x201c;<article-title>Autosub long range: A long range deep diving AUV for ocean monitoring</article-title>,&#x201d; in <conf-name>2012 IEEE/OES Autonomous Underwater Vehicles (AUV)</conf-name>. <conf-loc>Southampton, UK</conf-loc>. <fpage>1</fpage>&#x2013;<lpage>7</lpage>.</citation>
</ref>
<ref id="B22">
<citation citation-type="confproc">
<person-group person-group-type="author">
<name>
<surname>Goodfellow</surname> <given-names>I.</given-names>
</name>
<name>
<surname>Pouget-Abadie</surname> <given-names>J.</given-names>
</name>
<name>
<surname>Mirza</surname> <given-names>M.</given-names>
</name>
<name>
<surname>Xu</surname> <given-names>B.</given-names>
</name>
<name>
<surname>Warde-Farley</surname> <given-names>D.</given-names>
</name>
<name>
<surname>Ozair</surname> <given-names>S.</given-names>
</name>
<etal/>
</person-group> (<year>2014</year>). <article-title>Generative adversarial nets. Advances in neural information processing systems</article-title>, <conf-loc>Quebec, Canada</conf-loc>. <conf-date>June 2014</conf-date>.</citation>
</ref>
<ref id="B23">
<citation citation-type="journal">
<person-group person-group-type="author">
<name>
<surname>Harris</surname> <given-names>P. T.</given-names>
</name>
<name>
<surname>Macmillan-Lawler</surname> <given-names>M.</given-names>
</name>
<name>
<surname>Rupp</surname> <given-names>J.</given-names>
</name>
<name>
<surname>Baker</surname> <given-names>E. K.</given-names>
</name>
</person-group> (<year>2014</year>). <article-title>Geomorphology of the oceans</article-title>. <source>Mar. Geology</source> <volume>352</volume>, <fpage>4</fpage>&#x2013;<lpage>24</lpage>. doi: <pub-id pub-id-type="doi">10.1016/j.margeo.2014.01.011</pub-id>
</citation>
</ref>
<ref id="B24">
<citation citation-type="journal">
<person-group person-group-type="author">
<name>
<surname>He</surname> <given-names>K.</given-names>
</name>
<name>
<surname>Zhang</surname> <given-names>X.</given-names>
</name>
<name>
<surname>Ren</surname> <given-names>S.</given-names>
</name>
<name>
<surname>Sun</surname> <given-names>J.</given-names>
</name>
</person-group> (<year>2015</year>). <article-title>Spatial pyramid pooling in deep convolutional networks for visual recognition</article-title>. <source>IEEE Trans. Pattern Anal. Mach. Intell.</source> <volume>37</volume> (<issue>9</issue>), <fpage>1904</fpage>&#x2013;<lpage>1916</lpage>. doi:&#xa0;<pub-id pub-id-type="doi">10.1109/TPAMI.2015.2389824</pub-id>
</citation>
</ref>
<ref id="B25">
<citation citation-type="confproc">
<person-group person-group-type="author">
<name>
<surname>He</surname> <given-names>K.</given-names>
</name>
<name>
<surname>Zhang</surname> <given-names>X.</given-names>
</name>
<name>
<surname>Ren</surname> <given-names>S.</given-names>
</name>
<name>
<surname>Sun</surname> <given-names>J.</given-names>
</name>
</person-group> (<year>2016</year>). &#x201c;<article-title>Deep residual learning for image recognition</article-title>,&#x201d; in <conf-name>2016 IEEE Conference on Computer Vision and Pattern Recognition (CVPR)</conf-name>. <conf-loc>Las Vegas, NV, USA</conf-loc>. <fpage>770</fpage>&#x2013;<lpage>778</lpage>. doi:&#xa0;<pub-id pub-id-type="doi">10.1109/CVPR.2016.90</pub-id>
</citation>
</ref>
<ref id="B26">
<citation citation-type="journal">
<person-group person-group-type="author">
<name>
<surname>Hinton</surname> <given-names>G. E.</given-names>
</name>
<name>
<surname>Osindero</surname> <given-names>S.</given-names>
</name>
<name>
<surname>Teh</surname> <given-names>Y. W.</given-names>
</name>
</person-group> (<year>2006</year>). <article-title>A fast learning algorithm for deep belief nets</article-title>. <source>Neural Comput.</source> <volume>18</volume> (<issue>7</issue>), <fpage>1527</fpage>&#x2013;<lpage>1554</lpage>. doi: <pub-id pub-id-type="doi">10.1162/neco.2006.18.7.1527</pub-id>
</citation>
</ref>
<ref id="B27">
<citation citation-type="confproc">
<person-group person-group-type="author">
<name>
<surname>Hu</surname> <given-names>J.</given-names>
</name>
<name>
<surname>Shen</surname> <given-names>L.</given-names>
</name>
<name>
<surname>Sun</surname> <given-names>G.</given-names>
</name>
</person-group> (<year>2018</year>). &#x201c;<article-title>Squeeze-and-excitation networks</article-title>,&#x201d; in <conf-name>Proceedings of the IEEE conference on computer vision and pattern recognition</conf-name>. <conf-loc>Salt Lake City, UT, USA</conf-loc>. <fpage>7132</fpage>&#x2013;<lpage>7141</lpage>. doi:&#xa0;<pub-id pub-id-type="doi">10.1109/TPAMI.2019.2913372</pub-id>
</citation>
</ref>
<ref id="B28">
<citation citation-type="confproc">
<person-group person-group-type="author">
<name>
<surname>Huang</surname> <given-names>G.</given-names>
</name>
<name>
<surname>Liu</surname> <given-names>Z.</given-names>
</name>
<name>
<surname>van der Maaten</surname> <given-names>L.</given-names>
</name>
<name>
<surname>Weinberger</surname> <given-names>K. Q.</given-names>
</name>
</person-group> (<year>2017</year>). &#x201c;<article-title>Densely connected convolutional networks</article-title>,&#x201d; in <conf-name>Proceedings of the IEEE conference on computer vision and pattern recognition</conf-name>. <conf-loc>Honolulu, HI, USA</conf-loc>. <fpage>4700</fpage>&#x2013;<lpage>4708</lpage>.</citation>
</ref>
<ref id="B29">
<citation citation-type="book">
<person-group person-group-type="author">
<name>
<surname>Huvenne</surname> <given-names>V. A.</given-names>
</name>
<name>
<surname>Robert</surname> <given-names>K.</given-names>
</name>
<name>
<surname>Marsh</surname> <given-names>L.</given-names>
</name>
<name>
<surname>Lo Iacono</surname> <given-names>C.</given-names>
</name>
<name>
<surname>Le Bas</surname> <given-names>T.</given-names>
</name>
<name>
<surname>Wynn</surname> <given-names>R. B.</given-names>
</name>
</person-group> (<year>2018</year>). &#x201c;<article-title>Rovs and auvs</article-title>,&#x201d; in <source>Submarine geomorphology</source>. ed. <person-group person-group-type="editor">
<name>
<surname>Micallef</surname> <given-names>A.</given-names>
</name>
</person-group> (<publisher-loc>Berlin, Germany</publisher-loc>: <publisher-name>Springer, Cham</publisher-name>) <fpage>93</fpage>&#x2013;<lpage>108</lpage>.</citation>
</ref>
<ref id="B30">
<citation citation-type="journal">
<person-group person-group-type="author">
<name>
<surname>Jannasch</surname> <given-names>H. W.</given-names>
</name>
<name>
<surname>Mottl</surname> <given-names>M. J.</given-names>
</name>
</person-group> (<year>1985</year>). <article-title>Geomicrobiology of deep-sea hydrothermal vents</article-title>. <source>Science</source> <volume>229</volume> (<issue>4715</issue>), <fpage>717</fpage>&#x2013;<lpage>725</lpage>. doi: <pub-id pub-id-type="doi">10.1126/science.229.4715.717</pub-id>
</citation>
</ref>
<ref id="B31">
<citation citation-type="confproc">
<person-group person-group-type="author">
<name>
<surname>J&#xe9;gou</surname> <given-names>S.</given-names>
</name>
<name>
<surname>Drozdzal</surname> <given-names>M.</given-names>
</name>
<name>
<surname>V&#xe1;zquez</surname> <given-names>D.</given-names>
</name>
<name>
<surname>Romero</surname> <given-names>A.</given-names>
</name>
<name>
<surname>Bengio</surname> <given-names>Y.</given-names>
</name>
</person-group> (<year>2017</year>). &#x201c;<article-title>The one hundred layers tiramisu: fully convolutional denseNets for semantic segmentation</article-title>,&#x201d; in <conf-name>2017 IEEE Conference on Computer Vision and Pattern Recognition Workshops (CVPRW)</conf-name>. <conf-loc>Honolulu, HI, USA</conf-loc>. <fpage>1175</fpage>&#x2013;<lpage>1183</lpage>.</citation>
</ref>
<ref id="B32">
<citation citation-type="journal">
<person-group person-group-type="author">
<name>
<surname>Johnson</surname> <given-names>K. S.</given-names>
</name>
<name>
<surname>Childress</surname> <given-names>J. J.</given-names>
</name>
<name>
<surname>Beehler</surname> <given-names>C. L.</given-names>
</name>
<name>
<surname>Sakamoto</surname> <given-names>C. M.</given-names>
</name>
</person-group> (<year>1994</year>). <article-title>Biogeochemistry of hydrothermal vent mussel communities: the deep-sea analogue to the intertidal zone</article-title>. <source>Deep Sea Research Part I: Oceanographic Research Papers</source> <volume>41</volume> (<issue>7</issue>), <fpage>993</fpage>&#x2013;<lpage>1011</lpage>. doi: <pub-id pub-id-type="doi">10.1016/0967-0637(94)90015-9</pub-id>
</citation>
</ref>
<ref id="B33">
<citation citation-type="confproc">
<person-group person-group-type="author">
<name>
<surname>Karpathy</surname> <given-names>A.</given-names>
</name>
<name>
<surname>Toderici</surname> <given-names>G.</given-names>
</name>
<name>
<surname>Shetty</surname> <given-names>S.</given-names>
</name>
<name>
<surname>Leung</surname> <given-names>T.</given-names>
</name>
<name>
<surname>Sukthankar</surname> <given-names>R.</given-names>
</name>
<name>
<surname>Fei-Fei</surname> <given-names>L.</given-names>
</name>
</person-group> (<year>2014</year>). &#x201c;<article-title>Large-scale video classification with convolutional neural networks</article-title>,&#x201d; in <conf-name>Proceedings of the IEEE conference on Computer Vision and Pattern Recognition</conf-name>. <conf-loc>Columbus, OH, USA</conf-loc>. <fpage>1725</fpage>&#x2013;<lpage>1732</lpage>.</citation>
</ref>
<ref id="B34">
<citation citation-type="confproc">
<person-group person-group-type="author">
<name>
<surname>Kreso</surname> <given-names>I.</given-names>
</name>
<name>
<surname>Segvic</surname> <given-names>S.</given-names>
</name>
<name>
<surname>Krapac</surname> <given-names>J.</given-names>
</name>
</person-group> (<year>2017</year>). &#x201c;<article-title>Ladder-style densenets for semantic segmentation of large natural images</article-title>,&#x201d; in <conf-name>Proceedings of the IEEE International Conference on Computer Vision Workshops</conf-name>. <conf-loc>Venice, Italy</conf-loc>. <fpage>238</fpage>&#x2013;<lpage>245</lpage>.</citation>
</ref>
<ref id="B35">
<citation citation-type="journal">
<person-group person-group-type="author">
<name>
<surname>Krizhevsky</surname> <given-names>A.</given-names>
</name>
<name>
<surname>Sutskever</surname> <given-names>I.</given-names>
</name>
<name>
<surname>Hinton</surname> <given-names>G. E.</given-names>
</name>
</person-group> (<year>2012</year>). <article-title>Imagenet classification with deep convolutional neural networks</article-title>. <source>Adv. Neural Inf. Process. Syst.</source> <volume>25</volume>, <fpage>1097</fpage>&#x2013;<lpage>1105</lpage>. doi:&#xa0;<pub-id pub-id-type="doi">10.1145/3065386</pub-id>
</citation>
</ref>
<ref id="B36">
<citation citation-type="journal">
<person-group person-group-type="author">
<name>
<surname>Kuhnz</surname> <given-names>L. A.</given-names>
</name>
<name>
<surname>Ruhl</surname> <given-names>H. A.</given-names>
</name>
<name>
<surname>Huffard</surname> <given-names>C. L.</given-names>
</name>
<name>
<surname>Smith</surname> <given-names>K. L.</given-names>
<suffix>Jr.</suffix>
</name>
</person-group> (<year>2014</year>). <article-title>Rapid changes and long-term cycles in the benthic megafaunal community observed over 24 years in the abyssal northeast Pacific</article-title>. <source>Prog. Oceanography</source> <volume>124</volume>, <fpage>1</fpage>&#x2013;<lpage>11</lpage>. doi:&#xa0;<pub-id pub-id-type="doi">10.1016/j.pocean.2014.04.007</pub-id>
</citation>
</ref>
<ref id="B37">
<citation citation-type="book">
<person-group person-group-type="author">
<name>
<surname>Lamarche</surname> <given-names>G.</given-names>
</name>
<name>
<surname>Orpin</surname> <given-names>A. R.</given-names>
</name>
<name>
<surname>Mitchell</surname> <given-names>J. S.</given-names>
</name>
<name>
<surname>Pallentin</surname> <given-names>A.</given-names>
</name>
</person-group> (<year>2016</year>). &#x201c;<article-title>Benthic habitat mapping</article-title>,&#x201d; in <source>Biological sampling in the deep sea</source>. ed. <person-group person-group-type="editor">
<name>
<surname>Clark</surname> <given-names>M. R.</given-names>
</name>
</person-group> (<publisher-loc>New York, US</publisher-loc>: <publisher-name>John Wiley &amp; Sons</publisher-name>) <fpage>80</fpage>&#x2013;<lpage>102</lpage>. doi:&#xa0;<pub-id pub-id-type="doi">10.1002/9781118332535</pub-id>
</citation>
</ref>
<ref id="B38">
<citation citation-type="confproc">
<person-group person-group-type="author">
<name>
<surname>Lee</surname> <given-names>Y.</given-names>
</name>
<name>
<surname>Hwang</surname> <given-names>J.</given-names>
</name>
<name>
<surname>Lee</surname> <given-names>S.</given-names>
</name>
<name>
<surname>Bae</surname> <given-names>Y.</given-names>
</name>
<name>
<surname>Park</surname> <given-names>J.</given-names>
</name>
</person-group> (<year>2019</year>). &#x201c;<article-title>An energy and GPU-computation efficient backbone network for real-time object detection</article-title>,&#x201d; in <conf-name>2019 IEEE/CVF Conference on Computer Vision and Pattern Recognition Workshops (CVPRW)</conf-name>. <conf-loc>Long Beach, CA, USA</conf-loc>. <fpage>752</fpage>&#x2013;<lpage>760</lpage>.</citation>
</ref>
<ref id="B39">
<citation citation-type="confproc">
<person-group person-group-type="author">
<name>
<surname>Lin</surname> <given-names>G.</given-names>
</name>
<name>
<surname>Milan</surname> <given-names>A.</given-names>
</name>
<name>
<surname>Shen</surname> <given-names>C.</given-names>
</name>
<name>
<surname>Reid</surname> <given-names>I.</given-names>
</name>
</person-group> (<year>2017</year>). &#x201c;<article-title>Refinenet: Multi-path refinement networks for high-resolution semantic segmentation</article-title>,&#x201d; in <conf-name>Proceedings of the IEEE conference on computer vision and pattern recognition</conf-name>. <conf-loc>Honolulu, HI, USA</conf-loc>. <fpage>1925</fpage>&#x2013;<lpage>1934</lpage>.</citation>
</ref>
<ref id="B40">
<citation citation-type="confproc">
<person-group person-group-type="author">
<name>
<surname>Long</surname> <given-names>J.</given-names>
</name>
<name>
<surname>Shelhamer</surname> <given-names>E.</given-names>
</name>
<name>
<surname>Darrell</surname> <given-names>T.</given-names>
</name>
</person-group> (<year>2015</year>). &#x201c;<article-title>Fully convolutional networks for semantic segmentation</article-title>,&#x201d; in <conf-name>Proceedings of the IEEE conference on computer vision and pattern recognition</conf-name>. <conf-loc>Boston, MA, USA</conf-loc>. <fpage>3431</fpage>&#x2013;<lpage>3440</lpage>.</citation>
</ref>
<ref id="B41">
<citation citation-type="journal">
<person-group person-group-type="author">
<name>
<surname>Lonsdale</surname> <given-names>P.</given-names>
</name>
</person-group> (<year>1977</year>). <article-title>Clustering of suspension-feeding macrobenthos near abyssal hydrothermal vents at oceanic spreading centers</article-title>. <source>Deep Sea Res.</source> <volume>24</volume> (<issue>9</issue>), <fpage>857</fpage>&#x2013;<lpage>863</lpage>. doi:&#xa0;<pub-id pub-id-type="doi">10.1016/0146-6291(77)90478-7</pub-id>
</citation>
</ref>
<ref id="B42">
<citation citation-type="journal">
<person-group person-group-type="author">
<name>
<surname>Lu</surname> <given-names>H.</given-names>
</name>
<name>
<surname>Yang</surname> <given-names>R.</given-names>
</name>
<name>
<surname>Deng</surname> <given-names>Z.</given-names>
</name>
<name>
<surname>Zhang</surname> <given-names>Y.</given-names>
</name>
<name>
<surname>Gao</surname> <given-names>G.</given-names>
</name>
<name>
<surname>Lan</surname> <given-names>R.</given-names>
</name>
</person-group> (<year>2021</year>). <article-title>Chinese image captioning via fuzzy attention-based DenseNet-BiLSTM</article-title>. <source>ACM Trans. Multimedia Computing Communications Appl. (TOMM)</source> <volume>17</volume> (<issue>1s</issue>), <fpage>1</fpage>&#x2013;<lpage>18</lpage>. doi: <pub-id pub-id-type="doi">10.1145/3422668</pub-id>
</citation>
</ref>
<ref id="B43">
<citation citation-type="journal">
<person-group person-group-type="author">
<name>
<surname>Luther</surname> <given-names>G. W.</given-names>
</name>
<name>
<surname>Rozan</surname> <given-names>T. F.</given-names>
</name>
<name>
<surname>Taillefert</surname> <given-names>M.</given-names>
</name>
<name>
<surname>Nuzzio</surname> <given-names>D. B.</given-names>
</name>
<name>
<surname>Meo</surname> <given-names>C. D.</given-names>
</name>
<name>
<surname>Shank</surname> <given-names>T. M.</given-names>
</name>
<etal/>
</person-group>. (<year>2001</year>). <article-title>Chemical speciation drives hydrothermal vent ecology</article-title>. <source>Nature</source> <volume>410</volume>, <fpage>813</fpage>&#x2013;<lpage>816</lpage>. doi: <pub-id pub-id-type="doi">10.1038/35071069</pub-id>
</citation>
</ref>
<ref id="B44">
<citation citation-type="confproc">
<person-group person-group-type="author">
<name>
<surname>MacMillan</surname> <given-names>R. A.</given-names>
</name>
<name>
<surname>McNabb</surname> <given-names>D. H.</given-names>
</name>
<name>
<surname>Jones</surname> <given-names>R. K.</given-names>
</name>
</person-group> (<year>2000</year>). &#x201c;<article-title>Automated landform classification using DEMs: a conceptual framework for a multi-level, hierarchy of hydrologically and geomorphologicaly oriented physiographic mapping units</article-title>,&#x201d; in <conf-name>Proceedings of the 4th International Conference on Integrating GIS and Environmental, Modeling: Problems, Prospects and Research Needs</conf-name>. <conf-loc>Banff, Alberta, Canada</conf-loc>.</citation>
</ref>
<ref id="B45">
<citation citation-type="journal">
<person-group person-group-type="author">
<name>
<surname>Mahmud</surname> <given-names>M. R.</given-names>
</name>
<name>
<surname>Yusof</surname> <given-names>O. M.</given-names>
</name>
</person-group> (<year>2006</year>). <article-title>The implementation of post-processing data thinning for multibeam echo sounding data</article-title>. <source>Int. Symposium Exhibition Geoinformation</source>. <volume>19</volume>, <fpage>21</fpage>.</citation>
</ref>
<ref id="B46">
<citation citation-type="confproc">
<person-group person-group-type="author">
<name>
<surname>Marthiniussen</surname> <given-names>R.</given-names>
</name>
<name>
<surname>Vestgard</surname> <given-names>K.</given-names>
</name>
<name>
<surname>Klepaker</surname> <given-names>R. A.</given-names>
</name>
<name>
<surname>Storkersen</surname> <given-names>N.</given-names>
</name>
</person-group> (<year>2004</year>). &#x201c;<article-title>HUGIN-AUV concept and operational experiences to date</article-title>,&#x201d; in <conf-name>Oceans' 04 MTS/IEEE Techno-Ocean'04 (IEEE Cat. No. 04CH37600)</conf-name>. <conf-loc>Kobe, Japan</conf-loc>. Vol. <volume>2</volume>. <fpage>846</fpage>&#x2013;<lpage>850</lpage>.</citation>
</ref>
<ref id="B47">
<citation citation-type="confproc">
<person-group person-group-type="author">
<name>
<surname>Martin-Abadal</surname> <given-names>M.</given-names>
</name>
<name>
<surname>Riutort-Ozcariz</surname> <given-names>I.</given-names>
</name>
<name>
<surname>Oliver-Codina</surname> <given-names>G.</given-names>
</name>
<name>
<surname>Gonzalez-Cid</surname> <given-names>Y.</given-names>
</name>
</person-group> (<year>2019</year>). &#x201c;<article-title>A deep learning solution for Posidonia oceanica seafloor habitat multiclass recognition</article-title>,&#x201d; in <conf-name>OCEANS 2019-Marseille</conf-name> <conf-sponsor>IEEE</conf-sponsor>. <conf-loc>Marseille, France</conf-loc>. <fpage>1</fpage>&#x2013;<lpage>7</lpage>. doi:&#xa0;<pub-id pub-id-type="doi">10.1109/OCEANSE.2019.8867304</pub-id>
</citation>
</ref>
<ref id="B48">
<citation citation-type="journal">
<person-group person-group-type="author">
<name>
<surname>Masetti</surname> <given-names>G.</given-names>
</name>
<name>
<surname>Mayer</surname> <given-names>L. A.</given-names>
</name>
<name>
<surname>Ward</surname> <given-names>L. G.</given-names>
</name>
</person-group> (<year>2018</year>). <article-title>A bathymetry-and reflectivity-based approach for seafloor segmentation</article-title>. <source>Geosciences</source> <volume>8</volume> (<issue>1</issue>), <elocation-id>14</elocation-id>. doi:&#xa0;<pub-id pub-id-type="doi">10.3390/geosciences8010014</pub-id>
</citation>
</ref>
<ref id="B49">
<citation citation-type="journal">
<person-group person-group-type="author">
<name>
<surname>McCollom</surname> <given-names>T. M.</given-names>
</name>
<name>
<surname>Shock</surname> <given-names>E. L.</given-names>
</name>
</person-group> (<year>1997</year>). <article-title>Geochemical constraints on chemolithoautotrophic metabolism by microorganisms in seafloor hydrothermal systems</article-title>. <source>Geochimica cosmochimica Acta</source> <volume>61 20</volume>, <fpage>4375</fpage>&#x2013;<lpage>4391</lpage>. doi: <pub-id pub-id-type="doi">10.1016/S0016-7037(97)00241-X</pub-id>
</citation>
</ref>
<ref id="B50">
<citation citation-type="book">
<person-group person-group-type="author">
<name>
<surname>Millar</surname> <given-names>S. W. S.</given-names>
</name>
</person-group> (<year>2013</year>). &#x201c;<article-title>Spatial and temporal scales in geomorphology</article-title>,&#x201d; in <source>The Foundations of Geomorphology</source> (<publisher-loc>London</publisher-loc>: <publisher-name>Elsevier</publisher-name>), <fpage>130</fpage>&#x2013;<lpage>145</lpage>.</citation>
</ref>
<ref id="B51">
<citation citation-type="journal">
<person-group person-group-type="author">
<name>
<surname>Misiuk</surname> <given-names>B.</given-names>
</name>
<name>
<surname>Brown</surname> <given-names>C. J.</given-names>
</name>
</person-group> (<year>2022</year>). <article-title>Multiple imputation of multibeam angular response data for high resolution full coverage seabed mapping</article-title>. <source>Mar. Geophysical Res.</source> <volume>43</volume> (<issue>1</issue>), <fpage>7</fpage>. doi: <pub-id pub-id-type="doi">10.1007/s11001-022-09471-3</pub-id>
</citation>
</ref>
<ref id="B52">
<citation citation-type="journal">
<person-group person-group-type="author">
<name>
<surname>Nakamura</surname> <given-names>K.</given-names>
</name>
<name>
<surname>Takai</surname> <given-names>K.</given-names>
</name>
</person-group> (<year>2014</year>). <article-title>Theoretical constraints of physical and chemical properties of hydrothermal fluids on variations in chemolithotrophic microbial communities in seafloor hydrothermal systems</article-title>. <source>Prog. Earth Planet Sci.</source> <volume>1</volume>, <fpage>5</fpage>. doi: <pub-id pub-id-type="doi">10.1186/2197-4284-1-5</pub-id>
</citation>
</ref>
<ref id="B53">
<citation citation-type="journal">
<person-group person-group-type="author">
<name>
<surname>Neil</surname> <given-names>H.</given-names>
</name>
<name>
<surname>Lamarche</surname> <given-names>G.</given-names>
</name>
<name>
<surname>Mitchell</surname> <given-names>J.</given-names>
</name>
<name>
<surname>MacKay</surname> <given-names>K.</given-names>
</name>
<name>
<surname>Pallentin</surname> <given-names>A.</given-names>
</name>
</person-group> (<year>2019</year>). <article-title>Multibeam echosounder mapping to identify seafloor habitats</article-title>. In <conf-name>The 8th Land Administration Domain Model Workshop</conf-name>. <conf-loc>Kuala Lumpur, Malaysia</conf-loc>.</citation>
</ref>
<ref id="B54">
<citation citation-type="journal">
<person-group person-group-type="author">
<name>
<surname>Nishizawa</surname> <given-names>A.</given-names>
</name>
<name>
<surname>Kaneda</surname> <given-names>K.</given-names>
</name>
<name>
<surname>Watanabe</surname> <given-names>N.</given-names>
</name>
<name>
<surname>Oikawa</surname> <given-names>M.</given-names>
</name>
</person-group> (<year>2009</year>). <article-title>Seismic structure of the subducting seamounts on the trench axis: Erimo Seamount and Daiichi-Kashima Seamount, northern and southern ends of the Japan Trench</article-title>. <source>Earth Planets Space</source> <volume>61</volume> (<issue>3</issue>), <fpage>e5</fpage>&#x2013;<lpage>e8</lpage>. doi: <pub-id pub-id-type="doi">10.1186/BF03352912</pub-id>
</citation>
</ref>
<ref id="B55">
<citation citation-type="confproc">
<person-group person-group-type="author">
<name>
<surname>Panish</surname> <given-names>R.</given-names>
</name>
<name>
<surname>Taylor</surname> <given-names>M.</given-names>
</name>
</person-group> (<year>2011</year>). &#x201c;<article-title>Achieving high navigation accuracy using inertial navigation systems in autonomous underwater vehicles</article-title>,&#x201d; in <conf-name>OCEANS 2011 IEEE-Spain</conf-name> <conf-sponsor>IEEE</conf-sponsor>. <conf-loc>Santander, Spain</conf-loc>. <fpage>1</fpage>&#x2013;<lpage>7</lpage>. doi:&#xa0;<pub-id pub-id-type="doi">10.1109/Oceans-Spain.2011.6003517</pub-id>
</citation>
</ref>
<ref id="B56">
<citation citation-type="journal">
<person-group person-group-type="author">
<name>
<surname>Perez</surname> <given-names>J. A. A.</given-names>
</name>
<name>
<surname>Vizuete</surname> <given-names>R. S.</given-names>
</name>
<name>
<surname>Ramil</surname> <given-names>F.</given-names>
</name>
<name>
<surname>Castillo</surname> <given-names>S.</given-names>
</name>
</person-group> (<year>2022</year>). <article-title>FISH, cephalopods and associated habitats of the Discovery rise seamounts, Southeast Atlantic</article-title>. <source>Deep Sea Res. Part I: Oceanographic Res. Papers</source> <volume>188</volume>, <fpage>103849</fpage>. doi: <pub-id pub-id-type="doi">10.1016/j.dsr.2022.103849</pub-id>
</citation>
</ref>
<ref id="B57">
<citation citation-type="journal">
<person-group person-group-type="author">
<name>
<surname>Phillips</surname> <given-names>J. D.</given-names>
</name>
</person-group> (<year>1988</year>). <article-title>The role of spatial scale in geomorphic systems</article-title>. <source>Geographical Anal.</source> <volume>20</volume> (<issue>4</issue>), <fpage>308</fpage>&#x2013;<lpage>317</lpage>. doi: <pub-id pub-id-type="doi">10.1111/j.1538-4632.1988.tb00185.x</pub-id>
</citation>
</ref>
<ref id="B58">
<citation citation-type="journal">
<person-group person-group-type="author">
<name>
<surname>Pierdomenico</surname> <given-names>M.</given-names>
</name>
<name>
<surname>Guida</surname> <given-names>V. G.</given-names>
</name>
<name>
<surname>Macelloni</surname> <given-names>L.</given-names>
</name>
<name>
<surname>Chiocci</surname> <given-names>F. L.</given-names>
</name>
<name>
<surname>Rona</surname> <given-names>P. A.</given-names>
</name>
<name>
<surname>Scranton</surname> <given-names>M. I.</given-names>
</name>
<etal/>
</person-group>. (<year>2015</year>). <article-title>Sedimentary facies, geomorphic features and habitat distribution at the Hudson Canyon head from AUV multibeam data</article-title>. <source>Deep Sea Res. Part II: Topical Stud. Oceanography</source> <volume>121</volume>, <fpage>112</fpage>&#x2013;<lpage>125</lpage>. doi: <pub-id pub-id-type="doi">10.1016/j.dsr2.2015.04.016</pub-id>
</citation>
</ref>
<ref id="B59">
<citation citation-type="journal">
<person-group person-group-type="author">
<name>
<surname>Pillay</surname> <given-names>T.</given-names>
</name>
<name>
<surname>Cawthra</surname> <given-names>H. C.</given-names>
</name>
<name>
<surname>Lombard</surname> <given-names>A. T.</given-names>
</name>
</person-group> (<year>2020</year>). <article-title>Characterisation of seafloor substrate using advanced processing of multibeam bathymetry, backscatter, and sidescan sonar in Table Bay, South Africa</article-title>. <source>Mar. Geology</source> <volume>429</volume>, <fpage>106332</fpage>. doi:&#xa0;<pub-id pub-id-type="doi">10.1016/j.margeo.2020.106332</pub-id>
</citation>
</ref>
<ref id="B60">
<citation citation-type="confproc">
<person-group person-group-type="author">
<name>
<surname>Pizarro</surname> <given-names>O.</given-names>
</name>
<name>
<surname>Williams</surname> <given-names>S. B.</given-names>
</name>
<name>
<surname>Jakuba</surname> <given-names>M. V.</given-names>
</name>
<name>
<surname>Johnson-Roberson</surname> <given-names>M.</given-names>
</name>
<name>
<surname>Mahon</surname> <given-names>I.</given-names>
</name>
<name>
<surname>Bryson</surname> <given-names>M.</given-names>
</name>
<etal/>
</person-group>. (<year>2013</year>). &#x201c;<article-title>Benthic monitoring with robotic platforms&#x2014;the experience of Australia</article-title>,&#x201d; in <conf-name>2013 IEEE international underwater technology symposium (UT)</conf-name>. <conf-loc>Tokyo</conf-loc>. <fpage>1</fpage>&#x2013;<lpage>10</lpage>. doi:&#xa0;<pub-id pub-id-type="doi">10.1109/UT.2013.6519909</pub-id>
</citation>
</ref>
<ref id="B61">
<citation citation-type="journal">
<person-group person-group-type="author">
<name>
<surname>Qin</surname> <given-names>X.</given-names>
</name>
<name>
<surname>Luo</surname> <given-names>X.</given-names>
</name>
<name>
<surname>Wu</surname> <given-names>Z.</given-names>
</name>
<name>
<surname>Shang</surname> <given-names>J.</given-names>
</name>
</person-group> (<year>2021</year>). <article-title>Optimizing the sediment classification of small side-scan sonar images based on deep learning</article-title>. <source>IEEE Access</source> <volume>9</volume>, <fpage>29416</fpage>&#x2013;<lpage>29428</lpage>. doi: <pub-id pub-id-type="doi">10.1109/ACCESS.2021.3052206</pub-id>
</citation>
</ref>
<ref id="B62">
<citation citation-type="book">
<person-group person-group-type="author">
<name>
<surname>Rimavicius</surname> <given-names>T.</given-names>
</name>
<name>
<surname>Gelzinis</surname> <given-names>A.</given-names>
</name>
</person-group> (<year>2017</year>). &#x201c;<article-title>A comparison of the deep learning methods for solving seafloor image classification task</article-title>,&#x201d; in <source>Communications in Computer and Information Science</source>, ed. <person-group person-group-type="editor">
<name>
<surname>Mobasheri</surname> <given-names>A.</given-names>
</name>
</person-group> (<publisher-loc>Berlin, Germany</publisher-loc>: <publisher-name>Springer, Cham</publisher-name>), <fpage>442</fpage>&#x2013;<lpage>453</lpage>. doi:&#xa0;<pub-id pub-id-type="doi">10.1007/978-3-319-67642-5_37</pub-id>
</citation>
</ref>
<ref id="B63">
<citation citation-type="journal">
<person-group person-group-type="author">
<name>
<surname>Sarrazin</surname> <given-names>J.</given-names>
</name>
<name>
<surname>Juniper</surname> <given-names>S. K.</given-names>
</name>
<name>
<surname>Massoth</surname> <given-names>G. J.</given-names>
</name>
<name>
<surname>Legendre</surname> <given-names>P.</given-names>
</name>
</person-group> (<year>1999</year>). <article-title>Physical and chemical factors influencing species distributions on hydrothermal sulfide edifices of the Juan de Fuca Ridge, northeast Pacific</article-title>. <source>Mar. Ecol. Prog. Ser.</source> <volume>190</volume>, <fpage>89</fpage>&#x2013;<lpage>112</lpage>. doi: <pub-id pub-id-type="doi">10.3354/meps190089</pub-id>
</citation>
</ref>
<ref id="B64">
<citation citation-type="journal">
<person-group person-group-type="author">
<name>
<surname>Sawa</surname> <given-names>T.</given-names>
</name>
<name>
<surname>Aoki</surname> <given-names>T.</given-names>
</name>
<name>
<surname>Yamamoto</surname> <given-names>I.</given-names>
</name>
<name>
<surname>Tsukioka</surname> <given-names>S.</given-names>
</name>
<name>
<surname>Yoshida</surname> <given-names>H.</given-names>
</name>
<name>
<surname>Hyakudome</surname> <given-names>T.</given-names>
</name>
<etal/>
</person-group>. (<year>2005</year>). <article-title>Performance of the fuel cell underwater vehicle URASHIMA</article-title>. <source>Acoustical Sci. Technol.</source> <volume>26</volume> (<issue>3</issue>), <fpage>249</fpage>&#x2013;<lpage>257</lpage>. doi:&#xa0;<pub-id pub-id-type="doi">10.1250/ast.26.249</pub-id>
</citation>
</ref>
<ref id="B65">
<citation citation-type="journal">
<person-group person-group-type="author">
<name>
<surname>Shary</surname> <given-names>P. A.</given-names>
</name>
</person-group> (<year>1995</year>). <article-title>Land surface in gravity points classification by a complete system of curvatures</article-title>. <source>Math. Geol</source> <volume>27</volume> (<issue>3</issue>), <fpage>373</fpage>&#x2013;<lpage>390</lpage>. doi: <pub-id pub-id-type="doi">10.1007/BF02084608</pub-id>
</citation>
</ref>
<ref id="B66">
<citation citation-type="journal">
<person-group person-group-type="author">
<name>
<surname>Singh</surname> <given-names>H.</given-names>
</name>
<name>
<surname>Armstrong</surname> <given-names>R.</given-names>
</name>
<name>
<surname>Gilbes</surname> <given-names>F.</given-names>
</name>
<name>
<surname>Eustice</surname> <given-names>R.</given-names>
</name>
<name>
<surname>Roman</surname> <given-names>C.</given-names>
</name>
<name>
<surname>Pizarro</surname> <given-names>O.</given-names>
</name>
<etal/>
</person-group>. (<year>2004</year>). <article-title>Imaging coral I: imaging coral habitats with the SeaBED AUV</article-title>. <source>Subsurface Sens. Technol. Appl.</source> <volume>5</volume> (<issue>1</issue>), <fpage>25</fpage>&#x2013;<lpage>42</lpage>. doi: <pub-id pub-id-type="doi">10.1023/B:SSTA.0000018445.25977.f3</pub-id>
</citation>
</ref>
<ref id="B67">
<citation citation-type="journal">
<person-group person-group-type="author">
<name>
<surname>Smale</surname> <given-names>D. A.</given-names>
</name>
<name>
<surname>Kendrick</surname> <given-names>G. A.</given-names>
</name>
<name>
<surname>Harvey</surname> <given-names>E. S.</given-names>
</name>
<name>
<surname>Langlois</surname> <given-names>T. J.</given-names>
</name>
<name>
<surname>Hovey</surname> <given-names>R. K.</given-names>
</name>
<name>
<surname>Van Niel</surname>
</name>
<etal/>
</person-group>. (<year>2012</year>). <article-title>Regional-scale benthic monitoring for ecosystem-based fisheries management (EBFM) using an autonomous underwater vehicle (AUV)</article-title>. <source>ICES J. Mar. Science: J. du Conseil</source> <volume>69</volume>, <fpage>1108</fpage>&#x2013;<lpage>1118</lpage>. doi: <pub-id pub-id-type="doi">10.1093/icesjms/fss082</pub-id>
</citation>
</ref>
<ref id="B68">
<citation citation-type="journal">
<person-group person-group-type="author">
<name>
<surname>Snelgrove</surname> <given-names>P. V.</given-names>
</name>
</person-group> (<year>1998</year>). <article-title>The biodiversity of macrofaunal organisms in marine sediments</article-title>. <source>Biodiversity Conserv.</source> <volume>7</volume> (<issue>9</issue>), <fpage>1123</fpage>&#x2013;<lpage>1132</lpage>. doi: <pub-id pub-id-type="doi">10.1023/A:1008867313340</pub-id>
</citation>
</ref>
<ref id="B69">
<citation citation-type="journal">
<person-group person-group-type="author">
<name>
<surname>Swanborn</surname> <given-names>D. J. B.</given-names>
</name>
<name>
<surname>Huvenne</surname> <given-names>V. A. I.</given-names>
</name>
<name>
<surname>Malpas</surname> <given-names>T.</given-names>
</name>
<name>
<surname>Pittman</surname> <given-names>S. J.</given-names>
</name>
<name>
<surname>Rogers</surname> <given-names>A. D.</given-names>
</name>
<name>
<surname>Taylor</surname> <given-names>M. L.</given-names>
</name>
<etal/>
</person-group>. (<year>2023</year>). <article-title>Seamount seascape composition and configuration shape Southwest Indian Ridge fish assemblages</article-title>. <source>Deep Sea Res. Part I: Oceanographic Res. Papers</source> <volume>191</volume>, <fpage>103921</fpage>. doi: <pub-id pub-id-type="doi">10.1016/j.dsr.2022.103921</pub-id>
</citation>
</ref>
<ref id="B70">
<citation citation-type="confproc">
<person-group person-group-type="author">
<name>
<surname>Tamura</surname> <given-names>K.</given-names>
</name>
<name>
<surname>Aoki</surname> <given-names>T.</given-names>
</name>
<name>
<surname>Nakamura</surname> <given-names>T.</given-names>
</name>
<name>
<surname>Tsukioka</surname> <given-names>S.</given-names>
</name>
<name>
<surname>Murashima</surname> <given-names>T.</given-names>
</name>
<name>
<surname>Ochi</surname> <given-names>H.</given-names>
</name>
<etal/>
</person-group>. (<year>2000</year>). &#x201c;<article-title>The development of the AUV-urashima</article-title>,&#x201d; in <conf-name>OCEANS 2000 MTS/IEEE Conference and Exhibition. Conference Proceedings (Cat. No. 00CH37158)</conf-name> <conf-sponsor>IEEE</conf-sponsor>. <conf-loc>Providence, RI, USA</conf-loc>. Vol. <volume>1</volume>. <fpage>139</fpage>&#x2013;<lpage>146</lpage>.</citation>
</ref>
<ref id="B71">
<citation citation-type="journal">
<person-group person-group-type="author">
<name>
<surname>Teixeira</surname> <given-names>J. B.</given-names>
</name>
<name>
<surname>Martins</surname> <given-names>A. S.</given-names>
</name>
<name>
<surname>Pinheiro</surname> <given-names>H. T.</given-names>
</name>
<name>
<surname>Secchin</surname> <given-names>N. A.</given-names>
</name>
<name>
<surname>de Moura</surname> <given-names>R. L.</given-names>
</name>
<name>
<surname>Bastos</surname> <given-names>A. C.</given-names>
</name>
</person-group> (<year>2013</year>). <article-title>Traditional ecological knowledge and the mapping of benthic marine habitats</article-title>. <source>J. Environ. Manage.</source> <volume>115</volume>, <fpage>241</fpage>&#x2013;<lpage>250</lpage>. doi: <pub-id pub-id-type="doi">10.1016/j.jenvman.2012.11.020</pub-id>
</citation>
</ref>
<ref id="B72">
<citation citation-type="journal">
<person-group person-group-type="author">
<name>
<surname>Thornton</surname> <given-names>B.</given-names>
</name>
<name>
<surname>Bodenmann</surname> <given-names>A.</given-names>
</name>
<name>
<surname>Pizarro</surname> <given-names>O.</given-names>
</name>
<name>
<surname>Williams</surname> <given-names>S. B.</given-names>
</name>
<name>
<surname>Friedman</surname> <given-names>A.</given-names>
</name>
<name>
<surname>Nakajima</surname> <given-names>R.</given-names>
</name>
<etal/>
</person-group>. (<year>2016</year>). <article-title>Biometric assessment of deep-sea vent megabenthic communities using multi-resolution 3D image reconstructions</article-title>. <source>Deep Sea Res. Part I: Oceanographic Res. Papers</source> <volume>116</volume>, <fpage>200</fpage>&#x2013;<lpage>219</lpage>. doi:&#xa0;<pub-id pub-id-type="doi">10.1016/j.dsr.2016.08.009</pub-id>
</citation>
</ref>
<ref id="B73">
<citation citation-type="journal">
<person-group person-group-type="author">
<name>
<surname>Trenkel</surname> <given-names>V. M.</given-names>
</name>
<name>
<surname>Lorance</surname> <given-names>P.</given-names>
</name>
<name>
<surname>Mah&#xe9;vas</surname> <given-names>S.</given-names>
</name>
</person-group> (<year>2004</year>). <article-title>Do visual transects provide true population density estimates for deepwater fish</article-title>? <source>ICES J. Mar. Sci.</source> <volume>61</volume> (<issue>7</issue>), <fpage>1050</fpage>&#x2013;<lpage>1056</lpage>. doi:&#xa0;<pub-id pub-id-type="doi">10.1016/j.icesjms.2004.06.002</pub-id>
</citation>
</ref>
<ref id="B74">
<citation citation-type="journal">
<person-group person-group-type="author">
<name>
<surname>Urra</surname> <given-names>J.</given-names>
</name>
<name>
<surname>Palomino</surname> <given-names>D.</given-names>
</name>
<name>
<surname>Lozano</surname> <given-names>P.</given-names>
</name>
<name>
<surname>Gonz&#xe1;lez-Garc&#xed;a</surname> <given-names>E.</given-names>
</name>
<name>
<surname>Farias</surname> <given-names>C.</given-names>
</name>
<name>
<surname>Mateo-Ram&#xed;rez</surname> <given-names>&#xc1;.</given-names>
</name>
<etal/>
</person-group>. (<year>2021</year>). <article-title>Deep-sea habitat characterization using acoustic data and underwater imagery in Gazul mud volcano (Gulf of C&#xe1;diz, NE Atlantic)</article-title>. <source>Deep Sea Res. Part I: Oceanographic Res. Papers</source> <volume>169</volume>, <fpage>103458</fpage>. doi: <pub-id pub-id-type="doi">10.1016/j.dsr.2020.103458</pub-id>
</citation>
</ref>
<ref id="B75">
<citation citation-type="journal">
<person-group person-group-type="author">
<name>
<surname>Van Dover</surname> <given-names>C. L.</given-names>
</name>
<name>
<surname>German</surname> <given-names>C. R.</given-names>
</name>
<name>
<surname>Speer</surname> <given-names>K. G.</given-names>
</name>
<name>
<surname>Parson</surname> <given-names>L. M.</given-names>
</name>
<name>
<surname>Vrijenhoek</surname> <given-names>R. C.</given-names>
</name>
</person-group> (<year>2002</year>). <article-title>Evolution and biogeography of deep-sea vent and seep invertebrates</article-title>. <source>Science</source> <volume>295</volume>, <fpage>1253</fpage>&#x2013;<lpage>1257</lpage>. doi: <pub-id pub-id-type="doi">10.1126/science.1067361</pub-id>
</citation>
</ref>
<ref id="B76">
<citation citation-type="confproc">
<person-group person-group-type="author">
<name>
<surname>Vaswani</surname> <given-names>A.</given-names>
</name>
<name>
<surname>Shazeer</surname> <given-names>N. M.</given-names>
</name>
<name>
<surname>Parmar</surname> <given-names>N.</given-names>
</name>
<name>
<surname>Uszkoreit</surname> <given-names>J.</given-names>
</name>
<name>
<surname>Jones</surname> <given-names>L.</given-names>
</name>
<name>
<surname>Gomez</surname> <given-names>A. N.</given-names>
</name>
<etal/>
</person-group>. (<year>2017</year>). <source>Attention is All you Need</source> (ArXiv). <conf-name>NIPS'17: Proceedings of the 31st International Conference on Neural Information Processing Systems</conf-name>, <conf-loc>Long Beach. California, USA</conf-loc>.</citation>
</ref>
<ref id="B77">
<citation citation-type="confproc">
<person-group person-group-type="author">
<name>
<surname>Wang</surname> <given-names>R. J.</given-names>
</name>
<name>
<surname>Li</surname> <given-names>X.</given-names>
</name>
<name>
<surname>Ao</surname> <given-names>S.</given-names>
</name>
<name>
<surname>Ling</surname> <given-names>C. X.</given-names>
</name>
</person-group> (<year>2018</year>). <source>Pelee: A Real-Time Object Detection System on Mobile Devices</source> (<publisher-name>NeurIPS</publisher-name>). <conf-name>NIPS'18: Proceedings of the 32nd International Conference on Neural Information Processing Systems</conf-name>. <conf-loc>Montr&#xe9;al, Canada</conf-loc>. <fpage>3</fpage>&#x2013;<lpage>8</lpage>.</citation>
</ref>
<ref id="B78">
<citation citation-type="journal">
<person-group person-group-type="author">
<name>
<surname>Wang</surname> <given-names>A.</given-names>
</name>
<name>
<surname>Tao</surname> <given-names>C.</given-names>
</name>
<name>
<surname>Zhang</surname> <given-names>G.</given-names>
</name>
<name>
<surname>Shen</surname> <given-names>C.</given-names>
</name>
<name>
<surname>Liu</surname> <given-names>Y.</given-names>
</name>
</person-group> (<year>2021</year>b). <article-title>Seafloor classification based on deep-sea multibeam data&#x2014;Application to the Southwest Indian Ridge at 50.47&#xb0; E</article-title>. <source>J. Appl. Geophysics</source> <volume>185</volume>, <elocation-id>104259</elocation-id>. doi:&#xa0;<pub-id pub-id-type="doi">10.1016/j.jappgeo.2021.104259</pub-id>
</citation>
</ref>
<ref id="B79">
<citation citation-type="journal">
<person-group person-group-type="author">
<name>
<surname>Wang</surname> <given-names>M.</given-names>
</name>
<name>
<surname>Wu</surname> <given-names>Z.</given-names>
</name>
<name>
<surname>Best</surname> <given-names>J.</given-names>
</name>
<name>
<surname>Yang</surname> <given-names>F.</given-names>
</name>
<name>
<surname>Li</surname> <given-names>X.</given-names>
</name>
<name>
<surname>Zhao</surname> <given-names>D.</given-names>
</name>
<etal/>
</person-group>. (<year>2021</year>a). <article-title>Using multibeam backscatter strength to analyze the distribution of manganese nodules: A case study of seamounts in the Western Pacific Ocean</article-title>. <source>Appl. Acoustics</source> <volume>173</volume>, <fpage>107729</fpage>. doi: <pub-id pub-id-type="doi">10.1016/j.apacoust.2020.107729</pub-id>
</citation>
</ref>
<ref id="B80">
<citation citation-type="book">
<person-group person-group-type="author">
<name>
<surname>Williams</surname> <given-names>S. B.</given-names>
</name>
<name>
<surname>Pizarro</surname> <given-names>O.</given-names>
</name>
<name>
<surname>Jakuba</surname> <given-names>M.</given-names>
</name>
<name>
<surname>Barrett</surname> <given-names>N.</given-names>
</name>
</person-group> (<year>2010</year>). &#x201c;<article-title>AUV benthic habitat mapping in south eastern Tasmania</article-title>,&#x201d; in <source>Field and Service Robotics</source> (<publisher-loc>Berlin, Heidelberg</publisher-loc>: <publisher-name>Springer</publisher-name>), <fpage>275</fpage>&#x2013;<lpage>284</lpage>.</citation>
</ref>
<ref id="B81">
<citation citation-type="journal">
<person-group person-group-type="author">
<name>
<surname>Williams</surname> <given-names>S. B.</given-names>
</name>
<name>
<surname>Pizarro</surname> <given-names>O. R.</given-names>
</name>
<name>
<surname>Jakuba</surname> <given-names>M. V.</given-names>
</name>
<name>
<surname>Johnson</surname> <given-names>C. R.</given-names>
</name>
<name>
<surname>Barrett</surname> <given-names>N. S.</given-names>
</name>
<name>
<surname>Babcock</surname> <given-names>R. C.</given-names>
</name>
<etal/>
</person-group>. (<year>2012</year>). <article-title>Monitoring of benthic reference sites: using an autonomous underwater vehicle</article-title>. <source>IEEE Robotics Automation Magazine</source> <volume>19</volume> (<issue>1</issue>), <fpage>73</fpage>&#x2013;<lpage>84</lpage>. doi:&#xa0;<pub-id pub-id-type="doi">10.1109/MRA.2011.2181772</pub-id>
</citation>
</ref>
<ref id="B82">
<citation citation-type="journal">
<person-group person-group-type="author">
<name>
<surname>Wong</surname> <given-names>F. Y.</given-names>
</name>
</person-group> (<year>1973</year>). <article-title>Micro-scale geomorphology features</article-title>. <source>Photogrammetric Eng.</source> <volume>39</volume>.</citation>
</ref>
<ref id="B83">
<citation citation-type="confproc">
<person-group person-group-type="author">
<name>
<surname>Woo</surname> <given-names>S.</given-names>
</name>
<name>
<surname>Park</surname> <given-names>J.</given-names>
</name>
<name>
<surname>Lee</surname> <given-names>J. Y.</given-names>
</name>
<name>
<surname>Kweon</surname> <given-names>I. S.</given-names>
</name>
</person-group> (<year>2018</year>). &#x201c;<article-title>Cbam: Convolutional block attention module</article-title>,&#x201d; in <conf-name>Proceedings of the European conference on computer vision (ECCV)</conf-name>. <conf-loc>Munich, Germany</conf-loc>. <fpage>3</fpage>&#x2013;<lpage>19</lpage>.</citation>
</ref>
<ref id="B84">
<citation citation-type="journal">
<person-group person-group-type="author">
<name>
<surname>Wu</surname> <given-names>Z.</given-names>
</name>
<name>
<surname>Yang</surname> <given-names>F.</given-names>
</name>
<name>
<surname>Tang</surname> <given-names>Y.</given-names>
</name>
<name>
<surname>Wu</surname> <given-names>Z.</given-names>
</name>
<name>
<surname>Yang</surname> <given-names>F.</given-names>
</name>
<name>
<surname>Tang</surname> <given-names>Y.</given-names>
</name>
</person-group> (<year>2021</year>). <article-title>Multibeam bathymetric technology</article-title>. <source>High-resolution Seafloor Survey Appl.</source>, <fpage>21</fpage>&#x2013;<lpage>76</lpage>. doi: <pub-id pub-id-type="doi">10.1007/978-981-15-9750-3_2</pub-id>
</citation>
</ref>
<ref id="B85">
<citation citation-type="journal">
<person-group person-group-type="author">
<name>
<surname>Xiao</surname> <given-names>B.</given-names>
</name>
<name>
<surname>Yang</surname> <given-names>Z.</given-names>
</name>
<name>
<surname>Qiu</surname> <given-names>X.</given-names>
</name>
<name>
<surname>Xiao</surname> <given-names>J.</given-names>
</name>
<name>
<surname>Wang</surname> <given-names>G.</given-names>
</name>
<name>
<surname>Zeng</surname> <given-names>W.</given-names>
</name>
<etal/>
</person-group>. (<year>2021</year>). <article-title>PAM-DenseNet: A deep convolutional neural network for computer-aided COVID-19 diagnosis</article-title>. <source>IEEE Trans. Cybernetics</source> <volume>52</volume> (<issue>11</issue>), <fpage>12163</fpage>&#x2013;<lpage>12174</lpage>. doi: <pub-id pub-id-type="doi">10.1109/tcyb.2020.3042837</pub-id>
</citation>
</ref>
<ref id="B86">
<citation citation-type="confproc">
<person-group person-group-type="author">
<name>
<surname>Zhao</surname> <given-names>H.</given-names>
</name>
<name>
<surname>Shi</surname> <given-names>J.</given-names>
</name>
<name>
<surname>Qi</surname> <given-names>X.</given-names>
</name>
<name>
<surname>Wang</surname> <given-names>X.</given-names>
</name>
<name>
<surname>Jia</surname> <given-names>J.</given-names>
</name>
</person-group> (<year>2017</year>). &#x201c;<article-title>Pyramid scene parsing network</article-title>,&#x201d; in <conf-name>Proceedings of the IEEE conference on computer vision and pattern recognition</conf-name>. <conf-loc>Honolulu, HI, USA</conf-loc>. <fpage>2881</fpage>&#x2013;<lpage>2890</lpage>. doi:&#xa0;<pub-id pub-id-type="doi">10.1109/CVPR.2017.660</pub-id>
</citation>
</ref>
<ref id="B87">
<citation citation-type="journal">
<person-group person-group-type="author">
<name>
<surname>Zhao</surname> <given-names>D.</given-names>
</name>
<name>
<surname>Wu</surname> <given-names>Z.</given-names>
</name>
<name>
<surname>Zhou</surname> <given-names>J.</given-names>
</name>
<name>
<surname>Wang</surname> <given-names>M.</given-names>
</name>
<name>
<surname>Liu</surname> <given-names>Z.</given-names>
</name>
</person-group> (<year>2020</year>). <article-title>Parameter group optimization by combining CUBE with surface filtering and its application</article-title>. <source>J. Geodesy Geoinformation Sci.</source> <volume>3</volume> (<issue>2</issue>), <fpage>81</fpage>. doi: <pub-id pub-id-type="doi">10.11947/j.JGGS.2020.0209</pub-id>
</citation>
</ref>
<ref id="B88">
<citation citation-type="confproc">
<person-group person-group-type="author">
<name>
<surname>Zhu</surname> <given-names>Y.</given-names>
</name>
<name>
<surname>Newsam</surname> <given-names>S.</given-names>
</name>
</person-group> (<year>2017</year>). &#x201c;<article-title>Densenet for dense flow</article-title>,&#x201d; in <conf-name>2017 IEEE international conference on image processing (ICIP)</conf-name> <conf-sponsor>IEEE</conf-sponsor>. <conf-loc>Beijing, China</conf-loc>. <fpage>790</fpage>&#x2013;<lpage>794</lpage>.</citation>
</ref>
</ref-list>
</back>
</article>