<?xml version="1.0" encoding="UTF-8"?>
<!DOCTYPE article PUBLIC "-//NLM//DTD Journal Publishing DTD v2.3 20070202//EN" "journalpublishing.dtd">
<article article-type="research-article" dtd-version="2.3" xml:lang="EN" xmlns:mml="http://www.w3.org/1998/Math/MathML" xmlns:xlink="http://www.w3.org/1999/xlink">
<front>
<journal-meta>
<journal-id journal-id-type="publisher-id">Front. Big Data</journal-id>
<journal-title>Frontiers in Big Data</journal-title>
<abbrev-journal-title abbrev-type="pubmed">Front. Big Data</abbrev-journal-title>
<issn pub-type="epub">2624-909X</issn>
<publisher>
<publisher-name>Frontiers Media S.A.</publisher-name>
</publisher>
</journal-meta>
<article-meta>
<article-id pub-id-type="publisher-id">762899</article-id>
<article-id pub-id-type="doi">10.3389/fdata.2021.762899</article-id>
<article-categories>
<subj-group subj-group-type="heading">
<subject>Big Data</subject>
<subj-group>
<subject>Original Research</subject>
</subj-group>
</subj-group>
</article-categories>
<title-group>
<article-title>Towards Semantically-Rich Spatial Network Representation Learning <italic>via</italic> Automated Feature Topic Pairing</article-title>
<alt-title alt-title-type="left-running-head">Wang et&#x20;al.</alt-title>
<alt-title alt-title-type="right-running-head">Automated Feature-Topic Pairing for SRL</alt-title>
</title-group>
<contrib-group>
<contrib contrib-type="author">
<name>
<surname>Wang</surname>
<given-names>Dongjie</given-names>
</name>
<xref ref-type="aff" rid="aff1">
<sup>1</sup>
</xref>
<uri xlink:href="https://loop.frontiersin.org/people/1451567/overview"/>
</contrib>
<contrib contrib-type="author">
<name>
<surname>Liu</surname>
<given-names>Kunpeng</given-names>
</name>
<xref ref-type="aff" rid="aff1">
<sup>1</sup>
</xref>
</contrib>
<contrib contrib-type="author">
<name>
<surname>Mohaisen</surname>
<given-names>David</given-names>
</name>
<xref ref-type="aff" rid="aff1">
<sup>1</sup>
</xref>
<uri xlink:href="https://loop.frontiersin.org/people/561543/overview"/>
</contrib>
<contrib contrib-type="author">
<name>
<surname>Wang</surname>
<given-names>Pengyang</given-names>
</name>
<xref ref-type="aff" rid="aff2">
<sup>2</sup>
</xref>
</contrib>
<contrib contrib-type="author">
<name>
<surname>Lu</surname>
<given-names>Chang-Tien</given-names>
</name>
<xref ref-type="aff" rid="aff3">
<sup>3</sup>
</xref>
</contrib>
<contrib contrib-type="author" corresp="yes">
<name>
<surname>Fu</surname>
<given-names>Yanjie</given-names>
</name>
<xref ref-type="aff" rid="aff1">
<sup>1</sup>
</xref>
<xref ref-type="corresp" rid="c001">&#x2a;</xref>
</contrib>
</contrib-group>
<aff id="aff1">
<label>
<sup>1</sup>
</label>Computer Science Department, University of Central Florida, <addr-line>Orlando</addr-line>, <addr-line>FL</addr-line>, <country>United&#x20;States</country>
</aff>
<aff id="aff2">
<label>
<sup>2</sup>
</label>Computer Science Department, University of Macau, <addr-line>Macau</addr-line>, <country>China</country>
</aff>
<aff id="aff3">
<label>
<sup>3</sup>
</label>Computer Science Department, Virginia Tech Falls Church, <addr-line>Falls Church</addr-line>, <addr-line>VA</addr-line>, <country>United&#x20;States</country>
</aff>
<author-notes>
<fn fn-type="edited-by">
<p>
<bold>Edited by:</bold> <ext-link ext-link-type="uri" xlink:href="https://loop.frontiersin.org/people/1179065/overview">Xun Zhou</ext-link>, The University of Iowa, United&#x20;States</p>
</fn>
<fn fn-type="edited-by">
<p>
<bold>Reviewed by:</bold> <ext-link ext-link-type="uri" xlink:href="https://loop.frontiersin.org/people/641893/overview">Yiqun Xie</ext-link>, University of Minnesota Twin Cities, United&#x20;States</p>
<p>
<ext-link ext-link-type="uri" xlink:href="https://loop.frontiersin.org/people/1454265/overview">Amin Vahedian Khezerlou</ext-link>, University of Wisconsin&#x2013;Whitewater, United&#x20;States</p>
</fn>
<corresp id="c001">&#x2a;Correspondence: Yanjie Fu, <email>yanjie.fu@ucf.edu</email>
</corresp>
<fn fn-type="other">
<p>This article was submitted to Data Mining and Management, a section of the journal Frontiers in Big&#x20;Data</p>
</fn>
</author-notes>
<pub-date pub-type="epub">
<day>20</day>
<month>10</month>
<year>2021</year>
</pub-date>
<pub-date pub-type="collection">
<year>2021</year>
</pub-date>
<volume>4</volume>
<elocation-id>762899</elocation-id>
<history>
<date date-type="received">
<day>23</day>
<month>08</month>
<year>2021</year>
</date>
<date date-type="accepted">
<day>20</day>
<month>09</month>
<year>2021</year>
</date>
</history>
<permissions>
<copyright-statement>Copyright &#xa9; 2021 Wang, Liu, Mohaisen, Wang, Lu and Fu.</copyright-statement>
<copyright-year>2021</copyright-year>
<copyright-holder>Wang, Liu, Mohaisen, Wang, Lu and Fu</copyright-holder>
<license xlink:href="http://creativecommons.org/licenses/by/4.0/">
<p>This is an open-access article distributed under the terms of the Creative Commons Attribution License (CC BY). The use, distribution or reproduction in other forums is permitted, provided the original author(s) and the copyright owner(s) are credited and that the original publication in this journal is cited, in accordance with accepted academic practice. No use, distribution or reproduction is permitted which does not comply with these&#x20;terms.</p>
</license>
</permissions>
<abstract>
<p>Automated characterization of spatial data is a kind of critical geographical intelligence. As an emerging technique for characterization, spatial Representation Learning (SRL) uses deep neural networks (DNNs) to learn non-linear embedded features of spatial data for characterization. However, SRL extracts features by internal layers of DNNs, and thus suffers from lacking semantic labels. Texts of spatial entities, on the other hand, provide semantic understanding of latent feature labels, but is insensible to deep SRL models. How can we teach a SRL model to discover appropriate topic labels in texts and pair learned features with the labels? This paper formulates a new problem: feature-topic pairing, and proposes a novel Particle Swarm Optimization (PSO) based deep learning framework. Specifically, we formulate the feature-topic pairing problem into an automated alignment task between 1) a latent embedding feature space and 2) a textual semantic topic space. We decompose the alignment of the two spaces into: 1) point-wise alignment, denoting the correlation between a topic distribution and an embedding vector; 2) pair-wise alignment, denoting the consistency between a feature-feature similarity matrix and a topic-topic similarity matrix. We design a PSO based solver to simultaneously select an optimal set of topics and learn corresponding features based on the selected topics. We develop a closed loop algorithm to iterate between 1) minimizing losses of representation reconstruction and feature-topic alignment and 2) searching the best topics. Finally, we present extensive experiments to demonstrate the enhanced performance of our method.</p>
</abstract>
<kwd-group>
<kwd>feature-topic pairing</kwd>
<kwd>semantic space</kwd>
<kwd>spatial space</kwd>
<kwd>spatial representation learning</kwd>
<kwd>spatial graph</kwd>
</kwd-group>
<contract-num rid="cn001">1755946 I2040950 2006889&#x20;2045567</contract-num>
<contract-sponsor id="cn001">National Science Foundation<named-content content-type="fundref-id">10.13039/100000001</named-content>
</contract-sponsor>
</article-meta>
</front>
<body>
<sec id="s1">
<title>1 Introduction</title>
<p>Critical infrastructures (e.g., transportation networks, power networks, social networks, water supply networks) often consist of spatially distributed entities that interact with each other, and have generated massive spatial-networked behavior data. Analyzing such data can identify trends, forecast future behavior, and detect anomalies. To enable effective analysis, it is critical to desire a new capability of automated characterization that effectively extract feature vectors from spatio-networked&#x20;data.</p>
<p>As one of the emerging techniques, representation learning can be adapted to learn non-linear embedded features of spatial network data, which we call spatial representation learning (SRL). There has been a rich body in SRL, including node embedding, autoencoder, random walk, adversarial learning, generative learning based methods with spatial data (<xref ref-type="bibr" rid="B42">Wang and Li, 2017</xref>; <xref ref-type="bibr" rid="B44">Wang et&#x20;al., 2018a</xref>; <xref ref-type="bibr" rid="B45">Wang et&#x20;al., 2018b</xref>; <xref ref-type="bibr" rid="B7">Chandra et&#x20;al., 2019</xref>; <xref ref-type="bibr" rid="B18">Jean et&#x20;al., 2019</xref>; <xref ref-type="bibr" rid="B43">Wang et&#x20;al., 2019a</xref>; <xref ref-type="bibr" rid="B47">Wang et&#x20;al., 2019b</xref>; <xref ref-type="bibr" rid="B55">Zhang Y. et&#x20;al., 2019</xref>; <xref ref-type="bibr" rid="B35">Shan et&#x20;al., 2020</xref>; <xref ref-type="bibr" rid="B48">Wang et&#x20;al., 2020c</xref>; <xref ref-type="bibr" rid="B49">Wang et&#x20;al., 2020d</xref>; <xref ref-type="bibr" rid="B39">Wang et&#x20;al., 2021</xref>). Although these works achieved remarkable success, the model interpretability is still a big limitation that hinders these SRL methods from applying in more secure, fair, and rigorous scenarios.</p>
<p>Lacking model interpretability is possible to cause damaging or controversial consequences in incomplete scenarios that are not well-studied (<xref ref-type="bibr" rid="B8">Doshi-Velez and Kim, 2017</xref>). For instance, in the autonomous driving scenario, the end-to-end autopilot system brings high safety risks for drivers<xref ref-type="fn" rid="FN1">
<sup>1</sup>
</xref>. In 2015, Google&#x2019;s photo app classifies images of black people as gorillas, which exposes the limitation of algorithms<xref ref-type="fn" rid="FN2">
<sup>2</sup>
</xref>. More seriously, widely used crime prediction software prefers to provide higher risk scores of future crimes for black defendants<xref ref-type="fn" rid="FN3">
<sup>3</sup>
</xref>. Model interpretability is one of the most important approaches to overcome these limitations. Thus how to enhance the model interpretability attracts much attention of researchers (<xref ref-type="bibr" rid="B9">Elshawi et&#x20;al., 2019</xref>; <xref ref-type="bibr" rid="B14">Hong et&#x20;al., 2020</xref>; <xref ref-type="bibr" rid="B36">Stiglic et&#x20;al., 2020</xref>; <xref ref-type="bibr" rid="B30">Poursabzi-Sangdeh et&#x20;al., 2021</xref>). But, many existing works reflect that there is a trade-off between model performance and model interpretability (<xref ref-type="bibr" rid="B28">Mori and Uchihira, 2019</xref>; <xref ref-type="bibr" rid="B32">Saisubramanian et&#x20;al., 2020</xref>). Can we not only improve the model interpretability but also keep the model performance becomes the research point of this&#x20;paper.</p>
<p>To relieve the limitations of prior literature and expand the application scenarios of SRL approaches, a novel SRL model should understand not just which features are effective, but also what these effective features stand for. This issue relates to two tasks: 1) deep representation learning; 2) label generation and matching for latent embedded features. In response, we formulate the problem as a task of feature-topic pairing (<xref ref-type="fig" rid="F1">Figure&#x20;1</xref>), which is to align a latent embedding feature space, consisting of multiple latent features, and a textual semantic topic space, consisting of multiple topic labels during SRL. The basic idea is to teach a machine to extract topic labels from texts, and then pair the labels with learned features. To that end, we propose to develop a novel deep learning framework to unify feature learning, topic selection, feature-topic matching.</p>
<fig id="F1" position="float">
<label>FIGURE 1</label>
<caption>
<p>The motivation of the feature-topic pairing problem: bridging the gap between feature embedding space and topic semantic space in representation learning.</p>
</caption>
<graphic xlink:href="fdata-04-762899-g001.tif"/>
</fig>
<p>There are three unique challenges (<xref ref-type="fig" rid="F2">Figure&#x20;2</xref>) in addressing this problem. 1) Label Generation Challenge. The semantically-rich texts of spatial entities describe their types, functions, and attribute-related information. For instance, on a real estate website, the texts of a residential community describe crime rates and events, great school ratings, nearby transportation facilities, grocery stores, companies, and universities. These texts, if properly analyzed, will help to identify which underlying features truly attract residents to pay more to live. However, these spatial texts are all unstructured, how can we construct a textual semantic topic space for spatial entities to support feature-topic pairing? 2) Measurement Challenge. Be sure to note that we aim to teach a machine to automatically perform the automated pairing between embedded features and topic labels in a self-optimizing fashion. As a result, a measurement is needed to quantify the alignment or matching score between the topic label space and the embedding feature space, in order to guide the machine about how to search. However, there is no standard measurement for quantifying the topic-embedding space alignment. Thus, what does form of measurement should be adopted? And how can we integrate the suitable measurement into the whole self-optimizing framework? 3) Optimization Challenge. Since the model needs to decide an optimized topic label subset, the feature-topic pairing problem evolves multiple machine learning tasks, including feature learning, topic label selection, and feature-topic matching. If the three tasks are separately completed step by step, there is no guarantee that they are globally optimized. So, how can we develop a deep optimization framework to jointly and simultaneously unify the three tasks?</p>
<fig id="F2" position="float">
<label>FIGURE 2</label>
<caption>
<p>Key challenges and main tasks of the feature-topic pairing problem.</p>
</caption>
<graphic xlink:href="fdata-04-762899-g002.tif"/>
</fig>
<p>To solve the three challenges, we develop a new PSO-based framework (named AutoFTP) that enclose the optimizations of feature learning, topic selection, and feature-topic pairing in a loop. Specifically, our contributions are:<list list-type="simple">
<list-item>
<p>1) Formulating the feature-topic pairing problem. Motivated by lacking feature labels in SRL, we formulate and develop a new problem: feature topic pairing. In the proposed model, we propose a new strategy: we first let an optimizer to automatically select K topics; the optimizer then guides the representation learner to learn K latent features that optimally align with the K topics.</p>
</list-item>
<list-item>
<p>2) Generating candidate topic labels. We propose a three step mining method to generate candidate topic labels. Specifically, we first extract keywords from the texts of all spatial entities. Then, we learn keyword embedding feature vectors with a pre-trained word model (<xref ref-type="bibr" rid="B12">He, 2014</xref>). Finally, we cluster all keyword embeddings by maximizing inter-topic distances and minimizing intra-topic distances to generate topics as candidate feature labels.</p>
</list-item>
<list-item>
<p>3) Quantifying feature-topic alignment measurement. We identify two types of feature-topic alignments: 1) point-wise alignment, and 2) pair-wise alignment. First, the point-wise alignment is to describe the correlation between an embedding feature vector and a categorical topic distribution. In particular, we maximize the correlation so that the distance between the distribution of the embedding vector space and the distribution of the topic semantic vector space can be minimized. The underlying motivation of point-wise alignment is originated from the fact that: if a topic density is high in describing a spatial entity, the topic&#x2019;s corresponding feature value is expected to be large to co-vary with the topic density. In this way, we align the distribution covariance of the two spaces. Second, the pair-wise alignment is to describe the consistency between a feature-feature similarity matrix and a topic-topic similarity matrix. In particular, we use the feature-feature similarity graph to describe the topology of the latent embedding feature space, and use the topic-topic similarity graph to describe the topology of the textual semantic topic space. If the two spaces are aligned, the two graphs (represented by matrices) are similar as&#x20;well.</p>
</list-item>
<list-item>
<p>4) Optimization in the loop. We develop a Particle Swarm Optimization (PSO)-based algorithm. In this algorithm, we first simultaneously optimize the representation learning loss, point-wise alignment loss, pair-wise alignment loss, and downstream task loss as the feedback for PSO. Guided by the feedback, the PSO based algorithm selects a better K-sized topic subset for feature-topic pairing. In particular, based on the loss function value, PSO iteratively generates topic masks (i.e.,&#x20;0&#x2013;1 indicators to select or deselect) to search the optimal topics for space pairing until the learning objective converges. In this way, the PSO jointly achieves topic selection, feature-topic pairing, and latent feature learning.</p>
</list-item>
</list>
</p>
<p>Finally, we evaluate our method using Beijing&#x2019;s urban geography and mobility data. For comparison we implemented a broad range of other algorithms. Results showed that our method consistently outperformed the competing methods. We perform ablation study, interpretability, robustness check, stability, sensitivity to justify our technical insights.</p>
</sec>
<sec id="s2">
<title>2 Preliminaries and Problem Statement</title>
<p>In this section, we introduce key definitions of AutoFTP and the problem statement.</p>
<sec id="s2-1">
<title>2.1 Particle Swarm Optimization</title>
<p>PSO is a heuristic optimization algorithm that finds an optimal solution in a dynamic environment, by imitating the social activity of a flock of birds. <xref ref-type="fig" rid="F3">Figure&#x20;3</xref> shows the origin of PSO. A flock of eagles wants to capture a rabbit. To achieve the goal, all eagles exchange information related to the position of the rabbit. Each eagle updates its position based on its current status, velocity, the position where it knew is closest to the rabbit, and the position where the flock knew is closest to the rabbit, until the rabbit is captured.</p>
<fig id="F3" position="float">
<label>FIGURE 3</label>
<caption>
<p>The origin of PSO: a flock of eagles is preying on a rabbit. To capturing the rabbit quickly, each eagle records where is the closest position to the rabbit during its history exploration (pBest). Meanwhile, all eagles share the closest position to the rabbit where the flock knew (gBest). All eagles explore any position based on their velocity, pBest, and gBest until the rabbit is captured.</p>
</caption>
<graphic xlink:href="fdata-04-762899-g003.tif"/>
</fig>
<p>Similarly, solving the feature-topic pairing problem can be analogized as a task of searching the optimal matching solutions in a dynamic environment. Specifically, we can view the eagles as a set of binary topic selector, which are to select the optimized subset of topics from a candidate topic set for feature-topic pairing. The choices of these binary topic selectors are iteratively updated in order to converge into the ultimate most matched topic-feature pairs. During the iterative process, all the binary topic selectors jointly share the changes of objective function losses (i.e.,&#x20;the losses of representation construction, feature-topic alignment, and downstream predictive task), so the topic selector knows how to update the topic selection next&#x20;round.</p>
</sec>
<sec id="s2-2">
<title>2.2 Definitions</title>
<p>Definition 1: Spatial Entity. A spatial entity is a geographical concept that consists of a range (e.g. a circle area with a radius of 1 mile) and a location (i.e. the latitude and longitude of a center). The spatial entity also includes various Points-of-Interest (POIs) of different categories (e.g., buildings of education, shopping, medical, banking,&#x20;etc.).</p>
<p>Definition 2: Point-wise Alignment. To tackle feature-topic pairing, we assume there are 1) an embedding vector that describes the features of a spatial entity and 2) a corresponding topic distribution associated to a spatial entity, which are extracted by optimization. To achieve feature-topic alignment, we propose a point-wise alignment to describe the correlation between features and topics. <xref ref-type="fig" rid="F4">Figure&#x20;4A</xref> shows an example of point-wise alignment, we expect to maximize the correlation between the selected topic vector and the spatial embedding vector. The larger the correlation between the two vectors is, the larger the similarity between the two vectors larger&#x20;is.</p>
<fig id="F4" position="float">
<label>FIGURE 4</label>
<caption>
<p>Two kinds of space alignment.</p>
</caption>
<graphic xlink:href="fdata-04-762899-g004.tif"/>
</fig>
<p>Definition 3: Pair-wise Alignment. We propose another perspective (pair-wise) to model the feature-topic alignment. For each entity-entity pair, we compute their feature-feature similarity and topic-topic similarity, and obtain: 1) a topic-topic similarity matrix <bold>S</bold>; 2) a feature-feature similarity matrix <bold>S</bold>&#x2032;. We measure the consistency between the two matrices as the pairwise alignment.</p>
<p>
<xref ref-type="fig" rid="F4">Figure&#x20;4B</xref> shows an example of pair-wise alignment, we aim to let the topic-topic similarity matrix S as close as the feature-feature similarity matrix S&#x2032; possible.</p>
</sec>
<sec id="s2-3">
<title>2.3 The Feature-topic Pairing Problem</title>
<p>The feature-topic pairing problem aims to pair the latent features extracted by representation learning, with the explicit topics of texts of a spatial entity. Formally, given a set of <italic>N</italic> spatial entities, the <italic>n</italic>-th entity is described by multiple graphs (e.g., a POI-POI distance graph <inline-formula id="inf1">
<mml:math id="m1">
<mml:msubsup>
<mml:mrow>
<mml:mi>G</mml:mi>
</mml:mrow>
<mml:mrow>
<mml:mi>n</mml:mi>
</mml:mrow>
<mml:mrow>
<mml:mi>d</mml:mi>
</mml:mrow>
</mml:msubsup>
</mml:math>
</inline-formula> and a POI mobility connectivity <inline-formula id="inf2">
<mml:math id="m2">
<mml:msubsup>
<mml:mrow>
<mml:mi>G</mml:mi>
</mml:mrow>
<mml:mrow>
<mml:mi>n</mml:mi>
</mml:mrow>
<mml:mrow>
<mml:mi>m</mml:mi>
</mml:mrow>
</mml:msubsup>
</mml:math>
</inline-formula>, defined in <xref ref-type="sec" rid="s3-3">Section 3.3</xref>) and a topic distribution <bold>t</bold>
<sub>
<italic>n</italic>
</sub> extracted from textual descriptions <inline-formula id="inf3">
<mml:math id="m3">
<mml:msub>
<mml:mrow>
<mml:mi mathvariant="script">E</mml:mi>
</mml:mrow>
<mml:mrow>
<mml:mi>n</mml:mi>
</mml:mrow>
</mml:msub>
</mml:math>
</inline-formula>. Let <inline-formula id="inf4">
<mml:math id="m4">
<mml:msub>
<mml:mrow>
<mml:mover accent="true">
<mml:mrow>
<mml:mi mathvariant="bold">r</mml:mi>
</mml:mrow>
<mml:mo>&#x303;</mml:mo>
</mml:mover>
</mml:mrow>
<mml:mrow>
<mml:mi>n</mml:mi>
</mml:mrow>
</mml:msub>
</mml:math>
</inline-formula> be the embedding vector of the n-th entity. The objective is to optimize a function that measures representation loss and feature-topic alignment:<disp-formula id="e1">
<mml:math id="m5">
<mml:mrow>
<mml:mover accent="true">
<mml:mrow>
<mml:mi mathvariant="bold">R</mml:mi>
</mml:mrow>
<mml:mo>&#x303;</mml:mo>
</mml:mover>
</mml:mrow>
<mml:mo>&#x3d;</mml:mo>
<mml:msub>
<mml:mrow>
<mml:mi mathvariant="bold">a</mml:mi>
<mml:mi mathvariant="bold">r</mml:mi>
<mml:mi mathvariant="bold">g</mml:mi>
<mml:mi mathvariant="bold">m</mml:mi>
<mml:mi mathvariant="bold">a</mml:mi>
<mml:mi mathvariant="bold">x</mml:mi>
</mml:mrow>
<mml:mrow>
<mml:msub>
<mml:mrow>
<mml:mover accent="true">
<mml:mrow>
<mml:mi mathvariant="bold">r</mml:mi>
</mml:mrow>
<mml:mo>&#x303;</mml:mo>
</mml:mover>
</mml:mrow>
<mml:mrow>
<mml:mi>n</mml:mi>
</mml:mrow>
</mml:msub>
</mml:mrow>
</mml:msub>
<mml:munderover accentunder="false" accent="false">
<mml:mrow>
<mml:mo>&#x2211;</mml:mo>
</mml:mrow>
<mml:mrow>
<mml:mi>n</mml:mi>
<mml:mo>&#x3d;</mml:mo>
<mml:mn>1</mml:mn>
</mml:mrow>
<mml:mrow>
<mml:mi>N</mml:mi>
</mml:mrow>
</mml:munderover>
<mml:mi>f</mml:mi>
<mml:mfenced open="(" close=")">
<mml:mrow>
<mml:msub>
<mml:mrow>
<mml:mover accent="true">
<mml:mrow>
<mml:mi mathvariant="bold">r</mml:mi>
</mml:mrow>
<mml:mo>&#x303;</mml:mo>
</mml:mover>
</mml:mrow>
<mml:mrow>
<mml:mi>n</mml:mi>
</mml:mrow>
</mml:msub>
<mml:mo stretchy="false">&#x7c;</mml:mo>
<mml:msub>
<mml:mrow>
<mml:mi mathvariant="bold">t</mml:mi>
</mml:mrow>
<mml:mrow>
<mml:mi>n</mml:mi>
</mml:mrow>
</mml:msub>
<mml:mo>,</mml:mo>
<mml:msubsup>
<mml:mrow>
<mml:mi>G</mml:mi>
</mml:mrow>
<mml:mrow>
<mml:mi>n</mml:mi>
</mml:mrow>
<mml:mrow>
<mml:mi>d</mml:mi>
</mml:mrow>
</mml:msubsup>
<mml:mo>,</mml:mo>
<mml:msubsup>
<mml:mrow>
<mml:mi>G</mml:mi>
</mml:mrow>
<mml:mrow>
<mml:mi>n</mml:mi>
</mml:mrow>
<mml:mrow>
<mml:mi>m</mml:mi>
</mml:mrow>
</mml:msubsup>
<mml:mo>,</mml:mo>
<mml:mi>K</mml:mi>
</mml:mrow>
</mml:mfenced>
<mml:mo>,</mml:mo>
</mml:math>
<label>(1)</label>
</disp-formula>where <inline-formula id="inf5">
<mml:math id="m6">
<mml:mrow>
<mml:mover accent="true">
<mml:mrow>
<mml:mi mathvariant="bold">R</mml:mi>
</mml:mrow>
<mml:mo>&#x303;</mml:mo>
</mml:mover>
</mml:mrow>
<mml:mo>&#x3d;</mml:mo>
<mml:msubsup>
<mml:mrow>
<mml:mrow>
<mml:mo stretchy="false">{</mml:mo>
<mml:mrow>
<mml:msub>
<mml:mrow>
<mml:mover accent="true">
<mml:mrow>
<mml:mi mathvariant="bold">r</mml:mi>
</mml:mrow>
<mml:mo>&#x303;</mml:mo>
</mml:mover>
</mml:mrow>
<mml:mrow>
<mml:mi>n</mml:mi>
</mml:mrow>
</mml:msub>
</mml:mrow>
<mml:mo stretchy="false">}</mml:mo>
</mml:mrow>
</mml:mrow>
<mml:mrow>
<mml:mi>n</mml:mi>
<mml:mo>&#x3d;</mml:mo>
<mml:mn>1</mml:mn>
</mml:mrow>
<mml:mrow>
<mml:mi>N</mml:mi>
</mml:mrow>
</mml:msubsup>
<mml:mo>&#x2208;</mml:mo>
<mml:msup>
<mml:mrow>
<mml:mi mathvariant="double-struck">R</mml:mi>
</mml:mrow>
<mml:mrow>
<mml:mi>N</mml:mi>
<mml:mo>&#xd7;</mml:mo>
<mml:mi>K</mml:mi>
</mml:mrow>
</mml:msup>
</mml:math>
</inline-formula> are the embeddings of all spatial entities, <italic>K</italic> is the number features of an embedding vector.</p>
</sec>
</sec>
<sec id="s3">
<title>3 The Proposed Method-AutoFTP</title>
<p>In this section, we first introduce an overview of our AutoFTP framework, then present its technical details.</p>
<sec id="s3-1">
<title>3.1 Framework Overview</title>
<p>
<xref ref-type="fig" rid="F5">Figure&#x20;5</xref> shows our proposed framework. First, we construct a semantic topic space by extracting topic distribution from the corresponding texts of spatial entities. Then, we initialize a feature embedding space based on the geographical structures of spatial entities. Next, we utilize a PSO-based topic selector to select the optimal K topics for pairing with the spatial embeddings coming from the feature embedding space. During the pairing process, the losses of spatial representation learner, point-wise alignment, pair-wise alignment, and downstream tasks are regarded as feedback to update the topic selector for the next pairing iteration. With the development of the learning iteration, the feature embedding space aligns to the topic semantic space gradually. Finally, the learned spatial embeddings of AutoFTP are effective and semantically rich. Here, to validate the effectiveness of AutoFTP, we apply the framework to predict the real estate price (downstream tasks) of the residential communities (spatial entities) based on spatial embeddings of the communities. The more accurate the prediction is, the more effective the learned embedding is. In addition, the AutoFTP can be generalized to other spatial representation learning problems with graphs and&#x20;texts.</p>
<fig id="F5" position="float">
<label>FIGURE 5</label>
<caption>
<p>An overview of AutoFTP. In the framework, we first construct a topic semantic space based on the texts of spatial entities. Then, we initialize a embedding feature space based on the geographical structures of spatial entities. Later, we employ a PSO-based framework to conduct feature-topic pairing through jointly optimizing representation learning, point-wise alignment, pair-wise alignment, and downstream task over learning iterations.</p>
</caption>
<graphic xlink:href="fdata-04-762899-g005.tif"/>
</fig>
</sec>
<sec id="s3-2">
<title>3.2 Textual Topic Extraction</title>
<p>To derive the textual semantic topic space, we extract the topic distributions of spatial entities from texts generated by location based social networks. Traditional topic models, such as LDA (<xref ref-type="bibr" rid="B3">Blei et&#x20;al., 2003</xref>), PSLA (<xref ref-type="bibr" rid="B13">Hofmann, 2013</xref>), are implemented based on bag-of-words. These methods ignore word orders in sentences. To improve the performances of topic modeling, we employ a pre-trained deep word embedding model (<xref ref-type="bibr" rid="B12">He, 2014</xref>) to generate topics.</p>
<p>As illustrated in <xref ref-type="fig" rid="F6">Figure&#x20;6</xref>, we first collect the text descriptions of all entities. Besides, we extract keywords from texts using the TextRank algorithm (<xref ref-type="bibr" rid="B26">Mihalcea and Tarau, 2004</xref>) and leverage a pre-trained language model (<xref ref-type="bibr" rid="B12">He, 2014</xref>) to learn the corresponding word embedding of each keyword. Moreover, we exploit a Gaussian Mixture Model (GMM) to cluster the keyword embeddings into <italic>T</italic> topics. The clustering model provides a topic label for each keyword. To explain the labeling process, we take the <italic>i</italic>-th keyword&#x2019;s embedding vector <bold>x</bold>
<sub>
<italic>i</italic>
</sub> as an example. First, we assume that the <italic>T</italic> topics obey a Gaussian Mixture Distribution (GMD). Then we randomly initialize the parameters of GMD. Next, we use the Expectation Maximization (EM) algorithm to find the optimal parameters of the GMD. Finally, we calculate the probability of <bold>x</bold>
<sub>
<italic>i</italic>
</sub> (a.k.a., membership), belonging to each topic based on the GMD, and select the topic with the highest probability as the label of <bold>x</bold>
<sub>
<italic>i</italic>
</sub>. After that, we propose to construct the topic distribution vector of each spatial entity. In particular, for the <italic>n</italic>-th entity, the topic vector <bold>t</bold>
<sub>
<italic>n</italic>
</sub> is a <italic>T</italic> dimensional vector, where each dimension indicates a topic, and is filled by the number of associated keywords.</p>
<fig id="F6" position="float">
<label>FIGURE 6</label>
<caption>
<p>Obtaining the topic vectors of spatial entities.</p>
</caption>
<graphic xlink:href="fdata-04-762899-g006.tif"/>
</fig>
</sec>
<sec id="s3-3">
<title>3.3 Graph Extraction of Spatial Entities</title>
<p>In order to learn the embedding feature vectors of spatial entities, we propose to construct the graph-structured topology of each spatial entity. This is because there is inherent spatial autocorrelation between each two spatial entities, according to the geographical first law. We describe a spatial entity in terms of its POIs, by building two graphs. 1) POI-POI distance graph: denoted by <italic>G</italic>
<sup>
<italic>d</italic>
</sup>, where POI categories are nodes and the average distances between POI categories are edge weights. 2) <bold>POI-</bold>POI mobility graph: denoted by <italic>G</italic>
<sup>
<italic>m</italic>
</sup>, where nodes are POI categories, and edge weights are human mobility connectivity. The number of POI categories in this paper is <italic>M</italic>, and the two graphs are extracted via the method in (<xref ref-type="bibr" rid="B44">Wang et&#x20;al., 2018a</xref>). Specifically, we first use a parametric function to estimate POI visit probability based on a taxi GPS trace data: <inline-formula id="inf6">
<mml:math id="m7">
<mml:mi>P</mml:mi>
<mml:mrow>
<mml:mo stretchy="false">(</mml:mo>
<mml:mrow>
<mml:mi>&#x3c2;</mml:mi>
</mml:mrow>
<mml:mo stretchy="false">)</mml:mo>
</mml:mrow>
<mml:mo>&#x3d;</mml:mo>
<mml:mfrac>
<mml:mrow>
<mml:msub>
<mml:mrow>
<mml:mi>&#x3b2;</mml:mi>
</mml:mrow>
<mml:mrow>
<mml:mn>1</mml:mn>
</mml:mrow>
</mml:msub>
</mml:mrow>
<mml:mrow>
<mml:msub>
<mml:mrow>
<mml:mi>&#x3b2;</mml:mi>
</mml:mrow>
<mml:mrow>
<mml:mn>2</mml:mn>
</mml:mrow>
</mml:msub>
</mml:mrow>
</mml:mfrac>
<mml:mo>&#x22c5;</mml:mo>
<mml:mi>&#x3c2;</mml:mi>
<mml:mo>&#x22c5;</mml:mo>
<mml:mi>e</mml:mi>
<mml:mi>x</mml:mi>
<mml:mi>p</mml:mi>
<mml:mrow>
<mml:mo stretchy="false">(</mml:mo>
<mml:mrow>
<mml:mn>1</mml:mn>
<mml:mo>&#x2212;</mml:mo>
<mml:mfrac>
<mml:mrow>
<mml:mi>&#x3c2;</mml:mi>
</mml:mrow>
<mml:mrow>
<mml:msub>
<mml:mrow>
<mml:mi>&#x3b2;</mml:mi>
</mml:mrow>
<mml:mrow>
<mml:mn>2</mml:mn>
</mml:mrow>
</mml:msub>
</mml:mrow>
</mml:mfrac>
</mml:mrow>
<mml:mo stretchy="false">)</mml:mo>
</mml:mrow>
<mml:mo>,</mml:mo>
</mml:math>
</inline-formula> where <italic>&#x3c2;</italic> denotes the distance between a POI and a drop-off position in a taxi trace, <italic>&#x3b2;</italic>
<sub>1</sub> &#x3d;&#x2009; max<sub>
<italic>&#x3c2;</italic>
</sub>
<italic>P</italic>(<italic>&#x3c2;</italic>), and <italic>&#x3b2;</italic>
<sub>2</sub> &#x3d; arg&#x2009;max<sub>
<italic>&#x3c2;</italic>
</sub>
<italic>P</italic>(<italic>&#x3c2;</italic>). We calculate the visited probability of all POIs according to the formula. We sum up the probability of POIs belonging to the same POI category to calculate the visited probability of the POI category. Finally, we calculate the connectivity strength between POI categories as: <inline-formula id="inf7">
<mml:math id="m8">
<mml:msub>
<mml:mrow>
<mml:mi>C</mml:mi>
</mml:mrow>
<mml:mrow>
<mml:mover accent="true">
<mml:mrow>
<mml:mi>i</mml:mi>
<mml:mi>j</mml:mi>
</mml:mrow>
<mml:mo>&#x20d7;</mml:mo>
</mml:mover>
</mml:mrow>
</mml:msub>
<mml:mo>&#x3d;</mml:mo>
<mml:mfenced open="{" close="">
<mml:mrow>
<mml:mtable class="cases">
<mml:mtr>
<mml:mtd columnalign="left">
<mml:msub>
<mml:mrow>
<mml:mi>P</mml:mi>
</mml:mrow>
<mml:mrow>
<mml:mi>i</mml:mi>
</mml:mrow>
</mml:msub>
<mml:mo>&#x22c5;</mml:mo>
<mml:msub>
<mml:mrow>
<mml:mi>P</mml:mi>
</mml:mrow>
<mml:mrow>
<mml:mi>j</mml:mi>
</mml:mrow>
</mml:msub>
<mml:mo>,</mml:mo>
<mml:mspace width="1em"/>
<mml:mi>i</mml:mi>
<mml:mi>f</mml:mi>
<mml:mspace width="0.17em"/>
<mml:mspace width="0.3333em"/>
<mml:mi>i</mml:mi>
<mml:mo>&#x2260;</mml:mo>
<mml:mi>j</mml:mi>
<mml:mspace width="1em"/>
</mml:mtd>
</mml:mtr>
<mml:mtr>
<mml:mtd columnalign="left">
<mml:mn>0</mml:mn>
<mml:mo>,</mml:mo>
<mml:mspace width="1em"/>
<mml:mspace width="1em"/>
<mml:mspace width="1em"/>
<mml:mi>o</mml:mi>
<mml:mi>t</mml:mi>
<mml:mi>h</mml:mi>
<mml:mi>e</mml:mi>
<mml:mi>r</mml:mi>
<mml:mi>w</mml:mi>
<mml:mi>i</mml:mi>
<mml:mi>s</mml:mi>
<mml:mi>e</mml:mi>
<mml:mo>,</mml:mo>
<mml:mspace width="1em"/>
</mml:mtd>
</mml:mtr>
</mml:mtable>
</mml:mrow>
</mml:mfenced>
</mml:math>
</inline-formula> where <italic>P</italic>
<sub>
<italic>i</italic>
</sub> and <italic>P</italic>
<sub>
<italic>j</italic>
</sub> represent the visited probability of POI category <italic>i</italic> and POI category <italic>j</italic> respectively; <inline-formula id="inf8">
<mml:math id="m9">
<mml:msub>
<mml:mrow>
<mml:mi>C</mml:mi>
</mml:mrow>
<mml:mrow>
<mml:mover accent="true">
<mml:mrow>
<mml:mi>i</mml:mi>
<mml:mi>j</mml:mi>
</mml:mrow>
<mml:mo>&#x20d7;</mml:mo>
</mml:mover>
</mml:mrow>
</mml:msub>
</mml:math>
</inline-formula> indicates the connectivity between POI category <italic>i</italic> and&#x20;<italic>j</italic>.</p>
</sec>
<sec id="s3-4">
<title>3.4 Particle Swarm Optimization Based Feature-Topic Pairing</title>
<sec id="s3-4-1">
<title>3.4.1 Spatial Representation Learner</title>
<p>To learn the representations of spatial entities, we utilize the Graph Auto Encoder (GAE) (<xref ref-type="bibr" rid="B20">Kipf and Welling, 2016</xref>) to construct latent embedding space. Specifically, to learn the embedding feature vector of the <italic>n</italic>-th entity, the encoder has two GCN layers. The encoding calculation process can be formulated as follows:<disp-formula id="e2">
<mml:math id="m10">
<mml:mfenced open="{" close="">
<mml:mrow>
<mml:mtable class="cases">
<mml:mtr>
<mml:mtd columnalign="left">
<mml:msub>
<mml:mrow>
<mml:mover accent="true">
<mml:mrow>
<mml:mi mathvariant="bold">A</mml:mi>
</mml:mrow>
<mml:mo stretchy="false">&#x302;</mml:mo>
</mml:mover>
</mml:mrow>
<mml:mrow>
<mml:mi>n</mml:mi>
</mml:mrow>
</mml:msub>
<mml:mo>&#x3d;</mml:mo>
<mml:msub>
<mml:mrow>
<mml:mi mathvariant="bold">A</mml:mi>
</mml:mrow>
<mml:mrow>
<mml:mi>n</mml:mi>
</mml:mrow>
</mml:msub>
<mml:mo>&#x2b;</mml:mo>
<mml:msub>
<mml:mrow>
<mml:mi mathvariant="bold">I</mml:mi>
</mml:mrow>
<mml:mrow>
<mml:mi>n</mml:mi>
</mml:mrow>
</mml:msub>
<mml:mo>,</mml:mo>
<mml:mspace width="1em"/>
</mml:mtd>
</mml:mtr>
<mml:mtr>
<mml:mtd columnalign="left">
<mml:msub>
<mml:mrow>
<mml:mover accent="true">
<mml:mrow>
<mml:mi mathvariant="bold">A</mml:mi>
</mml:mrow>
<mml:mo>&#x303;</mml:mo>
</mml:mover>
</mml:mrow>
<mml:mrow>
<mml:mi>n</mml:mi>
</mml:mrow>
</mml:msub>
<mml:mo>&#x3d;</mml:mo>
<mml:msubsup>
<mml:mrow>
<mml:mover accent="true">
<mml:mrow>
<mml:mi mathvariant="bold-italic">D</mml:mi>
</mml:mrow>
<mml:mo stretchy="false">&#x302;</mml:mo>
</mml:mover>
</mml:mrow>
<mml:mrow>
<mml:mi>n</mml:mi>
</mml:mrow>
<mml:mrow>
<mml:mo>&#x2212;</mml:mo>
<mml:mfrac>
<mml:mn>1</mml:mn>
<mml:mn>2</mml:mn>
</mml:mfrac>
</mml:mrow>
</mml:msubsup>
<mml:msub>
<mml:mrow>
<mml:mover accent="true">
<mml:mrow>
<mml:mi mathvariant="bold-italic">A</mml:mi>
</mml:mrow>
<mml:mo stretchy="false">&#x302;</mml:mo>
</mml:mover>
</mml:mrow>
<mml:mrow>
<mml:mi>n</mml:mi>
</mml:mrow>
</mml:msub>
<mml:msubsup>
<mml:mrow>
<mml:mover accent="true">
<mml:mrow>
<mml:mi mathvariant="bold-italic">D</mml:mi>
</mml:mrow>
<mml:mo stretchy="false">&#x302;</mml:mo>
</mml:mover>
</mml:mrow>
<mml:mrow>
<mml:mi>n</mml:mi>
</mml:mrow>
<mml:mrow>
<mml:mo>&#x2212;</mml:mo>
<mml:mfrac>
<mml:mrow>
<mml:mn>1</mml:mn>
</mml:mrow>
<mml:mrow>
<mml:mn>2</mml:mn>
</mml:mrow>
</mml:mfrac>
</mml:mrow>
</mml:msubsup>
<mml:mo>,</mml:mo>
<mml:mspace width="1em"/>
</mml:mtd>
</mml:mtr>
<mml:mtr>
<mml:mtd columnalign="left">
<mml:msub>
<mml:mrow>
<mml:mi mathvariant="bold">z</mml:mi>
</mml:mrow>
<mml:mrow>
<mml:mi>n</mml:mi>
</mml:mrow>
</mml:msub>
<mml:mo>&#x3d;</mml:mo>
<mml:msub>
<mml:mrow>
<mml:mover accent="true">
<mml:mrow>
<mml:mi mathvariant="bold">A</mml:mi>
</mml:mrow>
<mml:mo>&#x303;</mml:mo>
</mml:mover>
</mml:mrow>
<mml:mrow>
<mml:mi>n</mml:mi>
</mml:mrow>
</mml:msub>
<mml:mtext>Relu</mml:mtext>
<mml:mfenced open="(" close=")">
<mml:mrow>
<mml:msub>
<mml:mrow>
<mml:mover accent="true">
<mml:mrow>
<mml:mi mathvariant="bold">A</mml:mi>
</mml:mrow>
<mml:mo>&#x303;</mml:mo>
</mml:mover>
</mml:mrow>
<mml:mrow>
<mml:mi>n</mml:mi>
</mml:mrow>
</mml:msub>
<mml:msub>
<mml:mrow>
<mml:mi mathvariant="bold">X</mml:mi>
</mml:mrow>
<mml:mrow>
<mml:mi>n</mml:mi>
</mml:mrow>
</mml:msub>
<mml:msubsup>
<mml:mrow>
<mml:mi mathvariant="bold">W</mml:mi>
</mml:mrow>
<mml:mrow>
<mml:mi>n</mml:mi>
</mml:mrow>
<mml:mrow>
<mml:mfenced open="(" close=")">
<mml:mrow>
<mml:mn>0</mml:mn>
</mml:mrow>
</mml:mfenced>
</mml:mrow>
</mml:msubsup>
</mml:mrow>
</mml:mfenced>
<mml:msubsup>
<mml:mrow>
<mml:mi mathvariant="bold">W</mml:mi>
</mml:mrow>
<mml:mrow>
<mml:mi>n</mml:mi>
</mml:mrow>
<mml:mrow>
<mml:mfenced open="(" close=")">
<mml:mrow>
<mml:mn>1</mml:mn>
</mml:mrow>
</mml:mfenced>
</mml:mrow>
</mml:msubsup>
<mml:mspace width="1em"/>
</mml:mtd>
</mml:mtr>
</mml:mtable>
</mml:mrow>
</mml:mfenced>
</mml:math>
<label>(2)</label>
</disp-formula>where <inline-formula id="inf9">
<mml:math id="m11">
<mml:msub>
<mml:mrow>
<mml:mi mathvariant="bold">A</mml:mi>
</mml:mrow>
<mml:mrow>
<mml:mi>n</mml:mi>
</mml:mrow>
</mml:msub>
<mml:mo>,</mml:mo>
<mml:msub>
<mml:mrow>
<mml:mi mathvariant="bold">I</mml:mi>
</mml:mrow>
<mml:mrow>
<mml:mi>n</mml:mi>
</mml:mrow>
</mml:msub>
<mml:mo>,</mml:mo>
<mml:msub>
<mml:mrow>
<mml:mover accent="true">
<mml:mrow>
<mml:mi mathvariant="bold-italic">A</mml:mi>
</mml:mrow>
<mml:mo>&#x303;</mml:mo>
</mml:mover>
</mml:mrow>
<mml:mrow>
<mml:mi>n</mml:mi>
</mml:mrow>
</mml:msub>
<mml:mo>,</mml:mo>
<mml:msub>
<mml:mrow>
<mml:mover accent="true">
<mml:mrow>
<mml:mi mathvariant="bold-italic">D</mml:mi>
</mml:mrow>
<mml:mo stretchy="false">&#x302;</mml:mo>
</mml:mover>
</mml:mrow>
<mml:mrow>
<mml:mi>n</mml:mi>
</mml:mrow>
</mml:msub>
</mml:math>
</inline-formula> own the same shape <inline-formula id="inf10">
<mml:math id="m12">
<mml:msup>
<mml:mrow>
<mml:mi mathvariant="double-struck">R</mml:mi>
</mml:mrow>
<mml:mrow>
<mml:mi>M</mml:mi>
<mml:mo>&#xd7;</mml:mo>
<mml:mi>M</mml:mi>
</mml:mrow>
</mml:msup>
</mml:math>
</inline-formula>. Moreover, <bold>A</bold>
<sub>
<italic>n</italic>
</sub> is the adjacency matrix, <bold>I</bold>
<sub>
<italic>n</italic>
</sub> is the identity matrix, <inline-formula id="inf11">
<mml:math id="m13">
<mml:msub>
<mml:mrow>
<mml:mover accent="true">
<mml:mrow>
<mml:mi mathvariant="bold-italic">A</mml:mi>
</mml:mrow>
<mml:mo>&#x303;</mml:mo>
</mml:mover>
</mml:mrow>
<mml:mrow>
<mml:mi>n</mml:mi>
</mml:mrow>
</mml:msub>
</mml:math>
</inline-formula> is the symmetrically normalized adjacency matrix, <inline-formula id="inf12">
<mml:math id="m14">
<mml:msub>
<mml:mrow>
<mml:mover accent="true">
<mml:mrow>
<mml:mi mathvariant="bold-italic">D</mml:mi>
</mml:mrow>
<mml:mo stretchy="false">&#x302;</mml:mo>
</mml:mover>
</mml:mrow>
<mml:mrow>
<mml:mi>n</mml:mi>
</mml:mrow>
</mml:msub>
</mml:math>
</inline-formula> is the degree matrix. In addition, <inline-formula id="inf13">
<mml:math id="m15">
<mml:msub>
<mml:mrow>
<mml:mi mathvariant="bold">X</mml:mi>
</mml:mrow>
<mml:mrow>
<mml:mi>n</mml:mi>
</mml:mrow>
</mml:msub>
<mml:mo>&#x2208;</mml:mo>
<mml:msup>
<mml:mrow>
<mml:mi mathvariant="double-struck">R</mml:mi>
</mml:mrow>
<mml:mrow>
<mml:mi>M</mml:mi>
<mml:mo>&#xd7;</mml:mo>
<mml:mi>U</mml:mi>
</mml:mrow>
</mml:msup>
</mml:math>
</inline-formula> is the feature matrix of the graph, in which <italic>U</italic> is the feature dimension; <inline-formula id="inf14">
<mml:math id="m16">
<mml:msubsup>
<mml:mrow>
<mml:mi mathvariant="bold">W</mml:mi>
</mml:mrow>
<mml:mrow>
<mml:mi>n</mml:mi>
</mml:mrow>
<mml:mrow>
<mml:mrow>
<mml:mo stretchy="false">(</mml:mo>
<mml:mrow>
<mml:mn>1</mml:mn>
</mml:mrow>
<mml:mo stretchy="false">)</mml:mo>
</mml:mrow>
</mml:mrow>
</mml:msubsup>
<mml:mo>&#x2208;</mml:mo>
<mml:msup>
<mml:mrow>
<mml:mi mathvariant="double-struck">R</mml:mi>
</mml:mrow>
<mml:mrow>
<mml:mi>U</mml:mi>
<mml:mo>&#xd7;</mml:mo>
<mml:mi>H</mml:mi>
</mml:mrow>
</mml:msup>
</mml:math>
</inline-formula> is the weight matrix of the first GCN layer, in which <italic>H</italic> is the output dimension of the layer; <inline-formula id="inf15">
<mml:math id="m17">
<mml:msubsup>
<mml:mrow>
<mml:mi mathvariant="bold">W</mml:mi>
</mml:mrow>
<mml:mrow>
<mml:mi>n</mml:mi>
</mml:mrow>
<mml:mrow>
<mml:mrow>
<mml:mo stretchy="false">(</mml:mo>
<mml:mrow>
<mml:mn>2</mml:mn>
</mml:mrow>
<mml:mo stretchy="false">)</mml:mo>
</mml:mrow>
</mml:mrow>
</mml:msubsup>
<mml:mo>&#x2208;</mml:mo>
<mml:msup>
<mml:mrow>
<mml:mi mathvariant="double-struck">R</mml:mi>
</mml:mrow>
<mml:mrow>
<mml:mi>H</mml:mi>
<mml:mo>&#xd7;</mml:mo>
<mml:mi>K</mml:mi>
</mml:mrow>
</mml:msup>
</mml:math>
</inline-formula> is the weight matrix of the second GCN layer; <inline-formula id="inf16">
<mml:math id="m18">
<mml:msub>
<mml:mrow>
<mml:mi mathvariant="bold">z</mml:mi>
</mml:mrow>
<mml:mrow>
<mml:mi>n</mml:mi>
</mml:mrow>
</mml:msub>
<mml:mo>&#x2208;</mml:mo>
<mml:msup>
<mml:mrow>
<mml:mi mathvariant="double-struck">R</mml:mi>
</mml:mrow>
<mml:mrow>
<mml:mi>M</mml:mi>
<mml:mo>&#xd7;</mml:mo>
<mml:mi>K</mml:mi>
</mml:mrow>
</mml:msup>
</mml:math>
</inline-formula> is the output embedding of the encoder. The decoder recovers the adjacency matrix according to <bold>z</bold>
<sub>
<italic>n</italic>
</sub>:<disp-formula id="e3">
<mml:math id="m19">
<mml:msup>
<mml:mrow>
<mml:mover accent="true">
<mml:mrow>
<mml:mi mathvariant="bold">A</mml:mi>
</mml:mrow>
<mml:mo stretchy="false">&#x302;</mml:mo>
</mml:mover>
</mml:mrow>
<mml:mrow>
<mml:mo>&#x2217;</mml:mo>
</mml:mrow>
</mml:msup>
<mml:mo>&#x3d;</mml:mo>
<mml:mi>&#x3c3;</mml:mi>
<mml:mfenced open="(" close=")">
<mml:mrow>
<mml:msub>
<mml:mrow>
<mml:mi mathvariant="bold">z</mml:mi>
</mml:mrow>
<mml:mrow>
<mml:mi>n</mml:mi>
</mml:mrow>
</mml:msub>
<mml:msubsup>
<mml:mrow>
<mml:mi mathvariant="bold">z</mml:mi>
</mml:mrow>
<mml:mrow>
<mml:mi>n</mml:mi>
</mml:mrow>
<mml:mrow>
<mml:mo>&#x2032;</mml:mo>
</mml:mrow>
</mml:msubsup>
</mml:mrow>
</mml:mfenced>
<mml:mo>.</mml:mo>
</mml:math>
<label>(3)</label>
</disp-formula>
</p>
<p>The optimization objective is to minimize the reconstruction loss between the original graph, denoted by the adjacency matrix <inline-formula id="inf17">
<mml:math id="m20">
<mml:msub>
<mml:mrow>
<mml:mover accent="true">
<mml:mrow>
<mml:mi mathvariant="bold">A</mml:mi>
</mml:mrow>
<mml:mo stretchy="false">&#x302;</mml:mo>
</mml:mover>
</mml:mrow>
<mml:mrow>
<mml:mi>n</mml:mi>
</mml:mrow>
</mml:msub>
</mml:math>
</inline-formula>, and the reconstructed graph, denoted by the adjacency matrix <inline-formula id="inf18">
<mml:math id="m21">
<mml:msubsup>
<mml:mrow>
<mml:mover accent="true">
<mml:mrow>
<mml:mi mathvariant="bold">A</mml:mi>
</mml:mrow>
<mml:mo stretchy="false">&#x302;</mml:mo>
</mml:mover>
</mml:mrow>
<mml:mrow>
<mml:mi>n</mml:mi>
</mml:mrow>
<mml:mrow>
<mml:mo>&#x2217;</mml:mo>
</mml:mrow>
</mml:msubsup>
</mml:math>
</inline-formula>:<disp-formula id="e4">
<mml:math id="m22">
<mml:msub>
<mml:mrow>
<mml:mi mathvariant="script">L</mml:mi>
</mml:mrow>
<mml:mrow>
<mml:mi>R</mml:mi>
</mml:mrow>
</mml:msub>
<mml:mo>&#x3d;</mml:mo>
<mml:munderover accentunder="false" accent="false">
<mml:mrow>
<mml:mo>&#x2211;</mml:mo>
</mml:mrow>
<mml:mrow>
<mml:mi>n</mml:mi>
<mml:mo>&#x3d;</mml:mo>
<mml:mn>1</mml:mn>
</mml:mrow>
<mml:mrow>
<mml:mi>N</mml:mi>
</mml:mrow>
</mml:munderover>
<mml:mo stretchy="false">&#x2016;</mml:mo>
<mml:mfenced open="(" close=")">
<mml:mrow>
<mml:msub>
<mml:mrow>
<mml:mover accent="true">
<mml:mrow>
<mml:mi mathvariant="bold">A</mml:mi>
</mml:mrow>
<mml:mo stretchy="false">&#x302;</mml:mo>
</mml:mover>
</mml:mrow>
<mml:mrow>
<mml:mi>n</mml:mi>
</mml:mrow>
</mml:msub>
<mml:mo>&#x2212;</mml:mo>
<mml:msubsup>
<mml:mrow>
<mml:mover accent="true">
<mml:mrow>
<mml:mi mathvariant="bold">A</mml:mi>
</mml:mrow>
<mml:mo stretchy="false">&#x302;</mml:mo>
</mml:mover>
</mml:mrow>
<mml:mrow>
<mml:mi>n</mml:mi>
</mml:mrow>
<mml:mrow>
<mml:mo>&#x2217;</mml:mo>
</mml:mrow>
</mml:msubsup>
</mml:mrow>
</mml:mfenced>
<mml:msup>
<mml:mrow>
<mml:mo stretchy="false">&#x2016;</mml:mo>
</mml:mrow>
<mml:mrow>
<mml:mn>2</mml:mn>
</mml:mrow>
</mml:msup>
</mml:math>
<label>(4)</label>
</disp-formula>
</p>
<p>We apply the GAE to the POI-POI distance graph <inline-formula id="inf19">
<mml:math id="m23">
<mml:msubsup>
<mml:mrow>
<mml:mi>G</mml:mi>
</mml:mrow>
<mml:mrow>
<mml:mi>n</mml:mi>
</mml:mrow>
<mml:mrow>
<mml:mi>d</mml:mi>
</mml:mrow>
</mml:msubsup>
</mml:math>
</inline-formula> and the POI-POI mobility graph <inline-formula id="inf20">
<mml:math id="m24">
<mml:msubsup>
<mml:mrow>
<mml:mi>G</mml:mi>
</mml:mrow>
<mml:mrow>
<mml:mi>n</mml:mi>
</mml:mrow>
<mml:mrow>
<mml:mi>m</mml:mi>
</mml:mrow>
</mml:msubsup>
</mml:math>
</inline-formula> of the <italic>n</italic>-th spatial entity. After that, we obtain the node representations of <inline-formula id="inf21">
<mml:math id="m25">
<mml:msubsup>
<mml:mrow>
<mml:mi>G</mml:mi>
</mml:mrow>
<mml:mrow>
<mml:mi>n</mml:mi>
</mml:mrow>
<mml:mrow>
<mml:mi>d</mml:mi>
</mml:mrow>
</mml:msubsup>
</mml:math>
</inline-formula> and <inline-formula id="inf22">
<mml:math id="m26">
<mml:msubsup>
<mml:mrow>
<mml:mi>G</mml:mi>
</mml:mrow>
<mml:mrow>
<mml:mi>n</mml:mi>
</mml:mrow>
<mml:mrow>
<mml:mi>m</mml:mi>
</mml:mrow>
</mml:msubsup>
</mml:math>
</inline-formula>, denoted by <inline-formula id="inf23">
<mml:math id="m27">
<mml:msub>
<mml:mrow>
<mml:mi mathvariant="bold">z</mml:mi>
</mml:mrow>
<mml:mrow>
<mml:msup>
<mml:mrow>
<mml:mi>n</mml:mi>
</mml:mrow>
<mml:mrow>
<mml:mi>d</mml:mi>
</mml:mrow>
</mml:msup>
</mml:mrow>
</mml:msub>
<mml:mo>&#x2208;</mml:mo>
<mml:msup>
<mml:mrow>
<mml:mi mathvariant="double-struck">R</mml:mi>
</mml:mrow>
<mml:mrow>
<mml:mi>M</mml:mi>
<mml:mo>&#xd7;</mml:mo>
<mml:mi>K</mml:mi>
</mml:mrow>
</mml:msup>
</mml:math>
</inline-formula> and <inline-formula id="inf24">
<mml:math id="m28">
<mml:msub>
<mml:mrow>
<mml:mi mathvariant="bold">z</mml:mi>
</mml:mrow>
<mml:mrow>
<mml:msup>
<mml:mrow>
<mml:mi>n</mml:mi>
</mml:mrow>
<mml:mrow>
<mml:mi>m</mml:mi>
</mml:mrow>
</mml:msup>
</mml:mrow>
</mml:msub>
<mml:mo>&#x2208;</mml:mo>
<mml:msup>
<mml:mrow>
<mml:mi mathvariant="double-struck">R</mml:mi>
</mml:mrow>
<mml:mrow>
<mml:mi>M</mml:mi>
<mml:mo>&#xd7;</mml:mo>
<mml:mi>K</mml:mi>
</mml:mrow>
</mml:msup>
</mml:math>
</inline-formula>. Then, we aggregate <inline-formula id="inf25">
<mml:math id="m29">
<mml:msub>
<mml:mrow>
<mml:mi mathvariant="bold">z</mml:mi>
</mml:mrow>
<mml:mrow>
<mml:msup>
<mml:mrow>
<mml:mi>n</mml:mi>
</mml:mrow>
<mml:mrow>
<mml:mi>d</mml:mi>
</mml:mrow>
</mml:msup>
</mml:mrow>
</mml:msub>
</mml:math>
</inline-formula> and <inline-formula id="inf26">
<mml:math id="m30">
<mml:msub>
<mml:mrow>
<mml:mi mathvariant="bold">z</mml:mi>
</mml:mrow>
<mml:mrow>
<mml:msup>
<mml:mrow>
<mml:mi>n</mml:mi>
</mml:mrow>
<mml:mrow>
<mml:mi>m</mml:mi>
</mml:mrow>
</mml:msup>
</mml:mrow>
</mml:msub>
</mml:math>
</inline-formula> by averaging all node embeddings together to attain the graph embedding of <inline-formula id="inf27">
<mml:math id="m31">
<mml:msubsup>
<mml:mrow>
<mml:mi>G</mml:mi>
</mml:mrow>
<mml:mrow>
<mml:mi>n</mml:mi>
</mml:mrow>
<mml:mrow>
<mml:mi>d</mml:mi>
</mml:mrow>
</mml:msubsup>
</mml:math>
</inline-formula> and <inline-formula id="inf28">
<mml:math id="m32">
<mml:msubsup>
<mml:mrow>
<mml:mi>G</mml:mi>
</mml:mrow>
<mml:mrow>
<mml:mi>n</mml:mi>
</mml:mrow>
<mml:mrow>
<mml:mi>m</mml:mi>
</mml:mrow>
</mml:msubsup>
</mml:math>
</inline-formula> respectively. Finally, we integrate the graph embeddings of <inline-formula id="inf29">
<mml:math id="m33">
<mml:msubsup>
<mml:mrow>
<mml:mi>G</mml:mi>
</mml:mrow>
<mml:mrow>
<mml:mi>n</mml:mi>
</mml:mrow>
<mml:mrow>
<mml:mi>d</mml:mi>
</mml:mrow>
</mml:msubsup>
</mml:math>
</inline-formula> and <inline-formula id="inf30">
<mml:math id="m34">
<mml:msubsup>
<mml:mrow>
<mml:mi>G</mml:mi>
</mml:mrow>
<mml:mrow>
<mml:mi>n</mml:mi>
</mml:mrow>
<mml:mrow>
<mml:mi>m</mml:mi>
</mml:mrow>
</mml:msubsup>
</mml:math>
</inline-formula> into the unified spatial embedding of the entity by averaging calculation, denoted by <inline-formula id="inf31">
<mml:math id="m35">
<mml:msub>
<mml:mrow>
<mml:mi mathvariant="bold">r</mml:mi>
</mml:mrow>
<mml:mrow>
<mml:mi>n</mml:mi>
</mml:mrow>
</mml:msub>
<mml:mo>&#x2208;</mml:mo>
<mml:msup>
<mml:mrow>
<mml:mi mathvariant="double-struck">R</mml:mi>
</mml:mrow>
<mml:mrow>
<mml:mi>K</mml:mi>
</mml:mrow>
</mml:msup>
</mml:math>
</inline-formula>.</p>
</sec>
<sec id="s3-4-2">
<title>3.4.2 Measuring the Alignment of Embedding and Semantic Spaces</title>
<p>To pair features with topics, we conduct space alignment from the point-wise and pair-wise perspectives. Referring to definitions <xref ref-type="sec" rid="s2-2">Section 2.2</xref> and <xref ref-type="sec" rid="s2-3">Section 2.3</xref>, we aim to align the topic semantic space and feature embedding space from the coordinate system and information contents respectively. During the aligning process, we minimize the point-wise alignment loss <inline-formula id="inf32">
<mml:math id="m36">
<mml:msub>
<mml:mrow>
<mml:mi mathvariant="script">L</mml:mi>
</mml:mrow>
<mml:mrow>
<mml:mi>P</mml:mi>
</mml:mrow>
</mml:msub>
</mml:math>
</inline-formula> and pair-wise alignment loss <inline-formula id="inf33">
<mml:math id="m37">
<mml:msub>
<mml:mrow>
<mml:mi mathvariant="script">L</mml:mi>
</mml:mrow>
<mml:mrow>
<mml:mi>C</mml:mi>
</mml:mrow>
</mml:msub>
</mml:math>
</inline-formula>. To be convenient, we take the <italic>n</italic>-th entity as an example to explain the calculation process.<list list-type="simple">
<list-item>
<p>1) Point-wise Alignment Loss: <inline-formula id="inf34">
<mml:math id="m38">
<mml:msub>
<mml:mrow>
<mml:mi mathvariant="script">L</mml:mi>
</mml:mrow>
<mml:mrow>
<mml:mi>P</mml:mi>
</mml:mrow>
</mml:msub>
</mml:math>
</inline-formula>
<italic>.</italic> We first select <italic>K</italic> values from the topic vector <bold>t</bold>
<sub>
<italic>n</italic>
</sub> as the vector <inline-formula id="inf35">
<mml:math id="m39">
<mml:msub>
<mml:mrow>
<mml:mover>
<mml:mrow>
<mml:mi mathvariant="bold">t</mml:mi>
</mml:mrow>
<mml:mo>&#x30c;</mml:mo>
</mml:mover>
</mml:mrow>
<mml:mrow>
<mml:mi>n</mml:mi>
</mml:mrow>
</mml:msub>
<mml:mo>&#x2208;</mml:mo>
<mml:msup>
<mml:mrow>
<mml:mi mathvariant="double-struck">R</mml:mi>
</mml:mrow>
<mml:mrow>
<mml:mi>K</mml:mi>
</mml:mrow>
</mml:msup>
</mml:math>
</inline-formula>, which contains the most representative semantics in the semantic space. Then, we maximize the correlation between <inline-formula id="inf36">
<mml:math id="m40">
<mml:msub>
<mml:mrow>
<mml:mover>
<mml:mrow>
<mml:mi mathvariant="bold">t</mml:mi>
</mml:mrow>
<mml:mo>&#x30c;</mml:mo>
</mml:mover>
</mml:mrow>
<mml:mrow>
<mml:mi>n</mml:mi>
</mml:mrow>
</mml:msub>
</mml:math>
</inline-formula> and the spatial embedding <bold>r</bold>
<sub>
<italic>n</italic>
</sub>, which is equal to minimize the negative correlation between the two vectors. The formula of the minimizing process as follows:</p>
</list-item>
</list>
<disp-formula id="e5">
<mml:math id="m41">
<mml:msub>
<mml:mrow>
<mml:mi mathvariant="script">L</mml:mi>
</mml:mrow>
<mml:mrow>
<mml:mi>P</mml:mi>
</mml:mrow>
</mml:msub>
<mml:mo>&#x3d;</mml:mo>
<mml:mo>&#x2212;</mml:mo>
<mml:munderover accentunder="false" accent="false">
<mml:mrow>
<mml:mo>&#x2211;</mml:mo>
</mml:mrow>
<mml:mrow>
<mml:mi>n</mml:mi>
<mml:mo>&#x3d;</mml:mo>
<mml:mn>1</mml:mn>
</mml:mrow>
<mml:mrow>
<mml:mi>N</mml:mi>
</mml:mrow>
</mml:munderover>
<mml:mfrac>
<mml:mrow>
<mml:mtext>cov</mml:mtext>
<mml:mfenced open="(" close=")">
<mml:mrow>
<mml:msub>
<mml:mrow>
<mml:mover>
<mml:mrow>
<mml:mi mathvariant="bold">t</mml:mi>
</mml:mrow>
<mml:mo>&#x30c;</mml:mo>
</mml:mover>
</mml:mrow>
<mml:mrow>
<mml:mi>n</mml:mi>
</mml:mrow>
</mml:msub>
<mml:mo>,</mml:mo>
<mml:msub>
<mml:mrow>
<mml:mi mathvariant="bold">r</mml:mi>
</mml:mrow>
<mml:mrow>
<mml:mi>n</mml:mi>
</mml:mrow>
</mml:msub>
</mml:mrow>
</mml:mfenced>
</mml:mrow>
<mml:mrow>
<mml:mi>&#x3b4;</mml:mi>
<mml:mfenced open="(" close=")">
<mml:mrow>
<mml:msub>
<mml:mrow>
<mml:mover>
<mml:mrow>
<mml:mi mathvariant="bold">t</mml:mi>
</mml:mrow>
<mml:mo>&#x30c;</mml:mo>
</mml:mover>
</mml:mrow>
<mml:mrow>
<mml:mi>n</mml:mi>
</mml:mrow>
</mml:msub>
</mml:mrow>
</mml:mfenced>
<mml:mi>&#x3b4;</mml:mi>
<mml:mfenced open="(" close=")">
<mml:mrow>
<mml:msub>
<mml:mrow>
<mml:mi mathvariant="bold">r</mml:mi>
</mml:mrow>
<mml:mrow>
<mml:mi>n</mml:mi>
</mml:mrow>
</mml:msub>
</mml:mrow>
</mml:mfenced>
</mml:mrow>
</mml:mfrac>
<mml:mo>,</mml:mo>
</mml:math>
<label>(5)</label>
</disp-formula>where cov(.) denotes the covariance calculation; <italic>&#x3b4;</italic>(.) denotes the standard deviation.<list list-type="simple">
<list-item>
<p>2) Pair-wise Alignment Loss: <inline-formula id="inf37">
<mml:math id="m42">
<mml:msub>
<mml:mrow>
<mml:mi mathvariant="script">L</mml:mi>
</mml:mrow>
<mml:mrow>
<mml:mi>C</mml:mi>
</mml:mrow>
</mml:msub>
</mml:math>
</inline-formula>
<italic>.</italic> We first construct the topic-topic similarity matrix <bold>S</bold> and the feature-feature similarity matrix <bold>S</bold>&#x2032;. Specifically, for <inline-formula id="inf38">
<mml:math id="m43">
<mml:mi mathvariant="bold">S</mml:mi>
<mml:mo>&#x2208;</mml:mo>
<mml:msup>
<mml:mrow>
<mml:mi mathvariant="double-struck">R</mml:mi>
</mml:mrow>
<mml:mrow>
<mml:mi>K</mml:mi>
<mml:mo>&#xd7;</mml:mo>
<mml:mi>K</mml:mi>
</mml:mrow>
</mml:msup>
</mml:math>
</inline-formula>, we calculate the similarity between any two topics. For <inline-formula id="inf39">
<mml:math id="m44">
<mml:msup>
<mml:mrow>
<mml:mi mathvariant="bold">S</mml:mi>
</mml:mrow>
<mml:mrow>
<mml:mo>&#x2032;</mml:mo>
</mml:mrow>
</mml:msup>
<mml:mo>&#x2208;</mml:mo>
<mml:msup>
<mml:mrow>
<mml:mi mathvariant="double-struck">R</mml:mi>
</mml:mrow>
<mml:mrow>
<mml:mi>K</mml:mi>
<mml:mo>&#xd7;</mml:mo>
<mml:mi>K</mml:mi>
</mml:mrow>
</mml:msup>
</mml:math>
</inline-formula>, we calculate the similarity between two features of spatial embeddings. We keep the pair-wise consistency between <bold>S</bold> and <bold>S</bold>&#x2032; by minimizing the Frobenius norm, as follows:</p>
</list-item>
</list>
<disp-formula id="e6">
<mml:math id="m45">
<mml:msub>
<mml:mrow>
<mml:mi mathvariant="script">L</mml:mi>
</mml:mrow>
<mml:mrow>
<mml:mi>C</mml:mi>
</mml:mrow>
</mml:msub>
<mml:mo>&#x3d;</mml:mo>
<mml:mo stretchy="false">&#x2016;</mml:mo>
<mml:mi mathvariant="bold">S</mml:mi>
<mml:mo>&#x2212;</mml:mo>
<mml:mrow>
<mml:mi mathvariant="bold">S</mml:mi>
</mml:mrow>
<mml:mo>&#x2032;</mml:mo>
<mml:msub>
<mml:mrow>
<mml:mo stretchy="false">&#x2016;</mml:mo>
</mml:mrow>
<mml:mrow>
<mml:mi>F</mml:mi>
</mml:mrow>
</mml:msub>
<mml:mo>.</mml:mo>
</mml:math>
<label>(6)</label>
</disp-formula>
</p>
</sec>
<sec id="s3-4-3">
<title>3.4.3 Supervised PSO For Automatic Topic Selection</title>
<p>As introduced above, we select <italic>K</italic> topics so the representation learner can learn a <italic>K</italic>-sized embedding vector in terms of <italic>K</italic> topics to achieve feature-topic alignment. However, how can the machine automatically identify the best <italic>K</italic> and select the most appropriate <italic>K</italic> topics?</p>
<p>A naive idea is that we can select <italic>K</italic> topics randomly at each iteration until we traverse all topic combinations and find the best topic subset based on the objective function. The searching process, however, is time-consuming and computationally expensive. Moreover, the topic selection problem belongs to the combinatorial optimization field, which is hard to solve by derivative-based optimization algorithms. Thus, a quickly and derivative-free optimization algorithm should be selected as our optimizer. Considering the high time complexity for traversing all possible subsets to find the optimal result, we propose to formulate the joint task of feature learning, topic selection, topic and feature matching into a PSO problem.</p>
<p>The PSO-based optimization framework is as illustrated in <xref ref-type="fig" rid="F7">Figure 7</xref>. Specifically, we first randomly initialize a number of particles in PSO, where a particle is a binary topic mask (i.e.,&#x20;the mask value of 1 indicates &#x201c;select&#x201d; and the mask value of 0 indicates &#x201c;deselect&#x201d;). In other words, a set of particles select a subset of topics. A multi-objective deep learning model, whose objective function includes the losses of graph reconstruction, semantic alignment, and the regression estimator in the downstream task, is trained to learn spatial representations, using each selected topic subset. As an application, we use the embedding of spatial entities (residential communities) to predict their real estate prices, and the loss of the regression model <inline-formula id="inf40">
<mml:math id="m46">
<mml:msub>
<mml:mrow>
<mml:mi mathvariant="script">L</mml:mi>
</mml:mrow>
<mml:mrow>
<mml:mi>R</mml:mi>
<mml:mi>e</mml:mi>
<mml:mi>g</mml:mi>
</mml:mrow>
</mml:msub>
</mml:math>
</inline-formula> is:<disp-formula id="e7">
<mml:math id="m47">
<mml:msub>
<mml:mrow>
<mml:mi mathvariant="script">L</mml:mi>
</mml:mrow>
<mml:mrow>
<mml:mi>R</mml:mi>
<mml:mi>e</mml:mi>
<mml:mi>g</mml:mi>
</mml:mrow>
</mml:msub>
<mml:mo>&#x3d;</mml:mo>
<mml:mfrac>
<mml:mrow>
<mml:mn>1</mml:mn>
</mml:mrow>
<mml:mrow>
<mml:mi>N</mml:mi>
</mml:mrow>
</mml:mfrac>
<mml:munderover accentunder="false" accent="false">
<mml:mrow>
<mml:mo>&#x2211;</mml:mo>
</mml:mrow>
<mml:mrow>
<mml:mi>n</mml:mi>
<mml:mo>&#x3d;</mml:mo>
<mml:mn>1</mml:mn>
</mml:mrow>
<mml:mrow>
<mml:mi>N</mml:mi>
</mml:mrow>
</mml:munderover>
<mml:msup>
<mml:mrow>
<mml:mfenced open="(" close=")">
<mml:mrow>
<mml:msub>
<mml:mrow>
<mml:mi>c</mml:mi>
</mml:mrow>
<mml:mrow>
<mml:mi>n</mml:mi>
</mml:mrow>
</mml:msub>
<mml:mo>&#x2212;</mml:mo>
<mml:msubsup>
<mml:mrow>
<mml:mi>c</mml:mi>
</mml:mrow>
<mml:mrow>
<mml:mi>n</mml:mi>
</mml:mrow>
<mml:mrow>
<mml:mo>&#x2217;</mml:mo>
</mml:mrow>
</mml:msubsup>
</mml:mrow>
</mml:mfenced>
</mml:mrow>
<mml:mrow>
<mml:mn>2</mml:mn>
</mml:mrow>
</mml:msup>
<mml:mo>,</mml:mo>
</mml:math>
<label>(7)</label>
</disp-formula>where <italic>c</italic>
<sub>
<italic>n</italic>
</sub> is the golden standard real estate price and <inline-formula id="inf41">
<mml:math id="m48">
<mml:msubsup>
<mml:mrow>
<mml:mi>c</mml:mi>
</mml:mrow>
<mml:mrow>
<mml:mi>n</mml:mi>
</mml:mrow>
<mml:mrow>
<mml:mo>&#x2217;</mml:mo>
</mml:mrow>
</mml:msubsup>
</mml:math>
</inline-formula> is the predicted price. Next, we calculate the fitness of each particle according to the total loss of the deep model. The fitness can be calculated by:<disp-formula id="e8">
<mml:math id="m49">
<mml:mi>F</mml:mi>
<mml:mi>i</mml:mi>
<mml:mi>t</mml:mi>
<mml:mi>n</mml:mi>
<mml:mi>e</mml:mi>
<mml:mi>s</mml:mi>
<mml:mi>s</mml:mi>
<mml:mo>&#x3d;</mml:mo>
<mml:msub>
<mml:mrow>
<mml:mi mathvariant="script">L</mml:mi>
</mml:mrow>
<mml:mrow>
<mml:mi>C</mml:mi>
</mml:mrow>
</mml:msub>
<mml:mo>&#x2b;</mml:mo>
<mml:msub>
<mml:mrow>
<mml:mi mathvariant="script">L</mml:mi>
</mml:mrow>
<mml:mrow>
<mml:mi>P</mml:mi>
</mml:mrow>
</mml:msub>
<mml:mo>&#x2b;</mml:mo>
<mml:msub>
<mml:mrow>
<mml:mi mathvariant="script">L</mml:mi>
</mml:mrow>
<mml:mrow>
<mml:mi>R</mml:mi>
</mml:mrow>
</mml:msub>
<mml:mo>&#x2b;</mml:mo>
<mml:msub>
<mml:mrow>
<mml:mi mathvariant="script">L</mml:mi>
</mml:mrow>
<mml:mrow>
<mml:mi>R</mml:mi>
<mml:mi>e</mml:mi>
<mml:mi>g</mml:mi>
</mml:mrow>
</mml:msub>
<mml:mo>.</mml:mo>
</mml:math>
<label>(8)</label>
</disp-formula>
</p>
<fig id="F7" position="float">
<label>FIGURE 7</label>
<caption>
<p>The optimization process of the PSO topic selector.</p>
</caption>
<graphic xlink:href="fdata-04-762899-g007.tif"/>
</fig>
<p>Then, we utilize the fitness to inform all particles how far they are from the best solution. Next, each particle moves forward to the solution based on not only its current status but also all particles&#x2019; movement. After the fitness value of PSO converges, PSO identifies the best topic subset. Finally, the semantically-rich embeddings of spatial entities, given by: <inline-formula id="inf42">
<mml:math id="m50">
<mml:mrow>
<mml:mover accent="true">
<mml:mrow>
<mml:mi mathvariant="bold">R</mml:mi>
</mml:mrow>
<mml:mo>&#x303;</mml:mo>
</mml:mover>
</mml:mrow>
<mml:mo>&#x3d;</mml:mo>
<mml:msubsup>
<mml:mrow>
<mml:mrow>
<mml:mo stretchy="false">{</mml:mo>
<mml:mrow>
<mml:msub>
<mml:mrow>
<mml:mover accent="true">
<mml:mrow>
<mml:mi mathvariant="bold">r</mml:mi>
</mml:mrow>
<mml:mo>&#x303;</mml:mo>
</mml:mover>
</mml:mrow>
<mml:mrow>
<mml:mi>n</mml:mi>
</mml:mrow>
</mml:msub>
</mml:mrow>
<mml:mo stretchy="false">}</mml:mo>
</mml:mrow>
</mml:mrow>
<mml:mrow>
<mml:mi>n</mml:mi>
<mml:mo>&#x3d;</mml:mo>
<mml:mn>1</mml:mn>
</mml:mrow>
<mml:mrow>
<mml:mi>N</mml:mi>
</mml:mrow>
</mml:msubsup>
</mml:math>
</inline-formula>.</p>
</sec>
</sec>
</sec>
<sec id="s4">
<title>4 Experimental Results</title>
<p>In this section, we present extensive experiments with real world data to answer the following research questions: Q1. How effective is our proposed AutoFTP in spatial representation learning? Q2. How effective is each loss function of AutoFTP? Are they all necessary for spatial representation learning? Q3. How well does AutoFTP work in space alignment? Do the learned spatial embeddings contain semantic meanings? Q4. How about the robustness of AutoFTP? Does it always outperform other baselines? Q5. How about the stability and sensitivity of AutoFTP?</p>
<sec id="s4-1">
<title>4.1 Experimental Setup</title>
<sec id="s4-1-1">
<title>4.1.1 Data Description</title>
<p>
<xref ref-type="table" rid="T1">Table&#x20;1</xref> shows the statistics of five data sources used in the experiments. Firstly, the taxi traces data describes the GPS trajectory of taxis in Beijing in 3&#xa0;months. The format of each trace record is &#x3c; trip id, distance, travel time, average speed, pick-up time, drop-off time, pick-up location, drop-off location &#x3e;. Secondly, the residential regions, texts, and real estate price data sources are crawled from <ext-link ext-link-type="uri" xlink:href="http://www.fang.com">www.fang.com</ext-link>. In&#x20;experiments, the residential regions are treated as spatial entities. The texts reflect the urban utilities and characteristics of spatial entities from multiple perspectives such as traffic condition, economic development, demographic situation, and etc. The real estate prices indicate the average value of the real estate of each spatial entity in 6&#xa0;months. Thirdly, the POIs are extracted from <ext-link ext-link-type="uri" xlink:href="http://www.dianping.com">www.dianping.com</ext-link>, which is a POI (small businesses such as restaurants, banks, gas stations, shopping markets) review website in China. Each POI is described in a format of &#x3c; POI id, POI category, latitude, longitude &#x3e;.</p>
<table-wrap id="T1" position="float">
<label>TABLE 1</label>
<caption>
<p>Statistics of the experimental&#x20;data.</p>
</caption>
<table>
<thead>
<tr>
<td align="left">Data sources</td>
<td align="center">Properties</td>
<td align="center">Statistics</td>
</tr>
</thead>
<tbody valign="top">
<tr>
<td rowspan="2" align="left">Taxi Traces</td>
<td align="left">Number of taxis</td>
<td align="left">13,597</td>
</tr>
<tr>
<td align="left">Time period</td>
<td align="left">Apr.&#x2013;Aug. 2012</td>
</tr>
<tr>
<td rowspan="2" align="left">Residential regions</td>
<td align="left">Number of residential regions</td>
<td align="left">2,990</td>
</tr>
<tr>
<td align="left">Time period of transactions</td>
<td align="left">04/2011&#x2013;09/2012</td>
</tr>
<tr>
<td rowspan="2" align="left">POIs</td>
<td align="left">Number of POIs</td>
<td align="left">328,668</td>
</tr>
<tr>
<td align="left">Number of POI categories</td>
<td align="left">20</td>
</tr>
<tr>
<td rowspan="2" align="left">Texts</td>
<td align="left">Number of textual descriptions</td>
<td align="left">2,990</td>
</tr>
<tr>
<td align="left">Time Period</td>
<td align="left">04/2011&#x2013;09/2012</td>
</tr>
<tr>
<td rowspan="2" align="left">Real Estate Prices</td>
<td align="left">Number of real estate prices</td>
<td align="left">41,753</td>
</tr>
<tr>
<td align="left">Time Period</td>
<td align="left">12/2011&#x2013;06/2012</td>
</tr>
</tbody>
</table>
</table-wrap>
</sec>
<sec id="s4-1-2">
<title>4.1.2 Application: Real Estate Price Prediction</title>
<p>Our proposed method (AutoFTP) can learn a list of vectorized representations for all spatial entities. Therefore, as a downstream application, we can apply these representations to train a regression model to predict the average real estate price of these spatial entities. Specifically, we first apply AutoFTP to learn a series of representations of spatial entities based on their geographical structural information and related text descriptions. Then, we build up a deep neural network (DNN) model for predicting average real estate price of each spatial entity according to its corresponding representation. To be convenient, we take the <italic>n</italic>-th spatial entity as an example to explain the regression model. The formulation of DNN is <inline-formula id="inf43">
<mml:math id="m51">
<mml:mi>f</mml:mi>
<mml:mrow>
<mml:mo stretchy="false">(</mml:mo>
<mml:mrow>
<mml:msub>
<mml:mrow>
<mml:mover accent="true">
<mml:mrow>
<mml:mi mathvariant="bold">r</mml:mi>
</mml:mrow>
<mml:mo>&#x303;</mml:mo>
</mml:mover>
</mml:mrow>
<mml:mrow>
<mml:mi>n</mml:mi>
</mml:mrow>
</mml:msub>
<mml:mo>,</mml:mo>
<mml:mi mathvariant="bold">w</mml:mi>
</mml:mrow>
<mml:mo stretchy="false">)</mml:mo>
</mml:mrow>
<mml:mo>&#x3d;</mml:mo>
<mml:mi mathvariant="bold">w</mml:mi>
<mml:mo>&#x22c5;</mml:mo>
<mml:mi>g</mml:mi>
<mml:mrow>
<mml:mo stretchy="false">(</mml:mo>
<mml:mrow>
<mml:msub>
<mml:mrow>
<mml:mover accent="true">
<mml:mrow>
<mml:mi mathvariant="bold">r</mml:mi>
</mml:mrow>
<mml:mo>&#x303;</mml:mo>
</mml:mover>
</mml:mrow>
<mml:mrow>
<mml:mi>n</mml:mi>
</mml:mrow>
</mml:msub>
</mml:mrow>
<mml:mo stretchy="false">)</mml:mo>
</mml:mrow>
<mml:mo>&#x2b;</mml:mo>
<mml:mi mathvariant="bold">b</mml:mi>
</mml:math>
</inline-formula>, where <inline-formula id="inf44">
<mml:math id="m52">
<mml:msub>
<mml:mrow>
<mml:mover accent="true">
<mml:mrow>
<mml:mi mathvariant="bold">r</mml:mi>
</mml:mrow>
<mml:mo>&#x303;</mml:mo>
</mml:mover>
</mml:mrow>
<mml:mrow>
<mml:mi>n</mml:mi>
</mml:mrow>
</mml:msub>
</mml:math>
</inline-formula> is the representation of the <italic>n</italic>-th spatial entity, <inline-formula id="inf45">
<mml:math id="m53">
<mml:mi>g</mml:mi>
<mml:mrow>
<mml:mo stretchy="false">(</mml:mo>
<mml:mrow>
<mml:msub>
<mml:mrow>
<mml:mover accent="true">
<mml:mrow>
<mml:mi mathvariant="bold">r</mml:mi>
</mml:mrow>
<mml:mo>&#x303;</mml:mo>
</mml:mover>
</mml:mrow>
<mml:mrow>
<mml:mi>n</mml:mi>
</mml:mrow>
</mml:msub>
</mml:mrow>
<mml:mo stretchy="false">)</mml:mo>
</mml:mrow>
</mml:math>
</inline-formula> is the nonlinear transformation of <inline-formula id="inf46">
<mml:math id="m54">
<mml:msub>
<mml:mrow>
<mml:mover accent="true">
<mml:mrow>
<mml:mi mathvariant="bold">r</mml:mi>
</mml:mrow>
<mml:mo>&#x303;</mml:mo>
</mml:mover>
</mml:mrow>
<mml:mrow>
<mml:mi>n</mml:mi>
</mml:mrow>
</mml:msub>
</mml:math>
</inline-formula>, <bold>w</bold> is the weight term, and <bold>b</bold> is the bias term. We want to minimize the difference between predicted price <inline-formula id="inf47">
<mml:math id="m55">
<mml:mi>f</mml:mi>
<mml:mrow>
<mml:mo stretchy="false">(</mml:mo>
<mml:mrow>
<mml:msub>
<mml:mrow>
<mml:mover accent="true">
<mml:mrow>
<mml:mi mathvariant="bold">r</mml:mi>
</mml:mrow>
<mml:mo>&#x303;</mml:mo>
</mml:mover>
</mml:mrow>
<mml:mrow>
<mml:mi>n</mml:mi>
</mml:mrow>
</mml:msub>
<mml:mo>,</mml:mo>
<mml:mi mathvariant="bold">w</mml:mi>
</mml:mrow>
<mml:mo stretchy="false">)</mml:mo>
</mml:mrow>
</mml:math>
</inline-formula> and real price <italic>y</italic>
<sub>
<italic>n</italic>
</sub>. Thus, the objective of the DNN is <inline-formula id="inf48">
<mml:math id="m56">
<mml:mtext>min</mml:mtext>
<mml:mfrac>
<mml:mrow>
<mml:mn>1</mml:mn>
</mml:mrow>
<mml:mrow>
<mml:mi>N</mml:mi>
</mml:mrow>
</mml:mfrac>
<mml:msubsup>
<mml:mrow>
<mml:mo movablelimits="false" form="prefix">&#x2211;</mml:mo>
</mml:mrow>
<mml:mrow>
<mml:mi>n</mml:mi>
<mml:mo>&#x3d;</mml:mo>
<mml:mn>1</mml:mn>
</mml:mrow>
<mml:mrow>
<mml:mi>N</mml:mi>
</mml:mrow>
</mml:msubsup>
<mml:msup>
<mml:mrow>
<mml:mrow>
<mml:mo stretchy="false">(</mml:mo>
<mml:mrow>
<mml:msub>
<mml:mrow>
<mml:mi>y</mml:mi>
</mml:mrow>
<mml:mrow>
<mml:mi>n</mml:mi>
</mml:mrow>
</mml:msub>
<mml:mo>&#x2212;</mml:mo>
<mml:mi>f</mml:mi>
<mml:mrow>
<mml:mo stretchy="false">(</mml:mo>
<mml:mrow>
<mml:msub>
<mml:mrow>
<mml:mover accent="true">
<mml:mrow>
<mml:mi mathvariant="bold">r</mml:mi>
</mml:mrow>
<mml:mo>&#x303;</mml:mo>
</mml:mover>
</mml:mrow>
<mml:mrow>
<mml:mi>n</mml:mi>
</mml:mrow>
</mml:msub>
<mml:mo>,</mml:mo>
<mml:mi mathvariant="bold">w</mml:mi>
</mml:mrow>
<mml:mo stretchy="false">)</mml:mo>
</mml:mrow>
</mml:mrow>
<mml:mo stretchy="false">)</mml:mo>
</mml:mrow>
</mml:mrow>
<mml:mrow>
<mml:mn>2</mml:mn>
</mml:mrow>
</mml:msup>
</mml:math>
</inline-formula>, where <italic>N</italic> is the total number of spatial entities.</p>
</sec>
<sec id="s4-1-3">
<title>4.1.3 Evaluation Metrics</title>
<p>We evaluated our method using a real estate price prediction task (<xref ref-type="sec" rid="s4-1-2">Section 4.1.2</xref>). We took the feature representation vectors of residential communities as inputs, and predicted their real estate prices. We compared the golden-standard prices <italic>y</italic>
<sub>
<italic>n</italic>
</sub> with the predicted prices <inline-formula id="inf49">
<mml:math id="m57">
<mml:msub>
<mml:mrow>
<mml:mover accent="true">
<mml:mrow>
<mml:mi>y</mml:mi>
</mml:mrow>
<mml:mo stretchy="false">&#x302;</mml:mo>
</mml:mover>
</mml:mrow>
<mml:mrow>
<mml:mi>n</mml:mi>
</mml:mrow>
</mml:msub>
</mml:math>
</inline-formula> in terms of four metrics: 1) <inline-formula id="inf50">
<mml:math id="m58">
<mml:mtext>RMSE</mml:mtext>
<mml:mo>&#x3d;</mml:mo>
<mml:msqrt>
<mml:mrow>
<mml:mfrac>
<mml:mrow>
<mml:mn>1</mml:mn>
</mml:mrow>
<mml:mrow>
<mml:mi>N</mml:mi>
</mml:mrow>
</mml:mfrac>
<mml:msubsup>
<mml:mrow>
<mml:mo movablelimits="false" form="prefix">&#x2211;</mml:mo>
</mml:mrow>
<mml:mrow>
<mml:mi>n</mml:mi>
<mml:mo>&#x3d;</mml:mo>
<mml:mn>1</mml:mn>
</mml:mrow>
<mml:mrow>
<mml:mi>N</mml:mi>
</mml:mrow>
</mml:msubsup>
<mml:msup>
<mml:mrow>
<mml:mrow>
<mml:mo stretchy="false">(</mml:mo>
<mml:mrow>
<mml:msub>
<mml:mrow>
<mml:mi>y</mml:mi>
</mml:mrow>
<mml:mrow>
<mml:mi>n</mml:mi>
</mml:mrow>
</mml:msub>
<mml:mo>&#x2212;</mml:mo>
<mml:msub>
<mml:mrow>
<mml:mover accent="true">
<mml:mrow>
<mml:mi>y</mml:mi>
</mml:mrow>
<mml:mo stretchy="false">&#x302;</mml:mo>
</mml:mover>
</mml:mrow>
<mml:mrow>
<mml:mi>n</mml:mi>
</mml:mrow>
</mml:msub>
</mml:mrow>
<mml:mo stretchy="false">)</mml:mo>
</mml:mrow>
</mml:mrow>
<mml:mrow>
<mml:mn>2</mml:mn>
</mml:mrow>
</mml:msup>
</mml:mrow>
</mml:msqrt>
</mml:math>
</inline-formula>; 2) <inline-formula id="inf51">
<mml:math id="m59">
<mml:mtext>MAE</mml:mtext>
<mml:mo>&#x3d;</mml:mo>
<mml:mfrac>
<mml:mrow>
<mml:mn>1</mml:mn>
</mml:mrow>
<mml:mrow>
<mml:mi>N</mml:mi>
</mml:mrow>
</mml:mfrac>
<mml:mtext>&#x2009;</mml:mtext>
<mml:msubsup>
<mml:mrow>
<mml:mo movablelimits="false" form="prefix">&#x2211;</mml:mo>
</mml:mrow>
<mml:mrow>
<mml:mi>n</mml:mi>
<mml:mo>&#x3d;</mml:mo>
<mml:mn>1</mml:mn>
</mml:mrow>
<mml:mrow>
<mml:mi>N</mml:mi>
</mml:mrow>
</mml:msubsup>
<mml:mfenced open="|" close="|">
<mml:mrow>
<mml:mrow>
<mml:mo stretchy="false">(</mml:mo>
<mml:mrow>
<mml:msub>
<mml:mrow>
<mml:mi>y</mml:mi>
</mml:mrow>
<mml:mrow>
<mml:mi>n</mml:mi>
</mml:mrow>
</mml:msub>
<mml:mo>&#x2212;</mml:mo>
<mml:msub>
<mml:mrow>
<mml:mover accent="true">
<mml:mrow>
<mml:mi>y</mml:mi>
</mml:mrow>
<mml:mo stretchy="false">&#x302;</mml:mo>
</mml:mover>
</mml:mrow>
<mml:mrow>
<mml:mi>n</mml:mi>
</mml:mrow>
</mml:msub>
</mml:mrow>
<mml:mo stretchy="false">)</mml:mo>
</mml:mrow>
</mml:mrow>
</mml:mfenced>
</mml:math>
</inline-formula>; 3) <inline-formula id="inf52">
<mml:math id="m60">
<mml:mtext>MAPE</mml:mtext>
<mml:mo>&#x3d;</mml:mo>
<mml:mfrac>
<mml:mrow>
<mml:mn>100</mml:mn>
</mml:mrow>
<mml:mrow>
<mml:mi>N</mml:mi>
</mml:mrow>
</mml:mfrac>
<mml:msubsup>
<mml:mrow>
<mml:mo movablelimits="false" form="prefix">&#x2211;</mml:mo>
</mml:mrow>
<mml:mrow>
<mml:mi>n</mml:mi>
<mml:mo>&#x3d;</mml:mo>
<mml:mn>1</mml:mn>
</mml:mrow>
<mml:mrow>
<mml:mi>N</mml:mi>
</mml:mrow>
</mml:msubsup>
<mml:mfenced open="|" close="|">
<mml:mrow>
<mml:mfrac>
<mml:mrow>
<mml:msub>
<mml:mrow>
<mml:mi>y</mml:mi>
</mml:mrow>
<mml:mrow>
<mml:mi>n</mml:mi>
</mml:mrow>
</mml:msub>
<mml:mo>&#x2212;</mml:mo>
<mml:mrow>
<mml:mover accent="true">
<mml:mrow>
<mml:msub>
<mml:mrow>
<mml:mi>y</mml:mi>
</mml:mrow>
<mml:mrow>
<mml:mi>n</mml:mi>
</mml:mrow>
</mml:msub>
</mml:mrow>
<mml:mo>&#x302;</mml:mo>
</mml:mover>
</mml:mrow>
</mml:mrow>
<mml:mrow>
<mml:msub>
<mml:mrow>
<mml:mi>y</mml:mi>
</mml:mrow>
<mml:mrow>
<mml:mi>n</mml:mi>
</mml:mrow>
</mml:msub>
</mml:mrow>
</mml:mfrac>
</mml:mrow>
</mml:mfenced>
</mml:math>
</inline-formula>; 4) <inline-formula id="inf53">
<mml:math id="m61">
<mml:mtext>MSLE</mml:mtext>
<mml:mo>&#x3d;</mml:mo>
<mml:mfrac>
<mml:mrow>
<mml:mn>1</mml:mn>
</mml:mrow>
<mml:mrow>
<mml:mi>N</mml:mi>
</mml:mrow>
</mml:mfrac>
<mml:msubsup>
<mml:mrow>
<mml:mo movablelimits="false" form="prefix">&#x2211;</mml:mo>
</mml:mrow>
<mml:mrow>
<mml:mi>n</mml:mi>
<mml:mo>&#x3d;</mml:mo>
<mml:mn>1</mml:mn>
</mml:mrow>
<mml:mrow>
<mml:mi>N</mml:mi>
</mml:mrow>
</mml:msubsup>
<mml:msup>
<mml:mrow>
<mml:mrow>
<mml:mo stretchy="false">(</mml:mo>
<mml:mrow>
<mml:mi>l</mml:mi>
<mml:mi>o</mml:mi>
<mml:mi>g</mml:mi>
<mml:mrow>
<mml:mo stretchy="false">(</mml:mo>
<mml:mrow>
<mml:mn>1</mml:mn>
<mml:mo>&#x2b;</mml:mo>
<mml:msub>
<mml:mrow>
<mml:mi>y</mml:mi>
</mml:mrow>
<mml:mrow>
<mml:mi>n</mml:mi>
</mml:mrow>
</mml:msub>
</mml:mrow>
<mml:mo stretchy="false">)</mml:mo>
</mml:mrow>
<mml:mo>&#x2212;</mml:mo>
<mml:mi>l</mml:mi>
<mml:mi>o</mml:mi>
<mml:mi>g</mml:mi>
<mml:mrow>
<mml:mo stretchy="false">(</mml:mo>
<mml:mrow>
<mml:msub>
<mml:mrow>
<mml:mover accent="true">
<mml:mrow>
<mml:mi>y</mml:mi>
</mml:mrow>
<mml:mo stretchy="false">&#x302;</mml:mo>
</mml:mover>
</mml:mrow>
<mml:mrow>
<mml:mi>n</mml:mi>
</mml:mrow>
</mml:msub>
<mml:mo>&#x2b;</mml:mo>
<mml:mn>1</mml:mn>
</mml:mrow>
<mml:mo stretchy="false">)</mml:mo>
</mml:mrow>
</mml:mrow>
<mml:mo stretchy="false">)</mml:mo>
</mml:mrow>
</mml:mrow>
<mml:mrow>
<mml:mn>2</mml:mn>
</mml:mrow>
</mml:msup>
</mml:math>
</inline-formula>. The regression loss and optimization algorithm are controlled to be the same. The lower the four metrics are, the more effective the spatial embedding features&#x20;are.</p>
</sec>
<sec id="s4-1-4">
<title>4.1.4 Baseline Algorithms</title>
<p>We compared our proposed method with seven widely-used and robust representation learning (embedding) methods as follows: 1) AttentionWalk (<xref ref-type="bibr" rid="B1">Abu-El-Haija et&#x20;al., 2018</xref>) utilizes a novel attention model to automatically learn the hyper-parameters of random-walk based network embedding methods, which improves the flexibility and performance of the model. We set the learning rate as 0.01, the regularization parameters as 0.5. 2) ProNE (<xref ref-type="bibr" rid="B54">Zhang J.&#x20;et&#x20;al., 2019</xref>) formulates the network embedding as sparse matrix factorization to improve the calculation speed, and conducts the propagation process in the spectrally modulated space to enhance the representation. We adopt the default parameter setting in (<xref ref-type="bibr" rid="B54">Zhang J.&#x20;et&#x20;al., 2019</xref>). 3) GatNE (<xref ref-type="bibr" rid="B6">Cen et&#x20;al., 2019</xref>) is a random-walk based network embedding method, which considers the information of different attributes of nodes to enhance the graph representation. We set the number of walks as 20, walk length as 10, window size as 5, patience as 5. 4) GAE (<xref ref-type="bibr" rid="B20">Kipf and Welling, 2016</xref>) utilizes GCN to learn the node representations in the encode-decoder paradigm by minimizing the reconstruction loss. We set the number of GCN layers as 2 and the learning rate as 0.0001. 5) DeepWalk (<xref ref-type="bibr" rid="B29">Perozzi et&#x20;al., 2014</xref>) is an extension of the word2vec model (<xref ref-type="bibr" rid="B27">Mikolov et&#x20;al., 2013</xref>), which brings the idea of truncated random walks to a network embedding scenario. We set the number of walks as 80, walk length as 10, and window size as 5. 6) Node2Vec (<xref ref-type="bibr" rid="B10">Grover and Leskovec, 2016</xref>) is an enhanced version of DeepWalk, which considers the homogeneity and structural equivalence of networks during embedding process. We set the number of walks as 80, walk length as 10, window size as 5, return parameter <italic>p</italic> as 0.25 and in-out parameter <italic>q</italic> as 4. 7) Struc2Vec (<xref ref-type="bibr" rid="B31">Ribeiro et&#x20;al., 2017</xref>) learns the node representation by considering the structural identity of nodes in the network. We set the number of walks as 80 and walk length as&#x20;10.</p>
<p>Besides, there are four losses in AutoFTP: reconstruction loss <inline-formula id="inf54">
<mml:math id="m62">
<mml:msub>
<mml:mrow>
<mml:mi mathvariant="script">L</mml:mi>
</mml:mrow>
<mml:mrow>
<mml:mi>R</mml:mi>
</mml:mrow>
</mml:msub>
</mml:math>
</inline-formula>, point-wise alignment loss <inline-formula id="inf55">
<mml:math id="m63">
<mml:msub>
<mml:mrow>
<mml:mi mathvariant="script">L</mml:mi>
</mml:mrow>
<mml:mrow>
<mml:mi>P</mml:mi>
</mml:mrow>
</mml:msub>
</mml:math>
</inline-formula>, pair-wise alignment loss <inline-formula id="inf56">
<mml:math id="m64">
<mml:msub>
<mml:mrow>
<mml:mi mathvariant="script">L</mml:mi>
</mml:mrow>
<mml:mrow>
<mml:mi>C</mml:mi>
</mml:mrow>
</mml:msub>
</mml:math>
</inline-formula>, and regression loss <inline-formula id="inf57">
<mml:math id="m65">
<mml:msub>
<mml:mrow>
<mml:mi mathvariant="script">L</mml:mi>
</mml:mrow>
<mml:mrow>
<mml:mi>R</mml:mi>
<mml:mi>e</mml:mi>
<mml:mi>g</mml:mi>
</mml:mrow>
</mml:msub>
</mml:math>
</inline-formula>. The four losses provide the optimization direction of AutoFTP. To study the benefits of each part, we develop four internal variants of AutoFTP: 1) <bold>AutoFTP</bold>
<sup>
<italic>R</italic>
</sup>, which only keeps <inline-formula id="inf58">
<mml:math id="m66">
<mml:msub>
<mml:mrow>
<mml:mi mathvariant="script">L</mml:mi>
</mml:mrow>
<mml:mrow>
<mml:mi>R</mml:mi>
</mml:mrow>
</mml:msub>
</mml:math>
</inline-formula> of AutoFTP; 2) <bold>AutoFTP</bold>
<sup>(<italic>R</italic>&#x2b;<italic>P</italic>)</sup>, which keeps <inline-formula id="inf59">
<mml:math id="m67">
<mml:msub>
<mml:mrow>
<mml:mi mathvariant="script">L</mml:mi>
</mml:mrow>
<mml:mrow>
<mml:mi>R</mml:mi>
</mml:mrow>
</mml:msub>
</mml:math>
</inline-formula> and <inline-formula id="inf60">
<mml:math id="m68">
<mml:msub>
<mml:mrow>
<mml:mi mathvariant="script">L</mml:mi>
</mml:mrow>
<mml:mrow>
<mml:mi>P</mml:mi>
</mml:mrow>
</mml:msub>
</mml:math>
</inline-formula> of AutoFTP; 3) <bold>AutoFTP</bold>
<sup>(<italic>R</italic>&#x2b;<italic>C</italic>)</sup>, which keeps <inline-formula id="inf61">
<mml:math id="m69">
<mml:msub>
<mml:mrow>
<mml:mi mathvariant="script">L</mml:mi>
</mml:mrow>
<mml:mrow>
<mml:mi>R</mml:mi>
</mml:mrow>
</mml:msub>
</mml:math>
</inline-formula> and <inline-formula id="inf62">
<mml:math id="m70">
<mml:msub>
<mml:mrow>
<mml:mi mathvariant="script">L</mml:mi>
</mml:mrow>
<mml:mrow>
<mml:mi>C</mml:mi>
</mml:mrow>
</mml:msub>
</mml:math>
</inline-formula> of AutoFTP; 4) <bold>AutoFTP</bold>
<sup>(<italic>R</italic>&#x2b;<italic>P</italic>&#x2b;<italic>C</italic>)</sup>, which keeps <inline-formula id="inf63">
<mml:math id="m71">
<mml:msub>
<mml:mrow>
<mml:mi mathvariant="script">L</mml:mi>
</mml:mrow>
<mml:mrow>
<mml:mi>R</mml:mi>
</mml:mrow>
</mml:msub>
</mml:math>
</inline-formula>, <inline-formula id="inf64">
<mml:math id="m72">
<mml:msub>
<mml:mrow>
<mml:mi mathvariant="script">L</mml:mi>
</mml:mrow>
<mml:mrow>
<mml:mi>P</mml:mi>
</mml:mrow>
</mml:msub>
</mml:math>
</inline-formula>, and <inline-formula id="inf65">
<mml:math id="m73">
<mml:msub>
<mml:mrow>
<mml:mi mathvariant="script">L</mml:mi>
</mml:mrow>
<mml:mrow>
<mml:mi>C</mml:mi>
</mml:mrow>
</mml:msub>
</mml:math>
</inline-formula> of AutoFTP. The dimension of embeddings in all models is&#x20;20.</p>
</sec>
<sec id="s4-1-5">
<title>4.1.5 Hyperparameters, Source Code, and Reproducibility</title>
<p>We detailed the hyperarameters and the steps of our algorithm in the <xref ref-type="app" rid="app1">Appendix</xref>. We released our code<xref ref-type="fn" rid="FN4">
<sup>4</sup>
</xref> to help to reproduce experimental results.</p>
</sec>
<sec id="s4-1-6">
<title>4.1.6 Environmental Settings</title>
<p>The experimental studies were conducted in the Ubuntu 18.04.3 LTS operating system, plus Intel(R) Core(TM) i9-9920X CPU@ 3.50GHz, 1 way SLI Titan RTX and 128GB of RAM, with the&#x20;framework of Python 3.7.4, Tensorflow 2.0.0, and Pyswarm 1.3.0.</p>
</sec>
</sec>
<sec id="s4-2">
<title>4.2 Overall Performance (Q1)</title>
<p>
<xref ref-type="table" rid="T2">Table&#x20;2</xref> shows the comparison of all the 11 models. As can be seen, AutoFTP, in overall, outperforms the baseline algorithms in terms of RMSE, MAE, MAPE and MSLE. A possible reason for this observation is that compared with other baseline algorithms, AutoFTP not just captures geographical structural information but also preserves rich semantics of spatial entity. Besides, the&#x20;regression estimator (the downstream task) of AutoFTP provides a clear learning direction (accuracy) for spatial representation learning. Thus, in the downstream predictive task, the spatial embedding features learned by AutoFTP beats all baselines.</p>
<table-wrap id="T2" position="float">
<label>TABLE 2</label>
<caption>
<p>Overall Performance with respect to RMSE, MAE, MAPE and MSLE. (The smaller value is, the better performance is).</p>
</caption>
<table>
<thead>
<tr>
<td align="left"/>
<td align="center">RMSE</td>
<td align="center">Outperform</td>
<td align="center">MAE</td>
<td align="center">Outperform</td>
<td align="center">MAPE</td>
<td align="center">Outperform</td>
<td align="center">MSLE</td>
<td align="center">Outperform</td>
</tr>
</thead>
<tbody valign="top">
<tr>
<td align="left">AutoFTP</td>
<td align="char" char=".">18.646</td>
<td align="center">&#x2014;</td>
<td align="char" char=".">16.192</td>
<td align="center">&#x2014;</td>
<td align="char" char=".">58.851</td>
<td align="center">&#x2014;</td>
<td align="char" char=".">0.2267</td>
<td align="center">&#x2014;</td>
</tr>
<tr>
<td align="left">AttentionWalk</td>
<td align="char" char=".">21.418</td>
<td align="center">&#x2b;14.9<italic>%</italic>
</td>
<td align="char" char=".">19.712</td>
<td align="center">&#x2b;21.7<italic>%</italic>
</td>
<td align="char" char=".">68.590</td>
<td align="center">&#x2b;16.6<italic>%</italic>
</td>
<td align="char" char=".">0.2907</td>
<td align="center">&#x2b;28.2<italic>%</italic>
</td>
</tr>
<tr>
<td align="left">ProNE</td>
<td align="char" char=".">21.830</td>
<td align="center">&#x2b;17.1<italic>%</italic>
</td>
<td align="char" char=".">19.929</td>
<td align="center">&#x2b;23.1<italic>%</italic>
</td>
<td align="char" char=".">69.188</td>
<td align="center">&#x2b;17.6<italic>%</italic>
</td>
<td align="char" char=".">0.2949</td>
<td align="center">&#x2b;30.1<italic>%</italic>
</td>
</tr>
<tr>
<td align="left">GatNE</td>
<td align="char" char=".">21.229</td>
<td align="center">&#x2b;13.9<italic>%</italic>
</td>
<td align="char" char=".">19.288</td>
<td align="center">&#x2b;19.1<italic>%</italic>
</td>
<td align="char" char=".">67.043</td>
<td align="center">&#x2b;13.9<italic>%</italic>
</td>
<td align="char" char=".">0.2854</td>
<td align="center">&#x2b;25.9<italic>%</italic>
</td>
</tr>
<tr>
<td align="left">GAE</td>
<td align="char" char=".">21.338</td>
<td align="center">&#x2b;14.4<italic>%</italic>
</td>
<td align="char" char=".">19.676</td>
<td align="center">&#x2b;21.5<italic>%</italic>
</td>
<td align="char" char=".">68.579</td>
<td align="center">&#x2b;16.5<italic>%</italic>
</td>
<td align="char" char=".">0.2894</td>
<td align="center">&#x2b;27.6<italic>%</italic>
</td>
</tr>
<tr>
<td align="left">DeepWalk</td>
<td align="char" char=".">23.561</td>
<td align="center">&#x2b;26.4<italic>%</italic>
</td>
<td align="char" char=".">21.987</td>
<td align="center">&#x2b;35.8<italic>%</italic>
</td>
<td align="char" char=".">76.038</td>
<td align="center">&#x2b;29.2<italic>%</italic>
</td>
<td align="char" char=".">0.3321</td>
<td align="center">&#x2b;46.5<italic>%</italic>
</td>
</tr>
<tr>
<td align="left">Node2Vec</td>
<td align="char" char=".">22.688</td>
<td align="center">&#x2b;21.7<italic>%</italic>
</td>
<td align="char" char=".">21.084</td>
<td align="center">&#x2b;30.2<italic>%</italic>
</td>
<td align="char" char=".">73.135</td>
<td align="center">&#x2b;24.3<italic>%</italic>
</td>
<td align="char" char=".">0.3152</td>
<td align="center">&#x2b;39.0<italic>%</italic>
</td>
</tr>
<tr>
<td align="left">Struc2Vec</td>
<td align="char" char=".">21.589</td>
<td align="center">&#x2b;15.8<italic>%</italic>
</td>
<td align="char" char=".">19.937</td>
<td align="center">&#x2b;23.1<italic>%</italic>
</td>
<td align="char" char=".">69.423</td>
<td align="center">&#x2b;17.9<italic>%</italic>
</td>
<td align="char" char=".">0.2942</td>
<td align="center">&#x2b;29.7<italic>%</italic>
</td>
</tr>
<tr>
<td align="left">AutoFTP<sup>
<italic>R</italic>
</sup>
</td>
<td align="char" char=".">21.965</td>
<td align="center">&#x2b;17.8<italic>%</italic>
</td>
<td align="char" char=".">20.283</td>
<td align="center">&#x2b;25.3<italic>%</italic>
</td>
<td align="char" char=".">70.991</td>
<td align="center">&#x2b;20.6<italic>%</italic>
</td>
<td align="char" char=".">0.2928</td>
<td align="center">&#x2b;29.1<italic>%</italic>
</td>
</tr>
<tr>
<td align="left">AutoFTP<sup>(<italic>R</italic>&#x2b;<italic>P</italic>)</sup>
</td>
<td align="char" char=".">20.509</td>
<td align="center">&#x2b;9.99<italic>%</italic>
</td>
<td align="char" char=".">18.921</td>
<td align="center">&#x2b;16.8<italic>%</italic>
</td>
<td align="char" char=".">66.477</td>
<td align="center">&#x2b;12.9<italic>%</italic>
</td>
<td align="char" char=".">0.2681</td>
<td align="center">&#x2b;18.3<italic>%</italic>
</td>
</tr>
<tr>
<td align="left">AutoFTP<sup>(<italic>R</italic>&#x2b;<italic>C</italic>)</sup>
</td>
<td align="char" char=".">21.014</td>
<td align="center">&#x2b;12.7<italic>%</italic>
</td>
<td align="char" char=".">19.413</td>
<td align="center">&#x2b;19.8<italic>%</italic>
</td>
<td align="char" char=".">67.920</td>
<td align="center">&#x2b;15.4<italic>%</italic>
</td>
<td align="char" char=".">0.2773</td>
<td align="center">&#x2b;22.3<italic>%</italic>
</td>
</tr>
<tr>
<td align="left">AutoFTP<sup>(<italic>R</italic>&#x2b;<italic>P</italic>&#x2b;<italic>C</italic>)</sup>
</td>
<td align="char" char=".">20.211</td>
<td align="center">&#x2b;8.39<italic>%</italic>
</td>
<td align="char" char=".">18.676</td>
<td align="center">&#x2b;15.3<italic>%</italic>
</td>
<td align="char" char=".">65.685</td>
<td align="center">&#x2b;11.6<italic>%</italic>
</td>
<td align="char" char=".">0.2636</td>
<td align="center">&#x2b;16.3<italic>%</italic>
</td>
</tr>
</tbody>
</table>
</table-wrap>
<p>In addition, another interesting observation is that among all baseline models, GatNE outperforms others in terms of all evaluation metrics. Such observation shows that GatNE considers different attributed information of nodes in spatial graphs of spatial entities. Thus, the spatial embedding features learned by GatNE are more effective compared with other baseline models. Moreover, after further observing <xref ref-type="table" rid="T2">Table&#x20;2</xref>, we can find that the predictive performances of GAE are better than most random-walk based approaches, except GatNE. Such observation indicates that the graph convolution-based methods (GAE, AutoFTP) are more suitable than the random-walk based approaches (other baselines) in modeling geographical structure information. In summary, the overall performance experiment shows the superiority and effectiveness of AutoFTP compared with other baseline models.</p>
</sec>
<sec id="s4-3">
<title>4.3 Study of AutoFTP Variants (Q2)</title>
<p>To validate the necessity of each loss of AutoFTP, we internally compared the performances of AutoFTP with the performances of the variants of AutoFTP. <xref ref-type="table" rid="T2">Table&#x20;2</xref> shows the ranking orders of the predictive accuracies of the compared methods are: AutoFTP &#x3e; AutoFTP<sup>(<italic>R</italic>&#x2b;<italic>P</italic>&#x2b;<italic>C</italic>)</sup> &#x3e; AutoFTP<sup>(<italic>R</italic>&#x2b;<italic>P</italic>)</sup> &#x3e; AutoFTP<sup>(<italic>R</italic>&#x2b;<italic>C</italic>)</sup> &#x3e; AutoFTP<sup>
<italic>R</italic>
</sup>. A potential interpretation for the observation is that with the increase of optimization objective (loss), AutoFTP captures more characteristics of spatial entities from representation learning, point-wise alignment, pair-wise alignment, and regression task. In addition, compared with AutoFTP<sup>(<italic>R</italic>&#x2b;<italic>P</italic>)</sup> and AutoFTP<sup>(<italic>R</italic>&#x2b;<italic>C</italic>)</sup>, we find that the predictive performance of AutoFTP<sup>(<italic>R</italic>&#x2b;<italic>P</italic>)</sup> is better than AutoFTP<sup>(<italic>R</italic>&#x2b;<italic>C</italic>)</sup>. A plausible reason for the observation is that the features of spatial entities captured by point-wise alignment are more indicative for spatial entities compared with them learned by pair-wise alignment. Moreover, another interesting observation is that AutoFTP outperforms other variants by a large margin. Such observation indicates that the regression loss <inline-formula id="inf66">
<mml:math id="m74">
<mml:msub>
<mml:mrow>
<mml:mi mathvariant="script">L</mml:mi>
</mml:mrow>
<mml:mrow>
<mml:mi>R</mml:mi>
<mml:mi>e</mml:mi>
<mml:mi>g</mml:mi>
</mml:mrow>
</mml:msub>
</mml:math>
</inline-formula> provides a clear optimization direction for AutoFTP, which preserves the features related to the downstream task into spatial embeddings. To sum up, the ablation study experiment demonstrates the four loss functions of AutoFTP are necessary for capturing the representative features in spatial entities during spatial representation learning process.</p>
</sec>
<sec id="s4-4">
<title>4.4 Study of the Interpretability of Spatial Embeddings (Q3)</title>
<p>The space alignment in AutoFTP is implemented from two perspectives: point-wise alignment and pair-wise alignment. The two kinds of alignment make the learned spatial embeddings contain more semantic meaning and interpretability.</p>
<sec id="s4-4-1">
<title>4.4.1 Study of the Point-wise Alignment</title>
<p>To analyze the point-wise alignment, we picked communities (spatial entities) 497, 1,043, 1,126, and 1,232 as examples to plot their corresponding embedding vectors against their corresponding topic vectors. Meanwhile, we extracted the topic names of the most significant 6 topics. <xref ref-type="fig" rid="F8">Figure&#x20;8</xref> shows AutoFTP keeps the point-wise consistency between the semantic feature space and the embedding space. Moreover, the learned spatial embeddings contain abundant semantic meanings. We can infer the urban functions for each community based on <xref ref-type="fig" rid="F8">Figure&#x20;8</xref>. For instance, the community &#x23;497 exhibits high weights on some specific topics, such as, functional facilities, general education, and construction materials. Such observation indicates that this community is probably a large residential area with well-decorated apartments and general education institutions. The community &#x23;1043 and &#x23;1126 all have high weights in entertainment, higher education, parks, etc. We can speculate that they are both residential regions nearby universities. This is because the facilities belonging to these topics indicates the two communities are very likely to be in a college town. For the community &#x23;1232, it exhibits high weights in district, entertainment and convenience related categories. We can infer that the community is a commercial district with many transportation facilities.</p>
<fig id="F8" position="float">
<label>FIGURE 8</label>
<caption>
<p>Illustration of Point-wise Alignment with sample communities.</p>
</caption>
<graphic xlink:href="fdata-04-762899-g008.tif"/>
</fig>
</sec>
<sec id="s4-4-2">
<title>4.4.2 Study of the Pair-wise Alignment</title>
<p>To observe the pair-wise alignment, we visualized the pair-wise topic similarity matrix and pair-wise feature matrix by heat map respectively. As illustrated in <xref ref-type="fig" rid="F9">Figure&#x20;9</xref>, we can find that the two matrices are similar with only minor differences. The observation indicates that the embedding feature space is well-matched with the semantic feature&#x20;space.</p>
<fig id="F9" position="float">
<label>FIGURE 9</label>
<caption>
<p>Illustration of pair-wise alignment.</p>
</caption>
<graphic xlink:href="fdata-04-762899-g009.tif"/>
</fig>
</sec>
<sec id="s4-4-3">
<title>4.4.3 Study of the Interpretability</title>
<p>The results of <xref ref-type="sec" rid="s4-4-1">section 4.4.1</xref> and <xref ref-type="sec" rid="s4-4-2">section 4.4.2</xref> shows that the feature embedding space and the topic semantic embedding space are aligned well. To study the interpretability of spatial embeddings further, we built up a tree model for real estate price prediction and then analyze the feature importance based on the semantic labels of the spatial embeddings. Specially, we exploited a random forest model to predict the real estate price of spatial entities based on the corresponding embeddings. Then, we collected the feature importance of the model as illustrated in <xref ref-type="fig" rid="F10">Figure&#x20;10</xref>. We can find that the semantic labels of top 5 dimensions in the embeddings that affects the real estate price prediction are &#x201c;Entertainment&#x201d;, &#x201c;Transportation&#x201d;, &#x201c;Security&#x201d;, &#x201c;Education&#x201d;, and &#x201c;Business&#x201d;. The three most representative keywords in each semantic label, as shown in <xref ref-type="table" rid="T3">Table&#x20;3</xref>. In common sense, the 5 semantic labels are the most important factors that people consider for buying an estate (<xref ref-type="bibr" rid="B4">Boiko et&#x20;al., 2020</xref>). In other words, they affect the real estate price heavily. Thus, the feature importance analysis experimental results are reasonable. In summary, this experiment validates that AutoFTP can select the most significant topic semantics for feature-topic automatically. In addition, the semantic labels of the spatial embeddings can be regarded as an auxiliary information to improve the interpretability of the embeddings.</p>
<fig id="F10" position="float">
<label>FIGURE 10</label>
<caption>
<p>Feature importance of spatial embeddings.</p>
</caption>
<graphic xlink:href="fdata-04-762899-g010.tif"/>
</fig>
<table-wrap id="T3" position="float">
<label>TABLE 3</label>
<caption>
<p>Semantic labels and top 3 keywords.</p>
</caption>
<table>
<thead>
<tr>
<td align="left">Semantic label</td>
<td align="center">Keywords</td>
</tr>
</thead>
<tbody valign="top">
<tr>
<td align="left">Entertainment</td>
<td align="center">Work Out, Tennis Court, Golf Court</td>
</tr>
<tr>
<td align="left">Transportation Facilities</td>
<td align="center">Highways, High Speed Rail, Bus Stations</td>
</tr>
<tr>
<td align="left">Security</td>
<td align="center">Surveillance, Firefighting, Emergency</td>
</tr>
<tr>
<td align="left">Education</td>
<td align="center">Primary School, High School, University</td>
</tr>
<tr>
<td align="left">Business</td>
<td align="center">Commercial Street, Canal, Satellite City</td>
</tr>
</tbody>
</table>
</table-wrap>
</sec>
</sec>
<sec id="s4-5">
<title>4.5 Robustness Check (Q4)</title>
<p>To evaluate the robustness of AutoFTP, we divided the embeddings into 5 groups (HaiDian, ChongWen, FengTai, ShiJingShan, FangShan) according to the geographical district of spatial entities. <xref ref-type="fig" rid="F11">Figure&#x20;11</xref> shows that AutoFTP consistently outperforms the baselines, and performs more stably than the baselines across the five districts. Such observation indicates that AutoFTP captures the unique local features of different spatial groups. There are two possible reasons for the observation: 1) the semantic alignment of AutoFTP injects the distinct semantic characteristics of spatial entities into the learned embeddings; and 2) the customized regression estimator provides a clear optimization objective for AutoFTP. Overall, the robustness check experiment demonstrates that AutoFTP outperforms other baseline models in not only the global zone but also each local spatial sub-areas.</p>
<fig id="F11" position="float">
<label>FIGURE 11</label>
<caption>
<p>Robustness check according to geographical district.</p>
</caption>
<graphic xlink:href="fdata-04-762899-g011.tif"/>
</fig>
</sec>
<sec id="s4-6">
<title>4.6 Study of the Stability and Sensitivity (Q5)</title>
<p>In this section, we fully evaluated the stability and parameter sensitivity of AutoFTP. We first examined the stability of AutoFTP by analyzing the training losses of AutoFTP and convergence of PSO optimization part. To observe the changing trend of each loss objectively, we scaled the value of losses into [0 &#x223c; 1] and visualized them in <xref ref-type="fig" rid="F12">Figure&#x20;12A</xref>. We can find that all losses (reconstruction loss <inline-formula id="inf67">
<mml:math id="m75">
<mml:msub>
<mml:mrow>
<mml:mi mathvariant="script">L</mml:mi>
</mml:mrow>
<mml:mrow>
<mml:mi>R</mml:mi>
</mml:mrow>
</mml:msub>
</mml:math>
</inline-formula>, regression loss <inline-formula id="inf68">
<mml:math id="m76">
<mml:msub>
<mml:mrow>
<mml:mi mathvariant="script">L</mml:mi>
</mml:mrow>
<mml:mrow>
<mml:mi>R</mml:mi>
<mml:mi>e</mml:mi>
<mml:mi>g</mml:mi>
</mml:mrow>
</mml:msub>
</mml:math>
</inline-formula>, point-wise loss <inline-formula id="inf69">
<mml:math id="m77">
<mml:msub>
<mml:mrow>
<mml:mi mathvariant="script">L</mml:mi>
</mml:mrow>
<mml:mrow>
<mml:mi>P</mml:mi>
</mml:mrow>
</mml:msub>
</mml:math>
</inline-formula>, pair-wise loss <inline-formula id="inf70">
<mml:math id="m78">
<mml:msub>
<mml:mrow>
<mml:mi mathvariant="script">L</mml:mi>
</mml:mrow>
<mml:mrow>
<mml:mi>C</mml:mi>
</mml:mrow>
</mml:msub>
</mml:math>
</inline-formula>) reach convergence over training iterations. Especially, <inline-formula id="inf71">
<mml:math id="m79">
<mml:msub>
<mml:mrow>
<mml:mi mathvariant="script">L</mml:mi>
</mml:mrow>
<mml:mrow>
<mml:mi>R</mml:mi>
</mml:mrow>
</mml:msub>
</mml:math>
</inline-formula> and <inline-formula id="inf72">
<mml:math id="m80">
<mml:msub>
<mml:mrow>
<mml:mi mathvariant="script">L</mml:mi>
</mml:mrow>
<mml:mrow>
<mml:mi>R</mml:mi>
<mml:mi>e</mml:mi>
<mml:mi>g</mml:mi>
</mml:mrow>
</mml:msub>
</mml:math>
</inline-formula> reach equilibrium quickly only after 10 epochs. This observation validates the training stability of AutoFTP. We also analyzed the convergence of PSO. As shown in <xref ref-type="fig" rid="F12">Figure&#x20;12B</xref>, the PSO optimization part reaches convergence after 65 epochs, which further indicates the stable performance of AutoFTP. For the parameter sensitivity evaluation, we investigated the influence of the parameter <italic>K</italic> (the dimension of final embeddings and the number of significant topics) for the model performance and the training time. The same to <xref ref-type="fig" rid="F12">Figure&#x20;12A</xref>, we scaled the value of all metrics into [0 &#x223c; 1] and visualized them in <xref ref-type="fig" rid="F12">Figure&#x20;12C</xref>. We can find that the value of <italic>K</italic> affects the model performance heavily. The observation is reasonable because <italic>K</italic> determines the information content of the final learned embeddings. The plots in <xref ref-type="fig" rid="F12">Figure&#x20;12D</xref> show that the larger <italic>K</italic> is, the longer the training time is. A potential reason for the observation is that the larger <italic>K</italic> means that we need to try more topic subsets for feature-topic pairing.</p>
<fig id="F12" position="float">
<label>FIGURE 12</label>
<caption>
<p>Study the stability and sensitivity of AutoFTP. <bold>(A)</bold> Training Losses. <bold>(B)</bold> Training PSO Cost. <bold>(C)</bold> Parameter Sensitivity. <bold>(D)</bold> K vs. Training Time.</p>
</caption>
<graphic xlink:href="fdata-04-762899-g012.tif"/>
</fig>
</sec>
</sec>
<sec id="s5">
<title>5 Related Work</title>
<p>Graph Representation Learning with Latent Semantics. Graph representation learning refers to techniques that preserve the structural information of a graph into a low-dimensional vector (<xref ref-type="bibr" rid="B38">Wang et&#x20;al., 2016</xref>; <xref ref-type="bibr" rid="B1">Abu-El-Haija et&#x20;al., 2018</xref>; <xref ref-type="bibr" rid="B54">Zhang J.&#x20;et&#x20;al., 2019</xref>; <xref ref-type="bibr" rid="B6">Cen et&#x20;al., 2019</xref>; <xref ref-type="bibr" rid="B46">Wang et&#x20;al., 2020b</xref>). However, owing to traditional graph representation learning models are implemented by deep neural networks, the learned embeddings lack interpretability. Recently, to overcome this limitation, researchers leveraged the texts related to graphs to learn semantically rich representations. For instance, Mai et&#x20;al. implemented an entity retrieval academic search engines that incorporate the text embedding and knowledge graph embedding for accelerating retrieving speed (<xref ref-type="bibr" rid="B25">Mai et&#x20;al., 2018</xref>). Xiao et&#x20;al. improved the semantic meaning of knowledge graph&#x2019;s embedding by integrating both graph triplets and textual descriptions of spatial entities (<xref ref-type="bibr" rid="B50">Xiao et&#x20;al., 2017</xref>). Different from these studies, in this paper, based on spatial entities data composing by spatial graphs and related texts, we propose a new representation learning framework that unifies feature embedding learning and feature-topic pairing together in a closed-loop manner by a PSO based optimization method.</p>
<p>Topic Models in Spatio-temporal Domain. Topic models aim to automatically cluster words and expressions patterns for characterizing documents (<xref ref-type="bibr" rid="B51">Xun et&#x20;al., 2017</xref>; <xref ref-type="bibr" rid="B23">Lee and Kang, 2018</xref>; <xref ref-type="bibr" rid="B15">Hu et&#x20;al., 2019</xref>). Recently, to understand the hidden semantics of spatial entities, many researchers applied topic models in the spatio-temporal data mining domain (<xref ref-type="bibr" rid="B57">Zheng et&#x20;al., 2017</xref>; <xref ref-type="bibr" rid="B17">Huang et&#x20;al., 2019</xref>; <xref ref-type="bibr" rid="B16">Huang et&#x20;al., 2020</xref>). For instance, Zhao et&#x20;al. discovered representative and interpretable human activity patterns from transit data automatically by a spatio-temporal topic model (<xref ref-type="bibr" rid="B56">Zhao et&#x20;al., 2020</xref>). Yao et&#x20;al. tracked spatio-temporal and semantic dynamics of urban geo-topics based on an improved dynamic topic model that embeds spatial factors of pairwise distances between tweets (<xref ref-type="bibr" rid="B53">Yao and Wang, 2020</xref>). These successful applications validate the effectiveness of topic models for extracting semantics in spatio-temporal domains. However, traditional topic models only focus on word frequency in texts but neglect the semantics of words. Recently, the success of many pre-trained language models (<xref ref-type="bibr" rid="B37">Vaswani et&#x20;al., 2017</xref>; <xref ref-type="bibr" rid="B19">Kenton and Toutanova, 2019</xref>; <xref ref-type="bibr" rid="B52">Yang et&#x20;al., 2019</xref>) brings hope for producing more reasonable topic distribution. Thus, in this paper, we employ a pre-trained language model to get the embeddings of keywords and utilize Gaussian Mixture Model to extract topic distribution based on the embeddings.</p>
<p>Explainable Artificial Intelligence (XAI) With artificial intelligence methods are applied in multiple scenarios successfully, how to improve the model explainability becomes a big challenge. In the traditional machine learning domain, researchers employ some simple models that own the explainability naturally such as linear models, decision trees, rule-based models, and etc to explain the modeling process (<xref ref-type="bibr" rid="B5">Burkart and Huber, 2021</xref>; <xref ref-type="bibr" rid="B21">Lakkaraju et&#x20;al., 2016</xref>; <xref ref-type="bibr" rid="B22">Lakkaraju et&#x20;al., 2017</xref>). For instance (<xref ref-type="bibr" rid="B24">Lundberg et&#x20;al., 2020</xref>), improved the global interpretability of tree models by combining many local feature explanations of each prediction and obtained good performance on three medical machine learning problems by applying these models (<xref ref-type="bibr" rid="B41">Wang and Rudin, 2015</xref>). provided a Bayesian framework for learning falling rule lists that do not rely on traditional greedy decision tree learning approaches to improve the explainability of classification models. Although these approaches can improve the model interpretability, the model performance often is sacrificed. Recently, the excellent predictive performance of deep learning models leads the techniques have been applied in many scenarios such as fraud detection, credit evaluation, healthcare, etc. But explainability is the key limitation of the deep learning models. To improve the model explainability, XAI on deep learning attracts much attention from researchers (<xref ref-type="bibr" rid="B11">Gunning, 2017</xref>; <xref ref-type="bibr" rid="B34">Selvaraju et&#x20;al., 2017</xref>; <xref ref-type="bibr" rid="B33">Samek and M&#xfc;ller, 2019</xref>; <xref ref-type="bibr" rid="B2">Agarwal et&#x20;al., 2020</xref>). For instance (<xref ref-type="bibr" rid="B34">Selvaraju et&#x20;al., 2017</xref>), proposed a gradient-weighted class activation mapping method to highlight the import regions in the image for predicting the concept. (<xref ref-type="bibr" rid="B2">Agarwal et&#x20;al., 2020</xref>). proposed neural additive models that learns a linear combination of neural networks for depicting the complex relationships between input features and the output. However, these models focus on studying the relationship between the embeddings and outputs, but cannot provide explicit semantic meanings. Different from these studies, we try to give explicit semantic labels for the learned embeddings through the alignment between the feature embedding space and topic semantic&#x20;space.</p>
<p>Comparison with Prior Literature As an emerging feature extraction technique, deep SRL has demonstrated the power in automated geographic and spatial feature extraction. However, SRL inherits drawbacks of traditional DNNs, such as: the embedding feature space lacks semantic interpretation. Texts can provide more interpretation, but spatial text mining has developed separately. Now, there is cross and increasing interests in both fields to benefit from the advances of the other. Our study targets at an unexplored area at the intersection between representation learning in geospatial data and topic label mining in texts. We develop and formulate a new problem: feature-topic pairing, to address the alignment challenges of the feature embedding space and the semantic topic space. The self-optimizing solution unifies representation learning, topic label selection, feature-topic matching in a PSO framework. This framework can be generalized to other integrated tasks, such as, representation learning integrated with not just topic based selection, but also causal selection, or other constrained selection over features, in various application senarios. This is how this study differentiates from and advances prior literature.</p>
</sec>
<sec id="s6">
<title>6 Conclusion</title>
<p>We presented a novel spatial representation learning (SRL) framework, namely AutoFTP. The spatial embeddings produced by traditional SRL models lack semantic meaning. To overcome this limitation, we formulated the feature-topic paring problem. We proposed a novel deep learning framework to unify representation learning, topic label selection, and feature-topic pairing. Specifically, we designed a segmentation-embedding-clustering method to generate candidate feature topic labels from texts. We developed an integrated measurement to measure the pointwise and pairwise alignment between topic label and embedding feature space. We devised a PSO based optimization algorithm to effectively solve the joint task of feature learning and feature-topic pairing. Our method integrated spatial graphs and associated texts to learn effective embedding features with visible labels. Extensive experiments demonstrated the effectiveness of AutoFTP by comparing it with other baseline models. The topic labels of the learned features were shown by many case studies and the feature importance analysis of a downstream task. For future work, we plan to extend our approach from geospatial networks to other applications that consist of graphs and texts, such as social media and software code safety.</p>
</sec>
</body>
<back>
<sec id="s7">
<title>Data Availability Statement</title>
<p>Publicly available datasets were analyzed in this study. This data can be found here: <ext-link ext-link-type="uri" xlink:href="https://www.dropbox.com/sh/woqh4qvuzq1788r/AAB5Vz1DSeJiLKxq-POHLMAVa?dl=0">https://www.dropbox.com/sh/woqh4qvuzq1788r/AAB5Vz1DSeJiLKxq-POHLMAVa?dl&#x3d;0</ext-link>.</p>
</sec>
<sec id="s8">
<title>Author Contributions</title>
<p>DW proposes the main idea, finishes major experiments, writes the paper. KL helps accomplish partial experiments and writes some paragraphs in the paper. DM helps improve the presentation of the paper PW helps modify some typos and errors in the paper C-TL improves the presentation and language of the paper YF improves the presentation of the paper and provides the experimental data and devices.</p>
</sec>
<sec id="s9">
<title>Funding</title>
<p>This research was partially supported by the National Science Foundation (NSF) via the grant numbers: 1755946, 2040950, 2006889, 2045567, 2141095.</p>
</sec>
<sec sec-type="COI-statement" id="s10">
<title>Conflict of Interest</title>
<p>The authors declare that the research was conducted in the absence of any commercial or financial relationships that could be construed as a potential conflict of interest.</p>
</sec>
<sec sec-type="disclaimer" id="s11">
<title>Publisher&#x2019;s Note</title>
<p>All claims expressed in this article are solely those of the authors and do not necessarily represent those of their affiliated organizations, or those of the publisher, the editors and the reviewers. Any product that may be evaluated in this article, or claim that may be made by its manufacturer, is not guaranteed or endorsed by the publisher.</p>
</sec>
<fn-group>
<fn id="FN1">
<label>1</label>
<p>
<ext-link ext-link-type="uri" xlink:href="https://www.brookings.edu/blog/techtank/2016/08/18/autopilot-fatality-reveals-risks-of-technology-testing/">https://www.brookings.edu/blog/techtank/2016/08/18/autopilot-fatality-reveals-risks-of-technology-testing/</ext-link>
</p>
</fn>
<fn id="FN2">
<label>2</label>
<p>
<ext-link ext-link-type="uri" xlink:href="https://www.wsj.com/articles/BL-DGB-42522">https://www.wsj.com/articles/BL-DGB-42522</ext-link>
</p>
</fn>
<fn id="FN3">
<label>3</label>
<p>
<ext-link ext-link-type="uri" xlink:href="https://www.propublica.org/article/machine-bias-risk-assessments-in-criminal-sentencing">https://www.propublica.org/article/machine-bias-risk-assessments-in-criminal-sentencing</ext-link>
</p>
</fn>
<fn id="FN4">
<label>4</label>
<p>
<ext-link ext-link-type="uri" xlink:href="https://www.dropbox.com/sh/cst7yegcp1yvwax/AAB_3xBtgGDod10ULgncI-4La?dl=0">https://www.dropbox.com/sh/cst7yegcp1yvwax/AAB_3xBtgGDod10ULgncI-4La?dl&#x3d;0</ext-link>
</p>
</fn>
</fn-group>
<ref-list>
<title>References</title>
<ref id="B1">
<citation citation-type="journal">
<person-group person-group-type="author">
<name>
<surname>Abu-El-Haija</surname>
<given-names>S.</given-names>
</name>
<name>
<surname>Perozzi</surname>
<given-names>B.</given-names>
</name>
<name>
<surname>Al-Rfou</surname>
<given-names>R.</given-names>
</name>
<name>
<surname>Alemi</surname>
<given-names>A. A.</given-names>
</name>
</person-group> (<year>2018</year>). <article-title>Watch Your Step: Learning Node Embeddings via Graph Attention</article-title>. <source>Adv. Neural Inf. Process. Syst.</source> <volume>31</volume>, <fpage>9180</fpage>&#x2013;<lpage>9190</lpage>. </citation>
</ref>
<ref id="B2">
<citation citation-type="book">
<person-group person-group-type="author">
<name>
<surname>Agarwal</surname>
<given-names>R.</given-names>
</name>
<name>
<surname>Frosst</surname>
<given-names>N.</given-names>
</name>
<name>
<surname>Zhang</surname>
<given-names>X.</given-names>
</name>
<name>
<surname>Caruana</surname>
<given-names>R.</given-names>
</name>
<name>
<surname>Hinton</surname>
<given-names>G. E.</given-names>
</name>
</person-group> (<year>2020</year>). <source>Neural Additive Models: Interpretable Machine Learning with Neural Nets</source>. </citation>
</ref>
<ref id="B3">
<citation citation-type="book">
<person-group person-group-type="author">
<name>
<surname>Blei</surname>
<given-names>D.</given-names>
</name>
<name>
<surname>Ng</surname>
<given-names>A.</given-names>
</name>
<name>
<surname>Jordan</surname>
<given-names>M.</given-names>
</name>
</person-group> (<year>2003</year>). <source>Latent Dirichlet Allocation Journal of Machine Learning Research</source>.<issue>3</issue> </citation>
</ref>
<ref id="B4">
<citation citation-type="book">
<person-group person-group-type="author">
<name>
<surname>Boiko</surname>
<given-names>D.</given-names>
</name>
<name>
<surname>Parygin</surname>
<given-names>D.</given-names>
</name>
<name>
<surname>Savina</surname>
<given-names>O.</given-names>
</name>
<name>
<surname>Golubev</surname>
<given-names>A.</given-names>
</name>
<name>
<surname>Zelenskiy</surname>
<given-names>I.</given-names>
</name>
<name>
<surname>Mityagin</surname>
<given-names>S.</given-names>
</name>
</person-group> (<year>2020</year>). &#x201c;<article-title>Approaches to Analysis of Factors Affecting the Residential Real Estate Bid Prices in Case of Open Data Use</article-title>,&#x201d; in <source>Electronic Governance and Open Society: Challenges in Eurasia</source>. Editors <person-group person-group-type="editor">
<name>
<surname>Chugunov</surname>
<given-names>A.</given-names>
</name>
<name>
<surname>Khodachek</surname>
<given-names>I.</given-names>
</name>
<name>
<surname>Misnikov</surname>
<given-names>Y.</given-names>
</name>
<name>
<surname>Trutnev</surname>
<given-names>D.</given-names>
</name>
</person-group> (<publisher-loc>Cham</publisher-loc>: <publisher-name>Springer International Publishing</publisher-name>), <fpage>360</fpage>&#x2013;<lpage>375</lpage>. <pub-id pub-id-type="doi">10.1007/978-3-030-39296-3_27</pub-id> </citation>
</ref>
<ref id="B5">
<citation citation-type="journal">
<person-group person-group-type="author">
<name>
<surname>Burkart</surname>
<given-names>N.</given-names>
</name>
<name>
<surname>Huber</surname>
<given-names>M. F.</given-names>
</name>
</person-group> (<year>2021</year>). <article-title>A Survey on the Explainability of Supervised Machine Learning</article-title>. <source>jair</source> <volume>70</volume>, <fpage>245</fpage>&#x2013;<lpage>317</lpage>. <pub-id pub-id-type="doi">10.1613/jair.1.12228</pub-id> </citation>
</ref>
<ref id="B6">
<citation citation-type="confproc">
<person-group person-group-type="author">
<name>
<surname>Cen</surname>
<given-names>Y.</given-names>
</name>
<name>
<surname>Zou</surname>
<given-names>X.</given-names>
</name>
<name>
<surname>Zhang</surname>
<given-names>J.</given-names>
</name>
<name>
<surname>Yang</surname>
<given-names>H.</given-names>
</name>
<name>
<surname>Zhou</surname>
<given-names>J.</given-names>
</name>
<name>
<surname>Tang</surname>
<given-names>J.</given-names>
</name>
</person-group> (<year>2019</year>). &#x201c;<article-title>Representation Learning for Attributed Multiplex Heterogeneous Network</article-title>,&#x201d; in <conf-name>Proceedings of the 25th ACM SIGKDD International Conference on Knowledge Discovery &#x26; Data Mining</conf-name>, <fpage>1358</fpage>&#x2013;<lpage>1368</lpage>. <pub-id pub-id-type="doi">10.1145/3292500.3330964</pub-id> </citation>
</ref>
<ref id="B7">
<citation citation-type="confproc">
<person-group person-group-type="author">
<name>
<surname>Chandra</surname>
<given-names>D. K.</given-names>
</name>
<name>
<surname>Wang</surname>
<given-names>P.</given-names>
</name>
<name>
<surname>Leopold</surname>
<given-names>J.</given-names>
</name>
<name>
<surname>Fu</surname>
<given-names>Y.</given-names>
</name>
</person-group> (<year>2019</year>). &#x201c;<article-title>Collective Representation Learning on Spatiotemporal Heterogeneous Information Networks</article-title>,&#x201d; in <conf-name>Proceedings of the 27th ACM SIGSPATIAL International Conference on Advances in Geographic Information Systems</conf-name>, <fpage>319</fpage>&#x2013;<lpage>328</lpage>. <pub-id pub-id-type="doi">10.1145/3347146.3359104</pub-id> </citation>
</ref>
<ref id="B8">
<citation citation-type="book">
<person-group person-group-type="author">
<name>
<surname>Doshi-Velez</surname>
<given-names>F.</given-names>
</name>
<name>
<surname>Kim</surname>
<given-names>B.</given-names>
</name>
</person-group> (<year>2017</year>). <source>Towards a Rigorous Science of Interpretable Machine Learning</source>. <comment>arXiv preprint arXiv:1702.08608</comment>. </citation>
</ref>
<ref id="B9">
<citation citation-type="journal">
<person-group person-group-type="author">
<name>
<surname>Elshawi</surname>
<given-names>R.</given-names>
</name>
<name>
<surname>Al-Mallah</surname>
<given-names>M. H.</given-names>
</name>
<name>
<surname>Sakr</surname>
<given-names>S.</given-names>
</name>
</person-group> (<year>2019</year>). <article-title>On the Interpretability of Machine Learning-Based Model for Predicting Hypertension</article-title>. <source>BMC Med. Inform. Decis. Mak</source>. <volume>19</volume>, <fpage>146</fpage>&#x2013;<lpage>232</lpage>. <pub-id pub-id-type="doi">10.1186/s12911-019-0874-0</pub-id> </citation>
</ref>
<ref id="B10">
<citation citation-type="confproc">
<person-group person-group-type="author">
<name>
<surname>Grover</surname>
<given-names>A.</given-names>
</name>
<name>
<surname>Leskovec</surname>
<given-names>J.</given-names>
</name>
</person-group> (<year>2016</year>). &#x201c;<article-title>node2vec: Scalable Feature Learning for Networks</article-title>,&#x201d; in <conf-name>Proceedings of the 22nd ACM SIGKDD international conference on Knowledge discovery and data mining</conf-name>, <conf-loc>New York, NY</conf-loc>, <conf-date>August 2016</conf-date>, <fpage>855</fpage>&#x2013;<lpage>864</lpage>. <pub-id pub-id-type="doi">10.1145/2939672.2939754</pub-id>
<source>KDD</source>
<volume>2016</volume> </citation>
</ref>
<ref id="B11">
<citation citation-type="book">
<person-group person-group-type="author">
<name>
<surname>Gunning</surname>
<given-names>D.</given-names>
</name>
</person-group> (<year>2017</year>). &#x201c;<article-title>Explainable Artificial Intelligence (Xai)</article-title>,&#x201d; in <source>
<italic>Defense Advanced Research Projects Agency (DARPA), Nd Web</italic> 2</source>. </citation>
</ref>
<ref id="B12">
<citation citation-type="book">
<person-group person-group-type="author">
<name>
<surname>He</surname>
<given-names>H.</given-names>
</name>
</person-group> (<year>2014</year>). <source>HanLP: Han Language Processing</source>. [<comment>Dataset</comment>]. </citation>
</ref>
<ref id="B13">
<citation citation-type="book">
<person-group person-group-type="author">
<name>
<surname>Hofmann</surname>
<given-names>T.</given-names>
</name>
</person-group> (<year>2013</year>). <source>Probabilistic Latent Semantic Analysis</source>. <comment>arXiv preprint arXiv:1301.6705</comment>. </citation>
</ref>
<ref id="B14">
<citation citation-type="journal">
<person-group person-group-type="author">
<name>
<surname>Hong</surname>
<given-names>S. R.</given-names>
</name>
<name>
<surname>Hullman</surname>
<given-names>J.</given-names>
</name>
<name>
<surname>Bertini</surname>
<given-names>E.</given-names>
</name>
</person-group> (<year>2020</year>). <article-title>Human Factors in Model Interpretability: Industry Practices, Challenges, and Needs</article-title>. <source>Proc. ACM Hum.-Comput. Interact.</source> <volume>4</volume>, <fpage>1</fpage>&#x2013;<lpage>26</lpage>. <pub-id pub-id-type="doi">10.1145/3392878</pub-id> </citation>
</ref>
<ref id="B15">
<citation citation-type="journal">
<person-group person-group-type="author">
<name>
<surname>Hu</surname>
<given-names>N.</given-names>
</name>
<name>
<surname>Zhang</surname>
<given-names>T.</given-names>
</name>
<name>
<surname>Gao</surname>
<given-names>B.</given-names>
</name>
<name>
<surname>Bose</surname>
<given-names>I.</given-names>
</name>
</person-group> (<year>2019</year>). <article-title>What Do Hotel Customers Complain about? Text Analysis Using Structural Topic Model</article-title>. <source>Tourism Manage.</source> <volume>72</volume>, <fpage>417</fpage>&#x2013;<lpage>426</lpage>. <pub-id pub-id-type="doi">10.1016/j.tourman.2019.01.002</pub-id> </citation>
</ref>
<ref id="B16">
<citation citation-type="journal">
<person-group person-group-type="author">
<name>
<surname>Huang</surname>
<given-names>L.</given-names>
</name>
<name>
<surname>Wen</surname>
<given-names>Y.</given-names>
</name>
<name>
<surname>Guo</surname>
<given-names>W.</given-names>
</name>
<name>
<surname>Zhu</surname>
<given-names>X.</given-names>
</name>
<name>
<surname>Zhou</surname>
<given-names>C.</given-names>
</name>
<name>
<surname>Zhang</surname>
<given-names>F.</given-names>
</name>
<etal/>
</person-group> (<year>2020</year>). <article-title>Mobility Pattern Analysis of Ship Trajectories Based on Semantic Transformation and Topic Model</article-title>. <source>Ocean Eng.</source> <volume>201</volume>, <fpage>107092</fpage>. <pub-id pub-id-type="doi">10.1016/j.oceaneng.2020.107092</pub-id> </citation>
</ref>
<ref id="B17">
<citation citation-type="journal">
<person-group person-group-type="author">
<name>
<surname>Huang</surname>
<given-names>Q.</given-names>
</name>
<name>
<surname>Huang</surname>
<given-names>C.</given-names>
</name>
<name>
<surname>Huang</surname>
<given-names>J.</given-names>
</name>
<name>
<surname>Fujita</surname>
<given-names>H.</given-names>
</name>
</person-group> (<year>2019</year>). <article-title>Adaptive Resource Prefetching with Spatial-Temporal and Topic Information for Educational Cloud Storage Systems</article-title>. <source>Knowledge-Based Syst.</source> <volume>181</volume>, <fpage>104791</fpage>. <pub-id pub-id-type="doi">10.1016/j.knosys.2019.05.034</pub-id> </citation>
</ref>
<ref id="B18">
<citation citation-type="confproc">
<person-group person-group-type="author">
<name>
<surname>Jean</surname>
<given-names>N.</given-names>
</name>
<name>
<surname>Wang</surname>
<given-names>S.</given-names>
</name>
<name>
<surname>Samar</surname>
<given-names>A.</given-names>
</name>
<name>
<surname>Azzari</surname>
<given-names>G.</given-names>
</name>
<name>
<surname>Lobell</surname>
<given-names>D.</given-names>
</name>
<name>
<surname>Ermon</surname>
<given-names>S.</given-names>
</name>
</person-group> (<year>2019</year>). &#x201c;<article-title>Tile2vec: Unsupervised Representation Learning for Spatially Distributed Data</article-title>,&#x201d; in <conf-name>Proceedings of the AAAI Conference on Artificial Intelligence</conf-name>, <volume>33</volume>. <fpage>3967</fpage>&#x2013;<lpage>3974</lpage>. <pub-id pub-id-type="doi">10.1609/aaai.v33i01.33013967</pub-id> </citation>
</ref>
<ref id="B19">
<citation citation-type="confproc">
<person-group person-group-type="author">
<name>
<surname>Kenton</surname>
<given-names>J.&#x20;D. M.-W. C.</given-names>
</name>
<name>
<surname>Toutanova</surname>
<given-names>L. K.</given-names>
</name>
</person-group> (<year>2019</year>). &#x201c;<article-title>Bert: Pre-training of Deep Bidirectional Transformers for Language Understanding</article-title>,&#x201d; in <conf-name>Proceedings of NAACL-HLT</conf-name>, <fpage>4171</fpage>&#x2013;<lpage>4186</lpage>. </citation>
</ref>
<ref id="B20">
<citation citation-type="book">
<person-group person-group-type="author">
<name>
<surname>Kipf</surname>
<given-names>T. N.</given-names>
</name>
<name>
<surname>Welling</surname>
<given-names>M.</given-names>
</name>
</person-group> (<year>2016</year>). <source>Variational Graph Auto-Encoders</source>. <comment>arXiv preprint arXiv:1611.07308</comment>. </citation>
</ref>
<ref id="B21">
<citation citation-type="confproc">
<person-group person-group-type="author">
<name>
<surname>Lakkaraju</surname>
<given-names>H.</given-names>
</name>
<name>
<surname>Bach</surname>
<given-names>S. H.</given-names>
</name>
<name>
<surname>Jure</surname>
<given-names>L.</given-names>
</name>
</person-group> (<year>2016</year>). &#x201c;<article-title>Interpretable Decision Sets: A Joint Framework for Description and Prediction</article-title>,&#x201d; in <conf-name>Proceedings of the 22nd ACM SIGKDD international conference on knowledge discovery and data mining</conf-name>, <volume>2016</volume>. <fpage>1675</fpage>&#x2013;<lpage>1684</lpage>. <pub-id pub-id-type="doi">10.1145/2939672.2939874</pub-id> </citation>
</ref>
<ref id="B22">
<citation citation-type="book">
<person-group person-group-type="author">
<name>
<surname>Lakkaraju</surname>
<given-names>H.</given-names>
</name>
<name>
<surname>Kamar</surname>
<given-names>E.</given-names>
</name>
<name>
<surname>Caruana</surname>
<given-names>R.</given-names>
</name>
<name>
<surname>Leskovec</surname>
<given-names>J.</given-names>
</name>
</person-group> (<year>2017</year>). <source>Interpretable &#x26; Explorable Approximations of Black Box Models</source>. <comment>arXiv preprint arXiv:1707.01154</comment>. </citation>
</ref>
<ref id="B23">
<citation citation-type="journal">
<person-group person-group-type="author">
<name>
<surname>Lee</surname>
<given-names>H.</given-names>
</name>
<name>
<surname>Kang</surname>
<given-names>P.</given-names>
</name>
</person-group> (<year>2018</year>). <article-title>Identifying Core Topics in Technology and Innovation Management Studies: A Topic Model Approach</article-title>. <source>J.&#x20;Technol. Transf</source>. <volume>43</volume>, <fpage>1291</fpage>&#x2013;<lpage>1317</lpage>. <pub-id pub-id-type="doi">10.1007/s10961-017-9561-4</pub-id> </citation>
</ref>
<ref id="B24">
<citation citation-type="journal">
<person-group person-group-type="author">
<name>
<surname>Lundberg</surname>
<given-names>S. M.</given-names>
</name>
<name>
<surname>Erion</surname>
<given-names>G.</given-names>
</name>
<name>
<surname>Chen</surname>
<given-names>H.</given-names>
</name>
<name>
<surname>DeGrave</surname>
<given-names>A.</given-names>
</name>
<name>
<surname>Prutkin</surname>
<given-names>J.&#x20;M.</given-names>
</name>
<name>
<surname>Nair</surname>
<given-names>B.</given-names>
</name>
<etal/>
</person-group> (<year>2020</year>). <article-title>From Local Explanations to Global Understanding with Explainable Ai for Trees</article-title>. <source>Nat. Mach. Intell.</source> <volume>2</volume>, <fpage>56</fpage>&#x2013;<lpage>67</lpage>. <pub-id pub-id-type="doi">10.1038/s42256-019-0138-9</pub-id> </citation>
</ref>
<ref id="B25">
<citation citation-type="book">
<person-group person-group-type="author">
<name>
<surname>Mai</surname>
<given-names>G.</given-names>
</name>
<name>
<surname>Janowicz</surname>
<given-names>K.</given-names>
</name>
<name>
<surname>Yan</surname>
<given-names>B.</given-names>
</name>
</person-group> (<year>2018</year>). &#x201c;<article-title>Combining Text Embedding and Knowledge Graph Embedding Techniques for Academic Search Engines</article-title>,&#x201d; in <source>Semdeep/NLIWoD@ ISWC</source>, <fpage>77</fpage>&#x2013;<lpage>88</lpage>. </citation>
</ref>
<ref id="B26">
<citation citation-type="confproc">
<person-group person-group-type="author">
<name>
<surname>Mihalcea</surname>
<given-names>R.</given-names>
</name>
<name>
<surname>Tarau</surname>
<given-names>P.</given-names>
</name>
</person-group> (<year>2004</year>). &#x201c;<article-title>Textrank: Bringing Order into Text</article-title>,&#x201d; in <conf-name>Proceedings of the 2004 conference on empirical methods in natural language processing</conf-name>. </citation>
</ref>
<ref id="B27">
<citation citation-type="book">
<person-group person-group-type="author">
<name>
<surname>Mikolov</surname>
<given-names>T.</given-names>
</name>
<name>
<surname>Sutskever</surname>
<given-names>I.</given-names>
</name>
<name>
<surname>Chen</surname>
<given-names>K.</given-names>
</name>
<name>
<surname>Corrado</surname>
<given-names>G. S.</given-names>
</name>
<name>
<surname>Dean</surname>
<given-names>J.</given-names>
</name>
</person-group> (<year>2013</year>). &#x201c;<article-title>Distributed Representations of Words and Phrases and Their Compositionality</article-title>,&#x201d; in <source>Advances in Neural Information Processing Systems</source>, <fpage>3111</fpage>&#x2013;<lpage>3119</lpage>. </citation>
</ref>
<ref id="B28">
<citation citation-type="journal">
<person-group person-group-type="author">
<name>
<surname>Mori</surname>
<given-names>T.</given-names>
</name>
<name>
<surname>Uchihira</surname>
<given-names>N.</given-names>
</name>
</person-group> (<year>2019</year>). <article-title>Balancing the Trade-Off between Accuracy and Interpretability in Software Defect Prediction</article-title>. <source>Empir. Softw. Eng</source>. <volume>24</volume>, <fpage>779</fpage>&#x2013;<lpage>825</lpage>. <pub-id pub-id-type="doi">10.1007/s10664-018-9638-1</pub-id> </citation>
</ref>
<ref id="B29">
<citation citation-type="confproc">
<person-group person-group-type="author">
<name>
<surname>Perozzi</surname>
<given-names>B.</given-names>
</name>
<name>
<surname>Al-Rfou</surname>
<given-names>R.</given-names>
</name>
<name>
<surname>Skiena</surname>
<given-names>S.</given-names>
</name>
</person-group> (<year>2014</year>). &#x201c;<article-title>Deepwalk: Online Learning of Social Representations</article-title>,&#x201d; in <conf-name>Proceedings of the 20th ACM SIGKDD international conference on Knowledge discovery and data mining</conf-name> (<publisher-loc>New York</publisher-loc>: <publisher-name>ACM</publisher-name>), <fpage>701</fpage>&#x2013;<lpage>710</lpage>. </citation>
</ref>
<ref id="B30">
<citation citation-type="confproc">
<person-group person-group-type="author">
<name>
<surname>Poursabzi-Sangdeh</surname>
<given-names>F.</given-names>
</name>
<name>
<surname>Goldstein</surname>
<given-names>D. G.</given-names>
</name>
<name>
<surname>Hofman</surname>
<given-names>J.&#x20;M.</given-names>
</name>
<name>
<surname>Wortman Vaughan</surname>
<given-names>J.&#x20;W.</given-names>
</name>
<name>
<surname>Wallach</surname>
<given-names>H.</given-names>
</name>
</person-group> (<year>2021</year>). &#x201c;<article-title>Manipulating and Measuring Model Interpretability</article-title>,&#x201d; in <conf-name>Proceedings of the 2021 CHI Conference on Human Factors in Computing Systems</conf-name>, <fpage>1</fpage>&#x2013;<lpage>52</lpage>. <pub-id pub-id-type="doi">10.1145/3411764.3445315</pub-id> </citation>
</ref>
<ref id="B31">
<citation citation-type="confproc">
<person-group person-group-type="author">
<name>
<surname>Ribeiro</surname>
<given-names>L. F.</given-names>
</name>
<name>
<surname>Saverese</surname>
<given-names>P. H.</given-names>
</name>
<name>
<surname>Figueiredo</surname>
<given-names>D. R.</given-names>
</name>
</person-group> (<year>2017</year>). &#x201c;<article-title>struc2vec: Learning Node Representations from Structural Identity</article-title>,&#x201d; in <conf-name>Proceedings of the 23rd ACM SIGKDD international conference on knowledge discovery and data mining</conf-name>, <fpage>385</fpage>&#x2013;<lpage>394</lpage>. </citation>
</ref>
<ref id="B32">
<citation citation-type="confproc">
<person-group person-group-type="author">
<name>
<surname>Saisubramanian</surname>
<given-names>S.</given-names>
</name>
<name>
<surname>Galhotra</surname>
<given-names>S.</given-names>
</name>
<name>
<surname>Zilberstein</surname>
<given-names>S.</given-names>
</name>
</person-group> (<year>2020</year>). &#x201c;<article-title>Balancing the Tradeoff between Clustering Value and Interpretability</article-title>,&#x201d; in <conf-name>Proceedings of the AAAI/ACM Conference on AI, Ethics and Society</conf-name> (<publisher-loc>New York</publisher-loc>: <publisher-name>ACM</publisher-name>), <fpage>351</fpage>&#x2013;<lpage>357</lpage>. <pub-id pub-id-type="doi">10.1145/3375627.3375843</pub-id> </citation>
</ref>
<ref id="B33">
<citation citation-type="confproc">
<person-group person-group-type="author">
<name>
<surname>Samek</surname>
<given-names>W.</given-names>
</name>
<name>
<surname>M&#xfc;ller</surname>
<given-names>K.-R.</given-names>
</name>
</person-group> (<year>2019</year>). &#x201c;<article-title>Towards Explainable Artificial Intelligence</article-title>,&#x201d; in <conf-name>Explainable AI: interpreting, explaining and visualizing deep learning</conf-name> (<publisher-name>Springer</publisher-name>), <fpage>5</fpage>&#x2013;<lpage>22</lpage>. <pub-id pub-id-type="doi">10.1007/978-3-030-28954-6_1</pub-id> </citation>
</ref>
<ref id="B34">
<citation citation-type="confproc">
<person-group person-group-type="author">
<name>
<surname>Selvaraju</surname>
<given-names>R. R.</given-names>
</name>
<name>
<surname>Cogswell</surname>
<given-names>M.</given-names>
</name>
<name>
<surname>Das</surname>
<given-names>A.</given-names>
</name>
<name>
<surname>Vedantam</surname>
<given-names>R.</given-names>
</name>
<name>
<surname>Parikh</surname>
<given-names>D.</given-names>
</name>
<name>
<surname>Batra</surname>
<given-names>D.</given-names>
</name>
</person-group> (<year>2017</year>). &#x201c;<article-title>Grad-cam: Visual Explanations from Deep Networks via Gradient-Based Localization</article-title>,&#x201d; in <conf-name>Proceedings of the IEEE international conference on computer vision</conf-name>, <fpage>618</fpage>&#x2013;<lpage>626</lpage>. <pub-id pub-id-type="doi">10.1109/iccv.2017.74</pub-id> </citation>
</ref>
<ref id="B35">
<citation citation-type="journal">
<person-group person-group-type="author">
<name>
<surname>Shan</surname>
<given-names>S.</given-names>
</name>
<name>
<surname>Li</surname>
<given-names>Z.</given-names>
</name>
<name>
<surname>Yang</surname>
<given-names>Q.</given-names>
</name>
<name>
<surname>Liu</surname>
<given-names>A.</given-names>
</name>
<name>
<surname>Zhao</surname>
<given-names>L.</given-names>
</name>
<name>
<surname>Liu</surname>
<given-names>G.</given-names>
</name>
<etal/>
</person-group> (<year>2020</year>). <article-title>Geographical Address Representation Learning for Address Matching</article-title>. <source>World Wide Web</source> <volume>23</volume>, <fpage>2005</fpage>&#x2013;<lpage>2022</lpage>. <pub-id pub-id-type="doi">10.1007/s11280-020-00782-2</pub-id> </citation>
</ref>
<ref id="B36">
<citation citation-type="journal">
<person-group person-group-type="author">
<name>
<surname>Stiglic</surname>
<given-names>G.</given-names>
</name>
<name>
<surname>Kocbek</surname>
<given-names>P.</given-names>
</name>
<name>
<surname>Fijacko</surname>
<given-names>N.</given-names>
</name>
<name>
<surname>Zitnik</surname>
<given-names>M.</given-names>
</name>
<name>
<surname>Verbert</surname>
<given-names>K.</given-names>
</name>
<name>
<surname>Cilar</surname>
<given-names>L.</given-names>
</name>
</person-group> (<year>2020</year>). <article-title>Interpretability of Machine Learning-Based Prediction Models in Healthcare</article-title>. <source>Wiley Interdiscip. Rev. Data Mining Knowledge Discov.</source> <volume>10</volume>, <fpage>e1379</fpage>. <pub-id pub-id-type="doi">10.1002/widm.1379</pub-id> </citation>
</ref>
<ref id="B37">
<citation citation-type="confproc">
<person-group person-group-type="author">
<name>
<surname>Vaswani</surname>
<given-names>A.</given-names>
</name>
<name>
<surname>Shazeer</surname>
<given-names>N.</given-names>
</name>
<name>
<surname>Parmar</surname>
<given-names>N.</given-names>
</name>
<name>
<surname>Uszkoreit</surname>
<given-names>J.</given-names>
</name>
<name>
<surname>Jones</surname>
<given-names>L.</given-names>
</name>
<name>
<surname>Gomez</surname>
<given-names>A. N.</given-names>
</name>
<etal/>
</person-group> (<year>2017</year>). &#x201c;<article-title>Attention Is All You Need</article-title>,&#x201d; in <conf-name>Proceedings of the 31st International Conference on Neural Information Processing Systems</conf-name>, <fpage>6000</fpage>&#x2013;<lpage>6010</lpage>. </citation>
</ref>
<ref id="B38">
<citation citation-type="confproc">
<person-group person-group-type="author">
<name>
<surname>Wang</surname>
<given-names>D.</given-names>
</name>
<name>
<surname>Cui</surname>
<given-names>P.</given-names>
</name>
<name>
<surname>Zhu</surname>
<given-names>W.</given-names>
</name>
</person-group> (<year>2016</year>). &#x201c;<article-title>Structural Deep Network Embedding</article-title>,&#x201d; in <conf-name>Proceedings of the 22nd ACM SIGKDD International Conference on Knowledge Discovery and Data Mining</conf-name> (<publisher-loc>New York</publisher-loc>: <publisher-name>ACM</publisher-name>), <fpage>1225</fpage>&#x2013;<lpage>1234</lpage>. <pub-id pub-id-type="doi">10.1145/2939672.2939753</pub-id> </citation>
</ref>
<ref id="B39">
<citation citation-type="confproc">
<person-group person-group-type="author">
<name>
<surname>Wang</surname>
<given-names>D.</given-names>
</name>
<name>
<surname>Wang</surname>
<given-names>P.</given-names>
</name>
<name>
<surname>Liu</surname>
<given-names>K.</given-names>
</name>
<name>
<surname>Zhou</surname>
<given-names>Y.</given-names>
</name>
<name>
<surname>Hughes</surname>
<given-names>C. E.</given-names>
</name>
<name>
<surname>Fu</surname>
<given-names>Y.</given-names>
</name>
</person-group> (<year>2021</year>). &#x201c;<article-title>Reinforced Imitative Graph Representation Learning for mobile User Profiling: An Adversarial Training Perspective</article-title>,&#x201d; in <conf-name>Proceedings of the AAAI Conference on Artificial Intelligence</conf-name>, <volume>35</volume>. <fpage>4410</fpage>&#x2013;<lpage>4417</lpage>. </citation>
</ref>
<ref id="B40">
<citation citation-type="confproc">
<person-group person-group-type="author">
<name>
<surname>Wang</surname>
<given-names>D.</given-names>
</name>
<name>
<surname>Wang</surname>
<given-names>P.</given-names>
</name>
<name>
<surname>Zhou</surname>
<given-names>J.</given-names>
</name>
<name>
<surname>Sun</surname>
<given-names>L.</given-names>
</name>
<name>
<surname>Du</surname>
<given-names>B.</given-names>
</name>
<name>
<surname>Fu</surname>
<given-names>Y.</given-names>
</name>
</person-group> (<year>2020a</year>). &#x201c;<article-title>Defending Water Treatment Networks: Exploiting Spatio-Temporal Effects for Cyber Attack Detection</article-title>,&#x201d; in <conf-name>2020 IEEE International Conference on Data Mining (ICDM)</conf-name> (<publisher-name>IEEE</publisher-name>), <fpage>32</fpage>&#x2013;<lpage>41</lpage>. <pub-id pub-id-type="doi">10.1109/icdm50108.2020.00012</pub-id> </citation>
</ref>
<ref id="B41">
<citation citation-type="confproc">
<person-group person-group-type="author">
<name>
<surname>Wang</surname>
<given-names>F.</given-names>
</name>
<name>
<surname>Rudin</surname>
<given-names>C.</given-names>
</name>
</person-group> (<year>2015</year>). &#x201c;<article-title>Falling Rule Lists</article-title>,&#x201d; in <conf-name>Proceedings of the Eighteenth International Conference on Artificial Intelligence and Statistics of&#x20;Proceedings of Machine Learning Research</conf-name>, <conf-loc>San Diego, California, USA</conf-loc>. Editors <person-group person-group-type="editor">
<name>
<surname>Lebanon</surname>
<given-names>G.</given-names>
</name>
<name>
<surname>Vishwanathan</surname>
<given-names>S. V. N.</given-names>
</name>
</person-group> (<publisher-loc>San Diego, CA</publisher-loc>: <publisher-name>PMLR</publisher-name>), <volume>38</volume>. <fpage>1013</fpage>&#x2013;<lpage>1022</lpage>. </citation>
</ref>
<ref id="B42">
<citation citation-type="confproc">
<person-group person-group-type="author">
<name>
<surname>Wang</surname>
<given-names>H.</given-names>
</name>
<name>
<surname>Li</surname>
<given-names>Z.</given-names>
</name>
</person-group> (<year>2017</year>). &#x201c;<article-title>Region Representation Learning via Mobility Flow</article-title>,&#x201d; in <conf-name>Proceedings of the 2017 ACM on Conference on Information and Knowledge Management</conf-name>, <fpage>237</fpage>&#x2013;<lpage>246</lpage>. <pub-id pub-id-type="doi">10.1145/3132847.3133006</pub-id> </citation>
</ref>
<ref id="B43">
<citation citation-type="confproc">
<person-group person-group-type="author">
<name>
<surname>Wang</surname>
<given-names>P.</given-names>
</name>
<name>
<surname>Fu</surname>
<given-names>Y.</given-names>
</name>
<name>
<surname>Xiong</surname>
<given-names>H.</given-names>
</name>
<name>
<surname>Li</surname>
<given-names>X.</given-names>
</name>
</person-group> (<year>2019a</year>). &#x201c;<article-title>Adversarial Substructured Representation Learning for mobile User Profiling</article-title>,&#x201d; in <conf-name>Proceedings of the 25th ACM SIGKDD International Conference on Knowledge Discovery &#x26; Data Mining</conf-name>, <fpage>130</fpage>&#x2013;<lpage>138</lpage>. <pub-id pub-id-type="doi">10.1145/3292500.3330869</pub-id> </citation>
</ref>
<ref id="B44">
<citation citation-type="journal">
<person-group person-group-type="author">
<name>
<surname>Wang</surname>
<given-names>P.</given-names>
</name>
<name>
<surname>Fu</surname>
<given-names>Y.</given-names>
</name>
<name>
<surname>Zhang</surname>
<given-names>J.</given-names>
</name>
<name>
<surname>Li</surname>
<given-names>X.</given-names>
</name>
<name>
<surname>Lin</surname>
<given-names>D.</given-names>
</name>
</person-group> (<year>2018a</year>). <article-title>Learning Urban Community Structures</article-title>. <source>ACM Trans. Intell. Syst. Technol.</source> <volume>9</volume>, <fpage>1</fpage>&#x2013;<lpage>28</lpage>. <pub-id pub-id-type="doi">10.1145/3209686</pub-id> </citation>
</ref>
<ref id="B45">
<citation citation-type="confproc">
<person-group person-group-type="author">
<name>
<surname>Wang</surname>
<given-names>P.</given-names>
</name>
<name>
<surname>Fu</surname>
<given-names>Y.</given-names>
</name>
<name>
<surname>Zhang</surname>
<given-names>J.</given-names>
</name>
<name>
<surname>Wang</surname>
<given-names>P.</given-names>
</name>
<name>
<surname>Zheng</surname>
<given-names>Y.</given-names>
</name>
<name>
<surname>Aggarwal</surname>
<given-names>C.</given-names>
</name>
</person-group> (<year>2018b</year>). &#x201c;<article-title>You Are How You Drive: Peer and Temporal-Aware Representation Learning for&#x20;Driving Behavior Analysis</article-title>,&#x201d; in <conf-name>Proceedings of the 24th ACM SIGKDD International Conference on Knowledge Discovery &#x26; Data Mining</conf-name>, <fpage>2457</fpage>&#x2013;<lpage>2466</lpage>. </citation>
</ref>
<ref id="B46">
<citation citation-type="book">
<person-group person-group-type="author">
<name>
<surname>Wang</surname>
<given-names>P.</given-names>
</name>
<name>
<surname>Fu</surname>
<given-names>Y.</given-names>
</name>
<name>
<surname>Zhou</surname>
<given-names>Y.</given-names>
</name>
<name>
<surname>Liu</surname>
<given-names>K.</given-names>
</name>
<name>
<surname>Li</surname>
<given-names>X.</given-names>
</name>
<name>
<surname>Hua</surname>
<given-names>K. A.</given-names>
</name>
</person-group> (<year>2020b</year>). &#x201c;<article-title>Exploiting Mutual Information for Substructure-Aware Graph Representation Learning</article-title>,&#x201d; in <source>IJCAI</source>, <fpage>3415</fpage>&#x2013;<lpage>3421</lpage>. <pub-id pub-id-type="doi">10.24963/ijcai.2020/472</pub-id> </citation>
</ref>
<ref id="B47">
<citation citation-type="book">
<person-group person-group-type="author">
<name>
<surname>Wang</surname>
<given-names>P.</given-names>
</name>
<name>
<surname>Li</surname>
<given-names>X.</given-names>
</name>
<name>
<surname>Zheng</surname>
<given-names>Y.</given-names>
</name>
<name>
<surname>Aggarwal</surname>
<given-names>C.</given-names>
</name>
<name>
<surname>Fu</surname>
<given-names>Y.</given-names>
</name>
</person-group> (<year>2019b</year>). &#x201c;<article-title>Spatiotemporal Representation Learning for Driving Behavior Analysis: A Joint Perspective of Peer and Temporal Dependencies</article-title>,&#x201d; in <source>IEEE Transactions on Knowledge and Data Engineering</source>. <pub-id pub-id-type="doi">10.1109/tkde.2019.2935203</pub-id> </citation>
</ref>
<ref id="B48">
<citation citation-type="confproc">
<person-group person-group-type="author">
<name>
<surname>Wang</surname>
<given-names>P.</given-names>
</name>
<name>
<surname>Liu</surname>
<given-names>K.</given-names>
</name>
<name>
<surname>Jiang</surname>
<given-names>L.</given-names>
</name>
<name>
<surname>Li</surname>
<given-names>X.</given-names>
</name>
<name>
<surname>Fu</surname>
<given-names>Y.</given-names>
</name>
</person-group> (<year>2020c</year>). &#x201c;<article-title>Incremental mobile User Profiling: Reinforcement Learning with Spatial Knowledge Graph for Modeling Event Streams</article-title>,&#x201d; in <conf-name>Proceedings of the 26th ACM SIGKDD International Conference on Knowledge Discovery &#x26; Data Mining</conf-name>, <fpage>853</fpage>&#x2013;<lpage>861</lpage>. <pub-id pub-id-type="doi">10.1145/3394486.3403128</pub-id> </citation>
</ref>
<ref id="B49">
<citation citation-type="journal">
<person-group person-group-type="author">
<name>
<surname>Wang</surname>
<given-names>S.</given-names>
</name>
<name>
<surname>Cao</surname>
<given-names>J.</given-names>
</name>
<name>
<surname>Chen</surname>
<given-names>H.</given-names>
</name>
<name>
<surname>Peng</surname>
<given-names>H.</given-names>
</name>
<name>
<surname>Huang</surname>
<given-names>Z.</given-names>
</name>
</person-group> (<year>2020d</year>). <article-title>SeqST-GAN</article-title>. <source>ACM Trans. Spat. Algorithms Syst.</source> <volume>6</volume>, <fpage>1</fpage>&#x2013;<lpage>24</lpage>. <pub-id pub-id-type="doi">10.1145/3378889</pub-id> </citation>
</ref>
<ref id="B50">
<citation citation-type="confproc">
<person-group person-group-type="author">
<name>
<surname>Xiao</surname>
<given-names>H.</given-names>
</name>
<name>
<surname>Huang</surname>
<given-names>M.</given-names>
</name>
<name>
<surname>Meng</surname>
<given-names>L.</given-names>
</name>
<name>
<surname>Zhu</surname>
<given-names>X.</given-names>
</name>
</person-group> (<year>2017</year>). &#x201c;<article-title>Ssp: Semantic Space Projection for Knowledge Graph Embedding with Text Descriptions</article-title>,&#x201d; in <conf-name>Thirty-First AAAI Conference on Artificial Intelligence</conf-name>. </citation>
</ref>
<ref id="B51">
<citation citation-type="confproc">
<person-group person-group-type="author">
<name>
<surname>Xun</surname>
<given-names>G.</given-names>
</name>
<name>
<surname>Li</surname>
<given-names>Y.</given-names>
</name>
<name>
<surname>Zhao</surname>
<given-names>W. X.</given-names>
</name>
<name>
<surname>Gao</surname>
<given-names>J.</given-names>
</name>
<name>
<surname>Zhang</surname>
<given-names>A.</given-names>
</name>
</person-group> (<year>2017</year>). &#x201c;<article-title>A Correlated Topic Model Using Word Embeddings</article-title>,&#x201d; in <conf-name>Proceedings of the Twenty-Sixth International Joint Conference on Artificial Intelligence</conf-name> (<publisher-loc>Melbourne</publisher-loc>: <publisher-name>IJCAI</publisher-name>), <fpage>4207</fpage>&#x2013;<lpage>4213</lpage>. <pub-id pub-id-type="doi">10.24963/ijcai.2017/588</pub-id> </citation>
</ref>
<ref id="B52">
<citation citation-type="journal">
<person-group person-group-type="author">
<name>
<surname>Yang</surname>
<given-names>Z.</given-names>
</name>
<name>
<surname>Dai</surname>
<given-names>Z.</given-names>
</name>
<name>
<surname>Yang</surname>
<given-names>Y.</given-names>
</name>
<name>
<surname>Carbonell</surname>
<given-names>J.</given-names>
</name>
<name>
<surname>Salakhutdinov</surname>
<given-names>R. R.</given-names>
</name>
<name>
<surname>Le</surname>
<given-names>Q. V.</given-names>
</name>
</person-group> (<year>2019</year>). <article-title>Xlnet: Generalized Autoregressive Pretraining for Language Understanding</article-title>. <source>Adv. Neural Inf. Process. Syst.</source> <volume>32</volume>, <fpage>5753</fpage>&#x2013;<lpage>5763</lpage>. </citation>
</ref>
<ref id="B53">
<citation citation-type="journal">
<person-group person-group-type="author">
<name>
<surname>Yao</surname>
<given-names>F.</given-names>
</name>
<name>
<surname>Wang</surname>
<given-names>Y.</given-names>
</name>
</person-group> (<year>2020</year>). <article-title>Tracking Urban Geo-Topics Based on Dynamic Topic Model</article-title>. <source>Comput. Environ. Urban Syst.</source> <volume>79</volume>, <fpage>101419</fpage>. <pub-id pub-id-type="doi">10.1016/j.compenvurbsys.2019.101419</pub-id> </citation>
</ref>
<ref id="B54">
<citation citation-type="journal">
<person-group person-group-type="author">
<name>
<surname>Zhang</surname>
<given-names>J.</given-names>
</name>
<name>
<surname>Dong</surname>
<given-names>Y.</given-names>
</name>
<name>
<surname>Wang</surname>
<given-names>Y.</given-names>
</name>
<name>
<surname>Tang</surname>
<given-names>J.</given-names>
</name>
<name>
<surname>Ding</surname>
<given-names>M.</given-names>
</name>
</person-group> (<year>2019a</year>). <article-title>Prone: Fast and Scalable Network Representation Learning</article-title>. <source>IJCAI</source> <volume>19</volume>, <fpage>4278</fpage>&#x2013;<lpage>4284</lpage>. <pub-id pub-id-type="doi">10.24963/ijcai.2019/594</pub-id> </citation>
</ref>
<ref id="B55">
<citation citation-type="confproc">
<person-group person-group-type="author">
<name>
<surname>Zhang</surname>
<given-names>Y.</given-names>
</name>
<name>
<surname>Fu</surname>
<given-names>Y.</given-names>
</name>
<name>
<surname>Wang</surname>
<given-names>P.</given-names>
</name>
<name>
<surname>Li</surname>
<given-names>X.</given-names>
</name>
<name>
<surname>Zheng</surname>
<given-names>Y.</given-names>
</name>
</person-group> (<year>2019b</year>). &#x201c;<article-title>Unifying Inter-region Autocorrelation and Intra-region Structures for Spatial Embedding via Collective Adversarial Learning</article-title>,&#x201d; in <conf-name>Proceedings of the 25th ACM SIGKDD International Conference on Knowledge Discovery &#x26; Data Mining</conf-name>, <fpage>1700</fpage>&#x2013;<lpage>1708</lpage>. <pub-id pub-id-type="doi">10.1145/3292500.3330972</pub-id> </citation>
</ref>
<ref id="B56">
<citation citation-type="journal">
<person-group person-group-type="author">
<name>
<surname>Zhao</surname>
<given-names>Z.</given-names>
</name>
<name>
<surname>Koutsopoulos</surname>
<given-names>H. N.</given-names>
</name>
<name>
<surname>Zhao</surname>
<given-names>J.</given-names>
</name>
</person-group> (<year>2020</year>). <article-title>Discovering Latent Activity Patterns from Transit Smart Card Data: A Spatiotemporal Topic Model</article-title>. <source>Transportation Res. C: Emerging Tech.</source> <volume>116</volume>, <fpage>102627</fpage>. <pub-id pub-id-type="doi">10.1016/j.trc.2020.102627</pub-id> </citation>
</ref>
<ref id="B57">
<citation citation-type="confproc">
<person-group person-group-type="author">
<name>
<surname>Zheng</surname>
<given-names>R.</given-names>
</name>
<name>
<surname>Liu</surname>
<given-names>Q.</given-names>
</name>
<name>
<surname>Rao</surname>
<given-names>W.</given-names>
</name>
<name>
<surname>Yuan</surname>
<given-names>M.</given-names>
</name>
<name>
<surname>Zeng</surname>
<given-names>J.</given-names>
</name>
<name>
<surname>Jin</surname>
<given-names>Z.</given-names>
</name>
</person-group> (<year>2017</year>). &#x201c;<article-title>Topic Model-Based Road Network Inference from Massive Trajectories</article-title>,&#x201d; in <conf-name>2017 18th IEEE International Conference on Mobile Data Management (MDM)</conf-name> (<publisher-name>IEEE</publisher-name>), <fpage>246</fpage>&#x2013;<lpage>255</lpage>. <pub-id pub-id-type="doi">10.1109/mdm.2017.41</pub-id> </citation>
</ref>
</ref-list>
<app-group>
<app id="app1">
<title>Appendix</title>
<sec>
<title>Reproducing the Algorithm</title>
<p>To claim AutoFTP clearly, we provide the pseudo-code of the learning process of AutoFTP. As illustrated in <bold>Algorithm 1</bold>, the framework includes three steps: (i) initializing the parameters of PSO, (ii) optimizing multiple objectives of AutoFTP, (iii) and outputting the final spatial embeddings. The framework takes topic vectors, POI-POI distance graphs, and POI-POI mobility graphs of spatial entities as input, and final semantically-rich embeddings as output.</p>
<p>For initializing the parameters of PSO (Line 1&#x2013;5 in <xref ref-type="table" rid="T4">Algorithm 1</xref>), we first generate <italic>M</italic> particles as a particle swarm. Then, we initialize the position (topic mask) and velocity of each particle. Specifically, we sample <italic>K</italic> values from the uniform distribution <italic>U</italic>(0, 1) as the position vector, and sample <italic>K</italic> values from the uniform distribution <italic>U</italic>(&#x2212;1, 1) as the velocity vector. Next, we update each particle&#x2019;s best known position (pBest) and the swarm&#x2019;s best known position (gBest) based on each particle&#x2019;s position.</p>
<table-wrap id="T4" position="float">
<label>Algorithm 1</label>
<caption>
<p>Automatic Feature-Topic Pairing (AutoFTP).</p>
</caption>
<table>
<tbody>
<tr>
<td>
<inline-graphic xlink:href="fdata-04-762899-fx1.tif"/>
</td>
</tr>
</tbody>
</table>
</table-wrap>
<p>For optimizing multiple objectives of AutoFTP (Line 6&#x2013;20 in <xref ref-type="table" rid="T4">Algorithm 1</xref>), we first check that if the optimization process achieves the termination conditions. If best topic mask is not found or the training iteration does not surpass the max iteration limitation, we optimize the objectives continually. Otherwise, we output the final spatial representations. During the optimization process, for one iteration, we utilize one particle to do feature-topic pairing. Specifically, we first update the velocity of the particle based on the old velocity, the gap between the current position and pBest, and the gap between the current position and gBest. Then, we generate a new position vector (topic mask) based on the velocity vector (Line 10&#x2013;11 in <xref ref-type="table" rid="T4">Algorithm 1</xref>). In the two lines, <italic>&#x3c9;</italic>, <italic>&#x3d5;</italic>
<sub>
<italic>p</italic>
</sub>, <italic>&#x3d5;</italic>
<sub>
<italic>g</italic>
</sub> are weights, and <italic>&#x3b3;</italic> is the learning rate of the corresponding items. Then, we filter <italic>K</italic> topics by the topic mask and generate the basic embedding of a spatial entity. In addition, we align the semantics of the <italic>K</italic> topics and the features of the basic embedding, and accomplish a downstream task simultaneously (Line 12&#x2013;15 in <xref ref-type="table" rid="T4">Algorithm 1</xref>). Moreover, we evaluate the performance of the particle, and update the value of pBest and gBest for next optimization iteration (Line 16&#x2013;20, in <xref ref-type="table" rid="T4">Algorithm 1</xref>).</p>
<p>For outputting final spatial embeddings (Line 21&#x2013;23 in <xref ref-type="table" rid="T4">Algorithm 1</xref>), we copy the learned spatial representation as the final semantically-rich representations of spatial entities.</p>
</sec>
</app>
</app-group>
</back>
</article>