<?xml version="1.0" encoding="UTF-8" standalone="no"?>
<!DOCTYPE article PUBLIC "-//NLM//DTD Journal Publishing DTD v2.3 20070202//EN" "journalpublishing.dtd">
<article xml:lang="EN" xmlns:mml="http://www.w3.org/1998/Math/MathML" xmlns:xlink="http://www.w3.org/1999/xlink" article-type="research-article">
<front>
<journal-meta>
<journal-id journal-id-type="publisher-id">Front. Psychiatry</journal-id>
<journal-title>Frontiers in Psychiatry</journal-title>
<abbrev-journal-title abbrev-type="pubmed">Front. Psychiatry</abbrev-journal-title>
<issn pub-type="epub">1664-0640</issn>
<publisher>
<publisher-name>Frontiers Media S.A.</publisher-name>
</publisher>
</journal-meta>
<article-meta>
<article-id pub-id-type="doi">10.3389/fpsyt.2022.1012980</article-id>
<article-categories>
<subj-group subj-group-type="heading">
<subject>Psychiatry</subject>
<subj-group>
<subject>Original Research</subject>
</subj-group>
</subj-group>
</article-categories>
<title-group>
<article-title>A graph neural network framework based on preference-aware graph diffusion for recommendation</article-title>
</title-group>
<contrib-group>
<contrib contrib-type="author">
<name><surname>Shu</surname> <given-names>Tao</given-names></name>
<xref ref-type="aff" rid="aff1"><sup>1</sup></xref>
</contrib>
<contrib contrib-type="author" corresp="yes">
<name><surname>Shi</surname> <given-names>Lei</given-names></name>
<xref ref-type="aff" rid="aff2"><sup>2</sup></xref>
<xref ref-type="corresp" rid="c001"><sup>&#x0002A;</sup></xref>
<uri xlink:href="http://loop.frontiersin.org/people/1837443/overview"/>
</contrib>
<contrib contrib-type="author">
<name><surname>Zhu</surname> <given-names>Chuangying</given-names></name>
<xref ref-type="aff" rid="aff3"><sup>3</sup></xref>
</contrib>
<contrib contrib-type="author">
<name><surname>Liu</surname> <given-names>Xia</given-names></name>
<xref ref-type="aff" rid="aff4"><sup>4</sup></xref>
<uri xlink:href="http://loop.frontiersin.org/people/1966718/overview"/>
</contrib>
</contrib-group>
<aff id="aff1"><sup>1</sup><institution>Information Technology Center, Sichuan Vocational and Technical College</institution>, <addr-line>Suining</addr-line>, <country>China</country></aff>
<aff id="aff2"><sup>2</sup><institution>State Key Laboratory of Media Convergence and Communication, Communication University of China</institution>, <addr-line>Beijing</addr-line>, <country>China</country></aff>
<aff id="aff3"><sup>3</sup><institution>Guangxi Key Laboratory of Trusted Software, Guilin University of Electronic Technology</institution>, <addr-line>Guilin</addr-line>, <country>China</country></aff>
<aff id="aff4"><sup>4</sup><institution>School of Physics and Electronic Information, Yantai University</institution>, <addr-line>Yantai</addr-line>, <country>China</country></aff>
<author-notes>
<fn fn-type="edited-by"><p>Edited by: Neeraj Kumar, Thapar Institute of Engineering &#x00026; Technology, India</p></fn>
<fn fn-type="edited-by"><p>Reviewed by: Yanwei Yu, Ocean University of China, China; Pengtao Lv, Henan University of Technology, China; Yujin Zhang, Shanghai University of Engineering Sciences, China; Ziyang Wu, Wuhan University of Technology, China; Bingjing Jia, Anhui Science and Technology University, China</p></fn>
<corresp id="c001">&#x0002A;Correspondence: Lei Shi <email>leikyshi&#x00040;qq.com</email></corresp>
<fn fn-type="other" id="fn001"><p>This article was submitted to Computational Psychiatry, a section of the journal Frontiers in Psychiatry</p></fn></author-notes>
<pub-date pub-type="epub">
<day>13</day>
<month>10</month>
<year>2022</year>
</pub-date>
<pub-date pub-type="collection">
<year>2022</year>
</pub-date>
<volume>13</volume>
<elocation-id>1012980</elocation-id>
<history>
<date date-type="received">
<day>24</day>
<month>08</month>
<year>2022</year>
</date>
<date date-type="accepted">
<day>12</day>
<month>09</month>
<year>2022</year>
</date>
</history>
<permissions>
<copyright-statement>Copyright &#x000A9; 2022 Shu, Shi, Zhu and Liu.</copyright-statement>
<copyright-year>2022</copyright-year>
<copyright-holder>Shu, Shi, Zhu and Liu</copyright-holder>
<license xlink:href="http://creativecommons.org/licenses/by/4.0/"><p>This is an open-access article distributed under the terms of the Creative Commons Attribution License (CC BY). The use, distribution or reproduction in other forums is permitted, provided the original author(s) and the copyright owner(s) are credited and that the original publication in this journal is cited, in accordance with accepted academic practice. No use, distribution or reproduction is permitted which does not comply with these terms.</p></license> </permissions>
<abstract>
<p>Transforming user check-in data into graph structure data is a popular and powerful way to analyze users&#x00027; behaviors in the field of recommendation. Graph-based deep learning methods such as graph embeddings and graph neural networks have shown promising performance on the task of point-of-interest recommendation in recent years. Despite effectiveness, existing methods fail to capture deep graph structural information, leading the suboptimal representations. In addition, they lack the ability of learning the influences of both global preference and user preference on the check-in behavior. To address the aforementioned issues, we propose a general framework based on preference-aware graph diffusion, named PGD. We first construct two types of graphs to represent the global preference and user preference. Then, we apply a graph diffusion process to capture the structural information of the generated graphs, resulting in weighted adjacency matrices. Finally, graph neural network-based backbones are introduced to learn the representations of users and POIs on weighted adjacency matrices. A learnable aggregation module is developed to learn the final representations from global preference and user preference adaptively. Extensive experiments on four real-world datasets demonstrate the superiority of PGD on POI recommendation, compared with the mainstream graph-based deep learning methods.</p></abstract>
<kwd-group>
<kwd>point-of-interest recommendation</kwd>
<kwd>user preference</kwd>
<kwd>graph convolutional network</kwd>
<kwd>temporal context</kwd>
<kwd>spatial context</kwd>
</kwd-group>
<counts>
<fig-count count="3"/>
<table-count count="10"/>
<equation-count count="8"/>
<ref-count count="35"/>
<page-count count="9"/>
<word-count count="6009"/>
</counts>
</article-meta>
</front>
<body>
<sec sec-type="intro" id="s1">
<title>Introduction</title>
<p>Location-based social networks (LBSNs) have attracted a large number of users to share their experience on the Internet in recent years. For example, users may submit comments about a restaurant when they visit that place in Yelp, a famous location-based social network (<xref ref-type="bibr" rid="B1">1</xref>, <xref ref-type="bibr" rid="B2">2</xref>). That restaurant is called point-of-interest (POI), which means a place that attracts a user&#x00027;s interest. As the scale of LBSNs increases, more and more users tend to record their activities on the <bold>p</bold>latform, accumulating enormous check-in data. Such large amount of data offers the opportunities to provide the personalized recommendation service for users when they do not know where to go. This recommendation service is called POI recommendation, which has been a popular service of an LBSN over the past decade.</p>
<p>In the field of POI recommendation, the activities of users are recorded as check-in records, which are usually represented by graph structured data. Thus, graph-based deep learning methods are popular and powerful tools to capture the user preference in this application scenario. A general idea of graph-based deep learning (<xref ref-type="bibr" rid="B3">3</xref>) methods is to transform the check-in records into a variety of graphs, such as the user&#x02013;timestamp graph, to model the user preference from various perspectives of factors (e.g., temporal influence). Graph embeddings (<xref ref-type="bibr" rid="B4">4</xref>, <xref ref-type="bibr" rid="B5">5</xref>) are typical methods for learning the representations of users and POIs. These methods use the techniques of graph representation learnings, such as Deepwalk (<xref ref-type="bibr" rid="B6">6</xref>) and LINE (<xref ref-type="bibr" rid="B7">7</xref>), to learn the latent representations of nodes in the generated graphs.</p>
<p>Despite effectiveness, existing methods share two common weaknesses:</p>
<list list-type="order">
<list-item><p>Hard to capture the deep structural relations of POIs from the generated graph. Existing methods mostly apply or design graph representation methods on the original generated graphs. Although effective for learning node representations, the generated graphs only hold on the relevance of POIs and their immediate neighbors, hard to preserve deep structural relations. Classical graph embeddings (<xref ref-type="bibr" rid="B6">6</xref>&#x02013;<xref ref-type="bibr" rid="B8">8</xref>) only pay attention to a limited range of neighbors. Even though stacking several graph neural networks can relieve this impact, the over-smoothing problem (<xref ref-type="bibr" rid="B9">9</xref>&#x02013;<xref ref-type="bibr" rid="B12">12</xref>) of graph neural networks will also lead to suboptimal representation.</p></list-item>
<list-item><p>Unable to learn presentations of users from global and personalized preferences. Graphs are constructed using check-in records of all users in most graph-based deep learning methods. Such graphs only preserve the global preference, ignoring the personalized preference of a unique user (<xref ref-type="bibr" rid="B13">13</xref>, <xref ref-type="bibr" rid="B14">14</xref>). This drawback could affect the model performance for personalized recommendation.</p></list-item>
</list>
<p>To address the aforementioned issues, we propose a general graph neural network framework for POI recommendation based on preference-aware graph diffusion, named PGD. We first construct two types of graphs to preserve global and personalized preferences, respectively. Then, we conduct the graph diffusion process on generated graphs to capture deep graph structural information, which resulted in a series of weighted matrices. Finally, a graph neural network-based backbone is applied to learn the representations of POIs according to the weighted matrices. We propose a learnable aggregation module to learn the user preference from both global and personalized aspects. We conduct extensive experiments on three widely used datasets from real-world LBSNs. The experimental results have demonstrated the superiority of PGD, compared with existing graph-based deep learning methods. The contributions of this article are as follows:</p>
<list list-type="bullet">
<list-item><p>We propose PGD, a general framework, for POI recommendation. The choice of a graph neural network as the backbone is arbitrary.</p></list-item>
<list-item><p>We conduct the graph diffusion process to capture deep structural information, which is neglected in most existing methods.</p></list-item>
<list-item><p>We propose a learnable aggregation module to learn the user preference from both global and personalized aspects adaptively.</p></list-item>
<list-item><p>We conduct extensive experiments on real-world datasets to validate effectiveness of the method. The results show that our proposed PGD outperforms existing graph-based deep learning methods.</p></list-item>
</list>
<p>The rest of the article is organized as follows: In Section Related work, we briefly review the related works on graph-based deep learning methods for POI recommendation. In Section Preliminaries, we provide some key definitions of terms used in this article, including the definitions of graphs and LBSNs. In Section Proposed framework, we detail our proposed method, including the key designs and learning methods of model parameters. In Section Experiments, we introduce the settings of experiments and report the results. Finally, we conclude this article and outline the future directions in Section Conclusion.</p>
</sec>
<sec id="s2">
<title>Related work</title>
<p>In this section, we review graph-based deep learning methods for the task of POI recommendation. The goal of graph-based deep learning methods, including graph embeddings and graph neural networks, is to learn the low-dimensional representation feature vectors of users and POIs from the graph-structured data generated by the check-in records of users. Then, the representation vectors are used to calculate the rank scores of all unobserved user&#x02013;POI pairs. Finally, the recommendation list is created according to the rank scores from high to low.</p>
<p>GeoMF (<xref ref-type="bibr" rid="B15">15</xref>) utilizes the geography of POIs to construct the potential regions to learn the influence of POI locations on user preference. Then, a learning method based on matrix decomposition is developed to learn the representation vectors of users and POIs. POI2Vec (<xref ref-type="bibr" rid="B16">16</xref>) leverages the rank-based embedding method to incorporate both the geographical influence and sequential transition influence. Geo-PFM (<xref ref-type="bibr" rid="B17">17</xref>) conducts the Poisson distribution to capture the user mobility behaviors and takes various factors into the model for learning user preferences precisely. GE (<xref ref-type="bibr" rid="B5">5</xref>) is one of the typical embedding-based methods for POI recommendation. GE first transforms the check-in records into four graphs to capture the features from the aspects of geography, time, check-in pattern, and semantics. Then, a joint training method is proposed to learn the representations from the aforementioned impact factors. STA (<xref ref-type="bibr" rid="B4">4</xref>) defines the spatiotemporal context, which combines the location and timestamp of check-in records. Such a novel definition of the context makes it possible to capture the characteristics of users&#x00027; check-in behaviors carefully. Based on this, STA utilizes the knowledge graph embedding method (<xref ref-type="bibr" rid="B18">18</xref>) to model the user preference through the translation-based methods. Zhang et al. (<xref ref-type="bibr" rid="B19">19</xref>) considered the category translation of check-in records and proposed a model named HCT to capture the dynamic preference of users according to the POIs and their categories. JLGE (<xref ref-type="bibr" rid="B20">20</xref>) uses a three-step strategy to learn the representation of users and POIs: First, JLGE constructs a series of graphs to represent the interactions of between users and various influence factors, such as temporal factors. Then, a graph embedding-based module (<xref ref-type="bibr" rid="B7">7</xref>) is applied to learn the representations of nodes. Finally, a ranking score function is used to calculate the scores of users and POIs according to the learned representation vectors. Xiong et al. (<xref ref-type="bibr" rid="B21">21</xref>) introduced the graph embeddings to jointly learn the representation vectors for different graphs to preserve the dynamic preference of users.</p>
<p>Despite their effectiveness, embedding-based methods are weak to learn more useful structural information from the check-in graphs. Thanks to the amazing ability of graph neural networks (GNNs) for learning the powerful representation from the graph-structured data, many related works have been proposed to introduce GNNs into the POI recommendation models in recent years. Wang et al. (<xref ref-type="bibr" rid="B22">22</xref>) utilized the GNNs to learn long- and short-term preferences of users according to the check-in graphs. Xu et al. (<xref ref-type="bibr" rid="B23">23</xref>) utilized the graph attention network (<xref ref-type="bibr" rid="B24">24</xref>) to learn the user preference from the POI and ROI levels. GGLR (<xref ref-type="bibr" rid="B25">25</xref>) leverages the graph neural network to learn the representations of POIs according to the newly defined two types of geographical influences: ingoing and outgoing influences. STP-UDGAT (<xref ref-type="bibr" rid="B26">26</xref>) develops a masked self-attention option based on the original graph attention network to exploit personalized user preferences. Zhang et al. (<xref ref-type="bibr" rid="B27">27</xref>) combined GNNs and long short-term memory (Bi-LSTM) to learn the user preference from the users&#x00027; sequential check-in behavior, involving geographical and temporal features. For more related works, we refer to the survey (<xref ref-type="bibr" rid="B28">28</xref>) about deep learning-based models for POI recommendation.</p>
<p>The aforementioned graph-based deep learning methods are conducted on the interaction networks generated by the check-in records. However, they ignore the deep structural information on such graph-structured data, causing them to learn the suboptimal representations of users and POIs. Our proposed framework PGD utilizes the graph diffusion process to preserve the structural information of the generated graphs, further improving the effectiveness of graph-based deep learning methods.</p>
</sec>
<sec id="s3">
<title>Preliminaries</title>
<sec>
<title>Definitions in LBSN</title>
<p>Suppose there are two sets <italic>U</italic> &#x0003D; {<italic>u</italic><sub>1</sub>, ..., <italic>u</italic><sub><italic>m</italic></sub>} and <italic>P</italic> &#x0003D; {<italic>p</italic><sub>1</sub>, ..., <italic>p</italic><sub><italic>n</italic></sub>} representing users and POIs in an LBSN. A POI <italic>p</italic><sub><italic>i</italic></sub> is associated with longitude and latitude coordinates, denoted <italic>l</italic><sub><italic>p</italic><sub><italic>i</italic></sub></sub>. Then, we have the following definitions:</p>
<p><bold>Definition 1 (Check-in record):</bold> Check-in records <italic>D</italic><sub><italic>u</italic></sub> are denoted by a tuple (<italic>u, p, l, t</italic>) that represents the check-in behavior of the user <italic>u</italic> who visited the POI <italic>p</italic> at the time <italic>t</italic> in the location <italic>l</italic>.</p>
<p><bold>Definition 2 (User&#x02013;POI graph):</bold> The user&#x02013;POI graph <italic>G</italic><sub><italic>up</italic></sub> &#x0003D; (<italic>V</italic><sub><italic>up</italic></sub>, <italic>E</italic><sub><italic>up</italic></sub>) is a bipartite graph whose node set consists of two disjoint parts <italic>V</italic><sub><italic>up</italic></sub> &#x0003D; <italic>U</italic>&#x0002B;<italic>P</italic>. <italic>E</italic><sub><italic>up</italic></sub> denotes the edge set. If the user <italic>u</italic> visited the POI <italic>p</italic>, there will be an edge between nodes <italic>u</italic> and <italic>p</italic>, reflecting users&#x00027; check-in records.</p>
<p><bold>Definition 3 (Global activity graph):</bold> The global activity graph <italic>G</italic><sub><italic>ga</italic></sub> &#x0003D; (<italic>V</italic><sub><italic>ga</italic></sub>, <italic>E</italic><sub><italic>ga</italic></sub>) is a POI-POI interaction graph, where <italic>V</italic><sub><italic>ga</italic></sub> &#x0003D; <italic>P</italic>. If a user first visits the POI<italic>p</italic><sub><italic>i</italic></sub> and then visits <italic>p</italic><sub><italic>j</italic></sub> within a time frame &#x00394;<italic>t</italic>, there will be an edge between nodes <italic>p</italic><sub><italic>i</italic></sub> and <italic>p</italic><sub><italic>j</italic></sub>. <italic>G</italic><sub><italic>ga</italic></sub> is a weighted graph that describes the check-in pattern of all users. The higher the frequency of <italic>p</italic><sub><italic>i</italic></sub> and <italic>p</italic><sub><italic>j</italic></sub>, the greater the weight of the edge <italic>e</italic><sub><italic>p</italic><sub><italic>i</italic></sub><italic>p</italic><sub><italic>j</italic></sub></sub>.</p>
<p><bold>Definition 4 (Personalized activity graph):</bold> The personalized activity graph <italic>G</italic><sub><italic>pa</italic></sub> &#x0003D; (<italic>V</italic><sub><italic>pa</italic></sub>, <italic>E</italic><sub><italic>pa</italic></sub>) is similar to <italic>G</italic><sub><italic>ga</italic></sub>. The difference between them is that <italic>G</italic><sub><italic>pa</italic></sub> is changed for each user, describing the check-in pattern of a unique user.</p>
</sec>
<sec>
<title>POI recommendation</title>
<p>Given the check-in records, the location <italic>l</italic>, and the timestamp <italic>t</italic>, the task of POI recommendation is generating a list of POIs {<italic>p</italic><sub>1</sub>, ..., <italic>p</italic><sub><italic>k</italic></sub>} for a user <italic>u</italic>, where <italic>k</italic> is the length of the recommendation list. These recommended POIs do not appear in the history check-in records of the user <italic>u</italic>.</p>
</sec>
</sec>
<sec id="s4">
<title>Proposed framework</title>
<p>In this section, we detail our proposed PGD. It consists of three stages: (1) generating the weight matrices based on graph diffusion, (2) learning the representations of users and POIs, and (3) optimizing the parameters.</p>
<sec>
<title>Graph diffusion operation</title>
<p>Most of the existing graph-based deep learning methods only utilize the information of immediate neighbors on the graphs generated by check-in records of users. For example, graph embeddings sample the node sequence based on the link relations between nodes. GNNs aggregate the information according to the adjacency matrix. Information from limited neighbors will lead to the suboptimal representations.</p>
<p>To address this problem and capture the graph structural information deeply, we conduct a graph diffusion operation on the generated graphs. We first produce a global activity graph to hold the global preference based on the definition in Section Definitions in LBSN. We produce a series of personalized activity graphs for users based on their unique check-in records to preserve the user preference according the definition in Section Definitions in LBSN.</p>
<p>Then, we define the graph diffusion process. Given a graph <italic>G</italic> and its corresponding adjacency matrix <italic>A</italic>, a generalized graph diffusion (<xref ref-type="bibr" rid="B29">29</xref>) operation is defined as (<xref ref-type="bibr" rid="B30">30</xref>, <xref ref-type="bibr" rid="B31">31</xref>) as follows:</p>
<disp-formula id="E1"><label>(1)</label><mml:math id="M1"><mml:mtable class="eqnarray" columnalign="left"><mml:mtr><mml:mtd><mml:mi>D</mml:mi><mml:mi>i</mml:mi><mml:mi>f</mml:mi><mml:mi>f</mml:mi><mml:mrow><mml:mo stretchy="false">(</mml:mo><mml:mrow><mml:mi>G</mml:mi></mml:mrow><mml:mo stretchy="false">)</mml:mo></mml:mrow><mml:mo>=</mml:mo><mml:mstyle displaystyle="true"><mml:munderover accentunder="false" accent="false"><mml:mrow><mml:mo>&#x02211;</mml:mo></mml:mrow><mml:mrow><mml:mi>x</mml:mi><mml:mo>=</mml:mo><mml:mn>0</mml:mn></mml:mrow><mml:mrow><mml:mi>&#x0221E;</mml:mi></mml:mrow></mml:munderover></mml:mstyle><mml:msub><mml:mrow><mml:mi>&#x003B8;</mml:mi></mml:mrow><mml:mrow><mml:mi>x</mml:mi></mml:mrow></mml:msub><mml:msup><mml:mrow><mml:mi>T</mml:mi></mml:mrow><mml:mrow><mml:mi>x</mml:mi></mml:mrow></mml:msup><mml:mo>.</mml:mo></mml:mtd></mml:mtr></mml:mtable></mml:math></disp-formula>
<p>where <italic>T</italic> is the transition matrix, produced by a normalized version of <italic>A</italic>, that is, symmetric normalization. Equation (1) is a general form. In practice, we apply the personalized PageRank to conduct the diffusion process by setting <inline-formula><mml:math id="M2"><mml:msub><mml:mrow><mml:mi>&#x003B8;</mml:mi></mml:mrow><mml:mrow><mml:mi>x</mml:mi></mml:mrow></mml:msub><mml:mstyle class="text"><mml:mtext class="textrm" mathvariant="normal">=</mml:mtext></mml:mstyle><mml:mi>&#x003B1;</mml:mi><mml:msup><mml:mrow><mml:mrow><mml:mo stretchy="false">(</mml:mo><mml:mrow><mml:mn>1</mml:mn><mml:mo>-</mml:mo><mml:mi>&#x003B1;</mml:mi></mml:mrow><mml:mo stretchy="false">)</mml:mo></mml:mrow></mml:mrow><mml:mrow><mml:mi>x</mml:mi></mml:mrow></mml:msup></mml:math></inline-formula>, where &#x003B1; &#x02208; (0, 1) denotes the teleport probability. Let <italic>S</italic> denote the result of <italic>Diff</italic>(<italic>G</italic>) and <italic>S</italic> a weighted graph, where the weight of an edge describes the structural information bias between two nodes on the graph. The large weight represents the strong topology similarity so that <italic>S</italic> preserves deeply structural information compared with the original adjacency matrix.</p>
<p>The motivation to conducting the diffusion operation is that the result of the diffusion process provides a more precise description of the similarity between two nodes, which is beneficial for learning the representations of users and POIs from the generated graphs based on check-in records. For each user, the adjacency matrix of the personalized activity graph is different from that of other users so that the resultant matrix of the diffusion matrix is also different, thus preserving the personalized preference of users.</p>
</sec>
<sec>
<title>GNN-based backbone</title>
<p>After the graph diffusion process, we obtain the weighted matrices of the generated graphs, <inline-formula><mml:math id="M3"><mml:msup><mml:mrow><mml:mi>S</mml:mi></mml:mrow><mml:mrow><mml:msub><mml:mrow><mml:mi>G</mml:mi></mml:mrow><mml:mrow><mml:mi>g</mml:mi><mml:mi>a</mml:mi></mml:mrow></mml:msub></mml:mrow></mml:msup></mml:math></inline-formula> from the global activity graph <italic>G</italic><sub><italic>ga</italic></sub> and <inline-formula><mml:math id="M4"><mml:msup><mml:mrow><mml:mi>S</mml:mi></mml:mrow><mml:mrow><mml:msub><mml:mrow><mml:mi>G</mml:mi></mml:mrow><mml:mrow><mml:mi>p</mml:mi><mml:mi>a</mml:mi></mml:mrow></mml:msub></mml:mrow></mml:msup></mml:math></inline-formula> from the personalized activity graph <italic>G</italic><sub><italic>ga</italic></sub>. We use the row normalization method to normalize them since we only consider the relations of the central node and their neighbors. We further learn the representations of users and POIs based on the aforementioned matrices through a GNN-based backbone.</p>
<p>For a user <italic>u</italic>, we have two matrices, <inline-formula><mml:math id="M5"><mml:msup><mml:mrow><mml:mi>S</mml:mi></mml:mrow><mml:mrow><mml:msub><mml:mrow><mml:mi>G</mml:mi></mml:mrow><mml:mrow><mml:mi>g</mml:mi><mml:mi>a</mml:mi></mml:mrow></mml:msub></mml:mrow></mml:msup></mml:math></inline-formula> and <inline-formula><mml:math id="M6"><mml:msup><mml:mrow><mml:mi>S</mml:mi></mml:mrow><mml:mrow><mml:msub><mml:mrow><mml:msup><mml:mrow><mml:mi>G</mml:mi></mml:mrow><mml:mrow><mml:msub><mml:mrow></mml:mrow><mml:mrow><mml:mi>u</mml:mi></mml:mrow></mml:msub></mml:mrow></mml:msup></mml:mrow><mml:mrow><mml:mi>p</mml:mi><mml:mi>a</mml:mi></mml:mrow></mml:msub></mml:mrow></mml:msup></mml:math></inline-formula>. These matrices preserve the relations of POIs from global and personalized perspectives. For learning the representation of POIs, we apply GNNs on the aforementioned matrices. It is noteworthy that the selection of GNNs is arbitrary, demonstrating the flexibility of our proposed method. In this article, we use two GNNs, GCN and GAT, to learn the representations of users and POIs. The GCN and GAT are popular and powerful GNNs for learning the node representations of graphs. Note that our proposed PGD is a general framework, and most GNNs could be introduced into PGD for POI recommendation.</p>
<p><bold>GCN</bold> (<xref ref-type="bibr" rid="B32">32</xref>): The GCN is a typical GNN that utilizes the first-order Laplace smoothing for aggregating the information from neighbors. A GCN layer is defined as follows:</p>
<disp-formula id="E2"><label>(2)</label><mml:math id="M7"><mml:mtable class="eqnarray" columnalign="left"><mml:mtr><mml:mtd><mml:msup><mml:mrow><mml:mtext>H</mml:mtext></mml:mrow><mml:mrow><mml:mrow><mml:mo stretchy="false">(</mml:mo><mml:mrow><mml:mi>l</mml:mi><mml:mo>&#x0002B;</mml:mo><mml:mn>1</mml:mn></mml:mrow><mml:mo stretchy="false">)</mml:mo></mml:mrow></mml:mrow></mml:msup><mml:mo>=</mml:mo><mml:mi>&#x003C3;</mml:mi><mml:mrow><mml:mo stretchy="false">(</mml:mo><mml:mrow><mml:mtext>S</mml:mtext><mml:msup><mml:mrow><mml:mtext>H</mml:mtext></mml:mrow><mml:mrow><mml:mrow><mml:mo stretchy="false">(</mml:mo><mml:mrow><mml:mi>l</mml:mi></mml:mrow><mml:mo stretchy="false">)</mml:mo></mml:mrow></mml:mrow></mml:msup><mml:msup><mml:mrow><mml:mtext>W</mml:mtext></mml:mrow><mml:mrow><mml:mrow><mml:mo stretchy="false">(</mml:mo><mml:mrow><mml:mi>l</mml:mi></mml:mrow><mml:mo stretchy="false">)</mml:mo></mml:mrow></mml:mrow></mml:msup></mml:mrow><mml:mo stretchy="false">)</mml:mo></mml:mrow><mml:mo>,</mml:mo></mml:mtd></mml:mtr></mml:mtable></mml:math></disp-formula>
<p>where <italic>H</italic> denotes the representations of POIs and <italic>W</italic> denotes the learnable parameter matrix. Since there are no raw features for POIs, we randomly use a matrix as the input of the first layer of the GCN.</p>
<p><bold>GAT</bold> (<xref ref-type="bibr" rid="B24">24</xref>): Different from the GCN that aggregates information based on the node degree, the GAT introduces the attention layer to guide the aggregation process. A GAT layer is defined as follows:</p>
<disp-formula id="E3"><label>(3)</label><mml:math id="M8"><mml:mtable class="eqnarray" columnalign="left"><mml:mtr><mml:mtd><mml:msup><mml:mrow><mml:mtext>H</mml:mtext></mml:mrow><mml:mrow><mml:mrow><mml:mo stretchy="false">(</mml:mo><mml:mrow><mml:mi>l</mml:mi><mml:mo>&#x0002B;</mml:mo><mml:mn>1</mml:mn></mml:mrow><mml:mo stretchy="false">)</mml:mo></mml:mrow></mml:mrow></mml:msup><mml:mo>=</mml:mo><mml:mi>&#x003C3;</mml:mi><mml:mrow><mml:mo stretchy="false">(</mml:mo><mml:mrow><mml:mrow><mml:mo stretchy="false">(</mml:mo><mml:mrow><mml:mtext>S</mml:mtext><mml:mo>&#x02299;</mml:mo><mml:mtext>M</mml:mtext></mml:mrow><mml:mo stretchy="false">)</mml:mo></mml:mrow><mml:msup><mml:mrow><mml:mtext>H</mml:mtext></mml:mrow><mml:mrow><mml:mrow><mml:mo stretchy="false">(</mml:mo><mml:mrow><mml:mi>l</mml:mi></mml:mrow><mml:mo stretchy="false">)</mml:mo></mml:mrow></mml:mrow></mml:msup><mml:msup><mml:mrow><mml:mtext>W</mml:mtext></mml:mrow><mml:mrow><mml:mrow><mml:mo stretchy="false">(</mml:mo><mml:mrow><mml:mi>l</mml:mi></mml:mrow><mml:mo stretchy="false">)</mml:mo></mml:mrow></mml:mrow></mml:msup></mml:mrow><mml:mo stretchy="false">)</mml:mo></mml:mrow><mml:mo>,</mml:mo></mml:mtd></mml:mtr></mml:mtable></mml:math></disp-formula>
<p>where <italic>M</italic> is the attention matrix of node pairs and &#x02299; denotes the element-wise multiplication. We modify the original GAT layer to introduce the diffusion matrix into the aggregation of the GAT.</p>
<p>After the GNN backbone, we obtain the representation of POIs from the global graph <inline-formula><mml:math id="M9"><mml:msup><mml:mrow><mml:mtext>S</mml:mtext></mml:mrow><mml:mrow><mml:msub><mml:mrow><mml:mi>G</mml:mi></mml:mrow><mml:mrow><mml:mi>g</mml:mi><mml:mi>a</mml:mi></mml:mrow></mml:msub></mml:mrow></mml:msup></mml:math></inline-formula>and personalized graph <inline-formula><mml:math id="M10"><mml:msup><mml:mrow><mml:mtext>S</mml:mtext></mml:mrow><mml:mrow><mml:msub><mml:mrow><mml:msup><mml:mrow><mml:mi>G</mml:mi></mml:mrow><mml:mrow><mml:msub><mml:mrow></mml:mrow><mml:mrow><mml:mi>u</mml:mi></mml:mrow></mml:msub></mml:mrow></mml:msup></mml:mrow><mml:mrow><mml:mi>p</mml:mi><mml:mi>a</mml:mi></mml:mrow></mml:msub></mml:mrow></mml:msup></mml:math></inline-formula>, denoted as <inline-formula><mml:math id="M11"><mml:msup><mml:mrow><mml:mtext>H</mml:mtext></mml:mrow><mml:mrow><mml:msub><mml:mrow><mml:mi>G</mml:mi></mml:mrow><mml:mrow><mml:mi>g</mml:mi><mml:mi>a</mml:mi></mml:mrow></mml:msub></mml:mrow></mml:msup></mml:math></inline-formula> and <inline-formula><mml:math id="M12"><mml:msup><mml:mrow><mml:mtext>H</mml:mtext></mml:mrow><mml:mrow><mml:msub><mml:mrow><mml:msup><mml:mrow><mml:mi>G</mml:mi></mml:mrow><mml:mrow><mml:mi>u</mml:mi></mml:mrow></mml:msup></mml:mrow><mml:mrow><mml:mi>p</mml:mi><mml:mi>a</mml:mi></mml:mrow></mml:msub></mml:mrow></mml:msup></mml:math></inline-formula>, respectively. We use <inline-formula><mml:math id="M13"><mml:msup><mml:mrow><mml:mtext>H</mml:mtext></mml:mrow><mml:mrow><mml:msub><mml:mrow><mml:mi>G</mml:mi></mml:mrow><mml:mrow><mml:mi>g</mml:mi><mml:mi>a</mml:mi></mml:mrow></mml:msub></mml:mrow></mml:msup></mml:math></inline-formula> as the final representations <italic>P</italic> for POIs for the reason that the global graph contains more information than the personalized graph.</p>
<p>For calculating the representations of users, we define a learnable aggregation module to learn the final representations. Suppose the visited list of POIs in the check-in records of the user <italic>u</italic> is <italic>C</italic> &#x0003D; {<italic>p</italic><sub>1</sub>, ..., <italic>p</italic><sub><italic>c</italic></sub>}, we develop the following strategy to learn the representation <italic>U</italic>:</p>
<disp-formula id="E4"><label>(4)</label><mml:math id="M14"><mml:mtable class="eqnarray" columnalign="left"><mml:mtr><mml:mtd><mml:msub><mml:mrow><mml:mtext>U</mml:mtext></mml:mrow><mml:mrow><mml:mi>u</mml:mi></mml:mrow></mml:msub><mml:mo>=</mml:mo><mml:mfrac><mml:mrow><mml:mn>1</mml:mn></mml:mrow><mml:mrow><mml:mo>|</mml:mo><mml:mi>C</mml:mi><mml:mo>|</mml:mo></mml:mrow></mml:mfrac><mml:mstyle displaystyle="true"><mml:munder class="msub"><mml:mrow><mml:mo>&#x02211;</mml:mo></mml:mrow><mml:mrow><mml:mi>p</mml:mi><mml:mo>&#x02208;</mml:mo><mml:mi>C</mml:mi></mml:mrow></mml:munder></mml:mstyle><mml:mi>L</mml:mi><mml:mi>A</mml:mi><mml:mrow><mml:mo stretchy="false">(</mml:mo><mml:mrow><mml:msup><mml:mrow><mml:mtext>H</mml:mtext></mml:mrow><mml:mrow><mml:msub><mml:mrow><mml:mi>G</mml:mi></mml:mrow><mml:mrow><mml:mi>g</mml:mi><mml:mi>a</mml:mi></mml:mrow></mml:msub></mml:mrow></mml:msup><mml:mo>,</mml:mo><mml:msup><mml:mrow><mml:mtext>H</mml:mtext></mml:mrow><mml:mrow><mml:msub><mml:mrow><mml:msup><mml:mrow><mml:mi>G</mml:mi></mml:mrow><mml:mrow><mml:mi>u</mml:mi></mml:mrow></mml:msup></mml:mrow><mml:mrow><mml:mi>p</mml:mi><mml:mi>a</mml:mi></mml:mrow></mml:msub></mml:mrow></mml:msup></mml:mrow><mml:mo stretchy="false">)</mml:mo></mml:mrow><mml:mo>,</mml:mo></mml:mtd></mml:mtr></mml:mtable></mml:math></disp-formula>
<disp-formula id="E5"><label>(5)</label><mml:math id="M15"><mml:mtable class="eqnarray" columnalign="left"><mml:mtr><mml:mtd><mml:mi>L</mml:mi><mml:mi>A</mml:mi><mml:mrow><mml:mo stretchy="false">(</mml:mo><mml:mrow><mml:msup><mml:mrow><mml:mi>H</mml:mi></mml:mrow><mml:mrow><mml:msub><mml:mrow><mml:mi>G</mml:mi></mml:mrow><mml:mrow><mml:mi>g</mml:mi><mml:mi>a</mml:mi></mml:mrow></mml:msub></mml:mrow></mml:msup><mml:mo>,</mml:mo><mml:msup><mml:mrow><mml:mi>H</mml:mi></mml:mrow><mml:mrow><mml:msub><mml:mrow><mml:msup><mml:mrow><mml:mi>G</mml:mi></mml:mrow><mml:mrow><mml:mi>u</mml:mi></mml:mrow></mml:msup></mml:mrow><mml:mrow><mml:mi>p</mml:mi><mml:mi>a</mml:mi></mml:mrow></mml:msub></mml:mrow></mml:msup></mml:mrow><mml:mo stretchy="false">)</mml:mo></mml:mrow><mml:mo>=</mml:mo><mml:msub><mml:mrow><mml:mi>&#x003B2;</mml:mi></mml:mrow><mml:mrow><mml:mi>g</mml:mi><mml:mi>a</mml:mi></mml:mrow></mml:msub><mml:mo>&#x000B7;</mml:mo><mml:msup><mml:mrow><mml:mtext>H</mml:mtext></mml:mrow><mml:mrow><mml:msub><mml:mrow><mml:mi>G</mml:mi></mml:mrow><mml:mrow><mml:mi>g</mml:mi><mml:mi>a</mml:mi></mml:mrow></mml:msub></mml:mrow></mml:msup><mml:mo>&#x0002B;</mml:mo><mml:msub><mml:mrow><mml:mi>&#x003B2;</mml:mi></mml:mrow><mml:mrow><mml:mi>p</mml:mi><mml:mi>a</mml:mi></mml:mrow></mml:msub><mml:mo>&#x000B7;</mml:mo><mml:msup><mml:mrow><mml:mtext>H</mml:mtext></mml:mrow><mml:mrow><mml:msub><mml:mrow><mml:msup><mml:mrow><mml:mi>G</mml:mi></mml:mrow><mml:mrow><mml:mi>u</mml:mi></mml:mrow></mml:msup></mml:mrow><mml:mrow><mml:mi>p</mml:mi><mml:mi>a</mml:mi></mml:mrow></mml:msub></mml:mrow></mml:msup><mml:mo>.</mml:mo></mml:mtd></mml:mtr></mml:mtable></mml:math></disp-formula>
<p>where <italic>LA</italic>(&#x000B7;) denotes the learnable aggregation module, and &#x003B2;<sub><italic>ga</italic></sub> and &#x003B2;<sub><italic>pa</italic></sub> are learnable scalars for calculating the representations of users adaptively. We further use the SoftMax function to guarantee the values of &#x003B2;<sub><italic>ga</italic></sub> and &#x003B2;<sub><italic>pa</italic></sub> are in the reasonable range.</p>
<p>Intuitively, the representation of a user comes from the global preference and personalized preference. The function <italic>LA</italic>(&#x000B7;) is capable of preserving the preferences from the previous two aspects by introducing the learnable aggregation factors.</p>
</sec>
<sec>
<title>Parameter optimization</title>
<p>To learn the parameters of the proposed model, we adopt the general optimization framework, Bayesian personalized ranking (<xref ref-type="bibr" rid="B33">33</xref>), for its wide usage in the field of recommendation (<xref ref-type="bibr" rid="B13">13</xref>, <xref ref-type="bibr" rid="B14">14</xref>, <xref ref-type="bibr" rid="B34">34</xref>). The objective function of proposed method is defined as follows:</p>
<disp-formula id="E6"><label>(6)</label><mml:math id="M16"><mml:mtable class="eqnarray" columnalign="left"><mml:mtr><mml:mtd><mml:mi>L</mml:mi><mml:mtext class="textrm" mathvariant="normal">=-</mml:mtext><mml:mstyle displaystyle="true"><mml:munderover accentunder="false" accent="false"><mml:mrow><mml:mo>&#x02211;</mml:mo></mml:mrow><mml:mrow><mml:mi>u</mml:mi><mml:mo>=</mml:mo><mml:mn>1</mml:mn></mml:mrow><mml:mrow><mml:mi>m</mml:mi></mml:mrow></mml:munderover></mml:mstyle><mml:mstyle displaystyle="true"><mml:munder class="msub"><mml:mrow><mml:mo>&#x02211;</mml:mo></mml:mrow><mml:mrow><mml:msub><mml:mrow><mml:mi>p</mml:mi></mml:mrow><mml:mrow><mml:mi>i</mml:mi></mml:mrow></mml:msub><mml:mo>&#x02208;</mml:mo><mml:msub><mml:mrow><mml:mi>D</mml:mi></mml:mrow><mml:mrow><mml:mi>u</mml:mi></mml:mrow></mml:msub></mml:mrow></mml:munder></mml:mstyle><mml:mstyle displaystyle="true"><mml:munder class="msub"><mml:mrow><mml:mo>&#x02211;</mml:mo></mml:mrow><mml:mrow><mml:msub><mml:mrow><mml:mi>p</mml:mi></mml:mrow><mml:mrow><mml:mi>j</mml:mi></mml:mrow></mml:msub><mml:mo>&#x02209;</mml:mo><mml:msub><mml:mrow><mml:mi>D</mml:mi></mml:mrow><mml:mrow><mml:mi>u</mml:mi></mml:mrow></mml:msub></mml:mrow></mml:munder></mml:mstyle><mml:mtext class="textrm" mathvariant="normal">In</mml:mtext><mml:mi>&#x003C6;</mml:mi><mml:mrow><mml:mo stretchy="false">(</mml:mo><mml:mrow><mml:msub><mml:mrow><mml:mtext>U</mml:mtext></mml:mrow><mml:mrow><mml:mi>u</mml:mi></mml:mrow></mml:msub><mml:mo>&#x000B7;</mml:mo><mml:msub><mml:mrow><mml:msup><mml:mrow><mml:mtext>P</mml:mtext></mml:mrow><mml:mrow><mml:mi>&#x003C4;</mml:mi></mml:mrow></mml:msup></mml:mrow><mml:mrow><mml:msub><mml:mrow><mml:mi>p</mml:mi></mml:mrow><mml:mrow><mml:mi>i</mml:mi></mml:mrow></mml:msub></mml:mrow></mml:msub><mml:mo>-</mml:mo><mml:msub><mml:mrow><mml:mtext>U</mml:mtext></mml:mrow><mml:mrow><mml:mi>u</mml:mi></mml:mrow></mml:msub><mml:mo>&#x000B7;</mml:mo><mml:msub><mml:mrow><mml:msup><mml:mrow><mml:mtext>P</mml:mtext></mml:mrow><mml:mrow><mml:mi>&#x003C4;</mml:mi></mml:mrow></mml:msup></mml:mrow><mml:mrow><mml:msub><mml:mrow><mml:mi>p</mml:mi></mml:mrow><mml:mrow><mml:mi>j</mml:mi></mml:mrow></mml:msub></mml:mrow></mml:msub></mml:mrow><mml:mo stretchy="false">)</mml:mo></mml:mrow><mml:mo>&#x0002B;</mml:mo><mml:mi>&#x003B6;</mml:mi><mml:mo>|</mml:mo><mml:mo>|</mml:mo><mml:mi>&#x00398;</mml:mi><mml:mo>|</mml:mo><mml:msup><mml:mrow><mml:mo>|</mml:mo></mml:mrow><mml:mrow><mml:mn>2</mml:mn></mml:mrow></mml:msup></mml:mtd></mml:mtr></mml:mtable></mml:math></disp-formula>
<p>where &#x003C6;(&#x000B7;) denotes the sigmoid function, &#x003B6; denotes the regularization coefficient, and &#x00398; denotes the parameters of PGD. By minimizing Equation (6) with the stochastic gradient descent algorithm, we can learn the representations for users and POIs.</p>
</sec>
</sec>
<sec id="s5">
<title>Experiments</title>
<p>In this section, we introduce the experiments conducted in this article. We first introduce the experimental settings, including datasets, evaluation metrics, and baselines. Then, we report the results of experiments and provide related analyses.</p>
<sec>
<title>Datasets</title>
<p>We use three popular real-world datasets, namely, Yelp (<xref ref-type="bibr" rid="B27">27</xref>), Foursquare (<xref ref-type="bibr" rid="B27">27</xref>) and Gowalla (<xref ref-type="bibr" rid="B27">27</xref>), for experiments in this article. These three datasets are collected from the famous LBSNs: Yelp, Foursquare, and Gowalla, respectively. For each dataset, we perform the data cleaning process and produce the check-in records, obeying the format described in Section Definitions in LBSN. In addition, we remove the users whose check-in records are &#x0003C; 20. We also remove the POIs whose visitors are &#x0003C; 20. The statistics of datasets are reported in <xref ref-type="table" rid="T1">Table 1</xref>.</p>
<table-wrap position="float" id="T1">
<label>Table 1</label>
<caption><p>Statistics of datasets.</p></caption>
<table frame="hsides" rules="groups">
<thead><tr>
<th valign="top" align="left"><bold>Dataset</bold></th>
<th valign="top" align="center"><bold>Users</bold></th>
<th valign="top" align="center"><bold>POIs</bold></th>
<th valign="top" align="center"><bold>Check-in records</bold></th>
</tr>
</thead>
<tbody>
<tr>
<td valign="top" align="left">Yelp</td>
<td valign="top" align="center">24,655</td>
<td valign="top" align="center">15,213</td>
<td valign="top" align="center">689,410</td>
</tr>
<tr>
<td valign="top" align="left">Foursquare</td>
<td valign="top" align="center">21,037</td>
<td valign="top" align="center">21,408</td>
<td valign="top" align="center">828,132</td>
</tr>
<tr>
<td valign="top" align="left">Gowalla</td>
<td valign="top" align="center">14,654</td>
<td valign="top" align="center">26,794</td>
<td valign="top" align="center">962,013</td>
</tr>
</tbody>
</table>
</table-wrap>
<p>We split each dataset into three sets according to the check-in timestamp: the former 60% is the train set, the latest 20% is the test set, and the remaining 20% is the validation set.</p>
</sec>
<sec>
<title>Evaluation metrics</title>
<p>In this article, we choose the widely used evaluation metrics, precision (<xref ref-type="bibr" rid="B27">27</xref>) and recall (<xref ref-type="bibr" rid="B35">35</xref>), to measure the recommendation performance of all models:</p>
<disp-formula id="E7"><label>(7)</label><mml:math id="M17"><mml:mtable class="eqnarray" columnalign="left"><mml:mtr><mml:mtd><mml:mi>P</mml:mi><mml:mi>r</mml:mi><mml:mi>e</mml:mi><mml:mi>c</mml:mi><mml:mi>i</mml:mi><mml:mi>s</mml:mi><mml:mi>i</mml:mi><mml:mi>o</mml:mi><mml:mi>n</mml:mi><mml:mo>=</mml:mo><mml:mfrac><mml:mrow><mml:mo>|</mml:mo><mml:msub><mml:mrow><mml:mi>D</mml:mi></mml:mrow><mml:mrow><mml:mi>t</mml:mi><mml:mi>e</mml:mi><mml:mi>s</mml:mi><mml:mi>t</mml:mi></mml:mrow></mml:msub><mml:mo>&#x02229;</mml:mo><mml:mi>T</mml:mi><mml:mi>o</mml:mi><mml:mi>p</mml:mi><mml:mtext>_</mml:mtext><mml:mi>k</mml:mi><mml:mo>|</mml:mo></mml:mrow><mml:mrow><mml:mo>|</mml:mo><mml:mi>T</mml:mi><mml:mi>o</mml:mi><mml:mi>p</mml:mi><mml:mtext>_</mml:mtext><mml:mi>k</mml:mi><mml:mo>|</mml:mo></mml:mrow></mml:mfrac></mml:mtd></mml:mtr></mml:mtable></mml:math></disp-formula>
<disp-formula id="E8"><label>(8)</label><mml:math id="M18"><mml:mtable class="eqnarray" columnalign="left"><mml:mtr><mml:mtd><mml:mi>R</mml:mi><mml:mi>e</mml:mi><mml:mi>c</mml:mi><mml:mi>a</mml:mi><mml:mi>l</mml:mi><mml:mi>l</mml:mi><mml:mo>=</mml:mo><mml:mfrac><mml:mrow><mml:mo>|</mml:mo><mml:msub><mml:mrow><mml:mi>D</mml:mi></mml:mrow><mml:mrow><mml:mi>t</mml:mi><mml:mi>e</mml:mi><mml:mi>s</mml:mi><mml:mi>t</mml:mi></mml:mrow></mml:msub><mml:mo>&#x02229;</mml:mo><mml:mi>T</mml:mi><mml:mi>o</mml:mi><mml:mi>p</mml:mi><mml:mtext>_</mml:mtext><mml:mi>k</mml:mi><mml:mo>|</mml:mo></mml:mrow><mml:mrow><mml:mo>|</mml:mo><mml:msub><mml:mrow><mml:mi>D</mml:mi></mml:mrow><mml:mrow><mml:mi>t</mml:mi><mml:mi>e</mml:mi><mml:mi>s</mml:mi><mml:mi>t</mml:mi></mml:mrow></mml:msub><mml:mo>|</mml:mo></mml:mrow></mml:mfrac><mml:mo>,</mml:mo></mml:mtd></mml:mtr></mml:mtable></mml:math></disp-formula>
<p>where <italic>D</italic><sub><italic>test</italic></sub> denotes the test set and <italic>Top</italic>_<italic>k</italic> denotes the recommendation list of POIs. We set the length of the list to 10 for experiments. Precision denotes the ratio of successfully recommended POIs in the recommendation list. Recall denotes the ratio of the ratio of successfully recommended POIs in all unvisited POIs.</p>
</sec>
<sec>
<title>Baselines</title>
<p>In this article, we select the following methods as the baselines for experiments:</p>
<p>GeoMF (<xref ref-type="bibr" rid="B15">15</xref>): GeoMF utilizes the latent factor model to capture the influence of geographical factors on the check-in behavior of users.</p>
<p>Geo-PFM (<xref ref-type="bibr" rid="B17">17</xref>): The geographical probabilistic factor model adopted Poisson distribution can effectively model the user mobility patterns by capturing the geographical influences.</p>
<p>POI2Vec (<xref ref-type="bibr" rid="B16">16</xref>): POI2Vec is a ranking-based model that utilizes the sequential influence of check-in records and jointly learns the preference of POIs and sequential transition.</p>
<p>GE (<xref ref-type="bibr" rid="B5">5</xref>): GE is a generic graph-based embedding model, which jointly captures the sequential effect, geographical influence, temporal cyclic effect, and semantic effect in a unified way.</p>
<p>STA (<xref ref-type="bibr" rid="B4">4</xref>) STA introduces the translation-based model to capture the spatiotemporal context for learning the check-in pattern of users.</p>
<p>For the proposed method PGD, we provide two variants implemented by GCN and GAT, namely, PGD-GCN and PGD-GAT, respectively.</p>
<p>For baselines, we use the recommended settings of the hyper-parameters from previous studies. For PGD, we use the grid search method to find the suitable values of the coefficient &#x003B6; of the regularization in Equation (6) and the learning rate <italic>lr</italic> of the optimizer. The research spaces are &#x003B6; &#x02208; {0.005, 0.001, 0.0005} and <italic>lr</italic> &#x02208; {0.01, 0.005, 0.001}. In this article, we set &#x003B6; &#x0003D; 0.0005 and <italic>lr</italic> &#x0003D; 0.001 for experiments.</p>
</sec>
<sec>
<title>Impact of time threshold</title>
<p>In this section, we study the influence of the time threshold &#x00394;<italic>t</italic>, determining the construction of <italic>G</italic><sub><italic>ga</italic></sub> and <italic>G</italic><sub><italic>pa</italic></sub>.</p>
<p>The time threshold &#x00394;<italic>t</italic> controls the density of the graph. If we set a small value &#x00394;<italic>t</italic>, we will get a relatively sparse graph, which means there are less interactions between POIs. Also, it is hard to learn the meaningful representations on a sparse graph. But if we set a large value &#x00394;<italic>t</italic>, the edges in the constructed graph are unable to accurately capture the relations of POIs from the check-in pattern of users. Thus, we conduct the experiments to study the influence of the time threshold &#x00394;<italic>t</italic>. The settings of three datasets are different due to the different check-in data. For Foursquare and Gowalla, we set &#x00394;<italic>t</italic> from {4, 8, ..., 24}. For Yelp, we set &#x00394;<italic>t</italic> from {24, 48, ..., 144}. The unit of &#x00394;<italic>t</italic> is hour; the reason is that Yelp is a reviewer dataset, and the check-in time is recorded by day. Foursquare and Gowalla are the real-time check-in datasets; thus, we have more information on the check-in time on these datasets. So, the value of &#x00394;<italic>t</italic> on Foursquare and Gowalla is smaller than that on Yelp. We use the GCN as the backbone of experiments. The results are reported in <xref ref-type="fig" rid="F1">Figures 1</xref>&#x02013;<xref ref-type="fig" rid="F3">3</xref>.</p>
<fig id="F1" position="float">
<label>Figure 1</label>
<caption><p>Impact of time threshold on Yelp.</p></caption>
<graphic mimetype="image" mime-subtype="tiff" xlink:href="fpsyt-13-1012980-g0001.tif"/>
</fig>
<fig id="F2" position="float">
<label>Figure 2</label>
<caption><p>The impact of time threshold on Foursquare.</p></caption>
<graphic mimetype="image" mime-subtype="tiff" xlink:href="fpsyt-13-1012980-g0002.tif"/>
</fig>
<fig id="F3" position="float">
<label>Figure 3</label>
<caption><p>Impact of time threshold on Gowalla.</p></caption>
<graphic mimetype="image" mime-subtype="tiff" xlink:href="fpsyt-13-1012980-g0003.tif"/>
</fig>
<p>From the results of <xref ref-type="fig" rid="F1">Figures 1</xref>&#x02013;<xref ref-type="fig" rid="F3">3</xref>, we can observe that the time threshold makes a great influence on the model performance. The reason is that the time threshold determines the quality of the generated graphs. A suitable time threshold is beneficial to construct a graph with high quality to describe the relations of POIs, further improving the model performance. We can also observe that the value of achieving the best performance is sensitive to the datasets. This is because different datasets exhibit different check-in patterns of users. Based on the results, we set &#x00394;<italic>t</italic> to 48 on Yelp. For Foursquare, we set it to 16. For Gowalla, we set it to 12.</p>
</sec>
<sec>
<title>Comparison of methods</title>
<p>We run all methods on three datasets with 10 random seeds and report the average of all evaluation metrics. The results are summarized in <xref ref-type="table" rid="T2">Tables 2</xref>&#x02013;<xref ref-type="table" rid="T4">4</xref>.</p>
<table-wrap position="float" id="T2">
<label>Table 2</label>
<caption><p>Results of all methods on Yelp.</p></caption>
<table frame="hsides" rules="groups">
<thead><tr>
<th valign="top" align="left"><bold>Method</bold></th>
<th valign="top" align="center"><bold>Precision</bold></th>
<th valign="top" align="center"><bold>Recall</bold></th>
</tr>
</thead>
<tbody>
<tr>
<td valign="top" align="left">GeoMF</td>
<td valign="top" align="center">0.0223</td>
<td valign="top" align="center">0.0327</td>
</tr>
<tr>
<td valign="top" align="left">Geo-PFM</td>
<td valign="top" align="center">0.0281</td>
<td valign="top" align="center">0.0415</td>
</tr>
<tr>
<td valign="top" align="left">POI2Vec</td>
<td valign="top" align="center">0.0354</td>
<td valign="top" align="center">0.0514</td>
</tr>
<tr>
<td valign="top" align="left">GE</td>
<td valign="top" align="center">0.0432</td>
<td valign="top" align="center">0.0643</td>
</tr>
<tr>
<td valign="top" align="left">STA</td>
<td valign="top" align="center">0.0439</td>
<td valign="top" align="center">0.0652</td>
</tr>
<tr>
<td valign="top" align="left">PGD-GCN</td>
<td valign="top" align="center">0.0481</td>
<td valign="top" align="center">0.0711</td>
</tr>
<tr>
<td valign="top" align="left">PGD-GAT</td>
<td valign="top" align="center"><bold>0.0509</bold></td>
<td valign="top" align="center"><bold>0.0754</bold></td>
</tr>
</tbody>
</table>
</table-wrap>
<table-wrap position="float" id="T3">
<label>Table 3</label>
<caption><p>Results of all methods on Foursquare.</p></caption>
<table frame="hsides" rules="groups">
<thead><tr>
<th valign="top" align="left"><bold>Method</bold></th>
<th valign="top" align="center"><bold>Precision</bold></th>
<th valign="top" align="center"><bold>Recall</bold></th>
</tr>
</thead>
<tbody>
<tr>
<td valign="top" align="left">GeoMF</td>
<td valign="top" align="center">0.0319</td>
<td valign="top" align="center">0.0423</td>
</tr>
<tr>
<td valign="top" align="left">Geo-PFM</td>
<td valign="top" align="center">0.0322</td>
<td valign="top" align="center">0.0441</td>
</tr>
<tr>
<td valign="top" align="left">POI2Vec</td>
<td valign="top" align="center">0.0373</td>
<td valign="top" align="center">0.0536</td>
</tr>
<tr>
<td valign="top" align="left">GE</td>
<td valign="top" align="center">0.0491</td>
<td valign="top" align="center">0.0693</td>
</tr>
<tr>
<td valign="top" align="left">STA</td>
<td valign="top" align="center">0.0521</td>
<td valign="top" align="center">0.0732</td>
</tr>
<tr>
<td valign="top" align="left">PGD-GCN</td>
<td valign="top" align="center">0.0587</td>
<td valign="top" align="center">0.0786</td>
</tr>
<tr>
<td valign="top" align="left">PGD-GAT</td>
<td valign="top" align="center">0.0595</td>
<td valign="top" align="center">0.0798</td>
</tr>
</tbody>
</table>
</table-wrap>
<table-wrap position="float" id="T4">
<label>Table 4</label>
<caption><p>Results of all methods on Gowalla.</p></caption>
<table frame="hsides" rules="groups">
<thead><tr>
<th valign="top" align="left"><bold>Method</bold></th>
<th valign="top" align="center"><bold>Precision</bold></th>
<th valign="top" align="center"><bold>Recall</bold></th>
</tr>
</thead>
<tbody>
<tr>
<td valign="top" align="left">GeoMF</td>
<td valign="top" align="center">0.0459</td>
<td valign="top" align="center">0.0636</td>
</tr>
<tr>
<td valign="top" align="left">Geo-PFM</td>
<td valign="top" align="center">0.0471</td>
<td valign="top" align="center">0.0678</td>
</tr>
<tr>
<td valign="top" align="left">POI2Vec</td>
<td valign="top" align="center">0.0644</td>
<td valign="top" align="center">0.0916</td>
</tr>
<tr>
<td valign="top" align="left">GE</td>
<td valign="top" align="center">0.0693</td>
<td valign="top" align="center">0.0983</td>
</tr>
<tr>
<td valign="top" align="left">STA</td>
<td valign="top" align="center">0.0702</td>
<td valign="top" align="center">0.0994</td>
</tr>
<tr>
<td valign="top" align="left">PGD-GCN</td>
<td valign="top" align="center">0.0733</td>
<td valign="top" align="center">0.1092</td>
</tr>
<tr>
<td valign="top" align="left">PGD-GAT</td>
<td valign="top" align="center">0.0745</td>
<td valign="top" align="center">0.1125</td>
</tr>
</tbody>
</table>
</table-wrap>
<p>From <xref ref-type="table" rid="T2">Tables 2</xref>&#x02013;<xref ref-type="table" rid="T4">4</xref>, we can observe that our proposed methods PGD-GCN and PGD-GAT consistently outperform other baselines, demonstrating the superiority of the proposed framework. In addition, PGD-GAT outperforms PGD-GCN, which indicates that introducing the attention mechanism benefits learning the representation vectors of users and POIs. Considering the best results of three datasets, the lowest one is from Yelp. This is because the dataset of Yelp is most sparse, compared with Foursquare and Gowalla. This phenomenon also implies that the data sparsity has a great influence on the performance of the POI recommendation task.</p>
</sec>
<sec>
<title>Ablation study</title>
<p>In this section, we first design ablation studies to measure the contribution of the proposed learnable aggregation module to the model performance. We also use the GCN as the backbone model. We propose two variants, PGD-GCN-0 and PGD-GCN-1. PGD-GCN-0 denotes that only the global preference is considered in the model. PGD-GCN-1 means that the learnable aggregation is removed. The results are reported in <xref ref-type="table" rid="T5">Tables 5</xref>&#x02013;<xref ref-type="table" rid="T7">7</xref>.</p>
<table-wrap position="float" id="T5">
<label>Table 5</label>
<caption><p>Results of variants on Yelp.</p></caption>
<table frame="hsides" rules="groups">
<thead><tr>
<th valign="top" align="left"><bold>Method</bold></th>
<th valign="top" align="center"><bold>Precision</bold></th>
<th valign="top" align="center"><bold>Recall</bold></th>
</tr>
</thead>
<tbody>
<tr>
<td valign="top" align="left">PGD-GCN-0</td>
<td valign="top" align="center">0.0442</td>
<td valign="top" align="center">0.0661</td>
</tr>
<tr>
<td valign="top" align="left">PGD-GCN-1</td>
<td valign="top" align="center">0.0451</td>
<td valign="top" align="center">0.0675</td>
</tr>
<tr>
<td valign="top" align="left">PGD-GCN</td>
<td valign="top" align="center">0.0481</td>
<td valign="top" align="center">0.0711</td>
</tr>
</tbody>
</table>
</table-wrap>
<table-wrap position="float" id="T6">
<label>Table 6</label>
<caption><p>Results of variants on Foursquare.</p></caption>
<table frame="hsides" rules="groups">
<thead><tr>
<th valign="top" align="left"><bold>Method</bold></th>
<th valign="top" align="center"><bold>Precision</bold></th>
<th valign="top" align="center"><bold>Recall</bold></th>
</tr>
</thead>
<tbody>
<tr>
<td valign="top" align="left">PGD-GCN-0</td>
<td valign="top" align="center">0.0512</td>
<td valign="top" align="center">0.0709</td>
</tr>
<tr>
<td valign="top" align="left">PGD-GCN-1</td>
<td valign="top" align="center">0.0541</td>
<td valign="top" align="center">0.0755</td>
</tr>
<tr>
<td valign="top" align="left">PGD-GCN</td>
<td valign="top" align="center">0.0587</td>
<td valign="top" align="center">0.0786</td>
</tr>
</tbody>
</table>
</table-wrap>
<table-wrap position="float" id="T7">
<label>Table 7</label>
<caption><p>Results of variants on Gowalla.</p></caption>
<table frame="hsides" rules="groups">
<thead><tr>
<th valign="top" align="left"><bold>Method</bold></th>
<th valign="top" align="center"><bold>Precision</bold></th>
<th valign="top" align="center"><bold>Recall</bold></th>
</tr>
</thead>
<tbody>
<tr>
<td valign="top" align="left">PGD-GCN-0</td>
<td valign="top" align="center">0.0708</td>
<td valign="top" align="center">0.0998</td>
</tr>
<tr>
<td valign="top" align="left">PGD-GCN-1</td>
<td valign="top" align="center">0.0721</td>
<td valign="top" align="center">0.1054</td>
</tr>
<tr>
<td valign="top" align="left">PGD-GCN</td>
<td valign="top" align="center">0.0733</td>
<td valign="top" align="center">0.1092</td>
</tr>
</tbody>
</table>
</table-wrap>
<p>The results of <xref ref-type="table" rid="T5">Tables 5</xref>&#x02013;<xref ref-type="table" rid="T7">7</xref> have demonstrated that our proposed learnable aggregation module is helpful to learn the precise representations of users.</p>
<p>Then, we design experiments to validate the effectiveness of the graph diffusion process. As mentioned before, the diffusion process is helpful to capture the deep graph structural information and further promote to learn the relations of POIs. We consider a variant of PGD where the graph diffusion process is removed, PGD-GCN-RW. The GCN backbone is also applied in experiments. The results are reported in <xref ref-type="table" rid="T8">Tables 8</xref>&#x02013;<xref ref-type="table" rid="T10">10</xref>.</p>
<table-wrap position="float" id="T8">
<label>Table 8</label>
<caption><p>Results of variants on Yelp.</p></caption>
<table frame="hsides" rules="groups">
<thead><tr>
<th valign="top" align="left"><bold>Method</bold></th>
<th valign="top" align="center"><bold>Precision</bold></th>
<th valign="top" align="center"><bold>Recall</bold></th>
</tr>
</thead>
<tbody>
<tr>
<td valign="top" align="left">PGD-GCN-RW</td>
<td valign="top" align="center">0.0436</td>
<td valign="top" align="center">0.0648</td>
</tr>
<tr>
<td valign="top" align="left">PGD-GCN</td>
<td valign="top" align="center">0.0481</td>
<td valign="top" align="center">0.0711</td>
</tr>
</tbody>
</table>
</table-wrap>
<table-wrap position="float" id="T9">
<label>Table 9</label>
<caption><p>Results of variants on Foursquare.</p></caption>
<table frame="hsides" rules="groups">
<thead><tr>
<th valign="top" align="left"><bold>Method</bold></th>
<th valign="top" align="center"><bold>Precision</bold></th>
<th valign="top" align="center"><bold>Recall</bold></th>
</tr>
</thead>
<tbody>
<tr>
<td valign="top" align="left">PGD-GCN-RW</td>
<td valign="top" align="center">0.0498</td>
<td valign="top" align="center">0.0699</td>
</tr>
<tr>
<td valign="top" align="left">PGD-GCN</td>
<td valign="top" align="center">0.0587</td>
<td valign="top" align="center">0.0786</td>
</tr>
</tbody>
</table>
</table-wrap>
<table-wrap position="float" id="T10">
<label>Table 10</label>
<caption><p>Results of variants on Gowalla.</p></caption>
<table frame="hsides" rules="groups">
<thead><tr>
<th valign="top" align="left"><bold>Method</bold></th>
<th valign="top" align="center"><bold>Precision</bold></th>
<th valign="top" align="center"><bold>Recall</bold></th>
</tr>
</thead>
<tbody>
<tr>
<td valign="top" align="left">PGD-GCN-RW</td>
<td valign="top" align="center">0.0703</td>
<td valign="top" align="center">0.0996</td>
</tr>
<tr>
<td valign="top" align="left">PGD-GCN</td>
<td valign="top" align="center">0.0733</td>
<td valign="top" align="center">0.1092</td>
</tr>
</tbody>
</table>
</table-wrap>
<p>The results from <xref ref-type="table" rid="T8">Tables 8</xref>&#x02013;<xref ref-type="table" rid="T10">10</xref> have proved that the graph diffusion process is necessary to learn the powerful representations of users and POIs. With the graph diffusion process, the performance of the model has been significantly improved.</p>
</sec>
<sec>
<title>Discussion of results</title>
<p>In the experiments, we first study the influence of the settings of the time threshold. The results show that a suitable value of the time threshold can help model improve the recommendation effectiveness. Then, we compare our proposed PGD with baselines on real-world datasets. The results indicate the superiority of PGD for the POI recommendation task. Finally, we conduct ablation studies to explore the gain of key designs of PGD, that is, graph diffusion and learnable aggregation module. The results show that all key designs are beneficial for improving the model performance.</p>
</sec>
</sec>
<sec sec-type="conclusions" id="s6">
<title>Conclusion</title>
<p>In this article, we propose a general GNN-based framework, named PGD. PGD first constructs two types of graphs to preserve the global and personalized preferences. Then, a graph diffusion process is applied to capture the deep graph structural information. Finally, a GNN-based backbone is developed to learn the representations of POIs. For the representations of users, we propose a learnable aggregation module to learn the features from both global and personalized aspects adaptively. We conduct extensive experiments on three real-world datasets. The experimental results show that our proposed method outperforms the mainstream POI recommendation methods.</p>
<p>PGD is a general framework, and it can utilize most GNNs to learn the representations of users and POIs and show its high flexibility. The superiority of PGD demonstrates that the graph diffusion process is beneficial for learning the powerful representations, which reveals that leveraging high-order structural relations is a crucial point for improving the model performance.</p>
<p>For the future directions, although PGD utilizes the graph diffusion process to preserve the structural information, it relies on the rich check-in records of users. It is hard to capture the relations of unobserved POIs based on the graph diffusion so that we plan to introduce various similarity-based techniques to estimate the semantic relevance between all POIs. Such pre-computed similarities are helpful to relieve the impact of data sparsity.</p>
</sec>
<sec sec-type="data-availability" id="s7">
<title>Data availability statement</title>
<p>All datasets can be downloaded from the following websites: <ext-link ext-link-type="uri" xlink:href="https://www.yelp.com/dataset">https://www.yelp.com/dataset</ext-link> (for Yelp), <ext-link ext-link-type="uri" xlink:href="https://sites.google.com/site/yangdingqi/home/foursquare-dataset">https://sites.google.com/site/yangdingqi/home/foursquare-dataset</ext-link> (for Foursquare), <ext-link ext-link-type="uri" xlink:href="http://snap.stanford.edu/data/loc-gowalla.html">http://snap.stanford.edu/data/loc-gowalla.html</ext-link> (for Gowalla).</p>
</sec>
<sec id="s8">
<title>Author contributions</title>
<p>TS and LS designed the overall framework and conceived the idea of this paper. TS analyzed the data using correlation algorithms. TS and CZ wrote the paper. XL helped in typesetting and revising the paper, and modified the English grammar. All authors contributed to the article and approved the submitted version.</p>
</sec>
<sec sec-type="funding-information" id="s9">
<title>Funding</title>
<p>This work was supported by the Fundamental Research Funds for the Central Universities (No. CUC220C011, CUC22GZ038) and Youth Fund Project of Guangxi Natural Science Foundation (No. 2021GXNSFBA196054).</p>
</sec>
<sec sec-type="COI-statement" id="conf1">
<title>Conflict of interest</title>
<p>The authors declare that the research was conducted in the absence of any commercial or financial relationships that could be construed as a potential conflict of interest.</p>
</sec>
<sec sec-type="disclaimer" id="s10">
<title>Publisher&#x00027;s note</title>
<p>All claims expressed in this article are solely those of the authors and do not necessarily represent those of their affiliated organizations, or those of the publisher, the editors and the reviewers. Any product that may be evaluated in this article, or claim that may be made by its manufacturer, is not guaranteed or endorsed by the publisher.</p>
</sec>
</body>
<back>

<ref-list>
<title>References</title>
<ref id="B1">
<label>1.</label>
<citation citation-type="journal"><person-group person-group-type="author"><name><surname>Shi</surname> <given-names>L</given-names></name> <name><surname>Du</surname> <given-names>J</given-names></name> <name><surname>Cheng</surname> <given-names>G</given-names></name> <name><surname>Liu</surname> <given-names>X</given-names></name> <name><surname>Xiong</surname> <given-names>Z</given-names></name> <name><surname>Luo</surname> <given-names>J</given-names></name></person-group>. <article-title>Cross-media search method based on complementary attention and generative adversarial network for social networks</article-title>. <source>Int J Intell Syst.</source> (<year>2022</year>) <volume>37</volume>:<fpage>4393</fpage>&#x02013;<lpage>416</lpage>. <pub-id pub-id-type="doi">10.1002/int.22723</pub-id></citation>
</ref>
<ref id="B2">
<label>2.</label>
<citation citation-type="journal"><person-group person-group-type="author"><name><surname>Shi</surname> <given-names>L</given-names></name> <name><surname>Song</surname> <given-names>G</given-names></name> <name><surname>Cheng</surname> <given-names>G</given-names></name> <name><surname>Liu</surname> <given-names>X</given-names></name></person-group>. <article-title>A user-based aggregation topic model for understanding user&#x00027;s preference and intention in social network</article-title>. <source>Neurocomputing.</source> (<year>2020</year>) <volume>413</volume>:<fpage>1</fpage>&#x02013;<lpage>13</lpage>. <pub-id pub-id-type="doi">10.1016/j.neucom.2020.06.099</pub-id></citation>
</ref>
<ref id="B3">
<label>3.</label>
<citation citation-type="journal"><person-group person-group-type="author"><name><surname>Zhang</surname> <given-names>Y</given-names></name> <name><surname>Yu</surname> <given-names>L</given-names></name> <name><surname>Fang</surname> <given-names>Z</given-names></name> <name><surname>Xiong</surname> <given-names>NN</given-names></name> <name><surname>Zhang</surname> <given-names>L</given-names></name> <name><surname>Tian</surname> <given-names>H</given-names></name></person-group>. <article-title>An end-to-end deep learning model for robust smooth filtering identification</article-title>. <source>Fut Gen Comput Syst.</source> (<year>2022</year>) <volume>127</volume>:<fpage>263</fpage>&#x02013;<lpage>75</lpage>. <pub-id pub-id-type="doi">10.1016/j.future.2021.09.004</pub-id></citation>
</ref>
<ref id="B4">
<label>4.</label>
<citation citation-type="journal"><person-group person-group-type="author"><name><surname>Qian</surname> <given-names>T</given-names></name> <name><surname>Liu</surname> <given-names>B</given-names></name> <name><surname>Nguyen</surname> <given-names>QVH</given-names></name> <name><surname>Yin</surname> <given-names>H</given-names></name></person-group>. <article-title>Spatiotemporal representation learning for translation-based POI recommendation</article-title>. <source>ACM Trans Infm Syst.</source> (<year>2019</year>) <volume>37</volume>:<fpage>1</fpage>&#x02013;<lpage>24</lpage>. <pub-id pub-id-type="doi">10.1145/3295499</pub-id></citation>
</ref>
<ref id="B5">
<label>5.</label>
<citation citation-type="journal"><person-group person-group-type="author"><name><surname>Xie</surname> <given-names>M</given-names></name> <name><surname>Yin</surname> <given-names>H</given-names></name> <name><surname>Wang</surname> <given-names>H</given-names></name> <name><surname>Xu</surname> <given-names>F</given-names></name> <name><surname>Chen</surname> <given-names>W</given-names></name> <name><surname>Wang</surname> <given-names>S</given-names></name></person-group>. <article-title>Learning graph-based poi embedding for location-based recommendation</article-title>. In: <source>Proceedings of the 25th ACM International on Conference on Information and Knowledge Management</source>. <publisher-loc>Indianapolis, IN</publisher-loc>: <publisher-name>Association for Computing Machinery</publisher-name> (<year>2016</year>). p. <fpage>15</fpage>&#x02013;<lpage>24</lpage>. <pub-id pub-id-type="doi">10.1145/2983323.2983711</pub-id></citation>
</ref>
<ref id="B6">
<label>6.</label>
<citation citation-type="journal"><person-group person-group-type="author"><name><surname>Perozzi</surname> <given-names>B</given-names></name> <name><surname>Al-Rfou</surname> <given-names>R</given-names></name> <name><surname>Skiena</surname> <given-names>S</given-names></name></person-group>. <article-title>Deepwalk: online learning of social representations</article-title>. In: <source>Proceedings of the 20th ACM SIGKDD International Conference on Knowledge Discovery and Data Mining</source>. <publisher-loc>New York, NY</publisher-loc>: <publisher-name>Association for Computing Machinery</publisher-name> (<year>2014</year>). p. <fpage>701</fpage>&#x02013;<lpage>10</lpage>. <pub-id pub-id-type="doi">10.1145/2623330.2623732</pub-id></citation>
</ref>
<ref id="B7">
<label>7.</label>
<citation citation-type="journal"><person-group person-group-type="author"><name><surname>Tang</surname> <given-names>J</given-names></name> <name><surname>Qu</surname> <given-names>M</given-names></name> <name><surname>Wang</surname> <given-names>M</given-names></name> <name><surname>Zhang</surname> <given-names>M</given-names></name> <name><surname>Yan</surname> <given-names>J</given-names></name> <name><surname>Mei</surname> <given-names>Q</given-names></name></person-group>. <article-title>Line: large-scale information network embedding</article-title>. In: <source>Proceedings of the 24th International Conference on World Wide Web</source>. <publisher-loc>Florence</publisher-loc>: <publisher-name>International World Wide Web Conferences Steering Committee</publisher-name> (<year>2015</year>). p. <fpage>1067</fpage>&#x02013;<lpage>77</lpage>. <pub-id pub-id-type="doi">10.1145/2736277.2741093</pub-id></citation>
</ref>
<ref id="B8">
<label>8.</label>
<citation citation-type="journal"><person-group person-group-type="author"><name><surname>Grover</surname> <given-names>A</given-names></name> <name><surname>Leskovec</surname> <given-names>J</given-names></name></person-group>. <article-title>node2vec: scalable feature learning for networks</article-title>. In: <source>Proceedings of the 22nd ACM SIGKDD International Conference on Knowledge Discovery and Data Mining</source>. (<year>2016</year>). p. <fpage>855</fpage>&#x02013;<lpage>64</lpage>. <pub-id pub-id-type="doi">10.1145/2939672.2939754</pub-id><pub-id pub-id-type="pmid">27853626</pub-id></citation></ref>
<ref id="B9">
<label>9.</label>
<citation citation-type="journal"><person-group person-group-type="author"><name><surname>Li</surname> <given-names>Q</given-names></name> <name><surname>Han</surname> <given-names>Z</given-names></name> <name><surname>Wu</surname> <given-names>XM</given-names></name></person-group>. <article-title>Deeper insights into graph convolutional networks for semi-supervised learning</article-title>. In: <source>Thirty-Second AAAI Conference on Artificial Intelligence</source>. (<year>2018</year>). p. <fpage>3538</fpage>&#x02013;<lpage>45</lpage>. <pub-id pub-id-type="doi">10.1609/aaai.v32i1.11604</pub-id></citation>
</ref>
<ref id="B10">
<label>10.</label>
<citation citation-type="journal"><person-group person-group-type="author"><name><surname>Chen</surname> <given-names>D</given-names></name> <name><surname>Lin</surname> <given-names>Y</given-names></name> <name><surname>Li</surname> <given-names>W</given-names></name> <name><surname>Li</surname> <given-names>P</given-names></name> <name><surname>Zhou</surname> <given-names>J</given-names></name> <name><surname>Sun</surname> <given-names>X</given-names></name></person-group>. <article-title>Measuring and relieving the over-smoothing problem for graph neural networks from the topological view</article-title>. In: P<italic>roceedings of the AAAI Conference on Artificial Intelligence</italic>. (<year>2020</year>). Vol. 34, p. <fpage>3438</fpage>&#x02013;<lpage>45</lpage>. <pub-id pub-id-type="doi">10.1609/aaai.v34i04.5747</pub-id></citation>
</ref>
<ref id="B11">
<label>11.</label>
<citation citation-type="journal"><person-group person-group-type="author"><name><surname>Zhou</surname> <given-names>K</given-names></name> <name><surname>Dong</surname> <given-names>Y</given-names></name> <name><surname>Wang</surname> <given-names>K</given-names></name> <name><surname>Lee</surname> <given-names>WS</given-names></name> <name><surname>Hooi</surname> <given-names>B</given-names></name> <name><surname>Xu</surname> <given-names>H</given-names></name> <etal/></person-group>. <article-title>Understanding and resolving performance degradation in deep graph convolutional networks</article-title>. In: <source>Proceedings of the 30th ACM International Conference on Information and Knowledge Management</source>. (<year>2021</year>). p. <fpage>2728</fpage>&#x02013;<lpage>37</lpage>. <pub-id pub-id-type="doi">10.1145/3459637.3482488</pub-id></citation>
</ref>
<ref id="B12">
<label>12.</label>
<citation citation-type="journal"><person-group person-group-type="author"><name><surname>Liu</surname> <given-names>M</given-names></name> <name><surname>Gao</surname> <given-names>H</given-names></name> <name><surname>Ji</surname> <given-names>S</given-names></name></person-group>. <article-title>Towards deeper graph neural networks</article-title>. In: <source>Proceedings of the 26th ACM SIGKDD International Conference on Knowledge Discovery and Data Mining</source>. (<year>2020</year>). p. <fpage>338</fpage>&#x02013;<lpage>48</lpage>. <pub-id pub-id-type="doi">10.1145/3394486.3403076</pub-id></citation>
</ref>
<ref id="B13">
<label>13.</label>
<citation citation-type="journal"><person-group person-group-type="author"><name><surname>Li</surname> <given-names>X</given-names></name> <name><surname>Jiang</surname> <given-names>M</given-names></name> <name><surname>Hong</surname> <given-names>H</given-names></name> <name><surname>Liao</surname> <given-names>L</given-names></name></person-group>. <article-title>A time-aware personalized point-of-interest recommendation via high-order tensor factorization</article-title>. <source>ACM Trans Inform Syst.</source> (<year>2017</year>) <volume>35</volume>:<fpage>1</fpage>&#x02013;<lpage>23</lpage>. <pub-id pub-id-type="doi">10.1145/3057283</pub-id></citation>
</ref>
<ref id="B14">
<label>14.</label>
<citation citation-type="journal"><person-group person-group-type="author"><name><surname>He</surname> <given-names>J</given-names></name> <name><surname>Li</surname> <given-names>X</given-names></name> <name><surname>Liao</surname> <given-names>L</given-names></name></person-group>. <article-title>Category-aware next point-of-interest recommendation via listwise bayesian personalized ranking</article-title>. In: <source>Twenty-Sixth International Joint Conference on Artificial Intelligence</source>. <publisher-loc>Melbourne, VIC</publisher-loc>: <publisher-name>IJCAI</publisher-name> (<year>2017</year>). Vol. 17, p. <fpage>1837</fpage>&#x02013;<lpage>43</lpage>. <pub-id pub-id-type="doi">10.24963/ijcai.2017/255</pub-id></citation>
</ref>
<ref id="B15">
<label>15.</label>
<citation citation-type="journal"><person-group person-group-type="author"><name><surname>Lian</surname> <given-names>D</given-names></name> <name><surname>Zhao</surname> <given-names>C</given-names></name> <name><surname>Xie</surname> <given-names>X</given-names></name> <name><surname>Sun</surname> <given-names>G</given-names></name> <name><surname>Chen</surname> <given-names>E</given-names></name> <name><surname>Rui</surname> <given-names>Y</given-names></name></person-group>. <article-title>GeoMF: joint geographical modeling and matrix factorization for point-of-interest recommendation</article-title>. In: <source>Proceedings of the 20th ACM SIGKDD International Conference on Knowledge Discovery and Data Mining</source>. (<year>2014</year>). p. <fpage>831</fpage>&#x02013;<lpage>40</lpage>. <pub-id pub-id-type="doi">10.1145/2623330.2623638</pub-id></citation>
</ref>
<ref id="B16">
<label>16.</label>
<citation citation-type="journal"><person-group person-group-type="author"><name><surname>Feng</surname> <given-names>S</given-names></name> <name><surname>Cong</surname> <given-names>G</given-names></name> <name><surname>An</surname> <given-names>B</given-names></name> <name><surname>Chee</surname> <given-names>YM</given-names></name></person-group>. <article-title>Poi2vec: Geographical latent representation for predicting future visitors</article-title>. In: <source>Thirty-First AAAI Conference on Artificial Intelligence</source>. (<year>2017</year>). p. <fpage>102</fpage>&#x02013;<lpage>8</lpage>. <pub-id pub-id-type="doi">10.1609/aaai.v31i1.10500</pub-id></citation>
</ref>
<ref id="B17">
<label>17.</label>
<citation citation-type="journal"><person-group person-group-type="author"><name><surname>Liu</surname> <given-names>B</given-names></name> <name><surname>Xiong</surname> <given-names>H</given-names></name> <name><surname>Papadimitriou</surname> <given-names>S</given-names></name> <name><surname>Fu</surname> <given-names>Y</given-names></name> <name><surname>Yao</surname> <given-names>Z</given-names></name></person-group>. <article-title>A general geographical probabilistic factor model for point of interest recommendation</article-title>. <source>IEEE Trans Knowl Data Eng.</source> (<year>2014</year>) <volume>27</volume>:<fpage>1167</fpage>&#x02013;<lpage>79</lpage>. <pub-id pub-id-type="doi">10.1109/TKDE.2014.2362525</pub-id></citation>
</ref>
<ref id="B18">
<label>18.</label>
<citation citation-type="journal"><person-group person-group-type="author"><name><surname>Lin</surname> <given-names>Y</given-names></name> <name><surname>Liu</surname> <given-names>Z</given-names></name> <name><surname>Sun</surname> <given-names>M</given-names></name> <name><surname>Liu</surname> <given-names>Y</given-names></name> <name><surname>Zhu</surname> <given-names>X</given-names></name></person-group>. <article-title>Learning entity and relation embeddings for knowledge graph completion</article-title>. In: <source>Twenty-Ninth AAAI Conference on Artificial Intelligence</source>. (<year>2015</year>). p. <fpage>2181</fpage>&#x02013;<lpage>7</lpage>. <pub-id pub-id-type="doi">10.1609/aaai.v29i1.9491</pub-id></citation>
</ref>
<ref id="B19">
<label>19.</label>
<citation citation-type="journal"><person-group person-group-type="author"><name><surname>Zhang</surname> <given-names>L</given-names></name> <name><surname>Sun</surname> <given-names>Z</given-names></name> <name><surname>Zhang</surname> <given-names>J</given-names></name> <name><surname>Kloeden</surname> <given-names>H</given-names></name> <name><surname>Klanner</surname> <given-names>F</given-names></name></person-group>. <article-title>Modeling hierarchical category transition for next POI recommendation with uncertain check-ins</article-title>. <source>Inf Sci.</source> (<year>2020</year>) <volume>515</volume>:<fpage>169</fpage>&#x02013;<lpage>90</lpage>. <pub-id pub-id-type="doi">10.1016/j.ins.2019.12.006</pub-id></citation>
</ref>
<ref id="B20">
<label>20.</label>
<citation citation-type="journal"><person-group person-group-type="author"><name><surname>Christoforidis</surname> <given-names>G</given-names></name> <name><surname>Kefalas</surname> <given-names>P</given-names></name> <name><surname>Papadopoulos</surname> <given-names>A</given-names></name> <name><surname>Manolopoulos</surname> <given-names>Y</given-names></name></person-group>. <article-title>Recommendation of points-of-interest using graph embeddings</article-title>. In: <source>2018 IEEE 5th International Conference on Data Science and Advanced Analytics</source>. (<year>2018</year>). p. <fpage>31</fpage>&#x02013;<lpage>40</lpage>. <pub-id pub-id-type="doi">10.1109/DSAA.2018.00013</pub-id></citation>
</ref>
<ref id="B21">
<label>21.</label>
<citation citation-type="journal"><person-group person-group-type="author"><name><surname>Xiong</surname> <given-names>X</given-names></name> <name><surname>Xiong</surname> <given-names>F</given-names></name> <name><surname>Zhao</surname> <given-names>J</given-names></name> <name><surname>Qiao</surname> <given-names>S</given-names></name> <name><surname>Li</surname> <given-names>Y</given-names></name> <name><surname>Zhao</surname> <given-names>Y</given-names></name></person-group>. <article-title>Dynamic discovery of favorite locations in spatio-temporal social networks</article-title>. <source>Inform Process Manag.</source> (<year>2020</year>) <volume>57</volume>:<fpage>102337</fpage>. <pub-id pub-id-type="doi">10.1016/j.ipm.2020.102337</pub-id></citation>
</ref>
<ref id="B22">
<label>22.</label>
<citation citation-type="journal"><person-group person-group-type="author"><name><surname>Wang</surname> <given-names>D</given-names></name> <name><surname>Wang</surname> <given-names>X</given-names></name> <name><surname>Xiang</surname> <given-names>Z</given-names></name> <name><surname>Yu</surname> <given-names>D</given-names></name> <name><surname>Deng</surname> <given-names>S</given-names></name> <name><surname>Xu</surname> <given-names>G</given-names></name></person-group>. <article-title>Attentive sequential model based on graph neural network for next poi recommendation</article-title>. <source>World Wide Web.</source> (<year>2021</year>) <volume>24</volume>:<fpage>2161</fpage>&#x02013;<lpage>84</lpage>. <pub-id pub-id-type="doi">10.1007/s11280-021-00961-9</pub-id></citation>
</ref>
<ref id="B23">
<label>23.</label>
<citation citation-type="journal"><person-group person-group-type="author"><name><surname>Xu</surname> <given-names>H</given-names></name> <name><surname>Wei</surname> <given-names>J</given-names></name> <name><surname>Yang</surname> <given-names>Z</given-names></name> <name><surname>Wang</surname> <given-names>J</given-names></name></person-group>. <article-title>Graph attentive network for region recommendation with poi-and roi-level attention</article-title>. In: <source>Asia-Pacific Web (APWeb) and Web-Age Information Management (WAIM) Joint International Conference on Web and Big Data</source>. (<year>2020</year>). p. <fpage>509</fpage>&#x02013;<lpage>16</lpage>. <pub-id pub-id-type="doi">10.1007/978-3-030-60259-8_37</pub-id></citation>
</ref>
<ref id="B24">
<label>24.</label>
<citation citation-type="journal"><person-group person-group-type="author"><name><surname>Veli&#x0010D;kovi&#x00107;</surname> <given-names>P</given-names></name> <name><surname>Cucurull</surname> <given-names>G</given-names></name> <name><surname>Casanova</surname> <given-names>A</given-names></name> <name><surname>Romero</surname> <given-names>A</given-names></name> <name><surname>Lio</surname> <given-names>P</given-names></name> <name><surname>Bengio</surname> <given-names>Y</given-names></name></person-group>. <article-title>Graph attention networks</article-title>. <source>arXiv preprint arXiv:1710.10903</source>. (<year>2017</year>). <pub-id pub-id-type="doi">10.48550/arXiv.1710.10903</pub-id></citation>
</ref>
<ref id="B25">
<label>25.</label>
<citation citation-type="journal"><person-group person-group-type="author"><name><surname>Chang</surname> <given-names>B</given-names></name> <name><surname>Jang</surname> <given-names>G</given-names></name> <name><surname>Kim</surname> <given-names>S</given-names></name> <name><surname>Kang</surname> <given-names>J</given-names></name></person-group>. <article-title>Learning graph-based geographical latent representation for point-of-interest recommendation</article-title>. In: <source>Proceedings of the 29th ACM International Conference on Information and Knowledge Management</source>. (<year>2020</year>). p. <fpage>135</fpage>&#x02013;<lpage>44</lpage>. <pub-id pub-id-type="doi">10.1145/3340531.3411905</pub-id></citation>
</ref>
<ref id="B26">
<label>26.</label>
<citation citation-type="journal"><person-group person-group-type="author"><name><surname>Lim</surname> <given-names>N</given-names></name> <name><surname>Hooi</surname> <given-names>B</given-names></name> <name><surname>Ng</surname> <given-names>SK</given-names></name> <name><surname>Wang</surname> <given-names>X</given-names></name> <name><surname>Goh</surname> <given-names>YL</given-names></name> <name><surname>Weng</surname> <given-names>R</given-names></name> <etal/></person-group>. <article-title>STP-UDGAT: spatial-temporal-preference user dimensional graph attention network for next POI recommendation</article-title>. In: <source>Proceedings of the 29th ACM International Conference on Information and Knowledge Management</source>. (<year>2020</year>). p. <fpage>845</fpage>&#x02013;<lpage>54</lpage>. <pub-id pub-id-type="doi">10.1145/3340531.3411876</pub-id></citation>
</ref>
<ref id="B27">
<label>27.</label>
<citation citation-type="journal"><person-group person-group-type="author"><name><surname>Zhang</surname> <given-names>J</given-names></name> <name><surname>Liu</surname> <given-names>X</given-names></name> <name><surname>Zhou</surname> <given-names>X</given-names></name> <name><surname>Chu</surname> <given-names>X</given-names></name></person-group>. <article-title>Leveraging graph neural networks for point-of-interest recommendations</article-title>. <source>Neurocomputing.</source> (<year>2021</year>) <volume>462</volume>:<fpage>1</fpage>&#x02013;<lpage>13</lpage>. <pub-id pub-id-type="doi">10.1016/j.neucom.2021.07.063</pub-id></citation>
</ref>
<ref id="B28">
<label>28.</label>
<citation citation-type="journal"><person-group person-group-type="author"><name><surname>Islam</surname> <given-names>MA</given-names></name> <name><surname>Mohammad</surname> <given-names>MM</given-names></name> <name><surname>Das</surname> <given-names>SSS</given-names></name> <name><surname>Ali</surname> <given-names>ME</given-names></name></person-group>. <article-title>A survey on deep learning based Point-of-Interest (POI) recommendations</article-title>. <source>Neurocomputing.</source> (<year>2022</year>) <volume>472</volume>:<fpage>306</fpage>&#x02013;<lpage>25</lpage>. <pub-id pub-id-type="doi">10.1016/j.neucom.2021.05.114</pub-id></citation>
</ref>
<ref id="B29">
<label>29.</label>
<citation citation-type="journal"><person-group person-group-type="author"><name><surname>Zhang</surname> <given-names>Y</given-names></name> <name><surname>Liu</surname> <given-names>T</given-names></name> <name><surname>Cattani</surname> <given-names>C</given-names></name> <name><surname>Cui</surname> <given-names>Q</given-names></name> <name><surname>Liu</surname> <given-names>S</given-names></name></person-group>. <article-title>Diffusion-based image inpainting forensics via weighted least squares filtering enhancement</article-title>. <source>Multimed Tools Appl.</source> (<year>2021</year>) <volume>80</volume>:<fpage>30725</fpage>&#x02013;<lpage>39</lpage>. <pub-id pub-id-type="doi">10.1007/s11042-021-10623-7</pub-id></citation>
</ref>
<ref id="B30">
<label>30.</label>
<citation citation-type="journal"><person-group person-group-type="author"><name><surname>Klicpera</surname> <given-names>J</given-names></name> <name><surname>Wei&#x000DF;enberger</surname> <given-names>S</given-names></name> <name><surname>G&#x000FC;nnemann</surname> <given-names>S</given-names></name></person-group>. <article-title>Diffusion improves graph learning</article-title>. <source>arXiv preprint arXiv:1911.05485</source>. (<year>2019</year>). <pub-id pub-id-type="doi">10.48550/arXiv.1911.05485</pub-id></citation>
</ref>
<ref id="B31">
<label>31.</label>
<citation citation-type="journal"><person-group person-group-type="author"><name><surname>Zhao</surname> <given-names>J</given-names></name> <name><surname>Dong</surname> <given-names>Y</given-names></name> <name><surname>Ding</surname> <given-names>M</given-names></name> <name><surname>Kharlamov</surname> <given-names>E</given-names></name> <name><surname>Tang</surname> <given-names>J</given-names></name></person-group>. <article-title>Adaptive diffusion in graph neural networks</article-title>. <source>Adv Neural Inf Process Syst.</source> (<year>2021</year>) <volume>34</volume>:<fpage>23321</fpage>&#x02013;<lpage>33</lpage>.</citation>
</ref>
<ref id="B32">
<label>32.</label>
<citation citation-type="journal"><person-group person-group-type="author"><name><surname>Kipf</surname> <given-names>TN</given-names></name> <name><surname>Welling</surname> <given-names>M</given-names></name></person-group>. <article-title>Semi-supervised classification with graph convolutional networks</article-title>. <source>arXiv preprint arXiv:1609.02907</source>. (<year>2016</year>). <pub-id pub-id-type="doi">10.48550/arXiv.1609.02907</pub-id></citation>
</ref>
<ref id="B33">
<label>33.</label>
<citation citation-type="journal"><person-group person-group-type="author"><collab>Rendle S Freudenthaler C Gantner Z and Schmidt-Thieme L</collab></person-group>. <article-title>BPR: bayesian personalized ranking from implicit feedback</article-title>. <source>arXiv preprint arXiv:1205.2618</source>. (<year>2012</year>). <pub-id pub-id-type="doi">10.48550/arXiv.1205.2618</pub-id></citation>
</ref>
<ref id="B34">
<label>34.</label>
<citation citation-type="journal"><person-group person-group-type="author"><name><surname>Jhamb</surname> <given-names>Y</given-names></name> <name><surname>Fang</surname> <given-names>Y</given-names></name></person-group>. <article-title>A dual-perspective latent factor model for group-aware social event recommendation</article-title>. <source>Inform Process Manag.</source> (<year>2017</year>) <volume>53</volume>:<fpage>559</fpage>&#x02013;<lpage>76</lpage>. <pub-id pub-id-type="doi">10.1016/j.ipm.2017.01.001</pub-id></citation>
</ref>
<ref id="B35">
<label>35.</label>
<citation citation-type="journal"><person-group person-group-type="author"><name><surname>Ji</surname> <given-names>W</given-names></name> <name><surname>Meng</surname> <given-names>X</given-names></name> <name><surname>Zhang</surname> <given-names>Y</given-names></name></person-group>. <article-title>STARec: adaptive learning with spatiotemporal and activity influence for POI recommendation</article-title>. <source>ACM Trans Inform Syst.</source> (<year>2021</year>) <volume>40</volume>:<fpage>1</fpage>&#x02013;<lpage>40</lpage>. <pub-id pub-id-type="doi">10.1145/3485631</pub-id></citation>
</ref>
</ref-list> 
</back>
</article>