<?xml version="1.0" encoding="UTF-8" standalone="no"?>
<!DOCTYPE article PUBLIC "-//NLM//DTD Journal Publishing DTD v2.3 20070202//EN" "journalpublishing.dtd">
<article xmlns:mml="http://www.w3.org/1998/Math/MathML" xmlns:xlink="http://www.w3.org/1999/xlink" article-type="research-article">
<front>
<journal-meta>
<journal-id journal-id-type="publisher-id">Front. Appl. Math. Stat.</journal-id>
<journal-title>Frontiers in Applied Mathematics and Statistics</journal-title>
<abbrev-journal-title abbrev-type="pubmed">Front. Appl. Math. Stat.</abbrev-journal-title>
<issn pub-type="epub">2297-4687</issn>
<publisher>
<publisher-name>Frontiers Media S.A.</publisher-name>
</publisher>
</journal-meta>
<article-meta>
<article-id pub-id-type="doi">10.3389/fams.2018.00001</article-id>
<article-categories>
<subj-group subj-group-type="heading">
<subject>Applied Mathematics and Statistics</subject>
<subj-group>
<subject>Original Research</subject>
</subj-group>
</subj-group>
</article-categories>
<title-group>
<article-title>Exact Heat Kernel on a Hypersphere and Its Applications in Kernel SVM</article-title>
</title-group>
<contrib-group>
<contrib contrib-type="author">
<name><surname>Zhao</surname> <given-names>Chenchao</given-names></name>
<xref ref-type="aff" rid="aff1"><sup>1</sup></xref>
<xref ref-type="aff" rid="aff2"><sup>2</sup></xref>
<uri xlink:href="http://loop.frontiersin.org/people/498725/overview"/>
</contrib>
<contrib contrib-type="author" corresp="yes">
<name><surname>Song</surname> <given-names>Jun S.</given-names></name>
<xref ref-type="aff" rid="aff1"><sup>1</sup></xref>
<xref ref-type="aff" rid="aff2"><sup>2</sup></xref>
<xref ref-type="author-notes" rid="fn001"><sup>&#x0002A;</sup></xref>
<uri xlink:href="http://loop.frontiersin.org/people/25538/overview"/>
</contrib>
</contrib-group>
<aff id="aff1"><sup>1</sup><institution>Department of Physics, University of Illinois at Urbana-Champaign</institution>, <addr-line>Urbana, IL</addr-line>, <country>United States</country></aff>
<aff id="aff2"><sup>2</sup><institution>Carl R. Woese Institute for Genomic Biology, University of Illinois at Urbana-Champaign</institution>, <addr-line>Urbana, IL</addr-line>, <country>United States</country></aff>
<author-notes>
<fn fn-type="edited-by"><p>Edited by: Yiming Ying, University at Albany (SUNY), United States</p></fn>
<fn fn-type="edited-by"><p>Reviewed by: Jun Fan, Hong Kong Baptist University, Hong Kong; Xin GUO, Hong Kong Polytechnic University, Hong Kong</p></fn>
<fn fn-type="corresp" id="fn001"><p>&#x0002A;Correspondence: Jun S. Song <email>songj&#x00040;illinois.edu</email></p></fn>
<fn fn-type="other" id="fn002"><p>This article was submitted to Mathematics of Computation and Data Science, a section of the journal Frontiers in Applied Mathematics and Statistics</p></fn></author-notes>
<pub-date pub-type="epub">
<day>24</day>
<month>01</month>
<year>2018</year>
</pub-date>
<pub-date pub-type="collection">
<year>2018</year>
</pub-date>
<volume>4</volume>
<elocation-id>1</elocation-id>
<history>
<date date-type="received">
<day>03</day>
<month>10</month>
<year>2017</year>
</date>
<date date-type="accepted">
<day>08</day>
<month>01</month>
<year>2018</year>
</date>
</history>
<permissions>
<copyright-statement>Copyright &#x000A9; 2018 Zhao and Song.</copyright-statement>
<copyright-year>2018</copyright-year>
<copyright-holder>Zhao and Song</copyright-holder>
<license xlink:href="http://creativecommons.org/licenses/by/4.0/"><p>This is an open-access article distributed under the terms of the Creative Commons Attribution License (CC BY). The use, distribution or reproduction in other forums is permitted, provided the original author(s) or licensor are credited and that the original publication in this journal is cited, in accordance with accepted academic practice. No use, distribution or reproduction is permitted which does not comply with these terms.</p></license>
</permissions>
<abstract><p>Many contemporary statistical learning methods assume a Euclidean feature space. This paper presents a method for defining similarity based on hyperspherical geometry and shows that it often improves the performance of support vector machine compared to other competing similarity measures. Specifically, the idea of using heat diffusion on a hypersphere to measure similarity has been previously proposed and tested by Lafferty and Lebanon [<xref ref-type="bibr" rid="B1">1</xref>], demonstrating promising results based on a heuristic heat kernel obtained from the zeroth order parametrix expansion; however, how well this heuristic kernel agrees with the exact hyperspherical heat kernel remains unknown. This paper presents a higher order parametrix expansion of the heat kernel on a unit hypersphere and discusses several problems associated with this expansion method. We then compare the heuristic kernel with an exact form of the heat kernel expressed in terms of a uniformly and absolutely convergent series in high-dimensional angular momentum eigenmodes. Being a natural measure of similarity between sample points dwelling on a hypersphere, the exact kernel often shows superior performance in kernel SVM classifications applied to text mining, tumor somatic mutation imputation, and stock market analysis.</p></abstract>
<kwd-group>
<kwd>heat kernel</kwd>
<kwd>support vector machine (SVM)</kwd>
<kwd>hyperspherical geometry</kwd>
<kwd>document classification</kwd>
<kwd>genomics</kwd>
<kwd>time series</kwd>
</kwd-group>
<contract-sponsor id="cn001">Sontag Foundation<named-content content-type="fundref-id">10.13039/100006064</named-content></contract-sponsor>
<counts>
<fig-count count="3"/>
<table-count count="4"/>
<equation-count count="31"/>
<ref-count count="38"/>
<page-count count="10"/>
<word-count count="6701"/>
</counts>
</article-meta>
</front>
<body>
<sec sec-type="intro" id="s1">
<title>1. Introduction</title>
<p>As the techniques for analyzing large data sets continue to grow, diverse quantitative sciences&#x02014;including computational biology, observation astronomy, and high energy physics&#x02014;are becoming increasingly data driven. Moreover, modern business decision making critically depends on quantitative analyses such as community detection and consumer behavior prediction. Consequently, statistical learning has become an indispensable tool for modern data analysis. Data acquired from various experiments are usually organized into an <italic>n</italic> &#x000D7; <italic>m</italic> matrix, where the number <italic>n</italic> of features typically far exceeds the number <italic>m</italic> of samples. In this view, the <italic>m</italic> samples, corresponding to the columns of the data matrix, are naturally interpreted as points in a high-dimensional feature space &#x0211D;<sup><italic>n</italic></sup>. Traditional statistical modeling approaches often lose their power when the feature dimension is high. To ameliorate this problem, Lafferty and Lebanon proposed a multinomial interpretation of non-negative feature vectors and an accompanying transformation of the multinomial simplex to a hypersphere, demonstrating that using the heat kernel on this hypersphere may improve the performance of kernel support vector machine (SVM) [<xref ref-type="bibr" rid="B2">2</xref>&#x02013;<xref ref-type="bibr" rid="B8">8</xref>]. Despite the interest that this idea has attracted, only approximate heat kernel is known to date. We here present an exact form of the heat kernel on a hypersphere of arbitrary dimension and study its performance in kernel SVM classifications of text mining, genomic, and stock price data sets.</p>
<p>To date, sparse data clouds have been extensively analyzed in the flat Euclidean space endowed with the <italic>L</italic><sup>2</sup>-norm using traditional statistical learning algorithms, including KMeans, hierarchical clustering, SVM, and neural network [<xref ref-type="bibr" rid="B2">2</xref>&#x02013;<xref ref-type="bibr" rid="B8">8</xref>]; however, the flat geometry of the Euclidean space often poses severe challenges in clustering and classification problems when the data clouds take non-trivial geometric shapes or class labels are spatially mixed. Manifold learning and kernel-based embedding methods attempt to address these challenges by estimating the intrinsic geometry of a putative submanifold from which the data points were sampled and by embedding the data into an abstract Hilbert space using a nonlinear map implicitly induced by the chosen kernel, respectively [<xref ref-type="bibr" rid="B9">9</xref>&#x02013;<xref ref-type="bibr" rid="B11">11</xref>]. The geometry of these curved spaces may then provide novel information about the structure and organization of original data points.</p>
<p>Heat equation on the data submanifold or transformed feature space offers an especially attractive idea of measuring similarity between data points by using the physical model of diffusion of relatedness (&#x0201C;heat&#x0201D;) on curved space, where the diffusion process is driven by the intrinsic geometry of the underlying space. Even though such diffusion process has been successfully approximated as a discrete-time, discrete-space random walk on complex networks, its continuous formulation is rarely analytically solvable and usually requires complicated asymptotic expansion techniques from differential geometry [<xref ref-type="bibr" rid="B12">12</xref>]. An analytic solution, if available, would thus provide a valuable opportunity for comparing its performance with approximate asymptotic solutions and rigorously testing the power of heat diffusion for geometric data analysis.</p>
<p>Given that a Riemannian manifold of dimension <italic>d</italic> is locally homeomorphic to &#x0211D;<sup><italic>d</italic></sup>, and that the heat kernel is a solution to the heat equation with a point source initial condition, one may assume in the short diffusion time limit (<italic>t</italic> &#x02193; 0) that most of the heat is localized within the vicinity of the initial point and that the heat kernel on a Riemannian manifold locally resembles the Euclidean heat kernel. This idea forms the motivation behind the parametrix expansion, where the heat kernel in curved space is approximated as a product of the Euclidean heat kernel in normal coordinates and an asymptotic series involving the diffusion time and normal coordinates. In particular, for a unit hypersphere, the parametrix expansion in the limit <italic>t</italic> &#x02193; 0 involves a modified Euclidean heat kernel with the Euclidean distance ||<bold>x</bold>|| replaced by the geodesic arc length &#x003B8;. Computing this parametrix expansion is, however, technically challenging; even when the computation is tractable, applying the approximation directly to high-dimensional clustering and classification problems may have limitations. For example, in order to be able to group samples robustly, one needs the diffusion time <italic>t</italic> to be not too small; otherwise, the sample relatedness may be highly localized and decay too fast away from each sample. Moreover, the leading order term in the asymptotic series is an increasing function of &#x003B8; and diverges as &#x003B8; approaches &#x003C0;, yielding an incorrect conclusion that two antipodal points are highly similar. For these reasons, the machine learning community has been using only the Euclidean diffusion term without the asymptotic series correction; how this resulting kernel, called the parametrix kernel [<xref ref-type="bibr" rid="B1">1</xref>], compares with the exact heat kernel on a hypersphere remains an outstanding question, which is addressed in this paper.</p>
<p>Analytically solving the diffusion equation on a Riemannian manifold is challenging [<xref ref-type="bibr" rid="B12">12</xref>&#x02013;<xref ref-type="bibr" rid="B14">14</xref>]. Unlike the discrete analogs&#x02014;such as spectral clustering [<xref ref-type="bibr" rid="B15">15</xref>] and diffusion map [<xref ref-type="bibr" rid="B16">16</xref>], where eigenvectors of a finite dimensional matrix can be easily obtained&#x02014;the eigenfunctions of the Laplace operator on a Riemannian manifold are usually intractable. Fortunately, the high degree of symmetry of a hypersphere allows the explicit construction of eigenfunctions, called hyperspherical harmonics, via the projection of homogeneous polynomials [<xref ref-type="bibr" rid="B17">17</xref>, <xref ref-type="bibr" rid="B18">18</xref>]. The exact heat kernel is then obtained as a convergent power series in these eigenfunctions. In this paper, we compare the analytic behavior of this exact heat kernel with that of the parametrix kernel and analyze their performance in classification.</p>
</sec>
<sec sec-type="results" id="s2">
<title>2. Results</title>
<p>The heat kernel is the fundamental solution to the heat equation (&#x02202;<sub><italic>t</italic></sub> &#x02212; &#x00394;<sub><italic>x</italic></sub>)<italic>u</italic>(<italic>x, t</italic>) &#x0003D; 0 with an initial point source [<xref ref-type="bibr" rid="B19">19</xref>], where &#x00394;<sub><italic>x</italic></sub> is the Laplace operator; the amount of heat emanating from the source that has diffused to a neighborhood during time <italic>t</italic> &#x0003E; 0 is used to measure the similarity between the source and proximal points. The heat conduction depends on the geometry of feature space, and the main idea behind the application of hyperspherical geometry to data analysis relies on the following map from a non-negative feature space to a unit hypersphere:</p>
<p>Definition 1. <italic>A hyperspherical map <inline-formula><mml:math id="M32"><mml:mi>&#x003C6;</mml:mi><mml:mo>:</mml:mo><mml:msubsup><mml:mrow><mml:mi>&#x0211D;</mml:mi></mml:mrow><mml:mrow><mml:mo>&#x02265;</mml:mo><mml:mn>0</mml:mn></mml:mrow><mml:mrow><mml:mi>n</mml:mi></mml:mrow></mml:msubsup><mml:mo>\</mml:mo><mml:mrow><mml:mo>{</mml:mo><mml:mrow><mml:mn>0</mml:mn></mml:mrow><mml:mo>}</mml:mo></mml:mrow><mml:mo>&#x02192;</mml:mo><mml:msup><mml:mrow><mml:mi>S</mml:mi></mml:mrow><mml:mrow><mml:mi>n</mml:mi><mml:mo>-</mml:mo><mml:mn>1</mml:mn></mml:mrow></mml:msup></mml:math></inline-formula> maps a vector</italic> <bold>x</bold>, <italic>with <italic>x</italic><sub><italic>i</italic></sub> &#x02265; 0 and <inline-formula><mml:math id="M33"><mml:munderover accentunder="false" accent="false"><mml:mrow><mml:mo>&#x02211;</mml:mo></mml:mrow><mml:mrow><mml:mi>i</mml:mi><mml:mo>=</mml:mo><mml:mn>1</mml:mn></mml:mrow><mml:mrow><mml:mi>n</mml:mi></mml:mrow></mml:munderover><mml:msub><mml:mrow><mml:mi>x</mml:mi></mml:mrow><mml:mrow><mml:mi>i</mml:mi></mml:mrow></mml:msub><mml:mo>&#x0003E;</mml:mo><mml:mn>0</mml:mn></mml:math></inline-formula>, to a unit vector <inline-formula><mml:math id="M34"><mml:mover accent="true"><mml:mrow><mml:mi>x</mml:mi></mml:mrow><mml:mo>^</mml:mo></mml:mover><mml:mo>&#x02208;</mml:mo><mml:msup><mml:mrow><mml:mi>S</mml:mi></mml:mrow><mml:mrow><mml:mi>n</mml:mi><mml:mo>-</mml:mo><mml:mn>1</mml:mn></mml:mrow></mml:msup></mml:math></inline-formula> where <inline-formula><mml:math id="M35"><mml:msub><mml:mrow><mml:mrow><mml:mo stretchy="false">(</mml:mo><mml:mrow><mml:mover accent="true"><mml:mrow><mml:mi>x</mml:mi></mml:mrow><mml:mo>^</mml:mo></mml:mover></mml:mrow><mml:mo stretchy="false">)</mml:mo></mml:mrow></mml:mrow><mml:mrow><mml:mi>i</mml:mi></mml:mrow></mml:msub><mml:mo>&#x02261;</mml:mo><mml:msqrt><mml:mrow><mml:msub><mml:mrow><mml:mi>x</mml:mi></mml:mrow><mml:mrow><mml:mi>i</mml:mi></mml:mrow></mml:msub><mml:mo>/</mml:mo><mml:munderover accentunder="false" accent="false"><mml:mrow><mml:mo>&#x02211;</mml:mo></mml:mrow><mml:mrow><mml:mi>j</mml:mi><mml:mo>=</mml:mo><mml:mn>1</mml:mn></mml:mrow><mml:mrow><mml:mi>n</mml:mi></mml:mrow></mml:munderover><mml:msub><mml:mrow><mml:mi>x</mml:mi></mml:mrow><mml:mrow><mml:mi>j</mml:mi></mml:mrow></mml:msub></mml:mrow></mml:msqrt></mml:math></inline-formula></italic>.</p>
<p>We will henceforth denote the image of a feature vector <bold>x</bold> under the hyperspherical map as <inline-formula><mml:math id="M36"><mml:mover accent="true"><mml:mrow><mml:mi>x</mml:mi></mml:mrow><mml:mo>^</mml:mo></mml:mover></mml:math></inline-formula>. The notion of neighborhood requires a well-defined measurement of distance on the hypersphere, which is naturally the great arc length&#x02014;the geodesic on a hypersphere. Both parametrix approximation and exact solution employ the great arc length, which is related to the following definition of cosine similarity:</p>
<p>Definition 2. <italic>The generic cosine similarity between two feature vectors</italic> <bold>x</bold>, <bold>y</bold> &#x02208; &#x0211D;<sup><italic>n</italic></sup>\{0} <italic>is</italic></p>
<disp-formula id="E1"><mml:math id="M1"><mml:mrow><mml:mi>cos</mml:mi><mml:mi>&#x003B8;</mml:mi><mml:mo>&#x02261;</mml:mo><mml:mfrac><mml:mrow><mml:mstyle mathvariant='bold' mathsize='normal'><mml:mi>x</mml:mi></mml:mstyle><mml:mo>&#x000B7;</mml:mo><mml:mstyle mathvariant='bold' mathsize='normal'><mml:mi>y</mml:mi></mml:mstyle></mml:mrow><mml:mrow><mml:mrow><mml:mo>&#x02016;</mml:mo><mml:mstyle mathvariant='bold' mathsize='normal'><mml:mi>x</mml:mi></mml:mstyle><mml:mo>&#x02016;</mml:mo></mml:mrow><mml:mrow><mml:mo>&#x02016;</mml:mo><mml:mstyle mathvariant='bold' mathsize='normal'><mml:mi>y</mml:mi></mml:mstyle><mml:mo>&#x02016;</mml:mo></mml:mrow></mml:mrow></mml:mfrac><mml:mo>,</mml:mo></mml:mrow></mml:math></disp-formula>
<p><italic>where ||&#x000B7;|| is the Euclidean <italic>L</italic><sup>2</sup>-norm, and &#x003B8; &#x02208; [0, &#x003C0;] is the great arc length on <italic>S</italic><sup><italic>n</italic> &#x02212; 1</sup>. For unit vectors <inline-formula><mml:math id="M37"><mml:mover accent="true"><mml:mrow><mml:mi>x</mml:mi></mml:mrow><mml:mo>^</mml:mo></mml:mover><mml:mo>=</mml:mo><mml:mi>&#x003C6;</mml:mi><mml:mrow><mml:mo stretchy="false">(</mml:mo><mml:mrow><mml:mstyle mathvariant="bold"><mml:mtext>x</mml:mtext></mml:mstyle></mml:mrow><mml:mo stretchy="false">)</mml:mo></mml:mrow></mml:math></inline-formula> and &#x00177; &#x0003D; &#x003C6;(<bold>y</bold>) obtained from non-negative feature vectors <inline-formula><mml:math id="M38"><mml:mstyle mathvariant="bold"><mml:mtext>x</mml:mtext></mml:mstyle><mml:mo>,</mml:mo><mml:mstyle mathvariant="bold"><mml:mtext>y</mml:mtext></mml:mstyle><mml:mo>&#x02208;</mml:mo><mml:msubsup><mml:mrow><mml:mi>&#x0211D;</mml:mi></mml:mrow><mml:mrow><mml:mo>&#x02265;</mml:mo><mml:mn>0</mml:mn></mml:mrow><mml:mrow><mml:mi>n</mml:mi></mml:mrow></mml:msubsup><mml:mo>\</mml:mo><mml:mrow><mml:mo>{</mml:mo><mml:mrow><mml:mn>0</mml:mn></mml:mrow><mml:mo>}</mml:mo></mml:mrow></mml:math></inline-formula> via the hyperspherical map, the cosine similarity reduces to the dot product <inline-formula><mml:math id="M39"><mml:mo class="qopname">cos</mml:mo><mml:mi>&#x003B8;</mml:mi><mml:mo>=</mml:mo><mml:mover accent="true"><mml:mrow><mml:mi>x</mml:mi></mml:mrow><mml:mo class="qopname">^</mml:mo></mml:mover><mml:mo>&#x000B7;</mml:mo><mml:mover accent="true"><mml:mrow><mml:mi>y</mml:mi></mml:mrow><mml:mo class="qopname">^</mml:mo></mml:mover></mml:math></inline-formula>; the non-negativity of</italic> <bold>x</bold> and <bold>y</bold> <italic>guarantees that &#x003B8; &#x02208; [0, &#x003C0;/2] in this case</italic>.</p>
<sec>
<title>2.1. Parametrix expansion</title>
<p>The parametrix kernel <italic>K</italic><sup>prx</sup> previously used in the literature is just a Gaussian RBF function with <inline-formula><mml:math id="M40"><mml:mi>&#x003B8;</mml:mi><mml:mo>=</mml:mo><mml:mo class="qopname">arccos</mml:mo><mml:mover accent="true"><mml:mrow><mml:mi>x</mml:mi></mml:mrow><mml:mo class="qopname">^</mml:mo></mml:mover><mml:mo>&#x000B7;</mml:mo><mml:mover accent="true"><mml:mrow><mml:mi>y</mml:mi></mml:mrow><mml:mo class="qopname">^</mml:mo></mml:mover></mml:math></inline-formula> as the radial distance [<xref ref-type="bibr" rid="B1">1</xref>]:</p>
<p>Definition 3. <italic>The parametrix kernel is a non-negative function</italic></p>
<disp-formula id="E2"><mml:math id="M2"><mml:mrow><mml:msup><mml:mi>K</mml:mi><mml:mrow><mml:mtext>prx</mml:mtext></mml:mrow></mml:msup><mml:mo stretchy='false'>(</mml:mo><mml:mover accent='true'><mml:mi>x</mml:mi><mml:mo>&#x0005E;</mml:mo></mml:mover><mml:mo>,</mml:mo><mml:mover accent='true'><mml:mi>y</mml:mi><mml:mo>&#x0005E;</mml:mo></mml:mover><mml:mo>;</mml:mo><mml:mi>t</mml:mi><mml:mo stretchy='false'>)</mml:mo><mml:mo>=</mml:mo><mml:msup><mml:mtext>e</mml:mtext><mml:mrow><mml:mo>&#x02212;</mml:mo><mml:mfrac><mml:mrow><mml:msup><mml:mrow><mml:mi>arccos</mml:mi></mml:mrow><mml:mn>2</mml:mn></mml:msup><mml:mover accent='true'><mml:mi>x</mml:mi><mml:mo>&#x0005E;</mml:mo></mml:mover><mml:mo>&#x000B7;</mml:mo><mml:mover accent='true'><mml:mi>y</mml:mi><mml:mo>&#x0005E;</mml:mo></mml:mover></mml:mrow><mml:mrow><mml:mn>4</mml:mn><mml:mi>t</mml:mi></mml:mrow></mml:mfrac></mml:mrow></mml:msup><mml:mo>=</mml:mo><mml:msup><mml:mtext>e</mml:mtext><mml:mrow><mml:mo>&#x02212;</mml:mo><mml:mfrac><mml:mrow><mml:msup><mml:mi>&#x003B8;</mml:mi><mml:mn>2</mml:mn></mml:msup></mml:mrow><mml:mrow><mml:mn>4</mml:mn><mml:mi>t</mml:mi></mml:mrow></mml:mfrac></mml:mrow></mml:msup><mml:mo>,</mml:mo></mml:mrow></mml:math></disp-formula>
<p><italic>defined for t &#x0003E; 0 and attaining global maximum 1 at</italic> &#x003B8; &#x0003D; 0.</p>
<p>Note that this kernel is assumed to be restricted to the positive orthant. The normalization factor <inline-formula><mml:math id="M41"><mml:msup><mml:mrow><mml:mrow><mml:mo stretchy="false">(</mml:mo><mml:mrow><mml:mn>4</mml:mn><mml:mi>&#x003C0;</mml:mi><mml:mi>t</mml:mi></mml:mrow><mml:mo stretchy="false">)</mml:mo></mml:mrow></mml:mrow><mml:mrow><mml:mo>-</mml:mo><mml:mfrac><mml:mrow><mml:mi>n</mml:mi><mml:mo>-</mml:mo><mml:mn>1</mml:mn></mml:mrow><mml:mrow><mml:mn>2</mml:mn></mml:mrow></mml:mfrac></mml:mrow></mml:msup></mml:math></inline-formula> is numerically unstable as <italic>t</italic> &#x02193; 0 and complicates hyperparameter tuning; as a global scaling factor of the kernel can be absorbed into the misclassification <italic>C</italic>-parameter in SVM, this overall normalization term is ignored in this paper. Importantly, the parametrix kernel <italic>K</italic><sup>prx</sup> is merely the Gaussian multiplicative factor without any asymptotic expansion terms in the full parametrix expansion <italic>G</italic><sup>prx</sup> of the heat kernel on a hypersphere [<xref ref-type="bibr" rid="B1">1</xref>, <xref ref-type="bibr" rid="B12">12</xref>], as described below. The leading order term in the full parametrix expansion was derived in Berger et al. [<xref ref-type="bibr" rid="B12">12</xref>] and quoted in Lafferty and Lebanon [<xref ref-type="bibr" rid="B1">1</xref>]. For the sake of completeness, we include the details here and also extend the calculation to higher order terms.</p>
<p>The Laplace operator on manifold <inline-formula><mml:math id="M42"><mml:mrow><mml:mi mathvariant="-tex-caligraphic">M</mml:mi></mml:mrow></mml:math></inline-formula> equiped with a Riemannian metric <italic>g</italic><sub>&#x003BC;&#x003BD;</sub> acts on a function <italic>f</italic> that depends only on the geodesic distance <italic>r</italic> from a fixed point as</p>
<disp-formula id="E3"><label>(1)</label><mml:math id="M3"><mml:mrow><mml:mi>&#x00394;</mml:mi><mml:mi>f</mml:mi><mml:mo stretchy='false'>(</mml:mo><mml:mi>r</mml:mi><mml:mo stretchy='false'>)</mml:mo><mml:mo>=</mml:mo><mml:msup><mml:mi>f</mml:mi><mml:mo>&#x02033;</mml:mo></mml:msup><mml:mo stretchy='false'>(</mml:mo><mml:mi>r</mml:mi><mml:mo stretchy='false'>)</mml:mo><mml:mo>+</mml:mo><mml:msup><mml:mrow><mml:mrow><mml:mo>(</mml:mo><mml:mrow><mml:mi>log</mml:mi><mml:msqrt><mml:mi>g</mml:mi></mml:msqrt></mml:mrow><mml:mo>)</mml:mo></mml:mrow></mml:mrow><mml:mo>&#x02032;</mml:mo></mml:msup><mml:msup><mml:mi>f</mml:mi><mml:mo>&#x02032;</mml:mo></mml:msup><mml:mo stretchy='false'>(</mml:mo><mml:mi>r</mml:mi><mml:mo stretchy='false'>)</mml:mo><mml:mo>,</mml:mo></mml:mrow></mml:math></disp-formula>
<p>where <italic>g</italic> &#x02261; det(<italic>g</italic><sub>&#x003BC;&#x003BD;</sub>) and &#x02032; denotes the radial derivative. Due to the nonvanishing metric derivative in Equation (1), the canonical diffusion function</p>
<disp-formula id="E4"><label>(2)</label><mml:math id="M4"><mml:mrow><mml:mi>G</mml:mi><mml:mo stretchy='false'>(</mml:mo><mml:mi>r</mml:mi><mml:mo>,</mml:mo><mml:mi>t</mml:mi><mml:mo stretchy='false'>)</mml:mo><mml:mo>=</mml:mo><mml:msup><mml:mrow><mml:mrow><mml:mo>(</mml:mo><mml:mrow><mml:mfrac><mml:mn>1</mml:mn><mml:mrow><mml:mn>4</mml:mn><mml:mi>&#x003C0;</mml:mi><mml:mi>t</mml:mi></mml:mrow></mml:mfrac></mml:mrow><mml:mo>)</mml:mo></mml:mrow></mml:mrow><mml:mrow><mml:mfrac><mml:mi>d</mml:mi><mml:mn>2</mml:mn></mml:mfrac></mml:mrow></mml:msup><mml:mi>exp</mml:mi><mml:mrow><mml:mo>(</mml:mo><mml:mrow><mml:mo>&#x02212;</mml:mo><mml:mfrac><mml:mrow><mml:msup><mml:mi>r</mml:mi><mml:mn>2</mml:mn></mml:msup></mml:mrow><mml:mrow><mml:mn>4</mml:mn><mml:mi>t</mml:mi></mml:mrow></mml:mfrac></mml:mrow><mml:mo>)</mml:mo></mml:mrow></mml:mrow></mml:math></disp-formula>
<p>does not satisfy the heat equation; that is, (&#x00394; &#x02212; &#x02202;<sub><italic>t</italic></sub>)<italic>G</italic>(<italic>r, t</italic>) &#x02260; 0 (Supplementary Material, section 2). For sufficiently small time <italic>t</italic> and geodesic distance <italic>r</italic>, the parametrix expansion of the heat kernel on a full hypersphere proposes an approximate solution</p>
<disp-formula id="E5"><mml:math id="M5"><mml:mrow><mml:msub><mml:mi>K</mml:mi><mml:mi>p</mml:mi></mml:msub><mml:mo stretchy='false'>(</mml:mo><mml:mi>r</mml:mi><mml:mo>,</mml:mo><mml:mi>t</mml:mi><mml:mo stretchy='false'>)</mml:mo><mml:mo>=</mml:mo><mml:mi>G</mml:mi><mml:mo stretchy='false'>(</mml:mo><mml:mi>r</mml:mi><mml:mo>,</mml:mo><mml:mi>t</mml:mi><mml:mo stretchy='false'>)</mml:mo><mml:mrow><mml:mo>(</mml:mo><mml:mrow><mml:msub><mml:mi>u</mml:mi><mml:mn>0</mml:mn></mml:msub><mml:mo stretchy='false'>(</mml:mo><mml:mi>r</mml:mi><mml:mo stretchy='false'>)</mml:mo><mml:mo>+</mml:mo><mml:msub><mml:mi>u</mml:mi><mml:mn>1</mml:mn></mml:msub><mml:mo stretchy='false'>(</mml:mo><mml:mi>r</mml:mi><mml:mo stretchy='false'>)</mml:mo><mml:mi>t</mml:mi><mml:mo>+</mml:mo><mml:msub><mml:mi>u</mml:mi><mml:mn>2</mml:mn></mml:msub><mml:mo stretchy='false'>(</mml:mo><mml:mi>r</mml:mi><mml:mo stretchy='false'>)</mml:mo><mml:msup><mml:mi>t</mml:mi><mml:mn>2</mml:mn></mml:msup><mml:mo>+</mml:mo><mml:mo>&#x022EF;</mml:mo><mml:mo>+</mml:mo><mml:msub><mml:mi>u</mml:mi><mml:mi>p</mml:mi></mml:msub><mml:mo stretchy='false'>(</mml:mo><mml:mi>r</mml:mi><mml:mo stretchy='false'>)</mml:mo><mml:msup><mml:mi>t</mml:mi><mml:mi>p</mml:mi></mml:msup></mml:mrow><mml:mo>)</mml:mo></mml:mrow><mml:mo>,</mml:mo></mml:mrow></mml:math></disp-formula>
<p>where the functions <italic>u</italic><sub><italic>i</italic></sub> should be found such that <italic>K</italic><sub><italic>p</italic></sub> satisfies the heat equation to order <italic>t</italic><sup><italic>p</italic>&#x02212;<italic>d</italic>/2</sup>, which is small for <italic>t</italic> &#x0226A; 1 and <italic>p</italic> &#x0003E; <italic>d</italic>/2; more precisely, we seek <italic>u</italic><sub><italic>i</italic></sub> such that</p>
<disp-formula id="E6"><label>(3)</label><mml:math id="M6"><mml:mrow><mml:mo stretchy='false'>(</mml:mo><mml:mi>&#x00394;</mml:mi><mml:mo>&#x02212;</mml:mo><mml:msub><mml:mo>&#x02202;</mml:mo><mml:mi>t</mml:mi></mml:msub><mml:mo stretchy='false'>)</mml:mo><mml:msub><mml:mi>K</mml:mi><mml:mi>p</mml:mi></mml:msub><mml:mo>=</mml:mo><mml:mi>G</mml:mi><mml:msup><mml:mi>t</mml:mi><mml:mi>p</mml:mi></mml:msup><mml:mi>&#x00394;</mml:mi><mml:msub><mml:mi>u</mml:mi><mml:mi>p</mml:mi></mml:msub><mml:mo>.</mml:mo></mml:mrow></mml:math></disp-formula>
<p>Taking the time derivative of <italic>K</italic><sub><italic>p</italic></sub> yields</p>
<disp-formula id="E7"><mml:math id="M7"><mml:mtable columnalign='left'><mml:mtr><mml:mtd><mml:msub><mml:mo>&#x02202;</mml:mo><mml:mi>t</mml:mi></mml:msub><mml:msub><mml:mi>K</mml:mi><mml:mi>p</mml:mi></mml:msub><mml:mo>=</mml:mo><mml:mi>G</mml:mi><mml:mtext>&#x000A0;</mml:mtext><mml:mo>&#x000B7;</mml:mo><mml:mtext>&#x000A0;</mml:mtext><mml:mrow><mml:mo>[</mml:mo><mml:mrow><mml:mrow><mml:mo>(</mml:mo><mml:mrow><mml:mo>&#x02212;</mml:mo><mml:mfrac><mml:mi>d</mml:mi><mml:mrow><mml:mn>2</mml:mn><mml:mi>t</mml:mi></mml:mrow></mml:mfrac><mml:mo>+</mml:mo><mml:mfrac><mml:mrow><mml:msup><mml:mi>r</mml:mi><mml:mn>2</mml:mn></mml:msup></mml:mrow><mml:mrow><mml:mn>4</mml:mn><mml:msup><mml:mi>t</mml:mi><mml:mn>2</mml:mn></mml:msup></mml:mrow></mml:mfrac></mml:mrow><mml:mo>)</mml:mo></mml:mrow><mml:mrow><mml:mo>(</mml:mo><mml:mrow><mml:msub><mml:mi>u</mml:mi><mml:mn>0</mml:mn></mml:msub><mml:mo>+</mml:mo><mml:msub><mml:mi>u</mml:mi><mml:mn>1</mml:mn></mml:msub><mml:mi>t</mml:mi><mml:mo>+</mml:mo><mml:msub><mml:mi>u</mml:mi><mml:mn>2</mml:mn></mml:msub><mml:msup><mml:mi>t</mml:mi><mml:mn>2</mml:mn></mml:msup><mml:mo>+</mml:mo><mml:mo>&#x022EF;</mml:mo><mml:mo>+</mml:mo><mml:msub><mml:mi>u</mml:mi><mml:mi>p</mml:mi></mml:msub><mml:msup><mml:mi>t</mml:mi><mml:mi>p</mml:mi></mml:msup></mml:mrow><mml:mo>)</mml:mo></mml:mrow></mml:mrow></mml:mrow></mml:mtd></mml:mtr><mml:mtr><mml:mtd><mml:mtext>&#x000A0;&#x000A0;&#x000A0;&#x000A0;&#x000A0;&#x000A0;&#x000A0;&#x000A0;&#x000A0;&#x000A0;&#x000A0;</mml:mtext><mml:mo>+</mml:mo><mml:mtext>&#x000A0;</mml:mtext><mml:mrow><mml:mrow><mml:mrow><mml:mo>(</mml:mo><mml:mrow><mml:msub><mml:mi>u</mml:mi><mml:mn>1</mml:mn></mml:msub><mml:mo>+</mml:mo><mml:mn>2</mml:mn><mml:msub><mml:mi>u</mml:mi><mml:mn>2</mml:mn></mml:msub><mml:mi>t</mml:mi><mml:mo>+</mml:mo><mml:mo>&#x022EF;</mml:mo><mml:mo>+</mml:mo><mml:mi>p</mml:mi><mml:msub><mml:mi>u</mml:mi><mml:mi>p</mml:mi></mml:msub><mml:msup><mml:mi>t</mml:mi><mml:mrow><mml:mi>p</mml:mi><mml:mo>&#x02212;</mml:mo><mml:mn>1</mml:mn></mml:mrow></mml:msup></mml:mrow><mml:mo>)</mml:mo></mml:mrow></mml:mrow><mml:mo>]</mml:mo></mml:mrow><mml:mo>,</mml:mo></mml:mtd></mml:mtr></mml:mtable></mml:math></disp-formula>
<p>while the Laplacian of <italic>K</italic><sub><italic>p</italic></sub> is</p>
<disp-formula id="E8"><mml:math id="M8"><mml:mtable columnalign='left'><mml:mtr><mml:mtd><mml:mi>&#x00394;</mml:mi><mml:msub><mml:mi>K</mml:mi><mml:mi>p</mml:mi></mml:msub><mml:mo>=</mml:mo><mml:mrow><mml:mo>(</mml:mo><mml:mrow><mml:msub><mml:mi>u</mml:mi><mml:mn>0</mml:mn></mml:msub><mml:mo>+</mml:mo><mml:msub><mml:mi>u</mml:mi><mml:mn>1</mml:mn></mml:msub><mml:mi>t</mml:mi><mml:mo>+</mml:mo><mml:mo>&#x022EF;</mml:mo><mml:mo>+</mml:mo><mml:msub><mml:mi>u</mml:mi><mml:mi>p</mml:mi></mml:msub><mml:msup><mml:mi>t</mml:mi><mml:mi>p</mml:mi></mml:msup></mml:mrow><mml:mo>)</mml:mo></mml:mrow><mml:mi>&#x00394;</mml:mi><mml:mi>G</mml:mi></mml:mtd></mml:mtr><mml:mtr><mml:mtd><mml:mtext>&#x000A0;&#x000A0;&#x000A0;&#x000A0;&#x000A0;&#x000A0;&#x000A0;&#x000A0;&#x000A0;&#x000A0;</mml:mtext><mml:mo>+</mml:mo><mml:mtext>&#x000A0;</mml:mtext><mml:mi>G</mml:mi><mml:mi>&#x00394;</mml:mi><mml:mrow><mml:mo>(</mml:mo><mml:mrow><mml:msub><mml:mi>u</mml:mi><mml:mn>0</mml:mn></mml:msub><mml:mo>+</mml:mo><mml:msub><mml:mi>u</mml:mi><mml:mn>1</mml:mn></mml:msub><mml:mi>t</mml:mi><mml:mo>+</mml:mo><mml:mo>&#x022EF;</mml:mo><mml:mo>+</mml:mo><mml:msub><mml:mi>u</mml:mi><mml:mi>p</mml:mi></mml:msub><mml:msup><mml:mi>t</mml:mi><mml:mi>p</mml:mi></mml:msup></mml:mrow><mml:mo>)</mml:mo></mml:mrow></mml:mtd></mml:mtr><mml:mtr><mml:mtd><mml:mtext>&#x000A0;&#x000A0;&#x000A0;&#x000A0;&#x000A0;&#x000A0;&#x000A0;&#x000A0;&#x000A0;&#x000A0;</mml:mtext><mml:mo>+</mml:mo><mml:mtext>&#x000A0;</mml:mtext><mml:mn>2</mml:mn><mml:msup><mml:mi>G</mml:mi><mml:mo>&#x02032;</mml:mo></mml:msup><mml:msup><mml:mrow><mml:mo>(</mml:mo><mml:mrow><mml:msub><mml:mi>u</mml:mi><mml:mn>0</mml:mn></mml:msub><mml:mo>+</mml:mo><mml:msub><mml:mi>u</mml:mi><mml:mn>1</mml:mn></mml:msub><mml:mi>t</mml:mi><mml:mo>+</mml:mo><mml:mo>&#x022EF;</mml:mo><mml:mo>+</mml:mo><mml:msub><mml:mi>u</mml:mi><mml:mi>p</mml:mi></mml:msub><mml:msup><mml:mi>t</mml:mi><mml:mi>p</mml:mi></mml:msup></mml:mrow><mml:mo>)</mml:mo></mml:mrow><mml:mo>&#x02032;</mml:mo></mml:msup><mml:mo>.</mml:mo></mml:mtd></mml:mtr></mml:mtable></mml:math></disp-formula>
<p>One can easily compute</p>
<disp-formula id="E9"><mml:math id="M9"><mml:mrow><mml:mi>&#x00394;</mml:mi><mml:mi>G</mml:mi><mml:mo>=</mml:mo><mml:mrow><mml:mo>[</mml:mo><mml:mrow><mml:mrow><mml:mo>(</mml:mo><mml:mrow><mml:mo>&#x02212;</mml:mo><mml:mfrac><mml:mn>1</mml:mn><mml:mrow><mml:mn>2</mml:mn><mml:mi>t</mml:mi></mml:mrow></mml:mfrac><mml:mo>+</mml:mo><mml:mfrac><mml:mrow><mml:msup><mml:mi>r</mml:mi><mml:mn>2</mml:mn></mml:msup></mml:mrow><mml:mrow><mml:mn>4</mml:mn><mml:msup><mml:mi>t</mml:mi><mml:mn>2</mml:mn></mml:msup></mml:mrow></mml:mfrac></mml:mrow><mml:mo>)</mml:mo></mml:mrow><mml:mo>&#x02212;</mml:mo><mml:mfrac><mml:mi>r</mml:mi><mml:mrow><mml:mn>2</mml:mn><mml:mi>t</mml:mi></mml:mrow></mml:mfrac><mml:msup><mml:mrow><mml:mo stretchy='false'>(</mml:mo><mml:mi>log</mml:mi><mml:msqrt><mml:mi>g</mml:mi></mml:msqrt><mml:mo stretchy='false'>)</mml:mo></mml:mrow><mml:mo>&#x02032;</mml:mo></mml:msup></mml:mrow><mml:mo>]</mml:mo></mml:mrow><mml:mi>G</mml:mi></mml:mrow></mml:math></disp-formula>
<p>and</p>
<disp-formula id="E10"><mml:math id="M10"><mml:mrow><mml:msup><mml:mi>G</mml:mi><mml:mo>&#x02032;</mml:mo></mml:msup><mml:msup><mml:mrow><mml:mrow><mml:mo>(</mml:mo><mml:mrow><mml:msub><mml:mi>u</mml:mi><mml:mn>0</mml:mn></mml:msub><mml:mo>+</mml:mo><mml:msub><mml:mi>u</mml:mi><mml:mn>1</mml:mn></mml:msub><mml:mi>t</mml:mi><mml:mo>+</mml:mo><mml:mo>&#x022EF;</mml:mo></mml:mrow><mml:mo>)</mml:mo></mml:mrow></mml:mrow><mml:mo>&#x02032;</mml:mo></mml:msup><mml:mo>=</mml:mo><mml:mo>&#x02212;</mml:mo><mml:mfrac><mml:mi>r</mml:mi><mml:mrow><mml:mn>2</mml:mn><mml:mi>t</mml:mi></mml:mrow></mml:mfrac><mml:mrow><mml:mo>(</mml:mo><mml:mrow><mml:msubsup><mml:mi>u</mml:mi><mml:mn>0</mml:mn><mml:mo>&#x02032;</mml:mo></mml:msubsup><mml:mo>+</mml:mo><mml:msubsup><mml:mi>u</mml:mi><mml:mn>1</mml:mn><mml:mo>&#x02032;</mml:mo></mml:msubsup><mml:mi>t</mml:mi><mml:mo>+</mml:mo><mml:mo>&#x022EF;</mml:mo></mml:mrow><mml:mo>)</mml:mo></mml:mrow><mml:mi>G</mml:mi><mml:mo>.</mml:mo></mml:mrow></mml:math></disp-formula>
<p>The left-hand side of Equation (3) is thus equal to <italic>G</italic> multiplied by</p>
<disp-formula id="E11"><mml:math id="M11"><mml:mtable columnalign='left'><mml:mtr><mml:mtd><mml:mrow><mml:mo>(</mml:mo><mml:mrow><mml:msub><mml:mi>u</mml:mi><mml:mn>0</mml:mn></mml:msub><mml:mo>+</mml:mo><mml:mo>&#x022EF;</mml:mo><mml:mo>+</mml:mo><mml:msub><mml:mi>u</mml:mi><mml:mi>p</mml:mi></mml:msub><mml:msup><mml:mi>t</mml:mi><mml:mi>p</mml:mi></mml:msup></mml:mrow><mml:mo>)</mml:mo></mml:mrow><mml:mrow><mml:mo>[</mml:mo><mml:mrow><mml:mo>&#x02212;</mml:mo><mml:mfrac><mml:mi>r</mml:mi><mml:mrow><mml:mn>2</mml:mn><mml:mi>t</mml:mi></mml:mrow></mml:mfrac><mml:mo stretchy='false'>(</mml:mo><mml:mi>log</mml:mi><mml:msqrt><mml:mi>g</mml:mi></mml:msqrt><mml:msup><mml:mo stretchy='false'>)</mml:mo><mml:mo>&#x02032;</mml:mo></mml:msup><mml:mo>+</mml:mo><mml:mfrac><mml:mrow><mml:mi>d</mml:mi><mml:mo>&#x02212;</mml:mo><mml:mn>1</mml:mn></mml:mrow><mml:mrow><mml:mn>2</mml:mn><mml:mi>t</mml:mi></mml:mrow></mml:mfrac></mml:mrow><mml:mo>]</mml:mo></mml:mrow><mml:mo>+</mml:mo><mml:mi>&#x00394;</mml:mi><mml:mrow><mml:mo>(</mml:mo><mml:mrow><mml:msub><mml:mi>u</mml:mi><mml:mn>0</mml:mn></mml:msub><mml:mo>+</mml:mo><mml:mo>&#x022EF;</mml:mo><mml:mo>+</mml:mo><mml:msub><mml:mi>u</mml:mi><mml:mi>p</mml:mi></mml:msub><mml:msup><mml:mi>t</mml:mi><mml:mi>p</mml:mi></mml:msup></mml:mrow><mml:mo>)</mml:mo></mml:mrow><mml:mo>+</mml:mo></mml:mtd></mml:mtr><mml:mtr><mml:mtd><mml:mtext>&#x000A0;&#x000A0;&#x000A0;&#x000A0;&#x000A0;&#x000A0;&#x000A0;&#x000A0;&#x000A0;&#x000A0;&#x000A0;&#x000A0;&#x000A0;&#x000A0;&#x000A0;&#x000A0;&#x000A0;&#x000A0;</mml:mtext><mml:mo>&#x02212;</mml:mo><mml:mfrac><mml:mi>r</mml:mi><mml:mi>t</mml:mi></mml:mfrac><mml:mrow><mml:mo>(</mml:mo><mml:mrow><mml:msubsup><mml:mi>u</mml:mi><mml:mn>0</mml:mn><mml:mo>&#x02032;</mml:mo></mml:msubsup><mml:mo>+</mml:mo><mml:mo>&#x022EF;</mml:mo><mml:mo>+</mml:mo><mml:msubsup><mml:mi>u</mml:mi><mml:mi>p</mml:mi><mml:mo>&#x02032;</mml:mo></mml:msubsup><mml:msup><mml:mi>t</mml:mi><mml:mi>p</mml:mi></mml:msup></mml:mrow><mml:mo>)</mml:mo></mml:mrow><mml:mo>&#x02212;</mml:mo><mml:mrow><mml:mo>(</mml:mo><mml:mrow><mml:msub><mml:mi>u</mml:mi><mml:mn>1</mml:mn></mml:msub><mml:mo>+</mml:mo><mml:mn>2</mml:mn><mml:msub><mml:mi>u</mml:mi><mml:mn>2</mml:mn></mml:msub><mml:mi>t</mml:mi><mml:mo>+</mml:mo><mml:mo>&#x022EF;</mml:mo><mml:mo>+</mml:mo><mml:mi>p</mml:mi><mml:msub><mml:mi>u</mml:mi><mml:mi>p</mml:mi></mml:msub><mml:msup><mml:mi>t</mml:mi><mml:mrow><mml:mi>p</mml:mi><mml:mo>&#x02212;</mml:mo><mml:mn>1</mml:mn></mml:mrow></mml:msup></mml:mrow><mml:mo>)</mml:mo></mml:mrow><mml:mo>,</mml:mo></mml:mtd></mml:mtr></mml:mtable></mml:math></disp-formula>
<p>and we need to solve for <italic>u</italic><sub><italic>i</italic></sub> such that all the coefficients of <italic>t</italic><sup><italic>q</italic></sup> in this expression, for <italic>q</italic> &#x0003C; <italic>p</italic>, vanish.</p>
<p>For <italic>q</italic> &#x0003D; &#x02212;1, we need to solve</p>
<disp-formula id="E12"><mml:math id="M12"><mml:mrow><mml:msub><mml:mi>u</mml:mi><mml:mn>0</mml:mn></mml:msub><mml:mfrac><mml:mi>r</mml:mi><mml:mn>2</mml:mn></mml:mfrac><mml:mrow><mml:mo>[</mml:mo><mml:mrow><mml:mo>&#x02212;</mml:mo><mml:mo stretchy='false'>(</mml:mo><mml:mi>log</mml:mi><mml:msqrt><mml:mi>g</mml:mi></mml:msqrt><mml:msup><mml:mo stretchy='false'>)</mml:mo><mml:mo>&#x02032;</mml:mo></mml:msup><mml:mo>+</mml:mo><mml:mfrac><mml:mrow><mml:mi>d</mml:mi><mml:mo>&#x02212;</mml:mo><mml:mn>1</mml:mn></mml:mrow><mml:mi>r</mml:mi></mml:mfrac></mml:mrow><mml:mo>]</mml:mo></mml:mrow><mml:mo>=</mml:mo><mml:mi>r</mml:mi><mml:msubsup><mml:mi>u</mml:mi><mml:mn>0</mml:mn><mml:mo>&#x02032;</mml:mo></mml:msubsup><mml:mtext>&#x000A0;</mml:mtext><mml:mo>,</mml:mo></mml:mrow></mml:math></disp-formula>
<p>or equivalently,</p>
<disp-formula id="E13"><mml:math id="M13"><mml:mrow><mml:msup><mml:mrow><mml:mrow><mml:mo>(</mml:mo><mml:mrow><mml:mi>log</mml:mi><mml:msub><mml:mi>u</mml:mi><mml:mn>0</mml:mn></mml:msub></mml:mrow><mml:mo>)</mml:mo></mml:mrow></mml:mrow><mml:mo>&#x02032;</mml:mo></mml:msup><mml:mo>=</mml:mo><mml:mo>&#x02212;</mml:mo><mml:mfrac><mml:mn>1</mml:mn><mml:mn>2</mml:mn></mml:mfrac><mml:mo stretchy='false'>(</mml:mo><mml:mi>log</mml:mi><mml:msqrt><mml:mi>g</mml:mi></mml:msqrt><mml:msup><mml:mo stretchy='false'>)</mml:mo><mml:mo>&#x02032;</mml:mo></mml:msup><mml:mo>+</mml:mo><mml:mfrac><mml:mrow><mml:mi>d</mml:mi><mml:mo>&#x02212;</mml:mo><mml:mn>1</mml:mn></mml:mrow><mml:mrow><mml:mn>2</mml:mn><mml:mi>r</mml:mi></mml:mrow></mml:mfrac><mml:mo>.</mml:mo></mml:mrow></mml:math></disp-formula>
<p>Integrating with respect to <italic>r</italic> yields</p>
<disp-formula id="E14"><mml:math id="M14"><mml:mrow><mml:mi>log</mml:mi><mml:msub><mml:mi>u</mml:mi><mml:mn>0</mml:mn></mml:msub><mml:mo>=</mml:mo><mml:mo>&#x02212;</mml:mo><mml:mfrac><mml:mn>1</mml:mn><mml:mn>2</mml:mn></mml:mfrac><mml:mrow><mml:mo>[</mml:mo><mml:mrow><mml:mi>log</mml:mi><mml:msqrt><mml:mi>g</mml:mi></mml:msqrt><mml:mo>&#x02212;</mml:mo><mml:mo stretchy='false'>(</mml:mo><mml:mi>d</mml:mi><mml:mo>&#x02212;</mml:mo><mml:mn>1</mml:mn><mml:mo stretchy='false'>)</mml:mo><mml:mi>log</mml:mi><mml:mi>r</mml:mi></mml:mrow><mml:mo>]</mml:mo></mml:mrow><mml:mo>+</mml:mo><mml:mtext>const</mml:mtext><mml:mo>.</mml:mo><mml:mo>,</mml:mo></mml:mrow></mml:math></disp-formula>
<p>where we implicitly take only the radial part of <inline-formula><mml:math id="M43"><mml:mo class="qopname">log</mml:mo><mml:msqrt><mml:mrow><mml:mi>g</mml:mi></mml:mrow></mml:msqrt></mml:math></inline-formula>. Thus, we get</p>
<disp-formula id="E15"><mml:math id="M15"><mml:mrow><mml:msub><mml:mi>u</mml:mi><mml:mn>0</mml:mn></mml:msub><mml:mo>=</mml:mo><mml:mtext>const</mml:mtext><mml:mo>.</mml:mo><mml:mo>&#x000D7;</mml:mo><mml:msup><mml:mrow><mml:mrow><mml:mo>(</mml:mo><mml:mrow><mml:mfrac><mml:mrow><mml:msqrt><mml:mi>g</mml:mi></mml:msqrt></mml:mrow><mml:mrow><mml:msup><mml:mi>r</mml:mi><mml:mrow><mml:mi>d</mml:mi><mml:mo>&#x02212;</mml:mo><mml:mn>1</mml:mn></mml:mrow></mml:msup></mml:mrow></mml:mfrac></mml:mrow><mml:mo>)</mml:mo></mml:mrow></mml:mrow><mml:mrow><mml:mo>&#x02212;</mml:mo><mml:mfrac><mml:mn>1</mml:mn><mml:mn>2</mml:mn></mml:mfrac></mml:mrow></mml:msup><mml:mo>&#x0221D;</mml:mo><mml:msup><mml:mrow><mml:mrow><mml:mo>(</mml:mo><mml:mrow><mml:mfrac><mml:mrow><mml:mi>sin</mml:mi><mml:mi>r</mml:mi></mml:mrow><mml:mi>r</mml:mi></mml:mfrac></mml:mrow><mml:mo>)</mml:mo></mml:mrow></mml:mrow><mml:mrow><mml:mo>&#x02212;</mml:mo><mml:mfrac><mml:mrow><mml:mi>d</mml:mi><mml:mo>&#x02212;</mml:mo><mml:mn>1</mml:mn></mml:mrow><mml:mn>2</mml:mn></mml:mfrac></mml:mrow></mml:msup></mml:mrow></mml:math></disp-formula>
<p>as the zeroth-order term in the parametrix expansion. Using this expression of <italic>u</italic><sub>0</sub>, the remaining terms become</p>
<disp-formula id="E16"><mml:math id="M16"><mml:mtable columnalign='left'><mml:mtr><mml:mtd><mml:mi>r</mml:mi><mml:mrow><mml:mo>[</mml:mo><mml:mrow><mml:mrow><mml:mo>(</mml:mo><mml:mrow><mml:msub><mml:mi>u</mml:mi><mml:mn>1</mml:mn></mml:msub><mml:mo>+</mml:mo><mml:msub><mml:mi>u</mml:mi><mml:mn>2</mml:mn></mml:msub><mml:mi>t</mml:mi><mml:mo>+</mml:mo><mml:mo>&#x022EF;</mml:mo></mml:mrow><mml:mo>)</mml:mo></mml:mrow><mml:mo stretchy='false'>(</mml:mo><mml:mi>log</mml:mi><mml:msub><mml:mi>u</mml:mi><mml:mn>0</mml:mn></mml:msub><mml:msup><mml:mo stretchy='false'>)</mml:mo><mml:mo>&#x02032;</mml:mo></mml:msup><mml:mo>&#x02212;</mml:mo><mml:mrow><mml:mo>(</mml:mo><mml:mrow><mml:msubsup><mml:mi>u</mml:mi><mml:mn>1</mml:mn><mml:mo>&#x02032;</mml:mo></mml:msubsup><mml:mo>+</mml:mo><mml:msubsup><mml:mi>u</mml:mi><mml:mn>2</mml:mn><mml:mo>&#x02032;</mml:mo></mml:msubsup><mml:mi>t</mml:mi><mml:mo>+</mml:mo><mml:mo>&#x022EF;</mml:mo></mml:mrow><mml:mo>)</mml:mo></mml:mrow></mml:mrow><mml:mo>]</mml:mo></mml:mrow><mml:mo>+</mml:mo></mml:mtd></mml:mtr><mml:mtr><mml:mtd><mml:mtext>&#x000A0;&#x000A0;&#x000A0;&#x000A0;&#x000A0;</mml:mtext><mml:mo>+</mml:mo><mml:mrow><mml:mo>(</mml:mo><mml:mrow><mml:mi>&#x00394;</mml:mi><mml:msub><mml:mi>u</mml:mi><mml:mn>0</mml:mn></mml:msub><mml:mo>+</mml:mo><mml:mi>t</mml:mi><mml:mi>&#x00394;</mml:mi><mml:msub><mml:mi>u</mml:mi><mml:mn>1</mml:mn></mml:msub><mml:mo>+</mml:mo><mml:mo>&#x022EF;</mml:mo></mml:mrow><mml:mo>)</mml:mo></mml:mrow><mml:mo>&#x02212;</mml:mo><mml:mrow><mml:mo>(</mml:mo><mml:mrow><mml:msub><mml:mi>u</mml:mi><mml:mn>1</mml:mn></mml:msub><mml:mo>+</mml:mo><mml:mn>2</mml:mn><mml:msub><mml:mi>u</mml:mi><mml:mn>2</mml:mn></mml:msub><mml:mi>t</mml:mi><mml:mo>+</mml:mo><mml:mo>&#x022EF;</mml:mo></mml:mrow><mml:mo>)</mml:mo></mml:mrow><mml:mo>,</mml:mo></mml:mtd></mml:mtr></mml:mtable></mml:math></disp-formula>
<p>and we obtain the recursion relation</p>
<disp-formula id="E17"><mml:math id="M17"><mml:mrow><mml:msub><mml:mi>u</mml:mi><mml:mrow><mml:mi>k</mml:mi><mml:mo>+</mml:mo><mml:mn>1</mml:mn></mml:mrow></mml:msub><mml:mo stretchy='false'>(</mml:mo><mml:mi>log</mml:mi><mml:msub><mml:mi>u</mml:mi><mml:mn>0</mml:mn></mml:msub><mml:msup><mml:mo stretchy='false'>)</mml:mo><mml:mo>&#x02032;</mml:mo></mml:msup><mml:mo>&#x02212;</mml:mo><mml:msubsup><mml:mi>u</mml:mi><mml:mrow><mml:mi>k</mml:mi><mml:mo>+</mml:mo><mml:mn>1</mml:mn></mml:mrow><mml:mo>&#x02032;</mml:mo></mml:msubsup><mml:mo>=</mml:mo><mml:mo>&#x02212;</mml:mo><mml:mfrac><mml:mrow><mml:mi>&#x00394;</mml:mi><mml:msub><mml:mi>u</mml:mi><mml:mi>k</mml:mi></mml:msub><mml:mo>&#x02212;</mml:mo><mml:mo stretchy='false'>(</mml:mo><mml:mi>k</mml:mi><mml:mo>+</mml:mo><mml:mn>1</mml:mn><mml:mo stretchy='false'>)</mml:mo><mml:msub><mml:mi>u</mml:mi><mml:mrow><mml:mi>k</mml:mi><mml:mo>+</mml:mo><mml:mn>1</mml:mn></mml:mrow></mml:msub></mml:mrow><mml:mi>r</mml:mi></mml:mfrac><mml:mo>.</mml:mo></mml:mrow></mml:math></disp-formula>
<p>Algebraic manipulations show that</p>
<disp-formula id="E18"><mml:math id="M18"><mml:mrow><mml:mo stretchy='false'>(</mml:mo><mml:mi>log</mml:mi><mml:msup><mml:mi>r</mml:mi><mml:mrow><mml:mi>k</mml:mi><mml:mo>+</mml:mo><mml:mn>1</mml:mn></mml:mrow></mml:msup><mml:mo>&#x02212;</mml:mo><mml:mi>log</mml:mi><mml:msub><mml:mi>u</mml:mi><mml:mn>0</mml:mn></mml:msub><mml:mo>+</mml:mo><mml:mi>log</mml:mi><mml:msub><mml:mi>u</mml:mi><mml:mrow><mml:mi>k</mml:mi><mml:mo>+</mml:mo><mml:mn>1</mml:mn></mml:mrow></mml:msub><mml:msup><mml:mo stretchy='false'>)</mml:mo><mml:mo>&#x02032;</mml:mo></mml:msup><mml:msub><mml:mi>u</mml:mi><mml:mrow><mml:mi>k</mml:mi><mml:mo>+</mml:mo><mml:mn>1</mml:mn></mml:mrow></mml:msub><mml:mo>=</mml:mo><mml:msup><mml:mi>r</mml:mi><mml:mrow><mml:mo>&#x02212;</mml:mo><mml:mn>1</mml:mn></mml:mrow></mml:msup><mml:mi>&#x00394;</mml:mi><mml:msub><mml:mi>u</mml:mi><mml:mi>k</mml:mi></mml:msub><mml:mo>,</mml:mo></mml:mrow></mml:math></disp-formula>
<p>from which we get</p>
<disp-formula id="E19"><mml:math id="M19"><mml:mrow><mml:msup><mml:mrow><mml:mrow><mml:mo>(</mml:mo><mml:mrow><mml:mfrac><mml:mrow><mml:msub><mml:mi>u</mml:mi><mml:mrow><mml:mi>k</mml:mi><mml:mo>+</mml:mo><mml:mn>1</mml:mn></mml:mrow></mml:msub><mml:msup><mml:mi>r</mml:mi><mml:mrow><mml:mi>k</mml:mi><mml:mo>+</mml:mo><mml:mn>1</mml:mn></mml:mrow></mml:msup></mml:mrow><mml:mrow><mml:msub><mml:mi>u</mml:mi><mml:mn>0</mml:mn></mml:msub></mml:mrow></mml:mfrac></mml:mrow><mml:mo>)</mml:mo></mml:mrow></mml:mrow><mml:mo>&#x02032;</mml:mo></mml:msup><mml:mo>=</mml:mo><mml:msup><mml:mi>r</mml:mi><mml:mrow><mml:mo stretchy='false'>(</mml:mo><mml:mi>k</mml:mi><mml:mo>+</mml:mo><mml:mn>1</mml:mn><mml:mo stretchy='false'>)</mml:mo><mml:mo>&#x02212;</mml:mo><mml:mn>1</mml:mn></mml:mrow></mml:msup><mml:msubsup><mml:mi>u</mml:mi><mml:mn>0</mml:mn><mml:mrow><mml:mo>&#x02212;</mml:mo><mml:mn>1</mml:mn></mml:mrow></mml:msubsup><mml:mi>&#x00394;</mml:mi><mml:msub><mml:mi>u</mml:mi><mml:mi>k</mml:mi></mml:msub><mml:mo>.</mml:mo></mml:mrow></mml:math></disp-formula>
<p>Integrating this equation and rearranging terms, we finally get</p>
<disp-formula id="E20"><label>(4)</label><mml:math id="M20"><mml:mrow><mml:msub><mml:mi>u</mml:mi><mml:mrow><mml:mi>k</mml:mi><mml:mo>+</mml:mo><mml:mn>1</mml:mn></mml:mrow></mml:msub><mml:mo>=</mml:mo><mml:msup><mml:mi>r</mml:mi><mml:mrow><mml:mo>&#x02212;</mml:mo><mml:mo stretchy='false'>(</mml:mo><mml:mi>k</mml:mi><mml:mo>+</mml:mo><mml:mn>1</mml:mn><mml:mo stretchy='false'>)</mml:mo></mml:mrow></mml:msup><mml:msub><mml:mi>u</mml:mi><mml:mn>0</mml:mn></mml:msub><mml:mstyle displaystyle='true'><mml:mrow><mml:msubsup><mml:mo>&#x0222B;</mml:mo><mml:mn>0</mml:mn><mml:mi>r</mml:mi></mml:msubsup><mml:mi>d</mml:mi></mml:mrow></mml:mstyle><mml:mover accent='true'><mml:mi>r</mml:mi><mml:mo>&#x002DC;</mml:mo></mml:mover><mml:mtext>&#x000A0;</mml:mtext><mml:msup><mml:mover accent='true'><mml:mi>r</mml:mi><mml:mo>&#x002DC;</mml:mo></mml:mover><mml:mi>k</mml:mi></mml:msup><mml:msubsup><mml:mi>u</mml:mi><mml:mn>0</mml:mn><mml:mrow><mml:mo>&#x02212;</mml:mo><mml:mn>1</mml:mn></mml:mrow></mml:msubsup><mml:mi>&#x00394;</mml:mi><mml:msub><mml:mi>u</mml:mi><mml:mi>k</mml:mi></mml:msub><mml:mo>.</mml:mo></mml:mrow></mml:math></disp-formula>
<p>Setting <italic>k</italic> &#x0003D; 0 in this recursion equation, we find the second correction term to be</p>
<disp-formula id="E21"><mml:math id="M21"><mml:mtable columnalign='left'><mml:mtr><mml:mtd><mml:msub><mml:mi>u</mml:mi><mml:mn>1</mml:mn></mml:msub><mml:mo>=</mml:mo><mml:mfrac><mml:mrow><mml:msub><mml:mi>u</mml:mi><mml:mn>0</mml:mn></mml:msub></mml:mrow><mml:mi>r</mml:mi></mml:mfrac><mml:mstyle displaystyle='true'><mml:mrow><mml:msubsup><mml:mo>&#x0222B;</mml:mo><mml:mn>0</mml:mn><mml:mi>r</mml:mi></mml:msubsup><mml:mi>d</mml:mi></mml:mrow></mml:mstyle><mml:mover accent='true'><mml:mi>r</mml:mi><mml:mo>&#x002DC;</mml:mo></mml:mover><mml:mtext>&#x000A0;</mml:mtext><mml:msubsup><mml:mi>u</mml:mi><mml:mn>0</mml:mn><mml:mrow><mml:mo>&#x02212;</mml:mo><mml:mn>1</mml:mn></mml:mrow></mml:msubsup><mml:mi>&#x00394;</mml:mi><mml:msub><mml:mi>u</mml:mi><mml:mn>0</mml:mn></mml:msub></mml:mtd></mml:mtr><mml:mtr><mml:mtd><mml:mtext>&#x000A0;&#x000A0;&#x000A0;</mml:mtext><mml:mo>=</mml:mo><mml:mfrac><mml:mrow><mml:msub><mml:mi>u</mml:mi><mml:mn>0</mml:mn></mml:msub></mml:mrow><mml:mi>r</mml:mi></mml:mfrac><mml:mstyle displaystyle='true'><mml:mrow><mml:msubsup><mml:mo>&#x0222B;</mml:mo><mml:mn>0</mml:mn><mml:mi>r</mml:mi></mml:msubsup><mml:mi>d</mml:mi></mml:mrow></mml:mstyle><mml:mover accent='true'><mml:mi>r</mml:mi><mml:mo>&#x002DC;</mml:mo></mml:mover><mml:mtext>&#x000A0;</mml:mtext><mml:msubsup><mml:mi>u</mml:mi><mml:mn>0</mml:mn><mml:mrow><mml:mo>&#x02212;</mml:mo><mml:mn>1</mml:mn></mml:mrow></mml:msubsup><mml:mrow><mml:mo>(</mml:mo><mml:mrow><mml:msubsup><mml:mi>u</mml:mi><mml:mn>0</mml:mn><mml:mrow><mml:mo>&#x02033;</mml:mo></mml:mrow></mml:msubsup><mml:mo>+</mml:mo><mml:msubsup><mml:mi>u</mml:mi><mml:mn>0</mml:mn><mml:mo>&#x02032;</mml:mo></mml:msubsup><mml:mo stretchy='false'>(</mml:mo><mml:mi>log</mml:mi><mml:msqrt><mml:mi>g</mml:mi></mml:msqrt><mml:msup><mml:mo stretchy='false'>)</mml:mo><mml:mo>&#x02032;</mml:mo></mml:msup></mml:mrow><mml:mo>)</mml:mo></mml:mrow><mml:mo>.</mml:mo></mml:mtd></mml:mtr></mml:mtable></mml:math></disp-formula>
<p>From our previously obtained solution for <italic>u</italic><sub>0</sub>, we find</p>
<disp-formula id="E22"><mml:math id="M22"><mml:mrow><mml:msubsup><mml:mi>u</mml:mi><mml:mn>0</mml:mn><mml:mo>&#x02032;</mml:mo></mml:msubsup><mml:mo>=</mml:mo><mml:mfrac><mml:mn>1</mml:mn><mml:mn>2</mml:mn></mml:mfrac><mml:mrow><mml:mo>(</mml:mo><mml:mrow><mml:mfrac><mml:mrow><mml:mi>d</mml:mi><mml:mo>&#x02212;</mml:mo><mml:mn>1</mml:mn></mml:mrow><mml:mi>r</mml:mi></mml:mfrac><mml:mo>&#x02212;</mml:mo><mml:mfrac><mml:msup><mml:mi>g</mml:mi><mml:mo>&#x02032;</mml:mo></mml:msup><mml:mrow><mml:mn>2</mml:mn><mml:mi>g</mml:mi></mml:mrow></mml:mfrac></mml:mrow><mml:mo>)</mml:mo></mml:mrow><mml:msub><mml:mi>u</mml:mi><mml:mn>0</mml:mn></mml:msub><mml:mo>,</mml:mo></mml:mrow></mml:math></disp-formula>
<p>and</p>
<disp-formula id="E23"><mml:math id="M23"><mml:mrow><mml:msubsup><mml:mi>u</mml:mi><mml:mn>0</mml:mn><mml:mrow><mml:mo>&#x02033;</mml:mo></mml:mrow></mml:msubsup><mml:mo>=</mml:mo><mml:mfrac><mml:mn>1</mml:mn><mml:mn>4</mml:mn></mml:mfrac><mml:mrow><mml:mo>[</mml:mo><mml:mrow><mml:mfrac><mml:mrow><mml:mo stretchy='false'>(</mml:mo><mml:mi>d</mml:mi><mml:mo>&#x02212;</mml:mo><mml:mn>1</mml:mn><mml:mo stretchy='false'>)</mml:mo><mml:mo stretchy='false'>(</mml:mo><mml:mi>d</mml:mi><mml:mo>&#x02212;</mml:mo><mml:mn>3</mml:mn><mml:mo stretchy='false'>)</mml:mo></mml:mrow><mml:mrow><mml:msup><mml:mi>r</mml:mi><mml:mn>2</mml:mn></mml:msup></mml:mrow></mml:mfrac><mml:mo>&#x02212;</mml:mo><mml:mfrac><mml:mrow><mml:msup><mml:mi>g</mml:mi><mml:mo>&#x02032;</mml:mo></mml:msup><mml:mo stretchy='false'>(</mml:mo><mml:mi>d</mml:mi><mml:mo>&#x02212;</mml:mo><mml:mn>1</mml:mn><mml:mo stretchy='false'>)</mml:mo></mml:mrow><mml:mrow><mml:mi>g</mml:mi><mml:mi>r</mml:mi></mml:mrow></mml:mfrac><mml:mo>&#x02212;</mml:mo><mml:mfrac><mml:msup><mml:mi>g</mml:mi><mml:mo>&#x02033;</mml:mo></mml:msup><mml:mi>g</mml:mi></mml:mfrac><mml:mo>+</mml:mo><mml:mfrac><mml:mn>5</mml:mn><mml:mn>4</mml:mn></mml:mfrac><mml:msup><mml:mrow><mml:mrow><mml:mo>(</mml:mo><mml:mrow><mml:mfrac><mml:msup><mml:mi>g</mml:mi><mml:mo>&#x02032;</mml:mo></mml:msup><mml:mi>g</mml:mi></mml:mfrac></mml:mrow><mml:mo>)</mml:mo></mml:mrow></mml:mrow><mml:mn>2</mml:mn></mml:msup></mml:mrow><mml:mo>]</mml:mo></mml:mrow><mml:msub><mml:mi>u</mml:mi><mml:mn>0</mml:mn></mml:msub><mml:mo>.</mml:mo></mml:mrow></mml:math></disp-formula>
<p>Substituting these expressions into the recursion relation for <italic>u</italic><sub>1</sub> yields</p>
<disp-formula id="E24"><mml:math id="M24"><mml:mrow><mml:msub><mml:mi>u</mml:mi><mml:mn>1</mml:mn></mml:msub><mml:mo>=</mml:mo><mml:mfrac><mml:mrow><mml:msub><mml:mi>u</mml:mi><mml:mn>0</mml:mn></mml:msub></mml:mrow><mml:mrow><mml:mn>4</mml:mn><mml:mi>r</mml:mi></mml:mrow></mml:mfrac><mml:mstyle displaystyle='true'><mml:mrow><mml:msubsup><mml:mo>&#x0222B;</mml:mo><mml:mn>0</mml:mn><mml:mi>r</mml:mi></mml:msubsup><mml:mi>d</mml:mi></mml:mrow></mml:mstyle><mml:mi>r</mml:mi><mml:mrow><mml:mo>[</mml:mo><mml:mrow><mml:mfrac><mml:mrow><mml:mo stretchy='false'>(</mml:mo><mml:mi>d</mml:mi><mml:mo>&#x02212;</mml:mo><mml:mn>1</mml:mn><mml:mo stretchy='false'>)</mml:mo><mml:mo stretchy='false'>(</mml:mo><mml:mi>d</mml:mi><mml:mo>&#x02212;</mml:mo><mml:mn>3</mml:mn><mml:mo stretchy='false'>)</mml:mo></mml:mrow><mml:mrow><mml:msup><mml:mi>r</mml:mi><mml:mn>2</mml:mn></mml:msup></mml:mrow></mml:mfrac><mml:mo>&#x02212;</mml:mo><mml:mfrac><mml:msup><mml:mi>g</mml:mi><mml:mo>&#x02033;</mml:mo></mml:msup><mml:mi>g</mml:mi></mml:mfrac><mml:mo>+</mml:mo><mml:mfrac><mml:mn>3</mml:mn><mml:mn>4</mml:mn></mml:mfrac><mml:msup><mml:mrow><mml:mrow><mml:mo>(</mml:mo><mml:mrow><mml:mfrac><mml:msup><mml:mi>g</mml:mi><mml:mo>&#x02032;</mml:mo></mml:msup><mml:mi>g</mml:mi></mml:mfrac></mml:mrow><mml:mo>)</mml:mo></mml:mrow></mml:mrow><mml:mn>2</mml:mn></mml:msup></mml:mrow><mml:mo>]</mml:mo></mml:mrow><mml:mo>.</mml:mo></mml:mrow></mml:math></disp-formula>
<p>For the hypersphere <italic>S</italic><sup><italic>d</italic></sup>, where <italic>d</italic> &#x02261; <italic>n</italic> &#x02212; 1 and <italic>g</italic> &#x0003D; const. &#x000D7; sin<sup>2(<italic>d</italic> &#x02212; 1)</sup><italic>r</italic>, we have</p>
<disp-formula id="E25"><mml:math id="M25"><mml:mrow><mml:mfrac><mml:msup><mml:mi>g</mml:mi><mml:mo>&#x02032;</mml:mo></mml:msup><mml:mi>g</mml:mi></mml:mfrac><mml:mo>=</mml:mo><mml:mfrac><mml:mrow><mml:mn>2</mml:mn><mml:mo stretchy='false'>(</mml:mo><mml:mi>d</mml:mi><mml:mo>&#x02212;</mml:mo><mml:mn>1</mml:mn><mml:mo stretchy='false'>)</mml:mo></mml:mrow><mml:mrow><mml:mi>tan</mml:mi><mml:mi>r</mml:mi></mml:mrow></mml:mfrac></mml:mrow></mml:math></disp-formula>
<p>and</p>
<disp-formula id="E26"><mml:math id="M26"><mml:mrow><mml:mfrac><mml:msup><mml:mi>g</mml:mi><mml:mo>&#x02033;</mml:mo></mml:msup><mml:mi>g</mml:mi></mml:mfrac><mml:mo>=</mml:mo><mml:mn>2</mml:mn><mml:mo stretchy='false'>(</mml:mo><mml:mi>d</mml:mi><mml:mo>&#x02212;</mml:mo><mml:mn>1</mml:mn><mml:mo stretchy='false'>)</mml:mo><mml:mrow><mml:mo>(</mml:mo><mml:mrow><mml:mfrac><mml:mrow><mml:mn>2</mml:mn><mml:mi>d</mml:mi><mml:mo>&#x02212;</mml:mo><mml:mn>3</mml:mn></mml:mrow><mml:mrow><mml:msup><mml:mrow><mml:mi>tan</mml:mi></mml:mrow><mml:mn>2</mml:mn></mml:msup><mml:mi>r</mml:mi></mml:mrow></mml:mfrac><mml:mo>&#x02212;</mml:mo><mml:mn>1</mml:mn></mml:mrow><mml:mo>)</mml:mo></mml:mrow><mml:mo>.</mml:mo></mml:mrow></mml:math></disp-formula>
<p>Thus,</p>
<disp-formula id="E27"><mml:math id="M27"><mml:mtable columnalign='left'><mml:mtr><mml:mtd><mml:msub><mml:mi>u</mml:mi><mml:mn>1</mml:mn></mml:msub><mml:mo>=</mml:mo><mml:mfrac><mml:mrow><mml:msub><mml:mi>u</mml:mi><mml:mn>0</mml:mn></mml:msub></mml:mrow><mml:mrow><mml:mn>4</mml:mn><mml:mi>r</mml:mi></mml:mrow></mml:mfrac><mml:mstyle displaystyle='true'><mml:mrow><mml:msubsup><mml:mo>&#x0222B;</mml:mo><mml:mn>0</mml:mn><mml:mi>r</mml:mi></mml:msubsup><mml:mi>d</mml:mi></mml:mrow></mml:mstyle><mml:mover accent='true'><mml:mi>r</mml:mi><mml:mo>&#x002DC;</mml:mo></mml:mover><mml:mrow><mml:mo>[</mml:mo><mml:mrow><mml:mfrac><mml:mrow><mml:mo stretchy='false'>(</mml:mo><mml:mi>d</mml:mi><mml:mo>&#x02212;</mml:mo><mml:mn>1</mml:mn><mml:mo stretchy='false'>)</mml:mo><mml:mo stretchy='false'>(</mml:mo><mml:mi>d</mml:mi><mml:mo>&#x02212;</mml:mo><mml:mn>3</mml:mn><mml:mo stretchy='false'>)</mml:mo></mml:mrow><mml:mrow><mml:msup><mml:mover accent='true'><mml:mi>r</mml:mi><mml:mo>&#x002DC;</mml:mo></mml:mover><mml:mn>2</mml:mn></mml:msup></mml:mrow></mml:mfrac><mml:mo>&#x02212;</mml:mo><mml:mo stretchy='false'>(</mml:mo><mml:mi>d</mml:mi><mml:mo>&#x02212;</mml:mo><mml:mn>1</mml:mn><mml:mo stretchy='false'>)</mml:mo><mml:mrow><mml:mo>(</mml:mo><mml:mrow><mml:mfrac><mml:mrow><mml:mi>d</mml:mi><mml:mo>&#x02212;</mml:mo><mml:mn>3</mml:mn></mml:mrow><mml:mrow><mml:msup><mml:mrow><mml:mi>tan</mml:mi></mml:mrow><mml:mn>2</mml:mn></mml:msup><mml:mover accent='true'><mml:mi>r</mml:mi><mml:mo>&#x002DC;</mml:mo></mml:mover></mml:mrow></mml:mfrac><mml:mo>&#x02212;</mml:mo><mml:mn>2</mml:mn></mml:mrow><mml:mo>)</mml:mo></mml:mrow></mml:mrow><mml:mo>]</mml:mo></mml:mrow></mml:mtd></mml:mtr><mml:mtr><mml:mtd><mml:mtext>&#x000A0;&#x000A0;&#x000A0;</mml:mtext><mml:mo>=</mml:mo><mml:mfrac><mml:mrow><mml:msub><mml:mi>u</mml:mi><mml:mn>0</mml:mn></mml:msub><mml:mo stretchy='false'>(</mml:mo><mml:mi>d</mml:mi><mml:mo>&#x02212;</mml:mo><mml:mn>1</mml:mn><mml:mo stretchy='false'>)</mml:mo></mml:mrow><mml:mrow><mml:mn>4</mml:mn><mml:msup><mml:mi>r</mml:mi><mml:mn>2</mml:mn></mml:msup></mml:mrow></mml:mfrac><mml:mrow><mml:mo>[</mml:mo><mml:mrow><mml:mn>3</mml:mn><mml:mo>&#x02212;</mml:mo><mml:mi>d</mml:mi><mml:mo>+</mml:mo><mml:mo stretchy='false'>(</mml:mo><mml:mi>d</mml:mi><mml:mo>&#x02212;</mml:mo><mml:mn>1</mml:mn><mml:mo stretchy='false'>)</mml:mo><mml:msup><mml:mi>r</mml:mi><mml:mn>2</mml:mn></mml:msup><mml:mo>+</mml:mo><mml:mo stretchy='false'>(</mml:mo><mml:mi>d</mml:mi><mml:mo>&#x02212;</mml:mo><mml:mn>3</mml:mn><mml:mo stretchy='false'>)</mml:mo><mml:mi>r</mml:mi><mml:mi>cot</mml:mi><mml:mi>r</mml:mi></mml:mrow><mml:mo>]</mml:mo></mml:mrow><mml:mo>.</mml:mo></mml:mtd></mml:mtr></mml:mtable></mml:math></disp-formula>
<p>Notice that <italic>u</italic><sub>1</sub>(<italic>r</italic>) &#x0003D; 0 when <italic>d</italic> &#x0003D; 1 and <italic>u</italic><sub>1</sub>(<italic>r</italic>) &#x0003D; <italic>u</italic><sub>0</sub>(<italic>r</italic>) when <italic>d</italic> &#x0003D; 3. For <italic>d</italic> &#x0003D; 2, <italic>u</italic><sub>1</sub>/<italic>u</italic><sub>0</sub> is an increasing function in <italic>r</italic> and diverges to &#x0221E; at <italic>r</italic> &#x0003D; &#x003C0;. By contrast, for <italic>d</italic> &#x0003E; 3, <italic>u</italic><sub>1</sub>/<italic>u</italic><sub>0</sub> is a decreasing function in <italic>r</italic> and diverges to &#x02212;&#x0221E; at <italic>r</italic> &#x0003D; &#x003C0;; <italic>u</italic><sub>1</sub>/<italic>u</italic><sub>0</sub> is relatively constant for <italic>r</italic> &#x0003C; &#x003C0; and starts to decrease rapidly only near &#x003C0;. Therefore, the first order correction is not able to remove the unphysical behavior near <italic>r</italic> &#x0003D; 0 in high dimensions where, according to the first order parametrix kernel, the surrounding area is hotter than the heat source.</p>
<p>Next, we apply Equation (4) again to obtain <italic>u</italic><sub>2</sub> as</p>
<disp-formula id="E28"><mml:math id="M28"><mml:mtable columnalign='left'><mml:mtr><mml:mtd><mml:msub><mml:mi>u</mml:mi><mml:mn>2</mml:mn></mml:msub><mml:mo>=</mml:mo><mml:mfrac><mml:mrow><mml:msub><mml:mi>u</mml:mi><mml:mn>0</mml:mn></mml:msub></mml:mrow><mml:mrow><mml:msup><mml:mi>r</mml:mi><mml:mn>2</mml:mn></mml:msup></mml:mrow></mml:mfrac><mml:mstyle displaystyle='true'><mml:mrow><mml:msubsup><mml:mo>&#x0222B;</mml:mo><mml:mn>0</mml:mn><mml:mi>r</mml:mi></mml:msubsup><mml:mi>d</mml:mi></mml:mrow></mml:mstyle><mml:mover accent='true'><mml:mi>r</mml:mi><mml:mo>&#x002DC;</mml:mo></mml:mover><mml:mtext>&#x000A0;</mml:mtext><mml:mover accent='true'><mml:mi>r</mml:mi><mml:mo>&#x002DC;</mml:mo></mml:mover><mml:msubsup><mml:mi>u</mml:mi><mml:mn>0</mml:mn><mml:mrow><mml:mo>&#x02212;</mml:mo><mml:mn>1</mml:mn></mml:mrow></mml:msubsup><mml:mi>&#x00394;</mml:mi><mml:msub><mml:mi>u</mml:mi><mml:mn>1</mml:mn></mml:msub></mml:mtd></mml:mtr><mml:mtr><mml:mtd><mml:mtext>&#x000A0;&#x000A0;&#x000A0;</mml:mtext><mml:mo>=</mml:mo><mml:mfrac><mml:mrow><mml:msub><mml:mi>u</mml:mi><mml:mn>0</mml:mn></mml:msub></mml:mrow><mml:mrow><mml:msup><mml:mi>r</mml:mi><mml:mn>2</mml:mn></mml:msup></mml:mrow></mml:mfrac><mml:mstyle displaystyle='true'><mml:mrow><mml:msubsup><mml:mo>&#x0222B;</mml:mo><mml:mn>0</mml:mn><mml:mi>r</mml:mi></mml:msubsup><mml:mi>d</mml:mi></mml:mrow></mml:mstyle><mml:mover accent='true'><mml:mi>r</mml:mi><mml:mo>&#x002DC;</mml:mo></mml:mover><mml:mtext>&#x000A0;</mml:mtext><mml:mover accent='true'><mml:mi>r</mml:mi><mml:mo>&#x002DC;</mml:mo></mml:mover><mml:msubsup><mml:mi>u</mml:mi><mml:mn>0</mml:mn><mml:mrow><mml:mo>&#x02212;</mml:mo><mml:mn>1</mml:mn></mml:mrow></mml:msubsup><mml:mrow><mml:mo>(</mml:mo><mml:mrow><mml:msubsup><mml:mi>u</mml:mi><mml:mn>0</mml:mn><mml:mrow><mml:mo>&#x02033;</mml:mo></mml:mrow></mml:msubsup><mml:mo>+</mml:mo><mml:msubsup><mml:mi>u</mml:mi><mml:mn>1</mml:mn><mml:mo>&#x02032;</mml:mo></mml:msubsup><mml:mo stretchy='false'>(</mml:mo><mml:mi>log</mml:mi><mml:msqrt><mml:mi>g</mml:mi></mml:msqrt><mml:msup><mml:mo stretchy='false'>)</mml:mo><mml:mo>&#x02032;</mml:mo></mml:msup></mml:mrow><mml:mo>)</mml:mo></mml:mrow><mml:mo>.</mml:mo></mml:mtd></mml:mtr></mml:mtable></mml:math></disp-formula>
<p>After some cumbersome algebraic manipulations, we find</p>
<disp-formula id="E29"><mml:math id="M29"><mml:mtable columnalign='left'><mml:mtr><mml:mtd><mml:mfrac><mml:mrow><mml:msub><mml:mi>u</mml:mi><mml:mn>2</mml:mn></mml:msub></mml:mrow><mml:mrow><mml:msub><mml:mi>u</mml:mi><mml:mn>0</mml:mn></mml:msub></mml:mrow></mml:mfrac><mml:mo>=</mml:mo><mml:mfrac><mml:mrow><mml:mi>d</mml:mi><mml:mo>&#x02212;</mml:mo><mml:mn>1</mml:mn></mml:mrow><mml:mrow><mml:mn>32</mml:mn></mml:mrow></mml:mfrac><mml:mrow><mml:mo>[</mml:mo><mml:mrow><mml:msup><mml:mrow><mml:mo stretchy='false'>(</mml:mo><mml:mi>d</mml:mi><mml:mo>&#x02212;</mml:mo><mml:mn>3</mml:mn><mml:mo stretchy='false'>)</mml:mo></mml:mrow><mml:mn>3</mml:mn></mml:msup><mml:mo>+</mml:mo><mml:mfrac><mml:mrow><mml:mo stretchy='false'>(</mml:mo><mml:mi>d</mml:mi><mml:mo>&#x02212;</mml:mo><mml:mn>3</mml:mn><mml:mo stretchy='false'>)</mml:mo><mml:mo stretchy='false'>(</mml:mo><mml:mi>d</mml:mi><mml:mo>&#x02212;</mml:mo><mml:mn>5</mml:mn><mml:mo stretchy='false'>)</mml:mo><mml:mo stretchy='false'>(</mml:mo><mml:mi>d</mml:mi><mml:mo>&#x02212;</mml:mo><mml:mn>7</mml:mn><mml:mo stretchy='false'>)</mml:mo></mml:mrow><mml:mrow><mml:msup><mml:mi>r</mml:mi><mml:mn>4</mml:mn></mml:msup></mml:mrow></mml:mfrac><mml:mo>&#x02212;</mml:mo><mml:mfrac><mml:mrow><mml:msup><mml:mrow><mml:mo stretchy='false'>(</mml:mo><mml:mi>d</mml:mi><mml:mo>&#x02212;</mml:mo><mml:mn>3</mml:mn><mml:mo stretchy='false'>)</mml:mo></mml:mrow><mml:mn>2</mml:mn></mml:msup><mml:mo stretchy='false'>(</mml:mo><mml:mi>d</mml:mi><mml:mo>&#x02212;</mml:mo><mml:mn>5</mml:mn><mml:mo stretchy='false'>)</mml:mo></mml:mrow><mml:mrow><mml:msup><mml:mi>r</mml:mi><mml:mn>3</mml:mn></mml:msup><mml:mi>tan</mml:mi><mml:mi>r</mml:mi></mml:mrow></mml:mfrac></mml:mrow></mml:mrow></mml:mtd></mml:mtr><mml:mtr><mml:mtd><mml:mtext>&#x000A0;&#x000A0;&#x000A0;&#x000A0;&#x000A0;&#x000A0;&#x000A0;</mml:mtext><mml:mrow><mml:mrow><mml:mo>+</mml:mo><mml:mtext>&#x000A0;</mml:mtext><mml:mfrac><mml:mrow><mml:mn>2</mml:mn><mml:msup><mml:mrow><mml:mo stretchy='false'>(</mml:mo><mml:mi>d</mml:mi><mml:mo>&#x02212;</mml:mo><mml:mn>1</mml:mn><mml:mo stretchy='false'>)</mml:mo></mml:mrow><mml:mn>2</mml:mn></mml:msup><mml:mo stretchy='false'>(</mml:mo><mml:mi>d</mml:mi><mml:mo>&#x02212;</mml:mo><mml:mn>3</mml:mn><mml:mo stretchy='false'>)</mml:mo></mml:mrow><mml:mrow><mml:mi>r</mml:mi><mml:mi>tan</mml:mi><mml:mi>r</mml:mi></mml:mrow></mml:mfrac><mml:mo>+</mml:mo><mml:mfrac><mml:mrow><mml:mo stretchy='false'>(</mml:mo><mml:mi>d</mml:mi><mml:mo>+</mml:mo><mml:mn>1</mml:mn><mml:mo stretchy='false'>)</mml:mo><mml:mo stretchy='false'>(</mml:mo><mml:mi>d</mml:mi><mml:mo>&#x02212;</mml:mo><mml:mn>3</mml:mn><mml:mo stretchy='false'>)</mml:mo><mml:mo stretchy='false'>(</mml:mo><mml:mi>d</mml:mi><mml:mo>&#x02212;</mml:mo><mml:mn>5</mml:mn><mml:mo stretchy='false'>)</mml:mo></mml:mrow><mml:mrow><mml:msup><mml:mi>r</mml:mi><mml:mn>2</mml:mn></mml:msup><mml:mi>sin</mml:mi><mml:mi>r</mml:mi></mml:mrow></mml:mfrac></mml:mrow><mml:mo>]</mml:mo></mml:mrow><mml:mo>.</mml:mo></mml:mtd></mml:mtr></mml:mtable></mml:math></disp-formula>
<p>Again, <italic>d</italic> &#x0003D; 1 and <italic>d</italic> &#x0003D; 3 are special dimensions, where <italic>u</italic><sub>2</sub>(<italic>r</italic>) &#x0003D; 0 for <italic>d</italic> &#x0003D; 1, and <italic>u</italic><sub>2</sub>(<italic>r</italic>) &#x0003D; <italic>u</italic><sub>0</sub>/2 for <italic>d</italic> &#x0003D; 3; for other dimensions, <italic>u</italic><sub>2</sub>(<italic>r</italic>) is singular at both <italic>r</italic> &#x0003D; 0 and &#x003C0;. Note that on <italic>S</italic><sup>1</sup>, the metric in geodesic polar coordinate is <italic>g</italic><sub>11</sub> &#x0003D; 1, so all parametrix expansion coefficients <italic>u</italic><sub><italic>k</italic></sub>(<italic>r</italic>) must vanish identically, as we have explicitly shown above.</p>
<p>Thus, the full <italic>G</italic><sup>prx</sup> defined on a hypersphere, where the geodesic distance <italic>r</italic> is just the arc length &#x003B8;, suffers from numerous problems. The zeroth order correction term <inline-formula><mml:math id="M44"><mml:msub><mml:mrow><mml:mi>u</mml:mi></mml:mrow><mml:mrow><mml:mn>0</mml:mn></mml:mrow></mml:msub><mml:mo>=</mml:mo><mml:msup><mml:mrow><mml:mrow><mml:mo stretchy="false">(</mml:mo><mml:mrow><mml:mo class="qopname">sin</mml:mo><mml:mi>&#x003B8;</mml:mi><mml:mo>/</mml:mo><mml:mi>&#x003B8;</mml:mi></mml:mrow><mml:mo stretchy="false">)</mml:mo></mml:mrow></mml:mrow><mml:mrow><mml:mo>-</mml:mo><mml:mfrac><mml:mrow><mml:mi>n</mml:mi><mml:mo>-</mml:mo><mml:mn>2</mml:mn></mml:mrow><mml:mrow><mml:mn>2</mml:mn></mml:mrow></mml:mfrac></mml:mrow></mml:msup></mml:math></inline-formula> diverges at &#x003B8; &#x0003D; &#x003C0;; this behavior is not a major problem if &#x003B8; is restricted to the range <inline-formula><mml:math id="M45"><mml:mrow><mml:mo>[</mml:mo><mml:mrow><mml:mn>0</mml:mn><mml:mo>,</mml:mo><mml:mfrac><mml:mrow><mml:mi>&#x003C0;</mml:mi></mml:mrow><mml:mrow><mml:mn>2</mml:mn></mml:mrow></mml:mfrac></mml:mrow><mml:mo>]</mml:mo></mml:mrow></mml:math></inline-formula>. Moreover, <italic>G</italic><sup>prx</sup> is also unphysical as &#x003B8; &#x02193; 0 when (<italic>n</italic> &#x02212; 2)<italic>t</italic> &#x0003E; 3; this condition on dimension and time is obtained by expanding <inline-formula><mml:math id="M46"><mml:msup><mml:mrow><mml:mtext>e</mml:mtext></mml:mrow><mml:mrow><mml:mo>-</mml:mo><mml:msup><mml:mrow><mml:mi>&#x003B8;</mml:mi></mml:mrow><mml:mrow><mml:mn>2</mml:mn></mml:mrow></mml:msup><mml:mo>/</mml:mo><mml:mn>4</mml:mn><mml:mi>t</mml:mi></mml:mrow></mml:msup><mml:mo>=</mml:mo><mml:mn>1</mml:mn><mml:mo>-</mml:mo><mml:mfrac><mml:mrow><mml:msup><mml:mrow><mml:mi>&#x003B8;</mml:mi></mml:mrow><mml:mrow><mml:mn>2</mml:mn></mml:mrow></mml:msup></mml:mrow><mml:mrow><mml:mn>4</mml:mn><mml:mi>t</mml:mi></mml:mrow></mml:mfrac><mml:mo>&#x0002B;</mml:mo><mml:mrow><mml:mi mathvariant="-tex-caligraphic">O</mml:mi></mml:mrow><mml:mrow><mml:mo stretchy="false">(</mml:mo><mml:mrow><mml:msup><mml:mrow><mml:mi>&#x003B8;</mml:mi></mml:mrow><mml:mrow><mml:mn>4</mml:mn></mml:mrow></mml:msup></mml:mrow><mml:mo stretchy="false">)</mml:mo></mml:mrow></mml:math></inline-formula> and <inline-formula><mml:math id="M47"><mml:msup><mml:mrow><mml:mrow><mml:mo stretchy="false">(</mml:mo><mml:mrow><mml:mo class="qopname">sin</mml:mo><mml:mi>&#x003B8;</mml:mi><mml:mo>/</mml:mo><mml:mi>&#x003B8;</mml:mi></mml:mrow><mml:mo stretchy="false">)</mml:mo></mml:mrow></mml:mrow><mml:mrow><mml:mo>-</mml:mo><mml:mfrac><mml:mrow><mml:mi>n</mml:mi><mml:mo>-</mml:mo><mml:mn>2</mml:mn></mml:mrow><mml:mrow><mml:mn>2</mml:mn></mml:mrow></mml:mfrac></mml:mrow></mml:msup><mml:mo>=</mml:mo><mml:mn>1</mml:mn><mml:mo>&#x0002B;</mml:mo><mml:mfrac><mml:mrow><mml:msup><mml:mrow><mml:mi>&#x003B8;</mml:mi></mml:mrow><mml:mrow><mml:mn>2</mml:mn></mml:mrow></mml:msup></mml:mrow><mml:mrow><mml:mn>12</mml:mn></mml:mrow></mml:mfrac><mml:mrow><mml:mo stretchy="false">(</mml:mo><mml:mrow><mml:mi>n</mml:mi><mml:mo>-</mml:mo><mml:mn>2</mml:mn></mml:mrow><mml:mo stretchy="false">)</mml:mo></mml:mrow><mml:mo>&#x0002B;</mml:mo><mml:mrow><mml:mi mathvariant="-tex-caligraphic">O</mml:mi></mml:mrow><mml:mrow><mml:mo stretchy="false">(</mml:mo><mml:mrow><mml:msup><mml:mrow><mml:mi>&#x003B8;</mml:mi></mml:mrow><mml:mrow><mml:mn>3</mml:mn></mml:mrow></mml:msup></mml:mrow><mml:mo stretchy="false">)</mml:mo></mml:mrow></mml:math></inline-formula>, and noting that the leading order &#x003B8;<sup>2</sup> term in the product of the two factors is a non-decreasing function of distance &#x003B8; when <inline-formula><mml:math id="M48"><mml:mfrac><mml:mrow><mml:mi>n</mml:mi><mml:mo>-</mml:mo><mml:mn>2</mml:mn></mml:mrow><mml:mrow><mml:mn>12</mml:mn></mml:mrow></mml:mfrac><mml:mo>&#x02265;</mml:mo><mml:mfrac><mml:mrow><mml:mn>1</mml:mn></mml:mrow><mml:mrow><mml:mn>4</mml:mn><mml:mi>t</mml:mi></mml:mrow></mml:mfrac></mml:math></inline-formula>, corresponding to the unphysical situation of nearby points being hotter than the heat source itself. As the feature dimension <italic>n</italic> is typically very large, the restriction (<italic>n</italic> &#x02212; 2)<italic>t</italic> &#x0003C; 3 implies that we need to take the diffusion time to be very small, thus making the similarity measure captured by <italic>G</italic><sup>prx</sup> decay too fast away from each data point for use in clustering applications. In this work, we further computed the first and second order correction terms, denoted <italic>u</italic><sub>1</sub> and <italic>u</italic><sub>2</sub> in Equations (5), (6), respectively. In high dimensions, the divergence of <italic>u</italic><sub>1</sub>/<italic>u</italic><sub>0</sub> and <italic>u</italic><sub>2</sub>/<italic>u</italic><sub>0</sub> at &#x003B8; &#x0003D; &#x003C0; is not a major problem, as we expect the expansion to be valid only in the vicinity &#x003B8; &#x02193; 0; however, the divergence of <italic>u</italic><sub>2</sub>/<italic>u</italic><sub>0</sub> at &#x003B8; &#x0003D; 0 (to &#x02212;&#x0221E; in high dimensions) is pathological, and thus, we truncate our approximation to <inline-formula><mml:math id="M49"><mml:mrow><mml:mi mathvariant="-tex-caligraphic">O</mml:mi></mml:mrow><mml:mrow><mml:mo stretchy="false">(</mml:mo><mml:mrow><mml:msup><mml:mrow><mml:mi>t</mml:mi></mml:mrow><mml:mrow><mml:mn>2</mml:mn></mml:mrow></mml:msup></mml:mrow><mml:mo stretchy="false">)</mml:mo></mml:mrow></mml:math></inline-formula>. Since <italic>u</italic><sub>1</sub>(&#x003B8;) is not able to correct the unphysical behavior of the parametrix kernel near &#x003B8; &#x0003D; 0 in high dimensions, we conclude that the parametrix approximation fails in high dimensions. Hence, the only remaining part of <italic>G</italic><sup>prx</sup> still applicable to SVM classification is the Gaussian factor, which is clearly not a heat kernel on the hypersphere. The failure of this perturbative expansion using the Euclidean heat kernel as a starting point suggests that diffusion in &#x0211D;<sup><italic>d</italic></sup> and <italic>S</italic><sup><italic>d</italic></sup> are fundamentally different and that the exact hyperspherical heat kernel derived from a non-perturbative approach will likely yield better insights into the diffusion process.</p>
</sec>
<sec>
<title>2.2. Exact hyperspherical heat kernel</title>
<p>By definition, the exact heat kernel <inline-formula><mml:math id="M50"><mml:msup><mml:mrow><mml:mi>G</mml:mi></mml:mrow><mml:mrow><mml:mtext>ext</mml:mtext></mml:mrow></mml:msup><mml:mrow><mml:mo stretchy="false">(</mml:mo><mml:mrow><mml:mover accent="true"><mml:mrow><mml:mi>x</mml:mi></mml:mrow><mml:mo>^</mml:mo></mml:mover><mml:mo>,</mml:mo><mml:mover accent="true"><mml:mrow><mml:mi>y</mml:mi></mml:mrow><mml:mo class="qopname">^</mml:mo></mml:mover><mml:mo>;</mml:mo><mml:mi>t</mml:mi></mml:mrow><mml:mo stretchy="false">)</mml:mo></mml:mrow></mml:math></inline-formula> is the fundamental solution to heat equation <inline-formula><mml:math id="M51"><mml:msub><mml:mrow><mml:mi>&#x02202;</mml:mi></mml:mrow><mml:mrow><mml:mi>t</mml:mi></mml:mrow></mml:msub><mml:mi>u</mml:mi><mml:mo>&#x0002B;</mml:mo><mml:msup><mml:mrow><mml:mover accent="true"><mml:mrow><mml:mi>L</mml:mi></mml:mrow><mml:mo>^</mml:mo></mml:mover></mml:mrow><mml:mrow><mml:mn>2</mml:mn></mml:mrow></mml:msup><mml:mi>u</mml:mi><mml:mo>=</mml:mo><mml:mn>0</mml:mn></mml:math></inline-formula> where <inline-formula><mml:math id="M52"><mml:mo>-</mml:mo><mml:msup><mml:mrow><mml:mover accent="true"><mml:mrow><mml:mi>L</mml:mi></mml:mrow><mml:mo>^</mml:mo></mml:mover></mml:mrow><mml:mrow><mml:mn>2</mml:mn></mml:mrow></mml:msup></mml:math></inline-formula> is the hyperspherical Laplacian [<xref ref-type="bibr" rid="B13">13</xref>, <xref ref-type="bibr" rid="B14">14</xref>, <xref ref-type="bibr" rid="B19">19</xref>, <xref ref-type="bibr" rid="B20">20</xref>]. In the language of operator theory, <inline-formula><mml:math id="M53"><mml:msup><mml:mrow><mml:mi>G</mml:mi></mml:mrow><mml:mrow><mml:mtext>ext</mml:mtext></mml:mrow></mml:msup><mml:mrow><mml:mo stretchy="false">(</mml:mo><mml:mrow><mml:mover accent="true"><mml:mrow><mml:mi>x</mml:mi></mml:mrow><mml:mo>^</mml:mo></mml:mover><mml:mo>,</mml:mo><mml:mover accent="true"><mml:mrow><mml:mi>y</mml:mi></mml:mrow><mml:mo class="qopname">^</mml:mo></mml:mover><mml:mo>;</mml:mo><mml:mi>t</mml:mi></mml:mrow><mml:mo stretchy="false">)</mml:mo></mml:mrow></mml:math></inline-formula> is an integral kernel, or Green&#x00027;s function, for the operator <inline-formula><mml:math id="M54"><mml:mo class="qopname">exp</mml:mo><mml:mrow><mml:mo>{</mml:mo><mml:mrow><mml:mo>-</mml:mo><mml:msup><mml:mrow><mml:mover accent="true"><mml:mrow><mml:mi>L</mml:mi></mml:mrow><mml:mo class="qopname">^</mml:mo></mml:mover></mml:mrow><mml:mrow><mml:mn>2</mml:mn></mml:mrow></mml:msup><mml:mi>t</mml:mi></mml:mrow><mml:mo>}</mml:mo></mml:mrow></mml:math></inline-formula> and has an associated eigenfunction expansion. Because <inline-formula><mml:math id="M55"><mml:msup><mml:mrow><mml:mover accent="true"><mml:mrow><mml:mi>L</mml:mi></mml:mrow><mml:mo>^</mml:mo></mml:mover></mml:mrow><mml:mrow><mml:mn>2</mml:mn></mml:mrow></mml:msup></mml:math></inline-formula> and <inline-formula><mml:math id="M56"><mml:mo class="qopname">exp</mml:mo><mml:mrow><mml:mo>{</mml:mo><mml:mrow><mml:mo>-</mml:mo><mml:msup><mml:mrow><mml:mover accent="true"><mml:mrow><mml:mi>L</mml:mi></mml:mrow><mml:mo class="qopname">^</mml:mo></mml:mover></mml:mrow><mml:mrow><mml:mn>2</mml:mn></mml:mrow></mml:msup><mml:mi>t</mml:mi></mml:mrow><mml:mo>}</mml:mo></mml:mrow></mml:math></inline-formula> share the same eigenfunctions, obtaining the eigenfunction expansion of <inline-formula><mml:math id="M57"><mml:msup><mml:mrow><mml:mi>G</mml:mi></mml:mrow><mml:mrow><mml:mtext>ext</mml:mtext></mml:mrow></mml:msup><mml:mrow><mml:mo stretchy="false">(</mml:mo><mml:mrow><mml:mover accent="true"><mml:mrow><mml:mi>x</mml:mi></mml:mrow><mml:mo>^</mml:mo></mml:mover><mml:mo>,</mml:mo><mml:mover accent="true"><mml:mrow><mml:mi>y</mml:mi></mml:mrow><mml:mo class="qopname">^</mml:mo></mml:mover><mml:mo>;</mml:mo><mml:mi>t</mml:mi></mml:mrow><mml:mo stretchy="false">)</mml:mo></mml:mrow></mml:math></inline-formula> amounts to solving for the complete basis of eigenfunctions of <inline-formula><mml:math id="M58"><mml:msup><mml:mrow><mml:mover accent="true"><mml:mrow><mml:mi>L</mml:mi></mml:mrow><mml:mo>^</mml:mo></mml:mover></mml:mrow><mml:mrow><mml:mn>2</mml:mn></mml:mrow></mml:msup></mml:math></inline-formula>. The spectral decomposition of the Laplacian is in turn facilitated by embedding <italic>S</italic><sup><italic>n</italic> &#x02212; 1</sup> in &#x0211D;<sup><italic>n</italic></sup> and utilizing the global rotational symmetry of <italic>S</italic><sup><italic>n</italic>&#x02212;1</sup> in &#x0211D;<sup><italic>n</italic></sup>. The Euclidean space harmonic functions, which are the solutions to the Laplace equation &#x02207;<sup>2</sup><italic>u</italic> &#x0003D; 0 in &#x0211D;<sup><italic>n</italic></sup>, can be projected to the unit hypersphere <italic>S</italic><sup><italic>n</italic>&#x02212;1</sup> through the usual separation of radial and angular variables [<xref ref-type="bibr" rid="B17">17</xref>, <xref ref-type="bibr" rid="B18">18</xref>]. In this formalism, the hyperspherical Laplacian <inline-formula><mml:math id="M59"><mml:mo>-</mml:mo><mml:msup><mml:mrow><mml:mover accent="true"><mml:mrow><mml:mi>L</mml:mi></mml:mrow><mml:mo>^</mml:mo></mml:mover></mml:mrow><mml:mrow><mml:mn>2</mml:mn></mml:mrow></mml:msup></mml:math></inline-formula> on <italic>S</italic><sup><italic>n</italic>&#x02212;1</sup> naturally arises as the angular part of the Euclidean Laplacian on &#x0211D;<sup><italic>n</italic></sup>, and <inline-formula><mml:math id="M60"><mml:msup><mml:mrow><mml:mover accent="true"><mml:mrow><mml:mi>L</mml:mi></mml:mrow><mml:mo>^</mml:mo></mml:mover></mml:mrow><mml:mrow><mml:mn>2</mml:mn></mml:mrow></mml:msup></mml:math></inline-formula> can be interpreted as the squared angular momentum operator in &#x0211D;<sup><italic>n</italic></sup> [<xref ref-type="bibr" rid="B18">18</xref>].</p>
<p>The resulting eigenfunctions of <inline-formula><mml:math id="M61"><mml:msup><mml:mrow><mml:mover accent="true"><mml:mrow><mml:mi>L</mml:mi></mml:mrow><mml:mo>^</mml:mo></mml:mover></mml:mrow><mml:mrow><mml:mn>2</mml:mn></mml:mrow></mml:msup></mml:math></inline-formula> are known as the hyperspherical harmonics and generalize the usual spherical harmonics in &#x0211D;<sup>3</sup> to higher dimensions. Each hyperspherical harmonic is equipped with a triplet of parameters or &#x0201C;quantum numbers&#x0201D; (&#x02113;, {<italic>m</italic><sub><italic>i</italic></sub>}, &#x003B1;): the degree &#x02113;, magnetic quantum numbers {<italic>m</italic><sub><italic>i</italic></sub>} and <inline-formula><mml:math id="M62"><mml:mi>&#x003B1;</mml:mi><mml:mo>=</mml:mo><mml:mfrac><mml:mrow><mml:mi>n</mml:mi></mml:mrow><mml:mrow><mml:mn>2</mml:mn></mml:mrow></mml:mfrac><mml:mo>-</mml:mo><mml:mn>1</mml:mn></mml:math></inline-formula>. In the eigenfunction expansion of <inline-formula><mml:math id="M63"><mml:mo class="qopname">exp</mml:mo><mml:mrow><mml:mo>{</mml:mo><mml:mrow><mml:mo>-</mml:mo><mml:msup><mml:mrow><mml:mover accent="true"><mml:mrow><mml:mi>L</mml:mi></mml:mrow><mml:mo class="qopname">^</mml:mo></mml:mover></mml:mrow><mml:mrow><mml:mn>2</mml:mn></mml:mrow></mml:msup><mml:mi>t</mml:mi></mml:mrow><mml:mo>}</mml:mo></mml:mrow></mml:math></inline-formula>, we use the addition theorem of hyperspherical harmonics to sum over the magnetic quantum number {<italic>m</italic><sub><italic>i</italic></sub>} and obtain the following result:</p>
<p><bold>Theorem 1</bold>. <italic>The exact hyperspherical heat kernel <inline-formula><mml:math id="M64"><mml:msup><mml:mrow><mml:mi>G</mml:mi></mml:mrow><mml:mrow><mml:mtext>ext</mml:mtext></mml:mrow></mml:msup><mml:mrow><mml:mo stretchy="false">(</mml:mo><mml:mrow><mml:mover accent="true"><mml:mrow><mml:mi>x</mml:mi></mml:mrow><mml:mo>^</mml:mo></mml:mover><mml:mo>,</mml:mo><mml:mover accent="true"><mml:mrow><mml:mi>y</mml:mi></mml:mrow><mml:mo class="qopname">^</mml:mo></mml:mover><mml:mo>;</mml:mo><mml:mi>t</mml:mi></mml:mrow><mml:mo stretchy="false">)</mml:mo></mml:mrow></mml:math></inline-formula> can be expanded as a uniformly and absolutely convergent power series</italic></p>
<disp-formula id="E30"><mml:math id="M30"><mml:mrow><mml:msup><mml:mi>G</mml:mi><mml:mrow><mml:mtext>ext</mml:mtext></mml:mrow></mml:msup><mml:mo stretchy='false'>(</mml:mo><mml:mover accent='true'><mml:mi>x</mml:mi><mml:mo>&#x0005E;</mml:mo></mml:mover><mml:mo>,</mml:mo><mml:mover accent='true'><mml:mi>y</mml:mi><mml:mo>&#x0005E;</mml:mo></mml:mover><mml:mo>;</mml:mo><mml:mi>t</mml:mi><mml:mo stretchy='false'>)</mml:mo><mml:mo>=</mml:mo><mml:mstyle displaystyle='true'><mml:munderover><mml:mo>&#x02211;</mml:mo><mml:mrow><mml:mi>&#x02113;</mml:mi><mml:mo>=</mml:mo><mml:mn>0</mml:mn></mml:mrow><mml:mi>&#x0221E;</mml:mi></mml:munderover><mml:mrow><mml:msup><mml:mtext>e</mml:mtext><mml:mrow><mml:mo>&#x02212;</mml:mo><mml:mi>&#x02113;</mml:mi><mml:mo stretchy='false'>(</mml:mo><mml:mi>&#x02113;</mml:mi><mml:mo>+</mml:mo><mml:mi>n</mml:mi><mml:mo>&#x02212;</mml:mo><mml:mn>2</mml:mn><mml:mo stretchy='false'>)</mml:mo><mml:mi>t</mml:mi></mml:mrow></mml:msup></mml:mrow></mml:mstyle><mml:mfrac><mml:mrow><mml:mn>2</mml:mn><mml:mi>&#x02113;</mml:mi><mml:mo>+</mml:mo><mml:mi>n</mml:mi><mml:mo>&#x02212;</mml:mo><mml:mn>2</mml:mn></mml:mrow><mml:mrow><mml:mi>n</mml:mi><mml:mo>&#x02212;</mml:mo><mml:mn>2</mml:mn></mml:mrow></mml:mfrac><mml:mfrac><mml:mn>1</mml:mn><mml:mrow><mml:msub><mml:mi>A</mml:mi><mml:mrow><mml:msup><mml:mi>S</mml:mi><mml:mrow><mml:mi>n</mml:mi><mml:mo>&#x02212;</mml:mo><mml:mn>1</mml:mn></mml:mrow></mml:msup></mml:mrow></mml:msub></mml:mrow></mml:mfrac><mml:msubsup><mml:mi>C</mml:mi><mml:mi>&#x02113;</mml:mi><mml:mrow><mml:mfrac><mml:mi>n</mml:mi><mml:mn>2</mml:mn></mml:mfrac><mml:mo>&#x02212;</mml:mo><mml:mn>1</mml:mn></mml:mrow></mml:msubsup><mml:mo stretchy='false'>(</mml:mo><mml:mover accent='true'><mml:mi>x</mml:mi><mml:mo>&#x0005E;</mml:mo></mml:mover><mml:mo>&#x000B7;</mml:mo><mml:mover accent='true'><mml:mi>y</mml:mi><mml:mo>&#x0005E;</mml:mo></mml:mover><mml:mo stretchy='false'>)</mml:mo></mml:mrow></mml:math></disp-formula>
<p><italic>in the interval <inline-formula><mml:math id="M65"><mml:mover accent="true"><mml:mrow><mml:mi>x</mml:mi></mml:mrow><mml:mo>^</mml:mo></mml:mover><mml:mo>&#x000B7;</mml:mo><mml:mover accent="true"><mml:mrow><mml:mi>y</mml:mi></mml:mrow><mml:mo class="qopname">^</mml:mo></mml:mover><mml:mo>&#x02208;</mml:mo><mml:mrow><mml:mo>[</mml:mo><mml:mrow><mml:mo>-</mml:mo><mml:mn>1</mml:mn><mml:mo>,</mml:mo><mml:mn>1</mml:mn></mml:mrow><mml:mo>]</mml:mo></mml:mrow></mml:math></inline-formula> and for <italic>t</italic> &#x0003E; 0, where <inline-formula><mml:math id="M66"><mml:msubsup><mml:mrow><mml:mi>C</mml:mi></mml:mrow><mml:mrow><mml:mi>&#x02113;</mml:mi></mml:mrow><mml:mrow><mml:mi>&#x003B1;</mml:mi></mml:mrow></mml:msubsup><mml:mrow><mml:mo stretchy="false">(</mml:mo><mml:mrow><mml:mi>w</mml:mi></mml:mrow><mml:mo stretchy="false">)</mml:mo></mml:mrow></mml:math></inline-formula> are the Gegenbauer polynomials and <inline-formula><mml:math id="M67"><mml:msub><mml:mrow><mml:mi>A</mml:mi></mml:mrow><mml:mrow><mml:msup><mml:mrow><mml:mi>S</mml:mi></mml:mrow><mml:mrow><mml:mi>n</mml:mi><mml:mo>-</mml:mo><mml:mn>1</mml:mn></mml:mrow></mml:msup></mml:mrow></mml:msub><mml:mo>=</mml:mo><mml:mfrac><mml:mrow><mml:mn>2</mml:mn><mml:msup><mml:mrow><mml:mi>&#x003C0;</mml:mi></mml:mrow><mml:mrow><mml:mfrac><mml:mrow><mml:mi>n</mml:mi></mml:mrow><mml:mrow><mml:mn>2</mml:mn></mml:mrow></mml:mfrac></mml:mrow></mml:msup></mml:mrow><mml:mrow><mml:mi>&#x00393;</mml:mi><mml:mrow><mml:mo stretchy="false">(</mml:mo><mml:mrow><mml:mfrac><mml:mrow><mml:mi>n</mml:mi></mml:mrow><mml:mrow><mml:mn>2</mml:mn></mml:mrow></mml:mfrac></mml:mrow><mml:mo stretchy="false">)</mml:mo></mml:mrow></mml:mrow></mml:mfrac></mml:math></inline-formula> is the surface area of <italic>S</italic><sup><italic>n</italic>&#x02212;1</sup>. Since the kernel depends on <inline-formula><mml:math id="M68"><mml:mover accent="true"><mml:mrow><mml:mi>x</mml:mi></mml:mrow><mml:mo>^</mml:mo></mml:mover></mml:math></inline-formula> and &#x00177; only through <inline-formula><mml:math id="M69"><mml:mover accent="true"><mml:mrow><mml:mi>x</mml:mi></mml:mrow><mml:mo>^</mml:mo></mml:mover><mml:mo>&#x000B7;</mml:mo><mml:mover accent="true"><mml:mrow><mml:mi>y</mml:mi></mml:mrow><mml:mo class="qopname">^</mml:mo></mml:mover></mml:math></inline-formula>, we will write <inline-formula><mml:math id="M70"><mml:msup><mml:mrow><mml:mi>G</mml:mi></mml:mrow><mml:mrow><mml:mtext>ext</mml:mtext></mml:mrow></mml:msup><mml:mrow><mml:mo stretchy="false">(</mml:mo><mml:mrow><mml:mover accent="true"><mml:mrow><mml:mi>x</mml:mi></mml:mrow><mml:mo>^</mml:mo></mml:mover><mml:mo>,</mml:mo><mml:mover accent="true"><mml:mrow><mml:mi>y</mml:mi></mml:mrow><mml:mo class="qopname">^</mml:mo></mml:mover><mml:mo>;</mml:mo><mml:mi>t</mml:mi></mml:mrow><mml:mo stretchy="false">)</mml:mo></mml:mrow><mml:mo>=</mml:mo><mml:msup><mml:mrow><mml:mi>G</mml:mi></mml:mrow><mml:mrow><mml:mtext>ext</mml:mtext></mml:mrow></mml:msup><mml:mrow><mml:mo stretchy="false">(</mml:mo><mml:mrow><mml:mover accent="true"><mml:mrow><mml:mi>x</mml:mi></mml:mrow><mml:mo>^</mml:mo></mml:mover><mml:mo>&#x000B7;</mml:mo><mml:mover accent="true"><mml:mrow><mml:mi>y</mml:mi></mml:mrow><mml:mo class="qopname">^</mml:mo></mml:mover><mml:mo>;</mml:mo><mml:mi>t</mml:mi></mml:mrow><mml:mo stretchy="false">)</mml:mo></mml:mrow></mml:math></inline-formula></italic>.</p>
<p>A proof of similar expansions can be found in standard textbooks on orthogonal polynomials and geometric analysis, e.g., [<xref ref-type="bibr" rid="B17">17</xref>]. For the sake of completeness, we include a proof in Supplementary Material section 2.5.3.</p>
<p>Note that the exact kernel <italic>G</italic><sup>ext</sup> is a Mercer kernel re-expressed by summing over the degenerate eigenstates indexed by {<italic>m</italic>}. As before, we will rescale the kernel by self-similarity and define:</p>
<p>Definition 4. <italic>The exact kernel <inline-formula><mml:math id="M71"><mml:msup><mml:mrow><mml:mi>K</mml:mi></mml:mrow><mml:mrow><mml:mtext>ext</mml:mtext></mml:mrow></mml:msup><mml:mrow><mml:mo stretchy="false">(</mml:mo><mml:mrow><mml:mover accent="true"><mml:mrow><mml:mi>x</mml:mi></mml:mrow><mml:mo>^</mml:mo></mml:mover><mml:mo>,</mml:mo><mml:mover accent="true"><mml:mrow><mml:mi>y</mml:mi></mml:mrow><mml:mo class="qopname">^</mml:mo></mml:mover><mml:mo>;</mml:mo><mml:mi>t</mml:mi></mml:mrow><mml:mo stretchy="false">)</mml:mo></mml:mrow></mml:math></inline-formula> is the exact heat kernel normalized by self-similarity:</italic></p>
<disp-formula id="E31"><mml:math id="M31"><mml:mrow><mml:msup><mml:mi>K</mml:mi><mml:mrow><mml:mtext>ext</mml:mtext></mml:mrow></mml:msup><mml:mo stretchy='false'>(</mml:mo><mml:mover accent='true'><mml:mi>x</mml:mi><mml:mo>&#x0005E;</mml:mo></mml:mover><mml:mo>,</mml:mo><mml:mover accent='true'><mml:mi>y</mml:mi><mml:mo>&#x0005E;</mml:mo></mml:mover><mml:mo>;</mml:mo><mml:mi>t</mml:mi><mml:mo stretchy='false'>)</mml:mo><mml:mo>=</mml:mo><mml:mfrac><mml:mrow><mml:msup><mml:mi>G</mml:mi><mml:mrow><mml:mtext>ext</mml:mtext></mml:mrow></mml:msup><mml:mo stretchy='false'>(</mml:mo><mml:mover accent='true'><mml:mi>x</mml:mi><mml:mo>&#x0005E;</mml:mo></mml:mover><mml:mo>&#x000B7;</mml:mo><mml:mover accent='true'><mml:mi>y</mml:mi><mml:mo>&#x0005E;</mml:mo></mml:mover><mml:mo>;</mml:mo><mml:mi>t</mml:mi><mml:mo stretchy='false'>)</mml:mo></mml:mrow><mml:mrow><mml:msup><mml:mi>G</mml:mi><mml:mrow><mml:mtext>ext</mml:mtext></mml:mrow></mml:msup><mml:mo stretchy='false'>(</mml:mo><mml:mn>1</mml:mn><mml:mo>;</mml:mo><mml:mi>t</mml:mi><mml:mo stretchy='false'>)</mml:mo></mml:mrow></mml:mfrac><mml:mo>,</mml:mo></mml:mrow></mml:math></disp-formula>
<p><italic>which is defined for <italic>t</italic> &#x0003E; 0, is non-negative, and attains global maximum 1 at <inline-formula><mml:math id="M72"><mml:mover accent="true"><mml:mrow><mml:mi>x</mml:mi></mml:mrow><mml:mo>^</mml:mo></mml:mover><mml:mo>&#x000B7;</mml:mo><mml:mover accent="true"><mml:mrow><mml:mi>y</mml:mi></mml:mrow><mml:mo class="qopname">^</mml:mo></mml:mover><mml:mo>=</mml:mo><mml:mn>1</mml:mn></mml:math></inline-formula></italic>.</p>
<p>Note that unlike <inline-formula><mml:math id="M73"><mml:msup><mml:mrow><mml:mi>K</mml:mi></mml:mrow><mml:mrow><mml:mtext>prx</mml:mtext></mml:mrow></mml:msup><mml:mrow><mml:mo stretchy="false">(</mml:mo><mml:mrow><mml:mover accent="true"><mml:mrow><mml:mi>x</mml:mi></mml:mrow><mml:mo>^</mml:mo></mml:mover><mml:mo>,</mml:mo><mml:mover accent="true"><mml:mrow><mml:mi>y</mml:mi></mml:mrow><mml:mo class="qopname">^</mml:mo></mml:mover><mml:mo>;</mml:mo><mml:mi>t</mml:mi></mml:mrow><mml:mo stretchy="false">)</mml:mo></mml:mrow></mml:math></inline-formula>, <inline-formula><mml:math id="M74"><mml:msup><mml:mrow><mml:mi>K</mml:mi></mml:mrow><mml:mrow><mml:mtext>ext</mml:mtext></mml:mrow></mml:msup><mml:mrow><mml:mo stretchy="false">(</mml:mo><mml:mrow><mml:mover accent="true"><mml:mrow><mml:mi>x</mml:mi></mml:mrow><mml:mo>^</mml:mo></mml:mover><mml:mo>,</mml:mo><mml:mover accent="true"><mml:mrow><mml:mi>y</mml:mi></mml:mrow><mml:mo class="qopname">^</mml:mo></mml:mover><mml:mo>;</mml:mo><mml:mi>t</mml:mi></mml:mrow><mml:mo stretchy="false">)</mml:mo></mml:mrow></mml:math></inline-formula> explicitly depends on the feature dimension <italic>n</italic>. In general, SVM kernel hyperparameter tuning can be computationally costly for a data set with both high feature dimension and large sample size. In particular, choosing an appropriate diffusion time scale is an important challenge. On the one hand, choosing a very large value of <italic>t</italic> will make the series converge rapidly; but, then, all points will become uniformly similar, and the kernel will not be very useful. On the other hand, a too small value of <italic>t</italic> will make most data pairs too dissimilar, again limiting the applicability of the kernel. In practice, we thus need a guideline for a finite time scale at which the degree of &#x0201C;self-relatedness&#x0201D; is not singular, but still larger than the &#x0201C;relatedness&#x0201D; averaged over the whole hypersphere. Examining the asymptotic behavior of the exact heat kernel in high feature dimension <italic>n</italic> shows that an appropriate time scale is <inline-formula><mml:math id="M75"><mml:mi>t</mml:mi><mml:mo>&#x0007E;</mml:mo><mml:mrow><mml:mi mathvariant="-tex-caligraphic">O</mml:mi></mml:mrow><mml:mrow><mml:mo stretchy="false">(</mml:mo><mml:mrow><mml:mo class="qopname">log</mml:mo><mml:mi>n</mml:mi><mml:mo>/</mml:mo><mml:mi>n</mml:mi></mml:mrow><mml:mo stretchy="false">)</mml:mo></mml:mrow></mml:math></inline-formula>; in this regime the numerical sum in Theorem 1 satisfies a stopping condition at low orders in &#x02113; and the sample points are in moderate diffusion proximity to each other so that they can be accurately classified (Supplementary Material, section 2.5.4).</p>
<p>Figure <xref ref-type="fig" rid="F1">1A</xref> illustrates the diffusion process captured by our exact kernel <inline-formula><mml:math id="M76"><mml:msup><mml:mrow><mml:mi>K</mml:mi></mml:mrow><mml:mrow><mml:mtext>ext</mml:mtext></mml:mrow></mml:msup><mml:mrow><mml:mo stretchy="false">(</mml:mo><mml:mrow><mml:mover accent="true"><mml:mrow><mml:mi>x</mml:mi></mml:mrow><mml:mo>^</mml:mo></mml:mover><mml:mo>,</mml:mo><mml:mover accent="true"><mml:mrow><mml:mi>y</mml:mi></mml:mrow><mml:mo class="qopname">^</mml:mo></mml:mover><mml:mo>;</mml:mo><mml:mi>t</mml:mi></mml:mrow><mml:mo stretchy="false">)</mml:mo></mml:mrow></mml:math></inline-formula> in three feature dimensions at time <italic>t</italic> &#x0003D; <italic>t</italic><sup>&#x0002A;</sup>log3/3, for <italic>t</italic><sup>&#x0002A;</sup> &#x0003D; 0.5, 1.0, 2.0. In Figure <xref ref-type="fig" rid="F1">1B</xref>, we systematically compared the behavior of (1) dimension-independent parametrix kernel <italic>K</italic><sup>prx</sup> at time <italic>t</italic> &#x0003D; 0.5, 1.0, 2.0 and (2) exact kernel <italic>K</italic><sup>ext</sup> on <italic>S</italic><sup><italic>n</italic>&#x02212;1</sup> at <italic>t</italic> &#x0003D; <italic>t</italic><sup>&#x0002A;</sup>log<italic>n</italic>/<italic>n</italic> for <italic>t</italic><sup>&#x0002A;</sup> &#x0003D; 0.5, 1.0, 2.0 and <italic>n</italic> &#x0003D; 3, 100, 200. By symmetry, the slope of <italic>K</italic><sup>ext</sup> vanished at the south pole &#x003B8; &#x0003D; &#x003C0; for any time <italic>t</italic> and dimension <italic>n</italic>. In sharp contrast, <italic>K</italic><sup>prx</sup> had a negative slope at &#x003B8; &#x0003D; &#x003C0;, again highlighting a singular behavior of the parametrix kernel. The &#x0201C;relatedness&#x0201D; measured by <italic>K</italic><sup>ext</sup> at the sweet spot <italic>t</italic> &#x0003D; log<italic>n</italic>/<italic>n</italic> was finite over the whole hypersphere with sufficient contrast between nearby and far away points. Moreover, the characteristic behavior of <italic>K</italic><sup>ext</sup> at <italic>t</italic> &#x0003D; log<italic>n</italic>/<italic>n</italic> did not change significantly for different values of the feature dimension <italic>n</italic>, confirming that the optimal <italic>t</italic> for many classification applications will likely reside near the &#x0201C;sweet spot&#x0201D; <italic>t</italic> &#x0003D; log<italic>n</italic>/<italic>n</italic>.</p>
<fig id="F1" position="float">
<label>Figure 1</label>
<caption><p><bold>(A)</bold> Color maps of the exact kernel <italic>K</italic><sup>ext</sup> on <italic>S</italic><sup>2</sup> at rescaled time <italic>t</italic>&#x0002A; &#x0003D; 0.5, 1.0, 2.0; the white paths are simulated random walks on <italic>S</italic><sup>2</sup> with the Monte Carlo time approximately equal to <italic>t</italic> &#x0003D; <italic>t</italic>&#x0002A;log3/3. <bold>(B)</bold> Plots of the parametrix kernel <italic>K</italic><sup>prx</sup> and exact kernel <italic>K</italic><sup>ext</sup> on <italic>S</italic><sup><italic>n</italic>&#x02212;1</sup>, for <italic>n</italic> &#x0003D; 3, 100, 200, as functions of the geodesic distance.</p></caption>
<graphic xlink:href="fams-04-00001-g0001.tif"/>
</fig>
</sec>
<sec>
<title>2.3. SVM classifications</title>
<p>Linear SVM seeks a separating hyperplane that maximizes the margin, i.e., the distance to the nearest data point. The primal formulation of SVM attempts to minimize the norm of the weight vector <bold>w</bold> that is normal to the separating hyperplane, subject to either hard or soft margin constraints. In the so-called Lagrange dual formulation of SVM, one applies the Representer Theorem to rewrite the weight as a linear combination of data points; in this set-up, the dot products of data points naturally appear, and kernel SVM replaces the dot product operation with a chosen kernel evaluation. The ultimate hope is that the data points will become linearly separable in the new feature space implicitly defined by the kernel.</p>
<p>We evaluated the performance of kernel SVM using the</p>
<list list-type="order">
<list-item><p>linear kernel <italic>K</italic><sup>lin</sup>(<bold>x</bold>, <bold>y</bold>) &#x0003D; <bold>x</bold> &#x000B7; <bold>y</bold>,</p></list-item>
<list-item><p>Gaussian RBF <italic>K</italic><sup>rbf</sup>(<bold>x</bold>, <bold>y</bold>; &#x003B3;) &#x0003D; exp{&#x02212;&#x003B3;|<bold>x</bold> &#x02212; <bold>y</bold>|<sup>2</sup>},</p></list-item>
<list-item><p>cosine kernel <inline-formula><mml:math id="M77"><mml:msup><mml:mrow><mml:mi>K</mml:mi></mml:mrow><mml:mrow><mml:mtext>cos</mml:mtext></mml:mrow></mml:msup><mml:mrow><mml:mo stretchy="false">(</mml:mo><mml:mrow><mml:mover accent="true"><mml:mrow><mml:mi>x</mml:mi></mml:mrow><mml:mo>^</mml:mo></mml:mover><mml:mo>,</mml:mo><mml:mover accent="true"><mml:mrow><mml:mi>y</mml:mi></mml:mrow><mml:mo class="qopname">^</mml:mo></mml:mover></mml:mrow><mml:mo stretchy="false">)</mml:mo></mml:mrow><mml:mo>=</mml:mo><mml:mover accent="true"><mml:mrow><mml:mi>x</mml:mi></mml:mrow><mml:mo>^</mml:mo></mml:mover><mml:mo>&#x000B7;</mml:mo><mml:mover accent="true"><mml:mrow><mml:mi>y</mml:mi></mml:mrow><mml:mo class="qopname">^</mml:mo></mml:mover></mml:math></inline-formula>,</p></list-item>
<list-item><p>parametrix kernel <inline-formula><mml:math id="M78"><mml:msup><mml:mrow><mml:mi>K</mml:mi></mml:mrow><mml:mrow><mml:mtext>prx</mml:mtext></mml:mrow></mml:msup><mml:mrow><mml:mo stretchy="false">(</mml:mo><mml:mrow><mml:mover accent="true"><mml:mrow><mml:mi>x</mml:mi></mml:mrow><mml:mo>^</mml:mo></mml:mover><mml:mo>,</mml:mo><mml:mover accent="true"><mml:mrow><mml:mi>y</mml:mi></mml:mrow><mml:mo class="qopname">^</mml:mo></mml:mover><mml:mo>;</mml:mo><mml:mi>t</mml:mi></mml:mrow><mml:mo stretchy="false">)</mml:mo></mml:mrow></mml:math></inline-formula>, and</p></list-item>
<list-item><p>exact kernel <inline-formula><mml:math id="M79"><mml:msup><mml:mrow><mml:mi>K</mml:mi></mml:mrow><mml:mrow><mml:mtext>ext</mml:mtext></mml:mrow></mml:msup><mml:mrow><mml:mo stretchy="false">(</mml:mo><mml:mrow><mml:mover accent="true"><mml:mrow><mml:mi>x</mml:mi></mml:mrow><mml:mo>^</mml:mo></mml:mover><mml:mo>,</mml:mo><mml:mover accent="true"><mml:mrow><mml:mi>y</mml:mi></mml:mrow><mml:mo class="qopname">^</mml:mo></mml:mover><mml:mo>;</mml:mo><mml:mi>t</mml:mi></mml:mrow><mml:mo stretchy="false">)</mml:mo></mml:mrow></mml:math></inline-formula>,</p></list-item>
</list>
<p>on two independent data sets: (1) WebKB data of websites from four universities (WebKB-4-University) [<xref ref-type="bibr" rid="B21">21</xref>], and (2) glioblastoma multiforme (GBM) mutation data from The Cancer Genome Atlas (TCGA) with 5-fold cross-validations (CV) (Supplementary Material, section 1). The WebKB-4-University data contained 4,199 documents in total comprising four classes: student (1,641), faculty (1,124), course (930), and project (504); in our analysis, however, we selected an equal number of representative samples from each class, so that the training and testing sets had balanced classes. Table <xref ref-type="table" rid="T1">1</xref> shows the average optimal prediction accuracy scores of the five kernels for a varying number of representative samples, using 393 most frequent word features (Supplementary Material, section 1). The exact kernel outperformed the Gaussian RBF and parametrix kernel, reducing the error by 41 &#x0007E; 45% and by 1 &#x0007E; 7%, respectively. Changing the feature dimension did not affect the performance much (Table <xref ref-type="table" rid="T2">2</xref>).</p>
<table-wrap position="float" id="T1">
<label>Table 1</label>
<caption><p>WebKB-4-University Document Classification.</p></caption>
<table frame="hsides" rules="groups">
<thead><tr>
<th valign="top" align="left"><bold><italic>m</italic><sub>r</sub></bold></th>
<th valign="top" align="center"><bold>lin (%)</bold></th>
<th valign="top" align="center"><bold>rbf (%)</bold></th>
<th valign="top" align="center"><bold>cos (%)</bold></th>
<th valign="top" align="center"><bold>prx (%)</bold></th>
<th valign="top" align="center"><bold>ext (%)</bold></th>
</tr>
</thead>
<tbody>
<tr>
<td valign="top" align="left">100</td>
<td valign="top" align="center">74.2</td>
<td valign="top" align="center">75.1</td>
<td valign="top" align="center">84.4</td>
<td valign="top" align="center">85.4</td>
<td valign="top" align="center"><bold>85.6</bold></td>
</tr>
<tr>
<td valign="top" align="left">200</td>
<td valign="top" align="center">80.9</td>
<td valign="top" align="center">82.0</td>
<td valign="top" align="center">89.2</td>
<td valign="top" align="center">89.6</td>
<td valign="top" align="center"><bold>89.9</bold></td>
</tr>
<tr>
<td valign="top" align="left">300</td>
<td valign="top" align="center">83.2</td>
<td valign="top" align="center">84.1</td>
<td valign="top" align="center">89.9</td>
<td valign="top" align="center">90.5</td>
<td valign="top" align="center"><bold>91.1</bold></td>
</tr>
<tr>
<td valign="top" align="left">400</td>
<td valign="top" align="center">86.7</td>
<td valign="top" align="center">86.1</td>
<td valign="top" align="center">91.3</td>
<td valign="top" align="center">91.7</td>
<td valign="top" align="center"><bold>92.3</bold></td>
</tr>
</tbody>
</table>
<table-wrap-foot>
<p><italic>Performance test on four-class (student, faculty, course, and project) classification of WebKB-4-University word count data with different number m<sub>r</sub> of representatives for each class, for m<sub>r</sub> &#x0003D; 100, 200, 300, 400. The entries show the average of optimal 5-fold cross-validation mean accuracy scores of five runs. The exact kernel (ext) reduced the error of parametrix kernel (prx) by 1 &#x0007E; 7% and the Gaussian RBF (rbf) by 41 &#x0007E; 45%; the cosine kernel (cos) also reduced the error of linear kernel (lin) by 34 &#x0007E; 43%. The bold values indicate the highest accuracy score in each row</italic>.</p>
</table-wrap-foot>
</table-wrap>
<table-wrap position="float" id="T2">
<label>Table 2</label>
<caption><p>WebKB-4-University Document Classification.</p></caption>
<table frame="hsides" rules="groups">
<thead><tr>
<th valign="top" align="left"><bold><italic>n</italic></bold></th>
<th valign="top" align="left"><bold><italic>m</italic><sub>r</sub></bold></th>
<th valign="top" align="center"><bold>lin (%)</bold></th>
<th valign="top" align="center"><bold>rbf (%)</bold></th>
<th valign="top" align="center"><bold>cos (%)</bold></th>
<th valign="top" align="center"><bold>prx (%)</bold></th>
<th valign="top" align="center"><bold>ext (%)</bold></th>
</tr>
</thead>
<tbody>
<tr>
<td valign="top" align="left">393</td>
<td valign="top" align="left">400</td>
<td valign="top" align="center">86.73</td>
<td valign="top" align="center">86.27</td>
<td valign="top" align="center">91.57</td>
<td valign="top" align="center">91.99</td>
<td valign="top" align="center"><bold>92.44</bold></td>
</tr>
<tr>
<td valign="top" align="left">726</td>
<td valign="top" align="left">400</td>
<td valign="top" align="center">86.78</td>
<td valign="top" align="center">86.95</td>
<td valign="top" align="center">92.62</td>
<td valign="top" align="center">92.91</td>
<td valign="top" align="center"><bold>93.00</bold></td>
</tr>
<tr>
<td valign="top" align="left">1,023</td>
<td valign="top" align="left">400</td>
<td valign="top" align="center">85.56</td>
<td valign="top" align="center">86.11</td>
<td valign="top" align="center">92.62</td>
<td valign="top" align="center">92.74</td>
<td valign="top" align="center"><bold>92.91</bold></td>
</tr>
<tr>
<td valign="top" align="left">1,312</td>
<td valign="top" align="left">400</td>
<td valign="top" align="center">85.78</td>
<td valign="top" align="center">86.75</td>
<td valign="top" align="center">92.56</td>
<td valign="top" align="center">92.81</td>
<td valign="top" align="center"><bold>93.03</bold></td>
</tr>
</tbody>
</table>
<table-wrap-foot>
<p><italic>Comparison of kernel SVMs on the WebKB-4-University data with a fixed sample size m<sub>r</sub>, but varying feature dimension n. To account for the randomness in selecting the representative samples using KMeans (Supplementary Material, section 1), we performed fives runs of representative selection, and then performed CV using the training and test sets obtained from each run. Finally, we averaged the five mean CV scores to assess the performance of each classifier on the imbalanced WebKB-4-University data set. The exact (ext) and cosine (cos) kernels outperformed the Gaussian RBF (rbf) and linear (lin) kernels in various feature dimensions n &#x0003D; 393,726,1,023, and 1,312, with fixed and balanced class size m<sub>r</sub> &#x0003D; 400. A word was selected as a feature if its total count was greater than 1/10, 1/20, 1/30 or 1/40 times the total number of web pages in the WebKB-4-University data set, with the different thresholds corresponding to the different rows in the table. The exact kernel reduced the errors of Gaussian RBF and parametrix kernels by 45 &#x0007E; 48% and 1 &#x0007E; 6%, respectively; the cosine kernel reduced the errors of linear kernel by 36 &#x0007E; 49%. The bold values indicate the highest accuracy score in each row</italic>.</p>
</table-wrap-foot>
</table-wrap>
<p>In the TCGA-GBM data, there were 497 samples, and we aimed to impute the mutation status of one gene&#x02014;i.e., mutant or wild-type&#x02014;from the mutation counts of other genes. For each imputation target, we first counted the number <italic>m</italic><sub>r</sub> of mutant samples and then selected an equal number of wild-type samples for 5-fold CV. Imputation tests were performed for top 102 imputable genes (Supplementary Material, section 1). Table <xref ref-type="table" rid="T3">3</xref> shows the average prediction accuracy scores for 5 biologically interesting genes known to be important for cancer [<xref ref-type="bibr" rid="B22">22</xref>]:</p>
<list list-type="order">
<list-item><p><italic>ZMYM4</italic> (<italic>m</italic><sub>r</sub> &#x0003D; 33) is implicated in an antiapoptotic activity; [<xref ref-type="bibr" rid="B23">23</xref>, <xref ref-type="bibr" rid="B24">24</xref>];</p></list-item>
<list-item><p><italic>ADGRB3</italic> (<italic>m</italic><sub>r</sub> &#x0003D; 37) is a brain-specific angiogenesis inhibitor [<xref ref-type="bibr" rid="B25">25</xref>&#x02013;<xref ref-type="bibr" rid="B27">27</xref>];</p></list-item>
<list-item><p><italic>NFX1</italic> (<italic>m</italic><sub>r</sub> &#x0003D; 42) is a repressor of <italic>hTERT</italic> transcription [<xref ref-type="bibr" rid="B28">28</xref>] and is thought to regulate inflammatory response [<xref ref-type="bibr" rid="B29">29</xref>];</p></list-item>
<list-item><p><italic>P2RX7</italic> (<italic>m</italic><sub>r</sub> &#x0003D; 48) encodes an ATP receptor which plays a key role in restricting tumor growth and metastases [<xref ref-type="bibr" rid="B30">30</xref>&#x02013;<xref ref-type="bibr" rid="B32">32</xref>];</p></list-item>
<list-item><p><italic>COL1A2</italic> (<italic>m</italic><sub>r</sub> &#x0003D; 61) is overexpressed in the medulloblastoma microenvironment and is a potential therapeutic target [<xref ref-type="bibr" rid="B33">33</xref>&#x02013;<xref ref-type="bibr" rid="B35">35</xref>].</p></list-item>
</list>
<table-wrap position="float" id="T3">
<label>Table 3</label>
<caption><p>TCGA-GBM Genotype Imputation.</p></caption>
<table frame="hsides" rules="groups">
<thead><tr>
<th/>
<th valign="top" align="center"><bold>lin (%)</bold></th>
<th valign="top" align="center"><bold>rbf (%)</bold></th>
<th valign="top" align="center"><bold>cos (%)</bold></th>
<th valign="top" align="center"><bold>prx (%)</bold></th>
<th valign="top" align="center"><bold>ext (%)</bold></th>
</tr>
</thead>
<tbody>
<tr>
<td valign="top" align="left"><italic>ZMYM4</italic></td>
<td valign="top" align="center">82.9</td>
<td valign="top" align="center">84.0</td>
<td valign="top" align="center">83.6</td>
<td valign="top" align="center">84.1</td>
<td valign="top" align="center"><bold>85.1</bold></td>
</tr>
<tr>
<td valign="top" align="left"><italic>ADGRB3</italic></td>
<td valign="top" align="center">75.7</td>
<td valign="top" align="center"><bold>81.0</bold></td>
<td valign="top" align="center">78.0</td>
<td valign="top" align="center">79.5</td>
<td valign="top" align="center">79.3</td>
</tr>
<tr>
<td valign="top" align="left"><italic>NFX1</italic></td>
<td valign="top" align="center">73.0</td>
<td valign="top" align="center">81.2</td>
<td valign="top" align="center">80.9</td>
<td valign="top" align="center"><bold>82.7</bold></td>
<td valign="top" align="center">82.5</td>
</tr>
<tr>
<td valign="top" align="left"><italic>P2RX7</italic></td>
<td valign="top" align="center">79.2</td>
<td valign="top" align="center">84.1</td>
<td valign="top" align="center"><bold>85.0</bold></td>
<td valign="top" align="center">84.0</td>
<td valign="top" align="center"><bold>85.0</bold></td>
</tr>
<tr>
<td valign="top" align="left"><italic>COL1A2</italic></td>
<td valign="top" align="center">68.4</td>
<td valign="top" align="center">70.5</td>
<td valign="top" align="center">72.9</td>
<td valign="top" align="center">73.9</td>
<td valign="top" align="center"><bold>74.2</bold></td>
</tr>
</tbody>
</table>
<table-wrap-foot>
<p><italic>Performance test on binary classification of mutant vs. wild-type in TCGA-GBM mutation count data. The rows are different genes, the mutation statuses of which were imputed using m<sub>r</sub> samples in each mutant and wild-type class. The entries show the average of optimal 5-fold cross-validation mean accuracy scores of five runs. The bold values indicate the highest accuracy score in each row</italic>.</p>
</table-wrap-foot>
</table-wrap>
<p>For the remaining genes, the exact kernel generally outperformed the linear, cosine and parametrix kernels (Figure <xref ref-type="fig" rid="F2">2</xref>). However, even though the exact kernel dramatically outperformed the Gaussian RBF in the WebKB-4-University classification problem, the advantage of the exact kernel in this mutation analysis was not evident (Figure <xref ref-type="fig" rid="F2">2</xref>). It is possible that the radial degree of freedom <inline-formula><mml:math id="M80"><mml:munderover accentunder="false" accent="false"><mml:mrow><mml:mo>&#x02211;</mml:mo></mml:mrow><mml:mrow><mml:mi>i</mml:mi><mml:mo>=</mml:mo><mml:mn>1</mml:mn></mml:mrow><mml:mrow><mml:mi>n</mml:mi></mml:mrow></mml:munderover><mml:msub><mml:mrow><mml:mi>x</mml:mi></mml:mrow><mml:mrow><mml:mi>i</mml:mi></mml:mrow></mml:msub></mml:math></inline-formula> in this case, corresponding to the genome-wide mutation load in each sample, contained important covariate information not captured by the hyperspherical heat kernel. The difference in accuracy between the hyperspherical kernels (cos, prx, and ext) and the Euclidean kernels (lin and rbf) also hinted some weak dependence on class size <italic>m</italic><sub>r</sub> (Figure <xref ref-type="fig" rid="F2">2</xref>), or equivalently the sample size <italic>m</italic> &#x0003D; 2<italic>m</italic><sub>r</sub>. In fact, the level of accuracy showed much stronger correlation with the &#x0201C;effective sample size&#x0201D; <inline-formula><mml:math id="M81"><mml:mover accent="true"><mml:mrow><mml:mi>m</mml:mi></mml:mrow><mml:mo>&#x0007E;</mml:mo></mml:mover></mml:math></inline-formula> related to the empirical Vapnik-Chervonenkis (VC) dimension [<xref ref-type="bibr" rid="B4">4</xref>, <xref ref-type="bibr" rid="B7">7</xref>, <xref ref-type="bibr" rid="B36">36</xref>&#x02013;<xref ref-type="bibr" rid="B38">38</xref>] of a kernel SVM classifier (Figures <xref ref-type="fig" rid="F3">3A&#x02013;E</xref>); moreover, the advantage of the exact kernel over the Guassian RBF kernel grew with the effective sample size ratio <inline-formula><mml:math id="M82"><mml:msub><mml:mrow><mml:mover accent="true"><mml:mrow><mml:mi>m</mml:mi></mml:mrow><mml:mo>&#x0007E;</mml:mo></mml:mover></mml:mrow><mml:mrow><mml:mtext>cos</mml:mtext></mml:mrow></mml:msub><mml:mo>/</mml:mo><mml:msub><mml:mrow><mml:mover accent="true"><mml:mrow><mml:mi>m</mml:mi></mml:mrow><mml:mo>&#x0007E;</mml:mo></mml:mover></mml:mrow><mml:mrow><mml:mtext>lin</mml:mtext></mml:mrow></mml:msub></mml:math></inline-formula> (Figure <xref ref-type="fig" rid="F3">3F</xref>, Supplementary Material, section 2.5.5).</p>
<fig id="F2" position="float">
<label>Figure 2</label>
<caption><p>Comparison of the classification accuracy of SVM using linear (lin), cosine (cos), Gaussian RBF (rbf), parametrix (prx), and exact (ext) kernels on TCGA mutation count data. The plots show the ratio of accuracy scores for two different kernels. For visualization purpose, we excluded one gene with <italic>m</italic><sub>r</sub> &#x0003D; 250. The ratios rbf/lin, prx/cos, and ext/cos were essentially constant in class size <italic>m</italic><sub>r</sub> and greater than 1; in other words, the Gaussian RBF (rbf) kernel outperformed the linear (lin) kernel, while the exact (ext) and parametrix (prx) kernels outperformed the cosine (cos) kernel uniformly over all values of class size <italic>m</italic><sub>r</sub>. However, the more negative slope in the linear fit of cos/lin hints that the accuracy scores of cosine and linear kernels may depend on the class size <italic>m</italic><sub>r</sub>; the exact kernel also tended to outperform Gaussian RBF kernel when <italic>m</italic><sub>r</sub> was small.</p></caption>
<graphic xlink:href="fams-04-00001-g0002.tif"/>
</fig>
<fig id="F3" position="float">
<label>Figure 3</label>
<caption><p><bold>(A)</bold> A strong linear relation is seen between the VC-bound for cosine kernel <inline-formula><mml:math id="M83"><mml:msubsup><mml:mrow><mml:mi>&#x003BC;</mml:mi></mml:mrow><mml:mrow><mml:mtext>VC</mml:mtext></mml:mrow><mml:mrow><mml:mo>&#x0002A;</mml:mo><mml:mo class="qopname">cos</mml:mo></mml:mrow></mml:msubsup></mml:math></inline-formula> and class size <italic>m</italic><sub>r</sub>. The dashed line marks <italic>y</italic> &#x0003D; <italic>x</italic>; the VC-bound for linear kernel, however, was a constant <inline-formula><mml:math id="M84"><mml:msubsup><mml:mrow><mml:mi>&#x003BC;</mml:mi></mml:mrow><mml:mrow><mml:mtext>VC</mml:mtext></mml:mrow><mml:mrow><mml:mo>&#x0002A;</mml:mo><mml:mtext>lin</mml:mtext></mml:mrow></mml:msubsup><mml:mo>=</mml:mo><mml:mn>439</mml:mn></mml:math></inline-formula>. <bold>(B&#x02013;E)</bold> The scatter plots of accuracy scores for cosine (cos), linear (lin), exact (ext), and Gaussian RBF (rbf) kernels vs. the effective sample size <inline-formula><mml:math id="M85"><mml:mover accent="true"><mml:mrow><mml:mi>m</mml:mi></mml:mrow><mml:mo>&#x0007E;</mml:mo></mml:mover><mml:mo>=</mml:mo><mml:mn>2</mml:mn><mml:msub><mml:mrow><mml:mi>m</mml:mi></mml:mrow><mml:mrow><mml:mtext>r</mml:mtext></mml:mrow></mml:msub><mml:mo>/</mml:mo><mml:msubsup><mml:mrow><mml:mi>&#x003BC;</mml:mi></mml:mrow><mml:mrow><mml:mtext>VC</mml:mtext></mml:mrow><mml:mrow><mml:mo>&#x0002A;</mml:mo></mml:mrow></mml:msubsup></mml:math></inline-formula>; the accuracy scores of exact and cosine kernels increased with the effective sample size, whereas those of Gaussian RBF and linear kernels tended to decrease with the effective sample size. <bold>(F)</bold> The ratio of ext vs. rbf accuracy scores is positively correlated with the ratio <inline-formula><mml:math id="M86"><mml:msub><mml:mrow><mml:mover accent="true"><mml:mrow><mml:mi>m</mml:mi></mml:mrow><mml:mo>&#x0007E;</mml:mo></mml:mover></mml:mrow><mml:mrow><mml:mtext>cos</mml:mtext></mml:mrow></mml:msub><mml:mo>/</mml:mo><mml:msub><mml:mrow><mml:mover accent="true"><mml:mrow><mml:mi>m</mml:mi></mml:mrow><mml:mo>&#x0007E;</mml:mo></mml:mover></mml:mrow><mml:mrow><mml:mtext>lin</mml:mtext></mml:mrow></mml:msub></mml:math></inline-formula> of effective sample sizes.</p></caption>
<graphic xlink:href="fams-04-00001-g0003.tif"/>
</fig>
<p>By construction, our definition of the hyperspherical map exploits only the positive portion of the whole hypersphere, where the parametrix and exact heat kernels seem to have similar performances. However, if we allow the data set to assume negative values, i.e., the feature space is the usual &#x0211D;<sup><italic>n</italic></sup>\{0} instead of <inline-formula><mml:math id="M87"><mml:msubsup><mml:mrow><mml:mi>&#x0211D;</mml:mi></mml:mrow><mml:mrow><mml:mo>&#x02265;</mml:mo><mml:mn>0</mml:mn></mml:mrow><mml:mrow><mml:mi>n</mml:mi></mml:mrow></mml:msubsup><mml:mo>\</mml:mo><mml:mrow><mml:mo>{</mml:mo><mml:mrow><mml:mn>0</mml:mn></mml:mrow><mml:mo>}</mml:mo></mml:mrow></mml:math></inline-formula>, then we may apply the usual projective map, where each vector in the Euclidean space is normalized by its <italic>L</italic><sup>2</sup>-norm. As shown in Figure <xref ref-type="fig" rid="F1">1B</xref>, the parametrix kernel is singular at &#x003B8; &#x0003D; &#x003C0; and qualitatively deviates from the exact kernel for large values of &#x003B8;. Thus, when data points populate the whole hypersphere, we expect to find more significant differences in performance between the exact and parametrix kernels. For example, Table <xref ref-type="table" rid="T4">4</xref> shows the kernel SVM classifications of 91 S&#x00026;P500 <italic>Financials</italic> stocks against 64 <italic>Information Technology</italic> stocks (<italic>m</italic> &#x0003D; 155) using their log-return instances between January 5, 2015 and November 18, 2016 as features. As long as the number of features was greater than sample size, <italic>n</italic> &#x0003E; <italic>m</italic>, the exact kernel outperformed all other kernels and reduced the error of Gaussian RBF by 29 &#x0007E; 51% and that of parametrix kernel by 17 &#x0007E; 51%.</p>
<table-wrap position="float" id="T4">
<label>Table 4</label>
<caption><p>S&#x00026;P500 Stock Classification.</p></caption>
<table frame="hsides" rules="groups">
<thead><tr>
<th valign="top" align="right"><bold><italic>n</italic></bold></th>
<th valign="top" align="right"><bold><italic>m</italic></bold></th>
<th valign="top" align="center"><bold>lin (%)</bold></th>
<th valign="top" align="center"><bold>rbf (%)</bold></th>
<th valign="top" align="center"><bold>cos (%)</bold></th>
<th valign="top" align="center"><bold>prx (%)</bold></th>
<th valign="top" align="center"><bold>ext (%)</bold></th>
</tr>
</thead>
<tbody>
<tr>
<td valign="top" align="right">475</td>
<td valign="top" align="right">155</td>
<td valign="top" align="center">98.06</td>
<td valign="top" align="center">98.69</td>
<td valign="top" align="center">98.69</td>
<td valign="top" align="center">98.69</td>
<td valign="top" align="center"><bold>99.35</bold></td>
</tr>
<tr>
<td valign="top" align="right">238</td>
<td valign="top" align="right">155</td>
<td valign="top" align="center">95.50</td>
<td valign="top" align="center">96.77</td>
<td valign="top" align="center">94.82</td>
<td valign="top" align="center">96.13</td>
<td valign="top" align="center"><bold>98.06</bold></td>
</tr>
<tr>
<td valign="top" align="right">159</td>
<td valign="top" align="right">155</td>
<td valign="top" align="center">94.86</td>
<td valign="top" align="center">95.48</td>
<td valign="top" align="center">95.48</td>
<td valign="top" align="center">96.13</td>
<td valign="top" align="center"><bold>96.79</bold></td>
</tr>
<tr>
<td valign="top" align="right">119</td>
<td valign="top" align="right">155</td>
<td valign="top" align="center">92.86</td>
<td valign="top" align="center">93.53</td>
<td valign="top" align="center">91.57</td>
<td valign="top" align="center"><bold>94.15</bold></td>
<td valign="top" align="center"><bold>94.15</bold></td>
</tr>
<tr>
<td valign="top" align="right">95</td>
<td valign="top" align="right">155</td>
<td valign="top" align="center">91.55</td>
<td valign="top" align="center"><bold>95.50</bold></td>
<td valign="top" align="center">94.19</td>
<td valign="top" align="center">94.15</td>
<td valign="top" align="center">94.79</td>
</tr>
</tbody>
</table>
<table-wrap-foot>
<p><italic>Classifications were performed on m &#x0003D; 155 stocks from S&#x00026;P500 companies: 91 Financial vs. 64 Information Technology. The 475 log-return instances between January 5, 2015 and November 18, 2016 were used as features. We uniformly subsampled the instances to generate variations in the feature dimension n. Here, we report the mean 5-fold CV accuracy score for each kernel. Although the two classes were slightly imbalanced, all scores were much larger than the &#x0201C;random score&#x0201D; 91/155 &#x02248; 58.7%, calculated from the majority class size and sample size. For n &#x0003E; m, the exact (ext) kernel outperformed all other kernels and reduced the errors of Gaussian RBF (rbf) and parametrix (prx) kernels by 29 &#x0007E; 51% and 17 &#x0007E; 51%, respectively. When n &#x0003C; m, the exact kernel started to lose its advantage over the Gaussian RBF kernel. The bold values indicate the highest accuracy score in each row</italic>.</p>
</table-wrap-foot>
</table-wrap>
</sec>
</sec>
<sec sec-type="discussion" id="s3">
<title>3. Discussion</title>
<p>This paper has constructed the exact hyperspherical heat kernel using the complete basis of high-dimensional angular momentum eigenfunctions and tested its performance in kernel SVM. We have shown that the exact kernel and cosine kernel, both of which employ the hyperspherical maps, often outperform the Gaussian RBF and linear kernels. The advantage of using hyperspherical kernels likely arises from the hyperspherical maps of feature space, and the exact kernel may further improve the decision boundary flexibility of the raw cosine kernel. To be specific, the hyperspherical maps remove the less informative radial degree of freedom in a nonlinear fashion and compactify the Euclidean feature space into a unit hypersphere where all data points may then be enclosed within a finite radius. By contrast, our numerical estimations using TCGA-GBM data show that for linear kernel SVM, the margin <italic>M</italic> tends to be much smaller than the data range <italic>R</italic> in order to accommodate the separation of strongly mixed data points of different class labels; as a result, the ratio <italic>R</italic>/<italic>M</italic> was much larger than that for cosine kernel SVM. This insight may be summarized by the fact that the upper bound on the empirical VC-dimension of linear kernel SVM tends to be much larger than that for cosine kernel SVM, especially in high dimensions, suggesting that the cosine kernel SVM is less sensitive to noise and more generalizable to unseen data. The exact kernel is equipped with an additional tunable hyperparameter, namely the diffusion time <italic>t</italic>, which adjusts the curvature of nonlinear decision boundary and thus adds to the advantage of hyperspherical maps. Moreover, the hyperspherical kernels often have larger effective sample sizes than their Euclidean counterparts and, thus, may be especially useful for analyzing data with a small sample size in high feature dimensions.</p>
<p>The failure of the parametrix expansion of heat kernel, especially in dimensions <italic>n</italic> &#x0226B; 3, signals a dramatic difference between diffusion in a non-compact space and that on a compact manifold. It remains to be examined how these differences in diffusion process, random walk and topology between non-compact Euclidean spaces and compact manifolds like a hypersphere help improve clustering performance as supported by the results of this paper.</p>
</sec>
<sec id="s4">
<title>Author contributions</title>
<p>JS conceived the project and supervised CZ who performed the calculations and analyses. CZ and JS wrote the manuscript.</p>
<sec>
<title>Conflict of interest statement</title>
<p>The authors declare that the research was conducted in the absence of any commercial or financial relationships that could be construed as a potential conflict of interest.</p>
</sec>
</sec>
</body>
<back>
<ack><p>We thank Alex Finnegan and Hu Jin for critical reading of the manuscript and helpful comments. We also thank Mohith Manjunath for his help with the TCGA data.</p>
</ack>
<sec sec-type="supplementary-material" id="s6">
<title>Supplementary material</title>
<p>The Supplementary Material for this article can be found online at: <ext-link ext-link-type="uri" xlink:href="https://www.frontiersin.org/articles/10.3389/fams.2018.00001/full#supplementary-material">https://www.frontiersin.org/articles/10.3389/fams.2018.00001/full#supplementary-material</ext-link></p>
<supplementary-material xlink:href="Presentation1.PDF" id="SM1" mimetype="application/pdf" xmlns:xlink="http://www.w3.org/1999/xlink"/>
</sec>
<ref-list>
<title>References</title>
<ref id="B1">
<label>1.</label>
<citation citation-type="journal"><person-group person-group-type="author"><name><surname>Lafferty</surname> <given-names>J</given-names></name> <name><surname>Lebanon</surname> <given-names>G</given-names></name></person-group>. <article-title>Diffusion kernels on statistical manifolds</article-title>. <source>J. Mach. Learn. Res.</source> (<year>2005</year>) <volume>6</volume>:<fpage>129</fpage>&#x02013;<lpage>63</lpage>.</citation></ref>
<ref id="B2">
<label>2.</label>
<citation citation-type="book"><person-group person-group-type="author"><name><surname>Hastie</surname> <given-names>T</given-names></name> <name><surname>Tibshirani</surname> <given-names>R</given-names></name> <name><surname>Friedman</surname> <given-names>J</given-names></name></person-group>. <source>The Elements of Statistical Learning</source>. Data Mining, Inference, and Prediction. <publisher-name>Springer Science and Business Media</publisher-name> (<year>2013</year>). Available online at: <ext-link ext-link-type="uri" xlink:href="http://books.google.com/books?id=yPfZBwAAQBAJ&#x00026;printsec=frontcover&#x00026;dq=The&#x0002B;elelment&#x0002B;of&#x0002B;statistical&#x0002B;learning&#x00026;hl=&#x00026;cd=4&#x00026;source=gbs_api">http://books.google.com/books?id=yPfZBwAAQBAJ&#x00026;printsec=frontcover&#x00026;dq=The&#x0002B;elelment&#x0002B;of&#x0002B;statistical&#x0002B;learning&#x00026;hl=&#x00026;cd=4&#x00026;source=gbs_api</ext-link></citation>
</ref>
<ref id="B3">
<label>3.</label>
<citation citation-type="book"><person-group person-group-type="author"><name><surname>Evgeniou</surname> <given-names>T</given-names></name> <name><surname>Pontil</surname> <given-names>M</given-names></name></person-group>. <article-title>Support vector machines: theory and applications</article-title>. In: <source>Machine Learning and Its Applications</source>. <publisher-loc>Berlin, Heidelberg</publisher-loc>: <publisher-name>Springer</publisher-name> (<year>2001</year>). p. <fpage>249</fpage>&#x02013;<lpage>57</lpage>. Available online at: <ext-link ext-link-type="uri" xlink:href="http://link.springer.com/10.1007/3-540-44673-7_12">http://link.springer.com/10.1007/3-540-44673-7_12</ext-link></citation>
</ref>
<ref id="B4">
<label>4.</label>
<citation citation-type="book"><person-group person-group-type="author"><name><surname>Boser</surname> <given-names>BE</given-names></name> <name><surname>Guyon</surname> <given-names>IM</given-names></name> <name><surname>Vapnik</surname> <given-names>VN</given-names></name></person-group>. <article-title>A Training Algorithm for Optimal Margin Classifiers</article-title>. <publisher-loc>New York, NY</publisher-loc>: <publisher-name>ACM</publisher-name> (<year>1992</year>). Available online at: <ext-link ext-link-type="uri" xlink:href="http://portal.acm.org/citation.cfm?doid=130385.130401">http://portal.acm.org/citation.cfm?doid=130385.130401</ext-link></citation>
</ref>
<ref id="B5">
<label>5.</label>
<citation citation-type="journal"><person-group person-group-type="author"><name><surname>Cortes</surname> <given-names>C</given-names></name> <name><surname>Vapnik</surname> <given-names>V</given-names></name></person-group>. <article-title>Support-vector networks</article-title>. <source>Mach Learn.</source> (<year>1995</year>) <volume>20</volume>:<fpage>273</fpage>&#x02013;<lpage>97</lpage>. <pub-id pub-id-type="doi">10.1007/BF00994018</pub-id></citation>
</ref>
<ref id="B6">
<label>6.</label>
<citation citation-type="journal"><person-group person-group-type="author"><name><surname>Freund</surname> <given-names>Y</given-names></name> <name><surname>Schapire</surname> <given-names>RE</given-names></name></person-group>. <article-title>Large margin classification using the perceptron algorithm</article-title>. <source>Mach Learn.</source> (<year>1999</year>) <volume>37</volume>:<fpage>277</fpage>&#x02013;<lpage>96</lpage>. <pub-id pub-id-type="doi">10.1023/A:1007662407062</pub-id></citation>
</ref>
<ref id="B7">
<label>7.</label>
<citation citation-type="web"><person-group person-group-type="author"><name><surname>Guyon</surname> <given-names>I</given-names></name> <name><surname>Boser</surname> <given-names>B</given-names></name> <name><surname>Vapnik</surname> <given-names>V</given-names></name></person-group>. <article-title>Automatic capacity tuning of very large VC-dimension classifiers</article-title>. In: <source>Advances in Neural Information Processing Systems.</source> (<year>1993</year>). p. <fpage>147</fpage>&#x02013;<lpage>55</lpage>. Available online at: <ext-link ext-link-type="uri" xlink:href="https://pdfs.semanticscholar.org/4c8e/d38dff557e9c762836a8a80808c64a8dc76e.pdf">https://pdfs.semanticscholar.org/4c8e/d38dff557e9c762836a8a80808c64a8dc76e.pdf</ext-link></citation>
</ref>
<ref id="B8">
<label>8.</label>
<citation citation-type="book"><person-group person-group-type="author"><name><surname>Kaufman</surname> <given-names>L</given-names></name> <name><surname>Rousseeuw</surname> <given-names>PJ</given-names></name></person-group>. <article-title>Finding groups in data</article-title>. In: <person-group person-group-type="editor"><name><surname>Kaufman</surname> <given-names>L</given-names></name> <name><surname>Rousseeuw</surname> <given-names>PJ</given-names></name></person-group>, editors. <source>An Introduction to Cluster Analysis.</source> <publisher-loc>Hoboken, NJ</publisher-loc>: <publisher-name>John Wiley &#x00026; Sons</publisher-name> (<year>2009</year>). Available online at: <ext-link ext-link-type="uri" xlink:href="http://doi.wiley.com/10.1002/9780470316801">http://doi.wiley.com/10.1002/9780470316801</ext-link>.</citation>
</ref>
<ref id="B9">
<label>9.</label>
<citation citation-type="journal"><person-group person-group-type="author"><name><surname>Belkin</surname> <given-names>M</given-names></name> <name><surname>Niyogi</surname> <given-names>P</given-names></name> <name><surname>Sindhwani</surname> <given-names>V</given-names></name></person-group>. <article-title>Manifold regularization: a Geometric framework for learning from labeled and unlabeled examples</article-title>. <source>J Mach Learn Res.</source> (<year>2006</year>) <volume>7</volume>:<fpage>2399</fpage>&#x02013;<lpage>434</lpage>.</citation>
</ref>
<ref id="B10">
<label>10.</label>
<citation citation-type="journal"><person-group person-group-type="author"><name><surname>Aronszajn</surname> <given-names>N</given-names></name></person-group>. <article-title>Theory of reproducing kernels</article-title>. <source>Trans Am Math Soc.</source> (<year>1950</year>) <volume>68</volume>:<fpage>337</fpage>. <pub-id pub-id-type="doi">10.1090/S0002-9947-1950-0051437-7</pub-id></citation></ref>
<ref id="B11">
<label>11.</label>
<citation citation-type="book"><person-group person-group-type="author"><name><surname>Paulsen</surname> <given-names>VI</given-names></name> <name><surname>Raghupathi</surname> <given-names>M</given-names></name></person-group>. <source>An Introduction to the Theory of Reproducing Kernel Hilbert Spaces (Cambridge Studies in Advanced Mathematics).</source> <publisher-name>Cambridge University Press</publisher-name> (<year>2016</year>). Available online at: <ext-link ext-link-type="uri" xlink:href="http://books.google.com/books?id=rEfGCwAAQBAJ&#x00026;printsec=frontcover&#x00026;dq=intitle:An&#x0002B;introduction&#x0002B;to&#x0002B;the&#x0002B;theory&#x0002B;of&#x0002B;reproducing&#x0002B;kernel&#x0002B;Hilbert&#x0002B;spaces&#x00026;hl=&#x00026;cd=1&#x00026;source=gbs_api">http://books.google.com/books?id=rEfGCwAAQBAJ&#x00026;printsec=frontcover&#x00026;dq=intitle:An&#x0002B;introduction&#x0002B;to&#x0002B;the&#x0002B;theory&#x0002B;of&#x0002B;reproducing&#x0002B;kernel&#x0002B;Hilbert&#x0002B;spaces&#x00026;hl=&#x00026;cd=1&#x00026;source=gbs_api</ext-link></citation></ref>
<ref id="B12">
<label>12.</label>
<citation citation-type="book"><person-group person-group-type="author"><name><surname>Berger</surname> <given-names>M</given-names></name> <name><surname>Gauduchon</surname> <given-names>P</given-names></name> <name><surname>Mazet</surname> <given-names>E</given-names></name></person-group>. <source>Le Spectre d&#x00027;une Vari&#x000E9;t&#x000E9; Riemannienne</source>. <publisher-name>Springer</publisher-name> (<year>1971</year>). Available online at: <ext-link ext-link-type="uri" xlink:href="http://scholar.google.com/scholar?q=related:bCbU4y_3CgsJ:scholar.google.com/&#x00026;hl=en&#x00026;num=20&#x00026;as_sdt=0,5">http://scholar.google.com/scholar?q=related:bCbU4y_3CgsJ:scholar.google.com/&#x00026;hl=en&#x00026;num=20&#x00026;as_sdt=0,5</ext-link></citation></ref>
<ref id="B13">
<label>13.</label>
<citation citation-type="web"><person-group person-group-type="author"><name><surname>Hsu</surname> <given-names>EP</given-names></name></person-group>. <source>Stochastic analysis on manifolds, Volume 38 of Graduate Studies in Mathematics</source>. American Mathematical Society (<year>2002</year>). Available online at: <ext-link ext-link-type="uri" xlink:href="http://scholar.google.com/scholar?q=related:r70JUPP2G14J:scholar.google.com/&#x00026;hl=en&#x00026;num=20&#x00026;as_sdt=0,5">http://scholar.google.com/scholar?q=related:r70JUPP2G14J:scholar.google.com/&#x00026;hl=en&#x00026;num=20&#x00026;as_sdt=0,5</ext-link>.</citation></ref>
<ref id="B14">
<label>14.</label>
<citation citation-type="web"><person-group person-group-type="author"><name><surname>Varopoulos</surname> <given-names>NT,. Random Walks Brownian Motion on Manifolds. Symposia Mathematica</given-names></name></person-group> (<year>1987</year>). Available online at: <ext-link ext-link-type="uri" xlink:href="http://scholar.google.com/scholar?q=related:oap-nC2spI8J:scholar.google.com/&#x00026;hl=en&#x00026;num=20&#x00026;as_sdt=0,5">http://scholar.google.com/scholar?q=related:oap-nC2spI8J:scholar.google.com/&#x00026;hl=en&#x00026;num=20&#x00026;as_sdt=0,5</ext-link></citation></ref>
<ref id="B15">
<label>15.</label>
<citation citation-type="web"><person-group person-group-type="author"><name><surname>Ng</surname> <given-names>A</given-names></name> <name><surname>Jordan</surname> <given-names>M</given-names></name> <name><surname>Weiss</surname> <given-names>Y</given-names></name> <name><surname>Dietterich</surname> <given-names>T</given-names></name> <name><surname>Becker</surname> <given-names>S</given-names></name></person-group>. <source>Advances in Neural Information Processing Systems, 14, Chapter On spectral clustering: analysis and an algorithm</source> (<year>2002</year>). Available online at: <ext-link ext-link-type="uri" xlink:href="http://scholar.google.com/scholar?q=related:N4YMyham6pUJ:scholar.google.com/&#x00026;hl=en&#x00026;num=20&#x00026;as_sdt=0,5">http://scholar.google.com/scholar?q=related:N4YMyham6pUJ:scholar.google.com/&#x00026;hl=en&#x00026;num=20&#x00026;as_sdt=0,5</ext-link></citation></ref>
<ref id="B16">
<label>16.</label>
<citation citation-type="journal"><person-group person-group-type="author"><name><surname>Coifman</surname> <given-names>RR</given-names></name> <name><surname>Lafon</surname> <given-names>S</given-names></name></person-group>. <article-title>Diffusion maps</article-title>. <source>Appl Comput Harmon Anal.</source> (<year>2006</year>) <volume>21</volume>:<fpage>5</fpage>&#x02013;<lpage>30</lpage>. <pub-id pub-id-type="doi">10.1016/j.acha.2006.04.006</pub-id></citation></ref>
<ref id="B17">
<label>17.</label>
<citation citation-type="book"><person-group person-group-type="author"><name><surname>Atkinson</surname> <given-names>K</given-names></name> <name><surname>Han</surname> <given-names>W</given-names></name></person-group>. <source>Spherical Harmonics and Approximations on the Unit Sphere: An Introduction</source>. <publisher-name>Springer Science &#x00026; Business Media</publisher-name> (<year>2012</year>). Available online at: <ext-link ext-link-type="uri" xlink:href="http://books.google.com/books?id=RnZ8SMAb6T8C&#x00026;printsec=frontcover&#x00026;dq=intitle:Spherical&#x0002B;harmonics&#x0002B;and&#x0002B;approximations&#x0002B;on&#x0002B;the&#x0002B;unit&#x0002B;sphere&#x0002B;an&#x0002B;introduction&#x00026;hl=&#x00026;cd=1&#x00026;source=gbs_api">http://books.google.com/books?id=RnZ8SMAb6T8C&#x00026;printsec=frontcover&#x00026;dq=intitle:Spherical&#x0002B;harmonics&#x0002B;and&#x0002B;approximations&#x0002B;on&#x0002B;the&#x0002B;unit&#x0002B;sphere&#x0002B;an&#x0002B;introduction&#x00026;hl=&#x00026;cd=1&#x00026;source=gbs_api</ext-link></citation></ref>
<ref id="B18">
<label>18.</label>
<citation citation-type="journal"><person-group person-group-type="author"><name><surname>Wen</surname> <given-names>ZY</given-names></name> <name><surname>Avery</surname> <given-names>J</given-names></name></person-group>. <article-title>Some properties of hyperspherical harmonics</article-title>. <source>J Math Phys.</source> (<year>1985</year>) <volume>26</volume>:<fpage>396</fpage>&#x02013;<lpage>9</lpage>. <pub-id pub-id-type="doi">10.1063/1.526621</pub-id></citation></ref>
<ref id="B19">
<label>19.</label>
<citation citation-type="book"><person-group person-group-type="author"><name><surname>Stone</surname> <given-names>M</given-names></name> <name><surname>Goldbart</surname> <given-names>P</given-names></name></person-group>. <source>Mathematics for Physics: A Guided Tour for Graduate Students</source>. <publisher-loc>Cambridge</publisher-loc>: <publisher-name>Cambridge University Press</publisher-name> (<year>2009</year>). <publisher-name>Cambridge University Press</publisher-name>. Available online at: <ext-link ext-link-type="uri" xlink:href="http://ebooks.cambridge.org/ref/id/CBO9780511627040">http://ebooks.cambridge.org/ref/id/CBO9780511627040</ext-link></citation></ref>
<ref id="B20">
<label>20.</label>
<citation citation-type="journal"><person-group person-group-type="author"><name><surname>Grigor&#x00027;yan</surname> <given-names>A</given-names></name></person-group>. <article-title>Analytic and geometric background of recurrence and non-explosion of the Brownian motion on Riemannian manifolds</article-title>. <source>Bull Am Math Soc.</source> (<year>1999</year>) <volume>36</volume>:<fpage>135</fpage>&#x02013;<lpage>249</lpage>. <pub-id pub-id-type="doi">10.1090/S0273-0979-99-00776-4</pub-id></citation>
</ref>
<ref id="B21">
<label>21.</label>
<citation citation-type="web"><person-group person-group-type="author"><name><surname>Craven</surname> <given-names>M</given-names></name> <name><surname>McCallum</surname> <given-names>A</given-names></name> <name><surname>PiPasquo</surname> <given-names>D</given-names></name> <name><surname>Mitchell</surname> <given-names>T</given-names></name></person-group>. <article-title>Learning to extract symbolic knowledge from the World Wide Web</article-title>. In: <source>Proceedings of the National Conference on Artificial Intelligence.</source> (<year>1998</year>). p. <fpage>509</fpage>&#x02013;<lpage>16</lpage>. Available online at: <ext-link ext-link-type="uri" xlink:href="http://oai.dtic.mil/oai/oai?verb=getRecord&#x00026;metadataPrefix=html&#x00026;identifier=ADA356047">http://oai.dtic.mil/oai/oai?verb=getRecord&#x00026;metadataPrefix=html&#x00026;identifier=ADA356047</ext-link></citation>
</ref>
<ref id="B22">
<label>22.</label>
<citation citation-type="journal"><person-group person-group-type="author"><name><surname>Hanahan</surname> <given-names>D</given-names></name> <name><surname>Weinberg</surname> <given-names>RA</given-names></name></person-group>. <article-title>Hallmarks of cancer: the next generation</article-title>. <source>Cell</source> (<year>2011</year>) <volume>144</volume>:<fpage>646</fpage>&#x02013;<lpage>74</lpage>. <pub-id pub-id-type="doi">10.1016/j.cell.2011.02.013</pub-id></citation>
</ref>
<ref id="B23">
<label>23.</label>
<citation citation-type="web"><person-group person-group-type="author"><name><surname>Smedley</surname> <given-names>D,. SHORT COMMUNICATION Cloning Mapping of Members of the MYM Family.</given-names></name></person-group> (<year>1999</year>). Available online at: <ext-link ext-link-type="uri" xlink:href="http://ac.els-cdn.com/S0888754399959189/1-s2.0-S0888754399959189-main.pdf?_tid=fd1183ce-9ae9-11e6-901c-00000aacb362&#x00026;acdnat=1477424224_0985e54b94244b55ec1558e901cb6211">http://ac.els-cdn.com/S0888754399959189/1-s2.0-S0888754399959189-main.pdf?_tid=fd1183ce-9ae9-11e6-901c-00000aacb362&#x00026;acdnat=1477424224_0985e54b94244b55ec1558e901cb6211</ext-link></citation>
</ref>
<ref id="B24">
<label>24.</label>
<citation citation-type="journal"><person-group person-group-type="author"><name><surname>Shchors</surname> <given-names>K</given-names></name> <name><surname>Yehiely</surname> <given-names>F</given-names></name> <name><surname>Kular</surname> <given-names>RK</given-names></name> <name><surname>Kotlo</surname> <given-names>KU</given-names></name> <name><surname>Brewer</surname> <given-names>G</given-names></name> <name><surname>Deiss</surname> <given-names>LP</given-names></name></person-group>. <article-title>Cell death inhibiting RNA (CDIR) derived from a 3&#x00027;-untranslated region binds AUF1 and heat shock protein 27</article-title>. <source>J Biol Chem.</source> (<year>2002</year>) <volume>277</volume>:<fpage>47061</fpage>&#x02013;<lpage>72</lpage>. <pub-id pub-id-type="doi">10.1074/jbc.M202272200</pub-id></citation>
</ref>
<ref id="B25">
<label>25.</label>
<citation citation-type="web"><person-group person-group-type="author"><name><surname>Zohrabian</surname> <given-names>VM</given-names></name> <name><surname>Nandu</surname> <given-names>H</given-names></name> <name><surname>Gulati</surname> <given-names>N</given-names></name></person-group>. <article-title>Gene Expression Profiling of Metastatic Brain Cancer</article-title>. Oncology Reports. (<year>2007</year>) Available online at: <ext-link ext-link-type="uri" xlink:href="https://www.spandidos-publications.com/or/18/2/321/download">https://www.spandidos-publications.com/or/18/2/321/download</ext-link></citation>
</ref>
<ref id="B26">
<label>26.</label>
<citation citation-type="journal"><person-group person-group-type="author"><name><surname>Kaur</surname> <given-names>B</given-names></name> <name><surname>Brat</surname> <given-names>DJ</given-names></name> <name><surname>Calkins</surname> <given-names>CC</given-names></name> <name><surname>Van</surname> <given-names>Meir EG</given-names></name></person-group>. <article-title>Brain angiogenesis inhibitor 1 is differentially expressed in normal brain and glioblastoma independently of p53 expression</article-title>. <source>Am J Pathol.</source> (<year>2010</year>) <volume>162</volume>:<fpage>19</fpage>&#x02013;<lpage>27</lpage>. <pub-id pub-id-type="doi">10.1016/S0002-9440(10)63794-7</pub-id></citation>
</ref>
<ref id="B27">
<label>27.</label>
<citation citation-type="journal"><person-group person-group-type="author"><name><surname>Hamann</surname> <given-names>J</given-names></name> <name><surname>Aust</surname> <given-names>G</given-names></name> <name><surname>Ara&#x000E7;</surname> <given-names>D</given-names></name> <name><surname>Engel</surname> <given-names>FB</given-names></name> <name><surname>Formstone</surname> <given-names>C</given-names></name> <name><surname>Fredriksson</surname> <given-names>R</given-names></name> <etal/></person-group>. <article-title>International union of basic and clinical pharmacology</article-title>. XCIV. adhesion G protein-coupled receptors. <source>Pharmacol Rev.</source> (<year>2015</year>) <volume>67</volume>:<fpage>338</fpage>&#x02013;<lpage>67</lpage>. <pub-id pub-id-type="doi">10.1124/pr.114.009647</pub-id></citation>
</ref>
<ref id="B28">
<label>28.</label>
<citation citation-type="journal"><person-group person-group-type="author"><name><surname>Yamashita</surname> <given-names>S</given-names></name> <name><surname>Fujii</surname> <given-names>K</given-names></name> <name><surname>Zhao</surname> <given-names>C</given-names></name> <name><surname>Takagi</surname> <given-names>H</given-names></name> <name><surname>Katakura</surname> <given-names>Y</given-names></name></person-group>. <article-title>Involvement of the NFX1-repressor complex in PKC-&#x003B4;-induced repression of hTERT transcription</article-title>. <source>J Biochem.</source> (<year>2016</year>) <volume>160</volume>:<fpage>309</fpage>&#x02013;<lpage>13</lpage>. <pub-id pub-id-type="doi">10.1093/jb/mvw038</pub-id></citation>
</ref>
<ref id="B29">
<label>29.</label>
<citation citation-type="journal"><person-group person-group-type="author"><name><surname>Song</surname> <given-names>Z</given-names></name> <name><surname>Krishna</surname> <given-names>S</given-names></name> <name><surname>Thanos</surname> <given-names>D</given-names></name> <name><surname>Strominger</surname> <given-names>JL</given-names></name> <name><surname>Ono</surname> <given-names>SJ</given-names></name></person-group>. <article-title>A novel cysteine-rich sequence-specific DNA-binding protein interacts with the conserved X-box motif of the human major histocompatibility complex class II genes via a repeated Cys-His domain and functions as a transcriptional repressor</article-title>. <source>J Exp Med.</source> (<year>1994</year>) <volume>180</volume>:<fpage>1763</fpage>&#x02013;<lpage>74</lpage>. <pub-id pub-id-type="doi">10.1084/jem.180.5.1763</pub-id></citation>
</ref>
<ref id="B30">
<label>30.</label>
<citation citation-type="journal"><person-group person-group-type="author"><name><surname>Adinolfi</surname> <given-names>E</given-names></name> <name><surname>Capece</surname> <given-names>M</given-names></name> <name><surname>Franceschini</surname> <given-names>A</given-names></name> <name><surname>Falzoni</surname> <given-names>S</given-names></name></person-group>. <article-title>Accelerated tumor progression in mice lacking the ATP receptor P2X7</article-title>. <source>Cancer Res.</source> (<year>2015</year>) <volume>75</volume>:<fpage>635</fpage>&#x02013;<lpage>44</lpage>. <pub-id pub-id-type="doi">10.1158/0008-5472.CAN-14-1259</pub-id></citation>
</ref>
<ref id="B31">
<label>31.</label>
<citation citation-type="journal"><person-group person-group-type="author"><name><surname>G&#x000F3;mez-Villafuertes</surname> <given-names>R</given-names></name> <name><surname>Garc&#x000ED;a-Huerta</surname> <given-names>P</given-names></name> <name><surname>D&#x000ED;az-Hern&#x000E1;ndez</surname> <given-names>JI</given-names></name> <name><surname>Miras-Portugal</surname> <given-names>MT</given-names></name></person-group>. <article-title>PI3K/Akt signaling pathway triggers P2X7 receptor expression as a pro-survival factor of neuroblastoma cells under limiting growth conditions</article-title>. <source>Nat Publ Group</source> (<year>2015</year>) <volume>5</volume>:<fpage>1</fpage>&#x02013;<lpage>15</lpage>. <pub-id pub-id-type="doi">10.1038/srep18417</pub-id></citation>
</ref>
<ref id="B32">
<label>32.</label>
<citation citation-type="journal"><person-group person-group-type="author"><name><surname>Li&#x000F1;&#x000E1;n-Rico</surname> <given-names>A</given-names></name> <name><surname>Turco</surname> <given-names>F</given-names></name> <name><surname>Ochoa-Cortes</surname> <given-names>F</given-names></name> <name><surname>Harzman</surname> <given-names>A</given-names></name> <name><surname>Needleman</surname> <given-names>BJ</given-names></name> <name><surname>Arsenescu</surname> <given-names>R</given-names></name> <etal/></person-group>. <article-title>Molecular signaling and dysfunction of the human reactive enteric glial cell phenotype</article-title>. <source>Inflam Bowel Dis.</source> (<year>2016</year>) <volume>22</volume>:<fpage>1812</fpage>&#x02013;<lpage>34</lpage>. <pub-id pub-id-type="doi">10.1097/MIB.0000000000000854</pub-id></citation>
</ref>
<ref id="B33">
<label>33.</label>
<citation citation-type="journal"><person-group person-group-type="author"><name><surname>Anderton</surname> <given-names>JA</given-names></name> <name><surname>Lindsey</surname> <given-names>JC</given-names></name></person-group>. <article-title>Global analysis of the medulloblastoma epigenome identifies disease-subgroup-specific inactivation of <italic>COL1A2</italic></article-title>. <source>Neuro Oncol.</source> (<year>2008</year>) <volume>10</volume>:<fpage>981</fpage>&#x02013;<lpage>94</lpage>. <pub-id pub-id-type="doi">10.1215/15228517-2008-048</pub-id></citation>
</ref>
<ref id="B34">
<label>34.</label>
<citation citation-type="journal"><person-group person-group-type="author"><name><surname>Liang</surname> <given-names>Y</given-names></name> <name><surname>Diehn</surname> <given-names>M</given-names></name> <name><surname>Bollen</surname> <given-names>AW</given-names></name> <name><surname>Israel</surname> <given-names>MA</given-names></name> <name><surname>Gupta</surname> <given-names>N</given-names></name></person-group>. <article-title>Type I collagen is overexpressed in medulloblastoma as a component of tumor microenvironment</article-title>. <source>J Neuro Oncol.</source> (<year>2007</year>) <volume>86</volume>:<fpage>133</fpage>&#x02013;<lpage>41</lpage>. <pub-id pub-id-type="doi">10.1007/s11060-007-9457-5</pub-id></citation>
</ref>
<ref id="B35">
<label>35.</label>
<citation citation-type="journal"><person-group person-group-type="author"><name><surname>Schwalbe</surname> <given-names>EC</given-names></name> <name><surname>Lindsey</surname> <given-names>JC</given-names></name> <name><surname>Straughton</surname> <given-names>D</given-names></name> <name><surname>Hogg</surname> <given-names>TL</given-names></name> <name><surname>Cole</surname> <given-names>M</given-names></name> <name><surname>Megahed</surname> <given-names>H</given-names></name> <etal/></person-group>. <article-title>Rapid diagnosis of medulloblastoma molecular subgroups</article-title>. <source>Clin Cancer Res.</source> (<year>2011</year>) <volume>17</volume>:<fpage>1883</fpage>&#x02013;<lpage>1894</lpage>. <pub-id pub-id-type="doi">10.1158/1078-0432.CCR-10-2210</pub-id></citation>
</ref>
<ref id="B36">
<label>36.</label>
<citation citation-type="book"><person-group person-group-type="author"><name><surname>Vapnik</surname> <given-names>VN</given-names></name></person-group>. <source>The Nature of Statistical Learning Theory</source>. <publisher-name>Springer Science &#x00026; Business Media</publisher-name> (<year>2013</year>). Available online at: <ext-link ext-link-type="uri" xlink:href="http://books.google.com/books?id=EoDSBwAAQBAJ&#x00026;printsec=frontcover&#x00026;dq=intitle:The&#x0002B;nature&#x0002B;of&#x0002B;statistical&#x0002B;learning&#x0002B;theory&#x00026;hl=&#x00026;cd=1&#x00026;source=gbs_api">http://books.google.com/books?id=EoDSBwAAQBAJ&#x00026;printsec=frontcover&#x00026;dq=intitle:The&#x0002B;nature&#x0002B;of&#x0002B;statistical&#x0002B;learning&#x0002B;theory&#x00026;hl=&#x00026;cd=1&#x00026;source=gbs_api</ext-link></citation>
</ref>
<ref id="B37">
<label>37.</label>
<citation citation-type="journal"><person-group person-group-type="author"><name><surname>Vapnik</surname> <given-names>V</given-names></name> <name><surname>Levin</surname> <given-names>E</given-names></name> <name><surname>Le</surname> <given-names>Cun Y</given-names></name></person-group>. <article-title>Measuring the VC-dimension of a learning machine</article-title>. <source>Neural Comput.</source> (<year>1994</year>) <volume>6</volume>:<fpage>851</fpage>&#x02013;<lpage>76</lpage>. <pub-id pub-id-type="doi">10.1162/neco.1994.6.5.851</pub-id></citation>
</ref>
<ref id="B38">
<label>38.</label>
<citation citation-type="book"><person-group person-group-type="author"><name><surname>Paliouras</surname> <given-names>G</given-names></name> <name><surname>Karkaletsis</surname> <given-names>V</given-names></name> <name><surname>Spyropoulos</surname> <given-names>CD</given-names></name></person-group>. <source>Machine Learning and Its Applications</source>. <publisher-name>Springer</publisher-name> (<year>2003</year>). Available online at: <ext-link ext-link-type="uri" xlink:href="http://books.google.com/books?id=pm5rCQAAQBAJ&#x00026;pg=PR4&#x00026;dq=intitle:Machine&#x0002B;learning&#x0002B;and&#x0002B;its&#x0002B;applications&#x0002B;advanced&#x0002B;lectures&#x00026;hl=&#x00026;cd=1&#x00026;source=gbs_api">http://books.google.com/books?id=pm5rCQAAQBAJ&#x00026;pg=PR4&#x00026;dq=intitle:Machine&#x0002B;learning&#x0002B;and&#x0002B;its&#x0002B;applications&#x0002B;advanced&#x0002B;lectures&#x00026;hl=&#x00026;cd=1&#x00026;source=gbs_api</ext-link></citation>
</ref>
</ref-list>
<fn-group>
<fn fn-type="financial-disclosure"><p><bold>Funding.</bold> This research was supported by a Distinguished Scientist Award from Sontag Foundation and the Grainger Engineering Breakthroughs Initiative.</p>
</fn>
</fn-group>
</back>
</article>
