<?xml version="1.0" encoding="UTF-8"?>
<!DOCTYPE article PUBLIC "-//NLM//DTD Journal Publishing DTD v2.3 20070202//EN" "journalpublishing.dtd">
<article article-type="research-article" dtd-version="2.3" xml:lang="EN" xmlns:mml="http://www.w3.org/1998/Math/MathML" xmlns:xlink="http://www.w3.org/1999/xlink">
<front>
<journal-meta>
<journal-id journal-id-type="publisher-id">Front. Appl. Math. Stat.</journal-id>
<journal-title>Frontiers in Applied Mathematics and Statistics</journal-title>
<abbrev-journal-title abbrev-type="pubmed">Front. Appl. Math. Stat.</abbrev-journal-title>
<issn pub-type="epub">2297-4687</issn>
<publisher>
<publisher-name>Frontiers Media S.A.</publisher-name>
</publisher>
</journal-meta>
<article-meta>
<article-id pub-id-type="publisher-id">787455</article-id>
<article-id pub-id-type="doi">10.3389/fams.2021.787455</article-id>
<article-categories>
<subj-group subj-group-type="heading">
<subject>Applied Mathematics and Statistics</subject>
<subj-group>
<subject>Original Research</subject>
</subj-group>
</subj-group>
</article-categories>
<title-group>
<article-title>Kernel Distance Measures for Time Series, Random Fields and Other Structured Data</article-title>
<alt-title alt-title-type="left-running-head">Das et&#x20;al.</alt-title>
<alt-title alt-title-type="right-running-head">Kernel Distance Measure</alt-title>
</title-group>
<contrib-group>
<contrib contrib-type="author" corresp="yes">
<name>
<surname>Das</surname>
<given-names>Srinjoy</given-names>
</name>
<xref ref-type="aff" rid="aff1">
<sup>1</sup>
</xref>
<xref ref-type="corresp" rid="c001">&#x2a;</xref>
<uri xlink:href="https://loop.frontiersin.org/people/1554657/overview"/>
</contrib>
<contrib contrib-type="author">
<name>
<surname>Mhaskar</surname>
<given-names>Hrushikesh N.</given-names>
</name>
<xref ref-type="aff" rid="aff2">
<sup>2</sup>
</xref>
<uri xlink:href="https://loop.frontiersin.org/people/281015/overview"/>
</contrib>
<contrib contrib-type="author">
<name>
<surname>Cloninger</surname>
<given-names>Alexander</given-names>
</name>
<xref ref-type="aff" rid="aff3">
<sup>3</sup>
</xref>
<uri xlink:href="https://loop.frontiersin.org/people/933579/overview"/>
</contrib>
</contrib-group>
<aff id="aff1">
<sup>1</sup>
<institution>School of Mathematical and Data Sciences, West Virginia University</institution>, <addr-line>Morgantown</addr-line>, <addr-line>WV</addr-line>, <country>United&#x20;States</country>
</aff>
<aff id="aff2">
<sup>2</sup>
<institution>Institute of Mathematical Sciences, Claremont Graduate University</institution>, <addr-line>Claremont</addr-line>, <addr-line>CA</addr-line>, <country>United&#x20;States</country>
</aff>
<aff id="aff3">
<sup>3</sup>
<institution>Department of Mathematics and Hagi&#x11f;olou Data Science Institute, University of California, San Diego</institution>, <addr-line>San Diego</addr-line>, <addr-line>CA</addr-line>, <country>United&#x20;States</country>
</aff>
<author-notes>
<fn fn-type="edited-by">
<p>
<bold>Edited by:</bold> <ext-link ext-link-type="uri" xlink:href="https://loop.frontiersin.org/people/358028/overview">Yiming Ying</ext-link>, University at Albany, United&#x20;States</p>
</fn>
<fn fn-type="edited-by">
<p>
<bold>Reviewed by:</bold> <ext-link ext-link-type="uri" xlink:href="https://loop.frontiersin.org/people/369818/overview">Xin Guo</ext-link>, Hong Kong Polytechnic University, Hong Kong SAR, China</p>
<p>
<ext-link ext-link-type="uri" xlink:href="https://loop.frontiersin.org/people/501483/overview">Yunwen Lei</ext-link>, Southern University of Science and Technology, China</p>
</fn>
<corresp id="c001">&#x2a;Correspondence: Srinjoy Das, <email>srinjoy.das@mail.wvu.edu</email>
</corresp>
<fn fn-type="other">
<p>This article was submitted to Mathematics of Computation and Data Science, a section of the journal Frontiers in Applied Mathematics and Statistics</p>
</fn>
</author-notes>
<pub-date pub-type="epub">
<day>22</day>
<month>12</month>
<year>2021</year>
</pub-date>
<pub-date pub-type="collection">
<year>2021</year>
</pub-date>
<volume>7</volume>
<elocation-id>787455</elocation-id>
<history>
<date date-type="received">
<day>30</day>
<month>09</month>
<year>2021</year>
</date>
<date date-type="accepted">
<day>16</day>
<month>11</month>
<year>2021</year>
</date>
</history>
<permissions>
<copyright-statement>Copyright &#xa9; 2021 Das, Mhaskar and Cloninger.</copyright-statement>
<copyright-year>2021</copyright-year>
<copyright-holder>Das, Mhaskar and Cloninger</copyright-holder>
<license xlink:href="http://creativecommons.org/licenses/by/4.0/">
<p>This is an open-access article distributed under the terms of the Creative Commons Attribution License (CC BY). The use, distribution or reproduction in other forums is permitted, provided the original author(s) and the copyright owner(s) are credited and that the original publication in this journal is cited, in accordance with accepted academic practice. No use, distribution or reproduction is permitted which does not comply with these&#x20;terms.</p>
</license>
</permissions>
<abstract>
<p>This paper introduces kdiff, a novel kernel-based measure for estimating distances between instances of time series, random fields and other forms of structured data. This measure is based on the idea of matching distributions that only overlap over a portion of their region of support. Our proposed measure is inspired by MPdist which has been previously proposed for such datasets and is constructed using Euclidean metrics, whereas kdiff is constructed using non-linear kernel distances. Also, kdiff accounts for both self and cross similarities across the instances and is defined using a lower quantile of the distance distribution. Comparing the cross similarity to self similarity allows for measures of similarity that are more robust to noise and partial occlusions of the relevant signals. Our proposed measure kdiff is a more general form of the well known kernel-based Maximum Mean Discrepancy distance estimated over the embeddings. Some theoretical results are provided for separability conditions using kdiff as a distance measure for clustering and classification problems where the embedding distributions can be modeled as two component mixtures. Applications are demonstrated for clustering of synthetic and real-life time series and image data, and the performance of kdiff is compared to competing distance measures for clustering.</p>
</abstract>
<kwd-group>
<kwd>Kernels</kwd>
<kwd>time series</kwd>
<kwd>statistical distances</kwd>
<kwd>random fields</kwd>
<kwd>Motif detection</kwd>
</kwd-group>
<contract-num rid="cn001">1819222 2012266&#x20;2012355</contract-num>
<contract-num rid="cn002">2196</contract-num>
<contract-num rid="cn003">W911NF2110218</contract-num>
<contract-sponsor id="cn001">National Science Foundation10.13039/100000001</contract-sponsor>
<contract-sponsor id="cn002">Russell Sage Foundation10.13039/100000935</contract-sponsor>
<contract-sponsor id="cn003">Army Research Office10.13039/100000183</contract-sponsor>
<contract-sponsor id="cn004">Intel Corporation10.13039/100002418</contract-sponsor>
</article-meta>
</front>
<body>
<sec id="s1">
<title>1 Introduction and Motivation</title>
<p>Clustering and classification tasks in applications such as time series and image processing are critically dependent on the distance measure used to identify similarities in the available data. Distance measures are named as such because they may not satisfy all assumptions of a metric (e.g., they may induce equivalence classes of time series that are distance zero from one another, they may not satisfy the triangle inequality). In such contexts, several distance measures have been proposed in the literature:<list list-type="simple">
<list-item>
<p>&#x2022; Point-to-point matching e.g. Euclidean distance or Dynamic Time Warping distance [<xref ref-type="bibr" rid="B1">1</xref>,&#x20;<xref ref-type="bibr" rid="B2">2</xref>]</p>
</list-item>
<list-item>
<p>&#x2022; Matching features of the series e.g. autocorrelation coefficients [<xref ref-type="bibr" rid="B3">3</xref>], Pearson correlation coefficients [<xref ref-type="bibr" rid="B4">4</xref>], periodograms [<xref ref-type="bibr" rid="B5">5</xref>], extreme value behavior&#x20;[<xref ref-type="bibr" rid="B6">6</xref>]</p>
</list-item>
<list-item>
<p>&#x2022; Number of matching subsequences in the series&#x20;[<xref ref-type="bibr" rid="B7">7</xref>]</p>
</list-item>
<list-item>
<p>&#x2022; Similarity of embedding distributions of the series&#x20;[<xref ref-type="bibr" rid="B8">8</xref>]</p>
</list-item>
</list>
</p>
<p>In this paper we consider distance measures for applications involving clustering, classification and related data mining tasks in time series, random fields and other forms of possibly non i. i.d data. In particular, we focus on problems where membership in a specific class is characterized by instances matching only over a portion of their region of support. In addition, the regions where such feature matching occurs may not be overlapping in time, or on the underlying grid of the random field. Distance measures must take these data characteristics into consideration when determining similarity in such applications. Previously MPdist has been proposed as a distance measure for such time series datasets [<xref ref-type="bibr" rid="B7">7</xref>] which match only over part of their region of support and is constructed using Euclidean metrics. Inspired by MPdist, we propose a new kernel-based distance measure kdiff for estimating distances between instances of such univariate and multivariate time series, random field and other types of structured&#x20;data.</p>
<p>For constructing kdiff, we first create sliding window based embeddings over the given time series or random fields. We then estimate a distance distribution by using a kernel-based distance measure between such embeddings over given pairs of data instances. Finally the distance measure used in clustering, classification and related tasks is defined by a pre-specified lower quantile of this distance distribution. This kernel-based distance measure based on such embeddings can also be constructed using the Reproducing Kernel Hilbert Space (RKHS) based Maximum Mean Discrepancy (MMD) previously discussed in [<xref ref-type="bibr" rid="B9">9</xref>]. Our kernel based measure kdiff can be considered as a more general distance compared to RKHS MMD for applications where class instances match only over a part of the region of support. More details about the connections between kdiff and RKHS MMD are provided later in the paper. We also note that the kernel construction in kdiff allows for data-dependent kernel construction similar to MMD [<xref ref-type="bibr" rid="B10">10</xref>&#x2013;<xref ref-type="bibr" rid="B12">12</xref>], though we focus on isotropic localized kernels in this work and compare to standard&#x20;MMD.</p>
<p>The rest of the paper is organized as follows. <xref ref-type="sec" rid="s2">Section 2</xref> outlines the main idea and motivation behind the construction of our distance measure kdiff. <xref ref-type="sec" rid="s2">Section 2</xref> also outlines some theoretical results for separability of data using kdiff as a distance measure for clustering, classification and related tasks by modeling the embedding distributions derived from the original data as two component mixtures. <xref ref-type="sec" rid="s3">Section 3</xref> outlines some practical considerations and data-driven strategies to determine optimal parameters for the algorithm to estimate kdiff. <xref ref-type="sec" rid="s4">Section 4</xref> presents simulation results using kdiff on both synthetic and real-life datasets and compares them with existing methods. Finally <xref ref-type="sec" rid="s5">Section 5</xref> outlines some conclusions and directions for future&#x20;work.</p>
</sec>
<sec id="s2">
<title>2 Main Idea</title>
<sec id="s2-1">
<title>2.1 Overview</title>
<p>Consider two real-valued datasets <inline-formula id="inf1">
<mml:math id="m1">
<mml:mrow>
<mml:mo stretchy="false">{</mml:mo>
<mml:mrow>
<mml:msub>
<mml:mrow>
<mml:mi mathvariant="bold">X</mml:mi>
</mml:mrow>
<mml:mrow>
<mml:munder accentunder="false">
<mml:mrow>
<mml:mi>t</mml:mi>
</mml:mrow>
<mml:mo accent="true">&#x332;</mml:mo>
</mml:munder>
</mml:mrow>
</mml:msub>
<mml:mo>,</mml:mo>
<mml:msub>
<mml:mrow>
<mml:mi mathvariant="bold">Y</mml:mi>
</mml:mrow>
<mml:mrow>
<mml:munder accentunder="false">
<mml:mrow>
<mml:mi>t</mml:mi>
</mml:mrow>
<mml:mo accent="true">&#x332;</mml:mo>
</mml:munder>
</mml:mrow>
</mml:msub>
<mml:mo>:</mml:mo>
<mml:munder accentunder="false">
<mml:mrow>
<mml:mi>t</mml:mi>
</mml:mrow>
<mml:mo accent="true">&#x332;</mml:mo>
</mml:munder>
<mml:mo>&#x2208;</mml:mo>
<mml:msup>
<mml:mrow>
<mml:mi mathvariant="double-struck">Z</mml:mi>
</mml:mrow>
<mml:mrow>
<mml:mi>k</mml:mi>
</mml:mrow>
</mml:msup>
</mml:mrow>
<mml:mo stretchy="false">}</mml:mo>
</mml:mrow>
</mml:math>
</inline-formula> defined over a <italic>k</italic>-dimensional index set. These may in general be vector-valued random variables, and therefore <inline-formula id="inf2">
<mml:math id="m2">
<mml:msub>
<mml:mrow>
<mml:mi mathvariant="bold">X</mml:mi>
</mml:mrow>
<mml:mrow>
<mml:munder accentunder="false">
<mml:mrow>
<mml:mi>t</mml:mi>
</mml:mrow>
<mml:mo accent="true">&#x332;</mml:mo>
</mml:munder>
</mml:mrow>
</mml:msub>
</mml:math>
</inline-formula> and <inline-formula id="inf3">
<mml:math id="m3">
<mml:msub>
<mml:mrow>
<mml:mi mathvariant="bold">Y</mml:mi>
</mml:mrow>
<mml:mrow>
<mml:munder accentunder="false">
<mml:mrow>
<mml:mi>t</mml:mi>
</mml:mrow>
<mml:mo accent="true">&#x332;</mml:mo>
</mml:munder>
</mml:mrow>
</mml:msub>
</mml:math>
</inline-formula> can be considered as either univariate or multivariate time series, random fields or other types of structured data. Our problem of interest is where instances of <inline-formula id="inf4">
<mml:math id="m4">
<mml:msub>
<mml:mrow>
<mml:mi mathvariant="bold">X</mml:mi>
</mml:mrow>
<mml:mrow>
<mml:munder accentunder="false">
<mml:mrow>
<mml:mi>t</mml:mi>
</mml:mrow>
<mml:mo accent="true">&#x332;</mml:mo>
</mml:munder>
</mml:mrow>
</mml:msub>
</mml:math>
</inline-formula> and <inline-formula id="inf5">
<mml:math id="m5">
<mml:msub>
<mml:mrow>
<mml:mi mathvariant="bold">Y</mml:mi>
</mml:mrow>
<mml:mrow>
<mml:munder accentunder="false">
<mml:mrow>
<mml:mi>t</mml:mi>
</mml:mrow>
<mml:mo accent="true">&#x332;</mml:mo>
</mml:munder>
</mml:mrow>
</mml:msub>
</mml:math>
</inline-formula> match with certain localized motifs <inline-formula id="inf6">
<mml:math id="m6">
<mml:mrow>
<mml:mo stretchy="false">{</mml:mo>
<mml:mrow>
<mml:msub>
<mml:mrow>
<mml:mi mathvariant="bold">X</mml:mi>
</mml:mrow>
<mml:mrow>
<mml:munder accentunder="false">
<mml:mrow>
<mml:mi>t</mml:mi>
</mml:mrow>
<mml:mo accent="true">&#x332;</mml:mo>
</mml:munder>
</mml:mrow>
</mml:msub>
<mml:mo>:</mml:mo>
<mml:mi>t</mml:mi>
<mml:mo>&#x2208;</mml:mo>
<mml:mi>S</mml:mi>
</mml:mrow>
<mml:mo stretchy="false">}</mml:mo>
</mml:mrow>
<mml:mo>&#x2248;</mml:mo>
<mml:mrow>
<mml:mo stretchy="false">{</mml:mo>
<mml:mrow>
<mml:msub>
<mml:mrow>
<mml:mi mathvariant="bold">Y</mml:mi>
</mml:mrow>
<mml:mrow>
<mml:munder accentunder="false">
<mml:mrow>
<mml:mi>t</mml:mi>
</mml:mrow>
<mml:mo accent="true">&#x332;</mml:mo>
</mml:munder>
</mml:mrow>
</mml:msub>
<mml:mo>:</mml:mo>
<mml:mi>t</mml:mi>
<mml:mo>&#x2208;</mml:mo>
<mml:msup>
<mml:mrow>
<mml:mi>S</mml:mi>
</mml:mrow>
<mml:mrow>
<mml:mo>&#x2032;</mml:mo>
</mml:mrow>
</mml:msup>
</mml:mrow>
<mml:mo stretchy="false">}</mml:mo>
</mml:mrow>
</mml:math>
</inline-formula> for small localized index sets <inline-formula id="inf7">
<mml:math id="m7">
<mml:mi>S</mml:mi>
<mml:mo>,</mml:mo>
<mml:msup>
<mml:mrow>
<mml:mi>S</mml:mi>
</mml:mrow>
<mml:mrow>
<mml:mo>&#x2032;</mml:mo>
</mml:mrow>
</mml:msup>
<mml:mo>&#x2282;</mml:mo>
<mml:msup>
<mml:mrow>
<mml:mi mathvariant="double-struck">Z</mml:mi>
</mml:mrow>
<mml:mrow>
<mml:mi>k</mml:mi>
</mml:mrow>
</mml:msup>
</mml:math>
</inline-formula>. For both the univariate and multivariate cases, we can embed these data sets into some corresponding point clouds <inline-formula id="inf8">
<mml:math id="m8">
<mml:mi mathvariant="bold">X</mml:mi>
<mml:mo>,</mml:mo>
<mml:mi mathvariant="bold">Y</mml:mi>
<mml:mo>&#x2282;</mml:mo>
<mml:msup>
<mml:mrow>
<mml:mi mathvariant="double-struck">R</mml:mi>
</mml:mrow>
<mml:mrow>
<mml:mi>L</mml:mi>
</mml:mrow>
</mml:msup>
</mml:math>
</inline-formula> via windowing with a size <italic>L</italic> window, where <italic>L</italic> can be determined from training or some other appropriate technique [<xref ref-type="bibr" rid="B8">8</xref>, <xref ref-type="bibr" rid="B13">13</xref>]. Once we have a window embedding of these data sets, we can define various distance measures on the resulting point clouds to define similarity between <inline-formula id="inf9">
<mml:math id="m9">
<mml:msub>
<mml:mrow>
<mml:mi mathvariant="bold">X</mml:mi>
</mml:mrow>
<mml:mrow>
<mml:munder accentunder="false">
<mml:mrow>
<mml:mi>t</mml:mi>
</mml:mrow>
<mml:mo accent="true">&#x332;</mml:mo>
</mml:munder>
</mml:mrow>
</mml:msub>
</mml:math>
</inline-formula> and <inline-formula id="inf10">
<mml:math id="m10">
<mml:msub>
<mml:mrow>
<mml:mi mathvariant="bold">Y</mml:mi>
</mml:mrow>
<mml:mrow>
<mml:munder accentunder="false">
<mml:mrow>
<mml:mi>t</mml:mi>
</mml:mrow>
<mml:mo accent="true">&#x332;</mml:mo>
</mml:munder>
</mml:mrow>
</mml:msub>
</mml:math>
</inline-formula>.</p>
<p>A distance measure that has been proposed previously to determine similarity between two such time series embedded point clouds constructed over <inline-formula id="inf11">
<mml:math id="m11">
<mml:msup>
<mml:mrow>
<mml:mi mathvariant="double-struck">R</mml:mi>
</mml:mrow>
<mml:mrow>
<mml:mi>L</mml:mi>
</mml:mrow>
</mml:msup>
</mml:math>
</inline-formula> is MPdist [<xref ref-type="bibr" rid="B7">7</xref>]. In this case, a cross-data distance measure, denoted <italic>D</italic>
<sup>2</sup>, can be constructed by using 1-nearest neighbor Euclidean distances between point clouds <bold>X</bold> and <bold>Y</bold> as below:<disp-formula id="e1">
<mml:math id="m12">
<mml:mi>d</mml:mi>
<mml:mfenced open="(" close=")">
<mml:mrow>
<mml:mi>x</mml:mi>
</mml:mrow>
</mml:mfenced>
<mml:mo>&#x3d;</mml:mo>
<mml:msub>
<mml:mrow>
<mml:mi>inf</mml:mi>
</mml:mrow>
<mml:mrow>
<mml:mi>y</mml:mi>
<mml:mo>&#x2208;</mml:mo>
<mml:mi mathvariant="bold">Y</mml:mi>
</mml:mrow>
</mml:msub>
<mml:mo stretchy="false">&#x2016;</mml:mo>
<mml:mi>x</mml:mi>
<mml:mo>&#x2212;</mml:mo>
<mml:mi>y</mml:mi>
<mml:mo stretchy="false">&#x2016;</mml:mo>
<mml:mo>,</mml:mo>
<mml:mspace width="0.3333em"/>
<mml:mspace width="0.3333em"/>
<mml:mo>&#x2200;</mml:mo>
<mml:mi>x</mml:mi>
<mml:mo>&#x2208;</mml:mo>
<mml:mi mathvariant="bold">X</mml:mi>
<mml:mi>d</mml:mi>
<mml:mfenced open="(" close=")">
<mml:mrow>
<mml:mi>y</mml:mi>
</mml:mrow>
</mml:mfenced>
<mml:mo>&#x3d;</mml:mo>
<mml:msub>
<mml:mrow>
<mml:mi>inf</mml:mi>
</mml:mrow>
<mml:mrow>
<mml:mi>x</mml:mi>
<mml:mo>&#x2208;</mml:mo>
<mml:mi mathvariant="bold">X</mml:mi>
</mml:mrow>
</mml:msub>
<mml:mo stretchy="false">&#x2016;</mml:mo>
<mml:mi>x</mml:mi>
<mml:mo>&#x2212;</mml:mo>
<mml:mi>y</mml:mi>
<mml:mo stretchy="false">&#x2016;</mml:mo>
<mml:mo>,</mml:mo>
<mml:mspace width="0.3333em"/>
<mml:mspace width="0.3333em"/>
<mml:mo>&#x2200;</mml:mo>
<mml:mi>y</mml:mi>
<mml:mo>&#x2208;</mml:mo>
<mml:mi mathvariant="bold">Y</mml:mi>
<mml:msup>
<mml:mrow>
<mml:mi>D</mml:mi>
</mml:mrow>
<mml:mrow>
<mml:mn>2</mml:mn>
</mml:mrow>
</mml:msup>
<mml:mo>&#x3d;</mml:mo>
<mml:mfenced open="{" close="}">
<mml:mrow>
<mml:msup>
<mml:mrow>
<mml:mi>d</mml:mi>
</mml:mrow>
<mml:mrow>
<mml:mn>2</mml:mn>
</mml:mrow>
</mml:msup>
<mml:mfenced open="(" close=")">
<mml:mrow>
<mml:mi>z</mml:mi>
</mml:mrow>
</mml:mfenced>
<mml:mo>:</mml:mo>
<mml:mi>z</mml:mi>
<mml:mo>&#x2208;</mml:mo>
<mml:mi mathvariant="bold">X</mml:mi>
<mml:mo>&#x222a;</mml:mo>
<mml:mi mathvariant="bold">Y</mml:mi>
</mml:mrow>
</mml:mfenced>
<mml:mo>.</mml:mo>
</mml:math>
<label>(1)</label>
</disp-formula>
</p>
<p>In [<xref ref-type="bibr" rid="B7">7</xref>], the distance measure MPdist was estimated for univariate time series by choosing the <italic>k</italic>th smallest element in the set <italic>D</italic>
<sup>2</sup>. In general, MPdist can be constructed using a lower quantile of the distance distribution&#x20;<italic>D</italic>
<sup>2</sup>.</p>
<p>Our proposed distance measure kdiff generalizes MPdist using a kernel-based construction, and by considering both cross-similarity and self-similarity. Similar to MPdist, we first construct sliding window based embeddings over the original data instances <inline-formula id="inf12">
<mml:math id="m15">
<mml:msub>
<mml:mrow>
<mml:mi mathvariant="bold">X</mml:mi>
</mml:mrow>
<mml:mrow>
<mml:munder accentunder="false">
<mml:mrow>
<mml:mi>t</mml:mi>
</mml:mrow>
<mml:mo accent="true">&#x332;</mml:mo>
</mml:munder>
</mml:mrow>
</mml:msub>
<mml:mo>,</mml:mo>
<mml:mspace width="0.3333em"/>
<mml:msub>
<mml:mrow>
<mml:mi mathvariant="bold">Y</mml:mi>
</mml:mrow>
<mml:mrow>
<mml:munder accentunder="false">
<mml:mrow>
<mml:mi>t</mml:mi>
</mml:mrow>
<mml:mo accent="true">&#x332;</mml:mo>
</mml:munder>
</mml:mrow>
</mml:msub>
</mml:math>
</inline-formula> and obtain corresponding point clouds <bold>X</bold> and <bold>Y</bold>. For MPdist the final distance is estimated based on cross-similarity between the embeddings <bold>X</bold>, <bold>Y</bold> as shown in <xref ref-type="disp-formula" rid="e1">Eq. 1</xref>. Our distance measure kdiff differs from this in two ways:<list list-type="simple">
<list-item>
<p>&#x2022; We use a kernel based similarity measure over the obtained sliding window based embeddings <bold>X</bold> and <bold>Y</bold> for kdiff instead of the Euclidean metric used in MPdist<bold>.</bold>
</p>
</list-item>
<list-item>
<p>&#x2022; For kdiff the final distance is estimated based on both self and cross-similarities between the embeddings <bold>X</bold>, <bold>Y</bold> respectively. The inclusion of self-similarity in the construction of kdiff as compared to only cross-similarity for MPdist leads to better clustering performance for data with reduced signal-to-noise ratio of the matching region versus the background. This is demonstrated empirically for both synthetic and real-life data in <xref ref-type="sec" rid="s4">Section&#x20;4</xref>.</p>
</list-item>
</list>
</p>
</sec>
<sec id="s2-2">
<title>2.2 The Construction of kdiff</title>
<p>To define our kdff statistic, we will begin with a discussion of general distributions defined on <inline-formula id="inf13">
<mml:math id="m16">
<mml:msup>
<mml:mrow>
<mml:mi mathvariant="double-struck">R</mml:mi>
</mml:mrow>
<mml:mrow>
<mml:mi>L</mml:mi>
</mml:mrow>
</mml:msup>
</mml:math>
</inline-formula>. For the purposes of this paper, these can be assumed to be the distributions that the finite samples <bold>X</bold> are drawn from (in a non-iid fashion) and stitched together to form the time series <inline-formula id="inf14">
<mml:math id="m17">
<mml:msub>
<mml:mrow>
<mml:mi mathvariant="bold">X</mml:mi>
</mml:mrow>
<mml:mrow>
<mml:munder accentunder="false">
<mml:mrow>
<mml:mi>t</mml:mi>
</mml:mrow>
<mml:mo accent="true">&#x332;</mml:mo>
</mml:munder>
</mml:mrow>
</mml:msub>
</mml:math>
</inline-formula> (respectively for <bold>Y</bold> and <inline-formula id="inf15">
<mml:math id="m18">
<mml:msub>
<mml:mrow>
<mml:mi mathvariant="bold">Y</mml:mi>
</mml:mrow>
<mml:mrow>
<mml:munder accentunder="false">
<mml:mrow>
<mml:mi>t</mml:mi>
</mml:mrow>
<mml:mo accent="true">&#x332;</mml:mo>
</mml:munder>
</mml:mrow>
</mml:msub>
</mml:math>
</inline-formula>).</p>
<p>In general, we can define the distributions on <inline-formula id="inf16">
<mml:math id="m19">
<mml:mi mathvariant="double-struck">X</mml:mi>
</mml:math>
</inline-formula>, which is a locally compact metric measure space with the metric <italic>&#x3c1;</italic> and a distinguished probability measure <italic>&#x3bd;</italic>&#x2a;. The term measure will denote a signed Borel measure. We introduce a fixed, positive definite kernel <inline-formula id="inf17">
<mml:math id="m20">
<mml:mi>K</mml:mi>
<mml:mo>:</mml:mo>
<mml:mi mathvariant="double-struck">X</mml:mi>
<mml:mo>&#xd7;</mml:mo>
<mml:mi mathvariant="double-struck">X</mml:mi>
<mml:mo>&#x2192;</mml:mo>
<mml:mrow>
<mml:mo stretchy="false">(</mml:mo>
<mml:mrow>
<mml:mn>0</mml:mn>
<mml:mo>,</mml:mo>
<mml:mi>&#x221e;</mml:mi>
</mml:mrow>
<mml:mo stretchy="false">)</mml:mo>
</mml:mrow>
</mml:math>
</inline-formula>, <inline-formula id="inf18">
<mml:math id="m21">
<mml:mi>K</mml:mi>
<mml:mo>&#x2208;</mml:mo>
<mml:mi>C</mml:mi>
<mml:mrow>
<mml:mo stretchy="false">(</mml:mo>
<mml:mrow>
<mml:mi mathvariant="double-struck">X</mml:mi>
<mml:mo>&#xd7;</mml:mo>
<mml:mi mathvariant="double-struck">X</mml:mi>
</mml:mrow>
<mml:mo stretchy="false">)</mml:mo>
</mml:mrow>
</mml:math>
</inline-formula>. Since the kernel is fixed, the mention of this kernel will be omitted from notations, although the kernel plays a crucial role in our theory. Given any signed measure (or positive measure having bounded variation) <italic>&#x3c4;</italic> on <inline-formula id="inf19">
<mml:math id="m22">
<mml:mi mathvariant="double-struck">X</mml:mi>
</mml:math>
</inline-formula>, we define the witness function of <italic>&#x3c4;</italic> by<disp-formula id="e2">
<mml:math id="m23">
<mml:mi>U</mml:mi>
<mml:mfenced open="(" close=")">
<mml:mrow>
<mml:mi>&#x3c4;</mml:mi>
</mml:mrow>
</mml:mfenced>
<mml:mfenced open="(" close=")">
<mml:mrow>
<mml:mi>z</mml:mi>
</mml:mrow>
</mml:mfenced>
<mml:mo>&#x3d;</mml:mo>
<mml:msub>
<mml:mrow>
<mml:mo>&#x222b;</mml:mo>
</mml:mrow>
<mml:mrow>
<mml:mi mathvariant="double-struck">X</mml:mi>
</mml:mrow>
</mml:msub>
<mml:mi>K</mml:mi>
<mml:mfenced open="(" close=")">
<mml:mrow>
<mml:mi>z</mml:mi>
<mml:mo>,</mml:mo>
<mml:mi>x</mml:mi>
</mml:mrow>
</mml:mfenced>
<mml:mi>d</mml:mi>
<mml:mi>&#x3c4;</mml:mi>
<mml:mfenced open="(" close=")">
<mml:mrow>
<mml:mi>x</mml:mi>
</mml:mrow>
</mml:mfenced>
<mml:mo>,</mml:mo>
<mml:mspace width="2em"/>
<mml:mi>z</mml:mi>
<mml:mo>&#x2208;</mml:mo>
<mml:mi mathvariant="double-struck">X</mml:mi>
<mml:mo>,</mml:mo>
</mml:math>
<label>(2)</label>
</disp-formula>and similarly the magnitude of the witness function<disp-formula id="e3">
<mml:math id="m24">
<mml:mi>T</mml:mi>
<mml:mfenced open="(" close=")">
<mml:mrow>
<mml:mi>&#x3c4;</mml:mi>
</mml:mrow>
</mml:mfenced>
<mml:mfenced open="(" close=")">
<mml:mrow>
<mml:mi>z</mml:mi>
</mml:mrow>
</mml:mfenced>
<mml:mo>&#x3d;</mml:mo>
<mml:mo stretchy="false">&#x7c;</mml:mo>
<mml:mi>U</mml:mi>
<mml:mfenced open="(" close=")">
<mml:mrow>
<mml:mi>&#x3c4;</mml:mi>
</mml:mrow>
</mml:mfenced>
<mml:mfenced open="(" close=")">
<mml:mrow>
<mml:mi>z</mml:mi>
</mml:mrow>
</mml:mfenced>
<mml:mo stretchy="false">&#x7c;</mml:mo>
<mml:mo>,</mml:mo>
<mml:mspace width="2em"/>
<mml:mi>z</mml:mi>
<mml:mo>&#x2208;</mml:mo>
<mml:mi mathvariant="double-struck">X</mml:mi>
<mml:mo>.</mml:mo>
</mml:math>
<label>(3)</label>
</disp-formula>
</p>
<p>In the context of defining a distance between <italic>&#x3bc;</italic>
<sub>1</sub>, <italic>&#x3bc;</italic>
<sub>2</sub>, we take <italic>&#x3c4;</italic> &#x3d; <italic>&#x3bc;</italic>
<sub>1</sub> &#x2212; <italic>&#x3bc;</italic>
<sub>2</sub>, which results in a witness function<disp-formula id="equ1">
<mml:math id="m25">
<mml:mi>U</mml:mi>
<mml:mfenced open="(" close=")">
<mml:mrow>
<mml:msub>
<mml:mrow>
<mml:mi>&#x3bc;</mml:mi>
</mml:mrow>
<mml:mrow>
<mml:mn>1</mml:mn>
</mml:mrow>
</mml:msub>
<mml:mo>&#x2212;</mml:mo>
<mml:msub>
<mml:mrow>
<mml:mi>&#x3bc;</mml:mi>
</mml:mrow>
<mml:mrow>
<mml:mn>2</mml:mn>
</mml:mrow>
</mml:msub>
</mml:mrow>
</mml:mfenced>
<mml:mfenced open="(" close=")">
<mml:mrow>
<mml:mi>z</mml:mi>
</mml:mrow>
</mml:mfenced>
<mml:mo>&#x3d;</mml:mo>
<mml:msub>
<mml:mrow>
<mml:mi mathvariant="double-struck">E</mml:mi>
</mml:mrow>
<mml:mrow>
<mml:mi>x</mml:mi>
<mml:mo>&#x223c;</mml:mo>
<mml:msub>
<mml:mrow>
<mml:mi>&#x3bc;</mml:mi>
</mml:mrow>
<mml:mrow>
<mml:mn>1</mml:mn>
</mml:mrow>
</mml:msub>
</mml:mrow>
</mml:msub>
<mml:mfenced open="[" close="]">
<mml:mrow>
<mml:mi>K</mml:mi>
<mml:mfenced open="(" close=")">
<mml:mrow>
<mml:mi>z</mml:mi>
<mml:mo>,</mml:mo>
<mml:mi>x</mml:mi>
</mml:mrow>
</mml:mfenced>
</mml:mrow>
</mml:mfenced>
<mml:mo>&#x2212;</mml:mo>
<mml:msub>
<mml:mrow>
<mml:mi mathvariant="double-struck">E</mml:mi>
</mml:mrow>
<mml:mrow>
<mml:mi>y</mml:mi>
<mml:mo>&#x223c;</mml:mo>
<mml:msub>
<mml:mrow>
<mml:mi>&#x3bc;</mml:mi>
</mml:mrow>
<mml:mrow>
<mml:mn>2</mml:mn>
</mml:mrow>
</mml:msub>
</mml:mrow>
</mml:msub>
<mml:mfenced open="[" close="]">
<mml:mrow>
<mml:mi>K</mml:mi>
<mml:mfenced open="(" close=")">
<mml:mrow>
<mml:mi>z</mml:mi>
<mml:mo>,</mml:mo>
<mml:mi>y</mml:mi>
</mml:mrow>
</mml:mfenced>
</mml:mrow>
</mml:mfenced>
<mml:mo>.</mml:mo>
</mml:math>
</disp-formula>
</p>
<p>To quantify where <italic>T</italic> (<italic>&#x3bc;</italic>
<sub>1</sub> &#x2212; <italic>&#x3bc;</italic>
<sub>2</sub>) is small, we define the cumulative distribution function (CDF) of a Borel measurable function <inline-formula id="inf20">
<mml:math id="m26">
<mml:mi>f</mml:mi>
<mml:mo>:</mml:mo>
<mml:mi mathvariant="double-struck">X</mml:mi>
<mml:mo>&#x2192;</mml:mo>
<mml:mi mathvariant="double-struck">R</mml:mi>
</mml:math>
</inline-formula> by<disp-formula id="e4">
<mml:math id="m27">
<mml:mi mathvariant="normal">&#x39b;</mml:mi>
<mml:mfenced open="(" close=")">
<mml:mrow>
<mml:mi>f</mml:mi>
</mml:mrow>
</mml:mfenced>
<mml:mfenced open="(" close=")">
<mml:mrow>
<mml:mi>t</mml:mi>
</mml:mrow>
</mml:mfenced>
<mml:mo>&#x3d;</mml:mo>
<mml:mi mathvariant="normal">&#x39b;</mml:mi>
<mml:mfenced open="(" close=")">
<mml:mrow>
<mml:msup>
<mml:mrow>
<mml:mi>&#x3bd;</mml:mi>
</mml:mrow>
<mml:mrow>
<mml:mo>&#x2a;</mml:mo>
</mml:mrow>
</mml:msup>
<mml:mo>;</mml:mo>
<mml:mi>f</mml:mi>
</mml:mrow>
</mml:mfenced>
<mml:mfenced open="(" close=")">
<mml:mrow>
<mml:mi>t</mml:mi>
</mml:mrow>
</mml:mfenced>
<mml:mo>&#x3d;</mml:mo>
<mml:msup>
<mml:mrow>
<mml:mi>&#x3bd;</mml:mi>
</mml:mrow>
<mml:mrow>
<mml:mo>&#x2a;</mml:mo>
</mml:mrow>
</mml:msup>
<mml:mfenced open="(" close=")">
<mml:mrow>
<mml:mfenced open="{" close="}">
<mml:mrow>
<mml:mi>z</mml:mi>
<mml:mo>:</mml:mo>
<mml:mo stretchy="false">&#x7c;</mml:mo>
<mml:mi>f</mml:mi>
<mml:mfenced open="(" close=")">
<mml:mrow>
<mml:mi>z</mml:mi>
</mml:mrow>
</mml:mfenced>
<mml:mo stretchy="false">&#x7c;</mml:mo>
<mml:mo>&#x3c;</mml:mo>
<mml:mi>t</mml:mi>
</mml:mrow>
</mml:mfenced>
</mml:mrow>
</mml:mfenced>
<mml:mo>,</mml:mo>
<mml:mspace width="2em"/>
<mml:mi>t</mml:mi>
<mml:mo>&#x2208;</mml:mo>
<mml:mfenced open="[" close=")">
<mml:mrow>
<mml:mn>0</mml:mn>
<mml:mo>,</mml:mo>
<mml:mi>&#x221e;</mml:mi>
</mml:mrow>
</mml:mfenced>
<mml:mo>,</mml:mo>
</mml:math>
<label>(4)</label>
</disp-formula>and its &#x201c;inverse&#x201d; CDF by<disp-formula id="e5">
<mml:math id="m28">
<mml:msup>
<mml:mrow>
<mml:mi>f</mml:mi>
</mml:mrow>
<mml:mrow>
<mml:mi>&#x23;</mml:mi>
</mml:mrow>
</mml:msup>
<mml:mfenced open="(" close=")">
<mml:mrow>
<mml:mi>u</mml:mi>
</mml:mrow>
</mml:mfenced>
<mml:mo>&#x3d;</mml:mo>
<mml:mi>sup</mml:mi>
<mml:mfenced open="{" close="}">
<mml:mrow>
<mml:mi>t</mml:mi>
<mml:mo>&#x2208;</mml:mo>
<mml:mfenced open="[" close="">
<mml:mrow>
<mml:mn>0</mml:mn>
</mml:mrow>
</mml:mfenced>
<mml:mo>,</mml:mo>
<mml:mfenced open="" close=")">
<mml:mrow>
<mml:mi>&#x221e;</mml:mi>
</mml:mrow>
</mml:mfenced>
<mml:mo>:</mml:mo>
<mml:mi mathvariant="normal">&#x39b;</mml:mi>
<mml:mfenced open="(" close=")">
<mml:mrow>
<mml:mi>f</mml:mi>
</mml:mrow>
</mml:mfenced>
<mml:mfenced open="(" close=")">
<mml:mrow>
<mml:mi>t</mml:mi>
</mml:mrow>
</mml:mfenced>
<mml:mo>&#x2264;</mml:mo>
<mml:mi>u</mml:mi>
</mml:mrow>
</mml:mfenced>
<mml:mo>,</mml:mo>
<mml:mspace width="2em"/>
<mml:mi>u</mml:mi>
<mml:mo>&#x2208;</mml:mo>
<mml:mfenced open="[" close="">
<mml:mrow>
<mml:mn>0</mml:mn>
</mml:mrow>
</mml:mfenced>
<mml:mo>,</mml:mo>
<mml:mfenced open="" close=")">
<mml:mrow>
<mml:mi>&#x221e;</mml:mi>
</mml:mrow>
</mml:mfenced>
<mml:mo>.</mml:mo>
</mml:math>
<label>(5)</label>
</disp-formula>
</p>
<p>Both &#x39b;(<italic>f</italic>) and <italic>f</italic>
<sup>
<italic>&#x23;</italic>
</sup> are non-decreasing functions, and <italic>f</italic>
<sup>
<italic>&#x23;</italic>
</sup>(<italic>u</italic>) defines the <bold>
<italic>u</italic>
</bold>-th quartile of&#x20;<italic>f</italic>.</p>
<p>Finally, we are prepared to define our kdiff distance between probability measures <italic>&#x3bc;</italic>
<sub>1</sub>, <italic>&#x3bc;</italic>
<sub>2</sub>. Having defined <italic>T</italic> (<italic>&#x3bc;</italic>
<sub>1</sub> &#x2212; <italic>&#x3bc;</italic>
<sub>2</sub>) (<italic>z</italic>), we now define kdiff to be the <italic>&#x3b1;</italic> quantile of <italic>T</italic> (<italic>&#x3bc;</italic>
<sub>1</sub> &#x2212; <italic>&#x3bc;</italic>
<sub>2</sub>),<disp-formula id="e6">
<mml:math id="m29">
<mml:mi mathvariant="bold">k</mml:mi>
<mml:mi mathvariant="bold">d</mml:mi>
<mml:mi mathvariant="bold">i</mml:mi>
<mml:mi mathvariant="bold">ff</mml:mi>
<mml:mfenced open="(" close=")">
<mml:mrow>
<mml:msub>
<mml:mrow>
<mml:mi>&#x3bc;</mml:mi>
</mml:mrow>
<mml:mrow>
<mml:mn>1</mml:mn>
</mml:mrow>
</mml:msub>
<mml:mo>,</mml:mo>
<mml:msub>
<mml:mrow>
<mml:mi>&#x3bc;</mml:mi>
</mml:mrow>
<mml:mrow>
<mml:mn>2</mml:mn>
</mml:mrow>
</mml:msub>
<mml:mo>;</mml:mo>
<mml:mi>&#x3b1;</mml:mi>
</mml:mrow>
</mml:mfenced>
<mml:mo>&#x3d;</mml:mo>
<mml:msup>
<mml:mrow>
<mml:mfenced open="(" close=")">
<mml:mrow>
<mml:mi>T</mml:mi>
<mml:mfenced open="(" close=")">
<mml:mrow>
<mml:msub>
<mml:mrow>
<mml:mi>&#x3bc;</mml:mi>
</mml:mrow>
<mml:mrow>
<mml:mn>1</mml:mn>
</mml:mrow>
</mml:msub>
<mml:mo>&#x2212;</mml:mo>
<mml:msub>
<mml:mrow>
<mml:mi>&#x3bc;</mml:mi>
</mml:mrow>
<mml:mrow>
<mml:mn>2</mml:mn>
</mml:mrow>
</mml:msub>
</mml:mrow>
</mml:mfenced>
</mml:mrow>
</mml:mfenced>
</mml:mrow>
<mml:mrow>
<mml:mi>&#x23;</mml:mi>
</mml:mrow>
</mml:msup>
<mml:mfenced open="(" close=")">
<mml:mrow>
<mml:mi>&#x3b1;</mml:mi>
</mml:mrow>
</mml:mfenced>
<mml:mo>,</mml:mo>
<mml:mspace width="2em"/>
<mml:mi>&#x3b1;</mml:mi>
<mml:mo>&#x2208;</mml:mo>
<mml:mfenced open="(" close=")">
<mml:mrow>
<mml:mn>0,1</mml:mn>
</mml:mrow>
</mml:mfenced>
<mml:mo>.</mml:mo>
</mml:math>
<label>(6)</label>
</disp-formula>
</p>
<p>The intuition of <xref ref-type="disp-formula" rid="e6">(6)</xref> is that, if <italic>&#x3bc;</italic>
<sub>1</sub> &#x3d; <italic>&#x3bc;</italic>
<sub>2</sub>, the resulting kdiff statistic will be zero. But beyond this, if <italic>T</italic> (<italic>&#x3bc;</italic>
<sub>1</sub> &#x2212; <italic>&#x3bc;</italic>
<sub>2</sub>) (<italic>z</italic>) &#x3d; 0 for a set <inline-formula id="inf21">
<mml:math id="m30">
<mml:mi>z</mml:mi>
<mml:mo>&#x2208;</mml:mo>
<mml:mi>A</mml:mi>
<mml:mo>&#x2282;</mml:mo>
<mml:mi mathvariant="double-struck">X</mml:mi>
</mml:math>
</inline-formula> such that <italic>&#x3bd;</italic>&#x2a;(<italic>A</italic>) &#x3e; 0, then for a localized enough kernel, there exists a quantile <italic>&#x3b1;</italic> for which we can still have the resulting kdiff statistic be close to zero. This allows us to match distributions that agree over partial support. This will be discussed more precisely in <xref ref-type="sec" rid="s2-3">Section&#x20;2.3</xref>.</p>
</sec>
<sec id="s2-3">
<title>2.3 Separability Theorems for kdiff</title>
<p>For the purposes of analyzing the kdiff statistic, we will focus on the setting of resolving mixture models of probabilities on <inline-formula id="inf22">
<mml:math id="m31">
<mml:mi mathvariant="double-struck">X</mml:mi>
</mml:math>
</inline-formula> when only one of the components agree. Accordingly, for any <italic>&#x3b4;</italic> &#x2208; (0, 1), we define <inline-formula id="inf23">
<mml:math id="m32">
<mml:msub>
<mml:mrow>
<mml:mi mathvariant="double-struck">P</mml:mi>
</mml:mrow>
<mml:mrow>
<mml:mi>&#x3b4;</mml:mi>
</mml:mrow>
</mml:msub>
</mml:math>
</inline-formula> to be the class of all probability measures <italic>&#x3bc;</italic> on <inline-formula id="inf24">
<mml:math id="m33">
<mml:mi mathvariant="double-struck">X</mml:mi>
</mml:math>
</inline-formula> which can be expressed as <italic>&#x3bc;</italic> &#x3d; <italic>&#x3b4;&#x3bc;</italic>
<sub>
<italic>F</italic>
</sub> &#x2b; (1 &#x2212; <italic>&#x3b4;</italic>)<italic>&#x3bc;</italic>
<sub>
<italic>B</italic>
</sub>, where <italic>&#x3bc;</italic>
<sub>
<italic>F</italic>
</sub> and <italic>&#x3bc;</italic>
<sub>
<italic>B</italic>
</sub> are probability measures on <inline-formula id="inf25">
<mml:math id="m34">
<mml:mi mathvariant="double-struck">X</mml:mi>
</mml:math>
</inline-formula>. With the applications in the paper in mind, we will refer to <italic>&#x3bc;</italic>
<sub>
<italic>F</italic>
</sub> as the <italic>foreground</italic> and <italic>&#x3bc;</italic>
<sub>
<italic>B</italic>
</sub> as the <italic>background</italic> probabilities. Our interest is in developing a test to see whether given two measures <italic>&#x3bc;</italic>
<sub>1</sub> and <italic>&#x3bc;</italic>
<sub>2</sub> in <inline-formula id="inf26">
<mml:math id="m35">
<mml:msub>
<mml:mrow>
<mml:mi mathvariant="double-struck">P</mml:mi>
</mml:mrow>
<mml:mrow>
<mml:mi>&#x3b4;</mml:mi>
</mml:mrow>
</mml:msub>
</mml:math>
</inline-formula>, the corresponding foreground components agree. Clearly, the same discussion could also apply to the case when we wish to focus on the background components with obvious changes. We also note that in general, <inline-formula id="inf27">
<mml:math id="m36">
<mml:msub>
<mml:mrow>
<mml:mi mathvariant="double-struck">P</mml:mi>
</mml:mrow>
<mml:mrow>
<mml:mi>&#x3b4;</mml:mi>
</mml:mrow>
</mml:msub>
</mml:math>
</inline-formula> is simply the space of probability measures, but it should be thought of as the space of mixtures of almost-disjoint measures. We will add the necessary assumptions on <italic>&#x3b4;</italic>, <italic>&#x3bc;</italic>
<sub>
<italic>F</italic>
</sub>, <italic>&#x3bc;</italic>
<sub>
<italic>B</italic>
</sub> in the main results&#x20;below.</p>
<p>We first present some preparatory material before reaching our desired statements. For any subset <inline-formula id="inf28">
<mml:math id="m37">
<mml:mi>A</mml:mi>
<mml:mo>&#x2286;</mml:mo>
<mml:mi mathvariant="double-struck">X</mml:mi>
</mml:math>
</inline-formula> and <inline-formula id="inf29">
<mml:math id="m38">
<mml:mi>x</mml:mi>
<mml:mo>&#x2208;</mml:mo>
<mml:mi mathvariant="double-struck">X</mml:mi>
</mml:math>
</inline-formula>, we define<disp-formula id="e7">
<mml:math id="m39">
<mml:mi mathvariant="sans-serif">d</mml:mi>
<mml:mi mathvariant="sans-serif">i</mml:mi>
<mml:mi mathvariant="sans-serif">s</mml:mi>
<mml:mi mathvariant="sans-serif">t</mml:mi>
<mml:mfenced open="(" close=")">
<mml:mrow>
<mml:mi>A</mml:mi>
<mml:mo>,</mml:mo>
<mml:mi>x</mml:mi>
</mml:mrow>
</mml:mfenced>
<mml:mo>&#x3d;</mml:mo>
<mml:msub>
<mml:mrow>
<mml:mi>inf</mml:mi>
</mml:mrow>
<mml:mrow>
<mml:mi>y</mml:mi>
<mml:mo>&#x2208;</mml:mo>
<mml:mi>A</mml:mi>
</mml:mrow>
</mml:msub>
<mml:mi>&#x3c1;</mml:mi>
<mml:mfenced open="(" close=")">
<mml:mrow>
<mml:mi>y</mml:mi>
<mml:mo>,</mml:mo>
<mml:mi>x</mml:mi>
</mml:mrow>
</mml:mfenced>
<mml:mo>.</mml:mo>
</mml:math>
<label>(7)</label>
</disp-formula>
</p>
<p>The support of a finite positive measure <italic>&#x3bc;</italic>, denoted by <inline-formula id="inf30">
<mml:math id="m40">
<mml:mi mathvariant="sans-serif">s</mml:mi>
<mml:mi mathvariant="sans-serif">u</mml:mi>
<mml:mi mathvariant="sans-serif">p</mml:mi>
<mml:mi mathvariant="sans-serif">p</mml:mi>
<mml:mrow>
<mml:mo stretchy="false">(</mml:mo>
<mml:mrow>
<mml:mi>&#x3bc;</mml:mi>
</mml:mrow>
<mml:mo stretchy="false">)</mml:mo>
</mml:mrow>
</mml:math>
</inline-formula> is the set of all <inline-formula id="inf31">
<mml:math id="m41">
<mml:mi>x</mml:mi>
<mml:mo>&#x2208;</mml:mo>
<mml:mi mathvariant="double-struck">X</mml:mi>
</mml:math>
</inline-formula> such that <italic>&#x3bc;</italic>(<italic>U</italic>) &#x3e; 0 for all open subsets <italic>U</italic> containing <italic>x</italic>. Clearly, <inline-formula id="inf32">
<mml:math id="m42">
<mml:mi mathvariant="sans-serif">s</mml:mi>
<mml:mi mathvariant="sans-serif">u</mml:mi>
<mml:mi mathvariant="sans-serif">p</mml:mi>
<mml:mi mathvariant="sans-serif">p</mml:mi>
<mml:mrow>
<mml:mo stretchy="false">(</mml:mo>
<mml:mrow>
<mml:mi>&#x3bc;</mml:mi>
</mml:mrow>
<mml:mo stretchy="false">)</mml:mo>
</mml:mrow>
</mml:math>
</inline-formula> is a closed set. If <italic>&#x3c3;</italic> is a non-zero signed measure and <italic>&#x3c3;</italic> &#x3d; <italic>&#x3c3;</italic>
<sup>&#x2b;</sup> &#x2212; <italic>&#x3c3;</italic>
<sup>&#x2212;</sup> is the Jordan decomposition of <italic>&#x3c3;</italic> then we define <inline-formula id="inf33">
<mml:math id="m43">
<mml:mi mathvariant="sans-serif">s</mml:mi>
<mml:mi mathvariant="sans-serif">u</mml:mi>
<mml:mi mathvariant="sans-serif">p</mml:mi>
<mml:mi mathvariant="sans-serif">p</mml:mi>
<mml:mrow>
<mml:mo stretchy="false">(</mml:mo>
<mml:mrow>
<mml:mi>&#x3c3;</mml:mi>
</mml:mrow>
<mml:mo stretchy="false">)</mml:mo>
</mml:mrow>
<mml:mo>&#x3d;</mml:mo>
<mml:mi mathvariant="sans-serif">s</mml:mi>
<mml:mi mathvariant="sans-serif">u</mml:mi>
<mml:mi mathvariant="sans-serif">p</mml:mi>
<mml:mi mathvariant="sans-serif">p</mml:mi>
<mml:mrow>
<mml:mo stretchy="false">(</mml:mo>
<mml:mrow>
<mml:msup>
<mml:mrow>
<mml:mi>&#x3c3;</mml:mi>
</mml:mrow>
<mml:mrow>
<mml:mo>&#x2b;</mml:mo>
</mml:mrow>
</mml:msup>
</mml:mrow>
<mml:mo stretchy="false">)</mml:mo>
</mml:mrow>
<mml:mo>&#x222a;</mml:mo>
<mml:mi mathvariant="sans-serif">s</mml:mi>
<mml:mi mathvariant="sans-serif">u</mml:mi>
<mml:mi mathvariant="sans-serif">p</mml:mi>
<mml:mi mathvariant="sans-serif">p</mml:mi>
<mml:mrow>
<mml:mo stretchy="false">(</mml:mo>
<mml:mrow>
<mml:msup>
<mml:mrow>
<mml:mi>&#x3c3;</mml:mi>
</mml:mrow>
<mml:mrow>
<mml:mo>&#x2212;</mml:mo>
</mml:mrow>
</mml:msup>
</mml:mrow>
<mml:mo stretchy="false">)</mml:mo>
</mml:mrow>
</mml:math>
</inline-formula>. If <inline-formula id="inf34">
<mml:math id="m44">
<mml:mi>f</mml:mi>
<mml:mo>:</mml:mo>
<mml:mi mathvariant="double-struck">X</mml:mi>
<mml:mo>&#x2192;</mml:mo>
<mml:mi mathvariant="double-struck">R</mml:mi>
</mml:math>
</inline-formula>, we define<disp-formula id="equ2">
<mml:math id="m45">
<mml:mo stretchy="false">&#x2016;</mml:mo>
<mml:mi>f</mml:mi>
<mml:msub>
<mml:mrow>
<mml:mo stretchy="false">&#x2016;</mml:mo>
</mml:mrow>
<mml:mrow>
<mml:mi>&#x221e;</mml:mi>
</mml:mrow>
</mml:msub>
<mml:mo>&#x3d;</mml:mo>
<mml:msub>
<mml:mrow>
<mml:mi>sup</mml:mi>
</mml:mrow>
<mml:mrow>
<mml:mi>x</mml:mi>
<mml:mo>&#x2208;</mml:mo>
<mml:mi mathvariant="double-struck">X</mml:mi>
</mml:mrow>
</mml:msub>
<mml:mo stretchy="false">&#x7c;</mml:mo>
<mml:mi>f</mml:mi>
<mml:mfenced open="(" close=")">
<mml:mrow>
<mml:mi>x</mml:mi>
</mml:mrow>
</mml:mfenced>
<mml:mo stretchy="false">&#x7c;</mml:mo>
<mml:mo>.</mml:mo>
</mml:math>
</disp-formula>
</p>
<p>The following lemma summarizes some important but easy properties of quantities &#x39b;(<italic>f</italic>) and <italic>f</italic>
<sup>
<italic>&#x23;</italic>
</sup> defined in <xref ref-type="disp-formula" rid="e4">Eqs. 4</xref>, <xref ref-type="disp-formula" rid="e5">5</xref> respectively.</p>
<p>
<statement content-type="lemma" id="Lemma_1">
<label>Lemma 1</label>
<p>(a) <italic>For</italic> <italic>t</italic>, <italic>u</italic>&#x20;&#x2208; [0, <italic>&#x221e;</italic>)<italic>,</italic>
<disp-formula id="e8">
<mml:math id="m46">
<mml:mi mathvariant="normal">&#x39b;</mml:mi>
<mml:mfenced open="(" close=")">
<mml:mrow>
<mml:mi>f</mml:mi>
</mml:mrow>
</mml:mfenced>
<mml:mfenced open="(" close=")">
<mml:mrow>
<mml:mi>t</mml:mi>
</mml:mrow>
</mml:mfenced>
<mml:mo>&#x2264;</mml:mo>
<mml:mi>u</mml:mi>
<mml:mo>&#x21d2;</mml:mo>
<mml:mi>t</mml:mi>
<mml:mo>&#x2264;</mml:mo>
<mml:msup>
<mml:mrow>
<mml:mi>f</mml:mi>
</mml:mrow>
<mml:mrow>
<mml:mi>&#x23;</mml:mi>
</mml:mrow>
</mml:msup>
<mml:mfenced open="(" close=")">
<mml:mrow>
<mml:mi>u</mml:mi>
</mml:mrow>
</mml:mfenced>
<mml:mo>,</mml:mo>
<mml:mspace width="2em"/>
<mml:mi>u</mml:mi>
<mml:mo>&#x2264;</mml:mo>
<mml:mi mathvariant="normal">&#x39b;</mml:mi>
<mml:mfenced open="(" close=")">
<mml:mrow>
<mml:mi>f</mml:mi>
</mml:mrow>
</mml:mfenced>
<mml:mfenced open="(" close=")">
<mml:mrow>
<mml:mi>t</mml:mi>
</mml:mrow>
</mml:mfenced>
<mml:mo>&#x21d2;</mml:mo>
<mml:msup>
<mml:mrow>
<mml:mi>f</mml:mi>
</mml:mrow>
<mml:mrow>
<mml:mi>&#x23;</mml:mi>
</mml:mrow>
</mml:msup>
<mml:mfenced open="(" close=")">
<mml:mrow>
<mml:mi>u</mml:mi>
</mml:mrow>
</mml:mfenced>
<mml:mo>&#x2264;</mml:mo>
<mml:mi>t</mml:mi>
<mml:mo>.</mml:mo>
</mml:math>
<label>(8)</label>
</disp-formula>
</p>
<p>(b) <italic>If</italic> <italic>&#x3f5;</italic> &#x3e; 0<italic>,</italic> <inline-formula id="inf35">
<mml:math id="m47">
<mml:mi>f</mml:mi>
<mml:mo>,</mml:mo>
<mml:mi>g</mml:mi>
<mml:mo>:</mml:mo>
<mml:mi mathvariant="double-struck">X</mml:mi>
<mml:mo>&#x2192;</mml:mo>
<mml:mi mathvariant="double-struck">R</mml:mi>
</mml:math>
</inline-formula>
<italic>, and</italic> &#x2016;<italic>f</italic>&#x20;&#x2212; <italic>g</italic>&#x2016;<sub>
<italic>&#x221e;</italic>
</sub> &#x2264; <italic>&#x3f5;</italic>
<italic>, then</italic> sup<sub>
<italic>u</italic>&#x2208;[0,<italic>&#x221e;</italic>)</sub>&#x7c;<italic>f</italic>
<sup>
<italic>&#x23;</italic>
</sup>(<italic>u</italic>) &#x2212; <italic>g</italic>
<sup>
<italic>&#x23;</italic>
</sup>(<italic>u</italic>)&#x7c; &#x2264; <italic>&#x3f5;</italic>
<italic>.</italic>
</p>
<p>Our goal is to investigate sufficient conditions on two measures in <inline-formula id="inf36">
<mml:math id="m48">
<mml:msub>
<mml:mrow>
<mml:mi mathvariant="double-struck">P</mml:mi>
</mml:mrow>
<mml:mrow>
<mml:mi>&#x3b4;</mml:mi>
</mml:mrow>
</mml:msub>
</mml:math>
</inline-formula> so that kdiff can distinguish if the foreground components of the measures are the same. For this purpose, we introduce some further notation, where we suppress the mention of certain quantities for brevity. Let <inline-formula id="inf37">
<mml:math id="m49">
<mml:msub>
<mml:mrow>
<mml:mi>&#x3bc;</mml:mi>
</mml:mrow>
<mml:mrow>
<mml:mi>j</mml:mi>
</mml:mrow>
</mml:msub>
<mml:mo>&#x3d;</mml:mo>
<mml:mi>&#x3b4;</mml:mi>
<mml:msub>
<mml:mrow>
<mml:mi>&#x3bc;</mml:mi>
</mml:mrow>
<mml:mrow>
<mml:mi>j</mml:mi>
<mml:mo>,</mml:mo>
<mml:mi>F</mml:mi>
</mml:mrow>
</mml:msub>
<mml:mo>&#x2b;</mml:mo>
<mml:mrow>
<mml:mo stretchy="false">(</mml:mo>
<mml:mrow>
<mml:mn>1</mml:mn>
<mml:mo>&#x2212;</mml:mo>
<mml:mi>&#x3b4;</mml:mi>
</mml:mrow>
<mml:mo stretchy="false">)</mml:mo>
</mml:mrow>
<mml:msub>
<mml:mrow>
<mml:mi>&#x3bc;</mml:mi>
</mml:mrow>
<mml:mrow>
<mml:mi>j</mml:mi>
<mml:mo>,</mml:mo>
<mml:mi>B</mml:mi>
</mml:mrow>
</mml:msub>
<mml:mo>&#x2208;</mml:mo>
<mml:msub>
<mml:mrow>
<mml:mi mathvariant="double-struck">P</mml:mi>
</mml:mrow>
<mml:mrow>
<mml:mi>&#x3b4;</mml:mi>
</mml:mrow>
</mml:msub>
</mml:math>
</inline-formula>, <italic>j</italic>&#x20;&#x3d; 1, 2, and <inline-formula id="inf38">
<mml:math id="m50">
<mml:msub>
<mml:mrow>
<mml:mi mathvariant="double-struck">S</mml:mi>
</mml:mrow>
<mml:mrow>
<mml:mi>F</mml:mi>
</mml:mrow>
</mml:msub>
<mml:mo>&#x3d;</mml:mo>
<mml:mi mathvariant="sans-serif">s</mml:mi>
<mml:mi mathvariant="sans-serif">u</mml:mi>
<mml:mi mathvariant="sans-serif">p</mml:mi>
<mml:mi mathvariant="sans-serif">p</mml:mi>
<mml:mrow>
<mml:mo stretchy="false">(</mml:mo>
<mml:mrow>
<mml:msub>
<mml:mrow>
<mml:mi>&#x3bc;</mml:mi>
</mml:mrow>
<mml:mrow>
<mml:mn>1</mml:mn>
<mml:mo>,</mml:mo>
<mml:mi>F</mml:mi>
</mml:mrow>
</mml:msub>
</mml:mrow>
<mml:mo stretchy="false">)</mml:mo>
</mml:mrow>
<mml:mo>&#x222a;</mml:mo>
<mml:mi mathvariant="sans-serif">s</mml:mi>
<mml:mi mathvariant="sans-serif">u</mml:mi>
<mml:mi mathvariant="sans-serif">p</mml:mi>
<mml:mi mathvariant="sans-serif">p</mml:mi>
<mml:mrow>
<mml:mo stretchy="false">(</mml:mo>
<mml:mrow>
<mml:msub>
<mml:mrow>
<mml:mi>&#x3bc;</mml:mi>
</mml:mrow>
<mml:mrow>
<mml:mn>2</mml:mn>
<mml:mo>,</mml:mo>
<mml:mi>F</mml:mi>
</mml:mrow>
</mml:msub>
</mml:mrow>
<mml:mo stretchy="false">)</mml:mo>
</mml:mrow>
</mml:math>
</inline-formula>, and we define <inline-formula id="inf39">
<mml:math id="m51">
<mml:msup>
<mml:mrow>
<mml:mi mathvariant="script">S</mml:mi>
</mml:mrow>
<mml:mrow>
<mml:mi>c</mml:mi>
</mml:mrow>
</mml:msup>
<mml:mo>&#x3d;</mml:mo>
<mml:mi mathvariant="double-struck">X</mml:mi>
<mml:mo>/</mml:mo>
<mml:mi mathvariant="script">S</mml:mi>
</mml:math>
</inline-formula>. We define for <italic>&#x3b7;</italic>, <italic>&#x3b8;</italic> &#x3e; 0,<disp-formula id="e9">
<mml:math id="m52">
<mml:mtable class="aligned">
<mml:mtr>
<mml:mtd columnalign="right">
<mml:msub>
<mml:mrow>
<mml:mi mathvariant="script">S</mml:mi>
</mml:mrow>
<mml:mrow>
<mml:mi>B</mml:mi>
</mml:mrow>
</mml:msub>
<mml:mfenced open="(" close=")">
<mml:mrow>
<mml:msub>
<mml:mrow>
<mml:mi>&#x3bc;</mml:mi>
</mml:mrow>
<mml:mrow>
<mml:mn>1</mml:mn>
</mml:mrow>
</mml:msub>
<mml:mo>,</mml:mo>
<mml:msub>
<mml:mrow>
<mml:mi>&#x3bc;</mml:mi>
</mml:mrow>
<mml:mrow>
<mml:mn>2</mml:mn>
</mml:mrow>
</mml:msub>
<mml:mo>;</mml:mo>
<mml:mi>&#x3b7;</mml:mi>
</mml:mrow>
</mml:mfenced>
</mml:mtd>
<mml:mtd columnalign="left">
<mml:mo>&#x3d;</mml:mo>
<mml:mfenced open="{" close="}">
<mml:mrow>
<mml:mi>z</mml:mi>
<mml:mo>&#x2208;</mml:mo>
<mml:mi mathvariant="double-struck">X</mml:mi>
<mml:mo>:</mml:mo>
<mml:mi>T</mml:mi>
<mml:mfenced open="(" close=")">
<mml:mrow>
<mml:msub>
<mml:mrow>
<mml:mi>&#x3bc;</mml:mi>
</mml:mrow>
<mml:mrow>
<mml:mn>1</mml:mn>
<mml:mo>,</mml:mo>
<mml:mi>B</mml:mi>
</mml:mrow>
</mml:msub>
<mml:mo>&#x2212;</mml:mo>
<mml:msub>
<mml:mrow>
<mml:mi>&#x3bc;</mml:mi>
</mml:mrow>
<mml:mrow>
<mml:mn>2</mml:mn>
<mml:mo>,</mml:mo>
<mml:mi>B</mml:mi>
</mml:mrow>
</mml:msub>
</mml:mrow>
</mml:mfenced>
<mml:mfenced open="(" close=")">
<mml:mrow>
<mml:mi>z</mml:mi>
</mml:mrow>
</mml:mfenced>
<mml:mo>&#x3c;</mml:mo>
<mml:mi>&#x3b7;</mml:mi>
</mml:mrow>
</mml:mfenced>
<mml:mo>,</mml:mo>
</mml:mtd>
<mml:mtd columnalign="right">
<mml:msub>
<mml:mrow>
<mml:mi>&#x3d5;</mml:mi>
</mml:mrow>
<mml:mrow>
<mml:mi>B</mml:mi>
</mml:mrow>
</mml:msub>
<mml:mfenced open="(" close=")">
<mml:mrow>
<mml:mi>&#x3b7;</mml:mi>
</mml:mrow>
</mml:mfenced>
</mml:mtd>
<mml:mtd columnalign="left">
<mml:mo>&#x3d;</mml:mo>
<mml:msup>
<mml:mrow>
<mml:mi>&#x3bd;</mml:mi>
</mml:mrow>
<mml:mrow>
<mml:mo>&#x2a;</mml:mo>
</mml:mrow>
</mml:msup>
<mml:mfenced open="(" close=")">
<mml:mrow>
<mml:msub>
<mml:mrow>
<mml:mi mathvariant="script">S</mml:mi>
</mml:mrow>
<mml:mrow>
<mml:mi>B</mml:mi>
</mml:mrow>
</mml:msub>
<mml:mfenced open="(" close=")">
<mml:mrow>
<mml:msub>
<mml:mrow>
<mml:mi>&#x3bc;</mml:mi>
</mml:mrow>
<mml:mrow>
<mml:mn>1</mml:mn>
</mml:mrow>
</mml:msub>
<mml:mo>,</mml:mo>
<mml:msub>
<mml:mrow>
<mml:mi>&#x3bc;</mml:mi>
</mml:mrow>
<mml:mrow>
<mml:mn>2</mml:mn>
</mml:mrow>
</mml:msub>
<mml:mo>;</mml:mo>
<mml:mi>&#x3b7;</mml:mi>
</mml:mrow>
</mml:mfenced>
</mml:mrow>
</mml:mfenced>
</mml:mtd>
</mml:mtr>
<mml:mtr>
<mml:mtd columnalign="right">
<mml:msub>
<mml:mrow>
<mml:mi mathvariant="script">S</mml:mi>
</mml:mrow>
<mml:mrow>
<mml:mi>F</mml:mi>
</mml:mrow>
</mml:msub>
<mml:mfenced open="(" close=")">
<mml:mrow>
<mml:msub>
<mml:mrow>
<mml:mi>&#x3bc;</mml:mi>
</mml:mrow>
<mml:mrow>
<mml:mn>1</mml:mn>
</mml:mrow>
</mml:msub>
<mml:mo>,</mml:mo>
<mml:msub>
<mml:mrow>
<mml:mi>&#x3bc;</mml:mi>
</mml:mrow>
<mml:mrow>
<mml:mn>2</mml:mn>
</mml:mrow>
</mml:msub>
<mml:mo>;</mml:mo>
<mml:mi>&#x3b7;</mml:mi>
</mml:mrow>
</mml:mfenced>
</mml:mtd>
<mml:mtd columnalign="left">
<mml:mo>&#x3d;</mml:mo>
<mml:mfenced open="{" close="}">
<mml:mrow>
<mml:mi>z</mml:mi>
<mml:mo>&#x2208;</mml:mo>
<mml:mi mathvariant="double-struck">X</mml:mi>
<mml:mo>:</mml:mo>
<mml:mi>T</mml:mi>
<mml:mfenced open="(" close=")">
<mml:mrow>
<mml:msub>
<mml:mrow>
<mml:mi>&#x3bc;</mml:mi>
</mml:mrow>
<mml:mrow>
<mml:mn>1</mml:mn>
<mml:mo>,</mml:mo>
<mml:mi>F</mml:mi>
</mml:mrow>
</mml:msub>
<mml:mo>&#x2212;</mml:mo>
<mml:msub>
<mml:mrow>
<mml:mi>&#x3bc;</mml:mi>
</mml:mrow>
<mml:mrow>
<mml:mn>2</mml:mn>
<mml:mo>,</mml:mo>
<mml:mi>F</mml:mi>
</mml:mrow>
</mml:msub>
</mml:mrow>
</mml:mfenced>
<mml:mfenced open="(" close=")">
<mml:mrow>
<mml:mi>z</mml:mi>
</mml:mrow>
</mml:mfenced>
<mml:mo>&#x3c;</mml:mo>
<mml:mi>&#x3b7;</mml:mi>
</mml:mrow>
</mml:mfenced>
<mml:mo>,</mml:mo>
</mml:mtd>
<mml:mtd columnalign="right">
<mml:msub>
<mml:mrow>
<mml:mi>&#x3d5;</mml:mi>
</mml:mrow>
<mml:mrow>
<mml:mi>F</mml:mi>
</mml:mrow>
</mml:msub>
<mml:mfenced open="(" close=")">
<mml:mrow>
<mml:mi>&#x3b7;</mml:mi>
</mml:mrow>
</mml:mfenced>
</mml:mtd>
<mml:mtd columnalign="left">
<mml:mo>&#x3d;</mml:mo>
<mml:msup>
<mml:mrow>
<mml:mi>&#x3bd;</mml:mi>
</mml:mrow>
<mml:mrow>
<mml:mo>&#x2a;</mml:mo>
</mml:mrow>
</mml:msup>
<mml:mfenced open="(" close=")">
<mml:mrow>
<mml:msub>
<mml:mrow>
<mml:mi mathvariant="script">S</mml:mi>
</mml:mrow>
<mml:mrow>
<mml:mi>F</mml:mi>
</mml:mrow>
</mml:msub>
<mml:mfenced open="(" close=")">
<mml:mrow>
<mml:msub>
<mml:mrow>
<mml:mi>&#x3bc;</mml:mi>
</mml:mrow>
<mml:mrow>
<mml:mn>1</mml:mn>
</mml:mrow>
</mml:msub>
<mml:mo>,</mml:mo>
<mml:msub>
<mml:mrow>
<mml:mi>&#x3bc;</mml:mi>
</mml:mrow>
<mml:mrow>
<mml:mn>2</mml:mn>
</mml:mrow>
</mml:msub>
<mml:mo>;</mml:mo>
<mml:mi>&#x3b7;</mml:mi>
</mml:mrow>
</mml:mfenced>
</mml:mrow>
</mml:mfenced>
</mml:mtd>
</mml:mtr>
<mml:mtr>
<mml:mtd columnalign="right">
<mml:mi mathvariant="script">G</mml:mi>
<mml:mfenced open="(" close=")">
<mml:mrow>
<mml:msub>
<mml:mrow>
<mml:mi>&#x3bc;</mml:mi>
</mml:mrow>
<mml:mrow>
<mml:mn>1</mml:mn>
</mml:mrow>
</mml:msub>
<mml:mo>,</mml:mo>
<mml:msub>
<mml:mrow>
<mml:mi>&#x3bc;</mml:mi>
</mml:mrow>
<mml:mrow>
<mml:mn>2</mml:mn>
</mml:mrow>
</mml:msub>
<mml:mo>;</mml:mo>
<mml:mi>&#x3b8;</mml:mi>
<mml:mo>,</mml:mo>
<mml:mi>&#x3b7;</mml:mi>
</mml:mrow>
</mml:mfenced>
</mml:mtd>
<mml:mtd columnalign="left">
<mml:mo>&#x3d;</mml:mo>
<mml:mfenced open="(" close=")">
<mml:mrow>
<mml:msubsup>
<mml:mrow>
<mml:mi mathvariant="script">S</mml:mi>
</mml:mrow>
<mml:mrow>
<mml:mi>F</mml:mi>
</mml:mrow>
<mml:mrow>
<mml:mi>c</mml:mi>
</mml:mrow>
</mml:msubsup>
<mml:mfenced open="(" close=")">
<mml:mrow>
<mml:mi>&#x3b8;</mml:mi>
</mml:mrow>
</mml:mfenced>
<mml:mo>&#x2229;</mml:mo>
<mml:msub>
<mml:mrow>
<mml:mi mathvariant="script">S</mml:mi>
</mml:mrow>
<mml:mrow>
<mml:mi>B</mml:mi>
</mml:mrow>
</mml:msub>
<mml:mfenced open="(" close=")">
<mml:mrow>
<mml:mi>&#x3b7;</mml:mi>
</mml:mrow>
</mml:mfenced>
</mml:mrow>
</mml:mfenced>
<mml:mo>&#x222a;</mml:mo>
<mml:mfenced open="(" close=")">
<mml:mrow>
<mml:msubsup>
<mml:mrow>
<mml:mi mathvariant="script">S</mml:mi>
</mml:mrow>
<mml:mrow>
<mml:mi>B</mml:mi>
</mml:mrow>
<mml:mrow>
<mml:mi>c</mml:mi>
</mml:mrow>
</mml:msubsup>
<mml:mfenced open="(" close=")">
<mml:mrow>
<mml:mi>&#x3b8;</mml:mi>
</mml:mrow>
</mml:mfenced>
<mml:mo>&#x2229;</mml:mo>
<mml:msub>
<mml:mrow>
<mml:mi mathvariant="script">S</mml:mi>
</mml:mrow>
<mml:mrow>
<mml:mi>F</mml:mi>
</mml:mrow>
</mml:msub>
<mml:mfenced open="(" close=")">
<mml:mrow>
<mml:mi>&#x3b7;</mml:mi>
</mml:mrow>
</mml:mfenced>
</mml:mrow>
</mml:mfenced>
<mml:mo>,</mml:mo>
</mml:mtd>
<mml:mtd columnalign="right">
<mml:mi>&#x3c8;</mml:mi>
<mml:mfenced open="(" close=")">
<mml:mrow>
<mml:mi>&#x3b8;</mml:mi>
<mml:mo>,</mml:mo>
<mml:mi>&#x3b7;</mml:mi>
</mml:mrow>
</mml:mfenced>
</mml:mtd>
<mml:mtd columnalign="left">
<mml:mo>&#x3d;</mml:mo>
<mml:msup>
<mml:mrow>
<mml:mi>&#x3bd;</mml:mi>
</mml:mrow>
<mml:mrow>
<mml:mo>&#x2a;</mml:mo>
</mml:mrow>
</mml:msup>
<mml:mfenced open="(" close=")">
<mml:mrow>
<mml:mi mathvariant="script">G</mml:mi>
<mml:mfenced open="(" close=")">
<mml:mrow>
<mml:msub>
<mml:mrow>
<mml:mi>&#x3bc;</mml:mi>
</mml:mrow>
<mml:mrow>
<mml:mn>1</mml:mn>
</mml:mrow>
</mml:msub>
<mml:mo>,</mml:mo>
<mml:msub>
<mml:mrow>
<mml:mi>&#x3bc;</mml:mi>
</mml:mrow>
<mml:mrow>
<mml:mn>2</mml:mn>
</mml:mrow>
</mml:msub>
<mml:mo>;</mml:mo>
<mml:mi>&#x3b8;</mml:mi>
<mml:mo>,</mml:mo>
<mml:mi>&#x3b7;</mml:mi>
</mml:mrow>
</mml:mfenced>
</mml:mrow>
</mml:mfenced>
</mml:mtd>
</mml:mtr>
</mml:mtable>
</mml:math>
<label>(9)</label>
</disp-formula>
</p>
</statement>
</p>
<p>
<statement content-type="theorem" id="Theorem_2">
<label>Theorem 2</label>
<p>
<italic>Let</italic> <inline-formula id="inf40">
<mml:math id="m53">
<mml:mi>&#x3b4;</mml:mi>
<mml:mo>&#x2208;</mml:mo>
<mml:mrow>
<mml:mo stretchy="false">(</mml:mo>
<mml:mrow>
<mml:mn>0</mml:mn>
<mml:mo>,</mml:mo>
<mml:mfrac>
<mml:mrow>
<mml:mn>1</mml:mn>
</mml:mrow>
<mml:mrow>
<mml:mn>2</mml:mn>
</mml:mrow>
</mml:mfrac>
</mml:mrow>
<mml:mo stretchy="false">)</mml:mo>
</mml:mrow>
</mml:math>
</inline-formula>
<italic>,</italic> <inline-formula id="inf41">
<mml:math id="m54">
<mml:msub>
<mml:mrow>
<mml:mi>&#x3bc;</mml:mi>
</mml:mrow>
<mml:mrow>
<mml:mi>j</mml:mi>
</mml:mrow>
</mml:msub>
<mml:mo>&#x3d;</mml:mo>
<mml:mi>&#x3b4;</mml:mi>
<mml:msub>
<mml:mrow>
<mml:mi>&#x3bc;</mml:mi>
</mml:mrow>
<mml:mrow>
<mml:mi>j</mml:mi>
<mml:mo>,</mml:mo>
<mml:mi>F</mml:mi>
</mml:mrow>
</mml:msub>
<mml:mo>&#x2b;</mml:mo>
<mml:mrow>
<mml:mo stretchy="false">(</mml:mo>
<mml:mrow>
<mml:mn>1</mml:mn>
<mml:mo>&#x2212;</mml:mo>
<mml:mi>&#x3b4;</mml:mi>
</mml:mrow>
<mml:mo stretchy="false">)</mml:mo>
</mml:mrow>
<mml:msub>
<mml:mrow>
<mml:mi>&#x3bc;</mml:mi>
</mml:mrow>
<mml:mrow>
<mml:mi>j</mml:mi>
<mml:mo>,</mml:mo>
<mml:mi>B</mml:mi>
</mml:mrow>
</mml:msub>
<mml:mo>&#x2208;</mml:mo>
<mml:msub>
<mml:mrow>
<mml:mi mathvariant="double-struck">P</mml:mi>
</mml:mrow>
<mml:mrow>
<mml:mi>&#x3b4;</mml:mi>
</mml:mrow>
</mml:msub>
<mml:mspace width="0.3333em"/>
<mml:mrow>
<mml:mo stretchy="false">(</mml:mo>
<mml:mrow>
<mml:mi>j</mml:mi>
<mml:mo>&#x3d;</mml:mo>
<mml:mn>1,2</mml:mn>
</mml:mrow>
<mml:mo stretchy="false">)</mml:mo>
</mml:mrow>
</mml:math>
</inline-formula>
<italic>.</italic>
<list list-type="simple">
<list-item>
<p>a) <italic>If</italic> <italic>&#x3b7;</italic> &#x3e; 0 <italic>and</italic> <italic>&#x3bc;</italic>
<sub>1,<italic>F</italic>
</sub> &#x3d; <italic>&#x3bc;</italic>
<sub>2,<italic>F</italic>
</sub> <italic>then for any</italic> <italic>&#x3b1;</italic> &#x2264; <italic>&#x3d5;</italic>
<sub>
<italic>B</italic>
</sub>(<italic>&#x3b7;</italic>)<italic>, we have</italic> kdiff (<italic>&#x3bc;</italic>
<sub>1</sub>,&#x20;<italic>&#x3bc;</italic>
<sub>2</sub>; <italic>&#x3b1;</italic>) &#x2264; (1 &#x2212; <italic>&#x3b4;</italic>)<italic>&#x3b7;</italic>
<italic>.</italic>
</p>
</list-item>
<list-item>
<p>b) <italic>If</italic> <italic>&#x3b7;</italic> &#x3e; 0 <italic>such that</italic> <inline-formula id="inf42">
<mml:math id="m55">
<mml:msub>
<mml:mrow>
<mml:mi>&#x3d5;</mml:mi>
</mml:mrow>
<mml:mrow>
<mml:mi>F</mml:mi>
</mml:mrow>
</mml:msub>
<mml:mfenced open="(" close=")">
<mml:mrow>
<mml:mfrac>
<mml:mrow>
<mml:mn>3</mml:mn>
<mml:mrow>
<mml:mo stretchy="false">(</mml:mo>
<mml:mrow>
<mml:mn>1</mml:mn>
<mml:mo>&#x2212;</mml:mo>
<mml:mi>&#x3b4;</mml:mi>
</mml:mrow>
<mml:mo stretchy="false">)</mml:mo>
</mml:mrow>
</mml:mrow>
<mml:mrow>
<mml:mi>&#x3b4;</mml:mi>
</mml:mrow>
</mml:mfrac>
<mml:mi>&#x3b7;</mml:mi>
</mml:mrow>
</mml:mfenced>
<mml:mo>&#x3c;</mml:mo>
<mml:mn>1</mml:mn>
</mml:math>
</inline-formula> <italic>and</italic> <inline-formula id="inf43">
<mml:math id="m56">
<mml:mi>&#x3c8;</mml:mi>
<mml:mfenced open="(" close=")">
<mml:mrow>
<mml:mfrac>
<mml:mrow>
<mml:mn>3</mml:mn>
<mml:mrow>
<mml:mo stretchy="false">(</mml:mo>
<mml:mrow>
<mml:mn>1</mml:mn>
<mml:mo>&#x2212;</mml:mo>
<mml:mi>&#x3b4;</mml:mi>
</mml:mrow>
<mml:mo stretchy="false">)</mml:mo>
</mml:mrow>
</mml:mrow>
<mml:mrow>
<mml:mi>&#x3b4;</mml:mi>
</mml:mrow>
</mml:mfrac>
<mml:mi>&#x3b7;</mml:mi>
<mml:mo>,</mml:mo>
<mml:mi>&#x3b7;</mml:mi>
</mml:mrow>
</mml:mfenced>
<mml:mo>&#x3e;</mml:mo>
<mml:mn>0</mml:mn>
</mml:math>
</inline-formula>
<italic>, then</italic> <italic>&#x3bc;</italic>
<sub>1,<italic>F</italic>
</sub> &#x2260; <italic>&#x3bc;</italic>
<sub>2,<italic>F</italic>
</sub> <italic>and for any</italic> <italic>&#x3b1;</italic> <italic>with</italic> <inline-formula id="inf44">
<mml:math id="m57">
<mml:mn>1</mml:mn>
<mml:mo>&#x2212;</mml:mo>
<mml:mi>&#x3c8;</mml:mi>
<mml:mfenced open="(" close=")">
<mml:mrow>
<mml:mfrac>
<mml:mrow>
<mml:mn>3</mml:mn>
<mml:mrow>
<mml:mo stretchy="false">(</mml:mo>
<mml:mrow>
<mml:mn>1</mml:mn>
<mml:mo>&#x2212;</mml:mo>
<mml:mi>&#x3b4;</mml:mi>
</mml:mrow>
<mml:mo stretchy="false">)</mml:mo>
</mml:mrow>
</mml:mrow>
<mml:mrow>
<mml:mi>&#x3b4;</mml:mi>
</mml:mrow>
</mml:mfrac>
<mml:mi>&#x3b7;</mml:mi>
<mml:mo>,</mml:mo>
<mml:mi>&#x3b7;</mml:mi>
</mml:mrow>
</mml:mfenced>
<mml:mo>&#x2264;</mml:mo>
<mml:mi>&#x3b1;</mml:mi>
</mml:math>
</inline-formula>
<italic>, we have</italic> kdiff (<italic>&#x3bc;</italic>
<sub>1</sub>, <italic>&#x3bc;</italic>
<sub>2</sub>; <italic>&#x3b1;</italic>) &#x2265; 2 (1 &#x2212; <italic>&#x3b4;</italic>)<italic>&#x3b7;</italic>
<italic>.</italic>
</p>
</list-item>
</list>
</p>
</statement>
</p>
<p>
<statement content-type="proof" id="uProof_1">
<label>Proof</label>
<p>To prove part (a), we observe that since <italic>&#x3bc;</italic>
<sub>1,<italic>F</italic>
</sub> &#x3d; <italic>&#x3bc;</italic>
<sub>2,<italic>F</italic>
</sub>, <italic>T</italic>(<italic>&#x3bc;</italic>
<sub>1</sub> &#x2212; <italic>&#x3bc;</italic>
<sub>2</sub>) (<italic>z</italic>) &#x3d; (1 &#x2212; <italic>&#x3b4;</italic>) &#x22c5; <italic>T</italic>(<italic>&#x3bc;</italic>
<sub>1,<italic>B</italic>
</sub> &#x2212; <italic>&#x3bc;</italic>
<sub>2,<italic>B</italic>
</sub>) (<italic>z</italic>) for all <inline-formula id="inf45">
<mml:math id="m58">
<mml:mi>z</mml:mi>
<mml:mo>&#x2208;</mml:mo>
<mml:mi mathvariant="double-struck">X</mml:mi>
</mml:math>
</inline-formula>. By definition (9),<disp-formula id="equ3">
<mml:math id="m59">
<mml:msub>
<mml:mrow>
<mml:mi mathvariant="script">S</mml:mi>
</mml:mrow>
<mml:mrow>
<mml:mi>B</mml:mi>
</mml:mrow>
</mml:msub>
<mml:mfenced open="(" close=")">
<mml:mrow>
<mml:msub>
<mml:mrow>
<mml:mi>&#x3bc;</mml:mi>
</mml:mrow>
<mml:mrow>
<mml:mn>1</mml:mn>
</mml:mrow>
</mml:msub>
<mml:mo>,</mml:mo>
<mml:msub>
<mml:mrow>
<mml:mi>&#x3bc;</mml:mi>
</mml:mrow>
<mml:mrow>
<mml:mn>2</mml:mn>
</mml:mrow>
</mml:msub>
<mml:mo>;</mml:mo>
<mml:mi>&#x3b7;</mml:mi>
</mml:mrow>
</mml:mfenced>
<mml:mo>&#x3d;</mml:mo>
<mml:mfenced open="{" close="}">
<mml:mrow>
<mml:mi>z</mml:mi>
<mml:mo>&#x2208;</mml:mo>
<mml:mi mathvariant="double-struck">X</mml:mi>
<mml:mo>:</mml:mo>
<mml:mi>T</mml:mi>
<mml:mfenced open="(" close=")">
<mml:mrow>
<mml:msub>
<mml:mrow>
<mml:mi>&#x3bc;</mml:mi>
</mml:mrow>
<mml:mrow>
<mml:mn>1</mml:mn>
<mml:mo>,</mml:mo>
<mml:mi>B</mml:mi>
</mml:mrow>
</mml:msub>
<mml:mo>&#x2212;</mml:mo>
<mml:msub>
<mml:mrow>
<mml:mi>&#x3bc;</mml:mi>
</mml:mrow>
<mml:mrow>
<mml:mn>2</mml:mn>
<mml:mo>,</mml:mo>
<mml:mi>B</mml:mi>
</mml:mrow>
</mml:msub>
</mml:mrow>
</mml:mfenced>
<mml:mfenced open="(" close=")">
<mml:mrow>
<mml:mi>z</mml:mi>
</mml:mrow>
</mml:mfenced>
<mml:mo>&#x3c;</mml:mo>
<mml:mi>&#x3b7;</mml:mi>
</mml:mrow>
</mml:mfenced>
<mml:mo>&#x3d;</mml:mo>
<mml:mfenced open="{" close="}">
<mml:mrow>
<mml:mi>z</mml:mi>
<mml:mo>&#x2208;</mml:mo>
<mml:mi mathvariant="double-struck">X</mml:mi>
<mml:mo>:</mml:mo>
<mml:mi>T</mml:mi>
<mml:mfenced open="(" close=")">
<mml:mrow>
<mml:msub>
<mml:mrow>
<mml:mi>&#x3bc;</mml:mi>
</mml:mrow>
<mml:mrow>
<mml:mn>1</mml:mn>
</mml:mrow>
</mml:msub>
<mml:mo>&#x2212;</mml:mo>
<mml:msub>
<mml:mrow>
<mml:mi>&#x3bc;</mml:mi>
</mml:mrow>
<mml:mrow>
<mml:mn>2</mml:mn>
</mml:mrow>
</mml:msub>
</mml:mrow>
</mml:mfenced>
<mml:mfenced open="(" close=")">
<mml:mrow>
<mml:mi>z</mml:mi>
</mml:mrow>
</mml:mfenced>
<mml:mo>&#x3c;</mml:mo>
<mml:mfenced open="(" close=")">
<mml:mrow>
<mml:mn>1</mml:mn>
<mml:mo>&#x2212;</mml:mo>
<mml:mi>&#x3b4;</mml:mi>
</mml:mrow>
</mml:mfenced>
<mml:mi>&#x3b7;</mml:mi>
</mml:mrow>
</mml:mfenced>
<mml:mo>.</mml:mo>
</mml:math>
</disp-formula>
</p>
<p>Therefore, <italic>&#x3d5;</italic>
<sub>
<italic>B</italic>
</sub>(<italic>&#x3b7;</italic>) &#x2264; &#x39b;(<italic>T</italic> (<italic>&#x3bc;</italic>
<sub>1</sub> &#x2212; <italic>&#x3bc;</italic>
<sub>2</sub>)) ((1 &#x2212; <italic>&#x3b4;</italic>)<italic>&#x3b7;</italic>). In view of <xref ref-type="disp-formula" rid="e8">(8)</xref>, this proves part&#x20;(a).</p>
<p>To prove part (b), we will write<disp-formula id="equ4">
<mml:math id="m60">
<mml:mi>&#x3b8;</mml:mi>
<mml:mo>&#x3d;</mml:mo>
<mml:mfrac>
<mml:mrow>
<mml:mn>3</mml:mn>
<mml:mfenced open="(" close=")">
<mml:mrow>
<mml:mn>1</mml:mn>
<mml:mo>&#x2212;</mml:mo>
<mml:mi>&#x3b4;</mml:mi>
</mml:mrow>
</mml:mfenced>
</mml:mrow>
<mml:mrow>
<mml:mi>&#x3b4;</mml:mi>
</mml:mrow>
</mml:mfrac>
<mml:mi>&#x3b7;</mml:mi>
<mml:mo>.</mml:mo>
</mml:math>
</disp-formula>
</p>
<p>Our hypothesis that <italic>&#x3d5;</italic>
<sub>
<italic>F</italic>
</sub>(<italic>&#x3b8;</italic>) &#x3c; 1 means that <italic>&#x3bc;</italic>
<sub>1,<italic>F</italic>
</sub> &#x2260; <italic>&#x3bc;</italic>
<sub>2,<italic>F</italic>
</sub> and <inline-formula id="inf46">
<mml:math id="m61">
<mml:msubsup>
<mml:mrow>
<mml:mi mathvariant="script">S</mml:mi>
</mml:mrow>
<mml:mrow>
<mml:mi>F</mml:mi>
</mml:mrow>
<mml:mrow>
<mml:mi>c</mml:mi>
</mml:mrow>
</mml:msubsup>
<mml:mrow>
<mml:mo stretchy="false">(</mml:mo>
<mml:mrow>
<mml:mi>&#x3b8;</mml:mi>
</mml:mrow>
<mml:mo stretchy="false">)</mml:mo>
</mml:mrow>
</mml:math>
</inline-formula> is nonempty. For all <inline-formula id="inf47">
<mml:math id="m62">
<mml:mi>z</mml:mi>
<mml:mo>&#x2208;</mml:mo>
<mml:msubsup>
<mml:mrow>
<mml:mi mathvariant="script">S</mml:mi>
</mml:mrow>
<mml:mrow>
<mml:mi>F</mml:mi>
</mml:mrow>
<mml:mrow>
<mml:mi>c</mml:mi>
</mml:mrow>
</mml:msubsup>
<mml:mrow>
<mml:mo stretchy="false">(</mml:mo>
<mml:mrow>
<mml:mi>&#x3b8;</mml:mi>
</mml:mrow>
<mml:mo stretchy="false">)</mml:mo>
</mml:mrow>
<mml:mo>&#x2229;</mml:mo>
<mml:msub>
<mml:mrow>
<mml:mi mathvariant="script">S</mml:mi>
</mml:mrow>
<mml:mrow>
<mml:mi>B</mml:mi>
</mml:mrow>
</mml:msub>
<mml:mrow>
<mml:mo stretchy="false">(</mml:mo>
<mml:mrow>
<mml:mi>&#x3b7;</mml:mi>
</mml:mrow>
<mml:mo stretchy="false">)</mml:mo>
</mml:mrow>
</mml:math>
</inline-formula>,<disp-formula id="equ5">
<mml:math id="m63">
<mml:mi>T</mml:mi>
<mml:mfenced open="(" close=")">
<mml:mrow>
<mml:msub>
<mml:mrow>
<mml:mi>&#x3bc;</mml:mi>
</mml:mrow>
<mml:mrow>
<mml:mn>1</mml:mn>
</mml:mrow>
</mml:msub>
<mml:mo>&#x2212;</mml:mo>
<mml:msub>
<mml:mrow>
<mml:mi>&#x3bc;</mml:mi>
</mml:mrow>
<mml:mrow>
<mml:mn>2</mml:mn>
</mml:mrow>
</mml:msub>
</mml:mrow>
</mml:mfenced>
<mml:mfenced open="(" close=")">
<mml:mrow>
<mml:mi>z</mml:mi>
</mml:mrow>
</mml:mfenced>
<mml:mo>&#x2265;</mml:mo>
<mml:mi>&#x3b4;</mml:mi>
<mml:mo stretchy="false">&#x7c;</mml:mo>
<mml:mi>U</mml:mi>
<mml:mfenced open="(" close=")">
<mml:mrow>
<mml:msub>
<mml:mrow>
<mml:mi>&#x3bc;</mml:mi>
</mml:mrow>
<mml:mrow>
<mml:mn>1</mml:mn>
<mml:mo>,</mml:mo>
<mml:mi>F</mml:mi>
</mml:mrow>
</mml:msub>
<mml:mo>&#x2212;</mml:mo>
<mml:msub>
<mml:mrow>
<mml:mi>&#x3bc;</mml:mi>
</mml:mrow>
<mml:mrow>
<mml:mn>2</mml:mn>
<mml:mo>,</mml:mo>
<mml:mi>F</mml:mi>
</mml:mrow>
</mml:msub>
</mml:mrow>
</mml:mfenced>
<mml:mfenced open="(" close=")">
<mml:mrow>
<mml:mi>z</mml:mi>
</mml:mrow>
</mml:mfenced>
<mml:mo stretchy="false">&#x7c;</mml:mo>
<mml:mo>&#x2212;</mml:mo>
<mml:mfenced open="(" close=")">
<mml:mrow>
<mml:mn>1</mml:mn>
<mml:mo>&#x2212;</mml:mo>
<mml:mi>&#x3b4;</mml:mi>
</mml:mrow>
</mml:mfenced>
<mml:mo stretchy="false">&#x7c;</mml:mo>
<mml:mi>U</mml:mi>
<mml:mfenced open="(" close=")">
<mml:mrow>
<mml:msub>
<mml:mrow>
<mml:mi>&#x3bc;</mml:mi>
</mml:mrow>
<mml:mrow>
<mml:mn>1</mml:mn>
<mml:mo>,</mml:mo>
<mml:mi>B</mml:mi>
</mml:mrow>
</mml:msub>
<mml:mo>&#x2212;</mml:mo>
<mml:msub>
<mml:mrow>
<mml:mi>&#x3bc;</mml:mi>
</mml:mrow>
<mml:mrow>
<mml:mn>2</mml:mn>
<mml:mo>,</mml:mo>
<mml:mi>B</mml:mi>
</mml:mrow>
</mml:msub>
</mml:mrow>
</mml:mfenced>
<mml:mfenced open="(" close=")">
<mml:mrow>
<mml:mi>z</mml:mi>
</mml:mrow>
</mml:mfenced>
<mml:mo stretchy="false">&#x7c;</mml:mo>
<mml:mo>&#x3e;</mml:mo>
<mml:mi>&#x3b4;</mml:mi>
<mml:mi>&#x3b8;</mml:mi>
<mml:mo>&#x2212;</mml:mo>
<mml:mfenced open="(" close=")">
<mml:mrow>
<mml:mn>1</mml:mn>
<mml:mo>&#x2212;</mml:mo>
<mml:mi>&#x3b4;</mml:mi>
</mml:mrow>
</mml:mfenced>
<mml:mi>&#x3b7;</mml:mi>
<mml:mo>&#x2265;</mml:mo>
<mml:mn>2</mml:mn>
<mml:mfenced open="(" close=")">
<mml:mrow>
<mml:mn>1</mml:mn>
<mml:mo>&#x2212;</mml:mo>
<mml:mi>&#x3b4;</mml:mi>
</mml:mrow>
</mml:mfenced>
<mml:mi>&#x3b7;</mml:mi>
<mml:mo>.</mml:mo>
</mml:math>
</disp-formula>
</p>
<p>Moreover, for <inline-formula id="inf48">
<mml:math id="m64">
<mml:mi>z</mml:mi>
<mml:mo>&#x2208;</mml:mo>
<mml:msubsup>
<mml:mrow>
<mml:mi mathvariant="script">S</mml:mi>
</mml:mrow>
<mml:mrow>
<mml:mi>B</mml:mi>
</mml:mrow>
<mml:mrow>
<mml:mi>c</mml:mi>
</mml:mrow>
</mml:msubsup>
<mml:mrow>
<mml:mo stretchy="false">(</mml:mo>
<mml:mrow>
<mml:mi>&#x3b8;</mml:mi>
</mml:mrow>
<mml:mo stretchy="false">)</mml:mo>
</mml:mrow>
<mml:mo>&#x2229;</mml:mo>
<mml:msub>
<mml:mrow>
<mml:mi mathvariant="script">S</mml:mi>
</mml:mrow>
<mml:mrow>
<mml:mi>F</mml:mi>
</mml:mrow>
</mml:msub>
<mml:mrow>
<mml:mo stretchy="false">(</mml:mo>
<mml:mrow>
<mml:mi>&#x3b7;</mml:mi>
</mml:mrow>
<mml:mo stretchy="false">)</mml:mo>
</mml:mrow>
</mml:math>
</inline-formula>, we also know that<disp-formula id="equ6">
<mml:math id="m65">
<mml:mi>T</mml:mi>
<mml:mfenced open="(" close=")">
<mml:mrow>
<mml:msub>
<mml:mrow>
<mml:mi>&#x3bc;</mml:mi>
</mml:mrow>
<mml:mrow>
<mml:mn>1</mml:mn>
</mml:mrow>
</mml:msub>
<mml:mo>&#x2212;</mml:mo>
<mml:msub>
<mml:mrow>
<mml:mi>&#x3bc;</mml:mi>
</mml:mrow>
<mml:mrow>
<mml:mn>2</mml:mn>
</mml:mrow>
</mml:msub>
</mml:mrow>
</mml:mfenced>
<mml:mfenced open="(" close=")">
<mml:mrow>
<mml:mi>z</mml:mi>
</mml:mrow>
</mml:mfenced>
<mml:mo>&#x2265;</mml:mo>
<mml:mfenced open="(" close=")">
<mml:mrow>
<mml:mn>1</mml:mn>
<mml:mo>&#x2212;</mml:mo>
<mml:mi>&#x3b4;</mml:mi>
</mml:mrow>
</mml:mfenced>
<mml:mo stretchy="false">&#x7c;</mml:mo>
<mml:mi>U</mml:mi>
<mml:mfenced open="(" close=")">
<mml:mrow>
<mml:msub>
<mml:mrow>
<mml:mi>&#x3bc;</mml:mi>
</mml:mrow>
<mml:mrow>
<mml:mn>1</mml:mn>
<mml:mo>,</mml:mo>
<mml:mi>B</mml:mi>
</mml:mrow>
</mml:msub>
<mml:mo>&#x2212;</mml:mo>
<mml:msub>
<mml:mrow>
<mml:mi>&#x3bc;</mml:mi>
</mml:mrow>
<mml:mrow>
<mml:mn>2</mml:mn>
<mml:mo>,</mml:mo>
<mml:mi>B</mml:mi>
</mml:mrow>
</mml:msub>
</mml:mrow>
</mml:mfenced>
<mml:mfenced open="(" close=")">
<mml:mrow>
<mml:mi>z</mml:mi>
</mml:mrow>
</mml:mfenced>
<mml:mo stretchy="false">&#x7c;</mml:mo>
<mml:mo>&#x2212;</mml:mo>
<mml:mi>&#x3b4;</mml:mi>
<mml:mo stretchy="false">&#x7c;</mml:mo>
<mml:mi>U</mml:mi>
<mml:mfenced open="(" close=")">
<mml:mrow>
<mml:msub>
<mml:mrow>
<mml:mi>&#x3bc;</mml:mi>
</mml:mrow>
<mml:mrow>
<mml:mn>1</mml:mn>
<mml:mo>,</mml:mo>
<mml:mi>F</mml:mi>
</mml:mrow>
</mml:msub>
<mml:mo>&#x2212;</mml:mo>
<mml:msub>
<mml:mrow>
<mml:mi>&#x3bc;</mml:mi>
</mml:mrow>
<mml:mrow>
<mml:mn>2</mml:mn>
<mml:mo>,</mml:mo>
<mml:mi>F</mml:mi>
</mml:mrow>
</mml:msub>
</mml:mrow>
</mml:mfenced>
<mml:mfenced open="(" close=")">
<mml:mrow>
<mml:mi>z</mml:mi>
</mml:mrow>
</mml:mfenced>
<mml:mo stretchy="false">&#x7c;</mml:mo>
<mml:mo>&#x2265;</mml:mo>
<mml:mfenced open="(" close=")">
<mml:mrow>
<mml:mn>1</mml:mn>
<mml:mo>&#x2212;</mml:mo>
<mml:mi>&#x3b4;</mml:mi>
</mml:mrow>
</mml:mfenced>
<mml:mi>&#x3b8;</mml:mi>
<mml:mo>&#x2212;</mml:mo>
<mml:mi>&#x3b4;</mml:mi>
<mml:mi>&#x3b7;</mml:mi>
<mml:mo>&#x2265;</mml:mo>
<mml:mfrac>
<mml:mrow>
<mml:mn>3</mml:mn>
<mml:msup>
<mml:mrow>
<mml:mfenced open="(" close=")">
<mml:mrow>
<mml:mn>1</mml:mn>
<mml:mo>&#x2212;</mml:mo>
<mml:mi>&#x3b4;</mml:mi>
</mml:mrow>
</mml:mfenced>
</mml:mrow>
<mml:mrow>
<mml:mn>2</mml:mn>
</mml:mrow>
</mml:msup>
<mml:mo>&#x2212;</mml:mo>
<mml:msup>
<mml:mrow>
<mml:mi>&#x3b4;</mml:mi>
</mml:mrow>
<mml:mrow>
<mml:mn>2</mml:mn>
</mml:mrow>
</mml:msup>
</mml:mrow>
<mml:mrow>
<mml:mi>&#x3b4;</mml:mi>
</mml:mrow>
</mml:mfrac>
<mml:mi>&#x3b7;</mml:mi>
<mml:mo>.</mml:mo>
</mml:math>
</disp-formula>
</p>
<p>Note that because <inline-formula id="inf49">
<mml:math id="m66">
<mml:mi>&#x3b4;</mml:mi>
<mml:mo>&#x3c;</mml:mo>
<mml:mfrac>
<mml:mrow>
<mml:mn>1</mml:mn>
</mml:mrow>
<mml:mrow>
<mml:mn>2</mml:mn>
</mml:mrow>
</mml:mfrac>
</mml:math>
</inline-formula>, we have <inline-formula id="inf50">
<mml:math id="m67">
<mml:mfrac>
<mml:mrow>
<mml:mn>3</mml:mn>
<mml:msup>
<mml:mrow>
<mml:mrow>
<mml:mo stretchy="false">(</mml:mo>
<mml:mrow>
<mml:mn>1</mml:mn>
<mml:mo>&#x2212;</mml:mo>
<mml:mi>&#x3b4;</mml:mi>
</mml:mrow>
<mml:mo stretchy="false">)</mml:mo>
</mml:mrow>
</mml:mrow>
<mml:mrow>
<mml:mn>2</mml:mn>
</mml:mrow>
</mml:msup>
<mml:mo>&#x2212;</mml:mo>
<mml:msup>
<mml:mrow>
<mml:mi>&#x3b4;</mml:mi>
</mml:mrow>
<mml:mrow>
<mml:mn>2</mml:mn>
</mml:mrow>
</mml:msup>
</mml:mrow>
<mml:mrow>
<mml:mi>&#x3b4;</mml:mi>
</mml:mrow>
</mml:mfrac>
<mml:mo>&#x3e;</mml:mo>
<mml:mn>2</mml:mn>
<mml:mrow>
<mml:mo stretchy="false">(</mml:mo>
<mml:mrow>
<mml:mn>1</mml:mn>
<mml:mo>&#x2212;</mml:mo>
<mml:mi>&#x3b4;</mml:mi>
</mml:mrow>
<mml:mo stretchy="false">)</mml:mo>
</mml:mrow>
</mml:math>
</inline-formula>. So, this means that<disp-formula id="equ7">
<mml:math id="m68">
<mml:mfenced open="{" close="}">
<mml:mrow>
<mml:mi>z</mml:mi>
<mml:mo>&#x2208;</mml:mo>
<mml:mi mathvariant="double-struck">X</mml:mi>
<mml:mo>:</mml:mo>
<mml:mi>T</mml:mi>
<mml:mfenced open="(" close=")">
<mml:mrow>
<mml:msub>
<mml:mrow>
<mml:mi>&#x3bc;</mml:mi>
</mml:mrow>
<mml:mrow>
<mml:mn>1</mml:mn>
</mml:mrow>
</mml:msub>
<mml:mo>&#x2212;</mml:mo>
<mml:msub>
<mml:mrow>
<mml:mi>&#x3bc;</mml:mi>
</mml:mrow>
<mml:mrow>
<mml:mn>2</mml:mn>
</mml:mrow>
</mml:msub>
</mml:mrow>
</mml:mfenced>
<mml:mfenced open="(" close=")">
<mml:mrow>
<mml:mi>z</mml:mi>
</mml:mrow>
</mml:mfenced>
<mml:mo>&#x3c;</mml:mo>
<mml:mn>2</mml:mn>
<mml:mfenced open="(" close=")">
<mml:mrow>
<mml:mn>1</mml:mn>
<mml:mo>&#x2212;</mml:mo>
<mml:mi>&#x3b4;</mml:mi>
</mml:mrow>
</mml:mfenced>
<mml:mi>&#x3b7;</mml:mi>
</mml:mrow>
</mml:mfenced>
<mml:mo>&#x2282;</mml:mo>
<mml:mfenced open="{" close="}">
<mml:mrow>
<mml:mi>z</mml:mi>
<mml:mo>&#x2208;</mml:mo>
<mml:mi mathvariant="double-struck">X</mml:mi>
<mml:mo>:</mml:mo>
<mml:mi>z</mml:mi>
<mml:mo>&#x2209;</mml:mo>
<mml:mi mathvariant="script">G</mml:mi>
<mml:mfenced open="(" close=")">
<mml:mrow>
<mml:mi>&#x3b8;</mml:mi>
<mml:mo>,</mml:mo>
<mml:mi>&#x3b7;</mml:mi>
</mml:mrow>
</mml:mfenced>
</mml:mrow>
</mml:mfenced>
<mml:mo>;</mml:mo>
</mml:math>
</disp-formula>
</p>
<p>This means that &#x39b;(<italic>T</italic> (<italic>&#x3bc;</italic>
<sub>1</sub> &#x2212; <italic>&#x3bc;</italic>
<sub>2</sub>)) (2 (1 &#x2212; <italic>&#x3b4;</italic>)<italic>&#x3b7;</italic>) &#x2264; 1&#x20;&#x2212; <italic>&#x3c8;</italic>(<italic>&#x3b8;</italic>, <italic>&#x3b7;</italic>). Since <italic>&#x3b1;</italic> &#x2265; 1&#x20;&#x2212; <italic>&#x3c8;</italic>(<italic>&#x3b8;</italic>, <italic>&#x3b7;</italic>), this estimate together with <xref ref-type="disp-formula" rid="e8">(8)</xref> leads to the conclusion in part&#x20;(b).</p>
<p>We wish to comment on the practicality of the constants <inline-formula id="inf51">
<mml:math id="m69">
<mml:msub>
<mml:mrow>
<mml:mi mathvariant="script">S</mml:mi>
</mml:mrow>
<mml:mrow>
<mml:mi>F</mml:mi>
</mml:mrow>
</mml:msub>
<mml:mrow>
<mml:mo stretchy="false">(</mml:mo>
<mml:mrow>
<mml:msub>
<mml:mrow>
<mml:mi>&#x3bc;</mml:mi>
</mml:mrow>
<mml:mrow>
<mml:mn>1</mml:mn>
</mml:mrow>
</mml:msub>
<mml:mo>,</mml:mo>
<mml:msub>
<mml:mrow>
<mml:mi>&#x3bc;</mml:mi>
</mml:mrow>
<mml:mrow>
<mml:mn>2</mml:mn>
</mml:mrow>
</mml:msub>
<mml:mo>;</mml:mo>
<mml:mi>&#x3b7;</mml:mi>
</mml:mrow>
<mml:mo stretchy="false">)</mml:mo>
</mml:mrow>
</mml:math>
</inline-formula>, <inline-formula id="inf52">
<mml:math id="m70">
<mml:msub>
<mml:mrow>
<mml:mi mathvariant="script">S</mml:mi>
</mml:mrow>
<mml:mrow>
<mml:mi>B</mml:mi>
</mml:mrow>
</mml:msub>
<mml:mrow>
<mml:mo stretchy="false">(</mml:mo>
<mml:mrow>
<mml:msub>
<mml:mrow>
<mml:mi>&#x3bc;</mml:mi>
</mml:mrow>
<mml:mrow>
<mml:mn>1</mml:mn>
</mml:mrow>
</mml:msub>
<mml:mo>,</mml:mo>
<mml:msub>
<mml:mrow>
<mml:mi>&#x3bc;</mml:mi>
</mml:mrow>
<mml:mrow>
<mml:mn>2</mml:mn>
</mml:mrow>
</mml:msub>
<mml:mo>;</mml:mo>
<mml:mi>&#x3b7;</mml:mi>
</mml:mrow>
<mml:mo stretchy="false">)</mml:mo>
</mml:mrow>
</mml:math>
</inline-formula> and <inline-formula id="inf53">
<mml:math id="m71">
<mml:mi mathvariant="script">G</mml:mi>
<mml:mrow>
<mml:mo stretchy="false">(</mml:mo>
<mml:mrow>
<mml:msub>
<mml:mrow>
<mml:mi>&#x3bc;</mml:mi>
</mml:mrow>
<mml:mrow>
<mml:mn>1</mml:mn>
</mml:mrow>
</mml:msub>
<mml:mo>,</mml:mo>
<mml:msub>
<mml:mrow>
<mml:mi>&#x3bc;</mml:mi>
</mml:mrow>
<mml:mrow>
<mml:mn>2</mml:mn>
</mml:mrow>
</mml:msub>
<mml:mo>;</mml:mo>
<mml:mi>&#x3b8;</mml:mi>
<mml:mo>,</mml:mo>
<mml:mi>&#x3b7;</mml:mi>
</mml:mrow>
<mml:mo stretchy="false">)</mml:mo>
</mml:mrow>
</mml:math>
</inline-formula>. We consider this with the simple setting where <italic>K</italic> is a compactly supported localized kernel (e.g., indicator function of an <italic>&#x3f5;</italic>-ball) in order to avoid the discussion of tails. We define the well-separated setting as the setting where <inline-formula id="inf54">
<mml:math id="m72">
<mml:mi mathvariant="sans-serif">d</mml:mi>
<mml:mi mathvariant="sans-serif">i</mml:mi>
<mml:mi mathvariant="sans-serif">s</mml:mi>
<mml:mi mathvariant="sans-serif">t</mml:mi>
<mml:mrow>
<mml:mo stretchy="false">(</mml:mo>
<mml:mrow>
<mml:mi mathvariant="sans-serif">s</mml:mi>
<mml:mi mathvariant="sans-serif">u</mml:mi>
<mml:mi mathvariant="sans-serif">p</mml:mi>
<mml:mi mathvariant="sans-serif">p</mml:mi>
<mml:mrow>
<mml:mo stretchy="false">(</mml:mo>
<mml:mrow>
<mml:msub>
<mml:mrow>
<mml:mi>&#x3bc;</mml:mi>
</mml:mrow>
<mml:mrow>
<mml:mn>1</mml:mn>
<mml:mo>,</mml:mo>
<mml:mi>B</mml:mi>
</mml:mrow>
</mml:msub>
</mml:mrow>
<mml:mo stretchy="false">)</mml:mo>
</mml:mrow>
<mml:mo>,</mml:mo>
<mml:mi mathvariant="sans-serif">s</mml:mi>
<mml:mi mathvariant="sans-serif">u</mml:mi>
<mml:mi mathvariant="sans-serif">p</mml:mi>
<mml:mi mathvariant="sans-serif">p</mml:mi>
<mml:mrow>
<mml:mo stretchy="false">(</mml:mo>
<mml:mrow>
<mml:msub>
<mml:mrow>
<mml:mi>&#x3bc;</mml:mi>
</mml:mrow>
<mml:mrow>
<mml:mn>2</mml:mn>
<mml:mo>,</mml:mo>
<mml:mi>B</mml:mi>
</mml:mrow>
</mml:msub>
</mml:mrow>
<mml:mo stretchy="false">)</mml:mo>
</mml:mrow>
</mml:mrow>
<mml:mo stretchy="false">)</mml:mo>
</mml:mrow>
<mml:mo>&#x3e;</mml:mo>
<mml:mi>&#x3f5;</mml:mi>
</mml:math>
</inline-formula> and <inline-formula id="inf55">
<mml:math id="m73">
<mml:mi mathvariant="sans-serif">d</mml:mi>
<mml:mi mathvariant="sans-serif">i</mml:mi>
<mml:mi mathvariant="sans-serif">s</mml:mi>
<mml:mi mathvariant="sans-serif">t</mml:mi>
<mml:mrow>
<mml:mo stretchy="false">(</mml:mo>
<mml:mrow>
<mml:mi mathvariant="sans-serif">s</mml:mi>
<mml:mi mathvariant="sans-serif">u</mml:mi>
<mml:mi mathvariant="sans-serif">p</mml:mi>
<mml:mi mathvariant="sans-serif">p</mml:mi>
<mml:mrow>
<mml:mo stretchy="false">(</mml:mo>
<mml:mrow>
<mml:msub>
<mml:mrow>
<mml:mi>&#x3bc;</mml:mi>
</mml:mrow>
<mml:mrow>
<mml:mn>1</mml:mn>
<mml:mo>,</mml:mo>
<mml:mi>B</mml:mi>
</mml:mrow>
</mml:msub>
</mml:mrow>
<mml:mo stretchy="false">)</mml:mo>
</mml:mrow>
<mml:mo>,</mml:mo>
<mml:mi mathvariant="sans-serif">s</mml:mi>
<mml:mi mathvariant="sans-serif">u</mml:mi>
<mml:mi mathvariant="sans-serif">p</mml:mi>
<mml:mi mathvariant="sans-serif">p</mml:mi>
<mml:mrow>
<mml:mo stretchy="false">(</mml:mo>
<mml:mrow>
<mml:msub>
<mml:mrow>
<mml:mi>&#x3bc;</mml:mi>
</mml:mrow>
<mml:mrow>
<mml:mi>i</mml:mi>
<mml:mo>,</mml:mo>
<mml:mi>F</mml:mi>
</mml:mrow>
</mml:msub>
</mml:mrow>
<mml:mo stretchy="false">)</mml:mo>
</mml:mrow>
</mml:mrow>
<mml:mo stretchy="false">)</mml:mo>
</mml:mrow>
<mml:mo>&#x3e;</mml:mo>
<mml:mi>&#x3f5;</mml:mi>
</mml:math>
</inline-formula>. For part (b), we&#x2019;ll also use <inline-formula id="inf56">
<mml:math id="m74">
<mml:mi mathvariant="sans-serif">d</mml:mi>
<mml:mi mathvariant="sans-serif">i</mml:mi>
<mml:mi mathvariant="sans-serif">s</mml:mi>
<mml:mi mathvariant="sans-serif">t</mml:mi>
<mml:mrow>
<mml:mo stretchy="false">(</mml:mo>
<mml:mrow>
<mml:mi mathvariant="sans-serif">s</mml:mi>
<mml:mi mathvariant="sans-serif">u</mml:mi>
<mml:mi mathvariant="sans-serif">p</mml:mi>
<mml:mi mathvariant="sans-serif">p</mml:mi>
<mml:mrow>
<mml:mo stretchy="false">(</mml:mo>
<mml:mrow>
<mml:msub>
<mml:mrow>
<mml:mi>&#x3bc;</mml:mi>
</mml:mrow>
<mml:mrow>
<mml:mn>1</mml:mn>
<mml:mo>,</mml:mo>
<mml:mi>F</mml:mi>
</mml:mrow>
</mml:msub>
</mml:mrow>
<mml:mo stretchy="false">)</mml:mo>
</mml:mrow>
<mml:mo>,</mml:mo>
<mml:mi mathvariant="sans-serif">s</mml:mi>
<mml:mi mathvariant="sans-serif">u</mml:mi>
<mml:mi mathvariant="sans-serif">p</mml:mi>
<mml:mi mathvariant="sans-serif">p</mml:mi>
<mml:mrow>
<mml:mo stretchy="false">(</mml:mo>
<mml:mrow>
<mml:msub>
<mml:mrow>
<mml:mi>&#x3bc;</mml:mi>
</mml:mrow>
<mml:mrow>
<mml:mn>2</mml:mn>
<mml:mo>,</mml:mo>
<mml:mi>F</mml:mi>
</mml:mrow>
</mml:msub>
</mml:mrow>
<mml:mo stretchy="false">)</mml:mo>
</mml:mrow>
</mml:mrow>
<mml:mo stretchy="false">)</mml:mo>
</mml:mrow>
<mml:mo>&#x3e;</mml:mo>
<mml:mi>&#x3f5;</mml:mi>
</mml:math>
</inline-formula> and all four measures are sufficiently concentrated, i.e.,<disp-formula id="equ8">
<mml:math id="m75">
<mml:msub>
<mml:mrow>
<mml:mi>&#x3bc;</mml:mi>
</mml:mrow>
<mml:mrow>
<mml:mi>i</mml:mi>
<mml:mo>,</mml:mo>
<mml:mi>F</mml:mi>
</mml:mrow>
</mml:msub>
<mml:mfenced open="(" close=")">
<mml:mrow>
<mml:mfenced open="{" close="}">
<mml:mrow>
<mml:mi>z</mml:mi>
<mml:mo>&#x2208;</mml:mo>
<mml:mi mathvariant="double-struck">X</mml:mi>
<mml:mo>:</mml:mo>
<mml:mi>T</mml:mi>
<mml:mfenced open="(" close=")">
<mml:mrow>
<mml:msub>
<mml:mrow>
<mml:mi>&#x3bc;</mml:mi>
</mml:mrow>
<mml:mrow>
<mml:mi>i</mml:mi>
<mml:mo>,</mml:mo>
<mml:mi>F</mml:mi>
</mml:mrow>
</mml:msub>
</mml:mrow>
</mml:mfenced>
<mml:mfenced open="(" close=")">
<mml:mrow>
<mml:mi>z</mml:mi>
</mml:mrow>
</mml:mfenced>
<mml:mo>&#x2265;</mml:mo>
<mml:mi>&#x3b8;</mml:mi>
</mml:mrow>
</mml:mfenced>
</mml:mrow>
</mml:mfenced>
<mml:mo>&#x2265;</mml:mo>
<mml:mn>1</mml:mn>
<mml:mo>&#x2212;</mml:mo>
<mml:mi>&#x3be;</mml:mi>
<mml:mspace width="2em"/>
<mml:mtext>&#x2009;and&#x2009;</mml:mtext>
<mml:mspace width="2em"/>
<mml:msub>
<mml:mrow>
<mml:mi>&#x3bc;</mml:mi>
</mml:mrow>
<mml:mrow>
<mml:mi>i</mml:mi>
<mml:mo>,</mml:mo>
<mml:mi>B</mml:mi>
</mml:mrow>
</mml:msub>
<mml:mfenced open="(" close=")">
<mml:mrow>
<mml:mfenced open="{" close="}">
<mml:mrow>
<mml:mi>z</mml:mi>
<mml:mo>&#x2208;</mml:mo>
<mml:mi mathvariant="double-struck">X</mml:mi>
<mml:mo>:</mml:mo>
<mml:mi>T</mml:mi>
<mml:mfenced open="(" close=")">
<mml:mrow>
<mml:msub>
<mml:mrow>
<mml:mi>&#x3bc;</mml:mi>
</mml:mrow>
<mml:mrow>
<mml:mi>i</mml:mi>
<mml:mo>,</mml:mo>
<mml:mi>B</mml:mi>
</mml:mrow>
</mml:msub>
</mml:mrow>
</mml:mfenced>
<mml:mfenced open="(" close=")">
<mml:mrow>
<mml:mi>z</mml:mi>
</mml:mrow>
</mml:mfenced>
<mml:mo>&#x2265;</mml:mo>
<mml:mi>&#x3b8;</mml:mi>
</mml:mrow>
</mml:mfenced>
</mml:mrow>
</mml:mfenced>
<mml:mo>&#x2265;</mml:mo>
<mml:mn>1</mml:mn>
<mml:mo>&#x2212;</mml:mo>
<mml:mi>&#x3be;</mml:mi>
<mml:mo>.</mml:mo>
</mml:math>
</disp-formula>
</p>
<p>We consider the results of Theorem 2 in the well-separated setting with <inline-formula id="inf57">
<mml:math id="m76">
<mml:msup>
<mml:mrow>
<mml:mi>&#x3bd;</mml:mi>
</mml:mrow>
<mml:mrow>
<mml:mo>&#x2a;</mml:mo>
</mml:mrow>
</mml:msup>
<mml:mo>&#x3d;</mml:mo>
<mml:mfrac>
<mml:mrow>
<mml:mn>1</mml:mn>
</mml:mrow>
<mml:mrow>
<mml:mn>2</mml:mn>
</mml:mrow>
</mml:mfrac>
<mml:mrow>
<mml:mo stretchy="false">(</mml:mo>
<mml:mrow>
<mml:msub>
<mml:mrow>
<mml:mi>&#x3bc;</mml:mi>
</mml:mrow>
<mml:mrow>
<mml:mn>1</mml:mn>
</mml:mrow>
</mml:msub>
<mml:mo>&#x2b;</mml:mo>
<mml:msub>
<mml:mrow>
<mml:mi>&#x3bc;</mml:mi>
</mml:mrow>
<mml:mrow>
<mml:mn>2</mml:mn>
</mml:mrow>
</mml:msub>
</mml:mrow>
<mml:mo stretchy="false">)</mml:mo>
</mml:mrow>
</mml:math>
</inline-formula>:<list list-type="simple">
<list-item>
<p>a)<inline-formula id="inf58">
<mml:math id="m77">
<mml:msub>
<mml:mrow>
<mml:mi mathvariant="script">S</mml:mi>
</mml:mrow>
<mml:mrow>
<mml:mi>B</mml:mi>
</mml:mrow>
</mml:msub>
<mml:mrow>
<mml:mo stretchy="false">(</mml:mo>
<mml:mrow>
<mml:msub>
<mml:mrow>
<mml:mi>&#x3bc;</mml:mi>
</mml:mrow>
<mml:mrow>
<mml:mn>1</mml:mn>
</mml:mrow>
</mml:msub>
<mml:mo>,</mml:mo>
<mml:msub>
<mml:mrow>
<mml:mi>&#x3bc;</mml:mi>
</mml:mrow>
<mml:mrow>
<mml:mn>2</mml:mn>
</mml:mrow>
</mml:msub>
<mml:mo>;</mml:mo>
<mml:mi>&#x3b7;</mml:mi>
</mml:mrow>
<mml:mo stretchy="false">)</mml:mo>
</mml:mrow>
</mml:math>
</inline-formula> measures how much the backgrounds overlap with one another. In this setting, <italic>&#x3d5;</italic>
<sub>
<italic>B</italic>
</sub>(<italic>&#x3b7;</italic>) &#x2265; <italic>&#x3b4;</italic> for any <italic>&#x3b7;</italic> &#x3e; 0. This is because <italic>T</italic> (<italic>&#x3bc;</italic>
<sub>1,<italic>B</italic>
</sub>&#x2212; <italic>&#x3bc;</italic>
<sub>2,<italic>B</italic>
</sub>) (<italic>z</italic>) &#x3d; 0 for all <inline-formula id="inf59">
<mml:math id="m78">
<mml:mi>z</mml:mi>
<mml:mo>&#x2208;</mml:mo>
<mml:mi mathvariant="sans-serif">s</mml:mi>
<mml:mi mathvariant="sans-serif">u</mml:mi>
<mml:mi mathvariant="sans-serif">p</mml:mi>
<mml:mi mathvariant="sans-serif">p</mml:mi>
<mml:mrow>
<mml:mo stretchy="false">(</mml:mo>
<mml:mrow>
<mml:msub>
<mml:mrow>
<mml:mi>&#x3bc;</mml:mi>
</mml:mrow>
<mml:mrow>
<mml:mi>i</mml:mi>
<mml:mo>,</mml:mo>
<mml:mi>F</mml:mi>
</mml:mrow>
</mml:msub>
</mml:mrow>
<mml:mo stretchy="false">)</mml:mo>
</mml:mrow>
</mml:math>
</inline-formula>, and thus <inline-formula id="inf60">
<mml:math id="m79">
<mml:mi>z</mml:mi>
<mml:mo>&#x2208;</mml:mo>
<mml:msub>
<mml:mrow>
<mml:mi mathvariant="script">S</mml:mi>
</mml:mrow>
<mml:mrow>
<mml:mi>B</mml:mi>
</mml:mrow>
</mml:msub>
<mml:mrow>
<mml:mo stretchy="false">(</mml:mo>
<mml:mrow>
<mml:msub>
<mml:mrow>
<mml:mi>&#x3bc;</mml:mi>
</mml:mrow>
<mml:mrow>
<mml:mn>1</mml:mn>
</mml:mrow>
</mml:msub>
<mml:mo>,</mml:mo>
<mml:msub>
<mml:mrow>
<mml:mi>&#x3bc;</mml:mi>
</mml:mrow>
<mml:mrow>
<mml:mn>2</mml:mn>
</mml:mrow>
</mml:msub>
<mml:mo>;</mml:mo>
<mml:mi>&#x3b7;</mml:mi>
</mml:mrow>
<mml:mo stretchy="false">)</mml:mo>
</mml:mrow>
</mml:math>
</inline-formula>. Since <inline-formula id="inf61">
<mml:math id="m80">
<mml:msup>
<mml:mrow>
<mml:mi>&#x3bd;</mml:mi>
</mml:mrow>
<mml:mrow>
<mml:mo>&#x2a;</mml:mo>
</mml:mrow>
</mml:msup>
<mml:mrow>
<mml:mo stretchy="false">(</mml:mo>
<mml:mrow>
<mml:mi mathvariant="sans-serif">s</mml:mi>
<mml:mi mathvariant="sans-serif">u</mml:mi>
<mml:mi mathvariant="sans-serif">p</mml:mi>
<mml:mi mathvariant="sans-serif">p</mml:mi>
<mml:mrow>
<mml:mo stretchy="false">(</mml:mo>
<mml:mrow>
<mml:msub>
<mml:mrow>
<mml:mi>&#x3bc;</mml:mi>
</mml:mrow>
<mml:mrow>
<mml:mn>1</mml:mn>
<mml:mo>,</mml:mo>
<mml:mi>F</mml:mi>
</mml:mrow>
</mml:msub>
</mml:mrow>
<mml:mo stretchy="false">)</mml:mo>
</mml:mrow>
<mml:mo>&#x222a;</mml:mo>
<mml:mi mathvariant="sans-serif">s</mml:mi>
<mml:mi mathvariant="sans-serif">u</mml:mi>
<mml:mi mathvariant="sans-serif">p</mml:mi>
<mml:mi mathvariant="sans-serif">p</mml:mi>
<mml:mrow>
<mml:mo stretchy="false">(</mml:mo>
<mml:mrow>
<mml:msub>
<mml:mrow>
<mml:mi>&#x3bc;</mml:mi>
</mml:mrow>
<mml:mrow>
<mml:mn>2</mml:mn>
<mml:mo>,</mml:mo>
<mml:mi>F</mml:mi>
</mml:mrow>
</mml:msub>
</mml:mrow>
<mml:mo stretchy="false">)</mml:mo>
</mml:mrow>
</mml:mrow>
<mml:mo stretchy="false">)</mml:mo>
</mml:mrow>
<mml:mo>&#x3d;</mml:mo>
<mml:mi>&#x3b4;</mml:mi>
</mml:math>
</inline-formula>, this lower bounds <italic>&#x3d5;</italic>
<sub>
<italic>B</italic>
</sub>(<italic>&#x3b7;</italic>). This means for any <italic>&#x3b1;</italic> &#x3c; <italic>&#x3b4;</italic>, kdiff (<italic>&#x3bc;</italic>
<sub>1</sub>, <italic>&#x3bc;</italic>
<sub>2</sub>; <italic>&#x3b1;</italic>) &#x3d;&#x20;0.</p>
</list-item>
<list-item>
<p>&#x2009;b) Because of the well-separated assumption, <inline-formula id="inf62">
<mml:math id="m81">
<mml:msubsup>
<mml:mrow>
<mml:mi mathvariant="script">S</mml:mi>
</mml:mrow>
<mml:mrow>
<mml:mi>F</mml:mi>
</mml:mrow>
<mml:mrow>
<mml:mi>c</mml:mi>
</mml:mrow>
</mml:msubsup>
<mml:mrow>
<mml:mo stretchy="false">(</mml:mo>
<mml:mrow>
<mml:mi>&#x3b8;</mml:mi>
</mml:mrow>
<mml:mo stretchy="false">)</mml:mo>
</mml:mrow>
<mml:mo>&#x2282;</mml:mo>
<mml:msub>
<mml:mrow>
<mml:mi mathvariant="script">S</mml:mi>
</mml:mrow>
<mml:mrow>
<mml:mi>B</mml:mi>
</mml:mrow>
</mml:msub>
<mml:mrow>
<mml:mo stretchy="false">(</mml:mo>
<mml:mrow>
<mml:mi>&#x3b7;</mml:mi>
</mml:mrow>
<mml:mo stretchy="false">)</mml:mo>
</mml:mrow>
</mml:math>
</inline-formula>. This means that everywhere the foregrounds are sufficiently concentrated, the backgrounds must be sufficiently small. Similarly, <inline-formula id="inf63">
<mml:math id="m82">
<mml:msubsup>
<mml:mrow>
<mml:mi mathvariant="script">S</mml:mi>
</mml:mrow>
<mml:mrow>
<mml:mi>B</mml:mi>
</mml:mrow>
<mml:mrow>
<mml:mi>c</mml:mi>
</mml:mrow>
</mml:msubsup>
<mml:mrow>
<mml:mo stretchy="false">(</mml:mo>
<mml:mrow>
<mml:mi>&#x3b8;</mml:mi>
</mml:mrow>
<mml:mo stretchy="false">)</mml:mo>
</mml:mrow>
<mml:mo>&#x2282;</mml:mo>
<mml:msub>
<mml:mrow>
<mml:mi mathvariant="script">S</mml:mi>
</mml:mrow>
<mml:mrow>
<mml:mi>F</mml:mi>
</mml:mrow>
</mml:msub>
<mml:mrow>
<mml:mo stretchy="false">(</mml:mo>
<mml:mrow>
<mml:mi>&#x3b7;</mml:mi>
</mml:mrow>
<mml:mo stretchy="false">)</mml:mo>
</mml:mrow>
</mml:math>
</inline-formula>. Furthermore, the sets <inline-formula id="inf64">
<mml:math id="m83">
<mml:msubsup>
<mml:mrow>
<mml:mi mathvariant="script">S</mml:mi>
</mml:mrow>
<mml:mrow>
<mml:mi>F</mml:mi>
</mml:mrow>
<mml:mrow>
<mml:mi>c</mml:mi>
</mml:mrow>
</mml:msubsup>
<mml:mrow>
<mml:mo stretchy="false">(</mml:mo>
<mml:mrow>
<mml:mi>&#x3b8;</mml:mi>
</mml:mrow>
<mml:mo stretchy="false">)</mml:mo>
</mml:mrow>
<mml:mo>&#x2229;</mml:mo>
<mml:msub>
<mml:mrow>
<mml:mi mathvariant="script">S</mml:mi>
</mml:mrow>
<mml:mrow>
<mml:mi>B</mml:mi>
</mml:mrow>
</mml:msub>
<mml:mrow>
<mml:mo stretchy="false">(</mml:mo>
<mml:mrow>
<mml:mi>&#x3b7;</mml:mi>
</mml:mrow>
<mml:mo stretchy="false">)</mml:mo>
</mml:mrow>
</mml:math>
</inline-formula> and <inline-formula id="inf65">
<mml:math id="m84">
<mml:msubsup>
<mml:mrow>
<mml:mi mathvariant="script">S</mml:mi>
</mml:mrow>
<mml:mrow>
<mml:mi>B</mml:mi>
</mml:mrow>
<mml:mrow>
<mml:mi>c</mml:mi>
</mml:mrow>
</mml:msubsup>
<mml:mrow>
<mml:mo stretchy="false">(</mml:mo>
<mml:mrow>
<mml:mi>&#x3b8;</mml:mi>
</mml:mrow>
<mml:mo stretchy="false">)</mml:mo>
</mml:mrow>
<mml:mo>&#x2229;</mml:mo>
<mml:msub>
<mml:mrow>
<mml:mi mathvariant="script">S</mml:mi>
</mml:mrow>
<mml:mrow>
<mml:mi>F</mml:mi>
</mml:mrow>
</mml:msub>
<mml:mrow>
<mml:mo stretchy="false">(</mml:mo>
<mml:mrow>
<mml:mi>&#x3b7;</mml:mi>
</mml:mrow>
<mml:mo stretchy="false">)</mml:mo>
</mml:mrow>
</mml:math>
</inline-formula> by definition are disjoint when <italic>&#x3b8;</italic> &#x3e; <italic>&#x3b7;</italic>. Thus we&#x20;have</p>
</list-item>
</list>
<disp-formula id="equ9">
<mml:math id="m85">
<mml:mtable class="align-star" columnalign="left">
<mml:mtr>
<mml:mtd columnalign="right">
<mml:mn>1</mml:mn>
<mml:mo>&#x2212;</mml:mo>
<mml:mi>&#x3c8;</mml:mi>
<mml:mfenced open="(" close=")">
<mml:mrow>
<mml:mi>&#x3b8;</mml:mi>
<mml:mo>,</mml:mo>
<mml:mi>&#x3b7;</mml:mi>
</mml:mrow>
</mml:mfenced>
</mml:mtd>
<mml:mtd columnalign="left">
<mml:mo>&#x3d;</mml:mo>
<mml:mn>1</mml:mn>
<mml:mo>&#x2212;</mml:mo>
<mml:msup>
<mml:mrow>
<mml:mi>&#x3bd;</mml:mi>
</mml:mrow>
<mml:mrow>
<mml:mo>&#x2a;</mml:mo>
</mml:mrow>
</mml:msup>
<mml:mfenced open="(" close=")">
<mml:mrow>
<mml:mfenced open="(" close=")">
<mml:mrow>
<mml:msubsup>
<mml:mrow>
<mml:mi mathvariant="script">S</mml:mi>
</mml:mrow>
<mml:mrow>
<mml:mi>F</mml:mi>
</mml:mrow>
<mml:mrow>
<mml:mi>c</mml:mi>
</mml:mrow>
</mml:msubsup>
<mml:mfenced open="(" close=")">
<mml:mrow>
<mml:mi>&#x3b8;</mml:mi>
</mml:mrow>
</mml:mfenced>
<mml:mo>&#x2229;</mml:mo>
<mml:msub>
<mml:mrow>
<mml:mi mathvariant="script">S</mml:mi>
</mml:mrow>
<mml:mrow>
<mml:mi>B</mml:mi>
</mml:mrow>
</mml:msub>
<mml:mfenced open="(" close=")">
<mml:mrow>
<mml:mi>&#x3b7;</mml:mi>
</mml:mrow>
</mml:mfenced>
</mml:mrow>
</mml:mfenced>
<mml:mo>&#x222a;</mml:mo>
<mml:mfenced open="(" close=")">
<mml:mrow>
<mml:msubsup>
<mml:mrow>
<mml:mi mathvariant="script">S</mml:mi>
</mml:mrow>
<mml:mrow>
<mml:mi>B</mml:mi>
</mml:mrow>
<mml:mrow>
<mml:mi>c</mml:mi>
</mml:mrow>
</mml:msubsup>
<mml:mfenced open="(" close=")">
<mml:mrow>
<mml:mi>&#x3b8;</mml:mi>
</mml:mrow>
</mml:mfenced>
<mml:mo>&#x2229;</mml:mo>
<mml:msub>
<mml:mrow>
<mml:mi mathvariant="script">S</mml:mi>
</mml:mrow>
<mml:mrow>
<mml:mi>F</mml:mi>
</mml:mrow>
</mml:msub>
<mml:mfenced open="(" close=")">
<mml:mrow>
<mml:mi>&#x3b7;</mml:mi>
</mml:mrow>
</mml:mfenced>
</mml:mrow>
</mml:mfenced>
</mml:mrow>
</mml:mfenced>
</mml:mtd>
</mml:mtr>
<mml:mtr>
<mml:mtd columnalign="right"/>
<mml:mtd columnalign="left">
<mml:mo>&#x3d;</mml:mo>
<mml:mn>1</mml:mn>
<mml:mo>&#x2212;</mml:mo>
<mml:msup>
<mml:mrow>
<mml:mi>&#x3bd;</mml:mi>
</mml:mrow>
<mml:mrow>
<mml:mo>&#x2a;</mml:mo>
</mml:mrow>
</mml:msup>
<mml:mfenced open="(" close=")">
<mml:mrow>
<mml:msubsup>
<mml:mrow>
<mml:mi mathvariant="script">S</mml:mi>
</mml:mrow>
<mml:mrow>
<mml:mi>F</mml:mi>
</mml:mrow>
<mml:mrow>
<mml:mi>c</mml:mi>
</mml:mrow>
</mml:msubsup>
<mml:mfenced open="(" close=")">
<mml:mrow>
<mml:mi>&#x3b8;</mml:mi>
</mml:mrow>
</mml:mfenced>
<mml:mo>&#x2229;</mml:mo>
<mml:msub>
<mml:mrow>
<mml:mi mathvariant="script">S</mml:mi>
</mml:mrow>
<mml:mrow>
<mml:mi>B</mml:mi>
</mml:mrow>
</mml:msub>
<mml:mfenced open="(" close=")">
<mml:mrow>
<mml:mi>&#x3b7;</mml:mi>
</mml:mrow>
</mml:mfenced>
</mml:mrow>
</mml:mfenced>
<mml:mo>&#x2212;</mml:mo>
<mml:msup>
<mml:mrow>
<mml:mi>&#x3bd;</mml:mi>
</mml:mrow>
<mml:mrow>
<mml:mo>&#x2a;</mml:mo>
</mml:mrow>
</mml:msup>
<mml:mfenced open="(" close=")">
<mml:mrow>
<mml:msubsup>
<mml:mrow>
<mml:mi mathvariant="script">S</mml:mi>
</mml:mrow>
<mml:mrow>
<mml:mi>B</mml:mi>
</mml:mrow>
<mml:mrow>
<mml:mi>c</mml:mi>
</mml:mrow>
</mml:msubsup>
<mml:mfenced open="(" close=")">
<mml:mrow>
<mml:mi>&#x3b8;</mml:mi>
</mml:mrow>
</mml:mfenced>
<mml:mo>&#x2229;</mml:mo>
<mml:msub>
<mml:mrow>
<mml:mi mathvariant="script">S</mml:mi>
</mml:mrow>
<mml:mrow>
<mml:mi>F</mml:mi>
</mml:mrow>
</mml:msub>
<mml:mfenced open="(" close=")">
<mml:mrow>
<mml:mi>&#x3b7;</mml:mi>
</mml:mrow>
</mml:mfenced>
</mml:mrow>
</mml:mfenced>
</mml:mtd>
</mml:mtr>
<mml:mtr>
<mml:mtd columnalign="right"/>
<mml:mtd columnalign="left">
<mml:mo>&#x3d;</mml:mo>
<mml:mn>1</mml:mn>
<mml:mo>&#x2212;</mml:mo>
<mml:msup>
<mml:mrow>
<mml:mi>&#x3bd;</mml:mi>
</mml:mrow>
<mml:mrow>
<mml:mo>&#x2a;</mml:mo>
</mml:mrow>
</mml:msup>
<mml:mfenced open="(" close=")">
<mml:mrow>
<mml:msubsup>
<mml:mrow>
<mml:mi mathvariant="script">S</mml:mi>
</mml:mrow>
<mml:mrow>
<mml:mi>F</mml:mi>
</mml:mrow>
<mml:mrow>
<mml:mi>c</mml:mi>
</mml:mrow>
</mml:msubsup>
<mml:mfenced open="(" close=")">
<mml:mrow>
<mml:mi>&#x3b8;</mml:mi>
</mml:mrow>
</mml:mfenced>
</mml:mrow>
</mml:mfenced>
<mml:mo>&#x2212;</mml:mo>
<mml:msup>
<mml:mrow>
<mml:mi>&#x3bd;</mml:mi>
</mml:mrow>
<mml:mrow>
<mml:mo>&#x2a;</mml:mo>
</mml:mrow>
</mml:msup>
<mml:mfenced open="(" close=")">
<mml:mrow>
<mml:msubsup>
<mml:mrow>
<mml:mi mathvariant="script">S</mml:mi>
</mml:mrow>
<mml:mrow>
<mml:mi>B</mml:mi>
</mml:mrow>
<mml:mrow>
<mml:mi>c</mml:mi>
</mml:mrow>
</mml:msubsup>
<mml:mfenced open="(" close=")">
<mml:mrow>
<mml:mi>&#x3b8;</mml:mi>
</mml:mrow>
</mml:mfenced>
</mml:mrow>
</mml:mfenced>
</mml:mtd>
</mml:mtr>
<mml:mtr>
<mml:mtd columnalign="right"/>
<mml:mtd columnalign="left">
<mml:mo>&#x2264;</mml:mo>
<mml:mn>1</mml:mn>
<mml:mo>&#x2212;</mml:mo>
<mml:mi>&#x3b4;</mml:mi>
<mml:mfenced open="(" close=")">
<mml:mrow>
<mml:mn>1</mml:mn>
<mml:mo>&#x2212;</mml:mo>
<mml:mi>&#x3be;</mml:mi>
</mml:mrow>
</mml:mfenced>
<mml:mo>&#x2212;</mml:mo>
<mml:mfenced open="(" close=")">
<mml:mrow>
<mml:mn>1</mml:mn>
<mml:mo>&#x2212;</mml:mo>
<mml:mi>&#x3b4;</mml:mi>
</mml:mrow>
</mml:mfenced>
<mml:mfenced open="(" close=")">
<mml:mrow>
<mml:mn>1</mml:mn>
<mml:mo>&#x2212;</mml:mo>
<mml:mi>&#x3be;</mml:mi>
</mml:mrow>
</mml:mfenced>
</mml:mtd>
</mml:mtr>
<mml:mtr>
<mml:mtd columnalign="right"/>
<mml:mtd columnalign="left">
<mml:mo>&#x3d;</mml:mo>
<mml:mi>&#x3be;</mml:mi>
<mml:mo>.</mml:mo>
</mml:mtd>
</mml:mtr>
</mml:mtable>
</mml:math>
</disp-formula>
</p>
<p>Thus we can choose <italic>&#x3b7;</italic>, <italic>&#x3b8;</italic> as large as possible to satisfy the assumptions, and even then for very small quantiles <italic>&#x3b1;</italic> &#x3e; <italic>&#x3be;</italic> we kdiff (<italic>&#x3bc;</italic>
<sub>1</sub>, <italic>&#x3bc;</italic>
<sub>2</sub>; <italic>&#x3b1;</italic>) &#x3e; 2 (1 &#x2212; <italic>&#x3b4;</italic>)<italic>&#x3b7;</italic>.</p>
<p>These above descriptions clarify the theorem in the simplest setting. When the foreground distributions are small but concentrated, and far from the separate backgrounds, then the hypothesis of <inline-formula id="inf66">
<mml:math id="m86">
<mml:msub>
<mml:mrow>
<mml:mi>&#x3bc;</mml:mi>
</mml:mrow>
<mml:mrow>
<mml:mn>1</mml:mn>
<mml:mo>,</mml:mo>
<mml:mi>F</mml:mi>
</mml:mrow>
</mml:msub>
<mml:mrow>
<mml:mover>
<mml:mrow>
<mml:mo>&#x3d;</mml:mo>
</mml:mrow>
<mml:mrow>
<mml:mrow>
<mml:mo>?</mml:mo>
</mml:mrow>
</mml:mrow>
</mml:mover>
</mml:mrow>
<mml:msub>
<mml:mrow>
<mml:mi>&#x3bc;</mml:mi>
</mml:mrow>
<mml:mrow>
<mml:mn>2</mml:mn>
<mml:mo>,</mml:mo>
<mml:mi>F</mml:mi>
</mml:mrow>
</mml:msub>
</mml:math>
</inline-formula> can be easily distinguished with kdiff for almost all <italic>&#x3b1;</italic> &#x3c; <italic>&#x3b4;</italic>.</p>
<p>In practice, of course, we need to estimate <bold>kidff</bold> (<italic>&#x3bc;</italic>
<sub>1</sub>, <italic>&#x3bc;</italic>
<sub>2</sub>; <italic>&#x3b1;</italic>) from samples taken from <italic>&#x3bc;</italic>
<sub>1</sub> and <italic>&#x3bc;</italic>
<sub>2</sub>. In turn, this necessitates an estimation of the witness function of probability measure from samples from this probability. We need to do this separately for <italic>&#x3bc;</italic>
<sub>1</sub> and <italic>&#x3bc;</italic>
<sub>2</sub>, but it is convenient to formulate the result for a generic probability measure <italic>&#x3bc;</italic>. To estimate the error in the resulting approximation, we need to stipulate some further conditions enumerated below. We will denote by <inline-formula id="inf67">
<mml:math id="m87">
<mml:msup>
<mml:mrow>
<mml:mi mathvariant="double-struck">S</mml:mi>
</mml:mrow>
<mml:mrow>
<mml:mo>&#x2a;</mml:mo>
</mml:mrow>
</mml:msup>
<mml:mo>&#x3d;</mml:mo>
<mml:mi mathvariant="sans-serif">s</mml:mi>
<mml:mi mathvariant="sans-serif">u</mml:mi>
<mml:mi mathvariant="sans-serif">p</mml:mi>
<mml:mi mathvariant="sans-serif">p</mml:mi>
<mml:mrow>
<mml:mo stretchy="false">(</mml:mo>
<mml:mrow>
<mml:mi>&#x3bc;</mml:mi>
</mml:mrow>
<mml:mo stretchy="false">)</mml:mo>
</mml:mrow>
</mml:math>
</inline-formula>.</p>
<p>
<bold>Essential compact support</bold> For any <italic>t</italic>&#x20;&#x3e; 0, there exists <italic>R</italic>(<italic>t</italic>) &#x3e; 0 such that<disp-formula id="e10">
<mml:math id="m88">
<mml:mi>K</mml:mi>
<mml:mfenced open="(" close=")">
<mml:mrow>
<mml:mi>x</mml:mi>
<mml:mo>,</mml:mo>
<mml:mi>y</mml:mi>
</mml:mrow>
</mml:mfenced>
<mml:mo>&#x2264;</mml:mo>
<mml:mi>t</mml:mi>
<mml:mo>,</mml:mo>
<mml:mspace width="2em"/>
<mml:mi>x</mml:mi>
<mml:mo>,</mml:mo>
<mml:mi>y</mml:mi>
<mml:mo>&#x2208;</mml:mo>
<mml:mi mathvariant="double-struck">X</mml:mi>
<mml:mo>,</mml:mo>
<mml:mspace width="0.3333em"/>
<mml:mi>&#x3c1;</mml:mi>
<mml:mfenced open="(" close=")">
<mml:mrow>
<mml:mi>x</mml:mi>
<mml:mo>,</mml:mo>
<mml:mi>y</mml:mi>
</mml:mrow>
</mml:mfenced>
<mml:mo>&#x2265;</mml:mo>
<mml:mi>R</mml:mi>
<mml:mfenced open="(" close=")">
<mml:mrow>
<mml:mi>t</mml:mi>
</mml:mrow>
</mml:mfenced>
<mml:mo>.</mml:mo>
</mml:math>
<label>(10)</label>
</disp-formula>
</p>
<p>
<bold>Covering property</bold> For <italic>t</italic>&#x20;&#x3e; 0, let <inline-formula id="inf68">
<mml:math id="m89">
<mml:mi mathvariant="double-struck">B</mml:mi>
<mml:mrow>
<mml:mo stretchy="false">(</mml:mo>
<mml:mrow>
<mml:msup>
<mml:mrow>
<mml:mi mathvariant="double-struck">S</mml:mi>
</mml:mrow>
<mml:mrow>
<mml:mo>&#x2a;</mml:mo>
</mml:mrow>
</mml:msup>
<mml:mo>,</mml:mo>
<mml:mi>t</mml:mi>
</mml:mrow>
<mml:mo stretchy="false">)</mml:mo>
</mml:mrow>
<mml:mo>&#x3d;</mml:mo>
<mml:mrow>
<mml:mo stretchy="false">{</mml:mo>
<mml:mrow>
<mml:mi>z</mml:mi>
<mml:mo>&#x2208;</mml:mo>
<mml:mi mathvariant="double-struck">X</mml:mi>
<mml:mo>:</mml:mo>
<mml:mi mathvariant="sans-serif">d</mml:mi>
<mml:mi mathvariant="sans-serif">i</mml:mi>
<mml:mi mathvariant="sans-serif">s</mml:mi>
<mml:mi mathvariant="sans-serif">t</mml:mi>
<mml:mrow>
<mml:mo stretchy="false">(</mml:mo>
<mml:mrow>
<mml:msup>
<mml:mrow>
<mml:mi mathvariant="double-struck">S</mml:mi>
</mml:mrow>
<mml:mrow>
<mml:mo>&#x2a;</mml:mo>
</mml:mrow>
</mml:msup>
<mml:mo>,</mml:mo>
<mml:mi>z</mml:mi>
</mml:mrow>
<mml:mo stretchy="false">)</mml:mo>
</mml:mrow>
<mml:mo>&#x2264;</mml:mo>
<mml:mi>R</mml:mi>
<mml:mrow>
<mml:mo stretchy="false">(</mml:mo>
<mml:mrow>
<mml:mi>t</mml:mi>
</mml:mrow>
<mml:mo stretchy="false">)</mml:mo>
</mml:mrow>
</mml:mrow>
<mml:mo stretchy="false">}</mml:mo>
</mml:mrow>
</mml:math>
</inline-formula>. There exist <italic>A</italic>, <italic>&#x3b2;</italic> &#x3e; 0 such that for any <italic>t</italic>&#x20;&#x3e; 0, the set <inline-formula id="inf69">
<mml:math id="m90">
<mml:mi mathvariant="double-struck">B</mml:mi>
<mml:mrow>
<mml:mo stretchy="false">(</mml:mo>
<mml:mrow>
<mml:mi mathvariant="double-struck">S</mml:mi>
<mml:mo>,</mml:mo>
<mml:mi>t</mml:mi>
</mml:mrow>
<mml:mo stretchy="false">)</mml:mo>
</mml:mrow>
</mml:math>
</inline-formula> is contained in the union of at most <italic>At</italic>
<sup>&#x2212;<italic>&#x3b2;</italic>
</sup> balls of radius &#x2264;&#x20;<italic>t</italic>.</p>
<p>
<bold>Lipschitz condition</bold> We have<disp-formula id="e11">
<mml:math id="m91">
<mml:msub>
<mml:mrow>
<mml:mi>max</mml:mi>
</mml:mrow>
<mml:mrow>
<mml:mfenced open="(" close=")">
<mml:mrow>
<mml:mi>x</mml:mi>
<mml:mo>,</mml:mo>
<mml:mi>y</mml:mi>
</mml:mrow>
</mml:mfenced>
<mml:mo>&#x2208;</mml:mo>
<mml:mi mathvariant="double-struck">X</mml:mi>
<mml:mo>&#xd7;</mml:mo>
<mml:mi mathvariant="double-struck">X</mml:mi>
</mml:mrow>
</mml:msub>
<mml:mi>K</mml:mi>
<mml:mfenced open="(" close=")">
<mml:mrow>
<mml:mi>x</mml:mi>
<mml:mo>,</mml:mo>
<mml:mi>y</mml:mi>
</mml:mrow>
</mml:mfenced>
<mml:mo>&#x2b;</mml:mo>
<mml:msub>
<mml:mrow>
<mml:mi>max</mml:mi>
</mml:mrow>
<mml:mrow>
<mml:mi>x</mml:mi>
<mml:mo>,</mml:mo>
<mml:msup>
<mml:mrow>
<mml:mi>x</mml:mi>
</mml:mrow>
<mml:mrow>
<mml:mo>&#x2032;</mml:mo>
</mml:mrow>
</mml:msup>
<mml:mo>,</mml:mo>
<mml:mi>y</mml:mi>
<mml:mo>,</mml:mo>
<mml:msup>
<mml:mrow>
<mml:mi>y</mml:mi>
</mml:mrow>
<mml:mrow>
<mml:mo>&#x2032;</mml:mo>
</mml:mrow>
</mml:msup>
<mml:mo>&#x2208;</mml:mo>
<mml:mi mathvariant="double-struck">X</mml:mi>
</mml:mrow>
</mml:msub>
<mml:mfenced open="{" close="}">
<mml:mrow>
<mml:mfrac>
<mml:mrow>
<mml:mo stretchy="false">&#x7c;</mml:mo>
<mml:mi>K</mml:mi>
<mml:mfenced open="(" close=")">
<mml:mrow>
<mml:mi>x</mml:mi>
<mml:mo>,</mml:mo>
<mml:mi>y</mml:mi>
</mml:mrow>
</mml:mfenced>
<mml:mo>&#x2212;</mml:mo>
<mml:mi>K</mml:mi>
<mml:mfenced open="(" close=")">
<mml:mrow>
<mml:msup>
<mml:mrow>
<mml:mi>x</mml:mi>
</mml:mrow>
<mml:mrow>
<mml:mo>&#x2032;</mml:mo>
</mml:mrow>
</mml:msup>
<mml:mo>,</mml:mo>
<mml:msup>
<mml:mrow>
<mml:mi>y</mml:mi>
</mml:mrow>
<mml:mrow>
<mml:mo>&#x2032;</mml:mo>
</mml:mrow>
</mml:msup>
</mml:mrow>
</mml:mfenced>
<mml:mo stretchy="false">&#x7c;</mml:mo>
</mml:mrow>
<mml:mrow>
<mml:mi>&#x3c1;</mml:mi>
<mml:mfenced open="(" close=")">
<mml:mrow>
<mml:mi>x</mml:mi>
<mml:mo>,</mml:mo>
<mml:msup>
<mml:mrow>
<mml:mi>x</mml:mi>
</mml:mrow>
<mml:mrow>
<mml:mo>&#x2032;</mml:mo>
</mml:mrow>
</mml:msup>
</mml:mrow>
</mml:mfenced>
<mml:mo>&#x2b;</mml:mo>
<mml:mi>&#x3c1;</mml:mi>
<mml:mfenced open="(" close=")">
<mml:mrow>
<mml:mi>y</mml:mi>
<mml:mo>,</mml:mo>
<mml:msup>
<mml:mrow>
<mml:mi>y</mml:mi>
</mml:mrow>
<mml:mrow>
<mml:mo>&#x2032;</mml:mo>
</mml:mrow>
</mml:msup>
</mml:mrow>
</mml:mfenced>
</mml:mrow>
</mml:mfrac>
</mml:mrow>
</mml:mfenced>
<mml:mo>&#x2264;</mml:mo>
<mml:mn>1</mml:mn>
<mml:mo>.</mml:mo>
</mml:math>
<label>(11)</label>
</disp-formula>
</p>
<p>Then H&#xf6;ffding&#x2019;s inequality leads to the following theorem.</p>
</statement>
</p>
<p>
<statement content-type="theorem" id="Theorem_3">
<label>Theorem 3</label>
<p>
<italic>Let</italic> <italic>&#x3f5;</italic> &#x3e; 0<italic>,</italic> <italic>M</italic>&#x20;&#x2265; 2 <italic>be an integer, and</italic> <italic>&#x3bc;</italic> <italic>be any probability measure on</italic> <inline-formula id="inf70">
<mml:math id="m92">
<mml:mi mathvariant="double-struck">X</mml:mi>
</mml:math>
</inline-formula> <italic>and</italic> {<italic>y</italic>
<sub>1</sub>, &#x2026; , <italic>y</italic>
<sub>
<italic>M</italic>
</sub>} <italic>be i.i.d. samples from</italic> <italic>&#x3bc;</italic>
<italic>. Then with</italic> <italic>&#x3bc;</italic>
<italic>-probability</italic> &#x2265; 1&#x20;&#x2212; <italic>&#x3f5;</italic>
<italic>, we have</italic>
<disp-formula id="e12">
<mml:math id="m93">
<mml:msub>
<mml:mrow>
<mml:mfenced open="&#x2016;" close="&#x2016;">
<mml:mrow>
<mml:mi>U</mml:mi>
<mml:mfenced open="(" close=")">
<mml:mrow>
<mml:mi>&#x3bc;</mml:mi>
</mml:mrow>
</mml:mfenced>
<mml:mfenced open="(" close=")">
<mml:mrow>
<mml:mo>&#x25e6;</mml:mo>
</mml:mrow>
</mml:mfenced>
<mml:mo>&#x2212;</mml:mo>
<mml:mfrac>
<mml:mrow>
<mml:mn>1</mml:mn>
</mml:mrow>
<mml:mrow>
<mml:mi>M</mml:mi>
</mml:mrow>
</mml:mfrac>
<mml:munderover accentunder="false" accent="false">
<mml:mrow>
<mml:mo>&#x2211;</mml:mo>
</mml:mrow>
<mml:mrow>
<mml:mi>j</mml:mi>
<mml:mo>&#x3d;</mml:mo>
<mml:mn>1</mml:mn>
</mml:mrow>
<mml:mrow>
<mml:mi>M</mml:mi>
</mml:mrow>
</mml:munderover>
<mml:mi>K</mml:mi>
<mml:mfenced open="(" close=")">
<mml:mrow>
<mml:mo>&#x25e6;</mml:mo>
<mml:mo>,</mml:mo>
<mml:msub>
<mml:mrow>
<mml:mi>y</mml:mi>
</mml:mrow>
<mml:mrow>
<mml:mi>j</mml:mi>
</mml:mrow>
</mml:msub>
</mml:mrow>
</mml:mfenced>
</mml:mrow>
</mml:mfenced>
</mml:mrow>
<mml:mrow>
<mml:mi>&#x221e;</mml:mi>
</mml:mrow>
</mml:msub>
<mml:mo>&#x2264;</mml:mo>
<mml:mn>2</mml:mn>
<mml:msup>
<mml:mrow>
<mml:mfenced open="{" close="}">
<mml:mrow>
<mml:mfrac>
<mml:mrow>
<mml:mi>log</mml:mi>
<mml:mfenced open="(" close=")">
<mml:mrow>
<mml:msup>
<mml:mrow>
<mml:mn>4</mml:mn>
</mml:mrow>
<mml:mrow>
<mml:mi>&#x3b2;</mml:mi>
<mml:mo>&#x2b;</mml:mo>
<mml:mn>1</mml:mn>
</mml:mrow>
</mml:msup>
<mml:mi>A</mml:mi>
<mml:mo>/</mml:mo>
<mml:mi>&#x3f5;</mml:mi>
</mml:mrow>
</mml:mfenced>
</mml:mrow>
<mml:mrow>
<mml:mi>M</mml:mi>
</mml:mrow>
</mml:mfrac>
</mml:mrow>
</mml:mfenced>
</mml:mrow>
<mml:mrow>
<mml:mn>1</mml:mn>
<mml:mo>/</mml:mo>
<mml:mn>2</mml:mn>
</mml:mrow>
</mml:msup>
<mml:mo>.</mml:mo>
</mml:math>
<label>(12)</label>
</disp-formula>
</p>
<p>The proof of Theorem 3 mirrors the results for the witness function in&#x20;[<xref ref-type="bibr" rid="B14">14</xref>].</p>
</statement>
</p>
</sec>
<sec id="s2-4">
<title>2.4 Conclusions From Separability Theorems</title>
<p>To illustrate the benefit of the above theory, we recall the MMD distance measure between two probability measures <italic>&#x3bc;</italic>
<sub>1</sub> and <italic>&#x3bc;</italic>
<sub>2</sub> defined by<disp-formula id="e13">
<mml:math id="m94">
<mml:msup>
<mml:mrow>
<mml:mtext>MMD</mml:mtext>
</mml:mrow>
<mml:mrow>
<mml:mn>2</mml:mn>
</mml:mrow>
</mml:msup>
<mml:mfenced open="(" close=")">
<mml:mrow>
<mml:msub>
<mml:mrow>
<mml:mi>&#x3bc;</mml:mi>
</mml:mrow>
<mml:mrow>
<mml:mn>1</mml:mn>
</mml:mrow>
</mml:msub>
<mml:mo>,</mml:mo>
<mml:msub>
<mml:mrow>
<mml:mi>&#x3bc;</mml:mi>
</mml:mrow>
<mml:mrow>
<mml:mn>2</mml:mn>
</mml:mrow>
</mml:msub>
</mml:mrow>
</mml:mfenced>
<mml:mo>&#x3d;</mml:mo>
<mml:msub>
<mml:mrow>
<mml:mo>&#x222b;</mml:mo>
</mml:mrow>
<mml:mrow>
<mml:mi mathvariant="double-struck">X</mml:mi>
</mml:mrow>
</mml:msub>
<mml:msub>
<mml:mrow>
<mml:mo>&#x222b;</mml:mo>
</mml:mrow>
<mml:mrow>
<mml:mi mathvariant="double-struck">X</mml:mi>
</mml:mrow>
</mml:msub>
<mml:mi>K</mml:mi>
<mml:mfenced open="(" close=")">
<mml:mrow>
<mml:mi>x</mml:mi>
<mml:mo>,</mml:mo>
<mml:mi>y</mml:mi>
</mml:mrow>
</mml:mfenced>
<mml:mfenced open="(" close=")">
<mml:mrow>
<mml:mi>d</mml:mi>
<mml:msub>
<mml:mrow>
<mml:mi>&#x3bc;</mml:mi>
</mml:mrow>
<mml:mrow>
<mml:mn>1</mml:mn>
</mml:mrow>
</mml:msub>
<mml:mfenced open="(" close=")">
<mml:mrow>
<mml:mi>x</mml:mi>
</mml:mrow>
</mml:mfenced>
<mml:mo>&#x2212;</mml:mo>
<mml:mi>d</mml:mi>
<mml:msub>
<mml:mrow>
<mml:mi>&#x3bc;</mml:mi>
</mml:mrow>
<mml:mrow>
<mml:mn>2</mml:mn>
</mml:mrow>
</mml:msub>
<mml:mfenced open="(" close=")">
<mml:mrow>
<mml:mi>x</mml:mi>
</mml:mrow>
</mml:mfenced>
</mml:mrow>
</mml:mfenced>
<mml:mfenced open="(" close=")">
<mml:mrow>
<mml:mi>d</mml:mi>
<mml:msub>
<mml:mrow>
<mml:mi>&#x3bc;</mml:mi>
</mml:mrow>
<mml:mrow>
<mml:mn>1</mml:mn>
</mml:mrow>
</mml:msub>
<mml:mfenced open="(" close=")">
<mml:mrow>
<mml:mi>y</mml:mi>
</mml:mrow>
</mml:mfenced>
<mml:mo>&#x2212;</mml:mo>
<mml:mi>d</mml:mi>
<mml:msub>
<mml:mrow>
<mml:mi>&#x3bc;</mml:mi>
</mml:mrow>
<mml:mrow>
<mml:mn>2</mml:mn>
</mml:mrow>
</mml:msub>
<mml:mfenced open="(" close=")">
<mml:mrow>
<mml:mi>y</mml:mi>
</mml:mrow>
</mml:mfenced>
</mml:mrow>
</mml:mfenced>
<mml:mo>.</mml:mo>
</mml:math>
<label>(13)</label>
</disp-formula>
</p>
<p>When <inline-formula id="inf71">
<mml:math id="m95">
<mml:msub>
<mml:mrow>
<mml:mi>&#x3bc;</mml:mi>
</mml:mrow>
<mml:mrow>
<mml:mn>1</mml:mn>
</mml:mrow>
</mml:msub>
<mml:mo>,</mml:mo>
<mml:msub>
<mml:mrow>
<mml:mi>&#x3bc;</mml:mi>
</mml:mrow>
<mml:mrow>
<mml:mn>2</mml:mn>
</mml:mrow>
</mml:msub>
<mml:mo>&#x2208;</mml:mo>
<mml:msub>
<mml:mrow>
<mml:mi mathvariant="double-struck">P</mml:mi>
</mml:mrow>
<mml:mrow>
<mml:mi>&#x3b4;</mml:mi>
</mml:mrow>
</mml:msub>
</mml:math>
</inline-formula> and the foreground components <italic>&#x3bc;</italic>
<sub>1,<italic>F</italic>
</sub> &#x3d; <italic>&#x3bc;</italic>
<sub>2,<italic>F</italic>
</sub> then <italic>&#x3bc;</italic>
<sub>1</sub> &#x2212; <italic>&#x3bc;</italic>
<sub>2</sub> &#x3d; (1 &#x2212; <italic>&#x3b4;</italic>) (<italic>&#x3bc;</italic>
<sub>1,<italic>B</italic>
</sub>&#x2212; <italic>&#x3bc;</italic>
<sub>2,<italic>B</italic>
</sub>) and<disp-formula id="e14">
<mml:math id="m96">
<mml:msup>
<mml:mrow>
<mml:mtext>MMD</mml:mtext>
</mml:mrow>
<mml:mrow>
<mml:mn>2</mml:mn>
</mml:mrow>
</mml:msup>
<mml:mfenced open="(" close=")">
<mml:mrow>
<mml:msub>
<mml:mrow>
<mml:mi>&#x3bc;</mml:mi>
</mml:mrow>
<mml:mrow>
<mml:mn>1</mml:mn>
</mml:mrow>
</mml:msub>
<mml:mo>,</mml:mo>
<mml:msub>
<mml:mrow>
<mml:mi>&#x3bc;</mml:mi>
</mml:mrow>
<mml:mrow>
<mml:mn>2</mml:mn>
</mml:mrow>
</mml:msub>
</mml:mrow>
</mml:mfenced>
<mml:mo>&#x3d;</mml:mo>
<mml:msup>
<mml:mrow>
<mml:mfenced open="(" close=")">
<mml:mrow>
<mml:mn>1</mml:mn>
<mml:mo>&#x2212;</mml:mo>
<mml:mi>&#x3b4;</mml:mi>
</mml:mrow>
</mml:mfenced>
</mml:mrow>
<mml:mrow>
<mml:mn>2</mml:mn>
</mml:mrow>
</mml:msup>
<mml:msup>
<mml:mrow>
<mml:mtext>MMD</mml:mtext>
</mml:mrow>
<mml:mrow>
<mml:mn>2</mml:mn>
</mml:mrow>
</mml:msup>
<mml:mfenced open="(" close=")">
<mml:mrow>
<mml:msub>
<mml:mrow>
<mml:mi>&#x3bc;</mml:mi>
</mml:mrow>
<mml:mrow>
<mml:mn>1</mml:mn>
<mml:mo>,</mml:mo>
<mml:mi>B</mml:mi>
</mml:mrow>
</mml:msub>
<mml:mo>,</mml:mo>
<mml:msub>
<mml:mrow>
<mml:mi>&#x3bc;</mml:mi>
</mml:mrow>
<mml:mrow>
<mml:mn>2</mml:mn>
<mml:mo>,</mml:mo>
<mml:mi>B</mml:mi>
</mml:mrow>
</mml:msub>
</mml:mrow>
</mml:mfenced>
<mml:mo>.</mml:mo>
</mml:math>
<label>(14)</label>
</disp-formula>
</p>
<p>Since <italic>K</italic> is a positive definite kernel, it is thus impossible for MMD<sup>2</sup> (<italic>&#x3bc;</italic>
<sub>1</sub>, <italic>&#x3bc;</italic>
<sub>2</sub>) &#x3d; 0 unless <italic>&#x3bc;</italic>
<sub>1,<italic>B</italic>
</sub> &#x3d; <italic>&#x3bc;</italic>
<sub>2,<italic>B</italic>
</sub>. One of the motivations for our construction is to devise a test statistic that can be arbitrarily small even if <italic>&#x3bc;</italic>
<sub>1,<italic>B</italic>
</sub> &#x2260; <italic>&#x3bc;</italic>
<sub>2,<italic>B</italic>
</sub>.</p>
<p>The results derived above provide certain insights regarding when it is possible to perform tasks such as clustering and classification of data using distance measures such as kdiff and <bold>MMD</bold> based on the characteristics of their foreground and background distributions. The results in Theorem 2 show that, provided the backgrounds are sufficiently separated, the kdiff statistic will be significantly smaller when <italic>&#x3bc;</italic>
<sub>1,<italic>F</italic>
</sub> &#x3d; <italic>&#x3bc;</italic>
<sub>2,<italic>F</italic>
</sub> than when <italic>&#x3bc;</italic>
<sub>1,<italic>F</italic>
</sub> and <italic>&#x3bc;</italic>
<sub>2,<italic>F</italic>
</sub> are separated.</p>
<p>This enables kdiff to perform accurate discrimination i.e. data belonging to the same class will be clustered correctly in this case. On the other hand, it is clear that even if <italic>&#x3bc;</italic>
<sub>1,<italic>F</italic>
</sub> &#x3d; <italic>&#x3bc;</italic>
<sub>2,<italic>F</italic>
</sub>, MMD<sup>2</sup> (<italic>&#x3bc;</italic>
<sub>1</sub>, <italic>&#x3bc;</italic>
<sub>2</sub>) will still be highly dependent on the backgrounds. In this paper we consider the case where data instances belonging to the same class have the same foreground but different background distributions. In such situations using synthetic and real life examples we demonstrate the comparative performance and effectiveness of kdiff for clustering tasks versus other distance measures including&#x20;MMD.</p>
<p>As a final note, we wish to mention the relationship between <bold>MMD</bold> and kdiff. It can be shown that <bold>MMD</bold>
<sup>2</sup> is the mean of the witness function with respect to <inline-formula id="inf72">
<mml:math id="m97">
<mml:msup>
<mml:mrow>
<mml:mi>&#x3bd;</mml:mi>
</mml:mrow>
<mml:mrow>
<mml:mo>&#x2a;</mml:mo>
</mml:mrow>
</mml:msup>
<mml:mo>&#x3d;</mml:mo>
<mml:mfrac>
<mml:mrow>
<mml:mn>1</mml:mn>
</mml:mrow>
<mml:mrow>
<mml:mn>2</mml:mn>
</mml:mrow>
</mml:mfrac>
<mml:mrow>
<mml:mo stretchy="false">(</mml:mo>
<mml:mrow>
<mml:msub>
<mml:mrow>
<mml:mi>&#x3bc;</mml:mi>
</mml:mrow>
<mml:mrow>
<mml:mn>1</mml:mn>
</mml:mrow>
</mml:msub>
<mml:mo>&#x2b;</mml:mo>
<mml:msub>
<mml:mrow>
<mml:mi>&#x3bc;</mml:mi>
</mml:mrow>
<mml:mrow>
<mml:mn>2</mml:mn>
</mml:mrow>
</mml:msub>
</mml:mrow>
<mml:mo stretchy="false">)</mml:mo>
</mml:mrow>
</mml:math>
</inline-formula>, <inline-formula id="inf73">
<mml:math id="m98">
<mml:msup>
<mml:mrow>
<mml:mi mathvariant="bold">M</mml:mi>
<mml:mi mathvariant="bold">M</mml:mi>
<mml:mi mathvariant="bold">D</mml:mi>
</mml:mrow>
<mml:mrow>
<mml:mn>2</mml:mn>
</mml:mrow>
</mml:msup>
<mml:mrow>
<mml:mo stretchy="false">(</mml:mo>
<mml:mrow>
<mml:msub>
<mml:mrow>
<mml:mi>&#x3bc;</mml:mi>
</mml:mrow>
<mml:mrow>
<mml:mn>1</mml:mn>
</mml:mrow>
</mml:msub>
<mml:mo>,</mml:mo>
<mml:msub>
<mml:mrow>
<mml:mi>&#x3bc;</mml:mi>
</mml:mrow>
<mml:mrow>
<mml:mn>2</mml:mn>
</mml:mrow>
</mml:msub>
</mml:mrow>
<mml:mo stretchy="false">)</mml:mo>
</mml:mrow>
<mml:mo>&#x3d;</mml:mo>
<mml:msub>
<mml:mrow>
<mml:mi mathvariant="double-struck">E</mml:mi>
</mml:mrow>
<mml:mrow>
<mml:mi>z</mml:mi>
<mml:mo>&#x223c;</mml:mo>
<mml:msup>
<mml:mrow>
<mml:mi>&#x3bd;</mml:mi>
</mml:mrow>
<mml:mrow>
<mml:mo>&#x2a;</mml:mo>
</mml:mrow>
</mml:msup>
</mml:mrow>
</mml:msub>
<mml:mfenced open="(" close=")">
<mml:mrow>
<mml:mo stretchy="false">&#x7c;</mml:mo>
<mml:mi>U</mml:mi>
<mml:mrow>
<mml:mo stretchy="false">(</mml:mo>
<mml:mrow>
<mml:msub>
<mml:mrow>
<mml:mi>&#x3bc;</mml:mi>
</mml:mrow>
<mml:mrow>
<mml:mn>1</mml:mn>
</mml:mrow>
</mml:msub>
<mml:mo>&#x2212;</mml:mo>
<mml:msub>
<mml:mrow>
<mml:mi>&#x3bc;</mml:mi>
</mml:mrow>
<mml:mrow>
<mml:mn>2</mml:mn>
</mml:mrow>
</mml:msub>
</mml:mrow>
<mml:mo stretchy="false">)</mml:mo>
</mml:mrow>
<mml:mrow>
<mml:mo stretchy="false">(</mml:mo>
<mml:mrow>
<mml:mi>z</mml:mi>
</mml:mrow>
<mml:mo stretchy="false">)</mml:mo>
</mml:mrow>
<mml:msup>
<mml:mrow>
<mml:mo stretchy="false">&#x7c;</mml:mo>
</mml:mrow>
<mml:mrow>
<mml:mn>2</mml:mn>
</mml:mrow>
</mml:msup>
</mml:mrow>
</mml:mfenced>
</mml:math>
</inline-formula> [<xref ref-type="bibr" rid="B11">11</xref>, <xref ref-type="bibr" rid="B15">15</xref>]. This is compared to our results for kdiff, or in particular kdiff<sup>2</sup>. Note that computing <inline-formula id="inf74">
<mml:math id="m99">
<mml:mi mathvariant="bold">k</mml:mi>
<mml:mi mathvariant="bold">d</mml:mi>
<mml:mi mathvariant="bold">i</mml:mi>
<mml:mi mathvariant="bold">ff</mml:mi>
<mml:msup>
<mml:mrow>
<mml:mrow>
<mml:mo stretchy="false">(</mml:mo>
<mml:mrow>
<mml:msub>
<mml:mrow>
<mml:mi>&#x3bc;</mml:mi>
</mml:mrow>
<mml:mrow>
<mml:mn>1</mml:mn>
</mml:mrow>
</mml:msub>
<mml:mo>,</mml:mo>
<mml:msub>
<mml:mrow>
<mml:mi>&#x3bc;</mml:mi>
</mml:mrow>
<mml:mrow>
<mml:mn>2</mml:mn>
</mml:mrow>
</mml:msub>
<mml:mo>;</mml:mo>
<mml:mi>&#x3b1;</mml:mi>
</mml:mrow>
<mml:mo stretchy="false">)</mml:mo>
</mml:mrow>
</mml:mrow>
<mml:mrow>
<mml:mn>2</mml:mn>
</mml:mrow>
</mml:msup>
</mml:math>
</inline-formula> is equivalent to computing kdiff on the square of the witness function <italic>T</italic> (<italic>&#x3bc;</italic>
<sub>1</sub> &#x2212; <italic>&#x3bc;</italic>
<sub>2</sub>) (<italic>z</italic>) &#x3d; &#x7c;<italic>U</italic> (<italic>&#x3bc;</italic>
<sub>1</sub> &#x2212; <italic>&#x3bc;</italic>
<sub>2</sub>) (<italic>z</italic>)&#x7c;<sup>2</sup>, since quantiles depend only on the ordering of the underlying function. This means the statistic kdiff<sup>2</sup> is simply taking the quantile of the square of the witness function, rather than the mean as in&#x20;<bold>MMD</bold>
<sup>2</sup>.</p>
</sec>
</sec>
<sec id="s3">
<title>3 Estimation of Algorithm Parameters</title>
<p>The following parameters are required for estimation of the distance measure kdiff:<list list-type="simple">
<list-item>
<p>&#x2022; Length of sliding window <bold>
<italic>SL</italic>
</bold> used to generate subsequences over given data (<bold>embedding dimension</bold>)</p>
</list-item>
<list-item>
<p>&#x2022; Kernel bandwidth (<bold>
<italic>&#x3c3;</italic>
</bold>) of the Gaussian kernel <inline-formula id="inf75">
<mml:math id="m100">
<mml:mi>k</mml:mi>
<mml:mrow>
<mml:mo stretchy="false">(</mml:mo>
<mml:mrow>
<mml:mi>x</mml:mi>
<mml:mo>,</mml:mo>
<mml:mi>y</mml:mi>
</mml:mrow>
<mml:mo stretchy="false">)</mml:mo>
</mml:mrow>
<mml:mo>&#x3d;</mml:mo>
<mml:msup>
<mml:mrow>
<mml:mi>e</mml:mi>
</mml:mrow>
<mml:mrow>
<mml:mo>&#x2212;</mml:mo>
<mml:mo stretchy="false">&#x2016;</mml:mo>
<mml:mi>x</mml:mi>
<mml:mo>&#x2212;</mml:mo>
<mml:mi>y</mml:mi>
<mml:msup>
<mml:mrow>
<mml:mo stretchy="false">&#x2016;</mml:mo>
</mml:mrow>
<mml:mrow>
<mml:mn>2</mml:mn>
</mml:mrow>
</mml:msup>
<mml:mo>/</mml:mo>
<mml:mn>2</mml:mn>
<mml:msup>
<mml:mrow>
<mml:mi>&#x3c3;</mml:mi>
</mml:mrow>
<mml:mrow>
<mml:mn>2</mml:mn>
</mml:mrow>
</mml:msup>
</mml:mrow>
</mml:msup>
</mml:math>
</inline-formula>
</p>
</list-item>
<list-item>
<p>&#x2022; Lower quantile <italic>&#x3b1;</italic> of the kernel-based distance distribution&#x20;<italic>T</italic>(<italic>z</italic>)</p>
</list-item>
</list>
</p>
<p>Determining <italic>SL</italic>:</p>
<p>In this paper we demonstrate the application of kdiff for clustering time series and random fields. The sliding window length <italic>SL</italic> is used to create subsequences (i.e. sliding window based embeddings) over such time series or random fields over which kdiff is estimated. The number of subsequences formed depend on <italic>SL</italic>, the number of points in the time series or random field and the dimensionality of the data under consideration. Some examples are given as below:<list list-type="simple">
<list-item>
<p>&#x2022; In case of a univariate time series of length <italic>n</italic> if each subsequence is of length <italic>L</italic>&#x20;&#x3d; <italic>SL</italic> then there are <italic>m</italic>&#x20;&#x3d; <italic>n</italic>&#x20;&#x2212; <italic>SL</italic> &#x2b; 1 embeddings</p>
</list-item>
<list-item>
<p>&#x2022; For a two dimensional <italic>n</italic> x <italic>n</italic> random field if each subsequence has dimension <italic>L</italic>&#x20;&#x3d; <italic>SL</italic> &#x2217; <italic>SL</italic> then there are <italic>m</italic>&#x20;&#x3d; (<italic>n</italic>&#x20;&#x2212; <italic>SL</italic> &#x2b; 1)<sup>2</sup> embeddings</p>
</list-item>
<list-item>
<p>&#x2022; For a p-variate time series if each subsequence is of length <italic>L</italic>&#x20;&#x3d; <italic>p</italic> &#x2217; <italic>SL</italic> then there are <italic>m</italic>&#x20;&#x3d; <italic>n</italic>&#x20;&#x2212; <italic>L</italic>&#x20;&#x2b; 1 embeddings</p>
</list-item>
</list>
</p>
<p>The distance measure kdiff is estimated over these <italic>m</italic> points in the <italic>L</italic> dimensional embedding space. It is necessary to determine an optimal value of <italic>SL</italic> to obtain accurate values of kdiff. Very small values of <italic>SL</italic> may result in erroneous identification of the region where the time series or random field under consideration match. For example embeddings obtained in this manner may result in two dissimilar time series containing noise related fluctuations over a small region identified as &#x201c;matching&#x201d;. On the other hand very high values of <italic>SL</italic> can lead to erroneous estimation of the distance distribution owing to less number of subsequences or sub-regions which results in incorrect estimates for kdiff. As an optimal tradeoff between these competing considerations we determine the value of <italic>SL</italic> based on the best clustering performance over a training set selected from the original&#x20;data.</p>
<p>Determining <italic>&#x3c3;</italic>: Since kdiff is a kernel-based similarity measure determination of the kernel bandwidth <italic>&#x3c3;</italic> is critical to the accuracy of estimation. In this case very small bandwidths for sliding window based embeddings <bold>X</bold> and <bold>Y</bold> derived from two corresponding time series or random fields can lead to incorrect estimates since only points in the immediate neighborhood of embeddings <bold>X</bold> and <bold>Y</bold> are considered in the estimation of the kdiff statistic. On the other hand very large bandwidths are also problematic since in this case any point <bold>Z</bold> becomes nearly equidistant from <bold>X</bold> and <bold>Y</bold> (here all points are considered in the embedding space), thereby causing the distance measure to lose sensitivity. To achieve a suitable tradeoff between these extremes we select <italic>&#x3c3;</italic> over a range of values of order equal to the k nearest neighbor distance over all points in the embedding space of <bold>Z</bold> &#x3d; <bold>X</bold> &#x222a; <bold>Y</bold> for a suitably chosen value of <italic>k</italic>. The optimal value of <italic>&#x3c3;</italic> is selected from this range based on the best clustering performance over a training set selected from the original&#x20;data.</p>
<p>Determining <italic>&#x3b1;</italic>: The distance measure kdiff is based on a lower quantile <italic>&#x3b1;</italic> of the estimated distance distribution over the embedding spaces of two time series or random fields. This quantile can be specified as a fraction of the total number of points in the distance distribution using either of the following methods below:</p>
<p>&#x2022; Based on exploratory data analysis, visual inspection or other methods if the extent of the matching portions of the time series, random fields or other data under investigation can be determined then <italic>&#x3b1;</italic> can be set as a fraction of the length or area of this matching region versus the overall span of the&#x20;data.</p>
<p>&#x2022; For high dimensional time series or random fields <italic>&#x3b1;</italic> can be determined from a range of values based on the best clustering performance over a training set selected from the original&#x20;data.</p>
</sec>
<sec id="s4">
<title>4 Numerical Work: Simulations and Real Data</title>
<p>The effectiveness of the our novel distance measure kdiff for comparing two sets of data which match only partially over their region of support is estimated using kmedoids clustering [<xref ref-type="bibr" rid="B16">16</xref>]. The kmedoids algorithm is a similar partitional clustering algorithm as kmeans which works by minimizing the distance between the points belonging to a cluster and the center of the cluster. However kmeans can work only with Euclidean distances or a distance measure which can be directly expressed in terms of Euclidean for example the cosine distance. In contrast the kmedoids algorithm can work with non Euclidean distance measures such as kdiff and is also advantageous because the obtained cluster centers belong to one of the input data points thereby leading to greater interpretability of the results. For these reasons in this paper we consider kmedoids clustering with <italic>k</italic>&#x20;&#x3d; 2 classes and measure the accuracy of clustering for distance measures kdiff, MMD [<xref ref-type="bibr" rid="B9">9</xref>], MPdist [<xref ref-type="bibr" rid="B7">7</xref>] and dtw [<xref ref-type="bibr" rid="B1">1</xref>, <xref ref-type="bibr" rid="B2">2</xref>] over synthetic and real time series and random field datasets as described in the following sections. Suitably chosen combinations of the parameters can be specified as described in <xref ref-type="sec" rid="s3">Section 3</xref> and the derived optimal values can then be used for measuring clustering performance with the test data using kdiff. Similar to kdiff, distances measures using Maximum Mean Discrepancy (MMD) and MPdist are computed by first creating subsequences over the original time series or random fields. In both these cases the length of the sliding window <italic>SL</italic> is determined based on the best clustering performance over a training set selected from the original data. Additionally for MMD which is also a kernel-based measure we determine the optimal kernel bandwidth (<italic>&#x3c3;</italic>) based on training.</p>
<p>In this work we consider two synthetic and one real-life datasets for measuring clustering performance with four distance measures kdiff, MMD, MPdist and dtwd (Dynamic Time Warping distance). For the synthetic datasets we generate the foregrounds and backgrounds as described in <xref ref-type="sec" rid="s2-3">Section 2.3</xref> using autoregressive models of order <italic>p</italic>, denoted as AR(<italic>p</italic>). These are models for a time series <italic>W</italic>
<sub>
<italic>t</italic>
</sub> generated by<disp-formula id="equ10">
<mml:math id="m101">
<mml:msub>
<mml:mrow>
<mml:mi>W</mml:mi>
</mml:mrow>
<mml:mrow>
<mml:mi>t</mml:mi>
</mml:mrow>
</mml:msub>
<mml:mo>&#x3d;</mml:mo>
<mml:munderover accentunder="false" accent="false">
<mml:mrow>
<mml:mo>&#x2211;</mml:mo>
</mml:mrow>
<mml:mrow>
<mml:mi>i</mml:mi>
<mml:mo>&#x3d;</mml:mo>
<mml:mn>1</mml:mn>
</mml:mrow>
<mml:mrow>
<mml:mi>p</mml:mi>
</mml:mrow>
</mml:munderover>
<mml:msub>
<mml:mrow>
<mml:mi>&#x3d5;</mml:mi>
</mml:mrow>
<mml:mrow>
<mml:mi>i</mml:mi>
</mml:mrow>
</mml:msub>
<mml:msub>
<mml:mrow>
<mml:mi>W</mml:mi>
</mml:mrow>
<mml:mrow>
<mml:mi>t</mml:mi>
<mml:mo>&#x2212;</mml:mo>
<mml:mi>i</mml:mi>
</mml:mrow>
</mml:msub>
<mml:mo>&#x2b;</mml:mo>
<mml:msub>
<mml:mrow>
<mml:mi>&#x3f5;</mml:mi>
</mml:mrow>
<mml:mrow>
<mml:mi>t</mml:mi>
</mml:mrow>
</mml:msub>
<mml:mo>,</mml:mo>
<mml:mspace width="2em"/>
<mml:mi>t</mml:mi>
<mml:mo>&#x3d;</mml:mo>
<mml:mi>p</mml:mi>
<mml:mo>&#x2b;</mml:mo>
<mml:mn>1</mml:mn>
<mml:mo>,</mml:mo>
<mml:mo>&#x2026;</mml:mo>
<mml:mo>.</mml:mo>
</mml:math>
</disp-formula>where <italic>&#x3d5;</italic>
<sub>1</sub>, &#x2026; , <italic>&#x3d5;</italic>
<sub>
<italic>p</italic>
</sub> are the <italic>p</italic> coefficients of the AR(<italic>p</italic>) model and <italic>&#x3f5;</italic>
<sub>
<italic>t</italic>
</sub> can be i. i.d. Gaussian errors. We perform 50 Monte Carlo runs over each dataset and in each run we randomly divide the data into training and test sets. For each set of training data we determine the optimal values of the algorithm parameters based on the best clustering performance. Following this we use these parameter values on the test data in each of the 50 runs. The final performance metric for a given distance measure is given by the total number of clustering errors for the test data over all 50 runs. The dtwclust package [<xref ref-type="bibr" rid="B17">17</xref>] of R 3.6.2 has been used for implementation of the kmedoids clustering algorithm and to evaluate the results of clustering.</p>
<p>As a techincal note, as MPdist and MMD are generally computed as squared distances, we similarly work with <inline-formula id="inf76">
<mml:math id="m102">
<mml:mi mathvariant="bold">k</mml:mi>
<mml:mi mathvariant="bold">d</mml:mi>
<mml:mi mathvariant="bold">i</mml:mi>
<mml:mi mathvariant="bold">ff</mml:mi>
<mml:msup>
<mml:mrow>
<mml:mrow>
<mml:mo stretchy="false">(</mml:mo>
<mml:mrow>
<mml:msub>
<mml:mrow>
<mml:mi>&#x3bc;</mml:mi>
</mml:mrow>
<mml:mrow>
<mml:mn>1</mml:mn>
</mml:mrow>
</mml:msub>
<mml:mo>,</mml:mo>
<mml:msub>
<mml:mrow>
<mml:mi>&#x3bc;</mml:mi>
</mml:mrow>
<mml:mrow>
<mml:mn>2</mml:mn>
</mml:mrow>
</mml:msub>
<mml:mo>;</mml:mo>
<mml:mi>&#x3b1;</mml:mi>
</mml:mrow>
<mml:mo stretchy="false">)</mml:mo>
</mml:mrow>
</mml:mrow>
<mml:mrow>
<mml:mn>2</mml:mn>
</mml:mrow>
</mml:msup>
</mml:math>
</inline-formula> as the distance between distributions. This is solely to ensure that the distances are based of Euclidean or kernel distances squares, and to ensure a fair comparison being fed into the kmedoids clustering algorithm. Also as mentioned previously, computing <inline-formula id="inf77">
<mml:math id="m103">
<mml:mi mathvariant="bold">k</mml:mi>
<mml:mi mathvariant="bold">d</mml:mi>
<mml:mi mathvariant="bold">i</mml:mi>
<mml:mi mathvariant="bold">ff</mml:mi>
<mml:msup>
<mml:mrow>
<mml:mrow>
<mml:mo stretchy="false">(</mml:mo>
<mml:mrow>
<mml:msub>
<mml:mrow>
<mml:mi>&#x3bc;</mml:mi>
</mml:mrow>
<mml:mrow>
<mml:mn>1</mml:mn>
</mml:mrow>
</mml:msub>
<mml:mo>,</mml:mo>
<mml:msub>
<mml:mrow>
<mml:mi>&#x3bc;</mml:mi>
</mml:mrow>
<mml:mrow>
<mml:mn>2</mml:mn>
</mml:mrow>
</mml:msub>
<mml:mo>;</mml:mo>
<mml:mi>&#x3b1;</mml:mi>
</mml:mrow>
<mml:mo stretchy="false">)</mml:mo>
</mml:mrow>
</mml:mrow>
<mml:mrow>
<mml:mn>2</mml:mn>
</mml:mrow>
</mml:msup>
</mml:math>
</inline-formula> is equivalent to computing kdiff on the square of the witness function &#x7c;<italic>U</italic> (<italic>&#x3bc;</italic>
<sub>1</sub> &#x2212; <italic>&#x3bc;</italic>
<sub>2</sub>) (<italic>z</italic>)&#x7c;<sup>2</sup>, since quantiles depend only on the ordering of the underlying function.</p>
<sec id="s4-1">
<title>4.1 Simulation: Matching Sub-regions in Univariate Time Series</title>
<p>Data <italic>Y</italic>
<sub>
<italic>i</italic>
</sub> for <italic>i</italic>&#x20;&#x3d; 1, 2, &#x2026; , 1,000 are simulated using the model (15). To generate this the series <italic>W</italic>
<sub>
<italic>i</italic>
</sub> are constructed via an AR (5) model driven by i. i.d errors <inline-formula id="inf78">
<mml:math id="m104">
<mml:mo>&#x223c;</mml:mo>
<mml:mi>N</mml:mi>
<mml:mrow>
<mml:mo stretchy="false">(</mml:mo>
<mml:mrow>
<mml:mn>0,1</mml:mn>
</mml:mrow>
<mml:mo stretchy="false">)</mml:mo>
</mml:mrow>
</mml:math>
</inline-formula>. The AR (5) coefficients are set to 0.5, 0.1, 0.1, 0.1, 0.1.<disp-formula id="e15">
<mml:math id="m105">
<mml:msub>
<mml:mrow>
<mml:mi>Y</mml:mi>
</mml:mrow>
<mml:mrow>
<mml:mi>i</mml:mi>
</mml:mrow>
</mml:msub>
<mml:mo>&#x3d;</mml:mo>
<mml:mi>&#x3bc;</mml:mi>
<mml:mo>&#x2b;</mml:mo>
<mml:msub>
<mml:mrow>
<mml:mi>W</mml:mi>
</mml:mrow>
<mml:mrow>
<mml:mi>i</mml:mi>
</mml:mrow>
</mml:msub>
</mml:math>
<label>(15)</label>
</disp-formula>
</p>
<p>Following this we form a <italic>background</italic> dataset <inline-formula id="inf79">
<mml:math id="m106">
<mml:msub>
<mml:mrow>
<mml:mi>X</mml:mi>
</mml:mrow>
<mml:mrow>
<mml:msub>
<mml:mrow>
<mml:mi>B</mml:mi>
</mml:mrow>
<mml:mrow>
<mml:mi>j</mml:mi>
</mml:mrow>
</mml:msub>
</mml:mrow>
</mml:msub>
</mml:math>
</inline-formula> by generating <italic>j</italic>&#x20;&#x3d; 1, 2, &#x2026; , 21 realizations of this data where the mean <italic>&#x3bc;</italic>
<sub>
<italic>j</italic>
</sub> for realization <italic>j</italic> is set as below:<disp-formula id="equ11">
<mml:math id="m107">
<mml:msub>
<mml:mrow>
<mml:mi>&#x3bc;</mml:mi>
</mml:mrow>
<mml:mrow>
<mml:mi>j</mml:mi>
</mml:mrow>
</mml:msub>
<mml:mo>&#x3d;</mml:mo>
<mml:mfenced open="{" close="">
<mml:mrow>
<mml:mtable class="cases">
<mml:mtr>
<mml:mtd columnalign="left">
<mml:mn>100</mml:mn>
<mml:mo>&#x2217;</mml:mo>
<mml:mi>j</mml:mi>
<mml:mspace width="1em"/>
</mml:mtd>
<mml:mtd columnalign="left">
<mml:mtext>if&#x2009;</mml:mtext>
<mml:mi>j</mml:mi>
<mml:mo>&#x2265;</mml:mo>
<mml:mn>1</mml:mn>
<mml:mspace width="0.3333em"/>
<mml:mtext>and&#x2009;</mml:mtext>
<mml:mspace width="0.3333em"/>
<mml:mi>j</mml:mi>
<mml:mo>&#x2264;</mml:mo>
<mml:mn>10</mml:mn>
</mml:mtd>
</mml:mtr>
<mml:mtr>
<mml:mtd columnalign="left">
<mml:mn>100</mml:mn>
<mml:mo>&#x2217;</mml:mo>
<mml:mfenced open="(" close=")">
<mml:mrow>
<mml:mn>10</mml:mn>
<mml:mo>&#x2212;</mml:mo>
<mml:mi>j</mml:mi>
</mml:mrow>
</mml:mfenced>
<mml:mspace width="1em"/>
</mml:mtd>
<mml:mtd columnalign="left">
<mml:mtext>if&#x2009;</mml:mtext>
<mml:mi>j</mml:mi>
<mml:mo>&#x2265;</mml:mo>
<mml:mn>11</mml:mn>
<mml:mspace width="0.3333em"/>
<mml:mtext>and&#x2009;</mml:mtext>
<mml:mspace width="0.3333em"/>
<mml:mi>j</mml:mi>
<mml:mo>&#x2264;</mml:mo>
<mml:mn>20</mml:mn>
</mml:mtd>
</mml:mtr>
<mml:mtr>
<mml:mtd columnalign="left">
<mml:mn>0</mml:mn>
<mml:mspace width="1em"/>
</mml:mtd>
<mml:mtd columnalign="left">
<mml:mtext>if&#x2009;</mml:mtext>
<mml:mi>j</mml:mi>
<mml:mo>&#x3d;</mml:mo>
<mml:mn>21</mml:mn>
</mml:mtd>
</mml:mtr>
</mml:mtable>
</mml:mrow>
</mml:mfenced>
</mml:math>
</disp-formula>
</p>
<p>Next we generate a dataset <italic>X</italic>
<sub>
<italic>F</italic>
</sub> consisting of 2&#x20;<italic>foregrounds</italic> <italic>X</italic>
<sub>
<italic>FA</italic>
</sub> and <italic>X</italic>
<sub>
<italic>FB</italic>
</sub> which enable forming the 2 classes to be considered for k-medoids clustering as follows. For foreground <italic>X</italic>
<sub>
<italic>FA</italic>
</sub> data <italic>Y</italic>
<sub>
<italic>i</italic>
</sub> for <italic>i</italic>&#x20;&#x3d; 1, 2, &#x2026; , 50 are simulated using the model (15). The series <italic>W</italic>
<sub>
<italic>i</italic>
</sub> is constructed via an AR (1) model driven by i. i.d errors <inline-formula id="inf80">
<mml:math id="m108">
<mml:mo>&#x223c;</mml:mo>
<mml:mi>N</mml:mi>
<mml:mrow>
<mml:mo stretchy="false">(</mml:mo>
<mml:mrow>
<mml:mn>0,1</mml:mn>
</mml:mrow>
<mml:mo stretchy="false">)</mml:mo>
</mml:mrow>
</mml:math>
</inline-formula>. The AR (1) coefficient is set to 0.1 and <italic>&#x3bc;</italic> &#x3d; 10. For foreground <italic>X</italic>
<sub>
<italic>FB</italic>
</sub> data <italic>Y</italic>
<sub>
<italic>i</italic>
</sub> for <italic>i</italic>&#x20;&#x3d; 1, 2, &#x2026; , 25 are simulated using the model (15). The series <italic>W</italic>
<sub>
<italic>i</italic>
</sub> is constructed via an AR (1) model driven by i. i.d errors <inline-formula id="inf81">
<mml:math id="m109">
<mml:mo>&#x223c;</mml:mo>
<mml:mi>N</mml:mi>
<mml:mrow>
<mml:mo stretchy="false">(</mml:mo>
<mml:mrow>
<mml:mn>0,1</mml:mn>
</mml:mrow>
<mml:mo stretchy="false">)</mml:mo>
</mml:mrow>
</mml:math>
</inline-formula>. The AR (1) coefficient is set to 0.1 and <italic>&#x3bc;</italic> &#x3d; &#x2212; 10. We then form the <italic>foreground</italic> dataset <inline-formula id="inf82">
<mml:math id="m110">
<mml:msub>
<mml:mrow>
<mml:mi>X</mml:mi>
</mml:mrow>
<mml:mrow>
<mml:msub>
<mml:mrow>
<mml:mi>F</mml:mi>
</mml:mrow>
<mml:mrow>
<mml:mi>j</mml:mi>
</mml:mrow>
</mml:msub>
</mml:mrow>
</mml:msub>
</mml:math>
</inline-formula> by generating <italic>j</italic>&#x20;&#x3d; 1, 2, &#x2026; , 21 realizations of this data as follows:<disp-formula id="equ12">
<mml:math id="m111">
<mml:msub>
<mml:mrow>
<mml:mi>X</mml:mi>
</mml:mrow>
<mml:mrow>
<mml:msub>
<mml:mrow>
<mml:mi>F</mml:mi>
</mml:mrow>
<mml:mrow>
<mml:mi>j</mml:mi>
</mml:mrow>
</mml:msub>
</mml:mrow>
</mml:msub>
<mml:mo>&#x3d;</mml:mo>
<mml:mfenced open="{" close="">
<mml:mrow>
<mml:mtable class="cases">
<mml:mtr>
<mml:mtd columnalign="left">
<mml:msub>
<mml:mrow>
<mml:mi>X</mml:mi>
</mml:mrow>
<mml:mrow>
<mml:mi>F</mml:mi>
<mml:mi>A</mml:mi>
</mml:mrow>
</mml:msub>
<mml:mspace width="1em"/>
</mml:mtd>
<mml:mtd columnalign="left">
<mml:mtext>if&#x2009;</mml:mtext>
<mml:mi>j</mml:mi>
<mml:mspace width="0.3333em"/>
<mml:mtext>mod&#x2009;</mml:mtext>
<mml:mn>2</mml:mn>
<mml:mtext>&#x2009;</mml:mtext>
<mml:mo>&#x3d;</mml:mo>
<mml:mn>1</mml:mn>
</mml:mtd>
</mml:mtr>
<mml:mtr>
<mml:mtd columnalign="left">
<mml:msub>
<mml:mrow>
<mml:mi>X</mml:mi>
</mml:mrow>
<mml:mrow>
<mml:mi>F</mml:mi>
<mml:mi>B</mml:mi>
</mml:mrow>
</mml:msub>
<mml:mspace width="1em"/>
</mml:mtd>
<mml:mtd columnalign="left">
<mml:mtext>if&#x2009;</mml:mtext>
<mml:mi>j</mml:mi>
<mml:mspace width="0.3333em"/>
<mml:mtext>mod&#x2009;</mml:mtext>
<mml:mn>2</mml:mn>
<mml:mtext>&#x2009;</mml:mtext>
<mml:mo>&#x3d;</mml:mo>
<mml:mn>0</mml:mn>
</mml:mtd>
</mml:mtr>
</mml:mtable>
</mml:mrow>
</mml:mfenced>
</mml:math>
</disp-formula>
</p>
<p>Finally the dataset used for clustering <italic>Z</italic>
<sub>
<italic>ij</italic>
</sub> where <italic>i</italic>&#x20;&#x3d; 1, 2, &#x2026; , 1,000 and <italic>j</italic>&#x20;&#x3d; 1, 2, &#x2026; , 21 is formed by mixing <italic>backgrounds</italic> <italic>X</italic>
<sub>
<italic>B</italic>
</sub> and <italic>foregrounds</italic> <italic>X</italic>
<sub>
<italic>F</italic>
</sub> as follows:<disp-formula id="equ13">
<mml:math id="m112">
<mml:msub>
<mml:mrow>
<mml:mi>Z</mml:mi>
</mml:mrow>
<mml:mrow>
<mml:mi>i</mml:mi>
<mml:mi>j</mml:mi>
</mml:mrow>
</mml:msub>
<mml:mo>&#x3d;</mml:mo>
<mml:mfenced open="{" close="">
<mml:mrow>
<mml:mtable class="cases">
<mml:mtr>
<mml:mtd columnalign="left">
<mml:munderover accentunder="false" accent="false">
<mml:mrow>
<mml:mo>&#x2211;</mml:mo>
</mml:mrow>
<mml:mrow>
<mml:mi>i</mml:mi>
<mml:mo>&#x3d;</mml:mo>
<mml:mn>1</mml:mn>
</mml:mrow>
<mml:mrow>
<mml:mn>50</mml:mn>
</mml:mrow>
</mml:munderover>
<mml:msub>
<mml:mrow>
<mml:mi>X</mml:mi>
</mml:mrow>
<mml:mrow>
<mml:msub>
<mml:mrow>
<mml:mi>F</mml:mi>
</mml:mrow>
<mml:mrow>
<mml:mi>i</mml:mi>
<mml:mi>j</mml:mi>
</mml:mrow>
</mml:msub>
</mml:mrow>
</mml:msub>
<mml:mo>&#x2b;</mml:mo>
<mml:munderover accentunder="false" accent="false">
<mml:mrow>
<mml:mo>&#x2211;</mml:mo>
</mml:mrow>
<mml:mrow>
<mml:mi>i</mml:mi>
<mml:mo>&#x3d;</mml:mo>
<mml:mn>51</mml:mn>
</mml:mrow>
<mml:mrow>
<mml:mn>1000</mml:mn>
</mml:mrow>
</mml:munderover>
<mml:msub>
<mml:mrow>
<mml:mi>X</mml:mi>
</mml:mrow>
<mml:mrow>
<mml:msub>
<mml:mrow>
<mml:mi>B</mml:mi>
</mml:mrow>
<mml:mrow>
<mml:mi>i</mml:mi>
<mml:mi>j</mml:mi>
</mml:mrow>
</mml:msub>
</mml:mrow>
</mml:msub>
<mml:mspace width="1em"/>
</mml:mtd>
<mml:mtd columnalign="left">
<mml:mtext>if&#x2009;</mml:mtext>
<mml:mi>j</mml:mi>
<mml:mspace width="0.3333em"/>
<mml:mtext>mod&#x2009;</mml:mtext>
<mml:mn>2</mml:mn>
<mml:mtext>&#x2009;</mml:mtext>
<mml:mo>&#x3d;</mml:mo>
<mml:mn>1</mml:mn>
</mml:mtd>
</mml:mtr>
<mml:mtr>
<mml:mtd columnalign="left">
<mml:munderover accentunder="false" accent="false">
<mml:mrow>
<mml:mo>&#x2211;</mml:mo>
</mml:mrow>
<mml:mrow>
<mml:mi>i</mml:mi>
<mml:mo>&#x3d;</mml:mo>
<mml:mn>1</mml:mn>
</mml:mrow>
<mml:mrow>
<mml:mn>25</mml:mn>
</mml:mrow>
</mml:munderover>
<mml:msub>
<mml:mrow>
<mml:mi>X</mml:mi>
</mml:mrow>
<mml:mrow>
<mml:msub>
<mml:mrow>
<mml:mi>F</mml:mi>
</mml:mrow>
<mml:mrow>
<mml:mi>i</mml:mi>
<mml:mi>j</mml:mi>
</mml:mrow>
</mml:msub>
</mml:mrow>
</mml:msub>
<mml:mo>&#x2b;</mml:mo>
<mml:munderover accentunder="false" accent="false">
<mml:mrow>
<mml:mo>&#x2211;</mml:mo>
</mml:mrow>
<mml:mrow>
<mml:mi>i</mml:mi>
<mml:mo>&#x3d;</mml:mo>
<mml:mn>26</mml:mn>
</mml:mrow>
<mml:mrow>
<mml:mn>1000</mml:mn>
</mml:mrow>
</mml:munderover>
<mml:msub>
<mml:mrow>
<mml:mi>X</mml:mi>
</mml:mrow>
<mml:mrow>
<mml:msub>
<mml:mrow>
<mml:mi>B</mml:mi>
</mml:mrow>
<mml:mrow>
<mml:mi>i</mml:mi>
<mml:mi>j</mml:mi>
</mml:mrow>
</mml:msub>
</mml:mrow>
</mml:msub>
<mml:mspace width="1em"/>
</mml:mtd>
<mml:mtd columnalign="left">
<mml:mtext>if&#x2009;</mml:mtext>
<mml:mi>j</mml:mi>
<mml:mspace width="0.3333em"/>
<mml:mtext>mod&#x2009;</mml:mtext>
<mml:mn>2</mml:mn>
<mml:mtext>&#x2009;</mml:mtext>
<mml:mo>&#x3d;</mml:mo>
<mml:mn>0</mml:mn>
</mml:mtd>
</mml:mtr>
</mml:mtable>
</mml:mrow>
</mml:mfenced>
</mml:math>
</disp-formula>
</p>
<p>The dataset <italic>Z</italic>
<sub>
<italic>ij</italic>
</sub> formed in this manner consists of two types of subregions (<italic>foregrounds</italic>) which define the two classes used for k-medoids clustering. We perform 50 random splits of the dataset <italic>Z</italic>
<sub>
<italic>ij</italic>
</sub> where each split consists of a training set of size 10 and a test set of size 11. The results for clustering are shown for the 4 distance measures in <xref ref-type="table" rid="T1">Table&#x20;1</xref>.</p>
<table-wrap id="T1" position="float">
<label>TABLE 1</label>
<caption>
<p>Clustering Performance for Univariate Time Series dataset with <italic>&#x3c4;</italic> &#x3d; 1</p>
</caption>
<table>
<thead valign="top">
<tr>
<th align="left">Distance measure</th>
<th align="center">Total number of errors</th>
<th align="center">Percent error</th>
</tr>
</thead>
<tbody valign="top">
<tr>
<td align="left">kdiff</td>
<td align="char" char=".">0</td>
<td align="char" char=".">0</td>
</tr>
<tr>
<td align="left">MMD</td>
<td align="char" char=".">219</td>
<td align="char" char=".">39.8</td>
</tr>
<tr>
<td align="left">MPdist</td>
<td align="char" char=".">0</td>
<td align="char" char=".">0</td>
</tr>
<tr>
<td align="left">Dtwd</td>
<td align="char" char=".">227</td>
<td align="char" char=".">41.2</td>
</tr>
</tbody>
</table>
</table-wrap>
<p>From the results it can be seen that both kdiff and MPdist produce the best clustering performance with 0 errors for this dataset. This is attributed to the fact that the subregions of interest are well defined for both classes and using suitable values of parameters determined from training it is possible to accurately cluster all the time series data into two separate groups. On the other hand the performance of MMD is inferior to both kdiff and MPdist because the <italic>backgrounds</italic> are well separated with different mean values for time series within and across the two classes. This results in time series even belonging to the same class to be placed in separate clusters when MMD is used as a distance measure. Similarly dtwd suffers from poor performance as this distance measure tends to place time series with smaller separation between the mean background values in the same cluster. However these may have distinct values for the <italic>foregrounds</italic> i.e.,&#x20;they can in general belong to different classes and as a result this causes errors during clustering.</p>
<p>
<italic>Noise</italic> <italic>robustness</italic> We explore the performance of the distance measures by considering noisy foregrounds. For foreground <italic>X</italic>
<sub>
<italic>FA</italic>
</sub> data <italic>Y</italic>
<sub>
<italic>i</italic>
</sub> for <italic>i</italic>&#x20;&#x3d; 1, 2, &#x2026; , 50 are simulated using the model (15). The series <italic>W</italic>
<sub>
<italic>i</italic>
</sub> is constructed via an AR (1) model driven by i. i.d errors <inline-formula id="inf83">
<mml:math id="m113">
<mml:mo>&#x223c;</mml:mo>
<mml:mi>N</mml:mi>
<mml:mrow>
<mml:mo stretchy="false">(</mml:mo>
<mml:mrow>
<mml:mn>0,100</mml:mn>
</mml:mrow>
<mml:mo stretchy="false">)</mml:mo>
</mml:mrow>
</mml:math>
</inline-formula>. The AR (1) coefficient is set to 0.1 and <italic>&#x3bc;</italic> &#x3d; 10. For foreground <italic>X</italic>
<sub>
<italic>FB</italic>
</sub> data <italic>Y</italic>
<sub>
<italic>i</italic>
</sub> for <italic>i</italic>&#x20;&#x3d; 1, 2, &#x2026; , 25 are simulated using the model (15). The series <italic>W</italic>
<sub>
<italic>i</italic>
</sub> is constructed via an AR (1) model driven by i. i.d errors <inline-formula id="inf84">
<mml:math id="m114">
<mml:mo>&#x223c;</mml:mo>
<mml:mi>N</mml:mi>
<mml:mrow>
<mml:mo stretchy="false">(</mml:mo>
<mml:mrow>
<mml:mn>0,1</mml:mn>
</mml:mrow>
<mml:mo stretchy="false">)</mml:mo>
</mml:mrow>
</mml:math>
</inline-formula>. The AR (1) coefficient is set to 0.1 and <italic>&#x3bc;</italic> &#x3d; &#x2212; 10. Following this the foreground datasets <inline-formula id="inf85">
<mml:math id="m115">
<mml:msub>
<mml:mrow>
<mml:mi>X</mml:mi>
</mml:mrow>
<mml:mrow>
<mml:msub>
<mml:mrow>
<mml:mi>F</mml:mi>
</mml:mrow>
<mml:mrow>
<mml:mi>j</mml:mi>
</mml:mrow>
</mml:msub>
</mml:mrow>
</mml:msub>
</mml:math>
</inline-formula> and the dataset used for clustering <italic>Z</italic>
<sub>
<italic>ij</italic>
</sub> where <italic>i</italic>&#x20;&#x3d; 1, 2, &#x2026; , 1,000 and <italic>j</italic>&#x20;&#x3d; 1, 2, &#x2026; , 21 are formed in the same manner as described earlier. We show example time series realizations for <italic>&#x3c4;</italic> &#x3d; 1 and 10 in <xref ref-type="fig" rid="F1">Figures 1</xref>, <xref ref-type="fig" rid="F2">2</xref> respectively. Each figure contains plots of two time series with mean &#x3d; &#x2212; 10, 10 as per the construction of foreground <italic>X</italic>
<sub>
<italic>FB</italic>
</sub> for the original and noisy case and show the relative separation between the realizations.</p>
<fig id="F1" position="float">
<label>FIGURE 1</label>
<caption>
<p>Foreground Time Series Realization with mean &#x3d; 10, &#x2212; 10 for <italic>&#x3c4;</italic> &#x3d; 1</p>
</caption>
<graphic xlink:href="fams-07-787455-g001.tif"/>
</fig>
<fig id="F2" position="float">
<label>FIGURE 2</label>
<caption>
<p>Foreground Time Series Realizations with mean &#x3d; 10, &#x2212; 10 for <italic>&#x3c4;</italic> &#x3d; 10.</p>
</caption>
<graphic xlink:href="fams-07-787455-g002.tif"/>
</fig>
<p>The results for clustering using these noisy foregrounds are shown in <xref ref-type="table" rid="T2">Table&#x20;2</xref>. The data shows empirically that as the noise level of the foreground increases kdiff is more resilient and performs better than MPdist. This is because after constructing sliding window based embeddings over the original data, MPdist is computed using Euclidean metric based cross-similarities between the embeddings whereas kdiff is estimated using kernel based self and cross similarities over the embeddings.</p>
<table-wrap id="T2" position="float">
<label>TABLE 2</label>
<caption>
<p>Clustering Performance for Univariate Time Series dataset with <italic>&#x3c4;</italic> &#x3d; 10</p>
</caption>
<table>
<thead valign="top">
<tr>
<th align="left">Distance measure</th>
<th align="center">Total number of errors</th>
<th align="center">Percent error</th>
</tr>
</thead>
<tbody valign="top">
<tr>
<td align="left">kdiff</td>
<td align="char" char=".">20</td>
<td align="char" char=".">3.6</td>
</tr>
<tr>
<td align="left">MMD</td>
<td align="char" char=".">219</td>
<td align="char" char=".">39.8</td>
</tr>
<tr>
<td align="left">MPdist</td>
<td align="char" char=".">64</td>
<td align="char" char=".">11.6</td>
</tr>
<tr>
<td align="left">Dtwd</td>
<td align="char" char=".">220</td>
<td align="char" char=".">40.0</td>
</tr>
</tbody>
</table>
</table-wrap>
</sec>
<sec id="s4-2">
<title>4.2 Simulation: Matching Sub-regions in 3-Dimensional Time Series in Spherical Coordinates</title>
<p>We generate a 3d multivariate <italic>background</italic> dataset <italic>s</italic>
<sub>
<italic>B</italic>
</sub> as follows. Data <inline-formula id="inf86">
<mml:math id="m116">
<mml:msub>
<mml:mrow>
<mml:mi>Y</mml:mi>
</mml:mrow>
<mml:mrow>
<mml:msub>
<mml:mrow>
<mml:mi>a</mml:mi>
</mml:mrow>
<mml:mrow>
<mml:mi>i</mml:mi>
</mml:mrow>
</mml:msub>
</mml:mrow>
</mml:msub>
</mml:math>
</inline-formula> for <italic>i</italic>&#x20;&#x3d; 1, 2, &#x2026; , 1,000 are simulated using the model (15). To generate this the series <italic>W</italic>
<sub>
<italic>i</italic>
</sub> is constructed via an AR (1) model driven by i. i.d errors <inline-formula id="inf87">
<mml:math id="m117">
<mml:mo>&#x223c;</mml:mo>
<mml:mi>N</mml:mi>
<mml:mrow>
<mml:mo stretchy="false">(</mml:mo>
<mml:mrow>
<mml:mn>0,1</mml:mn>
</mml:mrow>
<mml:mo stretchy="false">)</mml:mo>
</mml:mrow>
</mml:math>
</inline-formula>. The AR (1) coefficient is set to 0.1. Similarly data <inline-formula id="inf88">
<mml:math id="m118">
<mml:msub>
<mml:mrow>
<mml:mi>Y</mml:mi>
</mml:mrow>
<mml:mrow>
<mml:msub>
<mml:mrow>
<mml:mi>b</mml:mi>
</mml:mrow>
<mml:mrow>
<mml:mi>i</mml:mi>
</mml:mrow>
</mml:msub>
</mml:mrow>
</mml:msub>
</mml:math>
</inline-formula> for <italic>i</italic>&#x20;&#x3d; 1, 2, &#x2026; , 1,000 are simulated using the model (15). To generate this the series <italic>W</italic>
<sub>
<italic>i</italic>
</sub> is constructed via an AR (1) model driven by i. i.d errors <inline-formula id="inf89">
<mml:math id="m119">
<mml:mo>&#x223c;</mml:mo>
<mml:mi>N</mml:mi>
<mml:mrow>
<mml:mo stretchy="false">(</mml:mo>
<mml:mrow>
<mml:mn>0,1</mml:mn>
</mml:mrow>
<mml:mo stretchy="false">)</mml:mo>
</mml:mrow>
</mml:math>
</inline-formula>. The AR (1) coefficient is set to&#x20;0.1.</p>
<p>Following the generation of <italic>W</italic>
<sub>
<italic>i</italic>
</sub> values for the data <bold>Y</bold> &#x3d; (<italic>Y</italic>
<sub>
<italic>a</italic>
</sub>, <italic>Y</italic>
<sub>
<italic>b</italic>
</sub>) we form a <italic>background</italic> dataset <inline-formula id="inf90">
<mml:math id="m120">
<mml:msub>
<mml:mrow>
<mml:mi>X</mml:mi>
</mml:mrow>
<mml:mrow>
<mml:msub>
<mml:mrow>
<mml:mi>B</mml:mi>
</mml:mrow>
<mml:mrow>
<mml:mi>j</mml:mi>
</mml:mrow>
</mml:msub>
</mml:mrow>
</mml:msub>
</mml:math>
</inline-formula> in this 2d space by generating <italic>j</italic>&#x20;&#x3d; 1, 2, &#x2026; , 21 realizations of this data <bold>Y</bold> where the mean <italic>&#x3bc;</italic> for realization <italic>j</italic> of each pair is set as below:<disp-formula id="equ14">
<mml:math id="m121">
<mml:mi>&#x3bc;</mml:mi>
<mml:mo>&#x3d;</mml:mo>
<mml:mfenced open="{" close="">
<mml:mrow>
<mml:mtable class="cases">
<mml:mtr>
<mml:mtd columnalign="left">
<mml:mn>100</mml:mn>
<mml:mo>&#x2217;</mml:mo>
<mml:mi>j</mml:mi>
<mml:mspace width="1em"/>
</mml:mtd>
<mml:mtd columnalign="left">
<mml:mtext>if&#x2009;</mml:mtext>
<mml:mi>j</mml:mi>
<mml:mo>&#x2265;</mml:mo>
<mml:mn>1</mml:mn>
<mml:mspace width="0.3333em"/>
<mml:mtext>and&#x2009;</mml:mtext>
<mml:mspace width="0.3333em"/>
<mml:mi>j</mml:mi>
<mml:mo>&#x2264;</mml:mo>
<mml:mn>10</mml:mn>
</mml:mtd>
</mml:mtr>
<mml:mtr>
<mml:mtd columnalign="left">
<mml:mn>100</mml:mn>
<mml:mo>&#x2217;</mml:mo>
<mml:mfenced open="(" close=")">
<mml:mrow>
<mml:mn>10</mml:mn>
<mml:mo>&#x2212;</mml:mo>
<mml:mi>j</mml:mi>
</mml:mrow>
</mml:mfenced>
<mml:mspace width="1em"/>
</mml:mtd>
<mml:mtd columnalign="left">
<mml:mtext>if&#x2009;</mml:mtext>
<mml:mi>j</mml:mi>
<mml:mo>&#x2265;</mml:mo>
<mml:mn>11</mml:mn>
<mml:mspace width="0.3333em"/>
<mml:mtext>and&#x2009;</mml:mtext>
<mml:mspace width="0.3333em"/>
<mml:mi>j</mml:mi>
<mml:mo>&#x2264;</mml:mo>
<mml:mn>20</mml:mn>
</mml:mtd>
</mml:mtr>
<mml:mtr>
<mml:mtd columnalign="left">
<mml:mn>0</mml:mn>
<mml:mspace width="1em"/>
</mml:mtd>
<mml:mtd columnalign="left">
<mml:mtext>j</mml:mtext>
<mml:mo>&#x3d;</mml:mo>
<mml:mn>21</mml:mn>
</mml:mtd>
</mml:mtr>
</mml:mtable>
</mml:mrow>
</mml:mfenced>
</mml:math>
</disp-formula>
</p>
<p>Our next step involves transforming these 21 instances of the 2d backgrounds into a 3d spherical surface of radius 1 as described in the following steps. We first map each series <italic>Y</italic>
<sub>
<italic>a</italic>
</sub> and <italic>Y</italic>
<sub>
<italic>b</italic>
</sub> linearly into the region [0, <italic>&#x3c0;</italic>/2]. The corresponding mapped series are denoted as <inline-formula id="inf91">
<mml:math id="m122">
<mml:msub>
<mml:mrow>
<mml:mi>Y</mml:mi>
</mml:mrow>
<mml:mrow>
<mml:msub>
<mml:mrow>
<mml:mi>a</mml:mi>
</mml:mrow>
<mml:mrow>
<mml:mi>s</mml:mi>
</mml:mrow>
</mml:msub>
</mml:mrow>
</mml:msub>
</mml:math>
</inline-formula> and <inline-formula id="inf92">
<mml:math id="m123">
<mml:msub>
<mml:mrow>
<mml:mi>Y</mml:mi>
</mml:mrow>
<mml:mrow>
<mml:msub>
<mml:mrow>
<mml:mi>b</mml:mi>
</mml:mrow>
<mml:mrow>
<mml:mi>s</mml:mi>
</mml:mrow>
</mml:msub>
</mml:mrow>
</mml:msub>
</mml:math>
</inline-formula> respectively. To ensure that the <italic>backgrounds</italic> are clearly separated we divide the region [0, <italic>&#x3c0;</italic>/2] into 21 nonoverlapping partitions for this linear mapping. The final background dataset <italic>s</italic>
<sub>
<italic>B</italic>
</sub> &#x3d; {<italic>s</italic>
<sub>
<italic>a</italic>
</sub>, <italic>s</italic>
<sub>
<italic>b</italic>
</sub>, <italic>s</italic>
<sub>
<italic>c</italic>
</sub>} is derived using :<disp-formula id="e16">
<mml:math id="m124">
<mml:mtable class="aligned">
<mml:mtr>
<mml:mtd columnalign="right">
<mml:msub>
<mml:mrow>
<mml:mi>s</mml:mi>
</mml:mrow>
<mml:mrow>
<mml:mi>a</mml:mi>
</mml:mrow>
</mml:msub>
</mml:mtd>
<mml:mtd columnalign="left">
<mml:mo>&#x3d;</mml:mo>
<mml:mi>sin</mml:mi>
<mml:mfenced open="(" close=")">
<mml:mrow>
<mml:msub>
<mml:mrow>
<mml:mi>Y</mml:mi>
</mml:mrow>
<mml:mrow>
<mml:msub>
<mml:mrow>
<mml:mi>b</mml:mi>
</mml:mrow>
<mml:mrow>
<mml:mi>s</mml:mi>
</mml:mrow>
</mml:msub>
</mml:mrow>
</mml:msub>
</mml:mrow>
</mml:mfenced>
<mml:mo>&#x2217;</mml:mo>
<mml:mi>cos</mml:mi>
<mml:mfenced open="(" close=")">
<mml:mrow>
<mml:msub>
<mml:mrow>
<mml:mi>Y</mml:mi>
</mml:mrow>
<mml:mrow>
<mml:msub>
<mml:mrow>
<mml:mi>a</mml:mi>
</mml:mrow>
<mml:mrow>
<mml:mi>s</mml:mi>
</mml:mrow>
</mml:msub>
</mml:mrow>
</mml:msub>
</mml:mrow>
</mml:mfenced>
</mml:mtd>
</mml:mtr>
<mml:mtr>
<mml:mtd columnalign="right">
<mml:msub>
<mml:mrow>
<mml:mi>s</mml:mi>
</mml:mrow>
<mml:mrow>
<mml:mi>b</mml:mi>
</mml:mrow>
</mml:msub>
</mml:mtd>
<mml:mtd columnalign="left">
<mml:mo>&#x3d;</mml:mo>
<mml:mi>sin</mml:mi>
<mml:mfenced open="(" close=")">
<mml:mrow>
<mml:msub>
<mml:mrow>
<mml:mi>Y</mml:mi>
</mml:mrow>
<mml:mrow>
<mml:msub>
<mml:mrow>
<mml:mi>b</mml:mi>
</mml:mrow>
<mml:mrow>
<mml:mi>s</mml:mi>
</mml:mrow>
</mml:msub>
</mml:mrow>
</mml:msub>
</mml:mrow>
</mml:mfenced>
<mml:mo>&#x2217;</mml:mo>
<mml:mi>sin</mml:mi>
<mml:mfenced open="(" close=")">
<mml:mrow>
<mml:msub>
<mml:mrow>
<mml:mi>Y</mml:mi>
</mml:mrow>
<mml:mrow>
<mml:msub>
<mml:mrow>
<mml:mi>a</mml:mi>
</mml:mrow>
<mml:mrow>
<mml:mi>s</mml:mi>
</mml:mrow>
</mml:msub>
</mml:mrow>
</mml:msub>
</mml:mrow>
</mml:mfenced>
</mml:mtd>
</mml:mtr>
<mml:mtr>
<mml:mtd columnalign="right">
<mml:msub>
<mml:mrow>
<mml:mi>s</mml:mi>
</mml:mrow>
<mml:mrow>
<mml:mi>c</mml:mi>
</mml:mrow>
</mml:msub>
</mml:mtd>
<mml:mtd columnalign="left">
<mml:mo>&#x3d;</mml:mo>
<mml:mi>cos</mml:mi>
<mml:mfenced open="(" close=")">
<mml:mrow>
<mml:msub>
<mml:mrow>
<mml:mi>Y</mml:mi>
</mml:mrow>
<mml:mrow>
<mml:msub>
<mml:mrow>
<mml:mi>b</mml:mi>
</mml:mrow>
<mml:mrow>
<mml:mi>s</mml:mi>
</mml:mrow>
</mml:msub>
</mml:mrow>
</mml:msub>
</mml:mrow>
</mml:mfenced>
</mml:mtd>
</mml:mtr>
</mml:mtable>
</mml:math>
<label>(16)</label>
</disp-formula>
</p>
<p>Next we generate a 3d foreground dataset <italic>s</italic>
<sub>
<italic>F</italic>
</sub> consisting of 2&#x20;<italic>foregrounds</italic> <italic>s</italic>
<sub>
<italic>FA</italic>
</sub> and <italic>s</italic>
<sub>
<italic>FB</italic>
</sub> which will enable forming the 2 classes to be considered for k-medoids clustering as follows. Data <inline-formula id="inf93">
<mml:math id="m125">
<mml:msub>
<mml:mrow>
<mml:mi>Y</mml:mi>
</mml:mrow>
<mml:mrow>
<mml:msub>
<mml:mrow>
<mml:mi>a</mml:mi>
</mml:mrow>
<mml:mrow>
<mml:mi>i</mml:mi>
</mml:mrow>
</mml:msub>
</mml:mrow>
</mml:msub>
</mml:math>
</inline-formula> for <italic>i</italic>&#x20;&#x3d; 1, 2, &#x2026; , 50 are simulated using the model (15). To generate this the series <italic>W</italic>
<sub>
<italic>i</italic>
</sub> is constructed via an AR (1) model driven by i. i.d errors <inline-formula id="inf94">
<mml:math id="m126">
<mml:mo>&#x223c;</mml:mo>
<mml:mi>N</mml:mi>
<mml:mrow>
<mml:mo stretchy="false">(</mml:mo>
<mml:mrow>
<mml:mn>0,1</mml:mn>
</mml:mrow>
<mml:mo stretchy="false">)</mml:mo>
</mml:mrow>
</mml:math>
</inline-formula>. The AR (1) coefficient is set to 0.1 and <italic>&#x3bc;</italic> &#x3d; 10. Similarly data <inline-formula id="inf95">
<mml:math id="m127">
<mml:msub>
<mml:mrow>
<mml:mi>Y</mml:mi>
</mml:mrow>
<mml:mrow>
<mml:msub>
<mml:mrow>
<mml:mi>b</mml:mi>
</mml:mrow>
<mml:mrow>
<mml:mi>i</mml:mi>
</mml:mrow>
</mml:msub>
</mml:mrow>
</mml:msub>
</mml:math>
</inline-formula> for <italic>i</italic>&#x20;&#x3d; 1, 2, &#x2026; , 50 are simulated using the model (15). To generate this the series <italic>W</italic>
<sub>
<italic>i</italic>
</sub> is constructed via an AR (1) model driven by i. i.d errors <inline-formula id="inf96">
<mml:math id="m128">
<mml:mo>&#x223c;</mml:mo>
<mml:mi>N</mml:mi>
<mml:mrow>
<mml:mo stretchy="false">(</mml:mo>
<mml:mrow>
<mml:mn>0,1</mml:mn>
</mml:mrow>
<mml:mo stretchy="false">)</mml:mo>
</mml:mrow>
</mml:math>
</inline-formula>. The AR (1) coefficient is set to 0.1 and <italic>&#x3bc;</italic> &#x3d; 10. we linearly map the original 2d data (<italic>Y</italic>
<sub>
<italic>a</italic>
</sub>, <italic>Y</italic>
<sub>
<italic>b</italic>
</sub>) into the region [<italic>&#x3c0;</italic>/2, 5<italic>&#x3c0;</italic>/8] as <inline-formula id="inf97">
<mml:math id="m129">
<mml:mrow>
<mml:mo stretchy="false">(</mml:mo>
<mml:mrow>
<mml:msub>
<mml:mrow>
<mml:mi>Y</mml:mi>
</mml:mrow>
<mml:mrow>
<mml:msub>
<mml:mrow>
<mml:mi>a</mml:mi>
</mml:mrow>
<mml:mrow>
<mml:mi>s</mml:mi>
</mml:mrow>
</mml:msub>
</mml:mrow>
</mml:msub>
<mml:mo>,</mml:mo>
<mml:msub>
<mml:mrow>
<mml:mi>Y</mml:mi>
</mml:mrow>
<mml:mrow>
<mml:msub>
<mml:mrow>
<mml:mi>b</mml:mi>
</mml:mrow>
<mml:mrow>
<mml:mi>s</mml:mi>
</mml:mrow>
</mml:msub>
</mml:mrow>
</mml:msub>
</mml:mrow>
<mml:mo stretchy="false">)</mml:mo>
</mml:mrow>
</mml:math>
</inline-formula> and then perform the mapping as given in <xref ref-type="disp-formula" rid="e16">Eq. 16</xref> to form the <italic>foreground</italic> <italic>s</italic>
<sub>
<italic>FA</italic>
</sub>. The foreground <italic>s</italic>
<sub>
<italic>FB</italic>
</sub> is generated in a similar manner except that <italic>&#x3bc;</italic> &#x3d; &#x2212; 10 and the 2d series is linearly mapped to the region [3<italic>&#x3c0;</italic>/4, 7<italic>&#x3c0;</italic>/8]. We form the <italic>foreground</italic> dataset <inline-formula id="inf98">
<mml:math id="m130">
<mml:msub>
<mml:mrow>
<mml:mi>s</mml:mi>
</mml:mrow>
<mml:mrow>
<mml:msub>
<mml:mrow>
<mml:mi>F</mml:mi>
</mml:mrow>
<mml:mrow>
<mml:mi>j</mml:mi>
</mml:mrow>
</mml:msub>
</mml:mrow>
</mml:msub>
</mml:math>
</inline-formula> by generating <italic>j</italic>&#x20;&#x3d; 1, 2, &#x2026; , 21 realizations of this data as follows:<disp-formula id="equ15">
<mml:math id="m131">
<mml:msub>
<mml:mrow>
<mml:mi>s</mml:mi>
</mml:mrow>
<mml:mrow>
<mml:msub>
<mml:mrow>
<mml:mi>F</mml:mi>
</mml:mrow>
<mml:mrow>
<mml:mi>j</mml:mi>
</mml:mrow>
</mml:msub>
</mml:mrow>
</mml:msub>
<mml:mo>&#x3d;</mml:mo>
<mml:mfenced open="{" close="">
<mml:mrow>
<mml:mtable class="cases">
<mml:mtr>
<mml:mtd columnalign="left">
<mml:msub>
<mml:mrow>
<mml:mi>s</mml:mi>
</mml:mrow>
<mml:mrow>
<mml:mi>F</mml:mi>
<mml:mi>A</mml:mi>
</mml:mrow>
</mml:msub>
<mml:mspace width="1em"/>
</mml:mtd>
<mml:mtd columnalign="left">
<mml:mtext>if&#x2009;</mml:mtext>
<mml:mi>j</mml:mi>
<mml:mspace width="0.3333em"/>
<mml:mtext>mod&#x2009;</mml:mtext>
<mml:mn>2</mml:mn>
<mml:mtext>&#x2009;</mml:mtext>
<mml:mo>&#x3d;</mml:mo>
<mml:mn>1</mml:mn>
</mml:mtd>
</mml:mtr>
<mml:mtr>
<mml:mtd columnalign="left">
<mml:msub>
<mml:mrow>
<mml:mi>s</mml:mi>
</mml:mrow>
<mml:mrow>
<mml:mi>F</mml:mi>
<mml:mi>B</mml:mi>
</mml:mrow>
</mml:msub>
<mml:mspace width="1em"/>
</mml:mtd>
<mml:mtd columnalign="left">
<mml:mtext>if&#x2009;</mml:mtext>
<mml:mi>j</mml:mi>
<mml:mspace width="0.3333em"/>
<mml:mtext>mod&#x2009;</mml:mtext>
<mml:mn>2</mml:mn>
<mml:mtext>&#x2009;</mml:mtext>
<mml:mo>&#x3d;</mml:mo>
<mml:mn>0</mml:mn>
</mml:mtd>
</mml:mtr>
</mml:mtable>
</mml:mrow>
</mml:mfenced>
</mml:math>
</disp-formula>
</p>
<p>Finally the dataset used for clustering <italic>Z</italic>
<sub>
<italic>ij</italic>
</sub> where <italic>i</italic>&#x20;&#x3d; 1, 2, &#x2026; , 1,000 and <italic>j</italic>&#x20;&#x3d; 1, 2, &#x2026; , 21 is formed by mixing <italic>backgrounds</italic> <italic>s</italic>
<sub>
<italic>B</italic>
</sub> and <italic>foregrounds</italic> <italic>s</italic>
<sub>
<italic>F</italic>
</sub> as follows:<disp-formula id="equ16">
<mml:math id="m132">
<mml:msub>
<mml:mrow>
<mml:mi>Z</mml:mi>
</mml:mrow>
<mml:mrow>
<mml:mi>i</mml:mi>
<mml:mi>j</mml:mi>
</mml:mrow>
</mml:msub>
<mml:mo>&#x3d;</mml:mo>
<mml:mfenced open="{" close="">
<mml:mrow>
<mml:mtable class="cases">
<mml:mtr>
<mml:mtd columnalign="left">
<mml:munderover accentunder="false" accent="false">
<mml:mrow>
<mml:mo>&#x2211;</mml:mo>
</mml:mrow>
<mml:mrow>
<mml:mi>i</mml:mi>
<mml:mo>&#x3d;</mml:mo>
<mml:mn>1</mml:mn>
</mml:mrow>
<mml:mrow>
<mml:mn>50</mml:mn>
</mml:mrow>
</mml:munderover>
<mml:msub>
<mml:mrow>
<mml:mi>s</mml:mi>
</mml:mrow>
<mml:mrow>
<mml:msub>
<mml:mrow>
<mml:mi>F</mml:mi>
</mml:mrow>
<mml:mrow>
<mml:mi>i</mml:mi>
<mml:mi>j</mml:mi>
</mml:mrow>
</mml:msub>
</mml:mrow>
</mml:msub>
<mml:mo>&#x2b;</mml:mo>
<mml:munderover accentunder="false" accent="false">
<mml:mrow>
<mml:mo>&#x2211;</mml:mo>
</mml:mrow>
<mml:mrow>
<mml:mi>i</mml:mi>
<mml:mo>&#x3d;</mml:mo>
<mml:mn>51</mml:mn>
</mml:mrow>
<mml:mrow>
<mml:mn>1000</mml:mn>
</mml:mrow>
</mml:munderover>
<mml:msub>
<mml:mrow>
<mml:mi>s</mml:mi>
</mml:mrow>
<mml:mrow>
<mml:msub>
<mml:mrow>
<mml:mi>B</mml:mi>
</mml:mrow>
<mml:mrow>
<mml:mi>i</mml:mi>
<mml:mi>j</mml:mi>
</mml:mrow>
</mml:msub>
</mml:mrow>
</mml:msub>
<mml:mspace width="1em"/>
</mml:mtd>
<mml:mtd columnalign="left">
<mml:mtext>if&#x2009;</mml:mtext>
<mml:mi>j</mml:mi>
<mml:mspace width="0.3333em"/>
<mml:mtext>mod&#x2009;</mml:mtext>
<mml:mn>2</mml:mn>
<mml:mtext>&#x2009;</mml:mtext>
<mml:mo>&#x3d;</mml:mo>
<mml:mn>1</mml:mn>
</mml:mtd>
</mml:mtr>
<mml:mtr>
<mml:mtd columnalign="left">
<mml:munderover accentunder="false" accent="false">
<mml:mrow>
<mml:mo>&#x2211;</mml:mo>
</mml:mrow>
<mml:mrow>
<mml:mi>i</mml:mi>
<mml:mo>&#x3d;</mml:mo>
<mml:mn>1</mml:mn>
</mml:mrow>
<mml:mrow>
<mml:mn>25</mml:mn>
</mml:mrow>
</mml:munderover>
<mml:msub>
<mml:mrow>
<mml:mi>s</mml:mi>
</mml:mrow>
<mml:mrow>
<mml:msub>
<mml:mrow>
<mml:mi>F</mml:mi>
</mml:mrow>
<mml:mrow>
<mml:mi>i</mml:mi>
<mml:mi>j</mml:mi>
</mml:mrow>
</mml:msub>
</mml:mrow>
</mml:msub>
<mml:mo>&#x2b;</mml:mo>
<mml:munderover accentunder="false" accent="false">
<mml:mrow>
<mml:mo>&#x2211;</mml:mo>
</mml:mrow>
<mml:mrow>
<mml:mi>i</mml:mi>
<mml:mo>&#x3d;</mml:mo>
<mml:mn>26</mml:mn>
</mml:mrow>
<mml:mrow>
<mml:mn>1000</mml:mn>
</mml:mrow>
</mml:munderover>
<mml:msub>
<mml:mrow>
<mml:mi>s</mml:mi>
</mml:mrow>
<mml:mrow>
<mml:msub>
<mml:mrow>
<mml:mi>B</mml:mi>
</mml:mrow>
<mml:mrow>
<mml:mi>i</mml:mi>
<mml:mi>j</mml:mi>
</mml:mrow>
</mml:msub>
</mml:mrow>
</mml:msub>
<mml:mspace width="1em"/>
</mml:mtd>
<mml:mtd columnalign="left">
<mml:mtext>if&#x2009;</mml:mtext>
<mml:mi>j</mml:mi>
<mml:mspace width="0.3333em"/>
<mml:mtext>mod&#x2009;</mml:mtext>
<mml:mn>2</mml:mn>
<mml:mtext>&#x2009;</mml:mtext>
<mml:mo>&#x3d;</mml:mo>
<mml:mn>0</mml:mn>
</mml:mtd>
</mml:mtr>
</mml:mtable>
</mml:mrow>
</mml:mfenced>
</mml:math>
</disp-formula>
</p>
<p>The dataset <italic>Z</italic>
<sub>
<italic>ij</italic>
</sub> formed in this manner consists of two types of subregions (<italic>foregrounds</italic>) which define the two classes used for k-medoids clustering. An illustration of the data on such a spherical surface with 5&#x20;<italic>backgrounds</italic> and 2&#x20;<italic>foregrounds</italic> is shown in <xref ref-type="fig" rid="F3">Figure&#x20;3</xref>.</p>
<fig id="F3" position="float">
<label>FIGURE 3</label>
<caption>
<p>Illustration of data consisting of 5 backgrounds and 2 foregrounds on a spherical surface, similar colors indicate association of foregrounds with respective backgrounds.</p>
</caption>
<graphic xlink:href="fams-07-787455-g003.tif"/>
</fig>
<p>We perform 50 random splits of the dataset <italic>Z</italic>
<sub>
<italic>ij</italic>
</sub> where each split consists of a training set of size 10 and a test set of size 11. The results for clustering are shown for the 4 distance measures in <xref ref-type="table" rid="T3">Table&#x20;3</xref>.</p>
<table-wrap id="T3" position="float">
<label>TABLE 3</label>
<caption>
<p>Clustering Performance for 3d Multivariate Time Series dataset.</p>
</caption>
<table>
<thead valign="top">
<tr>
<th align="left">Distance measure</th>
<th align="center">Total number of errors</th>
<th align="center">Percent error</th>
</tr>
</thead>
<tbody valign="top">
<tr>
<td align="left">kdiff</td>
<td align="char" char=".">0</td>
<td align="char" char=".">0</td>
</tr>
<tr>
<td align="left">MMD</td>
<td align="char" char=".">225</td>
<td align="char" char=".">40.9</td>
</tr>
<tr>
<td align="left">MPdist</td>
<td align="char" char=".">141</td>
<td align="char" char=".">25.6</td>
</tr>
<tr>
<td align="left">Dtwd</td>
<td align="char" char=".">227</td>
<td align="char" char=".">41.3</td>
</tr>
</tbody>
</table>
</table-wrap>
<p>From the results it can be seen that kdiff produces the best clustering performance with 0 errors for this dataset. This is attributed to the fact that the subregions of interest are well defined for both classes and using suitable values of parameters determined from training it is possible to accurately cluster all the time series data into two separate groups. On the other hand the performance of MMD is inferior to kdiff because the backgrounds are well separated with different mean values for time series within and across the two classes. This results in time series even belonging to the same class to be placed in separate clusters when MMD is used as a distance measure. Similarly dtwd suffers from poor performance as this distance measure tends to put time series with smaller separation between the mean background values in the same cluster. However these may have distinct values for the foregrounds i.e. they can in general belong to different classes and as a result this causes errors during clustering. For this dataset the performance of MPdist is inferior to kdiff even though the former can find matching sub-regions with zero errors in the case of univariate time series. This difference is attributed to the nature of the spherical region over which the sub-region matching is done where the 1-nearest neighbor strategy employed by MPdist using Euclidean metrics to construct the distance distribution. In case of spherical surfaces it is necessary to use appropriate geodesic distances for nearest neighbor search as discussed in ([<xref ref-type="bibr" rid="B18">18</xref>]). This issue is resolved in kdiff which can find the matching subregion over a non Euclidean region which in this case is a spherical surface thereby giving the most accurate clustering results for this dataset.</p>
</sec>
<sec id="s4-3">
<title>4.3 Real Life Example: MNIST-M Dataset</title>
<p>The MNIST-M dataset used in [<xref ref-type="bibr" rid="B19">19</xref>, <xref ref-type="bibr" rid="B20">20</xref>] was selected as a real-life example to demonstrate the differences in clustering performance using the four distance measures kdiff, MMD, MPdist and dtwd. The MNIST-M dataset consists of MNIST digits [<xref ref-type="bibr" rid="B21">21</xref>] which are difference blended over patches selected from the BSDS500 database of color photos [<xref ref-type="bibr" rid="B22">22</xref>]. In our experiments where we consider k-medoid clustering over <italic>k</italic>&#x20;&#x3d; 2 classes we select 10 instances each of the MNIST digits 0 and 1 to be blended with a selection of background images to form our dataset MNIST-M-1. Since BSDS500 is a dataset of color images the components of this dataset are random fields whose dimensions are 28&#x20;&#xd7; 28&#x20;&#xd7; 3. We form our final dataset for clustering consisting of random fields with dimensions 28&#x20;&#xd7; 28 by averaging over all three channels. Examples of individual zero and one digits on different backgrounds for all three channels of MNIST-M-1 are shown in <xref ref-type="fig" rid="F4">Figures 4</xref>&#x2013;<xref ref-type="fig" rid="F7">7</xref>.</p>
<fig id="F4" position="float">
<label>FIGURE 4</label>
<caption>
<p>Example 1 of MNIST-M-1 digit&#x20;zero.</p>
</caption>
<graphic xlink:href="fams-07-787455-g004.tif"/>
</fig>
<fig id="F5" position="float">
<label>FIGURE 5</label>
<caption>
<p>Example 2 of MNIST-M-1 digit&#x20;zero.</p>
</caption>
<graphic xlink:href="fams-07-787455-g005.tif"/>
</fig>
<fig id="F6" position="float">
<label>FIGURE 6</label>
<caption>
<p>Example 1 of MNIST-M-1 digit&#x20;one.</p>
</caption>
<graphic xlink:href="fams-07-787455-g006.tif"/>
</fig>
<fig id="F7" position="float">
<label>FIGURE 7</label>
<caption>
<p>Example 2 of MNIST-M-1 digit&#x20;one.</p>
</caption>
<graphic xlink:href="fams-07-787455-g007.tif"/>
</fig>
<p>We perform 50 random splits of the dataset where each split consists of a training set of size 10 and a test set of size 10. The results for clustering are shown for the distance measures in <xref ref-type="table" rid="T4">Table&#x20;4</xref>.</p>
<table-wrap id="T4" position="float">
<label>TABLE 4</label>
<caption>
<p>Clustering performance for MNIST-M-1.</p>
</caption>
<table>
<thead valign="top">
<tr>
<th align="left">Distance measure</th>
<th align="center">Total number of errors</th>
<th align="center">Percent error</th>
</tr>
</thead>
<tbody valign="top">
<tr>
<td align="left">kdiff</td>
<td align="char" char=".">91</td>
<td align="char" char=".">18.2</td>
</tr>
<tr>
<td align="left">MMD</td>
<td align="char" char=".">131</td>
<td align="char" char=".">26.2</td>
</tr>
<tr>
<td align="left">MPdist</td>
<td align="char" char=".">68</td>
<td align="char" char=".">13.6</td>
</tr>
<tr>
<td align="left">Dtwd</td>
<td align="char" char=".">149</td>
<td align="char" char=".">29.8</td>
</tr>
</tbody>
</table>
</table-wrap>
<p>From the results it can be seen that for MNIST-M-1 MPdist somewhat outperforms our proposed distance measure kdiff however the latter is superior to both MMD and dtwd. Since in general the background statistics of the MNIST-M images are different, two images belonging to the same class can be placed in separate clusters when MMD is used as a distance measure and this causes MMD to underperform versus kdiff. Similarly dtwd suffers from poor performance as this distance measure tends to put images with smaller separation between the mean background values in the same cluster. However these may have distinct values for the foregrounds i.e. they can in general belong to different classes and as a result this causes errors during clustering.</p>
<p>
<italic>Noise</italic> <italic>robustness</italic> Following the discussion in <xref ref-type="sec" rid="s4-1">Section 4.1</xref> we explore the performance of the distance measures by considering a selection of noisy backgrounds from the BSDS500 database over which the same 10 instances of the MNIST digits 0 and 1 are blended to form a second version of our dataset called MNIST-M-2. Similar to the earlier case we form our final dataset for clustering consisting of random fields with dimensions 28&#x20;&#xd7; 28 by averaging over all three channels of the color image. Examples of individual zero and one digits on different backgrounds for a single channel are shown in <xref ref-type="fig" rid="F8">Figures 8</xref>&#x2013;<xref ref-type="fig" rid="F11">11</xref>. Note that these correspond to the same MNIST digits shown in <xref ref-type="fig" rid="F4">Figures 4</xref>&#x2013;<xref ref-type="fig" rid="F7">7</xref> however are blended with different backgrounds which have been chosen such that the distinguishability of the two classes is reduced.</p>
<fig id="F8" position="float">
<label>FIGURE 8</label>
<caption>
<p>Example 1 of MNIST-M-2 digit&#x20;zero.</p>
</caption>
<graphic xlink:href="fams-07-787455-g008.tif"/>
</fig>
<fig id="F9" position="float">
<label>FIGURE 9</label>
<caption>
<p>Example 2 of MNIST-M-2 digit&#x20;zero.</p>
</caption>
<graphic xlink:href="fams-07-787455-g009.tif"/>
</fig>
<fig id="F10" position="float">
<label>FIGURE 10</label>
<caption>
<p>Example 1 of MNIST-M-2 digit&#x20;one.</p>
</caption>
<graphic xlink:href="fams-07-787455-g010.tif"/>
</fig>
<fig id="F11" position="float">
<label>FIGURE 11</label>
<caption>
<p>Example 2 of MNIST-M-2 digit&#x20;one.</p>
</caption>
<graphic xlink:href="fams-07-787455-g011.tif"/>
</fig>
<p>We use the Kolmogorov-Smirnov (KS) test statistic to characterize the differences between the backgrounds (BSDS500 images) and the foregrounds (MNIST digits 0 and 1) as below:<list list-type="simple">
<list-item>
<p>&#x2022; The mean KS statistic between the distribution of the pixels where a digit 0 is present and the distribution of the pixel values which make up the background (KS-bg-fg-0)</p>
</list-item>
<list-item>
<p>&#x2022; The mean KS statistic between the distribution of the pixels where a digit 1 is present and the distribution of the pixel values which make up the background (KS-bg-fg-1)</p>
</list-item>
<list-item>
<p>&#x2022; The mean KS statistic between pairs of distributions which make up the corresponding backgrounds (KS-bg)</p>
</list-item>
</list>
</p>
<p>The KS values shown in <xref ref-type="table" rid="T5">Table&#x20;5</xref> confirm our visual intuition that the distinguishability of the foreground (MNIST 0 and 1 digits) and the background is less for MNIST-M-2 as compared to MNIST-M-1. Additionally it can be seen that for the noisier dataset MNIST-M-2 the separation between the background distribution of pixels is less than that of MNIST-M-1.</p>
<table-wrap id="T5" position="float">
<label>TABLE 5</label>
<caption>
<p>KS statistic for MNIST-M foregrounds and backgrounds.</p>
</caption>
<table>
<thead valign="top">
<tr>
<th align="left">Dataset</th>
<th align="center">KS-bg-fg-0</th>
<th align="center">KS-bg-fg-1</th>
<th align="center">KS-bg</th>
</tr>
</thead>
<tbody valign="top">
<tr>
<td align="left">MNIST-M-1</td>
<td align="char" char=".">0.998</td>
<td align="char" char=".">0.991</td>
<td align="char" char=".">0.358</td>
</tr>
<tr>
<td align="left">MNIST-M-2</td>
<td align="char" char=".">0.889</td>
<td align="char" char=".">0.754</td>
<td align="char" char=".">0.202</td>
</tr>
</tbody>
</table>
</table-wrap>
<p>We perform 50 random splits of the dataset <italic>Z</italic> where each split consists of a training set of size 10 and a test set of size 10. The results for clustering are shown for the distance measures in <xref ref-type="table" rid="T6">Table&#x20;6</xref>.</p>
<table-wrap id="T6" position="float">
<label>TABLE 6</label>
<caption>
<p>Clustering performance for MNIST-M</p>
</caption>
<table>
<thead valign="top">
<tr>
<th align="left">Distance measure</th>
<th align="center">Total number of errors</th>
<th align="center">Percent error</th>
</tr>
</thead>
<tbody valign="top">
<tr>
<td align="left">kdiff</td>
<td align="char" char=".">183</td>
<td align="char" char=".">36.6</td>
</tr>
<tr>
<td align="left">MMD</td>
<td align="char" char=".">186</td>
<td align="char" char=".">37.2</td>
</tr>
<tr>
<td align="left">MPdist</td>
<td align="char" char=".">197</td>
<td align="char" char=".">39.4</td>
</tr>
<tr>
<td align="left">Dtwd</td>
<td align="char" char=".">197</td>
<td align="char" char=".">39.4</td>
</tr>
</tbody>
</table>
</table-wrap>
<p>From the results it can be seen that for this noisy dataset the clustering accuracy results for all four distance measures are lower as expected, however kdiff slightly outperforms MPdist. As discussed in <xref ref-type="sec" rid="s4-1">Section 4.1</xref> this can be attributed to the fact that in such cases with a lower signal to noise ratio between the foreground and the background kdiff which is estimated using kernel based self and cross similarities over the embeddings can outperform MPdist which is computed using only Euclidean metric based cross-similarities over the embeddings. The expected noise characterizaion is confirmed by our KS statistic values of KS-bg-fg-0 and KS-bg-fg-1 in <xref ref-type="table" rid="T6">Table&#x20;6</xref>. Moreover the lower values of the KS statistic value KS-bg for MNIST-M-2 compared to MNIST-M-1 manifest in similar clustering performances of MMD and kdiff for MNIST-M-2 in contrast with the trends for MNIST-M-1.</p>
<p>Additional comments: For kdiff we used <italic>L</italic>&#x20;&#x3d; <italic>SL</italic> &#x2217; <italic>SL</italic> windows for capturing the image sub-regions leading to (<italic>n</italic>&#x20;&#x2212; <italic>SL</italic> &#x2b; 1)<sup>2</sup> embeddings which were subsequently &#x201c;flattened&#x201d; to form subsequences of size <italic>L</italic>&#x20;&#x3d; <italic>SL</italic>
<sup>2</sup> over which kdiff was estimated using a one dimensional Gaussian kernel. This process can be augmented by estimating kdiff with two dimensional anisotropic Gaussian kernels to improve performance. However this augmented method of kdiff estimation using a higher dimensional kernel with more parameters will significantly increase the computation time and implementation complexity. Note that in the case of MPdist flattening the subregion is not as much of an issue since it does not use kernel based estimations which need accurate bandwidths.</p>
</sec>
</sec>
<sec id="s5">
<title>5 Conclusions and Future Work</title>
<p>In this work we have proposed a kernel-based measure kdiff for estimating distances between time series, random fields and similar univariate or multivariate and possibly non-iid data. Such a distance measure can be used for clustering and classification in applications where data belonging to a given class match only partially over their region of support. In such cases kdiff is shown to outperform both Maximum Mean Discrepancy and Dynamic Time Warping based distance measures for both synthetic and real-life datasets. We also compare the performance of kdiff which is constructed using kernel-based embeddings over the given data versus MPdist which uses Euclidean distance based embeddings. In this case we empirically demonstrate that for data with high signal-to-noise ratio between the matching region and the background both kdiff and MPdist perform equally well for synthetic datasets and MPdist somewhat outperforms kdiff for real life MNIST-M data. For data where the foreground is less distinguishable versus the background kdiff outperforms MPdist for both synthetic and real-life datasets. Additionally for multivariate time series on a spherical manifold we show that kdiff outperforms MPdist because of its kernel-based construction which leads to superior performance in non Euclidean spaces. Our future work will focus on application of kdiff for applications on spherical manifolds such as text embedding [<xref ref-type="bibr" rid="B23">23</xref>] and hyperspectral imagery [<xref ref-type="bibr" rid="B18">18</xref>, <xref ref-type="bibr" rid="B24">24</xref>] as well as clustering and classification applications for time series and random fields with noisy motifs and foregrounds.</p>
</sec>
</body>
<back>
<sec id="s6">
<title>Data Availability Statement</title>
<p>Publicly available datasets were analyzed in this study. This data can be found here: Ganin et al., 2016 [<xref ref-type="bibr" rid="B19">19</xref>].</p>
</sec>
<sec id="s7">
<title>Author Contributions</title>
<p>SD and AC contributed to conception and design. All authors contributed to the theory. SD performed the computation and analysis, and wrote the first draft of the manuscript. All authors contributed to the manuscript revision.</p>
</sec>
<sec id="s8">
<title>Funding</title>
<p>This work was supported in part by the NSF awards CNS1730158, ACI-1540112, ACI-1541349, OAC-1826967, the University of California Office of the President, and the California Institute for Telecommunications and Information Technology&#x2019;s Qualcomm Institute (Calit2-QI). SD was partially supported by an Intel Sponsored Research grant. AC was supported by NSF awards 1819222, 2012266, Russel Sage Founation 2196, and an Intel Sponsored Research grant. HM is supported in part by NSF grant DMS 2012355 and ARO Grant W911NF2110218.</p>
</sec>
<sec sec-type="COI-statement" id="s9">
<title>Conflict of Interest</title>
<p>The authors declare that the research was conducted in the absence of any commercial or financial relationships that could be construed as a potential conflict of interest.</p>
</sec>
<sec sec-type="disclaimer" id="s10">
<title>Publisher&#x2019;s Note</title>
<p>All claims expressed in this article are solely those of the authors and do not necessarily represent those of their affiliated organizations, or those of the publisher, the editors and the reviewers. Any product that may be evaluated in this article, or claim that may be made by its manufacturer, is not guaranteed or endorsed by the publisher.</p>
</sec>
<ack>
<p>Thanks to CENIC for the 100Gpbs networks. The authors thank Siqiao Ruan for helpful discussions.</p>
</ack>
<ref-list>
<title>References</title>
<ref id="B1">
<label>1.</label>
<citation citation-type="book">
<person-group person-group-type="author">
<name>
<surname>Ratanamahatana</surname>
<given-names>CA</given-names>
</name>
<name>
<surname>Keogh</surname>
<given-names>E</given-names>
</name>
</person-group>. <article-title>Everything You Know about Dynamic Time Warping Is Wrong</article-title>. In: <conf-name>Third Workshop on Mining Temporal and Sequential Data, Vol. 32</conf-name>, <conf-loc>Seattle</conf-loc>. <publisher-name>Citeseer</publisher-name> (<year>2004</year>). </citation>
</ref>
<ref id="B2">
<label>2.</label>
<citation citation-type="journal">
<person-group person-group-type="author">
<name>
<surname>Keogh</surname>
<given-names>E</given-names>
</name>
<name>
<surname>Ratanamahatana</surname>
<given-names>CA</given-names>
</name>
</person-group>. <article-title>Exact Indexing of Dynamic Time Warping</article-title>. <source>Knowl Inf Syst</source> (<year>2005</year>) <volume>7</volume>(<issue>3</issue>):<fpage>358</fpage>&#x2013;<lpage>86</lpage>. <pub-id pub-id-type="doi">10.1007/s10115-004-0154-9</pub-id> </citation>
</ref>
<ref id="B3">
<label>3.</label>
<citation citation-type="journal">
<person-group person-group-type="author">
<name>
<surname>D&#x2019;Urso</surname>
<given-names>P</given-names>
</name>
<name>
<surname>Maharaj</surname>
<given-names>EA</given-names>
</name>
</person-group>. <article-title>Autocorrelation-based Fuzzy Clustering of Time Series</article-title>. <source>Fuzzy Sets Syst</source> (<year>2009</year>) <volume>160</volume>(<issue>24</issue>):<fpage>3565</fpage>&#x2013;<lpage>89</lpage>. </citation>
</ref>
<ref id="B4">
<label>4.</label>
<citation citation-type="journal">
<person-group person-group-type="author">
<name>
<surname>Golay</surname>
<given-names>X</given-names>
</name>
<name>
<surname>Kollias</surname>
<given-names>S</given-names>
</name>
<name>
<surname>Stoll</surname>
<given-names>G</given-names>
</name>
<name>
<surname>Meier</surname>
<given-names>D</given-names>
</name>
<name>
<surname>Valavanis</surname>
<given-names>A</given-names>
</name>
<name>
<surname>Boesiger</surname>
<given-names>P</given-names>
</name>
</person-group>. <article-title>A New Correlation-Based Fuzzy Logic Clustering Algorithm for Fmri</article-title>. <source>Magn Reson Med</source> (<year>1998</year>) <volume>40</volume>(<issue>2</issue>):<fpage>249</fpage>&#x2013;<lpage>60</lpage>. <pub-id pub-id-type="doi">10.1002/mrm.1910400211</pub-id> </citation>
</ref>
<ref id="B5">
<label>5.</label>
<citation citation-type="journal">
<person-group person-group-type="author">
<name>
<surname>Alonso</surname>
<given-names>AM</given-names>
</name>
<name>
<surname>Casado</surname>
<given-names>D</given-names>
</name>
<name>
<surname>L&#xf3;pez-Pintado</surname>
<given-names>S</given-names>
</name>
<name>
<surname>Romo</surname>
<given-names>J</given-names>
</name>
</person-group>. <article-title>Robust Functional Supervised Classification for Time Series</article-title>. <source>J&#x20;Classif</source> (<year>2014</year>) <volume>31</volume>(<issue>3</issue>):<fpage>325</fpage>&#x2013;<lpage>50</lpage>. <pub-id pub-id-type="doi">10.1007/s00357-014-9163-x</pub-id> </citation>
</ref>
<ref id="B6">
<label>6.</label>
<citation citation-type="journal">
<person-group person-group-type="author">
<name>
<surname>D&#x2019;Urso</surname>
<given-names>P</given-names>
</name>
<name>
<surname>Maharaj</surname>
<given-names>EA</given-names>
</name>
<name>
<surname>Alonso</surname>
<given-names>AM</given-names>
</name>
</person-group>. <article-title>Fuzzy Clustering of Time Series Using Extremes</article-title>. <source>Fuzzy Sets Syst</source> (<year>2017</year>) <volume>318</volume>:<fpage>56</fpage>&#x2013;<lpage>79</lpage>. </citation>
</ref>
<ref id="B7">
<label>7.</label>
<citation citation-type="book">
<person-group person-group-type="author">
<name>
<surname>Gharghabi</surname>
<given-names>S</given-names>
</name>
<name>
<surname>Imani</surname>
<given-names>S</given-names>
</name>
<name>
<surname>Bagnall</surname>
<given-names>A</given-names>
</name>
<name>
<surname>Darvishzadeh</surname>
<given-names>A</given-names>
</name>
<name>
<surname>Keogh</surname>
<given-names>E</given-names>
</name>
</person-group>. <article-title>Matrix Profile Xii: Mpdist: A Novel Time Series Distance Measure to Allow Data Mining in More Challenging Scenarios</article-title>. In: <conf-name>2018 IEEE International Conference on Data Mining (ICDM)</conf-name>. <publisher-name>IEEE</publisher-name> (<year>2018</year>). p. <fpage>965</fpage>&#x2013;<lpage>70</lpage>. <pub-id pub-id-type="doi">10.1109/icdm.2018.00119</pub-id> </citation>
</ref>
<ref id="B8">
<label>8.</label>
<citation citation-type="book">
<person-group person-group-type="author">
<name>
<surname>Brandmaier</surname>
<given-names>AM</given-names>
</name>
</person-group>. <source>Permutation Distribution Clustering and Structural Equation Model Trees</source>. <publisher-name>CRAN</publisher-name> (<year>2011</year>). </citation>
</ref>
<ref id="B9">
<label>9.</label>
<citation citation-type="journal">
<person-group person-group-type="author">
<name>
<surname>Gretton</surname>
<given-names>A</given-names>
</name>
<name>
<surname>Borgwardt</surname>
<given-names>KM</given-names>
</name>
<name>
<surname>Rasch</surname>
<given-names>MJ</given-names>
</name>
<name>
<surname>Sch&#xf6;lkopf</surname>
<given-names>B</given-names>
</name>
<name>
<surname>Smola</surname>
<given-names>A</given-names>
</name>
</person-group>. <article-title>A Kernel Two-Sample Test</article-title>. <source>J&#x20;Machine Learn Res</source> (<year>2012</year>) <volume>13</volume>:<fpage>723</fpage>&#x2013;<lpage>73</lpage>. </citation>
</ref>
<ref id="B10">
<label>10.</label>
<citation citation-type="book">
<person-group person-group-type="author">
<name>
<surname>Gretton</surname>
<given-names>A</given-names>
</name>
<name>
<surname>Sejdinovic</surname>
<given-names>D</given-names>
</name>
<name>
<surname>Strathmann</surname>
<given-names>H</given-names>
</name>
<name>
<surname>Balakrishnan</surname>
<given-names>S</given-names>
</name>
<name>
<surname>Pontil</surname>
<given-names>M</given-names>
</name>
<name>
<surname>Fukumizu</surname>
<given-names>K</given-names>
</name>
<etal/>
</person-group> <article-title>Optimal Kernel Choice for Large-Scale Two-Sample Tests</article-title>. In: <source>Advances in Neural Information Processing Systems</source>. <publisher-name>Citeseer</publisher-name> (<year>2012</year>). p. <fpage>1205</fpage>&#x2013;<lpage>13</lpage>. </citation>
</ref>
<ref id="B11">
<label>11.</label>
<citation citation-type="book">
<person-group person-group-type="author">
<name>
<surname>Szabo</surname>
<given-names>Z</given-names>
</name>
<name>
<surname>Chwialkowski</surname>
<given-names>K</given-names>
</name>
<name>
<surname>Gretton</surname>
<given-names>A</given-names>
</name>
<name>
<surname>Jitkrittum</surname>
<given-names>W</given-names>
</name>
</person-group>. <article-title>Interpretable Distribution Features with Maximum Testing Power</article-title>. In: <conf-name>Proceedings of the 30th International Conference on Neural Information Processing Systems</conf-name>. <publisher-name>IEEE</publisher-name> (<year>2016</year>). </citation>
</ref>
<ref id="B12">
<label>12.</label>
<citation citation-type="journal">
<person-group person-group-type="author">
<name>
<surname>Cheng</surname>
<given-names>X</given-names>
</name>
<name>
<surname>Cloninger</surname>
<given-names>A</given-names>
</name>
<name>
<surname>Coifman</surname>
<given-names>RR</given-names>
</name>
</person-group>. <article-title>Two-sample Statistics Based on Anisotropic Kernels</article-title>. <source>Inf Inference: A J&#x20;IMA</source> (<year>2020</year>) <volume>9</volume>(<issue>3</issue>):<fpage>677</fpage>&#x2013;<lpage>719</lpage>. <pub-id pub-id-type="doi">10.1093/imaiai/iaz018</pub-id> </citation>
</ref>
<ref id="B13">
<label>13.</label>
<citation citation-type="journal">
<person-group person-group-type="author">
<name>
<surname>Bandt</surname>
<given-names>C</given-names>
</name>
<name>
<surname>Pompe</surname>
<given-names>B</given-names>
</name>
</person-group>. <article-title>Permutation Entropy: a Natural Complexity Measure for Time Series</article-title>. <source>Phys Rev Lett</source> (<year>2002</year>) <volume>88</volume>(<issue>17</issue>):<fpage>174102</fpage>. <pub-id pub-id-type="doi">10.1103/physrevlett.88.174102</pub-id> </citation>
</ref>
<ref id="B14">
<label>14.</label>
<citation citation-type="journal">
<person-group person-group-type="author">
<name>
<surname>Mhaskar</surname>
<given-names>HN</given-names>
</name>
<name>
<surname>Cheng</surname>
<given-names>X</given-names>
</name>
<name>
<surname>Cloninger</surname>
<given-names>A</given-names>
</name>
</person-group>. <article-title>A Witness Function Based Construction of Discriminative Models Using Hermite Polynomials</article-title>. <source>Front Appl Math Stat</source> (<year>2020</year>) <volume>6</volume>:<fpage>31</fpage>. <pub-id pub-id-type="doi">10.3389/fams.2020.00031</pub-id> </citation>
</ref>
<ref id="B15">
<label>15.</label>
<citation citation-type="book">
<person-group person-group-type="author">
<name>
<surname>Cloninger</surname>
<given-names>A</given-names>
</name>
</person-group>. <source>Bounding the Error from Reference Set Kernel Maximum Mean Discrepancy</source>. <publisher-name>arXiv preprint arXiv:1812.04594</publisher-name> (<year>2018</year>). </citation>
</ref>
<ref id="B16">
<label>16.</label>
<citation citation-type="book">
<person-group person-group-type="author">
<name>
<surname>Kaufmann</surname>
<given-names>L</given-names>
</name>
<name>
<surname>Rousseeuw</surname>
<given-names>P</given-names>
</name>
</person-group>. <article-title>Clustering by Means of Medoids</article-title>. In: <conf-name>Proc. Statistical Data Analysis Based on the L1 Norm Conference</conf-name>. <publisher-name>Neuchatel</publisher-name> (<year>19871987</year>). p. <fpage>405</fpage>&#x2013;<lpage>16</lpage>. </citation>
</ref>
<ref id="B17">
<label>17.</label>
<citation citation-type="journal">
<person-group person-group-type="author">
<name>
<surname>Sard&#xe1;-Espinosa</surname>
<given-names>A</given-names>
</name>
</person-group>. <article-title>Comparing Time-Series Clustering Algorithms in R Using the Dtwclust Package</article-title>. <source>R Package Vignette</source> (<year>2017</year>) <volume>12</volume>:<fpage>41</fpage>. </citation>
</ref>
<ref id="B18">
<label>18.</label>
<citation citation-type="book">
<person-group person-group-type="author">
<name>
<surname>Lunga</surname>
<given-names>D</given-names>
</name>
<name>
<surname>Ersoy</surname>
<given-names>O</given-names>
</name>
</person-group>. <article-title>Spherical Nearest Neighbor Classification: Application to Hyperspectral Data</article-title>. In: <conf-name>International Workshop on Machine Learning and Data Mining in Pattern Recognition</conf-name>. <publisher-name>Springer</publisher-name> (<year>2011</year>). p. <fpage>170</fpage>&#x2013;<lpage>84</lpage>. <pub-id pub-id-type="doi">10.1007/978-3-642-23199-5_13</pub-id> </citation>
</ref>
<ref id="B19">
<label>19.</label>
<citation citation-type="journal">
<person-group person-group-type="author">
<name>
<surname>Ganin</surname>
<given-names>Y</given-names>
</name>
<name>
<surname>Ustinova</surname>
<given-names>E</given-names>
</name>
<name>
<surname>Ajakan</surname>
<given-names>H</given-names>
</name>
<name>
<surname>Germain</surname>
<given-names>P</given-names>
</name>
<name>
<surname>Larochelle</surname>
<given-names>H</given-names>
</name>
<name>
<surname>Laviolette</surname>
<given-names>F</given-names>
</name>
<etal/>
</person-group> <article-title>Domain-adversarial training of neural networks</article-title>, <source>J Machine Learn Res</source> <volume>17</volume>(<issue>1</issue>), pp. <fpage>2096</fpage> (2030, 2016). </citation>
</ref>
<ref id="B20">
<label>20.</label>
<citation citation-type="book">
<person-group person-group-type="author">
<name>
<surname>Haeusser</surname>
<given-names>P</given-names>
</name>
<name>
<surname>Frerix</surname>
<given-names>T</given-names>
</name>
<name>
<surname>Mordvintsev</surname>
<given-names>A</given-names>
</name>
<name>
<surname>Cremers</surname>
<given-names>D</given-names>
</name>
</person-group>. <article-title>Associative Domain Adaptation</article-title>. In: <conf-name>Proceedings of the IEEE International Conference on&#x20;Computer Vision</conf-name>. <publisher-name>IEEE</publisher-name> (<year>2017</year>). p. <fpage>2765</fpage>&#x2013;<lpage>73</lpage>. <pub-id pub-id-type="doi">10.1109/iccv.2017.301</pub-id> </citation>
</ref>
<ref id="B21">
<label>21.</label>
<citation citation-type="journal">
<person-group person-group-type="author">
<name>
<surname>LeCun</surname>
<given-names>Y</given-names>
</name>
<name>
<surname>Bottou</surname>
<given-names>L</given-names>
</name>
<name>
<surname>Bengio</surname>
<given-names>Y</given-names>
</name>
<name>
<surname>Haffner</surname>
<given-names>P</given-names>
</name>
</person-group>. <article-title>Gradient-based Learning Applied to Document Recognition</article-title>. <source>Proc IEEE</source> (<year>1998</year>) <volume>86</volume>(<issue>11</issue>):<fpage>2278</fpage>&#x2013;<lpage>324</lpage>. <pub-id pub-id-type="doi">10.1109/5.726791</pub-id> </citation>
</ref>
<ref id="B22">
<label>22.</label>
<citation citation-type="journal">
<person-group person-group-type="author">
<name>
<surname>Arbel&#xe1;ez</surname>
<given-names>P</given-names>
</name>
<name>
<surname>Maire</surname>
<given-names>M</given-names>
</name>
<name>
<surname>Fowlkes</surname>
<given-names>C</given-names>
</name>
<name>
<surname>Malik</surname>
<given-names>J</given-names>
</name>
</person-group>. <article-title>Contour Detection and Hierarchical Image Segmentation</article-title>. <source>IEEE Trans Pattern Anal Mach Intell</source> (<year>2010</year>) <volume>33</volume>(<issue>5</issue>):<fpage>898</fpage>&#x2013;<lpage>916</lpage>. <pub-id pub-id-type="doi">10.1109/TPAMI.2010.161</pub-id> </citation>
</ref>
<ref id="B23">
<label>23.</label>
<citation citation-type="book">
<person-group person-group-type="author">
<name>
<surname>Meng</surname>
<given-names>Y</given-names>
</name>
<name>
<surname>Huang</surname>
<given-names>J</given-names>
</name>
<name>
<surname>Wang</surname>
<given-names>G</given-names>
</name>
<name>
<surname>Zhang</surname>
<given-names>C</given-names>
</name>
<name>
<surname>Zhuang</surname>
<given-names>H</given-names>
</name>
<name>
<surname>Kaplan</surname>
<given-names>L</given-names>
</name>
<etal/>
</person-group> <article-title>Spherical Text Embedding</article-title>. In: <source>Advances in Neural Information Processing Systems</source>. <publisher-name>Springer</publisher-name> (<year>2019</year>). p. <fpage>8208</fpage>&#x2013;<lpage>17</lpage>. </citation>
</ref>
<ref id="B24">
<label>24.</label>
<citation citation-type="book">
<person-group person-group-type="author">
<name>
<surname>Lunga</surname>
<given-names>D</given-names>
</name>
<name>
<surname>Ersoy</surname>
<given-names>O</given-names>
</name>
</person-group>. <article-title>Unsupervised Classification of Hyperspectral Images on Spherical Manifolds</article-title>. In: <conf-name>Industrial Conference on Data Mining</conf-name>. <publisher-name>Springer</publisher-name> (<year>2011</year>). p. <fpage>134</fpage>&#x2013;<lpage>46</lpage>. <pub-id pub-id-type="doi">10.1007/978-3-642-23184-1_11</pub-id> </citation>
</ref>
</ref-list>
</back>
</article>