Skip to main content

ORIGINAL RESEARCH article

Front. Appl. Math. Stat., 08 November 2017
Sec. Mathematics of Computation and Data Science
Volume 3 - 2017 | https://doi.org/10.3389/fams.2017.00023

Regularized Kernel Algorithms for Support Estimation

  • 1INRIA—Sierra team—École Normale Supérieure, Paris, France
  • 2Laboratory for Computational and Statistical Learning, Istituto Italiano di Tecnologia, Genova, Italy
  • 3Dipartimento di Matematica, Università di Genova, Genova, Italy
  • 4DIBRIS, Università di Genova, Genova, Italy

In the framework of non-parametric support estimation, we study the statistical properties of a set estimator defined by means of Kernel Principal Component Analysis. Under a suitable assumption on the kernel, we prove that the algorithm is strongly consistent with respect to the Hausdorff distance. We also extend the above analysis to a larger class of set estimators defined in terms of a low-pass filter function. We finally provide numerical simulations on synthetic data to highlight the role of the hyper parameters, which affect the algorithm.

1. Introduction

A classical issue in statistics is support estimation, i.e., the problem of learning the support of a probability distribution from a set of points identically sampled according to the distribution. For example, the Devroye-Wise algorithm [1] estimates the support with the union of suitable balls centered in the training points. In the last two decades, many algorithms have been proposed and their statistical properties analyzed [114] and references therein.

An instance of the above setting, which plays an important role in applications, is the problem of novelty/anomaly detection, see Campos et al. [15] for an updated review. In this context, in Hoffmann [16] the author proposed an estimator based on Kernel Principal Component Analysis (KPCA), first introduced in Schölkopf et al. [17] in the context of dimensionality reduction. The algorithm was successfully tested in many applications from computer vision to biochemistry [1824]. In many of these examples the data are often represented by high dimensional vectors, but they actually live close to a nonlinear low dimensional submanifold of the original space, and the proposed estimator takes advantage of the fact that KPCA provides an efficient compression/dimensionality reduction of the original data [16, 17], whereas many classical set estimators refer to the dimension of the original space, as it happens for the Devroye-Wise algorithm.

In this paper we prove that KPCA is a consistent estimator of the support of the distribution with respect to the Hausdorff distance. The result is based on an intriguing property of the reproducing kernel, called separating condition, first introduced in De Vito et al. [25]. This assumption ensures that any closed subset of the original space is represented in the feature space by a linear subspace. We show that this property remains true if the data are recentered to have zero mean in the feature space. Together with the results in De Vito et al. [25], we conclude that the consistency of KPCA algorithm is preserved by recentering of the data, which can be regarded as a degree of freedom to improve the empirical performance of the algorithm in a specific application.

Our main contribution is sketched in the next subsection together with some basic properties of KPCA and some relevant previous works. In section 2, we describe the mathematical framework and the related notations. Section 3 introduces the spectral support estimator and informally discusses its main features, whereas its statistical properties and the meaning of the separating condition for the kernel are analyzed in section 4. Finally section 5 presents the effective algorithm to compute the decision function and discusses the role of the two meta-parameters based on the previous theoretical analysis. In the Appendix (Supplementary Material), we collect some technical results.

1.1. Sketch of the Main Result and Previous Works

In this section we sketch our main result by first recalling the construction of the KPCA estimator introduced in Hoffmann [16]. We have at disposal a training set {x1,,xn}Dd of n points independently sampled according to some probability distribution P. The input space D is a known compact subset of ℝd, but the probability distribution P is unknown and the goal is to estimate the support C of P from the empirical data. We recall that C is the smallest closed subset of D such that ℙ[C] = 1 and we stress that C is in general a proper subset of D, possibly of low dimension.

Classical Principal Component Analysis (PCA) is based on the construction of the vector space V spanned by the first m eigenvectors associated with the largest eigenvalues of the empirical covariance matrix

1ni=1n(xix¯)(xix¯),

where x¯=1ni=1nxi is the empirical mean. However, if the data do not live on an affine subspace, the set V is not a consistent estimator of the support. In order to take into account non-linear models, following the idea introduced in Schölkopf et al. [17] we consider a feature map Φ from the input space D into the corresponding feature space H, which is assumed to be a Hilbert space, and we replace the empirical covariance matrix with the empirical covariance operator

Tnc^=1ni=1n(Φ(xi)μ^n)(Φ(xi)μ^n),

where μ^n=1ni=1nΦ(xi) is the empirical mean in the feature space. As it happens in PCA, we consider the subspace V^m,n of H spanned by the first m- eigenvectors f^1,,f^m of Tnc^. According to the proposal in Hoffmann [16], we consider the following estimator of the support of the probability distribution P

C^n={xDμ^n+j=1mΦ(x)μ^n,f^jf^jΦ(x)τn},

where τn is a suitable threshold depending on the number of examples and

μ^n+j=1mΦ(x)μ^n,f^jf^j

is the projection of an arbitrary point xD onto the affine subspace μ^n+V^m,n. We show that, under a suitable assumption on the feature map, called separating property, C^n is a consistent estimator of C with respect to the Hausdorff distance between compact sets, see Theorem 3 of section 4.2.

The separating property was introduced in De Vito et al. [25] and it ensures that the feature space is rich enough to learn any closed subset of D. This assumption plays the same role of the notion of universal kernel [26] in supervised learning.

Moreover, following [25, 27] we extend the KPCA estimator to a class of learning algorithms defined in terms of a low-pass filter function rm(σ) acting on the spectrum of the covariance matrix and depending on a regularization parameter m ∈ ℕ. The projection of Φ^n(x) onto V^m,n is replaced by the vector

Φ^n,m(x)=j=1+rm(σ^j)Φ(x)μ^n,f^jf^j,

where {f^j}j is the family of eigenvectors of Tnc^ and {σ^j}j is the corresponding family of eigenvalues. The support is then estimated by the set

{xD||μ^n+Φ^n,m(x)Φ(x)||τn}.

Note that KPCA corresponds to the choice of the hard-cut off filter

rm(σ^j)={1im0i>m.

However, other filter functions can be considered, inspired by the theory of regularization for inverse problems [28] and by supervised learning algorithms [29, 30]. In this paper we show that the explicit computation of these spectral estimators reduces to a finite dimensional problem depending only on the kernel K(x, w) = 〈Φ(x), Φ(w)〉 associated with the feature map, as for KPCA. The computational properties of each learning algorithm depend on the choice of the low-pass filter rm(σ), which can be tuned to out-perform of some specific data set, see the discussion in Rudi et al. [31].

We conclude this section with two considerations. First, in De Vito et al. [25, 27] it is proven a consistency result for a similar estimator, where the subspace V^n,m is computed with respect to the non-centered covariance matrix in the feature space H, instead of the covariance matrix. In this paper we analyze the impact of recentering the data in the feature space H on the support estimation problem, see Theorem 1 below. This point of view is further analyzed in Rudi et al. [32, 33].

Finally note that, our consistency results are based on convergence rates of empirical subspaces to true subspaces of the covariance operator, see Theorem 2 below. The main difference between our result and the result in Blanchard et al. [34], is that we prove the consistency for the case when the dimension m = mn of the subspace V^m,n goes to infinity slowly enough. On the contrary, in their seminal paper [34] the authors analyze the most specific case when the dimension of the projection space is fixed.

2. Mathematical Assumptions

In this section we introduce the statistical model generating the data, the notion of separating feature map and the properties of the filter function. Furthermore, we show that KPCA can be seen as a filter function and we recall the main properties of the covariance operators.

We assume that the input space D is a bounded closed subset of ℝd. However, our results also hold true by replacing D with any compact metric space. We denote by d(x, w) the Euclidean distance |xw| between two points x, w ∈ ℝd and by dH(A, B) the Hausdorff distance between two compact subsets A,BD, explicitly given by

dH(A,B)=max{supxAd(x,B),supxBd(x,A)},

where d(x,A)=infwAd(x,w).

2.1. Statistical Model

The statistical model is described by a random vector X taking value in D. We denote by P the probability distribution of X, defined on the Borel σ-algebra of D, and by C the support of P.

Since the probability distribution P is unknown, so is its support. We aim to estimate C from a training set of empirical data, which are described by a family X1, …, Xn of random vectors, which are independent and identically distributed as X. More precisely, we are looking for a closed subset C^n=C^X1,,XnD, depending only on X1, …, Xn, but independent of P, such that

[limn+dH(C^n,C)=0]=1

for all probability distributions P. In the context of regression estimate, the above convergence is usually called universal strong consistency [35].

2.2. Mercer Feature Maps and Separating Condition

To define the estimator C^n we first map the data into a suitable feature space, so that the support C is represented by a linear subspace.

Assumption 1. Given a Hilbert space H, take Φ:DH satisfying the following properties:

(H1) the set Φ(D) is total in H, i.e.,

span¯{Φ^(x)xD}=H,

where span¯{·} denotes the closure of the linear span;

(H2) the map Φ is continuous.

The space H is called the feature space and the map Φ is called a Mercer feature map.

In the following the norm and scalar product of H are denoted by ||·|| and 〈··〉, respectively.

Assumptions (H1) and (H2) are standard for kernel methods, see Steinwart and Christmann [36]. We now briefly recall some basic consequences. First of all, the map K:D×D

K(x,w)=Φ(x)Φ(w)

is a Mercer kernel and we denote by HK the corresponding (separable) reproducing kernel Hilbert space, whose elements are continuous functions on D. Moreover, each element fH defines a function fΦHK by setting fΦ(x) = 〈f, Φ(x)〉 for all xD. Since Φ(D) is total in H, the linear map ffΦ is an isometry from H onto HK. In the following, with slight abuse of notation, we write f instead of fΦ, so that the elements fH are viewed as functions on D satisfying the reproducing property

f(x)=f,Φ(x)  xD.

Finally, since D is compact and Φ is continuous, it holds that

R=supxD||Φ(x)||2=supxDK(x,x)<+.    (1)

Following De Vito et al. [27], we call Φ a separating Mercer feature map if the following the separating property also holds true.

(H3) The map Φ is injective and for all closed subsets CD

Φ(C)=Φ(D)span¯{Φ(x)xC}.    (2)

It states that any closed subset CD is mapped by Φ onto the intersection of Φ(D) and the closed subspace span¯{Φ(x)xC}H. Examples of kernels satisfying the separating property are for Dd [27]:

• Sobolev kernels with smoothness index s>d2;

• the Abel/Laplacian kernel K(x, w) = e − γ|xw| with γ > 0;

• the ℓ1-kernel K(x,w)=e-γ|x-w|1, where |·|1 is the ℓ1-norm and γ > 0.

As shown in De Vito et al. [25], given a closed set C the equality (2) is equivalent to the condition that for every x0C there exists fH such that

f(x0)0  and  f(x)=0xC.    (3)

Clearly, an arbitrary Mercer feature map is not able to separate all the closed subsets, but only few of them. To better describe these sets, we introduce the elementary learnable sets, namely

Cf={xDf(x)=f,Φ(x)=0}, 

where fH. Clearly, Cf is closed and the equality (3) holds true. Furthermore the intersection of an arbitrary family of elementary learnable sets fFCf with FH satisfies (3), too. Conversely, if C is a set satisfying (2), select a maximal family {fj}jJ of orthonormal functions in H such that

fj(x)=fj,Φ(x)=0  xC, jJ,

i.e., a basis of the orthogonal complement of {Φ(x)xC}, then it is easy to prove that

C={xDfj,Φ(x)=0 jJ}=jJCfj,(4)

so that any set which is separating by Φ is the (possibly denumerable) intersection of elementary sets. Assumption (H3) is hence a requirement that the family of the elementary learnable sets, labeled by the elements of H, is rich enough to parameterize all the closed subsets of D by means of (4). In section 4.3 we present some examples.

The Gaussian kernel K(x, w) = e−γ|xw|2 is a popular choice in machine learning, however it is not separating. Indeed, since K is analytic, the elements of the corresponding reproducing kernel Hilbert space are analytic functions, too [36]. It is known that, given an analytic function f ≠ 0, the corresponding elementary learnable set Cf={xDf(x)=0} is a closed set whose interior is the empty set. Hence also the denumerable intersections have empty interior, so that K can not separate a support with not-empty interior. In Figure 1 we compare the decay behavior of the eigenvalues of the Laplacian and the Gaussian kernels.

FIGURE 1
www.frontiersin.org

Figure 1. Eigenvalues in logarithmic scale of the Covariance operator when the kernel is Abel (blue) and Gaussian (red) and the distribution is uniformly supported on the “8” curve in Figure 2. Note that the eigenvalue decay rate of the first operator has a polynomial behavior while the second has an exponential one.

2.3. Filter Function

The second building block is a low pass filter, we introduce to avoid that the estimator overfits the empirical data. The filter functions were first introduced in the context of inverse problem, see Engl et al. [28] and references therein, and in the context of supervised learning, see Lo Gerfo et al. [29] and Blanchard and Mucke [30].

We now fix some notations. For any fH, we denote by ff the rank one operator (ff)g = 〈gff. We recall that a bounded operator A on H is a Hilbert-Schmidt operator if for some (any) basis {fj}j the series ||A||22:=j||Afj||2 is finite, ||A||2 is called the Hilbert-Schmidt norm and ||A|| ≤ ||A||2, where ||·|| is the spectral norm. We denote by S2 the space of Hilbert-Schmidt operators, which is a separable Hilbert space under the scalar product AB2=jAfjBfj.

Assumption 2. A filter function is a sequence of functions rm:[0, R] → [0, 1], with m ∈ ℕ, satisfying

(H4) for any m ∈ ℕ, rm(0) = 0;

(H5) for all σ > 0, limm+rm(σ)=1;

(H6) for all m ∈ ℕ, there is Lm > 0 such that

|rm(σ)rm(σ)|Lm|σσ|,

i.e., rm is a Lipschitz function with Lipschitz constant Lm.

For fixed m, rm is a filter cutting the smallest eigenvalues (high frequencies). Indeed, (H4) and (H6) with σ′ = 0 give

|rm(σ)|Lm|σ|.    (5)

On the contrary, if m goes to infinity, by (H5) rm converges point-wisely to the Heaviside function

Θ(σ)={1σ>00σ=0.

Since rm(σ) converges to Θ(σ), which does not satisfy (5), we have that limm+Lm=+.

We fix the interval [0, R] as domain of the filter functions rm since the eigenvalues of the operators we are interested belong to [0, R], see (23).

Examples of filter functions are

• Tikhonov filter

rm(σ)=mσmσ+R  Lm=mR.

• Soft cut-off

rm(σ)={1σRmmσRσ<Rm  Lm=mR.

• Landweber iteration

rm(σ)=σRk=0m(1σR)k=1(1σR)m+1Lm=m+1R.

We recall a technical result, which is based on functional calculus for compact operators. If A is a positive Hilbert-Schmidt operator, Hilbert-Schmidt theorem (for compact self-adjoint operators) gives that there exist a basis {fj}j of H and a family {σj}j of positive numbers such that

A=jσjfjfj    Afj=σjfj.    (6)

If the spectral norm ||A||R, then all the eigenvalues σj belong to [0, R] and the spectral calculus defines rm(A) as the operator on H given by

rm(A)=jrm(σj)fjfj    rm(A)fj=rm(σj)fj.

With this definition each fj is still an eigenvector of rm(A), but the corresponding eigenvalue is shrunk to rmj). Proposition 1 in the Appendix in Supplementary Material summarizes the main properties of rm(A).

2.4. Kernel Principal Component Analysis

As anticipated in the introduction, the estimators we propose are a generalization of KPCA suggested by Hoffmann [16] in the context of novelty detection. In our framework this corresponds to the hard cut-off filter, i.e., by labeling the different eigenvalues of A in a decreasing order1 σ1 > σ2 > … > σm > σm+1 > …, the filter function is

rm(σ)={1σσm0σ<σm.

Clearly, rm satisfies (H4) and (H5), but (H6) does not hold. However, the Lipschitz assumption is needed only to prove the bound (21e) and, for the hard cut-off filter, rm(A) is simply the orthogonal projector onto the linear space spanned by the eigenvectors whose eigenvalues are bigger than σm+1. For such projections [37] proves the following bound

||rm(A)rm(A)||22σm+1σm||AA||2,

so that (21e) holds true with Lm=2σm+1-σm. Hence, our results also hold for hard cut-off filter at the price to have a Lipschitz constant Lm depending on the eigenvalues of A.

2.5. Covariance Operators

The third building block is made of the eigenvectors of the distribution dependent covariance operator and of its empirical version. The covariance operators are computed by first mapping the data in the feature space H.

As usual, we introduce two random variables Φ(X) and Φ(X) ⊗ Φ(X), taking value in H and in S2, respectively. Since Φ is continuous and X belongs to the compact subset D, both random variables are bounded. We set

μ=𝔼[Φ(X)]=DΦ(x)dP(x),    (7a)
T=𝔼[Φ(X)Φ(X)]=DΦ(x)Φ(x)dP(x),    (7b)
Tc=Tμμ,    (7c)

where the integrals are in the Bochner sense.

We denote by μ^n and T^n the empirical mean of Φ(X) and Φ(X) ⊗ Φ(X), respectively, and by Tnc^ the empirical covariance operator, respectively. Explicitly,

μ^n=1niΦ(Xi),    (8a)
T^n=1niΦ(Xi)Φ(Xi),    (8b)
Tnc^=T^nμ^nμ^n.    (8c)

The main properties of the covariance operator and its empirical version are summarized in Proposition 2 in the Appendix in Supplementary Material.

3. The Estimator

Now we are ready to construct the estimator, whose computational aspects are discussed in section 5. The set C^n is defined by the following three steps:

a) the points xD are mapped into the corresponding centered vectors Φ(x)-μ^nH, where the center is the empirical mean;

b) the operator rm(Tnc^) is applied to each vector Φ(x)-μ^n;

c) the point xD is assigned to C^n if the distance between rmn(T^nc)(Φ(x)-μ^n) and Φ(x)-μ^n is smaller than a threshold τ.

Explicitly we have that

C^n={xD||rmn(T^nc)(Φ(x)μ^n)(Φ(x)μ^n)||τn},    (9)

where τ = τn and m = mn are chosen as a function of the number n of training data.

With the choice of the hard cut-off filter, this reduces to the KPCA algorithm [16, 17]. Indeed, rm(T^nc) is the projection Qm onto the vector space spanned by the first m eigenvectors. Hence C^n is the set of points x whose image Φ(x)-μ^n is close to Qm. For an arbitrary filter function rm, Qm is replaced by rm(T^nc), which can be interpreted as a smooth version of Qm. Note that, in general, rm(T^nc) is not a projection.

In De Vito et al. [27] a different estimator is defined. In that paper the data are mapped in the feature space H without centering the points with respect to the empirical mean and the estimator is given by

C˜n={xD|Φ(x)rmn(T^n)Φ(x)Φ(x)|τn2},

where the filter function rm is as in the present work, but rmn(T^n) is defined in terms of the eigenvectors of the non-centered second momentum T^n. To compare the two estimators note that

||rm(T^nc)(Φ(x)μ^n)(Φ(x)μ^n)||2             =(Irm(T^nc))2(Φ(x)μ^n)Φ(x)μ^n             =(Irm(T^nc))(Φ(x)μ^n)Φ(x)μ^n,

where rm*(σ)=2rm(σ)-rm(σ)2, which is a filter function too, possibly with a Lipschitz constant Lm*2Lm. Note that for the hard cut-off filter rm*(σ)=rm(σ).

Though rmn(T^n) and rmn(T^nc) are different, both C^n and C~n converge to the support of the probability distribution P, provided that the separating property (H3) holds true. Hence, one has the freedom to choose if the empirical data have or not zero mean in the feature space.

4. Main Results

In this section, we prove that the estimator C^n we introduce is strongly consistent. To state our results, for each n ∈ ℕ, we fix an integer mn ∈ ℕ and set F^n:DH to be

F^n(x)=(Irmn(T^nc))(Φ(x)μ^n),

so that Equation (9) becomes

C^n={xD||F^n(x)||τn}.    (10

4.1. Spectral Characterization

First of all, we characterize the support of P by means of Qc, the orthogonal projector onto the null space of the distribution dependence covariance operator Tc. The following theorem will show that the centered feature map

Φc:DH  Φc(x)=Φ(x)μ

sends the support C onto the intersection of Φc(D) and the closed subspace (I-Qc)H, i.e,

Φc(C)=Φc(D)(IQc)H.

Theorem 1. Assume that Φ is a separating Mercer feature map, then

C={xDQc(Φ(x)μ)=0},    (11)

where Qc is the orthogonal projector onto the null space of the covariance operator Tc.

Proof. To prove the result we need some technical lemmas, we state and prove in the Appendix in Supplementary Material. Assume first that xD is such that Qc(Φ(x) − μ) = 0. Denoted by Q the orthogonal projection onto the null space of T, by Lemma 2 QQc = Q and = 0, so that

QΦ(x)=Q(Φ(x)μ)=QQc(Φ(x)μ)=0.

Hence Lemma 1 implies that xC.

Conversely, if xC, then as above Q(Φ(x) − μ) = 0. By Lemma 2 we have that Qc(1 − Q) = ||Qcμ||−2Qc μ ⊗ Qcμ. Hence it is enough to prove that

QcμΦ(x)μ=0QcΦ(x)=Qcμ,

which holds true by Lemma 3.

4.2. Consistency

Our first result is about the convergence of F^n.

Theorem 2. Assume that Φ is a Mercer feature map. Take the sequence {mn}n such that

limnmn=+,    (12a)
Lmnκnln n,    (12b)

for some constant κ > 0, then

[limnsupxD||F^n(x)Qc(Φ(x)μ)||=0]=1.    (13)

Proof. We first prove Equation (13). Set An=I-rmn(Tnc^). Given xD,

||An(Φ(x)μ^n)Qc(Φ(x)μ)||      ||(AnQc)(Φ(x)μ)||+||An(μμ^n)||      ||rmn(T^nc)rmn(Tc)||2||Φ(x)μ||  +||(rmn(Tc)(IQc))(Φ(x)μ)||+||An(μμ^n)|| 2RLmn||T^ncTc||2         +||(rmn(Tc)(IQc))(Φ(x)μ)||+ ||μμ^n||,

where the fourth line is due to (21e), the bound ||An||=supσ[0,1]|1-rm(σ)|1, and the fact that both Φ(x) and μ are bounded by R. By (12b) it follows that

supxD||An(Φ(x)μ^n)Qc(Φμ)||2Rκnln n||T^ncTc||2  + supxD||(rmn(Tc)(IQc))(Φ(x)μ)||||μμ^n||,

so that, taking into account (24a) and (24c), it holds that

limn+supxD||An(Φ(x)μ^n)Qc(Φμ)|| =0

almost surely, provided that limn→+∞∥(rmn(Tc) − (IQc))(Φ(x) − μ)∥ = 0. This last limit is a consequence of (21d) observing that {Φ(x)-μxD} is compact since D is compact and Φ is continuous.

We add some comments. Theorem 1 suggests that the consistency depends on the fact that the vector (I-rmn(T^nc))(Φ(x)-μ^n) is a good approximation of Qc(Φ(x)−μ). By the law of large numbers, T^n and μ^n converge to T and μ, respectively, and Equation (21d) implies that, if m is large enough, (Irm(T))(Φ(x)−μ) is closed to Qc(Φ(x)−μ). Hence, if mn is large enough, see condition (12a), we expect that rmn(T^nc) is close to rmn(Tc). However, this is true only if mn goes to infinity slowly enough, see condition (12b). The rate depends on the behavior of the Lipschitz constant Lm, which goes to infinity if m goes to infinity. For example, for Tikhonov filter a sufficient condition is that mn~n12-ϵ with ϵ > 0. With the right choice of mn, the empirical decision function F^n converges uniformly to the function F(x) = Qc(Φ(x) − μ), see Equation (13).

If the map Φ is separating, Theorem 1 gives that the zero level set of F is precisely the support C. However, if C is not learnable by Φ, i.e., the equality (2) does not hold, then the zero level set of F is bigger than C. For example, if D is connected, C has not-empty interior and Φ is the feature map associated with the Gaussian kernel, it is possible to prove that F is an analytic function, which is zero on an open set, hence it is zero on the whole space D. We note that, in real applications the difference between Gaussian and Abel kernel, which is separating, is not so big and in our experience the Gaussian kernel provides a reasonable estimator.

From now on we assume that Φ is separating, so that Theorem 1 holds true. However, the uniform convergence of F^n to F does not imply that the zero level sets of F^n converges to C = F−1(0) with respect to the Hausdorff distance. For example, with the Tikhonov filter F^n-1(0) is always the empty set. To overcome the problem, C^n is defined as the τn-neighborhood of the zero level set of F^n, where the threshold τm goes to zero slowly enough.

Define the data dependent parameter τ^n as

τ^n=max1in||F^n(Xi)||.    (14)

Since F^n[0,1], clearly τ^n[0,1] and the set estimator becomes

C^n={xD||F^n(x)||τ^n}.

The following result shows that C^n is a universal strongly consistent estimator of the support of the probability distribution P. Note that for KPCA the consistency is not universal since the choice of mn depends on some a-priori information about the decay of the eigenvalues of the covariance operator Tc, which depends on P.

Theorem 3. Assume that Φ is a separating Mercer feature map. Take the sequence {mn}n satisfying (12a)-(12b) and define τ^n by (14). Then

[limnτ^n=0]=1,    (15a)
[limndH(C^n,C)=0]=1.    (15b)

Proof. We first show Equation (15a). Set F(x) = Qc(Φ(x) − μ) and let E be the event on which F^n converges uniformly to F(x), and F be the event such that XiC for all i ≥ 1. Theorem 2 shows that ℙ[E] = 1 and, since C is the support, then ℙ[F] = 1. Take ω ∈ EF and fix ϵ > 0, then there exists n0 > 0 (possibly depending on ω and ϵ) such that for all nn0 |F^n(x)-F(x)|ϵ for all xD. By Theorem 1 F(x) = 0 for all xC and X1(ω), …, Xn(ω) ∈ C, it follows that |F^n(Xi(ω))|ϵ for all 1 ≤ in so that 0τ^n(ω)ϵ, so that the sequence τ^n(ω) goes to zero. Since ℙ[EF] = 1 Equation (15a) holds true.

We split the proof of Equation (15b) in two steps. We first show that with probability one limn+supxC^nd(x,C)=0. On the event EF, suppose, by contraction, that the sequence {supxC^nd(x,C)}n does not converge to zero. Possibly passing to a subsequence, for all n ∈ ℕ there exists xnC^n such that d(xn,C^n)ϵ0 for some fixed ϵ0 > 0. Since D is compact, possibly passing to a subsequence, {xn}n converges to x0D with d(x0, C) ≥ ϵ0. We claim that x0C. Indeed,

||Qc(Φ(x0)μ)||    ||Qc(Φ(x0)Φ(xn))||                                + ||F^n(xn)Qc(Φ(xn)μ)||+||F^n(xn)||                                     ||Qc(Φ(x0)Φ(xn))||                                 + supxD||F^n(x)Qc(Φ(x)μ)+τn,

since xnC^n means that ||F^n(xn)||τn. If n goes to infinity, since Φ is continuous and by the definition of E and F, the right side of the above inequality goes to zero, so that ||Qc(Φ(x0)-μ)||=0, i.e., by Theorem 1 we get x0C, which is a contraction since by construction d(x0, C) ≥ ϵ0 > 0.

We now prove that

limnsupxCd(x,C^n).

For any xD, set X1,n(x) to be a first neighbor of x in the training set {X1, …, Xn}. It is known that for all xC,

[limn+d(X1,n(x),x)=0]=1,    (16)

see for example Lemma 6.1 of Györfi et al. [35].

Choose a denumerable family {zj}jJ in C such that is dense in C. By Equation (16) there exists an event G with such that ℙ[G] = 1 and, on G, for all jJ

limn+d(X1,n(zj),zj)=0.

Fix ω ∈ G, we claim that limnsupxCd(x,C^n)=0. Observe that, by definition of τ^n, XiC^n for all 1 ≤ in and

supxCd(x,C^n)supxCmin1ind(x,Xi)=supxCd(X1,n(x),x),

so that it is enough to show that limn+supxCd(X1,n(x),x)=0.

Fix ϵ > 0. Since C is compact, there is a finite subset JϵJ such that {B(zj, ϵ)}jJϵ is a finite covering of C. Furthermore,

supxCd(X1,n(x),x)maxjJϵd(X1,n(zj),zj)+ϵ.    (17)

Indeed, fix xC, there exists an index jJϵ such that xB(zj, ϵ). By definition of first neighbor, clearly

d(X1,n(x),x)d(X1,n(zj),x),

so that by triangular inequality we get

d(X1,n(x),x)d(X1,n(zj),x)d(X1,n(zj),zj)+d(zj,x)                           d(X1,n(zj),zj)+ϵ                           maxjJϵd(X1,n(zj),zj)+ϵ.

Taking the supremum over C we get the claim. Since ω ∈ G and Jϵ is finite,

limn+maxjJϵd(X1,n(zj),zj)=0,

so that by Equation (17)

limsup n+supxCd(X1,n(x),x)ϵ.

Since ϵ is arbitrary, we get limn+supxCd(X1,n(x),x)=0, which implies that

limn+supxCd(x,C^n)=0.

Theorem 3 is an asymptotic result. Up to now, we are not able to provide finite sample bounds on dHn, C). It is possible to have finite sample bounds on ||F^n(x)-Qc(Φ(x)-μ)||, as in Theorem 7 of De Vito et al. [25] with the same kind of proof.

4.3. The Separating Condition

The following two examples clarify the notion of the separating condition.

Example 1. Let D be a compact subset of ℝ2, H=6 with the euclidean scalar product, and Φ:D6 be the feature map

Φ((x,y))=(x2,y2,2xy,2x,2y,1),

whose corresponding Mercer kernel is a polynomial kernel of degree two, explicitly given by

K(x1,y1;x2,y2)=(x1x2+y1y2+1)2.

Given a vector f=(f1,,f6), the corresponding elementary set is the conic

Cf,c={(x,y)Df1x2+f2y2+f32xy            +f42x+f52y+f6=0},

Conversely, all the conics are elementary sets. The family of all the intersections of at most five conics, i.e., the sets whose cartesian equation is a system of the form

{f11x2+f12y2+f132xy+f142x+f152y+f16=0f51x2+f52y2+f532xy+f542x+f552y+f56=0,

where f11, …, f56 ∈ ℝ.

Example 2. The data are the random vectors in ℝ2

Xi=(sin(aΘi+b),sin(cΘi+d)),

where a, c ∈ ℕ, b, d ∈ [0, 2π] and Θ1, …, Θn are independent random variables, each of them uniformly distributed on [0, 2π]. Setting D=[-1,1]2, clearly XiD and the support of their common probability distribution is the Lissajous curve

C=Lisa,b,c,d={(sin(aθ+b),sin(cθ+d))Dθ[0,2π]}.

Figure 2 shows two examples of Lissajous curves. As a filter function rm, we fix the hard cut-off filter where m is the number of eigenvectors corresponding to the highest eigenvalues we keep. We denote by Ĉnm,τ the corresponding estimator given by (10).

FIGURE 2
www.frontiersin.org

Figure 2. Examples of Lissajous curves for different values of the parameters. (Left) Lis1,0,1,π2. (Right) Lis2,0.11,1,0.3.

In the first two tests we use the polynomial kernel (18), so that the elementary learnable sets are conics. One can check that the rank of Tc is less or equal than 5. More precisely, if Lisa,b,c,d is a conic, the rank of Tc is 4 and we need to estimate five parameters, whereas if Lisa,b,c,d is not a conic, Lisa,b,c,d is not a learnable set and the rank of Tc is 5.

In the first test the data are sampled from the distribution supported on the circumference Lis1,0,1,π2 (see panel left of Figure 2). In Figure 3 we draw the set Ĉnm,τ for different values of m and τ when n varies. In this toy example n = 5 is enough to learn exactly the support, hence for each n = 2, …, 5 the corresponding values of mn and τn are mn = 1, 2, 3, 4 and τn = 0.01, 0.005, 0.005, 0.002.

FIGURE 3
www.frontiersin.org

Figure 3. From left to right, top to bottom. The set Ĉnm,τ with n, respectively 2, 3, 4, 5, m = 1, 2, 3, 4 and τ = 0.1, 0.005, 0.005, 0.002.

In the second test the data are sampled from the distribution supported on the curve Lis2,0.11,1,0.3, which is not a conic (see panel right of Figure 2). In Figure 4 we draw the set Ĉnm,τ for n = 10, 20, 50, 100, m = 4, and τ = 0.01. Clearly, C^n is not able to estimate Lis2,0.11,1,0.3.

FIGURE 4
www.frontiersin.org

Figure 4. The set Ĉn4,0.01 with n respectively 10, 20, 50, 100.

In the third test, we use the Abel kernel with the data sampled from the distribution supported on the curve Lis2,0.11,1,0.3 (see panel right of Figure 2). In Figure 5 we show the set Ĉnm,τ for n = 20, 50, 100, 500, m = 5, 20, 30, 50, and τ = 0.4, 0.35, 0.3, 0.2. According the fact that the kernel is separating, C^n is able to estimate Lis2,0.11,1,0.3 correctly.

FIGURE 5
www.frontiersin.org

Figure 5. From left to right and top to bottom. The learned set Ĉnm,τ with n respectively of 20, 50, 100, 500, and m = 5, 20, 30, 50, τ = 0.4, 0.35, 0.3, 0.2.

We now briefly discuss how to select the parameter mn and τn from the data. The goal of set-learning problem is to recover the support of the probability distribution generating the data by using the given input observations. Since no output is present, set-learning belongs to the category of unsupervised learning problems, for which there is not a general framework accounting for model selection. However there are some possible strategies (whose analysis is out of the scope of this paper). A first approach, we used in our simulations, is based on the monotonicity properties of Ĉnm,τ with respect to m, τ. More precisely, given f ∈ (0, 1), we select (the smallest) m and (the biggest) τ such that at most nf observed points belong to the the estimated set. It is possible to prove that this method is consistent when f tends to 1 as the number of observations increases. Another way to select the parameters consists in transforming the set-learning problem is a supervised one and then performing standard model selection techniques like cross validation. In particular set-learning can be casted in a classification problem by associating the observed example to the class +1 and by defining an auxiliary measure μ (e.g., uniform on a ball of interest in D) associated to −1, from which n i.i.d. points are drawn. It is possible to prove that this last method is consistent when μ(suppρ) = 0.

4.4. The Role of the Regularization

We now explain the role of the filter function. Given a training set X1, …, Xn of size n, the separating property (3) applied to the support of the empirical distribution gives that

{X1,,Xn}={xDQnc^(Φ(x)μ^n)=0},

where μ^n is the empirical mean and IQnc^ is the orthogonal projection onto the linear space spanned by the family {Φ(X1)-μ^n,,Φ(Xn)-μ^n}, which are the centered images of the examples. Hence, given a new point xD the condition |Qnc^(Φ(x)μ^n)||  τ with τ ≪ 1 is satisfied if only if x is close to one of the examples in the training set. Hence the naive estimator {xD|Qnc^(Φ(x)μn^)τ} overfits the data. Hence we would like to replace Qnc^ with an operator, which should be close to the identity on the linear subspace spanned by {Φ(X1)-μ^n,,Φ(Xn)-μ^n} and it should have a small range. To modulate the two requests, one can consider the following optimization problem

minAS2(1ni=1n(IA)(Φ(Xi)μ^n)2+λA22).

We note that if A is a projection its Hilbert-Schmidt norm ||A||2 is the square root of the dimension of the range of A. Since

1ni=1n||(IA)(Φ(Xi)μ^n)||2     =Tr((IA)(IA))1ni=1n(Φ(Xi)μ^n)(Φ(Xi)μ^n))     =Tr((IA)(IA)Tnc),

where Tr(A) is the trace, A is the transpose and ||A||2=Tr(AA) is the Hilbert-Schmidt norm, then

1ni=1n||(IA)(Φ(Xi)μ^n)||2+λ||A||22= Tr((IA)(IA)T^nc + λmAA),

and the optimal solution is given by

Aopt=Tnc^(Tnc^+λm)1,

i.e., Aopt is precisely the operator rm(Tnc^) with the Tikhonov filter rm(σ)=σσ+λ and λ=Rm. A different choice of the filter function rm corresponds to a different regularization of the least-square problem

min1ni=1n||(IA)(Φ(Xi)μ^n)||2.

5. The Kernel Machine

In this section we show that the computation of ||F^n(x)||, in terms of which is defined the estimator C^n, reduces to a finite dimensional problem, depending only on the Mercer kernel K, associated with the feature map. We introduce the centered sampling operator

Snc:n  (Sncf)i=f,Φ(Xi)μ^n,

whose transpose is given by

Snc:n  Sncv=i=1vi(Φ(Xi)μ^n),

where vi is the i-th entry of the column vector v ∈ ℝn. Hence, it holds that

1nSncSnc=Tnc^,1nSncSnc=(In11n)Knn(In11n),

where Kn is the n × n matrix whose (i, j)-entry is K(Xi, Xj) and In is the identity n × n matrix, so that the (i, j)-entry of SncSnc is

K(Xi,Xj)1nbK(Xi,Xb)1naK(Xa,Xj)+1n2a,bK(Xa,Xb).

Denoted by ℓ the rank of SncSnc, take the singular value decomposition of SncSnc/n, i.e.,

SncSncn=VΣV,

where V is an n × ℓ matrix whose columns vjn are the normalized eigenvectors, VV=I, and Σ is a diagonal ℓ × ℓ matrix with the strictly positive eigenvalues on the diagonal, i.e., Σ=diag(σ^1,,σ^). Set U=SncVΣ-12, regarded as operator from ℝ to H, then a simple calculation shows that

Tnc^=UΣU  rm(Tnc^)=Urm(Σ)U,

where rm(Σ) is the diagonal ℓ × ℓ matrix

rm(Σ)=diag(rm(σ^1),,rm(σ^)),

and the equation for rm(Tnc^) holds true since by assumption rm(0) = 0. Hence

||F^n(x)||2=(Irm(Tnc^))(Φ(x)μ^n)(Irm(Tnc^))(Φ(x)μ^n)                   =Φ(x)μ^nΦ(x)μ^n(2rm(Tnc^)                 rm(Tnc^)2)(Φ(x)μ^n)Φ(x)μ^n                   =Φ(x)μ^nΦ(x)μ^nU(2rm(Σ)                 rm(Σ)2)U(Φ(x)μ^n)Φ(x)μ^n                   =Φ(x)μ^nΦ(x)μ^nVΣ12(2rm(Σ)                  rm(Σ)2)Σ12VSnc(Φ(x)μ^n)Snc(Φ(x)μ^n)                   =w(x)v(x)Gmv(x),    (19)

where the real number w(x)=Φ(x)-μ^nΦ(x)-μ^n is

w(x)=K(x,x)2nbK(x,Xb)+1n2a,bK(Xa,Xb),

the i-th entry of the column vector v(x) ∈ ℝn is

v(x)i=(Snc(Φ(x)μ^n))i=K(x,Xi)1naK(Xa,x)           1nbK(Xi,Xb)+1n2a,bK(Xa,Xb),

the diagonal ℓ × ℓ matrix Rm(Σ)=Σ-1(2rm(Σ)-rm(Σ)2) is

Rm(Σ)=diag(2rm(σ^1)rm(σ^1)2σ^1,,2rm(σ^)rm(σ^)2σ^),

and the n × n-matrix Gm is

Gm=VRm(Σ)V.    (20)

In Algorithm 1 we list the corresponding MatLab Code.

www.frontiersin.org

Algorithm 1. Matlab code for Set Learning.

The above equations make clear that both F^n and C^n can be computed in terms of the singular value decomposition (V, Σ) of the n × n Gram matrix Kn and of the filter function rm, so that F^n belongs to the class of kernel methods and C^n is a plug-in estimator. For the hard cut-off filter, one simply has

Rm(Σ)ii={1σ^iσ^iσ^m0σ^i<σ^m.

For real applications, a delicate issue is the choice of the parameters m and τ, we refer to Rudi et al. [31] for a detailed discussion. Here, we add some simple remarks.

We first discuss the role of τ. According to (10), Ĉnm,τĈnm,τ whenever τ < τ′. We exemplify this behavior with the dataset of Example 2. The training set is sampled from the distribution supported on the curve Lis2,0.11,1,0.3 (see panel right of Figure 2) and we compute C^n with the Abel kernel, n = 100 and m ranging over 5, 10, 20, 50. Figure 6 shows the nested sets when τ runs in the associated color-bar.

FIGURE 6
www.frontiersin.org

Figure 6. From left to right and top to bottom: The family of sets Ĉ1005,τ,Ĉ10010,τ,Ĉ10020,τ,Ĉ10050,τ with τ varying as in the related colorbar.

Analyzing the role of m, we now show that, for the the hard cut-off filter, Ĉnm,τĈnm,τ whenever m′ ≤ m. Indeed, this filter satisfies rm(σ)rm(σ) and, since 0 ≤ rm(σ) ≤ 1, one has (1-rm(σ))2(1-rm(σ))2. Hence, denoted by {u^j}j a base of eigenvectors of Tnc^, it holds that

||(Irm(Tnc^))(Φ(x)μ^n)||2=j(1rm(σ^j))2Φ(x)μ^nu^j2                                                      j(1rm(σ^j))2Φ(x)μ^nu^j2                                                      = ||(Irm(Tnc^))(Φ(x)μ^n)||2.

Hence, for any point in xĈnm,τ,

 ||(Irm(Tnc^))(Φ(x)μ^n) ||2   ||(Irm(Tnc^))(Φ(x)μ^n) ||2τ2,

so that xĈnm,τ.

As above, we illustrate the different choices of m with the data sampled from the curve Lis2,0.11,1,0.3 and the Abel kernel where n = 100 and τ ranges over 0.25, 0.3, 0.4, 0.5. Figure 7 shows the nested sets when m runs in the associated color-bar.

FIGURE 7
www.frontiersin.org

Figure 7. From left to right and top to bottom: The family of sets Ĉ100m,0.25,Ĉ100m,0.3,Ĉ100m,0.4,Ĉ100m,0.5 with m varying as in the related colorbar.

6. Discussion

We presented a new class of set estimators, which are able to learn the support of an unknown probability distribution from a training set of random data. The set estimator is defined through a decision function, which can be seen as a novelty/anomality detection algorithm as in Schölkopf et al. [6].

The decision function we defined is a kernel machine. It is computed by the singular value decomposition of the empirical (kernel)-covariance matrix and by a low pass filter. An example of filter is the hard cut-off function and the corresponding decision function reduces to KPCA algorithm for novelty detection first introduced by Hoffmann [16]. However, we showed that it is possible to choose other low pass filters, as it was done for a class of supervised algorithms in the regression/classification setting [38].

Under some weak assumptions on the low pass filter, we proved that the corresponding set estimator is strongly consistent with respect to the Hausdorff distance, provided that the kernel satisfies a suitable separating condition, as it happens, for example, for the Abel kernel. Furthermore, by comparing Theorem 2 with a similar consistency result in De Vito et al. [27], it appears clear that the algorithm correctly learns the support both if the data have zero mean, as in our paper, and if the data are not centered, as in De Vito et al. [27]. On the contrary, if the separating property does not hold, the algorithm learns only the supports that are mapped into linear subspaces by the feature map defined by the kernel.

The set estimator we introduced depends on two parameters: the effective number m of eigenvectors defining the decision function and the thickness τ of the region estimating the support. The role of these parameters and of the separating property was briefly discussed by a few tests on toy data.

We finally observe that our class of set learning algorithms is very similar to classical kernel machines in supervised learning. So, in order to reduce both the computational cost and the memory requirements, there is the possibility to successfully implement some new advanced approximation techniques, for which there exist theoretical guarantees for the statistical learning setting. For example random features [39, 40], Nyström projections [41, 42] or mixed approaches with iterative regularization and preconditioning [43, 44].

Author Contributions

All authors listed have made a substantial, direct and intellectual contribution to the work, and approved it for publication.

Conflict of Interest Statement

The authors declare that the research was conducted in the absence of any commercial or financial relationships that could be construed as a potential conflict of interest.

Acknowledgments

ED is member of the Gruppo Nazionale per l'Analisi Matematica, la Probabilità e le loro Applicazioni (GNAMPA) of the Istituto Nazionale di Alta Matematica (INdAM).

Supplementary Material

The Supplementary Material for this article can be found online at: https://www.frontiersin.org/articles/10.3389/fams.2017.00023/full#supplementary-material

Footnotes

1. ^Here, the labeling is different from the one in (6), where the eigenvalues are repeated according to their multiplicity.

References

1. Devroye L, Wise GL. Detection of abnormal behavior via nonparametric estimation of the support. SIAM J Appl Math. (1980) 38:480–8. doi: 10.1137/0138038

CrossRef Full Text | Google Scholar

2. Korostelëv AP, Tsybakov AB. Minimax Theory of Image Reconstruction. New York, NY: Springer-Verlag (1993).

Google Scholar

3. Dümbgen L, Walther G. Rates of convergence for random approximations of convex sets. Adv Appl Probab. (1996) 28:384–93. doi: 10.2307/1428063

CrossRef Full Text | Google Scholar

4. Cuevas A, Fraiman R. A plug-in approach to support estimation. Ann Stat. (1997) 25:2300–12. doi: 10.1214/aos/1030741073

CrossRef Full Text | Google Scholar

5. Tsybakov AB. On nonparametric estimation of density level sets. Ann Stat. (1997) 25:948–69. doi: 10.1214/aos/1069362732

CrossRef Full Text | Google Scholar

6. Schölkopf B, Platt J, Shawe-Taylor J, Smola A, Williamson R. Estimating the support of a high-dimensional distribution. Neural Comput. (2001) 13:1443–71. doi: 10.1162/089976601750264965

CrossRef Full Text | Google Scholar

7. Cuevas A, Rodríguez-Casal A. Set estimation: an overview and some recent developments. In: Recent Advances and Trends in Nonparametric Statistics. Elsevier: Amsterdam (2003). p. 251–64.

8. Reitzner M. Random polytopes and the Efron-Stein jackknife inequality. Ann Probab. (2003) 31:2136–66. doi: 10.1214/aop/1068646381

CrossRef Full Text | Google Scholar

9. Steinwart I, Hush D, Scovel C. A classification framework for anomaly detection. J Mach Learn Res. (2005) 6:211–32.

Google Scholar

10. Vert R, Vert JP. Consistency and convergence rates of one-class SVMs and related algorithms. J Mach Learn Res. (2006) 7:817–54.

Google Scholar

11. Scott CD, Nowak RD. Learning minimum volume sets. J Mach Learn Res. (2006) 7:665–704.

Google Scholar

12. Biau G, Cadre B, Mason D, Pelletier B. Asymptotic normality in density support estimation. Electron J Probab. (2009) 91:2617–35.

Google Scholar

13. Cuevas A, Fraiman R. Set estimation. In: W. Kendall and I. Molchanov, editors. New Perspectives in Stochastic Geometry. Oxford: Oxford University Press (2010). p. 374–97.

14. Bobrowski O, Mukherjee S, Taylor JE. Topological consistency via kernel estimation. Bernoulli (2017) 23:288–328. doi: 10.3150/15-BEJ744

CrossRef Full Text | Google Scholar

15. Campos GO, Zimek A, Sander J, Campello RJ, Micenková B, Schubert E, et al. On the evaluation of unsupervised outlier detection: measures, datasets, and an empirical study. Data Min Knowl Discov. (2016) 30:891–927. doi: 10.1007/s10618-015-0444-8

CrossRef Full Text | Google Scholar

16. Hoffmann H. Kernel PCA for novelty detection. Pattern Recognit. (2007) 40:863–74. doi: 10.1016/j.patcog.2006.07.009

CrossRef Full Text | Google Scholar

17. Schölkopf B, Smola A, Müller KR. Nonlinear component analysis as a kernel eigenvalue problem. Neural Comput. (1998) 10:1299–319.

Google Scholar

18. Ristic B, La Scala B, Morelande M, Gordon N. Statistical analysis of motion patterns in AIS data: anomaly detection and motion prediction. In: 2008 11th International Conference on Information Fusion (2008). p. 1–7.

Google Scholar

19. Lee HJ, Cho S, Shin MS. Supporting diagnosis of attention-deficit hyperactive disorder with novelty detection. Artif Intell Med. (2008) 42:199–212. doi: 10.1016/j.artmed.2007.11.001

CrossRef Full Text | Google Scholar

20. Valero-Cuevas FJ, Hoffmann H, Kurse MU, Kutch JJ, Theodorou EA. Computational models for neuromuscular function. IEEE Rev Biomed Eng. (2009) 2:110–35. doi: 10.1109/RBME.2009.2034981

CrossRef Full Text | Google Scholar

21. He F, Yang JH, Li M, Xu JW. Research on nonlinear process monitoring and fault diagnosis based on kernel principal component analysis. Key Eng Mater. (2009) 413:583–90. doi: 10.4028/www.scientific.net/KEM.413-414.583

CrossRef Full Text | Google Scholar

22. Maestri ML, Cassanello MC, Horowitz GI. Kernel PCA performance in processes with multiple operation modes. Chem Prod Process Model. (2009) 4:1934–2659. doi: 10.2202/1934-2659.1383

CrossRef Full Text | Google Scholar

23. Cheng P, Li W, Ogunbona P. Kernel PCA of HOG features for posture detection. In: VCNZ'09. 24th International Conference on Image and Vision Computing New Zealand, 2009. Wellington (2009). p. 415–20.

Google Scholar

24. Sofman B, Bagnell JA, Stentz A. Anytime online novelty detection for vehicle safeguarding. In: 2010 IEEE International Conference on Robotics and Automation (ICRA). Pittsburgh, PA (2010). p. 1247–54.

Google Scholar

25. De Vito E, Rosasco L, Toigo A. A universally consistent spectral estimator for the support of a distribution. Appl Comput Harmonic Anal. (2014) 37:185–217. doi: 10.1016/j.acha.2013.11.003

CrossRef Full Text | Google Scholar

26. Steinwart I. On the influence of the kernel on the consistency of support vector machines. J Mach Learn Res. (2002) 2:67–93. doi: 10.1162/153244302760185252

CrossRef Full Text | Google Scholar

27. De Vito E, Rosasco L, Toigo A. Spectral regularization for support estimation. In: NIPS. Vancouver, BC (2010). p. 1–9.

Google Scholar

28. Engl HW, Hanke M, Neubauer A. Regularization of Inverse Problems. Vol. 375 of Mathematics and its Applications. Dordrecht: Kluwer Academic Publishers Group (1996).

29. Lo Gerfo L, Rosasco L, Odone F, De Vito E, Verri A. Spectral algorithms for supervised learning. Neural Comput. (2008) 20:1873–97. doi: 10.1162/neco.2008.05-07-517

CrossRef Full Text | Google Scholar

30. Blanchard G, Mücke N. Optimal rates for regularization of statistical inverse learning problems. In: Foundations of Computational Mathematics. (2017). Available online at: https://arxiv.org/abs/1604.04054

Google Scholar

31. Rudi A, Odone, F, De Vito E. Geometrical and computational aspects of Spectral Support Estimation for novelty detection. Pattern Recognit Lett. (2014) 36:107–16. doi: 10.1016/j.patrec.2013.09.025

CrossRef Full Text | Google Scholar

32. Rudi A, Canas, GD, Rosasco L. On the sample complexity of subspace learning. In: Burges CJC, Bottou L, Welling M, Ghahramani Z, Weinberger KQ, editors. Advances in Neural Information Processing Systems. Lake Tahoe: Neural Information Processing Systems Conference (2013). p. 2067–75.

Google Scholar

33. Rudi A, Canas GD, De Vito E, Rosasco L. Learning Sets and Subspaces. In: Suykens JAK, Signoretto M, and Argyriou A, editors. Regularization, Optimization, Kernels, and Support Vector Machines. Boca Raton, FL: Chapman and Hall/CRC (2014). p. 337.

Google Scholar

34. Blanchard G, Bousquet O, Zwald L. Statistical properties of kernel principal component analysis. Machine Learn. (2007) 66:259–94. doi: 10.1007/s10994-006-8886-2

CrossRef Full Text | Google Scholar

35. Györfi L, Kohler M, Krzy zak A, Walk H. A Distribution-Free Theory of Nonparametric Regression. Springer Series in Statistics. New York, NY: Springer-Verlag (2002).

Google Scholar

36. Steinwart I, Christmann A. Support Vector Machines. Information Science and Statistics. New York, NY: Springer (2008).

37. Zwald L, Blanchard G. On the Convergence of eigenspaces in kernel principal component analysis. In: Weiss Y, Schölkopf B, Platt J, editors. Advances in Neural Information Processing Systems 18. Cambridge, MA: MIT Press (2006). p. 1649–56.

Google Scholar

38. De Vito E, Rosasco L, Caponnetto A, De Giovannini U, Odone F. Learning from examples as an inverse problem. J Machine Learn Res. (2005) 6:883–904.

Google Scholar

39. Rahimi A, Recht B. Random features for large-scale kernel machines. In: Koller D, Schuurmans D, Bengio, Y, Bottou L, editors. Advances in Neural Information Processing Systems. Vancouver, BC: Neural Information Processing Systems Conference (2008). p. 1177–84.

Google Scholar

40. Rudi A, Camoriano R, Rosasco L. Generalization properties of learning with random features. arXiv preprint arXiv:160204474 (2016).

Google Scholar

41. Smola AJ, Schölkopf B. Sparse greedy matrix approximation for machine learning. In: Proceeding ICML '00 Proceedings of the Seventeenth International Conference on Machine Learning. Stanford, CA: Morgan Kaufmann (2000). p. 911–18.

Google Scholar

42. Rudi A, Camoriano R, Rosasco L. Less is more: nyström computational regularization. In: Cortes C, Lawrence ND, Lee DD, Sugiyama M, Garnett R, editors. Advances in Neural Information Processing Systems. Montreal, QC: Neural Information Processing Systems Conference (2015). p. 1657–1665.

Google Scholar

43. Camoriano R, Angles T, Rudi A, Rosasco L. NYTRO: when subsampling meets early stopping. In: Gretton A, Robert CC, editors. Artificial Intelligence and Statistics. Cadiz: Proceedings of Machine Learning Research (2016). p. 1403–11.

Google Scholar

44. Rudi A, Carratino L, Rosasco L. FALKON: an optimal large scale Kernel method. arXiv preprint arXiv:170510958 (2017).

Google Scholar

45. Folland G. A Course in Abstract Harmonic Analysis. Studies in Advanced Mathematics. Boca Raton, FL: CRC Press (1995).

Google Scholar

46. Birman MS, Solomyak M. Double operator integrals in a Hilbert space. Integr Equat Oper Theor. (2003) 47:131–68. doi: 10.1007/s00020-003-1157-8

CrossRef Full Text | Google Scholar

47. De Vito E, Umanità V, Villa S. A consistent algorithm to solve Lasso, elastic-net and Tikhonov regularization. J Complex. (2011) 27:188–200. doi: 10.1016/j.jco.2011.01.003

CrossRef Full Text | Google Scholar

Keywords: support estimation, Kernel PCA, novelty detection, dimensionality reduction, regularized Kernel methods

Citation: Rudi A, De Vito E, Verri A and Odone F (2017) Regularized Kernel Algorithms for Support Estimation. Front. Appl. Math. Stat. 3:23. doi: 10.3389/fams.2017.00023

Received: 11 August 2017; Accepted: 25 October 2017;
Published: 08 November 2017.

Edited by:

Ding-Xuan Zhou, City University of Hong Kong, Hong Kong

Reviewed by:

Ting Hu, Wuhan University, China
Sergiy Pereverzyev, University of Innsbruck, Austria
Qiang Wu, Middle Tennessee State University, United States

Copyright © 2017 Rudi, De Vito, Verri and Odone. This is an open-access article distributed under the terms of the Creative Commons Attribution License (CC BY). The use, distribution or reproduction in other forums is permitted, provided the original author(s) or licensor are credited and that the original publication in this journal is cited, in accordance with accepted academic practice. No use, distribution or reproduction is permitted which does not comply with these terms.

*Correspondence: Ernesto De Vito, devito@dima.unige.it

Download