quantization clustering banach

Upload: mario-cabrera

Post on 10-Apr-2018

233 views

Category:

Documents


0 download

TRANSCRIPT

  • 8/8/2019 Quantization Clustering Banach

    1/26

    L 1 -quantization and clustering in Banach spaces

    Thomas LALO E

    Institut de Mathematiques et de Modelisation de MontpellierUMR CNRS 5149, Equipe de Probabilites et Statistique

    Universite Montpellier II, Cc 051Place Eugene Bataillon, 34095 Montpellier Cedex 5, France

    [email protected]

    Abstract

    Let X be a random variable with distribution taking values ina Banach space H. First, we establish the existence of an optimalquantization of with respect to the L1 -distance. Second, we proposeseveral estimators of the optimal quantizer in the potentially innite-dimensional space

    H, with associated algorithms. Finally, we discuss

    practical results obtained from real-life data sets.

    Key-words and phrases: Quantization, clustering, L1-distance, Banachspace.

    1 Introduction

    Clustering consists in partitioning a data set into subsets (or clusters), sothat the data in each subset share some common trait. Proximity is de-termined according to some distance measure. For a thorough introductionof the subject, we refer to the book by Kaufman and Rousseeuw [14]. The

    origin of clustering goes back to 45 years ago, when some biologists and so-ciologists began to search for automatics methods to build different groupswith their data. Today, clustering is used in many elds. For example,in medical imaging, it can be used to differentiate between different typesof tissue and blood in a three dimensional image. Market researchers useit to partition the general population of consumers into market segmentsand to better understand the relationships between different groups of con-sumers/potential customers. There are also many different applications inarticial intelligence, sociology, medical research, or political sciences.

    1

  • 8/8/2019 Quantization Clustering Banach

    2/26

    In the present paper, the clustering method we investigate lays on the

    technique of quantization, commonly used in signal compression (Graf andLuschgy [12], Linder [17]). Given a normed space ( H, . ), a codebook (of size k) is dened by a subset C Hwith cardinality k. Then, each x His represented by a unique xCvia the function q,

    q : H Cx x,

    which is called a quantizer. Here we come back to the clustering, as wecreate clusters in the data by regrouping the observations which have thesame image by q.

    Denote by d the distance induced by the norm on H:d : H H R +

    (x, y) x y .In this paper, observations are modeled by a random variable X on H, withdistribution . The quality of the approximation of X by q(X ) is then givenby the distortion E d X, q(X ) . Thus the aim is to minimize E d X, q(X )among all possible quantizers. However, in practice, the distribution of the observations is unknown, and we only have at hand n independent ob-servations X 1, . . . , X n with the same distribution than X . The goal is thento minimize the empirical distortion:

    1n

    n

    i=1d(X i , q(X i )) .

    Since the early work of Hartigan [13] and Pollard [19, 20, 21], the perfor-mances of clustering have been considered by many authors. Convergenceproperties of the minimizer qn of the empirical distortion have been mostlystudied in the case when H= R d . Consistency of qn was shown by Pollard[19, 21] and Abaya and Wise [1]. Rates of convergence have been consideredby Pollard [20], Linder, Lugosi, and Zeger [18], Linder [17].

    As a matter of facts, in many practical problems, input data items are in theform of random functions (speech recordings, spectra, images) rather thanstandard vectors, and this casts the clustering problem into the general classof functional data analysis. Even though in practice such observations areobserved at discrete sampling points, the challenge in this context is to inferthe data structure by exploiting the innite-dimensional nature of the obser-vations. The last few years have witnessed important developments in boththe theory and practice of functional data analysis, and many traditionaldata analysis tools have been adapted to handle functional inputs. The

    2

  • 8/8/2019 Quantization Clustering Banach

    3/26

    book by Ramsay and Silverman [22] provides a comprehensive introduction

    to the area. Recently, Biau, Devroye, and Lugosi [2] gave some consistencyresults in Hilbert spaces and with a L2-based distortion.

    Thus, the rst novelty in this paper is to consider data taking place in aseparable and reexive Banach space, with no restriction on their dimen-sion. The second novelty is that we consider a L1-based distortion, whichleads to more robust estimators. For a discussion of the advantage of theL1-distance we refer the reader to the paper by Kemperman [15].

    This setup calls for substantially different arguments to prove results whichare known to be true when considering nite dimensional spaces and a L2-

    based distortion. In particular, specic notions will be required, such asweak topology (Dunford and Schwartz [10]), lower semi-continuity (Ekelandand Temam [10]) and entropy (Van der Vaart and Wellner [23]).

    The document is organized as follows. We rst provide the formal contextof quantization in Banach space in the rst part of Section 2. Then, wefocus on the problem of the existence of an optimal quantizer. In Sections 3and 4 we study two consistent estimators of this optimal quantizer, and weconfront them to real-life data in Section 5. Proofs are collected in AppendixA.

    2 Quantization in a Banach space

    2.1 General framework

    The fact that the closed bounded balls are not compact is a major problemwhen considering innite dimensional spaces. To overcome this, the classi-cal solution is to consider reexive spaces, i.e., spaces in which the closedbounded balls are compact for the weak topology (Dunford and Schwartz[9]). Thus, throughout the document, ( H, . ) will denote a reexive andseparable Banach space. We let X be a H-valued random variable withdistribution such as E X < .Given a set C= {yi}ki=1 of points in Hk , any Borel application q : H Ciscalled a quantizer. The set Cis called a codebook, and the yi , i = 1 , . . . , kare the centers of C. The error made by replacing X by q(X ) is measuredby the distortion:

    D (, q) = E d(X, q(X )) = H x q(x) (dx).Note that D (, q) < since E X < . For a given k, the aim is tominimize D (, .) among the set Qk of all possible k-quantizers. The optimal

    3

  • 8/8/2019 Quantization Clustering Banach

    4/26

    distortion is then dened by

    Dk () = inf qQkD(, q).

    When it exists, a quantizer q satisfying D(, q) = Dk() is said to be anoptimal quantizer.

    Any quantizer is characterized by its codebook C= {yi}ki=1 and a partitionof Hin cells S i = {xH: q(x) = yi}, i = 1 , . . . , k via the ruleq(x) = yi xS i .

    Thus, from now on, we will dene a quantizer by its codebook and its cells.Let us consider the particular family of Voronoi partitions, constructed bythe nearest neighbors rule. That is, for each center of the codebook, acell is constituted by the elements x Hwhich are the closest to him(Gersho and Gray [11]). A quantizer with such a partition is named anearest neighbor quantizer, and we denote by Qknn the set of all k-nearestneighbor quantizers. It can be easily proven (see Lemma 1 in Linder [17])that

    inf qQk

    D (, q) = inf qQknn

    D(, q).

    More precisely, given two quantizersq Qk

    andq Qknn

    with the samecodebook, we have

    D (, q ) D (, q).Therefore, in the following, we will restrict ourselves to nearest neighborquantizers.

    A complementary result (see Lemma 2 in Linder [17]) is that for a quantizerq with codebook Cand partition S , a quantizer q with the same partitionbut with a codebook dened by

    yi arg minyHE [ X y |X S i] , i = 1 , . . . , k ,

    satisesD (, q ) D (, q).

    From the two previous optimality results, on the codebook and associatedpartition, we can derive a simple algorithm in order to nd a good quantizer.This algorithm is called the Lloyd algorithm and based on the so-called Lloyditeration (Gersho and Gray [11], Chapter 6). The outline is as follows:

    1. Choose randomly an initial codebook;

    4

  • 8/8/2019 Quantization Clustering Banach

    5/26

    2. Given a codebook C m , build the associated Voronoi partition;

    3. Build C m +1 , the optimal codebook for the previous partition;

    4. Stop when the distortion no longer decreases.

    Unfortunately, this algorithm has two drawbacks: it depends on the initialcodebook chosen, and it does not necessarily converge to the optimal distor-tion. In Section 4 we will discuss an alternative to this algorithm, leadingto an optimal quantizer.

    2.2 Existence of an optimal quantizer

    The aim of this section is to show that the minimization problem of D (, q)has at least one solution. Recall that we consider only nearest neighbor quan-tizers, which can be entirely characterized by their codebook ( y1, . . . , y k),and set y k = ( y1, . . . , yk ).

    We denote byD (, q) = D (, y k )

    the associated distortion. Therefore our rst task is to prove that the func-tion D (, .) has at least one minimum, or, in other words, that there existsat least one optimal codebook.

    Theorem 2.1 Assume that His a reexive and separable Banach space.Then, the function D(, .) admits at least one minimum.Theoretically speaking, it is of interest to search for an optimal quantizer.To make the link with clustering, Theorem 2.1 states that there exists atleast one optimal repartition of the space Hin different clusters. The nextstep is to consider the statistical case, in which the distribution of X isunknown.

    3 A consistent estimator3.1 Construction and consistency

    In a statistical context, the distribution of X is unknown and we onlyhave at hand n random variables, X 1, . . . , X n , independent and distributedas X . Let the empirical measure n be dened as

    n (A) =1n

    n

    i=1

    1 [X iA],

    5

  • 8/8/2019 Quantization Clustering Banach

    6/26

    for any measurable set AH. For any quantizer q, the associated empiricaldistortion is then given by

    D (n , q) =1n

    n

    i=1

    X i q(X i) .

    An (empirical) quantizer qn = qn (., X 1, . . . , X n ) satisfying

    qn argminqQk

    n

    i=1

    X i q(X i )

    is said to be empirically optimal. In particular, if we set (with a slight abuse

    of notation) D (, qn ) = E [ X qn (X ) |X 1 . . . , X n ] ,we have

    D (n , qn ) = Dk (n ).

    From Theorem 2.1, we know that for every n, an empirically optimal quan-tizer always exists.

    The following theorem, which is an adaptation of Theorem 2 in Linder [17],establishes the asymptotic optimality of the quantizer qn with respect to thedistortion.

    Theorem 3.1 Assume that His a reexive and separable Banach space,and set k 1. Then, any sequence of empirically optimal k-quantizers(qn )n 1 satises

    limn

    D (, qn ) = Dk () a.s.

    and lim

    n D (n , qn ) = D

    k () a.s.

    3.2 Rate of convergence

    Most results in the literature concern the situation when H= R d and thedistortion is a L2-based one (Pollard [20], Linder, Lugosi, and Zeger [18],Linder [17]). For example, it is shown in [17] that if there exists T > 0 suchthat P [ X T ] = 1, then

    E D (, qn ) D() CT 2 k(d + 1) ln( k(d + 1))n ,where C > 0 is a universal constant.

    6

  • 8/8/2019 Quantization Clustering Banach

    7/26

    Recently, Biau, Devroye, and Lugosi [2] proved that when His an Hilbertspace, and the distortion is a L2-based one, then

    E D(, qn ) D() C kn ,

    where C > 0 is a universal constant.

    In the sequel, our goal is to establish a rate of convergence in a Banach spaceand with a L1-criterion. This will require some new notions.

    Let P (H) be the set of all probability measures on H.Denition 3.1 Let p

    [1, [.

    1. The L p-Wasserstein distance between , P (H) is dened by:

    p(, ) = inf X,Y

    E d X, Y p1

    p .

    2. A probability P (H) satises a transportation inequality T p() if there exists > 0 such that, for any probability P (H),

    p p(, )

    2

    H (

    |),

    where H (|) = H dd log dd d is the Kullback information be-tween and .

    Remarks:

    The L p-Wasserstein distance, also called L p-Kantorovich distance, isknown to be appropriate for the quantization problem (Graf and Luschgy,Section 3 [12]);

    For this choice of distance, in view of getting rates of convergence,the so-called transportation inequalities, or Talagrand inequalities, arewell designed (Ledoux [16]).

    Generally speaking, it is a difficult task to determine whether a probabilityP (H) satises a transportation inequality T p(). However, the problemis simpler when p = 1, as expressed in the theorem below proven in Djellout,Guillin, and Wu [7] (Theorem 2.3 and Section 1).

    7

  • 8/8/2019 Quantization Clustering Banach

    8/26

    Theorem 3.2 A probability P (H) satises a transportation inequality T 1() if and only if, for all < / 2,

    H e x y 2 d(x) < for one (and therefore for all) y in H.In the sequel, we will only consider the case p = 1, and we set = 1. Forany set H, let P () be the set of all probability measures on . Letalso N (r, ) be the smallest number of balls of radius r (for the metric )required to cover P (), that is

    N (r, )

    = inf nN s.t. x1, . . . , x n P () : ni=1 BP () (x i , r )P () ,

    where BP () (x i , r ) is the ball in P () centered at x i and with radius r (forthe metric ). The quantity ln( N (r, )) is the entropy of P () (Van derVaart and Wellner [23]).In the same way, let N (r, ) be the smallest number of balls of radius r/ 2required to cover , with respect to the metric of H.In order to state a rate of convergence for D(n ), we introduce the following

    assumptions:H1 : There exists > 0 such that satises a transportation inequalityT 1();

    H2 : Any closed bounded ball BHis totally bounded. That is, forall r > 0, N (r, B ) is nite.

    Note that H1 is satised for paths of stochastic differential equations

    dX t = b(X t )dt + s(X t )dW t ,

    where t

    [0, T ], T < , and b(.), s(.) satisfy suitable properties (Djellout,Guillin and Wu [7], Corollary 4.1). H2 is satised, for example, if His aSobolev space on a compact domain of R d (Cucker and Smale [6], example 3).From now on, BR stands for the ball of center 0 and radius R in H. Ac-cording to assumption H2 and Theorem A.1 in Bolley, Guillin, and Villani[4], there exists a positive constant C such that for all r, R > 0,

    N (r, B R ) CR

    r

    N (r/ 2,R )

    . (3.1)

    8

  • 8/8/2019 Quantization Clustering Banach

    9/26

    Theorem 3.3 Assume that His a reexive and separable Banach space,and H1 , H2 are satised. Then, for all < and > 0, there exist threepositive constants K , , and R1 such that if R = R1 max 1, 2, ln 1/ 2

    1/ 2

    and n K ln ( N (,B R )) / 2, we have:P [(, n ) ] e ( / 2)n

    2

    .

    Using the inequality

    D (, qn ) D() 2(, n ),we deduce the following corollary.

    Corollary 3.1 Assume that

    His a reexive and separable Banach space,

    and H1 , H2 are satised. Then, for all < and > 0, there exist threepositive constants K , , and R1 such that if R = R1 max 1, 2, ln 1/ 2

    1/ 2

    and n K ln ( N (,B R )) / 2, we have:P [D (, qn ) D (, q) ] e ( / 8)n

    2

    .

    Let Rbe the function from R + to R + dened by

    R(x) = R1 max 1, x2, ln 1/x 21/ 2 ,

    and denote Mthe function from R + to R + dened by

    M(x) = K ln N (x,B R (x)) /x2

    . (3.2)Theorem 3.4 below gives us the desired rate of convergence.

    Theorem 3.4 Assume that His a reexive and separable Banach space,H1 , H2 are satised, and Mis invertible on some interval ]0, a ]. Then,there exists C 0 > 0 such that

    E D (, qn ) D (, q) C 0 max( M 1(n), n 1/ 2).Note there is no restriction on the support of . In particular, we do notrequire that the support of is bounded. This is an important point, sincesuch an assumption is not veried, for example, by the distributions of clas-

    sical diffusion processes, yet widely used in stochastic modeling.

    Example: Suppose that assumptions H1 and H2 are satised. Considerthe example 3 in Cucker and Smale [6], in which His a Sobolev space on acompact domain set of R d . Using the entropy of the balls BR H(Cuckerand Smale [6]) and Theorem 3.4, we have

    E D (, qn ) D (, q) C

    (ln n)s/d,

    where C is a positive constant.

    9

  • 8/8/2019 Quantization Clustering Banach

    10/26

    3.3 Algorithm

    Calculating qn appears to be a NP -complete problem. In order to approxi-mate qn one can adapt the Lloyd Lloyd algorithm, which has been presentedin Section 2, to the statistical context in which we use n instead of . More-over, rather to calculate empirical medians in each cell, a possible solution isto consider medoids, i.e., centers taken within the sample {X 1, . . . , X n}. Formore details about the Lloyd algorithm and medoids, we refer the reader tothe book by Kaufman and Rousseeuw [14].

    However, this Lloyd algorithm with medoids has the same drawbacks as theLloyd algorithm presented in section 2: non optimality and dependence oninitial codebook. Thus, in the next section, we will present a new estimator,in order to overcome these drawbacks.

    4 Minimization on data

    4.1 Construction and Consistency

    The basic idea of the estimator presented in this section consists in search-ing the minimum of the empirical distortion D (n , .) within the sample

    {X 1, . . . , X n}. It is a generalization of a method of Cadre [5], who consid-ered the case k = 1 only. Formally, our estimator yk,n = ( y1,n , . . . , yk,n ) isdened by

    yk,n arg minz{X 1 ,...,X n }k

    D (n , z).

    Note . k a norm on Hk (as an example, for z = ( z1, . . . , z k )Hk , z k =maxi=1 ,...,k

    zi ), and BH k (z , r ) the associated closed ball in Hk centered at zand with radius r .

    Theorem 4.1 Assume that His a reexive and separable Banach space,and there exists yk an optimal codebook for , which satises

    > 0,P [(X 1, . . . , X k )BH k (y

    k , )] > 0. (4.1)

    Then,lim

    n D (, yk,n ) = D

    k () a.s.

    Remark: The condition (4.1) in Theorem 4.1 simply requires that the prob-ability that k observations fall in the neighborhood of yk is not zero. Thenecessity of this condition is easy to understand. Indeed, suppose there exists > 0 such that for all optimal codebook yk for , (X 1, . . . , X k ) /BH k (y

    k , )

    with probability 1. Then, by construction, D (, yk,n ) can not converge toDk().

    10

  • 8/8/2019 Quantization Clustering Banach

    11/26

    Theorem 4.2 Assume that His a reexive and separable Banach space,and (4.1) , H1 and H2 hold. Then, we have

    limn

    E D (, yk,n ) = Dk ().

    4.2 Rate of convergence

    The next theorem states that D (n , yn,k ) converges to Dk () at the same

    rate as Dk (n ). Remember that the function Mis dened in (3.2), and letyk be an optimal codebook for . For > 0 we set

    f (yk , ) = P (X 1, . . . , X k )BH k (yk , ) .

    We also introduce the assumption:H3 : There exist a decreasing function V : N R + and positiveconstants u,v,C such that

    max u

    0(1 f (yk , )) n/k d,

    +

    v(1 f (yk , )) n/k d V (n).

    Theorem 4.3 Assume that His a reexive and separable Banach space,and H1 and H2 are satised. Let yk be an optimal codebook for satisfying H3 . Then, if Mis invertible on some interval ]0, b] there exists a positiveconstant C 0 such that, for n large enough,

    E D , yk,n Dk() C 0 max( M 1(n), V (n), n/k 1/ 2).Remarks:

    Assumption H3 is a necessary one. Indeed, we will see in the proof of Theorem 4.3 that if H3 is not checked, there exists no decreasingfunction V 1 : N R + such that

    E D , yk,n Dk() V 1(n).

    Assumption H3 is satised if the following assumptions hold:H4 : There exists c1 > 0 such that f (yk , )

    1

    exp(

    2) for

    ]0, c1];H5 : There exists c2 > 0 such that f (yk , ) 1 exp(2) for[c2, + [.

    Assume that H4 and H5 are satised. Then we haveE D , yk,n Dk () C 0 max( M 1(n), n/k 1/ 2).

    That is, D(n , yn,k ) converges to Dk () at the same rate as D

    k(n ).

    Assumption H5 is satised if has a bounded support.

    11

  • 8/8/2019 Quantization Clustering Banach

    12/26

    4.3 Algorithms

    In order to calculate yk,n , we provide an algorithm which we will call Alteralgorithm. The outline is the following:

    1. List all possible codebooks, i.e., all possible k-tuple of data;

    2. Calculate the empirical distortion associated to the rst codebook;

    3. For each successive codebook, calculate the associated empirical dis-tortion. Each time a codebook has an associated empirical distortionsmaller than the previous, store the codebook;

    4. Return the codebook which has the smallest distortion.

    This algorithm overcomes the two drawbacks of the Lloyd algorithm: it doesnot depend on initial conditions and it converges to the optimal distortion.Unfortunately its complexity is o(nk ) and it is impossible to use it for highvalues of n or k.

    In order to overcome this complexity problem, we dene the Alter-fast iter-ation, working as follows:

    1. Select randomly n1 < n data in the whole data set ( n1 should besmall);

    2. Run the Alter algorithm on these n1 data (empirical distortions shouldbe calculated using the whole data set);

    3. Store the obtained codebook.

    Then we derive an accelerated version of the Alter algorithm, which we callAlter-fast algorithm. The outline is the following:

    1. Run n2 times the Alter-fast iteration ( n2 should be high);

    2. Select, among all the obtained codebooks, the one which minimizesthe associated empirical distortion (calculated using the whole data

    set).The Alter-fast algorithm provide a usable alternative for the Alter algorithm,in the same way as the Lloyd algorithm using medoids was an alternative tothe Lloyd algorithm. Its complexity is o(n2 nk1 ). We will see in the nextsection that the Alter-fast algorithm seems to perform almost as well as theAlter algorithm on real-life data.

    12

  • 8/8/2019 Quantization Clustering Banach

    13/26

    5 Application: speech recognition

    Here we use a part of the TIMIT database (http://www-stat.stanford.edu/tibs/ElemStatLearn/) . The data are log-periodograms corresponding torecording phonemes of 32 ms duration. We are interested in the discrimi-nation of ve speech frames corresponding to ve phonemes transcribed asfollows: sh as in she (872 items), dcl as in dark (757 items), iyas the vowel in she (1163 items), aa as the vowel in dark (695 items)and ao as the rst vowel in water (1022 items). The database is a multispeaker database. Each speaker is recorded at a 6 kHz sampling rate andwe retain only the rst 256 frequencies (see Figure 1).

    Figure 1: A sample of log-periodograms for ves phonemes.

    Thus the data consist of 4509 series of length 256. We compare here theLloyd and Alter-fast algorithms. We split the data into a learning and atesting set. The quantizer is constructed using only the rst set and its per-formance (i.e., the rate of good classication) is evaluated from the secondone. We give the rates of good classication associated to the codebooksselected by the Lloyd and and Alter-fast algorithms in Table 1. Recall that,for each center, a cluster includes the data which are closer to this centerthan to any other. Moreover we give the variance induced by the dependence

    13

  • 8/8/2019 Quantization Clustering Banach

    14/26

    on initial conditions: the initial codebook for the Lloyd algorithm, and the

    successive reduced data set for the Alter-fast algorithm. We note that the re-sults of the Alter-fast algorithm are better than those of the Lloyd algorithm.

    Algorithm Rate of good classicationLloyd 0.80 (var=0.0047)

    Alter-fast 0.84 (var=0.00014)

    Table 1: Rate of good classication with the ve phonemes.

    The phonemes ao and aa appear to be particularly difficult to classify.To illustrate this phenomenon, we confront the Lloyd, Alter, and Alter-fastalgorithms on these two phonemes only. The rates of good classication aregiven in Table 2 (note that we gave no variance for the Alter algorithm,since it does not depends on any initial condition). As expected, the resultsare not satisfactory. We note however that the Alter algorithm results aremore reliable than the Lloyd algorithm ones, and that the rates of goodclassication obtained from the Alter and Alter-fast algorithms are almostequivalent. We also note that we improve over the results of Bleakley [3](Chapter 2), who is using different SVM algorithms in a supervised learningcontext.

    Algorithm Rate of good classicationLloyd 0.64 (var=0.0031)Alter 0.71

    Alter-fast 0.68 (var=0.00015)Max. bin. kernel [3] 0.61Min. bin. kernel [3] 0.63

    Table 2: Rate of good classication of phonemes aa and ao.

    Finally, we provide a similar study by removing the phonemes aofromthe database (see Table 3). The results are signicantly better than thoseobtained with the whole database.

    Algorithm Rate of good classicationLloyd 0.87 (var=0.0032)

    Alter-fast 0.90 (var=0.0001)

    Table 3: Rate of good classication without the phoneme ao.

    14

  • 8/8/2019 Quantization Clustering Banach

    15/26

    6 Conclusion

    This paper thus provided an answer to the problem of functional L1-clustering: we rst proved that for any measure P (H) with nitemoment, an optimal quantization always exists (Theorem 2.1). Then weproposed a consistent estimator of q (Theorem 3.1), and we state its rateof convergence (Theorem 3.4). In order to offset the main drawbacks of theLloyd algorithm, we then proposed the Alter algorithm and its acceleratedversion, the Alter-fast algorithm. Finally, a confrontation of our algorithmson real-life data states the practical suitability of our theoretical results.

    One of the most interesting points in our results is that the assumptions

    we make are as light as possible. For example, we made no restriction onthe support of , and the assumptions H1 , H2 are satised in classicalstochastic modeling.

    15

  • 8/8/2019 Quantization Clustering Banach

    16/26

    A Appendix: Proofs

    A.1 Proof of Theorem 2.1

    Before we prove Theorem 2.1, we will need to introduce the following de-nition.

    Denition A.1 A function : HR is called lower semi-continuous for the weak topology (abbreviated weakly l.s.c.) if it satises one of the following equivalent conditions:

    (i) tR , {uH: (u) t}is closed for the weak topology.

    (ii)

    u

    H, lim inf

    uwu

    (u)

    (u) (where w

    note the weak convergence in

    H).For a proof of this equivalence and of the following proposition, we refer thereader to the book by Ekeland and Temam [10].

    Proposition A.1 With the notation of Denition A.1, the two following properties hold:

    (i) If is continuous and convex, then it is weakly l.s.c.

    (ii) If is weakly l.s.c. on a set which is compact for the weak topology,then has a minimum on .

    Lemma A.1 is a straightforward adaptation of the results proven in the rstpart of the proof of Theorem 1 in Linder [17].

    Lemma A.1 There exists A > 0 and k such that inf

    y kHkD(, y k ) = inf

    y B AD (, y ).

    For all x in H, we dene the functions gi,x : Hk R and gx : Hk R by:gi,x (y k ) = x

    yi ,

    andgx (y k) = min

    i=1 ,...,kgi,x (y k ).

    Lemma A.2 For any x in H, the function gx is weakly l.s.c. on Hk .Proof of Lemma A.2 For each x in H, the functions gi,x are continuousand convex, thus they are weakly l.s.c. according to Proposition A.1. Forall t in R , the sets

    y kHk : gi,x (y k ) t

    16

  • 8/8/2019 Quantization Clustering Banach

    17/26

    are then weakly closed. We deduce that

    y kHk : gx (y k ) t =k

    i=1

    y kHk : gi,x (y k) t

    is weakly closed. Lemma A.2 follows by using statement ( i) in DenitionA.1.

    Lemma A.3 The function D(, .) is weakly l.s.c. on Hk .Proof of Lemma A.3 For each ykHk , we can write:

    liminf y k wy k D(, y k ) = liminf y k wy k H gx (y k )(dx) H liminf y k wy k gx (y k )(dx)(by Fatous Lemma)

    H gx (yk )(dx)(by Lemma A.2 and statement ( ii ) in Denition A.1)= D(, yk ),

    which proves that D(, .) satises the condition ( ii ) of Denition A.1.

    We are now in a position to prove Theorem 2.1.

    Proof of Theorem 2.1 According to Lemma A.1, there exists R > 0 suchthat the inmum of D (, .) on Hk is also the inmum of D (, .) on B kR .Moreover, on the one hand B kR is compact for the weak topology, and on theother hand D(, .) is weakly l.s.c. according to Lemma A.3. Thus, accordingto Proposition A.1, the function D (, .) reaches its inmum on B kR .

    A.2 Proof of Theorem 3.3

    The proof is adapted from the proof of Theorem 1 by Bolley, Guillin, and

    Villani [4]. It can be decomposed in three steps:1. First, we show we can consider truncated version of the probability

    measures and n on the ball BR ;

    2. Then we cover the space P (BR ) by small balls of radius r ;3. Finally, we optimize the various parameters introduced in the proof.

    Each of the next three lemmas matches a step.

    17

  • 8/8/2019 Quantization Clustering Banach

    18/26

    Let R > 0. We consider R dened, for all Borel set AH, by:R [A] =

    [A BR ][BR ]

    = [A|BR ].Consider now the independent random variables {X i}ni=1 with distribution and {Y i}ni=1 with distribution R . We dene, for i n,

    X Ri =X i if X i RY i if X i > R.

    Let x be the Dirac measure at point x. The empirical measures n and Rnare dened by

    n = 1n

    n

    i=1X i and Rn = 1n

    n

    i=1X Ri .

    Note E = H exp( x2)(dx). Since we suppose that satises a T 1()-

    inequality, we have, for < / 2, E < .Lemma A.4 Let ]0, 1[, , > 0, 1]0, / 2[, and ] 1, / 2[. Then, for all R > max 1/ 2, 2/ 1 , we have

    P [ (n , ) > ] P R , Rn > 2E Re R2

    +exp

    n (1

    )

    E e( 1 )R

    2

    .

    Proof of Lemma A.4 For a xed > 0, we bound P [(, n ) > ] in func-tion of R and Rn . First, following the arguments of the proof of Theorem1.1 by Bolley, Guillin, and Villani (step 1) [4], it can be proven that for all < / 2 and R 1/ 2 ,

    (, R ) 2E Re R2

    . (A.1)

    Second, the probability measures n and Rn satisfy

    (n , Rn ) 1n

    n

    i=1

    X Ri X i 1n

    n

    i=1

    Z i ,

    where Z i = 2 X i 1 X i >R (i = 1 , . . . , n ). Using a similar argument as in theproof of Theorem 1.1 by Bolley, Guillin, and Villani (step 1) [4], we deducethat if , are positive and < / 2,

    P (n , Rn ) > exp n E e( 1 )R2

    . (A.2)

    The conclusion follows from (A.1), (A.2), and the triangular inequality for.

    18

  • 8/8/2019 Quantization Clustering Banach

    19/26

    Lemma A.5 Given ,, 1, 1 > 0 such that 1 < , ] 1, / 2[, and

    > 1, there exist positive constants 1, 2 < 1, K 1 and K 2 such that, for all R > max 1/ 2, 2/ 1 and > 0,

    P [(, n ) > ] N (1/ 2, B R )exp n22

    2 K 1R2e R2

    +exp n K 2 K 3e( 1 )R2

    ,

    where K 3 is a positive constant depending only on and 1.

    Proof of Lemma A.5 We start by proving that R satises a modiedT 1()-inequality. Let be a Borel set of

    P (BR ). Following the arguments

    of the proof of Theorem 1.1 of Bolley, Guillin, and Villani (step 2) [4], onemay write

    P [Rn ] exp n inf H ( |R ) . (A.3)From now on, we consider that P (BR ) is equipped with the distance . Con-sider > 0 and A a measurable subset of P (BR ). We set N A = N (/ 2, A).Then there exist N A balls B i , i = 1 , . . . , N A , covering A. Each of this ballsis convex and included in the -neighborhood A of A. Moreover, by as-sumption H2 , the balls B i are totally bounded.

    It is easily inferred from equation (A.3) thatP [Rn A] N A exp n inf A H ( |

    R ) . (A.4)

    Dene now

    A = P (BR ) : (, R ) 2E Re R2

    .

    According to the basic inequality

    a]0, 1[,C > 0 such that x, yR , (x y)2 (1 a)x2 Cy2, (A.5)

    we have, for any H,

    1 < , K > 0 such that H ( |R ) 12

    2(R , ) KR 2e R2

    .

    Thus, we can write

    A, H ( |R ) 12

    2(R , ) KR 2e R2

    12

    m2 KR 2e R2

    ,

    wherem = max 2E Re R

    2

    , 0 .

    19

  • 8/8/2019 Quantization Clustering Banach

    20/26

    From this and equation (A.4) we conclude that

    P (R , Rn ) 2E Re R2

    N A exp n12

    m2 KR 2e R2

    .

    (A.6)Now, given 2 < 1, it follows from (A.5) that there exist three positiveconstants 1, 1 and K 1 depending only on , 1, and 2 such that

    12

    m2 KR 2e R2

    22

    2 K 1R2e R2

    ,

    where = 1. This leads, together with (A.6), to

    P (R , Rn

    )

    2E Re R

    2

    N A exp

    n

    2

    22

    K 1R2e R

    2

    .

    (A.7)To bound N A , we observe that since AP (BR ),

    N A N (/ 2, B R ) = N (1/ 2, B R ).The conclusion follows by Lemma A.4 and inequality (A.7).

    The following lemma simplies the results of the previous.

    Lemma A.6 Let < , < / 2, and < . There exists 1 > 0 such that, for all > 0,

    P [(, n ) > ] exp 2 n 2 + exp n 2 ,as soon as

    R2 R2 max 1, 2, ln12

    and n K 4ln ( N (1/ 2, B R ))

    2,

    where R2 and K 4 are some positive constants depending on through and .

    Proof of Lemma A.6 On the one hand, under the assumptions and no-tation of Lemma A.5, we have, for all < 2,

    ln N (1/ 2, B R )exp n22

    2 K 1R2e R2

    n 2

    2(A.8)

    as soon as R, R/ ln(1/ 2) and n 2/ ln( N (1/ 2, BR )) are large enough (seethe third step of the proof of Theorem 1.1 by Bolley, Guillin, and Villani [4]).

    On the other hand, let < 2 < 1. We can choose such that K 2 = 2.With this choice we obtain

    exp n K 2 K 3e( 1 )R2

    = exp n 22 K 3e( 1 )R2

    ,

    20

  • 8/8/2019 Quantization Clustering Banach

    21/26

    which can be bounded by exp n 2 , for R and R2/ ln(1 / 2) large enough.This, together with (A.8), leads to the conclusion.

    Theorem 3.3 is then a straightforward consequence of Lemma A.6, noticingthat, for any K < min(( / 2), ) and n large enough, we have

    exp 2

    n 2 + exp n 2 exp Kn 2 .

    A.3 Proof of Theorem 3.4

    Let > 0 be small enough. According to Corollary 3.1 we have

    P [D (, qn ) D (, q) > ] e ( / 8)n2 ,

    as soon as n M(). Therefore we can write:E D (, qn ) D (, q) =

    +

    0P [D (, qn ) D(, q) > ]d

    = M 1 (n )

    0P [D(, qn ) D (, q) > ]d

    + +

    M 1 (n )P [D (, qn ) D (, q) > ]d

    M 1(n) + + 0 e ( / 8)n2

    d

    C 0 max( M 1(n), n 1/ 2),as desired.

    A.4 Proof of Theorem 4.1

    One can easily show that

    D (n , y

    k,n ) D (, y

    k,n ) (, n ). (A.9)Thus, by Lemma 4 in Linder [17] and Varadarajans Theorem [8], we deducethat:

    D(n , yk,n ) D(, yk,n ) 0 a.s. as n . (A.10)Let p n and z {X 1, . . . , X p}k . Since D (n , yk,n ) D (n , z ) and, by thelaw of large number, D(n , z) D (, z) a.s., we have

    lim supn

    D (n , yk,n ) D (, z ) a.s.

    21

  • 8/8/2019 Quantization Clustering Banach

    22/26

    From (A.10), we deduce that, for all p 1,limsup

    nD(, yk,n ) minz{X 1 ,...,X p }k D (, z). (A.11)

    Let us now evaluate the limit of the right-hand term in the equation (A.11)as p . Note, for > 0 and p 1,

    N ( p,) = z

    argminz{X 1 ,...,X p }k

    D (, z) BH k (yk , ),

    D(, z) D (, yk ) + 2 .

    Since,

    y k , y kHk , |D (, y k ) D (, y k )| y k y k k , we obtainN ( p,) [D (, y

    k ) D(, yk ) + ] =.

    Therefore as soon as p k,

    P minz{X 1 ,...,X p }k

    D (, z ) D (, yk ) > 2

    P N ( p,) + P z {X 1, . . . , X p}k , zBH k (yk , )

    P (X 1, . . . , X

    k)

    BH

    k (yk, )

    p/k

    = 1 P (X 1, . . . , X k )BH k (yk , ) p/k

    , (A.12)

    where . stands for the integer part function. Then, by the Borel-Cantellilemma,

    lim p

    minz{X 1 ,...,X p }k

    D (, z ) = D (, yk) a.s.

    This result, together with (A.11), leads to the conclusion.

    A.5 Proof of Theorem 4.2

    On the one hand we can write:

    D (, yk,n ) D()= D(, yk,n ) D (n , yk,n ) + D (n , yk,n ) D() |D(, yk,n ) D (n , yk,n )|+ |D(n , yk,n ) D()|(, n ) + |D(n , yk,n ) D()|,

    according to (A.9).

    22

  • 8/8/2019 Quantization Clustering Banach

    23/26

    On the other hand,

    limn

    D(n , yk,n ) = D() a.s.

    Moreover,

    D (n , yk,n ) = minz{X 1 ,...,X n }k

    1n

    n

    i=1

    min j =1 ,...,k

    X i z j

    1n

    n

    i=1

    X i X 1

    1

    n

    n

    i=1

    X i

    + X 1

    .

    Hence, D (n , yk,n ) is equi-integrable, which proves that it converges in L1.

    Finally, E (, n ) 0 by Theorem 3.3, and we deduce the proof of Theorem4.2.A.6 Proof of Theorem 4.3

    First we can write

    D(,y

    k,n ) D

    k () = D (,y

    k,n ) D(n ,y

    k,n )+ D (n , yk,n ) minz{X 1 ,...,X n }k D (, z)+ min

    z{X 1 ,...,X n }kD (, z) Dk().

    Then, according to Lemma 3 in Linder [17], we have

    D(, yk,n ) D (n , yk,n ) (, n )and

    D (n , yk,n ) minz{X 1 ,...,X n }k D (, z) (, n ).Thus,

    D (, yk,n ) Dk() 2(, n ) + minz{X 1 ,...,X n }k D (, z ) Dk (). (A.13)

    Moreover, according to the inequality (A.12), we have for n k:P min

    z{X 1 ,...,X n }kD (, z) Dk () 2 1 f (yk , )

    n/k.

    23

  • 8/8/2019 Quantization Clustering Banach

    24/26

    We deduce

    E minz{X 1 ,...,X n }k

    D(, z) Dk ()

    = +

    0P min

    z{X 1 ,...,X n }kD (, z) Dk () d

    2 +

    01 f (yk , )

    n/kd

    2 [0,u ][v, [ 1 f (yk , ) n/k d + v

    u1 f (yk , )

    n/kd

    2 2V (n) +

    v

    u1

    f (yk , )

    n/kd

    (according to assumption H3 )

    2 2V (n) + ( v u) n/k

    C max n/k 1/ 2, V (n) for n large enough ,where < 1 and C are some positive constants. Theorem 4.3 follows from(A.13), Theorem 3.3 and Theorem 3.4.

    24

  • 8/8/2019 Quantization Clustering Banach

    25/26

    References

    [1] E. Abaya and G. Wise. Convergence of vector quantizers with applica-tion to optimal quantization. SIAM Journal on Applied Mathematics ,44:183189, 1984.

    [2] G. Biau, L. Devroye, and G. Lugosi. On the performance of clustering inHilbert spaces. IEEE Transactions on Information Theory , 54:781790,2007.

    [3] K. Bleakley. Quelques Contributions `a lAnalyse Statistique et `a la Clas-sication des Graphes et des Courbes. Applications ` a lImmunobiologieet a la Reconstruction des Reseaux Biologiques . PhD thesis, UniversiteMontpellier II, 2007.

    [4] F. Bolley, A. Guillin, and C. Villani. Quantitative concentration in-equalities for empirical measures on non-compact spaces. Probability Theory and Related Fields , 137(3-4):541593, 2007.

    [5] B. Cadre. Convergent estimators for the L1-median of a Banach valuedrandom variable. Statistics , 35(4):509521, 2001.

    [6] F. Cucker and S. Smale. On the mathematical foundations of learn-ing. American Mathematical Society. Bulletin. New Series , 39(1):149(electronic), 2002.

    [7] H. Djellout, A. Guillin, and L. Wu. Transportation cost-informationinequalities and applications to random dynamical systems and diffu-sions. Annals of Probability , 32(3B):27022732, 2004.

    [8] R. M. Dudley. Real Analysis and Probability , volume 74 of CambridgeStudies in Advanced Mathematics . Cambridge University Press, Cam-bridge, 2002. Revised reprint of the 1989 original.

    [9] N. Dunford and J. T. Schwartz. Linear Operators. Part I . Wiley Clas-sics Library. John Wiley & Sons Inc., New York, 1988. General theory,With the assistance of William G. Bade and Robert G. Bartle, Reprintof the 1958 original, A Wiley-Interscience Publication.

    [10] I. Ekeland and R. Temam. Analyse Convexe et Problemes Variation-nels. Dunod, 1974. Collection Etudes Mathematiques.

    [11] A. Gersho and R. M. Gray. Vector Quantization and Signal Compres-sion . Kluwer Academic Publishers, Norwell, MA, USA, 1991.

    [12] S. Graf and H. Luschgy. Foundations of Quantization for Probability Distributions , volume 1730 of Lecture Notes in Mathematics . Springer-Verlag, Berlin, 2000.

    25

  • 8/8/2019 Quantization Clustering Banach

    26/26

    [13] J. A. Hartigan. Clustering Algorithms . John Wiley & Sons, New York-

    London-Sydney, 1975. Wiley Series in Probability and MathematicalStatistics.

    [14] L. Kaufman and P. J. Rousseeuw. Finding Groups in Data . Wiley Seriesin Probability and Mathematical Statistics: Applied Probability andStatistics. John Wiley & Sons Inc., New York, 1990. An introductionto cluster analysis, A Wiley-Interscience Publication.

    [15] J. H. B. Kemperman. The median of a nite measure on a Banach space.In Statistical data analysis based on the L1-norm and related methods(Neuch atel, 1987) , pages 217230. North-Holland, Amsterdam, 1987.

    [16] M. Ledoux. The Concentration of Measure Phenomenon , volume 89 of Mathematical Surveys and Monographs . American Mathematical Soci-ety, 2001.

    [17] T. Linder. Learning-theoretic methods in vector quantization. In Prin-ciples of nonparametric learning (Udine, 2001) , volume 434 of CISM Courses and Lectures , pages 163210. Springer, Vienna, 2002.

    [18] T. Linder, G. Lugosi, and K. Zeger. Rates of convergence in the sourcecoding theorem, in empirical quantizer design, and in universal lossysource coding. IEEE Transactions on Information Theory , 40:17281740, 1994.

    [19] D. Pollard. Strong consistency of k-means clustering. The Annals of Statistics , 9:135140, 1981.

    [20] D. Pollard. A central limit theorem for k-means clustering. The Annalsof Probability , 10:919926, 1982.

    [21] D. Pollard. Quantization and the method of k-means. IEEE Transac-tions on Information Theory , 28:199205, 1982.

    [22] J. O. Ramsay and B. W. Silverman. Functional Data Analysis . SpringerSeries in Statistics. Springer, New York, second edition, 2005.

    [23] A. W. Van der Vaart and J. A. Wellner. Weak Convergence and Empir-ical Processes . Springer Series in Statistics. Springer-Verlag, New York,1996. With applications to statistics.

    26