SlideShare una empresa de Scribd logo
1 de 40
Descargar para leer sin conexión
A Tutorial on Spectral Clustering


                                                    Chris Ding
                                   Computational Research Division
                                Lawrence Berkeley National Laboratory
                                       University of California

                     Supported by Office of Science, U.S. Dept. of Energy




Tutorial on Spectral Clustering, ICML 2004, Chris Ding © University of California   1
Some historical notes
         • Fiedler, 1973, 1975, graph Laplacian matrix
         • Donath & Hoffman, 1973, bounds
         • Pothen, Simon, Liou, 1990, Spectral graph
           partitioning (many related papers there after)
         • Hagen & Kahng, 1992, Ratio-cut
         • Chan, Schlag & Zien, multi-way Ratio-cut
         • Chung, 1997, Spectral graph theory book
         • Shi & Malik, 2000, Normalized Cut


Tutorial on Spectral Clustering, ICML 2004, Chris Ding © University of California   2
Spectral Gold-Rush of 2001
                                   9 papers on spectral clustering

     • Meila & Shi, AI-Stat 2001. Random Walk interpreation of
           Normalized Cut
     • Ding, He & Zha, KDD 2001. Perturbation analysis of Laplacian
                    matrix on sparsely connected graphs
     • Ng, Jordan & Weiss, NIPS 2001, K-means algorithm on the
           embeded eigen-space
     • Belkin & Niyogi, NIPS 2001. Spectral Embedding
     • Dhillon, KDD 2001, Bipartite graph clustering
     • Zha et al, CIKM 2001, Bipartite graph clustering
     • Zha et al, NIPS 2001. Spectral Relaxation of K-means
     • Ding et al, ICDM 2001. MinMaxCut, Uniqueness of relaxation.
     • Gu et al, K-way Relaxation of NormCut and MinMaxCut
Tutorial on Spectral Clustering, ICML 2004, Chris Ding © University of California   3
Part I: Basic Theory, 1973 – 2001




Tutorial on Spectral Clustering, ICML 2004, Chris Ding © University of California   4
Spectral Graph Partitioning

       MinCut: min cutsize
                                                              Constraint on sizes: |A| = |B|
       cutsize = # of cut edges




Tutorial on Spectral Clustering, ICML 2004, Chris Ding © University of California              5
2-way Spectral Graph Partitioning
                                                                               1 if i ∈ A
     Partition membership indicator:                                     qi = 
                                                                              − 1 if i ∈ B

                                       ∑
                                   1
      J = CutSize =                                 wij [qi − q j ]2
                                   4         i, j


                  ∑                                                         ∑
              1                                                         1
          =                    wij [qi2 + q 2 − 2qi q j ] =
                                            j                                            q [d δ    − wij ]q j
              4         i, j                                            2           i , j i i ij

              1 T
          =     q ( D − W )q
              2
       Relax indicators qi from discrete values to continuous values,
       the solution for min J(q) is given by the eigenvectors of

                                       ( D − W ) q = λq                                 (Fiedler, 1973, 1975)
                                                                                     (Pothen, Simon, Liou, 1990)
Tutorial on Spectral Clustering, ICML 2004, Chris Ding © University of California                                  6
Properties of Graph Laplacian

           Laplacian matrix of the Graph: L = D − W

         • L is semi-positive definite xT Lx ≥ 0 for any x.
         • First eigenvector is q1=(1,…,1)T = eT with λ1=0.
         • Second eigenvector q2 is the desired solution.
               • The smaller λ2, the better quality of the
               partitioning. Perturbation analysis gives
                           cutsize cutsize
                                  λ2 =
                                   +
                             | A|      |B|
          • Higher eigenvectors are also useful
Tutorial on Spectral Clustering, ICML 2004, Chris Ding © University of California   7
Recovering Partitions

          From the definition of cluster indicators:
          Partitions A, B are determined by:

              A = {i | q2 (i ) < 0}, B = {i | q2 (i ) ≥ 0}

          However, the objective function J(q) is
          insensitive to additive constant c :

                                               ∑
                            1
             J = CutSize =            w [( qi + c) − (q j + c)]2
                            4    i , j ij

           Thus, we sort q2 to increasing order, and cut in the
           middle point.
Tutorial on Spectral Clustering, ICML 2004, Chris Ding © University of California   8
Multi-way Graph Partitioning

         • Recursively applying the 2-way partitioning
            • Recursive 2-way partitioning
            • Using Kernigan-Lin to do local refinements
         • Using higher eigenvectors
            • Using q3 to further partitioning those obtained
            via q2.

         • Popular graph partitioning packages
            • Metis, Univ of Minnesota
               • Chaco, Sandia Nat’l Lab

Tutorial on Spectral Clustering, ICML 2004, Chris Ding © University of California   9
2-way Spectral Clustering

               • Undirected graphs (pairwise similarities)
               • Bipartite graphs (contingency tables)
               • Directed graphs (web graphs)




Tutorial on Spectral Clustering, ICML 2004, Chris Ding © University of California   10
Spectral Clustering
         min cutsize , without explicit size constraints
          But where to cut ?




            Need to balance sizes

Tutorial on Spectral Clustering, ICML 2004, Chris Ding © University of California   11
Clustering Objective Functions
                                                                                    s(A,B) =   ∑∑ w          ij

         • Ratio Cut                                                                           i∈A j∈B
                                                         s(A,B) s(A,B)
                              J Rcut (A,B) =                   +
                                                           |A|    |B|
         • Normalized Cut                                                                 dA =    ∑d     i
                                                                                                  i∈A
                                                       s( A, B ) s( A, B )
                            J Ncut ( A, B) =                    +
                                                          dA        dB
                                                               s ( A, B)            s ( A, B )
                                                 =                            +
                                                        s ( A, A) + s ( A, B ) s(B, B ) + s ( A, B )
         • Min-Max-Cut
                                                        s(A,B) s(A,B)
                              J MMC (A,B) =                   +
                                                        s(A,A) s(B,B)

Tutorial on Spectral Clustering, ICML 2004, Chris Ding © University of California                                 12
Ratio Cut                  (Hagen & Kahng, 1992)

        Min similarity between A , B:
                                                                             s(A,B) =   ∑∑ w
                                                                                        i∈ A j∈B
                                                                                                   ij


                                                               s(A,B) s(A,B)
       Size Balance J Rcut (A,B) =                                   +                   (Wei & Cheng, 1989)
                                                                 |A|    |B|

                                                                              
                                                                               n2 / n1n           if i ∈ A
       Cluster membership indicator:                                  q(i ) = 
                                                                              − n1 / n2 n
                                                                                                  if i ∈ B

      Normalization:                     q T q = 1, q T e = 0

        Substitute q leads to J Rcut (q) = q T ( D − W )q
        Now relax q, the by eigenvectornd eigenvector of L
          Solution given solution is 2

Tutorial on Spectral Clustering, ICML 2004, Chris Ding © University of California                             13
Normalized Cut (Shi & Malik, 1997)

           Min similarity between A & B: s(A,B) = ∑                               ∑ wij
                                                                          i∈ A j∈B
      Balance weights                                          s( A, B ) s( A, B )
                                   J Ncut ( A, B) =                     +
                                                                  dA        dB       dA =     ∑d
                                                                                              i∈A
                                                                                                        i


                                                        d B / d Ad
                                                                             if i ∈ A
     Cluster indicator:                        q(i ) = 
                                                       − d A / d B d
                                                                             if i ∈ B   d=   ∑d
                                                                                              i∈G
                                                                                                    i

      Normalization:                         q Dq = 1, q De = 0
                                               T                  T


      Substitute q leads to                         J Ncut (q) = q T ( D − W )q

       min q q T ( D − W )q + λ (q T Dq − 1)
        Solution is eigenvector of ( D − W )q = λDq
Tutorial on Spectral Clustering, ICML 2004, Chris Ding © University of California                           14
MinMaxCut (Ding et al 2001)
              Min similarity between A & B:                              s(A,B) =     ∑∑ w
                                                                                      i∈A j∈B
                                                                                                ij


                 Max similarity within A & B:                            s(A,A) =     ∑∑ w
                                                                                      i∈A j∈A
                                                                                                ij


                                              s(A,B) s(A,B)
                    J MMC(A,B) =                    +
                                              s(A,A) s(B,B)
                                                         
                                                          d B / d Ad               if i ∈ A
      Cluster indicator:                         q(i ) = 
                                                         − d A / d B d
                                                                                   if i ∈ B
      Substituting,
                                  1+ dB / d A                 1+ d A / dB                              q T Wq
              J MMC ( q) =                               +                             −2       Jm =
                                 Jm + dB / d A               Jm + d A / dB                             q T Dq
       Because               dJ MMC ( J m )
                                            <0                   min Jmmc ⇒ max Jm(q)
                                 dJ m
                        ⇒             Wq = ξDq                    ⇒             ( D − W )q = λDq
Tutorial on Spectral Clustering, ICML 2004, Chris Ding © University of California                               15
A simple example
                 2 dense clusters, with sparse connections
                 between them.
               Adjacency matrix                                             Eigenvector q2




Tutorial on Spectral Clustering, ICML 2004, Chris Ding © University of California            16
Comparison of Clustering Objectives
          • If clusters are well separated, all three give
            very similar and accurate results.
          • When clusters are marginally separated,
            NormCut and MinMaxCut give better results
          • When clusters overlap significantly,
            MinMaxCut tend to give more compact and
            balanced clusters.
                                                      s ( A, B)         s ( A, B)
                                J Ncut =                          +
                                               s ( A, A) + s(A, B) s(B, B) + s(A, B)

              Cluster Compactness ⇒                               max s ( A, A)
Tutorial on Spectral Clustering, ICML 2004, Chris Ding © University of California      17
2-way Clustering of Newsgroups

     Newsgroups                        RatioCut NormCut                             MinMaxCut
     Atheism          63.2 ± 16.2 97.2 ± 0.8                                        97.2 ± 1.1
     Comp.graphics
     Baseball         54.9 ± 2.5 74.4 ± 20.4                                        79.5 ± 11.0
     Hockey
     Politics.mideast 53.6 ± 3.1 57.5 ± 0.9                                         83.6 ± 2.5
     Politics.misc

Tutorial on Spectral Clustering, ICML 2004, Chris Ding © University of California                 18
Cluster Balance Analysis I:
                                Random Graph Model
       • Random graph: edges are randomly assigned
         with probability p: 0 ≤ p ≤ 1.
       • RatioCut & NormCut show no size dependence
                                           p | A || B | p | A || B |
                        J Rcut ( A, B) =               +             = np = constant
                                              | A|         |B|
                                      p | A || B |    p | A || B |      n
                  J Ncut ( A, B) =                  +                =      = constant
                                     p | A | (n − 1) p | B | ( n − 1) n − 1


       • MinMaxCut favors balanced clusters: |A|=|B|
                                        p | A || B |        p | A || B |     |B|      | A|
              J MMC ( A, B) =                          +                  =        +
                                     p | A | (| A | −1) p | B | (| B | −1) | A | −1 | B | −1

Tutorial on Spectral Clustering, ICML 2004, Chris Ding © University of California              19
2-way Clustering of Newsgroups
    Cluster Balance

    Eigenvector




      JNcut(i)



      JMMC(i)



Tutorial on Spectral Clustering, ICML 2004, Chris Ding © University of California   20
Cluster Balance Analysis II:
                                                         Large Overlap Case
                                                         s ( A, B )
                                        f =                                 > 0.5
                                              (1 / 2)[ s( A, A) + s( B, B)]

            Conditions for skewed cuts:
                                                                 1 1
                          NormCut : s(A,A) ≥ (                    − ) s ( A, B ) = s ( A, B) / 2
                                                                2f 2
                                                                    1
                           MinMaxCut : s(A,A) ≥                        s ( A, B) = s ( A, B )
                                                                    2f

            Thus MinMaxCut is much less prone to skewed cuts

Tutorial on Spectral Clustering, ICML 2004, Chris Ding © University of California                  21
Spectral Clustering of Bipartite Graphs

            Simultaneous clustering of rows and columns
            of a contingency table (adjacency matrix B )

            Examples of bipartite graphs
            • Information Retrieval: word-by-document matrix
            • Market basket data: transaction-by-item matrix
            • DNA Gene expression profiles
            • Protein vs protein-complex
Tutorial on Spectral Clustering, ICML 2004, Chris Ding © University of California   22
Spectral Clustering of Bipartite Graphs
          Simultaneous clustering of rows and columns
                     (adjacency matrix B )
                                                                                s ( BR1 ,C2 ) =   ∑ ∑b
                                                                                                  ri ∈R1c j ∈C 2
                                                                                                                   ij




                                                            min between-cluster sum of
                                                            xyz weights: s(R1,C2), s(R2,C1)
                                                            max within-cluster sum of xyz
                                  cut                       xyz weights: s(R1,C1), s(R2,C2)

                                          s ( BR1 ,C2 ) + s ( B R2 ,C1 )       s ( BR1 ,C2 ) + s ( B R2 ,C1 )
   J MMC (C1 , C 2 ; R1 , R2 ) =                                           +
                                                  2s ( B R1 ,C1 )                     2s ( BR2 ,C2 )
                                                                                           (Ding, AI-STAT 2003)
Tutorial on Spectral Clustering, ICML 2004, Chris Ding © University of California                                       23
Bipartite Graph Clustering
         Clustering indicators for rows and columns:
                          1 if ri ∈ R1                                     1 if ci ∈ C1
                f (i ) =                                         g (i ) = 
                         − 1 if ri ∈ R2                                   − 1 if ci ∈ C2

                    BR ,C            BR1 ,C2                0             B           f 
                 B= 1 1                                 W = T                       q= 
                                                                                          g
                    BR ,C            BR2 ,C2               B              0            
                    2 1                                                    
     Substitute and obtain
                                                                              s (W12 ) s (W12 )
                                        J MMC (C1 , C 2 ; R1 , R2 ) =                 +
                                                                              s (W11 ) s (W22 )
    f,g are determined by
                          D r           0            B   f    D                f 
                                      −                   = λ  r               
                         
                         
                                    Dc   B T
                                         
                                                            g
                                                         0   
                                                            
                                                                      
                                                                                   Dc  g 
                                                                                        
Tutorial on Spectral Clustering, ICML 2004, Chris Ding © University of California                 24
Clustering of Bipartite Graphs
       Let        ~                          u         D1 / 2 f 
                  B = Dr−1/ 2 BDc 1/ 2 , z =   = Dq =  r / 2 
                                 −
                                              
                                                         D1 g 
                                             v         c        
       We      obtain           ~
                        0 B  u            u 
                       ~           
                        B T 0  v  = λ  v 
                                               
                                           
                                                         m

                                                        ∑u λ v
                                               ~
       Solution is SVD:                        B=                    T
                                                                 k k k
                                                        k =1

                                                                                    (Zha et al, 2001, Dhillon, 2001)




Tutorial on Spectral Clustering, ICML 2004, Chris Ding © University of California                                      25
Clustering of Bipartite Graphs
        Recovering row clusters:
                     R1 = {ri , | f 2 (i) < z r }, R2 = {ri , | f 2 (i) ≥ z r },

         Recovering column clusters:
                     C1 = {ci , | g 2 (i ) < z c }, C 2 = {ci , | g 2 (i ) ≥ z c },

         zr=zc=0 are dividing points. Relaxation is
         invariant up to a constant shift.
         Algorithm: search for optimal points icut, jcut, let
         zr=f2(icut), zc= g2(jcut), such that J MMC (C1 , C2 ; R1 , R2 )
                                                                                    (Zha et al, 2001)
         is minimized.

Tutorial on Spectral Clustering, ICML 2004, Chris Ding © University of California                       26
Clustering of Directed Graphs
          Min directed edge weights between A & B:
                                                                s(A,B)=    ∑∑(w
                                                                           i∈A j∈B
                                                                                       ij   + w ji )

           Max directed edges within A & B:
                                                                  s(A,A)=    ∑∑(w
                                                                             i∈A j∈A
                                                                                            ij   + w ji )


      • Equivalent to deal with W = W + W T
                                                                   ~
                                      ~
      • All spectral methods apply to W
      • For example, web graphs clustered in such
        way
                                                                                (He, Ding, Zha, Simon, ICDM 2001)
Tutorial on Spectral Clustering, ICML 2004, Chris Ding © University of California                                   27
K-way Spectral Clustering
                        K≥2




Tutorial on Spectral Clustering, ICML 2004, Chris Ding © University of California   28
K-way Clustering Objectives

         • Ratio Cut
                                                         s(C k ,Cl )       s (C k ,Cl )             s(C k ,G − C k )
           J Rcut (C1 , , C K ) =           ∑
                                              
                                              
                                              k ,l 
                                                            |C k|
                                                                        +
                                                                               |Cl| 
                                                                                         =
                                                                                         
                                                                                              ∑
                                                                                              k
                                                                                                            |C k|


         • Normalized Cut
                                                         s(C k ,Cl )       s (C k ,Cl )             s(C k ,G − C k )
           J Ncut (C1 , , C K ) =           ∑
                                              
                                              
                                              k ,l 
                                                             dk
                                                                        +
                                                                                dl
                                                                                         =
                                                                                         
                                                                                         
                                                                                              ∑
                                                                                              k
                                                                                                             dk

         • Min-Max-Cut
                                                          s(C k ,Cl )       s (C k ,Cl )             s(C k ,G − C k )
           J MMC (C1 , , C K ) =              ∑  s(C , C ) + s(C , C )  = ∑
                                                 
                                                 
                                               k ,l          k    k
                                                                         
                                                                                 l   l           k
                                                                                                         s (C k , C k )

Tutorial on Spectral Clustering, ICML 2004, Chris Ding © University of California                                         29
K-way Spectral Relaxation

         • Prove that the solution lie in the subspace
           spanned by the first k eigenvectors
         • Ratio Cut
         • Normalized Cut
         • Min-Max-Cut




Tutorial on Spectral Clustering, ICML 2004, Chris Ding © University of California   30
K-way Spectral Relaxation
                                                                h1 = (1m1,0m 0,0 m 0)T
     Unsigned cluster indicators:
                                                                h2 = (0 m 0,1m1,0 m 0)T
                                                               mmm
       Re-write:                                                hk = (0 m 0,0 m 0,1m1)T

                                           h1 ( D − W )h1
                                            T
                                                                          hk ( D − W ) hk
                                                                           T
              J Rcut (h1 , , hk ) =              T
                                                                ++                  T
                                                 h1 h1                              hk hk

                                            h1 ( D − W )h1
                                             T
                                                                           hk ( D − W )hk
                                                                            T
               J Ncut (h1 , , hk ) =             T
                                                                 ++             T
                                                 h1 Dh1                         hk Dhk
                                             h1 ( D − W ) h1
                                              T
                                                                             hk ( D − W )hk
                                                                              T
               J MMC (h1 , , hk ) =               T
                                                                   ++               T
                                                  h1 Wh1                            hk Whk


Tutorial on Spectral Clustering, ICML 2004, Chris Ding © University of California             31
K-way Ratio Cut Spectral Relaxation
                                             -
     Unsigned cluster indicators: x = (0  0,11,0  0)T / n1/ 2
                                                                                    nk

                                   k                        k

      Re-write: J ( x , , x ) = xT ( D − W ) x +  + xT ( D − W ) x
                 Rcut 1     k     1            1       k             k

                            = Tr ( X T ( D − W ) X )                                X = ( x1 , , xk )

         Optimize : min Tr ( X T ( D − W ) X ), subject to X T X = I
                               X

     By K. Fan’s theorem, optimal solution is
     eigenvectors: X=(v1,v2, …, vk), (D-W)vk=λkvk
     and lower-bound
                         λ1 +  + λk ≤ min J Rcut ( x1 , , xk )
                                                                  (Chan, Schlag, Zien, 1994)

Tutorial on Spectral Clustering, ICML 2004, Chris Ding © University of California                        32
K-way Normalized Cut Spectral Relaxation
     Unsigned cluster indicators:
                                                                     -nk

                                            yk = D     1/ 2
                                                              (0 o 0,1o1,0o 0)T / || D1/ 2 hk ||
      Re-write:                                      ~                      ~
                   J Ncut ( y1 , , y k ) = y1 ( I − W ) y1 +  + y k ( I − W ) y k
                                             T                      T

                                     ~                             ~
                   = Tr (Y T ( I − W )Y )                         W = D −1/ 2WD −1/ 2
                                     ~
        Optimize : min Tr (Y T ( I − W )Y ), subject to Y T Y = I
                                Y
  By K. Fan’s theorem, optimal solution is
                                        ~
  eigenvectors: Y=(v1,v2, …, vk), ( I − W )vk = λk vk
                         ( D − W )u k = λk Du k ,                 u k = D −1 / 2 v k

                         λ1 + l + λk ≤ min J Ncut ( y1 , l, y k )                      (Gu, et al, 2001)

Tutorial on Spectral Clustering, ICML 2004, Chris Ding © University of California                          33
K-way Min-Max Cut Spectral Relaxation
     Unsigned cluster indicators:
                                                                                      ~
                                             y k = D1/ 2 hk / || D1/ 2 hk ||          W = D −1/ 2WD −1/ 2
      Re-write:
                                                 1                      1
             J MMC ( y1 , , y k ) =         T ~
                                                         ++        T ~
                                                                                 −k
                                            y1 W y1                yk W     yk
                                                        T ~
    Optimize : min J MMC (Y ), subject to Y T Y = I , y k Wy k  0.
                            Y

  Theorem. Optimal solution is by eigenvectors:
                          ~
  Y=(v1,v2, …, vk),      W v k = λk v k

                         k2
                                − k ≤ min J MMC ( y1 , m , y k )
                    λ1 + m + λk                                                         (Gu, et al, 2001)


Tutorial on Spectral Clustering, ICML 2004, Chris Ding © University of California                           34
K-way Spectral Clustering

          • Embedding (similar to PCA subspace approach)
                – Embed data points in the subspace of the K eigenvectors
                – Clustering embedded points using another algorithm, such as K-
                  means (Shi  Malik, Ng et al, Zha, et al)
          • Recursive 2-way clustering (standard graph partitioning)
                – If desired K is not power of 2, how optimcally to choose the next
                  sub-cluster to split? (Ding, et al 2002)
          • Both above approach do not use K-way clustering
            objective functions.
          • Refine the obtained clusters using the K-way clustering
            objective function typically improve the results (Ding et al
               2002).



Tutorial on Spectral Clustering, ICML 2004, Chris Ding © University of California     35
DNA Gene expression


     Lymphoma Cancer
     (Alizadeh et al, 2000)
                                              Genes


   Effects of feature selection:
     Select 900 genes out of
           4025 genes




Tutorial on Spectral Clustering, ICML 2004, Chris Ding © University of California Tissue sample   36
Lymphoma Cancer
       Tissue samples
      B cell lymphoma go thru
          different stages
                –3 cancer stages
                –3 normal stages
   Key question: can we detect
      them automatically ?




                                                                    PCA 2D Display
Tutorial on Spectral Clustering, ICML 2004, Chris Ding © University of California    37
Tutorial on Spectral Clustering, ICML 2004, Chris Ding © University of California   38
Brief summary of Part I
     •   Spectral graph partitioning as origin
     •   Clustering objective functions and solutions
     •   Extensions to bipartite and directed graphs
     •   Characteristics
           –   Principled approach
           –   Well-motivated objective functions
           –   Clear, un-ambiguous
           –   A framework of rich structures and contents
           –   Everything is proved rigorously (within the relaxation
               framework, i.e., using continuous approximation of the discrete
               variables)
     • Above results mostly done by 2001.
     • More to come in Part II
Tutorial on Spectral Clustering, ICML 2004, Chris Ding © University of California   39
icml2004 tutorial on spectral clustering part I

Más contenido relacionado

La actualidad más candente

A New Enhanced Method of Non Parametric power spectrum Estimation.
A New Enhanced Method of Non Parametric power spectrum Estimation.A New Enhanced Method of Non Parametric power spectrum Estimation.
A New Enhanced Method of Non Parametric power spectrum Estimation.CSCJournals
 
SPECTRAL SYNTHESIS PROBLEM FOR FOURIER ALGEBRAS
SPECTRAL SYNTHESIS PROBLEM FOR FOURIER ALGEBRASSPECTRAL SYNTHESIS PROBLEM FOR FOURIER ALGEBRAS
SPECTRAL SYNTHESIS PROBLEM FOR FOURIER ALGEBRASKunda Chowdaiah
 
study Streaming Multigrid For Gradient Domain Operations On Large Images
study Streaming Multigrid For Gradient Domain Operations On Large Imagesstudy Streaming Multigrid For Gradient Domain Operations On Large Images
study Streaming Multigrid For Gradient Domain Operations On Large ImagesChiamin Hsu
 
A new transformation into State Transition Algorithm for finding the global m...
A new transformation into State Transition Algorithm for finding the global m...A new transformation into State Transition Algorithm for finding the global m...
A new transformation into State Transition Algorithm for finding the global m...Michael_Chou
 
CVPR2010: Advanced ITinCVPR in a Nutshell: part 6: Mixtures
CVPR2010: Advanced ITinCVPR in a Nutshell: part 6: MixturesCVPR2010: Advanced ITinCVPR in a Nutshell: part 6: Mixtures
CVPR2010: Advanced ITinCVPR in a Nutshell: part 6: Mixtureszukun
 
Solving connectivity problems via basic Linear Algebra
Solving connectivity problems via basic Linear AlgebraSolving connectivity problems via basic Linear Algebra
Solving connectivity problems via basic Linear Algebracseiitgn
 
Weisfeiler and Leman Go Neural: Higher-order Graph Neural Networks
Weisfeiler and Leman Go Neural: Higher-order Graph Neural Networks Weisfeiler and Leman Go Neural: Higher-order Graph Neural Networks
Weisfeiler and Leman Go Neural: Higher-order Graph Neural Networks Christopher Morris
 
2014-06-20 Multinomial Logistic Regression with Apache Spark
2014-06-20 Multinomial Logistic Regression with Apache Spark2014-06-20 Multinomial Logistic Regression with Apache Spark
2014-06-20 Multinomial Logistic Regression with Apache SparkDB Tsai
 
Kernel for Chordal Vertex Deletion
Kernel for Chordal Vertex DeletionKernel for Chordal Vertex Deletion
Kernel for Chordal Vertex DeletionAkankshaAgrawal55
 
Glocalized Weisfeiler-Lehman Graph Kernels: Global-Local Feature Maps of Graphs
Glocalized Weisfeiler-Lehman Graph Kernels: Global-Local Feature Maps of Graphs Glocalized Weisfeiler-Lehman Graph Kernels: Global-Local Feature Maps of Graphs
Glocalized Weisfeiler-Lehman Graph Kernels: Global-Local Feature Maps of Graphs Christopher Morris
 
Ultrasound Modular Architecture
Ultrasound Modular ArchitectureUltrasound Modular Architecture
Ultrasound Modular ArchitectureJose Miguel Moreno
 
Recent Advances in Kernel-Based Graph Classification
Recent Advances in Kernel-Based Graph ClassificationRecent Advances in Kernel-Based Graph Classification
Recent Advances in Kernel-Based Graph ClassificationChristopher Morris
 
Minimax optimal alternating minimization \\ for kernel nonparametric tensor l...
Minimax optimal alternating minimization \\ for kernel nonparametric tensor l...Minimax optimal alternating minimization \\ for kernel nonparametric tensor l...
Minimax optimal alternating minimization \\ for kernel nonparametric tensor l...Taiji Suzuki
 
Principal component analysis and matrix factorizations for learning (part 1) ...
Principal component analysis and matrix factorizations for learning (part 1) ...Principal component analysis and matrix factorizations for learning (part 1) ...
Principal component analysis and matrix factorizations for learning (part 1) ...zukun
 
Ilya Shkredov – Subsets of Z/pZ with small Wiener norm and arithmetic progres...
Ilya Shkredov – Subsets of Z/pZ with small Wiener norm and arithmetic progres...Ilya Shkredov – Subsets of Z/pZ with small Wiener norm and arithmetic progres...
Ilya Shkredov – Subsets of Z/pZ with small Wiener norm and arithmetic progres...Yandex
 
Semi-empirical Monte Carlo optical-gain modelling of Nuclear Imaging scintill...
Semi-empirical Monte Carlo optical-gain modelling of Nuclear Imaging scintill...Semi-empirical Monte Carlo optical-gain modelling of Nuclear Imaging scintill...
Semi-empirical Monte Carlo optical-gain modelling of Nuclear Imaging scintill...Anax Fotopoulos
 
Closed-Form Solutions in Low-Rank Subspace Recovery Models and Their Implicat...
Closed-Form Solutions in Low-Rank Subspace Recovery Models and Their Implicat...Closed-Form Solutions in Low-Rank Subspace Recovery Models and Their Implicat...
Closed-Form Solutions in Low-Rank Subspace Recovery Models and Their Implicat...少华 白
 
Non-interacting and interacting Graphene in a strong uniform magnetic field
Non-interacting and interacting Graphene in a strong uniform magnetic fieldNon-interacting and interacting Graphene in a strong uniform magnetic field
Non-interacting and interacting Graphene in a strong uniform magnetic fieldAnkurDas60
 
Lec5 advanced-policy-gradient-methods
Lec5 advanced-policy-gradient-methodsLec5 advanced-policy-gradient-methods
Lec5 advanced-policy-gradient-methodsRonald Teo
 

La actualidad más candente (20)

A New Enhanced Method of Non Parametric power spectrum Estimation.
A New Enhanced Method of Non Parametric power spectrum Estimation.A New Enhanced Method of Non Parametric power spectrum Estimation.
A New Enhanced Method of Non Parametric power spectrum Estimation.
 
SPECTRAL SYNTHESIS PROBLEM FOR FOURIER ALGEBRAS
SPECTRAL SYNTHESIS PROBLEM FOR FOURIER ALGEBRASSPECTRAL SYNTHESIS PROBLEM FOR FOURIER ALGEBRAS
SPECTRAL SYNTHESIS PROBLEM FOR FOURIER ALGEBRAS
 
study Streaming Multigrid For Gradient Domain Operations On Large Images
study Streaming Multigrid For Gradient Domain Operations On Large Imagesstudy Streaming Multigrid For Gradient Domain Operations On Large Images
study Streaming Multigrid For Gradient Domain Operations On Large Images
 
A new transformation into State Transition Algorithm for finding the global m...
A new transformation into State Transition Algorithm for finding the global m...A new transformation into State Transition Algorithm for finding the global m...
A new transformation into State Transition Algorithm for finding the global m...
 
CVPR2010: Advanced ITinCVPR in a Nutshell: part 6: Mixtures
CVPR2010: Advanced ITinCVPR in a Nutshell: part 6: MixturesCVPR2010: Advanced ITinCVPR in a Nutshell: part 6: Mixtures
CVPR2010: Advanced ITinCVPR in a Nutshell: part 6: Mixtures
 
Solving connectivity problems via basic Linear Algebra
Solving connectivity problems via basic Linear AlgebraSolving connectivity problems via basic Linear Algebra
Solving connectivity problems via basic Linear Algebra
 
Introduction to DFT Part 2
Introduction to DFT Part 2Introduction to DFT Part 2
Introduction to DFT Part 2
 
Weisfeiler and Leman Go Neural: Higher-order Graph Neural Networks
Weisfeiler and Leman Go Neural: Higher-order Graph Neural Networks Weisfeiler and Leman Go Neural: Higher-order Graph Neural Networks
Weisfeiler and Leman Go Neural: Higher-order Graph Neural Networks
 
2014-06-20 Multinomial Logistic Regression with Apache Spark
2014-06-20 Multinomial Logistic Regression with Apache Spark2014-06-20 Multinomial Logistic Regression with Apache Spark
2014-06-20 Multinomial Logistic Regression with Apache Spark
 
Kernel for Chordal Vertex Deletion
Kernel for Chordal Vertex DeletionKernel for Chordal Vertex Deletion
Kernel for Chordal Vertex Deletion
 
Glocalized Weisfeiler-Lehman Graph Kernels: Global-Local Feature Maps of Graphs
Glocalized Weisfeiler-Lehman Graph Kernels: Global-Local Feature Maps of Graphs Glocalized Weisfeiler-Lehman Graph Kernels: Global-Local Feature Maps of Graphs
Glocalized Weisfeiler-Lehman Graph Kernels: Global-Local Feature Maps of Graphs
 
Ultrasound Modular Architecture
Ultrasound Modular ArchitectureUltrasound Modular Architecture
Ultrasound Modular Architecture
 
Recent Advances in Kernel-Based Graph Classification
Recent Advances in Kernel-Based Graph ClassificationRecent Advances in Kernel-Based Graph Classification
Recent Advances in Kernel-Based Graph Classification
 
Minimax optimal alternating minimization \\ for kernel nonparametric tensor l...
Minimax optimal alternating minimization \\ for kernel nonparametric tensor l...Minimax optimal alternating minimization \\ for kernel nonparametric tensor l...
Minimax optimal alternating minimization \\ for kernel nonparametric tensor l...
 
Principal component analysis and matrix factorizations for learning (part 1) ...
Principal component analysis and matrix factorizations for learning (part 1) ...Principal component analysis and matrix factorizations for learning (part 1) ...
Principal component analysis and matrix factorizations for learning (part 1) ...
 
Ilya Shkredov – Subsets of Z/pZ with small Wiener norm and arithmetic progres...
Ilya Shkredov – Subsets of Z/pZ with small Wiener norm and arithmetic progres...Ilya Shkredov – Subsets of Z/pZ with small Wiener norm and arithmetic progres...
Ilya Shkredov – Subsets of Z/pZ with small Wiener norm and arithmetic progres...
 
Semi-empirical Monte Carlo optical-gain modelling of Nuclear Imaging scintill...
Semi-empirical Monte Carlo optical-gain modelling of Nuclear Imaging scintill...Semi-empirical Monte Carlo optical-gain modelling of Nuclear Imaging scintill...
Semi-empirical Monte Carlo optical-gain modelling of Nuclear Imaging scintill...
 
Closed-Form Solutions in Low-Rank Subspace Recovery Models and Their Implicat...
Closed-Form Solutions in Low-Rank Subspace Recovery Models and Their Implicat...Closed-Form Solutions in Low-Rank Subspace Recovery Models and Their Implicat...
Closed-Form Solutions in Low-Rank Subspace Recovery Models and Their Implicat...
 
Non-interacting and interacting Graphene in a strong uniform magnetic field
Non-interacting and interacting Graphene in a strong uniform magnetic fieldNon-interacting and interacting Graphene in a strong uniform magnetic field
Non-interacting and interacting Graphene in a strong uniform magnetic field
 
Lec5 advanced-policy-gradient-methods
Lec5 advanced-policy-gradient-methodsLec5 advanced-policy-gradient-methods
Lec5 advanced-policy-gradient-methods
 

Similar a icml2004 tutorial on spectral clustering part I

icml2004 tutorial on spectral clustering part II
icml2004 tutorial on spectral clustering part IIicml2004 tutorial on spectral clustering part II
icml2004 tutorial on spectral clustering part IIzukun
 
Social Network Analysis
Social Network AnalysisSocial Network Analysis
Social Network Analysisrik0
 
WE4.L09 - MEAN-SHIFT AND HIERARCHICAL CLUSTERING FOR TEXTURED POLARIMETRIC SA...
WE4.L09 - MEAN-SHIFT AND HIERARCHICAL CLUSTERING FOR TEXTURED POLARIMETRIC SA...WE4.L09 - MEAN-SHIFT AND HIERARCHICAL CLUSTERING FOR TEXTURED POLARIMETRIC SA...
WE4.L09 - MEAN-SHIFT AND HIERARCHICAL CLUSTERING FOR TEXTURED POLARIMETRIC SA...grssieee
 
Numerical Linear Algebra for Data and Link Analysis.
Numerical Linear Algebra for Data and Link Analysis.Numerical Linear Algebra for Data and Link Analysis.
Numerical Linear Algebra for Data and Link Analysis.Leonid Zhukov
 
Community Detection
Community DetectionCommunity Detection
Community DetectionIlio Catallo
 
Mapping Ash Tree Colonization in an Agricultural Moutain Landscape_ Investiga...
Mapping Ash Tree Colonization in an Agricultural Moutain Landscape_ Investiga...Mapping Ash Tree Colonization in an Agricultural Moutain Landscape_ Investiga...
Mapping Ash Tree Colonization in an Agricultural Moutain Landscape_ Investiga...grssieee
 
Resolution-free community detection
Resolution-free community detectionResolution-free community detection
Resolution-free community detectionVincent Traag
 
Random Matrix Theory and Machine Learning - Part 4
Random Matrix Theory and Machine Learning - Part 4Random Matrix Theory and Machine Learning - Part 4
Random Matrix Theory and Machine Learning - Part 4Fabian Pedregosa
 
Divergence clustering
Divergence clusteringDivergence clustering
Divergence clusteringFrank Nielsen
 
Object Recognition with Deformable Models
Object Recognition with Deformable ModelsObject Recognition with Deformable Models
Object Recognition with Deformable Modelszukun
 
Elementary Landscape Decomposition of Combinatorial Optimization Problems
Elementary Landscape Decomposition of Combinatorial Optimization ProblemsElementary Landscape Decomposition of Combinatorial Optimization Problems
Elementary Landscape Decomposition of Combinatorial Optimization Problemsjfrchicanog
 
CS 354 Global Illumination
CS 354 Global IlluminationCS 354 Global Illumination
CS 354 Global IlluminationMark Kilgard
 
Introduction to second gradient theory of elasticity - Arjun Narayanan
Introduction to second gradient theory of elasticity - Arjun NarayananIntroduction to second gradient theory of elasticity - Arjun Narayanan
Introduction to second gradient theory of elasticity - Arjun NarayananArjun Narayanan
 
Physics of Algorithms Talk
Physics of Algorithms TalkPhysics of Algorithms Talk
Physics of Algorithms Talkjasonj383
 
Subspace Indexing on Grassmannian Manifold for Large Scale Visual Identification
Subspace Indexing on Grassmannian Manifold for Large Scale Visual IdentificationSubspace Indexing on Grassmannian Manifold for Large Scale Visual Identification
Subspace Indexing on Grassmannian Manifold for Large Scale Visual IdentificationUnited States Air Force Academy
 
Community structure in complex networks
Community structure in complex networksCommunity structure in complex networks
Community structure in complex networksVincent Traag
 
T. Popov - Drinfeld-Jimbo and Cremmer-Gervais Quantum Lie Algebras
T. Popov - Drinfeld-Jimbo and Cremmer-Gervais Quantum Lie AlgebrasT. Popov - Drinfeld-Jimbo and Cremmer-Gervais Quantum Lie Algebras
T. Popov - Drinfeld-Jimbo and Cremmer-Gervais Quantum Lie AlgebrasSEENET-MTP
 
Elementary Landscape Decomposition of Combinatorial Optimization Problems
Elementary Landscape Decomposition of Combinatorial Optimization ProblemsElementary Landscape Decomposition of Combinatorial Optimization Problems
Elementary Landscape Decomposition of Combinatorial Optimization Problemsjfrchicanog
 

Similar a icml2004 tutorial on spectral clustering part I (20)

icml2004 tutorial on spectral clustering part II
icml2004 tutorial on spectral clustering part IIicml2004 tutorial on spectral clustering part II
icml2004 tutorial on spectral clustering part II
 
Social Network Analysis
Social Network AnalysisSocial Network Analysis
Social Network Analysis
 
WE4.L09 - MEAN-SHIFT AND HIERARCHICAL CLUSTERING FOR TEXTURED POLARIMETRIC SA...
WE4.L09 - MEAN-SHIFT AND HIERARCHICAL CLUSTERING FOR TEXTURED POLARIMETRIC SA...WE4.L09 - MEAN-SHIFT AND HIERARCHICAL CLUSTERING FOR TEXTURED POLARIMETRIC SA...
WE4.L09 - MEAN-SHIFT AND HIERARCHICAL CLUSTERING FOR TEXTURED POLARIMETRIC SA...
 
Numerical Linear Algebra for Data and Link Analysis.
Numerical Linear Algebra for Data and Link Analysis.Numerical Linear Algebra for Data and Link Analysis.
Numerical Linear Algebra for Data and Link Analysis.
 
Community Detection
Community DetectionCommunity Detection
Community Detection
 
Mapping Ash Tree Colonization in an Agricultural Moutain Landscape_ Investiga...
Mapping Ash Tree Colonization in an Agricultural Moutain Landscape_ Investiga...Mapping Ash Tree Colonization in an Agricultural Moutain Landscape_ Investiga...
Mapping Ash Tree Colonization in an Agricultural Moutain Landscape_ Investiga...
 
Resolution-free community detection
Resolution-free community detectionResolution-free community detection
Resolution-free community detection
 
TunUp final presentation
TunUp final presentationTunUp final presentation
TunUp final presentation
 
Random Matrix Theory and Machine Learning - Part 4
Random Matrix Theory and Machine Learning - Part 4Random Matrix Theory and Machine Learning - Part 4
Random Matrix Theory and Machine Learning - Part 4
 
Divergence clustering
Divergence clusteringDivergence clustering
Divergence clustering
 
YSC 2013
YSC 2013YSC 2013
YSC 2013
 
Object Recognition with Deformable Models
Object Recognition with Deformable ModelsObject Recognition with Deformable Models
Object Recognition with Deformable Models
 
Elementary Landscape Decomposition of Combinatorial Optimization Problems
Elementary Landscape Decomposition of Combinatorial Optimization ProblemsElementary Landscape Decomposition of Combinatorial Optimization Problems
Elementary Landscape Decomposition of Combinatorial Optimization Problems
 
CS 354 Global Illumination
CS 354 Global IlluminationCS 354 Global Illumination
CS 354 Global Illumination
 
Introduction to second gradient theory of elasticity - Arjun Narayanan
Introduction to second gradient theory of elasticity - Arjun NarayananIntroduction to second gradient theory of elasticity - Arjun Narayanan
Introduction to second gradient theory of elasticity - Arjun Narayanan
 
Physics of Algorithms Talk
Physics of Algorithms TalkPhysics of Algorithms Talk
Physics of Algorithms Talk
 
Subspace Indexing on Grassmannian Manifold for Large Scale Visual Identification
Subspace Indexing on Grassmannian Manifold for Large Scale Visual IdentificationSubspace Indexing on Grassmannian Manifold for Large Scale Visual Identification
Subspace Indexing on Grassmannian Manifold for Large Scale Visual Identification
 
Community structure in complex networks
Community structure in complex networksCommunity structure in complex networks
Community structure in complex networks
 
T. Popov - Drinfeld-Jimbo and Cremmer-Gervais Quantum Lie Algebras
T. Popov - Drinfeld-Jimbo and Cremmer-Gervais Quantum Lie AlgebrasT. Popov - Drinfeld-Jimbo and Cremmer-Gervais Quantum Lie Algebras
T. Popov - Drinfeld-Jimbo and Cremmer-Gervais Quantum Lie Algebras
 
Elementary Landscape Decomposition of Combinatorial Optimization Problems
Elementary Landscape Decomposition of Combinatorial Optimization ProblemsElementary Landscape Decomposition of Combinatorial Optimization Problems
Elementary Landscape Decomposition of Combinatorial Optimization Problems
 

Más de zukun

My lyn tutorial 2009
My lyn tutorial 2009My lyn tutorial 2009
My lyn tutorial 2009zukun
 
ETHZ CV2012: Tutorial openCV
ETHZ CV2012: Tutorial openCVETHZ CV2012: Tutorial openCV
ETHZ CV2012: Tutorial openCVzukun
 
ETHZ CV2012: Information
ETHZ CV2012: InformationETHZ CV2012: Information
ETHZ CV2012: Informationzukun
 
Siwei lyu: natural image statistics
Siwei lyu: natural image statisticsSiwei lyu: natural image statistics
Siwei lyu: natural image statisticszukun
 
Lecture9 camera calibration
Lecture9 camera calibrationLecture9 camera calibration
Lecture9 camera calibrationzukun
 
Brunelli 2008: template matching techniques in computer vision
Brunelli 2008: template matching techniques in computer visionBrunelli 2008: template matching techniques in computer vision
Brunelli 2008: template matching techniques in computer visionzukun
 
Modern features-part-4-evaluation
Modern features-part-4-evaluationModern features-part-4-evaluation
Modern features-part-4-evaluationzukun
 
Modern features-part-3-software
Modern features-part-3-softwareModern features-part-3-software
Modern features-part-3-softwarezukun
 
Modern features-part-2-descriptors
Modern features-part-2-descriptorsModern features-part-2-descriptors
Modern features-part-2-descriptorszukun
 
Modern features-part-1-detectors
Modern features-part-1-detectorsModern features-part-1-detectors
Modern features-part-1-detectorszukun
 
Modern features-part-0-intro
Modern features-part-0-introModern features-part-0-intro
Modern features-part-0-introzukun
 
Lecture 02 internet video search
Lecture 02 internet video searchLecture 02 internet video search
Lecture 02 internet video searchzukun
 
Lecture 01 internet video search
Lecture 01 internet video searchLecture 01 internet video search
Lecture 01 internet video searchzukun
 
Lecture 03 internet video search
Lecture 03 internet video searchLecture 03 internet video search
Lecture 03 internet video searchzukun
 
Icml2012 tutorial representation_learning
Icml2012 tutorial representation_learningIcml2012 tutorial representation_learning
Icml2012 tutorial representation_learningzukun
 
Advances in discrete energy minimisation for computer vision
Advances in discrete energy minimisation for computer visionAdvances in discrete energy minimisation for computer vision
Advances in discrete energy minimisation for computer visionzukun
 
Gephi tutorial: quick start
Gephi tutorial: quick startGephi tutorial: quick start
Gephi tutorial: quick startzukun
 
EM algorithm and its application in probabilistic latent semantic analysis
EM algorithm and its application in probabilistic latent semantic analysisEM algorithm and its application in probabilistic latent semantic analysis
EM algorithm and its application in probabilistic latent semantic analysiszukun
 
Object recognition with pictorial structures
Object recognition with pictorial structuresObject recognition with pictorial structures
Object recognition with pictorial structureszukun
 
Iccv2011 learning spatiotemporal graphs of human activities
Iccv2011 learning spatiotemporal graphs of human activities Iccv2011 learning spatiotemporal graphs of human activities
Iccv2011 learning spatiotemporal graphs of human activities zukun
 

Más de zukun (20)

My lyn tutorial 2009
My lyn tutorial 2009My lyn tutorial 2009
My lyn tutorial 2009
 
ETHZ CV2012: Tutorial openCV
ETHZ CV2012: Tutorial openCVETHZ CV2012: Tutorial openCV
ETHZ CV2012: Tutorial openCV
 
ETHZ CV2012: Information
ETHZ CV2012: InformationETHZ CV2012: Information
ETHZ CV2012: Information
 
Siwei lyu: natural image statistics
Siwei lyu: natural image statisticsSiwei lyu: natural image statistics
Siwei lyu: natural image statistics
 
Lecture9 camera calibration
Lecture9 camera calibrationLecture9 camera calibration
Lecture9 camera calibration
 
Brunelli 2008: template matching techniques in computer vision
Brunelli 2008: template matching techniques in computer visionBrunelli 2008: template matching techniques in computer vision
Brunelli 2008: template matching techniques in computer vision
 
Modern features-part-4-evaluation
Modern features-part-4-evaluationModern features-part-4-evaluation
Modern features-part-4-evaluation
 
Modern features-part-3-software
Modern features-part-3-softwareModern features-part-3-software
Modern features-part-3-software
 
Modern features-part-2-descriptors
Modern features-part-2-descriptorsModern features-part-2-descriptors
Modern features-part-2-descriptors
 
Modern features-part-1-detectors
Modern features-part-1-detectorsModern features-part-1-detectors
Modern features-part-1-detectors
 
Modern features-part-0-intro
Modern features-part-0-introModern features-part-0-intro
Modern features-part-0-intro
 
Lecture 02 internet video search
Lecture 02 internet video searchLecture 02 internet video search
Lecture 02 internet video search
 
Lecture 01 internet video search
Lecture 01 internet video searchLecture 01 internet video search
Lecture 01 internet video search
 
Lecture 03 internet video search
Lecture 03 internet video searchLecture 03 internet video search
Lecture 03 internet video search
 
Icml2012 tutorial representation_learning
Icml2012 tutorial representation_learningIcml2012 tutorial representation_learning
Icml2012 tutorial representation_learning
 
Advances in discrete energy minimisation for computer vision
Advances in discrete energy minimisation for computer visionAdvances in discrete energy minimisation for computer vision
Advances in discrete energy minimisation for computer vision
 
Gephi tutorial: quick start
Gephi tutorial: quick startGephi tutorial: quick start
Gephi tutorial: quick start
 
EM algorithm and its application in probabilistic latent semantic analysis
EM algorithm and its application in probabilistic latent semantic analysisEM algorithm and its application in probabilistic latent semantic analysis
EM algorithm and its application in probabilistic latent semantic analysis
 
Object recognition with pictorial structures
Object recognition with pictorial structuresObject recognition with pictorial structures
Object recognition with pictorial structures
 
Iccv2011 learning spatiotemporal graphs of human activities
Iccv2011 learning spatiotemporal graphs of human activities Iccv2011 learning spatiotemporal graphs of human activities
Iccv2011 learning spatiotemporal graphs of human activities
 

Último

Histor y of HAM Radio presentation slide
Histor y of HAM Radio presentation slideHistor y of HAM Radio presentation slide
Histor y of HAM Radio presentation slidevu2urc
 
Mastering MySQL Database Architecture: Deep Dive into MySQL Shell and MySQL R...
Mastering MySQL Database Architecture: Deep Dive into MySQL Shell and MySQL R...Mastering MySQL Database Architecture: Deep Dive into MySQL Shell and MySQL R...
Mastering MySQL Database Architecture: Deep Dive into MySQL Shell and MySQL R...Miguel Araújo
 
How to Troubleshoot Apps for the Modern Connected Worker
How to Troubleshoot Apps for the Modern Connected WorkerHow to Troubleshoot Apps for the Modern Connected Worker
How to Troubleshoot Apps for the Modern Connected WorkerThousandEyes
 
Powerful Google developer tools for immediate impact! (2023-24 C)
Powerful Google developer tools for immediate impact! (2023-24 C)Powerful Google developer tools for immediate impact! (2023-24 C)
Powerful Google developer tools for immediate impact! (2023-24 C)wesley chun
 
EIS-Webinar-Prompt-Knowledge-Eng-2024-04-08.pptx
EIS-Webinar-Prompt-Knowledge-Eng-2024-04-08.pptxEIS-Webinar-Prompt-Knowledge-Eng-2024-04-08.pptx
EIS-Webinar-Prompt-Knowledge-Eng-2024-04-08.pptxEarley Information Science
 
Finology Group – Insurtech Innovation Award 2024
Finology Group – Insurtech Innovation Award 2024Finology Group – Insurtech Innovation Award 2024
Finology Group – Insurtech Innovation Award 2024The Digital Insurer
 
Driving Behavioral Change for Information Management through Data-Driven Gree...
Driving Behavioral Change for Information Management through Data-Driven Gree...Driving Behavioral Change for Information Management through Data-Driven Gree...
Driving Behavioral Change for Information Management through Data-Driven Gree...Enterprise Knowledge
 
Automating Google Workspace (GWS) & more with Apps Script
Automating Google Workspace (GWS) & more with Apps ScriptAutomating Google Workspace (GWS) & more with Apps Script
Automating Google Workspace (GWS) & more with Apps Scriptwesley chun
 
2024: Domino Containers - The Next Step. News from the Domino Container commu...
2024: Domino Containers - The Next Step. News from the Domino Container commu...2024: Domino Containers - The Next Step. News from the Domino Container commu...
2024: Domino Containers - The Next Step. News from the Domino Container commu...Martijn de Jong
 
Exploring the Future Potential of AI-Enabled Smartphone Processors
Exploring the Future Potential of AI-Enabled Smartphone ProcessorsExploring the Future Potential of AI-Enabled Smartphone Processors
Exploring the Future Potential of AI-Enabled Smartphone Processorsdebabhi2
 
Artificial Intelligence: Facts and Myths
Artificial Intelligence: Facts and MythsArtificial Intelligence: Facts and Myths
Artificial Intelligence: Facts and MythsJoaquim Jorge
 
Strategies for Landing an Oracle DBA Job as a Fresher
Strategies for Landing an Oracle DBA Job as a FresherStrategies for Landing an Oracle DBA Job as a Fresher
Strategies for Landing an Oracle DBA Job as a FresherRemote DBA Services
 
Tech Trends Report 2024 Future Today Institute.pdf
Tech Trends Report 2024 Future Today Institute.pdfTech Trends Report 2024 Future Today Institute.pdf
Tech Trends Report 2024 Future Today Institute.pdfhans926745
 
From Event to Action: Accelerate Your Decision Making with Real-Time Automation
From Event to Action: Accelerate Your Decision Making with Real-Time AutomationFrom Event to Action: Accelerate Your Decision Making with Real-Time Automation
From Event to Action: Accelerate Your Decision Making with Real-Time AutomationSafe Software
 
TrustArc Webinar - Stay Ahead of US State Data Privacy Law Developments
TrustArc Webinar - Stay Ahead of US State Data Privacy Law DevelopmentsTrustArc Webinar - Stay Ahead of US State Data Privacy Law Developments
TrustArc Webinar - Stay Ahead of US State Data Privacy Law DevelopmentsTrustArc
 
IAC 2024 - IA Fast Track to Search Focused AI Solutions
IAC 2024 - IA Fast Track to Search Focused AI SolutionsIAC 2024 - IA Fast Track to Search Focused AI Solutions
IAC 2024 - IA Fast Track to Search Focused AI SolutionsEnterprise Knowledge
 
Axa Assurance Maroc - Insurer Innovation Award 2024
Axa Assurance Maroc - Insurer Innovation Award 2024Axa Assurance Maroc - Insurer Innovation Award 2024
Axa Assurance Maroc - Insurer Innovation Award 2024The Digital Insurer
 
08448380779 Call Girls In Civil Lines Women Seeking Men
08448380779 Call Girls In Civil Lines Women Seeking Men08448380779 Call Girls In Civil Lines Women Seeking Men
08448380779 Call Girls In Civil Lines Women Seeking MenDelhi Call girls
 
Scaling API-first – The story of a global engineering organization
Scaling API-first – The story of a global engineering organizationScaling API-first – The story of a global engineering organization
Scaling API-first – The story of a global engineering organizationRadu Cotescu
 
The 7 Things I Know About Cyber Security After 25 Years | April 2024
The 7 Things I Know About Cyber Security After 25 Years | April 2024The 7 Things I Know About Cyber Security After 25 Years | April 2024
The 7 Things I Know About Cyber Security After 25 Years | April 2024Rafal Los
 

Último (20)

Histor y of HAM Radio presentation slide
Histor y of HAM Radio presentation slideHistor y of HAM Radio presentation slide
Histor y of HAM Radio presentation slide
 
Mastering MySQL Database Architecture: Deep Dive into MySQL Shell and MySQL R...
Mastering MySQL Database Architecture: Deep Dive into MySQL Shell and MySQL R...Mastering MySQL Database Architecture: Deep Dive into MySQL Shell and MySQL R...
Mastering MySQL Database Architecture: Deep Dive into MySQL Shell and MySQL R...
 
How to Troubleshoot Apps for the Modern Connected Worker
How to Troubleshoot Apps for the Modern Connected WorkerHow to Troubleshoot Apps for the Modern Connected Worker
How to Troubleshoot Apps for the Modern Connected Worker
 
Powerful Google developer tools for immediate impact! (2023-24 C)
Powerful Google developer tools for immediate impact! (2023-24 C)Powerful Google developer tools for immediate impact! (2023-24 C)
Powerful Google developer tools for immediate impact! (2023-24 C)
 
EIS-Webinar-Prompt-Knowledge-Eng-2024-04-08.pptx
EIS-Webinar-Prompt-Knowledge-Eng-2024-04-08.pptxEIS-Webinar-Prompt-Knowledge-Eng-2024-04-08.pptx
EIS-Webinar-Prompt-Knowledge-Eng-2024-04-08.pptx
 
Finology Group – Insurtech Innovation Award 2024
Finology Group – Insurtech Innovation Award 2024Finology Group – Insurtech Innovation Award 2024
Finology Group – Insurtech Innovation Award 2024
 
Driving Behavioral Change for Information Management through Data-Driven Gree...
Driving Behavioral Change for Information Management through Data-Driven Gree...Driving Behavioral Change for Information Management through Data-Driven Gree...
Driving Behavioral Change for Information Management through Data-Driven Gree...
 
Automating Google Workspace (GWS) & more with Apps Script
Automating Google Workspace (GWS) & more with Apps ScriptAutomating Google Workspace (GWS) & more with Apps Script
Automating Google Workspace (GWS) & more with Apps Script
 
2024: Domino Containers - The Next Step. News from the Domino Container commu...
2024: Domino Containers - The Next Step. News from the Domino Container commu...2024: Domino Containers - The Next Step. News from the Domino Container commu...
2024: Domino Containers - The Next Step. News from the Domino Container commu...
 
Exploring the Future Potential of AI-Enabled Smartphone Processors
Exploring the Future Potential of AI-Enabled Smartphone ProcessorsExploring the Future Potential of AI-Enabled Smartphone Processors
Exploring the Future Potential of AI-Enabled Smartphone Processors
 
Artificial Intelligence: Facts and Myths
Artificial Intelligence: Facts and MythsArtificial Intelligence: Facts and Myths
Artificial Intelligence: Facts and Myths
 
Strategies for Landing an Oracle DBA Job as a Fresher
Strategies for Landing an Oracle DBA Job as a FresherStrategies for Landing an Oracle DBA Job as a Fresher
Strategies for Landing an Oracle DBA Job as a Fresher
 
Tech Trends Report 2024 Future Today Institute.pdf
Tech Trends Report 2024 Future Today Institute.pdfTech Trends Report 2024 Future Today Institute.pdf
Tech Trends Report 2024 Future Today Institute.pdf
 
From Event to Action: Accelerate Your Decision Making with Real-Time Automation
From Event to Action: Accelerate Your Decision Making with Real-Time AutomationFrom Event to Action: Accelerate Your Decision Making with Real-Time Automation
From Event to Action: Accelerate Your Decision Making with Real-Time Automation
 
TrustArc Webinar - Stay Ahead of US State Data Privacy Law Developments
TrustArc Webinar - Stay Ahead of US State Data Privacy Law DevelopmentsTrustArc Webinar - Stay Ahead of US State Data Privacy Law Developments
TrustArc Webinar - Stay Ahead of US State Data Privacy Law Developments
 
IAC 2024 - IA Fast Track to Search Focused AI Solutions
IAC 2024 - IA Fast Track to Search Focused AI SolutionsIAC 2024 - IA Fast Track to Search Focused AI Solutions
IAC 2024 - IA Fast Track to Search Focused AI Solutions
 
Axa Assurance Maroc - Insurer Innovation Award 2024
Axa Assurance Maroc - Insurer Innovation Award 2024Axa Assurance Maroc - Insurer Innovation Award 2024
Axa Assurance Maroc - Insurer Innovation Award 2024
 
08448380779 Call Girls In Civil Lines Women Seeking Men
08448380779 Call Girls In Civil Lines Women Seeking Men08448380779 Call Girls In Civil Lines Women Seeking Men
08448380779 Call Girls In Civil Lines Women Seeking Men
 
Scaling API-first – The story of a global engineering organization
Scaling API-first – The story of a global engineering organizationScaling API-first – The story of a global engineering organization
Scaling API-first – The story of a global engineering organization
 
The 7 Things I Know About Cyber Security After 25 Years | April 2024
The 7 Things I Know About Cyber Security After 25 Years | April 2024The 7 Things I Know About Cyber Security After 25 Years | April 2024
The 7 Things I Know About Cyber Security After 25 Years | April 2024
 

icml2004 tutorial on spectral clustering part I

  • 1. A Tutorial on Spectral Clustering Chris Ding Computational Research Division Lawrence Berkeley National Laboratory University of California Supported by Office of Science, U.S. Dept. of Energy Tutorial on Spectral Clustering, ICML 2004, Chris Ding © University of California 1
  • 2. Some historical notes • Fiedler, 1973, 1975, graph Laplacian matrix • Donath & Hoffman, 1973, bounds • Pothen, Simon, Liou, 1990, Spectral graph partitioning (many related papers there after) • Hagen & Kahng, 1992, Ratio-cut • Chan, Schlag & Zien, multi-way Ratio-cut • Chung, 1997, Spectral graph theory book • Shi & Malik, 2000, Normalized Cut Tutorial on Spectral Clustering, ICML 2004, Chris Ding © University of California 2
  • 3. Spectral Gold-Rush of 2001 9 papers on spectral clustering • Meila & Shi, AI-Stat 2001. Random Walk interpreation of Normalized Cut • Ding, He & Zha, KDD 2001. Perturbation analysis of Laplacian matrix on sparsely connected graphs • Ng, Jordan & Weiss, NIPS 2001, K-means algorithm on the embeded eigen-space • Belkin & Niyogi, NIPS 2001. Spectral Embedding • Dhillon, KDD 2001, Bipartite graph clustering • Zha et al, CIKM 2001, Bipartite graph clustering • Zha et al, NIPS 2001. Spectral Relaxation of K-means • Ding et al, ICDM 2001. MinMaxCut, Uniqueness of relaxation. • Gu et al, K-way Relaxation of NormCut and MinMaxCut Tutorial on Spectral Clustering, ICML 2004, Chris Ding © University of California 3
  • 4. Part I: Basic Theory, 1973 – 2001 Tutorial on Spectral Clustering, ICML 2004, Chris Ding © University of California 4
  • 5. Spectral Graph Partitioning MinCut: min cutsize Constraint on sizes: |A| = |B| cutsize = # of cut edges Tutorial on Spectral Clustering, ICML 2004, Chris Ding © University of California 5
  • 6. 2-way Spectral Graph Partitioning  1 if i ∈ A Partition membership indicator: qi =  − 1 if i ∈ B ∑ 1 J = CutSize = wij [qi − q j ]2 4 i, j ∑ ∑ 1 1 = wij [qi2 + q 2 − 2qi q j ] = j q [d δ − wij ]q j 4 i, j 2 i , j i i ij 1 T = q ( D − W )q 2 Relax indicators qi from discrete values to continuous values, the solution for min J(q) is given by the eigenvectors of ( D − W ) q = λq (Fiedler, 1973, 1975) (Pothen, Simon, Liou, 1990) Tutorial on Spectral Clustering, ICML 2004, Chris Ding © University of California 6
  • 7. Properties of Graph Laplacian Laplacian matrix of the Graph: L = D − W • L is semi-positive definite xT Lx ≥ 0 for any x. • First eigenvector is q1=(1,…,1)T = eT with λ1=0. • Second eigenvector q2 is the desired solution. • The smaller λ2, the better quality of the partitioning. Perturbation analysis gives cutsize cutsize λ2 = + | A| |B| • Higher eigenvectors are also useful Tutorial on Spectral Clustering, ICML 2004, Chris Ding © University of California 7
  • 8. Recovering Partitions From the definition of cluster indicators: Partitions A, B are determined by: A = {i | q2 (i ) < 0}, B = {i | q2 (i ) ≥ 0} However, the objective function J(q) is insensitive to additive constant c : ∑ 1 J = CutSize = w [( qi + c) − (q j + c)]2 4 i , j ij Thus, we sort q2 to increasing order, and cut in the middle point. Tutorial on Spectral Clustering, ICML 2004, Chris Ding © University of California 8
  • 9. Multi-way Graph Partitioning • Recursively applying the 2-way partitioning • Recursive 2-way partitioning • Using Kernigan-Lin to do local refinements • Using higher eigenvectors • Using q3 to further partitioning those obtained via q2. • Popular graph partitioning packages • Metis, Univ of Minnesota • Chaco, Sandia Nat’l Lab Tutorial on Spectral Clustering, ICML 2004, Chris Ding © University of California 9
  • 10. 2-way Spectral Clustering • Undirected graphs (pairwise similarities) • Bipartite graphs (contingency tables) • Directed graphs (web graphs) Tutorial on Spectral Clustering, ICML 2004, Chris Ding © University of California 10
  • 11. Spectral Clustering min cutsize , without explicit size constraints But where to cut ? Need to balance sizes Tutorial on Spectral Clustering, ICML 2004, Chris Ding © University of California 11
  • 12. Clustering Objective Functions s(A,B) = ∑∑ w ij • Ratio Cut i∈A j∈B s(A,B) s(A,B) J Rcut (A,B) = + |A| |B| • Normalized Cut dA = ∑d i i∈A s( A, B ) s( A, B ) J Ncut ( A, B) = + dA dB s ( A, B) s ( A, B ) = + s ( A, A) + s ( A, B ) s(B, B ) + s ( A, B ) • Min-Max-Cut s(A,B) s(A,B) J MMC (A,B) = + s(A,A) s(B,B) Tutorial on Spectral Clustering, ICML 2004, Chris Ding © University of California 12
  • 13. Ratio Cut (Hagen & Kahng, 1992) Min similarity between A , B: s(A,B) = ∑∑ w i∈ A j∈B ij s(A,B) s(A,B) Size Balance J Rcut (A,B) = + (Wei & Cheng, 1989) |A| |B|   n2 / n1n if i ∈ A Cluster membership indicator: q(i ) =  − n1 / n2 n  if i ∈ B Normalization: q T q = 1, q T e = 0 Substitute q leads to J Rcut (q) = q T ( D − W )q Now relax q, the by eigenvectornd eigenvector of L Solution given solution is 2 Tutorial on Spectral Clustering, ICML 2004, Chris Ding © University of California 13
  • 14. Normalized Cut (Shi & Malik, 1997) Min similarity between A & B: s(A,B) = ∑ ∑ wij i∈ A j∈B Balance weights s( A, B ) s( A, B ) J Ncut ( A, B) = + dA dB dA = ∑d i∈A i  d B / d Ad  if i ∈ A Cluster indicator: q(i ) =  − d A / d B d  if i ∈ B d= ∑d i∈G i Normalization: q Dq = 1, q De = 0 T T Substitute q leads to J Ncut (q) = q T ( D − W )q min q q T ( D − W )q + λ (q T Dq − 1) Solution is eigenvector of ( D − W )q = λDq Tutorial on Spectral Clustering, ICML 2004, Chris Ding © University of California 14
  • 15. MinMaxCut (Ding et al 2001) Min similarity between A & B: s(A,B) = ∑∑ w i∈A j∈B ij Max similarity within A & B: s(A,A) = ∑∑ w i∈A j∈A ij s(A,B) s(A,B) J MMC(A,B) = + s(A,A) s(B,B)   d B / d Ad if i ∈ A Cluster indicator: q(i ) =  − d A / d B d  if i ∈ B Substituting, 1+ dB / d A 1+ d A / dB q T Wq J MMC ( q) = + −2 Jm = Jm + dB / d A Jm + d A / dB q T Dq Because dJ MMC ( J m ) <0 min Jmmc ⇒ max Jm(q) dJ m ⇒ Wq = ξDq ⇒ ( D − W )q = λDq Tutorial on Spectral Clustering, ICML 2004, Chris Ding © University of California 15
  • 16. A simple example 2 dense clusters, with sparse connections between them. Adjacency matrix Eigenvector q2 Tutorial on Spectral Clustering, ICML 2004, Chris Ding © University of California 16
  • 17. Comparison of Clustering Objectives • If clusters are well separated, all three give very similar and accurate results. • When clusters are marginally separated, NormCut and MinMaxCut give better results • When clusters overlap significantly, MinMaxCut tend to give more compact and balanced clusters. s ( A, B) s ( A, B) J Ncut = + s ( A, A) + s(A, B) s(B, B) + s(A, B) Cluster Compactness ⇒ max s ( A, A) Tutorial on Spectral Clustering, ICML 2004, Chris Ding © University of California 17
  • 18. 2-way Clustering of Newsgroups Newsgroups RatioCut NormCut MinMaxCut Atheism 63.2 ± 16.2 97.2 ± 0.8 97.2 ± 1.1 Comp.graphics Baseball 54.9 ± 2.5 74.4 ± 20.4 79.5 ± 11.0 Hockey Politics.mideast 53.6 ± 3.1 57.5 ± 0.9 83.6 ± 2.5 Politics.misc Tutorial on Spectral Clustering, ICML 2004, Chris Ding © University of California 18
  • 19. Cluster Balance Analysis I: Random Graph Model • Random graph: edges are randomly assigned with probability p: 0 ≤ p ≤ 1. • RatioCut & NormCut show no size dependence p | A || B | p | A || B | J Rcut ( A, B) = + = np = constant | A| |B| p | A || B | p | A || B | n J Ncut ( A, B) = + = = constant p | A | (n − 1) p | B | ( n − 1) n − 1 • MinMaxCut favors balanced clusters: |A|=|B| p | A || B | p | A || B | |B| | A| J MMC ( A, B) = + = + p | A | (| A | −1) p | B | (| B | −1) | A | −1 | B | −1 Tutorial on Spectral Clustering, ICML 2004, Chris Ding © University of California 19
  • 20. 2-way Clustering of Newsgroups Cluster Balance Eigenvector JNcut(i) JMMC(i) Tutorial on Spectral Clustering, ICML 2004, Chris Ding © University of California 20
  • 21. Cluster Balance Analysis II: Large Overlap Case s ( A, B ) f = > 0.5 (1 / 2)[ s( A, A) + s( B, B)] Conditions for skewed cuts: 1 1 NormCut : s(A,A) ≥ ( − ) s ( A, B ) = s ( A, B) / 2 2f 2 1 MinMaxCut : s(A,A) ≥ s ( A, B) = s ( A, B ) 2f Thus MinMaxCut is much less prone to skewed cuts Tutorial on Spectral Clustering, ICML 2004, Chris Ding © University of California 21
  • 22. Spectral Clustering of Bipartite Graphs Simultaneous clustering of rows and columns of a contingency table (adjacency matrix B ) Examples of bipartite graphs • Information Retrieval: word-by-document matrix • Market basket data: transaction-by-item matrix • DNA Gene expression profiles • Protein vs protein-complex Tutorial on Spectral Clustering, ICML 2004, Chris Ding © University of California 22
  • 23. Spectral Clustering of Bipartite Graphs Simultaneous clustering of rows and columns (adjacency matrix B ) s ( BR1 ,C2 ) = ∑ ∑b ri ∈R1c j ∈C 2 ij min between-cluster sum of xyz weights: s(R1,C2), s(R2,C1) max within-cluster sum of xyz cut xyz weights: s(R1,C1), s(R2,C2) s ( BR1 ,C2 ) + s ( B R2 ,C1 ) s ( BR1 ,C2 ) + s ( B R2 ,C1 ) J MMC (C1 , C 2 ; R1 , R2 ) = + 2s ( B R1 ,C1 ) 2s ( BR2 ,C2 ) (Ding, AI-STAT 2003) Tutorial on Spectral Clustering, ICML 2004, Chris Ding © University of California 23
  • 24. Bipartite Graph Clustering Clustering indicators for rows and columns:  1 if ri ∈ R1  1 if ci ∈ C1 f (i ) =  g (i ) =  − 1 if ri ∈ R2 − 1 if ci ∈ C2  BR ,C BR1 ,C2   0 B f  B= 1 1  W = T  q=  g  BR ,C BR2 ,C2  B 0    2 1    Substitute and obtain s (W12 ) s (W12 ) J MMC (C1 , C 2 ; R1 , R2 ) = + s (W11 ) s (W22 ) f,g are determined by  D r   0 B   f  D  f   −    = λ  r      Dc   B T    g 0       Dc  g    Tutorial on Spectral Clustering, ICML 2004, Chris Ding © University of California 24
  • 25. Clustering of Bipartite Graphs Let ~ u   D1 / 2 f  B = Dr−1/ 2 BDc 1/ 2 , z =   = Dq =  r / 2  −    D1 g  v  c  We obtain ~  0 B  u  u  ~    B T 0  v  = λ  v         m ∑u λ v ~ Solution is SVD: B= T k k k k =1 (Zha et al, 2001, Dhillon, 2001) Tutorial on Spectral Clustering, ICML 2004, Chris Ding © University of California 25
  • 26. Clustering of Bipartite Graphs Recovering row clusters: R1 = {ri , | f 2 (i) < z r }, R2 = {ri , | f 2 (i) ≥ z r }, Recovering column clusters: C1 = {ci , | g 2 (i ) < z c }, C 2 = {ci , | g 2 (i ) ≥ z c }, zr=zc=0 are dividing points. Relaxation is invariant up to a constant shift. Algorithm: search for optimal points icut, jcut, let zr=f2(icut), zc= g2(jcut), such that J MMC (C1 , C2 ; R1 , R2 ) (Zha et al, 2001) is minimized. Tutorial on Spectral Clustering, ICML 2004, Chris Ding © University of California 26
  • 27. Clustering of Directed Graphs Min directed edge weights between A & B: s(A,B)= ∑∑(w i∈A j∈B ij + w ji ) Max directed edges within A & B: s(A,A)= ∑∑(w i∈A j∈A ij + w ji ) • Equivalent to deal with W = W + W T ~ ~ • All spectral methods apply to W • For example, web graphs clustered in such way (He, Ding, Zha, Simon, ICDM 2001) Tutorial on Spectral Clustering, ICML 2004, Chris Ding © University of California 27
  • 28. K-way Spectral Clustering K≥2 Tutorial on Spectral Clustering, ICML 2004, Chris Ding © University of California 28
  • 29. K-way Clustering Objectives • Ratio Cut  s(C k ,Cl ) s (C k ,Cl )  s(C k ,G − C k ) J Rcut (C1 , , C K ) = ∑   k ,l |C k| + |Cl|  =  ∑ k |C k| • Normalized Cut  s(C k ,Cl ) s (C k ,Cl )  s(C k ,G − C k ) J Ncut (C1 , , C K ) = ∑   k ,l dk + dl =   ∑ k dk • Min-Max-Cut  s(C k ,Cl ) s (C k ,Cl )  s(C k ,G − C k ) J MMC (C1 , , C K ) = ∑  s(C , C ) + s(C , C )  = ∑   k ,l k k   l l k s (C k , C k ) Tutorial on Spectral Clustering, ICML 2004, Chris Ding © University of California 29
  • 30. K-way Spectral Relaxation • Prove that the solution lie in the subspace spanned by the first k eigenvectors • Ratio Cut • Normalized Cut • Min-Max-Cut Tutorial on Spectral Clustering, ICML 2004, Chris Ding © University of California 30
  • 31. K-way Spectral Relaxation h1 = (1m1,0m 0,0 m 0)T Unsigned cluster indicators: h2 = (0 m 0,1m1,0 m 0)T mmm Re-write: hk = (0 m 0,0 m 0,1m1)T h1 ( D − W )h1 T hk ( D − W ) hk T J Rcut (h1 , , hk ) = T ++ T h1 h1 hk hk h1 ( D − W )h1 T hk ( D − W )hk T J Ncut (h1 , , hk ) = T ++ T h1 Dh1 hk Dhk h1 ( D − W ) h1 T hk ( D − W )hk T J MMC (h1 , , hk ) = T ++ T h1 Wh1 hk Whk Tutorial on Spectral Clustering, ICML 2004, Chris Ding © University of California 31
  • 32. K-way Ratio Cut Spectral Relaxation - Unsigned cluster indicators: x = (0 0,11,0 0)T / n1/ 2 nk k k Re-write: J ( x , , x ) = xT ( D − W ) x + + xT ( D − W ) x Rcut 1 k 1 1 k k = Tr ( X T ( D − W ) X ) X = ( x1 , , xk ) Optimize : min Tr ( X T ( D − W ) X ), subject to X T X = I X By K. Fan’s theorem, optimal solution is eigenvectors: X=(v1,v2, …, vk), (D-W)vk=λkvk and lower-bound λ1 + + λk ≤ min J Rcut ( x1 , , xk ) (Chan, Schlag, Zien, 1994) Tutorial on Spectral Clustering, ICML 2004, Chris Ding © University of California 32
  • 33. K-way Normalized Cut Spectral Relaxation Unsigned cluster indicators: -nk yk = D 1/ 2 (0 o 0,1o1,0o 0)T / || D1/ 2 hk || Re-write: ~ ~ J Ncut ( y1 , , y k ) = y1 ( I − W ) y1 + + y k ( I − W ) y k T T ~ ~ = Tr (Y T ( I − W )Y ) W = D −1/ 2WD −1/ 2 ~ Optimize : min Tr (Y T ( I − W )Y ), subject to Y T Y = I Y By K. Fan’s theorem, optimal solution is ~ eigenvectors: Y=(v1,v2, …, vk), ( I − W )vk = λk vk ( D − W )u k = λk Du k , u k = D −1 / 2 v k λ1 + l + λk ≤ min J Ncut ( y1 , l, y k ) (Gu, et al, 2001) Tutorial on Spectral Clustering, ICML 2004, Chris Ding © University of California 33
  • 34. K-way Min-Max Cut Spectral Relaxation Unsigned cluster indicators: ~ y k = D1/ 2 hk / || D1/ 2 hk || W = D −1/ 2WD −1/ 2 Re-write: 1 1 J MMC ( y1 , , y k ) = T ~ ++ T ~ −k y1 W y1 yk W yk T ~ Optimize : min J MMC (Y ), subject to Y T Y = I , y k Wy k 0. Y Theorem. Optimal solution is by eigenvectors: ~ Y=(v1,v2, …, vk), W v k = λk v k k2 − k ≤ min J MMC ( y1 , m , y k ) λ1 + m + λk (Gu, et al, 2001) Tutorial on Spectral Clustering, ICML 2004, Chris Ding © University of California 34
  • 35. K-way Spectral Clustering • Embedding (similar to PCA subspace approach) – Embed data points in the subspace of the K eigenvectors – Clustering embedded points using another algorithm, such as K- means (Shi Malik, Ng et al, Zha, et al) • Recursive 2-way clustering (standard graph partitioning) – If desired K is not power of 2, how optimcally to choose the next sub-cluster to split? (Ding, et al 2002) • Both above approach do not use K-way clustering objective functions. • Refine the obtained clusters using the K-way clustering objective function typically improve the results (Ding et al 2002). Tutorial on Spectral Clustering, ICML 2004, Chris Ding © University of California 35
  • 36. DNA Gene expression Lymphoma Cancer (Alizadeh et al, 2000) Genes Effects of feature selection: Select 900 genes out of 4025 genes Tutorial on Spectral Clustering, ICML 2004, Chris Ding © University of California Tissue sample 36
  • 37. Lymphoma Cancer Tissue samples B cell lymphoma go thru different stages –3 cancer stages –3 normal stages Key question: can we detect them automatically ? PCA 2D Display Tutorial on Spectral Clustering, ICML 2004, Chris Ding © University of California 37
  • 38. Tutorial on Spectral Clustering, ICML 2004, Chris Ding © University of California 38
  • 39. Brief summary of Part I • Spectral graph partitioning as origin • Clustering objective functions and solutions • Extensions to bipartite and directed graphs • Characteristics – Principled approach – Well-motivated objective functions – Clear, un-ambiguous – A framework of rich structures and contents – Everything is proved rigorously (within the relaxation framework, i.e., using continuous approximation of the discrete variables) • Above results mostly done by 2001. • More to come in Part II Tutorial on Spectral Clustering, ICML 2004, Chris Ding © University of California 39