1. Beijing, China August 17, 2009 A Framework for Multi-objective Clustering and Its Application to Co-location Mining RachsudaJianthapthaksin, Christoph F. Eick and Ricardo Vilalta University of Houston, Texas, USA
2. Talk Outline What is unique about this work with respect to clustering? Multi-objective Clustering (MOC)—Objectives and an Architecture Clustering with Plug-in Fitness Functions Filling the Repository with Clusters Creating Final Clusterings Related Work Co-location Mining Case Study Conclusion and Future Work
3. 1. What is unique about this work with respect to clustering? Clustering algorithms that support plug-in fitness function are used. Clustering algorithms are run multiple times to create clusters. Clusters are stored in a repository that is updated on the fly; cluster generation is separated from creating the final clustering. The final clustering is created from the clusters in the repository based on user preferences. Our approach needs to seeks for alternative, overlapping clusters.
4.
5. Task: Find sets of clusters that a good with respect to two or more objectivesTexas Multi-Objective Clustering Dataset: (longitude,latitude,<concentrations>+)
6. Survey MOC Approach Clustering algorithms are run multiple times maximizing different subsets of objectives that are captured in compound fitness functions. Uses a repository to store promising candidates. Only clusters that satisfying two or more objectives are considered as candidates. After a sufficient number of clusters has been created, final clustering are generated based on user-preferences. 5
7. An Architecture for MOC S1 S2 Q’ Clustering Algorithm Goal-driven Fitness Function Generator A Spatial Dataset M X Cluster Summarization Unit Q’ M’ Storage Unit S3 S4 Steps in multi-run clustering: S1: Generate a compound fitness function. S2: Run a clustering algorithm. S3: Update the cluster repository M. S4: Summarize clusters discovered M’. 6
8. 3. Clustering with Plug-in Fitness Functions Motivation: Finding subgroups in geo-referenced datasets has many applications. However, in many applications the subgroups to be searched for do not share the characteristics considered by traditional clustering algorithms, such as cluster compactness and separation. Domain or task knowledge frequently imposes additional requirements concerning what constitutes a “good” subgroup. Consequently, it is desirable to develop clustering algorithms that provide plug-in fitness functions that allow domain experts to express desirable characteristics of subgroups they are looking for.
9. Current Suite of Spatial Clustering Algorithms Representative-based: SCEC, SRIDHCR, SPAM, CLEVER Grid-based: SCMRG Agglomerative: MOSAIC Density-based: SCDE, DCONTOUR (not really plug-in but some fitness functions can be simulated) Density-based Grid-based Agglomerative-based Representative-based Clustering Algorithms Remark: All algorithms partition a dataset into clusters by maximizing a reward-based, plug-in fitness function.
10. 4. Filling the Repository with Clusters Plug-in Reward functions Rewardq(x) are used to assess to which extend an objective q is satisfied for a cluster x. User defined thresholds q are used to determine if an objective q is satisfied by a cluster x (Rewardq(x)>q). Only clusters that satisfy 2 or more objectives are stored in the repository. Only non-dominated clusters are stored in the repository. Dominance relations only apply to pairs of clusters that have a certain degree of agreement (overlap) sim.
11. Dominance and Multi-Objective Clusters Dominance between clusters x and y with respect to multiple objectives Q. Dominance Constraint with Respect to the Repository 10
12. Compound Fitness Functions The goal-driven fitness function generator selects a subset Q’(Q) of the objectives Q and creates a compound fitness function qQ’relying on a penalty function approach [Baeck et al. 2000]. CmpReward(x)= (qQ’Rewardq(x)) * Penalty(Q’,x) 11
13. Updating the Cluster Repository M:= clusters in the repository X:= “new” clusters generated by a single run of the clustering algorithm 12
14. 5. Creating a Final Clustering Final clusterings are subsets of the clusters in the repository M. Inputs: The user provides her own individual objective function RewardU and a reward threshold U and cluster similarity threshold rem that indicates how much cluster overlap she likes to tolerate. Goal: Find XM that maximizes: subject to: 1. xXx’X (xx’ Similarity(x,x’)<rem) 2. xX(RewardU(x)>U) Our paper introduces MO-Dominance-guided Cluster Reduction algorithm (MO-DCR) to create the final clustering.
15. MO-Dominance-guided Cluster Reduction(MO-DCR) algorithm (MO-DCR) : a dominant cluster : dominated clusters A The algorithm loops over the following 2 steps until M is empty: Include dominant clusters D which are the highest reward clusters in M’ Remove D and their dominated clusters in the rem-proximity from M. B C Dominance graphs D Remark: AB RewardU(A)>RewardU(B) Similarity(A,B)> rem E F sim(A,B)=0.8 rem=0.5 M’ 0.7 0.6 A E A E F 14
16.
17. 7. Case Study: Co-location Mining Goal: Finding regional co-location patterns where high concentrations of Arsenic are co-located with a lot of other factors in Texas. Remark: Each binary co-location is treated as a single objective. Dataset: TWDB has monitored water quality and collected the data for 105,814 wells in Texas over last 25 years. we use a subset of Arsenic_10_avg data set: longitude and latitude, Arsenic (As), Molybdenum (Mo), Vanadium (V), Boron (B), Fluoride (F-), Chloride (Cl-), Sulfate (SO42-) and Total Dissolved Solids (TDS). 16
20. Experimental Results MOC is able to identify: Multi-objective clusters Alternative clusters e.g. Rank1 regions of (a) and Rank2 regions of (b) Nested clusters e.g. in (b) Rank3-5 regions are sub-regions of Rank1 region. Particularly discriminate among companion elements such as Vanadium (Rank3 region), or Chloride, Sulfate and Total Dissolved Solids (Rank4 region). (a) (b) Fig. 7.6 The top 5 regions and patterns with respect to two queries: query1={As,Mo} and query2={As,B} are shown in Figure (a) and (b), respectively. 19
21. 8. Conclusion and Future Work Building blocks for Future Multi-Objective Clustering Systems were provided in this work; namely: A dominance relation for problems in which only a subset of the objectives can be satisfied was introduced. Clustering algorithms with plug-in fitness functions and the capability to create compound fitness functions are excessively used in our approach. Initially, a repository of potentially useful clusters is generated based on a large set of objectives. Individualized, specific clusterings are then generated based on user preferences. The approach is highly generic and incorporates specific domain needs in form of single-objective fitness functions. The approach was evaluated in a case study and turned out more suitable than a single-objective clustering approach that was used for the same application in a previous paper [ACM-GIS 2008].
22. Challenges in Multi-objective Clustering (MOC) Find clusters that are individually good with respect to multiple objectives in an automated fashion. Provide search engine style capabilities to summarize final clustering obtained from multiple runs of clustering algorithms. 21
23. Traditional Clustering Algorithms & Fitness Functions Clustering Algorithms No Fitness Function Fixed Fitness Function Provides Plug-in Fitness Function Implicit Fitness Function DBSCAN Hierarchical Clustering CHAMELEON Our Work PAM K-Means Traditional clustering algorithms consider only domain independent and task independent characteristics to form a solution. Different domain tasks require different fitness functions. Traditional Clustering Algorithms 22
26. Interestingness of a Pattern Interestingness of a pattern B (e.g. B= {C, D, E}) for an object o, Interestingness of a pattern B for a region c, Remark: Purity (i(B,o)>0) measures the percentage of objects that exhibit pattern B in region c.
27. Characteristics of the Top5 Regions Table 7.7 Top 5 Regions Ranked by Reward of the Query {As,Mo} Table 7.8 Top 5 Regions Ranked by Reward of the Query {As, B} 26
28. Representative-based Clustering 2 Attribute1 1 3 Attribute2 4 Objective: Find a set of objects OR such that the clustering X obtained by using the objects in OR as representatives minimizes q(X). Properties: Cluster shapes are convex polygons Popular Algorithms: K-means. K-medoids
29. 5. CLEVER (ClustEring using representatiVEs and Randomized hill climbing) Is a representative-based, sometimes called prototype-based clustering algorithm Uses variable number of clusters and larger neighborhood sizes to battle premature termination and randomized hill climbing and adaptive sampling to reduce complexity. Searches for optimal number of clusters
Notas del editor
—radical depart from traditional clustering, MOC Relies on clustering algorithms that support plug-in fitness functions, and multi-run clustering with repository to generate a large number of clusters.It generates clusters satisfying two or more objectives (not clustering level), and supports different domain specific tasks assigned by users.Is incremental approach that collects and refines clusters on the fly, and the search for alternative clusters takes into consideration what clusters already have been generated, rewarding novelty.Support different domain specific tasks assigned by users.Provide search engine type capabilities to users, enabling them to query a large set of clusters with respect to different objectives and thresholds to generate final clusterings.
The architecture of the MOC system that we propose is depicted in the figure given; it consists of 4 main components: a clustering algorithm, storage unit, goal-driven fitness function generator and cluster summarization unit. Steps in MOC include: first, the goal-driven fitness function generator selects a new compound fitness function for the clustering algorithm, which generates a new clustering X in the second step. Third, the storage unit updates its repository M using the clusters in X. The algorithm iterates over these three steps until a large number of clusters has been obtained. Later, in the fourth step, the cluster summarization unit produces final clusters based on user preferences which are subsets of the clusters in M.
In general, M should only store non-dominated clusters, and algorithms that update M should not violate this constraint
In particular to step 1, the goal-driven fitness function generator selects pairs of single-objectivefitness functions (q, q’) with q, q’Qto create a compound fitness function Q’ by considering all combinations of two fitness functions in Q. In general, the compound fitness function qQ’ is the sums of the rewards for all objectives qQ’; however, it gives more reward to clusters xi that satisfy all objectives in order to motivate theclustering algorithm to seek for multi-objective clusters.
The algorithm loops over the following 2 steps until there is no cluster to be removed in the second step:1. Compute dominant clusters which are the highest quality clusters, and2. remove their dominated clusters in the rem-proximity.We illustrate MO-DCR algorithm using dominance graphs.In the first iteration A is a dominant cluster, and its dominated clusters B and C, which are very similar to A (in close proximity), are removed.In the second iteration D is found as a dominant cluster, and its dominated clusters E and F, are removed.Consequently, the final clustering consists of two clusters: A and D.
Related work are Multi-objective clustering based on evolutionary algorithms and scatter tabu search.Similar to MOEA, our approach evaluates cluster quality by using a combination of fitness functions. In contrast, we selectively store and refine clusters on the fly to maintain efficacy of storage management and cluster comparison.
We demonstrate the benefits of the proposed multi-objective clustering framework in a real world water pollution case study. The goal is to find regional co-location patterns of high concentrations of Arsenic with other factors in Texas.TWDB has monitored water quality and collected the data for 105,814 wells in Texas over last 25 years. For this experiment we used a water well dataset called Arsenic_10_avg that contains3 spatial attributes and other 8 non-spatial attributes of chemical concentrations including Arsenic concentrations
Challenges in Multi-objective Clustering includeFind clusters that are individually good with respect to multiple fitness functions in automated fashion.Provide search engine style capabilities to summarize final clustering obtained from multiple runs of clustering algorithms.
We can categorize clustering algorithms into 4 types based on the use of fitness functions.First, the clustering algorithms that do not use fitness function, for example, DBSCAN and hierarchical clustering.Second, the algorithms that use fitness function implicitly, for example, k-means.Third, the algorithms that use fixed fitness function, for example, PAM.And Forth, the algorithms that provide plug-in fitness function, for example, CHAMELEON and MOSAIC.Using traditional clustering algorithms for region discovery face significant limitations:As we just mentioned, traditional clustering algorithms consider only domain independent and task independent characteristics; for example. k-means algorithm considers cluster compactness to form a solution. However, in region discovery, such as hotspot discovery of high arsenic concentration, it requires the clustering algorithm to find clusters based on task specific characteristics. Therefore, clustering with plug-in fitness functions is desirable because it can capture what groupings are of particular interest to domain experts. 2) To the best of our knowledge, there is no single fitness function that effectively serves different domain tasks. Therefore, it is necessary to design task specific families of fitness functions.
Step 4 provides search engine type capabilities to users, enabling them to query a large set of clusters with respect to different objectives and thresholds to generate final clusterings.Goal of MO-Dominance-guided Cluster Reduction algorithm (MO-DCR) isto return a clustering that is good with respect to multiple objectivesselected by a user, and to remove clusters that are highly overlapping.
DCR algorithm is one of effective algorithms that generate final clustering. Normally we can straightforwardly produce the result by removing all dominated clusters. However, this way will remove too many clusters, for example, clusters A dominates and overlaps B, and B dominates and overlaps C. DCR does not remove C because it does not spatially overlap with A.Finally, MRC perform the tasks of parameter selections, finding alternative clusters and summarizing clusters in a highly automated fashion.