Loading...
Please wait, while we are loading the content...
Similar Documents
Comparative Study of Various Clustering Techniques in Data Mining
| Content Provider | Semantic Scholar |
|---|---|
| Author | Soni, Anu Goel, Mukta Goel, Rohit |
| Copyright Year | 2015 |
| Abstract | Data mining is used to find the hidden information pattern and relationship between the large data set which is very useful in decision making. Clustering is an automatic unsupervised learning technique which partitions a data set into several groups based on the principle of maximizing the intraclass similarity and minimizing the inter-class similarity. This paper analyze the three major clustering algorithms: Partition clustering, Hierarchical clustering and Density based clustering algorithm and compare the performances of these three major clustering algorithms I.INTRODUCTION Data Mining is one of the important steps for mining or extracting a great deal of information. It is designed to explore giant amount of information in search of consistent patterns and to validate the results by the detected patterns to the new subset of information. Clustering is a data mining technique of grouping set of data instances into multiple groups or clusters so that objects within the cluster are similar with each other, but are very different to objects in the other clusters. Homogenous data and Heterogeneous data are assessed based on the properties of the objects or instances. Homogenous data is contained in a cluster, but it is heterogeneous data from another cluster's data. A data’s cluster is chosen according to attribute values describing by objects .Clustering algorithms are used to data organisation, data categorization, data compression, construction of model and for detection of outliers etc. II. WHY CLUSTERING Data Clustering is one of the challenging mining techniques in the knowledge data discovery process. Clustering huge amount of data is a difficult task since the goal is to find a suitable partition in an unsupervised way i.e. without any prior knowledge trying to maximize the intra-cluster similarity and minimize inter-cluster similarity which in turn maintains high cluster cohesiveness. The instances are thereby organized into an efficient representation that characterizes the population being sampled. Thus the output of cluster analysis is the number of groups or clusters that form the structure of partitions, of the data set. By clustering the instances, we could reduce our search domain for recommendations as most of the users are interested in the instances corresponding to a few numbers of clusters. This could improve the result of time efficiency to a greater extent and would also help in identification of same news from different sources. The main motivation is to investigate possible improvements of the effectiveness of document III.TYPES OF CLUSTERING Many different clustering techniques have been defined in order to solve the problem from different perspective ,these techniques are: Partition Clustering Density based Clustering Hierarchical clustering A) PARTITION CLUSTERING Partition clustering is known to be the most popular class of clustering algorithm also said to be iterative relocation algorithm. These algorithms minimize a given clustering criterion by iteratively relocating data points between clusters until a optimal partition is attained. Partition clustering decomposes a data set into a set of disjoint clusters. Given a data set of N points, a partitioning method constructs K (N ≥ K) partitions of the data, with each partition representing a cluster. That is, it classifies the data into various K groups by satisfying the following requirements: (1) each group contains at least one point, and (2) each point belongs to exactly one group. Notice that for fuzzy partitioning, a point can belong to one or more than one group. This method is Proceedings of National Conference on Innovative Trends in Computer Science Engineering (ITCSE-2015) held at BRCMCET, Bahal on 4 April 2015 IJRRA ISSN: 2349-7688 151 effective for small to medium sized data sets. Examples of partitioning methods include k-means and k-medoids . K-MEANS ALGORITHM It is a centred based technique. This algorithm takes the input parameters k and partition a set of n objects into k clusters that the resulting object within cluster are of same kind whereas objects outside the cluster are of different kind. The method can be used by cluster to assign rank values to the cluster categorical data is statistical method .in reality; K mean is mainly based on the distance between the object and the cluster mean. Then it computes the new mean for each cluster .Here categorical data have been changed into numerical values by assigning rank to them. K-Means Algorithm Properties 1. There are always K clusters. 2. There is always at least one item in each cluster. 3. The clusters are non-hierarchical and they do not overlap. 4. Every member of a cluster is closer to its cluster than any other cluster 5. Closeness does not always involve the 'centre' of clusters. K-Means Algorithm Process 1. The dataset is partitioned into K clusters and the data points are randomly assigned to the clusters resulting in clusters that have roughly the same number of data points. 2. For each data point: 3. Calculate the distance from the data point to each cluster 4. If the data point is closest to its own cluster, leave it on its own position; otherwise move data points into the closest cluster. 5. Repeat the above step until a complete pass through all the data points result in no further movements of data points among clusters. At this point the clusters are still and the clustering process ends. 6. Initial partition choice can greatly affect the final clusters that result, in to the terms of inter-cluster and intra cluster distances and cohesion. K-MEDOIDS ALGORITHM This is a variation of the k-means algorithm and is less sensitive to outliers. In this instead of mean we use the actual object to represent the cluster, using one representative object per cluster. Clusters are generated by points which are close to respective methods. The function used for classification is a measure of dissimilarities of points in a cluster and their representative. The partitioning is done based on minimizing the sum if the dissimilarities between each object and its cluster representative. This criterion is called as absoluteerror criterion. N Sum of Absolute error=Σ Σ Dist (p, a) i=1 p ∈ Ci Where p represents an object in the data set and oi is the ith representative. N is the number of clusters .Two well-known types of k-medoids clustering are the PAM (Partitioning Around Medoids) and CLARA (Clustering Large Applications). B) HIERARCHICAL CLUSTERING A hierarchical decomposition of the given set of data objects is created in hierarchal clustering. Here tree of clusters are built. These trees are known as dendrogram . Every cluster node contains child clusters. Sibling clusters partition the points covered by their common parent into clusters. In hierarchical clustering we assign each item to a cluster such that if we have N items then we have N clusters. Find pair of clusters which are nearby and merge them into a single cluster. Find out the distance between new cluster and each of old clusters. Repeat above steps until all items are clustered into K no. of clusters. Hierarchal clustering is of two types: Agglomerative (bottom up)Agglomerative hierarchical clustering is a bottom-up clustering method .in this clustering technique clusters have sub-clusters, which their selves have sub-clusters, etc. It starts by letting each instance built its own cluster and recursively merges cluster into larger and larger clusters, until all the instances are in a single cluster or certain termination condition is satisfied. The single cluster becomes the hierarchy’s root i.e. upper most node. For the merging step, it finds the two different clusters that are closest to each other, and combines the two to built one cluster. Divisive (top down)A top-down clustering method and is less commonly used. It works in the same way as agglomerative clustering but this technique work in the opposite direction. This method starts with a single cluster containing all instances, and then successively divides into resulting clusters until only clusters of individual instances remain only. C) DENSITY BASED CLUSTERING Density-based clustering algorithms are use to build arbitrary-shaped clusters. In this approach, a cluster is regarded as a region in which the density of data objects exceeds a threshold. DBSCAN and OPTICS are typical algorithms of this type. DBSCAN Clustering DBSCAN (Density Based Spatial Clustering of Application with Noise).It grows clusters according to the density of neighbourhood instances. It is based on the concept of “density reachibility” and “density connect ability”, these two depends upon input parametersize of epsilon neighbourhood and minimum terms of local distribution of nearest neighbours. Proceedings of National Conference on Innovative Trends in Computer Science Engineering (ITCSE-2015) held at BRCMCET, Bahal on 4 April 2015 IJRRA ISSN: 2349-7688 152 Here size of neighbourhood and size of clusters are control by input parameters. It starts with an arbitrary starting point that has not been visited yet. The point’s e-neighbourhood is retrieved, and if it contains sufficiently many points, a cluster is started. Otherwise the point is labelled as noise. The number of point parameter impacts detection of outliers. DBSCAN targeting low-dimensional spatial data used DENCLUE algorithm DBSCAN separates data points into three classes: Core points: These are points that are at the internal side of a cluster. Border points: A border point is a point that is not a core point, but it falls within the boundary of a core point. Noise points: A noise point is any point that is not a core point or a border point. To find a cluster, DBSCAN starts with an arbitrary instance (p) in data set (D) and retrieves all instances of D with respect to Eps and Min Pts. The algorithm makes use of a spatial data structure(R*tree) to locate points within Eps distance from the core |
| File Format | PDF HTM / HTML |
| Alternate Webpage(s) | https://www.ijrra.net/NCITCSE2015/NCITCSE43.pdf |
| Language | English |
| Access Restriction | Open |
| Content Type | Text |
| Resource Type | Article |