Hôm nay, Thứ 7 27/04/24 23:28

Thời gian được tính theo giờ UTC + 7 Giờ




Tạo chủ đề mới Gửi bài trả lời  [ 1 bài viết ] 
Người gửi Nội dung
Gửi bàiĐã gửi: Thứ 3 28/11/23 16:35 
Ngoại tuyến

Ngày tham gia: Thứ 4 17/08/22 14:49
Bài viết: 12
Hierarchical clustering and k-means based clustering are two common methods that are used in data analysis as well as machine learning to cluster related data points. Both methods aim to identify clusters in a data set but they differ in the way they approach and the type of clusters they create. This article we'll examine the differences between hierarchical clustering and K-means clustering in depth.
Data Science Course in Pune

Hierarchical Clustering Hierarchical clustering can be described as an approach from the bottom up that is also referred to as agglomerative clumping. It begins by treating each data point as separate cluster. It then joins the most close clusters in a series of iterative steps until a single cluster is left. This process creates a hierarchical structure for clusters, which is often depicted as dendrograms.

Two primary kinds of hierarchical clustering:
Agglomerative clustering This starts by treating every data point being an individual cluster, and then gradually merges the clusters closest to it until there is only one cluster left. The merging is dependent on the measure of dissimilarity or similarity between clusters, including Euclidean distance, or correlation coefficients.

Dividesive Clustering The process begins with the entire set of the data points of the same cluster and splits them up into smaller clusters until every data point is located in their own group. This approach is more uncommon and more expensive computationally in comparison to agglomerative aggregation.

Hierarchical clustering doesn't need a predetermined number of clusters as it establishes a cluster hierarchy which allows for various levels of detail. It provides an illustration of the clustering process using the dendrogram. This could be helpful in exploratory analysis and finding the ideal quantity of clusters.

K-means Clustering: K means clustering is an iterative method of partitioning an entire dataset into a set quantity (k) of exclusive mutually bonded clusters. It's aim is to minimize the amount of distances that are squared between the points of data and their respective cluster centersoids. The algorithm operates in the following manner the following: Data Science Education in Pune
Initialization Choose randomly k points of data to be the initial centroids.
Assignment Each data object is assigned to the closest centroid using the distance metric, usually Euclidean distance.
Changes: The centroids are recalculated by taking the median for all points that are assigned for each cluster.
Repetition: Iterate steps 2 and 3 until convergence when the centroids don't change significantly, or until a maximum number of repetitions is reached.

K-means clustering needs that the amount of clusters that must be defined beforehand, which may be a problem. It is an algorithm that is more efficient in terms of computation as compared to hierarchical, which makes it suitable for large data sets.
Differentialities: Now that we have covered the fundamentals of hierarchical and k'means clustering Let's look at the major distinctions between these two approaches:

The nature of clusters Hierarchical clustering is the structure of clusters in a hierarchical manner which allows for various levels of granularity, and also recording complicated relationships. K-means clustering creates distinct non-overlapping clusters that are dependent on the amount of clusters defined. Data Science Courses in Pune

number of clusters The hierarchical model does not need an amount of clusters to be defined because it creates an dendrogram that is cut at various heights to create varying numbers of clusters. K-means clustering is, however is dependent on a predetermined amount of clusters.
Computing: Hierarchical clustering can be costly computationally, particularly for large amounts of data because the algorithm has to determine the pairwise distances between each data element. K-means clustering is generally more efficient in computational efficiency because of its iterative nature and less complicated distance calculation.

The shape of the cluster: Hierarchical clustering can manage clusters of different sizes and shapes, including non-convex and irregular clusters. K-means clustering presumes that clusters are isotropic and spherical. This means that it can be a challenge for clusters with different sizes or shapes.
Top


Đầu trang
 Xem thông tin cá nhân  
 
Hiển thị những bài viết cách đây:  Sắp xếp theo  
Tạo chủ đề mới Gửi bài trả lời  [ 1 bài viết ] 

Thời gian được tính theo giờ UTC + 7 Giờ


Ai đang trực tuyến?

Đang xem chuyên mục này: Bing [Bot]24 khách


Bạn không thể tạo chủ đề mới trong chuyên mục này.
Bạn không thể trả lời bài viết trong chuyên mục này.
Bạn không thể sửa những bài viết của mình trong chuyên mục này.
Bạn không thể xoá những bài viết của mình trong chuyên mục này.
Bạn không thể gửi tập tin đính kèm trong chuyên mục này.

Tìm kiếm với từ khoá:
Chuyển đến:  
Chứng nhận thanh toán bảo đảm

CÔNG TY TNHH DỊCH VỤ HÀNG HOÁ TRỰC TUYẾN

Diễn đàn sử dụng phần mềm phpBB® Forum


Chợ xây dựng Hà nội: Công ty TNHH Thương mại Dương Linh

Showroom: Số 1B, Ngõ 5, tổ 19, thị trấn Cầu Diễn, Huyện Từ Liêm, Thành phố Hà nội

Tel: 04.37737548; Fax: 04.38370082

Email Phòng kinh doanh: sale@choxaydung.vn

Chợ xây dựng Sài Gòn: Công ty TNHH SX-TM Đăng Hải

Địa chỉ: Số 140, Tô Hiến Thành, P15, Quận 10, Tp.HCM

Tel: 08.38620524; Fax: 08.38633011;

Email: saigonRep@choxaydung.vn