Agglomerative Clustering.
Recursively merges pair of clusters of sample data; uses linkage distance.
Read more in the User Guide.
The number of clusters to find. It must be None if distance_threshold is not None.
Metric used to compute the linkage. Can be “euclidean”, “l1”, “l2”, “manhattan”, “cosine”, or “precomputed”. If linkage is “ward”, only “euclidean” is accepted. If “precomputed”, a distance matrix is needed as input for the fit method. If connectivity is None, linkage is “single” and affinity is not “precomputed” any valid pairwise distance metric can be assigned.
Added in version 1.2.
Used to cache the output of the computation of the tree. By default, no caching is done. If a string is given, it is the path to the caching directory.
Connectivity matrix. Defines for each sample the neighboring samples following a given structure of the data. This can be a connectivity matrix itself or a callable that transforms the data into a connectivity matrix, such as derived from kneighbors_graph. Default is None, i.e, the hierarchical clustering algorithm is unstructured.
For an example of connectivity matrix using kneighbors_graph, see Agglomerative clustering with and without structure.
Stop early the construction of the tree at n_clusters. This is useful to decrease computation time if the number of clusters is not small compared to the number of samples. This option is useful only when specifying a connectivity matrix. Note also that when varying the number of clusters and using caching, it may be advantageous to compute the full tree. It must be True if distance_threshold is not None. By default compute_full_tree is “auto”, which is equivalent to True when distance_threshold is not None or that n_clusters is inferior to the maximum between 100 or 0.02 * n_samples. Otherwise, “auto” is equivalent to False.
Which linkage criterion to use. The linkage criterion determines which distance to use between sets of observation. The algorithm will merge the pairs of cluster that minimize this criterion.
Added in version 0.20: Added the ‘single’ option
For examples comparing different linkage criteria, see Comparing different hierarchical linkage methods on toy datasets.
The linkage distance threshold at or above which clusters will not be merged. If not None, n_clusters must be None and compute_full_tree must be True.
Added in version 0.21.
Computes distances between clusters even if distance_threshold is not used. This can be used to make dendrogram visualization, but introduces a computational and memory overhead.
Added in version 0.24.
For an example of dendrogram visualization, see Plot Hierarchical Clustering Dendrogram.
The number of clusters found by the algorithm. If distance_threshold=None, it will be equal to the given n_clusters.
Cluster labels for each point.
Number of leaves in the hierarchical tree.
The estimated number of connected components in the graph.
Added in version 0.21: n_connected_components_ was added to replace n_components_.
Number of features seen during fit.
Added in version 0.24.
n_features_in_,)
Names of features seen during fit. Defined only when X has feature names that are all strings.
Added in version 1.0.
The children of each non-leaf node. Values less than n_samples correspond to leaves of the tree which are the original samples. A node i greater than or equal to n_samples is a non-leaf node and has children children_[i - n_samples]. Alternatively at the i-th iteration, children[i][0] and children[i][1] are merged to form node n_samples + i.
Distances between nodes in the corresponding place in children_. Only computed if distance_threshold is used or compute_distances is set to True.
See also
FeatureAgglomerationAgglomerative clustering but for features instead of samples.
ward_treeHierarchical clustering with ward linkage.
>>> from sklearn.cluster import AgglomerativeClustering >>> import numpy as np >>> X = np.array([[1, 2], [1, 4], [1, 0], ... [4, 2], [4, 4], [4, 0]]) >>> clustering = AgglomerativeClustering().fit(X) >>> clustering AgglomerativeClustering() >>> clustering.labels_ array([1, 1, 1, 0, 0, 0])
Fit the hierarchical clustering from features, or distance matrix.
Training instances to cluster, or distances between instances if metric='precomputed'.
Not used, present here for API consistency by convention.
Returns the fitted instance.
Fit and return the result of each sample’s clustering assignment.
In addition to fitting, this method also return the result of the clustering assignment for each sample in the training set.
Training instances to cluster, or distances between instances if affinity='precomputed'.
Not used, present here for API consistency by convention.
Cluster labels.
Get metadata routing of this object.
Please check User Guide on how the routing mechanism works.
A MetadataRequest encapsulating routing information.
Get parameters for this estimator.
If True, will return the parameters for this estimator and contained subobjects that are estimators.
Parameter names mapped to their values.
Set the parameters of this estimator.
The method works on simple estimators as well as on nested objects (such as Pipeline). The latter have parameters of the form <component>__<parameter> so that it’s possible to update each component of a nested object.
Estimator parameters.
Estimator instance.
© 2007–2025 The scikit-learn developers
Licensed under the 3-clause BSD License.
https://scikit-learn.org/1.6/modules/generated/sklearn.cluster.AgglomerativeClustering.html