Take a look at scipy's fcluster function.
If M is a matrix of all of your feature vectors, this code snippet should
work.
You need to figure out what metric and algorithm work for you
from sklearn.metrics import pairwise_distance
from scipy.cluster import hierarchy
X = pairwise_distance(M, metric=metric)
Z = hierarchy.linkage(X, algo, metric=metric)
C = hierarchy.fcluster(Z,threshold, criterion="distance")
Best,
Uri Goren
On Tue, Jul 11, 2017 at 7:42 PM, Ariani A <[email protected]> wrote:
> Hi all,
> I want to perform agglomerative clustering, but I have no idea of number
> of clusters before hand. But I want that every cluster has at least 40
> data points in it. How can I apply this to sklearn.agglomerative clusteri
> ng?
> Should I use dendrogram and cut it somehow? I have no idea how to relate
> dendrogram to this and cutting it out. Any help will be appreciated!
> I have to use agglomerative clustering!
> Thanks,
> -Ariani
>
> _______________________________________________
> scikit-learn mailing list
> [email protected]
> https://mail.python.org/mailman/listinfo/scikit-learn
>
>
--
*Uri Goren,Software innovator*
*Phone: +972-507-649-650*
*EMail: [email protected] <[email protected]>*
*Linkedin: il.linkedin.com/in/ugoren/ <http://il.linkedin.com/in/ugoren/>*
_______________________________________________
scikit-learn mailing list
[email protected]
https://mail.python.org/mailman/listinfo/scikit-learn