However, there is some issue to frozen a fitted classifier. You can refer to:
https://github.com/scikit-learn/scikit-learn/issues/8370 with the associated discussion. On Wed, 19 Sep 2018 at 17:34, Guillaume Lemaître <g.lemaitr...@gmail.com> wrote: > > Ups I misread your comment. I don't think that we have currently a > mechanism to avoid cloning classifier internally. > On Wed, 19 Sep 2018 at 17:31, Guillaume Lemaître <g.lemaitr...@gmail.com> > wrote: > > > > You don't have anywhere in your class MyClassifier where you are > > calling base_classifier.fit(...) therefore when calling > > base_classifier.predict(...) it will let you know that you did not fit > > it. > > > > On Wed, 19 Sep 2018 at 16:43, Luiz Gustavo Hafemann <luiz...@gmail.com> > > wrote: > > > > > > Hello, > > > > > > I am one of the developers of a library for Dynamic Ensemble Selection > > > (DES) methods (the library is called DESlib), and we are currently > > > working to get the library fully compatible with scikit-learn (to submit > > > it to scikit-learn-contrib). We have "check_estimator" working for most > > > of the classes, but now I am having problems to make the classes > > > compatible with GridSearch / other CV functions. > > > > > > One of the main use cases of this library is to facilitate research on > > > this field, and this led to a design decision that the base classifiers > > > are fit by the user, and the DES methods receive a pool of base > > > classifiers that were already fit (this allow users to compare many DES > > > techniques with the same base classifiers). This is creating an issue > > > with GridSearch, since the clone method (defined in sklearn.base) is not > > > cloning the classes as we would like. It does a shallow (non-deep) copy > > > of the parameters, but we would like the pool of base classifiers to be > > > deep-copied. > > > > > > I analyzed this issue and I could not find a solution that does not > > > require changes on the scikit-learn code. Here is the sequence of steps > > > that cause the problem: > > > > > > GridSearchCV calls "clone" on the DES estimator (link) > > > The clone function calls the "get_params" function of the DES estimator > > > (link, line 60). We don't re-implement this function, so it gets all the > > > parameters, including the pool of classifiers (at this point, they are > > > still "fitted") > > > The clone function then clones each parameter with safe=False (line 62). > > > When cloning the pool of classifiers, the result is a pool that is not > > > "fitted" anymore. > > > > > > The problem is that, to my knowledge, there is no way for my classifier > > > to inform "clone" that a parameter should be always deep copied. I see > > > that other ensemble methods in sklearn always fit the base classifiers > > > within the "fit" method of the ensemble, so this problem does not happen > > > there. I would like to know if there is a solution for this problem while > > > having the base classifiers fitted elsewhere. > > > > > > Here is a short code that reproduces the issue: > > > > > > --------------------------- > > > > > > from sklearn.model_selection import GridSearchCV, train_test_split > > > from sklearn.base import BaseEstimator, ClassifierMixin > > > from sklearn.ensemble import BaggingClassifier > > > from sklearn.datasets import load_iris > > > > > > > > > class MyClassifier(BaseEstimator, ClassifierMixin): > > > def __init__(self, base_classifiers, k): > > > self.base_classifiers = base_classifiers # Base classifiers that > > > are already trained > > > self.k = k # Simulate a parameter that we want to do a grid > > > search on > > > > > > def fit(self, X_dsel, y_dsel): > > > pass # Here we would fit any parameters for the Dynamic > > > selection method, not the base classifiers > > > > > > def predict(self, X): > > > return self.base_classifiers.predict(X) # In practice the > > > methods would do something with the predictions of each classifier > > > > > > > > > X, y = load_iris(return_X_y=True) > > > X_train, X_dsel, y_train, y_dsel = train_test_split(X, y, test_size=0.5) > > > > > > base_classifiers = BaggingClassifier() > > > base_classifiers.fit(X_train, y_train) > > > > > > clf = MyClassifier(base_classifiers, k=1) > > > > > > params = {'k': [1, 3, 5, 7]} > > > grid = GridSearchCV(clf, params) > > > > > > grid.fit(X_dsel, y_dsel) # Raises error that the bagging classifiers are > > > not fitted > > > > > > --------------------------- > > > > > > Btw, here is the branch that we are using to make the library compatible > > > with sklearn: https://github.com/Menelau/DESlib/tree/sklearn-estimators. > > > The failing test related to this issue is in > > > https://github.com/Menelau/DESlib/blob/sklearn-estimators/deslib/tests/test_des_integration.py#L36 > > > > > > Thanks in advance for any help on this case, > > > > > > Luiz Gustavo Hafemann > > > > > > _______________________________________________ > > > scikit-learn mailing list > > > scikit-learn@python.org > > > https://mail.python.org/mailman/listinfo/scikit-learn > > > > > > > > -- > > Guillaume Lemaitre > > INRIA Saclay - Parietal team > > Center for Data Science Paris-Saclay > > https://glemaitre.github.io/ > > > > -- > Guillaume Lemaitre > INRIA Saclay - Parietal team > Center for Data Science Paris-Saclay > https://glemaitre.github.io/ -- Guillaume Lemaitre INRIA Saclay - Parietal team Center for Data Science Paris-Saclay https://glemaitre.github.io/ _______________________________________________ scikit-learn mailing list scikit-learn@python.org https://mail.python.org/mailman/listinfo/scikit-learn