Skip to content
New issue

Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.

By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.

Already on GitHub? Sign in to your account

Is there any plans to suuport GradientBoostingClassifier and CalibratedClassifier? #42

Open
odedfisc opened this issue Jan 3, 2019 · 4 comments

Comments

@odedfisc
Copy link

odedfisc commented Jan 3, 2019

No description provided.

@nok
Copy link
Owner

nok commented Jan 3, 2019

GradientBoostingClassifier: I will check and compare the final result of a GradientBoostingClassifier with other combined (ensemble) classifiers like ensemble.AdaBoostClassifier, ensemble.ExtraTreesClassifier or ensemble.RandomForestClassifier. If the computed results and data structure is the same, it will be simple to support and add this classifier.


CalibratedClassifierCV: The CalibratedClassifierCV is a cross validator estimator.

Probability calibration with isotonic regression or sigmoid.

See glossary entry for cross-validation estimator.

With this class, the base_estimator is fit on the train set of the cross->validation generator and the test set is used for calibration. The probabilities >for each of the folds are then averaged for prediction. In case that cv=”prefit” >is passed to init, it is assumed that base_estimator has been fitted already >and all data is used for calibration. Note that data for fitting the classifier >and for calibrating it must be disjoint.

So it's an encapsulation of the base_estimator. But today there is no automatic check, decapsulation and export of the base_estimator. But I will add it like the Pipeline, GridSearchCV or RandomizedSearchCV (source: Porter.py#L75-L87).

As a workaround you can try to assign the base estimator directly: Porter(clf.base_estimator)

@odedfisc
Copy link
Author

Many thanks

@beojan
Copy link

beojan commented Feb 20, 2019

Has there been any progress on this?

@loftalert
Copy link

I would also be interested in a port of XGBoost, an algorithm that regularly beats the other ensemble methods. After all, the package provides a sklearn API so it sort of belongs to the sklearn model zoo. Any plans on that?

Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment
Projects
None yet
Development

No branches or pull requests

4 participants