Skip to content
New issue

Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.

By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.

Already on GitHub? Sign in to your account

[ENH] Gradient Boosting: New widget #5160

Merged
merged 11 commits into from
Jan 15, 2021
Merged

[ENH] Gradient Boosting: New widget #5160

merged 11 commits into from
Jan 15, 2021

Conversation

VesnaT
Copy link
Contributor

@VesnaT VesnaT commented Jan 5, 2021

Issue

Implements #5107
Alternative to #5116

Description of changes
  • wrap sklearn gradient boosting classification and regression learners
  • wrap catboost classification and regression learners
  • wrap xgboost classification and regression learners
  • Gradient Boosting widget
Includes
  • Code changes
  • Tests
  • Documentation

@codecov
Copy link

codecov bot commented Jan 5, 2021

Codecov Report

Merging #5160 (bfffd6e) into master (ef789ee) will increase coverage by 0.20%.
The diff coverage is 97.80%.

@@            Coverage Diff             @@
##           master    #5160      +/-   ##
==========================================
+ Coverage   84.77%   84.98%   +0.20%     
==========================================
  Files         286      296      +10     
  Lines       60084    60643     +559     
==========================================
+ Hits        50938    51539     +601     
+ Misses       9146     9104      -42     

@VesnaT VesnaT force-pushed the gbtrees branch 2 times, most recently from 13b88c6 to f2da44e Compare January 6, 2021 08:41
@VesnaT VesnaT changed the title [WIP] Gradient Boosting: Wrapped models, widget [ENH] Gradient Boosting: Wrapped models, widget Jan 7, 2021
@lanzagar
Copy link
Contributor

lanzagar commented Jan 8, 2021

Works well for me.
I would suggest:

  • being explicit for the default "Gradient Boosting" method as well and changing to "Gradient Boosting (scikit-learn)" to be more consistent with other entries in the dropdown
  • changing the default num of trees for catboost from 500 to 100 so it is the same as all the other methods - catboost is already the slowest (might need additional investigation)
  • change default value of Replicable training to True. BTW is this really only available from scikit-learn, do xgboost and catboost not allow setting random seeds?

@VesnaT
Copy link
Contributor Author

VesnaT commented Jan 11, 2021

  • change default value of Replicable training to True. BTW is this really only available from scikit-learn, do xgboost and catboost not allow setting random seeds?

I think so. Both (xgboost and catboost) have 0 as default. If one passes None, it is changed to 0.

@lanzagar lanzagar changed the title [ENH] Gradient Boosting: Wrapped models, widget [ENH] Gradient Boosting: New widget Jan 15, 2021
@lanzagar lanzagar merged commit 1165816 into biolab:master Jan 15, 2021
Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment
Labels
None yet
Projects
None yet
Development

Successfully merging this pull request may close these issues.

2 participants