-
Notifications
You must be signed in to change notification settings - Fork 10
New issue
Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.
By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.
Already on GitHub? Sign in to your account
consider how we can bring more lightgbm parameters #23
Comments
For completeness, not all of the parameters you'd be interested in tweaking are available for tweaking. An important one comes to mind: |
Defaults might not be matching main LightGBM: #42 (comment) |
Btw, Caret has an interface to many R packages, just like MLJ. Might this be worth doing w/ your interface to MLJ? |
In implementation terms, doing this is likely to result in massive amounts of duplication of code. If you see recent commit ee01161 that we actually went in opposite direction to this, by removing object-wise distinction between binary-multiclass (which just corresponds to a change in the value for However, I agree that the missing support for the boosting parameter needs to be added. |
You’re probably right. xgbDART, xgbLinear, xgbTree |
Just to let you know, I haven't forgotten this. I've been buried with other work, but I'm hoping in the coming weeks to bundle a few features together and create a new release. |
i found these important parameters missing: boosting: defines the type of algorithm you want to run, default=gdbt max_cat_group: When the number of category is large, finding the split point on it is easily over-fitting. So LightGBM merges them into ‘max_cat_group’ groups, and finds the split points on the group boundaries, default:64 application: This is the most important parameter and specifies the application of your model, whether it is a regression problem or classification problem. LightGBM will by default consider model as a regression model.(while we have created two separately for regression and classification, i found where the two are defined under this) num_boost_round: Number of boosting iterations, typically 100+ ignore_column: same as categorical_features just instead of considering specific columns as categorical, it will completely ignore them. |
num_boost_round is an alias for num_iterations which we already have I am not yet sure what to do about supporting all of the aliases, because you cannot reasonably alias struct fields (that I know of), and reproducing their aliasing logic seems like a bit of a wasted effort |
As for application you can see it is supported: https://github.com/IQVIA-ML/LightGBM.jl/blob/master/src/estimators.jl#L7 |
#62 PR just merged, brings support for boosting parameter and those related to DART and GOSS |
#79 merged, brings quite a few additional parameters. Will become available hopefully soon when we cut the next release. |
https://lightgbm.readthedocs.io/en/latest/Parameters.html
The text was updated successfully, but these errors were encountered: