Skip to content
New issue

Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.

By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.

Already on GitHub? Sign in to your account

Please add CatBoost or any alternate package (pure Julia) which can beat it #992

Closed
MrBenzWorld opened this issue Nov 23, 2022 · 9 comments

Comments

@MrBenzWorld
Copy link

I'm committed to learn julia . I have tried MLJFLUX, BetaML, MLJ models, Evotrees etc...

but, nothing is giving similar to Catboost Performance and quality of results (MLJFLUX is near but computaionally expesnive).

Im testing for regression and I need to get high quality results for research pulication.

I appriciate entire julia team and MLJ. I like it.
I hope, you will consider my request.
Thank you

@ablaom
Copy link
Member

ablaom commented Nov 23, 2022

There is a discussion about this here: JuliaAI/CatBoost.jl#9 .

Seems a pity you are not able to get what you want from EvoTrees.jl, which should be similar to CatBoost but pure julia.
The main developer of EvoTrees.jl is quite active and, it seems to me, open to feature requests.

In my view, it's a better use of limited resources to improve pure Julia implementations than wrapping python/C implementations. And in the case of gradient tree boosters, we already have julia and MLJ interfaces for XGBoost and LightGBM. Do we really need a 4th tree booster?

Of course if someone is interested in an MLJ interface for CatBoost.jl, then I am happy to provide guidance.

@MrBenzWorld
Copy link
Author

MrBenzWorld commented Nov 24, 2022

I also thought, EvoTrees is a good option for Julia.

I request, please add EvoTrees with Treezen optimization with MLJ.

Also, show one example that, it increased the quality of results. Sorry for the discussion. But ,it is Julia learner problems. Once, if I get proper results, I will continue with it ,otherwise, I'm forced by python again.

You can use Boston, or AMES MLJ datasets for implementation.
@ablaom

@ablaom
Copy link
Member

ablaom commented Nov 24, 2022

I think these various tools are individually well-documented. If you have a specific tutorial you'd like to see, please make a request at https://github.com/JuliaAI/DataScienceTutorials.jl/issues

@MrBenzWorld
Copy link
Author

Yes, they are well documented individually. It is really helpful for beginners and it helped me a lot.

But there is a problem getting the best results.

We have chosen Julia for fast and best results.

MLJ is best with all tools ( specially with pure Julia packages) ,

MLJ+evotrees+Treezen (or Latin etc)
There is no example.
This example should show best performance with great accurate results. ( time and results ).

It should beat python based XGBoost ,LightGboost optuna optimization.

You have given an example of how to use it. But didn't show ,how to get best.

As a Julia learner, it is a suggestion.
Julia don't have access to kaggle ,but you have Julia cloud or colab . somewhere ,you can also share competition notebooks.

Thanks for your efforts and cooperation @ablaom

@MrBenzWorld
Copy link
Author

Yes, I will do the request as you mentioned.
Thank you.

@ablaom
Copy link
Member

ablaom commented Jan 18, 2023

JuliaAI/CatBoost.jl#16

@MrBenzWorld
Copy link
Author

Thank you very much .@ablaom

@ablaom
Copy link
Member

ablaom commented Jan 19, 2023

You're welcome. 'm just providing guidance. The main work is being carried out by @tylerjthomas9.

@ablaom
Copy link
Member

ablaom commented Feb 8, 2023

Closed as completed: https://github.com/JuliaAI/CatBoost.jl#mlj-example

@ablaom ablaom closed this as completed Feb 8, 2023
Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment
Labels
None yet
Projects
None yet
Development

No branches or pull requests

2 participants