Skip to content
New issue

Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.

By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.

Already on GitHub? Sign in to your account

Why should I use InfiniBoost? #3

Open
hrstoyanov opened this issue Jun 12, 2017 · 2 comments
Open

Why should I use InfiniBoost? #3

hrstoyanov opened this issue Jun 12, 2017 · 2 comments

Comments

@hrstoyanov
Copy link

I read the paper (thanks) but I am still puzzled - I don't see any ground-breaking improvements in precision or performance over RF or GB? What is the big benefit?

Thanks

@arogozhnikov
Copy link
Owner

arogozhnikov commented Jun 13, 2017

Hi Hristo,
indeed, there is no ground-breaking improvement. However in both cases algorithm was compared on the problems where competitors are very good and using tree hyper-parameters that are known to be appropriate for competitors (RF and GB are used with very different kinds of trees).
Since this is one algorithm, I think it is a good result.

While I'm far from insisting that someone should use this approach, the following points look very interesting to me

  • in this modification boosting becomes a converging algorithm. From theoretical point it is a nice property
  • we wanted to reproduce the behavior of random forest, when one can simply leave ensemble training for hours and not to worry that 'there are too many trees', and convergence made this possible
  • we replaced GB's shrinkage parameter with a capacity parameter, which characterizes the state the algorithm convergences to (thus, can be changed at any moment). Tuning of learning rate (+ number of trees) is a common procedure for GB, which we tried to avoid - in InfiniteBoost we introduced automated search for capacity. This works in some cases.

@hrstoyanov
Copy link
Author

Thanks for the clarifications!

Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment
Labels
None yet
Projects
None yet
Development

No branches or pull requests

2 participants