Gradient boosting on Vowpal Wabbit

只谈情不闲聊 提交于 2019-11-30 05:17:21

问题


Is there a way to use gradient boosting on regression using Vowpal Wabbit? I use various techniques that come with Vowpal Wabbit that are helpful. I want to try gradient boosting along with that, but I can't find a way to implement gradient boosting on VW.


回答1:


The idea of gradient boosting is that an ensemble model is built from black-box weak models. You can surely use VW as the black box, but note that VW does not offer decision trees, which are the most popular choice for the black-box weak models in boosting. Boosting in general decreases bias (and increases variance), so you should make sure that the VW models have low variance (no overfitting). See bias-variance tradeoff.

There are some reductions related to boosting and bagging in VW:

  • --autolink N adds a link function with polynomial N, which can be considered a simple way of boosting.
  • --log_multi K is an online boosting algorithm for K-class classification. See the paper. You can use it even for binary classification (K=2), but not for regression.
  • --bootstrap M M-way bootstrap by online importance resampling. Use --bs_type=vote for classification and --bs_type=mean for regression. Note that this is bagging, not boosting.
  • --boosting N (added on 2015-06-17) online boosting with N weak learners, see a theoretic paper


来源:https://stackoverflow.com/questions/30008991/gradient-boosting-on-vowpal-wabbit

易学教程内所有资源均来自网络或用户发布的内容,如有违反法律规定的内容欢迎反馈
该文章没有解决你所遇到的问题?点击提问,说说你的问题,让更多的人一起探讨吧!