Ensemble of different kinds of regressors using scikit-learn (or any other python framework)

后端 未结 4 1774
猫巷女王i
猫巷女王i 2021-01-30 05:43

I am trying to solve the regression task. I found out that 3 models are working nicely for different subsets of data: LassoLARS, SVR and Gradient Tree Boosting. I noticed that w

4条回答
  •  刺人心
    刺人心 (楼主)
    2021-01-30 06:22

    Late response, but I wanted to add one practical point for this sort of stacked regression approach (which I use this frequently in my work).

    You may want to choose an algorithm for the stacker which allows positive=True (for example, ElasticNet). I have found that, when you have one relatively stronger model, the unconstrained LinearRegression() model will often fit a larger positive coefficient to the stronger and a negative coefficient to the weaker model.

    Unless you actually believe that your weaker model has negative predictive power, this is not a helpful outcome. Very similar to having high multi-colinearity between features of a regular regression model. Causes all sorts of edge effects.

    This comment applies most significantly to noisy data situations. If you're aiming to get RSQ of 0.9-0.95-0.99, you'd probably want to throw out the model which was getting a negative weighting.

提交回复
热议问题