You signed in with another tab or window. Reload to refresh your session.You signed out in another tab or window. Reload to refresh your session.You switched accounts on another tab or window. Reload to refresh your session.Dismiss alert
Is there a way to change safely weights after each iteration based in some user custom logic?
This would be my use case:
A problem where the data is grouped by geographic area and the gradient boosting model consistently predicts some areas worse than others. I prefer to sacrifice some of the performance of the easier to predict zones in exchange for improving the predictions of the difficult zones, so reducing the variability of the errors per zone is a priority.
One approach is to fit a model and, based on the results, over-weight the 'difficult' regions, and do this several times.
What I want to do is introduce the process into the model fitting by doing upweighting after each iteration, evaluating all predictions by geographic area (the area is passed to dtrain with attr() and the performance of each area is assessed with a groupby on the absolute value of the prediction errors of the samples they include) and increasing the weight of the samples from the worst predicted areas.
It's the original idea of adaboost introduced into gradient boosting.
The text was updated successfully, but these errors were encountered:
Is there a way to change safely weights after each iteration based in some user custom logic?
This would be my use case:
A problem where the data is grouped by geographic area and the gradient boosting model consistently predicts some areas worse than others. I prefer to sacrifice some of the performance of the easier to predict zones in exchange for improving the predictions of the difficult zones, so reducing the variability of the errors per zone is a priority.
One approach is to fit a model and, based on the results, over-weight the 'difficult' regions, and do this several times.
What I want to do is introduce the process into the model fitting by doing upweighting after each iteration, evaluating all predictions by geographic area (the area is passed to dtrain with attr() and the performance of each area is assessed with a groupby on the absolute value of the prediction errors of the samples they include) and increasing the weight of the samples from the worst predicted areas.
It's the original idea of adaboost introduced into gradient boosting.
The text was updated successfully, but these errors were encountered: