美文网首页心理简友广场想法
机器学习:5.4 Stacking

机器学习:5.4 Stacking

作者: Cache_wood | 来源:发表于2022-04-17 20:57 被阅读0次

    @[toc]

    • Combine multiple base learners to reduce variance

      • Base learners can be different model types
      • Linearly combine base learners outputs by learned parameters
    • Widely used in competitions

    • bagging VS stacking

      • Bagging: bootstrap samples to get diversity

      • Stacking: different types of models extract different features

    Multi-layer Stacking

    • Stacking base learners in multiple levels to reduce bias
      • Can use a different set of base learners at each level
    • Upper levels (e.g. L2) are trained on the outputs of the level below (e.g. L1)
      • Concatenating original inputs helps

    Overfitting in Multi-layer Stacking

    • Train leaners from different levels on different data to alleviate
      overfitting

    • Split training data into A and B, train L_1 learners on A, run inference on B to generate training data for L_2 learners

    • Repeated k-fold bagging:

      • Train k models as in k-fold cross validation
      • Combine predictions of each model on out-of-fold data
      • Repeat step 1,2 by n times, average the n predictions of each example for the next level training

    相关文章

      网友评论

        本文标题:机器学习:5.4 Stacking

        本文链接:https://www.haomeiwen.com/subject/cbeosrtx.html