美文网首页
Paper Reading: Deep Networks wit

Paper Reading: Deep Networks wit

作者: Dorts | 来源:发表于2016-08-31 16:31 被阅读103次

    Written on 2016/08/31

    Application:

    This simple approach complements the recent success of **residual network ** to reduce training time and improve the test error.

    Challenge:

    1. Very deep models become worse at function approximation (called ** degradation ** problem) is not caused by overfitting, but caused by training signals vanishing.
    2. Effective and efficient training methods for very deep models need to be found.

    Problem:

    Motivated by ** ResNets ** which simplifies ** Highway Networks **, authors proposed a method new called Stochastic Depth to go a step further to reduce ResNet's test error and training time.

    Solution:

    1. Shrink the depth of a network during training, while keeping it unchanged during testing.
    2. By a survival probability, randomly dropping entire ResBlocks during training and by bypassing their transformations through skip connections.
    3. Survival probabilities can adopt uniform distribution or linear decay (better)

    Insights:

    1. This method(Stochastic depth) is designed for ResNet. Therefore, other networks without ResBlocks is not compatible with this method.
    2. This method can be regarded as an implicit model ensemble.
    3. A new more competitive method has been proposed (http://arxiv.org/pdf/1603.05027.pdf), which can be employed on deeper model and acquire lower test error.

    One word to summarize:

    This paper proposes a deep network with stochastic depth, a procedure to train very deep neural networks effectively and efficiently.

    相关文章

      网友评论

          本文标题:Paper Reading: Deep Networks wit

          本文链接:https://www.haomeiwen.com/subject/ewoqettx.html