美文网首页
Tensorflow(二) Residual Network原理

Tensorflow(二) Residual Network原理

作者: blackmanba_084b | 来源:发表于2018-12-06 14:36 被阅读0次

    首先附上原著的论文(Deep Residual Learning for Image Recognition)。 ResNet(Residual Neural Network)通过使用残差学习单元(Residual Unit),训练了152层深的神经网络,在ILSVRC 2015比赛中取得3.57%的top-5错误率。

    一、原理介绍

    1. 暴露的问题

    我们知道随着我们卷积层数的升高,我们会产生梯度消失/爆炸的问题,这里我们可以通过BN(Batch Normalization)可以简单理解为对图片进行归一化处理。但是当更深的网络能够开始收敛的时候,会暴露一个退化的问题,即随着网络深度的增加,准确率达到饱和,然后迅速的下降。但是这种退化不是过拟合引起的,并且当我们添加更多的层数,training error和test error都会很高。具体退化问题,目前还没有一个确切的解释(具体可以看这篇文章训练深度神经网络失败的罪魁祸首不是梯度消失,而是退化)。

    2. 解决问题的办法

    目前对于深层的卷积网络暴露的2个问题,分别是梯度消失/爆炸以及网络退化问题。针对第一个问题我们直接通过BN的方法来解决,第二个问题我们可以通过残差模块来解决。


    Residual learning: a building block.png

    原文中的作者是这么解释的

    In this paper, we address the degradation problem by introducing a deep residual learning framework. Instead of hoping each few stacked layers directly fit a desired underlying mapping, we explicitly let these layers fit a residual mapping. Formally, denoting the desired underlying mapping as H(x), we let the stacked nonlinear layers fit another mapping of F(x) := H(x)−x. The original mapping is recast into F(x)+x. We hypothesize that it
    is easier to optimize the residual mapping than to optimize the original, unreferenced mapping. To the extreme, if an identity mapping were optimal, it would be easier to push the residual to zero than to fit an identity mapping by a stack of nonlinear layers.

    简单解释就是某段神经网络输入是x, 期望输出是H(x)。我们可以得出输出结果是H(x) = F(x) + x。可以简单理解为F(x)就是输入x做卷积池化得出的结果,x就是原始的输入,最终将它们相加得出最终结果就是我们需要的, 其实我们可以抽象理解为这个unit不仅利用了卷积后的信息还利用了原始的输入信息,我们称右侧x的方式为shortcut

    3. 具体网络结构介绍
    deep network.png
    a. 一般我们做图像分类无外乎就是对图像做卷积-池化,如图中VGG-19。但是当我们扩充卷积层数的时候(如图中34-layer plain),会出现我们一开始介绍模型暴露的问题。
    b. 右图就是我们所谓的残差神经网络的结构图。同样的颜色filter的输出通道数是一致的,不同颜色filte输出通道数是不一致的。
    image.png
    细心的读者会发现我们的shortcut有的是用实线表示,有的使用虚线表示,这是什么原因呢?巧的是我们发现这些虚线出现的地方都是卷积颜色变化的地方,这是由于我们的多次卷积池化后的输出F(x)与输入x的通道数不一致,因此在虚线的地方是需要让其通道书保持一致才能执行我们矩阵相加的想法, Residual Network Structure.png

    二、代码介绍(以resnet50为例)

    resnet_v1.py
    resnet_utils.py
    下面主要来介绍一下resnet_v1.py

    # Copyright 2016 The TensorFlow Authors. All Rights Reserved.
    #
    # Licensed under the Apache License, Version 2.0 (the "License");
    # you may not use this file except in compliance with the License.
    # You may obtain a copy of the License at
    #
    # http://www.apache.org/licenses/LICENSE-2.0
    #
    # Unless required by applicable law or agreed to in writing, software
    # distributed under the License is distributed on an "AS IS" BASIS,
    # WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
    # See the License for the specific language governing permissions and
    # limitations under the License.
    # ==============================================================================
    """Contains definitions for the original form of Residual Networks.
    The 'v1' residual networks (ResNets) implemented in this module were proposed
    by:
    [1] Kaiming He, Xiangyu Zhang, Shaoqing Ren, Jian Sun
        Deep Residual Learning for Image Recognition. arXiv:1512.03385
    Other variants were introduced in:
    [2] Kaiming He, Xiangyu Zhang, Shaoqing Ren, Jian Sun
        Identity Mappings in Deep Residual Networks. arXiv: 1603.05027
    The networks defined in this module utilize the bottleneck building block of
    [1] with projection shortcuts only for increasing depths. They employ batch
    normalization *after* every weight layer. This is the architecture used by
    MSRA in the Imagenet and MSCOCO 2016 competition models ResNet-101 and
    ResNet-152. See [2; Fig. 1a] for a comparison between the current 'v1'
    architecture and the alternative 'v2' architecture of [2] which uses batch
    normalization *before* every weight layer in the so-called full pre-activation
    units.
    Typical use:
       from tensorflow.contrib.slim.nets import resnet_v1
    ResNet-101 for image classification into 1000 classes:
       # inputs has shape [batch, 224, 224, 3]
       with slim.arg_scope(resnet_v1.resnet_arg_scope()):
          net, end_points = resnet_v1.resnet_v1_101(inputs, 1000, is_training=False)
    ResNet-101 for semantic segmentation into 21 classes:
       # inputs has shape [batch, 513, 513, 3]
       with slim.arg_scope(resnet_v1.resnet_arg_scope()):
          net, end_points = resnet_v1.resnet_v1_101(inputs,
                                                    21,
                                                    is_training=False,
                                                    global_pool=False,
                                                    output_stride=16)
    """
    from __future__ import absolute_import
    from __future__ import division
    from __future__ import print_function
    
    import tensorflow as tf
    
    from nets import resnet_utils
    
    
    resnet_arg_scope = resnet_utils.resnet_arg_scope
    slim = tf.contrib.slim
    
    
    class NoOpScope(object):
      """No-op context manager."""
    
      def __enter__(self):
        return None
    
      def __exit__(self, exc_type, exc_value, traceback):
        return False
    
    
    @slim.add_arg_scope
    def bottleneck(inputs,
                   depth,
                   depth_bottleneck,
                   stride,
                   rate=1,
                   outputs_collections=None,
                   scope=None,
                   use_bounded_activations=False):
      """Bottleneck residual unit variant with BN after convolutions.
      This is the original residual unit proposed in [1]. See Fig. 1(a) of [2] for
      its definition. Note that we use here the bottleneck variant which has an
      extra bottleneck layer.
      When putting together two consecutive ResNet blocks that use this unit, one
      should use stride = 2 in the last unit of the first block.
      Args:
        inputs: A tensor of size [batch, height, width, channels].
        depth: The depth of the ResNet unit output.
        depth_bottleneck: The depth of the bottleneck layers.
        stride: The ResNet unit's stride. Determines the amount of downsampling of
          the units output compared to its input.
        rate: An integer, rate for atrous convolution.
        outputs_collections: Collection to add the ResNet unit output.
        scope: Optional variable_scope.
        use_bounded_activations: Whether or not to use bounded activations. Bounded
          activations better lend themselves to quantized inference.
      Returns:
        The ResNet unit's output.
      """
      with tf.variable_scope(scope, 'bottleneck_v1', [inputs]) as sc:
        depth_in = slim.utils.last_dimension(inputs.get_shape(), min_rank=4)
        if depth == depth_in:
          shortcut = resnet_utils.subsample(inputs, stride, 'shortcut')
        else:
          shortcut = slim.conv2d(
              inputs,
              depth, [1, 1],
              stride=stride,
              activation_fn=tf.nn.relu6 if use_bounded_activations else None,
              scope='shortcut')
    
        residual = slim.conv2d(inputs, depth_bottleneck, [1, 1], stride=1,
                               scope='conv1')
        residual = resnet_utils.conv2d_same(residual, depth_bottleneck, 3, stride,
                                            rate=rate, scope='conv2')
        residual = slim.conv2d(residual, depth, [1, 1], stride=1,
                               activation_fn=None, scope='conv3')
    
        if use_bounded_activations:
          # Use clip_by_value to simulate bandpass activation.
          residual = tf.clip_by_value(residual, -6.0, 6.0)
          output = tf.nn.relu6(shortcut + residual)
        else:
          output = tf.nn.relu(shortcut + residual)
    
        return slim.utils.collect_named_outputs(outputs_collections,
                                                sc.name,
                                                output)
    
    
    def resnet_v1(inputs,
                  blocks,
                  num_classes=None,
                  is_training=True,
                  global_pool=True,
                  output_stride=None,
                  include_root_block=True,
                  spatial_squeeze=True,
                  store_non_strided_activations=False,
                  reuse=None,
                  scope=None):
      """Generator for v1 ResNet models.
      This function generates a family of ResNet v1 models. See the resnet_v1_*()
      methods for specific model instantiations, obtained by selecting different
      block instantiations that produce ResNets of various depths.
      Training for image classification on Imagenet is usually done with [224, 224]
      inputs, resulting in [7, 7] feature maps at the output of the last ResNet
      block for the ResNets defined in [1] that have nominal stride equal to 32.
      However, for dense prediction tasks we advise that one uses inputs with
      spatial dimensions that are multiples of 32 plus 1, e.g., [321, 321]. In
      this case the feature maps at the ResNet output will have spatial shape
      [(height - 1) / output_stride + 1, (width - 1) / output_stride + 1]
      and corners exactly aligned with the input image corners, which greatly
      facilitates alignment of the features to the image. Using as input [225, 225]
      images results in [8, 8] feature maps at the output of the last ResNet block.
      For dense prediction tasks, the ResNet needs to run in fully-convolutional
      (FCN) mode and global_pool needs to be set to False. The ResNets in [1, 2] all
      have nominal stride equal to 32 and a good choice in FCN mode is to use
      output_stride=16 in order to increase the density of the computed features at
      small computational and memory overhead, cf. http://arxiv.org/abs/1606.00915.
      Args:
        inputs: A tensor of size [batch, height_in, width_in, channels].
        blocks: A list of length equal to the number of ResNet blocks. Each element
          is a resnet_utils.Block object describing the units in the block.
        num_classes: Number of predicted classes for classification tasks.
          If 0 or None, we return the features before the logit layer.
        is_training: whether batch_norm layers are in training mode. If this is set
          to None, the callers can specify slim.batch_norm's is_training parameter
          from an outer slim.arg_scope.
        global_pool: If True, we perform global average pooling before computing the
          logits. Set to True for image classification, False for dense prediction.
        output_stride: If None, then the output will be computed at the nominal
          network stride. If output_stride is not None, it specifies the requested
          ratio of input to output spatial resolution.
        include_root_block: If True, include the initial convolution followed by
          max-pooling, if False excludes it.
        spatial_squeeze: if True, logits is of shape [B, C], if false logits is
            of shape [B, 1, 1, C], where B is batch_size and C is number of classes.
            To use this parameter, the input images must be smaller than 300x300
            pixels, in which case the output logit layer does not contain spatial
            information and can be removed.
        store_non_strided_activations: If True, we compute non-strided (undecimated)
          activations at the last unit of each block and store them in the
          `outputs_collections` before subsampling them. This gives us access to
          higher resolution intermediate activations which are useful in some
          dense prediction problems but increases 4x the computation and memory cost
          at the last unit of each block.
        reuse: whether or not the network and its variables should be reused. To be
          able to reuse 'scope' must be given.
        scope: Optional variable_scope.
      Returns:
        net: A rank-4 tensor of size [batch, height_out, width_out, channels_out].
          If global_pool is False, then height_out and width_out are reduced by a
          factor of output_stride compared to the respective height_in and width_in,
          else both height_out and width_out equal one. If num_classes is 0 or None,
          then net is the output of the last ResNet block, potentially after global
          average pooling. If num_classes a non-zero integer, net contains the
          pre-softmax activations.
        end_points: A dictionary from components of the network to the corresponding
          activation.
      Raises:
        ValueError: If the target output_stride is not valid.
      """
      with tf.variable_scope(scope, 'resnet_v1', [inputs], reuse=reuse) as sc:
        end_points_collection = sc.original_name_scope + '_end_points'
        with slim.arg_scope([slim.conv2d, bottleneck,
                             resnet_utils.stack_blocks_dense],
                            outputs_collections=end_points_collection):
          with (slim.arg_scope([slim.batch_norm], is_training=is_training)
                if is_training is not None else NoOpScope()):
            net = inputs
            if include_root_block:
              if output_stride is not None:
                if output_stride % 4 != 0:
                  raise ValueError('The output_stride needs to be a multiple of 4.')
                output_stride /= 4
              net = resnet_utils.conv2d_same(net, 64, 7, stride=2, scope='conv1')
              net = slim.max_pool2d(net, [3, 3], stride=2, scope='pool1')
            net = resnet_utils.stack_blocks_dense(net, blocks, output_stride,
                                                  store_non_strided_activations)
            # Convert end_points_collection into a dictionary of end_points.
            end_points = slim.utils.convert_collection_to_dict(
                end_points_collection)
    
            if global_pool:
              # Global average pooling.
              net = tf.reduce_mean(net, [1, 2], name='pool5', keep_dims=True)
              end_points['global_pool'] = net
            if num_classes:
              net = slim.conv2d(net, num_classes, [1, 1], activation_fn=None,
                                normalizer_fn=None, scope='logits')
              end_points[sc.name + '/logits'] = net
              if spatial_squeeze:
                net = tf.squeeze(net, [1, 2], name='SpatialSqueeze')
                end_points[sc.name + '/spatial_squeeze'] = net
              end_points['predictions'] = slim.softmax(net, scope='predictions')
            return net, end_points
    resnet_v1.default_image_size = 224
    
    
    def resnet_v1_block(scope, base_depth, num_units, stride):
      """Helper function for creating a resnet_v1 bottleneck block.
      Args:
        scope: The scope of the block.
        base_depth: The depth of the bottleneck layer for each unit.
        num_units: The number of units in the block.
        stride: The stride of the block, implemented as a stride in the last unit.
          All other units have stride=1.
      Returns:
        A resnet_v1 bottleneck block.
      """
      return resnet_utils.Block(scope, bottleneck, [{
          'depth': base_depth * 4,
          'depth_bottleneck': base_depth,
          'stride': 1
      }] * (num_units - 1) + [{
          'depth': base_depth * 4,
          'depth_bottleneck': base_depth,
          'stride': stride
      }])
    
    
    def resnet_v1_50(inputs,
                     num_classes=None,
                     is_training=True,
                     global_pool=True,
                     output_stride=None,
                     spatial_squeeze=True,
                     store_non_strided_activations=False,
                     reuse=None,
                     scope='resnet_v1_50'):
      """ResNet-50 model of [1]. See resnet_v1() for arg and return description."""
      blocks = [
          resnet_v1_block('block1', base_depth=64, num_units=3, stride=2),
          resnet_v1_block('block2', base_depth=128, num_units=4, stride=2),
          resnet_v1_block('block3', base_depth=256, num_units=6, stride=2),
          resnet_v1_block('block4', base_depth=512, num_units=3, stride=1),
      ]
      return resnet_v1(inputs, blocks, num_classes, is_training,
                       global_pool=global_pool, output_stride=output_stride,
                       include_root_block=True, spatial_squeeze=spatial_squeeze,
                       store_non_strided_activations=store_non_strided_activations,
                       reuse=reuse, scope=scope)
    resnet_v1_50.default_image_size = resnet_v1.default_image_size
    
    
    def resnet_v1_101(inputs,
                      num_classes=None,
                      is_training=True,
                      global_pool=True,
                      output_stride=None,
                      spatial_squeeze=True,
                      store_non_strided_activations=False,
                      reuse=None,
                      scope='resnet_v1_101'):
      """ResNet-101 model of [1]. See resnet_v1() for arg and return description."""
      blocks = [
          resnet_v1_block('block1', base_depth=64, num_units=3, stride=2),
          resnet_v1_block('block2', base_depth=128, num_units=4, stride=2),
          resnet_v1_block('block3', base_depth=256, num_units=23, stride=2),
          resnet_v1_block('block4', base_depth=512, num_units=3, stride=1),
      ]
      return resnet_v1(inputs, blocks, num_classes, is_training,
                       global_pool=global_pool, output_stride=output_stride,
                       include_root_block=True, spatial_squeeze=spatial_squeeze,
                       store_non_strided_activations=store_non_strided_activations,
                       reuse=reuse, scope=scope)
    resnet_v1_101.default_image_size = resnet_v1.default_image_size
    
    
    def resnet_v1_152(inputs,
                      num_classes=None,
                      is_training=True,
                      global_pool=True,
                      output_stride=None,
                      store_non_strided_activations=False,
                      spatial_squeeze=True,
                      reuse=None,
                      scope='resnet_v1_152'):
      """ResNet-152 model of [1]. See resnet_v1() for arg and return description."""
      blocks = [
          resnet_v1_block('block1', base_depth=64, num_units=3, stride=2),
          resnet_v1_block('block2', base_depth=128, num_units=8, stride=2),
          resnet_v1_block('block3', base_depth=256, num_units=36, stride=2),
          resnet_v1_block('block4', base_depth=512, num_units=3, stride=1),
      ]
      return resnet_v1(inputs, blocks, num_classes, is_training,
                       global_pool=global_pool, output_stride=output_stride,
                       include_root_block=True, spatial_squeeze=spatial_squeeze,
                       store_non_strided_activations=store_non_strided_activations,
                       reuse=reuse, scope=scope)
    resnet_v1_152.default_image_size = resnet_v1.default_image_size
    
    
    def resnet_v1_200(inputs,
                      num_classes=None,
                      is_training=True,
                      global_pool=True,
                      output_stride=None,
                      store_non_strided_activations=False,
                      spatial_squeeze=True,
                      reuse=None,
                      scope='resnet_v1_200'):
      """ResNet-200 model of [2]. See resnet_v1() for arg and return description."""
      blocks = [
          resnet_v1_block('block1', base_depth=64, num_units=3, stride=2),
          resnet_v1_block('block2', base_depth=128, num_units=24, stride=2),
          resnet_v1_block('block3', base_depth=256, num_units=36, stride=2),
          resnet_v1_block('block4', base_depth=512, num_units=3, stride=1),
      ]
      return resnet_v1(inputs, blocks, num_classes, is_training,
                       global_pool=global_pool, output_stride=output_stride,
                       include_root_block=True, spatial_squeeze=spatial_squeeze,
                       store_non_strided_activations=store_non_strided_activations,
                       reuse=reuse, scope=scope)
    resnet_v1_200.default_image_size = resnet_v1.default_image_size
    

    这个代码我们分几个部分介绍,分别是bottleneck, resnet_v1_block, resnet_v1_50, resnet_v1

    1. 首先我们先看一下resnet_v1_block
    def resnet_v1_block(scope, base_depth, num_units, stride):
      """Helper function for creating a resnet_v1 bottleneck block.
    
      Args:
        scope: The scope of the block.
        base_depth: The depth of the bottleneck layer for each unit.
        num_units: The number of units in the block.
        stride: The stride of the block, implemented as a stride in the last unit.
          All other units have stride=1.
    
      Returns:
        A resnet_v1 bottleneck block.
      """
      return resnet_utils.Block(scope, bottleneck, [{
          'depth': base_depth * 4,
          'depth_bottleneck': base_depth,
          'stride': 1
      }] * (num_units - 1) + [{
          'depth': base_depth * 4,
          'depth_bottleneck': base_depth,
          'stride': stride
      }])
    

    从中我们可以得知,(num_units - 1)表示的是每一个v1_block第一个到倒数第二个unit都是按照stride为1,最后一个unit stride为2.至于为什么是2,我们可以在介绍resnet_v1_50介绍。

    1. 接下来我们来说明以下bottleneck.
      bottleneck其实是ResNet代码中最核心的模块。也就是对下面原理图的实现。它具体是如何实线的,这里我会很详细的介绍,因为这里也是整个resnet的核心。
      bottleneck.png
    def bottleneck(inputs,
                   depth,
                   depth_bottleneck,
                   stride,
                   rate=1,
                   outputs_collections=None,
                   scope=None,
                   use_bounded_activations=False):
          with tf.variable_scope(scope, 'bottleneck_v1', [inputs]) as sc:
            depth_in = slim.utils.last_dimension(inputs.get_shape(), min_rank=4)
            if depth == depth_in:
              shortcut = resnet_utils.subsample(inputs, stride, 'shortcut')
            else:
              shortcut = slim.conv2d(
                  inputs,
                  depth, [1, 1],
                  stride=stride,
                  activation_fn=tf.nn.relu6 if use_bounded_activations else None,
                  scope='shortcut')
    
            residual = slim.conv2d(inputs, depth_bottleneck, [1, 1], stride=1,
                                   scope='conv1')
            residual = resnet_utils.conv2d_same(residual, depth_bottleneck, 3, stride,
                                                rate=rate, scope='conv2')
            residual = slim.conv2d(residual, depth, [1, 1], stride=1,
                                   activation_fn=None, scope='conv3')
    
            if use_bounded_activations:
              # Use clip_by_value to simulate bandpass activation.
              residual = tf.clip_by_value(residual, -6.0, 6.0)
              output = tf.nn.relu6(shortcut + residual)
            else:
              output = tf.nn.relu(shortcut + residual)
    
            return slim.utils.collect_named_outputs(outputs_collections,
                                                    sc.name,
                                                output)
    

    (1). 参数介绍:
      inputs: 输入;
       depth: 输出通道数;
       depth_bottleneck: 残差的第1、2层输出通道数;
       stride:步长;
       rate:是做atrous convolution实现空洞卷积,空洞卷积的采样率(可忽略)
       outputs_collections收集残差网络单元作为输出(可忽略)
       scope 定义scope名称(可忽略)
       use_bounded_activations是否使用有界激活(可忽略)
      我们来结合下面的图来理解就很好理解上面的参数作用了。

    bottleneck.png
    (2). 我们在结合代码看以下这段代码的详细作用
    a. depth_in = slim.utils.last_dimension(inputs.get_shape(), min_rank=4)是用来获取输入的最后一个维度,这里min_rank=4通道数必须大于4.

    (3)shortcut路线

     if depth == depth_in:
          shortcut = resnet_utils.subsample(inputs, stride, 'shortcut')
        else:
          shortcut = slim.conv2d(
              inputs,
              depth, [1, 1],
              stride=stride,
              activation_fn=tf.nn.relu6 if use_bounded_activations else None,
              scope='shortcut')
    

    a. 如果输入通道数depth_in和输出通道数depth相同则使用subsample降采样的方式处理,定义如下:

    def subsample(inputs, factor, scope=None):
      """Subsamples the input along the spatial dimensions.
    
      Args:
        inputs: A `Tensor` of size [batch, height_in, width_in, channels].
        factor: The subsampling factor.
        scope: Optional variable_scope.
    
      Returns:
        output: A `Tensor` of size [batch, height_out, width_out, channels] with the
          input, either intact (if factor == 1) or subsampled (if factor > 1).
      """
      if factor == 1:
        return inputs
      else:
        return slim.max_pool2d(inputs, [1, 1], stride=factor, scope=scope)
    

    我们的输入就是输出,也就是说我们的feature map size以及我们的通道数(256)都没有发生变动。

    b. 如果我们depth_in与输出通道depth不相同则会走shortcut虚线渠道,这里我们会做1*1conv2d卷积,这里cond2d也就是做一个1*1卷积,stride为2的卷积。我们的feature size为\frac{featureSize-1}{2}+1=\frac{featureSize}{2}+1因为是向下取整所以我们的feature size为原来的\frac{1}{2}。并且我们的depth为256,所以我们的通道数为256。
    也就是说我们的通道数变为原来的4倍,feature map 变为原来的(\frac{1}{2}*\frac{1}{2})。

    (4)中间路线

        residual = slim.conv2d(inputs, depth_bottleneck, [1, 1], stride=1,
                               scope='conv1')
        residual = resnet_utils.conv2d_same(residual, depth_bottleneck, 3, stride,
                                            rate=rate, scope='conv2')
        residual = slim.conv2d(residual, depth, [1, 1], stride=1,
                               activation_fn=None, scope='conv3')
    

    中间分别是1*1卷积,3*3卷积,以及1*1卷积。这里为什么要这么做呢?我们以resnet50的配置参数为例说明。
    (a)我们图中中间这条路线在shortcut为实线的情况下

    • 在第一做卷积的时候stride=1, kernel size =[1,1]。根据我们的常识卷积完之后输出Size = \frac{OrginalSize - kernerSize+2Padding}{Stride}+1(这里的Padding为0)所以feature map的size没有发生变化),所以第一次卷积我们的通道数变为原来的\frac{1}{4}即64。但是我们的feature size没有发生变化。

    • 第二次做3*3con2d_same(说明如下)

    if stride == 1:
        return slim.conv2d(inputs, num_outputs, kernel_size, stride=1, rate=rate,
                           padding='SAME', scope=scope)
    

    也就是说第一次做kerner size=[3,3], stride=1,由于padding='SAME', 此时我们的feature map没有发生变化,但是通道数变为1.

    • 第三次我们再进行kernel size=[1,1], stride=1,此时我们知道当kernel size=stride=1的时候我们的feature map是没有发生变化的,但是我们的通道数也没有发生变化(256)。

    所以最终的feature map 与通道数都没有发生变化

    ★★★至于我们为什么要这么操作,这是由于训练时间太长,我们可以先将通道数变少再进行下面一步3*3的卷积。之后第三步的卷积是为了还原一开始通道数的尺寸。这么一看很明显了其实很就是为了第二步骤3*3的卷积提速,才先将输入缩小,卷积完成之后再进行还原,所以说我们的通道数是没有发生任何改变的,只是加速卷积速度。

    (b)我们图中中间这条路线在shortcut为虚线的情况下

    • 第一次做卷积和实现没有区别通道数变为原来的\frac{1}{4}即64。但是我们的feature size没有发生变化。
    • 第二次做conv2d_same,定义如下:
     Note that
         net = conv2d_same(inputs, num_outputs, 3, stride=stride)
      is equivalent to
         net = slim.conv2d(inputs, num_outputs, 3, stride=1, padding='SAME')
         net = subsample(net, factor=stride)
      whereas
         net = slim.conv2d(inputs, num_outputs, 3, stride=stride, padding='SAME')
    

    通道数没有变化(64),但是由于我们的stride=2,所以我们的feature size为\frac{featureSize-1}{2}+1=\frac{featureSize}{2}+1因为是向下取整所以我们的feature size为原来的\frac{1}{2}。即feature map 变为原来的(\frac{1}{2}*\frac{1}{2})。
    (c)第三次卷积因为stride size=filter size=1, 所以feature size不变,但是我们的通道数改变为原来的4倍。
    也就是说当shortcut为虚线的时候feature map变为原来的/frac{1}{4},但是通道数变为原来的4倍。

    通过上面的分析我们可以得出下面的结论:

    Feature 中间路线 shortcut路线 中间路线与shortcut路线是否可合并
    feature map大小(shortcut为实线) 没有变化 没有变化 可以
    通道数(shortcut为实线) 没有变化 没有变化 可以
    feature map大小(shortcut为虚线) 原来的\frac{1}{4} 原来的\frac{1}{4} 可以
    通道数(shortcut为虚线) 变为原来的4倍 变为原来的4倍 可以

    (6)所以最后我们通过output = tf.nn.relu(shortcut + residual)将我们的F(x) 与 x合并。

    1. 这里我们来介绍以下def resnet_v1_50
    def resnet_v1_50(inputs,
                     num_classes=None,
                     is_training=True,
                     global_pool=True,
                     output_stride=None,
                     spatial_squeeze=True,
                     store_non_strided_activations=False,
                     reuse=None,
                     scope='resnet_v1_50'):
      """ResNet-50 model of [1]. See resnet_v1() for arg and return description."""
      blocks = [
          resnet_v1_block('block1', base_depth=64, num_units=3, stride=2),
          resnet_v1_block('block2', base_depth=128, num_units=4, stride=2),
          resnet_v1_block('block3', base_depth=256, num_units=6, stride=2),
          resnet_v1_block('block4', base_depth=512, num_units=3, stride=1),
      ]
      return resnet_v1(inputs, blocks, num_classes, is_training,
                       global_pool=global_pool, output_stride=output_stride,
                       include_root_block=True, spatial_squeeze=spatial_squeeze,
                       store_non_strided_activations=store_non_strided_activations,
                       reuse=reuse, scope=scope)
    

    这里我们仅仅需要介绍以下blocks

      blocks = [
          resnet_v1_block('block1', base_depth=64, num_units=3, stride=2),
          resnet_v1_block('block2', base_depth=128, num_units=4, stride=2),
          resnet_v1_block('block3', base_depth=256, num_units=6, stride=2),
          resnet_v1_block('block4', base_depth=512, num_units=3, stride=1),
      ]
    

    base_depth是每一个单元bottleneck层的深度(可参考bottleneck.png),num_units表示为每一个unit的数量这里需要注意以下除了最后的block stride是1,其余的stride都是2。

    1. 这里我们要介绍的是resnet_v1, 这个函数是一个完整的ResNet网络结构。
    def resnet_v1(inputs,
                  blocks,
                  num_classes=None,
                  is_training=True,
                  global_pool=True,
                  output_stride=None,
                  include_root_block=True,
                  spatial_squeeze=True,
                  store_non_strided_activations=False,
                  reuse=None,
                  scope=None):
      """Generator for v1 ResNet models.
      This function generates a family of ResNet v1 models. See the resnet_v1_*()
      methods for specific model instantiations, obtained by selecting different
      block instantiations that produce ResNets of various depths.
      Training for image classification on Imagenet is usually done with [224, 224]
      inputs, resulting in [7, 7] feature maps at the output of the last ResNet
      block for the ResNets defined in [1] that have nominal stride equal to 32.
      However, for dense prediction tasks we advise that one uses inputs with
      spatial dimensions that are multiples of 32 plus 1, e.g., [321, 321]. In
      this case the feature maps at the ResNet output will have spatial shape
      [(height - 1) / output_stride + 1, (width - 1) / output_stride + 1]
      and corners exactly aligned with the input image corners, which greatly
      facilitates alignment of the features to the image. Using as input [225, 225]
      images results in [8, 8] feature maps at the output of the last ResNet block.
      For dense prediction tasks, the ResNet needs to run in fully-convolutional
      (FCN) mode and global_pool needs to be set to False. The ResNets in [1, 2] all
      have nominal stride equal to 32 and a good choice in FCN mode is to use
      output_stride=16 in order to increase the density of the computed features at
      small computational and memory overhead, cf. http://arxiv.org/abs/1606.00915.
      Args:
        inputs: A tensor of size [batch, height_in, width_in, channels].
        blocks: A list of length equal to the number of ResNet blocks. Each element
          is a resnet_utils.Block object describing the units in the block.
        num_classes: Number of predicted classes for classification tasks.
          If 0 or None, we return the features before the logit layer.
        is_training: whether batch_norm layers are in training mode. If this is set
          to None, the callers can specify slim.batch_norm's is_training parameter
          from an outer slim.arg_scope.
        global_pool: If True, we perform global average pooling before computing the
          logits. Set to True for image classification, False for dense prediction.
        output_stride: If None, then the output will be computed at the nominal
          network stride. If output_stride is not None, it specifies the requested
          ratio of input to output spatial resolution.
        include_root_block: If True, include the initial convolution followed by
          max-pooling, if False excludes it.
        spatial_squeeze: if True, logits is of shape [B, C], if false logits is
            of shape [B, 1, 1, C], where B is batch_size and C is number of classes.
            To use this parameter, the input images must be smaller than 300x300
            pixels, in which case the output logit layer does not contain spatial
            information and can be removed.
        store_non_strided_activations: If True, we compute non-strided (undecimated)
          activations at the last unit of each block and store them in the
          `outputs_collections` before subsampling them. This gives us access to
          higher resolution intermediate activations which are useful in some
          dense prediction problems but increases 4x the computation and memory cost
          at the last unit of each block.
        reuse: whether or not the network and its variables should be reused. To be
          able to reuse 'scope' must be given.
        scope: Optional variable_scope.
      Returns:
        net: A rank-4 tensor of size [batch, height_out, width_out, channels_out].
          If global_pool is False, then height_out and width_out are reduced by a
          factor of output_stride compared to the respective height_in and width_in,
          else both height_out and width_out equal one. If num_classes is 0 or None,
          then net is the output of the last ResNet block, potentially after global
          average pooling. If num_classes a non-zero integer, net contains the
          pre-softmax activations.
        end_points: A dictionary from components of the network to the corresponding
          activation.
      Raises:
        ValueError: If the target output_stride is not valid.
      """
      with tf.variable_scope(scope, 'resnet_v1', [inputs], reuse=reuse) as sc:
        end_points_collection = sc.original_name_scope + '_end_points'
        with slim.arg_scope([slim.conv2d, bottleneck,
                             resnet_utils.stack_blocks_dense],
                            outputs_collections=end_points_collection):
          with (slim.arg_scope([slim.batch_norm], is_training=is_training)
                if is_training is not None else NoOpScope()):
            net = inputs
            if include_root_block:
              if output_stride is not None:
                if output_stride % 4 != 0:
                  raise ValueError('The output_stride needs to be a multiple of 4.')
                output_stride /= 4
              net = resnet_utils.conv2d_same(net, 64, 7, stride=2, scope='conv1')
              net = slim.max_pool2d(net, [3, 3], stride=2, scope='pool1')
            net = resnet_utils.stack_blocks_dense(net, blocks, output_stride,
                                                  store_non_strided_activations)
            # Convert end_points_collection into a dictionary of end_points.
            end_points = slim.utils.convert_collection_to_dict(
                end_points_collection)
    
            if global_pool:
              # Global average pooling.
              net = tf.reduce_mean(net, [1, 2], name='pool5', keep_dims=True)
              end_points['global_pool'] = net
            if num_classes:
              net = slim.conv2d(net, num_classes, [1, 1], activation_fn=None,
                                normalizer_fn=None, scope='logits')
              end_points[sc.name + '/logits'] = net
              if spatial_squeeze:
                net = tf.squeeze(net, [1, 2], name='SpatialSqueeze')
                end_points[sc.name + '/spatial_squeeze'] = net
              end_points['predictions'] = slim.softmax(net, scope='predictions')
            return net, end_points
    resnet_v1.default_image_size = 224
    

    a. end_points_collecion = sc.original_name_scope + 'end_points'这是字符串用来命名collection的名字。
    b.

    net = resnet_utils.conv2d_same(net, 64, 7, stride=2, scope='conv1')
    net = slim.max_pool2d(net, [3, 3], stride=2, scope='pool1')
    

    这里是刚输入的图片进行kernel size[7,7], stride=2, 输出通道是64的卷积,在进行kernel size[3,3], stride=2的池化操作, 并将我们的feature map缩小为原来的\frac{1}{4}
    c. end_points = slim.utils.convert_collection_to_dict(end_points_collection)
    这里读取blocks数据结构,将一开始生成的池化结构与我们生成的block合并生成残差结构。
    d. end_points = slim.utils.convert_collection_to_dict( end_points_collection)
    这里是将我们的collection转换成dict。
    e. 下面我们来介绍一下下面的代码:

    if global_pool:
      # Global average pooling.
        net = tf.reduce_mean(net, [1, 2], name='pool5', keep_dims=True)
        end_points['global_pool'] = net
    if num_classes:
        net = slim.conv2d(net, num_classes, [1, 1], activation_fn=None, normalizer_fn=None, scope='logits')
        end_points[sc.name + '/logits'] = net
        if spatial_squeeze:
            net = tf.squeeze(net, [1, 2], name='SpatialSqueeze')
            end_points[sc.name + '/spatial_squeeze'] = net
        end_points['predictions'] = slim.softmax(net, scope='predictions')
    return net, end_points
    

    if global_pool:这里的意思就是全局池化,效率会比avg_pool要更很多,即对每个feature做平均池化,使每个feature输出一个值,这是得到[1,1,channel]大小的向量,即我们的全连接层net, 为了后期做分类使用。
    if num_classes:这里是不但输出一个全连接层也输出分类结果。但是我们在做finetune训练自己的模型的时候,我们要将我们的num_classes设置为None, 我们只需要得出我们的全连接层net即可。

    1. 最后我们来介绍一下resnet_arg_scope函数。
    def resnet_arg_scope(is_training=True,
                         weight_decay=0.0001,      # L2权重衰减速率
                         batch_norm_decay=0.997,   # BN的衰减速率
                         batch_norm_epsilon=1e-5,  # BN的epsilon默认1e-5
                         batch_norm_scale=True):   # BN的scale默认值
     
        batch_norm_params = {  # 定义batch normalization(标准化)的参数字典
            'is_training': is_training,
            # 是否是在训练模式,如果是在训练阶段,将会使用指数衰减函数(衰减系数为指定的decay),
            # 对moving_mean和moving_variance进行统计特性的动量更新,也就是进行使用指数衰减函数对均值和方
            # 差进行更新,而如果是在测试阶段,均值和方差就是固定不变的,是在训练阶段就求好的,在训练阶段,
            # 每个批的均值和方差的更新是加上了一个指数衰减函数,而最后求得的整个训练样本的均值和方差就是所
            # 有批的均值的均值,和所有批的方差的无偏估计
     
            'zero_debias_moving_mean': True,
            # 如果为True,将会创建一个新的变量对 'moving_mean/biased' and 'moving_mean/local_step',
            # 默认设置为False,将其设为True可以增加稳定性
     
            'decay': batch_norm_decay,             # Decay for the moving averages.
            # 该参数能够衡量使用指数衰减函数更新均值方差时,更新的速度,取值通常在0.999-0.99-0.9之间,值
            # 越小,代表更新速度越快,而值太大的话,有可能会导致均值方差更新太慢,而最后变成一个常量1,而
            # 这个值会导致模型性能较低很多.另外,如果出现过拟合时,也可以考虑增加均值和方差的更新速度,也
            # 就是减小decay
     
            'epsilon': batch_norm_epsilon,         # 就是在归一化时,除以方差时,防止方差为0而加上的一个数
            'scale': batch_norm_scale,
            'updates_collections': tf.GraphKeys.UPDATE_OPS,    
            # force in-place updates of mean and variance estimates
            # 该参数有一个默认值,ops.GraphKeys.UPDATE_OPS,当取默认值时,slim会在当前批训练完成后再更新均
            # 值和方差,这样会存在一个问题,就是当前批数据使用的均值和方差总是慢一拍,最后导致训练出来的模
            # 型性能较差。所以,一般需要将该值设为None,这样slim进行批处理时,会对均值和方差进行即时更新,
            # 批处理使用的就是最新的均值和方差。
            #
            # 另外,不论是即使更新还是一步训练后再对所有均值方差一起更新,对测试数据是没有影响的,即测试数
            # 据使用的都是保存的模型中的均值方差数据,但是如果你在训练中需要测试,而忘了将is_training这个值
            # 改成false,那么这批测试数据将会综合当前批数据的均值方差和训练数据的均值方差。而这样做应该是不
            # 正确的。
    

    最后我们附上Residual Network- 50的详细结构图


    Residual Network - 50.png

    参考:
    1.『TensorFlow』读书笔记_ResNet_V2

    1. resnet50结构图

    相关文章

      网友评论

          本文标题:Tensorflow(二) Residual Network原理

          本文链接:https://www.haomeiwen.com/subject/fpkwcqtx.html