小言_互联网的博客

人工智能学习:ResNet神经网络(8)

332人阅读  评论(0)

ResNet是一种非常有效的图像分类识别的模型,可以参考如下的链接
https://blog.csdn.net/qq_45649076/article/details/120494328

ResNet网络由残差(Residual)结构的基本模块构成,每一个基本模块包含几个卷积层。其中,除了网络的推理输出,基本模块的输入也被直接加成到模块的输出。这种设计可以防止网络在深度加大之后产生退化的现象。

通常所见的有Resnet-18,Resnet-34,Resnet-50,Resnet-152等多层神经网络。作为示例,以下通过Tensorflow来构建Resnet-18模型。

首先,导入需要的模块

import numpy as np

import tensorflow as tf
from tensorflow import keras
from keras import models, layers

import matplotlib.pyplot as plt

然后,定义Resnet的基本模块

# define class of basic block
class ResBlock(keras.Model):
    def __init__(self, filters, strides=1, down_sample=False):
        super().__init__()
        
        self.down_sample = down_sample
        
        self.conv1 = layers.Conv2D(filters, (3,3), strides=strides, padding='same', use_bias=False)
        self.bn1 = layers.BatchNormalization()
        self.relu1 = layers.Activation('relu')
        
        self.conv2 = layers.Conv2D(filters, (3,3), strides=1, padding='same', use_bias=False)
        self.bn2 = layers.BatchNormalization()

        if self.down_sample:
            self.down_conv = layers.Conv2D(filters, (1,1), strides=strides, padding='same', use_bias=False)
            self.down_bn = layers.BatchNormalization()
            
        self.relu2 = layers.Activation('relu')

    def call(self, inputs):
        net = self.conv1(inputs)
        net = self.bn1(net)
        net = self.relu1(net)

        net = self.conv2(net)
        net = self.bn2(net)

        # down sample inputs if dimension changes
        if self.down_sample:
            identity_out = self.down_conv(inputs)
            identity_out = self.down_bn(identity_out)
        else:
            identity_out = inputs

        net = self.relu2(net+identity_out)
        
        return net

 

ResBlock由两个卷积层组成,每一个卷积层后面跟BatchNormalization层和Relu层。输入层的数据与第二层的卷积输出相加,通过Relu层产生模块的输出。这里分为两类,如果模块中的第一个卷积层进行了stride>1(通常为2)的降维卷积,那么输入也需要进行kernel_size为1的降维操作。

然后根据ResBlock来构建Resnet网络

# define class of Resnet-18 model
class Resnet18(keras.Model):
    def __init__(self, initial_filters=64):
        # each item in block_list represent number of base blocks(ResBlock) in that block
        super().__init__()
        
        filters = initial_filters
        
        # input layers
        self.input_layer = models.Sequential([
            layers.Conv2D(filters, (3,3), strides=1, padding='same', use_bias=False),
            layers.BatchNormalization(),
            layers.Activation('relu')
        ])
        
        # first layers, no down sample, default input filters, 64
        self.layer1 = models.Sequential([
            ResBlock(filters),
            ResBlock(filters)
        ])
        
        # second layer, filters number doubled, 128
        filters *= 2
        self.layer2 = models.Sequential([
            ResBlock(filters, strides=2, down_sample=True),
            ResBlock(filters)
        ])
        
        # third layer, filters number 256
        filters *= 2
        self.layer3 = models.Sequential([
            ResBlock(filters, strides=2, down_sample=True),
            ResBlock(filters)
        ])

        # third layer, filters number 512
        filters *= 2
        self.layer4 = models.Sequential([
            ResBlock(filters, strides=2, down_sample=True),
            ResBlock(filters)
        ])

        # output layer
        self.output_layer = models.Sequential([
            layers.GlobalAveragePooling2D(),
            layers.Dense(10, activation='softmax')
        ])
        
    def call(self, inputs):
        # input layer
        net = self.input_layer(inputs)
        
        # Resnet layers
        net = self.layer1(net)
        net = self.layer2(net)
        net = self.layer3(net)
        net = self.layer4(net)
        
        # output layer
        net = self.output_layer(net)
        
        return net

 

Resnet18由一个输入层、4个中间层和一个输出层组成,每一个中间层包含两基本模块,除了第一个中间层,每一层的第一个ResBlock为降维操作,第二个ResBlock为同维的卷积操作。每到下一个中间层,卷积特征的个数倍增。输入层为一个卷积层和一个BN层、一个Relu层组成。有些地方用7x7的降维卷积层和池化层,这里准备用于尺寸较小的CIFAR-10数据集,不进行降维操作。输出层为一个均值池化层和全连接层的连接。

下面构建网络并在CIFAR-10数据集上进行测试。
构建Resnet18网络,选择优化器

# build model, Resnet18
model = Resnet18()
model.compile(optimizer='adam', loss='sparse_categorical_crossentropy', metrics=['sparse_categorical_accuracy'])

载入CIFAR-10数据

# test on CIFAR-10 data, load CIFAR-10 dataset
(train_images, train_labels), (test_images, test_labels) = keras.datasets.cifar10.load_data()

# train_images: 50000*32*32*3, train_labels: 50000*1
# test_images: 10000*32*32*3, test_labels: 10000*1

# pre-process data
train_input = train_images/255.0
test_input = test_images/255.0

train_output = train_labels
test_output = test_labels

定义数据处理器

# define data generator
data_generator = keras.preprocessing.image.ImageDataGenerator(
    featurewise_center=False,              #set mean of data to 0, by feature
    samplewise_center=False,               #set mean of data to 0, by sample
    featurewise_std_normalization=False,   #normalize by std, by feature
    samplewise_std_normalization=False,    #normalize by std, by sample
    zca_whitening=False,                   #zca whitening
    #zca_epsilon                           #zca epsilon, default 1e-6
    rotation_range=15,                     #degree of random rotation (integer, 0-180)
    width_shift_range=0.1,                 #probability of horizontal shift
    height_shift_range=0.1,                #probability of vertical shift
    horizontal_flip=True,                  #if random horizantal flip
    vertical_flip=False                    #if random vertical flip
)

data_generator.fit(train_input)

 

进行训练

# train, with batch size and epochs
epochs = 60
batch_size = 128

history = model.fit(data_generator.flow(train_input, train_output, batch_size=batch_size), epochs=epochs,
                      steps_per_epoch=len(train_input)//batch_size, validation_data=(test_input, test_output))

结果如下

Epoch 1/60
390/390 [==============================] - 47s 113ms/step - loss: 1.4509 - sparse_categorical_accuracy: 0.4840 - val_loss: 2.3173 - val_sparse_categorical_accuracy: 0.3550
Epoch 2/60
390/390 [==============================] - 43s 110ms/step - loss: 0.9640 - sparse_categorical_accuracy: 0.6601 - val_loss: 1.2524 - val_sparse_categorical_accuracy: 0.5868
Epoch 3/60
390/390 [==============================] - 43s 111ms/step - loss: 0.7742 - sparse_categorical_accuracy: 0.7273 - val_loss: 1.4201 - val_sparse_categorical_accuracy: 0.5968
...
Epoch 59/60
390/390 [==============================] - 43s 110ms/step - loss: 0.0483 - sparse_categorical_accuracy: 0.9824 - val_loss: 0.3880 - val_sparse_categorical_accuracy: 0.9106
Epoch 60/60
390/390 [==============================] - 43s 110ms/step - loss: 0.0501 - sparse_categorical_accuracy: 0.9830 - val_loss: 0.4432 - val_sparse_categorical_accuracy: 0.9006

最后达到训练精度98.3%,测试精度90.06%。
绘制训练曲线

# plot train history
loss = history.history['loss']
val_loss = history.history['val_loss']
acc = history.history['sparse_categorical_accuracy']
val_acc = history.history['val_sparse_categorical_accuracy']

plt.figure(figsize=(11,3.5))

plt.subplot(1,2,1)
plt.plot(loss, color='blue', label='train')
plt.plot(val_loss, color='red', label='test')
plt.ylabel('loss')
plt.legend()

plt.subplot(1,2,2)
plt.plot(acc, color='blue', label='train')
plt.plot(val_acc, color='red', label='test')
plt.ylabel('accuracy')
plt.legend()

 

结果如下


转载:https://blog.csdn.net/mbdong/article/details/127925280
查看评论
* 以上用户言论只代表其个人观点,不代表本网站的观点或立场