美文网首页TensorFlow HOWTO程序员
TensorFlow HOWTO 5.1 循环神经网络(时间序列

TensorFlow HOWTO 5.1 循环神经网络(时间序列

作者: 布客飞龙 | 来源:发表于2018-12-26 17:07 被阅读68次

    5.1 循环神经网络(时间序列)

    循环神经网络(RNN)用于建模带有时间关系的数据。它的架构是这样的。

    image

    在最基本的 RNN 中,单元(方框)中的操作和全连接层没什么区别,都是线性变换和激活。它完全可以看做多个全连接层的横向扩展。

    但是操作数量多了之后,就会有梯度消失和爆炸的问题,于是人们改良了 RNN 单元,添加了精巧的结构来避免这样问题。这是 RNN 的几种改良结构:

    image

    操作步骤

    导入所需的包。

    import tensorflow as tf
    import numpy as np
    import pandas as pd
    import matplotlib as mpl
    import matplotlib.pyplot as plt
    

    导入数据,并进行预处理。我们使用国际航班乘客数据集,由于它不存在于任何现有库中,我们需要先下载它。

    ts = pd.read_csv('international-airline-passengers.csv', usecols=[1], header=0).dropna().values.ravel()
    

    之后,我们需要将其转换为 RNN 单元可以接受的格式。可接受的格式是三维的,维度分别表示样本、时间和特征。我们需要一个窗口大小,表示几个历史值与当前值有关,然后我们按照它来切分时间序列,就能得到样本。

    我仅仅使用原始特征,也就是乘客数量。我并不是表明 RNN 预测得有多好,只是告诉大家 RNN 怎么用。以后有了更好的特征再补充。

    wnd_sz = 5
    ds = []
    for i in range(0, len(ts) - wnd_sz + 1):
        ds.append(ts[i:i + wnd_sz])
    ds = np.asarray(ds)
    
    x_ = ds[:, 0:wnd_sz - 1]
    y_ = ds[:, [wnd_sz - 1]]
    x_ = np.expand_dims(x_, 2)
    

    之后是训练集和测试集的划分。同样要注意绝对不能打乱。

    train_size = int(len(x_) * 0.7)
    x_train = x_[:train_size]
    y_train = y_[:train_size]
    x_test = x_[train_size:]
    y_test  = y_[train_size:]
    

    由于 RNN 单元里面使用了tanh激活函数,它对于数据的尺度非常敏感,我们需要将数据标准化。但是,如果我们遵循传统方法,就发现效果不是很好,RNN 甚至不能预测趋势。观察数据集就会发现,后面的数据明显大于前面的数据。如果我们使用训练集的均值和标准差来标准化,仍旧是如此。换句话说,对于训练集来说,测试集是“新的分布”。所以我们分别计算训练集和测试集的均值和标准差,并使用它们自己的均值和标准差,将训练集和测试集标准化。

    x_mu_train = x_train.mean(0)
    x_sigma_train = x_train.std(0)
    y_mu_train = y_train.mean(0)
    y_sigma_train = y_train.std(0)
    x_train = (x_train - x_mu_train) / x_sigma_train
    y_train = (y_train - y_mu_train) / y_sigma_train
    x_mu_test = x_test.mean(0)
    x_sigma_test = x_test.std(0)
    y_mu_test = y_test.mean(0)
    y_sigma_test = y_test.std(0)
    x_test = (x_test - x_mu_test) / x_sigma_test
    y_test = (y_test - y_mu_test) / y_sigma_test
    

    定义超参数。

    变量 含义
    n_step 时间步长
    n_input 样本特征数
    n_epoch 迭代数
    n_hidden 循环层的单元数
    n_output 输出层单元数
    lr 学习率
    n_step = wnd_sz - 1
    n_input = 1
    n_hidden = 4
    n_output = 1
    n_epoch = 10000
    lr = 0.1
    

    搭建模型。循环层之后添加了一个输出层,目的是把循环层输出的四个特征压缩为一个特征,与标签匹配。

    变量 含义
    x 输入
    y 真实标签
    cell 循环层
    w_l2 输出层的权重
    b_l2 输出层的偏置
    h_l1 循环层的输出
    h_l2 模型的输出
    x = tf.placeholder(tf.float64, [None, n_step, n_input])
    y = tf.placeholder(tf.float64, [None, n_output])
    cell = tf.nn.rnn_cell.GRUCell(n_hidden)
    w_l2 = tf.Variable(np.random.rand(n_hidden, n_output))
    b_l2 = tf.Variable(np.random.rand(1, n_output))
    
    h_l1, _ = tf.nn.dynamic_rnn(cell, x, dtype=tf.float64)
    h_l2 = h_l1[:, -1] @ w_l2 + b_l2
    

    定义 MSE 损失、优化操作、和 R 方度量指标。

    变量 含义
    loss 损失
    op 优化操作
    r_sqr R 方
    loss = tf.reduce_mean((h_l2 - y) ** 2)
    op = tf.train.AdamOptimizer(lr).minimize(loss)
    
    y_mean = tf.reduce_mean(y)
    r_sqr = 1 - tf.reduce_sum((y - h_l2) ** 2) / tf.reduce_sum((y - y_mean) ** 2)
    

    使用训练集训练模型。

    losses = []
    r_sqrs = []
    
    with tf.Session() as sess:
        sess.run(tf.global_variables_initializer())
        
        for e in range(n_epoch):
            _, loss_ = sess.run([op, loss], feed_dict={x: x_train, y: y_train})
            losses.append(loss_)
    

    使用测试集计算 R 方。

            r_sqr_ = sess.run(r_sqr, feed_dict={x: x_test, y: y_test})
            r_sqrs.append(r_sqr_)
    
    

    每一百步打印损失和度量值。

            if e % 100 == 0:
                print(f'epoch: {e}, loss: {loss_}, r_sqr: {r_sqr_}')
    

    得到模型对训练特征和测试特征的预测值。

        y_train_pred = sess.run(h_l2, feed_dict={x: x_train})
        y_test_pred = sess.run(h_l2, feed_dict={x: x_test})
        y_train_pred = y_train_pred * y_sigma_train + y_mu_train
        y_test_pred = y_test_pred * y_sigma_test + y_mu_test
    

    输出:

    epoch: 0, loss: 1.172066950288726, r_sqr: 0.08181895235767012
    epoch: 100, loss: 0.06762743481295605, r_sqr: 0.74114029701611
    epoch: 200, loss: 0.04335752219908887, r_sqr: 0.6723681622904742
    epoch: 300, loss: 0.030765678607308333, r_sqr: 0.6122519542608709
    epoch: 400, loss: 0.0255667258649117, r_sqr: 0.5849438991657725
    epoch: 500, loss: 0.027757059089794138, r_sqr: 0.5547220968537527
    epoch: 600, loss: 0.02280867896340781, r_sqr: 0.5770103850696798
    epoch: 700, loss: 0.021203888157209534, r_sqr: 0.5588744061563752
    epoch: 800, loss: 0.020631124908383643, r_sqr: 0.5340900532988054
    epoch: 900, loss: 0.01848138445620317, r_sqr: 0.4793500483558242
    epoch: 1000, loss: 0.01750574268333934, r_sqr: 0.4363987472239601
    epoch: 1100, loss: 0.019063501795946507, r_sqr: 0.39722141072773787
    epoch: 1200, loss: 0.0162686638744368, r_sqr: 0.3656020052876743
    epoch: 1300, loss: 0.01700178459519973, r_sqr: 0.34215312925023267
    epoch: 1400, loss: 0.01680496271967236, r_sqr: 0.32958240891225643
    epoch: 1500, loss: 0.022387361152892183, r_sqr: 0.311875062539985
    epoch: 1600, loss: 0.01692835079726568, r_sqr: 0.31973211945726576
    epoch: 1700, loss: 0.014981184565736732, r_sqr: 0.34878737027421214
    epoch: 1800, loss: 0.015259222044438638, r_sqr: 0.34762914874016715
    epoch: 1900, loss: 0.013233242381529384, r_sqr: 0.33428957917112656
    epoch: 2000, loss: 0.013406887787008756, r_sqr: 0.33925822033716413
    epoch: 2100, loss: 0.014663732662643054, r_sqr: 0.347748811545596
    epoch: 2200, loss: 0.013725021637547831, r_sqr: 0.3537942857251979
    epoch: 2300, loss: 0.012930256601236268, r_sqr: 0.34525246614475136
    epoch: 2400, loss: 0.013634394369979042, r_sqr: 0.31935143022735935
    epoch: 2500, loss: 0.012295407249325577, r_sqr: 0.32843801172553166
    epoch: 2600, loss: 0.012086876034542369, r_sqr: 0.2944659162449187
    epoch: 2700, loss: 0.011431132027934963, r_sqr: 0.3320304352651986
    epoch: 2800, loss: 0.026766484877065972, r_sqr: 0.3516603080525481
    epoch: 2900, loss: 0.013249484588141427, r_sqr: 0.32412481149383066
    epoch: 3000, loss: 0.01105305279694339, r_sqr: 0.3292310540476926
    epoch: 3100, loss: 0.011470806104939772, r_sqr: 0.3297084772513311
    epoch: 3200, loss: 0.011529391738337445, r_sqr: 0.32604568271503975
    epoch: 3300, loss: 0.016529603402840765, r_sqr: 0.3043582054899223
    epoch: 3400, loss: 0.013990350362048133, r_sqr: 0.3199655109188301
    epoch: 3500, loss: 0.01153881610786597, r_sqr: 0.3133999574164319
    epoch: 3600, loss: 0.012797147535482193, r_sqr: 0.31258610293795597
    epoch: 3700, loss: 0.011476855518680756, r_sqr: 0.34671798591970837
    epoch: 3800, loss: 0.010467519750448642, r_sqr: 0.3104500031552391
    epoch: 3900, loss: 0.010726962231936567, r_sqr: 0.3138785388392976
    epoch: 4000, loss: 0.011354515890079096, r_sqr: 0.3439112763768726
    epoch: 4100, loss: 0.010956935944379312, r_sqr: 0.33047301867737333
    epoch: 4200, loss: 0.010516217542845503, r_sqr: 0.3169866902021138
    epoch: 4300, loss: 0.017644784078390447, r_sqr: 0.30790429205962355
    epoch: 4400, loss: 0.010311798410482015, r_sqr: 0.33750001847417244
    epoch: 4500, loss: 0.012096887824821146, r_sqr: 0.3483536401639381
    epoch: 4600, loss: 0.010336452329588753, r_sqr: 0.333357260461554
    epoch: 4700, loss: 0.010768585264597813, r_sqr: 0.3286332805909209
    epoch: 4800, loss: 0.010187296826748472, r_sqr: 0.3508737658734946
    epoch: 4900, loss: 0.012830964380652418, r_sqr: 0.38883425472040245
    epoch: 5000, loss: 0.010034723893693883, r_sqr: 0.34865494006975606
    epoch: 5100, loss: 0.018500185229121037, r_sqr: 0.37866142043477613
    epoch: 5200, loss: 0.010063261243586386, r_sqr: 0.34191028337439755
    epoch: 5300, loss: 0.011189870387229503, r_sqr: 0.34997022979225245
    epoch: 5400, loss: 0.011487597291485683, r_sqr: 0.31996396612360156
    epoch: 5500, loss: 0.018130474919110774, r_sqr: 0.40163834751360294
    epoch: 5600, loss: 0.010631864046823009, r_sqr: 0.31888613030550417
    epoch: 5700, loss: 0.010235333409754856, r_sqr: 0.33371577618904724
    epoch: 5800, loss: 0.017054583875343695, r_sqr: 0.4573069550292207
    epoch: 5900, loss: 0.010375495082076958, r_sqr: 0.34263745746177443
    epoch: 6000, loss: 0.02899888987961715, r_sqr: 0.4731489130962029
    epoch: 6100, loss: 0.009635932593390793, r_sqr: 0.3377314247566131
    epoch: 6200, loss: 0.009756606958491803, r_sqr: 0.3433460486517317
    epoch: 6300, loss: 0.015849799410205617, r_sqr: 0.38774652539473653
    epoch: 6400, loss: 0.00952936061724416, r_sqr: 0.3316576525214012
    epoch: 6500, loss: 0.01726899798192304, r_sqr: 0.3273077681611758
    epoch: 6600, loss: 0.009523356787685604, r_sqr: 0.3307726070509249
    epoch: 6700, loss: 0.012766831260869482, r_sqr: 0.3307548009627649
    epoch: 6800, loss: 0.009289520442372325, r_sqr: 0.32373411787154416
    epoch: 6900, loss: 0.011158193836457308, r_sqr: 0.32997251352858914
    epoch: 7000, loss: 0.02869441623350593, r_sqr: 0.3338995201916093
    epoch: 7100, loss: 0.018485258063551675, r_sqr: 0.2938792287466744
    epoch: 7200, loss: 0.008861181401821593, r_sqr: 0.32004037549190045
    epoch: 7300, loss: 0.00867730190276807, r_sqr: 0.31307907559266057
    epoch: 7400, loss: 0.010568595192759116, r_sqr: 0.3007627324642683
    epoch: 7500, loss: 0.008931675261205094, r_sqr: 0.3355236299841886
    epoch: 7600, loss: 0.01437912261193081, r_sqr: 0.3118478993546139
    epoch: 7700, loss: 0.013553350799683056, r_sqr: 0.3492071104037119
    epoch: 7800, loss: 0.030876872968019105, r_sqr: 0.31738784411969434
    epoch: 7900, loss: 0.010089701030161817, r_sqr: 0.3301184791799566
    epoch: 8000, loss: 0.01147013359703656, r_sqr: 0.30306917655113186
    epoch: 8100, loss: 0.009855414853123119, r_sqr: 0.34582101555677947
    epoch: 8200, loss: 0.014423066317111677, r_sqr: 0.3470667522167451
    epoch: 8300, loss: 0.011664752107448714, r_sqr: 0.32157739869359203
    epoch: 8400, loss: 0.009646714124369984, r_sqr: 0.3355179490543708
    epoch: 8500, loss: 0.008793324000994002, r_sqr: 0.2839973244304941
    epoch: 8600, loss: 0.009740176108560508, r_sqr: 0.324717537310496
    epoch: 8700, loss: 0.008509925688666948, r_sqr: 0.29162404590044055
    epoch: 8800, loss: 0.00935785011763191, r_sqr: 0.2925641966682784
    epoch: 8900, loss: 0.010778117882046828, r_sqr: 0.2969009246659946
    epoch: 9000, loss: 0.008386197722733947, r_sqr: 0.31203296266480274
    epoch: 9100, loss: 0.01152050463672165, r_sqr: 0.32619935954711543
    epoch: 9200, loss: 0.008279818958514217, r_sqr: 0.3172741922458261
    epoch: 9300, loss: 0.00767664017767626, r_sqr: 0.33894394589120425
    epoch: 9400, loss: 0.0093665602478405, r_sqr: 0.3409878486332283
    epoch: 9500, loss: 0.014677227625938132, r_sqr: 0.3553809732953209
    epoch: 9600, loss: 0.008036392158721397, r_sqr: 0.33729133931823474
    epoch: 9700, loss: 0.011275509041766035, r_sqr: 0.30956775357004673
    epoch: 9800, loss: 0.00885348178403466, r_sqr: 0.3402377958439502
    epoch: 9900, loss: 0.009747783923937898, r_sqr: 0.3843754214763523
    

    绘制时间序列及其预测值。

    plt.figure()
    plt.plot(ts, label='Original')
    y_train_pred = np.concatenate([
        [np.nan] * n_input, 
        y_train_pred.ravel()
    ])
    y_test_pred = np.concatenate([
        [np.nan] * (n_input + train_size),
        y_test_pred.ravel()
    ])
    plt.plot(y_train_pred, label='y_train_pred')
    plt.plot(y_test_pred, label='y_test_pred')
    plt.legend()
    plt.show()
    
    image

    绘制训练集上的损失。

    plt.figure()
    plt.plot(losses)
    plt.title('Loss on Training Set')
    plt.xlabel('#epoch')
    plt.ylabel('MSE')
    plt.show()
    
    image

    绘制测试集上的 R 方。

    plt.figure()
    plt.plot(r_sqrs)
    plt.title('$R^2$ on Testing Set')
    plt.xlabel('#epoch')
    plt.ylabel('$R^2$')
    plt.show()
    
    image

    扩展阅读

    相关文章

      网友评论

        本文标题:TensorFlow HOWTO 5.1 循环神经网络(时间序列

        本文链接:https://www.haomeiwen.com/subject/cuxklqtx.html