• 循环神经网络(五)-LSTM进阶


    基础的LSTM模型,单隐层,隐层单神经元,而实际中一般需要更为复杂的网络结构

    下面借用手写数字的经典案例构造比较复杂的LSTM模型,并用代码实现。

    单隐层,隐层多神经元

    # -*- coding:utf-8 -*-
    import tensorflow as tf
    from tensorflow.examples.tutorials.mnist import input_data
    
    # 导入数据
    mnist = input_data.read_data_sets('MNIST_data',one_hot = True)
    
    training_iters = 50001
    batch_size = 100          # 批量大小
    
    n_inputs = 28
    n_steps = 28             # 序列长度
    n_hidden_number = 128    # 隐藏层神经元个数
    n_outputs = 10           # 输出层神经元个数
    
    x = tf.placeholder(tf.float32,[None,n_steps,n_inputs])      # 输入3维  样本数*序列长度*每个元素长度
    Y = tf.placeholder(tf.float32,[None,n_outputs])
    
    weights = {
        # 权重设定同全连接,这是单步权重
        # 每步权重共享
        # shape = (28,128)
        'in':tf.Variable(tf.random_normal([n_inputs,n_hidden_number])),
        # shape = (128,10)
        'out':tf.Variable(tf.random_normal([n_hidden_number,n_outputs]))}
    
    biases = {
        # shape = (128,)
        'in':tf.Variable(tf.constant(0.1,shape = [n_hidden_number,])),
        # shape = (10,)
        'out':tf.Variable(tf.constant(0.1,shape = [n_outputs,]))}
    
    
    def RNN(X,weights,biases):
        """ 这里是整个RNN的网络
        1. 输入x是整个序列的,即 [x1,x2,..xt]*samples
        2. 隐层的输出包含了序列中每个时间的输出
        """
        ### 输入层到核运算 ###
        # X shape = (100batch,28steps,28inputs) ==> (100batch*28steps,28inputs)
        X = tf.reshape(X,[-1,n_inputs])     # 之所以转成2维,是要与weight相乘
        # X_in shape ==> (100batch*28steps,128hidden)
        X_in = tf.matmul(X,weights['in'])+biases['in']
        # 这是所有样本每个时序都乘以weight,每个时序都变成了hidden_num长度,结果是所有样本每个时序的hidden_num长,此时所有样本混在一起
    
        # X_in shape ==> (100batch,28steps,128hidden)
        # 然后要重新划分成样本,每个样本有n_steps个时序,每个时序是hidden_num长,按样本送给隐层
        X_in = tf.reshape(X_in,[-1,n_steps,n_hidden_number])
    
        ### cell核内运算 ###
        ## 构建单个rnn cell,单隐层,n_hidden_number个神经元,横向网络会根据输入(n_steps)自动构建
        lstm_cell = tf.nn.rnn_cell.BasicLSTMCell(n_hidden_number,forget_bias = 1.0)
    
        # LSTM cell is divided into two parts-->(c_state,m_state)
        # 初始化s
        init_state = lstm_cell.zero_state(batch_size,dtype=tf.float32)
    
        # outputs states 都是隐层的输出,整个RNN所有时序的输出,但只是h,还没到o  o=vh+c(c是bias,不同于states的c,这是记忆单元)
        # outputs 是以三维矩阵的形式记录了所有样本所有时序的所有隐层神经元的输出,shape为[batch_size, timestep_size, hidden_size]
        # states 是最后时刻的c 和 h,c是记忆,shape为 [batch_size, 2, hidden_size]
            # 当然通常顺序是这样的[2, batch_size, hidden_size]
        outputs,states = tf.nn.dynamic_rnn(lstm_cell,X_in,initial_state=init_state,time_major=False)
    
        ### 核内运算到输出层 ###
        # states[1]就是所有样本最后时刻的h, 实际上 states[1] == outputs[:,-1,:]
        result = tf.matmul(states[1],weights['out'])+biases['out']  # result就是o
        return  result
    
    prediction = RNN(x,weights,biases)
    
    ### 后面所有的神经网络都大同小异
    loss = tf.reduce_mean(tf.nn.softmax_cross_entropy_with_logits(labels=Y, logits=prediction))
    train_step = tf.train.AdamOptimizer(1e-3).minimize(loss)
    correct_prediction = tf.equal(tf.argmax(prediction,1), tf.argmax(Y,1))
    accuracy = tf.reduce_mean(tf.cast(correct_prediction, tf.float32))
    
    init = tf.global_variables_initializer()
    with tf.Session() as sess:
        sess.run(init)
        step = 0
        while step*batch_size < training_iters:
            batch_xs,batch_ys = mnist.train.next_batch(batch_size)
            # batch_xs shape = [100,28,28]
            batch_xs = batch_xs.reshape([batch_size,n_steps,n_inputs])
            train_step.run(feed_dict={x:batch_xs,Y:batch_ys,})
            if step%50 == 0:
                train_accuracy = accuracy.eval(feed_dict={x:batch_xs,Y:batch_ys,})
                print("step", step, "training accuracy", train_accuracy)
            step += 1

    上面的网络大概是这样

     

     x1 x2 其实应该是 xt1 xt2

    多层隐层,隐层多节点

    layer_num = 2  # 隐层数 
    
    def clstm():
        lstm_cell = tf.nn.rnn_cell.BasicLSTMCell(num_units=hidden_size, forget_bias=1.0, state_is_tuple=True)
        lstm_cell = rnn.DropoutWrapper(cell=lstm_cell, input_keep_prob=1.0, output_keep_prob=keep_prob)
        return lstm_cell
    
    # 调用 MultiRNNCell 来实现多层 LSTM
    mlstm_cell = rnn.MultiRNNCell([clstm() for i in range(layer_num)], state_is_tuple=True)
    
    init_state = mlstm_cell.zero_state(batch_size, dtype=tf.float32)
    
    outputs, state = tf.nn.dynamic_rnn(mlstm_cell, inputs=X, initial_state=init_state, time_major=False)
    h_state = outputs[:, -1, :]  
    

    其他代码雷同 

    预测

    看看rnn是如何预测的

    import matplotlib.pyplot as plt
    current_y= mnist.train.labels[5]
    current_x=mnist.train.images[5]
    print(current_y)
    
    # 计算h current_x.shape
    = [-1, 784] current_y.shape = [-1, class_num] current_outputs = np.array(sess.run(outputs, feed_dict={_X: current_x, y:current_y, keep_prob: 1.0, batch_size: 1})) print(current_outputs.shape) # (1, timesteps, hidden_size) current_outputs.shape = [28, hidden_size] # 计算出参数 h_W = sess.run(W, feed_dict={_X:current_x, y: current_y, keep_prob: 1.0, batch_size: 1}) h_bias = sess.run(bias, feed_dict={_X:current_x, y: current_y, keep_prob: 1.0, batch_size: 1}) h_bias.shape = [-1, 10] bar_index = range(class_num) # 识别过程 for i in range(current_outputs.shape[0]): # current_outputs.shape[0] 是 squence_length plt.subplot(7, 4, i+1) current_h_shate = current_outputs[i, :].reshape([-1, hidden_size]) # 每个时刻的h current_formula=tf.nn.softmax(tf.matmul(current_h_shate, h_W) + h_bias) # 每个时刻的y pro = sess.run(current_formula) plt.bar(bar_index, pro[0], width=0.2 , align='center') plt.axis('off') plt.show()

    每一行显示了 4 个图,共有 7 行,每个图是一个y,表示了一行一行读取过程中,模型对字符的识别。

    可以看到,在只看到前面的几行像素时,模型根本认不出来是什么字符,随着看到的像素越来越多,最后就基本确定了它是字符 4.

  • 相关阅读:
    Android Hal 分析
    Android JNI 使用的数据结构JNINativeMethod详解
    MTK GPIO 一些理解
    DEVICE_ATTR
    DEVICE_ATTR
    内核驱动中常见的miscdevice、platform_device、platform_driver
    DEVICE_ATTR实例分析
    虚拟机安装Ubuntu14.04打开FireFox提示Server not found
    Android编译系统中的Kconfig,Makefile,.config编译系统浅析
    几款在线脑图制作工具
  • 原文地址:https://www.cnblogs.com/yanshw/p/10496290.html
Copyright © 2020-2023  润新知