1 #coding = utf-8
2 from tensorflow.examples.tutorials.mnist import input_data
3 import tensorflow as tf
4
5 mnist = input_data.read_data_sets("MNIST_data/", one_hot = True)
6 #创建一个IntercativeSession,这样后面的操作就无需指定Session
7 sess = tf.InteractiveSession()
8
9 #隐含层输出节点设置为300,(在此模型中隐含节点数设在200~1000结果区别不大)
10 in_units = 784
11 h1_units = 300
12 #利用tf.truncated_normal实现截断的正态分布,其标准差为0.1 [-1, 784]x[784, 300]
13 w1 = tf.Variable(tf.truncated_normal([in_units, h1_units], stddev=0.1))
14 b1 = tf.Variable(tf.zeros([h1_units]))
15 #[784, 300]x[300, 10]
16 w2 = tf.Variable(tf.zeros([h1_units, 10]))
17 b2 = tf.Variable(tf.zeros([10]))
18
19 #定义输入x,Dropout的比率keep_prob(通常在训练时小于1,而预测时等于1)
20 x = tf.placeholder(tf.float32, [None, in_units])
21 y_ = tf.placeholder(tf.float32, [None, 10])
22 keep_prob = tf.placeholder(tf.float32)
23
24 #hidden1:隐含层 y = relu(W1*x+b1)
25 hidden1 = tf.nn.relu(tf.matmul(x, w1) + b1)
26 '''调用tf.nn.dropout实现Dropout,keep_prob在训练时小于1,用于制造随机性,防止过拟合;
27 在预测时等于1,即使用全部特征来预测样本类别'''
28 hidden1_drop = tf.nn.dropout(hidden1, keep_prob)
29
30 #prediction
31 y = tf.nn.softmax(tf.matmul(hidden1_drop, w2)+b2)
32
33 cross_entropy = tf.reduce_mean(-tf.reduce_sum(y_*tf.log(y), reduction_indices=[1]))
34 trian_step = tf.train.AdagradOptimizer(0.3).minimize(cross_entropy)
35
36 tf.global_variables_initializer().run()
37
38 for i in range(3000):
39 batch_xs, batch_ys = mnist.train.next_batch(100)
40 trian_step.run({x: batch_xs, y_: batch_ys, keep_prob: 0.75})
41 #out correct prediction
42 correct_prediction = tf.equal(tf.argmax(y, 1), tf.argmax(y_, 1))
43 accuracy = tf.reduce_mean(tf.cast(correct_prediction, tf.float32))
44 if i % 500 == 0:
45 print(accuracy.eval({x: mnist.test.images, y_: mnist.test.labels, keep_prob: 1.0}))
1 0.2318
2 0.9584
3 0.9709
4 0.9761
5 0.9778
6 0.9782