Advertisement
Guest User

Untitled

a guest
Feb 20th, 2017
68
0
Never
Not a member of Pastebin yet? Sign Up, it unlocks many cool features!
text 0.61 KB | None | 0 0
  1. py_x = model(X, w_h, w_o)
  2. cost = tf.reduce_mean(tf.nn.softmax_cross_entropy_with_logits(py_x, Y))
  3. train_op = tf.train.GradientDescentOptimizer(learning_rate=0.01).minimize(cost)
  4. predict_op = tf.argmax(py_x, 1)
  5.  
  6. sess = tf.Session()
  7. init = tf.initialize_all_variables()
  8. sess.run(init)
  9.  
  10. for i in range(700):
  11. for start, end in zip(range(0, len(X_train),1 ), range(1, len(X_train),1)):
  12. sess.run(train_op, feed_dict={X: X_train[start:end], Y: y_train[start:end]})
  13. if i % 100 == 0:
  14. print (i, np.mean(np.argmax(y_test, axis=1) ==
  15. sess.run(predict_op, feed_dict={X: x_test, Y: y_test})))
Advertisement
Add Comment
Please, Sign In to add comment
Advertisement