Guest User

Untitled

a guest
Feb 21st, 2018
83
0
Never
Not a member of Pastebin yet? Sign Up, it unlocks many cool features!
text 3.94 KB | None | 0 0
  1. with tf.device("/gpu:0"):
  2. x = tf.placeholder(tf.float32, [None, feature_dim], name="input")
  3. y_ = tf.placeholder(tf.float32, [None, output_dim],name="output")
  4. phase = tf.placeholder(tf.bool, name='phase')
  5.  
  6. weights = {
  7. 'hidden1': tf.Variable(tf.random_normal([feature_dim, hidden_num_units1], stddev=1 / np.sqrt(feature_dim)), name="w1"),
  8. 'hidden2': tf.Variable(
  9. tf.random_normal([hidden_num_units1, hidden_num_units2], stddev=1 / np.sqrt(hidden_num_units1)), name="w2"),
  10. 'output': tf.Variable(
  11. tf.random_normal([hidden_num_units2, output_num_units], stddev=1 / np.sqrt(hidden_num_units2)), name="w3")
  12. }
  13.  
  14. biases = {
  15. 'hidden1': tf.Variable(tf.random_normal([hidden_num_units1], stddev=1 / np.sqrt(hidden_num_units1)), name="b1"),
  16. 'hidden2': tf.Variable(tf.random_normal([hidden_num_units2], stddev=1 / np.sqrt(hidden_num_units2)), name="b2"),
  17. 'output': tf.Variable(tf.random_normal([output_num_units], stddev=1 / np.sqrt(output_num_units)), name="b3")
  18. }
  19.  
  20. def network1(data):
  21. with tf.name_scope(name="layer_1"):
  22. h1 = tf.add(tf.matmul(data, weights['hidden1']), biases['hidden1'])
  23. h1_BN = tf.contrib.layers.batch_norm(h1,
  24. center=True, scale=True,
  25. is_training=phase,
  26. scope='hidden_layer1_bn')
  27. h1_relu = tf.nn.relu(h1_BN)
  28. with tf.name_scope(name="layer_2"):
  29. h2 = tf.add(tf.matmul(h1_relu, weights['hidden2']), biases['hidden2'])
  30. h2_BN = tf.contrib.layers.batch_norm(h2,
  31. center=True, scale=True,
  32. is_training=phase,
  33. scope='hidden_layer2_bn')
  34. h2_relu = tf.nn.relu(h2_BN)
  35.  
  36. with tf.name_scope(name="output_layer"):
  37. output_layer = tf.add(tf.matmul(h2_relu, weights['output']), biases['output'])
  38. output__layer_BN = tf.contrib.layers.batch_norm(output_layer,
  39. center=True, scale=True,
  40. is_training=phase,
  41. scope='output_bn')
  42. output = tf.sigmoid(output__layer_BN, name="f" )
  43. return output
  44.  
  45. def network1(data):
  46. weights = {
  47. 'hidden1': tf.Variable(tf.random_normal([feature_dim, hidden_num_units1], stddev=1 / np.sqrt(feature_dim)), name="w1"),
  48. 'hidden2': tf.Variable(
  49. tf.random_normal([hidden_num_units1, hidden_num_units2], stddev=1 / np.sqrt(hidden_num_units1)), name="w2"),
  50. 'output': tf.Variable(
  51. tf.random_normal([hidden_num_units2, output_num_units], stddev=1 / np.sqrt(hidden_num_units2)), name="w3")
  52. }
  53.  
  54. biases = {
  55. 'hidden1': tf.Variable(tf.random_normal([hidden_num_units1], stddev=1 / np.sqrt(hidden_num_units1)), name="b1"),
  56. 'hidden2': tf.Variable(tf.random_normal([hidden_num_units2], stddev=1 / np.sqrt(hidden_num_units2)), name="b2"),
  57. 'output': tf.Variable(tf.random_normal([output_num_units], stddev=1 / np.sqrt(output_num_units)), name="b3")
  58. }
  59.  
  60. with tf.name_scope(name = "layer_1"):
  61. hidden_layer1 = tf.add(tf.matmul(data, weights['hidden1']), biases['hidden1'])
  62. hidden_layer1 = tf.nn.relu(hidden_layer1)
  63. with tf.name_scope(name = "layer_1_dropout"):
  64. drop_out1 = tf.nn.dropout(hidden_layer1, keep_prob) # DROP-OUT here
  65. with tf.name_scope(name = "layer_2"):
  66. hidden_layer2 = tf.add(tf.matmul(drop_out1, weights['hidden2']), biases['hidden2'])
  67. hidden_layer2 = tf.nn.relu(hidden_layer2)
  68. with tf.name_scope(name = "layer_2_dropout"):
  69. drop_out2 = tf.nn.dropout(hidden_layer2, keep_prob) # DROP-OUT here
  70. with tf.name_scope(name = "output_layer"):
  71. output = tf.add(tf.matmul(drop_out2, weights['output']), biases['output'])
  72. output = tf.nn.sigmoid(output, name="f")
  73. return output
Add Comment
Please, Sign In to add comment