Guest User

Untitled

a guest
Jan 18th, 2018
69
0
Never
Not a member of Pastebin yet? Sign Up, it unlocks many cool features!
text 3.05 KB | None | 0 0
  1. name: "CIFAR10_full"
  2. layer {
  3. name: "cifar"
  4. type: "Data"
  5. top: "data"
  6. top: "label"
  7. include {
  8. phase: TRAIN
  9. }
  10. transform_param {
  11. mean_file: "examples/cifar10/mean.binaryproto"
  12. }
  13. data_param {
  14. source: "examples/cifar10/cifar10_train_lmdb"
  15. batch_size: 100
  16. backend: LMDB
  17. }
  18. }
  19. layer {
  20. name: "cifar"
  21. type: "Data"
  22. top: "data"
  23. top: "label"
  24. include {
  25. phase: TEST
  26. }
  27. transform_param {
  28. mean_file: "examples/cifar10/mean.binaryproto"
  29. }
  30. data_param {
  31. source: "examples/cifar10/cifar10_test_lmdb"
  32. batch_size: 100
  33. backend: LMDB
  34. }
  35. }
  36. layer {
  37. name: "conv1"
  38. type: "Convolution"
  39. bottom: "data"
  40. top: "conv1"
  41. param {
  42. lr_mult: 1
  43. }
  44. param {
  45. lr_mult: 2
  46. }
  47. convolution_param {
  48. num_output: 32
  49. pad: 2
  50. kernel_size: 5
  51. stride: 1
  52. weight_filler {
  53. type: "gaussian"
  54. std: 0.0001
  55. }
  56. bias_filler {
  57. type: "constant"
  58. }
  59. }
  60. }
  61. layer {
  62. name: "pool1"
  63. type: "Pooling"
  64. bottom: "conv1"
  65. top: "pool1"
  66. pooling_param {
  67. pool: MAX
  68. kernel_size: 3
  69. stride: 2
  70. }
  71. }
  72. layer {
  73. name: "relu1"
  74. type: "ReLU"
  75. bottom: "pool1"
  76. top: "pool1"
  77. }
  78. layer {
  79. name: "norm1"
  80. type: "LRN"
  81. bottom: "pool1"
  82. top: "norm1"
  83. lrn_param {
  84. local_size: 3
  85. alpha: 5e-05
  86. beta: 0.75
  87. norm_region: WITHIN_CHANNEL
  88. }
  89. }
  90. layer {
  91. name: "conv2"
  92. type: "Convolution"
  93. bottom: "norm1"
  94. top: "conv2"
  95. param {
  96. lr_mult: 1
  97. }
  98. param {
  99. lr_mult: 2
  100. }
  101. convolution_param {
  102. num_output: 32
  103. pad: 2
  104. kernel_size: 5
  105. stride: 1
  106. weight_filler {
  107. type: "gaussian"
  108. std: 0.01
  109. }
  110. bias_filler {
  111. type: "constant"
  112. }
  113. }
  114. }
  115. layer {
  116. name: "relu2"
  117. type: "ReLU"
  118. bottom: "conv2"
  119. top: "conv2"
  120. }
  121. layer {
  122. name: "pool2"
  123. type: "Pooling"
  124. bottom: "conv2"
  125. top: "pool2"
  126. pooling_param {
  127. pool: AVE
  128. kernel_size: 3
  129. stride: 2
  130. }
  131. }
  132. layer {
  133. name: "norm2"
  134. type: "LRN"
  135. bottom: "pool2"
  136. top: "norm2"
  137. lrn_param {
  138. local_size: 3
  139. alpha: 5e-05
  140. beta: 0.75
  141. norm_region: WITHIN_CHANNEL
  142. }
  143. }
  144. layer {
  145. name: "conv3"
  146. type: "Convolution"
  147. bottom: "norm2"
  148. top: "conv3"
  149. convolution_param {
  150. num_output: 64
  151. pad: 2
  152. kernel_size: 5
  153. stride: 1
  154. weight_filler {
  155. type: "gaussian"
  156. std: 0.01
  157. }
  158. bias_filler {
  159. type: "constant"
  160. }
  161. }
  162. }
  163. layer {
  164. name: "relu3"
  165. type: "ReLU"
  166. bottom: "conv3"
  167. top: "conv3"
  168. }
  169. layer {
  170. name: "pool3"
  171. type: "Pooling"
  172. bottom: "conv3"
  173. top: "pool3"
  174. pooling_param {
  175. pool: AVE
  176. kernel_size: 3
  177. stride: 2
  178. }
  179. }
  180. layer {
  181. name: "ip1"
  182. type: "InnerProduct"
  183. bottom: "pool3"
  184. top: "ip1"
  185. param {
  186. lr_mult: 1
  187. decay_mult: 250
  188. }
  189. param {
  190. lr_mult: 2
  191. decay_mult: 0
  192. }
  193. inner_product_param {
  194. num_output: 10
  195. weight_filler {
  196. type: "gaussian"
  197. std: 0.01
  198. }
  199. bias_filler {
  200. type: "constant"
  201. }
  202. }
  203. }
  204. layer {
  205. name: "accuracy"
  206. type: "Accuracy"
  207. bottom: "ip1"
  208. bottom: "label"
  209. top: "accuracy"
  210. include {
  211. phase: TEST
  212. }
  213. }
  214. layer {
  215. name: "loss"
  216. type: "SoftmaxWithLoss"
  217. bottom: "ip1"
  218. bottom: "label"
  219. top: "loss"
  220. }
Add Comment
Please, Sign In to add comment