Advertisement
Guest User

Untitled

a guest
Mar 28th, 2017
43
0
Never
Not a member of Pastebin yet? Sign Up, it unlocks many cool features!
text 5.35 KB | None | 0 0
  1. name: "pretrain_cnn_lstm_softmax"
  2. layer {
  3. name: "data"
  4. type: "Python"
  5. top: "data"
  6. top: "label"
  7. top: "clip_markers"
  8. python_param {
  9. module: "obj_input_layer"
  10. layer: "videoReadTrain"
  11. }
  12. include: { phase: TRAIN }
  13. }
  14.  
  15. layer {
  16. name: "data"
  17. type: "Python"
  18. top: "data"
  19. top: "label"
  20. top: "clip_markers"
  21. python_param {
  22. module: "obj_input_layer"
  23. layer: "videoReadTest"
  24. }
  25. include: { phase: TEST stage: "test-on-test" }
  26. }
  27.  
  28. layer {
  29. name: "conv1"
  30. type: "Convolution"
  31. bottom: "data"
  32. top: "conv1"
  33. param {
  34. lr_mult: 1
  35. decay_mult: 1
  36. }
  37. param {
  38. lr_mult: 2
  39. decay_mult: 0
  40. }
  41. convolution_param {
  42. num_output: 96
  43. kernel_size: 7
  44. stride: 2
  45. weight_filler {
  46. type: "gaussian"
  47. std: 0.01
  48. }
  49. bias_filler {
  50. type: "constant"
  51. value: 0.1
  52. }
  53. }
  54. }
  55. layer {
  56. name: "relu1"
  57. type: "ReLU"
  58. bottom: "conv1"
  59. top: "conv1"
  60. }
  61. layer {
  62. name: "pool1"
  63. type: "Pooling"
  64. bottom: "conv1"
  65. top: "pool1"
  66. pooling_param {
  67. pool: MAX
  68. kernel_size: 3
  69. stride: 2
  70. }
  71. }
  72. layer {
  73. name: "norm1"
  74. type: "LRN"
  75. bottom: "pool1"
  76. top: "norm1"
  77. lrn_param {
  78. local_size: 5
  79. alpha: 0.0001
  80. beta: 0.75
  81. }
  82. }
  83. layer {
  84. name: "conv2"
  85. type: "Convolution"
  86. bottom: "norm1"
  87. top: "conv2"
  88. param {
  89. lr_mult: 1
  90. decay_mult: 1
  91. }
  92. param {
  93. lr_mult: 2
  94. decay_mult: 0
  95. }
  96. convolution_param {
  97. num_output: 384
  98. kernel_size: 5
  99. group: 2
  100. stride: 2
  101. weight_filler {
  102. type: "gaussian"
  103. std: 0.01
  104. }
  105. bias_filler {
  106. type: "constant"
  107. value: 0.1
  108. }
  109. }
  110. }
  111. layer {
  112. name: "relu2"
  113. type: "ReLU"
  114. bottom: "conv2"
  115. top: "conv2"
  116. }
  117. layer {
  118. name: "pool2"
  119. type: "Pooling"
  120. bottom: "conv2"
  121. top: "pool2"
  122. pooling_param {
  123. pool: MAX
  124. kernel_size: 3
  125. stride: 2
  126. }
  127. }
  128. layer {
  129. name: "norm2"
  130. type: "LRN"
  131. bottom: "pool2"
  132. top: "norm2"
  133. lrn_param {
  134. local_size: 5
  135. alpha: 0.0001
  136. beta: 0.75
  137. }
  138. }
  139. layer {
  140. name: "conv3"
  141. type: "Convolution"
  142. bottom: "norm2"
  143. top: "conv3"
  144. param {
  145. lr_mult: 1
  146. decay_mult: 1
  147. }
  148. param {
  149. lr_mult: 2
  150. decay_mult: 0
  151. }
  152. convolution_param {
  153. num_output: 512
  154. pad: 1
  155. kernel_size: 3
  156. weight_filler {
  157. type: "gaussian"
  158. std: 0.01
  159. }
  160. bias_filler {
  161. type: "constant"
  162. value: 0.1
  163. }
  164. }
  165. }
  166. layer {
  167. name: "relu3"
  168. type: "ReLU"
  169. bottom: "conv3"
  170. top: "conv3"
  171. }
  172. layer {
  173. name: "conv4"
  174. type: "Convolution"
  175. bottom: "conv3"
  176. top: "conv4"
  177. param {
  178. lr_mult: 1
  179. decay_mult: 1
  180. }
  181. param {
  182. lr_mult: 2
  183. decay_mult: 0
  184. }
  185. convolution_param {
  186. num_output: 512
  187. pad: 1
  188. kernel_size: 3
  189. group: 2
  190. weight_filler {
  191. type: "gaussian"
  192. std: 0.01
  193. }
  194. bias_filler {
  195. type: "constant"
  196. value: 0.1
  197. }
  198. }
  199. }
  200. layer {
  201. name: "relu4"
  202. type: "ReLU"
  203. bottom: "conv4"
  204. top: "conv4"
  205. }
  206. layer {
  207. name: "conv5"
  208. type: "Convolution"
  209. bottom: "conv4"
  210. top: "conv5"
  211. param {
  212. lr_mult: 1
  213. decay_mult: 1
  214. }
  215. param {
  216. lr_mult: 2
  217. decay_mult: 0
  218. }
  219. convolution_param {
  220. num_output: 384
  221. pad: 1
  222. kernel_size: 3
  223. group: 2
  224. weight_filler {
  225. type: "gaussian"
  226. std: 0.01
  227. }
  228. bias_filler {
  229. type: "constant"
  230. value: 0.1
  231. }
  232. }
  233. }
  234. layer {
  235. name: "relu5"
  236. type: "ReLU"
  237. bottom: "conv5"
  238. top: "conv5"
  239. }
  240. layer {
  241. name: "pool5"
  242. type: "Pooling"
  243. bottom: "conv5"
  244. top: "pool5"
  245. pooling_param {
  246. pool: MAX
  247. kernel_size: 3
  248. stride: 2
  249. }
  250. }
  251. layer {
  252. name: "fc6"
  253. type: "InnerProduct"
  254. bottom: "pool5"
  255. top: "fc6"
  256. param {
  257. lr_mult: 1
  258. decay_mult: 1
  259. }
  260. param {
  261. lr_mult: 2
  262. decay_mult: 0
  263. }
  264. inner_product_param {
  265. num_output: 4096
  266. weight_filler {
  267. type: "gaussian"
  268. std: 0.01
  269. }
  270. bias_filler {
  271. type: "constant"
  272. value: 0.1
  273. }
  274. }
  275. }
  276. layer {
  277. name: "relu6"
  278. type: "ReLU"
  279. bottom: "fc6"
  280. top: "fc6"
  281. }
  282. layer {
  283. name: "drop6"
  284. type: "Dropout"
  285. bottom: "fc6"
  286. top: "fc6"
  287. dropout_param {
  288. dropout_ratio: 0.9
  289. }
  290. }
  291. layer{
  292. name: "reshape-data"
  293. type: "Reshape"
  294. bottom: "fc6"
  295. top: "fc6-reshape"
  296. reshape_param{
  297. shape{
  298. dim: 32 ################ parameter
  299. dim: 5 ################
  300. dim: 4096
  301. }
  302. }
  303. }
  304.  
  305. layer{
  306. name: "reshape-cm"
  307. type: "Reshape"
  308. bottom: "clip_markers"
  309. top: "reshape-cm"
  310. reshape_param{
  311. shape{
  312. dim: 30 ################# Parameter
  313. dim: 5 #################
  314. }
  315. }
  316. }
  317.  
  318. layer {
  319. name: "lstm1"
  320. type: "LSTM"
  321. bottom: "fc6-reshape"
  322. bottom: "reshape-cm"
  323. top: "lstm1"
  324. recurrent_param {
  325. num_output: 4096
  326. weight_filler {
  327. type: "uniform"
  328. min: -0.01
  329. max: 0.01
  330. }
  331. bias_filler {
  332. type: "constant"
  333. value: 0
  334. }
  335. }
  336. }
  337.  
  338. layer {
  339. name: "lstm1-drop"
  340. type: "Dropout"
  341. bottom: "lstm1"
  342. top: "lstm1-drop"
  343. dropout_param {
  344. dropout_ratio: 0.5
  345. }
  346. }
  347.  
  348. layer {
  349. name: "fc8-final"
  350. type: "InnerProduct"
  351. bottom: "lstm1-drop"
  352. top: "fc8-final"
  353. param {
  354. lr_mult: 10
  355. decay_mult: 1
  356. }
  357. param {
  358. lr_mult: 20
  359. decay_mult: 0
  360. }
  361. inner_product_param {
  362. num_output: 10 ######################
  363. weight_filler {
  364. type: "gaussian"
  365. std: 0.001
  366. }
  367. bias_filler {
  368. type: "constant"
  369. value: 0.1
  370. }
  371. }
  372. }
  373.  
  374. layer{
  375. name: "loss"
  376. type: "SoftmaxWithLoss"
  377. bottom: "fc8-final"
  378. bottom: "label"
  379. top: "loss"
  380. }
  381.  
  382. layer{
  383. name: "accuracy"
  384. type: "Accuracy"
  385. bottom: "fc8-final"
  386. bottom: "label"
  387. top: "accuracy"
  388. }
Advertisement
Add Comment
Please, Sign In to add comment
Advertisement