Advertisement
Guest User

Untitled

a guest
May 23rd, 2016
202
0
Never
Not a member of Pastebin yet? Sign Up, it unlocks many cool features!
text 5.74 KB | None | 0 0
  1. name: "singleFrame_flow"
  2. layer {
  3. name: "data"
  4. type: "ImageData"
  5. top: "data"
  6. top: "label"
  7. include {
  8. phase: TRAIN
  9. }
  10. transform_param {
  11. mirror: true
  12. crop_size: 227
  13. mean_value: 128
  14. mean_value: 128
  15. mean_value: 128
  16. flow: true
  17. }
  18. image_data_param {
  19. source: "ucf_xml_train_list.txt"
  20. root_folder: ""
  21. batch_size: 128
  22. min_height: 227
  23. min_width: 227
  24. }
  25. }
  26. layer {
  27. name: "data"
  28. type: "ImageData"
  29. top: "data"
  30. top: "label"
  31. include {
  32. phase: TEST
  33. stage: "test-on-test"
  34. }
  35. transform_param {
  36. mirror: true
  37. crop_size: 227
  38. mean_value: 128
  39. mean_value: 128
  40. mean_value: 128
  41. flow: true
  42. }
  43. image_data_param {
  44. source: "ucf_xml_test_list.txt"
  45. root_folder: ""
  46. batch_size: 128
  47. min_height: 227
  48. min_width: 227
  49. }
  50. }
  51. layer {
  52. name: "conv1"
  53. type: "Convolution"
  54. bottom: "data"
  55. top: "conv1"
  56. param {
  57. lr_mult: 1
  58. decay_mult: 1
  59. }
  60. param {
  61. lr_mult: 2
  62. decay_mult: 0
  63. }
  64. convolution_param {
  65. num_output: 96
  66. kernel_size: 7
  67. stride: 2
  68. weight_filler {
  69. type: "gaussian"
  70. std: 0.01
  71. }
  72. bias_filler {
  73. type: "constant"
  74. value: 0.1
  75. }
  76. }
  77. }
  78. layer {
  79. name: "relu1"
  80. type: "ReLU"
  81. bottom: "conv1"
  82. top: "conv1"
  83. }
  84. layer {
  85. name: "pool1"
  86. type: "Pooling"
  87. bottom: "conv1"
  88. top: "pool1"
  89. pooling_param {
  90. pool: MAX
  91. kernel_size: 3
  92. stride: 2
  93. }
  94. }
  95. layer {
  96. name: "norm1"
  97. type: "LRN"
  98. bottom: "pool1"
  99. top: "norm1"
  100. lrn_param {
  101. local_size: 5
  102. alpha: 0.0001
  103. beta: 0.75
  104. }
  105. }
  106. layer {
  107. name: "conv2"
  108. type: "Convolution"
  109. bottom: "norm1"
  110. top: "conv2"
  111. param {
  112. lr_mult: 1
  113. decay_mult: 1
  114. }
  115. param {
  116. lr_mult: 2
  117. decay_mult: 0
  118. }
  119. convolution_param {
  120. num_output: 384
  121. kernel_size: 5
  122. group: 2
  123. stride: 2
  124. weight_filler {
  125. type: "gaussian"
  126. std: 0.01
  127. }
  128. bias_filler {
  129. type: "constant"
  130. value: 0.1
  131. }
  132. }
  133. }
  134. layer {
  135. name: "relu2"
  136. type: "ReLU"
  137. bottom: "conv2"
  138. top: "conv2"
  139. }
  140. layer {
  141. name: "pool2"
  142. type: "Pooling"
  143. bottom: "conv2"
  144. top: "pool2"
  145. pooling_param {
  146. pool: MAX
  147. kernel_size: 3
  148. stride: 2
  149. }
  150. }
  151. layer {
  152. name: "norm2"
  153. type: "LRN"
  154. bottom: "pool2"
  155. top: "norm2"
  156. lrn_param {
  157. local_size: 5
  158. alpha: 0.0001
  159. beta: 0.75
  160. }
  161. }
  162. layer {
  163. name: "conv3"
  164. type: "Convolution"
  165. bottom: "norm2"
  166. top: "conv3"
  167. param {
  168. lr_mult: 1
  169. decay_mult: 1
  170. }
  171. param {
  172. lr_mult: 2
  173. decay_mult: 0
  174. }
  175. convolution_param {
  176. num_output: 512
  177. pad: 1
  178. kernel_size: 3
  179. weight_filler {
  180. type: "gaussian"
  181. std: 0.01
  182. }
  183. bias_filler {
  184. type: "constant"
  185. value: 0.1
  186. }
  187. }
  188. }
  189. layer {
  190. name: "relu3"
  191. type: "ReLU"
  192. bottom: "conv3"
  193. top: "conv3"
  194. }
  195. layer {
  196. name: "conv4"
  197. type: "Convolution"
  198. bottom: "conv3"
  199. top: "conv4"
  200. param {
  201. lr_mult: 1
  202. decay_mult: 1
  203. }
  204. param {
  205. lr_mult: 2
  206. decay_mult: 0
  207. }
  208. convolution_param {
  209. num_output: 512
  210. pad: 1
  211. kernel_size: 3
  212. group: 2
  213. weight_filler {
  214. type: "gaussian"
  215. std: 0.01
  216. }
  217. bias_filler {
  218. type: "constant"
  219. value: 0.1
  220. }
  221. }
  222. }
  223. layer {
  224. name: "relu4"
  225. type: "ReLU"
  226. bottom: "conv4"
  227. top: "conv4"
  228. }
  229. layer {
  230. name: "conv5"
  231. type: "Convolution"
  232. bottom: "conv4"
  233. top: "conv5"
  234. param {
  235. lr_mult: 1
  236. decay_mult: 1
  237. }
  238. param {
  239. lr_mult: 2
  240. decay_mult: 0
  241. }
  242. convolution_param {
  243. num_output: 384
  244. pad: 1
  245. kernel_size: 3
  246. group: 2
  247. weight_filler {
  248. type: "gaussian"
  249. std: 0.01
  250. }
  251. bias_filler {
  252. type: "constant"
  253. value: 0.1
  254. }
  255. }
  256. }
  257. layer {
  258. name: "relu5"
  259. type: "ReLU"
  260. bottom: "conv5"
  261. top: "conv5"
  262. }
  263. layer {
  264. name: "pool5"
  265. type: "Pooling"
  266. bottom: "conv5"
  267. top: "pool5"
  268. pooling_param {
  269. pool: MAX
  270. kernel_size: 3
  271. stride: 2
  272. }
  273. }
  274. layer {
  275. name: "fc6"
  276. type: "InnerProduct"
  277. bottom: "pool5"
  278. top: "fc6"
  279. param {
  280. lr_mult: 1
  281. decay_mult: 1
  282. }
  283. param {
  284. lr_mult: 2
  285. decay_mult: 0
  286. }
  287. inner_product_param {
  288. num_output: 4096
  289. weight_filler {
  290. type: "gaussian"
  291. std: 0.01
  292. }
  293. bias_filler {
  294. type: "constant"
  295. value: 0.1
  296. }
  297. }
  298. }
  299. layer {
  300. name: "relu6"
  301. type: "ReLU"
  302. bottom: "fc6"
  303. top: "fc6"
  304. }
  305. layer {
  306. name: "drop6"
  307. type: "Dropout"
  308. bottom: "fc6"
  309. top: "fc6"
  310. dropout_param {
  311. dropout_ratio: 0.5
  312. }
  313. }
  314. layer {
  315. name: "fc7"
  316. type: "InnerProduct"
  317. bottom: "fc6"
  318. top: "fc7"
  319. param {
  320. lr_mult: 1
  321. decay_mult: 1
  322. }
  323. param {
  324. lr_mult: 2
  325. decay_mult: 0
  326. }
  327. inner_product_param {
  328. num_output: 4096
  329. weight_filler {
  330. type: "gaussian"
  331. std: 0.01
  332. }
  333. bias_filler {
  334. type: "constant"
  335. value: 0.1
  336. }
  337. }
  338. }
  339. layer {
  340. name: "relu7"
  341. type: "ReLU"
  342. bottom: "fc7"
  343. top: "fc7"
  344. }
  345. layer {
  346. name: "drop7"
  347. type: "Dropout"
  348. bottom: "fc7"
  349. top: "fc7"
  350. dropout_param {
  351. dropout_ratio: 0.5
  352. }
  353. }
  354. layer {
  355. name: "fc8-ucf"
  356. type: "InnerProduct"
  357. bottom: "fc7"
  358. top: "fc8-ucf"
  359. param {
  360. lr_mult: 10
  361. decay_mult: 1
  362. }
  363. param {
  364. lr_mult: 20
  365. decay_mult: 0
  366. }
  367. inner_product_param {
  368. num_output: 101
  369. weight_filler {
  370. type: "gaussian"
  371. std: 0.01
  372. }
  373. bias_filler {
  374. type: "constant"
  375. value: 0
  376. }
  377. }
  378. }
  379. layer {
  380. name: "loss"
  381. type: "SoftmaxWithLoss"
  382. bottom: "fc8-ucf"
  383. bottom: "label"
  384. top: "loss"
  385. }
  386. layer {
  387. name: "accuracy"
  388. type: "Accuracy"
  389. bottom: "fc8-ucf"
  390. bottom: "label"
  391. top: "accuracy"
  392. }
Advertisement
Add Comment
Please, Sign In to add comment
Advertisement