Guest User

Untitled

a guest
Dec 4th, 2016
69
0
Never
Not a member of Pastebin yet? Sign Up, it unlocks many cool features!
text 6.33 KB | None | 0 0
  1. VGG_ILSVRC_19_layers_train_val.prototxt
  2. name: "VGG_ILSVRC_19_layers"
  3. layers {
  4. name: "data"
  5. type: DATA
  6. include {
  7. phase: TRAIN
  8. }
  9. transform_param {
  10. crop_size: 224
  11. mean_value: 104
  12. mean_value: 117
  13. mean_value: 123
  14. mirror: true
  15. }
  16. data_param {
  17. source: "data/ilsvrc12/ilsvrc12_train_lmdb"
  18. batch_size: 64
  19. backend: LMDB
  20. }
  21. top: "data"
  22. top: "label"
  23. }
  24. layers {
  25. name: "data"
  26. type: DATA
  27. include {
  28. phase: TEST
  29. }
  30. transform_param {
  31. crop_size: 224
  32. mean_value: 104
  33. mean_value: 117
  34. mean_value: 123
  35. mirror: false
  36. }
  37. data_param {
  38. source: "data/ilsvrc12/ilsvrc12_val_lmdb"
  39. batch_size: 50
  40. backend: LMDB
  41. }
  42. top: "data"
  43. top: "label"
  44. }
  45. layers {
  46. bottom: "data"
  47. top: "conv1_1"
  48. name: "conv1_1"
  49. type: CONVOLUTION
  50. convolution_param {
  51. num_output: 64
  52. pad: 1
  53. kernel_size: 3
  54. }
  55. }
  56. layers {
  57. bottom: "conv1_1"
  58. top: "conv1_1"
  59. name: "relu1_1"
  60. type: RELU
  61. }
  62. layers {
  63. bottom: "conv1_1"
  64. top: "conv1_2"
  65. name: "conv1_2"
  66. type: CONVOLUTION
  67. convolution_param {
  68. num_output: 64
  69. pad: 1
  70. kernel_size: 3
  71. }
  72. }
  73. layers {
  74. bottom: "conv1_2"
  75. top: "conv1_2"
  76. name: "relu1_2"
  77. type: RELU
  78. }
  79. layers {
  80. bottom: "conv1_2"
  81. top: "pool1"
  82. name: "pool1"
  83. type: POOLING
  84. pooling_param {
  85. pool: MAX
  86. kernel_size: 2
  87. stride: 2
  88. }
  89. }
  90. layers {
  91. bottom: "pool1"
  92. top: "conv2_1"
  93. name: "conv2_1"
  94. type: CONVOLUTION
  95. convolution_param {
  96. num_output: 128
  97. pad: 1
  98. kernel_size: 3
  99. }
  100. }
  101. layers {
  102. bottom: "conv2_1"
  103. top: "conv2_1"
  104. name: "relu2_1"
  105. type: RELU
  106. }
  107. layers {
  108. bottom: "conv2_1"
  109. top: "conv2_2"
  110. name: "conv2_2"
  111. type: CONVOLUTION
  112. convolution_param {
  113. num_output: 128
  114. pad: 1
  115. kernel_size: 3
  116. }
  117. }
  118. layers {
  119. bottom: "conv2_2"
  120. top: "conv2_2"
  121. name: "relu2_2"
  122. type: RELU
  123. }
  124. layers {
  125. bottom: "conv2_2"
  126. top: "pool2"
  127. name: "pool2"
  128. type: POOLING
  129. pooling_param {
  130. pool: MAX
  131. kernel_size: 2
  132. stride: 2
  133. }
  134. }
  135. layers {
  136. bottom: "pool2"
  137. top: "conv3_1"
  138. name: "conv3_1"
  139. type: CONVOLUTION
  140. convolution_param {
  141. num_output: 256
  142. pad: 1
  143. kernel_size: 3
  144. }
  145. }
  146. layers {
  147. bottom: "conv3_1"
  148. top: "conv3_1"
  149. name: "relu3_1"
  150. type: RELU
  151. }
  152. layers {
  153. bottom: "conv3_1"
  154. top: "conv3_2"
  155. name: "conv3_2"
  156. type: CONVOLUTION
  157. convolution_param {
  158. num_output: 256
  159. pad: 1
  160. kernel_size: 3
  161. }
  162. }
  163. layers {
  164. bottom: "conv3_2"
  165. top: "conv3_2"
  166. name: "relu3_2"
  167. type: RELU
  168. }
  169. layers {
  170. bottom: "conv3_2"
  171. top: "conv3_3"
  172. name: "conv3_3"
  173. type: CONVOLUTION
  174. convolution_param {
  175. num_output: 256
  176. pad: 1
  177. kernel_size: 3
  178. }
  179. }
  180. layers {
  181. bottom: "conv3_3"
  182. top: "conv3_3"
  183. name: "relu3_3"
  184. type: RELU
  185. }
  186. layers {
  187. bottom: "conv3_3"
  188. top: "conv3_4"
  189. name: "conv3_4"
  190. type: CONVOLUTION
  191. convolution_param {
  192. num_output: 256
  193. pad: 1
  194. kernel_size: 3
  195. }
  196. }
  197. layers {
  198. bottom: "conv3_4"
  199. top: "conv3_4"
  200. name: "relu3_4"
  201. type: RELU
  202. }
  203. layers {
  204. bottom: "conv3_4"
  205. top: "pool3"
  206. name: "pool3"
  207. type: POOLING
  208. pooling_param {
  209. pool: MAX
  210. kernel_size: 2
  211. stride: 2
  212. }
  213. }
  214. layers {
  215. bottom: "pool3"
  216. top: "conv4_1"
  217. name: "conv4_1"
  218. type: CONVOLUTION
  219. convolution_param {
  220. num_output: 512
  221. pad: 1
  222. kernel_size: 3
  223. }
  224. }
  225. layers {
  226. bottom: "conv4_1"
  227. top: "conv4_1"
  228. name: "relu4_1"
  229. type: RELU
  230. }
  231. layers {
  232. bottom: "conv4_1"
  233. top: "conv4_2"
  234. name: "conv4_2"
  235. type: CONVOLUTION
  236. convolution_param {
  237. num_output: 512
  238. pad: 1
  239. kernel_size: 3
  240. }
  241. }
  242. layers {
  243. bottom: "conv4_2"
  244. top: "conv4_2"
  245. name: "relu4_2"
  246. type: RELU
  247. }
  248. layers {
  249. bottom: "conv4_2"
  250. top: "conv4_3"
  251. name: "conv4_3"
  252. type: CONVOLUTION
  253. convolution_param {
  254. num_output: 512
  255. pad: 1
  256. kernel_size: 3
  257. }
  258. }
  259. layers {
  260. bottom: "conv4_3"
  261. top: "conv4_3"
  262. name: "relu4_3"
  263. type: RELU
  264. }
  265. layers {
  266. bottom: "conv4_3"
  267. top: "conv4_4"
  268. name: "conv4_4"
  269. type: CONVOLUTION
  270. convolution_param {
  271. num_output: 512
  272. pad: 1
  273. kernel_size: 3
  274. }
  275. }
  276. layers {
  277. bottom: "conv4_4"
  278. top: "conv4_4"
  279. name: "relu4_4"
  280. type: RELU
  281. }
  282. layers {
  283. bottom: "conv4_4"
  284. top: "pool4"
  285. name: "pool4"
  286. type: POOLING
  287. pooling_param {
  288. pool: MAX
  289. kernel_size: 2
  290. stride: 2
  291. }
  292. }
  293. layers {
  294. bottom: "pool4"
  295. top: "conv5_1"
  296. name: "conv5_1"
  297. type: CONVOLUTION
  298. convolution_param {
  299. num_output: 512
  300. pad: 1
  301. kernel_size: 3
  302. }
  303. }
  304. layers {
  305. bottom: "conv5_1"
  306. top: "conv5_1"
  307. name: "relu5_1"
  308. type: RELU
  309. }
  310. layers {
  311. bottom: "conv5_1"
  312. top: "conv5_2"
  313. name: "conv5_2"
  314. type: CONVOLUTION
  315. convolution_param {
  316. num_output: 512
  317. pad: 1
  318. kernel_size: 3
  319. }
  320. }
  321. layers {
  322. bottom: "conv5_2"
  323. top: "conv5_2"
  324. name: "relu5_2"
  325. type: RELU
  326. }
  327. layers {
  328. bottom: "conv5_2"
  329. top: "conv5_3"
  330. name: "conv5_3"
  331. type: CONVOLUTION
  332. convolution_param {
  333. num_output: 512
  334. pad: 1
  335. kernel_size: 3
  336. }
  337. }
  338. layers {
  339. bottom: "conv5_3"
  340. top: "conv5_3"
  341. name: "relu5_3"
  342. type: RELU
  343. }
  344. layers {
  345. bottom: "conv5_3"
  346. top: "conv5_4"
  347. name: "conv5_4"
  348. type: CONVOLUTION
  349. convolution_param {
  350. num_output: 512
  351. pad: 1
  352. kernel_size: 3
  353. }
  354. }
  355. layers {
  356. bottom: "conv5_4"
  357. top: "conv5_4"
  358. name: "relu5_4"
  359. type: RELU
  360. }
  361. layers {
  362. bottom: "conv5_4"
  363. top: "pool5"
  364. name: "pool5"
  365. type: POOLING
  366. pooling_param {
  367. pool: MAX
  368. kernel_size: 2
  369. stride: 2
  370. }
  371. }
  372. layers {
  373. bottom: "pool5"
  374. top: "fc6"
  375. name: "fc6"
  376. type: INNER_PRODUCT
  377. inner_product_param {
  378. num_output: 4096
  379. }
  380. }
  381. layers {
  382. bottom: "fc6"
  383. top: "fc6"
  384. name: "relu6"
  385. type: RELU
  386. }
  387. layers {
  388. bottom: "fc6"
  389. top: "fc6"
  390. name: "drop6"
  391. type: DROPOUT
  392. dropout_param {
  393. dropout_ratio: 0.5
  394. }
  395. }
  396. layers {
  397. bottom: "fc6"
  398. top: "fc7"
  399. name: "fc7"
  400. type: INNER_PRODUCT
  401. inner_product_param {
  402. num_output: 4096
  403. }
  404. }
  405. layers {
  406. bottom: "fc7"
  407. top: "fc7"
  408. name: "relu7"
  409. type: RELU
  410. }
  411. layers {
  412. bottom: "fc7"
  413. top: "fc7"
  414. name: "drop7"
  415. type: DROPOUT
  416. dropout_param {
  417. dropout_ratio: 0.5
  418. }
  419. }
  420. layers {
  421. name: "fc8"
  422. bottom: "fc7"
  423. top: "fc8"
  424. type: INNER_PRODUCT
  425. inner_product_param {
  426. num_output: 1000
  427. }
  428. }
  429. layers {
  430. name: "loss"
  431. type: SOFTMAX_LOSS
  432. bottom: "fc8"
  433. bottom: "label"
  434. top: "loss/loss"
  435. }
  436. layers {
  437. name: "accuracy/top1"
  438. type: ACCURACY
  439. bottom: "fc8"
  440. bottom: "label"
  441. top: "accuracy@1"
  442. include: { phase: TEST }
  443. accuracy_param {
  444. top_k: 1
  445. }
  446. }
  447. layers {
  448. name: "accuracy/top5"
  449. type: ACCURACY
  450. bottom: "fc8"
  451. bottom: "label"
  452. top: "accuracy@5"
  453. include: { phase: TEST }
  454. accuracy_param {
  455. top_k: 5
  456. }
  457. }
Add Comment
Please, Sign In to add comment