Advertisement
ashwinnair14

Untitled

Jan 13th, 2016
309
0
Never
Not a member of Pastebin yet? Sign Up, it unlocks many cool features!
text 39.27 KB | None | 0 0
  1. I0113 14:20:50.793929 85242 caffe.cpp:184] Using GPUs 0, 1, 2, 3, 4, 5, 6, 7
  2. I0113 14:20:51.195428 85242 solver.cpp:48] Initializing solver from parameters:
  3. test_iter: 50
  4. test_interval: 500
  5. base_lr: 0.01
  6. display: 500
  7. max_iter: 850000
  8. lr_policy: "step"
  9. gamma: 0.5
  10. momentum: 0.9
  11. weight_decay: 0.0005
  12. snapshot: 5000
  13. snapshot_prefix: "models/mv16f/mv_16_"
  14. solver_mode: GPU
  15. device_id: 0
  16. net: "models/mv16f/mv_train.prototxt"
  17. I0113 14:20:51.195653 85242 solver.cpp:91] Creating training net from net file: models/mv16f/mv_train.prototxt
  18. I0113 14:20:51.196537 85242 net.cpp:322] The NetState phase (0) differed from the phase (1) specified by a rule in layer data
  19. I0113 14:20:51.196575 85242 net.cpp:322] The NetState phase (0) differed from the phase (1) specified by a rule in layer accuracy
  20. I0113 14:20:51.196720 85242 net.cpp:49] Initializing net from parameters:
  21. name: "mv_16f"
  22. state {
  23. phase: TRAIN
  24. }
  25. layer {
  26. name: "data"
  27. type: "HDF5Data"
  28. top: "data"
  29. top: "label"
  30. include {
  31. phase: TRAIN
  32. }
  33. hdf5_data_param {
  34. source: "/home/fe/anilil/caffe/models/mv16f/train.txt"
  35. batch_size: 1000
  36. }
  37. }
  38. layer {
  39. name: "conv1"
  40. type: "Convolution"
  41. bottom: "data"
  42. top: "conv1"
  43. param {
  44. lr_mult: 1
  45. decay_mult: 1
  46. }
  47. param {
  48. lr_mult: 1
  49. decay_mult: 1
  50. }
  51. convolution_param {
  52. num_output: 64
  53. kernel_size: 3
  54. weight_filler {
  55. type: "xavier"
  56. }
  57. bias_filler {
  58. type: "constant"
  59. value: 0
  60. }
  61. }
  62. }
  63. layer {
  64. name: "relu1"
  65. type: "ReLU"
  66. bottom: "conv1"
  67. top: "conv1"
  68. }
  69. layer {
  70. name: "pool1"
  71. type: "Pooling"
  72. bottom: "conv1"
  73. top: "pool1"
  74. pooling_param {
  75. pool: MAX
  76. kernel_size: 2
  77. stride: 2
  78. }
  79. }
  80. layer {
  81. name: "conv2"
  82. type: "Convolution"
  83. bottom: "pool1"
  84. top: "conv2"
  85. param {
  86. lr_mult: 1
  87. decay_mult: 1
  88. }
  89. param {
  90. lr_mult: 1
  91. decay_mult: 1
  92. }
  93. convolution_param {
  94. num_output: 128
  95. kernel_size: 3
  96. stride: 1
  97. weight_filler {
  98. type: "xavier"
  99. }
  100. bias_filler {
  101. type: "constant"
  102. value: 0
  103. }
  104. }
  105. }
  106. layer {
  107. name: "relu2"
  108. type: "ReLU"
  109. bottom: "conv2"
  110. top: "conv2"
  111. }
  112. layer {
  113. name: "pool2"
  114. type: "Pooling"
  115. bottom: "conv2"
  116. top: "pool2"
  117. pooling_param {
  118. pool: MAX
  119. kernel_size: 2
  120. stride: 2
  121. }
  122. }
  123. layer {
  124. name: "conv3"
  125. type: "Convolution"
  126. bottom: "pool2"
  127. top: "conv3"
  128. param {
  129. lr_mult: 1
  130. decay_mult: 1
  131. }
  132. param {
  133. lr_mult: 1
  134. decay_mult: 1
  135. }
  136. convolution_param {
  137. num_output: 256
  138. kernel_size: 3
  139. stride: 1
  140. weight_filler {
  141. type: "xavier"
  142. }
  143. bias_filler {
  144. type: "constant"
  145. value: 0
  146. }
  147. }
  148. }
  149. layer {
  150. name: "relu3"
  151. type: "ReLU"
  152. bottom: "conv3"
  153. top: "conv3"
  154. }
  155. layer {
  156. name: "pool3"
  157. type: "Pooling"
  158. bottom: "conv3"
  159. top: "pool3"
  160. pooling_param {
  161. pool: MAX
  162. kernel_size: 2
  163. stride: 2
  164. }
  165. }
  166. layer {
  167. name: "conv4"
  168. type: "Convolution"
  169. bottom: "pool3"
  170. top: "conv4"
  171. param {
  172. lr_mult: 1
  173. decay_mult: 1
  174. }
  175. param {
  176. lr_mult: 1
  177. decay_mult: 1
  178. }
  179. convolution_param {
  180. num_output: 256
  181. kernel_size: 3
  182. stride: 1
  183. weight_filler {
  184. type: "xavier"
  185. }
  186. bias_filler {
  187. type: "constant"
  188. value: 0
  189. }
  190. }
  191. }
  192. layer {
  193. name: "relu4"
  194. type: "ReLU"
  195. bottom: "conv4"
  196. top: "conv4"
  197. }
  198. layer {
  199. name: "pool4"
  200. type: "Pooling"
  201. bottom: "conv4"
  202. top: "pool4"
  203. pooling_param {
  204. pool: MAX
  205. kernel_size: 2
  206. stride: 2
  207. }
  208. }
  209. layer {
  210. name: "fc6"
  211. type: "InnerProduct"
  212. bottom: "pool4"
  213. top: "fc6"
  214. param {
  215. lr_mult: 1
  216. decay_mult: 1
  217. }
  218. param {
  219. lr_mult: 2
  220. decay_mult: 0
  221. }
  222. inner_product_param {
  223. num_output: 2048
  224. weight_filler {
  225. type: "gaussian"
  226. std: 0.005
  227. }
  228. bias_filler {
  229. type: "constant"
  230. value: 0.1
  231. }
  232. }
  233. }
  234. layer {
  235. name: "relu6"
  236. type: "ReLU"
  237. bottom: "fc6"
  238. top: "fc6"
  239. }
  240. layer {
  241. name: "drop6"
  242. type: "Dropout"
  243. bottom: "fc6"
  244. top: "fc6"
  245. dropout_param {
  246. dropout_ratio: 0.5
  247. }
  248. }
  249. layer {
  250. name: "fc7"
  251. type: "InnerProduct"
  252. bottom: "fc6"
  253. top: "fc7"
  254. param {
  255. lr_mult: 1
  256. decay_mult: 1
  257. }
  258. param {
  259. lr_mult: 2
  260. decay_mult: 0
  261. }
  262. inner_product_param {
  263. num_output: 4096
  264. weight_filler {
  265. type: "gaussian"
  266. std: 0.005
  267. }
  268. bias_filler {
  269. type: "constant"
  270. value: 0.1
  271. }
  272. }
  273. }
  274. layer {
  275. name: "relu7"
  276. type: "ReLU"
  277. bottom: "fc7"
  278. top: "fc7"
  279. }
  280. layer {
  281. name: "drop7"
  282. type: "Dropout"
  283. bottom: "fc7"
  284. top: "fc7"
  285. dropout_param {
  286. dropout_ratio: 0.5
  287. }
  288. }
  289. layer {
  290. name: "fc8"
  291. type: "InnerProduct"
  292. bottom: "fc7"
  293. top: "fc8"
  294. param {
  295. lr_mult: 1
  296. decay_mult: 1
  297. }
  298. param {
  299. lr_mult: 2
  300. decay_mult: 0
  301. }
  302. inner_product_param {
  303. num_output: 101
  304. weight_filler {
  305. type: "xavier"
  306. }
  307. bias_filler {
  308. type: "constant"
  309. value: 0
  310. }
  311. }
  312. }
  313. layer {
  314. name: "loss"
  315. type: "SoftmaxWithLoss"
  316. bottom: "fc8"
  317. bottom: "label"
  318. top: "loss"
  319. }
  320. I0113 14:20:51.197294 85242 layer_factory.hpp:77] Creating layer data
  321. I0113 14:20:51.197355 85242 net.cpp:106] Creating Layer data
  322. I0113 14:20:51.197376 85242 net.cpp:411] data -> data
  323. I0113 14:20:51.197504 85242 net.cpp:411] data -> label
  324. I0113 14:20:51.198678 85242 hdf5.cpp:32] Datatype class: H5T_FLOAT
  325. I0113 14:20:52.063480 85242 net.cpp:150] Setting up data
  326. I0113 14:20:52.063524 85242 net.cpp:157] Top shape: 1000 48 58 58 (161472000)
  327. I0113 14:20:52.063534 85242 net.cpp:157] Top shape: 1000 1 (1000)
  328. I0113 14:20:52.063539 85242 net.cpp:165] Memory required for data: 645892000
  329. I0113 14:20:52.063563 85242 layer_factory.hpp:77] Creating layer conv1
  330. I0113 14:20:52.063654 85242 net.cpp:106] Creating Layer conv1
  331. I0113 14:20:52.063675 85242 net.cpp:454] conv1 <- data
  332. I0113 14:20:52.063720 85242 net.cpp:411] conv1 -> conv1
  333. I0113 14:20:52.192370 85242 net.cpp:150] Setting up conv1
  334. I0113 14:20:52.192412 85242 net.cpp:157] Top shape: 1000 64 56 56 (200704000)
  335. I0113 14:20:52.192416 85242 net.cpp:165] Memory required for data: 1448708000
  336. I0113 14:20:52.192497 85242 layer_factory.hpp:77] Creating layer relu1
  337. I0113 14:20:52.192543 85242 net.cpp:106] Creating Layer relu1
  338. I0113 14:20:52.192554 85242 net.cpp:454] relu1 <- conv1
  339. I0113 14:20:52.192579 85242 net.cpp:397] relu1 -> conv1 (in-place)
  340. I0113 14:20:52.192730 85242 net.cpp:150] Setting up relu1
  341. I0113 14:20:52.192740 85242 net.cpp:157] Top shape: 1000 64 56 56 (200704000)
  342. I0113 14:20:52.192744 85242 net.cpp:165] Memory required for data: 2251524000
  343. I0113 14:20:52.192750 85242 layer_factory.hpp:77] Creating layer pool1
  344. I0113 14:20:52.192781 85242 net.cpp:106] Creating Layer pool1
  345. I0113 14:20:52.192790 85242 net.cpp:454] pool1 <- conv1
  346. I0113 14:20:52.192808 85242 net.cpp:411] pool1 -> pool1
  347. I0113 14:20:52.193133 85242 net.cpp:150] Setting up pool1
  348. I0113 14:20:52.193152 85242 net.cpp:157] Top shape: 1000 64 28 28 (50176000)
  349. I0113 14:20:52.193157 85242 net.cpp:165] Memory required for data: 2452228000
  350. I0113 14:20:52.193164 85242 layer_factory.hpp:77] Creating layer conv2
  351. I0113 14:20:52.193200 85242 net.cpp:106] Creating Layer conv2
  352. I0113 14:20:52.193209 85242 net.cpp:454] conv2 <- pool1
  353. I0113 14:20:52.193230 85242 net.cpp:411] conv2 -> conv2
  354. I0113 14:20:52.199265 85242 net.cpp:150] Setting up conv2
  355. I0113 14:20:52.199278 85242 net.cpp:157] Top shape: 1000 128 26 26 (86528000)
  356. I0113 14:20:52.199283 85242 net.cpp:165] Memory required for data: 2798340000
  357. I0113 14:20:52.199306 85242 layer_factory.hpp:77] Creating layer relu2
  358. I0113 14:20:52.199322 85242 net.cpp:106] Creating Layer relu2
  359. I0113 14:20:52.199331 85242 net.cpp:454] relu2 <- conv2
  360. I0113 14:20:52.199345 85242 net.cpp:397] relu2 -> conv2 (in-place)
  361. I0113 14:20:52.199492 85242 net.cpp:150] Setting up relu2
  362. I0113 14:20:52.199502 85242 net.cpp:157] Top shape: 1000 128 26 26 (86528000)
  363. I0113 14:20:52.199507 85242 net.cpp:165] Memory required for data: 3144452000
  364. I0113 14:20:52.199560 85242 layer_factory.hpp:77] Creating layer pool2
  365. I0113 14:20:52.199578 85242 net.cpp:106] Creating Layer pool2
  366. I0113 14:20:52.199585 85242 net.cpp:454] pool2 <- conv2
  367. I0113 14:20:52.199605 85242 net.cpp:411] pool2 -> pool2
  368. I0113 14:20:52.199965 85242 net.cpp:150] Setting up pool2
  369. I0113 14:20:52.199980 85242 net.cpp:157] Top shape: 1000 128 13 13 (21632000)
  370. I0113 14:20:52.199985 85242 net.cpp:165] Memory required for data: 3230980000
  371. I0113 14:20:52.199990 85242 layer_factory.hpp:77] Creating layer conv3
  372. I0113 14:20:52.200013 85242 net.cpp:106] Creating Layer conv3
  373. I0113 14:20:52.200022 85242 net.cpp:454] conv3 <- pool2
  374. I0113 14:20:52.200042 85242 net.cpp:411] conv3 -> conv3
  375. I0113 14:20:52.216902 85242 net.cpp:150] Setting up conv3
  376. I0113 14:20:52.216912 85242 net.cpp:157] Top shape: 1000 256 11 11 (30976000)
  377. I0113 14:20:52.216917 85242 net.cpp:165] Memory required for data: 3354884000
  378. I0113 14:20:52.216938 85242 layer_factory.hpp:77] Creating layer relu3
  379. I0113 14:20:52.216958 85242 net.cpp:106] Creating Layer relu3
  380. I0113 14:20:52.216967 85242 net.cpp:454] relu3 <- conv3
  381. I0113 14:20:52.216982 85242 net.cpp:397] relu3 -> conv3 (in-place)
  382. I0113 14:20:52.217123 85242 net.cpp:150] Setting up relu3
  383. I0113 14:20:52.217133 85242 net.cpp:157] Top shape: 1000 256 11 11 (30976000)
  384. I0113 14:20:52.217139 85242 net.cpp:165] Memory required for data: 3478788000
  385. I0113 14:20:52.217145 85242 layer_factory.hpp:77] Creating layer pool3
  386. I0113 14:20:52.217160 85242 net.cpp:106] Creating Layer pool3
  387. I0113 14:20:52.217169 85242 net.cpp:454] pool3 <- conv3
  388. I0113 14:20:52.217185 85242 net.cpp:411] pool3 -> pool3
  389. I0113 14:20:52.217483 85242 net.cpp:150] Setting up pool3
  390. I0113 14:20:52.217496 85242 net.cpp:157] Top shape: 1000 256 6 6 (9216000)
  391. I0113 14:20:52.217501 85242 net.cpp:165] Memory required for data: 3515652000
  392. I0113 14:20:52.217507 85242 layer_factory.hpp:77] Creating layer conv4
  393. I0113 14:20:52.217530 85242 net.cpp:106] Creating Layer conv4
  394. I0113 14:20:52.217538 85242 net.cpp:454] conv4 <- pool3
  395. I0113 14:20:52.217558 85242 net.cpp:411] conv4 -> conv4
  396. I0113 14:20:52.249881 85242 net.cpp:150] Setting up conv4
  397. I0113 14:20:52.249891 85242 net.cpp:157] Top shape: 1000 256 4 4 (4096000)
  398. I0113 14:20:52.249896 85242 net.cpp:165] Memory required for data: 3532036000
  399. I0113 14:20:52.249908 85242 layer_factory.hpp:77] Creating layer relu4
  400. I0113 14:20:52.249922 85242 net.cpp:106] Creating Layer relu4
  401. I0113 14:20:52.249929 85242 net.cpp:454] relu4 <- conv4
  402. I0113 14:20:52.249944 85242 net.cpp:397] relu4 -> conv4 (in-place)
  403. I0113 14:20:52.250211 85242 net.cpp:150] Setting up relu4
  404. I0113 14:20:52.250222 85242 net.cpp:157] Top shape: 1000 256 4 4 (4096000)
  405. I0113 14:20:52.250227 85242 net.cpp:165] Memory required for data: 3548420000
  406. I0113 14:20:52.250233 85242 layer_factory.hpp:77] Creating layer pool4
  407. I0113 14:20:52.250252 85242 net.cpp:106] Creating Layer pool4
  408. I0113 14:20:52.250259 85242 net.cpp:454] pool4 <- conv4
  409. I0113 14:20:52.250277 85242 net.cpp:411] pool4 -> pool4
  410. I0113 14:20:52.250453 85242 net.cpp:150] Setting up pool4
  411. I0113 14:20:52.250466 85242 net.cpp:157] Top shape: 1000 256 2 2 (1024000)
  412. I0113 14:20:52.250471 85242 net.cpp:165] Memory required for data: 3552516000
  413. I0113 14:20:52.250478 85242 layer_factory.hpp:77] Creating layer fc6
  414. I0113 14:20:52.250510 85242 net.cpp:106] Creating Layer fc6
  415. I0113 14:20:52.250520 85242 net.cpp:454] fc6 <- pool4
  416. I0113 14:20:52.250538 85242 net.cpp:411] fc6 -> fc6
  417. I0113 14:20:52.431202 85242 net.cpp:150] Setting up fc6
  418. I0113 14:20:52.431219 85242 net.cpp:157] Top shape: 1000 2048 (2048000)
  419. I0113 14:20:52.431224 85242 net.cpp:165] Memory required for data: 3560708000
  420. I0113 14:20:52.431248 85242 layer_factory.hpp:77] Creating layer relu6
  421. I0113 14:20:52.431263 85242 net.cpp:106] Creating Layer relu6
  422. I0113 14:20:52.431272 85242 net.cpp:454] relu6 <- fc6
  423. I0113 14:20:52.431298 85242 net.cpp:397] relu6 -> fc6 (in-place)
  424. I0113 14:20:52.431664 85242 net.cpp:150] Setting up relu6
  425. I0113 14:20:52.431677 85242 net.cpp:157] Top shape: 1000 2048 (2048000)
  426. I0113 14:20:52.431682 85242 net.cpp:165] Memory required for data: 3568900000
  427. I0113 14:20:52.431689 85242 layer_factory.hpp:77] Creating layer drop6
  428. I0113 14:20:52.431717 85242 net.cpp:106] Creating Layer drop6
  429. I0113 14:20:52.431726 85242 net.cpp:454] drop6 <- fc6
  430. I0113 14:20:52.431743 85242 net.cpp:397] drop6 -> fc6 (in-place)
  431. I0113 14:20:52.431787 85242 net.cpp:150] Setting up drop6
  432. I0113 14:20:52.431795 85242 net.cpp:157] Top shape: 1000 2048 (2048000)
  433. I0113 14:20:52.431800 85242 net.cpp:165] Memory required for data: 3577092000
  434. I0113 14:20:52.431805 85242 layer_factory.hpp:77] Creating layer fc7
  435. I0113 14:20:52.431826 85242 net.cpp:106] Creating Layer fc7
  436. I0113 14:20:52.431834 85242 net.cpp:454] fc7 <- fc6
  437. I0113 14:20:52.431852 85242 net.cpp:411] fc7 -> fc7
  438. I0113 14:20:53.158080 85242 net.cpp:150] Setting up fc7
  439. I0113 14:20:53.158118 85242 net.cpp:157] Top shape: 1000 4096 (4096000)
  440. I0113 14:20:53.158121 85242 net.cpp:165] Memory required for data: 3593476000
  441. I0113 14:20:53.158146 85242 layer_factory.hpp:77] Creating layer relu7
  442. I0113 14:20:53.158180 85242 net.cpp:106] Creating Layer relu7
  443. I0113 14:20:53.158195 85242 net.cpp:454] relu7 <- fc7
  444. I0113 14:20:53.158221 85242 net.cpp:397] relu7 -> fc7 (in-place)
  445. I0113 14:20:53.158637 85242 net.cpp:150] Setting up relu7
  446. I0113 14:20:53.158648 85242 net.cpp:157] Top shape: 1000 4096 (4096000)
  447. I0113 14:20:53.158653 85242 net.cpp:165] Memory required for data: 3609860000
  448. I0113 14:20:53.158658 85242 layer_factory.hpp:77] Creating layer drop7
  449. I0113 14:20:53.158679 85242 net.cpp:106] Creating Layer drop7
  450. I0113 14:20:53.158690 85242 net.cpp:454] drop7 <- fc7
  451. I0113 14:20:53.158706 85242 net.cpp:397] drop7 -> fc7 (in-place)
  452. I0113 14:20:53.158751 85242 net.cpp:150] Setting up drop7
  453. I0113 14:20:53.158761 85242 net.cpp:157] Top shape: 1000 4096 (4096000)
  454. I0113 14:20:53.158763 85242 net.cpp:165] Memory required for data: 3626244000
  455. I0113 14:20:53.158768 85242 layer_factory.hpp:77] Creating layer fc8
  456. I0113 14:20:53.158788 85242 net.cpp:106] Creating Layer fc8
  457. I0113 14:20:53.158795 85242 net.cpp:454] fc8 <- fc7
  458. I0113 14:20:53.158817 85242 net.cpp:411] fc8 -> fc8
  459. I0113 14:20:53.181195 85242 net.cpp:150] Setting up fc8
  460. I0113 14:20:53.181215 85242 net.cpp:157] Top shape: 1000 101 (101000)
  461. I0113 14:20:53.181221 85242 net.cpp:165] Memory required for data: 3626648000
  462. I0113 14:20:53.181233 85242 layer_factory.hpp:77] Creating layer loss
  463. I0113 14:20:53.181262 85242 net.cpp:106] Creating Layer loss
  464. I0113 14:20:53.181282 85242 net.cpp:454] loss <- fc8
  465. I0113 14:20:53.181295 85242 net.cpp:454] loss <- label
  466. I0113 14:20:53.181313 85242 net.cpp:411] loss -> loss
  467. I0113 14:20:53.181342 85242 layer_factory.hpp:77] Creating layer loss
  468. I0113 14:20:53.181895 85242 net.cpp:150] Setting up loss
  469. I0113 14:20:53.181910 85242 net.cpp:157] Top shape: (1)
  470. I0113 14:20:53.181913 85242 net.cpp:160] with loss weight 1
  471. I0113 14:20:53.181944 85242 net.cpp:165] Memory required for data: 3626648004
  472. I0113 14:20:53.181956 85242 net.cpp:226] loss needs backward computation.
  473. I0113 14:20:53.181964 85242 net.cpp:226] fc8 needs backward computation.
  474. I0113 14:20:53.181969 85242 net.cpp:226] drop7 needs backward computation.
  475. I0113 14:20:53.181975 85242 net.cpp:226] relu7 needs backward computation.
  476. I0113 14:20:53.181980 85242 net.cpp:226] fc7 needs backward computation.
  477. I0113 14:20:53.181987 85242 net.cpp:226] drop6 needs backward computation.
  478. I0113 14:20:53.181991 85242 net.cpp:226] relu6 needs backward computation.
  479. I0113 14:20:53.181998 85242 net.cpp:226] fc6 needs backward computation.
  480. I0113 14:20:53.182003 85242 net.cpp:226] pool4 needs backward computation.
  481. I0113 14:20:53.182013 85242 net.cpp:226] relu4 needs backward computation.
  482. I0113 14:20:53.182018 85242 net.cpp:226] conv4 needs backward computation.
  483. I0113 14:20:53.182025 85242 net.cpp:226] pool3 needs backward computation.
  484. I0113 14:20:53.182032 85242 net.cpp:226] relu3 needs backward computation.
  485. I0113 14:20:53.182076 85242 net.cpp:226] conv3 needs backward computation.
  486. I0113 14:20:53.182085 85242 net.cpp:226] pool2 needs backward computation.
  487. I0113 14:20:53.182091 85242 net.cpp:226] relu2 needs backward computation.
  488. I0113 14:20:53.182097 85242 net.cpp:226] conv2 needs backward computation.
  489. I0113 14:20:53.182102 85242 net.cpp:226] pool1 needs backward computation.
  490. I0113 14:20:53.182111 85242 net.cpp:226] relu1 needs backward computation.
  491. I0113 14:20:53.182128 85242 net.cpp:226] conv1 needs backward computation.
  492. I0113 14:20:53.182135 85242 net.cpp:228] data does not need backward computation.
  493. I0113 14:20:53.182143 85242 net.cpp:270] This network produces output loss
  494. I0113 14:20:53.182194 85242 net.cpp:283] Network initialization done.
  495. I0113 14:20:53.183140 85242 solver.cpp:181] Creating test net (#0) specified by net file: models/mv16f/mv_train.prototxt
  496. I0113 14:20:53.183226 85242 net.cpp:322] The NetState phase (1) differed from the phase (0) specified by a rule in layer data
  497. I0113 14:20:53.183384 85242 net.cpp:49] Initializing net from parameters:
  498. name: "mv_16f"
  499. state {
  500. phase: TEST
  501. }
  502. layer {
  503. name: "data"
  504. type: "HDF5Data"
  505. top: "data"
  506. top: "label"
  507. include {
  508. phase: TEST
  509. }
  510. hdf5_data_param {
  511. source: "/home/fe/anilil/caffe/models/mv16f/test.txt"
  512. batch_size: 500
  513. }
  514. }
  515. layer {
  516. name: "conv1"
  517. type: "Convolution"
  518. bottom: "data"
  519. top: "conv1"
  520. param {
  521. lr_mult: 1
  522. decay_mult: 1
  523. }
  524. param {
  525. lr_mult: 1
  526. decay_mult: 1
  527. }
  528. convolution_param {
  529. num_output: 64
  530. kernel_size: 3
  531. weight_filler {
  532. type: "xavier"
  533. }
  534. bias_filler {
  535. type: "constant"
  536. value: 0
  537. }
  538. }
  539. }
  540. layer {
  541. name: "relu1"
  542. type: "ReLU"
  543. bottom: "conv1"
  544. top: "conv1"
  545. }
  546. layer {
  547. name: "pool1"
  548. type: "Pooling"
  549. bottom: "conv1"
  550. top: "pool1"
  551. pooling_param {
  552. pool: MAX
  553. kernel_size: 2
  554. stride: 2
  555. }
  556. }
  557. layer {
  558. name: "conv2"
  559. type: "Convolution"
  560. bottom: "pool1"
  561. top: "conv2"
  562. param {
  563. lr_mult: 1
  564. decay_mult: 1
  565. }
  566. param {
  567. lr_mult: 1
  568. decay_mult: 1
  569. }
  570. convolution_param {
  571. num_output: 128
  572. kernel_size: 3
  573. stride: 1
  574. weight_filler {
  575. type: "xavier"
  576. }
  577. bias_filler {
  578. type: "constant"
  579. value: 0
  580. }
  581. }
  582. }
  583. layer {
  584. name: "relu2"
  585. type: "ReLU"
  586. bottom: "conv2"
  587. top: "conv2"
  588. }
  589. layer {
  590. name: "pool2"
  591. type: "Pooling"
  592. bottom: "conv2"
  593. top: "pool2"
  594. pooling_param {
  595. pool: MAX
  596. kernel_size: 2
  597. stride: 2
  598. }
  599. }
  600. layer {
  601. name: "conv3"
  602. type: "Convolution"
  603. bottom: "pool2"
  604. top: "conv3"
  605. param {
  606. lr_mult: 1
  607. decay_mult: 1
  608. }
  609. param {
  610. lr_mult: 1
  611. decay_mult: 1
  612. }
  613. convolution_param {
  614. num_output: 256
  615. kernel_size: 3
  616. stride: 1
  617. weight_filler {
  618. type: "xavier"
  619. }
  620. bias_filler {
  621. type: "constant"
  622. value: 0
  623. }
  624. }
  625. }
  626. layer {
  627. name: "relu3"
  628. type: "ReLU"
  629. bottom: "conv3"
  630. top: "conv3"
  631. }
  632. layer {
  633. name: "pool3"
  634. type: "Pooling"
  635. bottom: "conv3"
  636. top: "pool3"
  637. pooling_param {
  638. pool: MAX
  639. kernel_size: 2
  640. stride: 2
  641. }
  642. }
  643. layer {
  644. name: "conv4"
  645. type: "Convolution"
  646. bottom: "pool3"
  647. top: "conv4"
  648. param {
  649. lr_mult: 1
  650. decay_mult: 1
  651. }
  652. param {
  653. lr_mult: 1
  654. decay_mult: 1
  655. }
  656. convolution_param {
  657. num_output: 256
  658. kernel_size: 3
  659. stride: 1
  660. weight_filler {
  661. type: "xavier"
  662. }
  663. bias_filler {
  664. type: "constant"
  665. value: 0
  666. }
  667. }
  668. }
  669. layer {
  670. name: "relu4"
  671. type: "ReLU"
  672. bottom: "conv4"
  673. top: "conv4"
  674. }
  675. layer {
  676. name: "pool4"
  677. type: "Pooling"
  678. bottom: "conv4"
  679. top: "pool4"
  680. pooling_param {
  681. pool: MAX
  682. kernel_size: 2
  683. stride: 2
  684. }
  685. }
  686. layer {
  687. name: "fc6"
  688. type: "InnerProduct"
  689. bottom: "pool4"
  690. top: "fc6"
  691. param {
  692. lr_mult: 1
  693. decay_mult: 1
  694. }
  695. param {
  696. lr_mult: 2
  697. decay_mult: 0
  698. }
  699. inner_product_param {
  700. num_output: 2048
  701. weight_filler {
  702. type: "gaussian"
  703. std: 0.005
  704. }
  705. bias_filler {
  706. type: "constant"
  707. value: 0.1
  708. }
  709. }
  710. }
  711. layer {
  712. name: "relu6"
  713. type: "ReLU"
  714. bottom: "fc6"
  715. top: "fc6"
  716. }
  717. layer {
  718. name: "drop6"
  719. type: "Dropout"
  720. bottom: "fc6"
  721. top: "fc6"
  722. dropout_param {
  723. dropout_ratio: 0.5
  724. }
  725. }
  726. layer {
  727. name: "fc7"
  728. type: "InnerProduct"
  729. bottom: "fc6"
  730. top: "fc7"
  731. param {
  732. lr_mult: 1
  733. decay_mult: 1
  734. }
  735. param {
  736. lr_mult: 2
  737. decay_mult: 0
  738. }
  739. inner_product_param {
  740. num_output: 4096
  741. weight_filler {
  742. type: "gaussian"
  743. std: 0.005
  744. }
  745. bias_filler {
  746. type: "constant"
  747. value: 0.1
  748. }
  749. }
  750. }
  751. layer {
  752. name: "relu7"
  753. type: "ReLU"
  754. bottom: "fc7"
  755. top: "fc7"
  756. }
  757. layer {
  758. name: "drop7"
  759. type: "Dropout"
  760. bottom: "fc7"
  761. top: "fc7"
  762. dropout_param {
  763. dropout_ratio: 0.5
  764. }
  765. }
  766. layer {
  767. name: "fc8"
  768. type: "InnerProduct"
  769. bottom: "fc7"
  770. top: "fc8"
  771. param {
  772. lr_mult: 1
  773. decay_mult: 1
  774. }
  775. param {
  776. lr_mult: 2
  777. decay_mult: 0
  778. }
  779. inner_product_param {
  780. num_output: 101
  781. weight_filler {
  782. type: "xavier"
  783. }
  784. bias_filler {
  785. type: "constant"
  786. value: 0
  787. }
  788. }
  789. }
  790. layer {
  791. name: "accuracy"
  792. type: "Accuracy"
  793. bottom: "fc8"
  794. bottom: "label"
  795. top: "accuracy"
  796. include {
  797. phase: TEST
  798. }
  799. }
  800. layer {
  801. name: "loss"
  802. type: "SoftmaxWithLoss"
  803. bottom: "fc8"
  804. bottom: "label"
  805. top: "loss"
  806. }
  807. I0113 14:20:53.183778 85242 layer_factory.hpp:77] Creating layer data
  808. I0113 14:20:53.183812 85242 net.cpp:106] Creating Layer data
  809. I0113 14:20:53.183820 85242 net.cpp:411] data -> data
  810. I0113 14:20:53.183842 85242 net.cpp:411] data -> label
  811. I0113 14:20:54.018275 85242 net.cpp:150] Setting up data
  812. I0113 14:20:54.018297 85242 net.cpp:157] Top shape: 500 48 58 58 (80736000)
  813. I0113 14:20:54.018306 85242 net.cpp:157] Top shape: 500 1 (500)
  814. I0113 14:20:54.018309 85242 net.cpp:165] Memory required for data: 322946000
  815. I0113 14:20:54.018332 85242 layer_factory.hpp:77] Creating layer label_data_1_split
  816. I0113 14:20:54.018378 85242 net.cpp:106] Creating Layer label_data_1_split
  817. I0113 14:20:54.018390 85242 net.cpp:454] label_data_1_split <- label
  818. I0113 14:20:54.018414 85242 net.cpp:411] label_data_1_split -> label_data_1_split_0
  819. I0113 14:20:54.018440 85242 net.cpp:411] label_data_1_split -> label_data_1_split_1
  820. I0113 14:20:54.018491 85242 net.cpp:150] Setting up label_data_1_split
  821. I0113 14:20:54.018501 85242 net.cpp:157] Top shape: 500 1 (500)
  822. I0113 14:20:54.018507 85242 net.cpp:157] Top shape: 500 1 (500)
  823. I0113 14:20:54.018509 85242 net.cpp:165] Memory required for data: 322950000
  824. I0113 14:20:54.018514 85242 layer_factory.hpp:77] Creating layer conv1
  825. I0113 14:20:54.018548 85242 net.cpp:106] Creating Layer conv1
  826. I0113 14:20:54.018556 85242 net.cpp:454] conv1 <- data
  827. I0113 14:20:54.018574 85242 net.cpp:411] conv1 -> conv1
  828. I0113 14:20:54.021152 85242 net.cpp:150] Setting up conv1
  829. I0113 14:20:54.021168 85242 net.cpp:157] Top shape: 500 64 56 56 (100352000)
  830. I0113 14:20:54.021173 85242 net.cpp:165] Memory required for data: 724358000
  831. I0113 14:20:54.021203 85242 layer_factory.hpp:77] Creating layer relu1
  832. I0113 14:20:54.021219 85242 net.cpp:106] Creating Layer relu1
  833. I0113 14:20:54.021226 85242 net.cpp:454] relu1 <- conv1
  834. I0113 14:20:54.021239 85242 net.cpp:397] relu1 -> conv1 (in-place)
  835. I0113 14:20:54.021386 85242 net.cpp:150] Setting up relu1
  836. I0113 14:20:54.021397 85242 net.cpp:157] Top shape: 500 64 56 56 (100352000)
  837. I0113 14:20:54.021402 85242 net.cpp:165] Memory required for data: 1125766000
  838. I0113 14:20:54.021407 85242 layer_factory.hpp:77] Creating layer pool1
  839. I0113 14:20:54.021428 85242 net.cpp:106] Creating Layer pool1
  840. I0113 14:20:54.021436 85242 net.cpp:454] pool1 <- conv1
  841. I0113 14:20:54.021452 85242 net.cpp:411] pool1 -> pool1
  842. I0113 14:20:54.021766 85242 net.cpp:150] Setting up pool1
  843. I0113 14:20:54.021780 85242 net.cpp:157] Top shape: 500 64 28 28 (25088000)
  844. I0113 14:20:54.021785 85242 net.cpp:165] Memory required for data: 1226118000
  845. I0113 14:20:54.021826 85242 layer_factory.hpp:77] Creating layer conv2
  846. I0113 14:20:54.021850 85242 net.cpp:106] Creating Layer conv2
  847. I0113 14:20:54.021858 85242 net.cpp:454] conv2 <- pool1
  848. I0113 14:20:54.021891 85242 net.cpp:411] conv2 -> conv2
  849. I0113 14:20:54.027290 85242 net.cpp:150] Setting up conv2
  850. I0113 14:20:54.027300 85242 net.cpp:157] Top shape: 500 128 26 26 (43264000)
  851. I0113 14:20:54.027305 85242 net.cpp:165] Memory required for data: 1399174000
  852. I0113 14:20:54.027326 85242 layer_factory.hpp:77] Creating layer relu2
  853. I0113 14:20:54.027341 85242 net.cpp:106] Creating Layer relu2
  854. I0113 14:20:54.027348 85242 net.cpp:454] relu2 <- conv2
  855. I0113 14:20:54.027364 85242 net.cpp:397] relu2 -> conv2 (in-place)
  856. I0113 14:20:54.027513 85242 net.cpp:150] Setting up relu2
  857. I0113 14:20:54.027524 85242 net.cpp:157] Top shape: 500 128 26 26 (43264000)
  858. I0113 14:20:54.027529 85242 net.cpp:165] Memory required for data: 1572230000
  859. I0113 14:20:54.027534 85242 layer_factory.hpp:77] Creating layer pool2
  860. I0113 14:20:54.027549 85242 net.cpp:106] Creating Layer pool2
  861. I0113 14:20:54.027556 85242 net.cpp:454] pool2 <- conv2
  862. I0113 14:20:54.027571 85242 net.cpp:411] pool2 -> pool2
  863. I0113 14:20:54.027881 85242 net.cpp:150] Setting up pool2
  864. I0113 14:20:54.027895 85242 net.cpp:157] Top shape: 500 128 13 13 (10816000)
  865. I0113 14:20:54.027900 85242 net.cpp:165] Memory required for data: 1615494000
  866. I0113 14:20:54.027906 85242 layer_factory.hpp:77] Creating layer conv3
  867. I0113 14:20:54.027931 85242 net.cpp:106] Creating Layer conv3
  868. I0113 14:20:54.027940 85242 net.cpp:454] conv3 <- pool2
  869. I0113 14:20:54.027961 85242 net.cpp:411] conv3 -> conv3
  870. I0113 14:20:54.045791 85242 net.cpp:150] Setting up conv3
  871. I0113 14:20:54.045800 85242 net.cpp:157] Top shape: 500 256 11 11 (15488000)
  872. I0113 14:20:54.045804 85242 net.cpp:165] Memory required for data: 1677446000
  873. I0113 14:20:54.045825 85242 layer_factory.hpp:77] Creating layer relu3
  874. I0113 14:20:54.045845 85242 net.cpp:106] Creating Layer relu3
  875. I0113 14:20:54.045852 85242 net.cpp:454] relu3 <- conv3
  876. I0113 14:20:54.045867 85242 net.cpp:397] relu3 -> conv3 (in-place)
  877. I0113 14:20:54.046027 85242 net.cpp:150] Setting up relu3
  878. I0113 14:20:54.046038 85242 net.cpp:157] Top shape: 500 256 11 11 (15488000)
  879. I0113 14:20:54.046041 85242 net.cpp:165] Memory required for data: 1739398000
  880. I0113 14:20:54.046046 85242 layer_factory.hpp:77] Creating layer pool3
  881. I0113 14:20:54.046059 85242 net.cpp:106] Creating Layer pool3
  882. I0113 14:20:54.046066 85242 net.cpp:454] pool3 <- conv3
  883. I0113 14:20:54.046082 85242 net.cpp:411] pool3 -> pool3
  884. I0113 14:20:54.046407 85242 net.cpp:150] Setting up pool3
  885. I0113 14:20:54.046422 85242 net.cpp:157] Top shape: 500 256 6 6 (4608000)
  886. I0113 14:20:54.046425 85242 net.cpp:165] Memory required for data: 1757830000
  887. I0113 14:20:54.046430 85242 layer_factory.hpp:77] Creating layer conv4
  888. I0113 14:20:54.046453 85242 net.cpp:106] Creating Layer conv4
  889. I0113 14:20:54.046463 85242 net.cpp:454] conv4 <- pool3
  890. I0113 14:20:54.046485 85242 net.cpp:411] conv4 -> conv4
  891. I0113 14:20:54.080294 85242 net.cpp:150] Setting up conv4
  892. I0113 14:20:54.080307 85242 net.cpp:157] Top shape: 500 256 4 4 (2048000)
  893. I0113 14:20:54.080309 85242 net.cpp:165] Memory required for data: 1766022000
  894. I0113 14:20:54.080327 85242 layer_factory.hpp:77] Creating layer relu4
  895. I0113 14:20:54.080349 85242 net.cpp:106] Creating Layer relu4
  896. I0113 14:20:54.080360 85242 net.cpp:454] relu4 <- conv4
  897. I0113 14:20:54.080379 85242 net.cpp:397] relu4 -> conv4 (in-place)
  898. I0113 14:20:54.080667 85242 net.cpp:150] Setting up relu4
  899. I0113 14:20:54.080678 85242 net.cpp:157] Top shape: 500 256 4 4 (2048000)
  900. I0113 14:20:54.080682 85242 net.cpp:165] Memory required for data: 1774214000
  901. I0113 14:20:54.080687 85242 layer_factory.hpp:77] Creating layer pool4
  902. I0113 14:20:54.080708 85242 net.cpp:106] Creating Layer pool4
  903. I0113 14:20:54.080742 85242 net.cpp:454] pool4 <- conv4
  904. I0113 14:20:54.080765 85242 net.cpp:411] pool4 -> pool4
  905. I0113 14:20:54.081223 85242 net.cpp:150] Setting up pool4
  906. I0113 14:20:54.081249 85242 net.cpp:157] Top shape: 500 256 2 2 (512000)
  907. I0113 14:20:54.081254 85242 net.cpp:165] Memory required for data: 1776262000
  908. I0113 14:20:54.081259 85242 layer_factory.hpp:77] Creating layer fc6
  909. I0113 14:20:54.081279 85242 net.cpp:106] Creating Layer fc6
  910. I0113 14:20:54.081286 85242 net.cpp:454] fc6 <- pool4
  911. I0113 14:20:54.081326 85242 net.cpp:411] fc6 -> fc6
  912. I0113 14:20:54.263216 85242 net.cpp:150] Setting up fc6
  913. I0113 14:20:54.263254 85242 net.cpp:157] Top shape: 500 2048 (1024000)
  914. I0113 14:20:54.263258 85242 net.cpp:165] Memory required for data: 1780358000
  915. I0113 14:20:54.263304 85242 layer_factory.hpp:77] Creating layer relu6
  916. I0113 14:20:54.263339 85242 net.cpp:106] Creating Layer relu6
  917. I0113 14:20:54.263353 85242 net.cpp:454] relu6 <- fc6
  918. I0113 14:20:54.263372 85242 net.cpp:397] relu6 -> fc6 (in-place)
  919. I0113 14:20:54.263597 85242 net.cpp:150] Setting up relu6
  920. I0113 14:20:54.263608 85242 net.cpp:157] Top shape: 500 2048 (1024000)
  921. I0113 14:20:54.263612 85242 net.cpp:165] Memory required for data: 1784454000
  922. I0113 14:20:54.263617 85242 layer_factory.hpp:77] Creating layer drop6
  923. I0113 14:20:54.263641 85242 net.cpp:106] Creating Layer drop6
  924. I0113 14:20:54.263650 85242 net.cpp:454] drop6 <- fc6
  925. I0113 14:20:54.263670 85242 net.cpp:397] drop6 -> fc6 (in-place)
  926. I0113 14:20:54.263713 85242 net.cpp:150] Setting up drop6
  927. I0113 14:20:54.263722 85242 net.cpp:157] Top shape: 500 2048 (1024000)
  928. I0113 14:20:54.263725 85242 net.cpp:165] Memory required for data: 1788550000
  929. I0113 14:20:54.263731 85242 layer_factory.hpp:77] Creating layer fc7
  930. I0113 14:20:54.263751 85242 net.cpp:106] Creating Layer fc7
  931. I0113 14:20:54.263758 85242 net.cpp:454] fc7 <- fc6
  932. I0113 14:20:54.263778 85242 net.cpp:411] fc7 -> fc7
  933. I0113 14:20:54.994442 85242 net.cpp:150] Setting up fc7
  934. I0113 14:20:54.994488 85242 net.cpp:157] Top shape: 500 4096 (2048000)
  935. I0113 14:20:54.994493 85242 net.cpp:165] Memory required for data: 1796742000
  936. I0113 14:20:54.994526 85242 layer_factory.hpp:77] Creating layer relu7
  937. I0113 14:20:54.994565 85242 net.cpp:106] Creating Layer relu7
  938. I0113 14:20:54.994581 85242 net.cpp:454] relu7 <- fc7
  939. I0113 14:20:54.994614 85242 net.cpp:397] relu7 -> fc7 (in-place)
  940. I0113 14:20:54.995162 85242 net.cpp:150] Setting up relu7
  941. I0113 14:20:54.995177 85242 net.cpp:157] Top shape: 500 4096 (2048000)
  942. I0113 14:20:54.995182 85242 net.cpp:165] Memory required for data: 1804934000
  943. I0113 14:20:54.995190 85242 layer_factory.hpp:77] Creating layer drop7
  944. I0113 14:20:54.995213 85242 net.cpp:106] Creating Layer drop7
  945. I0113 14:20:54.995223 85242 net.cpp:454] drop7 <- fc7
  946. I0113 14:20:54.995245 85242 net.cpp:397] drop7 -> fc7 (in-place)
  947. I0113 14:20:54.995296 85242 net.cpp:150] Setting up drop7
  948. I0113 14:20:54.995306 85242 net.cpp:157] Top shape: 500 4096 (2048000)
  949. I0113 14:20:54.995311 85242 net.cpp:165] Memory required for data: 1813126000
  950. I0113 14:20:54.995317 85242 layer_factory.hpp:77] Creating layer fc8
  951. I0113 14:20:54.995345 85242 net.cpp:106] Creating Layer fc8
  952. I0113 14:20:54.995357 85242 net.cpp:454] fc8 <- fc7
  953. I0113 14:20:54.995383 85242 net.cpp:411] fc8 -> fc8
  954. I0113 14:20:55.026443 85242 net.cpp:150] Setting up fc8
  955. I0113 14:20:55.026459 85242 net.cpp:157] Top shape: 500 101 (50500)
  956. I0113 14:20:55.026464 85242 net.cpp:165] Memory required for data: 1813328000
  957. I0113 14:20:55.026481 85242 layer_factory.hpp:77] Creating layer fc8_fc8_0_split
  958. I0113 14:20:55.026496 85242 net.cpp:106] Creating Layer fc8_fc8_0_split
  959. I0113 14:20:55.026506 85242 net.cpp:454] fc8_fc8_0_split <- fc8
  960. I0113 14:20:55.026526 85242 net.cpp:411] fc8_fc8_0_split -> fc8_fc8_0_split_0
  961. I0113 14:20:55.026559 85242 net.cpp:411] fc8_fc8_0_split -> fc8_fc8_0_split_1
  962. I0113 14:20:55.026643 85242 net.cpp:150] Setting up fc8_fc8_0_split
  963. I0113 14:20:55.026666 85242 net.cpp:157] Top shape: 500 101 (50500)
  964. I0113 14:20:55.026679 85242 net.cpp:157] Top shape: 500 101 (50500)
  965. I0113 14:20:55.026684 85242 net.cpp:165] Memory required for data: 1813732000
  966. I0113 14:20:55.026731 85242 layer_factory.hpp:77] Creating layer accuracy
  967. I0113 14:20:55.026760 85242 net.cpp:106] Creating Layer accuracy
  968. I0113 14:20:55.026782 85242 net.cpp:454] accuracy <- fc8_fc8_0_split_0
  969. I0113 14:20:55.026799 85242 net.cpp:454] accuracy <- label_data_1_split_0
  970. I0113 14:20:55.026815 85242 net.cpp:411] accuracy -> accuracy
  971. I0113 14:20:55.026870 85242 net.cpp:150] Setting up accuracy
  972. I0113 14:20:55.026881 85242 net.cpp:157] Top shape: (1)
  973. I0113 14:20:55.026887 85242 net.cpp:165] Memory required for data: 1813732004
  974. I0113 14:20:55.026906 85242 layer_factory.hpp:77] Creating layer loss
  975. I0113 14:20:55.026934 85242 net.cpp:106] Creating Layer loss
  976. I0113 14:20:55.026943 85242 net.cpp:454] loss <- fc8_fc8_0_split_1
  977. I0113 14:20:55.026959 85242 net.cpp:454] loss <- label_data_1_split_1
  978. I0113 14:20:55.026983 85242 net.cpp:411] loss -> loss
  979. I0113 14:20:55.027019 85242 layer_factory.hpp:77] Creating layer loss
  980. I0113 14:20:55.027576 85242 net.cpp:150] Setting up loss
  981. I0113 14:20:55.027591 85242 net.cpp:157] Top shape: (1)
  982. I0113 14:20:55.027597 85242 net.cpp:160] with loss weight 1
  983. I0113 14:20:55.027616 85242 net.cpp:165] Memory required for data: 1813732008
  984. I0113 14:20:55.027627 85242 net.cpp:226] loss needs backward computation.
  985. I0113 14:20:55.027640 85242 net.cpp:228] accuracy does not need backward computation.
  986. I0113 14:20:55.027649 85242 net.cpp:226] fc8_fc8_0_split needs backward computation.
  987. I0113 14:20:55.027657 85242 net.cpp:226] fc8 needs backward computation.
  988. I0113 14:20:55.027664 85242 net.cpp:226] drop7 needs backward computation.
  989. I0113 14:20:55.027672 85242 net.cpp:226] relu7 needs backward computation.
  990. I0113 14:20:55.027678 85242 net.cpp:226] fc7 needs backward computation.
  991. I0113 14:20:55.027685 85242 net.cpp:226] drop6 needs backward computation.
  992. I0113 14:20:55.027693 85242 net.cpp:226] relu6 needs backward computation.
  993. I0113 14:20:55.027699 85242 net.cpp:226] fc6 needs backward computation.
  994. I0113 14:20:55.027706 85242 net.cpp:226] pool4 needs backward computation.
  995. I0113 14:20:55.027714 85242 net.cpp:226] relu4 needs backward computation.
  996. I0113 14:20:55.027721 85242 net.cpp:226] conv4 needs backward computation.
  997. I0113 14:20:55.027729 85242 net.cpp:226] pool3 needs backward computation.
  998. I0113 14:20:55.027736 85242 net.cpp:226] relu3 needs backward computation.
  999. I0113 14:20:55.027743 85242 net.cpp:226] conv3 needs backward computation.
  1000. I0113 14:20:55.027750 85242 net.cpp:226] pool2 needs backward computation.
  1001. I0113 14:20:55.027758 85242 net.cpp:226] relu2 needs backward computation.
  1002. I0113 14:20:55.027765 85242 net.cpp:226] conv2 needs backward computation.
  1003. I0113 14:20:55.027772 85242 net.cpp:226] pool1 needs backward computation.
  1004. I0113 14:20:55.027779 85242 net.cpp:226] relu1 needs backward computation.
  1005. I0113 14:20:55.027786 85242 net.cpp:226] conv1 needs backward computation.
  1006. I0113 14:20:55.027796 85242 net.cpp:228] label_data_1_split does not need backward computation.
  1007. I0113 14:20:55.027804 85242 net.cpp:228] data does not need backward computation.
  1008. I0113 14:20:55.027812 85242 net.cpp:270] This network produces output accuracy
  1009. I0113 14:20:55.027822 85242 net.cpp:270] This network produces output loss
  1010. I0113 14:20:55.027869 85242 net.cpp:283] Network initialization done.
  1011. I0113 14:20:55.028033 85242 solver.cpp:60] Solver scaffolding done.
  1012. I0113 14:20:55.037928 85242 parallel.cpp:134] GPU board: 0:1
  1013. I0113 14:20:55.040472 85242 parallel.cpp:134] GPU board: 2:3
  1014. I0113 14:20:55.042877 85242 parallel.cpp:134] GPU board: 4:5
  1015. I0113 14:20:55.045142 85242 parallel.cpp:134] GPU board: 6:7
  1016. I0113 14:20:55.074800 85242 parallel.cpp:146] GPUs paired by boards, remaining: 0, 2, 4, 6
  1017. I0113 14:20:55.074898 85242 parallel.cpp:158] P2P pair: 0:2
  1018. I0113 14:20:55.074939 85242 parallel.cpp:158] P2P pair: 4:6
  1019. I0113 14:20:55.074967 85242 parallel.cpp:169] GPUs paired by P2P access, remaining: 0, 4
  1020. I0113 14:20:55.074978 85242 parallel.cpp:176] Remaining pair: 0:4
  1021. I0113 14:20:55.075013 85242 parallel.cpp:391] GPUs pairs 0:1, 2:3, 4:5, 6:7, 0:2, 4:6, 0:4
  1022. I0113 14:20:55.367068 85242 net.cpp:99] Sharing layer data from root net
  1023. I0113 14:20:55.367928 85242 net.cpp:143] Created top blob 0 (shape: 1000 48 58 58 (161472000)) for shared layer data
  1024. I0113 14:20:55.367974 85242 net.cpp:143] Created top blob 1 (shape: 1000 1 (1000)) for shared layer data
  1025. I0113 14:20:56.858564 85242 net.cpp:99] Sharing layer data from root net
  1026. I0113 14:20:56.860147 85242 net.cpp:143] Created top blob 0 (shape: 1000 48 58 58 (161472000)) for shared layer data
  1027. I0113 14:20:56.860211 85242 net.cpp:143] Created top blob 1 (shape: 1000 1 (1000)) for shared layer data
  1028. I0113 14:20:58.415379 85242 net.cpp:99] Sharing layer data from root net
  1029. I0113 14:20:58.416828 85242 net.cpp:143] Created top blob 0 (shape: 1000 48 58 58 (161472000)) for shared layer data
  1030. I0113 14:20:58.416924 85242 net.cpp:143] Created top blob 1 (shape: 1000 1 (1000)) for shared layer data
  1031. I0113 14:20:59.546455 85242 parallel.cpp:234] GPU 4 does not have p2p access to GPU 0
  1032. I0113 14:20:59.890323 85242 net.cpp:99] Sharing layer data from root net
  1033. I0113 14:20:59.891892 85242 net.cpp:143] Created top blob 0 (shape: 1000 48 58 58 (161472000)) for shared layer data
  1034. I0113 14:20:59.892002 85242 net.cpp:143] Created top blob 1 (shape: 1000 1 (1000)) for shared layer data
  1035. I0113 14:21:01.481375 85242 net.cpp:99] Sharing layer data from root net
  1036. I0113 14:21:01.482951 85242 net.cpp:143] Created top blob 0 (shape: 1000 48 58 58 (161472000)) for shared layer data
  1037. I0113 14:21:01.483057 85242 net.cpp:143] Created top blob 1 (shape: 1000 1 (1000)) for shared layer data
  1038. I0113 14:21:03.023627 85242 net.cpp:99] Sharing layer data from root net
  1039. I0113 14:21:03.025602 85242 net.cpp:143] Created top blob 0 (shape: 1000 48 58 58 (161472000)) for shared layer data
  1040. I0113 14:21:03.025787 85242 net.cpp:143] Created top blob 1 (shape: 1000 1 (1000)) for shared layer data
  1041. I0113 14:21:04.597707 85242 net.cpp:99] Sharing layer data from root net
  1042. I0113 14:21:04.600539 85242 net.cpp:143] Created top blob 0 (shape: 1000 48 58 58 (161472000)) for shared layer data
  1043. I0113 14:21:04.600782 85242 net.cpp:143] Created top blob 1 (shape: 1000 1 (1000)) for shared layer data
  1044. I0113 14:21:05.860424 85242 parallel.cpp:419] Starting Optimization
  1045. I0113 14:21:05.861312 85242 solver.cpp:288] Solving mv_16f
  1046. I0113 14:21:05.861361 85242 solver.cpp:289] Learning Rate Policy: step
  1047. I0113 14:21:05.862876 85242 solver.cpp:341] Iteration 0, Testing net (#0)
  1048. I0113 14:21:30.271030 85242 solver.cpp:409] Test net output #0: accuracy = 0.00872
  1049. I0113 14:21:30.271070 85242 solver.cpp:409] Test net output #1: loss = 4.62895 (* 1 = 4.62895 loss)
  1050. I0113 14:21:32.317018 85242 solver.cpp:237] Iteration 0, loss = 4.62663
  1051. I0113 14:21:32.317062 85242 solver.cpp:253] Train net output #0: loss = 4.62663 (* 1 = 4.62663 loss)
  1052. *** Aborted at 1452691298 (unix time) try "date -d @1452691298" if you are using GNU date ***
  1053. PC: @ 0x7fe7f65f1cbc caffe::SGDSolver<>::GetLearningRate()
  1054. *** SIGFPE (@0x7fe7f65f1cbc) received by PID 85242 (TID 0x7fe7f72057c0) from PID 18446744073548012732; stack trace: ***
  1055. @ 0x7fe7f49c0d40 (unknown)
  1056. @ 0x7fe7f65f1cbc caffe::SGDSolver<>::GetLearningRate()
  1057. @ 0x7fe7f65f2281 caffe::SGDSolver<>::ApplyUpdate()
  1058. @ 0x7fe7f65d967c caffe::Solver<>::Step()
  1059. @ 0x7fe7f65d8990 caffe::Solver<>::Solve()
  1060. @ 0x7fe7f673251e caffe::P2PSync<>::run()
  1061. @ 0x416aa6 train()
  1062. @ 0x418c9a main
  1063. @ 0x7fe7f49abec5 (unknown)
  1064. @ 0x415819 (unknown)
  1065. @ 0x0 (unknown)
Advertisement
Add Comment
Please, Sign In to add comment
Advertisement