Advertisement
Guest User

Untitled

a guest
Aug 22nd, 2017
91
0
Never
Not a member of Pastebin yet? Sign Up, it unlocks many cool features!
text 8.84 KB | None | 0 0
  1. layer {
  2. name: "data"
  3. type: "Input"
  4. top: "data"
  5. input_param { shape: { dim: dim: dim: dim: } }
  6. }
  7.  
  8. layer {
  9. name: "conv1"
  10. type: "Convolution"
  11. bottom: "data"
  12. top: "conv1"
  13. convolution_param {
  14. num_output:
  15. kernel_size:
  16. stride:
  17. pad:
  18. weight_filler {
  19. type: "xavier"
  20. }
  21. }
  22. }
  23. layer {
  24. name: "bn_conv1"
  25. type: "BatchNorm"
  26. bottom: "conv1"
  27. top: "conv1"
  28. batch_norm_param {
  29. use_global_stats: true
  30. }
  31. include {
  32. phase: TEST
  33. }
  34. }
  35. layer {
  36. name: "scale_conv1"
  37. type: "Scale"
  38. bottom: "conv1"
  39. top: "conv1"
  40. scale_param {
  41. bias_term: true
  42. }
  43. }
  44. layer {
  45. name: "relu_conv1"
  46. type: "ReLU"
  47. bottom: "conv1"
  48. top: "conv1"
  49. }
  50. layer {
  51. name: "conv2"
  52. type: "Convolution"
  53. bottom: "conv1"
  54. top: "conv2"
  55. convolution_param {
  56. num_output:
  57. kernel_size:
  58. pad:
  59. weight_filler {
  60. type: "xavier"
  61. }
  62. }
  63. }
  64. layer {
  65. name: "bn_conv2"
  66. type: "BatchNorm"
  67. bottom: "conv2"
  68. top: "conv2"
  69. batch_norm_param {
  70. use_global_stats: true
  71. }
  72. include {
  73. phase: TEST
  74. }
  75. }
  76. layer {
  77. name: "scale_conv2"
  78. type: "Scale"
  79. bottom: "conv2"
  80. top: "conv2"
  81. scale_param {
  82. bias_term: true
  83. }
  84. }
  85. layer {
  86. name: "relu_conv2"
  87. type: "ReLU"
  88. bottom: "conv2"
  89. top: "conv2"
  90. }
  91. layer {
  92. name: "pool2"
  93. type: "Pooling"
  94. bottom: "conv2"
  95. top: "pool2"
  96. pooling_param {
  97. pool: MAX
  98. kernel_size:
  99. stride:
  100. }
  101. }
  102. layer {
  103. name: "conv3"
  104. type: "Convolution"
  105. bottom: "pool2"
  106. top: "conv3"
  107. convolution_param {
  108. num_output:
  109. kernel_size:
  110. pad:
  111. weight_filler {
  112. type: "xavier"
  113. }
  114. }
  115. }
  116. layer {
  117. name: "bn_conv3"
  118. type: "BatchNorm"
  119. bottom: "conv3"
  120. top: "conv3"
  121. batch_norm_param {
  122. use_global_stats: true
  123. }
  124. include {
  125. phase: TEST
  126. }
  127. }
  128. layer {
  129. name: "scale_conv3"
  130. type: "Scale"
  131. bottom: "conv3"
  132. top: "conv3"
  133. scale_param {
  134. bias_term: true
  135. }
  136. }
  137. layer {
  138. name: "relu_conv3"
  139. type: "ReLU"
  140. bottom: "conv3"
  141. top: "conv3"
  142. }
  143. layer {
  144. name: "conv4"
  145. type: "Convolution"
  146. bottom: "conv3"
  147. top: "conv4"
  148. convolution_param {
  149. num_output:
  150. kernel_size:
  151. pad:
  152. weight_filler {
  153. type: "xavier"
  154. }
  155. }
  156. }
  157. layer {
  158. name: "bn_conv4"
  159. type: "BatchNorm"
  160. bottom: "conv4"
  161. top: "conv4"
  162. batch_norm_param {
  163. use_global_stats: true
  164. }
  165. include {
  166. phase: TEST
  167. }
  168. }
  169. layer {
  170. name: "scale_conv4"
  171. type: "Scale"
  172. bottom: "conv4"
  173. top: "conv4"
  174. scale_param {
  175. bias_term: true
  176. }
  177. }
  178. layer {
  179. name: "relu_conv4"
  180. type: "ReLU"
  181. bottom: "conv4"
  182. top: "conv4"
  183. }
  184. layer {
  185. name: "conv5"
  186. type: "Convolution"
  187. bottom: "conv4"
  188. top: "conv5"
  189. convolution_param {
  190. num_output:
  191. kernel_size: 3
  192. pad: 1
  193. weight_filler {
  194. type: "xavier"
  195. }
  196. }
  197. }
  198. layer {
  199. name: "bn_conv5"
  200. type: "BatchNorm"
  201. bottom: "conv5"
  202. top: "conv5"
  203. batch_norm_param {
  204. use_global_stats: true
  205. }
  206. include {
  207. phase: TEST
  208. }
  209. }
  210. layer {
  211. name: "scale_conv5"
  212. type: "Scale"
  213. bottom: "conv5"
  214. top: "conv5"
  215. scale_param {
  216. bias_term: true
  217. }
  218. }
  219. layer {
  220. name: "relu_conv5"
  221. type: "ReLU"
  222. bottom: "conv5"
  223. top: "conv5"
  224. }
  225. layer {
  226. name: "pool5"
  227. type: "Pooling"
  228. bottom: "conv5"
  229. top: "pool5"
  230. pooling_param {
  231. pool: MAX
  232. kernel_size:
  233. stride:
  234. }
  235. }
  236. layer {
  237. name: "conv6"
  238. type: "Convolution"
  239. bottom: "pool5"
  240. top: "conv6"
  241. convolution_param {
  242. num_output:
  243. kernel_size:
  244. pad:
  245. weight_filler {
  246. type: "xavier"
  247. }
  248. }
  249. }
  250. layer {
  251. name: "bn_conv6"
  252. type: "BatchNorm"
  253. bottom: "conv6"
  254. top: "conv6"
  255. batch_norm_param {
  256. use_global_stats: true
  257. }
  258. include {
  259. phase: TEST
  260. }
  261. }
  262. layer {
  263. name: "scale_conv6"
  264. type: "Scale"
  265. bottom: "conv6"
  266. top: "conv6"
  267. scale_param {
  268. bias_term: true
  269. }
  270. }
  271. layer {
  272. name: "relu_conv6"
  273. type: "ReLU"
  274. bottom: "conv6"
  275. top: "conv6"
  276. }
  277. layer {
  278. name: "conv7"
  279. type: "Convolution"
  280. bottom: "conv6"
  281. top: "conv7"
  282. convolution_param {
  283. num_output:
  284. kernel_size:
  285. pad:
  286. weight_filler {
  287. type: "xavier"
  288. }
  289. }
  290. }
  291.  
  292. layer {
  293. name: "bn_conv7"
  294. type: "BatchNorm"
  295. bottom: "conv7"
  296. top: "conv7"
  297. batch_norm_param {
  298. use_global_stats: true
  299. }
  300. include {
  301. phase: TEST
  302. }
  303. }
  304. layer {
  305. name: "scale_conv7"
  306. type: "Scale"
  307. bottom: "conv7"
  308. top: "conv7"
  309. scale_param {
  310. bias_term: true
  311. }
  312. }
  313. layer {
  314. name: "relu_con7"
  315. type: "ReLU"
  316. bottom: "conv7"
  317. top: "conv7"
  318. }
  319. layer {
  320. name: "conv8"
  321. type: "Convolution"
  322. bottom: "conv7"
  323. top: "conv8"
  324. convolution_param {
  325. num_output:
  326. kernel_size:
  327. pad:
  328. weight_filler {
  329. type: "xavier"
  330. }
  331. }
  332. }
  333. layer {
  334. name: "bn_conv8"
  335. type: "BatchNorm"
  336. bottom: "conv8"
  337. top: "conv8"
  338. batch_norm_param {
  339. use_global_stats: true
  340. }
  341. include {
  342. phase: TEST
  343. }
  344. }
  345. layer {
  346. name: "scale_conv8"
  347. type: "Scale"
  348. bottom: "conv8"
  349. top: "conv8"
  350. scale_param {
  351. bias_term: true
  352. }
  353. }
  354. layer {
  355. name: "relu_conv8"
  356. type: "ReLU"
  357. bottom: "conv8"
  358. top: "conv8"
  359. }
  360. layer {
  361. name: "pool8"
  362. type: "Pooling"
  363. bottom: "conv8"
  364. top: "pool8"
  365. pooling_param {
  366. pool: MAX
  367. kernel_size:
  368. stride:
  369. }
  370. }
  371. layer {
  372. name: "conv9"
  373. type: "Convolution"
  374. bottom: "pool8"
  375. top: "conv9"
  376. convolution_param {
  377. num_output:
  378. kernel_size:
  379. pad:
  380. weight_filler {
  381. type: "xavier"
  382. }
  383. }
  384. }
  385. layer {
  386. name: "bn_conv9"
  387. type: "BatchNorm"
  388. bottom: "conv9"
  389. top: "conv9"
  390. batch_norm_param {
  391. use_global_stats: true
  392. }
  393. include {
  394. phase: TEST
  395. }
  396. }
  397. layer {
  398. name: "scale_conv9"
  399. type: "Scale"
  400. bottom: "conv9"
  401. top: "conv9"
  402. scale_param {
  403. bias_term: true
  404. }
  405. }
  406. layer {
  407. name: "relu_conv9"
  408. type: "ReLU"
  409. bottom: "conv9"
  410. top: "conv9"
  411. }
  412. layer {
  413. name: "conv10"
  414. type: "Convolution"
  415. bottom: "conv9"
  416. top: "conv10"
  417. convolution_param {
  418. num_output:
  419. kernel_size:
  420. pad: 1
  421. weight_filler {
  422. type: "xavier"
  423. }
  424. }
  425. }
  426. layer {
  427. name: "bn_conv10"
  428. type: "BatchNorm"
  429. bottom: "conv10"
  430. top: "conv10"
  431. batch_norm_param {
  432. use_global_stats: true
  433. }
  434. include {
  435. phase: TEST
  436. }
  437. }
  438. layer {
  439. name: "scale_conv10"
  440. type: "Scale"
  441. bottom: "conv10"
  442. top: "conv10"
  443. scale_param {
  444. bias_term: true
  445. }
  446. }
  447. layer {
  448. name: "relu_con10"
  449. type: "ReLU"
  450. bottom: "conv10"
  451. top: "conv10"
  452. }
  453. layer {
  454. name: "conv11"
  455. type: "Convolution"
  456. bottom: "conv10"
  457. top: "conv11"
  458. convolution_param {
  459. num_output:
  460. kernel_size:
  461. pad:
  462. weight_filler {
  463. type: "xavier"
  464. }
  465. }
  466. }
  467. layer {
  468. name: "bn_conv11"
  469. type: "BatchNorm"
  470. bottom: "conv11"
  471. top: "conv11"
  472. batch_norm_param {
  473. use_global_stats: true
  474. }
  475. include {
  476. phase: TEST
  477. }
  478. }
  479. layer {
  480. name: "scale_conv11"
  481. type: "Scale"
  482. bottom: "conv11"
  483. top: "conv11"
  484. scale_param {
  485. bias_term: true
  486. }
  487. }
  488. layer {
  489. name: "relu_conv11"
  490. type: "ReLU"
  491. bottom: "conv11"
  492. top: "conv11"
  493. }
  494. layer {
  495. name: "pool11"
  496. type: "Pooling"
  497. bottom: "conv11"
  498. top: "pool11"
  499. pooling_param {
  500. pool: MAX
  501. kernel_size:
  502. stride:
  503. }
  504. }
  505. layer {
  506. name: "conv12"
  507. type: "Convolution"
  508. bottom: "pool11"
  509. top: "conv12"
  510. convolution_param {
  511. num_output:
  512. kernel_size:
  513. pad:
  514. weight_filler {
  515. type: "xavier"
  516. }
  517. }
  518. }
  519. layer {
  520. name: "bn_conv12"
  521. type: "BatchNorm"
  522. bottom: "conv12"
  523. top: "conv12"
  524. batch_norm_param {
  525. use_global_stats: true
  526. }
  527. include {
  528. phase: TEST
  529. }
  530. }
  531. layer {
  532. name: "scale_conv12"
  533. type: "Scale"
  534. bottom: "conv12"
  535. top: "conv12"
  536. scale_param {
  537. bias_term: true
  538. }
  539. }
  540. layer {
  541. name: "relu_conv12"
  542. type: "ReLU"
  543. bottom: "conv12"
  544. top: "conv12"
  545. }
  546. layer {
  547. name: "conv13"
  548. type: "Convolution"
  549. bottom: "conv12"
  550. top: "conv13"
  551. convolution_param {
  552. num_output:
  553. kernel_size:
  554. pad:
  555. weight_filler {
  556. type: "xavier"
  557. }
  558. }
  559. }
  560. layer {
  561. name: "bn_conv13"
  562. type: "BatchNorm"
  563. bottom: "conv13"
  564. top: "conv13"
  565. batch_norm_param {
  566. use_global_stats: true
  567. }
  568. include {
  569. phase: TEST
  570. }
  571. }
  572. layer {
  573. name: "scale_conv13"
  574. type: "Scale"
  575. bottom: "conv13"
  576. top: "conv13"
  577. scale_param {
  578. bias_term: true
  579. }
  580. }
  581. layer {
  582. name: "relu_con13"
  583. type: "ReLU"
  584. bottom: "conv13"
  585. top: "conv13"
  586. }
  587. layer {
  588. name: "conv14"
  589. type: "Convolution"
  590. bottom: "conv13"
  591. top: "conv14"
  592. convolution_param {
  593. num_output:
  594. kernel_size:
  595. pad:
  596. weight_filler {
  597. type: "xavier"
  598. }
  599. }
  600. }
  601. layer {
  602. name: "bn_conv14"
  603. type: "BatchNorm"
  604. bottom: "conv14"
  605. top: "conv14"
  606. batch_norm_param {
  607. use_global_stats: true
  608. }
  609. include {
  610. phase: TEST
  611. }
  612. }
  613. layer {
  614. name: "scale_conv14"
  615. type: "Scale"
  616. bottom: "conv14"
  617. top: "conv14"
  618. scale_param {
  619. bias_term: true
  620. }
  621. }
  622. layer {
  623. name: "relu_conv14"
  624. type: "ReLU"
  625. bottom: "conv14"
  626. top: "conv14"
  627. }
  628.  
  629. layer {
  630. name: "conv_final"
  631. type: "Convolution"
  632. bottom: "conv14"
  633. top: "conv_final"
  634. convolution_param {
  635. num_output:
  636. kernel_size:
  637. weight_filler {
  638. type: "gaussian"
  639. mean:
  640. std:
  641. }
  642. }
  643. }
  644. layer {
  645. name: "relu_conv_final"
  646. type: "ReLU"
  647. bottom: "conv_final"
  648. top: "conv_final"
  649. }
  650. layer {
  651. name: "pool_final"
  652. type: "Pooling"
  653. bottom: "conv_final"
  654. top: "pool_final"
  655. pooling_param {
  656. pool: AVE
  657. global_pooling: true
  658. }
  659. }
Advertisement
Add Comment
Please, Sign In to add comment
Advertisement